01/27/2022 18:45:28 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 01/27/2022 18:45:29 - WARNING - huggingface_hub.repository - Revision `colorful-plasma-1` does not exist. Created and checked out branch `colorful-plasma-1`. 01/27/2022 18:45:29 - WARNING - huggingface_hub.repository - 01/27/2022 18:45:44 - WARNING - datasets.builder - Using custom data configuration lvwerra___codeparrot-clean-train-a1efdd1059bd841d 01/27/2022 18:45:45 - WARNING - datasets.builder - Using custom data configuration lvwerra___codeparrot-clean-valid-a800eb55c299abc0 01/27/2022 18:46:29 - INFO - codeparrot_training - Step 0: {'lr': 0.0, 'samples': 192, 'steps': 0, 'loss/train': 12.299906730651855} 01/27/2022 18:47:44 - INFO - codeparrot_training - Step 1: {'lr': 2.5e-07, 'samples': 384, 'steps': 1, 'loss/train': 12.254416465759277} 01/27/2022 18:48:57 - INFO - codeparrot_training - Step 2: {'lr': 5e-07, 'samples': 576, 'steps': 2, 'loss/train': 12.296040058135986} 01/27/2022 18:49:02 - INFO - codeparrot_training - Step 3: {'lr': 7.5e-07, 'samples': 768, 'steps': 3, 'loss/train': 12.258198738098145} 01/27/2022 18:49:08 - INFO - codeparrot_training - Step 4: {'lr': 1e-06, 'samples': 960, 'steps': 4, 'loss/train': 12.26336145401001} 01/27/2022 18:49:12 - INFO - codeparrot_training - Step 5: {'lr': 1.25e-06, 'samples': 1152, 'steps': 5, 'loss/train': 12.287298202514648} 01/27/2022 18:49:17 - INFO - codeparrot_training - Step 6: {'lr': 1.5e-06, 'samples': 1344, 'steps': 6, 'loss/train': 12.187986373901367} 01/27/2022 18:49:21 - INFO - codeparrot_training - Step 7: {'lr': 1.75e-06, 'samples': 1536, 'steps': 7, 'loss/train': 12.255367755889893} 01/27/2022 18:49:25 - INFO - codeparrot_training - Step 8: {'lr': 2e-06, 'samples': 1728, 'steps': 8, 'loss/train': 12.240334510803223} 01/27/2022 18:49:30 - INFO - codeparrot_training - Step 9: {'lr': 2.25e-06, 'samples': 1920, 'steps': 9, 'loss/train': 12.035845756530762} 01/27/2022 18:49:34 - INFO - codeparrot_training - Step 10: {'lr': 2.5e-06, 'samples': 2112, 'steps': 10, 'loss/train': 12.013946056365967} 01/27/2022 18:49:38 - INFO - codeparrot_training - Step 11: {'lr': 2.75e-06, 'samples': 2304, 'steps': 11, 'loss/train': 12.014110565185547} 01/27/2022 18:49:43 - INFO - codeparrot_training - Step 12: {'lr': 3e-06, 'samples': 2496, 'steps': 12, 'loss/train': 11.882771015167236} 01/27/2022 18:49:47 - INFO - codeparrot_training - Step 13: {'lr': 3.25e-06, 'samples': 2688, 'steps': 13, 'loss/train': 11.914687871932983} 01/27/2022 18:49:53 - INFO - codeparrot_training - Step 14: {'lr': 3.5e-06, 'samples': 2880, 'steps': 14, 'loss/train': 11.83001446723938} 01/27/2022 18:49:57 - INFO - codeparrot_training - Step 15: {'lr': 3.75e-06, 'samples': 3072, 'steps': 15, 'loss/train': 11.810596704483032} 01/27/2022 18:50:01 - INFO - codeparrot_training - Step 16: {'lr': 4e-06, 'samples': 3264, 'steps': 16, 'loss/train': 11.734410524368286} 01/27/2022 18:50:05 - INFO - codeparrot_training - Step 17: {'lr': 4.250000000000001e-06, 'samples': 3456, 'steps': 17, 'loss/train': 11.65330696105957} 01/27/2022 18:50:10 - INFO - codeparrot_training - Step 18: {'lr': 4.5e-06, 'samples': 3648, 'steps': 18, 'loss/train': 11.715793132781982} 01/27/2022 18:50:15 - INFO - codeparrot_training - Step 19: {'lr': 4.75e-06, 'samples': 3840, 'steps': 19, 'loss/train': 11.387081623077393} 01/27/2022 18:50:19 - INFO - codeparrot_training - Step 20: {'lr': 5e-06, 'samples': 4032, 'steps': 20, 'loss/train': 11.467334747314453} 01/27/2022 18:50:23 - INFO - codeparrot_training - Step 21: {'lr': 5.2500000000000006e-06, 'samples': 4224, 'steps': 21, 'loss/train': 11.103084325790405} 01/27/2022 18:50:27 - INFO - codeparrot_training - Step 22: {'lr': 5.5e-06, 'samples': 4416, 'steps': 22, 'loss/train': 11.488559246063232} 01/27/2022 18:50:31 - INFO - codeparrot_training - Step 23: {'lr': 5.75e-06, 'samples': 4608, 'steps': 23, 'loss/train': 11.6660635471344} 01/27/2022 18:50:37 - INFO - codeparrot_training - Step 24: {'lr': 6e-06, 'samples': 4800, 'steps': 24, 'loss/train': 10.99052095413208} 01/27/2022 18:50:41 - INFO - codeparrot_training - Step 25: {'lr': 6.25e-06, 'samples': 4992, 'steps': 25, 'loss/train': 11.491893768310547} 01/27/2022 18:50:45 - INFO - codeparrot_training - Step 26: {'lr': 6.5e-06, 'samples': 5184, 'steps': 26, 'loss/train': 11.300984144210815} 01/27/2022 18:50:49 - INFO - codeparrot_training - Step 27: {'lr': 6.75e-06, 'samples': 5376, 'steps': 27, 'loss/train': 11.552963733673096} 01/27/2022 18:50:53 - INFO - codeparrot_training - Step 28: {'lr': 7e-06, 'samples': 5568, 'steps': 28, 'loss/train': 10.592391729354858} 01/27/2022 18:51:00 - INFO - codeparrot_training - Step 29: {'lr': 7.250000000000001e-06, 'samples': 5760, 'steps': 29, 'loss/train': 10.966165781021118} 01/27/2022 18:51:04 - INFO - codeparrot_training - Step 30: {'lr': 7.5e-06, 'samples': 5952, 'steps': 30, 'loss/train': 11.093848943710327} 01/27/2022 18:51:08 - INFO - codeparrot_training - Step 31: {'lr': 7.75e-06, 'samples': 6144, 'steps': 31, 'loss/train': 11.246557474136353} 01/27/2022 18:51:12 - INFO - codeparrot_training - Step 32: {'lr': 8e-06, 'samples': 6336, 'steps': 32, 'loss/train': 10.572290897369385} 01/27/2022 18:51:16 - INFO - codeparrot_training - Step 33: {'lr': 8.25e-06, 'samples': 6528, 'steps': 33, 'loss/train': 10.75559949874878} 01/27/2022 18:51:22 - INFO - codeparrot_training - Step 34: {'lr': 8.500000000000002e-06, 'samples': 6720, 'steps': 34, 'loss/train': 11.311209440231323} 01/27/2022 18:51:26 - INFO - codeparrot_training - Step 35: {'lr': 8.750000000000001e-06, 'samples': 6912, 'steps': 35, 'loss/train': 10.82191014289856} 01/27/2022 18:51:30 - INFO - codeparrot_training - Step 36: {'lr': 9e-06, 'samples': 7104, 'steps': 36, 'loss/train': 11.219964981079102} 01/27/2022 18:51:34 - INFO - codeparrot_training - Step 37: {'lr': 9.25e-06, 'samples': 7296, 'steps': 37, 'loss/train': 11.145844459533691} 01/27/2022 18:51:38 - INFO - codeparrot_training - Step 38: {'lr': 9.5e-06, 'samples': 7488, 'steps': 38, 'loss/train': 10.999979496002197} 01/27/2022 18:51:43 - INFO - codeparrot_training - Step 39: {'lr': 9.75e-06, 'samples': 7680, 'steps': 39, 'loss/train': 9.97054123878479} 01/27/2022 18:51:47 - INFO - codeparrot_training - Step 40: {'lr': 1e-05, 'samples': 7872, 'steps': 40, 'loss/train': 11.009016752243042} 01/27/2022 18:51:52 - INFO - codeparrot_training - Step 41: {'lr': 1.025e-05, 'samples': 8064, 'steps': 41, 'loss/train': 10.472270250320435} 01/27/2022 18:51:56 - INFO - codeparrot_training - Step 42: {'lr': 1.0500000000000001e-05, 'samples': 8256, 'steps': 42, 'loss/train': 10.424858093261719} 01/27/2022 18:52:00 - INFO - codeparrot_training - Step 43: {'lr': 1.0749999999999999e-05, 'samples': 8448, 'steps': 43, 'loss/train': 10.910295009613037} 01/27/2022 18:52:06 - INFO - codeparrot_training - Step 44: {'lr': 1.1e-05, 'samples': 8640, 'steps': 44, 'loss/train': 10.834372758865356} 01/27/2022 18:52:10 - INFO - codeparrot_training - Step 45: {'lr': 1.1249999999999999e-05, 'samples': 8832, 'steps': 45, 'loss/train': 11.367445707321167} 01/27/2022 18:52:15 - INFO - codeparrot_training - Step 46: {'lr': 1.15e-05, 'samples': 9024, 'steps': 46, 'loss/train': 10.566834926605225} 01/27/2022 18:52:19 - INFO - codeparrot_training - Step 47: {'lr': 1.1750000000000001e-05, 'samples': 9216, 'steps': 47, 'loss/train': 10.700668573379517} 01/27/2022 18:52:23 - INFO - codeparrot_training - Step 48: {'lr': 1.2e-05, 'samples': 9408, 'steps': 48, 'loss/train': 11.141412734985352} 01/27/2022 18:52:28 - INFO - codeparrot_training - Step 49: {'lr': 1.2250000000000001e-05, 'samples': 9600, 'steps': 49, 'loss/train': 11.026137113571167} 01/27/2022 18:52:32 - INFO - codeparrot_training - Step 50: {'lr': 1.25e-05, 'samples': 9792, 'steps': 50, 'loss/train': 11.223975419998169} 01/27/2022 18:52:36 - INFO - codeparrot_training - Step 51: {'lr': 1.275e-05, 'samples': 9984, 'steps': 51, 'loss/train': 10.629223108291626} 01/27/2022 18:52:40 - INFO - codeparrot_training - Step 52: {'lr': 1.3e-05, 'samples': 10176, 'steps': 52, 'loss/train': 10.660104274749756} 01/27/2022 18:52:45 - INFO - codeparrot_training - Step 53: {'lr': 1.325e-05, 'samples': 10368, 'steps': 53, 'loss/train': 10.400070190429688} 01/27/2022 18:52:50 - INFO - codeparrot_training - Step 54: {'lr': 1.35e-05, 'samples': 10560, 'steps': 54, 'loss/train': 11.049419403076172} 01/27/2022 18:52:54 - INFO - codeparrot_training - Step 55: {'lr': 1.375e-05, 'samples': 10752, 'steps': 55, 'loss/train': 10.336088418960571} 01/27/2022 18:52:58 - INFO - codeparrot_training - Step 56: {'lr': 1.4e-05, 'samples': 10944, 'steps': 56, 'loss/train': 11.02346920967102} 01/27/2022 18:53:02 - INFO - codeparrot_training - Step 57: {'lr': 1.425e-05, 'samples': 11136, 'steps': 57, 'loss/train': 10.886565685272217} 01/27/2022 18:53:06 - INFO - codeparrot_training - Step 58: {'lr': 1.4500000000000002e-05, 'samples': 11328, 'steps': 58, 'loss/train': 10.751394510269165} 01/27/2022 18:53:11 - INFO - codeparrot_training - Step 59: {'lr': 1.475e-05, 'samples': 11520, 'steps': 59, 'loss/train': 10.39724063873291} 01/27/2022 18:53:15 - INFO - codeparrot_training - Step 60: {'lr': 1.5e-05, 'samples': 11712, 'steps': 60, 'loss/train': 10.75915789604187} 01/27/2022 18:53:20 - INFO - codeparrot_training - Step 61: {'lr': 1.525e-05, 'samples': 11904, 'steps': 61, 'loss/train': 10.844947814941406} 01/27/2022 18:53:24 - INFO - codeparrot_training - Step 62: {'lr': 1.55e-05, 'samples': 12096, 'steps': 62, 'loss/train': 9.661353349685669} 01/27/2022 18:53:28 - INFO - codeparrot_training - Step 63: {'lr': 1.575e-05, 'samples': 12288, 'steps': 63, 'loss/train': 10.0857253074646} 01/27/2022 18:53:34 - INFO - codeparrot_training - Step 64: {'lr': 1.6e-05, 'samples': 12480, 'steps': 64, 'loss/train': 10.630635738372803} 01/27/2022 18:53:39 - INFO - codeparrot_training - Step 65: {'lr': 1.6250000000000002e-05, 'samples': 12672, 'steps': 65, 'loss/train': 11.000107526779175} 01/27/2022 18:53:43 - INFO - codeparrot_training - Step 66: {'lr': 1.65e-05, 'samples': 12864, 'steps': 66, 'loss/train': 10.835297584533691} 01/27/2022 18:53:47 - INFO - codeparrot_training - Step 67: {'lr': 1.675e-05, 'samples': 13056, 'steps': 67, 'loss/train': 10.615701913833618} 01/27/2022 18:53:51 - INFO - codeparrot_training - Step 68: {'lr': 1.7000000000000003e-05, 'samples': 13248, 'steps': 68, 'loss/train': 11.006195783615112} 01/27/2022 18:53:56 - INFO - codeparrot_training - Step 69: {'lr': 1.7250000000000003e-05, 'samples': 13440, 'steps': 69, 'loss/train': 11.055887460708618} 01/27/2022 18:54:00 - INFO - codeparrot_training - Step 70: {'lr': 1.7500000000000002e-05, 'samples': 13632, 'steps': 70, 'loss/train': 11.058547496795654} 01/27/2022 18:54:04 - INFO - codeparrot_training - Step 71: {'lr': 1.7749999999999998e-05, 'samples': 13824, 'steps': 71, 'loss/train': 11.03822636604309} 01/27/2022 18:54:09 - INFO - codeparrot_training - Step 72: {'lr': 1.8e-05, 'samples': 14016, 'steps': 72, 'loss/train': 10.50807809829712} 01/27/2022 18:54:13 - INFO - codeparrot_training - Step 73: {'lr': 1.825e-05, 'samples': 14208, 'steps': 73, 'loss/train': 10.766568660736084} 01/27/2022 18:54:19 - INFO - codeparrot_training - Step 74: {'lr': 1.85e-05, 'samples': 14400, 'steps': 74, 'loss/train': 10.553864479064941} 01/27/2022 18:54:23 - INFO - codeparrot_training - Step 75: {'lr': 1.875e-05, 'samples': 14592, 'steps': 75, 'loss/train': 10.21326470375061} 01/27/2022 18:54:27 - INFO - codeparrot_training - Step 76: {'lr': 1.9e-05, 'samples': 14784, 'steps': 76, 'loss/train': 10.913193941116333} 01/27/2022 18:54:32 - INFO - codeparrot_training - Step 77: {'lr': 1.925e-05, 'samples': 14976, 'steps': 77, 'loss/train': 11.222438335418701} 01/27/2022 18:54:37 - INFO - codeparrot_training - Step 78: {'lr': 1.95e-05, 'samples': 15168, 'steps': 78, 'loss/train': 11.160276889801025} 01/27/2022 18:54:41 - INFO - codeparrot_training - Step 79: {'lr': 1.975e-05, 'samples': 15360, 'steps': 79, 'loss/train': 10.69108772277832} 01/27/2022 18:54:45 - INFO - codeparrot_training - Step 80: {'lr': 2e-05, 'samples': 15552, 'steps': 80, 'loss/train': 11.075340270996094} 01/27/2022 18:54:49 - INFO - codeparrot_training - Step 81: {'lr': 2.025e-05, 'samples': 15744, 'steps': 81, 'loss/train': 10.32427453994751} 01/27/2022 18:54:53 - INFO - codeparrot_training - Step 82: {'lr': 2.05e-05, 'samples': 15936, 'steps': 82, 'loss/train': 11.106040477752686} 01/27/2022 18:54:57 - INFO - codeparrot_training - Step 83: {'lr': 2.0750000000000003e-05, 'samples': 16128, 'steps': 83, 'loss/train': 10.27891731262207} 01/27/2022 18:55:03 - INFO - codeparrot_training - Step 84: {'lr': 2.1000000000000002e-05, 'samples': 16320, 'steps': 84, 'loss/train': 10.17206597328186} 01/27/2022 18:55:07 - INFO - codeparrot_training - Step 85: {'lr': 2.125e-05, 'samples': 16512, 'steps': 85, 'loss/train': 10.83674669265747} 01/27/2022 18:55:11 - INFO - codeparrot_training - Step 86: {'lr': 2.1499999999999997e-05, 'samples': 16704, 'steps': 86, 'loss/train': 10.95323395729065} 01/27/2022 18:55:15 - INFO - codeparrot_training - Step 87: {'lr': 2.175e-05, 'samples': 16896, 'steps': 87, 'loss/train': 10.6928071975708} 01/27/2022 18:55:19 - INFO - codeparrot_training - Step 88: {'lr': 2.2e-05, 'samples': 17088, 'steps': 88, 'loss/train': 10.250412225723267} 01/27/2022 18:55:26 - INFO - codeparrot_training - Step 89: {'lr': 2.225e-05, 'samples': 17280, 'steps': 89, 'loss/train': 11.357476472854614} 01/27/2022 18:55:30 - INFO - codeparrot_training - Step 90: {'lr': 2.2499999999999998e-05, 'samples': 17472, 'steps': 90, 'loss/train': 10.97425103187561} 01/27/2022 18:55:34 - INFO - codeparrot_training - Step 91: {'lr': 2.275e-05, 'samples': 17664, 'steps': 91, 'loss/train': 10.546386480331421} 01/27/2022 18:55:38 - INFO - codeparrot_training - Step 92: {'lr': 2.3e-05, 'samples': 17856, 'steps': 92, 'loss/train': 10.648743152618408} 01/27/2022 18:55:43 - INFO - codeparrot_training - Step 93: {'lr': 2.325e-05, 'samples': 18048, 'steps': 93, 'loss/train': 10.244426965713501} 01/27/2022 18:55:48 - INFO - codeparrot_training - Step 94: {'lr': 2.3500000000000002e-05, 'samples': 18240, 'steps': 94, 'loss/train': 11.093276023864746} 01/27/2022 18:55:52 - INFO - codeparrot_training - Step 95: {'lr': 2.375e-05, 'samples': 18432, 'steps': 95, 'loss/train': 10.794178247451782} 01/27/2022 18:55:56 - INFO - codeparrot_training - Step 96: {'lr': 2.4e-05, 'samples': 18624, 'steps': 96, 'loss/train': 10.56095266342163} 01/27/2022 18:56:00 - INFO - codeparrot_training - Step 97: {'lr': 2.425e-05, 'samples': 18816, 'steps': 97, 'loss/train': 9.840100049972534} 01/27/2022 18:56:05 - INFO - codeparrot_training - Step 98: {'lr': 2.4500000000000003e-05, 'samples': 19008, 'steps': 98, 'loss/train': 10.543049097061157} 01/27/2022 18:56:09 - INFO - codeparrot_training - Step 99: {'lr': 2.4750000000000002e-05, 'samples': 19200, 'steps': 99, 'loss/train': 10.238038301467896} 01/27/2022 18:56:14 - INFO - codeparrot_training - Step 100: {'lr': 2.5e-05, 'samples': 19392, 'steps': 100, 'loss/train': 10.771350145339966} 01/27/2022 18:56:18 - INFO - codeparrot_training - Step 101: {'lr': 2.525e-05, 'samples': 19584, 'steps': 101, 'loss/train': 10.056172370910645} 01/27/2022 18:56:22 - INFO - codeparrot_training - Step 102: {'lr': 2.55e-05, 'samples': 19776, 'steps': 102, 'loss/train': 10.980010986328125} 01/27/2022 18:56:28 - INFO - codeparrot_training - Step 103: {'lr': 2.575e-05, 'samples': 19968, 'steps': 103, 'loss/train': 10.886353969573975} 01/27/2022 18:56:32 - INFO - codeparrot_training - Step 104: {'lr': 2.6e-05, 'samples': 20160, 'steps': 104, 'loss/train': 10.946210861206055} 01/27/2022 18:56:36 - INFO - codeparrot_training - Step 105: {'lr': 2.625e-05, 'samples': 20352, 'steps': 105, 'loss/train': 10.287086963653564} 01/27/2022 18:56:41 - INFO - codeparrot_training - Step 106: {'lr': 2.65e-05, 'samples': 20544, 'steps': 106, 'loss/train': 10.691130638122559} 01/27/2022 18:56:45 - INFO - codeparrot_training - Step 107: {'lr': 2.675e-05, 'samples': 20736, 'steps': 107, 'loss/train': 10.635468006134033} 01/27/2022 18:56:50 - INFO - codeparrot_training - Step 108: {'lr': 2.7e-05, 'samples': 20928, 'steps': 108, 'loss/train': 10.396392345428467} 01/27/2022 18:56:54 - INFO - codeparrot_training - Step 109: {'lr': 2.725e-05, 'samples': 21120, 'steps': 109, 'loss/train': 11.198972940444946} 01/27/2022 18:56:58 - INFO - codeparrot_training - Step 110: {'lr': 2.75e-05, 'samples': 21312, 'steps': 110, 'loss/train': 10.812159061431885} 01/27/2022 18:57:03 - INFO - codeparrot_training - Step 111: {'lr': 2.775e-05, 'samples': 21504, 'steps': 111, 'loss/train': 10.239061117172241} 01/27/2022 18:57:07 - INFO - codeparrot_training - Step 112: {'lr': 2.8e-05, 'samples': 21696, 'steps': 112, 'loss/train': 10.685145378112793} 01/27/2022 18:57:14 - INFO - codeparrot_training - Step 113: {'lr': 2.8250000000000002e-05, 'samples': 21888, 'steps': 113, 'loss/train': 10.541263103485107} 01/27/2022 18:57:18 - INFO - codeparrot_training - Step 114: {'lr': 2.85e-05, 'samples': 22080, 'steps': 114, 'loss/train': 10.467995882034302} 01/27/2022 18:57:22 - INFO - codeparrot_training - Step 115: {'lr': 2.875e-05, 'samples': 22272, 'steps': 115, 'loss/train': 10.09206748008728} 01/27/2022 18:57:26 - INFO - codeparrot_training - Step 116: {'lr': 2.9000000000000004e-05, 'samples': 22464, 'steps': 116, 'loss/train': 9.689193964004517} 01/27/2022 18:57:30 - INFO - codeparrot_training - Step 117: {'lr': 2.9250000000000003e-05, 'samples': 22656, 'steps': 117, 'loss/train': 9.968210935592651} 01/27/2022 18:57:34 - INFO - codeparrot_training - Step 118: {'lr': 2.95e-05, 'samples': 22848, 'steps': 118, 'loss/train': 10.019670009613037} 01/27/2022 18:57:39 - INFO - codeparrot_training - Step 119: {'lr': 2.9749999999999998e-05, 'samples': 23040, 'steps': 119, 'loss/train': 9.76928186416626} 01/27/2022 18:57:44 - INFO - codeparrot_training - Step 120: {'lr': 3e-05, 'samples': 23232, 'steps': 120, 'loss/train': 9.688311338424683} 01/27/2022 18:57:48 - INFO - codeparrot_training - Step 121: {'lr': 3.025e-05, 'samples': 23424, 'steps': 121, 'loss/train': 10.486538171768188} 01/27/2022 18:57:52 - INFO - codeparrot_training - Step 122: {'lr': 3.05e-05, 'samples': 23616, 'steps': 122, 'loss/train': 10.556225538253784} 01/27/2022 18:57:56 - INFO - codeparrot_training - Step 123: {'lr': 3.075e-05, 'samples': 23808, 'steps': 123, 'loss/train': 9.987429141998291} 01/27/2022 18:58:01 - INFO - codeparrot_training - Step 124: {'lr': 3.1e-05, 'samples': 24000, 'steps': 124, 'loss/train': 9.408846616744995} 01/27/2022 18:58:05 - INFO - codeparrot_training - Step 125: {'lr': 3.125e-05, 'samples': 24192, 'steps': 125, 'loss/train': 9.847419261932373} 01/27/2022 18:58:09 - INFO - codeparrot_training - Step 126: {'lr': 3.15e-05, 'samples': 24384, 'steps': 126, 'loss/train': 10.11429762840271} 01/27/2022 18:58:14 - INFO - codeparrot_training - Step 127: {'lr': 3.175e-05, 'samples': 24576, 'steps': 127, 'loss/train': 10.277008295059204} 01/27/2022 18:58:18 - INFO - codeparrot_training - Step 128: {'lr': 3.2e-05, 'samples': 24768, 'steps': 128, 'loss/train': 9.601389169692993} 01/27/2022 18:58:23 - INFO - codeparrot_training - Step 129: {'lr': 3.2250000000000005e-05, 'samples': 24960, 'steps': 129, 'loss/train': 10.083940029144287} 01/27/2022 18:58:27 - INFO - codeparrot_training - Step 130: {'lr': 3.2500000000000004e-05, 'samples': 25152, 'steps': 130, 'loss/train': 10.465167760848999} 01/27/2022 18:58:31 - INFO - codeparrot_training - Step 131: {'lr': 3.275e-05, 'samples': 25344, 'steps': 131, 'loss/train': 10.400367021560669} 01/27/2022 18:58:35 - INFO - codeparrot_training - Step 132: {'lr': 3.3e-05, 'samples': 25536, 'steps': 132, 'loss/train': 9.513238191604614} 01/27/2022 18:58:39 - INFO - codeparrot_training - Step 133: {'lr': 3.325e-05, 'samples': 25728, 'steps': 133, 'loss/train': 9.972669124603271} 01/27/2022 18:58:46 - INFO - codeparrot_training - Step 134: {'lr': 3.35e-05, 'samples': 25920, 'steps': 134, 'loss/train': 11.59807276725769} 01/27/2022 18:58:50 - INFO - codeparrot_training - Step 135: {'lr': 3.375e-05, 'samples': 26112, 'steps': 135, 'loss/train': 10.637120962142944} 01/27/2022 18:58:54 - INFO - codeparrot_training - Step 136: {'lr': 3.4000000000000007e-05, 'samples': 26304, 'steps': 136, 'loss/train': 9.751748085021973} 01/27/2022 18:58:59 - INFO - codeparrot_training - Step 137: {'lr': 3.4250000000000006e-05, 'samples': 26496, 'steps': 137, 'loss/train': 8.839701890945435} 01/27/2022 18:59:03 - INFO - codeparrot_training - Step 138: {'lr': 3.4500000000000005e-05, 'samples': 26688, 'steps': 138, 'loss/train': 10.428051710128784} 01/27/2022 18:59:08 - INFO - codeparrot_training - Step 139: {'lr': 3.4750000000000004e-05, 'samples': 26880, 'steps': 139, 'loss/train': 10.133004426956177} 01/27/2022 18:59:12 - INFO - codeparrot_training - Step 140: {'lr': 3.5000000000000004e-05, 'samples': 27072, 'steps': 140, 'loss/train': 6.793606281280518} 01/27/2022 18:59:16 - INFO - codeparrot_training - Step 141: {'lr': 3.5249999999999996e-05, 'samples': 27264, 'steps': 141, 'loss/train': 8.847809314727783} 01/27/2022 18:59:20 - INFO - codeparrot_training - Step 142: {'lr': 3.5499999999999996e-05, 'samples': 27456, 'steps': 142, 'loss/train': 10.743701219558716} 01/27/2022 18:59:24 - INFO - codeparrot_training - Step 143: {'lr': 3.5749999999999995e-05, 'samples': 27648, 'steps': 143, 'loss/train': 10.090747833251953} 01/27/2022 18:59:30 - INFO - codeparrot_training - Step 144: {'lr': 3.6e-05, 'samples': 27840, 'steps': 144, 'loss/train': 10.086535692214966} 01/27/2022 18:59:35 - INFO - codeparrot_training - Step 145: {'lr': 3.625e-05, 'samples': 28032, 'steps': 145, 'loss/train': 10.105350494384766} 01/27/2022 18:59:39 - INFO - codeparrot_training - Step 146: {'lr': 3.65e-05, 'samples': 28224, 'steps': 146, 'loss/train': 9.796883583068848} 01/27/2022 18:59:43 - INFO - codeparrot_training - Step 147: {'lr': 3.675e-05, 'samples': 28416, 'steps': 147, 'loss/train': 10.850471019744873} 01/27/2022 18:59:47 - INFO - codeparrot_training - Step 148: {'lr': 3.7e-05, 'samples': 28608, 'steps': 148, 'loss/train': 10.105576515197754} 01/27/2022 18:59:52 - INFO - codeparrot_training - Step 149: {'lr': 3.725e-05, 'samples': 28800, 'steps': 149, 'loss/train': 10.211700439453125} 01/27/2022 18:59:56 - INFO - codeparrot_training - Step 150: {'lr': 3.75e-05, 'samples': 28992, 'steps': 150, 'loss/train': 10.29758906364441} 01/27/2022 19:00:00 - INFO - codeparrot_training - Step 151: {'lr': 3.775e-05, 'samples': 29184, 'steps': 151, 'loss/train': 9.936643123626709} 01/27/2022 19:00:04 - INFO - codeparrot_training - Step 152: {'lr': 3.8e-05, 'samples': 29376, 'steps': 152, 'loss/train': 8.640991687774658} 01/27/2022 19:00:09 - INFO - codeparrot_training - Step 153: {'lr': 3.825e-05, 'samples': 29568, 'steps': 153, 'loss/train': 10.395740747451782} 01/27/2022 19:00:14 - INFO - codeparrot_training - Step 154: {'lr': 3.85e-05, 'samples': 29760, 'steps': 154, 'loss/train': 9.710185289382935} 01/27/2022 19:00:18 - INFO - codeparrot_training - Step 155: {'lr': 3.875e-05, 'samples': 29952, 'steps': 155, 'loss/train': 10.221109628677368} 01/27/2022 19:00:22 - INFO - codeparrot_training - Step 156: {'lr': 3.9e-05, 'samples': 30144, 'steps': 156, 'loss/train': 9.791813850402832} 01/27/2022 19:00:26 - INFO - codeparrot_training - Step 157: {'lr': 3.925e-05, 'samples': 30336, 'steps': 157, 'loss/train': 9.372159004211426} 01/27/2022 19:00:30 - INFO - codeparrot_training - Step 158: {'lr': 3.95e-05, 'samples': 30528, 'steps': 158, 'loss/train': 10.537101745605469} 01/27/2022 19:00:37 - INFO - codeparrot_training - Step 159: {'lr': 3.9750000000000004e-05, 'samples': 30720, 'steps': 159, 'loss/train': 8.887600421905518} 01/27/2022 19:00:41 - INFO - codeparrot_training - Step 160: {'lr': 4e-05, 'samples': 30912, 'steps': 160, 'loss/train': 9.904645442962646} 01/27/2022 19:00:45 - INFO - codeparrot_training - Step 161: {'lr': 4.025e-05, 'samples': 31104, 'steps': 161, 'loss/train': 9.94864010810852} 01/27/2022 19:00:49 - INFO - codeparrot_training - Step 162: {'lr': 4.05e-05, 'samples': 31296, 'steps': 162, 'loss/train': 9.892107725143433} 01/27/2022 19:00:53 - INFO - codeparrot_training - Step 163: {'lr': 4.075e-05, 'samples': 31488, 'steps': 163, 'loss/train': 10.019979000091553} 01/27/2022 19:00:58 - INFO - codeparrot_training - Step 164: {'lr': 4.1e-05, 'samples': 31680, 'steps': 164, 'loss/train': 9.86623764038086} 01/27/2022 19:01:03 - INFO - codeparrot_training - Step 165: {'lr': 4.125e-05, 'samples': 31872, 'steps': 165, 'loss/train': 9.841260194778442} 01/27/2022 19:01:07 - INFO - codeparrot_training - Step 166: {'lr': 4.1500000000000006e-05, 'samples': 32064, 'steps': 166, 'loss/train': 8.373968839645386} 01/27/2022 19:01:11 - INFO - codeparrot_training - Step 167: {'lr': 4.1750000000000005e-05, 'samples': 32256, 'steps': 167, 'loss/train': 9.584777355194092} 01/27/2022 19:01:15 - INFO - codeparrot_training - Step 168: {'lr': 4.2000000000000004e-05, 'samples': 32448, 'steps': 168, 'loss/train': 10.839874505996704} 01/27/2022 19:01:20 - INFO - codeparrot_training - Step 169: {'lr': 4.2250000000000004e-05, 'samples': 32640, 'steps': 169, 'loss/train': 10.583508253097534} 01/27/2022 19:01:24 - INFO - codeparrot_training - Step 170: {'lr': 4.25e-05, 'samples': 32832, 'steps': 170, 'loss/train': 9.085827112197876} 01/27/2022 19:01:28 - INFO - codeparrot_training - Step 171: {'lr': 4.275e-05, 'samples': 33024, 'steps': 171, 'loss/train': 9.918323278427124} 01/27/2022 19:01:32 - INFO - codeparrot_training - Step 172: {'lr': 4.2999999999999995e-05, 'samples': 33216, 'steps': 172, 'loss/train': 10.165422677993774} 01/27/2022 19:01:37 - INFO - codeparrot_training - Step 173: {'lr': 4.325e-05, 'samples': 33408, 'steps': 173, 'loss/train': 9.276403427124023} 01/27/2022 19:01:42 - INFO - codeparrot_training - Step 174: {'lr': 4.35e-05, 'samples': 33600, 'steps': 174, 'loss/train': 9.754451036453247} 01/27/2022 19:01:46 - INFO - codeparrot_training - Step 175: {'lr': 4.375e-05, 'samples': 33792, 'steps': 175, 'loss/train': 10.083123207092285} 01/27/2022 19:01:50 - INFO - codeparrot_training - Step 176: {'lr': 4.4e-05, 'samples': 33984, 'steps': 176, 'loss/train': 8.438665866851807} 01/27/2022 19:01:54 - INFO - codeparrot_training - Step 177: {'lr': 4.425e-05, 'samples': 34176, 'steps': 177, 'loss/train': 8.652661085128784} 01/27/2022 19:01:58 - INFO - codeparrot_training - Step 178: {'lr': 4.45e-05, 'samples': 34368, 'steps': 178, 'loss/train': 9.337483406066895} 01/27/2022 19:02:04 - INFO - codeparrot_training - Step 179: {'lr': 4.475e-05, 'samples': 34560, 'steps': 179, 'loss/train': 9.091297388076782} 01/27/2022 19:02:09 - INFO - codeparrot_training - Step 180: {'lr': 4.4999999999999996e-05, 'samples': 34752, 'steps': 180, 'loss/train': 9.52385401725769} 01/27/2022 19:02:13 - INFO - codeparrot_training - Step 181: {'lr': 4.525e-05, 'samples': 34944, 'steps': 181, 'loss/train': 9.355359077453613} 01/27/2022 19:02:17 - INFO - codeparrot_training - Step 182: {'lr': 4.55e-05, 'samples': 35136, 'steps': 182, 'loss/train': 8.646886110305786} 01/27/2022 19:02:21 - INFO - codeparrot_training - Step 183: {'lr': 4.575e-05, 'samples': 35328, 'steps': 183, 'loss/train': 10.042189836502075} 01/27/2022 19:02:26 - INFO - codeparrot_training - Step 184: {'lr': 4.6e-05, 'samples': 35520, 'steps': 184, 'loss/train': 10.683302164077759} 01/27/2022 19:02:30 - INFO - codeparrot_training - Step 185: {'lr': 4.625e-05, 'samples': 35712, 'steps': 185, 'loss/train': 8.979841232299805} 01/27/2022 19:02:35 - INFO - codeparrot_training - Step 186: {'lr': 4.65e-05, 'samples': 35904, 'steps': 186, 'loss/train': 9.779321908950806} 01/27/2022 19:02:39 - INFO - codeparrot_training - Step 187: {'lr': 4.675e-05, 'samples': 36096, 'steps': 187, 'loss/train': 9.665011167526245} 01/27/2022 19:02:43 - INFO - codeparrot_training - Step 188: {'lr': 4.7000000000000004e-05, 'samples': 36288, 'steps': 188, 'loss/train': 9.561795473098755} 01/27/2022 19:02:48 - INFO - codeparrot_training - Step 189: {'lr': 4.725e-05, 'samples': 36480, 'steps': 189, 'loss/train': 9.158857583999634} 01/27/2022 19:02:52 - INFO - codeparrot_training - Step 190: {'lr': 4.75e-05, 'samples': 36672, 'steps': 190, 'loss/train': 9.023682117462158} 01/27/2022 19:02:56 - INFO - codeparrot_training - Step 191: {'lr': 4.775e-05, 'samples': 36864, 'steps': 191, 'loss/train': 9.031155824661255} 01/27/2022 19:03:00 - INFO - codeparrot_training - Step 192: {'lr': 4.8e-05, 'samples': 37056, 'steps': 192, 'loss/train': 9.122292280197144} 01/27/2022 19:03:05 - INFO - codeparrot_training - Step 193: {'lr': 4.825e-05, 'samples': 37248, 'steps': 193, 'loss/train': 9.38718581199646} 01/27/2022 19:03:11 - INFO - codeparrot_training - Step 194: {'lr': 4.85e-05, 'samples': 37440, 'steps': 194, 'loss/train': 9.636369466781616} 01/27/2022 19:03:16 - INFO - codeparrot_training - Step 195: {'lr': 4.8750000000000006e-05, 'samples': 37632, 'steps': 195, 'loss/train': 8.828883647918701} 01/27/2022 19:03:20 - INFO - codeparrot_training - Step 196: {'lr': 4.9000000000000005e-05, 'samples': 37824, 'steps': 196, 'loss/train': 6.451952219009399} 01/27/2022 19:03:24 - INFO - codeparrot_training - Step 197: {'lr': 4.9250000000000004e-05, 'samples': 38016, 'steps': 197, 'loss/train': 6.7652199268341064} 01/27/2022 19:03:28 - INFO - codeparrot_training - Step 198: {'lr': 4.9500000000000004e-05, 'samples': 38208, 'steps': 198, 'loss/train': 9.880377531051636} 01/27/2022 19:03:32 - INFO - codeparrot_training - Step 199: {'lr': 4.975e-05, 'samples': 38400, 'steps': 199, 'loss/train': 9.147174596786499} 01/27/2022 19:03:37 - INFO - codeparrot_training - Step 200: {'lr': 5e-05, 'samples': 38592, 'steps': 200, 'loss/train': 8.676039218902588} 01/27/2022 19:03:41 - INFO - codeparrot_training - Step 201: {'lr': 5.025e-05, 'samples': 38784, 'steps': 201, 'loss/train': 9.533545017242432} 01/27/2022 19:03:46 - INFO - codeparrot_training - Step 202: {'lr': 5.05e-05, 'samples': 38976, 'steps': 202, 'loss/train': 9.315980672836304} 01/27/2022 19:03:50 - INFO - codeparrot_training - Step 203: {'lr': 5.075000000000001e-05, 'samples': 39168, 'steps': 203, 'loss/train': 10.182298421859741} 01/27/2022 19:03:54 - INFO - codeparrot_training - Step 204: {'lr': 5.1e-05, 'samples': 39360, 'steps': 204, 'loss/train': 10.115712404251099} 01/27/2022 19:04:00 - INFO - codeparrot_training - Step 205: {'lr': 5.125e-05, 'samples': 39552, 'steps': 205, 'loss/train': 8.276333570480347} 01/27/2022 19:04:04 - INFO - codeparrot_training - Step 206: {'lr': 5.15e-05, 'samples': 39744, 'steps': 206, 'loss/train': 9.16178011894226} 01/27/2022 19:04:08 - INFO - codeparrot_training - Step 207: {'lr': 5.175e-05, 'samples': 39936, 'steps': 207, 'loss/train': 9.451743364334106} 01/27/2022 19:04:12 - INFO - codeparrot_training - Step 208: {'lr': 5.2e-05, 'samples': 40128, 'steps': 208, 'loss/train': 10.303925514221191} 01/27/2022 19:04:17 - INFO - codeparrot_training - Step 209: {'lr': 5.2249999999999996e-05, 'samples': 40320, 'steps': 209, 'loss/train': 8.935240745544434} 01/27/2022 19:04:22 - INFO - codeparrot_training - Step 210: {'lr': 5.25e-05, 'samples': 40512, 'steps': 210, 'loss/train': 9.774219274520874} 01/27/2022 19:04:26 - INFO - codeparrot_training - Step 211: {'lr': 5.275e-05, 'samples': 40704, 'steps': 211, 'loss/train': 8.904966115951538} 01/27/2022 19:04:30 - INFO - codeparrot_training - Step 212: {'lr': 5.3e-05, 'samples': 40896, 'steps': 212, 'loss/train': 9.138321161270142} 01/27/2022 19:04:34 - INFO - codeparrot_training - Step 213: {'lr': 5.325e-05, 'samples': 41088, 'steps': 213, 'loss/train': 9.766228437423706} 01/27/2022 19:04:38 - INFO - codeparrot_training - Step 214: {'lr': 5.35e-05, 'samples': 41280, 'steps': 214, 'loss/train': 8.124200820922852} 01/27/2022 19:04:43 - INFO - codeparrot_training - Step 215: {'lr': 5.375e-05, 'samples': 41472, 'steps': 215, 'loss/train': 9.688585996627808} 01/27/2022 19:04:47 - INFO - codeparrot_training - Step 216: {'lr': 5.4e-05, 'samples': 41664, 'steps': 216, 'loss/train': 10.137500524520874} 01/27/2022 19:04:52 - INFO - codeparrot_training - Step 217: {'lr': 5.4250000000000004e-05, 'samples': 41856, 'steps': 217, 'loss/train': 9.659674644470215} 01/27/2022 19:04:56 - INFO - codeparrot_training - Step 218: {'lr': 5.45e-05, 'samples': 42048, 'steps': 218, 'loss/train': 8.457206726074219} 01/27/2022 19:05:00 - INFO - codeparrot_training - Step 219: {'lr': 5.475e-05, 'samples': 42240, 'steps': 219, 'loss/train': 9.459918737411499} 01/27/2022 19:05:06 - INFO - codeparrot_training - Step 220: {'lr': 5.5e-05, 'samples': 42432, 'steps': 220, 'loss/train': 8.549018383026123} 01/27/2022 19:05:10 - INFO - codeparrot_training - Step 221: {'lr': 5.525e-05, 'samples': 42624, 'steps': 221, 'loss/train': 10.017014265060425} 01/27/2022 19:05:15 - INFO - codeparrot_training - Step 222: {'lr': 5.55e-05, 'samples': 42816, 'steps': 222, 'loss/train': 9.32455587387085} 01/27/2022 19:05:19 - INFO - codeparrot_training - Step 223: {'lr': 5.575e-05, 'samples': 43008, 'steps': 223, 'loss/train': 9.075031757354736} 01/27/2022 19:05:23 - INFO - codeparrot_training - Step 224: {'lr': 5.6e-05, 'samples': 43200, 'steps': 224, 'loss/train': 8.99543023109436} 01/27/2022 19:05:28 - INFO - codeparrot_training - Step 225: {'lr': 5.6250000000000005e-05, 'samples': 43392, 'steps': 225, 'loss/train': 9.823913812637329} 01/27/2022 19:05:32 - INFO - codeparrot_training - Step 226: {'lr': 5.6500000000000005e-05, 'samples': 43584, 'steps': 226, 'loss/train': 9.353821277618408} 01/27/2022 19:05:37 - INFO - codeparrot_training - Step 227: {'lr': 5.6750000000000004e-05, 'samples': 43776, 'steps': 227, 'loss/train': 9.444689512252808} 01/27/2022 19:05:41 - INFO - codeparrot_training - Step 228: {'lr': 5.7e-05, 'samples': 43968, 'steps': 228, 'loss/train': 8.868521690368652} 01/27/2022 19:05:47 - INFO - codeparrot_training - Step 229: {'lr': 5.725e-05, 'samples': 44160, 'steps': 229, 'loss/train': 9.563207387924194} 01/27/2022 19:05:51 - INFO - codeparrot_training - Step 230: {'lr': 5.75e-05, 'samples': 44352, 'steps': 230, 'loss/train': 8.640009641647339} 01/27/2022 19:05:55 - INFO - codeparrot_training - Step 231: {'lr': 5.775e-05, 'samples': 44544, 'steps': 231, 'loss/train': 9.500942945480347} 01/27/2022 19:05:59 - INFO - codeparrot_training - Step 232: {'lr': 5.800000000000001e-05, 'samples': 44736, 'steps': 232, 'loss/train': 9.161518335342407} 01/27/2022 19:06:03 - INFO - codeparrot_training - Step 233: {'lr': 5.8250000000000006e-05, 'samples': 44928, 'steps': 233, 'loss/train': 11.825181484222412} 01/27/2022 19:06:09 - INFO - codeparrot_training - Step 234: {'lr': 5.8500000000000006e-05, 'samples': 45120, 'steps': 234, 'loss/train': 6.474182367324829} 01/27/2022 19:06:13 - INFO - codeparrot_training - Step 235: {'lr': 5.875e-05, 'samples': 45312, 'steps': 235, 'loss/train': 8.545825481414795} 01/27/2022 19:06:17 - INFO - codeparrot_training - Step 236: {'lr': 5.9e-05, 'samples': 45504, 'steps': 236, 'loss/train': 9.267122268676758} 01/27/2022 19:06:21 - INFO - codeparrot_training - Step 237: {'lr': 5.925e-05, 'samples': 45696, 'steps': 237, 'loss/train': 9.534893989562988} 01/27/2022 19:06:25 - INFO - codeparrot_training - Step 238: {'lr': 5.9499999999999996e-05, 'samples': 45888, 'steps': 238, 'loss/train': 9.004261493682861} 01/27/2022 19:06:30 - INFO - codeparrot_training - Step 239: {'lr': 5.9749999999999995e-05, 'samples': 46080, 'steps': 239, 'loss/train': 8.730738401412964} 01/27/2022 19:06:34 - INFO - codeparrot_training - Step 240: {'lr': 6e-05, 'samples': 46272, 'steps': 240, 'loss/train': 9.990759372711182} 01/27/2022 19:06:39 - INFO - codeparrot_training - Step 241: {'lr': 6.025e-05, 'samples': 46464, 'steps': 241, 'loss/train': 9.453759670257568} 01/27/2022 19:06:43 - INFO - codeparrot_training - Step 242: {'lr': 6.05e-05, 'samples': 46656, 'steps': 242, 'loss/train': 8.70737886428833} 01/27/2022 19:06:47 - INFO - codeparrot_training - Step 243: {'lr': 6.075e-05, 'samples': 46848, 'steps': 243, 'loss/train': 9.522861242294312} 01/27/2022 19:06:52 - INFO - codeparrot_training - Step 244: {'lr': 6.1e-05, 'samples': 47040, 'steps': 244, 'loss/train': 9.158050775527954} 01/27/2022 19:06:56 - INFO - codeparrot_training - Step 245: {'lr': 6.125e-05, 'samples': 47232, 'steps': 245, 'loss/train': 9.867875576019287} 01/27/2022 19:07:00 - INFO - codeparrot_training - Step 246: {'lr': 6.15e-05, 'samples': 47424, 'steps': 246, 'loss/train': 9.131139993667603} 01/27/2022 19:07:05 - INFO - codeparrot_training - Step 247: {'lr': 6.175e-05, 'samples': 47616, 'steps': 247, 'loss/train': 9.486006259918213} 01/27/2022 19:07:09 - INFO - codeparrot_training - Step 248: {'lr': 6.2e-05, 'samples': 47808, 'steps': 248, 'loss/train': 10.204999208450317} 01/27/2022 19:07:16 - INFO - codeparrot_training - Step 249: {'lr': 6.225e-05, 'samples': 48000, 'steps': 249, 'loss/train': 9.54790735244751} 01/27/2022 19:07:20 - INFO - codeparrot_training - Step 250: {'lr': 6.25e-05, 'samples': 48192, 'steps': 250, 'loss/train': 9.41062617301941} 01/27/2022 19:07:24 - INFO - codeparrot_training - Step 251: {'lr': 6.275000000000001e-05, 'samples': 48384, 'steps': 251, 'loss/train': 8.843762397766113} 01/27/2022 19:07:28 - INFO - codeparrot_training - Step 252: {'lr': 6.3e-05, 'samples': 48576, 'steps': 252, 'loss/train': 8.657515525817871} 01/27/2022 19:07:32 - INFO - codeparrot_training - Step 253: {'lr': 6.325e-05, 'samples': 48768, 'steps': 253, 'loss/train': 9.77532148361206} 01/27/2022 19:07:36 - INFO - codeparrot_training - Step 254: {'lr': 6.35e-05, 'samples': 48960, 'steps': 254, 'loss/train': 8.118179082870483} 01/27/2022 19:07:42 - INFO - codeparrot_training - Step 255: {'lr': 6.375e-05, 'samples': 49152, 'steps': 255, 'loss/train': 8.203963279724121} 01/27/2022 19:07:46 - INFO - codeparrot_training - Step 256: {'lr': 6.4e-05, 'samples': 49344, 'steps': 256, 'loss/train': 7.924534320831299} 01/27/2022 19:07:50 - INFO - codeparrot_training - Step 257: {'lr': 6.425e-05, 'samples': 49536, 'steps': 257, 'loss/train': 9.613861799240112} 01/27/2022 19:07:54 - INFO - codeparrot_training - Step 258: {'lr': 6.450000000000001e-05, 'samples': 49728, 'steps': 258, 'loss/train': 9.142173528671265} 01/27/2022 19:07:58 - INFO - codeparrot_training - Step 259: {'lr': 6.475e-05, 'samples': 49920, 'steps': 259, 'loss/train': 8.71092438697815} 01/27/2022 19:08:03 - INFO - codeparrot_training - Step 260: {'lr': 6.500000000000001e-05, 'samples': 50112, 'steps': 260, 'loss/train': 9.467320919036865} 01/27/2022 19:08:07 - INFO - codeparrot_training - Step 261: {'lr': 6.525e-05, 'samples': 50304, 'steps': 261, 'loss/train': 9.934967994689941} 01/27/2022 19:08:12 - INFO - codeparrot_training - Step 262: {'lr': 6.55e-05, 'samples': 50496, 'steps': 262, 'loss/train': 9.104984521865845} 01/27/2022 19:08:16 - INFO - codeparrot_training - Step 263: {'lr': 6.575e-05, 'samples': 50688, 'steps': 263, 'loss/train': 8.995105504989624} 01/27/2022 19:08:20 - INFO - codeparrot_training - Step 264: {'lr': 6.6e-05, 'samples': 50880, 'steps': 264, 'loss/train': 9.03219223022461} 01/27/2022 19:08:26 - INFO - codeparrot_training - Step 265: {'lr': 6.625000000000001e-05, 'samples': 51072, 'steps': 265, 'loss/train': 9.551888465881348} 01/27/2022 19:08:30 - INFO - codeparrot_training - Step 266: {'lr': 6.65e-05, 'samples': 51264, 'steps': 266, 'loss/train': 8.787272930145264} 01/27/2022 19:08:35 - INFO - codeparrot_training - Step 267: {'lr': 6.675000000000001e-05, 'samples': 51456, 'steps': 267, 'loss/train': 8.996274948120117} 01/27/2022 19:08:39 - INFO - codeparrot_training - Step 268: {'lr': 6.7e-05, 'samples': 51648, 'steps': 268, 'loss/train': 9.076965093612671} 01/27/2022 19:08:43 - INFO - codeparrot_training - Step 269: {'lr': 6.725000000000001e-05, 'samples': 51840, 'steps': 269, 'loss/train': 10.229608297348022} 01/27/2022 19:08:48 - INFO - codeparrot_training - Step 270: {'lr': 6.75e-05, 'samples': 52032, 'steps': 270, 'loss/train': 8.42824673652649} 01/27/2022 19:08:52 - INFO - codeparrot_training - Step 271: {'lr': 6.775000000000001e-05, 'samples': 52224, 'steps': 271, 'loss/train': 9.485559225082397} 01/27/2022 19:08:56 - INFO - codeparrot_training - Step 272: {'lr': 6.800000000000001e-05, 'samples': 52416, 'steps': 272, 'loss/train': 10.391708135604858} 01/27/2022 19:09:01 - INFO - codeparrot_training - Step 273: {'lr': 6.825e-05, 'samples': 52608, 'steps': 273, 'loss/train': 8.94865608215332} 01/27/2022 19:09:05 - INFO - codeparrot_training - Step 274: {'lr': 6.850000000000001e-05, 'samples': 52800, 'steps': 274, 'loss/train': 11.138488054275513} 01/27/2022 19:09:11 - INFO - codeparrot_training - Step 275: {'lr': 6.875e-05, 'samples': 52992, 'steps': 275, 'loss/train': 9.075019598007202} 01/27/2022 19:09:15 - INFO - codeparrot_training - Step 276: {'lr': 6.900000000000001e-05, 'samples': 53184, 'steps': 276, 'loss/train': 9.00141978263855} 01/27/2022 19:09:19 - INFO - codeparrot_training - Step 277: {'lr': 6.925e-05, 'samples': 53376, 'steps': 277, 'loss/train': 9.004418134689331} 01/27/2022 19:09:23 - INFO - codeparrot_training - Step 278: {'lr': 6.950000000000001e-05, 'samples': 53568, 'steps': 278, 'loss/train': 8.450300931930542} 01/27/2022 19:09:27 - INFO - codeparrot_training - Step 279: {'lr': 6.975e-05, 'samples': 53760, 'steps': 279, 'loss/train': 8.994419574737549} 01/27/2022 19:09:33 - INFO - codeparrot_training - Step 280: {'lr': 7.000000000000001e-05, 'samples': 53952, 'steps': 280, 'loss/train': 9.547023296356201} 01/27/2022 19:09:37 - INFO - codeparrot_training - Step 281: {'lr': 7.025000000000001e-05, 'samples': 54144, 'steps': 281, 'loss/train': 9.412967205047607} 01/27/2022 19:09:41 - INFO - codeparrot_training - Step 282: {'lr': 7.049999999999999e-05, 'samples': 54336, 'steps': 282, 'loss/train': 8.810139656066895} 01/27/2022 19:09:45 - INFO - codeparrot_training - Step 283: {'lr': 7.075e-05, 'samples': 54528, 'steps': 283, 'loss/train': 9.050954818725586} 01/27/2022 19:09:49 - INFO - codeparrot_training - Step 284: {'lr': 7.099999999999999e-05, 'samples': 54720, 'steps': 284, 'loss/train': 5.882340788841248} 01/27/2022 19:09:54 - INFO - codeparrot_training - Step 285: {'lr': 7.125e-05, 'samples': 54912, 'steps': 285, 'loss/train': 8.897455215454102} 01/27/2022 19:09:59 - INFO - codeparrot_training - Step 286: {'lr': 7.149999999999999e-05, 'samples': 55104, 'steps': 286, 'loss/train': 9.109829664230347} 01/27/2022 19:10:03 - INFO - codeparrot_training - Step 287: {'lr': 7.175e-05, 'samples': 55296, 'steps': 287, 'loss/train': 9.303452253341675} 01/27/2022 19:10:07 - INFO - codeparrot_training - Step 288: {'lr': 7.2e-05, 'samples': 55488, 'steps': 288, 'loss/train': 9.503397703170776} 01/27/2022 19:10:11 - INFO - codeparrot_training - Step 289: {'lr': 7.225e-05, 'samples': 55680, 'steps': 289, 'loss/train': 9.040723085403442} 01/27/2022 19:10:16 - INFO - codeparrot_training - Step 290: {'lr': 7.25e-05, 'samples': 55872, 'steps': 290, 'loss/train': 9.225221872329712} 01/27/2022 19:10:20 - INFO - codeparrot_training - Step 291: {'lr': 7.274999999999999e-05, 'samples': 56064, 'steps': 291, 'loss/train': 10.142202615737915} 01/27/2022 19:10:25 - INFO - codeparrot_training - Step 292: {'lr': 7.3e-05, 'samples': 56256, 'steps': 292, 'loss/train': 9.247089385986328} 01/27/2022 19:10:29 - INFO - codeparrot_training - Step 293: {'lr': 7.324999999999999e-05, 'samples': 56448, 'steps': 293, 'loss/train': 9.730626583099365} 01/27/2022 19:10:33 - INFO - codeparrot_training - Step 294: {'lr': 7.35e-05, 'samples': 56640, 'steps': 294, 'loss/train': 8.77634882926941} 01/27/2022 19:10:39 - INFO - codeparrot_training - Step 295: {'lr': 7.375e-05, 'samples': 56832, 'steps': 295, 'loss/train': 9.378878831863403} 01/27/2022 19:10:43 - INFO - codeparrot_training - Step 296: {'lr': 7.4e-05, 'samples': 57024, 'steps': 296, 'loss/train': 8.283158540725708} 01/27/2022 19:10:47 - INFO - codeparrot_training - Step 297: {'lr': 7.425e-05, 'samples': 57216, 'steps': 297, 'loss/train': 9.073874473571777} 01/27/2022 19:10:52 - INFO - codeparrot_training - Step 298: {'lr': 7.45e-05, 'samples': 57408, 'steps': 298, 'loss/train': 9.411186218261719} 01/27/2022 19:10:56 - INFO - codeparrot_training - Step 299: {'lr': 7.475e-05, 'samples': 57600, 'steps': 299, 'loss/train': 9.331010341644287} 01/27/2022 19:11:01 - INFO - codeparrot_training - Step 300: {'lr': 7.5e-05, 'samples': 57792, 'steps': 300, 'loss/train': 8.803576469421387} 01/27/2022 19:11:05 - INFO - codeparrot_training - Step 301: {'lr': 7.525e-05, 'samples': 57984, 'steps': 301, 'loss/train': 8.91267728805542} 01/27/2022 19:11:09 - INFO - codeparrot_training - Step 302: {'lr': 7.55e-05, 'samples': 58176, 'steps': 302, 'loss/train': 9.571316242218018} 01/27/2022 19:11:13 - INFO - codeparrot_training - Step 303: {'lr': 7.575e-05, 'samples': 58368, 'steps': 303, 'loss/train': 8.972235918045044} 01/27/2022 19:11:17 - INFO - codeparrot_training - Step 304: {'lr': 7.6e-05, 'samples': 58560, 'steps': 304, 'loss/train': 8.86762261390686} 01/27/2022 19:11:22 - INFO - codeparrot_training - Step 305: {'lr': 7.625e-05, 'samples': 58752, 'steps': 305, 'loss/train': 9.450607538223267} 01/27/2022 19:11:27 - INFO - codeparrot_training - Step 306: {'lr': 7.65e-05, 'samples': 58944, 'steps': 306, 'loss/train': 7.825889825820923} 01/27/2022 19:11:31 - INFO - codeparrot_training - Step 307: {'lr': 7.675e-05, 'samples': 59136, 'steps': 307, 'loss/train': 9.48786735534668} 01/27/2022 19:11:35 - INFO - codeparrot_training - Step 308: {'lr': 7.7e-05, 'samples': 59328, 'steps': 308, 'loss/train': 8.511699199676514} 01/27/2022 19:11:39 - INFO - codeparrot_training - Step 309: {'lr': 7.725000000000001e-05, 'samples': 59520, 'steps': 309, 'loss/train': 8.14856243133545} 01/27/2022 19:11:45 - INFO - codeparrot_training - Step 310: {'lr': 7.75e-05, 'samples': 59712, 'steps': 310, 'loss/train': 9.22744631767273} 01/27/2022 19:11:49 - INFO - codeparrot_training - Step 311: {'lr': 7.775e-05, 'samples': 59904, 'steps': 311, 'loss/train': 8.862165927886963} 01/27/2022 19:11:54 - INFO - codeparrot_training - Step 312: {'lr': 7.8e-05, 'samples': 60096, 'steps': 312, 'loss/train': 7.8139564990997314} 01/27/2022 19:11:58 - INFO - codeparrot_training - Step 313: {'lr': 7.825e-05, 'samples': 60288, 'steps': 313, 'loss/train': 9.201639890670776} 01/27/2022 19:12:02 - INFO - codeparrot_training - Step 314: {'lr': 7.85e-05, 'samples': 60480, 'steps': 314, 'loss/train': 9.275781154632568} 01/27/2022 19:12:07 - INFO - codeparrot_training - Step 315: {'lr': 7.875e-05, 'samples': 60672, 'steps': 315, 'loss/train': 8.995943069458008} 01/27/2022 19:12:11 - INFO - codeparrot_training - Step 316: {'lr': 7.9e-05, 'samples': 60864, 'steps': 316, 'loss/train': 8.69792890548706} 01/27/2022 19:12:15 - INFO - codeparrot_training - Step 317: {'lr': 7.925e-05, 'samples': 61056, 'steps': 317, 'loss/train': 9.844456672668457} 01/27/2022 19:12:20 - INFO - codeparrot_training - Step 318: {'lr': 7.950000000000001e-05, 'samples': 61248, 'steps': 318, 'loss/train': 10.125522136688232} 01/27/2022 19:12:24 - INFO - codeparrot_training - Step 319: {'lr': 7.975e-05, 'samples': 61440, 'steps': 319, 'loss/train': 9.283169746398926} 01/27/2022 19:12:30 - INFO - codeparrot_training - Step 320: {'lr': 8e-05, 'samples': 61632, 'steps': 320, 'loss/train': 9.539125442504883} 01/27/2022 19:12:34 - INFO - codeparrot_training - Step 321: {'lr': 8.025e-05, 'samples': 61824, 'steps': 321, 'loss/train': 8.35297966003418} 01/27/2022 19:12:38 - INFO - codeparrot_training - Step 322: {'lr': 8.05e-05, 'samples': 62016, 'steps': 322, 'loss/train': 9.01502251625061} 01/27/2022 19:12:43 - INFO - codeparrot_training - Step 323: {'lr': 8.075e-05, 'samples': 62208, 'steps': 323, 'loss/train': 8.88915753364563} 01/27/2022 19:12:47 - INFO - codeparrot_training - Step 324: {'lr': 8.1e-05, 'samples': 62400, 'steps': 324, 'loss/train': 9.352758407592773} 01/27/2022 19:12:52 - INFO - codeparrot_training - Step 325: {'lr': 8.125000000000001e-05, 'samples': 62592, 'steps': 325, 'loss/train': 8.851012945175171} 01/27/2022 19:12:56 - INFO - codeparrot_training - Step 326: {'lr': 8.15e-05, 'samples': 62784, 'steps': 326, 'loss/train': 8.044254541397095} 01/27/2022 19:13:00 - INFO - codeparrot_training - Step 327: {'lr': 8.175000000000001e-05, 'samples': 62976, 'steps': 327, 'loss/train': 8.919136047363281} 01/27/2022 19:13:04 - INFO - codeparrot_training - Step 328: {'lr': 8.2e-05, 'samples': 63168, 'steps': 328, 'loss/train': 7.897423267364502} 01/27/2022 19:13:09 - INFO - codeparrot_training - Step 329: {'lr': 8.225000000000001e-05, 'samples': 63360, 'steps': 329, 'loss/train': 8.732008695602417} 01/27/2022 19:13:14 - INFO - codeparrot_training - Step 330: {'lr': 8.25e-05, 'samples': 63552, 'steps': 330, 'loss/train': 8.978665351867676} 01/27/2022 19:13:18 - INFO - codeparrot_training - Step 331: {'lr': 8.275e-05, 'samples': 63744, 'steps': 331, 'loss/train': 8.693783283233643} 01/27/2022 19:13:22 - INFO - codeparrot_training - Step 332: {'lr': 8.300000000000001e-05, 'samples': 63936, 'steps': 332, 'loss/train': 10.519891262054443} 01/27/2022 19:13:27 - INFO - codeparrot_training - Step 333: {'lr': 8.325e-05, 'samples': 64128, 'steps': 333, 'loss/train': 9.274801254272461} 01/27/2022 19:13:31 - INFO - codeparrot_training - Step 334: {'lr': 8.350000000000001e-05, 'samples': 64320, 'steps': 334, 'loss/train': 9.015591859817505} 01/27/2022 19:13:36 - INFO - codeparrot_training - Step 335: {'lr': 8.375e-05, 'samples': 64512, 'steps': 335, 'loss/train': 8.601804971694946} 01/27/2022 19:13:40 - INFO - codeparrot_training - Step 336: {'lr': 8.400000000000001e-05, 'samples': 64704, 'steps': 336, 'loss/train': 9.10523271560669} 01/27/2022 19:13:44 - INFO - codeparrot_training - Step 337: {'lr': 8.425e-05, 'samples': 64896, 'steps': 337, 'loss/train': 9.47730302810669} 01/27/2022 19:13:49 - INFO - codeparrot_training - Step 338: {'lr': 8.450000000000001e-05, 'samples': 65088, 'steps': 338, 'loss/train': 8.80834150314331} 01/27/2022 19:13:53 - INFO - codeparrot_training - Step 339: {'lr': 8.475000000000001e-05, 'samples': 65280, 'steps': 339, 'loss/train': 8.973668575286865} 01/27/2022 19:14:00 - INFO - codeparrot_training - Step 340: {'lr': 8.5e-05, 'samples': 65472, 'steps': 340, 'loss/train': 8.7022705078125} 01/27/2022 19:14:04 - INFO - codeparrot_training - Step 341: {'lr': 8.525000000000001e-05, 'samples': 65664, 'steps': 341, 'loss/train': 9.048253297805786} 01/27/2022 19:14:08 - INFO - codeparrot_training - Step 342: {'lr': 8.55e-05, 'samples': 65856, 'steps': 342, 'loss/train': 8.496989965438843} 01/27/2022 19:14:12 - INFO - codeparrot_training - Step 343: {'lr': 8.575000000000001e-05, 'samples': 66048, 'steps': 343, 'loss/train': 10.007545709609985} 01/27/2022 19:14:16 - INFO - codeparrot_training - Step 344: {'lr': 8.599999999999999e-05, 'samples': 66240, 'steps': 344, 'loss/train': 9.02388310432434} 01/27/2022 19:14:21 - INFO - codeparrot_training - Step 345: {'lr': 8.625e-05, 'samples': 66432, 'steps': 345, 'loss/train': 8.611332893371582} 01/27/2022 19:14:25 - INFO - codeparrot_training - Step 346: {'lr': 8.65e-05, 'samples': 66624, 'steps': 346, 'loss/train': 8.02782654762268} 01/27/2022 19:14:30 - INFO - codeparrot_training - Step 347: {'lr': 8.675e-05, 'samples': 66816, 'steps': 347, 'loss/train': 8.79668641090393} 01/27/2022 19:14:34 - INFO - codeparrot_training - Step 348: {'lr': 8.7e-05, 'samples': 67008, 'steps': 348, 'loss/train': 8.362978219985962} 01/27/2022 19:14:38 - INFO - codeparrot_training - Step 349: {'lr': 8.724999999999999e-05, 'samples': 67200, 'steps': 349, 'loss/train': 8.863133668899536} 01/27/2022 19:14:43 - INFO - codeparrot_training - Step 350: {'lr': 8.75e-05, 'samples': 67392, 'steps': 350, 'loss/train': 8.88676929473877} 01/27/2022 19:14:47 - INFO - codeparrot_training - Step 351: {'lr': 8.774999999999999e-05, 'samples': 67584, 'steps': 351, 'loss/train': 9.247013568878174} 01/27/2022 19:14:51 - INFO - codeparrot_training - Step 352: {'lr': 8.8e-05, 'samples': 67776, 'steps': 352, 'loss/train': 7.664119720458984} 01/27/2022 19:14:56 - INFO - codeparrot_training - Step 353: {'lr': 8.824999999999999e-05, 'samples': 67968, 'steps': 353, 'loss/train': 9.151994705200195} 01/27/2022 19:15:00 - INFO - codeparrot_training - Step 354: {'lr': 8.85e-05, 'samples': 68160, 'steps': 354, 'loss/train': 9.227952003479004} 01/27/2022 19:15:06 - INFO - codeparrot_training - Step 355: {'lr': 8.875e-05, 'samples': 68352, 'steps': 355, 'loss/train': 9.120519876480103} 01/27/2022 19:15:10 - INFO - codeparrot_training - Step 356: {'lr': 8.9e-05, 'samples': 68544, 'steps': 356, 'loss/train': 9.297804594039917} 01/27/2022 19:15:14 - INFO - codeparrot_training - Step 357: {'lr': 8.925e-05, 'samples': 68736, 'steps': 357, 'loss/train': 8.624364852905273} 01/27/2022 19:15:18 - INFO - codeparrot_training - Step 358: {'lr': 8.95e-05, 'samples': 68928, 'steps': 358, 'loss/train': 9.515746593475342} 01/27/2022 19:15:23 - INFO - codeparrot_training - Step 359: {'lr': 8.975e-05, 'samples': 69120, 'steps': 359, 'loss/train': 8.978293418884277} 01/27/2022 19:15:27 - INFO - codeparrot_training - Step 360: {'lr': 8.999999999999999e-05, 'samples': 69312, 'steps': 360, 'loss/train': 8.702915668487549} 01/27/2022 19:15:32 - INFO - codeparrot_training - Step 361: {'lr': 9.025e-05, 'samples': 69504, 'steps': 361, 'loss/train': 9.636523246765137} 01/27/2022 19:15:36 - INFO - codeparrot_training - Step 362: {'lr': 9.05e-05, 'samples': 69696, 'steps': 362, 'loss/train': 9.489566087722778} 01/27/2022 19:15:40 - INFO - codeparrot_training - Step 363: {'lr': 9.075e-05, 'samples': 69888, 'steps': 363, 'loss/train': 9.47793173789978} 01/27/2022 19:15:44 - INFO - codeparrot_training - Step 364: {'lr': 9.1e-05, 'samples': 70080, 'steps': 364, 'loss/train': 8.561887979507446} 01/27/2022 19:15:48 - INFO - codeparrot_training - Step 365: {'lr': 9.125e-05, 'samples': 70272, 'steps': 365, 'loss/train': 8.943610668182373} 01/27/2022 19:15:54 - INFO - codeparrot_training - Step 366: {'lr': 9.15e-05, 'samples': 70464, 'steps': 366, 'loss/train': 9.22202754020691} 01/27/2022 19:15:59 - INFO - codeparrot_training - Step 367: {'lr': 9.175e-05, 'samples': 70656, 'steps': 367, 'loss/train': 8.819038152694702} 01/27/2022 19:16:03 - INFO - codeparrot_training - Step 368: {'lr': 9.2e-05, 'samples': 70848, 'steps': 368, 'loss/train': 8.843501329421997} 01/27/2022 19:16:07 - INFO - codeparrot_training - Step 369: {'lr': 9.225e-05, 'samples': 71040, 'steps': 369, 'loss/train': 9.003690004348755} 01/27/2022 19:16:12 - INFO - codeparrot_training - Step 370: {'lr': 9.25e-05, 'samples': 71232, 'steps': 370, 'loss/train': 7.789920330047607} 01/27/2022 19:16:16 - INFO - codeparrot_training - Step 371: {'lr': 9.275e-05, 'samples': 71424, 'steps': 371, 'loss/train': 8.740252017974854} 01/27/2022 19:16:20 - INFO - codeparrot_training - Step 372: {'lr': 9.3e-05, 'samples': 71616, 'steps': 372, 'loss/train': 9.126051664352417} 01/27/2022 19:16:24 - INFO - codeparrot_training - Step 373: {'lr': 9.325e-05, 'samples': 71808, 'steps': 373, 'loss/train': 10.538711786270142} 01/27/2022 19:16:29 - INFO - codeparrot_training - Step 374: {'lr': 9.35e-05, 'samples': 72000, 'steps': 374, 'loss/train': 11.539859533309937} 01/27/2022 19:16:33 - INFO - codeparrot_training - Step 375: {'lr': 9.375e-05, 'samples': 72192, 'steps': 375, 'loss/train': 4.924748182296753} 01/27/2022 19:16:38 - INFO - codeparrot_training - Step 376: {'lr': 9.400000000000001e-05, 'samples': 72384, 'steps': 376, 'loss/train': 10.32872200012207} 01/27/2022 19:16:42 - INFO - codeparrot_training - Step 377: {'lr': 9.425e-05, 'samples': 72576, 'steps': 377, 'loss/train': 7.9768452644348145} 01/27/2022 19:16:46 - INFO - codeparrot_training - Step 378: {'lr': 9.45e-05, 'samples': 72768, 'steps': 378, 'loss/train': 8.341711521148682} 01/27/2022 19:16:50 - INFO - codeparrot_training - Step 379: {'lr': 9.475e-05, 'samples': 72960, 'steps': 379, 'loss/train': 8.565402746200562} 01/27/2022 19:16:56 - INFO - codeparrot_training - Step 380: {'lr': 9.5e-05, 'samples': 73152, 'steps': 380, 'loss/train': 9.063901662826538} 01/27/2022 19:17:00 - INFO - codeparrot_training - Step 381: {'lr': 9.525e-05, 'samples': 73344, 'steps': 381, 'loss/train': 8.95740008354187} 01/27/2022 19:17:04 - INFO - codeparrot_training - Step 382: {'lr': 9.55e-05, 'samples': 73536, 'steps': 382, 'loss/train': 8.847960233688354} 01/27/2022 19:17:09 - INFO - codeparrot_training - Step 383: {'lr': 9.575000000000001e-05, 'samples': 73728, 'steps': 383, 'loss/train': 8.798057556152344} 01/27/2022 19:17:13 - INFO - codeparrot_training - Step 384: {'lr': 9.6e-05, 'samples': 73920, 'steps': 384, 'loss/train': 8.828840732574463} 01/27/2022 19:17:18 - INFO - codeparrot_training - Step 385: {'lr': 9.625000000000001e-05, 'samples': 74112, 'steps': 385, 'loss/train': 9.475341796875} 01/27/2022 19:17:22 - INFO - codeparrot_training - Step 386: {'lr': 9.65e-05, 'samples': 74304, 'steps': 386, 'loss/train': 8.664866209030151} 01/27/2022 19:17:26 - INFO - codeparrot_training - Step 387: {'lr': 9.675000000000001e-05, 'samples': 74496, 'steps': 387, 'loss/train': 9.094900846481323} 01/27/2022 19:17:30 - INFO - codeparrot_training - Step 388: {'lr': 9.7e-05, 'samples': 74688, 'steps': 388, 'loss/train': 8.839553833007812} 01/27/2022 19:17:34 - INFO - codeparrot_training - Step 389: {'lr': 9.725e-05, 'samples': 74880, 'steps': 389, 'loss/train': 8.480072736740112} 01/27/2022 19:17:39 - INFO - codeparrot_training - Step 390: {'lr': 9.750000000000001e-05, 'samples': 75072, 'steps': 390, 'loss/train': 9.995652437210083} 01/27/2022 19:17:44 - INFO - codeparrot_training - Step 391: {'lr': 9.775e-05, 'samples': 75264, 'steps': 391, 'loss/train': 9.592024326324463} 01/27/2022 19:17:48 - INFO - codeparrot_training - Step 392: {'lr': 9.800000000000001e-05, 'samples': 75456, 'steps': 392, 'loss/train': 8.603671073913574} 01/27/2022 19:17:52 - INFO - codeparrot_training - Step 393: {'lr': 9.825e-05, 'samples': 75648, 'steps': 393, 'loss/train': 9.834109783172607} 01/27/2022 19:17:56 - INFO - codeparrot_training - Step 394: {'lr': 9.850000000000001e-05, 'samples': 75840, 'steps': 394, 'loss/train': 9.401895761489868} 01/27/2022 19:18:02 - INFO - codeparrot_training - Step 395: {'lr': 9.875e-05, 'samples': 76032, 'steps': 395, 'loss/train': 8.424455165863037} 01/27/2022 19:18:06 - INFO - codeparrot_training - Step 396: {'lr': 9.900000000000001e-05, 'samples': 76224, 'steps': 396, 'loss/train': 8.191839694976807} 01/27/2022 19:18:10 - INFO - codeparrot_training - Step 397: {'lr': 9.925000000000001e-05, 'samples': 76416, 'steps': 397, 'loss/train': 9.516587018966675} 01/27/2022 19:18:14 - INFO - codeparrot_training - Step 398: {'lr': 9.95e-05, 'samples': 76608, 'steps': 398, 'loss/train': 8.904014110565186} 01/27/2022 19:18:19 - INFO - codeparrot_training - Step 399: {'lr': 9.975000000000001e-05, 'samples': 76800, 'steps': 399, 'loss/train': 8.627155780792236} 01/27/2022 19:18:24 - INFO - codeparrot_training - Step 400: {'lr': 0.0001, 'samples': 76992, 'steps': 400, 'loss/train': 7.955677270889282} 01/27/2022 19:18:28 - INFO - codeparrot_training - Step 401: {'lr': 0.00010025000000000001, 'samples': 77184, 'steps': 401, 'loss/train': 8.836859464645386} 01/27/2022 19:18:32 - INFO - codeparrot_training - Step 402: {'lr': 0.0001005, 'samples': 77376, 'steps': 402, 'loss/train': 8.888705492019653} 01/27/2022 19:18:36 - INFO - codeparrot_training - Step 403: {'lr': 0.00010075000000000001, 'samples': 77568, 'steps': 403, 'loss/train': 9.59162163734436} 01/27/2022 19:18:40 - INFO - codeparrot_training - Step 404: {'lr': 0.000101, 'samples': 77760, 'steps': 404, 'loss/train': 7.719418287277222} 01/27/2022 19:18:46 - INFO - codeparrot_training - Step 405: {'lr': 0.00010125000000000001, 'samples': 77952, 'steps': 405, 'loss/train': 8.82796025276184} 01/27/2022 19:18:50 - INFO - codeparrot_training - Step 406: {'lr': 0.00010150000000000001, 'samples': 78144, 'steps': 406, 'loss/train': 5.772486448287964} 01/27/2022 19:18:54 - INFO - codeparrot_training - Step 407: {'lr': 0.00010174999999999999, 'samples': 78336, 'steps': 407, 'loss/train': 7.790276527404785} 01/27/2022 19:18:58 - INFO - codeparrot_training - Step 408: {'lr': 0.000102, 'samples': 78528, 'steps': 408, 'loss/train': 9.21680474281311} 01/27/2022 19:19:02 - INFO - codeparrot_training - Step 409: {'lr': 0.00010224999999999999, 'samples': 78720, 'steps': 409, 'loss/train': 9.145365715026855} 01/27/2022 19:19:07 - INFO - codeparrot_training - Step 410: {'lr': 0.0001025, 'samples': 78912, 'steps': 410, 'loss/train': 8.65236496925354} 01/27/2022 19:19:11 - INFO - codeparrot_training - Step 411: {'lr': 0.00010274999999999999, 'samples': 79104, 'steps': 411, 'loss/train': 9.195783376693726} 01/27/2022 19:19:16 - INFO - codeparrot_training - Step 412: {'lr': 0.000103, 'samples': 79296, 'steps': 412, 'loss/train': 8.685220956802368} 01/27/2022 19:19:20 - INFO - codeparrot_training - Step 413: {'lr': 0.00010325, 'samples': 79488, 'steps': 413, 'loss/train': 8.50923228263855} 01/27/2022 19:19:24 - INFO - codeparrot_training - Step 414: {'lr': 0.0001035, 'samples': 79680, 'steps': 414, 'loss/train': 8.791507244110107} 01/27/2022 19:19:30 - INFO - codeparrot_training - Step 415: {'lr': 0.00010375, 'samples': 79872, 'steps': 415, 'loss/train': 8.846287965774536} 01/27/2022 19:19:34 - INFO - codeparrot_training - Step 416: {'lr': 0.000104, 'samples': 80064, 'steps': 416, 'loss/train': 9.453208208084106} 01/27/2022 19:19:38 - INFO - codeparrot_training - Step 417: {'lr': 0.00010425, 'samples': 80256, 'steps': 417, 'loss/train': 7.977580547332764} 01/27/2022 19:19:43 - INFO - codeparrot_training - Step 418: {'lr': 0.00010449999999999999, 'samples': 80448, 'steps': 418, 'loss/train': 8.50738263130188} 01/27/2022 19:19:47 - INFO - codeparrot_training - Step 419: {'lr': 0.00010475, 'samples': 80640, 'steps': 419, 'loss/train': 7.507714033126831} 01/27/2022 19:19:51 - INFO - codeparrot_training - Step 420: {'lr': 0.000105, 'samples': 80832, 'steps': 420, 'loss/train': 10.262629508972168} 01/27/2022 19:19:56 - INFO - codeparrot_training - Step 421: {'lr': 0.00010525, 'samples': 81024, 'steps': 421, 'loss/train': 8.580647706985474} 01/27/2022 19:20:00 - INFO - codeparrot_training - Step 422: {'lr': 0.0001055, 'samples': 81216, 'steps': 422, 'loss/train': 8.881326913833618} 01/27/2022 19:20:05 - INFO - codeparrot_training - Step 423: {'lr': 0.00010575, 'samples': 81408, 'steps': 423, 'loss/train': 10.195034265518188} 01/27/2022 19:20:09 - INFO - codeparrot_training - Step 424: {'lr': 0.000106, 'samples': 81600, 'steps': 424, 'loss/train': 8.339170217514038} 01/27/2022 19:20:15 - INFO - codeparrot_training - Step 425: {'lr': 0.00010625, 'samples': 81792, 'steps': 425, 'loss/train': 8.661192655563354} 01/27/2022 19:20:19 - INFO - codeparrot_training - Step 426: {'lr': 0.0001065, 'samples': 81984, 'steps': 426, 'loss/train': 8.773484230041504} 01/27/2022 19:20:23 - INFO - codeparrot_training - Step 427: {'lr': 0.00010675, 'samples': 82176, 'steps': 427, 'loss/train': 7.9621946811676025} 01/27/2022 19:20:27 - INFO - codeparrot_training - Step 428: {'lr': 0.000107, 'samples': 82368, 'steps': 428, 'loss/train': 9.662103652954102} 01/27/2022 19:20:31 - INFO - codeparrot_training - Step 429: {'lr': 0.00010725, 'samples': 82560, 'steps': 429, 'loss/train': 8.578760862350464} 01/27/2022 19:20:36 - INFO - codeparrot_training - Step 430: {'lr': 0.0001075, 'samples': 82752, 'steps': 430, 'loss/train': 9.057950019836426} 01/27/2022 19:20:41 - INFO - codeparrot_training - Step 431: {'lr': 0.00010775, 'samples': 82944, 'steps': 431, 'loss/train': 8.532402992248535} 01/27/2022 19:20:45 - INFO - codeparrot_training - Step 432: {'lr': 0.000108, 'samples': 83136, 'steps': 432, 'loss/train': 8.037179946899414} 01/27/2022 19:20:49 - INFO - codeparrot_training - Step 433: {'lr': 0.00010825, 'samples': 83328, 'steps': 433, 'loss/train': 8.34144401550293} 01/27/2022 19:20:53 - INFO - codeparrot_training - Step 434: {'lr': 0.00010850000000000001, 'samples': 83520, 'steps': 434, 'loss/train': 9.126363515853882} 01/27/2022 19:20:58 - INFO - codeparrot_training - Step 435: {'lr': 0.00010875, 'samples': 83712, 'steps': 435, 'loss/train': 8.610699892044067} 01/27/2022 19:21:02 - INFO - codeparrot_training - Step 436: {'lr': 0.000109, 'samples': 83904, 'steps': 436, 'loss/train': 9.06845998764038} 01/27/2022 19:21:06 - INFO - codeparrot_training - Step 437: {'lr': 0.00010925, 'samples': 84096, 'steps': 437, 'loss/train': 9.466124296188354} 01/27/2022 19:21:11 - INFO - codeparrot_training - Step 438: {'lr': 0.0001095, 'samples': 84288, 'steps': 438, 'loss/train': 9.236776113510132} 01/27/2022 19:21:15 - INFO - codeparrot_training - Step 439: {'lr': 0.00010975, 'samples': 84480, 'steps': 439, 'loss/train': 8.494872093200684} 01/27/2022 19:21:22 - INFO - codeparrot_training - Step 440: {'lr': 0.00011, 'samples': 84672, 'steps': 440, 'loss/train': 8.198070287704468} 01/27/2022 19:21:26 - INFO - codeparrot_training - Step 441: {'lr': 0.00011025, 'samples': 84864, 'steps': 441, 'loss/train': 8.65276837348938} 01/27/2022 19:21:30 - INFO - codeparrot_training - Step 442: {'lr': 0.0001105, 'samples': 85056, 'steps': 442, 'loss/train': 8.7622811794281} 01/27/2022 19:21:34 - INFO - codeparrot_training - Step 443: {'lr': 0.00011075000000000001, 'samples': 85248, 'steps': 443, 'loss/train': 9.05262565612793} 01/27/2022 19:21:38 - INFO - codeparrot_training - Step 444: {'lr': 0.000111, 'samples': 85440, 'steps': 444, 'loss/train': 8.239619493484497} 01/27/2022 19:21:43 - INFO - codeparrot_training - Step 445: {'lr': 0.00011125000000000001, 'samples': 85632, 'steps': 445, 'loss/train': 8.450908184051514} 01/27/2022 19:21:48 - INFO - codeparrot_training - Step 446: {'lr': 0.0001115, 'samples': 85824, 'steps': 446, 'loss/train': 9.082252979278564} 01/27/2022 19:21:52 - INFO - codeparrot_training - Step 447: {'lr': 0.00011175, 'samples': 86016, 'steps': 447, 'loss/train': 8.904988288879395} 01/27/2022 19:21:56 - INFO - codeparrot_training - Step 448: {'lr': 0.000112, 'samples': 86208, 'steps': 448, 'loss/train': 8.551026105880737} 01/27/2022 19:22:00 - INFO - codeparrot_training - Step 449: {'lr': 0.00011225, 'samples': 86400, 'steps': 449, 'loss/train': 8.441035509109497} 01/27/2022 19:22:06 - INFO - codeparrot_training - Step 450: {'lr': 0.00011250000000000001, 'samples': 86592, 'steps': 450, 'loss/train': 8.621562480926514} 01/27/2022 19:22:10 - INFO - codeparrot_training - Step 451: {'lr': 0.00011275, 'samples': 86784, 'steps': 451, 'loss/train': 8.748263597488403} 01/27/2022 19:22:14 - INFO - codeparrot_training - Step 452: {'lr': 0.00011300000000000001, 'samples': 86976, 'steps': 452, 'loss/train': 8.997128963470459} 01/27/2022 19:22:18 - INFO - codeparrot_training - Step 453: {'lr': 0.00011325, 'samples': 87168, 'steps': 453, 'loss/train': 8.09796667098999} 01/27/2022 19:22:22 - INFO - codeparrot_training - Step 454: {'lr': 0.00011350000000000001, 'samples': 87360, 'steps': 454, 'loss/train': 8.689520359039307} 01/27/2022 19:22:28 - INFO - codeparrot_training - Step 455: {'lr': 0.00011375, 'samples': 87552, 'steps': 455, 'loss/train': 9.379069805145264} 01/27/2022 19:22:32 - INFO - codeparrot_training - Step 456: {'lr': 0.000114, 'samples': 87744, 'steps': 456, 'loss/train': 8.803585052490234} 01/27/2022 19:22:36 - INFO - codeparrot_training - Step 457: {'lr': 0.00011425000000000001, 'samples': 87936, 'steps': 457, 'loss/train': 8.374651908874512} 01/27/2022 19:22:40 - INFO - codeparrot_training - Step 458: {'lr': 0.0001145, 'samples': 88128, 'steps': 458, 'loss/train': 8.781396389007568} 01/27/2022 19:22:44 - INFO - codeparrot_training - Step 459: {'lr': 0.00011475000000000001, 'samples': 88320, 'steps': 459, 'loss/train': 8.592355012893677} 01/27/2022 19:22:49 - INFO - codeparrot_training - Step 460: {'lr': 0.000115, 'samples': 88512, 'steps': 460, 'loss/train': 7.943728923797607} 01/27/2022 19:22:54 - INFO - codeparrot_training - Step 461: {'lr': 0.00011525000000000001, 'samples': 88704, 'steps': 461, 'loss/train': 8.678408145904541} 01/27/2022 19:22:58 - INFO - codeparrot_training - Step 462: {'lr': 0.0001155, 'samples': 88896, 'steps': 462, 'loss/train': 10.198154211044312} 01/27/2022 19:23:02 - INFO - codeparrot_training - Step 463: {'lr': 0.00011575000000000001, 'samples': 89088, 'steps': 463, 'loss/train': 8.712399244308472} 01/27/2022 19:23:06 - INFO - codeparrot_training - Step 464: {'lr': 0.00011600000000000001, 'samples': 89280, 'steps': 464, 'loss/train': 8.593222618103027} 01/27/2022 19:23:11 - INFO - codeparrot_training - Step 465: {'lr': 0.00011625, 'samples': 89472, 'steps': 465, 'loss/train': 8.657633543014526} 01/27/2022 19:23:15 - INFO - codeparrot_training - Step 466: {'lr': 0.00011650000000000001, 'samples': 89664, 'steps': 466, 'loss/train': 8.809774160385132} 01/27/2022 19:23:19 - INFO - codeparrot_training - Step 467: {'lr': 0.00011675, 'samples': 89856, 'steps': 467, 'loss/train': 8.131398439407349} 01/27/2022 19:23:24 - INFO - codeparrot_training - Step 468: {'lr': 0.00011700000000000001, 'samples': 90048, 'steps': 468, 'loss/train': 8.037885189056396} 01/27/2022 19:23:28 - INFO - codeparrot_training - Step 469: {'lr': 0.00011724999999999999, 'samples': 90240, 'steps': 469, 'loss/train': 9.223987340927124} 01/27/2022 19:23:34 - INFO - codeparrot_training - Step 470: {'lr': 0.0001175, 'samples': 90432, 'steps': 470, 'loss/train': 8.752360582351685} 01/27/2022 19:23:38 - INFO - codeparrot_training - Step 471: {'lr': 0.00011775, 'samples': 90624, 'steps': 471, 'loss/train': 8.001303434371948} 01/27/2022 19:23:42 - INFO - codeparrot_training - Step 472: {'lr': 0.000118, 'samples': 90816, 'steps': 472, 'loss/train': 8.158232688903809} 01/27/2022 19:23:46 - INFO - codeparrot_training - Step 473: {'lr': 0.00011825, 'samples': 91008, 'steps': 473, 'loss/train': 7.7227113246917725} 01/27/2022 19:23:50 - INFO - codeparrot_training - Step 474: {'lr': 0.0001185, 'samples': 91200, 'steps': 474, 'loss/train': 8.60695481300354} 01/27/2022 19:23:55 - INFO - codeparrot_training - Step 475: {'lr': 0.00011875, 'samples': 91392, 'steps': 475, 'loss/train': 8.559202909469604} 01/27/2022 19:24:00 - INFO - codeparrot_training - Step 476: {'lr': 0.00011899999999999999, 'samples': 91584, 'steps': 476, 'loss/train': 8.1037015914917} 01/27/2022 19:24:04 - INFO - codeparrot_training - Step 477: {'lr': 0.00011925, 'samples': 91776, 'steps': 477, 'loss/train': 8.52787184715271} 01/27/2022 19:24:08 - INFO - codeparrot_training - Step 478: {'lr': 0.00011949999999999999, 'samples': 91968, 'steps': 478, 'loss/train': 9.877085208892822} 01/27/2022 19:24:12 - INFO - codeparrot_training - Step 479: {'lr': 0.00011975, 'samples': 92160, 'steps': 479, 'loss/train': 8.822041511535645} 01/27/2022 19:24:17 - INFO - codeparrot_training - Step 480: {'lr': 0.00012, 'samples': 92352, 'steps': 480, 'loss/train': 8.334150552749634} 01/27/2022 19:24:21 - INFO - codeparrot_training - Step 481: {'lr': 0.00012025, 'samples': 92544, 'steps': 481, 'loss/train': 8.455609560012817} 01/27/2022 19:24:25 - INFO - codeparrot_training - Step 482: {'lr': 0.0001205, 'samples': 92736, 'steps': 482, 'loss/train': 9.207207441329956} 01/27/2022 19:24:30 - INFO - codeparrot_training - Step 483: {'lr': 0.00012075, 'samples': 92928, 'steps': 483, 'loss/train': 8.488374710083008} 01/27/2022 19:24:34 - INFO - codeparrot_training - Step 484: {'lr': 0.000121, 'samples': 93120, 'steps': 484, 'loss/train': 8.650288581848145} 01/27/2022 19:24:40 - INFO - codeparrot_training - Step 485: {'lr': 0.00012124999999999999, 'samples': 93312, 'steps': 485, 'loss/train': 9.08907437324524} 01/27/2022 19:24:44 - INFO - codeparrot_training - Step 486: {'lr': 0.0001215, 'samples': 93504, 'steps': 486, 'loss/train': 8.138011693954468} 01/27/2022 19:24:48 - INFO - codeparrot_training - Step 487: {'lr': 0.00012175, 'samples': 93696, 'steps': 487, 'loss/train': 8.505961418151855} 01/27/2022 19:24:52 - INFO - codeparrot_training - Step 488: {'lr': 0.000122, 'samples': 93888, 'steps': 488, 'loss/train': 8.104951858520508} 01/27/2022 19:24:56 - INFO - codeparrot_training - Step 489: {'lr': 0.00012225, 'samples': 94080, 'steps': 489, 'loss/train': 8.997678995132446} 01/27/2022 19:25:01 - INFO - codeparrot_training - Step 490: {'lr': 0.0001225, 'samples': 94272, 'steps': 490, 'loss/train': 9.484206676483154} 01/27/2022 19:25:05 - INFO - codeparrot_training - Step 491: {'lr': 0.00012275, 'samples': 94464, 'steps': 491, 'loss/train': 8.445098161697388} 01/27/2022 19:25:10 - INFO - codeparrot_training - Step 492: {'lr': 0.000123, 'samples': 94656, 'steps': 492, 'loss/train': 9.071303129196167} 01/27/2022 19:25:14 - INFO - codeparrot_training - Step 493: {'lr': 0.00012325000000000001, 'samples': 94848, 'steps': 493, 'loss/train': 9.275761842727661} 01/27/2022 19:25:18 - INFO - codeparrot_training - Step 494: {'lr': 0.0001235, 'samples': 95040, 'steps': 494, 'loss/train': 7.959945917129517} 01/27/2022 19:25:24 - INFO - codeparrot_training - Step 495: {'lr': 0.00012375, 'samples': 95232, 'steps': 495, 'loss/train': 8.65991735458374} 01/27/2022 19:25:28 - INFO - codeparrot_training - Step 496: {'lr': 0.000124, 'samples': 95424, 'steps': 496, 'loss/train': 7.939209222793579} 01/27/2022 19:25:32 - INFO - codeparrot_training - Step 497: {'lr': 0.00012425, 'samples': 95616, 'steps': 497, 'loss/train': 8.233192920684814} 01/27/2022 19:25:36 - INFO - codeparrot_training - Step 498: {'lr': 0.0001245, 'samples': 95808, 'steps': 498, 'loss/train': 8.595185995101929} 01/27/2022 19:25:41 - INFO - codeparrot_training - Step 499: {'lr': 0.00012475, 'samples': 96000, 'steps': 499, 'loss/train': 8.485043048858643} 01/27/2022 19:25:46 - INFO - codeparrot_training - Step 500: {'lr': 0.000125, 'samples': 96192, 'steps': 500, 'loss/train': 8.449069261550903} 01/27/2022 19:25:50 - INFO - codeparrot_training - Step 501: {'lr': 0.00012525, 'samples': 96384, 'steps': 501, 'loss/train': 8.081286907196045} 01/27/2022 19:25:54 - INFO - codeparrot_training - Step 502: {'lr': 0.00012550000000000001, 'samples': 96576, 'steps': 502, 'loss/train': 7.730855941772461} 01/27/2022 19:25:58 - INFO - codeparrot_training - Step 503: {'lr': 0.00012575, 'samples': 96768, 'steps': 503, 'loss/train': 8.047250032424927} 01/27/2022 19:26:02 - INFO - codeparrot_training - Step 504: {'lr': 0.000126, 'samples': 96960, 'steps': 504, 'loss/train': 8.251461267471313} 01/27/2022 19:26:07 - INFO - codeparrot_training - Step 505: {'lr': 0.00012625, 'samples': 97152, 'steps': 505, 'loss/train': 8.245980262756348} 01/27/2022 19:26:12 - INFO - codeparrot_training - Step 506: {'lr': 0.0001265, 'samples': 97344, 'steps': 506, 'loss/train': 9.07162356376648} 01/27/2022 19:26:16 - INFO - codeparrot_training - Step 507: {'lr': 0.00012675, 'samples': 97536, 'steps': 507, 'loss/train': 8.477697372436523} 01/27/2022 19:26:20 - INFO - codeparrot_training - Step 508: {'lr': 0.000127, 'samples': 97728, 'steps': 508, 'loss/train': 8.715029954910278} 01/27/2022 19:26:24 - INFO - codeparrot_training - Step 509: {'lr': 0.00012725, 'samples': 97920, 'steps': 509, 'loss/train': 8.291935443878174} 01/27/2022 19:26:30 - INFO - codeparrot_training - Step 510: {'lr': 0.0001275, 'samples': 98112, 'steps': 510, 'loss/train': 8.807954549789429} 01/27/2022 19:26:34 - INFO - codeparrot_training - Step 511: {'lr': 0.00012775000000000002, 'samples': 98304, 'steps': 511, 'loss/train': 7.690762996673584} 01/27/2022 19:26:39 - INFO - codeparrot_training - Step 512: {'lr': 0.000128, 'samples': 98496, 'steps': 512, 'loss/train': 9.079181671142578} 01/27/2022 19:26:43 - INFO - codeparrot_training - Step 513: {'lr': 0.00012825, 'samples': 98688, 'steps': 513, 'loss/train': 8.444893598556519} 01/27/2022 19:26:47 - INFO - codeparrot_training - Step 514: {'lr': 0.0001285, 'samples': 98880, 'steps': 514, 'loss/train': 8.17026972770691} 01/27/2022 19:26:52 - INFO - codeparrot_training - Step 515: {'lr': 0.00012875, 'samples': 99072, 'steps': 515, 'loss/train': 8.681805610656738} 01/27/2022 19:26:56 - INFO - codeparrot_training - Step 516: {'lr': 0.00012900000000000002, 'samples': 99264, 'steps': 516, 'loss/train': 8.015620708465576} 01/27/2022 19:27:00 - INFO - codeparrot_training - Step 517: {'lr': 0.00012925, 'samples': 99456, 'steps': 517, 'loss/train': 8.866118431091309} 01/27/2022 19:27:04 - INFO - codeparrot_training - Step 518: {'lr': 0.0001295, 'samples': 99648, 'steps': 518, 'loss/train': 7.207820892333984} 01/27/2022 19:27:09 - INFO - codeparrot_training - Step 519: {'lr': 0.00012975, 'samples': 99840, 'steps': 519, 'loss/train': 8.88998293876648} 01/27/2022 19:27:14 - INFO - codeparrot_training - Step 520: {'lr': 0.00013000000000000002, 'samples': 100032, 'steps': 520, 'loss/train': 8.746418237686157} 01/27/2022 19:27:19 - INFO - codeparrot_training - Step 521: {'lr': 0.00013025, 'samples': 100224, 'steps': 521, 'loss/train': 8.946112632751465} 01/27/2022 19:27:23 - INFO - codeparrot_training - Step 522: {'lr': 0.0001305, 'samples': 100416, 'steps': 522, 'loss/train': 8.354847192764282} 01/27/2022 19:27:27 - INFO - codeparrot_training - Step 523: {'lr': 0.00013075, 'samples': 100608, 'steps': 523, 'loss/train': 8.326250553131104} 01/27/2022 19:27:31 - INFO - codeparrot_training - Step 524: {'lr': 0.000131, 'samples': 100800, 'steps': 524, 'loss/train': 7.511366128921509} 01/27/2022 19:27:36 - INFO - codeparrot_training - Step 525: {'lr': 0.00013125000000000002, 'samples': 100992, 'steps': 525, 'loss/train': 7.890795707702637} 01/27/2022 19:27:40 - INFO - codeparrot_training - Step 526: {'lr': 0.0001315, 'samples': 101184, 'steps': 526, 'loss/train': 8.375786304473877} 01/27/2022 19:27:45 - INFO - codeparrot_training - Step 527: {'lr': 0.00013175, 'samples': 101376, 'steps': 527, 'loss/train': 8.021414995193481} 01/27/2022 19:27:49 - INFO - codeparrot_training - Step 528: {'lr': 0.000132, 'samples': 101568, 'steps': 528, 'loss/train': 7.28090500831604} 01/27/2022 19:27:53 - INFO - codeparrot_training - Step 529: {'lr': 0.00013225000000000002, 'samples': 101760, 'steps': 529, 'loss/train': 8.963454008102417} 01/27/2022 19:27:58 - INFO - codeparrot_training - Step 530: {'lr': 0.00013250000000000002, 'samples': 101952, 'steps': 530, 'loss/train': 8.001948595046997} 01/27/2022 19:28:02 - INFO - codeparrot_training - Step 531: {'lr': 0.00013275, 'samples': 102144, 'steps': 531, 'loss/train': 9.026639699935913} 01/27/2022 19:28:06 - INFO - codeparrot_training - Step 532: {'lr': 0.000133, 'samples': 102336, 'steps': 532, 'loss/train': 8.229212522506714} 01/27/2022 19:28:10 - INFO - codeparrot_training - Step 533: {'lr': 0.00013325, 'samples': 102528, 'steps': 533, 'loss/train': 8.387352705001831} 01/27/2022 19:28:15 - INFO - codeparrot_training - Step 534: {'lr': 0.00013350000000000002, 'samples': 102720, 'steps': 534, 'loss/train': 8.451747179031372} 01/27/2022 19:28:20 - INFO - codeparrot_training - Step 535: {'lr': 0.00013375, 'samples': 102912, 'steps': 535, 'loss/train': 7.593340873718262} 01/27/2022 19:28:24 - INFO - codeparrot_training - Step 536: {'lr': 0.000134, 'samples': 103104, 'steps': 536, 'loss/train': 8.485268354415894} 01/27/2022 19:28:28 - INFO - codeparrot_training - Step 537: {'lr': 0.00013425, 'samples': 103296, 'steps': 537, 'loss/train': 8.370349645614624} 01/27/2022 19:28:32 - INFO - codeparrot_training - Step 538: {'lr': 0.00013450000000000002, 'samples': 103488, 'steps': 538, 'loss/train': 8.465364217758179} 01/27/2022 19:28:36 - INFO - codeparrot_training - Step 539: {'lr': 0.00013475000000000002, 'samples': 103680, 'steps': 539, 'loss/train': 8.298682451248169} 01/27/2022 19:28:42 - INFO - codeparrot_training - Step 540: {'lr': 0.000135, 'samples': 103872, 'steps': 540, 'loss/train': 8.758976697921753} 01/27/2022 19:28:47 - INFO - codeparrot_training - Step 541: {'lr': 0.00013525, 'samples': 104064, 'steps': 541, 'loss/train': 8.58174991607666} 01/27/2022 19:28:51 - INFO - codeparrot_training - Step 542: {'lr': 0.00013550000000000001, 'samples': 104256, 'steps': 542, 'loss/train': 8.730417251586914} 01/27/2022 19:28:55 - INFO - codeparrot_training - Step 543: {'lr': 0.00013575000000000002, 'samples': 104448, 'steps': 543, 'loss/train': 8.091321229934692} 01/27/2022 19:28:59 - INFO - codeparrot_training - Step 544: {'lr': 0.00013600000000000003, 'samples': 104640, 'steps': 544, 'loss/train': 8.639739990234375} 01/27/2022 19:29:04 - INFO - codeparrot_training - Step 545: {'lr': 0.00013625, 'samples': 104832, 'steps': 545, 'loss/train': 8.174670696258545} 01/27/2022 19:29:08 - INFO - codeparrot_training - Step 546: {'lr': 0.0001365, 'samples': 105024, 'steps': 546, 'loss/train': 8.93849802017212} 01/27/2022 19:29:12 - INFO - codeparrot_training - Step 547: {'lr': 0.00013675000000000002, 'samples': 105216, 'steps': 547, 'loss/train': 7.690248012542725} 01/27/2022 19:29:17 - INFO - codeparrot_training - Step 548: {'lr': 0.00013700000000000002, 'samples': 105408, 'steps': 548, 'loss/train': 8.276523113250732} 01/27/2022 19:29:21 - INFO - codeparrot_training - Step 549: {'lr': 0.00013725, 'samples': 105600, 'steps': 549, 'loss/train': 8.273192882537842} 01/27/2022 19:29:26 - INFO - codeparrot_training - Step 550: {'lr': 0.0001375, 'samples': 105792, 'steps': 550, 'loss/train': 8.364015340805054} 01/27/2022 19:29:30 - INFO - codeparrot_training - Step 551: {'lr': 0.00013775000000000001, 'samples': 105984, 'steps': 551, 'loss/train': 9.108482837677002} 01/27/2022 19:29:34 - INFO - codeparrot_training - Step 552: {'lr': 0.00013800000000000002, 'samples': 106176, 'steps': 552, 'loss/train': 8.416153907775879} 01/27/2022 19:29:38 - INFO - codeparrot_training - Step 553: {'lr': 0.00013825000000000003, 'samples': 106368, 'steps': 553, 'loss/train': 8.693472862243652} 01/27/2022 19:29:42 - INFO - codeparrot_training - Step 554: {'lr': 0.0001385, 'samples': 106560, 'steps': 554, 'loss/train': 8.259191751480103} 01/27/2022 19:29:48 - INFO - codeparrot_training - Step 555: {'lr': 0.00013875, 'samples': 106752, 'steps': 555, 'loss/train': 8.928863525390625} 01/27/2022 19:29:52 - INFO - codeparrot_training - Step 556: {'lr': 0.00013900000000000002, 'samples': 106944, 'steps': 556, 'loss/train': 8.77770209312439} 01/27/2022 19:29:57 - INFO - codeparrot_training - Step 557: {'lr': 0.00013925000000000002, 'samples': 107136, 'steps': 557, 'loss/train': 7.560959100723267} 01/27/2022 19:30:01 - INFO - codeparrot_training - Step 558: {'lr': 0.0001395, 'samples': 107328, 'steps': 558, 'loss/train': 8.654487133026123} 01/27/2022 19:30:05 - INFO - codeparrot_training - Step 559: {'lr': 0.00013975, 'samples': 107520, 'steps': 559, 'loss/train': 8.728206396102905} 01/27/2022 19:30:10 - INFO - codeparrot_training - Step 560: {'lr': 0.00014000000000000001, 'samples': 107712, 'steps': 560, 'loss/train': 9.231706380844116} 01/27/2022 19:30:14 - INFO - codeparrot_training - Step 561: {'lr': 0.00014025000000000002, 'samples': 107904, 'steps': 561, 'loss/train': 8.957905769348145} 01/27/2022 19:30:18 - INFO - codeparrot_training - Step 562: {'lr': 0.00014050000000000003, 'samples': 108096, 'steps': 562, 'loss/train': 8.754714488983154} 01/27/2022 19:30:22 - INFO - codeparrot_training - Step 563: {'lr': 0.00014074999999999998, 'samples': 108288, 'steps': 563, 'loss/train': 8.369371891021729} 01/27/2022 19:30:27 - INFO - codeparrot_training - Step 564: {'lr': 0.00014099999999999998, 'samples': 108480, 'steps': 564, 'loss/train': 8.07359004020691} 01/27/2022 19:30:32 - INFO - codeparrot_training - Step 565: {'lr': 0.00014125, 'samples': 108672, 'steps': 565, 'loss/train': 9.256146669387817} 01/27/2022 19:30:36 - INFO - codeparrot_training - Step 566: {'lr': 0.0001415, 'samples': 108864, 'steps': 566, 'loss/train': 4.473942518234253} 01/27/2022 19:30:40 - INFO - codeparrot_training - Step 567: {'lr': 0.00014175, 'samples': 109056, 'steps': 567, 'loss/train': 7.521366119384766} 01/27/2022 19:30:44 - INFO - codeparrot_training - Step 568: {'lr': 0.00014199999999999998, 'samples': 109248, 'steps': 568, 'loss/train': 7.821649789810181} 01/27/2022 19:30:48 - INFO - codeparrot_training - Step 569: {'lr': 0.00014225, 'samples': 109440, 'steps': 569, 'loss/train': 9.155828475952148} 01/27/2022 19:30:54 - INFO - codeparrot_training - Step 570: {'lr': 0.0001425, 'samples': 109632, 'steps': 570, 'loss/train': 8.23076605796814} 01/27/2022 19:30:58 - INFO - codeparrot_training - Step 571: {'lr': 0.00014275, 'samples': 109824, 'steps': 571, 'loss/train': 8.648569107055664} 01/27/2022 19:31:03 - INFO - codeparrot_training - Step 572: {'lr': 0.00014299999999999998, 'samples': 110016, 'steps': 572, 'loss/train': 8.473631858825684} 01/27/2022 19:31:07 - INFO - codeparrot_training - Step 573: {'lr': 0.00014324999999999999, 'samples': 110208, 'steps': 573, 'loss/train': 6.631492853164673} 01/27/2022 19:31:12 - INFO - codeparrot_training - Step 574: {'lr': 0.0001435, 'samples': 110400, 'steps': 574, 'loss/train': 8.640091180801392} 01/27/2022 19:31:16 - INFO - codeparrot_training - Step 575: {'lr': 0.00014375, 'samples': 110592, 'steps': 575, 'loss/train': 8.769315004348755} 01/27/2022 19:31:20 - INFO - codeparrot_training - Step 576: {'lr': 0.000144, 'samples': 110784, 'steps': 576, 'loss/train': 8.279324054718018} 01/27/2022 19:31:25 - INFO - codeparrot_training - Step 577: {'lr': 0.00014424999999999998, 'samples': 110976, 'steps': 577, 'loss/train': 9.597347259521484} 01/27/2022 19:31:29 - INFO - codeparrot_training - Step 578: {'lr': 0.0001445, 'samples': 111168, 'steps': 578, 'loss/train': 8.302575588226318} 01/27/2022 19:31:34 - INFO - codeparrot_training - Step 579: {'lr': 0.00014475, 'samples': 111360, 'steps': 579, 'loss/train': 7.874625205993652} 01/27/2022 19:31:39 - INFO - codeparrot_training - Step 580: {'lr': 0.000145, 'samples': 111552, 'steps': 580, 'loss/train': 8.314260721206665} 01/27/2022 19:31:43 - INFO - codeparrot_training - Step 581: {'lr': 0.00014524999999999998, 'samples': 111744, 'steps': 581, 'loss/train': 8.078325748443604} 01/27/2022 19:31:47 - INFO - codeparrot_training - Step 582: {'lr': 0.00014549999999999999, 'samples': 111936, 'steps': 582, 'loss/train': 8.919520139694214} 01/27/2022 19:31:51 - INFO - codeparrot_training - Step 583: {'lr': 0.00014575, 'samples': 112128, 'steps': 583, 'loss/train': 8.303371667861938} 01/27/2022 19:31:55 - INFO - codeparrot_training - Step 584: {'lr': 0.000146, 'samples': 112320, 'steps': 584, 'loss/train': 7.914836883544922} 01/27/2022 19:32:01 - INFO - codeparrot_training - Step 585: {'lr': 0.00014625, 'samples': 112512, 'steps': 585, 'loss/train': 8.660935878753662} 01/27/2022 19:32:05 - INFO - codeparrot_training - Step 586: {'lr': 0.00014649999999999998, 'samples': 112704, 'steps': 586, 'loss/train': 7.782023906707764} 01/27/2022 19:32:09 - INFO - codeparrot_training - Step 587: {'lr': 0.00014675, 'samples': 112896, 'steps': 587, 'loss/train': 7.973663806915283} 01/27/2022 19:32:13 - INFO - codeparrot_training - Step 588: {'lr': 0.000147, 'samples': 113088, 'steps': 588, 'loss/train': 8.153911828994751} 01/27/2022 19:32:18 - INFO - codeparrot_training - Step 589: {'lr': 0.00014725, 'samples': 113280, 'steps': 589, 'loss/train': 8.781780481338501} 01/27/2022 19:32:23 - INFO - codeparrot_training - Step 590: {'lr': 0.0001475, 'samples': 113472, 'steps': 590, 'loss/train': 8.209387063980103} 01/27/2022 19:32:27 - INFO - codeparrot_training - Step 591: {'lr': 0.00014774999999999999, 'samples': 113664, 'steps': 591, 'loss/train': 7.8359527587890625} 01/27/2022 19:32:31 - INFO - codeparrot_training - Step 592: {'lr': 0.000148, 'samples': 113856, 'steps': 592, 'loss/train': 8.229139566421509} 01/27/2022 19:32:35 - INFO - codeparrot_training - Step 593: {'lr': 0.00014825, 'samples': 114048, 'steps': 593, 'loss/train': 8.467285394668579} 01/27/2022 19:32:40 - INFO - codeparrot_training - Step 594: {'lr': 0.0001485, 'samples': 114240, 'steps': 594, 'loss/train': 7.223497152328491} 01/27/2022 19:32:44 - INFO - codeparrot_training - Step 595: {'lr': 0.00014874999999999998, 'samples': 114432, 'steps': 595, 'loss/train': 8.965587615966797} 01/27/2022 19:32:49 - INFO - codeparrot_training - Step 596: {'lr': 0.000149, 'samples': 114624, 'steps': 596, 'loss/train': 8.973945379257202} 01/27/2022 19:32:53 - INFO - codeparrot_training - Step 597: {'lr': 0.00014925, 'samples': 114816, 'steps': 597, 'loss/train': 8.514985799789429} 01/27/2022 19:32:57 - INFO - codeparrot_training - Step 598: {'lr': 0.0001495, 'samples': 115008, 'steps': 598, 'loss/train': 9.305945634841919} 01/27/2022 19:33:03 - INFO - codeparrot_training - Step 599: {'lr': 0.00014975, 'samples': 115200, 'steps': 599, 'loss/train': 7.214632987976074} 01/27/2022 19:33:07 - INFO - codeparrot_training - Step 600: {'lr': 0.00015, 'samples': 115392, 'steps': 600, 'loss/train': 8.336686849594116} 01/27/2022 19:33:11 - INFO - codeparrot_training - Step 601: {'lr': 0.00015025, 'samples': 115584, 'steps': 601, 'loss/train': 7.75456166267395} 01/27/2022 19:33:15 - INFO - codeparrot_training - Step 602: {'lr': 0.0001505, 'samples': 115776, 'steps': 602, 'loss/train': 8.072105884552002} 01/27/2022 19:33:20 - INFO - codeparrot_training - Step 603: {'lr': 0.00015075, 'samples': 115968, 'steps': 603, 'loss/train': 8.81142282485962} 01/27/2022 19:33:25 - INFO - codeparrot_training - Step 604: {'lr': 0.000151, 'samples': 116160, 'steps': 604, 'loss/train': 7.754583120346069} 01/27/2022 19:33:29 - INFO - codeparrot_training - Step 605: {'lr': 0.00015125, 'samples': 116352, 'steps': 605, 'loss/train': 7.993920564651489} 01/27/2022 19:33:33 - INFO - codeparrot_training - Step 606: {'lr': 0.0001515, 'samples': 116544, 'steps': 606, 'loss/train': 7.54375147819519} 01/27/2022 19:33:37 - INFO - codeparrot_training - Step 607: {'lr': 0.00015175, 'samples': 116736, 'steps': 607, 'loss/train': 8.539509057998657} 01/27/2022 19:33:41 - INFO - codeparrot_training - Step 608: {'lr': 0.000152, 'samples': 116928, 'steps': 608, 'loss/train': 7.788200855255127} 01/27/2022 19:33:46 - INFO - codeparrot_training - Step 609: {'lr': 0.00015225, 'samples': 117120, 'steps': 609, 'loss/train': 8.278229713439941} 01/27/2022 19:33:51 - INFO - codeparrot_training - Step 610: {'lr': 0.0001525, 'samples': 117312, 'steps': 610, 'loss/train': 8.791933536529541} 01/27/2022 19:33:55 - INFO - codeparrot_training - Step 611: {'lr': 0.00015275, 'samples': 117504, 'steps': 611, 'loss/train': 8.514147520065308} 01/27/2022 19:33:59 - INFO - codeparrot_training - Step 612: {'lr': 0.000153, 'samples': 117696, 'steps': 612, 'loss/train': 8.430699348449707} 01/27/2022 19:34:03 - INFO - codeparrot_training - Step 613: {'lr': 0.00015325, 'samples': 117888, 'steps': 613, 'loss/train': 8.370607137680054} 01/27/2022 19:34:09 - INFO - codeparrot_training - Step 614: {'lr': 0.0001535, 'samples': 118080, 'steps': 614, 'loss/train': 8.25789499282837} 01/27/2022 19:34:13 - INFO - codeparrot_training - Step 615: {'lr': 0.00015375, 'samples': 118272, 'steps': 615, 'loss/train': 9.28782606124878} 01/27/2022 19:34:17 - INFO - codeparrot_training - Step 616: {'lr': 0.000154, 'samples': 118464, 'steps': 616, 'loss/train': 7.710235118865967} 01/27/2022 19:34:21 - INFO - codeparrot_training - Step 617: {'lr': 0.00015425, 'samples': 118656, 'steps': 617, 'loss/train': 9.011675834655762} 01/27/2022 19:34:26 - INFO - codeparrot_training - Step 618: {'lr': 0.00015450000000000001, 'samples': 118848, 'steps': 618, 'loss/train': 8.131665229797363} 01/27/2022 19:34:31 - INFO - codeparrot_training - Step 619: {'lr': 0.00015475, 'samples': 119040, 'steps': 619, 'loss/train': 7.657416343688965} 01/27/2022 19:34:35 - INFO - codeparrot_training - Step 620: {'lr': 0.000155, 'samples': 119232, 'steps': 620, 'loss/train': 8.05387258529663} 01/27/2022 19:34:39 - INFO - codeparrot_training - Step 621: {'lr': 0.00015525, 'samples': 119424, 'steps': 621, 'loss/train': 8.192591428756714} 01/27/2022 19:34:43 - INFO - codeparrot_training - Step 622: {'lr': 0.0001555, 'samples': 119616, 'steps': 622, 'loss/train': 8.283280849456787} 01/27/2022 19:34:47 - INFO - codeparrot_training - Step 623: {'lr': 0.00015575, 'samples': 119808, 'steps': 623, 'loss/train': 8.495567321777344} 01/27/2022 19:34:53 - INFO - codeparrot_training - Step 624: {'lr': 0.000156, 'samples': 120000, 'steps': 624, 'loss/train': 8.304754257202148} 01/27/2022 19:34:58 - INFO - codeparrot_training - Step 625: {'lr': 0.00015625, 'samples': 120192, 'steps': 625, 'loss/train': 8.422377347946167} 01/27/2022 19:35:02 - INFO - codeparrot_training - Step 626: {'lr': 0.0001565, 'samples': 120384, 'steps': 626, 'loss/train': 7.985602140426636} 01/27/2022 19:35:06 - INFO - codeparrot_training - Step 627: {'lr': 0.00015675000000000002, 'samples': 120576, 'steps': 627, 'loss/train': 7.165228843688965} 01/27/2022 19:35:10 - INFO - codeparrot_training - Step 628: {'lr': 0.000157, 'samples': 120768, 'steps': 628, 'loss/train': 8.297940015792847} 01/27/2022 19:35:15 - INFO - codeparrot_training - Step 629: {'lr': 0.00015725, 'samples': 120960, 'steps': 629, 'loss/train': 8.375863552093506} 01/27/2022 19:35:19 - INFO - codeparrot_training - Step 630: {'lr': 0.0001575, 'samples': 121152, 'steps': 630, 'loss/train': 9.098508596420288} 01/27/2022 19:35:23 - INFO - codeparrot_training - Step 631: {'lr': 0.00015775, 'samples': 121344, 'steps': 631, 'loss/train': 8.307096004486084} 01/27/2022 19:35:28 - INFO - codeparrot_training - Step 632: {'lr': 0.000158, 'samples': 121536, 'steps': 632, 'loss/train': 7.5440497398376465} 01/27/2022 19:35:32 - INFO - codeparrot_training - Step 633: {'lr': 0.00015825, 'samples': 121728, 'steps': 633, 'loss/train': 8.483729124069214} 01/27/2022 19:35:37 - INFO - codeparrot_training - Step 634: {'lr': 0.0001585, 'samples': 121920, 'steps': 634, 'loss/train': 8.692340612411499} 01/27/2022 19:35:41 - INFO - codeparrot_training - Step 635: {'lr': 0.00015875, 'samples': 122112, 'steps': 635, 'loss/train': 8.061896324157715} 01/27/2022 19:35:45 - INFO - codeparrot_training - Step 636: {'lr': 0.00015900000000000002, 'samples': 122304, 'steps': 636, 'loss/train': 8.079899311065674} 01/27/2022 19:35:49 - INFO - codeparrot_training - Step 637: {'lr': 0.00015925, 'samples': 122496, 'steps': 637, 'loss/train': 7.544781446456909} 01/27/2022 19:35:53 - INFO - codeparrot_training - Step 638: {'lr': 0.0001595, 'samples': 122688, 'steps': 638, 'loss/train': 9.011221647262573} 01/27/2022 19:35:59 - INFO - codeparrot_training - Step 639: {'lr': 0.00015975, 'samples': 122880, 'steps': 639, 'loss/train': 8.249982833862305} 01/27/2022 19:36:03 - INFO - codeparrot_training - Step 640: {'lr': 0.00016, 'samples': 123072, 'steps': 640, 'loss/train': 7.240501642227173} 01/27/2022 19:36:07 - INFO - codeparrot_training - Step 641: {'lr': 0.00016025000000000002, 'samples': 123264, 'steps': 641, 'loss/train': 8.318994998931885} 01/27/2022 19:36:12 - INFO - codeparrot_training - Step 642: {'lr': 0.0001605, 'samples': 123456, 'steps': 642, 'loss/train': 8.296253442764282} 01/27/2022 19:36:16 - INFO - codeparrot_training - Step 643: {'lr': 0.00016075, 'samples': 123648, 'steps': 643, 'loss/train': 9.265507936477661} 01/27/2022 19:36:21 - INFO - codeparrot_training - Step 644: {'lr': 0.000161, 'samples': 123840, 'steps': 644, 'loss/train': 8.830106019973755} 01/27/2022 19:36:25 - INFO - codeparrot_training - Step 645: {'lr': 0.00016125000000000002, 'samples': 124032, 'steps': 645, 'loss/train': 9.4315767288208} 01/27/2022 19:36:29 - INFO - codeparrot_training - Step 646: {'lr': 0.0001615, 'samples': 124224, 'steps': 646, 'loss/train': 5.333669543266296} 01/27/2022 19:36:33 - INFO - codeparrot_training - Step 647: {'lr': 0.00016175, 'samples': 124416, 'steps': 647, 'loss/train': 7.553824424743652} 01/27/2022 19:36:38 - INFO - codeparrot_training - Step 648: {'lr': 0.000162, 'samples': 124608, 'steps': 648, 'loss/train': 9.619696140289307} 01/27/2022 19:36:43 - INFO - codeparrot_training - Step 649: {'lr': 0.00016225000000000001, 'samples': 124800, 'steps': 649, 'loss/train': 8.453848600387573} 01/27/2022 19:36:47 - INFO - codeparrot_training - Step 650: {'lr': 0.00016250000000000002, 'samples': 124992, 'steps': 650, 'loss/train': 7.417596101760864} 01/27/2022 19:36:51 - INFO - codeparrot_training - Step 651: {'lr': 0.00016275, 'samples': 125184, 'steps': 651, 'loss/train': 7.024638891220093} 01/27/2022 19:36:55 - INFO - codeparrot_training - Step 652: {'lr': 0.000163, 'samples': 125376, 'steps': 652, 'loss/train': 8.656749486923218} 01/27/2022 19:37:00 - INFO - codeparrot_training - Step 653: {'lr': 0.00016325, 'samples': 125568, 'steps': 653, 'loss/train': 8.449368953704834} 01/27/2022 19:37:06 - INFO - codeparrot_training - Step 654: {'lr': 0.00016350000000000002, 'samples': 125760, 'steps': 654, 'loss/train': 8.034994125366211} 01/27/2022 19:37:10 - INFO - codeparrot_training - Step 655: {'lr': 0.00016375000000000002, 'samples': 125952, 'steps': 655, 'loss/train': 9.132595539093018} 01/27/2022 19:37:14 - INFO - codeparrot_training - Step 656: {'lr': 0.000164, 'samples': 126144, 'steps': 656, 'loss/train': 7.8157031536102295} 01/27/2022 19:37:18 - INFO - codeparrot_training - Step 657: {'lr': 0.00016425, 'samples': 126336, 'steps': 657, 'loss/train': 9.362704038619995} 01/27/2022 19:37:22 - INFO - codeparrot_training - Step 658: {'lr': 0.00016450000000000001, 'samples': 126528, 'steps': 658, 'loss/train': 8.65830659866333} 01/27/2022 19:37:27 - INFO - codeparrot_training - Step 659: {'lr': 0.00016475000000000002, 'samples': 126720, 'steps': 659, 'loss/train': 7.5368242263793945} 01/27/2022 19:37:31 - INFO - codeparrot_training - Step 660: {'lr': 0.000165, 'samples': 126912, 'steps': 660, 'loss/train': 7.6077353954315186} 01/27/2022 19:37:36 - INFO - codeparrot_training - Step 661: {'lr': 0.00016525, 'samples': 127104, 'steps': 661, 'loss/train': 7.764186143875122} 01/27/2022 19:37:40 - INFO - codeparrot_training - Step 662: {'lr': 0.0001655, 'samples': 127296, 'steps': 662, 'loss/train': 8.5371572971344} 01/27/2022 19:37:44 - INFO - codeparrot_training - Step 663: {'lr': 0.00016575000000000002, 'samples': 127488, 'steps': 663, 'loss/train': 8.563830614089966} 01/27/2022 19:37:49 - INFO - codeparrot_training - Step 664: {'lr': 0.00016600000000000002, 'samples': 127680, 'steps': 664, 'loss/train': 8.0364511013031} 01/27/2022 19:37:53 - INFO - codeparrot_training - Step 665: {'lr': 0.00016625, 'samples': 127872, 'steps': 665, 'loss/train': 7.877760887145996} 01/27/2022 19:37:57 - INFO - codeparrot_training - Step 666: {'lr': 0.0001665, 'samples': 128064, 'steps': 666, 'loss/train': 8.593442916870117} 01/27/2022 19:38:01 - INFO - codeparrot_training - Step 667: {'lr': 0.00016675000000000001, 'samples': 128256, 'steps': 667, 'loss/train': 7.1693115234375} 01/27/2022 19:38:07 - INFO - codeparrot_training - Step 668: {'lr': 0.00016700000000000002, 'samples': 128448, 'steps': 668, 'loss/train': 7.933293342590332} 01/27/2022 19:38:11 - INFO - codeparrot_training - Step 669: {'lr': 0.00016725000000000003, 'samples': 128640, 'steps': 669, 'loss/train': 7.103787660598755} 01/27/2022 19:38:15 - INFO - codeparrot_training - Step 670: {'lr': 0.0001675, 'samples': 128832, 'steps': 670, 'loss/train': 8.280439138412476} 01/27/2022 19:38:20 - INFO - codeparrot_training - Step 671: {'lr': 0.00016775, 'samples': 129024, 'steps': 671, 'loss/train': 8.499544858932495} 01/27/2022 19:38:24 - INFO - codeparrot_training - Step 672: {'lr': 0.00016800000000000002, 'samples': 129216, 'steps': 672, 'loss/train': 8.899912118911743} 01/27/2022 19:38:29 - INFO - codeparrot_training - Step 673: {'lr': 0.00016825000000000002, 'samples': 129408, 'steps': 673, 'loss/train': 8.211901187896729} 01/27/2022 19:38:33 - INFO - codeparrot_training - Step 674: {'lr': 0.0001685, 'samples': 129600, 'steps': 674, 'loss/train': 8.731487274169922} 01/27/2022 19:38:37 - INFO - codeparrot_training - Step 675: {'lr': 0.00016875, 'samples': 129792, 'steps': 675, 'loss/train': 6.801019191741943} 01/27/2022 19:38:42 - INFO - codeparrot_training - Step 676: {'lr': 0.00016900000000000002, 'samples': 129984, 'steps': 676, 'loss/train': 7.723647594451904} 01/27/2022 19:38:46 - INFO - codeparrot_training - Step 677: {'lr': 0.00016925000000000002, 'samples': 130176, 'steps': 677, 'loss/train': 7.8934242725372314} 01/27/2022 19:38:51 - INFO - codeparrot_training - Step 678: {'lr': 0.00016950000000000003, 'samples': 130368, 'steps': 678, 'loss/train': 7.5493855476379395} 01/27/2022 19:38:55 - INFO - codeparrot_training - Step 679: {'lr': 0.00016975, 'samples': 130560, 'steps': 679, 'loss/train': 8.79888367652893} 01/27/2022 19:38:59 - INFO - codeparrot_training - Step 680: {'lr': 0.00017, 'samples': 130752, 'steps': 680, 'loss/train': 7.605967283248901} 01/27/2022 19:39:03 - INFO - codeparrot_training - Step 681: {'lr': 0.00017025000000000002, 'samples': 130944, 'steps': 681, 'loss/train': 5.154707193374634} 01/27/2022 19:39:08 - INFO - codeparrot_training - Step 682: {'lr': 0.00017050000000000002, 'samples': 131136, 'steps': 682, 'loss/train': 8.063627243041992} 01/27/2022 19:39:13 - INFO - codeparrot_training - Step 683: {'lr': 0.00017075, 'samples': 131328, 'steps': 683, 'loss/train': 8.482172012329102} 01/27/2022 19:39:18 - INFO - codeparrot_training - Step 684: {'lr': 0.000171, 'samples': 131520, 'steps': 684, 'loss/train': 8.070060968399048} 01/27/2022 19:39:22 - INFO - codeparrot_training - Step 685: {'lr': 0.00017125000000000002, 'samples': 131712, 'steps': 685, 'loss/train': 4.122437953948975} 01/27/2022 19:39:26 - INFO - codeparrot_training - Step 686: {'lr': 0.00017150000000000002, 'samples': 131904, 'steps': 686, 'loss/train': 8.9731764793396} 01/27/2022 19:39:30 - INFO - codeparrot_training - Step 687: {'lr': 0.00017175000000000003, 'samples': 132096, 'steps': 687, 'loss/train': 8.004594326019287} 01/27/2022 19:39:35 - INFO - codeparrot_training - Step 688: {'lr': 0.00017199999999999998, 'samples': 132288, 'steps': 688, 'loss/train': 7.945577144622803} 01/27/2022 19:39:39 - INFO - codeparrot_training - Step 689: {'lr': 0.00017224999999999999, 'samples': 132480, 'steps': 689, 'loss/train': 8.377017974853516} 01/27/2022 19:39:44 - INFO - codeparrot_training - Step 690: {'lr': 0.0001725, 'samples': 132672, 'steps': 690, 'loss/train': 8.469246625900269} 01/27/2022 19:39:48 - INFO - codeparrot_training - Step 691: {'lr': 0.00017275, 'samples': 132864, 'steps': 691, 'loss/train': 8.472698450088501} 01/27/2022 19:39:52 - INFO - codeparrot_training - Step 692: {'lr': 0.000173, 'samples': 133056, 'steps': 692, 'loss/train': 8.58025074005127} 01/27/2022 19:39:58 - INFO - codeparrot_training - Step 693: {'lr': 0.00017324999999999998, 'samples': 133248, 'steps': 693, 'loss/train': 8.043803215026855} 01/27/2022 19:40:02 - INFO - codeparrot_training - Step 694: {'lr': 0.0001735, 'samples': 133440, 'steps': 694, 'loss/train': 7.3625078201293945} 01/27/2022 19:40:06 - INFO - codeparrot_training - Step 695: {'lr': 0.00017375, 'samples': 133632, 'steps': 695, 'loss/train': 8.26303482055664} 01/27/2022 19:40:10 - INFO - codeparrot_training - Step 696: {'lr': 0.000174, 'samples': 133824, 'steps': 696, 'loss/train': 7.4576239585876465} 01/27/2022 19:40:14 - INFO - codeparrot_training - Step 697: {'lr': 0.00017424999999999998, 'samples': 134016, 'steps': 697, 'loss/train': 8.191519975662231} 01/27/2022 19:40:19 - INFO - codeparrot_training - Step 698: {'lr': 0.00017449999999999999, 'samples': 134208, 'steps': 698, 'loss/train': 7.1545493602752686} 01/27/2022 19:40:23 - INFO - codeparrot_training - Step 699: {'lr': 0.00017475, 'samples': 134400, 'steps': 699, 'loss/train': 8.097326517105103} 01/27/2022 19:40:28 - INFO - codeparrot_training - Step 700: {'lr': 0.000175, 'samples': 134592, 'steps': 700, 'loss/train': 7.546249151229858} 01/27/2022 19:40:32 - INFO - codeparrot_training - Step 701: {'lr': 0.00017525, 'samples': 134784, 'steps': 701, 'loss/train': 8.151185989379883} 01/27/2022 19:40:36 - INFO - codeparrot_training - Step 702: {'lr': 0.00017549999999999998, 'samples': 134976, 'steps': 702, 'loss/train': 7.952397108078003} 01/27/2022 19:40:41 - INFO - codeparrot_training - Step 703: {'lr': 0.00017575, 'samples': 135168, 'steps': 703, 'loss/train': 8.549700736999512} 01/27/2022 19:40:45 - INFO - codeparrot_training - Step 704: {'lr': 0.000176, 'samples': 135360, 'steps': 704, 'loss/train': 8.339847564697266} 01/27/2022 19:40:49 - INFO - codeparrot_training - Step 705: {'lr': 0.00017625, 'samples': 135552, 'steps': 705, 'loss/train': 7.991495847702026} 01/27/2022 19:40:54 - INFO - codeparrot_training - Step 706: {'lr': 0.00017649999999999998, 'samples': 135744, 'steps': 706, 'loss/train': 7.92998743057251} 01/27/2022 19:40:58 - INFO - codeparrot_training - Step 707: {'lr': 0.00017675, 'samples': 135936, 'steps': 707, 'loss/train': 8.14457631111145} 01/27/2022 19:41:03 - INFO - codeparrot_training - Step 708: {'lr': 0.000177, 'samples': 136128, 'steps': 708, 'loss/train': 7.857103586196899} 01/27/2022 19:41:08 - INFO - codeparrot_training - Step 709: {'lr': 0.00017725, 'samples': 136320, 'steps': 709, 'loss/train': 8.489074230194092} 01/27/2022 19:41:12 - INFO - codeparrot_training - Step 710: {'lr': 0.0001775, 'samples': 136512, 'steps': 710, 'loss/train': 7.297561883926392} 01/27/2022 19:41:16 - INFO - codeparrot_training - Step 711: {'lr': 0.00017774999999999998, 'samples': 136704, 'steps': 711, 'loss/train': 7.275579929351807} 01/27/2022 19:41:20 - INFO - codeparrot_training - Step 712: {'lr': 0.000178, 'samples': 136896, 'steps': 712, 'loss/train': 7.603190660476685} 01/27/2022 19:41:25 - INFO - codeparrot_training - Step 713: {'lr': 0.00017825, 'samples': 137088, 'steps': 713, 'loss/train': 7.588614463806152} 01/27/2022 19:41:29 - INFO - codeparrot_training - Step 714: {'lr': 0.0001785, 'samples': 137280, 'steps': 714, 'loss/train': 7.02622389793396} 01/27/2022 19:41:33 - INFO - codeparrot_training - Step 715: {'lr': 0.00017875, 'samples': 137472, 'steps': 715, 'loss/train': 7.690698623657227} 01/27/2022 19:41:38 - INFO - codeparrot_training - Step 716: {'lr': 0.000179, 'samples': 137664, 'steps': 716, 'loss/train': 7.861556768417358} 01/27/2022 19:41:42 - INFO - codeparrot_training - Step 717: {'lr': 0.00017925, 'samples': 137856, 'steps': 717, 'loss/train': 7.842064619064331} 01/27/2022 19:41:47 - INFO - codeparrot_training - Step 718: {'lr': 0.0001795, 'samples': 138048, 'steps': 718, 'loss/train': 8.416612386703491} 01/27/2022 19:41:51 - INFO - codeparrot_training - Step 719: {'lr': 0.00017975, 'samples': 138240, 'steps': 719, 'loss/train': 7.464595556259155} 01/27/2022 19:41:55 - INFO - codeparrot_training - Step 720: {'lr': 0.00017999999999999998, 'samples': 138432, 'steps': 720, 'loss/train': 7.971174716949463} 01/27/2022 19:41:59 - INFO - codeparrot_training - Step 721: {'lr': 0.00018025, 'samples': 138624, 'steps': 721, 'loss/train': 7.8818700313568115} 01/27/2022 19:42:04 - INFO - codeparrot_training - Step 722: {'lr': 0.0001805, 'samples': 138816, 'steps': 722, 'loss/train': 7.1598801612854} 01/27/2022 19:42:09 - INFO - codeparrot_training - Step 723: {'lr': 0.00018075, 'samples': 139008, 'steps': 723, 'loss/train': 7.922802686691284} 01/27/2022 19:42:13 - INFO - codeparrot_training - Step 724: {'lr': 0.000181, 'samples': 139200, 'steps': 724, 'loss/train': 7.660142183303833} 01/27/2022 19:42:17 - INFO - codeparrot_training - Step 725: {'lr': 0.00018125, 'samples': 139392, 'steps': 725, 'loss/train': 7.76464319229126} 01/27/2022 19:42:21 - INFO - codeparrot_training - Step 726: {'lr': 0.0001815, 'samples': 139584, 'steps': 726, 'loss/train': 7.864576578140259} 01/27/2022 19:42:25 - INFO - codeparrot_training - Step 727: {'lr': 0.00018175, 'samples': 139776, 'steps': 727, 'loss/train': 7.511577129364014} 01/27/2022 19:42:32 - INFO - codeparrot_training - Step 728: {'lr': 0.000182, 'samples': 139968, 'steps': 728, 'loss/train': 7.740895986557007} 01/27/2022 19:42:36 - INFO - codeparrot_training - Step 729: {'lr': 0.00018225, 'samples': 140160, 'steps': 729, 'loss/train': 8.064425468444824} 01/27/2022 19:42:41 - INFO - codeparrot_training - Step 730: {'lr': 0.0001825, 'samples': 140352, 'steps': 730, 'loss/train': 8.026722192764282} 01/27/2022 19:42:45 - INFO - codeparrot_training - Step 731: {'lr': 0.00018275, 'samples': 140544, 'steps': 731, 'loss/train': 4.874718189239502} 01/27/2022 19:42:49 - INFO - codeparrot_training - Step 732: {'lr': 0.000183, 'samples': 140736, 'steps': 732, 'loss/train': 5.072027206420898} 01/27/2022 19:42:54 - INFO - codeparrot_training - Step 733: {'lr': 0.00018325, 'samples': 140928, 'steps': 733, 'loss/train': 7.2466349601745605} 01/27/2022 19:42:58 - INFO - codeparrot_training - Step 734: {'lr': 0.0001835, 'samples': 141120, 'steps': 734, 'loss/train': 8.898398637771606} 01/27/2022 19:43:02 - INFO - codeparrot_training - Step 735: {'lr': 0.00018375, 'samples': 141312, 'steps': 735, 'loss/train': 8.268721103668213} 01/27/2022 19:43:06 - INFO - codeparrot_training - Step 736: {'lr': 0.000184, 'samples': 141504, 'steps': 736, 'loss/train': 8.143299579620361} 01/27/2022 19:43:11 - INFO - codeparrot_training - Step 737: {'lr': 0.00018425, 'samples': 141696, 'steps': 737, 'loss/train': 7.380795478820801} 01/27/2022 19:43:16 - INFO - codeparrot_training - Step 738: {'lr': 0.0001845, 'samples': 141888, 'steps': 738, 'loss/train': 7.440507173538208} 01/27/2022 19:43:21 - INFO - codeparrot_training - Step 739: {'lr': 0.00018475, 'samples': 142080, 'steps': 739, 'loss/train': 7.575955152511597} 01/27/2022 19:43:25 - INFO - codeparrot_training - Step 740: {'lr': 0.000185, 'samples': 142272, 'steps': 740, 'loss/train': 8.205970287322998} 01/27/2022 19:43:29 - INFO - codeparrot_training - Step 741: {'lr': 0.00018525, 'samples': 142464, 'steps': 741, 'loss/train': 8.263526916503906} 01/27/2022 19:43:33 - INFO - codeparrot_training - Step 742: {'lr': 0.0001855, 'samples': 142656, 'steps': 742, 'loss/train': 6.766265630722046} 01/27/2022 19:43:38 - INFO - codeparrot_training - Step 743: {'lr': 0.00018575000000000002, 'samples': 142848, 'steps': 743, 'loss/train': 5.876844048500061} 01/27/2022 19:43:43 - INFO - codeparrot_training - Step 744: {'lr': 0.000186, 'samples': 143040, 'steps': 744, 'loss/train': 5.536075830459595} 01/27/2022 19:43:47 - INFO - codeparrot_training - Step 745: {'lr': 0.00018625, 'samples': 143232, 'steps': 745, 'loss/train': 7.311076641082764} 01/27/2022 19:43:51 - INFO - codeparrot_training - Step 746: {'lr': 0.0001865, 'samples': 143424, 'steps': 746, 'loss/train': 8.380584955215454} 01/27/2022 19:43:55 - INFO - codeparrot_training - Step 747: {'lr': 0.00018675, 'samples': 143616, 'steps': 747, 'loss/train': 8.695755243301392} 01/27/2022 19:44:00 - INFO - codeparrot_training - Step 748: {'lr': 0.000187, 'samples': 143808, 'steps': 748, 'loss/train': 6.71664834022522} 01/27/2022 19:44:04 - INFO - codeparrot_training - Step 749: {'lr': 0.00018725, 'samples': 144000, 'steps': 749, 'loss/train': 8.109562397003174} 01/27/2022 19:44:09 - INFO - codeparrot_training - Step 750: {'lr': 0.0001875, 'samples': 144192, 'steps': 750, 'loss/train': 5.441789388656616} 01/27/2022 19:44:13 - INFO - codeparrot_training - Step 751: {'lr': 0.00018775, 'samples': 144384, 'steps': 751, 'loss/train': 7.893102407455444} 01/27/2022 19:44:17 - INFO - codeparrot_training - Step 752: {'lr': 0.00018800000000000002, 'samples': 144576, 'steps': 752, 'loss/train': 8.077803611755371} 01/27/2022 19:44:23 - INFO - codeparrot_training - Step 753: {'lr': 0.00018825, 'samples': 144768, 'steps': 753, 'loss/train': 8.374654054641724} 01/27/2022 19:44:27 - INFO - codeparrot_training - Step 754: {'lr': 0.0001885, 'samples': 144960, 'steps': 754, 'loss/train': 7.729047775268555} 01/27/2022 19:44:31 - INFO - codeparrot_training - Step 755: {'lr': 0.00018875, 'samples': 145152, 'steps': 755, 'loss/train': 7.703528165817261} 01/27/2022 19:44:35 - INFO - codeparrot_training - Step 756: {'lr': 0.000189, 'samples': 145344, 'steps': 756, 'loss/train': 7.928225755691528} 01/27/2022 19:44:39 - INFO - codeparrot_training - Step 757: {'lr': 0.00018925, 'samples': 145536, 'steps': 757, 'loss/train': 7.509890556335449} 01/27/2022 19:44:44 - INFO - codeparrot_training - Step 758: {'lr': 0.0001895, 'samples': 145728, 'steps': 758, 'loss/train': 7.092700481414795} 01/27/2022 19:44:48 - INFO - codeparrot_training - Step 759: {'lr': 0.00018975, 'samples': 145920, 'steps': 759, 'loss/train': 6.214686155319214} 01/27/2022 19:44:53 - INFO - codeparrot_training - Step 760: {'lr': 0.00019, 'samples': 146112, 'steps': 760, 'loss/train': 8.178486585617065} 01/27/2022 19:44:57 - INFO - codeparrot_training - Step 761: {'lr': 0.00019025000000000002, 'samples': 146304, 'steps': 761, 'loss/train': 6.458388090133667} 01/27/2022 19:45:01 - INFO - codeparrot_training - Step 762: {'lr': 0.0001905, 'samples': 146496, 'steps': 762, 'loss/train': 7.590346097946167} 01/27/2022 19:45:06 - INFO - codeparrot_training - Step 763: {'lr': 0.00019075, 'samples': 146688, 'steps': 763, 'loss/train': 7.043601751327515} 01/27/2022 19:45:10 - INFO - codeparrot_training - Step 764: {'lr': 0.000191, 'samples': 146880, 'steps': 764, 'loss/train': 7.580701589584351} 01/27/2022 19:45:15 - INFO - codeparrot_training - Step 765: {'lr': 0.00019125000000000001, 'samples': 147072, 'steps': 765, 'loss/train': 7.34462571144104} 01/27/2022 19:45:19 - INFO - codeparrot_training - Step 766: {'lr': 0.00019150000000000002, 'samples': 147264, 'steps': 766, 'loss/train': 6.2960364818573} 01/27/2022 19:45:23 - INFO - codeparrot_training - Step 767: {'lr': 0.00019175, 'samples': 147456, 'steps': 767, 'loss/train': 7.760506868362427} 01/27/2022 19:45:28 - INFO - codeparrot_training - Step 768: {'lr': 0.000192, 'samples': 147648, 'steps': 768, 'loss/train': 7.030592679977417} 01/27/2022 19:45:33 - INFO - codeparrot_training - Step 769: {'lr': 0.00019225, 'samples': 147840, 'steps': 769, 'loss/train': 8.43232011795044} 01/27/2022 19:45:37 - INFO - codeparrot_training - Step 770: {'lr': 0.00019250000000000002, 'samples': 148032, 'steps': 770, 'loss/train': 7.337213516235352} 01/27/2022 19:45:41 - INFO - codeparrot_training - Step 771: {'lr': 0.00019275, 'samples': 148224, 'steps': 771, 'loss/train': 6.487716436386108} 01/27/2022 19:45:45 - INFO - codeparrot_training - Step 772: {'lr': 0.000193, 'samples': 148416, 'steps': 772, 'loss/train': 7.263144493103027} 01/27/2022 19:45:51 - INFO - codeparrot_training - Step 773: {'lr': 0.00019325, 'samples': 148608, 'steps': 773, 'loss/train': 4.697204232215881} 01/27/2022 19:45:55 - INFO - codeparrot_training - Step 774: {'lr': 0.00019350000000000001, 'samples': 148800, 'steps': 774, 'loss/train': 7.143262624740601} 01/27/2022 19:46:00 - INFO - codeparrot_training - Step 775: {'lr': 0.00019375000000000002, 'samples': 148992, 'steps': 775, 'loss/train': 7.7516984939575195} 01/27/2022 19:46:04 - INFO - codeparrot_training - Step 776: {'lr': 0.000194, 'samples': 149184, 'steps': 776, 'loss/train': 7.434524059295654} 01/27/2022 19:46:09 - INFO - codeparrot_training - Step 777: {'lr': 0.00019425, 'samples': 149376, 'steps': 777, 'loss/train': 7.52365779876709} 01/27/2022 19:46:13 - INFO - codeparrot_training - Step 778: {'lr': 0.0001945, 'samples': 149568, 'steps': 778, 'loss/train': 7.154605865478516} 01/27/2022 19:46:17 - INFO - codeparrot_training - Step 779: {'lr': 0.00019475000000000002, 'samples': 149760, 'steps': 779, 'loss/train': 7.896199464797974} 01/27/2022 19:46:21 - INFO - codeparrot_training - Step 780: {'lr': 0.00019500000000000002, 'samples': 149952, 'steps': 780, 'loss/train': 7.540104389190674} 01/27/2022 19:46:25 - INFO - codeparrot_training - Step 781: {'lr': 0.00019525, 'samples': 150144, 'steps': 781, 'loss/train': 8.11072325706482} 01/27/2022 19:46:31 - INFO - codeparrot_training - Step 782: {'lr': 0.0001955, 'samples': 150336, 'steps': 782, 'loss/train': 7.418931484222412} 01/27/2022 19:46:35 - INFO - codeparrot_training - Step 783: {'lr': 0.00019575000000000001, 'samples': 150528, 'steps': 783, 'loss/train': 6.519766330718994} 01/27/2022 19:46:40 - INFO - codeparrot_training - Step 784: {'lr': 0.00019600000000000002, 'samples': 150720, 'steps': 784, 'loss/train': 8.003017902374268} 01/27/2022 19:46:44 - INFO - codeparrot_training - Step 785: {'lr': 0.00019625, 'samples': 150912, 'steps': 785, 'loss/train': 7.556897878646851} 01/27/2022 19:46:48 - INFO - codeparrot_training - Step 786: {'lr': 0.0001965, 'samples': 151104, 'steps': 786, 'loss/train': 7.598862648010254} 01/27/2022 19:46:53 - INFO - codeparrot_training - Step 787: {'lr': 0.00019675, 'samples': 151296, 'steps': 787, 'loss/train': 8.38918948173523} 01/27/2022 19:46:57 - INFO - codeparrot_training - Step 788: {'lr': 0.00019700000000000002, 'samples': 151488, 'steps': 788, 'loss/train': 7.555492401123047} 01/27/2022 19:47:01 - INFO - codeparrot_training - Step 789: {'lr': 0.00019725000000000002, 'samples': 151680, 'steps': 789, 'loss/train': 7.571065664291382} 01/27/2022 19:47:05 - INFO - codeparrot_training - Step 790: {'lr': 0.0001975, 'samples': 151872, 'steps': 790, 'loss/train': 7.973852634429932} 01/27/2022 19:47:09 - INFO - codeparrot_training - Step 791: {'lr': 0.00019775, 'samples': 152064, 'steps': 791, 'loss/train': 7.589966297149658} 01/27/2022 19:47:15 - INFO - codeparrot_training - Step 792: {'lr': 0.00019800000000000002, 'samples': 152256, 'steps': 792, 'loss/train': 8.019420146942139} 01/27/2022 19:47:19 - INFO - codeparrot_training - Step 793: {'lr': 0.00019825000000000002, 'samples': 152448, 'steps': 793, 'loss/train': 3.786741614341736} 01/27/2022 19:47:23 - INFO - codeparrot_training - Step 794: {'lr': 0.00019850000000000003, 'samples': 152640, 'steps': 794, 'loss/train': 7.3002705574035645} 01/27/2022 19:47:27 - INFO - codeparrot_training - Step 795: {'lr': 0.00019875, 'samples': 152832, 'steps': 795, 'loss/train': 7.697554349899292} 01/27/2022 19:47:31 - INFO - codeparrot_training - Step 796: {'lr': 0.000199, 'samples': 153024, 'steps': 796, 'loss/train': 7.322354793548584} 01/27/2022 19:47:37 - INFO - codeparrot_training - Step 797: {'lr': 0.00019925000000000002, 'samples': 153216, 'steps': 797, 'loss/train': 7.714418649673462} 01/27/2022 19:47:41 - INFO - codeparrot_training - Step 798: {'lr': 0.00019950000000000002, 'samples': 153408, 'steps': 798, 'loss/train': 6.932586908340454} 01/27/2022 19:47:46 - INFO - codeparrot_training - Step 799: {'lr': 0.00019975, 'samples': 153600, 'steps': 799, 'loss/train': 7.1686248779296875} 01/27/2022 19:47:50 - INFO - codeparrot_training - Step 800: {'lr': 0.0002, 'samples': 153792, 'steps': 800, 'loss/train': 7.482051372528076} 01/27/2022 19:47:54 - INFO - codeparrot_training - Step 801: {'lr': 0.00020025000000000002, 'samples': 153984, 'steps': 801, 'loss/train': 9.389641284942627} 01/27/2022 19:47:59 - INFO - codeparrot_training - Step 802: {'lr': 0.00020050000000000002, 'samples': 154176, 'steps': 802, 'loss/train': 7.458050966262817} 01/27/2022 19:48:03 - INFO - codeparrot_training - Step 803: {'lr': 0.00020075000000000003, 'samples': 154368, 'steps': 803, 'loss/train': 7.7519567012786865} 01/27/2022 19:48:08 - INFO - codeparrot_training - Step 804: {'lr': 0.000201, 'samples': 154560, 'steps': 804, 'loss/train': 7.618620157241821} 01/27/2022 19:48:12 - INFO - codeparrot_training - Step 805: {'lr': 0.00020125, 'samples': 154752, 'steps': 805, 'loss/train': 8.397203207015991} 01/27/2022 19:48:16 - INFO - codeparrot_training - Step 806: {'lr': 0.00020150000000000002, 'samples': 154944, 'steps': 806, 'loss/train': 7.557582378387451} 01/27/2022 19:48:21 - INFO - codeparrot_training - Step 807: {'lr': 0.00020175000000000003, 'samples': 155136, 'steps': 807, 'loss/train': 7.578557252883911} 01/27/2022 19:48:25 - INFO - codeparrot_training - Step 808: {'lr': 0.000202, 'samples': 155328, 'steps': 808, 'loss/train': 7.775289058685303} 01/27/2022 19:48:29 - INFO - codeparrot_training - Step 809: {'lr': 0.00020225, 'samples': 155520, 'steps': 809, 'loss/train': 5.857353329658508} 01/27/2022 19:48:33 - INFO - codeparrot_training - Step 810: {'lr': 0.00020250000000000002, 'samples': 155712, 'steps': 810, 'loss/train': 7.948709964752197} 01/27/2022 19:48:38 - INFO - codeparrot_training - Step 811: {'lr': 0.00020275000000000002, 'samples': 155904, 'steps': 811, 'loss/train': 8.424277782440186} 01/27/2022 19:48:43 - INFO - codeparrot_training - Step 812: {'lr': 0.00020300000000000003, 'samples': 156096, 'steps': 812, 'loss/train': 7.227815866470337} 01/27/2022 19:48:47 - INFO - codeparrot_training - Step 813: {'lr': 0.00020324999999999998, 'samples': 156288, 'steps': 813, 'loss/train': 7.635960102081299} 01/27/2022 19:48:52 - INFO - codeparrot_training - Step 814: {'lr': 0.00020349999999999999, 'samples': 156480, 'steps': 814, 'loss/train': 7.400285482406616} 01/27/2022 19:48:56 - INFO - codeparrot_training - Step 815: {'lr': 0.00020375, 'samples': 156672, 'steps': 815, 'loss/train': 7.3520286083221436} 01/27/2022 19:49:00 - INFO - codeparrot_training - Step 816: {'lr': 0.000204, 'samples': 156864, 'steps': 816, 'loss/train': 7.535558223724365} 01/27/2022 19:49:05 - INFO - codeparrot_training - Step 817: {'lr': 0.00020425, 'samples': 157056, 'steps': 817, 'loss/train': 7.880458116531372} 01/27/2022 19:49:09 - INFO - codeparrot_training - Step 818: {'lr': 0.00020449999999999998, 'samples': 157248, 'steps': 818, 'loss/train': 6.829286813735962} 01/27/2022 19:49:13 - INFO - codeparrot_training - Step 819: {'lr': 0.00020475, 'samples': 157440, 'steps': 819, 'loss/train': 8.716244459152222} 01/27/2022 19:49:18 - INFO - codeparrot_training - Step 820: {'lr': 0.000205, 'samples': 157632, 'steps': 820, 'loss/train': 7.438692569732666} 01/27/2022 19:49:22 - INFO - codeparrot_training - Step 821: {'lr': 0.00020525, 'samples': 157824, 'steps': 821, 'loss/train': 7.752033233642578} 01/27/2022 19:49:27 - INFO - codeparrot_training - Step 822: {'lr': 0.00020549999999999998, 'samples': 158016, 'steps': 822, 'loss/train': 7.9090189933776855} 01/27/2022 19:49:31 - INFO - codeparrot_training - Step 823: {'lr': 0.00020575, 'samples': 158208, 'steps': 823, 'loss/train': 6.93665885925293} 01/27/2022 19:49:35 - INFO - codeparrot_training - Step 824: {'lr': 0.000206, 'samples': 158400, 'steps': 824, 'loss/train': 7.665543794631958} 01/27/2022 19:49:39 - INFO - codeparrot_training - Step 825: {'lr': 0.00020625, 'samples': 158592, 'steps': 825, 'loss/train': 6.250310897827148} 01/27/2022 19:49:43 - INFO - codeparrot_training - Step 826: {'lr': 0.0002065, 'samples': 158784, 'steps': 826, 'loss/train': 6.196201801300049} 01/27/2022 19:49:49 - INFO - codeparrot_training - Step 827: {'lr': 0.00020674999999999998, 'samples': 158976, 'steps': 827, 'loss/train': 6.98847770690918} 01/27/2022 19:49:53 - INFO - codeparrot_training - Step 828: {'lr': 0.000207, 'samples': 159168, 'steps': 828, 'loss/train': 7.437055349349976} 01/27/2022 19:49:57 - INFO - codeparrot_training - Step 829: {'lr': 0.00020725, 'samples': 159360, 'steps': 829, 'loss/train': 7.113067388534546} 01/27/2022 19:50:02 - INFO - codeparrot_training - Step 830: {'lr': 0.0002075, 'samples': 159552, 'steps': 830, 'loss/train': 7.686000108718872} 01/27/2022 19:50:06 - INFO - codeparrot_training - Step 831: {'lr': 0.00020774999999999998, 'samples': 159744, 'steps': 831, 'loss/train': 6.280385255813599} 01/27/2022 19:50:11 - INFO - codeparrot_training - Step 832: {'lr': 0.000208, 'samples': 159936, 'steps': 832, 'loss/train': 7.563260793685913} 01/27/2022 19:50:15 - INFO - codeparrot_training - Step 833: {'lr': 0.00020825, 'samples': 160128, 'steps': 833, 'loss/train': 7.371585130691528} 01/27/2022 19:50:19 - INFO - codeparrot_training - Step 834: {'lr': 0.0002085, 'samples': 160320, 'steps': 834, 'loss/train': 7.753119707107544} 01/27/2022 19:50:23 - INFO - codeparrot_training - Step 835: {'lr': 0.00020875, 'samples': 160512, 'steps': 835, 'loss/train': 7.26523232460022} 01/27/2022 19:50:28 - INFO - codeparrot_training - Step 836: {'lr': 0.00020899999999999998, 'samples': 160704, 'steps': 836, 'loss/train': 7.165149450302124} 01/27/2022 19:50:33 - INFO - codeparrot_training - Step 837: {'lr': 0.00020925, 'samples': 160896, 'steps': 837, 'loss/train': 5.2571539878845215} 01/27/2022 19:50:37 - INFO - codeparrot_training - Step 838: {'lr': 0.0002095, 'samples': 161088, 'steps': 838, 'loss/train': 7.704563140869141} 01/27/2022 19:50:41 - INFO - codeparrot_training - Step 839: {'lr': 0.00020975, 'samples': 161280, 'steps': 839, 'loss/train': 6.747865676879883} 01/27/2022 19:50:45 - INFO - codeparrot_training - Step 840: {'lr': 0.00021, 'samples': 161472, 'steps': 840, 'loss/train': 6.728604555130005} 01/27/2022 19:50:49 - INFO - codeparrot_training - Step 841: {'lr': 0.00021025, 'samples': 161664, 'steps': 841, 'loss/train': 6.855344295501709} 01/27/2022 19:50:56 - INFO - codeparrot_training - Step 842: {'lr': 0.0002105, 'samples': 161856, 'steps': 842, 'loss/train': 7.13779878616333} 01/27/2022 19:51:00 - INFO - codeparrot_training - Step 843: {'lr': 0.00021075, 'samples': 162048, 'steps': 843, 'loss/train': 6.714107036590576} 01/27/2022 19:51:04 - INFO - codeparrot_training - Step 844: {'lr': 0.000211, 'samples': 162240, 'steps': 844, 'loss/train': 7.078253746032715} 01/27/2022 19:51:08 - INFO - codeparrot_training - Step 845: {'lr': 0.00021124999999999998, 'samples': 162432, 'steps': 845, 'loss/train': 4.984176993370056} 01/27/2022 19:51:12 - INFO - codeparrot_training - Step 846: {'lr': 0.0002115, 'samples': 162624, 'steps': 846, 'loss/train': 4.37643027305603} 01/27/2022 19:51:17 - INFO - codeparrot_training - Step 847: {'lr': 0.00021175, 'samples': 162816, 'steps': 847, 'loss/train': 7.023126840591431} 01/27/2022 19:51:22 - INFO - codeparrot_training - Step 848: {'lr': 0.000212, 'samples': 163008, 'steps': 848, 'loss/train': 6.818269729614258} 01/27/2022 19:51:26 - INFO - codeparrot_training - Step 849: {'lr': 0.00021225, 'samples': 163200, 'steps': 849, 'loss/train': 7.410804033279419} 01/27/2022 19:51:30 - INFO - codeparrot_training - Step 850: {'lr': 0.0002125, 'samples': 163392, 'steps': 850, 'loss/train': 7.784402847290039} 01/27/2022 19:51:34 - INFO - codeparrot_training - Step 851: {'lr': 0.00021275, 'samples': 163584, 'steps': 851, 'loss/train': 7.730224370956421} 01/27/2022 19:51:40 - INFO - codeparrot_training - Step 852: {'lr': 0.000213, 'samples': 163776, 'steps': 852, 'loss/train': 6.38837742805481} 01/27/2022 19:51:44 - INFO - codeparrot_training - Step 853: {'lr': 0.00021325, 'samples': 163968, 'steps': 853, 'loss/train': 6.656601190567017} 01/27/2022 19:51:48 - INFO - codeparrot_training - Step 854: {'lr': 0.0002135, 'samples': 164160, 'steps': 854, 'loss/train': 8.349605083465576} 01/27/2022 19:51:52 - INFO - codeparrot_training - Step 855: {'lr': 0.00021375, 'samples': 164352, 'steps': 855, 'loss/train': 6.615616321563721} 01/27/2022 19:51:56 - INFO - codeparrot_training - Step 856: {'lr': 0.000214, 'samples': 164544, 'steps': 856, 'loss/train': 7.127137899398804} 01/27/2022 19:52:02 - INFO - codeparrot_training - Step 857: {'lr': 0.00021425, 'samples': 164736, 'steps': 857, 'loss/train': 7.894361972808838} 01/27/2022 19:52:06 - INFO - codeparrot_training - Step 858: {'lr': 0.0002145, 'samples': 164928, 'steps': 858, 'loss/train': 7.885131597518921} 01/27/2022 19:52:10 - INFO - codeparrot_training - Step 859: {'lr': 0.00021475, 'samples': 165120, 'steps': 859, 'loss/train': 6.882686376571655} 01/27/2022 19:52:15 - INFO - codeparrot_training - Step 860: {'lr': 0.000215, 'samples': 165312, 'steps': 860, 'loss/train': 7.567090272903442} 01/27/2022 19:52:19 - INFO - codeparrot_training - Step 861: {'lr': 0.00021525, 'samples': 165504, 'steps': 861, 'loss/train': 7.0731682777404785} 01/27/2022 19:52:24 - INFO - codeparrot_training - Step 862: {'lr': 0.0002155, 'samples': 165696, 'steps': 862, 'loss/train': 9.323542356491089} 01/27/2022 19:52:28 - INFO - codeparrot_training - Step 863: {'lr': 0.00021575, 'samples': 165888, 'steps': 863, 'loss/train': 6.99372124671936} 01/27/2022 19:52:32 - INFO - codeparrot_training - Step 864: {'lr': 0.000216, 'samples': 166080, 'steps': 864, 'loss/train': 7.365343809127808} 01/27/2022 19:52:36 - INFO - codeparrot_training - Step 865: {'lr': 0.00021625, 'samples': 166272, 'steps': 865, 'loss/train': 7.873505115509033} 01/27/2022 19:52:40 - INFO - codeparrot_training - Step 866: {'lr': 0.0002165, 'samples': 166464, 'steps': 866, 'loss/train': 7.0479912757873535} 01/27/2022 19:52:46 - INFO - codeparrot_training - Step 867: {'lr': 0.00021675, 'samples': 166656, 'steps': 867, 'loss/train': 7.002092599868774} 01/27/2022 19:52:50 - INFO - codeparrot_training - Step 868: {'lr': 0.00021700000000000002, 'samples': 166848, 'steps': 868, 'loss/train': 8.695030689239502} 01/27/2022 19:52:54 - INFO - codeparrot_training - Step 869: {'lr': 0.00021725, 'samples': 167040, 'steps': 869, 'loss/train': 6.47701621055603} 01/27/2022 19:52:58 - INFO - codeparrot_training - Step 870: {'lr': 0.0002175, 'samples': 167232, 'steps': 870, 'loss/train': 7.1184282302856445} 01/27/2022 19:53:02 - INFO - codeparrot_training - Step 871: {'lr': 0.00021775, 'samples': 167424, 'steps': 871, 'loss/train': 7.635532379150391} 01/27/2022 19:53:07 - INFO - codeparrot_training - Step 872: {'lr': 0.000218, 'samples': 167616, 'steps': 872, 'loss/train': 7.512827396392822} 01/27/2022 19:53:12 - INFO - codeparrot_training - Step 873: {'lr': 0.00021825, 'samples': 167808, 'steps': 873, 'loss/train': 7.669466257095337} 01/27/2022 19:53:16 - INFO - codeparrot_training - Step 874: {'lr': 0.0002185, 'samples': 168000, 'steps': 874, 'loss/train': 7.071858644485474} 01/27/2022 19:53:20 - INFO - codeparrot_training - Step 875: {'lr': 0.00021875, 'samples': 168192, 'steps': 875, 'loss/train': 7.268239259719849} 01/27/2022 19:53:24 - INFO - codeparrot_training - Step 876: {'lr': 0.000219, 'samples': 168384, 'steps': 876, 'loss/train': 8.0877685546875} 01/27/2022 19:53:30 - INFO - codeparrot_training - Step 877: {'lr': 0.00021925000000000002, 'samples': 168576, 'steps': 877, 'loss/train': 7.6133058071136475} 01/27/2022 19:53:34 - INFO - codeparrot_training - Step 878: {'lr': 0.0002195, 'samples': 168768, 'steps': 878, 'loss/train': 7.026730298995972} 01/27/2022 19:53:39 - INFO - codeparrot_training - Step 879: {'lr': 0.00021975, 'samples': 168960, 'steps': 879, 'loss/train': 7.413425445556641} 01/27/2022 19:53:43 - INFO - codeparrot_training - Step 880: {'lr': 0.00022, 'samples': 169152, 'steps': 880, 'loss/train': 8.85764193534851} 01/27/2022 19:53:47 - INFO - codeparrot_training - Step 881: {'lr': 0.00022025000000000001, 'samples': 169344, 'steps': 881, 'loss/train': 7.755910634994507} 01/27/2022 19:53:51 - INFO - codeparrot_training - Step 882: {'lr': 0.0002205, 'samples': 169536, 'steps': 882, 'loss/train': 7.811797857284546} 01/27/2022 19:53:56 - INFO - codeparrot_training - Step 883: {'lr': 0.00022075, 'samples': 169728, 'steps': 883, 'loss/train': 7.600443363189697} 01/27/2022 19:54:01 - INFO - codeparrot_training - Step 884: {'lr': 0.000221, 'samples': 169920, 'steps': 884, 'loss/train': 6.118984937667847} 01/27/2022 19:54:05 - INFO - codeparrot_training - Step 885: {'lr': 0.00022125, 'samples': 170112, 'steps': 885, 'loss/train': 7.171224117279053} 01/27/2022 19:54:09 - INFO - codeparrot_training - Step 886: {'lr': 0.00022150000000000002, 'samples': 170304, 'steps': 886, 'loss/train': 6.639321327209473} 01/27/2022 19:54:15 - INFO - codeparrot_training - Step 887: {'lr': 0.00022175, 'samples': 170496, 'steps': 887, 'loss/train': 7.177823066711426} 01/27/2022 19:54:19 - INFO - codeparrot_training - Step 888: {'lr': 0.000222, 'samples': 170688, 'steps': 888, 'loss/train': 7.860971689224243} 01/27/2022 19:54:23 - INFO - codeparrot_training - Step 889: {'lr': 0.00022225, 'samples': 170880, 'steps': 889, 'loss/train': 7.319984436035156} 01/27/2022 19:54:27 - INFO - codeparrot_training - Step 890: {'lr': 0.00022250000000000001, 'samples': 171072, 'steps': 890, 'loss/train': 6.704685688018799} 01/27/2022 19:54:31 - INFO - codeparrot_training - Step 891: {'lr': 0.00022275000000000002, 'samples': 171264, 'steps': 891, 'loss/train': 6.902200698852539} 01/27/2022 19:54:36 - INFO - codeparrot_training - Step 892: {'lr': 0.000223, 'samples': 171456, 'steps': 892, 'loss/train': 7.375280141830444} 01/27/2022 19:54:41 - INFO - codeparrot_training - Step 893: {'lr': 0.00022325, 'samples': 171648, 'steps': 893, 'loss/train': 5.083872556686401} 01/27/2022 19:54:45 - INFO - codeparrot_training - Step 894: {'lr': 0.0002235, 'samples': 171840, 'steps': 894, 'loss/train': 6.673138618469238} 01/27/2022 19:54:49 - INFO - codeparrot_training - Step 895: {'lr': 0.00022375000000000002, 'samples': 172032, 'steps': 895, 'loss/train': 7.516380786895752} 01/27/2022 19:54:53 - INFO - codeparrot_training - Step 896: {'lr': 0.000224, 'samples': 172224, 'steps': 896, 'loss/train': 6.994914293289185} 01/27/2022 19:54:58 - INFO - codeparrot_training - Step 897: {'lr': 0.00022425, 'samples': 172416, 'steps': 897, 'loss/train': 7.059888124465942} 01/27/2022 19:55:02 - INFO - codeparrot_training - Step 898: {'lr': 0.0002245, 'samples': 172608, 'steps': 898, 'loss/train': 3.5606542825698853} 01/27/2022 19:55:06 - INFO - codeparrot_training - Step 899: {'lr': 0.00022475000000000001, 'samples': 172800, 'steps': 899, 'loss/train': 8.262231588363647} 01/27/2022 19:55:11 - INFO - codeparrot_training - Step 900: {'lr': 0.00022500000000000002, 'samples': 172992, 'steps': 900, 'loss/train': 7.715365648269653} 01/27/2022 19:55:15 - INFO - codeparrot_training - Step 901: {'lr': 0.00022525, 'samples': 173184, 'steps': 901, 'loss/train': 6.664234399795532} 01/27/2022 19:55:21 - INFO - codeparrot_training - Step 902: {'lr': 0.0002255, 'samples': 173376, 'steps': 902, 'loss/train': 7.200341463088989} 01/27/2022 19:55:25 - INFO - codeparrot_training - Step 903: {'lr': 0.00022575, 'samples': 173568, 'steps': 903, 'loss/train': 7.4642908573150635} 01/27/2022 19:55:29 - INFO - codeparrot_training - Step 904: {'lr': 0.00022600000000000002, 'samples': 173760, 'steps': 904, 'loss/train': 6.824518203735352} 01/27/2022 19:55:33 - INFO - codeparrot_training - Step 905: {'lr': 0.00022625000000000002, 'samples': 173952, 'steps': 905, 'loss/train': 6.4327311515808105} 01/27/2022 19:55:37 - INFO - codeparrot_training - Step 906: {'lr': 0.0002265, 'samples': 174144, 'steps': 906, 'loss/train': 6.040751695632935} 01/27/2022 19:55:42 - INFO - codeparrot_training - Step 907: {'lr': 0.00022675, 'samples': 174336, 'steps': 907, 'loss/train': 4.96863055229187} 01/27/2022 19:55:47 - INFO - codeparrot_training - Step 908: {'lr': 0.00022700000000000002, 'samples': 174528, 'steps': 908, 'loss/train': 6.524607181549072} 01/27/2022 19:55:51 - INFO - codeparrot_training - Step 909: {'lr': 0.00022725000000000002, 'samples': 174720, 'steps': 909, 'loss/train': 6.859138011932373} 01/27/2022 19:55:55 - INFO - codeparrot_training - Step 910: {'lr': 0.0002275, 'samples': 174912, 'steps': 910, 'loss/train': 7.168780088424683} 01/27/2022 19:55:59 - INFO - codeparrot_training - Step 911: {'lr': 0.00022775, 'samples': 175104, 'steps': 911, 'loss/train': 6.728759050369263} 01/27/2022 19:56:05 - INFO - codeparrot_training - Step 912: {'lr': 0.000228, 'samples': 175296, 'steps': 912, 'loss/train': 7.605461597442627} 01/27/2022 19:56:09 - INFO - codeparrot_training - Step 913: {'lr': 0.00022825000000000002, 'samples': 175488, 'steps': 913, 'loss/train': 6.410759210586548} 01/27/2022 19:56:13 - INFO - codeparrot_training - Step 914: {'lr': 0.00022850000000000002, 'samples': 175680, 'steps': 914, 'loss/train': 7.014211893081665} 01/27/2022 19:56:17 - INFO - codeparrot_training - Step 915: {'lr': 0.00022875, 'samples': 175872, 'steps': 915, 'loss/train': 7.0388023853302} 01/27/2022 19:56:22 - INFO - codeparrot_training - Step 916: {'lr': 0.000229, 'samples': 176064, 'steps': 916, 'loss/train': 4.739097833633423} 01/27/2022 19:56:27 - INFO - codeparrot_training - Step 917: {'lr': 0.00022925000000000002, 'samples': 176256, 'steps': 917, 'loss/train': 7.574775695800781} 01/27/2022 19:56:31 - INFO - codeparrot_training - Step 918: {'lr': 0.00022950000000000002, 'samples': 176448, 'steps': 918, 'loss/train': 7.156859636306763} 01/27/2022 19:56:35 - INFO - codeparrot_training - Step 919: {'lr': 0.00022975000000000003, 'samples': 176640, 'steps': 919, 'loss/train': 7.780276536941528} 01/27/2022 19:56:39 - INFO - codeparrot_training - Step 920: {'lr': 0.00023, 'samples': 176832, 'steps': 920, 'loss/train': 6.601100921630859} 01/27/2022 19:56:43 - INFO - codeparrot_training - Step 921: {'lr': 0.00023025, 'samples': 177024, 'steps': 921, 'loss/train': 6.518356561660767} 01/27/2022 19:56:49 - INFO - codeparrot_training - Step 922: {'lr': 0.00023050000000000002, 'samples': 177216, 'steps': 922, 'loss/train': 7.5875279903411865} 01/27/2022 19:56:53 - INFO - codeparrot_training - Step 923: {'lr': 0.00023075000000000003, 'samples': 177408, 'steps': 923, 'loss/train': 6.655817270278931} 01/27/2022 19:56:57 - INFO - codeparrot_training - Step 924: {'lr': 0.000231, 'samples': 177600, 'steps': 924, 'loss/train': 6.846208333969116} 01/27/2022 19:57:01 - INFO - codeparrot_training - Step 925: {'lr': 0.00023125, 'samples': 177792, 'steps': 925, 'loss/train': 7.254077196121216} 01/27/2022 19:57:05 - INFO - codeparrot_training - Step 926: {'lr': 0.00023150000000000002, 'samples': 177984, 'steps': 926, 'loss/train': 7.18137788772583} 01/27/2022 19:57:10 - INFO - codeparrot_training - Step 927: {'lr': 0.00023175000000000002, 'samples': 178176, 'steps': 927, 'loss/train': 9.288630723953247} 01/27/2022 19:57:15 - INFO - codeparrot_training - Step 928: {'lr': 0.00023200000000000003, 'samples': 178368, 'steps': 928, 'loss/train': 7.154744625091553} 01/27/2022 19:57:19 - INFO - codeparrot_training - Step 929: {'lr': 0.00023225, 'samples': 178560, 'steps': 929, 'loss/train': 6.805961608886719} 01/27/2022 19:57:23 - INFO - codeparrot_training - Step 930: {'lr': 0.0002325, 'samples': 178752, 'steps': 930, 'loss/train': 7.899904489517212} 01/27/2022 19:57:27 - INFO - codeparrot_training - Step 931: {'lr': 0.00023275000000000002, 'samples': 178944, 'steps': 931, 'loss/train': 7.287409543991089} 01/27/2022 19:57:33 - INFO - codeparrot_training - Step 932: {'lr': 0.00023300000000000003, 'samples': 179136, 'steps': 932, 'loss/train': 6.028538703918457} 01/27/2022 19:57:37 - INFO - codeparrot_training - Step 933: {'lr': 0.00023325, 'samples': 179328, 'steps': 933, 'loss/train': 6.456415414810181} 01/27/2022 19:57:41 - INFO - codeparrot_training - Step 934: {'lr': 0.0002335, 'samples': 179520, 'steps': 934, 'loss/train': 8.036659955978394} 01/27/2022 19:57:45 - INFO - codeparrot_training - Step 935: {'lr': 0.00023375000000000002, 'samples': 179712, 'steps': 935, 'loss/train': 6.991081953048706} 01/27/2022 19:57:50 - INFO - codeparrot_training - Step 936: {'lr': 0.00023400000000000002, 'samples': 179904, 'steps': 936, 'loss/train': 6.907919883728027} 01/27/2022 19:57:55 - INFO - codeparrot_training - Step 937: {'lr': 0.00023425000000000003, 'samples': 180096, 'steps': 937, 'loss/train': 7.334412574768066} 01/27/2022 19:57:59 - INFO - codeparrot_training - Step 938: {'lr': 0.00023449999999999998, 'samples': 180288, 'steps': 938, 'loss/train': 6.904307842254639} 01/27/2022 19:58:03 - INFO - codeparrot_training - Step 939: {'lr': 0.00023475, 'samples': 180480, 'steps': 939, 'loss/train': 9.220777988433838} 01/27/2022 19:58:07 - INFO - codeparrot_training - Step 940: {'lr': 0.000235, 'samples': 180672, 'steps': 940, 'loss/train': 6.523514270782471} 01/27/2022 19:58:11 - INFO - codeparrot_training - Step 941: {'lr': 0.00023525, 'samples': 180864, 'steps': 941, 'loss/train': 7.42478084564209} 01/27/2022 19:58:16 - INFO - codeparrot_training - Step 942: {'lr': 0.0002355, 'samples': 181056, 'steps': 942, 'loss/train': 7.942927837371826} 01/27/2022 19:58:21 - INFO - codeparrot_training - Step 943: {'lr': 0.00023574999999999998, 'samples': 181248, 'steps': 943, 'loss/train': 6.2294347286224365} 01/27/2022 19:58:25 - INFO - codeparrot_training - Step 944: {'lr': 0.000236, 'samples': 181440, 'steps': 944, 'loss/train': 8.150972843170166} 01/27/2022 19:58:29 - INFO - codeparrot_training - Step 945: {'lr': 0.00023625, 'samples': 181632, 'steps': 945, 'loss/train': 7.299391508102417} 01/27/2022 19:58:33 - INFO - codeparrot_training - Step 946: {'lr': 0.0002365, 'samples': 181824, 'steps': 946, 'loss/train': 7.574323654174805} 01/27/2022 19:58:39 - INFO - codeparrot_training - Step 947: {'lr': 0.00023674999999999998, 'samples': 182016, 'steps': 947, 'loss/train': 6.772777318954468} 01/27/2022 19:58:43 - INFO - codeparrot_training - Step 948: {'lr': 0.000237, 'samples': 182208, 'steps': 948, 'loss/train': 6.088188171386719} 01/27/2022 19:58:47 - INFO - codeparrot_training - Step 949: {'lr': 0.00023725, 'samples': 182400, 'steps': 949, 'loss/train': 7.268367290496826} 01/27/2022 19:58:51 - INFO - codeparrot_training - Step 950: {'lr': 0.0002375, 'samples': 182592, 'steps': 950, 'loss/train': 7.064356327056885} 01/27/2022 19:58:56 - INFO - codeparrot_training - Step 951: {'lr': 0.00023775, 'samples': 182784, 'steps': 951, 'loss/train': 7.1734278202056885} 01/27/2022 19:59:01 - INFO - codeparrot_training - Step 952: {'lr': 0.00023799999999999998, 'samples': 182976, 'steps': 952, 'loss/train': 6.710713148117065} 01/27/2022 19:59:05 - INFO - codeparrot_training - Step 953: {'lr': 0.00023825, 'samples': 183168, 'steps': 953, 'loss/train': 8.174689292907715} 01/27/2022 19:59:09 - INFO - codeparrot_training - Step 954: {'lr': 0.0002385, 'samples': 183360, 'steps': 954, 'loss/train': 7.612285137176514} 01/27/2022 19:59:14 - INFO - codeparrot_training - Step 955: {'lr': 0.00023875, 'samples': 183552, 'steps': 955, 'loss/train': 7.401938438415527} 01/27/2022 19:59:18 - INFO - codeparrot_training - Step 956: {'lr': 0.00023899999999999998, 'samples': 183744, 'steps': 956, 'loss/train': 7.803273439407349} 01/27/2022 19:59:24 - INFO - codeparrot_training - Step 957: {'lr': 0.00023925, 'samples': 183936, 'steps': 957, 'loss/train': 6.716939449310303} 01/27/2022 19:59:28 - INFO - codeparrot_training - Step 958: {'lr': 0.0002395, 'samples': 184128, 'steps': 958, 'loss/train': 6.849352598190308} 01/27/2022 19:59:32 - INFO - codeparrot_training - Step 959: {'lr': 0.00023975, 'samples': 184320, 'steps': 959, 'loss/train': 7.120357275009155} 01/27/2022 19:59:36 - INFO - codeparrot_training - Step 960: {'lr': 0.00024, 'samples': 184512, 'steps': 960, 'loss/train': 9.059615850448608} 01/27/2022 19:59:40 - INFO - codeparrot_training - Step 961: {'lr': 0.00024024999999999999, 'samples': 184704, 'steps': 961, 'loss/train': 6.081015586853027} 01/27/2022 19:59:45 - INFO - codeparrot_training - Step 962: {'lr': 0.0002405, 'samples': 184896, 'steps': 962, 'loss/train': 7.1585633754730225} 01/27/2022 19:59:49 - INFO - codeparrot_training - Step 963: {'lr': 0.00024075, 'samples': 185088, 'steps': 963, 'loss/train': 6.374916315078735} 01/27/2022 19:59:54 - INFO - codeparrot_training - Step 964: {'lr': 0.000241, 'samples': 185280, 'steps': 964, 'loss/train': 7.46045994758606} 01/27/2022 19:59:58 - INFO - codeparrot_training - Step 965: {'lr': 0.00024125, 'samples': 185472, 'steps': 965, 'loss/train': 7.519346237182617} 01/27/2022 20:00:02 - INFO - codeparrot_training - Step 966: {'lr': 0.0002415, 'samples': 185664, 'steps': 966, 'loss/train': 6.693201541900635} 01/27/2022 20:00:07 - INFO - codeparrot_training - Step 967: {'lr': 0.00024175, 'samples': 185856, 'steps': 967, 'loss/train': 5.646547436714172} 01/27/2022 20:00:11 - INFO - codeparrot_training - Step 968: {'lr': 0.000242, 'samples': 186048, 'steps': 968, 'loss/train': 7.297076225280762} 01/27/2022 20:00:16 - INFO - codeparrot_training - Step 969: {'lr': 0.00024225, 'samples': 186240, 'steps': 969, 'loss/train': 6.749300479888916} 01/27/2022 20:00:20 - INFO - codeparrot_training - Step 970: {'lr': 0.00024249999999999999, 'samples': 186432, 'steps': 970, 'loss/train': 8.186639785766602} 01/27/2022 20:00:24 - INFO - codeparrot_training - Step 971: {'lr': 0.00024275, 'samples': 186624, 'steps': 971, 'loss/train': 3.0953779220581055} 01/27/2022 20:00:29 - INFO - codeparrot_training - Step 972: {'lr': 0.000243, 'samples': 186816, 'steps': 972, 'loss/train': 3.534411907196045} 01/27/2022 20:00:33 - INFO - codeparrot_training - Step 973: {'lr': 0.00024325, 'samples': 187008, 'steps': 973, 'loss/train': 7.451557874679565} 01/27/2022 20:00:38 - INFO - codeparrot_training - Step 974: {'lr': 0.0002435, 'samples': 187200, 'steps': 974, 'loss/train': 7.9868645668029785} 01/27/2022 20:00:42 - INFO - codeparrot_training - Step 975: {'lr': 0.00024375, 'samples': 187392, 'steps': 975, 'loss/train': 6.10376501083374} 01/27/2022 20:00:46 - INFO - codeparrot_training - Step 976: {'lr': 0.000244, 'samples': 187584, 'steps': 976, 'loss/train': 7.4672276973724365} 01/27/2022 20:00:52 - INFO - codeparrot_training - Step 977: {'lr': 0.00024425, 'samples': 187776, 'steps': 977, 'loss/train': 7.113274097442627} 01/27/2022 20:00:56 - INFO - codeparrot_training - Step 978: {'lr': 0.0002445, 'samples': 187968, 'steps': 978, 'loss/train': 7.51456618309021} 01/27/2022 20:01:00 - INFO - codeparrot_training - Step 979: {'lr': 0.00024475, 'samples': 188160, 'steps': 979, 'loss/train': 7.385890960693359} 01/27/2022 20:01:04 - INFO - codeparrot_training - Step 980: {'lr': 0.000245, 'samples': 188352, 'steps': 980, 'loss/train': 8.501586198806763} 01/27/2022 20:01:09 - INFO - codeparrot_training - Step 981: {'lr': 0.00024525, 'samples': 188544, 'steps': 981, 'loss/train': 7.369788408279419} 01/27/2022 20:01:14 - INFO - codeparrot_training - Step 982: {'lr': 0.0002455, 'samples': 188736, 'steps': 982, 'loss/train': 6.954960823059082} 01/27/2022 20:01:18 - INFO - codeparrot_training - Step 983: {'lr': 0.00024575, 'samples': 188928, 'steps': 983, 'loss/train': 6.792914628982544} 01/27/2022 20:01:22 - INFO - codeparrot_training - Step 984: {'lr': 0.000246, 'samples': 189120, 'steps': 984, 'loss/train': 3.656927704811096} 01/27/2022 20:01:27 - INFO - codeparrot_training - Step 985: {'lr': 0.00024625, 'samples': 189312, 'steps': 985, 'loss/train': 6.928876161575317} 01/27/2022 20:01:32 - INFO - codeparrot_training - Step 986: {'lr': 0.00024650000000000003, 'samples': 189504, 'steps': 986, 'loss/train': 7.014236211776733} 01/27/2022 20:01:36 - INFO - codeparrot_training - Step 987: {'lr': 0.00024675, 'samples': 189696, 'steps': 987, 'loss/train': 6.70487380027771} 01/27/2022 20:01:40 - INFO - codeparrot_training - Step 988: {'lr': 0.000247, 'samples': 189888, 'steps': 988, 'loss/train': 7.39477801322937} 01/27/2022 20:01:44 - INFO - codeparrot_training - Step 989: {'lr': 0.00024725, 'samples': 190080, 'steps': 989, 'loss/train': 7.157878875732422} 01/27/2022 20:01:48 - INFO - codeparrot_training - Step 990: {'lr': 0.0002475, 'samples': 190272, 'steps': 990, 'loss/train': 7.20275616645813} 01/27/2022 20:01:54 - INFO - codeparrot_training - Step 991: {'lr': 0.00024775, 'samples': 190464, 'steps': 991, 'loss/train': 6.944641828536987} 01/27/2022 20:01:58 - INFO - codeparrot_training - Step 992: {'lr': 0.000248, 'samples': 190656, 'steps': 992, 'loss/train': 6.4415974617004395} 01/27/2022 20:02:03 - INFO - codeparrot_training - Step 993: {'lr': 0.00024825, 'samples': 190848, 'steps': 993, 'loss/train': 6.851691484451294} 01/27/2022 20:02:07 - INFO - codeparrot_training - Step 994: {'lr': 0.0002485, 'samples': 191040, 'steps': 994, 'loss/train': 6.905003786087036} 01/27/2022 20:02:11 - INFO - codeparrot_training - Step 995: {'lr': 0.00024875, 'samples': 191232, 'steps': 995, 'loss/train': 8.069269180297852} 01/27/2022 20:02:16 - INFO - codeparrot_training - Step 996: {'lr': 0.000249, 'samples': 191424, 'steps': 996, 'loss/train': 6.3540308475494385} 01/27/2022 20:02:20 - INFO - codeparrot_training - Step 997: {'lr': 0.00024925, 'samples': 191616, 'steps': 997, 'loss/train': 6.985671043395996} 01/27/2022 20:02:24 - INFO - codeparrot_training - Step 998: {'lr': 0.0002495, 'samples': 191808, 'steps': 998, 'loss/train': 5.8150938749313354} 01/27/2022 20:02:29 - INFO - codeparrot_training - Step 999: {'lr': 0.00024975, 'samples': 192000, 'steps': 999, 'loss/train': 6.347761631011963} 01/27/2022 20:02:33 - INFO - codeparrot_training - Step 1000: {'lr': 0.00025, 'samples': 192192, 'steps': 1000, 'loss/train': 2.727059841156006} 01/27/2022 20:02:39 - INFO - codeparrot_training - Step 1001: {'lr': 0.00025025, 'samples': 192384, 'steps': 1001, 'loss/train': 8.42210340499878} 01/27/2022 20:02:43 - INFO - codeparrot_training - Step 1002: {'lr': 0.0002505, 'samples': 192576, 'steps': 1002, 'loss/train': 7.332793951034546} 01/27/2022 20:02:47 - INFO - codeparrot_training - Step 1003: {'lr': 0.00025075, 'samples': 192768, 'steps': 1003, 'loss/train': 7.334043502807617} 01/27/2022 20:02:51 - INFO - codeparrot_training - Step 1004: {'lr': 0.00025100000000000003, 'samples': 192960, 'steps': 1004, 'loss/train': 6.399794340133667} 01/27/2022 20:02:55 - INFO - codeparrot_training - Step 1005: {'lr': 0.00025124999999999995, 'samples': 193152, 'steps': 1005, 'loss/train': 6.915691137313843} 01/27/2022 20:03:00 - INFO - codeparrot_training - Step 1006: {'lr': 0.0002515, 'samples': 193344, 'steps': 1006, 'loss/train': 7.140349388122559} 01/27/2022 20:03:04 - INFO - codeparrot_training - Step 1007: {'lr': 0.00025174999999999997, 'samples': 193536, 'steps': 1007, 'loss/train': 7.182172536849976} 01/27/2022 20:03:09 - INFO - codeparrot_training - Step 1008: {'lr': 0.000252, 'samples': 193728, 'steps': 1008, 'loss/train': 5.296411871910095} 01/27/2022 20:03:13 - INFO - codeparrot_training - Step 1009: {'lr': 0.00025225, 'samples': 193920, 'steps': 1009, 'loss/train': 7.232731819152832} 01/27/2022 20:03:17 - INFO - codeparrot_training - Step 1010: {'lr': 0.0002525, 'samples': 194112, 'steps': 1010, 'loss/train': 6.959974765777588} 01/27/2022 20:03:22 - INFO - codeparrot_training - Step 1011: {'lr': 0.00025275, 'samples': 194304, 'steps': 1011, 'loss/train': 6.996084451675415} 01/27/2022 20:03:26 - INFO - codeparrot_training - Step 1012: {'lr': 0.000253, 'samples': 194496, 'steps': 1012, 'loss/train': 6.97163200378418} 01/27/2022 20:03:30 - INFO - codeparrot_training - Step 1013: {'lr': 0.00025325, 'samples': 194688, 'steps': 1013, 'loss/train': 6.342827081680298} 01/27/2022 20:03:35 - INFO - codeparrot_training - Step 1014: {'lr': 0.0002535, 'samples': 194880, 'steps': 1014, 'loss/train': 7.124094486236572} 01/27/2022 20:03:39 - INFO - codeparrot_training - Step 1015: {'lr': 0.00025374999999999996, 'samples': 195072, 'steps': 1015, 'loss/train': 8.34977102279663} 01/27/2022 20:03:45 - INFO - codeparrot_training - Step 1016: {'lr': 0.000254, 'samples': 195264, 'steps': 1016, 'loss/train': 7.086735248565674} 01/27/2022 20:03:49 - INFO - codeparrot_training - Step 1017: {'lr': 0.00025425, 'samples': 195456, 'steps': 1017, 'loss/train': 7.534659147262573} 01/27/2022 20:03:53 - INFO - codeparrot_training - Step 1018: {'lr': 0.0002545, 'samples': 195648, 'steps': 1018, 'loss/train': 7.8111891746521} 01/27/2022 20:03:57 - INFO - codeparrot_training - Step 1019: {'lr': 0.00025475, 'samples': 195840, 'steps': 1019, 'loss/train': 6.920222282409668} 01/27/2022 20:04:01 - INFO - codeparrot_training - Step 1020: {'lr': 0.000255, 'samples': 196032, 'steps': 1020, 'loss/train': 7.115208864212036} 01/27/2022 20:04:07 - INFO - codeparrot_training - Step 1021: {'lr': 0.00025525, 'samples': 196224, 'steps': 1021, 'loss/train': 3.0207903385162354} 01/27/2022 20:04:11 - INFO - codeparrot_training - Step 1022: {'lr': 0.00025550000000000003, 'samples': 196416, 'steps': 1022, 'loss/train': 4.931810975074768} 01/27/2022 20:04:15 - INFO - codeparrot_training - Step 1023: {'lr': 0.00025575, 'samples': 196608, 'steps': 1023, 'loss/train': 6.924686908721924} 01/27/2022 20:04:19 - INFO - codeparrot_training - Step 1024: {'lr': 0.000256, 'samples': 196800, 'steps': 1024, 'loss/train': 6.617191314697266} 01/27/2022 20:04:23 - INFO - codeparrot_training - Step 1025: {'lr': 0.00025624999999999997, 'samples': 196992, 'steps': 1025, 'loss/train': 6.591811895370483} 01/27/2022 20:04:28 - INFO - codeparrot_training - Step 1026: {'lr': 0.0002565, 'samples': 197184, 'steps': 1026, 'loss/train': 8.660568952560425} 01/27/2022 20:04:32 - INFO - codeparrot_training - Step 1027: {'lr': 0.00025675, 'samples': 197376, 'steps': 1027, 'loss/train': 6.5282721519470215} 01/27/2022 20:04:37 - INFO - codeparrot_training - Step 1028: {'lr': 0.000257, 'samples': 197568, 'steps': 1028, 'loss/train': 7.155653715133667} 01/27/2022 20:04:41 - INFO - codeparrot_training - Step 1029: {'lr': 0.00025725, 'samples': 197760, 'steps': 1029, 'loss/train': 6.496752977371216} 01/27/2022 20:04:45 - INFO - codeparrot_training - Step 1030: {'lr': 0.0002575, 'samples': 197952, 'steps': 1030, 'loss/train': 6.339601278305054} 01/27/2022 20:04:50 - INFO - codeparrot_training - Step 1031: {'lr': 0.00025775, 'samples': 198144, 'steps': 1031, 'loss/train': 6.180484056472778} 01/27/2022 20:04:54 - INFO - codeparrot_training - Step 1032: {'lr': 0.00025800000000000004, 'samples': 198336, 'steps': 1032, 'loss/train': 6.904130458831787} 01/27/2022 20:04:59 - INFO - codeparrot_training - Step 1033: {'lr': 0.00025824999999999996, 'samples': 198528, 'steps': 1033, 'loss/train': 6.179152965545654} 01/27/2022 20:05:03 - INFO - codeparrot_training - Step 1034: {'lr': 0.0002585, 'samples': 198720, 'steps': 1034, 'loss/train': 6.8922951221466064} 01/27/2022 20:05:09 - INFO - codeparrot_training - Step 1035: {'lr': 0.00025875, 'samples': 198912, 'steps': 1035, 'loss/train': 6.693593502044678} 01/27/2022 20:05:13 - INFO - codeparrot_training - Step 1036: {'lr': 0.000259, 'samples': 199104, 'steps': 1036, 'loss/train': 6.657398700714111} 01/27/2022 20:05:17 - INFO - codeparrot_training - Step 1037: {'lr': 0.00025925, 'samples': 199296, 'steps': 1037, 'loss/train': 5.598530888557434} 01/27/2022 20:05:21 - INFO - codeparrot_training - Step 1038: {'lr': 0.0002595, 'samples': 199488, 'steps': 1038, 'loss/train': 6.935786247253418} 01/27/2022 20:05:25 - INFO - codeparrot_training - Step 1039: {'lr': 0.00025975, 'samples': 199680, 'steps': 1039, 'loss/train': 5.9086010456085205} 01/27/2022 20:05:31 - INFO - codeparrot_training - Step 1040: {'lr': 0.00026000000000000003, 'samples': 199872, 'steps': 1040, 'loss/train': 2.6797045469284058} 01/27/2022 20:05:35 - INFO - codeparrot_training - Step 1041: {'lr': 0.00026025, 'samples': 200064, 'steps': 1041, 'loss/train': 6.955983638763428} 01/27/2022 20:05:39 - INFO - codeparrot_training - Step 1042: {'lr': 0.0002605, 'samples': 200256, 'steps': 1042, 'loss/train': 7.775304079055786} 01/27/2022 20:05:43 - INFO - codeparrot_training - Step 1043: {'lr': 0.00026074999999999997, 'samples': 200448, 'steps': 1043, 'loss/train': 8.124592781066895} 01/27/2022 20:05:47 - INFO - codeparrot_training - Step 1044: {'lr': 0.000261, 'samples': 200640, 'steps': 1044, 'loss/train': 7.032229900360107} 01/27/2022 20:05:52 - INFO - codeparrot_training - Step 1045: {'lr': 0.00026125, 'samples': 200832, 'steps': 1045, 'loss/train': 6.700995683670044} 01/27/2022 20:05:57 - INFO - codeparrot_training - Step 1046: {'lr': 0.0002615, 'samples': 201024, 'steps': 1046, 'loss/train': 5.590494990348816} 01/27/2022 20:06:01 - INFO - codeparrot_training - Step 1047: {'lr': 0.00026175, 'samples': 201216, 'steps': 1047, 'loss/train': 7.30798602104187} 01/27/2022 20:06:05 - INFO - codeparrot_training - Step 1048: {'lr': 0.000262, 'samples': 201408, 'steps': 1048, 'loss/train': 5.346248030662537} 01/27/2022 20:06:09 - INFO - codeparrot_training - Step 1049: {'lr': 0.00026225, 'samples': 201600, 'steps': 1049, 'loss/train': 7.299126148223877} 01/27/2022 20:06:15 - INFO - codeparrot_training - Step 1050: {'lr': 0.00026250000000000004, 'samples': 201792, 'steps': 1050, 'loss/train': 7.436676979064941} 01/27/2022 20:06:19 - INFO - codeparrot_training - Step 1051: {'lr': 0.00026274999999999996, 'samples': 201984, 'steps': 1051, 'loss/train': 6.254862070083618} 01/27/2022 20:06:23 - INFO - codeparrot_training - Step 1052: {'lr': 0.000263, 'samples': 202176, 'steps': 1052, 'loss/train': 6.72723126411438} 01/27/2022 20:06:27 - INFO - codeparrot_training - Step 1053: {'lr': 0.00026325, 'samples': 202368, 'steps': 1053, 'loss/train': 6.373230457305908} 01/27/2022 20:06:32 - INFO - codeparrot_training - Step 1054: {'lr': 0.0002635, 'samples': 202560, 'steps': 1054, 'loss/train': 7.470884799957275} 01/27/2022 20:06:37 - INFO - codeparrot_training - Step 1055: {'lr': 0.00026375, 'samples': 202752, 'steps': 1055, 'loss/train': 6.225339889526367} 01/27/2022 20:06:41 - INFO - codeparrot_training - Step 1056: {'lr': 0.000264, 'samples': 202944, 'steps': 1056, 'loss/train': 6.287844657897949} 01/27/2022 20:06:45 - INFO - codeparrot_training - Step 1057: {'lr': 0.00026425, 'samples': 203136, 'steps': 1057, 'loss/train': 5.9497575759887695} 01/27/2022 20:06:49 - INFO - codeparrot_training - Step 1058: {'lr': 0.00026450000000000003, 'samples': 203328, 'steps': 1058, 'loss/train': 6.595285177230835} 01/27/2022 20:06:53 - INFO - codeparrot_training - Step 1059: {'lr': 0.00026475, 'samples': 203520, 'steps': 1059, 'loss/train': 6.861415386199951} 01/27/2022 20:06:59 - INFO - codeparrot_training - Step 1060: {'lr': 0.00026500000000000004, 'samples': 203712, 'steps': 1060, 'loss/train': 6.762890338897705} 01/27/2022 20:07:03 - INFO - codeparrot_training - Step 1061: {'lr': 0.00026524999999999997, 'samples': 203904, 'steps': 1061, 'loss/train': 6.5421459674835205} 01/27/2022 20:07:08 - INFO - codeparrot_training - Step 1062: {'lr': 0.0002655, 'samples': 204096, 'steps': 1062, 'loss/train': 8.077380895614624} 01/27/2022 20:07:12 - INFO - codeparrot_training - Step 1063: {'lr': 0.00026575, 'samples': 204288, 'steps': 1063, 'loss/train': 6.763631343841553} 01/27/2022 20:07:16 - INFO - codeparrot_training - Step 1064: {'lr': 0.000266, 'samples': 204480, 'steps': 1064, 'loss/train': 7.018068552017212} 01/27/2022 20:07:21 - INFO - codeparrot_training - Step 1065: {'lr': 0.00026625, 'samples': 204672, 'steps': 1065, 'loss/train': 8.275073289871216} 01/27/2022 20:07:25 - INFO - codeparrot_training - Step 1066: {'lr': 0.0002665, 'samples': 204864, 'steps': 1066, 'loss/train': 6.306669473648071} 01/27/2022 20:07:29 - INFO - codeparrot_training - Step 1067: {'lr': 0.00026675, 'samples': 205056, 'steps': 1067, 'loss/train': 2.7302345037460327} 01/27/2022 20:07:33 - INFO - codeparrot_training - Step 1068: {'lr': 0.00026700000000000004, 'samples': 205248, 'steps': 1068, 'loss/train': 7.303512096405029} 01/27/2022 20:07:37 - INFO - codeparrot_training - Step 1069: {'lr': 0.00026725, 'samples': 205440, 'steps': 1069, 'loss/train': 6.571877717971802} 01/27/2022 20:07:43 - INFO - codeparrot_training - Step 1070: {'lr': 0.0002675, 'samples': 205632, 'steps': 1070, 'loss/train': 6.219856023788452} 01/27/2022 20:07:47 - INFO - codeparrot_training - Step 1071: {'lr': 0.00026775, 'samples': 205824, 'steps': 1071, 'loss/train': 6.711027145385742} 01/27/2022 20:07:51 - INFO - codeparrot_training - Step 1072: {'lr': 0.000268, 'samples': 206016, 'steps': 1072, 'loss/train': 6.044724225997925} 01/27/2022 20:07:55 - INFO - codeparrot_training - Step 1073: {'lr': 0.00026825, 'samples': 206208, 'steps': 1073, 'loss/train': 2.7096844911575317} 01/27/2022 20:07:59 - INFO - codeparrot_training - Step 1074: {'lr': 0.0002685, 'samples': 206400, 'steps': 1074, 'loss/train': 8.10952377319336} 01/27/2022 20:08:05 - INFO - codeparrot_training - Step 1075: {'lr': 0.00026875, 'samples': 206592, 'steps': 1075, 'loss/train': 7.001063346862793} 01/27/2022 20:08:09 - INFO - codeparrot_training - Step 1076: {'lr': 0.00026900000000000003, 'samples': 206784, 'steps': 1076, 'loss/train': 6.5374345779418945} 01/27/2022 20:08:13 - INFO - codeparrot_training - Step 1077: {'lr': 0.00026925, 'samples': 206976, 'steps': 1077, 'loss/train': 5.885529041290283} 01/27/2022 20:08:17 - INFO - codeparrot_training - Step 1078: {'lr': 0.00026950000000000005, 'samples': 207168, 'steps': 1078, 'loss/train': 6.198288917541504} 01/27/2022 20:08:22 - INFO - codeparrot_training - Step 1079: {'lr': 0.00026974999999999997, 'samples': 207360, 'steps': 1079, 'loss/train': 6.992666959762573} 01/27/2022 20:08:27 - INFO - codeparrot_training - Step 1080: {'lr': 0.00027, 'samples': 207552, 'steps': 1080, 'loss/train': 6.471394300460815} 01/27/2022 20:08:31 - INFO - codeparrot_training - Step 1081: {'lr': 0.00027025, 'samples': 207744, 'steps': 1081, 'loss/train': 6.9427735805511475} 01/27/2022 20:08:35 - INFO - codeparrot_training - Step 1082: {'lr': 0.0002705, 'samples': 207936, 'steps': 1082, 'loss/train': 7.291034460067749} 01/27/2022 20:08:40 - INFO - codeparrot_training - Step 1083: {'lr': 0.00027075, 'samples': 208128, 'steps': 1083, 'loss/train': 7.425044059753418} 01/27/2022 20:08:44 - INFO - codeparrot_training - Step 1084: {'lr': 0.00027100000000000003, 'samples': 208320, 'steps': 1084, 'loss/train': 6.592901945114136} 01/27/2022 20:08:49 - INFO - codeparrot_training - Step 1085: {'lr': 0.00027125, 'samples': 208512, 'steps': 1085, 'loss/train': 6.1694605350494385} 01/27/2022 20:08:53 - INFO - codeparrot_training - Step 1086: {'lr': 0.00027150000000000004, 'samples': 208704, 'steps': 1086, 'loss/train': 6.886749029159546} 01/27/2022 20:08:57 - INFO - codeparrot_training - Step 1087: {'lr': 0.00027175, 'samples': 208896, 'steps': 1087, 'loss/train': 6.601865530014038} 01/27/2022 20:09:01 - INFO - codeparrot_training - Step 1088: {'lr': 0.00027200000000000005, 'samples': 209088, 'steps': 1088, 'loss/train': 7.312464952468872} 01/27/2022 20:09:05 - INFO - codeparrot_training - Step 1089: {'lr': 0.00027225, 'samples': 209280, 'steps': 1089, 'loss/train': 7.383501291275024} 01/27/2022 20:09:10 - INFO - codeparrot_training - Step 1090: {'lr': 0.0002725, 'samples': 209472, 'steps': 1090, 'loss/train': 7.0864784717559814} 01/27/2022 20:09:15 - INFO - codeparrot_training - Step 1091: {'lr': 0.00027275, 'samples': 209664, 'steps': 1091, 'loss/train': 5.819867134094238} 01/27/2022 20:09:19 - INFO - codeparrot_training - Step 1092: {'lr': 0.000273, 'samples': 209856, 'steps': 1092, 'loss/train': 5.9724122285842896} 01/27/2022 20:09:23 - INFO - codeparrot_training - Step 1093: {'lr': 0.00027325, 'samples': 210048, 'steps': 1093, 'loss/train': 6.062058448791504} 01/27/2022 20:09:27 - INFO - codeparrot_training - Step 1094: {'lr': 0.00027350000000000003, 'samples': 210240, 'steps': 1094, 'loss/train': 5.281084299087524} 01/27/2022 20:09:33 - INFO - codeparrot_training - Step 1095: {'lr': 0.00027375, 'samples': 210432, 'steps': 1095, 'loss/train': 6.791861057281494} 01/27/2022 20:09:38 - INFO - codeparrot_training - Step 1096: {'lr': 0.00027400000000000005, 'samples': 210624, 'steps': 1096, 'loss/train': 5.105173945426941} 01/27/2022 20:09:42 - INFO - codeparrot_training - Step 1097: {'lr': 0.00027425, 'samples': 210816, 'steps': 1097, 'loss/train': 6.027201175689697} 01/27/2022 20:09:46 - INFO - codeparrot_training - Step 1098: {'lr': 0.0002745, 'samples': 211008, 'steps': 1098, 'loss/train': 4.949175596237183} 01/27/2022 20:09:50 - INFO - codeparrot_training - Step 1099: {'lr': 0.00027475, 'samples': 211200, 'steps': 1099, 'loss/train': 7.246063470840454} 01/27/2022 20:09:55 - INFO - codeparrot_training - Step 1100: {'lr': 0.000275, 'samples': 211392, 'steps': 1100, 'loss/train': 5.314031481742859} 01/27/2022 20:09:59 - INFO - codeparrot_training - Step 1101: {'lr': 0.00027525, 'samples': 211584, 'steps': 1101, 'loss/train': 7.240302801132202} 01/27/2022 20:10:03 - INFO - codeparrot_training - Step 1102: {'lr': 0.00027550000000000003, 'samples': 211776, 'steps': 1102, 'loss/train': 7.208335876464844} 01/27/2022 20:10:08 - INFO - codeparrot_training - Step 1103: {'lr': 0.00027575, 'samples': 211968, 'steps': 1103, 'loss/train': 6.179112195968628} 01/27/2022 20:10:12 - INFO - codeparrot_training - Step 1104: {'lr': 0.00027600000000000004, 'samples': 212160, 'steps': 1104, 'loss/train': 6.805338621139526} 01/27/2022 20:10:18 - INFO - codeparrot_training - Step 1105: {'lr': 0.00027625, 'samples': 212352, 'steps': 1105, 'loss/train': 5.8247305154800415} 01/27/2022 20:10:22 - INFO - codeparrot_training - Step 1106: {'lr': 0.00027650000000000005, 'samples': 212544, 'steps': 1106, 'loss/train': 7.093969345092773} 01/27/2022 20:10:26 - INFO - codeparrot_training - Step 1107: {'lr': 0.00027675, 'samples': 212736, 'steps': 1107, 'loss/train': 6.224668979644775} 01/27/2022 20:10:30 - INFO - codeparrot_training - Step 1108: {'lr': 0.000277, 'samples': 212928, 'steps': 1108, 'loss/train': 6.302997350692749} 01/27/2022 20:10:34 - INFO - codeparrot_training - Step 1109: {'lr': 0.00027725, 'samples': 213120, 'steps': 1109, 'loss/train': 7.30902886390686} 01/27/2022 20:10:39 - INFO - codeparrot_training - Step 1110: {'lr': 0.0002775, 'samples': 213312, 'steps': 1110, 'loss/train': 6.090117931365967} 01/27/2022 20:10:43 - INFO - codeparrot_training - Step 1111: {'lr': 0.00027775, 'samples': 213504, 'steps': 1111, 'loss/train': 6.622694492340088} 01/27/2022 20:10:48 - INFO - codeparrot_training - Step 1112: {'lr': 0.00027800000000000004, 'samples': 213696, 'steps': 1112, 'loss/train': 6.511759042739868} 01/27/2022 20:10:52 - INFO - codeparrot_training - Step 1113: {'lr': 0.00027825, 'samples': 213888, 'steps': 1113, 'loss/train': 6.381570339202881} 01/27/2022 20:10:56 - INFO - codeparrot_training - Step 1114: {'lr': 0.00027850000000000005, 'samples': 214080, 'steps': 1114, 'loss/train': 5.967792391777039} 01/27/2022 20:11:01 - INFO - codeparrot_training - Step 1115: {'lr': 0.00027875, 'samples': 214272, 'steps': 1115, 'loss/train': 8.538374662399292} 01/27/2022 20:11:05 - INFO - codeparrot_training - Step 1116: {'lr': 0.000279, 'samples': 214464, 'steps': 1116, 'loss/train': 5.950817227363586} 01/27/2022 20:11:09 - INFO - codeparrot_training - Step 1117: {'lr': 0.00027925, 'samples': 214656, 'steps': 1117, 'loss/train': 7.775706052780151} 01/27/2022 20:11:13 - INFO - codeparrot_training - Step 1118: {'lr': 0.0002795, 'samples': 214848, 'steps': 1118, 'loss/train': 6.328991889953613} 01/27/2022 20:11:18 - INFO - codeparrot_training - Step 1119: {'lr': 0.00027975, 'samples': 215040, 'steps': 1119, 'loss/train': 7.466660499572754} 01/27/2022 20:11:24 - INFO - codeparrot_training - Step 1120: {'lr': 0.00028000000000000003, 'samples': 215232, 'steps': 1120, 'loss/train': 6.498028993606567} 01/27/2022 20:11:28 - INFO - codeparrot_training - Step 1121: {'lr': 0.00028025, 'samples': 215424, 'steps': 1121, 'loss/train': 6.444835424423218} 01/27/2022 20:11:32 - INFO - codeparrot_training - Step 1122: {'lr': 0.00028050000000000004, 'samples': 215616, 'steps': 1122, 'loss/train': 6.599710464477539} 01/27/2022 20:11:36 - INFO - codeparrot_training - Step 1123: {'lr': 0.00028075, 'samples': 215808, 'steps': 1123, 'loss/train': 6.3989417552948} 01/27/2022 20:11:40 - INFO - codeparrot_training - Step 1124: {'lr': 0.00028100000000000005, 'samples': 216000, 'steps': 1124, 'loss/train': 7.14109468460083} 01/27/2022 20:11:45 - INFO - codeparrot_training - Step 1125: {'lr': 0.00028125000000000003, 'samples': 216192, 'steps': 1125, 'loss/train': 6.869213104248047} 01/27/2022 20:11:50 - INFO - codeparrot_training - Step 1126: {'lr': 0.00028149999999999996, 'samples': 216384, 'steps': 1126, 'loss/train': 6.086499452590942} 01/27/2022 20:11:54 - INFO - codeparrot_training - Step 1127: {'lr': 0.00028175, 'samples': 216576, 'steps': 1127, 'loss/train': 6.089702367782593} 01/27/2022 20:11:58 - INFO - codeparrot_training - Step 1128: {'lr': 0.00028199999999999997, 'samples': 216768, 'steps': 1128, 'loss/train': 6.265631675720215} 01/27/2022 20:12:02 - INFO - codeparrot_training - Step 1129: {'lr': 0.00028225, 'samples': 216960, 'steps': 1129, 'loss/train': 6.024996042251587} 01/27/2022 20:12:07 - INFO - codeparrot_training - Step 1130: {'lr': 0.0002825, 'samples': 217152, 'steps': 1130, 'loss/train': 6.219256639480591} 01/27/2022 20:12:11 - INFO - codeparrot_training - Step 1131: {'lr': 0.00028275, 'samples': 217344, 'steps': 1131, 'loss/train': 7.128313779830933} 01/27/2022 20:12:16 - INFO - codeparrot_training - Step 1132: {'lr': 0.000283, 'samples': 217536, 'steps': 1132, 'loss/train': 5.9960854053497314} 01/27/2022 20:12:20 - INFO - codeparrot_training - Step 1133: {'lr': 0.00028325000000000003, 'samples': 217728, 'steps': 1133, 'loss/train': 6.278270244598389} 01/27/2022 20:12:26 - INFO - codeparrot_training - Step 1134: {'lr': 0.0002835, 'samples': 217920, 'steps': 1134, 'loss/train': 5.583935737609863} 01/27/2022 20:12:30 - INFO - codeparrot_training - Step 1135: {'lr': 0.00028375, 'samples': 218112, 'steps': 1135, 'loss/train': 5.667274475097656} 01/27/2022 20:12:34 - INFO - codeparrot_training - Step 1136: {'lr': 0.00028399999999999996, 'samples': 218304, 'steps': 1136, 'loss/train': 5.555999279022217} 01/27/2022 20:12:38 - INFO - codeparrot_training - Step 1137: {'lr': 0.00028425, 'samples': 218496, 'steps': 1137, 'loss/train': 5.911423087120056} 01/27/2022 20:12:42 - INFO - codeparrot_training - Step 1138: {'lr': 0.0002845, 'samples': 218688, 'steps': 1138, 'loss/train': 6.839578628540039} 01/27/2022 20:12:47 - INFO - codeparrot_training - Step 1139: {'lr': 0.00028475, 'samples': 218880, 'steps': 1139, 'loss/train': 6.366803884506226} 01/27/2022 20:12:52 - INFO - codeparrot_training - Step 1140: {'lr': 0.000285, 'samples': 219072, 'steps': 1140, 'loss/train': 6.293874263763428} 01/27/2022 20:12:56 - INFO - codeparrot_training - Step 1141: {'lr': 0.00028525, 'samples': 219264, 'steps': 1141, 'loss/train': 6.04304051399231} 01/27/2022 20:13:00 - INFO - codeparrot_training - Step 1142: {'lr': 0.0002855, 'samples': 219456, 'steps': 1142, 'loss/train': 7.324260234832764} 01/27/2022 20:13:04 - INFO - codeparrot_training - Step 1143: {'lr': 0.00028575000000000003, 'samples': 219648, 'steps': 1143, 'loss/train': 6.565672874450684} 01/27/2022 20:13:09 - INFO - codeparrot_training - Step 1144: {'lr': 0.00028599999999999996, 'samples': 219840, 'steps': 1144, 'loss/train': 5.621720910072327} 01/27/2022 20:13:13 - INFO - codeparrot_training - Step 1145: {'lr': 0.00028625, 'samples': 220032, 'steps': 1145, 'loss/train': 7.351043701171875} 01/27/2022 20:13:18 - INFO - codeparrot_training - Step 1146: {'lr': 0.00028649999999999997, 'samples': 220224, 'steps': 1146, 'loss/train': 6.070271015167236} 01/27/2022 20:13:22 - INFO - codeparrot_training - Step 1147: {'lr': 0.00028675, 'samples': 220416, 'steps': 1147, 'loss/train': 8.005390405654907} 01/27/2022 20:13:26 - INFO - codeparrot_training - Step 1148: {'lr': 0.000287, 'samples': 220608, 'steps': 1148, 'loss/train': 6.678696155548096} 01/27/2022 20:13:32 - INFO - codeparrot_training - Step 1149: {'lr': 0.00028725, 'samples': 220800, 'steps': 1149, 'loss/train': 6.863654851913452} 01/27/2022 20:13:37 - INFO - codeparrot_training - Step 1150: {'lr': 0.0002875, 'samples': 220992, 'steps': 1150, 'loss/train': 6.061199426651001} 01/27/2022 20:13:41 - INFO - codeparrot_training - Step 1151: {'lr': 0.00028775000000000003, 'samples': 221184, 'steps': 1151, 'loss/train': 5.310739159584045} 01/27/2022 20:13:45 - INFO - codeparrot_training - Step 1152: {'lr': 0.000288, 'samples': 221376, 'steps': 1152, 'loss/train': 6.548009634017944} 01/27/2022 20:13:49 - INFO - codeparrot_training - Step 1153: {'lr': 0.00028825, 'samples': 221568, 'steps': 1153, 'loss/train': 4.907338500022888} 01/27/2022 20:13:53 - INFO - codeparrot_training - Step 1154: {'lr': 0.00028849999999999997, 'samples': 221760, 'steps': 1154, 'loss/train': 5.8318716287612915} 01/27/2022 20:13:58 - INFO - codeparrot_training - Step 1155: {'lr': 0.00028875, 'samples': 221952, 'steps': 1155, 'loss/train': 5.536755681037903} 01/27/2022 20:14:02 - INFO - codeparrot_training - Step 1156: {'lr': 0.000289, 'samples': 222144, 'steps': 1156, 'loss/train': 6.573516368865967} 01/27/2022 20:14:06 - INFO - codeparrot_training - Step 1157: {'lr': 0.00028925, 'samples': 222336, 'steps': 1157, 'loss/train': 9.802571296691895} 01/27/2022 20:14:11 - INFO - codeparrot_training - Step 1158: {'lr': 0.0002895, 'samples': 222528, 'steps': 1158, 'loss/train': 6.779712438583374} 01/27/2022 20:14:15 - INFO - codeparrot_training - Step 1159: {'lr': 0.00028975, 'samples': 222720, 'steps': 1159, 'loss/train': 5.3718788623809814} 01/27/2022 20:14:20 - INFO - codeparrot_training - Step 1160: {'lr': 0.00029, 'samples': 222912, 'steps': 1160, 'loss/train': 6.039066553115845} 01/27/2022 20:14:24 - INFO - codeparrot_training - Step 1161: {'lr': 0.00029025000000000003, 'samples': 223104, 'steps': 1161, 'loss/train': 6.990618467330933} 01/27/2022 20:14:28 - INFO - codeparrot_training - Step 1162: {'lr': 0.00029049999999999996, 'samples': 223296, 'steps': 1162, 'loss/train': 6.2342870235443115} 01/27/2022 20:14:32 - INFO - codeparrot_training - Step 1163: {'lr': 0.00029075, 'samples': 223488, 'steps': 1163, 'loss/train': 6.3156702518463135} 01/27/2022 20:14:37 - INFO - codeparrot_training - Step 1164: {'lr': 0.00029099999999999997, 'samples': 223680, 'steps': 1164, 'loss/train': 6.798289775848389} 01/27/2022 20:14:43 - INFO - codeparrot_training - Step 1165: {'lr': 0.00029125, 'samples': 223872, 'steps': 1165, 'loss/train': 6.069545030593872} 01/27/2022 20:14:47 - INFO - codeparrot_training - Step 1166: {'lr': 0.0002915, 'samples': 224064, 'steps': 1166, 'loss/train': 7.147094249725342} 01/27/2022 20:14:51 - INFO - codeparrot_training - Step 1167: {'lr': 0.00029175, 'samples': 224256, 'steps': 1167, 'loss/train': 6.279296636581421} 01/27/2022 20:14:55 - INFO - codeparrot_training - Step 1168: {'lr': 0.000292, 'samples': 224448, 'steps': 1168, 'loss/train': 1.8390909433364868} 01/27/2022 20:14:59 - INFO - codeparrot_training - Step 1169: {'lr': 0.00029225000000000003, 'samples': 224640, 'steps': 1169, 'loss/train': 6.530015230178833} 01/27/2022 20:15:05 - INFO - codeparrot_training - Step 1170: {'lr': 0.0002925, 'samples': 224832, 'steps': 1170, 'loss/train': 6.314259052276611} 01/27/2022 20:15:09 - INFO - codeparrot_training - Step 1171: {'lr': 0.00029275000000000004, 'samples': 225024, 'steps': 1171, 'loss/train': 5.534663915634155} 01/27/2022 20:15:13 - INFO - codeparrot_training - Step 1172: {'lr': 0.00029299999999999997, 'samples': 225216, 'steps': 1172, 'loss/train': 4.3329784870147705} 01/27/2022 20:15:17 - INFO - codeparrot_training - Step 1173: {'lr': 0.00029325, 'samples': 225408, 'steps': 1173, 'loss/train': 6.79440450668335} 01/27/2022 20:15:21 - INFO - codeparrot_training - Step 1174: {'lr': 0.0002935, 'samples': 225600, 'steps': 1174, 'loss/train': 6.60332179069519} 01/27/2022 20:15:26 - INFO - codeparrot_training - Step 1175: {'lr': 0.00029375, 'samples': 225792, 'steps': 1175, 'loss/train': 6.518944501876831} 01/27/2022 20:15:31 - INFO - codeparrot_training - Step 1176: {'lr': 0.000294, 'samples': 225984, 'steps': 1176, 'loss/train': 5.771566987037659} 01/27/2022 20:15:35 - INFO - codeparrot_training - Step 1177: {'lr': 0.00029425, 'samples': 226176, 'steps': 1177, 'loss/train': 5.747413158416748} 01/27/2022 20:15:39 - INFO - codeparrot_training - Step 1178: {'lr': 0.0002945, 'samples': 226368, 'steps': 1178, 'loss/train': 6.580724000930786} 01/27/2022 20:15:43 - INFO - codeparrot_training - Step 1179: {'lr': 0.00029475000000000004, 'samples': 226560, 'steps': 1179, 'loss/train': 4.977144241333008} 01/27/2022 20:15:49 - INFO - codeparrot_training - Step 1180: {'lr': 0.000295, 'samples': 226752, 'steps': 1180, 'loss/train': 6.731553554534912} 01/27/2022 20:15:53 - INFO - codeparrot_training - Step 1181: {'lr': 0.00029525, 'samples': 226944, 'steps': 1181, 'loss/train': 7.82481837272644} 01/27/2022 20:15:58 - INFO - codeparrot_training - Step 1182: {'lr': 0.00029549999999999997, 'samples': 227136, 'steps': 1182, 'loss/train': 6.1408281326293945} 01/27/2022 20:16:02 - INFO - codeparrot_training - Step 1183: {'lr': 0.00029575, 'samples': 227328, 'steps': 1183, 'loss/train': 5.9980491399765015} 01/27/2022 20:16:07 - INFO - codeparrot_training - Step 1184: {'lr': 0.000296, 'samples': 227520, 'steps': 1184, 'loss/train': 6.043874502182007} 01/27/2022 20:16:11 - INFO - codeparrot_training - Step 1185: {'lr': 0.00029625, 'samples': 227712, 'steps': 1185, 'loss/train': 4.868932127952576} 01/27/2022 20:16:15 - INFO - codeparrot_training - Step 1186: {'lr': 0.0002965, 'samples': 227904, 'steps': 1186, 'loss/train': 6.037048816680908} 01/27/2022 20:16:19 - INFO - codeparrot_training - Step 1187: {'lr': 0.00029675000000000003, 'samples': 228096, 'steps': 1187, 'loss/train': 6.044588327407837} 01/27/2022 20:16:24 - INFO - codeparrot_training - Step 1188: {'lr': 0.000297, 'samples': 228288, 'steps': 1188, 'loss/train': 6.625821590423584} 01/27/2022 20:16:28 - INFO - codeparrot_training - Step 1189: {'lr': 0.00029725000000000004, 'samples': 228480, 'steps': 1189, 'loss/train': 5.946741342544556} 01/27/2022 20:16:34 - INFO - codeparrot_training - Step 1190: {'lr': 0.00029749999999999997, 'samples': 228672, 'steps': 1190, 'loss/train': 6.38441276550293} 01/27/2022 20:16:38 - INFO - codeparrot_training - Step 1191: {'lr': 0.00029775, 'samples': 228864, 'steps': 1191, 'loss/train': 6.306578636169434} 01/27/2022 20:16:42 - INFO - codeparrot_training - Step 1192: {'lr': 0.000298, 'samples': 229056, 'steps': 1192, 'loss/train': 6.502519369125366} 01/27/2022 20:16:46 - INFO - codeparrot_training - Step 1193: {'lr': 0.00029825, 'samples': 229248, 'steps': 1193, 'loss/train': 6.899967670440674} 01/27/2022 20:16:51 - INFO - codeparrot_training - Step 1194: {'lr': 0.0002985, 'samples': 229440, 'steps': 1194, 'loss/train': 6.601760387420654} 01/27/2022 20:16:55 - INFO - codeparrot_training - Step 1195: {'lr': 0.00029875, 'samples': 229632, 'steps': 1195, 'loss/train': 6.1342384815216064} 01/27/2022 20:16:59 - INFO - codeparrot_training - Step 1196: {'lr': 0.000299, 'samples': 229824, 'steps': 1196, 'loss/train': 6.967541456222534} 01/27/2022 20:17:04 - INFO - codeparrot_training - Step 1197: {'lr': 0.00029925000000000004, 'samples': 230016, 'steps': 1197, 'loss/train': 6.511771202087402} 01/27/2022 20:17:08 - INFO - codeparrot_training - Step 1198: {'lr': 0.0002995, 'samples': 230208, 'steps': 1198, 'loss/train': 6.0717387199401855} 01/27/2022 20:17:13 - INFO - codeparrot_training - Step 1199: {'lr': 0.00029975000000000005, 'samples': 230400, 'steps': 1199, 'loss/train': 4.923880577087402} 01/27/2022 20:17:17 - INFO - codeparrot_training - Step 1200: {'lr': 0.0003, 'samples': 230592, 'steps': 1200, 'loss/train': 6.300751447677612} 01/27/2022 20:17:21 - INFO - codeparrot_training - Step 1201: {'lr': 0.00030025, 'samples': 230784, 'steps': 1201, 'loss/train': 5.998017311096191} 01/27/2022 20:17:25 - INFO - codeparrot_training - Step 1202: {'lr': 0.0003005, 'samples': 230976, 'steps': 1202, 'loss/train': 5.30766499042511} 01/27/2022 20:17:30 - INFO - codeparrot_training - Step 1203: {'lr': 0.00030075, 'samples': 231168, 'steps': 1203, 'loss/train': 6.3577001094818115} 01/27/2022 20:17:35 - INFO - codeparrot_training - Step 1204: {'lr': 0.000301, 'samples': 231360, 'steps': 1204, 'loss/train': 6.341496706008911} 01/27/2022 20:17:39 - INFO - codeparrot_training - Step 1205: {'lr': 0.00030125000000000003, 'samples': 231552, 'steps': 1205, 'loss/train': 7.376644134521484} 01/27/2022 20:17:43 - INFO - codeparrot_training - Step 1206: {'lr': 0.0003015, 'samples': 231744, 'steps': 1206, 'loss/train': 6.280597686767578} 01/27/2022 20:17:47 - INFO - codeparrot_training - Step 1207: {'lr': 0.00030175000000000004, 'samples': 231936, 'steps': 1207, 'loss/train': 6.872528314590454} 01/27/2022 20:17:51 - INFO - codeparrot_training - Step 1208: {'lr': 0.000302, 'samples': 232128, 'steps': 1208, 'loss/train': 6.061391115188599} 01/27/2022 20:17:57 - INFO - codeparrot_training - Step 1209: {'lr': 0.00030225, 'samples': 232320, 'steps': 1209, 'loss/train': 5.210599422454834} 01/27/2022 20:18:01 - INFO - codeparrot_training - Step 1210: {'lr': 0.0003025, 'samples': 232512, 'steps': 1210, 'loss/train': 5.790054202079773} 01/27/2022 20:18:06 - INFO - codeparrot_training - Step 1211: {'lr': 0.00030275, 'samples': 232704, 'steps': 1211, 'loss/train': 5.796718597412109} 01/27/2022 20:18:10 - INFO - codeparrot_training - Step 1212: {'lr': 0.000303, 'samples': 232896, 'steps': 1212, 'loss/train': 4.802600026130676} 01/27/2022 20:18:14 - INFO - codeparrot_training - Step 1213: {'lr': 0.00030325, 'samples': 233088, 'steps': 1213, 'loss/train': 6.613306045532227} 01/27/2022 20:18:19 - INFO - codeparrot_training - Step 1214: {'lr': 0.0003035, 'samples': 233280, 'steps': 1214, 'loss/train': 5.217711925506592} 01/27/2022 20:18:23 - INFO - codeparrot_training - Step 1215: {'lr': 0.00030375000000000004, 'samples': 233472, 'steps': 1215, 'loss/train': 5.135290503501892} 01/27/2022 20:18:28 - INFO - codeparrot_training - Step 1216: {'lr': 0.000304, 'samples': 233664, 'steps': 1216, 'loss/train': 5.579798698425293} 01/27/2022 20:18:32 - INFO - codeparrot_training - Step 1217: {'lr': 0.00030425000000000005, 'samples': 233856, 'steps': 1217, 'loss/train': 4.471961617469788} 01/27/2022 20:18:36 - INFO - codeparrot_training - Step 1218: {'lr': 0.0003045, 'samples': 234048, 'steps': 1218, 'loss/train': 5.69901180267334} 01/27/2022 20:18:41 - INFO - codeparrot_training - Step 1219: {'lr': 0.00030475, 'samples': 234240, 'steps': 1219, 'loss/train': 5.395376443862915} 01/27/2022 20:18:45 - INFO - codeparrot_training - Step 1220: {'lr': 0.000305, 'samples': 234432, 'steps': 1220, 'loss/train': 6.262244939804077} 01/27/2022 20:18:49 - INFO - codeparrot_training - Step 1221: {'lr': 0.00030525, 'samples': 234624, 'steps': 1221, 'loss/train': 5.507705926895142} 01/27/2022 20:18:54 - INFO - codeparrot_training - Step 1222: {'lr': 0.0003055, 'samples': 234816, 'steps': 1222, 'loss/train': 5.467741012573242} 01/27/2022 20:18:58 - INFO - codeparrot_training - Step 1223: {'lr': 0.00030575000000000003, 'samples': 235008, 'steps': 1223, 'loss/train': 7.247064113616943} 01/27/2022 20:19:04 - INFO - codeparrot_training - Step 1224: {'lr': 0.000306, 'samples': 235200, 'steps': 1224, 'loss/train': 5.343425273895264} 01/27/2022 20:19:08 - INFO - codeparrot_training - Step 1225: {'lr': 0.00030625000000000004, 'samples': 235392, 'steps': 1225, 'loss/train': 5.71028745174408} 01/27/2022 20:19:12 - INFO - codeparrot_training - Step 1226: {'lr': 0.0003065, 'samples': 235584, 'steps': 1226, 'loss/train': 5.828846454620361} 01/27/2022 20:19:16 - INFO - codeparrot_training - Step 1227: {'lr': 0.00030675, 'samples': 235776, 'steps': 1227, 'loss/train': 6.56818699836731} 01/27/2022 20:19:20 - INFO - codeparrot_training - Step 1228: {'lr': 0.000307, 'samples': 235968, 'steps': 1228, 'loss/train': 6.041790246963501} 01/27/2022 20:19:25 - INFO - codeparrot_training - Step 1229: {'lr': 0.00030725, 'samples': 236160, 'steps': 1229, 'loss/train': 6.172706365585327} 01/27/2022 20:19:29 - INFO - codeparrot_training - Step 1230: {'lr': 0.0003075, 'samples': 236352, 'steps': 1230, 'loss/train': 6.19190239906311} 01/27/2022 20:19:34 - INFO - codeparrot_training - Step 1231: {'lr': 0.00030775, 'samples': 236544, 'steps': 1231, 'loss/train': 6.197120904922485} 01/27/2022 20:19:38 - INFO - codeparrot_training - Step 1232: {'lr': 0.000308, 'samples': 236736, 'steps': 1232, 'loss/train': 6.165764093399048} 01/27/2022 20:19:42 - INFO - codeparrot_training - Step 1233: {'lr': 0.00030825000000000004, 'samples': 236928, 'steps': 1233, 'loss/train': 5.548623561859131} 01/27/2022 20:19:49 - INFO - codeparrot_training - Step 1234: {'lr': 0.0003085, 'samples': 237120, 'steps': 1234, 'loss/train': 7.033030986785889} 01/27/2022 20:19:53 - INFO - codeparrot_training - Step 1235: {'lr': 0.00030875000000000005, 'samples': 237312, 'steps': 1235, 'loss/train': 6.1687610149383545} 01/27/2022 20:19:57 - INFO - codeparrot_training - Step 1236: {'lr': 0.00030900000000000003, 'samples': 237504, 'steps': 1236, 'loss/train': 5.911296129226685} 01/27/2022 20:20:01 - INFO - codeparrot_training - Step 1237: {'lr': 0.00030925, 'samples': 237696, 'steps': 1237, 'loss/train': 5.624135255813599} 01/27/2022 20:20:05 - INFO - codeparrot_training - Step 1238: {'lr': 0.0003095, 'samples': 237888, 'steps': 1238, 'loss/train': 5.627055287361145} 01/27/2022 20:20:11 - INFO - codeparrot_training - Step 1239: {'lr': 0.00030975, 'samples': 238080, 'steps': 1239, 'loss/train': 7.038751602172852} 01/27/2022 20:20:15 - INFO - codeparrot_training - Step 1240: {'lr': 0.00031, 'samples': 238272, 'steps': 1240, 'loss/train': 4.588157057762146} 01/27/2022 20:20:19 - INFO - codeparrot_training - Step 1241: {'lr': 0.00031025000000000003, 'samples': 238464, 'steps': 1241, 'loss/train': 5.421356678009033} 01/27/2022 20:20:23 - INFO - codeparrot_training - Step 1242: {'lr': 0.0003105, 'samples': 238656, 'steps': 1242, 'loss/train': 5.470308423042297} 01/27/2022 20:20:27 - INFO - codeparrot_training - Step 1243: {'lr': 0.00031075000000000005, 'samples': 238848, 'steps': 1243, 'loss/train': 5.879084944725037} 01/27/2022 20:20:32 - INFO - codeparrot_training - Step 1244: {'lr': 0.000311, 'samples': 239040, 'steps': 1244, 'loss/train': 6.332687616348267} 01/27/2022 20:20:37 - INFO - codeparrot_training - Step 1245: {'lr': 0.00031125000000000006, 'samples': 239232, 'steps': 1245, 'loss/train': 5.388120174407959} 01/27/2022 20:20:41 - INFO - codeparrot_training - Step 1246: {'lr': 0.0003115, 'samples': 239424, 'steps': 1246, 'loss/train': 5.9562599658966064} 01/27/2022 20:20:45 - INFO - codeparrot_training - Step 1247: {'lr': 0.00031175, 'samples': 239616, 'steps': 1247, 'loss/train': 5.953097820281982} 01/27/2022 20:20:49 - INFO - codeparrot_training - Step 1248: {'lr': 0.000312, 'samples': 239808, 'steps': 1248, 'loss/train': 5.735149383544922} 01/27/2022 20:20:55 - INFO - codeparrot_training - Step 1249: {'lr': 0.00031225000000000003, 'samples': 240000, 'steps': 1249, 'loss/train': 5.868349313735962} 01/27/2022 20:20:59 - INFO - codeparrot_training - Step 1250: {'lr': 0.0003125, 'samples': 240192, 'steps': 1250, 'loss/train': 5.6702492237091064} 01/27/2022 20:21:03 - INFO - codeparrot_training - Step 1251: {'lr': 0.00031275, 'samples': 240384, 'steps': 1251, 'loss/train': 5.804202318191528} 01/27/2022 20:21:07 - INFO - codeparrot_training - Step 1252: {'lr': 0.000313, 'samples': 240576, 'steps': 1252, 'loss/train': 5.520266175270081} 01/27/2022 20:21:12 - INFO - codeparrot_training - Step 1253: {'lr': 0.00031325, 'samples': 240768, 'steps': 1253, 'loss/train': 6.052096366882324} 01/27/2022 20:21:17 - INFO - codeparrot_training - Step 1254: {'lr': 0.00031350000000000003, 'samples': 240960, 'steps': 1254, 'loss/train': 6.466270208358765} 01/27/2022 20:21:21 - INFO - codeparrot_training - Step 1255: {'lr': 0.00031374999999999996, 'samples': 241152, 'steps': 1255, 'loss/train': 5.624329090118408} 01/27/2022 20:21:25 - INFO - codeparrot_training - Step 1256: {'lr': 0.000314, 'samples': 241344, 'steps': 1256, 'loss/train': 6.50145149230957} 01/27/2022 20:21:29 - INFO - codeparrot_training - Step 1257: {'lr': 0.00031424999999999997, 'samples': 241536, 'steps': 1257, 'loss/train': 5.034597873687744} 01/27/2022 20:21:33 - INFO - codeparrot_training - Step 1258: {'lr': 0.0003145, 'samples': 241728, 'steps': 1258, 'loss/train': 5.814113616943359} 01/27/2022 20:21:38 - INFO - codeparrot_training - Step 1259: {'lr': 0.00031475, 'samples': 241920, 'steps': 1259, 'loss/train': 5.415977954864502} 01/27/2022 20:21:42 - INFO - codeparrot_training - Step 1260: {'lr': 0.000315, 'samples': 242112, 'steps': 1260, 'loss/train': 6.592894077301025} 01/27/2022 20:21:47 - INFO - codeparrot_training - Step 1261: {'lr': 0.00031525, 'samples': 242304, 'steps': 1261, 'loss/train': 5.853173732757568} 01/27/2022 20:21:51 - INFO - codeparrot_training - Step 1262: {'lr': 0.0003155, 'samples': 242496, 'steps': 1262, 'loss/train': 2.980490505695343} 01/27/2022 20:21:55 - INFO - codeparrot_training - Step 1263: {'lr': 0.00031575, 'samples': 242688, 'steps': 1263, 'loss/train': 5.863975882530212} 01/27/2022 20:22:00 - INFO - codeparrot_training - Step 1264: {'lr': 0.000316, 'samples': 242880, 'steps': 1264, 'loss/train': 6.6285810470581055} 01/27/2022 20:22:04 - INFO - codeparrot_training - Step 1265: {'lr': 0.00031624999999999996, 'samples': 243072, 'steps': 1265, 'loss/train': 7.1148762702941895} 01/27/2022 20:22:08 - INFO - codeparrot_training - Step 1266: {'lr': 0.0003165, 'samples': 243264, 'steps': 1266, 'loss/train': 6.865442991256714} 01/27/2022 20:22:12 - INFO - codeparrot_training - Step 1267: {'lr': 0.00031675, 'samples': 243456, 'steps': 1267, 'loss/train': 5.841933846473694} 01/27/2022 20:22:17 - INFO - codeparrot_training - Step 1268: {'lr': 0.000317, 'samples': 243648, 'steps': 1268, 'loss/train': 5.91564416885376} 01/27/2022 20:22:22 - INFO - codeparrot_training - Step 1269: {'lr': 0.00031725, 'samples': 243840, 'steps': 1269, 'loss/train': 5.927930474281311} 01/27/2022 20:22:27 - INFO - codeparrot_training - Step 1270: {'lr': 0.0003175, 'samples': 244032, 'steps': 1270, 'loss/train': 5.584486484527588} 01/27/2022 20:22:31 - INFO - codeparrot_training - Step 1271: {'lr': 0.00031775, 'samples': 244224, 'steps': 1271, 'loss/train': 5.189220070838928} 01/27/2022 20:22:35 - INFO - codeparrot_training - Step 1272: {'lr': 0.00031800000000000003, 'samples': 244416, 'steps': 1272, 'loss/train': 5.846536874771118} 01/27/2022 20:22:39 - INFO - codeparrot_training - Step 1273: {'lr': 0.00031825, 'samples': 244608, 'steps': 1273, 'loss/train': 5.705549240112305} 01/27/2022 20:22:44 - INFO - codeparrot_training - Step 1274: {'lr': 0.0003185, 'samples': 244800, 'steps': 1274, 'loss/train': 5.478667616844177} 01/27/2022 20:22:48 - INFO - codeparrot_training - Step 1275: {'lr': 0.00031874999999999997, 'samples': 244992, 'steps': 1275, 'loss/train': 4.732180595397949} 01/27/2022 20:22:52 - INFO - codeparrot_training - Step 1276: {'lr': 0.000319, 'samples': 245184, 'steps': 1276, 'loss/train': 7.347651958465576} 01/27/2022 20:22:57 - INFO - codeparrot_training - Step 1277: {'lr': 0.00031925, 'samples': 245376, 'steps': 1277, 'loss/train': 6.617639064788818} 01/27/2022 20:23:01 - INFO - codeparrot_training - Step 1278: {'lr': 0.0003195, 'samples': 245568, 'steps': 1278, 'loss/train': 6.381918668746948} 01/27/2022 20:23:07 - INFO - codeparrot_training - Step 1279: {'lr': 0.00031975, 'samples': 245760, 'steps': 1279, 'loss/train': 3.523518204689026} 01/27/2022 20:23:11 - INFO - codeparrot_training - Step 1280: {'lr': 0.00032, 'samples': 245952, 'steps': 1280, 'loss/train': 6.2709832191467285} 01/27/2022 20:23:15 - INFO - codeparrot_training - Step 1281: {'lr': 0.00032025, 'samples': 246144, 'steps': 1281, 'loss/train': 5.995697021484375} 01/27/2022 20:23:19 - INFO - codeparrot_training - Step 1282: {'lr': 0.00032050000000000004, 'samples': 246336, 'steps': 1282, 'loss/train': 10.030723571777344} 01/27/2022 20:23:23 - INFO - codeparrot_training - Step 1283: {'lr': 0.00032074999999999996, 'samples': 246528, 'steps': 1283, 'loss/train': 6.351902961730957} 01/27/2022 20:23:29 - INFO - codeparrot_training - Step 1284: {'lr': 0.000321, 'samples': 246720, 'steps': 1284, 'loss/train': 6.770330429077148} 01/27/2022 20:23:33 - INFO - codeparrot_training - Step 1285: {'lr': 0.00032125, 'samples': 246912, 'steps': 1285, 'loss/train': 6.45099663734436} 01/27/2022 20:23:37 - INFO - codeparrot_training - Step 1286: {'lr': 0.0003215, 'samples': 247104, 'steps': 1286, 'loss/train': 8.37216567993164} 01/27/2022 20:23:41 - INFO - codeparrot_training - Step 1287: {'lr': 0.00032175, 'samples': 247296, 'steps': 1287, 'loss/train': 5.633511185646057} 01/27/2022 20:23:46 - INFO - codeparrot_training - Step 1288: {'lr': 0.000322, 'samples': 247488, 'steps': 1288, 'loss/train': 6.591914176940918} 01/27/2022 20:23:51 - INFO - codeparrot_training - Step 1289: {'lr': 0.00032225, 'samples': 247680, 'steps': 1289, 'loss/train': 6.548358678817749} 01/27/2022 20:23:55 - INFO - codeparrot_training - Step 1290: {'lr': 0.00032250000000000003, 'samples': 247872, 'steps': 1290, 'loss/train': 6.10328221321106} 01/27/2022 20:23:59 - INFO - codeparrot_training - Step 1291: {'lr': 0.00032275, 'samples': 248064, 'steps': 1291, 'loss/train': 5.338511824607849} 01/27/2022 20:24:03 - INFO - codeparrot_training - Step 1292: {'lr': 0.000323, 'samples': 248256, 'steps': 1292, 'loss/train': 6.172074794769287} 01/27/2022 20:24:07 - INFO - codeparrot_training - Step 1293: {'lr': 0.00032324999999999997, 'samples': 248448, 'steps': 1293, 'loss/train': 6.089951992034912} 01/27/2022 20:24:14 - INFO - codeparrot_training - Step 1294: {'lr': 0.0003235, 'samples': 248640, 'steps': 1294, 'loss/train': 6.235123157501221} 01/27/2022 20:24:18 - INFO - codeparrot_training - Step 1295: {'lr': 0.00032375, 'samples': 248832, 'steps': 1295, 'loss/train': 4.142842411994934} 01/27/2022 20:24:22 - INFO - codeparrot_training - Step 1296: {'lr': 0.000324, 'samples': 249024, 'steps': 1296, 'loss/train': 5.903172612190247} 01/27/2022 20:24:26 - INFO - codeparrot_training - Step 1297: {'lr': 0.00032425, 'samples': 249216, 'steps': 1297, 'loss/train': 6.823473930358887} 01/27/2022 20:24:30 - INFO - codeparrot_training - Step 1298: {'lr': 0.00032450000000000003, 'samples': 249408, 'steps': 1298, 'loss/train': 7.181246995925903} 01/27/2022 20:24:37 - INFO - codeparrot_training - Step 1299: {'lr': 0.00032475, 'samples': 249600, 'steps': 1299, 'loss/train': 9.384151697158813} 01/27/2022 20:24:41 - INFO - codeparrot_training - Step 1300: {'lr': 0.00032500000000000004, 'samples': 249792, 'steps': 1300, 'loss/train': 6.890306711196899} 01/27/2022 20:24:45 - INFO - codeparrot_training - Step 1301: {'lr': 0.00032524999999999996, 'samples': 249984, 'steps': 1301, 'loss/train': 4.543152451515198} 01/27/2022 20:24:49 - INFO - codeparrot_training - Step 1302: {'lr': 0.0003255, 'samples': 250176, 'steps': 1302, 'loss/train': 4.94231379032135} 01/27/2022 20:24:53 - INFO - codeparrot_training - Step 1303: {'lr': 0.00032575, 'samples': 250368, 'steps': 1303, 'loss/train': 6.2204976081848145} 01/27/2022 20:24:58 - INFO - codeparrot_training - Step 1304: {'lr': 0.000326, 'samples': 250560, 'steps': 1304, 'loss/train': 4.890730619430542} 01/27/2022 20:25:03 - INFO - codeparrot_training - Step 1305: {'lr': 0.00032625, 'samples': 250752, 'steps': 1305, 'loss/train': 7.002712726593018} 01/27/2022 20:25:07 - INFO - codeparrot_training - Step 1306: {'lr': 0.0003265, 'samples': 250944, 'steps': 1306, 'loss/train': 6.253939390182495} 01/27/2022 20:25:11 - INFO - codeparrot_training - Step 1307: {'lr': 0.00032675, 'samples': 251136, 'steps': 1307, 'loss/train': 6.320581912994385} 01/27/2022 20:25:15 - INFO - codeparrot_training - Step 1308: {'lr': 0.00032700000000000003, 'samples': 251328, 'steps': 1308, 'loss/train': 5.838821053504944} 01/27/2022 20:25:19 - INFO - codeparrot_training - Step 1309: {'lr': 0.00032725, 'samples': 251520, 'steps': 1309, 'loss/train': 6.843127012252808} 01/27/2022 20:25:25 - INFO - codeparrot_training - Step 1310: {'lr': 0.00032750000000000005, 'samples': 251712, 'steps': 1310, 'loss/train': 6.715729236602783} 01/27/2022 20:25:29 - INFO - codeparrot_training - Step 1311: {'lr': 0.00032774999999999997, 'samples': 251904, 'steps': 1311, 'loss/train': 5.719773888587952} 01/27/2022 20:25:33 - INFO - codeparrot_training - Step 1312: {'lr': 0.000328, 'samples': 252096, 'steps': 1312, 'loss/train': 6.019824743270874} 01/27/2022 20:25:37 - INFO - codeparrot_training - Step 1313: {'lr': 0.00032825, 'samples': 252288, 'steps': 1313, 'loss/train': 5.300379395484924} 01/27/2022 20:25:42 - INFO - codeparrot_training - Step 1314: {'lr': 0.0003285, 'samples': 252480, 'steps': 1314, 'loss/train': 6.472275495529175} 01/27/2022 20:25:47 - INFO - codeparrot_training - Step 1315: {'lr': 0.00032875, 'samples': 252672, 'steps': 1315, 'loss/train': 4.7848674058914185} 01/27/2022 20:25:51 - INFO - codeparrot_training - Step 1316: {'lr': 0.00032900000000000003, 'samples': 252864, 'steps': 1316, 'loss/train': 5.480200409889221} 01/27/2022 20:25:55 - INFO - codeparrot_training - Step 1317: {'lr': 0.00032925, 'samples': 253056, 'steps': 1317, 'loss/train': 6.194598913192749} 01/27/2022 20:26:00 - INFO - codeparrot_training - Step 1318: {'lr': 0.00032950000000000004, 'samples': 253248, 'steps': 1318, 'loss/train': 5.316754102706909} 01/27/2022 20:26:04 - INFO - codeparrot_training - Step 1319: {'lr': 0.00032975, 'samples': 253440, 'steps': 1319, 'loss/train': 10.881548881530762} 01/27/2022 20:26:09 - INFO - codeparrot_training - Step 1320: {'lr': 0.00033, 'samples': 253632, 'steps': 1320, 'loss/train': 5.247634291648865} 01/27/2022 20:26:13 - INFO - codeparrot_training - Step 1321: {'lr': 0.00033025, 'samples': 253824, 'steps': 1321, 'loss/train': 5.323326945304871} 01/27/2022 20:26:17 - INFO - codeparrot_training - Step 1322: {'lr': 0.0003305, 'samples': 254016, 'steps': 1322, 'loss/train': 5.004830002784729} 01/27/2022 20:26:21 - INFO - codeparrot_training - Step 1323: {'lr': 0.00033075, 'samples': 254208, 'steps': 1323, 'loss/train': 5.998468637466431} 01/27/2022 20:26:25 - INFO - codeparrot_training - Step 1324: {'lr': 0.000331, 'samples': 254400, 'steps': 1324, 'loss/train': 5.258039832115173} 01/27/2022 20:26:31 - INFO - codeparrot_training - Step 1325: {'lr': 0.00033125, 'samples': 254592, 'steps': 1325, 'loss/train': 5.178244113922119} 01/27/2022 20:26:35 - INFO - codeparrot_training - Step 1326: {'lr': 0.00033150000000000003, 'samples': 254784, 'steps': 1326, 'loss/train': 6.739606618881226} 01/27/2022 20:26:39 - INFO - codeparrot_training - Step 1327: {'lr': 0.00033175, 'samples': 254976, 'steps': 1327, 'loss/train': 5.981183409690857} 01/27/2022 20:26:44 - INFO - codeparrot_training - Step 1328: {'lr': 0.00033200000000000005, 'samples': 255168, 'steps': 1328, 'loss/train': 5.995321869850159} 01/27/2022 20:26:48 - INFO - codeparrot_training - Step 1329: {'lr': 0.00033224999999999997, 'samples': 255360, 'steps': 1329, 'loss/train': 5.820279121398926} 01/27/2022 20:26:53 - INFO - codeparrot_training - Step 1330: {'lr': 0.0003325, 'samples': 255552, 'steps': 1330, 'loss/train': 5.64120626449585} 01/27/2022 20:26:57 - INFO - codeparrot_training - Step 1331: {'lr': 0.00033275, 'samples': 255744, 'steps': 1331, 'loss/train': 5.612629294395447} 01/27/2022 20:27:01 - INFO - codeparrot_training - Step 1332: {'lr': 0.000333, 'samples': 255936, 'steps': 1332, 'loss/train': 6.206796884536743} 01/27/2022 20:27:05 - INFO - codeparrot_training - Step 1333: {'lr': 0.00033325, 'samples': 256128, 'steps': 1333, 'loss/train': 4.314150452613831} 01/27/2022 20:27:09 - INFO - codeparrot_training - Step 1334: {'lr': 0.00033350000000000003, 'samples': 256320, 'steps': 1334, 'loss/train': 5.413541078567505} 01/27/2022 20:27:15 - INFO - codeparrot_training - Step 1335: {'lr': 0.00033375, 'samples': 256512, 'steps': 1335, 'loss/train': 5.380462288856506} 01/27/2022 20:27:19 - INFO - codeparrot_training - Step 1336: {'lr': 0.00033400000000000004, 'samples': 256704, 'steps': 1336, 'loss/train': 6.767319917678833} 01/27/2022 20:27:23 - INFO - codeparrot_training - Step 1337: {'lr': 0.00033425, 'samples': 256896, 'steps': 1337, 'loss/train': 6.24406099319458} 01/27/2022 20:27:27 - INFO - codeparrot_training - Step 1338: {'lr': 0.00033450000000000005, 'samples': 257088, 'steps': 1338, 'loss/train': 5.743507146835327} 01/27/2022 20:27:31 - INFO - codeparrot_training - Step 1339: {'lr': 0.00033475, 'samples': 257280, 'steps': 1339, 'loss/train': 5.89429771900177} 01/27/2022 20:27:38 - INFO - codeparrot_training - Step 1340: {'lr': 0.000335, 'samples': 257472, 'steps': 1340, 'loss/train': 4.837573170661926} 01/27/2022 20:27:42 - INFO - codeparrot_training - Step 1341: {'lr': 0.00033525, 'samples': 257664, 'steps': 1341, 'loss/train': 4.5707173347473145} 01/27/2022 20:27:46 - INFO - codeparrot_training - Step 1342: {'lr': 0.0003355, 'samples': 257856, 'steps': 1342, 'loss/train': 4.389951109886169} 01/27/2022 20:27:50 - INFO - codeparrot_training - Step 1343: {'lr': 0.00033575, 'samples': 258048, 'steps': 1343, 'loss/train': 5.474466562271118} 01/27/2022 20:27:54 - INFO - codeparrot_training - Step 1344: {'lr': 0.00033600000000000004, 'samples': 258240, 'steps': 1344, 'loss/train': 5.8221423625946045} 01/27/2022 20:27:59 - INFO - codeparrot_training - Step 1345: {'lr': 0.00033625, 'samples': 258432, 'steps': 1345, 'loss/train': 6.327499866485596} 01/27/2022 20:28:03 - INFO - codeparrot_training - Step 1346: {'lr': 0.00033650000000000005, 'samples': 258624, 'steps': 1346, 'loss/train': 5.9627180099487305} 01/27/2022 20:28:08 - INFO - codeparrot_training - Step 1347: {'lr': 0.00033675, 'samples': 258816, 'steps': 1347, 'loss/train': 6.49405574798584} 01/27/2022 20:28:12 - INFO - codeparrot_training - Step 1348: {'lr': 0.000337, 'samples': 259008, 'steps': 1348, 'loss/train': 5.677776575088501} 01/27/2022 20:28:16 - INFO - codeparrot_training - Step 1349: {'lr': 0.00033725, 'samples': 259200, 'steps': 1349, 'loss/train': 5.403262138366699} 01/27/2022 20:28:21 - INFO - codeparrot_training - Step 1350: {'lr': 0.0003375, 'samples': 259392, 'steps': 1350, 'loss/train': 5.227678656578064} 01/27/2022 20:28:25 - INFO - codeparrot_training - Step 1351: {'lr': 0.00033775, 'samples': 259584, 'steps': 1351, 'loss/train': 5.7738107442855835} 01/27/2022 20:28:29 - INFO - codeparrot_training - Step 1352: {'lr': 0.00033800000000000003, 'samples': 259776, 'steps': 1352, 'loss/train': 5.483787775039673} 01/27/2022 20:28:34 - INFO - codeparrot_training - Step 1353: {'lr': 0.00033825, 'samples': 259968, 'steps': 1353, 'loss/train': 6.146427154541016} 01/27/2022 20:28:38 - INFO - codeparrot_training - Step 1354: {'lr': 0.00033850000000000004, 'samples': 260160, 'steps': 1354, 'loss/train': 5.212450861930847} 01/27/2022 20:28:44 - INFO - codeparrot_training - Step 1355: {'lr': 0.00033875, 'samples': 260352, 'steps': 1355, 'loss/train': 6.173201322555542} 01/27/2022 20:28:48 - INFO - codeparrot_training - Step 1356: {'lr': 0.00033900000000000005, 'samples': 260544, 'steps': 1356, 'loss/train': 5.408001065254211} 01/27/2022 20:28:52 - INFO - codeparrot_training - Step 1357: {'lr': 0.00033925, 'samples': 260736, 'steps': 1357, 'loss/train': 5.412554740905762} 01/27/2022 20:28:56 - INFO - codeparrot_training - Step 1358: {'lr': 0.0003395, 'samples': 260928, 'steps': 1358, 'loss/train': 6.324632406234741} 01/27/2022 20:29:00 - INFO - codeparrot_training - Step 1359: {'lr': 0.00033975, 'samples': 261120, 'steps': 1359, 'loss/train': 5.624316215515137} 01/27/2022 20:29:05 - INFO - codeparrot_training - Step 1360: {'lr': 0.00034, 'samples': 261312, 'steps': 1360, 'loss/train': 6.95796275138855} 01/27/2022 20:29:09 - INFO - codeparrot_training - Step 1361: {'lr': 0.00034025, 'samples': 261504, 'steps': 1361, 'loss/train': 5.593010187149048} 01/27/2022 20:29:14 - INFO - codeparrot_training - Step 1362: {'lr': 0.00034050000000000004, 'samples': 261696, 'steps': 1362, 'loss/train': 5.221229910850525} 01/27/2022 20:29:18 - INFO - codeparrot_training - Step 1363: {'lr': 0.00034075, 'samples': 261888, 'steps': 1363, 'loss/train': 4.74666166305542} 01/27/2022 20:29:22 - INFO - codeparrot_training - Step 1364: {'lr': 0.00034100000000000005, 'samples': 262080, 'steps': 1364, 'loss/train': 5.112065434455872} 01/27/2022 20:29:29 - INFO - codeparrot_training - Step 1365: {'lr': 0.00034125000000000003, 'samples': 262272, 'steps': 1365, 'loss/train': 7.578925609588623} 01/27/2022 20:29:33 - INFO - codeparrot_training - Step 1366: {'lr': 0.0003415, 'samples': 262464, 'steps': 1366, 'loss/train': 4.9115309715271} 01/27/2022 20:29:37 - INFO - codeparrot_training - Step 1367: {'lr': 0.00034175, 'samples': 262656, 'steps': 1367, 'loss/train': 4.823127508163452} 01/27/2022 20:29:41 - INFO - codeparrot_training - Step 1368: {'lr': 0.000342, 'samples': 262848, 'steps': 1368, 'loss/train': 6.049104452133179} 01/27/2022 20:29:46 - INFO - codeparrot_training - Step 1369: {'lr': 0.00034225, 'samples': 263040, 'steps': 1369, 'loss/train': 6.076598882675171} 01/27/2022 20:29:51 - INFO - codeparrot_training - Step 1370: {'lr': 0.00034250000000000003, 'samples': 263232, 'steps': 1370, 'loss/train': 5.928333163261414} 01/27/2022 20:29:55 - INFO - codeparrot_training - Step 1371: {'lr': 0.00034275, 'samples': 263424, 'steps': 1371, 'loss/train': 5.8759496212005615} 01/27/2022 20:29:59 - INFO - codeparrot_training - Step 1372: {'lr': 0.00034300000000000004, 'samples': 263616, 'steps': 1372, 'loss/train': 5.896329402923584} 01/27/2022 20:30:03 - INFO - codeparrot_training - Step 1373: {'lr': 0.00034325, 'samples': 263808, 'steps': 1373, 'loss/train': 5.90148138999939} 01/27/2022 20:30:07 - INFO - codeparrot_training - Step 1374: {'lr': 0.00034350000000000006, 'samples': 264000, 'steps': 1374, 'loss/train': 4.09662938117981} 01/27/2022 20:30:12 - INFO - codeparrot_training - Step 1375: {'lr': 0.00034375, 'samples': 264192, 'steps': 1375, 'loss/train': 4.673898696899414} 01/27/2022 20:30:16 - INFO - codeparrot_training - Step 1376: {'lr': 0.00034399999999999996, 'samples': 264384, 'steps': 1376, 'loss/train': 5.130805134773254} 01/27/2022 20:30:21 - INFO - codeparrot_training - Step 1377: {'lr': 0.00034425, 'samples': 264576, 'steps': 1377, 'loss/train': 4.903575539588928} 01/27/2022 20:30:25 - INFO - codeparrot_training - Step 1378: {'lr': 0.00034449999999999997, 'samples': 264768, 'steps': 1378, 'loss/train': 5.7353843450546265} 01/27/2022 20:30:29 - INFO - codeparrot_training - Step 1379: {'lr': 0.00034475, 'samples': 264960, 'steps': 1379, 'loss/train': 5.375657558441162} 01/27/2022 20:30:35 - INFO - codeparrot_training - Step 1380: {'lr': 0.000345, 'samples': 265152, 'steps': 1380, 'loss/train': 5.550543308258057} 01/27/2022 20:30:39 - INFO - codeparrot_training - Step 1381: {'lr': 0.00034525, 'samples': 265344, 'steps': 1381, 'loss/train': 6.442796945571899} 01/27/2022 20:30:43 - INFO - codeparrot_training - Step 1382: {'lr': 0.0003455, 'samples': 265536, 'steps': 1382, 'loss/train': 5.595800757408142} 01/27/2022 20:30:47 - INFO - codeparrot_training - Step 1383: {'lr': 0.00034575000000000003, 'samples': 265728, 'steps': 1383, 'loss/train': 5.464699029922485} 01/27/2022 20:30:51 - INFO - codeparrot_training - Step 1384: {'lr': 0.000346, 'samples': 265920, 'steps': 1384, 'loss/train': 5.888701558113098} 01/27/2022 20:30:57 - INFO - codeparrot_training - Step 1385: {'lr': 0.00034625, 'samples': 266112, 'steps': 1385, 'loss/train': 5.0014214515686035} 01/27/2022 20:31:01 - INFO - codeparrot_training - Step 1386: {'lr': 0.00034649999999999997, 'samples': 266304, 'steps': 1386, 'loss/train': 5.541521787643433} 01/27/2022 20:31:05 - INFO - codeparrot_training - Step 1387: {'lr': 0.00034675, 'samples': 266496, 'steps': 1387, 'loss/train': 5.95023787021637} 01/27/2022 20:31:09 - INFO - codeparrot_training - Step 1388: {'lr': 0.000347, 'samples': 266688, 'steps': 1388, 'loss/train': 6.971344470977783} 01/27/2022 20:31:13 - INFO - codeparrot_training - Step 1389: {'lr': 0.00034725, 'samples': 266880, 'steps': 1389, 'loss/train': 5.894403576850891} 01/27/2022 20:31:19 - INFO - codeparrot_training - Step 1390: {'lr': 0.0003475, 'samples': 267072, 'steps': 1390, 'loss/train': 5.6230244636535645} 01/27/2022 20:31:23 - INFO - codeparrot_training - Step 1391: {'lr': 0.00034775, 'samples': 267264, 'steps': 1391, 'loss/train': 6.673340320587158} 01/27/2022 20:31:28 - INFO - codeparrot_training - Step 1392: {'lr': 0.000348, 'samples': 267456, 'steps': 1392, 'loss/train': 6.4198994636535645} 01/27/2022 20:31:32 - INFO - codeparrot_training - Step 1393: {'lr': 0.00034825000000000004, 'samples': 267648, 'steps': 1393, 'loss/train': 5.833487033843994} 01/27/2022 20:31:36 - INFO - codeparrot_training - Step 1394: {'lr': 0.00034849999999999996, 'samples': 267840, 'steps': 1394, 'loss/train': 5.105870246887207} 01/27/2022 20:31:41 - INFO - codeparrot_training - Step 1395: {'lr': 0.00034875, 'samples': 268032, 'steps': 1395, 'loss/train': 4.660842418670654} 01/27/2022 20:31:45 - INFO - codeparrot_training - Step 1396: {'lr': 0.00034899999999999997, 'samples': 268224, 'steps': 1396, 'loss/train': 4.761531829833984} 01/27/2022 20:31:49 - INFO - codeparrot_training - Step 1397: {'lr': 0.00034925, 'samples': 268416, 'steps': 1397, 'loss/train': 6.4013872146606445} 01/27/2022 20:31:53 - INFO - codeparrot_training - Step 1398: {'lr': 0.0003495, 'samples': 268608, 'steps': 1398, 'loss/train': 3.6321576833724976} 01/27/2022 20:31:58 - INFO - codeparrot_training - Step 1399: {'lr': 0.00034975, 'samples': 268800, 'steps': 1399, 'loss/train': 5.103322505950928} 01/27/2022 20:32:03 - INFO - codeparrot_training - Step 1400: {'lr': 0.00035, 'samples': 268992, 'steps': 1400, 'loss/train': 4.250860691070557} 01/27/2022 20:32:07 - INFO - codeparrot_training - Step 1401: {'lr': 0.00035025000000000003, 'samples': 269184, 'steps': 1401, 'loss/train': 6.224779844284058} 01/27/2022 20:32:11 - INFO - codeparrot_training - Step 1402: {'lr': 0.0003505, 'samples': 269376, 'steps': 1402, 'loss/train': 5.970402359962463} 01/27/2022 20:32:15 - INFO - codeparrot_training - Step 1403: {'lr': 0.00035075, 'samples': 269568, 'steps': 1403, 'loss/train': 5.812585473060608} 01/27/2022 20:32:19 - INFO - codeparrot_training - Step 1404: {'lr': 0.00035099999999999997, 'samples': 269760, 'steps': 1404, 'loss/train': 4.140394806861877} 01/27/2022 20:32:24 - INFO - codeparrot_training - Step 1405: {'lr': 0.00035125, 'samples': 269952, 'steps': 1405, 'loss/train': 4.795518279075623} 01/27/2022 20:32:29 - INFO - codeparrot_training - Step 1406: {'lr': 0.0003515, 'samples': 270144, 'steps': 1406, 'loss/train': 4.613077640533447} 01/27/2022 20:32:33 - INFO - codeparrot_training - Step 1407: {'lr': 0.00035175, 'samples': 270336, 'steps': 1407, 'loss/train': 5.554975390434265} 01/27/2022 20:32:37 - INFO - codeparrot_training - Step 1408: {'lr': 0.000352, 'samples': 270528, 'steps': 1408, 'loss/train': 5.505416393280029} 01/27/2022 20:32:41 - INFO - codeparrot_training - Step 1409: {'lr': 0.00035225, 'samples': 270720, 'steps': 1409, 'loss/train': 4.871671199798584} 01/27/2022 20:32:47 - INFO - codeparrot_training - Step 1410: {'lr': 0.0003525, 'samples': 270912, 'steps': 1410, 'loss/train': 5.588037371635437} 01/27/2022 20:32:51 - INFO - codeparrot_training - Step 1411: {'lr': 0.00035275000000000004, 'samples': 271104, 'steps': 1411, 'loss/train': 3.6950258016586304} 01/27/2022 20:32:56 - INFO - codeparrot_training - Step 1412: {'lr': 0.00035299999999999996, 'samples': 271296, 'steps': 1412, 'loss/train': 10.260623931884766} 01/27/2022 20:33:00 - INFO - codeparrot_training - Step 1413: {'lr': 0.00035325, 'samples': 271488, 'steps': 1413, 'loss/train': 5.850403904914856} 01/27/2022 20:33:04 - INFO - codeparrot_training - Step 1414: {'lr': 0.0003535, 'samples': 271680, 'steps': 1414, 'loss/train': 4.54086971282959} 01/27/2022 20:33:09 - INFO - codeparrot_training - Step 1415: {'lr': 0.00035375, 'samples': 271872, 'steps': 1415, 'loss/train': 4.318433046340942} 01/27/2022 20:33:13 - INFO - codeparrot_training - Step 1416: {'lr': 0.000354, 'samples': 272064, 'steps': 1416, 'loss/train': 4.870678424835205} 01/27/2022 20:33:17 - INFO - codeparrot_training - Step 1417: {'lr': 0.00035425, 'samples': 272256, 'steps': 1417, 'loss/train': 4.648704886436462} 01/27/2022 20:33:21 - INFO - codeparrot_training - Step 1418: {'lr': 0.0003545, 'samples': 272448, 'steps': 1418, 'loss/train': 6.076154708862305} 01/27/2022 20:33:26 - INFO - codeparrot_training - Step 1419: {'lr': 0.00035475000000000003, 'samples': 272640, 'steps': 1419, 'loss/train': 4.453089237213135} 01/27/2022 20:33:31 - INFO - codeparrot_training - Step 1420: {'lr': 0.000355, 'samples': 272832, 'steps': 1420, 'loss/train': 5.415349245071411} 01/27/2022 20:33:35 - INFO - codeparrot_training - Step 1421: {'lr': 0.00035525000000000004, 'samples': 273024, 'steps': 1421, 'loss/train': 5.648927450180054} 01/27/2022 20:33:39 - INFO - codeparrot_training - Step 1422: {'lr': 0.00035549999999999997, 'samples': 273216, 'steps': 1422, 'loss/train': 4.790592670440674} 01/27/2022 20:33:43 - INFO - codeparrot_training - Step 1423: {'lr': 0.00035575, 'samples': 273408, 'steps': 1423, 'loss/train': 4.946924686431885} 01/27/2022 20:33:47 - INFO - codeparrot_training - Step 1424: {'lr': 0.000356, 'samples': 273600, 'steps': 1424, 'loss/train': 6.102973222732544} 01/27/2022 20:33:53 - INFO - codeparrot_training - Step 1425: {'lr': 0.00035625, 'samples': 273792, 'steps': 1425, 'loss/train': 4.645643591880798} 01/27/2022 20:33:57 - INFO - codeparrot_training - Step 1426: {'lr': 0.0003565, 'samples': 273984, 'steps': 1426, 'loss/train': 5.433918356895447} 01/27/2022 20:34:02 - INFO - codeparrot_training - Step 1427: {'lr': 0.00035675, 'samples': 274176, 'steps': 1427, 'loss/train': 5.46021580696106} 01/27/2022 20:34:06 - INFO - codeparrot_training - Step 1428: {'lr': 0.000357, 'samples': 274368, 'steps': 1428, 'loss/train': 5.155351281166077} 01/27/2022 20:34:10 - INFO - codeparrot_training - Step 1429: {'lr': 0.00035725000000000004, 'samples': 274560, 'steps': 1429, 'loss/train': 5.724423050880432} 01/27/2022 20:34:15 - INFO - codeparrot_training - Step 1430: {'lr': 0.0003575, 'samples': 274752, 'steps': 1430, 'loss/train': 5.392196774482727} 01/27/2022 20:34:19 - INFO - codeparrot_training - Step 1431: {'lr': 0.00035775, 'samples': 274944, 'steps': 1431, 'loss/train': 2.5217714309692383} 01/27/2022 20:34:23 - INFO - codeparrot_training - Step 1432: {'lr': 0.000358, 'samples': 275136, 'steps': 1432, 'loss/train': 9.94968581199646} 01/27/2022 20:34:28 - INFO - codeparrot_training - Step 1433: {'lr': 0.00035825, 'samples': 275328, 'steps': 1433, 'loss/train': 12.064802169799805} 01/27/2022 20:34:32 - INFO - codeparrot_training - Step 1434: {'lr': 0.0003585, 'samples': 275520, 'steps': 1434, 'loss/train': 6.286809682846069} 01/27/2022 20:34:39 - INFO - codeparrot_training - Step 1435: {'lr': 0.00035875, 'samples': 275712, 'steps': 1435, 'loss/train': 6.113040447235107} 01/27/2022 20:34:43 - INFO - codeparrot_training - Step 1436: {'lr': 0.000359, 'samples': 275904, 'steps': 1436, 'loss/train': 5.559696793556213} 01/27/2022 20:34:47 - INFO - codeparrot_training - Step 1437: {'lr': 0.00035925000000000003, 'samples': 276096, 'steps': 1437, 'loss/train': 5.99433159828186} 01/27/2022 20:34:51 - INFO - codeparrot_training - Step 1438: {'lr': 0.0003595, 'samples': 276288, 'steps': 1438, 'loss/train': 5.255759239196777} 01/27/2022 20:34:55 - INFO - codeparrot_training - Step 1439: {'lr': 0.00035975000000000004, 'samples': 276480, 'steps': 1439, 'loss/train': 4.47011411190033} 01/27/2022 20:35:00 - INFO - codeparrot_training - Step 1440: {'lr': 0.00035999999999999997, 'samples': 276672, 'steps': 1440, 'loss/train': 5.940374851226807} 01/27/2022 20:35:05 - INFO - codeparrot_training - Step 1441: {'lr': 0.00036025, 'samples': 276864, 'steps': 1441, 'loss/train': 5.317480802536011} 01/27/2022 20:35:09 - INFO - codeparrot_training - Step 1442: {'lr': 0.0003605, 'samples': 277056, 'steps': 1442, 'loss/train': 5.7190375328063965} 01/27/2022 20:35:13 - INFO - codeparrot_training - Step 1443: {'lr': 0.00036075, 'samples': 277248, 'steps': 1443, 'loss/train': 4.8697017431259155} 01/27/2022 20:35:17 - INFO - codeparrot_training - Step 1444: {'lr': 0.000361, 'samples': 277440, 'steps': 1444, 'loss/train': 4.927966475486755} 01/27/2022 20:35:22 - INFO - codeparrot_training - Step 1445: {'lr': 0.00036125, 'samples': 277632, 'steps': 1445, 'loss/train': 5.465997934341431} 01/27/2022 20:35:26 - INFO - codeparrot_training - Step 1446: {'lr': 0.0003615, 'samples': 277824, 'steps': 1446, 'loss/train': 5.08867084980011} 01/27/2022 20:35:31 - INFO - codeparrot_training - Step 1447: {'lr': 0.00036175000000000004, 'samples': 278016, 'steps': 1447, 'loss/train': 4.6464879512786865} 01/27/2022 20:35:35 - INFO - codeparrot_training - Step 1448: {'lr': 0.000362, 'samples': 278208, 'steps': 1448, 'loss/train': 5.358344078063965} 01/27/2022 20:35:39 - INFO - codeparrot_training - Step 1449: {'lr': 0.00036225000000000005, 'samples': 278400, 'steps': 1449, 'loss/train': 5.268254399299622} 01/27/2022 20:35:45 - INFO - codeparrot_training - Step 1450: {'lr': 0.0003625, 'samples': 278592, 'steps': 1450, 'loss/train': 9.227090120315552} 01/27/2022 20:35:49 - INFO - codeparrot_training - Step 1451: {'lr': 0.00036275, 'samples': 278784, 'steps': 1451, 'loss/train': 5.206702709197998} 01/27/2022 20:35:53 - INFO - codeparrot_training - Step 1452: {'lr': 0.000363, 'samples': 278976, 'steps': 1452, 'loss/train': 6.03830623626709} 01/27/2022 20:35:57 - INFO - codeparrot_training - Step 1453: {'lr': 0.00036325, 'samples': 279168, 'steps': 1453, 'loss/train': 5.090924978256226} 01/27/2022 20:36:01 - INFO - codeparrot_training - Step 1454: {'lr': 0.0003635, 'samples': 279360, 'steps': 1454, 'loss/train': 5.084671139717102} 01/27/2022 20:36:06 - INFO - codeparrot_training - Step 1455: {'lr': 0.00036375000000000003, 'samples': 279552, 'steps': 1455, 'loss/train': 5.014504551887512} 01/27/2022 20:36:11 - INFO - codeparrot_training - Step 1456: {'lr': 0.000364, 'samples': 279744, 'steps': 1456, 'loss/train': 5.267293453216553} 01/27/2022 20:36:15 - INFO - codeparrot_training - Step 1457: {'lr': 0.00036425000000000004, 'samples': 279936, 'steps': 1457, 'loss/train': 4.5051562786102295} 01/27/2022 20:36:19 - INFO - codeparrot_training - Step 1458: {'lr': 0.0003645, 'samples': 280128, 'steps': 1458, 'loss/train': 5.632647156715393} 01/27/2022 20:36:23 - INFO - codeparrot_training - Step 1459: {'lr': 0.00036475, 'samples': 280320, 'steps': 1459, 'loss/train': 5.735796689987183} 01/27/2022 20:36:28 - INFO - codeparrot_training - Step 1460: {'lr': 0.000365, 'samples': 280512, 'steps': 1460, 'loss/train': 4.906913638114929} 01/27/2022 20:36:32 - INFO - codeparrot_training - Step 1461: {'lr': 0.00036525, 'samples': 280704, 'steps': 1461, 'loss/train': 6.09788703918457} 01/27/2022 20:36:37 - INFO - codeparrot_training - Step 1462: {'lr': 0.0003655, 'samples': 280896, 'steps': 1462, 'loss/train': 4.120104432106018} 01/27/2022 20:36:41 - INFO - codeparrot_training - Step 1463: {'lr': 0.00036575, 'samples': 281088, 'steps': 1463, 'loss/train': 5.0831029415130615} 01/27/2022 20:36:45 - INFO - codeparrot_training - Step 1464: {'lr': 0.000366, 'samples': 281280, 'steps': 1464, 'loss/train': 4.8359938859939575} 01/27/2022 20:36:50 - INFO - codeparrot_training - Step 1465: {'lr': 0.00036625000000000004, 'samples': 281472, 'steps': 1465, 'loss/train': 4.709892511367798} 01/27/2022 20:36:54 - INFO - codeparrot_training - Step 1466: {'lr': 0.0003665, 'samples': 281664, 'steps': 1466, 'loss/train': 4.493479371070862} 01/27/2022 20:36:58 - INFO - codeparrot_training - Step 1467: {'lr': 0.00036675000000000005, 'samples': 281856, 'steps': 1467, 'loss/train': 5.326740860939026} 01/27/2022 20:37:03 - INFO - codeparrot_training - Step 1468: {'lr': 0.000367, 'samples': 282048, 'steps': 1468, 'loss/train': 2.735465705394745} 01/27/2022 20:37:07 - INFO - codeparrot_training - Step 1469: {'lr': 0.00036725, 'samples': 282240, 'steps': 1469, 'loss/train': 5.649134159088135} 01/27/2022 20:37:13 - INFO - codeparrot_training - Step 1470: {'lr': 0.0003675, 'samples': 282432, 'steps': 1470, 'loss/train': 5.295011758804321} 01/27/2022 20:37:17 - INFO - codeparrot_training - Step 1471: {'lr': 0.00036775, 'samples': 282624, 'steps': 1471, 'loss/train': 6.056208372116089} 01/27/2022 20:37:21 - INFO - codeparrot_training - Step 1472: {'lr': 0.000368, 'samples': 282816, 'steps': 1472, 'loss/train': 5.210906982421875} 01/27/2022 20:37:25 - INFO - codeparrot_training - Step 1473: {'lr': 0.00036825000000000003, 'samples': 283008, 'steps': 1473, 'loss/train': 5.32011866569519} 01/27/2022 20:37:29 - INFO - codeparrot_training - Step 1474: {'lr': 0.0003685, 'samples': 283200, 'steps': 1474, 'loss/train': 5.910141706466675} 01/27/2022 20:37:34 - INFO - codeparrot_training - Step 1475: {'lr': 0.00036875000000000005, 'samples': 283392, 'steps': 1475, 'loss/train': 5.243043422698975} 01/27/2022 20:37:39 - INFO - codeparrot_training - Step 1476: {'lr': 0.000369, 'samples': 283584, 'steps': 1476, 'loss/train': 3.3713961839675903} 01/27/2022 20:37:43 - INFO - codeparrot_training - Step 1477: {'lr': 0.00036925, 'samples': 283776, 'steps': 1477, 'loss/train': 5.813286781311035} 01/27/2022 20:37:47 - INFO - codeparrot_training - Step 1478: {'lr': 0.0003695, 'samples': 283968, 'steps': 1478, 'loss/train': 3.6485652923583984} 01/27/2022 20:37:51 - INFO - codeparrot_training - Step 1479: {'lr': 0.00036975, 'samples': 284160, 'steps': 1479, 'loss/train': 4.983058333396912} 01/27/2022 20:37:57 - INFO - codeparrot_training - Step 1480: {'lr': 0.00037, 'samples': 284352, 'steps': 1480, 'loss/train': 5.243464350700378} 01/27/2022 20:38:01 - INFO - codeparrot_training - Step 1481: {'lr': 0.00037025000000000003, 'samples': 284544, 'steps': 1481, 'loss/train': 5.352209687232971} 01/27/2022 20:38:05 - INFO - codeparrot_training - Step 1482: {'lr': 0.0003705, 'samples': 284736, 'steps': 1482, 'loss/train': 5.5908472537994385} 01/27/2022 20:38:09 - INFO - codeparrot_training - Step 1483: {'lr': 0.00037075000000000004, 'samples': 284928, 'steps': 1483, 'loss/train': 5.441688895225525} 01/27/2022 20:38:14 - INFO - codeparrot_training - Step 1484: {'lr': 0.000371, 'samples': 285120, 'steps': 1484, 'loss/train': 5.308485746383667} 01/27/2022 20:38:19 - INFO - codeparrot_training - Step 1485: {'lr': 0.00037125000000000005, 'samples': 285312, 'steps': 1485, 'loss/train': 5.070591330528259} 01/27/2022 20:38:23 - INFO - codeparrot_training - Step 1486: {'lr': 0.00037150000000000003, 'samples': 285504, 'steps': 1486, 'loss/train': 5.304766058921814} 01/27/2022 20:38:27 - INFO - codeparrot_training - Step 1487: {'lr': 0.00037175, 'samples': 285696, 'steps': 1487, 'loss/train': 6.033056259155273} 01/27/2022 20:38:31 - INFO - codeparrot_training - Step 1488: {'lr': 0.000372, 'samples': 285888, 'steps': 1488, 'loss/train': 4.6105087995529175} 01/27/2022 20:38:35 - INFO - codeparrot_training - Step 1489: {'lr': 0.00037225, 'samples': 286080, 'steps': 1489, 'loss/train': 8.464864253997803} 01/27/2022 20:38:40 - INFO - codeparrot_training - Step 1490: {'lr': 0.0003725, 'samples': 286272, 'steps': 1490, 'loss/train': 4.637064456939697} 01/27/2022 20:38:45 - INFO - codeparrot_training - Step 1491: {'lr': 0.00037275000000000003, 'samples': 286464, 'steps': 1491, 'loss/train': 5.377440333366394} 01/27/2022 20:38:49 - INFO - codeparrot_training - Step 1492: {'lr': 0.000373, 'samples': 286656, 'steps': 1492, 'loss/train': 5.113956570625305} 01/27/2022 20:38:53 - INFO - codeparrot_training - Step 1493: {'lr': 0.00037325000000000005, 'samples': 286848, 'steps': 1493, 'loss/train': 5.29119086265564} 01/27/2022 20:38:57 - INFO - codeparrot_training - Step 1494: {'lr': 0.0003735, 'samples': 287040, 'steps': 1494, 'loss/train': 4.833268046379089} 01/27/2022 20:39:03 - INFO - codeparrot_training - Step 1495: {'lr': 0.00037375000000000006, 'samples': 287232, 'steps': 1495, 'loss/train': 5.805394649505615} 01/27/2022 20:39:07 - INFO - codeparrot_training - Step 1496: {'lr': 0.000374, 'samples': 287424, 'steps': 1496, 'loss/train': 5.163054585456848} 01/27/2022 20:39:11 - INFO - codeparrot_training - Step 1497: {'lr': 0.00037425, 'samples': 287616, 'steps': 1497, 'loss/train': 6.053295135498047} 01/27/2022 20:39:15 - INFO - codeparrot_training - Step 1498: {'lr': 0.0003745, 'samples': 287808, 'steps': 1498, 'loss/train': 5.041678547859192} 01/27/2022 20:39:20 - INFO - codeparrot_training - Step 1499: {'lr': 0.00037475000000000003, 'samples': 288000, 'steps': 1499, 'loss/train': 5.274630188941956} 01/27/2022 20:39:25 - INFO - codeparrot_training - Step 1500: {'lr': 0.000375, 'samples': 288192, 'steps': 1500, 'loss/train': 5.205589056015015} 01/27/2022 20:39:29 - INFO - codeparrot_training - Step 1501: {'lr': 0.00037525, 'samples': 288384, 'steps': 1501, 'loss/train': 5.572053551673889} 01/27/2022 20:39:33 - INFO - codeparrot_training - Step 1502: {'lr': 0.0003755, 'samples': 288576, 'steps': 1502, 'loss/train': 5.647122502326965} 01/27/2022 20:39:37 - INFO - codeparrot_training - Step 1503: {'lr': 0.00037575, 'samples': 288768, 'steps': 1503, 'loss/train': 2.705719470977783} 01/27/2022 20:39:41 - INFO - codeparrot_training - Step 1504: {'lr': 0.00037600000000000003, 'samples': 288960, 'steps': 1504, 'loss/train': 4.762568950653076} 01/27/2022 20:39:46 - INFO - codeparrot_training - Step 1505: {'lr': 0.00037624999999999996, 'samples': 289152, 'steps': 1505, 'loss/train': 6.0654966831207275} 01/27/2022 20:39:51 - INFO - codeparrot_training - Step 1506: {'lr': 0.0003765, 'samples': 289344, 'steps': 1506, 'loss/train': 6.051204442977905} 01/27/2022 20:39:55 - INFO - codeparrot_training - Step 1507: {'lr': 0.00037674999999999997, 'samples': 289536, 'steps': 1507, 'loss/train': 5.4785696268081665} 01/27/2022 20:39:59 - INFO - codeparrot_training - Step 1508: {'lr': 0.000377, 'samples': 289728, 'steps': 1508, 'loss/train': 6.292721271514893} 01/27/2022 20:40:03 - INFO - codeparrot_training - Step 1509: {'lr': 0.00037725, 'samples': 289920, 'steps': 1509, 'loss/train': 6.2328407764434814} 01/27/2022 20:40:09 - INFO - codeparrot_training - Step 1510: {'lr': 0.0003775, 'samples': 290112, 'steps': 1510, 'loss/train': 5.49200963973999} 01/27/2022 20:40:13 - INFO - codeparrot_training - Step 1511: {'lr': 0.00037775, 'samples': 290304, 'steps': 1511, 'loss/train': 4.554342269897461} 01/27/2022 20:40:17 - INFO - codeparrot_training - Step 1512: {'lr': 0.000378, 'samples': 290496, 'steps': 1512, 'loss/train': 5.183329939842224} 01/27/2022 20:40:21 - INFO - codeparrot_training - Step 1513: {'lr': 0.00037825, 'samples': 290688, 'steps': 1513, 'loss/train': 6.6806960105896} 01/27/2022 20:40:25 - INFO - codeparrot_training - Step 1514: {'lr': 0.0003785, 'samples': 290880, 'steps': 1514, 'loss/train': 5.068376183509827} 01/27/2022 20:40:31 - INFO - codeparrot_training - Step 1515: {'lr': 0.00037874999999999996, 'samples': 291072, 'steps': 1515, 'loss/train': 6.151756525039673} 01/27/2022 20:40:35 - INFO - codeparrot_training - Step 1516: {'lr': 0.000379, 'samples': 291264, 'steps': 1516, 'loss/train': 4.8808043003082275} 01/27/2022 20:40:39 - INFO - codeparrot_training - Step 1517: {'lr': 0.00037925, 'samples': 291456, 'steps': 1517, 'loss/train': 6.073041200637817} 01/27/2022 20:40:43 - INFO - codeparrot_training - Step 1518: {'lr': 0.0003795, 'samples': 291648, 'steps': 1518, 'loss/train': 7.414225101470947} 01/27/2022 20:40:47 - INFO - codeparrot_training - Step 1519: {'lr': 0.00037975, 'samples': 291840, 'steps': 1519, 'loss/train': 4.969941258430481} 01/27/2022 20:40:52 - INFO - codeparrot_training - Step 1520: {'lr': 0.00038, 'samples': 292032, 'steps': 1520, 'loss/train': 6.691567897796631} 01/27/2022 20:40:56 - INFO - codeparrot_training - Step 1521: {'lr': 0.00038025, 'samples': 292224, 'steps': 1521, 'loss/train': 5.593958258628845} 01/27/2022 20:41:01 - INFO - codeparrot_training - Step 1522: {'lr': 0.00038050000000000003, 'samples': 292416, 'steps': 1522, 'loss/train': 5.019861459732056} 01/27/2022 20:41:05 - INFO - codeparrot_training - Step 1523: {'lr': 0.00038075, 'samples': 292608, 'steps': 1523, 'loss/train': 5.610503911972046} 01/27/2022 20:41:09 - INFO - codeparrot_training - Step 1524: {'lr': 0.000381, 'samples': 292800, 'steps': 1524, 'loss/train': 5.230224251747131} 01/27/2022 20:41:14 - INFO - codeparrot_training - Step 1525: {'lr': 0.00038124999999999997, 'samples': 292992, 'steps': 1525, 'loss/train': 5.2162288427352905} 01/27/2022 20:41:18 - INFO - codeparrot_training - Step 1526: {'lr': 0.0003815, 'samples': 293184, 'steps': 1526, 'loss/train': 5.494348168373108} 01/27/2022 20:41:22 - INFO - codeparrot_training - Step 1527: {'lr': 0.00038175, 'samples': 293376, 'steps': 1527, 'loss/train': 5.9144861698150635} 01/27/2022 20:41:26 - INFO - codeparrot_training - Step 1528: {'lr': 0.000382, 'samples': 293568, 'steps': 1528, 'loss/train': 5.5439969301223755} 01/27/2022 20:41:30 - INFO - codeparrot_training - Step 1529: {'lr': 0.00038225, 'samples': 293760, 'steps': 1529, 'loss/train': 3.6012353897094727} 01/27/2022 20:41:37 - INFO - codeparrot_training - Step 1530: {'lr': 0.00038250000000000003, 'samples': 293952, 'steps': 1530, 'loss/train': 4.143694639205933} 01/27/2022 20:41:41 - INFO - codeparrot_training - Step 1531: {'lr': 0.00038275, 'samples': 294144, 'steps': 1531, 'loss/train': 4.783222675323486} 01/27/2022 20:41:46 - INFO - codeparrot_training - Step 1532: {'lr': 0.00038300000000000004, 'samples': 294336, 'steps': 1532, 'loss/train': 5.426028370857239} 01/27/2022 20:41:50 - INFO - codeparrot_training - Step 1533: {'lr': 0.00038324999999999996, 'samples': 294528, 'steps': 1533, 'loss/train': 4.787345051765442} 01/27/2022 20:41:54 - INFO - codeparrot_training - Step 1534: {'lr': 0.0003835, 'samples': 294720, 'steps': 1534, 'loss/train': 4.922302007675171} 01/27/2022 20:41:59 - INFO - codeparrot_training - Step 1535: {'lr': 0.00038375, 'samples': 294912, 'steps': 1535, 'loss/train': 3.7891005277633667} 01/27/2022 20:42:03 - INFO - codeparrot_training - Step 1536: {'lr': 0.000384, 'samples': 295104, 'steps': 1536, 'loss/train': 6.030781030654907} 01/27/2022 20:42:08 - INFO - codeparrot_training - Step 1537: {'lr': 0.00038425, 'samples': 295296, 'steps': 1537, 'loss/train': 4.749072074890137} 01/27/2022 20:42:12 - INFO - codeparrot_training - Step 1538: {'lr': 0.0003845, 'samples': 295488, 'steps': 1538, 'loss/train': 4.922683954238892} 01/27/2022 20:42:16 - INFO - codeparrot_training - Step 1539: {'lr': 0.00038475, 'samples': 295680, 'steps': 1539, 'loss/train': 3.9118831157684326} 01/27/2022 20:42:22 - INFO - codeparrot_training - Step 1540: {'lr': 0.00038500000000000003, 'samples': 295872, 'steps': 1540, 'loss/train': 7.51905369758606} 01/27/2022 20:42:26 - INFO - codeparrot_training - Step 1541: {'lr': 0.00038525, 'samples': 296064, 'steps': 1541, 'loss/train': 8.751177549362183} 01/27/2022 20:42:30 - INFO - codeparrot_training - Step 1542: {'lr': 0.0003855, 'samples': 296256, 'steps': 1542, 'loss/train': 4.647533655166626} 01/27/2022 20:42:34 - INFO - codeparrot_training - Step 1543: {'lr': 0.00038574999999999997, 'samples': 296448, 'steps': 1543, 'loss/train': 5.61114764213562} 01/27/2022 20:42:38 - INFO - codeparrot_training - Step 1544: {'lr': 0.000386, 'samples': 296640, 'steps': 1544, 'loss/train': 5.708655953407288} 01/27/2022 20:42:43 - INFO - codeparrot_training - Step 1545: {'lr': 0.00038625, 'samples': 296832, 'steps': 1545, 'loss/train': 5.464983701705933} 01/27/2022 20:42:48 - INFO - codeparrot_training - Step 1546: {'lr': 0.0003865, 'samples': 297024, 'steps': 1546, 'loss/train': 5.688818335533142} 01/27/2022 20:42:52 - INFO - codeparrot_training - Step 1547: {'lr': 0.00038675, 'samples': 297216, 'steps': 1547, 'loss/train': 5.926201701164246} 01/27/2022 20:42:56 - INFO - codeparrot_training - Step 1548: {'lr': 0.00038700000000000003, 'samples': 297408, 'steps': 1548, 'loss/train': 4.735667824745178} 01/27/2022 20:43:00 - INFO - codeparrot_training - Step 1549: {'lr': 0.00038725, 'samples': 297600, 'steps': 1549, 'loss/train': 4.420346260070801} 01/27/2022 20:43:05 - INFO - codeparrot_training - Step 1550: {'lr': 0.00038750000000000004, 'samples': 297792, 'steps': 1550, 'loss/train': 4.591160774230957} 01/27/2022 20:43:09 - INFO - codeparrot_training - Step 1551: {'lr': 0.00038774999999999997, 'samples': 297984, 'steps': 1551, 'loss/train': 4.2985639572143555} 01/27/2022 20:43:13 - INFO - codeparrot_training - Step 1552: {'lr': 0.000388, 'samples': 298176, 'steps': 1552, 'loss/train': 4.6727306842803955} 01/27/2022 20:43:18 - INFO - codeparrot_training - Step 1553: {'lr': 0.00038825, 'samples': 298368, 'steps': 1553, 'loss/train': 3.6735552549362183} 01/27/2022 20:43:22 - INFO - codeparrot_training - Step 1554: {'lr': 0.0003885, 'samples': 298560, 'steps': 1554, 'loss/train': 5.152446985244751} 01/27/2022 20:43:28 - INFO - codeparrot_training - Step 1555: {'lr': 0.00038875, 'samples': 298752, 'steps': 1555, 'loss/train': 5.147357225418091} 01/27/2022 20:43:32 - INFO - codeparrot_training - Step 1556: {'lr': 0.000389, 'samples': 298944, 'steps': 1556, 'loss/train': 4.523068070411682} 01/27/2022 20:43:36 - INFO - codeparrot_training - Step 1557: {'lr': 0.00038925, 'samples': 299136, 'steps': 1557, 'loss/train': 4.796229600906372} 01/27/2022 20:43:41 - INFO - codeparrot_training - Step 1558: {'lr': 0.00038950000000000003, 'samples': 299328, 'steps': 1558, 'loss/train': 4.981274843215942} 01/27/2022 20:43:46 - INFO - codeparrot_training - Step 1559: {'lr': 0.00038975, 'samples': 299520, 'steps': 1559, 'loss/train': 5.344164133071899} 01/27/2022 20:43:50 - INFO - codeparrot_training - Step 1560: {'lr': 0.00039000000000000005, 'samples': 299712, 'steps': 1560, 'loss/train': 3.658048152923584} 01/27/2022 20:43:54 - INFO - codeparrot_training - Step 1561: {'lr': 0.00039024999999999997, 'samples': 299904, 'steps': 1561, 'loss/train': 6.048384189605713} 01/27/2022 20:43:58 - INFO - codeparrot_training - Step 1562: {'lr': 0.0003905, 'samples': 300096, 'steps': 1562, 'loss/train': 4.9259244203567505} 01/27/2022 20:44:02 - INFO - codeparrot_training - Step 1563: {'lr': 0.00039075, 'samples': 300288, 'steps': 1563, 'loss/train': 4.429136037826538} 01/27/2022 20:44:08 - INFO - codeparrot_training - Step 1564: {'lr': 0.000391, 'samples': 300480, 'steps': 1564, 'loss/train': 4.8463979959487915} 01/27/2022 20:44:12 - INFO - codeparrot_training - Step 1565: {'lr': 0.00039125, 'samples': 300672, 'steps': 1565, 'loss/train': 4.949241399765015} 01/27/2022 20:44:16 - INFO - codeparrot_training - Step 1566: {'lr': 0.00039150000000000003, 'samples': 300864, 'steps': 1566, 'loss/train': 4.108010530471802} 01/27/2022 20:44:20 - INFO - codeparrot_training - Step 1567: {'lr': 0.00039175, 'samples': 301056, 'steps': 1567, 'loss/train': 4.8989574909210205} 01/27/2022 20:44:25 - INFO - codeparrot_training - Step 1568: {'lr': 0.00039200000000000004, 'samples': 301248, 'steps': 1568, 'loss/train': 4.339829206466675} 01/27/2022 20:44:30 - INFO - codeparrot_training - Step 1569: {'lr': 0.00039225, 'samples': 301440, 'steps': 1569, 'loss/train': 4.469141006469727} 01/27/2022 20:44:34 - INFO - codeparrot_training - Step 1570: {'lr': 0.0003925, 'samples': 301632, 'steps': 1570, 'loss/train': 3.9126977920532227} 01/27/2022 20:44:38 - INFO - codeparrot_training - Step 1571: {'lr': 0.00039275, 'samples': 301824, 'steps': 1571, 'loss/train': 5.024573922157288} 01/27/2022 20:44:42 - INFO - codeparrot_training - Step 1572: {'lr': 0.000393, 'samples': 302016, 'steps': 1572, 'loss/train': 4.451012134552002} 01/27/2022 20:44:47 - INFO - codeparrot_training - Step 1573: {'lr': 0.00039325, 'samples': 302208, 'steps': 1573, 'loss/train': 4.521331787109375} 01/27/2022 20:44:52 - INFO - codeparrot_training - Step 1574: {'lr': 0.0003935, 'samples': 302400, 'steps': 1574, 'loss/train': 4.5153197050094604} 01/27/2022 20:44:56 - INFO - codeparrot_training - Step 1575: {'lr': 0.00039375, 'samples': 302592, 'steps': 1575, 'loss/train': 4.996673941612244} 01/27/2022 20:45:00 - INFO - codeparrot_training - Step 1576: {'lr': 0.00039400000000000004, 'samples': 302784, 'steps': 1576, 'loss/train': 5.757978558540344} 01/27/2022 20:45:04 - INFO - codeparrot_training - Step 1577: {'lr': 0.00039425, 'samples': 302976, 'steps': 1577, 'loss/train': 5.083407282829285} 01/27/2022 20:45:08 - INFO - codeparrot_training - Step 1578: {'lr': 0.00039450000000000005, 'samples': 303168, 'steps': 1578, 'loss/train': 5.262720823287964} 01/27/2022 20:45:13 - INFO - codeparrot_training - Step 1579: {'lr': 0.00039474999999999997, 'samples': 303360, 'steps': 1579, 'loss/train': 5.140163898468018} 01/27/2022 20:45:18 - INFO - codeparrot_training - Step 1580: {'lr': 0.000395, 'samples': 303552, 'steps': 1580, 'loss/train': 5.56126070022583} 01/27/2022 20:45:22 - INFO - codeparrot_training - Step 1581: {'lr': 0.00039525, 'samples': 303744, 'steps': 1581, 'loss/train': 4.782194137573242} 01/27/2022 20:45:26 - INFO - codeparrot_training - Step 1582: {'lr': 0.0003955, 'samples': 303936, 'steps': 1582, 'loss/train': 4.439503312110901} 01/27/2022 20:45:30 - INFO - codeparrot_training - Step 1583: {'lr': 0.00039575, 'samples': 304128, 'steps': 1583, 'loss/train': 3.5207608938217163} 01/27/2022 20:45:36 - INFO - codeparrot_training - Step 1584: {'lr': 0.00039600000000000003, 'samples': 304320, 'steps': 1584, 'loss/train': 5.967398643493652} 01/27/2022 20:45:40 - INFO - codeparrot_training - Step 1585: {'lr': 0.00039625, 'samples': 304512, 'steps': 1585, 'loss/train': 5.810554504394531} 01/27/2022 20:45:45 - INFO - codeparrot_training - Step 1586: {'lr': 0.00039650000000000004, 'samples': 304704, 'steps': 1586, 'loss/train': 2.8902506232261658} 01/27/2022 20:45:49 - INFO - codeparrot_training - Step 1587: {'lr': 0.00039675, 'samples': 304896, 'steps': 1587, 'loss/train': 4.792583584785461} 01/27/2022 20:45:53 - INFO - codeparrot_training - Step 1588: {'lr': 0.00039700000000000005, 'samples': 305088, 'steps': 1588, 'loss/train': 4.8030846118927} 01/27/2022 20:45:58 - INFO - codeparrot_training - Step 1589: {'lr': 0.00039725, 'samples': 305280, 'steps': 1589, 'loss/train': 4.019704699516296} 01/27/2022 20:46:02 - INFO - codeparrot_training - Step 1590: {'lr': 0.0003975, 'samples': 305472, 'steps': 1590, 'loss/train': 4.39790153503418} 01/27/2022 20:46:06 - INFO - codeparrot_training - Step 1591: {'lr': 0.00039775, 'samples': 305664, 'steps': 1591, 'loss/train': 4.640955805778503} 01/27/2022 20:46:10 - INFO - codeparrot_training - Step 1592: {'lr': 0.000398, 'samples': 305856, 'steps': 1592, 'loss/train': 4.73585844039917} 01/27/2022 20:46:15 - INFO - codeparrot_training - Step 1593: {'lr': 0.00039825, 'samples': 306048, 'steps': 1593, 'loss/train': 4.972269058227539} 01/27/2022 20:46:20 - INFO - codeparrot_training - Step 1594: {'lr': 0.00039850000000000004, 'samples': 306240, 'steps': 1594, 'loss/train': 4.492441534996033} 01/27/2022 20:46:24 - INFO - codeparrot_training - Step 1595: {'lr': 0.00039875, 'samples': 306432, 'steps': 1595, 'loss/train': 3.3818349838256836} 01/27/2022 20:46:28 - INFO - codeparrot_training - Step 1596: {'lr': 0.00039900000000000005, 'samples': 306624, 'steps': 1596, 'loss/train': 5.988375306129456} 01/27/2022 20:46:32 - INFO - codeparrot_training - Step 1597: {'lr': 0.00039925000000000003, 'samples': 306816, 'steps': 1597, 'loss/train': 4.95640754699707} 01/27/2022 20:46:36 - INFO - codeparrot_training - Step 1598: {'lr': 0.0003995, 'samples': 307008, 'steps': 1598, 'loss/train': 4.50589907169342} 01/27/2022 20:46:42 - INFO - codeparrot_training - Step 1599: {'lr': 0.00039975, 'samples': 307200, 'steps': 1599, 'loss/train': 1.7140374183654785} 01/27/2022 20:46:46 - INFO - codeparrot_training - Step 1600: {'lr': 0.0004, 'samples': 307392, 'steps': 1600, 'loss/train': 4.591402530670166} 01/27/2022 20:46:51 - INFO - codeparrot_training - Step 1601: {'lr': 0.00040025, 'samples': 307584, 'steps': 1601, 'loss/train': 4.896166205406189} 01/27/2022 20:46:55 - INFO - codeparrot_training - Step 1602: {'lr': 0.00040050000000000003, 'samples': 307776, 'steps': 1602, 'loss/train': 3.837166428565979} 01/27/2022 20:46:59 - INFO - codeparrot_training - Step 1603: {'lr': 0.00040075, 'samples': 307968, 'steps': 1603, 'loss/train': 5.559482574462891} 01/27/2022 20:47:04 - INFO - codeparrot_training - Step 1604: {'lr': 0.00040100000000000004, 'samples': 308160, 'steps': 1604, 'loss/train': 3.984802007675171} 01/27/2022 20:47:08 - INFO - codeparrot_training - Step 1605: {'lr': 0.00040125, 'samples': 308352, 'steps': 1605, 'loss/train': 3.6002179384231567} 01/27/2022 20:47:12 - INFO - codeparrot_training - Step 1606: {'lr': 0.00040150000000000006, 'samples': 308544, 'steps': 1606, 'loss/train': 5.544490456581116} 01/27/2022 20:47:17 - INFO - codeparrot_training - Step 1607: {'lr': 0.00040175, 'samples': 308736, 'steps': 1607, 'loss/train': 5.94091272354126} 01/27/2022 20:47:21 - INFO - codeparrot_training - Step 1608: {'lr': 0.000402, 'samples': 308928, 'steps': 1608, 'loss/train': 6.148103713989258} 01/27/2022 20:47:27 - INFO - codeparrot_training - Step 1609: {'lr': 0.00040225, 'samples': 309120, 'steps': 1609, 'loss/train': 5.291676163673401} 01/27/2022 20:47:31 - INFO - codeparrot_training - Step 1610: {'lr': 0.0004025, 'samples': 309312, 'steps': 1610, 'loss/train': 4.492552042007446} 01/27/2022 20:47:35 - INFO - codeparrot_training - Step 1611: {'lr': 0.00040275, 'samples': 309504, 'steps': 1611, 'loss/train': 4.539552569389343} 01/27/2022 20:47:39 - INFO - codeparrot_training - Step 1612: {'lr': 0.00040300000000000004, 'samples': 309696, 'steps': 1612, 'loss/train': 5.3907119035720825} 01/27/2022 20:47:43 - INFO - codeparrot_training - Step 1613: {'lr': 0.00040325, 'samples': 309888, 'steps': 1613, 'loss/train': 4.658654093742371} 01/27/2022 20:47:48 - INFO - codeparrot_training - Step 1614: {'lr': 0.00040350000000000005, 'samples': 310080, 'steps': 1614, 'loss/train': 3.8750767707824707} 01/27/2022 20:47:53 - INFO - codeparrot_training - Step 1615: {'lr': 0.00040375000000000003, 'samples': 310272, 'steps': 1615, 'loss/train': 4.661151051521301} 01/27/2022 20:47:57 - INFO - codeparrot_training - Step 1616: {'lr': 0.000404, 'samples': 310464, 'steps': 1616, 'loss/train': 4.621346712112427} 01/27/2022 20:48:01 - INFO - codeparrot_training - Step 1617: {'lr': 0.00040425, 'samples': 310656, 'steps': 1617, 'loss/train': 6.432635307312012} 01/27/2022 20:48:05 - INFO - codeparrot_training - Step 1618: {'lr': 0.0004045, 'samples': 310848, 'steps': 1618, 'loss/train': 5.453046798706055} 01/27/2022 20:48:10 - INFO - codeparrot_training - Step 1619: {'lr': 0.00040475, 'samples': 311040, 'steps': 1619, 'loss/train': 4.535217046737671} 01/27/2022 20:48:14 - INFO - codeparrot_training - Step 1620: {'lr': 0.00040500000000000003, 'samples': 311232, 'steps': 1620, 'loss/train': 5.5862696170806885} 01/27/2022 20:48:19 - INFO - codeparrot_training - Step 1621: {'lr': 0.00040525, 'samples': 311424, 'steps': 1621, 'loss/train': 4.422206282615662} 01/27/2022 20:48:23 - INFO - codeparrot_training - Step 1622: {'lr': 0.00040550000000000004, 'samples': 311616, 'steps': 1622, 'loss/train': 4.436450958251953} 01/27/2022 20:48:27 - INFO - codeparrot_training - Step 1623: {'lr': 0.00040575, 'samples': 311808, 'steps': 1623, 'loss/train': 4.70411217212677} 01/27/2022 20:48:33 - INFO - codeparrot_training - Step 1624: {'lr': 0.00040600000000000006, 'samples': 312000, 'steps': 1624, 'loss/train': 6.878641605377197} 01/27/2022 20:48:37 - INFO - codeparrot_training - Step 1625: {'lr': 0.00040625000000000004, 'samples': 312192, 'steps': 1625, 'loss/train': 5.109841346740723} 01/27/2022 20:48:41 - INFO - codeparrot_training - Step 1626: {'lr': 0.00040649999999999996, 'samples': 312384, 'steps': 1626, 'loss/train': 3.6674916744232178} 01/27/2022 20:48:45 - INFO - codeparrot_training - Step 1627: {'lr': 0.00040675, 'samples': 312576, 'steps': 1627, 'loss/train': 5.621813178062439} 01/27/2022 20:48:49 - INFO - codeparrot_training - Step 1628: {'lr': 0.00040699999999999997, 'samples': 312768, 'steps': 1628, 'loss/train': 4.177264451980591} 01/27/2022 20:48:55 - INFO - codeparrot_training - Step 1629: {'lr': 0.00040725, 'samples': 312960, 'steps': 1629, 'loss/train': 4.30924665927887} 01/27/2022 20:48:59 - INFO - codeparrot_training - Step 1630: {'lr': 0.0004075, 'samples': 313152, 'steps': 1630, 'loss/train': 4.878913521766663} 01/27/2022 20:49:03 - INFO - codeparrot_training - Step 1631: {'lr': 0.00040775, 'samples': 313344, 'steps': 1631, 'loss/train': 4.684783101081848} 01/27/2022 20:49:07 - INFO - codeparrot_training - Step 1632: {'lr': 0.000408, 'samples': 313536, 'steps': 1632, 'loss/train': 4.662874817848206} 01/27/2022 20:49:11 - INFO - codeparrot_training - Step 1633: {'lr': 0.00040825000000000003, 'samples': 313728, 'steps': 1633, 'loss/train': 4.981182932853699} 01/27/2022 20:49:17 - INFO - codeparrot_training - Step 1634: {'lr': 0.0004085, 'samples': 313920, 'steps': 1634, 'loss/train': 4.151762008666992} 01/27/2022 20:49:21 - INFO - codeparrot_training - Step 1635: {'lr': 0.00040875, 'samples': 314112, 'steps': 1635, 'loss/train': 5.552863240242004} 01/27/2022 20:49:25 - INFO - codeparrot_training - Step 1636: {'lr': 0.00040899999999999997, 'samples': 314304, 'steps': 1636, 'loss/train': 5.81500518321991} 01/27/2022 20:49:29 - INFO - codeparrot_training - Step 1637: {'lr': 0.00040925, 'samples': 314496, 'steps': 1637, 'loss/train': 5.945162773132324} 01/27/2022 20:49:33 - INFO - codeparrot_training - Step 1638: {'lr': 0.0004095, 'samples': 314688, 'steps': 1638, 'loss/train': 5.293833374977112} 01/27/2022 20:49:39 - INFO - codeparrot_training - Step 1639: {'lr': 0.00040975, 'samples': 314880, 'steps': 1639, 'loss/train': 5.059976935386658} 01/27/2022 20:49:43 - INFO - codeparrot_training - Step 1640: {'lr': 0.00041, 'samples': 315072, 'steps': 1640, 'loss/train': 5.369177341461182} 01/27/2022 20:49:47 - INFO - codeparrot_training - Step 1641: {'lr': 0.00041025, 'samples': 315264, 'steps': 1641, 'loss/train': 5.4606417417526245} 01/27/2022 20:49:51 - INFO - codeparrot_training - Step 1642: {'lr': 0.0004105, 'samples': 315456, 'steps': 1642, 'loss/train': 4.924026846885681} 01/27/2022 20:49:58 - INFO - codeparrot_training - Step 1643: {'lr': 0.00041075000000000004, 'samples': 315648, 'steps': 1643, 'loss/train': 4.861361145973206} 01/27/2022 20:50:02 - INFO - codeparrot_training - Step 1644: {'lr': 0.00041099999999999996, 'samples': 315840, 'steps': 1644, 'loss/train': 4.4336124658584595} 01/27/2022 20:50:06 - INFO - codeparrot_training - Step 1645: {'lr': 0.00041125, 'samples': 316032, 'steps': 1645, 'loss/train': 6.497914552688599} 01/27/2022 20:50:10 - INFO - codeparrot_training - Step 1646: {'lr': 0.0004115, 'samples': 316224, 'steps': 1646, 'loss/train': 4.900483846664429} 01/27/2022 20:50:14 - INFO - codeparrot_training - Step 1647: {'lr': 0.00041175, 'samples': 316416, 'steps': 1647, 'loss/train': 5.458470582962036} 01/27/2022 20:50:18 - INFO - codeparrot_training - Step 1648: {'lr': 0.000412, 'samples': 316608, 'steps': 1648, 'loss/train': 8.15811538696289} 01/27/2022 20:50:24 - INFO - codeparrot_training - Step 1649: {'lr': 0.00041225, 'samples': 316800, 'steps': 1649, 'loss/train': 5.343349456787109} 01/27/2022 20:50:28 - INFO - codeparrot_training - Step 1650: {'lr': 0.0004125, 'samples': 316992, 'steps': 1650, 'loss/train': 4.865218877792358} 01/27/2022 20:50:32 - INFO - codeparrot_training - Step 1651: {'lr': 0.00041275000000000003, 'samples': 317184, 'steps': 1651, 'loss/train': 4.42691445350647} 01/27/2022 20:50:36 - INFO - codeparrot_training - Step 1652: {'lr': 0.000413, 'samples': 317376, 'steps': 1652, 'loss/train': 5.561220645904541} 01/27/2022 20:50:40 - INFO - codeparrot_training - Step 1653: {'lr': 0.00041325, 'samples': 317568, 'steps': 1653, 'loss/train': 3.0752928256988525} 01/27/2022 20:50:46 - INFO - codeparrot_training - Step 1654: {'lr': 0.00041349999999999997, 'samples': 317760, 'steps': 1654, 'loss/train': 4.584151983261108} 01/27/2022 20:50:50 - INFO - codeparrot_training - Step 1655: {'lr': 0.00041375, 'samples': 317952, 'steps': 1655, 'loss/train': 2.5737547874450684} 01/27/2022 20:50:55 - INFO - codeparrot_training - Step 1656: {'lr': 0.000414, 'samples': 318144, 'steps': 1656, 'loss/train': 4.079554796218872} 01/27/2022 20:50:59 - INFO - codeparrot_training - Step 1657: {'lr': 0.00041425, 'samples': 318336, 'steps': 1657, 'loss/train': 3.551210403442383} 01/27/2022 20:51:03 - INFO - codeparrot_training - Step 1658: {'lr': 0.0004145, 'samples': 318528, 'steps': 1658, 'loss/train': 5.817299008369446} 01/27/2022 20:51:09 - INFO - codeparrot_training - Step 1659: {'lr': 0.00041475, 'samples': 318720, 'steps': 1659, 'loss/train': 5.47824239730835} 01/27/2022 20:51:13 - INFO - codeparrot_training - Step 1660: {'lr': 0.000415, 'samples': 318912, 'steps': 1660, 'loss/train': 4.1034064292907715} 01/27/2022 20:51:18 - INFO - codeparrot_training - Step 1661: {'lr': 0.00041525000000000004, 'samples': 319104, 'steps': 1661, 'loss/train': 4.750672459602356} 01/27/2022 20:51:22 - INFO - codeparrot_training - Step 1662: {'lr': 0.00041549999999999996, 'samples': 319296, 'steps': 1662, 'loss/train': 4.089618802070618} 01/27/2022 20:51:26 - INFO - codeparrot_training - Step 1663: {'lr': 0.00041575, 'samples': 319488, 'steps': 1663, 'loss/train': 4.134867310523987} 01/27/2022 20:51:30 - INFO - codeparrot_training - Step 1664: {'lr': 0.000416, 'samples': 319680, 'steps': 1664, 'loss/train': 4.357253193855286} 01/27/2022 20:51:35 - INFO - codeparrot_training - Step 1665: {'lr': 0.00041625, 'samples': 319872, 'steps': 1665, 'loss/train': 5.185676693916321} 01/27/2022 20:51:39 - INFO - codeparrot_training - Step 1666: {'lr': 0.0004165, 'samples': 320064, 'steps': 1666, 'loss/train': 4.968478202819824} 01/27/2022 20:51:44 - INFO - codeparrot_training - Step 1667: {'lr': 0.00041675, 'samples': 320256, 'steps': 1667, 'loss/train': 10.678154468536377} 01/27/2022 20:51:48 - INFO - codeparrot_training - Step 1668: {'lr': 0.000417, 'samples': 320448, 'steps': 1668, 'loss/train': 6.057011604309082} 01/27/2022 20:51:52 - INFO - codeparrot_training - Step 1669: {'lr': 0.00041725000000000003, 'samples': 320640, 'steps': 1669, 'loss/train': 4.69346022605896} 01/27/2022 20:51:58 - INFO - codeparrot_training - Step 1670: {'lr': 0.0004175, 'samples': 320832, 'steps': 1670, 'loss/train': 5.976093292236328} 01/27/2022 20:52:02 - INFO - codeparrot_training - Step 1671: {'lr': 0.00041775000000000004, 'samples': 321024, 'steps': 1671, 'loss/train': 5.555848717689514} 01/27/2022 20:52:07 - INFO - codeparrot_training - Step 1672: {'lr': 0.00041799999999999997, 'samples': 321216, 'steps': 1672, 'loss/train': 5.629608750343323} 01/27/2022 20:52:11 - INFO - codeparrot_training - Step 1673: {'lr': 0.00041825, 'samples': 321408, 'steps': 1673, 'loss/train': 4.571760535240173} 01/27/2022 20:52:15 - INFO - codeparrot_training - Step 1674: {'lr': 0.0004185, 'samples': 321600, 'steps': 1674, 'loss/train': 3.648239493370056} 01/27/2022 20:52:20 - INFO - codeparrot_training - Step 1675: {'lr': 0.00041875, 'samples': 321792, 'steps': 1675, 'loss/train': 6.776950836181641} 01/27/2022 20:52:24 - INFO - codeparrot_training - Step 1676: {'lr': 0.000419, 'samples': 321984, 'steps': 1676, 'loss/train': 5.541209936141968} 01/27/2022 20:52:29 - INFO - codeparrot_training - Step 1677: {'lr': 0.00041925, 'samples': 322176, 'steps': 1677, 'loss/train': 4.438369274139404} 01/27/2022 20:52:33 - INFO - codeparrot_training - Step 1678: {'lr': 0.0004195, 'samples': 322368, 'steps': 1678, 'loss/train': 5.130235075950623} 01/27/2022 20:52:37 - INFO - codeparrot_training - Step 1679: {'lr': 0.00041975000000000004, 'samples': 322560, 'steps': 1679, 'loss/train': 4.956532716751099} 01/27/2022 20:52:42 - INFO - codeparrot_training - Step 1680: {'lr': 0.00042, 'samples': 322752, 'steps': 1680, 'loss/train': 7.478966474533081} 01/27/2022 20:52:46 - INFO - codeparrot_training - Step 1681: {'lr': 0.00042025, 'samples': 322944, 'steps': 1681, 'loss/train': 5.931625127792358} 01/27/2022 20:52:51 - INFO - codeparrot_training - Step 1682: {'lr': 0.0004205, 'samples': 323136, 'steps': 1682, 'loss/train': 4.653693079948425} 01/27/2022 20:52:55 - INFO - codeparrot_training - Step 1683: {'lr': 0.00042075, 'samples': 323328, 'steps': 1683, 'loss/train': 5.044461250305176} 01/27/2022 20:52:59 - INFO - codeparrot_training - Step 1684: {'lr': 0.000421, 'samples': 323520, 'steps': 1684, 'loss/train': 3.442354917526245} 01/27/2022 20:53:06 - INFO - codeparrot_training - Step 1685: {'lr': 0.00042125, 'samples': 323712, 'steps': 1685, 'loss/train': 4.211433291435242} 01/27/2022 20:53:10 - INFO - codeparrot_training - Step 1686: {'lr': 0.0004215, 'samples': 323904, 'steps': 1686, 'loss/train': 5.015124678611755} 01/27/2022 20:53:14 - INFO - codeparrot_training - Step 1687: {'lr': 0.00042175000000000003, 'samples': 324096, 'steps': 1687, 'loss/train': 5.645220637321472} 01/27/2022 20:53:18 - INFO - codeparrot_training - Step 1688: {'lr': 0.000422, 'samples': 324288, 'steps': 1688, 'loss/train': 5.594154596328735} 01/27/2022 20:53:23 - INFO - codeparrot_training - Step 1689: {'lr': 0.00042225000000000005, 'samples': 324480, 'steps': 1689, 'loss/train': 3.2915682792663574} 01/27/2022 20:53:28 - INFO - codeparrot_training - Step 1690: {'lr': 0.00042249999999999997, 'samples': 324672, 'steps': 1690, 'loss/train': 3.065605401992798} 01/27/2022 20:53:32 - INFO - codeparrot_training - Step 1691: {'lr': 0.00042275, 'samples': 324864, 'steps': 1691, 'loss/train': 4.275627493858337} 01/27/2022 20:53:36 - INFO - codeparrot_training - Step 1692: {'lr': 0.000423, 'samples': 325056, 'steps': 1692, 'loss/train': 5.2252349853515625} 01/27/2022 20:53:40 - INFO - codeparrot_training - Step 1693: {'lr': 0.00042325, 'samples': 325248, 'steps': 1693, 'loss/train': 4.2724692821502686} 01/27/2022 20:53:44 - INFO - codeparrot_training - Step 1694: {'lr': 0.0004235, 'samples': 325440, 'steps': 1694, 'loss/train': 5.012627720832825} 01/27/2022 20:53:49 - INFO - codeparrot_training - Step 1695: {'lr': 0.00042375000000000003, 'samples': 325632, 'steps': 1695, 'loss/train': 4.095845460891724} 01/27/2022 20:53:54 - INFO - codeparrot_training - Step 1696: {'lr': 0.000424, 'samples': 325824, 'steps': 1696, 'loss/train': 4.753535270690918} 01/27/2022 20:53:58 - INFO - codeparrot_training - Step 1697: {'lr': 0.00042425000000000004, 'samples': 326016, 'steps': 1697, 'loss/train': 4.817803144454956} 01/27/2022 20:54:02 - INFO - codeparrot_training - Step 1698: {'lr': 0.0004245, 'samples': 326208, 'steps': 1698, 'loss/train': 4.125654816627502} 01/27/2022 20:54:06 - INFO - codeparrot_training - Step 1699: {'lr': 0.00042475000000000005, 'samples': 326400, 'steps': 1699, 'loss/train': 4.899822950363159} 01/27/2022 20:54:12 - INFO - codeparrot_training - Step 1700: {'lr': 0.000425, 'samples': 326592, 'steps': 1700, 'loss/train': 4.947185754776001} 01/27/2022 20:54:16 - INFO - codeparrot_training - Step 1701: {'lr': 0.00042525, 'samples': 326784, 'steps': 1701, 'loss/train': 6.078960657119751} 01/27/2022 20:54:20 - INFO - codeparrot_training - Step 1702: {'lr': 0.0004255, 'samples': 326976, 'steps': 1702, 'loss/train': 5.877833604812622} 01/27/2022 20:54:25 - INFO - codeparrot_training - Step 1703: {'lr': 0.00042575, 'samples': 327168, 'steps': 1703, 'loss/train': 4.6492838859558105} 01/27/2022 20:54:29 - INFO - codeparrot_training - Step 1704: {'lr': 0.000426, 'samples': 327360, 'steps': 1704, 'loss/train': 5.065916419029236} 01/27/2022 20:54:34 - INFO - codeparrot_training - Step 1705: {'lr': 0.00042625000000000003, 'samples': 327552, 'steps': 1705, 'loss/train': 5.7648056745529175} 01/27/2022 20:54:38 - INFO - codeparrot_training - Step 1706: {'lr': 0.0004265, 'samples': 327744, 'steps': 1706, 'loss/train': 5.916955947875977} 01/27/2022 20:54:42 - INFO - codeparrot_training - Step 1707: {'lr': 0.00042675000000000005, 'samples': 327936, 'steps': 1707, 'loss/train': 4.648202061653137} 01/27/2022 20:54:46 - INFO - codeparrot_training - Step 1708: {'lr': 0.000427, 'samples': 328128, 'steps': 1708, 'loss/train': 4.2541598081588745} 01/27/2022 20:54:51 - INFO - codeparrot_training - Step 1709: {'lr': 0.00042725, 'samples': 328320, 'steps': 1709, 'loss/train': 4.176839590072632} 01/27/2022 20:54:56 - INFO - codeparrot_training - Step 1710: {'lr': 0.0004275, 'samples': 328512, 'steps': 1710, 'loss/train': 5.986522436141968} 01/27/2022 20:55:00 - INFO - codeparrot_training - Step 1711: {'lr': 0.00042775, 'samples': 328704, 'steps': 1711, 'loss/train': 3.9285460710525513} 01/27/2022 20:55:04 - INFO - codeparrot_training - Step 1712: {'lr': 0.000428, 'samples': 328896, 'steps': 1712, 'loss/train': 5.205624103546143} 01/27/2022 20:55:08 - INFO - codeparrot_training - Step 1713: {'lr': 0.00042825000000000003, 'samples': 329088, 'steps': 1713, 'loss/train': 5.124661445617676} 01/27/2022 20:55:12 - INFO - codeparrot_training - Step 1714: {'lr': 0.0004285, 'samples': 329280, 'steps': 1714, 'loss/train': 5.29037082195282} 01/27/2022 20:55:18 - INFO - codeparrot_training - Step 1715: {'lr': 0.00042875000000000004, 'samples': 329472, 'steps': 1715, 'loss/train': 3.200813055038452} 01/27/2022 20:55:22 - INFO - codeparrot_training - Step 1716: {'lr': 0.000429, 'samples': 329664, 'steps': 1716, 'loss/train': 5.33403217792511} 01/27/2022 20:55:27 - INFO - codeparrot_training - Step 1717: {'lr': 0.00042925000000000005, 'samples': 329856, 'steps': 1717, 'loss/train': 5.40205192565918} 01/27/2022 20:55:31 - INFO - codeparrot_training - Step 1718: {'lr': 0.0004295, 'samples': 330048, 'steps': 1718, 'loss/train': 5.164235830307007} 01/27/2022 20:55:35 - INFO - codeparrot_training - Step 1719: {'lr': 0.00042975, 'samples': 330240, 'steps': 1719, 'loss/train': 5.0945563316345215} 01/27/2022 20:55:40 - INFO - codeparrot_training - Step 1720: {'lr': 0.00043, 'samples': 330432, 'steps': 1720, 'loss/train': 5.805701851844788} 01/27/2022 20:55:44 - INFO - codeparrot_training - Step 1721: {'lr': 0.00043025, 'samples': 330624, 'steps': 1721, 'loss/train': 5.174437522888184} 01/27/2022 20:55:48 - INFO - codeparrot_training - Step 1722: {'lr': 0.0004305, 'samples': 330816, 'steps': 1722, 'loss/train': 5.244656324386597} 01/27/2022 20:55:53 - INFO - codeparrot_training - Step 1723: {'lr': 0.00043075000000000003, 'samples': 331008, 'steps': 1723, 'loss/train': 5.414682984352112} 01/27/2022 20:55:57 - INFO - codeparrot_training - Step 1724: {'lr': 0.000431, 'samples': 331200, 'steps': 1724, 'loss/train': 5.539515495300293} 01/27/2022 20:56:02 - INFO - codeparrot_training - Step 1725: {'lr': 0.00043125000000000005, 'samples': 331392, 'steps': 1725, 'loss/train': 4.8776267766952515} 01/27/2022 20:56:06 - INFO - codeparrot_training - Step 1726: {'lr': 0.0004315, 'samples': 331584, 'steps': 1726, 'loss/train': 5.9577155113220215} 01/27/2022 20:56:10 - INFO - codeparrot_training - Step 1727: {'lr': 0.00043175, 'samples': 331776, 'steps': 1727, 'loss/train': 5.266177296638489} 01/27/2022 20:56:14 - INFO - codeparrot_training - Step 1728: {'lr': 0.000432, 'samples': 331968, 'steps': 1728, 'loss/train': 4.860105514526367} 01/27/2022 20:56:18 - INFO - codeparrot_training - Step 1729: {'lr': 0.00043225, 'samples': 332160, 'steps': 1729, 'loss/train': 4.335245490074158} 01/27/2022 20:56:24 - INFO - codeparrot_training - Step 1730: {'lr': 0.0004325, 'samples': 332352, 'steps': 1730, 'loss/train': 6.060381174087524} 01/27/2022 20:56:29 - INFO - codeparrot_training - Step 1731: {'lr': 0.00043275000000000003, 'samples': 332544, 'steps': 1731, 'loss/train': 4.861485242843628} 01/27/2022 20:56:33 - INFO - codeparrot_training - Step 1732: {'lr': 0.000433, 'samples': 332736, 'steps': 1732, 'loss/train': 4.13783061504364} 01/27/2022 20:56:37 - INFO - codeparrot_training - Step 1733: {'lr': 0.00043325000000000004, 'samples': 332928, 'steps': 1733, 'loss/train': 4.61021625995636} 01/27/2022 20:56:41 - INFO - codeparrot_training - Step 1734: {'lr': 0.0004335, 'samples': 333120, 'steps': 1734, 'loss/train': 6.506906747817993} 01/27/2022 20:56:46 - INFO - codeparrot_training - Step 1735: {'lr': 0.00043375000000000005, 'samples': 333312, 'steps': 1735, 'loss/train': 4.977795481681824} 01/27/2022 20:56:50 - INFO - codeparrot_training - Step 1736: {'lr': 0.00043400000000000003, 'samples': 333504, 'steps': 1736, 'loss/train': 4.854881644248962} 01/27/2022 20:56:54 - INFO - codeparrot_training - Step 1737: {'lr': 0.00043425, 'samples': 333696, 'steps': 1737, 'loss/train': 4.509007215499878} 01/27/2022 20:56:59 - INFO - codeparrot_training - Step 1738: {'lr': 0.0004345, 'samples': 333888, 'steps': 1738, 'loss/train': 3.3236074447631836} 01/27/2022 20:57:03 - INFO - codeparrot_training - Step 1739: {'lr': 0.00043475, 'samples': 334080, 'steps': 1739, 'loss/train': 3.633428692817688} 01/27/2022 20:57:09 - INFO - codeparrot_training - Step 1740: {'lr': 0.000435, 'samples': 334272, 'steps': 1740, 'loss/train': 3.8261711597442627} 01/27/2022 20:57:13 - INFO - codeparrot_training - Step 1741: {'lr': 0.00043525000000000004, 'samples': 334464, 'steps': 1741, 'loss/train': 5.210784673690796} 01/27/2022 20:57:17 - INFO - codeparrot_training - Step 1742: {'lr': 0.0004355, 'samples': 334656, 'steps': 1742, 'loss/train': 4.622898817062378} 01/27/2022 20:57:21 - INFO - codeparrot_training - Step 1743: {'lr': 0.00043575000000000005, 'samples': 334848, 'steps': 1743, 'loss/train': 4.687637686729431} 01/27/2022 20:57:25 - INFO - codeparrot_training - Step 1744: {'lr': 0.000436, 'samples': 335040, 'steps': 1744, 'loss/train': 5.478589653968811} 01/27/2022 20:57:31 - INFO - codeparrot_training - Step 1745: {'lr': 0.00043625000000000006, 'samples': 335232, 'steps': 1745, 'loss/train': 5.851840496063232} 01/27/2022 20:57:35 - INFO - codeparrot_training - Step 1746: {'lr': 0.0004365, 'samples': 335424, 'steps': 1746, 'loss/train': 4.417731285095215} 01/27/2022 20:57:40 - INFO - codeparrot_training - Step 1747: {'lr': 0.00043675, 'samples': 335616, 'steps': 1747, 'loss/train': 5.634359836578369} 01/27/2022 20:57:44 - INFO - codeparrot_training - Step 1748: {'lr': 0.000437, 'samples': 335808, 'steps': 1748, 'loss/train': 4.898866653442383} 01/27/2022 20:57:48 - INFO - codeparrot_training - Step 1749: {'lr': 0.00043725000000000003, 'samples': 336000, 'steps': 1749, 'loss/train': 4.736881613731384} 01/27/2022 20:57:52 - INFO - codeparrot_training - Step 1750: {'lr': 0.0004375, 'samples': 336192, 'steps': 1750, 'loss/train': 4.807565331459045} 01/27/2022 20:57:57 - INFO - codeparrot_training - Step 1751: {'lr': 0.00043775, 'samples': 336384, 'steps': 1751, 'loss/train': 4.518897771835327} 01/27/2022 20:58:01 - INFO - codeparrot_training - Step 1752: {'lr': 0.000438, 'samples': 336576, 'steps': 1752, 'loss/train': 3.820875406265259} 01/27/2022 20:58:06 - INFO - codeparrot_training - Step 1753: {'lr': 0.00043825, 'samples': 336768, 'steps': 1753, 'loss/train': 5.065893173217773} 01/27/2022 20:58:10 - INFO - codeparrot_training - Step 1754: {'lr': 0.00043850000000000003, 'samples': 336960, 'steps': 1754, 'loss/train': 4.350127100944519} 01/27/2022 20:58:14 - INFO - codeparrot_training - Step 1755: {'lr': 0.00043874999999999996, 'samples': 337152, 'steps': 1755, 'loss/train': 4.935661554336548} 01/27/2022 20:58:19 - INFO - codeparrot_training - Step 1756: {'lr': 0.000439, 'samples': 337344, 'steps': 1756, 'loss/train': 5.391317367553711} 01/27/2022 20:58:23 - INFO - codeparrot_training - Step 1757: {'lr': 0.00043924999999999997, 'samples': 337536, 'steps': 1757, 'loss/train': 5.007919192314148} 01/27/2022 20:58:27 - INFO - codeparrot_training - Step 1758: {'lr': 0.0004395, 'samples': 337728, 'steps': 1758, 'loss/train': 3.904427647590637} 01/27/2022 20:58:31 - INFO - codeparrot_training - Step 1759: {'lr': 0.00043975, 'samples': 337920, 'steps': 1759, 'loss/train': 4.726148843765259} 01/27/2022 20:58:36 - INFO - codeparrot_training - Step 1760: {'lr': 0.00044, 'samples': 338112, 'steps': 1760, 'loss/train': 4.59481143951416} 01/27/2022 20:58:42 - INFO - codeparrot_training - Step 1761: {'lr': 0.00044025, 'samples': 338304, 'steps': 1761, 'loss/train': 4.144498586654663} 01/27/2022 20:58:46 - INFO - codeparrot_training - Step 1762: {'lr': 0.00044050000000000003, 'samples': 338496, 'steps': 1762, 'loss/train': 4.81161904335022} 01/27/2022 20:58:50 - INFO - codeparrot_training - Step 1763: {'lr': 0.00044075, 'samples': 338688, 'steps': 1763, 'loss/train': 4.207608461380005} 01/27/2022 20:58:54 - INFO - codeparrot_training - Step 1764: {'lr': 0.000441, 'samples': 338880, 'steps': 1764, 'loss/train': 5.70531964302063} 01/27/2022 20:58:58 - INFO - codeparrot_training - Step 1765: {'lr': 0.00044124999999999996, 'samples': 339072, 'steps': 1765, 'loss/train': 5.701893925666809} 01/27/2022 20:59:03 - INFO - codeparrot_training - Step 1766: {'lr': 0.0004415, 'samples': 339264, 'steps': 1766, 'loss/train': 4.631235122680664} 01/27/2022 20:59:08 - INFO - codeparrot_training - Step 1767: {'lr': 0.00044175, 'samples': 339456, 'steps': 1767, 'loss/train': 4.755450367927551} 01/27/2022 20:59:12 - INFO - codeparrot_training - Step 1768: {'lr': 0.000442, 'samples': 339648, 'steps': 1768, 'loss/train': 5.2073482275009155} 01/27/2022 20:59:16 - INFO - codeparrot_training - Step 1769: {'lr': 0.00044225, 'samples': 339840, 'steps': 1769, 'loss/train': 5.654919147491455} 01/27/2022 20:59:20 - INFO - codeparrot_training - Step 1770: {'lr': 0.0004425, 'samples': 340032, 'steps': 1770, 'loss/train': 5.131520748138428} 01/27/2022 20:59:25 - INFO - codeparrot_training - Step 1771: {'lr': 0.00044275, 'samples': 340224, 'steps': 1771, 'loss/train': 6.042468309402466} 01/27/2022 20:59:30 - INFO - codeparrot_training - Step 1772: {'lr': 0.00044300000000000003, 'samples': 340416, 'steps': 1772, 'loss/train': 3.4836251735687256} 01/27/2022 20:59:34 - INFO - codeparrot_training - Step 1773: {'lr': 0.00044325, 'samples': 340608, 'steps': 1773, 'loss/train': 4.968786835670471} 01/27/2022 20:59:38 - INFO - codeparrot_training - Step 1774: {'lr': 0.0004435, 'samples': 340800, 'steps': 1774, 'loss/train': 3.3600107431411743} 01/27/2022 20:59:42 - INFO - codeparrot_training - Step 1775: {'lr': 0.00044374999999999997, 'samples': 340992, 'steps': 1775, 'loss/train': 5.393645524978638} 01/27/2022 20:59:48 - INFO - codeparrot_training - Step 1776: {'lr': 0.000444, 'samples': 341184, 'steps': 1776, 'loss/train': 3.910504460334778} 01/27/2022 20:59:52 - INFO - codeparrot_training - Step 1777: {'lr': 0.00044425, 'samples': 341376, 'steps': 1777, 'loss/train': 5.942509889602661} 01/27/2022 20:59:57 - INFO - codeparrot_training - Step 1778: {'lr': 0.0004445, 'samples': 341568, 'steps': 1778, 'loss/train': 1.6830652356147766} 01/27/2022 21:00:01 - INFO - codeparrot_training - Step 1779: {'lr': 0.00044475, 'samples': 341760, 'steps': 1779, 'loss/train': 5.512582898139954} 01/27/2022 21:00:05 - INFO - codeparrot_training - Step 1780: {'lr': 0.00044500000000000003, 'samples': 341952, 'steps': 1780, 'loss/train': 4.7168920040130615} 01/27/2022 21:00:10 - INFO - codeparrot_training - Step 1781: {'lr': 0.00044525, 'samples': 342144, 'steps': 1781, 'loss/train': 5.519904613494873} 01/27/2022 21:00:14 - INFO - codeparrot_training - Step 1782: {'lr': 0.00044550000000000004, 'samples': 342336, 'steps': 1782, 'loss/train': 4.961467981338501} 01/27/2022 21:00:19 - INFO - codeparrot_training - Step 1783: {'lr': 0.00044574999999999997, 'samples': 342528, 'steps': 1783, 'loss/train': 4.779161095619202} 01/27/2022 21:00:23 - INFO - codeparrot_training - Step 1784: {'lr': 0.000446, 'samples': 342720, 'steps': 1784, 'loss/train': 3.9149551391601562} 01/27/2022 21:00:27 - INFO - codeparrot_training - Step 1785: {'lr': 0.00044625, 'samples': 342912, 'steps': 1785, 'loss/train': 5.313560128211975} 01/27/2022 21:00:33 - INFO - codeparrot_training - Step 1786: {'lr': 0.0004465, 'samples': 343104, 'steps': 1786, 'loss/train': 4.449593782424927} 01/27/2022 21:00:37 - INFO - codeparrot_training - Step 1787: {'lr': 0.00044675, 'samples': 343296, 'steps': 1787, 'loss/train': 4.626879930496216} 01/27/2022 21:00:41 - INFO - codeparrot_training - Step 1788: {'lr': 0.000447, 'samples': 343488, 'steps': 1788, 'loss/train': 4.868821978569031} 01/27/2022 21:00:45 - INFO - codeparrot_training - Step 1789: {'lr': 0.00044725, 'samples': 343680, 'steps': 1789, 'loss/train': 4.6601643562316895} 01/27/2022 21:00:49 - INFO - codeparrot_training - Step 1790: {'lr': 0.00044750000000000004, 'samples': 343872, 'steps': 1790, 'loss/train': 3.3959877490997314} 01/27/2022 21:00:54 - INFO - codeparrot_training - Step 1791: {'lr': 0.00044775, 'samples': 344064, 'steps': 1791, 'loss/train': 4.442063927650452} 01/27/2022 21:00:59 - INFO - codeparrot_training - Step 1792: {'lr': 0.000448, 'samples': 344256, 'steps': 1792, 'loss/train': 4.918506145477295} 01/27/2022 21:01:03 - INFO - codeparrot_training - Step 1793: {'lr': 0.00044824999999999997, 'samples': 344448, 'steps': 1793, 'loss/train': 5.150592684745789} 01/27/2022 21:01:07 - INFO - codeparrot_training - Step 1794: {'lr': 0.0004485, 'samples': 344640, 'steps': 1794, 'loss/train': 5.543230175971985} 01/27/2022 21:01:11 - INFO - codeparrot_training - Step 1795: {'lr': 0.00044875, 'samples': 344832, 'steps': 1795, 'loss/train': 5.858844995498657} 01/27/2022 21:01:16 - INFO - codeparrot_training - Step 1796: {'lr': 0.000449, 'samples': 345024, 'steps': 1796, 'loss/train': 3.9244412183761597} 01/27/2022 21:01:20 - INFO - codeparrot_training - Step 1797: {'lr': 0.00044925, 'samples': 345216, 'steps': 1797, 'loss/train': 5.390080690383911} 01/27/2022 21:01:25 - INFO - codeparrot_training - Step 1798: {'lr': 0.00044950000000000003, 'samples': 345408, 'steps': 1798, 'loss/train': 5.1660075187683105} 01/27/2022 21:01:29 - INFO - codeparrot_training - Step 1799: {'lr': 0.00044975, 'samples': 345600, 'steps': 1799, 'loss/train': 4.866254925727844} 01/27/2022 21:01:33 - INFO - codeparrot_training - Step 1800: {'lr': 0.00045000000000000004, 'samples': 345792, 'steps': 1800, 'loss/train': 5.002342700958252} 01/27/2022 21:01:39 - INFO - codeparrot_training - Step 1801: {'lr': 0.00045024999999999997, 'samples': 345984, 'steps': 1801, 'loss/train': 4.4176822900772095} 01/27/2022 21:01:43 - INFO - codeparrot_training - Step 1802: {'lr': 0.0004505, 'samples': 346176, 'steps': 1802, 'loss/train': 4.236628890037537} 01/27/2022 21:01:47 - INFO - codeparrot_training - Step 1803: {'lr': 0.00045075, 'samples': 346368, 'steps': 1803, 'loss/train': 2.723323702812195} 01/27/2022 21:01:51 - INFO - codeparrot_training - Step 1804: {'lr': 0.000451, 'samples': 346560, 'steps': 1804, 'loss/train': 4.8871750831604} 01/27/2022 21:01:55 - INFO - codeparrot_training - Step 1805: {'lr': 0.00045125, 'samples': 346752, 'steps': 1805, 'loss/train': 3.266984224319458} 01/27/2022 21:02:01 - INFO - codeparrot_training - Step 1806: {'lr': 0.0004515, 'samples': 346944, 'steps': 1806, 'loss/train': 4.703328251838684} 01/27/2022 21:02:05 - INFO - codeparrot_training - Step 1807: {'lr': 0.00045175, 'samples': 347136, 'steps': 1807, 'loss/train': 7.13761568069458} 01/27/2022 21:02:09 - INFO - codeparrot_training - Step 1808: {'lr': 0.00045200000000000004, 'samples': 347328, 'steps': 1808, 'loss/train': 5.49048113822937} 01/27/2022 21:02:13 - INFO - codeparrot_training - Step 1809: {'lr': 0.00045225, 'samples': 347520, 'steps': 1809, 'loss/train': 6.181141376495361} 01/27/2022 21:02:17 - INFO - codeparrot_training - Step 1810: {'lr': 0.00045250000000000005, 'samples': 347712, 'steps': 1810, 'loss/train': 4.769554853439331} 01/27/2022 21:02:23 - INFO - codeparrot_training - Step 1811: {'lr': 0.00045275, 'samples': 347904, 'steps': 1811, 'loss/train': 5.000910043716431} 01/27/2022 21:02:27 - INFO - codeparrot_training - Step 1812: {'lr': 0.000453, 'samples': 348096, 'steps': 1812, 'loss/train': 4.625191569328308} 01/27/2022 21:02:31 - INFO - codeparrot_training - Step 1813: {'lr': 0.00045325, 'samples': 348288, 'steps': 1813, 'loss/train': 7.105871915817261} 01/27/2022 21:02:35 - INFO - codeparrot_training - Step 1814: {'lr': 0.0004535, 'samples': 348480, 'steps': 1814, 'loss/train': 4.6342878341674805} 01/27/2022 21:02:39 - INFO - codeparrot_training - Step 1815: {'lr': 0.00045375, 'samples': 348672, 'steps': 1815, 'loss/train': 5.0025869607925415} 01/27/2022 21:02:44 - INFO - codeparrot_training - Step 1816: {'lr': 0.00045400000000000003, 'samples': 348864, 'steps': 1816, 'loss/train': 4.478776216506958} 01/27/2022 21:02:49 - INFO - codeparrot_training - Step 1817: {'lr': 0.00045425, 'samples': 349056, 'steps': 1817, 'loss/train': 3.7656140327453613} 01/27/2022 21:02:53 - INFO - codeparrot_training - Step 1818: {'lr': 0.00045450000000000004, 'samples': 349248, 'steps': 1818, 'loss/train': 4.221714019775391} 01/27/2022 21:02:57 - INFO - codeparrot_training - Step 1819: {'lr': 0.00045475, 'samples': 349440, 'steps': 1819, 'loss/train': 5.6670098304748535} 01/27/2022 21:03:01 - INFO - codeparrot_training - Step 1820: {'lr': 0.000455, 'samples': 349632, 'steps': 1820, 'loss/train': 6.225698947906494} 01/27/2022 21:03:07 - INFO - codeparrot_training - Step 1821: {'lr': 0.00045525, 'samples': 349824, 'steps': 1821, 'loss/train': 5.336203336715698} 01/27/2022 21:03:11 - INFO - codeparrot_training - Step 1822: {'lr': 0.0004555, 'samples': 350016, 'steps': 1822, 'loss/train': 4.444505095481873} 01/27/2022 21:03:15 - INFO - codeparrot_training - Step 1823: {'lr': 0.00045575, 'samples': 350208, 'steps': 1823, 'loss/train': 4.926588535308838} 01/27/2022 21:03:20 - INFO - codeparrot_training - Step 1824: {'lr': 0.000456, 'samples': 350400, 'steps': 1824, 'loss/train': 5.198822021484375} 01/27/2022 21:03:24 - INFO - codeparrot_training - Step 1825: {'lr': 0.00045625, 'samples': 350592, 'steps': 1825, 'loss/train': 5.063373327255249} 01/27/2022 21:03:29 - INFO - codeparrot_training - Step 1826: {'lr': 0.00045650000000000004, 'samples': 350784, 'steps': 1826, 'loss/train': 4.201614260673523} 01/27/2022 21:03:33 - INFO - codeparrot_training - Step 1827: {'lr': 0.00045675, 'samples': 350976, 'steps': 1827, 'loss/train': 4.0784443616867065} 01/27/2022 21:03:37 - INFO - codeparrot_training - Step 1828: {'lr': 0.00045700000000000005, 'samples': 351168, 'steps': 1828, 'loss/train': 3.9706603288650513} 01/27/2022 21:03:41 - INFO - codeparrot_training - Step 1829: {'lr': 0.00045725, 'samples': 351360, 'steps': 1829, 'loss/train': 4.465611577033997} 01/27/2022 21:03:46 - INFO - codeparrot_training - Step 1830: {'lr': 0.0004575, 'samples': 351552, 'steps': 1830, 'loss/train': 5.088142991065979} 01/27/2022 21:03:52 - INFO - codeparrot_training - Step 1831: {'lr': 0.00045775, 'samples': 351744, 'steps': 1831, 'loss/train': 4.200507760047913} 01/27/2022 21:03:56 - INFO - codeparrot_training - Step 1832: {'lr': 0.000458, 'samples': 351936, 'steps': 1832, 'loss/train': 3.968131184577942} 01/27/2022 21:04:00 - INFO - codeparrot_training - Step 1833: {'lr': 0.00045825, 'samples': 352128, 'steps': 1833, 'loss/train': 5.042641282081604} 01/27/2022 21:04:05 - INFO - codeparrot_training - Step 1834: {'lr': 0.00045850000000000003, 'samples': 352320, 'steps': 1834, 'loss/train': 4.397329330444336} 01/27/2022 21:04:09 - INFO - codeparrot_training - Step 1835: {'lr': 0.00045875, 'samples': 352512, 'steps': 1835, 'loss/train': 5.026055932044983} 01/27/2022 21:04:14 - INFO - codeparrot_training - Step 1836: {'lr': 0.00045900000000000004, 'samples': 352704, 'steps': 1836, 'loss/train': 3.6876357793807983} 01/27/2022 21:04:18 - INFO - codeparrot_training - Step 1837: {'lr': 0.00045925, 'samples': 352896, 'steps': 1837, 'loss/train': 6.1690993309021} 01/27/2022 21:04:22 - INFO - codeparrot_training - Step 1838: {'lr': 0.00045950000000000006, 'samples': 353088, 'steps': 1838, 'loss/train': 4.706387042999268} 01/27/2022 21:04:26 - INFO - codeparrot_training - Step 1839: {'lr': 0.00045975, 'samples': 353280, 'steps': 1839, 'loss/train': 6.05207633972168} 01/27/2022 21:04:31 - INFO - codeparrot_training - Step 1840: {'lr': 0.00046, 'samples': 353472, 'steps': 1840, 'loss/train': 3.8779070377349854} 01/27/2022 21:04:36 - INFO - codeparrot_training - Step 1841: {'lr': 0.00046025, 'samples': 353664, 'steps': 1841, 'loss/train': 4.576024532318115} 01/27/2022 21:04:40 - INFO - codeparrot_training - Step 1842: {'lr': 0.0004605, 'samples': 353856, 'steps': 1842, 'loss/train': 4.642486810684204} 01/27/2022 21:04:44 - INFO - codeparrot_training - Step 1843: {'lr': 0.00046075, 'samples': 354048, 'steps': 1843, 'loss/train': 4.1859272718429565} 01/27/2022 21:04:48 - INFO - codeparrot_training - Step 1844: {'lr': 0.00046100000000000004, 'samples': 354240, 'steps': 1844, 'loss/train': 4.481332540512085} 01/27/2022 21:04:52 - INFO - codeparrot_training - Step 1845: {'lr': 0.00046125, 'samples': 354432, 'steps': 1845, 'loss/train': 3.7361608743667603} 01/27/2022 21:04:58 - INFO - codeparrot_training - Step 1846: {'lr': 0.00046150000000000005, 'samples': 354624, 'steps': 1846, 'loss/train': 5.090752601623535} 01/27/2022 21:05:02 - INFO - codeparrot_training - Step 1847: {'lr': 0.00046175000000000003, 'samples': 354816, 'steps': 1847, 'loss/train': 5.028010010719299} 01/27/2022 21:05:06 - INFO - codeparrot_training - Step 1848: {'lr': 0.000462, 'samples': 355008, 'steps': 1848, 'loss/train': 4.967589497566223} 01/27/2022 21:05:10 - INFO - codeparrot_training - Step 1849: {'lr': 0.00046225, 'samples': 355200, 'steps': 1849, 'loss/train': 4.8270556926727295} 01/27/2022 21:05:15 - INFO - codeparrot_training - Step 1850: {'lr': 0.0004625, 'samples': 355392, 'steps': 1850, 'loss/train': 5.161792159080505} 01/27/2022 21:05:20 - INFO - codeparrot_training - Step 1851: {'lr': 0.00046275, 'samples': 355584, 'steps': 1851, 'loss/train': 5.122801423072815} 01/27/2022 21:05:24 - INFO - codeparrot_training - Step 1852: {'lr': 0.00046300000000000003, 'samples': 355776, 'steps': 1852, 'loss/train': 2.74101322889328} 01/27/2022 21:05:28 - INFO - codeparrot_training - Step 1853: {'lr': 0.00046325, 'samples': 355968, 'steps': 1853, 'loss/train': 5.113685131072998} 01/27/2022 21:05:32 - INFO - codeparrot_training - Step 1854: {'lr': 0.00046350000000000004, 'samples': 356160, 'steps': 1854, 'loss/train': 4.935330390930176} 01/27/2022 21:05:37 - INFO - codeparrot_training - Step 1855: {'lr': 0.00046375, 'samples': 356352, 'steps': 1855, 'loss/train': 4.784466862678528} 01/27/2022 21:05:42 - INFO - codeparrot_training - Step 1856: {'lr': 0.00046400000000000006, 'samples': 356544, 'steps': 1856, 'loss/train': 4.772746682167053} 01/27/2022 21:05:46 - INFO - codeparrot_training - Step 1857: {'lr': 0.00046425, 'samples': 356736, 'steps': 1857, 'loss/train': 6.764199256896973} 01/27/2022 21:05:50 - INFO - codeparrot_training - Step 1858: {'lr': 0.0004645, 'samples': 356928, 'steps': 1858, 'loss/train': 4.822646856307983} 01/27/2022 21:05:54 - INFO - codeparrot_training - Step 1859: {'lr': 0.00046475, 'samples': 357120, 'steps': 1859, 'loss/train': 3.762139320373535} 01/27/2022 21:05:58 - INFO - codeparrot_training - Step 1860: {'lr': 0.000465, 'samples': 357312, 'steps': 1860, 'loss/train': 5.29742968082428} 01/27/2022 21:06:04 - INFO - codeparrot_training - Step 1861: {'lr': 0.00046525, 'samples': 357504, 'steps': 1861, 'loss/train': 5.090372085571289} 01/27/2022 21:06:08 - INFO - codeparrot_training - Step 1862: {'lr': 0.00046550000000000004, 'samples': 357696, 'steps': 1862, 'loss/train': 4.783032774925232} 01/27/2022 21:06:12 - INFO - codeparrot_training - Step 1863: {'lr': 0.00046575, 'samples': 357888, 'steps': 1863, 'loss/train': 5.143263816833496} 01/27/2022 21:06:16 - INFO - codeparrot_training - Step 1864: {'lr': 0.00046600000000000005, 'samples': 358080, 'steps': 1864, 'loss/train': 4.6861878633499146} 01/27/2022 21:06:20 - INFO - codeparrot_training - Step 1865: {'lr': 0.00046625000000000003, 'samples': 358272, 'steps': 1865, 'loss/train': 3.656361937522888} 01/27/2022 21:06:26 - INFO - codeparrot_training - Step 1866: {'lr': 0.0004665, 'samples': 358464, 'steps': 1866, 'loss/train': 3.8588401079177856} 01/27/2022 21:06:30 - INFO - codeparrot_training - Step 1867: {'lr': 0.00046675, 'samples': 358656, 'steps': 1867, 'loss/train': 4.612795114517212} 01/27/2022 21:06:35 - INFO - codeparrot_training - Step 1868: {'lr': 0.000467, 'samples': 358848, 'steps': 1868, 'loss/train': 4.297463893890381} 01/27/2022 21:06:39 - INFO - codeparrot_training - Step 1869: {'lr': 0.00046725, 'samples': 359040, 'steps': 1869, 'loss/train': 4.389006972312927} 01/27/2022 21:06:43 - INFO - codeparrot_training - Step 1870: {'lr': 0.00046750000000000003, 'samples': 359232, 'steps': 1870, 'loss/train': 3.264990448951721} 01/27/2022 21:06:48 - INFO - codeparrot_training - Step 1871: {'lr': 0.00046775, 'samples': 359424, 'steps': 1871, 'loss/train': 5.188838839530945} 01/27/2022 21:06:52 - INFO - codeparrot_training - Step 1872: {'lr': 0.00046800000000000005, 'samples': 359616, 'steps': 1872, 'loss/train': 4.107481598854065} 01/27/2022 21:06:57 - INFO - codeparrot_training - Step 1873: {'lr': 0.00046825, 'samples': 359808, 'steps': 1873, 'loss/train': 4.440730690956116} 01/27/2022 21:07:01 - INFO - codeparrot_training - Step 1874: {'lr': 0.00046850000000000006, 'samples': 360000, 'steps': 1874, 'loss/train': 3.99187695980072} 01/27/2022 21:07:05 - INFO - codeparrot_training - Step 1875: {'lr': 0.00046875, 'samples': 360192, 'steps': 1875, 'loss/train': 4.760858416557312} 01/27/2022 21:07:11 - INFO - codeparrot_training - Step 1876: {'lr': 0.00046899999999999996, 'samples': 360384, 'steps': 1876, 'loss/train': 5.039731979370117} 01/27/2022 21:07:15 - INFO - codeparrot_training - Step 1877: {'lr': 0.00046925, 'samples': 360576, 'steps': 1877, 'loss/train': 4.0667442083358765} 01/27/2022 21:07:19 - INFO - codeparrot_training - Step 1878: {'lr': 0.0004695, 'samples': 360768, 'steps': 1878, 'loss/train': 4.332077264785767} 01/27/2022 21:07:23 - INFO - codeparrot_training - Step 1879: {'lr': 0.00046975, 'samples': 360960, 'steps': 1879, 'loss/train': 4.177701473236084} 01/27/2022 21:07:27 - INFO - codeparrot_training - Step 1880: {'lr': 0.00047, 'samples': 361152, 'steps': 1880, 'loss/train': 4.516612529754639} 01/27/2022 21:07:32 - INFO - codeparrot_training - Step 1881: {'lr': 0.00047025, 'samples': 361344, 'steps': 1881, 'loss/train': 3.8615119457244873} 01/27/2022 21:07:37 - INFO - codeparrot_training - Step 1882: {'lr': 0.0004705, 'samples': 361536, 'steps': 1882, 'loss/train': 5.216192007064819} 01/27/2022 21:07:41 - INFO - codeparrot_training - Step 1883: {'lr': 0.00047075000000000003, 'samples': 361728, 'steps': 1883, 'loss/train': 5.14946722984314} 01/27/2022 21:07:45 - INFO - codeparrot_training - Step 1884: {'lr': 0.000471, 'samples': 361920, 'steps': 1884, 'loss/train': 4.049653172492981} 01/27/2022 21:07:49 - INFO - codeparrot_training - Step 1885: {'lr': 0.00047125, 'samples': 362112, 'steps': 1885, 'loss/train': 4.4000043869018555} 01/27/2022 21:07:54 - INFO - codeparrot_training - Step 1886: {'lr': 0.00047149999999999997, 'samples': 362304, 'steps': 1886, 'loss/train': 4.617537975311279} 01/27/2022 21:07:58 - INFO - codeparrot_training - Step 1887: {'lr': 0.00047175, 'samples': 362496, 'steps': 1887, 'loss/train': 3.882035493850708} 01/27/2022 21:08:03 - INFO - codeparrot_training - Step 1888: {'lr': 0.000472, 'samples': 362688, 'steps': 1888, 'loss/train': 7.318262815475464} 01/27/2022 21:08:07 - INFO - codeparrot_training - Step 1889: {'lr': 0.00047225, 'samples': 362880, 'steps': 1889, 'loss/train': 5.3229228258132935} 01/27/2022 21:08:11 - INFO - codeparrot_training - Step 1890: {'lr': 0.0004725, 'samples': 363072, 'steps': 1890, 'loss/train': 4.344042062759399} 01/27/2022 21:08:17 - INFO - codeparrot_training - Step 1891: {'lr': 0.00047275, 'samples': 363264, 'steps': 1891, 'loss/train': 4.763678312301636} 01/27/2022 21:08:21 - INFO - codeparrot_training - Step 1892: {'lr': 0.000473, 'samples': 363456, 'steps': 1892, 'loss/train': 4.571455121040344} 01/27/2022 21:08:25 - INFO - codeparrot_training - Step 1893: {'lr': 0.00047325000000000004, 'samples': 363648, 'steps': 1893, 'loss/train': 5.195192098617554} 01/27/2022 21:08:29 - INFO - codeparrot_training - Step 1894: {'lr': 0.00047349999999999996, 'samples': 363840, 'steps': 1894, 'loss/train': 5.191724181175232} 01/27/2022 21:08:33 - INFO - codeparrot_training - Step 1895: {'lr': 0.00047375, 'samples': 364032, 'steps': 1895, 'loss/train': 4.533761501312256} 01/27/2022 21:08:38 - INFO - codeparrot_training - Step 1896: {'lr': 0.000474, 'samples': 364224, 'steps': 1896, 'loss/train': 4.539787173271179} 01/27/2022 21:08:43 - INFO - codeparrot_training - Step 1897: {'lr': 0.00047425, 'samples': 364416, 'steps': 1897, 'loss/train': 4.5197049379348755} 01/27/2022 21:08:47 - INFO - codeparrot_training - Step 1898: {'lr': 0.0004745, 'samples': 364608, 'steps': 1898, 'loss/train': 4.860765337944031} 01/27/2022 21:08:51 - INFO - codeparrot_training - Step 1899: {'lr': 0.00047475, 'samples': 364800, 'steps': 1899, 'loss/train': 4.127663969993591} 01/27/2022 21:08:55 - INFO - codeparrot_training - Step 1900: {'lr': 0.000475, 'samples': 364992, 'steps': 1900, 'loss/train': 4.609157681465149} 01/27/2022 21:09:00 - INFO - codeparrot_training - Step 1901: {'lr': 0.00047525000000000003, 'samples': 365184, 'steps': 1901, 'loss/train': 5.054586410522461} 01/27/2022 21:09:05 - INFO - codeparrot_training - Step 1902: {'lr': 0.0004755, 'samples': 365376, 'steps': 1902, 'loss/train': 4.420566558837891} 01/27/2022 21:09:09 - INFO - codeparrot_training - Step 1903: {'lr': 0.00047575, 'samples': 365568, 'steps': 1903, 'loss/train': 4.976064920425415} 01/27/2022 21:09:13 - INFO - codeparrot_training - Step 1904: {'lr': 0.00047599999999999997, 'samples': 365760, 'steps': 1904, 'loss/train': 4.838501930236816} 01/27/2022 21:09:19 - INFO - codeparrot_training - Step 1905: {'lr': 0.00047625, 'samples': 365952, 'steps': 1905, 'loss/train': 4.80051326751709} 01/27/2022 21:09:23 - INFO - codeparrot_training - Step 1906: {'lr': 0.0004765, 'samples': 366144, 'steps': 1906, 'loss/train': 4.284734487533569} 01/27/2022 21:09:27 - INFO - codeparrot_training - Step 1907: {'lr': 0.00047675, 'samples': 366336, 'steps': 1907, 'loss/train': 4.938504338264465} 01/27/2022 21:09:31 - INFO - codeparrot_training - Step 1908: {'lr': 0.000477, 'samples': 366528, 'steps': 1908, 'loss/train': 2.722438395023346} 01/27/2022 21:09:35 - INFO - codeparrot_training - Step 1909: {'lr': 0.00047725, 'samples': 366720, 'steps': 1909, 'loss/train': 3.8878129720687866} 01/27/2022 21:09:41 - INFO - codeparrot_training - Step 1910: {'lr': 0.0004775, 'samples': 366912, 'steps': 1910, 'loss/train': 4.722251772880554} 01/27/2022 21:09:45 - INFO - codeparrot_training - Step 1911: {'lr': 0.00047775000000000004, 'samples': 367104, 'steps': 1911, 'loss/train': 4.588999629020691} 01/27/2022 21:09:49 - INFO - codeparrot_training - Step 1912: {'lr': 0.00047799999999999996, 'samples': 367296, 'steps': 1912, 'loss/train': 4.008896827697754} 01/27/2022 21:09:53 - INFO - codeparrot_training - Step 1913: {'lr': 0.00047825, 'samples': 367488, 'steps': 1913, 'loss/train': 4.707217454910278} 01/27/2022 21:09:57 - INFO - codeparrot_training - Step 1914: {'lr': 0.0004785, 'samples': 367680, 'steps': 1914, 'loss/train': 3.5580400228500366} 01/27/2022 21:10:01 - INFO - codeparrot_training - Step 1915: {'lr': 0.00047875, 'samples': 367872, 'steps': 1915, 'loss/train': 4.5836663246154785} 01/27/2022 21:10:08 - INFO - codeparrot_training - Step 1916: {'lr': 0.000479, 'samples': 368064, 'steps': 1916, 'loss/train': 4.874203562736511} 01/27/2022 21:10:12 - INFO - codeparrot_training - Step 1917: {'lr': 0.00047925, 'samples': 368256, 'steps': 1917, 'loss/train': 4.514940619468689} 01/27/2022 21:10:16 - INFO - codeparrot_training - Step 1918: {'lr': 0.0004795, 'samples': 368448, 'steps': 1918, 'loss/train': 3.8254787921905518} 01/27/2022 21:10:20 - INFO - codeparrot_training - Step 1919: {'lr': 0.00047975000000000003, 'samples': 368640, 'steps': 1919, 'loss/train': 3.3733913898468018} 01/27/2022 21:10:24 - INFO - codeparrot_training - Step 1920: {'lr': 0.00048, 'samples': 368832, 'steps': 1920, 'loss/train': 4.963331937789917} 01/27/2022 21:10:28 - INFO - codeparrot_training - Step 1921: {'lr': 0.00048025000000000005, 'samples': 369024, 'steps': 1921, 'loss/train': 4.147143244743347} 01/27/2022 21:10:34 - INFO - codeparrot_training - Step 1922: {'lr': 0.00048049999999999997, 'samples': 369216, 'steps': 1922, 'loss/train': 4.811792492866516} 01/27/2022 21:10:39 - INFO - codeparrot_training - Step 1923: {'lr': 0.00048075, 'samples': 369408, 'steps': 1923, 'loss/train': 4.590798854827881} 01/27/2022 21:10:43 - INFO - codeparrot_training - Step 1924: {'lr': 0.000481, 'samples': 369600, 'steps': 1924, 'loss/train': 4.559292197227478} 01/27/2022 21:10:47 - INFO - codeparrot_training - Step 1925: {'lr': 0.00048125, 'samples': 369792, 'steps': 1925, 'loss/train': 4.20210599899292} 01/27/2022 21:10:52 - INFO - codeparrot_training - Step 1926: {'lr': 0.0004815, 'samples': 369984, 'steps': 1926, 'loss/train': 4.739063858985901} 01/27/2022 21:10:56 - INFO - codeparrot_training - Step 1927: {'lr': 0.00048175000000000003, 'samples': 370176, 'steps': 1927, 'loss/train': 4.388291716575623} 01/27/2022 21:11:00 - INFO - codeparrot_training - Step 1928: {'lr': 0.000482, 'samples': 370368, 'steps': 1928, 'loss/train': 4.732992053031921} 01/27/2022 21:11:05 - INFO - codeparrot_training - Step 1929: {'lr': 0.00048225000000000004, 'samples': 370560, 'steps': 1929, 'loss/train': 4.608087658882141} 01/27/2022 21:11:09 - INFO - codeparrot_training - Step 1930: {'lr': 0.0004825, 'samples': 370752, 'steps': 1930, 'loss/train': 5.404157638549805} 01/27/2022 21:11:14 - INFO - codeparrot_training - Step 1931: {'lr': 0.00048275, 'samples': 370944, 'steps': 1931, 'loss/train': 5.553591012954712} 01/27/2022 21:11:18 - INFO - codeparrot_training - Step 1932: {'lr': 0.000483, 'samples': 371136, 'steps': 1932, 'loss/train': 4.732909083366394} 01/27/2022 21:11:22 - INFO - codeparrot_training - Step 1933: {'lr': 0.00048325, 'samples': 371328, 'steps': 1933, 'loss/train': 5.583869576454163} 01/27/2022 21:11:26 - INFO - codeparrot_training - Step 1934: {'lr': 0.0004835, 'samples': 371520, 'steps': 1934, 'loss/train': 3.6927484273910522} 01/27/2022 21:11:31 - INFO - codeparrot_training - Step 1935: {'lr': 0.00048375, 'samples': 371712, 'steps': 1935, 'loss/train': 3.2610021829605103} 01/27/2022 21:11:36 - INFO - codeparrot_training - Step 1936: {'lr': 0.000484, 'samples': 371904, 'steps': 1936, 'loss/train': 3.924904704093933} 01/27/2022 21:11:41 - INFO - codeparrot_training - Step 1937: {'lr': 0.00048425000000000003, 'samples': 372096, 'steps': 1937, 'loss/train': 3.885619640350342} 01/27/2022 21:11:45 - INFO - codeparrot_training - Step 1938: {'lr': 0.0004845, 'samples': 372288, 'steps': 1938, 'loss/train': 4.519023656845093} 01/27/2022 21:11:49 - INFO - codeparrot_training - Step 1939: {'lr': 0.00048475000000000005, 'samples': 372480, 'steps': 1939, 'loss/train': 4.748381495475769} 01/27/2022 21:11:53 - INFO - codeparrot_training - Step 1940: {'lr': 0.00048499999999999997, 'samples': 372672, 'steps': 1940, 'loss/train': 4.6287617683410645} 01/27/2022 21:11:58 - INFO - codeparrot_training - Step 1941: {'lr': 0.00048525, 'samples': 372864, 'steps': 1941, 'loss/train': 4.685520887374878} 01/27/2022 21:12:02 - INFO - codeparrot_training - Step 1942: {'lr': 0.0004855, 'samples': 373056, 'steps': 1942, 'loss/train': 5.095425009727478} 01/27/2022 21:12:06 - INFO - codeparrot_training - Step 1943: {'lr': 0.00048575, 'samples': 373248, 'steps': 1943, 'loss/train': 5.088365077972412} 01/27/2022 21:12:11 - INFO - codeparrot_training - Step 1944: {'lr': 0.000486, 'samples': 373440, 'steps': 1944, 'loss/train': 1.493291974067688} 01/27/2022 21:12:15 - INFO - codeparrot_training - Step 1945: {'lr': 0.00048625000000000003, 'samples': 373632, 'steps': 1945, 'loss/train': 4.513129234313965} 01/27/2022 21:12:20 - INFO - codeparrot_training - Step 1946: {'lr': 0.0004865, 'samples': 373824, 'steps': 1946, 'loss/train': 5.02623438835144} 01/27/2022 21:12:24 - INFO - codeparrot_training - Step 1947: {'lr': 0.00048675000000000004, 'samples': 374016, 'steps': 1947, 'loss/train': 4.261286973953247} 01/27/2022 21:12:28 - INFO - codeparrot_training - Step 1948: {'lr': 0.000487, 'samples': 374208, 'steps': 1948, 'loss/train': 4.509857654571533} 01/27/2022 21:12:32 - INFO - codeparrot_training - Step 1949: {'lr': 0.00048725000000000005, 'samples': 374400, 'steps': 1949, 'loss/train': 3.888123035430908} 01/27/2022 21:12:37 - INFO - codeparrot_training - Step 1950: {'lr': 0.0004875, 'samples': 374592, 'steps': 1950, 'loss/train': 5.118421196937561} 01/27/2022 21:12:42 - INFO - codeparrot_training - Step 1951: {'lr': 0.00048775, 'samples': 374784, 'steps': 1951, 'loss/train': 4.300903558731079} 01/27/2022 21:12:47 - INFO - codeparrot_training - Step 1952: {'lr': 0.000488, 'samples': 374976, 'steps': 1952, 'loss/train': 4.937902808189392} 01/27/2022 21:12:51 - INFO - codeparrot_training - Step 1953: {'lr': 0.00048825, 'samples': 375168, 'steps': 1953, 'loss/train': 2.7438263297080994} 01/27/2022 21:12:55 - INFO - codeparrot_training - Step 1954: {'lr': 0.0004885, 'samples': 375360, 'steps': 1954, 'loss/train': 4.456546068191528} 01/27/2022 21:12:59 - INFO - codeparrot_training - Step 1955: {'lr': 0.00048875, 'samples': 375552, 'steps': 1955, 'loss/train': 3.7986055612564087} 01/27/2022 21:13:04 - INFO - codeparrot_training - Step 1956: {'lr': 0.000489, 'samples': 375744, 'steps': 1956, 'loss/train': 4.974090814590454} 01/27/2022 21:13:08 - INFO - codeparrot_training - Step 1957: {'lr': 0.00048925, 'samples': 375936, 'steps': 1957, 'loss/train': 3.625651717185974} 01/27/2022 21:13:12 - INFO - codeparrot_training - Step 1958: {'lr': 0.0004895, 'samples': 376128, 'steps': 1958, 'loss/train': 5.105894565582275} 01/27/2022 21:13:17 - INFO - codeparrot_training - Step 1959: {'lr': 0.0004897500000000001, 'samples': 376320, 'steps': 1959, 'loss/train': 4.603535413742065} 01/27/2022 21:13:21 - INFO - codeparrot_training - Step 1960: {'lr': 0.00049, 'samples': 376512, 'steps': 1960, 'loss/train': 4.694121837615967} 01/27/2022 21:13:27 - INFO - codeparrot_training - Step 1961: {'lr': 0.00049025, 'samples': 376704, 'steps': 1961, 'loss/train': 6.37076210975647} 01/27/2022 21:13:31 - INFO - codeparrot_training - Step 1962: {'lr': 0.0004905, 'samples': 376896, 'steps': 1962, 'loss/train': 4.068798780441284} 01/27/2022 21:13:36 - INFO - codeparrot_training - Step 1963: {'lr': 0.0004907500000000001, 'samples': 377088, 'steps': 1963, 'loss/train': 4.044296979904175} 01/27/2022 21:13:40 - INFO - codeparrot_training - Step 1964: {'lr': 0.000491, 'samples': 377280, 'steps': 1964, 'loss/train': 4.433956861495972} 01/27/2022 21:13:44 - INFO - codeparrot_training - Step 1965: {'lr': 0.00049125, 'samples': 377472, 'steps': 1965, 'loss/train': 3.865524172782898} 01/27/2022 21:13:49 - INFO - codeparrot_training - Step 1966: {'lr': 0.0004915, 'samples': 377664, 'steps': 1966, 'loss/train': 4.424926042556763} 01/27/2022 21:13:53 - INFO - codeparrot_training - Step 1967: {'lr': 0.00049175, 'samples': 377856, 'steps': 1967, 'loss/train': 4.4014259576797485} 01/27/2022 21:13:57 - INFO - codeparrot_training - Step 1968: {'lr': 0.000492, 'samples': 378048, 'steps': 1968, 'loss/train': 3.88840913772583} 01/27/2022 21:14:01 - INFO - codeparrot_training - Step 1969: {'lr': 0.0004922500000000001, 'samples': 378240, 'steps': 1969, 'loss/train': 5.5969297885894775} 01/27/2022 21:14:06 - INFO - codeparrot_training - Step 1970: {'lr': 0.0004925, 'samples': 378432, 'steps': 1970, 'loss/train': 4.795366644859314} 01/27/2022 21:14:11 - INFO - codeparrot_training - Step 1971: {'lr': 0.00049275, 'samples': 378624, 'steps': 1971, 'loss/train': 5.265831828117371} 01/27/2022 21:14:16 - INFO - codeparrot_training - Step 1972: {'lr': 0.0004930000000000001, 'samples': 378816, 'steps': 1972, 'loss/train': 3.9992154836654663} 01/27/2022 21:14:20 - INFO - codeparrot_training - Step 1973: {'lr': 0.00049325, 'samples': 379008, 'steps': 1973, 'loss/train': 4.666123151779175} 01/27/2022 21:14:24 - INFO - codeparrot_training - Step 1974: {'lr': 0.0004935, 'samples': 379200, 'steps': 1974, 'loss/train': 6.514744520187378} 01/27/2022 21:14:28 - INFO - codeparrot_training - Step 1975: {'lr': 0.00049375, 'samples': 379392, 'steps': 1975, 'loss/train': 3.383027672767639} 01/27/2022 21:14:33 - INFO - codeparrot_training - Step 1976: {'lr': 0.000494, 'samples': 379584, 'steps': 1976, 'loss/train': 4.652289032936096} 01/27/2022 21:14:37 - INFO - codeparrot_training - Step 1977: {'lr': 0.00049425, 'samples': 379776, 'steps': 1977, 'loss/train': 5.445010185241699} 01/27/2022 21:14:41 - INFO - codeparrot_training - Step 1978: {'lr': 0.0004945, 'samples': 379968, 'steps': 1978, 'loss/train': 3.688882827758789} 01/27/2022 21:14:46 - INFO - codeparrot_training - Step 1979: {'lr': 0.0004947500000000001, 'samples': 380160, 'steps': 1979, 'loss/train': 4.830942749977112} 01/27/2022 21:14:50 - INFO - codeparrot_training - Step 1980: {'lr': 0.000495, 'samples': 380352, 'steps': 1980, 'loss/train': 3.948821783065796} 01/27/2022 21:14:56 - INFO - codeparrot_training - Step 1981: {'lr': 0.00049525, 'samples': 380544, 'steps': 1981, 'loss/train': 4.744636058807373} 01/27/2022 21:15:00 - INFO - codeparrot_training - Step 1982: {'lr': 0.0004955, 'samples': 380736, 'steps': 1982, 'loss/train': 5.086475729942322} 01/27/2022 21:15:04 - INFO - codeparrot_training - Step 1983: {'lr': 0.00049575, 'samples': 380928, 'steps': 1983, 'loss/train': 5.406883120536804} 01/27/2022 21:15:08 - INFO - codeparrot_training - Step 1984: {'lr': 0.000496, 'samples': 381120, 'steps': 1984, 'loss/train': 4.5938379764556885} 01/27/2022 21:15:12 - INFO - codeparrot_training - Step 1985: {'lr': 0.0004962500000000001, 'samples': 381312, 'steps': 1985, 'loss/train': 3.1530808210372925} 01/27/2022 21:15:18 - INFO - codeparrot_training - Step 1986: {'lr': 0.0004965, 'samples': 381504, 'steps': 1986, 'loss/train': 4.398847460746765} 01/27/2022 21:15:22 - INFO - codeparrot_training - Step 1987: {'lr': 0.00049675, 'samples': 381696, 'steps': 1987, 'loss/train': 4.239651918411255} 01/27/2022 21:15:26 - INFO - codeparrot_training - Step 1988: {'lr': 0.000497, 'samples': 381888, 'steps': 1988, 'loss/train': 5.302281618118286} 01/27/2022 21:15:30 - INFO - codeparrot_training - Step 1989: {'lr': 0.0004972500000000001, 'samples': 382080, 'steps': 1989, 'loss/train': 4.933800101280212} 01/27/2022 21:15:34 - INFO - codeparrot_training - Step 1990: {'lr': 0.0004975, 'samples': 382272, 'steps': 1990, 'loss/train': 3.3851537704467773} 01/27/2022 21:15:40 - INFO - codeparrot_training - Step 1991: {'lr': 0.00049775, 'samples': 382464, 'steps': 1991, 'loss/train': 4.300583839416504} 01/27/2022 21:15:44 - INFO - codeparrot_training - Step 1992: {'lr': 0.000498, 'samples': 382656, 'steps': 1992, 'loss/train': 3.1776952743530273} 01/27/2022 21:15:48 - INFO - codeparrot_training - Step 1993: {'lr': 0.00049825, 'samples': 382848, 'steps': 1993, 'loss/train': 4.60379683971405} 01/27/2022 21:15:52 - INFO - codeparrot_training - Step 1994: {'lr': 0.0004985, 'samples': 383040, 'steps': 1994, 'loss/train': 4.928932785987854} 01/27/2022 21:15:56 - INFO - codeparrot_training - Step 1995: {'lr': 0.0004987500000000001, 'samples': 383232, 'steps': 1995, 'loss/train': 4.511248826980591} 01/27/2022 21:16:02 - INFO - codeparrot_training - Step 1996: {'lr': 0.000499, 'samples': 383424, 'steps': 1996, 'loss/train': 5.340387940406799} 01/27/2022 21:16:06 - INFO - codeparrot_training - Step 1997: {'lr': 0.00049925, 'samples': 383616, 'steps': 1997, 'loss/train': 4.429168224334717} 01/27/2022 21:16:10 - INFO - codeparrot_training - Step 1998: {'lr': 0.0004995, 'samples': 383808, 'steps': 1998, 'loss/train': 4.711615204811096} 01/27/2022 21:16:15 - INFO - codeparrot_training - Step 1999: {'lr': 0.0004997500000000001, 'samples': 384000, 'steps': 1999, 'loss/train': 4.689506649971008} 01/27/2022 21:16:15 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/27/2022 21:20:56 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py * [new branch] colorful-plasma-1 -> colorful-plasma-1 01/27/2022 21:22:08 - INFO - codeparrot_training - Step 2000: {'lr': 0.0005, 'samples': 384192, 'steps': 2000, 'loss/train': 5.073051452636719} 01/27/2022 21:22:14 - INFO - codeparrot_training - Step 2001: {'lr': 0.0004999999994645397, 'samples': 384384, 'steps': 2001, 'loss/train': 3.1000488996505737} 01/27/2022 21:22:18 - INFO - codeparrot_training - Step 2002: {'lr': 0.0004999999978581587, 'samples': 384576, 'steps': 2002, 'loss/train': 3.5170061588287354} 01/27/2022 21:22:22 - INFO - codeparrot_training - Step 2003: {'lr': 0.0004999999951808573, 'samples': 384768, 'steps': 2003, 'loss/train': 3.9817121028900146} 01/27/2022 21:22:26 - INFO - codeparrot_training - Step 2004: {'lr': 0.0004999999914326351, 'samples': 384960, 'steps': 2004, 'loss/train': 8.102889060974121} 01/27/2022 21:22:30 - INFO - codeparrot_training - Step 2005: {'lr': 0.0004999999866134924, 'samples': 385152, 'steps': 2005, 'loss/train': 4.138147830963135} 01/27/2022 21:22:36 - INFO - codeparrot_training - Step 2006: {'lr': 0.0004999999807234292, 'samples': 385344, 'steps': 2006, 'loss/train': 2.5845749974250793} 01/27/2022 21:22:40 - INFO - codeparrot_training - Step 2007: {'lr': 0.0004999999737624453, 'samples': 385536, 'steps': 2007, 'loss/train': 3.132956027984619} 01/27/2022 21:22:44 - INFO - codeparrot_training - Step 2008: {'lr': 0.0004999999657305411, 'samples': 385728, 'steps': 2008, 'loss/train': 4.171491980552673} 01/27/2022 21:22:48 - INFO - codeparrot_training - Step 2009: {'lr': 0.0004999999566277163, 'samples': 385920, 'steps': 2009, 'loss/train': 2.699103355407715} 01/27/2022 21:22:53 - INFO - codeparrot_training - Step 2010: {'lr': 0.0004999999464539711, 'samples': 386112, 'steps': 2010, 'loss/train': 4.613174200057983} 01/27/2022 21:22:58 - INFO - codeparrot_training - Step 2011: {'lr': 0.0004999999352093055, 'samples': 386304, 'steps': 2011, 'loss/train': 4.878560185432434} 01/27/2022 21:23:02 - INFO - codeparrot_training - Step 2012: {'lr': 0.0004999999228937196, 'samples': 386496, 'steps': 2012, 'loss/train': 2.9797768592834473} 01/27/2022 21:23:07 - INFO - codeparrot_training - Step 2013: {'lr': 0.0004999999095072135, 'samples': 386688, 'steps': 2013, 'loss/train': 3.9190346002578735} 01/27/2022 21:23:11 - INFO - codeparrot_training - Step 2014: {'lr': 0.0004999998950497869, 'samples': 386880, 'steps': 2014, 'loss/train': 4.584710597991943} 01/27/2022 21:23:15 - INFO - codeparrot_training - Step 2015: {'lr': 0.0004999998795214404, 'samples': 387072, 'steps': 2015, 'loss/train': 4.217143893241882} 01/27/2022 21:23:21 - INFO - codeparrot_training - Step 2016: {'lr': 0.0004999998629221736, 'samples': 387264, 'steps': 2016, 'loss/train': 3.7688881158828735} 01/27/2022 21:23:25 - INFO - codeparrot_training - Step 2017: {'lr': 0.0004999998452519869, 'samples': 387456, 'steps': 2017, 'loss/train': 4.648779273033142} 01/27/2022 21:23:30 - INFO - codeparrot_training - Step 2018: {'lr': 0.0004999998265108802, 'samples': 387648, 'steps': 2018, 'loss/train': 4.526477694511414} 01/27/2022 21:23:34 - INFO - codeparrot_training - Step 2019: {'lr': 0.0004999998066988537, 'samples': 387840, 'steps': 2019, 'loss/train': 4.835636615753174} 01/27/2022 21:23:38 - INFO - codeparrot_training - Step 2020: {'lr': 0.0004999997858159073, 'samples': 388032, 'steps': 2020, 'loss/train': 4.577046990394592} 01/27/2022 21:23:44 - INFO - codeparrot_training - Step 2021: {'lr': 0.0004999997638620412, 'samples': 388224, 'steps': 2021, 'loss/train': 4.602665662765503} 01/27/2022 21:23:48 - INFO - codeparrot_training - Step 2022: {'lr': 0.0004999997408372557, 'samples': 388416, 'steps': 2022, 'loss/train': 5.0575772523880005} 01/27/2022 21:23:52 - INFO - codeparrot_training - Step 2023: {'lr': 0.0004999997167415504, 'samples': 388608, 'steps': 2023, 'loss/train': 4.149248957633972} 01/27/2022 21:23:57 - INFO - codeparrot_training - Step 2024: {'lr': 0.0004999996915749259, 'samples': 388800, 'steps': 2024, 'loss/train': 7.0745837688446045} 01/27/2022 21:24:01 - INFO - codeparrot_training - Step 2025: {'lr': 0.0004999996653373821, 'samples': 388992, 'steps': 2025, 'loss/train': 4.072395086288452} 01/27/2022 21:24:07 - INFO - codeparrot_training - Step 2026: {'lr': 0.000499999638028919, 'samples': 389184, 'steps': 2026, 'loss/train': 3.2807092666625977} 01/27/2022 21:24:11 - INFO - codeparrot_training - Step 2027: {'lr': 0.0004999996096495369, 'samples': 389376, 'steps': 2027, 'loss/train': 5.0650516748428345} 01/27/2022 21:24:15 - INFO - codeparrot_training - Step 2028: {'lr': 0.0004999995801992359, 'samples': 389568, 'steps': 2028, 'loss/train': 4.054237246513367} 01/27/2022 21:24:20 - INFO - codeparrot_training - Step 2029: {'lr': 0.000499999549678016, 'samples': 389760, 'steps': 2029, 'loss/train': 4.451490640640259} 01/27/2022 21:24:24 - INFO - codeparrot_training - Step 2030: {'lr': 0.0004999995180858774, 'samples': 389952, 'steps': 2030, 'loss/train': 4.5836509466171265} 01/27/2022 21:24:29 - INFO - codeparrot_training - Step 2031: {'lr': 0.0004999994854228203, 'samples': 390144, 'steps': 2031, 'loss/train': 4.5753690004348755} 01/27/2022 21:24:33 - INFO - codeparrot_training - Step 2032: {'lr': 0.0004999994516888449, 'samples': 390336, 'steps': 2032, 'loss/train': 3.1286333799362183} 01/27/2022 21:24:37 - INFO - codeparrot_training - Step 2033: {'lr': 0.000499999416883951, 'samples': 390528, 'steps': 2033, 'loss/train': 3.4229815006256104} 01/27/2022 21:24:42 - INFO - codeparrot_training - Step 2034: {'lr': 0.0004999993810081391, 'samples': 390720, 'steps': 2034, 'loss/train': 4.54124128818512} 01/27/2022 21:24:46 - INFO - codeparrot_training - Step 2035: {'lr': 0.0004999993440614092, 'samples': 390912, 'steps': 2035, 'loss/train': 4.977101683616638} 01/27/2022 21:24:51 - INFO - codeparrot_training - Step 2036: {'lr': 0.0004999993060437616, 'samples': 391104, 'steps': 2036, 'loss/train': 3.207081913948059} 01/27/2022 21:24:55 - INFO - codeparrot_training - Step 2037: {'lr': 0.0004999992669551962, 'samples': 391296, 'steps': 2037, 'loss/train': 4.949477434158325} 01/27/2022 21:25:00 - INFO - codeparrot_training - Step 2038: {'lr': 0.0004999992267957135, 'samples': 391488, 'steps': 2038, 'loss/train': 4.599668025970459} 01/27/2022 21:25:04 - INFO - codeparrot_training - Step 2039: {'lr': 0.0004999991855653134, 'samples': 391680, 'steps': 2039, 'loss/train': 4.165125846862793} 01/27/2022 21:25:08 - INFO - codeparrot_training - Step 2040: {'lr': 0.0004999991432639963, 'samples': 391872, 'steps': 2040, 'loss/train': 3.061027765274048} 01/27/2022 21:25:14 - INFO - codeparrot_training - Step 2041: {'lr': 0.0004999990998917621, 'samples': 392064, 'steps': 2041, 'loss/train': 4.200434446334839} 01/27/2022 21:25:18 - INFO - codeparrot_training - Step 2042: {'lr': 0.0004999990554486111, 'samples': 392256, 'steps': 2042, 'loss/train': 2.973904073238373} 01/27/2022 21:25:22 - INFO - codeparrot_training - Step 2043: {'lr': 0.0004999990099345436, 'samples': 392448, 'steps': 2043, 'loss/train': 4.64750862121582} 01/27/2022 21:25:27 - INFO - codeparrot_training - Step 2044: {'lr': 0.0004999989633495597, 'samples': 392640, 'steps': 2044, 'loss/train': 3.329468250274658} 01/27/2022 21:25:31 - INFO - codeparrot_training - Step 2045: {'lr': 0.0004999989156936597, 'samples': 392832, 'steps': 2045, 'loss/train': 5.07621431350708} 01/27/2022 21:25:36 - INFO - codeparrot_training - Step 2046: {'lr': 0.0004999988669668437, 'samples': 393024, 'steps': 2046, 'loss/train': 3.6208301782608032} 01/27/2022 21:25:40 - INFO - codeparrot_training - Step 2047: {'lr': 0.0004999988171691119, 'samples': 393216, 'steps': 2047, 'loss/train': 5.139476895332336} 01/27/2022 21:25:45 - INFO - codeparrot_training - Step 2048: {'lr': 0.0004999987663004646, 'samples': 393408, 'steps': 2048, 'loss/train': 4.006467819213867} 01/27/2022 21:25:49 - INFO - codeparrot_training - Step 2049: {'lr': 0.0004999987143609019, 'samples': 393600, 'steps': 2049, 'loss/train': 3.8665266036987305} 01/27/2022 21:25:53 - INFO - codeparrot_training - Step 2050: {'lr': 0.0004999986613504242, 'samples': 393792, 'steps': 2050, 'loss/train': 4.6785889863967896} 01/27/2022 21:25:59 - INFO - codeparrot_training - Step 2051: {'lr': 0.0004999986072690315, 'samples': 393984, 'steps': 2051, 'loss/train': 2.875666379928589} 01/27/2022 21:26:03 - INFO - codeparrot_training - Step 2052: {'lr': 0.0004999985521167242, 'samples': 394176, 'steps': 2052, 'loss/train': 3.412823438644409} 01/27/2022 21:26:07 - INFO - codeparrot_training - Step 2053: {'lr': 0.0004999984958935025, 'samples': 394368, 'steps': 2053, 'loss/train': 4.369283080101013} 01/27/2022 21:26:12 - INFO - codeparrot_training - Step 2054: {'lr': 0.0004999984385993665, 'samples': 394560, 'steps': 2054, 'loss/train': 5.066033720970154} 01/27/2022 21:26:16 - INFO - codeparrot_training - Step 2055: {'lr': 0.0004999983802343168, 'samples': 394752, 'steps': 2055, 'loss/train': 4.661246180534363} 01/27/2022 21:26:21 - INFO - codeparrot_training - Step 2056: {'lr': 0.0004999983207983532, 'samples': 394944, 'steps': 2056, 'loss/train': 3.8914815187454224} 01/27/2022 21:26:25 - INFO - codeparrot_training - Step 2057: {'lr': 0.0004999982602914763, 'samples': 395136, 'steps': 2057, 'loss/train': 4.6746954917907715} 01/27/2022 21:26:30 - INFO - codeparrot_training - Step 2058: {'lr': 0.0004999981987136862, 'samples': 395328, 'steps': 2058, 'loss/train': 2.4994254112243652} 01/27/2022 21:26:34 - INFO - codeparrot_training - Step 2059: {'lr': 0.0004999981360649833, 'samples': 395520, 'steps': 2059, 'loss/train': 5.1682223081588745} 01/27/2022 21:26:38 - INFO - codeparrot_training - Step 2060: {'lr': 0.0004999980723453676, 'samples': 395712, 'steps': 2060, 'loss/train': 3.8289270401000977} 01/27/2022 21:26:43 - INFO - codeparrot_training - Step 2061: {'lr': 0.0004999980075548397, 'samples': 395904, 'steps': 2061, 'loss/train': 3.9833943843841553} 01/27/2022 21:26:47 - INFO - codeparrot_training - Step 2062: {'lr': 0.0004999979416933997, 'samples': 396096, 'steps': 2062, 'loss/train': 5.4817986488342285} 01/27/2022 21:26:52 - INFO - codeparrot_training - Step 2063: {'lr': 0.0004999978747610478, 'samples': 396288, 'steps': 2063, 'loss/train': 3.70340359210968} 01/27/2022 21:26:56 - INFO - codeparrot_training - Step 2064: {'lr': 0.0004999978067577843, 'samples': 396480, 'steps': 2064, 'loss/train': 3.2504950761795044} 01/27/2022 21:27:00 - INFO - codeparrot_training - Step 2065: {'lr': 0.0004999977376836098, 'samples': 396672, 'steps': 2065, 'loss/train': 3.695536494255066} 01/27/2022 21:27:06 - INFO - codeparrot_training - Step 2066: {'lr': 0.0004999976675385243, 'samples': 396864, 'steps': 2066, 'loss/train': 4.730744004249573} 01/27/2022 21:27:10 - INFO - codeparrot_training - Step 2067: {'lr': 0.0004999975963225282, 'samples': 397056, 'steps': 2067, 'loss/train': 4.419664263725281} 01/27/2022 21:27:15 - INFO - codeparrot_training - Step 2068: {'lr': 0.0004999975240356217, 'samples': 397248, 'steps': 2068, 'loss/train': 6.171286582946777} 01/27/2022 21:27:19 - INFO - codeparrot_training - Step 2069: {'lr': 0.0004999974506778053, 'samples': 397440, 'steps': 2069, 'loss/train': 3.7731549739837646} 01/27/2022 21:27:23 - INFO - codeparrot_training - Step 2070: {'lr': 0.0004999973762490792, 'samples': 397632, 'steps': 2070, 'loss/train': 4.054191470146179} 01/27/2022 21:27:28 - INFO - codeparrot_training - Step 2071: {'lr': 0.0004999973007494436, 'samples': 397824, 'steps': 2071, 'loss/train': 3.2632373571395874} 01/27/2022 21:27:32 - INFO - codeparrot_training - Step 2072: {'lr': 0.000499997224178899, 'samples': 398016, 'steps': 2072, 'loss/train': 2.8741047978401184} 01/27/2022 21:27:37 - INFO - codeparrot_training - Step 2073: {'lr': 0.0004999971465374457, 'samples': 398208, 'steps': 2073, 'loss/train': 5.296878933906555} 01/27/2022 21:27:41 - INFO - codeparrot_training - Step 2074: {'lr': 0.000499997067825084, 'samples': 398400, 'steps': 2074, 'loss/train': 5.080601692199707} 01/27/2022 21:27:45 - INFO - codeparrot_training - Step 2075: {'lr': 0.0004999969880418142, 'samples': 398592, 'steps': 2075, 'loss/train': 3.975107431411743} 01/27/2022 21:27:50 - INFO - codeparrot_training - Step 2076: {'lr': 0.0004999969071876367, 'samples': 398784, 'steps': 2076, 'loss/train': 4.16400146484375} 01/27/2022 21:27:55 - INFO - codeparrot_training - Step 2077: {'lr': 0.0004999968252625519, 'samples': 398976, 'steps': 2077, 'loss/train': 5.024808168411255} 01/27/2022 21:27:59 - INFO - codeparrot_training - Step 2078: {'lr': 0.00049999674226656, 'samples': 399168, 'steps': 2078, 'loss/train': 4.704577803611755} 01/27/2022 21:28:03 - INFO - codeparrot_training - Step 2079: {'lr': 0.0004999966581996616, 'samples': 399360, 'steps': 2079, 'loss/train': 4.41048789024353} 01/27/2022 21:28:07 - INFO - codeparrot_training - Step 2080: {'lr': 0.0004999965730618567, 'samples': 399552, 'steps': 2080, 'loss/train': 3.1104930639266968} 01/27/2022 21:28:13 - INFO - codeparrot_training - Step 2081: {'lr': 0.000499996486853146, 'samples': 399744, 'steps': 2081, 'loss/train': 11.17986273765564} 01/27/2022 21:28:17 - INFO - codeparrot_training - Step 2082: {'lr': 0.0004999963995735296, 'samples': 399936, 'steps': 2082, 'loss/train': 4.030752539634705} 01/27/2022 21:28:21 - INFO - codeparrot_training - Step 2083: {'lr': 0.0004999963112230081, 'samples': 400128, 'steps': 2083, 'loss/train': 3.9683425426483154} 01/27/2022 21:28:26 - INFO - codeparrot_training - Step 2084: {'lr': 0.0004999962218015818, 'samples': 400320, 'steps': 2084, 'loss/train': 5.263067722320557} 01/27/2022 21:28:30 - INFO - codeparrot_training - Step 2085: {'lr': 0.0004999961313092511, 'samples': 400512, 'steps': 2085, 'loss/train': 4.152779817581177} 01/27/2022 21:28:36 - INFO - codeparrot_training - Step 2086: {'lr': 0.0004999960397460162, 'samples': 400704, 'steps': 2086, 'loss/train': 4.800428867340088} 01/27/2022 21:28:40 - INFO - codeparrot_training - Step 2087: {'lr': 0.0004999959471118778, 'samples': 400896, 'steps': 2087, 'loss/train': 4.494580507278442} 01/27/2022 21:28:44 - INFO - codeparrot_training - Step 2088: {'lr': 0.000499995853406836, 'samples': 401088, 'steps': 2088, 'loss/train': 5.518758058547974} 01/27/2022 21:28:49 - INFO - codeparrot_training - Step 2089: {'lr': 0.0004999957586308914, 'samples': 401280, 'steps': 2089, 'loss/train': 4.084565877914429} 01/27/2022 21:28:53 - INFO - codeparrot_training - Step 2090: {'lr': 0.0004999956627840445, 'samples': 401472, 'steps': 2090, 'loss/train': 3.9640209674835205} 01/27/2022 21:28:58 - INFO - codeparrot_training - Step 2091: {'lr': 0.0004999955658662954, 'samples': 401664, 'steps': 2091, 'loss/train': 3.4384775161743164} 01/27/2022 21:29:02 - INFO - codeparrot_training - Step 2092: {'lr': 0.0004999954678776448, 'samples': 401856, 'steps': 2092, 'loss/train': 4.474225044250488} 01/27/2022 21:29:07 - INFO - codeparrot_training - Step 2093: {'lr': 0.0004999953688180929, 'samples': 402048, 'steps': 2093, 'loss/train': 4.536333203315735} 01/27/2022 21:29:11 - INFO - codeparrot_training - Step 2094: {'lr': 0.0004999952686876402, 'samples': 402240, 'steps': 2094, 'loss/train': 5.422452092170715} 01/27/2022 21:29:15 - INFO - codeparrot_training - Step 2095: {'lr': 0.0004999951674862872, 'samples': 402432, 'steps': 2095, 'loss/train': 3.445524215698242} 01/27/2022 21:29:20 - INFO - codeparrot_training - Step 2096: {'lr': 0.0004999950652140343, 'samples': 402624, 'steps': 2096, 'loss/train': 3.3091013431549072} 01/27/2022 21:29:24 - INFO - codeparrot_training - Step 2097: {'lr': 0.0004999949618708819, 'samples': 402816, 'steps': 2097, 'loss/train': 4.574344038963318} 01/27/2022 21:29:29 - INFO - codeparrot_training - Step 2098: {'lr': 0.0004999948574568305, 'samples': 403008, 'steps': 2098, 'loss/train': 4.306064486503601} 01/27/2022 21:29:33 - INFO - codeparrot_training - Step 2099: {'lr': 0.0004999947519718805, 'samples': 403200, 'steps': 2099, 'loss/train': 4.382693767547607} 01/27/2022 21:29:37 - INFO - codeparrot_training - Step 2100: {'lr': 0.0004999946454160324, 'samples': 403392, 'steps': 2100, 'loss/train': 3.848219633102417} 01/27/2022 21:29:43 - INFO - codeparrot_training - Step 2101: {'lr': 0.0004999945377892865, 'samples': 403584, 'steps': 2101, 'loss/train': 4.024955749511719} 01/27/2022 21:29:47 - INFO - codeparrot_training - Step 2102: {'lr': 0.0004999944290916434, 'samples': 403776, 'steps': 2102, 'loss/train': 2.8755378127098083} 01/27/2022 21:29:52 - INFO - codeparrot_training - Step 2103: {'lr': 0.0004999943193231037, 'samples': 403968, 'steps': 2103, 'loss/train': 4.409406423568726} 01/27/2022 21:29:56 - INFO - codeparrot_training - Step 2104: {'lr': 0.0004999942084836675, 'samples': 404160, 'steps': 2104, 'loss/train': 4.4839160442352295} 01/27/2022 21:30:00 - INFO - codeparrot_training - Step 2105: {'lr': 0.0004999940965733356, 'samples': 404352, 'steps': 2105, 'loss/train': 4.052880048751831} 01/27/2022 21:30:05 - INFO - codeparrot_training - Step 2106: {'lr': 0.0004999939835921085, 'samples': 404544, 'steps': 2106, 'loss/train': 3.1265920400619507} 01/27/2022 21:30:09 - INFO - codeparrot_training - Step 2107: {'lr': 0.0004999938695399864, 'samples': 404736, 'steps': 2107, 'loss/train': 4.27812659740448} 01/27/2022 21:30:14 - INFO - codeparrot_training - Step 2108: {'lr': 0.00049999375441697, 'samples': 404928, 'steps': 2108, 'loss/train': 3.2335166931152344} 01/27/2022 21:30:18 - INFO - codeparrot_training - Step 2109: {'lr': 0.0004999936382230597, 'samples': 405120, 'steps': 2109, 'loss/train': 4.09600567817688} 01/27/2022 21:30:22 - INFO - codeparrot_training - Step 2110: {'lr': 0.000499993520958256, 'samples': 405312, 'steps': 2110, 'loss/train': 4.117807030677795} 01/27/2022 21:30:29 - INFO - codeparrot_training - Step 2111: {'lr': 0.0004999934026225595, 'samples': 405504, 'steps': 2111, 'loss/train': 4.169109463691711} 01/27/2022 21:30:33 - INFO - codeparrot_training - Step 2112: {'lr': 0.0004999932832159708, 'samples': 405696, 'steps': 2112, 'loss/train': 4.156901836395264} 01/27/2022 21:30:37 - INFO - codeparrot_training - Step 2113: {'lr': 0.00049999316273849, 'samples': 405888, 'steps': 2113, 'loss/train': 4.685732960700989} 01/27/2022 21:30:41 - INFO - codeparrot_training - Step 2114: {'lr': 0.0004999930411901181, 'samples': 406080, 'steps': 2114, 'loss/train': 4.58092188835144} 01/27/2022 21:30:45 - INFO - codeparrot_training - Step 2115: {'lr': 0.0004999929185708551, 'samples': 406272, 'steps': 2115, 'loss/train': 3.422151803970337} 01/27/2022 21:30:50 - INFO - codeparrot_training - Step 2116: {'lr': 0.000499992794880702, 'samples': 406464, 'steps': 2116, 'loss/train': 5.382795453071594} 01/27/2022 21:30:55 - INFO - codeparrot_training - Step 2117: {'lr': 0.0004999926701196592, 'samples': 406656, 'steps': 2117, 'loss/train': 4.561665415763855} 01/27/2022 21:30:59 - INFO - codeparrot_training - Step 2118: {'lr': 0.0004999925442877271, 'samples': 406848, 'steps': 2118, 'loss/train': 5.569707155227661} 01/27/2022 21:31:03 - INFO - codeparrot_training - Step 2119: {'lr': 0.0004999924173849063, 'samples': 407040, 'steps': 2119, 'loss/train': 3.373836636543274} 01/27/2022 21:31:08 - INFO - codeparrot_training - Step 2120: {'lr': 0.0004999922894111975, 'samples': 407232, 'steps': 2120, 'loss/train': 4.534941673278809} 01/27/2022 21:31:12 - INFO - codeparrot_training - Step 2121: {'lr': 0.000499992160366601, 'samples': 407424, 'steps': 2121, 'loss/train': 4.071024298667908} 01/27/2022 21:31:17 - INFO - codeparrot_training - Step 2122: {'lr': 0.0004999920302511175, 'samples': 407616, 'steps': 2122, 'loss/train': 4.302779674530029} 01/27/2022 21:31:21 - INFO - codeparrot_training - Step 2123: {'lr': 0.0004999918990647474, 'samples': 407808, 'steps': 2123, 'loss/train': 3.247424840927124} 01/27/2022 21:31:26 - INFO - codeparrot_training - Step 2124: {'lr': 0.0004999917668074915, 'samples': 408000, 'steps': 2124, 'loss/train': 5.512786030769348} 01/27/2022 21:31:30 - INFO - codeparrot_training - Step 2125: {'lr': 0.0004999916334793503, 'samples': 408192, 'steps': 2125, 'loss/train': 4.6728222370147705} 01/27/2022 21:31:34 - INFO - codeparrot_training - Step 2126: {'lr': 0.0004999914990803242, 'samples': 408384, 'steps': 2126, 'loss/train': 4.196993350982666} 01/27/2022 21:31:40 - INFO - codeparrot_training - Step 2127: {'lr': 0.000499991363610414, 'samples': 408576, 'steps': 2127, 'loss/train': 3.7132787704467773} 01/27/2022 21:31:45 - INFO - codeparrot_training - Step 2128: {'lr': 0.0004999912270696202, 'samples': 408768, 'steps': 2128, 'loss/train': 4.194051146507263} 01/27/2022 21:31:49 - INFO - codeparrot_training - Step 2129: {'lr': 0.0004999910894579432, 'samples': 408960, 'steps': 2129, 'loss/train': 3.3209080696105957} 01/27/2022 21:31:53 - INFO - codeparrot_training - Step 2130: {'lr': 0.000499990950775384, 'samples': 409152, 'steps': 2130, 'loss/train': 4.156091809272766} 01/27/2022 21:31:57 - INFO - codeparrot_training - Step 2131: {'lr': 0.0004999908110219428, 'samples': 409344, 'steps': 2131, 'loss/train': 5.197344660758972} 01/27/2022 21:32:02 - INFO - codeparrot_training - Step 2132: {'lr': 0.0004999906701976203, 'samples': 409536, 'steps': 2132, 'loss/train': 4.95992910861969} 01/27/2022 21:32:07 - INFO - codeparrot_training - Step 2133: {'lr': 0.0004999905283024172, 'samples': 409728, 'steps': 2133, 'loss/train': 5.221668004989624} 01/27/2022 21:32:11 - INFO - codeparrot_training - Step 2134: {'lr': 0.0004999903853363341, 'samples': 409920, 'steps': 2134, 'loss/train': 4.242903828620911} 01/27/2022 21:32:15 - INFO - codeparrot_training - Step 2135: {'lr': 0.0004999902412993715, 'samples': 410112, 'steps': 2135, 'loss/train': 4.219412326812744} 01/27/2022 21:32:19 - INFO - codeparrot_training - Step 2136: {'lr': 0.0004999900961915302, 'samples': 410304, 'steps': 2136, 'loss/train': 3.7504652738571167} 01/27/2022 21:32:25 - INFO - codeparrot_training - Step 2137: {'lr': 0.0004999899500128107, 'samples': 410496, 'steps': 2137, 'loss/train': 4.167574167251587} 01/27/2022 21:32:29 - INFO - codeparrot_training - Step 2138: {'lr': 0.0004999898027632135, 'samples': 410688, 'steps': 2138, 'loss/train': 3.293210506439209} 01/27/2022 21:32:33 - INFO - codeparrot_training - Step 2139: {'lr': 0.0004999896544427394, 'samples': 410880, 'steps': 2139, 'loss/train': 3.8507834672927856} 01/27/2022 21:32:37 - INFO - codeparrot_training - Step 2140: {'lr': 0.0004999895050513891, 'samples': 411072, 'steps': 2140, 'loss/train': 3.4403539896011353} 01/27/2022 21:32:41 - INFO - codeparrot_training - Step 2141: {'lr': 0.0004999893545891631, 'samples': 411264, 'steps': 2141, 'loss/train': 5.375686526298523} 01/27/2022 21:32:47 - INFO - codeparrot_training - Step 2142: {'lr': 0.000499989203056062, 'samples': 411456, 'steps': 2142, 'loss/train': 3.810188412666321} 01/27/2022 21:32:51 - INFO - codeparrot_training - Step 2143: {'lr': 0.0004999890504520866, 'samples': 411648, 'steps': 2143, 'loss/train': 3.5443453788757324} 01/27/2022 21:32:55 - INFO - codeparrot_training - Step 2144: {'lr': 0.0004999888967772375, 'samples': 411840, 'steps': 2144, 'loss/train': 3.1384588479995728} 01/27/2022 21:32:59 - INFO - codeparrot_training - Step 2145: {'lr': 0.0004999887420315153, 'samples': 412032, 'steps': 2145, 'loss/train': 4.83708393573761} 01/27/2022 21:33:04 - INFO - codeparrot_training - Step 2146: {'lr': 0.0004999885862149207, 'samples': 412224, 'steps': 2146, 'loss/train': 4.070842266082764} 01/27/2022 21:33:10 - INFO - codeparrot_training - Step 2147: {'lr': 0.0004999884293274545, 'samples': 412416, 'steps': 2147, 'loss/train': 1.7194468975067139} 01/27/2022 21:33:14 - INFO - codeparrot_training - Step 2148: {'lr': 0.0004999882713691171, 'samples': 412608, 'steps': 2148, 'loss/train': 3.0897492170333862} 01/27/2022 21:33:18 - INFO - codeparrot_training - Step 2149: {'lr': 0.0004999881123399093, 'samples': 412800, 'steps': 2149, 'loss/train': 5.596706986427307} 01/27/2022 21:33:22 - INFO - codeparrot_training - Step 2150: {'lr': 0.000499987952239832, 'samples': 412992, 'steps': 2150, 'loss/train': 4.4542014598846436} 01/27/2022 21:33:26 - INFO - codeparrot_training - Step 2151: {'lr': 0.0004999877910688856, 'samples': 413184, 'steps': 2151, 'loss/train': 3.396170496940613} 01/27/2022 21:33:32 - INFO - codeparrot_training - Step 2152: {'lr': 0.0004999876288270708, 'samples': 413376, 'steps': 2152, 'loss/train': 2.9824742674827576} 01/27/2022 21:33:36 - INFO - codeparrot_training - Step 2153: {'lr': 0.0004999874655143886, 'samples': 413568, 'steps': 2153, 'loss/train': 3.2150237560272217} 01/27/2022 21:33:40 - INFO - codeparrot_training - Step 2154: {'lr': 0.0004999873011308393, 'samples': 413760, 'steps': 2154, 'loss/train': 2.878329813480377} 01/27/2022 21:33:45 - INFO - codeparrot_training - Step 2155: {'lr': 0.0004999871356764238, 'samples': 413952, 'steps': 2155, 'loss/train': 3.38191294670105} 01/27/2022 21:33:49 - INFO - codeparrot_training - Step 2156: {'lr': 0.0004999869691511428, 'samples': 414144, 'steps': 2156, 'loss/train': 3.8805960416793823} 01/27/2022 21:33:55 - INFO - codeparrot_training - Step 2157: {'lr': 0.000499986801554997, 'samples': 414336, 'steps': 2157, 'loss/train': 4.830006122589111} 01/27/2022 21:33:59 - INFO - codeparrot_training - Step 2158: {'lr': 0.0004999866328879871, 'samples': 414528, 'steps': 2158, 'loss/train': 2.327420711517334} 01/27/2022 21:34:03 - INFO - codeparrot_training - Step 2159: {'lr': 0.0004999864631501139, 'samples': 414720, 'steps': 2159, 'loss/train': 4.4707642793655396} 01/27/2022 21:34:07 - INFO - codeparrot_training - Step 2160: {'lr': 0.000499986292341378, 'samples': 414912, 'steps': 2160, 'loss/train': 4.192153573036194} 01/27/2022 21:34:12 - INFO - codeparrot_training - Step 2161: {'lr': 0.0004999861204617803, 'samples': 415104, 'steps': 2161, 'loss/train': 1.5697333216667175} 01/27/2022 21:34:17 - INFO - codeparrot_training - Step 2162: {'lr': 0.0004999859475113213, 'samples': 415296, 'steps': 2162, 'loss/train': 4.02592921257019} 01/27/2022 21:34:21 - INFO - codeparrot_training - Step 2163: {'lr': 0.0004999857734900021, 'samples': 415488, 'steps': 2163, 'loss/train': 3.940733313560486} 01/27/2022 21:34:25 - INFO - codeparrot_training - Step 2164: {'lr': 0.000499985598397823, 'samples': 415680, 'steps': 2164, 'loss/train': 4.399367451667786} 01/27/2022 21:34:30 - INFO - codeparrot_training - Step 2165: {'lr': 0.0004999854222347851, 'samples': 415872, 'steps': 2165, 'loss/train': 3.9896997213363647} 01/27/2022 21:34:34 - INFO - codeparrot_training - Step 2166: {'lr': 0.000499985245000889, 'samples': 416064, 'steps': 2166, 'loss/train': 3.1833064556121826} 01/27/2022 21:34:39 - INFO - codeparrot_training - Step 2167: {'lr': 0.0004999850666961355, 'samples': 416256, 'steps': 2167, 'loss/train': 3.050462007522583} 01/27/2022 21:34:44 - INFO - codeparrot_training - Step 2168: {'lr': 0.0004999848873205254, 'samples': 416448, 'steps': 2168, 'loss/train': 3.854872941970825} 01/27/2022 21:34:48 - INFO - codeparrot_training - Step 2169: {'lr': 0.0004999847068740593, 'samples': 416640, 'steps': 2169, 'loss/train': 4.035073399543762} 01/27/2022 21:34:52 - INFO - codeparrot_training - Step 2170: {'lr': 0.0004999845253567382, 'samples': 416832, 'steps': 2170, 'loss/train': 3.863155245780945} 01/27/2022 21:34:56 - INFO - codeparrot_training - Step 2171: {'lr': 0.0004999843427685627, 'samples': 417024, 'steps': 2171, 'loss/train': 4.088293075561523} 01/27/2022 21:35:02 - INFO - codeparrot_training - Step 2172: {'lr': 0.0004999841591095337, 'samples': 417216, 'steps': 2172, 'loss/train': 2.8361186385154724} 01/27/2022 21:35:06 - INFO - codeparrot_training - Step 2173: {'lr': 0.0004999839743796519, 'samples': 417408, 'steps': 2173, 'loss/train': 4.488118529319763} 01/27/2022 21:35:11 - INFO - codeparrot_training - Step 2174: {'lr': 0.0004999837885789182, 'samples': 417600, 'steps': 2174, 'loss/train': 4.699922561645508} 01/27/2022 21:35:15 - INFO - codeparrot_training - Step 2175: {'lr': 0.0004999836017073332, 'samples': 417792, 'steps': 2175, 'loss/train': 6.39881443977356} 01/27/2022 21:35:19 - INFO - codeparrot_training - Step 2176: {'lr': 0.000499983413764898, 'samples': 417984, 'steps': 2176, 'loss/train': 3.6661033630371094} 01/27/2022 21:35:24 - INFO - codeparrot_training - Step 2177: {'lr': 0.0004999832247516132, 'samples': 418176, 'steps': 2177, 'loss/train': 3.5030486583709717} 01/27/2022 21:35:29 - INFO - codeparrot_training - Step 2178: {'lr': 0.0004999830346674796, 'samples': 418368, 'steps': 2178, 'loss/train': 3.0596104860305786} 01/27/2022 21:35:33 - INFO - codeparrot_training - Step 2179: {'lr': 0.000499982843512498, 'samples': 418560, 'steps': 2179, 'loss/train': 3.8710895776748657} 01/27/2022 21:35:37 - INFO - codeparrot_training - Step 2180: {'lr': 0.0004999826512866693, 'samples': 418752, 'steps': 2180, 'loss/train': 3.066262364387512} 01/27/2022 21:35:41 - INFO - codeparrot_training - Step 2181: {'lr': 0.0004999824579899944, 'samples': 418944, 'steps': 2181, 'loss/train': 4.162054538726807} 01/27/2022 21:35:47 - INFO - codeparrot_training - Step 2182: {'lr': 0.000499982263622474, 'samples': 419136, 'steps': 2182, 'loss/train': 2.511970639228821} 01/27/2022 21:35:51 - INFO - codeparrot_training - Step 2183: {'lr': 0.0004999820681841088, 'samples': 419328, 'steps': 2183, 'loss/train': 4.345411419868469} 01/27/2022 21:35:55 - INFO - codeparrot_training - Step 2184: {'lr': 0.0004999818716748999, 'samples': 419520, 'steps': 2184, 'loss/train': 5.027921676635742} 01/27/2022 21:36:00 - INFO - codeparrot_training - Step 2185: {'lr': 0.0004999816740948481, 'samples': 419712, 'steps': 2185, 'loss/train': 4.7493427991867065} 01/27/2022 21:36:04 - INFO - codeparrot_training - Step 2186: {'lr': 0.0004999814754439542, 'samples': 419904, 'steps': 2186, 'loss/train': 3.238136887550354} 01/27/2022 21:36:09 - INFO - codeparrot_training - Step 2187: {'lr': 0.000499981275722219, 'samples': 420096, 'steps': 2187, 'loss/train': 3.8161407709121704} 01/27/2022 21:36:14 - INFO - codeparrot_training - Step 2188: {'lr': 0.0004999810749296434, 'samples': 420288, 'steps': 2188, 'loss/train': 4.119918465614319} 01/27/2022 21:36:18 - INFO - codeparrot_training - Step 2189: {'lr': 0.0004999808730662282, 'samples': 420480, 'steps': 2189, 'loss/train': 4.410783290863037} 01/27/2022 21:36:22 - INFO - codeparrot_training - Step 2190: {'lr': 0.0004999806701319743, 'samples': 420672, 'steps': 2190, 'loss/train': 5.142776370048523} 01/27/2022 21:36:27 - INFO - codeparrot_training - Step 2191: {'lr': 0.0004999804661268827, 'samples': 420864, 'steps': 2191, 'loss/train': 4.063244819641113} 01/27/2022 21:36:31 - INFO - codeparrot_training - Step 2192: {'lr': 0.0004999802610509541, 'samples': 421056, 'steps': 2192, 'loss/train': 4.947695016860962} 01/27/2022 21:36:36 - INFO - codeparrot_training - Step 2193: {'lr': 0.0004999800549041894, 'samples': 421248, 'steps': 2193, 'loss/train': 5.097684144973755} 01/27/2022 21:36:40 - INFO - codeparrot_training - Step 2194: {'lr': 0.0004999798476865895, 'samples': 421440, 'steps': 2194, 'loss/train': 4.061724901199341} 01/27/2022 21:36:44 - INFO - codeparrot_training - Step 2195: {'lr': 0.0004999796393981554, 'samples': 421632, 'steps': 2195, 'loss/train': 4.690984010696411} 01/27/2022 21:36:49 - INFO - codeparrot_training - Step 2196: {'lr': 0.0004999794300388879, 'samples': 421824, 'steps': 2196, 'loss/train': 3.912869095802307} 01/27/2022 21:36:54 - INFO - codeparrot_training - Step 2197: {'lr': 0.0004999792196087879, 'samples': 422016, 'steps': 2197, 'loss/train': 4.49416708946228} 01/27/2022 21:36:58 - INFO - codeparrot_training - Step 2198: {'lr': 0.0004999790081078562, 'samples': 422208, 'steps': 2198, 'loss/train': 4.248543977737427} 01/27/2022 21:37:02 - INFO - codeparrot_training - Step 2199: {'lr': 0.0004999787955360939, 'samples': 422400, 'steps': 2199, 'loss/train': 5.019331455230713} 01/27/2022 21:37:06 - INFO - codeparrot_training - Step 2200: {'lr': 0.0004999785818935018, 'samples': 422592, 'steps': 2200, 'loss/train': 3.692237377166748} 01/27/2022 21:37:12 - INFO - codeparrot_training - Step 2201: {'lr': 0.0004999783671800808, 'samples': 422784, 'steps': 2201, 'loss/train': 3.823124885559082} 01/27/2022 21:37:17 - INFO - codeparrot_training - Step 2202: {'lr': 0.0004999781513958318, 'samples': 422976, 'steps': 2202, 'loss/train': 3.1787166595458984} 01/27/2022 21:37:21 - INFO - codeparrot_training - Step 2203: {'lr': 0.000499977934540756, 'samples': 423168, 'steps': 2203, 'loss/train': 4.495790004730225} 01/27/2022 21:37:25 - INFO - codeparrot_training - Step 2204: {'lr': 0.0004999777166148539, 'samples': 423360, 'steps': 2204, 'loss/train': 4.097000241279602} 01/27/2022 21:37:29 - INFO - codeparrot_training - Step 2205: {'lr': 0.0004999774976181267, 'samples': 423552, 'steps': 2205, 'loss/train': 4.027989864349365} 01/27/2022 21:37:35 - INFO - codeparrot_training - Step 2206: {'lr': 0.0004999772775505753, 'samples': 423744, 'steps': 2206, 'loss/train': 3.7155221700668335} 01/27/2022 21:37:39 - INFO - codeparrot_training - Step 2207: {'lr': 0.0004999770564122005, 'samples': 423936, 'steps': 2207, 'loss/train': 4.0545841455459595} 01/27/2022 21:37:43 - INFO - codeparrot_training - Step 2208: {'lr': 0.0004999768342030035, 'samples': 424128, 'steps': 2208, 'loss/train': 3.0533941984176636} 01/27/2022 21:37:47 - INFO - codeparrot_training - Step 2209: {'lr': 0.0004999766109229851, 'samples': 424320, 'steps': 2209, 'loss/train': 4.600998759269714} 01/27/2022 21:37:52 - INFO - codeparrot_training - Step 2210: {'lr': 0.0004999763865721463, 'samples': 424512, 'steps': 2210, 'loss/train': 4.390320897102356} 01/27/2022 21:37:56 - INFO - codeparrot_training - Step 2211: {'lr': 0.000499976161150488, 'samples': 424704, 'steps': 2211, 'loss/train': 4.159342646598816} 01/27/2022 21:38:01 - INFO - codeparrot_training - Step 2212: {'lr': 0.0004999759346580111, 'samples': 424896, 'steps': 2212, 'loss/train': 4.8950114250183105} 01/27/2022 21:38:06 - INFO - codeparrot_training - Step 2213: {'lr': 0.0004999757070947168, 'samples': 425088, 'steps': 2213, 'loss/train': 4.534061908721924} 01/27/2022 21:38:10 - INFO - codeparrot_training - Step 2214: {'lr': 0.0004999754784606058, 'samples': 425280, 'steps': 2214, 'loss/train': 4.372095823287964} 01/27/2022 21:38:14 - INFO - codeparrot_training - Step 2215: {'lr': 0.0004999752487556794, 'samples': 425472, 'steps': 2215, 'loss/train': 2.819966197013855} 01/27/2022 21:38:20 - INFO - codeparrot_training - Step 2216: {'lr': 0.0004999750179799383, 'samples': 425664, 'steps': 2216, 'loss/train': 3.4332776069641113} 01/27/2022 21:38:25 - INFO - codeparrot_training - Step 2217: {'lr': 0.0004999747861333838, 'samples': 425856, 'steps': 2217, 'loss/train': 3.2841514348983765} 01/27/2022 21:38:29 - INFO - codeparrot_training - Step 2218: {'lr': 0.0004999745532160164, 'samples': 426048, 'steps': 2218, 'loss/train': 3.1857651472091675} 01/27/2022 21:38:33 - INFO - codeparrot_training - Step 2219: {'lr': 0.0004999743192278377, 'samples': 426240, 'steps': 2219, 'loss/train': 3.32416033744812} 01/27/2022 21:38:37 - INFO - codeparrot_training - Step 2220: {'lr': 0.0004999740841688481, 'samples': 426432, 'steps': 2220, 'loss/train': 2.687960386276245} 01/27/2022 21:38:42 - INFO - codeparrot_training - Step 2221: {'lr': 0.000499973848039049, 'samples': 426624, 'steps': 2221, 'loss/train': 4.315768361091614} 01/27/2022 21:38:47 - INFO - codeparrot_training - Step 2222: {'lr': 0.0004999736108384414, 'samples': 426816, 'steps': 2222, 'loss/train': 3.0894556045532227} 01/27/2022 21:38:51 - INFO - codeparrot_training - Step 2223: {'lr': 0.0004999733725670261, 'samples': 427008, 'steps': 2223, 'loss/train': 3.4190797805786133} 01/27/2022 21:38:55 - INFO - codeparrot_training - Step 2224: {'lr': 0.0004999731332248044, 'samples': 427200, 'steps': 2224, 'loss/train': 5.886992454528809} 01/27/2022 21:38:59 - INFO - codeparrot_training - Step 2225: {'lr': 0.0004999728928117771, 'samples': 427392, 'steps': 2225, 'loss/train': 2.758186161518097} 01/27/2022 21:39:05 - INFO - codeparrot_training - Step 2226: {'lr': 0.0004999726513279452, 'samples': 427584, 'steps': 2226, 'loss/train': 3.2254199981689453} 01/27/2022 21:39:09 - INFO - codeparrot_training - Step 2227: {'lr': 0.0004999724087733099, 'samples': 427776, 'steps': 2227, 'loss/train': 4.893525838851929} 01/27/2022 21:39:14 - INFO - codeparrot_training - Step 2228: {'lr': 0.0004999721651478723, 'samples': 427968, 'steps': 2228, 'loss/train': 4.315216541290283} 01/27/2022 21:39:18 - INFO - codeparrot_training - Step 2229: {'lr': 0.0004999719204516332, 'samples': 428160, 'steps': 2229, 'loss/train': 4.242341637611389} 01/27/2022 21:39:22 - INFO - codeparrot_training - Step 2230: {'lr': 0.0004999716746845937, 'samples': 428352, 'steps': 2230, 'loss/train': 3.5227521657943726} 01/27/2022 21:39:27 - INFO - codeparrot_training - Step 2231: {'lr': 0.0004999714278467551, 'samples': 428544, 'steps': 2231, 'loss/train': 4.605028867721558} 01/27/2022 21:39:32 - INFO - codeparrot_training - Step 2232: {'lr': 0.0004999711799381181, 'samples': 428736, 'steps': 2232, 'loss/train': 5.434418320655823} 01/27/2022 21:39:36 - INFO - codeparrot_training - Step 2233: {'lr': 0.000499970930958684, 'samples': 428928, 'steps': 2233, 'loss/train': 2.197646677494049} 01/27/2022 21:39:40 - INFO - codeparrot_training - Step 2234: {'lr': 0.0004999706809084538, 'samples': 429120, 'steps': 2234, 'loss/train': 4.3856213092803955} 01/27/2022 21:39:44 - INFO - codeparrot_training - Step 2235: {'lr': 0.0004999704297874287, 'samples': 429312, 'steps': 2235, 'loss/train': 4.406507849693298} 01/27/2022 21:39:50 - INFO - codeparrot_training - Step 2236: {'lr': 0.0004999701775956095, 'samples': 429504, 'steps': 2236, 'loss/train': 4.48380446434021} 01/27/2022 21:39:55 - INFO - codeparrot_training - Step 2237: {'lr': 0.0004999699243329975, 'samples': 429696, 'steps': 2237, 'loss/train': 4.110491037368774} 01/27/2022 21:39:59 - INFO - codeparrot_training - Step 2238: {'lr': 0.0004999696699995937, 'samples': 429888, 'steps': 2238, 'loss/train': 3.833233952522278} 01/27/2022 21:40:03 - INFO - codeparrot_training - Step 2239: {'lr': 0.0004999694145953992, 'samples': 430080, 'steps': 2239, 'loss/train': 4.385096311569214} 01/27/2022 21:40:07 - INFO - codeparrot_training - Step 2240: {'lr': 0.0004999691581204152, 'samples': 430272, 'steps': 2240, 'loss/train': 4.61331582069397} 01/27/2022 21:40:12 - INFO - codeparrot_training - Step 2241: {'lr': 0.0004999689005746426, 'samples': 430464, 'steps': 2241, 'loss/train': 3.8344792127609253} 01/27/2022 21:40:17 - INFO - codeparrot_training - Step 2242: {'lr': 0.0004999686419580827, 'samples': 430656, 'steps': 2242, 'loss/train': 4.537434339523315} 01/27/2022 21:40:22 - INFO - codeparrot_training - Step 2243: {'lr': 0.0004999683822707364, 'samples': 430848, 'steps': 2243, 'loss/train': 4.554895520210266} 01/27/2022 21:40:26 - INFO - codeparrot_training - Step 2244: {'lr': 0.0004999681215126049, 'samples': 431040, 'steps': 2244, 'loss/train': 3.915573835372925} 01/27/2022 21:40:30 - INFO - codeparrot_training - Step 2245: {'lr': 0.0004999678596836894, 'samples': 431232, 'steps': 2245, 'loss/train': 3.9647079706192017} 01/27/2022 21:40:34 - INFO - codeparrot_training - Step 2246: {'lr': 0.000499967596783991, 'samples': 431424, 'steps': 2246, 'loss/train': 2.395766258239746} 01/27/2022 21:40:40 - INFO - codeparrot_training - Step 2247: {'lr': 0.0004999673328135107, 'samples': 431616, 'steps': 2247, 'loss/train': 3.3133753538131714} 01/27/2022 21:40:44 - INFO - codeparrot_training - Step 2248: {'lr': 0.0004999670677722498, 'samples': 431808, 'steps': 2248, 'loss/train': 3.5399140119552612} 01/27/2022 21:40:48 - INFO - codeparrot_training - Step 2249: {'lr': 0.0004999668016602094, 'samples': 432000, 'steps': 2249, 'loss/train': 4.6503167152404785} 01/27/2022 21:40:52 - INFO - codeparrot_training - Step 2250: {'lr': 0.0004999665344773905, 'samples': 432192, 'steps': 2250, 'loss/train': 4.108303070068359} 01/27/2022 21:40:56 - INFO - codeparrot_training - Step 2251: {'lr': 0.0004999662662237943, 'samples': 432384, 'steps': 2251, 'loss/train': 3.100386142730713} 01/27/2022 21:41:02 - INFO - codeparrot_training - Step 2252: {'lr': 0.0004999659968994221, 'samples': 432576, 'steps': 2252, 'loss/train': 4.292020082473755} 01/27/2022 21:41:06 - INFO - codeparrot_training - Step 2253: {'lr': 0.0004999657265042748, 'samples': 432768, 'steps': 2253, 'loss/train': 3.8331745862960815} 01/27/2022 21:41:10 - INFO - codeparrot_training - Step 2254: {'lr': 0.0004999654550383539, 'samples': 432960, 'steps': 2254, 'loss/train': 3.1791422367095947} 01/27/2022 21:41:15 - INFO - codeparrot_training - Step 2255: {'lr': 0.0004999651825016603, 'samples': 433152, 'steps': 2255, 'loss/train': 4.2833744287490845} 01/27/2022 21:41:19 - INFO - codeparrot_training - Step 2256: {'lr': 0.0004999649088941951, 'samples': 433344, 'steps': 2256, 'loss/train': 5.396461129188538} 01/27/2022 21:41:25 - INFO - codeparrot_training - Step 2257: {'lr': 0.0004999646342159597, 'samples': 433536, 'steps': 2257, 'loss/train': 3.9162561893463135} 01/27/2022 21:41:29 - INFO - codeparrot_training - Step 2258: {'lr': 0.0004999643584669552, 'samples': 433728, 'steps': 2258, 'loss/train': 4.277682423591614} 01/27/2022 21:41:33 - INFO - codeparrot_training - Step 2259: {'lr': 0.0004999640816471827, 'samples': 433920, 'steps': 2259, 'loss/train': 4.237182855606079} 01/27/2022 21:41:37 - INFO - codeparrot_training - Step 2260: {'lr': 0.0004999638037566436, 'samples': 434112, 'steps': 2260, 'loss/train': 4.340367794036865} 01/27/2022 21:41:41 - INFO - codeparrot_training - Step 2261: {'lr': 0.0004999635247953387, 'samples': 434304, 'steps': 2261, 'loss/train': 3.99619460105896} 01/27/2022 21:41:47 - INFO - codeparrot_training - Step 2262: {'lr': 0.0004999632447632696, 'samples': 434496, 'steps': 2262, 'loss/train': 3.954289197921753} 01/27/2022 21:41:51 - INFO - codeparrot_training - Step 2263: {'lr': 0.0004999629636604372, 'samples': 434688, 'steps': 2263, 'loss/train': 4.621082425117493} 01/27/2022 21:41:55 - INFO - codeparrot_training - Step 2264: {'lr': 0.0004999626814868429, 'samples': 434880, 'steps': 2264, 'loss/train': 3.540034532546997} 01/27/2022 21:41:59 - INFO - codeparrot_training - Step 2265: {'lr': 0.0004999623982424879, 'samples': 435072, 'steps': 2265, 'loss/train': 5.133227705955505} 01/27/2022 21:42:04 - INFO - codeparrot_training - Step 2266: {'lr': 0.0004999621139273733, 'samples': 435264, 'steps': 2266, 'loss/train': 4.358781337738037} 01/27/2022 21:42:09 - INFO - codeparrot_training - Step 2267: {'lr': 0.0004999618285415004, 'samples': 435456, 'steps': 2267, 'loss/train': 4.226397514343262} 01/27/2022 21:42:13 - INFO - codeparrot_training - Step 2268: {'lr': 0.0004999615420848704, 'samples': 435648, 'steps': 2268, 'loss/train': 4.364398241043091} 01/27/2022 21:42:17 - INFO - codeparrot_training - Step 2269: {'lr': 0.0004999612545574845, 'samples': 435840, 'steps': 2269, 'loss/train': 4.872667193412781} 01/27/2022 21:42:22 - INFO - codeparrot_training - Step 2270: {'lr': 0.000499960965959344, 'samples': 436032, 'steps': 2270, 'loss/train': 4.2168402671813965} 01/27/2022 21:42:26 - INFO - codeparrot_training - Step 2271: {'lr': 0.0004999606762904501, 'samples': 436224, 'steps': 2271, 'loss/train': 4.718954801559448} 01/27/2022 21:42:32 - INFO - codeparrot_training - Step 2272: {'lr': 0.000499960385550804, 'samples': 436416, 'steps': 2272, 'loss/train': 4.129146695137024} 01/27/2022 21:42:36 - INFO - codeparrot_training - Step 2273: {'lr': 0.000499960093740407, 'samples': 436608, 'steps': 2273, 'loss/train': 4.521149754524231} 01/27/2022 21:42:40 - INFO - codeparrot_training - Step 2274: {'lr': 0.0004999598008592603, 'samples': 436800, 'steps': 2274, 'loss/train': 3.8882131576538086} 01/27/2022 21:42:44 - INFO - codeparrot_training - Step 2275: {'lr': 0.0004999595069073653, 'samples': 436992, 'steps': 2275, 'loss/train': 5.096085906028748} 01/27/2022 21:42:49 - INFO - codeparrot_training - Step 2276: {'lr': 0.0004999592118847229, 'samples': 437184, 'steps': 2276, 'loss/train': 3.9069589376449585} 01/27/2022 21:42:53 - INFO - codeparrot_training - Step 2277: {'lr': 0.0004999589157913348, 'samples': 437376, 'steps': 2277, 'loss/train': 3.7076679468154907} 01/27/2022 21:42:59 - INFO - codeparrot_training - Step 2278: {'lr': 0.0004999586186272021, 'samples': 437568, 'steps': 2278, 'loss/train': 4.450511813163757} 01/27/2022 21:43:03 - INFO - codeparrot_training - Step 2279: {'lr': 0.000499958320392326, 'samples': 437760, 'steps': 2279, 'loss/train': 3.341146945953369} 01/27/2022 21:43:07 - INFO - codeparrot_training - Step 2280: {'lr': 0.0004999580210867077, 'samples': 437952, 'steps': 2280, 'loss/train': 3.3035577535629272} 01/27/2022 21:43:12 - INFO - codeparrot_training - Step 2281: {'lr': 0.0004999577207103487, 'samples': 438144, 'steps': 2281, 'loss/train': 2.2278106212615967} 01/27/2022 21:43:17 - INFO - codeparrot_training - Step 2282: {'lr': 0.0004999574192632502, 'samples': 438336, 'steps': 2282, 'loss/train': 4.652247190475464} 01/27/2022 21:43:21 - INFO - codeparrot_training - Step 2283: {'lr': 0.0004999571167454135, 'samples': 438528, 'steps': 2283, 'loss/train': 4.119468569755554} 01/27/2022 21:43:25 - INFO - codeparrot_training - Step 2284: {'lr': 0.0004999568131568399, 'samples': 438720, 'steps': 2284, 'loss/train': 4.363968372344971} 01/27/2022 21:43:30 - INFO - codeparrot_training - Step 2285: {'lr': 0.0004999565084975306, 'samples': 438912, 'steps': 2285, 'loss/train': 5.252726912498474} 01/27/2022 21:43:34 - INFO - codeparrot_training - Step 2286: {'lr': 0.0004999562027674871, 'samples': 439104, 'steps': 2286, 'loss/train': 3.777079939842224} 01/27/2022 21:43:40 - INFO - codeparrot_training - Step 2287: {'lr': 0.0004999558959667105, 'samples': 439296, 'steps': 2287, 'loss/train': 5.156525015830994} 01/27/2022 21:43:44 - INFO - codeparrot_training - Step 2288: {'lr': 0.0004999555880952023, 'samples': 439488, 'steps': 2288, 'loss/train': 4.070854425430298} 01/27/2022 21:43:48 - INFO - codeparrot_training - Step 2289: {'lr': 0.0004999552791529637, 'samples': 439680, 'steps': 2289, 'loss/train': 3.9403356313705444} 01/27/2022 21:43:52 - INFO - codeparrot_training - Step 2290: {'lr': 0.000499954969139996, 'samples': 439872, 'steps': 2290, 'loss/train': 5.145655274391174} 01/27/2022 21:43:57 - INFO - codeparrot_training - Step 2291: {'lr': 0.0004999546580563006, 'samples': 440064, 'steps': 2291, 'loss/train': 4.152155756950378} 01/27/2022 21:44:02 - INFO - codeparrot_training - Step 2292: {'lr': 0.0004999543459018788, 'samples': 440256, 'steps': 2292, 'loss/train': 4.563095569610596} 01/27/2022 21:44:06 - INFO - codeparrot_training - Step 2293: {'lr': 0.000499954032676732, 'samples': 440448, 'steps': 2293, 'loss/train': 5.370435476303101} 01/27/2022 21:44:10 - INFO - codeparrot_training - Step 2294: {'lr': 0.0004999537183808614, 'samples': 440640, 'steps': 2294, 'loss/train': 4.10149884223938} 01/27/2022 21:44:15 - INFO - codeparrot_training - Step 2295: {'lr': 0.0004999534030142686, 'samples': 440832, 'steps': 2295, 'loss/train': 4.5764840841293335} 01/27/2022 21:44:19 - INFO - codeparrot_training - Step 2296: {'lr': 0.0004999530865769547, 'samples': 441024, 'steps': 2296, 'loss/train': 4.686656355857849} 01/27/2022 21:44:24 - INFO - codeparrot_training - Step 2297: {'lr': 0.0004999527690689212, 'samples': 441216, 'steps': 2297, 'loss/train': 3.8411028385162354} 01/27/2022 21:44:28 - INFO - codeparrot_training - Step 2298: {'lr': 0.0004999524504901694, 'samples': 441408, 'steps': 2298, 'loss/train': 3.359358072280884} 01/27/2022 21:44:33 - INFO - codeparrot_training - Step 2299: {'lr': 0.0004999521308407006, 'samples': 441600, 'steps': 2299, 'loss/train': 3.6689411401748657} 01/27/2022 21:44:37 - INFO - codeparrot_training - Step 2300: {'lr': 0.0004999518101205162, 'samples': 441792, 'steps': 2300, 'loss/train': 4.909021496772766} 01/27/2022 21:44:41 - INFO - codeparrot_training - Step 2301: {'lr': 0.0004999514883296176, 'samples': 441984, 'steps': 2301, 'loss/train': 4.669355750083923} 01/27/2022 21:44:47 - INFO - codeparrot_training - Step 2302: {'lr': 0.0004999511654680064, 'samples': 442176, 'steps': 2302, 'loss/train': 5.1595762968063354} 01/27/2022 21:44:51 - INFO - codeparrot_training - Step 2303: {'lr': 0.0004999508415356836, 'samples': 442368, 'steps': 2303, 'loss/train': 3.209192633628845} 01/27/2022 21:44:55 - INFO - codeparrot_training - Step 2304: {'lr': 0.0004999505165326509, 'samples': 442560, 'steps': 2304, 'loss/train': 1.6309452652931213} 01/27/2022 21:45:00 - INFO - codeparrot_training - Step 2305: {'lr': 0.0004999501904589095, 'samples': 442752, 'steps': 2305, 'loss/train': 4.5725566148757935} 01/27/2022 21:45:04 - INFO - codeparrot_training - Step 2306: {'lr': 0.0004999498633144608, 'samples': 442944, 'steps': 2306, 'loss/train': 3.955987572669983} 01/27/2022 21:45:09 - INFO - codeparrot_training - Step 2307: {'lr': 0.0004999495350993062, 'samples': 443136, 'steps': 2307, 'loss/train': 4.770353078842163} 01/27/2022 21:45:13 - INFO - codeparrot_training - Step 2308: {'lr': 0.0004999492058134473, 'samples': 443328, 'steps': 2308, 'loss/train': 1.5116295218467712} 01/27/2022 21:45:18 - INFO - codeparrot_training - Step 2309: {'lr': 0.0004999488754568853, 'samples': 443520, 'steps': 2309, 'loss/train': 4.79030442237854} 01/27/2022 21:45:22 - INFO - codeparrot_training - Step 2310: {'lr': 0.0004999485440296216, 'samples': 443712, 'steps': 2310, 'loss/train': 4.1325541734695435} 01/27/2022 21:45:26 - INFO - codeparrot_training - Step 2311: {'lr': 0.0004999482115316579, 'samples': 443904, 'steps': 2311, 'loss/train': 4.544613003730774} 01/27/2022 21:45:31 - INFO - codeparrot_training - Step 2312: {'lr': 0.0004999478779629953, 'samples': 444096, 'steps': 2312, 'loss/train': 3.7796047925949097} 01/27/2022 21:45:36 - INFO - codeparrot_training - Step 2313: {'lr': 0.0004999475433236354, 'samples': 444288, 'steps': 2313, 'loss/train': 5.734375476837158} 01/27/2022 21:45:40 - INFO - codeparrot_training - Step 2314: {'lr': 0.0004999472076135796, 'samples': 444480, 'steps': 2314, 'loss/train': 3.8440611362457275} 01/27/2022 21:45:44 - INFO - codeparrot_training - Step 2315: {'lr': 0.0004999468708328293, 'samples': 444672, 'steps': 2315, 'loss/train': 3.5759085416793823} 01/27/2022 21:45:48 - INFO - codeparrot_training - Step 2316: {'lr': 0.0004999465329813859, 'samples': 444864, 'steps': 2316, 'loss/train': 3.7775691747665405} 01/27/2022 21:45:53 - INFO - codeparrot_training - Step 2317: {'lr': 0.000499946194059251, 'samples': 445056, 'steps': 2317, 'loss/train': 4.263438105583191} 01/27/2022 21:45:58 - INFO - codeparrot_training - Step 2318: {'lr': 0.000499945854066426, 'samples': 445248, 'steps': 2318, 'loss/train': 3.8824506998062134} 01/27/2022 21:46:02 - INFO - codeparrot_training - Step 2319: {'lr': 0.0004999455130029123, 'samples': 445440, 'steps': 2319, 'loss/train': 2.856826901435852} 01/27/2022 21:46:06 - INFO - codeparrot_training - Step 2320: {'lr': 0.0004999451708687113, 'samples': 445632, 'steps': 2320, 'loss/train': 3.576704978942871} 01/27/2022 21:46:11 - INFO - codeparrot_training - Step 2321: {'lr': 0.0004999448276638247, 'samples': 445824, 'steps': 2321, 'loss/train': 0.45023949444293976} 01/27/2022 21:46:15 - INFO - codeparrot_training - Step 2322: {'lr': 0.0004999444833882538, 'samples': 446016, 'steps': 2322, 'loss/train': 4.177359938621521} 01/27/2022 21:46:21 - INFO - codeparrot_training - Step 2323: {'lr': 0.000499944138042, 'samples': 446208, 'steps': 2323, 'loss/train': 4.97665536403656} 01/27/2022 21:46:25 - INFO - codeparrot_training - Step 2324: {'lr': 0.000499943791625065, 'samples': 446400, 'steps': 2324, 'loss/train': 3.9742623567581177} 01/27/2022 21:46:29 - INFO - codeparrot_training - Step 2325: {'lr': 0.0004999434441374501, 'samples': 446592, 'steps': 2325, 'loss/train': 3.4932303428649902} 01/27/2022 21:46:34 - INFO - codeparrot_training - Step 2326: {'lr': 0.0004999430955791569, 'samples': 446784, 'steps': 2326, 'loss/train': 3.732313871383667} 01/27/2022 21:46:39 - INFO - codeparrot_training - Step 2327: {'lr': 0.0004999427459501868, 'samples': 446976, 'steps': 2327, 'loss/train': 4.555661201477051} 01/27/2022 21:46:43 - INFO - codeparrot_training - Step 2328: {'lr': 0.0004999423952505414, 'samples': 447168, 'steps': 2328, 'loss/train': 3.5799590349197388} 01/27/2022 21:46:47 - INFO - codeparrot_training - Step 2329: {'lr': 0.000499942043480222, 'samples': 447360, 'steps': 2329, 'loss/train': 3.3371862173080444} 01/27/2022 21:46:52 - INFO - codeparrot_training - Step 2330: {'lr': 0.0004999416906392303, 'samples': 447552, 'steps': 2330, 'loss/train': 5.070952892303467} 01/27/2022 21:46:56 - INFO - codeparrot_training - Step 2331: {'lr': 0.0004999413367275678, 'samples': 447744, 'steps': 2331, 'loss/train': 4.285721182823181} 01/27/2022 21:47:00 - INFO - codeparrot_training - Step 2332: {'lr': 0.000499940981745236, 'samples': 447936, 'steps': 2332, 'loss/train': 4.329686522483826} 01/27/2022 21:47:06 - INFO - codeparrot_training - Step 2333: {'lr': 0.0004999406256922365, 'samples': 448128, 'steps': 2333, 'loss/train': 4.356409549713135} 01/27/2022 21:47:11 - INFO - codeparrot_training - Step 2334: {'lr': 0.0004999402685685705, 'samples': 448320, 'steps': 2334, 'loss/train': 2.3250816464424133} 01/27/2022 21:47:15 - INFO - codeparrot_training - Step 2335: {'lr': 0.0004999399103742399, 'samples': 448512, 'steps': 2335, 'loss/train': 4.921323537826538} 01/27/2022 21:47:19 - INFO - codeparrot_training - Step 2336: {'lr': 0.000499939551109246, 'samples': 448704, 'steps': 2336, 'loss/train': 4.783559560775757} 01/27/2022 21:47:23 - INFO - codeparrot_training - Step 2337: {'lr': 0.0004999391907735905, 'samples': 448896, 'steps': 2337, 'loss/train': 6.433108806610107} 01/27/2022 21:47:29 - INFO - codeparrot_training - Step 2338: {'lr': 0.0004999388293672748, 'samples': 449088, 'steps': 2338, 'loss/train': 3.4580140113830566} 01/27/2022 21:47:33 - INFO - codeparrot_training - Step 2339: {'lr': 0.0004999384668903006, 'samples': 449280, 'steps': 2339, 'loss/train': 3.4873799085617065} 01/27/2022 21:47:37 - INFO - codeparrot_training - Step 2340: {'lr': 0.0004999381033426693, 'samples': 449472, 'steps': 2340, 'loss/train': 4.5058557987213135} 01/27/2022 21:47:42 - INFO - codeparrot_training - Step 2341: {'lr': 0.0004999377387243827, 'samples': 449664, 'steps': 2341, 'loss/train': 3.450191617012024} 01/27/2022 21:47:48 - INFO - codeparrot_training - Step 2342: {'lr': 0.0004999373730354419, 'samples': 449856, 'steps': 2342, 'loss/train': 5.02070939540863} 01/27/2022 21:47:52 - INFO - codeparrot_training - Step 2343: {'lr': 0.0004999370062758491, 'samples': 450048, 'steps': 2343, 'loss/train': 3.76923930644989} 01/27/2022 21:47:56 - INFO - codeparrot_training - Step 2344: {'lr': 0.0004999366384456052, 'samples': 450240, 'steps': 2344, 'loss/train': 4.175965905189514} 01/27/2022 21:48:00 - INFO - codeparrot_training - Step 2345: {'lr': 0.0004999362695447123, 'samples': 450432, 'steps': 2345, 'loss/train': 2.3904054164886475} 01/27/2022 21:48:05 - INFO - codeparrot_training - Step 2346: {'lr': 0.0004999358995731718, 'samples': 450624, 'steps': 2346, 'loss/train': 3.227553606033325} 01/27/2022 21:48:09 - INFO - codeparrot_training - Step 2347: {'lr': 0.0004999355285309851, 'samples': 450816, 'steps': 2347, 'loss/train': 4.60911762714386} 01/27/2022 21:48:15 - INFO - codeparrot_training - Step 2348: {'lr': 0.0004999351564181541, 'samples': 451008, 'steps': 2348, 'loss/train': 3.475814938545227} 01/27/2022 21:48:19 - INFO - codeparrot_training - Step 2349: {'lr': 0.0004999347832346802, 'samples': 451200, 'steps': 2349, 'loss/train': 4.815190672874451} 01/27/2022 21:48:23 - INFO - codeparrot_training - Step 2350: {'lr': 0.0004999344089805651, 'samples': 451392, 'steps': 2350, 'loss/train': 4.050331950187683} 01/27/2022 21:48:28 - INFO - codeparrot_training - Step 2351: {'lr': 0.0004999340336558104, 'samples': 451584, 'steps': 2351, 'loss/train': 4.562780141830444} 01/27/2022 21:48:32 - INFO - codeparrot_training - Step 2352: {'lr': 0.0004999336572604175, 'samples': 451776, 'steps': 2352, 'loss/train': 3.2937272787094116} 01/27/2022 21:48:37 - INFO - codeparrot_training - Step 2353: {'lr': 0.0004999332797943883, 'samples': 451968, 'steps': 2353, 'loss/train': 4.05780565738678} 01/27/2022 21:48:41 - INFO - codeparrot_training - Step 2354: {'lr': 0.0004999329012577243, 'samples': 452160, 'steps': 2354, 'loss/train': 2.3562183380126953} 01/27/2022 21:48:46 - INFO - codeparrot_training - Step 2355: {'lr': 0.000499932521650427, 'samples': 452352, 'steps': 2355, 'loss/train': 1.3660944700241089} 01/27/2022 21:48:50 - INFO - codeparrot_training - Step 2356: {'lr': 0.0004999321409724982, 'samples': 452544, 'steps': 2356, 'loss/train': 4.338561058044434} 01/27/2022 21:48:54 - INFO - codeparrot_training - Step 2357: {'lr': 0.0004999317592239395, 'samples': 452736, 'steps': 2357, 'loss/train': 3.2962907552719116} 01/27/2022 21:49:00 - INFO - codeparrot_training - Step 2358: {'lr': 0.0004999313764047525, 'samples': 452928, 'steps': 2358, 'loss/train': 4.912365674972534} 01/27/2022 21:49:04 - INFO - codeparrot_training - Step 2359: {'lr': 0.0004999309925149388, 'samples': 453120, 'steps': 2359, 'loss/train': 4.142649292945862} 01/27/2022 21:49:09 - INFO - codeparrot_training - Step 2360: {'lr': 0.0004999306075545002, 'samples': 453312, 'steps': 2360, 'loss/train': 5.436208605766296} 01/27/2022 21:49:13 - INFO - codeparrot_training - Step 2361: {'lr': 0.0004999302215234381, 'samples': 453504, 'steps': 2361, 'loss/train': 4.10815966129303} 01/27/2022 21:49:17 - INFO - codeparrot_training - Step 2362: {'lr': 0.0004999298344217543, 'samples': 453696, 'steps': 2362, 'loss/train': 3.3261154890060425} 01/27/2022 21:49:22 - INFO - codeparrot_training - Step 2363: {'lr': 0.0004999294462494506, 'samples': 453888, 'steps': 2363, 'loss/train': 4.990547060966492} 01/27/2022 21:49:27 - INFO - codeparrot_training - Step 2364: {'lr': 0.0004999290570065284, 'samples': 454080, 'steps': 2364, 'loss/train': 4.8315171003341675} 01/27/2022 21:49:31 - INFO - codeparrot_training - Step 2365: {'lr': 0.0004999286666929895, 'samples': 454272, 'steps': 2365, 'loss/train': 2.505073070526123} 01/27/2022 21:49:35 - INFO - codeparrot_training - Step 2366: {'lr': 0.0004999282753088356, 'samples': 454464, 'steps': 2366, 'loss/train': 3.182561159133911} 01/27/2022 21:49:40 - INFO - codeparrot_training - Step 2367: {'lr': 0.0004999278828540682, 'samples': 454656, 'steps': 2367, 'loss/train': 3.3684149980545044} 01/27/2022 21:49:45 - INFO - codeparrot_training - Step 2368: {'lr': 0.0004999274893286893, 'samples': 454848, 'steps': 2368, 'loss/train': 4.606690764427185} 01/27/2022 21:49:49 - INFO - codeparrot_training - Step 2369: {'lr': 0.0004999270947327003, 'samples': 455040, 'steps': 2369, 'loss/train': 4.47925078868866} 01/27/2022 21:49:54 - INFO - codeparrot_training - Step 2370: {'lr': 0.0004999266990661029, 'samples': 455232, 'steps': 2370, 'loss/train': 2.903558135032654} 01/27/2022 21:49:58 - INFO - codeparrot_training - Step 2371: {'lr': 0.0004999263023288989, 'samples': 455424, 'steps': 2371, 'loss/train': 4.508728623390198} 01/27/2022 21:50:02 - INFO - codeparrot_training - Step 2372: {'lr': 0.0004999259045210901, 'samples': 455616, 'steps': 2372, 'loss/train': 4.042861461639404} 01/27/2022 21:50:07 - INFO - codeparrot_training - Step 2373: {'lr': 0.000499925505642678, 'samples': 455808, 'steps': 2373, 'loss/train': 1.8325037956237793} 01/27/2022 21:50:11 - INFO - codeparrot_training - Step 2374: {'lr': 0.0004999251056936645, 'samples': 456000, 'steps': 2374, 'loss/train': 4.354261994361877} 01/27/2022 21:50:16 - INFO - codeparrot_training - Step 2375: {'lr': 0.000499924704674051, 'samples': 456192, 'steps': 2375, 'loss/train': 4.845682382583618} 01/27/2022 21:50:20 - INFO - codeparrot_training - Step 2376: {'lr': 0.0004999243025838396, 'samples': 456384, 'steps': 2376, 'loss/train': 3.321642279624939} 01/27/2022 21:50:24 - INFO - codeparrot_training - Step 2377: {'lr': 0.0004999238994230318, 'samples': 456576, 'steps': 2377, 'loss/train': 4.564930200576782} 01/27/2022 21:50:30 - INFO - codeparrot_training - Step 2378: {'lr': 0.0004999234951916293, 'samples': 456768, 'steps': 2378, 'loss/train': 3.9340224266052246} 01/27/2022 21:50:34 - INFO - codeparrot_training - Step 2379: {'lr': 0.0004999230898896341, 'samples': 456960, 'steps': 2379, 'loss/train': 3.2132309675216675} 01/27/2022 21:50:39 - INFO - codeparrot_training - Step 2380: {'lr': 0.0004999226835170476, 'samples': 457152, 'steps': 2380, 'loss/train': 2.599451780319214} 01/27/2022 21:50:43 - INFO - codeparrot_training - Step 2381: {'lr': 0.0004999222760738717, 'samples': 457344, 'steps': 2381, 'loss/train': 4.02343225479126} 01/27/2022 21:50:47 - INFO - codeparrot_training - Step 2382: {'lr': 0.0004999218675601081, 'samples': 457536, 'steps': 2382, 'loss/train': 4.476281046867371} 01/27/2022 21:50:51 - INFO - codeparrot_training - Step 2383: {'lr': 0.0004999214579757586, 'samples': 457728, 'steps': 2383, 'loss/train': 2.9488243460655212} 01/27/2022 21:50:57 - INFO - codeparrot_training - Step 2384: {'lr': 0.000499921047320825, 'samples': 457920, 'steps': 2384, 'loss/train': 3.688499093055725} 01/27/2022 21:51:01 - INFO - codeparrot_training - Step 2385: {'lr': 0.000499920635595309, 'samples': 458112, 'steps': 2385, 'loss/train': 3.9697787761688232} 01/27/2022 21:51:05 - INFO - codeparrot_training - Step 2386: {'lr': 0.0004999202227992122, 'samples': 458304, 'steps': 2386, 'loss/train': 4.334954738616943} 01/27/2022 21:51:09 - INFO - codeparrot_training - Step 2387: {'lr': 0.0004999198089325367, 'samples': 458496, 'steps': 2387, 'loss/train': 3.4274107217788696} 01/27/2022 21:51:15 - INFO - codeparrot_training - Step 2388: {'lr': 0.0004999193939952839, 'samples': 458688, 'steps': 2388, 'loss/train': 4.046189904212952} 01/27/2022 21:51:19 - INFO - codeparrot_training - Step 2389: {'lr': 0.000499918977987456, 'samples': 458880, 'steps': 2389, 'loss/train': 3.7755661010742188} 01/27/2022 21:51:23 - INFO - codeparrot_training - Step 2390: {'lr': 0.0004999185609090544, 'samples': 459072, 'steps': 2390, 'loss/train': 3.847935676574707} 01/27/2022 21:51:28 - INFO - codeparrot_training - Step 2391: {'lr': 0.0004999181427600811, 'samples': 459264, 'steps': 2391, 'loss/train': 3.2943402528762817} 01/27/2022 21:51:32 - INFO - codeparrot_training - Step 2392: {'lr': 0.0004999177235405378, 'samples': 459456, 'steps': 2392, 'loss/train': 2.9461246132850647} 01/27/2022 21:51:36 - INFO - codeparrot_training - Step 2393: {'lr': 0.0004999173032504264, 'samples': 459648, 'steps': 2393, 'loss/train': 3.804252505302429} 01/27/2022 21:51:42 - INFO - codeparrot_training - Step 2394: {'lr': 0.0004999168818897486, 'samples': 459840, 'steps': 2394, 'loss/train': 3.0164941549301147} 01/27/2022 21:51:46 - INFO - codeparrot_training - Step 2395: {'lr': 0.0004999164594585062, 'samples': 460032, 'steps': 2395, 'loss/train': 3.2965221405029297} 01/27/2022 21:51:50 - INFO - codeparrot_training - Step 2396: {'lr': 0.0004999160359567011, 'samples': 460224, 'steps': 2396, 'loss/train': 3.7124290466308594} 01/27/2022 21:51:55 - INFO - codeparrot_training - Step 2397: {'lr': 0.000499915611384335, 'samples': 460416, 'steps': 2397, 'loss/train': 3.563544273376465} 01/27/2022 21:51:59 - INFO - codeparrot_training - Step 2398: {'lr': 0.0004999151857414099, 'samples': 460608, 'steps': 2398, 'loss/train': 4.632796883583069} 01/27/2022 21:52:04 - INFO - codeparrot_training - Step 2399: {'lr': 0.0004999147590279273, 'samples': 460800, 'steps': 2399, 'loss/train': 4.094179630279541} 01/27/2022 21:52:08 - INFO - codeparrot_training - Step 2400: {'lr': 0.0004999143312438893, 'samples': 460992, 'steps': 2400, 'loss/train': 3.8708306550979614} 01/27/2022 21:52:13 - INFO - codeparrot_training - Step 2401: {'lr': 0.0004999139023892978, 'samples': 461184, 'steps': 2401, 'loss/train': 4.108440399169922} 01/27/2022 21:52:17 - INFO - codeparrot_training - Step 2402: {'lr': 0.0004999134724641543, 'samples': 461376, 'steps': 2402, 'loss/train': 3.6613558530807495} 01/27/2022 21:52:21 - INFO - codeparrot_training - Step 2403: {'lr': 0.000499913041468461, 'samples': 461568, 'steps': 2403, 'loss/train': 4.253682732582092} 01/27/2022 21:52:27 - INFO - codeparrot_training - Step 2404: {'lr': 0.0004999126094022195, 'samples': 461760, 'steps': 2404, 'loss/train': 4.821348309516907} 01/27/2022 21:52:31 - INFO - codeparrot_training - Step 2405: {'lr': 0.0004999121762654318, 'samples': 461952, 'steps': 2405, 'loss/train': 4.628625512123108} 01/27/2022 21:52:35 - INFO - codeparrot_training - Step 2406: {'lr': 0.0004999117420580996, 'samples': 462144, 'steps': 2406, 'loss/train': 4.2453285455703735} 01/27/2022 21:52:40 - INFO - codeparrot_training - Step 2407: {'lr': 0.0004999113067802249, 'samples': 462336, 'steps': 2407, 'loss/train': 4.648600459098816} 01/27/2022 21:52:44 - INFO - codeparrot_training - Step 2408: {'lr': 0.0004999108704318095, 'samples': 462528, 'steps': 2408, 'loss/train': 4.216181516647339} 01/27/2022 21:52:49 - INFO - codeparrot_training - Step 2409: {'lr': 0.0004999104330128553, 'samples': 462720, 'steps': 2409, 'loss/train': 1.6729423999786377} 01/27/2022 21:52:53 - INFO - codeparrot_training - Step 2410: {'lr': 0.0004999099945233641, 'samples': 462912, 'steps': 2410, 'loss/train': 3.226974606513977} 01/27/2022 21:52:58 - INFO - codeparrot_training - Step 2411: {'lr': 0.000499909554963338, 'samples': 463104, 'steps': 2411, 'loss/train': 4.7848920822143555} 01/27/2022 21:53:02 - INFO - codeparrot_training - Step 2412: {'lr': 0.0004999091143327786, 'samples': 463296, 'steps': 2412, 'loss/train': 5.150660276412964} 01/27/2022 21:53:07 - INFO - codeparrot_training - Step 2413: {'lr': 0.000499908672631688, 'samples': 463488, 'steps': 2413, 'loss/train': 3.0740586519241333} 01/27/2022 21:53:12 - INFO - codeparrot_training - Step 2414: {'lr': 0.0004999082298600679, 'samples': 463680, 'steps': 2414, 'loss/train': 3.8250564336776733} 01/27/2022 21:53:16 - INFO - codeparrot_training - Step 2415: {'lr': 0.0004999077860179204, 'samples': 463872, 'steps': 2415, 'loss/train': 4.1632150411605835} 01/27/2022 21:53:20 - INFO - codeparrot_training - Step 2416: {'lr': 0.0004999073411052472, 'samples': 464064, 'steps': 2416, 'loss/train': 3.9824284315109253} 01/27/2022 21:53:24 - INFO - codeparrot_training - Step 2417: {'lr': 0.0004999068951220503, 'samples': 464256, 'steps': 2417, 'loss/train': 4.129802227020264} 01/27/2022 21:53:30 - INFO - codeparrot_training - Step 2418: {'lr': 0.0004999064480683317, 'samples': 464448, 'steps': 2418, 'loss/train': 4.264164447784424} 01/27/2022 21:53:35 - INFO - codeparrot_training - Step 2419: {'lr': 0.0004999059999440932, 'samples': 464640, 'steps': 2419, 'loss/train': 4.17215359210968} 01/27/2022 21:53:39 - INFO - codeparrot_training - Step 2420: {'lr': 0.0004999055507493368, 'samples': 464832, 'steps': 2420, 'loss/train': 5.500573754310608} 01/27/2022 21:53:43 - INFO - codeparrot_training - Step 2421: {'lr': 0.0004999051004840642, 'samples': 465024, 'steps': 2421, 'loss/train': 3.6902164220809937} 01/27/2022 21:53:47 - INFO - codeparrot_training - Step 2422: {'lr': 0.0004999046491482777, 'samples': 465216, 'steps': 2422, 'loss/train': 4.315873146057129} 01/27/2022 21:53:53 - INFO - codeparrot_training - Step 2423: {'lr': 0.000499904196741979, 'samples': 465408, 'steps': 2423, 'loss/train': 4.6467365026474} 01/27/2022 21:53:57 - INFO - codeparrot_training - Step 2424: {'lr': 0.00049990374326517, 'samples': 465600, 'steps': 2424, 'loss/train': 2.570704400539398} 01/27/2022 21:54:01 - INFO - codeparrot_training - Step 2425: {'lr': 0.0004999032887178527, 'samples': 465792, 'steps': 2425, 'loss/train': 3.7520452737808228} 01/27/2022 21:54:05 - INFO - codeparrot_training - Step 2426: {'lr': 0.000499902833100029, 'samples': 465984, 'steps': 2426, 'loss/train': 3.6077520847320557} 01/27/2022 21:54:10 - INFO - codeparrot_training - Step 2427: {'lr': 0.0004999023764117011, 'samples': 466176, 'steps': 2427, 'loss/train': 3.520788788795471} 01/27/2022 21:54:15 - INFO - codeparrot_training - Step 2428: {'lr': 0.0004999019186528708, 'samples': 466368, 'steps': 2428, 'loss/train': 4.5310492515563965} 01/27/2022 21:54:19 - INFO - codeparrot_training - Step 2429: {'lr': 0.0004999014598235399, 'samples': 466560, 'steps': 2429, 'loss/train': 2.3430234789848328} 01/27/2022 21:54:23 - INFO - codeparrot_training - Step 2430: {'lr': 0.0004999009999237105, 'samples': 466752, 'steps': 2430, 'loss/train': 4.167592406272888} 01/27/2022 21:54:27 - INFO - codeparrot_training - Step 2431: {'lr': 0.0004999005389533846, 'samples': 466944, 'steps': 2431, 'loss/train': 3.287711262702942} 01/27/2022 21:54:32 - INFO - codeparrot_training - Step 2432: {'lr': 0.0004999000769125642, 'samples': 467136, 'steps': 2432, 'loss/train': 4.68183159828186} 01/27/2022 21:54:37 - INFO - codeparrot_training - Step 2433: {'lr': 0.0004998996138012512, 'samples': 467328, 'steps': 2433, 'loss/train': 3.5711395740509033} 01/27/2022 21:54:41 - INFO - codeparrot_training - Step 2434: {'lr': 0.0004998991496194475, 'samples': 467520, 'steps': 2434, 'loss/train': 4.3896260261535645} 01/27/2022 21:54:45 - INFO - codeparrot_training - Step 2435: {'lr': 0.0004998986843671552, 'samples': 467712, 'steps': 2435, 'loss/train': 4.409571647644043} 01/27/2022 21:54:50 - INFO - codeparrot_training - Step 2436: {'lr': 0.0004998982180443764, 'samples': 467904, 'steps': 2436, 'loss/train': 4.217297315597534} 01/27/2022 21:54:54 - INFO - codeparrot_training - Step 2437: {'lr': 0.000499897750651113, 'samples': 468096, 'steps': 2437, 'loss/train': 4.499282956123352} 01/27/2022 21:55:00 - INFO - codeparrot_training - Step 2438: {'lr': 0.0004998972821873668, 'samples': 468288, 'steps': 2438, 'loss/train': 4.770520806312561} 01/27/2022 21:55:04 - INFO - codeparrot_training - Step 2439: {'lr': 0.0004998968126531402, 'samples': 468480, 'steps': 2439, 'loss/train': 3.7966192960739136} 01/27/2022 21:55:08 - INFO - codeparrot_training - Step 2440: {'lr': 0.0004998963420484349, 'samples': 468672, 'steps': 2440, 'loss/train': 4.676024794578552} 01/27/2022 21:55:13 - INFO - codeparrot_training - Step 2441: {'lr': 0.0004998958703732532, 'samples': 468864, 'steps': 2441, 'loss/train': 3.3853397369384766} 01/27/2022 21:55:17 - INFO - codeparrot_training - Step 2442: {'lr': 0.0004998953976275966, 'samples': 469056, 'steps': 2442, 'loss/train': 4.965538144111633} 01/27/2022 21:55:22 - INFO - codeparrot_training - Step 2443: {'lr': 0.0004998949238114677, 'samples': 469248, 'steps': 2443, 'loss/train': 4.01248562335968} 01/27/2022 21:55:26 - INFO - codeparrot_training - Step 2444: {'lr': 0.0004998944489248683, 'samples': 469440, 'steps': 2444, 'loss/train': 1.8760035037994385} 01/27/2022 21:55:30 - INFO - codeparrot_training - Step 2445: {'lr': 0.0004998939729678004, 'samples': 469632, 'steps': 2445, 'loss/train': 4.395628809928894} 01/27/2022 21:55:35 - INFO - codeparrot_training - Step 2446: {'lr': 0.000499893495940266, 'samples': 469824, 'steps': 2446, 'loss/train': 3.9712321758270264} 01/27/2022 21:55:39 - INFO - codeparrot_training - Step 2447: {'lr': 0.0004998930178422673, 'samples': 470016, 'steps': 2447, 'loss/train': 4.217073082923889} 01/27/2022 21:55:45 - INFO - codeparrot_training - Step 2448: {'lr': 0.0004998925386738062, 'samples': 470208, 'steps': 2448, 'loss/train': 3.3345186710357666} 01/27/2022 21:55:49 - INFO - codeparrot_training - Step 2449: {'lr': 0.0004998920584348849, 'samples': 470400, 'steps': 2449, 'loss/train': 3.322903633117676} 01/27/2022 21:55:53 - INFO - codeparrot_training - Step 2450: {'lr': 0.0004998915771255053, 'samples': 470592, 'steps': 2450, 'loss/train': 2.442605495452881} 01/27/2022 21:55:58 - INFO - codeparrot_training - Step 2451: {'lr': 0.0004998910947456696, 'samples': 470784, 'steps': 2451, 'loss/train': 4.806025385856628} 01/27/2022 21:56:02 - INFO - codeparrot_training - Step 2452: {'lr': 0.0004998906112953797, 'samples': 470976, 'steps': 2452, 'loss/train': 3.1916459798812866} 01/27/2022 21:56:08 - INFO - codeparrot_training - Step 2453: {'lr': 0.0004998901267746379, 'samples': 471168, 'steps': 2453, 'loss/train': 4.781041145324707} 01/27/2022 21:56:12 - INFO - codeparrot_training - Step 2454: {'lr': 0.0004998896411834461, 'samples': 471360, 'steps': 2454, 'loss/train': 3.0455267429351807} 01/27/2022 21:56:16 - INFO - codeparrot_training - Step 2455: {'lr': 0.0004998891545218063, 'samples': 471552, 'steps': 2455, 'loss/train': 3.059239625930786} 01/27/2022 21:56:20 - INFO - codeparrot_training - Step 2456: {'lr': 0.0004998886667897209, 'samples': 471744, 'steps': 2456, 'loss/train': 4.288015007972717} 01/27/2022 21:56:25 - INFO - codeparrot_training - Step 2457: {'lr': 0.0004998881779871917, 'samples': 471936, 'steps': 2457, 'loss/train': 4.515865087509155} 01/27/2022 21:56:29 - INFO - codeparrot_training - Step 2458: {'lr': 0.0004998876881142208, 'samples': 472128, 'steps': 2458, 'loss/train': 4.69433319568634} 01/27/2022 21:56:34 - INFO - codeparrot_training - Step 2459: {'lr': 0.0004998871971708106, 'samples': 472320, 'steps': 2459, 'loss/train': 4.123122453689575} 01/27/2022 21:56:39 - INFO - codeparrot_training - Step 2460: {'lr': 0.0004998867051569627, 'samples': 472512, 'steps': 2460, 'loss/train': 5.067708849906921} 01/27/2022 21:56:43 - INFO - codeparrot_training - Step 2461: {'lr': 0.0004998862120726798, 'samples': 472704, 'steps': 2461, 'loss/train': 3.778199315071106} 01/27/2022 21:56:47 - INFO - codeparrot_training - Step 2462: {'lr': 0.0004998857179179636, 'samples': 472896, 'steps': 2462, 'loss/train': 3.850142240524292} 01/27/2022 21:56:51 - INFO - codeparrot_training - Step 2463: {'lr': 0.0004998852226928164, 'samples': 473088, 'steps': 2463, 'loss/train': 6.907233953475952} 01/27/2022 21:56:57 - INFO - codeparrot_training - Step 2464: {'lr': 0.0004998847263972401, 'samples': 473280, 'steps': 2464, 'loss/train': 4.568705320358276} 01/27/2022 21:57:01 - INFO - codeparrot_training - Step 2465: {'lr': 0.0004998842290312371, 'samples': 473472, 'steps': 2465, 'loss/train': 3.068029761314392} 01/27/2022 21:57:06 - INFO - codeparrot_training - Step 2466: {'lr': 0.0004998837305948094, 'samples': 473664, 'steps': 2466, 'loss/train': 5.072169542312622} 01/27/2022 21:57:10 - INFO - codeparrot_training - Step 2467: {'lr': 0.0004998832310879591, 'samples': 473856, 'steps': 2467, 'loss/train': 1.3504514694213867} 01/27/2022 21:57:14 - INFO - codeparrot_training - Step 2468: {'lr': 0.0004998827305106884, 'samples': 474048, 'steps': 2468, 'loss/train': 3.279212236404419} 01/27/2022 21:57:19 - INFO - codeparrot_training - Step 2469: {'lr': 0.0004998822288629995, 'samples': 474240, 'steps': 2469, 'loss/train': 4.057498097419739} 01/27/2022 21:57:24 - INFO - codeparrot_training - Step 2470: {'lr': 0.0004998817261448943, 'samples': 474432, 'steps': 2470, 'loss/train': 4.376158833503723} 01/27/2022 21:57:28 - INFO - codeparrot_training - Step 2471: {'lr': 0.0004998812223563754, 'samples': 474624, 'steps': 2471, 'loss/train': 4.363853931427002} 01/27/2022 21:57:32 - INFO - codeparrot_training - Step 2472: {'lr': 0.0004998807174974445, 'samples': 474816, 'steps': 2472, 'loss/train': 4.308769941329956} 01/27/2022 21:57:36 - INFO - codeparrot_training - Step 2473: {'lr': 0.0004998802115681039, 'samples': 475008, 'steps': 2473, 'loss/train': 2.5063183307647705} 01/27/2022 21:57:41 - INFO - codeparrot_training - Step 2474: {'lr': 0.000499879704568356, 'samples': 475200, 'steps': 2474, 'loss/train': 3.968892216682434} 01/27/2022 21:57:46 - INFO - codeparrot_training - Step 2475: {'lr': 0.0004998791964982026, 'samples': 475392, 'steps': 2475, 'loss/train': 4.156682252883911} 01/27/2022 21:57:50 - INFO - codeparrot_training - Step 2476: {'lr': 0.0004998786873576462, 'samples': 475584, 'steps': 2476, 'loss/train': 4.436222076416016} 01/27/2022 21:57:54 - INFO - codeparrot_training - Step 2477: {'lr': 0.0004998781771466889, 'samples': 475776, 'steps': 2477, 'loss/train': 3.581512212753296} 01/27/2022 21:57:58 - INFO - codeparrot_training - Step 2478: {'lr': 0.0004998776658653327, 'samples': 475968, 'steps': 2478, 'loss/train': 2.9615002870559692} 01/27/2022 21:58:04 - INFO - codeparrot_training - Step 2479: {'lr': 0.00049987715351358, 'samples': 476160, 'steps': 2479, 'loss/train': 3.1163949966430664} 01/27/2022 21:58:09 - INFO - codeparrot_training - Step 2480: {'lr': 0.0004998766400914329, 'samples': 476352, 'steps': 2480, 'loss/train': 6.358911037445068} 01/27/2022 21:58:13 - INFO - codeparrot_training - Step 2481: {'lr': 0.0004998761255988936, 'samples': 476544, 'steps': 2481, 'loss/train': 2.8623653054237366} 01/27/2022 21:58:17 - INFO - codeparrot_training - Step 2482: {'lr': 0.0004998756100359643, 'samples': 476736, 'steps': 2482, 'loss/train': 2.8394962549209595} 01/27/2022 21:58:21 - INFO - codeparrot_training - Step 2483: {'lr': 0.0004998750934026474, 'samples': 476928, 'steps': 2483, 'loss/train': 3.8055925369262695} 01/27/2022 21:58:27 - INFO - codeparrot_training - Step 2484: {'lr': 0.0004998745756989448, 'samples': 477120, 'steps': 2484, 'loss/train': 3.6500991582870483} 01/27/2022 21:58:31 - INFO - codeparrot_training - Step 2485: {'lr': 0.0004998740569248588, 'samples': 477312, 'steps': 2485, 'loss/train': 4.0976139307022095} 01/27/2022 21:58:35 - INFO - codeparrot_training - Step 2486: {'lr': 0.0004998735370803917, 'samples': 477504, 'steps': 2486, 'loss/train': 3.6242541074752808} 01/27/2022 21:58:39 - INFO - codeparrot_training - Step 2487: {'lr': 0.0004998730161655459, 'samples': 477696, 'steps': 2487, 'loss/train': 2.6186063289642334} 01/27/2022 21:58:44 - INFO - codeparrot_training - Step 2488: {'lr': 0.0004998724941803232, 'samples': 477888, 'steps': 2488, 'loss/train': 2.3333181738853455} 01/27/2022 21:58:49 - INFO - codeparrot_training - Step 2489: {'lr': 0.0004998719711247262, 'samples': 478080, 'steps': 2489, 'loss/train': 4.076439142227173} 01/27/2022 21:58:54 - INFO - codeparrot_training - Step 2490: {'lr': 0.0004998714469987571, 'samples': 478272, 'steps': 2490, 'loss/train': 3.8411089181900024} 01/27/2022 21:58:58 - INFO - codeparrot_training - Step 2491: {'lr': 0.000499870921802418, 'samples': 478464, 'steps': 2491, 'loss/train': 2.5435039401054382} 01/27/2022 21:59:02 - INFO - codeparrot_training - Step 2492: {'lr': 0.0004998703955357111, 'samples': 478656, 'steps': 2492, 'loss/train': 3.779982805252075} 01/27/2022 21:59:06 - INFO - codeparrot_training - Step 2493: {'lr': 0.0004998698681986389, 'samples': 478848, 'steps': 2493, 'loss/train': 4.640430808067322} 01/27/2022 21:59:11 - INFO - codeparrot_training - Step 2494: {'lr': 0.0004998693397912034, 'samples': 479040, 'steps': 2494, 'loss/train': 4.658910870552063} 01/27/2022 21:59:16 - INFO - codeparrot_training - Step 2495: {'lr': 0.0004998688103134072, 'samples': 479232, 'steps': 2495, 'loss/train': 4.104044437408447} 01/27/2022 21:59:20 - INFO - codeparrot_training - Step 2496: {'lr': 0.0004998682797652522, 'samples': 479424, 'steps': 2496, 'loss/train': 3.8574557304382324} 01/27/2022 21:59:24 - INFO - codeparrot_training - Step 2497: {'lr': 0.0004998677481467408, 'samples': 479616, 'steps': 2497, 'loss/train': 4.205096483230591} 01/27/2022 21:59:29 - INFO - codeparrot_training - Step 2498: {'lr': 0.0004998672154578754, 'samples': 479808, 'steps': 2498, 'loss/train': 4.375722885131836} 01/27/2022 21:59:33 - INFO - codeparrot_training - Step 2499: {'lr': 0.0004998666816986582, 'samples': 480000, 'steps': 2499, 'loss/train': 2.1045793890953064} 01/27/2022 21:59:39 - INFO - codeparrot_training - Step 2500: {'lr': 0.0004998661468690914, 'samples': 480192, 'steps': 2500, 'loss/train': 0.7901655435562134} 01/27/2022 21:59:43 - INFO - codeparrot_training - Step 2501: {'lr': 0.0004998656109691774, 'samples': 480384, 'steps': 2501, 'loss/train': 10.913561582565308} 01/27/2022 21:59:47 - INFO - codeparrot_training - Step 2502: {'lr': 0.0004998650739989185, 'samples': 480576, 'steps': 2502, 'loss/train': 4.127756953239441} 01/27/2022 21:59:51 - INFO - codeparrot_training - Step 2503: {'lr': 0.0004998645359583169, 'samples': 480768, 'steps': 2503, 'loss/train': 1.707682192325592} 01/27/2022 21:59:56 - INFO - codeparrot_training - Step 2504: {'lr': 0.0004998639968473751, 'samples': 480960, 'steps': 2504, 'loss/train': 3.4201841354370117} 01/27/2022 22:00:01 - INFO - codeparrot_training - Step 2505: {'lr': 0.0004998634566660952, 'samples': 481152, 'steps': 2505, 'loss/train': 1.9655376076698303} 01/27/2022 22:00:05 - INFO - codeparrot_training - Step 2506: {'lr': 0.0004998629154144795, 'samples': 481344, 'steps': 2506, 'loss/train': 3.6694071292877197} 01/27/2022 22:00:09 - INFO - codeparrot_training - Step 2507: {'lr': 0.0004998623730925305, 'samples': 481536, 'steps': 2507, 'loss/train': 2.5627350211143494} 01/27/2022 22:00:13 - INFO - codeparrot_training - Step 2508: {'lr': 0.0004998618297002504, 'samples': 481728, 'steps': 2508, 'loss/train': 4.185781359672546} 01/27/2022 22:00:18 - INFO - codeparrot_training - Step 2509: {'lr': 0.0004998612852376417, 'samples': 481920, 'steps': 2509, 'loss/train': 3.8090046644210815} 01/27/2022 22:00:24 - INFO - codeparrot_training - Step 2510: {'lr': 0.0004998607397047063, 'samples': 482112, 'steps': 2510, 'loss/train': 5.317846655845642} 01/27/2022 22:00:28 - INFO - codeparrot_training - Step 2511: {'lr': 0.0004998601931014471, 'samples': 482304, 'steps': 2511, 'loss/train': 3.5789759159088135} 01/27/2022 22:00:32 - INFO - codeparrot_training - Step 2512: {'lr': 0.0004998596454278661, 'samples': 482496, 'steps': 2512, 'loss/train': 3.0881956815719604} 01/27/2022 22:00:36 - INFO - codeparrot_training - Step 2513: {'lr': 0.0004998590966839657, 'samples': 482688, 'steps': 2513, 'loss/train': 3.9171230792999268} 01/27/2022 22:00:40 - INFO - codeparrot_training - Step 2514: {'lr': 0.0004998585468697482, 'samples': 482880, 'steps': 2514, 'loss/train': 3.58411180973053} 01/27/2022 22:00:46 - INFO - codeparrot_training - Step 2515: {'lr': 0.0004998579959852161, 'samples': 483072, 'steps': 2515, 'loss/train': 3.0324625968933105} 01/27/2022 22:00:50 - INFO - codeparrot_training - Step 2516: {'lr': 0.0004998574440303718, 'samples': 483264, 'steps': 2516, 'loss/train': 4.055911660194397} 01/27/2022 22:00:54 - INFO - codeparrot_training - Step 2517: {'lr': 0.0004998568910052173, 'samples': 483456, 'steps': 2517, 'loss/train': 3.103923797607422} 01/27/2022 22:00:58 - INFO - codeparrot_training - Step 2518: {'lr': 0.0004998563369097554, 'samples': 483648, 'steps': 2518, 'loss/train': 3.872380256652832} 01/27/2022 22:01:03 - INFO - codeparrot_training - Step 2519: {'lr': 0.0004998557817439882, 'samples': 483840, 'steps': 2519, 'loss/train': 4.8095011711120605} 01/27/2022 22:01:08 - INFO - codeparrot_training - Step 2520: {'lr': 0.0004998552255079182, 'samples': 484032, 'steps': 2520, 'loss/train': 4.317749977111816} 01/27/2022 22:01:12 - INFO - codeparrot_training - Step 2521: {'lr': 0.0004998546682015478, 'samples': 484224, 'steps': 2521, 'loss/train': 2.8663020730018616} 01/27/2022 22:01:16 - INFO - codeparrot_training - Step 2522: {'lr': 0.0004998541098248793, 'samples': 484416, 'steps': 2522, 'loss/train': 3.8832314014434814} 01/27/2022 22:01:20 - INFO - codeparrot_training - Step 2523: {'lr': 0.0004998535503779151, 'samples': 484608, 'steps': 2523, 'loss/train': 3.4319969415664673} 01/27/2022 22:01:26 - INFO - codeparrot_training - Step 2524: {'lr': 0.0004998529898606576, 'samples': 484800, 'steps': 2524, 'loss/train': 2.2001243233680725} 01/27/2022 22:01:30 - INFO - codeparrot_training - Step 2525: {'lr': 0.0004998524282731093, 'samples': 484992, 'steps': 2525, 'loss/train': 3.8647634983062744} 01/27/2022 22:01:35 - INFO - codeparrot_training - Step 2526: {'lr': 0.0004998518656152725, 'samples': 485184, 'steps': 2526, 'loss/train': 5.295224189758301} 01/27/2022 22:01:39 - INFO - codeparrot_training - Step 2527: {'lr': 0.0004998513018871498, 'samples': 485376, 'steps': 2527, 'loss/train': 3.3020514249801636} 01/27/2022 22:01:43 - INFO - codeparrot_training - Step 2528: {'lr': 0.0004998507370887433, 'samples': 485568, 'steps': 2528, 'loss/train': 7.026519298553467} 01/27/2022 22:01:47 - INFO - codeparrot_training - Step 2529: {'lr': 0.0004998501712200555, 'samples': 485760, 'steps': 2529, 'loss/train': 3.680317997932434} 01/27/2022 22:01:53 - INFO - codeparrot_training - Step 2530: {'lr': 0.000499849604281089, 'samples': 485952, 'steps': 2530, 'loss/train': 3.619542360305786} 01/27/2022 22:01:57 - INFO - codeparrot_training - Step 2531: {'lr': 0.0004998490362718462, 'samples': 486144, 'steps': 2531, 'loss/train': 3.6123558282852173} 01/27/2022 22:02:01 - INFO - codeparrot_training - Step 2532: {'lr': 0.0004998484671923293, 'samples': 486336, 'steps': 2532, 'loss/train': 3.5722296237945557} 01/27/2022 22:02:05 - INFO - codeparrot_training - Step 2533: {'lr': 0.000499847897042541, 'samples': 486528, 'steps': 2533, 'loss/train': 4.17948853969574} 01/27/2022 22:02:10 - INFO - codeparrot_training - Step 2534: {'lr': 0.0004998473258224837, 'samples': 486720, 'steps': 2534, 'loss/train': 3.891942858695984} 01/27/2022 22:02:16 - INFO - codeparrot_training - Step 2535: {'lr': 0.0004998467535321597, 'samples': 486912, 'steps': 2535, 'loss/train': 4.016491413116455} 01/27/2022 22:02:20 - INFO - codeparrot_training - Step 2536: {'lr': 0.0004998461801715716, 'samples': 487104, 'steps': 2536, 'loss/train': 4.18515944480896} 01/27/2022 22:02:25 - INFO - codeparrot_training - Step 2537: {'lr': 0.0004998456057407218, 'samples': 487296, 'steps': 2537, 'loss/train': 4.4984800815582275} 01/27/2022 22:02:29 - INFO - codeparrot_training - Step 2538: {'lr': 0.0004998450302396127, 'samples': 487488, 'steps': 2538, 'loss/train': 3.064322590827942} 01/27/2022 22:02:35 - INFO - codeparrot_training - Step 2539: {'lr': 0.0004998444536682469, 'samples': 487680, 'steps': 2539, 'loss/train': 4.184785008430481} 01/27/2022 22:02:39 - INFO - codeparrot_training - Step 2540: {'lr': 0.0004998438760266267, 'samples': 487872, 'steps': 2540, 'loss/train': 1.1161597967147827} 01/27/2022 22:02:43 - INFO - codeparrot_training - Step 2541: {'lr': 0.0004998432973147548, 'samples': 488064, 'steps': 2541, 'loss/train': 3.717744827270508} 01/27/2022 22:02:47 - INFO - codeparrot_training - Step 2542: {'lr': 0.0004998427175326335, 'samples': 488256, 'steps': 2542, 'loss/train': 3.361113667488098} 01/27/2022 22:02:51 - INFO - codeparrot_training - Step 2543: {'lr': 0.0004998421366802653, 'samples': 488448, 'steps': 2543, 'loss/train': 4.153737545013428} 01/27/2022 22:02:56 - INFO - codeparrot_training - Step 2544: {'lr': 0.0004998415547576527, 'samples': 488640, 'steps': 2544, 'loss/train': 3.4487850666046143} 01/27/2022 22:03:01 - INFO - codeparrot_training - Step 2545: {'lr': 0.0004998409717647983, 'samples': 488832, 'steps': 2545, 'loss/train': 3.696269989013672} 01/27/2022 22:03:05 - INFO - codeparrot_training - Step 2546: {'lr': 0.0004998403877017044, 'samples': 489024, 'steps': 2546, 'loss/train': 2.4057143926620483} 01/27/2022 22:03:09 - INFO - codeparrot_training - Step 2547: {'lr': 0.0004998398025683737, 'samples': 489216, 'steps': 2547, 'loss/train': 4.086504220962524} 01/27/2022 22:03:14 - INFO - codeparrot_training - Step 2548: {'lr': 0.0004998392163648085, 'samples': 489408, 'steps': 2548, 'loss/train': 4.091150164604187} 01/27/2022 22:03:18 - INFO - codeparrot_training - Step 2549: {'lr': 0.0004998386290910116, 'samples': 489600, 'steps': 2549, 'loss/train': 5.545863747596741} 01/27/2022 22:03:23 - INFO - codeparrot_training - Step 2550: {'lr': 0.0004998380407469853, 'samples': 489792, 'steps': 2550, 'loss/train': 4.664059281349182} 01/27/2022 22:03:27 - INFO - codeparrot_training - Step 2551: {'lr': 0.0004998374513327321, 'samples': 489984, 'steps': 2551, 'loss/train': 3.4121822118759155} 01/27/2022 22:03:31 - INFO - codeparrot_training - Step 2552: {'lr': 0.0004998368608482546, 'samples': 490176, 'steps': 2552, 'loss/train': 4.115047574043274} 01/27/2022 22:03:36 - INFO - codeparrot_training - Step 2553: {'lr': 0.0004998362692935553, 'samples': 490368, 'steps': 2553, 'loss/train': 3.6334980726242065} 01/27/2022 22:03:40 - INFO - codeparrot_training - Step 2554: {'lr': 0.0004998356766686368, 'samples': 490560, 'steps': 2554, 'loss/train': 4.242089509963989} 01/27/2022 22:03:46 - INFO - codeparrot_training - Step 2555: {'lr': 0.0004998350829735016, 'samples': 490752, 'steps': 2555, 'loss/train': 4.426425576210022} 01/27/2022 22:03:50 - INFO - codeparrot_training - Step 2556: {'lr': 0.0004998344882081522, 'samples': 490944, 'steps': 2556, 'loss/train': 3.9468719959259033} 01/27/2022 22:03:54 - INFO - codeparrot_training - Step 2557: {'lr': 0.0004998338923725913, 'samples': 491136, 'steps': 2557, 'loss/train': 1.8360502123832703} 01/27/2022 22:03:58 - INFO - codeparrot_training - Step 2558: {'lr': 0.0004998332954668211, 'samples': 491328, 'steps': 2558, 'loss/train': 3.387396454811096} 01/27/2022 22:04:03 - INFO - codeparrot_training - Step 2559: {'lr': 0.0004998326974908446, 'samples': 491520, 'steps': 2559, 'loss/train': 4.49105966091156} 01/27/2022 22:04:08 - INFO - codeparrot_training - Step 2560: {'lr': 0.0004998320984446641, 'samples': 491712, 'steps': 2560, 'loss/train': 3.6842232942581177} 01/27/2022 22:04:12 - INFO - codeparrot_training - Step 2561: {'lr': 0.0004998314983282821, 'samples': 491904, 'steps': 2561, 'loss/train': 4.140486717224121} 01/27/2022 22:04:17 - INFO - codeparrot_training - Step 2562: {'lr': 0.0004998308971417015, 'samples': 492096, 'steps': 2562, 'loss/train': 3.638977289199829} 01/27/2022 22:04:21 - INFO - codeparrot_training - Step 2563: {'lr': 0.0004998302948849246, 'samples': 492288, 'steps': 2563, 'loss/train': 4.24116575717926} 01/27/2022 22:04:25 - INFO - codeparrot_training - Step 2564: {'lr': 0.0004998296915579539, 'samples': 492480, 'steps': 2564, 'loss/train': 1.578011691570282} 01/27/2022 22:04:30 - INFO - codeparrot_training - Step 2565: {'lr': 0.0004998290871607924, 'samples': 492672, 'steps': 2565, 'loss/train': 4.370274782180786} 01/27/2022 22:04:35 - INFO - codeparrot_training - Step 2566: {'lr': 0.0004998284816934422, 'samples': 492864, 'steps': 2566, 'loss/train': 3.3584468364715576} 01/27/2022 22:04:39 - INFO - codeparrot_training - Step 2567: {'lr': 0.0004998278751559062, 'samples': 493056, 'steps': 2567, 'loss/train': 4.586533427238464} 01/27/2022 22:04:43 - INFO - codeparrot_training - Step 2568: {'lr': 0.0004998272675481868, 'samples': 493248, 'steps': 2568, 'loss/train': 3.9003167152404785} 01/27/2022 22:04:47 - INFO - codeparrot_training - Step 2569: {'lr': 0.0004998266588702869, 'samples': 493440, 'steps': 2569, 'loss/train': 2.8321604132652283} 01/27/2022 22:04:53 - INFO - codeparrot_training - Step 2570: {'lr': 0.0004998260491222088, 'samples': 493632, 'steps': 2570, 'loss/train': 4.147387146949768} 01/27/2022 22:04:57 - INFO - codeparrot_training - Step 2571: {'lr': 0.0004998254383039552, 'samples': 493824, 'steps': 2571, 'loss/train': 4.336243987083435} 01/27/2022 22:05:02 - INFO - codeparrot_training - Step 2572: {'lr': 0.0004998248264155288, 'samples': 494016, 'steps': 2572, 'loss/train': 2.6026548743247986} 01/27/2022 22:05:06 - INFO - codeparrot_training - Step 2573: {'lr': 0.0004998242134569322, 'samples': 494208, 'steps': 2573, 'loss/train': 1.26349076628685} 01/27/2022 22:05:10 - INFO - codeparrot_training - Step 2574: {'lr': 0.0004998235994281681, 'samples': 494400, 'steps': 2574, 'loss/train': 4.207420706748962} 01/27/2022 22:05:15 - INFO - codeparrot_training - Step 2575: {'lr': 0.0004998229843292388, 'samples': 494592, 'steps': 2575, 'loss/train': 4.135011434555054} 01/27/2022 22:05:19 - INFO - codeparrot_training - Step 2576: {'lr': 0.0004998223681601474, 'samples': 494784, 'steps': 2576, 'loss/train': 1.6260573863983154} 01/27/2022 22:05:24 - INFO - codeparrot_training - Step 2577: {'lr': 0.0004998217509208961, 'samples': 494976, 'steps': 2577, 'loss/train': 5.142782092094421} 01/27/2022 22:05:28 - INFO - codeparrot_training - Step 2578: {'lr': 0.0004998211326114878, 'samples': 495168, 'steps': 2578, 'loss/train': 4.755898475646973} 01/27/2022 22:05:32 - INFO - codeparrot_training - Step 2579: {'lr': 0.0004998205132319252, 'samples': 495360, 'steps': 2579, 'loss/train': 4.5527583360672} 01/27/2022 22:05:38 - INFO - codeparrot_training - Step 2580: {'lr': 0.0004998198927822108, 'samples': 495552, 'steps': 2580, 'loss/train': 4.439687490463257} 01/27/2022 22:05:42 - INFO - codeparrot_training - Step 2581: {'lr': 0.0004998192712623472, 'samples': 495744, 'steps': 2581, 'loss/train': 3.4460374116897583} 01/27/2022 22:05:47 - INFO - codeparrot_training - Step 2582: {'lr': 0.0004998186486723373, 'samples': 495936, 'steps': 2582, 'loss/train': 4.178836941719055} 01/27/2022 22:05:51 - INFO - codeparrot_training - Step 2583: {'lr': 0.0004998180250121836, 'samples': 496128, 'steps': 2583, 'loss/train': 4.267283320426941} 01/27/2022 22:05:55 - INFO - codeparrot_training - Step 2584: {'lr': 0.0004998174002818887, 'samples': 496320, 'steps': 2584, 'loss/train': 3.0308557748794556} 01/27/2022 22:06:00 - INFO - codeparrot_training - Step 2585: {'lr': 0.0004998167744814555, 'samples': 496512, 'steps': 2585, 'loss/train': 3.034871220588684} 01/27/2022 22:06:04 - INFO - codeparrot_training - Step 2586: {'lr': 0.0004998161476108864, 'samples': 496704, 'steps': 2586, 'loss/train': 2.1062439680099487} 01/27/2022 22:06:09 - INFO - codeparrot_training - Step 2587: {'lr': 0.0004998155196701845, 'samples': 496896, 'steps': 2587, 'loss/train': 3.888341546058655} 01/27/2022 22:06:13 - INFO - codeparrot_training - Step 2588: {'lr': 0.000499814890659352, 'samples': 497088, 'steps': 2588, 'loss/train': 3.3116204738616943} 01/27/2022 22:06:17 - INFO - codeparrot_training - Step 2589: {'lr': 0.000499814260578392, 'samples': 497280, 'steps': 2589, 'loss/train': 5.238370656967163} 01/27/2022 22:06:23 - INFO - codeparrot_training - Step 2590: {'lr': 0.000499813629427307, 'samples': 497472, 'steps': 2590, 'loss/train': 3.171543002128601} 01/27/2022 22:06:27 - INFO - codeparrot_training - Step 2591: {'lr': 0.0004998129972060998, 'samples': 497664, 'steps': 2591, 'loss/train': 3.9244322776794434} 01/27/2022 22:06:31 - INFO - codeparrot_training - Step 2592: {'lr': 0.000499812363914773, 'samples': 497856, 'steps': 2592, 'loss/train': 4.187557339668274} 01/27/2022 22:06:35 - INFO - codeparrot_training - Step 2593: {'lr': 0.0004998117295533292, 'samples': 498048, 'steps': 2593, 'loss/train': 5.390597105026245} 01/27/2022 22:06:40 - INFO - codeparrot_training - Step 2594: {'lr': 0.0004998110941217714, 'samples': 498240, 'steps': 2594, 'loss/train': 4.357524275779724} 01/27/2022 22:06:45 - INFO - codeparrot_training - Step 2595: {'lr': 0.0004998104576201022, 'samples': 498432, 'steps': 2595, 'loss/train': 4.269710898399353} 01/27/2022 22:06:50 - INFO - codeparrot_training - Step 2596: {'lr': 0.0004998098200483243, 'samples': 498624, 'steps': 2596, 'loss/train': 4.336485385894775} 01/27/2022 22:06:54 - INFO - codeparrot_training - Step 2597: {'lr': 0.0004998091814064405, 'samples': 498816, 'steps': 2597, 'loss/train': 3.6743767261505127} 01/27/2022 22:06:58 - INFO - codeparrot_training - Step 2598: {'lr': 0.0004998085416944534, 'samples': 499008, 'steps': 2598, 'loss/train': 4.197411775588989} 01/27/2022 22:07:02 - INFO - codeparrot_training - Step 2599: {'lr': 0.000499807900912366, 'samples': 499200, 'steps': 2599, 'loss/train': 3.6759703159332275} 01/27/2022 22:07:08 - INFO - codeparrot_training - Step 2600: {'lr': 0.0004998072590601808, 'samples': 499392, 'steps': 2600, 'loss/train': 3.382429361343384} 01/27/2022 22:07:12 - INFO - codeparrot_training - Step 2601: {'lr': 0.0004998066161379006, 'samples': 499584, 'steps': 2601, 'loss/train': 5.084058880805969} 01/27/2022 22:07:16 - INFO - codeparrot_training - Step 2602: {'lr': 0.0004998059721455281, 'samples': 499776, 'steps': 2602, 'loss/train': 4.575820326805115} 01/27/2022 22:07:20 - INFO - codeparrot_training - Step 2603: {'lr': 0.0004998053270830662, 'samples': 499968, 'steps': 2603, 'loss/train': 3.461800217628479} 01/27/2022 22:07:24 - INFO - codeparrot_training - Step 2604: {'lr': 0.0004998046809505176, 'samples': 500160, 'steps': 2604, 'loss/train': 5.468542098999023} 01/27/2022 22:07:30 - INFO - codeparrot_training - Step 2605: {'lr': 0.0004998040337478851, 'samples': 500352, 'steps': 2605, 'loss/train': 4.149477481842041} 01/27/2022 22:07:34 - INFO - codeparrot_training - Step 2606: {'lr': 0.0004998033854751715, 'samples': 500544, 'steps': 2606, 'loss/train': 3.72092342376709} 01/27/2022 22:07:38 - INFO - codeparrot_training - Step 2607: {'lr': 0.0004998027361323794, 'samples': 500736, 'steps': 2607, 'loss/train': 3.7759623527526855} 01/27/2022 22:07:42 - INFO - codeparrot_training - Step 2608: {'lr': 0.0004998020857195117, 'samples': 500928, 'steps': 2608, 'loss/train': 3.3530216217041016} 01/27/2022 22:07:46 - INFO - codeparrot_training - Step 2609: {'lr': 0.0004998014342365712, 'samples': 501120, 'steps': 2609, 'loss/train': 1.3033238053321838} 01/27/2022 22:07:52 - INFO - codeparrot_training - Step 2610: {'lr': 0.0004998007816835608, 'samples': 501312, 'steps': 2610, 'loss/train': 4.1862441301345825} 01/27/2022 22:07:56 - INFO - codeparrot_training - Step 2611: {'lr': 0.000499800128060483, 'samples': 501504, 'steps': 2611, 'loss/train': 3.342948317527771} 01/27/2022 22:08:00 - INFO - codeparrot_training - Step 2612: {'lr': 0.0004997994733673409, 'samples': 501696, 'steps': 2612, 'loss/train': 4.2212265729904175} 01/27/2022 22:08:04 - INFO - codeparrot_training - Step 2613: {'lr': 0.000499798817604137, 'samples': 501888, 'steps': 2613, 'loss/train': 4.161245942115784} 01/27/2022 22:08:09 - INFO - codeparrot_training - Step 2614: {'lr': 0.0004997981607708745, 'samples': 502080, 'steps': 2614, 'loss/train': 3.303670048713684} 01/27/2022 22:08:15 - INFO - codeparrot_training - Step 2615: {'lr': 0.0004997975028675558, 'samples': 502272, 'steps': 2615, 'loss/train': 2.642430603504181} 01/27/2022 22:08:19 - INFO - codeparrot_training - Step 2616: {'lr': 0.0004997968438941841, 'samples': 502464, 'steps': 2616, 'loss/train': 4.04421865940094} 01/27/2022 22:08:23 - INFO - codeparrot_training - Step 2617: {'lr': 0.0004997961838507619, 'samples': 502656, 'steps': 2617, 'loss/train': 4.33982527256012} 01/27/2022 22:08:27 - INFO - codeparrot_training - Step 2618: {'lr': 0.0004997955227372923, 'samples': 502848, 'steps': 2618, 'loss/train': 3.0536173582077026} 01/27/2022 22:08:31 - INFO - codeparrot_training - Step 2619: {'lr': 0.000499794860553778, 'samples': 503040, 'steps': 2619, 'loss/train': 4.82123064994812} 01/27/2022 22:08:37 - INFO - codeparrot_training - Step 2620: {'lr': 0.0004997941973002216, 'samples': 503232, 'steps': 2620, 'loss/train': 3.206822991371155} 01/27/2022 22:08:41 - INFO - codeparrot_training - Step 2621: {'lr': 0.0004997935329766265, 'samples': 503424, 'steps': 2621, 'loss/train': 2.3034971952438354} 01/27/2022 22:08:45 - INFO - codeparrot_training - Step 2622: {'lr': 0.000499792867582995, 'samples': 503616, 'steps': 2622, 'loss/train': 4.950564622879028} 01/27/2022 22:08:50 - INFO - codeparrot_training - Step 2623: {'lr': 0.0004997922011193303, 'samples': 503808, 'steps': 2623, 'loss/train': 4.228091597557068} 01/27/2022 22:08:54 - INFO - codeparrot_training - Step 2624: {'lr': 0.000499791533585635, 'samples': 504000, 'steps': 2624, 'loss/train': 3.932371973991394} 01/27/2022 22:08:59 - INFO - codeparrot_training - Step 2625: {'lr': 0.0004997908649819122, 'samples': 504192, 'steps': 2625, 'loss/train': 5.009784936904907} 01/27/2022 22:09:03 - INFO - codeparrot_training - Step 2626: {'lr': 0.0004997901953081646, 'samples': 504384, 'steps': 2626, 'loss/train': 4.088263034820557} 01/27/2022 22:09:08 - INFO - codeparrot_training - Step 2627: {'lr': 0.0004997895245643951, 'samples': 504576, 'steps': 2627, 'loss/train': 4.978622317314148} 01/27/2022 22:09:12 - INFO - codeparrot_training - Step 2628: {'lr': 0.0004997888527506067, 'samples': 504768, 'steps': 2628, 'loss/train': 3.6522445678710938} 01/27/2022 22:09:16 - INFO - codeparrot_training - Step 2629: {'lr': 0.000499788179866802, 'samples': 504960, 'steps': 2629, 'loss/train': 2.1755722761154175} 01/27/2022 22:09:22 - INFO - codeparrot_training - Step 2630: {'lr': 0.0004997875059129843, 'samples': 505152, 'steps': 2630, 'loss/train': 3.5425028800964355} 01/27/2022 22:09:26 - INFO - codeparrot_training - Step 2631: {'lr': 0.000499786830889156, 'samples': 505344, 'steps': 2631, 'loss/train': 4.069260835647583} 01/27/2022 22:09:30 - INFO - codeparrot_training - Step 2632: {'lr': 0.0004997861547953203, 'samples': 505536, 'steps': 2632, 'loss/train': 4.4538856744766235} 01/27/2022 22:09:35 - INFO - codeparrot_training - Step 2633: {'lr': 0.00049978547763148, 'samples': 505728, 'steps': 2633, 'loss/train': 3.6963436603546143} 01/27/2022 22:09:39 - INFO - codeparrot_training - Step 2634: {'lr': 0.0004997847993976381, 'samples': 505920, 'steps': 2634, 'loss/train': 2.547110438346863} 01/27/2022 22:09:44 - INFO - codeparrot_training - Step 2635: {'lr': 0.0004997841200937975, 'samples': 506112, 'steps': 2635, 'loss/train': 3.3031028509140015} 01/27/2022 22:09:48 - INFO - codeparrot_training - Step 2636: {'lr': 0.0004997834397199609, 'samples': 506304, 'steps': 2636, 'loss/train': 5.038434863090515} 01/27/2022 22:09:53 - INFO - codeparrot_training - Step 2637: {'lr': 0.0004997827582761315, 'samples': 506496, 'steps': 2637, 'loss/train': 4.431487083435059} 01/27/2022 22:09:57 - INFO - codeparrot_training - Step 2638: {'lr': 0.0004997820757623119, 'samples': 506688, 'steps': 2638, 'loss/train': 3.572938084602356} 01/27/2022 22:10:01 - INFO - codeparrot_training - Step 2639: {'lr': 0.0004997813921785054, 'samples': 506880, 'steps': 2639, 'loss/train': 3.134887218475342} 01/27/2022 22:10:07 - INFO - codeparrot_training - Step 2640: {'lr': 0.0004997807075247146, 'samples': 507072, 'steps': 2640, 'loss/train': 4.893571615219116} 01/27/2022 22:10:11 - INFO - codeparrot_training - Step 2641: {'lr': 0.0004997800218009426, 'samples': 507264, 'steps': 2641, 'loss/train': 3.4292385578155518} 01/27/2022 22:10:16 - INFO - codeparrot_training - Step 2642: {'lr': 0.0004997793350071923, 'samples': 507456, 'steps': 2642, 'loss/train': 3.556815505027771} 01/27/2022 22:10:20 - INFO - codeparrot_training - Step 2643: {'lr': 0.0004997786471434666, 'samples': 507648, 'steps': 2643, 'loss/train': 3.789656639099121} 01/27/2022 22:10:24 - INFO - codeparrot_training - Step 2644: {'lr': 0.0004997779582097686, 'samples': 507840, 'steps': 2644, 'loss/train': 3.2255219221115112} 01/27/2022 22:10:30 - INFO - codeparrot_training - Step 2645: {'lr': 0.0004997772682061011, 'samples': 508032, 'steps': 2645, 'loss/train': 3.9466195106506348} 01/27/2022 22:10:34 - INFO - codeparrot_training - Step 2646: {'lr': 0.000499776577132467, 'samples': 508224, 'steps': 2646, 'loss/train': 3.2947282791137695} 01/27/2022 22:10:38 - INFO - codeparrot_training - Step 2647: {'lr': 0.0004997758849888693, 'samples': 508416, 'steps': 2647, 'loss/train': 3.494605779647827} 01/27/2022 22:10:42 - INFO - codeparrot_training - Step 2648: {'lr': 0.0004997751917753113, 'samples': 508608, 'steps': 2648, 'loss/train': 5.536654829978943} 01/27/2022 22:10:47 - INFO - codeparrot_training - Step 2649: {'lr': 0.0004997744974917955, 'samples': 508800, 'steps': 2649, 'loss/train': 4.040855526924133} 01/27/2022 22:10:52 - INFO - codeparrot_training - Step 2650: {'lr': 0.0004997738021383252, 'samples': 508992, 'steps': 2650, 'loss/train': 3.4950671195983887} 01/27/2022 22:10:56 - INFO - codeparrot_training - Step 2651: {'lr': 0.000499773105714903, 'samples': 509184, 'steps': 2651, 'loss/train': 4.148135662078857} 01/27/2022 22:11:00 - INFO - codeparrot_training - Step 2652: {'lr': 0.0004997724082215323, 'samples': 509376, 'steps': 2652, 'loss/train': 2.869444191455841} 01/27/2022 22:11:05 - INFO - codeparrot_training - Step 2653: {'lr': 0.0004997717096582159, 'samples': 509568, 'steps': 2653, 'loss/train': 5.487991690635681} 01/27/2022 22:11:09 - INFO - codeparrot_training - Step 2654: {'lr': 0.0004997710100249568, 'samples': 509760, 'steps': 2654, 'loss/train': 4.232197880744934} 01/27/2022 22:11:15 - INFO - codeparrot_training - Step 2655: {'lr': 0.000499770309321758, 'samples': 509952, 'steps': 2655, 'loss/train': 3.649350643157959} 01/27/2022 22:11:19 - INFO - codeparrot_training - Step 2656: {'lr': 0.0004997696075486225, 'samples': 510144, 'steps': 2656, 'loss/train': 3.1563223600387573} 01/27/2022 22:11:23 - INFO - codeparrot_training - Step 2657: {'lr': 0.0004997689047055534, 'samples': 510336, 'steps': 2657, 'loss/train': 2.048428952693939} 01/27/2022 22:11:28 - INFO - codeparrot_training - Step 2658: {'lr': 0.0004997682007925535, 'samples': 510528, 'steps': 2658, 'loss/train': 4.255946159362793} 01/27/2022 22:11:32 - INFO - codeparrot_training - Step 2659: {'lr': 0.0004997674958096259, 'samples': 510720, 'steps': 2659, 'loss/train': 3.8421252965927124} 01/27/2022 22:11:37 - INFO - codeparrot_training - Step 2660: {'lr': 0.0004997667897567738, 'samples': 510912, 'steps': 2660, 'loss/train': 4.70853853225708} 01/27/2022 22:11:41 - INFO - codeparrot_training - Step 2661: {'lr': 0.000499766082634, 'samples': 511104, 'steps': 2661, 'loss/train': 3.8822550773620605} 01/27/2022 22:11:45 - INFO - codeparrot_training - Step 2662: {'lr': 0.0004997653744413076, 'samples': 511296, 'steps': 2662, 'loss/train': 4.238386631011963} 01/27/2022 22:11:50 - INFO - codeparrot_training - Step 2663: {'lr': 0.0004997646651786996, 'samples': 511488, 'steps': 2663, 'loss/train': 3.8904519081115723} 01/27/2022 22:11:54 - INFO - codeparrot_training - Step 2664: {'lr': 0.0004997639548461792, 'samples': 511680, 'steps': 2664, 'loss/train': 5.0052326917648315} 01/27/2022 22:12:00 - INFO - codeparrot_training - Step 2665: {'lr': 0.0004997632434437493, 'samples': 511872, 'steps': 2665, 'loss/train': 4.258145213127136} 01/27/2022 22:12:04 - INFO - codeparrot_training - Step 2666: {'lr': 0.0004997625309714129, 'samples': 512064, 'steps': 2666, 'loss/train': 3.812593460083008} 01/27/2022 22:12:09 - INFO - codeparrot_training - Step 2667: {'lr': 0.0004997618174291732, 'samples': 512256, 'steps': 2667, 'loss/train': 3.447083830833435} 01/27/2022 22:12:13 - INFO - codeparrot_training - Step 2668: {'lr': 0.0004997611028170332, 'samples': 512448, 'steps': 2668, 'loss/train': 4.475387692451477} 01/27/2022 22:12:17 - INFO - codeparrot_training - Step 2669: {'lr': 0.000499760387134996, 'samples': 512640, 'steps': 2669, 'loss/train': 4.630085349082947} 01/27/2022 22:12:23 - INFO - codeparrot_training - Step 2670: {'lr': 0.0004997596703830645, 'samples': 512832, 'steps': 2670, 'loss/train': 4.113433241844177} 01/27/2022 22:12:27 - INFO - codeparrot_training - Step 2671: {'lr': 0.0004997589525612418, 'samples': 513024, 'steps': 2671, 'loss/train': 4.441450595855713} 01/27/2022 22:12:31 - INFO - codeparrot_training - Step 2672: {'lr': 0.0004997582336695312, 'samples': 513216, 'steps': 2672, 'loss/train': 3.769567608833313} 01/27/2022 22:12:35 - INFO - codeparrot_training - Step 2673: {'lr': 0.0004997575137079355, 'samples': 513408, 'steps': 2673, 'loss/train': 4.3808183670043945} 01/27/2022 22:12:39 - INFO - codeparrot_training - Step 2674: {'lr': 0.0004997567926764581, 'samples': 513600, 'steps': 2674, 'loss/train': 3.791569232940674} 01/27/2022 22:12:44 - INFO - codeparrot_training - Step 2675: {'lr': 0.0004997560705751018, 'samples': 513792, 'steps': 2675, 'loss/train': 6.25655722618103} 01/27/2022 22:12:49 - INFO - codeparrot_training - Step 2676: {'lr': 0.0004997553474038698, 'samples': 513984, 'steps': 2676, 'loss/train': 3.3648844957351685} 01/27/2022 22:12:53 - INFO - codeparrot_training - Step 2677: {'lr': 0.0004997546231627652, 'samples': 514176, 'steps': 2677, 'loss/train': 4.6665648221969604} 01/27/2022 22:12:57 - INFO - codeparrot_training - Step 2678: {'lr': 0.0004997538978517912, 'samples': 514368, 'steps': 2678, 'loss/train': 4.091282844543457} 01/27/2022 22:13:01 - INFO - codeparrot_training - Step 2679: {'lr': 0.0004997531714709506, 'samples': 514560, 'steps': 2679, 'loss/train': 3.8542134761810303} 01/27/2022 22:13:07 - INFO - codeparrot_training - Step 2680: {'lr': 0.0004997524440202469, 'samples': 514752, 'steps': 2680, 'loss/train': 2.9880588054656982} 01/27/2022 22:13:12 - INFO - codeparrot_training - Step 2681: {'lr': 0.0004997517154996829, 'samples': 514944, 'steps': 2681, 'loss/train': 2.74605131149292} 01/27/2022 22:13:16 - INFO - codeparrot_training - Step 2682: {'lr': 0.000499750985909262, 'samples': 515136, 'steps': 2682, 'loss/train': 3.995102047920227} 01/27/2022 22:13:20 - INFO - codeparrot_training - Step 2683: {'lr': 0.0004997502552489871, 'samples': 515328, 'steps': 2683, 'loss/train': 4.067229151725769} 01/27/2022 22:13:24 - INFO - codeparrot_training - Step 2684: {'lr': 0.0004997495235188614, 'samples': 515520, 'steps': 2684, 'loss/train': 3.833909511566162} 01/27/2022 22:13:29 - INFO - codeparrot_training - Step 2685: {'lr': 0.0004997487907188881, 'samples': 515712, 'steps': 2685, 'loss/train': 4.449286580085754} 01/27/2022 22:13:34 - INFO - codeparrot_training - Step 2686: {'lr': 0.0004997480568490702, 'samples': 515904, 'steps': 2686, 'loss/train': 3.8874685764312744} 01/27/2022 22:13:38 - INFO - codeparrot_training - Step 2687: {'lr': 0.0004997473219094111, 'samples': 516096, 'steps': 2687, 'loss/train': 5.279076933860779} 01/27/2022 22:13:42 - INFO - codeparrot_training - Step 2688: {'lr': 0.0004997465858999136, 'samples': 516288, 'steps': 2688, 'loss/train': 4.158621668815613} 01/27/2022 22:13:46 - INFO - codeparrot_training - Step 2689: {'lr': 0.0004997458488205811, 'samples': 516480, 'steps': 2689, 'loss/train': 2.7155152559280396} 01/27/2022 22:13:52 - INFO - codeparrot_training - Step 2690: {'lr': 0.0004997451106714166, 'samples': 516672, 'steps': 2690, 'loss/train': 2.9373679161071777} 01/27/2022 22:13:56 - INFO - codeparrot_training - Step 2691: {'lr': 0.0004997443714524235, 'samples': 516864, 'steps': 2691, 'loss/train': 3.3744038343429565} 01/27/2022 22:14:01 - INFO - codeparrot_training - Step 2692: {'lr': 0.0004997436311636046, 'samples': 517056, 'steps': 2692, 'loss/train': 4.330641746520996} 01/27/2022 22:14:05 - INFO - codeparrot_training - Step 2693: {'lr': 0.0004997428898049635, 'samples': 517248, 'steps': 2693, 'loss/train': 4.134361624717712} 01/27/2022 22:14:09 - INFO - codeparrot_training - Step 2694: {'lr': 0.0004997421473765031, 'samples': 517440, 'steps': 2694, 'loss/train': 4.126322865486145} 01/27/2022 22:14:15 - INFO - codeparrot_training - Step 2695: {'lr': 0.0004997414038782266, 'samples': 517632, 'steps': 2695, 'loss/train': 4.587998628616333} 01/27/2022 22:14:19 - INFO - codeparrot_training - Step 2696: {'lr': 0.0004997406593101373, 'samples': 517824, 'steps': 2696, 'loss/train': 4.200737714767456} 01/27/2022 22:14:23 - INFO - codeparrot_training - Step 2697: {'lr': 0.0004997399136722383, 'samples': 518016, 'steps': 2697, 'loss/train': 3.6557693481445312} 01/27/2022 22:14:27 - INFO - codeparrot_training - Step 2698: {'lr': 0.0004997391669645327, 'samples': 518208, 'steps': 2698, 'loss/train': 3.7424354553222656} 01/27/2022 22:14:31 - INFO - codeparrot_training - Step 2699: {'lr': 0.0004997384191870239, 'samples': 518400, 'steps': 2699, 'loss/train': 3.6969913244247437} 01/27/2022 22:14:37 - INFO - codeparrot_training - Step 2700: {'lr': 0.000499737670339715, 'samples': 518592, 'steps': 2700, 'loss/train': 3.1224313974380493} 01/27/2022 22:14:41 - INFO - codeparrot_training - Step 2701: {'lr': 0.0004997369204226093, 'samples': 518784, 'steps': 2701, 'loss/train': 3.7508465051651} 01/27/2022 22:14:45 - INFO - codeparrot_training - Step 2702: {'lr': 0.0004997361694357098, 'samples': 518976, 'steps': 2702, 'loss/train': 3.3160969018936157} 01/27/2022 22:14:49 - INFO - codeparrot_training - Step 2703: {'lr': 0.00049973541737902, 'samples': 519168, 'steps': 2703, 'loss/train': 3.6532809734344482} 01/27/2022 22:14:54 - INFO - codeparrot_training - Step 2704: {'lr': 0.0004997346642525428, 'samples': 519360, 'steps': 2704, 'loss/train': 2.2008435130119324} 01/27/2022 22:14:59 - INFO - codeparrot_training - Step 2705: {'lr': 0.0004997339100562817, 'samples': 519552, 'steps': 2705, 'loss/train': 4.634685158729553} 01/27/2022 22:15:04 - INFO - codeparrot_training - Step 2706: {'lr': 0.0004997331547902398, 'samples': 519744, 'steps': 2706, 'loss/train': 2.9953388571739197} 01/27/2022 22:15:08 - INFO - codeparrot_training - Step 2707: {'lr': 0.0004997323984544204, 'samples': 519936, 'steps': 2707, 'loss/train': 3.82535183429718} 01/27/2022 22:15:12 - INFO - codeparrot_training - Step 2708: {'lr': 0.0004997316410488267, 'samples': 520128, 'steps': 2708, 'loss/train': 2.6371493339538574} 01/27/2022 22:15:16 - INFO - codeparrot_training - Step 2709: {'lr': 0.0004997308825734619, 'samples': 520320, 'steps': 2709, 'loss/train': 2.1445611119270325} 01/27/2022 22:15:22 - INFO - codeparrot_training - Step 2710: {'lr': 0.0004997301230283294, 'samples': 520512, 'steps': 2710, 'loss/train': 3.7710260152816772} 01/27/2022 22:15:26 - INFO - codeparrot_training - Step 2711: {'lr': 0.0004997293624134322, 'samples': 520704, 'steps': 2711, 'loss/train': 3.467681050300598} 01/27/2022 22:15:30 - INFO - codeparrot_training - Step 2712: {'lr': 0.0004997286007287738, 'samples': 520896, 'steps': 2712, 'loss/train': 4.576180100440979} 01/27/2022 22:15:34 - INFO - codeparrot_training - Step 2713: {'lr': 0.0004997278379743574, 'samples': 521088, 'steps': 2713, 'loss/train': 1.8164337873458862} 01/27/2022 22:15:39 - INFO - codeparrot_training - Step 2714: {'lr': 0.0004997270741501861, 'samples': 521280, 'steps': 2714, 'loss/train': 3.7472981214523315} 01/27/2022 22:15:44 - INFO - codeparrot_training - Step 2715: {'lr': 0.0004997263092562634, 'samples': 521472, 'steps': 2715, 'loss/train': 1.9072328209877014} 01/27/2022 22:15:48 - INFO - codeparrot_training - Step 2716: {'lr': 0.0004997255432925926, 'samples': 521664, 'steps': 2716, 'loss/train': 3.24021577835083} 01/27/2022 22:15:53 - INFO - codeparrot_training - Step 2717: {'lr': 0.0004997247762591766, 'samples': 521856, 'steps': 2717, 'loss/train': 4.49659538269043} 01/27/2022 22:15:57 - INFO - codeparrot_training - Step 2718: {'lr': 0.0004997240081560193, 'samples': 522048, 'steps': 2718, 'loss/train': 2.802269160747528} 01/27/2022 22:16:01 - INFO - codeparrot_training - Step 2719: {'lr': 0.0004997232389831234, 'samples': 522240, 'steps': 2719, 'loss/train': 2.5930275321006775} 01/27/2022 22:16:06 - INFO - codeparrot_training - Step 2720: {'lr': 0.0004997224687404926, 'samples': 522432, 'steps': 2720, 'loss/train': 3.177720308303833} 01/27/2022 22:16:11 - INFO - codeparrot_training - Step 2721: {'lr': 0.0004997216974281299, 'samples': 522624, 'steps': 2721, 'loss/train': 3.8938060998916626} 01/27/2022 22:16:15 - INFO - codeparrot_training - Step 2722: {'lr': 0.0004997209250460387, 'samples': 522816, 'steps': 2722, 'loss/train': 3.6972867250442505} 01/27/2022 22:16:19 - INFO - codeparrot_training - Step 2723: {'lr': 0.0004997201515942225, 'samples': 523008, 'steps': 2723, 'loss/train': 4.084325909614563} 01/27/2022 22:16:23 - INFO - codeparrot_training - Step 2724: {'lr': 0.0004997193770726844, 'samples': 523200, 'steps': 2724, 'loss/train': 2.6181854009628296} 01/27/2022 22:16:29 - INFO - codeparrot_training - Step 2725: {'lr': 0.0004997186014814278, 'samples': 523392, 'steps': 2725, 'loss/train': 3.6696828603744507} 01/27/2022 22:16:34 - INFO - codeparrot_training - Step 2726: {'lr': 0.000499717824820456, 'samples': 523584, 'steps': 2726, 'loss/train': 3.845784902572632} 01/27/2022 22:16:38 - INFO - codeparrot_training - Step 2727: {'lr': 0.0004997170470897723, 'samples': 523776, 'steps': 2727, 'loss/train': 2.163969933986664} 01/27/2022 22:16:42 - INFO - codeparrot_training - Step 2728: {'lr': 0.0004997162682893801, 'samples': 523968, 'steps': 2728, 'loss/train': 4.44156289100647} 01/27/2022 22:16:46 - INFO - codeparrot_training - Step 2729: {'lr': 0.0004997154884192827, 'samples': 524160, 'steps': 2729, 'loss/train': 3.21248996257782} 01/27/2022 22:16:52 - INFO - codeparrot_training - Step 2730: {'lr': 0.0004997147074794835, 'samples': 524352, 'steps': 2730, 'loss/train': 3.4793901443481445} 01/27/2022 22:16:56 - INFO - codeparrot_training - Step 2731: {'lr': 0.0004997139254699856, 'samples': 524544, 'steps': 2731, 'loss/train': 2.6209816932678223} 01/27/2022 22:17:00 - INFO - codeparrot_training - Step 2732: {'lr': 0.0004997131423907927, 'samples': 524736, 'steps': 2732, 'loss/train': 3.789306163787842} 01/27/2022 22:17:04 - INFO - codeparrot_training - Step 2733: {'lr': 0.000499712358241908, 'samples': 524928, 'steps': 2733, 'loss/train': 3.9580278396606445} 01/27/2022 22:17:08 - INFO - codeparrot_training - Step 2734: {'lr': 0.0004997115730233349, 'samples': 525120, 'steps': 2734, 'loss/train': 3.325494647026062} 01/27/2022 22:17:14 - INFO - codeparrot_training - Step 2735: {'lr': 0.0004997107867350765, 'samples': 525312, 'steps': 2735, 'loss/train': 4.323053956031799} 01/27/2022 22:17:19 - INFO - codeparrot_training - Step 2736: {'lr': 0.0004997099993771365, 'samples': 525504, 'steps': 2736, 'loss/train': 2.884681820869446} 01/27/2022 22:17:23 - INFO - codeparrot_training - Step 2737: {'lr': 0.0004997092109495181, 'samples': 525696, 'steps': 2737, 'loss/train': 3.38588547706604} 01/27/2022 22:17:27 - INFO - codeparrot_training - Step 2738: {'lr': 0.0004997084214522249, 'samples': 525888, 'steps': 2738, 'loss/train': 4.089487910270691} 01/27/2022 22:17:31 - INFO - codeparrot_training - Step 2739: {'lr': 0.0004997076308852599, 'samples': 526080, 'steps': 2739, 'loss/train': 2.8939009308815002} 01/27/2022 22:17:36 - INFO - codeparrot_training - Step 2740: {'lr': 0.0004997068392486268, 'samples': 526272, 'steps': 2740, 'loss/train': 3.1609532833099365} 01/27/2022 22:17:40 - INFO - codeparrot_training - Step 2741: {'lr': 0.0004997060465423288, 'samples': 526464, 'steps': 2741, 'loss/train': 4.385903477668762} 01/27/2022 22:17:45 - INFO - codeparrot_training - Step 2742: {'lr': 0.0004997052527663696, 'samples': 526656, 'steps': 2742, 'loss/train': 3.9015623331069946} 01/27/2022 22:17:49 - INFO - codeparrot_training - Step 2743: {'lr': 0.0004997044579207522, 'samples': 526848, 'steps': 2743, 'loss/train': 3.92998731136322} 01/27/2022 22:17:53 - INFO - codeparrot_training - Step 2744: {'lr': 0.0004997036620054803, 'samples': 527040, 'steps': 2744, 'loss/train': 3.9554754495620728} 01/27/2022 22:17:58 - INFO - codeparrot_training - Step 2745: {'lr': 0.0004997028650205572, 'samples': 527232, 'steps': 2745, 'loss/train': 2.5738536715507507} 01/27/2022 22:18:02 - INFO - codeparrot_training - Step 2746: {'lr': 0.0004997020669659862, 'samples': 527424, 'steps': 2746, 'loss/train': 3.850409746170044} 01/27/2022 22:18:07 - INFO - codeparrot_training - Step 2747: {'lr': 0.000499701267841771, 'samples': 527616, 'steps': 2747, 'loss/train': 3.6227227449417114} 01/27/2022 22:18:11 - INFO - codeparrot_training - Step 2748: {'lr': 0.0004997004676479147, 'samples': 527808, 'steps': 2748, 'loss/train': 2.9085259437561035} 01/27/2022 22:18:15 - INFO - codeparrot_training - Step 2749: {'lr': 0.0004996996663844209, 'samples': 528000, 'steps': 2749, 'loss/train': 4.931248784065247} 01/27/2022 22:18:21 - INFO - codeparrot_training - Step 2750: {'lr': 0.0004996988640512931, 'samples': 528192, 'steps': 2750, 'loss/train': 3.364799737930298} 01/27/2022 22:18:26 - INFO - codeparrot_training - Step 2751: {'lr': 0.0004996980606485346, 'samples': 528384, 'steps': 2751, 'loss/train': 3.55691134929657} 01/27/2022 22:18:30 - INFO - codeparrot_training - Step 2752: {'lr': 0.0004996972561761489, 'samples': 528576, 'steps': 2752, 'loss/train': 4.878429293632507} 01/27/2022 22:18:34 - INFO - codeparrot_training - Step 2753: {'lr': 0.0004996964506341395, 'samples': 528768, 'steps': 2753, 'loss/train': 3.063667416572571} 01/27/2022 22:18:38 - INFO - codeparrot_training - Step 2754: {'lr': 0.0004996956440225098, 'samples': 528960, 'steps': 2754, 'loss/train': 3.7838491201400757} 01/27/2022 22:18:44 - INFO - codeparrot_training - Step 2755: {'lr': 0.0004996948363412631, 'samples': 529152, 'steps': 2755, 'loss/train': 3.775256037712097} 01/27/2022 22:18:48 - INFO - codeparrot_training - Step 2756: {'lr': 0.0004996940275904031, 'samples': 529344, 'steps': 2756, 'loss/train': 3.2277567386627197} 01/27/2022 22:18:52 - INFO - codeparrot_training - Step 2757: {'lr': 0.0004996932177699332, 'samples': 529536, 'steps': 2757, 'loss/train': 4.417131543159485} 01/27/2022 22:18:56 - INFO - codeparrot_training - Step 2758: {'lr': 0.0004996924068798569, 'samples': 529728, 'steps': 2758, 'loss/train': 4.790768265724182} 01/27/2022 22:19:00 - INFO - codeparrot_training - Step 2759: {'lr': 0.0004996915949201775, 'samples': 529920, 'steps': 2759, 'loss/train': 3.02440345287323} 01/27/2022 22:19:06 - INFO - codeparrot_training - Step 2760: {'lr': 0.0004996907818908987, 'samples': 530112, 'steps': 2760, 'loss/train': 1.974201500415802} 01/27/2022 22:19:10 - INFO - codeparrot_training - Step 2761: {'lr': 0.0004996899677920238, 'samples': 530304, 'steps': 2761, 'loss/train': 4.243656635284424} 01/27/2022 22:19:15 - INFO - codeparrot_training - Step 2762: {'lr': 0.0004996891526235564, 'samples': 530496, 'steps': 2762, 'loss/train': 3.3810778856277466} 01/27/2022 22:19:19 - INFO - codeparrot_training - Step 2763: {'lr': 0.0004996883363854998, 'samples': 530688, 'steps': 2763, 'loss/train': 4.138623476028442} 01/27/2022 22:19:23 - INFO - codeparrot_training - Step 2764: {'lr': 0.0004996875190778579, 'samples': 530880, 'steps': 2764, 'loss/train': 2.153615176677704} 01/27/2022 22:19:28 - INFO - codeparrot_training - Step 2765: {'lr': 0.0004996867007006339, 'samples': 531072, 'steps': 2765, 'loss/train': 5.367157459259033} 01/27/2022 22:19:32 - INFO - codeparrot_training - Step 2766: {'lr': 0.0004996858812538312, 'samples': 531264, 'steps': 2766, 'loss/train': 7.367344379425049} 01/27/2022 22:19:37 - INFO - codeparrot_training - Step 2767: {'lr': 0.0004996850607374535, 'samples': 531456, 'steps': 2767, 'loss/train': 2.9494444727897644} 01/27/2022 22:19:41 - INFO - codeparrot_training - Step 2768: {'lr': 0.0004996842391515044, 'samples': 531648, 'steps': 2768, 'loss/train': 4.509234666824341} 01/27/2022 22:19:45 - INFO - codeparrot_training - Step 2769: {'lr': 0.0004996834164959872, 'samples': 531840, 'steps': 2769, 'loss/train': 4.219833970069885} 01/27/2022 22:19:50 - INFO - codeparrot_training - Step 2770: {'lr': 0.0004996825927709056, 'samples': 532032, 'steps': 2770, 'loss/train': 3.563303589820862} 01/27/2022 22:19:55 - INFO - codeparrot_training - Step 2771: {'lr': 0.0004996817679762631, 'samples': 532224, 'steps': 2771, 'loss/train': 3.3706880807876587} 01/27/2022 22:19:59 - INFO - codeparrot_training - Step 2772: {'lr': 0.000499680942112063, 'samples': 532416, 'steps': 2772, 'loss/train': 5.361120343208313} 01/27/2022 22:20:03 - INFO - codeparrot_training - Step 2773: {'lr': 0.0004996801151783092, 'samples': 532608, 'steps': 2773, 'loss/train': 4.673068642616272} 01/27/2022 22:20:07 - INFO - codeparrot_training - Step 2774: {'lr': 0.000499679287175005, 'samples': 532800, 'steps': 2774, 'loss/train': 4.105987071990967} 01/27/2022 22:20:14 - INFO - codeparrot_training - Step 2775: {'lr': 0.000499678458102154, 'samples': 532992, 'steps': 2775, 'loss/train': 3.586917757987976} 01/27/2022 22:20:18 - INFO - codeparrot_training - Step 2776: {'lr': 0.0004996776279597598, 'samples': 533184, 'steps': 2776, 'loss/train': 3.7293609380722046} 01/27/2022 22:20:22 - INFO - codeparrot_training - Step 2777: {'lr': 0.0004996767967478259, 'samples': 533376, 'steps': 2777, 'loss/train': 3.5921337604522705} 01/27/2022 22:20:26 - INFO - codeparrot_training - Step 2778: {'lr': 0.0004996759644663559, 'samples': 533568, 'steps': 2778, 'loss/train': 4.034047722816467} 01/27/2022 22:20:32 - INFO - codeparrot_training - Step 2779: {'lr': 0.0004996751311153535, 'samples': 533760, 'steps': 2779, 'loss/train': 2.838642418384552} 01/27/2022 22:20:36 - INFO - codeparrot_training - Step 2780: {'lr': 0.0004996742966948219, 'samples': 533952, 'steps': 2780, 'loss/train': 2.6877129077911377} 01/27/2022 22:20:40 - INFO - codeparrot_training - Step 2781: {'lr': 0.000499673461204765, 'samples': 534144, 'steps': 2781, 'loss/train': 3.574660062789917} 01/27/2022 22:20:44 - INFO - codeparrot_training - Step 2782: {'lr': 0.0004996726246451862, 'samples': 534336, 'steps': 2782, 'loss/train': 4.398756623268127} 01/27/2022 22:20:49 - INFO - codeparrot_training - Step 2783: {'lr': 0.0004996717870160892, 'samples': 534528, 'steps': 2783, 'loss/train': 3.235729694366455} 01/27/2022 22:20:54 - INFO - codeparrot_training - Step 2784: {'lr': 0.0004996709483174775, 'samples': 534720, 'steps': 2784, 'loss/train': 3.74263072013855} 01/27/2022 22:20:58 - INFO - codeparrot_training - Step 2785: {'lr': 0.0004996701085493547, 'samples': 534912, 'steps': 2785, 'loss/train': 4.100013256072998} 01/27/2022 22:21:02 - INFO - codeparrot_training - Step 2786: {'lr': 0.0004996692677117246, 'samples': 535104, 'steps': 2786, 'loss/train': 3.4193226099014282} 01/27/2022 22:21:07 - INFO - codeparrot_training - Step 2787: {'lr': 0.0004996684258045906, 'samples': 535296, 'steps': 2787, 'loss/train': 3.3135573863983154} 01/27/2022 22:21:11 - INFO - codeparrot_training - Step 2788: {'lr': 0.0004996675828279562, 'samples': 535488, 'steps': 2788, 'loss/train': 2.500487208366394} 01/27/2022 22:21:16 - INFO - codeparrot_training - Step 2789: {'lr': 0.0004996667387818254, 'samples': 535680, 'steps': 2789, 'loss/train': 4.53313136100769} 01/27/2022 22:21:21 - INFO - codeparrot_training - Step 2790: {'lr': 0.0004996658936662013, 'samples': 535872, 'steps': 2790, 'loss/train': 3.217542886734009} 01/27/2022 22:21:25 - INFO - codeparrot_training - Step 2791: {'lr': 0.0004996650474810879, 'samples': 536064, 'steps': 2791, 'loss/train': 2.963605284690857} 01/27/2022 22:21:29 - INFO - codeparrot_training - Step 2792: {'lr': 0.0004996642002264887, 'samples': 536256, 'steps': 2792, 'loss/train': 4.1138691902160645} 01/27/2022 22:21:33 - INFO - codeparrot_training - Step 2793: {'lr': 0.0004996633519024074, 'samples': 536448, 'steps': 2793, 'loss/train': 3.1590875387191772} 01/27/2022 22:21:38 - INFO - codeparrot_training - Step 2794: {'lr': 0.0004996625025088476, 'samples': 536640, 'steps': 2794, 'loss/train': 3.42733633518219} 01/27/2022 22:21:43 - INFO - codeparrot_training - Step 2795: {'lr': 0.0004996616520458128, 'samples': 536832, 'steps': 2795, 'loss/train': 4.2283430099487305} 01/27/2022 22:21:48 - INFO - codeparrot_training - Step 2796: {'lr': 0.0004996608005133068, 'samples': 537024, 'steps': 2796, 'loss/train': 3.4039024114608765} 01/27/2022 22:21:52 - INFO - codeparrot_training - Step 2797: {'lr': 0.0004996599479113333, 'samples': 537216, 'steps': 2797, 'loss/train': 3.98948872089386} 01/27/2022 22:21:56 - INFO - codeparrot_training - Step 2798: {'lr': 0.0004996590942398958, 'samples': 537408, 'steps': 2798, 'loss/train': 3.9280790090560913} 01/27/2022 22:22:00 - INFO - codeparrot_training - Step 2799: {'lr': 0.0004996582394989979, 'samples': 537600, 'steps': 2799, 'loss/train': 3.560028076171875} 01/27/2022 22:22:06 - INFO - codeparrot_training - Step 2800: {'lr': 0.0004996573836886434, 'samples': 537792, 'steps': 2800, 'loss/train': 3.910054922103882} 01/27/2022 22:22:10 - INFO - codeparrot_training - Step 2801: {'lr': 0.0004996565268088362, 'samples': 537984, 'steps': 2801, 'loss/train': 4.1111966371536255} 01/27/2022 22:22:14 - INFO - codeparrot_training - Step 2802: {'lr': 0.0004996556688595794, 'samples': 538176, 'steps': 2802, 'loss/train': 4.125572562217712} 01/27/2022 22:22:18 - INFO - codeparrot_training - Step 2803: {'lr': 0.0004996548098408772, 'samples': 538368, 'steps': 2803, 'loss/train': 2.7561448216438293} 01/27/2022 22:22:23 - INFO - codeparrot_training - Step 2804: {'lr': 0.0004996539497527329, 'samples': 538560, 'steps': 2804, 'loss/train': 2.235084056854248} 01/27/2022 22:22:28 - INFO - codeparrot_training - Step 2805: {'lr': 0.0004996530885951505, 'samples': 538752, 'steps': 2805, 'loss/train': 3.3882640600204468} 01/27/2022 22:22:33 - INFO - codeparrot_training - Step 2806: {'lr': 0.0004996522263681335, 'samples': 538944, 'steps': 2806, 'loss/train': 4.132438659667969} 01/27/2022 22:22:37 - INFO - codeparrot_training - Step 2807: {'lr': 0.0004996513630716856, 'samples': 539136, 'steps': 2807, 'loss/train': 3.9966888427734375} 01/27/2022 22:22:41 - INFO - codeparrot_training - Step 2808: {'lr': 0.0004996504987058105, 'samples': 539328, 'steps': 2808, 'loss/train': 2.4260494709014893} 01/27/2022 22:22:45 - INFO - codeparrot_training - Step 2809: {'lr': 0.000499649633270512, 'samples': 539520, 'steps': 2809, 'loss/train': 3.852301239967346} 01/27/2022 22:22:51 - INFO - codeparrot_training - Step 2810: {'lr': 0.0004996487667657938, 'samples': 539712, 'steps': 2810, 'loss/train': 2.72189462184906} 01/27/2022 22:22:55 - INFO - codeparrot_training - Step 2811: {'lr': 0.0004996478991916595, 'samples': 539904, 'steps': 2811, 'loss/train': 4.1588627099990845} 01/27/2022 22:22:59 - INFO - codeparrot_training - Step 2812: {'lr': 0.0004996470305481127, 'samples': 540096, 'steps': 2812, 'loss/train': 4.368572115898132} 01/27/2022 22:23:03 - INFO - codeparrot_training - Step 2813: {'lr': 0.0004996461608351575, 'samples': 540288, 'steps': 2813, 'loss/train': 3.9860748052597046} 01/27/2022 22:23:08 - INFO - codeparrot_training - Step 2814: {'lr': 0.0004996452900527974, 'samples': 540480, 'steps': 2814, 'loss/train': 4.211032390594482} 01/27/2022 22:23:13 - INFO - codeparrot_training - Step 2815: {'lr': 0.0004996444182010361, 'samples': 540672, 'steps': 2815, 'loss/train': 4.1571714878082275} 01/27/2022 22:23:17 - INFO - codeparrot_training - Step 2816: {'lr': 0.0004996435452798775, 'samples': 540864, 'steps': 2816, 'loss/train': 3.4233827590942383} 01/27/2022 22:23:22 - INFO - codeparrot_training - Step 2817: {'lr': 0.000499642671289325, 'samples': 541056, 'steps': 2817, 'loss/train': 4.1387282609939575} 01/27/2022 22:23:26 - INFO - codeparrot_training - Step 2818: {'lr': 0.0004996417962293828, 'samples': 541248, 'steps': 2818, 'loss/train': 4.448689341545105} 01/27/2022 22:23:30 - INFO - codeparrot_training - Step 2819: {'lr': 0.0004996409201000543, 'samples': 541440, 'steps': 2819, 'loss/train': 3.243462324142456} 01/27/2022 22:23:36 - INFO - codeparrot_training - Step 2820: {'lr': 0.0004996400429013434, 'samples': 541632, 'steps': 2820, 'loss/train': 3.5181963443756104} 01/27/2022 22:23:40 - INFO - codeparrot_training - Step 2821: {'lr': 0.0004996391646332537, 'samples': 541824, 'steps': 2821, 'loss/train': 3.339430332183838} 01/27/2022 22:23:44 - INFO - codeparrot_training - Step 2822: {'lr': 0.0004996382852957892, 'samples': 542016, 'steps': 2822, 'loss/train': 2.3520559072494507} 01/27/2022 22:23:49 - INFO - codeparrot_training - Step 2823: {'lr': 0.0004996374048889536, 'samples': 542208, 'steps': 2823, 'loss/train': 3.918536067008972} 01/27/2022 22:23:53 - INFO - codeparrot_training - Step 2824: {'lr': 0.0004996365234127506, 'samples': 542400, 'steps': 2824, 'loss/train': 3.928068995475769} 01/27/2022 22:23:58 - INFO - codeparrot_training - Step 2825: {'lr': 0.000499635640867184, 'samples': 542592, 'steps': 2825, 'loss/train': 4.114068388938904} 01/27/2022 22:24:02 - INFO - codeparrot_training - Step 2826: {'lr': 0.0004996347572522575, 'samples': 542784, 'steps': 2826, 'loss/train': 5.346673607826233} 01/27/2022 22:24:07 - INFO - codeparrot_training - Step 2827: {'lr': 0.000499633872567975, 'samples': 542976, 'steps': 2827, 'loss/train': 3.736755609512329} 01/27/2022 22:24:11 - INFO - codeparrot_training - Step 2828: {'lr': 0.0004996329868143404, 'samples': 543168, 'steps': 2828, 'loss/train': 3.39603853225708} 01/27/2022 22:24:15 - INFO - codeparrot_training - Step 2829: {'lr': 0.0004996320999913572, 'samples': 543360, 'steps': 2829, 'loss/train': 4.391223907470703} 01/27/2022 22:24:20 - INFO - codeparrot_training - Step 2830: {'lr': 0.0004996312120990293, 'samples': 543552, 'steps': 2830, 'loss/train': 4.00505518913269} 01/27/2022 22:24:24 - INFO - codeparrot_training - Step 2831: {'lr': 0.0004996303231373607, 'samples': 543744, 'steps': 2831, 'loss/train': 3.9200735092163086} 01/27/2022 22:24:29 - INFO - codeparrot_training - Step 2832: {'lr': 0.000499629433106355, 'samples': 543936, 'steps': 2832, 'loss/train': 3.171814441680908} 01/27/2022 22:24:33 - INFO - codeparrot_training - Step 2833: {'lr': 0.000499628542006016, 'samples': 544128, 'steps': 2833, 'loss/train': 3.996475338935852} 01/27/2022 22:24:37 - INFO - codeparrot_training - Step 2834: {'lr': 0.0004996276498363477, 'samples': 544320, 'steps': 2834, 'loss/train': 4.196946501731873} 01/27/2022 22:24:42 - INFO - codeparrot_training - Step 2835: {'lr': 0.0004996267565973538, 'samples': 544512, 'steps': 2835, 'loss/train': 5.5498799085617065} 01/27/2022 22:24:47 - INFO - codeparrot_training - Step 2836: {'lr': 0.0004996258622890381, 'samples': 544704, 'steps': 2836, 'loss/train': 3.899376153945923} 01/27/2022 22:24:51 - INFO - codeparrot_training - Step 2837: {'lr': 0.0004996249669114045, 'samples': 544896, 'steps': 2837, 'loss/train': 2.9255497455596924} 01/27/2022 22:24:55 - INFO - codeparrot_training - Step 2838: {'lr': 0.0004996240704644568, 'samples': 545088, 'steps': 2838, 'loss/train': 3.8894530534744263} 01/27/2022 22:24:59 - INFO - codeparrot_training - Step 2839: {'lr': 0.0004996231729481989, 'samples': 545280, 'steps': 2839, 'loss/train': 3.1907769441604614} 01/27/2022 22:25:05 - INFO - codeparrot_training - Step 2840: {'lr': 0.0004996222743626345, 'samples': 545472, 'steps': 2840, 'loss/train': 3.0539166927337646} 01/27/2022 22:25:09 - INFO - codeparrot_training - Step 2841: {'lr': 0.0004996213747077675, 'samples': 545664, 'steps': 2841, 'loss/train': 3.6018787622451782} 01/27/2022 22:25:14 - INFO - codeparrot_training - Step 2842: {'lr': 0.0004996204739836019, 'samples': 545856, 'steps': 2842, 'loss/train': 4.520619034767151} 01/27/2022 22:25:18 - INFO - codeparrot_training - Step 2843: {'lr': 0.0004996195721901415, 'samples': 546048, 'steps': 2843, 'loss/train': 2.521074950695038} 01/27/2022 22:25:22 - INFO - codeparrot_training - Step 2844: {'lr': 0.00049961866932739, 'samples': 546240, 'steps': 2844, 'loss/train': 2.8180822134017944} 01/27/2022 22:25:27 - INFO - codeparrot_training - Step 2845: {'lr': 0.0004996177653953514, 'samples': 546432, 'steps': 2845, 'loss/train': 3.5483983755111694} 01/27/2022 22:25:31 - INFO - codeparrot_training - Step 2846: {'lr': 0.0004996168603940296, 'samples': 546624, 'steps': 2846, 'loss/train': 4.164845108985901} 01/27/2022 22:25:36 - INFO - codeparrot_training - Step 2847: {'lr': 0.0004996159543234285, 'samples': 546816, 'steps': 2847, 'loss/train': 3.2032281160354614} 01/27/2022 22:25:40 - INFO - codeparrot_training - Step 2848: {'lr': 0.0004996150471835518, 'samples': 547008, 'steps': 2848, 'loss/train': 3.6209110021591187} 01/27/2022 22:25:44 - INFO - codeparrot_training - Step 2849: {'lr': 0.0004996141389744035, 'samples': 547200, 'steps': 2849, 'loss/train': 4.3613516092300415} 01/27/2022 22:25:50 - INFO - codeparrot_training - Step 2850: {'lr': 0.0004996132296959876, 'samples': 547392, 'steps': 2850, 'loss/train': 4.250658988952637} 01/27/2022 22:25:54 - INFO - codeparrot_training - Step 2851: {'lr': 0.0004996123193483076, 'samples': 547584, 'steps': 2851, 'loss/train': 2.9460246562957764} 01/27/2022 22:25:58 - INFO - codeparrot_training - Step 2852: {'lr': 0.000499611407931368, 'samples': 547776, 'steps': 2852, 'loss/train': 4.126279950141907} 01/27/2022 22:26:02 - INFO - codeparrot_training - Step 2853: {'lr': 0.0004996104954451722, 'samples': 547968, 'steps': 2853, 'loss/train': 2.853613793849945} 01/27/2022 22:26:06 - INFO - codeparrot_training - Step 2854: {'lr': 0.0004996095818897245, 'samples': 548160, 'steps': 2854, 'loss/train': 4.132460117340088} 01/27/2022 22:26:13 - INFO - codeparrot_training - Step 2855: {'lr': 0.0004996086672650284, 'samples': 548352, 'steps': 2855, 'loss/train': 3.609947919845581} 01/27/2022 22:26:17 - INFO - codeparrot_training - Step 2856: {'lr': 0.0004996077515710881, 'samples': 548544, 'steps': 2856, 'loss/train': 2.407172441482544} 01/27/2022 22:26:21 - INFO - codeparrot_training - Step 2857: {'lr': 0.0004996068348079075, 'samples': 548736, 'steps': 2857, 'loss/train': 3.29369580745697} 01/27/2022 22:26:25 - INFO - codeparrot_training - Step 2858: {'lr': 0.0004996059169754904, 'samples': 548928, 'steps': 2858, 'loss/train': 3.2653677463531494} 01/27/2022 22:26:31 - INFO - codeparrot_training - Step 2859: {'lr': 0.0004996049980738409, 'samples': 549120, 'steps': 2859, 'loss/train': 2.8425773978233337} 01/27/2022 22:26:35 - INFO - codeparrot_training - Step 2860: {'lr': 0.0004996040781029629, 'samples': 549312, 'steps': 2860, 'loss/train': 2.7872146368026733} 01/27/2022 22:26:39 - INFO - codeparrot_training - Step 2861: {'lr': 0.00049960315706286, 'samples': 549504, 'steps': 2861, 'loss/train': 4.788225531578064} 01/27/2022 22:26:43 - INFO - codeparrot_training - Step 2862: {'lr': 0.0004996022349535367, 'samples': 549696, 'steps': 2862, 'loss/train': 3.2519320249557495} 01/27/2022 22:26:48 - INFO - codeparrot_training - Step 2863: {'lr': 0.0004996013117749967, 'samples': 549888, 'steps': 2863, 'loss/train': 5.396276593208313} 01/27/2022 22:26:53 - INFO - codeparrot_training - Step 2864: {'lr': 0.0004996003875272438, 'samples': 550080, 'steps': 2864, 'loss/train': 2.228762984275818} 01/27/2022 22:26:58 - INFO - codeparrot_training - Step 2865: {'lr': 0.0004995994622102821, 'samples': 550272, 'steps': 2865, 'loss/train': 3.662929058074951} 01/27/2022 22:27:02 - INFO - codeparrot_training - Step 2866: {'lr': 0.0004995985358241156, 'samples': 550464, 'steps': 2866, 'loss/train': 3.833168864250183} 01/27/2022 22:27:06 - INFO - codeparrot_training - Step 2867: {'lr': 0.0004995976083687482, 'samples': 550656, 'steps': 2867, 'loss/train': 2.71035897731781} 01/27/2022 22:27:10 - INFO - codeparrot_training - Step 2868: {'lr': 0.000499596679844184, 'samples': 550848, 'steps': 2868, 'loss/train': 4.549469590187073} 01/27/2022 22:27:16 - INFO - codeparrot_training - Step 2869: {'lr': 0.0004995957502504268, 'samples': 551040, 'steps': 2869, 'loss/train': 3.8582843542099} 01/27/2022 22:27:20 - INFO - codeparrot_training - Step 2870: {'lr': 0.0004995948195874807, 'samples': 551232, 'steps': 2870, 'loss/train': 3.781654715538025} 01/27/2022 22:27:24 - INFO - codeparrot_training - Step 2871: {'lr': 0.0004995938878553496, 'samples': 551424, 'steps': 2871, 'loss/train': 2.529962718486786} 01/27/2022 22:27:28 - INFO - codeparrot_training - Step 2872: {'lr': 0.0004995929550540376, 'samples': 551616, 'steps': 2872, 'loss/train': 2.7379446029663086} 01/27/2022 22:27:32 - INFO - codeparrot_training - Step 2873: {'lr': 0.0004995920211835485, 'samples': 551808, 'steps': 2873, 'loss/train': 4.8426371812820435} 01/27/2022 22:27:38 - INFO - codeparrot_training - Step 2874: {'lr': 0.0004995910862438866, 'samples': 552000, 'steps': 2874, 'loss/train': 3.7835183143615723} 01/27/2022 22:27:42 - INFO - codeparrot_training - Step 2875: {'lr': 0.0004995901502350556, 'samples': 552192, 'steps': 2875, 'loss/train': 3.9660362005233765} 01/27/2022 22:27:46 - INFO - codeparrot_training - Step 2876: {'lr': 0.0004995892131570598, 'samples': 552384, 'steps': 2876, 'loss/train': 1.3182950913906097} 01/27/2022 22:27:50 - INFO - codeparrot_training - Step 2877: {'lr': 0.0004995882750099029, 'samples': 552576, 'steps': 2877, 'loss/train': 2.5387340784072876} 01/27/2022 22:27:55 - INFO - codeparrot_training - Step 2878: {'lr': 0.0004995873357935892, 'samples': 552768, 'steps': 2878, 'loss/train': 1.693309485912323} 01/27/2022 22:27:59 - INFO - codeparrot_training - Step 2879: {'lr': 0.0004995863955081226, 'samples': 552960, 'steps': 2879, 'loss/train': 3.563222050666809} 01/27/2022 22:28:05 - INFO - codeparrot_training - Step 2880: {'lr': 0.0004995854541535071, 'samples': 553152, 'steps': 2880, 'loss/train': 3.8985142707824707} 01/27/2022 22:28:10 - INFO - codeparrot_training - Step 2881: {'lr': 0.0004995845117297468, 'samples': 553344, 'steps': 2881, 'loss/train': 2.5785491466522217} 01/27/2022 22:28:14 - INFO - codeparrot_training - Step 2882: {'lr': 0.0004995835682368457, 'samples': 553536, 'steps': 2882, 'loss/train': 3.2982351779937744} 01/27/2022 22:28:18 - INFO - codeparrot_training - Step 2883: {'lr': 0.0004995826236748078, 'samples': 553728, 'steps': 2883, 'loss/train': 4.226249814033508} 01/27/2022 22:28:23 - INFO - codeparrot_training - Step 2884: {'lr': 0.0004995816780436372, 'samples': 553920, 'steps': 2884, 'loss/train': 4.1987621784210205} 01/27/2022 22:28:28 - INFO - codeparrot_training - Step 2885: {'lr': 0.0004995807313433379, 'samples': 554112, 'steps': 2885, 'loss/train': 3.487825870513916} 01/27/2022 22:28:32 - INFO - codeparrot_training - Step 2886: {'lr': 0.0004995797835739141, 'samples': 554304, 'steps': 2886, 'loss/train': 3.7370470762252808} 01/27/2022 22:28:36 - INFO - codeparrot_training - Step 2887: {'lr': 0.0004995788347353697, 'samples': 554496, 'steps': 2887, 'loss/train': 3.2748663425445557} 01/27/2022 22:28:40 - INFO - codeparrot_training - Step 2888: {'lr': 0.0004995778848277088, 'samples': 554688, 'steps': 2888, 'loss/train': 5.454961895942688} 01/27/2022 22:28:45 - INFO - codeparrot_training - Step 2889: {'lr': 0.0004995769338509357, 'samples': 554880, 'steps': 2889, 'loss/train': 4.108805894851685} 01/27/2022 22:28:50 - INFO - codeparrot_training - Step 2890: {'lr': 0.000499575981805054, 'samples': 555072, 'steps': 2890, 'loss/train': 1.9958842992782593} 01/27/2022 22:28:54 - INFO - codeparrot_training - Step 2891: {'lr': 0.000499575028690068, 'samples': 555264, 'steps': 2891, 'loss/train': 4.132209777832031} 01/27/2022 22:28:58 - INFO - codeparrot_training - Step 2892: {'lr': 0.000499574074505982, 'samples': 555456, 'steps': 2892, 'loss/train': 4.135448455810547} 01/27/2022 22:29:02 - INFO - codeparrot_training - Step 2893: {'lr': 0.0004995731192527999, 'samples': 555648, 'steps': 2893, 'loss/train': 3.8598403930664062} 01/27/2022 22:29:08 - INFO - codeparrot_training - Step 2894: {'lr': 0.0004995721629305258, 'samples': 555840, 'steps': 2894, 'loss/train': 3.9756020307540894} 01/27/2022 22:29:13 - INFO - codeparrot_training - Step 2895: {'lr': 0.0004995712055391638, 'samples': 556032, 'steps': 2895, 'loss/train': 3.8744713068008423} 01/27/2022 22:29:17 - INFO - codeparrot_training - Step 2896: {'lr': 0.000499570247078718, 'samples': 556224, 'steps': 2896, 'loss/train': 3.892065167427063} 01/27/2022 22:29:21 - INFO - codeparrot_training - Step 2897: {'lr': 0.0004995692875491925, 'samples': 556416, 'steps': 2897, 'loss/train': 2.598599910736084} 01/27/2022 22:29:25 - INFO - codeparrot_training - Step 2898: {'lr': 0.0004995683269505914, 'samples': 556608, 'steps': 2898, 'loss/train': 2.27533757686615} 01/27/2022 22:29:30 - INFO - codeparrot_training - Step 2899: {'lr': 0.000499567365282919, 'samples': 556800, 'steps': 2899, 'loss/train': 4.068802714347839} 01/27/2022 22:29:35 - INFO - codeparrot_training - Step 2900: {'lr': 0.000499566402546179, 'samples': 556992, 'steps': 2900, 'loss/train': 4.663514971733093} 01/27/2022 22:29:39 - INFO - codeparrot_training - Step 2901: {'lr': 0.0004995654387403758, 'samples': 557184, 'steps': 2901, 'loss/train': 3.900147557258606} 01/27/2022 22:29:43 - INFO - codeparrot_training - Step 2902: {'lr': 0.0004995644738655136, 'samples': 557376, 'steps': 2902, 'loss/train': 2.3868547081947327} 01/27/2022 22:29:47 - INFO - codeparrot_training - Step 2903: {'lr': 0.0004995635079215965, 'samples': 557568, 'steps': 2903, 'loss/train': 3.7674472332000732} 01/27/2022 22:29:53 - INFO - codeparrot_training - Step 2904: {'lr': 0.0004995625409086285, 'samples': 557760, 'steps': 2904, 'loss/train': 3.2461023330688477} 01/27/2022 22:29:57 - INFO - codeparrot_training - Step 2905: {'lr': 0.0004995615728266138, 'samples': 557952, 'steps': 2905, 'loss/train': 3.9776540994644165} 01/27/2022 22:30:01 - INFO - codeparrot_training - Step 2906: {'lr': 0.0004995606036755566, 'samples': 558144, 'steps': 2906, 'loss/train': 5.169573783874512} 01/27/2022 22:30:05 - INFO - codeparrot_training - Step 2907: {'lr': 0.000499559633455461, 'samples': 558336, 'steps': 2907, 'loss/train': 3.9271345138549805} 01/27/2022 22:30:10 - INFO - codeparrot_training - Step 2908: {'lr': 0.0004995586621663312, 'samples': 558528, 'steps': 2908, 'loss/train': 4.845523953437805} 01/27/2022 22:30:16 - INFO - codeparrot_training - Step 2909: {'lr': 0.0004995576898081713, 'samples': 558720, 'steps': 2909, 'loss/train': 3.3349721431732178} 01/27/2022 22:30:20 - INFO - codeparrot_training - Step 2910: {'lr': 0.0004995567163809855, 'samples': 558912, 'steps': 2910, 'loss/train': 3.8571313619613647} 01/27/2022 22:30:24 - INFO - codeparrot_training - Step 2911: {'lr': 0.000499555741884778, 'samples': 559104, 'steps': 2911, 'loss/train': 1.968533992767334} 01/27/2022 22:30:29 - INFO - codeparrot_training - Step 2912: {'lr': 0.000499554766319553, 'samples': 559296, 'steps': 2912, 'loss/train': 3.8853471279144287} 01/27/2022 22:30:33 - INFO - codeparrot_training - Step 2913: {'lr': 0.0004995537896853146, 'samples': 559488, 'steps': 2913, 'loss/train': 3.2280964851379395} 01/27/2022 22:30:39 - INFO - codeparrot_training - Step 2914: {'lr': 0.0004995528119820669, 'samples': 559680, 'steps': 2914, 'loss/train': 3.4252874851226807} 01/27/2022 22:30:43 - INFO - codeparrot_training - Step 2915: {'lr': 0.0004995518332098143, 'samples': 559872, 'steps': 2915, 'loss/train': 2.5934152007102966} 01/27/2022 22:30:48 - INFO - codeparrot_training - Step 2916: {'lr': 0.0004995508533685608, 'samples': 560064, 'steps': 2916, 'loss/train': 3.8311471939086914} 01/27/2022 22:30:52 - INFO - codeparrot_training - Step 2917: {'lr': 0.0004995498724583107, 'samples': 560256, 'steps': 2917, 'loss/train': 2.969388484954834} 01/27/2022 22:30:56 - INFO - codeparrot_training - Step 2918: {'lr': 0.0004995488904790682, 'samples': 560448, 'steps': 2918, 'loss/train': 4.2952083349227905} 01/27/2022 22:31:00 - INFO - codeparrot_training - Step 2919: {'lr': 0.0004995479074308375, 'samples': 560640, 'steps': 2919, 'loss/train': 3.7061036825180054} 01/27/2022 22:31:04 - INFO - codeparrot_training - Step 2920: {'lr': 0.0004995469233136228, 'samples': 560832, 'steps': 2920, 'loss/train': 2.7735646963119507} 01/27/2022 22:31:10 - INFO - codeparrot_training - Step 2921: {'lr': 0.0004995459381274284, 'samples': 561024, 'steps': 2921, 'loss/train': 3.543674111366272} 01/27/2022 22:31:14 - INFO - codeparrot_training - Step 2922: {'lr': 0.0004995449518722584, 'samples': 561216, 'steps': 2922, 'loss/train': 1.3542632460594177} 01/27/2022 22:31:18 - INFO - codeparrot_training - Step 2923: {'lr': 0.000499543964548117, 'samples': 561408, 'steps': 2923, 'loss/train': 3.5047380924224854} 01/27/2022 22:31:22 - INFO - codeparrot_training - Step 2924: {'lr': 0.0004995429761550086, 'samples': 561600, 'steps': 2924, 'loss/train': 3.153643012046814} 01/27/2022 22:31:27 - INFO - codeparrot_training - Step 2925: {'lr': 0.0004995419866929373, 'samples': 561792, 'steps': 2925, 'loss/train': 4.4142783880233765} 01/27/2022 22:31:33 - INFO - codeparrot_training - Step 2926: {'lr': 0.0004995409961619073, 'samples': 561984, 'steps': 2926, 'loss/train': 3.1004244089126587} 01/27/2022 22:31:37 - INFO - codeparrot_training - Step 2927: {'lr': 0.0004995400045619229, 'samples': 562176, 'steps': 2927, 'loss/train': 3.62332820892334} 01/27/2022 22:31:41 - INFO - codeparrot_training - Step 2928: {'lr': 0.0004995390118929885, 'samples': 562368, 'steps': 2928, 'loss/train': 2.869180977344513} 01/27/2022 22:31:45 - INFO - codeparrot_training - Step 2929: {'lr': 0.0004995380181551081, 'samples': 562560, 'steps': 2929, 'loss/train': 4.0865782499313354} 01/27/2022 22:31:49 - INFO - codeparrot_training - Step 2930: {'lr': 0.0004995370233482861, 'samples': 562752, 'steps': 2930, 'loss/train': 2.601468801498413} 01/27/2022 22:31:55 - INFO - codeparrot_training - Step 2931: {'lr': 0.0004995360274725267, 'samples': 562944, 'steps': 2931, 'loss/train': 4.557597756385803} 01/27/2022 22:31:59 - INFO - codeparrot_training - Step 2932: {'lr': 0.0004995350305278342, 'samples': 563136, 'steps': 2932, 'loss/train': 2.3992538452148438} 01/27/2022 22:32:03 - INFO - codeparrot_training - Step 2933: {'lr': 0.0004995340325142128, 'samples': 563328, 'steps': 2933, 'loss/train': 3.8933157920837402} 01/27/2022 22:32:07 - INFO - codeparrot_training - Step 2934: {'lr': 0.000499533033431667, 'samples': 563520, 'steps': 2934, 'loss/train': 3.4374157190322876} 01/27/2022 22:32:13 - INFO - codeparrot_training - Step 2935: {'lr': 0.0004995320332802008, 'samples': 563712, 'steps': 2935, 'loss/train': 2.7865546345710754} 01/27/2022 22:32:18 - INFO - codeparrot_training - Step 2936: {'lr': 0.0004995310320598187, 'samples': 563904, 'steps': 2936, 'loss/train': 3.5438597202301025} 01/27/2022 22:32:22 - INFO - codeparrot_training - Step 2937: {'lr': 0.0004995300297705248, 'samples': 564096, 'steps': 2937, 'loss/train': 1.516261339187622} 01/27/2022 22:32:26 - INFO - codeparrot_training - Step 2938: {'lr': 0.0004995290264123235, 'samples': 564288, 'steps': 2938, 'loss/train': 4.1150511503219604} 01/27/2022 22:32:30 - INFO - codeparrot_training - Step 2939: {'lr': 0.0004995280219852192, 'samples': 564480, 'steps': 2939, 'loss/train': 3.870937943458557} 01/27/2022 22:32:35 - INFO - codeparrot_training - Step 2940: {'lr': 0.000499527016489216, 'samples': 564672, 'steps': 2940, 'loss/train': 3.107169270515442} 01/27/2022 22:32:40 - INFO - codeparrot_training - Step 2941: {'lr': 0.0004995260099243182, 'samples': 564864, 'steps': 2941, 'loss/train': 3.7691949605941772} 01/27/2022 22:32:44 - INFO - codeparrot_training - Step 2942: {'lr': 0.0004995250022905303, 'samples': 565056, 'steps': 2942, 'loss/train': 2.5935997366905212} 01/27/2022 22:32:48 - INFO - codeparrot_training - Step 2943: {'lr': 0.0004995239935878565, 'samples': 565248, 'steps': 2943, 'loss/train': 4.232035875320435} 01/27/2022 22:32:52 - INFO - codeparrot_training - Step 2944: {'lr': 0.0004995229838163012, 'samples': 565440, 'steps': 2944, 'loss/train': 3.26818585395813} 01/27/2022 22:32:57 - INFO - codeparrot_training - Step 2945: {'lr': 0.0004995219729758687, 'samples': 565632, 'steps': 2945, 'loss/train': 3.1768230199813843} 01/27/2022 22:33:02 - INFO - codeparrot_training - Step 2946: {'lr': 0.0004995209610665632, 'samples': 565824, 'steps': 2946, 'loss/train': 3.9764442443847656} 01/27/2022 22:33:06 - INFO - codeparrot_training - Step 2947: {'lr': 0.0004995199480883892, 'samples': 566016, 'steps': 2947, 'loss/train': 2.7218459844589233} 01/27/2022 22:33:10 - INFO - codeparrot_training - Step 2948: {'lr': 0.0004995189340413509, 'samples': 566208, 'steps': 2948, 'loss/train': 4.8768779039382935} 01/27/2022 22:33:14 - INFO - codeparrot_training - Step 2949: {'lr': 0.0004995179189254528, 'samples': 566400, 'steps': 2949, 'loss/train': 3.7652066946029663} 01/27/2022 22:33:19 - INFO - codeparrot_training - Step 2950: {'lr': 0.000499516902740699, 'samples': 566592, 'steps': 2950, 'loss/train': 3.852784752845764} 01/27/2022 22:33:24 - INFO - codeparrot_training - Step 2951: {'lr': 0.0004995158854870942, 'samples': 566784, 'steps': 2951, 'loss/train': 2.6714184880256653} 01/27/2022 22:33:28 - INFO - codeparrot_training - Step 2952: {'lr': 0.0004995148671646426, 'samples': 566976, 'steps': 2952, 'loss/train': 4.151181578636169} 01/27/2022 22:33:32 - INFO - codeparrot_training - Step 2953: {'lr': 0.0004995138477733484, 'samples': 567168, 'steps': 2953, 'loss/train': 3.7211873531341553} 01/27/2022 22:33:36 - INFO - codeparrot_training - Step 2954: {'lr': 0.0004995128273132161, 'samples': 567360, 'steps': 2954, 'loss/train': 3.75957190990448} 01/27/2022 22:33:43 - INFO - codeparrot_training - Step 2955: {'lr': 0.0004995118057842502, 'samples': 567552, 'steps': 2955, 'loss/train': 3.513896942138672} 01/27/2022 22:33:47 - INFO - codeparrot_training - Step 2956: {'lr': 0.0004995107831864549, 'samples': 567744, 'steps': 2956, 'loss/train': 2.9298319816589355} 01/27/2022 22:33:51 - INFO - codeparrot_training - Step 2957: {'lr': 0.0004995097595198346, 'samples': 567936, 'steps': 2957, 'loss/train': 4.267438888549805} 01/27/2022 22:33:55 - INFO - codeparrot_training - Step 2958: {'lr': 0.0004995087347843938, 'samples': 568128, 'steps': 2958, 'loss/train': 3.4297406673431396} 01/27/2022 22:34:00 - INFO - codeparrot_training - Step 2959: {'lr': 0.0004995077089801368, 'samples': 568320, 'steps': 2959, 'loss/train': 3.8322240114212036} 01/27/2022 22:34:05 - INFO - codeparrot_training - Step 2960: {'lr': 0.0004995066821070679, 'samples': 568512, 'steps': 2960, 'loss/train': 3.808041214942932} 01/27/2022 22:34:09 - INFO - codeparrot_training - Step 2961: {'lr': 0.0004995056541651917, 'samples': 568704, 'steps': 2961, 'loss/train': 1.616515874862671} 01/27/2022 22:34:14 - INFO - codeparrot_training - Step 2962: {'lr': 0.0004995046251545125, 'samples': 568896, 'steps': 2962, 'loss/train': 1.9280683994293213} 01/27/2022 22:34:18 - INFO - codeparrot_training - Step 2963: {'lr': 0.0004995035950750346, 'samples': 569088, 'steps': 2963, 'loss/train': 4.614165902137756} 01/27/2022 22:34:22 - INFO - codeparrot_training - Step 2964: {'lr': 0.0004995025639267627, 'samples': 569280, 'steps': 2964, 'loss/train': 3.8764389753341675} 01/27/2022 22:34:27 - INFO - codeparrot_training - Step 2965: {'lr': 0.0004995015317097009, 'samples': 569472, 'steps': 2965, 'loss/train': 3.444812536239624} 01/27/2022 22:34:31 - INFO - codeparrot_training - Step 2966: {'lr': 0.0004995004984238538, 'samples': 569664, 'steps': 2966, 'loss/train': 3.615082025527954} 01/27/2022 22:34:36 - INFO - codeparrot_training - Step 2967: {'lr': 0.0004994994640692258, 'samples': 569856, 'steps': 2967, 'loss/train': 4.933567643165588} 01/27/2022 22:34:40 - INFO - codeparrot_training - Step 2968: {'lr': 0.0004994984286458213, 'samples': 570048, 'steps': 2968, 'loss/train': 3.705238938331604} 01/27/2022 22:34:44 - INFO - codeparrot_training - Step 2969: {'lr': 0.0004994973921536447, 'samples': 570240, 'steps': 2969, 'loss/train': 3.572673797607422} 01/27/2022 22:34:51 - INFO - codeparrot_training - Step 2970: {'lr': 0.0004994963545927006, 'samples': 570432, 'steps': 2970, 'loss/train': 3.703235149383545} 01/27/2022 22:34:55 - INFO - codeparrot_training - Step 2971: {'lr': 0.0004994953159629934, 'samples': 570624, 'steps': 2971, 'loss/train': 3.7666114568710327} 01/27/2022 22:34:59 - INFO - codeparrot_training - Step 2972: {'lr': 0.0004994942762645274, 'samples': 570816, 'steps': 2972, 'loss/train': 4.244257807731628} 01/27/2022 22:35:03 - INFO - codeparrot_training - Step 2973: {'lr': 0.000499493235497307, 'samples': 571008, 'steps': 2973, 'loss/train': 3.904697299003601} 01/27/2022 22:35:07 - INFO - codeparrot_training - Step 2974: {'lr': 0.000499492193661337, 'samples': 571200, 'steps': 2974, 'loss/train': 3.2192466259002686} 01/27/2022 22:35:13 - INFO - codeparrot_training - Step 2975: {'lr': 0.0004994911507566216, 'samples': 571392, 'steps': 2975, 'loss/train': 3.4934924840927124} 01/27/2022 22:35:17 - INFO - codeparrot_training - Step 2976: {'lr': 0.0004994901067831654, 'samples': 571584, 'steps': 2976, 'loss/train': 1.793217658996582} 01/27/2022 22:35:21 - INFO - codeparrot_training - Step 2977: {'lr': 0.0004994890617409728, 'samples': 571776, 'steps': 2977, 'loss/train': 3.414219617843628} 01/27/2022 22:35:25 - INFO - codeparrot_training - Step 2978: {'lr': 0.0004994880156300482, 'samples': 571968, 'steps': 2978, 'loss/train': 4.250257730484009} 01/27/2022 22:35:29 - INFO - codeparrot_training - Step 2979: {'lr': 0.0004994869684503962, 'samples': 572160, 'steps': 2979, 'loss/train': 3.072864532470703} 01/27/2022 22:35:36 - INFO - codeparrot_training - Step 2980: {'lr': 0.0004994859202020212, 'samples': 572352, 'steps': 2980, 'loss/train': 3.926766872406006} 01/27/2022 22:35:40 - INFO - codeparrot_training - Step 2981: {'lr': 0.0004994848708849279, 'samples': 572544, 'steps': 2981, 'loss/train': 3.2524412870407104} 01/27/2022 22:35:44 - INFO - codeparrot_training - Step 2982: {'lr': 0.0004994838204991205, 'samples': 572736, 'steps': 2982, 'loss/train': 3.1980990171432495} 01/27/2022 22:35:48 - INFO - codeparrot_training - Step 2983: {'lr': 0.0004994827690446036, 'samples': 572928, 'steps': 2983, 'loss/train': 3.213517427444458} 01/27/2022 22:35:52 - INFO - codeparrot_training - Step 2984: {'lr': 0.0004994817165213817, 'samples': 573120, 'steps': 2984, 'loss/train': 4.144993185997009} 01/27/2022 22:35:58 - INFO - codeparrot_training - Step 2985: {'lr': 0.0004994806629294594, 'samples': 573312, 'steps': 2985, 'loss/train': 5.2044432163238525} 01/27/2022 22:36:02 - INFO - codeparrot_training - Step 2986: {'lr': 0.0004994796082688413, 'samples': 573504, 'steps': 2986, 'loss/train': 2.3955450654029846} 01/27/2022 22:36:06 - INFO - codeparrot_training - Step 2987: {'lr': 0.0004994785525395316, 'samples': 573696, 'steps': 2987, 'loss/train': 3.572081208229065} 01/27/2022 22:36:10 - INFO - codeparrot_training - Step 2988: {'lr': 0.0004994774957415351, 'samples': 573888, 'steps': 2988, 'loss/train': 3.441437244415283} 01/27/2022 22:36:15 - INFO - codeparrot_training - Step 2989: {'lr': 0.0004994764378748562, 'samples': 574080, 'steps': 2989, 'loss/train': 4.187154650688171} 01/27/2022 22:36:20 - INFO - codeparrot_training - Step 2990: {'lr': 0.0004994753789394994, 'samples': 574272, 'steps': 2990, 'loss/train': 4.300530910491943} 01/27/2022 22:36:24 - INFO - codeparrot_training - Step 2991: {'lr': 0.0004994743189354694, 'samples': 574464, 'steps': 2991, 'loss/train': 4.1977046728134155} 01/27/2022 22:36:28 - INFO - codeparrot_training - Step 2992: {'lr': 0.0004994732578627706, 'samples': 574656, 'steps': 2992, 'loss/train': 3.907742500305176} 01/27/2022 22:36:33 - INFO - codeparrot_training - Step 2993: {'lr': 0.0004994721957214076, 'samples': 574848, 'steps': 2993, 'loss/train': 2.7197002172470093} 01/27/2022 22:36:37 - INFO - codeparrot_training - Step 2994: {'lr': 0.0004994711325113849, 'samples': 575040, 'steps': 2994, 'loss/train': 3.629009485244751} 01/27/2022 22:36:42 - INFO - codeparrot_training - Step 2995: {'lr': 0.000499470068232707, 'samples': 575232, 'steps': 2995, 'loss/train': 3.810677647590637} 01/27/2022 22:36:46 - INFO - codeparrot_training - Step 2996: {'lr': 0.0004994690028853787, 'samples': 575424, 'steps': 2996, 'loss/train': 4.0300551652908325} 01/27/2022 22:36:51 - INFO - codeparrot_training - Step 2997: {'lr': 0.0004994679364694043, 'samples': 575616, 'steps': 2997, 'loss/train': 2.3457874059677124} 01/27/2022 22:36:55 - INFO - codeparrot_training - Step 2998: {'lr': 0.0004994668689847885, 'samples': 575808, 'steps': 2998, 'loss/train': 3.075886845588684} 01/27/2022 22:36:59 - INFO - codeparrot_training - Step 2999: {'lr': 0.0004994658004315358, 'samples': 576000, 'steps': 2999, 'loss/train': 3.6470274925231934} 01/27/2022 22:37:05 - INFO - codeparrot_training - Step 3000: {'lr': 0.0004994647308096509, 'samples': 576192, 'steps': 3000, 'loss/train': 4.099261164665222} 01/27/2022 22:37:09 - INFO - codeparrot_training - Step 3001: {'lr': 0.0004994636601191383, 'samples': 576384, 'steps': 3001, 'loss/train': 3.8695517778396606} 01/27/2022 22:37:14 - INFO - codeparrot_training - Step 3002: {'lr': 0.0004994625883600025, 'samples': 576576, 'steps': 3002, 'loss/train': 3.0125287771224976} 01/27/2022 22:37:18 - INFO - codeparrot_training - Step 3003: {'lr': 0.0004994615155322483, 'samples': 576768, 'steps': 3003, 'loss/train': 4.476115107536316} 01/27/2022 22:37:22 - INFO - codeparrot_training - Step 3004: {'lr': 0.0004994604416358801, 'samples': 576960, 'steps': 3004, 'loss/train': 3.286051869392395} 01/27/2022 22:37:27 - INFO - codeparrot_training - Step 3005: {'lr': 0.0004994593666709027, 'samples': 577152, 'steps': 3005, 'loss/train': 4.241188287734985} 01/27/2022 22:37:31 - INFO - codeparrot_training - Step 3006: {'lr': 0.0004994582906373205, 'samples': 577344, 'steps': 3006, 'loss/train': 2.6923046708106995} 01/27/2022 22:37:36 - INFO - codeparrot_training - Step 3007: {'lr': 0.0004994572135351382, 'samples': 577536, 'steps': 3007, 'loss/train': 3.8148103952407837} 01/27/2022 22:37:40 - INFO - codeparrot_training - Step 3008: {'lr': 0.0004994561353643604, 'samples': 577728, 'steps': 3008, 'loss/train': 3.338180422782898} 01/27/2022 22:37:44 - INFO - codeparrot_training - Step 3009: {'lr': 0.0004994550561249917, 'samples': 577920, 'steps': 3009, 'loss/train': 3.4229832887649536} 01/27/2022 22:37:50 - INFO - codeparrot_training - Step 3010: {'lr': 0.0004994539758170367, 'samples': 578112, 'steps': 3010, 'loss/train': 4.0032044649124146} 01/27/2022 22:37:54 - INFO - codeparrot_training - Step 3011: {'lr': 0.0004994528944405002, 'samples': 578304, 'steps': 3011, 'loss/train': 3.9481945037841797} 01/27/2022 22:37:58 - INFO - codeparrot_training - Step 3012: {'lr': 0.0004994518119953867, 'samples': 578496, 'steps': 3012, 'loss/train': 3.9148975610733032} 01/27/2022 22:38:02 - INFO - codeparrot_training - Step 3013: {'lr': 0.0004994507284817009, 'samples': 578688, 'steps': 3013, 'loss/train': 4.116961598396301} 01/27/2022 22:38:06 - INFO - codeparrot_training - Step 3014: {'lr': 0.0004994496438994472, 'samples': 578880, 'steps': 3014, 'loss/train': 3.170509457588196} 01/27/2022 22:38:12 - INFO - codeparrot_training - Step 3015: {'lr': 0.0004994485582486306, 'samples': 579072, 'steps': 3015, 'loss/train': 3.763291597366333} 01/27/2022 22:38:17 - INFO - codeparrot_training - Step 3016: {'lr': 0.0004994474715292555, 'samples': 579264, 'steps': 3016, 'loss/train': 2.427133798599243} 01/27/2022 22:38:21 - INFO - codeparrot_training - Step 3017: {'lr': 0.0004994463837413268, 'samples': 579456, 'steps': 3017, 'loss/train': 3.513057589530945} 01/27/2022 22:38:25 - INFO - codeparrot_training - Step 3018: {'lr': 0.0004994452948848488, 'samples': 579648, 'steps': 3018, 'loss/train': 5.69235098361969} 01/27/2022 22:38:29 - INFO - codeparrot_training - Step 3019: {'lr': 0.0004994442049598265, 'samples': 579840, 'steps': 3019, 'loss/train': 3.9578261375427246} 01/27/2022 22:38:35 - INFO - codeparrot_training - Step 3020: {'lr': 0.0004994431139662643, 'samples': 580032, 'steps': 3020, 'loss/train': 2.6812461018562317} 01/27/2022 22:38:39 - INFO - codeparrot_training - Step 3021: {'lr': 0.0004994420219041671, 'samples': 580224, 'steps': 3021, 'loss/train': 2.8891228437423706} 01/27/2022 22:38:43 - INFO - codeparrot_training - Step 3022: {'lr': 0.0004994409287735394, 'samples': 580416, 'steps': 3022, 'loss/train': 4.094055891036987} 01/27/2022 22:38:47 - INFO - codeparrot_training - Step 3023: {'lr': 0.0004994398345743861, 'samples': 580608, 'steps': 3023, 'loss/train': 2.7287017107009888} 01/27/2022 22:38:51 - INFO - codeparrot_training - Step 3024: {'lr': 0.0004994387393067117, 'samples': 580800, 'steps': 3024, 'loss/train': 3.077560544013977} 01/27/2022 22:38:57 - INFO - codeparrot_training - Step 3025: {'lr': 0.0004994376429705208, 'samples': 580992, 'steps': 3025, 'loss/train': 4.408345341682434} 01/27/2022 22:39:01 - INFO - codeparrot_training - Step 3026: {'lr': 0.0004994365455658185, 'samples': 581184, 'steps': 3026, 'loss/train': 3.332540273666382} 01/27/2022 22:39:06 - INFO - codeparrot_training - Step 3027: {'lr': 0.000499435447092609, 'samples': 581376, 'steps': 3027, 'loss/train': 2.706877648830414} 01/27/2022 22:39:10 - INFO - codeparrot_training - Step 3028: {'lr': 0.0004994343475508974, 'samples': 581568, 'steps': 3028, 'loss/train': 3.805307149887085} 01/27/2022 22:39:14 - INFO - codeparrot_training - Step 3029: {'lr': 0.0004994332469406882, 'samples': 581760, 'steps': 3029, 'loss/train': 1.6086958050727844} 01/27/2022 22:39:19 - INFO - codeparrot_training - Step 3030: {'lr': 0.0004994321452619863, 'samples': 581952, 'steps': 3030, 'loss/train': 4.078457236289978} 01/27/2022 22:39:24 - INFO - codeparrot_training - Step 3031: {'lr': 0.0004994310425147962, 'samples': 582144, 'steps': 3031, 'loss/train': 3.533213496208191} 01/27/2022 22:39:28 - INFO - codeparrot_training - Step 3032: {'lr': 0.0004994299386991227, 'samples': 582336, 'steps': 3032, 'loss/train': 2.900184452533722} 01/27/2022 22:39:32 - INFO - codeparrot_training - Step 3033: {'lr': 0.0004994288338149705, 'samples': 582528, 'steps': 3033, 'loss/train': 3.541010856628418} 01/27/2022 22:39:36 - INFO - codeparrot_training - Step 3034: {'lr': 0.0004994277278623445, 'samples': 582720, 'steps': 3034, 'loss/train': 2.6927268505096436} 01/27/2022 22:39:42 - INFO - codeparrot_training - Step 3035: {'lr': 0.0004994266208412493, 'samples': 582912, 'steps': 3035, 'loss/train': 3.4738608598709106} 01/27/2022 22:39:46 - INFO - codeparrot_training - Step 3036: {'lr': 0.0004994255127516895, 'samples': 583104, 'steps': 3036, 'loss/train': 2.2878000140190125} 01/27/2022 22:39:50 - INFO - codeparrot_training - Step 3037: {'lr': 0.0004994244035936701, 'samples': 583296, 'steps': 3037, 'loss/train': 1.9735316634178162} 01/27/2022 22:39:54 - INFO - codeparrot_training - Step 3038: {'lr': 0.0004994232933671958, 'samples': 583488, 'steps': 3038, 'loss/train': 4.475515365600586} 01/27/2022 22:39:59 - INFO - codeparrot_training - Step 3039: {'lr': 0.0004994221820722713, 'samples': 583680, 'steps': 3039, 'loss/train': 1.9421718120574951} 01/27/2022 22:40:05 - INFO - codeparrot_training - Step 3040: {'lr': 0.0004994210697089013, 'samples': 583872, 'steps': 3040, 'loss/train': 3.6205755472183228} 01/27/2022 22:40:09 - INFO - codeparrot_training - Step 3041: {'lr': 0.0004994199562770907, 'samples': 584064, 'steps': 3041, 'loss/train': 1.8807876706123352} 01/27/2022 22:40:14 - INFO - codeparrot_training - Step 3042: {'lr': 0.0004994188417768443, 'samples': 584256, 'steps': 3042, 'loss/train': 2.0049217343330383} 01/27/2022 22:40:18 - INFO - codeparrot_training - Step 3043: {'lr': 0.0004994177262081666, 'samples': 584448, 'steps': 3043, 'loss/train': 3.9147130250930786} 01/27/2022 22:40:22 - INFO - codeparrot_training - Step 3044: {'lr': 0.0004994166095710626, 'samples': 584640, 'steps': 3044, 'loss/train': 3.74419105052948} 01/27/2022 22:40:28 - INFO - codeparrot_training - Step 3045: {'lr': 0.0004994154918655371, 'samples': 584832, 'steps': 3045, 'loss/train': 2.5503235459327698} 01/27/2022 22:40:32 - INFO - codeparrot_training - Step 3046: {'lr': 0.0004994143730915948, 'samples': 585024, 'steps': 3046, 'loss/train': 3.486863851547241} 01/27/2022 22:40:36 - INFO - codeparrot_training - Step 3047: {'lr': 0.0004994132532492406, 'samples': 585216, 'steps': 3047, 'loss/train': 4.9429110288619995} 01/27/2022 22:40:41 - INFO - codeparrot_training - Step 3048: {'lr': 0.0004994121323384791, 'samples': 585408, 'steps': 3048, 'loss/train': 3.881789445877075} 01/27/2022 22:40:45 - INFO - codeparrot_training - Step 3049: {'lr': 0.0004994110103593154, 'samples': 585600, 'steps': 3049, 'loss/train': 2.7051275968551636} 01/27/2022 22:40:49 - INFO - codeparrot_training - Step 3050: {'lr': 0.0004994098873117539, 'samples': 585792, 'steps': 3050, 'loss/train': 4.348789572715759} 01/27/2022 22:40:54 - INFO - codeparrot_training - Step 3051: {'lr': 0.0004994087631957998, 'samples': 585984, 'steps': 3051, 'loss/train': 4.234740972518921} 01/27/2022 22:40:59 - INFO - codeparrot_training - Step 3052: {'lr': 0.0004994076380114577, 'samples': 586176, 'steps': 3052, 'loss/train': 3.7790404558181763} 01/27/2022 22:41:03 - INFO - codeparrot_training - Step 3053: {'lr': 0.0004994065117587325, 'samples': 586368, 'steps': 3053, 'loss/train': 3.8583219051361084} 01/27/2022 22:41:07 - INFO - codeparrot_training - Step 3054: {'lr': 0.0004994053844376289, 'samples': 586560, 'steps': 3054, 'loss/train': 2.9955881237983704} 01/27/2022 22:41:12 - INFO - codeparrot_training - Step 3055: {'lr': 0.000499404256048152, 'samples': 586752, 'steps': 3055, 'loss/train': 3.1907533407211304} 01/27/2022 22:41:16 - INFO - codeparrot_training - Step 3056: {'lr': 0.0004994031265903063, 'samples': 586944, 'steps': 3056, 'loss/train': 4.42550003528595} 01/27/2022 22:41:21 - INFO - codeparrot_training - Step 3057: {'lr': 0.0004994019960640969, 'samples': 587136, 'steps': 3057, 'loss/train': 1.1688518822193146} 01/27/2022 22:41:25 - INFO - codeparrot_training - Step 3058: {'lr': 0.0004994008644695285, 'samples': 587328, 'steps': 3058, 'loss/train': 2.9855749011039734} 01/27/2022 22:41:29 - INFO - codeparrot_training - Step 3059: {'lr': 0.0004993997318066061, 'samples': 587520, 'steps': 3059, 'loss/train': 3.9997562170028687} 01/27/2022 22:41:35 - INFO - codeparrot_training - Step 3060: {'lr': 0.0004993985980753342, 'samples': 587712, 'steps': 3060, 'loss/train': 3.641914486885071} 01/27/2022 22:41:39 - INFO - codeparrot_training - Step 3061: {'lr': 0.0004993974632757181, 'samples': 587904, 'steps': 3061, 'loss/train': 3.582821488380432} 01/27/2022 22:41:43 - INFO - codeparrot_training - Step 3062: {'lr': 0.0004993963274077624, 'samples': 588096, 'steps': 3062, 'loss/train': 3.8367737531661987} 01/27/2022 22:41:48 - INFO - codeparrot_training - Step 3063: {'lr': 0.000499395190471472, 'samples': 588288, 'steps': 3063, 'loss/train': 3.2059510946273804} 01/27/2022 22:41:52 - INFO - codeparrot_training - Step 3064: {'lr': 0.0004993940524668518, 'samples': 588480, 'steps': 3064, 'loss/train': 1.8709221482276917} 01/27/2022 22:41:58 - INFO - codeparrot_training - Step 3065: {'lr': 0.0004993929133939067, 'samples': 588672, 'steps': 3065, 'loss/train': 4.3786139488220215} 01/27/2022 22:42:02 - INFO - codeparrot_training - Step 3066: {'lr': 0.0004993917732526416, 'samples': 588864, 'steps': 3066, 'loss/train': 3.2850261926651} 01/27/2022 22:42:06 - INFO - codeparrot_training - Step 3067: {'lr': 0.0004993906320430613, 'samples': 589056, 'steps': 3067, 'loss/train': 4.062875747680664} 01/27/2022 22:42:11 - INFO - codeparrot_training - Step 3068: {'lr': 0.0004993894897651706, 'samples': 589248, 'steps': 3068, 'loss/train': 3.65367329120636} 01/27/2022 22:42:15 - INFO - codeparrot_training - Step 3069: {'lr': 0.0004993883464189747, 'samples': 589440, 'steps': 3069, 'loss/train': 3.530601382255554} 01/27/2022 22:42:19 - INFO - codeparrot_training - Step 3070: {'lr': 0.0004993872020044781, 'samples': 589632, 'steps': 3070, 'loss/train': 2.6960513591766357} 01/27/2022 22:42:24 - INFO - codeparrot_training - Step 3071: {'lr': 0.0004993860565216861, 'samples': 589824, 'steps': 3071, 'loss/train': 3.976522207260132} 01/27/2022 22:42:28 - INFO - codeparrot_training - Step 3072: {'lr': 0.0004993849099706034, 'samples': 590016, 'steps': 3072, 'loss/train': 3.3037497997283936} 01/27/2022 22:42:33 - INFO - codeparrot_training - Step 3073: {'lr': 0.0004993837623512349, 'samples': 590208, 'steps': 3073, 'loss/train': 1.5774094462394714} 01/27/2022 22:42:37 - INFO - codeparrot_training - Step 3074: {'lr': 0.0004993826136635856, 'samples': 590400, 'steps': 3074, 'loss/train': 4.250776290893555} 01/27/2022 22:42:41 - INFO - codeparrot_training - Step 3075: {'lr': 0.0004993814639076602, 'samples': 590592, 'steps': 3075, 'loss/train': 4.688954830169678} 01/27/2022 22:42:47 - INFO - codeparrot_training - Step 3076: {'lr': 0.000499380313083464, 'samples': 590784, 'steps': 3076, 'loss/train': 3.3747775554656982} 01/27/2022 22:42:51 - INFO - codeparrot_training - Step 3077: {'lr': 0.0004993791611910017, 'samples': 590976, 'steps': 3077, 'loss/train': 4.090391993522644} 01/27/2022 22:42:55 - INFO - codeparrot_training - Step 3078: {'lr': 0.0004993780082302782, 'samples': 591168, 'steps': 3078, 'loss/train': 2.0767135620117188} 01/27/2022 22:43:00 - INFO - codeparrot_training - Step 3079: {'lr': 0.0004993768542012985, 'samples': 591360, 'steps': 3079, 'loss/train': 2.4533947706222534} 01/27/2022 22:43:04 - INFO - codeparrot_training - Step 3080: {'lr': 0.0004993756991040675, 'samples': 591552, 'steps': 3080, 'loss/train': 3.5006500482559204} 01/27/2022 22:43:09 - INFO - codeparrot_training - Step 3081: {'lr': 0.0004993745429385903, 'samples': 591744, 'steps': 3081, 'loss/train': 3.177215337753296} 01/27/2022 22:43:13 - INFO - codeparrot_training - Step 3082: {'lr': 0.0004993733857048717, 'samples': 591936, 'steps': 3082, 'loss/train': 4.526736259460449} 01/27/2022 22:43:18 - INFO - codeparrot_training - Step 3083: {'lr': 0.0004993722274029167, 'samples': 592128, 'steps': 3083, 'loss/train': 3.339797258377075} 01/27/2022 22:43:22 - INFO - codeparrot_training - Step 3084: {'lr': 0.0004993710680327301, 'samples': 592320, 'steps': 3084, 'loss/train': 2.7436006665229797} 01/27/2022 22:43:26 - INFO - codeparrot_training - Step 3085: {'lr': 0.0004993699075943172, 'samples': 592512, 'steps': 3085, 'loss/train': 2.7680971026420593} 01/27/2022 22:43:32 - INFO - codeparrot_training - Step 3086: {'lr': 0.0004993687460876829, 'samples': 592704, 'steps': 3086, 'loss/train': 3.6883556842803955} 01/27/2022 22:43:36 - INFO - codeparrot_training - Step 3087: {'lr': 0.0004993675835128319, 'samples': 592896, 'steps': 3087, 'loss/train': 3.3348881006240845} 01/27/2022 22:43:40 - INFO - codeparrot_training - Step 3088: {'lr': 0.0004993664198697694, 'samples': 593088, 'steps': 3088, 'loss/train': 3.6740258932113647} 01/27/2022 22:43:45 - INFO - codeparrot_training - Step 3089: {'lr': 0.0004993652551585003, 'samples': 593280, 'steps': 3089, 'loss/train': 2.9839687943458557} 01/27/2022 22:43:49 - INFO - codeparrot_training - Step 3090: {'lr': 0.0004993640893790298, 'samples': 593472, 'steps': 3090, 'loss/train': 3.6550605297088623} 01/27/2022 22:43:54 - INFO - codeparrot_training - Step 3091: {'lr': 0.0004993629225313625, 'samples': 593664, 'steps': 3091, 'loss/train': 3.985403537750244} 01/27/2022 22:43:58 - INFO - codeparrot_training - Step 3092: {'lr': 0.0004993617546155037, 'samples': 593856, 'steps': 3092, 'loss/train': 2.7620463967323303} 01/27/2022 22:44:03 - INFO - codeparrot_training - Step 3093: {'lr': 0.0004993605856314584, 'samples': 594048, 'steps': 3093, 'loss/train': 3.9262797832489014} 01/27/2022 22:44:07 - INFO - codeparrot_training - Step 3094: {'lr': 0.0004993594155792315, 'samples': 594240, 'steps': 3094, 'loss/train': 3.621980309486389} 01/27/2022 22:44:11 - INFO - codeparrot_training - Step 3095: {'lr': 0.000499358244458828, 'samples': 594432, 'steps': 3095, 'loss/train': 2.1281332969665527} 01/27/2022 22:44:16 - INFO - codeparrot_training - Step 3096: {'lr': 0.0004993570722702529, 'samples': 594624, 'steps': 3096, 'loss/train': 3.3281736373901367} 01/27/2022 22:44:21 - INFO - codeparrot_training - Step 3097: {'lr': 0.0004993558990135115, 'samples': 594816, 'steps': 3097, 'loss/train': 2.9377437829971313} 01/27/2022 22:44:25 - INFO - codeparrot_training - Step 3098: {'lr': 0.0004993547246886084, 'samples': 595008, 'steps': 3098, 'loss/train': 3.716862916946411} 01/27/2022 22:44:29 - INFO - codeparrot_training - Step 3099: {'lr': 0.0004993535492955488, 'samples': 595200, 'steps': 3099, 'loss/train': 3.655521869659424} 01/27/2022 22:44:33 - INFO - codeparrot_training - Step 3100: {'lr': 0.000499352372834338, 'samples': 595392, 'steps': 3100, 'loss/train': 3.7478599548339844} 01/27/2022 22:44:39 - INFO - codeparrot_training - Step 3101: {'lr': 0.0004993511953049807, 'samples': 595584, 'steps': 3101, 'loss/train': 4.10179603099823} 01/27/2022 22:44:43 - INFO - codeparrot_training - Step 3102: {'lr': 0.000499350016707482, 'samples': 595776, 'steps': 3102, 'loss/train': 3.328495502471924} 01/27/2022 22:44:47 - INFO - codeparrot_training - Step 3103: {'lr': 0.0004993488370418471, 'samples': 595968, 'steps': 3103, 'loss/train': 3.5054279565811157} 01/27/2022 22:44:52 - INFO - codeparrot_training - Step 3104: {'lr': 0.0004993476563080809, 'samples': 596160, 'steps': 3104, 'loss/train': 3.9740102291107178} 01/27/2022 22:44:56 - INFO - codeparrot_training - Step 3105: {'lr': 0.0004993464745061885, 'samples': 596352, 'steps': 3105, 'loss/train': 6.576958894729614} 01/27/2022 22:45:01 - INFO - codeparrot_training - Step 3106: {'lr': 0.0004993452916361751, 'samples': 596544, 'steps': 3106, 'loss/train': 2.818505644798279} 01/27/2022 22:45:05 - INFO - codeparrot_training - Step 3107: {'lr': 0.0004993441076980455, 'samples': 596736, 'steps': 3107, 'loss/train': 3.472429633140564} 01/27/2022 22:45:09 - INFO - codeparrot_training - Step 3108: {'lr': 0.0004993429226918051, 'samples': 596928, 'steps': 3108, 'loss/train': 4.0000526905059814} 01/27/2022 22:45:14 - INFO - codeparrot_training - Step 3109: {'lr': 0.0004993417366174586, 'samples': 597120, 'steps': 3109, 'loss/train': 3.1740692853927612} 01/27/2022 22:45:18 - INFO - codeparrot_training - Step 3110: {'lr': 0.0004993405494750113, 'samples': 597312, 'steps': 3110, 'loss/train': 3.8743293285369873} 01/27/2022 22:45:23 - INFO - codeparrot_training - Step 3111: {'lr': 0.0004993393612644683, 'samples': 597504, 'steps': 3111, 'loss/train': 4.11640191078186} 01/27/2022 22:45:27 - INFO - codeparrot_training - Step 3112: {'lr': 0.0004993381719858347, 'samples': 597696, 'steps': 3112, 'loss/train': 4.615322470664978} 01/27/2022 22:45:31 - INFO - codeparrot_training - Step 3113: {'lr': 0.0004993369816391156, 'samples': 597888, 'steps': 3113, 'loss/train': 3.90139639377594} 01/27/2022 22:45:36 - INFO - codeparrot_training - Step 3114: {'lr': 0.0004993357902243158, 'samples': 598080, 'steps': 3114, 'loss/train': 4.505835771560669} 01/27/2022 22:45:40 - INFO - codeparrot_training - Step 3115: {'lr': 0.0004993345977414408, 'samples': 598272, 'steps': 3115, 'loss/train': 3.4591695070266724} 01/27/2022 22:45:45 - INFO - codeparrot_training - Step 3116: {'lr': 0.0004993334041904957, 'samples': 598464, 'steps': 3116, 'loss/train': 3.745135188102722} 01/27/2022 22:45:49 - INFO - codeparrot_training - Step 3117: {'lr': 0.0004993322095714853, 'samples': 598656, 'steps': 3117, 'loss/train': 3.3211737871170044} 01/27/2022 22:45:53 - INFO - codeparrot_training - Step 3118: {'lr': 0.0004993310138844149, 'samples': 598848, 'steps': 3118, 'loss/train': 3.964070677757263} 01/27/2022 22:45:58 - INFO - codeparrot_training - Step 3119: {'lr': 0.0004993298171292896, 'samples': 599040, 'steps': 3119, 'loss/train': 4.374395728111267} 01/27/2022 22:46:02 - INFO - codeparrot_training - Step 3120: {'lr': 0.0004993286193061145, 'samples': 599232, 'steps': 3120, 'loss/train': 3.928914785385132} 01/27/2022 22:46:08 - INFO - codeparrot_training - Step 3121: {'lr': 0.0004993274204148949, 'samples': 599424, 'steps': 3121, 'loss/train': 2.9195226430892944} 01/27/2022 22:46:12 - INFO - codeparrot_training - Step 3122: {'lr': 0.0004993262204556356, 'samples': 599616, 'steps': 3122, 'loss/train': 4.014883875846863} 01/27/2022 22:46:16 - INFO - codeparrot_training - Step 3123: {'lr': 0.0004993250194283421, 'samples': 599808, 'steps': 3123, 'loss/train': 3.6228500604629517} 01/27/2022 22:46:21 - INFO - codeparrot_training - Step 3124: {'lr': 0.0004993238173330194, 'samples': 600000, 'steps': 3124, 'loss/train': 4.200058937072754} 01/27/2022 22:46:25 - INFO - codeparrot_training - Step 3125: {'lr': 0.0004993226141696725, 'samples': 600192, 'steps': 3125, 'loss/train': 3.367281675338745} 01/27/2022 22:46:30 - INFO - codeparrot_training - Step 3126: {'lr': 0.0004993214099383069, 'samples': 600384, 'steps': 3126, 'loss/train': 3.7214176654815674} 01/27/2022 22:46:34 - INFO - codeparrot_training - Step 3127: {'lr': 0.0004993202046389274, 'samples': 600576, 'steps': 3127, 'loss/train': 3.6612993478775024} 01/27/2022 22:46:38 - INFO - codeparrot_training - Step 3128: {'lr': 0.0004993189982715392, 'samples': 600768, 'steps': 3128, 'loss/train': 3.5526645183563232} 01/27/2022 22:46:42 - INFO - codeparrot_training - Step 3129: {'lr': 0.0004993177908361479, 'samples': 600960, 'steps': 3129, 'loss/train': 3.3352417945861816} 01/27/2022 22:46:47 - INFO - codeparrot_training - Step 3130: {'lr': 0.000499316582332758, 'samples': 601152, 'steps': 3130, 'loss/train': 3.4543944597244263} 01/27/2022 22:46:53 - INFO - codeparrot_training - Step 3131: {'lr': 0.0004993153727613753, 'samples': 601344, 'steps': 3131, 'loss/train': 4.843121409416199} 01/27/2022 22:46:57 - INFO - codeparrot_training - Step 3132: {'lr': 0.0004993141621220046, 'samples': 601536, 'steps': 3132, 'loss/train': 3.5010502338409424} 01/27/2022 22:47:01 - INFO - codeparrot_training - Step 3133: {'lr': 0.0004993129504146512, 'samples': 601728, 'steps': 3133, 'loss/train': 3.879149079322815} 01/27/2022 22:47:05 - INFO - codeparrot_training - Step 3134: {'lr': 0.0004993117376393203, 'samples': 601920, 'steps': 3134, 'loss/train': 4.798769116401672} 01/27/2022 22:47:09 - INFO - codeparrot_training - Step 3135: {'lr': 0.000499310523796017, 'samples': 602112, 'steps': 3135, 'loss/train': 2.993631362915039} 01/27/2022 22:47:14 - INFO - codeparrot_training - Step 3136: {'lr': 0.0004993093088847466, 'samples': 602304, 'steps': 3136, 'loss/train': 3.9609718322753906} 01/27/2022 22:47:19 - INFO - codeparrot_training - Step 3137: {'lr': 0.0004993080929055144, 'samples': 602496, 'steps': 3137, 'loss/train': 2.3357144594192505} 01/27/2022 22:47:23 - INFO - codeparrot_training - Step 3138: {'lr': 0.0004993068758583254, 'samples': 602688, 'steps': 3138, 'loss/train': 4.331887364387512} 01/27/2022 22:47:27 - INFO - codeparrot_training - Step 3139: {'lr': 0.0004993056577431849, 'samples': 602880, 'steps': 3139, 'loss/train': 3.8063879013061523} 01/27/2022 22:47:31 - INFO - codeparrot_training - Step 3140: {'lr': 0.0004993044385600982, 'samples': 603072, 'steps': 3140, 'loss/train': 4.360708594322205} 01/27/2022 22:47:37 - INFO - codeparrot_training - Step 3141: {'lr': 0.0004993032183090704, 'samples': 603264, 'steps': 3141, 'loss/train': 3.1006107330322266} 01/27/2022 22:47:41 - INFO - codeparrot_training - Step 3142: {'lr': 0.0004993019969901069, 'samples': 603456, 'steps': 3142, 'loss/train': 3.417270541191101} 01/27/2022 22:47:45 - INFO - codeparrot_training - Step 3143: {'lr': 0.0004993007746032126, 'samples': 603648, 'steps': 3143, 'loss/train': 3.6055076122283936} 01/27/2022 22:47:49 - INFO - codeparrot_training - Step 3144: {'lr': 0.000499299551148393, 'samples': 603840, 'steps': 3144, 'loss/train': 3.336031436920166} 01/27/2022 22:47:53 - INFO - codeparrot_training - Step 3145: {'lr': 0.0004992983266256533, 'samples': 604032, 'steps': 3145, 'loss/train': 5.530380606651306} 01/27/2022 22:47:59 - INFO - codeparrot_training - Step 3146: {'lr': 0.0004992971010349987, 'samples': 604224, 'steps': 3146, 'loss/train': 1.4570490717887878} 01/27/2022 22:48:04 - INFO - codeparrot_training - Step 3147: {'lr': 0.0004992958743764346, 'samples': 604416, 'steps': 3147, 'loss/train': 3.4211522340774536} 01/27/2022 22:48:08 - INFO - codeparrot_training - Step 3148: {'lr': 0.0004992946466499661, 'samples': 604608, 'steps': 3148, 'loss/train': 3.1918108463287354} 01/27/2022 22:48:12 - INFO - codeparrot_training - Step 3149: {'lr': 0.0004992934178555984, 'samples': 604800, 'steps': 3149, 'loss/train': 1.8312726616859436} 01/27/2022 22:48:16 - INFO - codeparrot_training - Step 3150: {'lr': 0.000499292187993337, 'samples': 604992, 'steps': 3150, 'loss/train': 3.971079468727112} 01/27/2022 22:48:21 - INFO - codeparrot_training - Step 3151: {'lr': 0.0004992909570631868, 'samples': 605184, 'steps': 3151, 'loss/train': 2.837827205657959} 01/27/2022 22:48:26 - INFO - codeparrot_training - Step 3152: {'lr': 0.0004992897250651535, 'samples': 605376, 'steps': 3152, 'loss/train': 3.3917956352233887} 01/27/2022 22:48:30 - INFO - codeparrot_training - Step 3153: {'lr': 0.0004992884919992421, 'samples': 605568, 'steps': 3153, 'loss/train': 4.141369700431824} 01/27/2022 22:48:34 - INFO - codeparrot_training - Step 3154: {'lr': 0.000499287257865458, 'samples': 605760, 'steps': 3154, 'loss/train': 2.868853211402893} 01/27/2022 22:48:38 - INFO - codeparrot_training - Step 3155: {'lr': 0.0004992860226638064, 'samples': 605952, 'steps': 3155, 'loss/train': 3.1843396425247192} 01/27/2022 22:48:44 - INFO - codeparrot_training - Step 3156: {'lr': 0.0004992847863942927, 'samples': 606144, 'steps': 3156, 'loss/train': 3.8941104412078857} 01/27/2022 22:48:48 - INFO - codeparrot_training - Step 3157: {'lr': 0.000499283549056922, 'samples': 606336, 'steps': 3157, 'loss/train': 3.139754891395569} 01/27/2022 22:48:52 - INFO - codeparrot_training - Step 3158: {'lr': 0.0004992823106516999, 'samples': 606528, 'steps': 3158, 'loss/train': 2.511710286140442} 01/27/2022 22:48:57 - INFO - codeparrot_training - Step 3159: {'lr': 0.0004992810711786314, 'samples': 606720, 'steps': 3159, 'loss/train': 4.265854239463806} 01/27/2022 22:49:01 - INFO - codeparrot_training - Step 3160: {'lr': 0.000499279830637722, 'samples': 606912, 'steps': 3160, 'loss/train': 3.902389883995056} 01/27/2022 22:49:06 - INFO - codeparrot_training - Step 3161: {'lr': 0.000499278589028977, 'samples': 607104, 'steps': 3161, 'loss/train': 3.587753176689148} 01/27/2022 22:49:10 - INFO - codeparrot_training - Step 3162: {'lr': 0.0004992773463524016, 'samples': 607296, 'steps': 3162, 'loss/train': 3.339136004447937} 01/27/2022 22:49:14 - INFO - codeparrot_training - Step 3163: {'lr': 0.0004992761026080013, 'samples': 607488, 'steps': 3163, 'loss/train': 3.2873003482818604} 01/27/2022 22:49:18 - INFO - codeparrot_training - Step 3164: {'lr': 0.0004992748577957812, 'samples': 607680, 'steps': 3164, 'loss/train': 2.931361198425293} 01/27/2022 22:49:23 - INFO - codeparrot_training - Step 3165: {'lr': 0.0004992736119157469, 'samples': 607872, 'steps': 3165, 'loss/train': 5.142746329307556} 01/27/2022 22:49:28 - INFO - codeparrot_training - Step 3166: {'lr': 0.0004992723649679035, 'samples': 608064, 'steps': 3166, 'loss/train': 3.7835644483566284} 01/27/2022 22:49:32 - INFO - codeparrot_training - Step 3167: {'lr': 0.0004992711169522565, 'samples': 608256, 'steps': 3167, 'loss/train': 2.177835166454315} 01/27/2022 22:49:36 - INFO - codeparrot_training - Step 3168: {'lr': 0.0004992698678688111, 'samples': 608448, 'steps': 3168, 'loss/train': 3.5858523845672607} 01/27/2022 22:49:40 - INFO - codeparrot_training - Step 3169: {'lr': 0.0004992686177175728, 'samples': 608640, 'steps': 3169, 'loss/train': 3.1543092727661133} 01/27/2022 22:49:45 - INFO - codeparrot_training - Step 3170: {'lr': 0.000499267366498547, 'samples': 608832, 'steps': 3170, 'loss/train': 1.8252933025360107} 01/27/2022 22:49:50 - INFO - codeparrot_training - Step 3171: {'lr': 0.0004992661142117388, 'samples': 609024, 'steps': 3171, 'loss/train': 2.272695779800415} 01/27/2022 22:49:54 - INFO - codeparrot_training - Step 3172: {'lr': 0.0004992648608571537, 'samples': 609216, 'steps': 3172, 'loss/train': 3.352020263671875} 01/27/2022 22:49:58 - INFO - codeparrot_training - Step 3173: {'lr': 0.0004992636064347971, 'samples': 609408, 'steps': 3173, 'loss/train': 3.3712631464004517} 01/27/2022 22:50:02 - INFO - codeparrot_training - Step 3174: {'lr': 0.0004992623509446746, 'samples': 609600, 'steps': 3174, 'loss/train': 2.917252242565155} 01/27/2022 22:50:06 - INFO - codeparrot_training - Step 3175: {'lr': 0.0004992610943867911, 'samples': 609792, 'steps': 3175, 'loss/train': 3.8034439086914062} 01/27/2022 22:50:12 - INFO - codeparrot_training - Step 3176: {'lr': 0.0004992598367611523, 'samples': 609984, 'steps': 3176, 'loss/train': 3.5890685319900513} 01/27/2022 22:50:16 - INFO - codeparrot_training - Step 3177: {'lr': 0.0004992585780677634, 'samples': 610176, 'steps': 3177, 'loss/train': 3.0811657905578613} 01/27/2022 22:50:21 - INFO - codeparrot_training - Step 3178: {'lr': 0.00049925731830663, 'samples': 610368, 'steps': 3178, 'loss/train': 3.6310683488845825} 01/27/2022 22:50:25 - INFO - codeparrot_training - Step 3179: {'lr': 0.0004992560574777574, 'samples': 610560, 'steps': 3179, 'loss/train': 4.938891291618347} 01/27/2022 22:50:29 - INFO - codeparrot_training - Step 3180: {'lr': 0.000499254795581151, 'samples': 610752, 'steps': 3180, 'loss/train': 4.729300618171692} 01/27/2022 22:50:34 - INFO - codeparrot_training - Step 3181: {'lr': 0.0004992535326168162, 'samples': 610944, 'steps': 3181, 'loss/train': 2.3547279238700867} 01/27/2022 22:50:38 - INFO - codeparrot_training - Step 3182: {'lr': 0.0004992522685847583, 'samples': 611136, 'steps': 3182, 'loss/train': 2.067015767097473} 01/27/2022 22:50:43 - INFO - codeparrot_training - Step 3183: {'lr': 0.000499251003484983, 'samples': 611328, 'steps': 3183, 'loss/train': 3.523358702659607} 01/27/2022 22:50:47 - INFO - codeparrot_training - Step 3184: {'lr': 0.0004992497373174955, 'samples': 611520, 'steps': 3184, 'loss/train': 3.110766649246216} 01/27/2022 22:50:51 - INFO - codeparrot_training - Step 3185: {'lr': 0.0004992484700823012, 'samples': 611712, 'steps': 3185, 'loss/train': 3.3937840461730957} 01/27/2022 22:50:56 - INFO - codeparrot_training - Step 3186: {'lr': 0.0004992472017794057, 'samples': 611904, 'steps': 3186, 'loss/train': 3.7202446460723877} 01/27/2022 22:51:00 - INFO - codeparrot_training - Step 3187: {'lr': 0.0004992459324088143, 'samples': 612096, 'steps': 3187, 'loss/train': 2.368560254573822} 01/27/2022 22:51:04 - INFO - codeparrot_training - Step 3188: {'lr': 0.0004992446619705324, 'samples': 612288, 'steps': 3188, 'loss/train': 3.047111749649048} 01/27/2022 22:51:09 - INFO - codeparrot_training - Step 3189: {'lr': 0.0004992433904645654, 'samples': 612480, 'steps': 3189, 'loss/train': 2.906568467617035} 01/27/2022 22:51:13 - INFO - codeparrot_training - Step 3190: {'lr': 0.0004992421178909191, 'samples': 612672, 'steps': 3190, 'loss/train': 3.9876526594161987} 01/27/2022 22:51:19 - INFO - codeparrot_training - Step 3191: {'lr': 0.0004992408442495986, 'samples': 612864, 'steps': 3191, 'loss/train': 3.797660708427429} 01/27/2022 22:51:23 - INFO - codeparrot_training - Step 3192: {'lr': 0.0004992395695406095, 'samples': 613056, 'steps': 3192, 'loss/train': 7.290163278579712} 01/27/2022 22:51:27 - INFO - codeparrot_training - Step 3193: {'lr': 0.0004992382937639572, 'samples': 613248, 'steps': 3193, 'loss/train': 2.689559519290924} 01/27/2022 22:51:31 - INFO - codeparrot_training - Step 3194: {'lr': 0.0004992370169196472, 'samples': 613440, 'steps': 3194, 'loss/train': 4.072935461997986} 01/27/2022 22:51:35 - INFO - codeparrot_training - Step 3195: {'lr': 0.000499235739007685, 'samples': 613632, 'steps': 3195, 'loss/train': 1.8981733918190002} 01/27/2022 22:51:41 - INFO - codeparrot_training - Step 3196: {'lr': 0.000499234460028076, 'samples': 613824, 'steps': 3196, 'loss/train': 2.2147963643074036} 01/27/2022 22:51:45 - INFO - codeparrot_training - Step 3197: {'lr': 0.0004992331799808258, 'samples': 614016, 'steps': 3197, 'loss/train': 3.2968332767486572} 01/27/2022 22:51:49 - INFO - codeparrot_training - Step 3198: {'lr': 0.0004992318988659396, 'samples': 614208, 'steps': 3198, 'loss/train': 2.56413996219635} 01/27/2022 22:51:53 - INFO - codeparrot_training - Step 3199: {'lr': 0.0004992306166834232, 'samples': 614400, 'steps': 3199, 'loss/train': 4.310112833976746} 01/27/2022 22:51:57 - INFO - codeparrot_training - Step 3200: {'lr': 0.000499229333433282, 'samples': 614592, 'steps': 3200, 'loss/train': 4.672325491905212} 01/27/2022 22:52:03 - INFO - codeparrot_training - Step 3201: {'lr': 0.0004992280491155214, 'samples': 614784, 'steps': 3201, 'loss/train': 2.869318127632141} 01/27/2022 22:52:07 - INFO - codeparrot_training - Step 3202: {'lr': 0.0004992267637301471, 'samples': 614976, 'steps': 3202, 'loss/train': 3.6695398092269897} 01/27/2022 22:52:11 - INFO - codeparrot_training - Step 3203: {'lr': 0.0004992254772771644, 'samples': 615168, 'steps': 3203, 'loss/train': 3.8642749786376953} 01/27/2022 22:52:15 - INFO - codeparrot_training - Step 3204: {'lr': 0.0004992241897565789, 'samples': 615360, 'steps': 3204, 'loss/train': 3.2001746892929077} 01/27/2022 22:52:20 - INFO - codeparrot_training - Step 3205: {'lr': 0.0004992229011683961, 'samples': 615552, 'steps': 3205, 'loss/train': 4.268836140632629} 01/27/2022 22:52:25 - INFO - codeparrot_training - Step 3206: {'lr': 0.0004992216115126216, 'samples': 615744, 'steps': 3206, 'loss/train': 2.9691967964172363} 01/27/2022 22:52:30 - INFO - codeparrot_training - Step 3207: {'lr': 0.0004992203207892607, 'samples': 615936, 'steps': 3207, 'loss/train': 2.684413433074951} 01/27/2022 22:52:34 - INFO - codeparrot_training - Step 3208: {'lr': 0.0004992190289983192, 'samples': 616128, 'steps': 3208, 'loss/train': 3.269659996032715} 01/27/2022 22:52:38 - INFO - codeparrot_training - Step 3209: {'lr': 0.0004992177361398026, 'samples': 616320, 'steps': 3209, 'loss/train': 0.9264598488807678} 01/27/2022 22:52:42 - INFO - codeparrot_training - Step 3210: {'lr': 0.0004992164422137162, 'samples': 616512, 'steps': 3210, 'loss/train': 3.36108934879303} 01/27/2022 22:52:48 - INFO - codeparrot_training - Step 3211: {'lr': 0.0004992151472200657, 'samples': 616704, 'steps': 3211, 'loss/train': 2.794287621974945} 01/27/2022 22:52:52 - INFO - codeparrot_training - Step 3212: {'lr': 0.0004992138511588567, 'samples': 616896, 'steps': 3212, 'loss/train': 2.891538441181183} 01/27/2022 22:52:56 - INFO - codeparrot_training - Step 3213: {'lr': 0.0004992125540300947, 'samples': 617088, 'steps': 3213, 'loss/train': 2.132103145122528} 01/27/2022 22:53:00 - INFO - codeparrot_training - Step 3214: {'lr': 0.0004992112558337852, 'samples': 617280, 'steps': 3214, 'loss/train': 3.530227303504944} 01/27/2022 22:53:04 - INFO - codeparrot_training - Step 3215: {'lr': 0.0004992099565699339, 'samples': 617472, 'steps': 3215, 'loss/train': 3.807012677192688} 01/27/2022 22:53:09 - INFO - codeparrot_training - Step 3216: {'lr': 0.0004992086562385462, 'samples': 617664, 'steps': 3216, 'loss/train': 3.4635279178619385} 01/27/2022 22:53:14 - INFO - codeparrot_training - Step 3217: {'lr': 0.0004992073548396277, 'samples': 617856, 'steps': 3217, 'loss/train': 3.2331418991088867} 01/27/2022 22:53:19 - INFO - codeparrot_training - Step 3218: {'lr': 0.0004992060523731842, 'samples': 618048, 'steps': 3218, 'loss/train': 2.524372100830078} 01/27/2022 22:53:23 - INFO - codeparrot_training - Step 3219: {'lr': 0.0004992047488392209, 'samples': 618240, 'steps': 3219, 'loss/train': 4.7458906173706055} 01/27/2022 22:53:27 - INFO - codeparrot_training - Step 3220: {'lr': 0.0004992034442377437, 'samples': 618432, 'steps': 3220, 'loss/train': 3.0180469751358032} 01/27/2022 22:53:31 - INFO - codeparrot_training - Step 3221: {'lr': 0.0004992021385687582, 'samples': 618624, 'steps': 3221, 'loss/train': 4.5531299114227295} 01/27/2022 22:53:36 - INFO - codeparrot_training - Step 3222: {'lr': 0.0004992008318322697, 'samples': 618816, 'steps': 3222, 'loss/train': 3.653615355491638} 01/27/2022 22:53:41 - INFO - codeparrot_training - Step 3223: {'lr': 0.000499199524028284, 'samples': 619008, 'steps': 3223, 'loss/train': 4.763666152954102} 01/27/2022 22:53:45 - INFO - codeparrot_training - Step 3224: {'lr': 0.0004991982151568066, 'samples': 619200, 'steps': 3224, 'loss/train': 3.160387873649597} 01/27/2022 22:53:49 - INFO - codeparrot_training - Step 3225: {'lr': 0.0004991969052178433, 'samples': 619392, 'steps': 3225, 'loss/train': 3.4723992347717285} 01/27/2022 22:53:53 - INFO - codeparrot_training - Step 3226: {'lr': 0.0004991955942113995, 'samples': 619584, 'steps': 3226, 'loss/train': 6.550274848937988} 01/27/2022 22:53:59 - INFO - codeparrot_training - Step 3227: {'lr': 0.0004991942821374809, 'samples': 619776, 'steps': 3227, 'loss/train': 3.782147526741028} 01/27/2022 22:54:03 - INFO - codeparrot_training - Step 3228: {'lr': 0.0004991929689960932, 'samples': 619968, 'steps': 3228, 'loss/train': 3.3800843954086304} 01/27/2022 22:54:07 - INFO - codeparrot_training - Step 3229: {'lr': 0.000499191654787242, 'samples': 620160, 'steps': 3229, 'loss/train': 3.116997241973877} 01/27/2022 22:54:11 - INFO - codeparrot_training - Step 3230: {'lr': 0.0004991903395109328, 'samples': 620352, 'steps': 3230, 'loss/train': 3.069319009780884} 01/27/2022 22:54:15 - INFO - codeparrot_training - Step 3231: {'lr': 0.0004991890231671712, 'samples': 620544, 'steps': 3231, 'loss/train': 3.976094126701355} 01/27/2022 22:54:21 - INFO - codeparrot_training - Step 3232: {'lr': 0.0004991877057559631, 'samples': 620736, 'steps': 3232, 'loss/train': 2.6137688755989075} 01/27/2022 22:54:25 - INFO - codeparrot_training - Step 3233: {'lr': 0.0004991863872773139, 'samples': 620928, 'steps': 3233, 'loss/train': 4.350563049316406} 01/27/2022 22:54:29 - INFO - codeparrot_training - Step 3234: {'lr': 0.0004991850677312295, 'samples': 621120, 'steps': 3234, 'loss/train': 3.8131706714630127} 01/27/2022 22:54:33 - INFO - codeparrot_training - Step 3235: {'lr': 0.0004991837471177152, 'samples': 621312, 'steps': 3235, 'loss/train': 3.776088237762451} 01/27/2022 22:54:37 - INFO - codeparrot_training - Step 3236: {'lr': 0.000499182425436777, 'samples': 621504, 'steps': 3236, 'loss/train': 2.5512571334838867} 01/27/2022 22:54:44 - INFO - codeparrot_training - Step 3237: {'lr': 0.0004991811026884203, 'samples': 621696, 'steps': 3237, 'loss/train': 4.177311301231384} 01/27/2022 22:54:48 - INFO - codeparrot_training - Step 3238: {'lr': 0.0004991797788726509, 'samples': 621888, 'steps': 3238, 'loss/train': 4.0945011377334595} 01/27/2022 22:54:52 - INFO - codeparrot_training - Step 3239: {'lr': 0.0004991784539894745, 'samples': 622080, 'steps': 3239, 'loss/train': 3.6664702892303467} 01/27/2022 22:54:56 - INFO - codeparrot_training - Step 3240: {'lr': 0.0004991771280388967, 'samples': 622272, 'steps': 3240, 'loss/train': 4.465333700180054} 01/27/2022 22:55:00 - INFO - codeparrot_training - Step 3241: {'lr': 0.0004991758010209232, 'samples': 622464, 'steps': 3241, 'loss/train': 3.065920829772949} 01/27/2022 22:55:05 - INFO - codeparrot_training - Step 3242: {'lr': 0.0004991744729355598, 'samples': 622656, 'steps': 3242, 'loss/train': 3.9764156341552734} 01/27/2022 22:55:10 - INFO - codeparrot_training - Step 3243: {'lr': 0.0004991731437828119, 'samples': 622848, 'steps': 3243, 'loss/train': 4.143669247627258} 01/27/2022 22:55:14 - INFO - codeparrot_training - Step 3244: {'lr': 0.0004991718135626855, 'samples': 623040, 'steps': 3244, 'loss/train': 3.174854278564453} 01/27/2022 22:55:18 - INFO - codeparrot_training - Step 3245: {'lr': 0.0004991704822751861, 'samples': 623232, 'steps': 3245, 'loss/train': 3.5363770723342896} 01/27/2022 22:55:22 - INFO - codeparrot_training - Step 3246: {'lr': 0.0004991691499203195, 'samples': 623424, 'steps': 3246, 'loss/train': 3.3921890258789062} 01/27/2022 22:55:28 - INFO - codeparrot_training - Step 3247: {'lr': 0.0004991678164980914, 'samples': 623616, 'steps': 3247, 'loss/train': 2.577237904071808} 01/27/2022 22:55:32 - INFO - codeparrot_training - Step 3248: {'lr': 0.0004991664820085074, 'samples': 623808, 'steps': 3248, 'loss/train': 3.5617200136184692} 01/27/2022 22:55:36 - INFO - codeparrot_training - Step 3249: {'lr': 0.0004991651464515735, 'samples': 624000, 'steps': 3249, 'loss/train': 3.082526206970215} 01/27/2022 22:55:40 - INFO - codeparrot_training - Step 3250: {'lr': 0.0004991638098272951, 'samples': 624192, 'steps': 3250, 'loss/train': 2.0896019339561462} 01/27/2022 22:55:45 - INFO - codeparrot_training - Step 3251: {'lr': 0.000499162472135678, 'samples': 624384, 'steps': 3251, 'loss/train': 3.2163448333740234} 01/27/2022 22:55:50 - INFO - codeparrot_training - Step 3252: {'lr': 0.0004991611333767281, 'samples': 624576, 'steps': 3252, 'loss/train': 3.725233554840088} 01/27/2022 22:55:54 - INFO - codeparrot_training - Step 3253: {'lr': 0.000499159793550451, 'samples': 624768, 'steps': 3253, 'loss/train': 3.1379220485687256} 01/27/2022 22:55:59 - INFO - codeparrot_training - Step 3254: {'lr': 0.0004991584526568524, 'samples': 624960, 'steps': 3254, 'loss/train': 3.0787521600723267} 01/27/2022 22:56:03 - INFO - codeparrot_training - Step 3255: {'lr': 0.0004991571106959383, 'samples': 625152, 'steps': 3255, 'loss/train': 4.1603089570999146} 01/27/2022 22:56:07 - INFO - codeparrot_training - Step 3256: {'lr': 0.000499155767667714, 'samples': 625344, 'steps': 3256, 'loss/train': 5.955940246582031} 01/27/2022 22:56:13 - INFO - codeparrot_training - Step 3257: {'lr': 0.0004991544235721857, 'samples': 625536, 'steps': 3257, 'loss/train': 2.75467586517334} 01/27/2022 22:56:17 - INFO - codeparrot_training - Step 3258: {'lr': 0.0004991530784093589, 'samples': 625728, 'steps': 3258, 'loss/train': 3.9024603366851807} 01/27/2022 22:56:21 - INFO - codeparrot_training - Step 3259: {'lr': 0.0004991517321792394, 'samples': 625920, 'steps': 3259, 'loss/train': 3.9703577756881714} 01/27/2022 22:56:25 - INFO - codeparrot_training - Step 3260: {'lr': 0.000499150384881833, 'samples': 626112, 'steps': 3260, 'loss/train': 3.2815099954605103} 01/27/2022 22:56:29 - INFO - codeparrot_training - Step 3261: {'lr': 0.0004991490365171454, 'samples': 626304, 'steps': 3261, 'loss/train': 2.836823344230652} 01/27/2022 22:56:35 - INFO - codeparrot_training - Step 3262: {'lr': 0.0004991476870851825, 'samples': 626496, 'steps': 3262, 'loss/train': 3.098743200302124} 01/27/2022 22:56:39 - INFO - codeparrot_training - Step 3263: {'lr': 0.0004991463365859501, 'samples': 626688, 'steps': 3263, 'loss/train': 2.0556047558784485} 01/27/2022 22:56:44 - INFO - codeparrot_training - Step 3264: {'lr': 0.0004991449850194538, 'samples': 626880, 'steps': 3264, 'loss/train': 4.199881196022034} 01/27/2022 22:56:48 - INFO - codeparrot_training - Step 3265: {'lr': 0.0004991436323856995, 'samples': 627072, 'steps': 3265, 'loss/train': 2.3612692952156067} 01/27/2022 22:56:52 - INFO - codeparrot_training - Step 3266: {'lr': 0.0004991422786846931, 'samples': 627264, 'steps': 3266, 'loss/train': 4.508329153060913} 01/27/2022 22:56:58 - INFO - codeparrot_training - Step 3267: {'lr': 0.0004991409239164401, 'samples': 627456, 'steps': 3267, 'loss/train': 4.221953988075256} 01/27/2022 22:57:02 - INFO - codeparrot_training - Step 3268: {'lr': 0.0004991395680809467, 'samples': 627648, 'steps': 3268, 'loss/train': 3.4669772386550903} 01/27/2022 22:57:06 - INFO - codeparrot_training - Step 3269: {'lr': 0.0004991382111782183, 'samples': 627840, 'steps': 3269, 'loss/train': 3.3413082361221313} 01/27/2022 22:57:10 - INFO - codeparrot_training - Step 3270: {'lr': 0.0004991368532082611, 'samples': 628032, 'steps': 3270, 'loss/train': 8.148984432220459} 01/27/2022 22:57:14 - INFO - codeparrot_training - Step 3271: {'lr': 0.0004991354941710806, 'samples': 628224, 'steps': 3271, 'loss/train': 3.3636428117752075} 01/27/2022 22:57:20 - INFO - codeparrot_training - Step 3272: {'lr': 0.0004991341340666828, 'samples': 628416, 'steps': 3272, 'loss/train': 3.1425018310546875} 01/27/2022 22:57:24 - INFO - codeparrot_training - Step 3273: {'lr': 0.0004991327728950736, 'samples': 628608, 'steps': 3273, 'loss/train': 3.4970297813415527} 01/27/2022 22:57:28 - INFO - codeparrot_training - Step 3274: {'lr': 0.0004991314106562586, 'samples': 628800, 'steps': 3274, 'loss/train': 3.6620371341705322} 01/27/2022 22:57:32 - INFO - codeparrot_training - Step 3275: {'lr': 0.0004991300473502437, 'samples': 628992, 'steps': 3275, 'loss/train': 2.9115793704986572} 01/27/2022 22:57:36 - INFO - codeparrot_training - Step 3276: {'lr': 0.0004991286829770348, 'samples': 629184, 'steps': 3276, 'loss/train': 2.6565751433372498} 01/27/2022 22:57:42 - INFO - codeparrot_training - Step 3277: {'lr': 0.0004991273175366378, 'samples': 629376, 'steps': 3277, 'loss/train': 3.1117032766342163} 01/27/2022 22:57:46 - INFO - codeparrot_training - Step 3278: {'lr': 0.0004991259510290584, 'samples': 629568, 'steps': 3278, 'loss/train': 3.7358075380325317} 01/27/2022 22:57:51 - INFO - codeparrot_training - Step 3279: {'lr': 0.0004991245834543025, 'samples': 629760, 'steps': 3279, 'loss/train': 3.793682813644409} 01/27/2022 22:57:55 - INFO - codeparrot_training - Step 3280: {'lr': 0.0004991232148123761, 'samples': 629952, 'steps': 3280, 'loss/train': 3.48506498336792} 01/27/2022 22:57:59 - INFO - codeparrot_training - Step 3281: {'lr': 0.0004991218451032849, 'samples': 630144, 'steps': 3281, 'loss/train': 3.7708364725112915} 01/27/2022 22:58:04 - INFO - codeparrot_training - Step 3282: {'lr': 0.0004991204743270348, 'samples': 630336, 'steps': 3282, 'loss/train': 2.586375117301941} 01/27/2022 22:58:08 - INFO - codeparrot_training - Step 3283: {'lr': 0.0004991191024836317, 'samples': 630528, 'steps': 3283, 'loss/train': 5.19451367855072} 01/27/2022 22:58:12 - INFO - codeparrot_training - Step 3284: {'lr': 0.0004991177295730815, 'samples': 630720, 'steps': 3284, 'loss/train': 2.0043638348579407} 01/27/2022 22:58:17 - INFO - codeparrot_training - Step 3285: {'lr': 0.0004991163555953901, 'samples': 630912, 'steps': 3285, 'loss/train': 2.813935339450836} 01/27/2022 22:58:21 - INFO - codeparrot_training - Step 3286: {'lr': 0.0004991149805505632, 'samples': 631104, 'steps': 3286, 'loss/train': 2.488706409931183} 01/27/2022 22:58:27 - INFO - codeparrot_training - Step 3287: {'lr': 0.0004991136044386069, 'samples': 631296, 'steps': 3287, 'loss/train': 3.01606822013855} 01/27/2022 22:58:31 - INFO - codeparrot_training - Step 3288: {'lr': 0.0004991122272595271, 'samples': 631488, 'steps': 3288, 'loss/train': 1.8791134357452393} 01/27/2022 22:58:35 - INFO - codeparrot_training - Step 3289: {'lr': 0.0004991108490133296, 'samples': 631680, 'steps': 3289, 'loss/train': 2.7844621539115906} 01/27/2022 22:58:39 - INFO - codeparrot_training - Step 3290: {'lr': 0.0004991094697000202, 'samples': 631872, 'steps': 3290, 'loss/train': 3.255859136581421} 01/27/2022 22:58:43 - INFO - codeparrot_training - Step 3291: {'lr': 0.000499108089319605, 'samples': 632064, 'steps': 3291, 'loss/train': 4.040838718414307} 01/27/2022 22:58:49 - INFO - codeparrot_training - Step 3292: {'lr': 0.0004991067078720899, 'samples': 632256, 'steps': 3292, 'loss/train': 3.486961841583252} 01/27/2022 22:58:53 - INFO - codeparrot_training - Step 3293: {'lr': 0.0004991053253574807, 'samples': 632448, 'steps': 3293, 'loss/train': 2.0824227333068848} 01/27/2022 22:58:57 - INFO - codeparrot_training - Step 3294: {'lr': 0.0004991039417757833, 'samples': 632640, 'steps': 3294, 'loss/train': 3.237493872642517} 01/27/2022 22:59:02 - INFO - codeparrot_training - Step 3295: {'lr': 0.0004991025571270039, 'samples': 632832, 'steps': 3295, 'loss/train': 2.3011229038238525} 01/27/2022 22:59:06 - INFO - codeparrot_training - Step 3296: {'lr': 0.000499101171411148, 'samples': 633024, 'steps': 3296, 'loss/train': 3.9045900106430054} 01/27/2022 22:59:11 - INFO - codeparrot_training - Step 3297: {'lr': 0.000499099784628222, 'samples': 633216, 'steps': 3297, 'loss/train': 3.546875238418579} 01/27/2022 22:59:15 - INFO - codeparrot_training - Step 3298: {'lr': 0.0004990983967782316, 'samples': 633408, 'steps': 3298, 'loss/train': 3.2949700355529785} 01/27/2022 22:59:20 - INFO - codeparrot_training - Step 3299: {'lr': 0.0004990970078611827, 'samples': 633600, 'steps': 3299, 'loss/train': 5.3286813497543335} 01/27/2022 22:59:24 - INFO - codeparrot_training - Step 3300: {'lr': 0.0004990956178770814, 'samples': 633792, 'steps': 3300, 'loss/train': 3.7921013832092285} 01/27/2022 22:59:28 - INFO - codeparrot_training - Step 3301: {'lr': 0.0004990942268259335, 'samples': 633984, 'steps': 3301, 'loss/train': 3.338190793991089} 01/27/2022 22:59:33 - INFO - codeparrot_training - Step 3302: {'lr': 0.000499092834707745, 'samples': 634176, 'steps': 3302, 'loss/train': 2.584092915058136} 01/27/2022 22:59:37 - INFO - codeparrot_training - Step 3303: {'lr': 0.000499091441522522, 'samples': 634368, 'steps': 3303, 'loss/train': 3.020998477935791} 01/27/2022 22:59:41 - INFO - codeparrot_training - Step 3304: {'lr': 0.0004990900472702702, 'samples': 634560, 'steps': 3304, 'loss/train': 3.5144609212875366} 01/27/2022 22:59:46 - INFO - codeparrot_training - Step 3305: {'lr': 0.0004990886519509959, 'samples': 634752, 'steps': 3305, 'loss/train': 3.9752962589263916} 01/27/2022 22:59:50 - INFO - codeparrot_training - Step 3306: {'lr': 0.0004990872555647048, 'samples': 634944, 'steps': 3306, 'loss/train': 3.625114917755127} 01/27/2022 22:59:56 - INFO - codeparrot_training - Step 3307: {'lr': 0.0004990858581114029, 'samples': 635136, 'steps': 3307, 'loss/train': 3.9631065130233765} 01/27/2022 23:00:00 - INFO - codeparrot_training - Step 3308: {'lr': 0.0004990844595910965, 'samples': 635328, 'steps': 3308, 'loss/train': 3.256378412246704} 01/27/2022 23:00:05 - INFO - codeparrot_training - Step 3309: {'lr': 0.0004990830600037912, 'samples': 635520, 'steps': 3309, 'loss/train': 3.3600457906723022} 01/27/2022 23:00:09 - INFO - codeparrot_training - Step 3310: {'lr': 0.0004990816593494933, 'samples': 635712, 'steps': 3310, 'loss/train': 0.4326363801956177} 01/27/2022 23:00:13 - INFO - codeparrot_training - Step 3311: {'lr': 0.0004990802576282085, 'samples': 635904, 'steps': 3311, 'loss/train': 3.461227297782898} 01/27/2022 23:00:18 - INFO - codeparrot_training - Step 3312: {'lr': 0.0004990788548399431, 'samples': 636096, 'steps': 3312, 'loss/train': 2.6101357340812683} 01/27/2022 23:00:22 - INFO - codeparrot_training - Step 3313: {'lr': 0.0004990774509847029, 'samples': 636288, 'steps': 3313, 'loss/train': 2.3357269763946533} 01/27/2022 23:00:26 - INFO - codeparrot_training - Step 3314: {'lr': 0.0004990760460624941, 'samples': 636480, 'steps': 3314, 'loss/train': 3.632524251937866} 01/27/2022 23:00:31 - INFO - codeparrot_training - Step 3315: {'lr': 0.0004990746400733225, 'samples': 636672, 'steps': 3315, 'loss/train': 2.9998303055763245} 01/27/2022 23:00:35 - INFO - codeparrot_training - Step 3316: {'lr': 0.0004990732330171943, 'samples': 636864, 'steps': 3316, 'loss/train': 3.449578642845154} 01/27/2022 23:00:41 - INFO - codeparrot_training - Step 3317: {'lr': 0.0004990718248941154, 'samples': 637056, 'steps': 3317, 'loss/train': 3.12689745426178} 01/27/2022 23:00:45 - INFO - codeparrot_training - Step 3318: {'lr': 0.0004990704157040919, 'samples': 637248, 'steps': 3318, 'loss/train': 9.551783323287964} 01/27/2022 23:00:49 - INFO - codeparrot_training - Step 3319: {'lr': 0.0004990690054471299, 'samples': 637440, 'steps': 3319, 'loss/train': 1.1904468834400177} 01/27/2022 23:00:54 - INFO - codeparrot_training - Step 3320: {'lr': 0.0004990675941232354, 'samples': 637632, 'steps': 3320, 'loss/train': 3.5280200242996216} 01/27/2022 23:00:58 - INFO - codeparrot_training - Step 3321: {'lr': 0.0004990661817324142, 'samples': 637824, 'steps': 3321, 'loss/train': 4.218076586723328} 01/27/2022 23:01:02 - INFO - codeparrot_training - Step 3322: {'lr': 0.0004990647682746727, 'samples': 638016, 'steps': 3322, 'loss/train': 10.603039026260376} 01/27/2022 23:01:07 - INFO - codeparrot_training - Step 3323: {'lr': 0.0004990633537500169, 'samples': 638208, 'steps': 3323, 'loss/train': 1.916492521762848} 01/27/2022 23:01:11 - INFO - codeparrot_training - Step 3324: {'lr': 0.0004990619381584527, 'samples': 638400, 'steps': 3324, 'loss/train': 3.6476937532424927} 01/27/2022 23:01:16 - INFO - codeparrot_training - Step 3325: {'lr': 0.0004990605214999862, 'samples': 638592, 'steps': 3325, 'loss/train': 1.679301917552948} 01/27/2022 23:01:20 - INFO - codeparrot_training - Step 3326: {'lr': 0.0004990591037746236, 'samples': 638784, 'steps': 3326, 'loss/train': 3.4153894186019897} 01/27/2022 23:01:24 - INFO - codeparrot_training - Step 3327: {'lr': 0.0004990576849823708, 'samples': 638976, 'steps': 3327, 'loss/train': 2.93931645154953} 01/27/2022 23:01:29 - INFO - codeparrot_training - Step 3328: {'lr': 0.000499056265123234, 'samples': 639168, 'steps': 3328, 'loss/train': 3.1938257217407227} 01/27/2022 23:01:34 - INFO - codeparrot_training - Step 3329: {'lr': 0.0004990548441972193, 'samples': 639360, 'steps': 3329, 'loss/train': 3.191433548927307} 01/27/2022 23:01:38 - INFO - codeparrot_training - Step 3330: {'lr': 0.0004990534222043325, 'samples': 639552, 'steps': 3330, 'loss/train': 2.964924395084381} 01/27/2022 23:01:42 - INFO - codeparrot_training - Step 3331: {'lr': 0.0004990519991445803, 'samples': 639744, 'steps': 3331, 'loss/train': 5.791944622993469} 01/27/2022 23:01:48 - INFO - codeparrot_training - Step 3332: {'lr': 0.0004990505750179682, 'samples': 639936, 'steps': 3332, 'loss/train': 3.1245524883270264} 01/27/2022 23:01:52 - INFO - codeparrot_training - Step 3333: {'lr': 0.0004990491498245024, 'samples': 640128, 'steps': 3333, 'loss/train': 4.204692363739014} 01/27/2022 23:01:56 - INFO - codeparrot_training - Step 3334: {'lr': 0.0004990477235641893, 'samples': 640320, 'steps': 3334, 'loss/train': 3.7206408977508545} 01/27/2022 23:02:01 - INFO - codeparrot_training - Step 3335: {'lr': 0.0004990462962370347, 'samples': 640512, 'steps': 3335, 'loss/train': 0.6476283520460129} 01/27/2022 23:02:05 - INFO - codeparrot_training - Step 3336: {'lr': 0.0004990448678430451, 'samples': 640704, 'steps': 3336, 'loss/train': 3.84457790851593} 01/27/2022 23:02:09 - INFO - codeparrot_training - Step 3337: {'lr': 0.0004990434383822261, 'samples': 640896, 'steps': 3337, 'loss/train': 5.362391710281372} 01/27/2022 23:02:14 - INFO - codeparrot_training - Step 3338: {'lr': 0.0004990420078545843, 'samples': 641088, 'steps': 3338, 'loss/train': 4.575841426849365} 01/27/2022 23:02:18 - INFO - codeparrot_training - Step 3339: {'lr': 0.0004990405762601254, 'samples': 641280, 'steps': 3339, 'loss/train': 3.8707669973373413} 01/27/2022 23:02:23 - INFO - codeparrot_training - Step 3340: {'lr': 0.000499039143598856, 'samples': 641472, 'steps': 3340, 'loss/train': 4.285722613334656} 01/27/2022 23:02:27 - INFO - codeparrot_training - Step 3341: {'lr': 0.0004990377098707818, 'samples': 641664, 'steps': 3341, 'loss/train': 4.5484442710876465} 01/27/2022 23:02:31 - INFO - codeparrot_training - Step 3342: {'lr': 0.0004990362750759092, 'samples': 641856, 'steps': 3342, 'loss/train': 4.009844541549683} 01/27/2022 23:02:36 - INFO - codeparrot_training - Step 3343: {'lr': 0.0004990348392142443, 'samples': 642048, 'steps': 3343, 'loss/train': 3.3439725637435913} 01/27/2022 23:02:40 - INFO - codeparrot_training - Step 3344: {'lr': 0.0004990334022857932, 'samples': 642240, 'steps': 3344, 'loss/train': 3.3714358806610107} 01/27/2022 23:02:44 - INFO - codeparrot_training - Step 3345: {'lr': 0.0004990319642905619, 'samples': 642432, 'steps': 3345, 'loss/train': 2.6660242080688477} 01/27/2022 23:02:49 - INFO - codeparrot_training - Step 3346: {'lr': 0.000499030525228557, 'samples': 642624, 'steps': 3346, 'loss/train': 4.28083062171936} 01/27/2022 23:02:53 - INFO - codeparrot_training - Step 3347: {'lr': 0.0004990290850997843, 'samples': 642816, 'steps': 3347, 'loss/train': 3.9202412366867065} 01/27/2022 23:02:59 - INFO - codeparrot_training - Step 3348: {'lr': 0.0004990276439042501, 'samples': 643008, 'steps': 3348, 'loss/train': 3.871934652328491} 01/27/2022 23:03:03 - INFO - codeparrot_training - Step 3349: {'lr': 0.0004990262016419606, 'samples': 643200, 'steps': 3349, 'loss/train': 2.6337472200393677} 01/27/2022 23:03:07 - INFO - codeparrot_training - Step 3350: {'lr': 0.0004990247583129218, 'samples': 643392, 'steps': 3350, 'loss/train': 1.9752715229988098} 01/27/2022 23:03:11 - INFO - codeparrot_training - Step 3351: {'lr': 0.00049902331391714, 'samples': 643584, 'steps': 3351, 'loss/train': 4.147114634513855} 01/27/2022 23:03:15 - INFO - codeparrot_training - Step 3352: {'lr': 0.0004990218684546216, 'samples': 643776, 'steps': 3352, 'loss/train': 3.626513957977295} 01/27/2022 23:03:20 - INFO - codeparrot_training - Step 3353: {'lr': 0.0004990204219253724, 'samples': 643968, 'steps': 3353, 'loss/train': 0.8796100616455078} 01/27/2022 23:03:25 - INFO - codeparrot_training - Step 3354: {'lr': 0.0004990189743293989, 'samples': 644160, 'steps': 3354, 'loss/train': 4.320117115974426} 01/27/2022 23:03:29 - INFO - codeparrot_training - Step 3355: {'lr': 0.0004990175256667071, 'samples': 644352, 'steps': 3355, 'loss/train': 3.7828670740127563} 01/27/2022 23:03:33 - INFO - codeparrot_training - Step 3356: {'lr': 0.0004990160759373033, 'samples': 644544, 'steps': 3356, 'loss/train': 3.1025705337524414} 01/27/2022 23:03:37 - INFO - codeparrot_training - Step 3357: {'lr': 0.0004990146251411938, 'samples': 644736, 'steps': 3357, 'loss/train': 3.2616931200027466} 01/27/2022 23:03:43 - INFO - codeparrot_training - Step 3358: {'lr': 0.0004990131732783846, 'samples': 644928, 'steps': 3358, 'loss/train': 4.419123530387878} 01/27/2022 23:03:47 - INFO - codeparrot_training - Step 3359: {'lr': 0.000499011720348882, 'samples': 645120, 'steps': 3359, 'loss/train': 1.04225492477417} 01/27/2022 23:03:51 - INFO - codeparrot_training - Step 3360: {'lr': 0.0004990102663526924, 'samples': 645312, 'steps': 3360, 'loss/train': 3.157815098762512} 01/27/2022 23:03:55 - INFO - codeparrot_training - Step 3361: {'lr': 0.0004990088112898219, 'samples': 645504, 'steps': 3361, 'loss/train': 4.3346439599990845} 01/27/2022 23:03:59 - INFO - codeparrot_training - Step 3362: {'lr': 0.0004990073551602766, 'samples': 645696, 'steps': 3362, 'loss/train': 3.3110697269439697} 01/27/2022 23:04:05 - INFO - codeparrot_training - Step 3363: {'lr': 0.000499005897964063, 'samples': 645888, 'steps': 3363, 'loss/train': 4.1797438859939575} 01/27/2022 23:04:09 - INFO - codeparrot_training - Step 3364: {'lr': 0.0004990044397011871, 'samples': 646080, 'steps': 3364, 'loss/train': 2.9986461997032166} 01/27/2022 23:04:14 - INFO - codeparrot_training - Step 3365: {'lr': 0.0004990029803716552, 'samples': 646272, 'steps': 3365, 'loss/train': 4.206450462341309} 01/27/2022 23:04:18 - INFO - codeparrot_training - Step 3366: {'lr': 0.0004990015199754736, 'samples': 646464, 'steps': 3366, 'loss/train': 3.627485990524292} 01/27/2022 23:04:22 - INFO - codeparrot_training - Step 3367: {'lr': 0.0004990000585126486, 'samples': 646656, 'steps': 3367, 'loss/train': 3.7052507400512695} 01/27/2022 23:04:27 - INFO - codeparrot_training - Step 3368: {'lr': 0.0004989985959831865, 'samples': 646848, 'steps': 3368, 'loss/train': 2.69439160823822} 01/27/2022 23:04:31 - INFO - codeparrot_training - Step 3369: {'lr': 0.0004989971323870934, 'samples': 647040, 'steps': 3369, 'loss/train': 2.5338674783706665} 01/27/2022 23:04:35 - INFO - codeparrot_training - Step 3370: {'lr': 0.0004989956677243757, 'samples': 647232, 'steps': 3370, 'loss/train': 3.7796316146850586} 01/27/2022 23:04:39 - INFO - codeparrot_training - Step 3371: {'lr': 0.0004989942019950395, 'samples': 647424, 'steps': 3371, 'loss/train': 2.5443313121795654} 01/27/2022 23:04:44 - INFO - codeparrot_training - Step 3372: {'lr': 0.0004989927351990912, 'samples': 647616, 'steps': 3372, 'loss/train': 2.255749762058258} 01/27/2022 23:04:49 - INFO - codeparrot_training - Step 3373: {'lr': 0.0004989912673365373, 'samples': 647808, 'steps': 3373, 'loss/train': 2.932197690010071} 01/27/2022 23:04:53 - INFO - codeparrot_training - Step 3374: {'lr': 0.0004989897984073837, 'samples': 648000, 'steps': 3374, 'loss/train': 2.9070895314216614} 01/27/2022 23:04:57 - INFO - codeparrot_training - Step 3375: {'lr': 0.000498988328411637, 'samples': 648192, 'steps': 3375, 'loss/train': 1.7143471240997314} 01/27/2022 23:05:01 - INFO - codeparrot_training - Step 3376: {'lr': 0.0004989868573493032, 'samples': 648384, 'steps': 3376, 'loss/train': 3.6251882314682007} 01/27/2022 23:05:08 - INFO - codeparrot_training - Step 3377: {'lr': 0.0004989853852203889, 'samples': 648576, 'steps': 3377, 'loss/train': 4.512553811073303} 01/27/2022 23:05:12 - INFO - codeparrot_training - Step 3378: {'lr': 0.0004989839120249002, 'samples': 648768, 'steps': 3378, 'loss/train': 4.414860248565674} 01/27/2022 23:05:16 - INFO - codeparrot_training - Step 3379: {'lr': 0.0004989824377628435, 'samples': 648960, 'steps': 3379, 'loss/train': 3.3405861854553223} 01/27/2022 23:05:20 - INFO - codeparrot_training - Step 3380: {'lr': 0.0004989809624342251, 'samples': 649152, 'steps': 3380, 'loss/train': 3.3856390714645386} 01/27/2022 23:05:24 - INFO - codeparrot_training - Step 3381: {'lr': 0.0004989794860390513, 'samples': 649344, 'steps': 3381, 'loss/train': 3.5672796964645386} 01/27/2022 23:05:29 - INFO - codeparrot_training - Step 3382: {'lr': 0.0004989780085773285, 'samples': 649536, 'steps': 3382, 'loss/train': 1.7507117986679077} 01/27/2022 23:05:34 - INFO - codeparrot_training - Step 3383: {'lr': 0.0004989765300490628, 'samples': 649728, 'steps': 3383, 'loss/train': 3.9254393577575684} 01/27/2022 23:05:38 - INFO - codeparrot_training - Step 3384: {'lr': 0.0004989750504542609, 'samples': 649920, 'steps': 3384, 'loss/train': 3.5861284732818604} 01/27/2022 23:05:42 - INFO - codeparrot_training - Step 3385: {'lr': 0.0004989735697929289, 'samples': 650112, 'steps': 3385, 'loss/train': 2.3819265961647034} 01/27/2022 23:05:46 - INFO - codeparrot_training - Step 3386: {'lr': 0.0004989720880650731, 'samples': 650304, 'steps': 3386, 'loss/train': 2.8588274717330933} 01/27/2022 23:05:52 - INFO - codeparrot_training - Step 3387: {'lr': 0.0004989706052707, 'samples': 650496, 'steps': 3387, 'loss/train': 4.172054886817932} 01/27/2022 23:05:56 - INFO - codeparrot_training - Step 3388: {'lr': 0.0004989691214098158, 'samples': 650688, 'steps': 3388, 'loss/train': 1.9619529247283936} 01/27/2022 23:06:00 - INFO - codeparrot_training - Step 3389: {'lr': 0.0004989676364824271, 'samples': 650880, 'steps': 3389, 'loss/train': 3.0781749486923218} 01/27/2022 23:06:04 - INFO - codeparrot_training - Step 3390: {'lr': 0.00049896615048854, 'samples': 651072, 'steps': 3390, 'loss/train': 3.346163749694824} 01/27/2022 23:06:09 - INFO - codeparrot_training - Step 3391: {'lr': 0.000498964663428161, 'samples': 651264, 'steps': 3391, 'loss/train': 3.3487257957458496} 01/27/2022 23:06:14 - INFO - codeparrot_training - Step 3392: {'lr': 0.0004989631753012964, 'samples': 651456, 'steps': 3392, 'loss/train': 3.582629442214966} 01/27/2022 23:06:18 - INFO - codeparrot_training - Step 3393: {'lr': 0.0004989616861079527, 'samples': 651648, 'steps': 3393, 'loss/train': 2.1526764035224915} 01/27/2022 23:06:22 - INFO - codeparrot_training - Step 3394: {'lr': 0.0004989601958481361, 'samples': 651840, 'steps': 3394, 'loss/train': 3.313825249671936} 01/27/2022 23:06:26 - INFO - codeparrot_training - Step 3395: {'lr': 0.000498958704521853, 'samples': 652032, 'steps': 3395, 'loss/train': 2.7753692865371704} 01/27/2022 23:06:30 - INFO - codeparrot_training - Step 3396: {'lr': 0.00049895721212911, 'samples': 652224, 'steps': 3396, 'loss/train': 3.4855188131332397} 01/27/2022 23:06:35 - INFO - codeparrot_training - Step 3397: {'lr': 0.0004989557186699133, 'samples': 652416, 'steps': 3397, 'loss/train': 4.026958465576172} 01/27/2022 23:06:40 - INFO - codeparrot_training - Step 3398: {'lr': 0.0004989542241442695, 'samples': 652608, 'steps': 3398, 'loss/train': 4.658165216445923} 01/27/2022 23:06:44 - INFO - codeparrot_training - Step 3399: {'lr': 0.0004989527285521846, 'samples': 652800, 'steps': 3399, 'loss/train': 0.9596365392208099} 01/27/2022 23:06:48 - INFO - codeparrot_training - Step 3400: {'lr': 0.0004989512318936654, 'samples': 652992, 'steps': 3400, 'loss/train': 3.8973026275634766} 01/27/2022 23:06:52 - INFO - codeparrot_training - Step 3401: {'lr': 0.0004989497341687182, 'samples': 653184, 'steps': 3401, 'loss/train': 3.291368007659912} 01/27/2022 23:06:58 - INFO - codeparrot_training - Step 3402: {'lr': 0.0004989482353773494, 'samples': 653376, 'steps': 3402, 'loss/train': 4.090349078178406} 01/27/2022 23:07:02 - INFO - codeparrot_training - Step 3403: {'lr': 0.0004989467355195653, 'samples': 653568, 'steps': 3403, 'loss/train': 3.621083378791809} 01/27/2022 23:07:06 - INFO - codeparrot_training - Step 3404: {'lr': 0.0004989452345953725, 'samples': 653760, 'steps': 3404, 'loss/train': 3.451648235321045} 01/27/2022 23:07:10 - INFO - codeparrot_training - Step 3405: {'lr': 0.0004989437326047774, 'samples': 653952, 'steps': 3405, 'loss/train': 1.5285170674324036} 01/27/2022 23:07:15 - INFO - codeparrot_training - Step 3406: {'lr': 0.0004989422295477863, 'samples': 654144, 'steps': 3406, 'loss/train': 2.7330719232559204} 01/27/2022 23:07:20 - INFO - codeparrot_training - Step 3407: {'lr': 0.0004989407254244058, 'samples': 654336, 'steps': 3407, 'loss/train': 2.951099216938019} 01/27/2022 23:07:24 - INFO - codeparrot_training - Step 3408: {'lr': 0.0004989392202346424, 'samples': 654528, 'steps': 3408, 'loss/train': 3.251931667327881} 01/27/2022 23:07:28 - INFO - codeparrot_training - Step 3409: {'lr': 0.0004989377139785022, 'samples': 654720, 'steps': 3409, 'loss/train': 2.0644372701644897} 01/27/2022 23:07:32 - INFO - codeparrot_training - Step 3410: {'lr': 0.000498936206655992, 'samples': 654912, 'steps': 3410, 'loss/train': 3.565981864929199} 01/27/2022 23:07:36 - INFO - codeparrot_training - Step 3411: {'lr': 0.0004989346982671181, 'samples': 655104, 'steps': 3411, 'loss/train': 2.994920790195465} 01/27/2022 23:07:41 - INFO - codeparrot_training - Step 3412: {'lr': 0.0004989331888118869, 'samples': 655296, 'steps': 3412, 'loss/train': 3.180545210838318} 01/27/2022 23:07:46 - INFO - codeparrot_training - Step 3413: {'lr': 0.0004989316782903052, 'samples': 655488, 'steps': 3413, 'loss/train': 3.4963327646255493} 01/27/2022 23:07:50 - INFO - codeparrot_training - Step 3414: {'lr': 0.0004989301667023791, 'samples': 655680, 'steps': 3414, 'loss/train': 2.9764702320098877} 01/27/2022 23:07:54 - INFO - codeparrot_training - Step 3415: {'lr': 0.0004989286540481152, 'samples': 655872, 'steps': 3415, 'loss/train': 3.6881414651870728} 01/27/2022 23:07:58 - INFO - codeparrot_training - Step 3416: {'lr': 0.00049892714032752, 'samples': 656064, 'steps': 3416, 'loss/train': 3.8308743238449097} 01/27/2022 23:08:03 - INFO - codeparrot_training - Step 3417: {'lr': 0.0004989256255406001, 'samples': 656256, 'steps': 3417, 'loss/train': 1.8341336846351624} 01/27/2022 23:08:08 - INFO - codeparrot_training - Step 3418: {'lr': 0.0004989241096873617, 'samples': 656448, 'steps': 3418, 'loss/train': 3.4872668981552124} 01/27/2022 23:08:12 - INFO - codeparrot_training - Step 3419: {'lr': 0.0004989225927678115, 'samples': 656640, 'steps': 3419, 'loss/train': 1.9531813859939575} 01/27/2022 23:08:16 - INFO - codeparrot_training - Step 3420: {'lr': 0.000498921074781956, 'samples': 656832, 'steps': 3420, 'loss/train': 2.8983201384544373} 01/27/2022 23:08:20 - INFO - codeparrot_training - Step 3421: {'lr': 0.0004989195557298016, 'samples': 657024, 'steps': 3421, 'loss/train': 0.9106407165527344} 01/27/2022 23:08:27 - INFO - codeparrot_training - Step 3422: {'lr': 0.0004989180356113549, 'samples': 657216, 'steps': 3422, 'loss/train': 7.040940999984741} 01/27/2022 23:08:31 - INFO - codeparrot_training - Step 3423: {'lr': 0.0004989165144266224, 'samples': 657408, 'steps': 3423, 'loss/train': 2.462111234664917} 01/27/2022 23:08:35 - INFO - codeparrot_training - Step 3424: {'lr': 0.0004989149921756105, 'samples': 657600, 'steps': 3424, 'loss/train': 4.8729987144470215} 01/27/2022 23:08:39 - INFO - codeparrot_training - Step 3425: {'lr': 0.0004989134688583259, 'samples': 657792, 'steps': 3425, 'loss/train': 4.579289674758911} 01/27/2022 23:08:43 - INFO - codeparrot_training - Step 3426: {'lr': 0.000498911944474775, 'samples': 657984, 'steps': 3426, 'loss/train': 2.307224214076996} 01/27/2022 23:08:49 - INFO - codeparrot_training - Step 3427: {'lr': 0.0004989104190249643, 'samples': 658176, 'steps': 3427, 'loss/train': 2.82904851436615} 01/27/2022 23:08:53 - INFO - codeparrot_training - Step 3428: {'lr': 0.0004989088925089005, 'samples': 658368, 'steps': 3428, 'loss/train': 3.2766906023025513} 01/27/2022 23:08:57 - INFO - codeparrot_training - Step 3429: {'lr': 0.00049890736492659, 'samples': 658560, 'steps': 3429, 'loss/train': 2.7911925315856934} 01/27/2022 23:09:01 - INFO - codeparrot_training - Step 3430: {'lr': 0.0004989058362780394, 'samples': 658752, 'steps': 3430, 'loss/train': 3.3826310634613037} 01/27/2022 23:09:05 - INFO - codeparrot_training - Step 3431: {'lr': 0.0004989043065632552, 'samples': 658944, 'steps': 3431, 'loss/train': 4.230495572090149} 01/27/2022 23:09:11 - INFO - codeparrot_training - Step 3432: {'lr': 0.0004989027757822441, 'samples': 659136, 'steps': 3432, 'loss/train': 2.6239094138145447} 01/27/2022 23:09:15 - INFO - codeparrot_training - Step 3433: {'lr': 0.0004989012439350124, 'samples': 659328, 'steps': 3433, 'loss/train': 3.9110870361328125} 01/27/2022 23:09:19 - INFO - codeparrot_training - Step 3434: {'lr': 0.0004988997110215668, 'samples': 659520, 'steps': 3434, 'loss/train': 3.4763413667678833} 01/27/2022 23:09:23 - INFO - codeparrot_training - Step 3435: {'lr': 0.0004988981770419141, 'samples': 659712, 'steps': 3435, 'loss/train': 3.108143448829651} 01/27/2022 23:09:27 - INFO - codeparrot_training - Step 3436: {'lr': 0.0004988966419960605, 'samples': 659904, 'steps': 3436, 'loss/train': 3.2024835348129272} 01/27/2022 23:09:33 - INFO - codeparrot_training - Step 3437: {'lr': 0.0004988951058840127, 'samples': 660096, 'steps': 3437, 'loss/train': 2.663639724254608} 01/27/2022 23:09:37 - INFO - codeparrot_training - Step 3438: {'lr': 0.0004988935687057773, 'samples': 660288, 'steps': 3438, 'loss/train': 3.4281277656555176} 01/27/2022 23:09:42 - INFO - codeparrot_training - Step 3439: {'lr': 0.0004988920304613609, 'samples': 660480, 'steps': 3439, 'loss/train': 4.22512149810791} 01/27/2022 23:09:46 - INFO - codeparrot_training - Step 3440: {'lr': 0.00049889049115077, 'samples': 660672, 'steps': 3440, 'loss/train': 1.8811814188957214} 01/27/2022 23:09:50 - INFO - codeparrot_training - Step 3441: {'lr': 0.0004988889507740113, 'samples': 660864, 'steps': 3441, 'loss/train': 3.4426270723342896} 01/27/2022 23:09:55 - INFO - codeparrot_training - Step 3442: {'lr': 0.0004988874093310914, 'samples': 661056, 'steps': 3442, 'loss/train': 4.008826732635498} 01/27/2022 23:09:59 - INFO - codeparrot_training - Step 3443: {'lr': 0.000498885866822017, 'samples': 661248, 'steps': 3443, 'loss/train': 3.332845687866211} 01/27/2022 23:10:03 - INFO - codeparrot_training - Step 3444: {'lr': 0.0004988843232467944, 'samples': 661440, 'steps': 3444, 'loss/train': 3.287359356880188} 01/27/2022 23:10:07 - INFO - codeparrot_training - Step 3445: {'lr': 0.0004988827786054304, 'samples': 661632, 'steps': 3445, 'loss/train': 3.9978811740875244} 01/27/2022 23:10:12 - INFO - codeparrot_training - Step 3446: {'lr': 0.0004988812328979317, 'samples': 661824, 'steps': 3446, 'loss/train': 2.3614965677261353} 01/27/2022 23:10:17 - INFO - codeparrot_training - Step 3447: {'lr': 0.0004988796861243046, 'samples': 662016, 'steps': 3447, 'loss/train': 3.8425272703170776} 01/27/2022 23:10:22 - INFO - codeparrot_training - Step 3448: {'lr': 0.0004988781382845562, 'samples': 662208, 'steps': 3448, 'loss/train': 3.1408199071884155} 01/27/2022 23:10:26 - INFO - codeparrot_training - Step 3449: {'lr': 0.0004988765893786929, 'samples': 662400, 'steps': 3449, 'loss/train': 10.743671178817749} 01/27/2022 23:10:30 - INFO - codeparrot_training - Step 3450: {'lr': 0.0004988750394067211, 'samples': 662592, 'steps': 3450, 'loss/train': 3.3941320180892944} 01/27/2022 23:10:34 - INFO - codeparrot_training - Step 3451: {'lr': 0.0004988734883686479, 'samples': 662784, 'steps': 3451, 'loss/train': 1.474985271692276} 01/27/2022 23:10:39 - INFO - codeparrot_training - Step 3452: {'lr': 0.0004988719362644795, 'samples': 662976, 'steps': 3452, 'loss/train': 3.621464967727661} 01/27/2022 23:10:43 - INFO - codeparrot_training - Step 3453: {'lr': 0.0004988703830942228, 'samples': 663168, 'steps': 3453, 'loss/train': 3.9251049757003784} 01/27/2022 23:10:47 - INFO - codeparrot_training - Step 3454: {'lr': 0.0004988688288578845, 'samples': 663360, 'steps': 3454, 'loss/train': 3.273497700691223} 01/27/2022 23:10:52 - INFO - codeparrot_training - Step 3455: {'lr': 0.0004988672735554711, 'samples': 663552, 'steps': 3455, 'loss/train': 4.1383055448532104} 01/27/2022 23:10:56 - INFO - codeparrot_training - Step 3456: {'lr': 0.0004988657171869893, 'samples': 663744, 'steps': 3456, 'loss/train': 2.8899133801460266} 01/27/2022 23:11:01 - INFO - codeparrot_training - Step 3457: {'lr': 0.0004988641597524458, 'samples': 663936, 'steps': 3457, 'loss/train': 2.274514138698578} 01/27/2022 23:11:05 - INFO - codeparrot_training - Step 3458: {'lr': 0.0004988626012518473, 'samples': 664128, 'steps': 3458, 'loss/train': 3.551068425178528} 01/27/2022 23:11:09 - INFO - codeparrot_training - Step 3459: {'lr': 0.0004988610416852004, 'samples': 664320, 'steps': 3459, 'loss/train': 1.9830549359321594} 01/27/2022 23:11:14 - INFO - codeparrot_training - Step 3460: {'lr': 0.0004988594810525118, 'samples': 664512, 'steps': 3460, 'loss/train': 3.55669105052948} 01/27/2022 23:11:18 - INFO - codeparrot_training - Step 3461: {'lr': 0.0004988579193537883, 'samples': 664704, 'steps': 3461, 'loss/train': 4.542608141899109} 01/27/2022 23:11:23 - INFO - codeparrot_training - Step 3462: {'lr': 0.0004988563565890364, 'samples': 664896, 'steps': 3462, 'loss/train': 3.5393736362457275} 01/27/2022 23:11:28 - INFO - codeparrot_training - Step 3463: {'lr': 0.000498854792758263, 'samples': 665088, 'steps': 3463, 'loss/train': 1.6769129633903503} 01/27/2022 23:11:32 - INFO - codeparrot_training - Step 3464: {'lr': 0.0004988532278614745, 'samples': 665280, 'steps': 3464, 'loss/train': 3.054471731185913} 01/27/2022 23:11:36 - INFO - codeparrot_training - Step 3465: {'lr': 0.0004988516618986779, 'samples': 665472, 'steps': 3465, 'loss/train': 4.2749855518341064} 01/27/2022 23:11:40 - INFO - codeparrot_training - Step 3466: {'lr': 0.0004988500948698799, 'samples': 665664, 'steps': 3466, 'loss/train': 3.685168504714966} 01/27/2022 23:11:45 - INFO - codeparrot_training - Step 3467: {'lr': 0.000498848526775087, 'samples': 665856, 'steps': 3467, 'loss/train': 2.155073404312134} 01/27/2022 23:11:49 - INFO - codeparrot_training - Step 3468: {'lr': 0.0004988469576143059, 'samples': 666048, 'steps': 3468, 'loss/train': 2.5664284229278564} 01/27/2022 23:11:54 - INFO - codeparrot_training - Step 3469: {'lr': 0.0004988453873875437, 'samples': 666240, 'steps': 3469, 'loss/train': 4.174206733703613} 01/27/2022 23:11:58 - INFO - codeparrot_training - Step 3470: {'lr': 0.0004988438160948068, 'samples': 666432, 'steps': 3470, 'loss/train': 2.9814807772636414} 01/27/2022 23:12:02 - INFO - codeparrot_training - Step 3471: {'lr': 0.000498842243736102, 'samples': 666624, 'steps': 3471, 'loss/train': 1.3721920251846313} 01/27/2022 23:12:08 - INFO - codeparrot_training - Step 3472: {'lr': 0.000498840670311436, 'samples': 666816, 'steps': 3472, 'loss/train': 3.2266337871551514} 01/27/2022 23:12:12 - INFO - codeparrot_training - Step 3473: {'lr': 0.0004988390958208156, 'samples': 667008, 'steps': 3473, 'loss/train': 4.41550612449646} 01/27/2022 23:12:16 - INFO - codeparrot_training - Step 3474: {'lr': 0.0004988375202642475, 'samples': 667200, 'steps': 3474, 'loss/train': 3.5146597623825073} 01/27/2022 23:12:20 - INFO - codeparrot_training - Step 3475: {'lr': 0.0004988359436417385, 'samples': 667392, 'steps': 3475, 'loss/train': 2.4276970624923706} 01/27/2022 23:12:24 - INFO - codeparrot_training - Step 3476: {'lr': 0.0004988343659532954, 'samples': 667584, 'steps': 3476, 'loss/train': 3.7230130434036255} 01/27/2022 23:12:29 - INFO - codeparrot_training - Step 3477: {'lr': 0.0004988327871989249, 'samples': 667776, 'steps': 3477, 'loss/train': 3.8168402910232544} 01/27/2022 23:12:33 - INFO - codeparrot_training - Step 3478: {'lr': 0.0004988312073786336, 'samples': 667968, 'steps': 3478, 'loss/train': 3.7969340085983276} 01/27/2022 23:12:38 - INFO - codeparrot_training - Step 3479: {'lr': 0.0004988296264924286, 'samples': 668160, 'steps': 3479, 'loss/train': 2.747434616088867} 01/27/2022 23:12:42 - INFO - codeparrot_training - Step 3480: {'lr': 0.0004988280445403164, 'samples': 668352, 'steps': 3480, 'loss/train': 3.953332543373108} 01/27/2022 23:12:46 - INFO - codeparrot_training - Step 3481: {'lr': 0.0004988264615223038, 'samples': 668544, 'steps': 3481, 'loss/train': 3.640150308609009} 01/27/2022 23:12:51 - INFO - codeparrot_training - Step 3482: {'lr': 0.0004988248774383978, 'samples': 668736, 'steps': 3482, 'loss/train': 3.922311544418335} 01/27/2022 23:12:55 - INFO - codeparrot_training - Step 3483: {'lr': 0.0004988232922886049, 'samples': 668928, 'steps': 3483, 'loss/train': 2.33420866727829} 01/27/2022 23:12:59 - INFO - codeparrot_training - Step 3484: {'lr': 0.0004988217060729321, 'samples': 669120, 'steps': 3484, 'loss/train': 3.172426700592041} 01/27/2022 23:13:04 - INFO - codeparrot_training - Step 3485: {'lr': 0.0004988201187913861, 'samples': 669312, 'steps': 3485, 'loss/train': 1.8050463795661926} 01/27/2022 23:13:08 - INFO - codeparrot_training - Step 3486: {'lr': 0.0004988185304439737, 'samples': 669504, 'steps': 3486, 'loss/train': 3.3263132572174072} 01/27/2022 23:13:13 - INFO - codeparrot_training - Step 3487: {'lr': 0.0004988169410307018, 'samples': 669696, 'steps': 3487, 'loss/train': 3.2248846292495728} 01/27/2022 23:13:17 - INFO - codeparrot_training - Step 3488: {'lr': 0.0004988153505515771, 'samples': 669888, 'steps': 3488, 'loss/train': 4.8504263162612915} 01/27/2022 23:13:21 - INFO - codeparrot_training - Step 3489: {'lr': 0.0004988137590066064, 'samples': 670080, 'steps': 3489, 'loss/train': 3.680567979812622} 01/27/2022 23:13:25 - INFO - codeparrot_training - Step 3490: {'lr': 0.0004988121663957966, 'samples': 670272, 'steps': 3490, 'loss/train': 3.2062182426452637} 01/27/2022 23:13:30 - INFO - codeparrot_training - Step 3491: {'lr': 0.0004988105727191546, 'samples': 670464, 'steps': 3491, 'loss/train': 3.146172881126404} 01/27/2022 23:13:35 - INFO - codeparrot_training - Step 3492: {'lr': 0.0004988089779766869, 'samples': 670656, 'steps': 3492, 'loss/train': 4.244987726211548} 01/27/2022 23:13:39 - INFO - codeparrot_training - Step 3493: {'lr': 0.0004988073821684006, 'samples': 670848, 'steps': 3493, 'loss/train': 1.7280623316764832} 01/27/2022 23:13:44 - INFO - codeparrot_training - Step 3494: {'lr': 0.0004988057852943025, 'samples': 671040, 'steps': 3494, 'loss/train': 4.565349698066711} 01/27/2022 23:13:48 - INFO - codeparrot_training - Step 3495: {'lr': 0.0004988041873543995, 'samples': 671232, 'steps': 3495, 'loss/train': 3.88287091255188} 01/27/2022 23:13:52 - INFO - codeparrot_training - Step 3496: {'lr': 0.0004988025883486983, 'samples': 671424, 'steps': 3496, 'loss/train': 4.156361818313599} 01/27/2022 23:13:57 - INFO - codeparrot_training - Step 3497: {'lr': 0.0004988009882772058, 'samples': 671616, 'steps': 3497, 'loss/train': 2.684070110321045} 01/27/2022 23:14:01 - INFO - codeparrot_training - Step 3498: {'lr': 0.0004987993871399289, 'samples': 671808, 'steps': 3498, 'loss/train': 2.2443559169769287} 01/27/2022 23:14:05 - INFO - codeparrot_training - Step 3499: {'lr': 0.0004987977849368744, 'samples': 672000, 'steps': 3499, 'loss/train': 3.697617530822754} 01/27/2022 23:14:10 - INFO - codeparrot_training - Step 3500: {'lr': 0.0004987961816680492, 'samples': 672192, 'steps': 3500, 'loss/train': 3.800976276397705} 01/27/2022 23:14:14 - INFO - codeparrot_training - Step 3501: {'lr': 0.0004987945773334602, 'samples': 672384, 'steps': 3501, 'loss/train': 3.97925341129303} 01/27/2022 23:14:19 - INFO - codeparrot_training - Step 3502: {'lr': 0.0004987929719331142, 'samples': 672576, 'steps': 3502, 'loss/train': 2.6393460631370544} 01/27/2022 23:14:23 - INFO - codeparrot_training - Step 3503: {'lr': 0.0004987913654670181, 'samples': 672768, 'steps': 3503, 'loss/train': 2.4846285581588745} 01/27/2022 23:14:28 - INFO - codeparrot_training - Step 3504: {'lr': 0.0004987897579351787, 'samples': 672960, 'steps': 3504, 'loss/train': 2.1982950568199158} 01/27/2022 23:14:32 - INFO - codeparrot_training - Step 3505: {'lr': 0.0004987881493376032, 'samples': 673152, 'steps': 3505, 'loss/train': 3.8646408319473267} 01/27/2022 23:14:36 - INFO - codeparrot_training - Step 3506: {'lr': 0.0004987865396742981, 'samples': 673344, 'steps': 3506, 'loss/train': 3.873883366584778} 01/27/2022 23:14:42 - INFO - codeparrot_training - Step 3507: {'lr': 0.0004987849289452705, 'samples': 673536, 'steps': 3507, 'loss/train': 3.086295247077942} 01/27/2022 23:14:46 - INFO - codeparrot_training - Step 3508: {'lr': 0.0004987833171505272, 'samples': 673728, 'steps': 3508, 'loss/train': 3.4967164993286133} 01/27/2022 23:14:50 - INFO - codeparrot_training - Step 3509: {'lr': 0.0004987817042900753, 'samples': 673920, 'steps': 3509, 'loss/train': 3.3129533529281616} 01/27/2022 23:14:54 - INFO - codeparrot_training - Step 3510: {'lr': 0.0004987800903639216, 'samples': 674112, 'steps': 3510, 'loss/train': 3.3099056482315063} 01/27/2022 23:14:58 - INFO - codeparrot_training - Step 3511: {'lr': 0.0004987784753720728, 'samples': 674304, 'steps': 3511, 'loss/train': 3.1991347074508667} 01/27/2022 23:15:04 - INFO - codeparrot_training - Step 3512: {'lr': 0.0004987768593145362, 'samples': 674496, 'steps': 3512, 'loss/train': 3.34154212474823} 01/27/2022 23:15:08 - INFO - codeparrot_training - Step 3513: {'lr': 0.0004987752421913185, 'samples': 674688, 'steps': 3513, 'loss/train': 1.8986746072769165} 01/27/2022 23:15:12 - INFO - codeparrot_training - Step 3514: {'lr': 0.0004987736240024264, 'samples': 674880, 'steps': 3514, 'loss/train': 4.832014203071594} 01/27/2022 23:15:16 - INFO - codeparrot_training - Step 3515: {'lr': 0.0004987720047478673, 'samples': 675072, 'steps': 3515, 'loss/train': 3.28797447681427} 01/27/2022 23:15:20 - INFO - codeparrot_training - Step 3516: {'lr': 0.000498770384427648, 'samples': 675264, 'steps': 3516, 'loss/train': 3.333463668823242} 01/27/2022 23:15:24 - INFO - codeparrot_training - Step 3517: {'lr': 0.0004987687630417753, 'samples': 675456, 'steps': 3517, 'loss/train': 3.711678385734558} 01/27/2022 23:15:30 - INFO - codeparrot_training - Step 3518: {'lr': 0.0004987671405902562, 'samples': 675648, 'steps': 3518, 'loss/train': 3.186054825782776} 01/27/2022 23:15:34 - INFO - codeparrot_training - Step 3519: {'lr': 0.0004987655170730976, 'samples': 675840, 'steps': 3519, 'loss/train': 2.9387855529785156} 01/27/2022 23:15:38 - INFO - codeparrot_training - Step 3520: {'lr': 0.0004987638924903066, 'samples': 676032, 'steps': 3520, 'loss/train': 3.509899377822876} 01/27/2022 23:15:43 - INFO - codeparrot_training - Step 3521: {'lr': 0.00049876226684189, 'samples': 676224, 'steps': 3521, 'loss/train': 3.258140802383423} 01/27/2022 23:15:47 - INFO - codeparrot_training - Step 3522: {'lr': 0.0004987606401278549, 'samples': 676416, 'steps': 3522, 'loss/train': 2.301091432571411} 01/27/2022 23:15:52 - INFO - codeparrot_training - Step 3523: {'lr': 0.0004987590123482082, 'samples': 676608, 'steps': 3523, 'loss/train': 2.4621371626853943} 01/27/2022 23:15:56 - INFO - codeparrot_training - Step 3524: {'lr': 0.0004987573835029569, 'samples': 676800, 'steps': 3524, 'loss/train': 3.325355887413025} 01/27/2022 23:16:00 - INFO - codeparrot_training - Step 3525: {'lr': 0.0004987557535921079, 'samples': 676992, 'steps': 3525, 'loss/train': 3.4502846002578735} 01/27/2022 23:16:05 - INFO - codeparrot_training - Step 3526: {'lr': 0.0004987541226156683, 'samples': 677184, 'steps': 3526, 'loss/train': 3.8051880598068237} 01/27/2022 23:16:10 - INFO - codeparrot_training - Step 3527: {'lr': 0.0004987524905736451, 'samples': 677376, 'steps': 3527, 'loss/train': 3.4025298357009888} 01/27/2022 23:16:14 - INFO - codeparrot_training - Step 3528: {'lr': 0.000498750857466045, 'samples': 677568, 'steps': 3528, 'loss/train': 3.8562358617782593} 01/27/2022 23:16:18 - INFO - codeparrot_training - Step 3529: {'lr': 0.0004987492232928753, 'samples': 677760, 'steps': 3529, 'loss/train': 4.421607971191406} 01/27/2022 23:16:22 - INFO - codeparrot_training - Step 3530: {'lr': 0.000498747588054143, 'samples': 677952, 'steps': 3530, 'loss/train': 2.5941959023475647} 01/27/2022 23:16:26 - INFO - codeparrot_training - Step 3531: {'lr': 0.0004987459517498549, 'samples': 678144, 'steps': 3531, 'loss/train': 3.3843272924423218} 01/27/2022 23:16:32 - INFO - codeparrot_training - Step 3532: {'lr': 0.0004987443143800182, 'samples': 678336, 'steps': 3532, 'loss/train': 4.402738809585571} 01/27/2022 23:16:37 - INFO - codeparrot_training - Step 3533: {'lr': 0.0004987426759446398, 'samples': 678528, 'steps': 3533, 'loss/train': 3.2776039838790894} 01/27/2022 23:16:41 - INFO - codeparrot_training - Step 3534: {'lr': 0.0004987410364437269, 'samples': 678720, 'steps': 3534, 'loss/train': 3.4921270608901978} 01/27/2022 23:16:45 - INFO - codeparrot_training - Step 3535: {'lr': 0.0004987393958772862, 'samples': 678912, 'steps': 3535, 'loss/train': 4.218548655509949} 01/27/2022 23:16:49 - INFO - codeparrot_training - Step 3536: {'lr': 0.0004987377542453251, 'samples': 679104, 'steps': 3536, 'loss/train': 2.694309711456299} 01/27/2022 23:16:54 - INFO - codeparrot_training - Step 3537: {'lr': 0.0004987361115478502, 'samples': 679296, 'steps': 3537, 'loss/train': 2.839621603488922} 01/27/2022 23:16:58 - INFO - codeparrot_training - Step 3538: {'lr': 0.000498734467784869, 'samples': 679488, 'steps': 3538, 'loss/train': 3.7593315839767456} 01/27/2022 23:17:02 - INFO - codeparrot_training - Step 3539: {'lr': 0.0004987328229563883, 'samples': 679680, 'steps': 3539, 'loss/train': 2.546055793762207} 01/27/2022 23:17:07 - INFO - codeparrot_training - Step 3540: {'lr': 0.0004987311770624151, 'samples': 679872, 'steps': 3540, 'loss/train': 4.630762338638306} 01/27/2022 23:17:11 - INFO - codeparrot_training - Step 3541: {'lr': 0.0004987295301029565, 'samples': 680064, 'steps': 3541, 'loss/train': 3.8485565185546875} 01/27/2022 23:17:16 - INFO - codeparrot_training - Step 3542: {'lr': 0.0004987278820780196, 'samples': 680256, 'steps': 3542, 'loss/train': 3.531485438346863} 01/27/2022 23:17:21 - INFO - codeparrot_training - Step 3543: {'lr': 0.0004987262329876114, 'samples': 680448, 'steps': 3543, 'loss/train': 3.425578236579895} 01/27/2022 23:17:25 - INFO - codeparrot_training - Step 3544: {'lr': 0.000498724582831739, 'samples': 680640, 'steps': 3544, 'loss/train': 3.2522231340408325} 01/27/2022 23:17:29 - INFO - codeparrot_training - Step 3545: {'lr': 0.0004987229316104095, 'samples': 680832, 'steps': 3545, 'loss/train': 3.152030825614929} 01/27/2022 23:17:33 - INFO - codeparrot_training - Step 3546: {'lr': 0.00049872127932363, 'samples': 681024, 'steps': 3546, 'loss/train': 2.369553565979004} 01/27/2022 23:17:37 - INFO - codeparrot_training - Step 3547: {'lr': 0.0004987196259714074, 'samples': 681216, 'steps': 3547, 'loss/train': 4.327896952629089} 01/27/2022 23:17:42 - INFO - codeparrot_training - Step 3548: {'lr': 0.000498717971553749, 'samples': 681408, 'steps': 3548, 'loss/train': 3.6664230823516846} 01/27/2022 23:17:47 - INFO - codeparrot_training - Step 3549: {'lr': 0.0004987163160706617, 'samples': 681600, 'steps': 3549, 'loss/train': 2.575393795967102} 01/27/2022 23:17:51 - INFO - codeparrot_training - Step 3550: {'lr': 0.0004987146595221527, 'samples': 681792, 'steps': 3550, 'loss/train': 4.040731072425842} 01/27/2022 23:17:55 - INFO - codeparrot_training - Step 3551: {'lr': 0.0004987130019082291, 'samples': 681984, 'steps': 3551, 'loss/train': 4.116085052490234} 01/27/2022 23:17:59 - INFO - codeparrot_training - Step 3552: {'lr': 0.000498711343228898, 'samples': 682176, 'steps': 3552, 'loss/train': 3.1829878091812134} 01/27/2022 23:18:05 - INFO - codeparrot_training - Step 3553: {'lr': 0.0004987096834841665, 'samples': 682368, 'steps': 3553, 'loss/train': 3.1399834156036377} 01/27/2022 23:18:09 - INFO - codeparrot_training - Step 3554: {'lr': 0.0004987080226740416, 'samples': 682560, 'steps': 3554, 'loss/train': 3.674511194229126} 01/27/2022 23:18:13 - INFO - codeparrot_training - Step 3555: {'lr': 0.0004987063607985305, 'samples': 682752, 'steps': 3555, 'loss/train': 3.9035321474075317} 01/27/2022 23:18:17 - INFO - codeparrot_training - Step 3556: {'lr': 0.0004987046978576404, 'samples': 682944, 'steps': 3556, 'loss/train': 3.8730475902557373} 01/27/2022 23:18:22 - INFO - codeparrot_training - Step 3557: {'lr': 0.0004987030338513783, 'samples': 683136, 'steps': 3557, 'loss/train': 4.527485132217407} 01/27/2022 23:18:27 - INFO - codeparrot_training - Step 3558: {'lr': 0.0004987013687797514, 'samples': 683328, 'steps': 3558, 'loss/train': 3.4452202320098877} 01/27/2022 23:18:31 - INFO - codeparrot_training - Step 3559: {'lr': 0.0004986997026427668, 'samples': 683520, 'steps': 3559, 'loss/train': 2.504199206829071} 01/27/2022 23:18:35 - INFO - codeparrot_training - Step 3560: {'lr': 0.0004986980354404316, 'samples': 683712, 'steps': 3560, 'loss/train': 5.979251861572266} 01/27/2022 23:18:39 - INFO - codeparrot_training - Step 3561: {'lr': 0.000498696367172753, 'samples': 683904, 'steps': 3561, 'loss/train': 3.1318774223327637} 01/27/2022 23:18:43 - INFO - codeparrot_training - Step 3562: {'lr': 0.0004986946978397382, 'samples': 684096, 'steps': 3562, 'loss/train': 3.0848429203033447} 01/27/2022 23:18:48 - INFO - codeparrot_training - Step 3563: {'lr': 0.0004986930274413942, 'samples': 684288, 'steps': 3563, 'loss/train': 2.5405579805374146} 01/27/2022 23:18:53 - INFO - codeparrot_training - Step 3564: {'lr': 0.0004986913559777283, 'samples': 684480, 'steps': 3564, 'loss/train': 4.111867904663086} 01/27/2022 23:18:57 - INFO - codeparrot_training - Step 3565: {'lr': 0.0004986896834487477, 'samples': 684672, 'steps': 3565, 'loss/train': 4.135714173316956} 01/27/2022 23:19:01 - INFO - codeparrot_training - Step 3566: {'lr': 0.0004986880098544593, 'samples': 684864, 'steps': 3566, 'loss/train': 4.388636112213135} 01/27/2022 23:19:05 - INFO - codeparrot_training - Step 3567: {'lr': 0.0004986863351948705, 'samples': 685056, 'steps': 3567, 'loss/train': 2.763412356376648} 01/27/2022 23:19:11 - INFO - codeparrot_training - Step 3568: {'lr': 0.0004986846594699883, 'samples': 685248, 'steps': 3568, 'loss/train': 3.6366244554519653} 01/27/2022 23:19:15 - INFO - codeparrot_training - Step 3569: {'lr': 0.0004986829826798202, 'samples': 685440, 'steps': 3569, 'loss/train': 3.740988850593567} 01/27/2022 23:19:19 - INFO - codeparrot_training - Step 3570: {'lr': 0.0004986813048243729, 'samples': 685632, 'steps': 3570, 'loss/train': 3.5944100618362427} 01/27/2022 23:19:23 - INFO - codeparrot_training - Step 3571: {'lr': 0.000498679625903654, 'samples': 685824, 'steps': 3571, 'loss/train': 3.1068109273910522} 01/27/2022 23:19:28 - INFO - codeparrot_training - Step 3572: {'lr': 0.0004986779459176706, 'samples': 686016, 'steps': 3572, 'loss/train': 3.551889181137085} 01/27/2022 23:19:33 - INFO - codeparrot_training - Step 3573: {'lr': 0.0004986762648664298, 'samples': 686208, 'steps': 3573, 'loss/train': 3.0169025659561157} 01/27/2022 23:19:37 - INFO - codeparrot_training - Step 3574: {'lr': 0.0004986745827499389, 'samples': 686400, 'steps': 3574, 'loss/train': 2.20835942029953} 01/27/2022 23:19:41 - INFO - codeparrot_training - Step 3575: {'lr': 0.0004986728995682049, 'samples': 686592, 'steps': 3575, 'loss/train': 3.0586230754852295} 01/27/2022 23:19:45 - INFO - codeparrot_training - Step 3576: {'lr': 0.0004986712153212352, 'samples': 686784, 'steps': 3576, 'loss/train': 4.005090951919556} 01/27/2022 23:19:49 - INFO - codeparrot_training - Step 3577: {'lr': 0.0004986695300090371, 'samples': 686976, 'steps': 3577, 'loss/train': 2.708167791366577} 01/27/2022 23:19:55 - INFO - codeparrot_training - Step 3578: {'lr': 0.0004986678436316175, 'samples': 687168, 'steps': 3578, 'loss/train': 4.1298744678497314} 01/27/2022 23:19:59 - INFO - codeparrot_training - Step 3579: {'lr': 0.000498666156188984, 'samples': 687360, 'steps': 3579, 'loss/train': 3.2527520656585693} 01/27/2022 23:20:03 - INFO - codeparrot_training - Step 3580: {'lr': 0.0004986644676811436, 'samples': 687552, 'steps': 3580, 'loss/train': 3.545482635498047} 01/27/2022 23:20:08 - INFO - codeparrot_training - Step 3581: {'lr': 0.0004986627781081035, 'samples': 687744, 'steps': 3581, 'loss/train': 2.820754051208496} 01/27/2022 23:20:12 - INFO - codeparrot_training - Step 3582: {'lr': 0.0004986610874698712, 'samples': 687936, 'steps': 3582, 'loss/train': 3.5730435848236084} 01/27/2022 23:20:17 - INFO - codeparrot_training - Step 3583: {'lr': 0.0004986593957664536, 'samples': 688128, 'steps': 3583, 'loss/train': 2.805766761302948} 01/27/2022 23:20:21 - INFO - codeparrot_training - Step 3584: {'lr': 0.0004986577029978581, 'samples': 688320, 'steps': 3584, 'loss/train': 1.873526930809021} 01/27/2022 23:20:25 - INFO - codeparrot_training - Step 3585: {'lr': 0.000498656009164092, 'samples': 688512, 'steps': 3585, 'loss/train': 3.6976583003997803} 01/27/2022 23:20:29 - INFO - codeparrot_training - Step 3586: {'lr': 0.0004986543142651625, 'samples': 688704, 'steps': 3586, 'loss/train': 2.7537968158721924} 01/27/2022 23:20:34 - INFO - codeparrot_training - Step 3587: {'lr': 0.0004986526183010769, 'samples': 688896, 'steps': 3587, 'loss/train': 2.1265609860420227} 01/27/2022 23:20:39 - INFO - codeparrot_training - Step 3588: {'lr': 0.0004986509212718425, 'samples': 689088, 'steps': 3588, 'loss/train': 1.7753849029541016} 01/27/2022 23:20:43 - INFO - codeparrot_training - Step 3589: {'lr': 0.0004986492231774664, 'samples': 689280, 'steps': 3589, 'loss/train': 4.477480173110962} 01/27/2022 23:20:47 - INFO - codeparrot_training - Step 3590: {'lr': 0.0004986475240179559, 'samples': 689472, 'steps': 3590, 'loss/train': 3.2664366960525513} 01/27/2022 23:20:51 - INFO - codeparrot_training - Step 3591: {'lr': 0.0004986458237933185, 'samples': 689664, 'steps': 3591, 'loss/train': 2.4888362288475037} 01/27/2022 23:20:55 - INFO - codeparrot_training - Step 3592: {'lr': 0.0004986441225035614, 'samples': 689856, 'steps': 3592, 'loss/train': 2.925217866897583} 01/27/2022 23:21:01 - INFO - codeparrot_training - Step 3593: {'lr': 0.0004986424201486918, 'samples': 690048, 'steps': 3593, 'loss/train': 2.741828978061676} 01/27/2022 23:21:05 - INFO - codeparrot_training - Step 3594: {'lr': 0.000498640716728717, 'samples': 690240, 'steps': 3594, 'loss/train': 1.2804584205150604} 01/27/2022 23:21:10 - INFO - codeparrot_training - Step 3595: {'lr': 0.0004986390122436443, 'samples': 690432, 'steps': 3595, 'loss/train': 4.271853446960449} 01/27/2022 23:21:14 - INFO - codeparrot_training - Step 3596: {'lr': 0.000498637306693481, 'samples': 690624, 'steps': 3596, 'loss/train': 3.617730975151062} 01/27/2022 23:21:18 - INFO - codeparrot_training - Step 3597: {'lr': 0.0004986356000782345, 'samples': 690816, 'steps': 3597, 'loss/train': 3.8264418840408325} 01/27/2022 23:21:23 - INFO - codeparrot_training - Step 3598: {'lr': 0.0004986338923979119, 'samples': 691008, 'steps': 3598, 'loss/train': 3.595512270927429} 01/27/2022 23:21:28 - INFO - codeparrot_training - Step 3599: {'lr': 0.0004986321836525209, 'samples': 691200, 'steps': 3599, 'loss/train': 3.0915430784225464} 01/27/2022 23:21:32 - INFO - codeparrot_training - Step 3600: {'lr': 0.0004986304738420684, 'samples': 691392, 'steps': 3600, 'loss/train': 4.40999972820282} 01/27/2022 23:21:36 - INFO - codeparrot_training - Step 3601: {'lr': 0.0004986287629665619, 'samples': 691584, 'steps': 3601, 'loss/train': 2.747349500656128} 01/27/2022 23:21:42 - INFO - codeparrot_training - Step 3602: {'lr': 0.0004986270510260087, 'samples': 691776, 'steps': 3602, 'loss/train': 1.6491653323173523} 01/27/2022 23:21:46 - INFO - codeparrot_training - Step 3603: {'lr': 0.0004986253380204163, 'samples': 691968, 'steps': 3603, 'loss/train': 3.30295193195343} 01/27/2022 23:21:50 - INFO - codeparrot_training - Step 3604: {'lr': 0.0004986236239497918, 'samples': 692160, 'steps': 3604, 'loss/train': 2.349478304386139} 01/27/2022 23:21:54 - INFO - codeparrot_training - Step 3605: {'lr': 0.0004986219088141426, 'samples': 692352, 'steps': 3605, 'loss/train': 3.333772659301758} 01/27/2022 23:21:58 - INFO - codeparrot_training - Step 3606: {'lr': 0.0004986201926134761, 'samples': 692544, 'steps': 3606, 'loss/train': 3.4731903076171875} 01/27/2022 23:22:03 - INFO - codeparrot_training - Step 3607: {'lr': 0.0004986184753477998, 'samples': 692736, 'steps': 3607, 'loss/train': 1.352493792772293} 01/27/2022 23:22:07 - INFO - codeparrot_training - Step 3608: {'lr': 0.0004986167570171208, 'samples': 692928, 'steps': 3608, 'loss/train': 3.309451460838318} 01/27/2022 23:22:12 - INFO - codeparrot_training - Step 3609: {'lr': 0.0004986150376214465, 'samples': 693120, 'steps': 3609, 'loss/train': 3.5920647382736206} 01/27/2022 23:22:16 - INFO - codeparrot_training - Step 3610: {'lr': 0.0004986133171607844, 'samples': 693312, 'steps': 3610, 'loss/train': 4.343097567558289} 01/27/2022 23:22:20 - INFO - codeparrot_training - Step 3611: {'lr': 0.0004986115956351417, 'samples': 693504, 'steps': 3611, 'loss/train': 2.9106886982917786} 01/27/2022 23:22:25 - INFO - codeparrot_training - Step 3612: {'lr': 0.000498609873044526, 'samples': 693696, 'steps': 3612, 'loss/train': 4.031979560852051} 01/27/2022 23:22:30 - INFO - codeparrot_training - Step 3613: {'lr': 0.0004986081493889444, 'samples': 693888, 'steps': 3613, 'loss/train': 3.9738088846206665} 01/27/2022 23:22:35 - INFO - codeparrot_training - Step 3614: {'lr': 0.0004986064246684046, 'samples': 694080, 'steps': 3614, 'loss/train': 2.6148873567581177} 01/27/2022 23:22:39 - INFO - codeparrot_training - Step 3615: {'lr': 0.0004986046988829136, 'samples': 694272, 'steps': 3615, 'loss/train': 3.7090208530426025} 01/27/2022 23:22:43 - INFO - codeparrot_training - Step 3616: {'lr': 0.0004986029720324791, 'samples': 694464, 'steps': 3616, 'loss/train': 3.2338428497314453} 01/27/2022 23:22:47 - INFO - codeparrot_training - Step 3617: {'lr': 0.0004986012441171085, 'samples': 694656, 'steps': 3617, 'loss/train': 4.382489204406738} 01/27/2022 23:22:52 - INFO - codeparrot_training - Step 3618: {'lr': 0.000498599515136809, 'samples': 694848, 'steps': 3618, 'loss/train': 2.6345930099487305} 01/27/2022 23:22:56 - INFO - codeparrot_training - Step 3619: {'lr': 0.0004985977850915882, 'samples': 695040, 'steps': 3619, 'loss/train': 3.1791930198669434} 01/27/2022 23:23:00 - INFO - codeparrot_training - Step 3620: {'lr': 0.0004985960539814534, 'samples': 695232, 'steps': 3620, 'loss/train': 3.186521530151367} 01/27/2022 23:23:05 - INFO - codeparrot_training - Step 3621: {'lr': 0.000498594321806412, 'samples': 695424, 'steps': 3621, 'loss/train': 3.739291548728943} 01/27/2022 23:23:09 - INFO - codeparrot_training - Step 3622: {'lr': 0.0004985925885664716, 'samples': 695616, 'steps': 3622, 'loss/train': 3.6443281173706055} 01/27/2022 23:23:15 - INFO - codeparrot_training - Step 3623: {'lr': 0.0004985908542616393, 'samples': 695808, 'steps': 3623, 'loss/train': 4.549105525016785} 01/27/2022 23:23:19 - INFO - codeparrot_training - Step 3624: {'lr': 0.0004985891188919229, 'samples': 696000, 'steps': 3624, 'loss/train': 3.013370990753174} 01/27/2022 23:23:23 - INFO - codeparrot_training - Step 3625: {'lr': 0.0004985873824573296, 'samples': 696192, 'steps': 3625, 'loss/train': 2.719703257083893} 01/27/2022 23:23:27 - INFO - codeparrot_training - Step 3626: {'lr': 0.0004985856449578667, 'samples': 696384, 'steps': 3626, 'loss/train': 3.757312774658203} 01/27/2022 23:23:31 - INFO - codeparrot_training - Step 3627: {'lr': 0.0004985839063935421, 'samples': 696576, 'steps': 3627, 'loss/train': 3.753095269203186} 01/27/2022 23:23:36 - INFO - codeparrot_training - Step 3628: {'lr': 0.0004985821667643628, 'samples': 696768, 'steps': 3628, 'loss/train': 2.5673203468322754} 01/27/2022 23:23:41 - INFO - codeparrot_training - Step 3629: {'lr': 0.0004985804260703364, 'samples': 696960, 'steps': 3629, 'loss/train': 4.613164901733398} 01/27/2022 23:23:45 - INFO - codeparrot_training - Step 3630: {'lr': 0.0004985786843114706, 'samples': 697152, 'steps': 3630, 'loss/train': 3.6172878742218018} 01/27/2022 23:23:49 - INFO - codeparrot_training - Step 3631: {'lr': 0.0004985769414877725, 'samples': 697344, 'steps': 3631, 'loss/train': 4.294896125793457} 01/27/2022 23:23:53 - INFO - codeparrot_training - Step 3632: {'lr': 0.0004985751975992497, 'samples': 697536, 'steps': 3632, 'loss/train': 4.491665840148926} 01/27/2022 23:23:58 - INFO - codeparrot_training - Step 3633: {'lr': 0.0004985734526459098, 'samples': 697728, 'steps': 3633, 'loss/train': 4.619938373565674} 01/27/2022 23:24:02 - INFO - codeparrot_training - Step 3634: {'lr': 0.0004985717066277601, 'samples': 697920, 'steps': 3634, 'loss/train': 2.1385390162467957} 01/27/2022 23:24:07 - INFO - codeparrot_training - Step 3635: {'lr': 0.0004985699595448081, 'samples': 698112, 'steps': 3635, 'loss/train': 3.546459674835205} 01/27/2022 23:24:11 - INFO - codeparrot_training - Step 3636: {'lr': 0.0004985682113970613, 'samples': 698304, 'steps': 3636, 'loss/train': 3.094549298286438} 01/27/2022 23:24:15 - INFO - codeparrot_training - Step 3637: {'lr': 0.0004985664621845273, 'samples': 698496, 'steps': 3637, 'loss/train': 3.4297817945480347} 01/27/2022 23:24:21 - INFO - codeparrot_training - Step 3638: {'lr': 0.0004985647119072135, 'samples': 698688, 'steps': 3638, 'loss/train': 3.3389618396759033} 01/27/2022 23:24:25 - INFO - codeparrot_training - Step 3639: {'lr': 0.0004985629605651273, 'samples': 698880, 'steps': 3639, 'loss/train': 3.822722554206848} 01/27/2022 23:24:29 - INFO - codeparrot_training - Step 3640: {'lr': 0.0004985612081582763, 'samples': 699072, 'steps': 3640, 'loss/train': 3.60599684715271} 01/27/2022 23:24:33 - INFO - codeparrot_training - Step 3641: {'lr': 0.0004985594546866682, 'samples': 699264, 'steps': 3641, 'loss/train': 2.988661766052246} 01/27/2022 23:24:38 - INFO - codeparrot_training - Step 3642: {'lr': 0.0004985577001503102, 'samples': 699456, 'steps': 3642, 'loss/train': 3.8399012088775635} 01/27/2022 23:24:42 - INFO - codeparrot_training - Step 3643: {'lr': 0.0004985559445492099, 'samples': 699648, 'steps': 3643, 'loss/train': 4.1877840757369995} 01/27/2022 23:24:47 - INFO - codeparrot_training - Step 3644: {'lr': 0.0004985541878833749, 'samples': 699840, 'steps': 3644, 'loss/train': 5.191302537918091} 01/27/2022 23:24:51 - INFO - codeparrot_training - Step 3645: {'lr': 0.0004985524301528127, 'samples': 700032, 'steps': 3645, 'loss/train': 5.0656421184539795} 01/27/2022 23:24:55 - INFO - codeparrot_training - Step 3646: {'lr': 0.0004985506713575307, 'samples': 700224, 'steps': 3646, 'loss/train': 3.123892664909363} 01/27/2022 23:25:00 - INFO - codeparrot_training - Step 3647: {'lr': 0.0004985489114975368, 'samples': 700416, 'steps': 3647, 'loss/train': 4.6742695569992065} 01/27/2022 23:25:05 - INFO - codeparrot_training - Step 3648: {'lr': 0.0004985471505728381, 'samples': 700608, 'steps': 3648, 'loss/train': 3.449152708053589} 01/27/2022 23:25:10 - INFO - codeparrot_training - Step 3649: {'lr': 0.0004985453885834423, 'samples': 700800, 'steps': 3649, 'loss/train': 3.681078314781189} 01/27/2022 23:25:14 - INFO - codeparrot_training - Step 3650: {'lr': 0.0004985436255293571, 'samples': 700992, 'steps': 3650, 'loss/train': 2.4295127391815186} 01/27/2022 23:25:18 - INFO - codeparrot_training - Step 3651: {'lr': 0.0004985418614105898, 'samples': 701184, 'steps': 3651, 'loss/train': 2.9959543347358704} 01/27/2022 23:25:22 - INFO - codeparrot_training - Step 3652: {'lr': 0.0004985400962271482, 'samples': 701376, 'steps': 3652, 'loss/train': 2.1120543479919434} 01/27/2022 23:25:27 - INFO - codeparrot_training - Step 3653: {'lr': 0.0004985383299790397, 'samples': 701568, 'steps': 3653, 'loss/train': 2.79321026802063} 01/27/2022 23:25:32 - INFO - codeparrot_training - Step 3654: {'lr': 0.0004985365626662719, 'samples': 701760, 'steps': 3654, 'loss/train': 3.099403738975525} 01/27/2022 23:25:36 - INFO - codeparrot_training - Step 3655: {'lr': 0.0004985347942888524, 'samples': 701952, 'steps': 3655, 'loss/train': 3.204015612602234} 01/27/2022 23:25:40 - INFO - codeparrot_training - Step 3656: {'lr': 0.0004985330248467888, 'samples': 702144, 'steps': 3656, 'loss/train': 4.6814693212509155} 01/27/2022 23:25:44 - INFO - codeparrot_training - Step 3657: {'lr': 0.0004985312543400886, 'samples': 702336, 'steps': 3657, 'loss/train': 4.130525708198547} 01/27/2022 23:25:49 - INFO - codeparrot_training - Step 3658: {'lr': 0.0004985294827687594, 'samples': 702528, 'steps': 3658, 'loss/train': 3.373075246810913} 01/27/2022 23:25:53 - INFO - codeparrot_training - Step 3659: {'lr': 0.0004985277101328088, 'samples': 702720, 'steps': 3659, 'loss/train': 3.534340739250183} 01/27/2022 23:25:58 - INFO - codeparrot_training - Step 3660: {'lr': 0.0004985259364322445, 'samples': 702912, 'steps': 3660, 'loss/train': 3.2291042804718018} 01/27/2022 23:26:02 - INFO - codeparrot_training - Step 3661: {'lr': 0.0004985241616670739, 'samples': 703104, 'steps': 3661, 'loss/train': 3.832661747932434} 01/27/2022 23:26:06 - INFO - codeparrot_training - Step 3662: {'lr': 0.0004985223858373048, 'samples': 703296, 'steps': 3662, 'loss/train': 3.306546449661255} 01/27/2022 23:26:12 - INFO - codeparrot_training - Step 3663: {'lr': 0.0004985206089429447, 'samples': 703488, 'steps': 3663, 'loss/train': 2.6208436489105225} 01/27/2022 23:26:16 - INFO - codeparrot_training - Step 3664: {'lr': 0.0004985188309840012, 'samples': 703680, 'steps': 3664, 'loss/train': 3.638591766357422} 01/27/2022 23:26:20 - INFO - codeparrot_training - Step 3665: {'lr': 0.0004985170519604819, 'samples': 703872, 'steps': 3665, 'loss/train': 4.063642144203186} 01/27/2022 23:26:24 - INFO - codeparrot_training - Step 3666: {'lr': 0.0004985152718723944, 'samples': 704064, 'steps': 3666, 'loss/train': 2.714570939540863} 01/27/2022 23:26:28 - INFO - codeparrot_training - Step 3667: {'lr': 0.0004985134907197466, 'samples': 704256, 'steps': 3667, 'loss/train': 3.011927604675293} 01/27/2022 23:26:33 - INFO - codeparrot_training - Step 3668: {'lr': 0.0004985117085025458, 'samples': 704448, 'steps': 3668, 'loss/train': 3.9770103693008423} 01/27/2022 23:26:38 - INFO - codeparrot_training - Step 3669: {'lr': 0.0004985099252207998, 'samples': 704640, 'steps': 3669, 'loss/train': 2.8204309344291687} 01/27/2022 23:26:42 - INFO - codeparrot_training - Step 3670: {'lr': 0.0004985081408745161, 'samples': 704832, 'steps': 3670, 'loss/train': 3.5869381427764893} 01/27/2022 23:26:46 - INFO - codeparrot_training - Step 3671: {'lr': 0.0004985063554637025, 'samples': 705024, 'steps': 3671, 'loss/train': 3.039162755012512} 01/27/2022 23:26:50 - INFO - codeparrot_training - Step 3672: {'lr': 0.0004985045689883665, 'samples': 705216, 'steps': 3672, 'loss/train': 3.579858183860779} 01/27/2022 23:26:55 - INFO - codeparrot_training - Step 3673: {'lr': 0.0004985027814485159, 'samples': 705408, 'steps': 3673, 'loss/train': 3.2261863946914673} 01/27/2022 23:26:59 - INFO - codeparrot_training - Step 3674: {'lr': 0.0004985009928441584, 'samples': 705600, 'steps': 3674, 'loss/train': 3.588984489440918} 01/27/2022 23:27:04 - INFO - codeparrot_training - Step 3675: {'lr': 0.0004984992031753014, 'samples': 705792, 'steps': 3675, 'loss/train': 3.2678768634796143} 01/27/2022 23:27:08 - INFO - codeparrot_training - Step 3676: {'lr': 0.0004984974124419528, 'samples': 705984, 'steps': 3676, 'loss/train': 1.4035325646400452} 01/27/2022 23:27:12 - INFO - codeparrot_training - Step 3677: {'lr': 0.0004984956206441201, 'samples': 706176, 'steps': 3677, 'loss/train': 2.9236008524894714} 01/27/2022 23:27:17 - INFO - codeparrot_training - Step 3678: {'lr': 0.0004984938277818112, 'samples': 706368, 'steps': 3678, 'loss/train': 2.169958233833313} 01/27/2022 23:27:21 - INFO - codeparrot_training - Step 3679: {'lr': 0.0004984920338550335, 'samples': 706560, 'steps': 3679, 'loss/train': 3.4971842765808105} 01/27/2022 23:27:25 - INFO - codeparrot_training - Step 3680: {'lr': 0.0004984902388637949, 'samples': 706752, 'steps': 3680, 'loss/train': 3.3180384635925293} 01/27/2022 23:27:29 - INFO - codeparrot_training - Step 3681: {'lr': 0.0004984884428081031, 'samples': 706944, 'steps': 3681, 'loss/train': 2.877655506134033} 01/27/2022 23:27:34 - INFO - codeparrot_training - Step 3682: {'lr': 0.0004984866456879657, 'samples': 707136, 'steps': 3682, 'loss/train': 3.787730097770691} 01/27/2022 23:27:39 - INFO - codeparrot_training - Step 3683: {'lr': 0.0004984848475033903, 'samples': 707328, 'steps': 3683, 'loss/train': 1.9699950814247131} 01/27/2022 23:27:44 - INFO - codeparrot_training - Step 3684: {'lr': 0.0004984830482543847, 'samples': 707520, 'steps': 3684, 'loss/train': 3.1449276208877563} 01/27/2022 23:27:48 - INFO - codeparrot_training - Step 3685: {'lr': 0.0004984812479409568, 'samples': 707712, 'steps': 3685, 'loss/train': 2.3164061307907104} 01/27/2022 23:27:52 - INFO - codeparrot_training - Step 3686: {'lr': 0.000498479446563114, 'samples': 707904, 'steps': 3686, 'loss/train': 4.058375716209412} 01/27/2022 23:27:56 - INFO - codeparrot_training - Step 3687: {'lr': 0.0004984776441208642, 'samples': 708096, 'steps': 3687, 'loss/train': 3.672360420227051} 01/27/2022 23:28:01 - INFO - codeparrot_training - Step 3688: {'lr': 0.000498475840614215, 'samples': 708288, 'steps': 3688, 'loss/train': 3.6309388875961304} 01/27/2022 23:28:05 - INFO - codeparrot_training - Step 3689: {'lr': 0.0004984740360431742, 'samples': 708480, 'steps': 3689, 'loss/train': 3.4967751502990723} 01/27/2022 23:28:10 - INFO - codeparrot_training - Step 3690: {'lr': 0.0004984722304077496, 'samples': 708672, 'steps': 3690, 'loss/train': 1.7336872816085815} 01/27/2022 23:28:14 - INFO - codeparrot_training - Step 3691: {'lr': 0.0004984704237079489, 'samples': 708864, 'steps': 3691, 'loss/train': 4.150395512580872} 01/27/2022 23:28:18 - INFO - codeparrot_training - Step 3692: {'lr': 0.0004984686159437798, 'samples': 709056, 'steps': 3692, 'loss/train': 2.880119204521179} 01/27/2022 23:28:23 - INFO - codeparrot_training - Step 3693: {'lr': 0.00049846680711525, 'samples': 709248, 'steps': 3693, 'loss/train': 2.4855052828788757} 01/27/2022 23:28:27 - INFO - codeparrot_training - Step 3694: {'lr': 0.0004984649972223673, 'samples': 709440, 'steps': 3694, 'loss/train': 3.181061625480652} 01/27/2022 23:28:31 - INFO - codeparrot_training - Step 3695: {'lr': 0.0004984631862651395, 'samples': 709632, 'steps': 3695, 'loss/train': 2.691763937473297} 01/27/2022 23:28:35 - INFO - codeparrot_training - Step 3696: {'lr': 0.0004984613742435742, 'samples': 709824, 'steps': 3696, 'loss/train': 2.522276222705841} 01/27/2022 23:28:40 - INFO - codeparrot_training - Step 3697: {'lr': 0.0004984595611576793, 'samples': 710016, 'steps': 3697, 'loss/train': 3.7794524431228638} 01/27/2022 23:28:45 - INFO - codeparrot_training - Step 3698: {'lr': 0.0004984577470074625, 'samples': 710208, 'steps': 3698, 'loss/train': 0.8813818395137787} 01/27/2022 23:28:49 - INFO - codeparrot_training - Step 3699: {'lr': 0.0004984559317929317, 'samples': 710400, 'steps': 3699, 'loss/train': 3.2851470708847046} 01/27/2022 23:28:54 - INFO - codeparrot_training - Step 3700: {'lr': 0.0004984541155140946, 'samples': 710592, 'steps': 3700, 'loss/train': 3.7922054529190063} 01/27/2022 23:28:58 - INFO - codeparrot_training - Step 3701: {'lr': 0.0004984522981709589, 'samples': 710784, 'steps': 3701, 'loss/train': 3.556020140647888} 01/27/2022 23:29:02 - INFO - codeparrot_training - Step 3702: {'lr': 0.0004984504797635324, 'samples': 710976, 'steps': 3702, 'loss/train': 3.5874931812286377} 01/27/2022 23:29:07 - INFO - codeparrot_training - Step 3703: {'lr': 0.000498448660291823, 'samples': 711168, 'steps': 3703, 'loss/train': 2.851816713809967} 01/27/2022 23:29:12 - INFO - codeparrot_training - Step 3704: {'lr': 0.0004984468397558384, 'samples': 711360, 'steps': 3704, 'loss/train': 3.48106849193573} 01/27/2022 23:29:16 - INFO - codeparrot_training - Step 3705: {'lr': 0.0004984450181555864, 'samples': 711552, 'steps': 3705, 'loss/train': 2.8658317923545837} 01/27/2022 23:29:20 - INFO - codeparrot_training - Step 3706: {'lr': 0.0004984431954910749, 'samples': 711744, 'steps': 3706, 'loss/train': 3.63983952999115} 01/27/2022 23:29:24 - INFO - codeparrot_training - Step 3707: {'lr': 0.0004984413717623117, 'samples': 711936, 'steps': 3707, 'loss/train': 2.8189845085144043} 01/27/2022 23:29:30 - INFO - codeparrot_training - Step 3708: {'lr': 0.0004984395469693044, 'samples': 712128, 'steps': 3708, 'loss/train': 3.5139466524124146} 01/27/2022 23:29:34 - INFO - codeparrot_training - Step 3709: {'lr': 0.000498437721112061, 'samples': 712320, 'steps': 3709, 'loss/train': 3.550000548362732} 01/27/2022 23:29:38 - INFO - codeparrot_training - Step 3710: {'lr': 0.0004984358941905894, 'samples': 712512, 'steps': 3710, 'loss/train': 3.6683539152145386} 01/27/2022 23:29:42 - INFO - codeparrot_training - Step 3711: {'lr': 0.0004984340662048972, 'samples': 712704, 'steps': 3711, 'loss/train': 2.799821197986603} 01/27/2022 23:29:46 - INFO - codeparrot_training - Step 3712: {'lr': 0.0004984322371549924, 'samples': 712896, 'steps': 3712, 'loss/train': 3.6901878118515015} 01/27/2022 23:29:51 - INFO - codeparrot_training - Step 3713: {'lr': 0.0004984304070408828, 'samples': 713088, 'steps': 3713, 'loss/train': 3.23579478263855} 01/27/2022 23:29:55 - INFO - codeparrot_training - Step 3714: {'lr': 0.0004984285758625761, 'samples': 713280, 'steps': 3714, 'loss/train': 3.723939299583435} 01/27/2022 23:30:00 - INFO - codeparrot_training - Step 3715: {'lr': 0.0004984267436200805, 'samples': 713472, 'steps': 3715, 'loss/train': 3.0842549800872803} 01/27/2022 23:30:04 - INFO - codeparrot_training - Step 3716: {'lr': 0.0004984249103134035, 'samples': 713664, 'steps': 3716, 'loss/train': 3.7217084169387817} 01/27/2022 23:30:08 - INFO - codeparrot_training - Step 3717: {'lr': 0.000498423075942553, 'samples': 713856, 'steps': 3717, 'loss/train': 1.5084366202354431} 01/27/2022 23:30:13 - INFO - codeparrot_training - Step 3718: {'lr': 0.0004984212405075369, 'samples': 714048, 'steps': 3718, 'loss/train': 2.02521675825119} 01/27/2022 23:30:17 - INFO - codeparrot_training - Step 3719: {'lr': 0.0004984194040083632, 'samples': 714240, 'steps': 3719, 'loss/train': 3.498010754585266} 01/27/2022 23:30:21 - INFO - codeparrot_training - Step 3720: {'lr': 0.0004984175664450397, 'samples': 714432, 'steps': 3720, 'loss/train': 4.892825961112976} 01/27/2022 23:30:25 - INFO - codeparrot_training - Step 3721: {'lr': 0.0004984157278175741, 'samples': 714624, 'steps': 3721, 'loss/train': 3.267124056816101} 01/27/2022 23:30:30 - INFO - codeparrot_training - Step 3722: {'lr': 0.0004984138881259744, 'samples': 714816, 'steps': 3722, 'loss/train': 1.3665830790996552} 01/27/2022 23:30:35 - INFO - codeparrot_training - Step 3723: {'lr': 0.0004984120473702486, 'samples': 715008, 'steps': 3723, 'loss/train': 3.5753045082092285} 01/27/2022 23:30:40 - INFO - codeparrot_training - Step 3724: {'lr': 0.0004984102055504044, 'samples': 715200, 'steps': 3724, 'loss/train': 3.259494423866272} 01/27/2022 23:30:44 - INFO - codeparrot_training - Step 3725: {'lr': 0.0004984083626664497, 'samples': 715392, 'steps': 3725, 'loss/train': 3.415322184562683} 01/27/2022 23:30:48 - INFO - codeparrot_training - Step 3726: {'lr': 0.0004984065187183925, 'samples': 715584, 'steps': 3726, 'loss/train': 2.4715635180473328} 01/27/2022 23:30:52 - INFO - codeparrot_training - Step 3727: {'lr': 0.0004984046737062407, 'samples': 715776, 'steps': 3727, 'loss/train': 2.864511787891388} 01/27/2022 23:30:57 - INFO - codeparrot_training - Step 3728: {'lr': 0.0004984028276300021, 'samples': 715968, 'steps': 3728, 'loss/train': 3.3400529623031616} 01/27/2022 23:31:02 - INFO - codeparrot_training - Step 3729: {'lr': 0.0004984009804896846, 'samples': 716160, 'steps': 3729, 'loss/train': 2.9856505393981934} 01/27/2022 23:31:06 - INFO - codeparrot_training - Step 3730: {'lr': 0.0004983991322852963, 'samples': 716352, 'steps': 3730, 'loss/train': 3.7067813873291016} 01/27/2022 23:31:10 - INFO - codeparrot_training - Step 3731: {'lr': 0.000498397283016845, 'samples': 716544, 'steps': 3731, 'loss/train': 2.669545590877533} 01/27/2022 23:31:14 - INFO - codeparrot_training - Step 3732: {'lr': 0.0004983954326843386, 'samples': 716736, 'steps': 3732, 'loss/train': 3.4887882471084595} 01/27/2022 23:31:19 - INFO - codeparrot_training - Step 3733: {'lr': 0.000498393581287785, 'samples': 716928, 'steps': 3733, 'loss/train': 4.6098597049713135} 01/27/2022 23:31:23 - INFO - codeparrot_training - Step 3734: {'lr': 0.0004983917288271921, 'samples': 717120, 'steps': 3734, 'loss/train': 0.9233699440956116} 01/27/2022 23:31:27 - INFO - codeparrot_training - Step 3735: {'lr': 0.0004983898753025681, 'samples': 717312, 'steps': 3735, 'loss/train': 2.8410290479660034} 01/27/2022 23:31:32 - INFO - codeparrot_training - Step 3736: {'lr': 0.0004983880207139205, 'samples': 717504, 'steps': 3736, 'loss/train': 1.4715054631233215} 01/27/2022 23:31:36 - INFO - codeparrot_training - Step 3737: {'lr': 0.0004983861650612577, 'samples': 717696, 'steps': 3737, 'loss/train': 2.609261691570282} 01/27/2022 23:31:41 - INFO - codeparrot_training - Step 3738: {'lr': 0.0004983843083445873, 'samples': 717888, 'steps': 3738, 'loss/train': 3.9454575777053833} 01/27/2022 23:31:46 - INFO - codeparrot_training - Step 3739: {'lr': 0.0004983824505639175, 'samples': 718080, 'steps': 3739, 'loss/train': 2.1958985924720764} 01/27/2022 23:31:50 - INFO - codeparrot_training - Step 3740: {'lr': 0.000498380591719256, 'samples': 718272, 'steps': 3740, 'loss/train': 2.9440258741378784} 01/27/2022 23:31:54 - INFO - codeparrot_training - Step 3741: {'lr': 0.0004983787318106111, 'samples': 718464, 'steps': 3741, 'loss/train': 3.0827962160110474} 01/27/2022 23:31:58 - INFO - codeparrot_training - Step 3742: {'lr': 0.0004983768708379905, 'samples': 718656, 'steps': 3742, 'loss/train': 3.3721368312835693} 01/27/2022 23:32:04 - INFO - codeparrot_training - Step 3743: {'lr': 0.0004983750088014023, 'samples': 718848, 'steps': 3743, 'loss/train': 5.009172677993774} 01/27/2022 23:32:08 - INFO - codeparrot_training - Step 3744: {'lr': 0.0004983731457008544, 'samples': 719040, 'steps': 3744, 'loss/train': 4.902856349945068} 01/27/2022 23:32:12 - INFO - codeparrot_training - Step 3745: {'lr': 0.0004983712815363548, 'samples': 719232, 'steps': 3745, 'loss/train': 4.671722173690796} 01/27/2022 23:32:16 - INFO - codeparrot_training - Step 3746: {'lr': 0.0004983694163079115, 'samples': 719424, 'steps': 3746, 'loss/train': 3.136746883392334} 01/27/2022 23:32:20 - INFO - codeparrot_training - Step 3747: {'lr': 0.0004983675500155325, 'samples': 719616, 'steps': 3747, 'loss/train': 2.4554505944252014} 01/27/2022 23:32:26 - INFO - codeparrot_training - Step 3748: {'lr': 0.0004983656826592258, 'samples': 719808, 'steps': 3748, 'loss/train': 2.703548312187195} 01/27/2022 23:32:30 - INFO - codeparrot_training - Step 3749: {'lr': 0.0004983638142389993, 'samples': 720000, 'steps': 3749, 'loss/train': 3.4358800649642944} 01/27/2022 23:32:34 - INFO - codeparrot_training - Step 3750: {'lr': 0.000498361944754861, 'samples': 720192, 'steps': 3750, 'loss/train': 2.470578610897064} 01/27/2022 23:32:38 - INFO - codeparrot_training - Step 3751: {'lr': 0.0004983600742068192, 'samples': 720384, 'steps': 3751, 'loss/train': 3.96330463886261} 01/27/2022 23:32:42 - INFO - codeparrot_training - Step 3752: {'lr': 0.0004983582025948816, 'samples': 720576, 'steps': 3752, 'loss/train': 3.0152292251586914} 01/27/2022 23:32:48 - INFO - codeparrot_training - Step 3753: {'lr': 0.0004983563299190564, 'samples': 720768, 'steps': 3753, 'loss/train': 2.3018000721931458} 01/27/2022 23:32:53 - INFO - codeparrot_training - Step 3754: {'lr': 0.0004983544561793515, 'samples': 720960, 'steps': 3754, 'loss/train': 3.5101507902145386} 01/27/2022 23:32:57 - INFO - codeparrot_training - Step 3755: {'lr': 0.000498352581375775, 'samples': 721152, 'steps': 3755, 'loss/train': 2.334016799926758} 01/27/2022 23:33:01 - INFO - codeparrot_training - Step 3756: {'lr': 0.0004983507055083349, 'samples': 721344, 'steps': 3756, 'loss/train': 5.316161155700684} 01/27/2022 23:33:05 - INFO - codeparrot_training - Step 3757: {'lr': 0.0004983488285770391, 'samples': 721536, 'steps': 3757, 'loss/train': 1.3412643671035767} 01/27/2022 23:33:10 - INFO - codeparrot_training - Step 3758: {'lr': 0.000498346950581896, 'samples': 721728, 'steps': 3758, 'loss/train': 3.2607851028442383} 01/27/2022 23:33:14 - INFO - codeparrot_training - Step 3759: {'lr': 0.0004983450715229132, 'samples': 721920, 'steps': 3759, 'loss/train': 2.7041496634483337} 01/27/2022 23:33:19 - INFO - codeparrot_training - Step 3760: {'lr': 0.000498343191400099, 'samples': 722112, 'steps': 3760, 'loss/train': 3.4964364767074585} 01/27/2022 23:33:23 - INFO - codeparrot_training - Step 3761: {'lr': 0.0004983413102134616, 'samples': 722304, 'steps': 3761, 'loss/train': 2.7899272441864014} 01/27/2022 23:33:27 - INFO - codeparrot_training - Step 3762: {'lr': 0.0004983394279630088, 'samples': 722496, 'steps': 3762, 'loss/train': 3.690846562385559} 01/27/2022 23:33:32 - INFO - codeparrot_training - Step 3763: {'lr': 0.0004983375446487488, 'samples': 722688, 'steps': 3763, 'loss/train': 2.9413910508155823} 01/27/2022 23:33:36 - INFO - codeparrot_training - Step 3764: {'lr': 0.0004983356602706895, 'samples': 722880, 'steps': 3764, 'loss/train': 3.7637661695480347} 01/27/2022 23:33:40 - INFO - codeparrot_training - Step 3765: {'lr': 0.0004983337748288391, 'samples': 723072, 'steps': 3765, 'loss/train': 1.8159717321395874} 01/27/2022 23:33:45 - INFO - codeparrot_training - Step 3766: {'lr': 0.0004983318883232058, 'samples': 723264, 'steps': 3766, 'loss/train': 2.4687357544898987} 01/27/2022 23:33:49 - INFO - codeparrot_training - Step 3767: {'lr': 0.0004983300007537974, 'samples': 723456, 'steps': 3767, 'loss/train': 2.1038990020751953} 01/27/2022 23:33:55 - INFO - codeparrot_training - Step 3768: {'lr': 0.0004983281121206222, 'samples': 723648, 'steps': 3768, 'loss/train': 3.3645458221435547} 01/27/2022 23:33:59 - INFO - codeparrot_training - Step 3769: {'lr': 0.0004983262224236882, 'samples': 723840, 'steps': 3769, 'loss/train': 4.820131301879883} 01/27/2022 23:34:03 - INFO - codeparrot_training - Step 3770: {'lr': 0.0004983243316630035, 'samples': 724032, 'steps': 3770, 'loss/train': 3.9073930978775024} 01/27/2022 23:34:07 - INFO - codeparrot_training - Step 3771: {'lr': 0.0004983224398385762, 'samples': 724224, 'steps': 3771, 'loss/train': 4.202261567115784} 01/27/2022 23:34:11 - INFO - codeparrot_training - Step 3772: {'lr': 0.0004983205469504144, 'samples': 724416, 'steps': 3772, 'loss/train': 3.0777947902679443} 01/27/2022 23:34:16 - INFO - codeparrot_training - Step 3773: {'lr': 0.0004983186529985263, 'samples': 724608, 'steps': 3773, 'loss/train': 3.687490940093994} 01/27/2022 23:34:21 - INFO - codeparrot_training - Step 3774: {'lr': 0.00049831675798292, 'samples': 724800, 'steps': 3774, 'loss/train': 3.419506072998047} 01/27/2022 23:34:25 - INFO - codeparrot_training - Step 3775: {'lr': 0.0004983148619036034, 'samples': 724992, 'steps': 3775, 'loss/train': 2.7007445096969604} 01/27/2022 23:34:29 - INFO - codeparrot_training - Step 3776: {'lr': 0.0004983129647605849, 'samples': 725184, 'steps': 3776, 'loss/train': 1.907267689704895} 01/27/2022 23:34:33 - INFO - codeparrot_training - Step 3777: {'lr': 0.0004983110665538724, 'samples': 725376, 'steps': 3777, 'loss/train': 4.0927112102508545} 01/27/2022 23:34:37 - INFO - codeparrot_training - Step 3778: {'lr': 0.0004983091672834742, 'samples': 725568, 'steps': 3778, 'loss/train': 5.756445407867432} 01/27/2022 23:34:42 - INFO - codeparrot_training - Step 3779: {'lr': 0.0004983072669493985, 'samples': 725760, 'steps': 3779, 'loss/train': 2.4398158192634583} 01/27/2022 23:34:47 - INFO - codeparrot_training - Step 3780: {'lr': 0.0004983053655516531, 'samples': 725952, 'steps': 3780, 'loss/train': 3.6190552711486816} 01/27/2022 23:34:51 - INFO - codeparrot_training - Step 3781: {'lr': 0.0004983034630902465, 'samples': 726144, 'steps': 3781, 'loss/train': 2.5198001861572266} 01/27/2022 23:34:55 - INFO - codeparrot_training - Step 3782: {'lr': 0.0004983015595651867, 'samples': 726336, 'steps': 3782, 'loss/train': 3.5458381175994873} 01/27/2022 23:34:59 - INFO - codeparrot_training - Step 3783: {'lr': 0.0004982996549764817, 'samples': 726528, 'steps': 3783, 'loss/train': 3.519631862640381} 01/27/2022 23:35:04 - INFO - codeparrot_training - Step 3784: {'lr': 0.0004982977493241399, 'samples': 726720, 'steps': 3784, 'loss/train': 3.423276901245117} 01/27/2022 23:35:09 - INFO - codeparrot_training - Step 3785: {'lr': 0.0004982958426081695, 'samples': 726912, 'steps': 3785, 'loss/train': 4.333733081817627} 01/27/2022 23:35:13 - INFO - codeparrot_training - Step 3786: {'lr': 0.0004982939348285784, 'samples': 727104, 'steps': 3786, 'loss/train': 3.5695109367370605} 01/27/2022 23:35:17 - INFO - codeparrot_training - Step 3787: {'lr': 0.000498292025985375, 'samples': 727296, 'steps': 3787, 'loss/train': 2.4297419786453247} 01/27/2022 23:35:23 - INFO - codeparrot_training - Step 3788: {'lr': 0.0004982901160785675, 'samples': 727488, 'steps': 3788, 'loss/train': 3.267141580581665} 01/27/2022 23:35:27 - INFO - codeparrot_training - Step 3789: {'lr': 0.0004982882051081639, 'samples': 727680, 'steps': 3789, 'loss/train': 3.234258770942688} 01/27/2022 23:35:31 - INFO - codeparrot_training - Step 3790: {'lr': 0.0004982862930741725, 'samples': 727872, 'steps': 3790, 'loss/train': 2.4388648867607117} 01/27/2022 23:35:35 - INFO - codeparrot_training - Step 3791: {'lr': 0.0004982843799766014, 'samples': 728064, 'steps': 3791, 'loss/train': 3.3438634872436523} 01/27/2022 23:35:39 - INFO - codeparrot_training - Step 3792: {'lr': 0.0004982824658154589, 'samples': 728256, 'steps': 3792, 'loss/train': 3.619312047958374} 01/27/2022 23:35:45 - INFO - codeparrot_training - Step 3793: {'lr': 0.000498280550590753, 'samples': 728448, 'steps': 3793, 'loss/train': 3.6061320304870605} 01/27/2022 23:35:49 - INFO - codeparrot_training - Step 3794: {'lr': 0.0004982786343024923, 'samples': 728640, 'steps': 3794, 'loss/train': 3.448552966117859} 01/27/2022 23:35:53 - INFO - codeparrot_training - Step 3795: {'lr': 0.0004982767169506847, 'samples': 728832, 'steps': 3795, 'loss/train': 2.884900689125061} 01/27/2022 23:35:57 - INFO - codeparrot_training - Step 3796: {'lr': 0.0004982747985353384, 'samples': 729024, 'steps': 3796, 'loss/train': 3.319574475288391} 01/27/2022 23:36:01 - INFO - codeparrot_training - Step 3797: {'lr': 0.0004982728790564616, 'samples': 729216, 'steps': 3797, 'loss/train': 2.73227459192276} 01/27/2022 23:36:07 - INFO - codeparrot_training - Step 3798: {'lr': 0.0004982709585140629, 'samples': 729408, 'steps': 3798, 'loss/train': 3.636817216873169} 01/27/2022 23:36:11 - INFO - codeparrot_training - Step 3799: {'lr': 0.0004982690369081501, 'samples': 729600, 'steps': 3799, 'loss/train': 2.8366037607192993} 01/27/2022 23:36:15 - INFO - codeparrot_training - Step 3800: {'lr': 0.0004982671142387316, 'samples': 729792, 'steps': 3800, 'loss/train': 3.79203200340271} 01/27/2022 23:36:19 - INFO - codeparrot_training - Step 3801: {'lr': 0.0004982651905058156, 'samples': 729984, 'steps': 3801, 'loss/train': 3.02475106716156} 01/27/2022 23:36:24 - INFO - codeparrot_training - Step 3802: {'lr': 0.0004982632657094104, 'samples': 730176, 'steps': 3802, 'loss/train': 2.984947979450226} 01/27/2022 23:36:29 - INFO - codeparrot_training - Step 3803: {'lr': 0.0004982613398495241, 'samples': 730368, 'steps': 3803, 'loss/train': 3.599941849708557} 01/27/2022 23:36:33 - INFO - codeparrot_training - Step 3804: {'lr': 0.0004982594129261652, 'samples': 730560, 'steps': 3804, 'loss/train': 3.491774082183838} 01/27/2022 23:36:37 - INFO - codeparrot_training - Step 3805: {'lr': 0.0004982574849393416, 'samples': 730752, 'steps': 3805, 'loss/train': 3.564940094947815} 01/27/2022 23:36:41 - INFO - codeparrot_training - Step 3806: {'lr': 0.000498255555889062, 'samples': 730944, 'steps': 3806, 'loss/train': 3.6254889965057373} 01/27/2022 23:36:45 - INFO - codeparrot_training - Step 3807: {'lr': 0.0004982536257753343, 'samples': 731136, 'steps': 3807, 'loss/train': 2.8893486857414246} 01/27/2022 23:36:51 - INFO - codeparrot_training - Step 3808: {'lr': 0.0004982516945981669, 'samples': 731328, 'steps': 3808, 'loss/train': 3.1902272701263428} 01/27/2022 23:36:55 - INFO - codeparrot_training - Step 3809: {'lr': 0.0004982497623575681, 'samples': 731520, 'steps': 3809, 'loss/train': 1.4235122501850128} 01/27/2022 23:36:59 - INFO - codeparrot_training - Step 3810: {'lr': 0.0004982478290535461, 'samples': 731712, 'steps': 3810, 'loss/train': 2.455996513366699} 01/27/2022 23:37:03 - INFO - codeparrot_training - Step 3811: {'lr': 0.0004982458946861093, 'samples': 731904, 'steps': 3811, 'loss/train': 2.539703071117401} 01/27/2022 23:37:07 - INFO - codeparrot_training - Step 3812: {'lr': 0.0004982439592552658, 'samples': 732096, 'steps': 3812, 'loss/train': 2.083168387413025} 01/27/2022 23:37:13 - INFO - codeparrot_training - Step 3813: {'lr': 0.0004982420227610242, 'samples': 732288, 'steps': 3813, 'loss/train': 3.159732699394226} 01/27/2022 23:37:17 - INFO - codeparrot_training - Step 3814: {'lr': 0.0004982400852033924, 'samples': 732480, 'steps': 3814, 'loss/train': 2.7244847416877747} 01/27/2022 23:37:21 - INFO - codeparrot_training - Step 3815: {'lr': 0.000498238146582379, 'samples': 732672, 'steps': 3815, 'loss/train': 1.9167087078094482} 01/27/2022 23:37:25 - INFO - codeparrot_training - Step 3816: {'lr': 0.0004982362068979921, 'samples': 732864, 'steps': 3816, 'loss/train': 2.782089650630951} 01/27/2022 23:37:30 - INFO - codeparrot_training - Step 3817: {'lr': 0.0004982342661502403, 'samples': 733056, 'steps': 3817, 'loss/train': 3.5628572702407837} 01/27/2022 23:37:35 - INFO - codeparrot_training - Step 3818: {'lr': 0.0004982323243391315, 'samples': 733248, 'steps': 3818, 'loss/train': 2.2699989080429077} 01/27/2022 23:37:39 - INFO - codeparrot_training - Step 3819: {'lr': 0.0004982303814646745, 'samples': 733440, 'steps': 3819, 'loss/train': 3.5574170351028442} 01/27/2022 23:37:43 - INFO - codeparrot_training - Step 3820: {'lr': 0.0004982284375268772, 'samples': 733632, 'steps': 3820, 'loss/train': 5.13874089717865} 01/27/2022 23:37:47 - INFO - codeparrot_training - Step 3821: {'lr': 0.0004982264925257481, 'samples': 733824, 'steps': 3821, 'loss/train': 3.1835049390792847} 01/27/2022 23:37:51 - INFO - codeparrot_training - Step 3822: {'lr': 0.0004982245464612955, 'samples': 734016, 'steps': 3822, 'loss/train': 3.032190442085266} 01/27/2022 23:37:57 - INFO - codeparrot_training - Step 3823: {'lr': 0.0004982225993335279, 'samples': 734208, 'steps': 3823, 'loss/train': 2.5306017994880676} 01/27/2022 23:38:01 - INFO - codeparrot_training - Step 3824: {'lr': 0.0004982206511424534, 'samples': 734400, 'steps': 3824, 'loss/train': 4.166053533554077} 01/27/2022 23:38:05 - INFO - codeparrot_training - Step 3825: {'lr': 0.0004982187018880805, 'samples': 734592, 'steps': 3825, 'loss/train': 4.093863129615784} 01/27/2022 23:38:09 - INFO - codeparrot_training - Step 3826: {'lr': 0.0004982167515704174, 'samples': 734784, 'steps': 3826, 'loss/train': 3.340131640434265} 01/27/2022 23:38:14 - INFO - codeparrot_training - Step 3827: {'lr': 0.0004982148001894727, 'samples': 734976, 'steps': 3827, 'loss/train': 2.648875594139099} 01/27/2022 23:38:19 - INFO - codeparrot_training - Step 3828: {'lr': 0.0004982128477452546, 'samples': 735168, 'steps': 3828, 'loss/train': 2.926435947418213} 01/27/2022 23:38:23 - INFO - codeparrot_training - Step 3829: {'lr': 0.0004982108942377713, 'samples': 735360, 'steps': 3829, 'loss/train': 3.9693102836608887} 01/27/2022 23:38:27 - INFO - codeparrot_training - Step 3830: {'lr': 0.0004982089396670316, 'samples': 735552, 'steps': 3830, 'loss/train': 2.2572383880615234} 01/27/2022 23:38:31 - INFO - codeparrot_training - Step 3831: {'lr': 0.0004982069840330435, 'samples': 735744, 'steps': 3831, 'loss/train': 3.856727957725525} 01/27/2022 23:38:35 - INFO - codeparrot_training - Step 3832: {'lr': 0.0004982050273358154, 'samples': 735936, 'steps': 3832, 'loss/train': 1.9903963208198547} 01/27/2022 23:38:40 - INFO - codeparrot_training - Step 3833: {'lr': 0.0004982030695753558, 'samples': 736128, 'steps': 3833, 'loss/train': 3.3012642860412598} 01/27/2022 23:38:45 - INFO - codeparrot_training - Step 3834: {'lr': 0.0004982011107516732, 'samples': 736320, 'steps': 3834, 'loss/train': 2.8103529810905457} 01/27/2022 23:38:49 - INFO - codeparrot_training - Step 3835: {'lr': 0.0004981991508647757, 'samples': 736512, 'steps': 3835, 'loss/train': 3.4237221479415894} 01/27/2022 23:38:53 - INFO - codeparrot_training - Step 3836: {'lr': 0.0004981971899146719, 'samples': 736704, 'steps': 3836, 'loss/train': 3.1177139282226562} 01/27/2022 23:38:57 - INFO - codeparrot_training - Step 3837: {'lr': 0.0004981952279013702, 'samples': 736896, 'steps': 3837, 'loss/train': 2.73541921377182} 01/27/2022 23:39:02 - INFO - codeparrot_training - Step 3838: {'lr': 0.0004981932648248789, 'samples': 737088, 'steps': 3838, 'loss/train': 3.311891198158264} 01/27/2022 23:39:06 - INFO - codeparrot_training - Step 3839: {'lr': 0.0004981913006852065, 'samples': 737280, 'steps': 3839, 'loss/train': 3.4509050846099854} 01/27/2022 23:39:11 - INFO - codeparrot_training - Step 3840: {'lr': 0.0004981893354823614, 'samples': 737472, 'steps': 3840, 'loss/train': 4.4206026792526245} 01/27/2022 23:39:15 - INFO - codeparrot_training - Step 3841: {'lr': 0.000498187369216352, 'samples': 737664, 'steps': 3841, 'loss/train': 2.0127652287483215} 01/27/2022 23:39:19 - INFO - codeparrot_training - Step 3842: {'lr': 0.0004981854018871867, 'samples': 737856, 'steps': 3842, 'loss/train': 3.0318703651428223} 01/27/2022 23:39:25 - INFO - codeparrot_training - Step 3843: {'lr': 0.0004981834334948738, 'samples': 738048, 'steps': 3843, 'loss/train': 2.997087299823761} 01/27/2022 23:39:29 - INFO - codeparrot_training - Step 3844: {'lr': 0.0004981814640394221, 'samples': 738240, 'steps': 3844, 'loss/train': 3.8425029516220093} 01/27/2022 23:39:33 - INFO - codeparrot_training - Step 3845: {'lr': 0.0004981794935208397, 'samples': 738432, 'steps': 3845, 'loss/train': 2.73724365234375} 01/27/2022 23:39:37 - INFO - codeparrot_training - Step 3846: {'lr': 0.0004981775219391352, 'samples': 738624, 'steps': 3846, 'loss/train': 2.5818800926208496} 01/27/2022 23:39:42 - INFO - codeparrot_training - Step 3847: {'lr': 0.000498175549294317, 'samples': 738816, 'steps': 3847, 'loss/train': 1.933349847793579} 01/27/2022 23:39:47 - INFO - codeparrot_training - Step 3848: {'lr': 0.0004981735755863934, 'samples': 739008, 'steps': 3848, 'loss/train': 3.4696422815322876} 01/27/2022 23:39:51 - INFO - codeparrot_training - Step 3849: {'lr': 0.0004981716008153732, 'samples': 739200, 'steps': 3849, 'loss/train': 3.9334473609924316} 01/27/2022 23:39:55 - INFO - codeparrot_training - Step 3850: {'lr': 0.0004981696249812646, 'samples': 739392, 'steps': 3850, 'loss/train': 3.3174819946289062} 01/27/2022 23:39:59 - INFO - codeparrot_training - Step 3851: {'lr': 0.0004981676480840761, 'samples': 739584, 'steps': 3851, 'loss/train': 3.1624478101730347} 01/27/2022 23:40:03 - INFO - codeparrot_training - Step 3852: {'lr': 0.0004981656701238162, 'samples': 739776, 'steps': 3852, 'loss/train': 4.5833845138549805} 01/27/2022 23:40:08 - INFO - codeparrot_training - Step 3853: {'lr': 0.0004981636911004934, 'samples': 739968, 'steps': 3853, 'loss/train': 2.721613883972168} 01/27/2022 23:40:13 - INFO - codeparrot_training - Step 3854: {'lr': 0.0004981617110141162, 'samples': 740160, 'steps': 3854, 'loss/train': 3.8722883462905884} 01/27/2022 23:40:17 - INFO - codeparrot_training - Step 3855: {'lr': 0.000498159729864693, 'samples': 740352, 'steps': 3855, 'loss/train': 3.0679757595062256} 01/27/2022 23:40:21 - INFO - codeparrot_training - Step 3856: {'lr': 0.0004981577476522323, 'samples': 740544, 'steps': 3856, 'loss/train': 3.41143262386322} 01/27/2022 23:40:25 - INFO - codeparrot_training - Step 3857: {'lr': 0.0004981557643767426, 'samples': 740736, 'steps': 3857, 'loss/train': 2.6269968152046204} 01/27/2022 23:40:31 - INFO - codeparrot_training - Step 3858: {'lr': 0.0004981537800382323, 'samples': 740928, 'steps': 3858, 'loss/train': 3.458041191101074} 01/27/2022 23:40:35 - INFO - codeparrot_training - Step 3859: {'lr': 0.0004981517946367102, 'samples': 741120, 'steps': 3859, 'loss/train': 3.354264736175537} 01/27/2022 23:40:39 - INFO - codeparrot_training - Step 3860: {'lr': 0.0004981498081721845, 'samples': 741312, 'steps': 3860, 'loss/train': 2.794378638267517} 01/27/2022 23:40:44 - INFO - codeparrot_training - Step 3861: {'lr': 0.0004981478206446638, 'samples': 741504, 'steps': 3861, 'loss/train': 2.6673824787139893} 01/27/2022 23:40:48 - INFO - codeparrot_training - Step 3862: {'lr': 0.0004981458320541567, 'samples': 741696, 'steps': 3862, 'loss/train': 3.094091534614563} 01/27/2022 23:40:53 - INFO - codeparrot_training - Step 3863: {'lr': 0.0004981438424006716, 'samples': 741888, 'steps': 3863, 'loss/train': 2.0630820393562317} 01/27/2022 23:40:57 - INFO - codeparrot_training - Step 3864: {'lr': 0.0004981418516842171, 'samples': 742080, 'steps': 3864, 'loss/train': 3.081325650215149} 01/27/2022 23:41:01 - INFO - codeparrot_training - Step 3865: {'lr': 0.0004981398599048018, 'samples': 742272, 'steps': 3865, 'loss/train': 3.1254862546920776} 01/27/2022 23:41:05 - INFO - codeparrot_training - Step 3866: {'lr': 0.000498137867062434, 'samples': 742464, 'steps': 3866, 'loss/train': 2.961456835269928} 01/27/2022 23:41:09 - INFO - codeparrot_training - Step 3867: {'lr': 0.0004981358731571223, 'samples': 742656, 'steps': 3867, 'loss/train': 2.8805347681045532} 01/27/2022 23:41:14 - INFO - codeparrot_training - Step 3868: {'lr': 0.0004981338781888755, 'samples': 742848, 'steps': 3868, 'loss/train': 2.4571040868759155} 01/27/2022 23:41:20 - INFO - codeparrot_training - Step 3869: {'lr': 0.0004981318821577018, 'samples': 743040, 'steps': 3869, 'loss/train': 1.8547165989875793} 01/27/2022 23:41:24 - INFO - codeparrot_training - Step 3870: {'lr': 0.00049812988506361, 'samples': 743232, 'steps': 3870, 'loss/train': 3.016051769256592} 01/27/2022 23:41:28 - INFO - codeparrot_training - Step 3871: {'lr': 0.0004981278869066085, 'samples': 743424, 'steps': 3871, 'loss/train': 2.6613965034484863} 01/27/2022 23:41:32 - INFO - codeparrot_training - Step 3872: {'lr': 0.000498125887686706, 'samples': 743616, 'steps': 3872, 'loss/train': 5.598769783973694} 01/27/2022 23:41:36 - INFO - codeparrot_training - Step 3873: {'lr': 0.0004981238874039109, 'samples': 743808, 'steps': 3873, 'loss/train': 3.9969820976257324} 01/27/2022 23:41:42 - INFO - codeparrot_training - Step 3874: {'lr': 0.0004981218860582319, 'samples': 744000, 'steps': 3874, 'loss/train': 4.209392309188843} 01/27/2022 23:41:46 - INFO - codeparrot_training - Step 3875: {'lr': 0.0004981198836496775, 'samples': 744192, 'steps': 3875, 'loss/train': 2.9711920022964478} 01/27/2022 23:41:50 - INFO - codeparrot_training - Step 3876: {'lr': 0.0004981178801782563, 'samples': 744384, 'steps': 3876, 'loss/train': 2.4235569834709167} 01/27/2022 23:41:54 - INFO - codeparrot_training - Step 3877: {'lr': 0.000498115875643977, 'samples': 744576, 'steps': 3877, 'loss/train': 2.1489697694778442} 01/27/2022 23:41:59 - INFO - codeparrot_training - Step 3878: {'lr': 0.0004981138700468479, 'samples': 744768, 'steps': 3878, 'loss/train': 2.870964825153351} 01/27/2022 23:42:04 - INFO - codeparrot_training - Step 3879: {'lr': 0.0004981118633868779, 'samples': 744960, 'steps': 3879, 'loss/train': 3.4458729028701782} 01/27/2022 23:42:08 - INFO - codeparrot_training - Step 3880: {'lr': 0.0004981098556640755, 'samples': 745152, 'steps': 3880, 'loss/train': 3.8212530612945557} 01/27/2022 23:42:12 - INFO - codeparrot_training - Step 3881: {'lr': 0.0004981078468784491, 'samples': 745344, 'steps': 3881, 'loss/train': 3.0449849367141724} 01/27/2022 23:42:16 - INFO - codeparrot_training - Step 3882: {'lr': 0.0004981058370300076, 'samples': 745536, 'steps': 3882, 'loss/train': 3.12955641746521} 01/27/2022 23:42:22 - INFO - codeparrot_training - Step 3883: {'lr': 0.0004981038261187594, 'samples': 745728, 'steps': 3883, 'loss/train': 3.1870826482772827} 01/27/2022 23:42:26 - INFO - codeparrot_training - Step 3884: {'lr': 0.0004981018141447133, 'samples': 745920, 'steps': 3884, 'loss/train': 2.9959647059440613} 01/27/2022 23:42:30 - INFO - codeparrot_training - Step 3885: {'lr': 0.0004980998011078776, 'samples': 746112, 'steps': 3885, 'loss/train': 3.3449374437332153} 01/27/2022 23:42:34 - INFO - codeparrot_training - Step 3886: {'lr': 0.0004980977870082613, 'samples': 746304, 'steps': 3886, 'loss/train': 2.0210521817207336} 01/27/2022 23:42:38 - INFO - codeparrot_training - Step 3887: {'lr': 0.0004980957718458729, 'samples': 746496, 'steps': 3887, 'loss/train': 2.4611796140670776} 01/27/2022 23:42:44 - INFO - codeparrot_training - Step 3888: {'lr': 0.0004980937556207207, 'samples': 746688, 'steps': 3888, 'loss/train': 4.613066554069519} 01/27/2022 23:42:48 - INFO - codeparrot_training - Step 3889: {'lr': 0.0004980917383328139, 'samples': 746880, 'steps': 3889, 'loss/train': 3.951477527618408} 01/27/2022 23:42:52 - INFO - codeparrot_training - Step 3890: {'lr': 0.0004980897199821609, 'samples': 747072, 'steps': 3890, 'loss/train': 3.9387370347976685} 01/27/2022 23:42:56 - INFO - codeparrot_training - Step 3891: {'lr': 0.0004980877005687701, 'samples': 747264, 'steps': 3891, 'loss/train': 3.3313082456588745} 01/27/2022 23:43:00 - INFO - codeparrot_training - Step 3892: {'lr': 0.0004980856800926506, 'samples': 747456, 'steps': 3892, 'loss/train': 3.2832648754119873} 01/27/2022 23:43:05 - INFO - codeparrot_training - Step 3893: {'lr': 0.0004980836585538107, 'samples': 747648, 'steps': 3893, 'loss/train': 3.3812384605407715} 01/27/2022 23:43:10 - INFO - codeparrot_training - Step 3894: {'lr': 0.0004980816359522592, 'samples': 747840, 'steps': 3894, 'loss/train': 1.7841612696647644} 01/27/2022 23:43:14 - INFO - codeparrot_training - Step 3895: {'lr': 0.0004980796122880048, 'samples': 748032, 'steps': 3895, 'loss/train': 2.3385223746299744} 01/27/2022 23:43:18 - INFO - codeparrot_training - Step 3896: {'lr': 0.000498077587561056, 'samples': 748224, 'steps': 3896, 'loss/train': 3.6282495260238647} 01/27/2022 23:43:22 - INFO - codeparrot_training - Step 3897: {'lr': 0.0004980755617714216, 'samples': 748416, 'steps': 3897, 'loss/train': 3.431385040283203} 01/27/2022 23:43:27 - INFO - codeparrot_training - Step 3898: {'lr': 0.0004980735349191104, 'samples': 748608, 'steps': 3898, 'loss/train': 3.511367082595825} 01/27/2022 23:43:31 - INFO - codeparrot_training - Step 3899: {'lr': 0.0004980715070041308, 'samples': 748800, 'steps': 3899, 'loss/train': 2.0389772057533264} 01/27/2022 23:43:36 - INFO - codeparrot_training - Step 3900: {'lr': 0.0004980694780264917, 'samples': 748992, 'steps': 3900, 'loss/train': 2.0728708505630493} 01/27/2022 23:43:40 - INFO - codeparrot_training - Step 3901: {'lr': 0.0004980674479862018, 'samples': 749184, 'steps': 3901, 'loss/train': 2.99200701713562} 01/27/2022 23:43:44 - INFO - codeparrot_training - Step 3902: {'lr': 0.0004980654168832697, 'samples': 749376, 'steps': 3902, 'loss/train': 3.7802481651306152} 01/27/2022 23:43:50 - INFO - codeparrot_training - Step 3903: {'lr': 0.0004980633847177041, 'samples': 749568, 'steps': 3903, 'loss/train': 2.8934903740882874} 01/27/2022 23:43:54 - INFO - codeparrot_training - Step 3904: {'lr': 0.0004980613514895135, 'samples': 749760, 'steps': 3904, 'loss/train': 4.249839663505554} 01/27/2022 23:43:58 - INFO - codeparrot_training - Step 3905: {'lr': 0.0004980593171987072, 'samples': 749952, 'steps': 3905, 'loss/train': 3.544609308242798} 01/27/2022 23:44:02 - INFO - codeparrot_training - Step 3906: {'lr': 0.0004980572818452934, 'samples': 750144, 'steps': 3906, 'loss/train': 2.0802465677261353} 01/27/2022 23:44:06 - INFO - codeparrot_training - Step 3907: {'lr': 0.0004980552454292809, 'samples': 750336, 'steps': 3907, 'loss/train': 3.9612321853637695} 01/27/2022 23:44:11 - INFO - codeparrot_training - Step 3908: {'lr': 0.0004980532079506786, 'samples': 750528, 'steps': 3908, 'loss/train': 2.8526996970176697} 01/27/2022 23:44:16 - INFO - codeparrot_training - Step 3909: {'lr': 0.0004980511694094951, 'samples': 750720, 'steps': 3909, 'loss/train': 3.7419272661209106} 01/27/2022 23:44:20 - INFO - codeparrot_training - Step 3910: {'lr': 0.0004980491298057392, 'samples': 750912, 'steps': 3910, 'loss/train': 3.3182544708251953} 01/27/2022 23:44:24 - INFO - codeparrot_training - Step 3911: {'lr': 0.0004980470891394194, 'samples': 751104, 'steps': 3911, 'loss/train': 3.3095412254333496} 01/27/2022 23:44:28 - INFO - codeparrot_training - Step 3912: {'lr': 0.0004980450474105448, 'samples': 751296, 'steps': 3912, 'loss/train': 3.7462642192840576} 01/27/2022 23:44:34 - INFO - codeparrot_training - Step 3913: {'lr': 0.000498043004619124, 'samples': 751488, 'steps': 3913, 'loss/train': 3.7428481578826904} 01/27/2022 23:44:39 - INFO - codeparrot_training - Step 3914: {'lr': 0.0004980409607651656, 'samples': 751680, 'steps': 3914, 'loss/train': 3.714506506919861} 01/27/2022 23:44:43 - INFO - codeparrot_training - Step 3915: {'lr': 0.0004980389158486786, 'samples': 751872, 'steps': 3915, 'loss/train': 1.8490049242973328} 01/27/2022 23:44:47 - INFO - codeparrot_training - Step 3916: {'lr': 0.0004980368698696716, 'samples': 752064, 'steps': 3916, 'loss/train': 3.320515751838684} 01/27/2022 23:44:51 - INFO - codeparrot_training - Step 3917: {'lr': 0.0004980348228281534, 'samples': 752256, 'steps': 3917, 'loss/train': 3.225396752357483} 01/27/2022 23:44:55 - INFO - codeparrot_training - Step 3918: {'lr': 0.0004980327747241329, 'samples': 752448, 'steps': 3918, 'loss/train': 0.7629964649677277} 01/27/2022 23:45:00 - INFO - codeparrot_training - Step 3919: {'lr': 0.0004980307255576185, 'samples': 752640, 'steps': 3919, 'loss/train': 3.6018630266189575} 01/27/2022 23:45:04 - INFO - codeparrot_training - Step 3920: {'lr': 0.0004980286753286195, 'samples': 752832, 'steps': 3920, 'loss/train': 2.782050669193268} 01/27/2022 23:45:09 - INFO - codeparrot_training - Step 3921: {'lr': 0.0004980266240371443, 'samples': 753024, 'steps': 3921, 'loss/train': 2.90959632396698} 01/27/2022 23:45:13 - INFO - codeparrot_training - Step 3922: {'lr': 0.0004980245716832018, 'samples': 753216, 'steps': 3922, 'loss/train': 3.541334867477417} 01/27/2022 23:45:17 - INFO - codeparrot_training - Step 3923: {'lr': 0.0004980225182668008, 'samples': 753408, 'steps': 3923, 'loss/train': 2.880419611930847} 01/27/2022 23:45:22 - INFO - codeparrot_training - Step 3924: {'lr': 0.00049802046378795, 'samples': 753600, 'steps': 3924, 'loss/train': 2.635993480682373} 01/27/2022 23:45:26 - INFO - codeparrot_training - Step 3925: {'lr': 0.0004980184082466583, 'samples': 753792, 'steps': 3925, 'loss/train': 2.049667775630951} 01/27/2022 23:45:31 - INFO - codeparrot_training - Step 3926: {'lr': 0.0004980163516429346, 'samples': 753984, 'steps': 3926, 'loss/train': 3.0327565670013428} 01/27/2022 23:45:35 - INFO - codeparrot_training - Step 3927: {'lr': 0.0004980142939767876, 'samples': 754176, 'steps': 3927, 'loss/train': 3.156273365020752} 01/27/2022 23:45:41 - INFO - codeparrot_training - Step 3928: {'lr': 0.000498012235248226, 'samples': 754368, 'steps': 3928, 'loss/train': 0.3211170360445976} 01/27/2022 23:45:45 - INFO - codeparrot_training - Step 3929: {'lr': 0.0004980101754572589, 'samples': 754560, 'steps': 3929, 'loss/train': 2.2922088503837585} 01/27/2022 23:45:49 - INFO - codeparrot_training - Step 3930: {'lr': 0.0004980081146038948, 'samples': 754752, 'steps': 3930, 'loss/train': 3.3566497564315796} 01/27/2022 23:45:53 - INFO - codeparrot_training - Step 3931: {'lr': 0.0004980060526881429, 'samples': 754944, 'steps': 3931, 'loss/train': 3.8356329202651978} 01/27/2022 23:45:57 - INFO - codeparrot_training - Step 3932: {'lr': 0.0004980039897100115, 'samples': 755136, 'steps': 3932, 'loss/train': 4.321481466293335} 01/27/2022 23:46:02 - INFO - codeparrot_training - Step 3933: {'lr': 0.0004980019256695101, 'samples': 755328, 'steps': 3933, 'loss/train': 4.378816366195679} 01/27/2022 23:46:07 - INFO - codeparrot_training - Step 3934: {'lr': 0.000497999860566647, 'samples': 755520, 'steps': 3934, 'loss/train': 2.6298898458480835} 01/27/2022 23:46:11 - INFO - codeparrot_training - Step 3935: {'lr': 0.0004979977944014313, 'samples': 755712, 'steps': 3935, 'loss/train': 3.3304349184036255} 01/27/2022 23:46:15 - INFO - codeparrot_training - Step 3936: {'lr': 0.0004979957271738718, 'samples': 755904, 'steps': 3936, 'loss/train': 3.894196629524231} 01/27/2022 23:46:19 - INFO - codeparrot_training - Step 3937: {'lr': 0.0004979936588839773, 'samples': 756096, 'steps': 3937, 'loss/train': 3.827898144721985} 01/27/2022 23:46:23 - INFO - codeparrot_training - Step 3938: {'lr': 0.0004979915895317567, 'samples': 756288, 'steps': 3938, 'loss/train': 1.7251405119895935} 01/27/2022 23:46:29 - INFO - codeparrot_training - Step 3939: {'lr': 0.000497989519117219, 'samples': 756480, 'steps': 3939, 'loss/train': 3.6852564811706543} 01/27/2022 23:46:33 - INFO - codeparrot_training - Step 3940: {'lr': 0.0004979874476403729, 'samples': 756672, 'steps': 3940, 'loss/train': 3.858055830001831} 01/27/2022 23:46:38 - INFO - codeparrot_training - Step 3941: {'lr': 0.0004979853751012273, 'samples': 756864, 'steps': 3941, 'loss/train': 1.9583439230918884} 01/27/2022 23:46:42 - INFO - codeparrot_training - Step 3942: {'lr': 0.0004979833014997911, 'samples': 757056, 'steps': 3942, 'loss/train': 3.937979578971863} 01/27/2022 23:46:46 - INFO - codeparrot_training - Step 3943: {'lr': 0.0004979812268360731, 'samples': 757248, 'steps': 3943, 'loss/train': 3.6720385551452637} 01/27/2022 23:46:51 - INFO - codeparrot_training - Step 3944: {'lr': 0.0004979791511100823, 'samples': 757440, 'steps': 3944, 'loss/train': 3.043034791946411} 01/27/2022 23:46:55 - INFO - codeparrot_training - Step 3945: {'lr': 0.0004979770743218276, 'samples': 757632, 'steps': 3945, 'loss/train': 2.847391963005066} 01/27/2022 23:47:00 - INFO - codeparrot_training - Step 3946: {'lr': 0.0004979749964713179, 'samples': 757824, 'steps': 3946, 'loss/train': 1.654130458831787} 01/27/2022 23:47:04 - INFO - codeparrot_training - Step 3947: {'lr': 0.000497972917558562, 'samples': 758016, 'steps': 3947, 'loss/train': 3.330405592918396} 01/27/2022 23:47:09 - INFO - codeparrot_training - Step 3948: {'lr': 0.0004979708375835688, 'samples': 758208, 'steps': 3948, 'loss/train': 2.7444185614585876} 01/27/2022 23:47:13 - INFO - codeparrot_training - Step 3949: {'lr': 0.0004979687565463475, 'samples': 758400, 'steps': 3949, 'loss/train': 3.4512906074523926} 01/27/2022 23:47:17 - INFO - codeparrot_training - Step 3950: {'lr': 0.0004979666744469065, 'samples': 758592, 'steps': 3950, 'loss/train': 3.400265336036682} 01/27/2022 23:47:22 - INFO - codeparrot_training - Step 3951: {'lr': 0.0004979645912852552, 'samples': 758784, 'steps': 3951, 'loss/train': 3.5110816955566406} 01/27/2022 23:47:26 - INFO - codeparrot_training - Step 3952: {'lr': 0.0004979625070614022, 'samples': 758976, 'steps': 3952, 'loss/train': 2.8682066202163696} 01/27/2022 23:47:31 - INFO - codeparrot_training - Step 3953: {'lr': 0.0004979604217753566, 'samples': 759168, 'steps': 3953, 'loss/train': 4.4882519245147705} 01/27/2022 23:47:35 - INFO - codeparrot_training - Step 3954: {'lr': 0.0004979583354271273, 'samples': 759360, 'steps': 3954, 'loss/train': 3.0694820880889893} 01/27/2022 23:47:39 - INFO - codeparrot_training - Step 3955: {'lr': 0.0004979562480167232, 'samples': 759552, 'steps': 3955, 'loss/train': 3.283119320869446} 01/27/2022 23:47:43 - INFO - codeparrot_training - Step 3956: {'lr': 0.0004979541595441534, 'samples': 759744, 'steps': 3956, 'loss/train': 2.5367807149887085} 01/27/2022 23:47:47 - INFO - codeparrot_training - Step 3957: {'lr': 0.0004979520700094265, 'samples': 759936, 'steps': 3957, 'loss/train': 1.8068341612815857} 01/27/2022 23:47:53 - INFO - codeparrot_training - Step 3958: {'lr': 0.0004979499794125518, 'samples': 760128, 'steps': 3958, 'loss/train': 3.7047300338745117} 01/27/2022 23:47:57 - INFO - codeparrot_training - Step 3959: {'lr': 0.0004979478877535382, 'samples': 760320, 'steps': 3959, 'loss/train': 4.066543936729431} 01/27/2022 23:48:02 - INFO - codeparrot_training - Step 3960: {'lr': 0.0004979457950323945, 'samples': 760512, 'steps': 3960, 'loss/train': 3.9472339153289795} 01/27/2022 23:48:06 - INFO - codeparrot_training - Step 3961: {'lr': 0.0004979437012491297, 'samples': 760704, 'steps': 3961, 'loss/train': 1.9013789892196655} 01/27/2022 23:48:10 - INFO - codeparrot_training - Step 3962: {'lr': 0.0004979416064037528, 'samples': 760896, 'steps': 3962, 'loss/train': 3.8283451795578003} 01/27/2022 23:48:16 - INFO - codeparrot_training - Step 3963: {'lr': 0.0004979395104962728, 'samples': 761088, 'steps': 3963, 'loss/train': 3.597359776496887} 01/27/2022 23:48:20 - INFO - codeparrot_training - Step 3964: {'lr': 0.0004979374135266987, 'samples': 761280, 'steps': 3964, 'loss/train': 3.4170591831207275} 01/27/2022 23:48:24 - INFO - codeparrot_training - Step 3965: {'lr': 0.0004979353154950394, 'samples': 761472, 'steps': 3965, 'loss/train': 2.903662919998169} 01/27/2022 23:48:29 - INFO - codeparrot_training - Step 3966: {'lr': 0.0004979332164013041, 'samples': 761664, 'steps': 3966, 'loss/train': 4.603860855102539} 01/27/2022 23:48:33 - INFO - codeparrot_training - Step 3967: {'lr': 0.0004979311162455015, 'samples': 761856, 'steps': 3967, 'loss/train': 2.8625945448875427} 01/27/2022 23:48:38 - INFO - codeparrot_training - Step 3968: {'lr': 0.0004979290150276407, 'samples': 762048, 'steps': 3968, 'loss/train': 3.141753673553467} 01/27/2022 23:48:42 - INFO - codeparrot_training - Step 3969: {'lr': 0.0004979269127477308, 'samples': 762240, 'steps': 3969, 'loss/train': 3.8081159591674805} 01/27/2022 23:48:46 - INFO - codeparrot_training - Step 3970: {'lr': 0.0004979248094057806, 'samples': 762432, 'steps': 3970, 'loss/train': 3.0425169467926025} 01/27/2022 23:48:51 - INFO - codeparrot_training - Step 3971: {'lr': 0.0004979227050017994, 'samples': 762624, 'steps': 3971, 'loss/train': 3.647522449493408} 01/27/2022 23:48:55 - INFO - codeparrot_training - Step 3972: {'lr': 0.000497920599535796, 'samples': 762816, 'steps': 3972, 'loss/train': 3.1273913383483887} 01/27/2022 23:49:00 - INFO - codeparrot_training - Step 3973: {'lr': 0.0004979184930077794, 'samples': 763008, 'steps': 3973, 'loss/train': 3.325735330581665} 01/27/2022 23:49:04 - INFO - codeparrot_training - Step 3974: {'lr': 0.0004979163854177588, 'samples': 763200, 'steps': 3974, 'loss/train': 3.63100004196167} 01/27/2022 23:49:08 - INFO - codeparrot_training - Step 3975: {'lr': 0.0004979142767657432, 'samples': 763392, 'steps': 3975, 'loss/train': 1.9502763748168945} 01/27/2022 23:49:12 - INFO - codeparrot_training - Step 3976: {'lr': 0.0004979121670517413, 'samples': 763584, 'steps': 3976, 'loss/train': 3.027304172515869} 01/27/2022 23:49:16 - INFO - codeparrot_training - Step 3977: {'lr': 0.0004979100562757626, 'samples': 763776, 'steps': 3977, 'loss/train': 3.8547295331954956} 01/27/2022 23:49:22 - INFO - codeparrot_training - Step 3978: {'lr': 0.0004979079444378159, 'samples': 763968, 'steps': 3978, 'loss/train': 2.9234517216682434} 01/27/2022 23:49:26 - INFO - codeparrot_training - Step 3979: {'lr': 0.0004979058315379103, 'samples': 764160, 'steps': 3979, 'loss/train': 3.528113007545471} 01/27/2022 23:49:31 - INFO - codeparrot_training - Step 3980: {'lr': 0.0004979037175760548, 'samples': 764352, 'steps': 3980, 'loss/train': 2.8164485692977905} 01/27/2022 23:49:35 - INFO - codeparrot_training - Step 3981: {'lr': 0.0004979016025522586, 'samples': 764544, 'steps': 3981, 'loss/train': 3.3351656198501587} 01/27/2022 23:49:39 - INFO - codeparrot_training - Step 3982: {'lr': 0.0004978994864665305, 'samples': 764736, 'steps': 3982, 'loss/train': 3.0262527465820312} 01/27/2022 23:49:44 - INFO - codeparrot_training - Step 3983: {'lr': 0.0004978973693188797, 'samples': 764928, 'steps': 3983, 'loss/train': 4.251277685165405} 01/27/2022 23:49:48 - INFO - codeparrot_training - Step 3984: {'lr': 0.0004978952511093155, 'samples': 765120, 'steps': 3984, 'loss/train': 2.4917977452278137} 01/27/2022 23:49:52 - INFO - codeparrot_training - Step 3985: {'lr': 0.0004978931318378465, 'samples': 765312, 'steps': 3985, 'loss/train': 2.0357927083969116} 01/27/2022 23:49:56 - INFO - codeparrot_training - Step 3986: {'lr': 0.0004978910115044822, 'samples': 765504, 'steps': 3986, 'loss/train': 3.755191683769226} 01/27/2022 23:50:01 - INFO - codeparrot_training - Step 3987: {'lr': 0.0004978888901092315, 'samples': 765696, 'steps': 3987, 'loss/train': 2.9696128964424133} 01/27/2022 23:50:06 - INFO - codeparrot_training - Step 3988: {'lr': 0.0004978867676521035, 'samples': 765888, 'steps': 3988, 'loss/train': 2.4763541221618652} 01/27/2022 23:50:11 - INFO - codeparrot_training - Step 3989: {'lr': 0.0004978846441331073, 'samples': 766080, 'steps': 3989, 'loss/train': 3.238136887550354} 01/27/2022 23:50:15 - INFO - codeparrot_training - Step 3990: {'lr': 0.000497882519552252, 'samples': 766272, 'steps': 3990, 'loss/train': 3.957271099090576} 01/27/2022 23:50:19 - INFO - codeparrot_training - Step 3991: {'lr': 0.0004978803939095466, 'samples': 766464, 'steps': 3991, 'loss/train': 2.806953012943268} 01/27/2022 23:50:23 - INFO - codeparrot_training - Step 3992: {'lr': 0.0004978782672050004, 'samples': 766656, 'steps': 3992, 'loss/train': 5.058438777923584} 01/27/2022 23:50:28 - INFO - codeparrot_training - Step 3993: {'lr': 0.0004978761394386224, 'samples': 766848, 'steps': 3993, 'loss/train': 4.709057807922363} 01/27/2022 23:50:32 - INFO - codeparrot_training - Step 3994: {'lr': 0.0004978740106104218, 'samples': 767040, 'steps': 3994, 'loss/train': 2.768343508243561} 01/27/2022 23:50:36 - INFO - codeparrot_training - Step 3995: {'lr': 0.0004978718807204076, 'samples': 767232, 'steps': 3995, 'loss/train': 2.804336965084076} 01/27/2022 23:50:41 - INFO - codeparrot_training - Step 3996: {'lr': 0.0004978697497685889, 'samples': 767424, 'steps': 3996, 'loss/train': 3.957491397857666} 01/27/2022 23:50:45 - INFO - codeparrot_training - Step 3997: {'lr': 0.0004978676177549749, 'samples': 767616, 'steps': 3997, 'loss/train': 3.629952549934387} 01/27/2022 23:50:50 - INFO - codeparrot_training - Step 3998: {'lr': 0.0004978654846795748, 'samples': 767808, 'steps': 3998, 'loss/train': 2.8321902751922607} 01/27/2022 23:50:54 - INFO - codeparrot_training - Step 3999: {'lr': 0.0004978633505423976, 'samples': 768000, 'steps': 3999, 'loss/train': 1.9902616739273071} 01/27/2022 23:50:54 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/27/2022 23:51:26 - WARNING - huggingface_hub.repository - Several commits (2) will be pushed upstream. 01/27/2022 23:51:26 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/27/2022 23:52:29 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 9481975..7415fde colorful-plasma-1 -> colorful-plasma-1 01/27/2022 23:52:35 - INFO - codeparrot_training - Step 4000: {'lr': 0.0004978612153434526, 'samples': 768192, 'steps': 4000, 'loss/train': 3.4668617248535156} 01/27/2022 23:52:39 - INFO - codeparrot_training - Step 4001: {'lr': 0.0004978590790827488, 'samples': 768384, 'steps': 4001, 'loss/train': 3.0691784620285034} 01/27/2022 23:52:44 - INFO - codeparrot_training - Step 4002: {'lr': 0.0004978569417602955, 'samples': 768576, 'steps': 4002, 'loss/train': 2.3476787209510803} 01/27/2022 23:52:50 - INFO - codeparrot_training - Step 4003: {'lr': 0.0004978548033761017, 'samples': 768768, 'steps': 4003, 'loss/train': 3.23464572429657} 01/27/2022 23:52:54 - INFO - codeparrot_training - Step 4004: {'lr': 0.0004978526639301766, 'samples': 768960, 'steps': 4004, 'loss/train': 3.557999610900879} 01/27/2022 23:52:58 - INFO - codeparrot_training - Step 4005: {'lr': 0.0004978505234225294, 'samples': 769152, 'steps': 4005, 'loss/train': 4.075190305709839} 01/27/2022 23:53:02 - INFO - codeparrot_training - Step 4006: {'lr': 0.0004978483818531693, 'samples': 769344, 'steps': 4006, 'loss/train': 3.6161255836486816} 01/27/2022 23:53:07 - INFO - codeparrot_training - Step 4007: {'lr': 0.0004978462392221054, 'samples': 769536, 'steps': 4007, 'loss/train': 4.30660343170166} 01/27/2022 23:53:12 - INFO - codeparrot_training - Step 4008: {'lr': 0.0004978440955293468, 'samples': 769728, 'steps': 4008, 'loss/train': 2.9322107434272766} 01/27/2022 23:53:16 - INFO - codeparrot_training - Step 4009: {'lr': 0.000497841950774903, 'samples': 769920, 'steps': 4009, 'loss/train': 3.6164917945861816} 01/27/2022 23:53:20 - INFO - codeparrot_training - Step 4010: {'lr': 0.0004978398049587828, 'samples': 770112, 'steps': 4010, 'loss/train': 3.0036925077438354} 01/27/2022 23:53:25 - INFO - codeparrot_training - Step 4011: {'lr': 0.0004978376580809957, 'samples': 770304, 'steps': 4011, 'loss/train': 4.209198832511902} 01/27/2022 23:53:29 - INFO - codeparrot_training - Step 4012: {'lr': 0.0004978355101415507, 'samples': 770496, 'steps': 4012, 'loss/train': 3.730727791786194} 01/27/2022 23:53:34 - INFO - codeparrot_training - Step 4013: {'lr': 0.0004978333611404571, 'samples': 770688, 'steps': 4013, 'loss/train': 2.533027946949005} 01/27/2022 23:53:38 - INFO - codeparrot_training - Step 4014: {'lr': 0.0004978312110777241, 'samples': 770880, 'steps': 4014, 'loss/train': 2.0896393060684204} 01/27/2022 23:53:42 - INFO - codeparrot_training - Step 4015: {'lr': 0.0004978290599533609, 'samples': 771072, 'steps': 4015, 'loss/train': 3.67752742767334} 01/27/2022 23:53:47 - INFO - codeparrot_training - Step 4016: {'lr': 0.0004978269077673766, 'samples': 771264, 'steps': 4016, 'loss/train': 1.1932446956634521} 01/27/2022 23:53:51 - INFO - codeparrot_training - Step 4017: {'lr': 0.0004978247545197806, 'samples': 771456, 'steps': 4017, 'loss/train': 3.309472918510437} 01/27/2022 23:53:56 - INFO - codeparrot_training - Step 4018: {'lr': 0.0004978226002105821, 'samples': 771648, 'steps': 4018, 'loss/train': 4.304401874542236} 01/27/2022 23:54:00 - INFO - codeparrot_training - Step 4019: {'lr': 0.0004978204448397902, 'samples': 771840, 'steps': 4019, 'loss/train': 2.0877009630203247} 01/27/2022 23:54:05 - INFO - codeparrot_training - Step 4020: {'lr': 0.0004978182884074142, 'samples': 772032, 'steps': 4020, 'loss/train': 4.092007398605347} 01/27/2022 23:54:09 - INFO - codeparrot_training - Step 4021: {'lr': 0.0004978161309134633, 'samples': 772224, 'steps': 4021, 'loss/train': 3.4227322340011597} 01/27/2022 23:54:13 - INFO - codeparrot_training - Step 4022: {'lr': 0.0004978139723579469, 'samples': 772416, 'steps': 4022, 'loss/train': 3.2083128690719604} 01/27/2022 23:54:19 - INFO - codeparrot_training - Step 4023: {'lr': 0.0004978118127408741, 'samples': 772608, 'steps': 4023, 'loss/train': 0.6562983244657516} 01/27/2022 23:54:24 - INFO - codeparrot_training - Step 4024: {'lr': 0.0004978096520622541, 'samples': 772800, 'steps': 4024, 'loss/train': 3.605013370513916} 01/27/2022 23:54:28 - INFO - codeparrot_training - Step 4025: {'lr': 0.0004978074903220964, 'samples': 772992, 'steps': 4025, 'loss/train': 3.402789831161499} 01/27/2022 23:54:32 - INFO - codeparrot_training - Step 4026: {'lr': 0.0004978053275204099, 'samples': 773184, 'steps': 4026, 'loss/train': 2.612524151802063} 01/27/2022 23:54:36 - INFO - codeparrot_training - Step 4027: {'lr': 0.0004978031636572042, 'samples': 773376, 'steps': 4027, 'loss/train': 3.792102098464966} 01/27/2022 23:54:42 - INFO - codeparrot_training - Step 4028: {'lr': 0.0004978009987324884, 'samples': 773568, 'steps': 4028, 'loss/train': 3.9628604650497437} 01/27/2022 23:54:46 - INFO - codeparrot_training - Step 4029: {'lr': 0.0004977988327462718, 'samples': 773760, 'steps': 4029, 'loss/train': 3.7592010498046875} 01/27/2022 23:54:50 - INFO - codeparrot_training - Step 4030: {'lr': 0.0004977966656985637, 'samples': 773952, 'steps': 4030, 'loss/train': 2.7272586822509766} 01/27/2022 23:54:54 - INFO - codeparrot_training - Step 4031: {'lr': 0.0004977944975893733, 'samples': 774144, 'steps': 4031, 'loss/train': 3.2865082025527954} 01/27/2022 23:54:59 - INFO - codeparrot_training - Step 4032: {'lr': 0.00049779232841871, 'samples': 774336, 'steps': 4032, 'loss/train': 3.4531302452087402} 01/27/2022 23:55:05 - INFO - codeparrot_training - Step 4033: {'lr': 0.0004977901581865831, 'samples': 774528, 'steps': 4033, 'loss/train': 4.624125838279724} 01/27/2022 23:55:09 - INFO - codeparrot_training - Step 4034: {'lr': 0.0004977879868930018, 'samples': 774720, 'steps': 4034, 'loss/train': 3.0678720474243164} 01/27/2022 23:55:13 - INFO - codeparrot_training - Step 4035: {'lr': 0.0004977858145379754, 'samples': 774912, 'steps': 4035, 'loss/train': 4.105309009552002} 01/27/2022 23:55:17 - INFO - codeparrot_training - Step 4036: {'lr': 0.0004977836411215133, 'samples': 775104, 'steps': 4036, 'loss/train': 2.837820589542389} 01/27/2022 23:55:22 - INFO - codeparrot_training - Step 4037: {'lr': 0.0004977814666436248, 'samples': 775296, 'steps': 4037, 'loss/train': 3.738779067993164} 01/27/2022 23:55:27 - INFO - codeparrot_training - Step 4038: {'lr': 0.0004977792911043191, 'samples': 775488, 'steps': 4038, 'loss/train': 3.064248561859131} 01/27/2022 23:55:31 - INFO - codeparrot_training - Step 4039: {'lr': 0.0004977771145036056, 'samples': 775680, 'steps': 4039, 'loss/train': 3.0197685956954956} 01/27/2022 23:55:35 - INFO - codeparrot_training - Step 4040: {'lr': 0.0004977749368414937, 'samples': 775872, 'steps': 4040, 'loss/train': 2.215318500995636} 01/27/2022 23:55:40 - INFO - codeparrot_training - Step 4041: {'lr': 0.0004977727581179926, 'samples': 776064, 'steps': 4041, 'loss/train': 2.119442403316498} 01/27/2022 23:55:44 - INFO - codeparrot_training - Step 4042: {'lr': 0.0004977705783331117, 'samples': 776256, 'steps': 4042, 'loss/train': 2.3120981454849243} 01/27/2022 23:55:49 - INFO - codeparrot_training - Step 4043: {'lr': 0.0004977683974868603, 'samples': 776448, 'steps': 4043, 'loss/train': 4.357568621635437} 01/27/2022 23:55:53 - INFO - codeparrot_training - Step 4044: {'lr': 0.0004977662155792478, 'samples': 776640, 'steps': 4044, 'loss/train': 4.1784210205078125} 01/27/2022 23:55:58 - INFO - codeparrot_training - Step 4045: {'lr': 0.0004977640326102834, 'samples': 776832, 'steps': 4045, 'loss/train': 2.8483824133872986} 01/27/2022 23:56:02 - INFO - codeparrot_training - Step 4046: {'lr': 0.0004977618485799767, 'samples': 777024, 'steps': 4046, 'loss/train': 10.301694631576538} 01/27/2022 23:56:06 - INFO - codeparrot_training - Step 4047: {'lr': 0.0004977596634883368, 'samples': 777216, 'steps': 4047, 'loss/train': 4.961359977722168} 01/27/2022 23:56:10 - INFO - codeparrot_training - Step 4048: {'lr': 0.0004977574773353732, 'samples': 777408, 'steps': 4048, 'loss/train': 2.74309641122818} 01/27/2022 23:56:16 - INFO - codeparrot_training - Step 4049: {'lr': 0.0004977552901210952, 'samples': 777600, 'steps': 4049, 'loss/train': 2.365483045578003} 01/27/2022 23:56:20 - INFO - codeparrot_training - Step 4050: {'lr': 0.0004977531018455124, 'samples': 777792, 'steps': 4050, 'loss/train': 3.5187467336654663} 01/27/2022 23:56:25 - INFO - codeparrot_training - Step 4051: {'lr': 0.0004977509125086338, 'samples': 777984, 'steps': 4051, 'loss/train': 3.2581897974014282} 01/27/2022 23:56:29 - INFO - codeparrot_training - Step 4052: {'lr': 0.000497748722110469, 'samples': 778176, 'steps': 4052, 'loss/train': 3.573821783065796} 01/27/2022 23:56:33 - INFO - codeparrot_training - Step 4053: {'lr': 0.0004977465306510273, 'samples': 778368, 'steps': 4053, 'loss/train': 2.5381826162338257} 01/27/2022 23:56:38 - INFO - codeparrot_training - Step 4054: {'lr': 0.0004977443381303182, 'samples': 778560, 'steps': 4054, 'loss/train': 2.5717576146125793} 01/27/2022 23:56:42 - INFO - codeparrot_training - Step 4055: {'lr': 0.000497742144548351, 'samples': 778752, 'steps': 4055, 'loss/train': 4.091739177703857} 01/27/2022 23:56:47 - INFO - codeparrot_training - Step 4056: {'lr': 0.0004977399499051351, 'samples': 778944, 'steps': 4056, 'loss/train': 3.130257725715637} 01/27/2022 23:56:51 - INFO - codeparrot_training - Step 4057: {'lr': 0.0004977377542006799, 'samples': 779136, 'steps': 4057, 'loss/train': 2.744827151298523} 01/27/2022 23:56:55 - INFO - codeparrot_training - Step 4058: {'lr': 0.0004977355574349949, 'samples': 779328, 'steps': 4058, 'loss/train': 3.595530867576599} 01/27/2022 23:57:00 - INFO - codeparrot_training - Step 4059: {'lr': 0.0004977333596080894, 'samples': 779520, 'steps': 4059, 'loss/train': 2.8797159790992737} 01/27/2022 23:57:04 - INFO - codeparrot_training - Step 4060: {'lr': 0.0004977311607199729, 'samples': 779712, 'steps': 4060, 'loss/train': 2.386080801486969} 01/27/2022 23:57:09 - INFO - codeparrot_training - Step 4061: {'lr': 0.0004977289607706547, 'samples': 779904, 'steps': 4061, 'loss/train': 3.2648141384124756} 01/27/2022 23:57:13 - INFO - codeparrot_training - Step 4062: {'lr': 0.0004977267597601443, 'samples': 780096, 'steps': 4062, 'loss/train': 2.91173779964447} 01/27/2022 23:57:17 - INFO - codeparrot_training - Step 4063: {'lr': 0.0004977245576884511, 'samples': 780288, 'steps': 4063, 'loss/train': 3.097644567489624} 01/27/2022 23:57:23 - INFO - codeparrot_training - Step 4064: {'lr': 0.0004977223545555847, 'samples': 780480, 'steps': 4064, 'loss/train': 3.017891764640808} 01/27/2022 23:57:27 - INFO - codeparrot_training - Step 4065: {'lr': 0.0004977201503615543, 'samples': 780672, 'steps': 4065, 'loss/train': 2.5534043312072754} 01/27/2022 23:57:32 - INFO - codeparrot_training - Step 4066: {'lr': 0.0004977179451063694, 'samples': 780864, 'steps': 4066, 'loss/train': 3.5307862758636475} 01/27/2022 23:57:36 - INFO - codeparrot_training - Step 4067: {'lr': 0.0004977157387900395, 'samples': 781056, 'steps': 4067, 'loss/train': 3.5122050046920776} 01/27/2022 23:57:41 - INFO - codeparrot_training - Step 4068: {'lr': 0.0004977135314125741, 'samples': 781248, 'steps': 4068, 'loss/train': 3.0230480432510376} 01/27/2022 23:57:45 - INFO - codeparrot_training - Step 4069: {'lr': 0.0004977113229739825, 'samples': 781440, 'steps': 4069, 'loss/train': 2.7292327880859375} 01/27/2022 23:57:50 - INFO - codeparrot_training - Step 4070: {'lr': 0.0004977091134742743, 'samples': 781632, 'steps': 4070, 'loss/train': 3.6481833457946777} 01/27/2022 23:57:54 - INFO - codeparrot_training - Step 4071: {'lr': 0.0004977069029134588, 'samples': 781824, 'steps': 4071, 'loss/train': 7.32629656791687} 01/27/2022 23:57:58 - INFO - codeparrot_training - Step 4072: {'lr': 0.0004977046912915458, 'samples': 782016, 'steps': 4072, 'loss/train': 3.6295191049575806} 01/27/2022 23:58:02 - INFO - codeparrot_training - Step 4073: {'lr': 0.0004977024786085444, 'samples': 782208, 'steps': 4073, 'loss/train': 4.729617476463318} 01/27/2022 23:58:08 - INFO - codeparrot_training - Step 4074: {'lr': 0.0004977002648644642, 'samples': 782400, 'steps': 4074, 'loss/train': 3.9214389324188232} 01/27/2022 23:58:13 - INFO - codeparrot_training - Step 4075: {'lr': 0.0004976980500593149, 'samples': 782592, 'steps': 4075, 'loss/train': 2.2802615761756897} 01/27/2022 23:58:17 - INFO - codeparrot_training - Step 4076: {'lr': 0.0004976958341931057, 'samples': 782784, 'steps': 4076, 'loss/train': 2.785670220851898} 01/27/2022 23:58:21 - INFO - codeparrot_training - Step 4077: {'lr': 0.0004976936172658462, 'samples': 782976, 'steps': 4077, 'loss/train': 2.7877222895622253} 01/27/2022 23:58:25 - INFO - codeparrot_training - Step 4078: {'lr': 0.0004976913992775459, 'samples': 783168, 'steps': 4078, 'loss/train': 2.0708239674568176} 01/27/2022 23:58:31 - INFO - codeparrot_training - Step 4079: {'lr': 0.0004976891802282143, 'samples': 783360, 'steps': 4079, 'loss/train': 3.2224870920181274} 01/27/2022 23:58:35 - INFO - codeparrot_training - Step 4080: {'lr': 0.0004976869601178609, 'samples': 783552, 'steps': 4080, 'loss/train': 3.1929874420166016} 01/27/2022 23:58:39 - INFO - codeparrot_training - Step 4081: {'lr': 0.0004976847389464952, 'samples': 783744, 'steps': 4081, 'loss/train': 2.7065736651420593} 01/27/2022 23:58:43 - INFO - codeparrot_training - Step 4082: {'lr': 0.0004976825167141268, 'samples': 783936, 'steps': 4082, 'loss/train': 3.579603910446167} 01/27/2022 23:58:47 - INFO - codeparrot_training - Step 4083: {'lr': 0.000497680293420765, 'samples': 784128, 'steps': 4083, 'loss/train': 1.7678022980690002} 01/27/2022 23:58:53 - INFO - codeparrot_training - Step 4084: {'lr': 0.0004976780690664196, 'samples': 784320, 'steps': 4084, 'loss/train': 3.6203194856643677} 01/27/2022 23:58:57 - INFO - codeparrot_training - Step 4085: {'lr': 0.0004976758436511, 'samples': 784512, 'steps': 4085, 'loss/train': 2.69931560754776} 01/27/2022 23:59:02 - INFO - codeparrot_training - Step 4086: {'lr': 0.0004976736171748156, 'samples': 784704, 'steps': 4086, 'loss/train': 2.4579055309295654} 01/27/2022 23:59:06 - INFO - codeparrot_training - Step 4087: {'lr': 0.0004976713896375762, 'samples': 784896, 'steps': 4087, 'loss/train': 3.3909534215927124} 01/27/2022 23:59:10 - INFO - codeparrot_training - Step 4088: {'lr': 0.0004976691610393911, 'samples': 785088, 'steps': 4088, 'loss/train': 2.8396283984184265} 01/27/2022 23:59:16 - INFO - codeparrot_training - Step 4089: {'lr': 0.0004976669313802701, 'samples': 785280, 'steps': 4089, 'loss/train': 3.300305485725403} 01/27/2022 23:59:20 - INFO - codeparrot_training - Step 4090: {'lr': 0.0004976647006602225, 'samples': 785472, 'steps': 4090, 'loss/train': 2.079449415206909} 01/27/2022 23:59:24 - INFO - codeparrot_training - Step 4091: {'lr': 0.0004976624688792581, 'samples': 785664, 'steps': 4091, 'loss/train': 2.602890729904175} 01/27/2022 23:59:28 - INFO - codeparrot_training - Step 4092: {'lr': 0.0004976602360373861, 'samples': 785856, 'steps': 4092, 'loss/train': 1.0574281811714172} 01/27/2022 23:59:33 - INFO - codeparrot_training - Step 4093: {'lr': 0.0004976580021346164, 'samples': 786048, 'steps': 4093, 'loss/train': 2.5764697194099426} 01/27/2022 23:59:39 - INFO - codeparrot_training - Step 4094: {'lr': 0.0004976557671709585, 'samples': 786240, 'steps': 4094, 'loss/train': 4.378450870513916} 01/27/2022 23:59:43 - INFO - codeparrot_training - Step 4095: {'lr': 0.0004976535311464219, 'samples': 786432, 'steps': 4095, 'loss/train': 3.497877359390259} 01/27/2022 23:59:47 - INFO - codeparrot_training - Step 4096: {'lr': 0.0004976512940610162, 'samples': 786624, 'steps': 4096, 'loss/train': 3.553464889526367} 01/27/2022 23:59:52 - INFO - codeparrot_training - Step 4097: {'lr': 0.0004976490559147511, 'samples': 786816, 'steps': 4097, 'loss/train': 2.090789794921875} 01/27/2022 23:59:56 - INFO - codeparrot_training - Step 4098: {'lr': 0.0004976468167076359, 'samples': 787008, 'steps': 4098, 'loss/train': 3.3767298460006714} 01/28/2022 00:00:01 - INFO - codeparrot_training - Step 4099: {'lr': 0.0004976445764396805, 'samples': 787200, 'steps': 4099, 'loss/train': 3.7743040323257446} 01/28/2022 00:00:05 - INFO - codeparrot_training - Step 4100: {'lr': 0.0004976423351108943, 'samples': 787392, 'steps': 4100, 'loss/train': 2.503924012184143} 01/28/2022 00:00:09 - INFO - codeparrot_training - Step 4101: {'lr': 0.0004976400927212871, 'samples': 787584, 'steps': 4101, 'loss/train': 2.8409632444381714} 01/28/2022 00:00:14 - INFO - codeparrot_training - Step 4102: {'lr': 0.0004976378492708681, 'samples': 787776, 'steps': 4102, 'loss/train': 3.4084185361862183} 01/28/2022 00:00:18 - INFO - codeparrot_training - Step 4103: {'lr': 0.0004976356047596475, 'samples': 787968, 'steps': 4103, 'loss/train': 3.6969401836395264} 01/28/2022 00:00:24 - INFO - codeparrot_training - Step 4104: {'lr': 0.0004976333591876344, 'samples': 788160, 'steps': 4104, 'loss/train': 1.6890592575073242} 01/28/2022 00:00:28 - INFO - codeparrot_training - Step 4105: {'lr': 0.0004976311125548387, 'samples': 788352, 'steps': 4105, 'loss/train': 3.1422826051712036} 01/28/2022 00:00:32 - INFO - codeparrot_training - Step 4106: {'lr': 0.00049762886486127, 'samples': 788544, 'steps': 4106, 'loss/train': 2.6949471831321716} 01/28/2022 00:00:36 - INFO - codeparrot_training - Step 4107: {'lr': 0.0004976266161069379, 'samples': 788736, 'steps': 4107, 'loss/train': 2.9497044682502747} 01/28/2022 00:00:41 - INFO - codeparrot_training - Step 4108: {'lr': 0.0004976243662918518, 'samples': 788928, 'steps': 4108, 'loss/train': 2.153401494026184} 01/28/2022 00:00:46 - INFO - codeparrot_training - Step 4109: {'lr': 0.0004976221154160217, 'samples': 789120, 'steps': 4109, 'loss/train': 1.6923261880874634} 01/28/2022 00:00:51 - INFO - codeparrot_training - Step 4110: {'lr': 0.0004976198634794571, 'samples': 789312, 'steps': 4110, 'loss/train': 2.799542784690857} 01/28/2022 00:00:55 - INFO - codeparrot_training - Step 4111: {'lr': 0.0004976176104821675, 'samples': 789504, 'steps': 4111, 'loss/train': 2.9907420873641968} 01/28/2022 00:00:59 - INFO - codeparrot_training - Step 4112: {'lr': 0.0004976153564241628, 'samples': 789696, 'steps': 4112, 'loss/train': 1.95523202419281} 01/28/2022 00:01:03 - INFO - codeparrot_training - Step 4113: {'lr': 0.0004976131013054526, 'samples': 789888, 'steps': 4113, 'loss/train': 3.627271056175232} 01/28/2022 00:01:09 - INFO - codeparrot_training - Step 4114: {'lr': 0.0004976108451260464, 'samples': 790080, 'steps': 4114, 'loss/train': 3.489960193634033} 01/28/2022 00:01:13 - INFO - codeparrot_training - Step 4115: {'lr': 0.000497608587885954, 'samples': 790272, 'steps': 4115, 'loss/train': 3.8638222217559814} 01/28/2022 00:01:17 - INFO - codeparrot_training - Step 4116: {'lr': 0.0004976063295851849, 'samples': 790464, 'steps': 4116, 'loss/train': 2.8666403889656067} 01/28/2022 00:01:22 - INFO - codeparrot_training - Step 4117: {'lr': 0.000497604070223749, 'samples': 790656, 'steps': 4117, 'loss/train': 3.121669292449951} 01/28/2022 00:01:26 - INFO - codeparrot_training - Step 4118: {'lr': 0.0004976018098016559, 'samples': 790848, 'steps': 4118, 'loss/train': 3.0682464838027954} 01/28/2022 00:01:32 - INFO - codeparrot_training - Step 4119: {'lr': 0.0004975995483189153, 'samples': 791040, 'steps': 4119, 'loss/train': 3.4053146839141846} 01/28/2022 00:01:36 - INFO - codeparrot_training - Step 4120: {'lr': 0.0004975972857755368, 'samples': 791232, 'steps': 4120, 'loss/train': 1.3697249293327332} 01/28/2022 00:01:40 - INFO - codeparrot_training - Step 4121: {'lr': 0.0004975950221715302, 'samples': 791424, 'steps': 4121, 'loss/train': 2.3150092363357544} 01/28/2022 00:01:44 - INFO - codeparrot_training - Step 4122: {'lr': 0.0004975927575069051, 'samples': 791616, 'steps': 4122, 'loss/train': 10.979862213134766} 01/28/2022 00:01:49 - INFO - codeparrot_training - Step 4123: {'lr': 0.0004975904917816713, 'samples': 791808, 'steps': 4123, 'loss/train': 3.231865167617798} 01/28/2022 00:01:54 - INFO - codeparrot_training - Step 4124: {'lr': 0.0004975882249958385, 'samples': 792000, 'steps': 4124, 'loss/train': 3.812573790550232} 01/28/2022 00:01:58 - INFO - codeparrot_training - Step 4125: {'lr': 0.0004975859571494162, 'samples': 792192, 'steps': 4125, 'loss/train': 2.652881920337677} 01/28/2022 00:02:02 - INFO - codeparrot_training - Step 4126: {'lr': 0.0004975836882424143, 'samples': 792384, 'steps': 4126, 'loss/train': 2.6136658787727356} 01/28/2022 00:02:07 - INFO - codeparrot_training - Step 4127: {'lr': 0.0004975814182748426, 'samples': 792576, 'steps': 4127, 'loss/train': 2.8187849521636963} 01/28/2022 00:02:11 - INFO - codeparrot_training - Step 4128: {'lr': 0.0004975791472467108, 'samples': 792768, 'steps': 4128, 'loss/train': 1.0012774765491486} 01/28/2022 00:02:16 - INFO - codeparrot_training - Step 4129: {'lr': 0.0004975768751580283, 'samples': 792960, 'steps': 4129, 'loss/train': 3.37856125831604} 01/28/2022 00:02:20 - INFO - codeparrot_training - Step 4130: {'lr': 0.0004975746020088052, 'samples': 793152, 'steps': 4130, 'loss/train': 3.470565676689148} 01/28/2022 00:02:24 - INFO - codeparrot_training - Step 4131: {'lr': 0.0004975723277990512, 'samples': 793344, 'steps': 4131, 'loss/train': 2.658418357372284} 01/28/2022 00:02:29 - INFO - codeparrot_training - Step 4132: {'lr': 0.0004975700525287758, 'samples': 793536, 'steps': 4132, 'loss/train': 3.7125080823898315} 01/28/2022 00:02:33 - INFO - codeparrot_training - Step 4133: {'lr': 0.0004975677761979891, 'samples': 793728, 'steps': 4133, 'loss/train': 3.996389865875244} 01/28/2022 00:02:39 - INFO - codeparrot_training - Step 4134: {'lr': 0.0004975654988067005, 'samples': 793920, 'steps': 4134, 'loss/train': 3.833632707595825} 01/28/2022 00:02:43 - INFO - codeparrot_training - Step 4135: {'lr': 0.00049756322035492, 'samples': 794112, 'steps': 4135, 'loss/train': 2.7541210055351257} 01/28/2022 00:02:47 - INFO - codeparrot_training - Step 4136: {'lr': 0.0004975609408426572, 'samples': 794304, 'steps': 4136, 'loss/train': 2.1259939670562744} 01/28/2022 00:02:52 - INFO - codeparrot_training - Step 4137: {'lr': 0.000497558660269922, 'samples': 794496, 'steps': 4137, 'loss/train': 3.424100160598755} 01/28/2022 00:02:56 - INFO - codeparrot_training - Step 4138: {'lr': 0.0004975563786367241, 'samples': 794688, 'steps': 4138, 'loss/train': 3.6539483070373535} 01/28/2022 00:03:01 - INFO - codeparrot_training - Step 4139: {'lr': 0.0004975540959430732, 'samples': 794880, 'steps': 4139, 'loss/train': 2.306133985519409} 01/28/2022 00:03:05 - INFO - codeparrot_training - Step 4140: {'lr': 0.0004975518121889793, 'samples': 795072, 'steps': 4140, 'loss/train': 3.211864471435547} 01/28/2022 00:03:09 - INFO - codeparrot_training - Step 4141: {'lr': 0.000497549527374452, 'samples': 795264, 'steps': 4141, 'loss/train': 2.056727349758148} 01/28/2022 00:03:14 - INFO - codeparrot_training - Step 4142: {'lr': 0.000497547241499501, 'samples': 795456, 'steps': 4142, 'loss/train': 3.3434911966323853} 01/28/2022 00:03:18 - INFO - codeparrot_training - Step 4143: {'lr': 0.0004975449545641364, 'samples': 795648, 'steps': 4143, 'loss/train': 2.2368809580802917} 01/28/2022 00:03:24 - INFO - codeparrot_training - Step 4144: {'lr': 0.0004975426665683678, 'samples': 795840, 'steps': 4144, 'loss/train': 4.567713975906372} 01/28/2022 00:03:28 - INFO - codeparrot_training - Step 4145: {'lr': 0.000497540377512205, 'samples': 796032, 'steps': 4145, 'loss/train': 3.38072669506073} 01/28/2022 00:03:33 - INFO - codeparrot_training - Step 4146: {'lr': 0.0004975380873956577, 'samples': 796224, 'steps': 4146, 'loss/train': 3.0132064819335938} 01/28/2022 00:03:37 - INFO - codeparrot_training - Step 4147: {'lr': 0.0004975357962187359, 'samples': 796416, 'steps': 4147, 'loss/train': 3.8639012575149536} 01/28/2022 00:03:41 - INFO - codeparrot_training - Step 4148: {'lr': 0.0004975335039814493, 'samples': 796608, 'steps': 4148, 'loss/train': 3.2491679191589355} 01/28/2022 00:03:46 - INFO - codeparrot_training - Step 4149: {'lr': 0.0004975312106838079, 'samples': 796800, 'steps': 4149, 'loss/train': 3.604928970336914} 01/28/2022 00:03:50 - INFO - codeparrot_training - Step 4150: {'lr': 0.0004975289163258214, 'samples': 796992, 'steps': 4150, 'loss/train': 3.4417508840560913} 01/28/2022 00:03:55 - INFO - codeparrot_training - Step 4151: {'lr': 0.0004975266209074995, 'samples': 797184, 'steps': 4151, 'loss/train': 2.8513898849487305} 01/28/2022 00:03:59 - INFO - codeparrot_training - Step 4152: {'lr': 0.0004975243244288522, 'samples': 797376, 'steps': 4152, 'loss/train': 1.5756443738937378} 01/28/2022 00:04:03 - INFO - codeparrot_training - Step 4153: {'lr': 0.0004975220268898893, 'samples': 797568, 'steps': 4153, 'loss/train': 2.8975263833999634} 01/28/2022 00:04:08 - INFO - codeparrot_training - Step 4154: {'lr': 0.0004975197282906207, 'samples': 797760, 'steps': 4154, 'loss/train': 1.8818628787994385} 01/28/2022 00:04:13 - INFO - codeparrot_training - Step 4155: {'lr': 0.0004975174286310562, 'samples': 797952, 'steps': 4155, 'loss/train': 3.252563953399658} 01/28/2022 00:04:17 - INFO - codeparrot_training - Step 4156: {'lr': 0.0004975151279112054, 'samples': 798144, 'steps': 4156, 'loss/train': 0.715496689081192} 01/28/2022 00:04:21 - INFO - codeparrot_training - Step 4157: {'lr': 0.0004975128261310787, 'samples': 798336, 'steps': 4157, 'loss/train': 3.9883675575256348} 01/28/2022 00:04:25 - INFO - codeparrot_training - Step 4158: {'lr': 0.0004975105232906854, 'samples': 798528, 'steps': 4158, 'loss/train': 1.6634883284568787} 01/28/2022 00:04:31 - INFO - codeparrot_training - Step 4159: {'lr': 0.0004975082193900357, 'samples': 798720, 'steps': 4159, 'loss/train': 3.0342679023742676} 01/28/2022 00:04:35 - INFO - codeparrot_training - Step 4160: {'lr': 0.0004975059144291394, 'samples': 798912, 'steps': 4160, 'loss/train': 4.069201111793518} 01/28/2022 00:04:39 - INFO - codeparrot_training - Step 4161: {'lr': 0.0004975036084080063, 'samples': 799104, 'steps': 4161, 'loss/train': 2.6430999040603638} 01/28/2022 00:04:43 - INFO - codeparrot_training - Step 4162: {'lr': 0.0004975013013266464, 'samples': 799296, 'steps': 4162, 'loss/train': 3.4096577167510986} 01/28/2022 00:04:47 - INFO - codeparrot_training - Step 4163: {'lr': 0.0004974989931850695, 'samples': 799488, 'steps': 4163, 'loss/train': 2.9196519255638123} 01/28/2022 00:04:53 - INFO - codeparrot_training - Step 4164: {'lr': 0.0004974966839832855, 'samples': 799680, 'steps': 4164, 'loss/train': 2.0015282034873962} 01/28/2022 00:04:57 - INFO - codeparrot_training - Step 4165: {'lr': 0.0004974943737213042, 'samples': 799872, 'steps': 4165, 'loss/train': 2.501125931739807} 01/28/2022 00:05:02 - INFO - codeparrot_training - Step 4166: {'lr': 0.0004974920623991356, 'samples': 800064, 'steps': 4166, 'loss/train': 2.7302294969558716} 01/28/2022 00:05:06 - INFO - codeparrot_training - Step 4167: {'lr': 0.0004974897500167898, 'samples': 800256, 'steps': 4167, 'loss/train': 1.5830102562904358} 01/28/2022 00:05:10 - INFO - codeparrot_training - Step 4168: {'lr': 0.0004974874365742763, 'samples': 800448, 'steps': 4168, 'loss/train': 3.7316783666610718} 01/28/2022 00:05:15 - INFO - codeparrot_training - Step 4169: {'lr': 0.0004974851220716053, 'samples': 800640, 'steps': 4169, 'loss/train': 3.955415725708008} 01/28/2022 00:05:20 - INFO - codeparrot_training - Step 4170: {'lr': 0.0004974828065087867, 'samples': 800832, 'steps': 4170, 'loss/train': 1.7044386863708496} 01/28/2022 00:05:24 - INFO - codeparrot_training - Step 4171: {'lr': 0.0004974804898858302, 'samples': 801024, 'steps': 4171, 'loss/train': 2.978027880191803} 01/28/2022 00:05:28 - INFO - codeparrot_training - Step 4172: {'lr': 0.0004974781722027459, 'samples': 801216, 'steps': 4172, 'loss/train': 1.654507040977478} 01/28/2022 00:05:32 - INFO - codeparrot_training - Step 4173: {'lr': 0.0004974758534595436, 'samples': 801408, 'steps': 4173, 'loss/train': 2.99757581949234} 01/28/2022 00:05:38 - INFO - codeparrot_training - Step 4174: {'lr': 0.0004974735336562335, 'samples': 801600, 'steps': 4174, 'loss/train': 3.004897356033325} 01/28/2022 00:05:42 - INFO - codeparrot_training - Step 4175: {'lr': 0.0004974712127928252, 'samples': 801792, 'steps': 4175, 'loss/train': 4.305615663528442} 01/28/2022 00:05:46 - INFO - codeparrot_training - Step 4176: {'lr': 0.000497468890869329, 'samples': 801984, 'steps': 4176, 'loss/train': 1.3944799304008484} 01/28/2022 00:05:50 - INFO - codeparrot_training - Step 4177: {'lr': 0.0004974665678857545, 'samples': 802176, 'steps': 4177, 'loss/train': 3.3582347631454468} 01/28/2022 00:05:54 - INFO - codeparrot_training - Step 4178: {'lr': 0.0004974642438421118, 'samples': 802368, 'steps': 4178, 'loss/train': 3.4732428789138794} 01/28/2022 00:06:00 - INFO - codeparrot_training - Step 4179: {'lr': 0.0004974619187384109, 'samples': 802560, 'steps': 4179, 'loss/train': 3.2671691179275513} 01/28/2022 00:06:05 - INFO - codeparrot_training - Step 4180: {'lr': 0.0004974595925746618, 'samples': 802752, 'steps': 4180, 'loss/train': 3.2326308488845825} 01/28/2022 00:06:09 - INFO - codeparrot_training - Step 4181: {'lr': 0.0004974572653508742, 'samples': 802944, 'steps': 4181, 'loss/train': 3.2105458974838257} 01/28/2022 00:06:13 - INFO - codeparrot_training - Step 4182: {'lr': 0.0004974549370670584, 'samples': 803136, 'steps': 4182, 'loss/train': 2.41939115524292} 01/28/2022 00:06:17 - INFO - codeparrot_training - Step 4183: {'lr': 0.0004974526077232242, 'samples': 803328, 'steps': 4183, 'loss/train': 3.6193220615386963} 01/28/2022 00:06:23 - INFO - codeparrot_training - Step 4184: {'lr': 0.0004974502773193815, 'samples': 803520, 'steps': 4184, 'loss/train': 3.0816493034362793} 01/28/2022 00:06:27 - INFO - codeparrot_training - Step 4185: {'lr': 0.0004974479458555405, 'samples': 803712, 'steps': 4185, 'loss/train': 3.1134181022644043} 01/28/2022 00:06:31 - INFO - codeparrot_training - Step 4186: {'lr': 0.000497445613331711, 'samples': 803904, 'steps': 4186, 'loss/train': 3.73685359954834} 01/28/2022 00:06:36 - INFO - codeparrot_training - Step 4187: {'lr': 0.0004974432797479032, 'samples': 804096, 'steps': 4187, 'loss/train': 3.3896448612213135} 01/28/2022 00:06:40 - INFO - codeparrot_training - Step 4188: {'lr': 0.0004974409451041268, 'samples': 804288, 'steps': 4188, 'loss/train': 3.2279698848724365} 01/28/2022 00:06:45 - INFO - codeparrot_training - Step 4189: {'lr': 0.0004974386094003921, 'samples': 804480, 'steps': 4189, 'loss/train': 2.7518261075019836} 01/28/2022 00:06:49 - INFO - codeparrot_training - Step 4190: {'lr': 0.0004974362726367089, 'samples': 804672, 'steps': 4190, 'loss/train': 3.4131789207458496} 01/28/2022 00:06:54 - INFO - codeparrot_training - Step 4191: {'lr': 0.0004974339348130873, 'samples': 804864, 'steps': 4191, 'loss/train': 2.951314151287079} 01/28/2022 00:06:58 - INFO - codeparrot_training - Step 4192: {'lr': 0.0004974315959295373, 'samples': 805056, 'steps': 4192, 'loss/train': 2.974318027496338} 01/28/2022 00:07:02 - INFO - codeparrot_training - Step 4193: {'lr': 0.0004974292559860688, 'samples': 805248, 'steps': 4193, 'loss/train': 3.3389317989349365} 01/28/2022 00:07:08 - INFO - codeparrot_training - Step 4194: {'lr': 0.0004974269149826921, 'samples': 805440, 'steps': 4194, 'loss/train': 3.161848783493042} 01/28/2022 00:07:12 - INFO - codeparrot_training - Step 4195: {'lr': 0.0004974245729194169, 'samples': 805632, 'steps': 4195, 'loss/train': 2.7896599173545837} 01/28/2022 00:07:17 - INFO - codeparrot_training - Step 4196: {'lr': 0.0004974222297962535, 'samples': 805824, 'steps': 4196, 'loss/train': 3.5538575649261475} 01/28/2022 00:07:21 - INFO - codeparrot_training - Step 4197: {'lr': 0.0004974198856132118, 'samples': 806016, 'steps': 4197, 'loss/train': 0.956319272518158} 01/28/2022 00:07:26 - INFO - codeparrot_training - Step 4198: {'lr': 0.0004974175403703019, 'samples': 806208, 'steps': 4198, 'loss/train': 2.2200735211372375} 01/28/2022 00:07:30 - INFO - codeparrot_training - Step 4199: {'lr': 0.0004974151940675338, 'samples': 806400, 'steps': 4199, 'loss/train': 1.7265116572380066} 01/28/2022 00:07:35 - INFO - codeparrot_training - Step 4200: {'lr': 0.0004974128467049176, 'samples': 806592, 'steps': 4200, 'loss/train': 2.915426731109619} 01/28/2022 00:07:39 - INFO - codeparrot_training - Step 4201: {'lr': 0.0004974104982824632, 'samples': 806784, 'steps': 4201, 'loss/train': 3.184661865234375} 01/28/2022 00:07:43 - INFO - codeparrot_training - Step 4202: {'lr': 0.0004974081488001809, 'samples': 806976, 'steps': 4202, 'loss/train': 3.0781803131103516} 01/28/2022 00:07:48 - INFO - codeparrot_training - Step 4203: {'lr': 0.0004974057982580806, 'samples': 807168, 'steps': 4203, 'loss/train': 3.5051082372665405} 01/28/2022 00:07:53 - INFO - codeparrot_training - Step 4204: {'lr': 0.0004974034466561725, 'samples': 807360, 'steps': 4204, 'loss/train': 3.1244384050369263} 01/28/2022 00:07:57 - INFO - codeparrot_training - Step 4205: {'lr': 0.0004974010939944667, 'samples': 807552, 'steps': 4205, 'loss/train': 2.909259617328644} 01/28/2022 00:08:01 - INFO - codeparrot_training - Step 4206: {'lr': 0.0004973987402729729, 'samples': 807744, 'steps': 4206, 'loss/train': 3.5375490188598633} 01/28/2022 00:08:06 - INFO - codeparrot_training - Step 4207: {'lr': 0.0004973963854917016, 'samples': 807936, 'steps': 4207, 'loss/train': 1.9392920136451721} 01/28/2022 00:08:11 - INFO - codeparrot_training - Step 4208: {'lr': 0.0004973940296506627, 'samples': 808128, 'steps': 4208, 'loss/train': 3.7655396461486816} 01/28/2022 00:08:16 - INFO - codeparrot_training - Step 4209: {'lr': 0.0004973916727498664, 'samples': 808320, 'steps': 4209, 'loss/train': 2.805968999862671} 01/28/2022 00:08:20 - INFO - codeparrot_training - Step 4210: {'lr': 0.0004973893147893227, 'samples': 808512, 'steps': 4210, 'loss/train': 2.7582555413246155} 01/28/2022 00:08:24 - INFO - codeparrot_training - Step 4211: {'lr': 0.0004973869557690417, 'samples': 808704, 'steps': 4211, 'loss/train': 1.0004096925258636} 01/28/2022 00:08:28 - INFO - codeparrot_training - Step 4212: {'lr': 0.0004973845956890336, 'samples': 808896, 'steps': 4212, 'loss/train': 3.156450390815735} 01/28/2022 00:08:33 - INFO - codeparrot_training - Step 4213: {'lr': 0.0004973822345493084, 'samples': 809088, 'steps': 4213, 'loss/train': 3.0673112869262695} 01/28/2022 00:08:38 - INFO - codeparrot_training - Step 4214: {'lr': 0.0004973798723498762, 'samples': 809280, 'steps': 4214, 'loss/train': 2.727433919906616} 01/28/2022 00:08:42 - INFO - codeparrot_training - Step 4215: {'lr': 0.0004973775090907473, 'samples': 809472, 'steps': 4215, 'loss/train': 3.52630877494812} 01/28/2022 00:08:46 - INFO - codeparrot_training - Step 4216: {'lr': 0.0004973751447719316, 'samples': 809664, 'steps': 4216, 'loss/train': 2.0875731110572815} 01/28/2022 00:08:50 - INFO - codeparrot_training - Step 4217: {'lr': 0.0004973727793934394, 'samples': 809856, 'steps': 4217, 'loss/train': 3.8472511768341064} 01/28/2022 00:08:56 - INFO - codeparrot_training - Step 4218: {'lr': 0.0004973704129552808, 'samples': 810048, 'steps': 4218, 'loss/train': 1.9011293649673462} 01/28/2022 00:09:00 - INFO - codeparrot_training - Step 4219: {'lr': 0.0004973680454574657, 'samples': 810240, 'steps': 4219, 'loss/train': 3.7694320678710938} 01/28/2022 00:09:04 - INFO - codeparrot_training - Step 4220: {'lr': 0.0004973656769000046, 'samples': 810432, 'steps': 4220, 'loss/train': 2.8754644989967346} 01/28/2022 00:09:08 - INFO - codeparrot_training - Step 4221: {'lr': 0.0004973633072829075, 'samples': 810624, 'steps': 4221, 'loss/train': 2.168702781200409} 01/28/2022 00:09:13 - INFO - codeparrot_training - Step 4222: {'lr': 0.0004973609366061845, 'samples': 810816, 'steps': 4222, 'loss/train': 3.4372200965881348} 01/28/2022 00:09:19 - INFO - codeparrot_training - Step 4223: {'lr': 0.0004973585648698457, 'samples': 811008, 'steps': 4223, 'loss/train': 3.6015021800994873} 01/28/2022 00:09:23 - INFO - codeparrot_training - Step 4224: {'lr': 0.0004973561920739015, 'samples': 811200, 'steps': 4224, 'loss/train': 3.0955653190612793} 01/28/2022 00:09:27 - INFO - codeparrot_training - Step 4225: {'lr': 0.0004973538182183618, 'samples': 811392, 'steps': 4225, 'loss/train': 2.735649347305298} 01/28/2022 00:09:31 - INFO - codeparrot_training - Step 4226: {'lr': 0.000497351443303237, 'samples': 811584, 'steps': 4226, 'loss/train': 3.283656120300293} 01/28/2022 00:09:36 - INFO - codeparrot_training - Step 4227: {'lr': 0.0004973490673285372, 'samples': 811776, 'steps': 4227, 'loss/train': 1.0418738722801208} 01/28/2022 00:09:41 - INFO - codeparrot_training - Step 4228: {'lr': 0.0004973466902942723, 'samples': 811968, 'steps': 4228, 'loss/train': 3.2676608562469482} 01/28/2022 00:09:45 - INFO - codeparrot_training - Step 4229: {'lr': 0.0004973443122004529, 'samples': 812160, 'steps': 4229, 'loss/train': 2.104238748550415} 01/28/2022 00:09:49 - INFO - codeparrot_training - Step 4230: {'lr': 0.0004973419330470891, 'samples': 812352, 'steps': 4230, 'loss/train': 2.603497803211212} 01/28/2022 00:09:54 - INFO - codeparrot_training - Step 4231: {'lr': 0.0004973395528341908, 'samples': 812544, 'steps': 4231, 'loss/train': 4.029016613960266} 01/28/2022 00:09:58 - INFO - codeparrot_training - Step 4232: {'lr': 0.0004973371715617685, 'samples': 812736, 'steps': 4232, 'loss/train': 2.1336176991462708} 01/28/2022 00:10:04 - INFO - codeparrot_training - Step 4233: {'lr': 0.0004973347892298322, 'samples': 812928, 'steps': 4233, 'loss/train': 2.903866410255432} 01/28/2022 00:10:08 - INFO - codeparrot_training - Step 4234: {'lr': 0.0004973324058383924, 'samples': 813120, 'steps': 4234, 'loss/train': 2.3196126222610474} 01/28/2022 00:10:12 - INFO - codeparrot_training - Step 4235: {'lr': 0.0004973300213874589, 'samples': 813312, 'steps': 4235, 'loss/train': 2.789303719997406} 01/28/2022 00:10:16 - INFO - codeparrot_training - Step 4236: {'lr': 0.0004973276358770422, 'samples': 813504, 'steps': 4236, 'loss/train': 2.95468407869339} 01/28/2022 00:10:21 - INFO - codeparrot_training - Step 4237: {'lr': 0.0004973252493071525, 'samples': 813696, 'steps': 4237, 'loss/train': 1.6037052869796753} 01/28/2022 00:10:26 - INFO - codeparrot_training - Step 4238: {'lr': 0.0004973228616777999, 'samples': 813888, 'steps': 4238, 'loss/train': 4.058254837989807} 01/28/2022 00:10:30 - INFO - codeparrot_training - Step 4239: {'lr': 0.0004973204729889946, 'samples': 814080, 'steps': 4239, 'loss/train': 2.597981035709381} 01/28/2022 00:10:34 - INFO - codeparrot_training - Step 4240: {'lr': 0.0004973180832407472, 'samples': 814272, 'steps': 4240, 'loss/train': 3.0699291229248047} 01/28/2022 00:10:38 - INFO - codeparrot_training - Step 4241: {'lr': 0.0004973156924330674, 'samples': 814464, 'steps': 4241, 'loss/train': 4.390836596488953} 01/28/2022 00:10:43 - INFO - codeparrot_training - Step 4242: {'lr': 0.0004973133005659658, 'samples': 814656, 'steps': 4242, 'loss/train': 2.93784499168396} 01/28/2022 00:10:48 - INFO - codeparrot_training - Step 4243: {'lr': 0.0004973109076394526, 'samples': 814848, 'steps': 4243, 'loss/train': 2.375382363796234} 01/28/2022 00:10:52 - INFO - codeparrot_training - Step 4244: {'lr': 0.0004973085136535379, 'samples': 815040, 'steps': 4244, 'loss/train': 2.2679116129875183} 01/28/2022 00:10:56 - INFO - codeparrot_training - Step 4245: {'lr': 0.000497306118608232, 'samples': 815232, 'steps': 4245, 'loss/train': 1.0900056660175323} 01/28/2022 00:11:01 - INFO - codeparrot_training - Step 4246: {'lr': 0.0004973037225035454, 'samples': 815424, 'steps': 4246, 'loss/train': 2.651662766933441} 01/28/2022 00:11:05 - INFO - codeparrot_training - Step 4247: {'lr': 0.0004973013253394881, 'samples': 815616, 'steps': 4247, 'loss/train': 3.9519245624542236} 01/28/2022 00:11:11 - INFO - codeparrot_training - Step 4248: {'lr': 0.0004972989271160705, 'samples': 815808, 'steps': 4248, 'loss/train': 2.52214515209198} 01/28/2022 00:11:15 - INFO - codeparrot_training - Step 4249: {'lr': 0.0004972965278333028, 'samples': 816000, 'steps': 4249, 'loss/train': 3.373993992805481} 01/28/2022 00:11:19 - INFO - codeparrot_training - Step 4250: {'lr': 0.0004972941274911952, 'samples': 816192, 'steps': 4250, 'loss/train': 2.2682341933250427} 01/28/2022 00:11:24 - INFO - codeparrot_training - Step 4251: {'lr': 0.0004972917260897583, 'samples': 816384, 'steps': 4251, 'loss/train': 3.153997778892517} 01/28/2022 00:11:28 - INFO - codeparrot_training - Step 4252: {'lr': 0.0004972893236290019, 'samples': 816576, 'steps': 4252, 'loss/train': 3.660583019256592} 01/28/2022 00:11:33 - INFO - codeparrot_training - Step 4253: {'lr': 0.0004972869201089367, 'samples': 816768, 'steps': 4253, 'loss/train': 3.616895914077759} 01/28/2022 00:11:37 - INFO - codeparrot_training - Step 4254: {'lr': 0.0004972845155295729, 'samples': 816960, 'steps': 4254, 'loss/train': 0.998762458562851} 01/28/2022 00:11:41 - INFO - codeparrot_training - Step 4255: {'lr': 0.0004972821098909207, 'samples': 817152, 'steps': 4255, 'loss/train': 3.1701500415802} 01/28/2022 00:11:46 - INFO - codeparrot_training - Step 4256: {'lr': 0.0004972797031929904, 'samples': 817344, 'steps': 4256, 'loss/train': 2.2221354246139526} 01/28/2022 00:11:50 - INFO - codeparrot_training - Step 4257: {'lr': 0.0004972772954357924, 'samples': 817536, 'steps': 4257, 'loss/train': 2.8309484124183655} 01/28/2022 00:11:55 - INFO - codeparrot_training - Step 4258: {'lr': 0.0004972748866193371, 'samples': 817728, 'steps': 4258, 'loss/train': 2.2935057878494263} 01/28/2022 00:11:59 - INFO - codeparrot_training - Step 4259: {'lr': 0.0004972724767436346, 'samples': 817920, 'steps': 4259, 'loss/train': 1.86215078830719} 01/28/2022 00:12:04 - INFO - codeparrot_training - Step 4260: {'lr': 0.0004972700658086954, 'samples': 818112, 'steps': 4260, 'loss/train': 1.7797220349311829} 01/28/2022 00:12:08 - INFO - codeparrot_training - Step 4261: {'lr': 0.0004972676538145298, 'samples': 818304, 'steps': 4261, 'loss/train': 2.5476842522621155} 01/28/2022 00:12:12 - INFO - codeparrot_training - Step 4262: {'lr': 0.0004972652407611479, 'samples': 818496, 'steps': 4262, 'loss/train': 3.103294014930725} 01/28/2022 00:12:18 - INFO - codeparrot_training - Step 4263: {'lr': 0.0004972628266485604, 'samples': 818688, 'steps': 4263, 'loss/train': 2.4207538962364197} 01/28/2022 00:12:22 - INFO - codeparrot_training - Step 4264: {'lr': 0.0004972604114767774, 'samples': 818880, 'steps': 4264, 'loss/train': 1.594680905342102} 01/28/2022 00:12:26 - INFO - codeparrot_training - Step 4265: {'lr': 0.0004972579952458092, 'samples': 819072, 'steps': 4265, 'loss/train': 2.458526372909546} 01/28/2022 00:12:30 - INFO - codeparrot_training - Step 4266: {'lr': 0.0004972555779556664, 'samples': 819264, 'steps': 4266, 'loss/train': 2.8098623156547546} 01/28/2022 00:12:35 - INFO - codeparrot_training - Step 4267: {'lr': 0.0004972531596063592, 'samples': 819456, 'steps': 4267, 'loss/train': 1.0761970281600952} 01/28/2022 00:12:41 - INFO - codeparrot_training - Step 4268: {'lr': 0.000497250740197898, 'samples': 819648, 'steps': 4268, 'loss/train': 2.120331287384033} 01/28/2022 00:12:45 - INFO - codeparrot_training - Step 4269: {'lr': 0.0004972483197302931, 'samples': 819840, 'steps': 4269, 'loss/train': 3.7808332443237305} 01/28/2022 00:12:49 - INFO - codeparrot_training - Step 4270: {'lr': 0.0004972458982035548, 'samples': 820032, 'steps': 4270, 'loss/train': 3.713672161102295} 01/28/2022 00:12:53 - INFO - codeparrot_training - Step 4271: {'lr': 0.0004972434756176937, 'samples': 820224, 'steps': 4271, 'loss/train': 3.7632973194122314} 01/28/2022 00:12:58 - INFO - codeparrot_training - Step 4272: {'lr': 0.0004972410519727201, 'samples': 820416, 'steps': 4272, 'loss/train': 2.668591797351837} 01/28/2022 00:13:03 - INFO - codeparrot_training - Step 4273: {'lr': 0.0004972386272686443, 'samples': 820608, 'steps': 4273, 'loss/train': 2.4389038681983948} 01/28/2022 00:13:07 - INFO - codeparrot_training - Step 4274: {'lr': 0.0004972362015054767, 'samples': 820800, 'steps': 4274, 'loss/train': 3.625276565551758} 01/28/2022 00:13:12 - INFO - codeparrot_training - Step 4275: {'lr': 0.0004972337746832278, 'samples': 820992, 'steps': 4275, 'loss/train': 0.8251339495182037} 01/28/2022 00:13:16 - INFO - codeparrot_training - Step 4276: {'lr': 0.0004972313468019077, 'samples': 821184, 'steps': 4276, 'loss/train': 3.882387399673462} 01/28/2022 00:13:20 - INFO - codeparrot_training - Step 4277: {'lr': 0.0004972289178615273, 'samples': 821376, 'steps': 4277, 'loss/train': 3.1740152835845947} 01/28/2022 00:13:26 - INFO - codeparrot_training - Step 4278: {'lr': 0.0004972264878620965, 'samples': 821568, 'steps': 4278, 'loss/train': 3.411125421524048} 01/28/2022 00:13:30 - INFO - codeparrot_training - Step 4279: {'lr': 0.000497224056803626, 'samples': 821760, 'steps': 4279, 'loss/train': 3.1039488315582275} 01/28/2022 00:13:34 - INFO - codeparrot_training - Step 4280: {'lr': 0.0004972216246861262, 'samples': 821952, 'steps': 4280, 'loss/train': 3.4773906469345093} 01/28/2022 00:13:38 - INFO - codeparrot_training - Step 4281: {'lr': 0.0004972191915096074, 'samples': 822144, 'steps': 4281, 'loss/train': 3.2249879837036133} 01/28/2022 00:13:43 - INFO - codeparrot_training - Step 4282: {'lr': 0.0004972167572740801, 'samples': 822336, 'steps': 4282, 'loss/train': 2.958386778831482} 01/28/2022 00:13:48 - INFO - codeparrot_training - Step 4283: {'lr': 0.0004972143219795547, 'samples': 822528, 'steps': 4283, 'loss/train': 2.2743422985076904} 01/28/2022 00:13:53 - INFO - codeparrot_training - Step 4284: {'lr': 0.0004972118856260416, 'samples': 822720, 'steps': 4284, 'loss/train': 2.5429569482803345} 01/28/2022 00:13:57 - INFO - codeparrot_training - Step 4285: {'lr': 0.0004972094482135514, 'samples': 822912, 'steps': 4285, 'loss/train': 3.671410918235779} 01/28/2022 00:14:01 - INFO - codeparrot_training - Step 4286: {'lr': 0.0004972070097420943, 'samples': 823104, 'steps': 4286, 'loss/train': 2.7792019844055176} 01/28/2022 00:14:05 - INFO - codeparrot_training - Step 4287: {'lr': 0.0004972045702116809, 'samples': 823296, 'steps': 4287, 'loss/train': 2.9588255882263184} 01/28/2022 00:14:11 - INFO - codeparrot_training - Step 4288: {'lr': 0.0004972021296223217, 'samples': 823488, 'steps': 4288, 'loss/train': 3.076267719268799} 01/28/2022 00:14:15 - INFO - codeparrot_training - Step 4289: {'lr': 0.0004971996879740271, 'samples': 823680, 'steps': 4289, 'loss/train': 2.7468754649162292} 01/28/2022 00:14:19 - INFO - codeparrot_training - Step 4290: {'lr': 0.0004971972452668074, 'samples': 823872, 'steps': 4290, 'loss/train': 3.5329291820526123} 01/28/2022 00:14:23 - INFO - codeparrot_training - Step 4291: {'lr': 0.0004971948015006732, 'samples': 824064, 'steps': 4291, 'loss/train': 2.935393810272217} 01/28/2022 00:14:28 - INFO - codeparrot_training - Step 4292: {'lr': 0.000497192356675635, 'samples': 824256, 'steps': 4292, 'loss/train': 3.7156684398651123} 01/28/2022 00:14:34 - INFO - codeparrot_training - Step 4293: {'lr': 0.0004971899107917033, 'samples': 824448, 'steps': 4293, 'loss/train': 3.61518931388855} 01/28/2022 00:14:38 - INFO - codeparrot_training - Step 4294: {'lr': 0.0004971874638488884, 'samples': 824640, 'steps': 4294, 'loss/train': 3.4388837814331055} 01/28/2022 00:14:42 - INFO - codeparrot_training - Step 4295: {'lr': 0.000497185015847201, 'samples': 824832, 'steps': 4295, 'loss/train': 3.165456175804138} 01/28/2022 00:14:46 - INFO - codeparrot_training - Step 4296: {'lr': 0.0004971825667866515, 'samples': 825024, 'steps': 4296, 'loss/train': 3.9295395612716675} 01/28/2022 00:14:51 - INFO - codeparrot_training - Step 4297: {'lr': 0.0004971801166672502, 'samples': 825216, 'steps': 4297, 'loss/train': 2.8070322275161743} 01/28/2022 00:14:56 - INFO - codeparrot_training - Step 4298: {'lr': 0.0004971776654890079, 'samples': 825408, 'steps': 4298, 'loss/train': 3.6585277318954468} 01/28/2022 00:15:00 - INFO - codeparrot_training - Step 4299: {'lr': 0.000497175213251935, 'samples': 825600, 'steps': 4299, 'loss/train': 3.354020118713379} 01/28/2022 00:15:04 - INFO - codeparrot_training - Step 4300: {'lr': 0.0004971727599560418, 'samples': 825792, 'steps': 4300, 'loss/train': 3.229189395904541} 01/28/2022 00:15:08 - INFO - codeparrot_training - Step 4301: {'lr': 0.0004971703056013392, 'samples': 825984, 'steps': 4301, 'loss/train': 3.44454824924469} 01/28/2022 00:15:13 - INFO - codeparrot_training - Step 4302: {'lr': 0.0004971678501878374, 'samples': 826176, 'steps': 4302, 'loss/train': 4.313030004501343} 01/28/2022 00:15:18 - INFO - codeparrot_training - Step 4303: {'lr': 0.000497165393715547, 'samples': 826368, 'steps': 4303, 'loss/train': 3.5911835432052612} 01/28/2022 00:15:22 - INFO - codeparrot_training - Step 4304: {'lr': 0.0004971629361844785, 'samples': 826560, 'steps': 4304, 'loss/train': 2.6549758315086365} 01/28/2022 00:15:26 - INFO - codeparrot_training - Step 4305: {'lr': 0.0004971604775946425, 'samples': 826752, 'steps': 4305, 'loss/train': 3.600569486618042} 01/28/2022 00:15:31 - INFO - codeparrot_training - Step 4306: {'lr': 0.0004971580179460495, 'samples': 826944, 'steps': 4306, 'loss/train': 2.3597259521484375} 01/28/2022 00:15:35 - INFO - codeparrot_training - Step 4307: {'lr': 0.0004971555572387101, 'samples': 827136, 'steps': 4307, 'loss/train': 2.5814900994300842} 01/28/2022 00:15:41 - INFO - codeparrot_training - Step 4308: {'lr': 0.0004971530954726346, 'samples': 827328, 'steps': 4308, 'loss/train': 1.8455341458320618} 01/28/2022 00:15:45 - INFO - codeparrot_training - Step 4309: {'lr': 0.0004971506326478339, 'samples': 827520, 'steps': 4309, 'loss/train': 3.015035033226013} 01/28/2022 00:15:49 - INFO - codeparrot_training - Step 4310: {'lr': 0.0004971481687643184, 'samples': 827712, 'steps': 4310, 'loss/train': 3.4874435663223267} 01/28/2022 00:15:53 - INFO - codeparrot_training - Step 4311: {'lr': 0.0004971457038220984, 'samples': 827904, 'steps': 4311, 'loss/train': 3.5517818927764893} 01/28/2022 00:15:58 - INFO - codeparrot_training - Step 4312: {'lr': 0.0004971432378211849, 'samples': 828096, 'steps': 4312, 'loss/train': 3.566014051437378} 01/28/2022 00:16:04 - INFO - codeparrot_training - Step 4313: {'lr': 0.0004971407707615881, 'samples': 828288, 'steps': 4313, 'loss/train': 3.2961437702178955} 01/28/2022 00:16:08 - INFO - codeparrot_training - Step 4314: {'lr': 0.0004971383026433189, 'samples': 828480, 'steps': 4314, 'loss/train': 4.680072784423828} 01/28/2022 00:16:12 - INFO - codeparrot_training - Step 4315: {'lr': 0.0004971358334663875, 'samples': 828672, 'steps': 4315, 'loss/train': 2.9275283217430115} 01/28/2022 00:16:17 - INFO - codeparrot_training - Step 4316: {'lr': 0.0004971333632308047, 'samples': 828864, 'steps': 4316, 'loss/train': 2.5763062834739685} 01/28/2022 00:16:21 - INFO - codeparrot_training - Step 4317: {'lr': 0.000497130891936581, 'samples': 829056, 'steps': 4317, 'loss/train': 3.724366307258606} 01/28/2022 00:16:26 - INFO - codeparrot_training - Step 4318: {'lr': 0.0004971284195837271, 'samples': 829248, 'steps': 4318, 'loss/train': 3.279073119163513} 01/28/2022 00:16:30 - INFO - codeparrot_training - Step 4319: {'lr': 0.0004971259461722536, 'samples': 829440, 'steps': 4319, 'loss/train': 3.327132225036621} 01/28/2022 00:16:35 - INFO - codeparrot_training - Step 4320: {'lr': 0.0004971234717021708, 'samples': 829632, 'steps': 4320, 'loss/train': 2.1655207872390747} 01/28/2022 00:16:39 - INFO - codeparrot_training - Step 4321: {'lr': 0.0004971209961734897, 'samples': 829824, 'steps': 4321, 'loss/train': 3.52341091632843} 01/28/2022 00:16:43 - INFO - codeparrot_training - Step 4322: {'lr': 0.0004971185195862207, 'samples': 830016, 'steps': 4322, 'loss/train': 3.6618164777755737} 01/28/2022 00:16:49 - INFO - codeparrot_training - Step 4323: {'lr': 0.0004971160419403744, 'samples': 830208, 'steps': 4323, 'loss/train': 3.7396109104156494} 01/28/2022 00:16:54 - INFO - codeparrot_training - Step 4324: {'lr': 0.0004971135632359614, 'samples': 830400, 'steps': 4324, 'loss/train': 3.819143772125244} 01/28/2022 00:16:58 - INFO - codeparrot_training - Step 4325: {'lr': 0.0004971110834729925, 'samples': 830592, 'steps': 4325, 'loss/train': 3.3618972301483154} 01/28/2022 00:17:02 - INFO - codeparrot_training - Step 4326: {'lr': 0.0004971086026514781, 'samples': 830784, 'steps': 4326, 'loss/train': 3.724114179611206} 01/28/2022 00:17:06 - INFO - codeparrot_training - Step 4327: {'lr': 0.0004971061207714289, 'samples': 830976, 'steps': 4327, 'loss/train': 2.1292170882225037} 01/28/2022 00:17:11 - INFO - codeparrot_training - Step 4328: {'lr': 0.0004971036378328556, 'samples': 831168, 'steps': 4328, 'loss/train': 2.020463526248932} 01/28/2022 00:17:16 - INFO - codeparrot_training - Step 4329: {'lr': 0.0004971011538357687, 'samples': 831360, 'steps': 4329, 'loss/train': 3.560525894165039} 01/28/2022 00:17:20 - INFO - codeparrot_training - Step 4330: {'lr': 0.000497098668780179, 'samples': 831552, 'steps': 4330, 'loss/train': 3.623387932777405} 01/28/2022 00:17:24 - INFO - codeparrot_training - Step 4331: {'lr': 0.000497096182666097, 'samples': 831744, 'steps': 4331, 'loss/train': 3.447218656539917} 01/28/2022 00:17:29 - INFO - codeparrot_training - Step 4332: {'lr': 0.0004970936954935334, 'samples': 831936, 'steps': 4332, 'loss/train': 3.328229069709778} 01/28/2022 00:17:33 - INFO - codeparrot_training - Step 4333: {'lr': 0.0004970912072624989, 'samples': 832128, 'steps': 4333, 'loss/train': 2.60619056224823} 01/28/2022 00:17:38 - INFO - codeparrot_training - Step 4334: {'lr': 0.0004970887179730041, 'samples': 832320, 'steps': 4334, 'loss/train': 1.4978318810462952} 01/28/2022 00:17:42 - INFO - codeparrot_training - Step 4335: {'lr': 0.0004970862276250599, 'samples': 832512, 'steps': 4335, 'loss/train': 3.5946807861328125} 01/28/2022 00:17:47 - INFO - codeparrot_training - Step 4336: {'lr': 0.0004970837362186766, 'samples': 832704, 'steps': 4336, 'loss/train': 3.2108652591705322} 01/28/2022 00:17:51 - INFO - codeparrot_training - Step 4337: {'lr': 0.0004970812437538649, 'samples': 832896, 'steps': 4337, 'loss/train': 3.1039645671844482} 01/28/2022 00:17:55 - INFO - codeparrot_training - Step 4338: {'lr': 0.0004970787502306357, 'samples': 833088, 'steps': 4338, 'loss/train': 2.0059157609939575} 01/28/2022 00:18:01 - INFO - codeparrot_training - Step 4339: {'lr': 0.0004970762556489996, 'samples': 833280, 'steps': 4339, 'loss/train': 11.941577196121216} 01/28/2022 00:18:05 - INFO - codeparrot_training - Step 4340: {'lr': 0.0004970737600089673, 'samples': 833472, 'steps': 4340, 'loss/train': 4.003241658210754} 01/28/2022 00:18:09 - INFO - codeparrot_training - Step 4341: {'lr': 0.0004970712633105496, 'samples': 833664, 'steps': 4341, 'loss/train': 2.0600441694259644} 01/28/2022 00:18:14 - INFO - codeparrot_training - Step 4342: {'lr': 0.0004970687655537568, 'samples': 833856, 'steps': 4342, 'loss/train': 4.904103398323059} 01/28/2022 00:18:18 - INFO - codeparrot_training - Step 4343: {'lr': 0.0004970662667386, 'samples': 834048, 'steps': 4343, 'loss/train': 3.006128668785095} 01/28/2022 00:18:23 - INFO - codeparrot_training - Step 4344: {'lr': 0.0004970637668650898, 'samples': 834240, 'steps': 4344, 'loss/train': 3.082721471786499} 01/28/2022 00:18:27 - INFO - codeparrot_training - Step 4345: {'lr': 0.0004970612659332368, 'samples': 834432, 'steps': 4345, 'loss/train': 2.2825241088867188} 01/28/2022 00:18:32 - INFO - codeparrot_training - Step 4346: {'lr': 0.0004970587639430518, 'samples': 834624, 'steps': 4346, 'loss/train': 3.989416480064392} 01/28/2022 00:18:36 - INFO - codeparrot_training - Step 4347: {'lr': 0.0004970562608945455, 'samples': 834816, 'steps': 4347, 'loss/train': 3.04880690574646} 01/28/2022 00:18:40 - INFO - codeparrot_training - Step 4348: {'lr': 0.0004970537567877286, 'samples': 835008, 'steps': 4348, 'loss/train': 3.439999580383301} 01/28/2022 00:18:46 - INFO - codeparrot_training - Step 4349: {'lr': 0.000497051251622612, 'samples': 835200, 'steps': 4349, 'loss/train': 3.0468528270721436} 01/28/2022 00:18:50 - INFO - codeparrot_training - Step 4350: {'lr': 0.0004970487453992062, 'samples': 835392, 'steps': 4350, 'loss/train': 3.0175838470458984} 01/28/2022 00:18:54 - INFO - codeparrot_training - Step 4351: {'lr': 0.000497046238117522, 'samples': 835584, 'steps': 4351, 'loss/train': 1.953701376914978} 01/28/2022 00:18:58 - INFO - codeparrot_training - Step 4352: {'lr': 0.0004970437297775702, 'samples': 835776, 'steps': 4352, 'loss/train': 1.7679278254508972} 01/28/2022 00:19:02 - INFO - codeparrot_training - Step 4353: {'lr': 0.0004970412203793614, 'samples': 835968, 'steps': 4353, 'loss/train': 3.2708523273468018} 01/28/2022 00:19:09 - INFO - codeparrot_training - Step 4354: {'lr': 0.0004970387099229066, 'samples': 836160, 'steps': 4354, 'loss/train': 2.719970226287842} 01/28/2022 00:19:13 - INFO - codeparrot_training - Step 4355: {'lr': 0.0004970361984082163, 'samples': 836352, 'steps': 4355, 'loss/train': 2.8983871936798096} 01/28/2022 00:19:17 - INFO - codeparrot_training - Step 4356: {'lr': 0.0004970336858353014, 'samples': 836544, 'steps': 4356, 'loss/train': 3.2478185892105103} 01/28/2022 00:19:21 - INFO - codeparrot_training - Step 4357: {'lr': 0.0004970311722041727, 'samples': 836736, 'steps': 4357, 'loss/train': 3.1037063598632812} 01/28/2022 00:19:26 - INFO - codeparrot_training - Step 4358: {'lr': 0.0004970286575148408, 'samples': 836928, 'steps': 4358, 'loss/train': 1.0570135116577148} 01/28/2022 00:19:31 - INFO - codeparrot_training - Step 4359: {'lr': 0.0004970261417673165, 'samples': 837120, 'steps': 4359, 'loss/train': 3.7720370292663574} 01/28/2022 00:19:35 - INFO - codeparrot_training - Step 4360: {'lr': 0.0004970236249616109, 'samples': 837312, 'steps': 4360, 'loss/train': 3.1642438173294067} 01/28/2022 00:19:39 - INFO - codeparrot_training - Step 4361: {'lr': 0.0004970211070977344, 'samples': 837504, 'steps': 4361, 'loss/train': 2.5017060041427612} 01/28/2022 00:19:44 - INFO - codeparrot_training - Step 4362: {'lr': 0.0004970185881756979, 'samples': 837696, 'steps': 4362, 'loss/train': 3.4531112909317017} 01/28/2022 00:19:48 - INFO - codeparrot_training - Step 4363: {'lr': 0.0004970160681955121, 'samples': 837888, 'steps': 4363, 'loss/train': 1.844340205192566} 01/28/2022 00:19:53 - INFO - codeparrot_training - Step 4364: {'lr': 0.0004970135471571881, 'samples': 838080, 'steps': 4364, 'loss/train': 3.1773691177368164} 01/28/2022 00:19:57 - INFO - codeparrot_training - Step 4365: {'lr': 0.0004970110250607364, 'samples': 838272, 'steps': 4365, 'loss/train': 0.6358417421579361} 01/28/2022 00:20:01 - INFO - codeparrot_training - Step 4366: {'lr': 0.000497008501906168, 'samples': 838464, 'steps': 4366, 'loss/train': 0.9202391803264618} 01/28/2022 00:20:06 - INFO - codeparrot_training - Step 4367: {'lr': 0.0004970059776934935, 'samples': 838656, 'steps': 4367, 'loss/train': 3.173258900642395} 01/28/2022 00:20:10 - INFO - codeparrot_training - Step 4368: {'lr': 0.0004970034524227238, 'samples': 838848, 'steps': 4368, 'loss/train': 3.5410666465759277} 01/28/2022 00:20:16 - INFO - codeparrot_training - Step 4369: {'lr': 0.0004970009260938698, 'samples': 839040, 'steps': 4369, 'loss/train': 2.776840031147003} 01/28/2022 00:20:20 - INFO - codeparrot_training - Step 4370: {'lr': 0.0004969983987069423, 'samples': 839232, 'steps': 4370, 'loss/train': 2.6675861477851868} 01/28/2022 00:20:25 - INFO - codeparrot_training - Step 4371: {'lr': 0.000496995870261952, 'samples': 839424, 'steps': 4371, 'loss/train': 3.0977386236190796} 01/28/2022 00:20:29 - INFO - codeparrot_training - Step 4372: {'lr': 0.0004969933407589098, 'samples': 839616, 'steps': 4372, 'loss/train': 2.8517303466796875} 01/28/2022 00:20:33 - INFO - codeparrot_training - Step 4373: {'lr': 0.0004969908101978267, 'samples': 839808, 'steps': 4373, 'loss/train': 3.8851908445358276} 01/28/2022 00:20:38 - INFO - codeparrot_training - Step 4374: {'lr': 0.0004969882785787133, 'samples': 840000, 'steps': 4374, 'loss/train': 1.0632749497890472} 01/28/2022 00:20:42 - INFO - codeparrot_training - Step 4375: {'lr': 0.0004969857459015807, 'samples': 840192, 'steps': 4375, 'loss/train': 1.7034839987754822} 01/28/2022 00:20:47 - INFO - codeparrot_training - Step 4376: {'lr': 0.0004969832121664394, 'samples': 840384, 'steps': 4376, 'loss/train': 2.3641353249549866} 01/28/2022 00:20:51 - INFO - codeparrot_training - Step 4377: {'lr': 0.0004969806773733004, 'samples': 840576, 'steps': 4377, 'loss/train': 4.016724228858948} 01/28/2022 00:20:55 - INFO - codeparrot_training - Step 4378: {'lr': 0.0004969781415221748, 'samples': 840768, 'steps': 4378, 'loss/train': 2.5370800495147705} 01/28/2022 00:21:01 - INFO - codeparrot_training - Step 4379: {'lr': 0.0004969756046130731, 'samples': 840960, 'steps': 4379, 'loss/train': 2.120908498764038} 01/28/2022 00:21:05 - INFO - codeparrot_training - Step 4380: {'lr': 0.0004969730666460065, 'samples': 841152, 'steps': 4380, 'loss/train': 2.7608118653297424} 01/28/2022 00:21:09 - INFO - codeparrot_training - Step 4381: {'lr': 0.0004969705276209856, 'samples': 841344, 'steps': 4381, 'loss/train': 3.667608618736267} 01/28/2022 00:21:14 - INFO - codeparrot_training - Step 4382: {'lr': 0.0004969679875380214, 'samples': 841536, 'steps': 4382, 'loss/train': 4.999453783035278} 01/28/2022 00:21:18 - INFO - codeparrot_training - Step 4383: {'lr': 0.0004969654463971247, 'samples': 841728, 'steps': 4383, 'loss/train': 1.8988780975341797} 01/28/2022 00:21:23 - INFO - codeparrot_training - Step 4384: {'lr': 0.0004969629041983065, 'samples': 841920, 'steps': 4384, 'loss/train': 2.7994077801704407} 01/28/2022 00:21:27 - INFO - codeparrot_training - Step 4385: {'lr': 0.0004969603609415777, 'samples': 842112, 'steps': 4385, 'loss/train': 3.420207381248474} 01/28/2022 00:21:32 - INFO - codeparrot_training - Step 4386: {'lr': 0.000496957816626949, 'samples': 842304, 'steps': 4386, 'loss/train': 3.299986481666565} 01/28/2022 00:21:36 - INFO - codeparrot_training - Step 4387: {'lr': 0.0004969552712544316, 'samples': 842496, 'steps': 4387, 'loss/train': 4.706005811691284} 01/28/2022 00:21:40 - INFO - codeparrot_training - Step 4388: {'lr': 0.0004969527248240361, 'samples': 842688, 'steps': 4388, 'loss/train': 1.5546830892562866} 01/28/2022 00:21:45 - INFO - codeparrot_training - Step 4389: {'lr': 0.0004969501773357736, 'samples': 842880, 'steps': 4389, 'loss/train': 3.7787153720855713} 01/28/2022 00:21:49 - INFO - codeparrot_training - Step 4390: {'lr': 0.000496947628789655, 'samples': 843072, 'steps': 4390, 'loss/train': 2.2422369718551636} 01/28/2022 00:21:54 - INFO - codeparrot_training - Step 4391: {'lr': 0.000496945079185691, 'samples': 843264, 'steps': 4391, 'loss/train': 1.7463916540145874} 01/28/2022 00:21:58 - INFO - codeparrot_training - Step 4392: {'lr': 0.0004969425285238928, 'samples': 843456, 'steps': 4392, 'loss/train': 1.5459033250808716} 01/28/2022 00:22:02 - INFO - codeparrot_training - Step 4393: {'lr': 0.0004969399768042713, 'samples': 843648, 'steps': 4393, 'loss/train': 2.6793975234031677} 01/28/2022 00:22:07 - INFO - codeparrot_training - Step 4394: {'lr': 0.0004969374240268373, 'samples': 843840, 'steps': 4394, 'loss/train': 2.8467666506767273} 01/28/2022 00:22:12 - INFO - codeparrot_training - Step 4395: {'lr': 0.0004969348701916018, 'samples': 844032, 'steps': 4395, 'loss/train': 3.283393621444702} 01/28/2022 00:22:16 - INFO - codeparrot_training - Step 4396: {'lr': 0.0004969323152985756, 'samples': 844224, 'steps': 4396, 'loss/train': 4.447515249252319} 01/28/2022 00:22:20 - INFO - codeparrot_training - Step 4397: {'lr': 0.0004969297593477699, 'samples': 844416, 'steps': 4397, 'loss/train': 3.667998790740967} 01/28/2022 00:22:24 - INFO - codeparrot_training - Step 4398: {'lr': 0.0004969272023391955, 'samples': 844608, 'steps': 4398, 'loss/train': 2.8955896496772766} 01/28/2022 00:22:30 - INFO - codeparrot_training - Step 4399: {'lr': 0.0004969246442728633, 'samples': 844800, 'steps': 4399, 'loss/train': 2.481488049030304} 01/28/2022 00:22:34 - INFO - codeparrot_training - Step 4400: {'lr': 0.0004969220851487844, 'samples': 844992, 'steps': 4400, 'loss/train': 3.4542503356933594} 01/28/2022 00:22:39 - INFO - codeparrot_training - Step 4401: {'lr': 0.0004969195249669697, 'samples': 845184, 'steps': 4401, 'loss/train': 2.092713475227356} 01/28/2022 00:22:43 - INFO - codeparrot_training - Step 4402: {'lr': 0.0004969169637274301, 'samples': 845376, 'steps': 4402, 'loss/train': 3.0373460054397583} 01/28/2022 00:22:47 - INFO - codeparrot_training - Step 4403: {'lr': 0.0004969144014301767, 'samples': 845568, 'steps': 4403, 'loss/train': 3.6220346689224243} 01/28/2022 00:22:53 - INFO - codeparrot_training - Step 4404: {'lr': 0.0004969118380752205, 'samples': 845760, 'steps': 4404, 'loss/train': 3.609298825263977} 01/28/2022 00:22:57 - INFO - codeparrot_training - Step 4405: {'lr': 0.0004969092736625722, 'samples': 845952, 'steps': 4405, 'loss/train': 3.202691674232483} 01/28/2022 00:23:01 - INFO - codeparrot_training - Step 4406: {'lr': 0.000496906708192243, 'samples': 846144, 'steps': 4406, 'loss/train': 3.3266851902008057} 01/28/2022 00:23:05 - INFO - codeparrot_training - Step 4407: {'lr': 0.000496904141664244, 'samples': 846336, 'steps': 4407, 'loss/train': 2.4952762126922607} 01/28/2022 00:23:09 - INFO - codeparrot_training - Step 4408: {'lr': 0.0004969015740785859, 'samples': 846528, 'steps': 4408, 'loss/train': 3.1667908430099487} 01/28/2022 00:23:15 - INFO - codeparrot_training - Step 4409: {'lr': 0.00049689900543528, 'samples': 846720, 'steps': 4409, 'loss/train': 3.3118250370025635} 01/28/2022 00:23:19 - INFO - codeparrot_training - Step 4410: {'lr': 0.0004968964357343371, 'samples': 846912, 'steps': 4410, 'loss/train': 2.8793166875839233} 01/28/2022 00:23:23 - INFO - codeparrot_training - Step 4411: {'lr': 0.0004968938649757682, 'samples': 847104, 'steps': 4411, 'loss/train': 2.6932872533798218} 01/28/2022 00:23:27 - INFO - codeparrot_training - Step 4412: {'lr': 0.0004968912931595845, 'samples': 847296, 'steps': 4412, 'loss/train': 3.7984617948532104} 01/28/2022 00:23:32 - INFO - codeparrot_training - Step 4413: {'lr': 0.0004968887202857968, 'samples': 847488, 'steps': 4413, 'loss/train': 2.382528305053711} 01/28/2022 00:23:37 - INFO - codeparrot_training - Step 4414: {'lr': 0.0004968861463544163, 'samples': 847680, 'steps': 4414, 'loss/train': 2.537481129169464} 01/28/2022 00:23:42 - INFO - codeparrot_training - Step 4415: {'lr': 0.0004968835713654538, 'samples': 847872, 'steps': 4415, 'loss/train': 2.024576246738434} 01/28/2022 00:23:46 - INFO - codeparrot_training - Step 4416: {'lr': 0.0004968809953189206, 'samples': 848064, 'steps': 4416, 'loss/train': 3.0420849323272705} 01/28/2022 00:23:50 - INFO - codeparrot_training - Step 4417: {'lr': 0.0004968784182148276, 'samples': 848256, 'steps': 4417, 'loss/train': 3.1856117248535156} 01/28/2022 00:23:54 - INFO - codeparrot_training - Step 4418: {'lr': 0.0004968758400531859, 'samples': 848448, 'steps': 4418, 'loss/train': 3.872304081916809} 01/28/2022 00:24:00 - INFO - codeparrot_training - Step 4419: {'lr': 0.0004968732608340064, 'samples': 848640, 'steps': 4419, 'loss/train': 3.5335299968719482} 01/28/2022 00:24:04 - INFO - codeparrot_training - Step 4420: {'lr': 0.0004968706805573002, 'samples': 848832, 'steps': 4420, 'loss/train': 2.8929609060287476} 01/28/2022 00:24:08 - INFO - codeparrot_training - Step 4421: {'lr': 0.0004968680992230785, 'samples': 849024, 'steps': 4421, 'loss/train': 2.962102174758911} 01/28/2022 00:24:13 - INFO - codeparrot_training - Step 4422: {'lr': 0.0004968655168313522, 'samples': 849216, 'steps': 4422, 'loss/train': 2.1203237771987915} 01/28/2022 00:24:17 - INFO - codeparrot_training - Step 4423: {'lr': 0.0004968629333821324, 'samples': 849408, 'steps': 4423, 'loss/train': 2.996235966682434} 01/28/2022 00:24:23 - INFO - codeparrot_training - Step 4424: {'lr': 0.0004968603488754302, 'samples': 849600, 'steps': 4424, 'loss/train': 5.3904300928115845} 01/28/2022 00:24:27 - INFO - codeparrot_training - Step 4425: {'lr': 0.0004968577633112566, 'samples': 849792, 'steps': 4425, 'loss/train': 1.8159754872322083} 01/28/2022 00:24:31 - INFO - codeparrot_training - Step 4426: {'lr': 0.0004968551766896228, 'samples': 849984, 'steps': 4426, 'loss/train': 3.405406951904297} 01/28/2022 00:24:35 - INFO - codeparrot_training - Step 4427: {'lr': 0.0004968525890105399, 'samples': 850176, 'steps': 4427, 'loss/train': 3.617675542831421} 01/28/2022 00:24:39 - INFO - codeparrot_training - Step 4428: {'lr': 0.0004968500002740187, 'samples': 850368, 'steps': 4428, 'loss/train': 3.7065389156341553} 01/28/2022 00:24:45 - INFO - codeparrot_training - Step 4429: {'lr': 0.0004968474104800706, 'samples': 850560, 'steps': 4429, 'loss/train': 2.5796669125556946} 01/28/2022 00:24:49 - INFO - codeparrot_training - Step 4430: {'lr': 0.0004968448196287066, 'samples': 850752, 'steps': 4430, 'loss/train': 3.830504894256592} 01/28/2022 00:24:53 - INFO - codeparrot_training - Step 4431: {'lr': 0.0004968422277199377, 'samples': 850944, 'steps': 4431, 'loss/train': 3.0582200288772583} 01/28/2022 00:24:58 - INFO - codeparrot_training - Step 4432: {'lr': 0.000496839634753775, 'samples': 851136, 'steps': 4432, 'loss/train': 3.0485265254974365} 01/28/2022 00:25:02 - INFO - codeparrot_training - Step 4433: {'lr': 0.0004968370407302299, 'samples': 851328, 'steps': 4433, 'loss/train': 3.084082245826721} 01/28/2022 00:25:07 - INFO - codeparrot_training - Step 4434: {'lr': 0.0004968344456493132, 'samples': 851520, 'steps': 4434, 'loss/train': 3.1412347555160522} 01/28/2022 00:25:11 - INFO - codeparrot_training - Step 4435: {'lr': 0.000496831849511036, 'samples': 851712, 'steps': 4435, 'loss/train': 3.2554246187210083} 01/28/2022 00:25:15 - INFO - codeparrot_training - Step 4436: {'lr': 0.0004968292523154096, 'samples': 851904, 'steps': 4436, 'loss/train': 2.9008739590644836} 01/28/2022 00:25:20 - INFO - codeparrot_training - Step 4437: {'lr': 0.0004968266540624452, 'samples': 852096, 'steps': 4437, 'loss/train': 3.050648331642151} 01/28/2022 00:25:24 - INFO - codeparrot_training - Step 4438: {'lr': 0.0004968240547521536, 'samples': 852288, 'steps': 4438, 'loss/train': 3.420424461364746} 01/28/2022 00:25:30 - INFO - codeparrot_training - Step 4439: {'lr': 0.0004968214543845463, 'samples': 852480, 'steps': 4439, 'loss/train': 2.799112021923065} 01/28/2022 00:25:34 - INFO - codeparrot_training - Step 4440: {'lr': 0.0004968188529596341, 'samples': 852672, 'steps': 4440, 'loss/train': 2.78484570980072} 01/28/2022 00:25:39 - INFO - codeparrot_training - Step 4441: {'lr': 0.0004968162504774284, 'samples': 852864, 'steps': 4441, 'loss/train': 2.807486593723297} 01/28/2022 00:25:43 - INFO - codeparrot_training - Step 4442: {'lr': 0.0004968136469379403, 'samples': 853056, 'steps': 4442, 'loss/train': 3.427432894706726} 01/28/2022 00:25:48 - INFO - codeparrot_training - Step 4443: {'lr': 0.0004968110423411808, 'samples': 853248, 'steps': 4443, 'loss/train': 3.1980947256088257} 01/28/2022 00:25:53 - INFO - codeparrot_training - Step 4444: {'lr': 0.0004968084366871612, 'samples': 853440, 'steps': 4444, 'loss/train': 2.4348506927490234} 01/28/2022 00:25:57 - INFO - codeparrot_training - Step 4445: {'lr': 0.0004968058299758926, 'samples': 853632, 'steps': 4445, 'loss/train': 3.5551235675811768} 01/28/2022 00:26:01 - INFO - codeparrot_training - Step 4446: {'lr': 0.0004968032222073863, 'samples': 853824, 'steps': 4446, 'loss/train': 3.8093533515930176} 01/28/2022 00:26:05 - INFO - codeparrot_training - Step 4447: {'lr': 0.0004968006133816532, 'samples': 854016, 'steps': 4447, 'loss/train': 2.8720831274986267} 01/28/2022 00:26:10 - INFO - codeparrot_training - Step 4448: {'lr': 0.0004967980034987048, 'samples': 854208, 'steps': 4448, 'loss/train': 2.7425436973571777} 01/28/2022 00:26:15 - INFO - codeparrot_training - Step 4449: {'lr': 0.0004967953925585521, 'samples': 854400, 'steps': 4449, 'loss/train': 1.7880620956420898} 01/28/2022 00:26:19 - INFO - codeparrot_training - Step 4450: {'lr': 0.0004967927805612063, 'samples': 854592, 'steps': 4450, 'loss/train': 2.869908034801483} 01/28/2022 00:26:23 - INFO - codeparrot_training - Step 4451: {'lr': 0.0004967901675066784, 'samples': 854784, 'steps': 4451, 'loss/train': 4.086017847061157} 01/28/2022 00:26:27 - INFO - codeparrot_training - Step 4452: {'lr': 0.0004967875533949801, 'samples': 854976, 'steps': 4452, 'loss/train': 2.592662036418915} 01/28/2022 00:26:33 - INFO - codeparrot_training - Step 4453: {'lr': 0.000496784938226122, 'samples': 855168, 'steps': 4453, 'loss/train': 2.9895989298820496} 01/28/2022 00:26:37 - INFO - codeparrot_training - Step 4454: {'lr': 0.0004967823220001158, 'samples': 855360, 'steps': 4454, 'loss/train': 2.4731701612472534} 01/28/2022 00:26:41 - INFO - codeparrot_training - Step 4455: {'lr': 0.0004967797047169724, 'samples': 855552, 'steps': 4455, 'loss/train': 3.4079235792160034} 01/28/2022 00:26:45 - INFO - codeparrot_training - Step 4456: {'lr': 0.0004967770863767031, 'samples': 855744, 'steps': 4456, 'loss/train': 3.224366784095764} 01/28/2022 00:26:50 - INFO - codeparrot_training - Step 4457: {'lr': 0.0004967744669793192, 'samples': 855936, 'steps': 4457, 'loss/train': 1.3957656919956207} 01/28/2022 00:26:56 - INFO - codeparrot_training - Step 4458: {'lr': 0.0004967718465248317, 'samples': 856128, 'steps': 4458, 'loss/train': 3.10901141166687} 01/28/2022 00:27:00 - INFO - codeparrot_training - Step 4459: {'lr': 0.000496769225013252, 'samples': 856320, 'steps': 4459, 'loss/train': 3.4071922302246094} 01/28/2022 00:27:04 - INFO - codeparrot_training - Step 4460: {'lr': 0.0004967666024445913, 'samples': 856512, 'steps': 4460, 'loss/train': 3.581870913505554} 01/28/2022 00:27:08 - INFO - codeparrot_training - Step 4461: {'lr': 0.000496763978818861, 'samples': 856704, 'steps': 4461, 'loss/train': 2.7650097012519836} 01/28/2022 00:27:12 - INFO - codeparrot_training - Step 4462: {'lr': 0.000496761354136072, 'samples': 856896, 'steps': 4462, 'loss/train': 3.547702431678772} 01/28/2022 00:27:18 - INFO - codeparrot_training - Step 4463: {'lr': 0.0004967587283962358, 'samples': 857088, 'steps': 4463, 'loss/train': 3.1626534461975098} 01/28/2022 00:27:22 - INFO - codeparrot_training - Step 4464: {'lr': 0.0004967561015993635, 'samples': 857280, 'steps': 4464, 'loss/train': 1.6259520649909973} 01/28/2022 00:27:26 - INFO - codeparrot_training - Step 4465: {'lr': 0.0004967534737454665, 'samples': 857472, 'steps': 4465, 'loss/train': 2.236378848552704} 01/28/2022 00:27:30 - INFO - codeparrot_training - Step 4466: {'lr': 0.000496750844834556, 'samples': 857664, 'steps': 4466, 'loss/train': 3.860809564590454} 01/28/2022 00:27:35 - INFO - codeparrot_training - Step 4467: {'lr': 0.000496748214866643, 'samples': 857856, 'steps': 4467, 'loss/train': 3.1254948377609253} 01/28/2022 00:27:40 - INFO - codeparrot_training - Step 4468: {'lr': 0.0004967455838417392, 'samples': 858048, 'steps': 4468, 'loss/train': 3.436797022819519} 01/28/2022 00:27:45 - INFO - codeparrot_training - Step 4469: {'lr': 0.0004967429517598556, 'samples': 858240, 'steps': 4469, 'loss/train': 3.5515480041503906} 01/28/2022 00:27:49 - INFO - codeparrot_training - Step 4470: {'lr': 0.0004967403186210036, 'samples': 858432, 'steps': 4470, 'loss/train': 2.744032859802246} 01/28/2022 00:27:53 - INFO - codeparrot_training - Step 4471: {'lr': 0.0004967376844251944, 'samples': 858624, 'steps': 4471, 'loss/train': 3.3202950954437256} 01/28/2022 00:27:57 - INFO - codeparrot_training - Step 4472: {'lr': 0.0004967350491724392, 'samples': 858816, 'steps': 4472, 'loss/train': 2.639150619506836} 01/28/2022 00:28:03 - INFO - codeparrot_training - Step 4473: {'lr': 0.0004967324128627495, 'samples': 859008, 'steps': 4473, 'loss/train': 2.6346073150634766} 01/28/2022 00:28:07 - INFO - codeparrot_training - Step 4474: {'lr': 0.0004967297754961365, 'samples': 859200, 'steps': 4474, 'loss/train': 3.995803713798523} 01/28/2022 00:28:11 - INFO - codeparrot_training - Step 4475: {'lr': 0.0004967271370726115, 'samples': 859392, 'steps': 4475, 'loss/train': 3.184984803199768} 01/28/2022 00:28:15 - INFO - codeparrot_training - Step 4476: {'lr': 0.0004967244975921857, 'samples': 859584, 'steps': 4476, 'loss/train': 2.9923505187034607} 01/28/2022 00:28:20 - INFO - codeparrot_training - Step 4477: {'lr': 0.0004967218570548706, 'samples': 859776, 'steps': 4477, 'loss/train': 2.8874621987342834} 01/28/2022 00:28:25 - INFO - codeparrot_training - Step 4478: {'lr': 0.0004967192154606774, 'samples': 859968, 'steps': 4478, 'loss/train': 3.561912775039673} 01/28/2022 00:28:29 - INFO - codeparrot_training - Step 4479: {'lr': 0.0004967165728096172, 'samples': 860160, 'steps': 4479, 'loss/train': 3.3828649520874023} 01/28/2022 00:28:33 - INFO - codeparrot_training - Step 4480: {'lr': 0.0004967139291017018, 'samples': 860352, 'steps': 4480, 'loss/train': 3.249027371406555} 01/28/2022 00:28:37 - INFO - codeparrot_training - Step 4481: {'lr': 0.0004967112843369423, 'samples': 860544, 'steps': 4481, 'loss/train': 3.54383647441864} 01/28/2022 00:28:42 - INFO - codeparrot_training - Step 4482: {'lr': 0.0004967086385153499, 'samples': 860736, 'steps': 4482, 'loss/train': 3.0651161670684814} 01/28/2022 00:28:48 - INFO - codeparrot_training - Step 4483: {'lr': 0.0004967059916369359, 'samples': 860928, 'steps': 4483, 'loss/train': 3.3770681619644165} 01/28/2022 00:28:52 - INFO - codeparrot_training - Step 4484: {'lr': 0.000496703343701712, 'samples': 861120, 'steps': 4484, 'loss/train': 2.3666887879371643} 01/28/2022 00:28:56 - INFO - codeparrot_training - Step 4485: {'lr': 0.0004967006947096892, 'samples': 861312, 'steps': 4485, 'loss/train': 3.060961604118347} 01/28/2022 00:29:00 - INFO - codeparrot_training - Step 4486: {'lr': 0.0004966980446608789, 'samples': 861504, 'steps': 4486, 'loss/train': 2.436271905899048} 01/28/2022 00:29:04 - INFO - codeparrot_training - Step 4487: {'lr': 0.0004966953935552925, 'samples': 861696, 'steps': 4487, 'loss/train': 2.1877350211143494} 01/28/2022 00:29:10 - INFO - codeparrot_training - Step 4488: {'lr': 0.0004966927413929415, 'samples': 861888, 'steps': 4488, 'loss/train': 3.8796579837799072} 01/28/2022 00:29:14 - INFO - codeparrot_training - Step 4489: {'lr': 0.0004966900881738371, 'samples': 862080, 'steps': 4489, 'loss/train': 2.4997318983078003} 01/28/2022 00:29:18 - INFO - codeparrot_training - Step 4490: {'lr': 0.0004966874338979907, 'samples': 862272, 'steps': 4490, 'loss/train': 2.1234260201454163} 01/28/2022 00:29:23 - INFO - codeparrot_training - Step 4491: {'lr': 0.0004966847785654136, 'samples': 862464, 'steps': 4491, 'loss/train': 3.0043888092041016} 01/28/2022 00:29:27 - INFO - codeparrot_training - Step 4492: {'lr': 0.0004966821221761173, 'samples': 862656, 'steps': 4492, 'loss/train': 1.6970166563987732} 01/28/2022 00:29:32 - INFO - codeparrot_training - Step 4493: {'lr': 0.0004966794647301131, 'samples': 862848, 'steps': 4493, 'loss/train': 3.5361067056655884} 01/28/2022 00:29:37 - INFO - codeparrot_training - Step 4494: {'lr': 0.0004966768062274125, 'samples': 863040, 'steps': 4494, 'loss/train': 2.940981924533844} 01/28/2022 00:29:41 - INFO - codeparrot_training - Step 4495: {'lr': 0.0004966741466680266, 'samples': 863232, 'steps': 4495, 'loss/train': 3.1167036294937134} 01/28/2022 00:29:45 - INFO - codeparrot_training - Step 4496: {'lr': 0.000496671486051967, 'samples': 863424, 'steps': 4496, 'loss/train': 2.8949203491210938} 01/28/2022 00:29:49 - INFO - codeparrot_training - Step 4497: {'lr': 0.0004966688243792452, 'samples': 863616, 'steps': 4497, 'loss/train': 1.4750939905643463} 01/28/2022 00:29:55 - INFO - codeparrot_training - Step 4498: {'lr': 0.0004966661616498724, 'samples': 863808, 'steps': 4498, 'loss/train': 3.064902663230896} 01/28/2022 00:29:59 - INFO - codeparrot_training - Step 4499: {'lr': 0.0004966634978638601, 'samples': 864000, 'steps': 4499, 'loss/train': 3.4542964696884155} 01/28/2022 00:30:04 - INFO - codeparrot_training - Step 4500: {'lr': 0.0004966608330212198, 'samples': 864192, 'steps': 4500, 'loss/train': 2.9656519889831543} 01/28/2022 00:30:08 - INFO - codeparrot_training - Step 4501: {'lr': 0.0004966581671219627, 'samples': 864384, 'steps': 4501, 'loss/train': 3.0733970403671265} 01/28/2022 00:30:12 - INFO - codeparrot_training - Step 4502: {'lr': 0.0004966555001661004, 'samples': 864576, 'steps': 4502, 'loss/train': 3.8482457399368286} 01/28/2022 00:30:17 - INFO - codeparrot_training - Step 4503: {'lr': 0.0004966528321536442, 'samples': 864768, 'steps': 4503, 'loss/train': 4.074586629867554} 01/28/2022 00:30:21 - INFO - codeparrot_training - Step 4504: {'lr': 0.0004966501630846057, 'samples': 864960, 'steps': 4504, 'loss/train': 3.8015635013580322} 01/28/2022 00:30:26 - INFO - codeparrot_training - Step 4505: {'lr': 0.000496647492958996, 'samples': 865152, 'steps': 4505, 'loss/train': 3.1058045625686646} 01/28/2022 00:30:30 - INFO - codeparrot_training - Step 4506: {'lr': 0.000496644821776827, 'samples': 865344, 'steps': 4506, 'loss/train': 2.9111679196357727} 01/28/2022 00:30:34 - INFO - codeparrot_training - Step 4507: {'lr': 0.0004966421495381098, 'samples': 865536, 'steps': 4507, 'loss/train': 3.1738622188568115} 01/28/2022 00:30:40 - INFO - codeparrot_training - Step 4508: {'lr': 0.0004966394762428559, 'samples': 865728, 'steps': 4508, 'loss/train': 1.508799970149994} 01/28/2022 00:30:44 - INFO - codeparrot_training - Step 4509: {'lr': 0.0004966368018910768, 'samples': 865920, 'steps': 4509, 'loss/train': 2.7724714279174805} 01/28/2022 00:30:48 - INFO - codeparrot_training - Step 4510: {'lr': 0.000496634126482784, 'samples': 866112, 'steps': 4510, 'loss/train': 1.8349390625953674} 01/28/2022 00:30:53 - INFO - codeparrot_training - Step 4511: {'lr': 0.000496631450017989, 'samples': 866304, 'steps': 4511, 'loss/train': 2.3126657009124756} 01/28/2022 00:30:57 - INFO - codeparrot_training - Step 4512: {'lr': 0.0004966287724967032, 'samples': 866496, 'steps': 4512, 'loss/train': 3.5421173572540283} 01/28/2022 00:31:02 - INFO - codeparrot_training - Step 4513: {'lr': 0.0004966260939189379, 'samples': 866688, 'steps': 4513, 'loss/train': 1.697727084159851} 01/28/2022 00:31:06 - INFO - codeparrot_training - Step 4514: {'lr': 0.0004966234142847048, 'samples': 866880, 'steps': 4514, 'loss/train': 2.5276517271995544} 01/28/2022 00:31:11 - INFO - codeparrot_training - Step 4515: {'lr': 0.0004966207335940153, 'samples': 867072, 'steps': 4515, 'loss/train': 3.133862257003784} 01/28/2022 00:31:15 - INFO - codeparrot_training - Step 4516: {'lr': 0.0004966180518468808, 'samples': 867264, 'steps': 4516, 'loss/train': 2.8779208660125732} 01/28/2022 00:31:21 - INFO - codeparrot_training - Step 4517: {'lr': 0.000496615369043313, 'samples': 867456, 'steps': 4517, 'loss/train': 3.603502035140991} 01/28/2022 00:31:25 - INFO - codeparrot_training - Step 4518: {'lr': 0.0004966126851833233, 'samples': 867648, 'steps': 4518, 'loss/train': 2.9627750515937805} 01/28/2022 00:31:30 - INFO - codeparrot_training - Step 4519: {'lr': 0.0004966100002669231, 'samples': 867840, 'steps': 4519, 'loss/train': 3.2161202430725098} 01/28/2022 00:31:34 - INFO - codeparrot_training - Step 4520: {'lr': 0.0004966073142941239, 'samples': 868032, 'steps': 4520, 'loss/train': 2.8958072662353516} 01/28/2022 00:31:38 - INFO - codeparrot_training - Step 4521: {'lr': 0.0004966046272649372, 'samples': 868224, 'steps': 4521, 'loss/train': 3.540958285331726} 01/28/2022 00:31:43 - INFO - codeparrot_training - Step 4522: {'lr': 0.0004966019391793748, 'samples': 868416, 'steps': 4522, 'loss/train': 3.1759207248687744} 01/28/2022 00:31:48 - INFO - codeparrot_training - Step 4523: {'lr': 0.0004965992500374479, 'samples': 868608, 'steps': 4523, 'loss/train': 3.459875464439392} 01/28/2022 00:31:52 - INFO - codeparrot_training - Step 4524: {'lr': 0.0004965965598391682, 'samples': 868800, 'steps': 4524, 'loss/train': 3.513285756111145} 01/28/2022 00:31:56 - INFO - codeparrot_training - Step 4525: {'lr': 0.000496593868584547, 'samples': 868992, 'steps': 4525, 'loss/train': 3.203561782836914} 01/28/2022 00:32:00 - INFO - codeparrot_training - Step 4526: {'lr': 0.0004965911762735961, 'samples': 869184, 'steps': 4526, 'loss/train': 2.8429877758026123} 01/28/2022 00:32:06 - INFO - codeparrot_training - Step 4527: {'lr': 0.0004965884829063268, 'samples': 869376, 'steps': 4527, 'loss/train': 2.686570465564728} 01/28/2022 00:32:11 - INFO - codeparrot_training - Step 4528: {'lr': 0.0004965857884827508, 'samples': 869568, 'steps': 4528, 'loss/train': 5.264647722244263} 01/28/2022 00:32:15 - INFO - codeparrot_training - Step 4529: {'lr': 0.0004965830930028795, 'samples': 869760, 'steps': 4529, 'loss/train': 6.035263538360596} 01/28/2022 00:32:19 - INFO - codeparrot_training - Step 4530: {'lr': 0.0004965803964667246, 'samples': 869952, 'steps': 4530, 'loss/train': 4.193016171455383} 01/28/2022 00:32:23 - INFO - codeparrot_training - Step 4531: {'lr': 0.0004965776988742976, 'samples': 870144, 'steps': 4531, 'loss/train': 3.70530366897583} 01/28/2022 00:32:29 - INFO - codeparrot_training - Step 4532: {'lr': 0.00049657500022561, 'samples': 870336, 'steps': 4532, 'loss/train': 3.341631531715393} 01/28/2022 00:32:33 - INFO - codeparrot_training - Step 4533: {'lr': 0.0004965723005206734, 'samples': 870528, 'steps': 4533, 'loss/train': 3.6197551488876343} 01/28/2022 00:32:37 - INFO - codeparrot_training - Step 4534: {'lr': 0.0004965695997594993, 'samples': 870720, 'steps': 4534, 'loss/train': 2.2640844583511353} 01/28/2022 00:32:41 - INFO - codeparrot_training - Step 4535: {'lr': 0.0004965668979420994, 'samples': 870912, 'steps': 4535, 'loss/train': 1.7499632835388184} 01/28/2022 00:32:46 - INFO - codeparrot_training - Step 4536: {'lr': 0.0004965641950684852, 'samples': 871104, 'steps': 4536, 'loss/train': 2.561503529548645} 01/28/2022 00:32:51 - INFO - codeparrot_training - Step 4537: {'lr': 0.0004965614911386683, 'samples': 871296, 'steps': 4537, 'loss/train': 3.0232200622558594} 01/28/2022 00:32:55 - INFO - codeparrot_training - Step 4538: {'lr': 0.0004965587861526602, 'samples': 871488, 'steps': 4538, 'loss/train': 1.021309643983841} 01/28/2022 00:32:59 - INFO - codeparrot_training - Step 4539: {'lr': 0.0004965560801104726, 'samples': 871680, 'steps': 4539, 'loss/train': 3.422972559928894} 01/28/2022 00:33:04 - INFO - codeparrot_training - Step 4540: {'lr': 0.000496553373012117, 'samples': 871872, 'steps': 4540, 'loss/train': 3.2339816093444824} 01/28/2022 00:33:08 - INFO - codeparrot_training - Step 4541: {'lr': 0.0004965506648576052, 'samples': 872064, 'steps': 4541, 'loss/train': 3.145135760307312} 01/28/2022 00:33:14 - INFO - codeparrot_training - Step 4542: {'lr': 0.0004965479556469485, 'samples': 872256, 'steps': 4542, 'loss/train': 3.359513998031616} 01/28/2022 00:33:18 - INFO - codeparrot_training - Step 4543: {'lr': 0.0004965452453801586, 'samples': 872448, 'steps': 4543, 'loss/train': 3.357060670852661} 01/28/2022 00:33:22 - INFO - codeparrot_training - Step 4544: {'lr': 0.0004965425340572472, 'samples': 872640, 'steps': 4544, 'loss/train': 2.1925529837608337} 01/28/2022 00:33:27 - INFO - codeparrot_training - Step 4545: {'lr': 0.0004965398216782258, 'samples': 872832, 'steps': 4545, 'loss/train': 10.271900653839111} 01/28/2022 00:33:31 - INFO - codeparrot_training - Step 4546: {'lr': 0.0004965371082431062, 'samples': 873024, 'steps': 4546, 'loss/train': 2.161949336528778} 01/28/2022 00:33:36 - INFO - codeparrot_training - Step 4547: {'lr': 0.0004965343937519, 'samples': 873216, 'steps': 4547, 'loss/train': 2.3420029878616333} 01/28/2022 00:33:40 - INFO - codeparrot_training - Step 4548: {'lr': 0.0004965316782046186, 'samples': 873408, 'steps': 4548, 'loss/train': 2.48169082403183} 01/28/2022 00:33:44 - INFO - codeparrot_training - Step 4549: {'lr': 0.0004965289616012739, 'samples': 873600, 'steps': 4549, 'loss/train': 2.7658352851867676} 01/28/2022 00:33:49 - INFO - codeparrot_training - Step 4550: {'lr': 0.0004965262439418772, 'samples': 873792, 'steps': 4550, 'loss/train': 3.3539743423461914} 01/28/2022 00:33:53 - INFO - codeparrot_training - Step 4551: {'lr': 0.0004965235252264405, 'samples': 873984, 'steps': 4551, 'loss/train': 2.3854583501815796} 01/28/2022 00:33:59 - INFO - codeparrot_training - Step 4552: {'lr': 0.0004965208054549753, 'samples': 874176, 'steps': 4552, 'loss/train': 2.5868242979049683} 01/28/2022 00:34:03 - INFO - codeparrot_training - Step 4553: {'lr': 0.0004965180846274931, 'samples': 874368, 'steps': 4553, 'loss/train': 3.1928189992904663} 01/28/2022 00:34:07 - INFO - codeparrot_training - Step 4554: {'lr': 0.0004965153627440058, 'samples': 874560, 'steps': 4554, 'loss/train': 2.6197094321250916} 01/28/2022 00:34:12 - INFO - codeparrot_training - Step 4555: {'lr': 0.000496512639804525, 'samples': 874752, 'steps': 4555, 'loss/train': 3.7905285358428955} 01/28/2022 00:34:16 - INFO - codeparrot_training - Step 4556: {'lr': 0.0004965099158090624, 'samples': 874944, 'steps': 4556, 'loss/train': 3.0258965492248535} 01/28/2022 00:34:21 - INFO - codeparrot_training - Step 4557: {'lr': 0.0004965071907576294, 'samples': 875136, 'steps': 4557, 'loss/train': 3.735707402229309} 01/28/2022 00:34:25 - INFO - codeparrot_training - Step 4558: {'lr': 0.000496504464650238, 'samples': 875328, 'steps': 4558, 'loss/train': 1.9800530076026917} 01/28/2022 00:34:30 - INFO - codeparrot_training - Step 4559: {'lr': 0.0004965017374868997, 'samples': 875520, 'steps': 4559, 'loss/train': 3.1129285097122192} 01/28/2022 00:34:34 - INFO - codeparrot_training - Step 4560: {'lr': 0.0004964990092676262, 'samples': 875712, 'steps': 4560, 'loss/train': 3.4200457334518433} 01/28/2022 00:34:38 - INFO - codeparrot_training - Step 4561: {'lr': 0.0004964962799924293, 'samples': 875904, 'steps': 4561, 'loss/train': 2.9622761607170105} 01/28/2022 00:34:43 - INFO - codeparrot_training - Step 4562: {'lr': 0.0004964935496613206, 'samples': 876096, 'steps': 4562, 'loss/train': 3.154034972190857} 01/28/2022 00:34:47 - INFO - codeparrot_training - Step 4563: {'lr': 0.0004964908182743117, 'samples': 876288, 'steps': 4563, 'loss/train': 2.301333010196686} 01/28/2022 00:34:52 - INFO - codeparrot_training - Step 4564: {'lr': 0.0004964880858314146, 'samples': 876480, 'steps': 4564, 'loss/train': 2.3071160316467285} 01/28/2022 00:34:56 - INFO - codeparrot_training - Step 4565: {'lr': 0.0004964853523326406, 'samples': 876672, 'steps': 4565, 'loss/train': 2.0809648633003235} 01/28/2022 00:35:00 - INFO - codeparrot_training - Step 4566: {'lr': 0.0004964826177780017, 'samples': 876864, 'steps': 4566, 'loss/train': 2.7504851818084717} 01/28/2022 00:35:05 - INFO - codeparrot_training - Step 4567: {'lr': 0.0004964798821675096, 'samples': 877056, 'steps': 4567, 'loss/train': 2.7331674098968506} 01/28/2022 00:35:10 - INFO - codeparrot_training - Step 4568: {'lr': 0.0004964771455011758, 'samples': 877248, 'steps': 4568, 'loss/train': 4.538552284240723} 01/28/2022 00:35:14 - INFO - codeparrot_training - Step 4569: {'lr': 0.0004964744077790123, 'samples': 877440, 'steps': 4569, 'loss/train': 3.1020952463150024} 01/28/2022 00:35:18 - INFO - codeparrot_training - Step 4570: {'lr': 0.0004964716690010306, 'samples': 877632, 'steps': 4570, 'loss/train': 2.590305805206299} 01/28/2022 00:35:22 - INFO - codeparrot_training - Step 4571: {'lr': 0.0004964689291672427, 'samples': 877824, 'steps': 4571, 'loss/train': 3.2090585231781006} 01/28/2022 00:35:28 - INFO - codeparrot_training - Step 4572: {'lr': 0.00049646618827766, 'samples': 878016, 'steps': 4572, 'loss/train': 3.909165143966675} 01/28/2022 00:35:32 - INFO - codeparrot_training - Step 4573: {'lr': 0.0004964634463322945, 'samples': 878208, 'steps': 4573, 'loss/train': 3.6268354654312134} 01/28/2022 00:35:36 - INFO - codeparrot_training - Step 4574: {'lr': 0.0004964607033311579, 'samples': 878400, 'steps': 4574, 'loss/train': 3.3308991193771362} 01/28/2022 00:35:41 - INFO - codeparrot_training - Step 4575: {'lr': 0.0004964579592742618, 'samples': 878592, 'steps': 4575, 'loss/train': 3.6826064586639404} 01/28/2022 00:35:45 - INFO - codeparrot_training - Step 4576: {'lr': 0.000496455214161618, 'samples': 878784, 'steps': 4576, 'loss/train': 2.270869195461273} 01/28/2022 00:35:50 - INFO - codeparrot_training - Step 4577: {'lr': 0.0004964524679932385, 'samples': 878976, 'steps': 4577, 'loss/train': 4.056484222412109} 01/28/2022 00:35:54 - INFO - codeparrot_training - Step 4578: {'lr': 0.0004964497207691349, 'samples': 879168, 'steps': 4578, 'loss/train': 3.3295018672943115} 01/28/2022 00:35:59 - INFO - codeparrot_training - Step 4579: {'lr': 0.0004964469724893188, 'samples': 879360, 'steps': 4579, 'loss/train': 3.1341530084609985} 01/28/2022 00:36:03 - INFO - codeparrot_training - Step 4580: {'lr': 0.0004964442231538023, 'samples': 879552, 'steps': 4580, 'loss/train': 1.3783425986766815} 01/28/2022 00:36:07 - INFO - codeparrot_training - Step 4581: {'lr': 0.0004964414727625968, 'samples': 879744, 'steps': 4581, 'loss/train': 3.7845908403396606} 01/28/2022 00:36:12 - INFO - codeparrot_training - Step 4582: {'lr': 0.0004964387213157143, 'samples': 879936, 'steps': 4582, 'loss/train': 2.8287715315818787} 01/28/2022 00:36:16 - INFO - codeparrot_training - Step 4583: {'lr': 0.0004964359688131667, 'samples': 880128, 'steps': 4583, 'loss/train': 2.3871195316314697} 01/28/2022 00:36:21 - INFO - codeparrot_training - Step 4584: {'lr': 0.0004964332152549657, 'samples': 880320, 'steps': 4584, 'loss/train': 3.824713110923767} 01/28/2022 00:36:25 - INFO - codeparrot_training - Step 4585: {'lr': 0.0004964304606411229, 'samples': 880512, 'steps': 4585, 'loss/train': 2.474182426929474} 01/28/2022 00:36:29 - INFO - codeparrot_training - Step 4586: {'lr': 0.0004964277049716503, 'samples': 880704, 'steps': 4586, 'loss/train': 3.6881006956100464} 01/28/2022 00:36:35 - INFO - codeparrot_training - Step 4587: {'lr': 0.0004964249482465597, 'samples': 880896, 'steps': 4587, 'loss/train': 2.9167460203170776} 01/28/2022 00:36:39 - INFO - codeparrot_training - Step 4588: {'lr': 0.0004964221904658629, 'samples': 881088, 'steps': 4588, 'loss/train': 3.148354411125183} 01/28/2022 00:36:43 - INFO - codeparrot_training - Step 4589: {'lr': 0.0004964194316295716, 'samples': 881280, 'steps': 4589, 'loss/train': 3.6843910217285156} 01/28/2022 00:36:48 - INFO - codeparrot_training - Step 4590: {'lr': 0.0004964166717376978, 'samples': 881472, 'steps': 4590, 'loss/train': 3.1427797079086304} 01/28/2022 00:36:52 - INFO - codeparrot_training - Step 4591: {'lr': 0.0004964139107902531, 'samples': 881664, 'steps': 4591, 'loss/train': 2.4412845969200134} 01/28/2022 00:36:57 - INFO - codeparrot_training - Step 4592: {'lr': 0.0004964111487872495, 'samples': 881856, 'steps': 4592, 'loss/train': 1.9604945182800293} 01/28/2022 00:37:01 - INFO - codeparrot_training - Step 4593: {'lr': 0.0004964083857286988, 'samples': 882048, 'steps': 4593, 'loss/train': 3.36323082447052} 01/28/2022 00:37:06 - INFO - codeparrot_training - Step 4594: {'lr': 0.0004964056216146129, 'samples': 882240, 'steps': 4594, 'loss/train': 2.2433568835258484} 01/28/2022 00:37:10 - INFO - codeparrot_training - Step 4595: {'lr': 0.0004964028564450034, 'samples': 882432, 'steps': 4595, 'loss/train': 1.9771009683609009} 01/28/2022 00:37:14 - INFO - codeparrot_training - Step 4596: {'lr': 0.0004964000902198824, 'samples': 882624, 'steps': 4596, 'loss/train': 3.673049211502075} 01/28/2022 00:37:20 - INFO - codeparrot_training - Step 4597: {'lr': 0.0004963973229392617, 'samples': 882816, 'steps': 4597, 'loss/train': 1.2041813135147095} 01/28/2022 00:37:25 - INFO - codeparrot_training - Step 4598: {'lr': 0.0004963945546031529, 'samples': 883008, 'steps': 4598, 'loss/train': 3.46042799949646} 01/28/2022 00:37:29 - INFO - codeparrot_training - Step 4599: {'lr': 0.0004963917852115683, 'samples': 883200, 'steps': 4599, 'loss/train': 1.6854155659675598} 01/28/2022 00:37:33 - INFO - codeparrot_training - Step 4600: {'lr': 0.0004963890147645194, 'samples': 883392, 'steps': 4600, 'loss/train': 1.405213326215744} 01/28/2022 00:37:37 - INFO - codeparrot_training - Step 4601: {'lr': 0.0004963862432620183, 'samples': 883584, 'steps': 4601, 'loss/train': 2.3152756690979004} 01/28/2022 00:37:43 - INFO - codeparrot_training - Step 4602: {'lr': 0.0004963834707040767, 'samples': 883776, 'steps': 4602, 'loss/train': 2.5340988636016846} 01/28/2022 00:37:47 - INFO - codeparrot_training - Step 4603: {'lr': 0.0004963806970907066, 'samples': 883968, 'steps': 4603, 'loss/train': 2.6156498193740845} 01/28/2022 00:37:51 - INFO - codeparrot_training - Step 4604: {'lr': 0.0004963779224219197, 'samples': 884160, 'steps': 4604, 'loss/train': 3.4204022884368896} 01/28/2022 00:37:55 - INFO - codeparrot_training - Step 4605: {'lr': 0.0004963751466977281, 'samples': 884352, 'steps': 4605, 'loss/train': 2.9553987979888916} 01/28/2022 00:38:00 - INFO - codeparrot_training - Step 4606: {'lr': 0.0004963723699181437, 'samples': 884544, 'steps': 4606, 'loss/train': 2.6215102672576904} 01/28/2022 00:38:06 - INFO - codeparrot_training - Step 4607: {'lr': 0.0004963695920831781, 'samples': 884736, 'steps': 4607, 'loss/train': 1.555953025817871} 01/28/2022 00:38:10 - INFO - codeparrot_training - Step 4608: {'lr': 0.0004963668131928436, 'samples': 884928, 'steps': 4608, 'loss/train': 2.787003993988037} 01/28/2022 00:38:14 - INFO - codeparrot_training - Step 4609: {'lr': 0.0004963640332471518, 'samples': 885120, 'steps': 4609, 'loss/train': 3.5804715156555176} 01/28/2022 00:38:18 - INFO - codeparrot_training - Step 4610: {'lr': 0.0004963612522461147, 'samples': 885312, 'steps': 4610, 'loss/train': 1.799586832523346} 01/28/2022 00:38:23 - INFO - codeparrot_training - Step 4611: {'lr': 0.0004963584701897443, 'samples': 885504, 'steps': 4611, 'loss/train': 4.184265732765198} 01/28/2022 00:38:27 - INFO - codeparrot_training - Step 4612: {'lr': 0.0004963556870780523, 'samples': 885696, 'steps': 4612, 'loss/train': 4.080895900726318} 01/28/2022 00:38:33 - INFO - codeparrot_training - Step 4613: {'lr': 0.0004963529029110509, 'samples': 885888, 'steps': 4613, 'loss/train': 3.707218050956726} 01/28/2022 00:38:37 - INFO - codeparrot_training - Step 4614: {'lr': 0.0004963501176887519, 'samples': 886080, 'steps': 4614, 'loss/train': 3.494336485862732} 01/28/2022 00:38:41 - INFO - codeparrot_training - Step 4615: {'lr': 0.000496347331411167, 'samples': 886272, 'steps': 4615, 'loss/train': 3.816191554069519} 01/28/2022 00:38:45 - INFO - codeparrot_training - Step 4616: {'lr': 0.0004963445440783086, 'samples': 886464, 'steps': 4616, 'loss/train': 2.761118710041046} 01/28/2022 00:38:49 - INFO - codeparrot_training - Step 4617: {'lr': 0.0004963417556901882, 'samples': 886656, 'steps': 4617, 'loss/train': 3.4401715993881226} 01/28/2022 00:38:56 - INFO - codeparrot_training - Step 4618: {'lr': 0.0004963389662468182, 'samples': 886848, 'steps': 4618, 'loss/train': 3.846843123435974} 01/28/2022 00:39:00 - INFO - codeparrot_training - Step 4619: {'lr': 0.0004963361757482101, 'samples': 887040, 'steps': 4619, 'loss/train': 3.230884552001953} 01/28/2022 00:39:04 - INFO - codeparrot_training - Step 4620: {'lr': 0.000496333384194376, 'samples': 887232, 'steps': 4620, 'loss/train': 4.186198711395264} 01/28/2022 00:39:08 - INFO - codeparrot_training - Step 4621: {'lr': 0.000496330591585328, 'samples': 887424, 'steps': 4621, 'loss/train': 2.7835278511047363} 01/28/2022 00:39:12 - INFO - codeparrot_training - Step 4622: {'lr': 0.0004963277979210779, 'samples': 887616, 'steps': 4622, 'loss/train': 3.9024603366851807} 01/28/2022 00:39:18 - INFO - codeparrot_training - Step 4623: {'lr': 0.0004963250032016379, 'samples': 887808, 'steps': 4623, 'loss/train': 3.69538414478302} 01/28/2022 00:39:22 - INFO - codeparrot_training - Step 4624: {'lr': 0.0004963222074270197, 'samples': 888000, 'steps': 4624, 'loss/train': 3.8345324993133545} 01/28/2022 00:39:26 - INFO - codeparrot_training - Step 4625: {'lr': 0.0004963194105972353, 'samples': 888192, 'steps': 4625, 'loss/train': 3.2497626543045044} 01/28/2022 00:39:30 - INFO - codeparrot_training - Step 4626: {'lr': 0.0004963166127122969, 'samples': 888384, 'steps': 4626, 'loss/train': 3.1927889585494995} 01/28/2022 00:39:35 - INFO - codeparrot_training - Step 4627: {'lr': 0.0004963138137722161, 'samples': 888576, 'steps': 4627, 'loss/train': 2.9334070086479187} 01/28/2022 00:39:40 - INFO - codeparrot_training - Step 4628: {'lr': 0.0004963110137770054, 'samples': 888768, 'steps': 4628, 'loss/train': 3.3819193840026855} 01/28/2022 00:39:44 - INFO - codeparrot_training - Step 4629: {'lr': 0.0004963082127266764, 'samples': 888960, 'steps': 4629, 'loss/train': 2.0242146849632263} 01/28/2022 00:39:48 - INFO - codeparrot_training - Step 4630: {'lr': 0.0004963054106212414, 'samples': 889152, 'steps': 4630, 'loss/train': 3.7883652448654175} 01/28/2022 00:39:53 - INFO - codeparrot_training - Step 4631: {'lr': 0.000496302607460712, 'samples': 889344, 'steps': 4631, 'loss/train': 2.956388533115387} 01/28/2022 00:39:57 - INFO - codeparrot_training - Step 4632: {'lr': 0.0004962998032451005, 'samples': 889536, 'steps': 4632, 'loss/train': 2.5736457109451294} 01/28/2022 00:40:03 - INFO - codeparrot_training - Step 4633: {'lr': 0.0004962969979744189, 'samples': 889728, 'steps': 4633, 'loss/train': 2.4626707434654236} 01/28/2022 00:40:07 - INFO - codeparrot_training - Step 4634: {'lr': 0.0004962941916486791, 'samples': 889920, 'steps': 4634, 'loss/train': 2.180883049964905} 01/28/2022 00:40:11 - INFO - codeparrot_training - Step 4635: {'lr': 0.0004962913842678934, 'samples': 890112, 'steps': 4635, 'loss/train': 3.738152861595154} 01/28/2022 00:40:16 - INFO - codeparrot_training - Step 4636: {'lr': 0.0004962885758320734, 'samples': 890304, 'steps': 4636, 'loss/train': 3.1911531686782837} 01/28/2022 00:40:20 - INFO - codeparrot_training - Step 4637: {'lr': 0.0004962857663412314, 'samples': 890496, 'steps': 4637, 'loss/train': 2.7855504155158997} 01/28/2022 00:40:26 - INFO - codeparrot_training - Step 4638: {'lr': 0.0004962829557953794, 'samples': 890688, 'steps': 4638, 'loss/train': 2.8012239933013916} 01/28/2022 00:40:30 - INFO - codeparrot_training - Step 4639: {'lr': 0.0004962801441945293, 'samples': 890880, 'steps': 4639, 'loss/train': 2.3829628229141235} 01/28/2022 00:40:34 - INFO - codeparrot_training - Step 4640: {'lr': 0.0004962773315386935, 'samples': 891072, 'steps': 4640, 'loss/train': 2.8871145844459534} 01/28/2022 00:40:38 - INFO - codeparrot_training - Step 4641: {'lr': 0.0004962745178278837, 'samples': 891264, 'steps': 4641, 'loss/train': 2.50970721244812} 01/28/2022 00:40:43 - INFO - codeparrot_training - Step 4642: {'lr': 0.000496271703062112, 'samples': 891456, 'steps': 4642, 'loss/train': 2.529011607170105} 01/28/2022 00:40:49 - INFO - codeparrot_training - Step 4643: {'lr': 0.0004962688872413906, 'samples': 891648, 'steps': 4643, 'loss/train': 3.0385751724243164} 01/28/2022 00:40:53 - INFO - codeparrot_training - Step 4644: {'lr': 0.0004962660703657315, 'samples': 891840, 'steps': 4644, 'loss/train': 3.6252601146698} 01/28/2022 00:40:57 - INFO - codeparrot_training - Step 4645: {'lr': 0.0004962632524351467, 'samples': 892032, 'steps': 4645, 'loss/train': 1.809964120388031} 01/28/2022 00:41:01 - INFO - codeparrot_training - Step 4646: {'lr': 0.0004962604334496483, 'samples': 892224, 'steps': 4646, 'loss/train': 3.0866256952285767} 01/28/2022 00:41:06 - INFO - codeparrot_training - Step 4647: {'lr': 0.0004962576134092485, 'samples': 892416, 'steps': 4647, 'loss/train': 2.510429263114929} 01/28/2022 00:41:11 - INFO - codeparrot_training - Step 4648: {'lr': 0.0004962547923139592, 'samples': 892608, 'steps': 4648, 'loss/train': 2.22296404838562} 01/28/2022 00:41:15 - INFO - codeparrot_training - Step 4649: {'lr': 0.0004962519701637926, 'samples': 892800, 'steps': 4649, 'loss/train': 3.357509136199951} 01/28/2022 00:41:19 - INFO - codeparrot_training - Step 4650: {'lr': 0.0004962491469587607, 'samples': 892992, 'steps': 4650, 'loss/train': 2.5638332962989807} 01/28/2022 00:41:24 - INFO - codeparrot_training - Step 4651: {'lr': 0.0004962463226988758, 'samples': 893184, 'steps': 4651, 'loss/train': 2.6207438707351685} 01/28/2022 00:41:28 - INFO - codeparrot_training - Step 4652: {'lr': 0.0004962434973841497, 'samples': 893376, 'steps': 4652, 'loss/train': 2.21720427274704} 01/28/2022 00:41:35 - INFO - codeparrot_training - Step 4653: {'lr': 0.0004962406710145946, 'samples': 893568, 'steps': 4653, 'loss/train': 2.7001326084136963} 01/28/2022 00:41:39 - INFO - codeparrot_training - Step 4654: {'lr': 0.0004962378435902228, 'samples': 893760, 'steps': 4654, 'loss/train': 1.9150301814079285} 01/28/2022 00:41:43 - INFO - codeparrot_training - Step 4655: {'lr': 0.0004962350151110461, 'samples': 893952, 'steps': 4655, 'loss/train': 3.710290789604187} 01/28/2022 00:41:47 - INFO - codeparrot_training - Step 4656: {'lr': 0.0004962321855770769, 'samples': 894144, 'steps': 4656, 'loss/train': 3.4380630254745483} 01/28/2022 00:41:52 - INFO - codeparrot_training - Step 4657: {'lr': 0.0004962293549883273, 'samples': 894336, 'steps': 4657, 'loss/train': 1.9617583751678467} 01/28/2022 00:41:57 - INFO - codeparrot_training - Step 4658: {'lr': 0.0004962265233448092, 'samples': 894528, 'steps': 4658, 'loss/train': 4.229194521903992} 01/28/2022 00:42:01 - INFO - codeparrot_training - Step 4659: {'lr': 0.0004962236906465349, 'samples': 894720, 'steps': 4659, 'loss/train': 2.119792342185974} 01/28/2022 00:42:05 - INFO - codeparrot_training - Step 4660: {'lr': 0.0004962208568935164, 'samples': 894912, 'steps': 4660, 'loss/train': 2.518196403980255} 01/28/2022 00:42:09 - INFO - codeparrot_training - Step 4661: {'lr': 0.000496218022085766, 'samples': 895104, 'steps': 4661, 'loss/train': 2.554018020629883} 01/28/2022 00:42:14 - INFO - codeparrot_training - Step 4662: {'lr': 0.0004962151862232958, 'samples': 895296, 'steps': 4662, 'loss/train': 2.93831866979599} 01/28/2022 00:42:19 - INFO - codeparrot_training - Step 4663: {'lr': 0.000496212349306118, 'samples': 895488, 'steps': 4663, 'loss/train': 2.987299919128418} 01/28/2022 00:42:23 - INFO - codeparrot_training - Step 4664: {'lr': 0.0004962095113342445, 'samples': 895680, 'steps': 4664, 'loss/train': 2.67551851272583} 01/28/2022 00:42:28 - INFO - codeparrot_training - Step 4665: {'lr': 0.0004962066723076878, 'samples': 895872, 'steps': 4665, 'loss/train': 3.538360118865967} 01/28/2022 00:42:32 - INFO - codeparrot_training - Step 4666: {'lr': 0.0004962038322264598, 'samples': 896064, 'steps': 4666, 'loss/train': 2.7710153460502625} 01/28/2022 00:42:36 - INFO - codeparrot_training - Step 4667: {'lr': 0.0004962009910905728, 'samples': 896256, 'steps': 4667, 'loss/train': 3.844488501548767} 01/28/2022 00:42:41 - INFO - codeparrot_training - Step 4668: {'lr': 0.0004961981489000389, 'samples': 896448, 'steps': 4668, 'loss/train': 2.2235727310180664} 01/28/2022 00:42:45 - INFO - codeparrot_training - Step 4669: {'lr': 0.0004961953056548703, 'samples': 896640, 'steps': 4669, 'loss/train': 3.775870442390442} 01/28/2022 00:42:50 - INFO - codeparrot_training - Step 4670: {'lr': 0.0004961924613550793, 'samples': 896832, 'steps': 4670, 'loss/train': 3.442336320877075} 01/28/2022 00:42:54 - INFO - codeparrot_training - Step 4671: {'lr': 0.0004961896160006778, 'samples': 897024, 'steps': 4671, 'loss/train': 2.218148946762085} 01/28/2022 00:42:58 - INFO - codeparrot_training - Step 4672: {'lr': 0.0004961867695916782, 'samples': 897216, 'steps': 4672, 'loss/train': 3.4477243423461914} 01/28/2022 00:43:04 - INFO - codeparrot_training - Step 4673: {'lr': 0.0004961839221280927, 'samples': 897408, 'steps': 4673, 'loss/train': 2.5450825095176697} 01/28/2022 00:43:08 - INFO - codeparrot_training - Step 4674: {'lr': 0.0004961810736099334, 'samples': 897600, 'steps': 4674, 'loss/train': 2.5500749945640564} 01/28/2022 00:43:12 - INFO - codeparrot_training - Step 4675: {'lr': 0.0004961782240372126, 'samples': 897792, 'steps': 4675, 'loss/train': 2.027247190475464} 01/28/2022 00:43:16 - INFO - codeparrot_training - Step 4676: {'lr': 0.0004961753734099425, 'samples': 897984, 'steps': 4676, 'loss/train': 2.2377832531929016} 01/28/2022 00:43:21 - INFO - codeparrot_training - Step 4677: {'lr': 0.0004961725217281352, 'samples': 898176, 'steps': 4677, 'loss/train': 3.9857393503189087} 01/28/2022 00:43:26 - INFO - codeparrot_training - Step 4678: {'lr': 0.0004961696689918029, 'samples': 898368, 'steps': 4678, 'loss/train': 2.6887367963790894} 01/28/2022 00:43:31 - INFO - codeparrot_training - Step 4679: {'lr': 0.0004961668152009581, 'samples': 898560, 'steps': 4679, 'loss/train': 3.8074657917022705} 01/28/2022 00:43:35 - INFO - codeparrot_training - Step 4680: {'lr': 0.0004961639603556127, 'samples': 898752, 'steps': 4680, 'loss/train': 3.304810881614685} 01/28/2022 00:43:39 - INFO - codeparrot_training - Step 4681: {'lr': 0.0004961611044557792, 'samples': 898944, 'steps': 4681, 'loss/train': 3.18235445022583} 01/28/2022 00:43:43 - INFO - codeparrot_training - Step 4682: {'lr': 0.0004961582475014695, 'samples': 899136, 'steps': 4682, 'loss/train': 1.7193508744239807} 01/28/2022 00:43:49 - INFO - codeparrot_training - Step 4683: {'lr': 0.0004961553894926961, 'samples': 899328, 'steps': 4683, 'loss/train': 2.475476324558258} 01/28/2022 00:43:53 - INFO - codeparrot_training - Step 4684: {'lr': 0.0004961525304294712, 'samples': 899520, 'steps': 4684, 'loss/train': 3.182825446128845} 01/28/2022 00:43:57 - INFO - codeparrot_training - Step 4685: {'lr': 0.000496149670311807, 'samples': 899712, 'steps': 4685, 'loss/train': 3.651124119758606} 01/28/2022 00:44:02 - INFO - codeparrot_training - Step 4686: {'lr': 0.0004961468091397158, 'samples': 899904, 'steps': 4686, 'loss/train': 2.6162198781967163} 01/28/2022 00:44:08 - INFO - codeparrot_training - Step 4687: {'lr': 0.0004961439469132098, 'samples': 900096, 'steps': 4687, 'loss/train': 3.952471375465393} 01/28/2022 00:44:12 - INFO - codeparrot_training - Step 4688: {'lr': 0.0004961410836323014, 'samples': 900288, 'steps': 4688, 'loss/train': 3.408446431159973} 01/28/2022 00:44:16 - INFO - codeparrot_training - Step 4689: {'lr': 0.0004961382192970027, 'samples': 900480, 'steps': 4689, 'loss/train': 1.4273311793804169} 01/28/2022 00:44:20 - INFO - codeparrot_training - Step 4690: {'lr': 0.0004961353539073258, 'samples': 900672, 'steps': 4690, 'loss/train': 2.544931948184967} 01/28/2022 00:44:24 - INFO - codeparrot_training - Step 4691: {'lr': 0.0004961324874632835, 'samples': 900864, 'steps': 4691, 'loss/train': 2.325681746006012} 01/28/2022 00:44:29 - INFO - codeparrot_training - Step 4692: {'lr': 0.0004961296199648877, 'samples': 901056, 'steps': 4692, 'loss/train': 3.773853063583374} 01/28/2022 00:44:34 - INFO - codeparrot_training - Step 4693: {'lr': 0.0004961267514121507, 'samples': 901248, 'steps': 4693, 'loss/train': 3.108606219291687} 01/28/2022 00:44:38 - INFO - codeparrot_training - Step 4694: {'lr': 0.0004961238818050849, 'samples': 901440, 'steps': 4694, 'loss/train': 3.683403253555298} 01/28/2022 00:44:43 - INFO - codeparrot_training - Step 4695: {'lr': 0.0004961210111437026, 'samples': 901632, 'steps': 4695, 'loss/train': 2.132631540298462} 01/28/2022 00:44:47 - INFO - codeparrot_training - Step 4696: {'lr': 0.0004961181394280159, 'samples': 901824, 'steps': 4696, 'loss/train': 2.6588380336761475} 01/28/2022 00:44:51 - INFO - codeparrot_training - Step 4697: {'lr': 0.0004961152666580373, 'samples': 902016, 'steps': 4697, 'loss/train': 1.9991497993469238} 01/28/2022 00:44:57 - INFO - codeparrot_training - Step 4698: {'lr': 0.0004961123928337791, 'samples': 902208, 'steps': 4698, 'loss/train': 1.2558602392673492} 01/28/2022 00:45:01 - INFO - codeparrot_training - Step 4699: {'lr': 0.0004961095179552535, 'samples': 902400, 'steps': 4699, 'loss/train': 2.9457249641418457} 01/28/2022 00:45:05 - INFO - codeparrot_training - Step 4700: {'lr': 0.0004961066420224729, 'samples': 902592, 'steps': 4700, 'loss/train': 3.721566081047058} 01/28/2022 00:45:09 - INFO - codeparrot_training - Step 4701: {'lr': 0.0004961037650354496, 'samples': 902784, 'steps': 4701, 'loss/train': 2.7450438737869263} 01/28/2022 00:45:14 - INFO - codeparrot_training - Step 4702: {'lr': 0.0004961008869941959, 'samples': 902976, 'steps': 4702, 'loss/train': 3.206966757774353} 01/28/2022 00:45:20 - INFO - codeparrot_training - Step 4703: {'lr': 0.0004960980078987241, 'samples': 903168, 'steps': 4703, 'loss/train': 3.2729666233062744} 01/28/2022 00:45:24 - INFO - codeparrot_training - Step 4704: {'lr': 0.0004960951277490467, 'samples': 903360, 'steps': 4704, 'loss/train': 3.147818684577942} 01/28/2022 00:45:28 - INFO - codeparrot_training - Step 4705: {'lr': 0.0004960922465451758, 'samples': 903552, 'steps': 4705, 'loss/train': 2.2145819664001465} 01/28/2022 00:45:33 - INFO - codeparrot_training - Step 4706: {'lr': 0.0004960893642871239, 'samples': 903744, 'steps': 4706, 'loss/train': 2.767132580280304} 01/28/2022 00:45:38 - INFO - codeparrot_training - Step 4707: {'lr': 0.0004960864809749034, 'samples': 903936, 'steps': 4707, 'loss/train': 2.4501625299453735} 01/28/2022 00:45:42 - INFO - codeparrot_training - Step 4708: {'lr': 0.0004960835966085264, 'samples': 904128, 'steps': 4708, 'loss/train': 2.966833233833313} 01/28/2022 00:45:46 - INFO - codeparrot_training - Step 4709: {'lr': 0.0004960807111880055, 'samples': 904320, 'steps': 4709, 'loss/train': 2.5536117553710938} 01/28/2022 00:45:50 - INFO - codeparrot_training - Step 4710: {'lr': 0.000496077824713353, 'samples': 904512, 'steps': 4710, 'loss/train': 3.986132025718689} 01/28/2022 00:45:55 - INFO - codeparrot_training - Step 4711: {'lr': 0.0004960749371845812, 'samples': 904704, 'steps': 4711, 'loss/train': 3.18798565864563} 01/28/2022 00:45:59 - INFO - codeparrot_training - Step 4712: {'lr': 0.0004960720486017025, 'samples': 904896, 'steps': 4712, 'loss/train': 3.8161540031433105} 01/28/2022 00:46:05 - INFO - codeparrot_training - Step 4713: {'lr': 0.0004960691589647292, 'samples': 905088, 'steps': 4713, 'loss/train': 1.6870425939559937} 01/28/2022 00:46:09 - INFO - codeparrot_training - Step 4714: {'lr': 0.0004960662682736739, 'samples': 905280, 'steps': 4714, 'loss/train': 3.3481374979019165} 01/28/2022 00:46:14 - INFO - codeparrot_training - Step 4715: {'lr': 0.0004960633765285487, 'samples': 905472, 'steps': 4715, 'loss/train': 4.809463977813721} 01/28/2022 00:46:18 - INFO - codeparrot_training - Step 4716: {'lr': 0.0004960604837293663, 'samples': 905664, 'steps': 4716, 'loss/train': 4.114566206932068} 01/28/2022 00:46:22 - INFO - codeparrot_training - Step 4717: {'lr': 0.0004960575898761388, 'samples': 905856, 'steps': 4717, 'loss/train': 3.089625835418701} 01/28/2022 00:46:28 - INFO - codeparrot_training - Step 4718: {'lr': 0.0004960546949688788, 'samples': 906048, 'steps': 4718, 'loss/train': 3.389636993408203} 01/28/2022 00:46:32 - INFO - codeparrot_training - Step 4719: {'lr': 0.0004960517990075985, 'samples': 906240, 'steps': 4719, 'loss/train': 1.9682210683822632} 01/28/2022 00:46:36 - INFO - codeparrot_training - Step 4720: {'lr': 0.0004960489019923105, 'samples': 906432, 'steps': 4720, 'loss/train': 3.148597240447998} 01/28/2022 00:46:40 - INFO - codeparrot_training - Step 4721: {'lr': 0.0004960460039230271, 'samples': 906624, 'steps': 4721, 'loss/train': 4.181139349937439} 01/28/2022 00:46:46 - INFO - codeparrot_training - Step 4722: {'lr': 0.0004960431047997608, 'samples': 906816, 'steps': 4722, 'loss/train': 2.502921402454376} 01/28/2022 00:46:50 - INFO - codeparrot_training - Step 4723: {'lr': 0.0004960402046225239, 'samples': 907008, 'steps': 4723, 'loss/train': 3.215916395187378} 01/28/2022 00:46:54 - INFO - codeparrot_training - Step 4724: {'lr': 0.0004960373033913289, 'samples': 907200, 'steps': 4724, 'loss/train': 3.0476757287979126} 01/28/2022 00:46:58 - INFO - codeparrot_training - Step 4725: {'lr': 0.0004960344011061882, 'samples': 907392, 'steps': 4725, 'loss/train': 4.172265887260437} 01/28/2022 00:47:02 - INFO - codeparrot_training - Step 4726: {'lr': 0.0004960314977671144, 'samples': 907584, 'steps': 4726, 'loss/train': 3.8514554500579834} 01/28/2022 00:47:08 - INFO - codeparrot_training - Step 4727: {'lr': 0.0004960285933741196, 'samples': 907776, 'steps': 4727, 'loss/train': 0.6622457653284073} 01/28/2022 00:47:12 - INFO - codeparrot_training - Step 4728: {'lr': 0.0004960256879272166, 'samples': 907968, 'steps': 4728, 'loss/train': 0.98743736743927} 01/28/2022 00:47:16 - INFO - codeparrot_training - Step 4729: {'lr': 0.0004960227814264175, 'samples': 908160, 'steps': 4729, 'loss/train': 2.354639947414398} 01/28/2022 00:47:20 - INFO - codeparrot_training - Step 4730: {'lr': 0.0004960198738717351, 'samples': 908352, 'steps': 4730, 'loss/train': 2.5649396181106567} 01/28/2022 00:47:25 - INFO - codeparrot_training - Step 4731: {'lr': 0.0004960169652631815, 'samples': 908544, 'steps': 4731, 'loss/train': 3.365182399749756} 01/28/2022 00:47:31 - INFO - codeparrot_training - Step 4732: {'lr': 0.0004960140556007695, 'samples': 908736, 'steps': 4732, 'loss/train': 2.5046077966690063} 01/28/2022 00:47:35 - INFO - codeparrot_training - Step 4733: {'lr': 0.0004960111448845114, 'samples': 908928, 'steps': 4733, 'loss/train': 2.9715312123298645} 01/28/2022 00:47:39 - INFO - codeparrot_training - Step 4734: {'lr': 0.0004960082331144195, 'samples': 909120, 'steps': 4734, 'loss/train': 2.549290180206299} 01/28/2022 00:47:43 - INFO - codeparrot_training - Step 4735: {'lr': 0.0004960053202905066, 'samples': 909312, 'steps': 4735, 'loss/train': 3.2284387350082397} 01/28/2022 00:47:48 - INFO - codeparrot_training - Step 4736: {'lr': 0.0004960024064127849, 'samples': 909504, 'steps': 4736, 'loss/train': 3.591287612915039} 01/28/2022 00:47:52 - INFO - codeparrot_training - Step 4737: {'lr': 0.0004959994914812671, 'samples': 909696, 'steps': 4737, 'loss/train': 1.0999500453472137} 01/28/2022 00:47:57 - INFO - codeparrot_training - Step 4738: {'lr': 0.0004959965754959656, 'samples': 909888, 'steps': 4738, 'loss/train': 1.9577459692955017} 01/28/2022 00:48:01 - INFO - codeparrot_training - Step 4739: {'lr': 0.0004959936584568928, 'samples': 910080, 'steps': 4739, 'loss/train': 3.4455567598342896} 01/28/2022 00:48:06 - INFO - codeparrot_training - Step 4740: {'lr': 0.0004959907403640614, 'samples': 910272, 'steps': 4740, 'loss/train': 3.1557776927948} 01/28/2022 00:48:10 - INFO - codeparrot_training - Step 4741: {'lr': 0.0004959878212174837, 'samples': 910464, 'steps': 4741, 'loss/train': 3.0131882429122925} 01/28/2022 00:48:14 - INFO - codeparrot_training - Step 4742: {'lr': 0.0004959849010171723, 'samples': 910656, 'steps': 4742, 'loss/train': 2.5712181329727173} 01/28/2022 00:48:19 - INFO - codeparrot_training - Step 4743: {'lr': 0.0004959819797631397, 'samples': 910848, 'steps': 4743, 'loss/train': 2.986541748046875} 01/28/2022 00:48:23 - INFO - codeparrot_training - Step 4744: {'lr': 0.0004959790574553984, 'samples': 911040, 'steps': 4744, 'loss/train': 2.3776440024375916} 01/28/2022 00:48:28 - INFO - codeparrot_training - Step 4745: {'lr': 0.000495976134093961, 'samples': 911232, 'steps': 4745, 'loss/train': 3.4718252420425415} 01/28/2022 00:48:32 - INFO - codeparrot_training - Step 4746: {'lr': 0.0004959732096788398, 'samples': 911424, 'steps': 4746, 'loss/train': 3.594449043273926} 01/28/2022 00:48:36 - INFO - codeparrot_training - Step 4747: {'lr': 0.0004959702842100475, 'samples': 911616, 'steps': 4747, 'loss/train': 2.23151957988739} 01/28/2022 00:48:42 - INFO - codeparrot_training - Step 4748: {'lr': 0.0004959673576875967, 'samples': 911808, 'steps': 4748, 'loss/train': 3.7101584672927856} 01/28/2022 00:48:47 - INFO - codeparrot_training - Step 4749: {'lr': 0.0004959644301114998, 'samples': 912000, 'steps': 4749, 'loss/train': 3.0652610063552856} 01/28/2022 00:48:51 - INFO - codeparrot_training - Step 4750: {'lr': 0.0004959615014817694, 'samples': 912192, 'steps': 4750, 'loss/train': 3.039864420890808} 01/28/2022 00:48:55 - INFO - codeparrot_training - Step 4751: {'lr': 0.000495958571798418, 'samples': 912384, 'steps': 4751, 'loss/train': 3.0419293642044067} 01/28/2022 00:48:59 - INFO - codeparrot_training - Step 4752: {'lr': 0.0004959556410614582, 'samples': 912576, 'steps': 4752, 'loss/train': 3.3134000301361084} 01/28/2022 00:49:04 - INFO - codeparrot_training - Step 4753: {'lr': 0.0004959527092709026, 'samples': 912768, 'steps': 4753, 'loss/train': 3.348997950553894} 01/28/2022 00:49:09 - INFO - codeparrot_training - Step 4754: {'lr': 0.0004959497764267636, 'samples': 912960, 'steps': 4754, 'loss/train': 2.9927889704704285} 01/28/2022 00:49:13 - INFO - codeparrot_training - Step 4755: {'lr': 0.0004959468425290537, 'samples': 913152, 'steps': 4755, 'loss/train': 1.2083352506160736} 01/28/2022 00:49:17 - INFO - codeparrot_training - Step 4756: {'lr': 0.0004959439075777858, 'samples': 913344, 'steps': 4756, 'loss/train': 2.273366332054138} 01/28/2022 00:49:21 - INFO - codeparrot_training - Step 4757: {'lr': 0.0004959409715729723, 'samples': 913536, 'steps': 4757, 'loss/train': 3.818446397781372} 01/28/2022 00:49:28 - INFO - codeparrot_training - Step 4758: {'lr': 0.0004959380345146258, 'samples': 913728, 'steps': 4758, 'loss/train': 3.4517011642456055} 01/28/2022 00:49:32 - INFO - codeparrot_training - Step 4759: {'lr': 0.0004959350964027588, 'samples': 913920, 'steps': 4759, 'loss/train': 2.935108959674835} 01/28/2022 00:49:36 - INFO - codeparrot_training - Step 4760: {'lr': 0.000495932157237384, 'samples': 914112, 'steps': 4760, 'loss/train': 3.178986668586731} 01/28/2022 00:49:41 - INFO - codeparrot_training - Step 4761: {'lr': 0.0004959292170185139, 'samples': 914304, 'steps': 4761, 'loss/train': 2.8763437271118164} 01/28/2022 00:49:45 - INFO - codeparrot_training - Step 4762: {'lr': 0.0004959262757461611, 'samples': 914496, 'steps': 4762, 'loss/train': 5.665349006652832} 01/28/2022 00:49:49 - INFO - codeparrot_training - Step 4763: {'lr': 0.0004959233334203382, 'samples': 914688, 'steps': 4763, 'loss/train': 2.865011215209961} 01/28/2022 00:49:54 - INFO - codeparrot_training - Step 4764: {'lr': 0.0004959203900410579, 'samples': 914880, 'steps': 4764, 'loss/train': 3.0992352962493896} 01/28/2022 00:49:59 - INFO - codeparrot_training - Step 4765: {'lr': 0.0004959174456083327, 'samples': 915072, 'steps': 4765, 'loss/train': 3.555457592010498} 01/28/2022 00:50:03 - INFO - codeparrot_training - Step 4766: {'lr': 0.0004959145001221752, 'samples': 915264, 'steps': 4766, 'loss/train': 3.081890344619751} 01/28/2022 00:50:07 - INFO - codeparrot_training - Step 4767: {'lr': 0.0004959115535825982, 'samples': 915456, 'steps': 4767, 'loss/train': 2.8430914878845215} 01/28/2022 00:50:12 - INFO - codeparrot_training - Step 4768: {'lr': 0.000495908605989614, 'samples': 915648, 'steps': 4768, 'loss/train': 5.060450077056885} 01/28/2022 00:50:17 - INFO - codeparrot_training - Step 4769: {'lr': 0.0004959056573432357, 'samples': 915840, 'steps': 4769, 'loss/train': 3.6583635807037354} 01/28/2022 00:50:21 - INFO - codeparrot_training - Step 4770: {'lr': 0.0004959027076434754, 'samples': 916032, 'steps': 4770, 'loss/train': 2.7508928775787354} 01/28/2022 00:50:25 - INFO - codeparrot_training - Step 4771: {'lr': 0.000495899756890346, 'samples': 916224, 'steps': 4771, 'loss/train': 2.590555429458618} 01/28/2022 00:50:29 - INFO - codeparrot_training - Step 4772: {'lr': 0.0004958968050838603, 'samples': 916416, 'steps': 4772, 'loss/train': 3.218960165977478} 01/28/2022 00:50:35 - INFO - codeparrot_training - Step 4773: {'lr': 0.0004958938522240306, 'samples': 916608, 'steps': 4773, 'loss/train': 3.045490264892578} 01/28/2022 00:50:39 - INFO - codeparrot_training - Step 4774: {'lr': 0.0004958908983108697, 'samples': 916800, 'steps': 4774, 'loss/train': 3.109159827232361} 01/28/2022 00:50:44 - INFO - codeparrot_training - Step 4775: {'lr': 0.0004958879433443903, 'samples': 916992, 'steps': 4775, 'loss/train': 4.156283855438232} 01/28/2022 00:50:48 - INFO - codeparrot_training - Step 4776: {'lr': 0.0004958849873246051, 'samples': 917184, 'steps': 4776, 'loss/train': 2.7021934390068054} 01/28/2022 00:50:52 - INFO - codeparrot_training - Step 4777: {'lr': 0.0004958820302515268, 'samples': 917376, 'steps': 4777, 'loss/train': 2.5707532167434692} 01/28/2022 00:50:57 - INFO - codeparrot_training - Step 4778: {'lr': 0.0004958790721251678, 'samples': 917568, 'steps': 4778, 'loss/train': 3.059100866317749} 01/28/2022 00:51:02 - INFO - codeparrot_training - Step 4779: {'lr': 0.000495876112945541, 'samples': 917760, 'steps': 4779, 'loss/train': 3.6144572496414185} 01/28/2022 00:51:06 - INFO - codeparrot_training - Step 4780: {'lr': 0.0004958731527126589, 'samples': 917952, 'steps': 4780, 'loss/train': 2.928152561187744} 01/28/2022 00:51:10 - INFO - codeparrot_training - Step 4781: {'lr': 0.0004958701914265344, 'samples': 918144, 'steps': 4781, 'loss/train': 3.3712334632873535} 01/28/2022 00:51:14 - INFO - codeparrot_training - Step 4782: {'lr': 0.0004958672290871799, 'samples': 918336, 'steps': 4782, 'loss/train': 3.677649736404419} 01/28/2022 00:51:19 - INFO - codeparrot_training - Step 4783: {'lr': 0.0004958642656946084, 'samples': 918528, 'steps': 4783, 'loss/train': 2.7804731726646423} 01/28/2022 00:51:24 - INFO - codeparrot_training - Step 4784: {'lr': 0.0004958613012488324, 'samples': 918720, 'steps': 4784, 'loss/train': 1.3148117065429688} 01/28/2022 00:51:28 - INFO - codeparrot_training - Step 4785: {'lr': 0.0004958583357498647, 'samples': 918912, 'steps': 4785, 'loss/train': 3.3579676151275635} 01/28/2022 00:51:32 - INFO - codeparrot_training - Step 4786: {'lr': 0.000495855369197718, 'samples': 919104, 'steps': 4786, 'loss/train': 3.2579190731048584} 01/28/2022 00:51:36 - INFO - codeparrot_training - Step 4787: {'lr': 0.0004958524015924048, 'samples': 919296, 'steps': 4787, 'loss/train': 3.430109739303589} 01/28/2022 00:51:41 - INFO - codeparrot_training - Step 4788: {'lr': 0.0004958494329339382, 'samples': 919488, 'steps': 4788, 'loss/train': 4.701287627220154} 01/28/2022 00:51:46 - INFO - codeparrot_training - Step 4789: {'lr': 0.0004958464632223306, 'samples': 919680, 'steps': 4789, 'loss/train': 3.162592649459839} 01/28/2022 00:51:50 - INFO - codeparrot_training - Step 4790: {'lr': 0.0004958434924575947, 'samples': 919872, 'steps': 4790, 'loss/train': 3.3744099140167236} 01/28/2022 00:51:54 - INFO - codeparrot_training - Step 4791: {'lr': 0.0004958405206397434, 'samples': 920064, 'steps': 4791, 'loss/train': 3.0344992876052856} 01/28/2022 00:51:58 - INFO - codeparrot_training - Step 4792: {'lr': 0.0004958375477687896, 'samples': 920256, 'steps': 4792, 'loss/train': 2.506931483745575} 01/28/2022 00:52:04 - INFO - codeparrot_training - Step 4793: {'lr': 0.0004958345738447456, 'samples': 920448, 'steps': 4793, 'loss/train': 2.3976184129714966} 01/28/2022 00:52:09 - INFO - codeparrot_training - Step 4794: {'lr': 0.0004958315988676244, 'samples': 920640, 'steps': 4794, 'loss/train': 2.97581148147583} 01/28/2022 00:52:13 - INFO - codeparrot_training - Step 4795: {'lr': 0.0004958286228374387, 'samples': 920832, 'steps': 4795, 'loss/train': 2.6462262868881226} 01/28/2022 00:52:17 - INFO - codeparrot_training - Step 4796: {'lr': 0.0004958256457542011, 'samples': 921024, 'steps': 4796, 'loss/train': 2.842157542705536} 01/28/2022 00:52:21 - INFO - codeparrot_training - Step 4797: {'lr': 0.0004958226676179246, 'samples': 921216, 'steps': 4797, 'loss/train': 3.3508872985839844} 01/28/2022 00:52:27 - INFO - codeparrot_training - Step 4798: {'lr': 0.0004958196884286218, 'samples': 921408, 'steps': 4798, 'loss/train': 3.628854990005493} 01/28/2022 00:52:31 - INFO - codeparrot_training - Step 4799: {'lr': 0.0004958167081863057, 'samples': 921600, 'steps': 4799, 'loss/train': 2.590636968612671} 01/28/2022 00:52:35 - INFO - codeparrot_training - Step 4800: {'lr': 0.0004958137268909887, 'samples': 921792, 'steps': 4800, 'loss/train': 3.2124252319335938} 01/28/2022 00:52:39 - INFO - codeparrot_training - Step 4801: {'lr': 0.0004958107445426838, 'samples': 921984, 'steps': 4801, 'loss/train': 2.672811985015869} 01/28/2022 00:52:43 - INFO - codeparrot_training - Step 4802: {'lr': 0.0004958077611414037, 'samples': 922176, 'steps': 4802, 'loss/train': 3.7049063444137573} 01/28/2022 00:52:49 - INFO - codeparrot_training - Step 4803: {'lr': 0.0004958047766871612, 'samples': 922368, 'steps': 4803, 'loss/train': 2.812734603881836} 01/28/2022 00:52:53 - INFO - codeparrot_training - Step 4804: {'lr': 0.000495801791179969, 'samples': 922560, 'steps': 4804, 'loss/train': 1.704936683177948} 01/28/2022 00:52:57 - INFO - codeparrot_training - Step 4805: {'lr': 0.0004957988046198401, 'samples': 922752, 'steps': 4805, 'loss/train': 1.8163431286811829} 01/28/2022 00:53:01 - INFO - codeparrot_training - Step 4806: {'lr': 0.0004957958170067872, 'samples': 922944, 'steps': 4806, 'loss/train': 2.8642247915267944} 01/28/2022 00:53:06 - INFO - codeparrot_training - Step 4807: {'lr': 0.000495792828340823, 'samples': 923136, 'steps': 4807, 'loss/train': 2.1682698726654053} 01/28/2022 00:53:11 - INFO - codeparrot_training - Step 4808: {'lr': 0.0004957898386219603, 'samples': 923328, 'steps': 4808, 'loss/train': 1.8175851702690125} 01/28/2022 00:53:16 - INFO - codeparrot_training - Step 4809: {'lr': 0.0004957868478502121, 'samples': 923520, 'steps': 4809, 'loss/train': 2.8374980092048645} 01/28/2022 00:53:20 - INFO - codeparrot_training - Step 4810: {'lr': 0.0004957838560255911, 'samples': 923712, 'steps': 4810, 'loss/train': 3.1261314153671265} 01/28/2022 00:53:24 - INFO - codeparrot_training - Step 4811: {'lr': 0.0004957808631481101, 'samples': 923904, 'steps': 4811, 'loss/train': 2.7526391744613647} 01/28/2022 00:53:28 - INFO - codeparrot_training - Step 4812: {'lr': 0.0004957778692177819, 'samples': 924096, 'steps': 4812, 'loss/train': 2.4343934655189514} 01/28/2022 00:53:33 - INFO - codeparrot_training - Step 4813: {'lr': 0.0004957748742346193, 'samples': 924288, 'steps': 4813, 'loss/train': 2.182641863822937} 01/28/2022 00:53:38 - INFO - codeparrot_training - Step 4814: {'lr': 0.0004957718781986352, 'samples': 924480, 'steps': 4814, 'loss/train': 2.6307027339935303} 01/28/2022 00:53:42 - INFO - codeparrot_training - Step 4815: {'lr': 0.0004957688811098425, 'samples': 924672, 'steps': 4815, 'loss/train': 2.3852312564849854} 01/28/2022 00:53:46 - INFO - codeparrot_training - Step 4816: {'lr': 0.0004957658829682539, 'samples': 924864, 'steps': 4816, 'loss/train': 2.3712369203567505} 01/28/2022 00:53:50 - INFO - codeparrot_training - Step 4817: {'lr': 0.0004957628837738823, 'samples': 925056, 'steps': 4817, 'loss/train': 3.3667023181915283} 01/28/2022 00:53:56 - INFO - codeparrot_training - Step 4818: {'lr': 0.0004957598835267405, 'samples': 925248, 'steps': 4818, 'loss/train': 3.784106969833374} 01/28/2022 00:54:00 - INFO - codeparrot_training - Step 4819: {'lr': 0.0004957568822268415, 'samples': 925440, 'steps': 4819, 'loss/train': 2.8389240503311157} 01/28/2022 00:54:05 - INFO - codeparrot_training - Step 4820: {'lr': 0.000495753879874198, 'samples': 925632, 'steps': 4820, 'loss/train': 2.9876180291175842} 01/28/2022 00:54:09 - INFO - codeparrot_training - Step 4821: {'lr': 0.0004957508764688227, 'samples': 925824, 'steps': 4821, 'loss/train': 1.9808260202407837} 01/28/2022 00:54:13 - INFO - codeparrot_training - Step 4822: {'lr': 0.000495747872010729, 'samples': 926016, 'steps': 4822, 'loss/train': 4.099175691604614} 01/28/2022 00:54:18 - INFO - codeparrot_training - Step 4823: {'lr': 0.0004957448664999293, 'samples': 926208, 'steps': 4823, 'loss/train': 2.742504894733429} 01/28/2022 00:54:23 - INFO - codeparrot_training - Step 4824: {'lr': 0.0004957418599364367, 'samples': 926400, 'steps': 4824, 'loss/train': 3.9024935960769653} 01/28/2022 00:54:27 - INFO - codeparrot_training - Step 4825: {'lr': 0.000495738852320264, 'samples': 926592, 'steps': 4825, 'loss/train': 3.3918317556381226} 01/28/2022 00:54:31 - INFO - codeparrot_training - Step 4826: {'lr': 0.000495735843651424, 'samples': 926784, 'steps': 4826, 'loss/train': 2.8856778144836426} 01/28/2022 00:54:35 - INFO - codeparrot_training - Step 4827: {'lr': 0.0004957328339299297, 'samples': 926976, 'steps': 4827, 'loss/train': 3.4965083599090576} 01/28/2022 00:54:41 - INFO - codeparrot_training - Step 4828: {'lr': 0.0004957298231557939, 'samples': 927168, 'steps': 4828, 'loss/train': 2.2348662614822388} 01/28/2022 00:54:45 - INFO - codeparrot_training - Step 4829: {'lr': 0.0004957268113290297, 'samples': 927360, 'steps': 4829, 'loss/train': 1.825361430644989} 01/28/2022 00:54:49 - INFO - codeparrot_training - Step 4830: {'lr': 0.0004957237984496499, 'samples': 927552, 'steps': 4830, 'loss/train': 3.6255176067352295} 01/28/2022 00:54:53 - INFO - codeparrot_training - Step 4831: {'lr': 0.0004957207845176673, 'samples': 927744, 'steps': 4831, 'loss/train': 3.1646887063980103} 01/28/2022 00:54:58 - INFO - codeparrot_training - Step 4832: {'lr': 0.0004957177695330948, 'samples': 927936, 'steps': 4832, 'loss/train': 2.6172527074813843} 01/28/2022 00:55:04 - INFO - codeparrot_training - Step 4833: {'lr': 0.0004957147534959455, 'samples': 928128, 'steps': 4833, 'loss/train': 1.9034274816513062} 01/28/2022 00:55:08 - INFO - codeparrot_training - Step 4834: {'lr': 0.0004957117364062321, 'samples': 928320, 'steps': 4834, 'loss/train': 2.2442740201950073} 01/28/2022 00:55:12 - INFO - codeparrot_training - Step 4835: {'lr': 0.0004957087182639678, 'samples': 928512, 'steps': 4835, 'loss/train': 2.3568933606147766} 01/28/2022 00:55:17 - INFO - codeparrot_training - Step 4836: {'lr': 0.0004957056990691653, 'samples': 928704, 'steps': 4836, 'loss/train': 3.0381245613098145} 01/28/2022 00:55:22 - INFO - codeparrot_training - Step 4837: {'lr': 0.0004957026788218377, 'samples': 928896, 'steps': 4837, 'loss/train': 2.405567228794098} 01/28/2022 00:55:26 - INFO - codeparrot_training - Step 4838: {'lr': 0.0004956996575219977, 'samples': 929088, 'steps': 4838, 'loss/train': 2.965156137943268} 01/28/2022 00:55:30 - INFO - codeparrot_training - Step 4839: {'lr': 0.0004956966351696584, 'samples': 929280, 'steps': 4839, 'loss/train': 2.6243467926979065} 01/28/2022 00:55:34 - INFO - codeparrot_training - Step 4840: {'lr': 0.0004956936117648329, 'samples': 929472, 'steps': 4840, 'loss/train': 3.8777986764907837} 01/28/2022 00:55:39 - INFO - codeparrot_training - Step 4841: {'lr': 0.0004956905873075338, 'samples': 929664, 'steps': 4841, 'loss/train': 2.824543297290802} 01/28/2022 00:55:44 - INFO - codeparrot_training - Step 4842: {'lr': 0.0004956875617977743, 'samples': 929856, 'steps': 4842, 'loss/train': 1.747384786605835} 01/28/2022 00:55:48 - INFO - codeparrot_training - Step 4843: {'lr': 0.0004956845352355674, 'samples': 930048, 'steps': 4843, 'loss/train': 3.3950575590133667} 01/28/2022 00:55:52 - INFO - codeparrot_training - Step 4844: {'lr': 0.0004956815076209257, 'samples': 930240, 'steps': 4844, 'loss/train': 2.6368523240089417} 01/28/2022 00:55:57 - INFO - codeparrot_training - Step 4845: {'lr': 0.0004956784789538626, 'samples': 930432, 'steps': 4845, 'loss/train': 3.0300217866897583} 01/28/2022 00:56:01 - INFO - codeparrot_training - Step 4846: {'lr': 0.000495675449234391, 'samples': 930624, 'steps': 4846, 'loss/train': 2.2821013927459717} 01/28/2022 00:56:06 - INFO - codeparrot_training - Step 4847: {'lr': 0.0004956724184625237, 'samples': 930816, 'steps': 4847, 'loss/train': 3.140247344970703} 01/28/2022 00:56:10 - INFO - codeparrot_training - Step 4848: {'lr': 0.0004956693866382738, 'samples': 931008, 'steps': 4848, 'loss/train': 3.1714242696762085} 01/28/2022 00:56:15 - INFO - codeparrot_training - Step 4849: {'lr': 0.0004956663537616542, 'samples': 931200, 'steps': 4849, 'loss/train': 2.8049654960632324} 01/28/2022 00:56:19 - INFO - codeparrot_training - Step 4850: {'lr': 0.000495663319832678, 'samples': 931392, 'steps': 4850, 'loss/train': 3.6536017656326294} 01/28/2022 00:56:23 - INFO - codeparrot_training - Step 4851: {'lr': 0.0004956602848513581, 'samples': 931584, 'steps': 4851, 'loss/train': 2.22873991727829} 01/28/2022 00:56:29 - INFO - codeparrot_training - Step 4852: {'lr': 0.0004956572488177075, 'samples': 931776, 'steps': 4852, 'loss/train': 3.3701913356781006} 01/28/2022 00:56:33 - INFO - codeparrot_training - Step 4853: {'lr': 0.0004956542117317393, 'samples': 931968, 'steps': 4853, 'loss/train': 2.2511531710624695} 01/28/2022 00:56:38 - INFO - codeparrot_training - Step 4854: {'lr': 0.0004956511735934665, 'samples': 932160, 'steps': 4854, 'loss/train': 2.307553768157959} 01/28/2022 00:56:42 - INFO - codeparrot_training - Step 4855: {'lr': 0.000495648134402902, 'samples': 932352, 'steps': 4855, 'loss/train': 1.7796882390975952} 01/28/2022 00:56:46 - INFO - codeparrot_training - Step 4856: {'lr': 0.0004956450941600589, 'samples': 932544, 'steps': 4856, 'loss/train': 3.0133126974105835} 01/28/2022 00:56:51 - INFO - codeparrot_training - Step 4857: {'lr': 0.0004956420528649504, 'samples': 932736, 'steps': 4857, 'loss/train': 2.235646963119507} 01/28/2022 00:56:56 - INFO - codeparrot_training - Step 4858: {'lr': 0.0004956390105175892, 'samples': 932928, 'steps': 4858, 'loss/train': 4.245454430580139} 01/28/2022 00:57:00 - INFO - codeparrot_training - Step 4859: {'lr': 0.0004956359671179885, 'samples': 933120, 'steps': 4859, 'loss/train': 2.687689483165741} 01/28/2022 00:57:04 - INFO - codeparrot_training - Step 4860: {'lr': 0.0004956329226661612, 'samples': 933312, 'steps': 4860, 'loss/train': 2.8987950682640076} 01/28/2022 00:57:08 - INFO - codeparrot_training - Step 4861: {'lr': 0.0004956298771621206, 'samples': 933504, 'steps': 4861, 'loss/train': 1.9850204586982727} 01/28/2022 00:57:15 - INFO - codeparrot_training - Step 4862: {'lr': 0.0004956268306058795, 'samples': 933696, 'steps': 4862, 'loss/train': 1.8619299530982971} 01/28/2022 00:57:19 - INFO - codeparrot_training - Step 4863: {'lr': 0.0004956237829974511, 'samples': 933888, 'steps': 4863, 'loss/train': 1.706199824810028} 01/28/2022 00:57:24 - INFO - codeparrot_training - Step 4864: {'lr': 0.0004956207343368485, 'samples': 934080, 'steps': 4864, 'loss/train': 3.1233723163604736} 01/28/2022 00:57:28 - INFO - codeparrot_training - Step 4865: {'lr': 0.0004956176846240845, 'samples': 934272, 'steps': 4865, 'loss/train': 3.130224823951721} 01/28/2022 00:57:32 - INFO - codeparrot_training - Step 4866: {'lr': 0.0004956146338591725, 'samples': 934464, 'steps': 4866, 'loss/train': 3.7729532718658447} 01/28/2022 00:57:36 - INFO - codeparrot_training - Step 4867: {'lr': 0.0004956115820421253, 'samples': 934656, 'steps': 4867, 'loss/train': 3.486369252204895} 01/28/2022 00:57:41 - INFO - codeparrot_training - Step 4868: {'lr': 0.000495608529172956, 'samples': 934848, 'steps': 4868, 'loss/train': 3.7931699752807617} 01/28/2022 00:57:46 - INFO - codeparrot_training - Step 4869: {'lr': 0.000495605475251678, 'samples': 935040, 'steps': 4869, 'loss/train': 3.4540740251541138} 01/28/2022 00:57:50 - INFO - codeparrot_training - Step 4870: {'lr': 0.000495602420278304, 'samples': 935232, 'steps': 4870, 'loss/train': 2.3947198390960693} 01/28/2022 00:57:54 - INFO - codeparrot_training - Step 4871: {'lr': 0.0004955993642528471, 'samples': 935424, 'steps': 4871, 'loss/train': 3.5297051668167114} 01/28/2022 00:57:58 - INFO - codeparrot_training - Step 4872: {'lr': 0.0004955963071753206, 'samples': 935616, 'steps': 4872, 'loss/train': 2.663811206817627} 01/28/2022 00:58:03 - INFO - codeparrot_training - Step 4873: {'lr': 0.0004955932490457375, 'samples': 935808, 'steps': 4873, 'loss/train': 3.160139322280884} 01/28/2022 00:58:08 - INFO - codeparrot_training - Step 4874: {'lr': 0.0004955901898641109, 'samples': 936000, 'steps': 4874, 'loss/train': 3.918057918548584} 01/28/2022 00:58:12 - INFO - codeparrot_training - Step 4875: {'lr': 0.000495587129630454, 'samples': 936192, 'steps': 4875, 'loss/train': 2.9777841567993164} 01/28/2022 00:58:17 - INFO - codeparrot_training - Step 4876: {'lr': 0.0004955840683447797, 'samples': 936384, 'steps': 4876, 'loss/train': 3.126895308494568} 01/28/2022 00:58:21 - INFO - codeparrot_training - Step 4877: {'lr': 0.0004955810060071012, 'samples': 936576, 'steps': 4877, 'loss/train': 3.0523810386657715} 01/28/2022 00:58:25 - INFO - codeparrot_training - Step 4878: {'lr': 0.0004955779426174318, 'samples': 936768, 'steps': 4878, 'loss/train': 4.152249813079834} 01/28/2022 00:58:31 - INFO - codeparrot_training - Step 4879: {'lr': 0.0004955748781757844, 'samples': 936960, 'steps': 4879, 'loss/train': 3.609367847442627} 01/28/2022 00:58:35 - INFO - codeparrot_training - Step 4880: {'lr': 0.0004955718126821722, 'samples': 937152, 'steps': 4880, 'loss/train': 2.8935091495513916} 01/28/2022 00:58:39 - INFO - codeparrot_training - Step 4881: {'lr': 0.0004955687461366083, 'samples': 937344, 'steps': 4881, 'loss/train': 5.787384510040283} 01/28/2022 00:58:44 - INFO - codeparrot_training - Step 4882: {'lr': 0.000495565678539106, 'samples': 937536, 'steps': 4882, 'loss/train': 2.051703929901123} 01/28/2022 00:58:48 - INFO - codeparrot_training - Step 4883: {'lr': 0.0004955626098896782, 'samples': 937728, 'steps': 4883, 'loss/train': 3.691472053527832} 01/28/2022 00:58:53 - INFO - codeparrot_training - Step 4884: {'lr': 0.0004955595401883381, 'samples': 937920, 'steps': 4884, 'loss/train': 2.8048728704452515} 01/28/2022 00:58:57 - INFO - codeparrot_training - Step 4885: {'lr': 0.0004955564694350989, 'samples': 938112, 'steps': 4885, 'loss/train': 3.26953125} 01/28/2022 00:59:02 - INFO - codeparrot_training - Step 4886: {'lr': 0.0004955533976299739, 'samples': 938304, 'steps': 4886, 'loss/train': 2.361892282962799} 01/28/2022 00:59:06 - INFO - codeparrot_training - Step 4887: {'lr': 0.000495550324772976, 'samples': 938496, 'steps': 4887, 'loss/train': 0.6641974747180939} 01/28/2022 00:59:10 - INFO - codeparrot_training - Step 4888: {'lr': 0.0004955472508641186, 'samples': 938688, 'steps': 4888, 'loss/train': 3.515652894973755} 01/28/2022 00:59:16 - INFO - codeparrot_training - Step 4889: {'lr': 0.0004955441759034146, 'samples': 938880, 'steps': 4889, 'loss/train': 3.0081886053085327} 01/28/2022 00:59:20 - INFO - codeparrot_training - Step 4890: {'lr': 0.0004955410998908774, 'samples': 939072, 'steps': 4890, 'loss/train': 2.500883996486664} 01/28/2022 00:59:24 - INFO - codeparrot_training - Step 4891: {'lr': 0.0004955380228265201, 'samples': 939264, 'steps': 4891, 'loss/train': 2.7559284567832947} 01/28/2022 00:59:29 - INFO - codeparrot_training - Step 4892: {'lr': 0.0004955349447103559, 'samples': 939456, 'steps': 4892, 'loss/train': 1.7305512428283691} 01/28/2022 00:59:33 - INFO - codeparrot_training - Step 4893: {'lr': 0.000495531865542398, 'samples': 939648, 'steps': 4893, 'loss/train': 4.635185480117798} 01/28/2022 00:59:38 - INFO - codeparrot_training - Step 4894: {'lr': 0.0004955287853226594, 'samples': 939840, 'steps': 4894, 'loss/train': 3.517961025238037} 01/28/2022 00:59:42 - INFO - codeparrot_training - Step 4895: {'lr': 0.0004955257040511534, 'samples': 940032, 'steps': 4895, 'loss/train': 3.5283243656158447} 01/28/2022 00:59:46 - INFO - codeparrot_training - Step 4896: {'lr': 0.0004955226217278934, 'samples': 940224, 'steps': 4896, 'loss/train': 3.6804556846618652} 01/28/2022 00:59:51 - INFO - codeparrot_training - Step 4897: {'lr': 0.0004955195383528926, 'samples': 940416, 'steps': 4897, 'loss/train': 2.485860228538513} 01/28/2022 00:59:55 - INFO - codeparrot_training - Step 4898: {'lr': 0.0004955164539261638, 'samples': 940608, 'steps': 4898, 'loss/train': 3.0831702947616577} 01/28/2022 01:00:00 - INFO - codeparrot_training - Step 4899: {'lr': 0.0004955133684477205, 'samples': 940800, 'steps': 4899, 'loss/train': 2.7609894275665283} 01/28/2022 01:00:04 - INFO - codeparrot_training - Step 4900: {'lr': 0.000495510281917576, 'samples': 940992, 'steps': 4900, 'loss/train': 2.9129270911216736} 01/28/2022 01:00:09 - INFO - codeparrot_training - Step 4901: {'lr': 0.0004955071943357433, 'samples': 941184, 'steps': 4901, 'loss/train': 2.8420393466949463} 01/28/2022 01:00:13 - INFO - codeparrot_training - Step 4902: {'lr': 0.0004955041057022358, 'samples': 941376, 'steps': 4902, 'loss/train': 3.326846122741699} 01/28/2022 01:00:17 - INFO - codeparrot_training - Step 4903: {'lr': 0.0004955010160170667, 'samples': 941568, 'steps': 4903, 'loss/train': 2.223512828350067} 01/28/2022 01:00:22 - INFO - codeparrot_training - Step 4904: {'lr': 0.0004954979252802491, 'samples': 941760, 'steps': 4904, 'loss/train': 2.9832944869995117} 01/28/2022 01:00:26 - INFO - codeparrot_training - Step 4905: {'lr': 0.0004954948334917965, 'samples': 941952, 'steps': 4905, 'loss/train': 3.800092577934265} 01/28/2022 01:00:31 - INFO - codeparrot_training - Step 4906: {'lr': 0.0004954917406517218, 'samples': 942144, 'steps': 4906, 'loss/train': 3.6834908723831177} 01/28/2022 01:00:35 - INFO - codeparrot_training - Step 4907: {'lr': 0.0004954886467600386, 'samples': 942336, 'steps': 4907, 'loss/train': 2.5084800124168396} 01/28/2022 01:00:39 - INFO - codeparrot_training - Step 4908: {'lr': 0.0004954855518167599, 'samples': 942528, 'steps': 4908, 'loss/train': 3.230789065361023} 01/28/2022 01:00:45 - INFO - codeparrot_training - Step 4909: {'lr': 0.000495482455821899, 'samples': 942720, 'steps': 4909, 'loss/train': 1.9065772891044617} 01/28/2022 01:00:49 - INFO - codeparrot_training - Step 4910: {'lr': 0.0004954793587754694, 'samples': 942912, 'steps': 4910, 'loss/train': 3.67567777633667} 01/28/2022 01:00:54 - INFO - codeparrot_training - Step 4911: {'lr': 0.000495476260677484, 'samples': 943104, 'steps': 4911, 'loss/train': 2.390337109565735} 01/28/2022 01:00:58 - INFO - codeparrot_training - Step 4912: {'lr': 0.0004954731615279563, 'samples': 943296, 'steps': 4912, 'loss/train': 1.9164751768112183} 01/28/2022 01:01:02 - INFO - codeparrot_training - Step 4913: {'lr': 0.0004954700613268995, 'samples': 943488, 'steps': 4913, 'loss/train': 1.682116985321045} 01/28/2022 01:01:07 - INFO - codeparrot_training - Step 4914: {'lr': 0.0004954669600743269, 'samples': 943680, 'steps': 4914, 'loss/train': 2.8326446413993835} 01/28/2022 01:01:12 - INFO - codeparrot_training - Step 4915: {'lr': 0.0004954638577702519, 'samples': 943872, 'steps': 4915, 'loss/train': 2.8024531602859497} 01/28/2022 01:01:16 - INFO - codeparrot_training - Step 4916: {'lr': 0.0004954607544146875, 'samples': 944064, 'steps': 4916, 'loss/train': 2.4824488162994385} 01/28/2022 01:01:20 - INFO - codeparrot_training - Step 4917: {'lr': 0.0004954576500076472, 'samples': 944256, 'steps': 4917, 'loss/train': 3.091994047164917} 01/28/2022 01:01:25 - INFO - codeparrot_training - Step 4918: {'lr': 0.0004954545445491444, 'samples': 944448, 'steps': 4918, 'loss/train': 2.6687121391296387} 01/28/2022 01:01:30 - INFO - codeparrot_training - Step 4919: {'lr': 0.0004954514380391921, 'samples': 944640, 'steps': 4919, 'loss/train': 1.6967436075210571} 01/28/2022 01:01:34 - INFO - codeparrot_training - Step 4920: {'lr': 0.0004954483304778039, 'samples': 944832, 'steps': 4920, 'loss/train': 3.1467851400375366} 01/28/2022 01:01:38 - INFO - codeparrot_training - Step 4921: {'lr': 0.0004954452218649929, 'samples': 945024, 'steps': 4921, 'loss/train': 2.6691445112228394} 01/28/2022 01:01:43 - INFO - codeparrot_training - Step 4922: {'lr': 0.0004954421122007727, 'samples': 945216, 'steps': 4922, 'loss/train': 3.3028271198272705} 01/28/2022 01:01:47 - INFO - codeparrot_training - Step 4923: {'lr': 0.0004954390014851563, 'samples': 945408, 'steps': 4923, 'loss/train': 3.143077254295349} 01/28/2022 01:01:53 - INFO - codeparrot_training - Step 4924: {'lr': 0.0004954358897181571, 'samples': 945600, 'steps': 4924, 'loss/train': 1.1373765170574188} 01/28/2022 01:01:57 - INFO - codeparrot_training - Step 4925: {'lr': 0.0004954327768997885, 'samples': 945792, 'steps': 4925, 'loss/train': 3.139392614364624} 01/28/2022 01:02:01 - INFO - codeparrot_training - Step 4926: {'lr': 0.0004954296630300638, 'samples': 945984, 'steps': 4926, 'loss/train': 3.3592203855514526} 01/28/2022 01:02:05 - INFO - codeparrot_training - Step 4927: {'lr': 0.0004954265481089965, 'samples': 946176, 'steps': 4927, 'loss/train': 3.1383694410324097} 01/28/2022 01:02:10 - INFO - codeparrot_training - Step 4928: {'lr': 0.0004954234321365998, 'samples': 946368, 'steps': 4928, 'loss/train': 3.1229413747787476} 01/28/2022 01:02:15 - INFO - codeparrot_training - Step 4929: {'lr': 0.0004954203151128868, 'samples': 946560, 'steps': 4929, 'loss/train': 2.5884456038475037} 01/28/2022 01:02:19 - INFO - codeparrot_training - Step 4930: {'lr': 0.0004954171970378713, 'samples': 946752, 'steps': 4930, 'loss/train': 3.212198853492737} 01/28/2022 01:02:23 - INFO - codeparrot_training - Step 4931: {'lr': 0.0004954140779115664, 'samples': 946944, 'steps': 4931, 'loss/train': 3.0822383165359497} 01/28/2022 01:02:27 - INFO - codeparrot_training - Step 4932: {'lr': 0.0004954109577339856, 'samples': 947136, 'steps': 4932, 'loss/train': 1.9040173888206482} 01/28/2022 01:02:32 - INFO - codeparrot_training - Step 4933: {'lr': 0.0004954078365051421, 'samples': 947328, 'steps': 4933, 'loss/train': 3.4057559967041016} 01/28/2022 01:02:38 - INFO - codeparrot_training - Step 4934: {'lr': 0.0004954047142250494, 'samples': 947520, 'steps': 4934, 'loss/train': 3.5429495573043823} 01/28/2022 01:02:42 - INFO - codeparrot_training - Step 4935: {'lr': 0.0004954015908937208, 'samples': 947712, 'steps': 4935, 'loss/train': 3.210487961769104} 01/28/2022 01:02:46 - INFO - codeparrot_training - Step 4936: {'lr': 0.0004953984665111697, 'samples': 947904, 'steps': 4936, 'loss/train': 2.654325306415558} 01/28/2022 01:02:51 - INFO - codeparrot_training - Step 4937: {'lr': 0.0004953953410774095, 'samples': 948096, 'steps': 4937, 'loss/train': 3.2515697479248047} 01/28/2022 01:02:56 - INFO - codeparrot_training - Step 4938: {'lr': 0.0004953922145924535, 'samples': 948288, 'steps': 4938, 'loss/train': 3.1722503900527954} 01/28/2022 01:03:00 - INFO - codeparrot_training - Step 4939: {'lr': 0.0004953890870563153, 'samples': 948480, 'steps': 4939, 'loss/train': 3.130582094192505} 01/28/2022 01:03:04 - INFO - codeparrot_training - Step 4940: {'lr': 0.0004953859584690081, 'samples': 948672, 'steps': 4940, 'loss/train': 2.1027191877365112} 01/28/2022 01:03:08 - INFO - codeparrot_training - Step 4941: {'lr': 0.0004953828288305454, 'samples': 948864, 'steps': 4941, 'loss/train': 2.2034996151924133} 01/28/2022 01:03:13 - INFO - codeparrot_training - Step 4942: {'lr': 0.0004953796981409407, 'samples': 949056, 'steps': 4942, 'loss/train': 3.5779999494552612} 01/28/2022 01:03:18 - INFO - codeparrot_training - Step 4943: {'lr': 0.0004953765664002071, 'samples': 949248, 'steps': 4943, 'loss/train': 3.655552625656128} 01/28/2022 01:03:22 - INFO - codeparrot_training - Step 4944: {'lr': 0.0004953734336083582, 'samples': 949440, 'steps': 4944, 'loss/train': 3.051100730895996} 01/28/2022 01:03:27 - INFO - codeparrot_training - Step 4945: {'lr': 0.0004953702997654076, 'samples': 949632, 'steps': 4945, 'loss/train': 3.164203405380249} 01/28/2022 01:03:31 - INFO - codeparrot_training - Step 4946: {'lr': 0.0004953671648713683, 'samples': 949824, 'steps': 4946, 'loss/train': 1.5033103823661804} 01/28/2022 01:03:35 - INFO - codeparrot_training - Step 4947: {'lr': 0.0004953640289262542, 'samples': 950016, 'steps': 4947, 'loss/train': 2.283093273639679} 01/28/2022 01:03:40 - INFO - codeparrot_training - Step 4948: {'lr': 0.0004953608919300784, 'samples': 950208, 'steps': 4948, 'loss/train': 1.7821453213691711} 01/28/2022 01:03:44 - INFO - codeparrot_training - Step 4949: {'lr': 0.0004953577538828546, 'samples': 950400, 'steps': 4949, 'loss/train': 3.0127322673797607} 01/28/2022 01:03:49 - INFO - codeparrot_training - Step 4950: {'lr': 0.0004953546147845959, 'samples': 950592, 'steps': 4950, 'loss/train': 3.3560296297073364} 01/28/2022 01:03:53 - INFO - codeparrot_training - Step 4951: {'lr': 0.0004953514746353161, 'samples': 950784, 'steps': 4951, 'loss/train': 3.439594030380249} 01/28/2022 01:03:57 - INFO - codeparrot_training - Step 4952: {'lr': 0.0004953483334350283, 'samples': 950976, 'steps': 4952, 'loss/train': 2.1462231874465942} 01/28/2022 01:04:03 - INFO - codeparrot_training - Step 4953: {'lr': 0.0004953451911837463, 'samples': 951168, 'steps': 4953, 'loss/train': 2.0510491132736206} 01/28/2022 01:04:07 - INFO - codeparrot_training - Step 4954: {'lr': 0.0004953420478814834, 'samples': 951360, 'steps': 4954, 'loss/train': 2.030945062637329} 01/28/2022 01:04:11 - INFO - codeparrot_training - Step 4955: {'lr': 0.000495338903528253, 'samples': 951552, 'steps': 4955, 'loss/train': 1.6427582502365112} 01/28/2022 01:04:16 - INFO - codeparrot_training - Step 4956: {'lr': 0.0004953357581240686, 'samples': 951744, 'steps': 4956, 'loss/train': 1.7370307445526123} 01/28/2022 01:04:20 - INFO - codeparrot_training - Step 4957: {'lr': 0.0004953326116689438, 'samples': 951936, 'steps': 4957, 'loss/train': 3.363122820854187} 01/28/2022 01:04:24 - INFO - codeparrot_training - Step 4958: {'lr': 0.000495329464162892, 'samples': 952128, 'steps': 4958, 'loss/train': 3.4379782676696777} 01/28/2022 01:04:29 - INFO - codeparrot_training - Step 4959: {'lr': 0.0004953263156059266, 'samples': 952320, 'steps': 4959, 'loss/train': 2.447881579399109} 01/28/2022 01:04:33 - INFO - codeparrot_training - Step 4960: {'lr': 0.0004953231659980613, 'samples': 952512, 'steps': 4960, 'loss/train': 1.607482373714447} 01/28/2022 01:04:38 - INFO - codeparrot_training - Step 4961: {'lr': 0.0004953200153393094, 'samples': 952704, 'steps': 4961, 'loss/train': 3.8963195085525513} 01/28/2022 01:04:42 - INFO - codeparrot_training - Step 4962: {'lr': 0.0004953168636296845, 'samples': 952896, 'steps': 4962, 'loss/train': 2.900760591030121} 01/28/2022 01:04:46 - INFO - codeparrot_training - Step 4963: {'lr': 0.0004953137108691999, 'samples': 953088, 'steps': 4963, 'loss/train': 2.238089382648468} 01/28/2022 01:04:51 - INFO - codeparrot_training - Step 4964: {'lr': 0.0004953105570578693, 'samples': 953280, 'steps': 4964, 'loss/train': 2.6513713002204895} 01/28/2022 01:04:56 - INFO - codeparrot_training - Step 4965: {'lr': 0.0004953074021957063, 'samples': 953472, 'steps': 4965, 'loss/train': 2.0937756299972534} 01/28/2022 01:05:00 - INFO - codeparrot_training - Step 4966: {'lr': 0.0004953042462827242, 'samples': 953664, 'steps': 4966, 'loss/train': 3.568637251853943} 01/28/2022 01:05:04 - INFO - codeparrot_training - Step 4967: {'lr': 0.0004953010893189365, 'samples': 953856, 'steps': 4967, 'loss/train': 2.398638904094696} 01/28/2022 01:05:08 - INFO - codeparrot_training - Step 4968: {'lr': 0.000495297931304357, 'samples': 954048, 'steps': 4968, 'loss/train': 1.8135298490524292} 01/28/2022 01:05:14 - INFO - codeparrot_training - Step 4969: {'lr': 0.000495294772238999, 'samples': 954240, 'steps': 4969, 'loss/train': 2.2856116890907288} 01/28/2022 01:05:18 - INFO - codeparrot_training - Step 4970: {'lr': 0.000495291612122876, 'samples': 954432, 'steps': 4970, 'loss/train': 2.5935473442077637} 01/28/2022 01:05:23 - INFO - codeparrot_training - Step 4971: {'lr': 0.0004952884509560017, 'samples': 954624, 'steps': 4971, 'loss/train': 1.4720814228057861} 01/28/2022 01:05:27 - INFO - codeparrot_training - Step 4972: {'lr': 0.0004952852887383895, 'samples': 954816, 'steps': 4972, 'loss/train': 2.5210949778556824} 01/28/2022 01:05:32 - INFO - codeparrot_training - Step 4973: {'lr': 0.0004952821254700531, 'samples': 955008, 'steps': 4973, 'loss/train': 2.808197557926178} 01/28/2022 01:05:37 - INFO - codeparrot_training - Step 4974: {'lr': 0.0004952789611510059, 'samples': 955200, 'steps': 4974, 'loss/train': 1.977605938911438} 01/28/2022 01:05:41 - INFO - codeparrot_training - Step 4975: {'lr': 0.0004952757957812615, 'samples': 955392, 'steps': 4975, 'loss/train': 1.4793849885463715} 01/28/2022 01:05:45 - INFO - codeparrot_training - Step 4976: {'lr': 0.0004952726293608335, 'samples': 955584, 'steps': 4976, 'loss/train': 4.3424752950668335} 01/28/2022 01:05:49 - INFO - codeparrot_training - Step 4977: {'lr': 0.0004952694618897354, 'samples': 955776, 'steps': 4977, 'loss/train': 3.7317806482315063} 01/28/2022 01:05:55 - INFO - codeparrot_training - Step 4978: {'lr': 0.0004952662933679809, 'samples': 955968, 'steps': 4978, 'loss/train': 2.2800984978675842} 01/28/2022 01:06:00 - INFO - codeparrot_training - Step 4979: {'lr': 0.0004952631237955835, 'samples': 956160, 'steps': 4979, 'loss/train': 2.3618892431259155} 01/28/2022 01:06:04 - INFO - codeparrot_training - Step 4980: {'lr': 0.0004952599531725567, 'samples': 956352, 'steps': 4980, 'loss/train': 3.3762402534484863} 01/28/2022 01:06:08 - INFO - codeparrot_training - Step 4981: {'lr': 0.0004952567814989141, 'samples': 956544, 'steps': 4981, 'loss/train': 3.2810815572738647} 01/28/2022 01:06:12 - INFO - codeparrot_training - Step 4982: {'lr': 0.0004952536087746693, 'samples': 956736, 'steps': 4982, 'loss/train': 3.5129964351654053} 01/28/2022 01:06:17 - INFO - codeparrot_training - Step 4983: {'lr': 0.000495250434999836, 'samples': 956928, 'steps': 4983, 'loss/train': 9.02471923828125} 01/28/2022 01:06:22 - INFO - codeparrot_training - Step 4984: {'lr': 0.0004952472601744277, 'samples': 957120, 'steps': 4984, 'loss/train': 1.5715667009353638} 01/28/2022 01:06:26 - INFO - codeparrot_training - Step 4985: {'lr': 0.000495244084298458, 'samples': 957312, 'steps': 4985, 'loss/train': 3.2517510652542114} 01/28/2022 01:06:30 - INFO - codeparrot_training - Step 4986: {'lr': 0.0004952409073719405, 'samples': 957504, 'steps': 4986, 'loss/train': 2.9353618025779724} 01/28/2022 01:06:34 - INFO - codeparrot_training - Step 4987: {'lr': 0.0004952377293948888, 'samples': 957696, 'steps': 4987, 'loss/train': 1.7365858554840088} 01/28/2022 01:06:40 - INFO - codeparrot_training - Step 4988: {'lr': 0.0004952345503673166, 'samples': 957888, 'steps': 4988, 'loss/train': 2.704836130142212} 01/28/2022 01:06:44 - INFO - codeparrot_training - Step 4989: {'lr': 0.0004952313702892375, 'samples': 958080, 'steps': 4989, 'loss/train': 2.239686369895935} 01/28/2022 01:06:48 - INFO - codeparrot_training - Step 4990: {'lr': 0.0004952281891606649, 'samples': 958272, 'steps': 4990, 'loss/train': 2.443862557411194} 01/28/2022 01:06:52 - INFO - codeparrot_training - Step 4991: {'lr': 0.0004952250069816127, 'samples': 958464, 'steps': 4991, 'loss/train': 2.787310302257538} 01/28/2022 01:06:57 - INFO - codeparrot_training - Step 4992: {'lr': 0.0004952218237520945, 'samples': 958656, 'steps': 4992, 'loss/train': 3.61231005191803} 01/28/2022 01:07:02 - INFO - codeparrot_training - Step 4993: {'lr': 0.0004952186394721239, 'samples': 958848, 'steps': 4993, 'loss/train': 2.8653600811958313} 01/28/2022 01:07:07 - INFO - codeparrot_training - Step 4994: {'lr': 0.0004952154541417144, 'samples': 959040, 'steps': 4994, 'loss/train': 4.417145490646362} 01/28/2022 01:07:11 - INFO - codeparrot_training - Step 4995: {'lr': 0.0004952122677608798, 'samples': 959232, 'steps': 4995, 'loss/train': 4.315163254737854} 01/28/2022 01:07:15 - INFO - codeparrot_training - Step 4996: {'lr': 0.0004952090803296337, 'samples': 959424, 'steps': 4996, 'loss/train': 1.5523139834403992} 01/28/2022 01:07:19 - INFO - codeparrot_training - Step 4997: {'lr': 0.0004952058918479899, 'samples': 959616, 'steps': 4997, 'loss/train': 3.268303871154785} 01/28/2022 01:07:25 - INFO - codeparrot_training - Step 4998: {'lr': 0.0004952027023159617, 'samples': 959808, 'steps': 4998, 'loss/train': 2.7825295329093933} 01/28/2022 01:07:29 - INFO - codeparrot_training - Step 4999: {'lr': 0.0004951995117335631, 'samples': 960000, 'steps': 4999, 'loss/train': 0.8443890810012817} 01/28/2022 01:07:33 - INFO - codeparrot_training - Step 5000: {'lr': 0.0004951963201008077, 'samples': 960192, 'steps': 5000, 'loss/train': 2.200170636177063} 01/28/2022 01:07:37 - INFO - codeparrot_training - Step 5001: {'lr': 0.000495193127417709, 'samples': 960384, 'steps': 5001, 'loss/train': 0.44876839220523834} 01/28/2022 01:07:42 - INFO - codeparrot_training - Step 5002: {'lr': 0.0004951899336842809, 'samples': 960576, 'steps': 5002, 'loss/train': 1.6333104372024536} 01/28/2022 01:07:47 - INFO - codeparrot_training - Step 5003: {'lr': 0.0004951867389005369, 'samples': 960768, 'steps': 5003, 'loss/train': 1.6933432817459106} 01/28/2022 01:07:51 - INFO - codeparrot_training - Step 5004: {'lr': 0.0004951835430664908, 'samples': 960960, 'steps': 5004, 'loss/train': 2.8271886706352234} 01/28/2022 01:07:55 - INFO - codeparrot_training - Step 5005: {'lr': 0.0004951803461821562, 'samples': 961152, 'steps': 5005, 'loss/train': 2.7272411584854126} 01/28/2022 01:08:00 - INFO - codeparrot_training - Step 5006: {'lr': 0.0004951771482475469, 'samples': 961344, 'steps': 5006, 'loss/train': 2.3385438323020935} 01/28/2022 01:08:04 - INFO - codeparrot_training - Step 5007: {'lr': 0.0004951739492626766, 'samples': 961536, 'steps': 5007, 'loss/train': 2.4105563163757324} 01/28/2022 01:08:10 - INFO - codeparrot_training - Step 5008: {'lr': 0.0004951707492275589, 'samples': 961728, 'steps': 5008, 'loss/train': 2.399742901325226} 01/28/2022 01:08:14 - INFO - codeparrot_training - Step 5009: {'lr': 0.0004951675481422075, 'samples': 961920, 'steps': 5009, 'loss/train': 3.516133189201355} 01/28/2022 01:08:18 - INFO - codeparrot_training - Step 5010: {'lr': 0.0004951643460066363, 'samples': 962112, 'steps': 5010, 'loss/train': 2.5928739309310913} 01/28/2022 01:08:23 - INFO - codeparrot_training - Step 5011: {'lr': 0.0004951611428208589, 'samples': 962304, 'steps': 5011, 'loss/train': 3.154770255088806} 01/28/2022 01:08:27 - INFO - codeparrot_training - Step 5012: {'lr': 0.0004951579385848889, 'samples': 962496, 'steps': 5012, 'loss/train': 3.697226643562317} 01/28/2022 01:08:32 - INFO - codeparrot_training - Step 5013: {'lr': 0.0004951547332987401, 'samples': 962688, 'steps': 5013, 'loss/train': 2.727745234966278} 01/28/2022 01:08:36 - INFO - codeparrot_training - Step 5014: {'lr': 0.0004951515269624265, 'samples': 962880, 'steps': 5014, 'loss/train': 4.099005103111267} 01/28/2022 01:08:40 - INFO - codeparrot_training - Step 5015: {'lr': 0.0004951483195759614, 'samples': 963072, 'steps': 5015, 'loss/train': 2.7492823004722595} 01/28/2022 01:08:45 - INFO - codeparrot_training - Step 5016: {'lr': 0.0004951451111393588, 'samples': 963264, 'steps': 5016, 'loss/train': 3.673882484436035} 01/28/2022 01:08:49 - INFO - codeparrot_training - Step 5017: {'lr': 0.0004951419016526324, 'samples': 963456, 'steps': 5017, 'loss/train': 4.310345649719238} 01/28/2022 01:08:54 - INFO - codeparrot_training - Step 5018: {'lr': 0.0004951386911157959, 'samples': 963648, 'steps': 5018, 'loss/train': 2.804174780845642} 01/28/2022 01:08:58 - INFO - codeparrot_training - Step 5019: {'lr': 0.0004951354795288631, 'samples': 963840, 'steps': 5019, 'loss/train': 3.203449845314026} 01/28/2022 01:09:03 - INFO - codeparrot_training - Step 5020: {'lr': 0.0004951322668918477, 'samples': 964032, 'steps': 5020, 'loss/train': 3.439982056617737} 01/28/2022 01:09:07 - INFO - codeparrot_training - Step 5021: {'lr': 0.0004951290532047637, 'samples': 964224, 'steps': 5021, 'loss/train': 2.55614572763443} 01/28/2022 01:09:11 - INFO - codeparrot_training - Step 5022: {'lr': 0.0004951258384676244, 'samples': 964416, 'steps': 5022, 'loss/train': 2.9474818110466003} 01/28/2022 01:09:17 - INFO - codeparrot_training - Step 5023: {'lr': 0.0004951226226804441, 'samples': 964608, 'steps': 5023, 'loss/train': 1.7938483357429504} 01/28/2022 01:09:21 - INFO - codeparrot_training - Step 5024: {'lr': 0.0004951194058432361, 'samples': 964800, 'steps': 5024, 'loss/train': 1.9312750697135925} 01/28/2022 01:09:25 - INFO - codeparrot_training - Step 5025: {'lr': 0.0004951161879560146, 'samples': 964992, 'steps': 5025, 'loss/train': 3.3742843866348267} 01/28/2022 01:09:29 - INFO - codeparrot_training - Step 5026: {'lr': 0.000495112969018793, 'samples': 965184, 'steps': 5026, 'loss/train': 2.7258020639419556} 01/28/2022 01:09:34 - INFO - codeparrot_training - Step 5027: {'lr': 0.0004951097490315853, 'samples': 965376, 'steps': 5027, 'loss/train': 1.815821349620819} 01/28/2022 01:09:39 - INFO - codeparrot_training - Step 5028: {'lr': 0.0004951065279944054, 'samples': 965568, 'steps': 5028, 'loss/train': 3.2975428104400635} 01/28/2022 01:09:43 - INFO - codeparrot_training - Step 5029: {'lr': 0.0004951033059072668, 'samples': 965760, 'steps': 5029, 'loss/train': 3.4694734811782837} 01/28/2022 01:09:47 - INFO - codeparrot_training - Step 5030: {'lr': 0.0004951000827701836, 'samples': 965952, 'steps': 5030, 'loss/train': 2.2466878294944763} 01/28/2022 01:09:51 - INFO - codeparrot_training - Step 5031: {'lr': 0.0004950968585831694, 'samples': 966144, 'steps': 5031, 'loss/train': 1.3600529730319977} 01/28/2022 01:09:56 - INFO - codeparrot_training - Step 5032: {'lr': 0.0004950936333462381, 'samples': 966336, 'steps': 5032, 'loss/train': 1.5137502551078796} 01/28/2022 01:10:01 - INFO - codeparrot_training - Step 5033: {'lr': 0.0004950904070594036, 'samples': 966528, 'steps': 5033, 'loss/train': 1.5761441588401794} 01/28/2022 01:10:05 - INFO - codeparrot_training - Step 5034: {'lr': 0.0004950871797226795, 'samples': 966720, 'steps': 5034, 'loss/train': 3.1063857078552246} 01/28/2022 01:10:09 - INFO - codeparrot_training - Step 5035: {'lr': 0.0004950839513360798, 'samples': 966912, 'steps': 5035, 'loss/train': 2.725900948047638} 01/28/2022 01:10:14 - INFO - codeparrot_training - Step 5036: {'lr': 0.0004950807218996182, 'samples': 967104, 'steps': 5036, 'loss/train': 2.472670555114746} 01/28/2022 01:10:18 - INFO - codeparrot_training - Step 5037: {'lr': 0.0004950774914133086, 'samples': 967296, 'steps': 5037, 'loss/train': 2.8557037711143494} 01/28/2022 01:10:24 - INFO - codeparrot_training - Step 5038: {'lr': 0.0004950742598771649, 'samples': 967488, 'steps': 5038, 'loss/train': 3.3863028287887573} 01/28/2022 01:10:28 - INFO - codeparrot_training - Step 5039: {'lr': 0.0004950710272912009, 'samples': 967680, 'steps': 5039, 'loss/train': 2.2745702862739563} 01/28/2022 01:10:32 - INFO - codeparrot_training - Step 5040: {'lr': 0.0004950677936554305, 'samples': 967872, 'steps': 5040, 'loss/train': 3.3772798776626587} 01/28/2022 01:10:36 - INFO - codeparrot_training - Step 5041: {'lr': 0.0004950645589698674, 'samples': 968064, 'steps': 5041, 'loss/train': 2.1654404997825623} 01/28/2022 01:10:41 - INFO - codeparrot_training - Step 5042: {'lr': 0.0004950613232345256, 'samples': 968256, 'steps': 5042, 'loss/train': 3.283331036567688} 01/28/2022 01:10:46 - INFO - codeparrot_training - Step 5043: {'lr': 0.0004950580864494188, 'samples': 968448, 'steps': 5043, 'loss/train': 2.884580075740814} 01/28/2022 01:10:51 - INFO - codeparrot_training - Step 5044: {'lr': 0.0004950548486145611, 'samples': 968640, 'steps': 5044, 'loss/train': 2.3259469270706177} 01/28/2022 01:10:55 - INFO - codeparrot_training - Step 5045: {'lr': 0.0004950516097299662, 'samples': 968832, 'steps': 5045, 'loss/train': 1.2597236037254333} 01/28/2022 01:10:59 - INFO - codeparrot_training - Step 5046: {'lr': 0.000495048369795648, 'samples': 969024, 'steps': 5046, 'loss/train': 2.6201958060264587} 01/28/2022 01:11:04 - INFO - codeparrot_training - Step 5047: {'lr': 0.0004950451288116204, 'samples': 969216, 'steps': 5047, 'loss/train': 3.1426631212234497} 01/28/2022 01:11:08 - INFO - codeparrot_training - Step 5048: {'lr': 0.0004950418867778973, 'samples': 969408, 'steps': 5048, 'loss/train': 2.742810845375061} 01/28/2022 01:11:13 - INFO - codeparrot_training - Step 5049: {'lr': 0.0004950386436944925, 'samples': 969600, 'steps': 5049, 'loss/train': 3.0922908782958984} 01/28/2022 01:11:17 - INFO - codeparrot_training - Step 5050: {'lr': 0.0004950353995614201, 'samples': 969792, 'steps': 5050, 'loss/train': 2.1935473680496216} 01/28/2022 01:11:21 - INFO - codeparrot_training - Step 5051: {'lr': 0.0004950321543786937, 'samples': 969984, 'steps': 5051, 'loss/train': 1.4589693546295166} 01/28/2022 01:11:27 - INFO - codeparrot_training - Step 5052: {'lr': 0.0004950289081463273, 'samples': 970176, 'steps': 5052, 'loss/train': 3.0610713958740234} 01/28/2022 01:11:31 - INFO - codeparrot_training - Step 5053: {'lr': 0.0004950256608643351, 'samples': 970368, 'steps': 5053, 'loss/train': 2.8343957662582397} 01/28/2022 01:11:35 - INFO - codeparrot_training - Step 5054: {'lr': 0.0004950224125327307, 'samples': 970560, 'steps': 5054, 'loss/train': 2.698632538318634} 01/28/2022 01:11:40 - INFO - codeparrot_training - Step 5055: {'lr': 0.000495019163151528, 'samples': 970752, 'steps': 5055, 'loss/train': 2.0482242107391357} 01/28/2022 01:11:44 - INFO - codeparrot_training - Step 5056: {'lr': 0.0004950159127207411, 'samples': 970944, 'steps': 5056, 'loss/train': 2.8202674984931946} 01/28/2022 01:11:49 - INFO - codeparrot_training - Step 5057: {'lr': 0.0004950126612403838, 'samples': 971136, 'steps': 5057, 'loss/train': 2.931077778339386} 01/28/2022 01:11:53 - INFO - codeparrot_training - Step 5058: {'lr': 0.00049500940871047, 'samples': 971328, 'steps': 5058, 'loss/train': 4.7674620151519775} 01/28/2022 01:11:58 - INFO - codeparrot_training - Step 5059: {'lr': 0.0004950061551310138, 'samples': 971520, 'steps': 5059, 'loss/train': 3.1543679237365723} 01/28/2022 01:12:02 - INFO - codeparrot_training - Step 5060: {'lr': 0.0004950029005020289, 'samples': 971712, 'steps': 5060, 'loss/train': 3.012622117996216} 01/28/2022 01:12:06 - INFO - codeparrot_training - Step 5061: {'lr': 0.0004949996448235294, 'samples': 971904, 'steps': 5061, 'loss/train': 3.12189781665802} 01/28/2022 01:12:11 - INFO - codeparrot_training - Step 5062: {'lr': 0.0004949963880955293, 'samples': 972096, 'steps': 5062, 'loss/train': 3.972660541534424} 01/28/2022 01:12:16 - INFO - codeparrot_training - Step 5063: {'lr': 0.0004949931303180424, 'samples': 972288, 'steps': 5063, 'loss/train': 5.000612854957581} 01/28/2022 01:12:20 - INFO - codeparrot_training - Step 5064: {'lr': 0.0004949898714910828, 'samples': 972480, 'steps': 5064, 'loss/train': 3.851137161254883} 01/28/2022 01:12:24 - INFO - codeparrot_training - Step 5065: {'lr': 0.0004949866116146643, 'samples': 972672, 'steps': 5065, 'loss/train': 2.9380284547805786} 01/28/2022 01:12:28 - INFO - codeparrot_training - Step 5066: {'lr': 0.000494983350688801, 'samples': 972864, 'steps': 5066, 'loss/train': 3.3480101823806763} 01/28/2022 01:12:34 - INFO - codeparrot_training - Step 5067: {'lr': 0.0004949800887135067, 'samples': 973056, 'steps': 5067, 'loss/train': 2.8463659286499023} 01/28/2022 01:12:38 - INFO - codeparrot_training - Step 5068: {'lr': 0.0004949768256887956, 'samples': 973248, 'steps': 5068, 'loss/train': 10.30515718460083} 01/28/2022 01:12:42 - INFO - codeparrot_training - Step 5069: {'lr': 0.0004949735616146816, 'samples': 973440, 'steps': 5069, 'loss/train': 0.5696472376585007} 01/28/2022 01:12:46 - INFO - codeparrot_training - Step 5070: {'lr': 0.0004949702964911787, 'samples': 973632, 'steps': 5070, 'loss/train': 4.607659578323364} 01/28/2022 01:12:51 - INFO - codeparrot_training - Step 5071: {'lr': 0.0004949670303183006, 'samples': 973824, 'steps': 5071, 'loss/train': 3.1772539615631104} 01/28/2022 01:12:55 - INFO - codeparrot_training - Step 5072: {'lr': 0.0004949637630960618, 'samples': 974016, 'steps': 5072, 'loss/train': 1.9166496992111206} 01/28/2022 01:13:01 - INFO - codeparrot_training - Step 5073: {'lr': 0.0004949604948244758, 'samples': 974208, 'steps': 5073, 'loss/train': 3.7034565210342407} 01/28/2022 01:13:06 - INFO - codeparrot_training - Step 5074: {'lr': 0.0004949572255035569, 'samples': 974400, 'steps': 5074, 'loss/train': 2.9666420817375183} 01/28/2022 01:13:10 - INFO - codeparrot_training - Step 5075: {'lr': 0.0004949539551333191, 'samples': 974592, 'steps': 5075, 'loss/train': 3.6851806640625} 01/28/2022 01:13:14 - INFO - codeparrot_training - Step 5076: {'lr': 0.0004949506837137763, 'samples': 974784, 'steps': 5076, 'loss/train': 3.186138868331909} 01/28/2022 01:13:18 - INFO - codeparrot_training - Step 5077: {'lr': 0.0004949474112449424, 'samples': 974976, 'steps': 5077, 'loss/train': 2.9104769825935364} 01/28/2022 01:13:24 - INFO - codeparrot_training - Step 5078: {'lr': 0.0004949441377268318, 'samples': 975168, 'steps': 5078, 'loss/train': 2.3596732020378113} 01/28/2022 01:13:28 - INFO - codeparrot_training - Step 5079: {'lr': 0.0004949408631594582, 'samples': 975360, 'steps': 5079, 'loss/train': 3.0721582174301147} 01/28/2022 01:13:32 - INFO - codeparrot_training - Step 5080: {'lr': 0.0004949375875428357, 'samples': 975552, 'steps': 5080, 'loss/train': 3.3537293672561646} 01/28/2022 01:13:36 - INFO - codeparrot_training - Step 5081: {'lr': 0.0004949343108769784, 'samples': 975744, 'steps': 5081, 'loss/train': 0.6237917393445969} 01/28/2022 01:13:40 - INFO - codeparrot_training - Step 5082: {'lr': 0.0004949310331619002, 'samples': 975936, 'steps': 5082, 'loss/train': 3.5127525329589844} 01/28/2022 01:13:47 - INFO - codeparrot_training - Step 5083: {'lr': 0.0004949277543976153, 'samples': 976128, 'steps': 5083, 'loss/train': 4.118197202682495} 01/28/2022 01:13:51 - INFO - codeparrot_training - Step 5084: {'lr': 0.0004949244745841377, 'samples': 976320, 'steps': 5084, 'loss/train': 2.8743204474449158} 01/28/2022 01:13:55 - INFO - codeparrot_training - Step 5085: {'lr': 0.0004949211937214814, 'samples': 976512, 'steps': 5085, 'loss/train': 1.6654030680656433} 01/28/2022 01:14:00 - INFO - codeparrot_training - Step 5086: {'lr': 0.0004949179118096604, 'samples': 976704, 'steps': 5086, 'loss/train': 2.4127694964408875} 01/28/2022 01:14:04 - INFO - codeparrot_training - Step 5087: {'lr': 0.0004949146288486889, 'samples': 976896, 'steps': 5087, 'loss/train': 3.9442198276519775} 01/28/2022 01:14:09 - INFO - codeparrot_training - Step 5088: {'lr': 0.0004949113448385809, 'samples': 977088, 'steps': 5088, 'loss/train': 3.6778539419174194} 01/28/2022 01:14:13 - INFO - codeparrot_training - Step 5089: {'lr': 0.0004949080597793505, 'samples': 977280, 'steps': 5089, 'loss/train': 1.3911442458629608} 01/28/2022 01:14:18 - INFO - codeparrot_training - Step 5090: {'lr': 0.0004949047736710116, 'samples': 977472, 'steps': 5090, 'loss/train': 2.6096511483192444} 01/28/2022 01:14:22 - INFO - codeparrot_training - Step 5091: {'lr': 0.0004949014865135786, 'samples': 977664, 'steps': 5091, 'loss/train': 3.4322354793548584} 01/28/2022 01:14:26 - INFO - codeparrot_training - Step 5092: {'lr': 0.0004948981983070652, 'samples': 977856, 'steps': 5092, 'loss/train': 2.906498372554779} 01/28/2022 01:14:31 - INFO - codeparrot_training - Step 5093: {'lr': 0.0004948949090514858, 'samples': 978048, 'steps': 5093, 'loss/train': 3.2582770586013794} 01/28/2022 01:14:36 - INFO - codeparrot_training - Step 5094: {'lr': 0.0004948916187468544, 'samples': 978240, 'steps': 5094, 'loss/train': 3.276276469230652} 01/28/2022 01:14:40 - INFO - codeparrot_training - Step 5095: {'lr': 0.000494888327393185, 'samples': 978432, 'steps': 5095, 'loss/train': 2.698039948940277} 01/28/2022 01:14:44 - INFO - codeparrot_training - Step 5096: {'lr': 0.0004948850349904919, 'samples': 978624, 'steps': 5096, 'loss/train': 3.3259352445602417} 01/28/2022 01:14:48 - INFO - codeparrot_training - Step 5097: {'lr': 0.000494881741538789, 'samples': 978816, 'steps': 5097, 'loss/train': 4.4115214347839355} 01/28/2022 01:14:54 - INFO - codeparrot_training - Step 5098: {'lr': 0.0004948784470380904, 'samples': 979008, 'steps': 5098, 'loss/train': 3.128661274909973} 01/28/2022 01:14:59 - INFO - codeparrot_training - Step 5099: {'lr': 0.0004948751514884103, 'samples': 979200, 'steps': 5099, 'loss/train': 3.3718360662460327} 01/28/2022 01:15:03 - INFO - codeparrot_training - Step 5100: {'lr': 0.0004948718548897628, 'samples': 979392, 'steps': 5100, 'loss/train': 1.9075983166694641} 01/28/2022 01:15:07 - INFO - codeparrot_training - Step 5101: {'lr': 0.0004948685572421621, 'samples': 979584, 'steps': 5101, 'loss/train': 2.98984158039093} 01/28/2022 01:15:11 - INFO - codeparrot_training - Step 5102: {'lr': 0.0004948652585456222, 'samples': 979776, 'steps': 5102, 'loss/train': 4.319680094718933} 01/28/2022 01:15:17 - INFO - codeparrot_training - Step 5103: {'lr': 0.0004948619588001574, 'samples': 979968, 'steps': 5103, 'loss/train': 3.019576907157898} 01/28/2022 01:15:21 - INFO - codeparrot_training - Step 5104: {'lr': 0.0004948586580057816, 'samples': 980160, 'steps': 5104, 'loss/train': 2.305060923099518} 01/28/2022 01:15:25 - INFO - codeparrot_training - Step 5105: {'lr': 0.0004948553561625091, 'samples': 980352, 'steps': 5105, 'loss/train': 2.3076637387275696} 01/28/2022 01:15:29 - INFO - codeparrot_training - Step 5106: {'lr': 0.000494852053270354, 'samples': 980544, 'steps': 5106, 'loss/train': 2.494162380695343} 01/28/2022 01:15:33 - INFO - codeparrot_training - Step 5107: {'lr': 0.0004948487493293305, 'samples': 980736, 'steps': 5107, 'loss/train': 3.2628239393234253} 01/28/2022 01:15:40 - INFO - codeparrot_training - Step 5108: {'lr': 0.0004948454443394527, 'samples': 980928, 'steps': 5108, 'loss/train': 3.0942463874816895} 01/28/2022 01:15:44 - INFO - codeparrot_training - Step 5109: {'lr': 0.0004948421383007347, 'samples': 981120, 'steps': 5109, 'loss/train': 2.3395294547080994} 01/28/2022 01:15:48 - INFO - codeparrot_training - Step 5110: {'lr': 0.0004948388312131908, 'samples': 981312, 'steps': 5110, 'loss/train': 4.722561836242676} 01/28/2022 01:15:52 - INFO - codeparrot_training - Step 5111: {'lr': 0.0004948355230768349, 'samples': 981504, 'steps': 5111, 'loss/train': 3.7846977710723877} 01/28/2022 01:15:56 - INFO - codeparrot_training - Step 5112: {'lr': 0.0004948322138916816, 'samples': 981696, 'steps': 5112, 'loss/train': 3.0582504272460938} 01/28/2022 01:16:02 - INFO - codeparrot_training - Step 5113: {'lr': 0.0004948289036577447, 'samples': 981888, 'steps': 5113, 'loss/train': 3.1991801261901855} 01/28/2022 01:16:06 - INFO - codeparrot_training - Step 5114: {'lr': 0.0004948255923750385, 'samples': 982080, 'steps': 5114, 'loss/train': 10.26242208480835} 01/28/2022 01:16:10 - INFO - codeparrot_training - Step 5115: {'lr': 0.0004948222800435773, 'samples': 982272, 'steps': 5115, 'loss/train': 2.8759689331054688} 01/28/2022 01:16:14 - INFO - codeparrot_training - Step 5116: {'lr': 0.0004948189666633752, 'samples': 982464, 'steps': 5116, 'loss/train': 3.726628303527832} 01/28/2022 01:16:19 - INFO - codeparrot_training - Step 5117: {'lr': 0.0004948156522344463, 'samples': 982656, 'steps': 5117, 'loss/train': 1.500143587589264} 01/28/2022 01:16:24 - INFO - codeparrot_training - Step 5118: {'lr': 0.0004948123367568049, 'samples': 982848, 'steps': 5118, 'loss/train': 2.7915367484092712} 01/28/2022 01:16:28 - INFO - codeparrot_training - Step 5119: {'lr': 0.0004948090202304652, 'samples': 983040, 'steps': 5119, 'loss/train': 2.0413143038749695} 01/28/2022 01:16:32 - INFO - codeparrot_training - Step 5120: {'lr': 0.0004948057026554415, 'samples': 983232, 'steps': 5120, 'loss/train': 3.055608630180359} 01/28/2022 01:16:37 - INFO - codeparrot_training - Step 5121: {'lr': 0.0004948023840317477, 'samples': 983424, 'steps': 5121, 'loss/train': 1.0137527883052826} 01/28/2022 01:16:41 - INFO - codeparrot_training - Step 5122: {'lr': 0.0004947990643593983, 'samples': 983616, 'steps': 5122, 'loss/train': 3.854940891265869} 01/28/2022 01:16:46 - INFO - codeparrot_training - Step 5123: {'lr': 0.0004947957436384076, 'samples': 983808, 'steps': 5123, 'loss/train': 3.589133620262146} 01/28/2022 01:16:50 - INFO - codeparrot_training - Step 5124: {'lr': 0.0004947924218687894, 'samples': 984000, 'steps': 5124, 'loss/train': 3.0967111587524414} 01/28/2022 01:16:54 - INFO - codeparrot_training - Step 5125: {'lr': 0.0004947890990505585, 'samples': 984192, 'steps': 5125, 'loss/train': 1.3540555536746979} 01/28/2022 01:16:59 - INFO - codeparrot_training - Step 5126: {'lr': 0.0004947857751837286, 'samples': 984384, 'steps': 5126, 'loss/train': 3.2725131511688232} 01/28/2022 01:17:03 - INFO - codeparrot_training - Step 5127: {'lr': 0.0004947824502683142, 'samples': 984576, 'steps': 5127, 'loss/train': 4.499201416969299} 01/28/2022 01:17:09 - INFO - codeparrot_training - Step 5128: {'lr': 0.0004947791243043296, 'samples': 984768, 'steps': 5128, 'loss/train': 0.8515472710132599} 01/28/2022 01:17:13 - INFO - codeparrot_training - Step 5129: {'lr': 0.0004947757972917889, 'samples': 984960, 'steps': 5129, 'loss/train': 3.1633747816085815} 01/28/2022 01:17:17 - INFO - codeparrot_training - Step 5130: {'lr': 0.0004947724692307064, 'samples': 985152, 'steps': 5130, 'loss/train': 3.6844249963760376} 01/28/2022 01:17:21 - INFO - codeparrot_training - Step 5131: {'lr': 0.0004947691401210963, 'samples': 985344, 'steps': 5131, 'loss/train': 3.715325117111206} 01/28/2022 01:17:26 - INFO - codeparrot_training - Step 5132: {'lr': 0.0004947658099629731, 'samples': 985536, 'steps': 5132, 'loss/train': 4.048071026802063} 01/28/2022 01:17:31 - INFO - codeparrot_training - Step 5133: {'lr': 0.0004947624787563507, 'samples': 985728, 'steps': 5133, 'loss/train': 3.271116256713867} 01/28/2022 01:17:36 - INFO - codeparrot_training - Step 5134: {'lr': 0.0004947591465012436, 'samples': 985920, 'steps': 5134, 'loss/train': 3.3424465656280518} 01/28/2022 01:17:40 - INFO - codeparrot_training - Step 5135: {'lr': 0.0004947558131976661, 'samples': 986112, 'steps': 5135, 'loss/train': 1.2020516395568848} 01/28/2022 01:17:44 - INFO - codeparrot_training - Step 5136: {'lr': 0.0004947524788456324, 'samples': 986304, 'steps': 5136, 'loss/train': 1.340097963809967} 01/28/2022 01:17:48 - INFO - codeparrot_training - Step 5137: {'lr': 0.0004947491434451569, 'samples': 986496, 'steps': 5137, 'loss/train': 3.676682710647583} 01/28/2022 01:17:53 - INFO - codeparrot_training - Step 5138: {'lr': 0.0004947458069962537, 'samples': 986688, 'steps': 5138, 'loss/train': 3.408611297607422} 01/28/2022 01:17:58 - INFO - codeparrot_training - Step 5139: {'lr': 0.0004947424694989371, 'samples': 986880, 'steps': 5139, 'loss/train': 2.8533512949943542} 01/28/2022 01:18:02 - INFO - codeparrot_training - Step 5140: {'lr': 0.0004947391309532216, 'samples': 987072, 'steps': 5140, 'loss/train': 2.7126410007476807} 01/28/2022 01:18:06 - INFO - codeparrot_training - Step 5141: {'lr': 0.0004947357913591213, 'samples': 987264, 'steps': 5141, 'loss/train': 3.0837106704711914} 01/28/2022 01:18:10 - INFO - codeparrot_training - Step 5142: {'lr': 0.0004947324507166505, 'samples': 987456, 'steps': 5142, 'loss/train': 2.823248863220215} 01/28/2022 01:18:17 - INFO - codeparrot_training - Step 5143: {'lr': 0.0004947291090258238, 'samples': 987648, 'steps': 5143, 'loss/train': 3.096943974494934} 01/28/2022 01:18:21 - INFO - codeparrot_training - Step 5144: {'lr': 0.0004947257662866551, 'samples': 987840, 'steps': 5144, 'loss/train': 4.234661936759949} 01/28/2022 01:18:25 - INFO - codeparrot_training - Step 5145: {'lr': 0.0004947224224991591, 'samples': 988032, 'steps': 5145, 'loss/train': 0.8866762518882751} 01/28/2022 01:18:30 - INFO - codeparrot_training - Step 5146: {'lr': 0.0004947190776633499, 'samples': 988224, 'steps': 5146, 'loss/train': 3.848039388656616} 01/28/2022 01:18:34 - INFO - codeparrot_training - Step 5147: {'lr': 0.0004947157317792418, 'samples': 988416, 'steps': 5147, 'loss/train': 2.704115152359009} 01/28/2022 01:18:39 - INFO - codeparrot_training - Step 5148: {'lr': 0.0004947123848468493, 'samples': 988608, 'steps': 5148, 'loss/train': 2.970757842063904} 01/28/2022 01:18:43 - INFO - codeparrot_training - Step 5149: {'lr': 0.0004947090368661866, 'samples': 988800, 'steps': 5149, 'loss/train': 3.0487385988235474} 01/28/2022 01:18:48 - INFO - codeparrot_training - Step 5150: {'lr': 0.0004947056878372681, 'samples': 988992, 'steps': 5150, 'loss/train': 2.371294140815735} 01/28/2022 01:18:52 - INFO - codeparrot_training - Step 5151: {'lr': 0.0004947023377601082, 'samples': 989184, 'steps': 5151, 'loss/train': 2.885875403881073} 01/28/2022 01:18:56 - INFO - codeparrot_training - Step 5152: {'lr': 0.0004946989866347211, 'samples': 989376, 'steps': 5152, 'loss/train': 2.2738616466522217} 01/28/2022 01:19:02 - INFO - codeparrot_training - Step 5153: {'lr': 0.0004946956344611212, 'samples': 989568, 'steps': 5153, 'loss/train': 2.970350682735443} 01/28/2022 01:19:06 - INFO - codeparrot_training - Step 5154: {'lr': 0.000494692281239323, 'samples': 989760, 'steps': 5154, 'loss/train': 3.2702611684799194} 01/28/2022 01:19:10 - INFO - codeparrot_training - Step 5155: {'lr': 0.0004946889269693408, 'samples': 989952, 'steps': 5155, 'loss/train': 2.304049551486969} 01/28/2022 01:19:14 - INFO - codeparrot_training - Step 5156: {'lr': 0.0004946855716511888, 'samples': 990144, 'steps': 5156, 'loss/train': 2.707904040813446} 01/28/2022 01:19:18 - INFO - codeparrot_training - Step 5157: {'lr': 0.0004946822152848816, 'samples': 990336, 'steps': 5157, 'loss/train': 1.3406290411949158} 01/28/2022 01:19:24 - INFO - codeparrot_training - Step 5158: {'lr': 0.0004946788578704335, 'samples': 990528, 'steps': 5158, 'loss/train': 3.6067124605178833} 01/28/2022 01:19:28 - INFO - codeparrot_training - Step 5159: {'lr': 0.0004946754994078588, 'samples': 990720, 'steps': 5159, 'loss/train': 4.930338978767395} 01/28/2022 01:19:33 - INFO - codeparrot_training - Step 5160: {'lr': 0.000494672139897172, 'samples': 990912, 'steps': 5160, 'loss/train': 2.9877004623413086} 01/28/2022 01:19:37 - INFO - codeparrot_training - Step 5161: {'lr': 0.0004946687793383874, 'samples': 991104, 'steps': 5161, 'loss/train': 3.9014604091644287} 01/28/2022 01:19:41 - INFO - codeparrot_training - Step 5162: {'lr': 0.0004946654177315194, 'samples': 991296, 'steps': 5162, 'loss/train': 2.948493003845215} 01/28/2022 01:19:47 - INFO - codeparrot_training - Step 5163: {'lr': 0.0004946620550765826, 'samples': 991488, 'steps': 5163, 'loss/train': 2.2218820452690125} 01/28/2022 01:19:51 - INFO - codeparrot_training - Step 5164: {'lr': 0.0004946586913735911, 'samples': 991680, 'steps': 5164, 'loss/train': 3.622636556625366} 01/28/2022 01:19:55 - INFO - codeparrot_training - Step 5165: {'lr': 0.0004946553266225595, 'samples': 991872, 'steps': 5165, 'loss/train': 2.891622483730316} 01/28/2022 01:19:59 - INFO - codeparrot_training - Step 5166: {'lr': 0.0004946519608235022, 'samples': 992064, 'steps': 5166, 'loss/train': 2.2800053358078003} 01/28/2022 01:20:05 - INFO - codeparrot_training - Step 5167: {'lr': 0.0004946485939764336, 'samples': 992256, 'steps': 5167, 'loss/train': 2.2698517441749573} 01/28/2022 01:20:09 - INFO - codeparrot_training - Step 5168: {'lr': 0.000494645226081368, 'samples': 992448, 'steps': 5168, 'loss/train': 2.4274985790252686} 01/28/2022 01:20:14 - INFO - codeparrot_training - Step 5169: {'lr': 0.00049464185713832, 'samples': 992640, 'steps': 5169, 'loss/train': 4.341811895370483} 01/28/2022 01:20:18 - INFO - codeparrot_training - Step 5170: {'lr': 0.000494638487147304, 'samples': 992832, 'steps': 5170, 'loss/train': 2.128409743309021} 01/28/2022 01:20:22 - INFO - codeparrot_training - Step 5171: {'lr': 0.0004946351161083344, 'samples': 993024, 'steps': 5171, 'loss/train': 2.9380284547805786} 01/28/2022 01:20:27 - INFO - codeparrot_training - Step 5172: {'lr': 0.0004946317440214257, 'samples': 993216, 'steps': 5172, 'loss/train': 1.5900088548660278} 01/28/2022 01:20:32 - INFO - codeparrot_training - Step 5173: {'lr': 0.000494628370886592, 'samples': 993408, 'steps': 5173, 'loss/train': 2.261054813861847} 01/28/2022 01:20:36 - INFO - codeparrot_training - Step 5174: {'lr': 0.0004946249967038483, 'samples': 993600, 'steps': 5174, 'loss/train': 3.028007984161377} 01/28/2022 01:20:40 - INFO - codeparrot_training - Step 5175: {'lr': 0.0004946216214732088, 'samples': 993792, 'steps': 5175, 'loss/train': 3.2503806352615356} 01/28/2022 01:20:45 - INFO - codeparrot_training - Step 5176: {'lr': 0.0004946182451946878, 'samples': 993984, 'steps': 5176, 'loss/train': 1.4384489357471466} 01/28/2022 01:20:50 - INFO - codeparrot_training - Step 5177: {'lr': 0.0004946148678683001, 'samples': 994176, 'steps': 5177, 'loss/train': 2.334482252597809} 01/28/2022 01:20:54 - INFO - codeparrot_training - Step 5178: {'lr': 0.0004946114894940599, 'samples': 994368, 'steps': 5178, 'loss/train': 2.203895330429077} 01/28/2022 01:20:58 - INFO - codeparrot_training - Step 5179: {'lr': 0.0004946081100719817, 'samples': 994560, 'steps': 5179, 'loss/train': 2.245194733142853} 01/28/2022 01:21:03 - INFO - codeparrot_training - Step 5180: {'lr': 0.00049460472960208, 'samples': 994752, 'steps': 5180, 'loss/train': 2.453680694103241} 01/28/2022 01:21:07 - INFO - codeparrot_training - Step 5181: {'lr': 0.0004946013480843694, 'samples': 994944, 'steps': 5181, 'loss/train': 3.124812126159668} 01/28/2022 01:21:12 - INFO - codeparrot_training - Step 5182: {'lr': 0.0004945979655188642, 'samples': 995136, 'steps': 5182, 'loss/train': 3.971431016921997} 01/28/2022 01:21:16 - INFO - codeparrot_training - Step 5183: {'lr': 0.0004945945819055791, 'samples': 995328, 'steps': 5183, 'loss/train': 3.670953869819641} 01/28/2022 01:21:21 - INFO - codeparrot_training - Step 5184: {'lr': 0.0004945911972445284, 'samples': 995520, 'steps': 5184, 'loss/train': 2.599625587463379} 01/28/2022 01:21:25 - INFO - codeparrot_training - Step 5185: {'lr': 0.0004945878115357267, 'samples': 995712, 'steps': 5185, 'loss/train': 1.0762430727481842} 01/28/2022 01:21:29 - INFO - codeparrot_training - Step 5186: {'lr': 0.0004945844247791886, 'samples': 995904, 'steps': 5186, 'loss/train': 2.904076874256134} 01/28/2022 01:21:35 - INFO - codeparrot_training - Step 5187: {'lr': 0.0004945810369749283, 'samples': 996096, 'steps': 5187, 'loss/train': 3.461276650428772} 01/28/2022 01:21:39 - INFO - codeparrot_training - Step 5188: {'lr': 0.0004945776481229605, 'samples': 996288, 'steps': 5188, 'loss/train': 2.6225962042808533} 01/28/2022 01:21:43 - INFO - codeparrot_training - Step 5189: {'lr': 0.0004945742582232999, 'samples': 996480, 'steps': 5189, 'loss/train': 2.76076877117157} 01/28/2022 01:21:48 - INFO - codeparrot_training - Step 5190: {'lr': 0.0004945708672759606, 'samples': 996672, 'steps': 5190, 'loss/train': 2.5059961080551147} 01/28/2022 01:21:52 - INFO - codeparrot_training - Step 5191: {'lr': 0.0004945674752809575, 'samples': 996864, 'steps': 5191, 'loss/train': 2.7673187255859375} 01/28/2022 01:21:57 - INFO - codeparrot_training - Step 5192: {'lr': 0.000494564082238305, 'samples': 997056, 'steps': 5192, 'loss/train': 3.1980639696121216} 01/28/2022 01:22:01 - INFO - codeparrot_training - Step 5193: {'lr': 0.0004945606881480176, 'samples': 997248, 'steps': 5193, 'loss/train': 2.9759986996650696} 01/28/2022 01:22:05 - INFO - codeparrot_training - Step 5194: {'lr': 0.0004945572930101098, 'samples': 997440, 'steps': 5194, 'loss/train': 3.0611658096313477} 01/28/2022 01:22:10 - INFO - codeparrot_training - Step 5195: {'lr': 0.0004945538968245964, 'samples': 997632, 'steps': 5195, 'loss/train': 3.5552637577056885} 01/28/2022 01:22:14 - INFO - codeparrot_training - Step 5196: {'lr': 0.0004945504995914917, 'samples': 997824, 'steps': 5196, 'loss/train': 3.4210113286972046} 01/28/2022 01:22:19 - INFO - codeparrot_training - Step 5197: {'lr': 0.0004945471013108102, 'samples': 998016, 'steps': 5197, 'loss/train': 3.2834300994873047} 01/28/2022 01:22:23 - INFO - codeparrot_training - Step 5198: {'lr': 0.0004945437019825668, 'samples': 998208, 'steps': 5198, 'loss/train': 3.400662660598755} 01/28/2022 01:22:27 - INFO - codeparrot_training - Step 5199: {'lr': 0.0004945403016067756, 'samples': 998400, 'steps': 5199, 'loss/train': 1.7824596762657166} 01/28/2022 01:22:32 - INFO - codeparrot_training - Step 5200: {'lr': 0.0004945369001834514, 'samples': 998592, 'steps': 5200, 'loss/train': 2.9071136713027954} 01/28/2022 01:22:36 - INFO - codeparrot_training - Step 5201: {'lr': 0.0004945334977126089, 'samples': 998784, 'steps': 5201, 'loss/train': 1.0154067277908325} 01/28/2022 01:22:42 - INFO - codeparrot_training - Step 5202: {'lr': 0.0004945300941942624, 'samples': 998976, 'steps': 5202, 'loss/train': 3.136623501777649} 01/28/2022 01:22:46 - INFO - codeparrot_training - Step 5203: {'lr': 0.0004945266896284268, 'samples': 999168, 'steps': 5203, 'loss/train': 3.111311674118042} 01/28/2022 01:22:50 - INFO - codeparrot_training - Step 5204: {'lr': 0.0004945232840151164, 'samples': 999360, 'steps': 5204, 'loss/train': 3.4208933115005493} 01/28/2022 01:22:54 - INFO - codeparrot_training - Step 5205: {'lr': 0.0004945198773543459, 'samples': 999552, 'steps': 5205, 'loss/train': 2.4573474526405334} 01/28/2022 01:22:59 - INFO - codeparrot_training - Step 5206: {'lr': 0.0004945164696461299, 'samples': 999744, 'steps': 5206, 'loss/train': 4.217990756034851} 01/28/2022 01:23:04 - INFO - codeparrot_training - Step 5207: {'lr': 0.000494513060890483, 'samples': 999936, 'steps': 5207, 'loss/train': 2.308526873588562} 01/28/2022 01:23:08 - INFO - codeparrot_training - Step 5208: {'lr': 0.0004945096510874197, 'samples': 1000128, 'steps': 5208, 'loss/train': 2.4599268436431885} 01/28/2022 01:23:12 - INFO - codeparrot_training - Step 5209: {'lr': 0.0004945062402369548, 'samples': 1000320, 'steps': 5209, 'loss/train': 3.818615198135376} 01/28/2022 01:23:17 - INFO - codeparrot_training - Step 5210: {'lr': 0.0004945028283391028, 'samples': 1000512, 'steps': 5210, 'loss/train': 1.5330119132995605} 01/28/2022 01:23:21 - INFO - codeparrot_training - Step 5211: {'lr': 0.0004944994153938783, 'samples': 1000704, 'steps': 5211, 'loss/train': 4.191638231277466} 01/28/2022 01:23:27 - INFO - codeparrot_training - Step 5212: {'lr': 0.0004944960014012959, 'samples': 1000896, 'steps': 5212, 'loss/train': 3.3885337114334106} 01/28/2022 01:23:31 - INFO - codeparrot_training - Step 5213: {'lr': 0.0004944925863613704, 'samples': 1001088, 'steps': 5213, 'loss/train': 2.897596836090088} 01/28/2022 01:23:35 - INFO - codeparrot_training - Step 5214: {'lr': 0.0004944891702741161, 'samples': 1001280, 'steps': 5214, 'loss/train': 2.316688656806946} 01/28/2022 01:23:39 - INFO - codeparrot_training - Step 5215: {'lr': 0.0004944857531395479, 'samples': 1001472, 'steps': 5215, 'loss/train': 3.252509593963623} 01/28/2022 01:23:44 - INFO - codeparrot_training - Step 5216: {'lr': 0.0004944823349576805, 'samples': 1001664, 'steps': 5216, 'loss/train': 0.5747925192117691} 01/28/2022 01:23:49 - INFO - codeparrot_training - Step 5217: {'lr': 0.0004944789157285283, 'samples': 1001856, 'steps': 5217, 'loss/train': 3.3089951276779175} 01/28/2022 01:23:53 - INFO - codeparrot_training - Step 5218: {'lr': 0.0004944754954521061, 'samples': 1002048, 'steps': 5218, 'loss/train': 2.1714141368865967} 01/28/2022 01:23:57 - INFO - codeparrot_training - Step 5219: {'lr': 0.0004944720741284285, 'samples': 1002240, 'steps': 5219, 'loss/train': 0.5347638577222824} 01/28/2022 01:24:01 - INFO - codeparrot_training - Step 5220: {'lr': 0.00049446865175751, 'samples': 1002432, 'steps': 5220, 'loss/train': 3.429529309272766} 01/28/2022 01:24:06 - INFO - codeparrot_training - Step 5221: {'lr': 0.0004944652283393656, 'samples': 1002624, 'steps': 5221, 'loss/train': 2.7706503868103027} 01/28/2022 01:24:11 - INFO - codeparrot_training - Step 5222: {'lr': 0.0004944618038740098, 'samples': 1002816, 'steps': 5222, 'loss/train': 2.902158200740814} 01/28/2022 01:24:15 - INFO - codeparrot_training - Step 5223: {'lr': 0.0004944583783614571, 'samples': 1003008, 'steps': 5223, 'loss/train': 3.1215662956237793} 01/28/2022 01:24:19 - INFO - codeparrot_training - Step 5224: {'lr': 0.0004944549518017225, 'samples': 1003200, 'steps': 5224, 'loss/train': 1.644219696521759} 01/28/2022 01:24:23 - INFO - codeparrot_training - Step 5225: {'lr': 0.0004944515241948204, 'samples': 1003392, 'steps': 5225, 'loss/train': 3.272623658180237} 01/28/2022 01:24:28 - INFO - codeparrot_training - Step 5226: {'lr': 0.0004944480955407656, 'samples': 1003584, 'steps': 5226, 'loss/train': 1.5614703297615051} 01/28/2022 01:24:34 - INFO - codeparrot_training - Step 5227: {'lr': 0.0004944446658395728, 'samples': 1003776, 'steps': 5227, 'loss/train': 3.3854323625564575} 01/28/2022 01:24:38 - INFO - codeparrot_training - Step 5228: {'lr': 0.0004944412350912567, 'samples': 1003968, 'steps': 5228, 'loss/train': 3.1595388650894165} 01/28/2022 01:24:42 - INFO - codeparrot_training - Step 5229: {'lr': 0.000494437803295832, 'samples': 1004160, 'steps': 5229, 'loss/train': 2.645333468914032} 01/28/2022 01:24:46 - INFO - codeparrot_training - Step 5230: {'lr': 0.0004944343704533133, 'samples': 1004352, 'steps': 5230, 'loss/train': 3.7090258598327637} 01/28/2022 01:24:51 - INFO - codeparrot_training - Step 5231: {'lr': 0.0004944309365637154, 'samples': 1004544, 'steps': 5231, 'loss/train': 1.7371759414672852} 01/28/2022 01:24:56 - INFO - codeparrot_training - Step 5232: {'lr': 0.000494427501627053, 'samples': 1004736, 'steps': 5232, 'loss/train': 2.937090218067169} 01/28/2022 01:25:00 - INFO - codeparrot_training - Step 5233: {'lr': 0.0004944240656433407, 'samples': 1004928, 'steps': 5233, 'loss/train': 2.2389197945594788} 01/28/2022 01:25:04 - INFO - codeparrot_training - Step 5234: {'lr': 0.0004944206286125935, 'samples': 1005120, 'steps': 5234, 'loss/train': 1.3674657940864563} 01/28/2022 01:25:09 - INFO - codeparrot_training - Step 5235: {'lr': 0.0004944171905348258, 'samples': 1005312, 'steps': 5235, 'loss/train': 3.985906720161438} 01/28/2022 01:25:13 - INFO - codeparrot_training - Step 5236: {'lr': 0.0004944137514100525, 'samples': 1005504, 'steps': 5236, 'loss/train': 2.4827048778533936} 01/28/2022 01:25:18 - INFO - codeparrot_training - Step 5237: {'lr': 0.0004944103112382883, 'samples': 1005696, 'steps': 5237, 'loss/train': 0.5184672921895981} 01/28/2022 01:25:23 - INFO - codeparrot_training - Step 5238: {'lr': 0.0004944068700195479, 'samples': 1005888, 'steps': 5238, 'loss/train': 1.462128907442093} 01/28/2022 01:25:27 - INFO - codeparrot_training - Step 5239: {'lr': 0.0004944034277538462, 'samples': 1006080, 'steps': 5239, 'loss/train': 3.306077241897583} 01/28/2022 01:25:31 - INFO - codeparrot_training - Step 5240: {'lr': 0.0004943999844411977, 'samples': 1006272, 'steps': 5240, 'loss/train': 3.0923455953598022} 01/28/2022 01:25:35 - INFO - codeparrot_training - Step 5241: {'lr': 0.0004943965400816173, 'samples': 1006464, 'steps': 5241, 'loss/train': 2.570419192314148} 01/28/2022 01:25:41 - INFO - codeparrot_training - Step 5242: {'lr': 0.0004943930946751197, 'samples': 1006656, 'steps': 5242, 'loss/train': 2.9839507341384888} 01/28/2022 01:25:45 - INFO - codeparrot_training - Step 5243: {'lr': 0.0004943896482217197, 'samples': 1006848, 'steps': 5243, 'loss/train': 3.0156548023223877} 01/28/2022 01:25:49 - INFO - codeparrot_training - Step 5244: {'lr': 0.0004943862007214322, 'samples': 1007040, 'steps': 5244, 'loss/train': 1.8161049485206604} 01/28/2022 01:25:53 - INFO - codeparrot_training - Step 5245: {'lr': 0.0004943827521742716, 'samples': 1007232, 'steps': 5245, 'loss/train': 2.548302412033081} 01/28/2022 01:25:57 - INFO - codeparrot_training - Step 5246: {'lr': 0.000494379302580253, 'samples': 1007424, 'steps': 5246, 'loss/train': 2.4707399010658264} 01/28/2022 01:26:03 - INFO - codeparrot_training - Step 5247: {'lr': 0.000494375851939391, 'samples': 1007616, 'steps': 5247, 'loss/train': 2.351127505302429} 01/28/2022 01:26:08 - INFO - codeparrot_training - Step 5248: {'lr': 0.0004943724002517005, 'samples': 1007808, 'steps': 5248, 'loss/train': 3.4751386642456055} 01/28/2022 01:26:12 - INFO - codeparrot_training - Step 5249: {'lr': 0.0004943689475171962, 'samples': 1008000, 'steps': 5249, 'loss/train': 2.2653059363365173} 01/28/2022 01:26:16 - INFO - codeparrot_training - Step 5250: {'lr': 0.000494365493735893, 'samples': 1008192, 'steps': 5250, 'loss/train': 2.821557104587555} 01/28/2022 01:26:20 - INFO - codeparrot_training - Step 5251: {'lr': 0.0004943620389078055, 'samples': 1008384, 'steps': 5251, 'loss/train': 1.6988620162010193} 01/28/2022 01:26:26 - INFO - codeparrot_training - Step 5252: {'lr': 0.0004943585830329487, 'samples': 1008576, 'steps': 5252, 'loss/train': 1.573540449142456} 01/28/2022 01:26:30 - INFO - codeparrot_training - Step 5253: {'lr': 0.0004943551261113373, 'samples': 1008768, 'steps': 5253, 'loss/train': 2.9190085530281067} 01/28/2022 01:26:34 - INFO - codeparrot_training - Step 5254: {'lr': 0.0004943516681429861, 'samples': 1008960, 'steps': 5254, 'loss/train': 1.3444799780845642} 01/28/2022 01:26:38 - INFO - codeparrot_training - Step 5255: {'lr': 0.0004943482091279101, 'samples': 1009152, 'steps': 5255, 'loss/train': 2.9909947514533997} 01/28/2022 01:26:43 - INFO - codeparrot_training - Step 5256: {'lr': 0.0004943447490661238, 'samples': 1009344, 'steps': 5256, 'loss/train': 2.4964240193367004} 01/28/2022 01:26:48 - INFO - codeparrot_training - Step 5257: {'lr': 0.0004943412879576422, 'samples': 1009536, 'steps': 5257, 'loss/train': 3.023536205291748} 01/28/2022 01:26:53 - INFO - codeparrot_training - Step 5258: {'lr': 0.0004943378258024802, 'samples': 1009728, 'steps': 5258, 'loss/train': 2.4861887097358704} 01/28/2022 01:26:57 - INFO - codeparrot_training - Step 5259: {'lr': 0.0004943343626006524, 'samples': 1009920, 'steps': 5259, 'loss/train': 2.6445685029029846} 01/28/2022 01:27:01 - INFO - codeparrot_training - Step 5260: {'lr': 0.000494330898352174, 'samples': 1010112, 'steps': 5260, 'loss/train': 2.6240004301071167} 01/28/2022 01:27:05 - INFO - codeparrot_training - Step 5261: {'lr': 0.0004943274330570594, 'samples': 1010304, 'steps': 5261, 'loss/train': 2.779309630393982} 01/28/2022 01:27:10 - INFO - codeparrot_training - Step 5262: {'lr': 0.0004943239667153237, 'samples': 1010496, 'steps': 5262, 'loss/train': 3.187579393386841} 01/28/2022 01:27:15 - INFO - codeparrot_training - Step 5263: {'lr': 0.0004943204993269818, 'samples': 1010688, 'steps': 5263, 'loss/train': 2.9026309847831726} 01/28/2022 01:27:19 - INFO - codeparrot_training - Step 5264: {'lr': 0.0004943170308920483, 'samples': 1010880, 'steps': 5264, 'loss/train': 3.4065674543380737} 01/28/2022 01:27:23 - INFO - codeparrot_training - Step 5265: {'lr': 0.0004943135614105384, 'samples': 1011072, 'steps': 5265, 'loss/train': 2.8855032920837402} 01/28/2022 01:27:27 - INFO - codeparrot_training - Step 5266: {'lr': 0.0004943100908824667, 'samples': 1011264, 'steps': 5266, 'loss/train': 2.7968342900276184} 01/28/2022 01:27:33 - INFO - codeparrot_training - Step 5267: {'lr': 0.0004943066193078482, 'samples': 1011456, 'steps': 5267, 'loss/train': 1.8648970127105713} 01/28/2022 01:27:37 - INFO - codeparrot_training - Step 5268: {'lr': 0.0004943031466866976, 'samples': 1011648, 'steps': 5268, 'loss/train': 3.2487645149230957} 01/28/2022 01:27:41 - INFO - codeparrot_training - Step 5269: {'lr': 0.00049429967301903, 'samples': 1011840, 'steps': 5269, 'loss/train': 2.827558994293213} 01/28/2022 01:27:45 - INFO - codeparrot_training - Step 5270: {'lr': 0.0004942961983048601, 'samples': 1012032, 'steps': 5270, 'loss/train': 2.867476522922516} 01/28/2022 01:27:50 - INFO - codeparrot_training - Step 5271: {'lr': 0.0004942927225442029, 'samples': 1012224, 'steps': 5271, 'loss/train': 3.2432777881622314} 01/28/2022 01:27:55 - INFO - codeparrot_training - Step 5272: {'lr': 0.0004942892457370732, 'samples': 1012416, 'steps': 5272, 'loss/train': 2.0637552738189697} 01/28/2022 01:27:59 - INFO - codeparrot_training - Step 5273: {'lr': 0.000494285767883486, 'samples': 1012608, 'steps': 5273, 'loss/train': 2.9277766942977905} 01/28/2022 01:28:04 - INFO - codeparrot_training - Step 5274: {'lr': 0.0004942822889834562, 'samples': 1012800, 'steps': 5274, 'loss/train': 3.2012135982513428} 01/28/2022 01:28:08 - INFO - codeparrot_training - Step 5275: {'lr': 0.0004942788090369985, 'samples': 1012992, 'steps': 5275, 'loss/train': 2.0310811400413513} 01/28/2022 01:28:12 - INFO - codeparrot_training - Step 5276: {'lr': 0.0004942753280441281, 'samples': 1013184, 'steps': 5276, 'loss/train': 3.084933042526245} 01/28/2022 01:28:18 - INFO - codeparrot_training - Step 5277: {'lr': 0.0004942718460048596, 'samples': 1013376, 'steps': 5277, 'loss/train': 3.0689198970794678} 01/28/2022 01:28:22 - INFO - codeparrot_training - Step 5278: {'lr': 0.0004942683629192082, 'samples': 1013568, 'steps': 5278, 'loss/train': 2.4501099586486816} 01/28/2022 01:28:26 - INFO - codeparrot_training - Step 5279: {'lr': 0.0004942648787871886, 'samples': 1013760, 'steps': 5279, 'loss/train': 4.092139363288879} 01/28/2022 01:28:31 - INFO - codeparrot_training - Step 5280: {'lr': 0.000494261393608816, 'samples': 1013952, 'steps': 5280, 'loss/train': 2.7169734835624695} 01/28/2022 01:28:35 - INFO - codeparrot_training - Step 5281: {'lr': 0.0004942579073841049, 'samples': 1014144, 'steps': 5281, 'loss/train': 3.0049660205841064} 01/28/2022 01:28:40 - INFO - codeparrot_training - Step 5282: {'lr': 0.0004942544201130706, 'samples': 1014336, 'steps': 5282, 'loss/train': 2.948359787464142} 01/28/2022 01:28:44 - INFO - codeparrot_training - Step 5283: {'lr': 0.000494250931795728, 'samples': 1014528, 'steps': 5283, 'loss/train': 2.055180788040161} 01/28/2022 01:28:48 - INFO - codeparrot_training - Step 5284: {'lr': 0.0004942474424320919, 'samples': 1014720, 'steps': 5284, 'loss/train': 1.7748903036117554} 01/28/2022 01:28:53 - INFO - codeparrot_training - Step 5285: {'lr': 0.0004942439520221774, 'samples': 1014912, 'steps': 5285, 'loss/train': 2.9643605947494507} 01/28/2022 01:28:57 - INFO - codeparrot_training - Step 5286: {'lr': 0.0004942404605659991, 'samples': 1015104, 'steps': 5286, 'loss/train': 2.950847804546356} 01/28/2022 01:29:02 - INFO - codeparrot_training - Step 5287: {'lr': 0.0004942369680635724, 'samples': 1015296, 'steps': 5287, 'loss/train': 2.552581787109375} 01/28/2022 01:29:06 - INFO - codeparrot_training - Step 5288: {'lr': 0.0004942334745149122, 'samples': 1015488, 'steps': 5288, 'loss/train': 3.504793167114258} 01/28/2022 01:29:11 - INFO - codeparrot_training - Step 5289: {'lr': 0.0004942299799200332, 'samples': 1015680, 'steps': 5289, 'loss/train': 2.825487792491913} 01/28/2022 01:29:15 - INFO - codeparrot_training - Step 5290: {'lr': 0.0004942264842789506, 'samples': 1015872, 'steps': 5290, 'loss/train': 1.841290533542633} 01/28/2022 01:29:19 - INFO - codeparrot_training - Step 5291: {'lr': 0.0004942229875916792, 'samples': 1016064, 'steps': 5291, 'loss/train': 2.8703380823135376} 01/28/2022 01:29:25 - INFO - codeparrot_training - Step 5292: {'lr': 0.0004942194898582341, 'samples': 1016256, 'steps': 5292, 'loss/train': 4.064197540283203} 01/28/2022 01:29:29 - INFO - codeparrot_training - Step 5293: {'lr': 0.0004942159910786303, 'samples': 1016448, 'steps': 5293, 'loss/train': 3.308609962463379} 01/28/2022 01:29:33 - INFO - codeparrot_training - Step 5294: {'lr': 0.0004942124912528827, 'samples': 1016640, 'steps': 5294, 'loss/train': 2.880803346633911} 01/28/2022 01:29:38 - INFO - codeparrot_training - Step 5295: {'lr': 0.0004942089903810064, 'samples': 1016832, 'steps': 5295, 'loss/train': 1.8778949975967407} 01/28/2022 01:29:42 - INFO - codeparrot_training - Step 5296: {'lr': 0.0004942054884630162, 'samples': 1017024, 'steps': 5296, 'loss/train': 2.837098717689514} 01/28/2022 01:29:47 - INFO - codeparrot_training - Step 5297: {'lr': 0.0004942019854989274, 'samples': 1017216, 'steps': 5297, 'loss/train': 3.140382170677185} 01/28/2022 01:29:52 - INFO - codeparrot_training - Step 5298: {'lr': 0.0004941984814887546, 'samples': 1017408, 'steps': 5298, 'loss/train': 3.6795294284820557} 01/28/2022 01:29:56 - INFO - codeparrot_training - Step 5299: {'lr': 0.0004941949764325133, 'samples': 1017600, 'steps': 5299, 'loss/train': 3.175606369972229} 01/28/2022 01:30:00 - INFO - codeparrot_training - Step 5300: {'lr': 0.0004941914703302181, 'samples': 1017792, 'steps': 5300, 'loss/train': 2.1397724747657776} 01/28/2022 01:30:04 - INFO - codeparrot_training - Step 5301: {'lr': 0.0004941879631818843, 'samples': 1017984, 'steps': 5301, 'loss/train': 3.3027169704437256} 01/28/2022 01:30:10 - INFO - codeparrot_training - Step 5302: {'lr': 0.0004941844549875267, 'samples': 1018176, 'steps': 5302, 'loss/train': 2.4668360352516174} 01/28/2022 01:30:14 - INFO - codeparrot_training - Step 5303: {'lr': 0.0004941809457471605, 'samples': 1018368, 'steps': 5303, 'loss/train': 3.575745105743408} 01/28/2022 01:30:18 - INFO - codeparrot_training - Step 5304: {'lr': 0.0004941774354608006, 'samples': 1018560, 'steps': 5304, 'loss/train': 3.202060103416443} 01/28/2022 01:30:23 - INFO - codeparrot_training - Step 5305: {'lr': 0.0004941739241284621, 'samples': 1018752, 'steps': 5305, 'loss/train': 2.2794567346572876} 01/28/2022 01:30:27 - INFO - codeparrot_training - Step 5306: {'lr': 0.0004941704117501601, 'samples': 1018944, 'steps': 5306, 'loss/train': 2.6880522966384888} 01/28/2022 01:30:33 - INFO - codeparrot_training - Step 5307: {'lr': 0.0004941668983259095, 'samples': 1019136, 'steps': 5307, 'loss/train': 0.5467876642942429} 01/28/2022 01:30:37 - INFO - codeparrot_training - Step 5308: {'lr': 0.0004941633838557256, 'samples': 1019328, 'steps': 5308, 'loss/train': 3.1908241510391235} 01/28/2022 01:30:41 - INFO - codeparrot_training - Step 5309: {'lr': 0.0004941598683396232, 'samples': 1019520, 'steps': 5309, 'loss/train': 3.2528157234191895} 01/28/2022 01:30:45 - INFO - codeparrot_training - Step 5310: {'lr': 0.0004941563517776174, 'samples': 1019712, 'steps': 5310, 'loss/train': 2.9022980332374573} 01/28/2022 01:30:49 - INFO - codeparrot_training - Step 5311: {'lr': 0.0004941528341697234, 'samples': 1019904, 'steps': 5311, 'loss/train': 2.664256989955902} 01/28/2022 01:30:55 - INFO - codeparrot_training - Step 5312: {'lr': 0.0004941493155159562, 'samples': 1020096, 'steps': 5312, 'loss/train': 10.236583471298218} 01/28/2022 01:30:59 - INFO - codeparrot_training - Step 5313: {'lr': 0.0004941457958163308, 'samples': 1020288, 'steps': 5313, 'loss/train': 2.252800405025482} 01/28/2022 01:31:04 - INFO - codeparrot_training - Step 5314: {'lr': 0.0004941422750708623, 'samples': 1020480, 'steps': 5314, 'loss/train': 3.2235699892044067} 01/28/2022 01:31:08 - INFO - codeparrot_training - Step 5315: {'lr': 0.0004941387532795659, 'samples': 1020672, 'steps': 5315, 'loss/train': 2.364248514175415} 01/28/2022 01:31:12 - INFO - codeparrot_training - Step 5316: {'lr': 0.0004941352304424566, 'samples': 1020864, 'steps': 5316, 'loss/train': 2.5890172719955444} 01/28/2022 01:31:18 - INFO - codeparrot_training - Step 5317: {'lr': 0.0004941317065595495, 'samples': 1021056, 'steps': 5317, 'loss/train': 3.7490698099136353} 01/28/2022 01:31:22 - INFO - codeparrot_training - Step 5318: {'lr': 0.0004941281816308596, 'samples': 1021248, 'steps': 5318, 'loss/train': 3.5645978450775146} 01/28/2022 01:31:26 - INFO - codeparrot_training - Step 5319: {'lr': 0.0004941246556564021, 'samples': 1021440, 'steps': 5319, 'loss/train': 3.291991710662842} 01/28/2022 01:31:31 - INFO - codeparrot_training - Step 5320: {'lr': 0.0004941211286361922, 'samples': 1021632, 'steps': 5320, 'loss/train': 3.0788036584854126} 01/28/2022 01:31:35 - INFO - codeparrot_training - Step 5321: {'lr': 0.0004941176005702448, 'samples': 1021824, 'steps': 5321, 'loss/train': 3.4686309099197388} 01/28/2022 01:31:40 - INFO - codeparrot_training - Step 5322: {'lr': 0.0004941140714585752, 'samples': 1022016, 'steps': 5322, 'loss/train': 2.601605772972107} 01/28/2022 01:31:44 - INFO - codeparrot_training - Step 5323: {'lr': 0.0004941105413011984, 'samples': 1022208, 'steps': 5323, 'loss/train': 2.9721314907073975} 01/28/2022 01:31:49 - INFO - codeparrot_training - Step 5324: {'lr': 0.0004941070100981295, 'samples': 1022400, 'steps': 5324, 'loss/train': 2.9418616890907288} 01/28/2022 01:31:53 - INFO - codeparrot_training - Step 5325: {'lr': 0.0004941034778493837, 'samples': 1022592, 'steps': 5325, 'loss/train': 1.7072229981422424} 01/28/2022 01:31:57 - INFO - codeparrot_training - Step 5326: {'lr': 0.0004940999445549762, 'samples': 1022784, 'steps': 5326, 'loss/train': 2.464781105518341} 01/28/2022 01:32:02 - INFO - codeparrot_training - Step 5327: {'lr': 0.0004940964102149219, 'samples': 1022976, 'steps': 5327, 'loss/train': 1.3386387526988983} 01/28/2022 01:32:06 - INFO - codeparrot_training - Step 5328: {'lr': 0.0004940928748292363, 'samples': 1023168, 'steps': 5328, 'loss/train': 4.126282453536987} 01/28/2022 01:32:11 - INFO - codeparrot_training - Step 5329: {'lr': 0.0004940893383979341, 'samples': 1023360, 'steps': 5329, 'loss/train': 3.109351873397827} 01/28/2022 01:32:15 - INFO - codeparrot_training - Step 5330: {'lr': 0.0004940858009210308, 'samples': 1023552, 'steps': 5330, 'loss/train': 3.099810004234314} 01/28/2022 01:32:19 - INFO - codeparrot_training - Step 5331: {'lr': 0.0004940822623985414, 'samples': 1023744, 'steps': 5331, 'loss/train': 1.5843912363052368} 01/28/2022 01:32:25 - INFO - codeparrot_training - Step 5332: {'lr': 0.0004940787228304811, 'samples': 1023936, 'steps': 5332, 'loss/train': 1.4593356549739838} 01/28/2022 01:32:29 - INFO - codeparrot_training - Step 5333: {'lr': 0.0004940751822168651, 'samples': 1024128, 'steps': 5333, 'loss/train': 2.258207380771637} 01/28/2022 01:32:34 - INFO - codeparrot_training - Step 5334: {'lr': 0.0004940716405577086, 'samples': 1024320, 'steps': 5334, 'loss/train': 3.9413416385650635} 01/28/2022 01:32:38 - INFO - codeparrot_training - Step 5335: {'lr': 0.0004940680978530265, 'samples': 1024512, 'steps': 5335, 'loss/train': 2.1533102989196777} 01/28/2022 01:32:42 - INFO - codeparrot_training - Step 5336: {'lr': 0.0004940645541028343, 'samples': 1024704, 'steps': 5336, 'loss/train': 2.735740542411804} 01/28/2022 01:32:47 - INFO - codeparrot_training - Step 5337: {'lr': 0.0004940610093071469, 'samples': 1024896, 'steps': 5337, 'loss/train': 3.0170141458511353} 01/28/2022 01:32:51 - INFO - codeparrot_training - Step 5338: {'lr': 0.0004940574634659798, 'samples': 1025088, 'steps': 5338, 'loss/train': 2.999441921710968} 01/28/2022 01:32:56 - INFO - codeparrot_training - Step 5339: {'lr': 0.000494053916579348, 'samples': 1025280, 'steps': 5339, 'loss/train': 2.7463055849075317} 01/28/2022 01:33:00 - INFO - codeparrot_training - Step 5340: {'lr': 0.0004940503686472667, 'samples': 1025472, 'steps': 5340, 'loss/train': 2.5248718857765198} 01/28/2022 01:33:04 - INFO - codeparrot_training - Step 5341: {'lr': 0.0004940468196697511, 'samples': 1025664, 'steps': 5341, 'loss/train': 2.6225607991218567} 01/28/2022 01:33:09 - INFO - codeparrot_training - Step 5342: {'lr': 0.0004940432696468164, 'samples': 1025856, 'steps': 5342, 'loss/train': 3.2310118675231934} 01/28/2022 01:33:14 - INFO - codeparrot_training - Step 5343: {'lr': 0.0004940397185784778, 'samples': 1026048, 'steps': 5343, 'loss/train': 3.442689299583435} 01/28/2022 01:33:18 - INFO - codeparrot_training - Step 5344: {'lr': 0.0004940361664647506, 'samples': 1026240, 'steps': 5344, 'loss/train': 2.6647818088531494} 01/28/2022 01:33:22 - INFO - codeparrot_training - Step 5345: {'lr': 0.0004940326133056499, 'samples': 1026432, 'steps': 5345, 'loss/train': 2.9968369603157043} 01/28/2022 01:33:26 - INFO - codeparrot_training - Step 5346: {'lr': 0.000494029059101191, 'samples': 1026624, 'steps': 5346, 'loss/train': 3.294980764389038} 01/28/2022 01:33:32 - INFO - codeparrot_training - Step 5347: {'lr': 0.0004940255038513891, 'samples': 1026816, 'steps': 5347, 'loss/train': 2.1327531337738037} 01/28/2022 01:33:36 - INFO - codeparrot_training - Step 5348: {'lr': 0.0004940219475562593, 'samples': 1027008, 'steps': 5348, 'loss/train': 3.248226284980774} 01/28/2022 01:33:40 - INFO - codeparrot_training - Step 5349: {'lr': 0.0004940183902158172, 'samples': 1027200, 'steps': 5349, 'loss/train': 1.467650055885315} 01/28/2022 01:33:45 - INFO - codeparrot_training - Step 5350: {'lr': 0.0004940148318300777, 'samples': 1027392, 'steps': 5350, 'loss/train': 1.923698902130127} 01/28/2022 01:33:49 - INFO - codeparrot_training - Step 5351: {'lr': 0.0004940112723990561, 'samples': 1027584, 'steps': 5351, 'loss/train': 3.737089991569519} 01/28/2022 01:33:55 - INFO - codeparrot_training - Step 5352: {'lr': 0.0004940077119227678, 'samples': 1027776, 'steps': 5352, 'loss/train': 2.2785741090774536} 01/28/2022 01:33:59 - INFO - codeparrot_training - Step 5353: {'lr': 0.0004940041504012279, 'samples': 1027968, 'steps': 5353, 'loss/train': 1.6892321705818176} 01/28/2022 01:34:03 - INFO - codeparrot_training - Step 5354: {'lr': 0.0004940005878344517, 'samples': 1028160, 'steps': 5354, 'loss/train': 2.9674283266067505} 01/28/2022 01:34:07 - INFO - codeparrot_training - Step 5355: {'lr': 0.0004939970242224544, 'samples': 1028352, 'steps': 5355, 'loss/train': 2.821284770965576} 01/28/2022 01:34:12 - INFO - codeparrot_training - Step 5356: {'lr': 0.0004939934595652513, 'samples': 1028544, 'steps': 5356, 'loss/train': 2.9866039752960205} 01/28/2022 01:34:17 - INFO - codeparrot_training - Step 5357: {'lr': 0.0004939898938628578, 'samples': 1028736, 'steps': 5357, 'loss/train': 3.5149773359298706} 01/28/2022 01:34:21 - INFO - codeparrot_training - Step 5358: {'lr': 0.000493986327115289, 'samples': 1028928, 'steps': 5358, 'loss/train': 3.6889750957489014} 01/28/2022 01:34:25 - INFO - codeparrot_training - Step 5359: {'lr': 0.0004939827593225602, 'samples': 1029120, 'steps': 5359, 'loss/train': 3.053527593612671} 01/28/2022 01:34:30 - INFO - codeparrot_training - Step 5360: {'lr': 0.0004939791904846869, 'samples': 1029312, 'steps': 5360, 'loss/train': 2.8351637721061707} 01/28/2022 01:34:34 - INFO - codeparrot_training - Step 5361: {'lr': 0.0004939756206016841, 'samples': 1029504, 'steps': 5361, 'loss/train': 3.6576586961746216} 01/28/2022 01:34:40 - INFO - codeparrot_training - Step 5362: {'lr': 0.0004939720496735672, 'samples': 1029696, 'steps': 5362, 'loss/train': 2.980075478553772} 01/28/2022 01:34:44 - INFO - codeparrot_training - Step 5363: {'lr': 0.0004939684777003516, 'samples': 1029888, 'steps': 5363, 'loss/train': 2.39290452003479} 01/28/2022 01:34:48 - INFO - codeparrot_training - Step 5364: {'lr': 0.0004939649046820524, 'samples': 1030080, 'steps': 5364, 'loss/train': 3.053941011428833} 01/28/2022 01:34:52 - INFO - codeparrot_training - Step 5365: {'lr': 0.0004939613306186851, 'samples': 1030272, 'steps': 5365, 'loss/train': 1.822664737701416} 01/28/2022 01:34:56 - INFO - codeparrot_training - Step 5366: {'lr': 0.0004939577555102649, 'samples': 1030464, 'steps': 5366, 'loss/train': 4.4408605098724365} 01/28/2022 01:35:02 - INFO - codeparrot_training - Step 5367: {'lr': 0.0004939541793568072, 'samples': 1030656, 'steps': 5367, 'loss/train': 2.903935968875885} 01/28/2022 01:35:06 - INFO - codeparrot_training - Step 5368: {'lr': 0.000493950602158327, 'samples': 1030848, 'steps': 5368, 'loss/train': 3.8786137104034424} 01/28/2022 01:35:10 - INFO - codeparrot_training - Step 5369: {'lr': 0.0004939470239148403, 'samples': 1031040, 'steps': 5369, 'loss/train': 3.1342148780822754} 01/28/2022 01:35:14 - INFO - codeparrot_training - Step 5370: {'lr': 0.0004939434446263617, 'samples': 1031232, 'steps': 5370, 'loss/train': 3.947035789489746} 01/28/2022 01:35:19 - INFO - codeparrot_training - Step 5371: {'lr': 0.000493939864292907, 'samples': 1031424, 'steps': 5371, 'loss/train': 2.924211323261261} 01/28/2022 01:35:24 - INFO - codeparrot_training - Step 5372: {'lr': 0.0004939362829144913, 'samples': 1031616, 'steps': 5372, 'loss/train': 2.7048568725585938} 01/28/2022 01:35:28 - INFO - codeparrot_training - Step 5373: {'lr': 0.00049393270049113, 'samples': 1031808, 'steps': 5373, 'loss/train': 2.181123375892639} 01/28/2022 01:35:32 - INFO - codeparrot_training - Step 5374: {'lr': 0.0004939291170228385, 'samples': 1032000, 'steps': 5374, 'loss/train': 4.498282313346863} 01/28/2022 01:35:36 - INFO - codeparrot_training - Step 5375: {'lr': 0.0004939255325096321, 'samples': 1032192, 'steps': 5375, 'loss/train': 2.3540621995925903} 01/28/2022 01:35:41 - INFO - codeparrot_training - Step 5376: {'lr': 0.0004939219469515262, 'samples': 1032384, 'steps': 5376, 'loss/train': 2.4104695916175842} 01/28/2022 01:35:47 - INFO - codeparrot_training - Step 5377: {'lr': 0.0004939183603485363, 'samples': 1032576, 'steps': 5377, 'loss/train': 2.3766093850135803} 01/28/2022 01:35:51 - INFO - codeparrot_training - Step 5378: {'lr': 0.0004939147727006773, 'samples': 1032768, 'steps': 5378, 'loss/train': 2.0129016637802124} 01/28/2022 01:35:55 - INFO - codeparrot_training - Step 5379: {'lr': 0.000493911184007965, 'samples': 1032960, 'steps': 5379, 'loss/train': 3.330846905708313} 01/28/2022 01:35:59 - INFO - codeparrot_training - Step 5380: {'lr': 0.0004939075942704147, 'samples': 1033152, 'steps': 5380, 'loss/train': 1.7827277183532715} 01/28/2022 01:36:04 - INFO - codeparrot_training - Step 5381: {'lr': 0.0004939040034880416, 'samples': 1033344, 'steps': 5381, 'loss/train': 2.6852661967277527} 01/28/2022 01:36:09 - INFO - codeparrot_training - Step 5382: {'lr': 0.0004939004116608612, 'samples': 1033536, 'steps': 5382, 'loss/train': 3.161369204521179} 01/28/2022 01:36:13 - INFO - codeparrot_training - Step 5383: {'lr': 0.000493896818788889, 'samples': 1033728, 'steps': 5383, 'loss/train': 2.076167106628418} 01/28/2022 01:36:17 - INFO - codeparrot_training - Step 5384: {'lr': 0.0004938932248721401, 'samples': 1033920, 'steps': 5384, 'loss/train': 2.744679808616638} 01/28/2022 01:36:22 - INFO - codeparrot_training - Step 5385: {'lr': 0.0004938896299106302, 'samples': 1034112, 'steps': 5385, 'loss/train': 2.3805474042892456} 01/28/2022 01:36:26 - INFO - codeparrot_training - Step 5386: {'lr': 0.0004938860339043746, 'samples': 1034304, 'steps': 5386, 'loss/train': 2.5925962328910828} 01/28/2022 01:36:32 - INFO - codeparrot_training - Step 5387: {'lr': 0.0004938824368533886, 'samples': 1034496, 'steps': 5387, 'loss/train': 3.2458012104034424} 01/28/2022 01:36:36 - INFO - codeparrot_training - Step 5388: {'lr': 0.0004938788387576878, 'samples': 1034688, 'steps': 5388, 'loss/train': 3.070823907852173} 01/28/2022 01:36:40 - INFO - codeparrot_training - Step 5389: {'lr': 0.0004938752396172873, 'samples': 1034880, 'steps': 5389, 'loss/train': 3.457394599914551} 01/28/2022 01:36:45 - INFO - codeparrot_training - Step 5390: {'lr': 0.0004938716394322028, 'samples': 1035072, 'steps': 5390, 'loss/train': 1.3794122636318207} 01/28/2022 01:36:49 - INFO - codeparrot_training - Step 5391: {'lr': 0.0004938680382024497, 'samples': 1035264, 'steps': 5391, 'loss/train': 3.2944836616516113} 01/28/2022 01:36:55 - INFO - codeparrot_training - Step 5392: {'lr': 0.0004938644359280433, 'samples': 1035456, 'steps': 5392, 'loss/train': 4.130287170410156} 01/28/2022 01:36:59 - INFO - codeparrot_training - Step 5393: {'lr': 0.000493860832608999, 'samples': 1035648, 'steps': 5393, 'loss/train': 2.0841532945632935} 01/28/2022 01:37:03 - INFO - codeparrot_training - Step 5394: {'lr': 0.0004938572282453326, 'samples': 1035840, 'steps': 5394, 'loss/train': 3.8426438570022583} 01/28/2022 01:37:08 - INFO - codeparrot_training - Step 5395: {'lr': 0.000493853622837059, 'samples': 1036032, 'steps': 5395, 'loss/train': 3.607234239578247} 01/28/2022 01:37:12 - INFO - codeparrot_training - Step 5396: {'lr': 0.000493850016384194, 'samples': 1036224, 'steps': 5396, 'loss/train': 3.06574022769928} 01/28/2022 01:37:16 - INFO - codeparrot_training - Step 5397: {'lr': 0.000493846408886753, 'samples': 1036416, 'steps': 5397, 'loss/train': 2.8427558541297913} 01/28/2022 01:37:21 - INFO - codeparrot_training - Step 5398: {'lr': 0.0004938428003447514, 'samples': 1036608, 'steps': 5398, 'loss/train': 2.634541869163513} 01/28/2022 01:37:25 - INFO - codeparrot_training - Step 5399: {'lr': 0.0004938391907582046, 'samples': 1036800, 'steps': 5399, 'loss/train': 4.6046222448349} 01/28/2022 01:37:30 - INFO - codeparrot_training - Step 5400: {'lr': 0.0004938355801271282, 'samples': 1036992, 'steps': 5400, 'loss/train': 2.5170936584472656} 01/28/2022 01:37:34 - INFO - codeparrot_training - Step 5401: {'lr': 0.0004938319684515375, 'samples': 1037184, 'steps': 5401, 'loss/train': 1.4280155003070831} 01/28/2022 01:37:38 - INFO - codeparrot_training - Step 5402: {'lr': 0.0004938283557314483, 'samples': 1037376, 'steps': 5402, 'loss/train': 3.261497139930725} 01/28/2022 01:37:44 - INFO - codeparrot_training - Step 5403: {'lr': 0.0004938247419668757, 'samples': 1037568, 'steps': 5403, 'loss/train': 2.935101628303528} 01/28/2022 01:37:48 - INFO - codeparrot_training - Step 5404: {'lr': 0.0004938211271578352, 'samples': 1037760, 'steps': 5404, 'loss/train': 1.720741331577301} 01/28/2022 01:37:52 - INFO - codeparrot_training - Step 5405: {'lr': 0.0004938175113043426, 'samples': 1037952, 'steps': 5405, 'loss/train': 3.3306477069854736} 01/28/2022 01:37:56 - INFO - codeparrot_training - Step 5406: {'lr': 0.0004938138944064131, 'samples': 1038144, 'steps': 5406, 'loss/train': 4.344101786613464} 01/28/2022 01:38:01 - INFO - codeparrot_training - Step 5407: {'lr': 0.0004938102764640624, 'samples': 1038336, 'steps': 5407, 'loss/train': 0.8386203646659851} 01/28/2022 01:38:07 - INFO - codeparrot_training - Step 5408: {'lr': 0.0004938066574773058, 'samples': 1038528, 'steps': 5408, 'loss/train': 0.7660645544528961} 01/28/2022 01:38:11 - INFO - codeparrot_training - Step 5409: {'lr': 0.000493803037446159, 'samples': 1038720, 'steps': 5409, 'loss/train': 2.5750914216041565} 01/28/2022 01:38:15 - INFO - codeparrot_training - Step 5410: {'lr': 0.0004937994163706374, 'samples': 1038912, 'steps': 5410, 'loss/train': 3.753603458404541} 01/28/2022 01:38:19 - INFO - codeparrot_training - Step 5411: {'lr': 0.0004937957942507564, 'samples': 1039104, 'steps': 5411, 'loss/train': 1.8631240725517273} 01/28/2022 01:38:23 - INFO - codeparrot_training - Step 5412: {'lr': 0.0004937921710865317, 'samples': 1039296, 'steps': 5412, 'loss/train': 2.180336058139801} 01/28/2022 01:38:29 - INFO - codeparrot_training - Step 5413: {'lr': 0.0004937885468779787, 'samples': 1039488, 'steps': 5413, 'loss/train': 2.515847325325012} 01/28/2022 01:38:33 - INFO - codeparrot_training - Step 5414: {'lr': 0.000493784921625113, 'samples': 1039680, 'steps': 5414, 'loss/train': 2.9716907143592834} 01/28/2022 01:38:37 - INFO - codeparrot_training - Step 5415: {'lr': 0.0004937812953279502, 'samples': 1039872, 'steps': 5415, 'loss/train': 3.7742786407470703} 01/28/2022 01:38:41 - INFO - codeparrot_training - Step 5416: {'lr': 0.0004937776679865057, 'samples': 1040064, 'steps': 5416, 'loss/train': 2.959122061729431} 01/28/2022 01:38:46 - INFO - codeparrot_training - Step 5417: {'lr': 0.000493774039600795, 'samples': 1040256, 'steps': 5417, 'loss/train': 2.7050732374191284} 01/28/2022 01:38:51 - INFO - codeparrot_training - Step 5418: {'lr': 0.0004937704101708338, 'samples': 1040448, 'steps': 5418, 'loss/train': 3.28727924823761} 01/28/2022 01:38:55 - INFO - codeparrot_training - Step 5419: {'lr': 0.0004937667796966374, 'samples': 1040640, 'steps': 5419, 'loss/train': 2.495517075061798} 01/28/2022 01:38:59 - INFO - codeparrot_training - Step 5420: {'lr': 0.0004937631481782218, 'samples': 1040832, 'steps': 5420, 'loss/train': 2.2501752376556396} 01/28/2022 01:39:04 - INFO - codeparrot_training - Step 5421: {'lr': 0.000493759515615602, 'samples': 1041024, 'steps': 5421, 'loss/train': 3.6354739665985107} 01/28/2022 01:39:08 - INFO - codeparrot_training - Step 5422: {'lr': 0.000493755882008794, 'samples': 1041216, 'steps': 5422, 'loss/train': 1.6325177550315857} 01/28/2022 01:39:14 - INFO - codeparrot_training - Step 5423: {'lr': 0.0004937522473578132, 'samples': 1041408, 'steps': 5423, 'loss/train': 1.2051466405391693} 01/28/2022 01:39:18 - INFO - codeparrot_training - Step 5424: {'lr': 0.0004937486116626752, 'samples': 1041600, 'steps': 5424, 'loss/train': 2.0218657851219177} 01/28/2022 01:39:22 - INFO - codeparrot_training - Step 5425: {'lr': 0.0004937449749233954, 'samples': 1041792, 'steps': 5425, 'loss/train': 2.8115347623825073} 01/28/2022 01:39:26 - INFO - codeparrot_training - Step 5426: {'lr': 0.0004937413371399897, 'samples': 1041984, 'steps': 5426, 'loss/train': 3.348379969596863} 01/28/2022 01:39:31 - INFO - codeparrot_training - Step 5427: {'lr': 0.0004937376983124734, 'samples': 1042176, 'steps': 5427, 'loss/train': 3.0259451866149902} 01/28/2022 01:39:36 - INFO - codeparrot_training - Step 5428: {'lr': 0.0004937340584408622, 'samples': 1042368, 'steps': 5428, 'loss/train': 2.273713231086731} 01/28/2022 01:39:40 - INFO - codeparrot_training - Step 5429: {'lr': 0.0004937304175251717, 'samples': 1042560, 'steps': 5429, 'loss/train': 3.5861088037490845} 01/28/2022 01:39:44 - INFO - codeparrot_training - Step 5430: {'lr': 0.0004937267755654174, 'samples': 1042752, 'steps': 5430, 'loss/train': 2.676105558872223} 01/28/2022 01:39:48 - INFO - codeparrot_training - Step 5431: {'lr': 0.0004937231325616152, 'samples': 1042944, 'steps': 5431, 'loss/train': 2.797670602798462} 01/28/2022 01:39:53 - INFO - codeparrot_training - Step 5432: {'lr': 0.0004937194885137803, 'samples': 1043136, 'steps': 5432, 'loss/train': 2.1813708543777466} 01/28/2022 01:39:59 - INFO - codeparrot_training - Step 5433: {'lr': 0.0004937158434219286, 'samples': 1043328, 'steps': 5433, 'loss/train': 3.1813570261001587} 01/28/2022 01:40:03 - INFO - codeparrot_training - Step 5434: {'lr': 0.0004937121972860755, 'samples': 1043520, 'steps': 5434, 'loss/train': 3.034653425216675} 01/28/2022 01:40:07 - INFO - codeparrot_training - Step 5435: {'lr': 0.0004937085501062369, 'samples': 1043712, 'steps': 5435, 'loss/train': 2.531430244445801} 01/28/2022 01:40:11 - INFO - codeparrot_training - Step 5436: {'lr': 0.0004937049018824282, 'samples': 1043904, 'steps': 5436, 'loss/train': 2.9388556480407715} 01/28/2022 01:40:15 - INFO - codeparrot_training - Step 5437: {'lr': 0.000493701252614665, 'samples': 1044096, 'steps': 5437, 'loss/train': 3.9670493602752686} 01/28/2022 01:40:21 - INFO - codeparrot_training - Step 5438: {'lr': 0.0004936976023029631, 'samples': 1044288, 'steps': 5438, 'loss/train': 1.8009737133979797} 01/28/2022 01:40:25 - INFO - codeparrot_training - Step 5439: {'lr': 0.000493693950947338, 'samples': 1044480, 'steps': 5439, 'loss/train': 2.349584698677063} 01/28/2022 01:40:29 - INFO - codeparrot_training - Step 5440: {'lr': 0.0004936902985478055, 'samples': 1044672, 'steps': 5440, 'loss/train': 2.76506245136261} 01/28/2022 01:40:33 - INFO - codeparrot_training - Step 5441: {'lr': 0.000493686645104381, 'samples': 1044864, 'steps': 5441, 'loss/train': 2.784539222717285} 01/28/2022 01:40:37 - INFO - codeparrot_training - Step 5442: {'lr': 0.0004936829906170804, 'samples': 1045056, 'steps': 5442, 'loss/train': 2.058957874774933} 01/28/2022 01:40:43 - INFO - codeparrot_training - Step 5443: {'lr': 0.0004936793350859192, 'samples': 1045248, 'steps': 5443, 'loss/train': 3.270264744758606} 01/28/2022 01:40:47 - INFO - codeparrot_training - Step 5444: {'lr': 0.0004936756785109131, 'samples': 1045440, 'steps': 5444, 'loss/train': 3.4982292652130127} 01/28/2022 01:40:51 - INFO - codeparrot_training - Step 5445: {'lr': 0.0004936720208920778, 'samples': 1045632, 'steps': 5445, 'loss/train': 3.887250065803528} 01/28/2022 01:40:55 - INFO - codeparrot_training - Step 5446: {'lr': 0.0004936683622294289, 'samples': 1045824, 'steps': 5446, 'loss/train': 3.6702178716659546} 01/28/2022 01:41:00 - INFO - codeparrot_training - Step 5447: {'lr': 0.0004936647025229822, 'samples': 1046016, 'steps': 5447, 'loss/train': 2.8561529517173767} 01/28/2022 01:41:06 - INFO - codeparrot_training - Step 5448: {'lr': 0.0004936610417727532, 'samples': 1046208, 'steps': 5448, 'loss/train': 4.014194369316101} 01/28/2022 01:41:10 - INFO - codeparrot_training - Step 5449: {'lr': 0.0004936573799787575, 'samples': 1046400, 'steps': 5449, 'loss/train': 3.3884421586990356} 01/28/2022 01:41:14 - INFO - codeparrot_training - Step 5450: {'lr': 0.0004936537171410112, 'samples': 1046592, 'steps': 5450, 'loss/train': 4.711328029632568} 01/28/2022 01:41:19 - INFO - codeparrot_training - Step 5451: {'lr': 0.0004936500532595297, 'samples': 1046784, 'steps': 5451, 'loss/train': 3.5632113218307495} 01/28/2022 01:41:23 - INFO - codeparrot_training - Step 5452: {'lr': 0.0004936463883343287, 'samples': 1046976, 'steps': 5452, 'loss/train': 2.3542195558547974} 01/28/2022 01:41:27 - INFO - codeparrot_training - Step 5453: {'lr': 0.000493642722365424, 'samples': 1047168, 'steps': 5453, 'loss/train': 3.074476718902588} 01/28/2022 01:41:32 - INFO - codeparrot_training - Step 5454: {'lr': 0.0004936390553528313, 'samples': 1047360, 'steps': 5454, 'loss/train': 3.199212670326233} 01/28/2022 01:41:36 - INFO - codeparrot_training - Step 5455: {'lr': 0.0004936353872965661, 'samples': 1047552, 'steps': 5455, 'loss/train': 3.2476229667663574} 01/28/2022 01:41:41 - INFO - codeparrot_training - Step 5456: {'lr': 0.0004936317181966443, 'samples': 1047744, 'steps': 5456, 'loss/train': 0.9184263646602631} 01/28/2022 01:41:45 - INFO - codeparrot_training - Step 5457: {'lr': 0.0004936280480530816, 'samples': 1047936, 'steps': 5457, 'loss/train': 2.99001806974411} 01/28/2022 01:41:49 - INFO - codeparrot_training - Step 5458: {'lr': 0.0004936243768658937, 'samples': 1048128, 'steps': 5458, 'loss/train': 2.69571715593338} 01/28/2022 01:41:55 - INFO - codeparrot_training - Step 5459: {'lr': 0.0004936207046350963, 'samples': 1048320, 'steps': 5459, 'loss/train': 2.939755439758301} 01/28/2022 01:41:59 - INFO - codeparrot_training - Step 5460: {'lr': 0.0004936170313607053, 'samples': 1048512, 'steps': 5460, 'loss/train': 4.627557635307312} 01/28/2022 01:42:03 - INFO - codeparrot_training - Step 5461: {'lr': 0.0004936133570427361, 'samples': 1048704, 'steps': 5461, 'loss/train': 1.6415870189666748} 01/28/2022 01:42:07 - INFO - codeparrot_training - Step 5462: {'lr': 0.0004936096816812046, 'samples': 1048896, 'steps': 5462, 'loss/train': 3.5766706466674805} 01/28/2022 01:42:12 - INFO - codeparrot_training - Step 5463: {'lr': 0.0004936060052761268, 'samples': 1049088, 'steps': 5463, 'loss/train': 2.219289243221283} 01/28/2022 01:42:17 - INFO - codeparrot_training - Step 5464: {'lr': 0.0004936023278275182, 'samples': 1049280, 'steps': 5464, 'loss/train': 3.1524667739868164} 01/28/2022 01:42:21 - INFO - codeparrot_training - Step 5465: {'lr': 0.0004935986493353944, 'samples': 1049472, 'steps': 5465, 'loss/train': 2.8288568258285522} 01/28/2022 01:42:26 - INFO - codeparrot_training - Step 5466: {'lr': 0.0004935949697997715, 'samples': 1049664, 'steps': 5466, 'loss/train': 3.109539270401001} 01/28/2022 01:42:30 - INFO - codeparrot_training - Step 5467: {'lr': 0.000493591289220665, 'samples': 1049856, 'steps': 5467, 'loss/train': 3.693967580795288} 01/28/2022 01:42:34 - INFO - codeparrot_training - Step 5468: {'lr': 0.0004935876075980908, 'samples': 1050048, 'steps': 5468, 'loss/train': 2.4388405680656433} 01/28/2022 01:42:40 - INFO - codeparrot_training - Step 5469: {'lr': 0.0004935839249320647, 'samples': 1050240, 'steps': 5469, 'loss/train': 3.8718398809432983} 01/28/2022 01:42:44 - INFO - codeparrot_training - Step 5470: {'lr': 0.0004935802412226024, 'samples': 1050432, 'steps': 5470, 'loss/train': 3.044331192970276} 01/28/2022 01:42:48 - INFO - codeparrot_training - Step 5471: {'lr': 0.0004935765564697195, 'samples': 1050624, 'steps': 5471, 'loss/train': 1.5272732377052307} 01/28/2022 01:42:53 - INFO - codeparrot_training - Step 5472: {'lr': 0.0004935728706734322, 'samples': 1050816, 'steps': 5472, 'loss/train': 2.4924237728118896} 01/28/2022 01:42:57 - INFO - codeparrot_training - Step 5473: {'lr': 0.000493569183833756, 'samples': 1051008, 'steps': 5473, 'loss/train': 1.6446738839149475} 01/28/2022 01:43:02 - INFO - codeparrot_training - Step 5474: {'lr': 0.0004935654959507068, 'samples': 1051200, 'steps': 5474, 'loss/train': 0.6840707659721375} 01/28/2022 01:43:06 - INFO - codeparrot_training - Step 5475: {'lr': 0.0004935618070243003, 'samples': 1051392, 'steps': 5475, 'loss/train': 3.770999550819397} 01/28/2022 01:43:11 - INFO - codeparrot_training - Step 5476: {'lr': 0.0004935581170545523, 'samples': 1051584, 'steps': 5476, 'loss/train': 2.3772250413894653} 01/28/2022 01:43:15 - INFO - codeparrot_training - Step 5477: {'lr': 0.0004935544260414787, 'samples': 1051776, 'steps': 5477, 'loss/train': 3.0680243968963623} 01/28/2022 01:43:19 - INFO - codeparrot_training - Step 5478: {'lr': 0.0004935507339850953, 'samples': 1051968, 'steps': 5478, 'loss/train': 3.3066705465316772} 01/28/2022 01:43:24 - INFO - codeparrot_training - Step 5479: {'lr': 0.0004935470408854179, 'samples': 1052160, 'steps': 5479, 'loss/train': 3.6336597204208374} 01/28/2022 01:43:28 - INFO - codeparrot_training - Step 5480: {'lr': 0.0004935433467424624, 'samples': 1052352, 'steps': 5480, 'loss/train': 3.151262640953064} 01/28/2022 01:43:33 - INFO - codeparrot_training - Step 5481: {'lr': 0.0004935396515562444, 'samples': 1052544, 'steps': 5481, 'loss/train': 3.098571538925171} 01/28/2022 01:43:37 - INFO - codeparrot_training - Step 5482: {'lr': 0.0004935359553267798, 'samples': 1052736, 'steps': 5482, 'loss/train': 3.5893993377685547} 01/28/2022 01:43:41 - INFO - codeparrot_training - Step 5483: {'lr': 0.0004935322580540847, 'samples': 1052928, 'steps': 5483, 'loss/train': 2.957851231098175} 01/28/2022 01:43:47 - INFO - codeparrot_training - Step 5484: {'lr': 0.0004935285597381747, 'samples': 1053120, 'steps': 5484, 'loss/train': 3.473876953125} 01/28/2022 01:43:51 - INFO - codeparrot_training - Step 5485: {'lr': 0.0004935248603790656, 'samples': 1053312, 'steps': 5485, 'loss/train': 3.159307837486267} 01/28/2022 01:43:55 - INFO - codeparrot_training - Step 5486: {'lr': 0.0004935211599767733, 'samples': 1053504, 'steps': 5486, 'loss/train': 2.982963502407074} 01/28/2022 01:44:00 - INFO - codeparrot_training - Step 5487: {'lr': 0.0004935174585313138, 'samples': 1053696, 'steps': 5487, 'loss/train': 1.3894259333610535} 01/28/2022 01:44:04 - INFO - codeparrot_training - Step 5488: {'lr': 0.0004935137560427027, 'samples': 1053888, 'steps': 5488, 'loss/train': 3.1207011938095093} 01/28/2022 01:44:09 - INFO - codeparrot_training - Step 5489: {'lr': 0.000493510052510956, 'samples': 1054080, 'steps': 5489, 'loss/train': 3.413411021232605} 01/28/2022 01:44:13 - INFO - codeparrot_training - Step 5490: {'lr': 0.0004935063479360897, 'samples': 1054272, 'steps': 5490, 'loss/train': 1.9846004247665405} 01/28/2022 01:44:18 - INFO - codeparrot_training - Step 5491: {'lr': 0.0004935026423181194, 'samples': 1054464, 'steps': 5491, 'loss/train': 2.849593698978424} 01/28/2022 01:44:22 - INFO - codeparrot_training - Step 5492: {'lr': 0.0004934989356570611, 'samples': 1054656, 'steps': 5492, 'loss/train': 2.840217411518097} 01/28/2022 01:44:26 - INFO - codeparrot_training - Step 5493: {'lr': 0.0004934952279529308, 'samples': 1054848, 'steps': 5493, 'loss/train': 2.6624667048454285} 01/28/2022 01:44:32 - INFO - codeparrot_training - Step 5494: {'lr': 0.0004934915192057441, 'samples': 1055040, 'steps': 5494, 'loss/train': 2.5498597025871277} 01/28/2022 01:44:36 - INFO - codeparrot_training - Step 5495: {'lr': 0.0004934878094155172, 'samples': 1055232, 'steps': 5495, 'loss/train': 3.389672040939331} 01/28/2022 01:44:40 - INFO - codeparrot_training - Step 5496: {'lr': 0.0004934840985822657, 'samples': 1055424, 'steps': 5496, 'loss/train': 3.5677539110183716} 01/28/2022 01:44:45 - INFO - codeparrot_training - Step 5497: {'lr': 0.0004934803867060058, 'samples': 1055616, 'steps': 5497, 'loss/train': 3.813566207885742} 01/28/2022 01:44:49 - INFO - codeparrot_training - Step 5498: {'lr': 0.0004934766737867531, 'samples': 1055808, 'steps': 5498, 'loss/train': 3.394246459007263} 01/28/2022 01:44:54 - INFO - codeparrot_training - Step 5499: {'lr': 0.0004934729598245237, 'samples': 1056000, 'steps': 5499, 'loss/train': 3.1436687707901} 01/28/2022 01:44:58 - INFO - codeparrot_training - Step 5500: {'lr': 0.0004934692448193334, 'samples': 1056192, 'steps': 5500, 'loss/train': 3.602718472480774} 01/28/2022 01:45:03 - INFO - codeparrot_training - Step 5501: {'lr': 0.0004934655287711982, 'samples': 1056384, 'steps': 5501, 'loss/train': 2.417172610759735} 01/28/2022 01:45:07 - INFO - codeparrot_training - Step 5502: {'lr': 0.0004934618116801341, 'samples': 1056576, 'steps': 5502, 'loss/train': 3.7492865324020386} 01/28/2022 01:45:11 - INFO - codeparrot_training - Step 5503: {'lr': 0.0004934580935461567, 'samples': 1056768, 'steps': 5503, 'loss/train': 1.8251236081123352} 01/28/2022 01:45:17 - INFO - codeparrot_training - Step 5504: {'lr': 0.0004934543743692822, 'samples': 1056960, 'steps': 5504, 'loss/train': 3.256978154182434} 01/28/2022 01:45:21 - INFO - codeparrot_training - Step 5505: {'lr': 0.0004934506541495265, 'samples': 1057152, 'steps': 5505, 'loss/train': 2.480801045894623} 01/28/2022 01:45:25 - INFO - codeparrot_training - Step 5506: {'lr': 0.0004934469328869056, 'samples': 1057344, 'steps': 5506, 'loss/train': 3.2042055130004883} 01/28/2022 01:45:29 - INFO - codeparrot_training - Step 5507: {'lr': 0.0004934432105814352, 'samples': 1057536, 'steps': 5507, 'loss/train': 3.157915949821472} 01/28/2022 01:45:33 - INFO - codeparrot_training - Step 5508: {'lr': 0.0004934394872331314, 'samples': 1057728, 'steps': 5508, 'loss/train': 2.7367773056030273} 01/28/2022 01:45:39 - INFO - codeparrot_training - Step 5509: {'lr': 0.0004934357628420101, 'samples': 1057920, 'steps': 5509, 'loss/train': 3.1543872356414795} 01/28/2022 01:45:44 - INFO - codeparrot_training - Step 5510: {'lr': 0.0004934320374080874, 'samples': 1058112, 'steps': 5510, 'loss/train': 1.2467890083789825} 01/28/2022 01:45:48 - INFO - codeparrot_training - Step 5511: {'lr': 0.000493428310931379, 'samples': 1058304, 'steps': 5511, 'loss/train': 2.851372718811035} 01/28/2022 01:45:52 - INFO - codeparrot_training - Step 5512: {'lr': 0.0004934245834119013, 'samples': 1058496, 'steps': 5512, 'loss/train': 2.6638705730438232} 01/28/2022 01:45:56 - INFO - codeparrot_training - Step 5513: {'lr': 0.0004934208548496697, 'samples': 1058688, 'steps': 5513, 'loss/train': 3.1677571535110474} 01/28/2022 01:46:01 - INFO - codeparrot_training - Step 5514: {'lr': 0.0004934171252447006, 'samples': 1058880, 'steps': 5514, 'loss/train': 2.4098328351974487} 01/28/2022 01:46:06 - INFO - codeparrot_training - Step 5515: {'lr': 0.0004934133945970097, 'samples': 1059072, 'steps': 5515, 'loss/train': 2.3432310819625854} 01/28/2022 01:46:10 - INFO - codeparrot_training - Step 5516: {'lr': 0.0004934096629066133, 'samples': 1059264, 'steps': 5516, 'loss/train': 3.383828043937683} 01/28/2022 01:46:14 - INFO - codeparrot_training - Step 5517: {'lr': 0.000493405930173527, 'samples': 1059456, 'steps': 5517, 'loss/train': 2.2924840450286865} 01/28/2022 01:46:18 - INFO - codeparrot_training - Step 5518: {'lr': 0.0004934021963977671, 'samples': 1059648, 'steps': 5518, 'loss/train': 2.11789333820343} 01/28/2022 01:46:24 - INFO - codeparrot_training - Step 5519: {'lr': 0.0004933984615793494, 'samples': 1059840, 'steps': 5519, 'loss/train': 2.344379961490631} 01/28/2022 01:46:29 - INFO - codeparrot_training - Step 5520: {'lr': 0.0004933947257182901, 'samples': 1060032, 'steps': 5520, 'loss/train': 3.100092887878418} 01/28/2022 01:46:33 - INFO - codeparrot_training - Step 5521: {'lr': 0.000493390988814605, 'samples': 1060224, 'steps': 5521, 'loss/train': 2.7676244974136353} 01/28/2022 01:46:37 - INFO - codeparrot_training - Step 5522: {'lr': 0.0004933872508683101, 'samples': 1060416, 'steps': 5522, 'loss/train': 2.616074860095978} 01/28/2022 01:46:41 - INFO - codeparrot_training - Step 5523: {'lr': 0.0004933835118794217, 'samples': 1060608, 'steps': 5523, 'loss/train': 10.909235715866089} 01/28/2022 01:46:46 - INFO - codeparrot_training - Step 5524: {'lr': 0.0004933797718479555, 'samples': 1060800, 'steps': 5524, 'loss/train': 2.9899678230285645} 01/28/2022 01:46:51 - INFO - codeparrot_training - Step 5525: {'lr': 0.0004933760307739277, 'samples': 1060992, 'steps': 5525, 'loss/train': 5.394487023353577} 01/28/2022 01:46:55 - INFO - codeparrot_training - Step 5526: {'lr': 0.0004933722886573542, 'samples': 1061184, 'steps': 5526, 'loss/train': 3.0045511722564697} 01/28/2022 01:46:59 - INFO - codeparrot_training - Step 5527: {'lr': 0.0004933685454982511, 'samples': 1061376, 'steps': 5527, 'loss/train': 3.034864068031311} 01/28/2022 01:47:03 - INFO - codeparrot_training - Step 5528: {'lr': 0.0004933648012966344, 'samples': 1061568, 'steps': 5528, 'loss/train': 3.120301365852356} 01/28/2022 01:47:09 - INFO - codeparrot_training - Step 5529: {'lr': 0.0004933610560525203, 'samples': 1061760, 'steps': 5529, 'loss/train': 2.838742733001709} 01/28/2022 01:47:13 - INFO - codeparrot_training - Step 5530: {'lr': 0.0004933573097659246, 'samples': 1061952, 'steps': 5530, 'loss/train': 1.5848336219787598} 01/28/2022 01:47:17 - INFO - codeparrot_training - Step 5531: {'lr': 0.0004933535624368634, 'samples': 1062144, 'steps': 5531, 'loss/train': 1.8774837255477905} 01/28/2022 01:47:21 - INFO - codeparrot_training - Step 5532: {'lr': 0.0004933498140653529, 'samples': 1062336, 'steps': 5532, 'loss/train': 1.5942553281784058} 01/28/2022 01:47:25 - INFO - codeparrot_training - Step 5533: {'lr': 0.0004933460646514092, 'samples': 1062528, 'steps': 5533, 'loss/train': 3.303270936012268} 01/28/2022 01:47:31 - INFO - codeparrot_training - Step 5534: {'lr': 0.000493342314195048, 'samples': 1062720, 'steps': 5534, 'loss/train': 3.3801158666610718} 01/28/2022 01:47:35 - INFO - codeparrot_training - Step 5535: {'lr': 0.0004933385626962858, 'samples': 1062912, 'steps': 5535, 'loss/train': 2.712700366973877} 01/28/2022 01:47:39 - INFO - codeparrot_training - Step 5536: {'lr': 0.0004933348101551383, 'samples': 1063104, 'steps': 5536, 'loss/train': 1.7251794934272766} 01/28/2022 01:47:43 - INFO - codeparrot_training - Step 5537: {'lr': 0.0004933310565716218, 'samples': 1063296, 'steps': 5537, 'loss/train': 3.0150368213653564} 01/28/2022 01:47:48 - INFO - codeparrot_training - Step 5538: {'lr': 0.0004933273019457524, 'samples': 1063488, 'steps': 5538, 'loss/train': 5.386903882026672} 01/28/2022 01:47:53 - INFO - codeparrot_training - Step 5539: {'lr': 0.0004933235462775459, 'samples': 1063680, 'steps': 5539, 'loss/train': 3.1341272592544556} 01/28/2022 01:47:58 - INFO - codeparrot_training - Step 5540: {'lr': 0.0004933197895670187, 'samples': 1063872, 'steps': 5540, 'loss/train': 2.9873244166374207} 01/28/2022 01:48:02 - INFO - codeparrot_training - Step 5541: {'lr': 0.0004933160318141869, 'samples': 1064064, 'steps': 5541, 'loss/train': 3.0948193073272705} 01/28/2022 01:48:06 - INFO - codeparrot_training - Step 5542: {'lr': 0.0004933122730190663, 'samples': 1064256, 'steps': 5542, 'loss/train': 3.3170703649520874} 01/28/2022 01:48:10 - INFO - codeparrot_training - Step 5543: {'lr': 0.0004933085131816733, 'samples': 1064448, 'steps': 5543, 'loss/train': 2.1715482473373413} 01/28/2022 01:48:16 - INFO - codeparrot_training - Step 5544: {'lr': 0.0004933047523020239, 'samples': 1064640, 'steps': 5544, 'loss/train': 2.6710453033447266} 01/28/2022 01:48:20 - INFO - codeparrot_training - Step 5545: {'lr': 0.0004933009903801341, 'samples': 1064832, 'steps': 5545, 'loss/train': 1.9877554178237915} 01/28/2022 01:48:24 - INFO - codeparrot_training - Step 5546: {'lr': 0.0004932972274160202, 'samples': 1065024, 'steps': 5546, 'loss/train': 2.4998350739479065} 01/28/2022 01:48:28 - INFO - codeparrot_training - Step 5547: {'lr': 0.0004932934634096982, 'samples': 1065216, 'steps': 5547, 'loss/train': 2.774084508419037} 01/28/2022 01:48:32 - INFO - codeparrot_training - Step 5548: {'lr': 0.0004932896983611843, 'samples': 1065408, 'steps': 5548, 'loss/train': 3.2313759326934814} 01/28/2022 01:48:38 - INFO - codeparrot_training - Step 5549: {'lr': 0.0004932859322704944, 'samples': 1065600, 'steps': 5549, 'loss/train': 1.3591049909591675} 01/28/2022 01:48:42 - INFO - codeparrot_training - Step 5550: {'lr': 0.000493282165137645, 'samples': 1065792, 'steps': 5550, 'loss/train': 2.410228192806244} 01/28/2022 01:48:46 - INFO - codeparrot_training - Step 5551: {'lr': 0.0004932783969626521, 'samples': 1065984, 'steps': 5551, 'loss/train': 2.9759475588798523} 01/28/2022 01:48:50 - INFO - codeparrot_training - Step 5552: {'lr': 0.0004932746277455317, 'samples': 1066176, 'steps': 5552, 'loss/train': 2.4711248874664307} 01/28/2022 01:48:55 - INFO - codeparrot_training - Step 5553: {'lr': 0.0004932708574863, 'samples': 1066368, 'steps': 5553, 'loss/train': 3.2120261192321777} 01/28/2022 01:49:01 - INFO - codeparrot_training - Step 5554: {'lr': 0.0004932670861849733, 'samples': 1066560, 'steps': 5554, 'loss/train': 3.2405329942703247} 01/28/2022 01:49:05 - INFO - codeparrot_training - Step 5555: {'lr': 0.0004932633138415675, 'samples': 1066752, 'steps': 5555, 'loss/train': 2.419900059700012} 01/28/2022 01:49:09 - INFO - codeparrot_training - Step 5556: {'lr': 0.000493259540456099, 'samples': 1066944, 'steps': 5556, 'loss/train': 3.308958649635315} 01/28/2022 01:49:13 - INFO - codeparrot_training - Step 5557: {'lr': 0.0004932557660285839, 'samples': 1067136, 'steps': 5557, 'loss/train': 1.4034833908081055} 01/28/2022 01:49:18 - INFO - codeparrot_training - Step 5558: {'lr': 0.0004932519905590383, 'samples': 1067328, 'steps': 5558, 'loss/train': 2.685313582420349} 01/28/2022 01:49:23 - INFO - codeparrot_training - Step 5559: {'lr': 0.0004932482140474785, 'samples': 1067520, 'steps': 5559, 'loss/train': 3.183581829071045} 01/28/2022 01:49:27 - INFO - codeparrot_training - Step 5560: {'lr': 0.0004932444364939204, 'samples': 1067712, 'steps': 5560, 'loss/train': 3.332339644432068} 01/28/2022 01:49:31 - INFO - codeparrot_training - Step 5561: {'lr': 0.0004932406578983806, 'samples': 1067904, 'steps': 5561, 'loss/train': 3.7386964559555054} 01/28/2022 01:49:35 - INFO - codeparrot_training - Step 5562: {'lr': 0.0004932368782608749, 'samples': 1068096, 'steps': 5562, 'loss/train': 3.0089231729507446} 01/28/2022 01:49:40 - INFO - codeparrot_training - Step 5563: {'lr': 0.0004932330975814198, 'samples': 1068288, 'steps': 5563, 'loss/train': 2.0007094144821167} 01/28/2022 01:49:45 - INFO - codeparrot_training - Step 5564: {'lr': 0.0004932293158600312, 'samples': 1068480, 'steps': 5564, 'loss/train': 3.4958009719848633} 01/28/2022 01:49:50 - INFO - codeparrot_training - Step 5565: {'lr': 0.0004932255330967255, 'samples': 1068672, 'steps': 5565, 'loss/train': 2.8621888160705566} 01/28/2022 01:49:54 - INFO - codeparrot_training - Step 5566: {'lr': 0.0004932217492915189, 'samples': 1068864, 'steps': 5566, 'loss/train': 3.200130343437195} 01/28/2022 01:49:58 - INFO - codeparrot_training - Step 5567: {'lr': 0.0004932179644444274, 'samples': 1069056, 'steps': 5567, 'loss/train': 1.9273483157157898} 01/28/2022 01:50:02 - INFO - codeparrot_training - Step 5568: {'lr': 0.0004932141785554676, 'samples': 1069248, 'steps': 5568, 'loss/train': 2.868194282054901} 01/28/2022 01:50:08 - INFO - codeparrot_training - Step 5569: {'lr': 0.0004932103916246553, 'samples': 1069440, 'steps': 5569, 'loss/train': 2.655991494655609} 01/28/2022 01:50:12 - INFO - codeparrot_training - Step 5570: {'lr': 0.000493206603652007, 'samples': 1069632, 'steps': 5570, 'loss/train': 3.6968071460723877} 01/28/2022 01:50:16 - INFO - codeparrot_training - Step 5571: {'lr': 0.0004932028146375388, 'samples': 1069824, 'steps': 5571, 'loss/train': 1.4478173553943634} 01/28/2022 01:50:20 - INFO - codeparrot_training - Step 5572: {'lr': 0.000493199024581267, 'samples': 1070016, 'steps': 5572, 'loss/train': 3.495955467224121} 01/28/2022 01:50:26 - INFO - codeparrot_training - Step 5573: {'lr': 0.0004931952334832077, 'samples': 1070208, 'steps': 5573, 'loss/train': 3.1394888162612915} 01/28/2022 01:50:30 - INFO - codeparrot_training - Step 5574: {'lr': 0.0004931914413433773, 'samples': 1070400, 'steps': 5574, 'loss/train': 3.2729365825653076} 01/28/2022 01:50:34 - INFO - codeparrot_training - Step 5575: {'lr': 0.0004931876481617921, 'samples': 1070592, 'steps': 5575, 'loss/train': 3.1981894969940186} 01/28/2022 01:50:38 - INFO - codeparrot_training - Step 5576: {'lr': 0.0004931838539384681, 'samples': 1070784, 'steps': 5576, 'loss/train': 3.205017328262329} 01/28/2022 01:50:43 - INFO - codeparrot_training - Step 5577: {'lr': 0.0004931800586734218, 'samples': 1070976, 'steps': 5577, 'loss/train': 2.581424653530121} 01/28/2022 01:50:49 - INFO - codeparrot_training - Step 5578: {'lr': 0.0004931762623666692, 'samples': 1071168, 'steps': 5578, 'loss/train': 3.172047972679138} 01/28/2022 01:50:53 - INFO - codeparrot_training - Step 5579: {'lr': 0.0004931724650182268, 'samples': 1071360, 'steps': 5579, 'loss/train': 2.742662787437439} 01/28/2022 01:50:57 - INFO - codeparrot_training - Step 5580: {'lr': 0.0004931686666281108, 'samples': 1071552, 'steps': 5580, 'loss/train': 2.675881505012512} 01/28/2022 01:51:02 - INFO - codeparrot_training - Step 5581: {'lr': 0.0004931648671963373, 'samples': 1071744, 'steps': 5581, 'loss/train': 3.8658331632614136} 01/28/2022 01:51:06 - INFO - codeparrot_training - Step 5582: {'lr': 0.000493161066722923, 'samples': 1071936, 'steps': 5582, 'loss/train': 3.013466477394104} 01/28/2022 01:51:10 - INFO - codeparrot_training - Step 5583: {'lr': 0.0004931572652078837, 'samples': 1072128, 'steps': 5583, 'loss/train': 3.8718327283859253} 01/28/2022 01:51:15 - INFO - codeparrot_training - Step 5584: {'lr': 0.0004931534626512359, 'samples': 1072320, 'steps': 5584, 'loss/train': 2.838295340538025} 01/28/2022 01:51:20 - INFO - codeparrot_training - Step 5585: {'lr': 0.0004931496590529959, 'samples': 1072512, 'steps': 5585, 'loss/train': 2.3975606560707092} 01/28/2022 01:51:24 - INFO - codeparrot_training - Step 5586: {'lr': 0.0004931458544131799, 'samples': 1072704, 'steps': 5586, 'loss/train': 1.8480529189109802} 01/28/2022 01:51:28 - INFO - codeparrot_training - Step 5587: {'lr': 0.0004931420487318044, 'samples': 1072896, 'steps': 5587, 'loss/train': 2.8963987827301025} 01/28/2022 01:51:32 - INFO - codeparrot_training - Step 5588: {'lr': 0.0004931382420088855, 'samples': 1073088, 'steps': 5588, 'loss/train': 3.179741621017456} 01/28/2022 01:51:38 - INFO - codeparrot_training - Step 5589: {'lr': 0.0004931344342444396, 'samples': 1073280, 'steps': 5589, 'loss/train': 4.178574085235596} 01/28/2022 01:51:42 - INFO - codeparrot_training - Step 5590: {'lr': 0.000493130625438483, 'samples': 1073472, 'steps': 5590, 'loss/train': 2.0320258140563965} 01/28/2022 01:51:46 - INFO - codeparrot_training - Step 5591: {'lr': 0.000493126815591032, 'samples': 1073664, 'steps': 5591, 'loss/train': 2.1247411966323853} 01/28/2022 01:51:50 - INFO - codeparrot_training - Step 5592: {'lr': 0.0004931230047021028, 'samples': 1073856, 'steps': 5592, 'loss/train': 6.337348937988281} 01/28/2022 01:51:55 - INFO - codeparrot_training - Step 5593: {'lr': 0.000493119192771712, 'samples': 1074048, 'steps': 5593, 'loss/train': 2.5036012530326843} 01/28/2022 01:52:00 - INFO - codeparrot_training - Step 5594: {'lr': 0.0004931153797998757, 'samples': 1074240, 'steps': 5594, 'loss/train': 2.6087565422058105} 01/28/2022 01:52:04 - INFO - codeparrot_training - Step 5595: {'lr': 0.0004931115657866103, 'samples': 1074432, 'steps': 5595, 'loss/train': 2.550797939300537} 01/28/2022 01:52:08 - INFO - codeparrot_training - Step 5596: {'lr': 0.0004931077507319322, 'samples': 1074624, 'steps': 5596, 'loss/train': 2.4681578278541565} 01/28/2022 01:52:13 - INFO - codeparrot_training - Step 5597: {'lr': 0.0004931039346358577, 'samples': 1074816, 'steps': 5597, 'loss/train': 2.247349977493286} 01/28/2022 01:52:17 - INFO - codeparrot_training - Step 5598: {'lr': 0.0004931001174984032, 'samples': 1075008, 'steps': 5598, 'loss/train': 2.9263896346092224} 01/28/2022 01:52:23 - INFO - codeparrot_training - Step 5599: {'lr': 0.0004930962993195848, 'samples': 1075200, 'steps': 5599, 'loss/train': 4.730226159095764} 01/28/2022 01:52:27 - INFO - codeparrot_training - Step 5600: {'lr': 0.0004930924800994192, 'samples': 1075392, 'steps': 5600, 'loss/train': 3.6889833211898804} 01/28/2022 01:52:31 - INFO - codeparrot_training - Step 5601: {'lr': 0.0004930886598379225, 'samples': 1075584, 'steps': 5601, 'loss/train': 3.1030004024505615} 01/28/2022 01:52:36 - INFO - codeparrot_training - Step 5602: {'lr': 0.0004930848385351112, 'samples': 1075776, 'steps': 5602, 'loss/train': 3.499601125717163} 01/28/2022 01:52:40 - INFO - codeparrot_training - Step 5603: {'lr': 0.0004930810161910017, 'samples': 1075968, 'steps': 5603, 'loss/train': 3.0240933895111084} 01/28/2022 01:52:45 - INFO - codeparrot_training - Step 5604: {'lr': 0.0004930771928056102, 'samples': 1076160, 'steps': 5604, 'loss/train': 2.381314516067505} 01/28/2022 01:52:49 - INFO - codeparrot_training - Step 5605: {'lr': 0.0004930733683789533, 'samples': 1076352, 'steps': 5605, 'loss/train': 2.481928825378418} 01/28/2022 01:52:54 - INFO - codeparrot_training - Step 5606: {'lr': 0.0004930695429110473, 'samples': 1076544, 'steps': 5606, 'loss/train': 2.9062331914901733} 01/28/2022 01:52:58 - INFO - codeparrot_training - Step 5607: {'lr': 0.0004930657164019085, 'samples': 1076736, 'steps': 5607, 'loss/train': 3.1121174097061157} 01/28/2022 01:53:02 - INFO - codeparrot_training - Step 5608: {'lr': 0.0004930618888515534, 'samples': 1076928, 'steps': 5608, 'loss/train': 2.5559951663017273} 01/28/2022 01:53:08 - INFO - codeparrot_training - Step 5609: {'lr': 0.0004930580602599983, 'samples': 1077120, 'steps': 5609, 'loss/train': 2.673461079597473} 01/28/2022 01:53:12 - INFO - codeparrot_training - Step 5610: {'lr': 0.0004930542306272596, 'samples': 1077312, 'steps': 5610, 'loss/train': 2.74900084733963} 01/28/2022 01:53:16 - INFO - codeparrot_training - Step 5611: {'lr': 0.0004930503999533538, 'samples': 1077504, 'steps': 5611, 'loss/train': 3.379238247871399} 01/28/2022 01:53:21 - INFO - codeparrot_training - Step 5612: {'lr': 0.0004930465682382973, 'samples': 1077696, 'steps': 5612, 'loss/train': 3.331598997116089} 01/28/2022 01:53:25 - INFO - codeparrot_training - Step 5613: {'lr': 0.0004930427354821064, 'samples': 1077888, 'steps': 5613, 'loss/train': 2.6084225177764893} 01/28/2022 01:53:30 - INFO - codeparrot_training - Step 5614: {'lr': 0.0004930389016847977, 'samples': 1078080, 'steps': 5614, 'loss/train': 2.3005292415618896} 01/28/2022 01:53:34 - INFO - codeparrot_training - Step 5615: {'lr': 0.0004930350668463874, 'samples': 1078272, 'steps': 5615, 'loss/train': 1.0181812942028046} 01/28/2022 01:53:39 - INFO - codeparrot_training - Step 5616: {'lr': 0.0004930312309668922, 'samples': 1078464, 'steps': 5616, 'loss/train': 2.6661665439605713} 01/28/2022 01:53:43 - INFO - codeparrot_training - Step 5617: {'lr': 0.0004930273940463283, 'samples': 1078656, 'steps': 5617, 'loss/train': 3.416642189025879} 01/28/2022 01:53:47 - INFO - codeparrot_training - Step 5618: {'lr': 0.0004930235560847121, 'samples': 1078848, 'steps': 5618, 'loss/train': 3.1339112520217896} 01/28/2022 01:53:53 - INFO - codeparrot_training - Step 5619: {'lr': 0.0004930197170820603, 'samples': 1079040, 'steps': 5619, 'loss/train': 3.070531725883484} 01/28/2022 01:53:57 - INFO - codeparrot_training - Step 5620: {'lr': 0.0004930158770383891, 'samples': 1079232, 'steps': 5620, 'loss/train': 3.389969229698181} 01/28/2022 01:54:01 - INFO - codeparrot_training - Step 5621: {'lr': 0.0004930120359537153, 'samples': 1079424, 'steps': 5621, 'loss/train': 2.53373783826828} 01/28/2022 01:54:05 - INFO - codeparrot_training - Step 5622: {'lr': 0.0004930081938280548, 'samples': 1079616, 'steps': 5622, 'loss/train': 3.7930591106414795} 01/28/2022 01:54:10 - INFO - codeparrot_training - Step 5623: {'lr': 0.0004930043506614245, 'samples': 1079808, 'steps': 5623, 'loss/train': 2.8752923011779785} 01/28/2022 01:54:14 - INFO - codeparrot_training - Step 5624: {'lr': 0.0004930005064538406, 'samples': 1080000, 'steps': 5624, 'loss/train': 0.8244012594223022} 01/28/2022 01:54:20 - INFO - codeparrot_training - Step 5625: {'lr': 0.0004929966612053199, 'samples': 1080192, 'steps': 5625, 'loss/train': 2.3320818543434143} 01/28/2022 01:54:25 - INFO - codeparrot_training - Step 5626: {'lr': 0.0004929928149158785, 'samples': 1080384, 'steps': 5626, 'loss/train': 5.638168573379517} 01/28/2022 01:54:29 - INFO - codeparrot_training - Step 5627: {'lr': 0.0004929889675855332, 'samples': 1080576, 'steps': 5627, 'loss/train': 2.63063907623291} 01/28/2022 01:54:33 - INFO - codeparrot_training - Step 5628: {'lr': 0.0004929851192143001, 'samples': 1080768, 'steps': 5628, 'loss/train': 2.6550339460372925} 01/28/2022 01:54:37 - INFO - codeparrot_training - Step 5629: {'lr': 0.0004929812698021961, 'samples': 1080960, 'steps': 5629, 'loss/train': 2.9492939114570618} 01/28/2022 01:54:43 - INFO - codeparrot_training - Step 5630: {'lr': 0.0004929774193492373, 'samples': 1081152, 'steps': 5630, 'loss/train': 2.799137771129608} 01/28/2022 01:54:47 - INFO - codeparrot_training - Step 5631: {'lr': 0.0004929735678554406, 'samples': 1081344, 'steps': 5631, 'loss/train': 2.440875828266144} 01/28/2022 01:54:51 - INFO - codeparrot_training - Step 5632: {'lr': 0.0004929697153208221, 'samples': 1081536, 'steps': 5632, 'loss/train': 3.126231551170349} 01/28/2022 01:54:56 - INFO - codeparrot_training - Step 5633: {'lr': 0.0004929658617453986, 'samples': 1081728, 'steps': 5633, 'loss/train': 2.7792747616767883} 01/28/2022 01:55:00 - INFO - codeparrot_training - Step 5634: {'lr': 0.0004929620071291865, 'samples': 1081920, 'steps': 5634, 'loss/train': 9.648634672164917} 01/28/2022 01:55:05 - INFO - codeparrot_training - Step 5635: {'lr': 0.0004929581514722023, 'samples': 1082112, 'steps': 5635, 'loss/train': 2.4870318174362183} 01/28/2022 01:55:09 - INFO - codeparrot_training - Step 5636: {'lr': 0.0004929542947744625, 'samples': 1082304, 'steps': 5636, 'loss/train': 3.176398515701294} 01/28/2022 01:55:13 - INFO - codeparrot_training - Step 5637: {'lr': 0.0004929504370359837, 'samples': 1082496, 'steps': 5637, 'loss/train': 0.16715367883443832} 01/28/2022 01:55:18 - INFO - codeparrot_training - Step 5638: {'lr': 0.0004929465782567824, 'samples': 1082688, 'steps': 5638, 'loss/train': 4.291820526123047} 01/28/2022 01:55:22 - INFO - codeparrot_training - Step 5639: {'lr': 0.000492942718436875, 'samples': 1082880, 'steps': 5639, 'loss/train': 2.1155576705932617} 01/28/2022 01:55:28 - INFO - codeparrot_training - Step 5640: {'lr': 0.0004929388575762782, 'samples': 1083072, 'steps': 5640, 'loss/train': 3.164024233818054} 01/28/2022 01:55:32 - INFO - codeparrot_training - Step 5641: {'lr': 0.0004929349956750085, 'samples': 1083264, 'steps': 5641, 'loss/train': 1.3187609016895294} 01/28/2022 01:55:36 - INFO - codeparrot_training - Step 5642: {'lr': 0.0004929311327330823, 'samples': 1083456, 'steps': 5642, 'loss/train': 2.4805485606193542} 01/28/2022 01:55:40 - INFO - codeparrot_training - Step 5643: {'lr': 0.0004929272687505163, 'samples': 1083648, 'steps': 5643, 'loss/train': 2.946821630001068} 01/28/2022 01:55:45 - INFO - codeparrot_training - Step 5644: {'lr': 0.0004929234037273271, 'samples': 1083840, 'steps': 5644, 'loss/train': 4.144093036651611} 01/28/2022 01:55:50 - INFO - codeparrot_training - Step 5645: {'lr': 0.0004929195376635311, 'samples': 1084032, 'steps': 5645, 'loss/train': 2.763274133205414} 01/28/2022 01:55:54 - INFO - codeparrot_training - Step 5646: {'lr': 0.000492915670559145, 'samples': 1084224, 'steps': 5646, 'loss/train': 3.0329779386520386} 01/28/2022 01:55:58 - INFO - codeparrot_training - Step 5647: {'lr': 0.0004929118024141853, 'samples': 1084416, 'steps': 5647, 'loss/train': 2.085492968559265} 01/28/2022 01:56:03 - INFO - codeparrot_training - Step 5648: {'lr': 0.0004929079332286685, 'samples': 1084608, 'steps': 5648, 'loss/train': 2.3570346236228943} 01/28/2022 01:56:07 - INFO - codeparrot_training - Step 5649: {'lr': 0.0004929040630026112, 'samples': 1084800, 'steps': 5649, 'loss/train': 2.893140435218811} 01/28/2022 01:56:13 - INFO - codeparrot_training - Step 5650: {'lr': 0.0004929001917360302, 'samples': 1084992, 'steps': 5650, 'loss/train': 1.7102605104446411} 01/28/2022 01:56:17 - INFO - codeparrot_training - Step 5651: {'lr': 0.0004928963194289419, 'samples': 1085184, 'steps': 5651, 'loss/train': 1.9072397947311401} 01/28/2022 01:56:21 - INFO - codeparrot_training - Step 5652: {'lr': 0.0004928924460813627, 'samples': 1085376, 'steps': 5652, 'loss/train': 3.203503131866455} 01/28/2022 01:56:26 - INFO - codeparrot_training - Step 5653: {'lr': 0.0004928885716933096, 'samples': 1085568, 'steps': 5653, 'loss/train': 3.4183162450790405} 01/28/2022 01:56:30 - INFO - codeparrot_training - Step 5654: {'lr': 0.0004928846962647988, 'samples': 1085760, 'steps': 5654, 'loss/train': 1.941984236240387} 01/28/2022 01:56:35 - INFO - codeparrot_training - Step 5655: {'lr': 0.0004928808197958472, 'samples': 1085952, 'steps': 5655, 'loss/train': 1.420243263244629} 01/28/2022 01:56:39 - INFO - codeparrot_training - Step 5656: {'lr': 0.0004928769422864712, 'samples': 1086144, 'steps': 5656, 'loss/train': 3.0286284685134888} 01/28/2022 01:56:43 - INFO - codeparrot_training - Step 5657: {'lr': 0.0004928730637366877, 'samples': 1086336, 'steps': 5657, 'loss/train': 2.493980884552002} 01/28/2022 01:56:48 - INFO - codeparrot_training - Step 5658: {'lr': 0.000492869184146513, 'samples': 1086528, 'steps': 5658, 'loss/train': 2.7588483095169067} 01/28/2022 01:56:52 - INFO - codeparrot_training - Step 5659: {'lr': 0.0004928653035159638, 'samples': 1086720, 'steps': 5659, 'loss/train': 1.3572562336921692} 01/28/2022 01:56:57 - INFO - codeparrot_training - Step 5660: {'lr': 0.0004928614218450568, 'samples': 1086912, 'steps': 5660, 'loss/train': 3.764301896095276} 01/28/2022 01:57:01 - INFO - codeparrot_training - Step 5661: {'lr': 0.0004928575391338085, 'samples': 1087104, 'steps': 5661, 'loss/train': 3.5239527225494385} 01/28/2022 01:57:06 - INFO - codeparrot_training - Step 5662: {'lr': 0.0004928536553822357, 'samples': 1087296, 'steps': 5662, 'loss/train': 3.46782124042511} 01/28/2022 01:57:10 - INFO - codeparrot_training - Step 5663: {'lr': 0.0004928497705903549, 'samples': 1087488, 'steps': 5663, 'loss/train': 2.6024338603019714} 01/28/2022 01:57:14 - INFO - codeparrot_training - Step 5664: {'lr': 0.0004928458847581828, 'samples': 1087680, 'steps': 5664, 'loss/train': 1.7810996174812317} 01/28/2022 01:57:19 - INFO - codeparrot_training - Step 5665: {'lr': 0.0004928419978857361, 'samples': 1087872, 'steps': 5665, 'loss/train': 2.9544244408607483} 01/28/2022 01:57:24 - INFO - codeparrot_training - Step 5666: {'lr': 0.0004928381099730314, 'samples': 1088064, 'steps': 5666, 'loss/train': 3.115535616874695} 01/28/2022 01:57:28 - INFO - codeparrot_training - Step 5667: {'lr': 0.0004928342210200853, 'samples': 1088256, 'steps': 5667, 'loss/train': 2.3063238859176636} 01/28/2022 01:57:32 - INFO - codeparrot_training - Step 5668: {'lr': 0.0004928303310269145, 'samples': 1088448, 'steps': 5668, 'loss/train': 3.831030249595642} 01/28/2022 01:57:36 - INFO - codeparrot_training - Step 5669: {'lr': 0.0004928264399935357, 'samples': 1088640, 'steps': 5669, 'loss/train': 3.052344560623169} 01/28/2022 01:57:42 - INFO - codeparrot_training - Step 5670: {'lr': 0.0004928225479199655, 'samples': 1088832, 'steps': 5670, 'loss/train': 2.744013011455536} 01/28/2022 01:57:47 - INFO - codeparrot_training - Step 5671: {'lr': 0.0004928186548062206, 'samples': 1089024, 'steps': 5671, 'loss/train': 3.184583902359009} 01/28/2022 01:57:51 - INFO - codeparrot_training - Step 5672: {'lr': 0.0004928147606523179, 'samples': 1089216, 'steps': 5672, 'loss/train': 2.919587552547455} 01/28/2022 01:57:55 - INFO - codeparrot_training - Step 5673: {'lr': 0.0004928108654582736, 'samples': 1089408, 'steps': 5673, 'loss/train': 2.4765953421592712} 01/28/2022 01:57:59 - INFO - codeparrot_training - Step 5674: {'lr': 0.0004928069692241048, 'samples': 1089600, 'steps': 5674, 'loss/train': 3.021830677986145} 01/28/2022 01:58:05 - INFO - codeparrot_training - Step 5675: {'lr': 0.000492803071949828, 'samples': 1089792, 'steps': 5675, 'loss/train': 3.412197232246399} 01/28/2022 01:58:09 - INFO - codeparrot_training - Step 5676: {'lr': 0.0004927991736354599, 'samples': 1089984, 'steps': 5676, 'loss/train': 2.9203766584396362} 01/28/2022 01:58:13 - INFO - codeparrot_training - Step 5677: {'lr': 0.0004927952742810173, 'samples': 1090176, 'steps': 5677, 'loss/train': 2.634082853794098} 01/28/2022 01:58:17 - INFO - codeparrot_training - Step 5678: {'lr': 0.0004927913738865167, 'samples': 1090368, 'steps': 5678, 'loss/train': 3.282035708427429} 01/28/2022 01:58:22 - INFO - codeparrot_training - Step 5679: {'lr': 0.0004927874724519751, 'samples': 1090560, 'steps': 5679, 'loss/train': 2.355598032474518} 01/28/2022 01:58:27 - INFO - codeparrot_training - Step 5680: {'lr': 0.000492783569977409, 'samples': 1090752, 'steps': 5680, 'loss/train': 3.327501654624939} 01/28/2022 01:58:31 - INFO - codeparrot_training - Step 5681: {'lr': 0.0004927796664628353, 'samples': 1090944, 'steps': 5681, 'loss/train': 6.940325975418091} 01/28/2022 01:58:36 - INFO - codeparrot_training - Step 5682: {'lr': 0.0004927757619082704, 'samples': 1091136, 'steps': 5682, 'loss/train': 3.969193696975708} 01/28/2022 01:58:40 - INFO - codeparrot_training - Step 5683: {'lr': 0.0004927718563137313, 'samples': 1091328, 'steps': 5683, 'loss/train': 4.111668348312378} 01/28/2022 01:58:44 - INFO - codeparrot_training - Step 5684: {'lr': 0.0004927679496792347, 'samples': 1091520, 'steps': 5684, 'loss/train': 3.123197078704834} 01/28/2022 01:58:48 - INFO - codeparrot_training - Step 5685: {'lr': 0.0004927640420047973, 'samples': 1091712, 'steps': 5685, 'loss/train': 2.306142568588257} 01/28/2022 01:58:55 - INFO - codeparrot_training - Step 5686: {'lr': 0.0004927601332904358, 'samples': 1091904, 'steps': 5686, 'loss/train': 3.108018636703491} 01/28/2022 01:58:59 - INFO - codeparrot_training - Step 5687: {'lr': 0.0004927562235361669, 'samples': 1092096, 'steps': 5687, 'loss/train': 3.331034302711487} 01/28/2022 01:59:03 - INFO - codeparrot_training - Step 5688: {'lr': 0.0004927523127420076, 'samples': 1092288, 'steps': 5688, 'loss/train': 1.7718620896339417} 01/28/2022 01:59:07 - INFO - codeparrot_training - Step 5689: {'lr': 0.0004927484009079743, 'samples': 1092480, 'steps': 5689, 'loss/train': 2.763248562812805} 01/28/2022 01:59:11 - INFO - codeparrot_training - Step 5690: {'lr': 0.000492744488034084, 'samples': 1092672, 'steps': 5690, 'loss/train': 2.5354095697402954} 01/28/2022 01:59:17 - INFO - codeparrot_training - Step 5691: {'lr': 0.0004927405741203534, 'samples': 1092864, 'steps': 5691, 'loss/train': 3.50847065448761} 01/28/2022 01:59:21 - INFO - codeparrot_training - Step 5692: {'lr': 0.0004927366591667993, 'samples': 1093056, 'steps': 5692, 'loss/train': 4.198776125907898} 01/28/2022 01:59:25 - INFO - codeparrot_training - Step 5693: {'lr': 0.0004927327431734383, 'samples': 1093248, 'steps': 5693, 'loss/train': 3.736510992050171} 01/28/2022 01:59:30 - INFO - codeparrot_training - Step 5694: {'lr': 0.0004927288261402875, 'samples': 1093440, 'steps': 5694, 'loss/train': 2.54536235332489} 01/28/2022 01:59:36 - INFO - codeparrot_training - Step 5695: {'lr': 0.0004927249080673633, 'samples': 1093632, 'steps': 5695, 'loss/train': 3.2761906385421753} 01/28/2022 01:59:40 - INFO - codeparrot_training - Step 5696: {'lr': 0.0004927209889546828, 'samples': 1093824, 'steps': 5696, 'loss/train': 3.390537142753601} 01/28/2022 01:59:44 - INFO - codeparrot_training - Step 5697: {'lr': 0.0004927170688022625, 'samples': 1094016, 'steps': 5697, 'loss/train': 2.9918076395988464} 01/28/2022 01:59:48 - INFO - codeparrot_training - Step 5698: {'lr': 0.0004927131476101195, 'samples': 1094208, 'steps': 5698, 'loss/train': 2.101992130279541} 01/28/2022 01:59:53 - INFO - codeparrot_training - Step 5699: {'lr': 0.0004927092253782704, 'samples': 1094400, 'steps': 5699, 'loss/train': 1.6339511275291443} 01/28/2022 01:59:58 - INFO - codeparrot_training - Step 5700: {'lr': 0.0004927053021067321, 'samples': 1094592, 'steps': 5700, 'loss/train': 1.7292221188545227} 01/28/2022 02:00:02 - INFO - codeparrot_training - Step 5701: {'lr': 0.0004927013777955212, 'samples': 1094784, 'steps': 5701, 'loss/train': 1.4606863260269165} 01/28/2022 02:00:06 - INFO - codeparrot_training - Step 5702: {'lr': 0.0004926974524446548, 'samples': 1094976, 'steps': 5702, 'loss/train': 3.003050208091736} 01/28/2022 02:00:10 - INFO - codeparrot_training - Step 5703: {'lr': 0.0004926935260541496, 'samples': 1095168, 'steps': 5703, 'loss/train': 1.976762294769287} 01/28/2022 02:00:15 - INFO - codeparrot_training - Step 5704: {'lr': 0.0004926895986240222, 'samples': 1095360, 'steps': 5704, 'loss/train': 1.673352599143982} 01/28/2022 02:00:20 - INFO - codeparrot_training - Step 5705: {'lr': 0.0004926856701542898, 'samples': 1095552, 'steps': 5705, 'loss/train': 2.6678824424743652} 01/28/2022 02:00:25 - INFO - codeparrot_training - Step 5706: {'lr': 0.000492681740644969, 'samples': 1095744, 'steps': 5706, 'loss/train': 2.6826714277267456} 01/28/2022 02:00:29 - INFO - codeparrot_training - Step 5707: {'lr': 0.0004926778100960767, 'samples': 1095936, 'steps': 5707, 'loss/train': 2.167461633682251} 01/28/2022 02:00:33 - INFO - codeparrot_training - Step 5708: {'lr': 0.0004926738785076297, 'samples': 1096128, 'steps': 5708, 'loss/train': 3.8154162168502808} 01/28/2022 02:00:37 - INFO - codeparrot_training - Step 5709: {'lr': 0.0004926699458796448, 'samples': 1096320, 'steps': 5709, 'loss/train': 1.5823071599006653} 01/28/2022 02:00:42 - INFO - codeparrot_training - Step 5710: {'lr': 0.0004926660122121391, 'samples': 1096512, 'steps': 5710, 'loss/train': 1.10056534409523} 01/28/2022 02:00:48 - INFO - codeparrot_training - Step 5711: {'lr': 0.0004926620775051291, 'samples': 1096704, 'steps': 5711, 'loss/train': 6.877254009246826} 01/28/2022 02:00:52 - INFO - codeparrot_training - Step 5712: {'lr': 0.0004926581417586318, 'samples': 1096896, 'steps': 5712, 'loss/train': 5.043222784996033} 01/28/2022 02:00:57 - INFO - codeparrot_training - Step 5713: {'lr': 0.0004926542049726642, 'samples': 1097088, 'steps': 5713, 'loss/train': 2.8719661831855774} 01/28/2022 02:01:01 - INFO - codeparrot_training - Step 5714: {'lr': 0.0004926502671472429, 'samples': 1097280, 'steps': 5714, 'loss/train': 1.3855798244476318} 01/28/2022 02:01:05 - INFO - codeparrot_training - Step 5715: {'lr': 0.000492646328282385, 'samples': 1097472, 'steps': 5715, 'loss/train': 2.7631521821022034} 01/28/2022 02:01:09 - INFO - codeparrot_training - Step 5716: {'lr': 0.0004926423883781073, 'samples': 1097664, 'steps': 5716, 'loss/train': 2.145338237285614} 01/28/2022 02:01:13 - INFO - codeparrot_training - Step 5717: {'lr': 0.0004926384474344265, 'samples': 1097856, 'steps': 5717, 'loss/train': 3.127984642982483} 01/28/2022 02:01:19 - INFO - codeparrot_training - Step 5718: {'lr': 0.0004926345054513598, 'samples': 1098048, 'steps': 5718, 'loss/train': 3.5002716779708862} 01/28/2022 02:01:23 - INFO - codeparrot_training - Step 5719: {'lr': 0.0004926305624289238, 'samples': 1098240, 'steps': 5719, 'loss/train': 2.5695735812187195} 01/28/2022 02:01:27 - INFO - codeparrot_training - Step 5720: {'lr': 0.0004926266183671356, 'samples': 1098432, 'steps': 5720, 'loss/train': 2.931149125099182} 01/28/2022 02:01:32 - INFO - codeparrot_training - Step 5721: {'lr': 0.000492622673266012, 'samples': 1098624, 'steps': 5721, 'loss/train': 2.422839939594269} 01/28/2022 02:01:37 - INFO - codeparrot_training - Step 5722: {'lr': 0.0004926187271255698, 'samples': 1098816, 'steps': 5722, 'loss/train': 3.392049193382263} 01/28/2022 02:01:41 - INFO - codeparrot_training - Step 5723: {'lr': 0.0004926147799458262, 'samples': 1099008, 'steps': 5723, 'loss/train': 3.16443407535553} 01/28/2022 02:01:45 - INFO - codeparrot_training - Step 5724: {'lr': 0.0004926108317267979, 'samples': 1099200, 'steps': 5724, 'loss/train': 1.6341918110847473} 01/28/2022 02:01:50 - INFO - codeparrot_training - Step 5725: {'lr': 0.0004926068824685017, 'samples': 1099392, 'steps': 5725, 'loss/train': 0.9492399394512177} 01/28/2022 02:01:54 - INFO - codeparrot_training - Step 5726: {'lr': 0.0004926029321709548, 'samples': 1099584, 'steps': 5726, 'loss/train': 3.7622534036636353} 01/28/2022 02:01:58 - INFO - codeparrot_training - Step 5727: {'lr': 0.0004925989808341738, 'samples': 1099776, 'steps': 5727, 'loss/train': 2.4790379405021667} 01/28/2022 02:02:04 - INFO - codeparrot_training - Step 5728: {'lr': 0.0004925950284581759, 'samples': 1099968, 'steps': 5728, 'loss/train': 2.821582317352295} 01/28/2022 02:02:08 - INFO - codeparrot_training - Step 5729: {'lr': 0.0004925910750429779, 'samples': 1100160, 'steps': 5729, 'loss/train': 3.602182388305664} 01/28/2022 02:02:12 - INFO - codeparrot_training - Step 5730: {'lr': 0.0004925871205885968, 'samples': 1100352, 'steps': 5730, 'loss/train': 2.2229182720184326} 01/28/2022 02:02:16 - INFO - codeparrot_training - Step 5731: {'lr': 0.0004925831650950495, 'samples': 1100544, 'steps': 5731, 'loss/train': 2.406689465045929} 01/28/2022 02:02:21 - INFO - codeparrot_training - Step 5732: {'lr': 0.000492579208562353, 'samples': 1100736, 'steps': 5732, 'loss/train': 1.7714738845825195} 01/28/2022 02:02:26 - INFO - codeparrot_training - Step 5733: {'lr': 0.0004925752509905241, 'samples': 1100928, 'steps': 5733, 'loss/train': 3.233237385749817} 01/28/2022 02:02:31 - INFO - codeparrot_training - Step 5734: {'lr': 0.0004925712923795799, 'samples': 1101120, 'steps': 5734, 'loss/train': 1.9109430313110352} 01/28/2022 02:02:35 - INFO - codeparrot_training - Step 5735: {'lr': 0.0004925673327295374, 'samples': 1101312, 'steps': 5735, 'loss/train': 2.6303991079330444} 01/28/2022 02:02:39 - INFO - codeparrot_training - Step 5736: {'lr': 0.0004925633720404132, 'samples': 1101504, 'steps': 5736, 'loss/train': 2.931621551513672} 01/28/2022 02:02:43 - INFO - codeparrot_training - Step 5737: {'lr': 0.0004925594103122248, 'samples': 1101696, 'steps': 5737, 'loss/train': 2.762357175350189} 01/28/2022 02:02:49 - INFO - codeparrot_training - Step 5738: {'lr': 0.0004925554475449888, 'samples': 1101888, 'steps': 5738, 'loss/train': 3.3677737712860107} 01/28/2022 02:02:53 - INFO - codeparrot_training - Step 5739: {'lr': 0.0004925514837387223, 'samples': 1102080, 'steps': 5739, 'loss/train': 2.890948534011841} 01/28/2022 02:02:57 - INFO - codeparrot_training - Step 5740: {'lr': 0.0004925475188934423, 'samples': 1102272, 'steps': 5740, 'loss/train': 2.4408481121063232} 01/28/2022 02:03:02 - INFO - codeparrot_training - Step 5741: {'lr': 0.0004925435530091656, 'samples': 1102464, 'steps': 5741, 'loss/train': 1.734926462173462} 01/28/2022 02:03:08 - INFO - codeparrot_training - Step 5742: {'lr': 0.0004925395860859096, 'samples': 1102656, 'steps': 5742, 'loss/train': 3.591549038887024} 01/28/2022 02:03:12 - INFO - codeparrot_training - Step 5743: {'lr': 0.0004925356181236908, 'samples': 1102848, 'steps': 5743, 'loss/train': 2.5554394125938416} 01/28/2022 02:03:16 - INFO - codeparrot_training - Step 5744: {'lr': 0.0004925316491225265, 'samples': 1103040, 'steps': 5744, 'loss/train': 2.9363536834716797} 01/28/2022 02:03:20 - INFO - codeparrot_training - Step 5745: {'lr': 0.0004925276790824336, 'samples': 1103232, 'steps': 5745, 'loss/train': 2.7330275774002075} 01/28/2022 02:03:25 - INFO - codeparrot_training - Step 5746: {'lr': 0.0004925237080034291, 'samples': 1103424, 'steps': 5746, 'loss/train': 2.530963182449341} 01/28/2022 02:03:29 - INFO - codeparrot_training - Step 5747: {'lr': 0.0004925197358855301, 'samples': 1103616, 'steps': 5747, 'loss/train': 2.7153952717781067} 01/28/2022 02:03:34 - INFO - codeparrot_training - Step 5748: {'lr': 0.0004925157627287536, 'samples': 1103808, 'steps': 5748, 'loss/train': 1.9603761434555054} 01/28/2022 02:03:38 - INFO - codeparrot_training - Step 5749: {'lr': 0.0004925117885331166, 'samples': 1104000, 'steps': 5749, 'loss/train': 3.310742497444153} 01/28/2022 02:03:43 - INFO - codeparrot_training - Step 5750: {'lr': 0.000492507813298636, 'samples': 1104192, 'steps': 5750, 'loss/train': 3.965185046195984} 01/28/2022 02:03:47 - INFO - codeparrot_training - Step 5751: {'lr': 0.000492503837025329, 'samples': 1104384, 'steps': 5751, 'loss/train': 3.6298463344573975} 01/28/2022 02:03:52 - INFO - codeparrot_training - Step 5752: {'lr': 0.0004924998597132125, 'samples': 1104576, 'steps': 5752, 'loss/train': 2.7767863869667053} 01/28/2022 02:03:56 - INFO - codeparrot_training - Step 5753: {'lr': 0.0004924958813623037, 'samples': 1104768, 'steps': 5753, 'loss/train': 3.276807188987732} 01/28/2022 02:04:01 - INFO - codeparrot_training - Step 5754: {'lr': 0.0004924919019726195, 'samples': 1104960, 'steps': 5754, 'loss/train': 2.3904420733451843} 01/28/2022 02:04:05 - INFO - codeparrot_training - Step 5755: {'lr': 0.000492487921544177, 'samples': 1105152, 'steps': 5755, 'loss/train': 4.0153809785842896} 01/28/2022 02:04:09 - INFO - codeparrot_training - Step 5756: {'lr': 0.0004924839400769932, 'samples': 1105344, 'steps': 5756, 'loss/train': 2.599227547645569} 01/28/2022 02:04:15 - INFO - codeparrot_training - Step 5757: {'lr': 0.0004924799575710852, 'samples': 1105536, 'steps': 5757, 'loss/train': 0.658332422375679} 01/28/2022 02:04:19 - INFO - codeparrot_training - Step 5758: {'lr': 0.0004924759740264701, 'samples': 1105728, 'steps': 5758, 'loss/train': 2.3217249512672424} 01/28/2022 02:04:23 - INFO - codeparrot_training - Step 5759: {'lr': 0.000492471989443165, 'samples': 1105920, 'steps': 5759, 'loss/train': 2.857750654220581} 01/28/2022 02:04:27 - INFO - codeparrot_training - Step 5760: {'lr': 0.0004924680038211868, 'samples': 1106112, 'steps': 5760, 'loss/train': 2.3163841366767883} 01/28/2022 02:04:32 - INFO - codeparrot_training - Step 5761: {'lr': 0.0004924640171605526, 'samples': 1106304, 'steps': 5761, 'loss/train': 3.4260538816452026} 01/28/2022 02:04:37 - INFO - codeparrot_training - Step 5762: {'lr': 0.0004924600294612796, 'samples': 1106496, 'steps': 5762, 'loss/train': 3.759766101837158} 01/28/2022 02:04:41 - INFO - codeparrot_training - Step 5763: {'lr': 0.0004924560407233848, 'samples': 1106688, 'steps': 5763, 'loss/train': 3.4588040113449097} 01/28/2022 02:04:45 - INFO - codeparrot_training - Step 5764: {'lr': 0.0004924520509468854, 'samples': 1106880, 'steps': 5764, 'loss/train': 1.6705368161201477} 01/28/2022 02:04:50 - INFO - codeparrot_training - Step 5765: {'lr': 0.0004924480601317982, 'samples': 1107072, 'steps': 5765, 'loss/train': 4.580164074897766} 01/28/2022 02:04:54 - INFO - codeparrot_training - Step 5766: {'lr': 0.0004924440682781407, 'samples': 1107264, 'steps': 5766, 'loss/train': 2.8328974843025208} 01/28/2022 02:05:00 - INFO - codeparrot_training - Step 5767: {'lr': 0.0004924400753859297, 'samples': 1107456, 'steps': 5767, 'loss/train': 4.392118334770203} 01/28/2022 02:05:04 - INFO - codeparrot_training - Step 5768: {'lr': 0.0004924360814551825, 'samples': 1107648, 'steps': 5768, 'loss/train': 1.9112082123756409} 01/28/2022 02:05:08 - INFO - codeparrot_training - Step 5769: {'lr': 0.000492432086485916, 'samples': 1107840, 'steps': 5769, 'loss/train': 4.024968981742859} 01/28/2022 02:05:12 - INFO - codeparrot_training - Step 5770: {'lr': 0.0004924280904781475, 'samples': 1108032, 'steps': 5770, 'loss/train': 2.7605552673339844} 01/28/2022 02:05:17 - INFO - codeparrot_training - Step 5771: {'lr': 0.0004924240934318939, 'samples': 1108224, 'steps': 5771, 'loss/train': 2.121771275997162} 01/28/2022 02:05:22 - INFO - codeparrot_training - Step 5772: {'lr': 0.0004924200953471727, 'samples': 1108416, 'steps': 5772, 'loss/train': 1.9272823333740234} 01/28/2022 02:05:26 - INFO - codeparrot_training - Step 5773: {'lr': 0.0004924160962240005, 'samples': 1108608, 'steps': 5773, 'loss/train': 1.0827971398830414} 01/28/2022 02:05:30 - INFO - codeparrot_training - Step 5774: {'lr': 0.0004924120960623949, 'samples': 1108800, 'steps': 5774, 'loss/train': 2.5055429935455322} 01/28/2022 02:05:35 - INFO - codeparrot_training - Step 5775: {'lr': 0.0004924080948623729, 'samples': 1108992, 'steps': 5775, 'loss/train': 3.973841071128845} 01/28/2022 02:05:39 - INFO - codeparrot_training - Step 5776: {'lr': 0.0004924040926239515, 'samples': 1109184, 'steps': 5776, 'loss/train': 3.891141414642334} 01/28/2022 02:05:44 - INFO - codeparrot_training - Step 5777: {'lr': 0.000492400089347148, 'samples': 1109376, 'steps': 5777, 'loss/train': 2.822625160217285} 01/28/2022 02:05:48 - INFO - codeparrot_training - Step 5778: {'lr': 0.0004923960850319794, 'samples': 1109568, 'steps': 5778, 'loss/train': 2.4638818502426147} 01/28/2022 02:05:53 - INFO - codeparrot_training - Step 5779: {'lr': 0.000492392079678463, 'samples': 1109760, 'steps': 5779, 'loss/train': 3.01317036151886} 01/28/2022 02:05:57 - INFO - codeparrot_training - Step 5780: {'lr': 0.0004923880732866159, 'samples': 1109952, 'steps': 5780, 'loss/train': 2.533119320869446} 01/28/2022 02:06:01 - INFO - codeparrot_training - Step 5781: {'lr': 0.0004923840658564553, 'samples': 1110144, 'steps': 5781, 'loss/train': 3.0891627073287964} 01/28/2022 02:06:05 - INFO - codeparrot_training - Step 5782: {'lr': 0.0004923800573879983, 'samples': 1110336, 'steps': 5782, 'loss/train': 3.318666100502014} 01/28/2022 02:06:11 - INFO - codeparrot_training - Step 5783: {'lr': 0.000492376047881262, 'samples': 1110528, 'steps': 5783, 'loss/train': 2.06441330909729} 01/28/2022 02:06:15 - INFO - codeparrot_training - Step 5784: {'lr': 0.0004923720373362638, 'samples': 1110720, 'steps': 5784, 'loss/train': 3.2557700872421265} 01/28/2022 02:06:19 - INFO - codeparrot_training - Step 5785: {'lr': 0.0004923680257530207, 'samples': 1110912, 'steps': 5785, 'loss/train': 2.66511994600296} 01/28/2022 02:06:23 - INFO - codeparrot_training - Step 5786: {'lr': 0.0004923640131315499, 'samples': 1111104, 'steps': 5786, 'loss/train': 3.788439989089966} 01/28/2022 02:06:29 - INFO - codeparrot_training - Step 5787: {'lr': 0.0004923599994718687, 'samples': 1111296, 'steps': 5787, 'loss/train': 3.9055556058883667} 01/28/2022 02:06:34 - INFO - codeparrot_training - Step 5788: {'lr': 0.0004923559847739941, 'samples': 1111488, 'steps': 5788, 'loss/train': 3.5730847120285034} 01/28/2022 02:06:38 - INFO - codeparrot_training - Step 5789: {'lr': 0.0004923519690379436, 'samples': 1111680, 'steps': 5789, 'loss/train': 3.1797409057617188} 01/28/2022 02:06:42 - INFO - codeparrot_training - Step 5790: {'lr': 0.0004923479522637341, 'samples': 1111872, 'steps': 5790, 'loss/train': 3.6151957511901855} 01/28/2022 02:06:46 - INFO - codeparrot_training - Step 5791: {'lr': 0.0004923439344513829, 'samples': 1112064, 'steps': 5791, 'loss/train': 1.904365360736847} 01/28/2022 02:06:51 - INFO - codeparrot_training - Step 5792: {'lr': 0.0004923399156009073, 'samples': 1112256, 'steps': 5792, 'loss/train': 3.0646458864212036} 01/28/2022 02:06:56 - INFO - codeparrot_training - Step 5793: {'lr': 0.0004923358957123245, 'samples': 1112448, 'steps': 5793, 'loss/train': 3.1497738361358643} 01/28/2022 02:07:00 - INFO - codeparrot_training - Step 5794: {'lr': 0.0004923318747856515, 'samples': 1112640, 'steps': 5794, 'loss/train': 3.248157262802124} 01/28/2022 02:07:04 - INFO - codeparrot_training - Step 5795: {'lr': 0.0004923278528209059, 'samples': 1112832, 'steps': 5795, 'loss/train': 2.4890264868736267} 01/28/2022 02:07:08 - INFO - codeparrot_training - Step 5796: {'lr': 0.0004923238298181047, 'samples': 1113024, 'steps': 5796, 'loss/train': 2.948649287223816} 01/28/2022 02:07:14 - INFO - codeparrot_training - Step 5797: {'lr': 0.0004923198057772651, 'samples': 1113216, 'steps': 5797, 'loss/train': 1.854840338230133} 01/28/2022 02:07:18 - INFO - codeparrot_training - Step 5798: {'lr': 0.0004923157806984044, 'samples': 1113408, 'steps': 5798, 'loss/train': 3.891411781311035} 01/28/2022 02:07:22 - INFO - codeparrot_training - Step 5799: {'lr': 0.0004923117545815398, 'samples': 1113600, 'steps': 5799, 'loss/train': 3.8032665252685547} 01/28/2022 02:07:26 - INFO - codeparrot_training - Step 5800: {'lr': 0.0004923077274266886, 'samples': 1113792, 'steps': 5800, 'loss/train': 1.613293468952179} 01/28/2022 02:07:31 - INFO - codeparrot_training - Step 5801: {'lr': 0.0004923036992338681, 'samples': 1113984, 'steps': 5801, 'loss/train': 2.7431706190109253} 01/28/2022 02:07:36 - INFO - codeparrot_training - Step 5802: {'lr': 0.0004922996700030954, 'samples': 1114176, 'steps': 5802, 'loss/train': 1.8835712671279907} 01/28/2022 02:07:41 - INFO - codeparrot_training - Step 5803: {'lr': 0.000492295639734388, 'samples': 1114368, 'steps': 5803, 'loss/train': 2.0053560733795166} 01/28/2022 02:07:45 - INFO - codeparrot_training - Step 5804: {'lr': 0.0004922916084277629, 'samples': 1114560, 'steps': 5804, 'loss/train': 2.915123999118805} 01/28/2022 02:07:49 - INFO - codeparrot_training - Step 5805: {'lr': 0.0004922875760832375, 'samples': 1114752, 'steps': 5805, 'loss/train': 2.9417468905448914} 01/28/2022 02:07:53 - INFO - codeparrot_training - Step 5806: {'lr': 0.000492283542700829, 'samples': 1114944, 'steps': 5806, 'loss/train': 3.5973644256591797} 01/28/2022 02:07:58 - INFO - codeparrot_training - Step 5807: {'lr': 0.0004922795082805549, 'samples': 1115136, 'steps': 5807, 'loss/train': 3.673802375793457} 01/28/2022 02:08:03 - INFO - codeparrot_training - Step 5808: {'lr': 0.0004922754728224322, 'samples': 1115328, 'steps': 5808, 'loss/train': 3.8085436820983887} 01/28/2022 02:08:07 - INFO - codeparrot_training - Step 5809: {'lr': 0.0004922714363264783, 'samples': 1115520, 'steps': 5809, 'loss/train': 2.164436101913452} 01/28/2022 02:08:11 - INFO - codeparrot_training - Step 5810: {'lr': 0.0004922673987927106, 'samples': 1115712, 'steps': 5810, 'loss/train': 1.0220831036567688} 01/28/2022 02:08:15 - INFO - codeparrot_training - Step 5811: {'lr': 0.0004922633602211462, 'samples': 1115904, 'steps': 5811, 'loss/train': 3.2357722520828247} 01/28/2022 02:08:21 - INFO - codeparrot_training - Step 5812: {'lr': 0.0004922593206118025, 'samples': 1116096, 'steps': 5812, 'loss/train': 3.0347378253936768} 01/28/2022 02:08:26 - INFO - codeparrot_training - Step 5813: {'lr': 0.0004922552799646968, 'samples': 1116288, 'steps': 5813, 'loss/train': 2.849492132663727} 01/28/2022 02:08:30 - INFO - codeparrot_training - Step 5814: {'lr': 0.0004922512382798463, 'samples': 1116480, 'steps': 5814, 'loss/train': 2.4001643657684326} 01/28/2022 02:08:34 - INFO - codeparrot_training - Step 5815: {'lr': 0.0004922471955572686, 'samples': 1116672, 'steps': 5815, 'loss/train': 5.304857969284058} 01/28/2022 02:08:38 - INFO - codeparrot_training - Step 5816: {'lr': 0.0004922431517969808, 'samples': 1116864, 'steps': 5816, 'loss/train': 3.0184046030044556} 01/28/2022 02:08:43 - INFO - codeparrot_training - Step 5817: {'lr': 0.0004922391069990002, 'samples': 1117056, 'steps': 5817, 'loss/train': 2.5229862928390503} 01/28/2022 02:08:48 - INFO - codeparrot_training - Step 5818: {'lr': 0.0004922350611633442, 'samples': 1117248, 'steps': 5818, 'loss/train': 2.376480460166931} 01/28/2022 02:08:52 - INFO - codeparrot_training - Step 5819: {'lr': 0.0004922310142900302, 'samples': 1117440, 'steps': 5819, 'loss/train': 0.6038575172424316} 01/28/2022 02:08:56 - INFO - codeparrot_training - Step 5820: {'lr': 0.0004922269663790753, 'samples': 1117632, 'steps': 5820, 'loss/train': 2.884299159049988} 01/28/2022 02:09:00 - INFO - codeparrot_training - Step 5821: {'lr': 0.0004922229174304971, 'samples': 1117824, 'steps': 5821, 'loss/train': 2.2521273493766785} 01/28/2022 02:09:06 - INFO - codeparrot_training - Step 5822: {'lr': 0.0004922188674443128, 'samples': 1118016, 'steps': 5822, 'loss/train': 1.6458730101585388} 01/28/2022 02:09:10 - INFO - codeparrot_training - Step 5823: {'lr': 0.0004922148164205398, 'samples': 1118208, 'steps': 5823, 'loss/train': 2.74300217628479} 01/28/2022 02:09:14 - INFO - codeparrot_training - Step 5824: {'lr': 0.0004922107643591954, 'samples': 1118400, 'steps': 5824, 'loss/train': 3.0984292030334473} 01/28/2022 02:09:19 - INFO - codeparrot_training - Step 5825: {'lr': 0.000492206711260297, 'samples': 1118592, 'steps': 5825, 'loss/train': 1.6867334246635437} 01/28/2022 02:09:23 - INFO - codeparrot_training - Step 5826: {'lr': 0.000492202657123862, 'samples': 1118784, 'steps': 5826, 'loss/train': 3.1231441497802734} 01/28/2022 02:09:29 - INFO - codeparrot_training - Step 5827: {'lr': 0.0004921986019499078, 'samples': 1118976, 'steps': 5827, 'loss/train': 2.912612557411194} 01/28/2022 02:09:33 - INFO - codeparrot_training - Step 5828: {'lr': 0.0004921945457384516, 'samples': 1119168, 'steps': 5828, 'loss/train': 3.5519131422042847} 01/28/2022 02:09:37 - INFO - codeparrot_training - Step 5829: {'lr': 0.0004921904884895108, 'samples': 1119360, 'steps': 5829, 'loss/train': 2.779664933681488} 01/28/2022 02:09:42 - INFO - codeparrot_training - Step 5830: {'lr': 0.000492186430203103, 'samples': 1119552, 'steps': 5830, 'loss/train': 3.0815012454986572} 01/28/2022 02:09:46 - INFO - codeparrot_training - Step 5831: {'lr': 0.0004921823708792453, 'samples': 1119744, 'steps': 5831, 'loss/train': 3.1538186073303223} 01/28/2022 02:09:51 - INFO - codeparrot_training - Step 5832: {'lr': 0.0004921783105179552, 'samples': 1119936, 'steps': 5832, 'loss/train': 1.2093158662319183} 01/28/2022 02:09:55 - INFO - codeparrot_training - Step 5833: {'lr': 0.0004921742491192502, 'samples': 1120128, 'steps': 5833, 'loss/train': 4.215378284454346} 01/28/2022 02:09:59 - INFO - codeparrot_training - Step 5834: {'lr': 0.0004921701866831477, 'samples': 1120320, 'steps': 5834, 'loss/train': 2.7640364170074463} 01/28/2022 02:10:04 - INFO - codeparrot_training - Step 5835: {'lr': 0.000492166123209665, 'samples': 1120512, 'steps': 5835, 'loss/train': 2.0828617215156555} 01/28/2022 02:10:08 - INFO - codeparrot_training - Step 5836: {'lr': 0.0004921620586988193, 'samples': 1120704, 'steps': 5836, 'loss/train': 3.1233190298080444} 01/28/2022 02:10:13 - INFO - codeparrot_training - Step 5837: {'lr': 0.0004921579931506285, 'samples': 1120896, 'steps': 5837, 'loss/train': 3.305760383605957} 01/28/2022 02:10:17 - INFO - codeparrot_training - Step 5838: {'lr': 0.0004921539265651096, 'samples': 1121088, 'steps': 5838, 'loss/train': 5.209257245063782} 01/28/2022 02:10:21 - INFO - codeparrot_training - Step 5839: {'lr': 0.0004921498589422803, 'samples': 1121280, 'steps': 5839, 'loss/train': 2.333466410636902} 01/28/2022 02:10:26 - INFO - codeparrot_training - Step 5840: {'lr': 0.0004921457902821578, 'samples': 1121472, 'steps': 5840, 'loss/train': 1.8971039056777954} 01/28/2022 02:10:30 - INFO - codeparrot_training - Step 5841: {'lr': 0.0004921417205847597, 'samples': 1121664, 'steps': 5841, 'loss/train': 2.5919541120529175} 01/28/2022 02:10:35 - INFO - codeparrot_training - Step 5842: {'lr': 0.0004921376498501032, 'samples': 1121856, 'steps': 5842, 'loss/train': 2.349085807800293} 01/28/2022 02:10:39 - INFO - codeparrot_training - Step 5843: {'lr': 0.000492133578078206, 'samples': 1122048, 'steps': 5843, 'loss/train': 1.7151278257369995} 01/28/2022 02:10:44 - INFO - codeparrot_training - Step 5844: {'lr': 0.0004921295052690855, 'samples': 1122240, 'steps': 5844, 'loss/train': 2.900335371494293} 01/28/2022 02:10:48 - INFO - codeparrot_training - Step 5845: {'lr': 0.000492125431422759, 'samples': 1122432, 'steps': 5845, 'loss/train': 2.3935166001319885} 01/28/2022 02:10:52 - INFO - codeparrot_training - Step 5846: {'lr': 0.0004921213565392441, 'samples': 1122624, 'steps': 5846, 'loss/train': 3.0676428079605103} 01/28/2022 02:10:58 - INFO - codeparrot_training - Step 5847: {'lr': 0.000492117280618558, 'samples': 1122816, 'steps': 5847, 'loss/train': 5.847450256347656} 01/28/2022 02:11:02 - INFO - codeparrot_training - Step 5848: {'lr': 0.0004921132036607186, 'samples': 1123008, 'steps': 5848, 'loss/train': 2.067874252796173} 01/28/2022 02:11:07 - INFO - codeparrot_training - Step 5849: {'lr': 0.0004921091256657429, 'samples': 1123200, 'steps': 5849, 'loss/train': 3.20949125289917} 01/28/2022 02:11:11 - INFO - codeparrot_training - Step 5850: {'lr': 0.0004921050466336487, 'samples': 1123392, 'steps': 5850, 'loss/train': 1.8135294914245605} 01/28/2022 02:11:15 - INFO - codeparrot_training - Step 5851: {'lr': 0.0004921009665644535, 'samples': 1123584, 'steps': 5851, 'loss/train': 1.6953874826431274} 01/28/2022 02:11:20 - INFO - codeparrot_training - Step 5852: {'lr': 0.0004920968854581745, 'samples': 1123776, 'steps': 5852, 'loss/train': 3.24246883392334} 01/28/2022 02:11:25 - INFO - codeparrot_training - Step 5853: {'lr': 0.0004920928033148292, 'samples': 1123968, 'steps': 5853, 'loss/train': 2.8555902242660522} 01/28/2022 02:11:29 - INFO - codeparrot_training - Step 5854: {'lr': 0.0004920887201344353, 'samples': 1124160, 'steps': 5854, 'loss/train': 2.784697651863098} 01/28/2022 02:11:33 - INFO - codeparrot_training - Step 5855: {'lr': 0.0004920846359170103, 'samples': 1124352, 'steps': 5855, 'loss/train': 3.2407450675964355} 01/28/2022 02:11:37 - INFO - codeparrot_training - Step 5856: {'lr': 0.0004920805506625714, 'samples': 1124544, 'steps': 5856, 'loss/train': 2.108177125453949} 01/28/2022 02:11:42 - INFO - codeparrot_training - Step 5857: {'lr': 0.0004920764643711364, 'samples': 1124736, 'steps': 5857, 'loss/train': 2.3930132389068604} 01/28/2022 02:11:47 - INFO - codeparrot_training - Step 5858: {'lr': 0.0004920723770427226, 'samples': 1124928, 'steps': 5858, 'loss/train': 2.963106393814087} 01/28/2022 02:11:51 - INFO - codeparrot_training - Step 5859: {'lr': 0.0004920682886773478, 'samples': 1125120, 'steps': 5859, 'loss/train': 2.6552878618240356} 01/28/2022 02:11:55 - INFO - codeparrot_training - Step 5860: {'lr': 0.000492064199275029, 'samples': 1125312, 'steps': 5860, 'loss/train': 3.4093533754348755} 01/28/2022 02:11:59 - INFO - codeparrot_training - Step 5861: {'lr': 0.0004920601088357844, 'samples': 1125504, 'steps': 5861, 'loss/train': 2.4352777004241943} 01/28/2022 02:12:05 - INFO - codeparrot_training - Step 5862: {'lr': 0.0004920560173596309, 'samples': 1125696, 'steps': 5862, 'loss/train': 3.7314802408218384} 01/28/2022 02:12:09 - INFO - codeparrot_training - Step 5863: {'lr': 0.0004920519248465864, 'samples': 1125888, 'steps': 5863, 'loss/train': 2.542195737361908} 01/28/2022 02:12:14 - INFO - codeparrot_training - Step 5864: {'lr': 0.0004920478312966683, 'samples': 1126080, 'steps': 5864, 'loss/train': 2.7279235124588013} 01/28/2022 02:12:18 - INFO - codeparrot_training - Step 5865: {'lr': 0.0004920437367098941, 'samples': 1126272, 'steps': 5865, 'loss/train': 2.2285419702529907} 01/28/2022 02:12:22 - INFO - codeparrot_training - Step 5866: {'lr': 0.0004920396410862815, 'samples': 1126464, 'steps': 5866, 'loss/train': 1.6092604994773865} 01/28/2022 02:12:27 - INFO - codeparrot_training - Step 5867: {'lr': 0.0004920355444258479, 'samples': 1126656, 'steps': 5867, 'loss/train': 3.302425503730774} 01/28/2022 02:12:32 - INFO - codeparrot_training - Step 5868: {'lr': 0.0004920314467286108, 'samples': 1126848, 'steps': 5868, 'loss/train': 2.397562086582184} 01/28/2022 02:12:36 - INFO - codeparrot_training - Step 5869: {'lr': 0.0004920273479945878, 'samples': 1127040, 'steps': 5869, 'loss/train': 3.214069962501526} 01/28/2022 02:12:40 - INFO - codeparrot_training - Step 5870: {'lr': 0.0004920232482237966, 'samples': 1127232, 'steps': 5870, 'loss/train': 3.868857979774475} 01/28/2022 02:12:44 - INFO - codeparrot_training - Step 5871: {'lr': 0.0004920191474162547, 'samples': 1127424, 'steps': 5871, 'loss/train': 3.2440541982650757} 01/28/2022 02:12:50 - INFO - codeparrot_training - Step 5872: {'lr': 0.0004920150455719795, 'samples': 1127616, 'steps': 5872, 'loss/train': 2.7832778692245483} 01/28/2022 02:12:54 - INFO - codeparrot_training - Step 5873: {'lr': 0.0004920109426909887, 'samples': 1127808, 'steps': 5873, 'loss/train': 2.101888597011566} 01/28/2022 02:12:59 - INFO - codeparrot_training - Step 5874: {'lr': 0.0004920068387733, 'samples': 1128000, 'steps': 5874, 'loss/train': 3.1614664793014526} 01/28/2022 02:13:03 - INFO - codeparrot_training - Step 5875: {'lr': 0.0004920027338189307, 'samples': 1128192, 'steps': 5875, 'loss/train': 2.05672949552536} 01/28/2022 02:13:07 - INFO - codeparrot_training - Step 5876: {'lr': 0.0004919986278278986, 'samples': 1128384, 'steps': 5876, 'loss/train': 2.607162594795227} 01/28/2022 02:13:13 - INFO - codeparrot_training - Step 5877: {'lr': 0.0004919945208002212, 'samples': 1128576, 'steps': 5877, 'loss/train': 10.069172859191895} 01/28/2022 02:13:17 - INFO - codeparrot_training - Step 5878: {'lr': 0.0004919904127359162, 'samples': 1128768, 'steps': 5878, 'loss/train': 0.75301194190979} 01/28/2022 02:13:21 - INFO - codeparrot_training - Step 5879: {'lr': 0.000491986303635001, 'samples': 1128960, 'steps': 5879, 'loss/train': 4.079577326774597} 01/28/2022 02:13:25 - INFO - codeparrot_training - Step 5880: {'lr': 0.0004919821934974933, 'samples': 1129152, 'steps': 5880, 'loss/train': 3.0554122924804688} 01/28/2022 02:13:29 - INFO - codeparrot_training - Step 5881: {'lr': 0.0004919780823234108, 'samples': 1129344, 'steps': 5881, 'loss/train': 2.602610170841217} 01/28/2022 02:13:35 - INFO - codeparrot_training - Step 5882: {'lr': 0.000491973970112771, 'samples': 1129536, 'steps': 5882, 'loss/train': 3.4515713453292847} 01/28/2022 02:13:39 - INFO - codeparrot_training - Step 5883: {'lr': 0.0004919698568655916, 'samples': 1129728, 'steps': 5883, 'loss/train': 2.144208312034607} 01/28/2022 02:13:43 - INFO - codeparrot_training - Step 5884: {'lr': 0.0004919657425818901, 'samples': 1129920, 'steps': 5884, 'loss/train': 3.2021284103393555} 01/28/2022 02:13:47 - INFO - codeparrot_training - Step 5885: {'lr': 0.0004919616272616842, 'samples': 1130112, 'steps': 5885, 'loss/train': 3.294089913368225} 01/28/2022 02:13:51 - INFO - codeparrot_training - Step 5886: {'lr': 0.0004919575109049915, 'samples': 1130304, 'steps': 5886, 'loss/train': 3.666390538215637} 01/28/2022 02:13:57 - INFO - codeparrot_training - Step 5887: {'lr': 0.0004919533935118296, 'samples': 1130496, 'steps': 5887, 'loss/train': 2.0149487257003784} 01/28/2022 02:14:02 - INFO - codeparrot_training - Step 5888: {'lr': 0.0004919492750822163, 'samples': 1130688, 'steps': 5888, 'loss/train': 3.34130859375} 01/28/2022 02:14:06 - INFO - codeparrot_training - Step 5889: {'lr': 0.0004919451556161692, 'samples': 1130880, 'steps': 5889, 'loss/train': 0.6609929502010345} 01/28/2022 02:14:10 - INFO - codeparrot_training - Step 5890: {'lr': 0.0004919410351137058, 'samples': 1131072, 'steps': 5890, 'loss/train': 2.839029908180237} 01/28/2022 02:14:14 - INFO - codeparrot_training - Step 5891: {'lr': 0.0004919369135748438, 'samples': 1131264, 'steps': 5891, 'loss/train': 3.1167715787887573} 01/28/2022 02:14:20 - INFO - codeparrot_training - Step 5892: {'lr': 0.0004919327909996008, 'samples': 1131456, 'steps': 5892, 'loss/train': 3.483919858932495} 01/28/2022 02:14:24 - INFO - codeparrot_training - Step 5893: {'lr': 0.0004919286673879948, 'samples': 1131648, 'steps': 5893, 'loss/train': 3.1574610471725464} 01/28/2022 02:14:29 - INFO - codeparrot_training - Step 5894: {'lr': 0.000491924542740043, 'samples': 1131840, 'steps': 5894, 'loss/train': 2.602450668811798} 01/28/2022 02:14:33 - INFO - codeparrot_training - Step 5895: {'lr': 0.0004919204170557634, 'samples': 1132032, 'steps': 5895, 'loss/train': 1.8437923192977905} 01/28/2022 02:14:37 - INFO - codeparrot_training - Step 5896: {'lr': 0.0004919162903351734, 'samples': 1132224, 'steps': 5896, 'loss/train': 2.4328558444976807} 01/28/2022 02:14:42 - INFO - codeparrot_training - Step 5897: {'lr': 0.000491912162578291, 'samples': 1132416, 'steps': 5897, 'loss/train': 2.017968714237213} 01/28/2022 02:14:46 - INFO - codeparrot_training - Step 5898: {'lr': 0.0004919080337851336, 'samples': 1132608, 'steps': 5898, 'loss/train': 3.6917967796325684} 01/28/2022 02:14:51 - INFO - codeparrot_training - Step 5899: {'lr': 0.000491903903955719, 'samples': 1132800, 'steps': 5899, 'loss/train': 2.8173587322235107} 01/28/2022 02:14:55 - INFO - codeparrot_training - Step 5900: {'lr': 0.0004918997730900649, 'samples': 1132992, 'steps': 5900, 'loss/train': 2.7084109783172607} 01/28/2022 02:14:59 - INFO - codeparrot_training - Step 5901: {'lr': 0.000491895641188189, 'samples': 1133184, 'steps': 5901, 'loss/train': 0.8295877575874329} 01/28/2022 02:15:05 - INFO - codeparrot_training - Step 5902: {'lr': 0.000491891508250109, 'samples': 1133376, 'steps': 5902, 'loss/train': 2.9686238765716553} 01/28/2022 02:15:09 - INFO - codeparrot_training - Step 5903: {'lr': 0.0004918873742758426, 'samples': 1133568, 'steps': 5903, 'loss/train': 2.6738850474357605} 01/28/2022 02:15:13 - INFO - codeparrot_training - Step 5904: {'lr': 0.0004918832392654074, 'samples': 1133760, 'steps': 5904, 'loss/train': 1.7403005361557007} 01/28/2022 02:15:17 - INFO - codeparrot_training - Step 5905: {'lr': 0.0004918791032188214, 'samples': 1133952, 'steps': 5905, 'loss/train': 3.1226695775985718} 01/28/2022 02:15:23 - INFO - codeparrot_training - Step 5906: {'lr': 0.0004918749661361019, 'samples': 1134144, 'steps': 5906, 'loss/train': 1.7388634085655212} 01/28/2022 02:15:27 - INFO - codeparrot_training - Step 5907: {'lr': 0.000491870828017267, 'samples': 1134336, 'steps': 5907, 'loss/train': 2.8519906997680664} 01/28/2022 02:15:32 - INFO - codeparrot_training - Step 5908: {'lr': 0.0004918666888623342, 'samples': 1134528, 'steps': 5908, 'loss/train': 2.975853145122528} 01/28/2022 02:15:36 - INFO - codeparrot_training - Step 5909: {'lr': 0.0004918625486713214, 'samples': 1134720, 'steps': 5909, 'loss/train': 3.937126636505127} 01/28/2022 02:15:40 - INFO - codeparrot_training - Step 5910: {'lr': 0.0004918584074442462, 'samples': 1134912, 'steps': 5910, 'loss/train': 3.5511492490768433} 01/28/2022 02:15:45 - INFO - codeparrot_training - Step 5911: {'lr': 0.0004918542651811263, 'samples': 1135104, 'steps': 5911, 'loss/train': 3.1420462131500244} 01/28/2022 02:15:49 - INFO - codeparrot_training - Step 5912: {'lr': 0.0004918501218819796, 'samples': 1135296, 'steps': 5912, 'loss/train': 1.9137330651283264} 01/28/2022 02:15:54 - INFO - codeparrot_training - Step 5913: {'lr': 0.0004918459775468238, 'samples': 1135488, 'steps': 5913, 'loss/train': 3.344046950340271} 01/28/2022 02:15:58 - INFO - codeparrot_training - Step 5914: {'lr': 0.0004918418321756766, 'samples': 1135680, 'steps': 5914, 'loss/train': 3.5410516262054443} 01/28/2022 02:16:02 - INFO - codeparrot_training - Step 5915: {'lr': 0.0004918376857685557, 'samples': 1135872, 'steps': 5915, 'loss/train': 2.217968165874481} 01/28/2022 02:16:08 - INFO - codeparrot_training - Step 5916: {'lr': 0.000491833538325479, 'samples': 1136064, 'steps': 5916, 'loss/train': 2.915977120399475} 01/28/2022 02:16:12 - INFO - codeparrot_training - Step 5917: {'lr': 0.0004918293898464643, 'samples': 1136256, 'steps': 5917, 'loss/train': 1.3719635009765625} 01/28/2022 02:16:16 - INFO - codeparrot_training - Step 5918: {'lr': 0.0004918252403315292, 'samples': 1136448, 'steps': 5918, 'loss/train': 2.2842437624931335} 01/28/2022 02:16:21 - INFO - codeparrot_training - Step 5919: {'lr': 0.0004918210897806916, 'samples': 1136640, 'steps': 5919, 'loss/train': 0.8830389976501465} 01/28/2022 02:16:25 - INFO - codeparrot_training - Step 5920: {'lr': 0.0004918169381939692, 'samples': 1136832, 'steps': 5920, 'loss/train': 3.4080870151519775} 01/28/2022 02:16:30 - INFO - codeparrot_training - Step 5921: {'lr': 0.0004918127855713799, 'samples': 1137024, 'steps': 5921, 'loss/train': 3.221192479133606} 01/28/2022 02:16:34 - INFO - codeparrot_training - Step 5922: {'lr': 0.0004918086319129413, 'samples': 1137216, 'steps': 5922, 'loss/train': 3.0333516597747803} 01/28/2022 02:16:39 - INFO - codeparrot_training - Step 5923: {'lr': 0.0004918044772186714, 'samples': 1137408, 'steps': 5923, 'loss/train': 2.3230249285697937} 01/28/2022 02:16:43 - INFO - codeparrot_training - Step 5924: {'lr': 0.0004918003214885877, 'samples': 1137600, 'steps': 5924, 'loss/train': 2.9069003462791443} 01/28/2022 02:16:47 - INFO - codeparrot_training - Step 5925: {'lr': 0.0004917961647227084, 'samples': 1137792, 'steps': 5925, 'loss/train': 3.3071987628936768} 01/28/2022 02:16:52 - INFO - codeparrot_training - Step 5926: {'lr': 0.0004917920069210511, 'samples': 1137984, 'steps': 5926, 'loss/train': 1.930332362651825} 01/28/2022 02:16:57 - INFO - codeparrot_training - Step 5927: {'lr': 0.0004917878480836336, 'samples': 1138176, 'steps': 5927, 'loss/train': 3.1911370754241943} 01/28/2022 02:17:01 - INFO - codeparrot_training - Step 5928: {'lr': 0.0004917836882104737, 'samples': 1138368, 'steps': 5928, 'loss/train': 1.5299171805381775} 01/28/2022 02:17:05 - INFO - codeparrot_training - Step 5929: {'lr': 0.0004917795273015892, 'samples': 1138560, 'steps': 5929, 'loss/train': 3.242104411125183} 01/28/2022 02:17:09 - INFO - codeparrot_training - Step 5930: {'lr': 0.0004917753653569981, 'samples': 1138752, 'steps': 5930, 'loss/train': 3.6192766427993774} 01/28/2022 02:17:15 - INFO - codeparrot_training - Step 5931: {'lr': 0.000491771202376718, 'samples': 1138944, 'steps': 5931, 'loss/train': 2.8199002146720886} 01/28/2022 02:17:19 - INFO - codeparrot_training - Step 5932: {'lr': 0.000491767038360767, 'samples': 1139136, 'steps': 5932, 'loss/train': 2.0272344946861267} 01/28/2022 02:17:23 - INFO - codeparrot_training - Step 5933: {'lr': 0.0004917628733091626, 'samples': 1139328, 'steps': 5933, 'loss/train': 2.3437764644622803} 01/28/2022 02:17:28 - INFO - codeparrot_training - Step 5934: {'lr': 0.000491758707221923, 'samples': 1139520, 'steps': 5934, 'loss/train': 3.650021553039551} 01/28/2022 02:17:32 - INFO - codeparrot_training - Step 5935: {'lr': 0.0004917545400990657, 'samples': 1139712, 'steps': 5935, 'loss/train': 2.122618317604065} 01/28/2022 02:17:37 - INFO - codeparrot_training - Step 5936: {'lr': 0.0004917503719406087, 'samples': 1139904, 'steps': 5936, 'loss/train': 3.7973885536193848} 01/28/2022 02:17:41 - INFO - codeparrot_training - Step 5937: {'lr': 0.00049174620274657, 'samples': 1140096, 'steps': 5937, 'loss/train': 2.4597871899604797} 01/28/2022 02:17:46 - INFO - codeparrot_training - Step 5938: {'lr': 0.0004917420325169673, 'samples': 1140288, 'steps': 5938, 'loss/train': 2.881145417690277} 01/28/2022 02:17:50 - INFO - codeparrot_training - Step 5939: {'lr': 0.0004917378612518185, 'samples': 1140480, 'steps': 5939, 'loss/train': 3.276169538497925} 01/28/2022 02:17:54 - INFO - codeparrot_training - Step 5940: {'lr': 0.0004917336889511414, 'samples': 1140672, 'steps': 5940, 'loss/train': 2.9432217478752136} 01/28/2022 02:17:59 - INFO - codeparrot_training - Step 5941: {'lr': 0.0004917295156149539, 'samples': 1140864, 'steps': 5941, 'loss/train': 1.244056910276413} 01/28/2022 02:18:03 - INFO - codeparrot_training - Step 5942: {'lr': 0.000491725341243274, 'samples': 1141056, 'steps': 5942, 'loss/train': 2.4044553637504578} 01/28/2022 02:18:08 - INFO - codeparrot_training - Step 5943: {'lr': 0.0004917211658361196, 'samples': 1141248, 'steps': 5943, 'loss/train': 3.5562790632247925} 01/28/2022 02:18:12 - INFO - codeparrot_training - Step 5944: {'lr': 0.0004917169893935083, 'samples': 1141440, 'steps': 5944, 'loss/train': 2.8870823979377747} 01/28/2022 02:18:16 - INFO - codeparrot_training - Step 5945: {'lr': 0.0004917128119154582, 'samples': 1141632, 'steps': 5945, 'loss/train': 2.873852491378784} 01/28/2022 02:18:24 - INFO - codeparrot_training - Step 5946: {'lr': 0.0004917086334019872, 'samples': 1141824, 'steps': 5946, 'loss/train': 1.274657517671585} 01/28/2022 02:18:28 - INFO - codeparrot_training - Step 5947: {'lr': 0.0004917044538531131, 'samples': 1142016, 'steps': 5947, 'loss/train': 1.4572659730911255} 01/28/2022 02:18:32 - INFO - codeparrot_training - Step 5948: {'lr': 0.000491700273268854, 'samples': 1142208, 'steps': 5948, 'loss/train': 3.432344913482666} 01/28/2022 02:18:36 - INFO - codeparrot_training - Step 5949: {'lr': 0.0004916960916492276, 'samples': 1142400, 'steps': 5949, 'loss/train': 3.249409317970276} 01/28/2022 02:18:40 - INFO - codeparrot_training - Step 5950: {'lr': 0.0004916919089942519, 'samples': 1142592, 'steps': 5950, 'loss/train': 3.4483240842819214} 01/28/2022 02:18:46 - INFO - codeparrot_training - Step 5951: {'lr': 0.0004916877253039448, 'samples': 1142784, 'steps': 5951, 'loss/train': 3.8045178651809692} 01/28/2022 02:18:50 - INFO - codeparrot_training - Step 5952: {'lr': 0.0004916835405783242, 'samples': 1142976, 'steps': 5952, 'loss/train': 3.162988543510437} 01/28/2022 02:18:54 - INFO - codeparrot_training - Step 5953: {'lr': 0.0004916793548174081, 'samples': 1143168, 'steps': 5953, 'loss/train': 2.8275105357170105} 01/28/2022 02:18:58 - INFO - codeparrot_training - Step 5954: {'lr': 0.0004916751680212145, 'samples': 1143360, 'steps': 5954, 'loss/train': 2.3208221197128296} 01/28/2022 02:19:03 - INFO - codeparrot_training - Step 5955: {'lr': 0.000491670980189761, 'samples': 1143552, 'steps': 5955, 'loss/train': 2.7698206901550293} 01/28/2022 02:19:10 - INFO - codeparrot_training - Step 5956: {'lr': 0.0004916667913230659, 'samples': 1143744, 'steps': 5956, 'loss/train': 1.877447783946991} 01/28/2022 02:19:14 - INFO - codeparrot_training - Step 5957: {'lr': 0.000491662601421147, 'samples': 1143936, 'steps': 5957, 'loss/train': 1.314138025045395} 01/28/2022 02:19:18 - INFO - codeparrot_training - Step 5958: {'lr': 0.0004916584104840222, 'samples': 1144128, 'steps': 5958, 'loss/train': 2.6855592727661133} 01/28/2022 02:19:22 - INFO - codeparrot_training - Step 5959: {'lr': 0.0004916542185117095, 'samples': 1144320, 'steps': 5959, 'loss/train': 3.3779118061065674} 01/28/2022 02:19:28 - INFO - codeparrot_training - Step 5960: {'lr': 0.0004916500255042268, 'samples': 1144512, 'steps': 5960, 'loss/train': 2.4503114819526672} 01/28/2022 02:19:32 - INFO - codeparrot_training - Step 5961: {'lr': 0.0004916458314615923, 'samples': 1144704, 'steps': 5961, 'loss/train': 1.9593923091888428} 01/28/2022 02:19:36 - INFO - codeparrot_training - Step 5962: {'lr': 0.0004916416363838237, 'samples': 1144896, 'steps': 5962, 'loss/train': 2.5353936553001404} 01/28/2022 02:19:40 - INFO - codeparrot_training - Step 5963: {'lr': 0.000491637440270939, 'samples': 1145088, 'steps': 5963, 'loss/train': 3.607673764228821} 01/28/2022 02:19:45 - INFO - codeparrot_training - Step 5964: {'lr': 0.0004916332431229562, 'samples': 1145280, 'steps': 5964, 'loss/train': 1.3740708231925964} 01/28/2022 02:19:49 - INFO - codeparrot_training - Step 5965: {'lr': 0.0004916290449398934, 'samples': 1145472, 'steps': 5965, 'loss/train': 1.5140195488929749} 01/28/2022 02:19:54 - INFO - codeparrot_training - Step 5966: {'lr': 0.0004916248457217686, 'samples': 1145664, 'steps': 5966, 'loss/train': 3.7657620906829834} 01/28/2022 02:19:59 - INFO - codeparrot_training - Step 5967: {'lr': 0.0004916206454685995, 'samples': 1145856, 'steps': 5967, 'loss/train': 2.3440248370170593} 01/28/2022 02:20:03 - INFO - codeparrot_training - Step 5968: {'lr': 0.0004916164441804044, 'samples': 1146048, 'steps': 5968, 'loss/train': 2.4932186007499695} 01/28/2022 02:20:07 - INFO - codeparrot_training - Step 5969: {'lr': 0.0004916122418572011, 'samples': 1146240, 'steps': 5969, 'loss/train': 2.1882011890411377} 01/28/2022 02:20:11 - INFO - codeparrot_training - Step 5970: {'lr': 0.0004916080384990077, 'samples': 1146432, 'steps': 5970, 'loss/train': 0.5732191354036331} 01/28/2022 02:20:18 - INFO - codeparrot_training - Step 5971: {'lr': 0.0004916038341058423, 'samples': 1146624, 'steps': 5971, 'loss/train': 3.0515867471694946} 01/28/2022 02:20:22 - INFO - codeparrot_training - Step 5972: {'lr': 0.0004915996286777226, 'samples': 1146816, 'steps': 5972, 'loss/train': 1.846244215965271} 01/28/2022 02:20:26 - INFO - codeparrot_training - Step 5973: {'lr': 0.0004915954222146669, 'samples': 1147008, 'steps': 5973, 'loss/train': 1.5028182864189148} 01/28/2022 02:20:31 - INFO - codeparrot_training - Step 5974: {'lr': 0.0004915912147166932, 'samples': 1147200, 'steps': 5974, 'loss/train': 2.5371962785720825} 01/28/2022 02:20:35 - INFO - codeparrot_training - Step 5975: {'lr': 0.0004915870061838193, 'samples': 1147392, 'steps': 5975, 'loss/train': 3.1063345670700073} 01/28/2022 02:20:41 - INFO - codeparrot_training - Step 5976: {'lr': 0.0004915827966160634, 'samples': 1147584, 'steps': 5976, 'loss/train': 2.914672315120697} 01/28/2022 02:20:45 - INFO - codeparrot_training - Step 5977: {'lr': 0.0004915785860134436, 'samples': 1147776, 'steps': 5977, 'loss/train': 0.9372382164001465} 01/28/2022 02:20:49 - INFO - codeparrot_training - Step 5978: {'lr': 0.0004915743743759779, 'samples': 1147968, 'steps': 5978, 'loss/train': 3.2123258113861084} 01/28/2022 02:20:53 - INFO - codeparrot_training - Step 5979: {'lr': 0.0004915701617036842, 'samples': 1148160, 'steps': 5979, 'loss/train': 3.375074028968811} 01/28/2022 02:21:00 - INFO - codeparrot_training - Step 5980: {'lr': 0.0004915659479965806, 'samples': 1148352, 'steps': 5980, 'loss/train': 3.6590033769607544} 01/28/2022 02:21:04 - INFO - codeparrot_training - Step 5981: {'lr': 0.0004915617332546852, 'samples': 1148544, 'steps': 5981, 'loss/train': 3.201115608215332} 01/28/2022 02:21:09 - INFO - codeparrot_training - Step 5982: {'lr': 0.0004915575174780161, 'samples': 1148736, 'steps': 5982, 'loss/train': 3.563143730163574} 01/28/2022 02:21:13 - INFO - codeparrot_training - Step 5983: {'lr': 0.0004915533006665912, 'samples': 1148928, 'steps': 5983, 'loss/train': 3.1464014053344727} 01/28/2022 02:21:17 - INFO - codeparrot_training - Step 5984: {'lr': 0.0004915490828204287, 'samples': 1149120, 'steps': 5984, 'loss/train': 2.7048035860061646} 01/28/2022 02:21:22 - INFO - codeparrot_training - Step 5985: {'lr': 0.0004915448639395466, 'samples': 1149312, 'steps': 5985, 'loss/train': 2.841071605682373} 01/28/2022 02:21:27 - INFO - codeparrot_training - Step 5986: {'lr': 0.0004915406440239631, 'samples': 1149504, 'steps': 5986, 'loss/train': 1.8809019327163696} 01/28/2022 02:21:31 - INFO - codeparrot_training - Step 5987: {'lr': 0.0004915364230736961, 'samples': 1149696, 'steps': 5987, 'loss/train': 3.030905842781067} 01/28/2022 02:21:35 - INFO - codeparrot_training - Step 5988: {'lr': 0.0004915322010887637, 'samples': 1149888, 'steps': 5988, 'loss/train': 3.030123710632324} 01/28/2022 02:21:39 - INFO - codeparrot_training - Step 5989: {'lr': 0.0004915279780691843, 'samples': 1150080, 'steps': 5989, 'loss/train': 2.787503242492676} 01/28/2022 02:21:44 - INFO - codeparrot_training - Step 5990: {'lr': 0.0004915237540149755, 'samples': 1150272, 'steps': 5990, 'loss/train': 3.2160000801086426} 01/28/2022 02:21:49 - INFO - codeparrot_training - Step 5991: {'lr': 0.0004915195289261557, 'samples': 1150464, 'steps': 5991, 'loss/train': 1.4503836929798126} 01/28/2022 02:21:53 - INFO - codeparrot_training - Step 5992: {'lr': 0.0004915153028027429, 'samples': 1150656, 'steps': 5992, 'loss/train': 3.1259032487869263} 01/28/2022 02:21:57 - INFO - codeparrot_training - Step 5993: {'lr': 0.0004915110756447552, 'samples': 1150848, 'steps': 5993, 'loss/train': 3.9006028175354004} 01/28/2022 02:22:01 - INFO - codeparrot_training - Step 5994: {'lr': 0.0004915068474522109, 'samples': 1151040, 'steps': 5994, 'loss/train': 2.8204160928726196} 01/28/2022 02:22:07 - INFO - codeparrot_training - Step 5995: {'lr': 0.0004915026182251278, 'samples': 1151232, 'steps': 5995, 'loss/train': 3.1630518436431885} 01/28/2022 02:22:11 - INFO - codeparrot_training - Step 5996: {'lr': 0.0004914983879635242, 'samples': 1151424, 'steps': 5996, 'loss/train': 1.7003368735313416} 01/28/2022 02:22:15 - INFO - codeparrot_training - Step 5997: {'lr': 0.0004914941566674183, 'samples': 1151616, 'steps': 5997, 'loss/train': 3.221074104309082} 01/28/2022 02:22:19 - INFO - codeparrot_training - Step 5998: {'lr': 0.0004914899243368279, 'samples': 1151808, 'steps': 5998, 'loss/train': 1.612649381160736} 01/28/2022 02:22:24 - INFO - codeparrot_training - Step 5999: {'lr': 0.0004914856909717715, 'samples': 1152000, 'steps': 5999, 'loss/train': 3.129716634750366} 01/28/2022 02:22:24 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/28/2022 12:53:15 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 01/28/2022 12:53:15 - WARNING - huggingface_hub.repository - Revision `expert-dust-2` does not exist. Created and checked out branch `expert-dust-2`. 01/28/2022 12:53:15 - WARNING - huggingface_hub.repository - 01/28/2022 12:53:30 - WARNING - datasets.builder - Using custom data configuration lvwerra___codeparrot-clean-train-a1efdd1059bd841d 01/28/2022 12:53:31 - WARNING - datasets.builder - Using custom data configuration lvwerra___codeparrot-clean-valid-a800eb55c299abc0 01/28/2022 12:54:16 - INFO - codeparrot_training - Step 0: {'lr': 0.0, 'samples': 192, 'steps': 0, 'loss/train': 3.179044246673584} 01/28/2022 12:55:37 - INFO - codeparrot_training - Step 1: {'lr': 2.5e-07, 'samples': 384, 'steps': 1, 'loss/train': 4.889771103858948} 01/28/2022 12:56:57 - INFO - codeparrot_training - Step 2: {'lr': 5e-07, 'samples': 576, 'steps': 2, 'loss/train': 3.6975746154785156} 01/28/2022 12:57:01 - INFO - codeparrot_training - Step 3: {'lr': 7.5e-07, 'samples': 768, 'steps': 3, 'loss/train': 5.370334267616272} 01/28/2022 12:57:08 - INFO - codeparrot_training - Step 4: {'lr': 1e-06, 'samples': 960, 'steps': 4, 'loss/train': 5.026500463485718} 01/28/2022 12:57:12 - INFO - codeparrot_training - Step 5: {'lr': 1.25e-06, 'samples': 1152, 'steps': 5, 'loss/train': 4.337551116943359} 01/28/2022 12:57:16 - INFO - codeparrot_training - Step 6: {'lr': 1.5e-06, 'samples': 1344, 'steps': 6, 'loss/train': 3.5363019704818726} 01/28/2022 12:57:21 - INFO - codeparrot_training - Step 7: {'lr': 1.75e-06, 'samples': 1536, 'steps': 7, 'loss/train': 4.8946956396102905} 01/28/2022 12:57:25 - INFO - codeparrot_training - Step 8: {'lr': 2e-06, 'samples': 1728, 'steps': 8, 'loss/train': 4.486309289932251} 01/28/2022 12:57:30 - INFO - codeparrot_training - Step 9: {'lr': 2.25e-06, 'samples': 1920, 'steps': 9, 'loss/train': 2.805045783519745} 01/28/2022 12:57:34 - INFO - codeparrot_training - Step 10: {'lr': 2.5e-06, 'samples': 2112, 'steps': 10, 'loss/train': 3.862419605255127} 01/28/2022 12:57:38 - INFO - codeparrot_training - Step 11: {'lr': 2.75e-06, 'samples': 2304, 'steps': 11, 'loss/train': 4.180359721183777} 01/28/2022 12:57:43 - INFO - codeparrot_training - Step 12: {'lr': 3e-06, 'samples': 2496, 'steps': 12, 'loss/train': 1.32671919465065} 01/28/2022 12:57:47 - INFO - codeparrot_training - Step 13: {'lr': 3.25e-06, 'samples': 2688, 'steps': 13, 'loss/train': 5.3362873792648315} 01/28/2022 12:57:54 - INFO - codeparrot_training - Step 14: {'lr': 3.5e-06, 'samples': 2880, 'steps': 14, 'loss/train': 4.131194829940796} 01/28/2022 12:57:58 - INFO - codeparrot_training - Step 15: {'lr': 3.75e-06, 'samples': 3072, 'steps': 15, 'loss/train': 4.904746413230896} 01/28/2022 12:58:02 - INFO - codeparrot_training - Step 16: {'lr': 4e-06, 'samples': 3264, 'steps': 16, 'loss/train': 4.231512665748596} 01/28/2022 12:58:06 - INFO - codeparrot_training - Step 17: {'lr': 4.250000000000001e-06, 'samples': 3456, 'steps': 17, 'loss/train': 3.7472927570343018} 01/28/2022 12:58:10 - INFO - codeparrot_training - Step 18: {'lr': 4.5e-06, 'samples': 3648, 'steps': 18, 'loss/train': 3.436768054962158} 01/28/2022 12:58:15 - INFO - codeparrot_training - Step 19: {'lr': 4.75e-06, 'samples': 3840, 'steps': 19, 'loss/train': 0.7554210126399994} 01/28/2022 12:58:19 - INFO - codeparrot_training - Step 20: {'lr': 5e-06, 'samples': 4032, 'steps': 20, 'loss/train': 3.5676480531692505} 01/28/2022 12:58:24 - INFO - codeparrot_training - Step 21: {'lr': 5.2500000000000006e-06, 'samples': 4224, 'steps': 21, 'loss/train': 1.501009225845337} 01/28/2022 12:58:28 - INFO - codeparrot_training - Step 22: {'lr': 5.5e-06, 'samples': 4416, 'steps': 22, 'loss/train': 3.9419503211975098} 01/28/2022 12:58:32 - INFO - codeparrot_training - Step 23: {'lr': 5.75e-06, 'samples': 4608, 'steps': 23, 'loss/train': 3.0857309103012085} 01/28/2022 12:58:37 - INFO - codeparrot_training - Step 24: {'lr': 6e-06, 'samples': 4800, 'steps': 24, 'loss/train': 3.599100351333618} 01/28/2022 12:58:41 - INFO - codeparrot_training - Step 25: {'lr': 6.25e-06, 'samples': 4992, 'steps': 25, 'loss/train': 4.2149587869644165} 01/28/2022 12:58:45 - INFO - codeparrot_training - Step 26: {'lr': 6.5e-06, 'samples': 5184, 'steps': 26, 'loss/train': 3.992274284362793} 01/28/2022 12:58:50 - INFO - codeparrot_training - Step 27: {'lr': 6.75e-06, 'samples': 5376, 'steps': 27, 'loss/train': 4.993232846260071} 01/28/2022 12:58:54 - INFO - codeparrot_training - Step 28: {'lr': 7e-06, 'samples': 5568, 'steps': 28, 'loss/train': 2.351971685886383} 01/28/2022 12:59:00 - INFO - codeparrot_training - Step 29: {'lr': 7.250000000000001e-06, 'samples': 5760, 'steps': 29, 'loss/train': 4.079081296920776} 01/28/2022 12:59:04 - INFO - codeparrot_training - Step 30: {'lr': 7.5e-06, 'samples': 5952, 'steps': 30, 'loss/train': 2.6876513957977295} 01/28/2022 12:59:09 - INFO - codeparrot_training - Step 31: {'lr': 7.75e-06, 'samples': 6144, 'steps': 31, 'loss/train': 2.663791537284851} 01/28/2022 12:59:13 - INFO - codeparrot_training - Step 32: {'lr': 8e-06, 'samples': 6336, 'steps': 32, 'loss/train': 5.024893283843994} 01/28/2022 12:59:17 - INFO - codeparrot_training - Step 33: {'lr': 8.25e-06, 'samples': 6528, 'steps': 33, 'loss/train': 3.3836392164230347} 01/28/2022 12:59:22 - INFO - codeparrot_training - Step 34: {'lr': 8.500000000000002e-06, 'samples': 6720, 'steps': 34, 'loss/train': 4.78788149356842} 01/28/2022 12:59:27 - INFO - codeparrot_training - Step 35: {'lr': 8.750000000000001e-06, 'samples': 6912, 'steps': 35, 'loss/train': 3.0112109184265137} 01/28/2022 12:59:31 - INFO - codeparrot_training - Step 36: {'lr': 9e-06, 'samples': 7104, 'steps': 36, 'loss/train': 3.644876003265381} 01/28/2022 12:59:35 - INFO - codeparrot_training - Step 37: {'lr': 9.25e-06, 'samples': 7296, 'steps': 37, 'loss/train': 2.7675840854644775} 01/28/2022 12:59:39 - INFO - codeparrot_training - Step 38: {'lr': 9.5e-06, 'samples': 7488, 'steps': 38, 'loss/train': 4.073085308074951} 01/28/2022 12:59:44 - INFO - codeparrot_training - Step 39: {'lr': 9.75e-06, 'samples': 7680, 'steps': 39, 'loss/train': 2.273773491382599} 01/28/2022 12:59:48 - INFO - codeparrot_training - Step 40: {'lr': 1e-05, 'samples': 7872, 'steps': 40, 'loss/train': 4.276880621910095} 01/28/2022 12:59:52 - INFO - codeparrot_training - Step 41: {'lr': 1.025e-05, 'samples': 8064, 'steps': 41, 'loss/train': 3.3449935913085938} 01/28/2022 12:59:57 - INFO - codeparrot_training - Step 42: {'lr': 1.0500000000000001e-05, 'samples': 8256, 'steps': 42, 'loss/train': 1.715063989162445} 01/28/2022 13:00:01 - INFO - codeparrot_training - Step 43: {'lr': 1.0749999999999999e-05, 'samples': 8448, 'steps': 43, 'loss/train': 3.7758289575576782} 01/28/2022 13:00:09 - INFO - codeparrot_training - Step 44: {'lr': 1.1e-05, 'samples': 8640, 'steps': 44, 'loss/train': 3.2152655124664307} 01/28/2022 13:00:13 - INFO - codeparrot_training - Step 45: {'lr': 1.1249999999999999e-05, 'samples': 8832, 'steps': 45, 'loss/train': 4.526304244995117} 01/28/2022 13:00:17 - INFO - codeparrot_training - Step 46: {'lr': 1.15e-05, 'samples': 9024, 'steps': 46, 'loss/train': 2.4934746623039246} 01/28/2022 13:00:21 - INFO - codeparrot_training - Step 47: {'lr': 1.1750000000000001e-05, 'samples': 9216, 'steps': 47, 'loss/train': 2.794708549976349} 01/28/2022 13:00:25 - INFO - codeparrot_training - Step 48: {'lr': 1.2e-05, 'samples': 9408, 'steps': 48, 'loss/train': 4.1755582094192505} 01/28/2022 13:00:30 - INFO - codeparrot_training - Step 49: {'lr': 1.2250000000000001e-05, 'samples': 9600, 'steps': 49, 'loss/train': 2.597469985485077} 01/28/2022 13:00:35 - INFO - codeparrot_training - Step 50: {'lr': 1.25e-05, 'samples': 9792, 'steps': 50, 'loss/train': 3.3644471168518066} 01/28/2022 13:00:39 - INFO - codeparrot_training - Step 51: {'lr': 1.275e-05, 'samples': 9984, 'steps': 51, 'loss/train': 1.8318015933036804} 01/28/2022 13:00:43 - INFO - codeparrot_training - Step 52: {'lr': 1.3e-05, 'samples': 10176, 'steps': 52, 'loss/train': 2.7238054275512695} 01/28/2022 13:00:47 - INFO - codeparrot_training - Step 53: {'lr': 1.325e-05, 'samples': 10368, 'steps': 53, 'loss/train': 1.2045309841632843} 01/28/2022 13:00:52 - INFO - codeparrot_training - Step 54: {'lr': 1.35e-05, 'samples': 10560, 'steps': 54, 'loss/train': 3.968233108520508} 01/28/2022 13:00:56 - INFO - codeparrot_training - Step 55: {'lr': 1.375e-05, 'samples': 10752, 'steps': 55, 'loss/train': 2.1083736419677734} 01/28/2022 13:01:01 - INFO - codeparrot_training - Step 56: {'lr': 1.4e-05, 'samples': 10944, 'steps': 56, 'loss/train': 3.1135400533676147} 01/28/2022 13:01:05 - INFO - codeparrot_training - Step 57: {'lr': 1.425e-05, 'samples': 11136, 'steps': 57, 'loss/train': 2.3143442273139954} 01/28/2022 13:01:09 - INFO - codeparrot_training - Step 58: {'lr': 1.4500000000000002e-05, 'samples': 11328, 'steps': 58, 'loss/train': 3.1513763666152954} 01/28/2022 13:01:14 - INFO - codeparrot_training - Step 59: {'lr': 1.475e-05, 'samples': 11520, 'steps': 59, 'loss/train': 2.0897560715675354} 01/28/2022 13:01:18 - INFO - codeparrot_training - Step 60: {'lr': 1.5e-05, 'samples': 11712, 'steps': 60, 'loss/train': 1.450158029794693} 01/28/2022 13:01:22 - INFO - codeparrot_training - Step 61: {'lr': 1.525e-05, 'samples': 11904, 'steps': 61, 'loss/train': 2.799786686897278} 01/28/2022 13:01:27 - INFO - codeparrot_training - Step 62: {'lr': 1.55e-05, 'samples': 12096, 'steps': 62, 'loss/train': 2.048109233379364} 01/28/2022 13:01:31 - INFO - codeparrot_training - Step 63: {'lr': 1.575e-05, 'samples': 12288, 'steps': 63, 'loss/train': 1.6185863614082336} 01/28/2022 13:01:38 - INFO - codeparrot_training - Step 64: {'lr': 1.6e-05, 'samples': 12480, 'steps': 64, 'loss/train': 2.163468897342682} 01/28/2022 13:01:42 - INFO - codeparrot_training - Step 65: {'lr': 1.6250000000000002e-05, 'samples': 12672, 'steps': 65, 'loss/train': 3.0222734212875366} 01/28/2022 13:01:46 - INFO - codeparrot_training - Step 66: {'lr': 1.65e-05, 'samples': 12864, 'steps': 66, 'loss/train': 2.523698329925537} 01/28/2022 13:01:50 - INFO - codeparrot_training - Step 67: {'lr': 1.675e-05, 'samples': 13056, 'steps': 67, 'loss/train': 1.986411452293396} 01/28/2022 13:01:54 - INFO - codeparrot_training - Step 68: {'lr': 1.7000000000000003e-05, 'samples': 13248, 'steps': 68, 'loss/train': 3.053613066673279} 01/28/2022 13:02:00 - INFO - codeparrot_training - Step 69: {'lr': 1.7250000000000003e-05, 'samples': 13440, 'steps': 69, 'loss/train': 1.8446375727653503} 01/28/2022 13:02:04 - INFO - codeparrot_training - Step 70: {'lr': 1.7500000000000002e-05, 'samples': 13632, 'steps': 70, 'loss/train': 2.6805474758148193} 01/28/2022 13:02:08 - INFO - codeparrot_training - Step 71: {'lr': 1.7749999999999998e-05, 'samples': 13824, 'steps': 71, 'loss/train': 2.880137264728546} 01/28/2022 13:02:12 - INFO - codeparrot_training - Step 72: {'lr': 1.8e-05, 'samples': 14016, 'steps': 72, 'loss/train': 3.41677987575531} 01/28/2022 13:02:16 - INFO - codeparrot_training - Step 73: {'lr': 1.825e-05, 'samples': 14208, 'steps': 73, 'loss/train': 2.2050638794898987} 01/28/2022 13:02:23 - INFO - codeparrot_training - Step 74: {'lr': 1.85e-05, 'samples': 14400, 'steps': 74, 'loss/train': 3.4777203798294067} 01/28/2022 13:02:27 - INFO - codeparrot_training - Step 75: {'lr': 1.875e-05, 'samples': 14592, 'steps': 75, 'loss/train': 2.9114425778388977} 01/28/2022 13:02:32 - INFO - codeparrot_training - Step 76: {'lr': 1.9e-05, 'samples': 14784, 'steps': 76, 'loss/train': 3.8636906147003174} 01/28/2022 13:02:36 - INFO - codeparrot_training - Step 77: {'lr': 1.925e-05, 'samples': 14976, 'steps': 77, 'loss/train': 3.2276644706726074} 01/28/2022 13:02:41 - INFO - codeparrot_training - Step 78: {'lr': 1.95e-05, 'samples': 15168, 'steps': 78, 'loss/train': 4.086126208305359} 01/28/2022 13:02:45 - INFO - codeparrot_training - Step 79: {'lr': 1.975e-05, 'samples': 15360, 'steps': 79, 'loss/train': 2.5836841464042664} 01/28/2022 13:02:49 - INFO - codeparrot_training - Step 80: {'lr': 2e-05, 'samples': 15552, 'steps': 80, 'loss/train': 3.178232431411743} 01/28/2022 13:02:53 - INFO - codeparrot_training - Step 81: {'lr': 2.025e-05, 'samples': 15744, 'steps': 81, 'loss/train': 1.7823808193206787} 01/28/2022 13:02:58 - INFO - codeparrot_training - Step 82: {'lr': 2.05e-05, 'samples': 15936, 'steps': 82, 'loss/train': 2.687070071697235} 01/28/2022 13:03:02 - INFO - codeparrot_training - Step 83: {'lr': 2.0750000000000003e-05, 'samples': 16128, 'steps': 83, 'loss/train': 1.0396304726600647} 01/28/2022 13:03:07 - INFO - codeparrot_training - Step 84: {'lr': 2.1000000000000002e-05, 'samples': 16320, 'steps': 84, 'loss/train': 1.7422910928726196} 01/28/2022 13:03:11 - INFO - codeparrot_training - Step 85: {'lr': 2.125e-05, 'samples': 16512, 'steps': 85, 'loss/train': 1.389947086572647} 01/28/2022 13:03:15 - INFO - codeparrot_training - Step 86: {'lr': 2.1499999999999997e-05, 'samples': 16704, 'steps': 86, 'loss/train': 3.1441233158111572} 01/28/2022 13:03:19 - INFO - codeparrot_training - Step 87: {'lr': 2.175e-05, 'samples': 16896, 'steps': 87, 'loss/train': 2.5782228112220764} 01/28/2022 13:03:24 - INFO - codeparrot_training - Step 88: {'lr': 2.2e-05, 'samples': 17088, 'steps': 88, 'loss/train': 1.3218863904476166} 01/28/2022 13:03:30 - INFO - codeparrot_training - Step 89: {'lr': 2.225e-05, 'samples': 17280, 'steps': 89, 'loss/train': 3.5424349308013916} 01/28/2022 13:03:34 - INFO - codeparrot_training - Step 90: {'lr': 2.2499999999999998e-05, 'samples': 17472, 'steps': 90, 'loss/train': 1.9482693672180176} 01/28/2022 13:03:39 - INFO - codeparrot_training - Step 91: {'lr': 2.275e-05, 'samples': 17664, 'steps': 91, 'loss/train': 2.814733564853668} 01/28/2022 13:03:43 - INFO - codeparrot_training - Step 92: {'lr': 2.3e-05, 'samples': 17856, 'steps': 92, 'loss/train': 2.916082262992859} 01/28/2022 13:03:47 - INFO - codeparrot_training - Step 93: {'lr': 2.325e-05, 'samples': 18048, 'steps': 93, 'loss/train': 1.2317577302455902} 01/28/2022 13:03:52 - INFO - codeparrot_training - Step 94: {'lr': 2.3500000000000002e-05, 'samples': 18240, 'steps': 94, 'loss/train': 3.3723506927490234} 01/28/2022 13:03:56 - INFO - codeparrot_training - Step 95: {'lr': 2.375e-05, 'samples': 18432, 'steps': 95, 'loss/train': 2.5385663509368896} 01/28/2022 13:04:01 - INFO - codeparrot_training - Step 96: {'lr': 2.4e-05, 'samples': 18624, 'steps': 96, 'loss/train': 2.389790117740631} 01/28/2022 13:04:05 - INFO - codeparrot_training - Step 97: {'lr': 2.425e-05, 'samples': 18816, 'steps': 97, 'loss/train': 1.5918365120887756} 01/28/2022 13:04:10 - INFO - codeparrot_training - Step 98: {'lr': 2.4500000000000003e-05, 'samples': 19008, 'steps': 98, 'loss/train': 2.6047056913375854} 01/28/2022 13:04:14 - INFO - codeparrot_training - Step 99: {'lr': 2.4750000000000002e-05, 'samples': 19200, 'steps': 99, 'loss/train': 2.6025636792182922} 01/28/2022 13:04:18 - INFO - codeparrot_training - Step 100: {'lr': 2.5e-05, 'samples': 19392, 'steps': 100, 'loss/train': 2.734703779220581} 01/28/2022 13:04:22 - INFO - codeparrot_training - Step 101: {'lr': 2.525e-05, 'samples': 19584, 'steps': 101, 'loss/train': 1.897295594215393} 01/28/2022 13:04:26 - INFO - codeparrot_training - Step 102: {'lr': 2.55e-05, 'samples': 19776, 'steps': 102, 'loss/train': 3.2490978240966797} 01/28/2022 13:04:33 - INFO - codeparrot_training - Step 103: {'lr': 2.575e-05, 'samples': 19968, 'steps': 103, 'loss/train': 2.876462996006012} 01/28/2022 13:04:37 - INFO - codeparrot_training - Step 104: {'lr': 2.6e-05, 'samples': 20160, 'steps': 104, 'loss/train': 2.6888392567634583} 01/28/2022 13:04:42 - INFO - codeparrot_training - Step 105: {'lr': 2.625e-05, 'samples': 20352, 'steps': 105, 'loss/train': 2.157119035720825} 01/28/2022 13:04:46 - INFO - codeparrot_training - Step 106: {'lr': 2.65e-05, 'samples': 20544, 'steps': 106, 'loss/train': 2.9895100593566895} 01/28/2022 13:04:50 - INFO - codeparrot_training - Step 107: {'lr': 2.675e-05, 'samples': 20736, 'steps': 107, 'loss/train': 2.2355427145957947} 01/28/2022 13:04:55 - INFO - codeparrot_training - Step 108: {'lr': 2.7e-05, 'samples': 20928, 'steps': 108, 'loss/train': 2.070980429649353} 01/28/2022 13:04:59 - INFO - codeparrot_training - Step 109: {'lr': 2.725e-05, 'samples': 21120, 'steps': 109, 'loss/train': 3.36741042137146} 01/28/2022 13:05:04 - INFO - codeparrot_training - Step 110: {'lr': 2.75e-05, 'samples': 21312, 'steps': 110, 'loss/train': 2.2960386872291565} 01/28/2022 13:05:08 - INFO - codeparrot_training - Step 111: {'lr': 2.775e-05, 'samples': 21504, 'steps': 111, 'loss/train': 1.2417257130146027} 01/28/2022 13:05:12 - INFO - codeparrot_training - Step 112: {'lr': 2.8e-05, 'samples': 21696, 'steps': 112, 'loss/train': 2.0917139053344727} 01/28/2022 13:05:20 - INFO - codeparrot_training - Step 113: {'lr': 2.8250000000000002e-05, 'samples': 21888, 'steps': 113, 'loss/train': 3.0688995122909546} 01/28/2022 13:05:24 - INFO - codeparrot_training - Step 114: {'lr': 2.85e-05, 'samples': 22080, 'steps': 114, 'loss/train': 2.2963083386421204} 01/28/2022 13:05:28 - INFO - codeparrot_training - Step 115: {'lr': 2.875e-05, 'samples': 22272, 'steps': 115, 'loss/train': 2.648421049118042} 01/28/2022 13:05:32 - INFO - codeparrot_training - Step 116: {'lr': 2.9000000000000004e-05, 'samples': 22464, 'steps': 116, 'loss/train': 1.5514199137687683} 01/28/2022 13:05:36 - INFO - codeparrot_training - Step 117: {'lr': 2.9250000000000003e-05, 'samples': 22656, 'steps': 117, 'loss/train': 4.397069692611694} 01/28/2022 13:05:40 - INFO - codeparrot_training - Step 118: {'lr': 2.95e-05, 'samples': 22848, 'steps': 118, 'loss/train': 2.3208956122398376} 01/28/2022 13:05:46 - INFO - codeparrot_training - Step 119: {'lr': 2.9749999999999998e-05, 'samples': 23040, 'steps': 119, 'loss/train': 1.5385152697563171} 01/28/2022 13:05:50 - INFO - codeparrot_training - Step 120: {'lr': 3e-05, 'samples': 23232, 'steps': 120, 'loss/train': 1.1449499130249023} 01/28/2022 13:05:54 - INFO - codeparrot_training - Step 121: {'lr': 3.025e-05, 'samples': 23424, 'steps': 121, 'loss/train': 3.054843306541443} 01/28/2022 13:05:58 - INFO - codeparrot_training - Step 122: {'lr': 3.05e-05, 'samples': 23616, 'steps': 122, 'loss/train': 2.37749183177948} 01/28/2022 13:06:02 - INFO - codeparrot_training - Step 123: {'lr': 3.075e-05, 'samples': 23808, 'steps': 123, 'loss/train': 2.101391851902008} 01/28/2022 13:06:07 - INFO - codeparrot_training - Step 124: {'lr': 3.1e-05, 'samples': 24000, 'steps': 124, 'loss/train': 0.6038169264793396} 01/28/2022 13:06:12 - INFO - codeparrot_training - Step 125: {'lr': 3.125e-05, 'samples': 24192, 'steps': 125, 'loss/train': 2.088426947593689} 01/28/2022 13:06:16 - INFO - codeparrot_training - Step 126: {'lr': 3.15e-05, 'samples': 24384, 'steps': 126, 'loss/train': 2.763307750225067} 01/28/2022 13:06:20 - INFO - codeparrot_training - Step 127: {'lr': 3.175e-05, 'samples': 24576, 'steps': 127, 'loss/train': 2.4774681329727173} 01/28/2022 13:06:24 - INFO - codeparrot_training - Step 128: {'lr': 3.2e-05, 'samples': 24768, 'steps': 128, 'loss/train': 2.5064316987991333} 01/28/2022 13:06:29 - INFO - codeparrot_training - Step 129: {'lr': 3.2250000000000005e-05, 'samples': 24960, 'steps': 129, 'loss/train': 2.1853771805763245} 01/28/2022 13:06:33 - INFO - codeparrot_training - Step 130: {'lr': 3.2500000000000004e-05, 'samples': 25152, 'steps': 130, 'loss/train': 2.8304572105407715} 01/28/2022 13:06:37 - INFO - codeparrot_training - Step 131: {'lr': 3.275e-05, 'samples': 25344, 'steps': 131, 'loss/train': 2.1926037669181824} 01/28/2022 13:06:42 - INFO - codeparrot_training - Step 132: {'lr': 3.3e-05, 'samples': 25536, 'steps': 132, 'loss/train': 2.542338788509369} 01/28/2022 13:06:46 - INFO - codeparrot_training - Step 133: {'lr': 3.325e-05, 'samples': 25728, 'steps': 133, 'loss/train': 2.3067288994789124} 01/28/2022 13:06:53 - INFO - codeparrot_training - Step 134: {'lr': 3.35e-05, 'samples': 25920, 'steps': 134, 'loss/train': 4.964722394943237} 01/28/2022 13:06:57 - INFO - codeparrot_training - Step 135: {'lr': 3.375e-05, 'samples': 26112, 'steps': 135, 'loss/train': 1.851418912410736} 01/28/2022 13:07:01 - INFO - codeparrot_training - Step 136: {'lr': 3.4000000000000007e-05, 'samples': 26304, 'steps': 136, 'loss/train': 1.762540876865387} 01/28/2022 13:07:05 - INFO - codeparrot_training - Step 137: {'lr': 3.4250000000000006e-05, 'samples': 26496, 'steps': 137, 'loss/train': 1.1407091617584229} 01/28/2022 13:07:10 - INFO - codeparrot_training - Step 138: {'lr': 3.4500000000000005e-05, 'samples': 26688, 'steps': 138, 'loss/train': 1.7459387183189392} 01/28/2022 13:07:15 - INFO - codeparrot_training - Step 139: {'lr': 3.4750000000000004e-05, 'samples': 26880, 'steps': 139, 'loss/train': 2.510267972946167} 01/28/2022 13:07:19 - INFO - codeparrot_training - Step 140: {'lr': 3.5000000000000004e-05, 'samples': 27072, 'steps': 140, 'loss/train': 1.379767656326294} 01/28/2022 13:07:23 - INFO - codeparrot_training - Step 141: {'lr': 3.5249999999999996e-05, 'samples': 27264, 'steps': 141, 'loss/train': 1.69757080078125} 01/28/2022 13:07:27 - INFO - codeparrot_training - Step 142: {'lr': 3.5499999999999996e-05, 'samples': 27456, 'steps': 142, 'loss/train': 3.3471726179122925} 01/28/2022 13:07:31 - INFO - codeparrot_training - Step 143: {'lr': 3.5749999999999995e-05, 'samples': 27648, 'steps': 143, 'loss/train': 1.7504510879516602} 01/28/2022 13:07:39 - INFO - codeparrot_training - Step 144: {'lr': 3.6e-05, 'samples': 27840, 'steps': 144, 'loss/train': 2.739710211753845} 01/28/2022 13:07:43 - INFO - codeparrot_training - Step 145: {'lr': 3.625e-05, 'samples': 28032, 'steps': 145, 'loss/train': 1.3888595402240753} 01/28/2022 13:07:47 - INFO - codeparrot_training - Step 146: {'lr': 3.65e-05, 'samples': 28224, 'steps': 146, 'loss/train': 1.75312077999115} 01/28/2022 13:07:51 - INFO - codeparrot_training - Step 147: {'lr': 3.675e-05, 'samples': 28416, 'steps': 147, 'loss/train': 2.441735029220581} 01/28/2022 13:07:55 - INFO - codeparrot_training - Step 148: {'lr': 3.7e-05, 'samples': 28608, 'steps': 148, 'loss/train': 2.997218370437622} 01/28/2022 13:08:00 - INFO - codeparrot_training - Step 149: {'lr': 3.725e-05, 'samples': 28800, 'steps': 149, 'loss/train': 1.257592499256134} 01/28/2022 13:08:05 - INFO - codeparrot_training - Step 150: {'lr': 3.75e-05, 'samples': 28992, 'steps': 150, 'loss/train': 3.246846556663513} 01/28/2022 13:08:09 - INFO - codeparrot_training - Step 151: {'lr': 3.775e-05, 'samples': 29184, 'steps': 151, 'loss/train': 2.535372734069824} 01/28/2022 13:08:13 - INFO - codeparrot_training - Step 152: {'lr': 3.8e-05, 'samples': 29376, 'steps': 152, 'loss/train': 1.0979348123073578} 01/28/2022 13:08:17 - INFO - codeparrot_training - Step 153: {'lr': 3.825e-05, 'samples': 29568, 'steps': 153, 'loss/train': 2.29162734746933} 01/28/2022 13:08:22 - INFO - codeparrot_training - Step 154: {'lr': 3.85e-05, 'samples': 29760, 'steps': 154, 'loss/train': 2.1003686785697937} 01/28/2022 13:08:26 - INFO - codeparrot_training - Step 155: {'lr': 3.875e-05, 'samples': 29952, 'steps': 155, 'loss/train': 2.7438379526138306} 01/28/2022 13:08:31 - INFO - codeparrot_training - Step 156: {'lr': 3.9e-05, 'samples': 30144, 'steps': 156, 'loss/train': 1.9371973872184753} 01/28/2022 13:08:35 - INFO - codeparrot_training - Step 157: {'lr': 3.925e-05, 'samples': 30336, 'steps': 157, 'loss/train': 2.4481106400489807} 01/28/2022 13:08:39 - INFO - codeparrot_training - Step 158: {'lr': 3.95e-05, 'samples': 30528, 'steps': 158, 'loss/train': 2.7821288108825684} 01/28/2022 13:08:46 - INFO - codeparrot_training - Step 159: {'lr': 3.9750000000000004e-05, 'samples': 30720, 'steps': 159, 'loss/train': 1.3834978938102722} 01/28/2022 13:08:50 - INFO - codeparrot_training - Step 160: {'lr': 4e-05, 'samples': 30912, 'steps': 160, 'loss/train': 2.443645656108856} 01/28/2022 13:08:54 - INFO - codeparrot_training - Step 161: {'lr': 4.025e-05, 'samples': 31104, 'steps': 161, 'loss/train': 1.3964619040489197} 01/28/2022 13:08:58 - INFO - codeparrot_training - Step 162: {'lr': 4.05e-05, 'samples': 31296, 'steps': 162, 'loss/train': 2.2152684330940247} 01/28/2022 13:09:03 - INFO - codeparrot_training - Step 163: {'lr': 4.075e-05, 'samples': 31488, 'steps': 163, 'loss/train': 2.8553239703178406} 01/28/2022 13:09:08 - INFO - codeparrot_training - Step 164: {'lr': 4.1e-05, 'samples': 31680, 'steps': 164, 'loss/train': 2.136304020881653} 01/28/2022 13:09:12 - INFO - codeparrot_training - Step 165: {'lr': 4.125e-05, 'samples': 31872, 'steps': 165, 'loss/train': 2.5264483094215393} 01/28/2022 13:09:16 - INFO - codeparrot_training - Step 166: {'lr': 4.1500000000000006e-05, 'samples': 32064, 'steps': 166, 'loss/train': 1.2165210843086243} 01/28/2022 13:09:20 - INFO - codeparrot_training - Step 167: {'lr': 4.1750000000000005e-05, 'samples': 32256, 'steps': 167, 'loss/train': 0.7394502013921738} 01/28/2022 13:09:24 - INFO - codeparrot_training - Step 168: {'lr': 4.2000000000000004e-05, 'samples': 32448, 'steps': 168, 'loss/train': 0.4713757038116455} 01/28/2022 13:09:29 - INFO - codeparrot_training - Step 169: {'lr': 4.2250000000000004e-05, 'samples': 32640, 'steps': 169, 'loss/train': 3.143898367881775} 01/28/2022 13:09:34 - INFO - codeparrot_training - Step 170: {'lr': 4.25e-05, 'samples': 32832, 'steps': 170, 'loss/train': 1.2985304296016693} 01/28/2022 13:09:38 - INFO - codeparrot_training - Step 171: {'lr': 4.275e-05, 'samples': 33024, 'steps': 171, 'loss/train': 2.660888135433197} 01/28/2022 13:09:42 - INFO - codeparrot_training - Step 172: {'lr': 4.2999999999999995e-05, 'samples': 33216, 'steps': 172, 'loss/train': 2.2592217922210693} 01/28/2022 13:09:46 - INFO - codeparrot_training - Step 173: {'lr': 4.325e-05, 'samples': 33408, 'steps': 173, 'loss/train': 1.6953336596488953} 01/28/2022 13:09:51 - INFO - codeparrot_training - Step 174: {'lr': 4.35e-05, 'samples': 33600, 'steps': 174, 'loss/train': 2.445850968360901} 01/28/2022 13:09:55 - INFO - codeparrot_training - Step 175: {'lr': 4.375e-05, 'samples': 33792, 'steps': 175, 'loss/train': 2.746981680393219} 01/28/2022 13:10:00 - INFO - codeparrot_training - Step 176: {'lr': 4.4e-05, 'samples': 33984, 'steps': 176, 'loss/train': 1.595458209514618} 01/28/2022 13:10:04 - INFO - codeparrot_training - Step 177: {'lr': 4.425e-05, 'samples': 34176, 'steps': 177, 'loss/train': 1.7864105701446533} 01/28/2022 13:10:08 - INFO - codeparrot_training - Step 178: {'lr': 4.45e-05, 'samples': 34368, 'steps': 178, 'loss/train': 2.025016486644745} 01/28/2022 13:10:15 - INFO - codeparrot_training - Step 179: {'lr': 4.475e-05, 'samples': 34560, 'steps': 179, 'loss/train': 2.3141865134239197} 01/28/2022 13:10:19 - INFO - codeparrot_training - Step 180: {'lr': 4.4999999999999996e-05, 'samples': 34752, 'steps': 180, 'loss/train': 2.201896905899048} 01/28/2022 13:10:23 - INFO - codeparrot_training - Step 181: {'lr': 4.525e-05, 'samples': 34944, 'steps': 181, 'loss/train': 1.7532691955566406} 01/28/2022 13:10:27 - INFO - codeparrot_training - Step 182: {'lr': 4.55e-05, 'samples': 35136, 'steps': 182, 'loss/train': 0.6213488727807999} 01/28/2022 13:10:31 - INFO - codeparrot_training - Step 183: {'lr': 4.575e-05, 'samples': 35328, 'steps': 183, 'loss/train': 1.7418964505195618} 01/28/2022 13:10:37 - INFO - codeparrot_training - Step 184: {'lr': 4.6e-05, 'samples': 35520, 'steps': 184, 'loss/train': 2.455689489841461} 01/28/2022 13:10:41 - INFO - codeparrot_training - Step 185: {'lr': 4.625e-05, 'samples': 35712, 'steps': 185, 'loss/train': 2.0197967290878296} 01/28/2022 13:10:45 - INFO - codeparrot_training - Step 186: {'lr': 4.65e-05, 'samples': 35904, 'steps': 186, 'loss/train': 2.6454811692237854} 01/28/2022 13:10:49 - INFO - codeparrot_training - Step 187: {'lr': 4.675e-05, 'samples': 36096, 'steps': 187, 'loss/train': 2.2441232800483704} 01/28/2022 13:10:53 - INFO - codeparrot_training - Step 188: {'lr': 4.7000000000000004e-05, 'samples': 36288, 'steps': 188, 'loss/train': 2.2492411136627197} 01/28/2022 13:10:58 - INFO - codeparrot_training - Step 189: {'lr': 4.725e-05, 'samples': 36480, 'steps': 189, 'loss/train': 2.7806480526924133} 01/28/2022 13:11:03 - INFO - codeparrot_training - Step 190: {'lr': 4.75e-05, 'samples': 36672, 'steps': 190, 'loss/train': 2.2955929040908813} 01/28/2022 13:11:07 - INFO - codeparrot_training - Step 191: {'lr': 4.775e-05, 'samples': 36864, 'steps': 191, 'loss/train': 2.0833232402801514} 01/28/2022 13:11:11 - INFO - codeparrot_training - Step 192: {'lr': 4.8e-05, 'samples': 37056, 'steps': 192, 'loss/train': 1.4081032276153564} 01/28/2022 13:11:15 - INFO - codeparrot_training - Step 193: {'lr': 4.825e-05, 'samples': 37248, 'steps': 193, 'loss/train': 1.2339371144771576} 01/28/2022 13:11:22 - INFO - codeparrot_training - Step 194: {'lr': 4.85e-05, 'samples': 37440, 'steps': 194, 'loss/train': 2.5167746543884277} 01/28/2022 13:11:27 - INFO - codeparrot_training - Step 195: {'lr': 4.8750000000000006e-05, 'samples': 37632, 'steps': 195, 'loss/train': 1.6063243746757507} 01/28/2022 13:11:31 - INFO - codeparrot_training - Step 196: {'lr': 4.9000000000000005e-05, 'samples': 37824, 'steps': 196, 'loss/train': 1.9755422472953796} 01/28/2022 13:11:35 - INFO - codeparrot_training - Step 197: {'lr': 4.9250000000000004e-05, 'samples': 38016, 'steps': 197, 'loss/train': 2.374062716960907} 01/28/2022 13:11:39 - INFO - codeparrot_training - Step 198: {'lr': 4.9500000000000004e-05, 'samples': 38208, 'steps': 198, 'loss/train': 2.317784070968628} 01/28/2022 13:11:43 - INFO - codeparrot_training - Step 199: {'lr': 4.975e-05, 'samples': 38400, 'steps': 199, 'loss/train': 2.3045952916145325} 01/28/2022 13:11:48 - INFO - codeparrot_training - Step 200: {'lr': 5e-05, 'samples': 38592, 'steps': 200, 'loss/train': 1.418749988079071} 01/28/2022 13:11:53 - INFO - codeparrot_training - Step 201: {'lr': 5.025e-05, 'samples': 38784, 'steps': 201, 'loss/train': 1.631143569946289} 01/28/2022 13:11:57 - INFO - codeparrot_training - Step 202: {'lr': 5.05e-05, 'samples': 38976, 'steps': 202, 'loss/train': 2.563623011112213} 01/28/2022 13:12:01 - INFO - codeparrot_training - Step 203: {'lr': 5.075000000000001e-05, 'samples': 39168, 'steps': 203, 'loss/train': 2.8652724623680115} 01/28/2022 13:12:05 - INFO - codeparrot_training - Step 204: {'lr': 5.1e-05, 'samples': 39360, 'steps': 204, 'loss/train': 2.67670476436615} 01/28/2022 13:12:12 - INFO - codeparrot_training - Step 205: {'lr': 5.125e-05, 'samples': 39552, 'steps': 205, 'loss/train': 0.7334272116422653} 01/28/2022 13:12:16 - INFO - codeparrot_training - Step 206: {'lr': 5.15e-05, 'samples': 39744, 'steps': 206, 'loss/train': 1.8160684704780579} 01/28/2022 13:12:21 - INFO - codeparrot_training - Step 207: {'lr': 5.175e-05, 'samples': 39936, 'steps': 207, 'loss/train': 2.314116060733795} 01/28/2022 13:12:25 - INFO - codeparrot_training - Step 208: {'lr': 5.2e-05, 'samples': 40128, 'steps': 208, 'loss/train': 2.081115424633026} 01/28/2022 13:12:29 - INFO - codeparrot_training - Step 209: {'lr': 5.2249999999999996e-05, 'samples': 40320, 'steps': 209, 'loss/train': 1.5594000220298767} 01/28/2022 13:12:34 - INFO - codeparrot_training - Step 210: {'lr': 5.25e-05, 'samples': 40512, 'steps': 210, 'loss/train': 2.8253384828567505} 01/28/2022 13:12:38 - INFO - codeparrot_training - Step 211: {'lr': 5.275e-05, 'samples': 40704, 'steps': 211, 'loss/train': 1.4428039491176605} 01/28/2022 13:12:42 - INFO - codeparrot_training - Step 212: {'lr': 5.3e-05, 'samples': 40896, 'steps': 212, 'loss/train': 2.298863410949707} 01/28/2022 13:12:46 - INFO - codeparrot_training - Step 213: {'lr': 5.325e-05, 'samples': 41088, 'steps': 213, 'loss/train': 2.1545201539993286} 01/28/2022 13:12:51 - INFO - codeparrot_training - Step 214: {'lr': 5.35e-05, 'samples': 41280, 'steps': 214, 'loss/train': 1.0137978494167328} 01/28/2022 13:12:56 - INFO - codeparrot_training - Step 215: {'lr': 5.375e-05, 'samples': 41472, 'steps': 215, 'loss/train': 2.675076484680176} 01/28/2022 13:13:00 - INFO - codeparrot_training - Step 216: {'lr': 5.4e-05, 'samples': 41664, 'steps': 216, 'loss/train': 2.0772820115089417} 01/28/2022 13:13:04 - INFO - codeparrot_training - Step 217: {'lr': 5.4250000000000004e-05, 'samples': 41856, 'steps': 217, 'loss/train': 2.3776509761810303} 01/28/2022 13:13:08 - INFO - codeparrot_training - Step 218: {'lr': 5.45e-05, 'samples': 42048, 'steps': 218, 'loss/train': 0.891341507434845} 01/28/2022 13:13:13 - INFO - codeparrot_training - Step 219: {'lr': 5.475e-05, 'samples': 42240, 'steps': 219, 'loss/train': 2.3276649713516235} 01/28/2022 13:13:19 - INFO - codeparrot_training - Step 220: {'lr': 5.5e-05, 'samples': 42432, 'steps': 220, 'loss/train': 1.3762422502040863} 01/28/2022 13:13:24 - INFO - codeparrot_training - Step 221: {'lr': 5.525e-05, 'samples': 42624, 'steps': 221, 'loss/train': 3.1256039142608643} 01/28/2022 13:13:28 - INFO - codeparrot_training - Step 222: {'lr': 5.55e-05, 'samples': 42816, 'steps': 222, 'loss/train': 2.19691264629364} 01/28/2022 13:13:32 - INFO - codeparrot_training - Step 223: {'lr': 5.575e-05, 'samples': 43008, 'steps': 223, 'loss/train': 2.2813870310783386} 01/28/2022 13:13:36 - INFO - codeparrot_training - Step 224: {'lr': 5.6e-05, 'samples': 43200, 'steps': 224, 'loss/train': 1.9025484323501587} 01/28/2022 13:13:41 - INFO - codeparrot_training - Step 225: {'lr': 5.6250000000000005e-05, 'samples': 43392, 'steps': 225, 'loss/train': 0.8648491501808167} 01/28/2022 13:13:46 - INFO - codeparrot_training - Step 226: {'lr': 5.6500000000000005e-05, 'samples': 43584, 'steps': 226, 'loss/train': 2.641293704509735} 01/28/2022 13:13:50 - INFO - codeparrot_training - Step 227: {'lr': 5.6750000000000004e-05, 'samples': 43776, 'steps': 227, 'loss/train': 1.987243115901947} 01/28/2022 13:13:54 - INFO - codeparrot_training - Step 228: {'lr': 5.7e-05, 'samples': 43968, 'steps': 228, 'loss/train': 2.377474844455719} 01/28/2022 13:14:01 - INFO - codeparrot_training - Step 229: {'lr': 5.725e-05, 'samples': 44160, 'steps': 229, 'loss/train': 2.3577218055725098} 01/28/2022 13:14:05 - INFO - codeparrot_training - Step 230: {'lr': 5.75e-05, 'samples': 44352, 'steps': 230, 'loss/train': 2.5242109894752502} 01/28/2022 13:14:09 - INFO - codeparrot_training - Step 231: {'lr': 5.775e-05, 'samples': 44544, 'steps': 231, 'loss/train': 2.576507270336151} 01/28/2022 13:14:13 - INFO - codeparrot_training - Step 232: {'lr': 5.800000000000001e-05, 'samples': 44736, 'steps': 232, 'loss/train': 2.0160496830940247} 01/28/2022 13:14:17 - INFO - codeparrot_training - Step 233: {'lr': 5.8250000000000006e-05, 'samples': 44928, 'steps': 233, 'loss/train': 8.470110654830933} 01/28/2022 13:14:23 - INFO - codeparrot_training - Step 234: {'lr': 5.8500000000000006e-05, 'samples': 45120, 'steps': 234, 'loss/train': 0.5271517932415009} 01/28/2022 13:14:27 - INFO - codeparrot_training - Step 235: {'lr': 5.875e-05, 'samples': 45312, 'steps': 235, 'loss/train': 1.8206296563148499} 01/28/2022 13:14:31 - INFO - codeparrot_training - Step 236: {'lr': 5.9e-05, 'samples': 45504, 'steps': 236, 'loss/train': 1.8314504027366638} 01/28/2022 13:14:35 - INFO - codeparrot_training - Step 237: {'lr': 5.925e-05, 'samples': 45696, 'steps': 237, 'loss/train': 2.513003647327423} 01/28/2022 13:14:39 - INFO - codeparrot_training - Step 238: {'lr': 5.9499999999999996e-05, 'samples': 45888, 'steps': 238, 'loss/train': 1.3541294038295746} 01/28/2022 13:14:44 - INFO - codeparrot_training - Step 239: {'lr': 5.9749999999999995e-05, 'samples': 46080, 'steps': 239, 'loss/train': 2.1684504747390747} 01/28/2022 13:14:49 - INFO - codeparrot_training - Step 240: {'lr': 6e-05, 'samples': 46272, 'steps': 240, 'loss/train': 3.3454166650772095} 01/28/2022 13:14:53 - INFO - codeparrot_training - Step 241: {'lr': 6.025e-05, 'samples': 46464, 'steps': 241, 'loss/train': 2.765377700328827} 01/28/2022 13:14:57 - INFO - codeparrot_training - Step 242: {'lr': 6.05e-05, 'samples': 46656, 'steps': 242, 'loss/train': 1.8727226257324219} 01/28/2022 13:15:01 - INFO - codeparrot_training - Step 243: {'lr': 6.075e-05, 'samples': 46848, 'steps': 243, 'loss/train': 2.726761758327484} 01/28/2022 13:15:06 - INFO - codeparrot_training - Step 244: {'lr': 6.1e-05, 'samples': 47040, 'steps': 244, 'loss/train': 1.147857517004013} 01/28/2022 13:15:10 - INFO - codeparrot_training - Step 245: {'lr': 6.125e-05, 'samples': 47232, 'steps': 245, 'loss/train': 3.1311804056167603} 01/28/2022 13:15:15 - INFO - codeparrot_training - Step 246: {'lr': 6.15e-05, 'samples': 47424, 'steps': 246, 'loss/train': 1.8552346229553223} 01/28/2022 13:15:19 - INFO - codeparrot_training - Step 247: {'lr': 6.175e-05, 'samples': 47616, 'steps': 247, 'loss/train': 2.0395506620407104} 01/28/2022 13:15:23 - INFO - codeparrot_training - Step 248: {'lr': 6.2e-05, 'samples': 47808, 'steps': 248, 'loss/train': 3.395920515060425} 01/28/2022 13:15:31 - INFO - codeparrot_training - Step 249: {'lr': 6.225e-05, 'samples': 48000, 'steps': 249, 'loss/train': 2.7890952229499817} 01/28/2022 13:15:35 - INFO - codeparrot_training - Step 250: {'lr': 6.25e-05, 'samples': 48192, 'steps': 250, 'loss/train': 2.481280267238617} 01/28/2022 13:15:39 - INFO - codeparrot_training - Step 251: {'lr': 6.275000000000001e-05, 'samples': 48384, 'steps': 251, 'loss/train': 1.9458904266357422} 01/28/2022 13:15:43 - INFO - codeparrot_training - Step 252: {'lr': 6.3e-05, 'samples': 48576, 'steps': 252, 'loss/train': 2.088134765625} 01/28/2022 13:15:47 - INFO - codeparrot_training - Step 253: {'lr': 6.325e-05, 'samples': 48768, 'steps': 253, 'loss/train': 2.216171443462372} 01/28/2022 13:15:52 - INFO - codeparrot_training - Step 254: {'lr': 6.35e-05, 'samples': 48960, 'steps': 254, 'loss/train': 1.2333862781524658} 01/28/2022 13:15:57 - INFO - codeparrot_training - Step 255: {'lr': 6.375e-05, 'samples': 49152, 'steps': 255, 'loss/train': 1.9478994011878967} 01/28/2022 13:16:01 - INFO - codeparrot_training - Step 256: {'lr': 6.4e-05, 'samples': 49344, 'steps': 256, 'loss/train': 1.372162789106369} 01/28/2022 13:16:05 - INFO - codeparrot_training - Step 257: {'lr': 6.425e-05, 'samples': 49536, 'steps': 257, 'loss/train': 2.602295994758606} 01/28/2022 13:16:09 - INFO - codeparrot_training - Step 258: {'lr': 6.450000000000001e-05, 'samples': 49728, 'steps': 258, 'loss/train': 2.4856991171836853} 01/28/2022 13:16:13 - INFO - codeparrot_training - Step 259: {'lr': 6.475e-05, 'samples': 49920, 'steps': 259, 'loss/train': 1.5183262825012207} 01/28/2022 13:16:19 - INFO - codeparrot_training - Step 260: {'lr': 6.500000000000001e-05, 'samples': 50112, 'steps': 260, 'loss/train': 0.7661848962306976} 01/28/2022 13:16:23 - INFO - codeparrot_training - Step 261: {'lr': 6.525e-05, 'samples': 50304, 'steps': 261, 'loss/train': 3.3317567110061646} 01/28/2022 13:16:27 - INFO - codeparrot_training - Step 262: {'lr': 6.55e-05, 'samples': 50496, 'steps': 262, 'loss/train': 1.8037270903587341} 01/28/2022 13:16:31 - INFO - codeparrot_training - Step 263: {'lr': 6.575e-05, 'samples': 50688, 'steps': 263, 'loss/train': 2.3626500964164734} 01/28/2022 13:16:35 - INFO - codeparrot_training - Step 264: {'lr': 6.6e-05, 'samples': 50880, 'steps': 264, 'loss/train': 2.5351778268814087} 01/28/2022 13:16:42 - INFO - codeparrot_training - Step 265: {'lr': 6.625000000000001e-05, 'samples': 51072, 'steps': 265, 'loss/train': 3.137581944465637} 01/28/2022 13:16:47 - INFO - codeparrot_training - Step 266: {'lr': 6.65e-05, 'samples': 51264, 'steps': 266, 'loss/train': 1.8716724514961243} 01/28/2022 13:16:51 - INFO - codeparrot_training - Step 267: {'lr': 6.675000000000001e-05, 'samples': 51456, 'steps': 267, 'loss/train': 1.8017098903656006} 01/28/2022 13:16:55 - INFO - codeparrot_training - Step 268: {'lr': 6.7e-05, 'samples': 51648, 'steps': 268, 'loss/train': 2.889967381954193} 01/28/2022 13:16:59 - INFO - codeparrot_training - Step 269: {'lr': 6.725000000000001e-05, 'samples': 51840, 'steps': 269, 'loss/train': 1.7229793667793274} 01/28/2022 13:17:04 - INFO - codeparrot_training - Step 270: {'lr': 6.75e-05, 'samples': 52032, 'steps': 270, 'loss/train': 1.4637369811534882} 01/28/2022 13:17:09 - INFO - codeparrot_training - Step 271: {'lr': 6.775000000000001e-05, 'samples': 52224, 'steps': 271, 'loss/train': 2.8115025758743286} 01/28/2022 13:17:13 - INFO - codeparrot_training - Step 272: {'lr': 6.800000000000001e-05, 'samples': 52416, 'steps': 272, 'loss/train': 2.779033899307251} 01/28/2022 13:17:17 - INFO - codeparrot_training - Step 273: {'lr': 6.825e-05, 'samples': 52608, 'steps': 273, 'loss/train': 1.7810704708099365} 01/28/2022 13:17:21 - INFO - codeparrot_training - Step 274: {'lr': 6.850000000000001e-05, 'samples': 52800, 'steps': 274, 'loss/train': 4.060012578964233} 01/28/2022 13:17:27 - INFO - codeparrot_training - Step 275: {'lr': 6.875e-05, 'samples': 52992, 'steps': 275, 'loss/train': 1.841540515422821} 01/28/2022 13:17:32 - INFO - codeparrot_training - Step 276: {'lr': 6.900000000000001e-05, 'samples': 53184, 'steps': 276, 'loss/train': 2.582349121570587} 01/28/2022 13:17:36 - INFO - codeparrot_training - Step 277: {'lr': 6.925e-05, 'samples': 53376, 'steps': 277, 'loss/train': 1.1392242014408112} 01/28/2022 13:17:40 - INFO - codeparrot_training - Step 278: {'lr': 6.950000000000001e-05, 'samples': 53568, 'steps': 278, 'loss/train': 0.390435129404068} 01/28/2022 13:17:44 - INFO - codeparrot_training - Step 279: {'lr': 6.975e-05, 'samples': 53760, 'steps': 279, 'loss/train': 1.2815892398357391} 01/28/2022 13:17:50 - INFO - codeparrot_training - Step 280: {'lr': 7.000000000000001e-05, 'samples': 53952, 'steps': 280, 'loss/train': 1.810163676738739} 01/28/2022 13:17:54 - INFO - codeparrot_training - Step 281: {'lr': 7.025000000000001e-05, 'samples': 54144, 'steps': 281, 'loss/train': 2.7389538288116455} 01/28/2022 13:17:58 - INFO - codeparrot_training - Step 282: {'lr': 7.049999999999999e-05, 'samples': 54336, 'steps': 282, 'loss/train': 2.225660026073456} 01/28/2022 13:18:02 - INFO - codeparrot_training - Step 283: {'lr': 7.075e-05, 'samples': 54528, 'steps': 283, 'loss/train': 1.847003996372223} 01/28/2022 13:18:06 - INFO - codeparrot_training - Step 284: {'lr': 7.099999999999999e-05, 'samples': 54720, 'steps': 284, 'loss/train': 2.1737850308418274} 01/28/2022 13:18:11 - INFO - codeparrot_training - Step 285: {'lr': 7.125e-05, 'samples': 54912, 'steps': 285, 'loss/train': 2.213922679424286} 01/28/2022 13:18:15 - INFO - codeparrot_training - Step 286: {'lr': 7.149999999999999e-05, 'samples': 55104, 'steps': 286, 'loss/train': 2.5635951161384583} 01/28/2022 13:18:20 - INFO - codeparrot_training - Step 287: {'lr': 7.175e-05, 'samples': 55296, 'steps': 287, 'loss/train': 2.2450772523880005} 01/28/2022 13:18:24 - INFO - codeparrot_training - Step 288: {'lr': 7.2e-05, 'samples': 55488, 'steps': 288, 'loss/train': 3.221435308456421} 01/28/2022 13:18:28 - INFO - codeparrot_training - Step 289: {'lr': 7.225e-05, 'samples': 55680, 'steps': 289, 'loss/train': 2.281312108039856} 01/28/2022 13:18:34 - INFO - codeparrot_training - Step 290: {'lr': 7.25e-05, 'samples': 55872, 'steps': 290, 'loss/train': 2.4072545170783997} 01/28/2022 13:18:38 - INFO - codeparrot_training - Step 291: {'lr': 7.274999999999999e-05, 'samples': 56064, 'steps': 291, 'loss/train': 1.6557587385177612} 01/28/2022 13:18:42 - INFO - codeparrot_training - Step 292: {'lr': 7.3e-05, 'samples': 56256, 'steps': 292, 'loss/train': 2.3484347462654114} 01/28/2022 13:18:46 - INFO - codeparrot_training - Step 293: {'lr': 7.324999999999999e-05, 'samples': 56448, 'steps': 293, 'loss/train': 2.9775270223617554} 01/28/2022 13:18:50 - INFO - codeparrot_training - Step 294: {'lr': 7.35e-05, 'samples': 56640, 'steps': 294, 'loss/train': 1.1433037519454956} 01/28/2022 13:18:58 - INFO - codeparrot_training - Step 295: {'lr': 7.375e-05, 'samples': 56832, 'steps': 295, 'loss/train': 1.8574823141098022} 01/28/2022 13:19:02 - INFO - codeparrot_training - Step 296: {'lr': 7.4e-05, 'samples': 57024, 'steps': 296, 'loss/train': 1.7895770072937012} 01/28/2022 13:19:06 - INFO - codeparrot_training - Step 297: {'lr': 7.425e-05, 'samples': 57216, 'steps': 297, 'loss/train': 2.022561728954315} 01/28/2022 13:19:10 - INFO - codeparrot_training - Step 298: {'lr': 7.45e-05, 'samples': 57408, 'steps': 298, 'loss/train': 2.517099916934967} 01/28/2022 13:19:14 - INFO - codeparrot_training - Step 299: {'lr': 7.475e-05, 'samples': 57600, 'steps': 299, 'loss/train': 2.313974440097809} 01/28/2022 13:19:19 - INFO - codeparrot_training - Step 300: {'lr': 7.5e-05, 'samples': 57792, 'steps': 300, 'loss/train': 2.4208420515060425} 01/28/2022 13:19:24 - INFO - codeparrot_training - Step 301: {'lr': 7.525e-05, 'samples': 57984, 'steps': 301, 'loss/train': 2.204310357570648} 01/28/2022 13:19:28 - INFO - codeparrot_training - Step 302: {'lr': 7.55e-05, 'samples': 58176, 'steps': 302, 'loss/train': 0.5202314257621765} 01/28/2022 13:19:32 - INFO - codeparrot_training - Step 303: {'lr': 7.575e-05, 'samples': 58368, 'steps': 303, 'loss/train': 2.4920074939727783} 01/28/2022 13:19:36 - INFO - codeparrot_training - Step 304: {'lr': 7.6e-05, 'samples': 58560, 'steps': 304, 'loss/train': 2.7514582872390747} 01/28/2022 13:19:41 - INFO - codeparrot_training - Step 305: {'lr': 7.625e-05, 'samples': 58752, 'steps': 305, 'loss/train': 2.034381866455078} 01/28/2022 13:19:45 - INFO - codeparrot_training - Step 306: {'lr': 7.65e-05, 'samples': 58944, 'steps': 306, 'loss/train': 1.742581307888031} 01/28/2022 13:19:49 - INFO - codeparrot_training - Step 307: {'lr': 7.675e-05, 'samples': 59136, 'steps': 307, 'loss/train': 2.4856508374214172} 01/28/2022 13:19:54 - INFO - codeparrot_training - Step 308: {'lr': 7.7e-05, 'samples': 59328, 'steps': 308, 'loss/train': 1.2047097086906433} 01/28/2022 13:19:58 - INFO - codeparrot_training - Step 309: {'lr': 7.725000000000001e-05, 'samples': 59520, 'steps': 309, 'loss/train': 1.5543610453605652} 01/28/2022 13:20:05 - INFO - codeparrot_training - Step 310: {'lr': 7.75e-05, 'samples': 59712, 'steps': 310, 'loss/train': 1.5788565874099731} 01/28/2022 13:20:09 - INFO - codeparrot_training - Step 311: {'lr': 7.775e-05, 'samples': 59904, 'steps': 311, 'loss/train': 1.8161746859550476} 01/28/2022 13:20:14 - INFO - codeparrot_training - Step 312: {'lr': 7.8e-05, 'samples': 60096, 'steps': 312, 'loss/train': 0.7483754754066467} 01/28/2022 13:20:18 - INFO - codeparrot_training - Step 313: {'lr': 7.825e-05, 'samples': 60288, 'steps': 313, 'loss/train': 2.3497719168663025} 01/28/2022 13:20:22 - INFO - codeparrot_training - Step 314: {'lr': 7.85e-05, 'samples': 60480, 'steps': 314, 'loss/train': 3.0754473209381104} 01/28/2022 13:20:27 - INFO - codeparrot_training - Step 315: {'lr': 7.875e-05, 'samples': 60672, 'steps': 315, 'loss/train': 1.910196840763092} 01/28/2022 13:20:31 - INFO - codeparrot_training - Step 316: {'lr': 7.9e-05, 'samples': 60864, 'steps': 316, 'loss/train': 2.241649031639099} 01/28/2022 13:20:35 - INFO - codeparrot_training - Step 317: {'lr': 7.925e-05, 'samples': 61056, 'steps': 317, 'loss/train': 3.8199355602264404} 01/28/2022 13:20:40 - INFO - codeparrot_training - Step 318: {'lr': 7.950000000000001e-05, 'samples': 61248, 'steps': 318, 'loss/train': 2.467266619205475} 01/28/2022 13:20:44 - INFO - codeparrot_training - Step 319: {'lr': 7.975e-05, 'samples': 61440, 'steps': 319, 'loss/train': 2.990099608898163} 01/28/2022 13:20:51 - INFO - codeparrot_training - Step 320: {'lr': 8e-05, 'samples': 61632, 'steps': 320, 'loss/train': 3.0172476768493652} 01/28/2022 13:20:55 - INFO - codeparrot_training - Step 321: {'lr': 8.025e-05, 'samples': 61824, 'steps': 321, 'loss/train': 1.8201200366020203} 01/28/2022 13:20:59 - INFO - codeparrot_training - Step 322: {'lr': 8.05e-05, 'samples': 62016, 'steps': 322, 'loss/train': 2.400550067424774} 01/28/2022 13:21:03 - INFO - codeparrot_training - Step 323: {'lr': 8.075e-05, 'samples': 62208, 'steps': 323, 'loss/train': 2.031210958957672} 01/28/2022 13:21:07 - INFO - codeparrot_training - Step 324: {'lr': 8.1e-05, 'samples': 62400, 'steps': 324, 'loss/train': 2.509629964828491} 01/28/2022 13:21:12 - INFO - codeparrot_training - Step 325: {'lr': 8.125000000000001e-05, 'samples': 62592, 'steps': 325, 'loss/train': 1.1312372088432312} 01/28/2022 13:21:17 - INFO - codeparrot_training - Step 326: {'lr': 8.15e-05, 'samples': 62784, 'steps': 326, 'loss/train': 1.8903037905693054} 01/28/2022 13:21:21 - INFO - codeparrot_training - Step 327: {'lr': 8.175000000000001e-05, 'samples': 62976, 'steps': 327, 'loss/train': 1.9436261057853699} 01/28/2022 13:21:25 - INFO - codeparrot_training - Step 328: {'lr': 8.2e-05, 'samples': 63168, 'steps': 328, 'loss/train': 0.6482186615467072} 01/28/2022 13:21:29 - INFO - codeparrot_training - Step 329: {'lr': 8.225000000000001e-05, 'samples': 63360, 'steps': 329, 'loss/train': 2.2017683386802673} 01/28/2022 13:21:35 - INFO - codeparrot_training - Step 330: {'lr': 8.25e-05, 'samples': 63552, 'steps': 330, 'loss/train': 2.331234097480774} 01/28/2022 13:21:39 - INFO - codeparrot_training - Step 331: {'lr': 8.275e-05, 'samples': 63744, 'steps': 331, 'loss/train': 2.096129894256592} 01/28/2022 13:21:43 - INFO - codeparrot_training - Step 332: {'lr': 8.300000000000001e-05, 'samples': 63936, 'steps': 332, 'loss/train': 2.484592616558075} 01/28/2022 13:21:47 - INFO - codeparrot_training - Step 333: {'lr': 8.325e-05, 'samples': 64128, 'steps': 333, 'loss/train': 1.504553496837616} 01/28/2022 13:21:51 - INFO - codeparrot_training - Step 334: {'lr': 8.350000000000001e-05, 'samples': 64320, 'steps': 334, 'loss/train': 1.5855562090873718} 01/28/2022 13:21:57 - INFO - codeparrot_training - Step 335: {'lr': 8.375e-05, 'samples': 64512, 'steps': 335, 'loss/train': 1.8586195707321167} 01/28/2022 13:22:01 - INFO - codeparrot_training - Step 336: {'lr': 8.400000000000001e-05, 'samples': 64704, 'steps': 336, 'loss/train': 2.4626998901367188} 01/28/2022 13:22:05 - INFO - codeparrot_training - Step 337: {'lr': 8.425e-05, 'samples': 64896, 'steps': 337, 'loss/train': 2.59538197517395} 01/28/2022 13:22:09 - INFO - codeparrot_training - Step 338: {'lr': 8.450000000000001e-05, 'samples': 65088, 'steps': 338, 'loss/train': 1.9972534775733948} 01/28/2022 13:22:13 - INFO - codeparrot_training - Step 339: {'lr': 8.475000000000001e-05, 'samples': 65280, 'steps': 339, 'loss/train': 1.6548653841018677} 01/28/2022 13:22:19 - INFO - codeparrot_training - Step 340: {'lr': 8.5e-05, 'samples': 65472, 'steps': 340, 'loss/train': 1.9769997596740723} 01/28/2022 13:22:23 - INFO - codeparrot_training - Step 341: {'lr': 8.525000000000001e-05, 'samples': 65664, 'steps': 341, 'loss/train': 2.2398714423179626} 01/28/2022 13:22:28 - INFO - codeparrot_training - Step 342: {'lr': 8.55e-05, 'samples': 65856, 'steps': 342, 'loss/train': 2.134919285774231} 01/28/2022 13:22:32 - INFO - codeparrot_training - Step 343: {'lr': 8.575000000000001e-05, 'samples': 66048, 'steps': 343, 'loss/train': 3.1586272716522217} 01/28/2022 13:22:36 - INFO - codeparrot_training - Step 344: {'lr': 8.599999999999999e-05, 'samples': 66240, 'steps': 344, 'loss/train': 2.379559278488159} 01/28/2022 13:22:41 - INFO - codeparrot_training - Step 345: {'lr': 8.625e-05, 'samples': 66432, 'steps': 345, 'loss/train': 2.109428286552429} 01/28/2022 13:22:45 - INFO - codeparrot_training - Step 346: {'lr': 8.65e-05, 'samples': 66624, 'steps': 346, 'loss/train': 1.7040767669677734} 01/28/2022 13:22:49 - INFO - codeparrot_training - Step 347: {'lr': 8.675e-05, 'samples': 66816, 'steps': 347, 'loss/train': 2.241216480731964} 01/28/2022 13:22:53 - INFO - codeparrot_training - Step 348: {'lr': 8.7e-05, 'samples': 67008, 'steps': 348, 'loss/train': 2.0094056725502014} 01/28/2022 13:22:58 - INFO - codeparrot_training - Step 349: {'lr': 8.724999999999999e-05, 'samples': 67200, 'steps': 349, 'loss/train': 2.4556962847709656} 01/28/2022 13:23:03 - INFO - codeparrot_training - Step 350: {'lr': 8.75e-05, 'samples': 67392, 'steps': 350, 'loss/train': 2.59530508518219} 01/28/2022 13:23:07 - INFO - codeparrot_training - Step 351: {'lr': 8.774999999999999e-05, 'samples': 67584, 'steps': 351, 'loss/train': 2.623638331890106} 01/28/2022 13:23:11 - INFO - codeparrot_training - Step 352: {'lr': 8.8e-05, 'samples': 67776, 'steps': 352, 'loss/train': 1.6449168920516968} 01/28/2022 13:23:15 - INFO - codeparrot_training - Step 353: {'lr': 8.824999999999999e-05, 'samples': 67968, 'steps': 353, 'loss/train': 2.1416271328926086} 01/28/2022 13:23:19 - INFO - codeparrot_training - Step 354: {'lr': 8.85e-05, 'samples': 68160, 'steps': 354, 'loss/train': 2.118185520172119} 01/28/2022 13:23:26 - INFO - codeparrot_training - Step 355: {'lr': 8.875e-05, 'samples': 68352, 'steps': 355, 'loss/train': 2.10310560464859} 01/28/2022 13:23:31 - INFO - codeparrot_training - Step 356: {'lr': 8.9e-05, 'samples': 68544, 'steps': 356, 'loss/train': 2.7734389901161194} 01/28/2022 13:23:35 - INFO - codeparrot_training - Step 357: {'lr': 8.925e-05, 'samples': 68736, 'steps': 357, 'loss/train': 2.062049925327301} 01/28/2022 13:23:39 - INFO - codeparrot_training - Step 358: {'lr': 8.95e-05, 'samples': 68928, 'steps': 358, 'loss/train': 3.362984776496887} 01/28/2022 13:23:43 - INFO - codeparrot_training - Step 359: {'lr': 8.975e-05, 'samples': 69120, 'steps': 359, 'loss/train': 2.5164751410484314} 01/28/2022 13:23:47 - INFO - codeparrot_training - Step 360: {'lr': 8.999999999999999e-05, 'samples': 69312, 'steps': 360, 'loss/train': 2.17100590467453} 01/28/2022 13:23:52 - INFO - codeparrot_training - Step 361: {'lr': 9.025e-05, 'samples': 69504, 'steps': 361, 'loss/train': 2.383537173271179} 01/28/2022 13:23:56 - INFO - codeparrot_training - Step 362: {'lr': 9.05e-05, 'samples': 69696, 'steps': 362, 'loss/train': 2.3970805406570435} 01/28/2022 13:24:01 - INFO - codeparrot_training - Step 363: {'lr': 9.075e-05, 'samples': 69888, 'steps': 363, 'loss/train': 2.8085650205612183} 01/28/2022 13:24:05 - INFO - codeparrot_training - Step 364: {'lr': 9.1e-05, 'samples': 70080, 'steps': 364, 'loss/train': 0.9615865051746368} 01/28/2022 13:24:09 - INFO - codeparrot_training - Step 365: {'lr': 9.125e-05, 'samples': 70272, 'steps': 365, 'loss/train': 1.9183196425437927} 01/28/2022 13:24:16 - INFO - codeparrot_training - Step 366: {'lr': 9.15e-05, 'samples': 70464, 'steps': 366, 'loss/train': 2.3761155009269714} 01/28/2022 13:24:21 - INFO - codeparrot_training - Step 367: {'lr': 9.175e-05, 'samples': 70656, 'steps': 367, 'loss/train': 1.073426753282547} 01/28/2022 13:24:25 - INFO - codeparrot_training - Step 368: {'lr': 9.2e-05, 'samples': 70848, 'steps': 368, 'loss/train': 1.969973087310791} 01/28/2022 13:24:29 - INFO - codeparrot_training - Step 369: {'lr': 9.225e-05, 'samples': 71040, 'steps': 369, 'loss/train': 2.59987735748291} 01/28/2022 13:24:34 - INFO - codeparrot_training - Step 370: {'lr': 9.25e-05, 'samples': 71232, 'steps': 370, 'loss/train': 1.669056236743927} 01/28/2022 13:24:38 - INFO - codeparrot_training - Step 371: {'lr': 9.275e-05, 'samples': 71424, 'steps': 371, 'loss/train': 1.404313713312149} 01/28/2022 13:24:42 - INFO - codeparrot_training - Step 372: {'lr': 9.3e-05, 'samples': 71616, 'steps': 372, 'loss/train': 2.7956053018569946} 01/28/2022 13:24:47 - INFO - codeparrot_training - Step 373: {'lr': 9.325e-05, 'samples': 71808, 'steps': 373, 'loss/train': 5.44849705696106} 01/28/2022 13:24:51 - INFO - codeparrot_training - Step 374: {'lr': 9.35e-05, 'samples': 72000, 'steps': 374, 'loss/train': 8.09075117111206} 01/28/2022 13:24:55 - INFO - codeparrot_training - Step 375: {'lr': 9.375e-05, 'samples': 72192, 'steps': 375, 'loss/train': 0.4709901362657547} 01/28/2022 13:25:00 - INFO - codeparrot_training - Step 376: {'lr': 9.400000000000001e-05, 'samples': 72384, 'steps': 376, 'loss/train': 2.5553566217422485} 01/28/2022 13:25:04 - INFO - codeparrot_training - Step 377: {'lr': 9.425e-05, 'samples': 72576, 'steps': 377, 'loss/train': 1.5268546342849731} 01/28/2022 13:25:08 - INFO - codeparrot_training - Step 378: {'lr': 9.45e-05, 'samples': 72768, 'steps': 378, 'loss/train': 1.832252025604248} 01/28/2022 13:25:12 - INFO - codeparrot_training - Step 379: {'lr': 9.475e-05, 'samples': 72960, 'steps': 379, 'loss/train': 2.602208733558655} 01/28/2022 13:25:19 - INFO - codeparrot_training - Step 380: {'lr': 9.5e-05, 'samples': 73152, 'steps': 380, 'loss/train': 2.220587432384491} 01/28/2022 13:25:23 - INFO - codeparrot_training - Step 381: {'lr': 9.525e-05, 'samples': 73344, 'steps': 381, 'loss/train': 2.0126200318336487} 01/28/2022 13:25:27 - INFO - codeparrot_training - Step 382: {'lr': 9.55e-05, 'samples': 73536, 'steps': 382, 'loss/train': 2.6490983963012695} 01/28/2022 13:25:31 - INFO - codeparrot_training - Step 383: {'lr': 9.575000000000001e-05, 'samples': 73728, 'steps': 383, 'loss/train': 1.9188519716262817} 01/28/2022 13:25:36 - INFO - codeparrot_training - Step 384: {'lr': 9.6e-05, 'samples': 73920, 'steps': 384, 'loss/train': 2.1151451468467712} 01/28/2022 13:25:41 - INFO - codeparrot_training - Step 385: {'lr': 9.625000000000001e-05, 'samples': 74112, 'steps': 385, 'loss/train': 2.536947548389435} 01/28/2022 13:25:45 - INFO - codeparrot_training - Step 386: {'lr': 9.65e-05, 'samples': 74304, 'steps': 386, 'loss/train': 2.7067641019821167} 01/28/2022 13:25:49 - INFO - codeparrot_training - Step 387: {'lr': 9.675000000000001e-05, 'samples': 74496, 'steps': 387, 'loss/train': 3.1732935905456543} 01/28/2022 13:25:53 - INFO - codeparrot_training - Step 388: {'lr': 9.7e-05, 'samples': 74688, 'steps': 388, 'loss/train': 2.006965756416321} 01/28/2022 13:25:58 - INFO - codeparrot_training - Step 389: {'lr': 9.725e-05, 'samples': 74880, 'steps': 389, 'loss/train': 1.8531373143196106} 01/28/2022 13:26:03 - INFO - codeparrot_training - Step 390: {'lr': 9.750000000000001e-05, 'samples': 75072, 'steps': 390, 'loss/train': 3.0755585432052612} 01/28/2022 13:26:07 - INFO - codeparrot_training - Step 391: {'lr': 9.775e-05, 'samples': 75264, 'steps': 391, 'loss/train': 2.0659557580947876} 01/28/2022 13:26:11 - INFO - codeparrot_training - Step 392: {'lr': 9.800000000000001e-05, 'samples': 75456, 'steps': 392, 'loss/train': 1.14596226811409} 01/28/2022 13:26:15 - INFO - codeparrot_training - Step 393: {'lr': 9.825e-05, 'samples': 75648, 'steps': 393, 'loss/train': 2.5506362915039062} 01/28/2022 13:26:19 - INFO - codeparrot_training - Step 394: {'lr': 9.850000000000001e-05, 'samples': 75840, 'steps': 394, 'loss/train': 2.6055253744125366} 01/28/2022 13:26:26 - INFO - codeparrot_training - Step 395: {'lr': 9.875e-05, 'samples': 76032, 'steps': 395, 'loss/train': 2.000922918319702} 01/28/2022 13:26:30 - INFO - codeparrot_training - Step 396: {'lr': 9.900000000000001e-05, 'samples': 76224, 'steps': 396, 'loss/train': 1.0836397111415863} 01/28/2022 13:26:34 - INFO - codeparrot_training - Step 397: {'lr': 9.925000000000001e-05, 'samples': 76416, 'steps': 397, 'loss/train': 2.526535749435425} 01/28/2022 13:26:38 - INFO - codeparrot_training - Step 398: {'lr': 9.95e-05, 'samples': 76608, 'steps': 398, 'loss/train': 2.303645968437195} 01/28/2022 13:26:42 - INFO - codeparrot_training - Step 399: {'lr': 9.975000000000001e-05, 'samples': 76800, 'steps': 399, 'loss/train': 1.985789179801941} 01/28/2022 13:26:48 - INFO - codeparrot_training - Step 400: {'lr': 0.0001, 'samples': 76992, 'steps': 400, 'loss/train': 1.34466353058815} 01/28/2022 13:26:52 - INFO - codeparrot_training - Step 401: {'lr': 0.00010025000000000001, 'samples': 77184, 'steps': 401, 'loss/train': 2.226090431213379} 01/28/2022 13:26:56 - INFO - codeparrot_training - Step 402: {'lr': 0.0001005, 'samples': 77376, 'steps': 402, 'loss/train': 1.430657297372818} 01/28/2022 13:27:00 - INFO - codeparrot_training - Step 403: {'lr': 0.00010075000000000001, 'samples': 77568, 'steps': 403, 'loss/train': 2.745595693588257} 01/28/2022 13:27:04 - INFO - codeparrot_training - Step 404: {'lr': 0.000101, 'samples': 77760, 'steps': 404, 'loss/train': 1.080320656299591} 01/28/2022 13:27:10 - INFO - codeparrot_training - Step 405: {'lr': 0.00010125000000000001, 'samples': 77952, 'steps': 405, 'loss/train': 2.3354228138923645} 01/28/2022 13:27:14 - INFO - codeparrot_training - Step 406: {'lr': 0.00010150000000000001, 'samples': 78144, 'steps': 406, 'loss/train': 2.291370928287506} 01/28/2022 13:27:18 - INFO - codeparrot_training - Step 407: {'lr': 0.00010174999999999999, 'samples': 78336, 'steps': 407, 'loss/train': 1.7008516788482666} 01/28/2022 13:27:22 - INFO - codeparrot_training - Step 408: {'lr': 0.000102, 'samples': 78528, 'steps': 408, 'loss/train': 2.51967054605484} 01/28/2022 13:27:26 - INFO - codeparrot_training - Step 409: {'lr': 0.00010224999999999999, 'samples': 78720, 'steps': 409, 'loss/train': 2.947802245616913} 01/28/2022 13:27:31 - INFO - codeparrot_training - Step 410: {'lr': 0.0001025, 'samples': 78912, 'steps': 410, 'loss/train': 2.009278178215027} 01/28/2022 13:27:35 - INFO - codeparrot_training - Step 411: {'lr': 0.00010274999999999999, 'samples': 79104, 'steps': 411, 'loss/train': 2.323540449142456} 01/28/2022 13:27:39 - INFO - codeparrot_training - Step 412: {'lr': 0.000103, 'samples': 79296, 'steps': 412, 'loss/train': 1.9923208951950073} 01/28/2022 13:27:44 - INFO - codeparrot_training - Step 413: {'lr': 0.00010325, 'samples': 79488, 'steps': 413, 'loss/train': 1.8653483390808105} 01/28/2022 13:27:48 - INFO - codeparrot_training - Step 414: {'lr': 0.0001035, 'samples': 79680, 'steps': 414, 'loss/train': 2.002540647983551} 01/28/2022 13:27:55 - INFO - codeparrot_training - Step 415: {'lr': 0.00010375, 'samples': 79872, 'steps': 415, 'loss/train': 2.548307418823242} 01/28/2022 13:27:59 - INFO - codeparrot_training - Step 416: {'lr': 0.000104, 'samples': 80064, 'steps': 416, 'loss/train': 0.30849865078926086} 01/28/2022 13:28:03 - INFO - codeparrot_training - Step 417: {'lr': 0.00010425, 'samples': 80256, 'steps': 417, 'loss/train': 1.1098234355449677} 01/28/2022 13:28:07 - INFO - codeparrot_training - Step 418: {'lr': 0.00010449999999999999, 'samples': 80448, 'steps': 418, 'loss/train': 2.236936390399933} 01/28/2022 13:28:11 - INFO - codeparrot_training - Step 419: {'lr': 0.00010475, 'samples': 80640, 'steps': 419, 'loss/train': 1.1480836272239685} 01/28/2022 13:28:15 - INFO - codeparrot_training - Step 420: {'lr': 0.000105, 'samples': 80832, 'steps': 420, 'loss/train': 3.086016297340393} 01/28/2022 13:28:21 - INFO - codeparrot_training - Step 421: {'lr': 0.00010525, 'samples': 81024, 'steps': 421, 'loss/train': 2.5556384325027466} 01/28/2022 13:28:25 - INFO - codeparrot_training - Step 422: {'lr': 0.0001055, 'samples': 81216, 'steps': 422, 'loss/train': 2.022362172603607} 01/28/2022 13:28:29 - INFO - codeparrot_training - Step 423: {'lr': 0.00010575, 'samples': 81408, 'steps': 423, 'loss/train': 4.700298428535461} 01/28/2022 13:28:33 - INFO - codeparrot_training - Step 424: {'lr': 0.000106, 'samples': 81600, 'steps': 424, 'loss/train': 1.3899056017398834} 01/28/2022 13:28:40 - INFO - codeparrot_training - Step 425: {'lr': 0.00010625, 'samples': 81792, 'steps': 425, 'loss/train': 2.117711842060089} 01/28/2022 13:28:44 - INFO - codeparrot_training - Step 426: {'lr': 0.0001065, 'samples': 81984, 'steps': 426, 'loss/train': 2.339418590068817} 01/28/2022 13:28:48 - INFO - codeparrot_training - Step 427: {'lr': 0.00010675, 'samples': 82176, 'steps': 427, 'loss/train': 1.9530190229415894} 01/28/2022 13:28:52 - INFO - codeparrot_training - Step 428: {'lr': 0.000107, 'samples': 82368, 'steps': 428, 'loss/train': 2.842305064201355} 01/28/2022 13:28:56 - INFO - codeparrot_training - Step 429: {'lr': 0.00010725, 'samples': 82560, 'steps': 429, 'loss/train': 0.9067325592041016} 01/28/2022 13:29:01 - INFO - codeparrot_training - Step 430: {'lr': 0.0001075, 'samples': 82752, 'steps': 430, 'loss/train': 2.254849433898926} 01/28/2022 13:29:06 - INFO - codeparrot_training - Step 431: {'lr': 0.00010775, 'samples': 82944, 'steps': 431, 'loss/train': 1.9067448377609253} 01/28/2022 13:29:10 - INFO - codeparrot_training - Step 432: {'lr': 0.000108, 'samples': 83136, 'steps': 432, 'loss/train': 1.7269633412361145} 01/28/2022 13:29:14 - INFO - codeparrot_training - Step 433: {'lr': 0.00010825, 'samples': 83328, 'steps': 433, 'loss/train': 2.1003241539001465} 01/28/2022 13:29:18 - INFO - codeparrot_training - Step 434: {'lr': 0.00010850000000000001, 'samples': 83520, 'steps': 434, 'loss/train': 2.33688747882843} 01/28/2022 13:29:23 - INFO - codeparrot_training - Step 435: {'lr': 0.00010875, 'samples': 83712, 'steps': 435, 'loss/train': 1.9798935055732727} 01/28/2022 13:29:27 - INFO - codeparrot_training - Step 436: {'lr': 0.000109, 'samples': 83904, 'steps': 436, 'loss/train': 2.5221866369247437} 01/28/2022 13:29:32 - INFO - codeparrot_training - Step 437: {'lr': 0.00010925, 'samples': 84096, 'steps': 437, 'loss/train': 2.6836801171302795} 01/28/2022 13:29:36 - INFO - codeparrot_training - Step 438: {'lr': 0.0001095, 'samples': 84288, 'steps': 438, 'loss/train': 2.5501572489738464} 01/28/2022 13:29:40 - INFO - codeparrot_training - Step 439: {'lr': 0.00010975, 'samples': 84480, 'steps': 439, 'loss/train': 1.7378119826316833} 01/28/2022 13:29:46 - INFO - codeparrot_training - Step 440: {'lr': 0.00011, 'samples': 84672, 'steps': 440, 'loss/train': 1.6253256797790527} 01/28/2022 13:29:50 - INFO - codeparrot_training - Step 441: {'lr': 0.00011025, 'samples': 84864, 'steps': 441, 'loss/train': 1.8613500595092773} 01/28/2022 13:29:54 - INFO - codeparrot_training - Step 442: {'lr': 0.0001105, 'samples': 85056, 'steps': 442, 'loss/train': 1.5432872772216797} 01/28/2022 13:29:58 - INFO - codeparrot_training - Step 443: {'lr': 0.00011075000000000001, 'samples': 85248, 'steps': 443, 'loss/train': 1.8352982997894287} 01/28/2022 13:30:03 - INFO - codeparrot_training - Step 444: {'lr': 0.000111, 'samples': 85440, 'steps': 444, 'loss/train': 1.8398144245147705} 01/28/2022 13:30:08 - INFO - codeparrot_training - Step 445: {'lr': 0.00011125000000000001, 'samples': 85632, 'steps': 445, 'loss/train': 2.1234737634658813} 01/28/2022 13:30:12 - INFO - codeparrot_training - Step 446: {'lr': 0.0001115, 'samples': 85824, 'steps': 446, 'loss/train': 2.040612280368805} 01/28/2022 13:30:16 - INFO - codeparrot_training - Step 447: {'lr': 0.00011175, 'samples': 86016, 'steps': 447, 'loss/train': 2.395890176296234} 01/28/2022 13:30:20 - INFO - codeparrot_training - Step 448: {'lr': 0.000112, 'samples': 86208, 'steps': 448, 'loss/train': 1.8500298857688904} 01/28/2022 13:30:24 - INFO - codeparrot_training - Step 449: {'lr': 0.00011225, 'samples': 86400, 'steps': 449, 'loss/train': 2.4372233748435974} 01/28/2022 13:30:31 - INFO - codeparrot_training - Step 450: {'lr': 0.00011250000000000001, 'samples': 86592, 'steps': 450, 'loss/train': 2.145418703556061} 01/28/2022 13:30:35 - INFO - codeparrot_training - Step 451: {'lr': 0.00011275, 'samples': 86784, 'steps': 451, 'loss/train': 2.6469987630844116} 01/28/2022 13:30:39 - INFO - codeparrot_training - Step 452: {'lr': 0.00011300000000000001, 'samples': 86976, 'steps': 452, 'loss/train': 1.5190070271492004} 01/28/2022 13:30:43 - INFO - codeparrot_training - Step 453: {'lr': 0.00011325, 'samples': 87168, 'steps': 453, 'loss/train': 1.7445648908615112} 01/28/2022 13:30:48 - INFO - codeparrot_training - Step 454: {'lr': 0.00011350000000000001, 'samples': 87360, 'steps': 454, 'loss/train': 2.5863465070724487} 01/28/2022 13:30:53 - INFO - codeparrot_training - Step 455: {'lr': 0.00011375, 'samples': 87552, 'steps': 455, 'loss/train': 3.1834391355514526} 01/28/2022 13:30:57 - INFO - codeparrot_training - Step 456: {'lr': 0.000114, 'samples': 87744, 'steps': 456, 'loss/train': 2.4723424315452576} 01/28/2022 13:31:01 - INFO - codeparrot_training - Step 457: {'lr': 0.00011425000000000001, 'samples': 87936, 'steps': 457, 'loss/train': 1.016595035791397} 01/28/2022 13:31:05 - INFO - codeparrot_training - Step 458: {'lr': 0.0001145, 'samples': 88128, 'steps': 458, 'loss/train': 1.9579001069068909} 01/28/2022 13:31:10 - INFO - codeparrot_training - Step 459: {'lr': 0.00011475000000000001, 'samples': 88320, 'steps': 459, 'loss/train': 1.9943790435791016} 01/28/2022 13:31:15 - INFO - codeparrot_training - Step 460: {'lr': 0.000115, 'samples': 88512, 'steps': 460, 'loss/train': 1.589141607284546} 01/28/2022 13:31:19 - INFO - codeparrot_training - Step 461: {'lr': 0.00011525000000000001, 'samples': 88704, 'steps': 461, 'loss/train': 1.9581166505813599} 01/28/2022 13:31:23 - INFO - codeparrot_training - Step 462: {'lr': 0.0001155, 'samples': 88896, 'steps': 462, 'loss/train': 2.523662567138672} 01/28/2022 13:31:27 - INFO - codeparrot_training - Step 463: {'lr': 0.00011575000000000001, 'samples': 89088, 'steps': 463, 'loss/train': 2.2411020398139954} 01/28/2022 13:31:31 - INFO - codeparrot_training - Step 464: {'lr': 0.00011600000000000001, 'samples': 89280, 'steps': 464, 'loss/train': 1.918980360031128} 01/28/2022 13:31:36 - INFO - codeparrot_training - Step 465: {'lr': 0.00011625, 'samples': 89472, 'steps': 465, 'loss/train': 2.039166569709778} 01/28/2022 13:31:41 - INFO - codeparrot_training - Step 466: {'lr': 0.00011650000000000001, 'samples': 89664, 'steps': 466, 'loss/train': 2.505170166492462} 01/28/2022 13:31:45 - INFO - codeparrot_training - Step 467: {'lr': 0.00011675, 'samples': 89856, 'steps': 467, 'loss/train': 1.4210465848445892} 01/28/2022 13:31:49 - INFO - codeparrot_training - Step 468: {'lr': 0.00011700000000000001, 'samples': 90048, 'steps': 468, 'loss/train': 2.003764808177948} 01/28/2022 13:31:53 - INFO - codeparrot_training - Step 469: {'lr': 0.00011724999999999999, 'samples': 90240, 'steps': 469, 'loss/train': 1.6631369590759277} 01/28/2022 13:32:00 - INFO - codeparrot_training - Step 470: {'lr': 0.0001175, 'samples': 90432, 'steps': 470, 'loss/train': 2.2279627919197083} 01/28/2022 13:32:04 - INFO - codeparrot_training - Step 471: {'lr': 0.00011775, 'samples': 90624, 'steps': 471, 'loss/train': 1.1695095598697662} 01/28/2022 13:32:08 - INFO - codeparrot_training - Step 472: {'lr': 0.000118, 'samples': 90816, 'steps': 472, 'loss/train': 1.9423595666885376} 01/28/2022 13:32:12 - INFO - codeparrot_training - Step 473: {'lr': 0.00011825, 'samples': 91008, 'steps': 473, 'loss/train': 1.1801507771015167} 01/28/2022 13:32:16 - INFO - codeparrot_training - Step 474: {'lr': 0.0001185, 'samples': 91200, 'steps': 474, 'loss/train': 1.6662245392799377} 01/28/2022 13:32:21 - INFO - codeparrot_training - Step 475: {'lr': 0.00011875, 'samples': 91392, 'steps': 475, 'loss/train': 2.1907097697257996} 01/28/2022 13:32:26 - INFO - codeparrot_training - Step 476: {'lr': 0.00011899999999999999, 'samples': 91584, 'steps': 476, 'loss/train': 2.5175941586494446} 01/28/2022 13:32:30 - INFO - codeparrot_training - Step 477: {'lr': 0.00011925, 'samples': 91776, 'steps': 477, 'loss/train': 0.7059400230646133} 01/28/2022 13:32:34 - INFO - codeparrot_training - Step 478: {'lr': 0.00011949999999999999, 'samples': 91968, 'steps': 478, 'loss/train': 2.5413649678230286} 01/28/2022 13:32:38 - INFO - codeparrot_training - Step 479: {'lr': 0.00011975, 'samples': 92160, 'steps': 479, 'loss/train': 2.275630474090576} 01/28/2022 13:32:43 - INFO - codeparrot_training - Step 480: {'lr': 0.00012, 'samples': 92352, 'steps': 480, 'loss/train': 1.866119384765625} 01/28/2022 13:32:47 - INFO - codeparrot_training - Step 481: {'lr': 0.00012025, 'samples': 92544, 'steps': 481, 'loss/train': 1.7266119718551636} 01/28/2022 13:32:52 - INFO - codeparrot_training - Step 482: {'lr': 0.0001205, 'samples': 92736, 'steps': 482, 'loss/train': 2.19597852230072} 01/28/2022 13:32:56 - INFO - codeparrot_training - Step 483: {'lr': 0.00012075, 'samples': 92928, 'steps': 483, 'loss/train': 2.622327446937561} 01/28/2022 13:33:00 - INFO - codeparrot_training - Step 484: {'lr': 0.000121, 'samples': 93120, 'steps': 484, 'loss/train': 2.2634063959121704} 01/28/2022 13:33:06 - INFO - codeparrot_training - Step 485: {'lr': 0.00012124999999999999, 'samples': 93312, 'steps': 485, 'loss/train': 2.522210955619812} 01/28/2022 13:33:10 - INFO - codeparrot_training - Step 486: {'lr': 0.0001215, 'samples': 93504, 'steps': 486, 'loss/train': 1.9755256175994873} 01/28/2022 13:33:14 - INFO - codeparrot_training - Step 487: {'lr': 0.00012175, 'samples': 93696, 'steps': 487, 'loss/train': 2.4433164596557617} 01/28/2022 13:33:19 - INFO - codeparrot_training - Step 488: {'lr': 0.000122, 'samples': 93888, 'steps': 488, 'loss/train': 1.7861722111701965} 01/28/2022 13:33:23 - INFO - codeparrot_training - Step 489: {'lr': 0.00012225, 'samples': 94080, 'steps': 489, 'loss/train': 2.7531570196151733} 01/28/2022 13:33:28 - INFO - codeparrot_training - Step 490: {'lr': 0.0001225, 'samples': 94272, 'steps': 490, 'loss/train': 2.5168349146842957} 01/28/2022 13:33:32 - INFO - codeparrot_training - Step 491: {'lr': 0.00012275, 'samples': 94464, 'steps': 491, 'loss/train': 2.293790817260742} 01/28/2022 13:33:36 - INFO - codeparrot_training - Step 492: {'lr': 0.000123, 'samples': 94656, 'steps': 492, 'loss/train': 2.551152527332306} 01/28/2022 13:33:40 - INFO - codeparrot_training - Step 493: {'lr': 0.00012325000000000001, 'samples': 94848, 'steps': 493, 'loss/train': 2.246843934059143} 01/28/2022 13:33:45 - INFO - codeparrot_training - Step 494: {'lr': 0.0001235, 'samples': 95040, 'steps': 494, 'loss/train': 1.8888823986053467} 01/28/2022 13:33:51 - INFO - codeparrot_training - Step 495: {'lr': 0.00012375, 'samples': 95232, 'steps': 495, 'loss/train': 1.067182570695877} 01/28/2022 13:33:55 - INFO - codeparrot_training - Step 496: {'lr': 0.000124, 'samples': 95424, 'steps': 496, 'loss/train': 1.6582483649253845} 01/28/2022 13:33:59 - INFO - codeparrot_training - Step 497: {'lr': 0.00012425, 'samples': 95616, 'steps': 497, 'loss/train': 2.5385130643844604} 01/28/2022 13:34:03 - INFO - codeparrot_training - Step 498: {'lr': 0.0001245, 'samples': 95808, 'steps': 498, 'loss/train': 2.833753287792206} 01/28/2022 13:34:08 - INFO - codeparrot_training - Step 499: {'lr': 0.00012475, 'samples': 96000, 'steps': 499, 'loss/train': 2.0532556772232056} 01/28/2022 13:34:13 - INFO - codeparrot_training - Step 500: {'lr': 0.000125, 'samples': 96192, 'steps': 500, 'loss/train': 2.557322144508362} 01/28/2022 13:34:17 - INFO - codeparrot_training - Step 501: {'lr': 0.00012525, 'samples': 96384, 'steps': 501, 'loss/train': 1.0035335719585419} 01/28/2022 13:34:21 - INFO - codeparrot_training - Step 502: {'lr': 0.00012550000000000001, 'samples': 96576, 'steps': 502, 'loss/train': 1.7553940415382385} 01/28/2022 13:34:25 - INFO - codeparrot_training - Step 503: {'lr': 0.00012575, 'samples': 96768, 'steps': 503, 'loss/train': 1.1455343663692474} 01/28/2022 13:34:29 - INFO - codeparrot_training - Step 504: {'lr': 0.000126, 'samples': 96960, 'steps': 504, 'loss/train': 1.9719264507293701} 01/28/2022 13:34:35 - INFO - codeparrot_training - Step 505: {'lr': 0.00012625, 'samples': 97152, 'steps': 505, 'loss/train': 2.4693276286125183} 01/28/2022 13:34:39 - INFO - codeparrot_training - Step 506: {'lr': 0.0001265, 'samples': 97344, 'steps': 506, 'loss/train': 2.0731043815612793} 01/28/2022 13:34:43 - INFO - codeparrot_training - Step 507: {'lr': 0.00012675, 'samples': 97536, 'steps': 507, 'loss/train': 1.5254591703414917} 01/28/2022 13:34:47 - INFO - codeparrot_training - Step 508: {'lr': 0.000127, 'samples': 97728, 'steps': 508, 'loss/train': 1.6457562446594238} 01/28/2022 13:34:51 - INFO - codeparrot_training - Step 509: {'lr': 0.00012725, 'samples': 97920, 'steps': 509, 'loss/train': 1.911628246307373} 01/28/2022 13:34:58 - INFO - codeparrot_training - Step 510: {'lr': 0.0001275, 'samples': 98112, 'steps': 510, 'loss/train': 1.9169179201126099} 01/28/2022 13:35:02 - INFO - codeparrot_training - Step 511: {'lr': 0.00012775000000000002, 'samples': 98304, 'steps': 511, 'loss/train': 1.7416259050369263} 01/28/2022 13:35:06 - INFO - codeparrot_training - Step 512: {'lr': 0.000128, 'samples': 98496, 'steps': 512, 'loss/train': 2.710155487060547} 01/28/2022 13:35:10 - INFO - codeparrot_training - Step 513: {'lr': 0.00012825, 'samples': 98688, 'steps': 513, 'loss/train': 1.9697399139404297} 01/28/2022 13:35:15 - INFO - codeparrot_training - Step 514: {'lr': 0.0001285, 'samples': 98880, 'steps': 514, 'loss/train': 1.9533600211143494} 01/28/2022 13:35:20 - INFO - codeparrot_training - Step 515: {'lr': 0.00012875, 'samples': 99072, 'steps': 515, 'loss/train': 2.605292558670044} 01/28/2022 13:35:24 - INFO - codeparrot_training - Step 516: {'lr': 0.00012900000000000002, 'samples': 99264, 'steps': 516, 'loss/train': 2.0502219200134277} 01/28/2022 13:35:28 - INFO - codeparrot_training - Step 517: {'lr': 0.00012925, 'samples': 99456, 'steps': 517, 'loss/train': 2.090924620628357} 01/28/2022 13:35:32 - INFO - codeparrot_training - Step 518: {'lr': 0.0001295, 'samples': 99648, 'steps': 518, 'loss/train': 1.0696645081043243} 01/28/2022 13:35:36 - INFO - codeparrot_training - Step 519: {'lr': 0.00012975, 'samples': 99840, 'steps': 519, 'loss/train': 1.9695006608963013} 01/28/2022 13:35:43 - INFO - codeparrot_training - Step 520: {'lr': 0.00013000000000000002, 'samples': 100032, 'steps': 520, 'loss/train': 2.400749623775482} 01/28/2022 13:35:47 - INFO - codeparrot_training - Step 521: {'lr': 0.00013025, 'samples': 100224, 'steps': 521, 'loss/train': 2.945523262023926} 01/28/2022 13:35:51 - INFO - codeparrot_training - Step 522: {'lr': 0.0001305, 'samples': 100416, 'steps': 522, 'loss/train': 1.7668529748916626} 01/28/2022 13:35:55 - INFO - codeparrot_training - Step 523: {'lr': 0.00013075, 'samples': 100608, 'steps': 523, 'loss/train': 2.4956894516944885} 01/28/2022 13:35:59 - INFO - codeparrot_training - Step 524: {'lr': 0.000131, 'samples': 100800, 'steps': 524, 'loss/train': 1.985107183456421} 01/28/2022 13:36:04 - INFO - codeparrot_training - Step 525: {'lr': 0.00013125000000000002, 'samples': 100992, 'steps': 525, 'loss/train': 2.349437177181244} 01/28/2022 13:36:09 - INFO - codeparrot_training - Step 526: {'lr': 0.0001315, 'samples': 101184, 'steps': 526, 'loss/train': 2.245360851287842} 01/28/2022 13:36:13 - INFO - codeparrot_training - Step 527: {'lr': 0.00013175, 'samples': 101376, 'steps': 527, 'loss/train': 1.464213252067566} 01/28/2022 13:36:17 - INFO - codeparrot_training - Step 528: {'lr': 0.000132, 'samples': 101568, 'steps': 528, 'loss/train': 0.34766723960638046} 01/28/2022 13:36:21 - INFO - codeparrot_training - Step 529: {'lr': 0.00013225000000000002, 'samples': 101760, 'steps': 529, 'loss/train': 2.9965327978134155} 01/28/2022 13:36:26 - INFO - codeparrot_training - Step 530: {'lr': 0.00013250000000000002, 'samples': 101952, 'steps': 530, 'loss/train': 2.038048267364502} 01/28/2022 13:36:30 - INFO - codeparrot_training - Step 531: {'lr': 0.00013275, 'samples': 102144, 'steps': 531, 'loss/train': 2.6090728640556335} 01/28/2022 13:36:35 - INFO - codeparrot_training - Step 532: {'lr': 0.000133, 'samples': 102336, 'steps': 532, 'loss/train': 1.7636557817459106} 01/28/2022 13:36:39 - INFO - codeparrot_training - Step 533: {'lr': 0.00013325, 'samples': 102528, 'steps': 533, 'loss/train': 2.100282669067383} 01/28/2022 13:36:43 - INFO - codeparrot_training - Step 534: {'lr': 0.00013350000000000002, 'samples': 102720, 'steps': 534, 'loss/train': 1.9676971435546875} 01/28/2022 13:36:48 - INFO - codeparrot_training - Step 535: {'lr': 0.00013375, 'samples': 102912, 'steps': 535, 'loss/train': 1.2690159380435944} 01/28/2022 13:36:52 - INFO - codeparrot_training - Step 536: {'lr': 0.000134, 'samples': 103104, 'steps': 536, 'loss/train': 2.6938878893852234} 01/28/2022 13:36:56 - INFO - codeparrot_training - Step 537: {'lr': 0.00013425, 'samples': 103296, 'steps': 537, 'loss/train': 2.315282106399536} 01/28/2022 13:37:01 - INFO - codeparrot_training - Step 538: {'lr': 0.00013450000000000002, 'samples': 103488, 'steps': 538, 'loss/train': 0.3548230156302452} 01/28/2022 13:37:05 - INFO - codeparrot_training - Step 539: {'lr': 0.00013475000000000002, 'samples': 103680, 'steps': 539, 'loss/train': 1.505185067653656} 01/28/2022 13:37:11 - INFO - codeparrot_training - Step 540: {'lr': 0.000135, 'samples': 103872, 'steps': 540, 'loss/train': 2.3433626890182495} 01/28/2022 13:37:15 - INFO - codeparrot_training - Step 541: {'lr': 0.00013525, 'samples': 104064, 'steps': 541, 'loss/train': 2.220704197883606} 01/28/2022 13:37:19 - INFO - codeparrot_training - Step 542: {'lr': 0.00013550000000000001, 'samples': 104256, 'steps': 542, 'loss/train': 1.8287013173103333} 01/28/2022 13:37:24 - INFO - codeparrot_training - Step 543: {'lr': 0.00013575000000000002, 'samples': 104448, 'steps': 543, 'loss/train': 1.16489177942276} 01/28/2022 13:37:28 - INFO - codeparrot_training - Step 544: {'lr': 0.00013600000000000003, 'samples': 104640, 'steps': 544, 'loss/train': 2.6590803265571594} 01/28/2022 13:37:33 - INFO - codeparrot_training - Step 545: {'lr': 0.00013625, 'samples': 104832, 'steps': 545, 'loss/train': 2.0241715908050537} 01/28/2022 13:37:37 - INFO - codeparrot_training - Step 546: {'lr': 0.0001365, 'samples': 105024, 'steps': 546, 'loss/train': 2.431113302707672} 01/28/2022 13:37:41 - INFO - codeparrot_training - Step 547: {'lr': 0.00013675000000000002, 'samples': 105216, 'steps': 547, 'loss/train': 1.7919319868087769} 01/28/2022 13:37:45 - INFO - codeparrot_training - Step 548: {'lr': 0.00013700000000000002, 'samples': 105408, 'steps': 548, 'loss/train': 1.962535321712494} 01/28/2022 13:37:50 - INFO - codeparrot_training - Step 549: {'lr': 0.00013725, 'samples': 105600, 'steps': 549, 'loss/train': 2.0620644092559814} 01/28/2022 13:37:55 - INFO - codeparrot_training - Step 550: {'lr': 0.0001375, 'samples': 105792, 'steps': 550, 'loss/train': 2.4056326746940613} 01/28/2022 13:37:59 - INFO - codeparrot_training - Step 551: {'lr': 0.00013775000000000001, 'samples': 105984, 'steps': 551, 'loss/train': 1.8909382224082947} 01/28/2022 13:38:03 - INFO - codeparrot_training - Step 552: {'lr': 0.00013800000000000002, 'samples': 106176, 'steps': 552, 'loss/train': 1.535703420639038} 01/28/2022 13:38:07 - INFO - codeparrot_training - Step 553: {'lr': 0.00013825000000000003, 'samples': 106368, 'steps': 553, 'loss/train': 2.2278385162353516} 01/28/2022 13:38:11 - INFO - codeparrot_training - Step 554: {'lr': 0.0001385, 'samples': 106560, 'steps': 554, 'loss/train': 1.7420064210891724} 01/28/2022 13:38:18 - INFO - codeparrot_training - Step 555: {'lr': 0.00013875, 'samples': 106752, 'steps': 555, 'loss/train': 2.375190496444702} 01/28/2022 13:38:22 - INFO - codeparrot_training - Step 556: {'lr': 0.00013900000000000002, 'samples': 106944, 'steps': 556, 'loss/train': 2.4950507283210754} 01/28/2022 13:38:26 - INFO - codeparrot_training - Step 557: {'lr': 0.00013925000000000002, 'samples': 107136, 'steps': 557, 'loss/train': 1.5562296509742737} 01/28/2022 13:38:30 - INFO - codeparrot_training - Step 558: {'lr': 0.0001395, 'samples': 107328, 'steps': 558, 'loss/train': 2.4424062967300415} 01/28/2022 13:38:34 - INFO - codeparrot_training - Step 559: {'lr': 0.00013975, 'samples': 107520, 'steps': 559, 'loss/train': 3.1070122718811035} 01/28/2022 13:38:39 - INFO - codeparrot_training - Step 560: {'lr': 0.00014000000000000001, 'samples': 107712, 'steps': 560, 'loss/train': 3.4953685998916626} 01/28/2022 13:38:44 - INFO - codeparrot_training - Step 561: {'lr': 0.00014025000000000002, 'samples': 107904, 'steps': 561, 'loss/train': 2.4157333374023438} 01/28/2022 13:38:48 - INFO - codeparrot_training - Step 562: {'lr': 0.00014050000000000003, 'samples': 108096, 'steps': 562, 'loss/train': 2.4815037846565247} 01/28/2022 13:38:52 - INFO - codeparrot_training - Step 563: {'lr': 0.00014074999999999998, 'samples': 108288, 'steps': 563, 'loss/train': 1.860853135585785} 01/28/2022 13:38:56 - INFO - codeparrot_training - Step 564: {'lr': 0.00014099999999999998, 'samples': 108480, 'steps': 564, 'loss/train': 2.395635187625885} 01/28/2022 13:39:01 - INFO - codeparrot_training - Step 565: {'lr': 0.00014125, 'samples': 108672, 'steps': 565, 'loss/train': 2.7822917103767395} 01/28/2022 13:39:05 - INFO - codeparrot_training - Step 566: {'lr': 0.0001415, 'samples': 108864, 'steps': 566, 'loss/train': 1.0121529400348663} 01/28/2022 13:39:10 - INFO - codeparrot_training - Step 567: {'lr': 0.00014175, 'samples': 109056, 'steps': 567, 'loss/train': 1.2860815823078156} 01/28/2022 13:39:14 - INFO - codeparrot_training - Step 568: {'lr': 0.00014199999999999998, 'samples': 109248, 'steps': 568, 'loss/train': 2.336336374282837} 01/28/2022 13:39:18 - INFO - codeparrot_training - Step 569: {'lr': 0.00014225, 'samples': 109440, 'steps': 569, 'loss/train': 2.524071514606476} 01/28/2022 13:39:24 - INFO - codeparrot_training - Step 570: {'lr': 0.0001425, 'samples': 109632, 'steps': 570, 'loss/train': 1.5741292834281921} 01/28/2022 13:39:29 - INFO - codeparrot_training - Step 571: {'lr': 0.00014275, 'samples': 109824, 'steps': 571, 'loss/train': 2.3421555161476135} 01/28/2022 13:39:33 - INFO - codeparrot_training - Step 572: {'lr': 0.00014299999999999998, 'samples': 110016, 'steps': 572, 'loss/train': 2.4181092381477356} 01/28/2022 13:39:37 - INFO - codeparrot_training - Step 573: {'lr': 0.00014324999999999999, 'samples': 110208, 'steps': 573, 'loss/train': 1.2129700183868408} 01/28/2022 13:39:42 - INFO - codeparrot_training - Step 574: {'lr': 0.0001435, 'samples': 110400, 'steps': 574, 'loss/train': 2.3288621306419373} 01/28/2022 13:39:46 - INFO - codeparrot_training - Step 575: {'lr': 0.00014375, 'samples': 110592, 'steps': 575, 'loss/train': 2.632317066192627} 01/28/2022 13:39:51 - INFO - codeparrot_training - Step 576: {'lr': 0.000144, 'samples': 110784, 'steps': 576, 'loss/train': 2.2250062823295593} 01/28/2022 13:39:55 - INFO - codeparrot_training - Step 577: {'lr': 0.00014424999999999998, 'samples': 110976, 'steps': 577, 'loss/train': 0.38584287464618683} 01/28/2022 13:39:59 - INFO - codeparrot_training - Step 578: {'lr': 0.0001445, 'samples': 111168, 'steps': 578, 'loss/train': 2.6866301894187927} 01/28/2022 13:40:05 - INFO - codeparrot_training - Step 579: {'lr': 0.00014475, 'samples': 111360, 'steps': 579, 'loss/train': 1.2802784442901611} 01/28/2022 13:40:09 - INFO - codeparrot_training - Step 580: {'lr': 0.000145, 'samples': 111552, 'steps': 580, 'loss/train': 1.2374301552772522} 01/28/2022 13:40:14 - INFO - codeparrot_training - Step 581: {'lr': 0.00014524999999999998, 'samples': 111744, 'steps': 581, 'loss/train': 1.6999842524528503} 01/28/2022 13:40:18 - INFO - codeparrot_training - Step 582: {'lr': 0.00014549999999999999, 'samples': 111936, 'steps': 582, 'loss/train': 2.119485318660736} 01/28/2022 13:40:22 - INFO - codeparrot_training - Step 583: {'lr': 0.00014575, 'samples': 112128, 'steps': 583, 'loss/train': 1.996980607509613} 01/28/2022 13:40:26 - INFO - codeparrot_training - Step 584: {'lr': 0.000146, 'samples': 112320, 'steps': 584, 'loss/train': 1.8928154110908508} 01/28/2022 13:40:31 - INFO - codeparrot_training - Step 585: {'lr': 0.00014625, 'samples': 112512, 'steps': 585, 'loss/train': 2.7012373208999634} 01/28/2022 13:40:36 - INFO - codeparrot_training - Step 586: {'lr': 0.00014649999999999998, 'samples': 112704, 'steps': 586, 'loss/train': 1.937915861606598} 01/28/2022 13:40:40 - INFO - codeparrot_training - Step 587: {'lr': 0.00014675, 'samples': 112896, 'steps': 587, 'loss/train': 1.3709938824176788} 01/28/2022 13:40:44 - INFO - codeparrot_training - Step 588: {'lr': 0.000147, 'samples': 113088, 'steps': 588, 'loss/train': 1.737755298614502} 01/28/2022 13:40:49 - INFO - codeparrot_training - Step 589: {'lr': 0.00014725, 'samples': 113280, 'steps': 589, 'loss/train': 2.3531811833381653} 01/28/2022 13:40:53 - INFO - codeparrot_training - Step 590: {'lr': 0.0001475, 'samples': 113472, 'steps': 590, 'loss/train': 1.6759954690933228} 01/28/2022 13:40:57 - INFO - codeparrot_training - Step 591: {'lr': 0.00014774999999999999, 'samples': 113664, 'steps': 591, 'loss/train': 1.5129391551017761} 01/28/2022 13:41:02 - INFO - codeparrot_training - Step 592: {'lr': 0.000148, 'samples': 113856, 'steps': 592, 'loss/train': 1.1192268133163452} 01/28/2022 13:41:06 - INFO - codeparrot_training - Step 593: {'lr': 0.00014825, 'samples': 114048, 'steps': 593, 'loss/train': 1.0748966932296753} 01/28/2022 13:41:11 - INFO - codeparrot_training - Step 594: {'lr': 0.0001485, 'samples': 114240, 'steps': 594, 'loss/train': 0.9071259498596191} 01/28/2022 13:41:15 - INFO - codeparrot_training - Step 595: {'lr': 0.00014874999999999998, 'samples': 114432, 'steps': 595, 'loss/train': 2.9671425819396973} 01/28/2022 13:41:19 - INFO - codeparrot_training - Step 596: {'lr': 0.000149, 'samples': 114624, 'steps': 596, 'loss/train': 2.400070309638977} 01/28/2022 13:41:24 - INFO - codeparrot_training - Step 597: {'lr': 0.00014925, 'samples': 114816, 'steps': 597, 'loss/train': 1.9464190006256104} 01/28/2022 13:41:28 - INFO - codeparrot_training - Step 598: {'lr': 0.0001495, 'samples': 115008, 'steps': 598, 'loss/train': 2.5532859563827515} 01/28/2022 13:41:34 - INFO - codeparrot_training - Step 599: {'lr': 0.00014975, 'samples': 115200, 'steps': 599, 'loss/train': 1.6116501688957214} 01/28/2022 13:41:38 - INFO - codeparrot_training - Step 600: {'lr': 0.00015, 'samples': 115392, 'steps': 600, 'loss/train': 2.051265299320221} 01/28/2022 13:41:42 - INFO - codeparrot_training - Step 601: {'lr': 0.00015025, 'samples': 115584, 'steps': 601, 'loss/train': 1.3858745098114014} 01/28/2022 13:41:47 - INFO - codeparrot_training - Step 602: {'lr': 0.0001505, 'samples': 115776, 'steps': 602, 'loss/train': 2.3164506554603577} 01/28/2022 13:41:51 - INFO - codeparrot_training - Step 603: {'lr': 0.00015075, 'samples': 115968, 'steps': 603, 'loss/train': 2.377302646636963} 01/28/2022 13:41:56 - INFO - codeparrot_training - Step 604: {'lr': 0.000151, 'samples': 116160, 'steps': 604, 'loss/train': 1.2618489861488342} 01/28/2022 13:42:00 - INFO - codeparrot_training - Step 605: {'lr': 0.00015125, 'samples': 116352, 'steps': 605, 'loss/train': 2.1234574913978577} 01/28/2022 13:42:04 - INFO - codeparrot_training - Step 606: {'lr': 0.0001515, 'samples': 116544, 'steps': 606, 'loss/train': 1.097367525100708} 01/28/2022 13:42:08 - INFO - codeparrot_training - Step 607: {'lr': 0.00015175, 'samples': 116736, 'steps': 607, 'loss/train': 2.438254237174988} 01/28/2022 13:42:13 - INFO - codeparrot_training - Step 608: {'lr': 0.000152, 'samples': 116928, 'steps': 608, 'loss/train': 0.9477005302906036} 01/28/2022 13:42:18 - INFO - codeparrot_training - Step 609: {'lr': 0.00015225, 'samples': 117120, 'steps': 609, 'loss/train': 2.601383149623871} 01/28/2022 13:42:22 - INFO - codeparrot_training - Step 610: {'lr': 0.0001525, 'samples': 117312, 'steps': 610, 'loss/train': 2.7112104892730713} 01/28/2022 13:42:26 - INFO - codeparrot_training - Step 611: {'lr': 0.00015275, 'samples': 117504, 'steps': 611, 'loss/train': 2.9145054817199707} 01/28/2022 13:42:30 - INFO - codeparrot_training - Step 612: {'lr': 0.000153, 'samples': 117696, 'steps': 612, 'loss/train': 1.7645321488380432} 01/28/2022 13:42:34 - INFO - codeparrot_training - Step 613: {'lr': 0.00015325, 'samples': 117888, 'steps': 613, 'loss/train': 2.312626004219055} 01/28/2022 13:42:41 - INFO - codeparrot_training - Step 614: {'lr': 0.0001535, 'samples': 118080, 'steps': 614, 'loss/train': 2.357422113418579} 01/28/2022 13:42:45 - INFO - codeparrot_training - Step 615: {'lr': 0.00015375, 'samples': 118272, 'steps': 615, 'loss/train': 2.432760536670685} 01/28/2022 13:42:49 - INFO - codeparrot_training - Step 616: {'lr': 0.000154, 'samples': 118464, 'steps': 616, 'loss/train': 1.920491337776184} 01/28/2022 13:42:53 - INFO - codeparrot_training - Step 617: {'lr': 0.00015425, 'samples': 118656, 'steps': 617, 'loss/train': 2.1638678312301636} 01/28/2022 13:42:57 - INFO - codeparrot_training - Step 618: {'lr': 0.00015450000000000001, 'samples': 118848, 'steps': 618, 'loss/train': 1.5627211332321167} 01/28/2022 13:43:03 - INFO - codeparrot_training - Step 619: {'lr': 0.00015475, 'samples': 119040, 'steps': 619, 'loss/train': 1.5436863899230957} 01/28/2022 13:43:07 - INFO - codeparrot_training - Step 620: {'lr': 0.000155, 'samples': 119232, 'steps': 620, 'loss/train': 1.648366391658783} 01/28/2022 13:43:11 - INFO - codeparrot_training - Step 621: {'lr': 0.00015525, 'samples': 119424, 'steps': 621, 'loss/train': 1.9625099301338196} 01/28/2022 13:43:15 - INFO - codeparrot_training - Step 622: {'lr': 0.0001555, 'samples': 119616, 'steps': 622, 'loss/train': 2.4296470284461975} 01/28/2022 13:43:19 - INFO - codeparrot_training - Step 623: {'lr': 0.00015575, 'samples': 119808, 'steps': 623, 'loss/train': 2.24530166387558} 01/28/2022 13:43:26 - INFO - codeparrot_training - Step 624: {'lr': 0.000156, 'samples': 120000, 'steps': 624, 'loss/train': 2.186740279197693} 01/28/2022 13:43:30 - INFO - codeparrot_training - Step 625: {'lr': 0.00015625, 'samples': 120192, 'steps': 625, 'loss/train': 2.57923024892807} 01/28/2022 13:43:34 - INFO - codeparrot_training - Step 626: {'lr': 0.0001565, 'samples': 120384, 'steps': 626, 'loss/train': 0.21514637768268585} 01/28/2022 13:43:38 - INFO - codeparrot_training - Step 627: {'lr': 0.00015675000000000002, 'samples': 120576, 'steps': 627, 'loss/train': 1.2782675921916962} 01/28/2022 13:43:42 - INFO - codeparrot_training - Step 628: {'lr': 0.000157, 'samples': 120768, 'steps': 628, 'loss/train': 1.5016379356384277} 01/28/2022 13:43:48 - INFO - codeparrot_training - Step 629: {'lr': 0.00015725, 'samples': 120960, 'steps': 629, 'loss/train': 2.411840558052063} 01/28/2022 13:43:52 - INFO - codeparrot_training - Step 630: {'lr': 0.0001575, 'samples': 121152, 'steps': 630, 'loss/train': 3.1171395778656006} 01/28/2022 13:43:56 - INFO - codeparrot_training - Step 631: {'lr': 0.00015775, 'samples': 121344, 'steps': 631, 'loss/train': 2.4070581793785095} 01/28/2022 13:44:00 - INFO - codeparrot_training - Step 632: {'lr': 0.000158, 'samples': 121536, 'steps': 632, 'loss/train': 2.004049837589264} 01/28/2022 13:44:04 - INFO - codeparrot_training - Step 633: {'lr': 0.00015825, 'samples': 121728, 'steps': 633, 'loss/train': 2.4875457286834717} 01/28/2022 13:44:09 - INFO - codeparrot_training - Step 634: {'lr': 0.0001585, 'samples': 121920, 'steps': 634, 'loss/train': 2.3557323217391968} 01/28/2022 13:44:14 - INFO - codeparrot_training - Step 635: {'lr': 0.00015875, 'samples': 122112, 'steps': 635, 'loss/train': 2.3731895685195923} 01/28/2022 13:44:18 - INFO - codeparrot_training - Step 636: {'lr': 0.00015900000000000002, 'samples': 122304, 'steps': 636, 'loss/train': 1.9418286681175232} 01/28/2022 13:44:22 - INFO - codeparrot_training - Step 637: {'lr': 0.00015925, 'samples': 122496, 'steps': 637, 'loss/train': 1.8106479048728943} 01/28/2022 13:44:26 - INFO - codeparrot_training - Step 638: {'lr': 0.0001595, 'samples': 122688, 'steps': 638, 'loss/train': 3.0402313470840454} 01/28/2022 13:44:32 - INFO - codeparrot_training - Step 639: {'lr': 0.00015975, 'samples': 122880, 'steps': 639, 'loss/train': 2.1514222025871277} 01/28/2022 13:44:36 - INFO - codeparrot_training - Step 640: {'lr': 0.00016, 'samples': 123072, 'steps': 640, 'loss/train': 1.9155203104019165} 01/28/2022 13:44:41 - INFO - codeparrot_training - Step 641: {'lr': 0.00016025000000000002, 'samples': 123264, 'steps': 641, 'loss/train': 2.1492696404457092} 01/28/2022 13:44:45 - INFO - codeparrot_training - Step 642: {'lr': 0.0001605, 'samples': 123456, 'steps': 642, 'loss/train': 1.9119831919670105} 01/28/2022 13:44:49 - INFO - codeparrot_training - Step 643: {'lr': 0.00016075, 'samples': 123648, 'steps': 643, 'loss/train': 1.6857231259346008} 01/28/2022 13:44:54 - INFO - codeparrot_training - Step 644: {'lr': 0.000161, 'samples': 123840, 'steps': 644, 'loss/train': 2.206142485141754} 01/28/2022 13:44:58 - INFO - codeparrot_training - Step 645: {'lr': 0.00016125000000000002, 'samples': 124032, 'steps': 645, 'loss/train': 2.237909495830536} 01/28/2022 13:45:02 - INFO - codeparrot_training - Step 646: {'lr': 0.0001615, 'samples': 124224, 'steps': 646, 'loss/train': 2.257291316986084} 01/28/2022 13:45:07 - INFO - codeparrot_training - Step 647: {'lr': 0.00016175, 'samples': 124416, 'steps': 647, 'loss/train': 2.5526073575019836} 01/28/2022 13:45:11 - INFO - codeparrot_training - Step 648: {'lr': 0.000162, 'samples': 124608, 'steps': 648, 'loss/train': 2.714849531650543} 01/28/2022 13:45:16 - INFO - codeparrot_training - Step 649: {'lr': 0.00016225000000000001, 'samples': 124800, 'steps': 649, 'loss/train': 2.0022963881492615} 01/28/2022 13:45:20 - INFO - codeparrot_training - Step 650: {'lr': 0.00016250000000000002, 'samples': 124992, 'steps': 650, 'loss/train': 1.6243588328361511} 01/28/2022 13:45:24 - INFO - codeparrot_training - Step 651: {'lr': 0.00016275, 'samples': 125184, 'steps': 651, 'loss/train': 1.1693198382854462} 01/28/2022 13:45:28 - INFO - codeparrot_training - Step 652: {'lr': 0.000163, 'samples': 125376, 'steps': 652, 'loss/train': 1.9620673656463623} 01/28/2022 13:45:33 - INFO - codeparrot_training - Step 653: {'lr': 0.00016325, 'samples': 125568, 'steps': 653, 'loss/train': 2.7315171360969543} 01/28/2022 13:45:39 - INFO - codeparrot_training - Step 654: {'lr': 0.00016350000000000002, 'samples': 125760, 'steps': 654, 'loss/train': 2.263129949569702} 01/28/2022 13:45:44 - INFO - codeparrot_training - Step 655: {'lr': 0.00016375000000000002, 'samples': 125952, 'steps': 655, 'loss/train': 2.9937681555747986} 01/28/2022 13:45:48 - INFO - codeparrot_training - Step 656: {'lr': 0.000164, 'samples': 126144, 'steps': 656, 'loss/train': 2.055426836013794} 01/28/2022 13:45:52 - INFO - codeparrot_training - Step 657: {'lr': 0.00016425, 'samples': 126336, 'steps': 657, 'loss/train': 1.3580561578273773} 01/28/2022 13:45:56 - INFO - codeparrot_training - Step 658: {'lr': 0.00016450000000000001, 'samples': 126528, 'steps': 658, 'loss/train': 2.403730630874634} 01/28/2022 13:46:01 - INFO - codeparrot_training - Step 659: {'lr': 0.00016475000000000002, 'samples': 126720, 'steps': 659, 'loss/train': 1.416071355342865} 01/28/2022 13:46:06 - INFO - codeparrot_training - Step 660: {'lr': 0.000165, 'samples': 126912, 'steps': 660, 'loss/train': 1.4284625351428986} 01/28/2022 13:46:10 - INFO - codeparrot_training - Step 661: {'lr': 0.00016525, 'samples': 127104, 'steps': 661, 'loss/train': 1.76523095369339} 01/28/2022 13:46:14 - INFO - codeparrot_training - Step 662: {'lr': 0.0001655, 'samples': 127296, 'steps': 662, 'loss/train': 2.2504477500915527} 01/28/2022 13:46:18 - INFO - codeparrot_training - Step 663: {'lr': 0.00016575000000000002, 'samples': 127488, 'steps': 663, 'loss/train': 2.462355673313141} 01/28/2022 13:46:23 - INFO - codeparrot_training - Step 664: {'lr': 0.00016600000000000002, 'samples': 127680, 'steps': 664, 'loss/train': 2.4343217611312866} 01/28/2022 13:46:27 - INFO - codeparrot_training - Step 665: {'lr': 0.00016625, 'samples': 127872, 'steps': 665, 'loss/train': 1.537813425064087} 01/28/2022 13:46:32 - INFO - codeparrot_training - Step 666: {'lr': 0.0001665, 'samples': 128064, 'steps': 666, 'loss/train': 2.2940849661827087} 01/28/2022 13:46:36 - INFO - codeparrot_training - Step 667: {'lr': 0.00016675000000000001, 'samples': 128256, 'steps': 667, 'loss/train': 0.7503932118415833} 01/28/2022 13:46:42 - INFO - codeparrot_training - Step 668: {'lr': 0.00016700000000000002, 'samples': 128448, 'steps': 668, 'loss/train': 1.8176835179328918} 01/28/2022 13:46:46 - INFO - codeparrot_training - Step 669: {'lr': 0.00016725000000000003, 'samples': 128640, 'steps': 669, 'loss/train': 1.193238079547882} 01/28/2022 13:46:51 - INFO - codeparrot_training - Step 670: {'lr': 0.0001675, 'samples': 128832, 'steps': 670, 'loss/train': 2.4515894651412964} 01/28/2022 13:46:55 - INFO - codeparrot_training - Step 671: {'lr': 0.00016775, 'samples': 129024, 'steps': 671, 'loss/train': 3.1075408458709717} 01/28/2022 13:46:59 - INFO - codeparrot_training - Step 672: {'lr': 0.00016800000000000002, 'samples': 129216, 'steps': 672, 'loss/train': 2.4856372475624084} 01/28/2022 13:47:04 - INFO - codeparrot_training - Step 673: {'lr': 0.00016825000000000002, 'samples': 129408, 'steps': 673, 'loss/train': 2.2740957140922546} 01/28/2022 13:47:08 - INFO - codeparrot_training - Step 674: {'lr': 0.0001685, 'samples': 129600, 'steps': 674, 'loss/train': 0.6187498569488525} 01/28/2022 13:47:13 - INFO - codeparrot_training - Step 675: {'lr': 0.00016875, 'samples': 129792, 'steps': 675, 'loss/train': 0.5071374624967575} 01/28/2022 13:47:17 - INFO - codeparrot_training - Step 676: {'lr': 0.00016900000000000002, 'samples': 129984, 'steps': 676, 'loss/train': 1.8388824462890625} 01/28/2022 13:47:21 - INFO - codeparrot_training - Step 677: {'lr': 0.00016925000000000002, 'samples': 130176, 'steps': 677, 'loss/train': 1.7039921879768372} 01/28/2022 13:47:26 - INFO - codeparrot_training - Step 678: {'lr': 0.00016950000000000003, 'samples': 130368, 'steps': 678, 'loss/train': 1.6443368196487427} 01/28/2022 13:47:30 - INFO - codeparrot_training - Step 679: {'lr': 0.00016975, 'samples': 130560, 'steps': 679, 'loss/train': 2.866596043109894} 01/28/2022 13:47:34 - INFO - codeparrot_training - Step 680: {'lr': 0.00017, 'samples': 130752, 'steps': 680, 'loss/train': 2.1567031145095825} 01/28/2022 13:47:38 - INFO - codeparrot_training - Step 681: {'lr': 0.00017025000000000002, 'samples': 130944, 'steps': 681, 'loss/train': 1.1343502700328827} 01/28/2022 13:47:43 - INFO - codeparrot_training - Step 682: {'lr': 0.00017050000000000002, 'samples': 131136, 'steps': 682, 'loss/train': 2.382563352584839} 01/28/2022 13:47:49 - INFO - codeparrot_training - Step 683: {'lr': 0.00017075, 'samples': 131328, 'steps': 683, 'loss/train': 2.6557987332344055} 01/28/2022 13:47:53 - INFO - codeparrot_training - Step 684: {'lr': 0.000171, 'samples': 131520, 'steps': 684, 'loss/train': 2.2218008637428284} 01/28/2022 13:47:57 - INFO - codeparrot_training - Step 685: {'lr': 0.00017125000000000002, 'samples': 131712, 'steps': 685, 'loss/train': 0.3219618648290634} 01/28/2022 13:48:01 - INFO - codeparrot_training - Step 686: {'lr': 0.00017150000000000002, 'samples': 131904, 'steps': 686, 'loss/train': 2.1139227747917175} 01/28/2022 13:48:06 - INFO - codeparrot_training - Step 687: {'lr': 0.00017175000000000003, 'samples': 132096, 'steps': 687, 'loss/train': 2.056875228881836} 01/28/2022 13:48:11 - INFO - codeparrot_training - Step 688: {'lr': 0.00017199999999999998, 'samples': 132288, 'steps': 688, 'loss/train': 2.5187333822250366} 01/28/2022 13:48:15 - INFO - codeparrot_training - Step 689: {'lr': 0.00017224999999999999, 'samples': 132480, 'steps': 689, 'loss/train': 2.5589182376861572} 01/28/2022 13:48:19 - INFO - codeparrot_training - Step 690: {'lr': 0.0001725, 'samples': 132672, 'steps': 690, 'loss/train': 1.7508323192596436} 01/28/2022 13:48:23 - INFO - codeparrot_training - Step 691: {'lr': 0.00017275, 'samples': 132864, 'steps': 691, 'loss/train': 2.8564528226852417} 01/28/2022 13:48:27 - INFO - codeparrot_training - Step 692: {'lr': 0.000173, 'samples': 133056, 'steps': 692, 'loss/train': 1.2361393868923187} 01/28/2022 13:48:33 - INFO - codeparrot_training - Step 693: {'lr': 0.00017324999999999998, 'samples': 133248, 'steps': 693, 'loss/train': 2.0843632221221924} 01/28/2022 13:48:38 - INFO - codeparrot_training - Step 694: {'lr': 0.0001735, 'samples': 133440, 'steps': 694, 'loss/train': 1.8635424971580505} 01/28/2022 13:48:42 - INFO - codeparrot_training - Step 695: {'lr': 0.00017375, 'samples': 133632, 'steps': 695, 'loss/train': 2.2758697271347046} 01/28/2022 13:48:46 - INFO - codeparrot_training - Step 696: {'lr': 0.000174, 'samples': 133824, 'steps': 696, 'loss/train': 2.1909876465797424} 01/28/2022 13:48:50 - INFO - codeparrot_training - Step 697: {'lr': 0.00017424999999999998, 'samples': 134016, 'steps': 697, 'loss/train': 2.010127544403076} 01/28/2022 13:48:55 - INFO - codeparrot_training - Step 698: {'lr': 0.00017449999999999999, 'samples': 134208, 'steps': 698, 'loss/train': 1.245522379875183} 01/28/2022 13:48:59 - INFO - codeparrot_training - Step 699: {'lr': 0.00017475, 'samples': 134400, 'steps': 699, 'loss/train': 2.3608108162879944} 01/28/2022 13:49:04 - INFO - codeparrot_training - Step 700: {'lr': 0.000175, 'samples': 134592, 'steps': 700, 'loss/train': 2.3515881299972534} 01/28/2022 13:49:08 - INFO - codeparrot_training - Step 701: {'lr': 0.00017525, 'samples': 134784, 'steps': 701, 'loss/train': 2.364215612411499} 01/28/2022 13:49:12 - INFO - codeparrot_training - Step 702: {'lr': 0.00017549999999999998, 'samples': 134976, 'steps': 702, 'loss/train': 2.6310383677482605} 01/28/2022 13:49:17 - INFO - codeparrot_training - Step 703: {'lr': 0.00017575, 'samples': 135168, 'steps': 703, 'loss/train': 2.053490102291107} 01/28/2022 13:49:21 - INFO - codeparrot_training - Step 704: {'lr': 0.000176, 'samples': 135360, 'steps': 704, 'loss/train': 1.8130384683609009} 01/28/2022 13:49:25 - INFO - codeparrot_training - Step 705: {'lr': 0.00017625, 'samples': 135552, 'steps': 705, 'loss/train': 2.0731971859931946} 01/28/2022 13:49:30 - INFO - codeparrot_training - Step 706: {'lr': 0.00017649999999999998, 'samples': 135744, 'steps': 706, 'loss/train': 1.7665809988975525} 01/28/2022 13:49:34 - INFO - codeparrot_training - Step 707: {'lr': 0.00017675, 'samples': 135936, 'steps': 707, 'loss/train': 3.3121590614318848} 01/28/2022 13:49:40 - INFO - codeparrot_training - Step 708: {'lr': 0.000177, 'samples': 136128, 'steps': 708, 'loss/train': 1.967294454574585} 01/28/2022 13:49:44 - INFO - codeparrot_training - Step 709: {'lr': 0.00017725, 'samples': 136320, 'steps': 709, 'loss/train': 1.8335312604904175} 01/28/2022 13:49:48 - INFO - codeparrot_training - Step 710: {'lr': 0.0001775, 'samples': 136512, 'steps': 710, 'loss/train': 1.8496663570404053} 01/28/2022 13:49:52 - INFO - codeparrot_training - Step 711: {'lr': 0.00017774999999999998, 'samples': 136704, 'steps': 711, 'loss/train': 2.1288585662841797} 01/28/2022 13:49:56 - INFO - codeparrot_training - Step 712: {'lr': 0.000178, 'samples': 136896, 'steps': 712, 'loss/train': 2.175538659095764} 01/28/2022 13:50:01 - INFO - codeparrot_training - Step 713: {'lr': 0.00017825, 'samples': 137088, 'steps': 713, 'loss/train': 1.683388888835907} 01/28/2022 13:50:06 - INFO - codeparrot_training - Step 714: {'lr': 0.0001785, 'samples': 137280, 'steps': 714, 'loss/train': 2.1068199276924133} 01/28/2022 13:50:10 - INFO - codeparrot_training - Step 715: {'lr': 0.00017875, 'samples': 137472, 'steps': 715, 'loss/train': 2.151412010192871} 01/28/2022 13:50:14 - INFO - codeparrot_training - Step 716: {'lr': 0.000179, 'samples': 137664, 'steps': 716, 'loss/train': 1.9954814314842224} 01/28/2022 13:50:18 - INFO - codeparrot_training - Step 717: {'lr': 0.00017925, 'samples': 137856, 'steps': 717, 'loss/train': 1.8740047216415405} 01/28/2022 13:50:23 - INFO - codeparrot_training - Step 718: {'lr': 0.0001795, 'samples': 138048, 'steps': 718, 'loss/train': 2.836313009262085} 01/28/2022 13:50:27 - INFO - codeparrot_training - Step 719: {'lr': 0.00017975, 'samples': 138240, 'steps': 719, 'loss/train': 2.4179506301879883} 01/28/2022 13:50:32 - INFO - codeparrot_training - Step 720: {'lr': 0.00017999999999999998, 'samples': 138432, 'steps': 720, 'loss/train': 1.003441035747528} 01/28/2022 13:50:36 - INFO - codeparrot_training - Step 721: {'lr': 0.00018025, 'samples': 138624, 'steps': 721, 'loss/train': 2.110181987285614} 01/28/2022 13:50:40 - INFO - codeparrot_training - Step 722: {'lr': 0.0001805, 'samples': 138816, 'steps': 722, 'loss/train': 1.2215899229049683} 01/28/2022 13:50:45 - INFO - codeparrot_training - Step 723: {'lr': 0.00018075, 'samples': 139008, 'steps': 723, 'loss/train': 2.227561354637146} 01/28/2022 13:50:49 - INFO - codeparrot_training - Step 724: {'lr': 0.000181, 'samples': 139200, 'steps': 724, 'loss/train': 0.8702154457569122} 01/28/2022 13:50:53 - INFO - codeparrot_training - Step 725: {'lr': 0.00018125, 'samples': 139392, 'steps': 725, 'loss/train': 2.0363819003105164} 01/28/2022 13:50:58 - INFO - codeparrot_training - Step 726: {'lr': 0.0001815, 'samples': 139584, 'steps': 726, 'loss/train': 1.328645020723343} 01/28/2022 13:51:02 - INFO - codeparrot_training - Step 727: {'lr': 0.00018175, 'samples': 139776, 'steps': 727, 'loss/train': 1.1624747514724731} 01/28/2022 13:51:09 - INFO - codeparrot_training - Step 728: {'lr': 0.000182, 'samples': 139968, 'steps': 728, 'loss/train': 1.9510837197303772} 01/28/2022 13:51:13 - INFO - codeparrot_training - Step 729: {'lr': 0.00018225, 'samples': 140160, 'steps': 729, 'loss/train': 1.3540838956832886} 01/28/2022 13:51:17 - INFO - codeparrot_training - Step 730: {'lr': 0.0001825, 'samples': 140352, 'steps': 730, 'loss/train': 1.785629689693451} 01/28/2022 13:51:22 - INFO - codeparrot_training - Step 731: {'lr': 0.00018275, 'samples': 140544, 'steps': 731, 'loss/train': 2.3768343329429626} 01/28/2022 13:51:26 - INFO - codeparrot_training - Step 732: {'lr': 0.000183, 'samples': 140736, 'steps': 732, 'loss/train': 2.1523295044898987} 01/28/2022 13:51:31 - INFO - codeparrot_training - Step 733: {'lr': 0.00018325, 'samples': 140928, 'steps': 733, 'loss/train': 1.305323839187622} 01/28/2022 13:51:35 - INFO - codeparrot_training - Step 734: {'lr': 0.0001835, 'samples': 141120, 'steps': 734, 'loss/train': 2.216181993484497} 01/28/2022 13:51:39 - INFO - codeparrot_training - Step 735: {'lr': 0.00018375, 'samples': 141312, 'steps': 735, 'loss/train': 1.8774415254592896} 01/28/2022 13:51:43 - INFO - codeparrot_training - Step 736: {'lr': 0.000184, 'samples': 141504, 'steps': 736, 'loss/train': 2.395267903804779} 01/28/2022 13:51:48 - INFO - codeparrot_training - Step 737: {'lr': 0.00018425, 'samples': 141696, 'steps': 737, 'loss/train': 2.4868631958961487} 01/28/2022 13:51:54 - INFO - codeparrot_training - Step 738: {'lr': 0.0001845, 'samples': 141888, 'steps': 738, 'loss/train': 2.3096947073936462} 01/28/2022 13:51:58 - INFO - codeparrot_training - Step 739: {'lr': 0.00018475, 'samples': 142080, 'steps': 739, 'loss/train': 1.2793242037296295} 01/28/2022 13:52:02 - INFO - codeparrot_training - Step 740: {'lr': 0.000185, 'samples': 142272, 'steps': 740, 'loss/train': 1.1945377886295319} 01/28/2022 13:52:06 - INFO - codeparrot_training - Step 741: {'lr': 0.00018525, 'samples': 142464, 'steps': 741, 'loss/train': 2.611294984817505} 01/28/2022 13:52:10 - INFO - codeparrot_training - Step 742: {'lr': 0.0001855, 'samples': 142656, 'steps': 742, 'loss/train': 1.6075279712677002} 01/28/2022 13:52:16 - INFO - codeparrot_training - Step 743: {'lr': 0.00018575000000000002, 'samples': 142848, 'steps': 743, 'loss/train': 0.28468088060617447} 01/28/2022 13:52:20 - INFO - codeparrot_training - Step 744: {'lr': 0.000186, 'samples': 143040, 'steps': 744, 'loss/train': 0.3475906401872635} 01/28/2022 13:52:24 - INFO - codeparrot_training - Step 745: {'lr': 0.00018625, 'samples': 143232, 'steps': 745, 'loss/train': 1.9002567529678345} 01/28/2022 13:52:28 - INFO - codeparrot_training - Step 746: {'lr': 0.0001865, 'samples': 143424, 'steps': 746, 'loss/train': 2.146664321422577} 01/28/2022 13:52:32 - INFO - codeparrot_training - Step 747: {'lr': 0.00018675, 'samples': 143616, 'steps': 747, 'loss/train': 2.1063637733459473} 01/28/2022 13:52:38 - INFO - codeparrot_training - Step 748: {'lr': 0.000187, 'samples': 143808, 'steps': 748, 'loss/train': 1.0725697875022888} 01/28/2022 13:52:42 - INFO - codeparrot_training - Step 749: {'lr': 0.00018725, 'samples': 144000, 'steps': 749, 'loss/train': 2.326325297355652} 01/28/2022 13:52:46 - INFO - codeparrot_training - Step 750: {'lr': 0.0001875, 'samples': 144192, 'steps': 750, 'loss/train': 0.3939068466424942} 01/28/2022 13:52:50 - INFO - codeparrot_training - Step 751: {'lr': 0.00018775, 'samples': 144384, 'steps': 751, 'loss/train': 2.0875954627990723} 01/28/2022 13:52:54 - INFO - codeparrot_training - Step 752: {'lr': 0.00018800000000000002, 'samples': 144576, 'steps': 752, 'loss/train': 2.3993967175483704} 01/28/2022 13:53:00 - INFO - codeparrot_training - Step 753: {'lr': 0.00018825, 'samples': 144768, 'steps': 753, 'loss/train': 3.2988767623901367} 01/28/2022 13:53:05 - INFO - codeparrot_training - Step 754: {'lr': 0.0001885, 'samples': 144960, 'steps': 754, 'loss/train': 2.5259474515914917} 01/28/2022 13:53:09 - INFO - codeparrot_training - Step 755: {'lr': 0.00018875, 'samples': 145152, 'steps': 755, 'loss/train': 1.3819931745529175} 01/28/2022 13:53:13 - INFO - codeparrot_training - Step 756: {'lr': 0.000189, 'samples': 145344, 'steps': 756, 'loss/train': 2.74090039730072} 01/28/2022 13:53:17 - INFO - codeparrot_training - Step 757: {'lr': 0.00018925, 'samples': 145536, 'steps': 757, 'loss/train': 2.4007617831230164} 01/28/2022 13:53:22 - INFO - codeparrot_training - Step 758: {'lr': 0.0001895, 'samples': 145728, 'steps': 758, 'loss/train': 1.7531501054763794} 01/28/2022 13:53:26 - INFO - codeparrot_training - Step 759: {'lr': 0.00018975, 'samples': 145920, 'steps': 759, 'loss/train': 1.9607099890708923} 01/28/2022 13:53:31 - INFO - codeparrot_training - Step 760: {'lr': 0.00019, 'samples': 146112, 'steps': 760, 'loss/train': 3.1573355197906494} 01/28/2022 13:53:35 - INFO - codeparrot_training - Step 761: {'lr': 0.00019025000000000002, 'samples': 146304, 'steps': 761, 'loss/train': 1.4960399866104126} 01/28/2022 13:53:39 - INFO - codeparrot_training - Step 762: {'lr': 0.0001905, 'samples': 146496, 'steps': 762, 'loss/train': 2.6072418093681335} 01/28/2022 13:53:44 - INFO - codeparrot_training - Step 763: {'lr': 0.00019075, 'samples': 146688, 'steps': 763, 'loss/train': 1.6823059916496277} 01/28/2022 13:53:49 - INFO - codeparrot_training - Step 764: {'lr': 0.000191, 'samples': 146880, 'steps': 764, 'loss/train': 2.3810341358184814} 01/28/2022 13:53:53 - INFO - codeparrot_training - Step 765: {'lr': 0.00019125000000000001, 'samples': 147072, 'steps': 765, 'loss/train': 1.8942965269088745} 01/28/2022 13:53:57 - INFO - codeparrot_training - Step 766: {'lr': 0.00019150000000000002, 'samples': 147264, 'steps': 766, 'loss/train': 0.8008798062801361} 01/28/2022 13:54:01 - INFO - codeparrot_training - Step 767: {'lr': 0.00019175, 'samples': 147456, 'steps': 767, 'loss/train': 2.4440735578536987} 01/28/2022 13:54:07 - INFO - codeparrot_training - Step 768: {'lr': 0.000192, 'samples': 147648, 'steps': 768, 'loss/train': 1.453626573085785} 01/28/2022 13:54:11 - INFO - codeparrot_training - Step 769: {'lr': 0.00019225, 'samples': 147840, 'steps': 769, 'loss/train': 2.397792398929596} 01/28/2022 13:54:15 - INFO - codeparrot_training - Step 770: {'lr': 0.00019250000000000002, 'samples': 148032, 'steps': 770, 'loss/train': 1.9776597619056702} 01/28/2022 13:54:19 - INFO - codeparrot_training - Step 771: {'lr': 0.00019275, 'samples': 148224, 'steps': 771, 'loss/train': 1.4877351522445679} 01/28/2022 13:54:23 - INFO - codeparrot_training - Step 772: {'lr': 0.000193, 'samples': 148416, 'steps': 772, 'loss/train': 2.0857322216033936} 01/28/2022 13:54:30 - INFO - codeparrot_training - Step 773: {'lr': 0.00019325, 'samples': 148608, 'steps': 773, 'loss/train': 0.4312707334756851} 01/28/2022 13:54:34 - INFO - codeparrot_training - Step 774: {'lr': 0.00019350000000000001, 'samples': 148800, 'steps': 774, 'loss/train': 1.2647781372070312} 01/28/2022 13:54:38 - INFO - codeparrot_training - Step 775: {'lr': 0.00019375000000000002, 'samples': 148992, 'steps': 775, 'loss/train': 1.962451994419098} 01/28/2022 13:54:42 - INFO - codeparrot_training - Step 776: {'lr': 0.000194, 'samples': 149184, 'steps': 776, 'loss/train': 1.943808138370514} 01/28/2022 13:54:47 - INFO - codeparrot_training - Step 777: {'lr': 0.00019425, 'samples': 149376, 'steps': 777, 'loss/train': 1.8323593139648438} 01/28/2022 13:54:51 - INFO - codeparrot_training - Step 778: {'lr': 0.0001945, 'samples': 149568, 'steps': 778, 'loss/train': 1.6993154883384705} 01/28/2022 13:54:55 - INFO - codeparrot_training - Step 779: {'lr': 0.00019475000000000002, 'samples': 149760, 'steps': 779, 'loss/train': 2.673709452152252} 01/28/2022 13:55:00 - INFO - codeparrot_training - Step 780: {'lr': 0.00019500000000000002, 'samples': 149952, 'steps': 780, 'loss/train': 2.29527086019516} 01/28/2022 13:55:04 - INFO - codeparrot_training - Step 781: {'lr': 0.00019525, 'samples': 150144, 'steps': 781, 'loss/train': 2.5113126039505005} 01/28/2022 13:55:10 - INFO - codeparrot_training - Step 782: {'lr': 0.0001955, 'samples': 150336, 'steps': 782, 'loss/train': 1.8993178009986877} 01/28/2022 13:55:14 - INFO - codeparrot_training - Step 783: {'lr': 0.00019575000000000001, 'samples': 150528, 'steps': 783, 'loss/train': 1.7865790128707886} 01/28/2022 13:55:19 - INFO - codeparrot_training - Step 784: {'lr': 0.00019600000000000002, 'samples': 150720, 'steps': 784, 'loss/train': 2.579177677631378} 01/28/2022 13:55:23 - INFO - codeparrot_training - Step 785: {'lr': 0.00019625, 'samples': 150912, 'steps': 785, 'loss/train': 2.2895668745040894} 01/28/2022 13:55:27 - INFO - codeparrot_training - Step 786: {'lr': 0.0001965, 'samples': 151104, 'steps': 786, 'loss/train': 2.4998415112495422} 01/28/2022 13:55:32 - INFO - codeparrot_training - Step 787: {'lr': 0.00019675, 'samples': 151296, 'steps': 787, 'loss/train': 2.8438369631767273} 01/28/2022 13:55:36 - INFO - codeparrot_training - Step 788: {'lr': 0.00019700000000000002, 'samples': 151488, 'steps': 788, 'loss/train': 2.2424651384353638} 01/28/2022 13:55:40 - INFO - codeparrot_training - Step 789: {'lr': 0.00019725000000000002, 'samples': 151680, 'steps': 789, 'loss/train': 2.666476786136627} 01/28/2022 13:55:45 - INFO - codeparrot_training - Step 790: {'lr': 0.0001975, 'samples': 151872, 'steps': 790, 'loss/train': 2.0350241661071777} 01/28/2022 13:55:49 - INFO - codeparrot_training - Step 791: {'lr': 0.00019775, 'samples': 152064, 'steps': 791, 'loss/train': 1.8862815499305725} 01/28/2022 13:55:54 - INFO - codeparrot_training - Step 792: {'lr': 0.00019800000000000002, 'samples': 152256, 'steps': 792, 'loss/train': 1.908484160900116} 01/28/2022 13:55:58 - INFO - codeparrot_training - Step 793: {'lr': 0.00019825000000000002, 'samples': 152448, 'steps': 793, 'loss/train': 0.45117898285388947} 01/28/2022 13:56:02 - INFO - codeparrot_training - Step 794: {'lr': 0.00019850000000000003, 'samples': 152640, 'steps': 794, 'loss/train': 2.614010274410248} 01/28/2022 13:56:06 - INFO - codeparrot_training - Step 795: {'lr': 0.00019875, 'samples': 152832, 'steps': 795, 'loss/train': 2.60962450504303} 01/28/2022 13:56:11 - INFO - codeparrot_training - Step 796: {'lr': 0.000199, 'samples': 153024, 'steps': 796, 'loss/train': 0.4116477370262146} 01/28/2022 13:56:17 - INFO - codeparrot_training - Step 797: {'lr': 0.00019925000000000002, 'samples': 153216, 'steps': 797, 'loss/train': 2.3355456590652466} 01/28/2022 13:56:21 - INFO - codeparrot_training - Step 798: {'lr': 0.00019950000000000002, 'samples': 153408, 'steps': 798, 'loss/train': 1.2737031877040863} 01/28/2022 13:56:25 - INFO - codeparrot_training - Step 799: {'lr': 0.00019975, 'samples': 153600, 'steps': 799, 'loss/train': 1.3683399260044098} 01/28/2022 13:56:30 - INFO - codeparrot_training - Step 800: {'lr': 0.0002, 'samples': 153792, 'steps': 800, 'loss/train': 1.414448618888855} 01/28/2022 13:56:34 - INFO - codeparrot_training - Step 801: {'lr': 0.00020025000000000002, 'samples': 153984, 'steps': 801, 'loss/train': 4.154987454414368} 01/28/2022 13:56:39 - INFO - codeparrot_training - Step 802: {'lr': 0.00020050000000000002, 'samples': 154176, 'steps': 802, 'loss/train': 2.5191761255264282} 01/28/2022 13:56:43 - INFO - codeparrot_training - Step 803: {'lr': 0.00020075000000000003, 'samples': 154368, 'steps': 803, 'loss/train': 1.8969095349311829} 01/28/2022 13:56:47 - INFO - codeparrot_training - Step 804: {'lr': 0.000201, 'samples': 154560, 'steps': 804, 'loss/train': 2.3058918714523315} 01/28/2022 13:56:51 - INFO - codeparrot_training - Step 805: {'lr': 0.00020125, 'samples': 154752, 'steps': 805, 'loss/train': 2.680654764175415} 01/28/2022 13:56:56 - INFO - codeparrot_training - Step 806: {'lr': 0.00020150000000000002, 'samples': 154944, 'steps': 806, 'loss/train': 1.9588572978973389} 01/28/2022 13:57:01 - INFO - codeparrot_training - Step 807: {'lr': 0.00020175000000000003, 'samples': 155136, 'steps': 807, 'loss/train': 2.012816369533539} 01/28/2022 13:57:05 - INFO - codeparrot_training - Step 808: {'lr': 0.000202, 'samples': 155328, 'steps': 808, 'loss/train': 1.887999415397644} 01/28/2022 13:57:09 - INFO - codeparrot_training - Step 809: {'lr': 0.00020225, 'samples': 155520, 'steps': 809, 'loss/train': 0.6191268861293793} 01/28/2022 13:57:13 - INFO - codeparrot_training - Step 810: {'lr': 0.00020250000000000002, 'samples': 155712, 'steps': 810, 'loss/train': 2.712753117084503} 01/28/2022 13:57:17 - INFO - codeparrot_training - Step 811: {'lr': 0.00020275000000000002, 'samples': 155904, 'steps': 811, 'loss/train': 3.355473518371582} 01/28/2022 13:57:24 - INFO - codeparrot_training - Step 812: {'lr': 0.00020300000000000003, 'samples': 156096, 'steps': 812, 'loss/train': 1.9651281237602234} 01/28/2022 13:57:28 - INFO - codeparrot_training - Step 813: {'lr': 0.00020324999999999998, 'samples': 156288, 'steps': 813, 'loss/train': 2.0714362263679504} 01/28/2022 13:57:32 - INFO - codeparrot_training - Step 814: {'lr': 0.00020349999999999999, 'samples': 156480, 'steps': 814, 'loss/train': 1.8547798991203308} 01/28/2022 13:57:36 - INFO - codeparrot_training - Step 815: {'lr': 0.00020375, 'samples': 156672, 'steps': 815, 'loss/train': 1.788557231426239} 01/28/2022 13:57:41 - INFO - codeparrot_training - Step 816: {'lr': 0.000204, 'samples': 156864, 'steps': 816, 'loss/train': 1.8209617137908936} 01/28/2022 13:57:46 - INFO - codeparrot_training - Step 817: {'lr': 0.00020425, 'samples': 157056, 'steps': 817, 'loss/train': 2.827017366886139} 01/28/2022 13:57:50 - INFO - codeparrot_training - Step 818: {'lr': 0.00020449999999999998, 'samples': 157248, 'steps': 818, 'loss/train': 1.0075352489948273} 01/28/2022 13:57:54 - INFO - codeparrot_training - Step 819: {'lr': 0.00020475, 'samples': 157440, 'steps': 819, 'loss/train': 2.7741583585739136} 01/28/2022 13:57:58 - INFO - codeparrot_training - Step 820: {'lr': 0.000205, 'samples': 157632, 'steps': 820, 'loss/train': 2.1799553632736206} 01/28/2022 13:58:03 - INFO - codeparrot_training - Step 821: {'lr': 0.00020525, 'samples': 157824, 'steps': 821, 'loss/train': 1.8148106932640076} 01/28/2022 13:58:08 - INFO - codeparrot_training - Step 822: {'lr': 0.00020549999999999998, 'samples': 158016, 'steps': 822, 'loss/train': 2.0223124623298645} 01/28/2022 13:58:12 - INFO - codeparrot_training - Step 823: {'lr': 0.00020575, 'samples': 158208, 'steps': 823, 'loss/train': 0.3983446955680847} 01/28/2022 13:58:16 - INFO - codeparrot_training - Step 824: {'lr': 0.000206, 'samples': 158400, 'steps': 824, 'loss/train': 2.5489456057548523} 01/28/2022 13:58:20 - INFO - codeparrot_training - Step 825: {'lr': 0.00020625, 'samples': 158592, 'steps': 825, 'loss/train': 1.0403689742088318} 01/28/2022 13:58:24 - INFO - codeparrot_training - Step 826: {'lr': 0.0002065, 'samples': 158784, 'steps': 826, 'loss/train': 1.494350552558899} 01/28/2022 13:58:31 - INFO - codeparrot_training - Step 827: {'lr': 0.00020674999999999998, 'samples': 158976, 'steps': 827, 'loss/train': 1.411444365978241} 01/28/2022 13:58:35 - INFO - codeparrot_training - Step 828: {'lr': 0.000207, 'samples': 159168, 'steps': 828, 'loss/train': 1.875558078289032} 01/28/2022 13:58:39 - INFO - codeparrot_training - Step 829: {'lr': 0.00020725, 'samples': 159360, 'steps': 829, 'loss/train': 2.001776933670044} 01/28/2022 13:58:43 - INFO - codeparrot_training - Step 830: {'lr': 0.0002075, 'samples': 159552, 'steps': 830, 'loss/train': 0.9597708284854889} 01/28/2022 13:58:47 - INFO - codeparrot_training - Step 831: {'lr': 0.00020774999999999998, 'samples': 159744, 'steps': 831, 'loss/train': 1.5234403610229492} 01/28/2022 13:58:52 - INFO - codeparrot_training - Step 832: {'lr': 0.000208, 'samples': 159936, 'steps': 832, 'loss/train': 2.019140303134918} 01/28/2022 13:58:57 - INFO - codeparrot_training - Step 833: {'lr': 0.00020825, 'samples': 160128, 'steps': 833, 'loss/train': 2.260460078716278} 01/28/2022 13:59:01 - INFO - codeparrot_training - Step 834: {'lr': 0.0002085, 'samples': 160320, 'steps': 834, 'loss/train': 1.7484446167945862} 01/28/2022 13:59:05 - INFO - codeparrot_training - Step 835: {'lr': 0.00020875, 'samples': 160512, 'steps': 835, 'loss/train': 2.249435842037201} 01/28/2022 13:59:09 - INFO - codeparrot_training - Step 836: {'lr': 0.00020899999999999998, 'samples': 160704, 'steps': 836, 'loss/train': 1.3577304482460022} 01/28/2022 13:59:15 - INFO - codeparrot_training - Step 837: {'lr': 0.00020925, 'samples': 160896, 'steps': 837, 'loss/train': 0.6003280878067017} 01/28/2022 13:59:19 - INFO - codeparrot_training - Step 838: {'lr': 0.0002095, 'samples': 161088, 'steps': 838, 'loss/train': 2.2564364075660706} 01/28/2022 13:59:23 - INFO - codeparrot_training - Step 839: {'lr': 0.00020975, 'samples': 161280, 'steps': 839, 'loss/train': 1.74202162027359} 01/28/2022 13:59:27 - INFO - codeparrot_training - Step 840: {'lr': 0.00021, 'samples': 161472, 'steps': 840, 'loss/train': 2.0347350239753723} 01/28/2022 13:59:31 - INFO - codeparrot_training - Step 841: {'lr': 0.00021025, 'samples': 161664, 'steps': 841, 'loss/train': 1.6626008749008179} 01/28/2022 13:59:38 - INFO - codeparrot_training - Step 842: {'lr': 0.0002105, 'samples': 161856, 'steps': 842, 'loss/train': 1.5262354016304016} 01/28/2022 13:59:42 - INFO - codeparrot_training - Step 843: {'lr': 0.00021075, 'samples': 162048, 'steps': 843, 'loss/train': 1.3404665887355804} 01/28/2022 13:59:46 - INFO - codeparrot_training - Step 844: {'lr': 0.000211, 'samples': 162240, 'steps': 844, 'loss/train': 1.5046443343162537} 01/28/2022 13:59:51 - INFO - codeparrot_training - Step 845: {'lr': 0.00021124999999999998, 'samples': 162432, 'steps': 845, 'loss/train': 1.9999444484710693} 01/28/2022 13:59:55 - INFO - codeparrot_training - Step 846: {'lr': 0.0002115, 'samples': 162624, 'steps': 846, 'loss/train': 1.0297354459762573} 01/28/2022 13:59:59 - INFO - codeparrot_training - Step 847: {'lr': 0.00021175, 'samples': 162816, 'steps': 847, 'loss/train': 1.5915196537971497} 01/28/2022 14:00:04 - INFO - codeparrot_training - Step 848: {'lr': 0.000212, 'samples': 163008, 'steps': 848, 'loss/train': 1.677103042602539} 01/28/2022 14:00:09 - INFO - codeparrot_training - Step 849: {'lr': 0.00021225, 'samples': 163200, 'steps': 849, 'loss/train': 2.068362772464752} 01/28/2022 14:00:13 - INFO - codeparrot_training - Step 850: {'lr': 0.0002125, 'samples': 163392, 'steps': 850, 'loss/train': 2.606583058834076} 01/28/2022 14:00:17 - INFO - codeparrot_training - Step 851: {'lr': 0.00021275, 'samples': 163584, 'steps': 851, 'loss/train': 2.0794715881347656} 01/28/2022 14:00:22 - INFO - codeparrot_training - Step 852: {'lr': 0.000213, 'samples': 163776, 'steps': 852, 'loss/train': 1.679269552230835} 01/28/2022 14:00:26 - INFO - codeparrot_training - Step 853: {'lr': 0.00021325, 'samples': 163968, 'steps': 853, 'loss/train': 1.684156894683838} 01/28/2022 14:00:30 - INFO - codeparrot_training - Step 854: {'lr': 0.0002135, 'samples': 164160, 'steps': 854, 'loss/train': 2.913556694984436} 01/28/2022 14:00:34 - INFO - codeparrot_training - Step 855: {'lr': 0.00021375, 'samples': 164352, 'steps': 855, 'loss/train': 1.0215583741664886} 01/28/2022 14:00:39 - INFO - codeparrot_training - Step 856: {'lr': 0.000214, 'samples': 164544, 'steps': 856, 'loss/train': 1.4885996282100677} 01/28/2022 14:00:45 - INFO - codeparrot_training - Step 857: {'lr': 0.00021425, 'samples': 164736, 'steps': 857, 'loss/train': 2.687958598136902} 01/28/2022 14:00:49 - INFO - codeparrot_training - Step 858: {'lr': 0.0002145, 'samples': 164928, 'steps': 858, 'loss/train': 2.189223289489746} 01/28/2022 14:00:53 - INFO - codeparrot_training - Step 859: {'lr': 0.00021475, 'samples': 165120, 'steps': 859, 'loss/train': 1.2806653082370758} 01/28/2022 14:00:57 - INFO - codeparrot_training - Step 860: {'lr': 0.000215, 'samples': 165312, 'steps': 860, 'loss/train': 1.780119001865387} 01/28/2022 14:01:01 - INFO - codeparrot_training - Step 861: {'lr': 0.00021525, 'samples': 165504, 'steps': 861, 'loss/train': 1.5507606267929077} 01/28/2022 14:01:06 - INFO - codeparrot_training - Step 862: {'lr': 0.0002155, 'samples': 165696, 'steps': 862, 'loss/train': 3.407323122024536} 01/28/2022 14:01:11 - INFO - codeparrot_training - Step 863: {'lr': 0.00021575, 'samples': 165888, 'steps': 863, 'loss/train': 1.8790640830993652} 01/28/2022 14:01:15 - INFO - codeparrot_training - Step 864: {'lr': 0.000216, 'samples': 166080, 'steps': 864, 'loss/train': 2.0040353536605835} 01/28/2022 14:01:19 - INFO - codeparrot_training - Step 865: {'lr': 0.00021625, 'samples': 166272, 'steps': 865, 'loss/train': 2.222499132156372} 01/28/2022 14:01:23 - INFO - codeparrot_training - Step 866: {'lr': 0.0002165, 'samples': 166464, 'steps': 866, 'loss/train': 2.053077757358551} 01/28/2022 14:01:28 - INFO - codeparrot_training - Step 867: {'lr': 0.00021675, 'samples': 166656, 'steps': 867, 'loss/train': 2.0784364342689514} 01/28/2022 14:01:32 - INFO - codeparrot_training - Step 868: {'lr': 0.00021700000000000002, 'samples': 166848, 'steps': 868, 'loss/train': 2.3811036944389343} 01/28/2022 14:01:37 - INFO - codeparrot_training - Step 869: {'lr': 0.00021725, 'samples': 167040, 'steps': 869, 'loss/train': 1.4527392983436584} 01/28/2022 14:01:41 - INFO - codeparrot_training - Step 870: {'lr': 0.0002175, 'samples': 167232, 'steps': 870, 'loss/train': 2.221083104610443} 01/28/2022 14:01:45 - INFO - codeparrot_training - Step 871: {'lr': 0.00021775, 'samples': 167424, 'steps': 871, 'loss/train': 1.972228467464447} 01/28/2022 14:01:50 - INFO - codeparrot_training - Step 872: {'lr': 0.000218, 'samples': 167616, 'steps': 872, 'loss/train': 2.286787748336792} 01/28/2022 14:01:54 - INFO - codeparrot_training - Step 873: {'lr': 0.00021825, 'samples': 167808, 'steps': 873, 'loss/train': 2.0016286969184875} 01/28/2022 14:01:58 - INFO - codeparrot_training - Step 874: {'lr': 0.0002185, 'samples': 168000, 'steps': 874, 'loss/train': 1.9257962107658386} 01/28/2022 14:02:03 - INFO - codeparrot_training - Step 875: {'lr': 0.00021875, 'samples': 168192, 'steps': 875, 'loss/train': 2.281014561653137} 01/28/2022 14:02:07 - INFO - codeparrot_training - Step 876: {'lr': 0.000219, 'samples': 168384, 'steps': 876, 'loss/train': 2.629209816455841} 01/28/2022 14:02:13 - INFO - codeparrot_training - Step 877: {'lr': 0.00021925000000000002, 'samples': 168576, 'steps': 877, 'loss/train': 2.3392494320869446} 01/28/2022 14:02:17 - INFO - codeparrot_training - Step 878: {'lr': 0.0002195, 'samples': 168768, 'steps': 878, 'loss/train': 2.3887833952903748} 01/28/2022 14:02:22 - INFO - codeparrot_training - Step 879: {'lr': 0.00021975, 'samples': 168960, 'steps': 879, 'loss/train': 2.3446547985076904} 01/28/2022 14:02:26 - INFO - codeparrot_training - Step 880: {'lr': 0.00022, 'samples': 169152, 'steps': 880, 'loss/train': 0.7433227300643921} 01/28/2022 14:02:30 - INFO - codeparrot_training - Step 881: {'lr': 0.00022025000000000001, 'samples': 169344, 'steps': 881, 'loss/train': 2.5716254711151123} 01/28/2022 14:02:34 - INFO - codeparrot_training - Step 882: {'lr': 0.0002205, 'samples': 169536, 'steps': 882, 'loss/train': 2.768708825111389} 01/28/2022 14:02:39 - INFO - codeparrot_training - Step 883: {'lr': 0.00022075, 'samples': 169728, 'steps': 883, 'loss/train': 1.771878182888031} 01/28/2022 14:02:44 - INFO - codeparrot_training - Step 884: {'lr': 0.000221, 'samples': 169920, 'steps': 884, 'loss/train': 1.4879236221313477} 01/28/2022 14:02:48 - INFO - codeparrot_training - Step 885: {'lr': 0.00022125, 'samples': 170112, 'steps': 885, 'loss/train': 2.529706120491028} 01/28/2022 14:02:52 - INFO - codeparrot_training - Step 886: {'lr': 0.00022150000000000002, 'samples': 170304, 'steps': 886, 'loss/train': 1.6054351329803467} 01/28/2022 14:02:58 - INFO - codeparrot_training - Step 887: {'lr': 0.00022175, 'samples': 170496, 'steps': 887, 'loss/train': 2.47254341840744} 01/28/2022 14:03:02 - INFO - codeparrot_training - Step 888: {'lr': 0.000222, 'samples': 170688, 'steps': 888, 'loss/train': 2.7468937039375305} 01/28/2022 14:03:07 - INFO - codeparrot_training - Step 889: {'lr': 0.00022225, 'samples': 170880, 'steps': 889, 'loss/train': 1.199430227279663} 01/28/2022 14:03:11 - INFO - codeparrot_training - Step 890: {'lr': 0.00022250000000000001, 'samples': 171072, 'steps': 890, 'loss/train': 1.7912914752960205} 01/28/2022 14:03:15 - INFO - codeparrot_training - Step 891: {'lr': 0.00022275000000000002, 'samples': 171264, 'steps': 891, 'loss/train': 1.6902914643287659} 01/28/2022 14:03:20 - INFO - codeparrot_training - Step 892: {'lr': 0.000223, 'samples': 171456, 'steps': 892, 'loss/train': 2.4049689173698425} 01/28/2022 14:03:24 - INFO - codeparrot_training - Step 893: {'lr': 0.00022325, 'samples': 171648, 'steps': 893, 'loss/train': 1.3512644469738007} 01/28/2022 14:03:28 - INFO - codeparrot_training - Step 894: {'lr': 0.0002235, 'samples': 171840, 'steps': 894, 'loss/train': 1.334088295698166} 01/28/2022 14:03:33 - INFO - codeparrot_training - Step 895: {'lr': 0.00022375000000000002, 'samples': 172032, 'steps': 895, 'loss/train': 2.4908400177955627} 01/28/2022 14:03:37 - INFO - codeparrot_training - Step 896: {'lr': 0.000224, 'samples': 172224, 'steps': 896, 'loss/train': 1.4749424457550049} 01/28/2022 14:03:42 - INFO - codeparrot_training - Step 897: {'lr': 0.00022425, 'samples': 172416, 'steps': 897, 'loss/train': 1.5301001071929932} 01/28/2022 14:03:46 - INFO - codeparrot_training - Step 898: {'lr': 0.0002245, 'samples': 172608, 'steps': 898, 'loss/train': 0.41856318712234497} 01/28/2022 14:03:50 - INFO - codeparrot_training - Step 899: {'lr': 0.00022475000000000001, 'samples': 172800, 'steps': 899, 'loss/train': 3.038217544555664} 01/28/2022 14:03:54 - INFO - codeparrot_training - Step 900: {'lr': 0.00022500000000000002, 'samples': 172992, 'steps': 900, 'loss/train': 2.5771729946136475} 01/28/2022 14:03:59 - INFO - codeparrot_training - Step 901: {'lr': 0.00022525, 'samples': 173184, 'steps': 901, 'loss/train': 2.2650965452194214} 01/28/2022 14:04:05 - INFO - codeparrot_training - Step 902: {'lr': 0.0002255, 'samples': 173376, 'steps': 902, 'loss/train': 2.9238787293434143} 01/28/2022 14:04:09 - INFO - codeparrot_training - Step 903: {'lr': 0.00022575, 'samples': 173568, 'steps': 903, 'loss/train': 2.5059000849723816} 01/28/2022 14:04:13 - INFO - codeparrot_training - Step 904: {'lr': 0.00022600000000000002, 'samples': 173760, 'steps': 904, 'loss/train': 2.163037598133087} 01/28/2022 14:04:17 - INFO - codeparrot_training - Step 905: {'lr': 0.00022625000000000002, 'samples': 173952, 'steps': 905, 'loss/train': 1.7128804922103882} 01/28/2022 14:04:22 - INFO - codeparrot_training - Step 906: {'lr': 0.0002265, 'samples': 174144, 'steps': 906, 'loss/train': 1.6359697580337524} 01/28/2022 14:04:27 - INFO - codeparrot_training - Step 907: {'lr': 0.00022675, 'samples': 174336, 'steps': 907, 'loss/train': 2.070010185241699} 01/28/2022 14:04:31 - INFO - codeparrot_training - Step 908: {'lr': 0.00022700000000000002, 'samples': 174528, 'steps': 908, 'loss/train': 2.0243589878082275} 01/28/2022 14:04:35 - INFO - codeparrot_training - Step 909: {'lr': 0.00022725000000000002, 'samples': 174720, 'steps': 909, 'loss/train': 1.6672611236572266} 01/28/2022 14:04:39 - INFO - codeparrot_training - Step 910: {'lr': 0.0002275, 'samples': 174912, 'steps': 910, 'loss/train': 2.3686044216156006} 01/28/2022 14:04:43 - INFO - codeparrot_training - Step 911: {'lr': 0.00022775, 'samples': 175104, 'steps': 911, 'loss/train': 2.3944530487060547} 01/28/2022 14:04:50 - INFO - codeparrot_training - Step 912: {'lr': 0.000228, 'samples': 175296, 'steps': 912, 'loss/train': 2.105957329273224} 01/28/2022 14:04:54 - INFO - codeparrot_training - Step 913: {'lr': 0.00022825000000000002, 'samples': 175488, 'steps': 913, 'loss/train': 1.0788872838020325} 01/28/2022 14:04:58 - INFO - codeparrot_training - Step 914: {'lr': 0.00022850000000000002, 'samples': 175680, 'steps': 914, 'loss/train': 2.0780652165412903} 01/28/2022 14:05:02 - INFO - codeparrot_training - Step 915: {'lr': 0.00022875, 'samples': 175872, 'steps': 915, 'loss/train': 0.9196842312812805} 01/28/2022 14:05:07 - INFO - codeparrot_training - Step 916: {'lr': 0.000229, 'samples': 176064, 'steps': 916, 'loss/train': 0.39963389933109283} 01/28/2022 14:05:12 - INFO - codeparrot_training - Step 917: {'lr': 0.00022925000000000002, 'samples': 176256, 'steps': 917, 'loss/train': 2.257414698600769} 01/28/2022 14:05:16 - INFO - codeparrot_training - Step 918: {'lr': 0.00022950000000000002, 'samples': 176448, 'steps': 918, 'loss/train': 2.2410600185394287} 01/28/2022 14:05:20 - INFO - codeparrot_training - Step 919: {'lr': 0.00022975000000000003, 'samples': 176640, 'steps': 919, 'loss/train': 1.985464632511139} 01/28/2022 14:05:24 - INFO - codeparrot_training - Step 920: {'lr': 0.00023, 'samples': 176832, 'steps': 920, 'loss/train': 1.9434914588928223} 01/28/2022 14:05:28 - INFO - codeparrot_training - Step 921: {'lr': 0.00023025, 'samples': 177024, 'steps': 921, 'loss/train': 1.194972664117813} 01/28/2022 14:05:34 - INFO - codeparrot_training - Step 922: {'lr': 0.00023050000000000002, 'samples': 177216, 'steps': 922, 'loss/train': 2.7834200263023376} 01/28/2022 14:05:38 - INFO - codeparrot_training - Step 923: {'lr': 0.00023075000000000003, 'samples': 177408, 'steps': 923, 'loss/train': 2.036443769931793} 01/28/2022 14:05:42 - INFO - codeparrot_training - Step 924: {'lr': 0.000231, 'samples': 177600, 'steps': 924, 'loss/train': 1.9771839380264282} 01/28/2022 14:05:46 - INFO - codeparrot_training - Step 925: {'lr': 0.00023125, 'samples': 177792, 'steps': 925, 'loss/train': 2.4894189834594727} 01/28/2022 14:05:50 - INFO - codeparrot_training - Step 926: {'lr': 0.00023150000000000002, 'samples': 177984, 'steps': 926, 'loss/train': 2.2887418270111084} 01/28/2022 14:05:56 - INFO - codeparrot_training - Step 927: {'lr': 0.00023175000000000002, 'samples': 178176, 'steps': 927, 'loss/train': 1.5233822464942932} 01/28/2022 14:06:00 - INFO - codeparrot_training - Step 928: {'lr': 0.00023200000000000003, 'samples': 178368, 'steps': 928, 'loss/train': 2.016639232635498} 01/28/2022 14:06:04 - INFO - codeparrot_training - Step 929: {'lr': 0.00023225, 'samples': 178560, 'steps': 929, 'loss/train': 2.223813772201538} 01/28/2022 14:06:08 - INFO - codeparrot_training - Step 930: {'lr': 0.0002325, 'samples': 178752, 'steps': 930, 'loss/train': 2.816200911998749} 01/28/2022 14:06:12 - INFO - codeparrot_training - Step 931: {'lr': 0.00023275000000000002, 'samples': 178944, 'steps': 931, 'loss/train': 2.375095009803772} 01/28/2022 14:06:18 - INFO - codeparrot_training - Step 932: {'lr': 0.00023300000000000003, 'samples': 179136, 'steps': 932, 'loss/train': 1.4614740908145905} 01/28/2022 14:06:22 - INFO - codeparrot_training - Step 933: {'lr': 0.00023325, 'samples': 179328, 'steps': 933, 'loss/train': 2.3723902702331543} 01/28/2022 14:06:27 - INFO - codeparrot_training - Step 934: {'lr': 0.0002335, 'samples': 179520, 'steps': 934, 'loss/train': 2.4305396676063538} 01/28/2022 14:06:31 - INFO - codeparrot_training - Step 935: {'lr': 0.00023375000000000002, 'samples': 179712, 'steps': 935, 'loss/train': 1.8231922388076782} 01/28/2022 14:06:35 - INFO - codeparrot_training - Step 936: {'lr': 0.00023400000000000002, 'samples': 179904, 'steps': 936, 'loss/train': 1.6477219462394714} 01/28/2022 14:06:40 - INFO - codeparrot_training - Step 937: {'lr': 0.00023425000000000003, 'samples': 180096, 'steps': 937, 'loss/train': 2.3454229831695557} 01/28/2022 14:06:44 - INFO - codeparrot_training - Step 938: {'lr': 0.00023449999999999998, 'samples': 180288, 'steps': 938, 'loss/train': 2.403246223926544} 01/28/2022 14:06:48 - INFO - codeparrot_training - Step 939: {'lr': 0.00023475, 'samples': 180480, 'steps': 939, 'loss/train': 2.058785140514374} 01/28/2022 14:06:53 - INFO - codeparrot_training - Step 940: {'lr': 0.000235, 'samples': 180672, 'steps': 940, 'loss/train': 1.7430635690689087} 01/28/2022 14:06:57 - INFO - codeparrot_training - Step 941: {'lr': 0.00023525, 'samples': 180864, 'steps': 941, 'loss/train': 2.117285370826721} 01/28/2022 14:07:02 - INFO - codeparrot_training - Step 942: {'lr': 0.0002355, 'samples': 181056, 'steps': 942, 'loss/train': 2.130128860473633} 01/28/2022 14:07:06 - INFO - codeparrot_training - Step 943: {'lr': 0.00023574999999999998, 'samples': 181248, 'steps': 943, 'loss/train': 1.0488569140434265} 01/28/2022 14:07:10 - INFO - codeparrot_training - Step 944: {'lr': 0.000236, 'samples': 181440, 'steps': 944, 'loss/train': 1.8373521566390991} 01/28/2022 14:07:14 - INFO - codeparrot_training - Step 945: {'lr': 0.00023625, 'samples': 181632, 'steps': 945, 'loss/train': 1.7738630175590515} 01/28/2022 14:07:19 - INFO - codeparrot_training - Step 946: {'lr': 0.0002365, 'samples': 181824, 'steps': 946, 'loss/train': 2.400554895401001} 01/28/2022 14:07:25 - INFO - codeparrot_training - Step 947: {'lr': 0.00023674999999999998, 'samples': 182016, 'steps': 947, 'loss/train': 1.9406023621559143} 01/28/2022 14:07:29 - INFO - codeparrot_training - Step 948: {'lr': 0.000237, 'samples': 182208, 'steps': 948, 'loss/train': 1.090250015258789} 01/28/2022 14:07:33 - INFO - codeparrot_training - Step 949: {'lr': 0.00023725, 'samples': 182400, 'steps': 949, 'loss/train': 2.052117884159088} 01/28/2022 14:07:38 - INFO - codeparrot_training - Step 950: {'lr': 0.0002375, 'samples': 182592, 'steps': 950, 'loss/train': 1.788548469543457} 01/28/2022 14:07:42 - INFO - codeparrot_training - Step 951: {'lr': 0.00023775, 'samples': 182784, 'steps': 951, 'loss/train': 2.4749749302864075} 01/28/2022 14:07:47 - INFO - codeparrot_training - Step 952: {'lr': 0.00023799999999999998, 'samples': 182976, 'steps': 952, 'loss/train': 2.05472069978714} 01/28/2022 14:07:52 - INFO - codeparrot_training - Step 953: {'lr': 0.00023825, 'samples': 183168, 'steps': 953, 'loss/train': 2.237941861152649} 01/28/2022 14:07:56 - INFO - codeparrot_training - Step 954: {'lr': 0.0002385, 'samples': 183360, 'steps': 954, 'loss/train': 2.536879777908325} 01/28/2022 14:08:00 - INFO - codeparrot_training - Step 955: {'lr': 0.00023875, 'samples': 183552, 'steps': 955, 'loss/train': 2.827202260494232} 01/28/2022 14:08:04 - INFO - codeparrot_training - Step 956: {'lr': 0.00023899999999999998, 'samples': 183744, 'steps': 956, 'loss/train': 2.5681278705596924} 01/28/2022 14:08:10 - INFO - codeparrot_training - Step 957: {'lr': 0.00023925, 'samples': 183936, 'steps': 957, 'loss/train': 1.8842350244522095} 01/28/2022 14:08:14 - INFO - codeparrot_training - Step 958: {'lr': 0.0002395, 'samples': 184128, 'steps': 958, 'loss/train': 1.7167965173721313} 01/28/2022 14:08:19 - INFO - codeparrot_training - Step 959: {'lr': 0.00023975, 'samples': 184320, 'steps': 959, 'loss/train': 1.9357423782348633} 01/28/2022 14:08:23 - INFO - codeparrot_training - Step 960: {'lr': 0.00024, 'samples': 184512, 'steps': 960, 'loss/train': 3.437917113304138} 01/28/2022 14:08:27 - INFO - codeparrot_training - Step 961: {'lr': 0.00024024999999999999, 'samples': 184704, 'steps': 961, 'loss/train': 1.518826961517334} 01/28/2022 14:08:32 - INFO - codeparrot_training - Step 962: {'lr': 0.0002405, 'samples': 184896, 'steps': 962, 'loss/train': 2.97124582529068} 01/28/2022 14:08:36 - INFO - codeparrot_training - Step 963: {'lr': 0.00024075, 'samples': 185088, 'steps': 963, 'loss/train': 1.9603344798088074} 01/28/2022 14:08:41 - INFO - codeparrot_training - Step 964: {'lr': 0.000241, 'samples': 185280, 'steps': 964, 'loss/train': 2.3115926384925842} 01/28/2022 14:08:45 - INFO - codeparrot_training - Step 965: {'lr': 0.00024125, 'samples': 185472, 'steps': 965, 'loss/train': 2.642190098762512} 01/28/2022 14:08:49 - INFO - codeparrot_training - Step 966: {'lr': 0.0002415, 'samples': 185664, 'steps': 966, 'loss/train': 1.372547060251236} 01/28/2022 14:08:54 - INFO - codeparrot_training - Step 967: {'lr': 0.00024175, 'samples': 185856, 'steps': 967, 'loss/train': 0.9638555645942688} 01/28/2022 14:08:59 - INFO - codeparrot_training - Step 968: {'lr': 0.000242, 'samples': 186048, 'steps': 968, 'loss/train': 2.255031645298004} 01/28/2022 14:09:03 - INFO - codeparrot_training - Step 969: {'lr': 0.00024225, 'samples': 186240, 'steps': 969, 'loss/train': 1.170596569776535} 01/28/2022 14:09:07 - INFO - codeparrot_training - Step 970: {'lr': 0.00024249999999999999, 'samples': 186432, 'steps': 970, 'loss/train': 2.940527379512787} 01/28/2022 14:09:11 - INFO - codeparrot_training - Step 971: {'lr': 0.00024275, 'samples': 186624, 'steps': 971, 'loss/train': 0.1604963354766369} 01/28/2022 14:09:17 - INFO - codeparrot_training - Step 972: {'lr': 0.000243, 'samples': 186816, 'steps': 972, 'loss/train': 0.3481811285018921} 01/28/2022 14:09:21 - INFO - codeparrot_training - Step 973: {'lr': 0.00024325, 'samples': 187008, 'steps': 973, 'loss/train': 2.578470468521118} 01/28/2022 14:09:25 - INFO - codeparrot_training - Step 974: {'lr': 0.0002435, 'samples': 187200, 'steps': 974, 'loss/train': 2.761125683784485} 01/28/2022 14:09:29 - INFO - codeparrot_training - Step 975: {'lr': 0.00024375, 'samples': 187392, 'steps': 975, 'loss/train': 1.2085129022598267} 01/28/2022 14:09:33 - INFO - codeparrot_training - Step 976: {'lr': 0.000244, 'samples': 187584, 'steps': 976, 'loss/train': 2.203972578048706} 01/28/2022 14:09:40 - INFO - codeparrot_training - Step 977: {'lr': 0.00024425, 'samples': 187776, 'steps': 977, 'loss/train': 2.558237135410309} 01/28/2022 14:09:44 - INFO - codeparrot_training - Step 978: {'lr': 0.0002445, 'samples': 187968, 'steps': 978, 'loss/train': 2.3359014987945557} 01/28/2022 14:09:48 - INFO - codeparrot_training - Step 979: {'lr': 0.00024475, 'samples': 188160, 'steps': 979, 'loss/train': 2.514530897140503} 01/28/2022 14:09:52 - INFO - codeparrot_training - Step 980: {'lr': 0.000245, 'samples': 188352, 'steps': 980, 'loss/train': 2.0911161303520203} 01/28/2022 14:09:56 - INFO - codeparrot_training - Step 981: {'lr': 0.00024525, 'samples': 188544, 'steps': 981, 'loss/train': 2.1678850650787354} 01/28/2022 14:10:02 - INFO - codeparrot_training - Step 982: {'lr': 0.0002455, 'samples': 188736, 'steps': 982, 'loss/train': 1.8934271335601807} 01/28/2022 14:10:06 - INFO - codeparrot_training - Step 983: {'lr': 0.00024575, 'samples': 188928, 'steps': 983, 'loss/train': 1.7833684086799622} 01/28/2022 14:10:10 - INFO - codeparrot_training - Step 984: {'lr': 0.000246, 'samples': 189120, 'steps': 984, 'loss/train': 0.32062308490276337} 01/28/2022 14:10:14 - INFO - codeparrot_training - Step 985: {'lr': 0.00024625, 'samples': 189312, 'steps': 985, 'loss/train': 1.548620581626892} 01/28/2022 14:10:20 - INFO - codeparrot_training - Step 986: {'lr': 0.00024650000000000003, 'samples': 189504, 'steps': 986, 'loss/train': 2.373738169670105} 01/28/2022 14:10:24 - INFO - codeparrot_training - Step 987: {'lr': 0.00024675, 'samples': 189696, 'steps': 987, 'loss/train': 2.1187296509742737} 01/28/2022 14:10:28 - INFO - codeparrot_training - Step 988: {'lr': 0.000247, 'samples': 189888, 'steps': 988, 'loss/train': 2.4755111932754517} 01/28/2022 14:10:32 - INFO - codeparrot_training - Step 989: {'lr': 0.00024725, 'samples': 190080, 'steps': 989, 'loss/train': 2.541538417339325} 01/28/2022 14:10:36 - INFO - codeparrot_training - Step 990: {'lr': 0.0002475, 'samples': 190272, 'steps': 990, 'loss/train': 2.328869104385376} 01/28/2022 14:10:43 - INFO - codeparrot_training - Step 991: {'lr': 0.00024775, 'samples': 190464, 'steps': 991, 'loss/train': 1.8598485589027405} 01/28/2022 14:10:47 - INFO - codeparrot_training - Step 992: {'lr': 0.000248, 'samples': 190656, 'steps': 992, 'loss/train': 1.9700358510017395} 01/28/2022 14:10:51 - INFO - codeparrot_training - Step 993: {'lr': 0.00024825, 'samples': 190848, 'steps': 993, 'loss/train': 1.9378893971443176} 01/28/2022 14:10:55 - INFO - codeparrot_training - Step 994: {'lr': 0.0002485, 'samples': 191040, 'steps': 994, 'loss/train': 2.5420188903808594} 01/28/2022 14:10:59 - INFO - codeparrot_training - Step 995: {'lr': 0.00024875, 'samples': 191232, 'steps': 995, 'loss/train': 2.8138840198516846} 01/28/2022 14:11:04 - INFO - codeparrot_training - Step 996: {'lr': 0.000249, 'samples': 191424, 'steps': 996, 'loss/train': 1.794360637664795} 01/28/2022 14:11:09 - INFO - codeparrot_training - Step 997: {'lr': 0.00024925, 'samples': 191616, 'steps': 997, 'loss/train': 2.1718207597732544} 01/28/2022 14:11:13 - INFO - codeparrot_training - Step 998: {'lr': 0.0002495, 'samples': 191808, 'steps': 998, 'loss/train': 0.9597846865653992} 01/28/2022 14:11:17 - INFO - codeparrot_training - Step 999: {'lr': 0.00024975, 'samples': 192000, 'steps': 999, 'loss/train': 1.6574327945709229} 01/28/2022 14:11:21 - INFO - codeparrot_training - Step 1000: {'lr': 0.00025, 'samples': 192192, 'steps': 1000, 'loss/train': 0.39816977083683014} 01/28/2022 14:11:27 - INFO - codeparrot_training - Step 1001: {'lr': 0.00025025, 'samples': 192384, 'steps': 1001, 'loss/train': 1.7129468321800232} 01/28/2022 14:11:31 - INFO - codeparrot_training - Step 1002: {'lr': 0.0002505, 'samples': 192576, 'steps': 1002, 'loss/train': 2.7859137654304504} 01/28/2022 14:11:36 - INFO - codeparrot_training - Step 1003: {'lr': 0.00025075, 'samples': 192768, 'steps': 1003, 'loss/train': 1.851826250553131} 01/28/2022 14:11:40 - INFO - codeparrot_training - Step 1004: {'lr': 0.00025100000000000003, 'samples': 192960, 'steps': 1004, 'loss/train': 2.2837146520614624} 01/28/2022 14:11:44 - INFO - codeparrot_training - Step 1005: {'lr': 0.00025124999999999995, 'samples': 193152, 'steps': 1005, 'loss/train': 1.3877654671669006} 01/28/2022 14:11:49 - INFO - codeparrot_training - Step 1006: {'lr': 0.0002515, 'samples': 193344, 'steps': 1006, 'loss/train': 1.5297165513038635} 01/28/2022 14:11:53 - INFO - codeparrot_training - Step 1007: {'lr': 0.00025174999999999997, 'samples': 193536, 'steps': 1007, 'loss/train': 2.5541785955429077} 01/28/2022 14:11:57 - INFO - codeparrot_training - Step 1008: {'lr': 0.000252, 'samples': 193728, 'steps': 1008, 'loss/train': 3.121661424636841} 01/28/2022 14:12:01 - INFO - codeparrot_training - Step 1009: {'lr': 0.00025225, 'samples': 193920, 'steps': 1009, 'loss/train': 2.461627721786499} 01/28/2022 14:12:06 - INFO - codeparrot_training - Step 1010: {'lr': 0.0002525, 'samples': 194112, 'steps': 1010, 'loss/train': 2.26544451713562} 01/28/2022 14:12:11 - INFO - codeparrot_training - Step 1011: {'lr': 0.00025275, 'samples': 194304, 'steps': 1011, 'loss/train': 2.1724873781204224} 01/28/2022 14:12:15 - INFO - codeparrot_training - Step 1012: {'lr': 0.000253, 'samples': 194496, 'steps': 1012, 'loss/train': 2.791079342365265} 01/28/2022 14:12:19 - INFO - codeparrot_training - Step 1013: {'lr': 0.00025325, 'samples': 194688, 'steps': 1013, 'loss/train': 1.707982063293457} 01/28/2022 14:12:24 - INFO - codeparrot_training - Step 1014: {'lr': 0.0002535, 'samples': 194880, 'steps': 1014, 'loss/train': 1.7306089997291565} 01/28/2022 14:12:28 - INFO - codeparrot_training - Step 1015: {'lr': 0.00025374999999999996, 'samples': 195072, 'steps': 1015, 'loss/train': 3.1959818601608276} 01/28/2022 14:12:34 - INFO - codeparrot_training - Step 1016: {'lr': 0.000254, 'samples': 195264, 'steps': 1016, 'loss/train': 2.206725776195526} 01/28/2022 14:12:38 - INFO - codeparrot_training - Step 1017: {'lr': 0.00025425, 'samples': 195456, 'steps': 1017, 'loss/train': 2.1099168062210083} 01/28/2022 14:12:42 - INFO - codeparrot_training - Step 1018: {'lr': 0.0002545, 'samples': 195648, 'steps': 1018, 'loss/train': 2.631308913230896} 01/28/2022 14:12:46 - INFO - codeparrot_training - Step 1019: {'lr': 0.00025475, 'samples': 195840, 'steps': 1019, 'loss/train': 2.523587644100189} 01/28/2022 14:12:51 - INFO - codeparrot_training - Step 1020: {'lr': 0.000255, 'samples': 196032, 'steps': 1020, 'loss/train': 2.4522632360458374} 01/28/2022 14:12:56 - INFO - codeparrot_training - Step 1021: {'lr': 0.00025525, 'samples': 196224, 'steps': 1021, 'loss/train': 0.18219225481152534} 01/28/2022 14:13:00 - INFO - codeparrot_training - Step 1022: {'lr': 0.00025550000000000003, 'samples': 196416, 'steps': 1022, 'loss/train': 0.8894366025924683} 01/28/2022 14:13:04 - INFO - codeparrot_training - Step 1023: {'lr': 0.00025575, 'samples': 196608, 'steps': 1023, 'loss/train': 1.7165992856025696} 01/28/2022 14:13:08 - INFO - codeparrot_training - Step 1024: {'lr': 0.000256, 'samples': 196800, 'steps': 1024, 'loss/train': 2.0367663502693176} 01/28/2022 14:13:12 - INFO - codeparrot_training - Step 1025: {'lr': 0.00025624999999999997, 'samples': 196992, 'steps': 1025, 'loss/train': 2.3185946345329285} 01/28/2022 14:13:17 - INFO - codeparrot_training - Step 1026: {'lr': 0.0002565, 'samples': 197184, 'steps': 1026, 'loss/train': 1.4287159144878387} 01/28/2022 14:13:22 - INFO - codeparrot_training - Step 1027: {'lr': 0.00025675, 'samples': 197376, 'steps': 1027, 'loss/train': 1.1328458189964294} 01/28/2022 14:13:26 - INFO - codeparrot_training - Step 1028: {'lr': 0.000257, 'samples': 197568, 'steps': 1028, 'loss/train': 2.249435842037201} 01/28/2022 14:13:30 - INFO - codeparrot_training - Step 1029: {'lr': 0.00025725, 'samples': 197760, 'steps': 1029, 'loss/train': 2.085183799266815} 01/28/2022 14:13:34 - INFO - codeparrot_training - Step 1030: {'lr': 0.0002575, 'samples': 197952, 'steps': 1030, 'loss/train': 2.149496912956238} 01/28/2022 14:13:40 - INFO - codeparrot_training - Step 1031: {'lr': 0.00025775, 'samples': 198144, 'steps': 1031, 'loss/train': 1.037769466638565} 01/28/2022 14:13:44 - INFO - codeparrot_training - Step 1032: {'lr': 0.00025800000000000004, 'samples': 198336, 'steps': 1032, 'loss/train': 1.3776796460151672} 01/28/2022 14:13:48 - INFO - codeparrot_training - Step 1033: {'lr': 0.00025824999999999996, 'samples': 198528, 'steps': 1033, 'loss/train': 1.66482013463974} 01/28/2022 14:13:52 - INFO - codeparrot_training - Step 1034: {'lr': 0.0002585, 'samples': 198720, 'steps': 1034, 'loss/train': 2.2700196504592896} 01/28/2022 14:13:58 - INFO - codeparrot_training - Step 1035: {'lr': 0.00025875, 'samples': 198912, 'steps': 1035, 'loss/train': 2.36336213350296} 01/28/2022 14:14:02 - INFO - codeparrot_training - Step 1036: {'lr': 0.000259, 'samples': 199104, 'steps': 1036, 'loss/train': 1.6398085355758667} 01/28/2022 14:14:07 - INFO - codeparrot_training - Step 1037: {'lr': 0.00025925, 'samples': 199296, 'steps': 1037, 'loss/train': 1.8467423915863037} 01/28/2022 14:14:11 - INFO - codeparrot_training - Step 1038: {'lr': 0.0002595, 'samples': 199488, 'steps': 1038, 'loss/train': 2.249245762825012} 01/28/2022 14:14:15 - INFO - codeparrot_training - Step 1039: {'lr': 0.00025975, 'samples': 199680, 'steps': 1039, 'loss/train': 1.9053810238838196} 01/28/2022 14:14:20 - INFO - codeparrot_training - Step 1040: {'lr': 0.00026000000000000003, 'samples': 199872, 'steps': 1040, 'loss/train': 0.5708346515893936} 01/28/2022 14:14:24 - INFO - codeparrot_training - Step 1041: {'lr': 0.00026025, 'samples': 200064, 'steps': 1041, 'loss/train': 2.495672821998596} 01/28/2022 14:14:29 - INFO - codeparrot_training - Step 1042: {'lr': 0.0002605, 'samples': 200256, 'steps': 1042, 'loss/train': 3.237251043319702} 01/28/2022 14:14:33 - INFO - codeparrot_training - Step 1043: {'lr': 0.00026074999999999997, 'samples': 200448, 'steps': 1043, 'loss/train': 2.2887749075889587} 01/28/2022 14:14:37 - INFO - codeparrot_training - Step 1044: {'lr': 0.000261, 'samples': 200640, 'steps': 1044, 'loss/train': 2.2949538230895996} 01/28/2022 14:14:42 - INFO - codeparrot_training - Step 1045: {'lr': 0.00026125, 'samples': 200832, 'steps': 1045, 'loss/train': 2.2994985580444336} 01/28/2022 14:14:46 - INFO - codeparrot_training - Step 1046: {'lr': 0.0002615, 'samples': 201024, 'steps': 1046, 'loss/train': 0.7112371176481247} 01/28/2022 14:14:50 - INFO - codeparrot_training - Step 1047: {'lr': 0.00026175, 'samples': 201216, 'steps': 1047, 'loss/train': 2.6505868434906006} 01/28/2022 14:14:55 - INFO - codeparrot_training - Step 1048: {'lr': 0.000262, 'samples': 201408, 'steps': 1048, 'loss/train': 1.562029480934143} 01/28/2022 14:14:59 - INFO - codeparrot_training - Step 1049: {'lr': 0.00026225, 'samples': 201600, 'steps': 1049, 'loss/train': 2.6525318026542664} 01/28/2022 14:15:07 - INFO - codeparrot_training - Step 1050: {'lr': 0.00026250000000000004, 'samples': 201792, 'steps': 1050, 'loss/train': 1.9723582863807678} 01/28/2022 14:15:11 - INFO - codeparrot_training - Step 1051: {'lr': 0.00026274999999999996, 'samples': 201984, 'steps': 1051, 'loss/train': 1.8843500018119812} 01/28/2022 14:15:15 - INFO - codeparrot_training - Step 1052: {'lr': 0.000263, 'samples': 202176, 'steps': 1052, 'loss/train': 1.870796263217926} 01/28/2022 14:15:19 - INFO - codeparrot_training - Step 1053: {'lr': 0.00026325, 'samples': 202368, 'steps': 1053, 'loss/train': 2.414433002471924} 01/28/2022 14:15:23 - INFO - codeparrot_training - Step 1054: {'lr': 0.0002635, 'samples': 202560, 'steps': 1054, 'loss/train': 2.5134161710739136} 01/28/2022 14:15:28 - INFO - codeparrot_training - Step 1055: {'lr': 0.00026375, 'samples': 202752, 'steps': 1055, 'loss/train': 1.9179829359054565} 01/28/2022 14:15:33 - INFO - codeparrot_training - Step 1056: {'lr': 0.000264, 'samples': 202944, 'steps': 1056, 'loss/train': 1.8453800082206726} 01/28/2022 14:15:37 - INFO - codeparrot_training - Step 1057: {'lr': 0.00026425, 'samples': 203136, 'steps': 1057, 'loss/train': 1.4905361831188202} 01/28/2022 14:15:41 - INFO - codeparrot_training - Step 1058: {'lr': 0.00026450000000000003, 'samples': 203328, 'steps': 1058, 'loss/train': 2.128647029399872} 01/28/2022 14:15:45 - INFO - codeparrot_training - Step 1059: {'lr': 0.00026475, 'samples': 203520, 'steps': 1059, 'loss/train': 2.187892735004425} 01/28/2022 14:15:51 - INFO - codeparrot_training - Step 1060: {'lr': 0.00026500000000000004, 'samples': 203712, 'steps': 1060, 'loss/train': 2.412875533103943} 01/28/2022 14:15:56 - INFO - codeparrot_training - Step 1061: {'lr': 0.00026524999999999997, 'samples': 203904, 'steps': 1061, 'loss/train': 1.9132726192474365} 01/28/2022 14:16:00 - INFO - codeparrot_training - Step 1062: {'lr': 0.0002655, 'samples': 204096, 'steps': 1062, 'loss/train': 1.0908856093883514} 01/28/2022 14:16:04 - INFO - codeparrot_training - Step 1063: {'lr': 0.00026575, 'samples': 204288, 'steps': 1063, 'loss/train': 1.9516666531562805} 01/28/2022 14:16:08 - INFO - codeparrot_training - Step 1064: {'lr': 0.000266, 'samples': 204480, 'steps': 1064, 'loss/train': 2.0072713494300842} 01/28/2022 14:16:13 - INFO - codeparrot_training - Step 1065: {'lr': 0.00026625, 'samples': 204672, 'steps': 1065, 'loss/train': 3.664784073829651} 01/28/2022 14:16:17 - INFO - codeparrot_training - Step 1066: {'lr': 0.0002665, 'samples': 204864, 'steps': 1066, 'loss/train': 1.419552505016327} 01/28/2022 14:16:22 - INFO - codeparrot_training - Step 1067: {'lr': 0.00026675, 'samples': 205056, 'steps': 1067, 'loss/train': 0.5007616728544235} 01/28/2022 14:16:26 - INFO - codeparrot_training - Step 1068: {'lr': 0.00026700000000000004, 'samples': 205248, 'steps': 1068, 'loss/train': 2.9083399772644043} 01/28/2022 14:16:30 - INFO - codeparrot_training - Step 1069: {'lr': 0.00026725, 'samples': 205440, 'steps': 1069, 'loss/train': 1.7998466491699219} 01/28/2022 14:16:35 - INFO - codeparrot_training - Step 1070: {'lr': 0.0002675, 'samples': 205632, 'steps': 1070, 'loss/train': 1.8166877031326294} 01/28/2022 14:16:39 - INFO - codeparrot_training - Step 1071: {'lr': 0.00026775, 'samples': 205824, 'steps': 1071, 'loss/train': 2.2733264565467834} 01/28/2022 14:16:43 - INFO - codeparrot_training - Step 1072: {'lr': 0.000268, 'samples': 206016, 'steps': 1072, 'loss/train': 1.6790778636932373} 01/28/2022 14:16:48 - INFO - codeparrot_training - Step 1073: {'lr': 0.00026825, 'samples': 206208, 'steps': 1073, 'loss/train': 0.6474594175815582} 01/28/2022 14:16:52 - INFO - codeparrot_training - Step 1074: {'lr': 0.0002685, 'samples': 206400, 'steps': 1074, 'loss/train': 3.7161948680877686} 01/28/2022 14:16:58 - INFO - codeparrot_training - Step 1075: {'lr': 0.00026875, 'samples': 206592, 'steps': 1075, 'loss/train': 2.4095218777656555} 01/28/2022 14:17:02 - INFO - codeparrot_training - Step 1076: {'lr': 0.00026900000000000003, 'samples': 206784, 'steps': 1076, 'loss/train': 1.8568680882453918} 01/28/2022 14:17:06 - INFO - codeparrot_training - Step 1077: {'lr': 0.00026925, 'samples': 206976, 'steps': 1077, 'loss/train': 1.1406211853027344} 01/28/2022 14:17:11 - INFO - codeparrot_training - Step 1078: {'lr': 0.00026950000000000005, 'samples': 207168, 'steps': 1078, 'loss/train': 0.6243577748537064} 01/28/2022 14:17:15 - INFO - codeparrot_training - Step 1079: {'lr': 0.00026974999999999997, 'samples': 207360, 'steps': 1079, 'loss/train': 2.3352359533309937} 01/28/2022 14:17:20 - INFO - codeparrot_training - Step 1080: {'lr': 0.00027, 'samples': 207552, 'steps': 1080, 'loss/train': 1.447355031967163} 01/28/2022 14:17:24 - INFO - codeparrot_training - Step 1081: {'lr': 0.00027025, 'samples': 207744, 'steps': 1081, 'loss/train': 2.088889181613922} 01/28/2022 14:17:29 - INFO - codeparrot_training - Step 1082: {'lr': 0.0002705, 'samples': 207936, 'steps': 1082, 'loss/train': 2.0433009266853333} 01/28/2022 14:17:33 - INFO - codeparrot_training - Step 1083: {'lr': 0.00027075, 'samples': 208128, 'steps': 1083, 'loss/train': 2.8213759660720825} 01/28/2022 14:17:37 - INFO - codeparrot_training - Step 1084: {'lr': 0.00027100000000000003, 'samples': 208320, 'steps': 1084, 'loss/train': 1.7189303040504456} 01/28/2022 14:17:42 - INFO - codeparrot_training - Step 1085: {'lr': 0.00027125, 'samples': 208512, 'steps': 1085, 'loss/train': 1.8216029405593872} 01/28/2022 14:17:46 - INFO - codeparrot_training - Step 1086: {'lr': 0.00027150000000000004, 'samples': 208704, 'steps': 1086, 'loss/train': 2.133514881134033} 01/28/2022 14:17:50 - INFO - codeparrot_training - Step 1087: {'lr': 0.00027175, 'samples': 208896, 'steps': 1087, 'loss/train': 2.365162432193756} 01/28/2022 14:17:54 - INFO - codeparrot_training - Step 1088: {'lr': 0.00027200000000000005, 'samples': 209088, 'steps': 1088, 'loss/train': 2.854024350643158} 01/28/2022 14:17:59 - INFO - codeparrot_training - Step 1089: {'lr': 0.00027225, 'samples': 209280, 'steps': 1089, 'loss/train': 2.615595281124115} 01/28/2022 14:18:04 - INFO - codeparrot_training - Step 1090: {'lr': 0.0002725, 'samples': 209472, 'steps': 1090, 'loss/train': 2.619783103466034} 01/28/2022 14:18:08 - INFO - codeparrot_training - Step 1091: {'lr': 0.00027275, 'samples': 209664, 'steps': 1091, 'loss/train': 2.2841153740882874} 01/28/2022 14:18:12 - INFO - codeparrot_training - Step 1092: {'lr': 0.000273, 'samples': 209856, 'steps': 1092, 'loss/train': 1.7424763441085815} 01/28/2022 14:18:16 - INFO - codeparrot_training - Step 1093: {'lr': 0.00027325, 'samples': 210048, 'steps': 1093, 'loss/train': 1.6202677488327026} 01/28/2022 14:18:20 - INFO - codeparrot_training - Step 1094: {'lr': 0.00027350000000000003, 'samples': 210240, 'steps': 1094, 'loss/train': 1.5558868646621704} 01/28/2022 14:18:27 - INFO - codeparrot_training - Step 1095: {'lr': 0.00027375, 'samples': 210432, 'steps': 1095, 'loss/train': 2.5102985501289368} 01/28/2022 14:18:31 - INFO - codeparrot_training - Step 1096: {'lr': 0.00027400000000000005, 'samples': 210624, 'steps': 1096, 'loss/train': 0.7001370638608932} 01/28/2022 14:18:35 - INFO - codeparrot_training - Step 1097: {'lr': 0.00027425, 'samples': 210816, 'steps': 1097, 'loss/train': 1.6671651005744934} 01/28/2022 14:18:39 - INFO - codeparrot_training - Step 1098: {'lr': 0.0002745, 'samples': 211008, 'steps': 1098, 'loss/train': 1.396962583065033} 01/28/2022 14:18:43 - INFO - codeparrot_training - Step 1099: {'lr': 0.00027475, 'samples': 211200, 'steps': 1099, 'loss/train': 2.585107684135437} 01/28/2022 14:18:48 - INFO - codeparrot_training - Step 1100: {'lr': 0.000275, 'samples': 211392, 'steps': 1100, 'loss/train': 1.5152605175971985} 01/28/2022 14:18:53 - INFO - codeparrot_training - Step 1101: {'lr': 0.00027525, 'samples': 211584, 'steps': 1101, 'loss/train': 2.3280705213546753} 01/28/2022 14:18:57 - INFO - codeparrot_training - Step 1102: {'lr': 0.00027550000000000003, 'samples': 211776, 'steps': 1102, 'loss/train': 2.813526928424835} 01/28/2022 14:19:01 - INFO - codeparrot_training - Step 1103: {'lr': 0.00027575, 'samples': 211968, 'steps': 1103, 'loss/train': 2.0946906208992004} 01/28/2022 14:19:05 - INFO - codeparrot_training - Step 1104: {'lr': 0.00027600000000000004, 'samples': 212160, 'steps': 1104, 'loss/train': 2.3015263080596924} 01/28/2022 14:19:11 - INFO - codeparrot_training - Step 1105: {'lr': 0.00027625, 'samples': 212352, 'steps': 1105, 'loss/train': 1.0499154031276703} 01/28/2022 14:19:15 - INFO - codeparrot_training - Step 1106: {'lr': 0.00027650000000000005, 'samples': 212544, 'steps': 1106, 'loss/train': 1.5928512811660767} 01/28/2022 14:19:20 - INFO - codeparrot_training - Step 1107: {'lr': 0.00027675, 'samples': 212736, 'steps': 1107, 'loss/train': 0.6067784875631332} 01/28/2022 14:19:24 - INFO - codeparrot_training - Step 1108: {'lr': 0.000277, 'samples': 212928, 'steps': 1108, 'loss/train': 2.222423493862152} 01/28/2022 14:19:28 - INFO - codeparrot_training - Step 1109: {'lr': 0.00027725, 'samples': 213120, 'steps': 1109, 'loss/train': 1.2546564638614655} 01/28/2022 14:19:33 - INFO - codeparrot_training - Step 1110: {'lr': 0.0002775, 'samples': 213312, 'steps': 1110, 'loss/train': 2.0215718150138855} 01/28/2022 14:19:37 - INFO - codeparrot_training - Step 1111: {'lr': 0.00027775, 'samples': 213504, 'steps': 1111, 'loss/train': 2.1070436239242554} 01/28/2022 14:19:41 - INFO - codeparrot_training - Step 1112: {'lr': 0.00027800000000000004, 'samples': 213696, 'steps': 1112, 'loss/train': 1.6528712511062622} 01/28/2022 14:19:46 - INFO - codeparrot_training - Step 1113: {'lr': 0.00027825, 'samples': 213888, 'steps': 1113, 'loss/train': 1.187238872051239} 01/28/2022 14:19:50 - INFO - codeparrot_training - Step 1114: {'lr': 0.00027850000000000005, 'samples': 214080, 'steps': 1114, 'loss/train': 1.9761300086975098} 01/28/2022 14:19:55 - INFO - codeparrot_training - Step 1115: {'lr': 0.00027875, 'samples': 214272, 'steps': 1115, 'loss/train': 4.130910515785217} 01/28/2022 14:19:59 - INFO - codeparrot_training - Step 1116: {'lr': 0.000279, 'samples': 214464, 'steps': 1116, 'loss/train': 1.7521902322769165} 01/28/2022 14:20:03 - INFO - codeparrot_training - Step 1117: {'lr': 0.00027925, 'samples': 214656, 'steps': 1117, 'loss/train': 3.095660448074341} 01/28/2022 14:20:07 - INFO - codeparrot_training - Step 1118: {'lr': 0.0002795, 'samples': 214848, 'steps': 1118, 'loss/train': 2.4569742679595947} 01/28/2022 14:20:12 - INFO - codeparrot_training - Step 1119: {'lr': 0.00027975, 'samples': 215040, 'steps': 1119, 'loss/train': 2.2170812487602234} 01/28/2022 14:20:18 - INFO - codeparrot_training - Step 1120: {'lr': 0.00028000000000000003, 'samples': 215232, 'steps': 1120, 'loss/train': 1.9206604957580566} 01/28/2022 14:20:22 - INFO - codeparrot_training - Step 1121: {'lr': 0.00028025, 'samples': 215424, 'steps': 1121, 'loss/train': 2.1181769371032715} 01/28/2022 14:20:27 - INFO - codeparrot_training - Step 1122: {'lr': 0.00028050000000000004, 'samples': 215616, 'steps': 1122, 'loss/train': 2.394699454307556} 01/28/2022 14:20:31 - INFO - codeparrot_training - Step 1123: {'lr': 0.00028075, 'samples': 215808, 'steps': 1123, 'loss/train': 2.0681358575820923} 01/28/2022 14:20:35 - INFO - codeparrot_training - Step 1124: {'lr': 0.00028100000000000005, 'samples': 216000, 'steps': 1124, 'loss/train': 2.164112627506256} 01/28/2022 14:20:40 - INFO - codeparrot_training - Step 1125: {'lr': 0.00028125000000000003, 'samples': 216192, 'steps': 1125, 'loss/train': 2.4953566789627075} 01/28/2022 14:20:44 - INFO - codeparrot_training - Step 1126: {'lr': 0.00028149999999999996, 'samples': 216384, 'steps': 1126, 'loss/train': 2.0171204209327698} 01/28/2022 14:20:48 - INFO - codeparrot_training - Step 1127: {'lr': 0.00028175, 'samples': 216576, 'steps': 1127, 'loss/train': 1.6230475902557373} 01/28/2022 14:20:53 - INFO - codeparrot_training - Step 1128: {'lr': 0.00028199999999999997, 'samples': 216768, 'steps': 1128, 'loss/train': 1.9569597244262695} 01/28/2022 14:20:57 - INFO - codeparrot_training - Step 1129: {'lr': 0.00028225, 'samples': 216960, 'steps': 1129, 'loss/train': 1.137268602848053} 01/28/2022 14:21:02 - INFO - codeparrot_training - Step 1130: {'lr': 0.0002825, 'samples': 217152, 'steps': 1130, 'loss/train': 1.9645442962646484} 01/28/2022 14:21:06 - INFO - codeparrot_training - Step 1131: {'lr': 0.00028275, 'samples': 217344, 'steps': 1131, 'loss/train': 1.6581682562828064} 01/28/2022 14:21:11 - INFO - codeparrot_training - Step 1132: {'lr': 0.000283, 'samples': 217536, 'steps': 1132, 'loss/train': 1.6082215905189514} 01/28/2022 14:21:15 - INFO - codeparrot_training - Step 1133: {'lr': 0.00028325000000000003, 'samples': 217728, 'steps': 1133, 'loss/train': 1.9482829570770264} 01/28/2022 14:21:23 - INFO - codeparrot_training - Step 1134: {'lr': 0.0002835, 'samples': 217920, 'steps': 1134, 'loss/train': 1.6671114563941956} 01/28/2022 14:21:27 - INFO - codeparrot_training - Step 1135: {'lr': 0.00028375, 'samples': 218112, 'steps': 1135, 'loss/train': 1.4944468438625336} 01/28/2022 14:21:31 - INFO - codeparrot_training - Step 1136: {'lr': 0.00028399999999999996, 'samples': 218304, 'steps': 1136, 'loss/train': 1.7960123419761658} 01/28/2022 14:21:35 - INFO - codeparrot_training - Step 1137: {'lr': 0.00028425, 'samples': 218496, 'steps': 1137, 'loss/train': 1.9311304092407227} 01/28/2022 14:21:39 - INFO - codeparrot_training - Step 1138: {'lr': 0.0002845, 'samples': 218688, 'steps': 1138, 'loss/train': 2.031345248222351} 01/28/2022 14:21:45 - INFO - codeparrot_training - Step 1139: {'lr': 0.00028475, 'samples': 218880, 'steps': 1139, 'loss/train': 2.0228909254074097} 01/28/2022 14:21:49 - INFO - codeparrot_training - Step 1140: {'lr': 0.000285, 'samples': 219072, 'steps': 1140, 'loss/train': 2.398287355899811} 01/28/2022 14:21:53 - INFO - codeparrot_training - Step 1141: {'lr': 0.00028525, 'samples': 219264, 'steps': 1141, 'loss/train': 1.9536155462265015} 01/28/2022 14:21:57 - INFO - codeparrot_training - Step 1142: {'lr': 0.0002855, 'samples': 219456, 'steps': 1142, 'loss/train': 2.905414581298828} 01/28/2022 14:22:01 - INFO - codeparrot_training - Step 1143: {'lr': 0.00028575000000000003, 'samples': 219648, 'steps': 1143, 'loss/train': 2.4513882994651794} 01/28/2022 14:22:07 - INFO - codeparrot_training - Step 1144: {'lr': 0.00028599999999999996, 'samples': 219840, 'steps': 1144, 'loss/train': 1.4090296626091003} 01/28/2022 14:22:11 - INFO - codeparrot_training - Step 1145: {'lr': 0.00028625, 'samples': 220032, 'steps': 1145, 'loss/train': 2.8324883580207825} 01/28/2022 14:22:15 - INFO - codeparrot_training - Step 1146: {'lr': 0.00028649999999999997, 'samples': 220224, 'steps': 1146, 'loss/train': 1.5323025584220886} 01/28/2022 14:22:19 - INFO - codeparrot_training - Step 1147: {'lr': 0.00028675, 'samples': 220416, 'steps': 1147, 'loss/train': 1.512565791606903} 01/28/2022 14:22:23 - INFO - codeparrot_training - Step 1148: {'lr': 0.000287, 'samples': 220608, 'steps': 1148, 'loss/train': 2.2020649909973145} 01/28/2022 14:22:31 - INFO - codeparrot_training - Step 1149: {'lr': 0.00028725, 'samples': 220800, 'steps': 1149, 'loss/train': 2.463794231414795} 01/28/2022 14:22:35 - INFO - codeparrot_training - Step 1150: {'lr': 0.0002875, 'samples': 220992, 'steps': 1150, 'loss/train': 2.0797138810157776} 01/28/2022 14:22:39 - INFO - codeparrot_training - Step 1151: {'lr': 0.00028775000000000003, 'samples': 221184, 'steps': 1151, 'loss/train': 1.522637128829956} 01/28/2022 14:22:43 - INFO - codeparrot_training - Step 1152: {'lr': 0.000288, 'samples': 221376, 'steps': 1152, 'loss/train': 2.42619788646698} 01/28/2022 14:22:47 - INFO - codeparrot_training - Step 1153: {'lr': 0.00028825, 'samples': 221568, 'steps': 1153, 'loss/train': 1.9129343032836914} 01/28/2022 14:22:51 - INFO - codeparrot_training - Step 1154: {'lr': 0.00028849999999999997, 'samples': 221760, 'steps': 1154, 'loss/train': 1.8711331486701965} 01/28/2022 14:22:56 - INFO - codeparrot_training - Step 1155: {'lr': 0.00028875, 'samples': 221952, 'steps': 1155, 'loss/train': 1.4060055613517761} 01/28/2022 14:23:01 - INFO - codeparrot_training - Step 1156: {'lr': 0.000289, 'samples': 222144, 'steps': 1156, 'loss/train': 2.0875688195228577} 01/28/2022 14:23:05 - INFO - codeparrot_training - Step 1157: {'lr': 0.00028925, 'samples': 222336, 'steps': 1157, 'loss/train': 1.3821549117565155} 01/28/2022 14:23:09 - INFO - codeparrot_training - Step 1158: {'lr': 0.0002895, 'samples': 222528, 'steps': 1158, 'loss/train': 1.9109625220298767} 01/28/2022 14:23:13 - INFO - codeparrot_training - Step 1159: {'lr': 0.00028975, 'samples': 222720, 'steps': 1159, 'loss/train': 0.9247423410415649} 01/28/2022 14:23:18 - INFO - codeparrot_training - Step 1160: {'lr': 0.00029, 'samples': 222912, 'steps': 1160, 'loss/train': 1.7462854385375977} 01/28/2022 14:23:23 - INFO - codeparrot_training - Step 1161: {'lr': 0.00029025000000000003, 'samples': 223104, 'steps': 1161, 'loss/train': 2.6918141841888428} 01/28/2022 14:23:27 - INFO - codeparrot_training - Step 1162: {'lr': 0.00029049999999999996, 'samples': 223296, 'steps': 1162, 'loss/train': 2.018281102180481} 01/28/2022 14:23:31 - INFO - codeparrot_training - Step 1163: {'lr': 0.00029075, 'samples': 223488, 'steps': 1163, 'loss/train': 2.0149283409118652} 01/28/2022 14:23:35 - INFO - codeparrot_training - Step 1164: {'lr': 0.00029099999999999997, 'samples': 223680, 'steps': 1164, 'loss/train': 2.026932120323181} 01/28/2022 14:23:41 - INFO - codeparrot_training - Step 1165: {'lr': 0.00029125, 'samples': 223872, 'steps': 1165, 'loss/train': 1.9492735862731934} 01/28/2022 14:23:46 - INFO - codeparrot_training - Step 1166: {'lr': 0.0002915, 'samples': 224064, 'steps': 1166, 'loss/train': 2.5035839080810547} 01/28/2022 14:23:50 - INFO - codeparrot_training - Step 1167: {'lr': 0.00029175, 'samples': 224256, 'steps': 1167, 'loss/train': 0.8822243213653564} 01/28/2022 14:23:54 - INFO - codeparrot_training - Step 1168: {'lr': 0.000292, 'samples': 224448, 'steps': 1168, 'loss/train': 0.12905460968613625} 01/28/2022 14:23:58 - INFO - codeparrot_training - Step 1169: {'lr': 0.00029225000000000003, 'samples': 224640, 'steps': 1169, 'loss/train': 1.9392811059951782} 01/28/2022 14:24:04 - INFO - codeparrot_training - Step 1170: {'lr': 0.0002925, 'samples': 224832, 'steps': 1170, 'loss/train': 1.5546244382858276} 01/28/2022 14:24:08 - INFO - codeparrot_training - Step 1171: {'lr': 0.00029275000000000004, 'samples': 225024, 'steps': 1171, 'loss/train': 1.1951023936271667} 01/28/2022 14:24:12 - INFO - codeparrot_training - Step 1172: {'lr': 0.00029299999999999997, 'samples': 225216, 'steps': 1172, 'loss/train': 0.8649976551532745} 01/28/2022 14:24:16 - INFO - codeparrot_training - Step 1173: {'lr': 0.00029325, 'samples': 225408, 'steps': 1173, 'loss/train': 2.031085968017578} 01/28/2022 14:24:20 - INFO - codeparrot_training - Step 1174: {'lr': 0.0002935, 'samples': 225600, 'steps': 1174, 'loss/train': 2.2030193209648132} 01/28/2022 14:24:25 - INFO - codeparrot_training - Step 1175: {'lr': 0.00029375, 'samples': 225792, 'steps': 1175, 'loss/train': 2.079620897769928} 01/28/2022 14:24:30 - INFO - codeparrot_training - Step 1176: {'lr': 0.000294, 'samples': 225984, 'steps': 1176, 'loss/train': 1.3593385219573975} 01/28/2022 14:24:34 - INFO - codeparrot_training - Step 1177: {'lr': 0.00029425, 'samples': 226176, 'steps': 1177, 'loss/train': 1.4963766932487488} 01/28/2022 14:24:38 - INFO - codeparrot_training - Step 1178: {'lr': 0.0002945, 'samples': 226368, 'steps': 1178, 'loss/train': 1.9959785342216492} 01/28/2022 14:24:42 - INFO - codeparrot_training - Step 1179: {'lr': 0.00029475000000000004, 'samples': 226560, 'steps': 1179, 'loss/train': 1.028979331254959} 01/28/2022 14:24:49 - INFO - codeparrot_training - Step 1180: {'lr': 0.000295, 'samples': 226752, 'steps': 1180, 'loss/train': 2.072916805744171} 01/28/2022 14:24:53 - INFO - codeparrot_training - Step 1181: {'lr': 0.00029525, 'samples': 226944, 'steps': 1181, 'loss/train': 2.045894980430603} 01/28/2022 14:24:57 - INFO - codeparrot_training - Step 1182: {'lr': 0.00029549999999999997, 'samples': 227136, 'steps': 1182, 'loss/train': 2.026222050189972} 01/28/2022 14:25:01 - INFO - codeparrot_training - Step 1183: {'lr': 0.00029575, 'samples': 227328, 'steps': 1183, 'loss/train': 2.2012311816215515} 01/28/2022 14:25:07 - INFO - codeparrot_training - Step 1184: {'lr': 0.000296, 'samples': 227520, 'steps': 1184, 'loss/train': 2.0714197754859924} 01/28/2022 14:25:11 - INFO - codeparrot_training - Step 1185: {'lr': 0.00029625, 'samples': 227712, 'steps': 1185, 'loss/train': 2.506234645843506} 01/28/2022 14:25:15 - INFO - codeparrot_training - Step 1186: {'lr': 0.0002965, 'samples': 227904, 'steps': 1186, 'loss/train': 1.7073501348495483} 01/28/2022 14:25:19 - INFO - codeparrot_training - Step 1187: {'lr': 0.00029675000000000003, 'samples': 228096, 'steps': 1187, 'loss/train': 2.3875452876091003} 01/28/2022 14:25:23 - INFO - codeparrot_training - Step 1188: {'lr': 0.000297, 'samples': 228288, 'steps': 1188, 'loss/train': 2.2774563431739807} 01/28/2022 14:25:27 - INFO - codeparrot_training - Step 1189: {'lr': 0.00029725000000000004, 'samples': 228480, 'steps': 1189, 'loss/train': 2.316476047039032} 01/28/2022 14:25:34 - INFO - codeparrot_training - Step 1190: {'lr': 0.00029749999999999997, 'samples': 228672, 'steps': 1190, 'loss/train': 2.426180899143219} 01/28/2022 14:25:38 - INFO - codeparrot_training - Step 1191: {'lr': 0.00029775, 'samples': 228864, 'steps': 1191, 'loss/train': 2.2409749031066895} 01/28/2022 14:25:42 - INFO - codeparrot_training - Step 1192: {'lr': 0.000298, 'samples': 229056, 'steps': 1192, 'loss/train': 2.3913220167160034} 01/28/2022 14:25:46 - INFO - codeparrot_training - Step 1193: {'lr': 0.00029825, 'samples': 229248, 'steps': 1193, 'loss/train': 1.7629244327545166} 01/28/2022 14:25:51 - INFO - codeparrot_training - Step 1194: {'lr': 0.0002985, 'samples': 229440, 'steps': 1194, 'loss/train': 1.996113359928131} 01/28/2022 14:25:56 - INFO - codeparrot_training - Step 1195: {'lr': 0.00029875, 'samples': 229632, 'steps': 1195, 'loss/train': 2.0356900691986084} 01/28/2022 14:26:00 - INFO - codeparrot_training - Step 1196: {'lr': 0.000299, 'samples': 229824, 'steps': 1196, 'loss/train': 2.0065242648124695} 01/28/2022 14:26:04 - INFO - codeparrot_training - Step 1197: {'lr': 0.00029925000000000004, 'samples': 230016, 'steps': 1197, 'loss/train': 2.469846189022064} 01/28/2022 14:26:08 - INFO - codeparrot_training - Step 1198: {'lr': 0.0002995, 'samples': 230208, 'steps': 1198, 'loss/train': 1.7728807926177979} 01/28/2022 14:26:13 - INFO - codeparrot_training - Step 1199: {'lr': 0.00029975000000000005, 'samples': 230400, 'steps': 1199, 'loss/train': 1.248712420463562} 01/28/2022 14:26:17 - INFO - codeparrot_training - Step 1200: {'lr': 0.0003, 'samples': 230592, 'steps': 1200, 'loss/train': 1.5993919372558594} 01/28/2022 14:26:22 - INFO - codeparrot_training - Step 1201: {'lr': 0.00030025, 'samples': 230784, 'steps': 1201, 'loss/train': 1.9246911406517029} 01/28/2022 14:26:26 - INFO - codeparrot_training - Step 1202: {'lr': 0.0003005, 'samples': 230976, 'steps': 1202, 'loss/train': 1.0685438811779022} 01/28/2022 14:26:30 - INFO - codeparrot_training - Step 1203: {'lr': 0.00030075, 'samples': 231168, 'steps': 1203, 'loss/train': 1.7633875608444214} 01/28/2022 14:26:35 - INFO - codeparrot_training - Step 1204: {'lr': 0.000301, 'samples': 231360, 'steps': 1204, 'loss/train': 1.6271365284919739} 01/28/2022 14:26:39 - INFO - codeparrot_training - Step 1205: {'lr': 0.00030125000000000003, 'samples': 231552, 'steps': 1205, 'loss/train': 2.708570122718811} 01/28/2022 14:26:43 - INFO - codeparrot_training - Step 1206: {'lr': 0.0003015, 'samples': 231744, 'steps': 1206, 'loss/train': 2.2029318809509277} 01/28/2022 14:26:48 - INFO - codeparrot_training - Step 1207: {'lr': 0.00030175000000000004, 'samples': 231936, 'steps': 1207, 'loss/train': 2.648502767086029} 01/28/2022 14:26:52 - INFO - codeparrot_training - Step 1208: {'lr': 0.000302, 'samples': 232128, 'steps': 1208, 'loss/train': 1.8947778940200806} 01/28/2022 14:26:58 - INFO - codeparrot_training - Step 1209: {'lr': 0.00030225, 'samples': 232320, 'steps': 1209, 'loss/train': 1.4863627552986145} 01/28/2022 14:27:02 - INFO - codeparrot_training - Step 1210: {'lr': 0.0003025, 'samples': 232512, 'steps': 1210, 'loss/train': 1.798529863357544} 01/28/2022 14:27:06 - INFO - codeparrot_training - Step 1211: {'lr': 0.00030275, 'samples': 232704, 'steps': 1211, 'loss/train': 1.7521644830703735} 01/28/2022 14:27:11 - INFO - codeparrot_training - Step 1212: {'lr': 0.000303, 'samples': 232896, 'steps': 1212, 'loss/train': 1.075796216726303} 01/28/2022 14:27:15 - INFO - codeparrot_training - Step 1213: {'lr': 0.00030325, 'samples': 233088, 'steps': 1213, 'loss/train': 2.2276135683059692} 01/28/2022 14:27:20 - INFO - codeparrot_training - Step 1214: {'lr': 0.0003035, 'samples': 233280, 'steps': 1214, 'loss/train': 1.6283690929412842} 01/28/2022 14:27:24 - INFO - codeparrot_training - Step 1215: {'lr': 0.00030375000000000004, 'samples': 233472, 'steps': 1215, 'loss/train': 2.640658736228943} 01/28/2022 14:27:28 - INFO - codeparrot_training - Step 1216: {'lr': 0.000304, 'samples': 233664, 'steps': 1216, 'loss/train': 1.7783694863319397} 01/28/2022 14:27:33 - INFO - codeparrot_training - Step 1217: {'lr': 0.00030425000000000005, 'samples': 233856, 'steps': 1217, 'loss/train': 0.5707189589738846} 01/28/2022 14:27:37 - INFO - codeparrot_training - Step 1218: {'lr': 0.0003045, 'samples': 234048, 'steps': 1218, 'loss/train': 1.9718114733695984} 01/28/2022 14:27:42 - INFO - codeparrot_training - Step 1219: {'lr': 0.00030475, 'samples': 234240, 'steps': 1219, 'loss/train': 1.4609184265136719} 01/28/2022 14:27:46 - INFO - codeparrot_training - Step 1220: {'lr': 0.000305, 'samples': 234432, 'steps': 1220, 'loss/train': 1.7563097476959229} 01/28/2022 14:27:50 - INFO - codeparrot_training - Step 1221: {'lr': 0.00030525, 'samples': 234624, 'steps': 1221, 'loss/train': 1.4921985268592834} 01/28/2022 14:27:54 - INFO - codeparrot_training - Step 1222: {'lr': 0.0003055, 'samples': 234816, 'steps': 1222, 'loss/train': 1.7588882446289062} 01/28/2022 14:27:58 - INFO - codeparrot_training - Step 1223: {'lr': 0.00030575000000000003, 'samples': 235008, 'steps': 1223, 'loss/train': 2.4149685502052307} 01/28/2022 14:28:05 - INFO - codeparrot_training - Step 1224: {'lr': 0.000306, 'samples': 235200, 'steps': 1224, 'loss/train': 1.5238671898841858} 01/28/2022 14:28:09 - INFO - codeparrot_training - Step 1225: {'lr': 0.00030625000000000004, 'samples': 235392, 'steps': 1225, 'loss/train': 2.1945383548736572} 01/28/2022 14:28:13 - INFO - codeparrot_training - Step 1226: {'lr': 0.0003065, 'samples': 235584, 'steps': 1226, 'loss/train': 2.148009181022644} 01/28/2022 14:28:17 - INFO - codeparrot_training - Step 1227: {'lr': 0.00030675, 'samples': 235776, 'steps': 1227, 'loss/train': 1.776505708694458} 01/28/2022 14:28:21 - INFO - codeparrot_training - Step 1228: {'lr': 0.000307, 'samples': 235968, 'steps': 1228, 'loss/train': 2.006635844707489} 01/28/2022 14:28:27 - INFO - codeparrot_training - Step 1229: {'lr': 0.00030725, 'samples': 236160, 'steps': 1229, 'loss/train': 2.0551148056983948} 01/28/2022 14:28:31 - INFO - codeparrot_training - Step 1230: {'lr': 0.0003075, 'samples': 236352, 'steps': 1230, 'loss/train': 2.2491883635520935} 01/28/2022 14:28:35 - INFO - codeparrot_training - Step 1231: {'lr': 0.00030775, 'samples': 236544, 'steps': 1231, 'loss/train': 1.9928281903266907} 01/28/2022 14:28:39 - INFO - codeparrot_training - Step 1232: {'lr': 0.000308, 'samples': 236736, 'steps': 1232, 'loss/train': 2.1883798241615295} 01/28/2022 14:28:43 - INFO - codeparrot_training - Step 1233: {'lr': 0.00030825000000000004, 'samples': 236928, 'steps': 1233, 'loss/train': 1.7304493188858032} 01/28/2022 14:28:49 - INFO - codeparrot_training - Step 1234: {'lr': 0.0003085, 'samples': 237120, 'steps': 1234, 'loss/train': 2.797524154186249} 01/28/2022 14:28:53 - INFO - codeparrot_training - Step 1235: {'lr': 0.00030875000000000005, 'samples': 237312, 'steps': 1235, 'loss/train': 2.330828905105591} 01/28/2022 14:28:57 - INFO - codeparrot_training - Step 1236: {'lr': 0.00030900000000000003, 'samples': 237504, 'steps': 1236, 'loss/train': 1.3517307043075562} 01/28/2022 14:29:02 - INFO - codeparrot_training - Step 1237: {'lr': 0.00030925, 'samples': 237696, 'steps': 1237, 'loss/train': 2.1212173104286194} 01/28/2022 14:29:06 - INFO - codeparrot_training - Step 1238: {'lr': 0.0003095, 'samples': 237888, 'steps': 1238, 'loss/train': 1.658824861049652} 01/28/2022 14:29:11 - INFO - codeparrot_training - Step 1239: {'lr': 0.00030975, 'samples': 238080, 'steps': 1239, 'loss/train': 2.3277708292007446} 01/28/2022 14:29:15 - INFO - codeparrot_training - Step 1240: {'lr': 0.00031, 'samples': 238272, 'steps': 1240, 'loss/train': 2.2680827379226685} 01/28/2022 14:29:19 - INFO - codeparrot_training - Step 1241: {'lr': 0.00031025000000000003, 'samples': 238464, 'steps': 1241, 'loss/train': 1.9836273193359375} 01/28/2022 14:29:23 - INFO - codeparrot_training - Step 1242: {'lr': 0.0003105, 'samples': 238656, 'steps': 1242, 'loss/train': 1.9074169993400574} 01/28/2022 14:29:28 - INFO - codeparrot_training - Step 1243: {'lr': 0.00031075000000000005, 'samples': 238848, 'steps': 1243, 'loss/train': 1.9596762657165527} 01/28/2022 14:29:33 - INFO - codeparrot_training - Step 1244: {'lr': 0.000311, 'samples': 239040, 'steps': 1244, 'loss/train': 2.069822072982788} 01/28/2022 14:29:37 - INFO - codeparrot_training - Step 1245: {'lr': 0.00031125000000000006, 'samples': 239232, 'steps': 1245, 'loss/train': 1.9424791932106018} 01/28/2022 14:29:41 - INFO - codeparrot_training - Step 1246: {'lr': 0.0003115, 'samples': 239424, 'steps': 1246, 'loss/train': 2.0992666482925415} 01/28/2022 14:29:45 - INFO - codeparrot_training - Step 1247: {'lr': 0.00031175, 'samples': 239616, 'steps': 1247, 'loss/train': 2.2403045296669006} 01/28/2022 14:29:50 - INFO - codeparrot_training - Step 1248: {'lr': 0.000312, 'samples': 239808, 'steps': 1248, 'loss/train': 1.8166873455047607} 01/28/2022 14:29:56 - INFO - codeparrot_training - Step 1249: {'lr': 0.00031225000000000003, 'samples': 240000, 'steps': 1249, 'loss/train': 1.8324059844017029} 01/28/2022 14:30:00 - INFO - codeparrot_training - Step 1250: {'lr': 0.0003125, 'samples': 240192, 'steps': 1250, 'loss/train': 2.144441306591034} 01/28/2022 14:30:04 - INFO - codeparrot_training - Step 1251: {'lr': 0.00031275, 'samples': 240384, 'steps': 1251, 'loss/train': 1.2343113720417023} 01/28/2022 14:30:08 - INFO - codeparrot_training - Step 1252: {'lr': 0.000313, 'samples': 240576, 'steps': 1252, 'loss/train': 1.9789615273475647} 01/28/2022 14:30:12 - INFO - codeparrot_training - Step 1253: {'lr': 0.00031325, 'samples': 240768, 'steps': 1253, 'loss/train': 2.2312560081481934} 01/28/2022 14:30:18 - INFO - codeparrot_training - Step 1254: {'lr': 0.00031350000000000003, 'samples': 240960, 'steps': 1254, 'loss/train': 2.2320646047592163} 01/28/2022 14:30:22 - INFO - codeparrot_training - Step 1255: {'lr': 0.00031374999999999996, 'samples': 241152, 'steps': 1255, 'loss/train': 2.1334245800971985} 01/28/2022 14:30:26 - INFO - codeparrot_training - Step 1256: {'lr': 0.000314, 'samples': 241344, 'steps': 1256, 'loss/train': 2.283311426639557} 01/28/2022 14:30:30 - INFO - codeparrot_training - Step 1257: {'lr': 0.00031424999999999997, 'samples': 241536, 'steps': 1257, 'loss/train': 1.394237458705902} 01/28/2022 14:30:34 - INFO - codeparrot_training - Step 1258: {'lr': 0.0003145, 'samples': 241728, 'steps': 1258, 'loss/train': 2.1438159942626953} 01/28/2022 14:30:39 - INFO - codeparrot_training - Step 1259: {'lr': 0.00031475, 'samples': 241920, 'steps': 1259, 'loss/train': 1.8382510542869568} 01/28/2022 14:30:44 - INFO - codeparrot_training - Step 1260: {'lr': 0.000315, 'samples': 242112, 'steps': 1260, 'loss/train': 2.5056676268577576} 01/28/2022 14:30:48 - INFO - codeparrot_training - Step 1261: {'lr': 0.00031525, 'samples': 242304, 'steps': 1261, 'loss/train': 1.3986004292964935} 01/28/2022 14:30:52 - INFO - codeparrot_training - Step 1262: {'lr': 0.0003155, 'samples': 242496, 'steps': 1262, 'loss/train': 0.3709464892745018} 01/28/2022 14:30:56 - INFO - codeparrot_training - Step 1263: {'lr': 0.00031575, 'samples': 242688, 'steps': 1263, 'loss/train': 1.9601438641548157} 01/28/2022 14:31:01 - INFO - codeparrot_training - Step 1264: {'lr': 0.000316, 'samples': 242880, 'steps': 1264, 'loss/train': 0.9880879819393158} 01/28/2022 14:31:05 - INFO - codeparrot_training - Step 1265: {'lr': 0.00031624999999999996, 'samples': 243072, 'steps': 1265, 'loss/train': 2.6647539138793945} 01/28/2022 14:31:10 - INFO - codeparrot_training - Step 1266: {'lr': 0.0003165, 'samples': 243264, 'steps': 1266, 'loss/train': 2.603080451488495} 01/28/2022 14:31:14 - INFO - codeparrot_training - Step 1267: {'lr': 0.00031675, 'samples': 243456, 'steps': 1267, 'loss/train': 1.8904933333396912} 01/28/2022 14:31:18 - INFO - codeparrot_training - Step 1268: {'lr': 0.000317, 'samples': 243648, 'steps': 1268, 'loss/train': 1.7316836714744568} 01/28/2022 14:31:24 - INFO - codeparrot_training - Step 1269: {'lr': 0.00031725, 'samples': 243840, 'steps': 1269, 'loss/train': 2.2894940972328186} 01/28/2022 14:31:28 - INFO - codeparrot_training - Step 1270: {'lr': 0.0003175, 'samples': 244032, 'steps': 1270, 'loss/train': 1.5143535733222961} 01/28/2022 14:31:32 - INFO - codeparrot_training - Step 1271: {'lr': 0.00031775, 'samples': 244224, 'steps': 1271, 'loss/train': 1.7681676149368286} 01/28/2022 14:31:37 - INFO - codeparrot_training - Step 1272: {'lr': 0.00031800000000000003, 'samples': 244416, 'steps': 1272, 'loss/train': 2.095627784729004} 01/28/2022 14:31:41 - INFO - codeparrot_training - Step 1273: {'lr': 0.00031825, 'samples': 244608, 'steps': 1273, 'loss/train': 1.8858548998832703} 01/28/2022 14:31:46 - INFO - codeparrot_training - Step 1274: {'lr': 0.0003185, 'samples': 244800, 'steps': 1274, 'loss/train': 1.6024004817008972} 01/28/2022 14:31:50 - INFO - codeparrot_training - Step 1275: {'lr': 0.00031874999999999997, 'samples': 244992, 'steps': 1275, 'loss/train': 1.3441583812236786} 01/28/2022 14:31:54 - INFO - codeparrot_training - Step 1276: {'lr': 0.000319, 'samples': 245184, 'steps': 1276, 'loss/train': 4.006978869438171} 01/28/2022 14:31:58 - INFO - codeparrot_training - Step 1277: {'lr': 0.00031925, 'samples': 245376, 'steps': 1277, 'loss/train': 2.839811146259308} 01/28/2022 14:32:03 - INFO - codeparrot_training - Step 1278: {'lr': 0.0003195, 'samples': 245568, 'steps': 1278, 'loss/train': 2.4529988765716553} 01/28/2022 14:32:08 - INFO - codeparrot_training - Step 1279: {'lr': 0.00031975, 'samples': 245760, 'steps': 1279, 'loss/train': 0.9646402895450592} 01/28/2022 14:32:13 - INFO - codeparrot_training - Step 1280: {'lr': 0.00032, 'samples': 245952, 'steps': 1280, 'loss/train': 3.6997357606887817} 01/28/2022 14:32:17 - INFO - codeparrot_training - Step 1281: {'lr': 0.00032025, 'samples': 246144, 'steps': 1281, 'loss/train': 2.2281416058540344} 01/28/2022 14:32:21 - INFO - codeparrot_training - Step 1282: {'lr': 0.00032050000000000004, 'samples': 246336, 'steps': 1282, 'loss/train': 1.5195629596710205} 01/28/2022 14:32:25 - INFO - codeparrot_training - Step 1283: {'lr': 0.00032074999999999996, 'samples': 246528, 'steps': 1283, 'loss/train': 1.9839765429496765} 01/28/2022 14:32:31 - INFO - codeparrot_training - Step 1284: {'lr': 0.000321, 'samples': 246720, 'steps': 1284, 'loss/train': 1.2186395823955536} 01/28/2022 14:32:36 - INFO - codeparrot_training - Step 1285: {'lr': 0.00032125, 'samples': 246912, 'steps': 1285, 'loss/train': 2.135055184364319} 01/28/2022 14:32:40 - INFO - codeparrot_training - Step 1286: {'lr': 0.0003215, 'samples': 247104, 'steps': 1286, 'loss/train': 2.6615227460861206} 01/28/2022 14:32:44 - INFO - codeparrot_training - Step 1287: {'lr': 0.00032175, 'samples': 247296, 'steps': 1287, 'loss/train': 1.6141528487205505} 01/28/2022 14:32:48 - INFO - codeparrot_training - Step 1288: {'lr': 0.000322, 'samples': 247488, 'steps': 1288, 'loss/train': 2.3701361417770386} 01/28/2022 14:32:53 - INFO - codeparrot_training - Step 1289: {'lr': 0.00032225, 'samples': 247680, 'steps': 1289, 'loss/train': 2.422580122947693} 01/28/2022 14:32:57 - INFO - codeparrot_training - Step 1290: {'lr': 0.00032250000000000003, 'samples': 247872, 'steps': 1290, 'loss/train': 1.765417456626892} 01/28/2022 14:33:02 - INFO - codeparrot_training - Step 1291: {'lr': 0.00032275, 'samples': 248064, 'steps': 1291, 'loss/train': 1.3590107560157776} 01/28/2022 14:33:06 - INFO - codeparrot_training - Step 1292: {'lr': 0.000323, 'samples': 248256, 'steps': 1292, 'loss/train': 1.7438267469406128} 01/28/2022 14:33:10 - INFO - codeparrot_training - Step 1293: {'lr': 0.00032324999999999997, 'samples': 248448, 'steps': 1293, 'loss/train': 2.316345512866974} 01/28/2022 14:33:16 - INFO - codeparrot_training - Step 1294: {'lr': 0.0003235, 'samples': 248640, 'steps': 1294, 'loss/train': 2.2638509273529053} 01/28/2022 14:33:21 - INFO - codeparrot_training - Step 1295: {'lr': 0.00032375, 'samples': 248832, 'steps': 1295, 'loss/train': 0.8593974709510803} 01/28/2022 14:33:25 - INFO - codeparrot_training - Step 1296: {'lr': 0.000324, 'samples': 249024, 'steps': 1296, 'loss/train': 2.4800893664360046} 01/28/2022 14:33:29 - INFO - codeparrot_training - Step 1297: {'lr': 0.00032425, 'samples': 249216, 'steps': 1297, 'loss/train': 2.8618574738502502} 01/28/2022 14:33:33 - INFO - codeparrot_training - Step 1298: {'lr': 0.00032450000000000003, 'samples': 249408, 'steps': 1298, 'loss/train': 3.4136245250701904} 01/28/2022 14:33:40 - INFO - codeparrot_training - Step 1299: {'lr': 0.00032475, 'samples': 249600, 'steps': 1299, 'loss/train': 6.189499139785767} 01/28/2022 14:33:44 - INFO - codeparrot_training - Step 1300: {'lr': 0.00032500000000000004, 'samples': 249792, 'steps': 1300, 'loss/train': 2.6553266644477844} 01/28/2022 14:33:48 - INFO - codeparrot_training - Step 1301: {'lr': 0.00032524999999999996, 'samples': 249984, 'steps': 1301, 'loss/train': 2.3256980180740356} 01/28/2022 14:33:52 - INFO - codeparrot_training - Step 1302: {'lr': 0.0003255, 'samples': 250176, 'steps': 1302, 'loss/train': 2.0929550528526306} 01/28/2022 14:33:56 - INFO - codeparrot_training - Step 1303: {'lr': 0.00032575, 'samples': 250368, 'steps': 1303, 'loss/train': 2.2351879477500916} 01/28/2022 14:34:01 - INFO - codeparrot_training - Step 1304: {'lr': 0.000326, 'samples': 250560, 'steps': 1304, 'loss/train': 1.607498288154602} 01/28/2022 14:34:06 - INFO - codeparrot_training - Step 1305: {'lr': 0.00032625, 'samples': 250752, 'steps': 1305, 'loss/train': 2.546508014202118} 01/28/2022 14:34:10 - INFO - codeparrot_training - Step 1306: {'lr': 0.0003265, 'samples': 250944, 'steps': 1306, 'loss/train': 2.697987914085388} 01/28/2022 14:34:14 - INFO - codeparrot_training - Step 1307: {'lr': 0.00032675, 'samples': 251136, 'steps': 1307, 'loss/train': 2.7276636958122253} 01/28/2022 14:34:18 - INFO - codeparrot_training - Step 1308: {'lr': 0.00032700000000000003, 'samples': 251328, 'steps': 1308, 'loss/train': 2.1076449751853943} 01/28/2022 14:34:22 - INFO - codeparrot_training - Step 1309: {'lr': 0.00032725, 'samples': 251520, 'steps': 1309, 'loss/train': 1.9426549673080444} 01/28/2022 14:34:29 - INFO - codeparrot_training - Step 1310: {'lr': 0.00032750000000000005, 'samples': 251712, 'steps': 1310, 'loss/train': 2.77842253446579} 01/28/2022 14:34:33 - INFO - codeparrot_training - Step 1311: {'lr': 0.00032774999999999997, 'samples': 251904, 'steps': 1311, 'loss/train': 2.3041967153549194} 01/28/2022 14:34:37 - INFO - codeparrot_training - Step 1312: {'lr': 0.000328, 'samples': 252096, 'steps': 1312, 'loss/train': 2.5988616943359375} 01/28/2022 14:34:41 - INFO - codeparrot_training - Step 1313: {'lr': 0.00032825, 'samples': 252288, 'steps': 1313, 'loss/train': 2.0474138259887695} 01/28/2022 14:34:45 - INFO - codeparrot_training - Step 1314: {'lr': 0.0003285, 'samples': 252480, 'steps': 1314, 'loss/train': 2.2288082242012024} 01/28/2022 14:34:51 - INFO - codeparrot_training - Step 1315: {'lr': 0.00032875, 'samples': 252672, 'steps': 1315, 'loss/train': 1.3138115108013153} 01/28/2022 14:34:55 - INFO - codeparrot_training - Step 1316: {'lr': 0.00032900000000000003, 'samples': 252864, 'steps': 1316, 'loss/train': 0.9061785936355591} 01/28/2022 14:34:59 - INFO - codeparrot_training - Step 1317: {'lr': 0.00032925, 'samples': 253056, 'steps': 1317, 'loss/train': 2.041993260383606} 01/28/2022 14:35:03 - INFO - codeparrot_training - Step 1318: {'lr': 0.00032950000000000004, 'samples': 253248, 'steps': 1318, 'loss/train': 1.9553212523460388} 01/28/2022 14:35:07 - INFO - codeparrot_training - Step 1319: {'lr': 0.00032975, 'samples': 253440, 'steps': 1319, 'loss/train': 2.995845079421997} 01/28/2022 14:35:13 - INFO - codeparrot_training - Step 1320: {'lr': 0.00033, 'samples': 253632, 'steps': 1320, 'loss/train': 0.7212120741605759} 01/28/2022 14:35:17 - INFO - codeparrot_training - Step 1321: {'lr': 0.00033025, 'samples': 253824, 'steps': 1321, 'loss/train': 1.9337984919548035} 01/28/2022 14:35:21 - INFO - codeparrot_training - Step 1322: {'lr': 0.0003305, 'samples': 254016, 'steps': 1322, 'loss/train': 1.747385859489441} 01/28/2022 14:35:25 - INFO - codeparrot_training - Step 1323: {'lr': 0.00033075, 'samples': 254208, 'steps': 1323, 'loss/train': 2.1172916293144226} 01/28/2022 14:35:29 - INFO - codeparrot_training - Step 1324: {'lr': 0.000331, 'samples': 254400, 'steps': 1324, 'loss/train': 1.8487502932548523} 01/28/2022 14:35:35 - INFO - codeparrot_training - Step 1325: {'lr': 0.00033125, 'samples': 254592, 'steps': 1325, 'loss/train': 1.3575683534145355} 01/28/2022 14:35:40 - INFO - codeparrot_training - Step 1326: {'lr': 0.00033150000000000003, 'samples': 254784, 'steps': 1326, 'loss/train': 2.272898554801941} 01/28/2022 14:35:44 - INFO - codeparrot_training - Step 1327: {'lr': 0.00033175, 'samples': 254976, 'steps': 1327, 'loss/train': 1.8226715326309204} 01/28/2022 14:35:48 - INFO - codeparrot_training - Step 1328: {'lr': 0.00033200000000000005, 'samples': 255168, 'steps': 1328, 'loss/train': 1.931670069694519} 01/28/2022 14:35:52 - INFO - codeparrot_training - Step 1329: {'lr': 0.00033224999999999997, 'samples': 255360, 'steps': 1329, 'loss/train': 2.3527586460113525} 01/28/2022 14:35:57 - INFO - codeparrot_training - Step 1330: {'lr': 0.0003325, 'samples': 255552, 'steps': 1330, 'loss/train': 1.7090086340904236} 01/28/2022 14:36:01 - INFO - codeparrot_training - Step 1331: {'lr': 0.00033275, 'samples': 255744, 'steps': 1331, 'loss/train': 2.0983681082725525} 01/28/2022 14:36:06 - INFO - codeparrot_training - Step 1332: {'lr': 0.000333, 'samples': 255936, 'steps': 1332, 'loss/train': 2.412836194038391} 01/28/2022 14:36:10 - INFO - codeparrot_training - Step 1333: {'lr': 0.00033325, 'samples': 256128, 'steps': 1333, 'loss/train': 1.30035799741745} 01/28/2022 14:36:14 - INFO - codeparrot_training - Step 1334: {'lr': 0.00033350000000000003, 'samples': 256320, 'steps': 1334, 'loss/train': 1.8489357233047485} 01/28/2022 14:36:19 - INFO - codeparrot_training - Step 1335: {'lr': 0.00033375, 'samples': 256512, 'steps': 1335, 'loss/train': 1.6000515818595886} 01/28/2022 14:36:23 - INFO - codeparrot_training - Step 1336: {'lr': 0.00033400000000000004, 'samples': 256704, 'steps': 1336, 'loss/train': 2.6565458178520203} 01/28/2022 14:36:27 - INFO - codeparrot_training - Step 1337: {'lr': 0.00033425, 'samples': 256896, 'steps': 1337, 'loss/train': 2.6835063099861145} 01/28/2022 14:36:32 - INFO - codeparrot_training - Step 1338: {'lr': 0.00033450000000000005, 'samples': 257088, 'steps': 1338, 'loss/train': 1.8902562260627747} 01/28/2022 14:36:36 - INFO - codeparrot_training - Step 1339: {'lr': 0.00033475, 'samples': 257280, 'steps': 1339, 'loss/train': 2.2435866594314575} 01/28/2022 14:36:43 - INFO - codeparrot_training - Step 1340: {'lr': 0.000335, 'samples': 257472, 'steps': 1340, 'loss/train': 1.6515862941741943} 01/28/2022 14:36:47 - INFO - codeparrot_training - Step 1341: {'lr': 0.00033525, 'samples': 257664, 'steps': 1341, 'loss/train': 1.3815561532974243} 01/28/2022 14:36:51 - INFO - codeparrot_training - Step 1342: {'lr': 0.0003355, 'samples': 257856, 'steps': 1342, 'loss/train': 1.2651248574256897} 01/28/2022 14:36:55 - INFO - codeparrot_training - Step 1343: {'lr': 0.00033575, 'samples': 258048, 'steps': 1343, 'loss/train': 2.5238158106803894} 01/28/2022 14:36:59 - INFO - codeparrot_training - Step 1344: {'lr': 0.00033600000000000004, 'samples': 258240, 'steps': 1344, 'loss/train': 1.9782784581184387} 01/28/2022 14:37:05 - INFO - codeparrot_training - Step 1345: {'lr': 0.00033625, 'samples': 258432, 'steps': 1345, 'loss/train': 2.1355027556419373} 01/28/2022 14:37:09 - INFO - codeparrot_training - Step 1346: {'lr': 0.00033650000000000005, 'samples': 258624, 'steps': 1346, 'loss/train': 2.072050452232361} 01/28/2022 14:37:13 - INFO - codeparrot_training - Step 1347: {'lr': 0.00033675, 'samples': 258816, 'steps': 1347, 'loss/train': 2.6958664655685425} 01/28/2022 14:37:17 - INFO - codeparrot_training - Step 1348: {'lr': 0.000337, 'samples': 259008, 'steps': 1348, 'loss/train': 0.5225539058446884} 01/28/2022 14:37:21 - INFO - codeparrot_training - Step 1349: {'lr': 0.00033725, 'samples': 259200, 'steps': 1349, 'loss/train': 2.016878843307495} 01/28/2022 14:37:26 - INFO - codeparrot_training - Step 1350: {'lr': 0.0003375, 'samples': 259392, 'steps': 1350, 'loss/train': 1.0599360466003418} 01/28/2022 14:37:31 - INFO - codeparrot_training - Step 1351: {'lr': 0.00033775, 'samples': 259584, 'steps': 1351, 'loss/train': 1.8771672248840332} 01/28/2022 14:37:35 - INFO - codeparrot_training - Step 1352: {'lr': 0.00033800000000000003, 'samples': 259776, 'steps': 1352, 'loss/train': 1.8408588767051697} 01/28/2022 14:37:39 - INFO - codeparrot_training - Step 1353: {'lr': 0.00033825, 'samples': 259968, 'steps': 1353, 'loss/train': 2.588968813419342} 01/28/2022 14:37:43 - INFO - codeparrot_training - Step 1354: {'lr': 0.00033850000000000004, 'samples': 260160, 'steps': 1354, 'loss/train': 2.027992844581604} 01/28/2022 14:37:49 - INFO - codeparrot_training - Step 1355: {'lr': 0.00033875, 'samples': 260352, 'steps': 1355, 'loss/train': 2.5865301489830017} 01/28/2022 14:37:54 - INFO - codeparrot_training - Step 1356: {'lr': 0.00033900000000000005, 'samples': 260544, 'steps': 1356, 'loss/train': 2.1816465854644775} 01/28/2022 14:37:58 - INFO - codeparrot_training - Step 1357: {'lr': 0.00033925, 'samples': 260736, 'steps': 1357, 'loss/train': 2.0678027272224426} 01/28/2022 14:38:02 - INFO - codeparrot_training - Step 1358: {'lr': 0.0003395, 'samples': 260928, 'steps': 1358, 'loss/train': 2.2306909561157227} 01/28/2022 14:38:06 - INFO - codeparrot_training - Step 1359: {'lr': 0.00033975, 'samples': 261120, 'steps': 1359, 'loss/train': 2.2895904779434204} 01/28/2022 14:38:11 - INFO - codeparrot_training - Step 1360: {'lr': 0.00034, 'samples': 261312, 'steps': 1360, 'loss/train': 2.860239565372467} 01/28/2022 14:38:15 - INFO - codeparrot_training - Step 1361: {'lr': 0.00034025, 'samples': 261504, 'steps': 1361, 'loss/train': 1.577426791191101} 01/28/2022 14:38:20 - INFO - codeparrot_training - Step 1362: {'lr': 0.00034050000000000004, 'samples': 261696, 'steps': 1362, 'loss/train': 1.8574272394180298} 01/28/2022 14:38:24 - INFO - codeparrot_training - Step 1363: {'lr': 0.00034075, 'samples': 261888, 'steps': 1363, 'loss/train': 0.7470297664403915} 01/28/2022 14:38:28 - INFO - codeparrot_training - Step 1364: {'lr': 0.00034100000000000005, 'samples': 262080, 'steps': 1364, 'loss/train': 1.7457994222640991} 01/28/2022 14:38:34 - INFO - codeparrot_training - Step 1365: {'lr': 0.00034125000000000003, 'samples': 262272, 'steps': 1365, 'loss/train': 4.521185874938965} 01/28/2022 14:38:38 - INFO - codeparrot_training - Step 1366: {'lr': 0.0003415, 'samples': 262464, 'steps': 1366, 'loss/train': 1.1644204258918762} 01/28/2022 14:38:42 - INFO - codeparrot_training - Step 1367: {'lr': 0.00034175, 'samples': 262656, 'steps': 1367, 'loss/train': 1.3718077540397644} 01/28/2022 14:38:47 - INFO - codeparrot_training - Step 1368: {'lr': 0.000342, 'samples': 262848, 'steps': 1368, 'loss/train': 2.42343932390213} 01/28/2022 14:38:51 - INFO - codeparrot_training - Step 1369: {'lr': 0.00034225, 'samples': 263040, 'steps': 1369, 'loss/train': 2.632986009120941} 01/28/2022 14:38:56 - INFO - codeparrot_training - Step 1370: {'lr': 0.00034250000000000003, 'samples': 263232, 'steps': 1370, 'loss/train': 2.4014114141464233} 01/28/2022 14:39:00 - INFO - codeparrot_training - Step 1371: {'lr': 0.00034275, 'samples': 263424, 'steps': 1371, 'loss/train': 2.4772748351097107} 01/28/2022 14:39:04 - INFO - codeparrot_training - Step 1372: {'lr': 0.00034300000000000004, 'samples': 263616, 'steps': 1372, 'loss/train': 1.9733619689941406} 01/28/2022 14:39:08 - INFO - codeparrot_training - Step 1373: {'lr': 0.00034325, 'samples': 263808, 'steps': 1373, 'loss/train': 2.1228017807006836} 01/28/2022 14:39:13 - INFO - codeparrot_training - Step 1374: {'lr': 0.00034350000000000006, 'samples': 264000, 'steps': 1374, 'loss/train': 1.4105159640312195} 01/28/2022 14:39:18 - INFO - codeparrot_training - Step 1375: {'lr': 0.00034375, 'samples': 264192, 'steps': 1375, 'loss/train': 1.4852900505065918} 01/28/2022 14:39:22 - INFO - codeparrot_training - Step 1376: {'lr': 0.00034399999999999996, 'samples': 264384, 'steps': 1376, 'loss/train': 1.2998045682907104} 01/28/2022 14:39:26 - INFO - codeparrot_training - Step 1377: {'lr': 0.00034425, 'samples': 264576, 'steps': 1377, 'loss/train': 1.629166603088379} 01/28/2022 14:39:30 - INFO - codeparrot_training - Step 1378: {'lr': 0.00034449999999999997, 'samples': 264768, 'steps': 1378, 'loss/train': 2.3005215525627136} 01/28/2022 14:39:34 - INFO - codeparrot_training - Step 1379: {'lr': 0.00034475, 'samples': 264960, 'steps': 1379, 'loss/train': 1.7424293160438538} 01/28/2022 14:39:41 - INFO - codeparrot_training - Step 1380: {'lr': 0.000345, 'samples': 265152, 'steps': 1380, 'loss/train': 1.8219273090362549} 01/28/2022 14:39:45 - INFO - codeparrot_training - Step 1381: {'lr': 0.00034525, 'samples': 265344, 'steps': 1381, 'loss/train': 2.8692728877067566} 01/28/2022 14:39:49 - INFO - codeparrot_training - Step 1382: {'lr': 0.0003455, 'samples': 265536, 'steps': 1382, 'loss/train': 2.0446810126304626} 01/28/2022 14:39:53 - INFO - codeparrot_training - Step 1383: {'lr': 0.00034575000000000003, 'samples': 265728, 'steps': 1383, 'loss/train': 2.2021061182022095} 01/28/2022 14:39:57 - INFO - codeparrot_training - Step 1384: {'lr': 0.000346, 'samples': 265920, 'steps': 1384, 'loss/train': 1.8561497926712036} 01/28/2022 14:40:03 - INFO - codeparrot_training - Step 1385: {'lr': 0.00034625, 'samples': 266112, 'steps': 1385, 'loss/train': 1.0708916187286377} 01/28/2022 14:40:07 - INFO - codeparrot_training - Step 1386: {'lr': 0.00034649999999999997, 'samples': 266304, 'steps': 1386, 'loss/train': 2.116901993751526} 01/28/2022 14:40:11 - INFO - codeparrot_training - Step 1387: {'lr': 0.00034675, 'samples': 266496, 'steps': 1387, 'loss/train': 2.4277337193489075} 01/28/2022 14:40:15 - INFO - codeparrot_training - Step 1388: {'lr': 0.000347, 'samples': 266688, 'steps': 1388, 'loss/train': 0.9782685041427612} 01/28/2022 14:40:19 - INFO - codeparrot_training - Step 1389: {'lr': 0.00034725, 'samples': 266880, 'steps': 1389, 'loss/train': 2.1834492087364197} 01/28/2022 14:40:26 - INFO - codeparrot_training - Step 1390: {'lr': 0.0003475, 'samples': 267072, 'steps': 1390, 'loss/train': 2.3380626440048218} 01/28/2022 14:40:30 - INFO - codeparrot_training - Step 1391: {'lr': 0.00034775, 'samples': 267264, 'steps': 1391, 'loss/train': 3.117113471031189} 01/28/2022 14:40:34 - INFO - codeparrot_training - Step 1392: {'lr': 0.000348, 'samples': 267456, 'steps': 1392, 'loss/train': 2.7118399143218994} 01/28/2022 14:40:38 - INFO - codeparrot_training - Step 1393: {'lr': 0.00034825000000000004, 'samples': 267648, 'steps': 1393, 'loss/train': 2.2237910628318787} 01/28/2022 14:40:42 - INFO - codeparrot_training - Step 1394: {'lr': 0.00034849999999999996, 'samples': 267840, 'steps': 1394, 'loss/train': 1.8954336047172546} 01/28/2022 14:40:47 - INFO - codeparrot_training - Step 1395: {'lr': 0.00034875, 'samples': 268032, 'steps': 1395, 'loss/train': 0.9795102775096893} 01/28/2022 14:40:51 - INFO - codeparrot_training - Step 1396: {'lr': 0.00034899999999999997, 'samples': 268224, 'steps': 1396, 'loss/train': 1.8414822220802307} 01/28/2022 14:40:56 - INFO - codeparrot_training - Step 1397: {'lr': 0.00034925, 'samples': 268416, 'steps': 1397, 'loss/train': 2.816260278224945} 01/28/2022 14:41:00 - INFO - codeparrot_training - Step 1398: {'lr': 0.0003495, 'samples': 268608, 'steps': 1398, 'loss/train': 0.859834760427475} 01/28/2022 14:41:04 - INFO - codeparrot_training - Step 1399: {'lr': 0.00034975, 'samples': 268800, 'steps': 1399, 'loss/train': 1.8614453673362732} 01/28/2022 14:41:09 - INFO - codeparrot_training - Step 1400: {'lr': 0.00035, 'samples': 268992, 'steps': 1400, 'loss/train': 0.31277579069137573} 01/28/2022 14:41:13 - INFO - codeparrot_training - Step 1401: {'lr': 0.00035025000000000003, 'samples': 269184, 'steps': 1401, 'loss/train': 2.5007164478302} 01/28/2022 14:41:17 - INFO - codeparrot_training - Step 1402: {'lr': 0.0003505, 'samples': 269376, 'steps': 1402, 'loss/train': 2.206383526325226} 01/28/2022 14:41:22 - INFO - codeparrot_training - Step 1403: {'lr': 0.00035075, 'samples': 269568, 'steps': 1403, 'loss/train': 2.4196372032165527} 01/28/2022 14:41:26 - INFO - codeparrot_training - Step 1404: {'lr': 0.00035099999999999997, 'samples': 269760, 'steps': 1404, 'loss/train': 1.476623386144638} 01/28/2022 14:41:31 - INFO - codeparrot_training - Step 1405: {'lr': 0.00035125, 'samples': 269952, 'steps': 1405, 'loss/train': 1.5099906921386719} 01/28/2022 14:41:35 - INFO - codeparrot_training - Step 1406: {'lr': 0.0003515, 'samples': 270144, 'steps': 1406, 'loss/train': 1.5855984091758728} 01/28/2022 14:41:39 - INFO - codeparrot_training - Step 1407: {'lr': 0.00035175, 'samples': 270336, 'steps': 1407, 'loss/train': 1.9899120926856995} 01/28/2022 14:41:43 - INFO - codeparrot_training - Step 1408: {'lr': 0.000352, 'samples': 270528, 'steps': 1408, 'loss/train': 2.0701995491981506} 01/28/2022 14:41:47 - INFO - codeparrot_training - Step 1409: {'lr': 0.00035225, 'samples': 270720, 'steps': 1409, 'loss/train': 0.5427952855825424} 01/28/2022 14:41:54 - INFO - codeparrot_training - Step 1410: {'lr': 0.0003525, 'samples': 270912, 'steps': 1410, 'loss/train': 1.9906865358352661} 01/28/2022 14:41:58 - INFO - codeparrot_training - Step 1411: {'lr': 0.00035275000000000004, 'samples': 271104, 'steps': 1411, 'loss/train': 0.9870024025440216} 01/28/2022 14:42:02 - INFO - codeparrot_training - Step 1412: {'lr': 0.00035299999999999996, 'samples': 271296, 'steps': 1412, 'loss/train': 4.5766496658325195} 01/28/2022 14:42:06 - INFO - codeparrot_training - Step 1413: {'lr': 0.00035325, 'samples': 271488, 'steps': 1413, 'loss/train': 2.3259868025779724} 01/28/2022 14:42:11 - INFO - codeparrot_training - Step 1414: {'lr': 0.0003535, 'samples': 271680, 'steps': 1414, 'loss/train': 1.3045651316642761} 01/28/2022 14:42:16 - INFO - codeparrot_training - Step 1415: {'lr': 0.00035375, 'samples': 271872, 'steps': 1415, 'loss/train': 1.2818263471126556} 01/28/2022 14:42:20 - INFO - codeparrot_training - Step 1416: {'lr': 0.000354, 'samples': 272064, 'steps': 1416, 'loss/train': 1.665628731250763} 01/28/2022 14:42:24 - INFO - codeparrot_training - Step 1417: {'lr': 0.00035425, 'samples': 272256, 'steps': 1417, 'loss/train': 1.380699098110199} 01/28/2022 14:42:29 - INFO - codeparrot_training - Step 1418: {'lr': 0.0003545, 'samples': 272448, 'steps': 1418, 'loss/train': 2.3854411840438843} 01/28/2022 14:42:33 - INFO - codeparrot_training - Step 1419: {'lr': 0.00035475000000000003, 'samples': 272640, 'steps': 1419, 'loss/train': 1.4741616547107697} 01/28/2022 14:42:38 - INFO - codeparrot_training - Step 1420: {'lr': 0.000355, 'samples': 272832, 'steps': 1420, 'loss/train': 1.7235773205757141} 01/28/2022 14:42:42 - INFO - codeparrot_training - Step 1421: {'lr': 0.00035525000000000004, 'samples': 273024, 'steps': 1421, 'loss/train': 2.261306583881378} 01/28/2022 14:42:46 - INFO - codeparrot_training - Step 1422: {'lr': 0.00035549999999999997, 'samples': 273216, 'steps': 1422, 'loss/train': 0.3395669460296631} 01/28/2022 14:42:50 - INFO - codeparrot_training - Step 1423: {'lr': 0.00035575, 'samples': 273408, 'steps': 1423, 'loss/train': 1.8638468384742737} 01/28/2022 14:42:54 - INFO - codeparrot_training - Step 1424: {'lr': 0.000356, 'samples': 273600, 'steps': 1424, 'loss/train': 2.5356557965278625} 01/28/2022 14:43:01 - INFO - codeparrot_training - Step 1425: {'lr': 0.00035625, 'samples': 273792, 'steps': 1425, 'loss/train': 1.4745907187461853} 01/28/2022 14:43:05 - INFO - codeparrot_training - Step 1426: {'lr': 0.0003565, 'samples': 273984, 'steps': 1426, 'loss/train': 2.21966814994812} 01/28/2022 14:43:09 - INFO - codeparrot_training - Step 1427: {'lr': 0.00035675, 'samples': 274176, 'steps': 1427, 'loss/train': 1.916796326637268} 01/28/2022 14:43:13 - INFO - codeparrot_training - Step 1428: {'lr': 0.000357, 'samples': 274368, 'steps': 1428, 'loss/train': 2.0884255170822144} 01/28/2022 14:43:18 - INFO - codeparrot_training - Step 1429: {'lr': 0.00035725000000000004, 'samples': 274560, 'steps': 1429, 'loss/train': 2.173219621181488} 01/28/2022 14:43:23 - INFO - codeparrot_training - Step 1430: {'lr': 0.0003575, 'samples': 274752, 'steps': 1430, 'loss/train': 1.588633954524994} 01/28/2022 14:43:27 - INFO - codeparrot_training - Step 1431: {'lr': 0.00035775, 'samples': 274944, 'steps': 1431, 'loss/train': 0.187926784157753} 01/28/2022 14:43:31 - INFO - codeparrot_training - Step 1432: {'lr': 0.000358, 'samples': 275136, 'steps': 1432, 'loss/train': 3.3244653940200806} 01/28/2022 14:43:35 - INFO - codeparrot_training - Step 1433: {'lr': 0.00035825, 'samples': 275328, 'steps': 1433, 'loss/train': 6.805822849273682} 01/28/2022 14:43:39 - INFO - codeparrot_training - Step 1434: {'lr': 0.0003585, 'samples': 275520, 'steps': 1434, 'loss/train': 2.645029127597809} 01/28/2022 14:43:45 - INFO - codeparrot_training - Step 1435: {'lr': 0.00035875, 'samples': 275712, 'steps': 1435, 'loss/train': 2.824445128440857} 01/28/2022 14:43:50 - INFO - codeparrot_training - Step 1436: {'lr': 0.000359, 'samples': 275904, 'steps': 1436, 'loss/train': 1.9784739017486572} 01/28/2022 14:43:54 - INFO - codeparrot_training - Step 1437: {'lr': 0.00035925000000000003, 'samples': 276096, 'steps': 1437, 'loss/train': 2.621967852115631} 01/28/2022 14:43:58 - INFO - codeparrot_training - Step 1438: {'lr': 0.0003595, 'samples': 276288, 'steps': 1438, 'loss/train': 1.8777088522911072} 01/28/2022 14:44:02 - INFO - codeparrot_training - Step 1439: {'lr': 0.00035975000000000004, 'samples': 276480, 'steps': 1439, 'loss/train': 1.238372415304184} 01/28/2022 14:44:07 - INFO - codeparrot_training - Step 1440: {'lr': 0.00035999999999999997, 'samples': 276672, 'steps': 1440, 'loss/train': 2.545883059501648} 01/28/2022 14:44:12 - INFO - codeparrot_training - Step 1441: {'lr': 0.00036025, 'samples': 276864, 'steps': 1441, 'loss/train': 2.162339150905609} 01/28/2022 14:44:16 - INFO - codeparrot_training - Step 1442: {'lr': 0.0003605, 'samples': 277056, 'steps': 1442, 'loss/train': 1.269875407218933} 01/28/2022 14:44:20 - INFO - codeparrot_training - Step 1443: {'lr': 0.00036075, 'samples': 277248, 'steps': 1443, 'loss/train': 1.8156200051307678} 01/28/2022 14:44:24 - INFO - codeparrot_training - Step 1444: {'lr': 0.000361, 'samples': 277440, 'steps': 1444, 'loss/train': 1.81924170255661} 01/28/2022 14:44:29 - INFO - codeparrot_training - Step 1445: {'lr': 0.00036125, 'samples': 277632, 'steps': 1445, 'loss/train': 1.4898181557655334} 01/28/2022 14:44:33 - INFO - codeparrot_training - Step 1446: {'lr': 0.0003615, 'samples': 277824, 'steps': 1446, 'loss/train': 1.7611811757087708} 01/28/2022 14:44:38 - INFO - codeparrot_training - Step 1447: {'lr': 0.00036175000000000004, 'samples': 278016, 'steps': 1447, 'loss/train': 1.6122309565544128} 01/28/2022 14:44:42 - INFO - codeparrot_training - Step 1448: {'lr': 0.000362, 'samples': 278208, 'steps': 1448, 'loss/train': 1.9820764660835266} 01/28/2022 14:44:46 - INFO - codeparrot_training - Step 1449: {'lr': 0.00036225000000000005, 'samples': 278400, 'steps': 1449, 'loss/train': 2.3526028990745544} 01/28/2022 14:44:52 - INFO - codeparrot_training - Step 1450: {'lr': 0.0003625, 'samples': 278592, 'steps': 1450, 'loss/train': 5.81442403793335} 01/28/2022 14:44:56 - INFO - codeparrot_training - Step 1451: {'lr': 0.00036275, 'samples': 278784, 'steps': 1451, 'loss/train': 2.156972587108612} 01/28/2022 14:45:00 - INFO - codeparrot_training - Step 1452: {'lr': 0.000363, 'samples': 278976, 'steps': 1452, 'loss/train': 2.5040354132652283} 01/28/2022 14:45:05 - INFO - codeparrot_training - Step 1453: {'lr': 0.00036325, 'samples': 279168, 'steps': 1453, 'loss/train': 2.108668327331543} 01/28/2022 14:45:09 - INFO - codeparrot_training - Step 1454: {'lr': 0.0003635, 'samples': 279360, 'steps': 1454, 'loss/train': 1.5375625491142273} 01/28/2022 14:45:14 - INFO - codeparrot_training - Step 1455: {'lr': 0.00036375000000000003, 'samples': 279552, 'steps': 1455, 'loss/train': 1.1961624026298523} 01/28/2022 14:45:18 - INFO - codeparrot_training - Step 1456: {'lr': 0.000364, 'samples': 279744, 'steps': 1456, 'loss/train': 1.9710862040519714} 01/28/2022 14:45:22 - INFO - codeparrot_training - Step 1457: {'lr': 0.00036425000000000004, 'samples': 279936, 'steps': 1457, 'loss/train': 1.3255140781402588} 01/28/2022 14:45:26 - INFO - codeparrot_training - Step 1458: {'lr': 0.0003645, 'samples': 280128, 'steps': 1458, 'loss/train': 1.753333568572998} 01/28/2022 14:45:31 - INFO - codeparrot_training - Step 1459: {'lr': 0.00036475, 'samples': 280320, 'steps': 1459, 'loss/train': 2.3563835620880127} 01/28/2022 14:45:36 - INFO - codeparrot_training - Step 1460: {'lr': 0.000365, 'samples': 280512, 'steps': 1460, 'loss/train': 1.7614874839782715} 01/28/2022 14:45:40 - INFO - codeparrot_training - Step 1461: {'lr': 0.00036525, 'samples': 280704, 'steps': 1461, 'loss/train': 2.208867609500885} 01/28/2022 14:45:44 - INFO - codeparrot_training - Step 1462: {'lr': 0.0003655, 'samples': 280896, 'steps': 1462, 'loss/train': 1.0331354141235352} 01/28/2022 14:45:48 - INFO - codeparrot_training - Step 1463: {'lr': 0.00036575, 'samples': 281088, 'steps': 1463, 'loss/train': 2.048400342464447} 01/28/2022 14:45:53 - INFO - codeparrot_training - Step 1464: {'lr': 0.000366, 'samples': 281280, 'steps': 1464, 'loss/train': 1.803429365158081} 01/28/2022 14:45:58 - INFO - codeparrot_training - Step 1465: {'lr': 0.00036625000000000004, 'samples': 281472, 'steps': 1465, 'loss/train': 1.9410638809204102} 01/28/2022 14:46:02 - INFO - codeparrot_training - Step 1466: {'lr': 0.0003665, 'samples': 281664, 'steps': 1466, 'loss/train': 1.6527809500694275} 01/28/2022 14:46:06 - INFO - codeparrot_training - Step 1467: {'lr': 0.00036675000000000005, 'samples': 281856, 'steps': 1467, 'loss/train': 2.056789219379425} 01/28/2022 14:46:10 - INFO - codeparrot_training - Step 1468: {'lr': 0.000367, 'samples': 282048, 'steps': 1468, 'loss/train': 0.2910330668091774} 01/28/2022 14:46:14 - INFO - codeparrot_training - Step 1469: {'lr': 0.00036725, 'samples': 282240, 'steps': 1469, 'loss/train': 2.420259654521942} 01/28/2022 14:46:21 - INFO - codeparrot_training - Step 1470: {'lr': 0.0003675, 'samples': 282432, 'steps': 1470, 'loss/train': 2.049958348274231} 01/28/2022 14:46:25 - INFO - codeparrot_training - Step 1471: {'lr': 0.00036775, 'samples': 282624, 'steps': 1471, 'loss/train': 2.6305498480796814} 01/28/2022 14:46:29 - INFO - codeparrot_training - Step 1472: {'lr': 0.000368, 'samples': 282816, 'steps': 1472, 'loss/train': 2.188101589679718} 01/28/2022 14:46:33 - INFO - codeparrot_training - Step 1473: {'lr': 0.00036825000000000003, 'samples': 283008, 'steps': 1473, 'loss/train': 2.267856538295746} 01/28/2022 14:46:38 - INFO - codeparrot_training - Step 1474: {'lr': 0.0003685, 'samples': 283200, 'steps': 1474, 'loss/train': 2.493461787700653} 01/28/2022 14:46:43 - INFO - codeparrot_training - Step 1475: {'lr': 0.00036875000000000005, 'samples': 283392, 'steps': 1475, 'loss/train': 2.1701560020446777} 01/28/2022 14:46:47 - INFO - codeparrot_training - Step 1476: {'lr': 0.000369, 'samples': 283584, 'steps': 1476, 'loss/train': 0.6946100145578384} 01/28/2022 14:46:51 - INFO - codeparrot_training - Step 1477: {'lr': 0.00036925, 'samples': 283776, 'steps': 1477, 'loss/train': 2.55524218082428} 01/28/2022 14:46:55 - INFO - codeparrot_training - Step 1478: {'lr': 0.0003695, 'samples': 283968, 'steps': 1478, 'loss/train': 1.431407868862152} 01/28/2022 14:46:59 - INFO - codeparrot_training - Step 1479: {'lr': 0.00036975, 'samples': 284160, 'steps': 1479, 'loss/train': 1.868682861328125} 01/28/2022 14:47:06 - INFO - codeparrot_training - Step 1480: {'lr': 0.00037, 'samples': 284352, 'steps': 1480, 'loss/train': 2.0764070749282837} 01/28/2022 14:47:10 - INFO - codeparrot_training - Step 1481: {'lr': 0.00037025000000000003, 'samples': 284544, 'steps': 1481, 'loss/train': 2.008040428161621} 01/28/2022 14:47:14 - INFO - codeparrot_training - Step 1482: {'lr': 0.0003705, 'samples': 284736, 'steps': 1482, 'loss/train': 2.2208288311958313} 01/28/2022 14:47:18 - INFO - codeparrot_training - Step 1483: {'lr': 0.00037075000000000004, 'samples': 284928, 'steps': 1483, 'loss/train': 2.450572192668915} 01/28/2022 14:47:22 - INFO - codeparrot_training - Step 1484: {'lr': 0.000371, 'samples': 285120, 'steps': 1484, 'loss/train': 2.3357478976249695} 01/28/2022 14:47:27 - INFO - codeparrot_training - Step 1485: {'lr': 0.00037125000000000005, 'samples': 285312, 'steps': 1485, 'loss/train': 2.0900675654411316} 01/28/2022 14:47:32 - INFO - codeparrot_training - Step 1486: {'lr': 0.00037150000000000003, 'samples': 285504, 'steps': 1486, 'loss/train': 2.0451951026916504} 01/28/2022 14:47:36 - INFO - codeparrot_training - Step 1487: {'lr': 0.00037175, 'samples': 285696, 'steps': 1487, 'loss/train': 2.0261921882629395} 01/28/2022 14:47:40 - INFO - codeparrot_training - Step 1488: {'lr': 0.000372, 'samples': 285888, 'steps': 1488, 'loss/train': 1.8156936764717102} 01/28/2022 14:47:44 - INFO - codeparrot_training - Step 1489: {'lr': 0.00037225, 'samples': 286080, 'steps': 1489, 'loss/train': 5.282412171363831} 01/28/2022 14:47:49 - INFO - codeparrot_training - Step 1490: {'lr': 0.0003725, 'samples': 286272, 'steps': 1490, 'loss/train': 1.2140560448169708} 01/28/2022 14:47:53 - INFO - codeparrot_training - Step 1491: {'lr': 0.00037275000000000003, 'samples': 286464, 'steps': 1491, 'loss/train': 2.3014842867851257} 01/28/2022 14:47:58 - INFO - codeparrot_training - Step 1492: {'lr': 0.000373, 'samples': 286656, 'steps': 1492, 'loss/train': 1.9697195291519165} 01/28/2022 14:48:02 - INFO - codeparrot_training - Step 1493: {'lr': 0.00037325000000000005, 'samples': 286848, 'steps': 1493, 'loss/train': 2.0082162022590637} 01/28/2022 14:48:06 - INFO - codeparrot_training - Step 1494: {'lr': 0.0003735, 'samples': 287040, 'steps': 1494, 'loss/train': 1.4459958672523499} 01/28/2022 14:48:12 - INFO - codeparrot_training - Step 1495: {'lr': 0.00037375000000000006, 'samples': 287232, 'steps': 1495, 'loss/train': 2.1181976795196533} 01/28/2022 14:48:16 - INFO - codeparrot_training - Step 1496: {'lr': 0.000374, 'samples': 287424, 'steps': 1496, 'loss/train': 2.017041027545929} 01/28/2022 14:48:21 - INFO - codeparrot_training - Step 1497: {'lr': 0.00037425, 'samples': 287616, 'steps': 1497, 'loss/train': 2.422980487346649} 01/28/2022 14:48:25 - INFO - codeparrot_training - Step 1498: {'lr': 0.0003745, 'samples': 287808, 'steps': 1498, 'loss/train': 2.0822367668151855} 01/28/2022 14:48:29 - INFO - codeparrot_training - Step 1499: {'lr': 0.00037475000000000003, 'samples': 288000, 'steps': 1499, 'loss/train': 1.9356159567832947} 01/28/2022 14:48:34 - INFO - codeparrot_training - Step 1500: {'lr': 0.000375, 'samples': 288192, 'steps': 1500, 'loss/train': 2.2283747792243958} 01/28/2022 14:48:38 - INFO - codeparrot_training - Step 1501: {'lr': 0.00037525, 'samples': 288384, 'steps': 1501, 'loss/train': 2.47530859708786} 01/28/2022 14:48:42 - INFO - codeparrot_training - Step 1502: {'lr': 0.0003755, 'samples': 288576, 'steps': 1502, 'loss/train': 2.23889297246933} 01/28/2022 14:48:46 - INFO - codeparrot_training - Step 1503: {'lr': 0.00037575, 'samples': 288768, 'steps': 1503, 'loss/train': 0.9162327647209167} 01/28/2022 14:48:51 - INFO - codeparrot_training - Step 1504: {'lr': 0.00037600000000000003, 'samples': 288960, 'steps': 1504, 'loss/train': 1.7732704281806946} 01/28/2022 14:48:56 - INFO - codeparrot_training - Step 1505: {'lr': 0.00037624999999999996, 'samples': 289152, 'steps': 1505, 'loss/train': 2.2343228459358215} 01/28/2022 14:49:00 - INFO - codeparrot_training - Step 1506: {'lr': 0.0003765, 'samples': 289344, 'steps': 1506, 'loss/train': 2.6932132244110107} 01/28/2022 14:49:04 - INFO - codeparrot_training - Step 1507: {'lr': 0.00037674999999999997, 'samples': 289536, 'steps': 1507, 'loss/train': 2.2434725761413574} 01/28/2022 14:49:08 - INFO - codeparrot_training - Step 1508: {'lr': 0.000377, 'samples': 289728, 'steps': 1508, 'loss/train': 2.698453903198242} 01/28/2022 14:49:12 - INFO - codeparrot_training - Step 1509: {'lr': 0.00037725, 'samples': 289920, 'steps': 1509, 'loss/train': 2.183669865131378} 01/28/2022 14:49:19 - INFO - codeparrot_training - Step 1510: {'lr': 0.0003775, 'samples': 290112, 'steps': 1510, 'loss/train': 2.2274338603019714} 01/28/2022 14:49:23 - INFO - codeparrot_training - Step 1511: {'lr': 0.00037775, 'samples': 290304, 'steps': 1511, 'loss/train': 1.710213840007782} 01/28/2022 14:49:27 - INFO - codeparrot_training - Step 1512: {'lr': 0.000378, 'samples': 290496, 'steps': 1512, 'loss/train': 2.0519365668296814} 01/28/2022 14:49:31 - INFO - codeparrot_training - Step 1513: {'lr': 0.00037825, 'samples': 290688, 'steps': 1513, 'loss/train': 1.6971901059150696} 01/28/2022 14:49:35 - INFO - codeparrot_training - Step 1514: {'lr': 0.0003785, 'samples': 290880, 'steps': 1514, 'loss/train': 2.0393988490104675} 01/28/2022 14:49:40 - INFO - codeparrot_training - Step 1515: {'lr': 0.00037874999999999996, 'samples': 291072, 'steps': 1515, 'loss/train': 3.0614705085754395} 01/28/2022 14:49:45 - INFO - codeparrot_training - Step 1516: {'lr': 0.000379, 'samples': 291264, 'steps': 1516, 'loss/train': 1.6879543662071228} 01/28/2022 14:49:49 - INFO - codeparrot_training - Step 1517: {'lr': 0.00037925, 'samples': 291456, 'steps': 1517, 'loss/train': 2.6892834305763245} 01/28/2022 14:49:53 - INFO - codeparrot_training - Step 1518: {'lr': 0.0003795, 'samples': 291648, 'steps': 1518, 'loss/train': 3.0960044860839844} 01/28/2022 14:49:57 - INFO - codeparrot_training - Step 1519: {'lr': 0.00037975, 'samples': 291840, 'steps': 1519, 'loss/train': 1.5748910307884216} 01/28/2022 14:50:02 - INFO - codeparrot_training - Step 1520: {'lr': 0.00038, 'samples': 292032, 'steps': 1520, 'loss/train': 3.128501772880554} 01/28/2022 14:50:06 - INFO - codeparrot_training - Step 1521: {'lr': 0.00038025, 'samples': 292224, 'steps': 1521, 'loss/train': 2.4408706426620483} 01/28/2022 14:50:11 - INFO - codeparrot_training - Step 1522: {'lr': 0.00038050000000000003, 'samples': 292416, 'steps': 1522, 'loss/train': 2.1497976779937744} 01/28/2022 14:50:15 - INFO - codeparrot_training - Step 1523: {'lr': 0.00038075, 'samples': 292608, 'steps': 1523, 'loss/train': 2.4055817127227783} 01/28/2022 14:50:19 - INFO - codeparrot_training - Step 1524: {'lr': 0.000381, 'samples': 292800, 'steps': 1524, 'loss/train': 2.161305785179138} 01/28/2022 14:50:24 - INFO - codeparrot_training - Step 1525: {'lr': 0.00038124999999999997, 'samples': 292992, 'steps': 1525, 'loss/train': 2.2210607528686523} 01/28/2022 14:50:28 - INFO - codeparrot_training - Step 1526: {'lr': 0.0003815, 'samples': 293184, 'steps': 1526, 'loss/train': 2.2888365983963013} 01/28/2022 14:50:32 - INFO - codeparrot_training - Step 1527: {'lr': 0.00038175, 'samples': 293376, 'steps': 1527, 'loss/train': 2.5658324360847473} 01/28/2022 14:50:36 - INFO - codeparrot_training - Step 1528: {'lr': 0.000382, 'samples': 293568, 'steps': 1528, 'loss/train': 2.258786916732788} 01/28/2022 14:50:41 - INFO - codeparrot_training - Step 1529: {'lr': 0.00038225, 'samples': 293760, 'steps': 1529, 'loss/train': 2.226496696472168} 01/28/2022 14:50:46 - INFO - codeparrot_training - Step 1530: {'lr': 0.00038250000000000003, 'samples': 293952, 'steps': 1530, 'loss/train': 1.3805146515369415} 01/28/2022 14:50:51 - INFO - codeparrot_training - Step 1531: {'lr': 0.00038275, 'samples': 294144, 'steps': 1531, 'loss/train': 1.6537193655967712} 01/28/2022 14:50:55 - INFO - codeparrot_training - Step 1532: {'lr': 0.00038300000000000004, 'samples': 294336, 'steps': 1532, 'loss/train': 2.1961310505867004} 01/28/2022 14:50:59 - INFO - codeparrot_training - Step 1533: {'lr': 0.00038324999999999996, 'samples': 294528, 'steps': 1533, 'loss/train': 1.9388673305511475} 01/28/2022 14:51:03 - INFO - codeparrot_training - Step 1534: {'lr': 0.0003835, 'samples': 294720, 'steps': 1534, 'loss/train': 1.9241986870765686} 01/28/2022 14:51:09 - INFO - codeparrot_training - Step 1535: {'lr': 0.00038375, 'samples': 294912, 'steps': 1535, 'loss/train': 1.3160555362701416} 01/28/2022 14:51:13 - INFO - codeparrot_training - Step 1536: {'lr': 0.000384, 'samples': 295104, 'steps': 1536, 'loss/train': 2.755121111869812} 01/28/2022 14:51:17 - INFO - codeparrot_training - Step 1537: {'lr': 0.00038425, 'samples': 295296, 'steps': 1537, 'loss/train': 1.6110917329788208} 01/28/2022 14:51:21 - INFO - codeparrot_training - Step 1538: {'lr': 0.0003845, 'samples': 295488, 'steps': 1538, 'loss/train': 1.6871750950813293} 01/28/2022 14:51:25 - INFO - codeparrot_training - Step 1539: {'lr': 0.00038475, 'samples': 295680, 'steps': 1539, 'loss/train': 1.5034343004226685} 01/28/2022 14:51:32 - INFO - codeparrot_training - Step 1540: {'lr': 0.00038500000000000003, 'samples': 295872, 'steps': 1540, 'loss/train': 3.7341595888137817} 01/28/2022 14:51:36 - INFO - codeparrot_training - Step 1541: {'lr': 0.00038525, 'samples': 296064, 'steps': 1541, 'loss/train': 4.917869567871094} 01/28/2022 14:51:40 - INFO - codeparrot_training - Step 1542: {'lr': 0.0003855, 'samples': 296256, 'steps': 1542, 'loss/train': 1.7549504041671753} 01/28/2022 14:51:44 - INFO - codeparrot_training - Step 1543: {'lr': 0.00038574999999999997, 'samples': 296448, 'steps': 1543, 'loss/train': 2.1063138842582703} 01/28/2022 14:51:48 - INFO - codeparrot_training - Step 1544: {'lr': 0.000386, 'samples': 296640, 'steps': 1544, 'loss/train': 2.4220043420791626} 01/28/2022 14:51:54 - INFO - codeparrot_training - Step 1545: {'lr': 0.00038625, 'samples': 296832, 'steps': 1545, 'loss/train': 2.3725165128707886} 01/28/2022 14:51:58 - INFO - codeparrot_training - Step 1546: {'lr': 0.0003865, 'samples': 297024, 'steps': 1546, 'loss/train': 2.071177661418915} 01/28/2022 14:52:02 - INFO - codeparrot_training - Step 1547: {'lr': 0.00038675, 'samples': 297216, 'steps': 1547, 'loss/train': 2.9652355313301086} 01/28/2022 14:52:06 - INFO - codeparrot_training - Step 1548: {'lr': 0.00038700000000000003, 'samples': 297408, 'steps': 1548, 'loss/train': 1.3161247372627258} 01/28/2022 14:52:10 - INFO - codeparrot_training - Step 1549: {'lr': 0.00038725, 'samples': 297600, 'steps': 1549, 'loss/train': 2.222547769546509} 01/28/2022 14:52:15 - INFO - codeparrot_training - Step 1550: {'lr': 0.00038750000000000004, 'samples': 297792, 'steps': 1550, 'loss/train': 1.5121116042137146} 01/28/2022 14:52:19 - INFO - codeparrot_training - Step 1551: {'lr': 0.00038774999999999997, 'samples': 297984, 'steps': 1551, 'loss/train': 1.4439716935157776} 01/28/2022 14:52:24 - INFO - codeparrot_training - Step 1552: {'lr': 0.000388, 'samples': 298176, 'steps': 1552, 'loss/train': 1.8827735781669617} 01/28/2022 14:52:28 - INFO - codeparrot_training - Step 1553: {'lr': 0.00038825, 'samples': 298368, 'steps': 1553, 'loss/train': 0.9111367464065552} 01/28/2022 14:52:32 - INFO - codeparrot_training - Step 1554: {'lr': 0.0003885, 'samples': 298560, 'steps': 1554, 'loss/train': 1.9800823330879211} 01/28/2022 14:52:39 - INFO - codeparrot_training - Step 1555: {'lr': 0.00038875, 'samples': 298752, 'steps': 1555, 'loss/train': 2.2939887642860413} 01/28/2022 14:52:43 - INFO - codeparrot_training - Step 1556: {'lr': 0.000389, 'samples': 298944, 'steps': 1556, 'loss/train': 1.6776552200317383} 01/28/2022 14:52:48 - INFO - codeparrot_training - Step 1557: {'lr': 0.00038925, 'samples': 299136, 'steps': 1557, 'loss/train': 2.02366840839386} 01/28/2022 14:52:52 - INFO - codeparrot_training - Step 1558: {'lr': 0.00038950000000000003, 'samples': 299328, 'steps': 1558, 'loss/train': 1.9503241181373596} 01/28/2022 14:52:57 - INFO - codeparrot_training - Step 1559: {'lr': 0.00038975, 'samples': 299520, 'steps': 1559, 'loss/train': 2.1594293117523193} 01/28/2022 14:53:01 - INFO - codeparrot_training - Step 1560: {'lr': 0.00039000000000000005, 'samples': 299712, 'steps': 1560, 'loss/train': 1.3216427564620972} 01/28/2022 14:53:05 - INFO - codeparrot_training - Step 1561: {'lr': 0.00039024999999999997, 'samples': 299904, 'steps': 1561, 'loss/train': 2.770711898803711} 01/28/2022 14:53:09 - INFO - codeparrot_training - Step 1562: {'lr': 0.0003905, 'samples': 300096, 'steps': 1562, 'loss/train': 1.7866830825805664} 01/28/2022 14:53:14 - INFO - codeparrot_training - Step 1563: {'lr': 0.00039075, 'samples': 300288, 'steps': 1563, 'loss/train': 1.454403966665268} 01/28/2022 14:53:20 - INFO - codeparrot_training - Step 1564: {'lr': 0.000391, 'samples': 300480, 'steps': 1564, 'loss/train': 1.819336473941803} 01/28/2022 14:53:24 - INFO - codeparrot_training - Step 1565: {'lr': 0.00039125, 'samples': 300672, 'steps': 1565, 'loss/train': 1.8993301391601562} 01/28/2022 14:53:28 - INFO - codeparrot_training - Step 1566: {'lr': 0.00039150000000000003, 'samples': 300864, 'steps': 1566, 'loss/train': 1.571977436542511} 01/28/2022 14:53:32 - INFO - codeparrot_training - Step 1567: {'lr': 0.00039175, 'samples': 301056, 'steps': 1567, 'loss/train': 2.043596863746643} 01/28/2022 14:53:36 - INFO - codeparrot_training - Step 1568: {'lr': 0.00039200000000000004, 'samples': 301248, 'steps': 1568, 'loss/train': 1.6348367929458618} 01/28/2022 14:53:41 - INFO - codeparrot_training - Step 1569: {'lr': 0.00039225, 'samples': 301440, 'steps': 1569, 'loss/train': 1.4605465829372406} 01/28/2022 14:53:46 - INFO - codeparrot_training - Step 1570: {'lr': 0.0003925, 'samples': 301632, 'steps': 1570, 'loss/train': 1.304221361875534} 01/28/2022 14:53:50 - INFO - codeparrot_training - Step 1571: {'lr': 0.00039275, 'samples': 301824, 'steps': 1571, 'loss/train': 1.7429915070533752} 01/28/2022 14:53:54 - INFO - codeparrot_training - Step 1572: {'lr': 0.000393, 'samples': 302016, 'steps': 1572, 'loss/train': 1.5400300025939941} 01/28/2022 14:53:58 - INFO - codeparrot_training - Step 1573: {'lr': 0.00039325, 'samples': 302208, 'steps': 1573, 'loss/train': 1.7717957496643066} 01/28/2022 14:54:03 - INFO - codeparrot_training - Step 1574: {'lr': 0.0003935, 'samples': 302400, 'steps': 1574, 'loss/train': 1.8246329426765442} 01/28/2022 14:54:07 - INFO - codeparrot_training - Step 1575: {'lr': 0.00039375, 'samples': 302592, 'steps': 1575, 'loss/train': 2.052974581718445} 01/28/2022 14:54:12 - INFO - codeparrot_training - Step 1576: {'lr': 0.00039400000000000004, 'samples': 302784, 'steps': 1576, 'loss/train': 2.9031431078910828} 01/28/2022 14:54:16 - INFO - codeparrot_training - Step 1577: {'lr': 0.00039425, 'samples': 302976, 'steps': 1577, 'loss/train': 1.974985420703888} 01/28/2022 14:54:20 - INFO - codeparrot_training - Step 1578: {'lr': 0.00039450000000000005, 'samples': 303168, 'steps': 1578, 'loss/train': 2.195119857788086} 01/28/2022 14:54:25 - INFO - codeparrot_training - Step 1579: {'lr': 0.00039474999999999997, 'samples': 303360, 'steps': 1579, 'loss/train': 2.2695658206939697} 01/28/2022 14:54:29 - INFO - codeparrot_training - Step 1580: {'lr': 0.000395, 'samples': 303552, 'steps': 1580, 'loss/train': 1.856881320476532} 01/28/2022 14:54:34 - INFO - codeparrot_training - Step 1581: {'lr': 0.00039525, 'samples': 303744, 'steps': 1581, 'loss/train': 1.7255542874336243} 01/28/2022 14:54:38 - INFO - codeparrot_training - Step 1582: {'lr': 0.0003955, 'samples': 303936, 'steps': 1582, 'loss/train': 1.7201820015907288} 01/28/2022 14:54:42 - INFO - codeparrot_training - Step 1583: {'lr': 0.00039575, 'samples': 304128, 'steps': 1583, 'loss/train': 1.3493105471134186} 01/28/2022 14:54:49 - INFO - codeparrot_training - Step 1584: {'lr': 0.00039600000000000003, 'samples': 304320, 'steps': 1584, 'loss/train': 2.6714154481887817} 01/28/2022 14:54:53 - INFO - codeparrot_training - Step 1585: {'lr': 0.00039625, 'samples': 304512, 'steps': 1585, 'loss/train': 2.430878520011902} 01/28/2022 14:54:57 - INFO - codeparrot_training - Step 1586: {'lr': 0.00039650000000000004, 'samples': 304704, 'steps': 1586, 'loss/train': 0.5384821146726608} 01/28/2022 14:55:01 - INFO - codeparrot_training - Step 1587: {'lr': 0.00039675, 'samples': 304896, 'steps': 1587, 'loss/train': 2.0079029202461243} 01/28/2022 14:55:05 - INFO - codeparrot_training - Step 1588: {'lr': 0.00039700000000000005, 'samples': 305088, 'steps': 1588, 'loss/train': 1.7119141817092896} 01/28/2022 14:55:10 - INFO - codeparrot_training - Step 1589: {'lr': 0.00039725, 'samples': 305280, 'steps': 1589, 'loss/train': 1.2689969837665558} 01/28/2022 14:55:15 - INFO - codeparrot_training - Step 1590: {'lr': 0.0003975, 'samples': 305472, 'steps': 1590, 'loss/train': 1.8829101920127869} 01/28/2022 14:55:19 - INFO - codeparrot_training - Step 1591: {'lr': 0.00039775, 'samples': 305664, 'steps': 1591, 'loss/train': 2.055941104888916} 01/28/2022 14:55:23 - INFO - codeparrot_training - Step 1592: {'lr': 0.000398, 'samples': 305856, 'steps': 1592, 'loss/train': 2.046999156475067} 01/28/2022 14:55:27 - INFO - codeparrot_training - Step 1593: {'lr': 0.00039825, 'samples': 306048, 'steps': 1593, 'loss/train': 2.1261996030807495} 01/28/2022 14:55:32 - INFO - codeparrot_training - Step 1594: {'lr': 0.00039850000000000004, 'samples': 306240, 'steps': 1594, 'loss/train': 1.079028457403183} 01/28/2022 14:55:37 - INFO - codeparrot_training - Step 1595: {'lr': 0.00039875, 'samples': 306432, 'steps': 1595, 'loss/train': 1.0516338050365448} 01/28/2022 14:55:41 - INFO - codeparrot_training - Step 1596: {'lr': 0.00039900000000000005, 'samples': 306624, 'steps': 1596, 'loss/train': 2.7042585611343384} 01/28/2022 14:55:45 - INFO - codeparrot_training - Step 1597: {'lr': 0.00039925000000000003, 'samples': 306816, 'steps': 1597, 'loss/train': 2.1264363527297974} 01/28/2022 14:55:49 - INFO - codeparrot_training - Step 1598: {'lr': 0.0003995, 'samples': 307008, 'steps': 1598, 'loss/train': 1.552732229232788} 01/28/2022 14:55:55 - INFO - codeparrot_training - Step 1599: {'lr': 0.00039975, 'samples': 307200, 'steps': 1599, 'loss/train': 0.6474548131227493} 01/28/2022 14:55:59 - INFO - codeparrot_training - Step 1600: {'lr': 0.0004, 'samples': 307392, 'steps': 1600, 'loss/train': 1.4992677569389343} 01/28/2022 14:56:04 - INFO - codeparrot_training - Step 1601: {'lr': 0.00040025, 'samples': 307584, 'steps': 1601, 'loss/train': 1.9685404300689697} 01/28/2022 14:56:08 - INFO - codeparrot_training - Step 1602: {'lr': 0.00040050000000000003, 'samples': 307776, 'steps': 1602, 'loss/train': 1.4442741572856903} 01/28/2022 14:56:12 - INFO - codeparrot_training - Step 1603: {'lr': 0.00040075, 'samples': 307968, 'steps': 1603, 'loss/train': 2.2493221163749695} 01/28/2022 14:56:17 - INFO - codeparrot_training - Step 1604: {'lr': 0.00040100000000000004, 'samples': 308160, 'steps': 1604, 'loss/train': 1.1082216203212738} 01/28/2022 14:56:21 - INFO - codeparrot_training - Step 1605: {'lr': 0.00040125, 'samples': 308352, 'steps': 1605, 'loss/train': 1.1565055847167969} 01/28/2022 14:56:25 - INFO - codeparrot_training - Step 1606: {'lr': 0.00040150000000000006, 'samples': 308544, 'steps': 1606, 'loss/train': 2.3579558730125427} 01/28/2022 14:56:29 - INFO - codeparrot_training - Step 1607: {'lr': 0.00040175, 'samples': 308736, 'steps': 1607, 'loss/train': 2.4444966316223145} 01/28/2022 14:56:34 - INFO - codeparrot_training - Step 1608: {'lr': 0.000402, 'samples': 308928, 'steps': 1608, 'loss/train': 2.744178593158722} 01/28/2022 14:56:40 - INFO - codeparrot_training - Step 1609: {'lr': 0.00040225, 'samples': 309120, 'steps': 1609, 'loss/train': 2.1844875812530518} 01/28/2022 14:56:44 - INFO - codeparrot_training - Step 1610: {'lr': 0.0004025, 'samples': 309312, 'steps': 1610, 'loss/train': 1.7995052933692932} 01/28/2022 14:56:48 - INFO - codeparrot_training - Step 1611: {'lr': 0.00040275, 'samples': 309504, 'steps': 1611, 'loss/train': 1.4684443473815918} 01/28/2022 14:56:52 - INFO - codeparrot_training - Step 1612: {'lr': 0.00040300000000000004, 'samples': 309696, 'steps': 1612, 'loss/train': 2.289213001728058} 01/28/2022 14:56:56 - INFO - codeparrot_training - Step 1613: {'lr': 0.00040325, 'samples': 309888, 'steps': 1613, 'loss/train': 1.5443111658096313} 01/28/2022 14:57:02 - INFO - codeparrot_training - Step 1614: {'lr': 0.00040350000000000005, 'samples': 310080, 'steps': 1614, 'loss/train': 1.3979035019874573} 01/28/2022 14:57:06 - INFO - codeparrot_training - Step 1615: {'lr': 0.00040375000000000003, 'samples': 310272, 'steps': 1615, 'loss/train': 2.060191333293915} 01/28/2022 14:57:10 - INFO - codeparrot_training - Step 1616: {'lr': 0.000404, 'samples': 310464, 'steps': 1616, 'loss/train': 1.7262323498725891} 01/28/2022 14:57:14 - INFO - codeparrot_training - Step 1617: {'lr': 0.00040425, 'samples': 310656, 'steps': 1617, 'loss/train': 3.0006362199783325} 01/28/2022 14:57:18 - INFO - codeparrot_training - Step 1618: {'lr': 0.0004045, 'samples': 310848, 'steps': 1618, 'loss/train': 2.334924817085266} 01/28/2022 14:57:24 - INFO - codeparrot_training - Step 1619: {'lr': 0.00040475, 'samples': 311040, 'steps': 1619, 'loss/train': 1.2173604369163513} 01/28/2022 14:57:28 - INFO - codeparrot_training - Step 1620: {'lr': 0.00040500000000000003, 'samples': 311232, 'steps': 1620, 'loss/train': 2.2780741453170776} 01/28/2022 14:57:32 - INFO - codeparrot_training - Step 1621: {'lr': 0.00040525, 'samples': 311424, 'steps': 1621, 'loss/train': 1.9695623517036438} 01/28/2022 14:57:36 - INFO - codeparrot_training - Step 1622: {'lr': 0.00040550000000000004, 'samples': 311616, 'steps': 1622, 'loss/train': 1.168434888124466} 01/28/2022 14:57:40 - INFO - codeparrot_training - Step 1623: {'lr': 0.00040575, 'samples': 311808, 'steps': 1623, 'loss/train': 1.8417105674743652} 01/28/2022 14:57:47 - INFO - codeparrot_training - Step 1624: {'lr': 0.00040600000000000006, 'samples': 312000, 'steps': 1624, 'loss/train': 3.285623073577881} 01/28/2022 14:57:51 - INFO - codeparrot_training - Step 1625: {'lr': 0.00040625000000000004, 'samples': 312192, 'steps': 1625, 'loss/train': 2.0142279267311096} 01/28/2022 14:57:55 - INFO - codeparrot_training - Step 1626: {'lr': 0.00040649999999999996, 'samples': 312384, 'steps': 1626, 'loss/train': 1.3140844702720642} 01/28/2022 14:57:59 - INFO - codeparrot_training - Step 1627: {'lr': 0.00040675, 'samples': 312576, 'steps': 1627, 'loss/train': 2.1175103187561035} 01/28/2022 14:58:03 - INFO - codeparrot_training - Step 1628: {'lr': 0.00040699999999999997, 'samples': 312768, 'steps': 1628, 'loss/train': 1.3695666790008545} 01/28/2022 14:58:09 - INFO - codeparrot_training - Step 1629: {'lr': 0.00040725, 'samples': 312960, 'steps': 1629, 'loss/train': 1.7023687362670898} 01/28/2022 14:58:13 - INFO - codeparrot_training - Step 1630: {'lr': 0.0004075, 'samples': 313152, 'steps': 1630, 'loss/train': 2.311078727245331} 01/28/2022 14:58:17 - INFO - codeparrot_training - Step 1631: {'lr': 0.00040775, 'samples': 313344, 'steps': 1631, 'loss/train': 2.030249834060669} 01/28/2022 14:58:21 - INFO - codeparrot_training - Step 1632: {'lr': 0.000408, 'samples': 313536, 'steps': 1632, 'loss/train': 2.1486525535583496} 01/28/2022 14:58:25 - INFO - codeparrot_training - Step 1633: {'lr': 0.00040825000000000003, 'samples': 313728, 'steps': 1633, 'loss/train': 1.766676127910614} 01/28/2022 14:58:31 - INFO - codeparrot_training - Step 1634: {'lr': 0.0004085, 'samples': 313920, 'steps': 1634, 'loss/train': 1.6279372572898865} 01/28/2022 14:58:35 - INFO - codeparrot_training - Step 1635: {'lr': 0.00040875, 'samples': 314112, 'steps': 1635, 'loss/train': 2.669340133666992} 01/28/2022 14:58:39 - INFO - codeparrot_training - Step 1636: {'lr': 0.00040899999999999997, 'samples': 314304, 'steps': 1636, 'loss/train': 2.5390260815620422} 01/28/2022 14:58:43 - INFO - codeparrot_training - Step 1637: {'lr': 0.00040925, 'samples': 314496, 'steps': 1637, 'loss/train': 1.9054071307182312} 01/28/2022 14:58:47 - INFO - codeparrot_training - Step 1638: {'lr': 0.0004095, 'samples': 314688, 'steps': 1638, 'loss/train': 2.244702458381653} 01/28/2022 14:58:53 - INFO - codeparrot_training - Step 1639: {'lr': 0.00040975, 'samples': 314880, 'steps': 1639, 'loss/train': 2.203292191028595} 01/28/2022 14:58:57 - INFO - codeparrot_training - Step 1640: {'lr': 0.00041, 'samples': 315072, 'steps': 1640, 'loss/train': 2.362157106399536} 01/28/2022 14:59:01 - INFO - codeparrot_training - Step 1641: {'lr': 0.00041025, 'samples': 315264, 'steps': 1641, 'loss/train': 2.3301604986190796} 01/28/2022 14:59:05 - INFO - codeparrot_training - Step 1642: {'lr': 0.0004105, 'samples': 315456, 'steps': 1642, 'loss/train': 2.079732298851013} 01/28/2022 14:59:12 - INFO - codeparrot_training - Step 1643: {'lr': 0.00041075000000000004, 'samples': 315648, 'steps': 1643, 'loss/train': 2.094905734062195} 01/28/2022 14:59:17 - INFO - codeparrot_training - Step 1644: {'lr': 0.00041099999999999996, 'samples': 315840, 'steps': 1644, 'loss/train': 0.8405170440673828} 01/28/2022 14:59:21 - INFO - codeparrot_training - Step 1645: {'lr': 0.00041125, 'samples': 316032, 'steps': 1645, 'loss/train': 2.3521072268486023} 01/28/2022 14:59:25 - INFO - codeparrot_training - Step 1646: {'lr': 0.0004115, 'samples': 316224, 'steps': 1646, 'loss/train': 1.6947895288467407} 01/28/2022 14:59:29 - INFO - codeparrot_training - Step 1647: {'lr': 0.00041175, 'samples': 316416, 'steps': 1647, 'loss/train': 2.3951185941696167} 01/28/2022 14:59:33 - INFO - codeparrot_training - Step 1648: {'lr': 0.000412, 'samples': 316608, 'steps': 1648, 'loss/train': 4.5501344203948975} 01/28/2022 14:59:39 - INFO - codeparrot_training - Step 1649: {'lr': 0.00041225, 'samples': 316800, 'steps': 1649, 'loss/train': 2.2973998188972473} 01/28/2022 14:59:43 - INFO - codeparrot_training - Step 1650: {'lr': 0.0004125, 'samples': 316992, 'steps': 1650, 'loss/train': 1.903662085533142} 01/28/2022 14:59:47 - INFO - codeparrot_training - Step 1651: {'lr': 0.00041275000000000003, 'samples': 317184, 'steps': 1651, 'loss/train': 1.7762771844863892} 01/28/2022 14:59:51 - INFO - codeparrot_training - Step 1652: {'lr': 0.000413, 'samples': 317376, 'steps': 1652, 'loss/train': 2.1064703464508057} 01/28/2022 14:59:55 - INFO - codeparrot_training - Step 1653: {'lr': 0.00041325, 'samples': 317568, 'steps': 1653, 'loss/train': 1.0577902793884277} 01/28/2022 15:00:01 - INFO - codeparrot_training - Step 1654: {'lr': 0.00041349999999999997, 'samples': 317760, 'steps': 1654, 'loss/train': 1.8391735553741455} 01/28/2022 15:00:06 - INFO - codeparrot_training - Step 1655: {'lr': 0.00041375, 'samples': 317952, 'steps': 1655, 'loss/train': 0.8554553389549255} 01/28/2022 15:00:10 - INFO - codeparrot_training - Step 1656: {'lr': 0.000414, 'samples': 318144, 'steps': 1656, 'loss/train': 1.835431694984436} 01/28/2022 15:00:14 - INFO - codeparrot_training - Step 1657: {'lr': 0.00041425, 'samples': 318336, 'steps': 1657, 'loss/train': 1.281811147928238} 01/28/2022 15:00:18 - INFO - codeparrot_training - Step 1658: {'lr': 0.0004145, 'samples': 318528, 'steps': 1658, 'loss/train': 2.6112366914749146} 01/28/2022 15:00:25 - INFO - codeparrot_training - Step 1659: {'lr': 0.00041475, 'samples': 318720, 'steps': 1659, 'loss/train': 2.0089561343193054} 01/28/2022 15:00:29 - INFO - codeparrot_training - Step 1660: {'lr': 0.000415, 'samples': 318912, 'steps': 1660, 'loss/train': 1.629371166229248} 01/28/2022 15:00:33 - INFO - codeparrot_training - Step 1661: {'lr': 0.00041525000000000004, 'samples': 319104, 'steps': 1661, 'loss/train': 1.8215864896774292} 01/28/2022 15:00:37 - INFO - codeparrot_training - Step 1662: {'lr': 0.00041549999999999996, 'samples': 319296, 'steps': 1662, 'loss/train': 1.7904415726661682} 01/28/2022 15:00:41 - INFO - codeparrot_training - Step 1663: {'lr': 0.00041575, 'samples': 319488, 'steps': 1663, 'loss/train': 1.8334554433822632} 01/28/2022 15:00:45 - INFO - codeparrot_training - Step 1664: {'lr': 0.000416, 'samples': 319680, 'steps': 1664, 'loss/train': 1.8769218921661377} 01/28/2022 15:00:51 - INFO - codeparrot_training - Step 1665: {'lr': 0.00041625, 'samples': 319872, 'steps': 1665, 'loss/train': 2.395482122898102} 01/28/2022 15:00:55 - INFO - codeparrot_training - Step 1666: {'lr': 0.0004165, 'samples': 320064, 'steps': 1666, 'loss/train': 1.7101351618766785} 01/28/2022 15:00:59 - INFO - codeparrot_training - Step 1667: {'lr': 0.00041675, 'samples': 320256, 'steps': 1667, 'loss/train': 3.880406141281128} 01/28/2022 15:01:03 - INFO - codeparrot_training - Step 1668: {'lr': 0.000417, 'samples': 320448, 'steps': 1668, 'loss/train': 2.570585310459137} 01/28/2022 15:01:07 - INFO - codeparrot_training - Step 1669: {'lr': 0.00041725000000000003, 'samples': 320640, 'steps': 1669, 'loss/train': 1.573073387145996} 01/28/2022 15:01:14 - INFO - codeparrot_training - Step 1670: {'lr': 0.0004175, 'samples': 320832, 'steps': 1670, 'loss/train': 2.4539020657539368} 01/28/2022 15:01:18 - INFO - codeparrot_training - Step 1671: {'lr': 0.00041775000000000004, 'samples': 321024, 'steps': 1671, 'loss/train': 2.210129678249359} 01/28/2022 15:01:22 - INFO - codeparrot_training - Step 1672: {'lr': 0.00041799999999999997, 'samples': 321216, 'steps': 1672, 'loss/train': 2.5319942235946655} 01/28/2022 15:01:26 - INFO - codeparrot_training - Step 1673: {'lr': 0.00041825, 'samples': 321408, 'steps': 1673, 'loss/train': 1.8601083755493164} 01/28/2022 15:01:30 - INFO - codeparrot_training - Step 1674: {'lr': 0.0004185, 'samples': 321600, 'steps': 1674, 'loss/train': 1.1732657253742218} 01/28/2022 15:01:36 - INFO - codeparrot_training - Step 1675: {'lr': 0.00041875, 'samples': 321792, 'steps': 1675, 'loss/train': 3.626062273979187} 01/28/2022 15:01:40 - INFO - codeparrot_training - Step 1676: {'lr': 0.000419, 'samples': 321984, 'steps': 1676, 'loss/train': 2.64785635471344} 01/28/2022 15:01:44 - INFO - codeparrot_training - Step 1677: {'lr': 0.00041925, 'samples': 322176, 'steps': 1677, 'loss/train': 1.6859341263771057} 01/28/2022 15:01:48 - INFO - codeparrot_training - Step 1678: {'lr': 0.0004195, 'samples': 322368, 'steps': 1678, 'loss/train': 2.2483758330345154} 01/28/2022 15:01:52 - INFO - codeparrot_training - Step 1679: {'lr': 0.00041975000000000004, 'samples': 322560, 'steps': 1679, 'loss/train': 1.9676912426948547} 01/28/2022 15:01:58 - INFO - codeparrot_training - Step 1680: {'lr': 0.00042, 'samples': 322752, 'steps': 1680, 'loss/train': 3.520582437515259} 01/28/2022 15:02:02 - INFO - codeparrot_training - Step 1681: {'lr': 0.00042025, 'samples': 322944, 'steps': 1681, 'loss/train': 2.914309501647949} 01/28/2022 15:02:06 - INFO - codeparrot_training - Step 1682: {'lr': 0.0004205, 'samples': 323136, 'steps': 1682, 'loss/train': 2.315524399280548} 01/28/2022 15:02:10 - INFO - codeparrot_training - Step 1683: {'lr': 0.00042075, 'samples': 323328, 'steps': 1683, 'loss/train': 2.43241024017334} 01/28/2022 15:02:14 - INFO - codeparrot_training - Step 1684: {'lr': 0.000421, 'samples': 323520, 'steps': 1684, 'loss/train': 0.9878323674201965} 01/28/2022 15:02:20 - INFO - codeparrot_training - Step 1685: {'lr': 0.00042125, 'samples': 323712, 'steps': 1685, 'loss/train': 1.6237521171569824} 01/28/2022 15:02:24 - INFO - codeparrot_training - Step 1686: {'lr': 0.0004215, 'samples': 323904, 'steps': 1686, 'loss/train': 2.0063527822494507} 01/28/2022 15:02:29 - INFO - codeparrot_training - Step 1687: {'lr': 0.00042175000000000003, 'samples': 324096, 'steps': 1687, 'loss/train': 2.4223408699035645} 01/28/2022 15:02:33 - INFO - codeparrot_training - Step 1688: {'lr': 0.000422, 'samples': 324288, 'steps': 1688, 'loss/train': 2.6415281295776367} 01/28/2022 15:02:37 - INFO - codeparrot_training - Step 1689: {'lr': 0.00042225000000000005, 'samples': 324480, 'steps': 1689, 'loss/train': 0.4096204340457916} 01/28/2022 15:02:42 - INFO - codeparrot_training - Step 1690: {'lr': 0.00042249999999999997, 'samples': 324672, 'steps': 1690, 'loss/train': 1.2357030808925629} 01/28/2022 15:02:46 - INFO - codeparrot_training - Step 1691: {'lr': 0.00042275, 'samples': 324864, 'steps': 1691, 'loss/train': 1.5208445191383362} 01/28/2022 15:02:50 - INFO - codeparrot_training - Step 1692: {'lr': 0.000423, 'samples': 325056, 'steps': 1692, 'loss/train': 2.450231194496155} 01/28/2022 15:02:55 - INFO - codeparrot_training - Step 1693: {'lr': 0.00042325, 'samples': 325248, 'steps': 1693, 'loss/train': 1.586899995803833} 01/28/2022 15:02:59 - INFO - codeparrot_training - Step 1694: {'lr': 0.0004235, 'samples': 325440, 'steps': 1694, 'loss/train': 2.2934728860855103} 01/28/2022 15:03:04 - INFO - codeparrot_training - Step 1695: {'lr': 0.00042375000000000003, 'samples': 325632, 'steps': 1695, 'loss/train': 0.9699786901473999} 01/28/2022 15:03:08 - INFO - codeparrot_training - Step 1696: {'lr': 0.000424, 'samples': 325824, 'steps': 1696, 'loss/train': 1.982442855834961} 01/28/2022 15:03:12 - INFO - codeparrot_training - Step 1697: {'lr': 0.00042425000000000004, 'samples': 326016, 'steps': 1697, 'loss/train': 1.996907114982605} 01/28/2022 15:03:16 - INFO - codeparrot_training - Step 1698: {'lr': 0.0004245, 'samples': 326208, 'steps': 1698, 'loss/train': 1.810570478439331} 01/28/2022 15:03:21 - INFO - codeparrot_training - Step 1699: {'lr': 0.00042475000000000005, 'samples': 326400, 'steps': 1699, 'loss/train': 2.0933877825737} 01/28/2022 15:03:27 - INFO - codeparrot_training - Step 1700: {'lr': 0.000425, 'samples': 326592, 'steps': 1700, 'loss/train': 2.357014775276184} 01/28/2022 15:03:31 - INFO - codeparrot_training - Step 1701: {'lr': 0.00042525, 'samples': 326784, 'steps': 1701, 'loss/train': 3.04356050491333} 01/28/2022 15:03:35 - INFO - codeparrot_training - Step 1702: {'lr': 0.0004255, 'samples': 326976, 'steps': 1702, 'loss/train': 2.8749386072158813} 01/28/2022 15:03:39 - INFO - codeparrot_training - Step 1703: {'lr': 0.00042575, 'samples': 327168, 'steps': 1703, 'loss/train': 1.8749803304672241} 01/28/2022 15:03:44 - INFO - codeparrot_training - Step 1704: {'lr': 0.000426, 'samples': 327360, 'steps': 1704, 'loss/train': 2.185858905315399} 01/28/2022 15:03:49 - INFO - codeparrot_training - Step 1705: {'lr': 0.00042625000000000003, 'samples': 327552, 'steps': 1705, 'loss/train': 2.8719656467437744} 01/28/2022 15:03:53 - INFO - codeparrot_training - Step 1706: {'lr': 0.0004265, 'samples': 327744, 'steps': 1706, 'loss/train': 2.5628678798675537} 01/28/2022 15:03:57 - INFO - codeparrot_training - Step 1707: {'lr': 0.00042675000000000005, 'samples': 327936, 'steps': 1707, 'loss/train': 2.082228183746338} 01/28/2022 15:04:01 - INFO - codeparrot_training - Step 1708: {'lr': 0.000427, 'samples': 328128, 'steps': 1708, 'loss/train': 1.5289018750190735} 01/28/2022 15:04:06 - INFO - codeparrot_training - Step 1709: {'lr': 0.00042725, 'samples': 328320, 'steps': 1709, 'loss/train': 1.3117757141590118} 01/28/2022 15:04:11 - INFO - codeparrot_training - Step 1710: {'lr': 0.0004275, 'samples': 328512, 'steps': 1710, 'loss/train': 2.728825092315674} 01/28/2022 15:04:15 - INFO - codeparrot_training - Step 1711: {'lr': 0.00042775, 'samples': 328704, 'steps': 1711, 'loss/train': 1.4139726161956787} 01/28/2022 15:04:19 - INFO - codeparrot_training - Step 1712: {'lr': 0.000428, 'samples': 328896, 'steps': 1712, 'loss/train': 2.34749436378479} 01/28/2022 15:04:23 - INFO - codeparrot_training - Step 1713: {'lr': 0.00042825000000000003, 'samples': 329088, 'steps': 1713, 'loss/train': 2.1273770928382874} 01/28/2022 15:04:27 - INFO - codeparrot_training - Step 1714: {'lr': 0.0004285, 'samples': 329280, 'steps': 1714, 'loss/train': 2.5097579956054688} 01/28/2022 15:04:34 - INFO - codeparrot_training - Step 1715: {'lr': 0.00042875000000000004, 'samples': 329472, 'steps': 1715, 'loss/train': 0.30567949265241623} 01/28/2022 15:04:38 - INFO - codeparrot_training - Step 1716: {'lr': 0.000429, 'samples': 329664, 'steps': 1716, 'loss/train': 2.365776479244232} 01/28/2022 15:04:42 - INFO - codeparrot_training - Step 1717: {'lr': 0.00042925000000000005, 'samples': 329856, 'steps': 1717, 'loss/train': 2.260462760925293} 01/28/2022 15:04:46 - INFO - codeparrot_training - Step 1718: {'lr': 0.0004295, 'samples': 330048, 'steps': 1718, 'loss/train': 2.4254602789878845} 01/28/2022 15:04:50 - INFO - codeparrot_training - Step 1719: {'lr': 0.00042975, 'samples': 330240, 'steps': 1719, 'loss/train': 2.3259803652763367} 01/28/2022 15:04:56 - INFO - codeparrot_training - Step 1720: {'lr': 0.00043, 'samples': 330432, 'steps': 1720, 'loss/train': 2.3502357602119446} 01/28/2022 15:05:00 - INFO - codeparrot_training - Step 1721: {'lr': 0.00043025, 'samples': 330624, 'steps': 1721, 'loss/train': 2.175258457660675} 01/28/2022 15:05:04 - INFO - codeparrot_training - Step 1722: {'lr': 0.0004305, 'samples': 330816, 'steps': 1722, 'loss/train': 2.3668336272239685} 01/28/2022 15:05:08 - INFO - codeparrot_training - Step 1723: {'lr': 0.00043075000000000003, 'samples': 331008, 'steps': 1723, 'loss/train': 2.3747372031211853} 01/28/2022 15:05:12 - INFO - codeparrot_training - Step 1724: {'lr': 0.000431, 'samples': 331200, 'steps': 1724, 'loss/train': 2.6278324127197266} 01/28/2022 15:05:17 - INFO - codeparrot_training - Step 1725: {'lr': 0.00043125000000000005, 'samples': 331392, 'steps': 1725, 'loss/train': 1.9630934000015259} 01/28/2022 15:05:21 - INFO - codeparrot_training - Step 1726: {'lr': 0.0004315, 'samples': 331584, 'steps': 1726, 'loss/train': 2.7423473596572876} 01/28/2022 15:05:26 - INFO - codeparrot_training - Step 1727: {'lr': 0.00043175, 'samples': 331776, 'steps': 1727, 'loss/train': 2.3206520676612854} 01/28/2022 15:05:30 - INFO - codeparrot_training - Step 1728: {'lr': 0.000432, 'samples': 331968, 'steps': 1728, 'loss/train': 2.275907278060913} 01/28/2022 15:05:34 - INFO - codeparrot_training - Step 1729: {'lr': 0.00043225, 'samples': 332160, 'steps': 1729, 'loss/train': 1.7995970249176025} 01/28/2022 15:05:41 - INFO - codeparrot_training - Step 1730: {'lr': 0.0004325, 'samples': 332352, 'steps': 1730, 'loss/train': 2.902667820453644} 01/28/2022 15:05:45 - INFO - codeparrot_training - Step 1731: {'lr': 0.00043275000000000003, 'samples': 332544, 'steps': 1731, 'loss/train': 1.9734790921211243} 01/28/2022 15:05:49 - INFO - codeparrot_training - Step 1732: {'lr': 0.000433, 'samples': 332736, 'steps': 1732, 'loss/train': 1.4128119349479675} 01/28/2022 15:05:53 - INFO - codeparrot_training - Step 1733: {'lr': 0.00043325000000000004, 'samples': 332928, 'steps': 1733, 'loss/train': 1.578961193561554} 01/28/2022 15:05:57 - INFO - codeparrot_training - Step 1734: {'lr': 0.0004335, 'samples': 333120, 'steps': 1734, 'loss/train': 2.5740053057670593} 01/28/2022 15:06:03 - INFO - codeparrot_training - Step 1735: {'lr': 0.00043375000000000005, 'samples': 333312, 'steps': 1735, 'loss/train': 1.8819192051887512} 01/28/2022 15:06:07 - INFO - codeparrot_training - Step 1736: {'lr': 0.00043400000000000003, 'samples': 333504, 'steps': 1736, 'loss/train': 1.927177369594574} 01/28/2022 15:06:11 - INFO - codeparrot_training - Step 1737: {'lr': 0.00043425, 'samples': 333696, 'steps': 1737, 'loss/train': 1.741814374923706} 01/28/2022 15:06:15 - INFO - codeparrot_training - Step 1738: {'lr': 0.0004345, 'samples': 333888, 'steps': 1738, 'loss/train': 0.773562490940094} 01/28/2022 15:06:19 - INFO - codeparrot_training - Step 1739: {'lr': 0.00043475, 'samples': 334080, 'steps': 1739, 'loss/train': 1.4254007935523987} 01/28/2022 15:06:26 - INFO - codeparrot_training - Step 1740: {'lr': 0.000435, 'samples': 334272, 'steps': 1740, 'loss/train': 1.614522099494934} 01/28/2022 15:06:30 - INFO - codeparrot_training - Step 1741: {'lr': 0.00043525000000000004, 'samples': 334464, 'steps': 1741, 'loss/train': 2.1455236673355103} 01/28/2022 15:06:34 - INFO - codeparrot_training - Step 1742: {'lr': 0.0004355, 'samples': 334656, 'steps': 1742, 'loss/train': 2.255546987056732} 01/28/2022 15:06:38 - INFO - codeparrot_training - Step 1743: {'lr': 0.00043575000000000005, 'samples': 334848, 'steps': 1743, 'loss/train': 1.6016448140144348} 01/28/2022 15:06:42 - INFO - codeparrot_training - Step 1744: {'lr': 0.000436, 'samples': 335040, 'steps': 1744, 'loss/train': 2.650581479072571} 01/28/2022 15:06:48 - INFO - codeparrot_training - Step 1745: {'lr': 0.00043625000000000006, 'samples': 335232, 'steps': 1745, 'loss/train': 2.922725200653076} 01/28/2022 15:06:52 - INFO - codeparrot_training - Step 1746: {'lr': 0.0004365, 'samples': 335424, 'steps': 1746, 'loss/train': 2.0202730894088745} 01/28/2022 15:06:57 - INFO - codeparrot_training - Step 1747: {'lr': 0.00043675, 'samples': 335616, 'steps': 1747, 'loss/train': 2.278334140777588} 01/28/2022 15:07:01 - INFO - codeparrot_training - Step 1748: {'lr': 0.000437, 'samples': 335808, 'steps': 1748, 'loss/train': 2.0173712968826294} 01/28/2022 15:07:05 - INFO - codeparrot_training - Step 1749: {'lr': 0.00043725000000000003, 'samples': 336000, 'steps': 1749, 'loss/train': 2.7109183073043823} 01/28/2022 15:07:09 - INFO - codeparrot_training - Step 1750: {'lr': 0.0004375, 'samples': 336192, 'steps': 1750, 'loss/train': 1.9394930005073547} 01/28/2022 15:07:14 - INFO - codeparrot_training - Step 1751: {'lr': 0.00043775, 'samples': 336384, 'steps': 1751, 'loss/train': 2.0573941469192505} 01/28/2022 15:07:18 - INFO - codeparrot_training - Step 1752: {'lr': 0.000438, 'samples': 336576, 'steps': 1752, 'loss/train': 1.3165290355682373} 01/28/2022 15:07:23 - INFO - codeparrot_training - Step 1753: {'lr': 0.00043825, 'samples': 336768, 'steps': 1753, 'loss/train': 2.2994107604026794} 01/28/2022 15:07:27 - INFO - codeparrot_training - Step 1754: {'lr': 0.00043850000000000003, 'samples': 336960, 'steps': 1754, 'loss/train': 1.7986255288124084} 01/28/2022 15:07:31 - INFO - codeparrot_training - Step 1755: {'lr': 0.00043874999999999996, 'samples': 337152, 'steps': 1755, 'loss/train': 2.4221463203430176} 01/28/2022 15:07:36 - INFO - codeparrot_training - Step 1756: {'lr': 0.000439, 'samples': 337344, 'steps': 1756, 'loss/train': 2.5629422664642334} 01/28/2022 15:07:40 - INFO - codeparrot_training - Step 1757: {'lr': 0.00043924999999999997, 'samples': 337536, 'steps': 1757, 'loss/train': 2.3310155868530273} 01/28/2022 15:07:44 - INFO - codeparrot_training - Step 1758: {'lr': 0.0004395, 'samples': 337728, 'steps': 1758, 'loss/train': 1.372418224811554} 01/28/2022 15:07:49 - INFO - codeparrot_training - Step 1759: {'lr': 0.00043975, 'samples': 337920, 'steps': 1759, 'loss/train': 2.0976839661598206} 01/28/2022 15:07:53 - INFO - codeparrot_training - Step 1760: {'lr': 0.00044, 'samples': 338112, 'steps': 1760, 'loss/train': 2.113537073135376} 01/28/2022 15:07:59 - INFO - codeparrot_training - Step 1761: {'lr': 0.00044025, 'samples': 338304, 'steps': 1761, 'loss/train': 1.1731322407722473} 01/28/2022 15:08:03 - INFO - codeparrot_training - Step 1762: {'lr': 0.00044050000000000003, 'samples': 338496, 'steps': 1762, 'loss/train': 2.0458574295043945} 01/28/2022 15:08:08 - INFO - codeparrot_training - Step 1763: {'lr': 0.00044075, 'samples': 338688, 'steps': 1763, 'loss/train': 1.7629807591438293} 01/28/2022 15:08:12 - INFO - codeparrot_training - Step 1764: {'lr': 0.000441, 'samples': 338880, 'steps': 1764, 'loss/train': 2.7237507104873657} 01/28/2022 15:08:16 - INFO - codeparrot_training - Step 1765: {'lr': 0.00044124999999999996, 'samples': 339072, 'steps': 1765, 'loss/train': 2.5689370036125183} 01/28/2022 15:08:21 - INFO - codeparrot_training - Step 1766: {'lr': 0.0004415, 'samples': 339264, 'steps': 1766, 'loss/train': 1.7629101276397705} 01/28/2022 15:08:25 - INFO - codeparrot_training - Step 1767: {'lr': 0.00044175, 'samples': 339456, 'steps': 1767, 'loss/train': 2.196295380592346} 01/28/2022 15:08:30 - INFO - codeparrot_training - Step 1768: {'lr': 0.000442, 'samples': 339648, 'steps': 1768, 'loss/train': 2.1143859028816223} 01/28/2022 15:08:34 - INFO - codeparrot_training - Step 1769: {'lr': 0.00044225, 'samples': 339840, 'steps': 1769, 'loss/train': 2.788992404937744} 01/28/2022 15:08:38 - INFO - codeparrot_training - Step 1770: {'lr': 0.0004425, 'samples': 340032, 'steps': 1770, 'loss/train': 2.2667550444602966} 01/28/2022 15:08:43 - INFO - codeparrot_training - Step 1771: {'lr': 0.00044275, 'samples': 340224, 'steps': 1771, 'loss/train': 1.475020855665207} 01/28/2022 15:08:47 - INFO - codeparrot_training - Step 1772: {'lr': 0.00044300000000000003, 'samples': 340416, 'steps': 1772, 'loss/train': 1.2330763041973114} 01/28/2022 15:08:52 - INFO - codeparrot_training - Step 1773: {'lr': 0.00044325, 'samples': 340608, 'steps': 1773, 'loss/train': 2.4245688915252686} 01/28/2022 15:08:56 - INFO - codeparrot_training - Step 1774: {'lr': 0.0004435, 'samples': 340800, 'steps': 1774, 'loss/train': 0.8312238156795502} 01/28/2022 15:09:00 - INFO - codeparrot_training - Step 1775: {'lr': 0.00044374999999999997, 'samples': 340992, 'steps': 1775, 'loss/train': 2.644107162952423} 01/28/2022 15:09:07 - INFO - codeparrot_training - Step 1776: {'lr': 0.000444, 'samples': 341184, 'steps': 1776, 'loss/train': 1.5467260479927063} 01/28/2022 15:09:11 - INFO - codeparrot_training - Step 1777: {'lr': 0.00044425, 'samples': 341376, 'steps': 1777, 'loss/train': 2.834625720977783} 01/28/2022 15:09:15 - INFO - codeparrot_training - Step 1778: {'lr': 0.0004445, 'samples': 341568, 'steps': 1778, 'loss/train': 0.598884254693985} 01/28/2022 15:09:19 - INFO - codeparrot_training - Step 1779: {'lr': 0.00044475, 'samples': 341760, 'steps': 1779, 'loss/train': 2.7426865696907043} 01/28/2022 15:09:23 - INFO - codeparrot_training - Step 1780: {'lr': 0.00044500000000000003, 'samples': 341952, 'steps': 1780, 'loss/train': 3.1601650714874268} 01/28/2022 15:09:28 - INFO - codeparrot_training - Step 1781: {'lr': 0.00044525, 'samples': 342144, 'steps': 1781, 'loss/train': 2.729690730571747} 01/28/2022 15:09:33 - INFO - codeparrot_training - Step 1782: {'lr': 0.00044550000000000004, 'samples': 342336, 'steps': 1782, 'loss/train': 2.0088767409324646} 01/28/2022 15:09:37 - INFO - codeparrot_training - Step 1783: {'lr': 0.00044574999999999997, 'samples': 342528, 'steps': 1783, 'loss/train': 2.195220172405243} 01/28/2022 15:09:41 - INFO - codeparrot_training - Step 1784: {'lr': 0.000446, 'samples': 342720, 'steps': 1784, 'loss/train': 1.7500455379486084} 01/28/2022 15:09:45 - INFO - codeparrot_training - Step 1785: {'lr': 0.00044625, 'samples': 342912, 'steps': 1785, 'loss/train': 2.4295756816864014} 01/28/2022 15:09:52 - INFO - codeparrot_training - Step 1786: {'lr': 0.0004465, 'samples': 343104, 'steps': 1786, 'loss/train': 1.9527581334114075} 01/28/2022 15:09:56 - INFO - codeparrot_training - Step 1787: {'lr': 0.00044675, 'samples': 343296, 'steps': 1787, 'loss/train': 2.1158979535102844} 01/28/2022 15:10:00 - INFO - codeparrot_training - Step 1788: {'lr': 0.000447, 'samples': 343488, 'steps': 1788, 'loss/train': 2.2558854818344116} 01/28/2022 15:10:04 - INFO - codeparrot_training - Step 1789: {'lr': 0.00044725, 'samples': 343680, 'steps': 1789, 'loss/train': 1.7354274988174438} 01/28/2022 15:10:08 - INFO - codeparrot_training - Step 1790: {'lr': 0.00044750000000000004, 'samples': 343872, 'steps': 1790, 'loss/train': 1.2520123422145844} 01/28/2022 15:10:13 - INFO - codeparrot_training - Step 1791: {'lr': 0.00044775, 'samples': 344064, 'steps': 1791, 'loss/train': 1.9029777646064758} 01/28/2022 15:10:18 - INFO - codeparrot_training - Step 1792: {'lr': 0.000448, 'samples': 344256, 'steps': 1792, 'loss/train': 2.2610605359077454} 01/28/2022 15:10:22 - INFO - codeparrot_training - Step 1793: {'lr': 0.00044824999999999997, 'samples': 344448, 'steps': 1793, 'loss/train': 2.4697925448417664} 01/28/2022 15:10:26 - INFO - codeparrot_training - Step 1794: {'lr': 0.0004485, 'samples': 344640, 'steps': 1794, 'loss/train': 2.3558549880981445} 01/28/2022 15:10:30 - INFO - codeparrot_training - Step 1795: {'lr': 0.00044875, 'samples': 344832, 'steps': 1795, 'loss/train': 2.600886583328247} 01/28/2022 15:10:35 - INFO - codeparrot_training - Step 1796: {'lr': 0.000449, 'samples': 345024, 'steps': 1796, 'loss/train': 1.7025368213653564} 01/28/2022 15:10:39 - INFO - codeparrot_training - Step 1797: {'lr': 0.00044925, 'samples': 345216, 'steps': 1797, 'loss/train': 2.6850298047065735} 01/28/2022 15:10:44 - INFO - codeparrot_training - Step 1798: {'lr': 0.00044950000000000003, 'samples': 345408, 'steps': 1798, 'loss/train': 2.2505434155464172} 01/28/2022 15:10:48 - INFO - codeparrot_training - Step 1799: {'lr': 0.00044975, 'samples': 345600, 'steps': 1799, 'loss/train': 2.317931056022644} 01/28/2022 15:10:52 - INFO - codeparrot_training - Step 1800: {'lr': 0.00045000000000000004, 'samples': 345792, 'steps': 1800, 'loss/train': 2.4370325803756714} 01/28/2022 15:10:58 - INFO - codeparrot_training - Step 1801: {'lr': 0.00045024999999999997, 'samples': 345984, 'steps': 1801, 'loss/train': 1.9124904870986938} 01/28/2022 15:11:02 - INFO - codeparrot_training - Step 1802: {'lr': 0.0004505, 'samples': 346176, 'steps': 1802, 'loss/train': 1.5768880248069763} 01/28/2022 15:11:06 - INFO - codeparrot_training - Step 1803: {'lr': 0.00045075, 'samples': 346368, 'steps': 1803, 'loss/train': 1.0997436046600342} 01/28/2022 15:11:11 - INFO - codeparrot_training - Step 1804: {'lr': 0.000451, 'samples': 346560, 'steps': 1804, 'loss/train': 2.1200843453407288} 01/28/2022 15:11:15 - INFO - codeparrot_training - Step 1805: {'lr': 0.00045125, 'samples': 346752, 'steps': 1805, 'loss/train': 0.8790491223335266} 01/28/2022 15:11:20 - INFO - codeparrot_training - Step 1806: {'lr': 0.0004515, 'samples': 346944, 'steps': 1806, 'loss/train': 2.0925493240356445} 01/28/2022 15:11:24 - INFO - codeparrot_training - Step 1807: {'lr': 0.00045175, 'samples': 347136, 'steps': 1807, 'loss/train': 3.6132770776748657} 01/28/2022 15:11:28 - INFO - codeparrot_training - Step 1808: {'lr': 0.00045200000000000004, 'samples': 347328, 'steps': 1808, 'loss/train': 2.4820961952209473} 01/28/2022 15:11:32 - INFO - codeparrot_training - Step 1809: {'lr': 0.00045225, 'samples': 347520, 'steps': 1809, 'loss/train': 3.2262940406799316} 01/28/2022 15:11:37 - INFO - codeparrot_training - Step 1810: {'lr': 0.00045250000000000005, 'samples': 347712, 'steps': 1810, 'loss/train': 2.2057530283927917} 01/28/2022 15:11:42 - INFO - codeparrot_training - Step 1811: {'lr': 0.00045275, 'samples': 347904, 'steps': 1811, 'loss/train': 2.502051293849945} 01/28/2022 15:11:46 - INFO - codeparrot_training - Step 1812: {'lr': 0.000453, 'samples': 348096, 'steps': 1812, 'loss/train': 2.2142679691314697} 01/28/2022 15:11:50 - INFO - codeparrot_training - Step 1813: {'lr': 0.00045325, 'samples': 348288, 'steps': 1813, 'loss/train': 3.9852951765060425} 01/28/2022 15:11:55 - INFO - codeparrot_training - Step 1814: {'lr': 0.0004535, 'samples': 348480, 'steps': 1814, 'loss/train': 1.9643903374671936} 01/28/2022 15:11:59 - INFO - codeparrot_training - Step 1815: {'lr': 0.00045375, 'samples': 348672, 'steps': 1815, 'loss/train': 2.3582332134246826} 01/28/2022 15:12:04 - INFO - codeparrot_training - Step 1816: {'lr': 0.00045400000000000003, 'samples': 348864, 'steps': 1816, 'loss/train': 1.931522011756897} 01/28/2022 15:12:08 - INFO - codeparrot_training - Step 1817: {'lr': 0.00045425, 'samples': 349056, 'steps': 1817, 'loss/train': 1.5595782995224} 01/28/2022 15:12:12 - INFO - codeparrot_training - Step 1818: {'lr': 0.00045450000000000004, 'samples': 349248, 'steps': 1818, 'loss/train': 1.6498337388038635} 01/28/2022 15:12:16 - INFO - codeparrot_training - Step 1819: {'lr': 0.00045475, 'samples': 349440, 'steps': 1819, 'loss/train': 2.3283090591430664} 01/28/2022 15:12:21 - INFO - codeparrot_training - Step 1820: {'lr': 0.000455, 'samples': 349632, 'steps': 1820, 'loss/train': 3.3872458934783936} 01/28/2022 15:12:27 - INFO - codeparrot_training - Step 1821: {'lr': 0.00045525, 'samples': 349824, 'steps': 1821, 'loss/train': 2.5988234281539917} 01/28/2022 15:12:31 - INFO - codeparrot_training - Step 1822: {'lr': 0.0004555, 'samples': 350016, 'steps': 1822, 'loss/train': 1.9621492624282837} 01/28/2022 15:12:35 - INFO - codeparrot_training - Step 1823: {'lr': 0.00045575, 'samples': 350208, 'steps': 1823, 'loss/train': 2.3275317549705505} 01/28/2022 15:12:40 - INFO - codeparrot_training - Step 1824: {'lr': 0.000456, 'samples': 350400, 'steps': 1824, 'loss/train': 2.390848159790039} 01/28/2022 15:12:44 - INFO - codeparrot_training - Step 1825: {'lr': 0.00045625, 'samples': 350592, 'steps': 1825, 'loss/train': 2.365892171859741} 01/28/2022 15:12:49 - INFO - codeparrot_training - Step 1826: {'lr': 0.00045650000000000004, 'samples': 350784, 'steps': 1826, 'loss/train': 1.8969102501869202} 01/28/2022 15:12:53 - INFO - codeparrot_training - Step 1827: {'lr': 0.00045675, 'samples': 350976, 'steps': 1827, 'loss/train': 1.8191596269607544} 01/28/2022 15:12:57 - INFO - codeparrot_training - Step 1828: {'lr': 0.00045700000000000005, 'samples': 351168, 'steps': 1828, 'loss/train': 1.6756017208099365} 01/28/2022 15:13:02 - INFO - codeparrot_training - Step 1829: {'lr': 0.00045725, 'samples': 351360, 'steps': 1829, 'loss/train': 2.0185545086860657} 01/28/2022 15:13:06 - INFO - codeparrot_training - Step 1830: {'lr': 0.0004575, 'samples': 351552, 'steps': 1830, 'loss/train': 2.2736778259277344} 01/28/2022 15:13:13 - INFO - codeparrot_training - Step 1831: {'lr': 0.00045775, 'samples': 351744, 'steps': 1831, 'loss/train': 2.231265127658844} 01/28/2022 15:13:17 - INFO - codeparrot_training - Step 1832: {'lr': 0.000458, 'samples': 351936, 'steps': 1832, 'loss/train': 1.5673556327819824} 01/28/2022 15:13:21 - INFO - codeparrot_training - Step 1833: {'lr': 0.00045825, 'samples': 352128, 'steps': 1833, 'loss/train': 2.290362775325775} 01/28/2022 15:13:25 - INFO - codeparrot_training - Step 1834: {'lr': 0.00045850000000000003, 'samples': 352320, 'steps': 1834, 'loss/train': 2.0408921241760254} 01/28/2022 15:13:29 - INFO - codeparrot_training - Step 1835: {'lr': 0.00045875, 'samples': 352512, 'steps': 1835, 'loss/train': 2.662816286087036} 01/28/2022 15:13:35 - INFO - codeparrot_training - Step 1836: {'lr': 0.00045900000000000004, 'samples': 352704, 'steps': 1836, 'loss/train': 1.523913860321045} 01/28/2022 15:13:39 - INFO - codeparrot_training - Step 1837: {'lr': 0.00045925, 'samples': 352896, 'steps': 1837, 'loss/train': 2.864295780658722} 01/28/2022 15:13:43 - INFO - codeparrot_training - Step 1838: {'lr': 0.00045950000000000006, 'samples': 353088, 'steps': 1838, 'loss/train': 1.9534723162651062} 01/28/2022 15:13:47 - INFO - codeparrot_training - Step 1839: {'lr': 0.00045975, 'samples': 353280, 'steps': 1839, 'loss/train': 3.1775182485580444} 01/28/2022 15:13:51 - INFO - codeparrot_training - Step 1840: {'lr': 0.00046, 'samples': 353472, 'steps': 1840, 'loss/train': 1.6611671447753906} 01/28/2022 15:13:56 - INFO - codeparrot_training - Step 1841: {'lr': 0.00046025, 'samples': 353664, 'steps': 1841, 'loss/train': 1.9205996990203857} 01/28/2022 15:14:01 - INFO - codeparrot_training - Step 1842: {'lr': 0.0004605, 'samples': 353856, 'steps': 1842, 'loss/train': 2.014385461807251} 01/28/2022 15:14:05 - INFO - codeparrot_training - Step 1843: {'lr': 0.00046075, 'samples': 354048, 'steps': 1843, 'loss/train': 1.74648517370224} 01/28/2022 15:14:09 - INFO - codeparrot_training - Step 1844: {'lr': 0.00046100000000000004, 'samples': 354240, 'steps': 1844, 'loss/train': 1.6978061199188232} 01/28/2022 15:14:13 - INFO - codeparrot_training - Step 1845: {'lr': 0.00046125, 'samples': 354432, 'steps': 1845, 'loss/train': 0.7862594425678253} 01/28/2022 15:14:19 - INFO - codeparrot_training - Step 1846: {'lr': 0.00046150000000000005, 'samples': 354624, 'steps': 1846, 'loss/train': 2.3629429936408997} 01/28/2022 15:14:23 - INFO - codeparrot_training - Step 1847: {'lr': 0.00046175000000000003, 'samples': 354816, 'steps': 1847, 'loss/train': 2.4384382367134094} 01/28/2022 15:14:28 - INFO - codeparrot_training - Step 1848: {'lr': 0.000462, 'samples': 355008, 'steps': 1848, 'loss/train': 2.0216946601867676} 01/28/2022 15:14:32 - INFO - codeparrot_training - Step 1849: {'lr': 0.00046225, 'samples': 355200, 'steps': 1849, 'loss/train': 2.008707582950592} 01/28/2022 15:14:36 - INFO - codeparrot_training - Step 1850: {'lr': 0.0004625, 'samples': 355392, 'steps': 1850, 'loss/train': 2.3486950993537903} 01/28/2022 15:14:41 - INFO - codeparrot_training - Step 1851: {'lr': 0.00046275, 'samples': 355584, 'steps': 1851, 'loss/train': 2.332076668739319} 01/28/2022 15:14:45 - INFO - codeparrot_training - Step 1852: {'lr': 0.00046300000000000003, 'samples': 355776, 'steps': 1852, 'loss/train': 0.9330092668533325} 01/28/2022 15:14:50 - INFO - codeparrot_training - Step 1853: {'lr': 0.00046325, 'samples': 355968, 'steps': 1853, 'loss/train': 2.0825924277305603} 01/28/2022 15:14:54 - INFO - codeparrot_training - Step 1854: {'lr': 0.00046350000000000004, 'samples': 356160, 'steps': 1854, 'loss/train': 2.330622375011444} 01/28/2022 15:14:58 - INFO - codeparrot_training - Step 1855: {'lr': 0.00046375, 'samples': 356352, 'steps': 1855, 'loss/train': 2.385727643966675} 01/28/2022 15:15:03 - INFO - codeparrot_training - Step 1856: {'lr': 0.00046400000000000006, 'samples': 356544, 'steps': 1856, 'loss/train': 0.6832497864961624} 01/28/2022 15:15:07 - INFO - codeparrot_training - Step 1857: {'lr': 0.00046425, 'samples': 356736, 'steps': 1857, 'loss/train': 3.539833903312683} 01/28/2022 15:15:11 - INFO - codeparrot_training - Step 1858: {'lr': 0.0004645, 'samples': 356928, 'steps': 1858, 'loss/train': 2.2786254286766052} 01/28/2022 15:15:16 - INFO - codeparrot_training - Step 1859: {'lr': 0.00046475, 'samples': 357120, 'steps': 1859, 'loss/train': 1.5281049013137817} 01/28/2022 15:15:20 - INFO - codeparrot_training - Step 1860: {'lr': 0.000465, 'samples': 357312, 'steps': 1860, 'loss/train': 2.505415141582489} 01/28/2022 15:15:25 - INFO - codeparrot_training - Step 1861: {'lr': 0.00046525, 'samples': 357504, 'steps': 1861, 'loss/train': 2.341139495372772} 01/28/2022 15:15:29 - INFO - codeparrot_training - Step 1862: {'lr': 0.00046550000000000004, 'samples': 357696, 'steps': 1862, 'loss/train': 2.225492298603058} 01/28/2022 15:15:34 - INFO - codeparrot_training - Step 1863: {'lr': 0.00046575, 'samples': 357888, 'steps': 1863, 'loss/train': 2.506253957748413} 01/28/2022 15:15:38 - INFO - codeparrot_training - Step 1864: {'lr': 0.00046600000000000005, 'samples': 358080, 'steps': 1864, 'loss/train': 2.166025221347809} 01/28/2022 15:15:42 - INFO - codeparrot_training - Step 1865: {'lr': 0.00046625000000000003, 'samples': 358272, 'steps': 1865, 'loss/train': 1.7458140850067139} 01/28/2022 15:15:48 - INFO - codeparrot_training - Step 1866: {'lr': 0.0004665, 'samples': 358464, 'steps': 1866, 'loss/train': 1.5841763019561768} 01/28/2022 15:15:53 - INFO - codeparrot_training - Step 1867: {'lr': 0.00046675, 'samples': 358656, 'steps': 1867, 'loss/train': 2.2836674451828003} 01/28/2022 15:15:57 - INFO - codeparrot_training - Step 1868: {'lr': 0.000467, 'samples': 358848, 'steps': 1868, 'loss/train': 1.9806527495384216} 01/28/2022 15:16:01 - INFO - codeparrot_training - Step 1869: {'lr': 0.00046725, 'samples': 359040, 'steps': 1869, 'loss/train': 1.9374698996543884} 01/28/2022 15:16:05 - INFO - codeparrot_training - Step 1870: {'lr': 0.00046750000000000003, 'samples': 359232, 'steps': 1870, 'loss/train': 0.8083462715148926} 01/28/2022 15:16:11 - INFO - codeparrot_training - Step 1871: {'lr': 0.00046775, 'samples': 359424, 'steps': 1871, 'loss/train': 2.2851065397262573} 01/28/2022 15:16:15 - INFO - codeparrot_training - Step 1872: {'lr': 0.00046800000000000005, 'samples': 359616, 'steps': 1872, 'loss/train': 1.7626469135284424} 01/28/2022 15:16:19 - INFO - codeparrot_training - Step 1873: {'lr': 0.00046825, 'samples': 359808, 'steps': 1873, 'loss/train': 2.020338535308838} 01/28/2022 15:16:23 - INFO - codeparrot_training - Step 1874: {'lr': 0.00046850000000000006, 'samples': 360000, 'steps': 1874, 'loss/train': 1.719109833240509} 01/28/2022 15:16:27 - INFO - codeparrot_training - Step 1875: {'lr': 0.00046875, 'samples': 360192, 'steps': 1875, 'loss/train': 2.1993414759635925} 01/28/2022 15:16:33 - INFO - codeparrot_training - Step 1876: {'lr': 0.00046899999999999996, 'samples': 360384, 'steps': 1876, 'loss/train': 2.6035741567611694} 01/28/2022 15:16:38 - INFO - codeparrot_training - Step 1877: {'lr': 0.00046925, 'samples': 360576, 'steps': 1877, 'loss/train': 2.008894979953766} 01/28/2022 15:16:42 - INFO - codeparrot_training - Step 1878: {'lr': 0.0004695, 'samples': 360768, 'steps': 1878, 'loss/train': 1.682439923286438} 01/28/2022 15:16:46 - INFO - codeparrot_training - Step 1879: {'lr': 0.00046975, 'samples': 360960, 'steps': 1879, 'loss/train': 1.5710864067077637} 01/28/2022 15:16:50 - INFO - codeparrot_training - Step 1880: {'lr': 0.00047, 'samples': 361152, 'steps': 1880, 'loss/train': 2.115239918231964} 01/28/2022 15:16:55 - INFO - codeparrot_training - Step 1881: {'lr': 0.00047025, 'samples': 361344, 'steps': 1881, 'loss/train': 1.6808903217315674} 01/28/2022 15:17:00 - INFO - codeparrot_training - Step 1882: {'lr': 0.0004705, 'samples': 361536, 'steps': 1882, 'loss/train': 2.5031254291534424} 01/28/2022 15:17:04 - INFO - codeparrot_training - Step 1883: {'lr': 0.00047075000000000003, 'samples': 361728, 'steps': 1883, 'loss/train': 2.463531196117401} 01/28/2022 15:17:08 - INFO - codeparrot_training - Step 1884: {'lr': 0.000471, 'samples': 361920, 'steps': 1884, 'loss/train': 1.57847660779953} 01/28/2022 15:17:12 - INFO - codeparrot_training - Step 1885: {'lr': 0.00047125, 'samples': 362112, 'steps': 1885, 'loss/train': 2.0253029465675354} 01/28/2022 15:17:17 - INFO - codeparrot_training - Step 1886: {'lr': 0.00047149999999999997, 'samples': 362304, 'steps': 1886, 'loss/train': 2.0553240180015564} 01/28/2022 15:17:21 - INFO - codeparrot_training - Step 1887: {'lr': 0.00047175, 'samples': 362496, 'steps': 1887, 'loss/train': 1.484438717365265} 01/28/2022 15:17:26 - INFO - codeparrot_training - Step 1888: {'lr': 0.000472, 'samples': 362688, 'steps': 1888, 'loss/train': 2.604108452796936} 01/28/2022 15:17:30 - INFO - codeparrot_training - Step 1889: {'lr': 0.00047225, 'samples': 362880, 'steps': 1889, 'loss/train': 2.267809510231018} 01/28/2022 15:17:34 - INFO - codeparrot_training - Step 1890: {'lr': 0.0004725, 'samples': 363072, 'steps': 1890, 'loss/train': 1.7722991108894348} 01/28/2022 15:17:40 - INFO - codeparrot_training - Step 1891: {'lr': 0.00047275, 'samples': 363264, 'steps': 1891, 'loss/train': 2.2410938143730164} 01/28/2022 15:17:44 - INFO - codeparrot_training - Step 1892: {'lr': 0.000473, 'samples': 363456, 'steps': 1892, 'loss/train': 1.861700713634491} 01/28/2022 15:17:48 - INFO - codeparrot_training - Step 1893: {'lr': 0.00047325000000000004, 'samples': 363648, 'steps': 1893, 'loss/train': 2.5292657017707825} 01/28/2022 15:17:53 - INFO - codeparrot_training - Step 1894: {'lr': 0.00047349999999999996, 'samples': 363840, 'steps': 1894, 'loss/train': 2.2569231390953064} 01/28/2022 15:17:57 - INFO - codeparrot_training - Step 1895: {'lr': 0.00047375, 'samples': 364032, 'steps': 1895, 'loss/train': 1.8880301713943481} 01/28/2022 15:18:02 - INFO - codeparrot_training - Step 1896: {'lr': 0.000474, 'samples': 364224, 'steps': 1896, 'loss/train': 2.038884401321411} 01/28/2022 15:18:06 - INFO - codeparrot_training - Step 1897: {'lr': 0.00047425, 'samples': 364416, 'steps': 1897, 'loss/train': 2.0557111501693726} 01/28/2022 15:18:10 - INFO - codeparrot_training - Step 1898: {'lr': 0.0004745, 'samples': 364608, 'steps': 1898, 'loss/train': 1.9232286214828491} 01/28/2022 15:18:14 - INFO - codeparrot_training - Step 1899: {'lr': 0.00047475, 'samples': 364800, 'steps': 1899, 'loss/train': 1.949855625629425} 01/28/2022 15:18:19 - INFO - codeparrot_training - Step 1900: {'lr': 0.000475, 'samples': 364992, 'steps': 1900, 'loss/train': 2.1276349425315857} 01/28/2022 15:18:24 - INFO - codeparrot_training - Step 1901: {'lr': 0.00047525000000000003, 'samples': 365184, 'steps': 1901, 'loss/train': 2.4334242939949036} 01/28/2022 15:18:28 - INFO - codeparrot_training - Step 1902: {'lr': 0.0004755, 'samples': 365376, 'steps': 1902, 'loss/train': 1.8637136220932007} 01/28/2022 15:18:32 - INFO - codeparrot_training - Step 1903: {'lr': 0.00047575, 'samples': 365568, 'steps': 1903, 'loss/train': 2.346198320388794} 01/28/2022 15:18:36 - INFO - codeparrot_training - Step 1904: {'lr': 0.00047599999999999997, 'samples': 365760, 'steps': 1904, 'loss/train': 2.134476900100708} 01/28/2022 15:18:43 - INFO - codeparrot_training - Step 1905: {'lr': 0.00047625, 'samples': 365952, 'steps': 1905, 'loss/train': 2.1911319494247437} 01/28/2022 15:18:47 - INFO - codeparrot_training - Step 1906: {'lr': 0.0004765, 'samples': 366144, 'steps': 1906, 'loss/train': 1.9768962264060974} 01/28/2022 15:18:51 - INFO - codeparrot_training - Step 1907: {'lr': 0.00047675, 'samples': 366336, 'steps': 1907, 'loss/train': 2.3271737694740295} 01/28/2022 15:18:55 - INFO - codeparrot_training - Step 1908: {'lr': 0.000477, 'samples': 366528, 'steps': 1908, 'loss/train': 0.9811719954013824} 01/28/2022 15:18:59 - INFO - codeparrot_training - Step 1909: {'lr': 0.00047725, 'samples': 366720, 'steps': 1909, 'loss/train': 1.7484862804412842} 01/28/2022 15:19:05 - INFO - codeparrot_training - Step 1910: {'lr': 0.0004775, 'samples': 366912, 'steps': 1910, 'loss/train': 2.330228626728058} 01/28/2022 15:19:09 - INFO - codeparrot_training - Step 1911: {'lr': 0.00047775000000000004, 'samples': 367104, 'steps': 1911, 'loss/train': 2.2461238503456116} 01/28/2022 15:19:13 - INFO - codeparrot_training - Step 1912: {'lr': 0.00047799999999999996, 'samples': 367296, 'steps': 1912, 'loss/train': 1.8187590837478638} 01/28/2022 15:19:17 - INFO - codeparrot_training - Step 1913: {'lr': 0.00047825, 'samples': 367488, 'steps': 1913, 'loss/train': 1.9331352710723877} 01/28/2022 15:19:21 - INFO - codeparrot_training - Step 1914: {'lr': 0.0004785, 'samples': 367680, 'steps': 1914, 'loss/train': 1.4992038309574127} 01/28/2022 15:19:25 - INFO - codeparrot_training - Step 1915: {'lr': 0.00047875, 'samples': 367872, 'steps': 1915, 'loss/train': 2.3997285962104797} 01/28/2022 15:19:32 - INFO - codeparrot_training - Step 1916: {'lr': 0.000479, 'samples': 368064, 'steps': 1916, 'loss/train': 2.373488187789917} 01/28/2022 15:19:36 - INFO - codeparrot_training - Step 1917: {'lr': 0.00047925, 'samples': 368256, 'steps': 1917, 'loss/train': 2.1079630851745605} 01/28/2022 15:19:40 - INFO - codeparrot_training - Step 1918: {'lr': 0.0004795, 'samples': 368448, 'steps': 1918, 'loss/train': 1.604847013950348} 01/28/2022 15:19:45 - INFO - codeparrot_training - Step 1919: {'lr': 0.00047975000000000003, 'samples': 368640, 'steps': 1919, 'loss/train': 1.4136720299720764} 01/28/2022 15:19:49 - INFO - codeparrot_training - Step 1920: {'lr': 0.00048, 'samples': 368832, 'steps': 1920, 'loss/train': 2.4885597825050354} 01/28/2022 15:19:53 - INFO - codeparrot_training - Step 1921: {'lr': 0.00048025000000000005, 'samples': 369024, 'steps': 1921, 'loss/train': 2.0177990198135376} 01/28/2022 15:19:59 - INFO - codeparrot_training - Step 1922: {'lr': 0.00048049999999999997, 'samples': 369216, 'steps': 1922, 'loss/train': 2.31428986787796} 01/28/2022 15:20:04 - INFO - codeparrot_training - Step 1923: {'lr': 0.00048075, 'samples': 369408, 'steps': 1923, 'loss/train': 1.8913630843162537} 01/28/2022 15:20:08 - INFO - codeparrot_training - Step 1924: {'lr': 0.000481, 'samples': 369600, 'steps': 1924, 'loss/train': 2.240315079689026} 01/28/2022 15:20:12 - INFO - codeparrot_training - Step 1925: {'lr': 0.00048125, 'samples': 369792, 'steps': 1925, 'loss/train': 1.903373658657074} 01/28/2022 15:20:17 - INFO - codeparrot_training - Step 1926: {'lr': 0.0004815, 'samples': 369984, 'steps': 1926, 'loss/train': 2.3228121399879456} 01/28/2022 15:20:21 - INFO - codeparrot_training - Step 1927: {'lr': 0.00048175000000000003, 'samples': 370176, 'steps': 1927, 'loss/train': 1.8365673422813416} 01/28/2022 15:20:25 - INFO - codeparrot_training - Step 1928: {'lr': 0.000482, 'samples': 370368, 'steps': 1928, 'loss/train': 2.251401722431183} 01/28/2022 15:20:30 - INFO - codeparrot_training - Step 1929: {'lr': 0.00048225000000000004, 'samples': 370560, 'steps': 1929, 'loss/train': 2.2396209239959717} 01/28/2022 15:20:34 - INFO - codeparrot_training - Step 1930: {'lr': 0.0004825, 'samples': 370752, 'steps': 1930, 'loss/train': 0.8318962454795837} 01/28/2022 15:20:39 - INFO - codeparrot_training - Step 1931: {'lr': 0.00048275, 'samples': 370944, 'steps': 1931, 'loss/train': 2.6039364337921143} 01/28/2022 15:20:43 - INFO - codeparrot_training - Step 1932: {'lr': 0.000483, 'samples': 371136, 'steps': 1932, 'loss/train': 2.2358558177948} 01/28/2022 15:20:47 - INFO - codeparrot_training - Step 1933: {'lr': 0.00048325, 'samples': 371328, 'steps': 1933, 'loss/train': 2.9010679721832275} 01/28/2022 15:20:52 - INFO - codeparrot_training - Step 1934: {'lr': 0.0004835, 'samples': 371520, 'steps': 1934, 'loss/train': 1.4107789993286133} 01/28/2022 15:20:56 - INFO - codeparrot_training - Step 1935: {'lr': 0.00048375, 'samples': 371712, 'steps': 1935, 'loss/train': 1.229703962802887} 01/28/2022 15:21:02 - INFO - codeparrot_training - Step 1936: {'lr': 0.000484, 'samples': 371904, 'steps': 1936, 'loss/train': 1.638334572315216} 01/28/2022 15:21:06 - INFO - codeparrot_training - Step 1937: {'lr': 0.00048425000000000003, 'samples': 372096, 'steps': 1937, 'loss/train': 1.6023758053779602} 01/28/2022 15:21:10 - INFO - codeparrot_training - Step 1938: {'lr': 0.0004845, 'samples': 372288, 'steps': 1938, 'loss/train': 2.2564027905464172} 01/28/2022 15:21:14 - INFO - codeparrot_training - Step 1939: {'lr': 0.00048475000000000005, 'samples': 372480, 'steps': 1939, 'loss/train': 2.090108871459961} 01/28/2022 15:21:19 - INFO - codeparrot_training - Step 1940: {'lr': 0.00048499999999999997, 'samples': 372672, 'steps': 1940, 'loss/train': 2.034739315509796} 01/28/2022 15:21:24 - INFO - codeparrot_training - Step 1941: {'lr': 0.00048525, 'samples': 372864, 'steps': 1941, 'loss/train': 2.192450702190399} 01/28/2022 15:21:28 - INFO - codeparrot_training - Step 1942: {'lr': 0.0004855, 'samples': 373056, 'steps': 1942, 'loss/train': 2.311851739883423} 01/28/2022 15:21:32 - INFO - codeparrot_training - Step 1943: {'lr': 0.00048575, 'samples': 373248, 'steps': 1943, 'loss/train': 2.4836182594299316} 01/28/2022 15:21:36 - INFO - codeparrot_training - Step 1944: {'lr': 0.000486, 'samples': 373440, 'steps': 1944, 'loss/train': 0.7596074044704437} 01/28/2022 15:21:40 - INFO - codeparrot_training - Step 1945: {'lr': 0.00048625000000000003, 'samples': 373632, 'steps': 1945, 'loss/train': 2.2612091302871704} 01/28/2022 15:21:46 - INFO - codeparrot_training - Step 1946: {'lr': 0.0004865, 'samples': 373824, 'steps': 1946, 'loss/train': 2.339252293109894} 01/28/2022 15:21:50 - INFO - codeparrot_training - Step 1947: {'lr': 0.00048675000000000004, 'samples': 374016, 'steps': 1947, 'loss/train': 1.8570465445518494} 01/28/2022 15:21:54 - INFO - codeparrot_training - Step 1948: {'lr': 0.000487, 'samples': 374208, 'steps': 1948, 'loss/train': 2.2299283146858215} 01/28/2022 15:21:58 - INFO - codeparrot_training - Step 1949: {'lr': 0.00048725000000000005, 'samples': 374400, 'steps': 1949, 'loss/train': 1.5235753655433655} 01/28/2022 15:22:02 - INFO - codeparrot_training - Step 1950: {'lr': 0.0004875, 'samples': 374592, 'steps': 1950, 'loss/train': 2.2883012294769287} 01/28/2022 15:22:09 - INFO - codeparrot_training - Step 1951: {'lr': 0.00048775, 'samples': 374784, 'steps': 1951, 'loss/train': 1.9050377011299133} 01/28/2022 15:22:13 - INFO - codeparrot_training - Step 1952: {'lr': 0.000488, 'samples': 374976, 'steps': 1952, 'loss/train': 2.455422878265381} 01/28/2022 15:22:17 - INFO - codeparrot_training - Step 1953: {'lr': 0.00048825, 'samples': 375168, 'steps': 1953, 'loss/train': 0.9972498714923859} 01/28/2022 15:22:21 - INFO - codeparrot_training - Step 1954: {'lr': 0.0004885, 'samples': 375360, 'steps': 1954, 'loss/train': 1.975938320159912} 01/28/2022 15:22:25 - INFO - codeparrot_training - Step 1955: {'lr': 0.00048875, 'samples': 375552, 'steps': 1955, 'loss/train': 1.527767837047577} 01/28/2022 15:22:31 - INFO - codeparrot_training - Step 1956: {'lr': 0.000489, 'samples': 375744, 'steps': 1956, 'loss/train': 2.4273703694343567} 01/28/2022 15:22:35 - INFO - codeparrot_training - Step 1957: {'lr': 0.00048925, 'samples': 375936, 'steps': 1957, 'loss/train': 1.6184406280517578} 01/28/2022 15:22:39 - INFO - codeparrot_training - Step 1958: {'lr': 0.0004895, 'samples': 376128, 'steps': 1958, 'loss/train': 2.6630428433418274} 01/28/2022 15:22:43 - INFO - codeparrot_training - Step 1959: {'lr': 0.0004897500000000001, 'samples': 376320, 'steps': 1959, 'loss/train': 2.161782681941986} 01/28/2022 15:22:47 - INFO - codeparrot_training - Step 1960: {'lr': 0.00049, 'samples': 376512, 'steps': 1960, 'loss/train': 2.1363739371299744} 01/28/2022 15:22:54 - INFO - codeparrot_training - Step 1961: {'lr': 0.00049025, 'samples': 376704, 'steps': 1961, 'loss/train': 3.115736961364746} 01/28/2022 15:22:58 - INFO - codeparrot_training - Step 1962: {'lr': 0.0004905, 'samples': 376896, 'steps': 1962, 'loss/train': 1.817329466342926} 01/28/2022 15:23:02 - INFO - codeparrot_training - Step 1963: {'lr': 0.0004907500000000001, 'samples': 377088, 'steps': 1963, 'loss/train': 1.8279869556427002} 01/28/2022 15:23:07 - INFO - codeparrot_training - Step 1964: {'lr': 0.000491, 'samples': 377280, 'steps': 1964, 'loss/train': 2.1692168712615967} 01/28/2022 15:23:11 - INFO - codeparrot_training - Step 1965: {'lr': 0.00049125, 'samples': 377472, 'steps': 1965, 'loss/train': 1.706683874130249} 01/28/2022 15:23:16 - INFO - codeparrot_training - Step 1966: {'lr': 0.0004915, 'samples': 377664, 'steps': 1966, 'loss/train': 2.158460319042206} 01/28/2022 15:23:20 - INFO - codeparrot_training - Step 1967: {'lr': 0.00049175, 'samples': 377856, 'steps': 1967, 'loss/train': 1.926735520362854} 01/28/2022 15:23:24 - INFO - codeparrot_training - Step 1968: {'lr': 0.000492, 'samples': 378048, 'steps': 1968, 'loss/train': 1.9206778407096863} 01/28/2022 15:23:28 - INFO - codeparrot_training - Step 1969: {'lr': 0.0004922500000000001, 'samples': 378240, 'steps': 1969, 'loss/train': 2.5992815494537354} 01/28/2022 15:23:33 - INFO - codeparrot_training - Step 1970: {'lr': 0.0004925, 'samples': 378432, 'steps': 1970, 'loss/train': 2.3113752007484436} 01/28/2022 15:23:39 - INFO - codeparrot_training - Step 1971: {'lr': 0.00049275, 'samples': 378624, 'steps': 1971, 'loss/train': 2.6710129380226135} 01/28/2022 15:23:43 - INFO - codeparrot_training - Step 1972: {'lr': 0.0004930000000000001, 'samples': 378816, 'steps': 1972, 'loss/train': 1.7548885345458984} 01/28/2022 15:23:47 - INFO - codeparrot_training - Step 1973: {'lr': 0.00049325, 'samples': 379008, 'steps': 1973, 'loss/train': 1.2484949827194214} 01/28/2022 15:23:51 - INFO - codeparrot_training - Step 1974: {'lr': 0.0004935, 'samples': 379200, 'steps': 1974, 'loss/train': 1.5888301134109497} 01/28/2022 15:23:55 - INFO - codeparrot_training - Step 1975: {'lr': 0.00049375, 'samples': 379392, 'steps': 1975, 'loss/train': 1.3845356404781342} 01/28/2022 15:24:00 - INFO - codeparrot_training - Step 1976: {'lr': 0.000494, 'samples': 379584, 'steps': 1976, 'loss/train': 2.257458508014679} 01/28/2022 15:24:05 - INFO - codeparrot_training - Step 1977: {'lr': 0.00049425, 'samples': 379776, 'steps': 1977, 'loss/train': 2.593573272228241} 01/28/2022 15:24:09 - INFO - codeparrot_training - Step 1978: {'lr': 0.0004945, 'samples': 379968, 'steps': 1978, 'loss/train': 1.6061355471611023} 01/28/2022 15:24:13 - INFO - codeparrot_training - Step 1979: {'lr': 0.0004947500000000001, 'samples': 380160, 'steps': 1979, 'loss/train': 2.2961500883102417} 01/28/2022 15:24:17 - INFO - codeparrot_training - Step 1980: {'lr': 0.000495, 'samples': 380352, 'steps': 1980, 'loss/train': 1.7744130492210388} 01/28/2022 15:24:23 - INFO - codeparrot_training - Step 1981: {'lr': 0.00049525, 'samples': 380544, 'steps': 1981, 'loss/train': 2.3064011335372925} 01/28/2022 15:24:28 - INFO - codeparrot_training - Step 1982: {'lr': 0.0004955, 'samples': 380736, 'steps': 1982, 'loss/train': 2.4494283199310303} 01/28/2022 15:24:32 - INFO - codeparrot_training - Step 1983: {'lr': 0.00049575, 'samples': 380928, 'steps': 1983, 'loss/train': 2.720816910266876} 01/28/2022 15:24:36 - INFO - codeparrot_training - Step 1984: {'lr': 0.000496, 'samples': 381120, 'steps': 1984, 'loss/train': 2.2167974710464478} 01/28/2022 15:24:40 - INFO - codeparrot_training - Step 1985: {'lr': 0.0004962500000000001, 'samples': 381312, 'steps': 1985, 'loss/train': 1.5309403538703918} 01/28/2022 15:24:46 - INFO - codeparrot_training - Step 1986: {'lr': 0.0004965, 'samples': 381504, 'steps': 1986, 'loss/train': 2.0925920605659485} 01/28/2022 15:24:50 - INFO - codeparrot_training - Step 1987: {'lr': 0.00049675, 'samples': 381696, 'steps': 1987, 'loss/train': 2.300881505012512} 01/28/2022 15:24:54 - INFO - codeparrot_training - Step 1988: {'lr': 0.000497, 'samples': 381888, 'steps': 1988, 'loss/train': 2.1129146218299866} 01/28/2022 15:24:58 - INFO - codeparrot_training - Step 1989: {'lr': 0.0004972500000000001, 'samples': 382080, 'steps': 1989, 'loss/train': 2.5438878536224365} 01/28/2022 15:25:02 - INFO - codeparrot_training - Step 1990: {'lr': 0.0004975, 'samples': 382272, 'steps': 1990, 'loss/train': 1.4437418282032013} 01/28/2022 15:25:07 - INFO - codeparrot_training - Step 1991: {'lr': 0.00049775, 'samples': 382464, 'steps': 1991, 'loss/train': 2.098858952522278} 01/28/2022 15:25:12 - INFO - codeparrot_training - Step 1992: {'lr': 0.000498, 'samples': 382656, 'steps': 1992, 'loss/train': 1.456356793642044} 01/28/2022 15:25:16 - INFO - codeparrot_training - Step 1993: {'lr': 0.00049825, 'samples': 382848, 'steps': 1993, 'loss/train': 2.171732246875763} 01/28/2022 15:25:20 - INFO - codeparrot_training - Step 1994: {'lr': 0.0004985, 'samples': 383040, 'steps': 1994, 'loss/train': 2.4507097005844116} 01/28/2022 15:25:24 - INFO - codeparrot_training - Step 1995: {'lr': 0.0004987500000000001, 'samples': 383232, 'steps': 1995, 'loss/train': 2.491869628429413} 01/28/2022 15:25:31 - INFO - codeparrot_training - Step 1996: {'lr': 0.000499, 'samples': 383424, 'steps': 1996, 'loss/train': 2.93396258354187} 01/28/2022 15:25:35 - INFO - codeparrot_training - Step 1997: {'lr': 0.00049925, 'samples': 383616, 'steps': 1997, 'loss/train': 2.2395620942115784} 01/28/2022 15:25:39 - INFO - codeparrot_training - Step 1998: {'lr': 0.0004995, 'samples': 383808, 'steps': 1998, 'loss/train': 2.186854362487793} 01/28/2022 15:25:43 - INFO - codeparrot_training - Step 1999: {'lr': 0.0004997500000000001, 'samples': 384000, 'steps': 1999, 'loss/train': 2.520771861076355} 01/28/2022 15:25:43 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/28/2022 15:28:40 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py * [new branch] expert-dust-2 -> expert-dust-2 01/28/2022 15:30:05 - INFO - codeparrot_training - Step 2000: {'lr': 0.0005, 'samples': 384192, 'steps': 2000, 'loss/train': 2.4648334980010986} 01/28/2022 15:30:11 - INFO - codeparrot_training - Step 2001: {'lr': 0.0004999999994645397, 'samples': 384384, 'steps': 2001, 'loss/train': 1.360277384519577} 01/28/2022 15:30:15 - INFO - codeparrot_training - Step 2002: {'lr': 0.0004999999978581587, 'samples': 384576, 'steps': 2002, 'loss/train': 1.5091118216514587} 01/28/2022 15:30:19 - INFO - codeparrot_training - Step 2003: {'lr': 0.0004999999951808573, 'samples': 384768, 'steps': 2003, 'loss/train': 1.934492290019989} 01/28/2022 15:30:23 - INFO - codeparrot_training - Step 2004: {'lr': 0.0004999999914326351, 'samples': 384960, 'steps': 2004, 'loss/train': 4.568555116653442} 01/28/2022 15:30:28 - INFO - codeparrot_training - Step 2005: {'lr': 0.0004999999866134924, 'samples': 385152, 'steps': 2005, 'loss/train': 1.8506308794021606} 01/28/2022 15:30:33 - INFO - codeparrot_training - Step 2006: {'lr': 0.0004999999807234292, 'samples': 385344, 'steps': 2006, 'loss/train': 0.8112654089927673} 01/28/2022 15:30:37 - INFO - codeparrot_training - Step 2007: {'lr': 0.0004999999737624453, 'samples': 385536, 'steps': 2007, 'loss/train': 1.3111443221569061} 01/28/2022 15:30:41 - INFO - codeparrot_training - Step 2008: {'lr': 0.0004999999657305411, 'samples': 385728, 'steps': 2008, 'loss/train': 1.8294358849525452} 01/28/2022 15:30:46 - INFO - codeparrot_training - Step 2009: {'lr': 0.0004999999566277163, 'samples': 385920, 'steps': 2009, 'loss/train': 0.904141366481781} 01/28/2022 15:30:50 - INFO - codeparrot_training - Step 2010: {'lr': 0.0004999999464539711, 'samples': 386112, 'steps': 2010, 'loss/train': 1.9541603922843933} 01/28/2022 15:30:56 - INFO - codeparrot_training - Step 2011: {'lr': 0.0004999999352093055, 'samples': 386304, 'steps': 2011, 'loss/train': 2.2489922046661377} 01/28/2022 15:31:00 - INFO - codeparrot_training - Step 2012: {'lr': 0.0004999999228937196, 'samples': 386496, 'steps': 2012, 'loss/train': 1.3898673355579376} 01/28/2022 15:31:04 - INFO - codeparrot_training - Step 2013: {'lr': 0.0004999999095072135, 'samples': 386688, 'steps': 2013, 'loss/train': 1.9303908348083496} 01/28/2022 15:31:08 - INFO - codeparrot_training - Step 2014: {'lr': 0.0004999998950497869, 'samples': 386880, 'steps': 2014, 'loss/train': 2.2241506576538086} 01/28/2022 15:31:13 - INFO - codeparrot_training - Step 2015: {'lr': 0.0004999998795214404, 'samples': 387072, 'steps': 2015, 'loss/train': 1.9885321855545044} 01/28/2022 15:31:19 - INFO - codeparrot_training - Step 2016: {'lr': 0.0004999998629221736, 'samples': 387264, 'steps': 2016, 'loss/train': 1.7596895098686218} 01/28/2022 15:31:24 - INFO - codeparrot_training - Step 2017: {'lr': 0.0004999998452519869, 'samples': 387456, 'steps': 2017, 'loss/train': 2.2404359579086304} 01/28/2022 15:31:28 - INFO - codeparrot_training - Step 2018: {'lr': 0.0004999998265108802, 'samples': 387648, 'steps': 2018, 'loss/train': 2.113161563873291} 01/28/2022 15:31:32 - INFO - codeparrot_training - Step 2019: {'lr': 0.0004999998066988537, 'samples': 387840, 'steps': 2019, 'loss/train': 2.4524288177490234} 01/28/2022 15:31:36 - INFO - codeparrot_training - Step 2020: {'lr': 0.0004999997858159073, 'samples': 388032, 'steps': 2020, 'loss/train': 2.4555426836013794} 01/28/2022 15:31:42 - INFO - codeparrot_training - Step 2021: {'lr': 0.0004999997638620412, 'samples': 388224, 'steps': 2021, 'loss/train': 2.3438721299171448} 01/28/2022 15:31:46 - INFO - codeparrot_training - Step 2022: {'lr': 0.0004999997408372557, 'samples': 388416, 'steps': 2022, 'loss/train': 1.9850549697875977} 01/28/2022 15:31:50 - INFO - codeparrot_training - Step 2023: {'lr': 0.0004999997167415504, 'samples': 388608, 'steps': 2023, 'loss/train': 1.9485357999801636} 01/28/2022 15:31:54 - INFO - codeparrot_training - Step 2024: {'lr': 0.0004999996915749259, 'samples': 388800, 'steps': 2024, 'loss/train': 4.096869349479675} 01/28/2022 15:31:59 - INFO - codeparrot_training - Step 2025: {'lr': 0.0004999996653373821, 'samples': 388992, 'steps': 2025, 'loss/train': 1.9876874685287476} 01/28/2022 15:32:05 - INFO - codeparrot_training - Step 2026: {'lr': 0.000499999638028919, 'samples': 389184, 'steps': 2026, 'loss/train': 1.3928376138210297} 01/28/2022 15:32:09 - INFO - codeparrot_training - Step 2027: {'lr': 0.0004999996096495369, 'samples': 389376, 'steps': 2027, 'loss/train': 2.572029769420624} 01/28/2022 15:32:14 - INFO - codeparrot_training - Step 2028: {'lr': 0.0004999995801992359, 'samples': 389568, 'steps': 2028, 'loss/train': 1.63018137216568} 01/28/2022 15:32:18 - INFO - codeparrot_training - Step 2029: {'lr': 0.000499999549678016, 'samples': 389760, 'steps': 2029, 'loss/train': 2.1157082319259644} 01/28/2022 15:32:22 - INFO - codeparrot_training - Step 2030: {'lr': 0.0004999995180858774, 'samples': 389952, 'steps': 2030, 'loss/train': 2.177930474281311} 01/28/2022 15:32:27 - INFO - codeparrot_training - Step 2031: {'lr': 0.0004999994854228203, 'samples': 390144, 'steps': 2031, 'loss/train': 2.1615849137306213} 01/28/2022 15:32:31 - INFO - codeparrot_training - Step 2032: {'lr': 0.0004999994516888449, 'samples': 390336, 'steps': 2032, 'loss/train': 0.9083207845687866} 01/28/2022 15:32:36 - INFO - codeparrot_training - Step 2033: {'lr': 0.000499999416883951, 'samples': 390528, 'steps': 2033, 'loss/train': 1.3825498223304749} 01/28/2022 15:32:40 - INFO - codeparrot_training - Step 2034: {'lr': 0.0004999993810081391, 'samples': 390720, 'steps': 2034, 'loss/train': 2.121063530445099} 01/28/2022 15:32:44 - INFO - codeparrot_training - Step 2035: {'lr': 0.0004999993440614092, 'samples': 390912, 'steps': 2035, 'loss/train': 2.1255966424942017} 01/28/2022 15:32:49 - INFO - codeparrot_training - Step 2036: {'lr': 0.0004999993060437616, 'samples': 391104, 'steps': 2036, 'loss/train': 1.4065226912498474} 01/28/2022 15:32:54 - INFO - codeparrot_training - Step 2037: {'lr': 0.0004999992669551962, 'samples': 391296, 'steps': 2037, 'loss/train': 2.4808266162872314} 01/28/2022 15:32:58 - INFO - codeparrot_training - Step 2038: {'lr': 0.0004999992267957135, 'samples': 391488, 'steps': 2038, 'loss/train': 2.2019158601760864} 01/28/2022 15:33:02 - INFO - codeparrot_training - Step 2039: {'lr': 0.0004999991855653134, 'samples': 391680, 'steps': 2039, 'loss/train': 1.7061637043952942} 01/28/2022 15:33:06 - INFO - codeparrot_training - Step 2040: {'lr': 0.0004999991432639963, 'samples': 391872, 'steps': 2040, 'loss/train': 1.2678300440311432} 01/28/2022 15:33:13 - INFO - codeparrot_training - Step 2041: {'lr': 0.0004999990998917621, 'samples': 392064, 'steps': 2041, 'loss/train': 1.9908993244171143} 01/28/2022 15:33:17 - INFO - codeparrot_training - Step 2042: {'lr': 0.0004999990554486111, 'samples': 392256, 'steps': 2042, 'loss/train': 1.3216642141342163} 01/28/2022 15:33:21 - INFO - codeparrot_training - Step 2043: {'lr': 0.0004999990099345436, 'samples': 392448, 'steps': 2043, 'loss/train': 2.3563621044158936} 01/28/2022 15:33:25 - INFO - codeparrot_training - Step 2044: {'lr': 0.0004999989633495597, 'samples': 392640, 'steps': 2044, 'loss/train': 1.45200115442276} 01/28/2022 15:33:30 - INFO - codeparrot_training - Step 2045: {'lr': 0.0004999989156936597, 'samples': 392832, 'steps': 2045, 'loss/train': 2.285312533378601} 01/28/2022 15:33:35 - INFO - codeparrot_training - Step 2046: {'lr': 0.0004999988669668437, 'samples': 393024, 'steps': 2046, 'loss/train': 1.6052595376968384} 01/28/2022 15:33:39 - INFO - codeparrot_training - Step 2047: {'lr': 0.0004999988171691119, 'samples': 393216, 'steps': 2047, 'loss/train': 2.519753873348236} 01/28/2022 15:33:43 - INFO - codeparrot_training - Step 2048: {'lr': 0.0004999987663004646, 'samples': 393408, 'steps': 2048, 'loss/train': 1.920848786830902} 01/28/2022 15:33:48 - INFO - codeparrot_training - Step 2049: {'lr': 0.0004999987143609019, 'samples': 393600, 'steps': 2049, 'loss/train': 1.762715756893158} 01/28/2022 15:33:52 - INFO - codeparrot_training - Step 2050: {'lr': 0.0004999986613504242, 'samples': 393792, 'steps': 2050, 'loss/train': 2.944675862789154} 01/28/2022 15:33:58 - INFO - codeparrot_training - Step 2051: {'lr': 0.0004999986072690315, 'samples': 393984, 'steps': 2051, 'loss/train': 1.127686321735382} 01/28/2022 15:34:02 - INFO - codeparrot_training - Step 2052: {'lr': 0.0004999985521167242, 'samples': 394176, 'steps': 2052, 'loss/train': 0.7359496057033539} 01/28/2022 15:34:07 - INFO - codeparrot_training - Step 2053: {'lr': 0.0004999984958935025, 'samples': 394368, 'steps': 2053, 'loss/train': 1.9092466235160828} 01/28/2022 15:34:11 - INFO - codeparrot_training - Step 2054: {'lr': 0.0004999984385993665, 'samples': 394560, 'steps': 2054, 'loss/train': 2.571120321750641} 01/28/2022 15:34:15 - INFO - codeparrot_training - Step 2055: {'lr': 0.0004999983802343168, 'samples': 394752, 'steps': 2055, 'loss/train': 1.6150295734405518} 01/28/2022 15:34:20 - INFO - codeparrot_training - Step 2056: {'lr': 0.0004999983207983532, 'samples': 394944, 'steps': 2056, 'loss/train': 1.8920574188232422} 01/28/2022 15:34:25 - INFO - codeparrot_training - Step 2057: {'lr': 0.0004999982602914763, 'samples': 395136, 'steps': 2057, 'loss/train': 2.401711642742157} 01/28/2022 15:34:29 - INFO - codeparrot_training - Step 2058: {'lr': 0.0004999981987136862, 'samples': 395328, 'steps': 2058, 'loss/train': 1.0922338664531708} 01/28/2022 15:34:33 - INFO - codeparrot_training - Step 2059: {'lr': 0.0004999981360649833, 'samples': 395520, 'steps': 2059, 'loss/train': 2.693662941455841} 01/28/2022 15:34:37 - INFO - codeparrot_training - Step 2060: {'lr': 0.0004999980723453676, 'samples': 395712, 'steps': 2060, 'loss/train': 1.789076328277588} 01/28/2022 15:34:42 - INFO - codeparrot_training - Step 2061: {'lr': 0.0004999980075548397, 'samples': 395904, 'steps': 2061, 'loss/train': 1.8108270764350891} 01/28/2022 15:34:47 - INFO - codeparrot_training - Step 2062: {'lr': 0.0004999979416933997, 'samples': 396096, 'steps': 2062, 'loss/train': 2.928839385509491} 01/28/2022 15:34:51 - INFO - codeparrot_training - Step 2063: {'lr': 0.0004999978747610478, 'samples': 396288, 'steps': 2063, 'loss/train': 1.8250860571861267} 01/28/2022 15:34:55 - INFO - codeparrot_training - Step 2064: {'lr': 0.0004999978067577843, 'samples': 396480, 'steps': 2064, 'loss/train': 1.4757256507873535} 01/28/2022 15:35:00 - INFO - codeparrot_training - Step 2065: {'lr': 0.0004999977376836098, 'samples': 396672, 'steps': 2065, 'loss/train': 1.6121065020561218} 01/28/2022 15:35:06 - INFO - codeparrot_training - Step 2066: {'lr': 0.0004999976675385243, 'samples': 396864, 'steps': 2066, 'loss/train': 2.3945472836494446} 01/28/2022 15:35:11 - INFO - codeparrot_training - Step 2067: {'lr': 0.0004999975963225282, 'samples': 397056, 'steps': 2067, 'loss/train': 1.8801158666610718} 01/28/2022 15:35:15 - INFO - codeparrot_training - Step 2068: {'lr': 0.0004999975240356217, 'samples': 397248, 'steps': 2068, 'loss/train': 3.3503966331481934} 01/28/2022 15:35:19 - INFO - codeparrot_training - Step 2069: {'lr': 0.0004999974506778053, 'samples': 397440, 'steps': 2069, 'loss/train': 1.8097543716430664} 01/28/2022 15:35:24 - INFO - codeparrot_training - Step 2070: {'lr': 0.0004999973762490792, 'samples': 397632, 'steps': 2070, 'loss/train': 2.1214240193367004} 01/28/2022 15:35:29 - INFO - codeparrot_training - Step 2071: {'lr': 0.0004999973007494436, 'samples': 397824, 'steps': 2071, 'loss/train': 1.5223910808563232} 01/28/2022 15:35:33 - INFO - codeparrot_training - Step 2072: {'lr': 0.000499997224178899, 'samples': 398016, 'steps': 2072, 'loss/train': 1.1178030967712402} 01/28/2022 15:35:37 - INFO - codeparrot_training - Step 2073: {'lr': 0.0004999971465374457, 'samples': 398208, 'steps': 2073, 'loss/train': 2.7821273803710938} 01/28/2022 15:35:42 - INFO - codeparrot_training - Step 2074: {'lr': 0.000499997067825084, 'samples': 398400, 'steps': 2074, 'loss/train': 2.7264904975891113} 01/28/2022 15:35:46 - INFO - codeparrot_training - Step 2075: {'lr': 0.0004999969880418142, 'samples': 398592, 'steps': 2075, 'loss/train': 2.08796489238739} 01/28/2022 15:35:51 - INFO - codeparrot_training - Step 2076: {'lr': 0.0004999969071876367, 'samples': 398784, 'steps': 2076, 'loss/train': 2.0501139163970947} 01/28/2022 15:35:55 - INFO - codeparrot_training - Step 2077: {'lr': 0.0004999968252625519, 'samples': 398976, 'steps': 2077, 'loss/train': 2.597481071949005} 01/28/2022 15:35:59 - INFO - codeparrot_training - Step 2078: {'lr': 0.00049999674226656, 'samples': 399168, 'steps': 2078, 'loss/train': 2.387266516685486} 01/28/2022 15:36:04 - INFO - codeparrot_training - Step 2079: {'lr': 0.0004999966581996616, 'samples': 399360, 'steps': 2079, 'loss/train': 2.1430370807647705} 01/28/2022 15:36:08 - INFO - codeparrot_training - Step 2080: {'lr': 0.0004999965730618567, 'samples': 399552, 'steps': 2080, 'loss/train': 1.2976401150226593} 01/28/2022 15:36:13 - INFO - codeparrot_training - Step 2081: {'lr': 0.000499996486853146, 'samples': 399744, 'steps': 2081, 'loss/train': 7.043699741363525} 01/28/2022 15:36:18 - INFO - codeparrot_training - Step 2082: {'lr': 0.0004999963995735296, 'samples': 399936, 'steps': 2082, 'loss/train': 1.9133990406990051} 01/28/2022 15:36:22 - INFO - codeparrot_training - Step 2083: {'lr': 0.0004999963112230081, 'samples': 400128, 'steps': 2083, 'loss/train': 1.9343887567520142} 01/28/2022 15:36:26 - INFO - codeparrot_training - Step 2084: {'lr': 0.0004999962218015818, 'samples': 400320, 'steps': 2084, 'loss/train': 2.6069172620773315} 01/28/2022 15:36:30 - INFO - codeparrot_training - Step 2085: {'lr': 0.0004999961313092511, 'samples': 400512, 'steps': 2085, 'loss/train': 1.6248759627342224} 01/28/2022 15:36:37 - INFO - codeparrot_training - Step 2086: {'lr': 0.0004999960397460162, 'samples': 400704, 'steps': 2086, 'loss/train': 2.190696895122528} 01/28/2022 15:36:41 - INFO - codeparrot_training - Step 2087: {'lr': 0.0004999959471118778, 'samples': 400896, 'steps': 2087, 'loss/train': 2.152191638946533} 01/28/2022 15:36:45 - INFO - codeparrot_training - Step 2088: {'lr': 0.000499995853406836, 'samples': 401088, 'steps': 2088, 'loss/train': 2.9412331581115723} 01/28/2022 15:36:49 - INFO - codeparrot_training - Step 2089: {'lr': 0.0004999957586308914, 'samples': 401280, 'steps': 2089, 'loss/train': 1.8365054726600647} 01/28/2022 15:36:54 - INFO - codeparrot_training - Step 2090: {'lr': 0.0004999956627840445, 'samples': 401472, 'steps': 2090, 'loss/train': 1.817452847957611} 01/28/2022 15:36:59 - INFO - codeparrot_training - Step 2091: {'lr': 0.0004999955658662954, 'samples': 401664, 'steps': 2091, 'loss/train': 1.901893436908722} 01/28/2022 15:37:03 - INFO - codeparrot_training - Step 2092: {'lr': 0.0004999954678776448, 'samples': 401856, 'steps': 2092, 'loss/train': 2.106295108795166} 01/28/2022 15:37:07 - INFO - codeparrot_training - Step 2093: {'lr': 0.0004999953688180929, 'samples': 402048, 'steps': 2093, 'loss/train': 2.3385645747184753} 01/28/2022 15:37:11 - INFO - codeparrot_training - Step 2094: {'lr': 0.0004999952686876402, 'samples': 402240, 'steps': 2094, 'loss/train': 2.74893057346344} 01/28/2022 15:37:16 - INFO - codeparrot_training - Step 2095: {'lr': 0.0004999951674862872, 'samples': 402432, 'steps': 2095, 'loss/train': 1.5605614185333252} 01/28/2022 15:37:21 - INFO - codeparrot_training - Step 2096: {'lr': 0.0004999950652140343, 'samples': 402624, 'steps': 2096, 'loss/train': 1.3050853908061981} 01/28/2022 15:37:25 - INFO - codeparrot_training - Step 2097: {'lr': 0.0004999949618708819, 'samples': 402816, 'steps': 2097, 'loss/train': 2.1385136246681213} 01/28/2022 15:37:29 - INFO - codeparrot_training - Step 2098: {'lr': 0.0004999948574568305, 'samples': 403008, 'steps': 2098, 'loss/train': 1.9483131766319275} 01/28/2022 15:37:34 - INFO - codeparrot_training - Step 2099: {'lr': 0.0004999947519718805, 'samples': 403200, 'steps': 2099, 'loss/train': 2.078092575073242} 01/28/2022 15:37:38 - INFO - codeparrot_training - Step 2100: {'lr': 0.0004999946454160324, 'samples': 403392, 'steps': 2100, 'loss/train': 1.8854087591171265} 01/28/2022 15:37:44 - INFO - codeparrot_training - Step 2101: {'lr': 0.0004999945377892865, 'samples': 403584, 'steps': 2101, 'loss/train': 1.9062979817390442} 01/28/2022 15:37:48 - INFO - codeparrot_training - Step 2102: {'lr': 0.0004999944290916434, 'samples': 403776, 'steps': 2102, 'loss/train': 0.25786393135786057} 01/28/2022 15:37:52 - INFO - codeparrot_training - Step 2103: {'lr': 0.0004999943193231037, 'samples': 403968, 'steps': 2103, 'loss/train': 2.2201942205429077} 01/28/2022 15:37:57 - INFO - codeparrot_training - Step 2104: {'lr': 0.0004999942084836675, 'samples': 404160, 'steps': 2104, 'loss/train': 2.093398869037628} 01/28/2022 15:38:01 - INFO - codeparrot_training - Step 2105: {'lr': 0.0004999940965733356, 'samples': 404352, 'steps': 2105, 'loss/train': 2.0284268260002136} 01/28/2022 15:38:06 - INFO - codeparrot_training - Step 2106: {'lr': 0.0004999939835921085, 'samples': 404544, 'steps': 2106, 'loss/train': 1.4982639849185944} 01/28/2022 15:38:10 - INFO - codeparrot_training - Step 2107: {'lr': 0.0004999938695399864, 'samples': 404736, 'steps': 2107, 'loss/train': 2.0829466581344604} 01/28/2022 15:38:15 - INFO - codeparrot_training - Step 2108: {'lr': 0.00049999375441697, 'samples': 404928, 'steps': 2108, 'loss/train': 1.4869933426380157} 01/28/2022 15:38:19 - INFO - codeparrot_training - Step 2109: {'lr': 0.0004999936382230597, 'samples': 405120, 'steps': 2109, 'loss/train': 1.8437515497207642} 01/28/2022 15:38:23 - INFO - codeparrot_training - Step 2110: {'lr': 0.000499993520958256, 'samples': 405312, 'steps': 2110, 'loss/train': 2.0545350909233093} 01/28/2022 15:38:30 - INFO - codeparrot_training - Step 2111: {'lr': 0.0004999934026225595, 'samples': 405504, 'steps': 2111, 'loss/train': 1.9734147191047668} 01/28/2022 15:38:34 - INFO - codeparrot_training - Step 2112: {'lr': 0.0004999932832159708, 'samples': 405696, 'steps': 2112, 'loss/train': 2.09454345703125} 01/28/2022 15:38:38 - INFO - codeparrot_training - Step 2113: {'lr': 0.00049999316273849, 'samples': 405888, 'steps': 2113, 'loss/train': 2.29085236787796} 01/28/2022 15:38:43 - INFO - codeparrot_training - Step 2114: {'lr': 0.0004999930411901181, 'samples': 406080, 'steps': 2114, 'loss/train': 2.3042680621147156} 01/28/2022 15:38:47 - INFO - codeparrot_training - Step 2115: {'lr': 0.0004999929185708551, 'samples': 406272, 'steps': 2115, 'loss/train': 1.3937213122844696} 01/28/2022 15:38:51 - INFO - codeparrot_training - Step 2116: {'lr': 0.000499992794880702, 'samples': 406464, 'steps': 2116, 'loss/train': 2.9675331115722656} 01/28/2022 15:38:56 - INFO - codeparrot_training - Step 2117: {'lr': 0.0004999926701196592, 'samples': 406656, 'steps': 2117, 'loss/train': 2.3369632959365845} 01/28/2022 15:39:01 - INFO - codeparrot_training - Step 2118: {'lr': 0.0004999925442877271, 'samples': 406848, 'steps': 2118, 'loss/train': 3.180352807044983} 01/28/2022 15:39:05 - INFO - codeparrot_training - Step 2119: {'lr': 0.0004999924173849063, 'samples': 407040, 'steps': 2119, 'loss/train': 1.479256957769394} 01/28/2022 15:39:09 - INFO - codeparrot_training - Step 2120: {'lr': 0.0004999922894111975, 'samples': 407232, 'steps': 2120, 'loss/train': 2.3040069937705994} 01/28/2022 15:39:13 - INFO - codeparrot_training - Step 2121: {'lr': 0.000499992160366601, 'samples': 407424, 'steps': 2121, 'loss/train': 1.957931399345398} 01/28/2022 15:39:19 - INFO - codeparrot_training - Step 2122: {'lr': 0.0004999920302511175, 'samples': 407616, 'steps': 2122, 'loss/train': 1.7469400763511658} 01/28/2022 15:39:23 - INFO - codeparrot_training - Step 2123: {'lr': 0.0004999918990647474, 'samples': 407808, 'steps': 2123, 'loss/train': 1.5121530890464783} 01/28/2022 15:39:27 - INFO - codeparrot_training - Step 2124: {'lr': 0.0004999917668074915, 'samples': 408000, 'steps': 2124, 'loss/train': 2.787859618663788} 01/28/2022 15:39:31 - INFO - codeparrot_training - Step 2125: {'lr': 0.0004999916334793503, 'samples': 408192, 'steps': 2125, 'loss/train': 2.2604493498802185} 01/28/2022 15:39:36 - INFO - codeparrot_training - Step 2126: {'lr': 0.0004999914990803242, 'samples': 408384, 'steps': 2126, 'loss/train': 2.0772737860679626} 01/28/2022 15:39:42 - INFO - codeparrot_training - Step 2127: {'lr': 0.000499991363610414, 'samples': 408576, 'steps': 2127, 'loss/train': 1.4657529294490814} 01/28/2022 15:39:46 - INFO - codeparrot_training - Step 2128: {'lr': 0.0004999912270696202, 'samples': 408768, 'steps': 2128, 'loss/train': 1.8421939015388489} 01/28/2022 15:39:51 - INFO - codeparrot_training - Step 2129: {'lr': 0.0004999910894579432, 'samples': 408960, 'steps': 2129, 'loss/train': 0.5818967968225479} 01/28/2022 15:39:55 - INFO - codeparrot_training - Step 2130: {'lr': 0.000499990950775384, 'samples': 409152, 'steps': 2130, 'loss/train': 1.9389642477035522} 01/28/2022 15:39:59 - INFO - codeparrot_training - Step 2131: {'lr': 0.0004999908110219428, 'samples': 409344, 'steps': 2131, 'loss/train': 2.5843276977539062} 01/28/2022 15:40:04 - INFO - codeparrot_training - Step 2132: {'lr': 0.0004999906701976203, 'samples': 409536, 'steps': 2132, 'loss/train': 2.568407714366913} 01/28/2022 15:40:09 - INFO - codeparrot_training - Step 2133: {'lr': 0.0004999905283024172, 'samples': 409728, 'steps': 2133, 'loss/train': 2.5085731744766235} 01/28/2022 15:40:13 - INFO - codeparrot_training - Step 2134: {'lr': 0.0004999903853363341, 'samples': 409920, 'steps': 2134, 'loss/train': 2.0518171191215515} 01/28/2022 15:40:17 - INFO - codeparrot_training - Step 2135: {'lr': 0.0004999902412993715, 'samples': 410112, 'steps': 2135, 'loss/train': 1.9156081080436707} 01/28/2022 15:40:21 - INFO - codeparrot_training - Step 2136: {'lr': 0.0004999900961915302, 'samples': 410304, 'steps': 2136, 'loss/train': 1.669373631477356} 01/28/2022 15:40:27 - INFO - codeparrot_training - Step 2137: {'lr': 0.0004999899500128107, 'samples': 410496, 'steps': 2137, 'loss/train': 2.0815570950508118} 01/28/2022 15:40:31 - INFO - codeparrot_training - Step 2138: {'lr': 0.0004999898027632135, 'samples': 410688, 'steps': 2138, 'loss/train': 1.1626326441764832} 01/28/2022 15:40:35 - INFO - codeparrot_training - Step 2139: {'lr': 0.0004999896544427394, 'samples': 410880, 'steps': 2139, 'loss/train': 1.7939379215240479} 01/28/2022 15:40:39 - INFO - codeparrot_training - Step 2140: {'lr': 0.0004999895050513891, 'samples': 411072, 'steps': 2140, 'loss/train': 1.2988113462924957} 01/28/2022 15:40:43 - INFO - codeparrot_training - Step 2141: {'lr': 0.0004999893545891631, 'samples': 411264, 'steps': 2141, 'loss/train': 2.825515329837799} 01/28/2022 15:40:49 - INFO - codeparrot_training - Step 2142: {'lr': 0.000499989203056062, 'samples': 411456, 'steps': 2142, 'loss/train': 1.7730841040611267} 01/28/2022 15:40:53 - INFO - codeparrot_training - Step 2143: {'lr': 0.0004999890504520866, 'samples': 411648, 'steps': 2143, 'loss/train': 1.7108275294303894} 01/28/2022 15:40:57 - INFO - codeparrot_training - Step 2144: {'lr': 0.0004999888967772375, 'samples': 411840, 'steps': 2144, 'loss/train': 1.2486152350902557} 01/28/2022 15:41:02 - INFO - codeparrot_training - Step 2145: {'lr': 0.0004999887420315153, 'samples': 412032, 'steps': 2145, 'loss/train': 2.6202152967453003} 01/28/2022 15:41:06 - INFO - codeparrot_training - Step 2146: {'lr': 0.0004999885862149207, 'samples': 412224, 'steps': 2146, 'loss/train': 1.8913323283195496} 01/28/2022 15:41:12 - INFO - codeparrot_training - Step 2147: {'lr': 0.0004999884293274545, 'samples': 412416, 'steps': 2147, 'loss/train': 0.7832641303539276} 01/28/2022 15:41:16 - INFO - codeparrot_training - Step 2148: {'lr': 0.0004999882713691171, 'samples': 412608, 'steps': 2148, 'loss/train': 1.77989262342453} 01/28/2022 15:41:20 - INFO - codeparrot_training - Step 2149: {'lr': 0.0004999881123399093, 'samples': 412800, 'steps': 2149, 'loss/train': 2.829009532928467} 01/28/2022 15:41:25 - INFO - codeparrot_training - Step 2150: {'lr': 0.000499987952239832, 'samples': 412992, 'steps': 2150, 'loss/train': 2.096156358718872} 01/28/2022 15:41:29 - INFO - codeparrot_training - Step 2151: {'lr': 0.0004999877910688856, 'samples': 413184, 'steps': 2151, 'loss/train': 1.507472813129425} 01/28/2022 15:41:34 - INFO - codeparrot_training - Step 2152: {'lr': 0.0004999876288270708, 'samples': 413376, 'steps': 2152, 'loss/train': 1.3206222653388977} 01/28/2022 15:41:39 - INFO - codeparrot_training - Step 2153: {'lr': 0.0004999874655143886, 'samples': 413568, 'steps': 2153, 'loss/train': 1.4811393320560455} 01/28/2022 15:41:43 - INFO - codeparrot_training - Step 2154: {'lr': 0.0004999873011308393, 'samples': 413760, 'steps': 2154, 'loss/train': 1.3359569013118744} 01/28/2022 15:41:47 - INFO - codeparrot_training - Step 2155: {'lr': 0.0004999871356764238, 'samples': 413952, 'steps': 2155, 'loss/train': 1.6113946437835693} 01/28/2022 15:41:51 - INFO - codeparrot_training - Step 2156: {'lr': 0.0004999869691511428, 'samples': 414144, 'steps': 2156, 'loss/train': 1.793084442615509} 01/28/2022 15:41:57 - INFO - codeparrot_training - Step 2157: {'lr': 0.000499986801554997, 'samples': 414336, 'steps': 2157, 'loss/train': 2.2589369416236877} 01/28/2022 15:42:01 - INFO - codeparrot_training - Step 2158: {'lr': 0.0004999866328879871, 'samples': 414528, 'steps': 2158, 'loss/train': 0.4259883910417557} 01/28/2022 15:42:06 - INFO - codeparrot_training - Step 2159: {'lr': 0.0004999864631501139, 'samples': 414720, 'steps': 2159, 'loss/train': 2.0754974484443665} 01/28/2022 15:42:10 - INFO - codeparrot_training - Step 2160: {'lr': 0.000499986292341378, 'samples': 414912, 'steps': 2160, 'loss/train': 2.1510087847709656} 01/28/2022 15:42:14 - INFO - codeparrot_training - Step 2161: {'lr': 0.0004999861204617803, 'samples': 415104, 'steps': 2161, 'loss/train': 0.6205110400915146} 01/28/2022 15:42:19 - INFO - codeparrot_training - Step 2162: {'lr': 0.0004999859475113213, 'samples': 415296, 'steps': 2162, 'loss/train': 1.4042243957519531} 01/28/2022 15:42:24 - INFO - codeparrot_training - Step 2163: {'lr': 0.0004999857734900021, 'samples': 415488, 'steps': 2163, 'loss/train': 1.8658095002174377} 01/28/2022 15:42:28 - INFO - codeparrot_training - Step 2164: {'lr': 0.000499985598397823, 'samples': 415680, 'steps': 2164, 'loss/train': 2.082563281059265} 01/28/2022 15:42:32 - INFO - codeparrot_training - Step 2165: {'lr': 0.0004999854222347851, 'samples': 415872, 'steps': 2165, 'loss/train': 1.750041425228119} 01/28/2022 15:42:36 - INFO - codeparrot_training - Step 2166: {'lr': 0.000499985245000889, 'samples': 416064, 'steps': 2166, 'loss/train': 1.477504849433899} 01/28/2022 15:42:42 - INFO - codeparrot_training - Step 2167: {'lr': 0.0004999850666961355, 'samples': 416256, 'steps': 2167, 'loss/train': 1.4194487929344177} 01/28/2022 15:42:46 - INFO - codeparrot_training - Step 2168: {'lr': 0.0004999848873205254, 'samples': 416448, 'steps': 2168, 'loss/train': 1.7474194765090942} 01/28/2022 15:42:50 - INFO - codeparrot_training - Step 2169: {'lr': 0.0004999847068740593, 'samples': 416640, 'steps': 2169, 'loss/train': 1.851754367351532} 01/28/2022 15:42:54 - INFO - codeparrot_training - Step 2170: {'lr': 0.0004999845253567382, 'samples': 416832, 'steps': 2170, 'loss/train': 1.5173689126968384} 01/28/2022 15:42:59 - INFO - codeparrot_training - Step 2171: {'lr': 0.0004999843427685627, 'samples': 417024, 'steps': 2171, 'loss/train': 2.003072440624237} 01/28/2022 15:43:05 - INFO - codeparrot_training - Step 2172: {'lr': 0.0004999841591095337, 'samples': 417216, 'steps': 2172, 'loss/train': 0.47783276438713074} 01/28/2022 15:43:10 - INFO - codeparrot_training - Step 2173: {'lr': 0.0004999839743796519, 'samples': 417408, 'steps': 2173, 'loss/train': 2.1769580841064453} 01/28/2022 15:43:14 - INFO - codeparrot_training - Step 2174: {'lr': 0.0004999837885789182, 'samples': 417600, 'steps': 2174, 'loss/train': 2.4043081998825073} 01/28/2022 15:43:18 - INFO - codeparrot_training - Step 2175: {'lr': 0.0004999836017073332, 'samples': 417792, 'steps': 2175, 'loss/train': 3.9788228273391724} 01/28/2022 15:43:23 - INFO - codeparrot_training - Step 2176: {'lr': 0.000499983413764898, 'samples': 417984, 'steps': 2176, 'loss/train': 1.8278219103813171} 01/28/2022 15:43:28 - INFO - codeparrot_training - Step 2177: {'lr': 0.0004999832247516132, 'samples': 418176, 'steps': 2177, 'loss/train': 1.6683207750320435} 01/28/2022 15:43:32 - INFO - codeparrot_training - Step 2178: {'lr': 0.0004999830346674796, 'samples': 418368, 'steps': 2178, 'loss/train': 1.16720312833786} 01/28/2022 15:43:37 - INFO - codeparrot_training - Step 2179: {'lr': 0.000499982843512498, 'samples': 418560, 'steps': 2179, 'loss/train': 1.8090879321098328} 01/28/2022 15:43:41 - INFO - codeparrot_training - Step 2180: {'lr': 0.0004999826512866693, 'samples': 418752, 'steps': 2180, 'loss/train': 1.4184968769550323} 01/28/2022 15:43:45 - INFO - codeparrot_training - Step 2181: {'lr': 0.0004999824579899944, 'samples': 418944, 'steps': 2181, 'loss/train': 1.7986210584640503} 01/28/2022 15:43:51 - INFO - codeparrot_training - Step 2182: {'lr': 0.000499982263622474, 'samples': 419136, 'steps': 2182, 'loss/train': 0.972728043794632} 01/28/2022 15:43:55 - INFO - codeparrot_training - Step 2183: {'lr': 0.0004999820681841088, 'samples': 419328, 'steps': 2183, 'loss/train': 2.100121021270752} 01/28/2022 15:44:00 - INFO - codeparrot_training - Step 2184: {'lr': 0.0004999818716748999, 'samples': 419520, 'steps': 2184, 'loss/train': 2.7601786851882935} 01/28/2022 15:44:04 - INFO - codeparrot_training - Step 2185: {'lr': 0.0004999816740948481, 'samples': 419712, 'steps': 2185, 'loss/train': 2.2432082891464233} 01/28/2022 15:44:08 - INFO - codeparrot_training - Step 2186: {'lr': 0.0004999814754439542, 'samples': 419904, 'steps': 2186, 'loss/train': 1.6851373314857483} 01/28/2022 15:44:13 - INFO - codeparrot_training - Step 2187: {'lr': 0.000499981275722219, 'samples': 420096, 'steps': 2187, 'loss/train': 1.7919908165931702} 01/28/2022 15:44:18 - INFO - codeparrot_training - Step 2188: {'lr': 0.0004999810749296434, 'samples': 420288, 'steps': 2188, 'loss/train': 2.2357996702194214} 01/28/2022 15:44:22 - INFO - codeparrot_training - Step 2189: {'lr': 0.0004999808730662282, 'samples': 420480, 'steps': 2189, 'loss/train': 2.2453479766845703} 01/28/2022 15:44:26 - INFO - codeparrot_training - Step 2190: {'lr': 0.0004999806701319743, 'samples': 420672, 'steps': 2190, 'loss/train': 2.6818124055862427} 01/28/2022 15:44:31 - INFO - codeparrot_training - Step 2191: {'lr': 0.0004999804661268827, 'samples': 420864, 'steps': 2191, 'loss/train': 1.8735188841819763} 01/28/2022 15:44:36 - INFO - codeparrot_training - Step 2192: {'lr': 0.0004999802610509541, 'samples': 421056, 'steps': 2192, 'loss/train': 2.3641823530197144} 01/28/2022 15:44:40 - INFO - codeparrot_training - Step 2193: {'lr': 0.0004999800549041894, 'samples': 421248, 'steps': 2193, 'loss/train': 2.5990995168685913} 01/28/2022 15:44:44 - INFO - codeparrot_training - Step 2194: {'lr': 0.0004999798476865895, 'samples': 421440, 'steps': 2194, 'loss/train': 2.1084700226783752} 01/28/2022 15:44:48 - INFO - codeparrot_training - Step 2195: {'lr': 0.0004999796393981554, 'samples': 421632, 'steps': 2195, 'loss/train': 2.4917827248573303} 01/28/2022 15:44:54 - INFO - codeparrot_training - Step 2196: {'lr': 0.0004999794300388879, 'samples': 421824, 'steps': 2196, 'loss/train': 1.9567508697509766} 01/28/2022 15:44:58 - INFO - codeparrot_training - Step 2197: {'lr': 0.0004999792196087879, 'samples': 422016, 'steps': 2197, 'loss/train': 2.1800875067710876} 01/28/2022 15:45:02 - INFO - codeparrot_training - Step 2198: {'lr': 0.0004999790081078562, 'samples': 422208, 'steps': 2198, 'loss/train': 1.992081642150879} 01/28/2022 15:45:06 - INFO - codeparrot_training - Step 2199: {'lr': 0.0004999787955360939, 'samples': 422400, 'steps': 2199, 'loss/train': 2.6228227615356445} 01/28/2022 15:45:11 - INFO - codeparrot_training - Step 2200: {'lr': 0.0004999785818935018, 'samples': 422592, 'steps': 2200, 'loss/train': 1.6961127519607544} 01/28/2022 15:45:17 - INFO - codeparrot_training - Step 2201: {'lr': 0.0004999783671800808, 'samples': 422784, 'steps': 2201, 'loss/train': 1.7837233543395996} 01/28/2022 15:45:21 - INFO - codeparrot_training - Step 2202: {'lr': 0.0004999781513958318, 'samples': 422976, 'steps': 2202, 'loss/train': 1.127033293247223} 01/28/2022 15:45:25 - INFO - codeparrot_training - Step 2203: {'lr': 0.000499977934540756, 'samples': 423168, 'steps': 2203, 'loss/train': 2.3698012232780457} 01/28/2022 15:45:29 - INFO - codeparrot_training - Step 2204: {'lr': 0.0004999777166148539, 'samples': 423360, 'steps': 2204, 'loss/train': 1.932809293270111} 01/28/2022 15:45:34 - INFO - codeparrot_training - Step 2205: {'lr': 0.0004999774976181267, 'samples': 423552, 'steps': 2205, 'loss/train': 2.0018935203552246} 01/28/2022 15:45:39 - INFO - codeparrot_training - Step 2206: {'lr': 0.0004999772775505753, 'samples': 423744, 'steps': 2206, 'loss/train': 1.7723352313041687} 01/28/2022 15:45:43 - INFO - codeparrot_training - Step 2207: {'lr': 0.0004999770564122005, 'samples': 423936, 'steps': 2207, 'loss/train': 2.1038333773612976} 01/28/2022 15:45:48 - INFO - codeparrot_training - Step 2208: {'lr': 0.0004999768342030035, 'samples': 424128, 'steps': 2208, 'loss/train': 1.4957935810089111} 01/28/2022 15:45:52 - INFO - codeparrot_training - Step 2209: {'lr': 0.0004999766109229851, 'samples': 424320, 'steps': 2209, 'loss/train': 2.2967782616615295} 01/28/2022 15:45:56 - INFO - codeparrot_training - Step 2210: {'lr': 0.0004999763865721463, 'samples': 424512, 'steps': 2210, 'loss/train': 2.2508904933929443} 01/28/2022 15:46:00 - INFO - codeparrot_training - Step 2211: {'lr': 0.000499976161150488, 'samples': 424704, 'steps': 2211, 'loss/train': 2.1112967133522034} 01/28/2022 15:46:06 - INFO - codeparrot_training - Step 2212: {'lr': 0.0004999759346580111, 'samples': 424896, 'steps': 2212, 'loss/train': 2.505953371524811} 01/28/2022 15:46:10 - INFO - codeparrot_training - Step 2213: {'lr': 0.0004999757070947168, 'samples': 425088, 'steps': 2213, 'loss/train': 3.0459762811660767} 01/28/2022 15:46:14 - INFO - codeparrot_training - Step 2214: {'lr': 0.0004999754784606058, 'samples': 425280, 'steps': 2214, 'loss/train': 2.322206497192383} 01/28/2022 15:46:19 - INFO - codeparrot_training - Step 2215: {'lr': 0.0004999752487556794, 'samples': 425472, 'steps': 2215, 'loss/train': 1.1713159382343292} 01/28/2022 15:46:25 - INFO - codeparrot_training - Step 2216: {'lr': 0.0004999750179799383, 'samples': 425664, 'steps': 2216, 'loss/train': 1.5454423427581787} 01/28/2022 15:46:29 - INFO - codeparrot_training - Step 2217: {'lr': 0.0004999747861333838, 'samples': 425856, 'steps': 2217, 'loss/train': 1.43303382396698} 01/28/2022 15:46:33 - INFO - codeparrot_training - Step 2218: {'lr': 0.0004999745532160164, 'samples': 426048, 'steps': 2218, 'loss/train': 1.3547088503837585} 01/28/2022 15:46:38 - INFO - codeparrot_training - Step 2219: {'lr': 0.0004999743192278377, 'samples': 426240, 'steps': 2219, 'loss/train': 1.5508933067321777} 01/28/2022 15:46:42 - INFO - codeparrot_training - Step 2220: {'lr': 0.0004999740841688481, 'samples': 426432, 'steps': 2220, 'loss/train': 0.930799126625061} 01/28/2022 15:46:47 - INFO - codeparrot_training - Step 2221: {'lr': 0.000499973848039049, 'samples': 426624, 'steps': 2221, 'loss/train': 1.9905629754066467} 01/28/2022 15:46:52 - INFO - codeparrot_training - Step 2222: {'lr': 0.0004999736108384414, 'samples': 426816, 'steps': 2222, 'loss/train': 1.1066862344741821} 01/28/2022 15:46:56 - INFO - codeparrot_training - Step 2223: {'lr': 0.0004999733725670261, 'samples': 427008, 'steps': 2223, 'loss/train': 1.6013107895851135} 01/28/2022 15:47:00 - INFO - codeparrot_training - Step 2224: {'lr': 0.0004999731332248044, 'samples': 427200, 'steps': 2224, 'loss/train': 2.8853284120559692} 01/28/2022 15:47:04 - INFO - codeparrot_training - Step 2225: {'lr': 0.0004999728928117771, 'samples': 427392, 'steps': 2225, 'loss/train': 1.1638171970844269} 01/28/2022 15:47:10 - INFO - codeparrot_training - Step 2226: {'lr': 0.0004999726513279452, 'samples': 427584, 'steps': 2226, 'loss/train': 1.4880283176898956} 01/28/2022 15:47:14 - INFO - codeparrot_training - Step 2227: {'lr': 0.0004999724087733099, 'samples': 427776, 'steps': 2227, 'loss/train': 2.458466649055481} 01/28/2022 15:47:19 - INFO - codeparrot_training - Step 2228: {'lr': 0.0004999721651478723, 'samples': 427968, 'steps': 2228, 'loss/train': 1.9671621322631836} 01/28/2022 15:47:23 - INFO - codeparrot_training - Step 2229: {'lr': 0.0004999719204516332, 'samples': 428160, 'steps': 2229, 'loss/train': 2.226933181285858} 01/28/2022 15:47:27 - INFO - codeparrot_training - Step 2230: {'lr': 0.0004999716746845937, 'samples': 428352, 'steps': 2230, 'loss/train': 1.558745563030243} 01/28/2022 15:47:32 - INFO - codeparrot_training - Step 2231: {'lr': 0.0004999714278467551, 'samples': 428544, 'steps': 2231, 'loss/train': 2.4303903579711914} 01/28/2022 15:47:37 - INFO - codeparrot_training - Step 2232: {'lr': 0.0004999711799381181, 'samples': 428736, 'steps': 2232, 'loss/train': 2.9010202288627625} 01/28/2022 15:47:41 - INFO - codeparrot_training - Step 2233: {'lr': 0.000499970930958684, 'samples': 428928, 'steps': 2233, 'loss/train': 0.5053063631057739} 01/28/2022 15:47:45 - INFO - codeparrot_training - Step 2234: {'lr': 0.0004999706809084538, 'samples': 429120, 'steps': 2234, 'loss/train': 2.1356948018074036} 01/28/2022 15:47:49 - INFO - codeparrot_training - Step 2235: {'lr': 0.0004999704297874287, 'samples': 429312, 'steps': 2235, 'loss/train': 2.296496093273163} 01/28/2022 15:47:56 - INFO - codeparrot_training - Step 2236: {'lr': 0.0004999701775956095, 'samples': 429504, 'steps': 2236, 'loss/train': 2.2119582295417786} 01/28/2022 15:48:00 - INFO - codeparrot_training - Step 2237: {'lr': 0.0004999699243329975, 'samples': 429696, 'steps': 2237, 'loss/train': 1.9724475145339966} 01/28/2022 15:48:04 - INFO - codeparrot_training - Step 2238: {'lr': 0.0004999696699995937, 'samples': 429888, 'steps': 2238, 'loss/train': 1.6753113269805908} 01/28/2022 15:48:08 - INFO - codeparrot_training - Step 2239: {'lr': 0.0004999694145953992, 'samples': 430080, 'steps': 2239, 'loss/train': 1.6820333003997803} 01/28/2022 15:48:13 - INFO - codeparrot_training - Step 2240: {'lr': 0.0004999691581204152, 'samples': 430272, 'steps': 2240, 'loss/train': 2.1714248657226562} 01/28/2022 15:48:17 - INFO - codeparrot_training - Step 2241: {'lr': 0.0004999689005746426, 'samples': 430464, 'steps': 2241, 'loss/train': 1.7601848244667053} 01/28/2022 15:48:23 - INFO - codeparrot_training - Step 2242: {'lr': 0.0004999686419580827, 'samples': 430656, 'steps': 2242, 'loss/train': 2.3687251210212708} 01/28/2022 15:48:27 - INFO - codeparrot_training - Step 2243: {'lr': 0.0004999683822707364, 'samples': 430848, 'steps': 2243, 'loss/train': 2.348948299884796} 01/28/2022 15:48:31 - INFO - codeparrot_training - Step 2244: {'lr': 0.0004999681215126049, 'samples': 431040, 'steps': 2244, 'loss/train': 1.899137556552887} 01/28/2022 15:48:36 - INFO - codeparrot_training - Step 2245: {'lr': 0.0004999678596836894, 'samples': 431232, 'steps': 2245, 'loss/train': 1.8764528632164001} 01/28/2022 15:48:40 - INFO - codeparrot_training - Step 2246: {'lr': 0.000499967596783991, 'samples': 431424, 'steps': 2246, 'loss/train': 0.9332020282745361} 01/28/2022 15:48:45 - INFO - codeparrot_training - Step 2247: {'lr': 0.0004999673328135107, 'samples': 431616, 'steps': 2247, 'loss/train': 1.4825587570667267} 01/28/2022 15:48:49 - INFO - codeparrot_training - Step 2248: {'lr': 0.0004999670677722498, 'samples': 431808, 'steps': 2248, 'loss/train': 1.6424530148506165} 01/28/2022 15:48:54 - INFO - codeparrot_training - Step 2249: {'lr': 0.0004999668016602094, 'samples': 432000, 'steps': 2249, 'loss/train': 2.4636935591697693} 01/28/2022 15:48:58 - INFO - codeparrot_training - Step 2250: {'lr': 0.0004999665344773905, 'samples': 432192, 'steps': 2250, 'loss/train': 2.066141366958618} 01/28/2022 15:49:02 - INFO - codeparrot_training - Step 2251: {'lr': 0.0004999662662237943, 'samples': 432384, 'steps': 2251, 'loss/train': 1.3799608647823334} 01/28/2022 15:49:08 - INFO - codeparrot_training - Step 2252: {'lr': 0.0004999659968994221, 'samples': 432576, 'steps': 2252, 'loss/train': 2.1375160217285156} 01/28/2022 15:49:12 - INFO - codeparrot_training - Step 2253: {'lr': 0.0004999657265042748, 'samples': 432768, 'steps': 2253, 'loss/train': 1.9579707384109497} 01/28/2022 15:49:17 - INFO - codeparrot_training - Step 2254: {'lr': 0.0004999654550383539, 'samples': 432960, 'steps': 2254, 'loss/train': 1.4252886772155762} 01/28/2022 15:49:21 - INFO - codeparrot_training - Step 2255: {'lr': 0.0004999651825016603, 'samples': 433152, 'steps': 2255, 'loss/train': 2.338835835456848} 01/28/2022 15:49:25 - INFO - codeparrot_training - Step 2256: {'lr': 0.0004999649088941951, 'samples': 433344, 'steps': 2256, 'loss/train': 2.726453483104706} 01/28/2022 15:49:31 - INFO - codeparrot_training - Step 2257: {'lr': 0.0004999646342159597, 'samples': 433536, 'steps': 2257, 'loss/train': 1.8097051978111267} 01/28/2022 15:49:35 - INFO - codeparrot_training - Step 2258: {'lr': 0.0004999643584669552, 'samples': 433728, 'steps': 2258, 'loss/train': 2.3766061663627625} 01/28/2022 15:49:40 - INFO - codeparrot_training - Step 2259: {'lr': 0.0004999640816471827, 'samples': 433920, 'steps': 2259, 'loss/train': 2.2201666831970215} 01/28/2022 15:49:44 - INFO - codeparrot_training - Step 2260: {'lr': 0.0004999638037566436, 'samples': 434112, 'steps': 2260, 'loss/train': 2.308919906616211} 01/28/2022 15:49:48 - INFO - codeparrot_training - Step 2261: {'lr': 0.0004999635247953387, 'samples': 434304, 'steps': 2261, 'loss/train': 2.0361803770065308} 01/28/2022 15:49:53 - INFO - codeparrot_training - Step 2262: {'lr': 0.0004999632447632696, 'samples': 434496, 'steps': 2262, 'loss/train': 1.7633916735649109} 01/28/2022 15:49:58 - INFO - codeparrot_training - Step 2263: {'lr': 0.0004999629636604372, 'samples': 434688, 'steps': 2263, 'loss/train': 2.3358097672462463} 01/28/2022 15:50:02 - INFO - codeparrot_training - Step 2264: {'lr': 0.0004999626814868429, 'samples': 434880, 'steps': 2264, 'loss/train': 1.763424575328827} 01/28/2022 15:50:06 - INFO - codeparrot_training - Step 2265: {'lr': 0.0004999623982424879, 'samples': 435072, 'steps': 2265, 'loss/train': 2.6878640055656433} 01/28/2022 15:50:11 - INFO - codeparrot_training - Step 2266: {'lr': 0.0004999621139273733, 'samples': 435264, 'steps': 2266, 'loss/train': 2.1830504536628723} 01/28/2022 15:50:16 - INFO - codeparrot_training - Step 2267: {'lr': 0.0004999618285415004, 'samples': 435456, 'steps': 2267, 'loss/train': 2.2113736867904663} 01/28/2022 15:50:20 - INFO - codeparrot_training - Step 2268: {'lr': 0.0004999615420848704, 'samples': 435648, 'steps': 2268, 'loss/train': 2.0343344807624817} 01/28/2022 15:50:24 - INFO - codeparrot_training - Step 2269: {'lr': 0.0004999612545574845, 'samples': 435840, 'steps': 2269, 'loss/train': 2.5675150752067566} 01/28/2022 15:50:29 - INFO - codeparrot_training - Step 2270: {'lr': 0.000499960965959344, 'samples': 436032, 'steps': 2270, 'loss/train': 2.134210467338562} 01/28/2022 15:50:33 - INFO - codeparrot_training - Step 2271: {'lr': 0.0004999606762904501, 'samples': 436224, 'steps': 2271, 'loss/train': 2.449031174182892} 01/28/2022 15:50:39 - INFO - codeparrot_training - Step 2272: {'lr': 0.000499960385550804, 'samples': 436416, 'steps': 2272, 'loss/train': 2.0309289693832397} 01/28/2022 15:50:43 - INFO - codeparrot_training - Step 2273: {'lr': 0.000499960093740407, 'samples': 436608, 'steps': 2273, 'loss/train': 2.27591872215271} 01/28/2022 15:50:47 - INFO - codeparrot_training - Step 2274: {'lr': 0.0004999598008592603, 'samples': 436800, 'steps': 2274, 'loss/train': 2.191078841686249} 01/28/2022 15:50:52 - INFO - codeparrot_training - Step 2275: {'lr': 0.0004999595069073653, 'samples': 436992, 'steps': 2275, 'loss/train': 2.7314443588256836} 01/28/2022 15:50:56 - INFO - codeparrot_training - Step 2276: {'lr': 0.0004999592118847229, 'samples': 437184, 'steps': 2276, 'loss/train': 1.916496455669403} 01/28/2022 15:51:00 - INFO - codeparrot_training - Step 2277: {'lr': 0.0004999589157913348, 'samples': 437376, 'steps': 2277, 'loss/train': 1.6970758438110352} 01/28/2022 15:51:06 - INFO - codeparrot_training - Step 2278: {'lr': 0.0004999586186272021, 'samples': 437568, 'steps': 2278, 'loss/train': 2.351752996444702} 01/28/2022 15:51:10 - INFO - codeparrot_training - Step 2279: {'lr': 0.000499958320392326, 'samples': 437760, 'steps': 2279, 'loss/train': 1.455099195241928} 01/28/2022 15:51:15 - INFO - codeparrot_training - Step 2280: {'lr': 0.0004999580210867077, 'samples': 437952, 'steps': 2280, 'loss/train': 1.4883931875228882} 01/28/2022 15:51:19 - INFO - codeparrot_training - Step 2281: {'lr': 0.0004999577207103487, 'samples': 438144, 'steps': 2281, 'loss/train': 0.7059273272752762} 01/28/2022 15:51:24 - INFO - codeparrot_training - Step 2282: {'lr': 0.0004999574192632502, 'samples': 438336, 'steps': 2282, 'loss/train': 2.4496334195137024} 01/28/2022 15:51:28 - INFO - codeparrot_training - Step 2283: {'lr': 0.0004999571167454135, 'samples': 438528, 'steps': 2283, 'loss/train': 2.089704215526581} 01/28/2022 15:51:33 - INFO - codeparrot_training - Step 2284: {'lr': 0.0004999568131568399, 'samples': 438720, 'steps': 2284, 'loss/train': 2.1186050176620483} 01/28/2022 15:51:37 - INFO - codeparrot_training - Step 2285: {'lr': 0.0004999565084975306, 'samples': 438912, 'steps': 2285, 'loss/train': 2.8124395608901978} 01/28/2022 15:51:41 - INFO - codeparrot_training - Step 2286: {'lr': 0.0004999562027674871, 'samples': 439104, 'steps': 2286, 'loss/train': 1.9156857132911682} 01/28/2022 15:51:47 - INFO - codeparrot_training - Step 2287: {'lr': 0.0004999558959667105, 'samples': 439296, 'steps': 2287, 'loss/train': 2.63397216796875} 01/28/2022 15:51:51 - INFO - codeparrot_training - Step 2288: {'lr': 0.0004999555880952023, 'samples': 439488, 'steps': 2288, 'loss/train': 2.040173828601837} 01/28/2022 15:51:56 - INFO - codeparrot_training - Step 2289: {'lr': 0.0004999552791529637, 'samples': 439680, 'steps': 2289, 'loss/train': 2.0011605620384216} 01/28/2022 15:52:00 - INFO - codeparrot_training - Step 2290: {'lr': 0.000499954969139996, 'samples': 439872, 'steps': 2290, 'loss/train': 2.8677581548690796} 01/28/2022 15:52:04 - INFO - codeparrot_training - Step 2291: {'lr': 0.0004999546580563006, 'samples': 440064, 'steps': 2291, 'loss/train': 2.079001486301422} 01/28/2022 15:52:09 - INFO - codeparrot_training - Step 2292: {'lr': 0.0004999543459018788, 'samples': 440256, 'steps': 2292, 'loss/train': 2.4769155979156494} 01/28/2022 15:52:14 - INFO - codeparrot_training - Step 2293: {'lr': 0.000499954032676732, 'samples': 440448, 'steps': 2293, 'loss/train': 3.1158989667892456} 01/28/2022 15:52:18 - INFO - codeparrot_training - Step 2294: {'lr': 0.0004999537183808614, 'samples': 440640, 'steps': 2294, 'loss/train': 2.104028820991516} 01/28/2022 15:52:22 - INFO - codeparrot_training - Step 2295: {'lr': 0.0004999534030142686, 'samples': 440832, 'steps': 2295, 'loss/train': 2.349847733974457} 01/28/2022 15:52:26 - INFO - codeparrot_training - Step 2296: {'lr': 0.0004999530865769547, 'samples': 441024, 'steps': 2296, 'loss/train': 2.419960677623749} 01/28/2022 15:52:32 - INFO - codeparrot_training - Step 2297: {'lr': 0.0004999527690689212, 'samples': 441216, 'steps': 2297, 'loss/train': 1.2991260588169098} 01/28/2022 15:52:36 - INFO - codeparrot_training - Step 2298: {'lr': 0.0004999524504901694, 'samples': 441408, 'steps': 2298, 'loss/train': 1.2792489230632782} 01/28/2022 15:52:40 - INFO - codeparrot_training - Step 2299: {'lr': 0.0004999521308407006, 'samples': 441600, 'steps': 2299, 'loss/train': 1.7332904934883118} 01/28/2022 15:52:44 - INFO - codeparrot_training - Step 2300: {'lr': 0.0004999518101205162, 'samples': 441792, 'steps': 2300, 'loss/train': 2.6804586052894592} 01/28/2022 15:52:49 - INFO - codeparrot_training - Step 2301: {'lr': 0.0004999514883296176, 'samples': 441984, 'steps': 2301, 'loss/train': 2.55648010969162} 01/28/2022 15:52:55 - INFO - codeparrot_training - Step 2302: {'lr': 0.0004999511654680064, 'samples': 442176, 'steps': 2302, 'loss/train': 2.8833478689193726} 01/28/2022 15:52:59 - INFO - codeparrot_training - Step 2303: {'lr': 0.0004999508415356836, 'samples': 442368, 'steps': 2303, 'loss/train': 1.2187355160713196} 01/28/2022 15:53:03 - INFO - codeparrot_training - Step 2304: {'lr': 0.0004999505165326509, 'samples': 442560, 'steps': 2304, 'loss/train': 0.36661794036626816} 01/28/2022 15:53:07 - INFO - codeparrot_training - Step 2305: {'lr': 0.0004999501904589095, 'samples': 442752, 'steps': 2305, 'loss/train': 2.1332216262817383} 01/28/2022 15:53:12 - INFO - codeparrot_training - Step 2306: {'lr': 0.0004999498633144608, 'samples': 442944, 'steps': 2306, 'loss/train': 1.7393141984939575} 01/28/2022 15:53:17 - INFO - codeparrot_training - Step 2307: {'lr': 0.0004999495350993062, 'samples': 443136, 'steps': 2307, 'loss/train': 2.513085901737213} 01/28/2022 15:53:21 - INFO - codeparrot_training - Step 2308: {'lr': 0.0004999492058134473, 'samples': 443328, 'steps': 2308, 'loss/train': 0.38183802366256714} 01/28/2022 15:53:26 - INFO - codeparrot_training - Step 2309: {'lr': 0.0004999488754568853, 'samples': 443520, 'steps': 2309, 'loss/train': 2.692461133003235} 01/28/2022 15:53:30 - INFO - codeparrot_training - Step 2310: {'lr': 0.0004999485440296216, 'samples': 443712, 'steps': 2310, 'loss/train': 2.1325364112854004} 01/28/2022 15:53:34 - INFO - codeparrot_training - Step 2311: {'lr': 0.0004999482115316579, 'samples': 443904, 'steps': 2311, 'loss/train': 2.1976595520973206} 01/28/2022 15:53:39 - INFO - codeparrot_training - Step 2312: {'lr': 0.0004999478779629953, 'samples': 444096, 'steps': 2312, 'loss/train': 1.8088667392730713} 01/28/2022 15:53:44 - INFO - codeparrot_training - Step 2313: {'lr': 0.0004999475433236354, 'samples': 444288, 'steps': 2313, 'loss/train': 2.82437002658844} 01/28/2022 15:53:48 - INFO - codeparrot_training - Step 2314: {'lr': 0.0004999472076135796, 'samples': 444480, 'steps': 2314, 'loss/train': 2.041930317878723} 01/28/2022 15:53:52 - INFO - codeparrot_training - Step 2315: {'lr': 0.0004999468708328293, 'samples': 444672, 'steps': 2315, 'loss/train': 1.39400714635849} 01/28/2022 15:53:56 - INFO - codeparrot_training - Step 2316: {'lr': 0.0004999465329813859, 'samples': 444864, 'steps': 2316, 'loss/train': 1.7947036027908325} 01/28/2022 15:54:01 - INFO - codeparrot_training - Step 2317: {'lr': 0.000499946194059251, 'samples': 445056, 'steps': 2317, 'loss/train': 2.135491132736206} 01/28/2022 15:54:06 - INFO - codeparrot_training - Step 2318: {'lr': 0.000499945854066426, 'samples': 445248, 'steps': 2318, 'loss/train': 2.046349346637726} 01/28/2022 15:54:10 - INFO - codeparrot_training - Step 2319: {'lr': 0.0004999455130029123, 'samples': 445440, 'steps': 2319, 'loss/train': 1.306700885295868} 01/28/2022 15:54:15 - INFO - codeparrot_training - Step 2320: {'lr': 0.0004999451708687113, 'samples': 445632, 'steps': 2320, 'loss/train': 1.8659525513648987} 01/28/2022 15:54:19 - INFO - codeparrot_training - Step 2321: {'lr': 0.0004999448276638247, 'samples': 445824, 'steps': 2321, 'loss/train': 0.33087144792079926} 01/28/2022 15:54:23 - INFO - codeparrot_training - Step 2322: {'lr': 0.0004999444833882538, 'samples': 446016, 'steps': 2322, 'loss/train': 1.9147224426269531} 01/28/2022 15:54:30 - INFO - codeparrot_training - Step 2323: {'lr': 0.000499944138042, 'samples': 446208, 'steps': 2323, 'loss/train': 2.6928239464759827} 01/28/2022 15:54:34 - INFO - codeparrot_training - Step 2324: {'lr': 0.000499943791625065, 'samples': 446400, 'steps': 2324, 'loss/train': 2.117751896381378} 01/28/2022 15:54:38 - INFO - codeparrot_training - Step 2325: {'lr': 0.0004999434441374501, 'samples': 446592, 'steps': 2325, 'loss/train': 1.635587990283966} 01/28/2022 15:54:42 - INFO - codeparrot_training - Step 2326: {'lr': 0.0004999430955791569, 'samples': 446784, 'steps': 2326, 'loss/train': 1.8259409666061401} 01/28/2022 15:54:48 - INFO - codeparrot_training - Step 2327: {'lr': 0.0004999427459501868, 'samples': 446976, 'steps': 2327, 'loss/train': 2.4233651161193848} 01/28/2022 15:54:52 - INFO - codeparrot_training - Step 2328: {'lr': 0.0004999423952505414, 'samples': 447168, 'steps': 2328, 'loss/train': 1.7904921770095825} 01/28/2022 15:54:57 - INFO - codeparrot_training - Step 2329: {'lr': 0.000499942043480222, 'samples': 447360, 'steps': 2329, 'loss/train': 1.5889288187026978} 01/28/2022 15:55:01 - INFO - codeparrot_training - Step 2330: {'lr': 0.0004999416906392303, 'samples': 447552, 'steps': 2330, 'loss/train': 2.737762749195099} 01/28/2022 15:55:05 - INFO - codeparrot_training - Step 2331: {'lr': 0.0004999413367275678, 'samples': 447744, 'steps': 2331, 'loss/train': 1.8589024543762207} 01/28/2022 15:55:09 - INFO - codeparrot_training - Step 2332: {'lr': 0.000499940981745236, 'samples': 447936, 'steps': 2332, 'loss/train': 2.1758771538734436} 01/28/2022 15:55:16 - INFO - codeparrot_training - Step 2333: {'lr': 0.0004999406256922365, 'samples': 448128, 'steps': 2333, 'loss/train': 2.118940830230713} 01/28/2022 15:55:20 - INFO - codeparrot_training - Step 2334: {'lr': 0.0004999402685685705, 'samples': 448320, 'steps': 2334, 'loss/train': 1.0319501459598541} 01/28/2022 15:55:24 - INFO - codeparrot_training - Step 2335: {'lr': 0.0004999399103742399, 'samples': 448512, 'steps': 2335, 'loss/train': 2.873024582862854} 01/28/2022 15:55:29 - INFO - codeparrot_training - Step 2336: {'lr': 0.000499939551109246, 'samples': 448704, 'steps': 2336, 'loss/train': 2.3813620805740356} 01/28/2022 15:55:33 - INFO - codeparrot_training - Step 2337: {'lr': 0.0004999391907735905, 'samples': 448896, 'steps': 2337, 'loss/train': 3.573099732398987} 01/28/2022 15:55:39 - INFO - codeparrot_training - Step 2338: {'lr': 0.0004999388293672748, 'samples': 449088, 'steps': 2338, 'loss/train': 1.4197577834129333} 01/28/2022 15:55:43 - INFO - codeparrot_training - Step 2339: {'lr': 0.0004999384668903006, 'samples': 449280, 'steps': 2339, 'loss/train': 1.7384478449821472} 01/28/2022 15:55:47 - INFO - codeparrot_training - Step 2340: {'lr': 0.0004999381033426693, 'samples': 449472, 'steps': 2340, 'loss/train': 2.4953293204307556} 01/28/2022 15:55:51 - INFO - codeparrot_training - Step 2341: {'lr': 0.0004999377387243827, 'samples': 449664, 'steps': 2341, 'loss/train': 1.5060916543006897} 01/28/2022 15:55:58 - INFO - codeparrot_training - Step 2342: {'lr': 0.0004999373730354419, 'samples': 449856, 'steps': 2342, 'loss/train': 2.8596308827400208} 01/28/2022 15:56:02 - INFO - codeparrot_training - Step 2343: {'lr': 0.0004999370062758491, 'samples': 450048, 'steps': 2343, 'loss/train': 1.7292308807373047} 01/28/2022 15:56:06 - INFO - codeparrot_training - Step 2344: {'lr': 0.0004999366384456052, 'samples': 450240, 'steps': 2344, 'loss/train': 2.0828661918640137} 01/28/2022 15:56:10 - INFO - codeparrot_training - Step 2345: {'lr': 0.0004999362695447123, 'samples': 450432, 'steps': 2345, 'loss/train': 0.922847181558609} 01/28/2022 15:56:15 - INFO - codeparrot_training - Step 2346: {'lr': 0.0004999358995731718, 'samples': 450624, 'steps': 2346, 'loss/train': 2.4418834447860718} 01/28/2022 15:56:19 - INFO - codeparrot_training - Step 2347: {'lr': 0.0004999355285309851, 'samples': 450816, 'steps': 2347, 'loss/train': 2.429480016231537} 01/28/2022 15:56:25 - INFO - codeparrot_training - Step 2348: {'lr': 0.0004999351564181541, 'samples': 451008, 'steps': 2348, 'loss/train': 1.7945460677146912} 01/28/2022 15:56:29 - INFO - codeparrot_training - Step 2349: {'lr': 0.0004999347832346802, 'samples': 451200, 'steps': 2349, 'loss/train': 2.6004475951194763} 01/28/2022 15:56:34 - INFO - codeparrot_training - Step 2350: {'lr': 0.0004999344089805651, 'samples': 451392, 'steps': 2350, 'loss/train': 1.714375913143158} 01/28/2022 15:56:38 - INFO - codeparrot_training - Step 2351: {'lr': 0.0004999340336558104, 'samples': 451584, 'steps': 2351, 'loss/train': 2.240787148475647} 01/28/2022 15:56:42 - INFO - codeparrot_training - Step 2352: {'lr': 0.0004999336572604175, 'samples': 451776, 'steps': 2352, 'loss/train': 2.0342647433280945} 01/28/2022 15:56:47 - INFO - codeparrot_training - Step 2353: {'lr': 0.0004999332797943883, 'samples': 451968, 'steps': 2353, 'loss/train': 1.8811023831367493} 01/28/2022 15:56:52 - INFO - codeparrot_training - Step 2354: {'lr': 0.0004999329012577243, 'samples': 452160, 'steps': 2354, 'loss/train': 1.0172638893127441} 01/28/2022 15:56:56 - INFO - codeparrot_training - Step 2355: {'lr': 0.000499932521650427, 'samples': 452352, 'steps': 2355, 'loss/train': 0.4616847038269043} 01/28/2022 15:57:00 - INFO - codeparrot_training - Step 2356: {'lr': 0.0004999321409724982, 'samples': 452544, 'steps': 2356, 'loss/train': 2.350193202495575} 01/28/2022 15:57:04 - INFO - codeparrot_training - Step 2357: {'lr': 0.0004999317592239395, 'samples': 452736, 'steps': 2357, 'loss/train': 1.610669732093811} 01/28/2022 15:57:10 - INFO - codeparrot_training - Step 2358: {'lr': 0.0004999313764047525, 'samples': 452928, 'steps': 2358, 'loss/train': 2.5830954909324646} 01/28/2022 15:57:15 - INFO - codeparrot_training - Step 2359: {'lr': 0.0004999309925149388, 'samples': 453120, 'steps': 2359, 'loss/train': 1.9938633441925049} 01/28/2022 15:57:19 - INFO - codeparrot_training - Step 2360: {'lr': 0.0004999306075545002, 'samples': 453312, 'steps': 2360, 'loss/train': 3.436068534851074} 01/28/2022 15:57:23 - INFO - codeparrot_training - Step 2361: {'lr': 0.0004999302215234381, 'samples': 453504, 'steps': 2361, 'loss/train': 2.225751221179962} 01/28/2022 15:57:27 - INFO - codeparrot_training - Step 2362: {'lr': 0.0004999298344217543, 'samples': 453696, 'steps': 2362, 'loss/train': 1.6788876056671143} 01/28/2022 15:57:33 - INFO - codeparrot_training - Step 2363: {'lr': 0.0004999294462494506, 'samples': 453888, 'steps': 2363, 'loss/train': 2.940725326538086} 01/28/2022 15:57:37 - INFO - codeparrot_training - Step 2364: {'lr': 0.0004999290570065284, 'samples': 454080, 'steps': 2364, 'loss/train': 2.611275315284729} 01/28/2022 15:57:41 - INFO - codeparrot_training - Step 2365: {'lr': 0.0004999286666929895, 'samples': 454272, 'steps': 2365, 'loss/train': 1.0834407806396484} 01/28/2022 15:57:45 - INFO - codeparrot_training - Step 2366: {'lr': 0.0004999282753088356, 'samples': 454464, 'steps': 2366, 'loss/train': 1.6226742267608643} 01/28/2022 15:57:50 - INFO - codeparrot_training - Step 2367: {'lr': 0.0004999278828540682, 'samples': 454656, 'steps': 2367, 'loss/train': 1.3196690082550049} 01/28/2022 15:57:55 - INFO - codeparrot_training - Step 2368: {'lr': 0.0004999274893286893, 'samples': 454848, 'steps': 2368, 'loss/train': 2.681888222694397} 01/28/2022 15:57:59 - INFO - codeparrot_training - Step 2369: {'lr': 0.0004999270947327003, 'samples': 455040, 'steps': 2369, 'loss/train': 2.308399200439453} 01/28/2022 15:58:04 - INFO - codeparrot_training - Step 2370: {'lr': 0.0004999266990661029, 'samples': 455232, 'steps': 2370, 'loss/train': 1.526474118232727} 01/28/2022 15:58:08 - INFO - codeparrot_training - Step 2371: {'lr': 0.0004999263023288989, 'samples': 455424, 'steps': 2371, 'loss/train': 2.5087591409683228} 01/28/2022 15:58:12 - INFO - codeparrot_training - Step 2372: {'lr': 0.0004999259045210901, 'samples': 455616, 'steps': 2372, 'loss/train': 2.0588690042495728} 01/28/2022 15:58:17 - INFO - codeparrot_training - Step 2373: {'lr': 0.000499925505642678, 'samples': 455808, 'steps': 2373, 'loss/train': 0.629629522562027} 01/28/2022 15:58:22 - INFO - codeparrot_training - Step 2374: {'lr': 0.0004999251056936645, 'samples': 456000, 'steps': 2374, 'loss/train': 2.282431483268738} 01/28/2022 15:58:26 - INFO - codeparrot_training - Step 2375: {'lr': 0.000499924704674051, 'samples': 456192, 'steps': 2375, 'loss/train': 2.6437357664108276} 01/28/2022 15:58:30 - INFO - codeparrot_training - Step 2376: {'lr': 0.0004999243025838396, 'samples': 456384, 'steps': 2376, 'loss/train': 1.5176909565925598} 01/28/2022 15:58:34 - INFO - codeparrot_training - Step 2377: {'lr': 0.0004999238994230318, 'samples': 456576, 'steps': 2377, 'loss/train': 2.3156336545944214} 01/28/2022 15:58:40 - INFO - codeparrot_training - Step 2378: {'lr': 0.0004999234951916293, 'samples': 456768, 'steps': 2378, 'loss/train': 1.9887304902076721} 01/28/2022 15:58:45 - INFO - codeparrot_training - Step 2379: {'lr': 0.0004999230898896341, 'samples': 456960, 'steps': 2379, 'loss/train': 1.5197601914405823} 01/28/2022 15:58:49 - INFO - codeparrot_training - Step 2380: {'lr': 0.0004999226835170476, 'samples': 457152, 'steps': 2380, 'loss/train': 1.3806594014167786} 01/28/2022 15:58:53 - INFO - codeparrot_training - Step 2381: {'lr': 0.0004999222760738717, 'samples': 457344, 'steps': 2381, 'loss/train': 2.2238497138023376} 01/28/2022 15:58:57 - INFO - codeparrot_training - Step 2382: {'lr': 0.0004999218675601081, 'samples': 457536, 'steps': 2382, 'loss/train': 2.3969414234161377} 01/28/2022 15:59:02 - INFO - codeparrot_training - Step 2383: {'lr': 0.0004999214579757586, 'samples': 457728, 'steps': 2383, 'loss/train': 1.604950726032257} 01/28/2022 15:59:07 - INFO - codeparrot_training - Step 2384: {'lr': 0.000499921047320825, 'samples': 457920, 'steps': 2384, 'loss/train': 1.7948859930038452} 01/28/2022 15:59:11 - INFO - codeparrot_training - Step 2385: {'lr': 0.000499920635595309, 'samples': 458112, 'steps': 2385, 'loss/train': 2.3057801127433777} 01/28/2022 15:59:16 - INFO - codeparrot_training - Step 2386: {'lr': 0.0004999202227992122, 'samples': 458304, 'steps': 2386, 'loss/train': 2.298172652721405} 01/28/2022 15:59:20 - INFO - codeparrot_training - Step 2387: {'lr': 0.0004999198089325367, 'samples': 458496, 'steps': 2387, 'loss/train': 1.748381495475769} 01/28/2022 15:59:25 - INFO - codeparrot_training - Step 2388: {'lr': 0.0004999193939952839, 'samples': 458688, 'steps': 2388, 'loss/train': 1.9492809176445007} 01/28/2022 15:59:30 - INFO - codeparrot_training - Step 2389: {'lr': 0.000499918977987456, 'samples': 458880, 'steps': 2389, 'loss/train': 1.992385447025299} 01/28/2022 15:59:34 - INFO - codeparrot_training - Step 2390: {'lr': 0.0004999185609090544, 'samples': 459072, 'steps': 2390, 'loss/train': 1.9693048596382141} 01/28/2022 15:59:38 - INFO - codeparrot_training - Step 2391: {'lr': 0.0004999181427600811, 'samples': 459264, 'steps': 2391, 'loss/train': 1.4081924557685852} 01/28/2022 15:59:42 - INFO - codeparrot_training - Step 2392: {'lr': 0.0004999177235405378, 'samples': 459456, 'steps': 2392, 'loss/train': 1.4817489981651306} 01/28/2022 15:59:47 - INFO - codeparrot_training - Step 2393: {'lr': 0.0004999173032504264, 'samples': 459648, 'steps': 2393, 'loss/train': 2.1611409187316895} 01/28/2022 15:59:53 - INFO - codeparrot_training - Step 2394: {'lr': 0.0004999168818897486, 'samples': 459840, 'steps': 2394, 'loss/train': 1.5825703740119934} 01/28/2022 15:59:57 - INFO - codeparrot_training - Step 2395: {'lr': 0.0004999164594585062, 'samples': 460032, 'steps': 2395, 'loss/train': 1.7731636762619019} 01/28/2022 16:00:01 - INFO - codeparrot_training - Step 2396: {'lr': 0.0004999160359567011, 'samples': 460224, 'steps': 2396, 'loss/train': 1.8471254110336304} 01/28/2022 16:00:05 - INFO - codeparrot_training - Step 2397: {'lr': 0.000499915611384335, 'samples': 460416, 'steps': 2397, 'loss/train': 1.7797526121139526} 01/28/2022 16:00:10 - INFO - codeparrot_training - Step 2398: {'lr': 0.0004999151857414099, 'samples': 460608, 'steps': 2398, 'loss/train': 2.4713834524154663} 01/28/2022 16:00:15 - INFO - codeparrot_training - Step 2399: {'lr': 0.0004999147590279273, 'samples': 460800, 'steps': 2399, 'loss/train': 2.0800974369049072} 01/28/2022 16:00:19 - INFO - codeparrot_training - Step 2400: {'lr': 0.0004999143312438893, 'samples': 460992, 'steps': 2400, 'loss/train': 2.035479247570038} 01/28/2022 16:00:23 - INFO - codeparrot_training - Step 2401: {'lr': 0.0004999139023892978, 'samples': 461184, 'steps': 2401, 'loss/train': 1.9140037894248962} 01/28/2022 16:00:28 - INFO - codeparrot_training - Step 2402: {'lr': 0.0004999134724641543, 'samples': 461376, 'steps': 2402, 'loss/train': 1.7598658204078674} 01/28/2022 16:00:32 - INFO - codeparrot_training - Step 2403: {'lr': 0.000499913041468461, 'samples': 461568, 'steps': 2403, 'loss/train': 2.2717053294181824} 01/28/2022 16:00:38 - INFO - codeparrot_training - Step 2404: {'lr': 0.0004999126094022195, 'samples': 461760, 'steps': 2404, 'loss/train': 2.6517934799194336} 01/28/2022 16:00:42 - INFO - codeparrot_training - Step 2405: {'lr': 0.0004999121762654318, 'samples': 461952, 'steps': 2405, 'loss/train': 2.3832271099090576} 01/28/2022 16:00:46 - INFO - codeparrot_training - Step 2406: {'lr': 0.0004999117420580996, 'samples': 462144, 'steps': 2406, 'loss/train': 2.098949611186981} 01/28/2022 16:00:51 - INFO - codeparrot_training - Step 2407: {'lr': 0.0004999113067802249, 'samples': 462336, 'steps': 2407, 'loss/train': 2.3856489658355713} 01/28/2022 16:00:55 - INFO - codeparrot_training - Step 2408: {'lr': 0.0004999108704318095, 'samples': 462528, 'steps': 2408, 'loss/train': 2.4236698150634766} 01/28/2022 16:01:00 - INFO - codeparrot_training - Step 2409: {'lr': 0.0004999104330128553, 'samples': 462720, 'steps': 2409, 'loss/train': 0.7455107867717743} 01/28/2022 16:01:04 - INFO - codeparrot_training - Step 2410: {'lr': 0.0004999099945233641, 'samples': 462912, 'steps': 2410, 'loss/train': 1.0929268598556519} 01/28/2022 16:01:08 - INFO - codeparrot_training - Step 2411: {'lr': 0.000499909554963338, 'samples': 463104, 'steps': 2411, 'loss/train': 2.555756449699402} 01/28/2022 16:01:13 - INFO - codeparrot_training - Step 2412: {'lr': 0.0004999091143327786, 'samples': 463296, 'steps': 2412, 'loss/train': 2.820684492588043} 01/28/2022 16:01:18 - INFO - codeparrot_training - Step 2413: {'lr': 0.000499908672631688, 'samples': 463488, 'steps': 2413, 'loss/train': 1.487872838973999} 01/28/2022 16:01:22 - INFO - codeparrot_training - Step 2414: {'lr': 0.0004999082298600679, 'samples': 463680, 'steps': 2414, 'loss/train': 1.8498721718788147} 01/28/2022 16:01:26 - INFO - codeparrot_training - Step 2415: {'lr': 0.0004999077860179204, 'samples': 463872, 'steps': 2415, 'loss/train': 2.052374303340912} 01/28/2022 16:01:31 - INFO - codeparrot_training - Step 2416: {'lr': 0.0004999073411052472, 'samples': 464064, 'steps': 2416, 'loss/train': 1.9604035019874573} 01/28/2022 16:01:35 - INFO - codeparrot_training - Step 2417: {'lr': 0.0004999068951220503, 'samples': 464256, 'steps': 2417, 'loss/train': 2.252175986766815} 01/28/2022 16:01:41 - INFO - codeparrot_training - Step 2418: {'lr': 0.0004999064480683317, 'samples': 464448, 'steps': 2418, 'loss/train': 2.352702498435974} 01/28/2022 16:01:45 - INFO - codeparrot_training - Step 2419: {'lr': 0.0004999059999440932, 'samples': 464640, 'steps': 2419, 'loss/train': 2.0527787804603577} 01/28/2022 16:01:49 - INFO - codeparrot_training - Step 2420: {'lr': 0.0004999055507493368, 'samples': 464832, 'steps': 2420, 'loss/train': 2.8373976945877075} 01/28/2022 16:01:54 - INFO - codeparrot_training - Step 2421: {'lr': 0.0004999051004840642, 'samples': 465024, 'steps': 2421, 'loss/train': 1.841233491897583} 01/28/2022 16:01:58 - INFO - codeparrot_training - Step 2422: {'lr': 0.0004999046491482777, 'samples': 465216, 'steps': 2422, 'loss/train': 2.221679449081421} 01/28/2022 16:02:03 - INFO - codeparrot_training - Step 2423: {'lr': 0.000499904196741979, 'samples': 465408, 'steps': 2423, 'loss/train': 2.535125970840454} 01/28/2022 16:02:07 - INFO - codeparrot_training - Step 2424: {'lr': 0.00049990374326517, 'samples': 465600, 'steps': 2424, 'loss/train': 1.17001211643219} 01/28/2022 16:02:12 - INFO - codeparrot_training - Step 2425: {'lr': 0.0004999032887178527, 'samples': 465792, 'steps': 2425, 'loss/train': 1.9143206477165222} 01/28/2022 16:02:16 - INFO - codeparrot_training - Step 2426: {'lr': 0.000499902833100029, 'samples': 465984, 'steps': 2426, 'loss/train': 1.8044258952140808} 01/28/2022 16:02:20 - INFO - codeparrot_training - Step 2427: {'lr': 0.0004999023764117011, 'samples': 466176, 'steps': 2427, 'loss/train': 1.6101393699645996} 01/28/2022 16:02:25 - INFO - codeparrot_training - Step 2428: {'lr': 0.0004999019186528708, 'samples': 466368, 'steps': 2428, 'loss/train': 2.3665273189544678} 01/28/2022 16:02:30 - INFO - codeparrot_training - Step 2429: {'lr': 0.0004999014598235399, 'samples': 466560, 'steps': 2429, 'loss/train': 0.8307017683982849} 01/28/2022 16:02:34 - INFO - codeparrot_training - Step 2430: {'lr': 0.0004999009999237105, 'samples': 466752, 'steps': 2430, 'loss/train': 2.152595043182373} 01/28/2022 16:02:38 - INFO - codeparrot_training - Step 2431: {'lr': 0.0004999005389533846, 'samples': 466944, 'steps': 2431, 'loss/train': 1.4179012477397919} 01/28/2022 16:02:42 - INFO - codeparrot_training - Step 2432: {'lr': 0.0004999000769125642, 'samples': 467136, 'steps': 2432, 'loss/train': 2.606311619281769} 01/28/2022 16:02:48 - INFO - codeparrot_training - Step 2433: {'lr': 0.0004998996138012512, 'samples': 467328, 'steps': 2433, 'loss/train': 1.8619306683540344} 01/28/2022 16:02:52 - INFO - codeparrot_training - Step 2434: {'lr': 0.0004998991496194475, 'samples': 467520, 'steps': 2434, 'loss/train': 2.2635939717292786} 01/28/2022 16:02:56 - INFO - codeparrot_training - Step 2435: {'lr': 0.0004998986843671552, 'samples': 467712, 'steps': 2435, 'loss/train': 2.3075430393218994} 01/28/2022 16:03:00 - INFO - codeparrot_training - Step 2436: {'lr': 0.0004998982180443764, 'samples': 467904, 'steps': 2436, 'loss/train': 2.1955840587615967} 01/28/2022 16:03:05 - INFO - codeparrot_training - Step 2437: {'lr': 0.000499897750651113, 'samples': 468096, 'steps': 2437, 'loss/train': 2.463289439678192} 01/28/2022 16:03:11 - INFO - codeparrot_training - Step 2438: {'lr': 0.0004998972821873668, 'samples': 468288, 'steps': 2438, 'loss/train': 2.6337071657180786} 01/28/2022 16:03:15 - INFO - codeparrot_training - Step 2439: {'lr': 0.0004998968126531402, 'samples': 468480, 'steps': 2439, 'loss/train': 2.0138742327690125} 01/28/2022 16:03:19 - INFO - codeparrot_training - Step 2440: {'lr': 0.0004998963420484349, 'samples': 468672, 'steps': 2440, 'loss/train': 2.019286572933197} 01/28/2022 16:03:24 - INFO - codeparrot_training - Step 2441: {'lr': 0.0004998958703732532, 'samples': 468864, 'steps': 2441, 'loss/train': 1.6449578404426575} 01/28/2022 16:03:28 - INFO - codeparrot_training - Step 2442: {'lr': 0.0004998953976275966, 'samples': 469056, 'steps': 2442, 'loss/train': 2.6808364391326904} 01/28/2022 16:03:33 - INFO - codeparrot_training - Step 2443: {'lr': 0.0004998949238114677, 'samples': 469248, 'steps': 2443, 'loss/train': 2.2493841648101807} 01/28/2022 16:03:38 - INFO - codeparrot_training - Step 2444: {'lr': 0.0004998944489248683, 'samples': 469440, 'steps': 2444, 'loss/train': 0.7781091928482056} 01/28/2022 16:03:42 - INFO - codeparrot_training - Step 2445: {'lr': 0.0004998939729678004, 'samples': 469632, 'steps': 2445, 'loss/train': 2.3587140440940857} 01/28/2022 16:03:46 - INFO - codeparrot_training - Step 2446: {'lr': 0.000499893495940266, 'samples': 469824, 'steps': 2446, 'loss/train': 2.1057443618774414} 01/28/2022 16:03:51 - INFO - codeparrot_training - Step 2447: {'lr': 0.0004998930178422673, 'samples': 470016, 'steps': 2447, 'loss/train': 2.284052789211273} 01/28/2022 16:03:57 - INFO - codeparrot_training - Step 2448: {'lr': 0.0004998925386738062, 'samples': 470208, 'steps': 2448, 'loss/train': 1.653443455696106} 01/28/2022 16:04:01 - INFO - codeparrot_training - Step 2449: {'lr': 0.0004998920584348849, 'samples': 470400, 'steps': 2449, 'loss/train': 1.3282309770584106} 01/28/2022 16:04:05 - INFO - codeparrot_training - Step 2450: {'lr': 0.0004998915771255053, 'samples': 470592, 'steps': 2450, 'loss/train': 1.1620477437973022} 01/28/2022 16:04:10 - INFO - codeparrot_training - Step 2451: {'lr': 0.0004998910947456696, 'samples': 470784, 'steps': 2451, 'loss/train': 2.749558746814728} 01/28/2022 16:04:14 - INFO - codeparrot_training - Step 2452: {'lr': 0.0004998906112953797, 'samples': 470976, 'steps': 2452, 'loss/train': 1.7733744978904724} 01/28/2022 16:04:20 - INFO - codeparrot_training - Step 2453: {'lr': 0.0004998901267746379, 'samples': 471168, 'steps': 2453, 'loss/train': 2.5705320239067078} 01/28/2022 16:04:24 - INFO - codeparrot_training - Step 2454: {'lr': 0.0004998896411834461, 'samples': 471360, 'steps': 2454, 'loss/train': 1.1187976598739624} 01/28/2022 16:04:29 - INFO - codeparrot_training - Step 2455: {'lr': 0.0004998891545218063, 'samples': 471552, 'steps': 2455, 'loss/train': 1.1101910769939423} 01/28/2022 16:04:33 - INFO - codeparrot_training - Step 2456: {'lr': 0.0004998886667897209, 'samples': 471744, 'steps': 2456, 'loss/train': 2.2819098830223083} 01/28/2022 16:04:37 - INFO - codeparrot_training - Step 2457: {'lr': 0.0004998881779871917, 'samples': 471936, 'steps': 2457, 'loss/train': 2.2428372502326965} 01/28/2022 16:04:41 - INFO - codeparrot_training - Step 2458: {'lr': 0.0004998876881142208, 'samples': 472128, 'steps': 2458, 'loss/train': 2.4393287301063538} 01/28/2022 16:04:47 - INFO - codeparrot_training - Step 2459: {'lr': 0.0004998871971708106, 'samples': 472320, 'steps': 2459, 'loss/train': 2.1626989245414734} 01/28/2022 16:04:51 - INFO - codeparrot_training - Step 2460: {'lr': 0.0004998867051569627, 'samples': 472512, 'steps': 2460, 'loss/train': 2.894318103790283} 01/28/2022 16:04:55 - INFO - codeparrot_training - Step 2461: {'lr': 0.0004998862120726798, 'samples': 472704, 'steps': 2461, 'loss/train': 1.8908191323280334} 01/28/2022 16:05:00 - INFO - codeparrot_training - Step 2462: {'lr': 0.0004998857179179636, 'samples': 472896, 'steps': 2462, 'loss/train': 1.8702898621559143} 01/28/2022 16:05:04 - INFO - codeparrot_training - Step 2463: {'lr': 0.0004998852226928164, 'samples': 473088, 'steps': 2463, 'loss/train': 4.001647353172302} 01/28/2022 16:05:10 - INFO - codeparrot_training - Step 2464: {'lr': 0.0004998847263972401, 'samples': 473280, 'steps': 2464, 'loss/train': 2.9411728978157043} 01/28/2022 16:05:14 - INFO - codeparrot_training - Step 2465: {'lr': 0.0004998842290312371, 'samples': 473472, 'steps': 2465, 'loss/train': 1.4933822751045227} 01/28/2022 16:05:18 - INFO - codeparrot_training - Step 2466: {'lr': 0.0004998837305948094, 'samples': 473664, 'steps': 2466, 'loss/train': 2.5456833243370056} 01/28/2022 16:05:23 - INFO - codeparrot_training - Step 2467: {'lr': 0.0004998832310879591, 'samples': 473856, 'steps': 2467, 'loss/train': 0.14495792984962463} 01/28/2022 16:05:27 - INFO - codeparrot_training - Step 2468: {'lr': 0.0004998827305106884, 'samples': 474048, 'steps': 2468, 'loss/train': 1.7202283143997192} 01/28/2022 16:05:32 - INFO - codeparrot_training - Step 2469: {'lr': 0.0004998822288629995, 'samples': 474240, 'steps': 2469, 'loss/train': 2.1604841351509094} 01/28/2022 16:05:36 - INFO - codeparrot_training - Step 2470: {'lr': 0.0004998817261448943, 'samples': 474432, 'steps': 2470, 'loss/train': 2.3022301197052} 01/28/2022 16:05:41 - INFO - codeparrot_training - Step 2471: {'lr': 0.0004998812223563754, 'samples': 474624, 'steps': 2471, 'loss/train': 2.331918776035309} 01/28/2022 16:05:45 - INFO - codeparrot_training - Step 2472: {'lr': 0.0004998807174974445, 'samples': 474816, 'steps': 2472, 'loss/train': 2.351798415184021} 01/28/2022 16:05:49 - INFO - codeparrot_training - Step 2473: {'lr': 0.0004998802115681039, 'samples': 475008, 'steps': 2473, 'loss/train': 0.5817186534404755} 01/28/2022 16:05:54 - INFO - codeparrot_training - Step 2474: {'lr': 0.000499879704568356, 'samples': 475200, 'steps': 2474, 'loss/train': 2.039862871170044} 01/28/2022 16:05:59 - INFO - codeparrot_training - Step 2475: {'lr': 0.0004998791964982026, 'samples': 475392, 'steps': 2475, 'loss/train': 2.310499906539917} 01/28/2022 16:06:03 - INFO - codeparrot_training - Step 2476: {'lr': 0.0004998786873576462, 'samples': 475584, 'steps': 2476, 'loss/train': 2.0561342239379883} 01/28/2022 16:06:07 - INFO - codeparrot_training - Step 2477: {'lr': 0.0004998781771466889, 'samples': 475776, 'steps': 2477, 'loss/train': 1.7828237414360046} 01/28/2022 16:06:11 - INFO - codeparrot_training - Step 2478: {'lr': 0.0004998776658653327, 'samples': 475968, 'steps': 2478, 'loss/train': 1.31217160820961} 01/28/2022 16:06:17 - INFO - codeparrot_training - Step 2479: {'lr': 0.00049987715351358, 'samples': 476160, 'steps': 2479, 'loss/train': 1.4749357402324677} 01/28/2022 16:06:22 - INFO - codeparrot_training - Step 2480: {'lr': 0.0004998766400914329, 'samples': 476352, 'steps': 2480, 'loss/train': 2.020177960395813} 01/28/2022 16:06:26 - INFO - codeparrot_training - Step 2481: {'lr': 0.0004998761255988936, 'samples': 476544, 'steps': 2481, 'loss/train': 0.8003562390804291} 01/28/2022 16:06:30 - INFO - codeparrot_training - Step 2482: {'lr': 0.0004998756100359643, 'samples': 476736, 'steps': 2482, 'loss/train': 1.3809986114501953} 01/28/2022 16:06:34 - INFO - codeparrot_training - Step 2483: {'lr': 0.0004998750934026474, 'samples': 476928, 'steps': 2483, 'loss/train': 1.9600444436073303} 01/28/2022 16:06:40 - INFO - codeparrot_training - Step 2484: {'lr': 0.0004998745756989448, 'samples': 477120, 'steps': 2484, 'loss/train': 1.9242598414421082} 01/28/2022 16:06:44 - INFO - codeparrot_training - Step 2485: {'lr': 0.0004998740569248588, 'samples': 477312, 'steps': 2485, 'loss/train': 2.317581295967102} 01/28/2022 16:06:48 - INFO - codeparrot_training - Step 2486: {'lr': 0.0004998735370803917, 'samples': 477504, 'steps': 2486, 'loss/train': 1.8216528296470642} 01/28/2022 16:06:52 - INFO - codeparrot_training - Step 2487: {'lr': 0.0004998730161655459, 'samples': 477696, 'steps': 2487, 'loss/train': 1.1060341000556946} 01/28/2022 16:06:57 - INFO - codeparrot_training - Step 2488: {'lr': 0.0004998724941803232, 'samples': 477888, 'steps': 2488, 'loss/train': 0.9793054461479187} 01/28/2022 16:07:03 - INFO - codeparrot_training - Step 2489: {'lr': 0.0004998719711247262, 'samples': 478080, 'steps': 2489, 'loss/train': 2.159582197666168} 01/28/2022 16:07:07 - INFO - codeparrot_training - Step 2490: {'lr': 0.0004998714469987571, 'samples': 478272, 'steps': 2490, 'loss/train': 1.9866694808006287} 01/28/2022 16:07:11 - INFO - codeparrot_training - Step 2491: {'lr': 0.000499870921802418, 'samples': 478464, 'steps': 2491, 'loss/train': 1.141652137041092} 01/28/2022 16:07:16 - INFO - codeparrot_training - Step 2492: {'lr': 0.0004998703955357111, 'samples': 478656, 'steps': 2492, 'loss/train': 1.9827995896339417} 01/28/2022 16:07:20 - INFO - codeparrot_training - Step 2493: {'lr': 0.0004998698681986389, 'samples': 478848, 'steps': 2493, 'loss/train': 2.509343683719635} 01/28/2022 16:07:24 - INFO - codeparrot_training - Step 2494: {'lr': 0.0004998693397912034, 'samples': 479040, 'steps': 2494, 'loss/train': 3.371426224708557} 01/28/2022 16:07:29 - INFO - codeparrot_training - Step 2495: {'lr': 0.0004998688103134072, 'samples': 479232, 'steps': 2495, 'loss/train': 2.203923761844635} 01/28/2022 16:07:34 - INFO - codeparrot_training - Step 2496: {'lr': 0.0004998682797652522, 'samples': 479424, 'steps': 2496, 'loss/train': 2.0834566354751587} 01/28/2022 16:07:38 - INFO - codeparrot_training - Step 2497: {'lr': 0.0004998677481467408, 'samples': 479616, 'steps': 2497, 'loss/train': 2.251102387905121} 01/28/2022 16:07:42 - INFO - codeparrot_training - Step 2498: {'lr': 0.0004998672154578754, 'samples': 479808, 'steps': 2498, 'loss/train': 2.260287344455719} 01/28/2022 16:07:46 - INFO - codeparrot_training - Step 2499: {'lr': 0.0004998666816986582, 'samples': 480000, 'steps': 2499, 'loss/train': 0.8864616751670837} 01/28/2022 16:07:52 - INFO - codeparrot_training - Step 2500: {'lr': 0.0004998661468690914, 'samples': 480192, 'steps': 2500, 'loss/train': 0.3659955561161041} 01/28/2022 16:07:56 - INFO - codeparrot_training - Step 2501: {'lr': 0.0004998656109691774, 'samples': 480384, 'steps': 2501, 'loss/train': 7.640167236328125} 01/28/2022 16:08:01 - INFO - codeparrot_training - Step 2502: {'lr': 0.0004998650739989185, 'samples': 480576, 'steps': 2502, 'loss/train': 2.275099217891693} 01/28/2022 16:08:05 - INFO - codeparrot_training - Step 2503: {'lr': 0.0004998645359583169, 'samples': 480768, 'steps': 2503, 'loss/train': 0.8551530539989471} 01/28/2022 16:08:09 - INFO - codeparrot_training - Step 2504: {'lr': 0.0004998639968473751, 'samples': 480960, 'steps': 2504, 'loss/train': 1.784681260585785} 01/28/2022 16:08:14 - INFO - codeparrot_training - Step 2505: {'lr': 0.0004998634566660952, 'samples': 481152, 'steps': 2505, 'loss/train': 0.8009107410907745} 01/28/2022 16:08:19 - INFO - codeparrot_training - Step 2506: {'lr': 0.0004998629154144795, 'samples': 481344, 'steps': 2506, 'loss/train': 1.6470192074775696} 01/28/2022 16:08:23 - INFO - codeparrot_training - Step 2507: {'lr': 0.0004998623730925305, 'samples': 481536, 'steps': 2507, 'loss/train': 1.2708532512187958} 01/28/2022 16:08:27 - INFO - codeparrot_training - Step 2508: {'lr': 0.0004998618297002504, 'samples': 481728, 'steps': 2508, 'loss/train': 2.210393786430359} 01/28/2022 16:08:31 - INFO - codeparrot_training - Step 2509: {'lr': 0.0004998612852376417, 'samples': 481920, 'steps': 2509, 'loss/train': 2.0113815665245056} 01/28/2022 16:08:37 - INFO - codeparrot_training - Step 2510: {'lr': 0.0004998607397047063, 'samples': 482112, 'steps': 2510, 'loss/train': 2.7478360533714294} 01/28/2022 16:08:42 - INFO - codeparrot_training - Step 2511: {'lr': 0.0004998601931014471, 'samples': 482304, 'steps': 2511, 'loss/train': 1.686794400215149} 01/28/2022 16:08:46 - INFO - codeparrot_training - Step 2512: {'lr': 0.0004998596454278661, 'samples': 482496, 'steps': 2512, 'loss/train': 1.399339646100998} 01/28/2022 16:08:50 - INFO - codeparrot_training - Step 2513: {'lr': 0.0004998590966839657, 'samples': 482688, 'steps': 2513, 'loss/train': 2.2570148706436157} 01/28/2022 16:08:54 - INFO - codeparrot_training - Step 2514: {'lr': 0.0004998585468697482, 'samples': 482880, 'steps': 2514, 'loss/train': 1.7745868563652039} 01/28/2022 16:09:00 - INFO - codeparrot_training - Step 2515: {'lr': 0.0004998579959852161, 'samples': 483072, 'steps': 2515, 'loss/train': 1.7465647459030151} 01/28/2022 16:09:04 - INFO - codeparrot_training - Step 2516: {'lr': 0.0004998574440303718, 'samples': 483264, 'steps': 2516, 'loss/train': 2.0553667545318604} 01/28/2022 16:09:08 - INFO - codeparrot_training - Step 2517: {'lr': 0.0004998568910052173, 'samples': 483456, 'steps': 2517, 'loss/train': 1.55695241689682} 01/28/2022 16:09:12 - INFO - codeparrot_training - Step 2518: {'lr': 0.0004998563369097554, 'samples': 483648, 'steps': 2518, 'loss/train': 2.138177990913391} 01/28/2022 16:09:16 - INFO - codeparrot_training - Step 2519: {'lr': 0.0004998557817439882, 'samples': 483840, 'steps': 2519, 'loss/train': 2.6709283590316772} 01/28/2022 16:09:22 - INFO - codeparrot_training - Step 2520: {'lr': 0.0004998552255079182, 'samples': 484032, 'steps': 2520, 'loss/train': 2.3491668105125427} 01/28/2022 16:09:26 - INFO - codeparrot_training - Step 2521: {'lr': 0.0004998546682015478, 'samples': 484224, 'steps': 2521, 'loss/train': 1.4795891046524048} 01/28/2022 16:09:30 - INFO - codeparrot_training - Step 2522: {'lr': 0.0004998541098248793, 'samples': 484416, 'steps': 2522, 'loss/train': 2.108799934387207} 01/28/2022 16:09:34 - INFO - codeparrot_training - Step 2523: {'lr': 0.0004998535503779151, 'samples': 484608, 'steps': 2523, 'loss/train': 1.7623048424720764} 01/28/2022 16:09:40 - INFO - codeparrot_training - Step 2524: {'lr': 0.0004998529898606576, 'samples': 484800, 'steps': 2524, 'loss/train': 1.1099495887756348} 01/28/2022 16:09:45 - INFO - codeparrot_training - Step 2525: {'lr': 0.0004998524282731093, 'samples': 484992, 'steps': 2525, 'loss/train': 2.142311453819275} 01/28/2022 16:09:49 - INFO - codeparrot_training - Step 2526: {'lr': 0.0004998518656152725, 'samples': 485184, 'steps': 2526, 'loss/train': 2.9388803243637085} 01/28/2022 16:09:53 - INFO - codeparrot_training - Step 2527: {'lr': 0.0004998513018871498, 'samples': 485376, 'steps': 2527, 'loss/train': 1.4957881271839142} 01/28/2022 16:09:57 - INFO - codeparrot_training - Step 2528: {'lr': 0.0004998507370887433, 'samples': 485568, 'steps': 2528, 'loss/train': 4.446305751800537} 01/28/2022 16:10:02 - INFO - codeparrot_training - Step 2529: {'lr': 0.0004998501712200555, 'samples': 485760, 'steps': 2529, 'loss/train': 1.7537546753883362} 01/28/2022 16:10:07 - INFO - codeparrot_training - Step 2530: {'lr': 0.000499849604281089, 'samples': 485952, 'steps': 2530, 'loss/train': 1.9594060778617859} 01/28/2022 16:10:11 - INFO - codeparrot_training - Step 2531: {'lr': 0.0004998490362718462, 'samples': 486144, 'steps': 2531, 'loss/train': 2.040957748889923} 01/28/2022 16:10:16 - INFO - codeparrot_training - Step 2532: {'lr': 0.0004998484671923293, 'samples': 486336, 'steps': 2532, 'loss/train': 1.7414976954460144} 01/28/2022 16:10:20 - INFO - codeparrot_training - Step 2533: {'lr': 0.000499847897042541, 'samples': 486528, 'steps': 2533, 'loss/train': 2.387184262275696} 01/28/2022 16:10:24 - INFO - codeparrot_training - Step 2534: {'lr': 0.0004998473258224837, 'samples': 486720, 'steps': 2534, 'loss/train': 2.054633617401123} 01/28/2022 16:10:31 - INFO - codeparrot_training - Step 2535: {'lr': 0.0004998467535321597, 'samples': 486912, 'steps': 2535, 'loss/train': 2.004441976547241} 01/28/2022 16:10:35 - INFO - codeparrot_training - Step 2536: {'lr': 0.0004998461801715716, 'samples': 487104, 'steps': 2536, 'loss/train': 2.5014209747314453} 01/28/2022 16:10:39 - INFO - codeparrot_training - Step 2537: {'lr': 0.0004998456057407218, 'samples': 487296, 'steps': 2537, 'loss/train': 2.6151645183563232} 01/28/2022 16:10:43 - INFO - codeparrot_training - Step 2538: {'lr': 0.0004998450302396127, 'samples': 487488, 'steps': 2538, 'loss/train': 1.4923443496227264} 01/28/2022 16:10:49 - INFO - codeparrot_training - Step 2539: {'lr': 0.0004998444536682469, 'samples': 487680, 'steps': 2539, 'loss/train': 2.3311296701431274} 01/28/2022 16:10:53 - INFO - codeparrot_training - Step 2540: {'lr': 0.0004998438760266267, 'samples': 487872, 'steps': 2540, 'loss/train': 0.6034493297338486} 01/28/2022 16:10:57 - INFO - codeparrot_training - Step 2541: {'lr': 0.0004998432973147548, 'samples': 488064, 'steps': 2541, 'loss/train': 1.9727556109428406} 01/28/2022 16:11:02 - INFO - codeparrot_training - Step 2542: {'lr': 0.0004998427175326335, 'samples': 488256, 'steps': 2542, 'loss/train': 1.729420244693756} 01/28/2022 16:11:06 - INFO - codeparrot_training - Step 2543: {'lr': 0.0004998421366802653, 'samples': 488448, 'steps': 2543, 'loss/train': 2.3278793692588806} 01/28/2022 16:11:10 - INFO - codeparrot_training - Step 2544: {'lr': 0.0004998415547576527, 'samples': 488640, 'steps': 2544, 'loss/train': 1.8590494394302368} 01/28/2022 16:11:15 - INFO - codeparrot_training - Step 2545: {'lr': 0.0004998409717647983, 'samples': 488832, 'steps': 2545, 'loss/train': 2.0246137976646423} 01/28/2022 16:11:20 - INFO - codeparrot_training - Step 2546: {'lr': 0.0004998403877017044, 'samples': 489024, 'steps': 2546, 'loss/train': 0.7779484391212463} 01/28/2022 16:11:24 - INFO - codeparrot_training - Step 2547: {'lr': 0.0004998398025683737, 'samples': 489216, 'steps': 2547, 'loss/train': 2.2727229595184326} 01/28/2022 16:11:28 - INFO - codeparrot_training - Step 2548: {'lr': 0.0004998392163648085, 'samples': 489408, 'steps': 2548, 'loss/train': 2.088809072971344} 01/28/2022 16:11:32 - INFO - codeparrot_training - Step 2549: {'lr': 0.0004998386290910116, 'samples': 489600, 'steps': 2549, 'loss/train': 3.2494232654571533} 01/28/2022 16:11:38 - INFO - codeparrot_training - Step 2550: {'lr': 0.0004998380407469853, 'samples': 489792, 'steps': 2550, 'loss/train': 2.565567433834076} 01/28/2022 16:11:42 - INFO - codeparrot_training - Step 2551: {'lr': 0.0004998374513327321, 'samples': 489984, 'steps': 2551, 'loss/train': 1.8055416941642761} 01/28/2022 16:11:46 - INFO - codeparrot_training - Step 2552: {'lr': 0.0004998368608482546, 'samples': 490176, 'steps': 2552, 'loss/train': 2.181822180747986} 01/28/2022 16:11:50 - INFO - codeparrot_training - Step 2553: {'lr': 0.0004998362692935553, 'samples': 490368, 'steps': 2553, 'loss/train': 1.9098538756370544} 01/28/2022 16:11:55 - INFO - codeparrot_training - Step 2554: {'lr': 0.0004998356766686368, 'samples': 490560, 'steps': 2554, 'loss/train': 2.113959789276123} 01/28/2022 16:12:01 - INFO - codeparrot_training - Step 2555: {'lr': 0.0004998350829735016, 'samples': 490752, 'steps': 2555, 'loss/train': 2.2886754870414734} 01/28/2022 16:12:05 - INFO - codeparrot_training - Step 2556: {'lr': 0.0004998344882081522, 'samples': 490944, 'steps': 2556, 'loss/train': 2.1074329018592834} 01/28/2022 16:12:09 - INFO - codeparrot_training - Step 2557: {'lr': 0.0004998338923725913, 'samples': 491136, 'steps': 2557, 'loss/train': 0.6363036632537842} 01/28/2022 16:12:13 - INFO - codeparrot_training - Step 2558: {'lr': 0.0004998332954668211, 'samples': 491328, 'steps': 2558, 'loss/train': 1.5987532138824463} 01/28/2022 16:12:18 - INFO - codeparrot_training - Step 2559: {'lr': 0.0004998326974908446, 'samples': 491520, 'steps': 2559, 'loss/train': 2.524575412273407} 01/28/2022 16:12:23 - INFO - codeparrot_training - Step 2560: {'lr': 0.0004998320984446641, 'samples': 491712, 'steps': 2560, 'loss/train': 1.8767731189727783} 01/28/2022 16:12:27 - INFO - codeparrot_training - Step 2561: {'lr': 0.0004998314983282821, 'samples': 491904, 'steps': 2561, 'loss/train': 2.3434976935386658} 01/28/2022 16:12:32 - INFO - codeparrot_training - Step 2562: {'lr': 0.0004998308971417015, 'samples': 492096, 'steps': 2562, 'loss/train': 1.960509181022644} 01/28/2022 16:12:36 - INFO - codeparrot_training - Step 2563: {'lr': 0.0004998302948849246, 'samples': 492288, 'steps': 2563, 'loss/train': 2.2585301399230957} 01/28/2022 16:12:40 - INFO - codeparrot_training - Step 2564: {'lr': 0.0004998296915579539, 'samples': 492480, 'steps': 2564, 'loss/train': 0.7170563489198685} 01/28/2022 16:12:46 - INFO - codeparrot_training - Step 2565: {'lr': 0.0004998290871607924, 'samples': 492672, 'steps': 2565, 'loss/train': 2.331052780151367} 01/28/2022 16:12:50 - INFO - codeparrot_training - Step 2566: {'lr': 0.0004998284816934422, 'samples': 492864, 'steps': 2566, 'loss/train': 1.8354588747024536} 01/28/2022 16:12:54 - INFO - codeparrot_training - Step 2567: {'lr': 0.0004998278751559062, 'samples': 493056, 'steps': 2567, 'loss/train': 2.5480191707611084} 01/28/2022 16:12:58 - INFO - codeparrot_training - Step 2568: {'lr': 0.0004998272675481868, 'samples': 493248, 'steps': 2568, 'loss/train': 2.121209442615509} 01/28/2022 16:13:03 - INFO - codeparrot_training - Step 2569: {'lr': 0.0004998266588702869, 'samples': 493440, 'steps': 2569, 'loss/train': 1.4433334171772003} 01/28/2022 16:13:09 - INFO - codeparrot_training - Step 2570: {'lr': 0.0004998260491222088, 'samples': 493632, 'steps': 2570, 'loss/train': 2.441839814186096} 01/28/2022 16:13:13 - INFO - codeparrot_training - Step 2571: {'lr': 0.0004998254383039552, 'samples': 493824, 'steps': 2571, 'loss/train': 2.347668170928955} 01/28/2022 16:13:17 - INFO - codeparrot_training - Step 2572: {'lr': 0.0004998248264155288, 'samples': 494016, 'steps': 2572, 'loss/train': 1.1915050148963928} 01/28/2022 16:13:22 - INFO - codeparrot_training - Step 2573: {'lr': 0.0004998242134569322, 'samples': 494208, 'steps': 2573, 'loss/train': 0.6510611772537231} 01/28/2022 16:13:26 - INFO - codeparrot_training - Step 2574: {'lr': 0.0004998235994281681, 'samples': 494400, 'steps': 2574, 'loss/train': 2.2850716710090637} 01/28/2022 16:13:31 - INFO - codeparrot_training - Step 2575: {'lr': 0.0004998229843292388, 'samples': 494592, 'steps': 2575, 'loss/train': 2.141515016555786} 01/28/2022 16:13:35 - INFO - codeparrot_training - Step 2576: {'lr': 0.0004998223681601474, 'samples': 494784, 'steps': 2576, 'loss/train': 0.8098731637001038} 01/28/2022 16:13:39 - INFO - codeparrot_training - Step 2577: {'lr': 0.0004998217509208961, 'samples': 494976, 'steps': 2577, 'loss/train': 3.099751353263855} 01/28/2022 16:13:44 - INFO - codeparrot_training - Step 2578: {'lr': 0.0004998211326114878, 'samples': 495168, 'steps': 2578, 'loss/train': 2.574513852596283} 01/28/2022 16:13:48 - INFO - codeparrot_training - Step 2579: {'lr': 0.0004998205132319252, 'samples': 495360, 'steps': 2579, 'loss/train': 2.643044114112854} 01/28/2022 16:13:54 - INFO - codeparrot_training - Step 2580: {'lr': 0.0004998198927822108, 'samples': 495552, 'steps': 2580, 'loss/train': 2.399567663669586} 01/28/2022 16:13:58 - INFO - codeparrot_training - Step 2581: {'lr': 0.0004998192712623472, 'samples': 495744, 'steps': 2581, 'loss/train': 1.8093359470367432} 01/28/2022 16:14:03 - INFO - codeparrot_training - Step 2582: {'lr': 0.0004998186486723373, 'samples': 495936, 'steps': 2582, 'loss/train': 2.1684170365333557} 01/28/2022 16:14:07 - INFO - codeparrot_training - Step 2583: {'lr': 0.0004998180250121836, 'samples': 496128, 'steps': 2583, 'loss/train': 2.4786645770072937} 01/28/2022 16:14:11 - INFO - codeparrot_training - Step 2584: {'lr': 0.0004998174002818887, 'samples': 496320, 'steps': 2584, 'loss/train': 1.4459193348884583} 01/28/2022 16:14:16 - INFO - codeparrot_training - Step 2585: {'lr': 0.0004998167744814555, 'samples': 496512, 'steps': 2585, 'loss/train': 1.5615110993385315} 01/28/2022 16:14:21 - INFO - codeparrot_training - Step 2586: {'lr': 0.0004998161476108864, 'samples': 496704, 'steps': 2586, 'loss/train': 0.7645061910152435} 01/28/2022 16:14:25 - INFO - codeparrot_training - Step 2587: {'lr': 0.0004998155196701845, 'samples': 496896, 'steps': 2587, 'loss/train': 2.0567725896835327} 01/28/2022 16:14:29 - INFO - codeparrot_training - Step 2588: {'lr': 0.000499814890659352, 'samples': 497088, 'steps': 2588, 'loss/train': 1.7481620907783508} 01/28/2022 16:14:33 - INFO - codeparrot_training - Step 2589: {'lr': 0.000499814260578392, 'samples': 497280, 'steps': 2589, 'loss/train': 3.0338133573532104} 01/28/2022 16:14:39 - INFO - codeparrot_training - Step 2590: {'lr': 0.000499813629427307, 'samples': 497472, 'steps': 2590, 'loss/train': 1.6267290115356445} 01/28/2022 16:14:43 - INFO - codeparrot_training - Step 2591: {'lr': 0.0004998129972060998, 'samples': 497664, 'steps': 2591, 'loss/train': 3.3577791452407837} 01/28/2022 16:14:47 - INFO - codeparrot_training - Step 2592: {'lr': 0.000499812363914773, 'samples': 497856, 'steps': 2592, 'loss/train': 2.441914200782776} 01/28/2022 16:14:52 - INFO - codeparrot_training - Step 2593: {'lr': 0.0004998117295533292, 'samples': 498048, 'steps': 2593, 'loss/train': 3.5788503885269165} 01/28/2022 16:14:56 - INFO - codeparrot_training - Step 2594: {'lr': 0.0004998110941217714, 'samples': 498240, 'steps': 2594, 'loss/train': 2.268131196498871} 01/28/2022 16:15:02 - INFO - codeparrot_training - Step 2595: {'lr': 0.0004998104576201022, 'samples': 498432, 'steps': 2595, 'loss/train': 2.049066960811615} 01/28/2022 16:15:06 - INFO - codeparrot_training - Step 2596: {'lr': 0.0004998098200483243, 'samples': 498624, 'steps': 2596, 'loss/train': 2.450718104839325} 01/28/2022 16:15:10 - INFO - codeparrot_training - Step 2597: {'lr': 0.0004998091814064405, 'samples': 498816, 'steps': 2597, 'loss/train': 1.9614644050598145} 01/28/2022 16:15:15 - INFO - codeparrot_training - Step 2598: {'lr': 0.0004998085416944534, 'samples': 499008, 'steps': 2598, 'loss/train': 1.9444865584373474} 01/28/2022 16:15:19 - INFO - codeparrot_training - Step 2599: {'lr': 0.000499807900912366, 'samples': 499200, 'steps': 2599, 'loss/train': 1.8972405195236206} 01/28/2022 16:15:24 - INFO - codeparrot_training - Step 2600: {'lr': 0.0004998072590601808, 'samples': 499392, 'steps': 2600, 'loss/train': 1.9391087293624878} 01/28/2022 16:15:28 - INFO - codeparrot_training - Step 2601: {'lr': 0.0004998066161379006, 'samples': 499584, 'steps': 2601, 'loss/train': 2.862732946872711} 01/28/2022 16:15:33 - INFO - codeparrot_training - Step 2602: {'lr': 0.0004998059721455281, 'samples': 499776, 'steps': 2602, 'loss/train': 2.364549458026886} 01/28/2022 16:15:37 - INFO - codeparrot_training - Step 2603: {'lr': 0.0004998053270830662, 'samples': 499968, 'steps': 2603, 'loss/train': 1.7850188612937927} 01/28/2022 16:15:41 - INFO - codeparrot_training - Step 2604: {'lr': 0.0004998046809505176, 'samples': 500160, 'steps': 2604, 'loss/train': 3.230522632598877} 01/28/2022 16:15:46 - INFO - codeparrot_training - Step 2605: {'lr': 0.0004998040337478851, 'samples': 500352, 'steps': 2605, 'loss/train': 2.1755869388580322} 01/28/2022 16:15:51 - INFO - codeparrot_training - Step 2606: {'lr': 0.0004998033854751715, 'samples': 500544, 'steps': 2606, 'loss/train': 1.8498555421829224} 01/28/2022 16:15:55 - INFO - codeparrot_training - Step 2607: {'lr': 0.0004998027361323794, 'samples': 500736, 'steps': 2607, 'loss/train': 1.9670745134353638} 01/28/2022 16:15:59 - INFO - codeparrot_training - Step 2608: {'lr': 0.0004998020857195117, 'samples': 500928, 'steps': 2608, 'loss/train': 1.6175240278244019} 01/28/2022 16:16:03 - INFO - codeparrot_training - Step 2609: {'lr': 0.0004998014342365712, 'samples': 501120, 'steps': 2609, 'loss/train': 0.640228807926178} 01/28/2022 16:16:09 - INFO - codeparrot_training - Step 2610: {'lr': 0.0004998007816835608, 'samples': 501312, 'steps': 2610, 'loss/train': 1.9003631472587585} 01/28/2022 16:16:13 - INFO - codeparrot_training - Step 2611: {'lr': 0.000499800128060483, 'samples': 501504, 'steps': 2611, 'loss/train': 1.7514342069625854} 01/28/2022 16:16:17 - INFO - codeparrot_training - Step 2612: {'lr': 0.0004997994733673409, 'samples': 501696, 'steps': 2612, 'loss/train': 2.410254120826721} 01/28/2022 16:16:21 - INFO - codeparrot_training - Step 2613: {'lr': 0.000499798817604137, 'samples': 501888, 'steps': 2613, 'loss/train': 2.4270150661468506} 01/28/2022 16:16:26 - INFO - codeparrot_training - Step 2614: {'lr': 0.0004997981607708745, 'samples': 502080, 'steps': 2614, 'loss/train': 1.4607843160629272} 01/28/2022 16:16:32 - INFO - codeparrot_training - Step 2615: {'lr': 0.0004997975028675558, 'samples': 502272, 'steps': 2615, 'loss/train': 1.0939720273017883} 01/28/2022 16:16:36 - INFO - codeparrot_training - Step 2616: {'lr': 0.0004997968438941841, 'samples': 502464, 'steps': 2616, 'loss/train': 1.57782644033432} 01/28/2022 16:16:40 - INFO - codeparrot_training - Step 2617: {'lr': 0.0004997961838507619, 'samples': 502656, 'steps': 2617, 'loss/train': 2.2761778235435486} 01/28/2022 16:16:44 - INFO - codeparrot_training - Step 2618: {'lr': 0.0004997955227372923, 'samples': 502848, 'steps': 2618, 'loss/train': 1.6494230031967163} 01/28/2022 16:16:49 - INFO - codeparrot_training - Step 2619: {'lr': 0.000499794860553778, 'samples': 503040, 'steps': 2619, 'loss/train': 2.412366807460785} 01/28/2022 16:16:54 - INFO - codeparrot_training - Step 2620: {'lr': 0.0004997941973002216, 'samples': 503232, 'steps': 2620, 'loss/train': 1.7622374296188354} 01/28/2022 16:16:58 - INFO - codeparrot_training - Step 2621: {'lr': 0.0004997935329766265, 'samples': 503424, 'steps': 2621, 'loss/train': 1.1536813974380493} 01/28/2022 16:17:02 - INFO - codeparrot_training - Step 2622: {'lr': 0.000499792867582995, 'samples': 503616, 'steps': 2622, 'loss/train': 2.769439995288849} 01/28/2022 16:17:07 - INFO - codeparrot_training - Step 2623: {'lr': 0.0004997922011193303, 'samples': 503808, 'steps': 2623, 'loss/train': 2.254555106163025} 01/28/2022 16:17:11 - INFO - codeparrot_training - Step 2624: {'lr': 0.000499791533585635, 'samples': 504000, 'steps': 2624, 'loss/train': 2.0879343152046204} 01/28/2022 16:17:16 - INFO - codeparrot_training - Step 2625: {'lr': 0.0004997908649819122, 'samples': 504192, 'steps': 2625, 'loss/train': 2.910595715045929} 01/28/2022 16:17:20 - INFO - codeparrot_training - Step 2626: {'lr': 0.0004997901953081646, 'samples': 504384, 'steps': 2626, 'loss/train': 2.1429598331451416} 01/28/2022 16:17:25 - INFO - codeparrot_training - Step 2627: {'lr': 0.0004997895245643951, 'samples': 504576, 'steps': 2627, 'loss/train': 2.6791089177131653} 01/28/2022 16:17:29 - INFO - codeparrot_training - Step 2628: {'lr': 0.0004997888527506067, 'samples': 504768, 'steps': 2628, 'loss/train': 1.8085602521896362} 01/28/2022 16:17:33 - INFO - codeparrot_training - Step 2629: {'lr': 0.000499788179866802, 'samples': 504960, 'steps': 2629, 'loss/train': 0.3804489076137543} 01/28/2022 16:17:39 - INFO - codeparrot_training - Step 2630: {'lr': 0.0004997875059129843, 'samples': 505152, 'steps': 2630, 'loss/train': 1.872795045375824} 01/28/2022 16:17:43 - INFO - codeparrot_training - Step 2631: {'lr': 0.000499786830889156, 'samples': 505344, 'steps': 2631, 'loss/train': 2.117417335510254} 01/28/2022 16:17:48 - INFO - codeparrot_training - Step 2632: {'lr': 0.0004997861547953203, 'samples': 505536, 'steps': 2632, 'loss/train': 2.5376967787742615} 01/28/2022 16:17:52 - INFO - codeparrot_training - Step 2633: {'lr': 0.00049978547763148, 'samples': 505728, 'steps': 2633, 'loss/train': 1.82765793800354} 01/28/2022 16:17:56 - INFO - codeparrot_training - Step 2634: {'lr': 0.0004997847993976381, 'samples': 505920, 'steps': 2634, 'loss/train': 1.1712586283683777} 01/28/2022 16:18:01 - INFO - codeparrot_training - Step 2635: {'lr': 0.0004997841200937975, 'samples': 506112, 'steps': 2635, 'loss/train': 1.3324507176876068} 01/28/2022 16:18:06 - INFO - codeparrot_training - Step 2636: {'lr': 0.0004997834397199609, 'samples': 506304, 'steps': 2636, 'loss/train': 2.865878462791443} 01/28/2022 16:18:10 - INFO - codeparrot_training - Step 2637: {'lr': 0.0004997827582761315, 'samples': 506496, 'steps': 2637, 'loss/train': 2.518528997898102} 01/28/2022 16:18:14 - INFO - codeparrot_training - Step 2638: {'lr': 0.0004997820757623119, 'samples': 506688, 'steps': 2638, 'loss/train': 1.8564201593399048} 01/28/2022 16:18:19 - INFO - codeparrot_training - Step 2639: {'lr': 0.0004997813921785054, 'samples': 506880, 'steps': 2639, 'loss/train': 1.498580664396286} 01/28/2022 16:18:25 - INFO - codeparrot_training - Step 2640: {'lr': 0.0004997807075247146, 'samples': 507072, 'steps': 2640, 'loss/train': 2.7186466455459595} 01/28/2022 16:18:29 - INFO - codeparrot_training - Step 2641: {'lr': 0.0004997800218009426, 'samples': 507264, 'steps': 2641, 'loss/train': 1.728000283241272} 01/28/2022 16:18:33 - INFO - codeparrot_training - Step 2642: {'lr': 0.0004997793350071923, 'samples': 507456, 'steps': 2642, 'loss/train': 2.043016791343689} 01/28/2022 16:18:38 - INFO - codeparrot_training - Step 2643: {'lr': 0.0004997786471434666, 'samples': 507648, 'steps': 2643, 'loss/train': 2.1043437123298645} 01/28/2022 16:18:42 - INFO - codeparrot_training - Step 2644: {'lr': 0.0004997779582097686, 'samples': 507840, 'steps': 2644, 'loss/train': 1.6447314620018005} 01/28/2022 16:18:48 - INFO - codeparrot_training - Step 2645: {'lr': 0.0004997772682061011, 'samples': 508032, 'steps': 2645, 'loss/train': 2.101845860481262} 01/28/2022 16:18:52 - INFO - codeparrot_training - Step 2646: {'lr': 0.000499776577132467, 'samples': 508224, 'steps': 2646, 'loss/train': 1.7915478944778442} 01/28/2022 16:18:56 - INFO - codeparrot_training - Step 2647: {'lr': 0.0004997758849888693, 'samples': 508416, 'steps': 2647, 'loss/train': 1.7444747686386108} 01/28/2022 16:19:00 - INFO - codeparrot_training - Step 2648: {'lr': 0.0004997751917753113, 'samples': 508608, 'steps': 2648, 'loss/train': 3.448694944381714} 01/28/2022 16:19:05 - INFO - codeparrot_training - Step 2649: {'lr': 0.0004997744974917955, 'samples': 508800, 'steps': 2649, 'loss/train': 1.8375964164733887} 01/28/2022 16:19:10 - INFO - codeparrot_training - Step 2650: {'lr': 0.0004997738021383252, 'samples': 508992, 'steps': 2650, 'loss/train': 1.8632524609565735} 01/28/2022 16:19:14 - INFO - codeparrot_training - Step 2651: {'lr': 0.000499773105714903, 'samples': 509184, 'steps': 2651, 'loss/train': 2.2386454939842224} 01/28/2022 16:19:18 - INFO - codeparrot_training - Step 2652: {'lr': 0.0004997724082215323, 'samples': 509376, 'steps': 2652, 'loss/train': 1.427517592906952} 01/28/2022 16:19:23 - INFO - codeparrot_training - Step 2653: {'lr': 0.0004997717096582159, 'samples': 509568, 'steps': 2653, 'loss/train': 2.922325909137726} 01/28/2022 16:19:27 - INFO - codeparrot_training - Step 2654: {'lr': 0.0004997710100249568, 'samples': 509760, 'steps': 2654, 'loss/train': 2.257430613040924} 01/28/2022 16:19:33 - INFO - codeparrot_training - Step 2655: {'lr': 0.000499770309321758, 'samples': 509952, 'steps': 2655, 'loss/train': 1.885106384754181} 01/28/2022 16:19:37 - INFO - codeparrot_training - Step 2656: {'lr': 0.0004997696075486225, 'samples': 510144, 'steps': 2656, 'loss/train': 1.6421306133270264} 01/28/2022 16:19:42 - INFO - codeparrot_training - Step 2657: {'lr': 0.0004997689047055534, 'samples': 510336, 'steps': 2657, 'loss/train': 0.9498263597488403} 01/28/2022 16:19:46 - INFO - codeparrot_training - Step 2658: {'lr': 0.0004997682007925535, 'samples': 510528, 'steps': 2658, 'loss/train': 2.333040475845337} 01/28/2022 16:19:50 - INFO - codeparrot_training - Step 2659: {'lr': 0.0004997674958096259, 'samples': 510720, 'steps': 2659, 'loss/train': 2.098331093788147} 01/28/2022 16:19:55 - INFO - codeparrot_training - Step 2660: {'lr': 0.0004997667897567738, 'samples': 510912, 'steps': 2660, 'loss/train': 2.547235429286957} 01/28/2022 16:19:59 - INFO - codeparrot_training - Step 2661: {'lr': 0.000499766082634, 'samples': 511104, 'steps': 2661, 'loss/train': 2.1579891443252563} 01/28/2022 16:20:04 - INFO - codeparrot_training - Step 2662: {'lr': 0.0004997653744413076, 'samples': 511296, 'steps': 2662, 'loss/train': 2.26847380399704} 01/28/2022 16:20:08 - INFO - codeparrot_training - Step 2663: {'lr': 0.0004997646651786996, 'samples': 511488, 'steps': 2663, 'loss/train': 2.0927542448043823} 01/28/2022 16:20:12 - INFO - codeparrot_training - Step 2664: {'lr': 0.0004997639548461792, 'samples': 511680, 'steps': 2664, 'loss/train': 2.8180262446403503} 01/28/2022 16:20:19 - INFO - codeparrot_training - Step 2665: {'lr': 0.0004997632434437493, 'samples': 511872, 'steps': 2665, 'loss/train': 2.4077548384666443} 01/28/2022 16:20:23 - INFO - codeparrot_training - Step 2666: {'lr': 0.0004997625309714129, 'samples': 512064, 'steps': 2666, 'loss/train': 1.9166130423545837} 01/28/2022 16:20:27 - INFO - codeparrot_training - Step 2667: {'lr': 0.0004997618174291732, 'samples': 512256, 'steps': 2667, 'loss/train': 1.8597795367240906} 01/28/2022 16:20:31 - INFO - codeparrot_training - Step 2668: {'lr': 0.0004997611028170332, 'samples': 512448, 'steps': 2668, 'loss/train': 2.5336959958076477} 01/28/2022 16:20:36 - INFO - codeparrot_training - Step 2669: {'lr': 0.000499760387134996, 'samples': 512640, 'steps': 2669, 'loss/train': 2.7683746218681335} 01/28/2022 16:20:41 - INFO - codeparrot_training - Step 2670: {'lr': 0.0004997596703830645, 'samples': 512832, 'steps': 2670, 'loss/train': 2.200792372226715} 01/28/2022 16:20:45 - INFO - codeparrot_training - Step 2671: {'lr': 0.0004997589525612418, 'samples': 513024, 'steps': 2671, 'loss/train': 2.0939801931381226} 01/28/2022 16:20:50 - INFO - codeparrot_training - Step 2672: {'lr': 0.0004997582336695312, 'samples': 513216, 'steps': 2672, 'loss/train': 2.0695745944976807} 01/28/2022 16:20:54 - INFO - codeparrot_training - Step 2673: {'lr': 0.0004997575137079355, 'samples': 513408, 'steps': 2673, 'loss/train': 2.3521310091018677} 01/28/2022 16:20:58 - INFO - codeparrot_training - Step 2674: {'lr': 0.0004997567926764581, 'samples': 513600, 'steps': 2674, 'loss/train': 1.878406584262848} 01/28/2022 16:21:03 - INFO - codeparrot_training - Step 2675: {'lr': 0.0004997560705751018, 'samples': 513792, 'steps': 2675, 'loss/train': 3.7658225297927856} 01/28/2022 16:21:07 - INFO - codeparrot_training - Step 2676: {'lr': 0.0004997553474038698, 'samples': 513984, 'steps': 2676, 'loss/train': 1.651928186416626} 01/28/2022 16:21:12 - INFO - codeparrot_training - Step 2677: {'lr': 0.0004997546231627652, 'samples': 514176, 'steps': 2677, 'loss/train': 2.568650186061859} 01/28/2022 16:21:16 - INFO - codeparrot_training - Step 2678: {'lr': 0.0004997538978517912, 'samples': 514368, 'steps': 2678, 'loss/train': 2.068412125110626} 01/28/2022 16:21:20 - INFO - codeparrot_training - Step 2679: {'lr': 0.0004997531714709506, 'samples': 514560, 'steps': 2679, 'loss/train': 2.0681426525115967} 01/28/2022 16:21:26 - INFO - codeparrot_training - Step 2680: {'lr': 0.0004997524440202469, 'samples': 514752, 'steps': 2680, 'loss/train': 1.6707261800765991} 01/28/2022 16:21:30 - INFO - codeparrot_training - Step 2681: {'lr': 0.0004997517154996829, 'samples': 514944, 'steps': 2681, 'loss/train': 1.348948895931244} 01/28/2022 16:21:35 - INFO - codeparrot_training - Step 2682: {'lr': 0.000499750985909262, 'samples': 515136, 'steps': 2682, 'loss/train': 2.14546662569046} 01/28/2022 16:21:39 - INFO - codeparrot_training - Step 2683: {'lr': 0.0004997502552489871, 'samples': 515328, 'steps': 2683, 'loss/train': 2.1376789212226868} 01/28/2022 16:21:43 - INFO - codeparrot_training - Step 2684: {'lr': 0.0004997495235188614, 'samples': 515520, 'steps': 2684, 'loss/train': 2.0676991939544678} 01/28/2022 16:21:48 - INFO - codeparrot_training - Step 2685: {'lr': 0.0004997487907188881, 'samples': 515712, 'steps': 2685, 'loss/train': 2.3761491179466248} 01/28/2022 16:21:53 - INFO - codeparrot_training - Step 2686: {'lr': 0.0004997480568490702, 'samples': 515904, 'steps': 2686, 'loss/train': 2.014072895050049} 01/28/2022 16:21:57 - INFO - codeparrot_training - Step 2687: {'lr': 0.0004997473219094111, 'samples': 516096, 'steps': 2687, 'loss/train': 3.2953094244003296} 01/28/2022 16:22:01 - INFO - codeparrot_training - Step 2688: {'lr': 0.0004997465858999136, 'samples': 516288, 'steps': 2688, 'loss/train': 2.2752951979637146} 01/28/2022 16:22:05 - INFO - codeparrot_training - Step 2689: {'lr': 0.0004997458488205811, 'samples': 516480, 'steps': 2689, 'loss/train': 1.4360565841197968} 01/28/2022 16:22:11 - INFO - codeparrot_training - Step 2690: {'lr': 0.0004997451106714166, 'samples': 516672, 'steps': 2690, 'loss/train': 1.4898558855056763} 01/28/2022 16:22:16 - INFO - codeparrot_training - Step 2691: {'lr': 0.0004997443714524235, 'samples': 516864, 'steps': 2691, 'loss/train': 1.7347660660743713} 01/28/2022 16:22:20 - INFO - codeparrot_training - Step 2692: {'lr': 0.0004997436311636046, 'samples': 517056, 'steps': 2692, 'loss/train': 2.3223971128463745} 01/28/2022 16:22:24 - INFO - codeparrot_training - Step 2693: {'lr': 0.0004997428898049635, 'samples': 517248, 'steps': 2693, 'loss/train': 2.258293926715851} 01/28/2022 16:22:28 - INFO - codeparrot_training - Step 2694: {'lr': 0.0004997421473765031, 'samples': 517440, 'steps': 2694, 'loss/train': 2.108013331890106} 01/28/2022 16:22:34 - INFO - codeparrot_training - Step 2695: {'lr': 0.0004997414038782266, 'samples': 517632, 'steps': 2695, 'loss/train': 2.540432631969452} 01/28/2022 16:22:38 - INFO - codeparrot_training - Step 2696: {'lr': 0.0004997406593101373, 'samples': 517824, 'steps': 2696, 'loss/train': 2.451370418071747} 01/28/2022 16:22:42 - INFO - codeparrot_training - Step 2697: {'lr': 0.0004997399136722383, 'samples': 518016, 'steps': 2697, 'loss/train': 1.886450171470642} 01/28/2022 16:22:47 - INFO - codeparrot_training - Step 2698: {'lr': 0.0004997391669645327, 'samples': 518208, 'steps': 2698, 'loss/train': 2.06521475315094} 01/28/2022 16:22:51 - INFO - codeparrot_training - Step 2699: {'lr': 0.0004997384191870239, 'samples': 518400, 'steps': 2699, 'loss/train': 2.121032953262329} 01/28/2022 16:22:56 - INFO - codeparrot_training - Step 2700: {'lr': 0.000499737670339715, 'samples': 518592, 'steps': 2700, 'loss/train': 1.4900512397289276} 01/28/2022 16:23:01 - INFO - codeparrot_training - Step 2701: {'lr': 0.0004997369204226093, 'samples': 518784, 'steps': 2701, 'loss/train': 1.9326138496398926} 01/28/2022 16:23:05 - INFO - codeparrot_training - Step 2702: {'lr': 0.0004997361694357098, 'samples': 518976, 'steps': 2702, 'loss/train': 1.8083946704864502} 01/28/2022 16:23:09 - INFO - codeparrot_training - Step 2703: {'lr': 0.00049973541737902, 'samples': 519168, 'steps': 2703, 'loss/train': 1.9337190985679626} 01/28/2022 16:23:13 - INFO - codeparrot_training - Step 2704: {'lr': 0.0004997346642525428, 'samples': 519360, 'steps': 2704, 'loss/train': 1.1574938893318176} 01/28/2022 16:23:19 - INFO - codeparrot_training - Step 2705: {'lr': 0.0004997339100562817, 'samples': 519552, 'steps': 2705, 'loss/train': 2.7682287096977234} 01/28/2022 16:23:24 - INFO - codeparrot_training - Step 2706: {'lr': 0.0004997331547902398, 'samples': 519744, 'steps': 2706, 'loss/train': 1.5839243531227112} 01/28/2022 16:23:28 - INFO - codeparrot_training - Step 2707: {'lr': 0.0004997323984544204, 'samples': 519936, 'steps': 2707, 'loss/train': 1.9704397916793823} 01/28/2022 16:23:32 - INFO - codeparrot_training - Step 2708: {'lr': 0.0004997316410488267, 'samples': 520128, 'steps': 2708, 'loss/train': 1.3901554942131042} 01/28/2022 16:23:36 - INFO - codeparrot_training - Step 2709: {'lr': 0.0004997308825734619, 'samples': 520320, 'steps': 2709, 'loss/train': 0.8993675708770752} 01/28/2022 16:23:42 - INFO - codeparrot_training - Step 2710: {'lr': 0.0004997301230283294, 'samples': 520512, 'steps': 2710, 'loss/train': 1.8798651695251465} 01/28/2022 16:23:46 - INFO - codeparrot_training - Step 2711: {'lr': 0.0004997293624134322, 'samples': 520704, 'steps': 2711, 'loss/train': 1.8825028538703918} 01/28/2022 16:23:50 - INFO - codeparrot_training - Step 2712: {'lr': 0.0004997286007287738, 'samples': 520896, 'steps': 2712, 'loss/train': 2.6718183159828186} 01/28/2022 16:23:54 - INFO - codeparrot_training - Step 2713: {'lr': 0.0004997278379743574, 'samples': 521088, 'steps': 2713, 'loss/train': 1.0196424722671509} 01/28/2022 16:23:59 - INFO - codeparrot_training - Step 2714: {'lr': 0.0004997270741501861, 'samples': 521280, 'steps': 2714, 'loss/train': 2.0649912357330322} 01/28/2022 16:24:04 - INFO - codeparrot_training - Step 2715: {'lr': 0.0004997263092562634, 'samples': 521472, 'steps': 2715, 'loss/train': 0.8769536018371582} 01/28/2022 16:24:08 - INFO - codeparrot_training - Step 2716: {'lr': 0.0004997255432925926, 'samples': 521664, 'steps': 2716, 'loss/train': 1.5284947156906128} 01/28/2022 16:24:12 - INFO - codeparrot_training - Step 2717: {'lr': 0.0004997247762591766, 'samples': 521856, 'steps': 2717, 'loss/train': 2.7457977533340454} 01/28/2022 16:24:17 - INFO - codeparrot_training - Step 2718: {'lr': 0.0004997240081560193, 'samples': 522048, 'steps': 2718, 'loss/train': 1.4502962529659271} 01/28/2022 16:24:21 - INFO - codeparrot_training - Step 2719: {'lr': 0.0004997232389831234, 'samples': 522240, 'steps': 2719, 'loss/train': 1.3146903812885284} 01/28/2022 16:24:26 - INFO - codeparrot_training - Step 2720: {'lr': 0.0004997224687404926, 'samples': 522432, 'steps': 2720, 'loss/train': 1.6217742562294006} 01/28/2022 16:24:30 - INFO - codeparrot_training - Step 2721: {'lr': 0.0004997216974281299, 'samples': 522624, 'steps': 2721, 'loss/train': 2.086282432079315} 01/28/2022 16:24:34 - INFO - codeparrot_training - Step 2722: {'lr': 0.0004997209250460387, 'samples': 522816, 'steps': 2722, 'loss/train': 1.964655876159668} 01/28/2022 16:24:39 - INFO - codeparrot_training - Step 2723: {'lr': 0.0004997201515942225, 'samples': 523008, 'steps': 2723, 'loss/train': 2.2889075875282288} 01/28/2022 16:24:43 - INFO - codeparrot_training - Step 2724: {'lr': 0.0004997193770726844, 'samples': 523200, 'steps': 2724, 'loss/train': 1.1919106543064117} 01/28/2022 16:24:49 - INFO - codeparrot_training - Step 2725: {'lr': 0.0004997186014814278, 'samples': 523392, 'steps': 2725, 'loss/train': 1.9861056804656982} 01/28/2022 16:24:53 - INFO - codeparrot_training - Step 2726: {'lr': 0.000499717824820456, 'samples': 523584, 'steps': 2726, 'loss/train': 2.1911065578460693} 01/28/2022 16:24:57 - INFO - codeparrot_training - Step 2727: {'lr': 0.0004997170470897723, 'samples': 523776, 'steps': 2727, 'loss/train': 0.9948188960552216} 01/28/2022 16:25:02 - INFO - codeparrot_training - Step 2728: {'lr': 0.0004997162682893801, 'samples': 523968, 'steps': 2728, 'loss/train': 2.5558347702026367} 01/28/2022 16:25:06 - INFO - codeparrot_training - Step 2729: {'lr': 0.0004997154884192827, 'samples': 524160, 'steps': 2729, 'loss/train': 1.8130205869674683} 01/28/2022 16:25:11 - INFO - codeparrot_training - Step 2730: {'lr': 0.0004997147074794835, 'samples': 524352, 'steps': 2730, 'loss/train': 1.9604941606521606} 01/28/2022 16:25:15 - INFO - codeparrot_training - Step 2731: {'lr': 0.0004997139254699856, 'samples': 524544, 'steps': 2731, 'loss/train': 1.2913083136081696} 01/28/2022 16:25:20 - INFO - codeparrot_training - Step 2732: {'lr': 0.0004997131423907927, 'samples': 524736, 'steps': 2732, 'loss/train': 1.9600263833999634} 01/28/2022 16:25:24 - INFO - codeparrot_training - Step 2733: {'lr': 0.000499712358241908, 'samples': 524928, 'steps': 2733, 'loss/train': 2.250646412372589} 01/28/2022 16:25:28 - INFO - codeparrot_training - Step 2734: {'lr': 0.0004997115730233349, 'samples': 525120, 'steps': 2734, 'loss/train': 1.8751700520515442} 01/28/2022 16:25:34 - INFO - codeparrot_training - Step 2735: {'lr': 0.0004997107867350765, 'samples': 525312, 'steps': 2735, 'loss/train': 2.3322967886924744} 01/28/2022 16:25:38 - INFO - codeparrot_training - Step 2736: {'lr': 0.0004997099993771365, 'samples': 525504, 'steps': 2736, 'loss/train': 1.4520982503890991} 01/28/2022 16:25:42 - INFO - codeparrot_training - Step 2737: {'lr': 0.0004997092109495181, 'samples': 525696, 'steps': 2737, 'loss/train': 1.8552128076553345} 01/28/2022 16:25:47 - INFO - codeparrot_training - Step 2738: {'lr': 0.0004997084214522249, 'samples': 525888, 'steps': 2738, 'loss/train': 2.2442834973335266} 01/28/2022 16:25:51 - INFO - codeparrot_training - Step 2739: {'lr': 0.0004997076308852599, 'samples': 526080, 'steps': 2739, 'loss/train': 1.5381844639778137} 01/28/2022 16:25:56 - INFO - codeparrot_training - Step 2740: {'lr': 0.0004997068392486268, 'samples': 526272, 'steps': 2740, 'loss/train': 1.3290838301181793} 01/28/2022 16:26:00 - INFO - codeparrot_training - Step 2741: {'lr': 0.0004997060465423288, 'samples': 526464, 'steps': 2741, 'loss/train': 2.591016411781311} 01/28/2022 16:26:04 - INFO - codeparrot_training - Step 2742: {'lr': 0.0004997052527663696, 'samples': 526656, 'steps': 2742, 'loss/train': 2.1460373997688293} 01/28/2022 16:26:09 - INFO - codeparrot_training - Step 2743: {'lr': 0.0004997044579207522, 'samples': 526848, 'steps': 2743, 'loss/train': 2.1138734221458435} 01/28/2022 16:26:13 - INFO - codeparrot_training - Step 2744: {'lr': 0.0004997036620054803, 'samples': 527040, 'steps': 2744, 'loss/train': 2.213067054748535} 01/28/2022 16:26:18 - INFO - codeparrot_training - Step 2745: {'lr': 0.0004997028650205572, 'samples': 527232, 'steps': 2745, 'loss/train': 1.2338629066944122} 01/28/2022 16:26:22 - INFO - codeparrot_training - Step 2746: {'lr': 0.0004997020669659862, 'samples': 527424, 'steps': 2746, 'loss/train': 2.147730052471161} 01/28/2022 16:26:27 - INFO - codeparrot_training - Step 2747: {'lr': 0.000499701267841771, 'samples': 527616, 'steps': 2747, 'loss/train': 1.746177077293396} 01/28/2022 16:26:31 - INFO - codeparrot_training - Step 2748: {'lr': 0.0004997004676479147, 'samples': 527808, 'steps': 2748, 'loss/train': 1.5588904023170471} 01/28/2022 16:26:35 - INFO - codeparrot_training - Step 2749: {'lr': 0.0004996996663844209, 'samples': 528000, 'steps': 2749, 'loss/train': 2.9353235363960266} 01/28/2022 16:26:41 - INFO - codeparrot_training - Step 2750: {'lr': 0.0004996988640512931, 'samples': 528192, 'steps': 2750, 'loss/train': 1.850597083568573} 01/28/2022 16:26:46 - INFO - codeparrot_training - Step 2751: {'lr': 0.0004996980606485346, 'samples': 528384, 'steps': 2751, 'loss/train': 1.9960981607437134} 01/28/2022 16:26:50 - INFO - codeparrot_training - Step 2752: {'lr': 0.0004996972561761489, 'samples': 528576, 'steps': 2752, 'loss/train': 2.591677665710449} 01/28/2022 16:26:54 - INFO - codeparrot_training - Step 2753: {'lr': 0.0004996964506341395, 'samples': 528768, 'steps': 2753, 'loss/train': 1.36570805311203} 01/28/2022 16:26:58 - INFO - codeparrot_training - Step 2754: {'lr': 0.0004996956440225098, 'samples': 528960, 'steps': 2754, 'loss/train': 2.0245930552482605} 01/28/2022 16:27:04 - INFO - codeparrot_training - Step 2755: {'lr': 0.0004996948363412631, 'samples': 529152, 'steps': 2755, 'loss/train': 2.110448956489563} 01/28/2022 16:27:08 - INFO - codeparrot_training - Step 2756: {'lr': 0.0004996940275904031, 'samples': 529344, 'steps': 2756, 'loss/train': 1.7926263213157654} 01/28/2022 16:27:12 - INFO - codeparrot_training - Step 2757: {'lr': 0.0004996932177699332, 'samples': 529536, 'steps': 2757, 'loss/train': 2.5705204010009766} 01/28/2022 16:27:16 - INFO - codeparrot_training - Step 2758: {'lr': 0.0004996924068798569, 'samples': 529728, 'steps': 2758, 'loss/train': 2.7287739515304565} 01/28/2022 16:27:21 - INFO - codeparrot_training - Step 2759: {'lr': 0.0004996915949201775, 'samples': 529920, 'steps': 2759, 'loss/train': 1.6445637345314026} 01/28/2022 16:27:27 - INFO - codeparrot_training - Step 2760: {'lr': 0.0004996907818908987, 'samples': 530112, 'steps': 2760, 'loss/train': 0.7681612372398376} 01/28/2022 16:27:31 - INFO - codeparrot_training - Step 2761: {'lr': 0.0004996899677920238, 'samples': 530304, 'steps': 2761, 'loss/train': 2.69989550113678} 01/28/2022 16:27:35 - INFO - codeparrot_training - Step 2762: {'lr': 0.0004996891526235564, 'samples': 530496, 'steps': 2762, 'loss/train': 1.901284396648407} 01/28/2022 16:27:39 - INFO - codeparrot_training - Step 2763: {'lr': 0.0004996883363854998, 'samples': 530688, 'steps': 2763, 'loss/train': 2.3511396646499634} 01/28/2022 16:27:44 - INFO - codeparrot_training - Step 2764: {'lr': 0.0004996875190778579, 'samples': 530880, 'steps': 2764, 'loss/train': 0.9045647084712982} 01/28/2022 16:27:49 - INFO - codeparrot_training - Step 2765: {'lr': 0.0004996867007006339, 'samples': 531072, 'steps': 2765, 'loss/train': 3.0353411436080933} 01/28/2022 16:27:53 - INFO - codeparrot_training - Step 2766: {'lr': 0.0004996858812538312, 'samples': 531264, 'steps': 2766, 'loss/train': 3.5546289682388306} 01/28/2022 16:27:57 - INFO - codeparrot_training - Step 2767: {'lr': 0.0004996850607374535, 'samples': 531456, 'steps': 2767, 'loss/train': 1.4684287011623383} 01/28/2022 16:28:01 - INFO - codeparrot_training - Step 2768: {'lr': 0.0004996842391515044, 'samples': 531648, 'steps': 2768, 'loss/train': 2.594182848930359} 01/28/2022 16:28:06 - INFO - codeparrot_training - Step 2769: {'lr': 0.0004996834164959872, 'samples': 531840, 'steps': 2769, 'loss/train': 2.4215848445892334} 01/28/2022 16:28:11 - INFO - codeparrot_training - Step 2770: {'lr': 0.0004996825927709056, 'samples': 532032, 'steps': 2770, 'loss/train': 1.8357607126235962} 01/28/2022 16:28:15 - INFO - codeparrot_training - Step 2771: {'lr': 0.0004996817679762631, 'samples': 532224, 'steps': 2771, 'loss/train': 1.7396194338798523} 01/28/2022 16:28:19 - INFO - codeparrot_training - Step 2772: {'lr': 0.000499680942112063, 'samples': 532416, 'steps': 2772, 'loss/train': 2.7076892852783203} 01/28/2022 16:28:24 - INFO - codeparrot_training - Step 2773: {'lr': 0.0004996801151783092, 'samples': 532608, 'steps': 2773, 'loss/train': 2.2883614897727966} 01/28/2022 16:28:28 - INFO - codeparrot_training - Step 2774: {'lr': 0.000499679287175005, 'samples': 532800, 'steps': 2774, 'loss/train': 2.1263174414634705} 01/28/2022 16:28:34 - INFO - codeparrot_training - Step 2775: {'lr': 0.000499678458102154, 'samples': 532992, 'steps': 2775, 'loss/train': 1.8826244473457336} 01/28/2022 16:28:38 - INFO - codeparrot_training - Step 2776: {'lr': 0.0004996776279597598, 'samples': 533184, 'steps': 2776, 'loss/train': 1.898419976234436} 01/28/2022 16:28:43 - INFO - codeparrot_training - Step 2777: {'lr': 0.0004996767967478259, 'samples': 533376, 'steps': 2777, 'loss/train': 1.9361404180526733} 01/28/2022 16:28:47 - INFO - codeparrot_training - Step 2778: {'lr': 0.0004996759644663559, 'samples': 533568, 'steps': 2778, 'loss/train': 2.1591153144836426} 01/28/2022 16:28:52 - INFO - codeparrot_training - Step 2779: {'lr': 0.0004996751311153535, 'samples': 533760, 'steps': 2779, 'loss/train': 1.4338898956775665} 01/28/2022 16:28:56 - INFO - codeparrot_training - Step 2780: {'lr': 0.0004996742966948219, 'samples': 533952, 'steps': 2780, 'loss/train': 1.3936613202095032} 01/28/2022 16:29:01 - INFO - codeparrot_training - Step 2781: {'lr': 0.000499673461204765, 'samples': 534144, 'steps': 2781, 'loss/train': 1.9776057600975037} 01/28/2022 16:29:05 - INFO - codeparrot_training - Step 2782: {'lr': 0.0004996726246451862, 'samples': 534336, 'steps': 2782, 'loss/train': 2.320054292678833} 01/28/2022 16:29:09 - INFO - codeparrot_training - Step 2783: {'lr': 0.0004996717870160892, 'samples': 534528, 'steps': 2783, 'loss/train': 1.738418161869049} 01/28/2022 16:29:14 - INFO - codeparrot_training - Step 2784: {'lr': 0.0004996709483174775, 'samples': 534720, 'steps': 2784, 'loss/train': 2.074714243412018} 01/28/2022 16:29:19 - INFO - codeparrot_training - Step 2785: {'lr': 0.0004996701085493547, 'samples': 534912, 'steps': 2785, 'loss/train': 2.499727427959442} 01/28/2022 16:29:23 - INFO - codeparrot_training - Step 2786: {'lr': 0.0004996692677117246, 'samples': 535104, 'steps': 2786, 'loss/train': 1.7445756196975708} 01/28/2022 16:29:27 - INFO - codeparrot_training - Step 2787: {'lr': 0.0004996684258045906, 'samples': 535296, 'steps': 2787, 'loss/train': 1.7865069508552551} 01/28/2022 16:29:31 - INFO - codeparrot_training - Step 2788: {'lr': 0.0004996675828279562, 'samples': 535488, 'steps': 2788, 'loss/train': 1.2505236268043518} 01/28/2022 16:29:37 - INFO - codeparrot_training - Step 2789: {'lr': 0.0004996667387818254, 'samples': 535680, 'steps': 2789, 'loss/train': 2.2369109988212585} 01/28/2022 16:29:41 - INFO - codeparrot_training - Step 2790: {'lr': 0.0004996658936662013, 'samples': 535872, 'steps': 2790, 'loss/train': 0.4596449285745621} 01/28/2022 16:29:45 - INFO - codeparrot_training - Step 2791: {'lr': 0.0004996650474810879, 'samples': 536064, 'steps': 2791, 'loss/train': 1.672118067741394} 01/28/2022 16:29:50 - INFO - codeparrot_training - Step 2792: {'lr': 0.0004996642002264887, 'samples': 536256, 'steps': 2792, 'loss/train': 2.161102831363678} 01/28/2022 16:29:54 - INFO - codeparrot_training - Step 2793: {'lr': 0.0004996633519024074, 'samples': 536448, 'steps': 2793, 'loss/train': 1.6042850017547607} 01/28/2022 16:29:58 - INFO - codeparrot_training - Step 2794: {'lr': 0.0004996625025088476, 'samples': 536640, 'steps': 2794, 'loss/train': 1.889855146408081} 01/28/2022 16:30:04 - INFO - codeparrot_training - Step 2795: {'lr': 0.0004996616520458128, 'samples': 536832, 'steps': 2795, 'loss/train': 2.655649244785309} 01/28/2022 16:30:08 - INFO - codeparrot_training - Step 2796: {'lr': 0.0004996608005133068, 'samples': 537024, 'steps': 2796, 'loss/train': 1.7159573435783386} 01/28/2022 16:30:13 - INFO - codeparrot_training - Step 2797: {'lr': 0.0004996599479113333, 'samples': 537216, 'steps': 2797, 'loss/train': 2.1065202355384827} 01/28/2022 16:30:17 - INFO - codeparrot_training - Step 2798: {'lr': 0.0004996590942398958, 'samples': 537408, 'steps': 2798, 'loss/train': 2.268958032131195} 01/28/2022 16:30:21 - INFO - codeparrot_training - Step 2799: {'lr': 0.0004996582394989979, 'samples': 537600, 'steps': 2799, 'loss/train': 1.97426837682724} 01/28/2022 16:30:26 - INFO - codeparrot_training - Step 2800: {'lr': 0.0004996573836886434, 'samples': 537792, 'steps': 2800, 'loss/train': 2.1956436038017273} 01/28/2022 16:30:31 - INFO - codeparrot_training - Step 2801: {'lr': 0.0004996565268088362, 'samples': 537984, 'steps': 2801, 'loss/train': 2.244502365589142} 01/28/2022 16:30:35 - INFO - codeparrot_training - Step 2802: {'lr': 0.0004996556688595794, 'samples': 538176, 'steps': 2802, 'loss/train': 2.234749138355255} 01/28/2022 16:30:39 - INFO - codeparrot_training - Step 2803: {'lr': 0.0004996548098408772, 'samples': 538368, 'steps': 2803, 'loss/train': 1.3311946392059326} 01/28/2022 16:30:43 - INFO - codeparrot_training - Step 2804: {'lr': 0.0004996539497527329, 'samples': 538560, 'steps': 2804, 'loss/train': 1.1796533167362213} 01/28/2022 16:30:49 - INFO - codeparrot_training - Step 2805: {'lr': 0.0004996530885951505, 'samples': 538752, 'steps': 2805, 'loss/train': 1.6457176208496094} 01/28/2022 16:30:54 - INFO - codeparrot_training - Step 2806: {'lr': 0.0004996522263681335, 'samples': 538944, 'steps': 2806, 'loss/train': 2.318039059638977} 01/28/2022 16:30:58 - INFO - codeparrot_training - Step 2807: {'lr': 0.0004996513630716856, 'samples': 539136, 'steps': 2807, 'loss/train': 2.153924524784088} 01/28/2022 16:31:02 - INFO - codeparrot_training - Step 2808: {'lr': 0.0004996504987058105, 'samples': 539328, 'steps': 2808, 'loss/train': 1.252970427274704} 01/28/2022 16:31:06 - INFO - codeparrot_training - Step 2809: {'lr': 0.000499649633270512, 'samples': 539520, 'steps': 2809, 'loss/train': 1.858011782169342} 01/28/2022 16:31:12 - INFO - codeparrot_training - Step 2810: {'lr': 0.0004996487667657938, 'samples': 539712, 'steps': 2810, 'loss/train': 1.3921073377132416} 01/28/2022 16:31:16 - INFO - codeparrot_training - Step 2811: {'lr': 0.0004996478991916595, 'samples': 539904, 'steps': 2811, 'loss/train': 2.180249869823456} 01/28/2022 16:31:21 - INFO - codeparrot_training - Step 2812: {'lr': 0.0004996470305481127, 'samples': 540096, 'steps': 2812, 'loss/train': 1.83964204788208} 01/28/2022 16:31:25 - INFO - codeparrot_training - Step 2813: {'lr': 0.0004996461608351575, 'samples': 540288, 'steps': 2813, 'loss/train': 2.277601718902588} 01/28/2022 16:31:29 - INFO - codeparrot_training - Step 2814: {'lr': 0.0004996452900527974, 'samples': 540480, 'steps': 2814, 'loss/train': 2.218710243701935} 01/28/2022 16:31:35 - INFO - codeparrot_training - Step 2815: {'lr': 0.0004996444182010361, 'samples': 540672, 'steps': 2815, 'loss/train': 2.3407756090164185} 01/28/2022 16:31:39 - INFO - codeparrot_training - Step 2816: {'lr': 0.0004996435452798775, 'samples': 540864, 'steps': 2816, 'loss/train': 1.8040637969970703} 01/28/2022 16:31:43 - INFO - codeparrot_training - Step 2817: {'lr': 0.000499642671289325, 'samples': 541056, 'steps': 2817, 'loss/train': 2.332928717136383} 01/28/2022 16:31:48 - INFO - codeparrot_training - Step 2818: {'lr': 0.0004996417962293828, 'samples': 541248, 'steps': 2818, 'loss/train': 2.488299250602722} 01/28/2022 16:31:52 - INFO - codeparrot_training - Step 2819: {'lr': 0.0004996409201000543, 'samples': 541440, 'steps': 2819, 'loss/train': 1.539486050605774} 01/28/2022 16:31:58 - INFO - codeparrot_training - Step 2820: {'lr': 0.0004996400429013434, 'samples': 541632, 'steps': 2820, 'loss/train': 1.9699162244796753} 01/28/2022 16:32:02 - INFO - codeparrot_training - Step 2821: {'lr': 0.0004996391646332537, 'samples': 541824, 'steps': 2821, 'loss/train': 1.7568830251693726} 01/28/2022 16:32:07 - INFO - codeparrot_training - Step 2822: {'lr': 0.0004996382852957892, 'samples': 542016, 'steps': 2822, 'loss/train': 1.0723114907741547} 01/28/2022 16:32:11 - INFO - codeparrot_training - Step 2823: {'lr': 0.0004996374048889536, 'samples': 542208, 'steps': 2823, 'loss/train': 1.8896563053131104} 01/28/2022 16:32:15 - INFO - codeparrot_training - Step 2824: {'lr': 0.0004996365234127506, 'samples': 542400, 'steps': 2824, 'loss/train': 2.1826520562171936} 01/28/2022 16:32:20 - INFO - codeparrot_training - Step 2825: {'lr': 0.000499635640867184, 'samples': 542592, 'steps': 2825, 'loss/train': 2.191642999649048} 01/28/2022 16:32:24 - INFO - codeparrot_training - Step 2826: {'lr': 0.0004996347572522575, 'samples': 542784, 'steps': 2826, 'loss/train': 3.231441020965576} 01/28/2022 16:32:29 - INFO - codeparrot_training - Step 2827: {'lr': 0.000499633872567975, 'samples': 542976, 'steps': 2827, 'loss/train': 2.1178232431411743} 01/28/2022 16:32:33 - INFO - codeparrot_training - Step 2828: {'lr': 0.0004996329868143404, 'samples': 543168, 'steps': 2828, 'loss/train': 1.8793314099311829} 01/28/2022 16:32:37 - INFO - codeparrot_training - Step 2829: {'lr': 0.0004996320999913572, 'samples': 543360, 'steps': 2829, 'loss/train': 2.484030783176422} 01/28/2022 16:32:42 - INFO - codeparrot_training - Step 2830: {'lr': 0.0004996312120990293, 'samples': 543552, 'steps': 2830, 'loss/train': 2.257442057132721} 01/28/2022 16:32:47 - INFO - codeparrot_training - Step 2831: {'lr': 0.0004996303231373607, 'samples': 543744, 'steps': 2831, 'loss/train': 2.03982013463974} 01/28/2022 16:32:51 - INFO - codeparrot_training - Step 2832: {'lr': 0.000499629433106355, 'samples': 543936, 'steps': 2832, 'loss/train': 1.5137275457382202} 01/28/2022 16:32:55 - INFO - codeparrot_training - Step 2833: {'lr': 0.000499628542006016, 'samples': 544128, 'steps': 2833, 'loss/train': 2.22450989484787} 01/28/2022 16:32:59 - INFO - codeparrot_training - Step 2834: {'lr': 0.0004996276498363477, 'samples': 544320, 'steps': 2834, 'loss/train': 2.285310924053192} 01/28/2022 16:33:05 - INFO - codeparrot_training - Step 2835: {'lr': 0.0004996267565973538, 'samples': 544512, 'steps': 2835, 'loss/train': 3.074856162071228} 01/28/2022 16:33:09 - INFO - codeparrot_training - Step 2836: {'lr': 0.0004996258622890381, 'samples': 544704, 'steps': 2836, 'loss/train': 1.7031387090682983} 01/28/2022 16:33:13 - INFO - codeparrot_training - Step 2837: {'lr': 0.0004996249669114045, 'samples': 544896, 'steps': 2837, 'loss/train': 1.4582170844078064} 01/28/2022 16:33:17 - INFO - codeparrot_training - Step 2838: {'lr': 0.0004996240704644568, 'samples': 545088, 'steps': 2838, 'loss/train': 2.256009042263031} 01/28/2022 16:33:22 - INFO - codeparrot_training - Step 2839: {'lr': 0.0004996231729481989, 'samples': 545280, 'steps': 2839, 'loss/train': 1.8009098768234253} 01/28/2022 16:33:28 - INFO - codeparrot_training - Step 2840: {'lr': 0.0004996222743626345, 'samples': 545472, 'steps': 2840, 'loss/train': 1.692776620388031} 01/28/2022 16:33:32 - INFO - codeparrot_training - Step 2841: {'lr': 0.0004996213747077675, 'samples': 545664, 'steps': 2841, 'loss/train': 1.974419116973877} 01/28/2022 16:33:36 - INFO - codeparrot_training - Step 2842: {'lr': 0.0004996204739836019, 'samples': 545856, 'steps': 2842, 'loss/train': 2.616937816143036} 01/28/2022 16:33:40 - INFO - codeparrot_training - Step 2843: {'lr': 0.0004996195721901415, 'samples': 546048, 'steps': 2843, 'loss/train': 1.3022858798503876} 01/28/2022 16:33:45 - INFO - codeparrot_training - Step 2844: {'lr': 0.00049961866932739, 'samples': 546240, 'steps': 2844, 'loss/train': 1.7318195700645447} 01/28/2022 16:33:50 - INFO - codeparrot_training - Step 2845: {'lr': 0.0004996177653953514, 'samples': 546432, 'steps': 2845, 'loss/train': 1.9821351170539856} 01/28/2022 16:33:54 - INFO - codeparrot_training - Step 2846: {'lr': 0.0004996168603940296, 'samples': 546624, 'steps': 2846, 'loss/train': 2.3859381079673767} 01/28/2022 16:33:58 - INFO - codeparrot_training - Step 2847: {'lr': 0.0004996159543234285, 'samples': 546816, 'steps': 2847, 'loss/train': 1.6182310581207275} 01/28/2022 16:34:03 - INFO - codeparrot_training - Step 2848: {'lr': 0.0004996150471835518, 'samples': 547008, 'steps': 2848, 'loss/train': 1.893602192401886} 01/28/2022 16:34:07 - INFO - codeparrot_training - Step 2849: {'lr': 0.0004996141389744035, 'samples': 547200, 'steps': 2849, 'loss/train': 2.563316524028778} 01/28/2022 16:34:12 - INFO - codeparrot_training - Step 2850: {'lr': 0.0004996132296959876, 'samples': 547392, 'steps': 2850, 'loss/train': 2.3209891319274902} 01/28/2022 16:34:17 - INFO - codeparrot_training - Step 2851: {'lr': 0.0004996123193483076, 'samples': 547584, 'steps': 2851, 'loss/train': 1.3020582497119904} 01/28/2022 16:34:21 - INFO - codeparrot_training - Step 2852: {'lr': 0.000499611407931368, 'samples': 547776, 'steps': 2852, 'loss/train': 2.3723625540733337} 01/28/2022 16:34:25 - INFO - codeparrot_training - Step 2853: {'lr': 0.0004996104954451722, 'samples': 547968, 'steps': 2853, 'loss/train': 1.5510971546173096} 01/28/2022 16:34:29 - INFO - codeparrot_training - Step 2854: {'lr': 0.0004996095818897245, 'samples': 548160, 'steps': 2854, 'loss/train': 2.4096415042877197} 01/28/2022 16:34:36 - INFO - codeparrot_training - Step 2855: {'lr': 0.0004996086672650284, 'samples': 548352, 'steps': 2855, 'loss/train': 1.9392317533493042} 01/28/2022 16:34:40 - INFO - codeparrot_training - Step 2856: {'lr': 0.0004996077515710881, 'samples': 548544, 'steps': 2856, 'loss/train': 0.9726923704147339} 01/28/2022 16:34:44 - INFO - codeparrot_training - Step 2857: {'lr': 0.0004996068348079075, 'samples': 548736, 'steps': 2857, 'loss/train': 1.821592926979065} 01/28/2022 16:34:48 - INFO - codeparrot_training - Step 2858: {'lr': 0.0004996059169754904, 'samples': 548928, 'steps': 2858, 'loss/train': 1.7374666929244995} 01/28/2022 16:34:54 - INFO - codeparrot_training - Step 2859: {'lr': 0.0004996049980738409, 'samples': 549120, 'steps': 2859, 'loss/train': 1.4436023533344269} 01/28/2022 16:34:58 - INFO - codeparrot_training - Step 2860: {'lr': 0.0004996040781029629, 'samples': 549312, 'steps': 2860, 'loss/train': 1.027565985918045} 01/28/2022 16:35:02 - INFO - codeparrot_training - Step 2861: {'lr': 0.00049960315706286, 'samples': 549504, 'steps': 2861, 'loss/train': 2.884691834449768} 01/28/2022 16:35:07 - INFO - codeparrot_training - Step 2862: {'lr': 0.0004996022349535367, 'samples': 549696, 'steps': 2862, 'loss/train': 1.6841740608215332} 01/28/2022 16:35:11 - INFO - codeparrot_training - Step 2863: {'lr': 0.0004996013117749967, 'samples': 549888, 'steps': 2863, 'loss/train': 2.9225317239761353} 01/28/2022 16:35:17 - INFO - codeparrot_training - Step 2864: {'lr': 0.0004996003875272438, 'samples': 550080, 'steps': 2864, 'loss/train': 1.0054290890693665} 01/28/2022 16:35:21 - INFO - codeparrot_training - Step 2865: {'lr': 0.0004995994622102821, 'samples': 550272, 'steps': 2865, 'loss/train': 1.998960256576538} 01/28/2022 16:35:25 - INFO - codeparrot_training - Step 2866: {'lr': 0.0004995985358241156, 'samples': 550464, 'steps': 2866, 'loss/train': 1.9550630450248718} 01/28/2022 16:35:30 - INFO - codeparrot_training - Step 2867: {'lr': 0.0004995976083687482, 'samples': 550656, 'steps': 2867, 'loss/train': 1.3940053582191467} 01/28/2022 16:35:34 - INFO - codeparrot_training - Step 2868: {'lr': 0.000499596679844184, 'samples': 550848, 'steps': 2868, 'loss/train': 2.4737293124198914} 01/28/2022 16:35:39 - INFO - codeparrot_training - Step 2869: {'lr': 0.0004995957502504268, 'samples': 551040, 'steps': 2869, 'loss/train': 2.221999704837799} 01/28/2022 16:35:43 - INFO - codeparrot_training - Step 2870: {'lr': 0.0004995948195874807, 'samples': 551232, 'steps': 2870, 'loss/train': 1.8506085276603699} 01/28/2022 16:35:48 - INFO - codeparrot_training - Step 2871: {'lr': 0.0004995938878553496, 'samples': 551424, 'steps': 2871, 'loss/train': 1.3165763318538666} 01/28/2022 16:35:52 - INFO - codeparrot_training - Step 2872: {'lr': 0.0004995929550540376, 'samples': 551616, 'steps': 2872, 'loss/train': 1.3801055252552032} 01/28/2022 16:35:56 - INFO - codeparrot_training - Step 2873: {'lr': 0.0004995920211835485, 'samples': 551808, 'steps': 2873, 'loss/train': 2.508486270904541} 01/28/2022 16:36:01 - INFO - codeparrot_training - Step 2874: {'lr': 0.0004995910862438866, 'samples': 552000, 'steps': 2874, 'loss/train': 2.3562939763069153} 01/28/2022 16:36:06 - INFO - codeparrot_training - Step 2875: {'lr': 0.0004995901502350556, 'samples': 552192, 'steps': 2875, 'loss/train': 2.3556865453720093} 01/28/2022 16:36:10 - INFO - codeparrot_training - Step 2876: {'lr': 0.0004995892131570598, 'samples': 552384, 'steps': 2876, 'loss/train': 0.27973971515893936} 01/28/2022 16:36:14 - INFO - codeparrot_training - Step 2877: {'lr': 0.0004995882750099029, 'samples': 552576, 'steps': 2877, 'loss/train': 1.0645923614501953} 01/28/2022 16:36:18 - INFO - codeparrot_training - Step 2878: {'lr': 0.0004995873357935892, 'samples': 552768, 'steps': 2878, 'loss/train': 1.1502640843391418} 01/28/2022 16:36:23 - INFO - codeparrot_training - Step 2879: {'lr': 0.0004995863955081226, 'samples': 552960, 'steps': 2879, 'loss/train': 1.9027118682861328} 01/28/2022 16:36:29 - INFO - codeparrot_training - Step 2880: {'lr': 0.0004995854541535071, 'samples': 553152, 'steps': 2880, 'loss/train': 2.1384336948394775} 01/28/2022 16:36:33 - INFO - codeparrot_training - Step 2881: {'lr': 0.0004995845117297468, 'samples': 553344, 'steps': 2881, 'loss/train': 0.3919220119714737} 01/28/2022 16:36:37 - INFO - codeparrot_training - Step 2882: {'lr': 0.0004995835682368457, 'samples': 553536, 'steps': 2882, 'loss/train': 1.85762357711792} 01/28/2022 16:36:42 - INFO - codeparrot_training - Step 2883: {'lr': 0.0004995826236748078, 'samples': 553728, 'steps': 2883, 'loss/train': 2.5810643434524536} 01/28/2022 16:36:47 - INFO - codeparrot_training - Step 2884: {'lr': 0.0004995816780436372, 'samples': 553920, 'steps': 2884, 'loss/train': 2.1905052065849304} 01/28/2022 16:36:51 - INFO - codeparrot_training - Step 2885: {'lr': 0.0004995807313433379, 'samples': 554112, 'steps': 2885, 'loss/train': 1.9962236881256104} 01/28/2022 16:36:55 - INFO - codeparrot_training - Step 2886: {'lr': 0.0004995797835739141, 'samples': 554304, 'steps': 2886, 'loss/train': 2.0348832607269287} 01/28/2022 16:37:00 - INFO - codeparrot_training - Step 2887: {'lr': 0.0004995788347353697, 'samples': 554496, 'steps': 2887, 'loss/train': 1.9146262407302856} 01/28/2022 16:37:04 - INFO - codeparrot_training - Step 2888: {'lr': 0.0004995778848277088, 'samples': 554688, 'steps': 2888, 'loss/train': 3.1724460124969482} 01/28/2022 16:37:09 - INFO - codeparrot_training - Step 2889: {'lr': 0.0004995769338509357, 'samples': 554880, 'steps': 2889, 'loss/train': 2.254588544368744} 01/28/2022 16:37:13 - INFO - codeparrot_training - Step 2890: {'lr': 0.000499575981805054, 'samples': 555072, 'steps': 2890, 'loss/train': 0.7363916784524918} 01/28/2022 16:37:18 - INFO - codeparrot_training - Step 2891: {'lr': 0.000499575028690068, 'samples': 555264, 'steps': 2891, 'loss/train': 2.39937686920166} 01/28/2022 16:37:22 - INFO - codeparrot_training - Step 2892: {'lr': 0.000499574074505982, 'samples': 555456, 'steps': 2892, 'loss/train': 2.5374906063079834} 01/28/2022 16:37:26 - INFO - codeparrot_training - Step 2893: {'lr': 0.0004995731192527999, 'samples': 555648, 'steps': 2893, 'loss/train': 2.1845374703407288} 01/28/2022 16:37:32 - INFO - codeparrot_training - Step 2894: {'lr': 0.0004995721629305258, 'samples': 555840, 'steps': 2894, 'loss/train': 2.9041205048561096} 01/28/2022 16:37:37 - INFO - codeparrot_training - Step 2895: {'lr': 0.0004995712055391638, 'samples': 556032, 'steps': 2895, 'loss/train': 2.1397064924240112} 01/28/2022 16:37:41 - INFO - codeparrot_training - Step 2896: {'lr': 0.000499570247078718, 'samples': 556224, 'steps': 2896, 'loss/train': 2.179419457912445} 01/28/2022 16:37:45 - INFO - codeparrot_training - Step 2897: {'lr': 0.0004995692875491925, 'samples': 556416, 'steps': 2897, 'loss/train': 1.389340728521347} 01/28/2022 16:37:49 - INFO - codeparrot_training - Step 2898: {'lr': 0.0004995683269505914, 'samples': 556608, 'steps': 2898, 'loss/train': 1.2850660979747772} 01/28/2022 16:37:55 - INFO - codeparrot_training - Step 2899: {'lr': 0.000499567365282919, 'samples': 556800, 'steps': 2899, 'loss/train': 2.2509841918945312} 01/28/2022 16:37:59 - INFO - codeparrot_training - Step 2900: {'lr': 0.000499566402546179, 'samples': 556992, 'steps': 2900, 'loss/train': 2.5901012420654297} 01/28/2022 16:38:03 - INFO - codeparrot_training - Step 2901: {'lr': 0.0004995654387403758, 'samples': 557184, 'steps': 2901, 'loss/train': 2.24283367395401} 01/28/2022 16:38:07 - INFO - codeparrot_training - Step 2902: {'lr': 0.0004995644738655136, 'samples': 557376, 'steps': 2902, 'loss/train': 1.1247256994247437} 01/28/2022 16:38:12 - INFO - codeparrot_training - Step 2903: {'lr': 0.0004995635079215965, 'samples': 557568, 'steps': 2903, 'loss/train': 2.2556771636009216} 01/28/2022 16:38:17 - INFO - codeparrot_training - Step 2904: {'lr': 0.0004995625409086285, 'samples': 557760, 'steps': 2904, 'loss/train': 1.742056667804718} 01/28/2022 16:38:21 - INFO - codeparrot_training - Step 2905: {'lr': 0.0004995615728266138, 'samples': 557952, 'steps': 2905, 'loss/train': 2.373321533203125} 01/28/2022 16:38:26 - INFO - codeparrot_training - Step 2906: {'lr': 0.0004995606036755566, 'samples': 558144, 'steps': 2906, 'loss/train': 2.9740211963653564} 01/28/2022 16:38:30 - INFO - codeparrot_training - Step 2907: {'lr': 0.000499559633455461, 'samples': 558336, 'steps': 2907, 'loss/train': 2.284411668777466} 01/28/2022 16:38:34 - INFO - codeparrot_training - Step 2908: {'lr': 0.0004995586621663312, 'samples': 558528, 'steps': 2908, 'loss/train': 2.9178099632263184} 01/28/2022 16:38:40 - INFO - codeparrot_training - Step 2909: {'lr': 0.0004995576898081713, 'samples': 558720, 'steps': 2909, 'loss/train': 1.858697533607483} 01/28/2022 16:38:45 - INFO - codeparrot_training - Step 2910: {'lr': 0.0004995567163809855, 'samples': 558912, 'steps': 2910, 'loss/train': 2.4842689633369446} 01/28/2022 16:38:49 - INFO - codeparrot_training - Step 2911: {'lr': 0.000499555741884778, 'samples': 559104, 'steps': 2911, 'loss/train': 0.9934489130973816} 01/28/2022 16:38:53 - INFO - codeparrot_training - Step 2912: {'lr': 0.000499554766319553, 'samples': 559296, 'steps': 2912, 'loss/train': 2.091419756412506} 01/28/2022 16:38:57 - INFO - codeparrot_training - Step 2913: {'lr': 0.0004995537896853146, 'samples': 559488, 'steps': 2913, 'loss/train': 1.7886933088302612} 01/28/2022 16:39:04 - INFO - codeparrot_training - Step 2914: {'lr': 0.0004995528119820669, 'samples': 559680, 'steps': 2914, 'loss/train': 1.8129297494888306} 01/28/2022 16:39:08 - INFO - codeparrot_training - Step 2915: {'lr': 0.0004995518332098143, 'samples': 559872, 'steps': 2915, 'loss/train': 1.4002951383590698} 01/28/2022 16:39:12 - INFO - codeparrot_training - Step 2916: {'lr': 0.0004995508533685608, 'samples': 560064, 'steps': 2916, 'loss/train': 1.9801769256591797} 01/28/2022 16:39:16 - INFO - codeparrot_training - Step 2917: {'lr': 0.0004995498724583107, 'samples': 560256, 'steps': 2917, 'loss/train': 1.5900921821594238} 01/28/2022 16:39:21 - INFO - codeparrot_training - Step 2918: {'lr': 0.0004995488904790682, 'samples': 560448, 'steps': 2918, 'loss/train': 2.5114728212356567} 01/28/2022 16:39:25 - INFO - codeparrot_training - Step 2919: {'lr': 0.0004995479074308375, 'samples': 560640, 'steps': 2919, 'loss/train': 2.1310940980911255} 01/28/2022 16:39:29 - INFO - codeparrot_training - Step 2920: {'lr': 0.0004995469233136228, 'samples': 560832, 'steps': 2920, 'loss/train': 1.1199265122413635} 01/28/2022 16:39:35 - INFO - codeparrot_training - Step 2921: {'lr': 0.0004995459381274284, 'samples': 561024, 'steps': 2921, 'loss/train': 1.8998326063156128} 01/28/2022 16:39:39 - INFO - codeparrot_training - Step 2922: {'lr': 0.0004995449518722584, 'samples': 561216, 'steps': 2922, 'loss/train': 0.7439377158880234} 01/28/2022 16:39:43 - INFO - codeparrot_training - Step 2923: {'lr': 0.000499543964548117, 'samples': 561408, 'steps': 2923, 'loss/train': 1.9861809611320496} 01/28/2022 16:39:47 - INFO - codeparrot_training - Step 2924: {'lr': 0.0004995429761550086, 'samples': 561600, 'steps': 2924, 'loss/train': 1.6547080278396606} 01/28/2022 16:39:52 - INFO - codeparrot_training - Step 2925: {'lr': 0.0004995419866929373, 'samples': 561792, 'steps': 2925, 'loss/train': 2.654215693473816} 01/28/2022 16:39:58 - INFO - codeparrot_training - Step 2926: {'lr': 0.0004995409961619073, 'samples': 561984, 'steps': 2926, 'loss/train': 1.7043424844741821} 01/28/2022 16:40:02 - INFO - codeparrot_training - Step 2927: {'lr': 0.0004995400045619229, 'samples': 562176, 'steps': 2927, 'loss/train': 1.8829516768455505} 01/28/2022 16:40:06 - INFO - codeparrot_training - Step 2928: {'lr': 0.0004995390118929885, 'samples': 562368, 'steps': 2928, 'loss/train': 1.5804103016853333} 01/28/2022 16:40:10 - INFO - codeparrot_training - Step 2929: {'lr': 0.0004995380181551081, 'samples': 562560, 'steps': 2929, 'loss/train': 2.258978247642517} 01/28/2022 16:40:15 - INFO - codeparrot_training - Step 2930: {'lr': 0.0004995370233482861, 'samples': 562752, 'steps': 2930, 'loss/train': 1.3215689063072205} 01/28/2022 16:40:20 - INFO - codeparrot_training - Step 2931: {'lr': 0.0004995360274725267, 'samples': 562944, 'steps': 2931, 'loss/train': 2.5808477997779846} 01/28/2022 16:40:24 - INFO - codeparrot_training - Step 2932: {'lr': 0.0004995350305278342, 'samples': 563136, 'steps': 2932, 'loss/train': 1.2689289450645447} 01/28/2022 16:40:28 - INFO - codeparrot_training - Step 2933: {'lr': 0.0004995340325142128, 'samples': 563328, 'steps': 2933, 'loss/train': 2.3649179935455322} 01/28/2022 16:40:33 - INFO - codeparrot_training - Step 2934: {'lr': 0.000499533033431667, 'samples': 563520, 'steps': 2934, 'loss/train': 1.8373205065727234} 01/28/2022 16:40:39 - INFO - codeparrot_training - Step 2935: {'lr': 0.0004995320332802008, 'samples': 563712, 'steps': 2935, 'loss/train': 1.4113383293151855} 01/28/2022 16:40:43 - INFO - codeparrot_training - Step 2936: {'lr': 0.0004995310320598187, 'samples': 563904, 'steps': 2936, 'loss/train': 1.8881356716156006} 01/28/2022 16:40:47 - INFO - codeparrot_training - Step 2937: {'lr': 0.0004995300297705248, 'samples': 564096, 'steps': 2937, 'loss/train': 0.4425578713417053} 01/28/2022 16:40:52 - INFO - codeparrot_training - Step 2938: {'lr': 0.0004995290264123235, 'samples': 564288, 'steps': 2938, 'loss/train': 2.3556191325187683} 01/28/2022 16:40:56 - INFO - codeparrot_training - Step 2939: {'lr': 0.0004995280219852192, 'samples': 564480, 'steps': 2939, 'loss/train': 2.3133543133735657} 01/28/2022 16:41:01 - INFO - codeparrot_training - Step 2940: {'lr': 0.000499527016489216, 'samples': 564672, 'steps': 2940, 'loss/train': 1.663754403591156} 01/28/2022 16:41:05 - INFO - codeparrot_training - Step 2941: {'lr': 0.0004995260099243182, 'samples': 564864, 'steps': 2941, 'loss/train': 1.7597651481628418} 01/28/2022 16:41:10 - INFO - codeparrot_training - Step 2942: {'lr': 0.0004995250022905303, 'samples': 565056, 'steps': 2942, 'loss/train': 1.3473418056964874} 01/28/2022 16:41:14 - INFO - codeparrot_training - Step 2943: {'lr': 0.0004995239935878565, 'samples': 565248, 'steps': 2943, 'loss/train': 2.399268329143524} 01/28/2022 16:41:18 - INFO - codeparrot_training - Step 2944: {'lr': 0.0004995229838163012, 'samples': 565440, 'steps': 2944, 'loss/train': 1.8467538356781006} 01/28/2022 16:41:23 - INFO - codeparrot_training - Step 2945: {'lr': 0.0004995219729758687, 'samples': 565632, 'steps': 2945, 'loss/train': 1.6483824849128723} 01/28/2022 16:41:28 - INFO - codeparrot_training - Step 2946: {'lr': 0.0004995209610665632, 'samples': 565824, 'steps': 2946, 'loss/train': 2.2865456342697144} 01/28/2022 16:41:32 - INFO - codeparrot_training - Step 2947: {'lr': 0.0004995199480883892, 'samples': 566016, 'steps': 2947, 'loss/train': 1.2955635488033295} 01/28/2022 16:41:36 - INFO - codeparrot_training - Step 2948: {'lr': 0.0004995189340413509, 'samples': 566208, 'steps': 2948, 'loss/train': 2.830091893672943} 01/28/2022 16:41:40 - INFO - codeparrot_training - Step 2949: {'lr': 0.0004995179189254528, 'samples': 566400, 'steps': 2949, 'loss/train': 2.2014350295066833} 01/28/2022 16:41:46 - INFO - codeparrot_training - Step 2950: {'lr': 0.000499516902740699, 'samples': 566592, 'steps': 2950, 'loss/train': 2.2442763447761536} 01/28/2022 16:41:50 - INFO - codeparrot_training - Step 2951: {'lr': 0.0004995158854870942, 'samples': 566784, 'steps': 2951, 'loss/train': 1.362839788198471} 01/28/2022 16:41:54 - INFO - codeparrot_training - Step 2952: {'lr': 0.0004995148671646426, 'samples': 566976, 'steps': 2952, 'loss/train': 2.4621355533599854} 01/28/2022 16:41:58 - INFO - codeparrot_training - Step 2953: {'lr': 0.0004995138477733484, 'samples': 567168, 'steps': 2953, 'loss/train': 2.161022901535034} 01/28/2022 16:42:03 - INFO - codeparrot_training - Step 2954: {'lr': 0.0004995128273132161, 'samples': 567360, 'steps': 2954, 'loss/train': 2.102894604206085} 01/28/2022 16:42:09 - INFO - codeparrot_training - Step 2955: {'lr': 0.0004995118057842502, 'samples': 567552, 'steps': 2955, 'loss/train': 1.972687840461731} 01/28/2022 16:42:13 - INFO - codeparrot_training - Step 2956: {'lr': 0.0004995107831864549, 'samples': 567744, 'steps': 2956, 'loss/train': 1.8054664134979248} 01/28/2022 16:42:18 - INFO - codeparrot_training - Step 2957: {'lr': 0.0004995097595198346, 'samples': 567936, 'steps': 2957, 'loss/train': 2.433100461959839} 01/28/2022 16:42:22 - INFO - codeparrot_training - Step 2958: {'lr': 0.0004995087347843938, 'samples': 568128, 'steps': 2958, 'loss/train': 1.718910813331604} 01/28/2022 16:42:26 - INFO - codeparrot_training - Step 2959: {'lr': 0.0004995077089801368, 'samples': 568320, 'steps': 2959, 'loss/train': 2.1867674589157104} 01/28/2022 16:42:32 - INFO - codeparrot_training - Step 2960: {'lr': 0.0004995066821070679, 'samples': 568512, 'steps': 2960, 'loss/train': 2.1810414791107178} 01/28/2022 16:42:36 - INFO - codeparrot_training - Step 2961: {'lr': 0.0004995056541651917, 'samples': 568704, 'steps': 2961, 'loss/train': 0.566377803683281} 01/28/2022 16:42:40 - INFO - codeparrot_training - Step 2962: {'lr': 0.0004995046251545125, 'samples': 568896, 'steps': 2962, 'loss/train': 0.9904441237449646} 01/28/2022 16:42:45 - INFO - codeparrot_training - Step 2963: {'lr': 0.0004995035950750346, 'samples': 569088, 'steps': 2963, 'loss/train': 2.7370609045028687} 01/28/2022 16:42:49 - INFO - codeparrot_training - Step 2964: {'lr': 0.0004995025639267627, 'samples': 569280, 'steps': 2964, 'loss/train': 2.1700819730758667} 01/28/2022 16:42:54 - INFO - codeparrot_training - Step 2965: {'lr': 0.0004995015317097009, 'samples': 569472, 'steps': 2965, 'loss/train': 1.7794541716575623} 01/28/2022 16:42:58 - INFO - codeparrot_training - Step 2966: {'lr': 0.0004995004984238538, 'samples': 569664, 'steps': 2966, 'loss/train': 1.884186565876007} 01/28/2022 16:43:03 - INFO - codeparrot_training - Step 2967: {'lr': 0.0004994994640692258, 'samples': 569856, 'steps': 2967, 'loss/train': 2.7323813438415527} 01/28/2022 16:43:07 - INFO - codeparrot_training - Step 2968: {'lr': 0.0004994984286458213, 'samples': 570048, 'steps': 2968, 'loss/train': 2.0588809847831726} 01/28/2022 16:43:11 - INFO - codeparrot_training - Step 2969: {'lr': 0.0004994973921536447, 'samples': 570240, 'steps': 2969, 'loss/train': 1.8983049988746643} 01/28/2022 16:43:17 - INFO - codeparrot_training - Step 2970: {'lr': 0.0004994963545927006, 'samples': 570432, 'steps': 2970, 'loss/train': 1.9048671126365662} 01/28/2022 16:43:22 - INFO - codeparrot_training - Step 2971: {'lr': 0.0004994953159629934, 'samples': 570624, 'steps': 2971, 'loss/train': 2.088205397129059} 01/28/2022 16:43:26 - INFO - codeparrot_training - Step 2972: {'lr': 0.0004994942762645274, 'samples': 570816, 'steps': 2972, 'loss/train': 2.490156412124634} 01/28/2022 16:43:30 - INFO - codeparrot_training - Step 2973: {'lr': 0.000499493235497307, 'samples': 571008, 'steps': 2973, 'loss/train': 1.9818339943885803} 01/28/2022 16:43:35 - INFO - codeparrot_training - Step 2974: {'lr': 0.000499492193661337, 'samples': 571200, 'steps': 2974, 'loss/train': 1.7155131697654724} 01/28/2022 16:43:40 - INFO - codeparrot_training - Step 2975: {'lr': 0.0004994911507566216, 'samples': 571392, 'steps': 2975, 'loss/train': 1.9309676885604858} 01/28/2022 16:43:44 - INFO - codeparrot_training - Step 2976: {'lr': 0.0004994901067831654, 'samples': 571584, 'steps': 2976, 'loss/train': 0.5065793395042419} 01/28/2022 16:43:48 - INFO - codeparrot_training - Step 2977: {'lr': 0.0004994890617409728, 'samples': 571776, 'steps': 2977, 'loss/train': 1.9305151104927063} 01/28/2022 16:43:53 - INFO - codeparrot_training - Step 2978: {'lr': 0.0004994880156300482, 'samples': 571968, 'steps': 2978, 'loss/train': 2.1882371306419373} 01/28/2022 16:43:57 - INFO - codeparrot_training - Step 2979: {'lr': 0.0004994869684503962, 'samples': 572160, 'steps': 2979, 'loss/train': 1.6841468811035156} 01/28/2022 16:44:03 - INFO - codeparrot_training - Step 2980: {'lr': 0.0004994859202020212, 'samples': 572352, 'steps': 2980, 'loss/train': 2.3812195658683777} 01/28/2022 16:44:07 - INFO - codeparrot_training - Step 2981: {'lr': 0.0004994848708849279, 'samples': 572544, 'steps': 2981, 'loss/train': 1.8124891519546509} 01/28/2022 16:44:11 - INFO - codeparrot_training - Step 2982: {'lr': 0.0004994838204991205, 'samples': 572736, 'steps': 2982, 'loss/train': 1.6956626772880554} 01/28/2022 16:44:16 - INFO - codeparrot_training - Step 2983: {'lr': 0.0004994827690446036, 'samples': 572928, 'steps': 2983, 'loss/train': 1.8358270525932312} 01/28/2022 16:44:20 - INFO - codeparrot_training - Step 2984: {'lr': 0.0004994817165213817, 'samples': 573120, 'steps': 2984, 'loss/train': 2.462141275405884} 01/28/2022 16:44:25 - INFO - codeparrot_training - Step 2985: {'lr': 0.0004994806629294594, 'samples': 573312, 'steps': 2985, 'loss/train': 3.1178126335144043} 01/28/2022 16:44:30 - INFO - codeparrot_training - Step 2986: {'lr': 0.0004994796082688413, 'samples': 573504, 'steps': 2986, 'loss/train': 1.1809747517108917} 01/28/2022 16:44:34 - INFO - codeparrot_training - Step 2987: {'lr': 0.0004994785525395316, 'samples': 573696, 'steps': 2987, 'loss/train': 2.001829147338867} 01/28/2022 16:44:38 - INFO - codeparrot_training - Step 2988: {'lr': 0.0004994774957415351, 'samples': 573888, 'steps': 2988, 'loss/train': 1.8745642304420471} 01/28/2022 16:44:42 - INFO - codeparrot_training - Step 2989: {'lr': 0.0004994764378748562, 'samples': 574080, 'steps': 2989, 'loss/train': 2.3776538372039795} 01/28/2022 16:44:48 - INFO - codeparrot_training - Step 2990: {'lr': 0.0004994753789394994, 'samples': 574272, 'steps': 2990, 'loss/train': 2.663105607032776} 01/28/2022 16:44:52 - INFO - codeparrot_training - Step 2991: {'lr': 0.0004994743189354694, 'samples': 574464, 'steps': 2991, 'loss/train': 2.436762750148773} 01/28/2022 16:44:56 - INFO - codeparrot_training - Step 2992: {'lr': 0.0004994732578627706, 'samples': 574656, 'steps': 2992, 'loss/train': 2.058025896549225} 01/28/2022 16:45:00 - INFO - codeparrot_training - Step 2993: {'lr': 0.0004994721957214076, 'samples': 574848, 'steps': 2993, 'loss/train': 1.4403668940067291} 01/28/2022 16:45:05 - INFO - codeparrot_training - Step 2994: {'lr': 0.0004994711325113849, 'samples': 575040, 'steps': 2994, 'loss/train': 1.9889464974403381} 01/28/2022 16:45:10 - INFO - codeparrot_training - Step 2995: {'lr': 0.000499470068232707, 'samples': 575232, 'steps': 2995, 'loss/train': 2.2382195591926575} 01/28/2022 16:45:14 - INFO - codeparrot_training - Step 2996: {'lr': 0.0004994690028853787, 'samples': 575424, 'steps': 2996, 'loss/train': 2.3541966676712036} 01/28/2022 16:45:18 - INFO - codeparrot_training - Step 2997: {'lr': 0.0004994679364694043, 'samples': 575616, 'steps': 2997, 'loss/train': 1.215524822473526} 01/28/2022 16:45:23 - INFO - codeparrot_training - Step 2998: {'lr': 0.0004994668689847885, 'samples': 575808, 'steps': 2998, 'loss/train': 1.6470564007759094} 01/28/2022 16:45:27 - INFO - codeparrot_training - Step 2999: {'lr': 0.0004994658004315358, 'samples': 576000, 'steps': 2999, 'loss/train': 1.9825705289840698} 01/28/2022 16:45:33 - INFO - codeparrot_training - Step 3000: {'lr': 0.0004994647308096509, 'samples': 576192, 'steps': 3000, 'loss/train': 2.265605092048645} 01/28/2022 16:45:37 - INFO - codeparrot_training - Step 3001: {'lr': 0.0004994636601191383, 'samples': 576384, 'steps': 3001, 'loss/train': 2.125622570514679} 01/28/2022 16:45:41 - INFO - codeparrot_training - Step 3002: {'lr': 0.0004994625883600025, 'samples': 576576, 'steps': 3002, 'loss/train': 1.7405256628990173} 01/28/2022 16:45:46 - INFO - codeparrot_training - Step 3003: {'lr': 0.0004994615155322483, 'samples': 576768, 'steps': 3003, 'loss/train': 2.5948145985603333} 01/28/2022 16:45:50 - INFO - codeparrot_training - Step 3004: {'lr': 0.0004994604416358801, 'samples': 576960, 'steps': 3004, 'loss/train': 1.7635526061058044} 01/28/2022 16:45:55 - INFO - codeparrot_training - Step 3005: {'lr': 0.0004994593666709027, 'samples': 577152, 'steps': 3005, 'loss/train': 2.291983723640442} 01/28/2022 16:45:59 - INFO - codeparrot_training - Step 3006: {'lr': 0.0004994582906373205, 'samples': 577344, 'steps': 3006, 'loss/train': 1.256087601184845} 01/28/2022 16:46:04 - INFO - codeparrot_training - Step 3007: {'lr': 0.0004994572135351382, 'samples': 577536, 'steps': 3007, 'loss/train': 1.9366126656532288} 01/28/2022 16:46:08 - INFO - codeparrot_training - Step 3008: {'lr': 0.0004994561353643604, 'samples': 577728, 'steps': 3008, 'loss/train': 1.8658323884010315} 01/28/2022 16:46:12 - INFO - codeparrot_training - Step 3009: {'lr': 0.0004994550561249917, 'samples': 577920, 'steps': 3009, 'loss/train': 1.9148864150047302} 01/28/2022 16:46:17 - INFO - codeparrot_training - Step 3010: {'lr': 0.0004994539758170367, 'samples': 578112, 'steps': 3010, 'loss/train': 2.2906413674354553} 01/28/2022 16:46:22 - INFO - codeparrot_training - Step 3011: {'lr': 0.0004994528944405002, 'samples': 578304, 'steps': 3011, 'loss/train': 2.2445104122161865} 01/28/2022 16:46:26 - INFO - codeparrot_training - Step 3012: {'lr': 0.0004994518119953867, 'samples': 578496, 'steps': 3012, 'loss/train': 2.2146998047828674} 01/28/2022 16:46:30 - INFO - codeparrot_training - Step 3013: {'lr': 0.0004994507284817009, 'samples': 578688, 'steps': 3013, 'loss/train': 1.9498173594474792} 01/28/2022 16:46:34 - INFO - codeparrot_training - Step 3014: {'lr': 0.0004994496438994472, 'samples': 578880, 'steps': 3014, 'loss/train': 1.6114731431007385} 01/28/2022 16:46:41 - INFO - codeparrot_training - Step 3015: {'lr': 0.0004994485582486306, 'samples': 579072, 'steps': 3015, 'loss/train': 2.1035817861557007} 01/28/2022 16:46:45 - INFO - codeparrot_training - Step 3016: {'lr': 0.0004994474715292555, 'samples': 579264, 'steps': 3016, 'loss/train': 1.3639193773269653} 01/28/2022 16:46:49 - INFO - codeparrot_training - Step 3017: {'lr': 0.0004994463837413268, 'samples': 579456, 'steps': 3017, 'loss/train': 2.0114586353302} 01/28/2022 16:46:53 - INFO - codeparrot_training - Step 3018: {'lr': 0.0004994452948848488, 'samples': 579648, 'steps': 3018, 'loss/train': 3.593832850456238} 01/28/2022 16:46:58 - INFO - codeparrot_training - Step 3019: {'lr': 0.0004994442049598265, 'samples': 579840, 'steps': 3019, 'loss/train': 2.3607337474823} 01/28/2022 16:47:03 - INFO - codeparrot_training - Step 3020: {'lr': 0.0004994431139662643, 'samples': 580032, 'steps': 3020, 'loss/train': 1.461512267589569} 01/28/2022 16:47:07 - INFO - codeparrot_training - Step 3021: {'lr': 0.0004994420219041671, 'samples': 580224, 'steps': 3021, 'loss/train': 1.5286023616790771} 01/28/2022 16:47:12 - INFO - codeparrot_training - Step 3022: {'lr': 0.0004994409287735394, 'samples': 580416, 'steps': 3022, 'loss/train': 2.4724280834198} 01/28/2022 16:47:16 - INFO - codeparrot_training - Step 3023: {'lr': 0.0004994398345743861, 'samples': 580608, 'steps': 3023, 'loss/train': 1.618725299835205} 01/28/2022 16:47:20 - INFO - codeparrot_training - Step 3024: {'lr': 0.0004994387393067117, 'samples': 580800, 'steps': 3024, 'loss/train': 1.594572365283966} 01/28/2022 16:47:26 - INFO - codeparrot_training - Step 3025: {'lr': 0.0004994376429705208, 'samples': 580992, 'steps': 3025, 'loss/train': 2.5987924933433533} 01/28/2022 16:47:30 - INFO - codeparrot_training - Step 3026: {'lr': 0.0004994365455658185, 'samples': 581184, 'steps': 3026, 'loss/train': 1.6660500168800354} 01/28/2022 16:47:35 - INFO - codeparrot_training - Step 3027: {'lr': 0.000499435447092609, 'samples': 581376, 'steps': 3027, 'loss/train': 1.4087245166301727} 01/28/2022 16:47:39 - INFO - codeparrot_training - Step 3028: {'lr': 0.0004994343475508974, 'samples': 581568, 'steps': 3028, 'loss/train': 2.115963935852051} 01/28/2022 16:47:43 - INFO - codeparrot_training - Step 3029: {'lr': 0.0004994332469406882, 'samples': 581760, 'steps': 3029, 'loss/train': 0.679526075720787} 01/28/2022 16:47:48 - INFO - codeparrot_training - Step 3030: {'lr': 0.0004994321452619863, 'samples': 581952, 'steps': 3030, 'loss/train': 2.3093680143356323} 01/28/2022 16:47:53 - INFO - codeparrot_training - Step 3031: {'lr': 0.0004994310425147962, 'samples': 582144, 'steps': 3031, 'loss/train': 2.040815770626068} 01/28/2022 16:47:57 - INFO - codeparrot_training - Step 3032: {'lr': 0.0004994299386991227, 'samples': 582336, 'steps': 3032, 'loss/train': 1.51594877243042} 01/28/2022 16:48:01 - INFO - codeparrot_training - Step 3033: {'lr': 0.0004994288338149705, 'samples': 582528, 'steps': 3033, 'loss/train': 1.9795644879341125} 01/28/2022 16:48:05 - INFO - codeparrot_training - Step 3034: {'lr': 0.0004994277278623445, 'samples': 582720, 'steps': 3034, 'loss/train': 1.3814513683319092} 01/28/2022 16:48:11 - INFO - codeparrot_training - Step 3035: {'lr': 0.0004994266208412493, 'samples': 582912, 'steps': 3035, 'loss/train': 2.093902051448822} 01/28/2022 16:48:15 - INFO - codeparrot_training - Step 3036: {'lr': 0.0004994255127516895, 'samples': 583104, 'steps': 3036, 'loss/train': 0.5901720374822617} 01/28/2022 16:48:19 - INFO - codeparrot_training - Step 3037: {'lr': 0.0004994244035936701, 'samples': 583296, 'steps': 3037, 'loss/train': 0.8802470862865448} 01/28/2022 16:48:23 - INFO - codeparrot_training - Step 3038: {'lr': 0.0004994232933671958, 'samples': 583488, 'steps': 3038, 'loss/train': 2.6897900104522705} 01/28/2022 16:48:28 - INFO - codeparrot_training - Step 3039: {'lr': 0.0004994221820722713, 'samples': 583680, 'steps': 3039, 'loss/train': 0.8509560227394104} 01/28/2022 16:48:34 - INFO - codeparrot_training - Step 3040: {'lr': 0.0004994210697089013, 'samples': 583872, 'steps': 3040, 'loss/train': 2.123354136943817} 01/28/2022 16:48:38 - INFO - codeparrot_training - Step 3041: {'lr': 0.0004994199562770907, 'samples': 584064, 'steps': 3041, 'loss/train': 0.954163134098053} 01/28/2022 16:48:42 - INFO - codeparrot_training - Step 3042: {'lr': 0.0004994188417768443, 'samples': 584256, 'steps': 3042, 'loss/train': 0.6375311762094498} 01/28/2022 16:48:47 - INFO - codeparrot_training - Step 3043: {'lr': 0.0004994177262081666, 'samples': 584448, 'steps': 3043, 'loss/train': 2.3575016856193542} 01/28/2022 16:48:51 - INFO - codeparrot_training - Step 3044: {'lr': 0.0004994166095710626, 'samples': 584640, 'steps': 3044, 'loss/train': 1.962426245212555} 01/28/2022 16:48:57 - INFO - codeparrot_training - Step 3045: {'lr': 0.0004994154918655371, 'samples': 584832, 'steps': 3045, 'loss/train': 1.4810615479946136} 01/28/2022 16:49:01 - INFO - codeparrot_training - Step 3046: {'lr': 0.0004994143730915948, 'samples': 585024, 'steps': 3046, 'loss/train': 2.0043416619300842} 01/28/2022 16:49:05 - INFO - codeparrot_training - Step 3047: {'lr': 0.0004994132532492406, 'samples': 585216, 'steps': 3047, 'loss/train': 3.0205389261245728} 01/28/2022 16:49:10 - INFO - codeparrot_training - Step 3048: {'lr': 0.0004994121323384791, 'samples': 585408, 'steps': 3048, 'loss/train': 2.3999011516571045} 01/28/2022 16:49:14 - INFO - codeparrot_training - Step 3049: {'lr': 0.0004994110103593154, 'samples': 585600, 'steps': 3049, 'loss/train': 1.6242434978485107} 01/28/2022 16:49:18 - INFO - codeparrot_training - Step 3050: {'lr': 0.0004994098873117539, 'samples': 585792, 'steps': 3050, 'loss/train': 2.513813853263855} 01/28/2022 16:49:24 - INFO - codeparrot_training - Step 3051: {'lr': 0.0004994087631957998, 'samples': 585984, 'steps': 3051, 'loss/train': 2.365417242050171} 01/28/2022 16:49:28 - INFO - codeparrot_training - Step 3052: {'lr': 0.0004994076380114577, 'samples': 586176, 'steps': 3052, 'loss/train': 2.050715982913971} 01/28/2022 16:49:32 - INFO - codeparrot_training - Step 3053: {'lr': 0.0004994065117587325, 'samples': 586368, 'steps': 3053, 'loss/train': 1.9812723398208618} 01/28/2022 16:49:36 - INFO - codeparrot_training - Step 3054: {'lr': 0.0004994053844376289, 'samples': 586560, 'steps': 3054, 'loss/train': 1.7464283108711243} 01/28/2022 16:49:42 - INFO - codeparrot_training - Step 3055: {'lr': 0.000499404256048152, 'samples': 586752, 'steps': 3055, 'loss/train': 1.6523188948631287} 01/28/2022 16:49:46 - INFO - codeparrot_training - Step 3056: {'lr': 0.0004994031265903063, 'samples': 586944, 'steps': 3056, 'loss/train': 2.5107627511024475} 01/28/2022 16:49:50 - INFO - codeparrot_training - Step 3057: {'lr': 0.0004994019960640969, 'samples': 587136, 'steps': 3057, 'loss/train': 0.6378641724586487} 01/28/2022 16:49:54 - INFO - codeparrot_training - Step 3058: {'lr': 0.0004994008644695285, 'samples': 587328, 'steps': 3058, 'loss/train': 1.53237783908844} 01/28/2022 16:49:59 - INFO - codeparrot_training - Step 3059: {'lr': 0.0004993997318066061, 'samples': 587520, 'steps': 3059, 'loss/train': 2.21854966878891} 01/28/2022 16:50:05 - INFO - codeparrot_training - Step 3060: {'lr': 0.0004993985980753342, 'samples': 587712, 'steps': 3060, 'loss/train': 1.994708776473999} 01/28/2022 16:50:10 - INFO - codeparrot_training - Step 3061: {'lr': 0.0004993974632757181, 'samples': 587904, 'steps': 3061, 'loss/train': 1.9744064211845398} 01/28/2022 16:50:14 - INFO - codeparrot_training - Step 3062: {'lr': 0.0004993963274077624, 'samples': 588096, 'steps': 3062, 'loss/train': 1.0452337861061096} 01/28/2022 16:50:18 - INFO - codeparrot_training - Step 3063: {'lr': 0.000499395190471472, 'samples': 588288, 'steps': 3063, 'loss/train': 1.5658793449401855} 01/28/2022 16:50:22 - INFO - codeparrot_training - Step 3064: {'lr': 0.0004993940524668518, 'samples': 588480, 'steps': 3064, 'loss/train': 0.8983260691165924} 01/28/2022 16:50:29 - INFO - codeparrot_training - Step 3065: {'lr': 0.0004993929133939067, 'samples': 588672, 'steps': 3065, 'loss/train': 2.5434529781341553} 01/28/2022 16:50:33 - INFO - codeparrot_training - Step 3066: {'lr': 0.0004993917732526416, 'samples': 588864, 'steps': 3066, 'loss/train': 1.8929720520973206} 01/28/2022 16:50:37 - INFO - codeparrot_training - Step 3067: {'lr': 0.0004993906320430613, 'samples': 589056, 'steps': 3067, 'loss/train': 2.514933407306671} 01/28/2022 16:50:41 - INFO - codeparrot_training - Step 3068: {'lr': 0.0004993894897651706, 'samples': 589248, 'steps': 3068, 'loss/train': 2.0280940532684326} 01/28/2022 16:50:46 - INFO - codeparrot_training - Step 3069: {'lr': 0.0004993883464189747, 'samples': 589440, 'steps': 3069, 'loss/train': 1.8753392100334167} 01/28/2022 16:50:50 - INFO - codeparrot_training - Step 3070: {'lr': 0.0004993872020044781, 'samples': 589632, 'steps': 3070, 'loss/train': 1.5017495155334473} 01/28/2022 16:50:55 - INFO - codeparrot_training - Step 3071: {'lr': 0.0004993860565216861, 'samples': 589824, 'steps': 3071, 'loss/train': 2.232458174228668} 01/28/2022 16:50:59 - INFO - codeparrot_training - Step 3072: {'lr': 0.0004993849099706034, 'samples': 590016, 'steps': 3072, 'loss/train': 1.7917394042015076} 01/28/2022 16:51:04 - INFO - codeparrot_training - Step 3073: {'lr': 0.0004993837623512349, 'samples': 590208, 'steps': 3073, 'loss/train': 0.8198211193084717} 01/28/2022 16:51:08 - INFO - codeparrot_training - Step 3074: {'lr': 0.0004993826136635856, 'samples': 590400, 'steps': 3074, 'loss/train': 2.5715854167938232} 01/28/2022 16:51:12 - INFO - codeparrot_training - Step 3075: {'lr': 0.0004993814639076602, 'samples': 590592, 'steps': 3075, 'loss/train': 3.0926796197891235} 01/28/2022 16:51:18 - INFO - codeparrot_training - Step 3076: {'lr': 0.000499380313083464, 'samples': 590784, 'steps': 3076, 'loss/train': 1.9421589374542236} 01/28/2022 16:51:22 - INFO - codeparrot_training - Step 3077: {'lr': 0.0004993791611910017, 'samples': 590976, 'steps': 3077, 'loss/train': 2.3654000759124756} 01/28/2022 16:51:27 - INFO - codeparrot_training - Step 3078: {'lr': 0.0004993780082302782, 'samples': 591168, 'steps': 3078, 'loss/train': 1.0235314965248108} 01/28/2022 16:51:31 - INFO - codeparrot_training - Step 3079: {'lr': 0.0004993768542012985, 'samples': 591360, 'steps': 3079, 'loss/train': 1.333871841430664} 01/28/2022 16:51:35 - INFO - codeparrot_training - Step 3080: {'lr': 0.0004993756991040675, 'samples': 591552, 'steps': 3080, 'loss/train': 2.2003363966941833} 01/28/2022 16:51:40 - INFO - codeparrot_training - Step 3081: {'lr': 0.0004993745429385903, 'samples': 591744, 'steps': 3081, 'loss/train': 1.8355772495269775} 01/28/2022 16:51:45 - INFO - codeparrot_training - Step 3082: {'lr': 0.0004993733857048717, 'samples': 591936, 'steps': 3082, 'loss/train': 2.6800026297569275} 01/28/2022 16:51:49 - INFO - codeparrot_training - Step 3083: {'lr': 0.0004993722274029167, 'samples': 592128, 'steps': 3083, 'loss/train': 1.8787348866462708} 01/28/2022 16:51:53 - INFO - codeparrot_training - Step 3084: {'lr': 0.0004993710680327301, 'samples': 592320, 'steps': 3084, 'loss/train': 1.4393125176429749} 01/28/2022 16:51:57 - INFO - codeparrot_training - Step 3085: {'lr': 0.0004993699075943172, 'samples': 592512, 'steps': 3085, 'loss/train': 1.3121527433395386} 01/28/2022 16:52:03 - INFO - codeparrot_training - Step 3086: {'lr': 0.0004993687460876829, 'samples': 592704, 'steps': 3086, 'loss/train': 2.0437061190605164} 01/28/2022 16:52:08 - INFO - codeparrot_training - Step 3087: {'lr': 0.0004993675835128319, 'samples': 592896, 'steps': 3087, 'loss/train': 1.7931175231933594} 01/28/2022 16:52:12 - INFO - codeparrot_training - Step 3088: {'lr': 0.0004993664198697694, 'samples': 593088, 'steps': 3088, 'loss/train': 2.0387189984321594} 01/28/2022 16:52:16 - INFO - codeparrot_training - Step 3089: {'lr': 0.0004993652551585003, 'samples': 593280, 'steps': 3089, 'loss/train': 1.3897220492362976} 01/28/2022 16:52:20 - INFO - codeparrot_training - Step 3090: {'lr': 0.0004993640893790298, 'samples': 593472, 'steps': 3090, 'loss/train': 2.101107358932495} 01/28/2022 16:52:26 - INFO - codeparrot_training - Step 3091: {'lr': 0.0004993629225313625, 'samples': 593664, 'steps': 3091, 'loss/train': 2.296230733394623} 01/28/2022 16:52:30 - INFO - codeparrot_training - Step 3092: {'lr': 0.0004993617546155037, 'samples': 593856, 'steps': 3092, 'loss/train': 1.4453896880149841} 01/28/2022 16:52:34 - INFO - codeparrot_training - Step 3093: {'lr': 0.0004993605856314584, 'samples': 594048, 'steps': 3093, 'loss/train': 2.370059072971344} 01/28/2022 16:52:38 - INFO - codeparrot_training - Step 3094: {'lr': 0.0004993594155792315, 'samples': 594240, 'steps': 3094, 'loss/train': 1.9586672186851501} 01/28/2022 16:52:43 - INFO - codeparrot_training - Step 3095: {'lr': 0.000499358244458828, 'samples': 594432, 'steps': 3095, 'loss/train': 1.1443345248699188} 01/28/2022 16:52:48 - INFO - codeparrot_training - Step 3096: {'lr': 0.0004993570722702529, 'samples': 594624, 'steps': 3096, 'loss/train': 1.813184916973114} 01/28/2022 16:52:52 - INFO - codeparrot_training - Step 3097: {'lr': 0.0004993558990135115, 'samples': 594816, 'steps': 3097, 'loss/train': 1.4820351004600525} 01/28/2022 16:52:57 - INFO - codeparrot_training - Step 3098: {'lr': 0.0004993547246886084, 'samples': 595008, 'steps': 3098, 'loss/train': 2.1849648356437683} 01/28/2022 16:53:01 - INFO - codeparrot_training - Step 3099: {'lr': 0.0004993535492955488, 'samples': 595200, 'steps': 3099, 'loss/train': 2.0222400426864624} 01/28/2022 16:53:05 - INFO - codeparrot_training - Step 3100: {'lr': 0.000499352372834338, 'samples': 595392, 'steps': 3100, 'loss/train': 2.1664817333221436} 01/28/2022 16:53:11 - INFO - codeparrot_training - Step 3101: {'lr': 0.0004993511953049807, 'samples': 595584, 'steps': 3101, 'loss/train': 2.3002209663391113} 01/28/2022 16:53:15 - INFO - codeparrot_training - Step 3102: {'lr': 0.000499350016707482, 'samples': 595776, 'steps': 3102, 'loss/train': 1.963348388671875} 01/28/2022 16:53:20 - INFO - codeparrot_training - Step 3103: {'lr': 0.0004993488370418471, 'samples': 595968, 'steps': 3103, 'loss/train': 1.9579771757125854} 01/28/2022 16:53:24 - INFO - codeparrot_training - Step 3104: {'lr': 0.0004993476563080809, 'samples': 596160, 'steps': 3104, 'loss/train': 2.142146408557892} 01/28/2022 16:53:28 - INFO - codeparrot_training - Step 3105: {'lr': 0.0004993464745061885, 'samples': 596352, 'steps': 3105, 'loss/train': 4.049476504325867} 01/28/2022 16:53:33 - INFO - codeparrot_training - Step 3106: {'lr': 0.0004993452916361751, 'samples': 596544, 'steps': 3106, 'loss/train': 1.6600207686424255} 01/28/2022 16:53:38 - INFO - codeparrot_training - Step 3107: {'lr': 0.0004993441076980455, 'samples': 596736, 'steps': 3107, 'loss/train': 1.9200233817100525} 01/28/2022 16:53:42 - INFO - codeparrot_training - Step 3108: {'lr': 0.0004993429226918051, 'samples': 596928, 'steps': 3108, 'loss/train': 2.19151908159256} 01/28/2022 16:53:46 - INFO - codeparrot_training - Step 3109: {'lr': 0.0004993417366174586, 'samples': 597120, 'steps': 3109, 'loss/train': 1.751364827156067} 01/28/2022 16:53:50 - INFO - codeparrot_training - Step 3110: {'lr': 0.0004993405494750113, 'samples': 597312, 'steps': 3110, 'loss/train': 2.3060145378112793} 01/28/2022 16:53:56 - INFO - codeparrot_training - Step 3111: {'lr': 0.0004993393612644683, 'samples': 597504, 'steps': 3111, 'loss/train': 2.300027310848236} 01/28/2022 16:54:00 - INFO - codeparrot_training - Step 3112: {'lr': 0.0004993381719858347, 'samples': 597696, 'steps': 3112, 'loss/train': 2.775234282016754} 01/28/2022 16:54:04 - INFO - codeparrot_training - Step 3113: {'lr': 0.0004993369816391156, 'samples': 597888, 'steps': 3113, 'loss/train': 2.1231120228767395} 01/28/2022 16:54:09 - INFO - codeparrot_training - Step 3114: {'lr': 0.0004993357902243158, 'samples': 598080, 'steps': 3114, 'loss/train': 2.558064579963684} 01/28/2022 16:54:13 - INFO - codeparrot_training - Step 3115: {'lr': 0.0004993345977414408, 'samples': 598272, 'steps': 3115, 'loss/train': 1.9296228289604187} 01/28/2022 16:54:18 - INFO - codeparrot_training - Step 3116: {'lr': 0.0004993334041904957, 'samples': 598464, 'steps': 3116, 'loss/train': 2.122667670249939} 01/28/2022 16:54:22 - INFO - codeparrot_training - Step 3117: {'lr': 0.0004993322095714853, 'samples': 598656, 'steps': 3117, 'loss/train': 1.8460464477539062} 01/28/2022 16:54:27 - INFO - codeparrot_training - Step 3118: {'lr': 0.0004993310138844149, 'samples': 598848, 'steps': 3118, 'loss/train': 2.1707940101623535} 01/28/2022 16:54:31 - INFO - codeparrot_training - Step 3119: {'lr': 0.0004993298171292896, 'samples': 599040, 'steps': 3119, 'loss/train': 2.576118528842926} 01/28/2022 16:54:35 - INFO - codeparrot_training - Step 3120: {'lr': 0.0004993286193061145, 'samples': 599232, 'steps': 3120, 'loss/train': 2.2718823552131653} 01/28/2022 16:54:41 - INFO - codeparrot_training - Step 3121: {'lr': 0.0004993274204148949, 'samples': 599424, 'steps': 3121, 'loss/train': 1.6897282004356384} 01/28/2022 16:54:46 - INFO - codeparrot_training - Step 3122: {'lr': 0.0004993262204556356, 'samples': 599616, 'steps': 3122, 'loss/train': 2.186732053756714} 01/28/2022 16:54:50 - INFO - codeparrot_training - Step 3123: {'lr': 0.0004993250194283421, 'samples': 599808, 'steps': 3123, 'loss/train': 1.9195438027381897} 01/28/2022 16:54:54 - INFO - codeparrot_training - Step 3124: {'lr': 0.0004993238173330194, 'samples': 600000, 'steps': 3124, 'loss/train': 2.380008101463318} 01/28/2022 16:54:59 - INFO - codeparrot_training - Step 3125: {'lr': 0.0004993226141696725, 'samples': 600192, 'steps': 3125, 'loss/train': 1.8099936246871948} 01/28/2022 16:55:04 - INFO - codeparrot_training - Step 3126: {'lr': 0.0004993214099383069, 'samples': 600384, 'steps': 3126, 'loss/train': 2.1202130913734436} 01/28/2022 16:55:08 - INFO - codeparrot_training - Step 3127: {'lr': 0.0004993202046389274, 'samples': 600576, 'steps': 3127, 'loss/train': 2.1381710171699524} 01/28/2022 16:55:12 - INFO - codeparrot_training - Step 3128: {'lr': 0.0004993189982715392, 'samples': 600768, 'steps': 3128, 'loss/train': 1.9770607352256775} 01/28/2022 16:55:17 - INFO - codeparrot_training - Step 3129: {'lr': 0.0004993177908361479, 'samples': 600960, 'steps': 3129, 'loss/train': 1.7584730386734009} 01/28/2022 16:55:21 - INFO - codeparrot_training - Step 3130: {'lr': 0.000499316582332758, 'samples': 601152, 'steps': 3130, 'loss/train': 2.031623125076294} 01/28/2022 16:55:27 - INFO - codeparrot_training - Step 3131: {'lr': 0.0004993153727613753, 'samples': 601344, 'steps': 3131, 'loss/train': 2.8003923296928406} 01/28/2022 16:55:31 - INFO - codeparrot_training - Step 3132: {'lr': 0.0004993141621220046, 'samples': 601536, 'steps': 3132, 'loss/train': 1.9725889563560486} 01/28/2022 16:55:35 - INFO - codeparrot_training - Step 3133: {'lr': 0.0004993129504146512, 'samples': 601728, 'steps': 3133, 'loss/train': 2.2429820895195007} 01/28/2022 16:55:40 - INFO - codeparrot_training - Step 3134: {'lr': 0.0004993117376393203, 'samples': 601920, 'steps': 3134, 'loss/train': 2.924464702606201} 01/28/2022 16:55:44 - INFO - codeparrot_training - Step 3135: {'lr': 0.000499310523796017, 'samples': 602112, 'steps': 3135, 'loss/train': 1.7336978316307068} 01/28/2022 16:55:49 - INFO - codeparrot_training - Step 3136: {'lr': 0.0004993093088847466, 'samples': 602304, 'steps': 3136, 'loss/train': 2.2609536051750183} 01/28/2022 16:55:53 - INFO - codeparrot_training - Step 3137: {'lr': 0.0004993080929055144, 'samples': 602496, 'steps': 3137, 'loss/train': 1.3223595321178436} 01/28/2022 16:55:57 - INFO - codeparrot_training - Step 3138: {'lr': 0.0004993068758583254, 'samples': 602688, 'steps': 3138, 'loss/train': 2.262670397758484} 01/28/2022 16:56:02 - INFO - codeparrot_training - Step 3139: {'lr': 0.0004993056577431849, 'samples': 602880, 'steps': 3139, 'loss/train': 2.093923509120941} 01/28/2022 16:56:06 - INFO - codeparrot_training - Step 3140: {'lr': 0.0004993044385600982, 'samples': 603072, 'steps': 3140, 'loss/train': 2.5338610410690308} 01/28/2022 16:56:11 - INFO - codeparrot_training - Step 3141: {'lr': 0.0004993032183090704, 'samples': 603264, 'steps': 3141, 'loss/train': 1.7147786021232605} 01/28/2022 16:56:16 - INFO - codeparrot_training - Step 3142: {'lr': 0.0004993019969901069, 'samples': 603456, 'steps': 3142, 'loss/train': 2.0301833152770996} 01/28/2022 16:56:20 - INFO - codeparrot_training - Step 3143: {'lr': 0.0004993007746032126, 'samples': 603648, 'steps': 3143, 'loss/train': 2.101225197315216} 01/28/2022 16:56:24 - INFO - codeparrot_training - Step 3144: {'lr': 0.000499299551148393, 'samples': 603840, 'steps': 3144, 'loss/train': 1.856275498867035} 01/28/2022 16:56:28 - INFO - codeparrot_training - Step 3145: {'lr': 0.0004992983266256533, 'samples': 604032, 'steps': 3145, 'loss/train': 3.5087571144104004} 01/28/2022 16:56:35 - INFO - codeparrot_training - Step 3146: {'lr': 0.0004992971010349987, 'samples': 604224, 'steps': 3146, 'loss/train': 0.5995797961950302} 01/28/2022 16:56:39 - INFO - codeparrot_training - Step 3147: {'lr': 0.0004992958743764346, 'samples': 604416, 'steps': 3147, 'loss/train': 1.819128155708313} 01/28/2022 16:56:43 - INFO - codeparrot_training - Step 3148: {'lr': 0.0004992946466499661, 'samples': 604608, 'steps': 3148, 'loss/train': 1.739919126033783} 01/28/2022 16:56:47 - INFO - codeparrot_training - Step 3149: {'lr': 0.0004992934178555984, 'samples': 604800, 'steps': 3149, 'loss/train': 0.9967224597930908} 01/28/2022 16:56:52 - INFO - codeparrot_training - Step 3150: {'lr': 0.000499292187993337, 'samples': 604992, 'steps': 3150, 'loss/train': 2.3949385285377502} 01/28/2022 16:56:57 - INFO - codeparrot_training - Step 3151: {'lr': 0.0004992909570631868, 'samples': 605184, 'steps': 3151, 'loss/train': 1.5502774715423584} 01/28/2022 16:57:01 - INFO - codeparrot_training - Step 3152: {'lr': 0.0004992897250651535, 'samples': 605376, 'steps': 3152, 'loss/train': 1.9983623027801514} 01/28/2022 16:57:05 - INFO - codeparrot_training - Step 3153: {'lr': 0.0004992884919992421, 'samples': 605568, 'steps': 3153, 'loss/train': 2.3601393699645996} 01/28/2022 16:57:10 - INFO - codeparrot_training - Step 3154: {'lr': 0.000499287257865458, 'samples': 605760, 'steps': 3154, 'loss/train': 1.3316857516765594} 01/28/2022 16:57:14 - INFO - codeparrot_training - Step 3155: {'lr': 0.0004992860226638064, 'samples': 605952, 'steps': 3155, 'loss/train': 1.7809469103813171} 01/28/2022 16:57:20 - INFO - codeparrot_training - Step 3156: {'lr': 0.0004992847863942927, 'samples': 606144, 'steps': 3156, 'loss/train': 1.9474035501480103} 01/28/2022 16:57:24 - INFO - codeparrot_training - Step 3157: {'lr': 0.000499283549056922, 'samples': 606336, 'steps': 3157, 'loss/train': 1.7894298434257507} 01/28/2022 16:57:28 - INFO - codeparrot_training - Step 3158: {'lr': 0.0004992823106516999, 'samples': 606528, 'steps': 3158, 'loss/train': 1.2890046536922455} 01/28/2022 16:57:33 - INFO - codeparrot_training - Step 3159: {'lr': 0.0004992810711786314, 'samples': 606720, 'steps': 3159, 'loss/train': 2.43354070186615} 01/28/2022 16:57:37 - INFO - codeparrot_training - Step 3160: {'lr': 0.000499279830637722, 'samples': 606912, 'steps': 3160, 'loss/train': 2.210454046726227} 01/28/2022 16:57:42 - INFO - codeparrot_training - Step 3161: {'lr': 0.000499278589028977, 'samples': 607104, 'steps': 3161, 'loss/train': 2.0348533987998962} 01/28/2022 16:57:46 - INFO - codeparrot_training - Step 3162: {'lr': 0.0004992773463524016, 'samples': 607296, 'steps': 3162, 'loss/train': 1.859728217124939} 01/28/2022 16:57:51 - INFO - codeparrot_training - Step 3163: {'lr': 0.0004992761026080013, 'samples': 607488, 'steps': 3163, 'loss/train': 1.8319780826568604} 01/28/2022 16:57:55 - INFO - codeparrot_training - Step 3164: {'lr': 0.0004992748577957812, 'samples': 607680, 'steps': 3164, 'loss/train': 1.7331987619400024} 01/28/2022 16:57:59 - INFO - codeparrot_training - Step 3165: {'lr': 0.0004992736119157469, 'samples': 607872, 'steps': 3165, 'loss/train': 3.022161841392517} 01/28/2022 16:58:04 - INFO - codeparrot_training - Step 3166: {'lr': 0.0004992723649679035, 'samples': 608064, 'steps': 3166, 'loss/train': 2.0934773683547974} 01/28/2022 16:58:09 - INFO - codeparrot_training - Step 3167: {'lr': 0.0004992711169522565, 'samples': 608256, 'steps': 3167, 'loss/train': 1.2101448476314545} 01/28/2022 16:58:13 - INFO - codeparrot_training - Step 3168: {'lr': 0.0004992698678688111, 'samples': 608448, 'steps': 3168, 'loss/train': 2.1100974082946777} 01/28/2022 16:58:17 - INFO - codeparrot_training - Step 3169: {'lr': 0.0004992686177175728, 'samples': 608640, 'steps': 3169, 'loss/train': 1.7137736678123474} 01/28/2022 16:58:21 - INFO - codeparrot_training - Step 3170: {'lr': 0.000499267366498547, 'samples': 608832, 'steps': 3170, 'loss/train': 0.9504860043525696} 01/28/2022 16:58:26 - INFO - codeparrot_training - Step 3171: {'lr': 0.0004992661142117388, 'samples': 609024, 'steps': 3171, 'loss/train': 1.3148620426654816} 01/28/2022 16:58:31 - INFO - codeparrot_training - Step 3172: {'lr': 0.0004992648608571537, 'samples': 609216, 'steps': 3172, 'loss/train': 1.8452407121658325} 01/28/2022 16:58:35 - INFO - codeparrot_training - Step 3173: {'lr': 0.0004992636064347971, 'samples': 609408, 'steps': 3173, 'loss/train': 1.8695425987243652} 01/28/2022 16:58:39 - INFO - codeparrot_training - Step 3174: {'lr': 0.0004992623509446746, 'samples': 609600, 'steps': 3174, 'loss/train': 1.607708215713501} 01/28/2022 16:58:43 - INFO - codeparrot_training - Step 3175: {'lr': 0.0004992610943867911, 'samples': 609792, 'steps': 3175, 'loss/train': 2.0636883974075317} 01/28/2022 16:58:49 - INFO - codeparrot_training - Step 3176: {'lr': 0.0004992598367611523, 'samples': 609984, 'steps': 3176, 'loss/train': 1.9925578236579895} 01/28/2022 16:58:54 - INFO - codeparrot_training - Step 3177: {'lr': 0.0004992585780677634, 'samples': 610176, 'steps': 3177, 'loss/train': 1.7187384366989136} 01/28/2022 16:58:58 - INFO - codeparrot_training - Step 3178: {'lr': 0.00049925731830663, 'samples': 610368, 'steps': 3178, 'loss/train': 2.034905791282654} 01/28/2022 16:59:02 - INFO - codeparrot_training - Step 3179: {'lr': 0.0004992560574777574, 'samples': 610560, 'steps': 3179, 'loss/train': 2.8247097730636597} 01/28/2022 16:59:06 - INFO - codeparrot_training - Step 3180: {'lr': 0.000499254795581151, 'samples': 610752, 'steps': 3180, 'loss/train': 2.795116424560547} 01/28/2022 16:59:12 - INFO - codeparrot_training - Step 3181: {'lr': 0.0004992535326168162, 'samples': 610944, 'steps': 3181, 'loss/train': 1.0584003031253815} 01/28/2022 16:59:16 - INFO - codeparrot_training - Step 3182: {'lr': 0.0004992522685847583, 'samples': 611136, 'steps': 3182, 'loss/train': 1.0693454146385193} 01/28/2022 16:59:20 - INFO - codeparrot_training - Step 3183: {'lr': 0.000499251003484983, 'samples': 611328, 'steps': 3183, 'loss/train': 2.011275351047516} 01/28/2022 16:59:24 - INFO - codeparrot_training - Step 3184: {'lr': 0.0004992497373174955, 'samples': 611520, 'steps': 3184, 'loss/train': 1.7303332686424255} 01/28/2022 16:59:29 - INFO - codeparrot_training - Step 3185: {'lr': 0.0004992484700823012, 'samples': 611712, 'steps': 3185, 'loss/train': 1.8550685048103333} 01/28/2022 16:59:34 - INFO - codeparrot_training - Step 3186: {'lr': 0.0004992472017794057, 'samples': 611904, 'steps': 3186, 'loss/train': 2.348506271839142} 01/28/2022 16:59:38 - INFO - codeparrot_training - Step 3187: {'lr': 0.0004992459324088143, 'samples': 612096, 'steps': 3187, 'loss/train': 1.21722874045372} 01/28/2022 16:59:42 - INFO - codeparrot_training - Step 3188: {'lr': 0.0004992446619705324, 'samples': 612288, 'steps': 3188, 'loss/train': 1.6308563947677612} 01/28/2022 16:59:47 - INFO - codeparrot_training - Step 3189: {'lr': 0.0004992433904645654, 'samples': 612480, 'steps': 3189, 'loss/train': 1.5778852701187134} 01/28/2022 16:59:51 - INFO - codeparrot_training - Step 3190: {'lr': 0.0004992421178909191, 'samples': 612672, 'steps': 3190, 'loss/train': 2.238235831260681} 01/28/2022 16:59:57 - INFO - codeparrot_training - Step 3191: {'lr': 0.0004992408442495986, 'samples': 612864, 'steps': 3191, 'loss/train': 2.2689347863197327} 01/28/2022 17:00:01 - INFO - codeparrot_training - Step 3192: {'lr': 0.0004992395695406095, 'samples': 613056, 'steps': 3192, 'loss/train': 2.674326717853546} 01/28/2022 17:00:05 - INFO - codeparrot_training - Step 3193: {'lr': 0.0004992382937639572, 'samples': 613248, 'steps': 3193, 'loss/train': 1.5914947986602783} 01/28/2022 17:00:10 - INFO - codeparrot_training - Step 3194: {'lr': 0.0004992370169196472, 'samples': 613440, 'steps': 3194, 'loss/train': 2.311042070388794} 01/28/2022 17:00:14 - INFO - codeparrot_training - Step 3195: {'lr': 0.000499235739007685, 'samples': 613632, 'steps': 3195, 'loss/train': 0.8782148659229279} 01/28/2022 17:00:19 - INFO - codeparrot_training - Step 3196: {'lr': 0.000499234460028076, 'samples': 613824, 'steps': 3196, 'loss/train': 0.5374535322189331} 01/28/2022 17:00:23 - INFO - codeparrot_training - Step 3197: {'lr': 0.0004992331799808258, 'samples': 614016, 'steps': 3197, 'loss/train': 1.8969776630401611} 01/28/2022 17:00:27 - INFO - codeparrot_training - Step 3198: {'lr': 0.0004992318988659396, 'samples': 614208, 'steps': 3198, 'loss/train': 1.3651802837848663} 01/28/2022 17:00:32 - INFO - codeparrot_training - Step 3199: {'lr': 0.0004992306166834232, 'samples': 614400, 'steps': 3199, 'loss/train': 2.2697920203208923} 01/28/2022 17:00:36 - INFO - codeparrot_training - Step 3200: {'lr': 0.000499229333433282, 'samples': 614592, 'steps': 3200, 'loss/train': 2.97292560338974} 01/28/2022 17:00:42 - INFO - codeparrot_training - Step 3201: {'lr': 0.0004992280491155214, 'samples': 614784, 'steps': 3201, 'loss/train': 1.5301106572151184} 01/28/2022 17:00:46 - INFO - codeparrot_training - Step 3202: {'lr': 0.0004992267637301471, 'samples': 614976, 'steps': 3202, 'loss/train': 2.0785632133483887} 01/28/2022 17:00:50 - INFO - codeparrot_training - Step 3203: {'lr': 0.0004992254772771644, 'samples': 615168, 'steps': 3203, 'loss/train': 2.2770169973373413} 01/28/2022 17:00:54 - INFO - codeparrot_training - Step 3204: {'lr': 0.0004992241897565789, 'samples': 615360, 'steps': 3204, 'loss/train': 1.7247494459152222} 01/28/2022 17:00:59 - INFO - codeparrot_training - Step 3205: {'lr': 0.0004992229011683961, 'samples': 615552, 'steps': 3205, 'loss/train': 2.550598382949829} 01/28/2022 17:01:05 - INFO - codeparrot_training - Step 3206: {'lr': 0.0004992216115126216, 'samples': 615744, 'steps': 3206, 'loss/train': 1.630627155303955} 01/28/2022 17:01:09 - INFO - codeparrot_training - Step 3207: {'lr': 0.0004992203207892607, 'samples': 615936, 'steps': 3207, 'loss/train': 1.3964444696903229} 01/28/2022 17:01:13 - INFO - codeparrot_training - Step 3208: {'lr': 0.0004992190289983192, 'samples': 616128, 'steps': 3208, 'loss/train': 1.9288185238838196} 01/28/2022 17:01:17 - INFO - codeparrot_training - Step 3209: {'lr': 0.0004992177361398026, 'samples': 616320, 'steps': 3209, 'loss/train': 0.4505711495876312} 01/28/2022 17:01:22 - INFO - codeparrot_training - Step 3210: {'lr': 0.0004992164422137162, 'samples': 616512, 'steps': 3210, 'loss/train': 2.150437116622925} 01/28/2022 17:01:27 - INFO - codeparrot_training - Step 3211: {'lr': 0.0004992151472200657, 'samples': 616704, 'steps': 3211, 'loss/train': 1.6219443082809448} 01/28/2022 17:01:32 - INFO - codeparrot_training - Step 3212: {'lr': 0.0004992138511588567, 'samples': 616896, 'steps': 3212, 'loss/train': 1.4008557200431824} 01/28/2022 17:01:36 - INFO - codeparrot_training - Step 3213: {'lr': 0.0004992125540300947, 'samples': 617088, 'steps': 3213, 'loss/train': 1.087882786989212} 01/28/2022 17:01:40 - INFO - codeparrot_training - Step 3214: {'lr': 0.0004992112558337852, 'samples': 617280, 'steps': 3214, 'loss/train': 2.2497262358665466} 01/28/2022 17:01:44 - INFO - codeparrot_training - Step 3215: {'lr': 0.0004992099565699339, 'samples': 617472, 'steps': 3215, 'loss/train': 2.127381920814514} 01/28/2022 17:01:49 - INFO - codeparrot_training - Step 3216: {'lr': 0.0004992086562385462, 'samples': 617664, 'steps': 3216, 'loss/train': 1.9017881155014038} 01/28/2022 17:01:55 - INFO - codeparrot_training - Step 3217: {'lr': 0.0004992073548396277, 'samples': 617856, 'steps': 3217, 'loss/train': 1.8568595051765442} 01/28/2022 17:01:59 - INFO - codeparrot_training - Step 3218: {'lr': 0.0004992060523731842, 'samples': 618048, 'steps': 3218, 'loss/train': 1.363954335451126} 01/28/2022 17:02:03 - INFO - codeparrot_training - Step 3219: {'lr': 0.0004992047488392209, 'samples': 618240, 'steps': 3219, 'loss/train': 2.7096351385116577} 01/28/2022 17:02:07 - INFO - codeparrot_training - Step 3220: {'lr': 0.0004992034442377437, 'samples': 618432, 'steps': 3220, 'loss/train': 1.689037263393402} 01/28/2022 17:02:12 - INFO - codeparrot_training - Step 3221: {'lr': 0.0004992021385687582, 'samples': 618624, 'steps': 3221, 'loss/train': 2.7237828969955444} 01/28/2022 17:02:17 - INFO - codeparrot_training - Step 3222: {'lr': 0.0004992008318322697, 'samples': 618816, 'steps': 3222, 'loss/train': 2.085135340690613} 01/28/2022 17:02:21 - INFO - codeparrot_training - Step 3223: {'lr': 0.000499199524028284, 'samples': 619008, 'steps': 3223, 'loss/train': 3.0377039909362793} 01/28/2022 17:02:25 - INFO - codeparrot_training - Step 3224: {'lr': 0.0004991982151568066, 'samples': 619200, 'steps': 3224, 'loss/train': 1.819304645061493} 01/28/2022 17:02:30 - INFO - codeparrot_training - Step 3225: {'lr': 0.0004991969052178433, 'samples': 619392, 'steps': 3225, 'loss/train': 1.8873367309570312} 01/28/2022 17:02:34 - INFO - codeparrot_training - Step 3226: {'lr': 0.0004991955942113995, 'samples': 619584, 'steps': 3226, 'loss/train': 4.911519527435303} 01/28/2022 17:02:39 - INFO - codeparrot_training - Step 3227: {'lr': 0.0004991942821374809, 'samples': 619776, 'steps': 3227, 'loss/train': 2.647854208946228} 01/28/2022 17:02:43 - INFO - codeparrot_training - Step 3228: {'lr': 0.0004991929689960932, 'samples': 619968, 'steps': 3228, 'loss/train': 1.9167938232421875} 01/28/2022 17:02:48 - INFO - codeparrot_training - Step 3229: {'lr': 0.000499191654787242, 'samples': 620160, 'steps': 3229, 'loss/train': 1.5446830987930298} 01/28/2022 17:02:52 - INFO - codeparrot_training - Step 3230: {'lr': 0.0004991903395109328, 'samples': 620352, 'steps': 3230, 'loss/train': 1.6821311116218567} 01/28/2022 17:02:56 - INFO - codeparrot_training - Step 3231: {'lr': 0.0004991890231671712, 'samples': 620544, 'steps': 3231, 'loss/train': 2.3745124340057373} 01/28/2022 17:03:01 - INFO - codeparrot_training - Step 3232: {'lr': 0.0004991877057559631, 'samples': 620736, 'steps': 3232, 'loss/train': 1.3021137714385986} 01/28/2022 17:03:06 - INFO - codeparrot_training - Step 3233: {'lr': 0.0004991863872773139, 'samples': 620928, 'steps': 3233, 'loss/train': 2.6681143641471863} 01/28/2022 17:03:10 - INFO - codeparrot_training - Step 3234: {'lr': 0.0004991850677312295, 'samples': 621120, 'steps': 3234, 'loss/train': 2.273876667022705} 01/28/2022 17:03:14 - INFO - codeparrot_training - Step 3235: {'lr': 0.0004991837471177152, 'samples': 621312, 'steps': 3235, 'loss/train': 2.222374141216278} 01/28/2022 17:03:18 - INFO - codeparrot_training - Step 3236: {'lr': 0.000499182425436777, 'samples': 621504, 'steps': 3236, 'loss/train': 1.4676686525344849} 01/28/2022 17:03:25 - INFO - codeparrot_training - Step 3237: {'lr': 0.0004991811026884203, 'samples': 621696, 'steps': 3237, 'loss/train': 2.4720484614372253} 01/28/2022 17:03:29 - INFO - codeparrot_training - Step 3238: {'lr': 0.0004991797788726509, 'samples': 621888, 'steps': 3238, 'loss/train': 2.255162000656128} 01/28/2022 17:03:33 - INFO - codeparrot_training - Step 3239: {'lr': 0.0004991784539894745, 'samples': 622080, 'steps': 3239, 'loss/train': 2.2207964658737183} 01/28/2022 17:03:38 - INFO - codeparrot_training - Step 3240: {'lr': 0.0004991771280388967, 'samples': 622272, 'steps': 3240, 'loss/train': 2.696568489074707} 01/28/2022 17:03:42 - INFO - codeparrot_training - Step 3241: {'lr': 0.0004991758010209232, 'samples': 622464, 'steps': 3241, 'loss/train': 1.7082914113998413} 01/28/2022 17:03:47 - INFO - codeparrot_training - Step 3242: {'lr': 0.0004991744729355598, 'samples': 622656, 'steps': 3242, 'loss/train': 2.3900349140167236} 01/28/2022 17:03:51 - INFO - codeparrot_training - Step 3243: {'lr': 0.0004991731437828119, 'samples': 622848, 'steps': 3243, 'loss/train': 1.9676356315612793} 01/28/2022 17:03:55 - INFO - codeparrot_training - Step 3244: {'lr': 0.0004991718135626855, 'samples': 623040, 'steps': 3244, 'loss/train': 1.825311541557312} 01/28/2022 17:04:00 - INFO - codeparrot_training - Step 3245: {'lr': 0.0004991704822751861, 'samples': 623232, 'steps': 3245, 'loss/train': 2.150765597820282} 01/28/2022 17:04:04 - INFO - codeparrot_training - Step 3246: {'lr': 0.0004991691499203195, 'samples': 623424, 'steps': 3246, 'loss/train': 1.927377998828888} 01/28/2022 17:04:10 - INFO - codeparrot_training - Step 3247: {'lr': 0.0004991678164980914, 'samples': 623616, 'steps': 3247, 'loss/train': 1.2465636134147644} 01/28/2022 17:04:14 - INFO - codeparrot_training - Step 3248: {'lr': 0.0004991664820085074, 'samples': 623808, 'steps': 3248, 'loss/train': 2.0710984468460083} 01/28/2022 17:04:18 - INFO - codeparrot_training - Step 3249: {'lr': 0.0004991651464515735, 'samples': 624000, 'steps': 3249, 'loss/train': 1.8042903542518616} 01/28/2022 17:04:22 - INFO - codeparrot_training - Step 3250: {'lr': 0.0004991638098272951, 'samples': 624192, 'steps': 3250, 'loss/train': 1.1404131352901459} 01/28/2022 17:04:27 - INFO - codeparrot_training - Step 3251: {'lr': 0.000499162472135678, 'samples': 624384, 'steps': 3251, 'loss/train': 1.778752326965332} 01/28/2022 17:04:33 - INFO - codeparrot_training - Step 3252: {'lr': 0.0004991611333767281, 'samples': 624576, 'steps': 3252, 'loss/train': 2.1574453711509705} 01/28/2022 17:04:37 - INFO - codeparrot_training - Step 3253: {'lr': 0.000499159793550451, 'samples': 624768, 'steps': 3253, 'loss/train': 1.7791308760643005} 01/28/2022 17:04:41 - INFO - codeparrot_training - Step 3254: {'lr': 0.0004991584526568524, 'samples': 624960, 'steps': 3254, 'loss/train': 1.737656056880951} 01/28/2022 17:04:45 - INFO - codeparrot_training - Step 3255: {'lr': 0.0004991571106959383, 'samples': 625152, 'steps': 3255, 'loss/train': 2.277083158493042} 01/28/2022 17:04:50 - INFO - codeparrot_training - Step 3256: {'lr': 0.000499155767667714, 'samples': 625344, 'steps': 3256, 'loss/train': 3.958219885826111} 01/28/2022 17:04:55 - INFO - codeparrot_training - Step 3257: {'lr': 0.0004991544235721857, 'samples': 625536, 'steps': 3257, 'loss/train': 1.6013585329055786} 01/28/2022 17:04:59 - INFO - codeparrot_training - Step 3258: {'lr': 0.0004991530784093589, 'samples': 625728, 'steps': 3258, 'loss/train': 2.2681750059127808} 01/28/2022 17:05:04 - INFO - codeparrot_training - Step 3259: {'lr': 0.0004991517321792394, 'samples': 625920, 'steps': 3259, 'loss/train': 2.193070113658905} 01/28/2022 17:05:08 - INFO - codeparrot_training - Step 3260: {'lr': 0.000499150384881833, 'samples': 626112, 'steps': 3260, 'loss/train': 1.7734951972961426} 01/28/2022 17:05:12 - INFO - codeparrot_training - Step 3261: {'lr': 0.0004991490365171454, 'samples': 626304, 'steps': 3261, 'loss/train': 1.6101155877113342} 01/28/2022 17:05:18 - INFO - codeparrot_training - Step 3262: {'lr': 0.0004991476870851825, 'samples': 626496, 'steps': 3262, 'loss/train': 1.8168498873710632} 01/28/2022 17:05:23 - INFO - codeparrot_training - Step 3263: {'lr': 0.0004991463365859501, 'samples': 626688, 'steps': 3263, 'loss/train': 1.2220890820026398} 01/28/2022 17:05:27 - INFO - codeparrot_training - Step 3264: {'lr': 0.0004991449850194538, 'samples': 626880, 'steps': 3264, 'loss/train': 2.3585540056228638} 01/28/2022 17:05:31 - INFO - codeparrot_training - Step 3265: {'lr': 0.0004991436323856995, 'samples': 627072, 'steps': 3265, 'loss/train': 1.2493788599967957} 01/28/2022 17:05:35 - INFO - codeparrot_training - Step 3266: {'lr': 0.0004991422786846931, 'samples': 627264, 'steps': 3266, 'loss/train': 2.65859591960907} 01/28/2022 17:05:41 - INFO - codeparrot_training - Step 3267: {'lr': 0.0004991409239164401, 'samples': 627456, 'steps': 3267, 'loss/train': 2.7589917182922363} 01/28/2022 17:05:45 - INFO - codeparrot_training - Step 3268: {'lr': 0.0004991395680809467, 'samples': 627648, 'steps': 3268, 'loss/train': 1.9395491480827332} 01/28/2022 17:05:49 - INFO - codeparrot_training - Step 3269: {'lr': 0.0004991382111782183, 'samples': 627840, 'steps': 3269, 'loss/train': 1.9686625599861145} 01/28/2022 17:05:54 - INFO - codeparrot_training - Step 3270: {'lr': 0.0004991368532082611, 'samples': 628032, 'steps': 3270, 'loss/train': 5.738460659980774} 01/28/2022 17:05:58 - INFO - codeparrot_training - Step 3271: {'lr': 0.0004991354941710806, 'samples': 628224, 'steps': 3271, 'loss/train': 1.8478147387504578} 01/28/2022 17:06:03 - INFO - codeparrot_training - Step 3272: {'lr': 0.0004991341340666828, 'samples': 628416, 'steps': 3272, 'loss/train': 1.7940187454223633} 01/28/2022 17:06:08 - INFO - codeparrot_training - Step 3273: {'lr': 0.0004991327728950736, 'samples': 628608, 'steps': 3273, 'loss/train': 2.01948219537735} 01/28/2022 17:06:12 - INFO - codeparrot_training - Step 3274: {'lr': 0.0004991314106562586, 'samples': 628800, 'steps': 3274, 'loss/train': 1.9770069122314453} 01/28/2022 17:06:16 - INFO - codeparrot_training - Step 3275: {'lr': 0.0004991300473502437, 'samples': 628992, 'steps': 3275, 'loss/train': 1.5398367047309875} 01/28/2022 17:06:20 - INFO - codeparrot_training - Step 3276: {'lr': 0.0004991286829770348, 'samples': 629184, 'steps': 3276, 'loss/train': 1.4643183946609497} 01/28/2022 17:06:26 - INFO - codeparrot_training - Step 3277: {'lr': 0.0004991273175366378, 'samples': 629376, 'steps': 3277, 'loss/train': 1.6179712414741516} 01/28/2022 17:06:31 - INFO - codeparrot_training - Step 3278: {'lr': 0.0004991259510290584, 'samples': 629568, 'steps': 3278, 'loss/train': 2.192616105079651} 01/28/2022 17:06:35 - INFO - codeparrot_training - Step 3279: {'lr': 0.0004991245834543025, 'samples': 629760, 'steps': 3279, 'loss/train': 2.176278233528137} 01/28/2022 17:06:39 - INFO - codeparrot_training - Step 3280: {'lr': 0.0004991232148123761, 'samples': 629952, 'steps': 3280, 'loss/train': 1.9814168214797974} 01/28/2022 17:06:43 - INFO - codeparrot_training - Step 3281: {'lr': 0.0004991218451032849, 'samples': 630144, 'steps': 3281, 'loss/train': 2.3228818774223328} 01/28/2022 17:06:49 - INFO - codeparrot_training - Step 3282: {'lr': 0.0004991204743270348, 'samples': 630336, 'steps': 3282, 'loss/train': 1.5748702883720398} 01/28/2022 17:06:53 - INFO - codeparrot_training - Step 3283: {'lr': 0.0004991191024836317, 'samples': 630528, 'steps': 3283, 'loss/train': 3.1816656589508057} 01/28/2022 17:06:57 - INFO - codeparrot_training - Step 3284: {'lr': 0.0004991177295730815, 'samples': 630720, 'steps': 3284, 'loss/train': 1.1091086268424988} 01/28/2022 17:07:01 - INFO - codeparrot_training - Step 3285: {'lr': 0.0004991163555953901, 'samples': 630912, 'steps': 3285, 'loss/train': 1.6157383918762207} 01/28/2022 17:07:06 - INFO - codeparrot_training - Step 3286: {'lr': 0.0004991149805505632, 'samples': 631104, 'steps': 3286, 'loss/train': 1.40871861577034} 01/28/2022 17:07:11 - INFO - codeparrot_training - Step 3287: {'lr': 0.0004991136044386069, 'samples': 631296, 'steps': 3287, 'loss/train': 1.7828680872917175} 01/28/2022 17:07:16 - INFO - codeparrot_training - Step 3288: {'lr': 0.0004991122272595271, 'samples': 631488, 'steps': 3288, 'loss/train': 0.8940243422985077} 01/28/2022 17:07:20 - INFO - codeparrot_training - Step 3289: {'lr': 0.0004991108490133296, 'samples': 631680, 'steps': 3289, 'loss/train': 1.5825551748275757} 01/28/2022 17:07:24 - INFO - codeparrot_training - Step 3290: {'lr': 0.0004991094697000202, 'samples': 631872, 'steps': 3290, 'loss/train': 1.9638692736625671} 01/28/2022 17:07:28 - INFO - codeparrot_training - Step 3291: {'lr': 0.000499108089319605, 'samples': 632064, 'steps': 3291, 'loss/train': 2.3120752573013306} 01/28/2022 17:07:34 - INFO - codeparrot_training - Step 3292: {'lr': 0.0004991067078720899, 'samples': 632256, 'steps': 3292, 'loss/train': 2.1565809845924377} 01/28/2022 17:07:38 - INFO - codeparrot_training - Step 3293: {'lr': 0.0004991053253574807, 'samples': 632448, 'steps': 3293, 'loss/train': 0.9434214234352112} 01/28/2022 17:07:43 - INFO - codeparrot_training - Step 3294: {'lr': 0.0004991039417757833, 'samples': 632640, 'steps': 3294, 'loss/train': 1.8608840703964233} 01/28/2022 17:07:47 - INFO - codeparrot_training - Step 3295: {'lr': 0.0004991025571270039, 'samples': 632832, 'steps': 3295, 'loss/train': 1.3081474900245667} 01/28/2022 17:07:51 - INFO - codeparrot_training - Step 3296: {'lr': 0.000499101171411148, 'samples': 633024, 'steps': 3296, 'loss/train': 2.5257423520088196} 01/28/2022 17:07:56 - INFO - codeparrot_training - Step 3297: {'lr': 0.000499099784628222, 'samples': 633216, 'steps': 3297, 'loss/train': 2.0969755053520203} 01/28/2022 17:08:01 - INFO - codeparrot_training - Step 3298: {'lr': 0.0004990983967782316, 'samples': 633408, 'steps': 3298, 'loss/train': 1.8612867593765259} 01/28/2022 17:08:05 - INFO - codeparrot_training - Step 3299: {'lr': 0.0004990970078611827, 'samples': 633600, 'steps': 3299, 'loss/train': 3.3692450523376465} 01/28/2022 17:08:09 - INFO - codeparrot_training - Step 3300: {'lr': 0.0004990956178770814, 'samples': 633792, 'steps': 3300, 'loss/train': 2.242097318172455} 01/28/2022 17:08:13 - INFO - codeparrot_training - Step 3301: {'lr': 0.0004990942268259335, 'samples': 633984, 'steps': 3301, 'loss/train': 2.0008413791656494} 01/28/2022 17:08:19 - INFO - codeparrot_training - Step 3302: {'lr': 0.000499092834707745, 'samples': 634176, 'steps': 3302, 'loss/train': 1.3897422552108765} 01/28/2022 17:08:23 - INFO - codeparrot_training - Step 3303: {'lr': 0.000499091441522522, 'samples': 634368, 'steps': 3303, 'loss/train': 1.7571174502372742} 01/28/2022 17:08:27 - INFO - codeparrot_training - Step 3304: {'lr': 0.0004990900472702702, 'samples': 634560, 'steps': 3304, 'loss/train': 1.7641659379005432} 01/28/2022 17:08:31 - INFO - codeparrot_training - Step 3305: {'lr': 0.0004990886519509959, 'samples': 634752, 'steps': 3305, 'loss/train': 2.1614758372306824} 01/28/2022 17:08:36 - INFO - codeparrot_training - Step 3306: {'lr': 0.0004990872555647048, 'samples': 634944, 'steps': 3306, 'loss/train': 2.0856850147247314} 01/28/2022 17:08:42 - INFO - codeparrot_training - Step 3307: {'lr': 0.0004990858581114029, 'samples': 635136, 'steps': 3307, 'loss/train': 2.321623742580414} 01/28/2022 17:08:46 - INFO - codeparrot_training - Step 3308: {'lr': 0.0004990844595910965, 'samples': 635328, 'steps': 3308, 'loss/train': 1.7495031952857971} 01/28/2022 17:08:51 - INFO - codeparrot_training - Step 3309: {'lr': 0.0004990830600037912, 'samples': 635520, 'steps': 3309, 'loss/train': 1.766127347946167} 01/28/2022 17:08:55 - INFO - codeparrot_training - Step 3310: {'lr': 0.0004990816593494933, 'samples': 635712, 'steps': 3310, 'loss/train': 0.09718338400125504} 01/28/2022 17:08:59 - INFO - codeparrot_training - Step 3311: {'lr': 0.0004990802576282085, 'samples': 635904, 'steps': 3311, 'loss/train': 1.8830480575561523} 01/28/2022 17:09:04 - INFO - codeparrot_training - Step 3312: {'lr': 0.0004990788548399431, 'samples': 636096, 'steps': 3312, 'loss/train': 1.4995155930519104} 01/28/2022 17:09:08 - INFO - codeparrot_training - Step 3313: {'lr': 0.0004990774509847029, 'samples': 636288, 'steps': 3313, 'loss/train': 1.1878799200057983} 01/28/2022 17:09:13 - INFO - codeparrot_training - Step 3314: {'lr': 0.0004990760460624941, 'samples': 636480, 'steps': 3314, 'loss/train': 2.0652807354927063} 01/28/2022 17:09:17 - INFO - codeparrot_training - Step 3315: {'lr': 0.0004990746400733225, 'samples': 636672, 'steps': 3315, 'loss/train': 1.7448808550834656} 01/28/2022 17:09:21 - INFO - codeparrot_training - Step 3316: {'lr': 0.0004990732330171943, 'samples': 636864, 'steps': 3316, 'loss/train': 1.867782711982727} 01/28/2022 17:09:28 - INFO - codeparrot_training - Step 3317: {'lr': 0.0004990718248941154, 'samples': 637056, 'steps': 3317, 'loss/train': 1.7459224462509155} 01/28/2022 17:09:32 - INFO - codeparrot_training - Step 3318: {'lr': 0.0004990704157040919, 'samples': 637248, 'steps': 3318, 'loss/train': 7.12051248550415} 01/28/2022 17:09:36 - INFO - codeparrot_training - Step 3319: {'lr': 0.0004990690054471299, 'samples': 637440, 'steps': 3319, 'loss/train': 0.3939906656742096} 01/28/2022 17:09:40 - INFO - codeparrot_training - Step 3320: {'lr': 0.0004990675941232354, 'samples': 637632, 'steps': 3320, 'loss/train': 2.0432499647140503} 01/28/2022 17:09:44 - INFO - codeparrot_training - Step 3321: {'lr': 0.0004990661817324142, 'samples': 637824, 'steps': 3321, 'loss/train': 2.5348645448684692} 01/28/2022 17:09:49 - INFO - codeparrot_training - Step 3322: {'lr': 0.0004990647682746727, 'samples': 638016, 'steps': 3322, 'loss/train': 8.626235246658325} 01/28/2022 17:09:54 - INFO - codeparrot_training - Step 3323: {'lr': 0.0004990633537500169, 'samples': 638208, 'steps': 3323, 'loss/train': 1.013859361410141} 01/28/2022 17:09:58 - INFO - codeparrot_training - Step 3324: {'lr': 0.0004990619381584527, 'samples': 638400, 'steps': 3324, 'loss/train': 2.1353155374526978} 01/28/2022 17:10:03 - INFO - codeparrot_training - Step 3325: {'lr': 0.0004990605214999862, 'samples': 638592, 'steps': 3325, 'loss/train': 0.8531077802181244} 01/28/2022 17:10:07 - INFO - codeparrot_training - Step 3326: {'lr': 0.0004990591037746236, 'samples': 638784, 'steps': 3326, 'loss/train': 2.036843776702881} 01/28/2022 17:10:11 - INFO - codeparrot_training - Step 3327: {'lr': 0.0004990576849823708, 'samples': 638976, 'steps': 3327, 'loss/train': 1.5980284810066223} 01/28/2022 17:10:17 - INFO - codeparrot_training - Step 3328: {'lr': 0.000499056265123234, 'samples': 639168, 'steps': 3328, 'loss/train': 2.093025326728821} 01/28/2022 17:10:21 - INFO - codeparrot_training - Step 3329: {'lr': 0.0004990548441972193, 'samples': 639360, 'steps': 3329, 'loss/train': 1.8807333111763} 01/28/2022 17:10:25 - INFO - codeparrot_training - Step 3330: {'lr': 0.0004990534222043325, 'samples': 639552, 'steps': 3330, 'loss/train': 1.6589241027832031} 01/28/2022 17:10:29 - INFO - codeparrot_training - Step 3331: {'lr': 0.0004990519991445803, 'samples': 639744, 'steps': 3331, 'loss/train': 3.618744134902954} 01/28/2022 17:10:35 - INFO - codeparrot_training - Step 3332: {'lr': 0.0004990505750179682, 'samples': 639936, 'steps': 3332, 'loss/train': 1.8519964814186096} 01/28/2022 17:10:40 - INFO - codeparrot_training - Step 3333: {'lr': 0.0004990491498245024, 'samples': 640128, 'steps': 3333, 'loss/train': 2.599949598312378} 01/28/2022 17:10:44 - INFO - codeparrot_training - Step 3334: {'lr': 0.0004990477235641893, 'samples': 640320, 'steps': 3334, 'loss/train': 2.171311855316162} 01/28/2022 17:10:48 - INFO - codeparrot_training - Step 3335: {'lr': 0.0004990462962370347, 'samples': 640512, 'steps': 3335, 'loss/train': 0.3502933010458946} 01/28/2022 17:10:52 - INFO - codeparrot_training - Step 3336: {'lr': 0.0004990448678430451, 'samples': 640704, 'steps': 3336, 'loss/train': 2.3106245398521423} 01/28/2022 17:10:57 - INFO - codeparrot_training - Step 3337: {'lr': 0.0004990434383822261, 'samples': 640896, 'steps': 3337, 'loss/train': 3.699058771133423} 01/28/2022 17:11:02 - INFO - codeparrot_training - Step 3338: {'lr': 0.0004990420078545843, 'samples': 641088, 'steps': 3338, 'loss/train': 2.982689917087555} 01/28/2022 17:11:06 - INFO - codeparrot_training - Step 3339: {'lr': 0.0004990405762601254, 'samples': 641280, 'steps': 3339, 'loss/train': 2.237691879272461} 01/28/2022 17:11:10 - INFO - codeparrot_training - Step 3340: {'lr': 0.000499039143598856, 'samples': 641472, 'steps': 3340, 'loss/train': 2.4430684447288513} 01/28/2022 17:11:15 - INFO - codeparrot_training - Step 3341: {'lr': 0.0004990377098707818, 'samples': 641664, 'steps': 3341, 'loss/train': 2.863072156906128} 01/28/2022 17:11:19 - INFO - codeparrot_training - Step 3342: {'lr': 0.0004990362750759092, 'samples': 641856, 'steps': 3342, 'loss/train': 2.449101448059082} 01/28/2022 17:11:24 - INFO - codeparrot_training - Step 3343: {'lr': 0.0004990348392142443, 'samples': 642048, 'steps': 3343, 'loss/train': 1.7457863688468933} 01/28/2022 17:11:28 - INFO - codeparrot_training - Step 3344: {'lr': 0.0004990334022857932, 'samples': 642240, 'steps': 3344, 'loss/train': 1.7449695467948914} 01/28/2022 17:11:32 - INFO - codeparrot_training - Step 3345: {'lr': 0.0004990319642905619, 'samples': 642432, 'steps': 3345, 'loss/train': 1.2565177381038666} 01/28/2022 17:11:37 - INFO - codeparrot_training - Step 3346: {'lr': 0.000499030525228557, 'samples': 642624, 'steps': 3346, 'loss/train': 2.397253632545471} 01/28/2022 17:11:41 - INFO - codeparrot_training - Step 3347: {'lr': 0.0004990290850997843, 'samples': 642816, 'steps': 3347, 'loss/train': 2.4518641233444214} 01/28/2022 17:11:47 - INFO - codeparrot_training - Step 3348: {'lr': 0.0004990276439042501, 'samples': 643008, 'steps': 3348, 'loss/train': 2.1496564149856567} 01/28/2022 17:11:51 - INFO - codeparrot_training - Step 3349: {'lr': 0.0004990262016419606, 'samples': 643200, 'steps': 3349, 'loss/train': 1.4485524594783783} 01/28/2022 17:11:55 - INFO - codeparrot_training - Step 3350: {'lr': 0.0004990247583129218, 'samples': 643392, 'steps': 3350, 'loss/train': 0.9809221029281616} 01/28/2022 17:11:59 - INFO - codeparrot_training - Step 3351: {'lr': 0.00049902331391714, 'samples': 643584, 'steps': 3351, 'loss/train': 2.5715999007225037} 01/28/2022 17:12:04 - INFO - codeparrot_training - Step 3352: {'lr': 0.0004990218684546216, 'samples': 643776, 'steps': 3352, 'loss/train': 1.9910452365875244} 01/28/2022 17:12:09 - INFO - codeparrot_training - Step 3353: {'lr': 0.0004990204219253724, 'samples': 643968, 'steps': 3353, 'loss/train': 0.48977574706077576} 01/28/2022 17:12:13 - INFO - codeparrot_training - Step 3354: {'lr': 0.0004990189743293989, 'samples': 644160, 'steps': 3354, 'loss/train': 2.4454148411750793} 01/28/2022 17:12:17 - INFO - codeparrot_training - Step 3355: {'lr': 0.0004990175256667071, 'samples': 644352, 'steps': 3355, 'loss/train': 2.2507004141807556} 01/28/2022 17:12:22 - INFO - codeparrot_training - Step 3356: {'lr': 0.0004990160759373033, 'samples': 644544, 'steps': 3356, 'loss/train': 1.8122470378875732} 01/28/2022 17:12:26 - INFO - codeparrot_training - Step 3357: {'lr': 0.0004990146251411938, 'samples': 644736, 'steps': 3357, 'loss/train': 1.8180001974105835} 01/28/2022 17:12:31 - INFO - codeparrot_training - Step 3358: {'lr': 0.0004990131732783846, 'samples': 644928, 'steps': 3358, 'loss/train': 2.752441942691803} 01/28/2022 17:12:36 - INFO - codeparrot_training - Step 3359: {'lr': 0.000499011720348882, 'samples': 645120, 'steps': 3359, 'loss/train': 0.3092534467577934} 01/28/2022 17:12:40 - INFO - codeparrot_training - Step 3360: {'lr': 0.0004990102663526924, 'samples': 645312, 'steps': 3360, 'loss/train': 1.8060044646263123} 01/28/2022 17:12:44 - INFO - codeparrot_training - Step 3361: {'lr': 0.0004990088112898219, 'samples': 645504, 'steps': 3361, 'loss/train': 2.484902322292328} 01/28/2022 17:12:48 - INFO - codeparrot_training - Step 3362: {'lr': 0.0004990073551602766, 'samples': 645696, 'steps': 3362, 'loss/train': 1.9503466486930847} 01/28/2022 17:12:54 - INFO - codeparrot_training - Step 3363: {'lr': 0.000499005897964063, 'samples': 645888, 'steps': 3363, 'loss/train': 2.644558310508728} 01/28/2022 17:12:58 - INFO - codeparrot_training - Step 3364: {'lr': 0.0004990044397011871, 'samples': 646080, 'steps': 3364, 'loss/train': 1.6711657047271729} 01/28/2022 17:13:02 - INFO - codeparrot_training - Step 3365: {'lr': 0.0004990029803716552, 'samples': 646272, 'steps': 3365, 'loss/train': 2.541266083717346} 01/28/2022 17:13:07 - INFO - codeparrot_training - Step 3366: {'lr': 0.0004990015199754736, 'samples': 646464, 'steps': 3366, 'loss/train': 2.147647976875305} 01/28/2022 17:13:11 - INFO - codeparrot_training - Step 3367: {'lr': 0.0004990000585126486, 'samples': 646656, 'steps': 3367, 'loss/train': 2.2044734358787537} 01/28/2022 17:13:16 - INFO - codeparrot_training - Step 3368: {'lr': 0.0004989985959831865, 'samples': 646848, 'steps': 3368, 'loss/train': 1.4757499694824219} 01/28/2022 17:13:20 - INFO - codeparrot_training - Step 3369: {'lr': 0.0004989971323870934, 'samples': 647040, 'steps': 3369, 'loss/train': 1.3800580501556396} 01/28/2022 17:13:24 - INFO - codeparrot_training - Step 3370: {'lr': 0.0004989956677243757, 'samples': 647232, 'steps': 3370, 'loss/train': 2.216049313545227} 01/28/2022 17:13:28 - INFO - codeparrot_training - Step 3371: {'lr': 0.0004989942019950395, 'samples': 647424, 'steps': 3371, 'loss/train': 1.4356082081794739} 01/28/2022 17:13:33 - INFO - codeparrot_training - Step 3372: {'lr': 0.0004989927351990912, 'samples': 647616, 'steps': 3372, 'loss/train': 1.1736498177051544} 01/28/2022 17:13:38 - INFO - codeparrot_training - Step 3373: {'lr': 0.0004989912673365373, 'samples': 647808, 'steps': 3373, 'loss/train': 1.657048523426056} 01/28/2022 17:13:42 - INFO - codeparrot_training - Step 3374: {'lr': 0.0004989897984073837, 'samples': 648000, 'steps': 3374, 'loss/train': 1.699204444885254} 01/28/2022 17:13:46 - INFO - codeparrot_training - Step 3375: {'lr': 0.000498988328411637, 'samples': 648192, 'steps': 3375, 'loss/train': 0.9371498823165894} 01/28/2022 17:13:51 - INFO - codeparrot_training - Step 3376: {'lr': 0.0004989868573493032, 'samples': 648384, 'steps': 3376, 'loss/train': 2.044402241706848} 01/28/2022 17:13:57 - INFO - codeparrot_training - Step 3377: {'lr': 0.0004989853852203889, 'samples': 648576, 'steps': 3377, 'loss/train': 2.697217583656311} 01/28/2022 17:14:01 - INFO - codeparrot_training - Step 3378: {'lr': 0.0004989839120249002, 'samples': 648768, 'steps': 3378, 'loss/train': 2.605935573577881} 01/28/2022 17:14:05 - INFO - codeparrot_training - Step 3379: {'lr': 0.0004989824377628435, 'samples': 648960, 'steps': 3379, 'loss/train': 1.8733398914337158} 01/28/2022 17:14:09 - INFO - codeparrot_training - Step 3380: {'lr': 0.0004989809624342251, 'samples': 649152, 'steps': 3380, 'loss/train': 2.111239492893219} 01/28/2022 17:14:14 - INFO - codeparrot_training - Step 3381: {'lr': 0.0004989794860390513, 'samples': 649344, 'steps': 3381, 'loss/train': 2.102760672569275} 01/28/2022 17:14:18 - INFO - codeparrot_training - Step 3382: {'lr': 0.0004989780085773285, 'samples': 649536, 'steps': 3382, 'loss/train': 0.7222752571105957} 01/28/2022 17:14:23 - INFO - codeparrot_training - Step 3383: {'lr': 0.0004989765300490628, 'samples': 649728, 'steps': 3383, 'loss/train': 2.256364166736603} 01/28/2022 17:14:27 - INFO - codeparrot_training - Step 3384: {'lr': 0.0004989750504542609, 'samples': 649920, 'steps': 3384, 'loss/train': 2.0207850337028503} 01/28/2022 17:14:31 - INFO - codeparrot_training - Step 3385: {'lr': 0.0004989735697929289, 'samples': 650112, 'steps': 3385, 'loss/train': 1.3482733368873596} 01/28/2022 17:14:36 - INFO - codeparrot_training - Step 3386: {'lr': 0.0004989720880650731, 'samples': 650304, 'steps': 3386, 'loss/train': 1.6406158804893494} 01/28/2022 17:14:41 - INFO - codeparrot_training - Step 3387: {'lr': 0.0004989706052707, 'samples': 650496, 'steps': 3387, 'loss/train': 2.279993176460266} 01/28/2022 17:14:46 - INFO - codeparrot_training - Step 3388: {'lr': 0.0004989691214098158, 'samples': 650688, 'steps': 3388, 'loss/train': 1.1091732680797577} 01/28/2022 17:14:50 - INFO - codeparrot_training - Step 3389: {'lr': 0.0004989676364824271, 'samples': 650880, 'steps': 3389, 'loss/train': 1.8113831877708435} 01/28/2022 17:14:54 - INFO - codeparrot_training - Step 3390: {'lr': 0.00049896615048854, 'samples': 651072, 'steps': 3390, 'loss/train': 1.9315161108970642} 01/28/2022 17:14:58 - INFO - codeparrot_training - Step 3391: {'lr': 0.000498964663428161, 'samples': 651264, 'steps': 3391, 'loss/train': 1.8700686693191528} 01/28/2022 17:15:03 - INFO - codeparrot_training - Step 3392: {'lr': 0.0004989631753012964, 'samples': 651456, 'steps': 3392, 'loss/train': 2.2014142870903015} 01/28/2022 17:15:07 - INFO - codeparrot_training - Step 3393: {'lr': 0.0004989616861079527, 'samples': 651648, 'steps': 3393, 'loss/train': 1.1970796287059784} 01/28/2022 17:15:12 - INFO - codeparrot_training - Step 3394: {'lr': 0.0004989601958481361, 'samples': 651840, 'steps': 3394, 'loss/train': 1.9505082964897156} 01/28/2022 17:15:16 - INFO - codeparrot_training - Step 3395: {'lr': 0.000498958704521853, 'samples': 652032, 'steps': 3395, 'loss/train': 1.5633834600448608} 01/28/2022 17:15:20 - INFO - codeparrot_training - Step 3396: {'lr': 0.00049895721212911, 'samples': 652224, 'steps': 3396, 'loss/train': 2.0146341919898987} 01/28/2022 17:15:25 - INFO - codeparrot_training - Step 3397: {'lr': 0.0004989557186699133, 'samples': 652416, 'steps': 3397, 'loss/train': 2.4102854132652283} 01/28/2022 17:15:29 - INFO - codeparrot_training - Step 3398: {'lr': 0.0004989542241442695, 'samples': 652608, 'steps': 3398, 'loss/train': 2.6903244853019714} 01/28/2022 17:15:34 - INFO - codeparrot_training - Step 3399: {'lr': 0.0004989527285521846, 'samples': 652800, 'steps': 3399, 'loss/train': 0.5938628911972046} 01/28/2022 17:15:38 - INFO - codeparrot_training - Step 3400: {'lr': 0.0004989512318936654, 'samples': 652992, 'steps': 3400, 'loss/train': 2.4004417061805725} 01/28/2022 17:15:42 - INFO - codeparrot_training - Step 3401: {'lr': 0.0004989497341687182, 'samples': 653184, 'steps': 3401, 'loss/train': 1.9415683150291443} 01/28/2022 17:15:48 - INFO - codeparrot_training - Step 3402: {'lr': 0.0004989482353773494, 'samples': 653376, 'steps': 3402, 'loss/train': 2.366325795650482} 01/28/2022 17:15:52 - INFO - codeparrot_training - Step 3403: {'lr': 0.0004989467355195653, 'samples': 653568, 'steps': 3403, 'loss/train': 2.2178640961647034} 01/28/2022 17:15:56 - INFO - codeparrot_training - Step 3404: {'lr': 0.0004989452345953725, 'samples': 653760, 'steps': 3404, 'loss/train': 2.1000795364379883} 01/28/2022 17:16:01 - INFO - codeparrot_training - Step 3405: {'lr': 0.0004989437326047774, 'samples': 653952, 'steps': 3405, 'loss/train': 0.847919225692749} 01/28/2022 17:16:05 - INFO - codeparrot_training - Step 3406: {'lr': 0.0004989422295477863, 'samples': 654144, 'steps': 3406, 'loss/train': 1.3154742121696472} 01/28/2022 17:16:10 - INFO - codeparrot_training - Step 3407: {'lr': 0.0004989407254244058, 'samples': 654336, 'steps': 3407, 'loss/train': 1.6152822375297546} 01/28/2022 17:16:14 - INFO - codeparrot_training - Step 3408: {'lr': 0.0004989392202346424, 'samples': 654528, 'steps': 3408, 'loss/train': 1.9891594648361206} 01/28/2022 17:16:18 - INFO - codeparrot_training - Step 3409: {'lr': 0.0004989377139785022, 'samples': 654720, 'steps': 3409, 'loss/train': 1.2172547578811646} 01/28/2022 17:16:22 - INFO - codeparrot_training - Step 3410: {'lr': 0.000498936206655992, 'samples': 654912, 'steps': 3410, 'loss/train': 1.9717960953712463} 01/28/2022 17:16:27 - INFO - codeparrot_training - Step 3411: {'lr': 0.0004989346982671181, 'samples': 655104, 'steps': 3411, 'loss/train': 1.6844238638877869} 01/28/2022 17:16:32 - INFO - codeparrot_training - Step 3412: {'lr': 0.0004989331888118869, 'samples': 655296, 'steps': 3412, 'loss/train': 2.0002121329307556} 01/28/2022 17:16:36 - INFO - codeparrot_training - Step 3413: {'lr': 0.0004989316782903052, 'samples': 655488, 'steps': 3413, 'loss/train': 2.023241400718689} 01/28/2022 17:16:40 - INFO - codeparrot_training - Step 3414: {'lr': 0.0004989301667023791, 'samples': 655680, 'steps': 3414, 'loss/train': 1.5562575459480286} 01/28/2022 17:16:44 - INFO - codeparrot_training - Step 3415: {'lr': 0.0004989286540481152, 'samples': 655872, 'steps': 3415, 'loss/train': 2.3314077258110046} 01/28/2022 17:16:49 - INFO - codeparrot_training - Step 3416: {'lr': 0.00049892714032752, 'samples': 656064, 'steps': 3416, 'loss/train': 2.2406092286109924} 01/28/2022 17:16:54 - INFO - codeparrot_training - Step 3417: {'lr': 0.0004989256255406001, 'samples': 656256, 'steps': 3417, 'loss/train': 1.0912992060184479} 01/28/2022 17:16:58 - INFO - codeparrot_training - Step 3418: {'lr': 0.0004989241096873617, 'samples': 656448, 'steps': 3418, 'loss/train': 2.0118232369422913} 01/28/2022 17:17:02 - INFO - codeparrot_training - Step 3419: {'lr': 0.0004989225927678115, 'samples': 656640, 'steps': 3419, 'loss/train': 1.0999403893947601} 01/28/2022 17:17:06 - INFO - codeparrot_training - Step 3420: {'lr': 0.000498921074781956, 'samples': 656832, 'steps': 3420, 'loss/train': 1.6501961946487427} 01/28/2022 17:17:11 - INFO - codeparrot_training - Step 3421: {'lr': 0.0004989195557298016, 'samples': 657024, 'steps': 3421, 'loss/train': 0.49570874869823456} 01/28/2022 17:17:17 - INFO - codeparrot_training - Step 3422: {'lr': 0.0004989180356113549, 'samples': 657216, 'steps': 3422, 'loss/train': 4.759301662445068} 01/28/2022 17:17:22 - INFO - codeparrot_training - Step 3423: {'lr': 0.0004989165144266224, 'samples': 657408, 'steps': 3423, 'loss/train': 0.9608408510684967} 01/28/2022 17:17:26 - INFO - codeparrot_training - Step 3424: {'lr': 0.0004989149921756105, 'samples': 657600, 'steps': 3424, 'loss/train': 2.7822346687316895} 01/28/2022 17:17:30 - INFO - codeparrot_training - Step 3425: {'lr': 0.0004989134688583259, 'samples': 657792, 'steps': 3425, 'loss/train': 2.8751474618911743} 01/28/2022 17:17:34 - INFO - codeparrot_training - Step 3426: {'lr': 0.000498911944474775, 'samples': 657984, 'steps': 3426, 'loss/train': 0.822127103805542} 01/28/2022 17:17:39 - INFO - codeparrot_training - Step 3427: {'lr': 0.0004989104190249643, 'samples': 658176, 'steps': 3427, 'loss/train': 1.7179160714149475} 01/28/2022 17:17:43 - INFO - codeparrot_training - Step 3428: {'lr': 0.0004989088925089005, 'samples': 658368, 'steps': 3428, 'loss/train': 1.9321238994598389} 01/28/2022 17:17:48 - INFO - codeparrot_training - Step 3429: {'lr': 0.00049890736492659, 'samples': 658560, 'steps': 3429, 'loss/train': 1.537576675415039} 01/28/2022 17:17:52 - INFO - codeparrot_training - Step 3430: {'lr': 0.0004989058362780394, 'samples': 658752, 'steps': 3430, 'loss/train': 2.026431441307068} 01/28/2022 17:17:56 - INFO - codeparrot_training - Step 3431: {'lr': 0.0004989043065632552, 'samples': 658944, 'steps': 3431, 'loss/train': 2.4100053906440735} 01/28/2022 17:18:01 - INFO - codeparrot_training - Step 3432: {'lr': 0.0004989027757822441, 'samples': 659136, 'steps': 3432, 'loss/train': 1.2026768624782562} 01/28/2022 17:18:05 - INFO - codeparrot_training - Step 3433: {'lr': 0.0004989012439350124, 'samples': 659328, 'steps': 3433, 'loss/train': 2.3470977544784546} 01/28/2022 17:18:10 - INFO - codeparrot_training - Step 3434: {'lr': 0.0004988997110215668, 'samples': 659520, 'steps': 3434, 'loss/train': 2.129858136177063} 01/28/2022 17:18:14 - INFO - codeparrot_training - Step 3435: {'lr': 0.0004988981770419141, 'samples': 659712, 'steps': 3435, 'loss/train': 1.9553396701812744} 01/28/2022 17:18:18 - INFO - codeparrot_training - Step 3436: {'lr': 0.0004988966419960605, 'samples': 659904, 'steps': 3436, 'loss/train': 1.8121781945228577} 01/28/2022 17:18:24 - INFO - codeparrot_training - Step 3437: {'lr': 0.0004988951058840127, 'samples': 660096, 'steps': 3437, 'loss/train': 1.4100684821605682} 01/28/2022 17:18:28 - INFO - codeparrot_training - Step 3438: {'lr': 0.0004988935687057773, 'samples': 660288, 'steps': 3438, 'loss/train': 2.096659541130066} 01/28/2022 17:18:32 - INFO - codeparrot_training - Step 3439: {'lr': 0.0004988920304613609, 'samples': 660480, 'steps': 3439, 'loss/train': 2.6210821866989136} 01/28/2022 17:18:37 - INFO - codeparrot_training - Step 3440: {'lr': 0.00049889049115077, 'samples': 660672, 'steps': 3440, 'loss/train': 1.0324238240718842} 01/28/2022 17:18:41 - INFO - codeparrot_training - Step 3441: {'lr': 0.0004988889507740113, 'samples': 660864, 'steps': 3441, 'loss/train': 2.098244369029999} 01/28/2022 17:18:46 - INFO - codeparrot_training - Step 3442: {'lr': 0.0004988874093310914, 'samples': 661056, 'steps': 3442, 'loss/train': 2.313957989215851} 01/28/2022 17:18:50 - INFO - codeparrot_training - Step 3443: {'lr': 0.000498885866822017, 'samples': 661248, 'steps': 3443, 'loss/train': 2.0055649280548096} 01/28/2022 17:18:54 - INFO - codeparrot_training - Step 3444: {'lr': 0.0004988843232467944, 'samples': 661440, 'steps': 3444, 'loss/train': 2.2558842301368713} 01/28/2022 17:18:58 - INFO - codeparrot_training - Step 3445: {'lr': 0.0004988827786054304, 'samples': 661632, 'steps': 3445, 'loss/train': 2.2999207377433777} 01/28/2022 17:19:03 - INFO - codeparrot_training - Step 3446: {'lr': 0.0004988812328979317, 'samples': 661824, 'steps': 3446, 'loss/train': 1.2451170086860657} 01/28/2022 17:19:09 - INFO - codeparrot_training - Step 3447: {'lr': 0.0004988796861243046, 'samples': 662016, 'steps': 3447, 'loss/train': 2.215763211250305} 01/28/2022 17:19:13 - INFO - codeparrot_training - Step 3448: {'lr': 0.0004988781382845562, 'samples': 662208, 'steps': 3448, 'loss/train': 1.8719658851623535} 01/28/2022 17:19:17 - INFO - codeparrot_training - Step 3449: {'lr': 0.0004988765893786929, 'samples': 662400, 'steps': 3449, 'loss/train': 8.216511726379395} 01/28/2022 17:19:21 - INFO - codeparrot_training - Step 3450: {'lr': 0.0004988750394067211, 'samples': 662592, 'steps': 3450, 'loss/train': 2.0635483860969543} 01/28/2022 17:19:25 - INFO - codeparrot_training - Step 3451: {'lr': 0.0004988734883686479, 'samples': 662784, 'steps': 3451, 'loss/train': 0.820305347442627} 01/28/2022 17:19:30 - INFO - codeparrot_training - Step 3452: {'lr': 0.0004988719362644795, 'samples': 662976, 'steps': 3452, 'loss/train': 2.3089969754219055} 01/28/2022 17:19:35 - INFO - codeparrot_training - Step 3453: {'lr': 0.0004988703830942228, 'samples': 663168, 'steps': 3453, 'loss/train': 2.1474772095680237} 01/28/2022 17:19:39 - INFO - codeparrot_training - Step 3454: {'lr': 0.0004988688288578845, 'samples': 663360, 'steps': 3454, 'loss/train': 1.9353860020637512} 01/28/2022 17:19:43 - INFO - codeparrot_training - Step 3455: {'lr': 0.0004988672735554711, 'samples': 663552, 'steps': 3455, 'loss/train': 2.6051323413848877} 01/28/2022 17:19:47 - INFO - codeparrot_training - Step 3456: {'lr': 0.0004988657171869893, 'samples': 663744, 'steps': 3456, 'loss/train': 1.6557095646858215} 01/28/2022 17:19:52 - INFO - codeparrot_training - Step 3457: {'lr': 0.0004988641597524458, 'samples': 663936, 'steps': 3457, 'loss/train': 1.3730342388153076} 01/28/2022 17:19:57 - INFO - codeparrot_training - Step 3458: {'lr': 0.0004988626012518473, 'samples': 664128, 'steps': 3458, 'loss/train': 2.1816437244415283} 01/28/2022 17:20:01 - INFO - codeparrot_training - Step 3459: {'lr': 0.0004988610416852004, 'samples': 664320, 'steps': 3459, 'loss/train': 1.0457918643951416} 01/28/2022 17:20:05 - INFO - codeparrot_training - Step 3460: {'lr': 0.0004988594810525118, 'samples': 664512, 'steps': 3460, 'loss/train': 2.09352707862854} 01/28/2022 17:20:09 - INFO - codeparrot_training - Step 3461: {'lr': 0.0004988579193537883, 'samples': 664704, 'steps': 3461, 'loss/train': 2.7683043479919434} 01/28/2022 17:20:15 - INFO - codeparrot_training - Step 3462: {'lr': 0.0004988563565890364, 'samples': 664896, 'steps': 3462, 'loss/train': 2.1884363293647766} 01/28/2022 17:20:19 - INFO - codeparrot_training - Step 3463: {'lr': 0.000498854792758263, 'samples': 665088, 'steps': 3463, 'loss/train': 0.8884752988815308} 01/28/2022 17:20:24 - INFO - codeparrot_training - Step 3464: {'lr': 0.0004988532278614745, 'samples': 665280, 'steps': 3464, 'loss/train': 1.853257656097412} 01/28/2022 17:20:28 - INFO - codeparrot_training - Step 3465: {'lr': 0.0004988516618986779, 'samples': 665472, 'steps': 3465, 'loss/train': 2.421216666698456} 01/28/2022 17:20:32 - INFO - codeparrot_training - Step 3466: {'lr': 0.0004988500948698799, 'samples': 665664, 'steps': 3466, 'loss/train': 2.1682414412498474} 01/28/2022 17:20:37 - INFO - codeparrot_training - Step 3467: {'lr': 0.000498848526775087, 'samples': 665856, 'steps': 3467, 'loss/train': 1.221519112586975} 01/28/2022 17:20:41 - INFO - codeparrot_training - Step 3468: {'lr': 0.0004988469576143059, 'samples': 666048, 'steps': 3468, 'loss/train': 1.2538594007492065} 01/28/2022 17:20:46 - INFO - codeparrot_training - Step 3469: {'lr': 0.0004988453873875437, 'samples': 666240, 'steps': 3469, 'loss/train': 2.4184325337409973} 01/28/2022 17:20:50 - INFO - codeparrot_training - Step 3470: {'lr': 0.0004988438160948068, 'samples': 666432, 'steps': 3470, 'loss/train': 1.6453801989555359} 01/28/2022 17:20:54 - INFO - codeparrot_training - Step 3471: {'lr': 0.000498842243736102, 'samples': 666624, 'steps': 3471, 'loss/train': 0.6958514302968979} 01/28/2022 17:21:00 - INFO - codeparrot_training - Step 3472: {'lr': 0.000498840670311436, 'samples': 666816, 'steps': 3472, 'loss/train': 1.878269612789154} 01/28/2022 17:21:04 - INFO - codeparrot_training - Step 3473: {'lr': 0.0004988390958208156, 'samples': 667008, 'steps': 3473, 'loss/train': 2.635384261608124} 01/28/2022 17:21:08 - INFO - codeparrot_training - Step 3474: {'lr': 0.0004988375202642475, 'samples': 667200, 'steps': 3474, 'loss/train': 2.095986485481262} 01/28/2022 17:21:12 - INFO - codeparrot_training - Step 3475: {'lr': 0.0004988359436417385, 'samples': 667392, 'steps': 3475, 'loss/train': 1.4500085413455963} 01/28/2022 17:21:16 - INFO - codeparrot_training - Step 3476: {'lr': 0.0004988343659532954, 'samples': 667584, 'steps': 3476, 'loss/train': 2.2020825147628784} 01/28/2022 17:21:22 - INFO - codeparrot_training - Step 3477: {'lr': 0.0004988327871989249, 'samples': 667776, 'steps': 3477, 'loss/train': 2.2821121215820312} 01/28/2022 17:21:26 - INFO - codeparrot_training - Step 3478: {'lr': 0.0004988312073786336, 'samples': 667968, 'steps': 3478, 'loss/train': 2.328925609588623} 01/28/2022 17:21:30 - INFO - codeparrot_training - Step 3479: {'lr': 0.0004988296264924286, 'samples': 668160, 'steps': 3479, 'loss/train': 1.5067128539085388} 01/28/2022 17:21:34 - INFO - codeparrot_training - Step 3480: {'lr': 0.0004988280445403164, 'samples': 668352, 'steps': 3480, 'loss/train': 2.264088749885559} 01/28/2022 17:21:38 - INFO - codeparrot_training - Step 3481: {'lr': 0.0004988264615223038, 'samples': 668544, 'steps': 3481, 'loss/train': 2.068879008293152} 01/28/2022 17:21:43 - INFO - codeparrot_training - Step 3482: {'lr': 0.0004988248774383978, 'samples': 668736, 'steps': 3482, 'loss/train': 2.2325246930122375} 01/28/2022 17:21:48 - INFO - codeparrot_training - Step 3483: {'lr': 0.0004988232922886049, 'samples': 668928, 'steps': 3483, 'loss/train': 1.262401968240738} 01/28/2022 17:21:52 - INFO - codeparrot_training - Step 3484: {'lr': 0.0004988217060729321, 'samples': 669120, 'steps': 3484, 'loss/train': 1.694058895111084} 01/28/2022 17:21:56 - INFO - codeparrot_training - Step 3485: {'lr': 0.0004988201187913861, 'samples': 669312, 'steps': 3485, 'loss/train': 0.8507605791091919} 01/28/2022 17:22:00 - INFO - codeparrot_training - Step 3486: {'lr': 0.0004988185304439737, 'samples': 669504, 'steps': 3486, 'loss/train': 2.0036494731903076} 01/28/2022 17:22:06 - INFO - codeparrot_training - Step 3487: {'lr': 0.0004988169410307018, 'samples': 669696, 'steps': 3487, 'loss/train': 1.7545385956764221} 01/28/2022 17:22:10 - INFO - codeparrot_training - Step 3488: {'lr': 0.0004988153505515771, 'samples': 669888, 'steps': 3488, 'loss/train': 3.0729278326034546} 01/28/2022 17:22:14 - INFO - codeparrot_training - Step 3489: {'lr': 0.0004988137590066064, 'samples': 670080, 'steps': 3489, 'loss/train': 2.0414328575134277} 01/28/2022 17:22:18 - INFO - codeparrot_training - Step 3490: {'lr': 0.0004988121663957966, 'samples': 670272, 'steps': 3490, 'loss/train': 1.881406545639038} 01/28/2022 17:22:22 - INFO - codeparrot_training - Step 3491: {'lr': 0.0004988105727191546, 'samples': 670464, 'steps': 3491, 'loss/train': 1.555256187915802} 01/28/2022 17:22:28 - INFO - codeparrot_training - Step 3492: {'lr': 0.0004988089779766869, 'samples': 670656, 'steps': 3492, 'loss/train': 2.8615981936454773} 01/28/2022 17:22:32 - INFO - codeparrot_training - Step 3493: {'lr': 0.0004988073821684006, 'samples': 670848, 'steps': 3493, 'loss/train': 0.9363803565502167} 01/28/2022 17:22:37 - INFO - codeparrot_training - Step 3494: {'lr': 0.0004988057852943025, 'samples': 671040, 'steps': 3494, 'loss/train': 2.8265901803970337} 01/28/2022 17:22:41 - INFO - codeparrot_training - Step 3495: {'lr': 0.0004988041873543995, 'samples': 671232, 'steps': 3495, 'loss/train': 2.261656165122986} 01/28/2022 17:22:45 - INFO - codeparrot_training - Step 3496: {'lr': 0.0004988025883486983, 'samples': 671424, 'steps': 3496, 'loss/train': 2.454525053501129} 01/28/2022 17:22:50 - INFO - codeparrot_training - Step 3497: {'lr': 0.0004988009882772058, 'samples': 671616, 'steps': 3497, 'loss/train': 1.3802669048309326} 01/28/2022 17:22:54 - INFO - codeparrot_training - Step 3498: {'lr': 0.0004987993871399289, 'samples': 671808, 'steps': 3498, 'loss/train': 1.1648378670215607} 01/28/2022 17:22:58 - INFO - codeparrot_training - Step 3499: {'lr': 0.0004987977849368744, 'samples': 672000, 'steps': 3499, 'loss/train': 2.22487610578537} 01/28/2022 17:23:03 - INFO - codeparrot_training - Step 3500: {'lr': 0.0004987961816680492, 'samples': 672192, 'steps': 3500, 'loss/train': 2.2888280153274536} 01/28/2022 17:23:07 - INFO - codeparrot_training - Step 3501: {'lr': 0.0004987945773334602, 'samples': 672384, 'steps': 3501, 'loss/train': 2.454994261264801} 01/28/2022 17:23:12 - INFO - codeparrot_training - Step 3502: {'lr': 0.0004987929719331142, 'samples': 672576, 'steps': 3502, 'loss/train': 1.6597580909729004} 01/28/2022 17:23:16 - INFO - codeparrot_training - Step 3503: {'lr': 0.0004987913654670181, 'samples': 672768, 'steps': 3503, 'loss/train': 1.313587725162506} 01/28/2022 17:23:21 - INFO - codeparrot_training - Step 3504: {'lr': 0.0004987897579351787, 'samples': 672960, 'steps': 3504, 'loss/train': 1.2457574307918549} 01/28/2022 17:23:25 - INFO - codeparrot_training - Step 3505: {'lr': 0.0004987881493376032, 'samples': 673152, 'steps': 3505, 'loss/train': 2.4781667590141296} 01/28/2022 17:23:29 - INFO - codeparrot_training - Step 3506: {'lr': 0.0004987865396742981, 'samples': 673344, 'steps': 3506, 'loss/train': 2.2639524936676025} 01/28/2022 17:23:35 - INFO - codeparrot_training - Step 3507: {'lr': 0.0004987849289452705, 'samples': 673536, 'steps': 3507, 'loss/train': 1.966713309288025} 01/28/2022 17:23:39 - INFO - codeparrot_training - Step 3508: {'lr': 0.0004987833171505272, 'samples': 673728, 'steps': 3508, 'loss/train': 2.0188087821006775} 01/28/2022 17:23:44 - INFO - codeparrot_training - Step 3509: {'lr': 0.0004987817042900753, 'samples': 673920, 'steps': 3509, 'loss/train': 1.961076557636261} 01/28/2022 17:23:48 - INFO - codeparrot_training - Step 3510: {'lr': 0.0004987800903639216, 'samples': 674112, 'steps': 3510, 'loss/train': 1.9246158599853516} 01/28/2022 17:23:52 - INFO - codeparrot_training - Step 3511: {'lr': 0.0004987784753720728, 'samples': 674304, 'steps': 3511, 'loss/train': 2.066120982170105} 01/28/2022 17:23:58 - INFO - codeparrot_training - Step 3512: {'lr': 0.0004987768593145362, 'samples': 674496, 'steps': 3512, 'loss/train': 1.97176855802536} 01/28/2022 17:24:02 - INFO - codeparrot_training - Step 3513: {'lr': 0.0004987752421913185, 'samples': 674688, 'steps': 3513, 'loss/train': 1.1635562181472778} 01/28/2022 17:24:06 - INFO - codeparrot_training - Step 3514: {'lr': 0.0004987736240024264, 'samples': 674880, 'steps': 3514, 'loss/train': 2.856773614883423} 01/28/2022 17:24:10 - INFO - codeparrot_training - Step 3515: {'lr': 0.0004987720047478673, 'samples': 675072, 'steps': 3515, 'loss/train': 1.8944619297981262} 01/28/2022 17:24:14 - INFO - codeparrot_training - Step 3516: {'lr': 0.000498770384427648, 'samples': 675264, 'steps': 3516, 'loss/train': 1.9441768527030945} 01/28/2022 17:24:18 - INFO - codeparrot_training - Step 3517: {'lr': 0.0004987687630417753, 'samples': 675456, 'steps': 3517, 'loss/train': 2.1731613278388977} 01/28/2022 17:24:24 - INFO - codeparrot_training - Step 3518: {'lr': 0.0004987671405902562, 'samples': 675648, 'steps': 3518, 'loss/train': 1.8716518878936768} 01/28/2022 17:24:28 - INFO - codeparrot_training - Step 3519: {'lr': 0.0004987655170730976, 'samples': 675840, 'steps': 3519, 'loss/train': 1.5928863286972046} 01/28/2022 17:24:33 - INFO - codeparrot_training - Step 3520: {'lr': 0.0004987638924903066, 'samples': 676032, 'steps': 3520, 'loss/train': 2.178442418575287} 01/28/2022 17:24:37 - INFO - codeparrot_training - Step 3521: {'lr': 0.00049876226684189, 'samples': 676224, 'steps': 3521, 'loss/train': 1.9673836827278137} 01/28/2022 17:24:41 - INFO - codeparrot_training - Step 3522: {'lr': 0.0004987606401278549, 'samples': 676416, 'steps': 3522, 'loss/train': 1.2856073677539825} 01/28/2022 17:24:46 - INFO - codeparrot_training - Step 3523: {'lr': 0.0004987590123482082, 'samples': 676608, 'steps': 3523, 'loss/train': 1.3490085303783417} 01/28/2022 17:24:51 - INFO - codeparrot_training - Step 3524: {'lr': 0.0004987573835029569, 'samples': 676800, 'steps': 3524, 'loss/train': 2.041038393974304} 01/28/2022 17:24:55 - INFO - codeparrot_training - Step 3525: {'lr': 0.0004987557535921079, 'samples': 676992, 'steps': 3525, 'loss/train': 2.1192485690116882} 01/28/2022 17:24:59 - INFO - codeparrot_training - Step 3526: {'lr': 0.0004987541226156683, 'samples': 677184, 'steps': 3526, 'loss/train': 2.408555030822754} 01/28/2022 17:25:04 - INFO - codeparrot_training - Step 3527: {'lr': 0.0004987524905736451, 'samples': 677376, 'steps': 3527, 'loss/train': 2.0117856860160828} 01/28/2022 17:25:08 - INFO - codeparrot_training - Step 3528: {'lr': 0.000498750857466045, 'samples': 677568, 'steps': 3528, 'loss/train': 2.2586987614631653} 01/28/2022 17:25:12 - INFO - codeparrot_training - Step 3529: {'lr': 0.0004987492232928753, 'samples': 677760, 'steps': 3529, 'loss/train': 2.772103250026703} 01/28/2022 17:25:17 - INFO - codeparrot_training - Step 3530: {'lr': 0.000498747588054143, 'samples': 677952, 'steps': 3530, 'loss/train': 1.503096878528595} 01/28/2022 17:25:21 - INFO - codeparrot_training - Step 3531: {'lr': 0.0004987459517498549, 'samples': 678144, 'steps': 3531, 'loss/train': 1.8849028944969177} 01/28/2022 17:25:27 - INFO - codeparrot_training - Step 3532: {'lr': 0.0004987443143800182, 'samples': 678336, 'steps': 3532, 'loss/train': 2.6522567868232727} 01/28/2022 17:25:31 - INFO - codeparrot_training - Step 3533: {'lr': 0.0004987426759446398, 'samples': 678528, 'steps': 3533, 'loss/train': 1.9293556809425354} 01/28/2022 17:25:35 - INFO - codeparrot_training - Step 3534: {'lr': 0.0004987410364437269, 'samples': 678720, 'steps': 3534, 'loss/train': 2.1038201451301575} 01/28/2022 17:25:39 - INFO - codeparrot_training - Step 3535: {'lr': 0.0004987393958772862, 'samples': 678912, 'steps': 3535, 'loss/train': 2.707227408885956} 01/28/2022 17:25:44 - INFO - codeparrot_training - Step 3536: {'lr': 0.0004987377542453251, 'samples': 679104, 'steps': 3536, 'loss/train': 1.2146193087100983} 01/28/2022 17:25:49 - INFO - codeparrot_training - Step 3537: {'lr': 0.0004987361115478502, 'samples': 679296, 'steps': 3537, 'loss/train': 1.7102131247520447} 01/28/2022 17:25:53 - INFO - codeparrot_training - Step 3538: {'lr': 0.000498734467784869, 'samples': 679488, 'steps': 3538, 'loss/train': 2.2323004603385925} 01/28/2022 17:25:57 - INFO - codeparrot_training - Step 3539: {'lr': 0.0004987328229563883, 'samples': 679680, 'steps': 3539, 'loss/train': 1.3509880006313324} 01/28/2022 17:26:01 - INFO - codeparrot_training - Step 3540: {'lr': 0.0004987311770624151, 'samples': 679872, 'steps': 3540, 'loss/train': 2.9493541717529297} 01/28/2022 17:26:05 - INFO - codeparrot_training - Step 3541: {'lr': 0.0004987295301029565, 'samples': 680064, 'steps': 3541, 'loss/train': 2.4177491068840027} 01/28/2022 17:26:11 - INFO - codeparrot_training - Step 3542: {'lr': 0.0004987278820780196, 'samples': 680256, 'steps': 3542, 'loss/train': 1.9248953461647034} 01/28/2022 17:26:15 - INFO - codeparrot_training - Step 3543: {'lr': 0.0004987262329876114, 'samples': 680448, 'steps': 3543, 'loss/train': 2.053938388824463} 01/28/2022 17:26:20 - INFO - codeparrot_training - Step 3544: {'lr': 0.000498724582831739, 'samples': 680640, 'steps': 3544, 'loss/train': 1.5591952800750732} 01/28/2022 17:26:24 - INFO - codeparrot_training - Step 3545: {'lr': 0.0004987229316104095, 'samples': 680832, 'steps': 3545, 'loss/train': 1.8720828294754028} 01/28/2022 17:26:28 - INFO - codeparrot_training - Step 3546: {'lr': 0.00049872127932363, 'samples': 681024, 'steps': 3546, 'loss/train': 1.135939121246338} 01/28/2022 17:26:32 - INFO - codeparrot_training - Step 3547: {'lr': 0.0004987196259714074, 'samples': 681216, 'steps': 3547, 'loss/train': 2.7304592728614807} 01/28/2022 17:26:37 - INFO - codeparrot_training - Step 3548: {'lr': 0.000498717971553749, 'samples': 681408, 'steps': 3548, 'loss/train': 2.4203174114227295} 01/28/2022 17:26:41 - INFO - codeparrot_training - Step 3549: {'lr': 0.0004987163160706617, 'samples': 681600, 'steps': 3549, 'loss/train': 1.5480026006698608} 01/28/2022 17:26:46 - INFO - codeparrot_training - Step 3550: {'lr': 0.0004987146595221527, 'samples': 681792, 'steps': 3550, 'loss/train': 2.5356470346450806} 01/28/2022 17:26:50 - INFO - codeparrot_training - Step 3551: {'lr': 0.0004987130019082291, 'samples': 681984, 'steps': 3551, 'loss/train': 2.669632315635681} 01/28/2022 17:26:54 - INFO - codeparrot_training - Step 3552: {'lr': 0.000498711343228898, 'samples': 682176, 'steps': 3552, 'loss/train': 1.917211890220642} 01/28/2022 17:27:00 - INFO - codeparrot_training - Step 3553: {'lr': 0.0004987096834841665, 'samples': 682368, 'steps': 3553, 'loss/train': 1.8631346225738525} 01/28/2022 17:27:04 - INFO - codeparrot_training - Step 3554: {'lr': 0.0004987080226740416, 'samples': 682560, 'steps': 3554, 'loss/train': 2.172384023666382} 01/28/2022 17:27:08 - INFO - codeparrot_training - Step 3555: {'lr': 0.0004987063607985305, 'samples': 682752, 'steps': 3555, 'loss/train': 2.402872145175934} 01/28/2022 17:27:12 - INFO - codeparrot_training - Step 3556: {'lr': 0.0004987046978576404, 'samples': 682944, 'steps': 3556, 'loss/train': 2.363262176513672} 01/28/2022 17:27:17 - INFO - codeparrot_training - Step 3557: {'lr': 0.0004987030338513783, 'samples': 683136, 'steps': 3557, 'loss/train': 3.022042751312256} 01/28/2022 17:27:22 - INFO - codeparrot_training - Step 3558: {'lr': 0.0004987013687797514, 'samples': 683328, 'steps': 3558, 'loss/train': 2.347598612308502} 01/28/2022 17:27:26 - INFO - codeparrot_training - Step 3559: {'lr': 0.0004986997026427668, 'samples': 683520, 'steps': 3559, 'loss/train': 1.3807637393474579} 01/28/2022 17:27:30 - INFO - codeparrot_training - Step 3560: {'lr': 0.0004986980354404316, 'samples': 683712, 'steps': 3560, 'loss/train': 4.451919078826904} 01/28/2022 17:27:34 - INFO - codeparrot_training - Step 3561: {'lr': 0.000498696367172753, 'samples': 683904, 'steps': 3561, 'loss/train': 1.8853599429130554} 01/28/2022 17:27:39 - INFO - codeparrot_training - Step 3562: {'lr': 0.0004986946978397382, 'samples': 684096, 'steps': 3562, 'loss/train': 1.8823188543319702} 01/28/2022 17:27:44 - INFO - codeparrot_training - Step 3563: {'lr': 0.0004986930274413942, 'samples': 684288, 'steps': 3563, 'loss/train': 1.570859670639038} 01/28/2022 17:27:48 - INFO - codeparrot_training - Step 3564: {'lr': 0.0004986913559777283, 'samples': 684480, 'steps': 3564, 'loss/train': 2.436706781387329} 01/28/2022 17:27:52 - INFO - codeparrot_training - Step 3565: {'lr': 0.0004986896834487477, 'samples': 684672, 'steps': 3565, 'loss/train': 2.6730154752731323} 01/28/2022 17:27:56 - INFO - codeparrot_training - Step 3566: {'lr': 0.0004986880098544593, 'samples': 684864, 'steps': 3566, 'loss/train': 2.602716386318207} 01/28/2022 17:28:00 - INFO - codeparrot_training - Step 3567: {'lr': 0.0004986863351948705, 'samples': 685056, 'steps': 3567, 'loss/train': 1.553360939025879} 01/28/2022 17:28:07 - INFO - codeparrot_training - Step 3568: {'lr': 0.0004986846594699883, 'samples': 685248, 'steps': 3568, 'loss/train': 2.2541545629501343} 01/28/2022 17:28:11 - INFO - codeparrot_training - Step 3569: {'lr': 0.0004986829826798202, 'samples': 685440, 'steps': 3569, 'loss/train': 2.1277944445610046} 01/28/2022 17:28:15 - INFO - codeparrot_training - Step 3570: {'lr': 0.0004986813048243729, 'samples': 685632, 'steps': 3570, 'loss/train': 2.0282819867134094} 01/28/2022 17:28:19 - INFO - codeparrot_training - Step 3571: {'lr': 0.000498679625903654, 'samples': 685824, 'steps': 3571, 'loss/train': 1.8572051525115967} 01/28/2022 17:28:23 - INFO - codeparrot_training - Step 3572: {'lr': 0.0004986779459176706, 'samples': 686016, 'steps': 3572, 'loss/train': 2.1808943152427673} 01/28/2022 17:28:29 - INFO - codeparrot_training - Step 3573: {'lr': 0.0004986762648664298, 'samples': 686208, 'steps': 3573, 'loss/train': 1.7285395860671997} 01/28/2022 17:28:33 - INFO - codeparrot_training - Step 3574: {'lr': 0.0004986745827499389, 'samples': 686400, 'steps': 3574, 'loss/train': 1.2598965167999268} 01/28/2022 17:28:37 - INFO - codeparrot_training - Step 3575: {'lr': 0.0004986728995682049, 'samples': 686592, 'steps': 3575, 'loss/train': 1.7704875469207764} 01/28/2022 17:28:41 - INFO - codeparrot_training - Step 3576: {'lr': 0.0004986712153212352, 'samples': 686784, 'steps': 3576, 'loss/train': 2.437831699848175} 01/28/2022 17:28:45 - INFO - codeparrot_training - Step 3577: {'lr': 0.0004986695300090371, 'samples': 686976, 'steps': 3577, 'loss/train': 1.7063786387443542} 01/28/2022 17:28:51 - INFO - codeparrot_training - Step 3578: {'lr': 0.0004986678436316175, 'samples': 687168, 'steps': 3578, 'loss/train': 2.5085978507995605} 01/28/2022 17:28:55 - INFO - codeparrot_training - Step 3579: {'lr': 0.000498666156188984, 'samples': 687360, 'steps': 3579, 'loss/train': 2.0038994550704956} 01/28/2022 17:28:59 - INFO - codeparrot_training - Step 3580: {'lr': 0.0004986644676811436, 'samples': 687552, 'steps': 3580, 'loss/train': 2.188195824623108} 01/28/2022 17:29:04 - INFO - codeparrot_training - Step 3581: {'lr': 0.0004986627781081035, 'samples': 687744, 'steps': 3581, 'loss/train': 1.6065173149108887} 01/28/2022 17:29:08 - INFO - codeparrot_training - Step 3582: {'lr': 0.0004986610874698712, 'samples': 687936, 'steps': 3582, 'loss/train': 1.934566855430603} 01/28/2022 17:29:13 - INFO - codeparrot_training - Step 3583: {'lr': 0.0004986593957664536, 'samples': 688128, 'steps': 3583, 'loss/train': 1.555329144001007} 01/28/2022 17:29:17 - INFO - codeparrot_training - Step 3584: {'lr': 0.0004986577029978581, 'samples': 688320, 'steps': 3584, 'loss/train': 1.1975299715995789} 01/28/2022 17:29:21 - INFO - codeparrot_training - Step 3585: {'lr': 0.000498656009164092, 'samples': 688512, 'steps': 3585, 'loss/train': 2.3361076712608337} 01/28/2022 17:29:25 - INFO - codeparrot_training - Step 3586: {'lr': 0.0004986543142651625, 'samples': 688704, 'steps': 3586, 'loss/train': 1.5799049735069275} 01/28/2022 17:29:30 - INFO - codeparrot_training - Step 3587: {'lr': 0.0004986526183010769, 'samples': 688896, 'steps': 3587, 'loss/train': 1.1862407326698303} 01/28/2022 17:29:35 - INFO - codeparrot_training - Step 3588: {'lr': 0.0004986509212718425, 'samples': 689088, 'steps': 3588, 'loss/train': 0.8755650222301483} 01/28/2022 17:29:39 - INFO - codeparrot_training - Step 3589: {'lr': 0.0004986492231774664, 'samples': 689280, 'steps': 3589, 'loss/train': 2.7363327741622925} 01/28/2022 17:29:43 - INFO - codeparrot_training - Step 3590: {'lr': 0.0004986475240179559, 'samples': 689472, 'steps': 3590, 'loss/train': 1.9445895552635193} 01/28/2022 17:29:47 - INFO - codeparrot_training - Step 3591: {'lr': 0.0004986458237933185, 'samples': 689664, 'steps': 3591, 'loss/train': 1.343209594488144} 01/28/2022 17:29:51 - INFO - codeparrot_training - Step 3592: {'lr': 0.0004986441225035614, 'samples': 689856, 'steps': 3592, 'loss/train': 1.7274851202964783} 01/28/2022 17:29:57 - INFO - codeparrot_training - Step 3593: {'lr': 0.0004986424201486918, 'samples': 690048, 'steps': 3593, 'loss/train': 1.5080228447914124} 01/28/2022 17:30:02 - INFO - codeparrot_training - Step 3594: {'lr': 0.000498640716728717, 'samples': 690240, 'steps': 3594, 'loss/train': 0.6507744044065475} 01/28/2022 17:30:06 - INFO - codeparrot_training - Step 3595: {'lr': 0.0004986390122436443, 'samples': 690432, 'steps': 3595, 'loss/train': 2.6285637617111206} 01/28/2022 17:30:10 - INFO - codeparrot_training - Step 3596: {'lr': 0.000498637306693481, 'samples': 690624, 'steps': 3596, 'loss/train': 2.1375412344932556} 01/28/2022 17:30:14 - INFO - codeparrot_training - Step 3597: {'lr': 0.0004986356000782345, 'samples': 690816, 'steps': 3597, 'loss/train': 2.3925302624702454} 01/28/2022 17:30:20 - INFO - codeparrot_training - Step 3598: {'lr': 0.0004986338923979119, 'samples': 691008, 'steps': 3598, 'loss/train': 2.1305049061775208} 01/28/2022 17:30:24 - INFO - codeparrot_training - Step 3599: {'lr': 0.0004986321836525209, 'samples': 691200, 'steps': 3599, 'loss/train': 1.749606192111969} 01/28/2022 17:30:28 - INFO - codeparrot_training - Step 3600: {'lr': 0.0004986304738420684, 'samples': 691392, 'steps': 3600, 'loss/train': 2.6453123688697815} 01/28/2022 17:30:32 - INFO - codeparrot_training - Step 3601: {'lr': 0.0004986287629665619, 'samples': 691584, 'steps': 3601, 'loss/train': 1.5482770800590515} 01/28/2022 17:30:38 - INFO - codeparrot_training - Step 3602: {'lr': 0.0004986270510260087, 'samples': 691776, 'steps': 3602, 'loss/train': 0.7207265943288803} 01/28/2022 17:30:43 - INFO - codeparrot_training - Step 3603: {'lr': 0.0004986253380204163, 'samples': 691968, 'steps': 3603, 'loss/train': 1.9804787635803223} 01/28/2022 17:30:47 - INFO - codeparrot_training - Step 3604: {'lr': 0.0004986236239497918, 'samples': 692160, 'steps': 3604, 'loss/train': 1.2621471583843231} 01/28/2022 17:30:51 - INFO - codeparrot_training - Step 3605: {'lr': 0.0004986219088141426, 'samples': 692352, 'steps': 3605, 'loss/train': 1.918094515800476} 01/28/2022 17:30:55 - INFO - codeparrot_training - Step 3606: {'lr': 0.0004986201926134761, 'samples': 692544, 'steps': 3606, 'loss/train': 2.0796931385993958} 01/28/2022 17:30:59 - INFO - codeparrot_training - Step 3607: {'lr': 0.0004986184753477998, 'samples': 692736, 'steps': 3607, 'loss/train': 1.4132406413555145} 01/28/2022 17:31:03 - INFO - codeparrot_training - Step 3608: {'lr': 0.0004986167570171208, 'samples': 692928, 'steps': 3608, 'loss/train': 2.075551450252533} 01/28/2022 17:31:09 - INFO - codeparrot_training - Step 3609: {'lr': 0.0004986150376214465, 'samples': 693120, 'steps': 3609, 'loss/train': 2.024173378944397} 01/28/2022 17:31:13 - INFO - codeparrot_training - Step 3610: {'lr': 0.0004986133171607844, 'samples': 693312, 'steps': 3610, 'loss/train': 2.5197002291679382} 01/28/2022 17:31:17 - INFO - codeparrot_training - Step 3611: {'lr': 0.0004986115956351417, 'samples': 693504, 'steps': 3611, 'loss/train': 1.674448549747467} 01/28/2022 17:31:21 - INFO - codeparrot_training - Step 3612: {'lr': 0.000498609873044526, 'samples': 693696, 'steps': 3612, 'loss/train': 2.1849648356437683} 01/28/2022 17:31:27 - INFO - codeparrot_training - Step 3613: {'lr': 0.0004986081493889444, 'samples': 693888, 'steps': 3613, 'loss/train': 2.2315324544906616} 01/28/2022 17:31:31 - INFO - codeparrot_training - Step 3614: {'lr': 0.0004986064246684046, 'samples': 694080, 'steps': 3614, 'loss/train': 1.423894464969635} 01/28/2022 17:31:35 - INFO - codeparrot_training - Step 3615: {'lr': 0.0004986046988829136, 'samples': 694272, 'steps': 3615, 'loss/train': 2.392770767211914} 01/28/2022 17:31:40 - INFO - codeparrot_training - Step 3616: {'lr': 0.0004986029720324791, 'samples': 694464, 'steps': 3616, 'loss/train': 1.9225929379463196} 01/28/2022 17:31:44 - INFO - codeparrot_training - Step 3617: {'lr': 0.0004986012441171085, 'samples': 694656, 'steps': 3617, 'loss/train': 2.8271169662475586} 01/28/2022 17:31:49 - INFO - codeparrot_training - Step 3618: {'lr': 0.000498599515136809, 'samples': 694848, 'steps': 3618, 'loss/train': 1.35988050699234} 01/28/2022 17:31:53 - INFO - codeparrot_training - Step 3619: {'lr': 0.0004985977850915882, 'samples': 695040, 'steps': 3619, 'loss/train': 1.7911748886108398} 01/28/2022 17:31:57 - INFO - codeparrot_training - Step 3620: {'lr': 0.0004985960539814534, 'samples': 695232, 'steps': 3620, 'loss/train': 1.902105689048767} 01/28/2022 17:32:01 - INFO - codeparrot_training - Step 3621: {'lr': 0.000498594321806412, 'samples': 695424, 'steps': 3621, 'loss/train': 2.370338201522827} 01/28/2022 17:32:06 - INFO - codeparrot_training - Step 3622: {'lr': 0.0004985925885664716, 'samples': 695616, 'steps': 3622, 'loss/train': 2.1530306339263916} 01/28/2022 17:32:11 - INFO - codeparrot_training - Step 3623: {'lr': 0.0004985908542616393, 'samples': 695808, 'steps': 3623, 'loss/train': 2.9010974764823914} 01/28/2022 17:32:16 - INFO - codeparrot_training - Step 3624: {'lr': 0.0004985891188919229, 'samples': 696000, 'steps': 3624, 'loss/train': 1.7743636965751648} 01/28/2022 17:32:20 - INFO - codeparrot_training - Step 3625: {'lr': 0.0004985873824573296, 'samples': 696192, 'steps': 3625, 'loss/train': 1.656348466873169} 01/28/2022 17:32:24 - INFO - codeparrot_training - Step 3626: {'lr': 0.0004985856449578667, 'samples': 696384, 'steps': 3626, 'loss/train': 2.268064498901367} 01/28/2022 17:32:28 - INFO - codeparrot_training - Step 3627: {'lr': 0.0004985839063935421, 'samples': 696576, 'steps': 3627, 'loss/train': 2.291037619113922} 01/28/2022 17:32:33 - INFO - codeparrot_training - Step 3628: {'lr': 0.0004985821667643628, 'samples': 696768, 'steps': 3628, 'loss/train': 1.5988454818725586} 01/28/2022 17:32:38 - INFO - codeparrot_training - Step 3629: {'lr': 0.0004985804260703364, 'samples': 696960, 'steps': 3629, 'loss/train': 2.8681452870368958} 01/28/2022 17:32:42 - INFO - codeparrot_training - Step 3630: {'lr': 0.0004985786843114706, 'samples': 697152, 'steps': 3630, 'loss/train': 2.182974636554718} 01/28/2022 17:32:46 - INFO - codeparrot_training - Step 3631: {'lr': 0.0004985769414877725, 'samples': 697344, 'steps': 3631, 'loss/train': 2.804019033908844} 01/28/2022 17:32:50 - INFO - codeparrot_training - Step 3632: {'lr': 0.0004985751975992497, 'samples': 697536, 'steps': 3632, 'loss/train': 2.829109311103821} 01/28/2022 17:32:55 - INFO - codeparrot_training - Step 3633: {'lr': 0.0004985734526459098, 'samples': 697728, 'steps': 3633, 'loss/train': 2.9049071073532104} 01/28/2022 17:32:59 - INFO - codeparrot_training - Step 3634: {'lr': 0.0004985717066277601, 'samples': 697920, 'steps': 3634, 'loss/train': 1.2158917486667633} 01/28/2022 17:33:04 - INFO - codeparrot_training - Step 3635: {'lr': 0.0004985699595448081, 'samples': 698112, 'steps': 3635, 'loss/train': 2.144705593585968} 01/28/2022 17:33:08 - INFO - codeparrot_training - Step 3636: {'lr': 0.0004985682113970613, 'samples': 698304, 'steps': 3636, 'loss/train': 1.8022863864898682} 01/28/2022 17:33:12 - INFO - codeparrot_training - Step 3637: {'lr': 0.0004985664621845273, 'samples': 698496, 'steps': 3637, 'loss/train': 2.1817420721054077} 01/28/2022 17:33:18 - INFO - codeparrot_training - Step 3638: {'lr': 0.0004985647119072135, 'samples': 698688, 'steps': 3638, 'loss/train': 1.9371340870857239} 01/28/2022 17:33:22 - INFO - codeparrot_training - Step 3639: {'lr': 0.0004985629605651273, 'samples': 698880, 'steps': 3639, 'loss/train': 2.223214566707611} 01/28/2022 17:33:27 - INFO - codeparrot_training - Step 3640: {'lr': 0.0004985612081582763, 'samples': 699072, 'steps': 3640, 'loss/train': 2.132153570652008} 01/28/2022 17:33:31 - INFO - codeparrot_training - Step 3641: {'lr': 0.0004985594546866682, 'samples': 699264, 'steps': 3641, 'loss/train': 1.7434671521186829} 01/28/2022 17:33:35 - INFO - codeparrot_training - Step 3642: {'lr': 0.0004985577001503102, 'samples': 699456, 'steps': 3642, 'loss/train': 2.2649402618408203} 01/28/2022 17:33:39 - INFO - codeparrot_training - Step 3643: {'lr': 0.0004985559445492099, 'samples': 699648, 'steps': 3643, 'loss/train': 2.3690911531448364} 01/28/2022 17:33:45 - INFO - codeparrot_training - Step 3644: {'lr': 0.0004985541878833749, 'samples': 699840, 'steps': 3644, 'loss/train': 3.465091824531555} 01/28/2022 17:33:49 - INFO - codeparrot_training - Step 3645: {'lr': 0.0004985524301528127, 'samples': 700032, 'steps': 3645, 'loss/train': 3.3333009481430054} 01/28/2022 17:33:53 - INFO - codeparrot_training - Step 3646: {'lr': 0.0004985506713575307, 'samples': 700224, 'steps': 3646, 'loss/train': 1.8782431483268738} 01/28/2022 17:33:57 - INFO - codeparrot_training - Step 3647: {'lr': 0.0004985489114975368, 'samples': 700416, 'steps': 3647, 'loss/train': 2.7525025606155396} 01/28/2022 17:34:03 - INFO - codeparrot_training - Step 3648: {'lr': 0.0004985471505728381, 'samples': 700608, 'steps': 3648, 'loss/train': 2.1115052103996277} 01/28/2022 17:34:07 - INFO - codeparrot_training - Step 3649: {'lr': 0.0004985453885834423, 'samples': 700800, 'steps': 3649, 'loss/train': 2.203114628791809} 01/28/2022 17:34:12 - INFO - codeparrot_training - Step 3650: {'lr': 0.0004985436255293571, 'samples': 700992, 'steps': 3650, 'loss/train': 1.362264633178711} 01/28/2022 17:34:16 - INFO - codeparrot_training - Step 3651: {'lr': 0.0004985418614105898, 'samples': 701184, 'steps': 3651, 'loss/train': 1.7582332491874695} 01/28/2022 17:34:20 - INFO - codeparrot_training - Step 3652: {'lr': 0.0004985400962271482, 'samples': 701376, 'steps': 3652, 'loss/train': 1.203687161207199} 01/28/2022 17:34:25 - INFO - codeparrot_training - Step 3653: {'lr': 0.0004985383299790397, 'samples': 701568, 'steps': 3653, 'loss/train': 1.6173604130744934} 01/28/2022 17:34:29 - INFO - codeparrot_training - Step 3654: {'lr': 0.0004985365626662719, 'samples': 701760, 'steps': 3654, 'loss/train': 1.7053468823432922} 01/28/2022 17:34:33 - INFO - codeparrot_training - Step 3655: {'lr': 0.0004985347942888524, 'samples': 701952, 'steps': 3655, 'loss/train': 1.8465458750724792} 01/28/2022 17:34:37 - INFO - codeparrot_training - Step 3656: {'lr': 0.0004985330248467888, 'samples': 702144, 'steps': 3656, 'loss/train': 3.0318092107772827} 01/28/2022 17:34:42 - INFO - codeparrot_training - Step 3657: {'lr': 0.0004985312543400886, 'samples': 702336, 'steps': 3657, 'loss/train': 2.545952260494232} 01/28/2022 17:34:47 - INFO - codeparrot_training - Step 3658: {'lr': 0.0004985294827687594, 'samples': 702528, 'steps': 3658, 'loss/train': 2.0528027415275574} 01/28/2022 17:34:51 - INFO - codeparrot_training - Step 3659: {'lr': 0.0004985277101328088, 'samples': 702720, 'steps': 3659, 'loss/train': 2.1628611087799072} 01/28/2022 17:34:55 - INFO - codeparrot_training - Step 3660: {'lr': 0.0004985259364322445, 'samples': 702912, 'steps': 3660, 'loss/train': 1.9102531671524048} 01/28/2022 17:34:59 - INFO - codeparrot_training - Step 3661: {'lr': 0.0004985241616670739, 'samples': 703104, 'steps': 3661, 'loss/train': 2.2545855045318604} 01/28/2022 17:35:03 - INFO - codeparrot_training - Step 3662: {'lr': 0.0004985223858373048, 'samples': 703296, 'steps': 3662, 'loss/train': 2.001529812812805} 01/28/2022 17:35:09 - INFO - codeparrot_training - Step 3663: {'lr': 0.0004985206089429447, 'samples': 703488, 'steps': 3663, 'loss/train': 1.3329559564590454} 01/28/2022 17:35:14 - INFO - codeparrot_training - Step 3664: {'lr': 0.0004985188309840012, 'samples': 703680, 'steps': 3664, 'loss/train': 2.1713746190071106} 01/28/2022 17:35:18 - INFO - codeparrot_training - Step 3665: {'lr': 0.0004985170519604819, 'samples': 703872, 'steps': 3665, 'loss/train': 2.6433370113372803} 01/28/2022 17:35:22 - INFO - codeparrot_training - Step 3666: {'lr': 0.0004985152718723944, 'samples': 704064, 'steps': 3666, 'loss/train': 1.2106235325336456} 01/28/2022 17:35:26 - INFO - codeparrot_training - Step 3667: {'lr': 0.0004985134907197466, 'samples': 704256, 'steps': 3667, 'loss/train': 1.7211533188819885} 01/28/2022 17:35:31 - INFO - codeparrot_training - Step 3668: {'lr': 0.0004985117085025458, 'samples': 704448, 'steps': 3668, 'loss/train': 2.303402066230774} 01/28/2022 17:35:36 - INFO - codeparrot_training - Step 3669: {'lr': 0.0004985099252207998, 'samples': 704640, 'steps': 3669, 'loss/train': 1.5848402380943298} 01/28/2022 17:35:40 - INFO - codeparrot_training - Step 3670: {'lr': 0.0004985081408745161, 'samples': 704832, 'steps': 3670, 'loss/train': 2.2251495122909546} 01/28/2022 17:35:44 - INFO - codeparrot_training - Step 3671: {'lr': 0.0004985063554637025, 'samples': 705024, 'steps': 3671, 'loss/train': 1.795073926448822} 01/28/2022 17:35:48 - INFO - codeparrot_training - Step 3672: {'lr': 0.0004985045689883665, 'samples': 705216, 'steps': 3672, 'loss/train': 2.1556647419929504} 01/28/2022 17:35:53 - INFO - codeparrot_training - Step 3673: {'lr': 0.0004985027814485159, 'samples': 705408, 'steps': 3673, 'loss/train': 1.8931989669799805} 01/28/2022 17:35:57 - INFO - codeparrot_training - Step 3674: {'lr': 0.0004985009928441584, 'samples': 705600, 'steps': 3674, 'loss/train': 2.116041898727417} 01/28/2022 17:36:02 - INFO - codeparrot_training - Step 3675: {'lr': 0.0004984992031753014, 'samples': 705792, 'steps': 3675, 'loss/train': 1.7962530255317688} 01/28/2022 17:36:06 - INFO - codeparrot_training - Step 3676: {'lr': 0.0004984974124419528, 'samples': 705984, 'steps': 3676, 'loss/train': 0.8067978322505951} 01/28/2022 17:36:10 - INFO - codeparrot_training - Step 3677: {'lr': 0.0004984956206441201, 'samples': 706176, 'steps': 3677, 'loss/train': 1.812877893447876} 01/28/2022 17:36:15 - INFO - codeparrot_training - Step 3678: {'lr': 0.0004984938277818112, 'samples': 706368, 'steps': 3678, 'loss/train': 1.1391072571277618} 01/28/2022 17:36:19 - INFO - codeparrot_training - Step 3679: {'lr': 0.0004984920338550335, 'samples': 706560, 'steps': 3679, 'loss/train': 2.2032442688941956} 01/28/2022 17:36:24 - INFO - codeparrot_training - Step 3680: {'lr': 0.0004984902388637949, 'samples': 706752, 'steps': 3680, 'loss/train': 1.921127736568451} 01/28/2022 17:36:28 - INFO - codeparrot_training - Step 3681: {'lr': 0.0004984884428081031, 'samples': 706944, 'steps': 3681, 'loss/train': 1.4936216175556183} 01/28/2022 17:36:32 - INFO - codeparrot_training - Step 3682: {'lr': 0.0004984866456879657, 'samples': 707136, 'steps': 3682, 'loss/train': 2.242888927459717} 01/28/2022 17:36:38 - INFO - codeparrot_training - Step 3683: {'lr': 0.0004984848475033903, 'samples': 707328, 'steps': 3683, 'loss/train': 0.6298561692237854} 01/28/2022 17:36:42 - INFO - codeparrot_training - Step 3684: {'lr': 0.0004984830482543847, 'samples': 707520, 'steps': 3684, 'loss/train': 1.8779847621917725} 01/28/2022 17:36:46 - INFO - codeparrot_training - Step 3685: {'lr': 0.0004984812479409568, 'samples': 707712, 'steps': 3685, 'loss/train': 1.2749084830284119} 01/28/2022 17:36:50 - INFO - codeparrot_training - Step 3686: {'lr': 0.000498479446563114, 'samples': 707904, 'steps': 3686, 'loss/train': 2.332263171672821} 01/28/2022 17:36:55 - INFO - codeparrot_training - Step 3687: {'lr': 0.0004984776441208642, 'samples': 708096, 'steps': 3687, 'loss/train': 2.080693781375885} 01/28/2022 17:37:00 - INFO - codeparrot_training - Step 3688: {'lr': 0.000498475840614215, 'samples': 708288, 'steps': 3688, 'loss/train': 2.2358733415603638} 01/28/2022 17:37:04 - INFO - codeparrot_training - Step 3689: {'lr': 0.0004984740360431742, 'samples': 708480, 'steps': 3689, 'loss/train': 2.110926568508148} 01/28/2022 17:37:08 - INFO - codeparrot_training - Step 3690: {'lr': 0.0004984722304077496, 'samples': 708672, 'steps': 3690, 'loss/train': 0.6674544811248779} 01/28/2022 17:37:12 - INFO - codeparrot_training - Step 3691: {'lr': 0.0004984704237079489, 'samples': 708864, 'steps': 3691, 'loss/train': 2.5373185873031616} 01/28/2022 17:37:16 - INFO - codeparrot_training - Step 3692: {'lr': 0.0004984686159437798, 'samples': 709056, 'steps': 3692, 'loss/train': 1.8438209295272827} 01/28/2022 17:37:22 - INFO - codeparrot_training - Step 3693: {'lr': 0.00049846680711525, 'samples': 709248, 'steps': 3693, 'loss/train': 1.4482027888298035} 01/28/2022 17:37:26 - INFO - codeparrot_training - Step 3694: {'lr': 0.0004984649972223673, 'samples': 709440, 'steps': 3694, 'loss/train': 1.8845000267028809} 01/28/2022 17:37:30 - INFO - codeparrot_training - Step 3695: {'lr': 0.0004984631862651395, 'samples': 709632, 'steps': 3695, 'loss/train': 1.5666573643684387} 01/28/2022 17:37:34 - INFO - codeparrot_training - Step 3696: {'lr': 0.0004984613742435742, 'samples': 709824, 'steps': 3696, 'loss/train': 1.4273261725902557} 01/28/2022 17:37:38 - INFO - codeparrot_training - Step 3697: {'lr': 0.0004984595611576793, 'samples': 710016, 'steps': 3697, 'loss/train': 2.3325782418251038} 01/28/2022 17:37:45 - INFO - codeparrot_training - Step 3698: {'lr': 0.0004984577470074625, 'samples': 710208, 'steps': 3698, 'loss/train': 0.2415599673986435} 01/28/2022 17:37:49 - INFO - codeparrot_training - Step 3699: {'lr': 0.0004984559317929317, 'samples': 710400, 'steps': 3699, 'loss/train': 1.9260213375091553} 01/28/2022 17:37:53 - INFO - codeparrot_training - Step 3700: {'lr': 0.0004984541155140946, 'samples': 710592, 'steps': 3700, 'loss/train': 2.238744378089905} 01/28/2022 17:37:57 - INFO - codeparrot_training - Step 3701: {'lr': 0.0004984522981709589, 'samples': 710784, 'steps': 3701, 'loss/train': 2.148831367492676} 01/28/2022 17:38:01 - INFO - codeparrot_training - Step 3702: {'lr': 0.0004984504797635324, 'samples': 710976, 'steps': 3702, 'loss/train': 2.2069287300109863} 01/28/2022 17:38:07 - INFO - codeparrot_training - Step 3703: {'lr': 0.000498448660291823, 'samples': 711168, 'steps': 3703, 'loss/train': 1.6544435620307922} 01/28/2022 17:38:11 - INFO - codeparrot_training - Step 3704: {'lr': 0.0004984468397558384, 'samples': 711360, 'steps': 3704, 'loss/train': 1.9243800044059753} 01/28/2022 17:38:15 - INFO - codeparrot_training - Step 3705: {'lr': 0.0004984450181555864, 'samples': 711552, 'steps': 3705, 'loss/train': 1.5731942653656006} 01/28/2022 17:38:19 - INFO - codeparrot_training - Step 3706: {'lr': 0.0004984431954910749, 'samples': 711744, 'steps': 3706, 'loss/train': 1.9941006302833557} 01/28/2022 17:38:24 - INFO - codeparrot_training - Step 3707: {'lr': 0.0004984413717623117, 'samples': 711936, 'steps': 3707, 'loss/train': 1.5425974130630493} 01/28/2022 17:38:29 - INFO - codeparrot_training - Step 3708: {'lr': 0.0004984395469693044, 'samples': 712128, 'steps': 3708, 'loss/train': 2.1088669896125793} 01/28/2022 17:38:34 - INFO - codeparrot_training - Step 3709: {'lr': 0.000498437721112061, 'samples': 712320, 'steps': 3709, 'loss/train': 2.117095470428467} 01/28/2022 17:38:38 - INFO - codeparrot_training - Step 3710: {'lr': 0.0004984358941905894, 'samples': 712512, 'steps': 3710, 'loss/train': 2.244738221168518} 01/28/2022 17:38:42 - INFO - codeparrot_training - Step 3711: {'lr': 0.0004984340662048972, 'samples': 712704, 'steps': 3711, 'loss/train': 1.645984411239624} 01/28/2022 17:38:46 - INFO - codeparrot_training - Step 3712: {'lr': 0.0004984322371549924, 'samples': 712896, 'steps': 3712, 'loss/train': 2.058500826358795} 01/28/2022 17:38:51 - INFO - codeparrot_training - Step 3713: {'lr': 0.0004984304070408828, 'samples': 713088, 'steps': 3713, 'loss/train': 1.9290311336517334} 01/28/2022 17:38:55 - INFO - codeparrot_training - Step 3714: {'lr': 0.0004984285758625761, 'samples': 713280, 'steps': 3714, 'loss/train': 2.246881663799286} 01/28/2022 17:39:00 - INFO - codeparrot_training - Step 3715: {'lr': 0.0004984267436200805, 'samples': 713472, 'steps': 3715, 'loss/train': 1.7560427784919739} 01/28/2022 17:39:04 - INFO - codeparrot_training - Step 3716: {'lr': 0.0004984249103134035, 'samples': 713664, 'steps': 3716, 'loss/train': 2.3098100423812866} 01/28/2022 17:39:08 - INFO - codeparrot_training - Step 3717: {'lr': 0.000498423075942553, 'samples': 713856, 'steps': 3717, 'loss/train': 0.7465713769197464} 01/28/2022 17:39:13 - INFO - codeparrot_training - Step 3718: {'lr': 0.0004984212405075369, 'samples': 714048, 'steps': 3718, 'loss/train': 1.0667465329170227} 01/28/2022 17:39:17 - INFO - codeparrot_training - Step 3719: {'lr': 0.0004984194040083632, 'samples': 714240, 'steps': 3719, 'loss/train': 2.1922295093536377} 01/28/2022 17:39:21 - INFO - codeparrot_training - Step 3720: {'lr': 0.0004984175664450397, 'samples': 714432, 'steps': 3720, 'loss/train': 3.264833092689514} 01/28/2022 17:39:25 - INFO - codeparrot_training - Step 3721: {'lr': 0.0004984157278175741, 'samples': 714624, 'steps': 3721, 'loss/train': 1.9622856974601746} 01/28/2022 17:39:30 - INFO - codeparrot_training - Step 3722: {'lr': 0.0004984138881259744, 'samples': 714816, 'steps': 3722, 'loss/train': 0.7516137063503265} 01/28/2022 17:39:36 - INFO - codeparrot_training - Step 3723: {'lr': 0.0004984120473702486, 'samples': 715008, 'steps': 3723, 'loss/train': 2.036876857280731} 01/28/2022 17:39:40 - INFO - codeparrot_training - Step 3724: {'lr': 0.0004984102055504044, 'samples': 715200, 'steps': 3724, 'loss/train': 1.9053964018821716} 01/28/2022 17:39:44 - INFO - codeparrot_training - Step 3725: {'lr': 0.0004984083626664497, 'samples': 715392, 'steps': 3725, 'loss/train': 2.1257981657981873} 01/28/2022 17:39:48 - INFO - codeparrot_training - Step 3726: {'lr': 0.0004984065187183925, 'samples': 715584, 'steps': 3726, 'loss/train': 1.4331915378570557} 01/28/2022 17:39:52 - INFO - codeparrot_training - Step 3727: {'lr': 0.0004984046737062407, 'samples': 715776, 'steps': 3727, 'loss/train': 1.6146145462989807} 01/28/2022 17:39:58 - INFO - codeparrot_training - Step 3728: {'lr': 0.0004984028276300021, 'samples': 715968, 'steps': 3728, 'loss/train': 2.068008005619049} 01/28/2022 17:40:02 - INFO - codeparrot_training - Step 3729: {'lr': 0.0004984009804896846, 'samples': 716160, 'steps': 3729, 'loss/train': 1.7406451106071472} 01/28/2022 17:40:06 - INFO - codeparrot_training - Step 3730: {'lr': 0.0004983991322852963, 'samples': 716352, 'steps': 3730, 'loss/train': 2.149251937866211} 01/28/2022 17:40:10 - INFO - codeparrot_training - Step 3731: {'lr': 0.000498397283016845, 'samples': 716544, 'steps': 3731, 'loss/train': 1.5975408554077148} 01/28/2022 17:40:15 - INFO - codeparrot_training - Step 3732: {'lr': 0.0004983954326843386, 'samples': 716736, 'steps': 3732, 'loss/train': 2.0848812460899353} 01/28/2022 17:40:20 - INFO - codeparrot_training - Step 3733: {'lr': 0.000498393581287785, 'samples': 716928, 'steps': 3733, 'loss/train': 2.6619858741760254} 01/28/2022 17:40:24 - INFO - codeparrot_training - Step 3734: {'lr': 0.0004983917288271921, 'samples': 717120, 'steps': 3734, 'loss/train': 0.46627990901470184} 01/28/2022 17:40:28 - INFO - codeparrot_training - Step 3735: {'lr': 0.0004983898753025681, 'samples': 717312, 'steps': 3735, 'loss/train': 1.5366965532302856} 01/28/2022 17:40:32 - INFO - codeparrot_training - Step 3736: {'lr': 0.0004983880207139205, 'samples': 717504, 'steps': 3736, 'loss/train': 0.7628850638866425} 01/28/2022 17:40:36 - INFO - codeparrot_training - Step 3737: {'lr': 0.0004983861650612577, 'samples': 717696, 'steps': 3737, 'loss/train': 1.520175039768219} 01/28/2022 17:40:42 - INFO - codeparrot_training - Step 3738: {'lr': 0.0004983843083445873, 'samples': 717888, 'steps': 3738, 'loss/train': 2.332996129989624} 01/28/2022 17:40:47 - INFO - codeparrot_training - Step 3739: {'lr': 0.0004983824505639175, 'samples': 718080, 'steps': 3739, 'loss/train': 1.2384587824344635} 01/28/2022 17:40:51 - INFO - codeparrot_training - Step 3740: {'lr': 0.000498380591719256, 'samples': 718272, 'steps': 3740, 'loss/train': 1.7678001523017883} 01/28/2022 17:40:55 - INFO - codeparrot_training - Step 3741: {'lr': 0.0004983787318106111, 'samples': 718464, 'steps': 3741, 'loss/train': 1.8438055515289307} 01/28/2022 17:40:59 - INFO - codeparrot_training - Step 3742: {'lr': 0.0004983768708379905, 'samples': 718656, 'steps': 3742, 'loss/train': 1.7904914617538452} 01/28/2022 17:41:05 - INFO - codeparrot_training - Step 3743: {'lr': 0.0004983750088014023, 'samples': 718848, 'steps': 3743, 'loss/train': 3.177743911743164} 01/28/2022 17:41:09 - INFO - codeparrot_training - Step 3744: {'lr': 0.0004983731457008544, 'samples': 719040, 'steps': 3744, 'loss/train': 3.0906976461410522} 01/28/2022 17:41:13 - INFO - codeparrot_training - Step 3745: {'lr': 0.0004983712815363548, 'samples': 719232, 'steps': 3745, 'loss/train': 2.898280441761017} 01/28/2022 17:41:17 - INFO - codeparrot_training - Step 3746: {'lr': 0.0004983694163079115, 'samples': 719424, 'steps': 3746, 'loss/train': 1.8809208869934082} 01/28/2022 17:41:21 - INFO - codeparrot_training - Step 3747: {'lr': 0.0004983675500155325, 'samples': 719616, 'steps': 3747, 'loss/train': 1.413057804107666} 01/28/2022 17:41:27 - INFO - codeparrot_training - Step 3748: {'lr': 0.0004983656826592258, 'samples': 719808, 'steps': 3748, 'loss/train': 1.4952103793621063} 01/28/2022 17:41:31 - INFO - codeparrot_training - Step 3749: {'lr': 0.0004983638142389993, 'samples': 720000, 'steps': 3749, 'loss/train': 2.008021831512451} 01/28/2022 17:41:35 - INFO - codeparrot_training - Step 3750: {'lr': 0.000498361944754861, 'samples': 720192, 'steps': 3750, 'loss/train': 1.3358238637447357} 01/28/2022 17:41:39 - INFO - codeparrot_training - Step 3751: {'lr': 0.0004983600742068192, 'samples': 720384, 'steps': 3751, 'loss/train': 2.369372248649597} 01/28/2022 17:41:44 - INFO - codeparrot_training - Step 3752: {'lr': 0.0004983582025948816, 'samples': 720576, 'steps': 3752, 'loss/train': 1.720060408115387} 01/28/2022 17:41:50 - INFO - codeparrot_training - Step 3753: {'lr': 0.0004983563299190564, 'samples': 720768, 'steps': 3753, 'loss/train': 1.32444566488266} 01/28/2022 17:41:54 - INFO - codeparrot_training - Step 3754: {'lr': 0.0004983544561793515, 'samples': 720960, 'steps': 3754, 'loss/train': 2.0864685773849487} 01/28/2022 17:41:59 - INFO - codeparrot_training - Step 3755: {'lr': 0.000498352581375775, 'samples': 721152, 'steps': 3755, 'loss/train': 1.2873390018939972} 01/28/2022 17:42:03 - INFO - codeparrot_training - Step 3756: {'lr': 0.0004983507055083349, 'samples': 721344, 'steps': 3756, 'loss/train': 3.7877966165542603} 01/28/2022 17:42:07 - INFO - codeparrot_training - Step 3757: {'lr': 0.0004983488285770391, 'samples': 721536, 'steps': 3757, 'loss/train': 0.47211652994155884} 01/28/2022 17:42:12 - INFO - codeparrot_training - Step 3758: {'lr': 0.000498346950581896, 'samples': 721728, 'steps': 3758, 'loss/train': 1.9862726926803589} 01/28/2022 17:42:16 - INFO - codeparrot_training - Step 3759: {'lr': 0.0004983450715229132, 'samples': 721920, 'steps': 3759, 'loss/train': 1.5227411985397339} 01/28/2022 17:42:21 - INFO - codeparrot_training - Step 3760: {'lr': 0.000498343191400099, 'samples': 722112, 'steps': 3760, 'loss/train': 2.16923189163208} 01/28/2022 17:42:25 - INFO - codeparrot_training - Step 3761: {'lr': 0.0004983413102134616, 'samples': 722304, 'steps': 3761, 'loss/train': 1.6531500220298767} 01/28/2022 17:42:29 - INFO - codeparrot_training - Step 3762: {'lr': 0.0004983394279630088, 'samples': 722496, 'steps': 3762, 'loss/train': 2.340050160884857} 01/28/2022 17:42:34 - INFO - codeparrot_training - Step 3763: {'lr': 0.0004983375446487488, 'samples': 722688, 'steps': 3763, 'loss/train': 1.8294095993041992} 01/28/2022 17:42:38 - INFO - codeparrot_training - Step 3764: {'lr': 0.0004983356602706895, 'samples': 722880, 'steps': 3764, 'loss/train': 2.265474021434784} 01/28/2022 17:42:42 - INFO - codeparrot_training - Step 3765: {'lr': 0.0004983337748288391, 'samples': 723072, 'steps': 3765, 'loss/train': 0.688498467206955} 01/28/2022 17:42:47 - INFO - codeparrot_training - Step 3766: {'lr': 0.0004983318883232058, 'samples': 723264, 'steps': 3766, 'loss/train': 1.4047125577926636} 01/28/2022 17:42:51 - INFO - codeparrot_training - Step 3767: {'lr': 0.0004983300007537974, 'samples': 723456, 'steps': 3767, 'loss/train': 1.198085904121399} 01/28/2022 17:42:57 - INFO - codeparrot_training - Step 3768: {'lr': 0.0004983281121206222, 'samples': 723648, 'steps': 3768, 'loss/train': 2.0418235659599304} 01/28/2022 17:43:01 - INFO - codeparrot_training - Step 3769: {'lr': 0.0004983262224236882, 'samples': 723840, 'steps': 3769, 'loss/train': 2.5745871663093567} 01/28/2022 17:43:05 - INFO - codeparrot_training - Step 3770: {'lr': 0.0004983243316630035, 'samples': 724032, 'steps': 3770, 'loss/train': 2.16656094789505} 01/28/2022 17:43:10 - INFO - codeparrot_training - Step 3771: {'lr': 0.0004983224398385762, 'samples': 724224, 'steps': 3771, 'loss/train': 2.440315067768097} 01/28/2022 17:43:14 - INFO - codeparrot_training - Step 3772: {'lr': 0.0004983205469504144, 'samples': 724416, 'steps': 3772, 'loss/train': 1.839138686656952} 01/28/2022 17:43:18 - INFO - codeparrot_training - Step 3773: {'lr': 0.0004983186529985263, 'samples': 724608, 'steps': 3773, 'loss/train': 2.2199671268463135} 01/28/2022 17:43:23 - INFO - codeparrot_training - Step 3774: {'lr': 0.00049831675798292, 'samples': 724800, 'steps': 3774, 'loss/train': 2.0251622200012207} 01/28/2022 17:43:27 - INFO - codeparrot_training - Step 3775: {'lr': 0.0004983148619036034, 'samples': 724992, 'steps': 3775, 'loss/train': 1.5986772179603577} 01/28/2022 17:43:32 - INFO - codeparrot_training - Step 3776: {'lr': 0.0004983129647605849, 'samples': 725184, 'steps': 3776, 'loss/train': 1.0930625796318054} 01/28/2022 17:43:36 - INFO - codeparrot_training - Step 3777: {'lr': 0.0004983110665538724, 'samples': 725376, 'steps': 3777, 'loss/train': 2.3454240560531616} 01/28/2022 17:43:40 - INFO - codeparrot_training - Step 3778: {'lr': 0.0004983091672834742, 'samples': 725568, 'steps': 3778, 'loss/train': 3.8706629276275635} 01/28/2022 17:43:45 - INFO - codeparrot_training - Step 3779: {'lr': 0.0004983072669493985, 'samples': 725760, 'steps': 3779, 'loss/train': 1.4655917286872864} 01/28/2022 17:43:49 - INFO - codeparrot_training - Step 3780: {'lr': 0.0004983053655516531, 'samples': 725952, 'steps': 3780, 'loss/train': 2.2732200622558594} 01/28/2022 17:43:53 - INFO - codeparrot_training - Step 3781: {'lr': 0.0004983034630902465, 'samples': 726144, 'steps': 3781, 'loss/train': 1.435823768377304} 01/28/2022 17:43:58 - INFO - codeparrot_training - Step 3782: {'lr': 0.0004983015595651867, 'samples': 726336, 'steps': 3782, 'loss/train': 2.0714730620384216} 01/28/2022 17:44:02 - INFO - codeparrot_training - Step 3783: {'lr': 0.0004982996549764817, 'samples': 726528, 'steps': 3783, 'loss/train': 2.138375759124756} 01/28/2022 17:44:07 - INFO - codeparrot_training - Step 3784: {'lr': 0.0004982977493241399, 'samples': 726720, 'steps': 3784, 'loss/train': 2.1043090224266052} 01/28/2022 17:44:11 - INFO - codeparrot_training - Step 3785: {'lr': 0.0004982958426081695, 'samples': 726912, 'steps': 3785, 'loss/train': 2.5355862379074097} 01/28/2022 17:44:15 - INFO - codeparrot_training - Step 3786: {'lr': 0.0004982939348285784, 'samples': 727104, 'steps': 3786, 'loss/train': 2.027514159679413} 01/28/2022 17:44:20 - INFO - codeparrot_training - Step 3787: {'lr': 0.000498292025985375, 'samples': 727296, 'steps': 3787, 'loss/train': 1.270609974861145} 01/28/2022 17:44:26 - INFO - codeparrot_training - Step 3788: {'lr': 0.0004982901160785675, 'samples': 727488, 'steps': 3788, 'loss/train': 1.9297910928726196} 01/28/2022 17:44:30 - INFO - codeparrot_training - Step 3789: {'lr': 0.0004982882051081639, 'samples': 727680, 'steps': 3789, 'loss/train': 1.8214080333709717} 01/28/2022 17:44:34 - INFO - codeparrot_training - Step 3790: {'lr': 0.0004982862930741725, 'samples': 727872, 'steps': 3790, 'loss/train': 1.3796598315238953} 01/28/2022 17:44:38 - INFO - codeparrot_training - Step 3791: {'lr': 0.0004982843799766014, 'samples': 728064, 'steps': 3791, 'loss/train': 1.9348713755607605} 01/28/2022 17:44:42 - INFO - codeparrot_training - Step 3792: {'lr': 0.0004982824658154589, 'samples': 728256, 'steps': 3792, 'loss/train': 2.102189362049103} 01/28/2022 17:44:48 - INFO - codeparrot_training - Step 3793: {'lr': 0.000498280550590753, 'samples': 728448, 'steps': 3793, 'loss/train': 2.1231388449668884} 01/28/2022 17:44:52 - INFO - codeparrot_training - Step 3794: {'lr': 0.0004982786343024923, 'samples': 728640, 'steps': 3794, 'loss/train': 2.076042115688324} 01/28/2022 17:44:56 - INFO - codeparrot_training - Step 3795: {'lr': 0.0004982767169506847, 'samples': 728832, 'steps': 3795, 'loss/train': 1.6783141493797302} 01/28/2022 17:45:00 - INFO - codeparrot_training - Step 3796: {'lr': 0.0004982747985353384, 'samples': 729024, 'steps': 3796, 'loss/train': 1.980941355228424} 01/28/2022 17:45:04 - INFO - codeparrot_training - Step 3797: {'lr': 0.0004982728790564616, 'samples': 729216, 'steps': 3797, 'loss/train': 1.5098227858543396} 01/28/2022 17:45:10 - INFO - codeparrot_training - Step 3798: {'lr': 0.0004982709585140629, 'samples': 729408, 'steps': 3798, 'loss/train': 2.306586742401123} 01/28/2022 17:45:15 - INFO - codeparrot_training - Step 3799: {'lr': 0.0004982690369081501, 'samples': 729600, 'steps': 3799, 'loss/train': 1.4896886944770813} 01/28/2022 17:45:19 - INFO - codeparrot_training - Step 3800: {'lr': 0.0004982671142387316, 'samples': 729792, 'steps': 3800, 'loss/train': 2.3522937297821045} 01/28/2022 17:45:23 - INFO - codeparrot_training - Step 3801: {'lr': 0.0004982651905058156, 'samples': 729984, 'steps': 3801, 'loss/train': 1.7128758430480957} 01/28/2022 17:45:27 - INFO - codeparrot_training - Step 3802: {'lr': 0.0004982632657094104, 'samples': 730176, 'steps': 3802, 'loss/train': 1.8128719925880432} 01/28/2022 17:45:32 - INFO - codeparrot_training - Step 3803: {'lr': 0.0004982613398495241, 'samples': 730368, 'steps': 3803, 'loss/train': 2.2410072684288025} 01/28/2022 17:45:37 - INFO - codeparrot_training - Step 3804: {'lr': 0.0004982594129261652, 'samples': 730560, 'steps': 3804, 'loss/train': 2.0297423601150513} 01/28/2022 17:45:41 - INFO - codeparrot_training - Step 3805: {'lr': 0.0004982574849393416, 'samples': 730752, 'steps': 3805, 'loss/train': 2.297536253929138} 01/28/2022 17:45:45 - INFO - codeparrot_training - Step 3806: {'lr': 0.000498255555889062, 'samples': 730944, 'steps': 3806, 'loss/train': 2.1866177916526794} 01/28/2022 17:45:49 - INFO - codeparrot_training - Step 3807: {'lr': 0.0004982536257753343, 'samples': 731136, 'steps': 3807, 'loss/train': 1.640033483505249} 01/28/2022 17:45:54 - INFO - codeparrot_training - Step 3808: {'lr': 0.0004982516945981669, 'samples': 731328, 'steps': 3808, 'loss/train': 1.7697540521621704} 01/28/2022 17:45:58 - INFO - codeparrot_training - Step 3809: {'lr': 0.0004982497623575681, 'samples': 731520, 'steps': 3809, 'loss/train': 0.8217978179454803} 01/28/2022 17:46:03 - INFO - codeparrot_training - Step 3810: {'lr': 0.0004982478290535461, 'samples': 731712, 'steps': 3810, 'loss/train': 1.3635838329792023} 01/28/2022 17:46:07 - INFO - codeparrot_training - Step 3811: {'lr': 0.0004982458946861093, 'samples': 731904, 'steps': 3811, 'loss/train': 1.5739628076553345} 01/28/2022 17:46:11 - INFO - codeparrot_training - Step 3812: {'lr': 0.0004982439592552658, 'samples': 732096, 'steps': 3812, 'loss/train': 1.1610338687896729} 01/28/2022 17:46:17 - INFO - codeparrot_training - Step 3813: {'lr': 0.0004982420227610242, 'samples': 732288, 'steps': 3813, 'loss/train': 1.8905478715896606} 01/28/2022 17:46:21 - INFO - codeparrot_training - Step 3814: {'lr': 0.0004982400852033924, 'samples': 732480, 'steps': 3814, 'loss/train': 1.5202329754829407} 01/28/2022 17:46:25 - INFO - codeparrot_training - Step 3815: {'lr': 0.000498238146582379, 'samples': 732672, 'steps': 3815, 'loss/train': 1.132884442806244} 01/28/2022 17:46:30 - INFO - codeparrot_training - Step 3816: {'lr': 0.0004982362068979921, 'samples': 732864, 'steps': 3816, 'loss/train': 1.6015470027923584} 01/28/2022 17:46:34 - INFO - codeparrot_training - Step 3817: {'lr': 0.0004982342661502403, 'samples': 733056, 'steps': 3817, 'loss/train': 2.1905731558799744} 01/28/2022 17:46:39 - INFO - codeparrot_training - Step 3818: {'lr': 0.0004982323243391315, 'samples': 733248, 'steps': 3818, 'loss/train': 1.2202920019626617} 01/28/2022 17:46:43 - INFO - codeparrot_training - Step 3819: {'lr': 0.0004982303814646745, 'samples': 733440, 'steps': 3819, 'loss/train': 2.15052992105484} 01/28/2022 17:46:47 - INFO - codeparrot_training - Step 3820: {'lr': 0.0004982284375268772, 'samples': 733632, 'steps': 3820, 'loss/train': 3.136966109275818} 01/28/2022 17:46:52 - INFO - codeparrot_training - Step 3821: {'lr': 0.0004982264925257481, 'samples': 733824, 'steps': 3821, 'loss/train': 1.8669962882995605} 01/28/2022 17:46:56 - INFO - codeparrot_training - Step 3822: {'lr': 0.0004982245464612955, 'samples': 734016, 'steps': 3822, 'loss/train': 1.8245417475700378} 01/28/2022 17:47:02 - INFO - codeparrot_training - Step 3823: {'lr': 0.0004982225993335279, 'samples': 734208, 'steps': 3823, 'loss/train': 1.4613733291625977} 01/28/2022 17:47:06 - INFO - codeparrot_training - Step 3824: {'lr': 0.0004982206511424534, 'samples': 734400, 'steps': 3824, 'loss/train': 2.441583752632141} 01/28/2022 17:47:10 - INFO - codeparrot_training - Step 3825: {'lr': 0.0004982187018880805, 'samples': 734592, 'steps': 3825, 'loss/train': 2.471556007862091} 01/28/2022 17:47:14 - INFO - codeparrot_training - Step 3826: {'lr': 0.0004982167515704174, 'samples': 734784, 'steps': 3826, 'loss/train': 2.0200252532958984} 01/28/2022 17:47:19 - INFO - codeparrot_training - Step 3827: {'lr': 0.0004982148001894727, 'samples': 734976, 'steps': 3827, 'loss/train': 1.5454351902008057} 01/28/2022 17:47:24 - INFO - codeparrot_training - Step 3828: {'lr': 0.0004982128477452546, 'samples': 735168, 'steps': 3828, 'loss/train': 1.6735652089118958} 01/28/2022 17:47:28 - INFO - codeparrot_training - Step 3829: {'lr': 0.0004982108942377713, 'samples': 735360, 'steps': 3829, 'loss/train': 2.317869544029236} 01/28/2022 17:47:32 - INFO - codeparrot_training - Step 3830: {'lr': 0.0004982089396670316, 'samples': 735552, 'steps': 3830, 'loss/train': 1.312457263469696} 01/28/2022 17:47:36 - INFO - codeparrot_training - Step 3831: {'lr': 0.0004982069840330435, 'samples': 735744, 'steps': 3831, 'loss/train': 2.402065694332123} 01/28/2022 17:47:40 - INFO - codeparrot_training - Step 3832: {'lr': 0.0004982050273358154, 'samples': 735936, 'steps': 3832, 'loss/train': 1.1077712774276733} 01/28/2022 17:47:46 - INFO - codeparrot_training - Step 3833: {'lr': 0.0004982030695753558, 'samples': 736128, 'steps': 3833, 'loss/train': 1.9854519367218018} 01/28/2022 17:47:50 - INFO - codeparrot_training - Step 3834: {'lr': 0.0004982011107516732, 'samples': 736320, 'steps': 3834, 'loss/train': 1.5382275581359863} 01/28/2022 17:47:54 - INFO - codeparrot_training - Step 3835: {'lr': 0.0004981991508647757, 'samples': 736512, 'steps': 3835, 'loss/train': 2.0655822157859802} 01/28/2022 17:47:58 - INFO - codeparrot_training - Step 3836: {'lr': 0.0004981971899146719, 'samples': 736704, 'steps': 3836, 'loss/train': 1.8704429268836975} 01/28/2022 17:48:02 - INFO - codeparrot_training - Step 3837: {'lr': 0.0004981952279013702, 'samples': 736896, 'steps': 3837, 'loss/train': 1.7009042501449585} 01/28/2022 17:48:08 - INFO - codeparrot_training - Step 3838: {'lr': 0.0004981932648248789, 'samples': 737088, 'steps': 3838, 'loss/train': 2.071100413799286} 01/28/2022 17:48:12 - INFO - codeparrot_training - Step 3839: {'lr': 0.0004981913006852065, 'samples': 737280, 'steps': 3839, 'loss/train': 1.983356237411499} 01/28/2022 17:48:16 - INFO - codeparrot_training - Step 3840: {'lr': 0.0004981893354823614, 'samples': 737472, 'steps': 3840, 'loss/train': 2.6442684531211853} 01/28/2022 17:48:20 - INFO - codeparrot_training - Step 3841: {'lr': 0.000498187369216352, 'samples': 737664, 'steps': 3841, 'loss/train': 1.16935595870018} 01/28/2022 17:48:24 - INFO - codeparrot_training - Step 3842: {'lr': 0.0004981854018871867, 'samples': 737856, 'steps': 3842, 'loss/train': 1.8718578815460205} 01/28/2022 17:48:30 - INFO - codeparrot_training - Step 3843: {'lr': 0.0004981834334948738, 'samples': 738048, 'steps': 3843, 'loss/train': 1.6393024921417236} 01/28/2022 17:48:35 - INFO - codeparrot_training - Step 3844: {'lr': 0.0004981814640394221, 'samples': 738240, 'steps': 3844, 'loss/train': 2.4034103751182556} 01/28/2022 17:48:39 - INFO - codeparrot_training - Step 3845: {'lr': 0.0004981794935208397, 'samples': 738432, 'steps': 3845, 'loss/train': 1.6567516922950745} 01/28/2022 17:48:43 - INFO - codeparrot_training - Step 3846: {'lr': 0.0004981775219391352, 'samples': 738624, 'steps': 3846, 'loss/train': 1.479652851819992} 01/28/2022 17:48:47 - INFO - codeparrot_training - Step 3847: {'lr': 0.000498175549294317, 'samples': 738816, 'steps': 3847, 'loss/train': 0.9961211085319519} 01/28/2022 17:48:52 - INFO - codeparrot_training - Step 3848: {'lr': 0.0004981735755863934, 'samples': 739008, 'steps': 3848, 'loss/train': 2.1062888503074646} 01/28/2022 17:48:57 - INFO - codeparrot_training - Step 3849: {'lr': 0.0004981716008153732, 'samples': 739200, 'steps': 3849, 'loss/train': 2.4359137415885925} 01/28/2022 17:49:01 - INFO - codeparrot_training - Step 3850: {'lr': 0.0004981696249812646, 'samples': 739392, 'steps': 3850, 'loss/train': 2.08172607421875} 01/28/2022 17:49:05 - INFO - codeparrot_training - Step 3851: {'lr': 0.0004981676480840761, 'samples': 739584, 'steps': 3851, 'loss/train': 1.9133241176605225} 01/28/2022 17:49:09 - INFO - codeparrot_training - Step 3852: {'lr': 0.0004981656701238162, 'samples': 739776, 'steps': 3852, 'loss/train': 2.905331254005432} 01/28/2022 17:49:14 - INFO - codeparrot_training - Step 3853: {'lr': 0.0004981636911004934, 'samples': 739968, 'steps': 3853, 'loss/train': 1.5084856152534485} 01/28/2022 17:49:18 - INFO - codeparrot_training - Step 3854: {'lr': 0.0004981617110141162, 'samples': 740160, 'steps': 3854, 'loss/train': 2.3061073422431946} 01/28/2022 17:49:22 - INFO - codeparrot_training - Step 3855: {'lr': 0.000498159729864693, 'samples': 740352, 'steps': 3855, 'loss/train': 1.857278287410736} 01/28/2022 17:49:27 - INFO - codeparrot_training - Step 3856: {'lr': 0.0004981577476522323, 'samples': 740544, 'steps': 3856, 'loss/train': 2.003214955329895} 01/28/2022 17:49:31 - INFO - codeparrot_training - Step 3857: {'lr': 0.0004981557643767426, 'samples': 740736, 'steps': 3857, 'loss/train': 1.4581747949123383} 01/28/2022 17:49:37 - INFO - codeparrot_training - Step 3858: {'lr': 0.0004981537800382323, 'samples': 740928, 'steps': 3858, 'loss/train': 1.8745765686035156} 01/28/2022 17:49:41 - INFO - codeparrot_training - Step 3859: {'lr': 0.0004981517946367102, 'samples': 741120, 'steps': 3859, 'loss/train': 2.052324950695038} 01/28/2022 17:49:46 - INFO - codeparrot_training - Step 3860: {'lr': 0.0004981498081721845, 'samples': 741312, 'steps': 3860, 'loss/train': 1.547675371170044} 01/28/2022 17:49:50 - INFO - codeparrot_training - Step 3861: {'lr': 0.0004981478206446638, 'samples': 741504, 'steps': 3861, 'loss/train': 1.7392938137054443} 01/28/2022 17:49:54 - INFO - codeparrot_training - Step 3862: {'lr': 0.0004981458320541567, 'samples': 741696, 'steps': 3862, 'loss/train': 1.8118463158607483} 01/28/2022 17:49:59 - INFO - codeparrot_training - Step 3863: {'lr': 0.0004981438424006716, 'samples': 741888, 'steps': 3863, 'loss/train': 1.1096261143684387} 01/28/2022 17:50:04 - INFO - codeparrot_training - Step 3864: {'lr': 0.0004981418516842171, 'samples': 742080, 'steps': 3864, 'loss/train': 1.7204040884971619} 01/28/2022 17:50:08 - INFO - codeparrot_training - Step 3865: {'lr': 0.0004981398599048018, 'samples': 742272, 'steps': 3865, 'loss/train': 1.9135589003562927} 01/28/2022 17:50:12 - INFO - codeparrot_training - Step 3866: {'lr': 0.000498137867062434, 'samples': 742464, 'steps': 3866, 'loss/train': 1.8485246300697327} 01/28/2022 17:50:16 - INFO - codeparrot_training - Step 3867: {'lr': 0.0004981358731571223, 'samples': 742656, 'steps': 3867, 'loss/train': 1.808117151260376} 01/28/2022 17:50:20 - INFO - codeparrot_training - Step 3868: {'lr': 0.0004981338781888755, 'samples': 742848, 'steps': 3868, 'loss/train': 1.4464999437332153} 01/28/2022 17:50:27 - INFO - codeparrot_training - Step 3869: {'lr': 0.0004981318821577018, 'samples': 743040, 'steps': 3869, 'loss/train': 0.9338452219963074} 01/28/2022 17:50:31 - INFO - codeparrot_training - Step 3870: {'lr': 0.00049812988506361, 'samples': 743232, 'steps': 3870, 'loss/train': 1.8003024458885193} 01/28/2022 17:50:35 - INFO - codeparrot_training - Step 3871: {'lr': 0.0004981278869066085, 'samples': 743424, 'steps': 3871, 'loss/train': 1.165239840745926} 01/28/2022 17:50:39 - INFO - codeparrot_training - Step 3872: {'lr': 0.000498125887686706, 'samples': 743616, 'steps': 3872, 'loss/train': 3.8366771936416626} 01/28/2022 17:50:43 - INFO - codeparrot_training - Step 3873: {'lr': 0.0004981238874039109, 'samples': 743808, 'steps': 3873, 'loss/train': 2.434667229652405} 01/28/2022 17:50:49 - INFO - codeparrot_training - Step 3874: {'lr': 0.0004981218860582319, 'samples': 744000, 'steps': 3874, 'loss/train': 2.6069117188453674} 01/28/2022 17:50:53 - INFO - codeparrot_training - Step 3875: {'lr': 0.0004981198836496775, 'samples': 744192, 'steps': 3875, 'loss/train': 1.7253445386886597} 01/28/2022 17:50:57 - INFO - codeparrot_training - Step 3876: {'lr': 0.0004981178801782563, 'samples': 744384, 'steps': 3876, 'loss/train': 1.5178323984146118} 01/28/2022 17:51:01 - INFO - codeparrot_training - Step 3877: {'lr': 0.000498115875643977, 'samples': 744576, 'steps': 3877, 'loss/train': 1.093361735343933} 01/28/2022 17:51:07 - INFO - codeparrot_training - Step 3878: {'lr': 0.0004981138700468479, 'samples': 744768, 'steps': 3878, 'loss/train': 1.7087464928627014} 01/28/2022 17:51:11 - INFO - codeparrot_training - Step 3879: {'lr': 0.0004981118633868779, 'samples': 744960, 'steps': 3879, 'loss/train': 2.0750439763069153} 01/28/2022 17:51:15 - INFO - codeparrot_training - Step 3880: {'lr': 0.0004981098556640755, 'samples': 745152, 'steps': 3880, 'loss/train': 2.3083688020706177} 01/28/2022 17:51:19 - INFO - codeparrot_training - Step 3881: {'lr': 0.0004981078468784491, 'samples': 745344, 'steps': 3881, 'loss/train': 1.8179799914360046} 01/28/2022 17:51:23 - INFO - codeparrot_training - Step 3882: {'lr': 0.0004981058370300076, 'samples': 745536, 'steps': 3882, 'loss/train': 1.7902568578720093} 01/28/2022 17:51:29 - INFO - codeparrot_training - Step 3883: {'lr': 0.0004981038261187594, 'samples': 745728, 'steps': 3883, 'loss/train': 1.927725613117218} 01/28/2022 17:51:34 - INFO - codeparrot_training - Step 3884: {'lr': 0.0004981018141447133, 'samples': 745920, 'steps': 3884, 'loss/train': 1.5930064916610718} 01/28/2022 17:51:38 - INFO - codeparrot_training - Step 3885: {'lr': 0.0004980998011078776, 'samples': 746112, 'steps': 3885, 'loss/train': 2.0757336616516113} 01/28/2022 17:51:42 - INFO - codeparrot_training - Step 3886: {'lr': 0.0004980977870082613, 'samples': 746304, 'steps': 3886, 'loss/train': 1.1829858720302582} 01/28/2022 17:51:46 - INFO - codeparrot_training - Step 3887: {'lr': 0.0004980957718458729, 'samples': 746496, 'steps': 3887, 'loss/train': 1.1576427519321442} 01/28/2022 17:51:51 - INFO - codeparrot_training - Step 3888: {'lr': 0.0004980937556207207, 'samples': 746688, 'steps': 3888, 'loss/train': 2.893006145954132} 01/28/2022 17:51:55 - INFO - codeparrot_training - Step 3889: {'lr': 0.0004980917383328139, 'samples': 746880, 'steps': 3889, 'loss/train': 2.497904598712921} 01/28/2022 17:52:00 - INFO - codeparrot_training - Step 3890: {'lr': 0.0004980897199821609, 'samples': 747072, 'steps': 3890, 'loss/train': 2.4781224131584167} 01/28/2022 17:52:04 - INFO - codeparrot_training - Step 3891: {'lr': 0.0004980877005687701, 'samples': 747264, 'steps': 3891, 'loss/train': 1.972620964050293} 01/28/2022 17:52:08 - INFO - codeparrot_training - Step 3892: {'lr': 0.0004980856800926506, 'samples': 747456, 'steps': 3892, 'loss/train': 1.9653453826904297} 01/28/2022 17:52:13 - INFO - codeparrot_training - Step 3893: {'lr': 0.0004980836585538107, 'samples': 747648, 'steps': 3893, 'loss/train': 1.9667390584945679} 01/28/2022 17:52:17 - INFO - codeparrot_training - Step 3894: {'lr': 0.0004980816359522592, 'samples': 747840, 'steps': 3894, 'loss/train': 0.9755584001541138} 01/28/2022 17:52:22 - INFO - codeparrot_training - Step 3895: {'lr': 0.0004980796122880048, 'samples': 748032, 'steps': 3895, 'loss/train': 1.4374012649059296} 01/28/2022 17:52:26 - INFO - codeparrot_training - Step 3896: {'lr': 0.000498077587561056, 'samples': 748224, 'steps': 3896, 'loss/train': 1.9217174649238586} 01/28/2022 17:52:30 - INFO - codeparrot_training - Step 3897: {'lr': 0.0004980755617714216, 'samples': 748416, 'steps': 3897, 'loss/train': 1.8404147028923035} 01/28/2022 17:52:35 - INFO - codeparrot_training - Step 3898: {'lr': 0.0004980735349191104, 'samples': 748608, 'steps': 3898, 'loss/train': 2.1829633712768555} 01/28/2022 17:52:39 - INFO - codeparrot_training - Step 3899: {'lr': 0.0004980715070041308, 'samples': 748800, 'steps': 3899, 'loss/train': 1.2026150822639465} 01/28/2022 17:52:44 - INFO - codeparrot_training - Step 3900: {'lr': 0.0004980694780264917, 'samples': 748992, 'steps': 3900, 'loss/train': 1.131614774465561} 01/28/2022 17:52:48 - INFO - codeparrot_training - Step 3901: {'lr': 0.0004980674479862018, 'samples': 749184, 'steps': 3901, 'loss/train': 1.9206358194351196} 01/28/2022 17:52:52 - INFO - codeparrot_training - Step 3902: {'lr': 0.0004980654168832697, 'samples': 749376, 'steps': 3902, 'loss/train': 2.6090787649154663} 01/28/2022 17:52:58 - INFO - codeparrot_training - Step 3903: {'lr': 0.0004980633847177041, 'samples': 749568, 'steps': 3903, 'loss/train': 1.7265456318855286} 01/28/2022 17:53:02 - INFO - codeparrot_training - Step 3904: {'lr': 0.0004980613514895135, 'samples': 749760, 'steps': 3904, 'loss/train': 2.5427765250205994} 01/28/2022 17:53:07 - INFO - codeparrot_training - Step 3905: {'lr': 0.0004980593171987072, 'samples': 749952, 'steps': 3905, 'loss/train': 2.1516337394714355} 01/28/2022 17:53:11 - INFO - codeparrot_training - Step 3906: {'lr': 0.0004980572818452934, 'samples': 750144, 'steps': 3906, 'loss/train': 1.1633750796318054} 01/28/2022 17:53:15 - INFO - codeparrot_training - Step 3907: {'lr': 0.0004980552454292809, 'samples': 750336, 'steps': 3907, 'loss/train': 2.496533453464508} 01/28/2022 17:53:20 - INFO - codeparrot_training - Step 3908: {'lr': 0.0004980532079506786, 'samples': 750528, 'steps': 3908, 'loss/train': 1.7206290364265442} 01/28/2022 17:53:24 - INFO - codeparrot_training - Step 3909: {'lr': 0.0004980511694094951, 'samples': 750720, 'steps': 3909, 'loss/train': 2.4004061222076416} 01/28/2022 17:53:29 - INFO - codeparrot_training - Step 3910: {'lr': 0.0004980491298057392, 'samples': 750912, 'steps': 3910, 'loss/train': 1.9838640689849854} 01/28/2022 17:53:33 - INFO - codeparrot_training - Step 3911: {'lr': 0.0004980470891394194, 'samples': 751104, 'steps': 3911, 'loss/train': 1.938628613948822} 01/28/2022 17:53:37 - INFO - codeparrot_training - Step 3912: {'lr': 0.0004980450474105448, 'samples': 751296, 'steps': 3912, 'loss/train': 2.3048068284988403} 01/28/2022 17:53:43 - INFO - codeparrot_training - Step 3913: {'lr': 0.000498043004619124, 'samples': 751488, 'steps': 3913, 'loss/train': 2.534775674343109} 01/28/2022 17:53:48 - INFO - codeparrot_training - Step 3914: {'lr': 0.0004980409607651656, 'samples': 751680, 'steps': 3914, 'loss/train': 2.2255786657333374} 01/28/2022 17:53:52 - INFO - codeparrot_training - Step 3915: {'lr': 0.0004980389158486786, 'samples': 751872, 'steps': 3915, 'loss/train': 0.9173478484153748} 01/28/2022 17:53:56 - INFO - codeparrot_training - Step 3916: {'lr': 0.0004980368698696716, 'samples': 752064, 'steps': 3916, 'loss/train': 2.127991497516632} 01/28/2022 17:54:00 - INFO - codeparrot_training - Step 3917: {'lr': 0.0004980348228281534, 'samples': 752256, 'steps': 3917, 'loss/train': 1.9022414088249207} 01/28/2022 17:54:04 - INFO - codeparrot_training - Step 3918: {'lr': 0.0004980327747241329, 'samples': 752448, 'steps': 3918, 'loss/train': 0.26545824855566025} 01/28/2022 17:54:09 - INFO - codeparrot_training - Step 3919: {'lr': 0.0004980307255576185, 'samples': 752640, 'steps': 3919, 'loss/train': 2.111508786678314} 01/28/2022 17:54:14 - INFO - codeparrot_training - Step 3920: {'lr': 0.0004980286753286195, 'samples': 752832, 'steps': 3920, 'loss/train': 1.6098525524139404} 01/28/2022 17:54:18 - INFO - codeparrot_training - Step 3921: {'lr': 0.0004980266240371443, 'samples': 753024, 'steps': 3921, 'loss/train': 1.7781044840812683} 01/28/2022 17:54:22 - INFO - codeparrot_training - Step 3922: {'lr': 0.0004980245716832018, 'samples': 753216, 'steps': 3922, 'loss/train': 2.237361788749695} 01/28/2022 17:54:26 - INFO - codeparrot_training - Step 3923: {'lr': 0.0004980225182668008, 'samples': 753408, 'steps': 3923, 'loss/train': 1.7071977853775024} 01/28/2022 17:54:31 - INFO - codeparrot_training - Step 3924: {'lr': 0.00049802046378795, 'samples': 753600, 'steps': 3924, 'loss/train': 1.4804002940654755} 01/28/2022 17:54:36 - INFO - codeparrot_training - Step 3925: {'lr': 0.0004980184082466583, 'samples': 753792, 'steps': 3925, 'loss/train': 1.14259672164917} 01/28/2022 17:54:40 - INFO - codeparrot_training - Step 3926: {'lr': 0.0004980163516429346, 'samples': 753984, 'steps': 3926, 'loss/train': 1.8966840505599976} 01/28/2022 17:54:44 - INFO - codeparrot_training - Step 3927: {'lr': 0.0004980142939767876, 'samples': 754176, 'steps': 3927, 'loss/train': 1.9021035432815552} 01/28/2022 17:54:50 - INFO - codeparrot_training - Step 3928: {'lr': 0.000498012235248226, 'samples': 754368, 'steps': 3928, 'loss/train': 0.10219147056341171} 01/28/2022 17:54:54 - INFO - codeparrot_training - Step 3929: {'lr': 0.0004980101754572589, 'samples': 754560, 'steps': 3929, 'loss/train': 1.3296090066432953} 01/28/2022 17:54:58 - INFO - codeparrot_training - Step 3930: {'lr': 0.0004980081146038948, 'samples': 754752, 'steps': 3930, 'loss/train': 1.911730706691742} 01/28/2022 17:55:03 - INFO - codeparrot_training - Step 3931: {'lr': 0.0004980060526881429, 'samples': 754944, 'steps': 3931, 'loss/train': 2.319004476070404} 01/28/2022 17:55:07 - INFO - codeparrot_training - Step 3932: {'lr': 0.0004980039897100115, 'samples': 755136, 'steps': 3932, 'loss/train': 2.751273214817047} 01/28/2022 17:55:12 - INFO - codeparrot_training - Step 3933: {'lr': 0.0004980019256695101, 'samples': 755328, 'steps': 3933, 'loss/train': 2.8066141605377197} 01/28/2022 17:55:16 - INFO - codeparrot_training - Step 3934: {'lr': 0.000497999860566647, 'samples': 755520, 'steps': 3934, 'loss/train': 1.6347396969795227} 01/28/2022 17:55:20 - INFO - codeparrot_training - Step 3935: {'lr': 0.0004979977944014313, 'samples': 755712, 'steps': 3935, 'loss/train': 2.014053761959076} 01/28/2022 17:55:25 - INFO - codeparrot_training - Step 3936: {'lr': 0.0004979957271738718, 'samples': 755904, 'steps': 3936, 'loss/train': 2.324941635131836} 01/28/2022 17:55:29 - INFO - codeparrot_training - Step 3937: {'lr': 0.0004979936588839773, 'samples': 756096, 'steps': 3937, 'loss/train': 2.1910396814346313} 01/28/2022 17:55:33 - INFO - codeparrot_training - Step 3938: {'lr': 0.0004979915895317567, 'samples': 756288, 'steps': 3938, 'loss/train': 0.5106831192970276} 01/28/2022 17:55:39 - INFO - codeparrot_training - Step 3939: {'lr': 0.000497989519117219, 'samples': 756480, 'steps': 3939, 'loss/train': 2.223763167858124} 01/28/2022 17:55:43 - INFO - codeparrot_training - Step 3940: {'lr': 0.0004979874476403729, 'samples': 756672, 'steps': 3940, 'loss/train': 2.3481014370918274} 01/28/2022 17:55:47 - INFO - codeparrot_training - Step 3941: {'lr': 0.0004979853751012273, 'samples': 756864, 'steps': 3941, 'loss/train': 1.1338672935962677} 01/28/2022 17:55:52 - INFO - codeparrot_training - Step 3942: {'lr': 0.0004979833014997911, 'samples': 757056, 'steps': 3942, 'loss/train': 2.1265187859535217} 01/28/2022 17:55:56 - INFO - codeparrot_training - Step 3943: {'lr': 0.0004979812268360731, 'samples': 757248, 'steps': 3943, 'loss/train': 2.223349928855896} 01/28/2022 17:56:01 - INFO - codeparrot_training - Step 3944: {'lr': 0.0004979791511100823, 'samples': 757440, 'steps': 3944, 'loss/train': 1.8944846391677856} 01/28/2022 17:56:05 - INFO - codeparrot_training - Step 3945: {'lr': 0.0004979770743218276, 'samples': 757632, 'steps': 3945, 'loss/train': 1.6984906196594238} 01/28/2022 17:56:10 - INFO - codeparrot_training - Step 3946: {'lr': 0.0004979749964713179, 'samples': 757824, 'steps': 3946, 'loss/train': 0.9422746896743774} 01/28/2022 17:56:14 - INFO - codeparrot_training - Step 3947: {'lr': 0.000497972917558562, 'samples': 758016, 'steps': 3947, 'loss/train': 2.0814929008483887} 01/28/2022 17:56:19 - INFO - codeparrot_training - Step 3948: {'lr': 0.0004979708375835688, 'samples': 758208, 'steps': 3948, 'loss/train': 1.6004717946052551} 01/28/2022 17:56:23 - INFO - codeparrot_training - Step 3949: {'lr': 0.0004979687565463475, 'samples': 758400, 'steps': 3949, 'loss/train': 1.7893479466438293} 01/28/2022 17:56:27 - INFO - codeparrot_training - Step 3950: {'lr': 0.0004979666744469065, 'samples': 758592, 'steps': 3950, 'loss/train': 2.025412380695343} 01/28/2022 17:56:32 - INFO - codeparrot_training - Step 3951: {'lr': 0.0004979645912852552, 'samples': 758784, 'steps': 3951, 'loss/train': 2.162503480911255} 01/28/2022 17:56:36 - INFO - codeparrot_training - Step 3952: {'lr': 0.0004979625070614022, 'samples': 758976, 'steps': 3952, 'loss/train': 1.7173821330070496} 01/28/2022 17:56:41 - INFO - codeparrot_training - Step 3953: {'lr': 0.0004979604217753566, 'samples': 759168, 'steps': 3953, 'loss/train': 2.8510614037513733} 01/28/2022 17:56:45 - INFO - codeparrot_training - Step 3954: {'lr': 0.0004979583354271273, 'samples': 759360, 'steps': 3954, 'loss/train': 1.9384594559669495} 01/28/2022 17:56:49 - INFO - codeparrot_training - Step 3955: {'lr': 0.0004979562480167232, 'samples': 759552, 'steps': 3955, 'loss/train': 1.9094563722610474} 01/28/2022 17:56:53 - INFO - codeparrot_training - Step 3956: {'lr': 0.0004979541595441534, 'samples': 759744, 'steps': 3956, 'loss/train': 1.422773391008377} 01/28/2022 17:56:58 - INFO - codeparrot_training - Step 3957: {'lr': 0.0004979520700094265, 'samples': 759936, 'steps': 3957, 'loss/train': 0.9266721904277802} 01/28/2022 17:57:03 - INFO - codeparrot_training - Step 3958: {'lr': 0.0004979499794125518, 'samples': 760128, 'steps': 3958, 'loss/train': 2.1288442611694336} 01/28/2022 17:57:08 - INFO - codeparrot_training - Step 3959: {'lr': 0.0004979478877535382, 'samples': 760320, 'steps': 3959, 'loss/train': 2.620185077190399} 01/28/2022 17:57:12 - INFO - codeparrot_training - Step 3960: {'lr': 0.0004979457950323945, 'samples': 760512, 'steps': 3960, 'loss/train': 2.3999215364456177} 01/28/2022 17:57:16 - INFO - codeparrot_training - Step 3961: {'lr': 0.0004979437012491297, 'samples': 760704, 'steps': 3961, 'loss/train': 1.0778078734874725} 01/28/2022 17:57:20 - INFO - codeparrot_training - Step 3962: {'lr': 0.0004979416064037528, 'samples': 760896, 'steps': 3962, 'loss/train': 2.1957762837409973} 01/28/2022 17:57:27 - INFO - codeparrot_training - Step 3963: {'lr': 0.0004979395104962728, 'samples': 761088, 'steps': 3963, 'loss/train': 2.2769320607185364} 01/28/2022 17:57:31 - INFO - codeparrot_training - Step 3964: {'lr': 0.0004979374135266987, 'samples': 761280, 'steps': 3964, 'loss/train': 1.9273332953453064} 01/28/2022 17:57:36 - INFO - codeparrot_training - Step 3965: {'lr': 0.0004979353154950394, 'samples': 761472, 'steps': 3965, 'loss/train': 1.726534366607666} 01/28/2022 17:57:40 - INFO - codeparrot_training - Step 3966: {'lr': 0.0004979332164013041, 'samples': 761664, 'steps': 3966, 'loss/train': 2.9027257561683655} 01/28/2022 17:57:44 - INFO - codeparrot_training - Step 3967: {'lr': 0.0004979311162455015, 'samples': 761856, 'steps': 3967, 'loss/train': 1.612653136253357} 01/28/2022 17:57:49 - INFO - codeparrot_training - Step 3968: {'lr': 0.0004979290150276407, 'samples': 762048, 'steps': 3968, 'loss/train': 1.8574129343032837} 01/28/2022 17:57:53 - INFO - codeparrot_training - Step 3969: {'lr': 0.0004979269127477308, 'samples': 762240, 'steps': 3969, 'loss/train': 2.232189416885376} 01/28/2022 17:57:58 - INFO - codeparrot_training - Step 3970: {'lr': 0.0004979248094057806, 'samples': 762432, 'steps': 3970, 'loss/train': 1.8178887963294983} 01/28/2022 17:58:02 - INFO - codeparrot_training - Step 3971: {'lr': 0.0004979227050017994, 'samples': 762624, 'steps': 3971, 'loss/train': 2.1406951546669006} 01/28/2022 17:58:06 - INFO - codeparrot_training - Step 3972: {'lr': 0.000497920599535796, 'samples': 762816, 'steps': 3972, 'loss/train': 1.9343889355659485} 01/28/2022 17:58:11 - INFO - codeparrot_training - Step 3973: {'lr': 0.0004979184930077794, 'samples': 763008, 'steps': 3973, 'loss/train': 1.9937823414802551} 01/28/2022 17:58:15 - INFO - codeparrot_training - Step 3974: {'lr': 0.0004979163854177588, 'samples': 763200, 'steps': 3974, 'loss/train': 2.2167967557907104} 01/28/2022 17:58:20 - INFO - codeparrot_training - Step 3975: {'lr': 0.0004979142767657432, 'samples': 763392, 'steps': 3975, 'loss/train': 1.138877034187317} 01/28/2022 17:58:24 - INFO - codeparrot_training - Step 3976: {'lr': 0.0004979121670517413, 'samples': 763584, 'steps': 3976, 'loss/train': 1.78169846534729} 01/28/2022 17:58:28 - INFO - codeparrot_training - Step 3977: {'lr': 0.0004979100562757626, 'samples': 763776, 'steps': 3977, 'loss/train': 2.129475474357605} 01/28/2022 17:58:34 - INFO - codeparrot_training - Step 3978: {'lr': 0.0004979079444378159, 'samples': 763968, 'steps': 3978, 'loss/train': 1.6194303631782532} 01/28/2022 17:58:38 - INFO - codeparrot_training - Step 3979: {'lr': 0.0004979058315379103, 'samples': 764160, 'steps': 3979, 'loss/train': 2.041813910007477} 01/28/2022 17:58:42 - INFO - codeparrot_training - Step 3980: {'lr': 0.0004979037175760548, 'samples': 764352, 'steps': 3980, 'loss/train': 1.7203370332717896} 01/28/2022 17:58:47 - INFO - codeparrot_training - Step 3981: {'lr': 0.0004979016025522586, 'samples': 764544, 'steps': 3981, 'loss/train': 1.8187174201011658} 01/28/2022 17:58:51 - INFO - codeparrot_training - Step 3982: {'lr': 0.0004978994864665305, 'samples': 764736, 'steps': 3982, 'loss/train': 1.7419863939285278} 01/28/2022 17:58:56 - INFO - codeparrot_training - Step 3983: {'lr': 0.0004978973693188797, 'samples': 764928, 'steps': 3983, 'loss/train': 2.6656227707862854} 01/28/2022 17:59:00 - INFO - codeparrot_training - Step 3984: {'lr': 0.0004978952511093155, 'samples': 765120, 'steps': 3984, 'loss/train': 1.464068055152893} 01/28/2022 17:59:04 - INFO - codeparrot_training - Step 3985: {'lr': 0.0004978931318378465, 'samples': 765312, 'steps': 3985, 'loss/train': 0.9005316495895386} 01/28/2022 17:59:08 - INFO - codeparrot_training - Step 3986: {'lr': 0.0004978910115044822, 'samples': 765504, 'steps': 3986, 'loss/train': 2.193739414215088} 01/28/2022 17:59:13 - INFO - codeparrot_training - Step 3987: {'lr': 0.0004978888901092315, 'samples': 765696, 'steps': 3987, 'loss/train': 1.7289653420448303} 01/28/2022 17:59:18 - INFO - codeparrot_training - Step 3988: {'lr': 0.0004978867676521035, 'samples': 765888, 'steps': 3988, 'loss/train': 1.4668865203857422} 01/28/2022 17:59:23 - INFO - codeparrot_training - Step 3989: {'lr': 0.0004978846441331073, 'samples': 766080, 'steps': 3989, 'loss/train': 1.8611608743667603} 01/28/2022 17:59:27 - INFO - codeparrot_training - Step 3990: {'lr': 0.000497882519552252, 'samples': 766272, 'steps': 3990, 'loss/train': 2.323066234588623} 01/28/2022 17:59:31 - INFO - codeparrot_training - Step 3991: {'lr': 0.0004978803939095466, 'samples': 766464, 'steps': 3991, 'loss/train': 1.446106106042862} 01/28/2022 17:59:35 - INFO - codeparrot_training - Step 3992: {'lr': 0.0004978782672050004, 'samples': 766656, 'steps': 3992, 'loss/train': 3.0235061645507812} 01/28/2022 17:59:40 - INFO - codeparrot_training - Step 3993: {'lr': 0.0004978761394386224, 'samples': 766848, 'steps': 3993, 'loss/train': 3.029785394668579} 01/28/2022 17:59:45 - INFO - codeparrot_training - Step 3994: {'lr': 0.0004978740106104218, 'samples': 767040, 'steps': 3994, 'loss/train': 1.6478347778320312} 01/28/2022 17:59:49 - INFO - codeparrot_training - Step 3995: {'lr': 0.0004978718807204076, 'samples': 767232, 'steps': 3995, 'loss/train': 1.7435545921325684} 01/28/2022 17:59:53 - INFO - codeparrot_training - Step 3996: {'lr': 0.0004978697497685889, 'samples': 767424, 'steps': 3996, 'loss/train': 2.579038918018341} 01/28/2022 17:59:57 - INFO - codeparrot_training - Step 3997: {'lr': 0.0004978676177549749, 'samples': 767616, 'steps': 3997, 'loss/train': 2.2389814853668213} 01/28/2022 18:00:02 - INFO - codeparrot_training - Step 3998: {'lr': 0.0004978654846795748, 'samples': 767808, 'steps': 3998, 'loss/train': 1.5715302228927612} 01/28/2022 18:00:07 - INFO - codeparrot_training - Step 3999: {'lr': 0.0004978633505423976, 'samples': 768000, 'steps': 3999, 'loss/train': 1.1117471158504486} 01/28/2022 18:00:07 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/28/2022 18:00:40 - WARNING - huggingface_hub.repository - Several commits (2) will be pushed upstream. 01/28/2022 18:00:40 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/28/2022 18:02:07 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 7313137..4e36df4 expert-dust-2 -> expert-dust-2 01/28/2022 18:02:12 - INFO - codeparrot_training - Step 4000: {'lr': 0.0004978612153434526, 'samples': 768192, 'steps': 4000, 'loss/train': 2.065965235233307} 01/28/2022 18:02:17 - INFO - codeparrot_training - Step 4001: {'lr': 0.0004978590790827488, 'samples': 768384, 'steps': 4001, 'loss/train': 2.017137587070465} 01/28/2022 18:02:21 - INFO - codeparrot_training - Step 4002: {'lr': 0.0004978569417602955, 'samples': 768576, 'steps': 4002, 'loss/train': 1.3672796487808228} 01/28/2022 18:02:27 - INFO - codeparrot_training - Step 4003: {'lr': 0.0004978548033761017, 'samples': 768768, 'steps': 4003, 'loss/train': 1.947236180305481} 01/28/2022 18:02:32 - INFO - codeparrot_training - Step 4004: {'lr': 0.0004978526639301766, 'samples': 768960, 'steps': 4004, 'loss/train': 2.1703197956085205} 01/28/2022 18:02:36 - INFO - codeparrot_training - Step 4005: {'lr': 0.0004978505234225294, 'samples': 769152, 'steps': 4005, 'loss/train': 2.4987359046936035} 01/28/2022 18:02:40 - INFO - codeparrot_training - Step 4006: {'lr': 0.0004978483818531693, 'samples': 769344, 'steps': 4006, 'loss/train': 2.1793325543403625} 01/28/2022 18:02:44 - INFO - codeparrot_training - Step 4007: {'lr': 0.0004978462392221054, 'samples': 769536, 'steps': 4007, 'loss/train': 2.7289686799049377} 01/28/2022 18:02:50 - INFO - codeparrot_training - Step 4008: {'lr': 0.0004978440955293468, 'samples': 769728, 'steps': 4008, 'loss/train': 1.8527601957321167} 01/28/2022 18:02:54 - INFO - codeparrot_training - Step 4009: {'lr': 0.000497841950774903, 'samples': 769920, 'steps': 4009, 'loss/train': 2.210744082927704} 01/28/2022 18:02:58 - INFO - codeparrot_training - Step 4010: {'lr': 0.0004978398049587828, 'samples': 770112, 'steps': 4010, 'loss/train': 1.8376838564872742} 01/28/2022 18:03:02 - INFO - codeparrot_training - Step 4011: {'lr': 0.0004978376580809957, 'samples': 770304, 'steps': 4011, 'loss/train': 2.7833086252212524} 01/28/2022 18:03:07 - INFO - codeparrot_training - Step 4012: {'lr': 0.0004978355101415507, 'samples': 770496, 'steps': 4012, 'loss/train': 2.2030885219573975} 01/28/2022 18:03:12 - INFO - codeparrot_training - Step 4013: {'lr': 0.0004978333611404571, 'samples': 770688, 'steps': 4013, 'loss/train': 1.5947436690330505} 01/28/2022 18:03:16 - INFO - codeparrot_training - Step 4014: {'lr': 0.0004978312110777241, 'samples': 770880, 'steps': 4014, 'loss/train': 1.106988251209259} 01/28/2022 18:03:20 - INFO - codeparrot_training - Step 4015: {'lr': 0.0004978290599533609, 'samples': 771072, 'steps': 4015, 'loss/train': 2.1578423380851746} 01/28/2022 18:03:25 - INFO - codeparrot_training - Step 4016: {'lr': 0.0004978269077673766, 'samples': 771264, 'steps': 4016, 'loss/train': 0.47042790055274963} 01/28/2022 18:03:29 - INFO - codeparrot_training - Step 4017: {'lr': 0.0004978247545197806, 'samples': 771456, 'steps': 4017, 'loss/train': 2.102486729621887} 01/28/2022 18:03:34 - INFO - codeparrot_training - Step 4018: {'lr': 0.0004978226002105821, 'samples': 771648, 'steps': 4018, 'loss/train': 2.675241708755493} 01/28/2022 18:03:38 - INFO - codeparrot_training - Step 4019: {'lr': 0.0004978204448397902, 'samples': 771840, 'steps': 4019, 'loss/train': 1.1712086498737335} 01/28/2022 18:03:43 - INFO - codeparrot_training - Step 4020: {'lr': 0.0004978182884074142, 'samples': 772032, 'steps': 4020, 'loss/train': 1.3026979565620422} 01/28/2022 18:03:47 - INFO - codeparrot_training - Step 4021: {'lr': 0.0004978161309134633, 'samples': 772224, 'steps': 4021, 'loss/train': 2.066890060901642} 01/28/2022 18:03:51 - INFO - codeparrot_training - Step 4022: {'lr': 0.0004978139723579469, 'samples': 772416, 'steps': 4022, 'loss/train': 1.8305127024650574} 01/28/2022 18:03:57 - INFO - codeparrot_training - Step 4023: {'lr': 0.0004978118127408741, 'samples': 772608, 'steps': 4023, 'loss/train': 0.3988284319639206} 01/28/2022 18:04:02 - INFO - codeparrot_training - Step 4024: {'lr': 0.0004978096520622541, 'samples': 772800, 'steps': 4024, 'loss/train': 2.041198432445526} 01/28/2022 18:04:06 - INFO - codeparrot_training - Step 4025: {'lr': 0.0004978074903220964, 'samples': 772992, 'steps': 4025, 'loss/train': 1.8327203392982483} 01/28/2022 18:04:10 - INFO - codeparrot_training - Step 4026: {'lr': 0.0004978053275204099, 'samples': 773184, 'steps': 4026, 'loss/train': 1.6116229891777039} 01/28/2022 18:04:14 - INFO - codeparrot_training - Step 4027: {'lr': 0.0004978031636572042, 'samples': 773376, 'steps': 4027, 'loss/train': 2.338808834552765} 01/28/2022 18:04:20 - INFO - codeparrot_training - Step 4028: {'lr': 0.0004978009987324884, 'samples': 773568, 'steps': 4028, 'loss/train': 2.586752951145172} 01/28/2022 18:04:24 - INFO - codeparrot_training - Step 4029: {'lr': 0.0004977988327462718, 'samples': 773760, 'steps': 4029, 'loss/train': 2.405592441558838} 01/28/2022 18:04:29 - INFO - codeparrot_training - Step 4030: {'lr': 0.0004977966656985637, 'samples': 773952, 'steps': 4030, 'loss/train': 1.6012389063835144} 01/28/2022 18:04:33 - INFO - codeparrot_training - Step 4031: {'lr': 0.0004977944975893733, 'samples': 774144, 'steps': 4031, 'loss/train': 2.082677721977234} 01/28/2022 18:04:37 - INFO - codeparrot_training - Step 4032: {'lr': 0.00049779232841871, 'samples': 774336, 'steps': 4032, 'loss/train': 2.1791033148765564} 01/28/2022 18:04:43 - INFO - codeparrot_training - Step 4033: {'lr': 0.0004977901581865831, 'samples': 774528, 'steps': 4033, 'loss/train': 2.7753522992134094} 01/28/2022 18:04:48 - INFO - codeparrot_training - Step 4034: {'lr': 0.0004977879868930018, 'samples': 774720, 'steps': 4034, 'loss/train': 1.4569023549556732} 01/28/2022 18:04:52 - INFO - codeparrot_training - Step 4035: {'lr': 0.0004977858145379754, 'samples': 774912, 'steps': 4035, 'loss/train': 2.536212980747223} 01/28/2022 18:04:56 - INFO - codeparrot_training - Step 4036: {'lr': 0.0004977836411215133, 'samples': 775104, 'steps': 4036, 'loss/train': 1.7875185012817383} 01/28/2022 18:05:00 - INFO - codeparrot_training - Step 4037: {'lr': 0.0004977814666436248, 'samples': 775296, 'steps': 4037, 'loss/train': 2.2038960456848145} 01/28/2022 18:05:05 - INFO - codeparrot_training - Step 4038: {'lr': 0.0004977792911043191, 'samples': 775488, 'steps': 4038, 'loss/train': 1.7666882872581482} 01/28/2022 18:05:10 - INFO - codeparrot_training - Step 4039: {'lr': 0.0004977771145036056, 'samples': 775680, 'steps': 4039, 'loss/train': 1.8483272194862366} 01/28/2022 18:05:14 - INFO - codeparrot_training - Step 4040: {'lr': 0.0004977749368414937, 'samples': 775872, 'steps': 4040, 'loss/train': 1.3048843145370483} 01/28/2022 18:05:18 - INFO - codeparrot_training - Step 4041: {'lr': 0.0004977727581179926, 'samples': 776064, 'steps': 4041, 'loss/train': 1.032491683959961} 01/28/2022 18:05:22 - INFO - codeparrot_training - Step 4042: {'lr': 0.0004977705783331117, 'samples': 776256, 'steps': 4042, 'loss/train': 1.3291069865226746} 01/28/2022 18:05:28 - INFO - codeparrot_training - Step 4043: {'lr': 0.0004977683974868603, 'samples': 776448, 'steps': 4043, 'loss/train': 2.747146725654602} 01/28/2022 18:05:32 - INFO - codeparrot_training - Step 4044: {'lr': 0.0004977662155792478, 'samples': 776640, 'steps': 4044, 'loss/train': 2.506002724170685} 01/28/2022 18:05:36 - INFO - codeparrot_training - Step 4045: {'lr': 0.0004977640326102834, 'samples': 776832, 'steps': 4045, 'loss/train': 1.757701814174652} 01/28/2022 18:05:41 - INFO - codeparrot_training - Step 4046: {'lr': 0.0004977618485799767, 'samples': 777024, 'steps': 4046, 'loss/train': 7.467373609542847} 01/28/2022 18:05:45 - INFO - codeparrot_training - Step 4047: {'lr': 0.0004977596634883368, 'samples': 777216, 'steps': 4047, 'loss/train': 3.35400652885437} 01/28/2022 18:05:49 - INFO - codeparrot_training - Step 4048: {'lr': 0.0004977574773353732, 'samples': 777408, 'steps': 4048, 'loss/train': 1.5756193399429321} 01/28/2022 18:05:55 - INFO - codeparrot_training - Step 4049: {'lr': 0.0004977552901210952, 'samples': 777600, 'steps': 4049, 'loss/train': 1.4140589833259583} 01/28/2022 18:06:00 - INFO - codeparrot_training - Step 4050: {'lr': 0.0004977531018455124, 'samples': 777792, 'steps': 4050, 'loss/train': 2.1815507411956787} 01/28/2022 18:06:04 - INFO - codeparrot_training - Step 4051: {'lr': 0.0004977509125086338, 'samples': 777984, 'steps': 4051, 'loss/train': 2.0535255074501038} 01/28/2022 18:06:08 - INFO - codeparrot_training - Step 4052: {'lr': 0.000497748722110469, 'samples': 778176, 'steps': 4052, 'loss/train': 2.2021275758743286} 01/28/2022 18:06:12 - INFO - codeparrot_training - Step 4053: {'lr': 0.0004977465306510273, 'samples': 778368, 'steps': 4053, 'loss/train': 1.5761223435401917} 01/28/2022 18:06:17 - INFO - codeparrot_training - Step 4054: {'lr': 0.0004977443381303182, 'samples': 778560, 'steps': 4054, 'loss/train': 1.5920079946517944} 01/28/2022 18:06:22 - INFO - codeparrot_training - Step 4055: {'lr': 0.000497742144548351, 'samples': 778752, 'steps': 4055, 'loss/train': 2.446621835231781} 01/28/2022 18:06:26 - INFO - codeparrot_training - Step 4056: {'lr': 0.0004977399499051351, 'samples': 778944, 'steps': 4056, 'loss/train': 1.8542110919952393} 01/28/2022 18:06:30 - INFO - codeparrot_training - Step 4057: {'lr': 0.0004977377542006799, 'samples': 779136, 'steps': 4057, 'loss/train': 1.4728157222270966} 01/28/2022 18:06:34 - INFO - codeparrot_training - Step 4058: {'lr': 0.0004977355574349949, 'samples': 779328, 'steps': 4058, 'loss/train': 2.331186354160309} 01/28/2022 18:06:40 - INFO - codeparrot_training - Step 4059: {'lr': 0.0004977333596080894, 'samples': 779520, 'steps': 4059, 'loss/train': 1.7447156310081482} 01/28/2022 18:06:44 - INFO - codeparrot_training - Step 4060: {'lr': 0.0004977311607199729, 'samples': 779712, 'steps': 4060, 'loss/train': 1.3923713564872742} 01/28/2022 18:06:48 - INFO - codeparrot_training - Step 4061: {'lr': 0.0004977289607706547, 'samples': 779904, 'steps': 4061, 'loss/train': 1.9490638375282288} 01/28/2022 18:06:52 - INFO - codeparrot_training - Step 4062: {'lr': 0.0004977267597601443, 'samples': 780096, 'steps': 4062, 'loss/train': 1.7280072569847107} 01/28/2022 18:06:57 - INFO - codeparrot_training - Step 4063: {'lr': 0.0004977245576884511, 'samples': 780288, 'steps': 4063, 'loss/train': 1.8545714020729065} 01/28/2022 18:07:03 - INFO - codeparrot_training - Step 4064: {'lr': 0.0004977223545555847, 'samples': 780480, 'steps': 4064, 'loss/train': 1.750346839427948} 01/28/2022 18:07:07 - INFO - codeparrot_training - Step 4065: {'lr': 0.0004977201503615543, 'samples': 780672, 'steps': 4065, 'loss/train': 1.4848122596740723} 01/28/2022 18:07:11 - INFO - codeparrot_training - Step 4066: {'lr': 0.0004977179451063694, 'samples': 780864, 'steps': 4066, 'loss/train': 2.2263944149017334} 01/28/2022 18:07:16 - INFO - codeparrot_training - Step 4067: {'lr': 0.0004977157387900395, 'samples': 781056, 'steps': 4067, 'loss/train': 2.2165226340293884} 01/28/2022 18:07:21 - INFO - codeparrot_training - Step 4068: {'lr': 0.0004977135314125741, 'samples': 781248, 'steps': 4068, 'loss/train': 1.8652592897415161} 01/28/2022 18:07:25 - INFO - codeparrot_training - Step 4069: {'lr': 0.0004977113229739825, 'samples': 781440, 'steps': 4069, 'loss/train': 1.6025129556655884} 01/28/2022 18:07:29 - INFO - codeparrot_training - Step 4070: {'lr': 0.0004977091134742743, 'samples': 781632, 'steps': 4070, 'loss/train': 2.232076942920685} 01/28/2022 18:07:34 - INFO - codeparrot_training - Step 4071: {'lr': 0.0004977069029134588, 'samples': 781824, 'steps': 4071, 'loss/train': 4.76726496219635} 01/28/2022 18:07:38 - INFO - codeparrot_training - Step 4072: {'lr': 0.0004977046912915458, 'samples': 782016, 'steps': 4072, 'loss/train': 2.085834324359894} 01/28/2022 18:07:42 - INFO - codeparrot_training - Step 4073: {'lr': 0.0004977024786085444, 'samples': 782208, 'steps': 4073, 'loss/train': 3.2885584831237793} 01/28/2022 18:07:48 - INFO - codeparrot_training - Step 4074: {'lr': 0.0004977002648644642, 'samples': 782400, 'steps': 4074, 'loss/train': 2.1826189756393433} 01/28/2022 18:07:52 - INFO - codeparrot_training - Step 4075: {'lr': 0.0004976980500593149, 'samples': 782592, 'steps': 4075, 'loss/train': 0.9357400238513947} 01/28/2022 18:07:57 - INFO - codeparrot_training - Step 4076: {'lr': 0.0004976958341931057, 'samples': 782784, 'steps': 4076, 'loss/train': 1.6471722722053528} 01/28/2022 18:08:01 - INFO - codeparrot_training - Step 4077: {'lr': 0.0004976936172658462, 'samples': 782976, 'steps': 4077, 'loss/train': 1.5928325057029724} 01/28/2022 18:08:05 - INFO - codeparrot_training - Step 4078: {'lr': 0.0004976913992775459, 'samples': 783168, 'steps': 4078, 'loss/train': 0.976947695016861} 01/28/2022 18:08:10 - INFO - codeparrot_training - Step 4079: {'lr': 0.0004976891802282143, 'samples': 783360, 'steps': 4079, 'loss/train': 1.884019911289215} 01/28/2022 18:08:15 - INFO - codeparrot_training - Step 4080: {'lr': 0.0004976869601178609, 'samples': 783552, 'steps': 4080, 'loss/train': 1.8797079920768738} 01/28/2022 18:08:19 - INFO - codeparrot_training - Step 4081: {'lr': 0.0004976847389464952, 'samples': 783744, 'steps': 4081, 'loss/train': 1.5573441982269287} 01/28/2022 18:08:23 - INFO - codeparrot_training - Step 4082: {'lr': 0.0004976825167141268, 'samples': 783936, 'steps': 4082, 'loss/train': 2.268274426460266} 01/28/2022 18:08:27 - INFO - codeparrot_training - Step 4083: {'lr': 0.000497680293420765, 'samples': 784128, 'steps': 4083, 'loss/train': 1.0259564816951752} 01/28/2022 18:08:33 - INFO - codeparrot_training - Step 4084: {'lr': 0.0004976780690664196, 'samples': 784320, 'steps': 4084, 'loss/train': 2.1221349835395813} 01/28/2022 18:08:37 - INFO - codeparrot_training - Step 4085: {'lr': 0.0004976758436511, 'samples': 784512, 'steps': 4085, 'loss/train': 1.7157617211341858} 01/28/2022 18:08:41 - INFO - codeparrot_training - Step 4086: {'lr': 0.0004976736171748156, 'samples': 784704, 'steps': 4086, 'loss/train': 1.5096704363822937} 01/28/2022 18:08:45 - INFO - codeparrot_training - Step 4087: {'lr': 0.0004976713896375762, 'samples': 784896, 'steps': 4087, 'loss/train': 1.9774680733680725} 01/28/2022 18:08:50 - INFO - codeparrot_training - Step 4088: {'lr': 0.0004976691610393911, 'samples': 785088, 'steps': 4088, 'loss/train': 1.6342803239822388} 01/28/2022 18:08:55 - INFO - codeparrot_training - Step 4089: {'lr': 0.0004976669313802701, 'samples': 785280, 'steps': 4089, 'loss/train': 2.0317867398262024} 01/28/2022 18:09:00 - INFO - codeparrot_training - Step 4090: {'lr': 0.0004976647006602225, 'samples': 785472, 'steps': 4090, 'loss/train': 1.2674315571784973} 01/28/2022 18:09:04 - INFO - codeparrot_training - Step 4091: {'lr': 0.0004976624688792581, 'samples': 785664, 'steps': 4091, 'loss/train': 1.6068849563598633} 01/28/2022 18:09:08 - INFO - codeparrot_training - Step 4092: {'lr': 0.0004976602360373861, 'samples': 785856, 'steps': 4092, 'loss/train': 0.3687591031193733} 01/28/2022 18:09:12 - INFO - codeparrot_training - Step 4093: {'lr': 0.0004976580021346164, 'samples': 786048, 'steps': 4093, 'loss/train': 1.5734290480613708} 01/28/2022 18:09:19 - INFO - codeparrot_training - Step 4094: {'lr': 0.0004976557671709585, 'samples': 786240, 'steps': 4094, 'loss/train': 2.66477108001709} 01/28/2022 18:09:23 - INFO - codeparrot_training - Step 4095: {'lr': 0.0004976535311464219, 'samples': 786432, 'steps': 4095, 'loss/train': 2.1654219031333923} 01/28/2022 18:09:28 - INFO - codeparrot_training - Step 4096: {'lr': 0.0004976512940610162, 'samples': 786624, 'steps': 4096, 'loss/train': 2.0862380862236023} 01/28/2022 18:09:32 - INFO - codeparrot_training - Step 4097: {'lr': 0.0004976490559147511, 'samples': 786816, 'steps': 4097, 'loss/train': 1.0763318538665771} 01/28/2022 18:09:36 - INFO - codeparrot_training - Step 4098: {'lr': 0.0004976468167076359, 'samples': 787008, 'steps': 4098, 'loss/train': 1.9571009874343872} 01/28/2022 18:09:41 - INFO - codeparrot_training - Step 4099: {'lr': 0.0004976445764396805, 'samples': 787200, 'steps': 4099, 'loss/train': 2.3144508004188538} 01/28/2022 18:09:45 - INFO - codeparrot_training - Step 4100: {'lr': 0.0004976423351108943, 'samples': 787392, 'steps': 4100, 'loss/train': 1.428719848394394} 01/28/2022 18:09:50 - INFO - codeparrot_training - Step 4101: {'lr': 0.0004976400927212871, 'samples': 787584, 'steps': 4101, 'loss/train': 1.6553826928138733} 01/28/2022 18:09:54 - INFO - codeparrot_training - Step 4102: {'lr': 0.0004976378492708681, 'samples': 787776, 'steps': 4102, 'loss/train': 2.0725765228271484} 01/28/2022 18:09:58 - INFO - codeparrot_training - Step 4103: {'lr': 0.0004976356047596475, 'samples': 787968, 'steps': 4103, 'loss/train': 2.256852149963379} 01/28/2022 18:10:04 - INFO - codeparrot_training - Step 4104: {'lr': 0.0004976333591876344, 'samples': 788160, 'steps': 4104, 'loss/train': 0.8542262613773346} 01/28/2022 18:10:08 - INFO - codeparrot_training - Step 4105: {'lr': 0.0004976311125548387, 'samples': 788352, 'steps': 4105, 'loss/train': 1.9526220560073853} 01/28/2022 18:10:13 - INFO - codeparrot_training - Step 4106: {'lr': 0.00049762886486127, 'samples': 788544, 'steps': 4106, 'loss/train': 1.5848949551582336} 01/28/2022 18:10:17 - INFO - codeparrot_training - Step 4107: {'lr': 0.0004976266161069379, 'samples': 788736, 'steps': 4107, 'loss/train': 1.719644844532013} 01/28/2022 18:10:21 - INFO - codeparrot_training - Step 4108: {'lr': 0.0004976243662918518, 'samples': 788928, 'steps': 4108, 'loss/train': 1.1946769952774048} 01/28/2022 18:10:27 - INFO - codeparrot_training - Step 4109: {'lr': 0.0004976221154160217, 'samples': 789120, 'steps': 4109, 'loss/train': 0.7125687897205353} 01/28/2022 18:10:31 - INFO - codeparrot_training - Step 4110: {'lr': 0.0004976198634794571, 'samples': 789312, 'steps': 4110, 'loss/train': 1.7094005942344666} 01/28/2022 18:10:35 - INFO - codeparrot_training - Step 4111: {'lr': 0.0004976176104821675, 'samples': 789504, 'steps': 4111, 'loss/train': 1.83381986618042} 01/28/2022 18:10:40 - INFO - codeparrot_training - Step 4112: {'lr': 0.0004976153564241628, 'samples': 789696, 'steps': 4112, 'loss/train': 1.122286856174469} 01/28/2022 18:10:44 - INFO - codeparrot_training - Step 4113: {'lr': 0.0004976131013054526, 'samples': 789888, 'steps': 4113, 'loss/train': 2.1235654950141907} 01/28/2022 18:10:49 - INFO - codeparrot_training - Step 4114: {'lr': 0.0004976108451260464, 'samples': 790080, 'steps': 4114, 'loss/train': 2.126357853412628} 01/28/2022 18:10:53 - INFO - codeparrot_training - Step 4115: {'lr': 0.000497608587885954, 'samples': 790272, 'steps': 4115, 'loss/train': 2.371664822101593} 01/28/2022 18:10:58 - INFO - codeparrot_training - Step 4116: {'lr': 0.0004976063295851849, 'samples': 790464, 'steps': 4116, 'loss/train': 1.6318241357803345} 01/28/2022 18:11:02 - INFO - codeparrot_training - Step 4117: {'lr': 0.000497604070223749, 'samples': 790656, 'steps': 4117, 'loss/train': 1.8979488015174866} 01/28/2022 18:11:06 - INFO - codeparrot_training - Step 4118: {'lr': 0.0004976018098016559, 'samples': 790848, 'steps': 4118, 'loss/train': 1.8572142720222473} 01/28/2022 18:11:12 - INFO - codeparrot_training - Step 4119: {'lr': 0.0004975995483189153, 'samples': 791040, 'steps': 4119, 'loss/train': 2.231271743774414} 01/28/2022 18:11:17 - INFO - codeparrot_training - Step 4120: {'lr': 0.0004975972857755368, 'samples': 791232, 'steps': 4120, 'loss/train': 0.8286586403846741} 01/28/2022 18:11:21 - INFO - codeparrot_training - Step 4121: {'lr': 0.0004975950221715302, 'samples': 791424, 'steps': 4121, 'loss/train': 1.3393613398075104} 01/28/2022 18:11:25 - INFO - codeparrot_training - Step 4122: {'lr': 0.0004975927575069051, 'samples': 791616, 'steps': 4122, 'loss/train': 7.802510976791382} 01/28/2022 18:11:29 - INFO - codeparrot_training - Step 4123: {'lr': 0.0004975904917816713, 'samples': 791808, 'steps': 4123, 'loss/train': 1.9192208647727966} 01/28/2022 18:11:34 - INFO - codeparrot_training - Step 4124: {'lr': 0.0004975882249958385, 'samples': 792000, 'steps': 4124, 'loss/train': 2.311534881591797} 01/28/2022 18:11:39 - INFO - codeparrot_training - Step 4125: {'lr': 0.0004975859571494162, 'samples': 792192, 'steps': 4125, 'loss/train': 1.5334534049034119} 01/28/2022 18:11:43 - INFO - codeparrot_training - Step 4126: {'lr': 0.0004975836882424143, 'samples': 792384, 'steps': 4126, 'loss/train': 1.3952805697917938} 01/28/2022 18:11:47 - INFO - codeparrot_training - Step 4127: {'lr': 0.0004975814182748426, 'samples': 792576, 'steps': 4127, 'loss/train': 1.666450560092926} 01/28/2022 18:11:51 - INFO - codeparrot_training - Step 4128: {'lr': 0.0004975791472467108, 'samples': 792768, 'steps': 4128, 'loss/train': 0.4891131520271301} 01/28/2022 18:11:57 - INFO - codeparrot_training - Step 4129: {'lr': 0.0004975768751580283, 'samples': 792960, 'steps': 4129, 'loss/train': 2.0238837003707886} 01/28/2022 18:12:01 - INFO - codeparrot_training - Step 4130: {'lr': 0.0004975746020088052, 'samples': 793152, 'steps': 4130, 'loss/train': 2.0302605628967285} 01/28/2022 18:12:05 - INFO - codeparrot_training - Step 4131: {'lr': 0.0004975723277990512, 'samples': 793344, 'steps': 4131, 'loss/train': 1.554624617099762} 01/28/2022 18:12:09 - INFO - codeparrot_training - Step 4132: {'lr': 0.0004975700525287758, 'samples': 793536, 'steps': 4132, 'loss/train': 2.205920398235321} 01/28/2022 18:12:14 - INFO - codeparrot_training - Step 4133: {'lr': 0.0004975677761979891, 'samples': 793728, 'steps': 4133, 'loss/train': 2.4972869753837585} 01/28/2022 18:12:20 - INFO - codeparrot_training - Step 4134: {'lr': 0.0004975654988067005, 'samples': 793920, 'steps': 4134, 'loss/train': 2.367594301700592} 01/28/2022 18:12:24 - INFO - codeparrot_training - Step 4135: {'lr': 0.00049756322035492, 'samples': 794112, 'steps': 4135, 'loss/train': 1.6503864526748657} 01/28/2022 18:12:28 - INFO - codeparrot_training - Step 4136: {'lr': 0.0004975609408426572, 'samples': 794304, 'steps': 4136, 'loss/train': 1.4519838094711304} 01/28/2022 18:12:33 - INFO - codeparrot_training - Step 4137: {'lr': 0.000497558660269922, 'samples': 794496, 'steps': 4137, 'loss/train': 2.0689428448677063} 01/28/2022 18:12:37 - INFO - codeparrot_training - Step 4138: {'lr': 0.0004975563786367241, 'samples': 794688, 'steps': 4138, 'loss/train': 2.256630778312683} 01/28/2022 18:12:42 - INFO - codeparrot_training - Step 4139: {'lr': 0.0004975540959430732, 'samples': 794880, 'steps': 4139, 'loss/train': 1.469999760389328} 01/28/2022 18:12:46 - INFO - codeparrot_training - Step 4140: {'lr': 0.0004975518121889793, 'samples': 795072, 'steps': 4140, 'loss/train': 1.9348202347755432} 01/28/2022 18:12:51 - INFO - codeparrot_training - Step 4141: {'lr': 0.000497549527374452, 'samples': 795264, 'steps': 4141, 'loss/train': 1.1433814465999603} 01/28/2022 18:12:55 - INFO - codeparrot_training - Step 4142: {'lr': 0.000497547241499501, 'samples': 795456, 'steps': 4142, 'loss/train': 1.9876511693000793} 01/28/2022 18:12:59 - INFO - codeparrot_training - Step 4143: {'lr': 0.0004975449545641364, 'samples': 795648, 'steps': 4143, 'loss/train': 1.2126650512218475} 01/28/2022 18:13:05 - INFO - codeparrot_training - Step 4144: {'lr': 0.0004975426665683678, 'samples': 795840, 'steps': 4144, 'loss/train': 2.8557230830192566} 01/28/2022 18:13:09 - INFO - codeparrot_training - Step 4145: {'lr': 0.000497540377512205, 'samples': 796032, 'steps': 4145, 'loss/train': 2.0783565044403076} 01/28/2022 18:13:14 - INFO - codeparrot_training - Step 4146: {'lr': 0.0004975380873956577, 'samples': 796224, 'steps': 4146, 'loss/train': 1.8175835609436035} 01/28/2022 18:13:18 - INFO - codeparrot_training - Step 4147: {'lr': 0.0004975357962187359, 'samples': 796416, 'steps': 4147, 'loss/train': 2.4155609607696533} 01/28/2022 18:13:22 - INFO - codeparrot_training - Step 4148: {'lr': 0.0004975335039814493, 'samples': 796608, 'steps': 4148, 'loss/train': 2.0906794667243958} 01/28/2022 18:13:27 - INFO - codeparrot_training - Step 4149: {'lr': 0.0004975312106838079, 'samples': 796800, 'steps': 4149, 'loss/train': 2.0748928785324097} 01/28/2022 18:13:32 - INFO - codeparrot_training - Step 4150: {'lr': 0.0004975289163258214, 'samples': 796992, 'steps': 4150, 'loss/train': 2.051795482635498} 01/28/2022 18:13:36 - INFO - codeparrot_training - Step 4151: {'lr': 0.0004975266209074995, 'samples': 797184, 'steps': 4151, 'loss/train': 1.7643926739692688} 01/28/2022 18:13:40 - INFO - codeparrot_training - Step 4152: {'lr': 0.0004975243244288522, 'samples': 797376, 'steps': 4152, 'loss/train': 0.9418076276779175} 01/28/2022 18:13:44 - INFO - codeparrot_training - Step 4153: {'lr': 0.0004975220268898893, 'samples': 797568, 'steps': 4153, 'loss/train': 1.8470179438591003} 01/28/2022 18:13:50 - INFO - codeparrot_training - Step 4154: {'lr': 0.0004975197282906207, 'samples': 797760, 'steps': 4154, 'loss/train': 1.0949116051197052} 01/28/2022 18:13:54 - INFO - codeparrot_training - Step 4155: {'lr': 0.0004975174286310562, 'samples': 797952, 'steps': 4155, 'loss/train': 2.0704652667045593} 01/28/2022 18:13:58 - INFO - codeparrot_training - Step 4156: {'lr': 0.0004975151279112054, 'samples': 798144, 'steps': 4156, 'loss/train': 0.276897557079792} 01/28/2022 18:14:03 - INFO - codeparrot_training - Step 4157: {'lr': 0.0004975128261310787, 'samples': 798336, 'steps': 4157, 'loss/train': 2.554901361465454} 01/28/2022 18:14:07 - INFO - codeparrot_training - Step 4158: {'lr': 0.0004975105232906854, 'samples': 798528, 'steps': 4158, 'loss/train': 0.8907318413257599} 01/28/2022 18:14:12 - INFO - codeparrot_training - Step 4159: {'lr': 0.0004975082193900357, 'samples': 798720, 'steps': 4159, 'loss/train': 1.8916479349136353} 01/28/2022 18:14:16 - INFO - codeparrot_training - Step 4160: {'lr': 0.0004975059144291394, 'samples': 798912, 'steps': 4160, 'loss/train': 2.5337968468666077} 01/28/2022 18:14:21 - INFO - codeparrot_training - Step 4161: {'lr': 0.0004975036084080063, 'samples': 799104, 'steps': 4161, 'loss/train': 1.6110464930534363} 01/28/2022 18:14:25 - INFO - codeparrot_training - Step 4162: {'lr': 0.0004975013013266464, 'samples': 799296, 'steps': 4162, 'loss/train': 2.010028123855591} 01/28/2022 18:14:29 - INFO - codeparrot_training - Step 4163: {'lr': 0.0004974989931850695, 'samples': 799488, 'steps': 4163, 'loss/train': 1.7518449425697327} 01/28/2022 18:14:35 - INFO - codeparrot_training - Step 4164: {'lr': 0.0004974966839832855, 'samples': 799680, 'steps': 4164, 'loss/train': 1.2096283435821533} 01/28/2022 18:14:39 - INFO - codeparrot_training - Step 4165: {'lr': 0.0004974943737213042, 'samples': 799872, 'steps': 4165, 'loss/train': 1.535805881023407} 01/28/2022 18:14:44 - INFO - codeparrot_training - Step 4166: {'lr': 0.0004974920623991356, 'samples': 800064, 'steps': 4166, 'loss/train': 1.5770663022994995} 01/28/2022 18:14:48 - INFO - codeparrot_training - Step 4167: {'lr': 0.0004974897500167898, 'samples': 800256, 'steps': 4167, 'loss/train': 0.9402774274349213} 01/28/2022 18:14:52 - INFO - codeparrot_training - Step 4168: {'lr': 0.0004974874365742763, 'samples': 800448, 'steps': 4168, 'loss/train': 2.2418843507766724} 01/28/2022 18:14:57 - INFO - codeparrot_training - Step 4169: {'lr': 0.0004974851220716053, 'samples': 800640, 'steps': 4169, 'loss/train': 2.570425808429718} 01/28/2022 18:15:02 - INFO - codeparrot_training - Step 4170: {'lr': 0.0004974828065087867, 'samples': 800832, 'steps': 4170, 'loss/train': 1.064268797636032} 01/28/2022 18:15:06 - INFO - codeparrot_training - Step 4171: {'lr': 0.0004974804898858302, 'samples': 801024, 'steps': 4171, 'loss/train': 1.8046842813491821} 01/28/2022 18:15:10 - INFO - codeparrot_training - Step 4172: {'lr': 0.0004974781722027459, 'samples': 801216, 'steps': 4172, 'loss/train': 0.8525471091270447} 01/28/2022 18:15:14 - INFO - codeparrot_training - Step 4173: {'lr': 0.0004974758534595436, 'samples': 801408, 'steps': 4173, 'loss/train': 1.794366180896759} 01/28/2022 18:15:20 - INFO - codeparrot_training - Step 4174: {'lr': 0.0004974735336562335, 'samples': 801600, 'steps': 4174, 'loss/train': 1.788723886013031} 01/28/2022 18:15:24 - INFO - codeparrot_training - Step 4175: {'lr': 0.0004974712127928252, 'samples': 801792, 'steps': 4175, 'loss/train': 2.7301247119903564} 01/28/2022 18:15:28 - INFO - codeparrot_training - Step 4176: {'lr': 0.000497468890869329, 'samples': 801984, 'steps': 4176, 'loss/train': 0.7005524933338165} 01/28/2022 18:15:33 - INFO - codeparrot_training - Step 4177: {'lr': 0.0004974665678857545, 'samples': 802176, 'steps': 4177, 'loss/train': 2.1072344183921814} 01/28/2022 18:15:37 - INFO - codeparrot_training - Step 4178: {'lr': 0.0004974642438421118, 'samples': 802368, 'steps': 4178, 'loss/train': 2.1286447048187256} 01/28/2022 18:15:43 - INFO - codeparrot_training - Step 4179: {'lr': 0.0004974619187384109, 'samples': 802560, 'steps': 4179, 'loss/train': 2.077796995639801} 01/28/2022 18:15:47 - INFO - codeparrot_training - Step 4180: {'lr': 0.0004974595925746618, 'samples': 802752, 'steps': 4180, 'loss/train': 1.976605474948883} 01/28/2022 18:15:51 - INFO - codeparrot_training - Step 4181: {'lr': 0.0004974572653508742, 'samples': 802944, 'steps': 4181, 'loss/train': 1.999528706073761} 01/28/2022 18:15:55 - INFO - codeparrot_training - Step 4182: {'lr': 0.0004974549370670584, 'samples': 803136, 'steps': 4182, 'loss/train': 1.4221134781837463} 01/28/2022 18:16:00 - INFO - codeparrot_training - Step 4183: {'lr': 0.0004974526077232242, 'samples': 803328, 'steps': 4183, 'loss/train': 2.168429911136627} 01/28/2022 18:16:05 - INFO - codeparrot_training - Step 4184: {'lr': 0.0004974502773193815, 'samples': 803520, 'steps': 4184, 'loss/train': 1.90384179353714} 01/28/2022 18:16:09 - INFO - codeparrot_training - Step 4185: {'lr': 0.0004974479458555405, 'samples': 803712, 'steps': 4185, 'loss/train': 1.9623318314552307} 01/28/2022 18:16:14 - INFO - codeparrot_training - Step 4186: {'lr': 0.000497445613331711, 'samples': 803904, 'steps': 4186, 'loss/train': 2.33586984872818} 01/28/2022 18:16:18 - INFO - codeparrot_training - Step 4187: {'lr': 0.0004974432797479032, 'samples': 804096, 'steps': 4187, 'loss/train': 2.098836600780487} 01/28/2022 18:16:22 - INFO - codeparrot_training - Step 4188: {'lr': 0.0004974409451041268, 'samples': 804288, 'steps': 4188, 'loss/train': 2.074431896209717} 01/28/2022 18:16:28 - INFO - codeparrot_training - Step 4189: {'lr': 0.0004974386094003921, 'samples': 804480, 'steps': 4189, 'loss/train': 1.703148365020752} 01/28/2022 18:16:32 - INFO - codeparrot_training - Step 4190: {'lr': 0.0004974362726367089, 'samples': 804672, 'steps': 4190, 'loss/train': 2.193408966064453} 01/28/2022 18:16:37 - INFO - codeparrot_training - Step 4191: {'lr': 0.0004974339348130873, 'samples': 804864, 'steps': 4191, 'loss/train': 1.7612364292144775} 01/28/2022 18:16:41 - INFO - codeparrot_training - Step 4192: {'lr': 0.0004974315959295373, 'samples': 805056, 'steps': 4192, 'loss/train': 1.860961139202118} 01/28/2022 18:16:45 - INFO - codeparrot_training - Step 4193: {'lr': 0.0004974292559860688, 'samples': 805248, 'steps': 4193, 'loss/train': 2.0324177742004395} 01/28/2022 18:16:51 - INFO - codeparrot_training - Step 4194: {'lr': 0.0004974269149826921, 'samples': 805440, 'steps': 4194, 'loss/train': 1.9896923303604126} 01/28/2022 18:16:56 - INFO - codeparrot_training - Step 4195: {'lr': 0.0004974245729194169, 'samples': 805632, 'steps': 4195, 'loss/train': 1.6302164196968079} 01/28/2022 18:17:00 - INFO - codeparrot_training - Step 4196: {'lr': 0.0004974222297962535, 'samples': 805824, 'steps': 4196, 'loss/train': 2.2475180625915527} 01/28/2022 18:17:04 - INFO - codeparrot_training - Step 4197: {'lr': 0.0004974198856132118, 'samples': 806016, 'steps': 4197, 'loss/train': 0.1939864084124565} 01/28/2022 18:17:10 - INFO - codeparrot_training - Step 4198: {'lr': 0.0004974175403703019, 'samples': 806208, 'steps': 4198, 'loss/train': 1.3878464698791504} 01/28/2022 18:17:14 - INFO - codeparrot_training - Step 4199: {'lr': 0.0004974151940675338, 'samples': 806400, 'steps': 4199, 'loss/train': 1.014393836259842} 01/28/2022 18:17:18 - INFO - codeparrot_training - Step 4200: {'lr': 0.0004974128467049176, 'samples': 806592, 'steps': 4200, 'loss/train': 1.7710243463516235} 01/28/2022 18:17:22 - INFO - codeparrot_training - Step 4201: {'lr': 0.0004974104982824632, 'samples': 806784, 'steps': 4201, 'loss/train': 1.9142764806747437} 01/28/2022 18:17:27 - INFO - codeparrot_training - Step 4202: {'lr': 0.0004974081488001809, 'samples': 806976, 'steps': 4202, 'loss/train': 1.815666675567627} 01/28/2022 18:17:32 - INFO - codeparrot_training - Step 4203: {'lr': 0.0004974057982580806, 'samples': 807168, 'steps': 4203, 'loss/train': 2.199504017829895} 01/28/2022 18:17:36 - INFO - codeparrot_training - Step 4204: {'lr': 0.0004974034466561725, 'samples': 807360, 'steps': 4204, 'loss/train': 1.8956011533737183} 01/28/2022 18:17:40 - INFO - codeparrot_training - Step 4205: {'lr': 0.0004974010939944667, 'samples': 807552, 'steps': 4205, 'loss/train': 1.8378204703330994} 01/28/2022 18:17:45 - INFO - codeparrot_training - Step 4206: {'lr': 0.0004973987402729729, 'samples': 807744, 'steps': 4206, 'loss/train': 2.1381195187568665} 01/28/2022 18:17:49 - INFO - codeparrot_training - Step 4207: {'lr': 0.0004973963854917016, 'samples': 807936, 'steps': 4207, 'loss/train': 1.0993779301643372} 01/28/2022 18:17:55 - INFO - codeparrot_training - Step 4208: {'lr': 0.0004973940296506627, 'samples': 808128, 'steps': 4208, 'loss/train': 2.3120062351226807} 01/28/2022 18:17:59 - INFO - codeparrot_training - Step 4209: {'lr': 0.0004973916727498664, 'samples': 808320, 'steps': 4209, 'loss/train': 1.6275474429130554} 01/28/2022 18:18:04 - INFO - codeparrot_training - Step 4210: {'lr': 0.0004973893147893227, 'samples': 808512, 'steps': 4210, 'loss/train': 1.6839330196380615} 01/28/2022 18:18:08 - INFO - codeparrot_training - Step 4211: {'lr': 0.0004973869557690417, 'samples': 808704, 'steps': 4211, 'loss/train': 0.4404802769422531} 01/28/2022 18:18:12 - INFO - codeparrot_training - Step 4212: {'lr': 0.0004973845956890336, 'samples': 808896, 'steps': 4212, 'loss/train': 1.772487759590149} 01/28/2022 18:18:17 - INFO - codeparrot_training - Step 4213: {'lr': 0.0004973822345493084, 'samples': 809088, 'steps': 4213, 'loss/train': 1.8542472124099731} 01/28/2022 18:18:22 - INFO - codeparrot_training - Step 4214: {'lr': 0.0004973798723498762, 'samples': 809280, 'steps': 4214, 'loss/train': 1.6414126753807068} 01/28/2022 18:18:26 - INFO - codeparrot_training - Step 4215: {'lr': 0.0004973775090907473, 'samples': 809472, 'steps': 4215, 'loss/train': 2.3575326204299927} 01/28/2022 18:18:30 - INFO - codeparrot_training - Step 4216: {'lr': 0.0004973751447719316, 'samples': 809664, 'steps': 4216, 'loss/train': 1.2251670062541962} 01/28/2022 18:18:34 - INFO - codeparrot_training - Step 4217: {'lr': 0.0004973727793934394, 'samples': 809856, 'steps': 4217, 'loss/train': 2.3781697154045105} 01/28/2022 18:18:40 - INFO - codeparrot_training - Step 4218: {'lr': 0.0004973704129552808, 'samples': 810048, 'steps': 4218, 'loss/train': 1.0531780421733856} 01/28/2022 18:18:44 - INFO - codeparrot_training - Step 4219: {'lr': 0.0004973680454574657, 'samples': 810240, 'steps': 4219, 'loss/train': 2.369584858417511} 01/28/2022 18:18:48 - INFO - codeparrot_training - Step 4220: {'lr': 0.0004973656769000046, 'samples': 810432, 'steps': 4220, 'loss/train': 1.734473168849945} 01/28/2022 18:18:52 - INFO - codeparrot_training - Step 4221: {'lr': 0.0004973633072829075, 'samples': 810624, 'steps': 4221, 'loss/train': 1.175313413143158} 01/28/2022 18:18:57 - INFO - codeparrot_training - Step 4222: {'lr': 0.0004973609366061845, 'samples': 810816, 'steps': 4222, 'loss/train': 2.2656952142715454} 01/28/2022 18:19:03 - INFO - codeparrot_training - Step 4223: {'lr': 0.0004973585648698457, 'samples': 811008, 'steps': 4223, 'loss/train': 2.3476245403289795} 01/28/2022 18:19:07 - INFO - codeparrot_training - Step 4224: {'lr': 0.0004973561920739015, 'samples': 811200, 'steps': 4224, 'loss/train': 1.7951910495758057} 01/28/2022 18:19:11 - INFO - codeparrot_training - Step 4225: {'lr': 0.0004973538182183618, 'samples': 811392, 'steps': 4225, 'loss/train': 1.6248770356178284} 01/28/2022 18:19:15 - INFO - codeparrot_training - Step 4226: {'lr': 0.000497351443303237, 'samples': 811584, 'steps': 4226, 'loss/train': 2.19542795419693} 01/28/2022 18:19:20 - INFO - codeparrot_training - Step 4227: {'lr': 0.0004973490673285372, 'samples': 811776, 'steps': 4227, 'loss/train': 0.3080916479229927} 01/28/2022 18:19:25 - INFO - codeparrot_training - Step 4228: {'lr': 0.0004973466902942723, 'samples': 811968, 'steps': 4228, 'loss/train': 2.0282903909683228} 01/28/2022 18:19:29 - INFO - codeparrot_training - Step 4229: {'lr': 0.0004973443122004529, 'samples': 812160, 'steps': 4229, 'loss/train': 1.3778345882892609} 01/28/2022 18:19:33 - INFO - codeparrot_training - Step 4230: {'lr': 0.0004973419330470891, 'samples': 812352, 'steps': 4230, 'loss/train': 1.5410040020942688} 01/28/2022 18:19:38 - INFO - codeparrot_training - Step 4231: {'lr': 0.0004973395528341908, 'samples': 812544, 'steps': 4231, 'loss/train': 2.4849547147750854} 01/28/2022 18:19:42 - INFO - codeparrot_training - Step 4232: {'lr': 0.0004973371715617685, 'samples': 812736, 'steps': 4232, 'loss/train': 1.190185010433197} 01/28/2022 18:19:48 - INFO - codeparrot_training - Step 4233: {'lr': 0.0004973347892298322, 'samples': 812928, 'steps': 4233, 'loss/train': 1.774814486503601} 01/28/2022 18:19:52 - INFO - codeparrot_training - Step 4234: {'lr': 0.0004973324058383924, 'samples': 813120, 'steps': 4234, 'loss/train': 1.2775864005088806} 01/28/2022 18:19:56 - INFO - codeparrot_training - Step 4235: {'lr': 0.0004973300213874589, 'samples': 813312, 'steps': 4235, 'loss/train': 1.6862043142318726} 01/28/2022 18:20:01 - INFO - codeparrot_training - Step 4236: {'lr': 0.0004973276358770422, 'samples': 813504, 'steps': 4236, 'loss/train': 1.7243582010269165} 01/28/2022 18:20:05 - INFO - codeparrot_training - Step 4237: {'lr': 0.0004973252493071525, 'samples': 813696, 'steps': 4237, 'loss/train': 0.7972845733165741} 01/28/2022 18:20:10 - INFO - codeparrot_training - Step 4238: {'lr': 0.0004973228616777999, 'samples': 813888, 'steps': 4238, 'loss/train': 2.426549971103668} 01/28/2022 18:20:14 - INFO - codeparrot_training - Step 4239: {'lr': 0.0004973204729889946, 'samples': 814080, 'steps': 4239, 'loss/train': 1.5684394240379333} 01/28/2022 18:20:18 - INFO - codeparrot_training - Step 4240: {'lr': 0.0004973180832407472, 'samples': 814272, 'steps': 4240, 'loss/train': 1.8019772171974182} 01/28/2022 18:20:23 - INFO - codeparrot_training - Step 4241: {'lr': 0.0004973156924330674, 'samples': 814464, 'steps': 4241, 'loss/train': 2.962576389312744} 01/28/2022 18:20:27 - INFO - codeparrot_training - Step 4242: {'lr': 0.0004973133005659658, 'samples': 814656, 'steps': 4242, 'loss/train': 1.8073191046714783} 01/28/2022 18:20:32 - INFO - codeparrot_training - Step 4243: {'lr': 0.0004973109076394526, 'samples': 814848, 'steps': 4243, 'loss/train': 1.4314425587654114} 01/28/2022 18:20:37 - INFO - codeparrot_training - Step 4244: {'lr': 0.0004973085136535379, 'samples': 815040, 'steps': 4244, 'loss/train': 1.2575687170028687} 01/28/2022 18:20:41 - INFO - codeparrot_training - Step 4245: {'lr': 0.000497306118608232, 'samples': 815232, 'steps': 4245, 'loss/train': 0.5393557548522949} 01/28/2022 18:20:45 - INFO - codeparrot_training - Step 4246: {'lr': 0.0004973037225035454, 'samples': 815424, 'steps': 4246, 'loss/train': 1.5122923851013184} 01/28/2022 18:20:49 - INFO - codeparrot_training - Step 4247: {'lr': 0.0004973013253394881, 'samples': 815616, 'steps': 4247, 'loss/train': 2.422505021095276} 01/28/2022 18:20:55 - INFO - codeparrot_training - Step 4248: {'lr': 0.0004972989271160705, 'samples': 815808, 'steps': 4248, 'loss/train': 1.4713726043701172} 01/28/2022 18:21:00 - INFO - codeparrot_training - Step 4249: {'lr': 0.0004972965278333028, 'samples': 816000, 'steps': 4249, 'loss/train': 1.9325870275497437} 01/28/2022 18:21:04 - INFO - codeparrot_training - Step 4250: {'lr': 0.0004972941274911952, 'samples': 816192, 'steps': 4250, 'loss/train': 1.186111718416214} 01/28/2022 18:21:08 - INFO - codeparrot_training - Step 4251: {'lr': 0.0004972917260897583, 'samples': 816384, 'steps': 4251, 'loss/train': 1.880176305770874} 01/28/2022 18:21:12 - INFO - codeparrot_training - Step 4252: {'lr': 0.0004972893236290019, 'samples': 816576, 'steps': 4252, 'loss/train': 2.220681667327881} 01/28/2022 18:21:18 - INFO - codeparrot_training - Step 4253: {'lr': 0.0004972869201089367, 'samples': 816768, 'steps': 4253, 'loss/train': 2.149201512336731} 01/28/2022 18:21:22 - INFO - codeparrot_training - Step 4254: {'lr': 0.0004972845155295729, 'samples': 816960, 'steps': 4254, 'loss/train': 0.3672722429037094} 01/28/2022 18:21:26 - INFO - codeparrot_training - Step 4255: {'lr': 0.0004972821098909207, 'samples': 817152, 'steps': 4255, 'loss/train': 1.9639427661895752} 01/28/2022 18:21:30 - INFO - codeparrot_training - Step 4256: {'lr': 0.0004972797031929904, 'samples': 817344, 'steps': 4256, 'loss/train': 1.356442004442215} 01/28/2022 18:21:34 - INFO - codeparrot_training - Step 4257: {'lr': 0.0004972772954357924, 'samples': 817536, 'steps': 4257, 'loss/train': 1.7277284860610962} 01/28/2022 18:21:40 - INFO - codeparrot_training - Step 4258: {'lr': 0.0004972748866193371, 'samples': 817728, 'steps': 4258, 'loss/train': 1.3503383696079254} 01/28/2022 18:21:44 - INFO - codeparrot_training - Step 4259: {'lr': 0.0004972724767436346, 'samples': 817920, 'steps': 4259, 'loss/train': 1.016992449760437} 01/28/2022 18:21:48 - INFO - codeparrot_training - Step 4260: {'lr': 0.0004972700658086954, 'samples': 818112, 'steps': 4260, 'loss/train': 1.0673084557056427} 01/28/2022 18:21:53 - INFO - codeparrot_training - Step 4261: {'lr': 0.0004972676538145298, 'samples': 818304, 'steps': 4261, 'loss/train': 1.563287079334259} 01/28/2022 18:21:57 - INFO - codeparrot_training - Step 4262: {'lr': 0.0004972652407611479, 'samples': 818496, 'steps': 4262, 'loss/train': 1.6412909030914307} 01/28/2022 18:22:02 - INFO - codeparrot_training - Step 4263: {'lr': 0.0004972628266485604, 'samples': 818688, 'steps': 4263, 'loss/train': 1.454592525959015} 01/28/2022 18:22:07 - INFO - codeparrot_training - Step 4264: {'lr': 0.0004972604114767774, 'samples': 818880, 'steps': 4264, 'loss/train': 0.8945094645023346} 01/28/2022 18:22:11 - INFO - codeparrot_training - Step 4265: {'lr': 0.0004972579952458092, 'samples': 819072, 'steps': 4265, 'loss/train': 1.5085694789886475} 01/28/2022 18:22:15 - INFO - codeparrot_training - Step 4266: {'lr': 0.0004972555779556664, 'samples': 819264, 'steps': 4266, 'loss/train': 1.8131094574928284} 01/28/2022 18:22:19 - INFO - codeparrot_training - Step 4267: {'lr': 0.0004972531596063592, 'samples': 819456, 'steps': 4267, 'loss/train': 0.47457951307296753} 01/28/2022 18:22:25 - INFO - codeparrot_training - Step 4268: {'lr': 0.000497250740197898, 'samples': 819648, 'steps': 4268, 'loss/train': 1.1849910020828247} 01/28/2022 18:22:29 - INFO - codeparrot_training - Step 4269: {'lr': 0.0004972483197302931, 'samples': 819840, 'steps': 4269, 'loss/train': 2.3531309366226196} 01/28/2022 18:22:34 - INFO - codeparrot_training - Step 4270: {'lr': 0.0004972458982035548, 'samples': 820032, 'steps': 4270, 'loss/train': 2.2915897965431213} 01/28/2022 18:22:38 - INFO - codeparrot_training - Step 4271: {'lr': 0.0004972434756176937, 'samples': 820224, 'steps': 4271, 'loss/train': 1.9941027760505676} 01/28/2022 18:22:42 - INFO - codeparrot_training - Step 4272: {'lr': 0.0004972410519727201, 'samples': 820416, 'steps': 4272, 'loss/train': 1.5515252351760864} 01/28/2022 18:22:47 - INFO - codeparrot_training - Step 4273: {'lr': 0.0004972386272686443, 'samples': 820608, 'steps': 4273, 'loss/train': 1.397973507642746} 01/28/2022 18:22:52 - INFO - codeparrot_training - Step 4274: {'lr': 0.0004972362015054767, 'samples': 820800, 'steps': 4274, 'loss/train': 2.3193780183792114} 01/28/2022 18:22:56 - INFO - codeparrot_training - Step 4275: {'lr': 0.0004972337746832278, 'samples': 820992, 'steps': 4275, 'loss/train': 0.4514671415090561} 01/28/2022 18:23:00 - INFO - codeparrot_training - Step 4276: {'lr': 0.0004972313468019077, 'samples': 821184, 'steps': 4276, 'loss/train': 2.467438817024231} 01/28/2022 18:23:04 - INFO - codeparrot_training - Step 4277: {'lr': 0.0004972289178615273, 'samples': 821376, 'steps': 4277, 'loss/train': 1.9827754497528076} 01/28/2022 18:23:10 - INFO - codeparrot_training - Step 4278: {'lr': 0.0004972264878620965, 'samples': 821568, 'steps': 4278, 'loss/train': 2.0124839544296265} 01/28/2022 18:23:14 - INFO - codeparrot_training - Step 4279: {'lr': 0.000497224056803626, 'samples': 821760, 'steps': 4279, 'loss/train': 1.97199547290802} 01/28/2022 18:23:18 - INFO - codeparrot_training - Step 4280: {'lr': 0.0004972216246861262, 'samples': 821952, 'steps': 4280, 'loss/train': 2.2137503027915955} 01/28/2022 18:23:23 - INFO - codeparrot_training - Step 4281: {'lr': 0.0004972191915096074, 'samples': 822144, 'steps': 4281, 'loss/train': 1.8611117005348206} 01/28/2022 18:23:27 - INFO - codeparrot_training - Step 4282: {'lr': 0.0004972167572740801, 'samples': 822336, 'steps': 4282, 'loss/train': 1.9306930303573608} 01/28/2022 18:23:33 - INFO - codeparrot_training - Step 4283: {'lr': 0.0004972143219795547, 'samples': 822528, 'steps': 4283, 'loss/train': 1.4302887618541718} 01/28/2022 18:23:37 - INFO - codeparrot_training - Step 4284: {'lr': 0.0004972118856260416, 'samples': 822720, 'steps': 4284, 'loss/train': 1.4846226274967194} 01/28/2022 18:23:41 - INFO - codeparrot_training - Step 4285: {'lr': 0.0004972094482135514, 'samples': 822912, 'steps': 4285, 'loss/train': 2.255715787410736} 01/28/2022 18:23:46 - INFO - codeparrot_training - Step 4286: {'lr': 0.0004972070097420943, 'samples': 823104, 'steps': 4286, 'loss/train': 1.6899470686912537} 01/28/2022 18:23:50 - INFO - codeparrot_training - Step 4287: {'lr': 0.0004972045702116809, 'samples': 823296, 'steps': 4287, 'loss/train': 1.6604109406471252} 01/28/2022 18:23:55 - INFO - codeparrot_training - Step 4288: {'lr': 0.0004972021296223217, 'samples': 823488, 'steps': 4288, 'loss/train': 1.617037832736969} 01/28/2022 18:24:00 - INFO - codeparrot_training - Step 4289: {'lr': 0.0004971996879740271, 'samples': 823680, 'steps': 4289, 'loss/train': 1.3717956840991974} 01/28/2022 18:24:04 - INFO - codeparrot_training - Step 4290: {'lr': 0.0004971972452668074, 'samples': 823872, 'steps': 4290, 'loss/train': 2.1996066570281982} 01/28/2022 18:24:08 - INFO - codeparrot_training - Step 4291: {'lr': 0.0004971948015006732, 'samples': 824064, 'steps': 4291, 'loss/train': 1.8753413558006287} 01/28/2022 18:24:12 - INFO - codeparrot_training - Step 4292: {'lr': 0.000497192356675635, 'samples': 824256, 'steps': 4292, 'loss/train': 2.42938631772995} 01/28/2022 18:24:18 - INFO - codeparrot_training - Step 4293: {'lr': 0.0004971899107917033, 'samples': 824448, 'steps': 4293, 'loss/train': 2.33437442779541} 01/28/2022 18:24:23 - INFO - codeparrot_training - Step 4294: {'lr': 0.0004971874638488884, 'samples': 824640, 'steps': 4294, 'loss/train': 2.200483024120331} 01/28/2022 18:24:27 - INFO - codeparrot_training - Step 4295: {'lr': 0.000497185015847201, 'samples': 824832, 'steps': 4295, 'loss/train': 1.9701105952262878} 01/28/2022 18:24:31 - INFO - codeparrot_training - Step 4296: {'lr': 0.0004971825667866515, 'samples': 825024, 'steps': 4296, 'loss/train': 2.4606302976608276} 01/28/2022 18:24:35 - INFO - codeparrot_training - Step 4297: {'lr': 0.0004971801166672502, 'samples': 825216, 'steps': 4297, 'loss/train': 1.666476845741272} 01/28/2022 18:24:40 - INFO - codeparrot_training - Step 4298: {'lr': 0.0004971776654890079, 'samples': 825408, 'steps': 4298, 'loss/train': 2.165467858314514} 01/28/2022 18:24:45 - INFO - codeparrot_training - Step 4299: {'lr': 0.000497175213251935, 'samples': 825600, 'steps': 4299, 'loss/train': 2.058509409427643} 01/28/2022 18:24:49 - INFO - codeparrot_training - Step 4300: {'lr': 0.0004971727599560418, 'samples': 825792, 'steps': 4300, 'loss/train': 2.0176252126693726} 01/28/2022 18:24:53 - INFO - codeparrot_training - Step 4301: {'lr': 0.0004971703056013392, 'samples': 825984, 'steps': 4301, 'loss/train': 2.0957826375961304} 01/28/2022 18:24:57 - INFO - codeparrot_training - Step 4302: {'lr': 0.0004971678501878374, 'samples': 826176, 'steps': 4302, 'loss/train': 2.684676468372345} 01/28/2022 18:25:03 - INFO - codeparrot_training - Step 4303: {'lr': 0.000497165393715547, 'samples': 826368, 'steps': 4303, 'loss/train': 2.2643508911132812} 01/28/2022 18:25:07 - INFO - codeparrot_training - Step 4304: {'lr': 0.0004971629361844785, 'samples': 826560, 'steps': 4304, 'loss/train': 1.3548289239406586} 01/28/2022 18:25:11 - INFO - codeparrot_training - Step 4305: {'lr': 0.0004971604775946425, 'samples': 826752, 'steps': 4305, 'loss/train': 2.2426196336746216} 01/28/2022 18:25:15 - INFO - codeparrot_training - Step 4306: {'lr': 0.0004971580179460495, 'samples': 826944, 'steps': 4306, 'loss/train': 1.3519046902656555} 01/28/2022 18:25:20 - INFO - codeparrot_training - Step 4307: {'lr': 0.0004971555572387101, 'samples': 827136, 'steps': 4307, 'loss/train': 1.5256998538970947} 01/28/2022 18:25:26 - INFO - codeparrot_training - Step 4308: {'lr': 0.0004971530954726346, 'samples': 827328, 'steps': 4308, 'loss/train': 1.1294580101966858} 01/28/2022 18:25:30 - INFO - codeparrot_training - Step 4309: {'lr': 0.0004971506326478339, 'samples': 827520, 'steps': 4309, 'loss/train': 1.8778917789459229} 01/28/2022 18:25:34 - INFO - codeparrot_training - Step 4310: {'lr': 0.0004971481687643184, 'samples': 827712, 'steps': 4310, 'loss/train': 2.1596744656562805} 01/28/2022 18:25:38 - INFO - codeparrot_training - Step 4311: {'lr': 0.0004971457038220984, 'samples': 827904, 'steps': 4311, 'loss/train': 2.1702716946601868} 01/28/2022 18:25:43 - INFO - codeparrot_training - Step 4312: {'lr': 0.0004971432378211849, 'samples': 828096, 'steps': 4312, 'loss/train': 2.396085798740387} 01/28/2022 18:25:49 - INFO - codeparrot_training - Step 4313: {'lr': 0.0004971407707615881, 'samples': 828288, 'steps': 4313, 'loss/train': 1.9805139899253845} 01/28/2022 18:25:53 - INFO - codeparrot_training - Step 4314: {'lr': 0.0004971383026433189, 'samples': 828480, 'steps': 4314, 'loss/train': 2.835638165473938} 01/28/2022 18:25:57 - INFO - codeparrot_training - Step 4315: {'lr': 0.0004971358334663875, 'samples': 828672, 'steps': 4315, 'loss/train': 1.889420986175537} 01/28/2022 18:26:02 - INFO - codeparrot_training - Step 4316: {'lr': 0.0004971333632308047, 'samples': 828864, 'steps': 4316, 'loss/train': 1.5242825746536255} 01/28/2022 18:26:06 - INFO - codeparrot_training - Step 4317: {'lr': 0.000497130891936581, 'samples': 829056, 'steps': 4317, 'loss/train': 2.317616879940033} 01/28/2022 18:26:11 - INFO - codeparrot_training - Step 4318: {'lr': 0.0004971284195837271, 'samples': 829248, 'steps': 4318, 'loss/train': 1.98831707239151} 01/28/2022 18:26:15 - INFO - codeparrot_training - Step 4319: {'lr': 0.0004971259461722536, 'samples': 829440, 'steps': 4319, 'loss/train': 2.1543622612953186} 01/28/2022 18:26:20 - INFO - codeparrot_training - Step 4320: {'lr': 0.0004971234717021708, 'samples': 829632, 'steps': 4320, 'loss/train': 1.182509958744049} 01/28/2022 18:26:24 - INFO - codeparrot_training - Step 4321: {'lr': 0.0004971209961734897, 'samples': 829824, 'steps': 4321, 'loss/train': 2.10719096660614} 01/28/2022 18:26:28 - INFO - codeparrot_training - Step 4322: {'lr': 0.0004971185195862207, 'samples': 830016, 'steps': 4322, 'loss/train': 2.145713210105896} 01/28/2022 18:26:35 - INFO - codeparrot_training - Step 4323: {'lr': 0.0004971160419403744, 'samples': 830208, 'steps': 4323, 'loss/train': 2.302400529384613} 01/28/2022 18:26:39 - INFO - codeparrot_training - Step 4324: {'lr': 0.0004971135632359614, 'samples': 830400, 'steps': 4324, 'loss/train': 2.3677300214767456} 01/28/2022 18:26:43 - INFO - codeparrot_training - Step 4325: {'lr': 0.0004971110834729925, 'samples': 830592, 'steps': 4325, 'loss/train': 2.1235504746437073} 01/28/2022 18:26:48 - INFO - codeparrot_training - Step 4326: {'lr': 0.0004971086026514781, 'samples': 830784, 'steps': 4326, 'loss/train': 2.3699238896369934} 01/28/2022 18:26:52 - INFO - codeparrot_training - Step 4327: {'lr': 0.0004971061207714289, 'samples': 830976, 'steps': 4327, 'loss/train': 1.419160634279251} 01/28/2022 18:26:56 - INFO - codeparrot_training - Step 4328: {'lr': 0.0004971036378328556, 'samples': 831168, 'steps': 4328, 'loss/train': 0.6711495816707611} 01/28/2022 18:27:01 - INFO - codeparrot_training - Step 4329: {'lr': 0.0004971011538357687, 'samples': 831360, 'steps': 4329, 'loss/train': 2.24004989862442} 01/28/2022 18:27:06 - INFO - codeparrot_training - Step 4330: {'lr': 0.000497098668780179, 'samples': 831552, 'steps': 4330, 'loss/train': 2.213118016719818} 01/28/2022 18:27:10 - INFO - codeparrot_training - Step 4331: {'lr': 0.000497096182666097, 'samples': 831744, 'steps': 4331, 'loss/train': 2.094253420829773} 01/28/2022 18:27:14 - INFO - codeparrot_training - Step 4332: {'lr': 0.0004970936954935334, 'samples': 831936, 'steps': 4332, 'loss/train': 2.1478264331817627} 01/28/2022 18:27:18 - INFO - codeparrot_training - Step 4333: {'lr': 0.0004970912072624989, 'samples': 832128, 'steps': 4333, 'loss/train': 1.5369000434875488} 01/28/2022 18:27:24 - INFO - codeparrot_training - Step 4334: {'lr': 0.0004970887179730041, 'samples': 832320, 'steps': 4334, 'loss/train': 0.7974211871623993} 01/28/2022 18:27:28 - INFO - codeparrot_training - Step 4335: {'lr': 0.0004970862276250599, 'samples': 832512, 'steps': 4335, 'loss/train': 2.1673545241355896} 01/28/2022 18:27:32 - INFO - codeparrot_training - Step 4336: {'lr': 0.0004970837362186766, 'samples': 832704, 'steps': 4336, 'loss/train': 1.9868159294128418} 01/28/2022 18:27:37 - INFO - codeparrot_training - Step 4337: {'lr': 0.0004970812437538649, 'samples': 832896, 'steps': 4337, 'loss/train': 1.8472681045532227} 01/28/2022 18:27:41 - INFO - codeparrot_training - Step 4338: {'lr': 0.0004970787502306357, 'samples': 833088, 'steps': 4338, 'loss/train': 1.1641145646572113} 01/28/2022 18:27:47 - INFO - codeparrot_training - Step 4339: {'lr': 0.0004970762556489996, 'samples': 833280, 'steps': 4339, 'loss/train': 9.177886962890625} 01/28/2022 18:27:51 - INFO - codeparrot_training - Step 4340: {'lr': 0.0004970737600089673, 'samples': 833472, 'steps': 4340, 'loss/train': 2.5624648332595825} 01/28/2022 18:27:55 - INFO - codeparrot_training - Step 4341: {'lr': 0.0004970712633105496, 'samples': 833664, 'steps': 4341, 'loss/train': 1.2503962218761444} 01/28/2022 18:28:00 - INFO - codeparrot_training - Step 4342: {'lr': 0.0004970687655537568, 'samples': 833856, 'steps': 4342, 'loss/train': 3.135720133781433} 01/28/2022 18:28:04 - INFO - codeparrot_training - Step 4343: {'lr': 0.0004970662667386, 'samples': 834048, 'steps': 4343, 'loss/train': 1.7642512321472168} 01/28/2022 18:28:09 - INFO - codeparrot_training - Step 4344: {'lr': 0.0004970637668650898, 'samples': 834240, 'steps': 4344, 'loss/train': 1.811410367488861} 01/28/2022 18:28:13 - INFO - codeparrot_training - Step 4345: {'lr': 0.0004970612659332368, 'samples': 834432, 'steps': 4345, 'loss/train': 1.3995953500270844} 01/28/2022 18:28:17 - INFO - codeparrot_training - Step 4346: {'lr': 0.0004970587639430518, 'samples': 834624, 'steps': 4346, 'loss/train': 2.4977332949638367} 01/28/2022 18:28:22 - INFO - codeparrot_training - Step 4347: {'lr': 0.0004970562608945455, 'samples': 834816, 'steps': 4347, 'loss/train': 1.8963763117790222} 01/28/2022 18:28:26 - INFO - codeparrot_training - Step 4348: {'lr': 0.0004970537567877286, 'samples': 835008, 'steps': 4348, 'loss/train': 2.1972888708114624} 01/28/2022 18:28:31 - INFO - codeparrot_training - Step 4349: {'lr': 0.000497051251622612, 'samples': 835200, 'steps': 4349, 'loss/train': 2.0055967569351196} 01/28/2022 18:28:36 - INFO - codeparrot_training - Step 4350: {'lr': 0.0004970487453992062, 'samples': 835392, 'steps': 4350, 'loss/train': 1.8927335143089294} 01/28/2022 18:28:40 - INFO - codeparrot_training - Step 4351: {'lr': 0.000497046238117522, 'samples': 835584, 'steps': 4351, 'loss/train': 0.8977977633476257} 01/28/2022 18:28:44 - INFO - codeparrot_training - Step 4352: {'lr': 0.0004970437297775702, 'samples': 835776, 'steps': 4352, 'loss/train': 0.9665324985980988} 01/28/2022 18:28:48 - INFO - codeparrot_training - Step 4353: {'lr': 0.0004970412203793614, 'samples': 835968, 'steps': 4353, 'loss/train': 2.094414532184601} 01/28/2022 18:28:55 - INFO - codeparrot_training - Step 4354: {'lr': 0.0004970387099229066, 'samples': 836160, 'steps': 4354, 'loss/train': 1.742137849330902} 01/28/2022 18:28:59 - INFO - codeparrot_training - Step 4355: {'lr': 0.0004970361984082163, 'samples': 836352, 'steps': 4355, 'loss/train': 1.888991117477417} 01/28/2022 18:29:03 - INFO - codeparrot_training - Step 4356: {'lr': 0.0004970336858353014, 'samples': 836544, 'steps': 4356, 'loss/train': 1.8831380009651184} 01/28/2022 18:29:08 - INFO - codeparrot_training - Step 4357: {'lr': 0.0004970311722041727, 'samples': 836736, 'steps': 4357, 'loss/train': 1.9244211316108704} 01/28/2022 18:29:12 - INFO - codeparrot_training - Step 4358: {'lr': 0.0004970286575148408, 'samples': 836928, 'steps': 4358, 'loss/train': 0.40095658600330353} 01/28/2022 18:29:18 - INFO - codeparrot_training - Step 4359: {'lr': 0.0004970261417673165, 'samples': 837120, 'steps': 4359, 'loss/train': 2.370512008666992} 01/28/2022 18:29:22 - INFO - codeparrot_training - Step 4360: {'lr': 0.0004970236249616109, 'samples': 837312, 'steps': 4360, 'loss/train': 1.7971587181091309} 01/28/2022 18:29:27 - INFO - codeparrot_training - Step 4361: {'lr': 0.0004970211070977344, 'samples': 837504, 'steps': 4361, 'loss/train': 1.5058518648147583} 01/28/2022 18:29:31 - INFO - codeparrot_training - Step 4362: {'lr': 0.0004970185881756979, 'samples': 837696, 'steps': 4362, 'loss/train': 2.1920161843299866} 01/28/2022 18:29:35 - INFO - codeparrot_training - Step 4363: {'lr': 0.0004970160681955121, 'samples': 837888, 'steps': 4363, 'loss/train': 0.9785981476306915} 01/28/2022 18:29:40 - INFO - codeparrot_training - Step 4364: {'lr': 0.0004970135471571881, 'samples': 838080, 'steps': 4364, 'loss/train': 1.9152293801307678} 01/28/2022 18:29:45 - INFO - codeparrot_training - Step 4365: {'lr': 0.0004970110250607364, 'samples': 838272, 'steps': 4365, 'loss/train': 0.2226792722940445} 01/28/2022 18:29:49 - INFO - codeparrot_training - Step 4366: {'lr': 0.000497008501906168, 'samples': 838464, 'steps': 4366, 'loss/train': 0.31972362846136093} 01/28/2022 18:29:53 - INFO - codeparrot_training - Step 4367: {'lr': 0.0004970059776934935, 'samples': 838656, 'steps': 4367, 'loss/train': 1.9075058698654175} 01/28/2022 18:29:57 - INFO - codeparrot_training - Step 4368: {'lr': 0.0004970034524227238, 'samples': 838848, 'steps': 4368, 'loss/train': 2.2584916949272156} 01/28/2022 18:30:04 - INFO - codeparrot_training - Step 4369: {'lr': 0.0004970009260938698, 'samples': 839040, 'steps': 4369, 'loss/train': 1.5408178567886353} 01/28/2022 18:30:08 - INFO - codeparrot_training - Step 4370: {'lr': 0.0004969983987069423, 'samples': 839232, 'steps': 4370, 'loss/train': 1.6089794039726257} 01/28/2022 18:30:12 - INFO - codeparrot_training - Step 4371: {'lr': 0.000496995870261952, 'samples': 839424, 'steps': 4371, 'loss/train': 1.9870385527610779} 01/28/2022 18:30:16 - INFO - codeparrot_training - Step 4372: {'lr': 0.0004969933407589098, 'samples': 839616, 'steps': 4372, 'loss/train': 1.5960686802864075} 01/28/2022 18:30:21 - INFO - codeparrot_training - Step 4373: {'lr': 0.0004969908101978267, 'samples': 839808, 'steps': 4373, 'loss/train': 2.4804006814956665} 01/28/2022 18:30:26 - INFO - codeparrot_training - Step 4374: {'lr': 0.0004969882785787133, 'samples': 840000, 'steps': 4374, 'loss/train': 0.534230500459671} 01/28/2022 18:30:30 - INFO - codeparrot_training - Step 4375: {'lr': 0.0004969857459015807, 'samples': 840192, 'steps': 4375, 'loss/train': 1.0790318548679352} 01/28/2022 18:30:35 - INFO - codeparrot_training - Step 4376: {'lr': 0.0004969832121664394, 'samples': 840384, 'steps': 4376, 'loss/train': 1.3785315155982971} 01/28/2022 18:30:39 - INFO - codeparrot_training - Step 4377: {'lr': 0.0004969806773733004, 'samples': 840576, 'steps': 4377, 'loss/train': 2.5786734223365784} 01/28/2022 18:30:43 - INFO - codeparrot_training - Step 4378: {'lr': 0.0004969781415221748, 'samples': 840768, 'steps': 4378, 'loss/train': 1.4646990895271301} 01/28/2022 18:30:49 - INFO - codeparrot_training - Step 4379: {'lr': 0.0004969756046130731, 'samples': 840960, 'steps': 4379, 'loss/train': 1.3043719232082367} 01/28/2022 18:30:53 - INFO - codeparrot_training - Step 4380: {'lr': 0.0004969730666460065, 'samples': 841152, 'steps': 4380, 'loss/train': 1.7680858969688416} 01/28/2022 18:30:57 - INFO - codeparrot_training - Step 4381: {'lr': 0.0004969705276209856, 'samples': 841344, 'steps': 4381, 'loss/train': 2.284079611301422} 01/28/2022 18:31:02 - INFO - codeparrot_training - Step 4382: {'lr': 0.0004969679875380214, 'samples': 841536, 'steps': 4382, 'loss/train': 3.348679304122925} 01/28/2022 18:31:06 - INFO - codeparrot_training - Step 4383: {'lr': 0.0004969654463971247, 'samples': 841728, 'steps': 4383, 'loss/train': 1.06427863240242} 01/28/2022 18:31:11 - INFO - codeparrot_training - Step 4384: {'lr': 0.0004969629041983065, 'samples': 841920, 'steps': 4384, 'loss/train': 1.6232367753982544} 01/28/2022 18:31:16 - INFO - codeparrot_training - Step 4385: {'lr': 0.0004969603609415777, 'samples': 842112, 'steps': 4385, 'loss/train': 2.0846341252326965} 01/28/2022 18:31:20 - INFO - codeparrot_training - Step 4386: {'lr': 0.000496957816626949, 'samples': 842304, 'steps': 4386, 'loss/train': 2.136211931705475} 01/28/2022 18:31:24 - INFO - codeparrot_training - Step 4387: {'lr': 0.0004969552712544316, 'samples': 842496, 'steps': 4387, 'loss/train': 3.0306705236434937} 01/28/2022 18:31:28 - INFO - codeparrot_training - Step 4388: {'lr': 0.0004969527248240361, 'samples': 842688, 'steps': 4388, 'loss/train': 0.904362827539444} 01/28/2022 18:31:33 - INFO - codeparrot_training - Step 4389: {'lr': 0.0004969501773357736, 'samples': 842880, 'steps': 4389, 'loss/train': 2.395689010620117} 01/28/2022 18:31:38 - INFO - codeparrot_training - Step 4390: {'lr': 0.000496947628789655, 'samples': 843072, 'steps': 4390, 'loss/train': 1.0398978888988495} 01/28/2022 18:31:42 - INFO - codeparrot_training - Step 4391: {'lr': 0.000496945079185691, 'samples': 843264, 'steps': 4391, 'loss/train': 1.139833241701126} 01/28/2022 18:31:46 - INFO - codeparrot_training - Step 4392: {'lr': 0.0004969425285238928, 'samples': 843456, 'steps': 4392, 'loss/train': 0.915759801864624} 01/28/2022 18:31:50 - INFO - codeparrot_training - Step 4393: {'lr': 0.0004969399768042713, 'samples': 843648, 'steps': 4393, 'loss/train': 1.6948798298835754} 01/28/2022 18:31:56 - INFO - codeparrot_training - Step 4394: {'lr': 0.0004969374240268373, 'samples': 843840, 'steps': 4394, 'loss/train': 1.7305197715759277} 01/28/2022 18:32:00 - INFO - codeparrot_training - Step 4395: {'lr': 0.0004969348701916018, 'samples': 844032, 'steps': 4395, 'loss/train': 1.9504673480987549} 01/28/2022 18:32:04 - INFO - codeparrot_training - Step 4396: {'lr': 0.0004969323152985756, 'samples': 844224, 'steps': 4396, 'loss/train': 2.7450132966041565} 01/28/2022 18:32:08 - INFO - codeparrot_training - Step 4397: {'lr': 0.0004969297593477699, 'samples': 844416, 'steps': 4397, 'loss/train': 2.442277193069458} 01/28/2022 18:32:13 - INFO - codeparrot_training - Step 4398: {'lr': 0.0004969272023391955, 'samples': 844608, 'steps': 4398, 'loss/train': 1.7448431253433228} 01/28/2022 18:32:19 - INFO - codeparrot_training - Step 4399: {'lr': 0.0004969246442728633, 'samples': 844800, 'steps': 4399, 'loss/train': 1.5139941573143005} 01/28/2022 18:32:23 - INFO - codeparrot_training - Step 4400: {'lr': 0.0004969220851487844, 'samples': 844992, 'steps': 4400, 'loss/train': 2.213906764984131} 01/28/2022 18:32:27 - INFO - codeparrot_training - Step 4401: {'lr': 0.0004969195249669697, 'samples': 845184, 'steps': 4401, 'loss/train': 1.249854862689972} 01/28/2022 18:32:31 - INFO - codeparrot_training - Step 4402: {'lr': 0.0004969169637274301, 'samples': 845376, 'steps': 4402, 'loss/train': 1.7416445016860962} 01/28/2022 18:32:36 - INFO - codeparrot_training - Step 4403: {'lr': 0.0004969144014301767, 'samples': 845568, 'steps': 4403, 'loss/train': 1.9886625409126282} 01/28/2022 18:32:41 - INFO - codeparrot_training - Step 4404: {'lr': 0.0004969118380752205, 'samples': 845760, 'steps': 4404, 'loss/train': 2.221530497074127} 01/28/2022 18:32:45 - INFO - codeparrot_training - Step 4405: {'lr': 0.0004969092736625722, 'samples': 845952, 'steps': 4405, 'loss/train': 0.7798857986927032} 01/28/2022 18:32:50 - INFO - codeparrot_training - Step 4406: {'lr': 0.000496906708192243, 'samples': 846144, 'steps': 4406, 'loss/train': 1.9431450963020325} 01/28/2022 18:32:54 - INFO - codeparrot_training - Step 4407: {'lr': 0.000496904141664244, 'samples': 846336, 'steps': 4407, 'loss/train': 1.6069529056549072} 01/28/2022 18:32:58 - INFO - codeparrot_training - Step 4408: {'lr': 0.0004969015740785859, 'samples': 846528, 'steps': 4408, 'loss/train': 1.9514164924621582} 01/28/2022 18:33:03 - INFO - codeparrot_training - Step 4409: {'lr': 0.00049689900543528, 'samples': 846720, 'steps': 4409, 'loss/train': 2.117719531059265} 01/28/2022 18:33:07 - INFO - codeparrot_training - Step 4410: {'lr': 0.0004968964357343371, 'samples': 846912, 'steps': 4410, 'loss/train': 1.5115993022918701} 01/28/2022 18:33:12 - INFO - codeparrot_training - Step 4411: {'lr': 0.0004968938649757682, 'samples': 847104, 'steps': 4411, 'loss/train': 1.660585105419159} 01/28/2022 18:33:16 - INFO - codeparrot_training - Step 4412: {'lr': 0.0004968912931595845, 'samples': 847296, 'steps': 4412, 'loss/train': 2.4056711196899414} 01/28/2022 18:33:20 - INFO - codeparrot_training - Step 4413: {'lr': 0.0004968887202857968, 'samples': 847488, 'steps': 4413, 'loss/train': 1.4676493406295776} 01/28/2022 18:33:26 - INFO - codeparrot_training - Step 4414: {'lr': 0.0004968861463544163, 'samples': 847680, 'steps': 4414, 'loss/train': 1.551247000694275} 01/28/2022 18:33:30 - INFO - codeparrot_training - Step 4415: {'lr': 0.0004968835713654538, 'samples': 847872, 'steps': 4415, 'loss/train': 1.1038231551647186} 01/28/2022 18:33:35 - INFO - codeparrot_training - Step 4416: {'lr': 0.0004968809953189206, 'samples': 848064, 'steps': 4416, 'loss/train': 1.8623732328414917} 01/28/2022 18:33:39 - INFO - codeparrot_training - Step 4417: {'lr': 0.0004968784182148276, 'samples': 848256, 'steps': 4417, 'loss/train': 1.9481161236763} 01/28/2022 18:33:43 - INFO - codeparrot_training - Step 4418: {'lr': 0.0004968758400531859, 'samples': 848448, 'steps': 4418, 'loss/train': 2.283594310283661} 01/28/2022 18:33:49 - INFO - codeparrot_training - Step 4419: {'lr': 0.0004968732608340064, 'samples': 848640, 'steps': 4419, 'loss/train': 2.2433055639266968} 01/28/2022 18:33:53 - INFO - codeparrot_training - Step 4420: {'lr': 0.0004968706805573002, 'samples': 848832, 'steps': 4420, 'loss/train': 1.7242571711540222} 01/28/2022 18:33:58 - INFO - codeparrot_training - Step 4421: {'lr': 0.0004968680992230785, 'samples': 849024, 'steps': 4421, 'loss/train': 1.856023371219635} 01/28/2022 18:34:02 - INFO - codeparrot_training - Step 4422: {'lr': 0.0004968655168313522, 'samples': 849216, 'steps': 4422, 'loss/train': 0.8410206735134125} 01/28/2022 18:34:06 - INFO - codeparrot_training - Step 4423: {'lr': 0.0004968629333821324, 'samples': 849408, 'steps': 4423, 'loss/train': 1.7829033136367798} 01/28/2022 18:34:12 - INFO - codeparrot_training - Step 4424: {'lr': 0.0004968603488754302, 'samples': 849600, 'steps': 4424, 'loss/train': 3.923138737678528} 01/28/2022 18:34:16 - INFO - codeparrot_training - Step 4425: {'lr': 0.0004968577633112566, 'samples': 849792, 'steps': 4425, 'loss/train': 1.0798248052597046} 01/28/2022 18:34:21 - INFO - codeparrot_training - Step 4426: {'lr': 0.0004968551766896228, 'samples': 849984, 'steps': 4426, 'loss/train': 2.150588035583496} 01/28/2022 18:34:25 - INFO - codeparrot_training - Step 4427: {'lr': 0.0004968525890105399, 'samples': 850176, 'steps': 4427, 'loss/train': 2.196832537651062} 01/28/2022 18:34:29 - INFO - codeparrot_training - Step 4428: {'lr': 0.0004968500002740187, 'samples': 850368, 'steps': 4428, 'loss/train': 1.7876665592193604} 01/28/2022 18:34:34 - INFO - codeparrot_training - Step 4429: {'lr': 0.0004968474104800706, 'samples': 850560, 'steps': 4429, 'loss/train': 1.581382691860199} 01/28/2022 18:34:39 - INFO - codeparrot_training - Step 4430: {'lr': 0.0004968448196287066, 'samples': 850752, 'steps': 4430, 'loss/train': 2.3476316928863525} 01/28/2022 18:34:43 - INFO - codeparrot_training - Step 4431: {'lr': 0.0004968422277199377, 'samples': 850944, 'steps': 4431, 'loss/train': 1.9100305438041687} 01/28/2022 18:34:47 - INFO - codeparrot_training - Step 4432: {'lr': 0.000496839634753775, 'samples': 851136, 'steps': 4432, 'loss/train': 1.979985773563385} 01/28/2022 18:34:51 - INFO - codeparrot_training - Step 4433: {'lr': 0.0004968370407302299, 'samples': 851328, 'steps': 4433, 'loss/train': 1.8400885462760925} 01/28/2022 18:34:57 - INFO - codeparrot_training - Step 4434: {'lr': 0.0004968344456493132, 'samples': 851520, 'steps': 4434, 'loss/train': 1.9279945492744446} 01/28/2022 18:35:01 - INFO - codeparrot_training - Step 4435: {'lr': 0.000496831849511036, 'samples': 851712, 'steps': 4435, 'loss/train': 1.9168270826339722} 01/28/2022 18:35:05 - INFO - codeparrot_training - Step 4436: {'lr': 0.0004968292523154096, 'samples': 851904, 'steps': 4436, 'loss/train': 1.7281320691108704} 01/28/2022 18:35:10 - INFO - codeparrot_training - Step 4437: {'lr': 0.0004968266540624452, 'samples': 852096, 'steps': 4437, 'loss/train': 1.9494121670722961} 01/28/2022 18:35:14 - INFO - codeparrot_training - Step 4438: {'lr': 0.0004968240547521536, 'samples': 852288, 'steps': 4438, 'loss/train': 2.1619260907173157} 01/28/2022 18:35:20 - INFO - codeparrot_training - Step 4439: {'lr': 0.0004968214543845463, 'samples': 852480, 'steps': 4439, 'loss/train': 1.7024021744728088} 01/28/2022 18:35:24 - INFO - codeparrot_training - Step 4440: {'lr': 0.0004968188529596341, 'samples': 852672, 'steps': 4440, 'loss/train': 1.7269989252090454} 01/28/2022 18:35:29 - INFO - codeparrot_training - Step 4441: {'lr': 0.0004968162504774284, 'samples': 852864, 'steps': 4441, 'loss/train': 1.7746991515159607} 01/28/2022 18:35:33 - INFO - codeparrot_training - Step 4442: {'lr': 0.0004968136469379403, 'samples': 853056, 'steps': 4442, 'loss/train': 2.2735792994499207} 01/28/2022 18:35:38 - INFO - codeparrot_training - Step 4443: {'lr': 0.0004968110423411808, 'samples': 853248, 'steps': 4443, 'loss/train': 1.9928745031356812} 01/28/2022 18:35:43 - INFO - codeparrot_training - Step 4444: {'lr': 0.0004968084366871612, 'samples': 853440, 'steps': 4444, 'loss/train': 1.490264743566513} 01/28/2022 18:35:47 - INFO - codeparrot_training - Step 4445: {'lr': 0.0004968058299758926, 'samples': 853632, 'steps': 4445, 'loss/train': 2.217683494091034} 01/28/2022 18:35:51 - INFO - codeparrot_training - Step 4446: {'lr': 0.0004968032222073863, 'samples': 853824, 'steps': 4446, 'loss/train': 2.46413254737854} 01/28/2022 18:35:55 - INFO - codeparrot_training - Step 4447: {'lr': 0.0004968006133816532, 'samples': 854016, 'steps': 4447, 'loss/train': 1.743005096912384} 01/28/2022 18:36:01 - INFO - codeparrot_training - Step 4448: {'lr': 0.0004967980034987048, 'samples': 854208, 'steps': 4448, 'loss/train': 1.6528122425079346} 01/28/2022 18:36:05 - INFO - codeparrot_training - Step 4449: {'lr': 0.0004967953925585521, 'samples': 854400, 'steps': 4449, 'loss/train': 0.6276872903108597} 01/28/2022 18:36:09 - INFO - codeparrot_training - Step 4450: {'lr': 0.0004967927805612063, 'samples': 854592, 'steps': 4450, 'loss/train': 1.6625686883926392} 01/28/2022 18:36:14 - INFO - codeparrot_training - Step 4451: {'lr': 0.0004967901675066784, 'samples': 854784, 'steps': 4451, 'loss/train': 2.6740956902503967} 01/28/2022 18:36:18 - INFO - codeparrot_training - Step 4452: {'lr': 0.0004967875533949801, 'samples': 854976, 'steps': 4452, 'loss/train': 1.568530261516571} 01/28/2022 18:36:23 - INFO - codeparrot_training - Step 4453: {'lr': 0.000496784938226122, 'samples': 855168, 'steps': 4453, 'loss/train': 1.8030819296836853} 01/28/2022 18:36:27 - INFO - codeparrot_training - Step 4454: {'lr': 0.0004967823220001158, 'samples': 855360, 'steps': 4454, 'loss/train': 1.4958639442920685} 01/28/2022 18:36:32 - INFO - codeparrot_training - Step 4455: {'lr': 0.0004967797047169724, 'samples': 855552, 'steps': 4455, 'loss/train': 2.2730796933174133} 01/28/2022 18:36:36 - INFO - codeparrot_training - Step 4456: {'lr': 0.0004967770863767031, 'samples': 855744, 'steps': 4456, 'loss/train': 2.0650983452796936} 01/28/2022 18:36:40 - INFO - codeparrot_training - Step 4457: {'lr': 0.0004967744669793192, 'samples': 855936, 'steps': 4457, 'loss/train': 0.7200533598661423} 01/28/2022 18:36:46 - INFO - codeparrot_training - Step 4458: {'lr': 0.0004967718465248317, 'samples': 856128, 'steps': 4458, 'loss/train': 1.9788070321083069} 01/28/2022 18:36:50 - INFO - codeparrot_training - Step 4459: {'lr': 0.000496769225013252, 'samples': 856320, 'steps': 4459, 'loss/train': 2.1889697313308716} 01/28/2022 18:36:55 - INFO - codeparrot_training - Step 4460: {'lr': 0.0004967666024445913, 'samples': 856512, 'steps': 4460, 'loss/train': 2.1559048891067505} 01/28/2022 18:36:59 - INFO - codeparrot_training - Step 4461: {'lr': 0.000496763978818861, 'samples': 856704, 'steps': 4461, 'loss/train': 1.8269190788269043} 01/28/2022 18:37:03 - INFO - codeparrot_training - Step 4462: {'lr': 0.000496761354136072, 'samples': 856896, 'steps': 4462, 'loss/train': 2.1769076585769653} 01/28/2022 18:37:08 - INFO - codeparrot_training - Step 4463: {'lr': 0.0004967587283962358, 'samples': 857088, 'steps': 4463, 'loss/train': 2.1479249596595764} 01/28/2022 18:37:13 - INFO - codeparrot_training - Step 4464: {'lr': 0.0004967561015993635, 'samples': 857280, 'steps': 4464, 'loss/train': 0.9434415400028229} 01/28/2022 18:37:17 - INFO - codeparrot_training - Step 4465: {'lr': 0.0004967534737454665, 'samples': 857472, 'steps': 4465, 'loss/train': 1.2985457181930542} 01/28/2022 18:37:21 - INFO - codeparrot_training - Step 4466: {'lr': 0.000496750844834556, 'samples': 857664, 'steps': 4466, 'loss/train': 2.4467203617095947} 01/28/2022 18:37:25 - INFO - codeparrot_training - Step 4467: {'lr': 0.000496748214866643, 'samples': 857856, 'steps': 4467, 'loss/train': 1.8920363187789917} 01/28/2022 18:37:31 - INFO - codeparrot_training - Step 4468: {'lr': 0.0004967455838417392, 'samples': 858048, 'steps': 4468, 'loss/train': 2.180608034133911} 01/28/2022 18:37:36 - INFO - codeparrot_training - Step 4469: {'lr': 0.0004967429517598556, 'samples': 858240, 'steps': 4469, 'loss/train': 2.180857300758362} 01/28/2022 18:37:40 - INFO - codeparrot_training - Step 4470: {'lr': 0.0004967403186210036, 'samples': 858432, 'steps': 4470, 'loss/train': 1.791697919368744} 01/28/2022 18:37:44 - INFO - codeparrot_training - Step 4471: {'lr': 0.0004967376844251944, 'samples': 858624, 'steps': 4471, 'loss/train': 1.9167383909225464} 01/28/2022 18:37:48 - INFO - codeparrot_training - Step 4472: {'lr': 0.0004967350491724392, 'samples': 858816, 'steps': 4472, 'loss/train': 1.62879878282547} 01/28/2022 18:37:54 - INFO - codeparrot_training - Step 4473: {'lr': 0.0004967324128627495, 'samples': 859008, 'steps': 4473, 'loss/train': 1.6367349028587341} 01/28/2022 18:37:58 - INFO - codeparrot_training - Step 4474: {'lr': 0.0004967297754961365, 'samples': 859200, 'steps': 4474, 'loss/train': 2.342597544193268} 01/28/2022 18:38:02 - INFO - codeparrot_training - Step 4475: {'lr': 0.0004967271370726115, 'samples': 859392, 'steps': 4475, 'loss/train': 2.034091293811798} 01/28/2022 18:38:07 - INFO - codeparrot_training - Step 4476: {'lr': 0.0004967244975921857, 'samples': 859584, 'steps': 4476, 'loss/train': 1.820469617843628} 01/28/2022 18:38:11 - INFO - codeparrot_training - Step 4477: {'lr': 0.0004967218570548706, 'samples': 859776, 'steps': 4477, 'loss/train': 1.8418861627578735} 01/28/2022 18:38:16 - INFO - codeparrot_training - Step 4478: {'lr': 0.0004967192154606774, 'samples': 859968, 'steps': 4478, 'loss/train': 2.321178138256073} 01/28/2022 18:38:20 - INFO - codeparrot_training - Step 4479: {'lr': 0.0004967165728096172, 'samples': 860160, 'steps': 4479, 'loss/train': 2.1346555352211} 01/28/2022 18:38:25 - INFO - codeparrot_training - Step 4480: {'lr': 0.0004967139291017018, 'samples': 860352, 'steps': 4480, 'loss/train': 2.08349472284317} 01/28/2022 18:38:29 - INFO - codeparrot_training - Step 4481: {'lr': 0.0004967112843369423, 'samples': 860544, 'steps': 4481, 'loss/train': 2.2005327343940735} 01/28/2022 18:38:33 - INFO - codeparrot_training - Step 4482: {'lr': 0.0004967086385153499, 'samples': 860736, 'steps': 4482, 'loss/train': 1.8513988852500916} 01/28/2022 18:38:39 - INFO - codeparrot_training - Step 4483: {'lr': 0.0004967059916369359, 'samples': 860928, 'steps': 4483, 'loss/train': 2.1546115279197693} 01/28/2022 18:38:44 - INFO - codeparrot_training - Step 4484: {'lr': 0.000496703343701712, 'samples': 861120, 'steps': 4484, 'loss/train': 1.413329154253006} 01/28/2022 18:38:48 - INFO - codeparrot_training - Step 4485: {'lr': 0.0004967006947096892, 'samples': 861312, 'steps': 4485, 'loss/train': 1.9910629391670227} 01/28/2022 18:38:52 - INFO - codeparrot_training - Step 4486: {'lr': 0.0004966980446608789, 'samples': 861504, 'steps': 4486, 'loss/train': 1.5479630827903748} 01/28/2022 18:38:56 - INFO - codeparrot_training - Step 4487: {'lr': 0.0004966953935552925, 'samples': 861696, 'steps': 4487, 'loss/train': 1.3478399813175201} 01/28/2022 18:39:02 - INFO - codeparrot_training - Step 4488: {'lr': 0.0004966927413929415, 'samples': 861888, 'steps': 4488, 'loss/train': 2.5168413519859314} 01/28/2022 18:39:06 - INFO - codeparrot_training - Step 4489: {'lr': 0.0004966900881738371, 'samples': 862080, 'steps': 4489, 'loss/train': 1.3936504125595093} 01/28/2022 18:39:10 - INFO - codeparrot_training - Step 4490: {'lr': 0.0004966874338979907, 'samples': 862272, 'steps': 4490, 'loss/train': 1.350424736738205} 01/28/2022 18:39:15 - INFO - codeparrot_training - Step 4491: {'lr': 0.0004966847785654136, 'samples': 862464, 'steps': 4491, 'loss/train': 1.8597596883773804} 01/28/2022 18:39:19 - INFO - codeparrot_training - Step 4492: {'lr': 0.0004966821221761173, 'samples': 862656, 'steps': 4492, 'loss/train': 0.8575358390808105} 01/28/2022 18:39:24 - INFO - codeparrot_training - Step 4493: {'lr': 0.0004966794647301131, 'samples': 862848, 'steps': 4493, 'loss/train': 2.3077815771102905} 01/28/2022 18:39:29 - INFO - codeparrot_training - Step 4494: {'lr': 0.0004966768062274125, 'samples': 863040, 'steps': 4494, 'loss/train': 1.865513026714325} 01/28/2022 18:39:33 - INFO - codeparrot_training - Step 4495: {'lr': 0.0004966741466680266, 'samples': 863232, 'steps': 4495, 'loss/train': 2.018867254257202} 01/28/2022 18:39:37 - INFO - codeparrot_training - Step 4496: {'lr': 0.000496671486051967, 'samples': 863424, 'steps': 4496, 'loss/train': 1.844614326953888} 01/28/2022 18:39:41 - INFO - codeparrot_training - Step 4497: {'lr': 0.0004966688243792452, 'samples': 863616, 'steps': 4497, 'loss/train': 0.6246204525232315} 01/28/2022 18:39:47 - INFO - codeparrot_training - Step 4498: {'lr': 0.0004966661616498724, 'samples': 863808, 'steps': 4498, 'loss/train': 1.9611473679542542} 01/28/2022 18:39:52 - INFO - codeparrot_training - Step 4499: {'lr': 0.0004966634978638601, 'samples': 864000, 'steps': 4499, 'loss/train': 2.1656565070152283} 01/28/2022 18:39:56 - INFO - codeparrot_training - Step 4500: {'lr': 0.0004966608330212198, 'samples': 864192, 'steps': 4500, 'loss/train': 1.8364537954330444} 01/28/2022 18:40:00 - INFO - codeparrot_training - Step 4501: {'lr': 0.0004966581671219627, 'samples': 864384, 'steps': 4501, 'loss/train': 1.9520387649536133} 01/28/2022 18:40:04 - INFO - codeparrot_training - Step 4502: {'lr': 0.0004966555001661004, 'samples': 864576, 'steps': 4502, 'loss/train': 2.3177404403686523} 01/28/2022 18:40:10 - INFO - codeparrot_training - Step 4503: {'lr': 0.0004966528321536442, 'samples': 864768, 'steps': 4503, 'loss/train': 2.5363685488700867} 01/28/2022 18:40:14 - INFO - codeparrot_training - Step 4504: {'lr': 0.0004966501630846057, 'samples': 864960, 'steps': 4504, 'loss/train': 2.372959613800049} 01/28/2022 18:40:18 - INFO - codeparrot_training - Step 4505: {'lr': 0.000496647492958996, 'samples': 865152, 'steps': 4505, 'loss/train': 2.0310911536216736} 01/28/2022 18:40:22 - INFO - codeparrot_training - Step 4506: {'lr': 0.000496644821776827, 'samples': 865344, 'steps': 4506, 'loss/train': 1.8769218921661377} 01/28/2022 18:40:27 - INFO - codeparrot_training - Step 4507: {'lr': 0.0004966421495381098, 'samples': 865536, 'steps': 4507, 'loss/train': 2.05849152803421} 01/28/2022 18:40:32 - INFO - codeparrot_training - Step 4508: {'lr': 0.0004966394762428559, 'samples': 865728, 'steps': 4508, 'loss/train': 0.7913011908531189} 01/28/2022 18:40:37 - INFO - codeparrot_training - Step 4509: {'lr': 0.0004966368018910768, 'samples': 865920, 'steps': 4509, 'loss/train': 1.6731505393981934} 01/28/2022 18:40:41 - INFO - codeparrot_training - Step 4510: {'lr': 0.000496634126482784, 'samples': 866112, 'steps': 4510, 'loss/train': 1.1790187060832977} 01/28/2022 18:40:45 - INFO - codeparrot_training - Step 4511: {'lr': 0.000496631450017989, 'samples': 866304, 'steps': 4511, 'loss/train': 1.4259402751922607} 01/28/2022 18:40:49 - INFO - codeparrot_training - Step 4512: {'lr': 0.0004966287724967032, 'samples': 866496, 'steps': 4512, 'loss/train': 2.0573136806488037} 01/28/2022 18:40:55 - INFO - codeparrot_training - Step 4513: {'lr': 0.0004966260939189379, 'samples': 866688, 'steps': 4513, 'loss/train': 0.8983004093170166} 01/28/2022 18:40:59 - INFO - codeparrot_training - Step 4514: {'lr': 0.0004966234142847048, 'samples': 866880, 'steps': 4514, 'loss/train': 1.5738987922668457} 01/28/2022 18:41:03 - INFO - codeparrot_training - Step 4515: {'lr': 0.0004966207335940153, 'samples': 867072, 'steps': 4515, 'loss/train': 1.9794241189956665} 01/28/2022 18:41:08 - INFO - codeparrot_training - Step 4516: {'lr': 0.0004966180518468808, 'samples': 867264, 'steps': 4516, 'loss/train': 1.7013028264045715} 01/28/2022 18:41:14 - INFO - codeparrot_training - Step 4517: {'lr': 0.000496615369043313, 'samples': 867456, 'steps': 4517, 'loss/train': 2.2610456943511963} 01/28/2022 18:41:18 - INFO - codeparrot_training - Step 4518: {'lr': 0.0004966126851833233, 'samples': 867648, 'steps': 4518, 'loss/train': 1.788068175315857} 01/28/2022 18:41:22 - INFO - codeparrot_training - Step 4519: {'lr': 0.0004966100002669231, 'samples': 867840, 'steps': 4519, 'loss/train': 2.027806341648102} 01/28/2022 18:41:26 - INFO - codeparrot_training - Step 4520: {'lr': 0.0004966073142941239, 'samples': 868032, 'steps': 4520, 'loss/train': 1.7552745938301086} 01/28/2022 18:41:31 - INFO - codeparrot_training - Step 4521: {'lr': 0.0004966046272649372, 'samples': 868224, 'steps': 4521, 'loss/train': 2.309660017490387} 01/28/2022 18:41:36 - INFO - codeparrot_training - Step 4522: {'lr': 0.0004966019391793748, 'samples': 868416, 'steps': 4522, 'loss/train': 1.9935277104377747} 01/28/2022 18:41:40 - INFO - codeparrot_training - Step 4523: {'lr': 0.0004965992500374479, 'samples': 868608, 'steps': 4523, 'loss/train': 2.0903059244155884} 01/28/2022 18:41:44 - INFO - codeparrot_training - Step 4524: {'lr': 0.0004965965598391682, 'samples': 868800, 'steps': 4524, 'loss/train': 2.236828923225403} 01/28/2022 18:41:49 - INFO - codeparrot_training - Step 4525: {'lr': 0.000496593868584547, 'samples': 868992, 'steps': 4525, 'loss/train': 2.1278942227363586} 01/28/2022 18:41:53 - INFO - codeparrot_training - Step 4526: {'lr': 0.0004965911762735961, 'samples': 869184, 'steps': 4526, 'loss/train': 1.8178821802139282} 01/28/2022 18:41:59 - INFO - codeparrot_training - Step 4527: {'lr': 0.0004965884829063268, 'samples': 869376, 'steps': 4527, 'loss/train': 1.719883382320404} 01/28/2022 18:42:03 - INFO - codeparrot_training - Step 4528: {'lr': 0.0004965857884827508, 'samples': 869568, 'steps': 4528, 'loss/train': 3.6583807468414307} 01/28/2022 18:42:07 - INFO - codeparrot_training - Step 4529: {'lr': 0.0004965830930028795, 'samples': 869760, 'steps': 4529, 'loss/train': 3.3125925064086914} 01/28/2022 18:42:12 - INFO - codeparrot_training - Step 4530: {'lr': 0.0004965803964667246, 'samples': 869952, 'steps': 4530, 'loss/train': 2.7054659128189087} 01/28/2022 18:42:16 - INFO - codeparrot_training - Step 4531: {'lr': 0.0004965776988742976, 'samples': 870144, 'steps': 4531, 'loss/train': 2.311285972595215} 01/28/2022 18:42:22 - INFO - codeparrot_training - Step 4532: {'lr': 0.00049657500022561, 'samples': 870336, 'steps': 4532, 'loss/train': 2.181011438369751} 01/28/2022 18:42:26 - INFO - codeparrot_training - Step 4533: {'lr': 0.0004965723005206734, 'samples': 870528, 'steps': 4533, 'loss/train': 2.017977297306061} 01/28/2022 18:42:30 - INFO - codeparrot_training - Step 4534: {'lr': 0.0004965695997594993, 'samples': 870720, 'steps': 4534, 'loss/train': 1.321081817150116} 01/28/2022 18:42:34 - INFO - codeparrot_training - Step 4535: {'lr': 0.0004965668979420994, 'samples': 870912, 'steps': 4535, 'loss/train': 0.7834985554218292} 01/28/2022 18:42:39 - INFO - codeparrot_training - Step 4536: {'lr': 0.0004965641950684852, 'samples': 871104, 'steps': 4536, 'loss/train': 1.6003303527832031} 01/28/2022 18:42:44 - INFO - codeparrot_training - Step 4537: {'lr': 0.0004965614911386683, 'samples': 871296, 'steps': 4537, 'loss/train': 1.9181426167488098} 01/28/2022 18:42:48 - INFO - codeparrot_training - Step 4538: {'lr': 0.0004965587861526602, 'samples': 871488, 'steps': 4538, 'loss/train': 0.5561443716287613} 01/28/2022 18:42:52 - INFO - codeparrot_training - Step 4539: {'lr': 0.0004965560801104726, 'samples': 871680, 'steps': 4539, 'loss/train': 2.1171138882637024} 01/28/2022 18:42:57 - INFO - codeparrot_training - Step 4540: {'lr': 0.000496553373012117, 'samples': 871872, 'steps': 4540, 'loss/train': 1.976488173007965} 01/28/2022 18:43:01 - INFO - codeparrot_training - Step 4541: {'lr': 0.0004965506648576052, 'samples': 872064, 'steps': 4541, 'loss/train': 1.9645442962646484} 01/28/2022 18:43:07 - INFO - codeparrot_training - Step 4542: {'lr': 0.0004965479556469485, 'samples': 872256, 'steps': 4542, 'loss/train': 2.025690972805023} 01/28/2022 18:43:11 - INFO - codeparrot_training - Step 4543: {'lr': 0.0004965452453801586, 'samples': 872448, 'steps': 4543, 'loss/train': 2.1421000957489014} 01/28/2022 18:43:16 - INFO - codeparrot_training - Step 4544: {'lr': 0.0004965425340572472, 'samples': 872640, 'steps': 4544, 'loss/train': 1.2842435538768768} 01/28/2022 18:43:20 - INFO - codeparrot_training - Step 4545: {'lr': 0.0004965398216782258, 'samples': 872832, 'steps': 4545, 'loss/train': 7.5161497592926025} 01/28/2022 18:43:24 - INFO - codeparrot_training - Step 4546: {'lr': 0.0004965371082431062, 'samples': 873024, 'steps': 4546, 'loss/train': 1.3233047425746918} 01/28/2022 18:43:29 - INFO - codeparrot_training - Step 4547: {'lr': 0.0004965343937519, 'samples': 873216, 'steps': 4547, 'loss/train': 1.4365069270133972} 01/28/2022 18:43:34 - INFO - codeparrot_training - Step 4548: {'lr': 0.0004965316782046186, 'samples': 873408, 'steps': 4548, 'loss/train': 1.4975279867649078} 01/28/2022 18:43:38 - INFO - codeparrot_training - Step 4549: {'lr': 0.0004965289616012739, 'samples': 873600, 'steps': 4549, 'loss/train': 1.6588649153709412} 01/28/2022 18:43:42 - INFO - codeparrot_training - Step 4550: {'lr': 0.0004965262439418772, 'samples': 873792, 'steps': 4550, 'loss/train': 2.1272799968719482} 01/28/2022 18:43:46 - INFO - codeparrot_training - Step 4551: {'lr': 0.0004965235252264405, 'samples': 873984, 'steps': 4551, 'loss/train': 1.4017311930656433} 01/28/2022 18:43:53 - INFO - codeparrot_training - Step 4552: {'lr': 0.0004965208054549753, 'samples': 874176, 'steps': 4552, 'loss/train': 1.5602275729179382} 01/28/2022 18:43:57 - INFO - codeparrot_training - Step 4553: {'lr': 0.0004965180846274931, 'samples': 874368, 'steps': 4553, 'loss/train': 1.9895575046539307} 01/28/2022 18:44:01 - INFO - codeparrot_training - Step 4554: {'lr': 0.0004965153627440058, 'samples': 874560, 'steps': 4554, 'loss/train': 1.654666543006897} 01/28/2022 18:44:05 - INFO - codeparrot_training - Step 4555: {'lr': 0.000496512639804525, 'samples': 874752, 'steps': 4555, 'loss/train': 2.4396016001701355} 01/28/2022 18:44:10 - INFO - codeparrot_training - Step 4556: {'lr': 0.0004965099158090624, 'samples': 874944, 'steps': 4556, 'loss/train': 1.3603473901748657} 01/28/2022 18:44:15 - INFO - codeparrot_training - Step 4557: {'lr': 0.0004965071907576294, 'samples': 875136, 'steps': 4557, 'loss/train': 2.3936866521835327} 01/28/2022 18:44:19 - INFO - codeparrot_training - Step 4558: {'lr': 0.000496504464650238, 'samples': 875328, 'steps': 4558, 'loss/train': 1.0614933371543884} 01/28/2022 18:44:23 - INFO - codeparrot_training - Step 4559: {'lr': 0.0004965017374868997, 'samples': 875520, 'steps': 4559, 'loss/train': 1.8870547413825989} 01/28/2022 18:44:28 - INFO - codeparrot_training - Step 4560: {'lr': 0.0004964990092676262, 'samples': 875712, 'steps': 4560, 'loss/train': 2.1475608944892883} 01/28/2022 18:44:32 - INFO - codeparrot_training - Step 4561: {'lr': 0.0004964962799924293, 'samples': 875904, 'steps': 4561, 'loss/train': 1.8375974893569946} 01/28/2022 18:44:37 - INFO - codeparrot_training - Step 4562: {'lr': 0.0004964935496613206, 'samples': 876096, 'steps': 4562, 'loss/train': 2.003977954387665} 01/28/2022 18:44:41 - INFO - codeparrot_training - Step 4563: {'lr': 0.0004964908182743117, 'samples': 876288, 'steps': 4563, 'loss/train': 1.4138993918895721} 01/28/2022 18:44:46 - INFO - codeparrot_training - Step 4564: {'lr': 0.0004964880858314146, 'samples': 876480, 'steps': 4564, 'loss/train': 1.4496392905712128} 01/28/2022 18:44:50 - INFO - codeparrot_training - Step 4565: {'lr': 0.0004964853523326406, 'samples': 876672, 'steps': 4565, 'loss/train': 1.285059928894043} 01/28/2022 18:44:54 - INFO - codeparrot_training - Step 4566: {'lr': 0.0004964826177780017, 'samples': 876864, 'steps': 4566, 'loss/train': 1.7824081778526306} 01/28/2022 18:44:59 - INFO - codeparrot_training - Step 4567: {'lr': 0.0004964798821675096, 'samples': 877056, 'steps': 4567, 'loss/train': 1.6625095009803772} 01/28/2022 18:45:04 - INFO - codeparrot_training - Step 4568: {'lr': 0.0004964771455011758, 'samples': 877248, 'steps': 4568, 'loss/train': 3.007254481315613} 01/28/2022 18:45:08 - INFO - codeparrot_training - Step 4569: {'lr': 0.0004964744077790123, 'samples': 877440, 'steps': 4569, 'loss/train': 1.9206948280334473} 01/28/2022 18:45:12 - INFO - codeparrot_training - Step 4570: {'lr': 0.0004964716690010306, 'samples': 877632, 'steps': 4570, 'loss/train': 1.7204525470733643} 01/28/2022 18:45:16 - INFO - codeparrot_training - Step 4571: {'lr': 0.0004964689291672427, 'samples': 877824, 'steps': 4571, 'loss/train': 1.8928150534629822} 01/28/2022 18:45:22 - INFO - codeparrot_training - Step 4572: {'lr': 0.00049646618827766, 'samples': 878016, 'steps': 4572, 'loss/train': 2.466345191001892} 01/28/2022 18:45:27 - INFO - codeparrot_training - Step 4573: {'lr': 0.0004964634463322945, 'samples': 878208, 'steps': 4573, 'loss/train': 2.2652639150619507} 01/28/2022 18:45:31 - INFO - codeparrot_training - Step 4574: {'lr': 0.0004964607033311579, 'samples': 878400, 'steps': 4574, 'loss/train': 2.1831262707710266} 01/28/2022 18:45:35 - INFO - codeparrot_training - Step 4575: {'lr': 0.0004964579592742618, 'samples': 878592, 'steps': 4575, 'loss/train': 2.4241670966148376} 01/28/2022 18:45:39 - INFO - codeparrot_training - Step 4576: {'lr': 0.000496455214161618, 'samples': 878784, 'steps': 4576, 'loss/train': 1.4577971398830414} 01/28/2022 18:45:45 - INFO - codeparrot_training - Step 4577: {'lr': 0.0004964524679932385, 'samples': 878976, 'steps': 4577, 'loss/train': 2.840828776359558} 01/28/2022 18:45:49 - INFO - codeparrot_training - Step 4578: {'lr': 0.0004964497207691349, 'samples': 879168, 'steps': 4578, 'loss/train': 2.13342422246933} 01/28/2022 18:45:53 - INFO - codeparrot_training - Step 4579: {'lr': 0.0004964469724893188, 'samples': 879360, 'steps': 4579, 'loss/train': 2.030818283557892} 01/28/2022 18:45:57 - INFO - codeparrot_training - Step 4580: {'lr': 0.0004964442231538023, 'samples': 879552, 'steps': 4580, 'loss/train': 0.9036380052566528} 01/28/2022 18:46:02 - INFO - codeparrot_training - Step 4581: {'lr': 0.0004964414727625968, 'samples': 879744, 'steps': 4581, 'loss/train': 2.6627737283706665} 01/28/2022 18:46:07 - INFO - codeparrot_training - Step 4582: {'lr': 0.0004964387213157143, 'samples': 879936, 'steps': 4582, 'loss/train': 1.6237762570381165} 01/28/2022 18:46:11 - INFO - codeparrot_training - Step 4583: {'lr': 0.0004964359688131667, 'samples': 880128, 'steps': 4583, 'loss/train': 1.372402310371399} 01/28/2022 18:46:16 - INFO - codeparrot_training - Step 4584: {'lr': 0.0004964332152549657, 'samples': 880320, 'steps': 4584, 'loss/train': 2.453890800476074} 01/28/2022 18:46:20 - INFO - codeparrot_training - Step 4585: {'lr': 0.0004964304606411229, 'samples': 880512, 'steps': 4585, 'loss/train': 1.5362775921821594} 01/28/2022 18:46:24 - INFO - codeparrot_training - Step 4586: {'lr': 0.0004964277049716503, 'samples': 880704, 'steps': 4586, 'loss/train': 2.441802978515625} 01/28/2022 18:46:30 - INFO - codeparrot_training - Step 4587: {'lr': 0.0004964249482465597, 'samples': 880896, 'steps': 4587, 'loss/train': 1.8003780841827393} 01/28/2022 18:46:34 - INFO - codeparrot_training - Step 4588: {'lr': 0.0004964221904658629, 'samples': 881088, 'steps': 4588, 'loss/train': 1.8577871918678284} 01/28/2022 18:46:39 - INFO - codeparrot_training - Step 4589: {'lr': 0.0004964194316295716, 'samples': 881280, 'steps': 4589, 'loss/train': 2.2688518166542053} 01/28/2022 18:46:43 - INFO - codeparrot_training - Step 4590: {'lr': 0.0004964166717376978, 'samples': 881472, 'steps': 4590, 'loss/train': 1.8729196786880493} 01/28/2022 18:46:47 - INFO - codeparrot_training - Step 4591: {'lr': 0.0004964139107902531, 'samples': 881664, 'steps': 4591, 'loss/train': 1.5796796679496765} 01/28/2022 18:46:52 - INFO - codeparrot_training - Step 4592: {'lr': 0.0004964111487872495, 'samples': 881856, 'steps': 4592, 'loss/train': 1.206094354391098} 01/28/2022 18:46:57 - INFO - codeparrot_training - Step 4593: {'lr': 0.0004964083857286988, 'samples': 882048, 'steps': 4593, 'loss/train': 1.9018798470497131} 01/28/2022 18:47:01 - INFO - codeparrot_training - Step 4594: {'lr': 0.0004964056216146129, 'samples': 882240, 'steps': 4594, 'loss/train': 1.4679525196552277} 01/28/2022 18:47:05 - INFO - codeparrot_training - Step 4595: {'lr': 0.0004964028564450034, 'samples': 882432, 'steps': 4595, 'loss/train': 1.2134357392787933} 01/28/2022 18:47:09 - INFO - codeparrot_training - Step 4596: {'lr': 0.0004964000902198824, 'samples': 882624, 'steps': 4596, 'loss/train': 2.3021700382232666} 01/28/2022 18:47:16 - INFO - codeparrot_training - Step 4597: {'lr': 0.0004963973229392617, 'samples': 882816, 'steps': 4597, 'loss/train': 0.6805270314216614} 01/28/2022 18:47:20 - INFO - codeparrot_training - Step 4598: {'lr': 0.0004963945546031529, 'samples': 883008, 'steps': 4598, 'loss/train': 2.214299261569977} 01/28/2022 18:47:24 - INFO - codeparrot_training - Step 4599: {'lr': 0.0004963917852115683, 'samples': 883200, 'steps': 4599, 'loss/train': 0.8779729306697845} 01/28/2022 18:47:28 - INFO - codeparrot_training - Step 4600: {'lr': 0.0004963890147645194, 'samples': 883392, 'steps': 4600, 'loss/train': 0.7526542246341705} 01/28/2022 18:47:33 - INFO - codeparrot_training - Step 4601: {'lr': 0.0004963862432620183, 'samples': 883584, 'steps': 4601, 'loss/train': 0.9177334606647491} 01/28/2022 18:47:38 - INFO - codeparrot_training - Step 4602: {'lr': 0.0004963834707040767, 'samples': 883776, 'steps': 4602, 'loss/train': 1.509323000907898} 01/28/2022 18:47:42 - INFO - codeparrot_training - Step 4603: {'lr': 0.0004963806970907066, 'samples': 883968, 'steps': 4603, 'loss/train': 1.6451305747032166} 01/28/2022 18:47:46 - INFO - codeparrot_training - Step 4604: {'lr': 0.0004963779224219197, 'samples': 884160, 'steps': 4604, 'loss/train': 2.1026267409324646} 01/28/2022 18:47:51 - INFO - codeparrot_training - Step 4605: {'lr': 0.0004963751466977281, 'samples': 884352, 'steps': 4605, 'loss/train': 1.807510256767273} 01/28/2022 18:47:55 - INFO - codeparrot_training - Step 4606: {'lr': 0.0004963723699181437, 'samples': 884544, 'steps': 4606, 'loss/train': 1.5971179604530334} 01/28/2022 18:48:01 - INFO - codeparrot_training - Step 4607: {'lr': 0.0004963695920831781, 'samples': 884736, 'steps': 4607, 'loss/train': 0.9028041064739227} 01/28/2022 18:48:05 - INFO - codeparrot_training - Step 4608: {'lr': 0.0004963668131928436, 'samples': 884928, 'steps': 4608, 'loss/train': 1.747477948665619} 01/28/2022 18:48:10 - INFO - codeparrot_training - Step 4609: {'lr': 0.0004963640332471518, 'samples': 885120, 'steps': 4609, 'loss/train': 2.2076223492622375} 01/28/2022 18:48:14 - INFO - codeparrot_training - Step 4610: {'lr': 0.0004963612522461147, 'samples': 885312, 'steps': 4610, 'loss/train': 1.067127764225006} 01/28/2022 18:48:18 - INFO - codeparrot_training - Step 4611: {'lr': 0.0004963584701897443, 'samples': 885504, 'steps': 4611, 'loss/train': 1.986553430557251} 01/28/2022 18:48:22 - INFO - codeparrot_training - Step 4612: {'lr': 0.0004963556870780523, 'samples': 885696, 'steps': 4612, 'loss/train': 2.6937661170959473} 01/28/2022 18:48:28 - INFO - codeparrot_training - Step 4613: {'lr': 0.0004963529029110509, 'samples': 885888, 'steps': 4613, 'loss/train': 2.157647967338562} 01/28/2022 18:48:32 - INFO - codeparrot_training - Step 4614: {'lr': 0.0004963501176887519, 'samples': 886080, 'steps': 4614, 'loss/train': 2.147068440914154} 01/28/2022 18:48:37 - INFO - codeparrot_training - Step 4615: {'lr': 0.000496347331411167, 'samples': 886272, 'steps': 4615, 'loss/train': 2.4590944647789} 01/28/2022 18:48:41 - INFO - codeparrot_training - Step 4616: {'lr': 0.0004963445440783086, 'samples': 886464, 'steps': 4616, 'loss/train': 1.6318028569221497} 01/28/2022 18:48:45 - INFO - codeparrot_training - Step 4617: {'lr': 0.0004963417556901882, 'samples': 886656, 'steps': 4617, 'loss/train': 2.1738113164901733} 01/28/2022 18:48:53 - INFO - codeparrot_training - Step 4618: {'lr': 0.0004963389662468182, 'samples': 886848, 'steps': 4618, 'loss/train': 2.3618693947792053} 01/28/2022 18:48:57 - INFO - codeparrot_training - Step 4619: {'lr': 0.0004963361757482101, 'samples': 887040, 'steps': 4619, 'loss/train': 2.055828273296356} 01/28/2022 18:49:01 - INFO - codeparrot_training - Step 4620: {'lr': 0.000496333384194376, 'samples': 887232, 'steps': 4620, 'loss/train': 2.70833283662796} 01/28/2022 18:49:05 - INFO - codeparrot_training - Step 4621: {'lr': 0.000496330591585328, 'samples': 887424, 'steps': 4621, 'loss/train': 1.692827045917511} 01/28/2022 18:49:10 - INFO - codeparrot_training - Step 4622: {'lr': 0.0004963277979210779, 'samples': 887616, 'steps': 4622, 'loss/train': 2.3906931281089783} 01/28/2022 18:49:15 - INFO - codeparrot_training - Step 4623: {'lr': 0.0004963250032016379, 'samples': 887808, 'steps': 4623, 'loss/train': 2.355844259262085} 01/28/2022 18:49:19 - INFO - codeparrot_training - Step 4624: {'lr': 0.0004963222074270197, 'samples': 888000, 'steps': 4624, 'loss/train': 2.3493348956108093} 01/28/2022 18:49:24 - INFO - codeparrot_training - Step 4625: {'lr': 0.0004963194105972353, 'samples': 888192, 'steps': 4625, 'loss/train': 2.029644191265106} 01/28/2022 18:49:28 - INFO - codeparrot_training - Step 4626: {'lr': 0.0004963166127122969, 'samples': 888384, 'steps': 4626, 'loss/train': 2.0485939979553223} 01/28/2022 18:49:32 - INFO - codeparrot_training - Step 4627: {'lr': 0.0004963138137722161, 'samples': 888576, 'steps': 4627, 'loss/train': 1.7251179814338684} 01/28/2022 18:49:37 - INFO - codeparrot_training - Step 4628: {'lr': 0.0004963110137770054, 'samples': 888768, 'steps': 4628, 'loss/train': 2.145281195640564} 01/28/2022 18:49:42 - INFO - codeparrot_training - Step 4629: {'lr': 0.0004963082127266764, 'samples': 888960, 'steps': 4629, 'loss/train': 1.290560781955719} 01/28/2022 18:49:46 - INFO - codeparrot_training - Step 4630: {'lr': 0.0004963054106212414, 'samples': 889152, 'steps': 4630, 'loss/train': 2.503536880016327} 01/28/2022 18:49:50 - INFO - codeparrot_training - Step 4631: {'lr': 0.000496302607460712, 'samples': 889344, 'steps': 4631, 'loss/train': 1.8747693300247192} 01/28/2022 18:49:54 - INFO - codeparrot_training - Step 4632: {'lr': 0.0004962998032451005, 'samples': 889536, 'steps': 4632, 'loss/train': 1.660361409187317} 01/28/2022 18:50:01 - INFO - codeparrot_training - Step 4633: {'lr': 0.0004962969979744189, 'samples': 889728, 'steps': 4633, 'loss/train': 1.5161912441253662} 01/28/2022 18:50:05 - INFO - codeparrot_training - Step 4634: {'lr': 0.0004962941916486791, 'samples': 889920, 'steps': 4634, 'loss/train': 1.12159064412117} 01/28/2022 18:50:09 - INFO - codeparrot_training - Step 4635: {'lr': 0.0004962913842678934, 'samples': 890112, 'steps': 4635, 'loss/train': 2.28110933303833} 01/28/2022 18:50:13 - INFO - codeparrot_training - Step 4636: {'lr': 0.0004962885758320734, 'samples': 890304, 'steps': 4636, 'loss/train': 2.0436469316482544} 01/28/2022 18:50:18 - INFO - codeparrot_training - Step 4637: {'lr': 0.0004962857663412314, 'samples': 890496, 'steps': 4637, 'loss/train': 1.6924946308135986} 01/28/2022 18:50:23 - INFO - codeparrot_training - Step 4638: {'lr': 0.0004962829557953794, 'samples': 890688, 'steps': 4638, 'loss/train': 1.7249125242233276} 01/28/2022 18:50:28 - INFO - codeparrot_training - Step 4639: {'lr': 0.0004962801441945293, 'samples': 890880, 'steps': 4639, 'loss/train': 1.4591057896614075} 01/28/2022 18:50:32 - INFO - codeparrot_training - Step 4640: {'lr': 0.0004962773315386935, 'samples': 891072, 'steps': 4640, 'loss/train': 1.8764217495918274} 01/28/2022 18:50:36 - INFO - codeparrot_training - Step 4641: {'lr': 0.0004962745178278837, 'samples': 891264, 'steps': 4641, 'loss/train': 1.3364755511283875} 01/28/2022 18:50:40 - INFO - codeparrot_training - Step 4642: {'lr': 0.000496271703062112, 'samples': 891456, 'steps': 4642, 'loss/train': 1.5508959889411926} 01/28/2022 18:50:47 - INFO - codeparrot_training - Step 4643: {'lr': 0.0004962688872413906, 'samples': 891648, 'steps': 4643, 'loss/train': 1.7844775915145874} 01/28/2022 18:50:51 - INFO - codeparrot_training - Step 4644: {'lr': 0.0004962660703657315, 'samples': 891840, 'steps': 4644, 'loss/train': 2.24796599149704} 01/28/2022 18:50:55 - INFO - codeparrot_training - Step 4645: {'lr': 0.0004962632524351467, 'samples': 892032, 'steps': 4645, 'loss/train': 1.056949943304062} 01/28/2022 18:51:00 - INFO - codeparrot_training - Step 4646: {'lr': 0.0004962604334496483, 'samples': 892224, 'steps': 4646, 'loss/train': 1.9296897053718567} 01/28/2022 18:51:04 - INFO - codeparrot_training - Step 4647: {'lr': 0.0004962576134092485, 'samples': 892416, 'steps': 4647, 'loss/train': 1.1204775273799896} 01/28/2022 18:51:09 - INFO - codeparrot_training - Step 4648: {'lr': 0.0004962547923139592, 'samples': 892608, 'steps': 4648, 'loss/train': 1.3604376912117004} 01/28/2022 18:51:13 - INFO - codeparrot_training - Step 4649: {'lr': 0.0004962519701637926, 'samples': 892800, 'steps': 4649, 'loss/train': 1.973893404006958} 01/28/2022 18:51:17 - INFO - codeparrot_training - Step 4650: {'lr': 0.0004962491469587607, 'samples': 892992, 'steps': 4650, 'loss/train': 1.5566293001174927} 01/28/2022 18:51:22 - INFO - codeparrot_training - Step 4651: {'lr': 0.0004962463226988758, 'samples': 893184, 'steps': 4651, 'loss/train': 1.5795426964759827} 01/28/2022 18:51:26 - INFO - codeparrot_training - Step 4652: {'lr': 0.0004962434973841497, 'samples': 893376, 'steps': 4652, 'loss/train': 1.2389181554317474} 01/28/2022 18:51:32 - INFO - codeparrot_training - Step 4653: {'lr': 0.0004962406710145946, 'samples': 893568, 'steps': 4653, 'loss/train': 1.644378125667572} 01/28/2022 18:51:36 - INFO - codeparrot_training - Step 4654: {'lr': 0.0004962378435902228, 'samples': 893760, 'steps': 4654, 'loss/train': 1.0566630363464355} 01/28/2022 18:51:40 - INFO - codeparrot_training - Step 4655: {'lr': 0.0004962350151110461, 'samples': 893952, 'steps': 4655, 'loss/train': 2.311233937740326} 01/28/2022 18:51:45 - INFO - codeparrot_training - Step 4656: {'lr': 0.0004962321855770769, 'samples': 894144, 'steps': 4656, 'loss/train': 2.005272388458252} 01/28/2022 18:51:49 - INFO - codeparrot_training - Step 4657: {'lr': 0.0004962293549883273, 'samples': 894336, 'steps': 4657, 'loss/train': 1.1389159262180328} 01/28/2022 18:51:54 - INFO - codeparrot_training - Step 4658: {'lr': 0.0004962265233448092, 'samples': 894528, 'steps': 4658, 'loss/train': 2.6812788248062134} 01/28/2022 18:51:58 - INFO - codeparrot_training - Step 4659: {'lr': 0.0004962236906465349, 'samples': 894720, 'steps': 4659, 'loss/train': 1.173781156539917} 01/28/2022 18:52:02 - INFO - codeparrot_training - Step 4660: {'lr': 0.0004962208568935164, 'samples': 894912, 'steps': 4660, 'loss/train': 1.4393253922462463} 01/28/2022 18:52:07 - INFO - codeparrot_training - Step 4661: {'lr': 0.000496218022085766, 'samples': 895104, 'steps': 4661, 'loss/train': 1.5622837543487549} 01/28/2022 18:52:11 - INFO - codeparrot_training - Step 4662: {'lr': 0.0004962151862232958, 'samples': 895296, 'steps': 4662, 'loss/train': 1.8467532992362976} 01/28/2022 18:52:17 - INFO - codeparrot_training - Step 4663: {'lr': 0.000496212349306118, 'samples': 895488, 'steps': 4663, 'loss/train': 1.7693976759910583} 01/28/2022 18:52:21 - INFO - codeparrot_training - Step 4664: {'lr': 0.0004962095113342445, 'samples': 895680, 'steps': 4664, 'loss/train': 1.6778106093406677} 01/28/2022 18:52:25 - INFO - codeparrot_training - Step 4665: {'lr': 0.0004962066723076878, 'samples': 895872, 'steps': 4665, 'loss/train': 2.298068940639496} 01/28/2022 18:52:29 - INFO - codeparrot_training - Step 4666: {'lr': 0.0004962038322264598, 'samples': 896064, 'steps': 4666, 'loss/train': 1.8520735502243042} 01/28/2022 18:52:34 - INFO - codeparrot_training - Step 4667: {'lr': 0.0004962009910905728, 'samples': 896256, 'steps': 4667, 'loss/train': 2.455179512500763} 01/28/2022 18:52:39 - INFO - codeparrot_training - Step 4668: {'lr': 0.0004961981489000389, 'samples': 896448, 'steps': 4668, 'loss/train': 1.3842593729496002} 01/28/2022 18:52:43 - INFO - codeparrot_training - Step 4669: {'lr': 0.0004961953056548703, 'samples': 896640, 'steps': 4669, 'loss/train': 2.4419293999671936} 01/28/2022 18:52:47 - INFO - codeparrot_training - Step 4670: {'lr': 0.0004961924613550793, 'samples': 896832, 'steps': 4670, 'loss/train': 2.1022151112556458} 01/28/2022 18:52:52 - INFO - codeparrot_training - Step 4671: {'lr': 0.0004961896160006778, 'samples': 897024, 'steps': 4671, 'loss/train': 1.3901905417442322} 01/28/2022 18:52:56 - INFO - codeparrot_training - Step 4672: {'lr': 0.0004961867695916782, 'samples': 897216, 'steps': 4672, 'loss/train': 2.186062216758728} 01/28/2022 18:53:01 - INFO - codeparrot_training - Step 4673: {'lr': 0.0004961839221280927, 'samples': 897408, 'steps': 4673, 'loss/train': 1.6170905828475952} 01/28/2022 18:53:06 - INFO - codeparrot_training - Step 4674: {'lr': 0.0004961810736099334, 'samples': 897600, 'steps': 4674, 'loss/train': 1.6201590299606323} 01/28/2022 18:53:10 - INFO - codeparrot_training - Step 4675: {'lr': 0.0004961782240372126, 'samples': 897792, 'steps': 4675, 'loss/train': 1.1070431470870972} 01/28/2022 18:53:14 - INFO - codeparrot_training - Step 4676: {'lr': 0.0004961753734099425, 'samples': 897984, 'steps': 4676, 'loss/train': 1.3322149515151978} 01/28/2022 18:53:18 - INFO - codeparrot_training - Step 4677: {'lr': 0.0004961725217281352, 'samples': 898176, 'steps': 4677, 'loss/train': 2.4725911617279053} 01/28/2022 18:53:24 - INFO - codeparrot_training - Step 4678: {'lr': 0.0004961696689918029, 'samples': 898368, 'steps': 4678, 'loss/train': 1.755039095878601} 01/28/2022 18:53:29 - INFO - codeparrot_training - Step 4679: {'lr': 0.0004961668152009581, 'samples': 898560, 'steps': 4679, 'loss/train': 2.433667480945587} 01/28/2022 18:53:33 - INFO - codeparrot_training - Step 4680: {'lr': 0.0004961639603556127, 'samples': 898752, 'steps': 4680, 'loss/train': 2.114495873451233} 01/28/2022 18:53:37 - INFO - codeparrot_training - Step 4681: {'lr': 0.0004961611044557792, 'samples': 898944, 'steps': 4681, 'loss/train': 1.9990044236183167} 01/28/2022 18:53:41 - INFO - codeparrot_training - Step 4682: {'lr': 0.0004961582475014695, 'samples': 899136, 'steps': 4682, 'loss/train': 1.0056543946266174} 01/28/2022 18:53:47 - INFO - codeparrot_training - Step 4683: {'lr': 0.0004961553894926961, 'samples': 899328, 'steps': 4683, 'loss/train': 1.5419161319732666} 01/28/2022 18:53:51 - INFO - codeparrot_training - Step 4684: {'lr': 0.0004961525304294712, 'samples': 899520, 'steps': 4684, 'loss/train': 1.9881933331489563} 01/28/2022 18:53:55 - INFO - codeparrot_training - Step 4685: {'lr': 0.000496149670311807, 'samples': 899712, 'steps': 4685, 'loss/train': 2.2963762879371643} 01/28/2022 18:54:00 - INFO - codeparrot_training - Step 4686: {'lr': 0.0004961468091397158, 'samples': 899904, 'steps': 4686, 'loss/train': 1.386005938053131} 01/28/2022 18:54:06 - INFO - codeparrot_training - Step 4687: {'lr': 0.0004961439469132098, 'samples': 900096, 'steps': 4687, 'loss/train': 2.562486469745636} 01/28/2022 18:54:10 - INFO - codeparrot_training - Step 4688: {'lr': 0.0004961410836323014, 'samples': 900288, 'steps': 4688, 'loss/train': 2.004485249519348} 01/28/2022 18:54:15 - INFO - codeparrot_training - Step 4689: {'lr': 0.0004961382192970027, 'samples': 900480, 'steps': 4689, 'loss/train': 0.7815653085708618} 01/28/2022 18:54:19 - INFO - codeparrot_training - Step 4690: {'lr': 0.0004961353539073258, 'samples': 900672, 'steps': 4690, 'loss/train': 1.5341446995735168} 01/28/2022 18:54:23 - INFO - codeparrot_training - Step 4691: {'lr': 0.0004961324874632835, 'samples': 900864, 'steps': 4691, 'loss/train': 1.3662261664867401} 01/28/2022 18:54:27 - INFO - codeparrot_training - Step 4692: {'lr': 0.0004961296199648877, 'samples': 901056, 'steps': 4692, 'loss/train': 2.452177405357361} 01/28/2022 18:54:33 - INFO - codeparrot_training - Step 4693: {'lr': 0.0004961267514121507, 'samples': 901248, 'steps': 4693, 'loss/train': 1.8780591487884521} 01/28/2022 18:54:37 - INFO - codeparrot_training - Step 4694: {'lr': 0.0004961238818050849, 'samples': 901440, 'steps': 4694, 'loss/train': 2.422157943248749} 01/28/2022 18:54:41 - INFO - codeparrot_training - Step 4695: {'lr': 0.0004961210111437026, 'samples': 901632, 'steps': 4695, 'loss/train': 1.3439653515815735} 01/28/2022 18:54:46 - INFO - codeparrot_training - Step 4696: {'lr': 0.0004961181394280159, 'samples': 901824, 'steps': 4696, 'loss/train': 1.6756884455680847} 01/28/2022 18:54:50 - INFO - codeparrot_training - Step 4697: {'lr': 0.0004961152666580373, 'samples': 902016, 'steps': 4697, 'loss/train': 1.236877977848053} 01/28/2022 18:54:55 - INFO - codeparrot_training - Step 4698: {'lr': 0.0004961123928337791, 'samples': 902208, 'steps': 4698, 'loss/train': 0.7794237434864044} 01/28/2022 18:55:00 - INFO - codeparrot_training - Step 4699: {'lr': 0.0004961095179552535, 'samples': 902400, 'steps': 4699, 'loss/train': 1.766785204410553} 01/28/2022 18:55:04 - INFO - codeparrot_training - Step 4700: {'lr': 0.0004961066420224729, 'samples': 902592, 'steps': 4700, 'loss/train': 2.3346042037010193} 01/28/2022 18:55:08 - INFO - codeparrot_training - Step 4701: {'lr': 0.0004961037650354496, 'samples': 902784, 'steps': 4701, 'loss/train': 1.674684226512909} 01/28/2022 18:55:12 - INFO - codeparrot_training - Step 4702: {'lr': 0.0004961008869941959, 'samples': 902976, 'steps': 4702, 'loss/train': 2.090745985507965} 01/28/2022 18:55:19 - INFO - codeparrot_training - Step 4703: {'lr': 0.0004960980078987241, 'samples': 903168, 'steps': 4703, 'loss/train': 2.0237515568733215} 01/28/2022 18:55:23 - INFO - codeparrot_training - Step 4704: {'lr': 0.0004960951277490467, 'samples': 903360, 'steps': 4704, 'loss/train': 2.0085160732269287} 01/28/2022 18:55:28 - INFO - codeparrot_training - Step 4705: {'lr': 0.0004960922465451758, 'samples': 903552, 'steps': 4705, 'loss/train': 1.1282781064510345} 01/28/2022 18:55:32 - INFO - codeparrot_training - Step 4706: {'lr': 0.0004960893642871239, 'samples': 903744, 'steps': 4706, 'loss/train': 1.7484825253486633} 01/28/2022 18:55:37 - INFO - codeparrot_training - Step 4707: {'lr': 0.0004960864809749034, 'samples': 903936, 'steps': 4707, 'loss/train': 1.4784815311431885} 01/28/2022 18:55:41 - INFO - codeparrot_training - Step 4708: {'lr': 0.0004960835966085264, 'samples': 904128, 'steps': 4708, 'loss/train': 1.8030345439910889} 01/28/2022 18:55:46 - INFO - codeparrot_training - Step 4709: {'lr': 0.0004960807111880055, 'samples': 904320, 'steps': 4709, 'loss/train': 1.459822028875351} 01/28/2022 18:55:50 - INFO - codeparrot_training - Step 4710: {'lr': 0.000496077824713353, 'samples': 904512, 'steps': 4710, 'loss/train': 2.609183371067047} 01/28/2022 18:55:54 - INFO - codeparrot_training - Step 4711: {'lr': 0.0004960749371845812, 'samples': 904704, 'steps': 4711, 'loss/train': 1.9909179210662842} 01/28/2022 18:55:59 - INFO - codeparrot_training - Step 4712: {'lr': 0.0004960720486017025, 'samples': 904896, 'steps': 4712, 'loss/train': 2.3461629152297974} 01/28/2022 18:56:05 - INFO - codeparrot_training - Step 4713: {'lr': 0.0004960691589647292, 'samples': 905088, 'steps': 4713, 'loss/train': 0.9840753972530365} 01/28/2022 18:56:09 - INFO - codeparrot_training - Step 4714: {'lr': 0.0004960662682736739, 'samples': 905280, 'steps': 4714, 'loss/train': 2.169280529022217} 01/28/2022 18:56:14 - INFO - codeparrot_training - Step 4715: {'lr': 0.0004960633765285487, 'samples': 905472, 'steps': 4715, 'loss/train': 2.954653322696686} 01/28/2022 18:56:18 - INFO - codeparrot_training - Step 4716: {'lr': 0.0004960604837293663, 'samples': 905664, 'steps': 4716, 'loss/train': 2.5806164145469666} 01/28/2022 18:56:22 - INFO - codeparrot_training - Step 4717: {'lr': 0.0004960575898761388, 'samples': 905856, 'steps': 4717, 'loss/train': 1.923590898513794} 01/28/2022 18:56:27 - INFO - codeparrot_training - Step 4718: {'lr': 0.0004960546949688788, 'samples': 906048, 'steps': 4718, 'loss/train': 2.217213749885559} 01/28/2022 18:56:32 - INFO - codeparrot_training - Step 4719: {'lr': 0.0004960517990075985, 'samples': 906240, 'steps': 4719, 'loss/train': 1.22993803024292} 01/28/2022 18:56:36 - INFO - codeparrot_training - Step 4720: {'lr': 0.0004960489019923105, 'samples': 906432, 'steps': 4720, 'loss/train': 1.8372772336006165} 01/28/2022 18:56:40 - INFO - codeparrot_training - Step 4721: {'lr': 0.0004960460039230271, 'samples': 906624, 'steps': 4721, 'loss/train': 2.5661324858665466} 01/28/2022 18:56:46 - INFO - codeparrot_training - Step 4722: {'lr': 0.0004960431047997608, 'samples': 906816, 'steps': 4722, 'loss/train': 1.486649215221405} 01/28/2022 18:56:50 - INFO - codeparrot_training - Step 4723: {'lr': 0.0004960402046225239, 'samples': 907008, 'steps': 4723, 'loss/train': 1.997354507446289} 01/28/2022 18:56:54 - INFO - codeparrot_training - Step 4724: {'lr': 0.0004960373033913289, 'samples': 907200, 'steps': 4724, 'loss/train': 1.7508507370948792} 01/28/2022 18:56:58 - INFO - codeparrot_training - Step 4725: {'lr': 0.0004960344011061882, 'samples': 907392, 'steps': 4725, 'loss/train': 2.8043925762176514} 01/28/2022 18:57:03 - INFO - codeparrot_training - Step 4726: {'lr': 0.0004960314977671144, 'samples': 907584, 'steps': 4726, 'loss/train': 2.349417507648468} 01/28/2022 18:57:08 - INFO - codeparrot_training - Step 4727: {'lr': 0.0004960285933741196, 'samples': 907776, 'steps': 4727, 'loss/train': 0.38763874769210815} 01/28/2022 18:57:12 - INFO - codeparrot_training - Step 4728: {'lr': 0.0004960256879272166, 'samples': 907968, 'steps': 4728, 'loss/train': 0.5348479002714157} 01/28/2022 18:57:16 - INFO - codeparrot_training - Step 4729: {'lr': 0.0004960227814264175, 'samples': 908160, 'steps': 4729, 'loss/train': 1.4473612010478973} 01/28/2022 18:57:21 - INFO - codeparrot_training - Step 4730: {'lr': 0.0004960198738717351, 'samples': 908352, 'steps': 4730, 'loss/train': 1.4488929212093353} 01/28/2022 18:57:25 - INFO - codeparrot_training - Step 4731: {'lr': 0.0004960169652631815, 'samples': 908544, 'steps': 4731, 'loss/train': 2.0156405568122864} 01/28/2022 18:57:32 - INFO - codeparrot_training - Step 4732: {'lr': 0.0004960140556007695, 'samples': 908736, 'steps': 4732, 'loss/train': 1.4586882591247559} 01/28/2022 18:57:36 - INFO - codeparrot_training - Step 4733: {'lr': 0.0004960111448845114, 'samples': 908928, 'steps': 4733, 'loss/train': 1.6348856091499329} 01/28/2022 18:57:40 - INFO - codeparrot_training - Step 4734: {'lr': 0.0004960082331144195, 'samples': 909120, 'steps': 4734, 'loss/train': 1.3271476328372955} 01/28/2022 18:57:44 - INFO - codeparrot_training - Step 4735: {'lr': 0.0004960053202905066, 'samples': 909312, 'steps': 4735, 'loss/train': 2.075933575630188} 01/28/2022 18:57:49 - INFO - codeparrot_training - Step 4736: {'lr': 0.0004960024064127849, 'samples': 909504, 'steps': 4736, 'loss/train': 2.3662837743759155} 01/28/2022 18:57:53 - INFO - codeparrot_training - Step 4737: {'lr': 0.0004959994914812671, 'samples': 909696, 'steps': 4737, 'loss/train': 0.6959213018417358} 01/28/2022 18:57:58 - INFO - codeparrot_training - Step 4738: {'lr': 0.0004959965754959656, 'samples': 909888, 'steps': 4738, 'loss/train': 1.1586846113204956} 01/28/2022 18:58:02 - INFO - codeparrot_training - Step 4739: {'lr': 0.0004959936584568928, 'samples': 910080, 'steps': 4739, 'loss/train': 2.0850952863693237} 01/28/2022 18:58:06 - INFO - codeparrot_training - Step 4740: {'lr': 0.0004959907403640614, 'samples': 910272, 'steps': 4740, 'loss/train': 1.9924424886703491} 01/28/2022 18:58:11 - INFO - codeparrot_training - Step 4741: {'lr': 0.0004959878212174837, 'samples': 910464, 'steps': 4741, 'loss/train': 1.8843727111816406} 01/28/2022 18:58:15 - INFO - codeparrot_training - Step 4742: {'lr': 0.0004959849010171723, 'samples': 910656, 'steps': 4742, 'loss/train': 1.5525822043418884} 01/28/2022 18:58:21 - INFO - codeparrot_training - Step 4743: {'lr': 0.0004959819797631397, 'samples': 910848, 'steps': 4743, 'loss/train': 1.816737949848175} 01/28/2022 18:58:25 - INFO - codeparrot_training - Step 4744: {'lr': 0.0004959790574553984, 'samples': 911040, 'steps': 4744, 'loss/train': 1.345628947019577} 01/28/2022 18:58:29 - INFO - codeparrot_training - Step 4745: {'lr': 0.000495976134093961, 'samples': 911232, 'steps': 4745, 'loss/train': 2.020986557006836} 01/28/2022 18:58:34 - INFO - codeparrot_training - Step 4746: {'lr': 0.0004959732096788398, 'samples': 911424, 'steps': 4746, 'loss/train': 2.3766599893569946} 01/28/2022 18:58:38 - INFO - codeparrot_training - Step 4747: {'lr': 0.0004959702842100475, 'samples': 911616, 'steps': 4747, 'loss/train': 1.444633662700653} 01/28/2022 18:58:44 - INFO - codeparrot_training - Step 4748: {'lr': 0.0004959673576875967, 'samples': 911808, 'steps': 4748, 'loss/train': 2.3275551795959473} 01/28/2022 18:58:49 - INFO - codeparrot_training - Step 4749: {'lr': 0.0004959644301114998, 'samples': 912000, 'steps': 4749, 'loss/train': 1.9813150763511658} 01/28/2022 18:58:53 - INFO - codeparrot_training - Step 4750: {'lr': 0.0004959615014817694, 'samples': 912192, 'steps': 4750, 'loss/train': 1.7799773812294006} 01/28/2022 18:58:57 - INFO - codeparrot_training - Step 4751: {'lr': 0.000495958571798418, 'samples': 912384, 'steps': 4751, 'loss/train': 1.9275190830230713} 01/28/2022 18:59:01 - INFO - codeparrot_training - Step 4752: {'lr': 0.0004959556410614582, 'samples': 912576, 'steps': 4752, 'loss/train': 2.099505364894867} 01/28/2022 18:59:06 - INFO - codeparrot_training - Step 4753: {'lr': 0.0004959527092709026, 'samples': 912768, 'steps': 4753, 'loss/train': 2.0410079956054688} 01/28/2022 18:59:11 - INFO - codeparrot_training - Step 4754: {'lr': 0.0004959497764267636, 'samples': 912960, 'steps': 4754, 'loss/train': 1.9472821354866028} 01/28/2022 18:59:15 - INFO - codeparrot_training - Step 4755: {'lr': 0.0004959468425290537, 'samples': 913152, 'steps': 4755, 'loss/train': 0.7063070833683014} 01/28/2022 18:59:19 - INFO - codeparrot_training - Step 4756: {'lr': 0.0004959439075777858, 'samples': 913344, 'steps': 4756, 'loss/train': 1.4187124371528625} 01/28/2022 18:59:24 - INFO - codeparrot_training - Step 4757: {'lr': 0.0004959409715729723, 'samples': 913536, 'steps': 4757, 'loss/train': 2.3713756799697876} 01/28/2022 18:59:31 - INFO - codeparrot_training - Step 4758: {'lr': 0.0004959380345146258, 'samples': 913728, 'steps': 4758, 'loss/train': 2.8111029267311096} 01/28/2022 18:59:35 - INFO - codeparrot_training - Step 4759: {'lr': 0.0004959350964027588, 'samples': 913920, 'steps': 4759, 'loss/train': 1.6550792455673218} 01/28/2022 18:59:39 - INFO - codeparrot_training - Step 4760: {'lr': 0.000495932157237384, 'samples': 914112, 'steps': 4760, 'loss/train': 1.954468846321106} 01/28/2022 18:59:43 - INFO - codeparrot_training - Step 4761: {'lr': 0.0004959292170185139, 'samples': 914304, 'steps': 4761, 'loss/train': 1.8791059255599976} 01/28/2022 18:59:48 - INFO - codeparrot_training - Step 4762: {'lr': 0.0004959262757461611, 'samples': 914496, 'steps': 4762, 'loss/train': 3.2326215505599976} 01/28/2022 18:59:52 - INFO - codeparrot_training - Step 4763: {'lr': 0.0004959233334203382, 'samples': 914688, 'steps': 4763, 'loss/train': 1.7478793859481812} 01/28/2022 18:59:57 - INFO - codeparrot_training - Step 4764: {'lr': 0.0004959203900410579, 'samples': 914880, 'steps': 4764, 'loss/train': 1.8916136026382446} 01/28/2022 19:00:01 - INFO - codeparrot_training - Step 4765: {'lr': 0.0004959174456083327, 'samples': 915072, 'steps': 4765, 'loss/train': 2.2674002051353455} 01/28/2022 19:00:06 - INFO - codeparrot_training - Step 4766: {'lr': 0.0004959145001221752, 'samples': 915264, 'steps': 4766, 'loss/train': 1.902198851108551} 01/28/2022 19:00:10 - INFO - codeparrot_training - Step 4767: {'lr': 0.0004959115535825982, 'samples': 915456, 'steps': 4767, 'loss/train': 1.7440947890281677} 01/28/2022 19:00:15 - INFO - codeparrot_training - Step 4768: {'lr': 0.000495908605989614, 'samples': 915648, 'steps': 4768, 'loss/train': 3.329767942428589} 01/28/2022 19:00:19 - INFO - codeparrot_training - Step 4769: {'lr': 0.0004959056573432357, 'samples': 915840, 'steps': 4769, 'loss/train': 2.4508379101753235} 01/28/2022 19:00:24 - INFO - codeparrot_training - Step 4770: {'lr': 0.0004959027076434754, 'samples': 916032, 'steps': 4770, 'loss/train': 1.7546367645263672} 01/28/2022 19:00:28 - INFO - codeparrot_training - Step 4771: {'lr': 0.000495899756890346, 'samples': 916224, 'steps': 4771, 'loss/train': 1.5716968774795532} 01/28/2022 19:00:32 - INFO - codeparrot_training - Step 4772: {'lr': 0.0004958968050838603, 'samples': 916416, 'steps': 4772, 'loss/train': 1.9419652819633484} 01/28/2022 19:00:38 - INFO - codeparrot_training - Step 4773: {'lr': 0.0004958938522240306, 'samples': 916608, 'steps': 4773, 'loss/train': 2.0639310479164124} 01/28/2022 19:00:42 - INFO - codeparrot_training - Step 4774: {'lr': 0.0004958908983108697, 'samples': 916800, 'steps': 4774, 'loss/train': 1.7481552958488464} 01/28/2022 19:00:47 - INFO - codeparrot_training - Step 4775: {'lr': 0.0004958879433443903, 'samples': 916992, 'steps': 4775, 'loss/train': 2.683569073677063} 01/28/2022 19:00:51 - INFO - codeparrot_training - Step 4776: {'lr': 0.0004958849873246051, 'samples': 917184, 'steps': 4776, 'loss/train': 1.6940404772758484} 01/28/2022 19:00:55 - INFO - codeparrot_training - Step 4777: {'lr': 0.0004958820302515268, 'samples': 917376, 'steps': 4777, 'loss/train': 1.5743388533592224} 01/28/2022 19:01:01 - INFO - codeparrot_training - Step 4778: {'lr': 0.0004958790721251678, 'samples': 917568, 'steps': 4778, 'loss/train': 1.719227135181427} 01/28/2022 19:01:05 - INFO - codeparrot_training - Step 4779: {'lr': 0.000495876112945541, 'samples': 917760, 'steps': 4779, 'loss/train': 2.3288092017173767} 01/28/2022 19:01:09 - INFO - codeparrot_training - Step 4780: {'lr': 0.0004958731527126589, 'samples': 917952, 'steps': 4780, 'loss/train': 1.8508472442626953} 01/28/2022 19:01:13 - INFO - codeparrot_training - Step 4781: {'lr': 0.0004958701914265344, 'samples': 918144, 'steps': 4781, 'loss/train': 2.1331767439842224} 01/28/2022 19:01:17 - INFO - codeparrot_training - Step 4782: {'lr': 0.0004958672290871799, 'samples': 918336, 'steps': 4782, 'loss/train': 2.353658616542816} 01/28/2022 19:01:23 - INFO - codeparrot_training - Step 4783: {'lr': 0.0004958642656946084, 'samples': 918528, 'steps': 4783, 'loss/train': 1.654792070388794} 01/28/2022 19:01:27 - INFO - codeparrot_training - Step 4784: {'lr': 0.0004958613012488324, 'samples': 918720, 'steps': 4784, 'loss/train': 0.8116044402122498} 01/28/2022 19:01:31 - INFO - codeparrot_training - Step 4785: {'lr': 0.0004958583357498647, 'samples': 918912, 'steps': 4785, 'loss/train': 2.117419123649597} 01/28/2022 19:01:35 - INFO - codeparrot_training - Step 4786: {'lr': 0.000495855369197718, 'samples': 919104, 'steps': 4786, 'loss/train': 2.0733906626701355} 01/28/2022 19:01:40 - INFO - codeparrot_training - Step 4787: {'lr': 0.0004958524015924048, 'samples': 919296, 'steps': 4787, 'loss/train': 2.2008216977119446} 01/28/2022 19:01:45 - INFO - codeparrot_training - Step 4788: {'lr': 0.0004958494329339382, 'samples': 919488, 'steps': 4788, 'loss/train': 2.938650369644165} 01/28/2022 19:01:49 - INFO - codeparrot_training - Step 4789: {'lr': 0.0004958464632223306, 'samples': 919680, 'steps': 4789, 'loss/train': 2.0349690914154053} 01/28/2022 19:01:53 - INFO - codeparrot_training - Step 4790: {'lr': 0.0004958434924575947, 'samples': 919872, 'steps': 4790, 'loss/train': 2.1530123949050903} 01/28/2022 19:01:58 - INFO - codeparrot_training - Step 4791: {'lr': 0.0004958405206397434, 'samples': 920064, 'steps': 4791, 'loss/train': 1.762939453125} 01/28/2022 19:02:02 - INFO - codeparrot_training - Step 4792: {'lr': 0.0004958375477687896, 'samples': 920256, 'steps': 4792, 'loss/train': 1.6141095757484436} 01/28/2022 19:02:08 - INFO - codeparrot_training - Step 4793: {'lr': 0.0004958345738447456, 'samples': 920448, 'steps': 4793, 'loss/train': 1.479662150144577} 01/28/2022 19:02:12 - INFO - codeparrot_training - Step 4794: {'lr': 0.0004958315988676244, 'samples': 920640, 'steps': 4794, 'loss/train': 1.9335134625434875} 01/28/2022 19:02:16 - INFO - codeparrot_training - Step 4795: {'lr': 0.0004958286228374387, 'samples': 920832, 'steps': 4795, 'loss/train': 1.6583143472671509} 01/28/2022 19:02:21 - INFO - codeparrot_training - Step 4796: {'lr': 0.0004958256457542011, 'samples': 921024, 'steps': 4796, 'loss/train': 1.723211646080017} 01/28/2022 19:02:25 - INFO - codeparrot_training - Step 4797: {'lr': 0.0004958226676179246, 'samples': 921216, 'steps': 4797, 'loss/train': 2.1168930530548096} 01/28/2022 19:02:30 - INFO - codeparrot_training - Step 4798: {'lr': 0.0004958196884286218, 'samples': 921408, 'steps': 4798, 'loss/train': 2.3911930918693542} 01/28/2022 19:02:35 - INFO - codeparrot_training - Step 4799: {'lr': 0.0004958167081863057, 'samples': 921600, 'steps': 4799, 'loss/train': 1.7090451121330261} 01/28/2022 19:02:39 - INFO - codeparrot_training - Step 4800: {'lr': 0.0004958137268909887, 'samples': 921792, 'steps': 4800, 'loss/train': 1.9705735445022583} 01/28/2022 19:02:43 - INFO - codeparrot_training - Step 4801: {'lr': 0.0004958107445426838, 'samples': 921984, 'steps': 4801, 'loss/train': 1.6670352816581726} 01/28/2022 19:02:47 - INFO - codeparrot_training - Step 4802: {'lr': 0.0004958077611414037, 'samples': 922176, 'steps': 4802, 'loss/train': 2.2616515159606934} 01/28/2022 19:02:53 - INFO - codeparrot_training - Step 4803: {'lr': 0.0004958047766871612, 'samples': 922368, 'steps': 4803, 'loss/train': 1.7644538283348083} 01/28/2022 19:02:57 - INFO - codeparrot_training - Step 4804: {'lr': 0.000495801791179969, 'samples': 922560, 'steps': 4804, 'loss/train': 0.7912818789482117} 01/28/2022 19:03:01 - INFO - codeparrot_training - Step 4805: {'lr': 0.0004957988046198401, 'samples': 922752, 'steps': 4805, 'loss/train': 1.1019141376018524} 01/28/2022 19:03:06 - INFO - codeparrot_training - Step 4806: {'lr': 0.0004957958170067872, 'samples': 922944, 'steps': 4806, 'loss/train': 1.8365693092346191} 01/28/2022 19:03:10 - INFO - codeparrot_training - Step 4807: {'lr': 0.000495792828340823, 'samples': 923136, 'steps': 4807, 'loss/train': 1.3434186279773712} 01/28/2022 19:03:16 - INFO - codeparrot_training - Step 4808: {'lr': 0.0004957898386219603, 'samples': 923328, 'steps': 4808, 'loss/train': 0.9793606996536255} 01/28/2022 19:03:20 - INFO - codeparrot_training - Step 4809: {'lr': 0.0004957868478502121, 'samples': 923520, 'steps': 4809, 'loss/train': 1.8006450533866882} 01/28/2022 19:03:25 - INFO - codeparrot_training - Step 4810: {'lr': 0.0004957838560255911, 'samples': 923712, 'steps': 4810, 'loss/train': 2.041714668273926} 01/28/2022 19:03:29 - INFO - codeparrot_training - Step 4811: {'lr': 0.0004957808631481101, 'samples': 923904, 'steps': 4811, 'loss/train': 1.7974990010261536} 01/28/2022 19:03:33 - INFO - codeparrot_training - Step 4812: {'lr': 0.0004957778692177819, 'samples': 924096, 'steps': 4812, 'loss/train': 1.5188319683074951} 01/28/2022 19:03:38 - INFO - codeparrot_training - Step 4813: {'lr': 0.0004957748742346193, 'samples': 924288, 'steps': 4813, 'loss/train': 1.2809396088123322} 01/28/2022 19:03:43 - INFO - codeparrot_training - Step 4814: {'lr': 0.0004957718781986352, 'samples': 924480, 'steps': 4814, 'loss/train': 1.6420184969902039} 01/28/2022 19:03:47 - INFO - codeparrot_training - Step 4815: {'lr': 0.0004957688811098425, 'samples': 924672, 'steps': 4815, 'loss/train': 1.3203451931476593} 01/28/2022 19:03:51 - INFO - codeparrot_training - Step 4816: {'lr': 0.0004957658829682539, 'samples': 924864, 'steps': 4816, 'loss/train': 1.417262703180313} 01/28/2022 19:03:55 - INFO - codeparrot_training - Step 4817: {'lr': 0.0004957628837738823, 'samples': 925056, 'steps': 4817, 'loss/train': 2.279658079147339} 01/28/2022 19:04:02 - INFO - codeparrot_training - Step 4818: {'lr': 0.0004957598835267405, 'samples': 925248, 'steps': 4818, 'loss/train': 2.388337254524231} 01/28/2022 19:04:06 - INFO - codeparrot_training - Step 4819: {'lr': 0.0004957568822268415, 'samples': 925440, 'steps': 4819, 'loss/train': 1.7979757189750671} 01/28/2022 19:04:10 - INFO - codeparrot_training - Step 4820: {'lr': 0.000495753879874198, 'samples': 925632, 'steps': 4820, 'loss/train': 1.9564763903617859} 01/28/2022 19:04:14 - INFO - codeparrot_training - Step 4821: {'lr': 0.0004957508764688227, 'samples': 925824, 'steps': 4821, 'loss/train': 1.27923783659935} 01/28/2022 19:04:19 - INFO - codeparrot_training - Step 4822: {'lr': 0.000495747872010729, 'samples': 926016, 'steps': 4822, 'loss/train': 2.30619478225708} 01/28/2022 19:04:24 - INFO - codeparrot_training - Step 4823: {'lr': 0.0004957448664999293, 'samples': 926208, 'steps': 4823, 'loss/train': 1.6548432111740112} 01/28/2022 19:04:28 - INFO - codeparrot_training - Step 4824: {'lr': 0.0004957418599364367, 'samples': 926400, 'steps': 4824, 'loss/train': 2.6122562885284424} 01/28/2022 19:04:32 - INFO - codeparrot_training - Step 4825: {'lr': 0.000495738852320264, 'samples': 926592, 'steps': 4825, 'loss/train': 2.1793225407600403} 01/28/2022 19:04:36 - INFO - codeparrot_training - Step 4826: {'lr': 0.000495735843651424, 'samples': 926784, 'steps': 4826, 'loss/train': 1.8313291668891907} 01/28/2022 19:04:41 - INFO - codeparrot_training - Step 4827: {'lr': 0.0004957328339299297, 'samples': 926976, 'steps': 4827, 'loss/train': 2.156960964202881} 01/28/2022 19:04:46 - INFO - codeparrot_training - Step 4828: {'lr': 0.0004957298231557939, 'samples': 927168, 'steps': 4828, 'loss/train': 1.307847797870636} 01/28/2022 19:04:50 - INFO - codeparrot_training - Step 4829: {'lr': 0.0004957268113290297, 'samples': 927360, 'steps': 4829, 'loss/train': 1.1150791347026825} 01/28/2022 19:04:55 - INFO - codeparrot_training - Step 4830: {'lr': 0.0004957237984496499, 'samples': 927552, 'steps': 4830, 'loss/train': 2.2194299697875977} 01/28/2022 19:04:59 - INFO - codeparrot_training - Step 4831: {'lr': 0.0004957207845176673, 'samples': 927744, 'steps': 4831, 'loss/train': 1.9037759900093079} 01/28/2022 19:05:03 - INFO - codeparrot_training - Step 4832: {'lr': 0.0004957177695330948, 'samples': 927936, 'steps': 4832, 'loss/train': 1.6625783443450928} 01/28/2022 19:05:10 - INFO - codeparrot_training - Step 4833: {'lr': 0.0004957147534959455, 'samples': 928128, 'steps': 4833, 'loss/train': 1.1298426389694214} 01/28/2022 19:05:14 - INFO - codeparrot_training - Step 4834: {'lr': 0.0004957117364062321, 'samples': 928320, 'steps': 4834, 'loss/train': 1.3311621844768524} 01/28/2022 19:05:19 - INFO - codeparrot_training - Step 4835: {'lr': 0.0004957087182639678, 'samples': 928512, 'steps': 4835, 'loss/train': 1.3567043244838715} 01/28/2022 19:05:23 - INFO - codeparrot_training - Step 4836: {'lr': 0.0004957056990691653, 'samples': 928704, 'steps': 4836, 'loss/train': 1.851756513118744} 01/28/2022 19:05:28 - INFO - codeparrot_training - Step 4837: {'lr': 0.0004957026788218377, 'samples': 928896, 'steps': 4837, 'loss/train': 1.4013554155826569} 01/28/2022 19:05:32 - INFO - codeparrot_training - Step 4838: {'lr': 0.0004956996575219977, 'samples': 929088, 'steps': 4838, 'loss/train': 1.942375659942627} 01/28/2022 19:05:37 - INFO - codeparrot_training - Step 4839: {'lr': 0.0004956966351696584, 'samples': 929280, 'steps': 4839, 'loss/train': 1.787551760673523} 01/28/2022 19:05:41 - INFO - codeparrot_training - Step 4840: {'lr': 0.0004956936117648329, 'samples': 929472, 'steps': 4840, 'loss/train': 2.5732626914978027} 01/28/2022 19:05:45 - INFO - codeparrot_training - Step 4841: {'lr': 0.0004956905873075338, 'samples': 929664, 'steps': 4841, 'loss/train': 1.611294150352478} 01/28/2022 19:05:51 - INFO - codeparrot_training - Step 4842: {'lr': 0.0004956875617977743, 'samples': 929856, 'steps': 4842, 'loss/train': 1.0774834156036377} 01/28/2022 19:05:55 - INFO - codeparrot_training - Step 4843: {'lr': 0.0004956845352355674, 'samples': 930048, 'steps': 4843, 'loss/train': 2.125356674194336} 01/28/2022 19:05:59 - INFO - codeparrot_training - Step 4844: {'lr': 0.0004956815076209257, 'samples': 930240, 'steps': 4844, 'loss/train': 1.5127151012420654} 01/28/2022 19:06:03 - INFO - codeparrot_training - Step 4845: {'lr': 0.0004956784789538626, 'samples': 930432, 'steps': 4845, 'loss/train': 1.818418800830841} 01/28/2022 19:06:08 - INFO - codeparrot_training - Step 4846: {'lr': 0.000495675449234391, 'samples': 930624, 'steps': 4846, 'loss/train': 1.4129148423671722} 01/28/2022 19:06:13 - INFO - codeparrot_training - Step 4847: {'lr': 0.0004956724184625237, 'samples': 930816, 'steps': 4847, 'loss/train': 2.0240832567214966} 01/28/2022 19:06:17 - INFO - codeparrot_training - Step 4848: {'lr': 0.0004956693866382738, 'samples': 931008, 'steps': 4848, 'loss/train': 2.0907803177833557} 01/28/2022 19:06:21 - INFO - codeparrot_training - Step 4849: {'lr': 0.0004956663537616542, 'samples': 931200, 'steps': 4849, 'loss/train': 1.7869825959205627} 01/28/2022 19:06:26 - INFO - codeparrot_training - Step 4850: {'lr': 0.000495663319832678, 'samples': 931392, 'steps': 4850, 'loss/train': 2.225042939186096} 01/28/2022 19:06:30 - INFO - codeparrot_training - Step 4851: {'lr': 0.0004956602848513581, 'samples': 931584, 'steps': 4851, 'loss/train': 1.4084435105323792} 01/28/2022 19:06:36 - INFO - codeparrot_training - Step 4852: {'lr': 0.0004956572488177075, 'samples': 931776, 'steps': 4852, 'loss/train': 2.1482614874839783} 01/28/2022 19:06:40 - INFO - codeparrot_training - Step 4853: {'lr': 0.0004956542117317393, 'samples': 931968, 'steps': 4853, 'loss/train': 1.41716730594635} 01/28/2022 19:06:44 - INFO - codeparrot_training - Step 4854: {'lr': 0.0004956511735934665, 'samples': 932160, 'steps': 4854, 'loss/train': 1.4204772412776947} 01/28/2022 19:06:49 - INFO - codeparrot_training - Step 4855: {'lr': 0.000495648134402902, 'samples': 932352, 'steps': 4855, 'loss/train': 1.074988692998886} 01/28/2022 19:06:53 - INFO - codeparrot_training - Step 4856: {'lr': 0.0004956450941600589, 'samples': 932544, 'steps': 4856, 'loss/train': 1.9822155833244324} 01/28/2022 19:06:58 - INFO - codeparrot_training - Step 4857: {'lr': 0.0004956420528649504, 'samples': 932736, 'steps': 4857, 'loss/train': 1.3167532682418823} 01/28/2022 19:07:03 - INFO - codeparrot_training - Step 4858: {'lr': 0.0004956390105175892, 'samples': 932928, 'steps': 4858, 'loss/train': 2.684695065021515} 01/28/2022 19:07:07 - INFO - codeparrot_training - Step 4859: {'lr': 0.0004956359671179885, 'samples': 933120, 'steps': 4859, 'loss/train': 1.6436421275138855} 01/28/2022 19:07:11 - INFO - codeparrot_training - Step 4860: {'lr': 0.0004956329226661612, 'samples': 933312, 'steps': 4860, 'loss/train': 1.7758135199546814} 01/28/2022 19:07:15 - INFO - codeparrot_training - Step 4861: {'lr': 0.0004956298771621206, 'samples': 933504, 'steps': 4861, 'loss/train': 1.1713523268699646} 01/28/2022 19:07:23 - INFO - codeparrot_training - Step 4862: {'lr': 0.0004956268306058795, 'samples': 933696, 'steps': 4862, 'loss/train': 1.111701786518097} 01/28/2022 19:07:27 - INFO - codeparrot_training - Step 4863: {'lr': 0.0004956237829974511, 'samples': 933888, 'steps': 4863, 'loss/train': 1.0213948488235474} 01/28/2022 19:07:31 - INFO - codeparrot_training - Step 4864: {'lr': 0.0004956207343368485, 'samples': 934080, 'steps': 4864, 'loss/train': 2.032146155834198} 01/28/2022 19:07:35 - INFO - codeparrot_training - Step 4865: {'lr': 0.0004956176846240845, 'samples': 934272, 'steps': 4865, 'loss/train': 1.927709698677063} 01/28/2022 19:07:40 - INFO - codeparrot_training - Step 4866: {'lr': 0.0004956146338591725, 'samples': 934464, 'steps': 4866, 'loss/train': 2.442746579647064} 01/28/2022 19:07:44 - INFO - codeparrot_training - Step 4867: {'lr': 0.0004956115820421253, 'samples': 934656, 'steps': 4867, 'loss/train': 2.2309606075286865} 01/28/2022 19:07:48 - INFO - codeparrot_training - Step 4868: {'lr': 0.000495608529172956, 'samples': 934848, 'steps': 4868, 'loss/train': 2.4768855571746826} 01/28/2022 19:07:54 - INFO - codeparrot_training - Step 4869: {'lr': 0.000495605475251678, 'samples': 935040, 'steps': 4869, 'loss/train': 2.038381040096283} 01/28/2022 19:07:58 - INFO - codeparrot_training - Step 4870: {'lr': 0.000495602420278304, 'samples': 935232, 'steps': 4870, 'loss/train': 1.629505455493927} 01/28/2022 19:08:02 - INFO - codeparrot_training - Step 4871: {'lr': 0.0004955993642528471, 'samples': 935424, 'steps': 4871, 'loss/train': 2.2458804845809937} 01/28/2022 19:08:06 - INFO - codeparrot_training - Step 4872: {'lr': 0.0004955963071753206, 'samples': 935616, 'steps': 4872, 'loss/train': 1.6396105885505676} 01/28/2022 19:08:11 - INFO - codeparrot_training - Step 4873: {'lr': 0.0004955932490457375, 'samples': 935808, 'steps': 4873, 'loss/train': 2.0302376747131348} 01/28/2022 19:08:16 - INFO - codeparrot_training - Step 4874: {'lr': 0.0004955901898641109, 'samples': 936000, 'steps': 4874, 'loss/train': 2.678380250930786} 01/28/2022 19:08:20 - INFO - codeparrot_training - Step 4875: {'lr': 0.000495587129630454, 'samples': 936192, 'steps': 4875, 'loss/train': 1.899935245513916} 01/28/2022 19:08:25 - INFO - codeparrot_training - Step 4876: {'lr': 0.0004955840683447797, 'samples': 936384, 'steps': 4876, 'loss/train': 1.87450110912323} 01/28/2022 19:08:29 - INFO - codeparrot_training - Step 4877: {'lr': 0.0004955810060071012, 'samples': 936576, 'steps': 4877, 'loss/train': 2.0140727162361145} 01/28/2022 19:08:33 - INFO - codeparrot_training - Step 4878: {'lr': 0.0004955779426174318, 'samples': 936768, 'steps': 4878, 'loss/train': 2.7046748399734497} 01/28/2022 19:08:39 - INFO - codeparrot_training - Step 4879: {'lr': 0.0004955748781757844, 'samples': 936960, 'steps': 4879, 'loss/train': 2.2505990266799927} 01/28/2022 19:08:44 - INFO - codeparrot_training - Step 4880: {'lr': 0.0004955718126821722, 'samples': 937152, 'steps': 4880, 'loss/train': 1.8742595314979553} 01/28/2022 19:08:48 - INFO - codeparrot_training - Step 4881: {'lr': 0.0004955687461366083, 'samples': 937344, 'steps': 4881, 'loss/train': 4.209123015403748} 01/28/2022 19:08:52 - INFO - codeparrot_training - Step 4882: {'lr': 0.000495565678539106, 'samples': 937536, 'steps': 4882, 'loss/train': 1.2894991636276245} 01/28/2022 19:08:56 - INFO - codeparrot_training - Step 4883: {'lr': 0.0004955626098896782, 'samples': 937728, 'steps': 4883, 'loss/train': 2.272673785686493} 01/28/2022 19:09:02 - INFO - codeparrot_training - Step 4884: {'lr': 0.0004955595401883381, 'samples': 937920, 'steps': 4884, 'loss/train': 1.779371738433838} 01/28/2022 19:09:06 - INFO - codeparrot_training - Step 4885: {'lr': 0.0004955564694350989, 'samples': 938112, 'steps': 4885, 'loss/train': 2.0291311740875244} 01/28/2022 19:09:10 - INFO - codeparrot_training - Step 4886: {'lr': 0.0004955533976299739, 'samples': 938304, 'steps': 4886, 'loss/train': 1.500319004058838} 01/28/2022 19:09:14 - INFO - codeparrot_training - Step 4887: {'lr': 0.000495550324772976, 'samples': 938496, 'steps': 4887, 'loss/train': 0.37245771288871765} 01/28/2022 19:09:19 - INFO - codeparrot_training - Step 4888: {'lr': 0.0004955472508641186, 'samples': 938688, 'steps': 4888, 'loss/train': 2.2891783118247986} 01/28/2022 19:09:25 - INFO - codeparrot_training - Step 4889: {'lr': 0.0004955441759034146, 'samples': 938880, 'steps': 4889, 'loss/train': 1.7923205494880676} 01/28/2022 19:09:29 - INFO - codeparrot_training - Step 4890: {'lr': 0.0004955410998908774, 'samples': 939072, 'steps': 4890, 'loss/train': 1.5926701426506042} 01/28/2022 19:09:33 - INFO - codeparrot_training - Step 4891: {'lr': 0.0004955380228265201, 'samples': 939264, 'steps': 4891, 'loss/train': 1.726300835609436} 01/28/2022 19:09:38 - INFO - codeparrot_training - Step 4892: {'lr': 0.0004955349447103559, 'samples': 939456, 'steps': 4892, 'loss/train': 1.037818193435669} 01/28/2022 19:09:42 - INFO - codeparrot_training - Step 4893: {'lr': 0.000495531865542398, 'samples': 939648, 'steps': 4893, 'loss/train': 3.0555245876312256} 01/28/2022 19:09:47 - INFO - codeparrot_training - Step 4894: {'lr': 0.0004955287853226594, 'samples': 939840, 'steps': 4894, 'loss/train': 2.1385996341705322} 01/28/2022 19:09:51 - INFO - codeparrot_training - Step 4895: {'lr': 0.0004955257040511534, 'samples': 940032, 'steps': 4895, 'loss/train': 2.2855830788612366} 01/28/2022 19:09:55 - INFO - codeparrot_training - Step 4896: {'lr': 0.0004955226217278934, 'samples': 940224, 'steps': 4896, 'loss/train': 2.2767693400382996} 01/28/2022 19:10:00 - INFO - codeparrot_training - Step 4897: {'lr': 0.0004955195383528926, 'samples': 940416, 'steps': 4897, 'loss/train': 1.6123334169387817} 01/28/2022 19:10:04 - INFO - codeparrot_training - Step 4898: {'lr': 0.0004955164539261638, 'samples': 940608, 'steps': 4898, 'loss/train': 1.9056419134140015} 01/28/2022 19:10:09 - INFO - codeparrot_training - Step 4899: {'lr': 0.0004955133684477205, 'samples': 940800, 'steps': 4899, 'loss/train': 1.6337944865226746} 01/28/2022 19:10:13 - INFO - codeparrot_training - Step 4900: {'lr': 0.000495510281917576, 'samples': 940992, 'steps': 4900, 'loss/train': 1.9086320400238037} 01/28/2022 19:10:18 - INFO - codeparrot_training - Step 4901: {'lr': 0.0004955071943357433, 'samples': 941184, 'steps': 4901, 'loss/train': 1.6769068837165833} 01/28/2022 19:10:22 - INFO - codeparrot_training - Step 4902: {'lr': 0.0004955041057022358, 'samples': 941376, 'steps': 4902, 'loss/train': 2.279150068759918} 01/28/2022 19:10:26 - INFO - codeparrot_training - Step 4903: {'lr': 0.0004955010160170667, 'samples': 941568, 'steps': 4903, 'loss/train': 1.4114436507225037} 01/28/2022 19:10:31 - INFO - codeparrot_training - Step 4904: {'lr': 0.0004954979252802491, 'samples': 941760, 'steps': 4904, 'loss/train': 1.8511893153190613} 01/28/2022 19:10:36 - INFO - codeparrot_training - Step 4905: {'lr': 0.0004954948334917965, 'samples': 941952, 'steps': 4905, 'loss/train': 2.461565673351288} 01/28/2022 19:10:40 - INFO - codeparrot_training - Step 4906: {'lr': 0.0004954917406517218, 'samples': 942144, 'steps': 4906, 'loss/train': 2.3447923064231873} 01/28/2022 19:10:44 - INFO - codeparrot_training - Step 4907: {'lr': 0.0004954886467600386, 'samples': 942336, 'steps': 4907, 'loss/train': 1.579925537109375} 01/28/2022 19:10:48 - INFO - codeparrot_training - Step 4908: {'lr': 0.0004954855518167599, 'samples': 942528, 'steps': 4908, 'loss/train': 2.1506075263023376} 01/28/2022 19:10:55 - INFO - codeparrot_training - Step 4909: {'lr': 0.000495482455821899, 'samples': 942720, 'steps': 4909, 'loss/train': 1.0382846295833588} 01/28/2022 19:10:59 - INFO - codeparrot_training - Step 4910: {'lr': 0.0004954793587754694, 'samples': 942912, 'steps': 4910, 'loss/train': 2.440907299518585} 01/28/2022 19:11:03 - INFO - codeparrot_training - Step 4911: {'lr': 0.000495476260677484, 'samples': 943104, 'steps': 4911, 'loss/train': 1.4886771440505981} 01/28/2022 19:11:07 - INFO - codeparrot_training - Step 4912: {'lr': 0.0004954731615279563, 'samples': 943296, 'steps': 4912, 'loss/train': 1.1934105455875397} 01/28/2022 19:11:12 - INFO - codeparrot_training - Step 4913: {'lr': 0.0004954700613268995, 'samples': 943488, 'steps': 4913, 'loss/train': 1.099903017282486} 01/28/2022 19:11:17 - INFO - codeparrot_training - Step 4914: {'lr': 0.0004954669600743269, 'samples': 943680, 'steps': 4914, 'loss/train': 1.8018688559532166} 01/28/2022 19:11:21 - INFO - codeparrot_training - Step 4915: {'lr': 0.0004954638577702519, 'samples': 943872, 'steps': 4915, 'loss/train': 1.6037612557411194} 01/28/2022 19:11:26 - INFO - codeparrot_training - Step 4916: {'lr': 0.0004954607544146875, 'samples': 944064, 'steps': 4916, 'loss/train': 1.5543222427368164} 01/28/2022 19:11:30 - INFO - codeparrot_training - Step 4917: {'lr': 0.0004954576500076472, 'samples': 944256, 'steps': 4917, 'loss/train': 2.0257043838500977} 01/28/2022 19:11:34 - INFO - codeparrot_training - Step 4918: {'lr': 0.0004954545445491444, 'samples': 944448, 'steps': 4918, 'loss/train': 1.6202369928359985} 01/28/2022 19:11:40 - INFO - codeparrot_training - Step 4919: {'lr': 0.0004954514380391921, 'samples': 944640, 'steps': 4919, 'loss/train': 1.0987283885478973} 01/28/2022 19:11:44 - INFO - codeparrot_training - Step 4920: {'lr': 0.0004954483304778039, 'samples': 944832, 'steps': 4920, 'loss/train': 1.9318088293075562} 01/28/2022 19:11:48 - INFO - codeparrot_training - Step 4921: {'lr': 0.0004954452218649929, 'samples': 945024, 'steps': 4921, 'loss/train': 1.6268582940101624} 01/28/2022 19:11:53 - INFO - codeparrot_training - Step 4922: {'lr': 0.0004954421122007727, 'samples': 945216, 'steps': 4922, 'loss/train': 2.13489693403244} 01/28/2022 19:11:57 - INFO - codeparrot_training - Step 4923: {'lr': 0.0004954390014851563, 'samples': 945408, 'steps': 4923, 'loss/train': 2.0071704983711243} 01/28/2022 19:12:03 - INFO - codeparrot_training - Step 4924: {'lr': 0.0004954358897181571, 'samples': 945600, 'steps': 4924, 'loss/train': 0.330019511282444} 01/28/2022 19:12:07 - INFO - codeparrot_training - Step 4925: {'lr': 0.0004954327768997885, 'samples': 945792, 'steps': 4925, 'loss/train': 1.9746562242507935} 01/28/2022 19:12:11 - INFO - codeparrot_training - Step 4926: {'lr': 0.0004954296630300638, 'samples': 945984, 'steps': 4926, 'loss/train': 2.2918304800987244} 01/28/2022 19:12:16 - INFO - codeparrot_training - Step 4927: {'lr': 0.0004954265481089965, 'samples': 946176, 'steps': 4927, 'loss/train': 2.058346152305603} 01/28/2022 19:12:20 - INFO - codeparrot_training - Step 4928: {'lr': 0.0004954234321365998, 'samples': 946368, 'steps': 4928, 'loss/train': 1.9695231914520264} 01/28/2022 19:12:25 - INFO - codeparrot_training - Step 4929: {'lr': 0.0004954203151128868, 'samples': 946560, 'steps': 4929, 'loss/train': 1.713664948940277} 01/28/2022 19:12:29 - INFO - codeparrot_training - Step 4930: {'lr': 0.0004954171970378713, 'samples': 946752, 'steps': 4930, 'loss/train': 2.2058104276657104} 01/28/2022 19:12:34 - INFO - codeparrot_training - Step 4931: {'lr': 0.0004954140779115664, 'samples': 946944, 'steps': 4931, 'loss/train': 1.8913494944572449} 01/28/2022 19:12:38 - INFO - codeparrot_training - Step 4932: {'lr': 0.0004954109577339856, 'samples': 947136, 'steps': 4932, 'loss/train': 0.9317089319229126} 01/28/2022 19:12:42 - INFO - codeparrot_training - Step 4933: {'lr': 0.0004954078365051421, 'samples': 947328, 'steps': 4933, 'loss/train': 2.1919530630111694} 01/28/2022 19:12:48 - INFO - codeparrot_training - Step 4934: {'lr': 0.0004954047142250494, 'samples': 947520, 'steps': 4934, 'loss/train': 2.4128390550613403} 01/28/2022 19:12:53 - INFO - codeparrot_training - Step 4935: {'lr': 0.0004954015908937208, 'samples': 947712, 'steps': 4935, 'loss/train': 2.030644655227661} 01/28/2022 19:12:57 - INFO - codeparrot_training - Step 4936: {'lr': 0.0004953984665111697, 'samples': 947904, 'steps': 4936, 'loss/train': 1.7337021231651306} 01/28/2022 19:13:01 - INFO - codeparrot_training - Step 4937: {'lr': 0.0004953953410774095, 'samples': 948096, 'steps': 4937, 'loss/train': 2.04888117313385} 01/28/2022 19:13:06 - INFO - codeparrot_training - Step 4938: {'lr': 0.0004953922145924535, 'samples': 948288, 'steps': 4938, 'loss/train': 2.045209050178528} 01/28/2022 19:13:11 - INFO - codeparrot_training - Step 4939: {'lr': 0.0004953890870563153, 'samples': 948480, 'steps': 4939, 'loss/train': 1.9438211917877197} 01/28/2022 19:13:15 - INFO - codeparrot_training - Step 4940: {'lr': 0.0004953859584690081, 'samples': 948672, 'steps': 4940, 'loss/train': 1.2705778777599335} 01/28/2022 19:13:19 - INFO - codeparrot_training - Step 4941: {'lr': 0.0004953828288305454, 'samples': 948864, 'steps': 4941, 'loss/train': 1.356948047876358} 01/28/2022 19:13:23 - INFO - codeparrot_training - Step 4942: {'lr': 0.0004953796981409407, 'samples': 949056, 'steps': 4942, 'loss/train': 2.3097753524780273} 01/28/2022 19:13:29 - INFO - codeparrot_training - Step 4943: {'lr': 0.0004953765664002071, 'samples': 949248, 'steps': 4943, 'loss/train': 2.463018536567688} 01/28/2022 19:13:33 - INFO - codeparrot_training - Step 4944: {'lr': 0.0004953734336083582, 'samples': 949440, 'steps': 4944, 'loss/train': 1.9005655646324158} 01/28/2022 19:13:38 - INFO - codeparrot_training - Step 4945: {'lr': 0.0004953702997654076, 'samples': 949632, 'steps': 4945, 'loss/train': 1.9394440054893494} 01/28/2022 19:13:42 - INFO - codeparrot_training - Step 4946: {'lr': 0.0004953671648713683, 'samples': 949824, 'steps': 4946, 'loss/train': 0.9154317677021027} 01/28/2022 19:13:46 - INFO - codeparrot_training - Step 4947: {'lr': 0.0004953640289262542, 'samples': 950016, 'steps': 4947, 'loss/train': 1.285516083240509} 01/28/2022 19:13:51 - INFO - codeparrot_training - Step 4948: {'lr': 0.0004953608919300784, 'samples': 950208, 'steps': 4948, 'loss/train': 0.8075573444366455} 01/28/2022 19:13:56 - INFO - codeparrot_training - Step 4949: {'lr': 0.0004953577538828546, 'samples': 950400, 'steps': 4949, 'loss/train': 1.9778836369514465} 01/28/2022 19:14:00 - INFO - codeparrot_training - Step 4950: {'lr': 0.0004953546147845959, 'samples': 950592, 'steps': 4950, 'loss/train': 2.1990031599998474} 01/28/2022 19:14:04 - INFO - codeparrot_training - Step 4951: {'lr': 0.0004953514746353161, 'samples': 950784, 'steps': 4951, 'loss/train': 2.2748336791992188} 01/28/2022 19:14:08 - INFO - codeparrot_training - Step 4952: {'lr': 0.0004953483334350283, 'samples': 950976, 'steps': 4952, 'loss/train': 1.354928344488144} 01/28/2022 19:14:15 - INFO - codeparrot_training - Step 4953: {'lr': 0.0004953451911837463, 'samples': 951168, 'steps': 4953, 'loss/train': 1.368551641702652} 01/28/2022 19:14:19 - INFO - codeparrot_training - Step 4954: {'lr': 0.0004953420478814834, 'samples': 951360, 'steps': 4954, 'loss/train': 1.3864031732082367} 01/28/2022 19:14:23 - INFO - codeparrot_training - Step 4955: {'lr': 0.000495338903528253, 'samples': 951552, 'steps': 4955, 'loss/train': 1.008187472820282} 01/28/2022 19:14:27 - INFO - codeparrot_training - Step 4956: {'lr': 0.0004953357581240686, 'samples': 951744, 'steps': 4956, 'loss/train': 1.004983127117157} 01/28/2022 19:14:32 - INFO - codeparrot_training - Step 4957: {'lr': 0.0004953326116689438, 'samples': 951936, 'steps': 4957, 'loss/train': 2.1897575855255127} 01/28/2022 19:14:36 - INFO - codeparrot_training - Step 4958: {'lr': 0.000495329464162892, 'samples': 952128, 'steps': 4958, 'loss/train': 2.2510803937911987} 01/28/2022 19:14:41 - INFO - codeparrot_training - Step 4959: {'lr': 0.0004953263156059266, 'samples': 952320, 'steps': 4959, 'loss/train': 1.456426978111267} 01/28/2022 19:14:45 - INFO - codeparrot_training - Step 4960: {'lr': 0.0004953231659980613, 'samples': 952512, 'steps': 4960, 'loss/train': 0.9214720129966736} 01/28/2022 19:14:50 - INFO - codeparrot_training - Step 4961: {'lr': 0.0004953200153393094, 'samples': 952704, 'steps': 4961, 'loss/train': 2.5490428805351257} 01/28/2022 19:14:54 - INFO - codeparrot_training - Step 4962: {'lr': 0.0004953168636296845, 'samples': 952896, 'steps': 4962, 'loss/train': 1.8680135607719421} 01/28/2022 19:14:58 - INFO - codeparrot_training - Step 4963: {'lr': 0.0004953137108691999, 'samples': 953088, 'steps': 4963, 'loss/train': 1.3226303458213806} 01/28/2022 19:15:03 - INFO - codeparrot_training - Step 4964: {'lr': 0.0004953105570578693, 'samples': 953280, 'steps': 4964, 'loss/train': 1.6365559101104736} 01/28/2022 19:15:07 - INFO - codeparrot_training - Step 4965: {'lr': 0.0004953074021957063, 'samples': 953472, 'steps': 4965, 'loss/train': 1.2778513133525848} 01/28/2022 19:15:12 - INFO - codeparrot_training - Step 4966: {'lr': 0.0004953042462827242, 'samples': 953664, 'steps': 4966, 'loss/train': 2.3353443145751953} 01/28/2022 19:15:16 - INFO - codeparrot_training - Step 4967: {'lr': 0.0004953010893189365, 'samples': 953856, 'steps': 4967, 'loss/train': 1.4399802088737488} 01/28/2022 19:15:20 - INFO - codeparrot_training - Step 4968: {'lr': 0.000495297931304357, 'samples': 954048, 'steps': 4968, 'loss/train': 1.1585785746574402} 01/28/2022 19:15:27 - INFO - codeparrot_training - Step 4969: {'lr': 0.000495294772238999, 'samples': 954240, 'steps': 4969, 'loss/train': 1.4138157963752747} 01/28/2022 19:15:31 - INFO - codeparrot_training - Step 4970: {'lr': 0.000495291612122876, 'samples': 954432, 'steps': 4970, 'loss/train': 1.5843542218208313} 01/28/2022 19:15:35 - INFO - codeparrot_training - Step 4971: {'lr': 0.0004952884509560017, 'samples': 954624, 'steps': 4971, 'loss/train': 0.87392458319664} 01/28/2022 19:15:39 - INFO - codeparrot_training - Step 4972: {'lr': 0.0004952852887383895, 'samples': 954816, 'steps': 4972, 'loss/train': 1.662691354751587} 01/28/2022 19:15:45 - INFO - codeparrot_training - Step 4973: {'lr': 0.0004952821254700531, 'samples': 955008, 'steps': 4973, 'loss/train': 1.777236521244049} 01/28/2022 19:15:49 - INFO - codeparrot_training - Step 4974: {'lr': 0.0004952789611510059, 'samples': 955200, 'steps': 4974, 'loss/train': 1.1198010742664337} 01/28/2022 19:15:53 - INFO - codeparrot_training - Step 4975: {'lr': 0.0004952757957812615, 'samples': 955392, 'steps': 4975, 'loss/train': 0.9586921334266663} 01/28/2022 19:15:57 - INFO - codeparrot_training - Step 4976: {'lr': 0.0004952726293608335, 'samples': 955584, 'steps': 4976, 'loss/train': 3.069520354270935} 01/28/2022 19:16:02 - INFO - codeparrot_training - Step 4977: {'lr': 0.0004952694618897354, 'samples': 955776, 'steps': 4977, 'loss/train': 2.419533133506775} 01/28/2022 19:16:08 - INFO - codeparrot_training - Step 4978: {'lr': 0.0004952662933679809, 'samples': 955968, 'steps': 4978, 'loss/train': 1.5383393168449402} 01/28/2022 19:16:12 - INFO - codeparrot_training - Step 4979: {'lr': 0.0004952631237955835, 'samples': 956160, 'steps': 4979, 'loss/train': 1.494953691959381} 01/28/2022 19:16:16 - INFO - codeparrot_training - Step 4980: {'lr': 0.0004952599531725567, 'samples': 956352, 'steps': 4980, 'loss/train': 2.2507063150405884} 01/28/2022 19:16:21 - INFO - codeparrot_training - Step 4981: {'lr': 0.0004952567814989141, 'samples': 956544, 'steps': 4981, 'loss/train': 1.993992805480957} 01/28/2022 19:16:25 - INFO - codeparrot_training - Step 4982: {'lr': 0.0004952536087746693, 'samples': 956736, 'steps': 4982, 'loss/train': 2.296269714832306} 01/28/2022 19:16:30 - INFO - codeparrot_training - Step 4983: {'lr': 0.000495250434999836, 'samples': 956928, 'steps': 4983, 'loss/train': 7.199799299240112} 01/28/2022 19:16:34 - INFO - codeparrot_training - Step 4984: {'lr': 0.0004952472601744277, 'samples': 957120, 'steps': 4984, 'loss/train': 0.934024840593338} 01/28/2022 19:16:39 - INFO - codeparrot_training - Step 4985: {'lr': 0.000495244084298458, 'samples': 957312, 'steps': 4985, 'loss/train': 2.096753418445587} 01/28/2022 19:16:43 - INFO - codeparrot_training - Step 4986: {'lr': 0.0004952409073719405, 'samples': 957504, 'steps': 4986, 'loss/train': 1.8753118515014648} 01/28/2022 19:16:47 - INFO - codeparrot_training - Step 4987: {'lr': 0.0004952377293948888, 'samples': 957696, 'steps': 4987, 'loss/train': 1.0382585227489471} 01/28/2022 19:16:52 - INFO - codeparrot_training - Step 4988: {'lr': 0.0004952345503673166, 'samples': 957888, 'steps': 4988, 'loss/train': 1.733128309249878} 01/28/2022 19:16:57 - INFO - codeparrot_training - Step 4989: {'lr': 0.0004952313702892375, 'samples': 958080, 'steps': 4989, 'loss/train': 1.4868644177913666} 01/28/2022 19:17:01 - INFO - codeparrot_training - Step 4990: {'lr': 0.0004952281891606649, 'samples': 958272, 'steps': 4990, 'loss/train': 1.4206523895263672} 01/28/2022 19:17:05 - INFO - codeparrot_training - Step 4991: {'lr': 0.0004952250069816127, 'samples': 958464, 'steps': 4991, 'loss/train': 1.7766863107681274} 01/28/2022 19:17:09 - INFO - codeparrot_training - Step 4992: {'lr': 0.0004952218237520945, 'samples': 958656, 'steps': 4992, 'loss/train': 2.199880599975586} 01/28/2022 19:17:15 - INFO - codeparrot_training - Step 4993: {'lr': 0.0004952186394721239, 'samples': 958848, 'steps': 4993, 'loss/train': 1.829498827457428} 01/28/2022 19:17:20 - INFO - codeparrot_training - Step 4994: {'lr': 0.0004952154541417144, 'samples': 959040, 'steps': 4994, 'loss/train': 2.7728809118270874} 01/28/2022 19:17:24 - INFO - codeparrot_training - Step 4995: {'lr': 0.0004952122677608798, 'samples': 959232, 'steps': 4995, 'loss/train': 2.8148422837257385} 01/28/2022 19:17:28 - INFO - codeparrot_training - Step 4996: {'lr': 0.0004952090803296337, 'samples': 959424, 'steps': 4996, 'loss/train': 0.9233001172542572} 01/28/2022 19:17:32 - INFO - codeparrot_training - Step 4997: {'lr': 0.0004952058918479899, 'samples': 959616, 'steps': 4997, 'loss/train': 2.081289231777191} 01/28/2022 19:17:38 - INFO - codeparrot_training - Step 4998: {'lr': 0.0004952027023159617, 'samples': 959808, 'steps': 4998, 'loss/train': 1.6866828203201294} 01/28/2022 19:17:42 - INFO - codeparrot_training - Step 4999: {'lr': 0.0004951995117335631, 'samples': 960000, 'steps': 4999, 'loss/train': 0.415274977684021} 01/28/2022 19:17:46 - INFO - codeparrot_training - Step 5000: {'lr': 0.0004951963201008077, 'samples': 960192, 'steps': 5000, 'loss/train': 1.361593633890152} 01/28/2022 19:17:51 - INFO - codeparrot_training - Step 5001: {'lr': 0.000495193127417709, 'samples': 960384, 'steps': 5001, 'loss/train': 0.1645497903227806} 01/28/2022 19:17:55 - INFO - codeparrot_training - Step 5002: {'lr': 0.0004951899336842809, 'samples': 960576, 'steps': 5002, 'loss/train': 0.6850746721029282} 01/28/2022 19:18:00 - INFO - codeparrot_training - Step 5003: {'lr': 0.0004951867389005369, 'samples': 960768, 'steps': 5003, 'loss/train': 1.0662437081336975} 01/28/2022 19:18:04 - INFO - codeparrot_training - Step 5004: {'lr': 0.0004951835430664908, 'samples': 960960, 'steps': 5004, 'loss/train': 1.8086321353912354} 01/28/2022 19:18:08 - INFO - codeparrot_training - Step 5005: {'lr': 0.0004951803461821562, 'samples': 961152, 'steps': 5005, 'loss/train': 1.7925805449485779} 01/28/2022 19:18:13 - INFO - codeparrot_training - Step 5006: {'lr': 0.0004951771482475469, 'samples': 961344, 'steps': 5006, 'loss/train': 1.2061094641685486} 01/28/2022 19:18:17 - INFO - codeparrot_training - Step 5007: {'lr': 0.0004951739492626766, 'samples': 961536, 'steps': 5007, 'loss/train': 1.4803897440433502} 01/28/2022 19:18:23 - INFO - codeparrot_training - Step 5008: {'lr': 0.0004951707492275589, 'samples': 961728, 'steps': 5008, 'loss/train': 1.578259527683258} 01/28/2022 19:18:28 - INFO - codeparrot_training - Step 5009: {'lr': 0.0004951675481422075, 'samples': 961920, 'steps': 5009, 'loss/train': 2.2890620827674866} 01/28/2022 19:18:32 - INFO - codeparrot_training - Step 5010: {'lr': 0.0004951643460066363, 'samples': 962112, 'steps': 5010, 'loss/train': 1.696587324142456} 01/28/2022 19:18:36 - INFO - codeparrot_training - Step 5011: {'lr': 0.0004951611428208589, 'samples': 962304, 'steps': 5011, 'loss/train': 1.9920040369033813} 01/28/2022 19:18:40 - INFO - codeparrot_training - Step 5012: {'lr': 0.0004951579385848889, 'samples': 962496, 'steps': 5012, 'loss/train': 2.375100553035736} 01/28/2022 19:18:46 - INFO - codeparrot_training - Step 5013: {'lr': 0.0004951547332987401, 'samples': 962688, 'steps': 5013, 'loss/train': 1.7013019323349} 01/28/2022 19:18:50 - INFO - codeparrot_training - Step 5014: {'lr': 0.0004951515269624265, 'samples': 962880, 'steps': 5014, 'loss/train': 2.7181928157806396} 01/28/2022 19:18:54 - INFO - codeparrot_training - Step 5015: {'lr': 0.0004951483195759614, 'samples': 963072, 'steps': 5015, 'loss/train': 1.7454553842544556} 01/28/2022 19:18:58 - INFO - codeparrot_training - Step 5016: {'lr': 0.0004951451111393588, 'samples': 963264, 'steps': 5016, 'loss/train': 2.359606683254242} 01/28/2022 19:19:03 - INFO - codeparrot_training - Step 5017: {'lr': 0.0004951419016526324, 'samples': 963456, 'steps': 5017, 'loss/train': 2.846675455570221} 01/28/2022 19:19:08 - INFO - codeparrot_training - Step 5018: {'lr': 0.0004951386911157959, 'samples': 963648, 'steps': 5018, 'loss/train': 1.7940330505371094} 01/28/2022 19:19:12 - INFO - codeparrot_training - Step 5019: {'lr': 0.0004951354795288631, 'samples': 963840, 'steps': 5019, 'loss/train': 2.0751700401306152} 01/28/2022 19:19:16 - INFO - codeparrot_training - Step 5020: {'lr': 0.0004951322668918477, 'samples': 964032, 'steps': 5020, 'loss/train': 2.2957749366760254} 01/28/2022 19:19:21 - INFO - codeparrot_training - Step 5021: {'lr': 0.0004951290532047637, 'samples': 964224, 'steps': 5021, 'loss/train': 1.5457334518432617} 01/28/2022 19:19:25 - INFO - codeparrot_training - Step 5022: {'lr': 0.0004951258384676244, 'samples': 964416, 'steps': 5022, 'loss/train': 1.7666172981262207} 01/28/2022 19:19:31 - INFO - codeparrot_training - Step 5023: {'lr': 0.0004951226226804441, 'samples': 964608, 'steps': 5023, 'loss/train': 1.0560647249221802} 01/28/2022 19:19:35 - INFO - codeparrot_training - Step 5024: {'lr': 0.0004951194058432361, 'samples': 964800, 'steps': 5024, 'loss/train': 1.191064327955246} 01/28/2022 19:19:39 - INFO - codeparrot_training - Step 5025: {'lr': 0.0004951161879560146, 'samples': 964992, 'steps': 5025, 'loss/train': 2.194292664527893} 01/28/2022 19:19:44 - INFO - codeparrot_training - Step 5026: {'lr': 0.000495112969018793, 'samples': 965184, 'steps': 5026, 'loss/train': 1.717703640460968} 01/28/2022 19:19:48 - INFO - codeparrot_training - Step 5027: {'lr': 0.0004951097490315853, 'samples': 965376, 'steps': 5027, 'loss/train': 1.0982399582862854} 01/28/2022 19:19:53 - INFO - codeparrot_training - Step 5028: {'lr': 0.0004951065279944054, 'samples': 965568, 'steps': 5028, 'loss/train': 2.112459361553192} 01/28/2022 19:19:57 - INFO - codeparrot_training - Step 5029: {'lr': 0.0004951033059072668, 'samples': 965760, 'steps': 5029, 'loss/train': 2.265931785106659} 01/28/2022 19:20:02 - INFO - codeparrot_training - Step 5030: {'lr': 0.0004951000827701836, 'samples': 965952, 'steps': 5030, 'loss/train': 1.3246483504772186} 01/28/2022 19:20:06 - INFO - codeparrot_training - Step 5031: {'lr': 0.0004950968585831694, 'samples': 966144, 'steps': 5031, 'loss/train': 0.8296273648738861} 01/28/2022 19:20:10 - INFO - codeparrot_training - Step 5032: {'lr': 0.0004950936333462381, 'samples': 966336, 'steps': 5032, 'loss/train': 0.9240882396697998} 01/28/2022 19:20:15 - INFO - codeparrot_training - Step 5033: {'lr': 0.0004950904070594036, 'samples': 966528, 'steps': 5033, 'loss/train': 1.0281144082546234} 01/28/2022 19:20:20 - INFO - codeparrot_training - Step 5034: {'lr': 0.0004950871797226795, 'samples': 966720, 'steps': 5034, 'loss/train': 2.0220739245414734} 01/28/2022 19:20:24 - INFO - codeparrot_training - Step 5035: {'lr': 0.0004950839513360798, 'samples': 966912, 'steps': 5035, 'loss/train': 1.1441034972667694} 01/28/2022 19:20:28 - INFO - codeparrot_training - Step 5036: {'lr': 0.0004950807218996182, 'samples': 967104, 'steps': 5036, 'loss/train': 1.4762544929981232} 01/28/2022 19:20:33 - INFO - codeparrot_training - Step 5037: {'lr': 0.0004950774914133086, 'samples': 967296, 'steps': 5037, 'loss/train': 1.8021265268325806} 01/28/2022 19:20:39 - INFO - codeparrot_training - Step 5038: {'lr': 0.0004950742598771649, 'samples': 967488, 'steps': 5038, 'loss/train': 2.0446808338165283} 01/28/2022 19:20:43 - INFO - codeparrot_training - Step 5039: {'lr': 0.0004950710272912009, 'samples': 967680, 'steps': 5039, 'loss/train': 1.3769340813159943} 01/28/2022 19:20:47 - INFO - codeparrot_training - Step 5040: {'lr': 0.0004950677936554305, 'samples': 967872, 'steps': 5040, 'loss/train': 2.133668839931488} 01/28/2022 19:20:52 - INFO - codeparrot_training - Step 5041: {'lr': 0.0004950645589698674, 'samples': 968064, 'steps': 5041, 'loss/train': 1.3035879135131836} 01/28/2022 19:20:56 - INFO - codeparrot_training - Step 5042: {'lr': 0.0004950613232345256, 'samples': 968256, 'steps': 5042, 'loss/train': 2.1601556539535522} 01/28/2022 19:21:02 - INFO - codeparrot_training - Step 5043: {'lr': 0.0004950580864494188, 'samples': 968448, 'steps': 5043, 'loss/train': 1.8693750500679016} 01/28/2022 19:21:06 - INFO - codeparrot_training - Step 5044: {'lr': 0.0004950548486145611, 'samples': 968640, 'steps': 5044, 'loss/train': 1.580689251422882} 01/28/2022 19:21:10 - INFO - codeparrot_training - Step 5045: {'lr': 0.0004950516097299662, 'samples': 968832, 'steps': 5045, 'loss/train': 0.7646372616291046} 01/28/2022 19:21:14 - INFO - codeparrot_training - Step 5046: {'lr': 0.000495048369795648, 'samples': 969024, 'steps': 5046, 'loss/train': 1.5394163131713867} 01/28/2022 19:21:20 - INFO - codeparrot_training - Step 5047: {'lr': 0.0004950451288116204, 'samples': 969216, 'steps': 5047, 'loss/train': 1.9810810089111328} 01/28/2022 19:21:24 - INFO - codeparrot_training - Step 5048: {'lr': 0.0004950418867778973, 'samples': 969408, 'steps': 5048, 'loss/train': 1.7422460317611694} 01/28/2022 19:21:28 - INFO - codeparrot_training - Step 5049: {'lr': 0.0004950386436944925, 'samples': 969600, 'steps': 5049, 'loss/train': 1.8199887871742249} 01/28/2022 19:21:32 - INFO - codeparrot_training - Step 5050: {'lr': 0.0004950353995614201, 'samples': 969792, 'steps': 5050, 'loss/train': 1.4718903601169586} 01/28/2022 19:21:37 - INFO - codeparrot_training - Step 5051: {'lr': 0.0004950321543786937, 'samples': 969984, 'steps': 5051, 'loss/train': 0.7761697769165039} 01/28/2022 19:21:43 - INFO - codeparrot_training - Step 5052: {'lr': 0.0004950289081463273, 'samples': 970176, 'steps': 5052, 'loss/train': 1.874461054801941} 01/28/2022 19:21:47 - INFO - codeparrot_training - Step 5053: {'lr': 0.0004950256608643351, 'samples': 970368, 'steps': 5053, 'loss/train': 1.8847750425338745} 01/28/2022 19:21:51 - INFO - codeparrot_training - Step 5054: {'lr': 0.0004950224125327307, 'samples': 970560, 'steps': 5054, 'loss/train': 1.725230097770691} 01/28/2022 19:21:55 - INFO - codeparrot_training - Step 5055: {'lr': 0.000495019163151528, 'samples': 970752, 'steps': 5055, 'loss/train': 1.2920324206352234} 01/28/2022 19:22:00 - INFO - codeparrot_training - Step 5056: {'lr': 0.0004950159127207411, 'samples': 970944, 'steps': 5056, 'loss/train': 1.8299360275268555} 01/28/2022 19:22:05 - INFO - codeparrot_training - Step 5057: {'lr': 0.0004950126612403838, 'samples': 971136, 'steps': 5057, 'loss/train': 1.8627979159355164} 01/28/2022 19:22:09 - INFO - codeparrot_training - Step 5058: {'lr': 0.00049500940871047, 'samples': 971328, 'steps': 5058, 'loss/train': 2.4872501492500305} 01/28/2022 19:22:13 - INFO - codeparrot_training - Step 5059: {'lr': 0.0004950061551310138, 'samples': 971520, 'steps': 5059, 'loss/train': 2.023664116859436} 01/28/2022 19:22:18 - INFO - codeparrot_training - Step 5060: {'lr': 0.0004950029005020289, 'samples': 971712, 'steps': 5060, 'loss/train': 1.9093505144119263} 01/28/2022 19:22:22 - INFO - codeparrot_training - Step 5061: {'lr': 0.0004949996448235294, 'samples': 971904, 'steps': 5061, 'loss/train': 2.001578450202942} 01/28/2022 19:22:27 - INFO - codeparrot_training - Step 5062: {'lr': 0.0004949963880955293, 'samples': 972096, 'steps': 5062, 'loss/train': 2.483379900455475} 01/28/2022 19:22:32 - INFO - codeparrot_training - Step 5063: {'lr': 0.0004949931303180424, 'samples': 972288, 'steps': 5063, 'loss/train': 2.9964609146118164} 01/28/2022 19:22:36 - INFO - codeparrot_training - Step 5064: {'lr': 0.0004949898714910828, 'samples': 972480, 'steps': 5064, 'loss/train': 2.441536009311676} 01/28/2022 19:22:40 - INFO - codeparrot_training - Step 5065: {'lr': 0.0004949866116146643, 'samples': 972672, 'steps': 5065, 'loss/train': 1.9071670174598694} 01/28/2022 19:22:44 - INFO - codeparrot_training - Step 5066: {'lr': 0.000494983350688801, 'samples': 972864, 'steps': 5066, 'loss/train': 2.215230882167816} 01/28/2022 19:22:50 - INFO - codeparrot_training - Step 5067: {'lr': 0.0004949800887135067, 'samples': 973056, 'steps': 5067, 'loss/train': 1.824753999710083} 01/28/2022 19:22:54 - INFO - codeparrot_training - Step 5068: {'lr': 0.0004949768256887956, 'samples': 973248, 'steps': 5068, 'loss/train': 7.613106966018677} 01/28/2022 19:22:58 - INFO - codeparrot_training - Step 5069: {'lr': 0.0004949735616146816, 'samples': 973440, 'steps': 5069, 'loss/train': 0.24754010885953903} 01/28/2022 19:23:02 - INFO - codeparrot_training - Step 5070: {'lr': 0.0004949702964911787, 'samples': 973632, 'steps': 5070, 'loss/train': 3.1015409231185913} 01/28/2022 19:23:07 - INFO - codeparrot_training - Step 5071: {'lr': 0.0004949670303183006, 'samples': 973824, 'steps': 5071, 'loss/train': 2.1193374395370483} 01/28/2022 19:23:11 - INFO - codeparrot_training - Step 5072: {'lr': 0.0004949637630960618, 'samples': 974016, 'steps': 5072, 'loss/train': 1.1435066163539886} 01/28/2022 19:23:17 - INFO - codeparrot_training - Step 5073: {'lr': 0.0004949604948244758, 'samples': 974208, 'steps': 5073, 'loss/train': 2.5058875679969788} 01/28/2022 19:23:21 - INFO - codeparrot_training - Step 5074: {'lr': 0.0004949572255035569, 'samples': 974400, 'steps': 5074, 'loss/train': 1.8771684765815735} 01/28/2022 19:23:25 - INFO - codeparrot_training - Step 5075: {'lr': 0.0004949539551333191, 'samples': 974592, 'steps': 5075, 'loss/train': 2.4672186970710754} 01/28/2022 19:23:30 - INFO - codeparrot_training - Step 5076: {'lr': 0.0004949506837137763, 'samples': 974784, 'steps': 5076, 'loss/train': 2.1207869052886963} 01/28/2022 19:23:34 - INFO - codeparrot_training - Step 5077: {'lr': 0.0004949474112449424, 'samples': 974976, 'steps': 5077, 'loss/train': 1.9653639793395996} 01/28/2022 19:23:39 - INFO - codeparrot_training - Step 5078: {'lr': 0.0004949441377268318, 'samples': 975168, 'steps': 5078, 'loss/train': 1.6707138419151306} 01/28/2022 19:23:44 - INFO - codeparrot_training - Step 5079: {'lr': 0.0004949408631594582, 'samples': 975360, 'steps': 5079, 'loss/train': 2.043950378894806} 01/28/2022 19:23:48 - INFO - codeparrot_training - Step 5080: {'lr': 0.0004949375875428357, 'samples': 975552, 'steps': 5080, 'loss/train': 2.124589741230011} 01/28/2022 19:23:52 - INFO - codeparrot_training - Step 5081: {'lr': 0.0004949343108769784, 'samples': 975744, 'steps': 5081, 'loss/train': 0.24565760046243668} 01/28/2022 19:23:56 - INFO - codeparrot_training - Step 5082: {'lr': 0.0004949310331619002, 'samples': 975936, 'steps': 5082, 'loss/train': 2.365135610103607} 01/28/2022 19:24:03 - INFO - codeparrot_training - Step 5083: {'lr': 0.0004949277543976153, 'samples': 976128, 'steps': 5083, 'loss/train': 2.677825391292572} 01/28/2022 19:24:07 - INFO - codeparrot_training - Step 5084: {'lr': 0.0004949244745841377, 'samples': 976320, 'steps': 5084, 'loss/train': 1.8629291653633118} 01/28/2022 19:24:12 - INFO - codeparrot_training - Step 5085: {'lr': 0.0004949211937214814, 'samples': 976512, 'steps': 5085, 'loss/train': 0.7684786319732666} 01/28/2022 19:24:16 - INFO - codeparrot_training - Step 5086: {'lr': 0.0004949179118096604, 'samples': 976704, 'steps': 5086, 'loss/train': 1.55228590965271} 01/28/2022 19:24:20 - INFO - codeparrot_training - Step 5087: {'lr': 0.0004949146288486889, 'samples': 976896, 'steps': 5087, 'loss/train': 2.4592453837394714} 01/28/2022 19:24:25 - INFO - codeparrot_training - Step 5088: {'lr': 0.0004949113448385809, 'samples': 977088, 'steps': 5088, 'loss/train': 2.00001060962677} 01/28/2022 19:24:30 - INFO - codeparrot_training - Step 5089: {'lr': 0.0004949080597793505, 'samples': 977280, 'steps': 5089, 'loss/train': 0.7711944580078125} 01/28/2022 19:24:34 - INFO - codeparrot_training - Step 5090: {'lr': 0.0004949047736710116, 'samples': 977472, 'steps': 5090, 'loss/train': 1.7209028005599976} 01/28/2022 19:24:38 - INFO - codeparrot_training - Step 5091: {'lr': 0.0004949014865135786, 'samples': 977664, 'steps': 5091, 'loss/train': 2.3150208592414856} 01/28/2022 19:24:42 - INFO - codeparrot_training - Step 5092: {'lr': 0.0004948981983070652, 'samples': 977856, 'steps': 5092, 'loss/train': 1.8795830011367798} 01/28/2022 19:24:48 - INFO - codeparrot_training - Step 5093: {'lr': 0.0004948949090514858, 'samples': 978048, 'steps': 5093, 'loss/train': 2.0479751229286194} 01/28/2022 19:24:52 - INFO - codeparrot_training - Step 5094: {'lr': 0.0004948916187468544, 'samples': 978240, 'steps': 5094, 'loss/train': 2.0464466214179993} 01/28/2022 19:24:56 - INFO - codeparrot_training - Step 5095: {'lr': 0.000494888327393185, 'samples': 978432, 'steps': 5095, 'loss/train': 1.842247724533081} 01/28/2022 19:25:00 - INFO - codeparrot_training - Step 5096: {'lr': 0.0004948850349904919, 'samples': 978624, 'steps': 5096, 'loss/train': 2.183404862880707} 01/28/2022 19:25:05 - INFO - codeparrot_training - Step 5097: {'lr': 0.000494881741538789, 'samples': 978816, 'steps': 5097, 'loss/train': 3.0621964931488037} 01/28/2022 19:25:11 - INFO - codeparrot_training - Step 5098: {'lr': 0.0004948784470380904, 'samples': 979008, 'steps': 5098, 'loss/train': 1.9946520924568176} 01/28/2022 19:25:15 - INFO - codeparrot_training - Step 5099: {'lr': 0.0004948751514884103, 'samples': 979200, 'steps': 5099, 'loss/train': 2.156618356704712} 01/28/2022 19:25:20 - INFO - codeparrot_training - Step 5100: {'lr': 0.0004948718548897628, 'samples': 979392, 'steps': 5100, 'loss/train': 1.135040670633316} 01/28/2022 19:25:24 - INFO - codeparrot_training - Step 5101: {'lr': 0.0004948685572421621, 'samples': 979584, 'steps': 5101, 'loss/train': 1.7748767137527466} 01/28/2022 19:25:28 - INFO - codeparrot_training - Step 5102: {'lr': 0.0004948652585456222, 'samples': 979776, 'steps': 5102, 'loss/train': 2.910196602344513} 01/28/2022 19:25:33 - INFO - codeparrot_training - Step 5103: {'lr': 0.0004948619588001574, 'samples': 979968, 'steps': 5103, 'loss/train': 1.9178103804588318} 01/28/2022 19:25:38 - INFO - codeparrot_training - Step 5104: {'lr': 0.0004948586580057816, 'samples': 980160, 'steps': 5104, 'loss/train': 1.472118079662323} 01/28/2022 19:25:42 - INFO - codeparrot_training - Step 5105: {'lr': 0.0004948553561625091, 'samples': 980352, 'steps': 5105, 'loss/train': 1.4210083186626434} 01/28/2022 19:25:46 - INFO - codeparrot_training - Step 5106: {'lr': 0.000494852053270354, 'samples': 980544, 'steps': 5106, 'loss/train': 1.5824593305587769} 01/28/2022 19:25:50 - INFO - codeparrot_training - Step 5107: {'lr': 0.0004948487493293305, 'samples': 980736, 'steps': 5107, 'loss/train': 1.903372585773468} 01/28/2022 19:25:57 - INFO - codeparrot_training - Step 5108: {'lr': 0.0004948454443394527, 'samples': 980928, 'steps': 5108, 'loss/train': 1.9363101124763489} 01/28/2022 19:26:01 - INFO - codeparrot_training - Step 5109: {'lr': 0.0004948421383007347, 'samples': 981120, 'steps': 5109, 'loss/train': 1.4690378308296204} 01/28/2022 19:26:05 - INFO - codeparrot_training - Step 5110: {'lr': 0.0004948388312131908, 'samples': 981312, 'steps': 5110, 'loss/train': 2.7684770822525024} 01/28/2022 19:26:09 - INFO - codeparrot_training - Step 5111: {'lr': 0.0004948355230768349, 'samples': 981504, 'steps': 5111, 'loss/train': 2.4903284311294556} 01/28/2022 19:26:14 - INFO - codeparrot_training - Step 5112: {'lr': 0.0004948322138916816, 'samples': 981696, 'steps': 5112, 'loss/train': 2.0127511024475098} 01/28/2022 19:26:19 - INFO - codeparrot_training - Step 5113: {'lr': 0.0004948289036577447, 'samples': 981888, 'steps': 5113, 'loss/train': 2.1120187640190125} 01/28/2022 19:26:23 - INFO - codeparrot_training - Step 5114: {'lr': 0.0004948255923750385, 'samples': 982080, 'steps': 5114, 'loss/train': 7.524540424346924} 01/28/2022 19:26:27 - INFO - codeparrot_training - Step 5115: {'lr': 0.0004948222800435773, 'samples': 982272, 'steps': 5115, 'loss/train': 1.8806878924369812} 01/28/2022 19:26:32 - INFO - codeparrot_training - Step 5116: {'lr': 0.0004948189666633752, 'samples': 982464, 'steps': 5116, 'loss/train': 2.4038026928901672} 01/28/2022 19:26:36 - INFO - codeparrot_training - Step 5117: {'lr': 0.0004948156522344463, 'samples': 982656, 'steps': 5117, 'loss/train': 0.7915661931037903} 01/28/2022 19:26:41 - INFO - codeparrot_training - Step 5118: {'lr': 0.0004948123367568049, 'samples': 982848, 'steps': 5118, 'loss/train': 1.6620388627052307} 01/28/2022 19:26:45 - INFO - codeparrot_training - Step 5119: {'lr': 0.0004948090202304652, 'samples': 983040, 'steps': 5119, 'loss/train': 1.2451496422290802} 01/28/2022 19:26:50 - INFO - codeparrot_training - Step 5120: {'lr': 0.0004948057026554415, 'samples': 983232, 'steps': 5120, 'loss/train': 1.9309369325637817} 01/28/2022 19:26:54 - INFO - codeparrot_training - Step 5121: {'lr': 0.0004948023840317477, 'samples': 983424, 'steps': 5121, 'loss/train': 0.4008413851261139} 01/28/2022 19:26:58 - INFO - codeparrot_training - Step 5122: {'lr': 0.0004947990643593983, 'samples': 983616, 'steps': 5122, 'loss/train': 2.5840176343917847} 01/28/2022 19:27:03 - INFO - codeparrot_training - Step 5123: {'lr': 0.0004947957436384076, 'samples': 983808, 'steps': 5123, 'loss/train': 2.4384777545928955} 01/28/2022 19:27:08 - INFO - codeparrot_training - Step 5124: {'lr': 0.0004947924218687894, 'samples': 984000, 'steps': 5124, 'loss/train': 1.9153794050216675} 01/28/2022 19:27:12 - INFO - codeparrot_training - Step 5125: {'lr': 0.0004947890990505585, 'samples': 984192, 'steps': 5125, 'loss/train': 0.8294419348239899} 01/28/2022 19:27:16 - INFO - codeparrot_training - Step 5126: {'lr': 0.0004947857751837286, 'samples': 984384, 'steps': 5126, 'loss/train': 1.959950029850006} 01/28/2022 19:27:20 - INFO - codeparrot_training - Step 5127: {'lr': 0.0004947824502683142, 'samples': 984576, 'steps': 5127, 'loss/train': 3.0210299491882324} 01/28/2022 19:27:27 - INFO - codeparrot_training - Step 5128: {'lr': 0.0004947791243043296, 'samples': 984768, 'steps': 5128, 'loss/train': 0.3757064491510391} 01/28/2022 19:27:31 - INFO - codeparrot_training - Step 5129: {'lr': 0.0004947757972917889, 'samples': 984960, 'steps': 5129, 'loss/train': 2.025180995464325} 01/28/2022 19:27:35 - INFO - codeparrot_training - Step 5130: {'lr': 0.0004947724692307064, 'samples': 985152, 'steps': 5130, 'loss/train': 2.460550010204315} 01/28/2022 19:27:39 - INFO - codeparrot_training - Step 5131: {'lr': 0.0004947691401210963, 'samples': 985344, 'steps': 5131, 'loss/train': 2.404591977596283} 01/28/2022 19:27:43 - INFO - codeparrot_training - Step 5132: {'lr': 0.0004947658099629731, 'samples': 985536, 'steps': 5132, 'loss/train': 2.829473912715912} 01/28/2022 19:27:49 - INFO - codeparrot_training - Step 5133: {'lr': 0.0004947624787563507, 'samples': 985728, 'steps': 5133, 'loss/train': 2.1343401074409485} 01/28/2022 19:27:54 - INFO - codeparrot_training - Step 5134: {'lr': 0.0004947591465012436, 'samples': 985920, 'steps': 5134, 'loss/train': 2.1696245670318604} 01/28/2022 19:27:58 - INFO - codeparrot_training - Step 5135: {'lr': 0.0004947558131976661, 'samples': 986112, 'steps': 5135, 'loss/train': 0.5274547934532166} 01/28/2022 19:28:02 - INFO - codeparrot_training - Step 5136: {'lr': 0.0004947524788456324, 'samples': 986304, 'steps': 5136, 'loss/train': 0.724431037902832} 01/28/2022 19:28:06 - INFO - codeparrot_training - Step 5137: {'lr': 0.0004947491434451569, 'samples': 986496, 'steps': 5137, 'loss/train': 2.4185691475868225} 01/28/2022 19:28:12 - INFO - codeparrot_training - Step 5138: {'lr': 0.0004947458069962537, 'samples': 986688, 'steps': 5138, 'loss/train': 2.253922998905182} 01/28/2022 19:28:16 - INFO - codeparrot_training - Step 5139: {'lr': 0.0004947424694989371, 'samples': 986880, 'steps': 5139, 'loss/train': 1.822223424911499} 01/28/2022 19:28:20 - INFO - codeparrot_training - Step 5140: {'lr': 0.0004947391309532216, 'samples': 987072, 'steps': 5140, 'loss/train': 1.5179357528686523} 01/28/2022 19:28:24 - INFO - codeparrot_training - Step 5141: {'lr': 0.0004947357913591213, 'samples': 987264, 'steps': 5141, 'loss/train': 2.0842804312705994} 01/28/2022 19:28:29 - INFO - codeparrot_training - Step 5142: {'lr': 0.0004947324507166505, 'samples': 987456, 'steps': 5142, 'loss/train': 1.831163227558136} 01/28/2022 19:28:35 - INFO - codeparrot_training - Step 5143: {'lr': 0.0004947291090258238, 'samples': 987648, 'steps': 5143, 'loss/train': 1.9249995946884155} 01/28/2022 19:28:39 - INFO - codeparrot_training - Step 5144: {'lr': 0.0004947257662866551, 'samples': 987840, 'steps': 5144, 'loss/train': 2.7025145888328552} 01/28/2022 19:28:44 - INFO - codeparrot_training - Step 5145: {'lr': 0.0004947224224991591, 'samples': 988032, 'steps': 5145, 'loss/train': 0.4471743553876877} 01/28/2022 19:28:48 - INFO - codeparrot_training - Step 5146: {'lr': 0.0004947190776633499, 'samples': 988224, 'steps': 5146, 'loss/train': 2.60390442609787} 01/28/2022 19:28:52 - INFO - codeparrot_training - Step 5147: {'lr': 0.0004947157317792418, 'samples': 988416, 'steps': 5147, 'loss/train': 1.7990421652793884} 01/28/2022 19:28:58 - INFO - codeparrot_training - Step 5148: {'lr': 0.0004947123848468493, 'samples': 988608, 'steps': 5148, 'loss/train': 1.9503387808799744} 01/28/2022 19:29:02 - INFO - codeparrot_training - Step 5149: {'lr': 0.0004947090368661866, 'samples': 988800, 'steps': 5149, 'loss/train': 1.9459009766578674} 01/28/2022 19:29:06 - INFO - codeparrot_training - Step 5150: {'lr': 0.0004947056878372681, 'samples': 988992, 'steps': 5150, 'loss/train': 1.535616159439087} 01/28/2022 19:29:10 - INFO - codeparrot_training - Step 5151: {'lr': 0.0004947023377601082, 'samples': 989184, 'steps': 5151, 'loss/train': 1.8471938967704773} 01/28/2022 19:29:15 - INFO - codeparrot_training - Step 5152: {'lr': 0.0004946989866347211, 'samples': 989376, 'steps': 5152, 'loss/train': 1.513242244720459} 01/28/2022 19:29:20 - INFO - codeparrot_training - Step 5153: {'lr': 0.0004946956344611212, 'samples': 989568, 'steps': 5153, 'loss/train': 1.8656253218650818} 01/28/2022 19:29:25 - INFO - codeparrot_training - Step 5154: {'lr': 0.000494692281239323, 'samples': 989760, 'steps': 5154, 'loss/train': 2.148451566696167} 01/28/2022 19:29:29 - INFO - codeparrot_training - Step 5155: {'lr': 0.0004946889269693408, 'samples': 989952, 'steps': 5155, 'loss/train': 1.4972282946109772} 01/28/2022 19:29:33 - INFO - codeparrot_training - Step 5156: {'lr': 0.0004946855716511888, 'samples': 990144, 'steps': 5156, 'loss/train': 1.6730316281318665} 01/28/2022 19:29:37 - INFO - codeparrot_training - Step 5157: {'lr': 0.0004946822152848816, 'samples': 990336, 'steps': 5157, 'loss/train': 0.7885200083255768} 01/28/2022 19:29:43 - INFO - codeparrot_training - Step 5158: {'lr': 0.0004946788578704335, 'samples': 990528, 'steps': 5158, 'loss/train': 2.3521331548690796} 01/28/2022 19:29:47 - INFO - codeparrot_training - Step 5159: {'lr': 0.0004946754994078588, 'samples': 990720, 'steps': 5159, 'loss/train': 3.391951560974121} 01/28/2022 19:29:52 - INFO - codeparrot_training - Step 5160: {'lr': 0.000494672139897172, 'samples': 990912, 'steps': 5160, 'loss/train': 1.8596178889274597} 01/28/2022 19:29:56 - INFO - codeparrot_training - Step 5161: {'lr': 0.0004946687793383874, 'samples': 991104, 'steps': 5161, 'loss/train': 2.5087738037109375} 01/28/2022 19:30:00 - INFO - codeparrot_training - Step 5162: {'lr': 0.0004946654177315194, 'samples': 991296, 'steps': 5162, 'loss/train': 1.7929480075836182} 01/28/2022 19:30:06 - INFO - codeparrot_training - Step 5163: {'lr': 0.0004946620550765826, 'samples': 991488, 'steps': 5163, 'loss/train': 1.3849167823791504} 01/28/2022 19:30:10 - INFO - codeparrot_training - Step 5164: {'lr': 0.0004946586913735911, 'samples': 991680, 'steps': 5164, 'loss/train': 2.3183562755584717} 01/28/2022 19:30:14 - INFO - codeparrot_training - Step 5165: {'lr': 0.0004946553266225595, 'samples': 991872, 'steps': 5165, 'loss/train': 1.8203274607658386} 01/28/2022 19:30:19 - INFO - codeparrot_training - Step 5166: {'lr': 0.0004946519608235022, 'samples': 992064, 'steps': 5166, 'loss/train': 1.2249955236911774} 01/28/2022 19:30:25 - INFO - codeparrot_training - Step 5167: {'lr': 0.0004946485939764336, 'samples': 992256, 'steps': 5167, 'loss/train': 1.4525141716003418} 01/28/2022 19:30:29 - INFO - codeparrot_training - Step 5168: {'lr': 0.000494645226081368, 'samples': 992448, 'steps': 5168, 'loss/train': 1.594765841960907} 01/28/2022 19:30:33 - INFO - codeparrot_training - Step 5169: {'lr': 0.00049464185713832, 'samples': 992640, 'steps': 5169, 'loss/train': 2.7858447432518005} 01/28/2022 19:30:37 - INFO - codeparrot_training - Step 5170: {'lr': 0.000494638487147304, 'samples': 992832, 'steps': 5170, 'loss/train': 1.2131820023059845} 01/28/2022 19:30:42 - INFO - codeparrot_training - Step 5171: {'lr': 0.0004946351161083344, 'samples': 993024, 'steps': 5171, 'loss/train': 1.9395076632499695} 01/28/2022 19:30:47 - INFO - codeparrot_training - Step 5172: {'lr': 0.0004946317440214257, 'samples': 993216, 'steps': 5172, 'loss/train': 1.0556532740592957} 01/28/2022 19:30:51 - INFO - codeparrot_training - Step 5173: {'lr': 0.000494628370886592, 'samples': 993408, 'steps': 5173, 'loss/train': 1.3513891696929932} 01/28/2022 19:30:55 - INFO - codeparrot_training - Step 5174: {'lr': 0.0004946249967038483, 'samples': 993600, 'steps': 5174, 'loss/train': 1.967602550983429} 01/28/2022 19:31:00 - INFO - codeparrot_training - Step 5175: {'lr': 0.0004946216214732088, 'samples': 993792, 'steps': 5175, 'loss/train': 2.0528120398521423} 01/28/2022 19:31:04 - INFO - codeparrot_training - Step 5176: {'lr': 0.0004946182451946878, 'samples': 993984, 'steps': 5176, 'loss/train': 0.9663905203342438} 01/28/2022 19:31:09 - INFO - codeparrot_training - Step 5177: {'lr': 0.0004946148678683001, 'samples': 994176, 'steps': 5177, 'loss/train': 1.395135372877121} 01/28/2022 19:31:13 - INFO - codeparrot_training - Step 5178: {'lr': 0.0004946114894940599, 'samples': 994368, 'steps': 5178, 'loss/train': 1.3825467824935913} 01/28/2022 19:31:18 - INFO - codeparrot_training - Step 5179: {'lr': 0.0004946081100719817, 'samples': 994560, 'steps': 5179, 'loss/train': 1.3722838461399078} 01/28/2022 19:31:22 - INFO - codeparrot_training - Step 5180: {'lr': 0.00049460472960208, 'samples': 994752, 'steps': 5180, 'loss/train': 1.473355919122696} 01/28/2022 19:31:26 - INFO - codeparrot_training - Step 5181: {'lr': 0.0004946013480843694, 'samples': 994944, 'steps': 5181, 'loss/train': 2.032467484474182} 01/28/2022 19:31:31 - INFO - codeparrot_training - Step 5182: {'lr': 0.0004945979655188642, 'samples': 995136, 'steps': 5182, 'loss/train': 2.5697184205055237} 01/28/2022 19:31:36 - INFO - codeparrot_training - Step 5183: {'lr': 0.0004945945819055791, 'samples': 995328, 'steps': 5183, 'loss/train': 2.3347785472869873} 01/28/2022 19:31:40 - INFO - codeparrot_training - Step 5184: {'lr': 0.0004945911972445284, 'samples': 995520, 'steps': 5184, 'loss/train': 1.6072912216186523} 01/28/2022 19:31:44 - INFO - codeparrot_training - Step 5185: {'lr': 0.0004945878115357267, 'samples': 995712, 'steps': 5185, 'loss/train': 0.6293092221021652} 01/28/2022 19:31:49 - INFO - codeparrot_training - Step 5186: {'lr': 0.0004945844247791886, 'samples': 995904, 'steps': 5186, 'loss/train': 1.7698500752449036} 01/28/2022 19:31:54 - INFO - codeparrot_training - Step 5187: {'lr': 0.0004945810369749283, 'samples': 996096, 'steps': 5187, 'loss/train': 2.240746021270752} 01/28/2022 19:31:59 - INFO - codeparrot_training - Step 5188: {'lr': 0.0004945776481229605, 'samples': 996288, 'steps': 5188, 'loss/train': 1.6575536727905273} 01/28/2022 19:32:03 - INFO - codeparrot_training - Step 5189: {'lr': 0.0004945742582232999, 'samples': 996480, 'steps': 5189, 'loss/train': 1.6774125695228577} 01/28/2022 19:32:07 - INFO - codeparrot_training - Step 5190: {'lr': 0.0004945708672759606, 'samples': 996672, 'steps': 5190, 'loss/train': 1.5093829035758972} 01/28/2022 19:32:11 - INFO - codeparrot_training - Step 5191: {'lr': 0.0004945674752809575, 'samples': 996864, 'steps': 5191, 'loss/train': 1.7823706269264221} 01/28/2022 19:32:17 - INFO - codeparrot_training - Step 5192: {'lr': 0.000494564082238305, 'samples': 997056, 'steps': 5192, 'loss/train': 2.040696680545807} 01/28/2022 19:32:21 - INFO - codeparrot_training - Step 5193: {'lr': 0.0004945606881480176, 'samples': 997248, 'steps': 5193, 'loss/train': 1.89966881275177} 01/28/2022 19:32:26 - INFO - codeparrot_training - Step 5194: {'lr': 0.0004945572930101098, 'samples': 997440, 'steps': 5194, 'loss/train': 1.95679771900177} 01/28/2022 19:32:30 - INFO - codeparrot_training - Step 5195: {'lr': 0.0004945538968245964, 'samples': 997632, 'steps': 5195, 'loss/train': 2.3166661262512207} 01/28/2022 19:32:34 - INFO - codeparrot_training - Step 5196: {'lr': 0.0004945504995914917, 'samples': 997824, 'steps': 5196, 'loss/train': 2.152831792831421} 01/28/2022 19:32:39 - INFO - codeparrot_training - Step 5197: {'lr': 0.0004945471013108102, 'samples': 998016, 'steps': 5197, 'loss/train': 2.05087673664093} 01/28/2022 19:32:43 - INFO - codeparrot_training - Step 5198: {'lr': 0.0004945437019825668, 'samples': 998208, 'steps': 5198, 'loss/train': 2.146640717983246} 01/28/2022 19:32:48 - INFO - codeparrot_training - Step 5199: {'lr': 0.0004945403016067756, 'samples': 998400, 'steps': 5199, 'loss/train': 1.1242139339447021} 01/28/2022 19:32:52 - INFO - codeparrot_training - Step 5200: {'lr': 0.0004945369001834514, 'samples': 998592, 'steps': 5200, 'loss/train': 1.8533159494400024} 01/28/2022 19:32:56 - INFO - codeparrot_training - Step 5201: {'lr': 0.0004945334977126089, 'samples': 998784, 'steps': 5201, 'loss/train': 0.4972397983074188} 01/28/2022 19:33:02 - INFO - codeparrot_training - Step 5202: {'lr': 0.0004945300941942624, 'samples': 998976, 'steps': 5202, 'loss/train': 2.1297488808631897} 01/28/2022 19:33:07 - INFO - codeparrot_training - Step 5203: {'lr': 0.0004945266896284268, 'samples': 999168, 'steps': 5203, 'loss/train': 1.9853897094726562} 01/28/2022 19:33:11 - INFO - codeparrot_training - Step 5204: {'lr': 0.0004945232840151164, 'samples': 999360, 'steps': 5204, 'loss/train': 2.212350904941559} 01/28/2022 19:33:15 - INFO - codeparrot_training - Step 5205: {'lr': 0.0004945198773543459, 'samples': 999552, 'steps': 5205, 'loss/train': 1.5481941103935242} 01/28/2022 19:33:19 - INFO - codeparrot_training - Step 5206: {'lr': 0.0004945164696461299, 'samples': 999744, 'steps': 5206, 'loss/train': 2.6626306772232056} 01/28/2022 19:33:25 - INFO - codeparrot_training - Step 5207: {'lr': 0.000494513060890483, 'samples': 999936, 'steps': 5207, 'loss/train': 1.5099021792411804} 01/28/2022 19:33:29 - INFO - codeparrot_training - Step 5208: {'lr': 0.0004945096510874197, 'samples': 1000128, 'steps': 5208, 'loss/train': 1.505727767944336} 01/28/2022 19:33:33 - INFO - codeparrot_training - Step 5209: {'lr': 0.0004945062402369548, 'samples': 1000320, 'steps': 5209, 'loss/train': 2.39948308467865} 01/28/2022 19:33:37 - INFO - codeparrot_training - Step 5210: {'lr': 0.0004945028283391028, 'samples': 1000512, 'steps': 5210, 'loss/train': 0.9828777015209198} 01/28/2022 19:33:42 - INFO - codeparrot_training - Step 5211: {'lr': 0.0004944994153938783, 'samples': 1000704, 'steps': 5211, 'loss/train': 2.740654706954956} 01/28/2022 19:33:48 - INFO - codeparrot_training - Step 5212: {'lr': 0.0004944960014012959, 'samples': 1000896, 'steps': 5212, 'loss/train': 2.2221193313598633} 01/28/2022 19:33:52 - INFO - codeparrot_training - Step 5213: {'lr': 0.0004944925863613704, 'samples': 1001088, 'steps': 5213, 'loss/train': 1.8521592020988464} 01/28/2022 19:33:56 - INFO - codeparrot_training - Step 5214: {'lr': 0.0004944891702741161, 'samples': 1001280, 'steps': 5214, 'loss/train': 1.4479941129684448} 01/28/2022 19:34:01 - INFO - codeparrot_training - Step 5215: {'lr': 0.0004944857531395479, 'samples': 1001472, 'steps': 5215, 'loss/train': 2.1184800267219543} 01/28/2022 19:34:05 - INFO - codeparrot_training - Step 5216: {'lr': 0.0004944823349576805, 'samples': 1001664, 'steps': 5216, 'loss/train': 0.14584679156541824} 01/28/2022 19:34:10 - INFO - codeparrot_training - Step 5217: {'lr': 0.0004944789157285283, 'samples': 1001856, 'steps': 5217, 'loss/train': 2.164377272129059} 01/28/2022 19:34:14 - INFO - codeparrot_training - Step 5218: {'lr': 0.0004944754954521061, 'samples': 1002048, 'steps': 5218, 'loss/train': 1.3422649204730988} 01/28/2022 19:34:19 - INFO - codeparrot_training - Step 5219: {'lr': 0.0004944720741284285, 'samples': 1002240, 'steps': 5219, 'loss/train': 0.39746855199337006} 01/28/2022 19:34:23 - INFO - codeparrot_training - Step 5220: {'lr': 0.00049446865175751, 'samples': 1002432, 'steps': 5220, 'loss/train': 2.2967217564582825} 01/28/2022 19:34:27 - INFO - codeparrot_training - Step 5221: {'lr': 0.0004944652283393656, 'samples': 1002624, 'steps': 5221, 'loss/train': 1.8099598288536072} 01/28/2022 19:34:32 - INFO - codeparrot_training - Step 5222: {'lr': 0.0004944618038740098, 'samples': 1002816, 'steps': 5222, 'loss/train': 1.904653787612915} 01/28/2022 19:34:37 - INFO - codeparrot_training - Step 5223: {'lr': 0.0004944583783614571, 'samples': 1003008, 'steps': 5223, 'loss/train': 1.6545463800430298} 01/28/2022 19:34:41 - INFO - codeparrot_training - Step 5224: {'lr': 0.0004944549518017225, 'samples': 1003200, 'steps': 5224, 'loss/train': 0.9443089663982391} 01/28/2022 19:34:45 - INFO - codeparrot_training - Step 5225: {'lr': 0.0004944515241948204, 'samples': 1003392, 'steps': 5225, 'loss/train': 2.284703493118286} 01/28/2022 19:34:49 - INFO - codeparrot_training - Step 5226: {'lr': 0.0004944480955407656, 'samples': 1003584, 'steps': 5226, 'loss/train': 0.8277635872364044} 01/28/2022 19:34:56 - INFO - codeparrot_training - Step 5227: {'lr': 0.0004944446658395728, 'samples': 1003776, 'steps': 5227, 'loss/train': 2.2885254621505737} 01/28/2022 19:35:00 - INFO - codeparrot_training - Step 5228: {'lr': 0.0004944412350912567, 'samples': 1003968, 'steps': 5228, 'loss/train': 2.023768723011017} 01/28/2022 19:35:04 - INFO - codeparrot_training - Step 5229: {'lr': 0.000494437803295832, 'samples': 1004160, 'steps': 5229, 'loss/train': 1.6939149498939514} 01/28/2022 19:35:08 - INFO - codeparrot_training - Step 5230: {'lr': 0.0004944343704533133, 'samples': 1004352, 'steps': 5230, 'loss/train': 2.270028233528137} 01/28/2022 19:35:12 - INFO - codeparrot_training - Step 5231: {'lr': 0.0004944309365637154, 'samples': 1004544, 'steps': 5231, 'loss/train': 0.998669296503067} 01/28/2022 19:35:18 - INFO - codeparrot_training - Step 5232: {'lr': 0.000494427501627053, 'samples': 1004736, 'steps': 5232, 'loss/train': 1.8771075010299683} 01/28/2022 19:35:22 - INFO - codeparrot_training - Step 5233: {'lr': 0.0004944240656433407, 'samples': 1004928, 'steps': 5233, 'loss/train': 1.2262314856052399} 01/28/2022 19:35:26 - INFO - codeparrot_training - Step 5234: {'lr': 0.0004944206286125935, 'samples': 1005120, 'steps': 5234, 'loss/train': 0.8112762272357941} 01/28/2022 19:35:30 - INFO - codeparrot_training - Step 5235: {'lr': 0.0004944171905348258, 'samples': 1005312, 'steps': 5235, 'loss/train': 2.6885915994644165} 01/28/2022 19:35:35 - INFO - codeparrot_training - Step 5236: {'lr': 0.0004944137514100525, 'samples': 1005504, 'steps': 5236, 'loss/train': 1.5185580253601074} 01/28/2022 19:35:41 - INFO - codeparrot_training - Step 5237: {'lr': 0.0004944103112382883, 'samples': 1005696, 'steps': 5237, 'loss/train': 0.2328147366642952} 01/28/2022 19:35:45 - INFO - codeparrot_training - Step 5238: {'lr': 0.0004944068700195479, 'samples': 1005888, 'steps': 5238, 'loss/train': 0.9030638337135315} 01/28/2022 19:35:49 - INFO - codeparrot_training - Step 5239: {'lr': 0.0004944034277538462, 'samples': 1006080, 'steps': 5239, 'loss/train': 2.051572859287262} 01/28/2022 19:35:53 - INFO - codeparrot_training - Step 5240: {'lr': 0.0004943999844411977, 'samples': 1006272, 'steps': 5240, 'loss/train': 1.8884865045547485} 01/28/2022 19:35:58 - INFO - codeparrot_training - Step 5241: {'lr': 0.0004943965400816173, 'samples': 1006464, 'steps': 5241, 'loss/train': 1.5949076414108276} 01/28/2022 19:36:03 - INFO - codeparrot_training - Step 5242: {'lr': 0.0004943930946751197, 'samples': 1006656, 'steps': 5242, 'loss/train': 1.8740419149398804} 01/28/2022 19:36:07 - INFO - codeparrot_training - Step 5243: {'lr': 0.0004943896482217197, 'samples': 1006848, 'steps': 5243, 'loss/train': 1.9013893604278564} 01/28/2022 19:36:11 - INFO - codeparrot_training - Step 5244: {'lr': 0.0004943862007214322, 'samples': 1007040, 'steps': 5244, 'loss/train': 1.118051290512085} 01/28/2022 19:36:16 - INFO - codeparrot_training - Step 5245: {'lr': 0.0004943827521742716, 'samples': 1007232, 'steps': 5245, 'loss/train': 1.5395121574401855} 01/28/2022 19:36:20 - INFO - codeparrot_training - Step 5246: {'lr': 0.000494379302580253, 'samples': 1007424, 'steps': 5246, 'loss/train': 1.687314212322235} 01/28/2022 19:36:26 - INFO - codeparrot_training - Step 5247: {'lr': 0.000494375851939391, 'samples': 1007616, 'steps': 5247, 'loss/train': 1.462194174528122} 01/28/2022 19:36:30 - INFO - codeparrot_training - Step 5248: {'lr': 0.0004943724002517005, 'samples': 1007808, 'steps': 5248, 'loss/train': 2.1483108401298523} 01/28/2022 19:36:35 - INFO - codeparrot_training - Step 5249: {'lr': 0.0004943689475171962, 'samples': 1008000, 'steps': 5249, 'loss/train': 1.3415307104587555} 01/28/2022 19:36:39 - INFO - codeparrot_training - Step 5250: {'lr': 0.000494365493735893, 'samples': 1008192, 'steps': 5250, 'loss/train': 1.7313939929008484} 01/28/2022 19:36:43 - INFO - codeparrot_training - Step 5251: {'lr': 0.0004943620389078055, 'samples': 1008384, 'steps': 5251, 'loss/train': 0.9752267003059387} 01/28/2022 19:36:48 - INFO - codeparrot_training - Step 5252: {'lr': 0.0004943585830329487, 'samples': 1008576, 'steps': 5252, 'loss/train': 0.8294287025928497} 01/28/2022 19:36:53 - INFO - codeparrot_training - Step 5253: {'lr': 0.0004943551261113373, 'samples': 1008768, 'steps': 5253, 'loss/train': 1.7982911467552185} 01/28/2022 19:36:57 - INFO - codeparrot_training - Step 5254: {'lr': 0.0004943516681429861, 'samples': 1008960, 'steps': 5254, 'loss/train': 0.8453375101089478} 01/28/2022 19:37:01 - INFO - codeparrot_training - Step 5255: {'lr': 0.0004943482091279101, 'samples': 1009152, 'steps': 5255, 'loss/train': 1.9097726941108704} 01/28/2022 19:37:05 - INFO - codeparrot_training - Step 5256: {'lr': 0.0004943447490661238, 'samples': 1009344, 'steps': 5256, 'loss/train': 1.6965340375900269} 01/28/2022 19:37:11 - INFO - codeparrot_training - Step 5257: {'lr': 0.0004943412879576422, 'samples': 1009536, 'steps': 5257, 'loss/train': 1.9056481719017029} 01/28/2022 19:37:16 - INFO - codeparrot_training - Step 5258: {'lr': 0.0004943378258024802, 'samples': 1009728, 'steps': 5258, 'loss/train': 1.588899314403534} 01/28/2022 19:37:20 - INFO - codeparrot_training - Step 5259: {'lr': 0.0004943343626006524, 'samples': 1009920, 'steps': 5259, 'loss/train': 1.6973837614059448} 01/28/2022 19:37:24 - INFO - codeparrot_training - Step 5260: {'lr': 0.000494330898352174, 'samples': 1010112, 'steps': 5260, 'loss/train': 1.6470907330513} 01/28/2022 19:37:28 - INFO - codeparrot_training - Step 5261: {'lr': 0.0004943274330570594, 'samples': 1010304, 'steps': 5261, 'loss/train': 1.7088659405708313} 01/28/2022 19:37:33 - INFO - codeparrot_training - Step 5262: {'lr': 0.0004943239667153237, 'samples': 1010496, 'steps': 5262, 'loss/train': 2.118900954723358} 01/28/2022 19:37:38 - INFO - codeparrot_training - Step 5263: {'lr': 0.0004943204993269818, 'samples': 1010688, 'steps': 5263, 'loss/train': 1.8171443939208984} 01/28/2022 19:37:42 - INFO - codeparrot_training - Step 5264: {'lr': 0.0004943170308920483, 'samples': 1010880, 'steps': 5264, 'loss/train': 2.2411844730377197} 01/28/2022 19:37:46 - INFO - codeparrot_training - Step 5265: {'lr': 0.0004943135614105384, 'samples': 1011072, 'steps': 5265, 'loss/train': 1.890578269958496} 01/28/2022 19:37:51 - INFO - codeparrot_training - Step 5266: {'lr': 0.0004943100908824667, 'samples': 1011264, 'steps': 5266, 'loss/train': 1.8032877445220947} 01/28/2022 19:37:56 - INFO - codeparrot_training - Step 5267: {'lr': 0.0004943066193078482, 'samples': 1011456, 'steps': 5267, 'loss/train': 1.1803985238075256} 01/28/2022 19:38:00 - INFO - codeparrot_training - Step 5268: {'lr': 0.0004943031466866976, 'samples': 1011648, 'steps': 5268, 'loss/train': 1.9089152812957764} 01/28/2022 19:38:04 - INFO - codeparrot_training - Step 5269: {'lr': 0.00049429967301903, 'samples': 1011840, 'steps': 5269, 'loss/train': 1.8218767046928406} 01/28/2022 19:38:09 - INFO - codeparrot_training - Step 5270: {'lr': 0.0004942961983048601, 'samples': 1012032, 'steps': 5270, 'loss/train': 1.8298010230064392} 01/28/2022 19:38:13 - INFO - codeparrot_training - Step 5271: {'lr': 0.0004942927225442029, 'samples': 1012224, 'steps': 5271, 'loss/train': 2.0203641057014465} 01/28/2022 19:38:18 - INFO - codeparrot_training - Step 5272: {'lr': 0.0004942892457370732, 'samples': 1012416, 'steps': 5272, 'loss/train': 1.2058332860469818} 01/28/2022 19:38:23 - INFO - codeparrot_training - Step 5273: {'lr': 0.000494285767883486, 'samples': 1012608, 'steps': 5273, 'loss/train': 1.8424934148788452} 01/28/2022 19:38:27 - INFO - codeparrot_training - Step 5274: {'lr': 0.0004942822889834562, 'samples': 1012800, 'steps': 5274, 'loss/train': 2.1038219332695007} 01/28/2022 19:38:31 - INFO - codeparrot_training - Step 5275: {'lr': 0.0004942788090369985, 'samples': 1012992, 'steps': 5275, 'loss/train': 1.245940625667572} 01/28/2022 19:38:35 - INFO - codeparrot_training - Step 5276: {'lr': 0.0004942753280441281, 'samples': 1013184, 'steps': 5276, 'loss/train': 1.9712272882461548} 01/28/2022 19:38:42 - INFO - codeparrot_training - Step 5277: {'lr': 0.0004942718460048596, 'samples': 1013376, 'steps': 5277, 'loss/train': 2.0484648942947388} 01/28/2022 19:38:47 - INFO - codeparrot_training - Step 5278: {'lr': 0.0004942683629192082, 'samples': 1013568, 'steps': 5278, 'loss/train': 1.579810380935669} 01/28/2022 19:38:51 - INFO - codeparrot_training - Step 5279: {'lr': 0.0004942648787871886, 'samples': 1013760, 'steps': 5279, 'loss/train': 2.6779063940048218} 01/28/2022 19:38:55 - INFO - codeparrot_training - Step 5280: {'lr': 0.000494261393608816, 'samples': 1013952, 'steps': 5280, 'loss/train': 1.7333012223243713} 01/28/2022 19:38:59 - INFO - codeparrot_training - Step 5281: {'lr': 0.0004942579073841049, 'samples': 1014144, 'steps': 5281, 'loss/train': 1.978298306465149} 01/28/2022 19:39:05 - INFO - codeparrot_training - Step 5282: {'lr': 0.0004942544201130706, 'samples': 1014336, 'steps': 5282, 'loss/train': 1.6975200176239014} 01/28/2022 19:39:09 - INFO - codeparrot_training - Step 5283: {'lr': 0.000494250931795728, 'samples': 1014528, 'steps': 5283, 'loss/train': 1.2195044159889221} 01/28/2022 19:39:13 - INFO - codeparrot_training - Step 5284: {'lr': 0.0004942474424320919, 'samples': 1014720, 'steps': 5284, 'loss/train': 1.1663104891777039} 01/28/2022 19:39:17 - INFO - codeparrot_training - Step 5285: {'lr': 0.0004942439520221774, 'samples': 1014912, 'steps': 5285, 'loss/train': 2.0777058005332947} 01/28/2022 19:39:21 - INFO - codeparrot_training - Step 5286: {'lr': 0.0004942404605659991, 'samples': 1015104, 'steps': 5286, 'loss/train': 2.033539831638336} 01/28/2022 19:39:27 - INFO - codeparrot_training - Step 5287: {'lr': 0.0004942369680635724, 'samples': 1015296, 'steps': 5287, 'loss/train': 1.6526952981948853} 01/28/2022 19:39:31 - INFO - codeparrot_training - Step 5288: {'lr': 0.0004942334745149122, 'samples': 1015488, 'steps': 5288, 'loss/train': 2.186837911605835} 01/28/2022 19:39:36 - INFO - codeparrot_training - Step 5289: {'lr': 0.0004942299799200332, 'samples': 1015680, 'steps': 5289, 'loss/train': 1.757222056388855} 01/28/2022 19:39:40 - INFO - codeparrot_training - Step 5290: {'lr': 0.0004942264842789506, 'samples': 1015872, 'steps': 5290, 'loss/train': 1.0925669074058533} 01/28/2022 19:39:44 - INFO - codeparrot_training - Step 5291: {'lr': 0.0004942229875916792, 'samples': 1016064, 'steps': 5291, 'loss/train': 1.8544090390205383} 01/28/2022 19:39:50 - INFO - codeparrot_training - Step 5292: {'lr': 0.0004942194898582341, 'samples': 1016256, 'steps': 5292, 'loss/train': 2.5947314500808716} 01/28/2022 19:39:55 - INFO - codeparrot_training - Step 5293: {'lr': 0.0004942159910786303, 'samples': 1016448, 'steps': 5293, 'loss/train': 2.1164719462394714} 01/28/2022 19:39:59 - INFO - codeparrot_training - Step 5294: {'lr': 0.0004942124912528827, 'samples': 1016640, 'steps': 5294, 'loss/train': 1.8420880436897278} 01/28/2022 19:40:03 - INFO - codeparrot_training - Step 5295: {'lr': 0.0004942089903810064, 'samples': 1016832, 'steps': 5295, 'loss/train': 1.0966133773326874} 01/28/2022 19:40:07 - INFO - codeparrot_training - Step 5296: {'lr': 0.0004942054884630162, 'samples': 1017024, 'steps': 5296, 'loss/train': 1.7782711386680603} 01/28/2022 19:40:13 - INFO - codeparrot_training - Step 5297: {'lr': 0.0004942019854989274, 'samples': 1017216, 'steps': 5297, 'loss/train': 1.9599087238311768} 01/28/2022 19:40:17 - INFO - codeparrot_training - Step 5298: {'lr': 0.0004941984814887546, 'samples': 1017408, 'steps': 5298, 'loss/train': 2.4417548775672913} 01/28/2022 19:40:21 - INFO - codeparrot_training - Step 5299: {'lr': 0.0004941949764325133, 'samples': 1017600, 'steps': 5299, 'loss/train': 2.016587734222412} 01/28/2022 19:40:25 - INFO - codeparrot_training - Step 5300: {'lr': 0.0004941914703302181, 'samples': 1017792, 'steps': 5300, 'loss/train': 1.2473958134651184} 01/28/2022 19:40:30 - INFO - codeparrot_training - Step 5301: {'lr': 0.0004941879631818843, 'samples': 1017984, 'steps': 5301, 'loss/train': 2.112401783466339} 01/28/2022 19:40:35 - INFO - codeparrot_training - Step 5302: {'lr': 0.0004941844549875267, 'samples': 1018176, 'steps': 5302, 'loss/train': 1.5030303597450256} 01/28/2022 19:40:39 - INFO - codeparrot_training - Step 5303: {'lr': 0.0004941809457471605, 'samples': 1018368, 'steps': 5303, 'loss/train': 2.321302056312561} 01/28/2022 19:40:43 - INFO - codeparrot_training - Step 5304: {'lr': 0.0004941774354608006, 'samples': 1018560, 'steps': 5304, 'loss/train': 2.070923924446106} 01/28/2022 19:40:48 - INFO - codeparrot_training - Step 5305: {'lr': 0.0004941739241284621, 'samples': 1018752, 'steps': 5305, 'loss/train': 1.503393530845642} 01/28/2022 19:40:52 - INFO - codeparrot_training - Step 5306: {'lr': 0.0004941704117501601, 'samples': 1018944, 'steps': 5306, 'loss/train': 1.715615451335907} 01/28/2022 19:40:58 - INFO - codeparrot_training - Step 5307: {'lr': 0.0004941668983259095, 'samples': 1019136, 'steps': 5307, 'loss/train': 0.21725358814001083} 01/28/2022 19:41:02 - INFO - codeparrot_training - Step 5308: {'lr': 0.0004941633838557256, 'samples': 1019328, 'steps': 5308, 'loss/train': 2.0232263803482056} 01/28/2022 19:41:07 - INFO - codeparrot_training - Step 5309: {'lr': 0.0004941598683396232, 'samples': 1019520, 'steps': 5309, 'loss/train': 2.145451784133911} 01/28/2022 19:41:11 - INFO - codeparrot_training - Step 5310: {'lr': 0.0004941563517776174, 'samples': 1019712, 'steps': 5310, 'loss/train': 1.9018730521202087} 01/28/2022 19:41:15 - INFO - codeparrot_training - Step 5311: {'lr': 0.0004941528341697234, 'samples': 1019904, 'steps': 5311, 'loss/train': 1.6089298725128174} 01/28/2022 19:41:21 - INFO - codeparrot_training - Step 5312: {'lr': 0.0004941493155159562, 'samples': 1020096, 'steps': 5312, 'loss/train': 7.240490198135376} 01/28/2022 19:41:25 - INFO - codeparrot_training - Step 5313: {'lr': 0.0004941457958163308, 'samples': 1020288, 'steps': 5313, 'loss/train': 1.332964450120926} 01/28/2022 19:41:29 - INFO - codeparrot_training - Step 5314: {'lr': 0.0004941422750708623, 'samples': 1020480, 'steps': 5314, 'loss/train': 1.9467272758483887} 01/28/2022 19:41:34 - INFO - codeparrot_training - Step 5315: {'lr': 0.0004941387532795659, 'samples': 1020672, 'steps': 5315, 'loss/train': 1.4939576983451843} 01/28/2022 19:41:38 - INFO - codeparrot_training - Step 5316: {'lr': 0.0004941352304424566, 'samples': 1020864, 'steps': 5316, 'loss/train': 1.6551493406295776} 01/28/2022 19:41:44 - INFO - codeparrot_training - Step 5317: {'lr': 0.0004941317065595495, 'samples': 1021056, 'steps': 5317, 'loss/train': 2.4848583340644836} 01/28/2022 19:41:48 - INFO - codeparrot_training - Step 5318: {'lr': 0.0004941281816308596, 'samples': 1021248, 'steps': 5318, 'loss/train': 2.2584872245788574} 01/28/2022 19:41:52 - INFO - codeparrot_training - Step 5319: {'lr': 0.0004941246556564021, 'samples': 1021440, 'steps': 5319, 'loss/train': 2.0055951476097107} 01/28/2022 19:41:57 - INFO - codeparrot_training - Step 5320: {'lr': 0.0004941211286361922, 'samples': 1021632, 'steps': 5320, 'loss/train': 1.9835878014564514} 01/28/2022 19:42:01 - INFO - codeparrot_training - Step 5321: {'lr': 0.0004941176005702448, 'samples': 1021824, 'steps': 5321, 'loss/train': 2.2340288758277893} 01/28/2022 19:42:06 - INFO - codeparrot_training - Step 5322: {'lr': 0.0004941140714585752, 'samples': 1022016, 'steps': 5322, 'loss/train': 1.7202674746513367} 01/28/2022 19:42:11 - INFO - codeparrot_training - Step 5323: {'lr': 0.0004941105413011984, 'samples': 1022208, 'steps': 5323, 'loss/train': 1.9389606714248657} 01/28/2022 19:42:15 - INFO - codeparrot_training - Step 5324: {'lr': 0.0004941070100981295, 'samples': 1022400, 'steps': 5324, 'loss/train': 1.8786449432373047} 01/28/2022 19:42:19 - INFO - codeparrot_training - Step 5325: {'lr': 0.0004941034778493837, 'samples': 1022592, 'steps': 5325, 'loss/train': 1.0126223266124725} 01/28/2022 19:42:23 - INFO - codeparrot_training - Step 5326: {'lr': 0.0004940999445549762, 'samples': 1022784, 'steps': 5326, 'loss/train': 1.4807537198066711} 01/28/2022 19:42:29 - INFO - codeparrot_training - Step 5327: {'lr': 0.0004940964102149219, 'samples': 1022976, 'steps': 5327, 'loss/train': 0.8088995218276978} 01/28/2022 19:42:33 - INFO - codeparrot_training - Step 5328: {'lr': 0.0004940928748292363, 'samples': 1023168, 'steps': 5328, 'loss/train': 2.509723663330078} 01/28/2022 19:42:37 - INFO - codeparrot_training - Step 5329: {'lr': 0.0004940893383979341, 'samples': 1023360, 'steps': 5329, 'loss/train': 2.2538865208625793} 01/28/2022 19:42:41 - INFO - codeparrot_training - Step 5330: {'lr': 0.0004940858009210308, 'samples': 1023552, 'steps': 5330, 'loss/train': 2.0624329447746277} 01/28/2022 19:42:46 - INFO - codeparrot_training - Step 5331: {'lr': 0.0004940822623985414, 'samples': 1023744, 'steps': 5331, 'loss/train': 0.9667960703372955} 01/28/2022 19:42:52 - INFO - codeparrot_training - Step 5332: {'lr': 0.0004940787228304811, 'samples': 1023936, 'steps': 5332, 'loss/train': 0.8687771558761597} 01/28/2022 19:42:56 - INFO - codeparrot_training - Step 5333: {'lr': 0.0004940751822168651, 'samples': 1024128, 'steps': 5333, 'loss/train': 1.4232876598834991} 01/28/2022 19:43:00 - INFO - codeparrot_training - Step 5334: {'lr': 0.0004940716405577086, 'samples': 1024320, 'steps': 5334, 'loss/train': 2.665561616420746} 01/28/2022 19:43:05 - INFO - codeparrot_training - Step 5335: {'lr': 0.0004940680978530265, 'samples': 1024512, 'steps': 5335, 'loss/train': 1.4631505608558655} 01/28/2022 19:43:09 - INFO - codeparrot_training - Step 5336: {'lr': 0.0004940645541028343, 'samples': 1024704, 'steps': 5336, 'loss/train': 1.7942714095115662} 01/28/2022 19:43:14 - INFO - codeparrot_training - Step 5337: {'lr': 0.0004940610093071469, 'samples': 1024896, 'steps': 5337, 'loss/train': 1.9364136457443237} 01/28/2022 19:43:18 - INFO - codeparrot_training - Step 5338: {'lr': 0.0004940574634659798, 'samples': 1025088, 'steps': 5338, 'loss/train': 1.9691539406776428} 01/28/2022 19:43:23 - INFO - codeparrot_training - Step 5339: {'lr': 0.000494053916579348, 'samples': 1025280, 'steps': 5339, 'loss/train': 1.8156035542488098} 01/28/2022 19:43:27 - INFO - codeparrot_training - Step 5340: {'lr': 0.0004940503686472667, 'samples': 1025472, 'steps': 5340, 'loss/train': 1.4933525919914246} 01/28/2022 19:43:31 - INFO - codeparrot_training - Step 5341: {'lr': 0.0004940468196697511, 'samples': 1025664, 'steps': 5341, 'loss/train': 1.6111562848091125} 01/28/2022 19:43:36 - INFO - codeparrot_training - Step 5342: {'lr': 0.0004940432696468164, 'samples': 1025856, 'steps': 5342, 'loss/train': 2.192757546901703} 01/28/2022 19:43:41 - INFO - codeparrot_training - Step 5343: {'lr': 0.0004940397185784778, 'samples': 1026048, 'steps': 5343, 'loss/train': 2.3075737953186035} 01/28/2022 19:43:45 - INFO - codeparrot_training - Step 5344: {'lr': 0.0004940361664647506, 'samples': 1026240, 'steps': 5344, 'loss/train': 1.7806915640830994} 01/28/2022 19:43:49 - INFO - codeparrot_training - Step 5345: {'lr': 0.0004940326133056499, 'samples': 1026432, 'steps': 5345, 'loss/train': 1.9882718324661255} 01/28/2022 19:43:54 - INFO - codeparrot_training - Step 5346: {'lr': 0.000494029059101191, 'samples': 1026624, 'steps': 5346, 'loss/train': 2.121248424053192} 01/28/2022 19:43:59 - INFO - codeparrot_training - Step 5347: {'lr': 0.0004940255038513891, 'samples': 1026816, 'steps': 5347, 'loss/train': 1.267199456691742} 01/28/2022 19:44:04 - INFO - codeparrot_training - Step 5348: {'lr': 0.0004940219475562593, 'samples': 1027008, 'steps': 5348, 'loss/train': 2.1723893880844116} 01/28/2022 19:44:08 - INFO - codeparrot_training - Step 5349: {'lr': 0.0004940183902158172, 'samples': 1027200, 'steps': 5349, 'loss/train': 0.9526903331279755} 01/28/2022 19:44:12 - INFO - codeparrot_training - Step 5350: {'lr': 0.0004940148318300777, 'samples': 1027392, 'steps': 5350, 'loss/train': 1.2409481406211853} 01/28/2022 19:44:16 - INFO - codeparrot_training - Step 5351: {'lr': 0.0004940112723990561, 'samples': 1027584, 'steps': 5351, 'loss/train': 2.509082794189453} 01/28/2022 19:44:22 - INFO - codeparrot_training - Step 5352: {'lr': 0.0004940077119227678, 'samples': 1027776, 'steps': 5352, 'loss/train': 1.4298806190490723} 01/28/2022 19:44:27 - INFO - codeparrot_training - Step 5353: {'lr': 0.0004940041504012279, 'samples': 1027968, 'steps': 5353, 'loss/train': 1.0986594557762146} 01/28/2022 19:44:31 - INFO - codeparrot_training - Step 5354: {'lr': 0.0004940005878344517, 'samples': 1028160, 'steps': 5354, 'loss/train': 1.8753581643104553} 01/28/2022 19:44:35 - INFO - codeparrot_training - Step 5355: {'lr': 0.0004939970242224544, 'samples': 1028352, 'steps': 5355, 'loss/train': 1.8483749628067017} 01/28/2022 19:44:39 - INFO - codeparrot_training - Step 5356: {'lr': 0.0004939934595652513, 'samples': 1028544, 'steps': 5356, 'loss/train': 1.9680195450782776} 01/28/2022 19:44:45 - INFO - codeparrot_training - Step 5357: {'lr': 0.0004939898938628578, 'samples': 1028736, 'steps': 5357, 'loss/train': 2.236750602722168} 01/28/2022 19:44:49 - INFO - codeparrot_training - Step 5358: {'lr': 0.000493986327115289, 'samples': 1028928, 'steps': 5358, 'loss/train': 2.3931080102920532} 01/28/2022 19:44:53 - INFO - codeparrot_training - Step 5359: {'lr': 0.0004939827593225602, 'samples': 1029120, 'steps': 5359, 'loss/train': 2.0060527324676514} 01/28/2022 19:44:57 - INFO - codeparrot_training - Step 5360: {'lr': 0.0004939791904846869, 'samples': 1029312, 'steps': 5360, 'loss/train': 1.9108797311782837} 01/28/2022 19:45:02 - INFO - codeparrot_training - Step 5361: {'lr': 0.0004939756206016841, 'samples': 1029504, 'steps': 5361, 'loss/train': 2.3991073966026306} 01/28/2022 19:45:09 - INFO - codeparrot_training - Step 5362: {'lr': 0.0004939720496735672, 'samples': 1029696, 'steps': 5362, 'loss/train': 1.9782428741455078} 01/28/2022 19:45:13 - INFO - codeparrot_training - Step 5363: {'lr': 0.0004939684777003516, 'samples': 1029888, 'steps': 5363, 'loss/train': 1.48750439286232} 01/28/2022 19:45:17 - INFO - codeparrot_training - Step 5364: {'lr': 0.0004939649046820524, 'samples': 1030080, 'steps': 5364, 'loss/train': 1.9401184916496277} 01/28/2022 19:45:21 - INFO - codeparrot_training - Step 5365: {'lr': 0.0004939613306186851, 'samples': 1030272, 'steps': 5365, 'loss/train': 1.0963197648525238} 01/28/2022 19:45:26 - INFO - codeparrot_training - Step 5366: {'lr': 0.0004939577555102649, 'samples': 1030464, 'steps': 5366, 'loss/train': 2.9184540510177612} 01/28/2022 19:45:31 - INFO - codeparrot_training - Step 5367: {'lr': 0.0004939541793568072, 'samples': 1030656, 'steps': 5367, 'loss/train': 1.808834195137024} 01/28/2022 19:45:35 - INFO - codeparrot_training - Step 5368: {'lr': 0.000493950602158327, 'samples': 1030848, 'steps': 5368, 'loss/train': 2.537820518016815} 01/28/2022 19:45:39 - INFO - codeparrot_training - Step 5369: {'lr': 0.0004939470239148403, 'samples': 1031040, 'steps': 5369, 'loss/train': 2.0156450271606445} 01/28/2022 19:45:44 - INFO - codeparrot_training - Step 5370: {'lr': 0.0004939434446263617, 'samples': 1031232, 'steps': 5370, 'loss/train': 2.6960672736167908} 01/28/2022 19:45:48 - INFO - codeparrot_training - Step 5371: {'lr': 0.000493939864292907, 'samples': 1031424, 'steps': 5371, 'loss/train': 1.8689353466033936} 01/28/2022 19:45:53 - INFO - codeparrot_training - Step 5372: {'lr': 0.0004939362829144913, 'samples': 1031616, 'steps': 5372, 'loss/train': 1.767040729522705} 01/28/2022 19:45:57 - INFO - codeparrot_training - Step 5373: {'lr': 0.00049393270049113, 'samples': 1031808, 'steps': 5373, 'loss/train': 1.3593265414237976} 01/28/2022 19:46:02 - INFO - codeparrot_training - Step 5374: {'lr': 0.0004939291170228385, 'samples': 1032000, 'steps': 5374, 'loss/train': 3.180564522743225} 01/28/2022 19:46:06 - INFO - codeparrot_training - Step 5375: {'lr': 0.0004939255325096321, 'samples': 1032192, 'steps': 5375, 'loss/train': 1.481909304857254} 01/28/2022 19:46:10 - INFO - codeparrot_training - Step 5376: {'lr': 0.0004939219469515262, 'samples': 1032384, 'steps': 5376, 'loss/train': 1.5546208620071411} 01/28/2022 19:46:16 - INFO - codeparrot_training - Step 5377: {'lr': 0.0004939183603485363, 'samples': 1032576, 'steps': 5377, 'loss/train': 1.5352638959884644} 01/28/2022 19:46:21 - INFO - codeparrot_training - Step 5378: {'lr': 0.0004939147727006773, 'samples': 1032768, 'steps': 5378, 'loss/train': 1.2631924152374268} 01/28/2022 19:46:25 - INFO - codeparrot_training - Step 5379: {'lr': 0.000493911184007965, 'samples': 1032960, 'steps': 5379, 'loss/train': 2.1281749606132507} 01/28/2022 19:46:29 - INFO - codeparrot_training - Step 5380: {'lr': 0.0004939075942704147, 'samples': 1033152, 'steps': 5380, 'loss/train': 1.193877249956131} 01/28/2022 19:46:33 - INFO - codeparrot_training - Step 5381: {'lr': 0.0004939040034880416, 'samples': 1033344, 'steps': 5381, 'loss/train': 1.7373322248458862} 01/28/2022 19:46:39 - INFO - codeparrot_training - Step 5382: {'lr': 0.0004939004116608612, 'samples': 1033536, 'steps': 5382, 'loss/train': 2.2404377460479736} 01/28/2022 19:46:43 - INFO - codeparrot_training - Step 5383: {'lr': 0.000493896818788889, 'samples': 1033728, 'steps': 5383, 'loss/train': 1.2975583970546722} 01/28/2022 19:46:47 - INFO - codeparrot_training - Step 5384: {'lr': 0.0004938932248721401, 'samples': 1033920, 'steps': 5384, 'loss/train': 1.718063235282898} 01/28/2022 19:46:51 - INFO - codeparrot_training - Step 5385: {'lr': 0.0004938896299106302, 'samples': 1034112, 'steps': 5385, 'loss/train': 1.5145370364189148} 01/28/2022 19:46:56 - INFO - codeparrot_training - Step 5386: {'lr': 0.0004938860339043746, 'samples': 1034304, 'steps': 5386, 'loss/train': 1.5919141173362732} 01/28/2022 19:47:02 - INFO - codeparrot_training - Step 5387: {'lr': 0.0004938824368533886, 'samples': 1034496, 'steps': 5387, 'loss/train': 2.1600456833839417} 01/28/2022 19:47:06 - INFO - codeparrot_training - Step 5388: {'lr': 0.0004938788387576878, 'samples': 1034688, 'steps': 5388, 'loss/train': 2.02515184879303} 01/28/2022 19:47:10 - INFO - codeparrot_training - Step 5389: {'lr': 0.0004938752396172873, 'samples': 1034880, 'steps': 5389, 'loss/train': 2.2219101190567017} 01/28/2022 19:47:14 - INFO - codeparrot_training - Step 5390: {'lr': 0.0004938716394322028, 'samples': 1035072, 'steps': 5390, 'loss/train': 0.7101819813251495} 01/28/2022 19:47:19 - INFO - codeparrot_training - Step 5391: {'lr': 0.0004938680382024497, 'samples': 1035264, 'steps': 5391, 'loss/train': 2.1355621218681335} 01/28/2022 19:47:25 - INFO - codeparrot_training - Step 5392: {'lr': 0.0004938644359280433, 'samples': 1035456, 'steps': 5392, 'loss/train': 2.763470470905304} 01/28/2022 19:47:29 - INFO - codeparrot_training - Step 5393: {'lr': 0.000493860832608999, 'samples': 1035648, 'steps': 5393, 'loss/train': 1.3246988654136658} 01/28/2022 19:47:34 - INFO - codeparrot_training - Step 5394: {'lr': 0.0004938572282453326, 'samples': 1035840, 'steps': 5394, 'loss/train': 2.5263891220092773} 01/28/2022 19:47:38 - INFO - codeparrot_training - Step 5395: {'lr': 0.000493853622837059, 'samples': 1036032, 'steps': 5395, 'loss/train': 1.9496973752975464} 01/28/2022 19:47:42 - INFO - codeparrot_training - Step 5396: {'lr': 0.000493850016384194, 'samples': 1036224, 'steps': 5396, 'loss/train': 1.8252968788146973} 01/28/2022 19:47:46 - INFO - codeparrot_training - Step 5397: {'lr': 0.000493846408886753, 'samples': 1036416, 'steps': 5397, 'loss/train': 1.7316187620162964} 01/28/2022 19:47:52 - INFO - codeparrot_training - Step 5398: {'lr': 0.0004938428003447514, 'samples': 1036608, 'steps': 5398, 'loss/train': 1.4854398965835571} 01/28/2022 19:47:56 - INFO - codeparrot_training - Step 5399: {'lr': 0.0004938391907582046, 'samples': 1036800, 'steps': 5399, 'loss/train': 2.8953757882118225} 01/28/2022 19:48:00 - INFO - codeparrot_training - Step 5400: {'lr': 0.0004938355801271282, 'samples': 1036992, 'steps': 5400, 'loss/train': 1.5433925986289978} 01/28/2022 19:48:04 - INFO - codeparrot_training - Step 5401: {'lr': 0.0004938319684515375, 'samples': 1037184, 'steps': 5401, 'loss/train': 0.9232403039932251} 01/28/2022 19:48:09 - INFO - codeparrot_training - Step 5402: {'lr': 0.0004938283557314483, 'samples': 1037376, 'steps': 5402, 'loss/train': 2.187393844127655} 01/28/2022 19:48:14 - INFO - codeparrot_training - Step 5403: {'lr': 0.0004938247419668757, 'samples': 1037568, 'steps': 5403, 'loss/train': 1.8679391741752625} 01/28/2022 19:48:18 - INFO - codeparrot_training - Step 5404: {'lr': 0.0004938211271578352, 'samples': 1037760, 'steps': 5404, 'loss/train': 1.0621342062950134} 01/28/2022 19:48:22 - INFO - codeparrot_training - Step 5405: {'lr': 0.0004938175113043426, 'samples': 1037952, 'steps': 5405, 'loss/train': 2.1068660616874695} 01/28/2022 19:48:27 - INFO - codeparrot_training - Step 5406: {'lr': 0.0004938138944064131, 'samples': 1038144, 'steps': 5406, 'loss/train': 3.037199020385742} 01/28/2022 19:48:31 - INFO - codeparrot_training - Step 5407: {'lr': 0.0004938102764640624, 'samples': 1038336, 'steps': 5407, 'loss/train': 0.8664507865905762} 01/28/2022 19:48:37 - INFO - codeparrot_training - Step 5408: {'lr': 0.0004938066574773058, 'samples': 1038528, 'steps': 5408, 'loss/train': 0.4740748107433319} 01/28/2022 19:48:41 - INFO - codeparrot_training - Step 5409: {'lr': 0.000493803037446159, 'samples': 1038720, 'steps': 5409, 'loss/train': 1.5287445187568665} 01/28/2022 19:48:45 - INFO - codeparrot_training - Step 5410: {'lr': 0.0004937994163706374, 'samples': 1038912, 'steps': 5410, 'loss/train': 2.5096782445907593} 01/28/2022 19:48:50 - INFO - codeparrot_training - Step 5411: {'lr': 0.0004937957942507564, 'samples': 1039104, 'steps': 5411, 'loss/train': 1.1851178705692291} 01/28/2022 19:48:54 - INFO - codeparrot_training - Step 5412: {'lr': 0.0004937921710865317, 'samples': 1039296, 'steps': 5412, 'loss/train': 1.4012742340564728} 01/28/2022 19:48:59 - INFO - codeparrot_training - Step 5413: {'lr': 0.0004937885468779787, 'samples': 1039488, 'steps': 5413, 'loss/train': 1.5621986389160156} 01/28/2022 19:49:03 - INFO - codeparrot_training - Step 5414: {'lr': 0.000493784921625113, 'samples': 1039680, 'steps': 5414, 'loss/train': 1.89851313829422} 01/28/2022 19:49:08 - INFO - codeparrot_training - Step 5415: {'lr': 0.0004937812953279502, 'samples': 1039872, 'steps': 5415, 'loss/train': 2.5315911769866943} 01/28/2022 19:49:12 - INFO - codeparrot_training - Step 5416: {'lr': 0.0004937776679865057, 'samples': 1040064, 'steps': 5416, 'loss/train': 1.839145839214325} 01/28/2022 19:49:16 - INFO - codeparrot_training - Step 5417: {'lr': 0.000493774039600795, 'samples': 1040256, 'steps': 5417, 'loss/train': 1.7538086771965027} 01/28/2022 19:49:21 - INFO - codeparrot_training - Step 5418: {'lr': 0.0004937704101708338, 'samples': 1040448, 'steps': 5418, 'loss/train': 2.1803049445152283} 01/28/2022 19:49:26 - INFO - codeparrot_training - Step 5419: {'lr': 0.0004937667796966374, 'samples': 1040640, 'steps': 5419, 'loss/train': 1.625920593738556} 01/28/2022 19:49:30 - INFO - codeparrot_training - Step 5420: {'lr': 0.0004937631481782218, 'samples': 1040832, 'steps': 5420, 'loss/train': 1.46329066157341} 01/28/2022 19:49:34 - INFO - codeparrot_training - Step 5421: {'lr': 0.000493759515615602, 'samples': 1041024, 'steps': 5421, 'loss/train': 2.4270917773246765} 01/28/2022 19:49:38 - INFO - codeparrot_training - Step 5422: {'lr': 0.000493755882008794, 'samples': 1041216, 'steps': 5422, 'loss/train': 0.9795501530170441} 01/28/2022 19:49:44 - INFO - codeparrot_training - Step 5423: {'lr': 0.0004937522473578132, 'samples': 1041408, 'steps': 5423, 'loss/train': 0.6541394144296646} 01/28/2022 19:49:48 - INFO - codeparrot_training - Step 5424: {'lr': 0.0004937486116626752, 'samples': 1041600, 'steps': 5424, 'loss/train': 1.3387995958328247} 01/28/2022 19:49:53 - INFO - codeparrot_training - Step 5425: {'lr': 0.0004937449749233954, 'samples': 1041792, 'steps': 5425, 'loss/train': 1.8771845698356628} 01/28/2022 19:49:57 - INFO - codeparrot_training - Step 5426: {'lr': 0.0004937413371399897, 'samples': 1041984, 'steps': 5426, 'loss/train': 2.1340527534484863} 01/28/2022 19:50:01 - INFO - codeparrot_training - Step 5427: {'lr': 0.0004937376983124734, 'samples': 1042176, 'steps': 5427, 'loss/train': 1.9813923239707947} 01/28/2022 19:50:06 - INFO - codeparrot_training - Step 5428: {'lr': 0.0004937340584408622, 'samples': 1042368, 'steps': 5428, 'loss/train': 1.4166599214076996} 01/28/2022 19:50:11 - INFO - codeparrot_training - Step 5429: {'lr': 0.0004937304175251717, 'samples': 1042560, 'steps': 5429, 'loss/train': 2.3900535106658936} 01/28/2022 19:50:15 - INFO - codeparrot_training - Step 5430: {'lr': 0.0004937267755654174, 'samples': 1042752, 'steps': 5430, 'loss/train': 1.7393651604652405} 01/28/2022 19:50:19 - INFO - codeparrot_training - Step 5431: {'lr': 0.0004937231325616152, 'samples': 1042944, 'steps': 5431, 'loss/train': 1.858784258365631} 01/28/2022 19:50:23 - INFO - codeparrot_training - Step 5432: {'lr': 0.0004937194885137803, 'samples': 1043136, 'steps': 5432, 'loss/train': 1.4428308606147766} 01/28/2022 19:50:30 - INFO - codeparrot_training - Step 5433: {'lr': 0.0004937158434219286, 'samples': 1043328, 'steps': 5433, 'loss/train': 2.0745100378990173} 01/28/2022 19:50:34 - INFO - codeparrot_training - Step 5434: {'lr': 0.0004937121972860755, 'samples': 1043520, 'steps': 5434, 'loss/train': 2.00647509098053} 01/28/2022 19:50:38 - INFO - codeparrot_training - Step 5435: {'lr': 0.0004937085501062369, 'samples': 1043712, 'steps': 5435, 'loss/train': 1.666491687297821} 01/28/2022 19:50:42 - INFO - codeparrot_training - Step 5436: {'lr': 0.0004937049018824282, 'samples': 1043904, 'steps': 5436, 'loss/train': 1.7295053601264954} 01/28/2022 19:50:46 - INFO - codeparrot_training - Step 5437: {'lr': 0.000493701252614665, 'samples': 1044096, 'steps': 5437, 'loss/train': 2.600728154182434} 01/28/2022 19:50:52 - INFO - codeparrot_training - Step 5438: {'lr': 0.0004936976023029631, 'samples': 1044288, 'steps': 5438, 'loss/train': 1.0260854065418243} 01/28/2022 19:50:56 - INFO - codeparrot_training - Step 5439: {'lr': 0.000493693950947338, 'samples': 1044480, 'steps': 5439, 'loss/train': 1.4669084250926971} 01/28/2022 19:51:00 - INFO - codeparrot_training - Step 5440: {'lr': 0.0004936902985478055, 'samples': 1044672, 'steps': 5440, 'loss/train': 1.6938053369522095} 01/28/2022 19:51:04 - INFO - codeparrot_training - Step 5441: {'lr': 0.000493686645104381, 'samples': 1044864, 'steps': 5441, 'loss/train': 1.7762230038642883} 01/28/2022 19:51:09 - INFO - codeparrot_training - Step 5442: {'lr': 0.0004936829906170804, 'samples': 1045056, 'steps': 5442, 'loss/train': 1.3110668063163757} 01/28/2022 19:51:14 - INFO - codeparrot_training - Step 5443: {'lr': 0.0004936793350859192, 'samples': 1045248, 'steps': 5443, 'loss/train': 2.178020417690277} 01/28/2022 19:51:18 - INFO - codeparrot_training - Step 5444: {'lr': 0.0004936756785109131, 'samples': 1045440, 'steps': 5444, 'loss/train': 2.168190121650696} 01/28/2022 19:51:22 - INFO - codeparrot_training - Step 5445: {'lr': 0.0004936720208920778, 'samples': 1045632, 'steps': 5445, 'loss/train': 2.7614051699638367} 01/28/2022 19:51:27 - INFO - codeparrot_training - Step 5446: {'lr': 0.0004936683622294289, 'samples': 1045824, 'steps': 5446, 'loss/train': 2.4832661747932434} 01/28/2022 19:51:31 - INFO - codeparrot_training - Step 5447: {'lr': 0.0004936647025229822, 'samples': 1046016, 'steps': 5447, 'loss/train': 1.8834537863731384} 01/28/2022 19:51:37 - INFO - codeparrot_training - Step 5448: {'lr': 0.0004936610417727532, 'samples': 1046208, 'steps': 5448, 'loss/train': 2.667386054992676} 01/28/2022 19:51:42 - INFO - codeparrot_training - Step 5449: {'lr': 0.0004936573799787575, 'samples': 1046400, 'steps': 5449, 'loss/train': 2.220189929008484} 01/28/2022 19:51:46 - INFO - codeparrot_training - Step 5450: {'lr': 0.0004936537171410112, 'samples': 1046592, 'steps': 5450, 'loss/train': 2.896672010421753} 01/28/2022 19:51:50 - INFO - codeparrot_training - Step 5451: {'lr': 0.0004936500532595297, 'samples': 1046784, 'steps': 5451, 'loss/train': 2.246707320213318} 01/28/2022 19:51:54 - INFO - codeparrot_training - Step 5452: {'lr': 0.0004936463883343287, 'samples': 1046976, 'steps': 5452, 'loss/train': 1.5467097759246826} 01/28/2022 19:51:59 - INFO - codeparrot_training - Step 5453: {'lr': 0.000493642722365424, 'samples': 1047168, 'steps': 5453, 'loss/train': 2.316272020339966} 01/28/2022 19:52:04 - INFO - codeparrot_training - Step 5454: {'lr': 0.0004936390553528313, 'samples': 1047360, 'steps': 5454, 'loss/train': 2.2303004264831543} 01/28/2022 19:52:08 - INFO - codeparrot_training - Step 5455: {'lr': 0.0004936353872965661, 'samples': 1047552, 'steps': 5455, 'loss/train': 2.1817708611488342} 01/28/2022 19:52:12 - INFO - codeparrot_training - Step 5456: {'lr': 0.0004936317181966443, 'samples': 1047744, 'steps': 5456, 'loss/train': 0.6324372589588165} 01/28/2022 19:52:17 - INFO - codeparrot_training - Step 5457: {'lr': 0.0004936280480530816, 'samples': 1047936, 'steps': 5457, 'loss/train': 1.9669421911239624} 01/28/2022 19:52:21 - INFO - codeparrot_training - Step 5458: {'lr': 0.0004936243768658937, 'samples': 1048128, 'steps': 5458, 'loss/train': 1.8238930106163025} 01/28/2022 19:52:27 - INFO - codeparrot_training - Step 5459: {'lr': 0.0004936207046350963, 'samples': 1048320, 'steps': 5459, 'loss/train': 2.01541006565094} 01/28/2022 19:52:31 - INFO - codeparrot_training - Step 5460: {'lr': 0.0004936170313607053, 'samples': 1048512, 'steps': 5460, 'loss/train': 3.337817430496216} 01/28/2022 19:52:35 - INFO - codeparrot_training - Step 5461: {'lr': 0.0004936133570427361, 'samples': 1048704, 'steps': 5461, 'loss/train': 0.9891588985919952} 01/28/2022 19:52:39 - INFO - codeparrot_training - Step 5462: {'lr': 0.0004936096816812046, 'samples': 1048896, 'steps': 5462, 'loss/train': 2.334756553173065} 01/28/2022 19:52:44 - INFO - codeparrot_training - Step 5463: {'lr': 0.0004936060052761268, 'samples': 1049088, 'steps': 5463, 'loss/train': 1.442282795906067} 01/28/2022 19:52:49 - INFO - codeparrot_training - Step 5464: {'lr': 0.0004936023278275182, 'samples': 1049280, 'steps': 5464, 'loss/train': 2.054155111312866} 01/28/2022 19:52:54 - INFO - codeparrot_training - Step 5465: {'lr': 0.0004935986493353944, 'samples': 1049472, 'steps': 5465, 'loss/train': 1.8986603021621704} 01/28/2022 19:52:58 - INFO - codeparrot_training - Step 5466: {'lr': 0.0004935949697997715, 'samples': 1049664, 'steps': 5466, 'loss/train': 2.104651629924774} 01/28/2022 19:53:02 - INFO - codeparrot_training - Step 5467: {'lr': 0.000493591289220665, 'samples': 1049856, 'steps': 5467, 'loss/train': 2.5334123969078064} 01/28/2022 19:53:06 - INFO - codeparrot_training - Step 5468: {'lr': 0.0004935876075980908, 'samples': 1050048, 'steps': 5468, 'loss/train': 1.4617767333984375} 01/28/2022 19:53:12 - INFO - codeparrot_training - Step 5469: {'lr': 0.0004935839249320647, 'samples': 1050240, 'steps': 5469, 'loss/train': 2.7102851271629333} 01/28/2022 19:53:16 - INFO - codeparrot_training - Step 5470: {'lr': 0.0004935802412226024, 'samples': 1050432, 'steps': 5470, 'loss/train': 1.944946825504303} 01/28/2022 19:53:21 - INFO - codeparrot_training - Step 5471: {'lr': 0.0004935765564697195, 'samples': 1050624, 'steps': 5471, 'loss/train': 0.9653279185295105} 01/28/2022 19:53:25 - INFO - codeparrot_training - Step 5472: {'lr': 0.0004935728706734322, 'samples': 1050816, 'steps': 5472, 'loss/train': 1.6738294959068298} 01/28/2022 19:53:29 - INFO - codeparrot_training - Step 5473: {'lr': 0.000493569183833756, 'samples': 1051008, 'steps': 5473, 'loss/train': 1.082538664340973} 01/28/2022 19:53:34 - INFO - codeparrot_training - Step 5474: {'lr': 0.0004935654959507068, 'samples': 1051200, 'steps': 5474, 'loss/train': 0.36381838470697403} 01/28/2022 19:53:39 - INFO - codeparrot_training - Step 5475: {'lr': 0.0004935618070243003, 'samples': 1051392, 'steps': 5475, 'loss/train': 2.5959776043891907} 01/28/2022 19:53:43 - INFO - codeparrot_training - Step 5476: {'lr': 0.0004935581170545523, 'samples': 1051584, 'steps': 5476, 'loss/train': 1.5313156843185425} 01/28/2022 19:53:47 - INFO - codeparrot_training - Step 5477: {'lr': 0.0004935544260414787, 'samples': 1051776, 'steps': 5477, 'loss/train': 2.0240820050239563} 01/28/2022 19:53:51 - INFO - codeparrot_training - Step 5478: {'lr': 0.0004935507339850953, 'samples': 1051968, 'steps': 5478, 'loss/train': 1.8698739409446716} 01/28/2022 19:53:57 - INFO - codeparrot_training - Step 5479: {'lr': 0.0004935470408854179, 'samples': 1052160, 'steps': 5479, 'loss/train': 2.3223207592964172} 01/28/2022 19:54:01 - INFO - codeparrot_training - Step 5480: {'lr': 0.0004935433467424624, 'samples': 1052352, 'steps': 5480, 'loss/train': 2.1226932406425476} 01/28/2022 19:54:05 - INFO - codeparrot_training - Step 5481: {'lr': 0.0004935396515562444, 'samples': 1052544, 'steps': 5481, 'loss/train': 1.9536341428756714} 01/28/2022 19:54:09 - INFO - codeparrot_training - Step 5482: {'lr': 0.0004935359553267798, 'samples': 1052736, 'steps': 5482, 'loss/train': 2.264603018760681} 01/28/2022 19:54:14 - INFO - codeparrot_training - Step 5483: {'lr': 0.0004935322580540847, 'samples': 1052928, 'steps': 5483, 'loss/train': 1.9058507680892944} 01/28/2022 19:54:20 - INFO - codeparrot_training - Step 5484: {'lr': 0.0004935285597381747, 'samples': 1053120, 'steps': 5484, 'loss/train': 2.3591973781585693} 01/28/2022 19:54:24 - INFO - codeparrot_training - Step 5485: {'lr': 0.0004935248603790656, 'samples': 1053312, 'steps': 5485, 'loss/train': 2.09721976518631} 01/28/2022 19:54:28 - INFO - codeparrot_training - Step 5486: {'lr': 0.0004935211599767733, 'samples': 1053504, 'steps': 5486, 'loss/train': 1.951465129852295} 01/28/2022 19:54:32 - INFO - codeparrot_training - Step 5487: {'lr': 0.0004935174585313138, 'samples': 1053696, 'steps': 5487, 'loss/train': 0.8720037639141083} 01/28/2022 19:54:37 - INFO - codeparrot_training - Step 5488: {'lr': 0.0004935137560427027, 'samples': 1053888, 'steps': 5488, 'loss/train': 2.0840622782707214} 01/28/2022 19:54:42 - INFO - codeparrot_training - Step 5489: {'lr': 0.000493510052510956, 'samples': 1054080, 'steps': 5489, 'loss/train': 2.0771838426589966} 01/28/2022 19:54:46 - INFO - codeparrot_training - Step 5490: {'lr': 0.0004935063479360897, 'samples': 1054272, 'steps': 5490, 'loss/train': 1.2942313849925995} 01/28/2022 19:54:50 - INFO - codeparrot_training - Step 5491: {'lr': 0.0004935026423181194, 'samples': 1054464, 'steps': 5491, 'loss/train': 1.830862283706665} 01/28/2022 19:54:54 - INFO - codeparrot_training - Step 5492: {'lr': 0.0004934989356570611, 'samples': 1054656, 'steps': 5492, 'loss/train': 0.79268878698349} 01/28/2022 19:54:59 - INFO - codeparrot_training - Step 5493: {'lr': 0.0004934952279529308, 'samples': 1054848, 'steps': 5493, 'loss/train': 1.6854388117790222} 01/28/2022 19:55:05 - INFO - codeparrot_training - Step 5494: {'lr': 0.0004934915192057441, 'samples': 1055040, 'steps': 5494, 'loss/train': 1.5749167799949646} 01/28/2022 19:55:09 - INFO - codeparrot_training - Step 5495: {'lr': 0.0004934878094155172, 'samples': 1055232, 'steps': 5495, 'loss/train': 2.1883127689361572} 01/28/2022 19:55:13 - INFO - codeparrot_training - Step 5496: {'lr': 0.0004934840985822657, 'samples': 1055424, 'steps': 5496, 'loss/train': 2.3940060138702393} 01/28/2022 19:55:18 - INFO - codeparrot_training - Step 5497: {'lr': 0.0004934803867060058, 'samples': 1055616, 'steps': 5497, 'loss/train': 2.6621673703193665} 01/28/2022 19:55:22 - INFO - codeparrot_training - Step 5498: {'lr': 0.0004934766737867531, 'samples': 1055808, 'steps': 5498, 'loss/train': 2.1981003284454346} 01/28/2022 19:55:27 - INFO - codeparrot_training - Step 5499: {'lr': 0.0004934729598245237, 'samples': 1056000, 'steps': 5499, 'loss/train': 2.145538330078125} 01/28/2022 19:55:31 - INFO - codeparrot_training - Step 5500: {'lr': 0.0004934692448193334, 'samples': 1056192, 'steps': 5500, 'loss/train': 2.334928572177887} 01/28/2022 19:55:36 - INFO - codeparrot_training - Step 5501: {'lr': 0.0004934655287711982, 'samples': 1056384, 'steps': 5501, 'loss/train': 1.522062599658966} 01/28/2022 19:55:40 - INFO - codeparrot_training - Step 5502: {'lr': 0.0004934618116801341, 'samples': 1056576, 'steps': 5502, 'loss/train': 2.6020116806030273} 01/28/2022 19:55:44 - INFO - codeparrot_training - Step 5503: {'lr': 0.0004934580935461567, 'samples': 1056768, 'steps': 5503, 'loss/train': 1.170915126800537} 01/28/2022 19:55:50 - INFO - codeparrot_training - Step 5504: {'lr': 0.0004934543743692822, 'samples': 1056960, 'steps': 5504, 'loss/train': 2.0459774136543274} 01/28/2022 19:55:54 - INFO - codeparrot_training - Step 5505: {'lr': 0.0004934506541495265, 'samples': 1057152, 'steps': 5505, 'loss/train': 1.4520345032215118} 01/28/2022 19:55:58 - INFO - codeparrot_training - Step 5506: {'lr': 0.0004934469328869056, 'samples': 1057344, 'steps': 5506, 'loss/train': 2.131978690624237} 01/28/2022 19:56:02 - INFO - codeparrot_training - Step 5507: {'lr': 0.0004934432105814352, 'samples': 1057536, 'steps': 5507, 'loss/train': 1.980684220790863} 01/28/2022 19:56:07 - INFO - codeparrot_training - Step 5508: {'lr': 0.0004934394872331314, 'samples': 1057728, 'steps': 5508, 'loss/train': 1.7959147095680237} 01/28/2022 19:56:13 - INFO - codeparrot_training - Step 5509: {'lr': 0.0004934357628420101, 'samples': 1057920, 'steps': 5509, 'loss/train': 2.0164042711257935} 01/28/2022 19:56:17 - INFO - codeparrot_training - Step 5510: {'lr': 0.0004934320374080874, 'samples': 1058112, 'steps': 5510, 'loss/train': 0.7430146336555481} 01/28/2022 19:56:21 - INFO - codeparrot_training - Step 5511: {'lr': 0.000493428310931379, 'samples': 1058304, 'steps': 5511, 'loss/train': 1.613284170627594} 01/28/2022 19:56:25 - INFO - codeparrot_training - Step 5512: {'lr': 0.0004934245834119013, 'samples': 1058496, 'steps': 5512, 'loss/train': 1.6746270060539246} 01/28/2022 19:56:30 - INFO - codeparrot_training - Step 5513: {'lr': 0.0004934208548496697, 'samples': 1058688, 'steps': 5513, 'loss/train': 2.008958637714386} 01/28/2022 19:56:35 - INFO - codeparrot_training - Step 5514: {'lr': 0.0004934171252447006, 'samples': 1058880, 'steps': 5514, 'loss/train': 1.574621558189392} 01/28/2022 19:56:39 - INFO - codeparrot_training - Step 5515: {'lr': 0.0004934133945970097, 'samples': 1059072, 'steps': 5515, 'loss/train': 1.4372543692588806} 01/28/2022 19:56:43 - INFO - codeparrot_training - Step 5516: {'lr': 0.0004934096629066133, 'samples': 1059264, 'steps': 5516, 'loss/train': 2.1689997911453247} 01/28/2022 19:56:48 - INFO - codeparrot_training - Step 5517: {'lr': 0.000493405930173527, 'samples': 1059456, 'steps': 5517, 'loss/train': 1.4921403229236603} 01/28/2022 19:56:52 - INFO - codeparrot_training - Step 5518: {'lr': 0.0004934021963977671, 'samples': 1059648, 'steps': 5518, 'loss/train': 1.4115065038204193} 01/28/2022 19:56:58 - INFO - codeparrot_training - Step 5519: {'lr': 0.0004933984615793494, 'samples': 1059840, 'steps': 5519, 'loss/train': 1.572351336479187} 01/28/2022 19:57:02 - INFO - codeparrot_training - Step 5520: {'lr': 0.0004933947257182901, 'samples': 1060032, 'steps': 5520, 'loss/train': 1.9667776823043823} 01/28/2022 19:57:07 - INFO - codeparrot_training - Step 5521: {'lr': 0.000493390988814605, 'samples': 1060224, 'steps': 5521, 'loss/train': 1.8331108689308167} 01/28/2022 19:57:11 - INFO - codeparrot_training - Step 5522: {'lr': 0.0004933872508683101, 'samples': 1060416, 'steps': 5522, 'loss/train': 1.6847952604293823} 01/28/2022 19:57:15 - INFO - codeparrot_training - Step 5523: {'lr': 0.0004933835118794217, 'samples': 1060608, 'steps': 5523, 'loss/train': 7.970631122589111} 01/28/2022 19:57:20 - INFO - codeparrot_training - Step 5524: {'lr': 0.0004933797718479555, 'samples': 1060800, 'steps': 5524, 'loss/train': 1.8658887147903442} 01/28/2022 19:57:25 - INFO - codeparrot_training - Step 5525: {'lr': 0.0004933760307739277, 'samples': 1060992, 'steps': 5525, 'loss/train': 4.181741237640381} 01/28/2022 19:57:29 - INFO - codeparrot_training - Step 5526: {'lr': 0.0004933722886573542, 'samples': 1061184, 'steps': 5526, 'loss/train': 2.103058397769928} 01/28/2022 19:57:34 - INFO - codeparrot_training - Step 5527: {'lr': 0.0004933685454982511, 'samples': 1061376, 'steps': 5527, 'loss/train': 1.9876023530960083} 01/28/2022 19:57:38 - INFO - codeparrot_training - Step 5528: {'lr': 0.0004933648012966344, 'samples': 1061568, 'steps': 5528, 'loss/train': 2.071341097354889} 01/28/2022 19:57:44 - INFO - codeparrot_training - Step 5529: {'lr': 0.0004933610560525203, 'samples': 1061760, 'steps': 5529, 'loss/train': 1.8482281565666199} 01/28/2022 19:57:48 - INFO - codeparrot_training - Step 5530: {'lr': 0.0004933573097659246, 'samples': 1061952, 'steps': 5530, 'loss/train': 0.8627452254295349} 01/28/2022 19:57:52 - INFO - codeparrot_training - Step 5531: {'lr': 0.0004933535624368634, 'samples': 1062144, 'steps': 5531, 'loss/train': 1.307642251253128} 01/28/2022 19:57:56 - INFO - codeparrot_training - Step 5532: {'lr': 0.0004933498140653529, 'samples': 1062336, 'steps': 5532, 'loss/train': 0.9882270991802216} 01/28/2022 19:58:01 - INFO - codeparrot_training - Step 5533: {'lr': 0.0004933460646514092, 'samples': 1062528, 'steps': 5533, 'loss/train': 2.2823885679244995} 01/28/2022 19:58:06 - INFO - codeparrot_training - Step 5534: {'lr': 0.000493342314195048, 'samples': 1062720, 'steps': 5534, 'loss/train': 2.0572469830513} 01/28/2022 19:58:10 - INFO - codeparrot_training - Step 5535: {'lr': 0.0004933385626962858, 'samples': 1062912, 'steps': 5535, 'loss/train': 1.7729262113571167} 01/28/2022 19:58:15 - INFO - codeparrot_training - Step 5536: {'lr': 0.0004933348101551383, 'samples': 1063104, 'steps': 5536, 'loss/train': 1.0633388757705688} 01/28/2022 19:58:19 - INFO - codeparrot_training - Step 5537: {'lr': 0.0004933310565716218, 'samples': 1063296, 'steps': 5537, 'loss/train': 2.0125608444213867} 01/28/2022 19:58:23 - INFO - codeparrot_training - Step 5538: {'lr': 0.0004933273019457524, 'samples': 1063488, 'steps': 5538, 'loss/train': 3.934409022331238} 01/28/2022 19:58:29 - INFO - codeparrot_training - Step 5539: {'lr': 0.0004933235462775459, 'samples': 1063680, 'steps': 5539, 'loss/train': 2.0939499735832214} 01/28/2022 19:58:34 - INFO - codeparrot_training - Step 5540: {'lr': 0.0004933197895670187, 'samples': 1063872, 'steps': 5540, 'loss/train': 2.127410888671875} 01/28/2022 19:58:38 - INFO - codeparrot_training - Step 5541: {'lr': 0.0004933160318141869, 'samples': 1064064, 'steps': 5541, 'loss/train': 1.8905081748962402} 01/28/2022 19:58:42 - INFO - codeparrot_training - Step 5542: {'lr': 0.0004933122730190663, 'samples': 1064256, 'steps': 5542, 'loss/train': 2.102356731891632} 01/28/2022 19:58:46 - INFO - codeparrot_training - Step 5543: {'lr': 0.0004933085131816733, 'samples': 1064448, 'steps': 5543, 'loss/train': 1.4022752344608307} 01/28/2022 19:58:51 - INFO - codeparrot_training - Step 5544: {'lr': 0.0004933047523020239, 'samples': 1064640, 'steps': 5544, 'loss/train': 1.7077531814575195} 01/28/2022 19:58:56 - INFO - codeparrot_training - Step 5545: {'lr': 0.0004933009903801341, 'samples': 1064832, 'steps': 5545, 'loss/train': 1.2213071286678314} 01/28/2022 19:59:00 - INFO - codeparrot_training - Step 5546: {'lr': 0.0004932972274160202, 'samples': 1065024, 'steps': 5546, 'loss/train': 1.6125850081443787} 01/28/2022 19:59:04 - INFO - codeparrot_training - Step 5547: {'lr': 0.0004932934634096982, 'samples': 1065216, 'steps': 5547, 'loss/train': 1.918335199356079} 01/28/2022 19:59:08 - INFO - codeparrot_training - Step 5548: {'lr': 0.0004932896983611843, 'samples': 1065408, 'steps': 5548, 'loss/train': 2.20750093460083} 01/28/2022 19:59:14 - INFO - codeparrot_training - Step 5549: {'lr': 0.0004932859322704944, 'samples': 1065600, 'steps': 5549, 'loss/train': 0.5757663398981094} 01/28/2022 19:59:18 - INFO - codeparrot_training - Step 5550: {'lr': 0.000493282165137645, 'samples': 1065792, 'steps': 5550, 'loss/train': 1.5139120817184448} 01/28/2022 19:59:22 - INFO - codeparrot_training - Step 5551: {'lr': 0.0004932783969626521, 'samples': 1065984, 'steps': 5551, 'loss/train': 1.911604642868042} 01/28/2022 19:59:26 - INFO - codeparrot_training - Step 5552: {'lr': 0.0004932746277455317, 'samples': 1066176, 'steps': 5552, 'loss/train': 1.5811325311660767} 01/28/2022 19:59:31 - INFO - codeparrot_training - Step 5553: {'lr': 0.0004932708574863, 'samples': 1066368, 'steps': 5553, 'loss/train': 2.130254030227661} 01/28/2022 19:59:37 - INFO - codeparrot_training - Step 5554: {'lr': 0.0004932670861849733, 'samples': 1066560, 'steps': 5554, 'loss/train': 2.07811439037323} 01/28/2022 19:59:41 - INFO - codeparrot_training - Step 5555: {'lr': 0.0004932633138415675, 'samples': 1066752, 'steps': 5555, 'loss/train': 1.5323703289031982} 01/28/2022 19:59:45 - INFO - codeparrot_training - Step 5556: {'lr': 0.000493259540456099, 'samples': 1066944, 'steps': 5556, 'loss/train': 2.1342662572860718} 01/28/2022 19:59:50 - INFO - codeparrot_training - Step 5557: {'lr': 0.0004932557660285839, 'samples': 1067136, 'steps': 5557, 'loss/train': 0.8015129864215851} 01/28/2022 19:59:54 - INFO - codeparrot_training - Step 5558: {'lr': 0.0004932519905590383, 'samples': 1067328, 'steps': 5558, 'loss/train': 1.7409886121749878} 01/28/2022 19:59:59 - INFO - codeparrot_training - Step 5559: {'lr': 0.0004932482140474785, 'samples': 1067520, 'steps': 5559, 'loss/train': 2.13798451423645} 01/28/2022 20:00:03 - INFO - codeparrot_training - Step 5560: {'lr': 0.0004932444364939204, 'samples': 1067712, 'steps': 5560, 'loss/train': 2.2894599437713623} 01/28/2022 20:00:08 - INFO - codeparrot_training - Step 5561: {'lr': 0.0004932406578983806, 'samples': 1067904, 'steps': 5561, 'loss/train': 2.502536416053772} 01/28/2022 20:00:12 - INFO - codeparrot_training - Step 5562: {'lr': 0.0004932368782608749, 'samples': 1068096, 'steps': 5562, 'loss/train': 1.7101757526397705} 01/28/2022 20:00:16 - INFO - codeparrot_training - Step 5563: {'lr': 0.0004932330975814198, 'samples': 1068288, 'steps': 5563, 'loss/train': 1.1254394352436066} 01/28/2022 20:00:22 - INFO - codeparrot_training - Step 5564: {'lr': 0.0004932293158600312, 'samples': 1068480, 'steps': 5564, 'loss/train': 2.25522780418396} 01/28/2022 20:00:27 - INFO - codeparrot_training - Step 5565: {'lr': 0.0004932255330967255, 'samples': 1068672, 'steps': 5565, 'loss/train': 1.8391225934028625} 01/28/2022 20:00:31 - INFO - codeparrot_training - Step 5566: {'lr': 0.0004932217492915189, 'samples': 1068864, 'steps': 5566, 'loss/train': 2.1621857285499573} 01/28/2022 20:00:35 - INFO - codeparrot_training - Step 5567: {'lr': 0.0004932179644444274, 'samples': 1069056, 'steps': 5567, 'loss/train': 1.06518092751503} 01/28/2022 20:00:39 - INFO - codeparrot_training - Step 5568: {'lr': 0.0004932141785554676, 'samples': 1069248, 'steps': 5568, 'loss/train': 1.5376760959625244} 01/28/2022 20:00:45 - INFO - codeparrot_training - Step 5569: {'lr': 0.0004932103916246553, 'samples': 1069440, 'steps': 5569, 'loss/train': 1.6680884957313538} 01/28/2022 20:00:49 - INFO - codeparrot_training - Step 5570: {'lr': 0.000493206603652007, 'samples': 1069632, 'steps': 5570, 'loss/train': 2.544128894805908} 01/28/2022 20:00:53 - INFO - codeparrot_training - Step 5571: {'lr': 0.0004932028146375388, 'samples': 1069824, 'steps': 5571, 'loss/train': 0.7926463186740875} 01/28/2022 20:00:58 - INFO - codeparrot_training - Step 5572: {'lr': 0.000493199024581267, 'samples': 1070016, 'steps': 5572, 'loss/train': 2.3207799196243286} 01/28/2022 20:01:03 - INFO - codeparrot_training - Step 5573: {'lr': 0.0004931952334832077, 'samples': 1070208, 'steps': 5573, 'loss/train': 2.0394394397735596} 01/28/2022 20:01:07 - INFO - codeparrot_training - Step 5574: {'lr': 0.0004931914413433773, 'samples': 1070400, 'steps': 5574, 'loss/train': 2.208224058151245} 01/28/2022 20:01:12 - INFO - codeparrot_training - Step 5575: {'lr': 0.0004931876481617921, 'samples': 1070592, 'steps': 5575, 'loss/train': 1.982364535331726} 01/28/2022 20:01:16 - INFO - codeparrot_training - Step 5576: {'lr': 0.0004931838539384681, 'samples': 1070784, 'steps': 5576, 'loss/train': 2.0911155939102173} 01/28/2022 20:01:20 - INFO - codeparrot_training - Step 5577: {'lr': 0.0004931800586734218, 'samples': 1070976, 'steps': 5577, 'loss/train': 1.6421467065811157} 01/28/2022 20:01:27 - INFO - codeparrot_training - Step 5578: {'lr': 0.0004931762623666692, 'samples': 1071168, 'steps': 5578, 'loss/train': 2.162404239177704} 01/28/2022 20:01:31 - INFO - codeparrot_training - Step 5579: {'lr': 0.0004931724650182268, 'samples': 1071360, 'steps': 5579, 'loss/train': 1.7146010398864746} 01/28/2022 20:01:35 - INFO - codeparrot_training - Step 5580: {'lr': 0.0004931686666281108, 'samples': 1071552, 'steps': 5580, 'loss/train': 1.6699937582015991} 01/28/2022 20:01:39 - INFO - codeparrot_training - Step 5581: {'lr': 0.0004931648671963373, 'samples': 1071744, 'steps': 5581, 'loss/train': 2.8467630743980408} 01/28/2022 20:01:44 - INFO - codeparrot_training - Step 5582: {'lr': 0.000493161066722923, 'samples': 1071936, 'steps': 5582, 'loss/train': 1.9511879682540894} 01/28/2022 20:01:48 - INFO - codeparrot_training - Step 5583: {'lr': 0.0004931572652078837, 'samples': 1072128, 'steps': 5583, 'loss/train': 2.513049066066742} 01/28/2022 20:01:53 - INFO - codeparrot_training - Step 5584: {'lr': 0.0004931534626512359, 'samples': 1072320, 'steps': 5584, 'loss/train': 1.8999921083450317} 01/28/2022 20:01:57 - INFO - codeparrot_training - Step 5585: {'lr': 0.0004931496590529959, 'samples': 1072512, 'steps': 5585, 'loss/train': 1.55728679895401} 01/28/2022 20:02:02 - INFO - codeparrot_training - Step 5586: {'lr': 0.0004931458544131799, 'samples': 1072704, 'steps': 5586, 'loss/train': 1.268993765115738} 01/28/2022 20:02:06 - INFO - codeparrot_training - Step 5587: {'lr': 0.0004931420487318044, 'samples': 1072896, 'steps': 5587, 'loss/train': 1.9598214626312256} 01/28/2022 20:02:10 - INFO - codeparrot_training - Step 5588: {'lr': 0.0004931382420088855, 'samples': 1073088, 'steps': 5588, 'loss/train': 2.145939588546753} 01/28/2022 20:02:15 - INFO - codeparrot_training - Step 5589: {'lr': 0.0004931344342444396, 'samples': 1073280, 'steps': 5589, 'loss/train': 2.820390522480011} 01/28/2022 20:02:20 - INFO - codeparrot_training - Step 5590: {'lr': 0.000493130625438483, 'samples': 1073472, 'steps': 5590, 'loss/train': 1.7992082834243774} 01/28/2022 20:02:24 - INFO - codeparrot_training - Step 5591: {'lr': 0.000493126815591032, 'samples': 1073664, 'steps': 5591, 'loss/train': 1.4377597868442535} 01/28/2022 20:02:28 - INFO - codeparrot_training - Step 5592: {'lr': 0.0004931230047021028, 'samples': 1073856, 'steps': 5592, 'loss/train': 4.586284875869751} 01/28/2022 20:02:32 - INFO - codeparrot_training - Step 5593: {'lr': 0.000493119192771712, 'samples': 1074048, 'steps': 5593, 'loss/train': 1.4688831567764282} 01/28/2022 20:02:38 - INFO - codeparrot_training - Step 5594: {'lr': 0.0004931153797998757, 'samples': 1074240, 'steps': 5594, 'loss/train': 1.5857187509536743} 01/28/2022 20:02:42 - INFO - codeparrot_training - Step 5595: {'lr': 0.0004931115657866103, 'samples': 1074432, 'steps': 5595, 'loss/train': 1.612208604812622} 01/28/2022 20:02:46 - INFO - codeparrot_training - Step 5596: {'lr': 0.0004931077507319322, 'samples': 1074624, 'steps': 5596, 'loss/train': 1.5756912231445312} 01/28/2022 20:02:51 - INFO - codeparrot_training - Step 5597: {'lr': 0.0004931039346358577, 'samples': 1074816, 'steps': 5597, 'loss/train': 1.5591807961463928} 01/28/2022 20:02:55 - INFO - codeparrot_training - Step 5598: {'lr': 0.0004931001174984032, 'samples': 1075008, 'steps': 5598, 'loss/train': 1.9131388664245605} 01/28/2022 20:03:01 - INFO - codeparrot_training - Step 5599: {'lr': 0.0004930962993195848, 'samples': 1075200, 'steps': 5599, 'loss/train': 3.2119710445404053} 01/28/2022 20:03:06 - INFO - codeparrot_training - Step 5600: {'lr': 0.0004930924800994192, 'samples': 1075392, 'steps': 5600, 'loss/train': 2.3548051714897156} 01/28/2022 20:03:10 - INFO - codeparrot_training - Step 5601: {'lr': 0.0004930886598379225, 'samples': 1075584, 'steps': 5601, 'loss/train': 2.04967600107193} 01/28/2022 20:03:14 - INFO - codeparrot_training - Step 5602: {'lr': 0.0004930848385351112, 'samples': 1075776, 'steps': 5602, 'loss/train': 2.4474152326583862} 01/28/2022 20:03:18 - INFO - codeparrot_training - Step 5603: {'lr': 0.0004930810161910017, 'samples': 1075968, 'steps': 5603, 'loss/train': 2.015217661857605} 01/28/2022 20:03:24 - INFO - codeparrot_training - Step 5604: {'lr': 0.0004930771928056102, 'samples': 1076160, 'steps': 5604, 'loss/train': 1.5535022020339966} 01/28/2022 20:03:28 - INFO - codeparrot_training - Step 5605: {'lr': 0.0004930733683789533, 'samples': 1076352, 'steps': 5605, 'loss/train': 1.5635823011398315} 01/28/2022 20:03:32 - INFO - codeparrot_training - Step 5606: {'lr': 0.0004930695429110473, 'samples': 1076544, 'steps': 5606, 'loss/train': 1.8263969421386719} 01/28/2022 20:03:36 - INFO - codeparrot_training - Step 5607: {'lr': 0.0004930657164019085, 'samples': 1076736, 'steps': 5607, 'loss/train': 2.0450729727745056} 01/28/2022 20:03:41 - INFO - codeparrot_training - Step 5608: {'lr': 0.0004930618888515534, 'samples': 1076928, 'steps': 5608, 'loss/train': 1.6227770447731018} 01/28/2022 20:03:47 - INFO - codeparrot_training - Step 5609: {'lr': 0.0004930580602599983, 'samples': 1077120, 'steps': 5609, 'loss/train': 1.686212182044983} 01/28/2022 20:03:51 - INFO - codeparrot_training - Step 5610: {'lr': 0.0004930542306272596, 'samples': 1077312, 'steps': 5610, 'loss/train': 1.6889145970344543} 01/28/2022 20:03:55 - INFO - codeparrot_training - Step 5611: {'lr': 0.0004930503999533538, 'samples': 1077504, 'steps': 5611, 'loss/train': 2.264702260494232} 01/28/2022 20:04:00 - INFO - codeparrot_training - Step 5612: {'lr': 0.0004930465682382973, 'samples': 1077696, 'steps': 5612, 'loss/train': 2.1826486587524414} 01/28/2022 20:04:04 - INFO - codeparrot_training - Step 5613: {'lr': 0.0004930427354821064, 'samples': 1077888, 'steps': 5613, 'loss/train': 1.6174067258834839} 01/28/2022 20:04:09 - INFO - codeparrot_training - Step 5614: {'lr': 0.0004930389016847977, 'samples': 1078080, 'steps': 5614, 'loss/train': 1.5034008622169495} 01/28/2022 20:04:13 - INFO - codeparrot_training - Step 5615: {'lr': 0.0004930350668463874, 'samples': 1078272, 'steps': 5615, 'loss/train': 0.44450703263282776} 01/28/2022 20:04:18 - INFO - codeparrot_training - Step 5616: {'lr': 0.0004930312309668922, 'samples': 1078464, 'steps': 5616, 'loss/train': 1.7653495073318481} 01/28/2022 20:04:22 - INFO - codeparrot_training - Step 5617: {'lr': 0.0004930273940463283, 'samples': 1078656, 'steps': 5617, 'loss/train': 2.2073437571525574} 01/28/2022 20:04:26 - INFO - codeparrot_training - Step 5618: {'lr': 0.0004930235560847121, 'samples': 1078848, 'steps': 5618, 'loss/train': 2.0823474526405334} 01/28/2022 20:04:32 - INFO - codeparrot_training - Step 5619: {'lr': 0.0004930197170820603, 'samples': 1079040, 'steps': 5619, 'loss/train': 1.9839824438095093} 01/28/2022 20:04:36 - INFO - codeparrot_training - Step 5620: {'lr': 0.0004930158770383891, 'samples': 1079232, 'steps': 5620, 'loss/train': 2.206943929195404} 01/28/2022 20:04:40 - INFO - codeparrot_training - Step 5621: {'lr': 0.0004930120359537153, 'samples': 1079424, 'steps': 5621, 'loss/train': 1.7115746140480042} 01/28/2022 20:04:44 - INFO - codeparrot_training - Step 5622: {'lr': 0.0004930081938280548, 'samples': 1079616, 'steps': 5622, 'loss/train': 2.549786925315857} 01/28/2022 20:04:49 - INFO - codeparrot_training - Step 5623: {'lr': 0.0004930043506614245, 'samples': 1079808, 'steps': 5623, 'loss/train': 1.7871896624565125} 01/28/2022 20:04:53 - INFO - codeparrot_training - Step 5624: {'lr': 0.0004930005064538406, 'samples': 1080000, 'steps': 5624, 'loss/train': 0.5292388200759888} 01/28/2022 20:05:00 - INFO - codeparrot_training - Step 5625: {'lr': 0.0004929966612053199, 'samples': 1080192, 'steps': 5625, 'loss/train': 1.4804711937904358} 01/28/2022 20:05:04 - INFO - codeparrot_training - Step 5626: {'lr': 0.0004929928149158785, 'samples': 1080384, 'steps': 5626, 'loss/train': 3.849830746650696} 01/28/2022 20:05:08 - INFO - codeparrot_training - Step 5627: {'lr': 0.0004929889675855332, 'samples': 1080576, 'steps': 5627, 'loss/train': 1.7788535356521606} 01/28/2022 20:05:12 - INFO - codeparrot_training - Step 5628: {'lr': 0.0004929851192143001, 'samples': 1080768, 'steps': 5628, 'loss/train': 1.7526694536209106} 01/28/2022 20:05:17 - INFO - codeparrot_training - Step 5629: {'lr': 0.0004929812698021961, 'samples': 1080960, 'steps': 5629, 'loss/train': 2.04110187292099} 01/28/2022 20:05:22 - INFO - codeparrot_training - Step 5630: {'lr': 0.0004929774193492373, 'samples': 1081152, 'steps': 5630, 'loss/train': 1.7416870594024658} 01/28/2022 20:05:27 - INFO - codeparrot_training - Step 5631: {'lr': 0.0004929735678554406, 'samples': 1081344, 'steps': 5631, 'loss/train': 1.6671715378761292} 01/28/2022 20:05:31 - INFO - codeparrot_training - Step 5632: {'lr': 0.0004929697153208221, 'samples': 1081536, 'steps': 5632, 'loss/train': 2.0458168387413025} 01/28/2022 20:05:35 - INFO - codeparrot_training - Step 5633: {'lr': 0.0004929658617453986, 'samples': 1081728, 'steps': 5633, 'loss/train': 1.7623277306556702} 01/28/2022 20:05:39 - INFO - codeparrot_training - Step 5634: {'lr': 0.0004929620071291865, 'samples': 1081920, 'steps': 5634, 'loss/train': 7.065927028656006} 01/28/2022 20:05:45 - INFO - codeparrot_training - Step 5635: {'lr': 0.0004929581514722023, 'samples': 1082112, 'steps': 5635, 'loss/train': 1.57441645860672} 01/28/2022 20:05:49 - INFO - codeparrot_training - Step 5636: {'lr': 0.0004929542947744625, 'samples': 1082304, 'steps': 5636, 'loss/train': 2.232717275619507} 01/28/2022 20:05:53 - INFO - codeparrot_training - Step 5637: {'lr': 0.0004929504370359837, 'samples': 1082496, 'steps': 5637, 'loss/train': 0.14042961224913597} 01/28/2022 20:05:57 - INFO - codeparrot_training - Step 5638: {'lr': 0.0004929465782567824, 'samples': 1082688, 'steps': 5638, 'loss/train': 2.7733590602874756} 01/28/2022 20:06:01 - INFO - codeparrot_training - Step 5639: {'lr': 0.000492942718436875, 'samples': 1082880, 'steps': 5639, 'loss/train': 1.4264169931411743} 01/28/2022 20:06:08 - INFO - codeparrot_training - Step 5640: {'lr': 0.0004929388575762782, 'samples': 1083072, 'steps': 5640, 'loss/train': 2.106772720813751} 01/28/2022 20:06:12 - INFO - codeparrot_training - Step 5641: {'lr': 0.0004929349956750085, 'samples': 1083264, 'steps': 5641, 'loss/train': 0.6992417424917221} 01/28/2022 20:06:16 - INFO - codeparrot_training - Step 5642: {'lr': 0.0004929311327330823, 'samples': 1083456, 'steps': 5642, 'loss/train': 1.5088359117507935} 01/28/2022 20:06:20 - INFO - codeparrot_training - Step 5643: {'lr': 0.0004929272687505163, 'samples': 1083648, 'steps': 5643, 'loss/train': 1.9927750825881958} 01/28/2022 20:06:25 - INFO - codeparrot_training - Step 5644: {'lr': 0.0004929234037273271, 'samples': 1083840, 'steps': 5644, 'loss/train': 2.6345622539520264} 01/28/2022 20:06:30 - INFO - codeparrot_training - Step 5645: {'lr': 0.0004929195376635311, 'samples': 1084032, 'steps': 5645, 'loss/train': 1.838677704334259} 01/28/2022 20:06:34 - INFO - codeparrot_training - Step 5646: {'lr': 0.000492915670559145, 'samples': 1084224, 'steps': 5646, 'loss/train': 2.0267200469970703} 01/28/2022 20:06:38 - INFO - codeparrot_training - Step 5647: {'lr': 0.0004929118024141853, 'samples': 1084416, 'steps': 5647, 'loss/train': 1.3692093193531036} 01/28/2022 20:06:43 - INFO - codeparrot_training - Step 5648: {'lr': 0.0004929079332286685, 'samples': 1084608, 'steps': 5648, 'loss/train': 1.5709285140037537} 01/28/2022 20:06:47 - INFO - codeparrot_training - Step 5649: {'lr': 0.0004929040630026112, 'samples': 1084800, 'steps': 5649, 'loss/train': 1.9128960371017456} 01/28/2022 20:06:53 - INFO - codeparrot_training - Step 5650: {'lr': 0.0004929001917360302, 'samples': 1084992, 'steps': 5650, 'loss/train': 1.087609738111496} 01/28/2022 20:06:57 - INFO - codeparrot_training - Step 5651: {'lr': 0.0004928963194289419, 'samples': 1085184, 'steps': 5651, 'loss/train': 1.2305847108364105} 01/28/2022 20:07:02 - INFO - codeparrot_training - Step 5652: {'lr': 0.0004928924460813627, 'samples': 1085376, 'steps': 5652, 'loss/train': 2.1067519783973694} 01/28/2022 20:07:06 - INFO - codeparrot_training - Step 5653: {'lr': 0.0004928885716933096, 'samples': 1085568, 'steps': 5653, 'loss/train': 2.3597286343574524} 01/28/2022 20:07:10 - INFO - codeparrot_training - Step 5654: {'lr': 0.0004928846962647988, 'samples': 1085760, 'steps': 5654, 'loss/train': 1.3417605757713318} 01/28/2022 20:07:15 - INFO - codeparrot_training - Step 5655: {'lr': 0.0004928808197958472, 'samples': 1085952, 'steps': 5655, 'loss/train': 0.9289421439170837} 01/28/2022 20:07:20 - INFO - codeparrot_training - Step 5656: {'lr': 0.0004928769422864712, 'samples': 1086144, 'steps': 5656, 'loss/train': 2.007850706577301} 01/28/2022 20:07:24 - INFO - codeparrot_training - Step 5657: {'lr': 0.0004928730637366877, 'samples': 1086336, 'steps': 5657, 'loss/train': 1.6206876039505005} 01/28/2022 20:07:28 - INFO - codeparrot_training - Step 5658: {'lr': 0.000492869184146513, 'samples': 1086528, 'steps': 5658, 'loss/train': 1.9199861884117126} 01/28/2022 20:07:32 - INFO - codeparrot_training - Step 5659: {'lr': 0.0004928653035159638, 'samples': 1086720, 'steps': 5659, 'loss/train': 0.6624079048633575} 01/28/2022 20:07:38 - INFO - codeparrot_training - Step 5660: {'lr': 0.0004928614218450568, 'samples': 1086912, 'steps': 5660, 'loss/train': 2.511889636516571} 01/28/2022 20:07:42 - INFO - codeparrot_training - Step 5661: {'lr': 0.0004928575391338085, 'samples': 1087104, 'steps': 5661, 'loss/train': 2.42636239528656} 01/28/2022 20:07:46 - INFO - codeparrot_training - Step 5662: {'lr': 0.0004928536553822357, 'samples': 1087296, 'steps': 5662, 'loss/train': 2.30062597990036} 01/28/2022 20:07:51 - INFO - codeparrot_training - Step 5663: {'lr': 0.0004928497705903549, 'samples': 1087488, 'steps': 5663, 'loss/train': 1.7146772146224976} 01/28/2022 20:07:55 - INFO - codeparrot_training - Step 5664: {'lr': 0.0004928458847581828, 'samples': 1087680, 'steps': 5664, 'loss/train': 1.2527234852313995} 01/28/2022 20:08:00 - INFO - codeparrot_training - Step 5665: {'lr': 0.0004928419978857361, 'samples': 1087872, 'steps': 5665, 'loss/train': 1.7375457286834717} 01/28/2022 20:08:04 - INFO - codeparrot_training - Step 5666: {'lr': 0.0004928381099730314, 'samples': 1088064, 'steps': 5666, 'loss/train': 2.0544426441192627} 01/28/2022 20:08:09 - INFO - codeparrot_training - Step 5667: {'lr': 0.0004928342210200853, 'samples': 1088256, 'steps': 5667, 'loss/train': 1.5121893882751465} 01/28/2022 20:08:13 - INFO - codeparrot_training - Step 5668: {'lr': 0.0004928303310269145, 'samples': 1088448, 'steps': 5668, 'loss/train': 2.6343042254447937} 01/28/2022 20:08:17 - INFO - codeparrot_training - Step 5669: {'lr': 0.0004928264399935357, 'samples': 1088640, 'steps': 5669, 'loss/train': 1.9586938619613647} 01/28/2022 20:08:23 - INFO - codeparrot_training - Step 5670: {'lr': 0.0004928225479199655, 'samples': 1088832, 'steps': 5670, 'loss/train': 1.7767769694328308} 01/28/2022 20:08:28 - INFO - codeparrot_training - Step 5671: {'lr': 0.0004928186548062206, 'samples': 1089024, 'steps': 5671, 'loss/train': 2.1441312432289124} 01/28/2022 20:08:32 - INFO - codeparrot_training - Step 5672: {'lr': 0.0004928147606523179, 'samples': 1089216, 'steps': 5672, 'loss/train': 1.959338665008545} 01/28/2022 20:08:36 - INFO - codeparrot_training - Step 5673: {'lr': 0.0004928108654582736, 'samples': 1089408, 'steps': 5673, 'loss/train': 1.6262730360031128} 01/28/2022 20:08:40 - INFO - codeparrot_training - Step 5674: {'lr': 0.0004928069692241048, 'samples': 1089600, 'steps': 5674, 'loss/train': 1.8987649083137512} 01/28/2022 20:08:46 - INFO - codeparrot_training - Step 5675: {'lr': 0.000492803071949828, 'samples': 1089792, 'steps': 5675, 'loss/train': 2.245052218437195} 01/28/2022 20:08:50 - INFO - codeparrot_training - Step 5676: {'lr': 0.0004927991736354599, 'samples': 1089984, 'steps': 5676, 'loss/train': 2.0397972464561462} 01/28/2022 20:08:54 - INFO - codeparrot_training - Step 5677: {'lr': 0.0004927952742810173, 'samples': 1090176, 'steps': 5677, 'loss/train': 1.7639175653457642} 01/28/2022 20:08:59 - INFO - codeparrot_training - Step 5678: {'lr': 0.0004927913738865167, 'samples': 1090368, 'steps': 5678, 'loss/train': 2.216033935546875} 01/28/2022 20:09:03 - INFO - codeparrot_training - Step 5679: {'lr': 0.0004927874724519751, 'samples': 1090560, 'steps': 5679, 'loss/train': 1.4625580608844757} 01/28/2022 20:09:09 - INFO - codeparrot_training - Step 5680: {'lr': 0.000492783569977409, 'samples': 1090752, 'steps': 5680, 'loss/train': 2.156005024909973} 01/28/2022 20:09:13 - INFO - codeparrot_training - Step 5681: {'lr': 0.0004927796664628353, 'samples': 1090944, 'steps': 5681, 'loss/train': 5.026713609695435} 01/28/2022 20:09:17 - INFO - codeparrot_training - Step 5682: {'lr': 0.0004927757619082704, 'samples': 1091136, 'steps': 5682, 'loss/train': 2.7026050686836243} 01/28/2022 20:09:21 - INFO - codeparrot_training - Step 5683: {'lr': 0.0004927718563137313, 'samples': 1091328, 'steps': 5683, 'loss/train': 2.8367221355438232} 01/28/2022 20:09:26 - INFO - codeparrot_training - Step 5684: {'lr': 0.0004927679496792347, 'samples': 1091520, 'steps': 5684, 'loss/train': 2.106757879257202} 01/28/2022 20:09:30 - INFO - codeparrot_training - Step 5685: {'lr': 0.0004927640420047973, 'samples': 1091712, 'steps': 5685, 'loss/train': 1.5864785313606262} 01/28/2022 20:09:36 - INFO - codeparrot_training - Step 5686: {'lr': 0.0004927601332904358, 'samples': 1091904, 'steps': 5686, 'loss/train': 2.113446056842804} 01/28/2022 20:09:40 - INFO - codeparrot_training - Step 5687: {'lr': 0.0004927562235361669, 'samples': 1092096, 'steps': 5687, 'loss/train': 2.1247512102127075} 01/28/2022 20:09:45 - INFO - codeparrot_training - Step 5688: {'lr': 0.0004927523127420076, 'samples': 1092288, 'steps': 5688, 'loss/train': 1.1593314707279205} 01/28/2022 20:09:49 - INFO - codeparrot_training - Step 5689: {'lr': 0.0004927484009079743, 'samples': 1092480, 'steps': 5689, 'loss/train': 1.6312211751937866} 01/28/2022 20:09:53 - INFO - codeparrot_training - Step 5690: {'lr': 0.000492744488034084, 'samples': 1092672, 'steps': 5690, 'loss/train': 1.6575361490249634} 01/28/2022 20:09:59 - INFO - codeparrot_training - Step 5691: {'lr': 0.0004927405741203534, 'samples': 1092864, 'steps': 5691, 'loss/train': 2.2930403351783752} 01/28/2022 20:10:03 - INFO - codeparrot_training - Step 5692: {'lr': 0.0004927366591667993, 'samples': 1093056, 'steps': 5692, 'loss/train': 2.7972537875175476} 01/28/2022 20:10:07 - INFO - codeparrot_training - Step 5693: {'lr': 0.0004927327431734383, 'samples': 1093248, 'steps': 5693, 'loss/train': 2.316022217273712} 01/28/2022 20:10:12 - INFO - codeparrot_training - Step 5694: {'lr': 0.0004927288261402875, 'samples': 1093440, 'steps': 5694, 'loss/train': 1.6661233305931091} 01/28/2022 20:10:18 - INFO - codeparrot_training - Step 5695: {'lr': 0.0004927249080673633, 'samples': 1093632, 'steps': 5695, 'loss/train': 2.176627457141876} 01/28/2022 20:10:22 - INFO - codeparrot_training - Step 5696: {'lr': 0.0004927209889546828, 'samples': 1093824, 'steps': 5696, 'loss/train': 2.1871621012687683} 01/28/2022 20:10:26 - INFO - codeparrot_training - Step 5697: {'lr': 0.0004927170688022625, 'samples': 1094016, 'steps': 5697, 'loss/train': 1.9072796702384949} 01/28/2022 20:10:30 - INFO - codeparrot_training - Step 5698: {'lr': 0.0004927131476101195, 'samples': 1094208, 'steps': 5698, 'loss/train': 1.36478653550148} 01/28/2022 20:10:35 - INFO - codeparrot_training - Step 5699: {'lr': 0.0004927092253782704, 'samples': 1094400, 'steps': 5699, 'loss/train': 1.0377649068832397} 01/28/2022 20:10:40 - INFO - codeparrot_training - Step 5700: {'lr': 0.0004927053021067321, 'samples': 1094592, 'steps': 5700, 'loss/train': 1.0253476202487946} 01/28/2022 20:10:44 - INFO - codeparrot_training - Step 5701: {'lr': 0.0004927013777955212, 'samples': 1094784, 'steps': 5701, 'loss/train': 0.8377898633480072} 01/28/2022 20:10:48 - INFO - codeparrot_training - Step 5702: {'lr': 0.0004926974524446548, 'samples': 1094976, 'steps': 5702, 'loss/train': 1.7954224348068237} 01/28/2022 20:10:53 - INFO - codeparrot_training - Step 5703: {'lr': 0.0004926935260541496, 'samples': 1095168, 'steps': 5703, 'loss/train': 1.2886857390403748} 01/28/2022 20:10:57 - INFO - codeparrot_training - Step 5704: {'lr': 0.0004926895986240222, 'samples': 1095360, 'steps': 5704, 'loss/train': 1.0878377258777618} 01/28/2022 20:11:03 - INFO - codeparrot_training - Step 5705: {'lr': 0.0004926856701542898, 'samples': 1095552, 'steps': 5705, 'loss/train': 1.7708389163017273} 01/28/2022 20:11:07 - INFO - codeparrot_training - Step 5706: {'lr': 0.000492681740644969, 'samples': 1095744, 'steps': 5706, 'loss/train': 1.757194697856903} 01/28/2022 20:11:11 - INFO - codeparrot_training - Step 5707: {'lr': 0.0004926778100960767, 'samples': 1095936, 'steps': 5707, 'loss/train': 1.4022138118743896} 01/28/2022 20:11:15 - INFO - codeparrot_training - Step 5708: {'lr': 0.0004926738785076297, 'samples': 1096128, 'steps': 5708, 'loss/train': 2.8659825325012207} 01/28/2022 20:11:20 - INFO - codeparrot_training - Step 5709: {'lr': 0.0004926699458796448, 'samples': 1096320, 'steps': 5709, 'loss/train': 1.1088743805885315} 01/28/2022 20:11:24 - INFO - codeparrot_training - Step 5710: {'lr': 0.0004926660122121391, 'samples': 1096512, 'steps': 5710, 'loss/train': 0.6382728070020676} 01/28/2022 20:11:31 - INFO - codeparrot_training - Step 5711: {'lr': 0.0004926620775051291, 'samples': 1096704, 'steps': 5711, 'loss/train': 4.922843098640442} 01/28/2022 20:11:35 - INFO - codeparrot_training - Step 5712: {'lr': 0.0004926581417586318, 'samples': 1096896, 'steps': 5712, 'loss/train': 3.833094835281372} 01/28/2022 20:11:39 - INFO - codeparrot_training - Step 5713: {'lr': 0.0004926542049726642, 'samples': 1097088, 'steps': 5713, 'loss/train': 1.7583872079849243} 01/28/2022 20:11:44 - INFO - codeparrot_training - Step 5714: {'lr': 0.0004926502671472429, 'samples': 1097280, 'steps': 5714, 'loss/train': 0.7224616706371307} 01/28/2022 20:11:48 - INFO - codeparrot_training - Step 5715: {'lr': 0.000492646328282385, 'samples': 1097472, 'steps': 5715, 'loss/train': 1.7170143127441406} 01/28/2022 20:11:52 - INFO - codeparrot_training - Step 5716: {'lr': 0.0004926423883781073, 'samples': 1097664, 'steps': 5716, 'loss/train': 1.1520699262619019} 01/28/2022 20:11:56 - INFO - codeparrot_training - Step 5717: {'lr': 0.0004926384474344265, 'samples': 1097856, 'steps': 5717, 'loss/train': 1.931938111782074} 01/28/2022 20:12:02 - INFO - codeparrot_training - Step 5718: {'lr': 0.0004926345054513598, 'samples': 1098048, 'steps': 5718, 'loss/train': 2.255440056324005} 01/28/2022 20:12:06 - INFO - codeparrot_training - Step 5719: {'lr': 0.0004926305624289238, 'samples': 1098240, 'steps': 5719, 'loss/train': 1.5945294499397278} 01/28/2022 20:12:10 - INFO - codeparrot_training - Step 5720: {'lr': 0.0004926266183671356, 'samples': 1098432, 'steps': 5720, 'loss/train': 1.8335598707199097} 01/28/2022 20:12:15 - INFO - codeparrot_training - Step 5721: {'lr': 0.000492622673266012, 'samples': 1098624, 'steps': 5721, 'loss/train': 1.5550745129585266} 01/28/2022 20:12:20 - INFO - codeparrot_training - Step 5722: {'lr': 0.0004926187271255698, 'samples': 1098816, 'steps': 5722, 'loss/train': 2.172136902809143} 01/28/2022 20:12:24 - INFO - codeparrot_training - Step 5723: {'lr': 0.0004926147799458262, 'samples': 1099008, 'steps': 5723, 'loss/train': 2.067994236946106} 01/28/2022 20:12:28 - INFO - codeparrot_training - Step 5724: {'lr': 0.0004926108317267979, 'samples': 1099200, 'steps': 5724, 'loss/train': 1.0465304553508759} 01/28/2022 20:12:33 - INFO - codeparrot_training - Step 5725: {'lr': 0.0004926068824685017, 'samples': 1099392, 'steps': 5725, 'loss/train': 0.3912148028612137} 01/28/2022 20:12:37 - INFO - codeparrot_training - Step 5726: {'lr': 0.0004926029321709548, 'samples': 1099584, 'steps': 5726, 'loss/train': 2.516352117061615} 01/28/2022 20:12:41 - INFO - codeparrot_training - Step 5727: {'lr': 0.0004925989808341738, 'samples': 1099776, 'steps': 5727, 'loss/train': 1.5104066133499146} 01/28/2022 20:12:47 - INFO - codeparrot_training - Step 5728: {'lr': 0.0004925950284581759, 'samples': 1099968, 'steps': 5728, 'loss/train': 1.8289545178413391} 01/28/2022 20:12:51 - INFO - codeparrot_training - Step 5729: {'lr': 0.0004925910750429779, 'samples': 1100160, 'steps': 5729, 'loss/train': 2.2942264080047607} 01/28/2022 20:12:55 - INFO - codeparrot_training - Step 5730: {'lr': 0.0004925871205885968, 'samples': 1100352, 'steps': 5730, 'loss/train': 1.4267241954803467} 01/28/2022 20:13:00 - INFO - codeparrot_training - Step 5731: {'lr': 0.0004925831650950495, 'samples': 1100544, 'steps': 5731, 'loss/train': 1.5425789952278137} 01/28/2022 20:13:04 - INFO - codeparrot_training - Step 5732: {'lr': 0.000492579208562353, 'samples': 1100736, 'steps': 5732, 'loss/train': 1.1576204001903534} 01/28/2022 20:13:10 - INFO - codeparrot_training - Step 5733: {'lr': 0.0004925752509905241, 'samples': 1100928, 'steps': 5733, 'loss/train': 2.139143228530884} 01/28/2022 20:13:14 - INFO - codeparrot_training - Step 5734: {'lr': 0.0004925712923795799, 'samples': 1101120, 'steps': 5734, 'loss/train': 1.3044677674770355} 01/28/2022 20:13:18 - INFO - codeparrot_training - Step 5735: {'lr': 0.0004925673327295374, 'samples': 1101312, 'steps': 5735, 'loss/train': 1.7040714025497437} 01/28/2022 20:13:23 - INFO - codeparrot_training - Step 5736: {'lr': 0.0004925633720404132, 'samples': 1101504, 'steps': 5736, 'loss/train': 1.976845622062683} 01/28/2022 20:13:27 - INFO - codeparrot_training - Step 5737: {'lr': 0.0004925594103122248, 'samples': 1101696, 'steps': 5737, 'loss/train': 1.7763762474060059} 01/28/2022 20:13:32 - INFO - codeparrot_training - Step 5738: {'lr': 0.0004925554475449888, 'samples': 1101888, 'steps': 5738, 'loss/train': 2.246658504009247} 01/28/2022 20:13:37 - INFO - codeparrot_training - Step 5739: {'lr': 0.0004925514837387223, 'samples': 1102080, 'steps': 5739, 'loss/train': 2.049126148223877} 01/28/2022 20:13:41 - INFO - codeparrot_training - Step 5740: {'lr': 0.0004925475188934423, 'samples': 1102272, 'steps': 5740, 'loss/train': 1.1274425089359283} 01/28/2022 20:13:45 - INFO - codeparrot_training - Step 5741: {'lr': 0.0004925435530091656, 'samples': 1102464, 'steps': 5741, 'loss/train': 1.0902821123600006} 01/28/2022 20:13:52 - INFO - codeparrot_training - Step 5742: {'lr': 0.0004925395860859096, 'samples': 1102656, 'steps': 5742, 'loss/train': 2.3506967425346375} 01/28/2022 20:13:56 - INFO - codeparrot_training - Step 5743: {'lr': 0.0004925356181236908, 'samples': 1102848, 'steps': 5743, 'loss/train': 1.324884295463562} 01/28/2022 20:14:00 - INFO - codeparrot_training - Step 5744: {'lr': 0.0004925316491225265, 'samples': 1103040, 'steps': 5744, 'loss/train': 1.992804229259491} 01/28/2022 20:14:04 - INFO - codeparrot_training - Step 5745: {'lr': 0.0004925276790824336, 'samples': 1103232, 'steps': 5745, 'loss/train': 1.8629606366157532} 01/28/2022 20:14:09 - INFO - codeparrot_training - Step 5746: {'lr': 0.0004925237080034291, 'samples': 1103424, 'steps': 5746, 'loss/train': 1.7363630533218384} 01/28/2022 20:14:13 - INFO - codeparrot_training - Step 5747: {'lr': 0.0004925197358855301, 'samples': 1103616, 'steps': 5747, 'loss/train': 1.8442425727844238} 01/28/2022 20:14:18 - INFO - codeparrot_training - Step 5748: {'lr': 0.0004925157627287536, 'samples': 1103808, 'steps': 5748, 'loss/train': 1.2481496930122375} 01/28/2022 20:14:23 - INFO - codeparrot_training - Step 5749: {'lr': 0.0004925117885331166, 'samples': 1104000, 'steps': 5749, 'loss/train': 2.199500262737274} 01/28/2022 20:14:27 - INFO - codeparrot_training - Step 5750: {'lr': 0.000492507813298636, 'samples': 1104192, 'steps': 5750, 'loss/train': 2.703912913799286} 01/28/2022 20:14:31 - INFO - codeparrot_training - Step 5751: {'lr': 0.000492503837025329, 'samples': 1104384, 'steps': 5751, 'loss/train': 2.3216893672943115} 01/28/2022 20:14:36 - INFO - codeparrot_training - Step 5752: {'lr': 0.0004924998597132125, 'samples': 1104576, 'steps': 5752, 'loss/train': 1.8645261526107788} 01/28/2022 20:14:41 - INFO - codeparrot_training - Step 5753: {'lr': 0.0004924958813623037, 'samples': 1104768, 'steps': 5753, 'loss/train': 2.1293440461158752} 01/28/2022 20:14:45 - INFO - codeparrot_training - Step 5754: {'lr': 0.0004924919019726195, 'samples': 1104960, 'steps': 5754, 'loss/train': 1.528338611125946} 01/28/2022 20:14:49 - INFO - codeparrot_training - Step 5755: {'lr': 0.000492487921544177, 'samples': 1105152, 'steps': 5755, 'loss/train': 2.5816535353660583} 01/28/2022 20:14:53 - INFO - codeparrot_training - Step 5756: {'lr': 0.0004924839400769932, 'samples': 1105344, 'steps': 5756, 'loss/train': 1.7389276027679443} 01/28/2022 20:14:59 - INFO - codeparrot_training - Step 5757: {'lr': 0.0004924799575710852, 'samples': 1105536, 'steps': 5757, 'loss/train': 0.3824111670255661} 01/28/2022 20:15:04 - INFO - codeparrot_training - Step 5758: {'lr': 0.0004924759740264701, 'samples': 1105728, 'steps': 5758, 'loss/train': 1.5165749788284302} 01/28/2022 20:15:08 - INFO - codeparrot_training - Step 5759: {'lr': 0.000492471989443165, 'samples': 1105920, 'steps': 5759, 'loss/train': 1.3808351755142212} 01/28/2022 20:15:12 - INFO - codeparrot_training - Step 5760: {'lr': 0.0004924680038211868, 'samples': 1106112, 'steps': 5760, 'loss/train': 1.531145453453064} 01/28/2022 20:15:16 - INFO - codeparrot_training - Step 5761: {'lr': 0.0004924640171605526, 'samples': 1106304, 'steps': 5761, 'loss/train': 2.2176831364631653} 01/28/2022 20:15:22 - INFO - codeparrot_training - Step 5762: {'lr': 0.0004924600294612796, 'samples': 1106496, 'steps': 5762, 'loss/train': 2.654338002204895} 01/28/2022 20:15:26 - INFO - codeparrot_training - Step 5763: {'lr': 0.0004924560407233848, 'samples': 1106688, 'steps': 5763, 'loss/train': 2.283902585506439} 01/28/2022 20:15:30 - INFO - codeparrot_training - Step 5764: {'lr': 0.0004924520509468854, 'samples': 1106880, 'steps': 5764, 'loss/train': 0.9551873803138733} 01/28/2022 20:15:34 - INFO - codeparrot_training - Step 5765: {'lr': 0.0004924480601317982, 'samples': 1107072, 'steps': 5765, 'loss/train': 3.1603481769561768} 01/28/2022 20:15:39 - INFO - codeparrot_training - Step 5766: {'lr': 0.0004924440682781407, 'samples': 1107264, 'steps': 5766, 'loss/train': 1.8133360147476196} 01/28/2022 20:15:45 - INFO - codeparrot_training - Step 5767: {'lr': 0.0004924400753859297, 'samples': 1107456, 'steps': 5767, 'loss/train': 2.857934832572937} 01/28/2022 20:15:49 - INFO - codeparrot_training - Step 5768: {'lr': 0.0004924360814551825, 'samples': 1107648, 'steps': 5768, 'loss/train': 1.24388587474823} 01/28/2022 20:15:53 - INFO - codeparrot_training - Step 5769: {'lr': 0.000492432086485916, 'samples': 1107840, 'steps': 5769, 'loss/train': 2.8137223720550537} 01/28/2022 20:15:57 - INFO - codeparrot_training - Step 5770: {'lr': 0.0004924280904781475, 'samples': 1108032, 'steps': 5770, 'loss/train': 1.813486933708191} 01/28/2022 20:16:02 - INFO - codeparrot_training - Step 5771: {'lr': 0.0004924240934318939, 'samples': 1108224, 'steps': 5771, 'loss/train': 1.3101150691509247} 01/28/2022 20:16:07 - INFO - codeparrot_training - Step 5772: {'lr': 0.0004924200953471727, 'samples': 1108416, 'steps': 5772, 'loss/train': 1.3311118483543396} 01/28/2022 20:16:11 - INFO - codeparrot_training - Step 5773: {'lr': 0.0004924160962240005, 'samples': 1108608, 'steps': 5773, 'loss/train': 0.6125645488500595} 01/28/2022 20:16:16 - INFO - codeparrot_training - Step 5774: {'lr': 0.0004924120960623949, 'samples': 1108800, 'steps': 5774, 'loss/train': 1.558993935585022} 01/28/2022 20:16:20 - INFO - codeparrot_training - Step 5775: {'lr': 0.0004924080948623729, 'samples': 1108992, 'steps': 5775, 'loss/train': 2.644357681274414} 01/28/2022 20:16:24 - INFO - codeparrot_training - Step 5776: {'lr': 0.0004924040926239515, 'samples': 1109184, 'steps': 5776, 'loss/train': 2.8318832516670227} 01/28/2022 20:16:29 - INFO - codeparrot_training - Step 5777: {'lr': 0.000492400089347148, 'samples': 1109376, 'steps': 5777, 'loss/train': 1.8293083906173706} 01/28/2022 20:16:34 - INFO - codeparrot_training - Step 5778: {'lr': 0.0004923960850319794, 'samples': 1109568, 'steps': 5778, 'loss/train': 1.632811188697815} 01/28/2022 20:16:38 - INFO - codeparrot_training - Step 5779: {'lr': 0.000492392079678463, 'samples': 1109760, 'steps': 5779, 'loss/train': 1.9766762852668762} 01/28/2022 20:16:42 - INFO - codeparrot_training - Step 5780: {'lr': 0.0004923880732866159, 'samples': 1109952, 'steps': 5780, 'loss/train': 1.6707238554954529} 01/28/2022 20:16:46 - INFO - codeparrot_training - Step 5781: {'lr': 0.0004923840658564553, 'samples': 1110144, 'steps': 5781, 'loss/train': 1.9121475219726562} 01/28/2022 20:16:51 - INFO - codeparrot_training - Step 5782: {'lr': 0.0004923800573879983, 'samples': 1110336, 'steps': 5782, 'loss/train': 2.142524242401123} 01/28/2022 20:16:56 - INFO - codeparrot_training - Step 5783: {'lr': 0.000492376047881262, 'samples': 1110528, 'steps': 5783, 'loss/train': 1.3498700559139252} 01/28/2022 20:17:00 - INFO - codeparrot_training - Step 5784: {'lr': 0.0004923720373362638, 'samples': 1110720, 'steps': 5784, 'loss/train': 2.1436926126480103} 01/28/2022 20:17:05 - INFO - codeparrot_training - Step 5785: {'lr': 0.0004923680257530207, 'samples': 1110912, 'steps': 5785, 'loss/train': 1.7841389179229736} 01/28/2022 20:17:09 - INFO - codeparrot_training - Step 5786: {'lr': 0.0004923640131315499, 'samples': 1111104, 'steps': 5786, 'loss/train': 2.4976603388786316} 01/28/2022 20:17:15 - INFO - codeparrot_training - Step 5787: {'lr': 0.0004923599994718687, 'samples': 1111296, 'steps': 5787, 'loss/train': 2.5168265104293823} 01/28/2022 20:17:19 - INFO - codeparrot_training - Step 5788: {'lr': 0.0004923559847739941, 'samples': 1111488, 'steps': 5788, 'loss/train': 2.399801552295685} 01/28/2022 20:17:23 - INFO - codeparrot_training - Step 5789: {'lr': 0.0004923519690379436, 'samples': 1111680, 'steps': 5789, 'loss/train': 2.186178982257843} 01/28/2022 20:17:28 - INFO - codeparrot_training - Step 5790: {'lr': 0.0004923479522637341, 'samples': 1111872, 'steps': 5790, 'loss/train': 2.4611613750457764} 01/28/2022 20:17:32 - INFO - codeparrot_training - Step 5791: {'lr': 0.0004923439344513829, 'samples': 1112064, 'steps': 5791, 'loss/train': 1.126900166273117} 01/28/2022 20:17:37 - INFO - codeparrot_training - Step 5792: {'lr': 0.0004923399156009073, 'samples': 1112256, 'steps': 5792, 'loss/train': 1.8940762281417847} 01/28/2022 20:17:41 - INFO - codeparrot_training - Step 5793: {'lr': 0.0004923358957123245, 'samples': 1112448, 'steps': 5793, 'loss/train': 2.1554428339004517} 01/28/2022 20:17:46 - INFO - codeparrot_training - Step 5794: {'lr': 0.0004923318747856515, 'samples': 1112640, 'steps': 5794, 'loss/train': 1.9291378855705261} 01/28/2022 20:17:50 - INFO - codeparrot_training - Step 5795: {'lr': 0.0004923278528209059, 'samples': 1112832, 'steps': 5795, 'loss/train': 1.5932334065437317} 01/28/2022 20:17:54 - INFO - codeparrot_training - Step 5796: {'lr': 0.0004923238298181047, 'samples': 1113024, 'steps': 5796, 'loss/train': 1.824604332447052} 01/28/2022 20:18:00 - INFO - codeparrot_training - Step 5797: {'lr': 0.0004923198057772651, 'samples': 1113216, 'steps': 5797, 'loss/train': 1.176780492067337} 01/28/2022 20:18:04 - INFO - codeparrot_training - Step 5798: {'lr': 0.0004923157806984044, 'samples': 1113408, 'steps': 5798, 'loss/train': 2.6337543725967407} 01/28/2022 20:18:08 - INFO - codeparrot_training - Step 5799: {'lr': 0.0004923117545815398, 'samples': 1113600, 'steps': 5799, 'loss/train': 2.527872920036316} 01/28/2022 20:18:12 - INFO - codeparrot_training - Step 5800: {'lr': 0.0004923077274266886, 'samples': 1113792, 'steps': 5800, 'loss/train': 0.969960629940033} 01/28/2022 20:18:17 - INFO - codeparrot_training - Step 5801: {'lr': 0.0004923036992338681, 'samples': 1113984, 'steps': 5801, 'loss/train': 1.652945637702942} 01/28/2022 20:18:23 - INFO - codeparrot_training - Step 5802: {'lr': 0.0004922996700030954, 'samples': 1114176, 'steps': 5802, 'loss/train': 1.1006910502910614} 01/28/2022 20:18:27 - INFO - codeparrot_training - Step 5803: {'lr': 0.000492295639734388, 'samples': 1114368, 'steps': 5803, 'loss/train': 1.2005000710487366} 01/28/2022 20:18:32 - INFO - codeparrot_training - Step 5804: {'lr': 0.0004922916084277629, 'samples': 1114560, 'steps': 5804, 'loss/train': 1.9412044286727905} 01/28/2022 20:18:36 - INFO - codeparrot_training - Step 5805: {'lr': 0.0004922875760832375, 'samples': 1114752, 'steps': 5805, 'loss/train': 1.9460034370422363} 01/28/2022 20:18:40 - INFO - codeparrot_training - Step 5806: {'lr': 0.000492283542700829, 'samples': 1114944, 'steps': 5806, 'loss/train': 2.2485415935516357} 01/28/2022 20:18:45 - INFO - codeparrot_training - Step 5807: {'lr': 0.0004922795082805549, 'samples': 1115136, 'steps': 5807, 'loss/train': 2.4137737154960632} 01/28/2022 20:18:50 - INFO - codeparrot_training - Step 5808: {'lr': 0.0004922754728224322, 'samples': 1115328, 'steps': 5808, 'loss/train': 2.614463210105896} 01/28/2022 20:18:54 - INFO - codeparrot_training - Step 5809: {'lr': 0.0004922714363264783, 'samples': 1115520, 'steps': 5809, 'loss/train': 1.3832262754440308} 01/28/2022 20:18:58 - INFO - codeparrot_training - Step 5810: {'lr': 0.0004922673987927106, 'samples': 1115712, 'steps': 5810, 'loss/train': 0.6709170788526535} 01/28/2022 20:19:02 - INFO - codeparrot_training - Step 5811: {'lr': 0.0004922633602211462, 'samples': 1115904, 'steps': 5811, 'loss/train': 2.222984254360199} 01/28/2022 20:19:09 - INFO - codeparrot_training - Step 5812: {'lr': 0.0004922593206118025, 'samples': 1116096, 'steps': 5812, 'loss/train': 1.9105752110481262} 01/28/2022 20:19:13 - INFO - codeparrot_training - Step 5813: {'lr': 0.0004922552799646968, 'samples': 1116288, 'steps': 5813, 'loss/train': 1.8674826622009277} 01/28/2022 20:19:17 - INFO - codeparrot_training - Step 5814: {'lr': 0.0004922512382798463, 'samples': 1116480, 'steps': 5814, 'loss/train': 1.4977085888385773} 01/28/2022 20:19:21 - INFO - codeparrot_training - Step 5815: {'lr': 0.0004922471955572686, 'samples': 1116672, 'steps': 5815, 'loss/train': 3.6830577850341797} 01/28/2022 20:19:26 - INFO - codeparrot_training - Step 5816: {'lr': 0.0004922431517969808, 'samples': 1116864, 'steps': 5816, 'loss/train': 2.092954158782959} 01/28/2022 20:19:31 - INFO - codeparrot_training - Step 5817: {'lr': 0.0004922391069990002, 'samples': 1117056, 'steps': 5817, 'loss/train': 1.7946111559867859} 01/28/2022 20:19:35 - INFO - codeparrot_training - Step 5818: {'lr': 0.0004922350611633442, 'samples': 1117248, 'steps': 5818, 'loss/train': 1.5327996611595154} 01/28/2022 20:19:40 - INFO - codeparrot_training - Step 5819: {'lr': 0.0004922310142900302, 'samples': 1117440, 'steps': 5819, 'loss/train': 0.21776124089956284} 01/28/2022 20:19:44 - INFO - codeparrot_training - Step 5820: {'lr': 0.0004922269663790753, 'samples': 1117632, 'steps': 5820, 'loss/train': 1.9140644073486328} 01/28/2022 20:19:48 - INFO - codeparrot_training - Step 5821: {'lr': 0.0004922229174304971, 'samples': 1117824, 'steps': 5821, 'loss/train': 1.4156506061553955} 01/28/2022 20:19:54 - INFO - codeparrot_training - Step 5822: {'lr': 0.0004922188674443128, 'samples': 1118016, 'steps': 5822, 'loss/train': 1.1225794851779938} 01/28/2022 20:19:58 - INFO - codeparrot_training - Step 5823: {'lr': 0.0004922148164205398, 'samples': 1118208, 'steps': 5823, 'loss/train': 1.8788414597511292} 01/28/2022 20:20:02 - INFO - codeparrot_training - Step 5824: {'lr': 0.0004922107643591954, 'samples': 1118400, 'steps': 5824, 'loss/train': 1.7995301485061646} 01/28/2022 20:20:06 - INFO - codeparrot_training - Step 5825: {'lr': 0.000492206711260297, 'samples': 1118592, 'steps': 5825, 'loss/train': 1.1315126717090607} 01/28/2022 20:20:11 - INFO - codeparrot_training - Step 5826: {'lr': 0.000492202657123862, 'samples': 1118784, 'steps': 5826, 'loss/train': 2.002004384994507} 01/28/2022 20:20:17 - INFO - codeparrot_training - Step 5827: {'lr': 0.0004921986019499078, 'samples': 1118976, 'steps': 5827, 'loss/train': 1.9159353375434875} 01/28/2022 20:20:21 - INFO - codeparrot_training - Step 5828: {'lr': 0.0004921945457384516, 'samples': 1119168, 'steps': 5828, 'loss/train': 2.3863953351974487} 01/28/2022 20:20:25 - INFO - codeparrot_training - Step 5829: {'lr': 0.0004921904884895108, 'samples': 1119360, 'steps': 5829, 'loss/train': 1.949224054813385} 01/28/2022 20:20:30 - INFO - codeparrot_training - Step 5830: {'lr': 0.000492186430203103, 'samples': 1119552, 'steps': 5830, 'loss/train': 2.115087926387787} 01/28/2022 20:20:34 - INFO - codeparrot_training - Step 5831: {'lr': 0.0004921823708792453, 'samples': 1119744, 'steps': 5831, 'loss/train': 2.043090283870697} 01/28/2022 20:20:39 - INFO - codeparrot_training - Step 5832: {'lr': 0.0004921783105179552, 'samples': 1119936, 'steps': 5832, 'loss/train': 0.798082709312439} 01/28/2022 20:20:43 - INFO - codeparrot_training - Step 5833: {'lr': 0.0004921742491192502, 'samples': 1120128, 'steps': 5833, 'loss/train': 2.7235963940620422} 01/28/2022 20:20:48 - INFO - codeparrot_training - Step 5834: {'lr': 0.0004921701866831477, 'samples': 1120320, 'steps': 5834, 'loss/train': 1.8814542889595032} 01/28/2022 20:20:52 - INFO - codeparrot_training - Step 5835: {'lr': 0.000492166123209665, 'samples': 1120512, 'steps': 5835, 'loss/train': 1.367554396390915} 01/28/2022 20:20:56 - INFO - codeparrot_training - Step 5836: {'lr': 0.0004921620586988193, 'samples': 1120704, 'steps': 5836, 'loss/train': 1.901207149028778} 01/28/2022 20:21:01 - INFO - codeparrot_training - Step 5837: {'lr': 0.0004921579931506285, 'samples': 1120896, 'steps': 5837, 'loss/train': 2.1509045362472534} 01/28/2022 20:21:06 - INFO - codeparrot_training - Step 5838: {'lr': 0.0004921539265651096, 'samples': 1121088, 'steps': 5838, 'loss/train': 3.552248954772949} 01/28/2022 20:21:10 - INFO - codeparrot_training - Step 5839: {'lr': 0.0004921498589422803, 'samples': 1121280, 'steps': 5839, 'loss/train': 1.4628905653953552} 01/28/2022 20:21:14 - INFO - codeparrot_training - Step 5840: {'lr': 0.0004921457902821578, 'samples': 1121472, 'steps': 5840, 'loss/train': 1.2144729495048523} 01/28/2022 20:21:18 - INFO - codeparrot_training - Step 5841: {'lr': 0.0004921417205847597, 'samples': 1121664, 'steps': 5841, 'loss/train': 1.7282298803329468} 01/28/2022 20:21:24 - INFO - codeparrot_training - Step 5842: {'lr': 0.0004921376498501032, 'samples': 1121856, 'steps': 5842, 'loss/train': 1.5490000247955322} 01/28/2022 20:21:28 - INFO - codeparrot_training - Step 5843: {'lr': 0.000492133578078206, 'samples': 1122048, 'steps': 5843, 'loss/train': 1.0920668542385101} 01/28/2022 20:21:32 - INFO - codeparrot_training - Step 5844: {'lr': 0.0004921295052690855, 'samples': 1122240, 'steps': 5844, 'loss/train': 1.9413626790046692} 01/28/2022 20:21:36 - INFO - codeparrot_training - Step 5845: {'lr': 0.000492125431422759, 'samples': 1122432, 'steps': 5845, 'loss/train': 1.5016334652900696} 01/28/2022 20:21:41 - INFO - codeparrot_training - Step 5846: {'lr': 0.0004921213565392441, 'samples': 1122624, 'steps': 5846, 'loss/train': 2.1399418115615845} 01/28/2022 20:21:47 - INFO - codeparrot_training - Step 5847: {'lr': 0.000492117280618558, 'samples': 1122816, 'steps': 5847, 'loss/train': 3.980440378189087} 01/28/2022 20:21:51 - INFO - codeparrot_training - Step 5848: {'lr': 0.0004921132036607186, 'samples': 1123008, 'steps': 5848, 'loss/train': 1.3769336342811584} 01/28/2022 20:21:55 - INFO - codeparrot_training - Step 5849: {'lr': 0.0004921091256657429, 'samples': 1123200, 'steps': 5849, 'loss/train': 2.1941322684288025} 01/28/2022 20:21:59 - INFO - codeparrot_training - Step 5850: {'lr': 0.0004921050466336487, 'samples': 1123392, 'steps': 5850, 'loss/train': 1.1364826261997223} 01/28/2022 20:22:04 - INFO - codeparrot_training - Step 5851: {'lr': 0.0004921009665644535, 'samples': 1123584, 'steps': 5851, 'loss/train': 1.075954109430313} 01/28/2022 20:22:09 - INFO - codeparrot_training - Step 5852: {'lr': 0.0004920968854581745, 'samples': 1123776, 'steps': 5852, 'loss/train': 2.1560270190238953} 01/28/2022 20:22:13 - INFO - codeparrot_training - Step 5853: {'lr': 0.0004920928033148292, 'samples': 1123968, 'steps': 5853, 'loss/train': 1.8925913572311401} 01/28/2022 20:22:18 - INFO - codeparrot_training - Step 5854: {'lr': 0.0004920887201344353, 'samples': 1124160, 'steps': 5854, 'loss/train': 1.6841252446174622} 01/28/2022 20:22:22 - INFO - codeparrot_training - Step 5855: {'lr': 0.0004920846359170103, 'samples': 1124352, 'steps': 5855, 'loss/train': 2.1927546858787537} 01/28/2022 20:22:26 - INFO - codeparrot_training - Step 5856: {'lr': 0.0004920805506625714, 'samples': 1124544, 'steps': 5856, 'loss/train': 1.3019864559173584} 01/28/2022 20:22:31 - INFO - codeparrot_training - Step 5857: {'lr': 0.0004920764643711364, 'samples': 1124736, 'steps': 5857, 'loss/train': 1.4951403737068176} 01/28/2022 20:22:36 - INFO - codeparrot_training - Step 5858: {'lr': 0.0004920723770427226, 'samples': 1124928, 'steps': 5858, 'loss/train': 1.9024288058280945} 01/28/2022 20:22:40 - INFO - codeparrot_training - Step 5859: {'lr': 0.0004920682886773478, 'samples': 1125120, 'steps': 5859, 'loss/train': 1.8200923204421997} 01/28/2022 20:22:44 - INFO - codeparrot_training - Step 5860: {'lr': 0.000492064199275029, 'samples': 1125312, 'steps': 5860, 'loss/train': 2.3658544421195984} 01/28/2022 20:22:48 - INFO - codeparrot_training - Step 5861: {'lr': 0.0004920601088357844, 'samples': 1125504, 'steps': 5861, 'loss/train': 1.5898659825325012} 01/28/2022 20:22:54 - INFO - codeparrot_training - Step 5862: {'lr': 0.0004920560173596309, 'samples': 1125696, 'steps': 5862, 'loss/train': 2.5098851323127747} 01/28/2022 20:22:59 - INFO - codeparrot_training - Step 5863: {'lr': 0.0004920519248465864, 'samples': 1125888, 'steps': 5863, 'loss/train': 1.625409185886383} 01/28/2022 20:23:03 - INFO - codeparrot_training - Step 5864: {'lr': 0.0004920478312966683, 'samples': 1126080, 'steps': 5864, 'loss/train': 1.8453744649887085} 01/28/2022 20:23:07 - INFO - codeparrot_training - Step 5865: {'lr': 0.0004920437367098941, 'samples': 1126272, 'steps': 5865, 'loss/train': 1.362774521112442} 01/28/2022 20:23:11 - INFO - codeparrot_training - Step 5866: {'lr': 0.0004920396410862815, 'samples': 1126464, 'steps': 5866, 'loss/train': 1.061270534992218} 01/28/2022 20:23:17 - INFO - codeparrot_training - Step 5867: {'lr': 0.0004920355444258479, 'samples': 1126656, 'steps': 5867, 'loss/train': 2.1989232301712036} 01/28/2022 20:23:21 - INFO - codeparrot_training - Step 5868: {'lr': 0.0004920314467286108, 'samples': 1126848, 'steps': 5868, 'loss/train': 1.6289459466934204} 01/28/2022 20:23:25 - INFO - codeparrot_training - Step 5869: {'lr': 0.0004920273479945878, 'samples': 1127040, 'steps': 5869, 'loss/train': 2.0368160605430603} 01/28/2022 20:23:29 - INFO - codeparrot_training - Step 5870: {'lr': 0.0004920232482237966, 'samples': 1127232, 'steps': 5870, 'loss/train': 2.640232264995575} 01/28/2022 20:23:34 - INFO - codeparrot_training - Step 5871: {'lr': 0.0004920191474162547, 'samples': 1127424, 'steps': 5871, 'loss/train': 2.1136448979377747} 01/28/2022 20:23:40 - INFO - codeparrot_training - Step 5872: {'lr': 0.0004920150455719795, 'samples': 1127616, 'steps': 5872, 'loss/train': 1.813682734966278} 01/28/2022 20:23:44 - INFO - codeparrot_training - Step 5873: {'lr': 0.0004920109426909887, 'samples': 1127808, 'steps': 5873, 'loss/train': 1.3451357781887054} 01/28/2022 20:23:48 - INFO - codeparrot_training - Step 5874: {'lr': 0.0004920068387733, 'samples': 1128000, 'steps': 5874, 'loss/train': 2.006288766860962} 01/28/2022 20:23:53 - INFO - codeparrot_training - Step 5875: {'lr': 0.0004920027338189307, 'samples': 1128192, 'steps': 5875, 'loss/train': 1.3159213364124298} 01/28/2022 20:23:57 - INFO - codeparrot_training - Step 5876: {'lr': 0.0004919986278278986, 'samples': 1128384, 'steps': 5876, 'loss/train': 1.6988874077796936} 01/28/2022 20:24:02 - INFO - codeparrot_training - Step 5877: {'lr': 0.0004919945208002212, 'samples': 1128576, 'steps': 5877, 'loss/train': 7.374750137329102} 01/28/2022 20:24:07 - INFO - codeparrot_training - Step 5878: {'lr': 0.0004919904127359162, 'samples': 1128768, 'steps': 5878, 'loss/train': 0.3620244786143303} 01/28/2022 20:24:11 - INFO - codeparrot_training - Step 5879: {'lr': 0.000491986303635001, 'samples': 1128960, 'steps': 5879, 'loss/train': 2.7401375770568848} 01/28/2022 20:24:15 - INFO - codeparrot_training - Step 5880: {'lr': 0.0004919821934974933, 'samples': 1129152, 'steps': 5880, 'loss/train': 2.0164239406585693} 01/28/2022 20:24:19 - INFO - codeparrot_training - Step 5881: {'lr': 0.0004919780823234108, 'samples': 1129344, 'steps': 5881, 'loss/train': 1.6630252003669739} 01/28/2022 20:24:24 - INFO - codeparrot_training - Step 5882: {'lr': 0.000491973970112771, 'samples': 1129536, 'steps': 5882, 'loss/train': 2.3150594830513} 01/28/2022 20:24:29 - INFO - codeparrot_training - Step 5883: {'lr': 0.0004919698568655916, 'samples': 1129728, 'steps': 5883, 'loss/train': 1.3901929557323456} 01/28/2022 20:24:33 - INFO - codeparrot_training - Step 5884: {'lr': 0.0004919657425818901, 'samples': 1129920, 'steps': 5884, 'loss/train': 2.246487557888031} 01/28/2022 20:24:38 - INFO - codeparrot_training - Step 5885: {'lr': 0.0004919616272616842, 'samples': 1130112, 'steps': 5885, 'loss/train': 2.2143362760543823} 01/28/2022 20:24:42 - INFO - codeparrot_training - Step 5886: {'lr': 0.0004919575109049915, 'samples': 1130304, 'steps': 5886, 'loss/train': 2.5048890709877014} 01/28/2022 20:24:48 - INFO - codeparrot_training - Step 5887: {'lr': 0.0004919533935118296, 'samples': 1130496, 'steps': 5887, 'loss/train': 1.266768604516983} 01/28/2022 20:24:52 - INFO - codeparrot_training - Step 5888: {'lr': 0.0004919492750822163, 'samples': 1130688, 'steps': 5888, 'loss/train': 2.2240952253341675} 01/28/2022 20:24:57 - INFO - codeparrot_training - Step 5889: {'lr': 0.0004919451556161692, 'samples': 1130880, 'steps': 5889, 'loss/train': 0.2399195283651352} 01/28/2022 20:25:01 - INFO - codeparrot_training - Step 5890: {'lr': 0.0004919410351137058, 'samples': 1131072, 'steps': 5890, 'loss/train': 1.8970540165901184} 01/28/2022 20:25:05 - INFO - codeparrot_training - Step 5891: {'lr': 0.0004919369135748438, 'samples': 1131264, 'steps': 5891, 'loss/train': 2.0140621662139893} 01/28/2022 20:25:11 - INFO - codeparrot_training - Step 5892: {'lr': 0.0004919327909996008, 'samples': 1131456, 'steps': 5892, 'loss/train': 2.215809166431427} 01/28/2022 20:25:15 - INFO - codeparrot_training - Step 5893: {'lr': 0.0004919286673879948, 'samples': 1131648, 'steps': 5893, 'loss/train': 2.1678959727287292} 01/28/2022 20:25:20 - INFO - codeparrot_training - Step 5894: {'lr': 0.000491924542740043, 'samples': 1131840, 'steps': 5894, 'loss/train': 1.75865238904953} 01/28/2022 20:25:24 - INFO - codeparrot_training - Step 5895: {'lr': 0.0004919204170557634, 'samples': 1132032, 'steps': 5895, 'loss/train': 1.1624529361724854} 01/28/2022 20:25:28 - INFO - codeparrot_training - Step 5896: {'lr': 0.0004919162903351734, 'samples': 1132224, 'steps': 5896, 'loss/train': 1.5457729697227478} 01/28/2022 20:25:33 - INFO - codeparrot_training - Step 5897: {'lr': 0.000491912162578291, 'samples': 1132416, 'steps': 5897, 'loss/train': 1.2920285761356354} 01/28/2022 20:25:37 - INFO - codeparrot_training - Step 5898: {'lr': 0.0004919080337851336, 'samples': 1132608, 'steps': 5898, 'loss/train': 2.467585802078247} 01/28/2022 20:25:42 - INFO - codeparrot_training - Step 5899: {'lr': 0.000491903903955719, 'samples': 1132800, 'steps': 5899, 'loss/train': 1.8126845955848694} 01/28/2022 20:25:46 - INFO - codeparrot_training - Step 5900: {'lr': 0.0004918997730900649, 'samples': 1132992, 'steps': 5900, 'loss/train': 1.8292969465255737} 01/28/2022 20:25:50 - INFO - codeparrot_training - Step 5901: {'lr': 0.000491895641188189, 'samples': 1133184, 'steps': 5901, 'loss/train': 0.5142768323421478} 01/28/2022 20:25:56 - INFO - codeparrot_training - Step 5902: {'lr': 0.000491891508250109, 'samples': 1133376, 'steps': 5902, 'loss/train': 1.8876212239265442} 01/28/2022 20:26:00 - INFO - codeparrot_training - Step 5903: {'lr': 0.0004918873742758426, 'samples': 1133568, 'steps': 5903, 'loss/train': 1.7543964385986328} 01/28/2022 20:26:04 - INFO - codeparrot_training - Step 5904: {'lr': 0.0004918832392654074, 'samples': 1133760, 'steps': 5904, 'loss/train': 1.03409805893898} 01/28/2022 20:26:08 - INFO - codeparrot_training - Step 5905: {'lr': 0.0004918791032188214, 'samples': 1133952, 'steps': 5905, 'loss/train': 1.9729245901107788} 01/28/2022 20:26:15 - INFO - codeparrot_training - Step 5906: {'lr': 0.0004918749661361019, 'samples': 1134144, 'steps': 5906, 'loss/train': 1.0244839489459991} 01/28/2022 20:26:19 - INFO - codeparrot_training - Step 5907: {'lr': 0.000491870828017267, 'samples': 1134336, 'steps': 5907, 'loss/train': 1.7850723266601562} 01/28/2022 20:26:23 - INFO - codeparrot_training - Step 5908: {'lr': 0.0004918666888623342, 'samples': 1134528, 'steps': 5908, 'loss/train': 1.9451104402542114} 01/28/2022 20:26:27 - INFO - codeparrot_training - Step 5909: {'lr': 0.0004918625486713214, 'samples': 1134720, 'steps': 5909, 'loss/train': 2.558558464050293} 01/28/2022 20:26:32 - INFO - codeparrot_training - Step 5910: {'lr': 0.0004918584074442462, 'samples': 1134912, 'steps': 5910, 'loss/train': 2.3624478578567505} 01/28/2022 20:26:37 - INFO - codeparrot_training - Step 5911: {'lr': 0.0004918542651811263, 'samples': 1135104, 'steps': 5911, 'loss/train': 2.000655949115753} 01/28/2022 20:26:41 - INFO - codeparrot_training - Step 5912: {'lr': 0.0004918501218819796, 'samples': 1135296, 'steps': 5912, 'loss/train': 1.307930052280426} 01/28/2022 20:26:45 - INFO - codeparrot_training - Step 5913: {'lr': 0.0004918459775468238, 'samples': 1135488, 'steps': 5913, 'loss/train': 2.311002731323242} 01/28/2022 20:26:50 - INFO - codeparrot_training - Step 5914: {'lr': 0.0004918418321756766, 'samples': 1135680, 'steps': 5914, 'loss/train': 2.3805786967277527} 01/28/2022 20:26:54 - INFO - codeparrot_training - Step 5915: {'lr': 0.0004918376857685557, 'samples': 1135872, 'steps': 5915, 'loss/train': 1.3991844356060028} 01/28/2022 20:27:00 - INFO - codeparrot_training - Step 5916: {'lr': 0.000491833538325479, 'samples': 1136064, 'steps': 5916, 'loss/train': 1.9315102100372314} 01/28/2022 20:27:04 - INFO - codeparrot_training - Step 5917: {'lr': 0.0004918293898464643, 'samples': 1136256, 'steps': 5917, 'loss/train': 0.8476210534572601} 01/28/2022 20:27:09 - INFO - codeparrot_training - Step 5918: {'lr': 0.0004918252403315292, 'samples': 1136448, 'steps': 5918, 'loss/train': 1.3554220497608185} 01/28/2022 20:27:13 - INFO - codeparrot_training - Step 5919: {'lr': 0.0004918210897806916, 'samples': 1136640, 'steps': 5919, 'loss/train': 0.5131009072065353} 01/28/2022 20:27:17 - INFO - codeparrot_training - Step 5920: {'lr': 0.0004918169381939692, 'samples': 1136832, 'steps': 5920, 'loss/train': 2.2731680274009705} 01/28/2022 20:27:22 - INFO - codeparrot_training - Step 5921: {'lr': 0.0004918127855713799, 'samples': 1137024, 'steps': 5921, 'loss/train': 2.211876153945923} 01/28/2022 20:27:27 - INFO - codeparrot_training - Step 5922: {'lr': 0.0004918086319129413, 'samples': 1137216, 'steps': 5922, 'loss/train': 1.9275568127632141} 01/28/2022 20:27:31 - INFO - codeparrot_training - Step 5923: {'lr': 0.0004918044772186714, 'samples': 1137408, 'steps': 5923, 'loss/train': 1.5278434753417969} 01/28/2022 20:27:35 - INFO - codeparrot_training - Step 5924: {'lr': 0.0004918003214885877, 'samples': 1137600, 'steps': 5924, 'loss/train': 1.8585192561149597} 01/28/2022 20:27:39 - INFO - codeparrot_training - Step 5925: {'lr': 0.0004917961647227084, 'samples': 1137792, 'steps': 5925, 'loss/train': 1.6934655904769897} 01/28/2022 20:27:45 - INFO - codeparrot_training - Step 5926: {'lr': 0.0004917920069210511, 'samples': 1137984, 'steps': 5926, 'loss/train': 1.1456973552703857} 01/28/2022 20:27:49 - INFO - codeparrot_training - Step 5927: {'lr': 0.0004917878480836336, 'samples': 1138176, 'steps': 5927, 'loss/train': 2.169240117073059} 01/28/2022 20:27:53 - INFO - codeparrot_training - Step 5928: {'lr': 0.0004917836882104737, 'samples': 1138368, 'steps': 5928, 'loss/train': 1.0387997031211853} 01/28/2022 20:27:57 - INFO - codeparrot_training - Step 5929: {'lr': 0.0004917795273015892, 'samples': 1138560, 'steps': 5929, 'loss/train': 2.070933222770691} 01/28/2022 20:28:02 - INFO - codeparrot_training - Step 5930: {'lr': 0.0004917753653569981, 'samples': 1138752, 'steps': 5930, 'loss/train': 2.297995448112488} 01/28/2022 20:28:07 - INFO - codeparrot_training - Step 5931: {'lr': 0.000491771202376718, 'samples': 1138944, 'steps': 5931, 'loss/train': 1.8889476656913757} 01/28/2022 20:28:12 - INFO - codeparrot_training - Step 5932: {'lr': 0.000491767038360767, 'samples': 1139136, 'steps': 5932, 'loss/train': 1.3531013131141663} 01/28/2022 20:28:16 - INFO - codeparrot_training - Step 5933: {'lr': 0.0004917628733091626, 'samples': 1139328, 'steps': 5933, 'loss/train': 1.5585055947303772} 01/28/2022 20:28:20 - INFO - codeparrot_training - Step 5934: {'lr': 0.000491758707221923, 'samples': 1139520, 'steps': 5934, 'loss/train': 3.3240966796875} 01/28/2022 20:28:24 - INFO - codeparrot_training - Step 5935: {'lr': 0.0004917545400990657, 'samples': 1139712, 'steps': 5935, 'loss/train': 1.3687883913516998} 01/28/2022 20:28:30 - INFO - codeparrot_training - Step 5936: {'lr': 0.0004917503719406087, 'samples': 1139904, 'steps': 5936, 'loss/train': 2.487292170524597} 01/28/2022 20:28:34 - INFO - codeparrot_training - Step 5937: {'lr': 0.00049174620274657, 'samples': 1140096, 'steps': 5937, 'loss/train': 1.49821937084198} 01/28/2022 20:28:38 - INFO - codeparrot_training - Step 5938: {'lr': 0.0004917420325169673, 'samples': 1140288, 'steps': 5938, 'loss/train': 1.9358242750167847} 01/28/2022 20:28:43 - INFO - codeparrot_training - Step 5939: {'lr': 0.0004917378612518185, 'samples': 1140480, 'steps': 5939, 'loss/train': 2.2086848616600037} 01/28/2022 20:28:47 - INFO - codeparrot_training - Step 5940: {'lr': 0.0004917336889511414, 'samples': 1140672, 'steps': 5940, 'loss/train': 1.8746890425682068} 01/28/2022 20:28:52 - INFO - codeparrot_training - Step 5941: {'lr': 0.0004917295156149539, 'samples': 1140864, 'steps': 5941, 'loss/train': 0.7260902971029282} 01/28/2022 20:28:56 - INFO - codeparrot_training - Step 5942: {'lr': 0.000491725341243274, 'samples': 1141056, 'steps': 5942, 'loss/train': 1.5390965938568115} 01/28/2022 20:29:01 - INFO - codeparrot_training - Step 5943: {'lr': 0.0004917211658361196, 'samples': 1141248, 'steps': 5943, 'loss/train': 2.2438266277313232} 01/28/2022 20:29:05 - INFO - codeparrot_training - Step 5944: {'lr': 0.0004917169893935083, 'samples': 1141440, 'steps': 5944, 'loss/train': 1.7518054246902466} 01/28/2022 20:29:09 - INFO - codeparrot_training - Step 5945: {'lr': 0.0004917128119154582, 'samples': 1141632, 'steps': 5945, 'loss/train': 1.8325148820877075} 01/28/2022 20:29:15 - INFO - codeparrot_training - Step 5946: {'lr': 0.0004917086334019872, 'samples': 1141824, 'steps': 5946, 'loss/train': 0.7979634404182434} 01/28/2022 20:29:20 - INFO - codeparrot_training - Step 5947: {'lr': 0.0004917044538531131, 'samples': 1142016, 'steps': 5947, 'loss/train': 0.9056333899497986} 01/28/2022 20:29:24 - INFO - codeparrot_training - Step 5948: {'lr': 0.000491700273268854, 'samples': 1142208, 'steps': 5948, 'loss/train': 2.3625399470329285} 01/28/2022 20:29:28 - INFO - codeparrot_training - Step 5949: {'lr': 0.0004916960916492276, 'samples': 1142400, 'steps': 5949, 'loss/train': 2.07241290807724} 01/28/2022 20:29:32 - INFO - codeparrot_training - Step 5950: {'lr': 0.0004916919089942519, 'samples': 1142592, 'steps': 5950, 'loss/train': 2.3605931997299194} 01/28/2022 20:29:38 - INFO - codeparrot_training - Step 5951: {'lr': 0.0004916877253039448, 'samples': 1142784, 'steps': 5951, 'loss/train': 2.540294587612152} 01/28/2022 20:29:42 - INFO - codeparrot_training - Step 5952: {'lr': 0.0004916835405783242, 'samples': 1142976, 'steps': 5952, 'loss/train': 1.9512380361557007} 01/28/2022 20:29:46 - INFO - codeparrot_training - Step 5953: {'lr': 0.0004916793548174081, 'samples': 1143168, 'steps': 5953, 'loss/train': 1.7478695511817932} 01/28/2022 20:29:51 - INFO - codeparrot_training - Step 5954: {'lr': 0.0004916751680212145, 'samples': 1143360, 'steps': 5954, 'loss/train': 1.5331546068191528} 01/28/2022 20:29:55 - INFO - codeparrot_training - Step 5955: {'lr': 0.000491670980189761, 'samples': 1143552, 'steps': 5955, 'loss/train': 1.7844293117523193} 01/28/2022 20:30:01 - INFO - codeparrot_training - Step 5956: {'lr': 0.0004916667913230659, 'samples': 1143744, 'steps': 5956, 'loss/train': 1.2715246081352234} 01/28/2022 20:30:05 - INFO - codeparrot_training - Step 5957: {'lr': 0.000491662601421147, 'samples': 1143936, 'steps': 5957, 'loss/train': 0.7663291096687317} 01/28/2022 20:30:10 - INFO - codeparrot_training - Step 5958: {'lr': 0.0004916584104840222, 'samples': 1144128, 'steps': 5958, 'loss/train': 1.632607340812683} 01/28/2022 20:30:14 - INFO - codeparrot_training - Step 5959: {'lr': 0.0004916542185117095, 'samples': 1144320, 'steps': 5959, 'loss/train': 2.226007103919983} 01/28/2022 20:30:19 - INFO - codeparrot_training - Step 5960: {'lr': 0.0004916500255042268, 'samples': 1144512, 'steps': 5960, 'loss/train': 1.5690017938613892} 01/28/2022 20:30:24 - INFO - codeparrot_training - Step 5961: {'lr': 0.0004916458314615923, 'samples': 1144704, 'steps': 5961, 'loss/train': 1.2932104468345642} 01/28/2022 20:30:28 - INFO - codeparrot_training - Step 5962: {'lr': 0.0004916416363838237, 'samples': 1144896, 'steps': 5962, 'loss/train': 1.72049081325531} 01/28/2022 20:30:32 - INFO - codeparrot_training - Step 5963: {'lr': 0.000491637440270939, 'samples': 1145088, 'steps': 5963, 'loss/train': 2.835364580154419} 01/28/2022 20:30:36 - INFO - codeparrot_training - Step 5964: {'lr': 0.0004916332431229562, 'samples': 1145280, 'steps': 5964, 'loss/train': 0.8161697387695312} 01/28/2022 20:30:41 - INFO - codeparrot_training - Step 5965: {'lr': 0.0004916290449398934, 'samples': 1145472, 'steps': 5965, 'loss/train': 0.771015465259552} 01/28/2022 20:30:46 - INFO - codeparrot_training - Step 5966: {'lr': 0.0004916248457217686, 'samples': 1145664, 'steps': 5966, 'loss/train': 2.4733200073242188} 01/28/2022 20:30:50 - INFO - codeparrot_training - Step 5967: {'lr': 0.0004916206454685995, 'samples': 1145856, 'steps': 5967, 'loss/train': 1.4784178733825684} 01/28/2022 20:30:54 - INFO - codeparrot_training - Step 5968: {'lr': 0.0004916164441804044, 'samples': 1146048, 'steps': 5968, 'loss/train': 1.6776198148727417} 01/28/2022 20:30:59 - INFO - codeparrot_training - Step 5969: {'lr': 0.0004916122418572011, 'samples': 1146240, 'steps': 5969, 'loss/train': 1.4001700580120087} 01/28/2022 20:31:03 - INFO - codeparrot_training - Step 5970: {'lr': 0.0004916080384990077, 'samples': 1146432, 'steps': 5970, 'loss/train': 0.23980654031038284} 01/28/2022 20:31:09 - INFO - codeparrot_training - Step 5971: {'lr': 0.0004916038341058423, 'samples': 1146624, 'steps': 5971, 'loss/train': 2.0760018825531006} 01/28/2022 20:31:13 - INFO - codeparrot_training - Step 5972: {'lr': 0.0004915996286777226, 'samples': 1146816, 'steps': 5972, 'loss/train': 1.2239448130130768} 01/28/2022 20:31:18 - INFO - codeparrot_training - Step 5973: {'lr': 0.0004915954222146669, 'samples': 1147008, 'steps': 5973, 'loss/train': 1.0069970190525055} 01/28/2022 20:31:22 - INFO - codeparrot_training - Step 5974: {'lr': 0.0004915912147166932, 'samples': 1147200, 'steps': 5974, 'loss/train': 1.617512583732605} 01/28/2022 20:31:26 - INFO - codeparrot_training - Step 5975: {'lr': 0.0004915870061838193, 'samples': 1147392, 'steps': 5975, 'loss/train': 1.9822234511375427} 01/28/2022 20:31:32 - INFO - codeparrot_training - Step 5976: {'lr': 0.0004915827966160634, 'samples': 1147584, 'steps': 5976, 'loss/train': 1.950512945652008} 01/28/2022 20:31:36 - INFO - codeparrot_training - Step 5977: {'lr': 0.0004915785860134436, 'samples': 1147776, 'steps': 5977, 'loss/train': 0.2587753012776375} 01/28/2022 20:31:40 - INFO - codeparrot_training - Step 5978: {'lr': 0.0004915743743759779, 'samples': 1147968, 'steps': 5978, 'loss/train': 2.0273292660713196} 01/28/2022 20:31:45 - INFO - codeparrot_training - Step 5979: {'lr': 0.0004915701617036842, 'samples': 1148160, 'steps': 5979, 'loss/train': 2.0867777466773987} 01/28/2022 20:31:51 - INFO - codeparrot_training - Step 5980: {'lr': 0.0004915659479965806, 'samples': 1148352, 'steps': 5980, 'loss/train': 2.377518117427826} 01/28/2022 20:31:55 - INFO - codeparrot_training - Step 5981: {'lr': 0.0004915617332546852, 'samples': 1148544, 'steps': 5981, 'loss/train': 2.113312840461731} 01/28/2022 20:31:59 - INFO - codeparrot_training - Step 5982: {'lr': 0.0004915575174780161, 'samples': 1148736, 'steps': 5982, 'loss/train': 2.2778831720352173} 01/28/2022 20:32:04 - INFO - codeparrot_training - Step 5983: {'lr': 0.0004915533006665912, 'samples': 1148928, 'steps': 5983, 'loss/train': 2.0714576840400696} 01/28/2022 20:32:08 - INFO - codeparrot_training - Step 5984: {'lr': 0.0004915490828204287, 'samples': 1149120, 'steps': 5984, 'loss/train': 1.7459054589271545} 01/28/2022 20:32:13 - INFO - codeparrot_training - Step 5985: {'lr': 0.0004915448639395466, 'samples': 1149312, 'steps': 5985, 'loss/train': 1.8360026478767395} 01/28/2022 20:32:17 - INFO - codeparrot_training - Step 5986: {'lr': 0.0004915406440239631, 'samples': 1149504, 'steps': 5986, 'loss/train': 1.2412113547325134} 01/28/2022 20:32:22 - INFO - codeparrot_training - Step 5987: {'lr': 0.0004915364230736961, 'samples': 1149696, 'steps': 5987, 'loss/train': 1.9646004438400269} 01/28/2022 20:32:26 - INFO - codeparrot_training - Step 5988: {'lr': 0.0004915322010887637, 'samples': 1149888, 'steps': 5988, 'loss/train': 2.050501048564911} 01/28/2022 20:32:30 - INFO - codeparrot_training - Step 5989: {'lr': 0.0004915279780691843, 'samples': 1150080, 'steps': 5989, 'loss/train': 1.756539523601532} 01/28/2022 20:32:36 - INFO - codeparrot_training - Step 5990: {'lr': 0.0004915237540149755, 'samples': 1150272, 'steps': 5990, 'loss/train': 2.125460207462311} 01/28/2022 20:32:40 - INFO - codeparrot_training - Step 5991: {'lr': 0.0004915195289261557, 'samples': 1150464, 'steps': 5991, 'loss/train': 0.8524556457996368} 01/28/2022 20:32:44 - INFO - codeparrot_training - Step 5992: {'lr': 0.0004915153028027429, 'samples': 1150656, 'steps': 5992, 'loss/train': 1.9587838053703308} 01/28/2022 20:32:48 - INFO - codeparrot_training - Step 5993: {'lr': 0.0004915110756447552, 'samples': 1150848, 'steps': 5993, 'loss/train': 2.427842438220978} 01/28/2022 20:32:53 - INFO - codeparrot_training - Step 5994: {'lr': 0.0004915068474522109, 'samples': 1151040, 'steps': 5994, 'loss/train': 1.8376988768577576} 01/28/2022 20:32:58 - INFO - codeparrot_training - Step 5995: {'lr': 0.0004915026182251278, 'samples': 1151232, 'steps': 5995, 'loss/train': 2.086914360523224} 01/28/2022 20:33:02 - INFO - codeparrot_training - Step 5996: {'lr': 0.0004914983879635242, 'samples': 1151424, 'steps': 5996, 'loss/train': 1.0613655745983124} 01/28/2022 20:33:06 - INFO - codeparrot_training - Step 5997: {'lr': 0.0004914941566674183, 'samples': 1151616, 'steps': 5997, 'loss/train': 1.8911764025688171} 01/28/2022 20:33:11 - INFO - codeparrot_training - Step 5998: {'lr': 0.0004914899243368279, 'samples': 1151808, 'steps': 5998, 'loss/train': 0.9964997470378876} 01/28/2022 20:33:15 - INFO - codeparrot_training - Step 5999: {'lr': 0.0004914856909717715, 'samples': 1152000, 'steps': 5999, 'loss/train': 2.0546990633010864} 01/28/2022 20:33:15 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/28/2022 20:33:48 - WARNING - huggingface_hub.repository - Several commits (3) will be pushed upstream. 01/28/2022 20:33:48 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/28/2022 20:34:54 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 4e36df4..ece0f24 expert-dust-2 -> expert-dust-2 01/28/2022 20:35:01 - INFO - codeparrot_training - Step 6000: {'lr': 0.0004914814565722671, 'samples': 1152192, 'steps': 6000, 'loss/train': 1.1461258828639984} 01/28/2022 20:35:05 - INFO - codeparrot_training - Step 6001: {'lr': 0.0004914772211383327, 'samples': 1152384, 'steps': 6001, 'loss/train': 1.793827772140503} 01/28/2022 20:35:09 - INFO - codeparrot_training - Step 6002: {'lr': 0.0004914729846699867, 'samples': 1152576, 'steps': 6002, 'loss/train': 1.6545943021774292} 01/28/2022 20:35:14 - INFO - codeparrot_training - Step 6003: {'lr': 0.000491468747167247, 'samples': 1152768, 'steps': 6003, 'loss/train': 1.2947983145713806} 01/28/2022 20:35:18 - INFO - codeparrot_training - Step 6004: {'lr': 0.0004914645086301319, 'samples': 1152960, 'steps': 6004, 'loss/train': 2.372019052505493} 01/28/2022 20:35:23 - INFO - codeparrot_training - Step 6005: {'lr': 0.0004914602690586596, 'samples': 1153152, 'steps': 6005, 'loss/train': 1.0780168175697327} 01/28/2022 20:35:27 - INFO - codeparrot_training - Step 6006: {'lr': 0.0004914560284528481, 'samples': 1153344, 'steps': 6006, 'loss/train': 2.0050907135009766} 01/28/2022 20:35:32 - INFO - codeparrot_training - Step 6007: {'lr': 0.0004914517868127156, 'samples': 1153536, 'steps': 6007, 'loss/train': 1.5638819932937622} 01/28/2022 20:35:36 - INFO - codeparrot_training - Step 6008: {'lr': 0.0004914475441382804, 'samples': 1153728, 'steps': 6008, 'loss/train': 1.9928131699562073} 01/28/2022 20:35:40 - INFO - codeparrot_training - Step 6009: {'lr': 0.0004914433004295605, 'samples': 1153920, 'steps': 6009, 'loss/train': 1.7628443241119385} 01/28/2022 20:35:45 - INFO - codeparrot_training - Step 6010: {'lr': 0.0004914390556865743, 'samples': 1154112, 'steps': 6010, 'loss/train': 1.5074468851089478} 01/28/2022 20:35:50 - INFO - codeparrot_training - Step 6011: {'lr': 0.0004914348099093398, 'samples': 1154304, 'steps': 6011, 'loss/train': 1.9457672238349915} 01/28/2022 20:35:54 - INFO - codeparrot_training - Step 6012: {'lr': 0.0004914305630978751, 'samples': 1154496, 'steps': 6012, 'loss/train': 1.47553950548172} 01/28/2022 20:35:58 - INFO - codeparrot_training - Step 6013: {'lr': 0.0004914263152521987, 'samples': 1154688, 'steps': 6013, 'loss/train': 2.237191915512085} 01/28/2022 20:36:02 - INFO - codeparrot_training - Step 6014: {'lr': 0.0004914220663723286, 'samples': 1154880, 'steps': 6014, 'loss/train': 1.7350614666938782} 01/28/2022 20:36:09 - INFO - codeparrot_training - Step 6015: {'lr': 0.0004914178164582829, 'samples': 1155072, 'steps': 6015, 'loss/train': 2.060547351837158} 01/28/2022 20:36:13 - INFO - codeparrot_training - Step 6016: {'lr': 0.0004914135655100801, 'samples': 1155264, 'steps': 6016, 'loss/train': 1.4957183003425598} 01/28/2022 20:36:18 - INFO - codeparrot_training - Step 6017: {'lr': 0.0004914093135277381, 'samples': 1155456, 'steps': 6017, 'loss/train': 1.2213007807731628} 01/28/2022 20:36:22 - INFO - codeparrot_training - Step 6018: {'lr': 0.0004914050605112753, 'samples': 1155648, 'steps': 6018, 'loss/train': 1.7390998005867004} 01/28/2022 20:36:26 - INFO - codeparrot_training - Step 6019: {'lr': 0.00049140080646071, 'samples': 1155840, 'steps': 6019, 'loss/train': 1.0334317088127136} 01/28/2022 20:36:31 - INFO - codeparrot_training - Step 6020: {'lr': 0.0004913965513760601, 'samples': 1156032, 'steps': 6020, 'loss/train': 2.216709852218628} 01/28/2022 20:36:36 - INFO - codeparrot_training - Step 6021: {'lr': 0.0004913922952573442, 'samples': 1156224, 'steps': 6021, 'loss/train': 1.8101980090141296} 01/28/2022 20:36:40 - INFO - codeparrot_training - Step 6022: {'lr': 0.0004913880381045803, 'samples': 1156416, 'steps': 6022, 'loss/train': 1.2511427700519562} 01/28/2022 20:36:44 - INFO - codeparrot_training - Step 6023: {'lr': 0.0004913837799177867, 'samples': 1156608, 'steps': 6023, 'loss/train': 4.813077449798584} 01/28/2022 20:36:48 - INFO - codeparrot_training - Step 6024: {'lr': 0.0004913795206969815, 'samples': 1156800, 'steps': 6024, 'loss/train': 1.9106170535087585} 01/28/2022 20:36:54 - INFO - codeparrot_training - Step 6025: {'lr': 0.0004913752604421833, 'samples': 1156992, 'steps': 6025, 'loss/train': 1.7181251049041748} 01/28/2022 20:36:58 - INFO - codeparrot_training - Step 6026: {'lr': 0.0004913709991534099, 'samples': 1157184, 'steps': 6026, 'loss/train': 3.1423813104629517} 01/28/2022 20:37:02 - INFO - codeparrot_training - Step 6027: {'lr': 0.00049136673683068, 'samples': 1157376, 'steps': 6027, 'loss/train': 1.0476156771183014} 01/28/2022 20:37:07 - INFO - codeparrot_training - Step 6028: {'lr': 0.0004913624734740115, 'samples': 1157568, 'steps': 6028, 'loss/train': 1.604585587978363} 01/28/2022 20:37:13 - INFO - codeparrot_training - Step 6029: {'lr': 0.0004913582090834229, 'samples': 1157760, 'steps': 6029, 'loss/train': 1.8966549038887024} 01/28/2022 20:37:17 - INFO - codeparrot_training - Step 6030: {'lr': 0.0004913539436589323, 'samples': 1157952, 'steps': 6030, 'loss/train': 0.9039782881736755} 01/28/2022 20:37:21 - INFO - codeparrot_training - Step 6031: {'lr': 0.0004913496772005581, 'samples': 1158144, 'steps': 6031, 'loss/train': 1.8801158666610718} 01/28/2022 20:37:26 - INFO - codeparrot_training - Step 6032: {'lr': 0.0004913454097083185, 'samples': 1158336, 'steps': 6032, 'loss/train': 2.367067873477936} 01/28/2022 20:37:30 - INFO - codeparrot_training - Step 6033: {'lr': 0.0004913411411822318, 'samples': 1158528, 'steps': 6033, 'loss/train': 1.0416067242622375} 01/28/2022 20:37:35 - INFO - codeparrot_training - Step 6034: {'lr': 0.0004913368716223162, 'samples': 1158720, 'steps': 6034, 'loss/train': 1.7832968831062317} 01/28/2022 20:37:39 - INFO - codeparrot_training - Step 6035: {'lr': 0.0004913326010285902, 'samples': 1158912, 'steps': 6035, 'loss/train': 0.7944119274616241} 01/28/2022 20:37:44 - INFO - codeparrot_training - Step 6036: {'lr': 0.0004913283294010719, 'samples': 1159104, 'steps': 6036, 'loss/train': 2.023652493953705} 01/28/2022 20:37:48 - INFO - codeparrot_training - Step 6037: {'lr': 0.0004913240567397797, 'samples': 1159296, 'steps': 6037, 'loss/train': 1.475430428981781} 01/28/2022 20:37:52 - INFO - codeparrot_training - Step 6038: {'lr': 0.0004913197830447319, 'samples': 1159488, 'steps': 6038, 'loss/train': 0.8599489331245422} 01/28/2022 20:37:59 - INFO - codeparrot_training - Step 6039: {'lr': 0.0004913155083159467, 'samples': 1159680, 'steps': 6039, 'loss/train': 0.7831673920154572} 01/28/2022 20:38:03 - INFO - codeparrot_training - Step 6040: {'lr': 0.0004913112325534426, 'samples': 1159872, 'steps': 6040, 'loss/train': 6.935019493103027} 01/28/2022 20:38:07 - INFO - codeparrot_training - Step 6041: {'lr': 0.0004913069557572376, 'samples': 1160064, 'steps': 6041, 'loss/train': 7.645186185836792} 01/28/2022 20:38:11 - INFO - codeparrot_training - Step 6042: {'lr': 0.0004913026779273504, 'samples': 1160256, 'steps': 6042, 'loss/train': 2.0677176117897034} 01/28/2022 20:38:16 - INFO - codeparrot_training - Step 6043: {'lr': 0.0004912983990637992, 'samples': 1160448, 'steps': 6043, 'loss/train': 1.1484251618385315} 01/28/2022 20:38:21 - INFO - codeparrot_training - Step 6044: {'lr': 0.0004912941191666021, 'samples': 1160640, 'steps': 6044, 'loss/train': 2.0978912115097046} 01/28/2022 20:38:25 - INFO - codeparrot_training - Step 6045: {'lr': 0.0004912898382357777, 'samples': 1160832, 'steps': 6045, 'loss/train': 1.7513524889945984} 01/28/2022 20:38:29 - INFO - codeparrot_training - Step 6046: {'lr': 0.0004912855562713443, 'samples': 1161024, 'steps': 6046, 'loss/train': 1.758177101612091} 01/28/2022 20:38:34 - INFO - codeparrot_training - Step 6047: {'lr': 0.0004912812732733201, 'samples': 1161216, 'steps': 6047, 'loss/train': 1.9367802143096924} 01/28/2022 20:38:38 - INFO - codeparrot_training - Step 6048: {'lr': 0.0004912769892417236, 'samples': 1161408, 'steps': 6048, 'loss/train': 2.01489919424057} 01/28/2022 20:38:43 - INFO - codeparrot_training - Step 6049: {'lr': 0.000491272704176573, 'samples': 1161600, 'steps': 6049, 'loss/train': 2.582838714122772} 01/28/2022 20:38:48 - INFO - codeparrot_training - Step 6050: {'lr': 0.0004912684180778869, 'samples': 1161792, 'steps': 6050, 'loss/train': 2.1242281794548035} 01/28/2022 20:38:52 - INFO - codeparrot_training - Step 6051: {'lr': 0.0004912641309456834, 'samples': 1161984, 'steps': 6051, 'loss/train': 1.31065034866333} 01/28/2022 20:38:56 - INFO - codeparrot_training - Step 6052: {'lr': 0.000491259842779981, 'samples': 1162176, 'steps': 6052, 'loss/train': 2.39087176322937} 01/28/2022 20:39:00 - INFO - codeparrot_training - Step 6053: {'lr': 0.0004912555535807981, 'samples': 1162368, 'steps': 6053, 'loss/train': 2.1323418617248535} 01/28/2022 20:39:06 - INFO - codeparrot_training - Step 6054: {'lr': 0.0004912512633481529, 'samples': 1162560, 'steps': 6054, 'loss/train': 1.8342220187187195} 01/28/2022 20:39:11 - INFO - codeparrot_training - Step 6055: {'lr': 0.0004912469720820639, 'samples': 1162752, 'steps': 6055, 'loss/train': 2.839962601661682} 01/28/2022 20:39:15 - INFO - codeparrot_training - Step 6056: {'lr': 0.0004912426797825495, 'samples': 1162944, 'steps': 6056, 'loss/train': 1.8129472732543945} 01/28/2022 20:39:19 - INFO - codeparrot_training - Step 6057: {'lr': 0.0004912383864496281, 'samples': 1163136, 'steps': 6057, 'loss/train': 1.7860980033874512} 01/28/2022 20:39:23 - INFO - codeparrot_training - Step 6058: {'lr': 0.0004912340920833182, 'samples': 1163328, 'steps': 6058, 'loss/train': 2.5604324340820312} 01/28/2022 20:39:29 - INFO - codeparrot_training - Step 6059: {'lr': 0.0004912297966836378, 'samples': 1163520, 'steps': 6059, 'loss/train': 2.141771972179413} 01/28/2022 20:39:33 - INFO - codeparrot_training - Step 6060: {'lr': 0.0004912255002506057, 'samples': 1163712, 'steps': 6060, 'loss/train': 1.6526294946670532} 01/28/2022 20:39:37 - INFO - codeparrot_training - Step 6061: {'lr': 0.00049122120278424, 'samples': 1163904, 'steps': 6061, 'loss/train': 2.198725461959839} 01/28/2022 20:39:42 - INFO - codeparrot_training - Step 6062: {'lr': 0.0004912169042845595, 'samples': 1164096, 'steps': 6062, 'loss/train': 1.4352972507476807} 01/28/2022 20:39:46 - INFO - codeparrot_training - Step 6063: {'lr': 0.0004912126047515821, 'samples': 1164288, 'steps': 6063, 'loss/train': 1.2546006739139557} 01/28/2022 20:39:51 - INFO - codeparrot_training - Step 6064: {'lr': 0.0004912083041853267, 'samples': 1164480, 'steps': 6064, 'loss/train': 2.434203028678894} 01/28/2022 20:39:55 - INFO - codeparrot_training - Step 6065: {'lr': 0.0004912040025858114, 'samples': 1164672, 'steps': 6065, 'loss/train': 1.7870694994926453} 01/28/2022 20:40:00 - INFO - codeparrot_training - Step 6066: {'lr': 0.0004911996999530548, 'samples': 1164864, 'steps': 6066, 'loss/train': 1.833459198474884} 01/28/2022 20:40:04 - INFO - codeparrot_training - Step 6067: {'lr': 0.0004911953962870754, 'samples': 1165056, 'steps': 6067, 'loss/train': 1.2691670358181} 01/28/2022 20:40:08 - INFO - codeparrot_training - Step 6068: {'lr': 0.0004911910915878913, 'samples': 1165248, 'steps': 6068, 'loss/train': 2.177170515060425} 01/28/2022 20:40:13 - INFO - codeparrot_training - Step 6069: {'lr': 0.0004911867858555212, 'samples': 1165440, 'steps': 6069, 'loss/train': 2.230476200580597} 01/28/2022 20:40:18 - INFO - codeparrot_training - Step 6070: {'lr': 0.0004911824790899836, 'samples': 1165632, 'steps': 6070, 'loss/train': 2.082961320877075} 01/28/2022 20:40:22 - INFO - codeparrot_training - Step 6071: {'lr': 0.0004911781712912968, 'samples': 1165824, 'steps': 6071, 'loss/train': 2.78812175989151} 01/28/2022 20:40:26 - INFO - codeparrot_training - Step 6072: {'lr': 0.0004911738624594793, 'samples': 1166016, 'steps': 6072, 'loss/train': 1.5950524806976318} 01/28/2022 20:40:30 - INFO - codeparrot_training - Step 6073: {'lr': 0.0004911695525945494, 'samples': 1166208, 'steps': 6073, 'loss/train': 1.489719808101654} 01/28/2022 20:40:35 - INFO - codeparrot_training - Step 6074: {'lr': 0.0004911652416965259, 'samples': 1166400, 'steps': 6074, 'loss/train': 2.7648534178733826} 01/28/2022 20:40:41 - INFO - codeparrot_training - Step 6075: {'lr': 0.000491160929765427, 'samples': 1166592, 'steps': 6075, 'loss/train': 2.9330363273620605} 01/28/2022 20:40:45 - INFO - codeparrot_training - Step 6076: {'lr': 0.0004911566168012714, 'samples': 1166784, 'steps': 6076, 'loss/train': 1.8041326403617859} 01/28/2022 20:40:49 - INFO - codeparrot_training - Step 6077: {'lr': 0.0004911523028040772, 'samples': 1166976, 'steps': 6077, 'loss/train': 2.427924871444702} 01/28/2022 20:40:54 - INFO - codeparrot_training - Step 6078: {'lr': 0.0004911479877738633, 'samples': 1167168, 'steps': 6078, 'loss/train': 1.713216483592987} 01/28/2022 20:40:58 - INFO - codeparrot_training - Step 6079: {'lr': 0.0004911436717106478, 'samples': 1167360, 'steps': 6079, 'loss/train': 1.1418323814868927} 01/28/2022 20:41:03 - INFO - codeparrot_training - Step 6080: {'lr': 0.0004911393546144495, 'samples': 1167552, 'steps': 6080, 'loss/train': 1.4632920920848846} 01/28/2022 20:41:08 - INFO - codeparrot_training - Step 6081: {'lr': 0.0004911350364852868, 'samples': 1167744, 'steps': 6081, 'loss/train': 1.43265500664711} 01/28/2022 20:41:12 - INFO - codeparrot_training - Step 6082: {'lr': 0.0004911307173231782, 'samples': 1167936, 'steps': 6082, 'loss/train': 0.9298369288444519} 01/28/2022 20:41:16 - INFO - codeparrot_training - Step 6083: {'lr': 0.000491126397128142, 'samples': 1168128, 'steps': 6083, 'loss/train': 1.7416797280311584} 01/28/2022 20:41:22 - INFO - codeparrot_training - Step 6084: {'lr': 0.0004911220759001971, 'samples': 1168320, 'steps': 6084, 'loss/train': 1.991996169090271} 01/28/2022 20:41:26 - INFO - codeparrot_training - Step 6085: {'lr': 0.0004911177536393616, 'samples': 1168512, 'steps': 6085, 'loss/train': 1.8118886947631836} 01/28/2022 20:41:31 - INFO - codeparrot_training - Step 6086: {'lr': 0.0004911134303456543, 'samples': 1168704, 'steps': 6086, 'loss/train': 1.9882146120071411} 01/28/2022 20:41:35 - INFO - codeparrot_training - Step 6087: {'lr': 0.0004911091060190937, 'samples': 1168896, 'steps': 6087, 'loss/train': 2.3673863410949707} 01/28/2022 20:41:39 - INFO - codeparrot_training - Step 6088: {'lr': 0.0004911047806596981, 'samples': 1169088, 'steps': 6088, 'loss/train': 1.0594293773174286} 01/28/2022 20:41:44 - INFO - codeparrot_training - Step 6089: {'lr': 0.0004911004542674863, 'samples': 1169280, 'steps': 6089, 'loss/train': 2.116032600402832} 01/28/2022 20:41:49 - INFO - codeparrot_training - Step 6090: {'lr': 0.0004910961268424766, 'samples': 1169472, 'steps': 6090, 'loss/train': 2.1293681859970093} 01/28/2022 20:41:53 - INFO - codeparrot_training - Step 6091: {'lr': 0.0004910917983846877, 'samples': 1169664, 'steps': 6091, 'loss/train': 1.2698120176792145} 01/28/2022 20:41:57 - INFO - codeparrot_training - Step 6092: {'lr': 0.0004910874688941381, 'samples': 1169856, 'steps': 6092, 'loss/train': 1.9277885556221008} 01/28/2022 20:42:01 - INFO - codeparrot_training - Step 6093: {'lr': 0.0004910831383708464, 'samples': 1170048, 'steps': 6093, 'loss/train': 2.1205363869667053} 01/28/2022 20:42:07 - INFO - codeparrot_training - Step 6094: {'lr': 0.000491078806814831, 'samples': 1170240, 'steps': 6094, 'loss/train': 2.344042718410492} 01/28/2022 20:42:11 - INFO - codeparrot_training - Step 6095: {'lr': 0.0004910744742261106, 'samples': 1170432, 'steps': 6095, 'loss/train': 1.3061645328998566} 01/28/2022 20:42:15 - INFO - codeparrot_training - Step 6096: {'lr': 0.0004910701406047037, 'samples': 1170624, 'steps': 6096, 'loss/train': 2.169794976711273} 01/28/2022 20:42:19 - INFO - codeparrot_training - Step 6097: {'lr': 0.0004910658059506289, 'samples': 1170816, 'steps': 6097, 'loss/train': 1.8281872272491455} 01/28/2022 20:42:24 - INFO - codeparrot_training - Step 6098: {'lr': 0.0004910614702639045, 'samples': 1171008, 'steps': 6098, 'loss/train': 1.8676049709320068} 01/28/2022 20:42:30 - INFO - codeparrot_training - Step 6099: {'lr': 0.0004910571335445496, 'samples': 1171200, 'steps': 6099, 'loss/train': 2.4590173959732056} 01/28/2022 20:42:34 - INFO - codeparrot_training - Step 6100: {'lr': 0.0004910527957925823, 'samples': 1171392, 'steps': 6100, 'loss/train': 2.0829055309295654} 01/28/2022 20:42:38 - INFO - codeparrot_training - Step 6101: {'lr': 0.0004910484570080215, 'samples': 1171584, 'steps': 6101, 'loss/train': 1.7787119150161743} 01/28/2022 20:42:42 - INFO - codeparrot_training - Step 6102: {'lr': 0.0004910441171908855, 'samples': 1171776, 'steps': 6102, 'loss/train': 1.119347870349884} 01/28/2022 20:42:47 - INFO - codeparrot_training - Step 6103: {'lr': 0.0004910397763411931, 'samples': 1171968, 'steps': 6103, 'loss/train': 1.3723326623439789} 01/28/2022 20:42:52 - INFO - codeparrot_training - Step 6104: {'lr': 0.000491035434458963, 'samples': 1172160, 'steps': 6104, 'loss/train': 2.188839018344879} 01/28/2022 20:42:56 - INFO - codeparrot_training - Step 6105: {'lr': 0.0004910310915442135, 'samples': 1172352, 'steps': 6105, 'loss/train': 1.1884676814079285} 01/28/2022 20:43:00 - INFO - codeparrot_training - Step 6106: {'lr': 0.0004910267475969633, 'samples': 1172544, 'steps': 6106, 'loss/train': 1.55589097738266} 01/28/2022 20:43:05 - INFO - codeparrot_training - Step 6107: {'lr': 0.000491022402617231, 'samples': 1172736, 'steps': 6107, 'loss/train': 3.071161150932312} 01/28/2022 20:43:09 - INFO - codeparrot_training - Step 6108: {'lr': 0.0004910180566050354, 'samples': 1172928, 'steps': 6108, 'loss/train': 1.9011729955673218} 01/28/2022 20:43:15 - INFO - codeparrot_training - Step 6109: {'lr': 0.0004910137095603949, 'samples': 1173120, 'steps': 6109, 'loss/train': 1.4731808602809906} 01/28/2022 20:43:19 - INFO - codeparrot_training - Step 6110: {'lr': 0.0004910093614833282, 'samples': 1173312, 'steps': 6110, 'loss/train': 1.4441472887992859} 01/28/2022 20:43:23 - INFO - codeparrot_training - Step 6111: {'lr': 0.000491005012373854, 'samples': 1173504, 'steps': 6111, 'loss/train': 2.0637359619140625} 01/28/2022 20:43:28 - INFO - codeparrot_training - Step 6112: {'lr': 0.0004910006622319908, 'samples': 1173696, 'steps': 6112, 'loss/train': 2.290369749069214} 01/28/2022 20:43:32 - INFO - codeparrot_training - Step 6113: {'lr': 0.0004909963110577573, 'samples': 1173888, 'steps': 6113, 'loss/train': 1.3259535133838654} 01/28/2022 20:43:37 - INFO - codeparrot_training - Step 6114: {'lr': 0.000490991958851172, 'samples': 1174080, 'steps': 6114, 'loss/train': 1.898880958557129} 01/28/2022 20:43:41 - INFO - codeparrot_training - Step 6115: {'lr': 0.0004909876056122538, 'samples': 1174272, 'steps': 6115, 'loss/train': 2.0320302844047546} 01/28/2022 20:43:46 - INFO - codeparrot_training - Step 6116: {'lr': 0.0004909832513410213, 'samples': 1174464, 'steps': 6116, 'loss/train': 1.4605750143527985} 01/28/2022 20:43:50 - INFO - codeparrot_training - Step 6117: {'lr': 0.000490978896037493, 'samples': 1174656, 'steps': 6117, 'loss/train': 1.4277098178863525} 01/28/2022 20:43:54 - INFO - codeparrot_training - Step 6118: {'lr': 0.0004909745397016876, 'samples': 1174848, 'steps': 6118, 'loss/train': 1.230183631181717} 01/28/2022 20:43:59 - INFO - codeparrot_training - Step 6119: {'lr': 0.0004909701823336238, 'samples': 1175040, 'steps': 6119, 'loss/train': 2.66631156206131} 01/28/2022 20:44:04 - INFO - codeparrot_training - Step 6120: {'lr': 0.0004909658239333202, 'samples': 1175232, 'steps': 6120, 'loss/train': 1.8361458778381348} 01/28/2022 20:44:08 - INFO - codeparrot_training - Step 6121: {'lr': 0.0004909614645007956, 'samples': 1175424, 'steps': 6121, 'loss/train': 1.7496984601020813} 01/28/2022 20:44:12 - INFO - codeparrot_training - Step 6122: {'lr': 0.0004909571040360686, 'samples': 1175616, 'steps': 6122, 'loss/train': 1.6784638166427612} 01/28/2022 20:44:16 - INFO - codeparrot_training - Step 6123: {'lr': 0.0004909527425391579, 'samples': 1175808, 'steps': 6123, 'loss/train': 1.0162133574485779} 01/28/2022 20:44:22 - INFO - codeparrot_training - Step 6124: {'lr': 0.0004909483800100822, 'samples': 1176000, 'steps': 6124, 'loss/train': 2.413476884365082} 01/28/2022 20:44:26 - INFO - codeparrot_training - Step 6125: {'lr': 0.00049094401644886, 'samples': 1176192, 'steps': 6125, 'loss/train': 1.3162972033023834} 01/28/2022 20:44:30 - INFO - codeparrot_training - Step 6126: {'lr': 0.0004909396518555102, 'samples': 1176384, 'steps': 6126, 'loss/train': 1.2119333446025848} 01/28/2022 20:44:34 - INFO - codeparrot_training - Step 6127: {'lr': 0.0004909352862300514, 'samples': 1176576, 'steps': 6127, 'loss/train': 1.148722618818283} 01/28/2022 20:44:38 - INFO - codeparrot_training - Step 6128: {'lr': 0.0004909309195725024, 'samples': 1176768, 'steps': 6128, 'loss/train': 2.1220863461494446} 01/28/2022 20:44:45 - INFO - codeparrot_training - Step 6129: {'lr': 0.0004909265518828819, 'samples': 1176960, 'steps': 6129, 'loss/train': 2.459510028362274} 01/28/2022 20:44:49 - INFO - codeparrot_training - Step 6130: {'lr': 0.0004909221831612085, 'samples': 1177152, 'steps': 6130, 'loss/train': 2.603480815887451} 01/28/2022 20:44:53 - INFO - codeparrot_training - Step 6131: {'lr': 0.000490917813407501, 'samples': 1177344, 'steps': 6131, 'loss/train': 1.296525925397873} 01/28/2022 20:44:57 - INFO - codeparrot_training - Step 6132: {'lr': 0.0004909134426217779, 'samples': 1177536, 'steps': 6132, 'loss/train': 1.8711140155792236} 01/28/2022 20:45:02 - INFO - codeparrot_training - Step 6133: {'lr': 0.0004909090708040583, 'samples': 1177728, 'steps': 6133, 'loss/train': 1.5920010209083557} 01/28/2022 20:45:07 - INFO - codeparrot_training - Step 6134: {'lr': 0.0004909046979543608, 'samples': 1177920, 'steps': 6134, 'loss/train': 1.7131401300430298} 01/28/2022 20:45:11 - INFO - codeparrot_training - Step 6135: {'lr': 0.000490900324072704, 'samples': 1178112, 'steps': 6135, 'loss/train': 1.214518278837204} 01/28/2022 20:45:15 - INFO - codeparrot_training - Step 6136: {'lr': 0.0004908959491591065, 'samples': 1178304, 'steps': 6136, 'loss/train': 1.3849202692508698} 01/28/2022 20:45:20 - INFO - codeparrot_training - Step 6137: {'lr': 0.0004908915732135874, 'samples': 1178496, 'steps': 6137, 'loss/train': 1.590735912322998} 01/28/2022 20:45:24 - INFO - codeparrot_training - Step 6138: {'lr': 0.0004908871962361654, 'samples': 1178688, 'steps': 6138, 'loss/train': 2.317778527736664} 01/28/2022 20:45:29 - INFO - codeparrot_training - Step 6139: {'lr': 0.0004908828182268591, 'samples': 1178880, 'steps': 6139, 'loss/train': 1.136342167854309} 01/28/2022 20:45:33 - INFO - codeparrot_training - Step 6140: {'lr': 0.0004908784391856872, 'samples': 1179072, 'steps': 6140, 'loss/train': 2.0879886746406555} 01/28/2022 20:45:38 - INFO - codeparrot_training - Step 6141: {'lr': 0.0004908740591126686, 'samples': 1179264, 'steps': 6141, 'loss/train': 2.260420024394989} 01/28/2022 20:45:42 - INFO - codeparrot_training - Step 6142: {'lr': 0.000490869678007822, 'samples': 1179456, 'steps': 6142, 'loss/train': 2.108934760093689} 01/28/2022 20:45:46 - INFO - codeparrot_training - Step 6143: {'lr': 0.0004908652958711663, 'samples': 1179648, 'steps': 6143, 'loss/train': 0.8295986652374268} 01/28/2022 20:45:52 - INFO - codeparrot_training - Step 6144: {'lr': 0.00049086091270272, 'samples': 1179840, 'steps': 6144, 'loss/train': 2.1576390266418457} 01/28/2022 20:45:57 - INFO - codeparrot_training - Step 6145: {'lr': 0.0004908565285025021, 'samples': 1180032, 'steps': 6145, 'loss/train': 2.053144633769989} 01/28/2022 20:46:01 - INFO - codeparrot_training - Step 6146: {'lr': 0.0004908521432705312, 'samples': 1180224, 'steps': 6146, 'loss/train': 1.6323966979980469} 01/28/2022 20:46:05 - INFO - codeparrot_training - Step 6147: {'lr': 0.0004908477570068263, 'samples': 1180416, 'steps': 6147, 'loss/train': 1.8245576620101929} 01/28/2022 20:46:09 - INFO - codeparrot_training - Step 6148: {'lr': 0.0004908433697114062, 'samples': 1180608, 'steps': 6148, 'loss/train': 1.8078959584236145} 01/28/2022 20:46:15 - INFO - codeparrot_training - Step 6149: {'lr': 0.0004908389813842894, 'samples': 1180800, 'steps': 6149, 'loss/train': 2.0596362948417664} 01/28/2022 20:46:19 - INFO - codeparrot_training - Step 6150: {'lr': 0.0004908345920254949, 'samples': 1180992, 'steps': 6150, 'loss/train': 1.8391767740249634} 01/28/2022 20:46:23 - INFO - codeparrot_training - Step 6151: {'lr': 0.0004908302016350416, 'samples': 1181184, 'steps': 6151, 'loss/train': 1.3176518976688385} 01/28/2022 20:46:27 - INFO - codeparrot_training - Step 6152: {'lr': 0.0004908258102129481, 'samples': 1181376, 'steps': 6152, 'loss/train': 1.371506541967392} 01/28/2022 20:46:32 - INFO - codeparrot_training - Step 6153: {'lr': 0.0004908214177592334, 'samples': 1181568, 'steps': 6153, 'loss/train': 1.9160004258155823} 01/28/2022 20:46:38 - INFO - codeparrot_training - Step 6154: {'lr': 0.000490817024273916, 'samples': 1181760, 'steps': 6154, 'loss/train': 2.178509831428528} 01/28/2022 20:46:42 - INFO - codeparrot_training - Step 6155: {'lr': 0.0004908126297570152, 'samples': 1181952, 'steps': 6155, 'loss/train': 2.2499061226844788} 01/28/2022 20:46:46 - INFO - codeparrot_training - Step 6156: {'lr': 0.0004908082342085494, 'samples': 1182144, 'steps': 6156, 'loss/train': 2.2159525752067566} 01/28/2022 20:46:51 - INFO - codeparrot_training - Step 6157: {'lr': 0.0004908038376285375, 'samples': 1182336, 'steps': 6157, 'loss/train': 1.8702753782272339} 01/28/2022 20:46:55 - INFO - codeparrot_training - Step 6158: {'lr': 0.0004907994400169986, 'samples': 1182528, 'steps': 6158, 'loss/train': 1.2403540313243866} 01/28/2022 20:47:00 - INFO - codeparrot_training - Step 6159: {'lr': 0.0004907950413739514, 'samples': 1182720, 'steps': 6159, 'loss/train': 1.4655900299549103} 01/28/2022 20:47:04 - INFO - codeparrot_training - Step 6160: {'lr': 0.0004907906416994146, 'samples': 1182912, 'steps': 6160, 'loss/train': 1.9345943927764893} 01/28/2022 20:47:09 - INFO - codeparrot_training - Step 6161: {'lr': 0.0004907862409934071, 'samples': 1183104, 'steps': 6161, 'loss/train': 2.7631680965423584} 01/28/2022 20:47:13 - INFO - codeparrot_training - Step 6162: {'lr': 0.0004907818392559479, 'samples': 1183296, 'steps': 6162, 'loss/train': 1.806086003780365} 01/28/2022 20:47:17 - INFO - codeparrot_training - Step 6163: {'lr': 0.0004907774364870557, 'samples': 1183488, 'steps': 6163, 'loss/train': 1.3564863502979279} 01/28/2022 20:47:23 - INFO - codeparrot_training - Step 6164: {'lr': 0.0004907730326867495, 'samples': 1183680, 'steps': 6164, 'loss/train': 2.0523308515548706} 01/28/2022 20:47:27 - INFO - codeparrot_training - Step 6165: {'lr': 0.0004907686278550479, 'samples': 1183872, 'steps': 6165, 'loss/train': 1.6346270442008972} 01/28/2022 20:47:31 - INFO - codeparrot_training - Step 6166: {'lr': 0.0004907642219919701, 'samples': 1184064, 'steps': 6166, 'loss/train': 2.2835991382598877} 01/28/2022 20:47:36 - INFO - codeparrot_training - Step 6167: {'lr': 0.0004907598150975348, 'samples': 1184256, 'steps': 6167, 'loss/train': 2.193592071533203} 01/28/2022 20:47:40 - INFO - codeparrot_training - Step 6168: {'lr': 0.0004907554071717609, 'samples': 1184448, 'steps': 6168, 'loss/train': 1.0643502473831177} 01/28/2022 20:47:46 - INFO - codeparrot_training - Step 6169: {'lr': 0.0004907509982146673, 'samples': 1184640, 'steps': 6169, 'loss/train': 2.1033480763435364} 01/28/2022 20:47:50 - INFO - codeparrot_training - Step 6170: {'lr': 0.0004907465882262728, 'samples': 1184832, 'steps': 6170, 'loss/train': 2.374295175075531} 01/28/2022 20:47:55 - INFO - codeparrot_training - Step 6171: {'lr': 0.0004907421772065965, 'samples': 1185024, 'steps': 6171, 'loss/train': 0.8154372274875641} 01/28/2022 20:47:59 - INFO - codeparrot_training - Step 6172: {'lr': 0.000490737765155657, 'samples': 1185216, 'steps': 6172, 'loss/train': 1.1217349469661713} 01/28/2022 20:48:03 - INFO - codeparrot_training - Step 6173: {'lr': 0.0004907333520734734, 'samples': 1185408, 'steps': 6173, 'loss/train': 1.8832342028617859} 01/28/2022 20:48:08 - INFO - codeparrot_training - Step 6174: {'lr': 0.0004907289379600646, 'samples': 1185600, 'steps': 6174, 'loss/train': 1.6979050040245056} 01/28/2022 20:48:13 - INFO - codeparrot_training - Step 6175: {'lr': 0.0004907245228154495, 'samples': 1185792, 'steps': 6175, 'loss/train': 2.853751301765442} 01/28/2022 20:48:17 - INFO - codeparrot_training - Step 6176: {'lr': 0.0004907201066396469, 'samples': 1185984, 'steps': 6176, 'loss/train': 1.6263883709907532} 01/28/2022 20:48:21 - INFO - codeparrot_training - Step 6177: {'lr': 0.0004907156894326758, 'samples': 1186176, 'steps': 6177, 'loss/train': 1.9007826447486877} 01/28/2022 20:48:25 - INFO - codeparrot_training - Step 6178: {'lr': 0.0004907112711945552, 'samples': 1186368, 'steps': 6178, 'loss/train': 1.6991711854934692} 01/28/2022 20:48:30 - INFO - codeparrot_training - Step 6179: {'lr': 0.000490706851925304, 'samples': 1186560, 'steps': 6179, 'loss/train': 1.1915735006332397} 01/28/2022 20:48:35 - INFO - codeparrot_training - Step 6180: {'lr': 0.0004907024316249408, 'samples': 1186752, 'steps': 6180, 'loss/train': 1.7583641409873962} 01/28/2022 20:48:39 - INFO - codeparrot_training - Step 6181: {'lr': 0.0004906980102934852, 'samples': 1186944, 'steps': 6181, 'loss/train': 1.464445799589157} 01/28/2022 20:48:43 - INFO - codeparrot_training - Step 6182: {'lr': 0.0004906935879309555, 'samples': 1187136, 'steps': 6182, 'loss/train': 1.8592597246170044} 01/28/2022 20:48:47 - INFO - codeparrot_training - Step 6183: {'lr': 0.0004906891645373709, 'samples': 1187328, 'steps': 6183, 'loss/train': 1.3199090659618378} 01/28/2022 20:48:53 - INFO - codeparrot_training - Step 6184: {'lr': 0.0004906847401127504, 'samples': 1187520, 'steps': 6184, 'loss/train': 2.1238426566123962} 01/28/2022 20:48:57 - INFO - codeparrot_training - Step 6185: {'lr': 0.0004906803146571129, 'samples': 1187712, 'steps': 6185, 'loss/train': 1.464550405740738} 01/28/2022 20:49:01 - INFO - codeparrot_training - Step 6186: {'lr': 0.0004906758881704774, 'samples': 1187904, 'steps': 6186, 'loss/train': 1.3733772039413452} 01/28/2022 20:49:05 - INFO - codeparrot_training - Step 6187: {'lr': 0.0004906714606528628, 'samples': 1188096, 'steps': 6187, 'loss/train': 2.245308995246887} 01/28/2022 20:49:10 - INFO - codeparrot_training - Step 6188: {'lr': 0.0004906670321042881, 'samples': 1188288, 'steps': 6188, 'loss/train': 2.6493483781814575} 01/28/2022 20:49:16 - INFO - codeparrot_training - Step 6189: {'lr': 0.0004906626025247722, 'samples': 1188480, 'steps': 6189, 'loss/train': 1.5607309341430664} 01/28/2022 20:49:20 - INFO - codeparrot_training - Step 6190: {'lr': 0.000490658171914334, 'samples': 1188672, 'steps': 6190, 'loss/train': 1.9328924417495728} 01/28/2022 20:49:24 - INFO - codeparrot_training - Step 6191: {'lr': 0.0004906537402729928, 'samples': 1188864, 'steps': 6191, 'loss/train': 2.0198619961738586} 01/28/2022 20:49:29 - INFO - codeparrot_training - Step 6192: {'lr': 0.0004906493076007675, 'samples': 1189056, 'steps': 6192, 'loss/train': 2.7645369172096252} 01/28/2022 20:49:34 - INFO - codeparrot_training - Step 6193: {'lr': 0.0004906448738976768, 'samples': 1189248, 'steps': 6193, 'loss/train': 1.1120708584785461} 01/28/2022 20:49:38 - INFO - codeparrot_training - Step 6194: {'lr': 0.0004906404391637397, 'samples': 1189440, 'steps': 6194, 'loss/train': 1.5854031443595886} 01/28/2022 20:49:43 - INFO - codeparrot_training - Step 6195: {'lr': 0.0004906360033989758, 'samples': 1189632, 'steps': 6195, 'loss/train': 2.3329043984413147} 01/28/2022 20:49:47 - INFO - codeparrot_training - Step 6196: {'lr': 0.0004906315666034034, 'samples': 1189824, 'steps': 6196, 'loss/train': 0.1283990889787674} 01/28/2022 20:49:51 - INFO - codeparrot_training - Step 6197: {'lr': 0.0004906271287770418, 'samples': 1190016, 'steps': 6197, 'loss/train': 1.007532924413681} 01/28/2022 20:49:57 - INFO - codeparrot_training - Step 6198: {'lr': 0.00049062268991991, 'samples': 1190208, 'steps': 6198, 'loss/train': 1.2413560152053833} 01/28/2022 20:50:02 - INFO - codeparrot_training - Step 6199: {'lr': 0.0004906182500320269, 'samples': 1190400, 'steps': 6199, 'loss/train': 1.3050008118152618} 01/28/2022 20:50:06 - INFO - codeparrot_training - Step 6200: {'lr': 0.0004906138091134118, 'samples': 1190592, 'steps': 6200, 'loss/train': 1.504494845867157} 01/28/2022 20:50:10 - INFO - codeparrot_training - Step 6201: {'lr': 0.0004906093671640836, 'samples': 1190784, 'steps': 6201, 'loss/train': 1.9256803393363953} 01/28/2022 20:50:14 - INFO - codeparrot_training - Step 6202: {'lr': 0.0004906049241840612, 'samples': 1190976, 'steps': 6202, 'loss/train': 1.860343337059021} 01/28/2022 20:50:20 - INFO - codeparrot_training - Step 6203: {'lr': 0.0004906004801733635, 'samples': 1191168, 'steps': 6203, 'loss/train': 2.9831560850143433} 01/28/2022 20:50:24 - INFO - codeparrot_training - Step 6204: {'lr': 0.0004905960351320099, 'samples': 1191360, 'steps': 6204, 'loss/train': 1.5336627960205078} 01/28/2022 20:50:28 - INFO - codeparrot_training - Step 6205: {'lr': 0.0004905915890600194, 'samples': 1191552, 'steps': 6205, 'loss/train': 3.002922534942627} 01/28/2022 20:50:32 - INFO - codeparrot_training - Step 6206: {'lr': 0.0004905871419574107, 'samples': 1191744, 'steps': 6206, 'loss/train': 2.081798493862152} 01/28/2022 20:50:37 - INFO - codeparrot_training - Step 6207: {'lr': 0.0004905826938242032, 'samples': 1191936, 'steps': 6207, 'loss/train': 1.903232753276825} 01/28/2022 20:50:42 - INFO - codeparrot_training - Step 6208: {'lr': 0.0004905782446604158, 'samples': 1192128, 'steps': 6208, 'loss/train': 1.8810392618179321} 01/28/2022 20:50:46 - INFO - codeparrot_training - Step 6209: {'lr': 0.0004905737944660676, 'samples': 1192320, 'steps': 6209, 'loss/train': 1.6743623614311218} 01/28/2022 20:50:50 - INFO - codeparrot_training - Step 6210: {'lr': 0.0004905693432411777, 'samples': 1192512, 'steps': 6210, 'loss/train': 1.9760679602622986} 01/28/2022 20:50:55 - INFO - codeparrot_training - Step 6211: {'lr': 0.0004905648909857652, 'samples': 1192704, 'steps': 6211, 'loss/train': 1.3650614619255066} 01/28/2022 20:50:59 - INFO - codeparrot_training - Step 6212: {'lr': 0.0004905604376998489, 'samples': 1192896, 'steps': 6212, 'loss/train': 2.4471189379692078} 01/28/2022 20:51:05 - INFO - codeparrot_training - Step 6213: {'lr': 0.0004905559833834482, 'samples': 1193088, 'steps': 6213, 'loss/train': 1.6654606461524963} 01/28/2022 20:51:09 - INFO - codeparrot_training - Step 6214: {'lr': 0.000490551528036582, 'samples': 1193280, 'steps': 6214, 'loss/train': 2.361317217350006} 01/28/2022 20:51:14 - INFO - codeparrot_training - Step 6215: {'lr': 0.0004905470716592695, 'samples': 1193472, 'steps': 6215, 'loss/train': 1.5515761971473694} 01/28/2022 20:51:18 - INFO - codeparrot_training - Step 6216: {'lr': 0.0004905426142515296, 'samples': 1193664, 'steps': 6216, 'loss/train': 1.4608095288276672} 01/28/2022 20:51:22 - INFO - codeparrot_training - Step 6217: {'lr': 0.0004905381558133817, 'samples': 1193856, 'steps': 6217, 'loss/train': 0.8059877157211304} 01/28/2022 20:51:27 - INFO - codeparrot_training - Step 6218: {'lr': 0.0004905336963448446, 'samples': 1194048, 'steps': 6218, 'loss/train': 2.225888729095459} 01/28/2022 20:51:32 - INFO - codeparrot_training - Step 6219: {'lr': 0.0004905292358459375, 'samples': 1194240, 'steps': 6219, 'loss/train': 2.2219380140304565} 01/28/2022 20:51:36 - INFO - codeparrot_training - Step 6220: {'lr': 0.0004905247743166796, 'samples': 1194432, 'steps': 6220, 'loss/train': 1.700050950050354} 01/28/2022 20:51:40 - INFO - codeparrot_training - Step 6221: {'lr': 0.0004905203117570899, 'samples': 1194624, 'steps': 6221, 'loss/train': 2.325361132621765} 01/28/2022 20:51:44 - INFO - codeparrot_training - Step 6222: {'lr': 0.0004905158481671876, 'samples': 1194816, 'steps': 6222, 'loss/train': 0.28317877650260925} 01/28/2022 20:51:50 - INFO - codeparrot_training - Step 6223: {'lr': 0.0004905113835469918, 'samples': 1195008, 'steps': 6223, 'loss/train': 1.0421966314315796} 01/28/2022 20:51:54 - INFO - codeparrot_training - Step 6224: {'lr': 0.0004905069178965214, 'samples': 1195200, 'steps': 6224, 'loss/train': 1.6534940600395203} 01/28/2022 20:51:58 - INFO - codeparrot_training - Step 6225: {'lr': 0.0004905024512157959, 'samples': 1195392, 'steps': 6225, 'loss/train': 1.379248559474945} 01/28/2022 20:52:02 - INFO - codeparrot_training - Step 6226: {'lr': 0.0004904979835048343, 'samples': 1195584, 'steps': 6226, 'loss/train': 2.402962625026703} 01/28/2022 20:52:07 - INFO - codeparrot_training - Step 6227: {'lr': 0.0004904935147636557, 'samples': 1195776, 'steps': 6227, 'loss/train': 1.491287648677826} 01/28/2022 20:52:13 - INFO - codeparrot_training - Step 6228: {'lr': 0.0004904890449922792, 'samples': 1195968, 'steps': 6228, 'loss/train': 1.3753075003623962} 01/28/2022 20:52:17 - INFO - codeparrot_training - Step 6229: {'lr': 0.0004904845741907241, 'samples': 1196160, 'steps': 6229, 'loss/train': 1.6140666604042053} 01/28/2022 20:52:21 - INFO - codeparrot_training - Step 6230: {'lr': 0.0004904801023590094, 'samples': 1196352, 'steps': 6230, 'loss/train': 3.301070809364319} 01/28/2022 20:52:25 - INFO - codeparrot_training - Step 6231: {'lr': 0.0004904756294971541, 'samples': 1196544, 'steps': 6231, 'loss/train': 2.707369565963745} 01/28/2022 20:52:30 - INFO - codeparrot_training - Step 6232: {'lr': 0.0004904711556051778, 'samples': 1196736, 'steps': 6232, 'loss/train': 1.8719333410263062} 01/28/2022 20:52:35 - INFO - codeparrot_training - Step 6233: {'lr': 0.0004904666806830992, 'samples': 1196928, 'steps': 6233, 'loss/train': 4.032824277877808} 01/28/2022 20:52:39 - INFO - codeparrot_training - Step 6234: {'lr': 0.0004904622047309379, 'samples': 1197120, 'steps': 6234, 'loss/train': 1.858205795288086} 01/28/2022 20:52:43 - INFO - codeparrot_training - Step 6235: {'lr': 0.0004904577277487129, 'samples': 1197312, 'steps': 6235, 'loss/train': 2.256463408470154} 01/28/2022 20:52:48 - INFO - codeparrot_training - Step 6236: {'lr': 0.0004904532497364432, 'samples': 1197504, 'steps': 6236, 'loss/train': 1.9797534942626953} 01/28/2022 20:52:52 - INFO - codeparrot_training - Step 6237: {'lr': 0.0004904487706941481, 'samples': 1197696, 'steps': 6237, 'loss/train': 2.565258264541626} 01/28/2022 20:52:58 - INFO - codeparrot_training - Step 6238: {'lr': 0.000490444290621847, 'samples': 1197888, 'steps': 6238, 'loss/train': 2.0547242760658264} 01/28/2022 20:53:02 - INFO - codeparrot_training - Step 6239: {'lr': 0.0004904398095195588, 'samples': 1198080, 'steps': 6239, 'loss/train': 1.4449584782123566} 01/28/2022 20:53:07 - INFO - codeparrot_training - Step 6240: {'lr': 0.0004904353273873028, 'samples': 1198272, 'steps': 6240, 'loss/train': 1.7752355933189392} 01/28/2022 20:53:11 - INFO - codeparrot_training - Step 6241: {'lr': 0.0004904308442250983, 'samples': 1198464, 'steps': 6241, 'loss/train': 2.104846179485321} 01/28/2022 20:53:15 - INFO - codeparrot_training - Step 6242: {'lr': 0.0004904263600329643, 'samples': 1198656, 'steps': 6242, 'loss/train': 0.5480470508337021} 01/28/2022 20:53:21 - INFO - codeparrot_training - Step 6243: {'lr': 0.0004904218748109201, 'samples': 1198848, 'steps': 6243, 'loss/train': 2.6313096284866333} 01/28/2022 20:53:26 - INFO - codeparrot_training - Step 6244: {'lr': 0.000490417388558985, 'samples': 1199040, 'steps': 6244, 'loss/train': 1.655142366886139} 01/28/2022 20:53:30 - INFO - codeparrot_training - Step 6245: {'lr': 0.0004904129012771782, 'samples': 1199232, 'steps': 6245, 'loss/train': 1.67819344997406} 01/28/2022 20:53:34 - INFO - codeparrot_training - Step 6246: {'lr': 0.0004904084129655188, 'samples': 1199424, 'steps': 6246, 'loss/train': 2.3707576990127563} 01/28/2022 20:53:39 - INFO - codeparrot_training - Step 6247: {'lr': 0.000490403923624026, 'samples': 1199616, 'steps': 6247, 'loss/train': 1.4639443159103394} 01/28/2022 20:53:44 - INFO - codeparrot_training - Step 6248: {'lr': 0.0004903994332527193, 'samples': 1199808, 'steps': 6248, 'loss/train': 1.8284255862236023} 01/28/2022 20:53:48 - INFO - codeparrot_training - Step 6249: {'lr': 0.0004903949418516178, 'samples': 1200000, 'steps': 6249, 'loss/train': 1.8193062543869019} 01/28/2022 20:53:53 - INFO - codeparrot_training - Step 6250: {'lr': 0.0004903904494207405, 'samples': 1200192, 'steps': 6250, 'loss/train': 2.042046546936035} 01/28/2022 20:53:57 - INFO - codeparrot_training - Step 6251: {'lr': 0.000490385955960107, 'samples': 1200384, 'steps': 6251, 'loss/train': 0.330249086022377} 01/28/2022 20:54:01 - INFO - codeparrot_training - Step 6252: {'lr': 0.0004903814614697363, 'samples': 1200576, 'steps': 6252, 'loss/train': 1.8231140971183777} 01/28/2022 20:54:07 - INFO - codeparrot_training - Step 6253: {'lr': 0.0004903769659496478, 'samples': 1200768, 'steps': 6253, 'loss/train': 2.3314080834388733} 01/28/2022 20:54:11 - INFO - codeparrot_training - Step 6254: {'lr': 0.0004903724693998607, 'samples': 1200960, 'steps': 6254, 'loss/train': 2.0564653873443604} 01/28/2022 20:54:15 - INFO - codeparrot_training - Step 6255: {'lr': 0.0004903679718203942, 'samples': 1201152, 'steps': 6255, 'loss/train': 2.225826859474182} 01/28/2022 20:54:19 - INFO - codeparrot_training - Step 6256: {'lr': 0.0004903634732112678, 'samples': 1201344, 'steps': 6256, 'loss/train': 2.0649853348731995} 01/28/2022 20:54:23 - INFO - codeparrot_training - Step 6257: {'lr': 0.0004903589735725004, 'samples': 1201536, 'steps': 6257, 'loss/train': 2.4588840007781982} 01/28/2022 20:54:29 - INFO - codeparrot_training - Step 6258: {'lr': 0.0004903544729041116, 'samples': 1201728, 'steps': 6258, 'loss/train': 2.0849602818489075} 01/28/2022 20:54:34 - INFO - codeparrot_training - Step 6259: {'lr': 0.0004903499712061206, 'samples': 1201920, 'steps': 6259, 'loss/train': 0.7238479256629944} 01/28/2022 20:54:38 - INFO - codeparrot_training - Step 6260: {'lr': 0.0004903454684785465, 'samples': 1202112, 'steps': 6260, 'loss/train': 1.0564101934432983} 01/28/2022 20:54:42 - INFO - codeparrot_training - Step 6261: {'lr': 0.0004903409647214088, 'samples': 1202304, 'steps': 6261, 'loss/train': 2.762028694152832} 01/28/2022 20:54:46 - INFO - codeparrot_training - Step 6262: {'lr': 0.0004903364599347268, 'samples': 1202496, 'steps': 6262, 'loss/train': 2.3094683289527893} 01/28/2022 20:54:52 - INFO - codeparrot_training - Step 6263: {'lr': 0.0004903319541185196, 'samples': 1202688, 'steps': 6263, 'loss/train': 1.3088912665843964} 01/28/2022 20:54:56 - INFO - codeparrot_training - Step 6264: {'lr': 0.0004903274472728067, 'samples': 1202880, 'steps': 6264, 'loss/train': 1.2403918504714966} 01/28/2022 20:55:00 - INFO - codeparrot_training - Step 6265: {'lr': 0.0004903229393976073, 'samples': 1203072, 'steps': 6265, 'loss/train': 1.6566980481147766} 01/28/2022 20:55:04 - INFO - codeparrot_training - Step 6266: {'lr': 0.0004903184304929408, 'samples': 1203264, 'steps': 6266, 'loss/train': 2.684384822845459} 01/28/2022 20:55:10 - INFO - codeparrot_training - Step 6267: {'lr': 0.0004903139205588264, 'samples': 1203456, 'steps': 6267, 'loss/train': 1.6623319387435913} 01/28/2022 20:55:14 - INFO - codeparrot_training - Step 6268: {'lr': 0.0004903094095952834, 'samples': 1203648, 'steps': 6268, 'loss/train': 1.8824116587638855} 01/28/2022 20:55:18 - INFO - codeparrot_training - Step 6269: {'lr': 0.0004903048976023313, 'samples': 1203840, 'steps': 6269, 'loss/train': 1.8839455246925354} 01/28/2022 20:55:22 - INFO - codeparrot_training - Step 6270: {'lr': 0.0004903003845799893, 'samples': 1204032, 'steps': 6270, 'loss/train': 0.38941052556037903} 01/28/2022 20:55:27 - INFO - codeparrot_training - Step 6271: {'lr': 0.0004902958705282767, 'samples': 1204224, 'steps': 6271, 'loss/train': 0.3530512601137161} 01/28/2022 20:55:33 - INFO - codeparrot_training - Step 6272: {'lr': 0.000490291355447213, 'samples': 1204416, 'steps': 6272, 'loss/train': 1.7265851497650146} 01/28/2022 20:55:37 - INFO - codeparrot_training - Step 6273: {'lr': 0.0004902868393368174, 'samples': 1204608, 'steps': 6273, 'loss/train': 2.3026492595672607} 01/28/2022 20:55:41 - INFO - codeparrot_training - Step 6274: {'lr': 0.0004902823221971092, 'samples': 1204800, 'steps': 6274, 'loss/train': 1.8013454675674438} 01/28/2022 20:55:45 - INFO - codeparrot_training - Step 6275: {'lr': 0.000490277804028108, 'samples': 1204992, 'steps': 6275, 'loss/train': 1.2412504255771637} 01/28/2022 20:55:50 - INFO - codeparrot_training - Step 6276: {'lr': 0.0004902732848298328, 'samples': 1205184, 'steps': 6276, 'loss/train': 1.570410668849945} 01/28/2022 20:55:55 - INFO - codeparrot_training - Step 6277: {'lr': 0.0004902687646023032, 'samples': 1205376, 'steps': 6277, 'loss/train': 1.8568745255470276} 01/28/2022 20:55:59 - INFO - codeparrot_training - Step 6278: {'lr': 0.0004902642433455385, 'samples': 1205568, 'steps': 6278, 'loss/train': 1.9701796174049377} 01/28/2022 20:56:04 - INFO - codeparrot_training - Step 6279: {'lr': 0.0004902597210595581, 'samples': 1205760, 'steps': 6279, 'loss/train': 0.5994447469711304} 01/28/2022 20:56:08 - INFO - codeparrot_training - Step 6280: {'lr': 0.0004902551977443813, 'samples': 1205952, 'steps': 6280, 'loss/train': 1.5713751912117004} 01/28/2022 20:56:12 - INFO - codeparrot_training - Step 6281: {'lr': 0.0004902506734000276, 'samples': 1206144, 'steps': 6281, 'loss/train': 1.3932976126670837} 01/28/2022 20:56:18 - INFO - codeparrot_training - Step 6282: {'lr': 0.0004902461480265163, 'samples': 1206336, 'steps': 6282, 'loss/train': 1.7141624093055725} 01/28/2022 20:56:22 - INFO - codeparrot_training - Step 6283: {'lr': 0.0004902416216238667, 'samples': 1206528, 'steps': 6283, 'loss/train': 1.46879044175148} 01/28/2022 20:56:27 - INFO - codeparrot_training - Step 6284: {'lr': 0.0004902370941920984, 'samples': 1206720, 'steps': 6284, 'loss/train': 2.0743521451950073} 01/28/2022 20:56:31 - INFO - codeparrot_training - Step 6285: {'lr': 0.0004902325657312306, 'samples': 1206912, 'steps': 6285, 'loss/train': 2.4608895778656006} 01/28/2022 20:56:35 - INFO - codeparrot_training - Step 6286: {'lr': 0.0004902280362412828, 'samples': 1207104, 'steps': 6286, 'loss/train': 1.3326162099838257} 01/28/2022 20:56:41 - INFO - codeparrot_training - Step 6287: {'lr': 0.0004902235057222743, 'samples': 1207296, 'steps': 6287, 'loss/train': 0.3862214684486389} 01/28/2022 20:56:45 - INFO - codeparrot_training - Step 6288: {'lr': 0.0004902189741742246, 'samples': 1207488, 'steps': 6288, 'loss/train': 0.7672557234764099} 01/28/2022 20:56:49 - INFO - codeparrot_training - Step 6289: {'lr': 0.0004902144415971532, 'samples': 1207680, 'steps': 6289, 'loss/train': 2.0635589361190796} 01/28/2022 20:56:53 - INFO - codeparrot_training - Step 6290: {'lr': 0.0004902099079910794, 'samples': 1207872, 'steps': 6290, 'loss/train': 1.0571885704994202} 01/28/2022 20:56:58 - INFO - codeparrot_training - Step 6291: {'lr': 0.0004902053733560225, 'samples': 1208064, 'steps': 6291, 'loss/train': 0.6772766411304474} 01/28/2022 20:57:02 - INFO - codeparrot_training - Step 6292: {'lr': 0.0004902008376920021, 'samples': 1208256, 'steps': 6292, 'loss/train': 3.3749345541000366} 01/28/2022 20:57:07 - INFO - codeparrot_training - Step 6293: {'lr': 0.0004901963009990376, 'samples': 1208448, 'steps': 6293, 'loss/train': 1.8110032081604004} 01/28/2022 20:57:11 - INFO - codeparrot_training - Step 6294: {'lr': 0.0004901917632771485, 'samples': 1208640, 'steps': 6294, 'loss/train': 1.4534622430801392} 01/28/2022 20:57:15 - INFO - codeparrot_training - Step 6295: {'lr': 0.000490187224526354, 'samples': 1208832, 'steps': 6295, 'loss/train': 1.9657548666000366} 01/28/2022 20:57:20 - INFO - codeparrot_training - Step 6296: {'lr': 0.0004901826847466738, 'samples': 1209024, 'steps': 6296, 'loss/train': 1.5850149393081665} 01/28/2022 20:57:26 - INFO - codeparrot_training - Step 6297: {'lr': 0.0004901781439381272, 'samples': 1209216, 'steps': 6297, 'loss/train': 1.4797134697437286} 01/28/2022 20:57:30 - INFO - codeparrot_training - Step 6298: {'lr': 0.0004901736021007337, 'samples': 1209408, 'steps': 6298, 'loss/train': 1.800644338130951} 01/28/2022 20:57:34 - INFO - codeparrot_training - Step 6299: {'lr': 0.0004901690592345127, 'samples': 1209600, 'steps': 6299, 'loss/train': 1.8253079652786255} 01/28/2022 20:57:38 - INFO - codeparrot_training - Step 6300: {'lr': 0.0004901645153394838, 'samples': 1209792, 'steps': 6300, 'loss/train': 1.236857682466507} 01/28/2022 20:57:43 - INFO - codeparrot_training - Step 6301: {'lr': 0.0004901599704156664, 'samples': 1209984, 'steps': 6301, 'loss/train': 1.141487181186676} 01/28/2022 20:57:47 - INFO - codeparrot_training - Step 6302: {'lr': 0.00049015542446308, 'samples': 1210176, 'steps': 6302, 'loss/train': 0.512945830821991} 01/28/2022 20:57:53 - INFO - codeparrot_training - Step 6303: {'lr': 0.0004901508774817438, 'samples': 1210368, 'steps': 6303, 'loss/train': 2.062897503376007} 01/28/2022 20:57:57 - INFO - codeparrot_training - Step 6304: {'lr': 0.0004901463294716776, 'samples': 1210560, 'steps': 6304, 'loss/train': 1.8032750487327576} 01/28/2022 20:58:02 - INFO - codeparrot_training - Step 6305: {'lr': 0.0004901417804329008, 'samples': 1210752, 'steps': 6305, 'loss/train': 2.2147879600524902} 01/28/2022 20:58:06 - INFO - codeparrot_training - Step 6306: {'lr': 0.0004901372303654329, 'samples': 1210944, 'steps': 6306, 'loss/train': 2.5724653601646423} 01/28/2022 20:58:10 - INFO - codeparrot_training - Step 6307: {'lr': 0.0004901326792692934, 'samples': 1211136, 'steps': 6307, 'loss/train': 0.6255944073200226} 01/28/2022 20:58:15 - INFO - codeparrot_training - Step 6308: {'lr': 0.0004901281271445016, 'samples': 1211328, 'steps': 6308, 'loss/train': 1.509781837463379} 01/28/2022 20:58:20 - INFO - codeparrot_training - Step 6309: {'lr': 0.0004901235739910772, 'samples': 1211520, 'steps': 6309, 'loss/train': 1.5349275469779968} 01/28/2022 20:58:24 - INFO - codeparrot_training - Step 6310: {'lr': 0.0004901190198090397, 'samples': 1211712, 'steps': 6310, 'loss/train': 2.4303449392318726} 01/28/2022 20:58:28 - INFO - codeparrot_training - Step 6311: {'lr': 0.0004901144645984086, 'samples': 1211904, 'steps': 6311, 'loss/train': 1.1435196697711945} 01/28/2022 20:58:32 - INFO - codeparrot_training - Step 6312: {'lr': 0.0004901099083592034, 'samples': 1212096, 'steps': 6312, 'loss/train': 1.7517446279525757} 01/28/2022 20:58:38 - INFO - codeparrot_training - Step 6313: {'lr': 0.0004901053510914434, 'samples': 1212288, 'steps': 6313, 'loss/train': 1.7168768048286438} 01/28/2022 20:58:42 - INFO - codeparrot_training - Step 6314: {'lr': 0.0004901007927951485, 'samples': 1212480, 'steps': 6314, 'loss/train': 2.1275664567947388} 01/28/2022 20:58:46 - INFO - codeparrot_training - Step 6315: {'lr': 0.000490096233470338, 'samples': 1212672, 'steps': 6315, 'loss/train': 1.4942576587200165} 01/28/2022 20:58:51 - INFO - codeparrot_training - Step 6316: {'lr': 0.0004900916731170314, 'samples': 1212864, 'steps': 6316, 'loss/train': 1.3421196341514587} 01/28/2022 20:58:55 - INFO - codeparrot_training - Step 6317: {'lr': 0.0004900871117352484, 'samples': 1213056, 'steps': 6317, 'loss/train': 2.240492284297943} 01/28/2022 20:59:01 - INFO - codeparrot_training - Step 6318: {'lr': 0.0004900825493250084, 'samples': 1213248, 'steps': 6318, 'loss/train': 2.2874438166618347} 01/28/2022 20:59:05 - INFO - codeparrot_training - Step 6319: {'lr': 0.000490077985886331, 'samples': 1213440, 'steps': 6319, 'loss/train': 2.2914302945137024} 01/28/2022 20:59:09 - INFO - codeparrot_training - Step 6320: {'lr': 0.0004900734214192358, 'samples': 1213632, 'steps': 6320, 'loss/train': 0.426875039935112} 01/28/2022 20:59:14 - INFO - codeparrot_training - Step 6321: {'lr': 0.0004900688559237422, 'samples': 1213824, 'steps': 6321, 'loss/train': 2.136794149875641} 01/28/2022 20:59:18 - INFO - codeparrot_training - Step 6322: {'lr': 0.0004900642893998699, 'samples': 1214016, 'steps': 6322, 'loss/train': 1.4453448951244354} 01/28/2022 20:59:23 - INFO - codeparrot_training - Step 6323: {'lr': 0.0004900597218476385, 'samples': 1214208, 'steps': 6323, 'loss/train': 2.111542224884033} 01/28/2022 20:59:27 - INFO - codeparrot_training - Step 6324: {'lr': 0.0004900551532670673, 'samples': 1214400, 'steps': 6324, 'loss/train': 1.4067336916923523} 01/28/2022 20:59:32 - INFO - codeparrot_training - Step 6325: {'lr': 0.0004900505836581763, 'samples': 1214592, 'steps': 6325, 'loss/train': 1.964145004749298} 01/28/2022 20:59:36 - INFO - codeparrot_training - Step 6326: {'lr': 0.0004900460130209845, 'samples': 1214784, 'steps': 6326, 'loss/train': 1.9351606965065002} 01/28/2022 20:59:40 - INFO - codeparrot_training - Step 6327: {'lr': 0.000490041441355512, 'samples': 1214976, 'steps': 6327, 'loss/train': 1.7443140149116516} 01/28/2022 20:59:46 - INFO - codeparrot_training - Step 6328: {'lr': 0.0004900368686617783, 'samples': 1215168, 'steps': 6328, 'loss/train': 1.236956387758255} 01/28/2022 20:59:51 - INFO - codeparrot_training - Step 6329: {'lr': 0.0004900322949398026, 'samples': 1215360, 'steps': 6329, 'loss/train': 0.9128149151802063} 01/28/2022 20:59:55 - INFO - codeparrot_training - Step 6330: {'lr': 0.000490027720189605, 'samples': 1215552, 'steps': 6330, 'loss/train': 0.5841621458530426} 01/28/2022 20:59:59 - INFO - codeparrot_training - Step 6331: {'lr': 0.0004900231444112047, 'samples': 1215744, 'steps': 6331, 'loss/train': 1.773202657699585} 01/28/2022 21:00:03 - INFO - codeparrot_training - Step 6332: {'lr': 0.0004900185676046214, 'samples': 1215936, 'steps': 6332, 'loss/train': 1.1673296391963959} 01/28/2022 21:00:09 - INFO - codeparrot_training - Step 6333: {'lr': 0.0004900139897698751, 'samples': 1216128, 'steps': 6333, 'loss/train': 2.043804109096527} 01/28/2022 21:00:13 - INFO - codeparrot_training - Step 6334: {'lr': 0.0004900094109069848, 'samples': 1216320, 'steps': 6334, 'loss/train': 0.36973996460437775} 01/28/2022 21:00:18 - INFO - codeparrot_training - Step 6335: {'lr': 0.0004900048310159705, 'samples': 1216512, 'steps': 6335, 'loss/train': 2.6170778274536133} 01/28/2022 21:00:22 - INFO - codeparrot_training - Step 6336: {'lr': 0.0004900002500968516, 'samples': 1216704, 'steps': 6336, 'loss/train': 2.295755445957184} 01/28/2022 21:00:26 - INFO - codeparrot_training - Step 6337: {'lr': 0.000489995668149648, 'samples': 1216896, 'steps': 6337, 'loss/train': 1.6333284974098206} 01/28/2022 21:00:30 - INFO - codeparrot_training - Step 6338: {'lr': 0.0004899910851743791, 'samples': 1217088, 'steps': 6338, 'loss/train': 1.699441909790039} 01/28/2022 21:00:36 - INFO - codeparrot_training - Step 6339: {'lr': 0.0004899865011710646, 'samples': 1217280, 'steps': 6339, 'loss/train': 1.7434312105178833} 01/28/2022 21:00:40 - INFO - codeparrot_training - Step 6340: {'lr': 0.0004899819161397241, 'samples': 1217472, 'steps': 6340, 'loss/train': 1.3081441819667816} 01/28/2022 21:00:44 - INFO - codeparrot_training - Step 6341: {'lr': 0.0004899773300803774, 'samples': 1217664, 'steps': 6341, 'loss/train': 2.0862836837768555} 01/28/2022 21:00:49 - INFO - codeparrot_training - Step 6342: {'lr': 0.0004899727429930438, 'samples': 1217856, 'steps': 6342, 'loss/train': 1.94777512550354} 01/28/2022 21:00:53 - INFO - codeparrot_training - Step 6343: {'lr': 0.0004899681548777434, 'samples': 1218048, 'steps': 6343, 'loss/train': 2.1078291535377502} 01/28/2022 21:00:58 - INFO - codeparrot_training - Step 6344: {'lr': 0.0004899635657344954, 'samples': 1218240, 'steps': 6344, 'loss/train': 1.3604019284248352} 01/28/2022 21:01:02 - INFO - codeparrot_training - Step 6345: {'lr': 0.0004899589755633198, 'samples': 1218432, 'steps': 6345, 'loss/train': 1.8241233229637146} 01/28/2022 21:01:07 - INFO - codeparrot_training - Step 6346: {'lr': 0.0004899543843642362, 'samples': 1218624, 'steps': 6346, 'loss/train': 1.7638605237007141} 01/28/2022 21:01:11 - INFO - codeparrot_training - Step 6347: {'lr': 0.0004899497921372641, 'samples': 1218816, 'steps': 6347, 'loss/train': 1.9779446125030518} 01/28/2022 21:01:15 - INFO - codeparrot_training - Step 6348: {'lr': 0.0004899451988824233, 'samples': 1219008, 'steps': 6348, 'loss/train': 1.8830344676971436} 01/28/2022 21:01:22 - INFO - codeparrot_training - Step 6349: {'lr': 0.0004899406045997336, 'samples': 1219200, 'steps': 6349, 'loss/train': 1.281896710395813} 01/28/2022 21:01:26 - INFO - codeparrot_training - Step 6350: {'lr': 0.0004899360092892143, 'samples': 1219392, 'steps': 6350, 'loss/train': 1.9688063263893127} 01/28/2022 21:01:30 - INFO - codeparrot_training - Step 6351: {'lr': 0.0004899314129508855, 'samples': 1219584, 'steps': 6351, 'loss/train': 2.442830801010132} 01/28/2022 21:01:34 - INFO - codeparrot_training - Step 6352: {'lr': 0.0004899268155847667, 'samples': 1219776, 'steps': 6352, 'loss/train': 1.7059789896011353} 01/28/2022 21:01:39 - INFO - codeparrot_training - Step 6353: {'lr': 0.0004899222171908776, 'samples': 1219968, 'steps': 6353, 'loss/train': 1.0786448121070862} 01/28/2022 21:01:44 - INFO - codeparrot_training - Step 6354: {'lr': 0.0004899176177692379, 'samples': 1220160, 'steps': 6354, 'loss/train': 2.182374358177185} 01/28/2022 21:01:48 - INFO - codeparrot_training - Step 6355: {'lr': 0.0004899130173198672, 'samples': 1220352, 'steps': 6355, 'loss/train': 2.1761176586151123} 01/28/2022 21:01:53 - INFO - codeparrot_training - Step 6356: {'lr': 0.0004899084158427855, 'samples': 1220544, 'steps': 6356, 'loss/train': 1.6773825287818909} 01/28/2022 21:01:57 - INFO - codeparrot_training - Step 6357: {'lr': 0.0004899038133380121, 'samples': 1220736, 'steps': 6357, 'loss/train': 3.4501705169677734} 01/28/2022 21:02:02 - INFO - codeparrot_training - Step 6358: {'lr': 0.0004898992098055671, 'samples': 1220928, 'steps': 6358, 'loss/train': 0.9733407497406006} 01/28/2022 21:02:07 - INFO - codeparrot_training - Step 6359: {'lr': 0.00048989460524547, 'samples': 1221120, 'steps': 6359, 'loss/train': 1.7358216047286987} 01/28/2022 21:02:11 - INFO - codeparrot_training - Step 6360: {'lr': 0.0004898899996577407, 'samples': 1221312, 'steps': 6360, 'loss/train': 1.9879785776138306} 01/28/2022 21:02:15 - INFO - codeparrot_training - Step 6361: {'lr': 0.0004898853930423987, 'samples': 1221504, 'steps': 6361, 'loss/train': 0.7329676598310471} 01/28/2022 21:02:19 - INFO - codeparrot_training - Step 6362: {'lr': 0.0004898807853994639, 'samples': 1221696, 'steps': 6362, 'loss/train': 0.8119922876358032} 01/28/2022 21:02:25 - INFO - codeparrot_training - Step 6363: {'lr': 0.000489876176728956, 'samples': 1221888, 'steps': 6363, 'loss/train': 2.2188958525657654} 01/28/2022 21:02:30 - INFO - codeparrot_training - Step 6364: {'lr': 0.0004898715670308947, 'samples': 1222080, 'steps': 6364, 'loss/train': 1.5268843173980713} 01/28/2022 21:02:34 - INFO - codeparrot_training - Step 6365: {'lr': 0.0004898669563052997, 'samples': 1222272, 'steps': 6365, 'loss/train': 0.9428963363170624} 01/28/2022 21:02:38 - INFO - codeparrot_training - Step 6366: {'lr': 0.0004898623445521909, 'samples': 1222464, 'steps': 6366, 'loss/train': 1.9491974115371704} 01/28/2022 21:02:42 - INFO - codeparrot_training - Step 6367: {'lr': 0.000489857731771588, 'samples': 1222656, 'steps': 6367, 'loss/train': 1.7752134203910828} 01/28/2022 21:02:48 - INFO - codeparrot_training - Step 6368: {'lr': 0.0004898531179635108, 'samples': 1222848, 'steps': 6368, 'loss/train': 1.3962070941925049} 01/28/2022 21:02:52 - INFO - codeparrot_training - Step 6369: {'lr': 0.0004898485031279788, 'samples': 1223040, 'steps': 6369, 'loss/train': 2.3484385013580322} 01/28/2022 21:02:56 - INFO - codeparrot_training - Step 6370: {'lr': 0.0004898438872650121, 'samples': 1223232, 'steps': 6370, 'loss/train': 0.4622957110404968} 01/28/2022 21:03:00 - INFO - codeparrot_training - Step 6371: {'lr': 0.0004898392703746304, 'samples': 1223424, 'steps': 6371, 'loss/train': 1.8255963921546936} 01/28/2022 21:03:05 - INFO - codeparrot_training - Step 6372: {'lr': 0.0004898346524568533, 'samples': 1223616, 'steps': 6372, 'loss/train': 1.4728496074676514} 01/28/2022 21:03:11 - INFO - codeparrot_training - Step 6373: {'lr': 0.0004898300335117008, 'samples': 1223808, 'steps': 6373, 'loss/train': 2.535050332546234} 01/28/2022 21:03:15 - INFO - codeparrot_training - Step 6374: {'lr': 0.0004898254135391926, 'samples': 1224000, 'steps': 6374, 'loss/train': 2.3329151272773743} 01/28/2022 21:03:19 - INFO - codeparrot_training - Step 6375: {'lr': 0.0004898207925393485, 'samples': 1224192, 'steps': 6375, 'loss/train': 1.3264693915843964} 01/28/2022 21:03:23 - INFO - codeparrot_training - Step 6376: {'lr': 0.0004898161705121882, 'samples': 1224384, 'steps': 6376, 'loss/train': 1.316944420337677} 01/28/2022 21:03:28 - INFO - codeparrot_training - Step 6377: {'lr': 0.0004898115474577315, 'samples': 1224576, 'steps': 6377, 'loss/train': 2.68101042509079} 01/28/2022 21:03:33 - INFO - codeparrot_training - Step 6378: {'lr': 0.0004898069233759985, 'samples': 1224768, 'steps': 6378, 'loss/train': 2.1156111359596252} 01/28/2022 21:03:37 - INFO - codeparrot_training - Step 6379: {'lr': 0.0004898022982670085, 'samples': 1224960, 'steps': 6379, 'loss/train': 1.4623793363571167} 01/28/2022 21:03:42 - INFO - codeparrot_training - Step 6380: {'lr': 0.0004897976721307818, 'samples': 1225152, 'steps': 6380, 'loss/train': 3.038652777671814} 01/28/2022 21:03:46 - INFO - codeparrot_training - Step 6381: {'lr': 0.000489793044967338, 'samples': 1225344, 'steps': 6381, 'loss/train': 1.7493769526481628} 01/28/2022 21:03:50 - INFO - codeparrot_training - Step 6382: {'lr': 0.0004897884167766969, 'samples': 1225536, 'steps': 6382, 'loss/train': 1.5387917160987854} 01/28/2022 21:03:56 - INFO - codeparrot_training - Step 6383: {'lr': 0.0004897837875588784, 'samples': 1225728, 'steps': 6383, 'loss/train': 1.4192462861537933} 01/28/2022 21:04:00 - INFO - codeparrot_training - Step 6384: {'lr': 0.0004897791573139022, 'samples': 1225920, 'steps': 6384, 'loss/train': 2.272040069103241} 01/28/2022 21:04:04 - INFO - codeparrot_training - Step 6385: {'lr': 0.0004897745260417884, 'samples': 1226112, 'steps': 6385, 'loss/train': 0.7607881128787994} 01/28/2022 21:04:08 - INFO - codeparrot_training - Step 6386: {'lr': 0.0004897698937425566, 'samples': 1226304, 'steps': 6386, 'loss/train': 2.4117146730422974} 01/28/2022 21:04:13 - INFO - codeparrot_training - Step 6387: {'lr': 0.0004897652604162266, 'samples': 1226496, 'steps': 6387, 'loss/train': 1.943731963634491} 01/28/2022 21:04:19 - INFO - codeparrot_training - Step 6388: {'lr': 0.0004897606260628184, 'samples': 1226688, 'steps': 6388, 'loss/train': 2.0199195742607117} 01/28/2022 21:04:23 - INFO - codeparrot_training - Step 6389: {'lr': 0.0004897559906823518, 'samples': 1226880, 'steps': 6389, 'loss/train': 1.7645108699798584} 01/28/2022 21:04:27 - INFO - codeparrot_training - Step 6390: {'lr': 0.0004897513542748468, 'samples': 1227072, 'steps': 6390, 'loss/train': 0.18203243985772133} 01/28/2022 21:04:31 - INFO - codeparrot_training - Step 6391: {'lr': 0.0004897467168403231, 'samples': 1227264, 'steps': 6391, 'loss/train': 1.2689986824989319} 01/28/2022 21:04:36 - INFO - codeparrot_training - Step 6392: {'lr': 0.0004897420783788006, 'samples': 1227456, 'steps': 6392, 'loss/train': 2.1158748865127563} 01/28/2022 21:04:41 - INFO - codeparrot_training - Step 6393: {'lr': 0.0004897374388902991, 'samples': 1227648, 'steps': 6393, 'loss/train': 2.5214381217956543} 01/28/2022 21:04:45 - INFO - codeparrot_training - Step 6394: {'lr': 0.0004897327983748385, 'samples': 1227840, 'steps': 6394, 'loss/train': 1.4766999185085297} 01/28/2022 21:04:50 - INFO - codeparrot_training - Step 6395: {'lr': 0.0004897281568324387, 'samples': 1228032, 'steps': 6395, 'loss/train': 1.7244404554367065} 01/28/2022 21:04:54 - INFO - codeparrot_training - Step 6396: {'lr': 0.0004897235142631197, 'samples': 1228224, 'steps': 6396, 'loss/train': 1.0341411530971527} 01/28/2022 21:04:58 - INFO - codeparrot_training - Step 6397: {'lr': 0.0004897188706669012, 'samples': 1228416, 'steps': 6397, 'loss/train': 1.9957927465438843} 01/28/2022 21:05:03 - INFO - codeparrot_training - Step 6398: {'lr': 0.0004897142260438032, 'samples': 1228608, 'steps': 6398, 'loss/train': 1.8569304943084717} 01/28/2022 21:05:08 - INFO - codeparrot_training - Step 6399: {'lr': 0.0004897095803938456, 'samples': 1228800, 'steps': 6399, 'loss/train': 0.4849046319723129} 01/28/2022 21:05:12 - INFO - codeparrot_training - Step 6400: {'lr': 0.0004897049337170483, 'samples': 1228992, 'steps': 6400, 'loss/train': 1.4317325949668884} 01/28/2022 21:05:16 - INFO - codeparrot_training - Step 6401: {'lr': 0.0004897002860134311, 'samples': 1229184, 'steps': 6401, 'loss/train': 1.4618272483348846} 01/28/2022 21:05:20 - INFO - codeparrot_training - Step 6402: {'lr': 0.0004896956372830141, 'samples': 1229376, 'steps': 6402, 'loss/train': 0.7153126001358032} 01/28/2022 21:05:26 - INFO - codeparrot_training - Step 6403: {'lr': 0.000489690987525817, 'samples': 1229568, 'steps': 6403, 'loss/train': 3.9635403156280518} 01/28/2022 21:05:30 - INFO - codeparrot_training - Step 6404: {'lr': 0.0004896863367418598, 'samples': 1229760, 'steps': 6404, 'loss/train': 1.7332255840301514} 01/28/2022 21:05:34 - INFO - codeparrot_training - Step 6405: {'lr': 0.0004896816849311625, 'samples': 1229952, 'steps': 6405, 'loss/train': 2.73608136177063} 01/28/2022 21:05:38 - INFO - codeparrot_training - Step 6406: {'lr': 0.000489677032093745, 'samples': 1230144, 'steps': 6406, 'loss/train': 1.9827572107315063} 01/28/2022 21:05:43 - INFO - codeparrot_training - Step 6407: {'lr': 0.0004896723782296272, 'samples': 1230336, 'steps': 6407, 'loss/train': 2.562765955924988} 01/28/2022 21:05:49 - INFO - codeparrot_training - Step 6408: {'lr': 0.0004896677233388289, 'samples': 1230528, 'steps': 6408, 'loss/train': 2.1936927437782288} 01/28/2022 21:05:53 - INFO - codeparrot_training - Step 6409: {'lr': 0.0004896630674213703, 'samples': 1230720, 'steps': 6409, 'loss/train': 1.8328789472579956} 01/28/2022 21:05:57 - INFO - codeparrot_training - Step 6410: {'lr': 0.0004896584104772712, 'samples': 1230912, 'steps': 6410, 'loss/train': 1.3346597850322723} 01/28/2022 21:06:01 - INFO - codeparrot_training - Step 6411: {'lr': 0.0004896537525065516, 'samples': 1231104, 'steps': 6411, 'loss/train': 1.8108825087547302} 01/28/2022 21:06:07 - INFO - codeparrot_training - Step 6412: {'lr': 0.0004896490935092314, 'samples': 1231296, 'steps': 6412, 'loss/train': 1.8419049382209778} 01/28/2022 21:06:11 - INFO - codeparrot_training - Step 6413: {'lr': 0.0004896444334853305, 'samples': 1231488, 'steps': 6413, 'loss/train': 1.6988977789878845} 01/28/2022 21:06:15 - INFO - codeparrot_training - Step 6414: {'lr': 0.000489639772434869, 'samples': 1231680, 'steps': 6414, 'loss/train': 1.9424616694450378} 01/28/2022 21:06:19 - INFO - codeparrot_training - Step 6415: {'lr': 0.0004896351103578669, 'samples': 1231872, 'steps': 6415, 'loss/train': 1.3140671253204346} 01/28/2022 21:06:24 - INFO - codeparrot_training - Step 6416: {'lr': 0.0004896304472543439, 'samples': 1232064, 'steps': 6416, 'loss/train': 1.6505118012428284} 01/28/2022 21:06:29 - INFO - codeparrot_training - Step 6417: {'lr': 0.0004896257831243204, 'samples': 1232256, 'steps': 6417, 'loss/train': 1.278368979692459} 01/28/2022 21:06:33 - INFO - codeparrot_training - Step 6418: {'lr': 0.0004896211179678159, 'samples': 1232448, 'steps': 6418, 'loss/train': 1.3479240238666534} 01/28/2022 21:06:38 - INFO - codeparrot_training - Step 6419: {'lr': 0.0004896164517848508, 'samples': 1232640, 'steps': 6419, 'loss/train': 1.8846044540405273} 01/28/2022 21:06:42 - INFO - codeparrot_training - Step 6420: {'lr': 0.0004896117845754448, 'samples': 1232832, 'steps': 6420, 'loss/train': 2.373608708381653} 01/28/2022 21:06:46 - INFO - codeparrot_training - Step 6421: {'lr': 0.0004896071163396179, 'samples': 1233024, 'steps': 6421, 'loss/train': 1.9555744528770447} 01/28/2022 21:06:50 - INFO - codeparrot_training - Step 6422: {'lr': 0.0004896024470773904, 'samples': 1233216, 'steps': 6422, 'loss/train': 1.696311593055725} 01/28/2022 21:06:56 - INFO - codeparrot_training - Step 6423: {'lr': 0.000489597776788782, 'samples': 1233408, 'steps': 6423, 'loss/train': 1.6513014435768127} 01/28/2022 21:07:00 - INFO - codeparrot_training - Step 6424: {'lr': 0.0004895931054738128, 'samples': 1233600, 'steps': 6424, 'loss/train': 2.1440162658691406} 01/28/2022 21:07:05 - INFO - codeparrot_training - Step 6425: {'lr': 0.0004895884331325028, 'samples': 1233792, 'steps': 6425, 'loss/train': 4.334014892578125} 01/28/2022 21:07:09 - INFO - codeparrot_training - Step 6426: {'lr': 0.0004895837597648721, 'samples': 1233984, 'steps': 6426, 'loss/train': 1.438298910856247} 01/28/2022 21:07:13 - INFO - codeparrot_training - Step 6427: {'lr': 0.0004895790853709406, 'samples': 1234176, 'steps': 6427, 'loss/train': 1.6857474446296692} 01/28/2022 21:07:18 - INFO - codeparrot_training - Step 6428: {'lr': 0.0004895744099507284, 'samples': 1234368, 'steps': 6428, 'loss/train': 2.433416783809662} 01/28/2022 21:07:23 - INFO - codeparrot_training - Step 6429: {'lr': 0.0004895697335042555, 'samples': 1234560, 'steps': 6429, 'loss/train': 2.7074148058891296} 01/28/2022 21:07:27 - INFO - codeparrot_training - Step 6430: {'lr': 0.0004895650560315419, 'samples': 1234752, 'steps': 6430, 'loss/train': 1.5415785312652588} 01/28/2022 21:07:31 - INFO - codeparrot_training - Step 6431: {'lr': 0.0004895603775326077, 'samples': 1234944, 'steps': 6431, 'loss/train': 1.3148362934589386} 01/28/2022 21:07:35 - INFO - codeparrot_training - Step 6432: {'lr': 0.0004895556980074729, 'samples': 1235136, 'steps': 6432, 'loss/train': 2.0456424951553345} 01/28/2022 21:07:42 - INFO - codeparrot_training - Step 6433: {'lr': 0.0004895510174561576, 'samples': 1235328, 'steps': 6433, 'loss/train': 1.664986789226532} 01/28/2022 21:07:46 - INFO - codeparrot_training - Step 6434: {'lr': 0.0004895463358786818, 'samples': 1235520, 'steps': 6434, 'loss/train': 2.2226391434669495} 01/28/2022 21:07:50 - INFO - codeparrot_training - Step 6435: {'lr': 0.0004895416532750655, 'samples': 1235712, 'steps': 6435, 'loss/train': 2.2547507286071777} 01/28/2022 21:07:55 - INFO - codeparrot_training - Step 6436: {'lr': 0.0004895369696453289, 'samples': 1235904, 'steps': 6436, 'loss/train': 1.6217824816703796} 01/28/2022 21:08:00 - INFO - codeparrot_training - Step 6437: {'lr': 0.0004895322849894918, 'samples': 1236096, 'steps': 6437, 'loss/train': 2.4038591980934143} 01/28/2022 21:08:04 - INFO - codeparrot_training - Step 6438: {'lr': 0.0004895275993075747, 'samples': 1236288, 'steps': 6438, 'loss/train': 2.5704075694084167} 01/28/2022 21:08:08 - INFO - codeparrot_training - Step 6439: {'lr': 0.0004895229125995973, 'samples': 1236480, 'steps': 6439, 'loss/train': 2.180099666118622} 01/28/2022 21:08:13 - INFO - codeparrot_training - Step 6440: {'lr': 0.0004895182248655798, 'samples': 1236672, 'steps': 6440, 'loss/train': 1.685462772846222} 01/28/2022 21:08:17 - INFO - codeparrot_training - Step 6441: {'lr': 0.0004895135361055422, 'samples': 1236864, 'steps': 6441, 'loss/train': 1.3715640306472778} 01/28/2022 21:08:22 - INFO - codeparrot_training - Step 6442: {'lr': 0.0004895088463195049, 'samples': 1237056, 'steps': 6442, 'loss/train': 1.707685947418213} 01/28/2022 21:08:26 - INFO - codeparrot_training - Step 6443: {'lr': 0.0004895041555074875, 'samples': 1237248, 'steps': 6443, 'loss/train': 1.566550076007843} 01/28/2022 21:08:31 - INFO - codeparrot_training - Step 6444: {'lr': 0.0004894994636695105, 'samples': 1237440, 'steps': 6444, 'loss/train': 1.153410404920578} 01/28/2022 21:08:35 - INFO - codeparrot_training - Step 6445: {'lr': 0.0004894947708055938, 'samples': 1237632, 'steps': 6445, 'loss/train': 2.412983000278473} 01/28/2022 21:08:39 - INFO - codeparrot_training - Step 6446: {'lr': 0.0004894900769157576, 'samples': 1237824, 'steps': 6446, 'loss/train': 1.5000599026679993} 01/28/2022 21:08:43 - INFO - codeparrot_training - Step 6447: {'lr': 0.0004894853820000219, 'samples': 1238016, 'steps': 6447, 'loss/train': 1.8162690997123718} 01/28/2022 21:08:50 - INFO - codeparrot_training - Step 6448: {'lr': 0.000489480686058407, 'samples': 1238208, 'steps': 6448, 'loss/train': 1.8741429448127747} 01/28/2022 21:08:54 - INFO - codeparrot_training - Step 6449: {'lr': 0.0004894759890909326, 'samples': 1238400, 'steps': 6449, 'loss/train': 1.7325477004051208} 01/28/2022 21:08:58 - INFO - codeparrot_training - Step 6450: {'lr': 0.0004894712910976193, 'samples': 1238592, 'steps': 6450, 'loss/train': 2.1330193877220154} 01/28/2022 21:09:03 - INFO - codeparrot_training - Step 6451: {'lr': 0.000489466592078487, 'samples': 1238784, 'steps': 6451, 'loss/train': 1.9003021717071533} 01/28/2022 21:09:08 - INFO - codeparrot_training - Step 6452: {'lr': 0.0004894618920335558, 'samples': 1238976, 'steps': 6452, 'loss/train': 2.4576593041419983} 01/28/2022 21:09:12 - INFO - codeparrot_training - Step 6453: {'lr': 0.000489457190962846, 'samples': 1239168, 'steps': 6453, 'loss/train': 1.6473172903060913} 01/28/2022 21:09:16 - INFO - codeparrot_training - Step 6454: {'lr': 0.0004894524888663776, 'samples': 1239360, 'steps': 6454, 'loss/train': 1.2408634722232819} 01/28/2022 21:09:21 - INFO - codeparrot_training - Step 6455: {'lr': 0.0004894477857441707, 'samples': 1239552, 'steps': 6455, 'loss/train': 2.51364004611969} 01/28/2022 21:09:25 - INFO - codeparrot_training - Step 6456: {'lr': 0.0004894430815962456, 'samples': 1239744, 'steps': 6456, 'loss/train': 1.3375118672847748} 01/28/2022 21:09:30 - INFO - codeparrot_training - Step 6457: {'lr': 0.0004894383764226224, 'samples': 1239936, 'steps': 6457, 'loss/train': 1.9098318815231323} 01/28/2022 21:09:34 - INFO - codeparrot_training - Step 6458: {'lr': 0.0004894336702233212, 'samples': 1240128, 'steps': 6458, 'loss/train': 2.805087447166443} 01/28/2022 21:09:39 - INFO - codeparrot_training - Step 6459: {'lr': 0.0004894289629983621, 'samples': 1240320, 'steps': 6459, 'loss/train': 2.5303462743759155} 01/28/2022 21:09:43 - INFO - codeparrot_training - Step 6460: {'lr': 0.0004894242547477654, 'samples': 1240512, 'steps': 6460, 'loss/train': 1.8042476177215576} 01/28/2022 21:09:47 - INFO - codeparrot_training - Step 6461: {'lr': 0.0004894195454715512, 'samples': 1240704, 'steps': 6461, 'loss/train': 1.8446175456047058} 01/28/2022 21:09:52 - INFO - codeparrot_training - Step 6462: {'lr': 0.0004894148351697398, 'samples': 1240896, 'steps': 6462, 'loss/train': 1.9800497889518738} 01/28/2022 21:09:57 - INFO - codeparrot_training - Step 6463: {'lr': 0.0004894101238423512, 'samples': 1241088, 'steps': 6463, 'loss/train': 1.3808968663215637} 01/28/2022 21:10:01 - INFO - codeparrot_training - Step 6464: {'lr': 0.0004894054114894055, 'samples': 1241280, 'steps': 6464, 'loss/train': 1.497871220111847} 01/28/2022 21:10:05 - INFO - codeparrot_training - Step 6465: {'lr': 0.0004894006981109232, 'samples': 1241472, 'steps': 6465, 'loss/train': 1.0663918554782867} 01/28/2022 21:10:09 - INFO - codeparrot_training - Step 6466: {'lr': 0.0004893959837069243, 'samples': 1241664, 'steps': 6466, 'loss/train': 1.4283922612667084} 01/28/2022 21:10:16 - INFO - codeparrot_training - Step 6467: {'lr': 0.0004893912682774291, 'samples': 1241856, 'steps': 6467, 'loss/train': 1.7515836954116821} 01/28/2022 21:10:20 - INFO - codeparrot_training - Step 6468: {'lr': 0.0004893865518224576, 'samples': 1242048, 'steps': 6468, 'loss/train': 2.1855136156082153} 01/28/2022 21:10:24 - INFO - codeparrot_training - Step 6469: {'lr': 0.0004893818343420302, 'samples': 1242240, 'steps': 6469, 'loss/train': 4.659189820289612} 01/28/2022 21:10:28 - INFO - codeparrot_training - Step 6470: {'lr': 0.000489377115836167, 'samples': 1242432, 'steps': 6470, 'loss/train': 1.670910358428955} 01/28/2022 21:10:33 - INFO - codeparrot_training - Step 6471: {'lr': 0.0004893723963048882, 'samples': 1242624, 'steps': 6471, 'loss/train': 2.016356348991394} 01/28/2022 21:10:38 - INFO - codeparrot_training - Step 6472: {'lr': 0.0004893676757482142, 'samples': 1242816, 'steps': 6472, 'loss/train': 2.0768975615501404} 01/28/2022 21:10:42 - INFO - codeparrot_training - Step 6473: {'lr': 0.0004893629541661649, 'samples': 1243008, 'steps': 6473, 'loss/train': 1.4563744068145752} 01/28/2022 21:10:46 - INFO - codeparrot_training - Step 6474: {'lr': 0.0004893582315587608, 'samples': 1243200, 'steps': 6474, 'loss/train': 2.1458845138549805} 01/28/2022 21:10:51 - INFO - codeparrot_training - Step 6475: {'lr': 0.0004893535079260221, 'samples': 1243392, 'steps': 6475, 'loss/train': 2.37124764919281} 01/28/2022 21:10:55 - INFO - codeparrot_training - Step 6476: {'lr': 0.000489348783267969, 'samples': 1243584, 'steps': 6476, 'loss/train': 1.8999252319335938} 01/28/2022 21:11:01 - INFO - codeparrot_training - Step 6477: {'lr': 0.0004893440575846215, 'samples': 1243776, 'steps': 6477, 'loss/train': 7.4031500816345215} 01/28/2022 21:11:05 - INFO - codeparrot_training - Step 6478: {'lr': 0.0004893393308760002, 'samples': 1243968, 'steps': 6478, 'loss/train': 1.7448785305023193} 01/28/2022 21:11:10 - INFO - codeparrot_training - Step 6479: {'lr': 0.0004893346031421253, 'samples': 1244160, 'steps': 6479, 'loss/train': 0.5633837431669235} 01/28/2022 21:11:14 - INFO - codeparrot_training - Step 6480: {'lr': 0.0004893298743830168, 'samples': 1244352, 'steps': 6480, 'loss/train': 1.859718382358551} 01/28/2022 21:11:18 - INFO - codeparrot_training - Step 6481: {'lr': 0.0004893251445986952, 'samples': 1244544, 'steps': 6481, 'loss/train': 3.2311660051345825} 01/28/2022 21:11:23 - INFO - codeparrot_training - Step 6482: {'lr': 0.0004893204137891807, 'samples': 1244736, 'steps': 6482, 'loss/train': 2.061392605304718} 01/28/2022 21:11:28 - INFO - codeparrot_training - Step 6483: {'lr': 0.0004893156819544935, 'samples': 1244928, 'steps': 6483, 'loss/train': 2.8433758020401} 01/28/2022 21:11:32 - INFO - codeparrot_training - Step 6484: {'lr': 0.0004893109490946539, 'samples': 1245120, 'steps': 6484, 'loss/train': 1.4737059473991394} 01/28/2022 21:11:36 - INFO - codeparrot_training - Step 6485: {'lr': 0.0004893062152096821, 'samples': 1245312, 'steps': 6485, 'loss/train': 2.1034398078918457} 01/28/2022 21:11:40 - INFO - codeparrot_training - Step 6486: {'lr': 0.0004893014802995985, 'samples': 1245504, 'steps': 6486, 'loss/train': 3.210375666618347} 01/28/2022 21:11:46 - INFO - codeparrot_training - Step 6487: {'lr': 0.0004892967443644235, 'samples': 1245696, 'steps': 6487, 'loss/train': 1.667368769645691} 01/28/2022 21:11:50 - INFO - codeparrot_training - Step 6488: {'lr': 0.0004892920074041771, 'samples': 1245888, 'steps': 6488, 'loss/train': 1.6712939143180847} 01/28/2022 21:11:54 - INFO - codeparrot_training - Step 6489: {'lr': 0.0004892872694188797, 'samples': 1246080, 'steps': 6489, 'loss/train': 2.5397016406059265} 01/28/2022 21:11:58 - INFO - codeparrot_training - Step 6490: {'lr': 0.0004892825304085517, 'samples': 1246272, 'steps': 6490, 'loss/train': 1.8997333645820618} 01/28/2022 21:12:03 - INFO - codeparrot_training - Step 6491: {'lr': 0.0004892777903732133, 'samples': 1246464, 'steps': 6491, 'loss/train': 1.474893182516098} 01/28/2022 21:12:09 - INFO - codeparrot_training - Step 6492: {'lr': 0.0004892730493128848, 'samples': 1246656, 'steps': 6492, 'loss/train': 1.5429868698120117} 01/28/2022 21:12:13 - INFO - codeparrot_training - Step 6493: {'lr': 0.0004892683072275865, 'samples': 1246848, 'steps': 6493, 'loss/train': 1.594190776348114} 01/28/2022 21:12:17 - INFO - codeparrot_training - Step 6494: {'lr': 0.0004892635641173389, 'samples': 1247040, 'steps': 6494, 'loss/train': 1.7626469135284424} 01/28/2022 21:12:21 - INFO - codeparrot_training - Step 6495: {'lr': 0.0004892588199821619, 'samples': 1247232, 'steps': 6495, 'loss/train': 1.7269214987754822} 01/28/2022 21:12:26 - INFO - codeparrot_training - Step 6496: {'lr': 0.0004892540748220763, 'samples': 1247424, 'steps': 6496, 'loss/train': 2.0129286646842957} 01/28/2022 21:12:31 - INFO - codeparrot_training - Step 6497: {'lr': 0.0004892493286371022, 'samples': 1247616, 'steps': 6497, 'loss/train': 2.3667408227920532} 01/28/2022 21:12:35 - INFO - codeparrot_training - Step 6498: {'lr': 0.00048924458142726, 'samples': 1247808, 'steps': 6498, 'loss/train': 2.1639066338539124} 01/28/2022 21:12:40 - INFO - codeparrot_training - Step 6499: {'lr': 0.0004892398331925698, 'samples': 1248000, 'steps': 6499, 'loss/train': 1.9906225204467773} 01/28/2022 21:12:44 - INFO - codeparrot_training - Step 6500: {'lr': 0.0004892350839330522, 'samples': 1248192, 'steps': 6500, 'loss/train': 2.092500686645508} 01/28/2022 21:12:48 - INFO - codeparrot_training - Step 6501: {'lr': 0.0004892303336487275, 'samples': 1248384, 'steps': 6501, 'loss/train': 2.1602960228919983} 01/28/2022 21:12:54 - INFO - codeparrot_training - Step 6502: {'lr': 0.000489225582339616, 'samples': 1248576, 'steps': 6502, 'loss/train': 0.5679697394371033} 01/28/2022 21:12:58 - INFO - codeparrot_training - Step 6503: {'lr': 0.000489220830005738, 'samples': 1248768, 'steps': 6503, 'loss/train': 3.042763352394104} 01/28/2022 21:13:03 - INFO - codeparrot_training - Step 6504: {'lr': 0.0004892160766471141, 'samples': 1248960, 'steps': 6504, 'loss/train': 1.2213653326034546} 01/28/2022 21:13:07 - INFO - codeparrot_training - Step 6505: {'lr': 0.0004892113222637643, 'samples': 1249152, 'steps': 6505, 'loss/train': 1.5660558342933655} 01/28/2022 21:13:11 - INFO - codeparrot_training - Step 6506: {'lr': 0.0004892065668557093, 'samples': 1249344, 'steps': 6506, 'loss/train': 1.1215996742248535} 01/28/2022 21:13:16 - INFO - codeparrot_training - Step 6507: {'lr': 0.0004892018104229692, 'samples': 1249536, 'steps': 6507, 'loss/train': 1.92274671792984} 01/28/2022 21:13:20 - INFO - codeparrot_training - Step 6508: {'lr': 0.0004891970529655646, 'samples': 1249728, 'steps': 6508, 'loss/train': 1.891595721244812} 01/28/2022 21:13:25 - INFO - codeparrot_training - Step 6509: {'lr': 0.0004891922944835158, 'samples': 1249920, 'steps': 6509, 'loss/train': 1.8186573386192322} 01/28/2022 21:13:29 - INFO - codeparrot_training - Step 6510: {'lr': 0.000489187534976843, 'samples': 1250112, 'steps': 6510, 'loss/train': 1.8164511322975159} 01/28/2022 21:13:33 - INFO - codeparrot_training - Step 6511: {'lr': 0.0004891827744455668, 'samples': 1250304, 'steps': 6511, 'loss/train': 1.9322077631950378} 01/28/2022 21:13:38 - INFO - codeparrot_training - Step 6512: {'lr': 0.0004891780128897077, 'samples': 1250496, 'steps': 6512, 'loss/train': 1.3762456476688385} 01/28/2022 21:13:43 - INFO - codeparrot_training - Step 6513: {'lr': 0.0004891732503092858, 'samples': 1250688, 'steps': 6513, 'loss/train': 1.3664002418518066} 01/28/2022 21:13:47 - INFO - codeparrot_training - Step 6514: {'lr': 0.0004891684867043216, 'samples': 1250880, 'steps': 6514, 'loss/train': 1.8552961349487305} 01/28/2022 21:13:51 - INFO - codeparrot_training - Step 6515: {'lr': 0.0004891637220748356, 'samples': 1251072, 'steps': 6515, 'loss/train': 1.1033794283866882} 01/28/2022 21:13:55 - INFO - codeparrot_training - Step 6516: {'lr': 0.0004891589564208482, 'samples': 1251264, 'steps': 6516, 'loss/train': 1.5049710273742676} 01/28/2022 21:14:01 - INFO - codeparrot_training - Step 6517: {'lr': 0.0004891541897423798, 'samples': 1251456, 'steps': 6517, 'loss/train': 2.3742406368255615} 01/28/2022 21:14:05 - INFO - codeparrot_training - Step 6518: {'lr': 0.0004891494220394507, 'samples': 1251648, 'steps': 6518, 'loss/train': 2.3539223670959473} 01/28/2022 21:14:09 - INFO - codeparrot_training - Step 6519: {'lr': 0.0004891446533120815, 'samples': 1251840, 'steps': 6519, 'loss/train': 1.6611964702606201} 01/28/2022 21:14:14 - INFO - codeparrot_training - Step 6520: {'lr': 0.0004891398835602925, 'samples': 1252032, 'steps': 6520, 'loss/train': 1.671263337135315} 01/28/2022 21:14:20 - INFO - codeparrot_training - Step 6521: {'lr': 0.0004891351127841041, 'samples': 1252224, 'steps': 6521, 'loss/train': 1.6268618702888489} 01/28/2022 21:14:24 - INFO - codeparrot_training - Step 6522: {'lr': 0.0004891303409835369, 'samples': 1252416, 'steps': 6522, 'loss/train': 1.8079538941383362} 01/28/2022 21:14:28 - INFO - codeparrot_training - Step 6523: {'lr': 0.0004891255681586113, 'samples': 1252608, 'steps': 6523, 'loss/train': 2.208880841732025} 01/28/2022 21:14:33 - INFO - codeparrot_training - Step 6524: {'lr': 0.0004891207943093476, 'samples': 1252800, 'steps': 6524, 'loss/train': 2.2086008191108704} 01/28/2022 21:14:37 - INFO - codeparrot_training - Step 6525: {'lr': 0.0004891160194357663, 'samples': 1252992, 'steps': 6525, 'loss/train': 1.8936476111412048} 01/28/2022 21:14:41 - INFO - codeparrot_training - Step 6526: {'lr': 0.0004891112435378881, 'samples': 1253184, 'steps': 6526, 'loss/train': 3.8618277311325073} 01/28/2022 21:14:47 - INFO - codeparrot_training - Step 6527: {'lr': 0.0004891064666157331, 'samples': 1253376, 'steps': 6527, 'loss/train': 1.1263669431209564} 01/28/2022 21:14:51 - INFO - codeparrot_training - Step 6528: {'lr': 0.0004891016886693219, 'samples': 1253568, 'steps': 6528, 'loss/train': 1.6819905638694763} 01/28/2022 21:14:55 - INFO - codeparrot_training - Step 6529: {'lr': 0.0004890969096986751, 'samples': 1253760, 'steps': 6529, 'loss/train': 1.2489294111728668} 01/28/2022 21:15:00 - INFO - codeparrot_training - Step 6530: {'lr': 0.000489092129703813, 'samples': 1253952, 'steps': 6530, 'loss/train': 2.2616883516311646} 01/28/2022 21:15:04 - INFO - codeparrot_training - Step 6531: {'lr': 0.0004890873486847561, 'samples': 1254144, 'steps': 6531, 'loss/train': 1.372316300868988} 01/28/2022 21:15:09 - INFO - codeparrot_training - Step 6532: {'lr': 0.000489082566641525, 'samples': 1254336, 'steps': 6532, 'loss/train': 2.125433921813965} 01/28/2022 21:15:14 - INFO - codeparrot_training - Step 6533: {'lr': 0.00048907778357414, 'samples': 1254528, 'steps': 6533, 'loss/train': 2.3855981826782227} 01/28/2022 21:15:18 - INFO - codeparrot_training - Step 6534: {'lr': 0.0004890729994826218, 'samples': 1254720, 'steps': 6534, 'loss/train': 1.51729416847229} 01/28/2022 21:15:22 - INFO - codeparrot_training - Step 6535: {'lr': 0.0004890682143669908, 'samples': 1254912, 'steps': 6535, 'loss/train': 2.895258128643036} 01/28/2022 21:15:28 - INFO - codeparrot_training - Step 6536: {'lr': 0.0004890634282272673, 'samples': 1255104, 'steps': 6536, 'loss/train': 1.133732557296753} 01/28/2022 21:15:32 - INFO - codeparrot_training - Step 6537: {'lr': 0.0004890586410634722, 'samples': 1255296, 'steps': 6537, 'loss/train': 1.9406594038009644} 01/28/2022 21:15:37 - INFO - codeparrot_training - Step 6538: {'lr': 0.0004890538528756256, 'samples': 1255488, 'steps': 6538, 'loss/train': 2.4951800107955933} 01/28/2022 21:15:41 - INFO - codeparrot_training - Step 6539: {'lr': 0.0004890490636637484, 'samples': 1255680, 'steps': 6539, 'loss/train': 1.3532183468341827} 01/28/2022 21:15:45 - INFO - codeparrot_training - Step 6540: {'lr': 0.0004890442734278608, 'samples': 1255872, 'steps': 6540, 'loss/train': 1.7075867056846619} 01/28/2022 21:15:51 - INFO - codeparrot_training - Step 6541: {'lr': 0.0004890394821679834, 'samples': 1256064, 'steps': 6541, 'loss/train': 2.169894576072693} 01/28/2022 21:15:55 - INFO - codeparrot_training - Step 6542: {'lr': 0.0004890346898841369, 'samples': 1256256, 'steps': 6542, 'loss/train': 2.2676514387130737} 01/28/2022 21:15:59 - INFO - codeparrot_training - Step 6543: {'lr': 0.0004890298965763416, 'samples': 1256448, 'steps': 6543, 'loss/train': 1.5254833102226257} 01/28/2022 21:16:03 - INFO - codeparrot_training - Step 6544: {'lr': 0.0004890251022446181, 'samples': 1256640, 'steps': 6544, 'loss/train': 2.2732629776000977} 01/28/2022 21:16:08 - INFO - codeparrot_training - Step 6545: {'lr': 0.000489020306888987, 'samples': 1256832, 'steps': 6545, 'loss/train': 2.261505603790283} 01/28/2022 21:16:12 - INFO - codeparrot_training - Step 6546: {'lr': 0.0004890155105094688, 'samples': 1257024, 'steps': 6546, 'loss/train': 1.0000962316989899} 01/28/2022 21:16:18 - INFO - codeparrot_training - Step 6547: {'lr': 0.0004890107131060841, 'samples': 1257216, 'steps': 6547, 'loss/train': 2.0763813257217407} 01/28/2022 21:16:22 - INFO - codeparrot_training - Step 6548: {'lr': 0.0004890059146788532, 'samples': 1257408, 'steps': 6548, 'loss/train': 1.0535034835338593} 01/28/2022 21:16:26 - INFO - codeparrot_training - Step 6549: {'lr': 0.000489001115227797, 'samples': 1257600, 'steps': 6549, 'loss/train': 2.3296603560447693} 01/28/2022 21:16:31 - INFO - codeparrot_training - Step 6550: {'lr': 0.000488996314752936, 'samples': 1257792, 'steps': 6550, 'loss/train': 1.32500758767128} 01/28/2022 21:16:35 - INFO - codeparrot_training - Step 6551: {'lr': 0.0004889915132542906, 'samples': 1257984, 'steps': 6551, 'loss/train': 2.003603160381317} 01/28/2022 21:16:40 - INFO - codeparrot_training - Step 6552: {'lr': 0.0004889867107318814, 'samples': 1258176, 'steps': 6552, 'loss/train': 1.5086016654968262} 01/28/2022 21:16:44 - INFO - codeparrot_training - Step 6553: {'lr': 0.0004889819071857291, 'samples': 1258368, 'steps': 6553, 'loss/train': 1.7823931574821472} 01/28/2022 21:16:48 - INFO - codeparrot_training - Step 6554: {'lr': 0.0004889771026158541, 'samples': 1258560, 'steps': 6554, 'loss/train': 1.8753216862678528} 01/28/2022 21:16:53 - INFO - codeparrot_training - Step 6555: {'lr': 0.0004889722970222772, 'samples': 1258752, 'steps': 6555, 'loss/train': 1.762839138507843} 01/28/2022 21:16:57 - INFO - codeparrot_training - Step 6556: {'lr': 0.0004889674904050188, 'samples': 1258944, 'steps': 6556, 'loss/train': 2.2769240140914917} 01/28/2022 21:17:02 - INFO - codeparrot_training - Step 6557: {'lr': 0.0004889626827640994, 'samples': 1259136, 'steps': 6557, 'loss/train': 1.8732633590698242} 01/28/2022 21:17:07 - INFO - codeparrot_training - Step 6558: {'lr': 0.00048895787409954, 'samples': 1259328, 'steps': 6558, 'loss/train': 0.9132753610610962} 01/28/2022 21:17:11 - INFO - codeparrot_training - Step 6559: {'lr': 0.0004889530644113608, 'samples': 1259520, 'steps': 6559, 'loss/train': 2.2243024706840515} 01/28/2022 21:17:15 - INFO - codeparrot_training - Step 6560: {'lr': 0.0004889482536995825, 'samples': 1259712, 'steps': 6560, 'loss/train': 2.238119065761566} 01/28/2022 21:17:22 - INFO - codeparrot_training - Step 6561: {'lr': 0.0004889434419642259, 'samples': 1259904, 'steps': 6561, 'loss/train': 2.2367767095565796} 01/28/2022 21:17:26 - INFO - codeparrot_training - Step 6562: {'lr': 0.0004889386292053114, 'samples': 1260096, 'steps': 6562, 'loss/train': 0.6049950867891312} 01/28/2022 21:17:30 - INFO - codeparrot_training - Step 6563: {'lr': 0.0004889338154228596, 'samples': 1260288, 'steps': 6563, 'loss/train': 2.272634983062744} 01/28/2022 21:17:35 - INFO - codeparrot_training - Step 6564: {'lr': 0.0004889290006168913, 'samples': 1260480, 'steps': 6564, 'loss/train': 1.8057577013969421} 01/28/2022 21:17:39 - INFO - codeparrot_training - Step 6565: {'lr': 0.0004889241847874271, 'samples': 1260672, 'steps': 6565, 'loss/train': 0.5818013995885849} 01/28/2022 21:17:44 - INFO - codeparrot_training - Step 6566: {'lr': 0.0004889193679344874, 'samples': 1260864, 'steps': 6566, 'loss/train': 2.21922504901886} 01/28/2022 21:17:48 - INFO - codeparrot_training - Step 6567: {'lr': 0.0004889145500580932, 'samples': 1261056, 'steps': 6567, 'loss/train': 1.211152195930481} 01/28/2022 21:17:53 - INFO - codeparrot_training - Step 6568: {'lr': 0.0004889097311582647, 'samples': 1261248, 'steps': 6568, 'loss/train': 1.899402916431427} 01/28/2022 21:17:57 - INFO - codeparrot_training - Step 6569: {'lr': 0.000488904911235023, 'samples': 1261440, 'steps': 6569, 'loss/train': 1.8812153935432434} 01/28/2022 21:18:01 - INFO - codeparrot_training - Step 6570: {'lr': 0.0004889000902883883, 'samples': 1261632, 'steps': 6570, 'loss/train': 1.2143139839172363} 01/28/2022 21:18:06 - INFO - codeparrot_training - Step 6571: {'lr': 0.0004888952683183816, 'samples': 1261824, 'steps': 6571, 'loss/train': 1.8461408615112305} 01/28/2022 21:18:11 - INFO - codeparrot_training - Step 6572: {'lr': 0.0004888904453250233, 'samples': 1262016, 'steps': 6572, 'loss/train': 1.8151039481163025} 01/28/2022 21:18:15 - INFO - codeparrot_training - Step 6573: {'lr': 0.0004888856213083343, 'samples': 1262208, 'steps': 6573, 'loss/train': 1.8081884980201721} 01/28/2022 21:18:19 - INFO - codeparrot_training - Step 6574: {'lr': 0.0004888807962683353, 'samples': 1262400, 'steps': 6574, 'loss/train': 1.7927019596099854} 01/28/2022 21:18:23 - INFO - codeparrot_training - Step 6575: {'lr': 0.0004888759702050466, 'samples': 1262592, 'steps': 6575, 'loss/train': 1.5738975405693054} 01/28/2022 21:18:29 - INFO - codeparrot_training - Step 6576: {'lr': 0.0004888711431184892, 'samples': 1262784, 'steps': 6576, 'loss/train': 1.660001277923584} 01/28/2022 21:18:33 - INFO - codeparrot_training - Step 6577: {'lr': 0.0004888663150086835, 'samples': 1262976, 'steps': 6577, 'loss/train': 1.9444935321807861} 01/28/2022 21:18:37 - INFO - codeparrot_training - Step 6578: {'lr': 0.0004888614858756505, 'samples': 1263168, 'steps': 6578, 'loss/train': 1.554964005947113} 01/28/2022 21:18:41 - INFO - codeparrot_training - Step 6579: {'lr': 0.0004888566557194107, 'samples': 1263360, 'steps': 6579, 'loss/train': 1.963272750377655} 01/28/2022 21:18:46 - INFO - codeparrot_training - Step 6580: {'lr': 0.0004888518245399849, 'samples': 1263552, 'steps': 6580, 'loss/train': 2.059575855731964} 01/28/2022 21:18:52 - INFO - codeparrot_training - Step 6581: {'lr': 0.0004888469923373937, 'samples': 1263744, 'steps': 6581, 'loss/train': 1.629931926727295} 01/28/2022 21:18:56 - INFO - codeparrot_training - Step 6582: {'lr': 0.0004888421591116578, 'samples': 1263936, 'steps': 6582, 'loss/train': 2.666151702404022} 01/28/2022 21:19:00 - INFO - codeparrot_training - Step 6583: {'lr': 0.000488837324862798, 'samples': 1264128, 'steps': 6583, 'loss/train': 0.7870572209358215} 01/28/2022 21:19:05 - INFO - codeparrot_training - Step 6584: {'lr': 0.0004888324895908349, 'samples': 1264320, 'steps': 6584, 'loss/train': 1.2430365085601807} 01/28/2022 21:19:09 - INFO - codeparrot_training - Step 6585: {'lr': 0.0004888276532957892, 'samples': 1264512, 'steps': 6585, 'loss/train': 1.7094933986663818} 01/28/2022 21:19:14 - INFO - codeparrot_training - Step 6586: {'lr': 0.0004888228159776818, 'samples': 1264704, 'steps': 6586, 'loss/train': 2.6233518719673157} 01/28/2022 21:19:18 - INFO - codeparrot_training - Step 6587: {'lr': 0.0004888179776365331, 'samples': 1264896, 'steps': 6587, 'loss/train': 1.9971622824668884} 01/28/2022 21:19:23 - INFO - codeparrot_training - Step 6588: {'lr': 0.0004888131382723641, 'samples': 1265088, 'steps': 6588, 'loss/train': 1.883696973323822} 01/28/2022 21:19:27 - INFO - codeparrot_training - Step 6589: {'lr': 0.0004888082978851954, 'samples': 1265280, 'steps': 6589, 'loss/train': 1.7943981885910034} 01/28/2022 21:19:31 - INFO - codeparrot_training - Step 6590: {'lr': 0.000488803456475048, 'samples': 1265472, 'steps': 6590, 'loss/train': 2.7565237283706665} 01/28/2022 21:19:37 - INFO - codeparrot_training - Step 6591: {'lr': 0.0004887986140419422, 'samples': 1265664, 'steps': 6591, 'loss/train': 1.5739370584487915} 01/28/2022 21:19:41 - INFO - codeparrot_training - Step 6592: {'lr': 0.000488793770585899, 'samples': 1265856, 'steps': 6592, 'loss/train': 2.2649009227752686} 01/28/2022 21:19:45 - INFO - codeparrot_training - Step 6593: {'lr': 0.0004887889261069392, 'samples': 1266048, 'steps': 6593, 'loss/train': 1.2406368255615234} 01/28/2022 21:19:50 - INFO - codeparrot_training - Step 6594: {'lr': 0.0004887840806050834, 'samples': 1266240, 'steps': 6594, 'loss/train': 2.0816518664360046} 01/28/2022 21:19:54 - INFO - codeparrot_training - Step 6595: {'lr': 0.0004887792340803524, 'samples': 1266432, 'steps': 6595, 'loss/train': 1.229609727859497} 01/28/2022 21:19:59 - INFO - codeparrot_training - Step 6596: {'lr': 0.000488774386532767, 'samples': 1266624, 'steps': 6596, 'loss/train': 1.4899669289588928} 01/28/2022 21:20:03 - INFO - codeparrot_training - Step 6597: {'lr': 0.0004887695379623481, 'samples': 1266816, 'steps': 6597, 'loss/train': 2.1980754733085632} 01/28/2022 21:20:08 - INFO - codeparrot_training - Step 6598: {'lr': 0.000488764688369116, 'samples': 1267008, 'steps': 6598, 'loss/train': 2.2445163130760193} 01/28/2022 21:20:12 - INFO - codeparrot_training - Step 6599: {'lr': 0.000488759837753092, 'samples': 1267200, 'steps': 6599, 'loss/train': 1.3134705126285553} 01/28/2022 21:20:17 - INFO - codeparrot_training - Step 6600: {'lr': 0.0004887549861142967, 'samples': 1267392, 'steps': 6600, 'loss/train': 1.8573946952819824} 01/28/2022 21:20:21 - INFO - codeparrot_training - Step 6601: {'lr': 0.0004887501334527507, 'samples': 1267584, 'steps': 6601, 'loss/train': 2.032783806324005} 01/28/2022 21:20:26 - INFO - codeparrot_training - Step 6602: {'lr': 0.000488745279768475, 'samples': 1267776, 'steps': 6602, 'loss/train': 1.7456510066986084} 01/28/2022 21:20:30 - INFO - codeparrot_training - Step 6603: {'lr': 0.0004887404250614904, 'samples': 1267968, 'steps': 6603, 'loss/train': 1.482249230146408} 01/28/2022 21:20:34 - INFO - codeparrot_training - Step 6604: {'lr': 0.0004887355693318176, 'samples': 1268160, 'steps': 6604, 'loss/train': 2.0676766633987427} 01/28/2022 21:20:38 - INFO - codeparrot_training - Step 6605: {'lr': 0.0004887307125794775, 'samples': 1268352, 'steps': 6605, 'loss/train': 2.2857829928398132} 01/28/2022 21:20:45 - INFO - codeparrot_training - Step 6606: {'lr': 0.0004887258548044907, 'samples': 1268544, 'steps': 6606, 'loss/train': 1.4984574615955353} 01/28/2022 21:20:49 - INFO - codeparrot_training - Step 6607: {'lr': 0.0004887209960068782, 'samples': 1268736, 'steps': 6607, 'loss/train': 1.637018859386444} 01/28/2022 21:20:53 - INFO - codeparrot_training - Step 6608: {'lr': 0.0004887161361866607, 'samples': 1268928, 'steps': 6608, 'loss/train': 1.9249185919761658} 01/28/2022 21:20:58 - INFO - codeparrot_training - Step 6609: {'lr': 0.0004887112753438592, 'samples': 1269120, 'steps': 6609, 'loss/train': 0.15247907862067223} 01/28/2022 21:21:03 - INFO - codeparrot_training - Step 6610: {'lr': 0.0004887064134784943, 'samples': 1269312, 'steps': 6610, 'loss/train': 1.7174143195152283} 01/28/2022 21:21:07 - INFO - codeparrot_training - Step 6611: {'lr': 0.0004887015505905869, 'samples': 1269504, 'steps': 6611, 'loss/train': 1.9348155856132507} 01/28/2022 21:21:11 - INFO - codeparrot_training - Step 6612: {'lr': 0.0004886966866801579, 'samples': 1269696, 'steps': 6612, 'loss/train': 2.022350013256073} 01/28/2022 21:21:15 - INFO - codeparrot_training - Step 6613: {'lr': 0.0004886918217472281, 'samples': 1269888, 'steps': 6613, 'loss/train': 2.1166135668754578} 01/28/2022 21:21:20 - INFO - codeparrot_training - Step 6614: {'lr': 0.0004886869557918183, 'samples': 1270080, 'steps': 6614, 'loss/train': 2.303636848926544} 01/28/2022 21:21:25 - INFO - codeparrot_training - Step 6615: {'lr': 0.0004886820888139494, 'samples': 1270272, 'steps': 6615, 'loss/train': 1.8889356851577759} 01/28/2022 21:21:29 - INFO - codeparrot_training - Step 6616: {'lr': 0.0004886772208136422, 'samples': 1270464, 'steps': 6616, 'loss/train': 1.3051186501979828} 01/28/2022 21:21:33 - INFO - codeparrot_training - Step 6617: {'lr': 0.0004886723517909176, 'samples': 1270656, 'steps': 6617, 'loss/train': 1.8547219634056091} 01/28/2022 21:21:38 - INFO - codeparrot_training - Step 6618: {'lr': 0.0004886674817457964, 'samples': 1270848, 'steps': 6618, 'loss/train': 1.6430697441101074} 01/28/2022 21:21:42 - INFO - codeparrot_training - Step 6619: {'lr': 0.0004886626106782995, 'samples': 1271040, 'steps': 6619, 'loss/train': 1.3391913771629333} 01/28/2022 21:21:48 - INFO - codeparrot_training - Step 6620: {'lr': 0.0004886577385884478, 'samples': 1271232, 'steps': 6620, 'loss/train': 2.2737228870391846} 01/28/2022 21:21:52 - INFO - codeparrot_training - Step 6621: {'lr': 0.0004886528654762621, 'samples': 1271424, 'steps': 6621, 'loss/train': 1.7816528677940369} 01/28/2022 21:21:57 - INFO - codeparrot_training - Step 6622: {'lr': 0.0004886479913417633, 'samples': 1271616, 'steps': 6622, 'loss/train': 1.0920462012290955} 01/28/2022 21:22:01 - INFO - codeparrot_training - Step 6623: {'lr': 0.0004886431161849722, 'samples': 1271808, 'steps': 6623, 'loss/train': 1.7292999029159546} 01/28/2022 21:22:05 - INFO - codeparrot_training - Step 6624: {'lr': 0.0004886382400059099, 'samples': 1272000, 'steps': 6624, 'loss/train': 1.957766354084015} 01/28/2022 21:22:10 - INFO - codeparrot_training - Step 6625: {'lr': 0.0004886333628045972, 'samples': 1272192, 'steps': 6625, 'loss/train': 2.1785743832588196} 01/28/2022 21:22:15 - INFO - codeparrot_training - Step 6626: {'lr': 0.0004886284845810548, 'samples': 1272384, 'steps': 6626, 'loss/train': 1.9044150710105896} 01/28/2022 21:22:19 - INFO - codeparrot_training - Step 6627: {'lr': 0.0004886236053353038, 'samples': 1272576, 'steps': 6627, 'loss/train': 1.7090048789978027} 01/28/2022 21:22:23 - INFO - codeparrot_training - Step 6628: {'lr': 0.000488618725067365, 'samples': 1272768, 'steps': 6628, 'loss/train': 1.92683744430542} 01/28/2022 21:22:27 - INFO - codeparrot_training - Step 6629: {'lr': 0.0004886138437772594, 'samples': 1272960, 'steps': 6629, 'loss/train': 1.3896322846412659} 01/28/2022 21:22:33 - INFO - codeparrot_training - Step 6630: {'lr': 0.0004886089614650078, 'samples': 1273152, 'steps': 6630, 'loss/train': 1.845430076122284} 01/28/2022 21:22:37 - INFO - codeparrot_training - Step 6631: {'lr': 0.0004886040781306313, 'samples': 1273344, 'steps': 6631, 'loss/train': 0.772089421749115} 01/28/2022 21:22:41 - INFO - codeparrot_training - Step 6632: {'lr': 0.0004885991937741506, 'samples': 1273536, 'steps': 6632, 'loss/train': 2.0195335149765015} 01/28/2022 21:22:46 - INFO - codeparrot_training - Step 6633: {'lr': 0.0004885943083955868, 'samples': 1273728, 'steps': 6633, 'loss/train': 1.1966529786586761} 01/28/2022 21:22:50 - INFO - codeparrot_training - Step 6634: {'lr': 0.0004885894219949607, 'samples': 1273920, 'steps': 6634, 'loss/train': 1.3025528490543365} 01/28/2022 21:22:56 - INFO - codeparrot_training - Step 6635: {'lr': 0.0004885845345722932, 'samples': 1274112, 'steps': 6635, 'loss/train': 2.2146449089050293} 01/28/2022 21:23:00 - INFO - codeparrot_training - Step 6636: {'lr': 0.0004885796461276055, 'samples': 1274304, 'steps': 6636, 'loss/train': 0.347844198346138} 01/28/2022 21:23:05 - INFO - codeparrot_training - Step 6637: {'lr': 0.0004885747566609182, 'samples': 1274496, 'steps': 6637, 'loss/train': 1.685600996017456} 01/28/2022 21:23:09 - INFO - codeparrot_training - Step 6638: {'lr': 0.0004885698661722524, 'samples': 1274688, 'steps': 6638, 'loss/train': 1.8078467845916748} 01/28/2022 21:23:13 - INFO - codeparrot_training - Step 6639: {'lr': 0.0004885649746616291, 'samples': 1274880, 'steps': 6639, 'loss/train': 1.0664794743061066} 01/28/2022 21:23:18 - INFO - codeparrot_training - Step 6640: {'lr': 0.0004885600821290692, 'samples': 1275072, 'steps': 6640, 'loss/train': 1.8018561601638794} 01/28/2022 21:23:23 - INFO - codeparrot_training - Step 6641: {'lr': 0.0004885551885745937, 'samples': 1275264, 'steps': 6641, 'loss/train': 2.1246684193611145} 01/28/2022 21:23:27 - INFO - codeparrot_training - Step 6642: {'lr': 0.0004885502939982235, 'samples': 1275456, 'steps': 6642, 'loss/train': 2.9235599040985107} 01/28/2022 21:23:31 - INFO - codeparrot_training - Step 6643: {'lr': 0.0004885453983999795, 'samples': 1275648, 'steps': 6643, 'loss/train': 2.157860577106476} 01/28/2022 21:23:35 - INFO - codeparrot_training - Step 6644: {'lr': 0.0004885405017798828, 'samples': 1275840, 'steps': 6644, 'loss/train': 1.5508521795272827} 01/28/2022 21:23:41 - INFO - codeparrot_training - Step 6645: {'lr': 0.0004885356041379544, 'samples': 1276032, 'steps': 6645, 'loss/train': 1.9251542687416077} 01/28/2022 21:23:46 - INFO - codeparrot_training - Step 6646: {'lr': 0.0004885307054742151, 'samples': 1276224, 'steps': 6646, 'loss/train': 1.8930798768997192} 01/28/2022 21:23:50 - INFO - codeparrot_training - Step 6647: {'lr': 0.0004885258057886861, 'samples': 1276416, 'steps': 6647, 'loss/train': 1.7503507733345032} 01/28/2022 21:23:54 - INFO - codeparrot_training - Step 6648: {'lr': 0.0004885209050813882, 'samples': 1276608, 'steps': 6648, 'loss/train': 2.5383989810943604} 01/28/2022 21:23:58 - INFO - codeparrot_training - Step 6649: {'lr': 0.0004885160033523426, 'samples': 1276800, 'steps': 6649, 'loss/train': 2.841685116291046} 01/28/2022 21:24:04 - INFO - codeparrot_training - Step 6650: {'lr': 0.0004885111006015701, 'samples': 1276992, 'steps': 6650, 'loss/train': 1.1134015023708344} 01/28/2022 21:24:08 - INFO - codeparrot_training - Step 6651: {'lr': 0.0004885061968290919, 'samples': 1277184, 'steps': 6651, 'loss/train': 2.3148252367973328} 01/28/2022 21:24:12 - INFO - codeparrot_training - Step 6652: {'lr': 0.0004885012920349287, 'samples': 1277376, 'steps': 6652, 'loss/train': 1.6730713248252869} 01/28/2022 21:24:16 - INFO - codeparrot_training - Step 6653: {'lr': 0.0004884963862191018, 'samples': 1277568, 'steps': 6653, 'loss/train': 1.455674260854721} 01/28/2022 21:24:21 - INFO - codeparrot_training - Step 6654: {'lr': 0.0004884914793816321, 'samples': 1277760, 'steps': 6654, 'loss/train': 1.2774618566036224} 01/28/2022 21:24:26 - INFO - codeparrot_training - Step 6655: {'lr': 0.0004884865715225407, 'samples': 1277952, 'steps': 6655, 'loss/train': 2.7267597913742065} 01/28/2022 21:24:30 - INFO - codeparrot_training - Step 6656: {'lr': 0.0004884816626418484, 'samples': 1278144, 'steps': 6656, 'loss/train': 1.7574480772018433} 01/28/2022 21:24:34 - INFO - codeparrot_training - Step 6657: {'lr': 0.0004884767527395765, 'samples': 1278336, 'steps': 6657, 'loss/train': 1.7485958933830261} 01/28/2022 21:24:39 - INFO - codeparrot_training - Step 6658: {'lr': 0.0004884718418157459, 'samples': 1278528, 'steps': 6658, 'loss/train': 1.4027094841003418} 01/28/2022 21:24:43 - INFO - codeparrot_training - Step 6659: {'lr': 0.0004884669298703775, 'samples': 1278720, 'steps': 6659, 'loss/train': 2.3988624215126038} 01/28/2022 21:24:49 - INFO - codeparrot_training - Step 6660: {'lr': 0.0004884620169034927, 'samples': 1278912, 'steps': 6660, 'loss/train': 2.2183797955513} 01/28/2022 21:24:53 - INFO - codeparrot_training - Step 6661: {'lr': 0.0004884571029151123, 'samples': 1279104, 'steps': 6661, 'loss/train': 2.5254576802253723} 01/28/2022 21:24:57 - INFO - codeparrot_training - Step 6662: {'lr': 0.0004884521879052573, 'samples': 1279296, 'steps': 6662, 'loss/train': 2.66202574968338} 01/28/2022 21:25:02 - INFO - codeparrot_training - Step 6663: {'lr': 0.000488447271873949, 'samples': 1279488, 'steps': 6663, 'loss/train': 2.150633454322815} 01/28/2022 21:25:06 - INFO - codeparrot_training - Step 6664: {'lr': 0.0004884423548212082, 'samples': 1279680, 'steps': 6664, 'loss/train': 2.2862150073051453} 01/28/2022 21:25:11 - INFO - codeparrot_training - Step 6665: {'lr': 0.000488437436747056, 'samples': 1279872, 'steps': 6665, 'loss/train': 1.8229948282241821} 01/28/2022 21:25:15 - INFO - codeparrot_training - Step 6666: {'lr': 0.0004884325176515137, 'samples': 1280064, 'steps': 6666, 'loss/train': 1.4301426708698273} 01/28/2022 21:25:20 - INFO - codeparrot_training - Step 6667: {'lr': 0.000488427597534602, 'samples': 1280256, 'steps': 6667, 'loss/train': 1.7490813732147217} 01/28/2022 21:25:24 - INFO - codeparrot_training - Step 6668: {'lr': 0.0004884226763963423, 'samples': 1280448, 'steps': 6668, 'loss/train': 1.4364216327667236} 01/28/2022 21:25:28 - INFO - codeparrot_training - Step 6669: {'lr': 0.0004884177542367556, 'samples': 1280640, 'steps': 6669, 'loss/train': 2.2078848481178284} 01/28/2022 21:25:34 - INFO - codeparrot_training - Step 6670: {'lr': 0.0004884128310558628, 'samples': 1280832, 'steps': 6670, 'loss/train': 0.6168211251497269} 01/28/2022 21:25:38 - INFO - codeparrot_training - Step 6671: {'lr': 0.0004884079068536853, 'samples': 1281024, 'steps': 6671, 'loss/train': 2.190905213356018} 01/28/2022 21:25:42 - INFO - codeparrot_training - Step 6672: {'lr': 0.000488402981630244, 'samples': 1281216, 'steps': 6672, 'loss/train': 2.338712453842163} 01/28/2022 21:25:46 - INFO - codeparrot_training - Step 6673: {'lr': 0.00048839805538556, 'samples': 1281408, 'steps': 6673, 'loss/train': 2.051582157611847} 01/28/2022 21:25:51 - INFO - codeparrot_training - Step 6674: {'lr': 0.0004883931281196544, 'samples': 1281600, 'steps': 6674, 'loss/train': 2.8532060980796814} 01/28/2022 21:25:57 - INFO - codeparrot_training - Step 6675: {'lr': 0.0004883881998325484, 'samples': 1281792, 'steps': 6675, 'loss/train': 1.433409422636032} 01/28/2022 21:26:01 - INFO - codeparrot_training - Step 6676: {'lr': 0.000488383270524263, 'samples': 1281984, 'steps': 6676, 'loss/train': 1.8494735956192017} 01/28/2022 21:26:06 - INFO - codeparrot_training - Step 6677: {'lr': 0.0004883783401948194, 'samples': 1282176, 'steps': 6677, 'loss/train': 1.568288505077362} 01/28/2022 21:26:10 - INFO - codeparrot_training - Step 6678: {'lr': 0.0004883734088442387, 'samples': 1282368, 'steps': 6678, 'loss/train': 1.918893277645111} 01/28/2022 21:26:14 - INFO - codeparrot_training - Step 6679: {'lr': 0.0004883684764725419, 'samples': 1282560, 'steps': 6679, 'loss/train': 1.542746901512146} 01/28/2022 21:26:19 - INFO - codeparrot_training - Step 6680: {'lr': 0.0004883635430797502, 'samples': 1282752, 'steps': 6680, 'loss/train': 1.973146140575409} 01/28/2022 21:26:24 - INFO - codeparrot_training - Step 6681: {'lr': 0.000488358608665885, 'samples': 1282944, 'steps': 6681, 'loss/train': 1.7671319246292114} 01/28/2022 21:26:28 - INFO - codeparrot_training - Step 6682: {'lr': 0.000488353673230967, 'samples': 1283136, 'steps': 6682, 'loss/train': 2.012208938598633} 01/28/2022 21:26:32 - INFO - codeparrot_training - Step 6683: {'lr': 0.0004883487367750177, 'samples': 1283328, 'steps': 6683, 'loss/train': 1.7733437418937683} 01/28/2022 21:26:36 - INFO - codeparrot_training - Step 6684: {'lr': 0.0004883437992980581, 'samples': 1283520, 'steps': 6684, 'loss/train': 0.5757096111774445} 01/28/2022 21:26:42 - INFO - codeparrot_training - Step 6685: {'lr': 0.0004883388608001093, 'samples': 1283712, 'steps': 6685, 'loss/train': 2.1613043546676636} 01/28/2022 21:26:46 - INFO - codeparrot_training - Step 6686: {'lr': 0.0004883339212811924, 'samples': 1283904, 'steps': 6686, 'loss/train': 1.4855190217494965} 01/28/2022 21:26:50 - INFO - codeparrot_training - Step 6687: {'lr': 0.0004883289807413288, 'samples': 1284096, 'steps': 6687, 'loss/train': 2.1371689438819885} 01/28/2022 21:26:54 - INFO - codeparrot_training - Step 6688: {'lr': 0.0004883240391805394, 'samples': 1284288, 'steps': 6688, 'loss/train': 2.3050947189331055} 01/28/2022 21:26:59 - INFO - codeparrot_training - Step 6689: {'lr': 0.0004883190965988455, 'samples': 1284480, 'steps': 6689, 'loss/train': 2.1470173001289368} 01/28/2022 21:27:05 - INFO - codeparrot_training - Step 6690: {'lr': 0.0004883141529962683, 'samples': 1284672, 'steps': 6690, 'loss/train': 2.180880904197693} 01/28/2022 21:27:09 - INFO - codeparrot_training - Step 6691: {'lr': 0.000488309208372829, 'samples': 1284864, 'steps': 6691, 'loss/train': 1.923158884048462} 01/28/2022 21:27:13 - INFO - codeparrot_training - Step 6692: {'lr': 0.0004883042627285488, 'samples': 1285056, 'steps': 6692, 'loss/train': 1.7016656398773193} 01/28/2022 21:27:18 - INFO - codeparrot_training - Step 6693: {'lr': 0.0004882993160634487, 'samples': 1285248, 'steps': 6693, 'loss/train': 1.9784882068634033} 01/28/2022 21:27:22 - INFO - codeparrot_training - Step 6694: {'lr': 0.0004882943683775499, 'samples': 1285440, 'steps': 6694, 'loss/train': 1.9727139472961426} 01/28/2022 21:27:28 - INFO - codeparrot_training - Step 6695: {'lr': 0.0004882894196708738, 'samples': 1285632, 'steps': 6695, 'loss/train': 2.8483616709709167} 01/28/2022 21:27:32 - INFO - codeparrot_training - Step 6696: {'lr': 0.0004882844699434415, 'samples': 1285824, 'steps': 6696, 'loss/train': 1.8943492770195007} 01/28/2022 21:27:36 - INFO - codeparrot_training - Step 6697: {'lr': 0.0004882795191952741, 'samples': 1286016, 'steps': 6697, 'loss/train': 1.640011489391327} 01/28/2022 21:27:40 - INFO - codeparrot_training - Step 6698: {'lr': 0.0004882745674263931, 'samples': 1286208, 'steps': 6698, 'loss/train': 1.9801511764526367} 01/28/2022 21:27:45 - INFO - codeparrot_training - Step 6699: {'lr': 0.00048826961463681936, 'samples': 1286400, 'steps': 6699, 'loss/train': 0.7194711863994598} 01/28/2022 21:27:49 - INFO - codeparrot_training - Step 6700: {'lr': 0.00048826466082657426, 'samples': 1286592, 'steps': 6700, 'loss/train': 3.672668695449829} 01/28/2022 21:27:54 - INFO - codeparrot_training - Step 6701: {'lr': 0.000488259705995679, 'samples': 1286784, 'steps': 6701, 'loss/train': 0.4663064181804657} 01/28/2022 21:27:58 - INFO - codeparrot_training - Step 6702: {'lr': 0.0004882547501441549, 'samples': 1286976, 'steps': 6702, 'loss/train': 1.6043667197227478} 01/28/2022 21:28:03 - INFO - codeparrot_training - Step 6703: {'lr': 0.000488249793272023, 'samples': 1287168, 'steps': 6703, 'loss/train': 2.1690012216567993} 01/28/2022 21:28:07 - INFO - codeparrot_training - Step 6704: {'lr': 0.0004882448353793048, 'samples': 1287360, 'steps': 6704, 'loss/train': 2.455866515636444} 01/28/2022 21:28:11 - INFO - codeparrot_training - Step 6705: {'lr': 0.0004882398764660212, 'samples': 1287552, 'steps': 6705, 'loss/train': 1.9972677826881409} 01/28/2022 21:28:17 - INFO - codeparrot_training - Step 6706: {'lr': 0.00048823491653219366, 'samples': 1287744, 'steps': 6706, 'loss/train': 0.6890159100294113} 01/28/2022 21:28:21 - INFO - codeparrot_training - Step 6707: {'lr': 0.00048822995557784343, 'samples': 1287936, 'steps': 6707, 'loss/train': 2.2469537258148193} 01/28/2022 21:28:26 - INFO - codeparrot_training - Step 6708: {'lr': 0.00048822499360299165, 'samples': 1288128, 'steps': 6708, 'loss/train': 0.8989681005477905} 01/28/2022 21:28:30 - INFO - codeparrot_training - Step 6709: {'lr': 0.00048822003060765973, 'samples': 1288320, 'steps': 6709, 'loss/train': 2.6885915994644165} 01/28/2022 21:28:34 - INFO - codeparrot_training - Step 6710: {'lr': 0.00048821506659186875, 'samples': 1288512, 'steps': 6710, 'loss/train': 2.395683467388153} 01/28/2022 21:28:40 - INFO - codeparrot_training - Step 6711: {'lr': 0.0004882101015556402, 'samples': 1288704, 'steps': 6711, 'loss/train': 1.4090709686279297} 01/28/2022 21:28:44 - INFO - codeparrot_training - Step 6712: {'lr': 0.00048820513549899507, 'samples': 1288896, 'steps': 6712, 'loss/train': 2.1325723528862} 01/28/2022 21:28:48 - INFO - codeparrot_training - Step 6713: {'lr': 0.00048820016842195487, 'samples': 1289088, 'steps': 6713, 'loss/train': 1.471300184726715} 01/28/2022 21:28:52 - INFO - codeparrot_training - Step 6714: {'lr': 0.0004881952003245408, 'samples': 1289280, 'steps': 6714, 'loss/train': 0.955351710319519} 01/28/2022 21:28:58 - INFO - codeparrot_training - Step 6715: {'lr': 0.00048819023120677405, 'samples': 1289472, 'steps': 6715, 'loss/train': 1.6608892679214478} 01/28/2022 21:29:02 - INFO - codeparrot_training - Step 6716: {'lr': 0.000488185261068676, 'samples': 1289664, 'steps': 6716, 'loss/train': 1.4514335989952087} 01/28/2022 21:29:06 - INFO - codeparrot_training - Step 6717: {'lr': 0.000488180289910268, 'samples': 1289856, 'steps': 6717, 'loss/train': 2.2617915272712708} 01/28/2022 21:29:11 - INFO - codeparrot_training - Step 6718: {'lr': 0.0004881753177315711, 'samples': 1290048, 'steps': 6718, 'loss/train': 0.9404345154762268} 01/28/2022 21:29:15 - INFO - codeparrot_training - Step 6719: {'lr': 0.0004881703445326069, 'samples': 1290240, 'steps': 6719, 'loss/train': 2.1270039081573486} 01/28/2022 21:29:21 - INFO - codeparrot_training - Step 6720: {'lr': 0.0004881653703133966, 'samples': 1290432, 'steps': 6720, 'loss/train': 1.8548927307128906} 01/28/2022 21:29:25 - INFO - codeparrot_training - Step 6721: {'lr': 0.00048816039507396135, 'samples': 1290624, 'steps': 6721, 'loss/train': 2.1796587109565735} 01/28/2022 21:29:29 - INFO - codeparrot_training - Step 6722: {'lr': 0.00048815541881432273, 'samples': 1290816, 'steps': 6722, 'loss/train': 1.7105414271354675} 01/28/2022 21:29:34 - INFO - codeparrot_training - Step 6723: {'lr': 0.00048815044153450185, 'samples': 1291008, 'steps': 6723, 'loss/train': 1.3269887566566467} 01/28/2022 21:29:38 - INFO - codeparrot_training - Step 6724: {'lr': 0.00048814546323452013, 'samples': 1291200, 'steps': 6724, 'loss/train': 2.352002263069153} 01/28/2022 21:29:43 - INFO - codeparrot_training - Step 6725: {'lr': 0.0004881404839143988, 'samples': 1291392, 'steps': 6725, 'loss/train': 1.424065500497818} 01/28/2022 21:29:47 - INFO - codeparrot_training - Step 6726: {'lr': 0.00048813550357415937, 'samples': 1291584, 'steps': 6726, 'loss/train': 2.076464831829071} 01/28/2022 21:29:52 - INFO - codeparrot_training - Step 6727: {'lr': 0.00048813052221382294, 'samples': 1291776, 'steps': 6727, 'loss/train': 0.6641915738582611} 01/28/2022 21:29:56 - INFO - codeparrot_training - Step 6728: {'lr': 0.000488125539833411, 'samples': 1291968, 'steps': 6728, 'loss/train': 1.626701295375824} 01/28/2022 21:30:00 - INFO - codeparrot_training - Step 6729: {'lr': 0.0004881205564329449, 'samples': 1292160, 'steps': 6729, 'loss/train': 2.446094334125519} 01/28/2022 21:30:05 - INFO - codeparrot_training - Step 6730: {'lr': 0.00048811557201244594, 'samples': 1292352, 'steps': 6730, 'loss/train': 0.9329860210418701} 01/28/2022 21:30:10 - INFO - codeparrot_training - Step 6731: {'lr': 0.0004881105865719355, 'samples': 1292544, 'steps': 6731, 'loss/train': 2.1296648383140564} 01/28/2022 21:30:14 - INFO - codeparrot_training - Step 6732: {'lr': 0.00048810560011143485, 'samples': 1292736, 'steps': 6732, 'loss/train': 1.935976803302765} 01/28/2022 21:30:18 - INFO - codeparrot_training - Step 6733: {'lr': 0.0004881006126309654, 'samples': 1292928, 'steps': 6733, 'loss/train': 1.770770788192749} 01/28/2022 21:30:22 - INFO - codeparrot_training - Step 6734: {'lr': 0.00048809562413054864, 'samples': 1293120, 'steps': 6734, 'loss/train': 1.6424941420555115} 01/28/2022 21:30:28 - INFO - codeparrot_training - Step 6735: {'lr': 0.00048809063461020575, 'samples': 1293312, 'steps': 6735, 'loss/train': 1.4489050805568695} 01/28/2022 21:30:33 - INFO - codeparrot_training - Step 6736: {'lr': 0.0004880856440699582, 'samples': 1293504, 'steps': 6736, 'loss/train': 1.963383972644806} 01/28/2022 21:30:37 - INFO - codeparrot_training - Step 6737: {'lr': 0.00048808065250982737, 'samples': 1293696, 'steps': 6737, 'loss/train': 2.348617136478424} 01/28/2022 21:30:41 - INFO - codeparrot_training - Step 6738: {'lr': 0.0004880756599298346, 'samples': 1293888, 'steps': 6738, 'loss/train': 3.039188861846924} 01/28/2022 21:30:46 - INFO - codeparrot_training - Step 6739: {'lr': 0.0004880706663300013, 'samples': 1294080, 'steps': 6739, 'loss/train': 0.8565664887428284} 01/28/2022 21:30:51 - INFO - codeparrot_training - Step 6740: {'lr': 0.0004880656717103489, 'samples': 1294272, 'steps': 6740, 'loss/train': 1.0802547633647919} 01/28/2022 21:30:55 - INFO - codeparrot_training - Step 6741: {'lr': 0.00048806067607089866, 'samples': 1294464, 'steps': 6741, 'loss/train': 1.4194985926151276} 01/28/2022 21:30:59 - INFO - codeparrot_training - Step 6742: {'lr': 0.00048805567941167215, 'samples': 1294656, 'steps': 6742, 'loss/train': 1.3969162702560425} 01/28/2022 21:31:04 - INFO - codeparrot_training - Step 6743: {'lr': 0.0004880506817326907, 'samples': 1294848, 'steps': 6743, 'loss/train': 0.9893994927406311} 01/28/2022 21:31:08 - INFO - codeparrot_training - Step 6744: {'lr': 0.0004880456830339757, 'samples': 1295040, 'steps': 6744, 'loss/train': 1.5091320276260376} 01/28/2022 21:31:13 - INFO - codeparrot_training - Step 6745: {'lr': 0.00048804068331554864, 'samples': 1295232, 'steps': 6745, 'loss/train': 2.090860426425934} 01/28/2022 21:31:17 - INFO - codeparrot_training - Step 6746: {'lr': 0.00048803568257743083, 'samples': 1295424, 'steps': 6746, 'loss/train': 1.5188294649124146} 01/28/2022 21:31:22 - INFO - codeparrot_training - Step 6747: {'lr': 0.00048803068081964375, 'samples': 1295616, 'steps': 6747, 'loss/train': 0.5683256238698959} 01/28/2022 21:31:26 - INFO - codeparrot_training - Step 6748: {'lr': 0.00048802567804220875, 'samples': 1295808, 'steps': 6748, 'loss/train': 0.2607727199792862} 01/28/2022 21:31:30 - INFO - codeparrot_training - Step 6749: {'lr': 0.0004880206742451474, 'samples': 1296000, 'steps': 6749, 'loss/train': 2.0900296568870544} 01/28/2022 21:31:37 - INFO - codeparrot_training - Step 6750: {'lr': 0.0004880156694284811, 'samples': 1296192, 'steps': 6750, 'loss/train': 2.3963592052459717} 01/28/2022 21:31:41 - INFO - codeparrot_training - Step 6751: {'lr': 0.00048801066359223117, 'samples': 1296384, 'steps': 6751, 'loss/train': 1.5917546153068542} 01/28/2022 21:31:45 - INFO - codeparrot_training - Step 6752: {'lr': 0.00048800565673641917, 'samples': 1296576, 'steps': 6752, 'loss/train': 1.8071814179420471} 01/28/2022 21:31:50 - INFO - codeparrot_training - Step 6753: {'lr': 0.00048800064886106654, 'samples': 1296768, 'steps': 6753, 'loss/train': 2.2348714470863342} 01/28/2022 21:31:54 - INFO - codeparrot_training - Step 6754: {'lr': 0.0004879956399661947, 'samples': 1296960, 'steps': 6754, 'loss/train': 1.8270573019981384} 01/28/2022 21:31:58 - INFO - codeparrot_training - Step 6755: {'lr': 0.000487990630051825, 'samples': 1297152, 'steps': 6755, 'loss/train': 1.8311267495155334} 01/28/2022 21:32:04 - INFO - codeparrot_training - Step 6756: {'lr': 0.00048798561911797913, 'samples': 1297344, 'steps': 6756, 'loss/train': 2.770773410797119} 01/28/2022 21:32:08 - INFO - codeparrot_training - Step 6757: {'lr': 0.0004879806071646784, 'samples': 1297536, 'steps': 6757, 'loss/train': 1.5359925627708435} 01/28/2022 21:32:12 - INFO - codeparrot_training - Step 6758: {'lr': 0.00048797559419194427, 'samples': 1297728, 'steps': 6758, 'loss/train': 2.2317867279052734} 01/28/2022 21:32:16 - INFO - codeparrot_training - Step 6759: {'lr': 0.00048797058019979837, 'samples': 1297920, 'steps': 6759, 'loss/train': 1.8125642538070679} 01/28/2022 21:32:21 - INFO - codeparrot_training - Step 6760: {'lr': 0.00048796556518826195, 'samples': 1298112, 'steps': 6760, 'loss/train': 1.2751059830188751} 01/28/2022 21:32:27 - INFO - codeparrot_training - Step 6761: {'lr': 0.00048796054915735664, 'samples': 1298304, 'steps': 6761, 'loss/train': 2.247613549232483} 01/28/2022 21:32:31 - INFO - codeparrot_training - Step 6762: {'lr': 0.00048795553210710397, 'samples': 1298496, 'steps': 6762, 'loss/train': 2.0586483478546143} 01/28/2022 21:32:35 - INFO - codeparrot_training - Step 6763: {'lr': 0.00048795051403752534, 'samples': 1298688, 'steps': 6763, 'loss/train': 1.686017632484436} 01/28/2022 21:32:40 - INFO - codeparrot_training - Step 6764: {'lr': 0.0004879454949486422, 'samples': 1298880, 'steps': 6764, 'loss/train': 1.3503889739513397} 01/28/2022 21:32:44 - INFO - codeparrot_training - Step 6765: {'lr': 0.00048794047484047615, 'samples': 1299072, 'steps': 6765, 'loss/train': 1.6658227443695068} 01/28/2022 21:32:49 - INFO - codeparrot_training - Step 6766: {'lr': 0.00048793545371304863, 'samples': 1299264, 'steps': 6766, 'loss/train': 2.0740079283714294} 01/28/2022 21:32:53 - INFO - codeparrot_training - Step 6767: {'lr': 0.0004879304315663813, 'samples': 1299456, 'steps': 6767, 'loss/train': 0.7558821737766266} 01/28/2022 21:32:58 - INFO - codeparrot_training - Step 6768: {'lr': 0.00048792540840049544, 'samples': 1299648, 'steps': 6768, 'loss/train': 0.716082975268364} 01/28/2022 21:33:02 - INFO - codeparrot_training - Step 6769: {'lr': 0.00048792038421541266, 'samples': 1299840, 'steps': 6769, 'loss/train': 2.2337486743927} 01/28/2022 21:33:06 - INFO - codeparrot_training - Step 6770: {'lr': 0.0004879153590111546, 'samples': 1300032, 'steps': 6770, 'loss/train': 1.8919431567192078} 01/28/2022 21:33:11 - INFO - codeparrot_training - Step 6771: {'lr': 0.0004879103327877426, 'samples': 1300224, 'steps': 6771, 'loss/train': 2.3284642696380615} 01/28/2022 21:33:16 - INFO - codeparrot_training - Step 6772: {'lr': 0.0004879053055451983, 'samples': 1300416, 'steps': 6772, 'loss/train': 2.739547312259674} 01/28/2022 21:33:20 - INFO - codeparrot_training - Step 6773: {'lr': 0.00048790027728354323, 'samples': 1300608, 'steps': 6773, 'loss/train': 2.05403333902359} 01/28/2022 21:33:24 - INFO - codeparrot_training - Step 6774: {'lr': 0.0004878952480027989, 'samples': 1300800, 'steps': 6774, 'loss/train': 1.1549271047115326} 01/28/2022 21:33:28 - INFO - codeparrot_training - Step 6775: {'lr': 0.0004878902177029869, 'samples': 1300992, 'steps': 6775, 'loss/train': 1.5935121774673462} 01/28/2022 21:33:34 - INFO - codeparrot_training - Step 6776: {'lr': 0.0004878851863841287, 'samples': 1301184, 'steps': 6776, 'loss/train': 2.490097224712372} 01/28/2022 21:33:38 - INFO - codeparrot_training - Step 6777: {'lr': 0.00048788015404624597, 'samples': 1301376, 'steps': 6777, 'loss/train': 0.7781098186969757} 01/28/2022 21:33:42 - INFO - codeparrot_training - Step 6778: {'lr': 0.0004878751206893601, 'samples': 1301568, 'steps': 6778, 'loss/train': 2.180437982082367} 01/28/2022 21:33:47 - INFO - codeparrot_training - Step 6779: {'lr': 0.0004878700863134928, 'samples': 1301760, 'steps': 6779, 'loss/train': 2.694803237915039} 01/28/2022 21:33:51 - INFO - codeparrot_training - Step 6780: {'lr': 0.00048786505091866564, 'samples': 1301952, 'steps': 6780, 'loss/train': 1.9280893206596375} 01/28/2022 21:33:57 - INFO - codeparrot_training - Step 6781: {'lr': 0.0004878600145049001, 'samples': 1302144, 'steps': 6781, 'loss/train': 2.2020376324653625} 01/28/2022 21:34:02 - INFO - codeparrot_training - Step 6782: {'lr': 0.0004878549770722177, 'samples': 1302336, 'steps': 6782, 'loss/train': 0.7752707898616791} 01/28/2022 21:34:06 - INFO - codeparrot_training - Step 6783: {'lr': 0.0004878499386206402, 'samples': 1302528, 'steps': 6783, 'loss/train': 1.4802357852458954} 01/28/2022 21:34:10 - INFO - codeparrot_training - Step 6784: {'lr': 0.000487844899150189, 'samples': 1302720, 'steps': 6784, 'loss/train': 2.2384554147720337} 01/28/2022 21:34:14 - INFO - codeparrot_training - Step 6785: {'lr': 0.0004878398586608859, 'samples': 1302912, 'steps': 6785, 'loss/train': 2.7584869265556335} 01/28/2022 21:34:19 - INFO - codeparrot_training - Step 6786: {'lr': 0.0004878348171527523, 'samples': 1303104, 'steps': 6786, 'loss/train': 2.7762449383735657} 01/28/2022 21:34:24 - INFO - codeparrot_training - Step 6787: {'lr': 0.0004878297746258099, 'samples': 1303296, 'steps': 6787, 'loss/train': 0.7519464790821075} 01/28/2022 21:34:28 - INFO - codeparrot_training - Step 6788: {'lr': 0.0004878247310800802, 'samples': 1303488, 'steps': 6788, 'loss/train': 2.0548827052116394} 01/28/2022 21:34:32 - INFO - codeparrot_training - Step 6789: {'lr': 0.0004878196865155849, 'samples': 1303680, 'steps': 6789, 'loss/train': 4.50877046585083} 01/28/2022 21:34:36 - INFO - codeparrot_training - Step 6790: {'lr': 0.0004878146409323456, 'samples': 1303872, 'steps': 6790, 'loss/train': 1.5731273889541626} 01/28/2022 21:34:42 - INFO - codeparrot_training - Step 6791: {'lr': 0.00048780959433038386, 'samples': 1304064, 'steps': 6791, 'loss/train': 2.2494235038757324} 01/28/2022 21:34:47 - INFO - codeparrot_training - Step 6792: {'lr': 0.00048780454670972127, 'samples': 1304256, 'steps': 6792, 'loss/train': 0.12440521270036697} 01/28/2022 21:34:51 - INFO - codeparrot_training - Step 6793: {'lr': 0.00048779949807037967, 'samples': 1304448, 'steps': 6793, 'loss/train': 1.7426710724830627} 01/28/2022 21:34:55 - INFO - codeparrot_training - Step 6794: {'lr': 0.0004877944484123804, 'samples': 1304640, 'steps': 6794, 'loss/train': 1.5688002705574036} 01/28/2022 21:34:59 - INFO - codeparrot_training - Step 6795: {'lr': 0.00048778939773574525, 'samples': 1304832, 'steps': 6795, 'loss/train': 1.0492188334465027} 01/28/2022 21:35:04 - INFO - codeparrot_training - Step 6796: {'lr': 0.0004877843460404959, 'samples': 1305024, 'steps': 6796, 'loss/train': 2.0260793566703796} 01/28/2022 21:35:10 - INFO - codeparrot_training - Step 6797: {'lr': 0.00048777929332665385, 'samples': 1305216, 'steps': 6797, 'loss/train': 2.5871633291244507} 01/28/2022 21:35:14 - INFO - codeparrot_training - Step 6798: {'lr': 0.00048777423959424083, 'samples': 1305408, 'steps': 6798, 'loss/train': 2.1253257393836975} 01/28/2022 21:35:19 - INFO - codeparrot_training - Step 6799: {'lr': 0.00048776918484327847, 'samples': 1305600, 'steps': 6799, 'loss/train': 1.714741587638855} 01/28/2022 21:35:23 - INFO - codeparrot_training - Step 6800: {'lr': 0.0004877641290737884, 'samples': 1305792, 'steps': 6800, 'loss/train': 1.4753307402133942} 01/28/2022 21:35:27 - INFO - codeparrot_training - Step 6801: {'lr': 0.0004877590722857923, 'samples': 1305984, 'steps': 6801, 'loss/train': 1.386944442987442} 01/28/2022 21:35:32 - INFO - codeparrot_training - Step 6802: {'lr': 0.00048775401447931187, 'samples': 1306176, 'steps': 6802, 'loss/train': 2.0277457237243652} 01/28/2022 21:35:37 - INFO - codeparrot_training - Step 6803: {'lr': 0.0004877489556543687, 'samples': 1306368, 'steps': 6803, 'loss/train': 1.683811604976654} 01/28/2022 21:35:41 - INFO - codeparrot_training - Step 6804: {'lr': 0.00048774389581098454, 'samples': 1306560, 'steps': 6804, 'loss/train': 1.6461777091026306} 01/28/2022 21:35:45 - INFO - codeparrot_training - Step 6805: {'lr': 0.00048773883494918096, 'samples': 1306752, 'steps': 6805, 'loss/train': 2.1824015378952026} 01/28/2022 21:35:49 - INFO - codeparrot_training - Step 6806: {'lr': 0.0004877337730689797, 'samples': 1306944, 'steps': 6806, 'loss/train': 1.8033354878425598} 01/28/2022 21:35:55 - INFO - codeparrot_training - Step 6807: {'lr': 0.00048772871017040256, 'samples': 1307136, 'steps': 6807, 'loss/train': 1.7072643041610718} 01/28/2022 21:36:00 - INFO - codeparrot_training - Step 6808: {'lr': 0.000487723646253471, 'samples': 1307328, 'steps': 6808, 'loss/train': 1.979580044746399} 01/28/2022 21:36:04 - INFO - codeparrot_training - Step 6809: {'lr': 0.00048771858131820684, 'samples': 1307520, 'steps': 6809, 'loss/train': 1.7581602931022644} 01/28/2022 21:36:08 - INFO - codeparrot_training - Step 6810: {'lr': 0.0004877135153646318, 'samples': 1307712, 'steps': 6810, 'loss/train': 2.39318186044693} 01/28/2022 21:36:12 - INFO - codeparrot_training - Step 6811: {'lr': 0.0004877084483927675, 'samples': 1307904, 'steps': 6811, 'loss/train': 1.0069898664951324} 01/28/2022 21:36:18 - INFO - codeparrot_training - Step 6812: {'lr': 0.00048770338040263574, 'samples': 1308096, 'steps': 6812, 'loss/train': 2.122567892074585} 01/28/2022 21:36:22 - INFO - codeparrot_training - Step 6813: {'lr': 0.00048769831139425815, 'samples': 1308288, 'steps': 6813, 'loss/train': 2.7325592637062073} 01/28/2022 21:36:26 - INFO - codeparrot_training - Step 6814: {'lr': 0.0004876932413676565, 'samples': 1308480, 'steps': 6814, 'loss/train': 2.0885506868362427} 01/28/2022 21:36:30 - INFO - codeparrot_training - Step 6815: {'lr': 0.0004876881703228524, 'samples': 1308672, 'steps': 6815, 'loss/train': 1.0202890634536743} 01/28/2022 21:36:35 - INFO - codeparrot_training - Step 6816: {'lr': 0.0004876830982598677, 'samples': 1308864, 'steps': 6816, 'loss/train': 0.49281343817710876} 01/28/2022 21:36:40 - INFO - codeparrot_training - Step 6817: {'lr': 0.0004876780251787241, 'samples': 1309056, 'steps': 6817, 'loss/train': 1.6098001599311829} 01/28/2022 21:36:44 - INFO - codeparrot_training - Step 6818: {'lr': 0.0004876729510794433, 'samples': 1309248, 'steps': 6818, 'loss/train': 1.9699932932853699} 01/28/2022 21:36:49 - INFO - codeparrot_training - Step 6819: {'lr': 0.00048766787596204704, 'samples': 1309440, 'steps': 6819, 'loss/train': 1.4975980818271637} 01/28/2022 21:36:53 - INFO - codeparrot_training - Step 6820: {'lr': 0.000487662799826557, 'samples': 1309632, 'steps': 6820, 'loss/train': 2.103273868560791} 01/28/2022 21:36:59 - INFO - codeparrot_training - Step 6821: {'lr': 0.00048765772267299513, 'samples': 1309824, 'steps': 6821, 'loss/train': 2.598573088645935} 01/28/2022 21:37:03 - INFO - codeparrot_training - Step 6822: {'lr': 0.00048765264450138297, 'samples': 1310016, 'steps': 6822, 'loss/train': 2.4833022952079773} 01/28/2022 21:37:08 - INFO - codeparrot_training - Step 6823: {'lr': 0.00048764756531174237, 'samples': 1310208, 'steps': 6823, 'loss/train': 3.116583466529846} 01/28/2022 21:37:12 - INFO - codeparrot_training - Step 6824: {'lr': 0.000487642485104095, 'samples': 1310400, 'steps': 6824, 'loss/train': 0.26353953033685684} 01/28/2022 21:37:16 - INFO - codeparrot_training - Step 6825: {'lr': 0.0004876374038784627, 'samples': 1310592, 'steps': 6825, 'loss/train': 2.221232235431671} 01/28/2022 21:37:22 - INFO - codeparrot_training - Step 6826: {'lr': 0.0004876323216348673, 'samples': 1310784, 'steps': 6826, 'loss/train': 1.5530844926834106} 01/28/2022 21:37:26 - INFO - codeparrot_training - Step 6827: {'lr': 0.0004876272383733304, 'samples': 1310976, 'steps': 6827, 'loss/train': 2.24614816904068} 01/28/2022 21:37:30 - INFO - codeparrot_training - Step 6828: {'lr': 0.0004876221540938739, 'samples': 1311168, 'steps': 6828, 'loss/train': 0.8048956096172333} 01/28/2022 21:37:34 - INFO - codeparrot_training - Step 6829: {'lr': 0.00048761706879651956, 'samples': 1311360, 'steps': 6829, 'loss/train': 2.073075771331787} 01/28/2022 21:37:39 - INFO - codeparrot_training - Step 6830: {'lr': 0.00048761198248128913, 'samples': 1311552, 'steps': 6830, 'loss/train': 1.2313981354236603} 01/28/2022 21:37:43 - INFO - codeparrot_training - Step 6831: {'lr': 0.00048760689514820444, 'samples': 1311744, 'steps': 6831, 'loss/train': 1.1299197971820831} 01/28/2022 21:37:49 - INFO - codeparrot_training - Step 6832: {'lr': 0.0004876018067972872, 'samples': 1311936, 'steps': 6832, 'loss/train': 0.19076033681631088} 01/28/2022 21:37:53 - INFO - codeparrot_training - Step 6833: {'lr': 0.00048759671742855935, 'samples': 1312128, 'steps': 6833, 'loss/train': 1.945297122001648} 01/28/2022 21:37:57 - INFO - codeparrot_training - Step 6834: {'lr': 0.00048759162704204253, 'samples': 1312320, 'steps': 6834, 'loss/train': 2.044318199157715} 01/28/2022 21:38:01 - INFO - codeparrot_training - Step 6835: {'lr': 0.0004875865356377587, 'samples': 1312512, 'steps': 6835, 'loss/train': 1.9777575731277466} 01/28/2022 21:38:06 - INFO - codeparrot_training - Step 6836: {'lr': 0.0004875814432157295, 'samples': 1312704, 'steps': 6836, 'loss/train': 0.9618055522441864} 01/28/2022 21:38:11 - INFO - codeparrot_training - Step 6837: {'lr': 0.0004875763497759769, 'samples': 1312896, 'steps': 6837, 'loss/train': 1.6489412784576416} 01/28/2022 21:38:15 - INFO - codeparrot_training - Step 6838: {'lr': 0.00048757125531852263, 'samples': 1313088, 'steps': 6838, 'loss/train': 2.607424736022949} 01/28/2022 21:38:19 - INFO - codeparrot_training - Step 6839: {'lr': 0.00048756615984338857, 'samples': 1313280, 'steps': 6839, 'loss/train': 0.5270792841911316} 01/28/2022 21:38:24 - INFO - codeparrot_training - Step 6840: {'lr': 0.0004875610633505965, 'samples': 1313472, 'steps': 6840, 'loss/train': 1.3913813531398773} 01/28/2022 21:38:28 - INFO - codeparrot_training - Step 6841: {'lr': 0.00048755596584016824, 'samples': 1313664, 'steps': 6841, 'loss/train': 1.131107121706009} 01/28/2022 21:38:34 - INFO - codeparrot_training - Step 6842: {'lr': 0.0004875508673121257, 'samples': 1313856, 'steps': 6842, 'loss/train': 2.7837849855422974} 01/28/2022 21:38:38 - INFO - codeparrot_training - Step 6843: {'lr': 0.00048754576776649066, 'samples': 1314048, 'steps': 6843, 'loss/train': 1.844286561012268} 01/28/2022 21:38:42 - INFO - codeparrot_training - Step 6844: {'lr': 0.000487540667203285, 'samples': 1314240, 'steps': 6844, 'loss/train': 2.3928143978118896} 01/28/2022 21:38:47 - INFO - codeparrot_training - Step 6845: {'lr': 0.0004875355656225305, 'samples': 1314432, 'steps': 6845, 'loss/train': 1.442175954580307} 01/28/2022 21:38:51 - INFO - codeparrot_training - Step 6846: {'lr': 0.0004875304630242491, 'samples': 1314624, 'steps': 6846, 'loss/train': 1.9528021216392517} 01/28/2022 21:38:56 - INFO - codeparrot_training - Step 6847: {'lr': 0.00048752535940846267, 'samples': 1314816, 'steps': 6847, 'loss/train': 1.6282312273979187} 01/28/2022 21:39:01 - INFO - codeparrot_training - Step 6848: {'lr': 0.0004875202547751929, 'samples': 1315008, 'steps': 6848, 'loss/train': 0.837057888507843} 01/28/2022 21:39:05 - INFO - codeparrot_training - Step 6849: {'lr': 0.00048751514912446185, 'samples': 1315200, 'steps': 6849, 'loss/train': 1.633419156074524} 01/28/2022 21:39:09 - INFO - codeparrot_training - Step 6850: {'lr': 0.0004875100424562914, 'samples': 1315392, 'steps': 6850, 'loss/train': 2.2438756227493286} 01/28/2022 21:39:13 - INFO - codeparrot_training - Step 6851: {'lr': 0.0004875049347707032, 'samples': 1315584, 'steps': 6851, 'loss/train': 2.061512768268585} 01/28/2022 21:39:19 - INFO - codeparrot_training - Step 6852: {'lr': 0.00048749982606771934, 'samples': 1315776, 'steps': 6852, 'loss/train': 2.557405650615692} 01/28/2022 21:39:23 - INFO - codeparrot_training - Step 6853: {'lr': 0.00048749471634736163, 'samples': 1315968, 'steps': 6853, 'loss/train': 1.4821065366268158} 01/28/2022 21:39:27 - INFO - codeparrot_training - Step 6854: {'lr': 0.0004874896056096521, 'samples': 1316160, 'steps': 6854, 'loss/train': 1.9811064004898071} 01/28/2022 21:39:32 - INFO - codeparrot_training - Step 6855: {'lr': 0.0004874844938546123, 'samples': 1316352, 'steps': 6855, 'loss/train': 1.9593313336372375} 01/28/2022 21:39:36 - INFO - codeparrot_training - Step 6856: {'lr': 0.0004874793810822644, 'samples': 1316544, 'steps': 6856, 'loss/train': 2.795668601989746} 01/28/2022 21:39:42 - INFO - codeparrot_training - Step 6857: {'lr': 0.00048747426729263036, 'samples': 1316736, 'steps': 6857, 'loss/train': 2.0267701148986816} 01/28/2022 21:39:46 - INFO - codeparrot_training - Step 6858: {'lr': 0.0004874691524857318, 'samples': 1316928, 'steps': 6858, 'loss/train': 1.8961331248283386} 01/28/2022 21:39:51 - INFO - codeparrot_training - Step 6859: {'lr': 0.00048746403666159087, 'samples': 1317120, 'steps': 6859, 'loss/train': 1.8852707147598267} 01/28/2022 21:39:55 - INFO - codeparrot_training - Step 6860: {'lr': 0.0004874589198202294, 'samples': 1317312, 'steps': 6860, 'loss/train': 1.92848539352417} 01/28/2022 21:39:59 - INFO - codeparrot_training - Step 6861: {'lr': 0.0004874538019616693, 'samples': 1317504, 'steps': 6861, 'loss/train': 1.6026751399040222} 01/28/2022 21:40:05 - INFO - codeparrot_training - Step 6862: {'lr': 0.0004874486830859326, 'samples': 1317696, 'steps': 6862, 'loss/train': 0.905624270439148} 01/28/2022 21:40:09 - INFO - codeparrot_training - Step 6863: {'lr': 0.0004874435631930411, 'samples': 1317888, 'steps': 6863, 'loss/train': 2.1058693528175354} 01/28/2022 21:40:13 - INFO - codeparrot_training - Step 6864: {'lr': 0.0004874384422830167, 'samples': 1318080, 'steps': 6864, 'loss/train': 2.0091984272003174} 01/28/2022 21:40:17 - INFO - codeparrot_training - Step 6865: {'lr': 0.0004874333203558815, 'samples': 1318272, 'steps': 6865, 'loss/train': 2.8024388551712036} 01/28/2022 21:40:22 - INFO - codeparrot_training - Step 6866: {'lr': 0.0004874281974116573, 'samples': 1318464, 'steps': 6866, 'loss/train': 2.585586190223694} 01/28/2022 21:40:28 - INFO - codeparrot_training - Step 6867: {'lr': 0.0004874230734503661, 'samples': 1318656, 'steps': 6867, 'loss/train': 0.981174498796463} 01/28/2022 21:40:32 - INFO - codeparrot_training - Step 6868: {'lr': 0.00048741794847202984, 'samples': 1318848, 'steps': 6868, 'loss/train': 1.3078725636005402} 01/28/2022 21:40:36 - INFO - codeparrot_training - Step 6869: {'lr': 0.00048741282247667054, 'samples': 1319040, 'steps': 6869, 'loss/train': 2.535730004310608} 01/28/2022 21:40:40 - INFO - codeparrot_training - Step 6870: {'lr': 0.00048740769546431, 'samples': 1319232, 'steps': 6870, 'loss/train': 0.9855654537677765} 01/28/2022 21:40:45 - INFO - codeparrot_training - Step 6871: {'lr': 0.0004874025674349704, 'samples': 1319424, 'steps': 6871, 'loss/train': 1.93938410282135} 01/28/2022 21:40:50 - INFO - codeparrot_training - Step 6872: {'lr': 0.00048739743838867344, 'samples': 1319616, 'steps': 6872, 'loss/train': 1.6077336072921753} 01/28/2022 21:40:54 - INFO - codeparrot_training - Step 6873: {'lr': 0.0004873923083254413, 'samples': 1319808, 'steps': 6873, 'loss/train': 2.2571688294410706} 01/28/2022 21:40:58 - INFO - codeparrot_training - Step 6874: {'lr': 0.0004873871772452959, 'samples': 1320000, 'steps': 6874, 'loss/train': 1.833693265914917} 01/28/2022 21:41:03 - INFO - codeparrot_training - Step 6875: {'lr': 0.00048738204514825917, 'samples': 1320192, 'steps': 6875, 'loss/train': 2.1192310452461243} 01/28/2022 21:41:07 - INFO - codeparrot_training - Step 6876: {'lr': 0.0004873769120343532, 'samples': 1320384, 'steps': 6876, 'loss/train': 1.9238834381103516} 01/28/2022 21:41:12 - INFO - codeparrot_training - Step 6877: {'lr': 0.0004873717779035999, 'samples': 1320576, 'steps': 6877, 'loss/train': 0.4073866009712219} 01/28/2022 21:41:16 - INFO - codeparrot_training - Step 6878: {'lr': 0.00048736664275602124, 'samples': 1320768, 'steps': 6878, 'loss/train': 1.0034732222557068} 01/28/2022 21:41:20 - INFO - codeparrot_training - Step 6879: {'lr': 0.00048736150659163925, 'samples': 1320960, 'steps': 6879, 'loss/train': 1.8907458186149597} 01/28/2022 21:41:25 - INFO - codeparrot_training - Step 6880: {'lr': 0.000487356369410476, 'samples': 1321152, 'steps': 6880, 'loss/train': 3.899169087409973} 01/28/2022 21:41:29 - INFO - codeparrot_training - Step 6881: {'lr': 0.00048735123121255335, 'samples': 1321344, 'steps': 6881, 'loss/train': 1.8699184656143188} 01/28/2022 21:41:35 - INFO - codeparrot_training - Step 6882: {'lr': 0.0004873460919978935, 'samples': 1321536, 'steps': 6882, 'loss/train': 2.3130460381507874} 01/28/2022 21:41:39 - INFO - codeparrot_training - Step 6883: {'lr': 0.00048734095176651825, 'samples': 1321728, 'steps': 6883, 'loss/train': 1.8388960361480713} 01/28/2022 21:41:43 - INFO - codeparrot_training - Step 6884: {'lr': 0.00048733581051844976, 'samples': 1321920, 'steps': 6884, 'loss/train': 1.4296292066574097} 01/28/2022 21:41:47 - INFO - codeparrot_training - Step 6885: {'lr': 0.0004873306682537101, 'samples': 1322112, 'steps': 6885, 'loss/train': 1.832992672920227} 01/28/2022 21:41:52 - INFO - codeparrot_training - Step 6886: {'lr': 0.0004873255249723211, 'samples': 1322304, 'steps': 6886, 'loss/train': 2.094925045967102} 01/28/2022 21:41:57 - INFO - codeparrot_training - Step 6887: {'lr': 0.000487320380674305, 'samples': 1322496, 'steps': 6887, 'loss/train': 1.5830503106117249} 01/28/2022 21:42:01 - INFO - codeparrot_training - Step 6888: {'lr': 0.0004873152353596837, 'samples': 1322688, 'steps': 6888, 'loss/train': 2.0722192525863647} 01/28/2022 21:42:06 - INFO - codeparrot_training - Step 6889: {'lr': 0.00048731008902847927, 'samples': 1322880, 'steps': 6889, 'loss/train': 1.9631924629211426} 01/28/2022 21:42:10 - INFO - codeparrot_training - Step 6890: {'lr': 0.0004873049416807138, 'samples': 1323072, 'steps': 6890, 'loss/train': 1.7288013696670532} 01/28/2022 21:42:15 - INFO - codeparrot_training - Step 6891: {'lr': 0.00048729979331640927, 'samples': 1323264, 'steps': 6891, 'loss/train': 1.6180656552314758} 01/28/2022 21:42:20 - INFO - codeparrot_training - Step 6892: {'lr': 0.0004872946439355879, 'samples': 1323456, 'steps': 6892, 'loss/train': 1.5381909012794495} 01/28/2022 21:42:24 - INFO - codeparrot_training - Step 6893: {'lr': 0.0004872894935382715, 'samples': 1323648, 'steps': 6893, 'loss/train': 2.2095925211906433} 01/28/2022 21:42:28 - INFO - codeparrot_training - Step 6894: {'lr': 0.00048728434212448233, 'samples': 1323840, 'steps': 6894, 'loss/train': 1.6953907012939453} 01/28/2022 21:42:32 - INFO - codeparrot_training - Step 6895: {'lr': 0.0004872791896942423, 'samples': 1324032, 'steps': 6895, 'loss/train': 2.019526720046997} 01/28/2022 21:42:38 - INFO - codeparrot_training - Step 6896: {'lr': 0.0004872740362475737, 'samples': 1324224, 'steps': 6896, 'loss/train': 1.8060715198516846} 01/28/2022 21:42:42 - INFO - codeparrot_training - Step 6897: {'lr': 0.00048726888178449835, 'samples': 1324416, 'steps': 6897, 'loss/train': 0.9333827197551727} 01/28/2022 21:42:46 - INFO - codeparrot_training - Step 6898: {'lr': 0.00048726372630503845, 'samples': 1324608, 'steps': 6898, 'loss/train': 0.927013635635376} 01/28/2022 21:42:50 - INFO - codeparrot_training - Step 6899: {'lr': 0.00048725856980921616, 'samples': 1324800, 'steps': 6899, 'loss/train': 2.127669095993042} 01/28/2022 21:42:55 - INFO - codeparrot_training - Step 6900: {'lr': 0.0004872534122970535, 'samples': 1324992, 'steps': 6900, 'loss/train': 1.5851266980171204} 01/28/2022 21:43:01 - INFO - codeparrot_training - Step 6901: {'lr': 0.00048724825376857253, 'samples': 1325184, 'steps': 6901, 'loss/train': 2.1369155645370483} 01/28/2022 21:43:05 - INFO - codeparrot_training - Step 6902: {'lr': 0.0004872430942237953, 'samples': 1325376, 'steps': 6902, 'loss/train': 0.4624437689781189} 01/28/2022 21:43:09 - INFO - codeparrot_training - Step 6903: {'lr': 0.0004872379336627441, 'samples': 1325568, 'steps': 6903, 'loss/train': 1.326988935470581} 01/28/2022 21:43:13 - INFO - codeparrot_training - Step 6904: {'lr': 0.0004872327720854409, 'samples': 1325760, 'steps': 6904, 'loss/train': 2.4157519340515137} 01/28/2022 21:43:18 - INFO - codeparrot_training - Step 6905: {'lr': 0.0004872276094919078, 'samples': 1325952, 'steps': 6905, 'loss/train': 2.0344455242156982} 01/28/2022 21:43:23 - INFO - codeparrot_training - Step 6906: {'lr': 0.00048722244588216695, 'samples': 1326144, 'steps': 6906, 'loss/train': 1.7477146983146667} 01/28/2022 21:43:27 - INFO - codeparrot_training - Step 6907: {'lr': 0.00048721728125624054, 'samples': 1326336, 'steps': 6907, 'loss/train': 2.357297658920288} 01/28/2022 21:43:32 - INFO - codeparrot_training - Step 6908: {'lr': 0.0004872121156141506, 'samples': 1326528, 'steps': 6908, 'loss/train': 0.9845839440822601} 01/28/2022 21:43:36 - INFO - codeparrot_training - Step 6909: {'lr': 0.0004872069489559192, 'samples': 1326720, 'steps': 6909, 'loss/train': 1.741697072982788} 01/28/2022 21:43:40 - INFO - codeparrot_training - Step 6910: {'lr': 0.00048720178128156856, 'samples': 1326912, 'steps': 6910, 'loss/train': 1.985187828540802} 01/28/2022 21:43:46 - INFO - codeparrot_training - Step 6911: {'lr': 0.00048719661259112086, 'samples': 1327104, 'steps': 6911, 'loss/train': 2.123471975326538} 01/28/2022 21:43:51 - INFO - codeparrot_training - Step 6912: {'lr': 0.0004871914428845982, 'samples': 1327296, 'steps': 6912, 'loss/train': 1.6560899019241333} 01/28/2022 21:43:55 - INFO - codeparrot_training - Step 6913: {'lr': 0.0004871862721620227, 'samples': 1327488, 'steps': 6913, 'loss/train': 1.7318199276924133} 01/28/2022 21:43:59 - INFO - codeparrot_training - Step 6914: {'lr': 0.0004871811004234165, 'samples': 1327680, 'steps': 6914, 'loss/train': 1.8640508651733398} 01/28/2022 21:44:03 - INFO - codeparrot_training - Step 6915: {'lr': 0.0004871759276688018, 'samples': 1327872, 'steps': 6915, 'loss/train': 0.54746975004673} 01/28/2022 21:44:09 - INFO - codeparrot_training - Step 6916: {'lr': 0.00048717075389820074, 'samples': 1328064, 'steps': 6916, 'loss/train': 1.5141441822052002} 01/28/2022 21:44:13 - INFO - codeparrot_training - Step 6917: {'lr': 0.0004871655791116355, 'samples': 1328256, 'steps': 6917, 'loss/train': 1.6258560419082642} 01/28/2022 21:44:17 - INFO - codeparrot_training - Step 6918: {'lr': 0.00048716040330912816, 'samples': 1328448, 'steps': 6918, 'loss/train': 0.9045562148094177} 01/28/2022 21:44:21 - INFO - codeparrot_training - Step 6919: {'lr': 0.000487155226490701, 'samples': 1328640, 'steps': 6919, 'loss/train': 1.4358006119728088} 01/28/2022 21:44:26 - INFO - codeparrot_training - Step 6920: {'lr': 0.0004871500486563761, 'samples': 1328832, 'steps': 6920, 'loss/train': 1.7664159536361694} 01/28/2022 21:44:31 - INFO - codeparrot_training - Step 6921: {'lr': 0.00048714486980617577, 'samples': 1329024, 'steps': 6921, 'loss/train': 2.706457257270813} 01/28/2022 21:44:35 - INFO - codeparrot_training - Step 6922: {'lr': 0.00048713968994012216, 'samples': 1329216, 'steps': 6922, 'loss/train': 2.040251612663269} 01/28/2022 21:44:39 - INFO - codeparrot_training - Step 6923: {'lr': 0.00048713450905823736, 'samples': 1329408, 'steps': 6923, 'loss/train': 2.0922651886940002} 01/28/2022 21:44:44 - INFO - codeparrot_training - Step 6924: {'lr': 0.0004871293271605436, 'samples': 1329600, 'steps': 6924, 'loss/train': 2.145052671432495} 01/28/2022 21:44:48 - INFO - codeparrot_training - Step 6925: {'lr': 0.00048712414424706315, 'samples': 1329792, 'steps': 6925, 'loss/train': 1.6842741966247559} 01/28/2022 21:44:54 - INFO - codeparrot_training - Step 6926: {'lr': 0.0004871189603178181, 'samples': 1329984, 'steps': 6926, 'loss/train': 1.7318070530891418} 01/28/2022 21:44:58 - INFO - codeparrot_training - Step 6927: {'lr': 0.00048711377537283073, 'samples': 1330176, 'steps': 6927, 'loss/train': 2.9376107454299927} 01/28/2022 21:45:02 - INFO - codeparrot_training - Step 6928: {'lr': 0.0004871085894121233, 'samples': 1330368, 'steps': 6928, 'loss/train': 2.0796972513198853} 01/28/2022 21:45:07 - INFO - codeparrot_training - Step 6929: {'lr': 0.00048710340243571796, 'samples': 1330560, 'steps': 6929, 'loss/train': 1.3777454495429993} 01/28/2022 21:45:11 - INFO - codeparrot_training - Step 6930: {'lr': 0.0004870982144436369, 'samples': 1330752, 'steps': 6930, 'loss/train': 2.3483887910842896} 01/28/2022 21:45:16 - INFO - codeparrot_training - Step 6931: {'lr': 0.0004870930254359023, 'samples': 1330944, 'steps': 6931, 'loss/train': 1.9519439935684204} 01/28/2022 21:45:21 - INFO - codeparrot_training - Step 6932: {'lr': 0.00048708783541253655, 'samples': 1331136, 'steps': 6932, 'loss/train': 1.9243828654289246} 01/28/2022 21:45:25 - INFO - codeparrot_training - Step 6933: {'lr': 0.0004870826443735618, 'samples': 1331328, 'steps': 6933, 'loss/train': 1.5705966353416443} 01/28/2022 21:45:29 - INFO - codeparrot_training - Step 6934: {'lr': 0.0004870774523190003, 'samples': 1331520, 'steps': 6934, 'loss/train': 1.0829355418682098} 01/28/2022 21:45:33 - INFO - codeparrot_training - Step 6935: {'lr': 0.00048707225924887423, 'samples': 1331712, 'steps': 6935, 'loss/train': 1.7158842086791992} 01/28/2022 21:45:39 - INFO - codeparrot_training - Step 6936: {'lr': 0.0004870670651632059, 'samples': 1331904, 'steps': 6936, 'loss/train': 1.9408223032951355} 01/28/2022 21:45:44 - INFO - codeparrot_training - Step 6937: {'lr': 0.0004870618700620175, 'samples': 1332096, 'steps': 6937, 'loss/train': 2.4935892820358276} 01/28/2022 21:45:48 - INFO - codeparrot_training - Step 6938: {'lr': 0.0004870566739453314, 'samples': 1332288, 'steps': 6938, 'loss/train': 2.4328866004943848} 01/28/2022 21:45:52 - INFO - codeparrot_training - Step 6939: {'lr': 0.00048705147681316974, 'samples': 1332480, 'steps': 6939, 'loss/train': 1.628383755683899} 01/28/2022 21:45:56 - INFO - codeparrot_training - Step 6940: {'lr': 0.00048704627866555486, 'samples': 1332672, 'steps': 6940, 'loss/train': 0.9795419275760651} 01/28/2022 21:46:02 - INFO - codeparrot_training - Step 6941: {'lr': 0.00048704107950250887, 'samples': 1332864, 'steps': 6941, 'loss/train': 1.880135715007782} 01/28/2022 21:46:06 - INFO - codeparrot_training - Step 6942: {'lr': 0.0004870358793240543, 'samples': 1333056, 'steps': 6942, 'loss/train': 1.3330166637897491} 01/28/2022 21:46:10 - INFO - codeparrot_training - Step 6943: {'lr': 0.00048703067813021323, 'samples': 1333248, 'steps': 6943, 'loss/train': 2.0824225544929504} 01/28/2022 21:46:15 - INFO - codeparrot_training - Step 6944: {'lr': 0.000487025475921008, 'samples': 1333440, 'steps': 6944, 'loss/train': 2.3675442337989807} 01/28/2022 21:46:19 - INFO - codeparrot_training - Step 6945: {'lr': 0.0004870202726964609, 'samples': 1333632, 'steps': 6945, 'loss/train': 1.1467465460300446} 01/28/2022 21:46:23 - INFO - codeparrot_training - Step 6946: {'lr': 0.0004870150684565943, 'samples': 1333824, 'steps': 6946, 'loss/train': 1.9766297936439514} 01/28/2022 21:46:28 - INFO - codeparrot_training - Step 6947: {'lr': 0.00048700986320143026, 'samples': 1334016, 'steps': 6947, 'loss/train': 1.0468285381793976} 01/28/2022 21:46:33 - INFO - codeparrot_training - Step 6948: {'lr': 0.0004870046569309913, 'samples': 1334208, 'steps': 6948, 'loss/train': 0.6109134703874588} 01/28/2022 21:46:37 - INFO - codeparrot_training - Step 6949: {'lr': 0.0004869994496452996, 'samples': 1334400, 'steps': 6949, 'loss/train': 2.470545530319214} 01/28/2022 21:46:41 - INFO - codeparrot_training - Step 6950: {'lr': 0.0004869942413443776, 'samples': 1334592, 'steps': 6950, 'loss/train': 2.087283432483673} 01/28/2022 21:46:45 - INFO - codeparrot_training - Step 6951: {'lr': 0.0004869890320282475, 'samples': 1334784, 'steps': 6951, 'loss/train': 1.101293921470642} 01/28/2022 21:46:51 - INFO - codeparrot_training - Step 6952: {'lr': 0.0004869838216969316, 'samples': 1334976, 'steps': 6952, 'loss/train': 1.5730619430541992} 01/28/2022 21:46:55 - INFO - codeparrot_training - Step 6953: {'lr': 0.0004869786103504523, 'samples': 1335168, 'steps': 6953, 'loss/train': 2.22749662399292} 01/28/2022 21:46:59 - INFO - codeparrot_training - Step 6954: {'lr': 0.0004869733979888319, 'samples': 1335360, 'steps': 6954, 'loss/train': 1.788072645664215} 01/28/2022 21:47:04 - INFO - codeparrot_training - Step 6955: {'lr': 0.00048696818461209265, 'samples': 1335552, 'steps': 6955, 'loss/train': 2.2376827597618103} 01/28/2022 21:47:10 - INFO - codeparrot_training - Step 6956: {'lr': 0.0004869629702202569, 'samples': 1335744, 'steps': 6956, 'loss/train': 2.6824344992637634} 01/28/2022 21:47:14 - INFO - codeparrot_training - Step 6957: {'lr': 0.0004869577548133471, 'samples': 1335936, 'steps': 6957, 'loss/train': 1.68721604347229} 01/28/2022 21:47:18 - INFO - codeparrot_training - Step 6958: {'lr': 0.00048695253839138553, 'samples': 1336128, 'steps': 6958, 'loss/train': 2.46365225315094} 01/28/2022 21:47:23 - INFO - codeparrot_training - Step 6959: {'lr': 0.0004869473209543945, 'samples': 1336320, 'steps': 6959, 'loss/train': 1.7300691604614258} 01/28/2022 21:47:27 - INFO - codeparrot_training - Step 6960: {'lr': 0.00048694210250239646, 'samples': 1336512, 'steps': 6960, 'loss/train': 1.7770408987998962} 01/28/2022 21:47:32 - INFO - codeparrot_training - Step 6961: {'lr': 0.0004869368830354136, 'samples': 1336704, 'steps': 6961, 'loss/train': 1.9586668610572815} 01/28/2022 21:47:36 - INFO - codeparrot_training - Step 6962: {'lr': 0.00048693166255346843, 'samples': 1336896, 'steps': 6962, 'loss/train': 2.0974826216697693} 01/28/2022 21:47:41 - INFO - codeparrot_training - Step 6963: {'lr': 0.0004869264410565832, 'samples': 1337088, 'steps': 6963, 'loss/train': 2.099624991416931} 01/28/2022 21:47:45 - INFO - codeparrot_training - Step 6964: {'lr': 0.00048692121854478033, 'samples': 1337280, 'steps': 6964, 'loss/train': 1.5984020233154297} 01/28/2022 21:47:49 - INFO - codeparrot_training - Step 6965: {'lr': 0.00048691599501808223, 'samples': 1337472, 'steps': 6965, 'loss/train': 1.9992996454238892} 01/28/2022 21:47:54 - INFO - codeparrot_training - Step 6966: {'lr': 0.0004869107704765112, 'samples': 1337664, 'steps': 6966, 'loss/train': 2.1756282448768616} 01/28/2022 21:47:59 - INFO - codeparrot_training - Step 6967: {'lr': 0.00048690554492008967, 'samples': 1337856, 'steps': 6967, 'loss/train': 1.135731965303421} 01/28/2022 21:48:03 - INFO - codeparrot_training - Step 6968: {'lr': 0.00048690031834884004, 'samples': 1338048, 'steps': 6968, 'loss/train': 1.1250666081905365} 01/28/2022 21:48:07 - INFO - codeparrot_training - Step 6969: {'lr': 0.0004868950907627846, 'samples': 1338240, 'steps': 6969, 'loss/train': 1.7826637029647827} 01/28/2022 21:48:11 - INFO - codeparrot_training - Step 6970: {'lr': 0.00048688986216194585, 'samples': 1338432, 'steps': 6970, 'loss/train': 1.929042398929596} 01/28/2022 21:48:18 - INFO - codeparrot_training - Step 6971: {'lr': 0.0004868846325463462, 'samples': 1338624, 'steps': 6971, 'loss/train': 0.7427905350923538} 01/28/2022 21:48:22 - INFO - codeparrot_training - Step 6972: {'lr': 0.000486879401916008, 'samples': 1338816, 'steps': 6972, 'loss/train': 1.4362073242664337} 01/28/2022 21:48:26 - INFO - codeparrot_training - Step 6973: {'lr': 0.0004868741702709536, 'samples': 1339008, 'steps': 6973, 'loss/train': 1.424689382314682} 01/28/2022 21:48:30 - INFO - codeparrot_training - Step 6974: {'lr': 0.0004868689376112055, 'samples': 1339200, 'steps': 6974, 'loss/train': 1.8257809281349182} 01/28/2022 21:48:35 - INFO - codeparrot_training - Step 6975: {'lr': 0.000486863703936786, 'samples': 1339392, 'steps': 6975, 'loss/train': 7.795653820037842} 01/28/2022 21:48:39 - INFO - codeparrot_training - Step 6976: {'lr': 0.0004868584692477178, 'samples': 1339584, 'steps': 6976, 'loss/train': 1.3379575610160828} 01/28/2022 21:48:44 - INFO - codeparrot_training - Step 6977: {'lr': 0.000486853233544023, 'samples': 1339776, 'steps': 6977, 'loss/train': 1.9999328255653381} 01/28/2022 21:48:49 - INFO - codeparrot_training - Step 6978: {'lr': 0.0004868479968257241, 'samples': 1339968, 'steps': 6978, 'loss/train': 2.209173560142517} 01/28/2022 21:48:53 - INFO - codeparrot_training - Step 6979: {'lr': 0.0004868427590928437, 'samples': 1340160, 'steps': 6979, 'loss/train': 1.0986370146274567} 01/28/2022 21:48:57 - INFO - codeparrot_training - Step 6980: {'lr': 0.0004868375203454041, 'samples': 1340352, 'steps': 6980, 'loss/train': 1.8376836776733398} 01/28/2022 21:49:01 - INFO - codeparrot_training - Step 6981: {'lr': 0.0004868322805834278, 'samples': 1340544, 'steps': 6981, 'loss/train': 1.5684036612510681} 01/28/2022 21:49:08 - INFO - codeparrot_training - Step 6982: {'lr': 0.0004868270398069371, 'samples': 1340736, 'steps': 6982, 'loss/train': 1.262591689825058} 01/28/2022 21:49:12 - INFO - codeparrot_training - Step 6983: {'lr': 0.0004868217980159546, 'samples': 1340928, 'steps': 6983, 'loss/train': 2.1724830865859985} 01/28/2022 21:49:17 - INFO - codeparrot_training - Step 6984: {'lr': 0.0004868165552105028, 'samples': 1341120, 'steps': 6984, 'loss/train': 2.0298948884010315} 01/28/2022 21:49:21 - INFO - codeparrot_training - Step 6985: {'lr': 0.000486811311390604, 'samples': 1341312, 'steps': 6985, 'loss/train': 1.9523169994354248} 01/28/2022 21:49:25 - INFO - codeparrot_training - Step 6986: {'lr': 0.0004868060665562808, 'samples': 1341504, 'steps': 6986, 'loss/train': 1.716623604297638} 01/28/2022 21:49:29 - INFO - codeparrot_training - Step 6987: {'lr': 0.0004868008207075555, 'samples': 1341696, 'steps': 6987, 'loss/train': 2.5651445388793945} 01/28/2022 21:49:35 - INFO - codeparrot_training - Step 6988: {'lr': 0.0004867955738444508, 'samples': 1341888, 'steps': 6988, 'loss/train': 1.4555760025978088} 01/28/2022 21:49:39 - INFO - codeparrot_training - Step 6989: {'lr': 0.000486790325966989, 'samples': 1342080, 'steps': 6989, 'loss/train': 1.5023971796035767} 01/28/2022 21:49:43 - INFO - codeparrot_training - Step 6990: {'lr': 0.0004867850770751926, 'samples': 1342272, 'steps': 6990, 'loss/train': 2.2808908224105835} 01/28/2022 21:49:47 - INFO - codeparrot_training - Step 6991: {'lr': 0.00048677982716908416, 'samples': 1342464, 'steps': 6991, 'loss/train': 1.674548864364624} 01/28/2022 21:49:52 - INFO - codeparrot_training - Step 6992: {'lr': 0.0004867745762486861, 'samples': 1342656, 'steps': 6992, 'loss/train': 1.5410690903663635} 01/28/2022 21:49:57 - INFO - codeparrot_training - Step 6993: {'lr': 0.0004867693243140209, 'samples': 1342848, 'steps': 6993, 'loss/train': 1.9829635620117188} 01/28/2022 21:50:01 - INFO - codeparrot_training - Step 6994: {'lr': 0.0004867640713651112, 'samples': 1343040, 'steps': 6994, 'loss/train': 1.6193899512290955} 01/28/2022 21:50:06 - INFO - codeparrot_training - Step 6995: {'lr': 0.0004867588174019794, 'samples': 1343232, 'steps': 6995, 'loss/train': 2.1771181225776672} 01/28/2022 21:50:10 - INFO - codeparrot_training - Step 6996: {'lr': 0.00048675356242464785, 'samples': 1343424, 'steps': 6996, 'loss/train': 1.9371305108070374} 01/28/2022 21:50:14 - INFO - codeparrot_training - Step 6997: {'lr': 0.0004867483064331394, 'samples': 1343616, 'steps': 6997, 'loss/train': 2.359727919101715} 01/28/2022 21:50:20 - INFO - codeparrot_training - Step 6998: {'lr': 0.00048674304942747626, 'samples': 1343808, 'steps': 6998, 'loss/train': 3.0525466203689575} 01/28/2022 21:50:24 - INFO - codeparrot_training - Step 6999: {'lr': 0.0004867377914076811, 'samples': 1344000, 'steps': 6999, 'loss/train': 2.1387208700180054} 01/28/2022 21:50:29 - INFO - codeparrot_training - Step 7000: {'lr': 0.00048673253237377644, 'samples': 1344192, 'steps': 7000, 'loss/train': 1.3656534254550934} 01/28/2022 21:50:33 - INFO - codeparrot_training - Step 7001: {'lr': 0.00048672727232578476, 'samples': 1344384, 'steps': 7001, 'loss/train': 1.6647638082504272} 01/28/2022 21:50:37 - INFO - codeparrot_training - Step 7002: {'lr': 0.0004867220112637286, 'samples': 1344576, 'steps': 7002, 'loss/train': 2.037169396877289} 01/28/2022 21:50:42 - INFO - codeparrot_training - Step 7003: {'lr': 0.00048671674918763055, 'samples': 1344768, 'steps': 7003, 'loss/train': 2.592536687850952} 01/28/2022 21:50:47 - INFO - codeparrot_training - Step 7004: {'lr': 0.00048671148609751307, 'samples': 1344960, 'steps': 7004, 'loss/train': 2.4137551188468933} 01/28/2022 21:50:51 - INFO - codeparrot_training - Step 7005: {'lr': 0.0004867062219933988, 'samples': 1345152, 'steps': 7005, 'loss/train': 2.292167901992798} 01/28/2022 21:50:55 - INFO - codeparrot_training - Step 7006: {'lr': 0.00048670095687531023, 'samples': 1345344, 'steps': 7006, 'loss/train': 1.4660016596317291} 01/28/2022 21:50:59 - INFO - codeparrot_training - Step 7007: {'lr': 0.0004866956907432699, 'samples': 1345536, 'steps': 7007, 'loss/train': 2.3791247606277466} 01/28/2022 21:51:05 - INFO - codeparrot_training - Step 7008: {'lr': 0.00048669042359730043, 'samples': 1345728, 'steps': 7008, 'loss/train': 0.1700092926621437} 01/28/2022 21:51:09 - INFO - codeparrot_training - Step 7009: {'lr': 0.00048668515543742426, 'samples': 1345920, 'steps': 7009, 'loss/train': 1.6299414038658142} 01/28/2022 21:51:13 - INFO - codeparrot_training - Step 7010: {'lr': 0.0004866798862636641, 'samples': 1346112, 'steps': 7010, 'loss/train': 2.154057741165161} 01/28/2022 21:51:17 - INFO - codeparrot_training - Step 7011: {'lr': 0.0004866746160760425, 'samples': 1346304, 'steps': 7011, 'loss/train': 1.8644240498542786} 01/28/2022 21:51:22 - INFO - codeparrot_training - Step 7012: {'lr': 0.0004866693448745819, 'samples': 1346496, 'steps': 7012, 'loss/train': 2.3087337613105774} 01/28/2022 21:51:27 - INFO - codeparrot_training - Step 7013: {'lr': 0.000486664072659305, 'samples': 1346688, 'steps': 7013, 'loss/train': 2.7950071692466736} 01/28/2022 21:51:32 - INFO - codeparrot_training - Step 7014: {'lr': 0.0004866587994302344, 'samples': 1346880, 'steps': 7014, 'loss/train': 1.7892445921897888} 01/28/2022 21:51:36 - INFO - codeparrot_training - Step 7015: {'lr': 0.0004866535251873926, 'samples': 1347072, 'steps': 7015, 'loss/train': 1.131672888994217} 01/28/2022 21:51:40 - INFO - codeparrot_training - Step 7016: {'lr': 0.0004866482499308023, 'samples': 1347264, 'steps': 7016, 'loss/train': 2.0089945793151855} 01/28/2022 21:51:46 - INFO - codeparrot_training - Step 7017: {'lr': 0.000486642973660486, 'samples': 1347456, 'steps': 7017, 'loss/train': 2.4636148810386658} 01/28/2022 21:51:50 - INFO - codeparrot_training - Step 7018: {'lr': 0.00048663769637646636, 'samples': 1347648, 'steps': 7018, 'loss/train': 2.534807503223419} 01/28/2022 21:51:54 - INFO - codeparrot_training - Step 7019: {'lr': 0.000486632418078766, 'samples': 1347840, 'steps': 7019, 'loss/train': 1.4812643229961395} 01/28/2022 21:51:59 - INFO - codeparrot_training - Step 7020: {'lr': 0.0004866271387674075, 'samples': 1348032, 'steps': 7020, 'loss/train': 2.6734326481819153} 01/28/2022 21:52:03 - INFO - codeparrot_training - Step 7021: {'lr': 0.00048662185844241347, 'samples': 1348224, 'steps': 7021, 'loss/train': 1.0476641356945038} 01/28/2022 21:52:07 - INFO - codeparrot_training - Step 7022: {'lr': 0.00048661657710380647, 'samples': 1348416, 'steps': 7022, 'loss/train': 0.745004877448082} 01/28/2022 21:52:12 - INFO - codeparrot_training - Step 7023: {'lr': 0.00048661129475160926, 'samples': 1348608, 'steps': 7023, 'loss/train': 1.9377200603485107} 01/28/2022 21:52:17 - INFO - codeparrot_training - Step 7024: {'lr': 0.00048660601138584436, 'samples': 1348800, 'steps': 7024, 'loss/train': 1.624030888080597} 01/28/2022 21:52:21 - INFO - codeparrot_training - Step 7025: {'lr': 0.00048660072700653446, 'samples': 1348992, 'steps': 7025, 'loss/train': 1.370413988828659} 01/28/2022 21:52:25 - INFO - codeparrot_training - Step 7026: {'lr': 0.0004865954416137022, 'samples': 1349184, 'steps': 7026, 'loss/train': 2.1064846515655518} 01/28/2022 21:52:29 - INFO - codeparrot_training - Step 7027: {'lr': 0.0004865901552073701, 'samples': 1349376, 'steps': 7027, 'loss/train': 1.5730726718902588} 01/28/2022 21:52:35 - INFO - codeparrot_training - Step 7028: {'lr': 0.00048658486778756097, 'samples': 1349568, 'steps': 7028, 'loss/train': 2.1491957902908325} 01/28/2022 21:52:40 - INFO - codeparrot_training - Step 7029: {'lr': 0.00048657957935429734, 'samples': 1349760, 'steps': 7029, 'loss/train': 1.7626141905784607} 01/28/2022 21:52:44 - INFO - codeparrot_training - Step 7030: {'lr': 0.000486574289907602, 'samples': 1349952, 'steps': 7030, 'loss/train': 3.265774726867676} 01/28/2022 21:52:48 - INFO - codeparrot_training - Step 7031: {'lr': 0.0004865689994474974, 'samples': 1350144, 'steps': 7031, 'loss/train': 1.106330394744873} 01/28/2022 21:52:52 - INFO - codeparrot_training - Step 7032: {'lr': 0.00048656370797400643, 'samples': 1350336, 'steps': 7032, 'loss/train': 1.8893777132034302} 01/28/2022 21:52:58 - INFO - codeparrot_training - Step 7033: {'lr': 0.00048655841548715163, 'samples': 1350528, 'steps': 7033, 'loss/train': 1.6673786044120789} 01/28/2022 21:53:02 - INFO - codeparrot_training - Step 7034: {'lr': 0.00048655312198695567, 'samples': 1350720, 'steps': 7034, 'loss/train': 2.46494060754776} 01/28/2022 21:53:06 - INFO - codeparrot_training - Step 7035: {'lr': 0.00048654782747344126, 'samples': 1350912, 'steps': 7035, 'loss/train': 1.2482382953166962} 01/28/2022 21:53:11 - INFO - codeparrot_training - Step 7036: {'lr': 0.00048654253194663113, 'samples': 1351104, 'steps': 7036, 'loss/train': 1.9625009894371033} 01/28/2022 21:53:15 - INFO - codeparrot_training - Step 7037: {'lr': 0.0004865372354065478, 'samples': 1351296, 'steps': 7037, 'loss/train': 3.96699857711792} 01/28/2022 21:53:21 - INFO - codeparrot_training - Step 7038: {'lr': 0.00048653193785321415, 'samples': 1351488, 'steps': 7038, 'loss/train': 1.5397211909294128} 01/28/2022 21:53:25 - INFO - codeparrot_training - Step 7039: {'lr': 0.00048652663928665273, 'samples': 1351680, 'steps': 7039, 'loss/train': 1.7886310815811157} 01/28/2022 21:53:29 - INFO - codeparrot_training - Step 7040: {'lr': 0.00048652133970688633, 'samples': 1351872, 'steps': 7040, 'loss/train': 1.3158945143222809} 01/28/2022 21:53:33 - INFO - codeparrot_training - Step 7041: {'lr': 0.0004865160391139376, 'samples': 1352064, 'steps': 7041, 'loss/train': 1.376347839832306} 01/28/2022 21:53:38 - INFO - codeparrot_training - Step 7042: {'lr': 0.0004865107375078293, 'samples': 1352256, 'steps': 7042, 'loss/train': 2.262649655342102} 01/28/2022 21:53:44 - INFO - codeparrot_training - Step 7043: {'lr': 0.000486505434888584, 'samples': 1352448, 'steps': 7043, 'loss/train': 1.50988107919693} 01/28/2022 21:53:48 - INFO - codeparrot_training - Step 7044: {'lr': 0.0004865001312562246, 'samples': 1352640, 'steps': 7044, 'loss/train': 1.5746957659721375} 01/28/2022 21:53:52 - INFO - codeparrot_training - Step 7045: {'lr': 0.0004864948266107737, 'samples': 1352832, 'steps': 7045, 'loss/train': 2.233256757259369} 01/28/2022 21:53:56 - INFO - codeparrot_training - Step 7046: {'lr': 0.0004864895209522541, 'samples': 1353024, 'steps': 7046, 'loss/train': 2.5707467794418335} 01/28/2022 21:54:01 - INFO - codeparrot_training - Step 7047: {'lr': 0.00048648421428068843, 'samples': 1353216, 'steps': 7047, 'loss/train': 1.1979120969772339} 01/28/2022 21:54:06 - INFO - codeparrot_training - Step 7048: {'lr': 0.0004864789065960995, 'samples': 1353408, 'steps': 7048, 'loss/train': 1.0897018611431122} 01/28/2022 21:54:10 - INFO - codeparrot_training - Step 7049: {'lr': 0.00048647359789851, 'samples': 1353600, 'steps': 7049, 'loss/train': 1.572974145412445} 01/28/2022 21:54:14 - INFO - codeparrot_training - Step 7050: {'lr': 0.00048646828818794274, 'samples': 1353792, 'steps': 7050, 'loss/train': 2.0547171235084534} 01/28/2022 21:54:19 - INFO - codeparrot_training - Step 7051: {'lr': 0.00048646297746442044, 'samples': 1353984, 'steps': 7051, 'loss/train': 2.1418656706809998} 01/28/2022 21:54:23 - INFO - codeparrot_training - Step 7052: {'lr': 0.0004864576657279658, 'samples': 1354176, 'steps': 7052, 'loss/train': 1.4922170341014862} 01/28/2022 21:54:28 - INFO - codeparrot_training - Step 7053: {'lr': 0.0004864523529786016, 'samples': 1354368, 'steps': 7053, 'loss/train': 1.343011200428009} 01/28/2022 21:54:32 - INFO - codeparrot_training - Step 7054: {'lr': 0.0004864470392163506, 'samples': 1354560, 'steps': 7054, 'loss/train': 0.08010959252715111} 01/28/2022 21:54:37 - INFO - codeparrot_training - Step 7055: {'lr': 0.0004864417244412355, 'samples': 1354752, 'steps': 7055, 'loss/train': 1.413124144077301} 01/28/2022 21:54:41 - INFO - codeparrot_training - Step 7056: {'lr': 0.0004864364086532792, 'samples': 1354944, 'steps': 7056, 'loss/train': 1.8036172986030579} 01/28/2022 21:54:45 - INFO - codeparrot_training - Step 7057: {'lr': 0.00048643109185250445, 'samples': 1355136, 'steps': 7057, 'loss/train': 1.4468455016613007} 01/28/2022 21:54:51 - INFO - codeparrot_training - Step 7058: {'lr': 0.0004864257740389338, 'samples': 1355328, 'steps': 7058, 'loss/train': 1.8173627257347107} 01/28/2022 21:54:55 - INFO - codeparrot_training - Step 7059: {'lr': 0.00048642045521259044, 'samples': 1355520, 'steps': 7059, 'loss/train': 2.279808282852173} 01/28/2022 21:54:59 - INFO - codeparrot_training - Step 7060: {'lr': 0.0004864151353734968, 'samples': 1355712, 'steps': 7060, 'loss/train': 2.106466054916382} 01/28/2022 21:55:04 - INFO - codeparrot_training - Step 7061: {'lr': 0.0004864098145216758, 'samples': 1355904, 'steps': 7061, 'loss/train': 1.5356571078300476} 01/28/2022 21:55:08 - INFO - codeparrot_training - Step 7062: {'lr': 0.0004864044926571503, 'samples': 1356096, 'steps': 7062, 'loss/train': 0.8805637657642365} 01/28/2022 21:55:13 - INFO - codeparrot_training - Step 7063: {'lr': 0.00048639916977994286, 'samples': 1356288, 'steps': 7063, 'loss/train': 1.887129306793213} 01/28/2022 21:55:18 - INFO - codeparrot_training - Step 7064: {'lr': 0.0004863938458900765, 'samples': 1356480, 'steps': 7064, 'loss/train': 2.059837281703949} 01/28/2022 21:55:22 - INFO - codeparrot_training - Step 7065: {'lr': 0.000486388520987574, 'samples': 1356672, 'steps': 7065, 'loss/train': 1.7386425733566284} 01/28/2022 21:55:26 - INFO - codeparrot_training - Step 7066: {'lr': 0.0004863831950724582, 'samples': 1356864, 'steps': 7066, 'loss/train': 2.1557908058166504} 01/28/2022 21:55:30 - INFO - codeparrot_training - Step 7067: {'lr': 0.00048637786814475175, 'samples': 1357056, 'steps': 7067, 'loss/train': 1.2045327723026276} 01/28/2022 21:55:36 - INFO - codeparrot_training - Step 7068: {'lr': 0.0004863725402044776, 'samples': 1357248, 'steps': 7068, 'loss/train': 1.6327176690101624} 01/28/2022 21:55:41 - INFO - codeparrot_training - Step 7069: {'lr': 0.00048636721125165855, 'samples': 1357440, 'steps': 7069, 'loss/train': 3.0947048664093018} 01/28/2022 21:55:45 - INFO - codeparrot_training - Step 7070: {'lr': 0.0004863618812863174, 'samples': 1357632, 'steps': 7070, 'loss/train': 2.020423650741577} 01/28/2022 21:55:49 - INFO - codeparrot_training - Step 7071: {'lr': 0.0004863565503084771, 'samples': 1357824, 'steps': 7071, 'loss/train': 2.4270517230033875} 01/28/2022 21:55:53 - INFO - codeparrot_training - Step 7072: {'lr': 0.0004863512183181603, 'samples': 1358016, 'steps': 7072, 'loss/train': 1.7636897563934326} 01/28/2022 21:55:59 - INFO - codeparrot_training - Step 7073: {'lr': 0.0004863458853153899, 'samples': 1358208, 'steps': 7073, 'loss/train': 1.8421109318733215} 01/28/2022 21:56:03 - INFO - codeparrot_training - Step 7074: {'lr': 0.00048634055130018886, 'samples': 1358400, 'steps': 7074, 'loss/train': 1.9527202248573303} 01/28/2022 21:56:07 - INFO - codeparrot_training - Step 7075: {'lr': 0.00048633521627257993, 'samples': 1358592, 'steps': 7075, 'loss/train': 2.3027926683425903} 01/28/2022 21:56:12 - INFO - codeparrot_training - Step 7076: {'lr': 0.00048632988023258596, 'samples': 1358784, 'steps': 7076, 'loss/train': 1.599617600440979} 01/28/2022 21:56:16 - INFO - codeparrot_training - Step 7077: {'lr': 0.0004863245431802298, 'samples': 1358976, 'steps': 7077, 'loss/train': 1.9218159914016724} 01/28/2022 21:56:20 - INFO - codeparrot_training - Step 7078: {'lr': 0.0004863192051155344, 'samples': 1359168, 'steps': 7078, 'loss/train': 1.372210532426834} 01/28/2022 21:56:25 - INFO - codeparrot_training - Step 7079: {'lr': 0.0004863138660385225, 'samples': 1359360, 'steps': 7079, 'loss/train': 1.8662527799606323} 01/28/2022 21:56:29 - INFO - codeparrot_training - Step 7080: {'lr': 0.00048630852594921703, 'samples': 1359552, 'steps': 7080, 'loss/train': 0.9514298737049103} 01/28/2022 21:56:34 - INFO - codeparrot_training - Step 7081: {'lr': 0.00048630318484764093, 'samples': 1359744, 'steps': 7081, 'loss/train': 0.32527126371860504} 01/28/2022 21:56:38 - INFO - codeparrot_training - Step 7082: {'lr': 0.000486297842733817, 'samples': 1359936, 'steps': 7082, 'loss/train': 1.5968132615089417} 01/28/2022 21:56:42 - INFO - codeparrot_training - Step 7083: {'lr': 0.0004862924996077682, 'samples': 1360128, 'steps': 7083, 'loss/train': 1.5118167400360107} 01/28/2022 21:56:48 - INFO - codeparrot_training - Step 7084: {'lr': 0.0004862871554695173, 'samples': 1360320, 'steps': 7084, 'loss/train': 1.3191305100917816} 01/28/2022 21:56:52 - INFO - codeparrot_training - Step 7085: {'lr': 0.00048628181031908725, 'samples': 1360512, 'steps': 7085, 'loss/train': 0.18172327056527138} 01/28/2022 21:56:56 - INFO - codeparrot_training - Step 7086: {'lr': 0.00048627646415650094, 'samples': 1360704, 'steps': 7086, 'loss/train': 1.7829272747039795} 01/28/2022 21:57:00 - INFO - codeparrot_training - Step 7087: {'lr': 0.0004862711169817813, 'samples': 1360896, 'steps': 7087, 'loss/train': 1.9519028663635254} 01/28/2022 21:57:06 - INFO - codeparrot_training - Step 7088: {'lr': 0.0004862657687949512, 'samples': 1361088, 'steps': 7088, 'loss/train': 1.894947588443756} 01/28/2022 21:57:10 - INFO - codeparrot_training - Step 7089: {'lr': 0.0004862604195960336, 'samples': 1361280, 'steps': 7089, 'loss/train': 1.9013993740081787} 01/28/2022 21:57:15 - INFO - codeparrot_training - Step 7090: {'lr': 0.00048625506938505136, 'samples': 1361472, 'steps': 7090, 'loss/train': 1.5234671831130981} 01/28/2022 21:57:19 - INFO - codeparrot_training - Step 7091: {'lr': 0.00048624971816202747, 'samples': 1361664, 'steps': 7091, 'loss/train': 1.6111085414886475} 01/28/2022 21:57:23 - INFO - codeparrot_training - Step 7092: {'lr': 0.0004862443659269848, 'samples': 1361856, 'steps': 7092, 'loss/train': 1.364600032567978} 01/28/2022 21:57:29 - INFO - codeparrot_training - Step 7093: {'lr': 0.00048623901267994625, 'samples': 1362048, 'steps': 7093, 'loss/train': 2.417465329170227} 01/28/2022 21:57:33 - INFO - codeparrot_training - Step 7094: {'lr': 0.00048623365842093483, 'samples': 1362240, 'steps': 7094, 'loss/train': 1.6305765509605408} 01/28/2022 21:57:37 - INFO - codeparrot_training - Step 7095: {'lr': 0.00048622830314997334, 'samples': 1362432, 'steps': 7095, 'loss/train': 1.9253187775611877} 01/28/2022 21:57:41 - INFO - codeparrot_training - Step 7096: {'lr': 0.0004862229468670849, 'samples': 1362624, 'steps': 7096, 'loss/train': 1.863988995552063} 01/28/2022 21:57:45 - INFO - codeparrot_training - Step 7097: {'lr': 0.0004862175895722923, 'samples': 1362816, 'steps': 7097, 'loss/train': 1.5544542074203491} 01/28/2022 21:57:51 - INFO - codeparrot_training - Step 7098: {'lr': 0.0004862122312656186, 'samples': 1363008, 'steps': 7098, 'loss/train': 1.3683363497257233} 01/28/2022 21:57:56 - INFO - codeparrot_training - Step 7099: {'lr': 0.0004862068719470867, 'samples': 1363200, 'steps': 7099, 'loss/train': 1.602832317352295} 01/28/2022 21:58:00 - INFO - codeparrot_training - Step 7100: {'lr': 0.00048620151161671955, 'samples': 1363392, 'steps': 7100, 'loss/train': 1.7655197381973267} 01/28/2022 21:58:04 - INFO - codeparrot_training - Step 7101: {'lr': 0.0004861961502745401, 'samples': 1363584, 'steps': 7101, 'loss/train': 2.3576653003692627} 01/28/2022 21:58:08 - INFO - codeparrot_training - Step 7102: {'lr': 0.00048619078792057135, 'samples': 1363776, 'steps': 7102, 'loss/train': 0.9661048650741577} 01/28/2022 21:58:14 - INFO - codeparrot_training - Step 7103: {'lr': 0.00048618542455483625, 'samples': 1363968, 'steps': 7103, 'loss/train': 2.2672335505485535} 01/28/2022 21:58:18 - INFO - codeparrot_training - Step 7104: {'lr': 0.0004861800601773579, 'samples': 1364160, 'steps': 7104, 'loss/train': 1.9024924635887146} 01/28/2022 21:58:22 - INFO - codeparrot_training - Step 7105: {'lr': 0.00048617469478815905, 'samples': 1364352, 'steps': 7105, 'loss/train': 1.4490043222904205} 01/28/2022 21:58:26 - INFO - codeparrot_training - Step 7106: {'lr': 0.00048616932838726286, 'samples': 1364544, 'steps': 7106, 'loss/train': 1.8345749974250793} 01/28/2022 21:58:31 - INFO - codeparrot_training - Step 7107: {'lr': 0.0004861639609746923, 'samples': 1364736, 'steps': 7107, 'loss/train': 2.084303319454193} 01/28/2022 21:58:37 - INFO - codeparrot_training - Step 7108: {'lr': 0.0004861585925504702, 'samples': 1364928, 'steps': 7108, 'loss/train': 1.3798681497573853} 01/28/2022 21:58:41 - INFO - codeparrot_training - Step 7109: {'lr': 0.00048615322311461973, 'samples': 1365120, 'steps': 7109, 'loss/train': 2.124642491340637} 01/28/2022 21:58:45 - INFO - codeparrot_training - Step 7110: {'lr': 0.0004861478526671639, 'samples': 1365312, 'steps': 7110, 'loss/train': 2.134616196155548} 01/28/2022 21:58:50 - INFO - codeparrot_training - Step 7111: {'lr': 0.0004861424812081256, 'samples': 1365504, 'steps': 7111, 'loss/train': 2.3425508737564087} 01/28/2022 21:58:54 - INFO - codeparrot_training - Step 7112: {'lr': 0.0004861371087375279, 'samples': 1365696, 'steps': 7112, 'loss/train': 2.0995128750801086} 01/28/2022 21:58:58 - INFO - codeparrot_training - Step 7113: {'lr': 0.0004861317352553938, 'samples': 1365888, 'steps': 7113, 'loss/train': 1.8707816004753113} 01/28/2022 21:59:04 - INFO - codeparrot_training - Step 7114: {'lr': 0.0004861263607617463, 'samples': 1366080, 'steps': 7114, 'loss/train': 1.283555120229721} 01/28/2022 21:59:08 - INFO - codeparrot_training - Step 7115: {'lr': 0.00048612098525660855, 'samples': 1366272, 'steps': 7115, 'loss/train': 0.9334357380867004} 01/28/2022 21:59:13 - INFO - codeparrot_training - Step 7116: {'lr': 0.00048611560874000335, 'samples': 1366464, 'steps': 7116, 'loss/train': 2.0675257444381714} 01/28/2022 21:59:17 - INFO - codeparrot_training - Step 7117: {'lr': 0.000486110231211954, 'samples': 1366656, 'steps': 7117, 'loss/train': 1.9830406308174133} 01/28/2022 21:59:21 - INFO - codeparrot_training - Step 7118: {'lr': 0.0004861048526724833, 'samples': 1366848, 'steps': 7118, 'loss/train': 0.8444624841213226} 01/28/2022 21:59:26 - INFO - codeparrot_training - Step 7119: {'lr': 0.00048609947312161435, 'samples': 1367040, 'steps': 7119, 'loss/train': 1.7543376088142395} 01/28/2022 21:59:31 - INFO - codeparrot_training - Step 7120: {'lr': 0.0004860940925593703, 'samples': 1367232, 'steps': 7120, 'loss/train': 1.9470768570899963} 01/28/2022 21:59:35 - INFO - codeparrot_training - Step 7121: {'lr': 0.0004860887109857741, 'samples': 1367424, 'steps': 7121, 'loss/train': 2.6260101199150085} 01/28/2022 21:59:39 - INFO - codeparrot_training - Step 7122: {'lr': 0.0004860833284008488, 'samples': 1367616, 'steps': 7122, 'loss/train': 2.4098179936408997} 01/28/2022 21:59:43 - INFO - codeparrot_training - Step 7123: {'lr': 0.00048607794480461753, 'samples': 1367808, 'steps': 7123, 'loss/train': 1.928726613521576} 01/28/2022 21:59:49 - INFO - codeparrot_training - Step 7124: {'lr': 0.00048607256019710327, 'samples': 1368000, 'steps': 7124, 'loss/train': 1.0833118557929993} 01/28/2022 21:59:53 - INFO - codeparrot_training - Step 7125: {'lr': 0.0004860671745783292, 'samples': 1368192, 'steps': 7125, 'loss/train': 1.9334349632263184} 01/28/2022 21:59:57 - INFO - codeparrot_training - Step 7126: {'lr': 0.0004860617879483182, 'samples': 1368384, 'steps': 7126, 'loss/train': 2.6339553594589233} 01/28/2022 22:00:01 - INFO - codeparrot_training - Step 7127: {'lr': 0.0004860564003070935, 'samples': 1368576, 'steps': 7127, 'loss/train': 1.9384961128234863} 01/28/2022 22:00:06 - INFO - codeparrot_training - Step 7128: {'lr': 0.00048605101165467813, 'samples': 1368768, 'steps': 7128, 'loss/train': 0.22776809334754944} 01/28/2022 22:00:12 - INFO - codeparrot_training - Step 7129: {'lr': 0.00048604562199109524, 'samples': 1368960, 'steps': 7129, 'loss/train': 3.0494667291641235} 01/28/2022 22:00:16 - INFO - codeparrot_training - Step 7130: {'lr': 0.00048604023131636784, 'samples': 1369152, 'steps': 7130, 'loss/train': 1.9649595022201538} 01/28/2022 22:00:21 - INFO - codeparrot_training - Step 7131: {'lr': 0.00048603483963051896, 'samples': 1369344, 'steps': 7131, 'loss/train': 1.6689303517341614} 01/28/2022 22:00:25 - INFO - codeparrot_training - Step 7132: {'lr': 0.0004860294469335719, 'samples': 1369536, 'steps': 7132, 'loss/train': 2.039207875728607} 01/28/2022 22:00:29 - INFO - codeparrot_training - Step 7133: {'lr': 0.00048602405322554956, 'samples': 1369728, 'steps': 7133, 'loss/train': 1.4469637870788574} 01/28/2022 22:00:34 - INFO - codeparrot_training - Step 7134: {'lr': 0.00048601865850647516, 'samples': 1369920, 'steps': 7134, 'loss/train': 2.174152672290802} 01/28/2022 22:00:39 - INFO - codeparrot_training - Step 7135: {'lr': 0.0004860132627763717, 'samples': 1370112, 'steps': 7135, 'loss/train': 1.5490643978118896} 01/28/2022 22:00:43 - INFO - codeparrot_training - Step 7136: {'lr': 0.0004860078660352625, 'samples': 1370304, 'steps': 7136, 'loss/train': 1.679890751838684} 01/28/2022 22:00:47 - INFO - codeparrot_training - Step 7137: {'lr': 0.0004860024682831704, 'samples': 1370496, 'steps': 7137, 'loss/train': 1.2228533327579498} 01/28/2022 22:00:51 - INFO - codeparrot_training - Step 7138: {'lr': 0.0004859970695201187, 'samples': 1370688, 'steps': 7138, 'loss/train': 0.9921416938304901} 01/28/2022 22:00:57 - INFO - codeparrot_training - Step 7139: {'lr': 0.00048599166974613053, 'samples': 1370880, 'steps': 7139, 'loss/train': 1.7634447813034058} 01/28/2022 22:01:01 - INFO - codeparrot_training - Step 7140: {'lr': 0.000485986268961229, 'samples': 1371072, 'steps': 7140, 'loss/train': 0.8348543643951416} 01/28/2022 22:01:05 - INFO - codeparrot_training - Step 7141: {'lr': 0.0004859808671654372, 'samples': 1371264, 'steps': 7141, 'loss/train': 1.5129041075706482} 01/28/2022 22:01:09 - INFO - codeparrot_training - Step 7142: {'lr': 0.00048597546435877824, 'samples': 1371456, 'steps': 7142, 'loss/train': 1.482890009880066} 01/28/2022 22:01:14 - INFO - codeparrot_training - Step 7143: {'lr': 0.0004859700605412754, 'samples': 1371648, 'steps': 7143, 'loss/train': 1.2275487184524536} 01/28/2022 22:01:19 - INFO - codeparrot_training - Step 7144: {'lr': 0.0004859646557129517, 'samples': 1371840, 'steps': 7144, 'loss/train': 2.1765353679656982} 01/28/2022 22:01:23 - INFO - codeparrot_training - Step 7145: {'lr': 0.0004859592498738304, 'samples': 1372032, 'steps': 7145, 'loss/train': 0.7883630096912384} 01/28/2022 22:01:27 - INFO - codeparrot_training - Step 7146: {'lr': 0.00048595384302393453, 'samples': 1372224, 'steps': 7146, 'loss/train': 1.0598529875278473} 01/28/2022 22:01:32 - INFO - codeparrot_training - Step 7147: {'lr': 0.00048594843516328734, 'samples': 1372416, 'steps': 7147, 'loss/train': 2.345004916191101} 01/28/2022 22:01:36 - INFO - codeparrot_training - Step 7148: {'lr': 0.000485943026291912, 'samples': 1372608, 'steps': 7148, 'loss/train': 1.8975511193275452} 01/28/2022 22:01:42 - INFO - codeparrot_training - Step 7149: {'lr': 0.0004859376164098317, 'samples': 1372800, 'steps': 7149, 'loss/train': 1.7204920649528503} 01/28/2022 22:01:46 - INFO - codeparrot_training - Step 7150: {'lr': 0.0004859322055170695, 'samples': 1372992, 'steps': 7150, 'loss/train': 2.207105576992035} 01/28/2022 22:01:50 - INFO - codeparrot_training - Step 7151: {'lr': 0.00048592679361364867, 'samples': 1373184, 'steps': 7151, 'loss/train': 1.4314918220043182} 01/28/2022 22:01:55 - INFO - codeparrot_training - Step 7152: {'lr': 0.00048592138069959235, 'samples': 1373376, 'steps': 7152, 'loss/train': 0.7538942992687225} 01/28/2022 22:01:59 - INFO - codeparrot_training - Step 7153: {'lr': 0.0004859159667749238, 'samples': 1373568, 'steps': 7153, 'loss/train': 1.8930798768997192} 01/28/2022 22:02:05 - INFO - codeparrot_training - Step 7154: {'lr': 0.000485910551839666, 'samples': 1373760, 'steps': 7154, 'loss/train': 2.2112302780151367} 01/28/2022 22:02:09 - INFO - codeparrot_training - Step 7155: {'lr': 0.0004859051358938425, 'samples': 1373952, 'steps': 7155, 'loss/train': 0.5774776339530945} 01/28/2022 22:02:13 - INFO - codeparrot_training - Step 7156: {'lr': 0.00048589971893747626, 'samples': 1374144, 'steps': 7156, 'loss/train': 2.190277397632599} 01/28/2022 22:02:18 - INFO - codeparrot_training - Step 7157: {'lr': 0.0004858943009705905, 'samples': 1374336, 'steps': 7157, 'loss/train': 1.9146637916564941} 01/28/2022 22:02:22 - INFO - codeparrot_training - Step 7158: {'lr': 0.00048588888199320847, 'samples': 1374528, 'steps': 7158, 'loss/train': 1.9217051267623901} 01/28/2022 22:02:26 - INFO - codeparrot_training - Step 7159: {'lr': 0.0004858834620053534, 'samples': 1374720, 'steps': 7159, 'loss/train': 1.8947938084602356} 01/28/2022 22:02:32 - INFO - codeparrot_training - Step 7160: {'lr': 0.0004858780410070484, 'samples': 1374912, 'steps': 7160, 'loss/train': 1.535796046257019} 01/28/2022 22:02:36 - INFO - codeparrot_training - Step 7161: {'lr': 0.0004858726189983168, 'samples': 1375104, 'steps': 7161, 'loss/train': 1.1320677995681763} 01/28/2022 22:02:41 - INFO - codeparrot_training - Step 7162: {'lr': 0.00048586719597918185, 'samples': 1375296, 'steps': 7162, 'loss/train': 1.676721453666687} 01/28/2022 22:02:45 - INFO - codeparrot_training - Step 7163: {'lr': 0.0004858617719496667, 'samples': 1375488, 'steps': 7163, 'loss/train': 2.1195536255836487} 01/28/2022 22:02:49 - INFO - codeparrot_training - Step 7164: {'lr': 0.0004858563469097946, 'samples': 1375680, 'steps': 7164, 'loss/train': 2.134684145450592} 01/28/2022 22:02:54 - INFO - codeparrot_training - Step 7165: {'lr': 0.0004858509208595888, 'samples': 1375872, 'steps': 7165, 'loss/train': 2.634005069732666} 01/28/2022 22:02:59 - INFO - codeparrot_training - Step 7166: {'lr': 0.0004858454937990726, 'samples': 1376064, 'steps': 7166, 'loss/train': 2.6452499628067017} 01/28/2022 22:03:03 - INFO - codeparrot_training - Step 7167: {'lr': 0.0004858400657282691, 'samples': 1376256, 'steps': 7167, 'loss/train': 2.0088611841201782} 01/28/2022 22:03:07 - INFO - codeparrot_training - Step 7168: {'lr': 0.00048583463664720174, 'samples': 1376448, 'steps': 7168, 'loss/train': 2.565321922302246} 01/28/2022 22:03:11 - INFO - codeparrot_training - Step 7169: {'lr': 0.00048582920655589366, 'samples': 1376640, 'steps': 7169, 'loss/train': 1.6382803916931152} 01/28/2022 22:03:17 - INFO - codeparrot_training - Step 7170: {'lr': 0.0004858237754543681, 'samples': 1376832, 'steps': 7170, 'loss/train': 2.3748326897621155} 01/28/2022 22:03:21 - INFO - codeparrot_training - Step 7171: {'lr': 0.0004858183433426484, 'samples': 1377024, 'steps': 7171, 'loss/train': 2.102761209011078} 01/28/2022 22:03:25 - INFO - codeparrot_training - Step 7172: {'lr': 0.0004858129102207578, 'samples': 1377216, 'steps': 7172, 'loss/train': 1.4401347041130066} 01/28/2022 22:03:29 - INFO - codeparrot_training - Step 7173: {'lr': 0.00048580747608871955, 'samples': 1377408, 'steps': 7173, 'loss/train': 2.191692531108856} 01/28/2022 22:03:34 - INFO - codeparrot_training - Step 7174: {'lr': 0.000485802040946557, 'samples': 1377600, 'steps': 7174, 'loss/train': 1.321407526731491} 01/28/2022 22:03:39 - INFO - codeparrot_training - Step 7175: {'lr': 0.00048579660479429335, 'samples': 1377792, 'steps': 7175, 'loss/train': 1.5685222148895264} 01/28/2022 22:03:44 - INFO - codeparrot_training - Step 7176: {'lr': 0.00048579116763195184, 'samples': 1377984, 'steps': 7176, 'loss/train': 2.2021263241767883} 01/28/2022 22:03:48 - INFO - codeparrot_training - Step 7177: {'lr': 0.00048578572945955594, 'samples': 1378176, 'steps': 7177, 'loss/train': 0.4799528270959854} 01/28/2022 22:03:52 - INFO - codeparrot_training - Step 7178: {'lr': 0.00048578029027712883, 'samples': 1378368, 'steps': 7178, 'loss/train': 1.682886779308319} 01/28/2022 22:03:56 - INFO - codeparrot_training - Step 7179: {'lr': 0.0004857748500846938, 'samples': 1378560, 'steps': 7179, 'loss/train': 1.5837467908859253} 01/28/2022 22:04:02 - INFO - codeparrot_training - Step 7180: {'lr': 0.0004857694088822742, 'samples': 1378752, 'steps': 7180, 'loss/train': 1.3870718479156494} 01/28/2022 22:04:06 - INFO - codeparrot_training - Step 7181: {'lr': 0.00048576396666989333, 'samples': 1378944, 'steps': 7181, 'loss/train': 2.2491719126701355} 01/28/2022 22:04:10 - INFO - codeparrot_training - Step 7182: {'lr': 0.0004857585234475745, 'samples': 1379136, 'steps': 7182, 'loss/train': 1.5819075107574463} 01/28/2022 22:04:15 - INFO - codeparrot_training - Step 7183: {'lr': 0.00048575307921534095, 'samples': 1379328, 'steps': 7183, 'loss/train': 1.3135530352592468} 01/28/2022 22:04:21 - INFO - codeparrot_training - Step 7184: {'lr': 0.0004857476339732161, 'samples': 1379520, 'steps': 7184, 'loss/train': 1.5424413084983826} 01/28/2022 22:04:25 - INFO - codeparrot_training - Step 7185: {'lr': 0.0004857421877212233, 'samples': 1379712, 'steps': 7185, 'loss/train': 1.8228188753128052} 01/28/2022 22:04:29 - INFO - codeparrot_training - Step 7186: {'lr': 0.00048573674045938577, 'samples': 1379904, 'steps': 7186, 'loss/train': 2.2843170762062073} 01/28/2022 22:04:33 - INFO - codeparrot_training - Step 7187: {'lr': 0.00048573129218772686, 'samples': 1380096, 'steps': 7187, 'loss/train': 7.450794696807861} 01/28/2022 22:04:38 - INFO - codeparrot_training - Step 7188: {'lr': 0.00048572584290627, 'samples': 1380288, 'steps': 7188, 'loss/train': 1.2417068481445312} 01/28/2022 22:04:43 - INFO - codeparrot_training - Step 7189: {'lr': 0.00048572039261503855, 'samples': 1380480, 'steps': 7189, 'loss/train': 1.7667973637580872} 01/28/2022 22:04:47 - INFO - codeparrot_training - Step 7190: {'lr': 0.00048571494131405567, 'samples': 1380672, 'steps': 7190, 'loss/train': 0.8149871528148651} 01/28/2022 22:04:51 - INFO - codeparrot_training - Step 7191: {'lr': 0.0004857094890033449, 'samples': 1380864, 'steps': 7191, 'loss/train': 1.8366943001747131} 01/28/2022 22:04:56 - INFO - codeparrot_training - Step 7192: {'lr': 0.0004857040356829295, 'samples': 1381056, 'steps': 7192, 'loss/train': 2.3628848791122437} 01/28/2022 22:05:00 - INFO - codeparrot_training - Step 7193: {'lr': 0.00048569858135283285, 'samples': 1381248, 'steps': 7193, 'loss/train': 1.978869080543518} 01/28/2022 22:05:05 - INFO - codeparrot_training - Step 7194: {'lr': 0.00048569312601307827, 'samples': 1381440, 'steps': 7194, 'loss/train': 1.4818036258220673} 01/28/2022 22:05:09 - INFO - codeparrot_training - Step 7195: {'lr': 0.00048568766966368925, 'samples': 1381632, 'steps': 7195, 'loss/train': 1.5566743612289429} 01/28/2022 22:05:14 - INFO - codeparrot_training - Step 7196: {'lr': 0.00048568221230468905, 'samples': 1381824, 'steps': 7196, 'loss/train': 1.4876950979232788} 01/28/2022 22:05:18 - INFO - codeparrot_training - Step 7197: {'lr': 0.0004856767539361011, 'samples': 1382016, 'steps': 7197, 'loss/train': 2.947879135608673} 01/28/2022 22:05:22 - INFO - codeparrot_training - Step 7198: {'lr': 0.0004856712945579488, 'samples': 1382208, 'steps': 7198, 'loss/train': 1.9695637822151184} 01/28/2022 22:05:28 - INFO - codeparrot_training - Step 7199: {'lr': 0.00048566583417025553, 'samples': 1382400, 'steps': 7199, 'loss/train': 2.2236024141311646} 01/28/2022 22:05:32 - INFO - codeparrot_training - Step 7200: {'lr': 0.00048566037277304465, 'samples': 1382592, 'steps': 7200, 'loss/train': 1.6526774168014526} 01/28/2022 22:05:36 - INFO - codeparrot_training - Step 7201: {'lr': 0.00048565491036633946, 'samples': 1382784, 'steps': 7201, 'loss/train': 2.0723095536231995} 01/28/2022 22:05:40 - INFO - codeparrot_training - Step 7202: {'lr': 0.00048564944695016356, 'samples': 1382976, 'steps': 7202, 'loss/train': 1.6459890604019165} 01/28/2022 22:05:45 - INFO - codeparrot_training - Step 7203: {'lr': 0.00048564398252454026, 'samples': 1383168, 'steps': 7203, 'loss/train': 5.051412105560303} 01/28/2022 22:05:51 - INFO - codeparrot_training - Step 7204: {'lr': 0.0004856385170894929, 'samples': 1383360, 'steps': 7204, 'loss/train': 1.888017475605011} 01/28/2022 22:05:55 - INFO - codeparrot_training - Step 7205: {'lr': 0.00048563305064504503, 'samples': 1383552, 'steps': 7205, 'loss/train': 2.0024142265319824} 01/28/2022 22:05:59 - INFO - codeparrot_training - Step 7206: {'lr': 0.00048562758319121996, 'samples': 1383744, 'steps': 7206, 'loss/train': 2.4758428931236267} 01/28/2022 22:06:04 - INFO - codeparrot_training - Step 7207: {'lr': 0.00048562211472804115, 'samples': 1383936, 'steps': 7207, 'loss/train': 2.257304012775421} 01/28/2022 22:06:08 - INFO - codeparrot_training - Step 7208: {'lr': 0.000485616645255532, 'samples': 1384128, 'steps': 7208, 'loss/train': 1.6444743275642395} 01/28/2022 22:06:13 - INFO - codeparrot_training - Step 7209: {'lr': 0.00048561117477371595, 'samples': 1384320, 'steps': 7209, 'loss/train': 1.7590066194534302} 01/28/2022 22:06:17 - INFO - codeparrot_training - Step 7210: {'lr': 0.0004856057032826165, 'samples': 1384512, 'steps': 7210, 'loss/train': 1.7763546109199524} 01/28/2022 22:06:22 - INFO - codeparrot_training - Step 7211: {'lr': 0.000485600230782257, 'samples': 1384704, 'steps': 7211, 'loss/train': 1.6747974157333374} 01/28/2022 22:06:26 - INFO - codeparrot_training - Step 7212: {'lr': 0.00048559475727266086, 'samples': 1384896, 'steps': 7212, 'loss/train': 2.3647756576538086} 01/28/2022 22:06:30 - INFO - codeparrot_training - Step 7213: {'lr': 0.00048558928275385167, 'samples': 1385088, 'steps': 7213, 'loss/train': 1.5478352308273315} 01/28/2022 22:06:36 - INFO - codeparrot_training - Step 7214: {'lr': 0.00048558380722585283, 'samples': 1385280, 'steps': 7214, 'loss/train': 1.288204550743103} 01/28/2022 22:06:40 - INFO - codeparrot_training - Step 7215: {'lr': 0.00048557833068868766, 'samples': 1385472, 'steps': 7215, 'loss/train': 2.3347286581993103} 01/28/2022 22:06:45 - INFO - codeparrot_training - Step 7216: {'lr': 0.00048557285314237975, 'samples': 1385664, 'steps': 7216, 'loss/train': 2.4694608449935913} 01/28/2022 22:06:49 - INFO - codeparrot_training - Step 7217: {'lr': 0.0004855673745869526, 'samples': 1385856, 'steps': 7217, 'loss/train': 1.9475850462913513} 01/28/2022 22:06:53 - INFO - codeparrot_training - Step 7218: {'lr': 0.00048556189502242956, 'samples': 1386048, 'steps': 7218, 'loss/train': 2.3936185240745544} 01/28/2022 22:06:57 - INFO - codeparrot_training - Step 7219: {'lr': 0.00048555641444883424, 'samples': 1386240, 'steps': 7219, 'loss/train': 2.1144858598709106} 01/28/2022 22:07:04 - INFO - codeparrot_training - Step 7220: {'lr': 0.00048555093286618996, 'samples': 1386432, 'steps': 7220, 'loss/train': 0.06621174328029156} 01/28/2022 22:07:08 - INFO - codeparrot_training - Step 7221: {'lr': 0.00048554545027452035, 'samples': 1386624, 'steps': 7221, 'loss/train': 2.427358567714691} 01/28/2022 22:07:12 - INFO - codeparrot_training - Step 7222: {'lr': 0.00048553996667384877, 'samples': 1386816, 'steps': 7222, 'loss/train': 2.200589418411255} 01/28/2022 22:07:17 - INFO - codeparrot_training - Step 7223: {'lr': 0.00048553448206419876, 'samples': 1387008, 'steps': 7223, 'loss/train': 1.6658735275268555} 01/28/2022 22:07:21 - INFO - codeparrot_training - Step 7224: {'lr': 0.0004855289964455938, 'samples': 1387200, 'steps': 7224, 'loss/train': 2.0993642807006836} 01/28/2022 22:07:25 - INFO - codeparrot_training - Step 7225: {'lr': 0.0004855235098180575, 'samples': 1387392, 'steps': 7225, 'loss/train': 1.6609252095222473} 01/28/2022 22:07:31 - INFO - codeparrot_training - Step 7226: {'lr': 0.00048551802218161315, 'samples': 1387584, 'steps': 7226, 'loss/train': 1.407860666513443} 01/28/2022 22:07:35 - INFO - codeparrot_training - Step 7227: {'lr': 0.00048551253353628444, 'samples': 1387776, 'steps': 7227, 'loss/train': 1.6751487851142883} 01/28/2022 22:07:39 - INFO - codeparrot_training - Step 7228: {'lr': 0.0004855070438820949, 'samples': 1387968, 'steps': 7228, 'loss/train': 1.6628104448318481} 01/28/2022 22:07:43 - INFO - codeparrot_training - Step 7229: {'lr': 0.0004855015532190679, 'samples': 1388160, 'steps': 7229, 'loss/train': 2.0817553997039795} 01/28/2022 22:07:48 - INFO - codeparrot_training - Step 7230: {'lr': 0.0004854960615472269, 'samples': 1388352, 'steps': 7230, 'loss/train': 0.8061258494853973} 01/28/2022 22:07:54 - INFO - codeparrot_training - Step 7231: {'lr': 0.0004854905688665957, 'samples': 1388544, 'steps': 7231, 'loss/train': 1.3839351832866669} 01/28/2022 22:07:58 - INFO - codeparrot_training - Step 7232: {'lr': 0.00048548507517719766, 'samples': 1388736, 'steps': 7232, 'loss/train': 1.574982762336731} 01/28/2022 22:08:02 - INFO - codeparrot_training - Step 7233: {'lr': 0.00048547958047905635, 'samples': 1388928, 'steps': 7233, 'loss/train': 1.4451613426208496} 01/28/2022 22:08:07 - INFO - codeparrot_training - Step 7234: {'lr': 0.00048547408477219524, 'samples': 1389120, 'steps': 7234, 'loss/train': 1.8920509815216064} 01/28/2022 22:08:11 - INFO - codeparrot_training - Step 7235: {'lr': 0.00048546858805663797, 'samples': 1389312, 'steps': 7235, 'loss/train': 1.5343515872955322} 01/28/2022 22:08:16 - INFO - codeparrot_training - Step 7236: {'lr': 0.000485463090332408, 'samples': 1389504, 'steps': 7236, 'loss/train': 1.8083560466766357} 01/28/2022 22:08:21 - INFO - codeparrot_training - Step 7237: {'lr': 0.0004854575915995289, 'samples': 1389696, 'steps': 7237, 'loss/train': 2.8872488737106323} 01/28/2022 22:08:25 - INFO - codeparrot_training - Step 7238: {'lr': 0.0004854520918580243, 'samples': 1389888, 'steps': 7238, 'loss/train': 1.9828773736953735} 01/28/2022 22:08:29 - INFO - codeparrot_training - Step 7239: {'lr': 0.00048544659110791766, 'samples': 1390080, 'steps': 7239, 'loss/train': 1.988381266593933} 01/28/2022 22:08:34 - INFO - codeparrot_training - Step 7240: {'lr': 0.0004854410893492326, 'samples': 1390272, 'steps': 7240, 'loss/train': 1.545345962047577} 01/28/2022 22:08:39 - INFO - codeparrot_training - Step 7241: {'lr': 0.00048543558658199266, 'samples': 1390464, 'steps': 7241, 'loss/train': 1.9880767464637756} 01/28/2022 22:08:43 - INFO - codeparrot_training - Step 7242: {'lr': 0.0004854300828062215, 'samples': 1390656, 'steps': 7242, 'loss/train': 1.9404845237731934} 01/28/2022 22:08:47 - INFO - codeparrot_training - Step 7243: {'lr': 0.0004854245780219425, 'samples': 1390848, 'steps': 7243, 'loss/train': 2.2095571160316467} 01/28/2022 22:08:51 - INFO - codeparrot_training - Step 7244: {'lr': 0.00048541907222917946, 'samples': 1391040, 'steps': 7244, 'loss/train': 1.7296348214149475} 01/28/2022 22:08:58 - INFO - codeparrot_training - Step 7245: {'lr': 0.0004854135654279558, 'samples': 1391232, 'steps': 7245, 'loss/train': 1.600294053554535} 01/28/2022 22:09:02 - INFO - codeparrot_training - Step 7246: {'lr': 0.0004854080576182952, 'samples': 1391424, 'steps': 7246, 'loss/train': 1.805122196674347} 01/28/2022 22:09:06 - INFO - codeparrot_training - Step 7247: {'lr': 0.00048540254880022126, 'samples': 1391616, 'steps': 7247, 'loss/train': 1.881267786026001} 01/28/2022 22:09:11 - INFO - codeparrot_training - Step 7248: {'lr': 0.00048539703897375753, 'samples': 1391808, 'steps': 7248, 'loss/train': 1.9045223593711853} 01/28/2022 22:09:15 - INFO - codeparrot_training - Step 7249: {'lr': 0.0004853915281389276, 'samples': 1392000, 'steps': 7249, 'loss/train': 1.321014940738678} 01/28/2022 22:09:19 - INFO - codeparrot_training - Step 7250: {'lr': 0.0004853860162957552, 'samples': 1392192, 'steps': 7250, 'loss/train': 0.7304272055625916} 01/28/2022 22:09:25 - INFO - codeparrot_training - Step 7251: {'lr': 0.00048538050344426375, 'samples': 1392384, 'steps': 7251, 'loss/train': 1.9712174534797668} 01/28/2022 22:09:29 - INFO - codeparrot_training - Step 7252: {'lr': 0.0004853749895844771, 'samples': 1392576, 'steps': 7252, 'loss/train': 1.449680507183075} 01/28/2022 22:09:33 - INFO - codeparrot_training - Step 7253: {'lr': 0.00048536947471641855, 'samples': 1392768, 'steps': 7253, 'loss/train': 2.125328779220581} 01/28/2022 22:09:37 - INFO - codeparrot_training - Step 7254: {'lr': 0.00048536395884011207, 'samples': 1392960, 'steps': 7254, 'loss/train': 4.405300140380859} 01/28/2022 22:09:42 - INFO - codeparrot_training - Step 7255: {'lr': 0.00048535844195558104, 'samples': 1393152, 'steps': 7255, 'loss/train': 2.461840867996216} 01/28/2022 22:09:47 - INFO - codeparrot_training - Step 7256: {'lr': 0.0004853529240628493, 'samples': 1393344, 'steps': 7256, 'loss/train': 1.6640517711639404} 01/28/2022 22:09:51 - INFO - codeparrot_training - Step 7257: {'lr': 0.0004853474051619402, 'samples': 1393536, 'steps': 7257, 'loss/train': 1.928017258644104} 01/28/2022 22:09:55 - INFO - codeparrot_training - Step 7258: {'lr': 0.0004853418852528776, 'samples': 1393728, 'steps': 7258, 'loss/train': 7.106718063354492} 01/28/2022 22:09:59 - INFO - codeparrot_training - Step 7259: {'lr': 0.00048533636433568505, 'samples': 1393920, 'steps': 7259, 'loss/train': 2.264686346054077} 01/28/2022 22:10:04 - INFO - codeparrot_training - Step 7260: {'lr': 0.00048533084241038637, 'samples': 1394112, 'steps': 7260, 'loss/train': 1.3505584001541138} 01/28/2022 22:10:09 - INFO - codeparrot_training - Step 7261: {'lr': 0.00048532531947700496, 'samples': 1394304, 'steps': 7261, 'loss/train': 1.6925366520881653} 01/28/2022 22:10:13 - INFO - codeparrot_training - Step 7262: {'lr': 0.00048531979553556473, 'samples': 1394496, 'steps': 7262, 'loss/train': 1.33279350399971} 01/28/2022 22:10:18 - INFO - codeparrot_training - Step 7263: {'lr': 0.0004853142705860891, 'samples': 1394688, 'steps': 7263, 'loss/train': 1.682803988456726} 01/28/2022 22:10:22 - INFO - codeparrot_training - Step 7264: {'lr': 0.00048530874462860194, 'samples': 1394880, 'steps': 7264, 'loss/train': 1.7415960431098938} 01/28/2022 22:10:26 - INFO - codeparrot_training - Step 7265: {'lr': 0.0004853032176631268, 'samples': 1395072, 'steps': 7265, 'loss/train': 1.8573684096336365} 01/28/2022 22:10:33 - INFO - codeparrot_training - Step 7266: {'lr': 0.0004852976896896874, 'samples': 1395264, 'steps': 7266, 'loss/train': 2.0858808159828186} 01/28/2022 22:10:37 - INFO - codeparrot_training - Step 7267: {'lr': 0.0004852921607083074, 'samples': 1395456, 'steps': 7267, 'loss/train': 1.565801203250885} 01/28/2022 22:10:41 - INFO - codeparrot_training - Step 7268: {'lr': 0.00048528663071901047, 'samples': 1395648, 'steps': 7268, 'loss/train': 1.5315744280815125} 01/28/2022 22:10:45 - INFO - codeparrot_training - Step 7269: {'lr': 0.00048528109972182043, 'samples': 1395840, 'steps': 7269, 'loss/train': 2.388579547405243} 01/28/2022 22:10:50 - INFO - codeparrot_training - Step 7270: {'lr': 0.0004852755677167607, 'samples': 1396032, 'steps': 7270, 'loss/train': 2.026220977306366} 01/28/2022 22:10:55 - INFO - codeparrot_training - Step 7271: {'lr': 0.00048527003470385534, 'samples': 1396224, 'steps': 7271, 'loss/train': 2.579102575778961} 01/28/2022 22:10:59 - INFO - codeparrot_training - Step 7272: {'lr': 0.0004852645006831278, 'samples': 1396416, 'steps': 7272, 'loss/train': 1.2895922362804413} 01/28/2022 22:11:03 - INFO - codeparrot_training - Step 7273: {'lr': 0.00048525896565460177, 'samples': 1396608, 'steps': 7273, 'loss/train': 1.038468450307846} 01/28/2022 22:11:08 - INFO - codeparrot_training - Step 7274: {'lr': 0.00048525342961830106, 'samples': 1396800, 'steps': 7274, 'loss/train': 1.137799233198166} 01/28/2022 22:11:12 - INFO - codeparrot_training - Step 7275: {'lr': 0.0004852478925742494, 'samples': 1396992, 'steps': 7275, 'loss/train': 1.823871374130249} 01/28/2022 22:11:18 - INFO - codeparrot_training - Step 7276: {'lr': 0.0004852423545224704, 'samples': 1397184, 'steps': 7276, 'loss/train': 1.4384577870368958} 01/28/2022 22:11:22 - INFO - codeparrot_training - Step 7277: {'lr': 0.00048523681546298793, 'samples': 1397376, 'steps': 7277, 'loss/train': 1.5279105305671692} 01/28/2022 22:11:26 - INFO - codeparrot_training - Step 7278: {'lr': 0.0004852312753958256, 'samples': 1397568, 'steps': 7278, 'loss/train': 1.9813562035560608} 01/28/2022 22:11:30 - INFO - codeparrot_training - Step 7279: {'lr': 0.00048522573432100715, 'samples': 1397760, 'steps': 7279, 'loss/train': 2.049864113330841} 01/28/2022 22:11:35 - INFO - codeparrot_training - Step 7280: {'lr': 0.0004852201922385564, 'samples': 1397952, 'steps': 7280, 'loss/train': 1.4113108813762665} 01/28/2022 22:11:40 - INFO - codeparrot_training - Step 7281: {'lr': 0.000485214649148497, 'samples': 1398144, 'steps': 7281, 'loss/train': 1.781259298324585} 01/28/2022 22:11:44 - INFO - codeparrot_training - Step 7282: {'lr': 0.00048520910505085274, 'samples': 1398336, 'steps': 7282, 'loss/train': 1.8181332349777222} 01/28/2022 22:11:49 - INFO - codeparrot_training - Step 7283: {'lr': 0.0004852035599456474, 'samples': 1398528, 'steps': 7283, 'loss/train': 2.132734715938568} 01/28/2022 22:11:53 - INFO - codeparrot_training - Step 7284: {'lr': 0.0004851980138329046, 'samples': 1398720, 'steps': 7284, 'loss/train': 1.6520435214042664} 01/28/2022 22:11:57 - INFO - codeparrot_training - Step 7285: {'lr': 0.00048519246671264825, 'samples': 1398912, 'steps': 7285, 'loss/train': 1.9489785432815552} 01/28/2022 22:12:03 - INFO - codeparrot_training - Step 7286: {'lr': 0.0004851869185849021, 'samples': 1399104, 'steps': 7286, 'loss/train': 1.8516032695770264} 01/28/2022 22:12:07 - INFO - codeparrot_training - Step 7287: {'lr': 0.0004851813694496898, 'samples': 1399296, 'steps': 7287, 'loss/train': 1.3110092282295227} 01/28/2022 22:12:11 - INFO - codeparrot_training - Step 7288: {'lr': 0.00048517581930703526, 'samples': 1399488, 'steps': 7288, 'loss/train': 2.1962292194366455} 01/28/2022 22:12:16 - INFO - codeparrot_training - Step 7289: {'lr': 0.0004851702681569621, 'samples': 1399680, 'steps': 7289, 'loss/train': 1.3914101421833038} 01/28/2022 22:12:20 - INFO - codeparrot_training - Step 7290: {'lr': 0.0004851647159994943, 'samples': 1399872, 'steps': 7290, 'loss/train': 1.3145704865455627} 01/28/2022 22:12:25 - INFO - codeparrot_training - Step 7291: {'lr': 0.00048515916283465546, 'samples': 1400064, 'steps': 7291, 'loss/train': 1.9734594225883484} 01/28/2022 22:12:29 - INFO - codeparrot_training - Step 7292: {'lr': 0.00048515360866246943, 'samples': 1400256, 'steps': 7292, 'loss/train': 1.6796016097068787} 01/28/2022 22:12:34 - INFO - codeparrot_training - Step 7293: {'lr': 0.00048514805348296, 'samples': 1400448, 'steps': 7293, 'loss/train': 2.469951331615448} 01/28/2022 22:12:38 - INFO - codeparrot_training - Step 7294: {'lr': 0.000485142497296151, 'samples': 1400640, 'steps': 7294, 'loss/train': 1.3300376236438751} 01/28/2022 22:12:42 - INFO - codeparrot_training - Step 7295: {'lr': 0.00048513694010206623, 'samples': 1400832, 'steps': 7295, 'loss/train': 1.8883459568023682} 01/28/2022 22:12:48 - INFO - codeparrot_training - Step 7296: {'lr': 0.0004851313819007295, 'samples': 1401024, 'steps': 7296, 'loss/train': 2.227198898792267} 01/28/2022 22:12:52 - INFO - codeparrot_training - Step 7297: {'lr': 0.0004851258226921645, 'samples': 1401216, 'steps': 7297, 'loss/train': 2.1870553493499756} 01/28/2022 22:12:56 - INFO - codeparrot_training - Step 7298: {'lr': 0.0004851202624763952, 'samples': 1401408, 'steps': 7298, 'loss/train': 2.1217387318611145} 01/28/2022 22:13:00 - INFO - codeparrot_training - Step 7299: {'lr': 0.0004851147012534453, 'samples': 1401600, 'steps': 7299, 'loss/train': 1.7337468266487122} 01/28/2022 22:13:05 - INFO - codeparrot_training - Step 7300: {'lr': 0.00048510913902333875, 'samples': 1401792, 'steps': 7300, 'loss/train': 2.4790141582489014} 01/28/2022 22:13:10 - INFO - codeparrot_training - Step 7301: {'lr': 0.0004851035757860992, 'samples': 1401984, 'steps': 7301, 'loss/train': 1.437045156955719} 01/28/2022 22:13:15 - INFO - codeparrot_training - Step 7302: {'lr': 0.0004850980115417507, 'samples': 1402176, 'steps': 7302, 'loss/train': 2.0269371271133423} 01/28/2022 22:13:19 - INFO - codeparrot_training - Step 7303: {'lr': 0.0004850924462903169, 'samples': 1402368, 'steps': 7303, 'loss/train': 2.004817485809326} 01/28/2022 22:13:23 - INFO - codeparrot_training - Step 7304: {'lr': 0.0004850868800318218, 'samples': 1402560, 'steps': 7304, 'loss/train': 2.254274368286133} 01/28/2022 22:13:28 - INFO - codeparrot_training - Step 7305: {'lr': 0.00048508131276628905, 'samples': 1402752, 'steps': 7305, 'loss/train': 0.814289778470993} 01/28/2022 22:13:33 - INFO - codeparrot_training - Step 7306: {'lr': 0.0004850757444937426, 'samples': 1402944, 'steps': 7306, 'loss/train': 1.4574404954910278} 01/28/2022 22:13:37 - INFO - codeparrot_training - Step 7307: {'lr': 0.00048507017521420636, 'samples': 1403136, 'steps': 7307, 'loss/train': 1.8944304585456848} 01/28/2022 22:13:41 - INFO - codeparrot_training - Step 7308: {'lr': 0.0004850646049277041, 'samples': 1403328, 'steps': 7308, 'loss/train': 2.421739876270294} 01/28/2022 22:13:46 - INFO - codeparrot_training - Step 7309: {'lr': 0.00048505903363425974, 'samples': 1403520, 'steps': 7309, 'loss/train': 1.3645992279052734} 01/28/2022 22:13:50 - INFO - codeparrot_training - Step 7310: {'lr': 0.0004850534613338972, 'samples': 1403712, 'steps': 7310, 'loss/train': 1.6340113878250122} 01/28/2022 22:13:55 - INFO - codeparrot_training - Step 7311: {'lr': 0.00048504788802664013, 'samples': 1403904, 'steps': 7311, 'loss/train': 1.4596776366233826} 01/28/2022 22:13:59 - INFO - codeparrot_training - Step 7312: {'lr': 0.00048504231371251255, 'samples': 1404096, 'steps': 7312, 'loss/train': 1.027408629655838} 01/28/2022 22:14:04 - INFO - codeparrot_training - Step 7313: {'lr': 0.0004850367383915384, 'samples': 1404288, 'steps': 7313, 'loss/train': 2.4257469177246094} 01/28/2022 22:14:08 - INFO - codeparrot_training - Step 7314: {'lr': 0.00048503116206374147, 'samples': 1404480, 'steps': 7314, 'loss/train': 2.0524717569351196} 01/28/2022 22:14:12 - INFO - codeparrot_training - Step 7315: {'lr': 0.00048502558472914573, 'samples': 1404672, 'steps': 7315, 'loss/train': 2.2179940938949585} 01/28/2022 22:14:18 - INFO - codeparrot_training - Step 7316: {'lr': 0.00048502000638777487, 'samples': 1404864, 'steps': 7316, 'loss/train': 0.5614043623209} 01/28/2022 22:14:22 - INFO - codeparrot_training - Step 7317: {'lr': 0.000485014427039653, 'samples': 1405056, 'steps': 7317, 'loss/train': 1.713200032711029} 01/28/2022 22:14:27 - INFO - codeparrot_training - Step 7318: {'lr': 0.00048500884668480407, 'samples': 1405248, 'steps': 7318, 'loss/train': 1.6405202150344849} 01/28/2022 22:14:31 - INFO - codeparrot_training - Step 7319: {'lr': 0.00048500326532325167, 'samples': 1405440, 'steps': 7319, 'loss/train': 2.4586468935012817} 01/28/2022 22:14:35 - INFO - codeparrot_training - Step 7320: {'lr': 0.00048499768295502, 'samples': 1405632, 'steps': 7320, 'loss/train': 1.2086463868618011} 01/28/2022 22:14:41 - INFO - codeparrot_training - Step 7321: {'lr': 0.0004849920995801329, 'samples': 1405824, 'steps': 7321, 'loss/train': 2.560108244419098} 01/28/2022 22:14:45 - INFO - codeparrot_training - Step 7322: {'lr': 0.00048498651519861426, 'samples': 1406016, 'steps': 7322, 'loss/train': 1.7587299942970276} 01/28/2022 22:14:49 - INFO - codeparrot_training - Step 7323: {'lr': 0.00048498092981048797, 'samples': 1406208, 'steps': 7323, 'loss/train': 1.8633370399475098} 01/28/2022 22:14:54 - INFO - codeparrot_training - Step 7324: {'lr': 0.000484975343415778, 'samples': 1406400, 'steps': 7324, 'loss/train': 1.9784950017929077} 01/28/2022 22:14:58 - INFO - codeparrot_training - Step 7325: {'lr': 0.00048496975601450835, 'samples': 1406592, 'steps': 7325, 'loss/train': 1.9648271799087524} 01/28/2022 22:15:04 - INFO - codeparrot_training - Step 7326: {'lr': 0.0004849641676067027, 'samples': 1406784, 'steps': 7326, 'loss/train': 0.15263549610972404} 01/28/2022 22:15:08 - INFO - codeparrot_training - Step 7327: {'lr': 0.0004849585781923853, 'samples': 1406976, 'steps': 7327, 'loss/train': 1.057133138179779} 01/28/2022 22:15:12 - INFO - codeparrot_training - Step 7328: {'lr': 0.00048495298777157994, 'samples': 1407168, 'steps': 7328, 'loss/train': 1.8456303477287292} 01/28/2022 22:15:16 - INFO - codeparrot_training - Step 7329: {'lr': 0.00048494739634431057, 'samples': 1407360, 'steps': 7329, 'loss/train': 3.236833691596985} 01/28/2022 22:15:20 - INFO - codeparrot_training - Step 7330: {'lr': 0.00048494180391060114, 'samples': 1407552, 'steps': 7330, 'loss/train': 1.540281057357788} 01/28/2022 22:15:26 - INFO - codeparrot_training - Step 7331: {'lr': 0.0004849362104704756, 'samples': 1407744, 'steps': 7331, 'loss/train': 2.0997143983840942} 01/28/2022 22:15:30 - INFO - codeparrot_training - Step 7332: {'lr': 0.00048493061602395803, 'samples': 1407936, 'steps': 7332, 'loss/train': 1.6050787568092346} 01/28/2022 22:15:34 - INFO - codeparrot_training - Step 7333: {'lr': 0.0004849250205710722, 'samples': 1408128, 'steps': 7333, 'loss/train': 1.994981825351715} 01/28/2022 22:15:38 - INFO - codeparrot_training - Step 7334: {'lr': 0.0004849194241118423, 'samples': 1408320, 'steps': 7334, 'loss/train': 0.9108814001083374} 01/28/2022 22:15:43 - INFO - codeparrot_training - Step 7335: {'lr': 0.0004849138266462921, 'samples': 1408512, 'steps': 7335, 'loss/train': 2.0776652097702026} 01/28/2022 22:15:49 - INFO - codeparrot_training - Step 7336: {'lr': 0.0004849082281744457, 'samples': 1408704, 'steps': 7336, 'loss/train': 1.9485484957695007} 01/28/2022 22:15:53 - INFO - codeparrot_training - Step 7337: {'lr': 0.00048490262869632693, 'samples': 1408896, 'steps': 7337, 'loss/train': 1.3633893728256226} 01/28/2022 22:15:57 - INFO - codeparrot_training - Step 7338: {'lr': 0.00048489702821196003, 'samples': 1409088, 'steps': 7338, 'loss/train': 1.6258655190467834} 01/28/2022 22:16:01 - INFO - codeparrot_training - Step 7339: {'lr': 0.0004848914267213688, 'samples': 1409280, 'steps': 7339, 'loss/train': 1.803956687450409} 01/28/2022 22:16:06 - INFO - codeparrot_training - Step 7340: {'lr': 0.00048488582422457726, 'samples': 1409472, 'steps': 7340, 'loss/train': 1.4189373850822449} 01/28/2022 22:16:11 - INFO - codeparrot_training - Step 7341: {'lr': 0.0004848802207216094, 'samples': 1409664, 'steps': 7341, 'loss/train': 1.2891477942466736} 01/28/2022 22:16:15 - INFO - codeparrot_training - Step 7342: {'lr': 0.0004848746162124894, 'samples': 1409856, 'steps': 7342, 'loss/train': 0.9317125082015991} 01/28/2022 22:16:19 - INFO - codeparrot_training - Step 7343: {'lr': 0.00048486901069724097, 'samples': 1410048, 'steps': 7343, 'loss/train': 1.8937901258468628} 01/28/2022 22:16:24 - INFO - codeparrot_training - Step 7344: {'lr': 0.0004848634041758884, 'samples': 1410240, 'steps': 7344, 'loss/train': 1.0949013233184814} 01/28/2022 22:16:28 - INFO - codeparrot_training - Step 7345: {'lr': 0.00048485779664845553, 'samples': 1410432, 'steps': 7345, 'loss/train': 2.15058034658432} 01/28/2022 22:16:34 - INFO - codeparrot_training - Step 7346: {'lr': 0.0004848521881149664, 'samples': 1410624, 'steps': 7346, 'loss/train': 1.517503023147583} 01/28/2022 22:16:38 - INFO - codeparrot_training - Step 7347: {'lr': 0.00048484657857544513, 'samples': 1410816, 'steps': 7347, 'loss/train': 1.9373023509979248} 01/28/2022 22:16:43 - INFO - codeparrot_training - Step 7348: {'lr': 0.0004848409680299156, 'samples': 1411008, 'steps': 7348, 'loss/train': 1.6165897250175476} 01/28/2022 22:16:47 - INFO - codeparrot_training - Step 7349: {'lr': 0.00048483535647840206, 'samples': 1411200, 'steps': 7349, 'loss/train': 1.2651565968990326} 01/28/2022 22:16:52 - INFO - codeparrot_training - Step 7350: {'lr': 0.00048482974392092827, 'samples': 1411392, 'steps': 7350, 'loss/train': 0.7122036516666412} 01/28/2022 22:16:56 - INFO - codeparrot_training - Step 7351: {'lr': 0.0004848241303575185, 'samples': 1411584, 'steps': 7351, 'loss/train': 1.4859498739242554} 01/28/2022 22:17:01 - INFO - codeparrot_training - Step 7352: {'lr': 0.0004848185157881968, 'samples': 1411776, 'steps': 7352, 'loss/train': 1.01617830991745} 01/28/2022 22:17:05 - INFO - codeparrot_training - Step 7353: {'lr': 0.0004848129002129871, 'samples': 1411968, 'steps': 7353, 'loss/train': 0.41811226308345795} 01/28/2022 22:17:09 - INFO - codeparrot_training - Step 7354: {'lr': 0.0004848072836319134, 'samples': 1412160, 'steps': 7354, 'loss/train': 1.3520253002643585} 01/28/2022 22:17:14 - INFO - codeparrot_training - Step 7355: {'lr': 0.000484801666045, 'samples': 1412352, 'steps': 7355, 'loss/train': 1.7908431887626648} 01/28/2022 22:17:19 - INFO - codeparrot_training - Step 7356: {'lr': 0.0004847960474522707, 'samples': 1412544, 'steps': 7356, 'loss/train': 7.32532525062561} 01/28/2022 22:17:23 - INFO - codeparrot_training - Step 7357: {'lr': 0.00048479042785374974, 'samples': 1412736, 'steps': 7357, 'loss/train': 2.37112820148468} 01/28/2022 22:17:27 - INFO - codeparrot_training - Step 7358: {'lr': 0.0004847848072494611, 'samples': 1412928, 'steps': 7358, 'loss/train': 0.7590128481388092} 01/28/2022 22:17:31 - INFO - codeparrot_training - Step 7359: {'lr': 0.0004847791856394289, 'samples': 1413120, 'steps': 7359, 'loss/train': 1.5493571162223816} 01/28/2022 22:17:37 - INFO - codeparrot_training - Step 7360: {'lr': 0.00048477356302367724, 'samples': 1413312, 'steps': 7360, 'loss/train': 1.7893280982971191} 01/28/2022 22:17:42 - INFO - codeparrot_training - Step 7361: {'lr': 0.00048476793940223026, 'samples': 1413504, 'steps': 7361, 'loss/train': 2.0786109566688538} 01/28/2022 22:17:46 - INFO - codeparrot_training - Step 7362: {'lr': 0.0004847623147751119, 'samples': 1413696, 'steps': 7362, 'loss/train': 2.2182002663612366} 01/28/2022 22:17:50 - INFO - codeparrot_training - Step 7363: {'lr': 0.00048475668914234636, 'samples': 1413888, 'steps': 7363, 'loss/train': 2.075882613658905} 01/28/2022 22:17:54 - INFO - codeparrot_training - Step 7364: {'lr': 0.0004847510625039577, 'samples': 1414080, 'steps': 7364, 'loss/train': 1.8669251203536987} 01/28/2022 22:18:00 - INFO - codeparrot_training - Step 7365: {'lr': 0.00048474543485997005, 'samples': 1414272, 'steps': 7365, 'loss/train': 1.4952559769153595} 01/28/2022 22:18:04 - INFO - codeparrot_training - Step 7366: {'lr': 0.00048473980621040744, 'samples': 1414464, 'steps': 7366, 'loss/train': 1.893983781337738} 01/28/2022 22:18:08 - INFO - codeparrot_training - Step 7367: {'lr': 0.00048473417655529405, 'samples': 1414656, 'steps': 7367, 'loss/train': 2.021942675113678} 01/28/2022 22:18:12 - INFO - codeparrot_training - Step 7368: {'lr': 0.000484728545894654, 'samples': 1414848, 'steps': 7368, 'loss/train': 1.8956192135810852} 01/28/2022 22:18:17 - INFO - codeparrot_training - Step 7369: {'lr': 0.00048472291422851135, 'samples': 1415040, 'steps': 7369, 'loss/train': 2.622327983379364} 01/28/2022 22:18:23 - INFO - codeparrot_training - Step 7370: {'lr': 0.00048471728155689034, 'samples': 1415232, 'steps': 7370, 'loss/train': 2.2352548241615295} 01/28/2022 22:18:27 - INFO - codeparrot_training - Step 7371: {'lr': 0.000484711647879815, 'samples': 1415424, 'steps': 7371, 'loss/train': 1.8163467049598694} 01/28/2022 22:18:31 - INFO - codeparrot_training - Step 7372: {'lr': 0.00048470601319730946, 'samples': 1415616, 'steps': 7372, 'loss/train': 1.87450110912323} 01/28/2022 22:18:35 - INFO - codeparrot_training - Step 7373: {'lr': 0.00048470037750939795, 'samples': 1415808, 'steps': 7373, 'loss/train': 0.6259899884462357} 01/28/2022 22:18:40 - INFO - codeparrot_training - Step 7374: {'lr': 0.0004846947408161045, 'samples': 1416000, 'steps': 7374, 'loss/train': 2.117339551448822} 01/28/2022 22:18:45 - INFO - codeparrot_training - Step 7375: {'lr': 0.0004846891031174533, 'samples': 1416192, 'steps': 7375, 'loss/train': 1.7243297696113586} 01/28/2022 22:18:49 - INFO - codeparrot_training - Step 7376: {'lr': 0.00048468346441346853, 'samples': 1416384, 'steps': 7376, 'loss/train': 1.3961366415023804} 01/28/2022 22:18:53 - INFO - codeparrot_training - Step 7377: {'lr': 0.00048467782470417434, 'samples': 1416576, 'steps': 7377, 'loss/train': 1.5589848160743713} 01/28/2022 22:18:58 - INFO - codeparrot_training - Step 7378: {'lr': 0.0004846721839895948, 'samples': 1416768, 'steps': 7378, 'loss/train': 1.958004891872406} 01/28/2022 22:19:02 - INFO - codeparrot_training - Step 7379: {'lr': 0.00048466654226975414, 'samples': 1416960, 'steps': 7379, 'loss/train': 1.3537148237228394} 01/28/2022 22:19:07 - INFO - codeparrot_training - Step 7380: {'lr': 0.00048466089954467663, 'samples': 1417152, 'steps': 7380, 'loss/train': 2.054919719696045} 01/28/2022 22:19:11 - INFO - codeparrot_training - Step 7381: {'lr': 0.0004846552558143863, 'samples': 1417344, 'steps': 7381, 'loss/train': 1.6499311923980713} 01/28/2022 22:19:16 - INFO - codeparrot_training - Step 7382: {'lr': 0.00048464961107890734, 'samples': 1417536, 'steps': 7382, 'loss/train': 1.3603270053863525} 01/28/2022 22:19:20 - INFO - codeparrot_training - Step 7383: {'lr': 0.00048464396533826396, 'samples': 1417728, 'steps': 7383, 'loss/train': 0.8966135680675507} 01/28/2022 22:19:24 - INFO - codeparrot_training - Step 7384: {'lr': 0.0004846383185924803, 'samples': 1417920, 'steps': 7384, 'loss/train': 1.2502878606319427} 01/28/2022 22:19:30 - INFO - codeparrot_training - Step 7385: {'lr': 0.0004846326708415806, 'samples': 1418112, 'steps': 7385, 'loss/train': 1.7012311220169067} 01/28/2022 22:19:34 - INFO - codeparrot_training - Step 7386: {'lr': 0.00048462702208558906, 'samples': 1418304, 'steps': 7386, 'loss/train': 2.393767833709717} 01/28/2022 22:19:38 - INFO - codeparrot_training - Step 7387: {'lr': 0.0004846213723245299, 'samples': 1418496, 'steps': 7387, 'loss/train': 2.02734375} 01/28/2022 22:19:42 - INFO - codeparrot_training - Step 7388: {'lr': 0.00048461572155842725, 'samples': 1418688, 'steps': 7388, 'loss/train': 2.2270352840423584} 01/28/2022 22:19:46 - INFO - codeparrot_training - Step 7389: {'lr': 0.0004846100697873054, 'samples': 1418880, 'steps': 7389, 'loss/train': 1.690770149230957} 01/28/2022 22:19:52 - INFO - codeparrot_training - Step 7390: {'lr': 0.0004846044170111884, 'samples': 1419072, 'steps': 7390, 'loss/train': 0.8018357455730438} 01/28/2022 22:19:57 - INFO - codeparrot_training - Step 7391: {'lr': 0.00048459876323010063, 'samples': 1419264, 'steps': 7391, 'loss/train': 1.2935790717601776} 01/28/2022 22:20:01 - INFO - codeparrot_training - Step 7392: {'lr': 0.00048459310844406624, 'samples': 1419456, 'steps': 7392, 'loss/train': 2.261606276035309} 01/28/2022 22:20:05 - INFO - codeparrot_training - Step 7393: {'lr': 0.0004845874526531095, 'samples': 1419648, 'steps': 7393, 'loss/train': 2.1813769340515137} 01/28/2022 22:20:10 - INFO - codeparrot_training - Step 7394: {'lr': 0.0004845817958572546, 'samples': 1419840, 'steps': 7394, 'loss/train': 1.7180360555648804} 01/28/2022 22:20:15 - INFO - codeparrot_training - Step 7395: {'lr': 0.0004845761380565257, 'samples': 1420032, 'steps': 7395, 'loss/train': 0.26878997683525085} 01/28/2022 22:20:19 - INFO - codeparrot_training - Step 7396: {'lr': 0.0004845704792509472, 'samples': 1420224, 'steps': 7396, 'loss/train': 1.7861441373825073} 01/28/2022 22:20:23 - INFO - codeparrot_training - Step 7397: {'lr': 0.0004845648194405432, 'samples': 1420416, 'steps': 7397, 'loss/train': 1.1806467175483704} 01/28/2022 22:20:27 - INFO - codeparrot_training - Step 7398: {'lr': 0.00048455915862533804, 'samples': 1420608, 'steps': 7398, 'loss/train': 1.9956997632980347} 01/28/2022 22:20:32 - INFO - codeparrot_training - Step 7399: {'lr': 0.0004845534968053559, 'samples': 1420800, 'steps': 7399, 'loss/train': 1.8377584218978882} 01/28/2022 22:20:37 - INFO - codeparrot_training - Step 7400: {'lr': 0.0004845478339806211, 'samples': 1420992, 'steps': 7400, 'loss/train': 1.8387256264686584} 01/28/2022 22:20:41 - INFO - codeparrot_training - Step 7401: {'lr': 0.0004845421701511578, 'samples': 1421184, 'steps': 7401, 'loss/train': 2.4193623661994934} 01/28/2022 22:20:46 - INFO - codeparrot_training - Step 7402: {'lr': 0.0004845365053169903, 'samples': 1421376, 'steps': 7402, 'loss/train': 2.0631378293037415} 01/28/2022 22:20:50 - INFO - codeparrot_training - Step 7403: {'lr': 0.0004845308394781429, 'samples': 1421568, 'steps': 7403, 'loss/train': 2.547844648361206} 01/28/2022 22:20:54 - INFO - codeparrot_training - Step 7404: {'lr': 0.0004845251726346399, 'samples': 1421760, 'steps': 7404, 'loss/train': 1.571685791015625} 01/28/2022 22:21:00 - INFO - codeparrot_training - Step 7405: {'lr': 0.0004845195047865055, 'samples': 1421952, 'steps': 7405, 'loss/train': 1.5617124438285828} 01/28/2022 22:21:04 - INFO - codeparrot_training - Step 7406: {'lr': 0.00048451383593376394, 'samples': 1422144, 'steps': 7406, 'loss/train': 0.9629844725131989} 01/28/2022 22:21:08 - INFO - codeparrot_training - Step 7407: {'lr': 0.0004845081660764397, 'samples': 1422336, 'steps': 7407, 'loss/train': 2.1559168696403503} 01/28/2022 22:21:13 - INFO - codeparrot_training - Step 7408: {'lr': 0.0004845024952145569, 'samples': 1422528, 'steps': 7408, 'loss/train': 1.8898013234138489} 01/28/2022 22:21:17 - INFO - codeparrot_training - Step 7409: {'lr': 0.00048449682334813983, 'samples': 1422720, 'steps': 7409, 'loss/train': 1.3699261844158173} 01/28/2022 22:21:23 - INFO - codeparrot_training - Step 7410: {'lr': 0.00048449115047721286, 'samples': 1422912, 'steps': 7410, 'loss/train': 2.0004780292510986} 01/28/2022 22:21:27 - INFO - codeparrot_training - Step 7411: {'lr': 0.00048448547660180034, 'samples': 1423104, 'steps': 7411, 'loss/train': 0.1366470493376255} 01/28/2022 22:21:32 - INFO - codeparrot_training - Step 7412: {'lr': 0.0004844798017219264, 'samples': 1423296, 'steps': 7412, 'loss/train': 2.1586897373199463} 01/28/2022 22:21:36 - INFO - codeparrot_training - Step 7413: {'lr': 0.00048447412583761543, 'samples': 1423488, 'steps': 7413, 'loss/train': 1.9170091152191162} 01/28/2022 22:21:40 - INFO - codeparrot_training - Step 7414: {'lr': 0.00048446844894889173, 'samples': 1423680, 'steps': 7414, 'loss/train': 0.9147261679172516} 01/28/2022 22:21:44 - INFO - codeparrot_training - Step 7415: {'lr': 0.00048446277105577973, 'samples': 1423872, 'steps': 7415, 'loss/train': 1.5076754093170166} 01/28/2022 22:21:51 - INFO - codeparrot_training - Step 7416: {'lr': 0.0004844570921583037, 'samples': 1424064, 'steps': 7416, 'loss/train': 2.093391001224518} 01/28/2022 22:21:55 - INFO - codeparrot_training - Step 7417: {'lr': 0.00048445141225648785, 'samples': 1424256, 'steps': 7417, 'loss/train': 1.3722014129161835} 01/28/2022 22:21:59 - INFO - codeparrot_training - Step 7418: {'lr': 0.00048444573135035665, 'samples': 1424448, 'steps': 7418, 'loss/train': 1.0399615466594696} 01/28/2022 22:22:03 - INFO - codeparrot_training - Step 7419: {'lr': 0.00048444004943993434, 'samples': 1424640, 'steps': 7419, 'loss/train': 0.988579273223877} 01/28/2022 22:22:08 - INFO - codeparrot_training - Step 7420: {'lr': 0.0004844343665252453, 'samples': 1424832, 'steps': 7420, 'loss/train': 0.9731236696243286} 01/28/2022 22:22:13 - INFO - codeparrot_training - Step 7421: {'lr': 0.0004844286826063139, 'samples': 1425024, 'steps': 7421, 'loss/train': 2.1179051399230957} 01/28/2022 22:22:17 - INFO - codeparrot_training - Step 7422: {'lr': 0.0004844229976831645, 'samples': 1425216, 'steps': 7422, 'loss/train': 2.429266333580017} 01/28/2022 22:22:21 - INFO - codeparrot_training - Step 7423: {'lr': 0.00048441731175582136, 'samples': 1425408, 'steps': 7423, 'loss/train': 1.8940960764884949} 01/28/2022 22:22:25 - INFO - codeparrot_training - Step 7424: {'lr': 0.0004844116248243089, 'samples': 1425600, 'steps': 7424, 'loss/train': 1.7724798917770386} 01/28/2022 22:22:31 - INFO - codeparrot_training - Step 7425: {'lr': 0.00048440593688865155, 'samples': 1425792, 'steps': 7425, 'loss/train': 1.5984553098678589} 01/28/2022 22:22:35 - INFO - codeparrot_training - Step 7426: {'lr': 0.0004844002479488735, 'samples': 1425984, 'steps': 7426, 'loss/train': 2.0792938470840454} 01/28/2022 22:22:39 - INFO - codeparrot_training - Step 7427: {'lr': 0.0004843945580049992, 'samples': 1426176, 'steps': 7427, 'loss/train': 1.5905326008796692} 01/28/2022 22:22:43 - INFO - codeparrot_training - Step 7428: {'lr': 0.0004843888670570531, 'samples': 1426368, 'steps': 7428, 'loss/train': 1.7966074347496033} 01/28/2022 22:22:48 - INFO - codeparrot_training - Step 7429: {'lr': 0.00048438317510505954, 'samples': 1426560, 'steps': 7429, 'loss/train': 1.180633306503296} 01/28/2022 22:22:54 - INFO - codeparrot_training - Step 7430: {'lr': 0.0004843774821490429, 'samples': 1426752, 'steps': 7430, 'loss/train': 1.6049346327781677} 01/28/2022 22:22:58 - INFO - codeparrot_training - Step 7431: {'lr': 0.0004843717881890275, 'samples': 1426944, 'steps': 7431, 'loss/train': 1.6988266110420227} 01/28/2022 22:23:02 - INFO - codeparrot_training - Step 7432: {'lr': 0.0004843660932250378, 'samples': 1427136, 'steps': 7432, 'loss/train': 2.105782449245453} 01/28/2022 22:23:06 - INFO - codeparrot_training - Step 7433: {'lr': 0.0004843603972570981, 'samples': 1427328, 'steps': 7433, 'loss/train': 0.15475577116012573} 01/28/2022 22:23:11 - INFO - codeparrot_training - Step 7434: {'lr': 0.00048435470028523295, 'samples': 1427520, 'steps': 7434, 'loss/train': 1.7810068130493164} 01/28/2022 22:23:16 - INFO - codeparrot_training - Step 7435: {'lr': 0.00048434900230946666, 'samples': 1427712, 'steps': 7435, 'loss/train': 1.6312009692192078} 01/28/2022 22:23:21 - INFO - codeparrot_training - Step 7436: {'lr': 0.0004843433033298237, 'samples': 1427904, 'steps': 7436, 'loss/train': 2.6908058524131775} 01/28/2022 22:23:25 - INFO - codeparrot_training - Step 7437: {'lr': 0.00048433760334632835, 'samples': 1428096, 'steps': 7437, 'loss/train': 2.5062578916549683} 01/28/2022 22:23:29 - INFO - codeparrot_training - Step 7438: {'lr': 0.0004843319023590052, 'samples': 1428288, 'steps': 7438, 'loss/train': 0.6011989116668701} 01/28/2022 22:23:33 - INFO - codeparrot_training - Step 7439: {'lr': 0.0004843262003678786, 'samples': 1428480, 'steps': 7439, 'loss/train': 1.8243640065193176} 01/28/2022 22:23:39 - INFO - codeparrot_training - Step 7440: {'lr': 0.0004843204973729729, 'samples': 1428672, 'steps': 7440, 'loss/train': 1.745950162410736} 01/28/2022 22:23:43 - INFO - codeparrot_training - Step 7441: {'lr': 0.0004843147933743126, 'samples': 1428864, 'steps': 7441, 'loss/train': 1.4317364394664764} 01/28/2022 22:23:47 - INFO - codeparrot_training - Step 7442: {'lr': 0.0004843090883719222, 'samples': 1429056, 'steps': 7442, 'loss/train': 1.1217006146907806} 01/28/2022 22:23:51 - INFO - codeparrot_training - Step 7443: {'lr': 0.00048430338236582596, 'samples': 1429248, 'steps': 7443, 'loss/train': 1.2264333665370941} 01/28/2022 22:23:56 - INFO - codeparrot_training - Step 7444: {'lr': 0.0004842976753560485, 'samples': 1429440, 'steps': 7444, 'loss/train': 2.5417558550834656} 01/28/2022 22:24:01 - INFO - codeparrot_training - Step 7445: {'lr': 0.00048429196734261413, 'samples': 1429632, 'steps': 7445, 'loss/train': 2.1188650131225586} 01/28/2022 22:24:05 - INFO - codeparrot_training - Step 7446: {'lr': 0.00048428625832554754, 'samples': 1429824, 'steps': 7446, 'loss/train': 2.3278014063835144} 01/28/2022 22:24:10 - INFO - codeparrot_training - Step 7447: {'lr': 0.0004842805483048728, 'samples': 1430016, 'steps': 7447, 'loss/train': 2.0794885754585266} 01/28/2022 22:24:14 - INFO - codeparrot_training - Step 7448: {'lr': 0.0004842748372806147, 'samples': 1430208, 'steps': 7448, 'loss/train': 0.41108693182468414} 01/28/2022 22:24:18 - INFO - codeparrot_training - Step 7449: {'lr': 0.0004842691252527976, 'samples': 1430400, 'steps': 7449, 'loss/train': 1.7376405000686646} 01/28/2022 22:24:24 - INFO - codeparrot_training - Step 7450: {'lr': 0.00048426341222144586, 'samples': 1430592, 'steps': 7450, 'loss/train': 0.04065669793635607} 01/28/2022 22:24:29 - INFO - codeparrot_training - Step 7451: {'lr': 0.00048425769818658416, 'samples': 1430784, 'steps': 7451, 'loss/train': 3.342632532119751} 01/28/2022 22:24:33 - INFO - codeparrot_training - Step 7452: {'lr': 0.0004842519831482368, 'samples': 1430976, 'steps': 7452, 'loss/train': 1.6852530241012573} 01/28/2022 22:24:37 - INFO - codeparrot_training - Step 7453: {'lr': 0.00048424626710642836, 'samples': 1431168, 'steps': 7453, 'loss/train': 1.9965149760246277} 01/28/2022 22:24:41 - INFO - codeparrot_training - Step 7454: {'lr': 0.0004842405500611833, 'samples': 1431360, 'steps': 7454, 'loss/train': 1.813050627708435} 01/28/2022 22:24:47 - INFO - codeparrot_training - Step 7455: {'lr': 0.00048423483201252604, 'samples': 1431552, 'steps': 7455, 'loss/train': 4.426190614700317} 01/28/2022 22:24:51 - INFO - codeparrot_training - Step 7456: {'lr': 0.0004842291129604812, 'samples': 1431744, 'steps': 7456, 'loss/train': 1.642305850982666} 01/28/2022 22:24:55 - INFO - codeparrot_training - Step 7457: {'lr': 0.0004842233929050732, 'samples': 1431936, 'steps': 7457, 'loss/train': 1.8199977278709412} 01/28/2022 22:24:59 - INFO - codeparrot_training - Step 7458: {'lr': 0.00048421767184632657, 'samples': 1432128, 'steps': 7458, 'loss/train': 2.0560185313224792} 01/28/2022 22:25:04 - INFO - codeparrot_training - Step 7459: {'lr': 0.00048421194978426574, 'samples': 1432320, 'steps': 7459, 'loss/train': 1.359655648469925} 01/28/2022 22:25:10 - INFO - codeparrot_training - Step 7460: {'lr': 0.00048420622671891533, 'samples': 1432512, 'steps': 7460, 'loss/train': 1.878211498260498} 01/28/2022 22:25:14 - INFO - codeparrot_training - Step 7461: {'lr': 0.0004842005026502999, 'samples': 1432704, 'steps': 7461, 'loss/train': 2.1003423929214478} 01/28/2022 22:25:18 - INFO - codeparrot_training - Step 7462: {'lr': 0.00048419477757844376, 'samples': 1432896, 'steps': 7462, 'loss/train': 1.440024733543396} 01/28/2022 22:25:22 - INFO - codeparrot_training - Step 7463: {'lr': 0.00048418905150337166, 'samples': 1433088, 'steps': 7463, 'loss/train': 0.08628412149846554} 01/28/2022 22:25:27 - INFO - codeparrot_training - Step 7464: {'lr': 0.00048418332442510794, 'samples': 1433280, 'steps': 7464, 'loss/train': 2.041046619415283} 01/28/2022 22:25:31 - INFO - codeparrot_training - Step 7465: {'lr': 0.00048417759634367726, 'samples': 1433472, 'steps': 7465, 'loss/train': 1.9887371063232422} 01/28/2022 22:25:36 - INFO - codeparrot_training - Step 7466: {'lr': 0.00048417186725910414, 'samples': 1433664, 'steps': 7466, 'loss/train': 1.4535754323005676} 01/28/2022 22:25:41 - INFO - codeparrot_training - Step 7467: {'lr': 0.000484166137171413, 'samples': 1433856, 'steps': 7467, 'loss/train': 2.603290557861328} 01/28/2022 22:25:45 - INFO - codeparrot_training - Step 7468: {'lr': 0.0004841604060806286, 'samples': 1434048, 'steps': 7468, 'loss/train': 1.5153774619102478} 01/28/2022 22:25:49 - INFO - codeparrot_training - Step 7469: {'lr': 0.00048415467398677534, 'samples': 1434240, 'steps': 7469, 'loss/train': 1.9272315502166748} 01/28/2022 22:25:55 - INFO - codeparrot_training - Step 7470: {'lr': 0.0004841489408898778, 'samples': 1434432, 'steps': 7470, 'loss/train': 1.7467585802078247} 01/28/2022 22:25:59 - INFO - codeparrot_training - Step 7471: {'lr': 0.0004841432067899605, 'samples': 1434624, 'steps': 7471, 'loss/train': 0.9200525879859924} 01/28/2022 22:26:03 - INFO - codeparrot_training - Step 7472: {'lr': 0.0004841374716870481, 'samples': 1434816, 'steps': 7472, 'loss/train': 1.49938103556633} 01/28/2022 22:26:07 - INFO - codeparrot_training - Step 7473: {'lr': 0.0004841317355811651, 'samples': 1435008, 'steps': 7473, 'loss/train': 1.8464797139167786} 01/28/2022 22:26:11 - INFO - codeparrot_training - Step 7474: {'lr': 0.00048412599847233613, 'samples': 1435200, 'steps': 7474, 'loss/train': 1.8073723912239075} 01/28/2022 22:26:18 - INFO - codeparrot_training - Step 7475: {'lr': 0.0004841202603605857, 'samples': 1435392, 'steps': 7475, 'loss/train': 2.522166430950165} 01/28/2022 22:26:22 - INFO - codeparrot_training - Step 7476: {'lr': 0.0004841145212459384, 'samples': 1435584, 'steps': 7476, 'loss/train': 2.389186441898346} 01/28/2022 22:26:26 - INFO - codeparrot_training - Step 7477: {'lr': 0.0004841087811284188, 'samples': 1435776, 'steps': 7477, 'loss/train': 1.8203778862953186} 01/28/2022 22:26:30 - INFO - codeparrot_training - Step 7478: {'lr': 0.0004841030400080516, 'samples': 1435968, 'steps': 7478, 'loss/train': 2.455769419670105} 01/28/2022 22:26:34 - INFO - codeparrot_training - Step 7479: {'lr': 0.00048409729788486127, 'samples': 1436160, 'steps': 7479, 'loss/train': 1.3128783702850342} 01/28/2022 22:26:40 - INFO - codeparrot_training - Step 7480: {'lr': 0.00048409155475887244, 'samples': 1436352, 'steps': 7480, 'loss/train': 0.9580954313278198} 01/28/2022 22:26:44 - INFO - codeparrot_training - Step 7481: {'lr': 0.00048408581063010973, 'samples': 1436544, 'steps': 7481, 'loss/train': 2.063178777694702} 01/28/2022 22:26:48 - INFO - codeparrot_training - Step 7482: {'lr': 0.00048408006549859777, 'samples': 1436736, 'steps': 7482, 'loss/train': 1.0290648937225342} 01/28/2022 22:26:53 - INFO - codeparrot_training - Step 7483: {'lr': 0.00048407431936436116, 'samples': 1436928, 'steps': 7483, 'loss/train': 1.7495681047439575} 01/28/2022 22:26:57 - INFO - codeparrot_training - Step 7484: {'lr': 0.0004840685722274244, 'samples': 1437120, 'steps': 7484, 'loss/train': 2.0953243374824524} 01/28/2022 22:27:02 - INFO - codeparrot_training - Step 7485: {'lr': 0.00048406282408781226, 'samples': 1437312, 'steps': 7485, 'loss/train': 1.813527524471283} 01/28/2022 22:27:06 - INFO - codeparrot_training - Step 7486: {'lr': 0.0004840570749455493, 'samples': 1437504, 'steps': 7486, 'loss/train': 1.5705819725990295} 01/28/2022 22:27:10 - INFO - codeparrot_training - Step 7487: {'lr': 0.00048405132480066015, 'samples': 1437696, 'steps': 7487, 'loss/train': 1.5741891860961914} 01/28/2022 22:27:15 - INFO - codeparrot_training - Step 7488: {'lr': 0.00048404557365316946, 'samples': 1437888, 'steps': 7488, 'loss/train': 2.457449197769165} 01/28/2022 22:27:19 - INFO - codeparrot_training - Step 7489: {'lr': 0.00048403982150310184, 'samples': 1438080, 'steps': 7489, 'loss/train': 1.877382516860962} 01/28/2022 22:27:26 - INFO - codeparrot_training - Step 7490: {'lr': 0.0004840340683504819, 'samples': 1438272, 'steps': 7490, 'loss/train': 2.4221085906028748} 01/28/2022 22:27:30 - INFO - codeparrot_training - Step 7491: {'lr': 0.0004840283141953343, 'samples': 1438464, 'steps': 7491, 'loss/train': 2.148221969604492} 01/28/2022 22:27:34 - INFO - codeparrot_training - Step 7492: {'lr': 0.0004840225590376839, 'samples': 1438656, 'steps': 7492, 'loss/train': 7.680069923400879} 01/28/2022 22:27:38 - INFO - codeparrot_training - Step 7493: {'lr': 0.000484016802877555, 'samples': 1438848, 'steps': 7493, 'loss/train': 2.7755793929100037} 01/28/2022 22:27:43 - INFO - codeparrot_training - Step 7494: {'lr': 0.00048401104571497245, 'samples': 1439040, 'steps': 7494, 'loss/train': 1.823330819606781} 01/28/2022 22:27:47 - INFO - codeparrot_training - Step 7495: {'lr': 0.00048400528754996086, 'samples': 1439232, 'steps': 7495, 'loss/train': 2.529412865638733} 01/28/2022 22:27:52 - INFO - codeparrot_training - Step 7496: {'lr': 0.000483999528382545, 'samples': 1439424, 'steps': 7496, 'loss/train': 2.202912926673889} 01/28/2022 22:27:57 - INFO - codeparrot_training - Step 7497: {'lr': 0.00048399376821274943, 'samples': 1439616, 'steps': 7497, 'loss/train': 2.229615569114685} 01/28/2022 22:28:01 - INFO - codeparrot_training - Step 7498: {'lr': 0.00048398800704059887, 'samples': 1439808, 'steps': 7498, 'loss/train': 2.2777944803237915} 01/28/2022 22:28:05 - INFO - codeparrot_training - Step 7499: {'lr': 0.000483982244866118, 'samples': 1440000, 'steps': 7499, 'loss/train': 2.240984559059143} 01/28/2022 22:28:09 - INFO - codeparrot_training - Step 7500: {'lr': 0.00048397648168933144, 'samples': 1440192, 'steps': 7500, 'loss/train': 2.1670528650283813} 01/28/2022 22:28:15 - INFO - codeparrot_training - Step 7501: {'lr': 0.00048397071751026395, 'samples': 1440384, 'steps': 7501, 'loss/train': 2.0516130924224854} 01/28/2022 22:28:20 - INFO - codeparrot_training - Step 7502: {'lr': 0.00048396495232894024, 'samples': 1440576, 'steps': 7502, 'loss/train': 2.0819878578186035} 01/28/2022 22:28:24 - INFO - codeparrot_training - Step 7503: {'lr': 0.0004839591861453849, 'samples': 1440768, 'steps': 7503, 'loss/train': 1.6715653538703918} 01/28/2022 22:28:28 - INFO - codeparrot_training - Step 7504: {'lr': 0.00048395341895962277, 'samples': 1440960, 'steps': 7504, 'loss/train': 1.5330618023872375} 01/28/2022 22:28:33 - INFO - codeparrot_training - Step 7505: {'lr': 0.0004839476507716784, 'samples': 1441152, 'steps': 7505, 'loss/train': 1.7690377235412598} 01/28/2022 22:28:38 - INFO - codeparrot_training - Step 7506: {'lr': 0.0004839418815815766, 'samples': 1441344, 'steps': 7506, 'loss/train': 1.3872244656085968} 01/28/2022 22:28:42 - INFO - codeparrot_training - Step 7507: {'lr': 0.0004839361113893421, 'samples': 1441536, 'steps': 7507, 'loss/train': 2.3562101125717163} 01/28/2022 22:28:46 - INFO - codeparrot_training - Step 7508: {'lr': 0.0004839303401949996, 'samples': 1441728, 'steps': 7508, 'loss/train': 1.6429322361946106} 01/28/2022 22:28:50 - INFO - codeparrot_training - Step 7509: {'lr': 0.00048392456799857374, 'samples': 1441920, 'steps': 7509, 'loss/train': 2.5660316348075867} 01/28/2022 22:28:56 - INFO - codeparrot_training - Step 7510: {'lr': 0.0004839187948000893, 'samples': 1442112, 'steps': 7510, 'loss/train': 2.60730117559433} 01/28/2022 22:29:00 - INFO - codeparrot_training - Step 7511: {'lr': 0.0004839130205995711, 'samples': 1442304, 'steps': 7511, 'loss/train': 1.8940140008926392} 01/28/2022 22:29:05 - INFO - codeparrot_training - Step 7512: {'lr': 0.0004839072453970438, 'samples': 1442496, 'steps': 7512, 'loss/train': 1.9873151779174805} 01/28/2022 22:29:09 - INFO - codeparrot_training - Step 7513: {'lr': 0.00048390146919253206, 'samples': 1442688, 'steps': 7513, 'loss/train': 1.7296535968780518} 01/28/2022 22:29:13 - INFO - codeparrot_training - Step 7514: {'lr': 0.0004838956919860607, 'samples': 1442880, 'steps': 7514, 'loss/train': 2.180100202560425} 01/28/2022 22:29:19 - INFO - codeparrot_training - Step 7515: {'lr': 0.0004838899137776545, 'samples': 1443072, 'steps': 7515, 'loss/train': 0.7923544049263} 01/28/2022 22:29:23 - INFO - codeparrot_training - Step 7516: {'lr': 0.00048388413456733814, 'samples': 1443264, 'steps': 7516, 'loss/train': 0.4764890670776367} 01/28/2022 22:29:28 - INFO - codeparrot_training - Step 7517: {'lr': 0.0004838783543551365, 'samples': 1443456, 'steps': 7517, 'loss/train': 2.6911368370056152} 01/28/2022 22:29:32 - INFO - codeparrot_training - Step 7518: {'lr': 0.0004838725731410742, 'samples': 1443648, 'steps': 7518, 'loss/train': 0.3495289608836174} 01/28/2022 22:29:36 - INFO - codeparrot_training - Step 7519: {'lr': 0.00048386679092517605, 'samples': 1443840, 'steps': 7519, 'loss/train': 0.1114000529050827} 01/28/2022 22:29:42 - INFO - codeparrot_training - Step 7520: {'lr': 0.00048386100770746686, 'samples': 1444032, 'steps': 7520, 'loss/train': 0.7297073006629944} 01/28/2022 22:29:47 - INFO - codeparrot_training - Step 7521: {'lr': 0.00048385522348797134, 'samples': 1444224, 'steps': 7521, 'loss/train': 1.8170233368873596} 01/28/2022 22:29:51 - INFO - codeparrot_training - Step 7522: {'lr': 0.0004838494382667143, 'samples': 1444416, 'steps': 7522, 'loss/train': 1.9692690968513489} 01/28/2022 22:29:55 - INFO - codeparrot_training - Step 7523: {'lr': 0.0004838436520437205, 'samples': 1444608, 'steps': 7523, 'loss/train': 2.262737989425659} 01/28/2022 22:29:59 - INFO - codeparrot_training - Step 7524: {'lr': 0.00048383786481901483, 'samples': 1444800, 'steps': 7524, 'loss/train': 2.2000628113746643} 01/28/2022 22:30:05 - INFO - codeparrot_training - Step 7525: {'lr': 0.00048383207659262196, 'samples': 1444992, 'steps': 7525, 'loss/train': 1.630771815776825} 01/28/2022 22:30:09 - INFO - codeparrot_training - Step 7526: {'lr': 0.0004838262873645667, 'samples': 1445184, 'steps': 7526, 'loss/train': 1.6003975868225098} 01/28/2022 22:30:13 - INFO - codeparrot_training - Step 7527: {'lr': 0.00048382049713487383, 'samples': 1445376, 'steps': 7527, 'loss/train': 2.0824685096740723} 01/28/2022 22:30:17 - INFO - codeparrot_training - Step 7528: {'lr': 0.00048381470590356835, 'samples': 1445568, 'steps': 7528, 'loss/train': 1.8230438232421875} 01/28/2022 22:30:22 - INFO - codeparrot_training - Step 7529: {'lr': 0.00048380891367067483, 'samples': 1445760, 'steps': 7529, 'loss/train': 2.2220757007598877} 01/28/2022 22:30:28 - INFO - codeparrot_training - Step 7530: {'lr': 0.0004838031204362181, 'samples': 1445952, 'steps': 7530, 'loss/train': 1.3804227411746979} 01/28/2022 22:30:32 - INFO - codeparrot_training - Step 7531: {'lr': 0.0004837973262002231, 'samples': 1446144, 'steps': 7531, 'loss/train': 1.9573418498039246} 01/28/2022 22:30:36 - INFO - codeparrot_training - Step 7532: {'lr': 0.0004837915309627146, 'samples': 1446336, 'steps': 7532, 'loss/train': 1.5118496417999268} 01/28/2022 22:30:41 - INFO - codeparrot_training - Step 7533: {'lr': 0.00048378573472371744, 'samples': 1446528, 'steps': 7533, 'loss/train': 1.6596550941467285} 01/28/2022 22:30:45 - INFO - codeparrot_training - Step 7534: {'lr': 0.0004837799374832564, 'samples': 1446720, 'steps': 7534, 'loss/train': 2.1571070551872253} 01/28/2022 22:30:50 - INFO - codeparrot_training - Step 7535: {'lr': 0.0004837741392413563, 'samples': 1446912, 'steps': 7535, 'loss/train': 1.2938447892665863} 01/28/2022 22:30:54 - INFO - codeparrot_training - Step 7536: {'lr': 0.000483768339998042, 'samples': 1447104, 'steps': 7536, 'loss/train': 2.778359055519104} 01/28/2022 22:30:59 - INFO - codeparrot_training - Step 7537: {'lr': 0.0004837625397533385, 'samples': 1447296, 'steps': 7537, 'loss/train': 3.091795563697815} 01/28/2022 22:31:03 - INFO - codeparrot_training - Step 7538: {'lr': 0.00048375673850727043, 'samples': 1447488, 'steps': 7538, 'loss/train': 0.839022696018219} 01/28/2022 22:31:07 - INFO - codeparrot_training - Step 7539: {'lr': 0.00048375093625986274, 'samples': 1447680, 'steps': 7539, 'loss/train': 2.5157427191734314} 01/28/2022 22:31:13 - INFO - codeparrot_training - Step 7540: {'lr': 0.0004837451330111402, 'samples': 1447872, 'steps': 7540, 'loss/train': 2.498226284980774} 01/28/2022 22:31:17 - INFO - codeparrot_training - Step 7541: {'lr': 0.0004837393287611278, 'samples': 1448064, 'steps': 7541, 'loss/train': 1.7254092693328857} 01/28/2022 22:31:21 - INFO - codeparrot_training - Step 7542: {'lr': 0.0004837335235098503, 'samples': 1448256, 'steps': 7542, 'loss/train': 1.5735209584236145} 01/28/2022 22:31:25 - INFO - codeparrot_training - Step 7543: {'lr': 0.0004837277172573326, 'samples': 1448448, 'steps': 7543, 'loss/train': 1.985711395740509} 01/28/2022 22:31:30 - INFO - codeparrot_training - Step 7544: {'lr': 0.00048372191000359955, 'samples': 1448640, 'steps': 7544, 'loss/train': 1.6974921226501465} 01/28/2022 22:31:36 - INFO - codeparrot_training - Step 7545: {'lr': 0.00048371610174867614, 'samples': 1448832, 'steps': 7545, 'loss/train': 0.1427701972424984} 01/28/2022 22:31:40 - INFO - codeparrot_training - Step 7546: {'lr': 0.00048371029249258716, 'samples': 1449024, 'steps': 7546, 'loss/train': 1.6805071234703064} 01/28/2022 22:31:44 - INFO - codeparrot_training - Step 7547: {'lr': 0.0004837044822353574, 'samples': 1449216, 'steps': 7547, 'loss/train': 1.5190921425819397} 01/28/2022 22:31:48 - INFO - codeparrot_training - Step 7548: {'lr': 0.0004836986709770119, 'samples': 1449408, 'steps': 7548, 'loss/train': 1.5347515940666199} 01/28/2022 22:31:53 - INFO - codeparrot_training - Step 7549: {'lr': 0.00048369285871757554, 'samples': 1449600, 'steps': 7549, 'loss/train': 0.20276691019535065} 01/28/2022 22:31:58 - INFO - codeparrot_training - Step 7550: {'lr': 0.0004836870454570731, 'samples': 1449792, 'steps': 7550, 'loss/train': 2.650163412094116} 01/28/2022 22:32:02 - INFO - codeparrot_training - Step 7551: {'lr': 0.00048368123119552965, 'samples': 1449984, 'steps': 7551, 'loss/train': 1.3959356546401978} 01/28/2022 22:32:06 - INFO - codeparrot_training - Step 7552: {'lr': 0.00048367541593296996, 'samples': 1450176, 'steps': 7552, 'loss/train': 1.5328579545021057} 01/28/2022 22:32:11 - INFO - codeparrot_training - Step 7553: {'lr': 0.00048366959966941893, 'samples': 1450368, 'steps': 7553, 'loss/train': 2.4381285309791565} 01/28/2022 22:32:15 - INFO - codeparrot_training - Step 7554: {'lr': 0.0004836637824049016, 'samples': 1450560, 'steps': 7554, 'loss/train': 1.446521669626236} 01/28/2022 22:32:20 - INFO - codeparrot_training - Step 7555: {'lr': 0.00048365796413944284, 'samples': 1450752, 'steps': 7555, 'loss/train': 0.9838560819625854} 01/28/2022 22:32:24 - INFO - codeparrot_training - Step 7556: {'lr': 0.00048365214487306753, 'samples': 1450944, 'steps': 7556, 'loss/train': 1.6538887023925781} 01/28/2022 22:32:29 - INFO - codeparrot_training - Step 7557: {'lr': 0.0004836463246058006, 'samples': 1451136, 'steps': 7557, 'loss/train': 1.430130958557129} 01/28/2022 22:32:33 - INFO - codeparrot_training - Step 7558: {'lr': 0.0004836405033376671, 'samples': 1451328, 'steps': 7558, 'loss/train': 1.6941279172897339} 01/28/2022 22:32:37 - INFO - codeparrot_training - Step 7559: {'lr': 0.00048363468106869177, 'samples': 1451520, 'steps': 7559, 'loss/train': 1.3078963458538055} 01/28/2022 22:32:43 - INFO - codeparrot_training - Step 7560: {'lr': 0.00048362885779889967, 'samples': 1451712, 'steps': 7560, 'loss/train': 0.7435941249132156} 01/28/2022 22:32:47 - INFO - codeparrot_training - Step 7561: {'lr': 0.0004836230335283158, 'samples': 1451904, 'steps': 7561, 'loss/train': 1.8818113803863525} 01/28/2022 22:32:51 - INFO - codeparrot_training - Step 7562: {'lr': 0.00048361720825696494, 'samples': 1452096, 'steps': 7562, 'loss/train': 2.05969101190567} 01/28/2022 22:32:56 - INFO - codeparrot_training - Step 7563: {'lr': 0.0004836113819848722, 'samples': 1452288, 'steps': 7563, 'loss/train': 1.2995484173297882} 01/28/2022 22:33:00 - INFO - codeparrot_training - Step 7564: {'lr': 0.0004836055547120625, 'samples': 1452480, 'steps': 7564, 'loss/train': 2.3304982781410217} 01/28/2022 22:33:06 - INFO - codeparrot_training - Step 7565: {'lr': 0.0004835997264385607, 'samples': 1452672, 'steps': 7565, 'loss/train': 0.9428735375404358} 01/28/2022 22:33:10 - INFO - codeparrot_training - Step 7566: {'lr': 0.0004835938971643919, 'samples': 1452864, 'steps': 7566, 'loss/train': 1.4525774717330933} 01/28/2022 22:33:14 - INFO - codeparrot_training - Step 7567: {'lr': 0.000483588066889581, 'samples': 1453056, 'steps': 7567, 'loss/train': 1.392434298992157} 01/28/2022 22:33:19 - INFO - codeparrot_training - Step 7568: {'lr': 0.00048358223561415306, 'samples': 1453248, 'steps': 7568, 'loss/train': 0.7651273012161255} 01/28/2022 22:33:23 - INFO - codeparrot_training - Step 7569: {'lr': 0.0004835764033381329, 'samples': 1453440, 'steps': 7569, 'loss/train': 1.12410369515419} 01/28/2022 22:33:28 - INFO - codeparrot_training - Step 7570: {'lr': 0.00048357057006154566, 'samples': 1453632, 'steps': 7570, 'loss/train': 0.23775138705968857} 01/28/2022 22:33:32 - INFO - codeparrot_training - Step 7571: {'lr': 0.0004835647357844162, 'samples': 1453824, 'steps': 7571, 'loss/train': 2.445335626602173} 01/28/2022 22:33:37 - INFO - codeparrot_training - Step 7572: {'lr': 0.00048355890050676966, 'samples': 1454016, 'steps': 7572, 'loss/train': 2.207042098045349} 01/28/2022 22:33:41 - INFO - codeparrot_training - Step 7573: {'lr': 0.0004835530642286309, 'samples': 1454208, 'steps': 7573, 'loss/train': 0.7985773086547852} 01/28/2022 22:33:45 - INFO - codeparrot_training - Step 7574: {'lr': 0.000483547226950025, 'samples': 1454400, 'steps': 7574, 'loss/train': 1.3126773834228516} 01/28/2022 22:33:51 - INFO - codeparrot_training - Step 7575: {'lr': 0.00048354138867097695, 'samples': 1454592, 'steps': 7575, 'loss/train': 2.2328310012817383} 01/28/2022 22:33:55 - INFO - codeparrot_training - Step 7576: {'lr': 0.00048353554939151167, 'samples': 1454784, 'steps': 7576, 'loss/train': 2.916848659515381} 01/28/2022 22:33:59 - INFO - codeparrot_training - Step 7577: {'lr': 0.00048352970911165434, 'samples': 1454976, 'steps': 7577, 'loss/train': 1.7927228808403015} 01/28/2022 22:34:03 - INFO - codeparrot_training - Step 7578: {'lr': 0.0004835238678314299, 'samples': 1455168, 'steps': 7578, 'loss/train': 1.2764263451099396} 01/28/2022 22:34:08 - INFO - codeparrot_training - Step 7579: {'lr': 0.00048351802555086335, 'samples': 1455360, 'steps': 7579, 'loss/train': 1.585975706577301} 01/28/2022 22:34:14 - INFO - codeparrot_training - Step 7580: {'lr': 0.0004835121822699796, 'samples': 1455552, 'steps': 7580, 'loss/train': 1.110478699207306} 01/28/2022 22:34:18 - INFO - codeparrot_training - Step 7581: {'lr': 0.00048350633798880397, 'samples': 1455744, 'steps': 7581, 'loss/train': 1.1437289714813232} 01/28/2022 22:34:22 - INFO - codeparrot_training - Step 7582: {'lr': 0.0004835004927073613, 'samples': 1455936, 'steps': 7582, 'loss/train': 1.0856116712093353} 01/28/2022 22:34:26 - INFO - codeparrot_training - Step 7583: {'lr': 0.0004834946464256766, 'samples': 1456128, 'steps': 7583, 'loss/train': 1.303738296031952} 01/28/2022 22:34:31 - INFO - codeparrot_training - Step 7584: {'lr': 0.00048348879914377504, 'samples': 1456320, 'steps': 7584, 'loss/train': 1.2596091628074646} 01/28/2022 22:34:36 - INFO - codeparrot_training - Step 7585: {'lr': 0.0004834829508616816, 'samples': 1456512, 'steps': 7585, 'loss/train': 2.319208323955536} 01/28/2022 22:34:41 - INFO - codeparrot_training - Step 7586: {'lr': 0.00048347710157942126, 'samples': 1456704, 'steps': 7586, 'loss/train': 1.148472636938095} 01/28/2022 22:34:45 - INFO - codeparrot_training - Step 7587: {'lr': 0.00048347125129701924, 'samples': 1456896, 'steps': 7587, 'loss/train': 2.1557815074920654} 01/28/2022 22:34:49 - INFO - codeparrot_training - Step 7588: {'lr': 0.00048346540001450045, 'samples': 1457088, 'steps': 7588, 'loss/train': 1.2487527430057526} 01/28/2022 22:34:53 - INFO - codeparrot_training - Step 7589: {'lr': 0.0004834595477318901, 'samples': 1457280, 'steps': 7589, 'loss/train': 2.0195242166519165} 01/28/2022 22:35:00 - INFO - codeparrot_training - Step 7590: {'lr': 0.00048345369444921315, 'samples': 1457472, 'steps': 7590, 'loss/train': 1.7341464757919312} 01/28/2022 22:35:04 - INFO - codeparrot_training - Step 7591: {'lr': 0.00048344784016649467, 'samples': 1457664, 'steps': 7591, 'loss/train': 1.9889937043190002} 01/28/2022 22:35:08 - INFO - codeparrot_training - Step 7592: {'lr': 0.0004834419848837598, 'samples': 1457856, 'steps': 7592, 'loss/train': 2.1033888459205627} 01/28/2022 22:35:12 - INFO - codeparrot_training - Step 7593: {'lr': 0.0004834361286010336, 'samples': 1458048, 'steps': 7593, 'loss/train': 2.2138078808784485} 01/28/2022 22:35:17 - INFO - codeparrot_training - Step 7594: {'lr': 0.0004834302713183411, 'samples': 1458240, 'steps': 7594, 'loss/train': 2.471737325191498} 01/28/2022 22:35:22 - INFO - codeparrot_training - Step 7595: {'lr': 0.0004834244130357075, 'samples': 1458432, 'steps': 7595, 'loss/train': 1.7811030149459839} 01/28/2022 22:35:26 - INFO - codeparrot_training - Step 7596: {'lr': 0.0004834185537531578, 'samples': 1458624, 'steps': 7596, 'loss/train': 2.2136950492858887} 01/28/2022 22:35:31 - INFO - codeparrot_training - Step 7597: {'lr': 0.00048341269347071717, 'samples': 1458816, 'steps': 7597, 'loss/train': 2.0266231298446655} 01/28/2022 22:35:35 - INFO - codeparrot_training - Step 7598: {'lr': 0.00048340683218841066, 'samples': 1459008, 'steps': 7598, 'loss/train': 1.6809852719306946} 01/28/2022 22:35:39 - INFO - codeparrot_training - Step 7599: {'lr': 0.00048340096990626336, 'samples': 1459200, 'steps': 7599, 'loss/train': 2.189228653907776} 01/28/2022 22:35:44 - INFO - codeparrot_training - Step 7600: {'lr': 0.00048339510662430044, 'samples': 1459392, 'steps': 7600, 'loss/train': 2.099264681339264} 01/28/2022 22:35:49 - INFO - codeparrot_training - Step 7601: {'lr': 0.000483389242342547, 'samples': 1459584, 'steps': 7601, 'loss/train': 1.9301995038986206} 01/28/2022 22:35:53 - INFO - codeparrot_training - Step 7602: {'lr': 0.00048338337706102817, 'samples': 1459776, 'steps': 7602, 'loss/train': 1.828373372554779} 01/28/2022 22:35:57 - INFO - codeparrot_training - Step 7603: {'lr': 0.00048337751077976907, 'samples': 1459968, 'steps': 7603, 'loss/train': 2.0640040040016174} 01/28/2022 22:36:01 - INFO - codeparrot_training - Step 7604: {'lr': 0.0004833716434987948, 'samples': 1460160, 'steps': 7604, 'loss/train': 0.9098332822322845} 01/28/2022 22:36:07 - INFO - codeparrot_training - Step 7605: {'lr': 0.0004833657752181305, 'samples': 1460352, 'steps': 7605, 'loss/train': 1.2731823921203613} 01/28/2022 22:36:12 - INFO - codeparrot_training - Step 7606: {'lr': 0.00048335990593780133, 'samples': 1460544, 'steps': 7606, 'loss/train': 2.821999490261078} 01/28/2022 22:36:16 - INFO - codeparrot_training - Step 7607: {'lr': 0.00048335403565783245, 'samples': 1460736, 'steps': 7607, 'loss/train': 2.095900297164917} 01/28/2022 22:36:20 - INFO - codeparrot_training - Step 7608: {'lr': 0.0004833481643782489, 'samples': 1460928, 'steps': 7608, 'loss/train': 2.5228464603424072} 01/28/2022 22:36:24 - INFO - codeparrot_training - Step 7609: {'lr': 0.000483342292099076, 'samples': 1461120, 'steps': 7609, 'loss/train': 2.259542226791382} 01/28/2022 22:36:30 - INFO - codeparrot_training - Step 7610: {'lr': 0.0004833364188203387, 'samples': 1461312, 'steps': 7610, 'loss/train': 2.046951413154602} 01/28/2022 22:36:34 - INFO - codeparrot_training - Step 7611: {'lr': 0.0004833305445420624, 'samples': 1461504, 'steps': 7611, 'loss/train': 2.0676101446151733} 01/28/2022 22:36:38 - INFO - codeparrot_training - Step 7612: {'lr': 0.0004833246692642721, 'samples': 1461696, 'steps': 7612, 'loss/train': 2.2351083755493164} 01/28/2022 22:36:42 - INFO - codeparrot_training - Step 7613: {'lr': 0.000483318792986993, 'samples': 1461888, 'steps': 7613, 'loss/train': 1.0787769556045532} 01/28/2022 22:36:47 - INFO - codeparrot_training - Step 7614: {'lr': 0.00048331291571025026, 'samples': 1462080, 'steps': 7614, 'loss/train': 1.0074488818645477} 01/28/2022 22:36:52 - INFO - codeparrot_training - Step 7615: {'lr': 0.0004833070374340691, 'samples': 1462272, 'steps': 7615, 'loss/train': 0.744707778096199} 01/28/2022 22:36:56 - INFO - codeparrot_training - Step 7616: {'lr': 0.00048330115815847465, 'samples': 1462464, 'steps': 7616, 'loss/train': 2.0056480765342712} 01/28/2022 22:37:00 - INFO - codeparrot_training - Step 7617: {'lr': 0.00048329527788349216, 'samples': 1462656, 'steps': 7617, 'loss/train': 1.8483728170394897} 01/28/2022 22:37:05 - INFO - codeparrot_training - Step 7618: {'lr': 0.0004832893966091467, 'samples': 1462848, 'steps': 7618, 'loss/train': 2.0836254358291626} 01/28/2022 22:37:09 - INFO - codeparrot_training - Step 7619: {'lr': 0.00048328351433546364, 'samples': 1463040, 'steps': 7619, 'loss/train': 2.151071012020111} 01/28/2022 22:37:14 - INFO - codeparrot_training - Step 7620: {'lr': 0.000483277631062468, 'samples': 1463232, 'steps': 7620, 'loss/train': 1.5321480631828308} 01/28/2022 22:37:18 - INFO - codeparrot_training - Step 7621: {'lr': 0.00048327174679018515, 'samples': 1463424, 'steps': 7621, 'loss/train': 2.0201579332351685} 01/28/2022 22:37:23 - INFO - codeparrot_training - Step 7622: {'lr': 0.00048326586151864015, 'samples': 1463616, 'steps': 7622, 'loss/train': 1.3510719537734985} 01/28/2022 22:37:27 - INFO - codeparrot_training - Step 7623: {'lr': 0.00048325997524785826, 'samples': 1463808, 'steps': 7623, 'loss/train': 1.5088720321655273} 01/28/2022 22:37:31 - INFO - codeparrot_training - Step 7624: {'lr': 0.00048325408797786476, 'samples': 1464000, 'steps': 7624, 'loss/train': 1.505541980266571} 01/28/2022 22:37:37 - INFO - codeparrot_training - Step 7625: {'lr': 0.00048324819970868473, 'samples': 1464192, 'steps': 7625, 'loss/train': 1.371246099472046} 01/28/2022 22:37:41 - INFO - codeparrot_training - Step 7626: {'lr': 0.0004832423104403435, 'samples': 1464384, 'steps': 7626, 'loss/train': 1.5940561294555664} 01/28/2022 22:37:46 - INFO - codeparrot_training - Step 7627: {'lr': 0.0004832364201728663, 'samples': 1464576, 'steps': 7627, 'loss/train': 1.8071674704551697} 01/28/2022 22:37:50 - INFO - codeparrot_training - Step 7628: {'lr': 0.0004832305289062784, 'samples': 1464768, 'steps': 7628, 'loss/train': 0.545773833990097} 01/28/2022 22:37:55 - INFO - codeparrot_training - Step 7629: {'lr': 0.0004832246366406049, 'samples': 1464960, 'steps': 7629, 'loss/train': 1.3329648971557617} 01/28/2022 22:38:00 - INFO - codeparrot_training - Step 7630: {'lr': 0.00048321874337587105, 'samples': 1465152, 'steps': 7630, 'loss/train': 0.13086963817477226} 01/28/2022 22:38:04 - INFO - codeparrot_training - Step 7631: {'lr': 0.0004832128491121023, 'samples': 1465344, 'steps': 7631, 'loss/train': 2.2306732535362244} 01/28/2022 22:38:08 - INFO - codeparrot_training - Step 7632: {'lr': 0.00048320695384932366, 'samples': 1465536, 'steps': 7632, 'loss/train': 1.906261146068573} 01/28/2022 22:38:12 - INFO - codeparrot_training - Step 7633: {'lr': 0.0004832010575875605, 'samples': 1465728, 'steps': 7633, 'loss/train': 1.6117826700210571} 01/28/2022 22:38:19 - INFO - codeparrot_training - Step 7634: {'lr': 0.0004831951603268381, 'samples': 1465920, 'steps': 7634, 'loss/train': 1.6585305333137512} 01/28/2022 22:38:23 - INFO - codeparrot_training - Step 7635: {'lr': 0.0004831892620671816, 'samples': 1466112, 'steps': 7635, 'loss/train': 1.096581906080246} 01/28/2022 22:38:27 - INFO - codeparrot_training - Step 7636: {'lr': 0.0004831833628086164, 'samples': 1466304, 'steps': 7636, 'loss/train': 1.5495581030845642} 01/28/2022 22:38:32 - INFO - codeparrot_training - Step 7637: {'lr': 0.0004831774625511677, 'samples': 1466496, 'steps': 7637, 'loss/train': 1.5744929909706116} 01/28/2022 22:38:36 - INFO - codeparrot_training - Step 7638: {'lr': 0.00048317156129486086, 'samples': 1466688, 'steps': 7638, 'loss/train': 1.7068410515785217} 01/28/2022 22:38:41 - INFO - codeparrot_training - Step 7639: {'lr': 0.000483165659039721, 'samples': 1466880, 'steps': 7639, 'loss/train': 1.784538209438324} 01/28/2022 22:38:46 - INFO - codeparrot_training - Step 7640: {'lr': 0.0004831597557857735, 'samples': 1467072, 'steps': 7640, 'loss/train': 1.5989812016487122} 01/28/2022 22:38:50 - INFO - codeparrot_training - Step 7641: {'lr': 0.0004831538515330437, 'samples': 1467264, 'steps': 7641, 'loss/train': 1.179499089717865} 01/28/2022 22:38:54 - INFO - codeparrot_training - Step 7642: {'lr': 0.0004831479462815568, 'samples': 1467456, 'steps': 7642, 'loss/train': 0.970570832490921} 01/28/2022 22:38:58 - INFO - codeparrot_training - Step 7643: {'lr': 0.00048314204003133815, 'samples': 1467648, 'steps': 7643, 'loss/train': 1.4335698187351227} 01/28/2022 22:39:04 - INFO - codeparrot_training - Step 7644: {'lr': 0.00048313613278241305, 'samples': 1467840, 'steps': 7644, 'loss/train': 1.6008383631706238} 01/28/2022 22:39:08 - INFO - codeparrot_training - Step 7645: {'lr': 0.0004831302245348068, 'samples': 1468032, 'steps': 7645, 'loss/train': 1.9492895007133484} 01/28/2022 22:39:12 - INFO - codeparrot_training - Step 7646: {'lr': 0.0004831243152885446, 'samples': 1468224, 'steps': 7646, 'loss/train': 0.15524186566472054} 01/28/2022 22:39:16 - INFO - codeparrot_training - Step 7647: {'lr': 0.0004831184050436519, 'samples': 1468416, 'steps': 7647, 'loss/train': 1.5976957082748413} 01/28/2022 22:39:21 - INFO - codeparrot_training - Step 7648: {'lr': 0.000483112493800154, 'samples': 1468608, 'steps': 7648, 'loss/train': 1.8435468077659607} 01/28/2022 22:39:27 - INFO - codeparrot_training - Step 7649: {'lr': 0.0004831065815580762, 'samples': 1468800, 'steps': 7649, 'loss/train': 1.8089028596878052} 01/28/2022 22:39:31 - INFO - codeparrot_training - Step 7650: {'lr': 0.0004831006683174438, 'samples': 1468992, 'steps': 7650, 'loss/train': 1.911927044391632} 01/28/2022 22:39:35 - INFO - codeparrot_training - Step 7651: {'lr': 0.0004830947540782822, 'samples': 1469184, 'steps': 7651, 'loss/train': 1.3630084097385406} 01/28/2022 22:39:39 - INFO - codeparrot_training - Step 7652: {'lr': 0.0004830888388406166, 'samples': 1469376, 'steps': 7652, 'loss/train': 2.332460939884186} 01/28/2022 22:39:44 - INFO - codeparrot_training - Step 7653: {'lr': 0.0004830829226044725, 'samples': 1469568, 'steps': 7653, 'loss/train': 1.4058417677879333} 01/28/2022 22:39:49 - INFO - codeparrot_training - Step 7654: {'lr': 0.0004830770053698752, 'samples': 1469760, 'steps': 7654, 'loss/train': 1.757643699645996} 01/28/2022 22:39:53 - INFO - codeparrot_training - Step 7655: {'lr': 0.00048307108713684994, 'samples': 1469952, 'steps': 7655, 'loss/train': 1.1485302150249481} 01/28/2022 22:39:57 - INFO - codeparrot_training - Step 7656: {'lr': 0.00048306516790542223, 'samples': 1470144, 'steps': 7656, 'loss/train': 1.6652133464813232} 01/28/2022 22:40:02 - INFO - codeparrot_training - Step 7657: {'lr': 0.00048305924767561725, 'samples': 1470336, 'steps': 7657, 'loss/train': 2.4320003986358643} 01/28/2022 22:40:06 - INFO - codeparrot_training - Step 7658: {'lr': 0.00048305332644746053, 'samples': 1470528, 'steps': 7658, 'loss/train': 1.729971170425415} 01/28/2022 22:40:11 - INFO - codeparrot_training - Step 7659: {'lr': 0.0004830474042209774, 'samples': 1470720, 'steps': 7659, 'loss/train': 1.7930299043655396} 01/28/2022 22:40:15 - INFO - codeparrot_training - Step 7660: {'lr': 0.00048304148099619304, 'samples': 1470912, 'steps': 7660, 'loss/train': 1.502970278263092} 01/28/2022 22:40:20 - INFO - codeparrot_training - Step 7661: {'lr': 0.0004830355567731331, 'samples': 1471104, 'steps': 7661, 'loss/train': 2.3268465399742126} 01/28/2022 22:40:24 - INFO - codeparrot_training - Step 7662: {'lr': 0.0004830296315518228, 'samples': 1471296, 'steps': 7662, 'loss/train': 2.167696237564087} 01/28/2022 22:40:28 - INFO - codeparrot_training - Step 7663: {'lr': 0.00048302370533228754, 'samples': 1471488, 'steps': 7663, 'loss/train': 1.1125187873840332} 01/28/2022 22:40:34 - INFO - codeparrot_training - Step 7664: {'lr': 0.00048301777811455274, 'samples': 1471680, 'steps': 7664, 'loss/train': 1.0776237845420837} 01/28/2022 22:40:38 - INFO - codeparrot_training - Step 7665: {'lr': 0.0004830118498986438, 'samples': 1471872, 'steps': 7665, 'loss/train': 1.5104824304580688} 01/28/2022 22:40:42 - INFO - codeparrot_training - Step 7666: {'lr': 0.000483005920684586, 'samples': 1472064, 'steps': 7666, 'loss/train': 1.0913309454917908} 01/28/2022 22:40:47 - INFO - codeparrot_training - Step 7667: {'lr': 0.0004829999904724049, 'samples': 1472256, 'steps': 7667, 'loss/train': 1.2299893498420715} 01/28/2022 22:40:51 - INFO - codeparrot_training - Step 7668: {'lr': 0.0004829940592621258, 'samples': 1472448, 'steps': 7668, 'loss/train': 2.0583600997924805} 01/28/2022 22:40:57 - INFO - codeparrot_training - Step 7669: {'lr': 0.00048298812705377414, 'samples': 1472640, 'steps': 7669, 'loss/train': 1.8201343417167664} 01/28/2022 22:41:01 - INFO - codeparrot_training - Step 7670: {'lr': 0.0004829821938473753, 'samples': 1472832, 'steps': 7670, 'loss/train': 1.3837048709392548} 01/28/2022 22:41:05 - INFO - codeparrot_training - Step 7671: {'lr': 0.0004829762596429548, 'samples': 1473024, 'steps': 7671, 'loss/train': 2.0861465334892273} 01/28/2022 22:41:09 - INFO - codeparrot_training - Step 7672: {'lr': 0.0004829703244405379, 'samples': 1473216, 'steps': 7672, 'loss/train': 1.9123100638389587} 01/28/2022 22:41:15 - INFO - codeparrot_training - Step 7673: {'lr': 0.0004829643882401501, 'samples': 1473408, 'steps': 7673, 'loss/train': 2.0520023703575134} 01/28/2022 22:41:19 - INFO - codeparrot_training - Step 7674: {'lr': 0.0004829584510418169, 'samples': 1473600, 'steps': 7674, 'loss/train': 2.083825170993805} 01/28/2022 22:41:23 - INFO - codeparrot_training - Step 7675: {'lr': 0.00048295251284556363, 'samples': 1473792, 'steps': 7675, 'loss/train': 2.0979470014572144} 01/28/2022 22:41:28 - INFO - codeparrot_training - Step 7676: {'lr': 0.0004829465736514157, 'samples': 1473984, 'steps': 7676, 'loss/train': 1.6458991169929504} 01/28/2022 22:41:32 - INFO - codeparrot_training - Step 7677: {'lr': 0.00048294063345939877, 'samples': 1474176, 'steps': 7677, 'loss/train': 1.558353066444397} 01/28/2022 22:41:36 - INFO - codeparrot_training - Step 7678: {'lr': 0.000482934692269538, 'samples': 1474368, 'steps': 7678, 'loss/train': 2.4510346055030823} 01/28/2022 22:41:42 - INFO - codeparrot_training - Step 7679: {'lr': 0.00048292875008185896, 'samples': 1474560, 'steps': 7679, 'loss/train': 2.1202942728996277} 01/28/2022 22:41:47 - INFO - codeparrot_training - Step 7680: {'lr': 0.0004829228068963872, 'samples': 1474752, 'steps': 7680, 'loss/train': 1.7196636199951172} 01/28/2022 22:41:51 - INFO - codeparrot_training - Step 7681: {'lr': 0.00048291686271314816, 'samples': 1474944, 'steps': 7681, 'loss/train': 1.6791096925735474} 01/28/2022 22:41:55 - INFO - codeparrot_training - Step 7682: {'lr': 0.0004829109175321671, 'samples': 1475136, 'steps': 7682, 'loss/train': 1.0374304354190826} 01/28/2022 22:42:00 - INFO - codeparrot_training - Step 7683: {'lr': 0.00048290497135346965, 'samples': 1475328, 'steps': 7683, 'loss/train': 2.4853352308273315} 01/28/2022 22:42:04 - INFO - codeparrot_training - Step 7684: {'lr': 0.0004828990241770813, 'samples': 1475520, 'steps': 7684, 'loss/train': 1.907017171382904} 01/28/2022 22:42:09 - INFO - codeparrot_training - Step 7685: {'lr': 0.0004828930760030275, 'samples': 1475712, 'steps': 7685, 'loss/train': 1.3256767094135284} 01/28/2022 22:42:13 - INFO - codeparrot_training - Step 7686: {'lr': 0.0004828871268313337, 'samples': 1475904, 'steps': 7686, 'loss/train': 1.9390672445297241} 01/28/2022 22:42:17 - INFO - codeparrot_training - Step 7687: {'lr': 0.0004828811766620254, 'samples': 1476096, 'steps': 7687, 'loss/train': 1.6032158732414246} 01/28/2022 22:42:22 - INFO - codeparrot_training - Step 7688: {'lr': 0.00048287522549512806, 'samples': 1476288, 'steps': 7688, 'loss/train': 1.6149219274520874} 01/28/2022 22:42:27 - INFO - codeparrot_training - Step 7689: {'lr': 0.0004828692733306672, 'samples': 1476480, 'steps': 7689, 'loss/train': 1.1240260004997253} 01/28/2022 22:42:31 - INFO - codeparrot_training - Step 7690: {'lr': 0.0004828633201686684, 'samples': 1476672, 'steps': 7690, 'loss/train': 2.6254663467407227} 01/28/2022 22:42:35 - INFO - codeparrot_training - Step 7691: {'lr': 0.00048285736600915696, 'samples': 1476864, 'steps': 7691, 'loss/train': 1.912727415561676} 01/28/2022 22:42:40 - INFO - codeparrot_training - Step 7692: {'lr': 0.00048285141085215857, 'samples': 1477056, 'steps': 7692, 'loss/train': 1.6403388977050781} 01/28/2022 22:42:46 - INFO - codeparrot_training - Step 7693: {'lr': 0.0004828454546976987, 'samples': 1477248, 'steps': 7693, 'loss/train': 1.146596074104309} 01/28/2022 22:42:50 - INFO - codeparrot_training - Step 7694: {'lr': 0.00048283949754580283, 'samples': 1477440, 'steps': 7694, 'loss/train': 2.2760385274887085} 01/28/2022 22:42:54 - INFO - codeparrot_training - Step 7695: {'lr': 0.00048283353939649644, 'samples': 1477632, 'steps': 7695, 'loss/train': 1.9863781929016113} 01/28/2022 22:42:58 - INFO - codeparrot_training - Step 7696: {'lr': 0.0004828275802498051, 'samples': 1477824, 'steps': 7696, 'loss/train': 1.9073404669761658} 01/28/2022 22:43:03 - INFO - codeparrot_training - Step 7697: {'lr': 0.0004828216201057544, 'samples': 1478016, 'steps': 7697, 'loss/train': 0.6161861568689346} 01/28/2022 22:43:08 - INFO - codeparrot_training - Step 7698: {'lr': 0.00048281565896436966, 'samples': 1478208, 'steps': 7698, 'loss/train': 2.0065916776657104} 01/28/2022 22:43:12 - INFO - codeparrot_training - Step 7699: {'lr': 0.0004828096968256767, 'samples': 1478400, 'steps': 7699, 'loss/train': 2.8394059538841248} 01/28/2022 22:43:16 - INFO - codeparrot_training - Step 7700: {'lr': 0.00048280373368970086, 'samples': 1478592, 'steps': 7700, 'loss/train': 1.3395940661430359} 01/28/2022 22:43:21 - INFO - codeparrot_training - Step 7701: {'lr': 0.0004827977695564678, 'samples': 1478784, 'steps': 7701, 'loss/train': 1.50167316198349} 01/28/2022 22:43:25 - INFO - codeparrot_training - Step 7702: {'lr': 0.000482791804426003, 'samples': 1478976, 'steps': 7702, 'loss/train': 1.83144611120224} 01/28/2022 22:43:30 - INFO - codeparrot_training - Step 7703: {'lr': 0.00048278583829833207, 'samples': 1479168, 'steps': 7703, 'loss/train': 0.6240895986557007} 01/28/2022 22:43:34 - INFO - codeparrot_training - Step 7704: {'lr': 0.00048277987117348043, 'samples': 1479360, 'steps': 7704, 'loss/train': 1.8228356838226318} 01/28/2022 22:43:39 - INFO - codeparrot_training - Step 7705: {'lr': 0.00048277390305147386, 'samples': 1479552, 'steps': 7705, 'loss/train': 1.8017064929008484} 01/28/2022 22:43:43 - INFO - codeparrot_training - Step 7706: {'lr': 0.0004827679339323377, 'samples': 1479744, 'steps': 7706, 'loss/train': 1.7246071100234985} 01/28/2022 22:43:47 - INFO - codeparrot_training - Step 7707: {'lr': 0.0004827619638160977, 'samples': 1479936, 'steps': 7707, 'loss/train': 2.2096978425979614} 01/28/2022 22:43:53 - INFO - codeparrot_training - Step 7708: {'lr': 0.00048275599270277927, 'samples': 1480128, 'steps': 7708, 'loss/train': 1.6857830286026} 01/28/2022 22:43:57 - INFO - codeparrot_training - Step 7709: {'lr': 0.00048275002059240815, 'samples': 1480320, 'steps': 7709, 'loss/train': 1.7076392769813538} 01/28/2022 22:44:02 - INFO - codeparrot_training - Step 7710: {'lr': 0.00048274404748500975, 'samples': 1480512, 'steps': 7710, 'loss/train': 1.7544835209846497} 01/28/2022 22:44:06 - INFO - codeparrot_training - Step 7711: {'lr': 0.0004827380733806099, 'samples': 1480704, 'steps': 7711, 'loss/train': 1.0808563828468323} 01/28/2022 22:44:10 - INFO - codeparrot_training - Step 7712: {'lr': 0.0004827320982792339, 'samples': 1480896, 'steps': 7712, 'loss/train': 1.575758457183838} 01/28/2022 22:44:15 - INFO - codeparrot_training - Step 7713: {'lr': 0.0004827261221809076, 'samples': 1481088, 'steps': 7713, 'loss/train': 1.9134576916694641} 01/28/2022 22:44:20 - INFO - codeparrot_training - Step 7714: {'lr': 0.00048272014508565645, 'samples': 1481280, 'steps': 7714, 'loss/train': 1.7152517437934875} 01/28/2022 22:44:24 - INFO - codeparrot_training - Step 7715: {'lr': 0.00048271416699350613, 'samples': 1481472, 'steps': 7715, 'loss/train': 1.9636364579200745} 01/28/2022 22:44:28 - INFO - codeparrot_training - Step 7716: {'lr': 0.0004827081879044821, 'samples': 1481664, 'steps': 7716, 'loss/train': 1.947320580482483} 01/28/2022 22:44:32 - INFO - codeparrot_training - Step 7717: {'lr': 0.00048270220781861025, 'samples': 1481856, 'steps': 7717, 'loss/train': 2.0533327460289} 01/28/2022 22:44:38 - INFO - codeparrot_training - Step 7718: {'lr': 0.000482696226735916, 'samples': 1482048, 'steps': 7718, 'loss/train': 1.920486867427826} 01/28/2022 22:44:42 - INFO - codeparrot_training - Step 7719: {'lr': 0.00048269024465642487, 'samples': 1482240, 'steps': 7719, 'loss/train': 1.9186987280845642} 01/28/2022 22:44:47 - INFO - codeparrot_training - Step 7720: {'lr': 0.00048268426158016274, 'samples': 1482432, 'steps': 7720, 'loss/train': 1.0960519909858704} 01/28/2022 22:44:51 - INFO - codeparrot_training - Step 7721: {'lr': 0.0004826782775071551, 'samples': 1482624, 'steps': 7721, 'loss/train': 1.291634202003479} 01/28/2022 22:44:56 - INFO - codeparrot_training - Step 7722: {'lr': 0.00048267229243742753, 'samples': 1482816, 'steps': 7722, 'loss/train': 2.1986337304115295} 01/28/2022 22:45:01 - INFO - codeparrot_training - Step 7723: {'lr': 0.00048266630637100585, 'samples': 1483008, 'steps': 7723, 'loss/train': 2.0472949147224426} 01/28/2022 22:45:05 - INFO - codeparrot_training - Step 7724: {'lr': 0.00048266031930791555, 'samples': 1483200, 'steps': 7724, 'loss/train': 1.8299279808998108} 01/28/2022 22:45:09 - INFO - codeparrot_training - Step 7725: {'lr': 0.00048265433124818226, 'samples': 1483392, 'steps': 7725, 'loss/train': 1.8744022250175476} 01/28/2022 22:45:13 - INFO - codeparrot_training - Step 7726: {'lr': 0.00048264834219183175, 'samples': 1483584, 'steps': 7726, 'loss/train': 1.1937549412250519} 01/28/2022 22:45:18 - INFO - codeparrot_training - Step 7727: {'lr': 0.00048264235213888964, 'samples': 1483776, 'steps': 7727, 'loss/train': 1.6685122847557068} 01/28/2022 22:45:23 - INFO - codeparrot_training - Step 7728: {'lr': 0.00048263636108938153, 'samples': 1483968, 'steps': 7728, 'loss/train': 1.583180844783783} 01/28/2022 22:45:27 - INFO - codeparrot_training - Step 7729: {'lr': 0.0004826303690433331, 'samples': 1484160, 'steps': 7729, 'loss/train': 1.9071539640426636} 01/28/2022 22:45:31 - INFO - codeparrot_training - Step 7730: {'lr': 0.0004826243760007701, 'samples': 1484352, 'steps': 7730, 'loss/train': 1.847033679485321} 01/28/2022 22:45:35 - INFO - codeparrot_training - Step 7731: {'lr': 0.00048261838196171804, 'samples': 1484544, 'steps': 7731, 'loss/train': 1.4648543894290924} 01/28/2022 22:45:41 - INFO - codeparrot_training - Step 7732: {'lr': 0.0004826123869262028, 'samples': 1484736, 'steps': 7732, 'loss/train': 2.3530304431915283} 01/28/2022 22:45:45 - INFO - codeparrot_training - Step 7733: {'lr': 0.0004826063908942499, 'samples': 1484928, 'steps': 7733, 'loss/train': 1.7940094470977783} 01/28/2022 22:45:49 - INFO - codeparrot_training - Step 7734: {'lr': 0.00048260039386588513, 'samples': 1485120, 'steps': 7734, 'loss/train': 1.5739502906799316} 01/28/2022 22:45:54 - INFO - codeparrot_training - Step 7735: {'lr': 0.00048259439584113405, 'samples': 1485312, 'steps': 7735, 'loss/train': 1.9988577961921692} 01/28/2022 22:45:58 - INFO - codeparrot_training - Step 7736: {'lr': 0.00048258839682002253, 'samples': 1485504, 'steps': 7736, 'loss/train': 1.3989956974983215} 01/28/2022 22:46:04 - INFO - codeparrot_training - Step 7737: {'lr': 0.0004825823968025761, 'samples': 1485696, 'steps': 7737, 'loss/train': 1.6699812412261963} 01/28/2022 22:46:08 - INFO - codeparrot_training - Step 7738: {'lr': 0.0004825763957888206, 'samples': 1485888, 'steps': 7738, 'loss/train': 2.154433786869049} 01/28/2022 22:46:12 - INFO - codeparrot_training - Step 7739: {'lr': 0.00048257039377878165, 'samples': 1486080, 'steps': 7739, 'loss/train': 1.2044959366321564} 01/28/2022 22:46:17 - INFO - codeparrot_training - Step 7740: {'lr': 0.00048256439077248495, 'samples': 1486272, 'steps': 7740, 'loss/train': 2.2673356533050537} 01/28/2022 22:46:21 - INFO - codeparrot_training - Step 7741: {'lr': 0.00048255838676995624, 'samples': 1486464, 'steps': 7741, 'loss/train': 3.2484050989151} 01/28/2022 22:46:26 - INFO - codeparrot_training - Step 7742: {'lr': 0.00048255238177122127, 'samples': 1486656, 'steps': 7742, 'loss/train': 1.277833342552185} 01/28/2022 22:46:30 - INFO - codeparrot_training - Step 7743: {'lr': 0.0004825463757763058, 'samples': 1486848, 'steps': 7743, 'loss/train': 2.257102847099304} 01/28/2022 22:46:35 - INFO - codeparrot_training - Step 7744: {'lr': 0.00048254036878523537, 'samples': 1487040, 'steps': 7744, 'loss/train': 1.8327194452285767} 01/28/2022 22:46:39 - INFO - codeparrot_training - Step 7745: {'lr': 0.00048253436079803594, 'samples': 1487232, 'steps': 7745, 'loss/train': 1.3854607343673706} 01/28/2022 22:46:43 - INFO - codeparrot_training - Step 7746: {'lr': 0.0004825283518147331, 'samples': 1487424, 'steps': 7746, 'loss/train': 1.096999078989029} 01/28/2022 22:46:49 - INFO - codeparrot_training - Step 7747: {'lr': 0.00048252234183535265, 'samples': 1487616, 'steps': 7747, 'loss/train': 2.279948115348816} 01/28/2022 22:46:53 - INFO - codeparrot_training - Step 7748: {'lr': 0.0004825163308599203, 'samples': 1487808, 'steps': 7748, 'loss/train': 0.9434197247028351} 01/28/2022 22:46:57 - INFO - codeparrot_training - Step 7749: {'lr': 0.0004825103188884619, 'samples': 1488000, 'steps': 7749, 'loss/train': 1.7352585196495056} 01/28/2022 22:47:01 - INFO - codeparrot_training - Step 7750: {'lr': 0.000482504305921003, 'samples': 1488192, 'steps': 7750, 'loss/train': 1.465461641550064} 01/28/2022 22:47:06 - INFO - codeparrot_training - Step 7751: {'lr': 0.00048249829195756954, 'samples': 1488384, 'steps': 7751, 'loss/train': 1.6755974292755127} 01/28/2022 22:47:12 - INFO - codeparrot_training - Step 7752: {'lr': 0.0004824922769981873, 'samples': 1488576, 'steps': 7752, 'loss/train': 1.0165978074073792} 01/28/2022 22:47:16 - INFO - codeparrot_training - Step 7753: {'lr': 0.0004824862610428819, 'samples': 1488768, 'steps': 7753, 'loss/train': 2.283262610435486} 01/28/2022 22:47:20 - INFO - codeparrot_training - Step 7754: {'lr': 0.0004824802440916792, 'samples': 1488960, 'steps': 7754, 'loss/train': 1.613469421863556} 01/28/2022 22:47:24 - INFO - codeparrot_training - Step 7755: {'lr': 0.0004824742261446049, 'samples': 1489152, 'steps': 7755, 'loss/train': 1.856442153453827} 01/28/2022 22:47:29 - INFO - codeparrot_training - Step 7756: {'lr': 0.0004824682072016849, 'samples': 1489344, 'steps': 7756, 'loss/train': 0.5758839100599289} 01/28/2022 22:47:34 - INFO - codeparrot_training - Step 7757: {'lr': 0.00048246218726294486, 'samples': 1489536, 'steps': 7757, 'loss/train': 0.8990780711174011} 01/28/2022 22:47:38 - INFO - codeparrot_training - Step 7758: {'lr': 0.0004824561663284107, 'samples': 1489728, 'steps': 7758, 'loss/train': 0.10528450459241867} 01/28/2022 22:47:43 - INFO - codeparrot_training - Step 7759: {'lr': 0.0004824501443981081, 'samples': 1489920, 'steps': 7759, 'loss/train': 1.3882920742034912} 01/28/2022 22:47:47 - INFO - codeparrot_training - Step 7760: {'lr': 0.00048244412147206283, 'samples': 1490112, 'steps': 7760, 'loss/train': 1.7979297637939453} 01/28/2022 22:47:51 - INFO - codeparrot_training - Step 7761: {'lr': 0.00048243809755030086, 'samples': 1490304, 'steps': 7761, 'loss/train': 6.5081799030303955} 01/28/2022 22:47:57 - INFO - codeparrot_training - Step 7762: {'lr': 0.00048243207263284785, 'samples': 1490496, 'steps': 7762, 'loss/train': 1.7764880061149597} 01/28/2022 22:48:01 - INFO - codeparrot_training - Step 7763: {'lr': 0.0004824260467197296, 'samples': 1490688, 'steps': 7763, 'loss/train': 1.58386892080307} 01/28/2022 22:48:06 - INFO - codeparrot_training - Step 7764: {'lr': 0.000482420019810972, 'samples': 1490880, 'steps': 7764, 'loss/train': 1.7825480103492737} 01/28/2022 22:48:10 - INFO - codeparrot_training - Step 7765: {'lr': 0.00048241399190660086, 'samples': 1491072, 'steps': 7765, 'loss/train': 0.40012241899967194} 01/28/2022 22:48:14 - INFO - codeparrot_training - Step 7766: {'lr': 0.0004824079630066419, 'samples': 1491264, 'steps': 7766, 'loss/train': 1.6184123754501343} 01/28/2022 22:48:19 - INFO - codeparrot_training - Step 7767: {'lr': 0.0004824019331111211, 'samples': 1491456, 'steps': 7767, 'loss/train': 2.223420739173889} 01/28/2022 22:48:24 - INFO - codeparrot_training - Step 7768: {'lr': 0.0004823959022200642, 'samples': 1491648, 'steps': 7768, 'loss/train': 2.2066274285316467} 01/28/2022 22:48:28 - INFO - codeparrot_training - Step 7769: {'lr': 0.00048238987033349706, 'samples': 1491840, 'steps': 7769, 'loss/train': 1.4459846913814545} 01/28/2022 22:48:32 - INFO - codeparrot_training - Step 7770: {'lr': 0.0004823838374514455, 'samples': 1492032, 'steps': 7770, 'loss/train': 1.1931372284889221} 01/28/2022 22:48:36 - INFO - codeparrot_training - Step 7771: {'lr': 0.00048237780357393535, 'samples': 1492224, 'steps': 7771, 'loss/train': 1.3314989805221558} 01/28/2022 22:48:42 - INFO - codeparrot_training - Step 7772: {'lr': 0.00048237176870099256, 'samples': 1492416, 'steps': 7772, 'loss/train': 2.0279041528701782} 01/28/2022 22:48:46 - INFO - codeparrot_training - Step 7773: {'lr': 0.0004823657328326427, 'samples': 1492608, 'steps': 7773, 'loss/train': 0.14337273314595222} 01/28/2022 22:48:50 - INFO - codeparrot_training - Step 7774: {'lr': 0.000482359695968912, 'samples': 1492800, 'steps': 7774, 'loss/train': 1.78187495470047} 01/28/2022 22:48:55 - INFO - codeparrot_training - Step 7775: {'lr': 0.0004823536581098261, 'samples': 1492992, 'steps': 7775, 'loss/train': 1.6278226375579834} 01/28/2022 22:48:59 - INFO - codeparrot_training - Step 7776: {'lr': 0.00048234761925541094, 'samples': 1493184, 'steps': 7776, 'loss/train': 2.4147530794143677} 01/28/2022 22:49:05 - INFO - codeparrot_training - Step 7777: {'lr': 0.0004823415794056923, 'samples': 1493376, 'steps': 7777, 'loss/train': 1.8701283931732178} 01/28/2022 22:49:09 - INFO - codeparrot_training - Step 7778: {'lr': 0.00048233553856069617, 'samples': 1493568, 'steps': 7778, 'loss/train': 1.5410022139549255} 01/28/2022 22:49:13 - INFO - codeparrot_training - Step 7779: {'lr': 0.00048232949672044834, 'samples': 1493760, 'steps': 7779, 'loss/train': 1.5734660625457764} 01/28/2022 22:49:17 - INFO - codeparrot_training - Step 7780: {'lr': 0.0004823234538849747, 'samples': 1493952, 'steps': 7780, 'loss/train': 0.3798581063747406} 01/28/2022 22:49:22 - INFO - codeparrot_training - Step 7781: {'lr': 0.0004823174100543012, 'samples': 1494144, 'steps': 7781, 'loss/train': 2.848923683166504} 01/28/2022 22:49:27 - INFO - codeparrot_training - Step 7782: {'lr': 0.0004823113652284536, 'samples': 1494336, 'steps': 7782, 'loss/train': 1.7794327139854431} 01/28/2022 22:49:31 - INFO - codeparrot_training - Step 7783: {'lr': 0.00048230531940745793, 'samples': 1494528, 'steps': 7783, 'loss/train': 1.786421298980713} 01/28/2022 22:49:36 - INFO - codeparrot_training - Step 7784: {'lr': 0.0004822992725913401, 'samples': 1494720, 'steps': 7784, 'loss/train': 2.3301817774772644} 01/28/2022 22:49:40 - INFO - codeparrot_training - Step 7785: {'lr': 0.00048229322478012584, 'samples': 1494912, 'steps': 7785, 'loss/train': 1.716612696647644} 01/28/2022 22:49:44 - INFO - codeparrot_training - Step 7786: {'lr': 0.0004822871759738412, 'samples': 1495104, 'steps': 7786, 'loss/train': 1.6549567580223083} 01/28/2022 22:49:50 - INFO - codeparrot_training - Step 7787: {'lr': 0.0004822811261725121, 'samples': 1495296, 'steps': 7787, 'loss/train': 0.22413452714681625} 01/28/2022 22:49:54 - INFO - codeparrot_training - Step 7788: {'lr': 0.0004822750753761644, 'samples': 1495488, 'steps': 7788, 'loss/train': 1.9909876585006714} 01/28/2022 22:49:58 - INFO - codeparrot_training - Step 7789: {'lr': 0.00048226902358482405, 'samples': 1495680, 'steps': 7789, 'loss/train': 1.2735357284545898} 01/28/2022 22:50:02 - INFO - codeparrot_training - Step 7790: {'lr': 0.0004822629707985169, 'samples': 1495872, 'steps': 7790, 'loss/train': 1.441390335559845} 01/28/2022 22:50:09 - INFO - codeparrot_training - Step 7791: {'lr': 0.00048225691701726895, 'samples': 1496064, 'steps': 7791, 'loss/train': 2.358195126056671} 01/28/2022 22:50:13 - INFO - codeparrot_training - Step 7792: {'lr': 0.00048225086224110614, 'samples': 1496256, 'steps': 7792, 'loss/train': 1.622381329536438} 01/28/2022 22:50:17 - INFO - codeparrot_training - Step 7793: {'lr': 0.00048224480647005437, 'samples': 1496448, 'steps': 7793, 'loss/train': 3.1240246295928955} 01/28/2022 22:50:22 - INFO - codeparrot_training - Step 7794: {'lr': 0.0004822387497041396, 'samples': 1496640, 'steps': 7794, 'loss/train': 2.4948830008506775} 01/28/2022 22:50:26 - INFO - codeparrot_training - Step 7795: {'lr': 0.00048223269194338776, 'samples': 1496832, 'steps': 7795, 'loss/train': 0.6531528979539871} 01/28/2022 22:50:30 - INFO - codeparrot_training - Step 7796: {'lr': 0.0004822266331878248, 'samples': 1497024, 'steps': 7796, 'loss/train': 2.4236943125724792} 01/28/2022 22:50:36 - INFO - codeparrot_training - Step 7797: {'lr': 0.0004822205734374767, 'samples': 1497216, 'steps': 7797, 'loss/train': 1.6128732562065125} 01/28/2022 22:50:40 - INFO - codeparrot_training - Step 7798: {'lr': 0.00048221451269236937, 'samples': 1497408, 'steps': 7798, 'loss/train': 0.3412386327981949} 01/28/2022 22:50:44 - INFO - codeparrot_training - Step 7799: {'lr': 0.0004822084509525289, 'samples': 1497600, 'steps': 7799, 'loss/train': 1.8485793471336365} 01/28/2022 22:50:48 - INFO - codeparrot_training - Step 7800: {'lr': 0.0004822023882179811, 'samples': 1497792, 'steps': 7800, 'loss/train': 1.4283592700958252} 01/28/2022 22:50:53 - INFO - codeparrot_training - Step 7801: {'lr': 0.00048219632448875195, 'samples': 1497984, 'steps': 7801, 'loss/train': 1.6105186343193054} 01/28/2022 22:50:58 - INFO - codeparrot_training - Step 7802: {'lr': 0.0004821902597648675, 'samples': 1498176, 'steps': 7802, 'loss/train': 1.3194547891616821} 01/28/2022 22:51:02 - INFO - codeparrot_training - Step 7803: {'lr': 0.0004821841940463538, 'samples': 1498368, 'steps': 7803, 'loss/train': 1.1962145268917084} 01/28/2022 22:51:07 - INFO - codeparrot_training - Step 7804: {'lr': 0.0004821781273332366, 'samples': 1498560, 'steps': 7804, 'loss/train': 1.6840953826904297} 01/28/2022 22:51:11 - INFO - codeparrot_training - Step 7805: {'lr': 0.00048217205962554214, 'samples': 1498752, 'steps': 7805, 'loss/train': 2.8171550631523132} 01/28/2022 22:51:15 - INFO - codeparrot_training - Step 7806: {'lr': 0.0004821659909232963, 'samples': 1498944, 'steps': 7806, 'loss/train': 1.954178810119629} 01/28/2022 22:51:21 - INFO - codeparrot_training - Step 7807: {'lr': 0.000482159921226525, 'samples': 1499136, 'steps': 7807, 'loss/train': 1.6521581411361694} 01/28/2022 22:51:26 - INFO - codeparrot_training - Step 7808: {'lr': 0.00048215385053525434, 'samples': 1499328, 'steps': 7808, 'loss/train': 2.250862419605255} 01/28/2022 22:51:30 - INFO - codeparrot_training - Step 7809: {'lr': 0.0004821477788495103, 'samples': 1499520, 'steps': 7809, 'loss/train': 0.5414690226316452} 01/28/2022 22:51:34 - INFO - codeparrot_training - Step 7810: {'lr': 0.0004821417061693189, 'samples': 1499712, 'steps': 7810, 'loss/train': 1.5846124291419983} 01/28/2022 22:51:38 - INFO - codeparrot_training - Step 7811: {'lr': 0.00048213563249470615, 'samples': 1499904, 'steps': 7811, 'loss/train': 1.892525553703308} 01/28/2022 22:51:44 - INFO - codeparrot_training - Step 7812: {'lr': 0.00048212955782569805, 'samples': 1500096, 'steps': 7812, 'loss/train': 2.114070475101471} 01/28/2022 22:51:48 - INFO - codeparrot_training - Step 7813: {'lr': 0.00048212348216232064, 'samples': 1500288, 'steps': 7813, 'loss/train': 1.8503935933113098} 01/28/2022 22:51:52 - INFO - codeparrot_training - Step 7814: {'lr': 0.0004821174055045999, 'samples': 1500480, 'steps': 7814, 'loss/train': 2.086417078971863} 01/28/2022 22:51:56 - INFO - codeparrot_training - Step 7815: {'lr': 0.000482111327852562, 'samples': 1500672, 'steps': 7815, 'loss/train': 1.1593791246414185} 01/28/2022 22:52:00 - INFO - codeparrot_training - Step 7816: {'lr': 0.0004821052492062328, 'samples': 1500864, 'steps': 7816, 'loss/train': 1.2584977447986603} 01/28/2022 22:52:06 - INFO - codeparrot_training - Step 7817: {'lr': 0.0004820991695656385, 'samples': 1501056, 'steps': 7817, 'loss/train': 2.026071310043335} 01/28/2022 22:52:10 - INFO - codeparrot_training - Step 7818: {'lr': 0.00048209308893080495, 'samples': 1501248, 'steps': 7818, 'loss/train': 0.7760607898235321} 01/28/2022 22:52:14 - INFO - codeparrot_training - Step 7819: {'lr': 0.00048208700730175834, 'samples': 1501440, 'steps': 7819, 'loss/train': 2.013343870639801} 01/28/2022 22:52:18 - INFO - codeparrot_training - Step 7820: {'lr': 0.0004820809246785247, 'samples': 1501632, 'steps': 7820, 'loss/train': 1.920718789100647} 01/28/2022 22:52:23 - INFO - codeparrot_training - Step 7821: {'lr': 0.00048207484106113, 'samples': 1501824, 'steps': 7821, 'loss/train': 0.6130354553461075} 01/28/2022 22:52:29 - INFO - codeparrot_training - Step 7822: {'lr': 0.0004820687564496005, 'samples': 1502016, 'steps': 7822, 'loss/train': 2.269716203212738} 01/28/2022 22:52:33 - INFO - codeparrot_training - Step 7823: {'lr': 0.00048206267084396204, 'samples': 1502208, 'steps': 7823, 'loss/train': 2.88648122549057} 01/28/2022 22:52:37 - INFO - codeparrot_training - Step 7824: {'lr': 0.0004820565842442408, 'samples': 1502400, 'steps': 7824, 'loss/train': 1.3071382641792297} 01/28/2022 22:52:41 - INFO - codeparrot_training - Step 7825: {'lr': 0.00048205049665046287, 'samples': 1502592, 'steps': 7825, 'loss/train': 1.5897930264472961} 01/28/2022 22:52:46 - INFO - codeparrot_training - Step 7826: {'lr': 0.0004820444080626543, 'samples': 1502784, 'steps': 7826, 'loss/train': 1.9271344542503357} 01/28/2022 22:52:51 - INFO - codeparrot_training - Step 7827: {'lr': 0.00048203831848084115, 'samples': 1502976, 'steps': 7827, 'loss/train': 4.036298632621765} 01/28/2022 22:52:55 - INFO - codeparrot_training - Step 7828: {'lr': 0.0004820322279050495, 'samples': 1503168, 'steps': 7828, 'loss/train': 1.6507419347763062} 01/28/2022 22:52:59 - INFO - codeparrot_training - Step 7829: {'lr': 0.00048202613633530555, 'samples': 1503360, 'steps': 7829, 'loss/train': 2.2848540544509888} 01/28/2022 22:53:04 - INFO - codeparrot_training - Step 7830: {'lr': 0.00048202004377163524, 'samples': 1503552, 'steps': 7830, 'loss/train': 1.7544211149215698} 01/28/2022 22:53:08 - INFO - codeparrot_training - Step 7831: {'lr': 0.00048201395021406476, 'samples': 1503744, 'steps': 7831, 'loss/train': 1.7418156266212463} 01/28/2022 22:53:13 - INFO - codeparrot_training - Step 7832: {'lr': 0.0004820078556626202, 'samples': 1503936, 'steps': 7832, 'loss/train': 2.183461010456085} 01/28/2022 22:53:17 - INFO - codeparrot_training - Step 7833: {'lr': 0.0004820017601173276, 'samples': 1504128, 'steps': 7833, 'loss/train': 0.1441500149667263} 01/28/2022 22:53:22 - INFO - codeparrot_training - Step 7834: {'lr': 0.00048199566357821314, 'samples': 1504320, 'steps': 7834, 'loss/train': 2.0947242379188538} 01/28/2022 22:53:26 - INFO - codeparrot_training - Step 7835: {'lr': 0.00048198956604530297, 'samples': 1504512, 'steps': 7835, 'loss/train': 1.6909832954406738} 01/28/2022 22:53:32 - INFO - codeparrot_training - Step 7836: {'lr': 0.0004819834675186231, 'samples': 1504704, 'steps': 7836, 'loss/train': 1.6142538785934448} 01/28/2022 22:53:36 - INFO - codeparrot_training - Step 7837: {'lr': 0.0004819773679981998, 'samples': 1504896, 'steps': 7837, 'loss/train': 2.6545302271842957} 01/28/2022 22:53:40 - INFO - codeparrot_training - Step 7838: {'lr': 0.0004819712674840591, 'samples': 1505088, 'steps': 7838, 'loss/train': 1.731411337852478} 01/28/2022 22:53:45 - INFO - codeparrot_training - Step 7839: {'lr': 0.00048196516597622706, 'samples': 1505280, 'steps': 7839, 'loss/train': 2.591281235218048} 01/28/2022 22:53:49 - INFO - codeparrot_training - Step 7840: {'lr': 0.00048195906347473, 'samples': 1505472, 'steps': 7840, 'loss/train': 1.4167825877666473} 01/28/2022 22:53:54 - INFO - codeparrot_training - Step 7841: {'lr': 0.00048195295997959393, 'samples': 1505664, 'steps': 7841, 'loss/train': 2.2072616815567017} 01/28/2022 22:53:58 - INFO - codeparrot_training - Step 7842: {'lr': 0.00048194685549084507, 'samples': 1505856, 'steps': 7842, 'loss/train': 1.796847939491272} 01/28/2022 22:54:03 - INFO - codeparrot_training - Step 7843: {'lr': 0.00048194075000850944, 'samples': 1506048, 'steps': 7843, 'loss/train': 2.3094820976257324} 01/28/2022 22:54:07 - INFO - codeparrot_training - Step 7844: {'lr': 0.0004819346435326134, 'samples': 1506240, 'steps': 7844, 'loss/train': 1.0581664144992828} 01/28/2022 22:54:11 - INFO - codeparrot_training - Step 7845: {'lr': 0.000481928536063183, 'samples': 1506432, 'steps': 7845, 'loss/train': 2.415346920490265} 01/28/2022 22:54:17 - INFO - codeparrot_training - Step 7846: {'lr': 0.0004819224276002443, 'samples': 1506624, 'steps': 7846, 'loss/train': 1.8997651934623718} 01/28/2022 22:54:21 - INFO - codeparrot_training - Step 7847: {'lr': 0.0004819163181438236, 'samples': 1506816, 'steps': 7847, 'loss/train': 1.134177803993225} 01/28/2022 22:54:26 - INFO - codeparrot_training - Step 7848: {'lr': 0.000481910207693947, 'samples': 1507008, 'steps': 7848, 'loss/train': 2.0054853558540344} 01/28/2022 22:54:30 - INFO - codeparrot_training - Step 7849: {'lr': 0.0004819040962506408, 'samples': 1507200, 'steps': 7849, 'loss/train': 1.7577536702156067} 01/28/2022 22:54:34 - INFO - codeparrot_training - Step 7850: {'lr': 0.000481897983813931, 'samples': 1507392, 'steps': 7850, 'loss/train': 2.3485684990882874} 01/28/2022 22:54:39 - INFO - codeparrot_training - Step 7851: {'lr': 0.00048189187038384396, 'samples': 1507584, 'steps': 7851, 'loss/train': 1.566707968711853} 01/28/2022 22:54:43 - INFO - codeparrot_training - Step 7852: {'lr': 0.00048188575596040575, 'samples': 1507776, 'steps': 7852, 'loss/train': 1.599834680557251} 01/28/2022 22:54:48 - INFO - codeparrot_training - Step 7853: {'lr': 0.00048187964054364254, 'samples': 1507968, 'steps': 7853, 'loss/train': 1.2272822856903076} 01/28/2022 22:54:52 - INFO - codeparrot_training - Step 7854: {'lr': 0.0004818735241335807, 'samples': 1508160, 'steps': 7854, 'loss/train': 1.8299937844276428} 01/28/2022 22:54:56 - INFO - codeparrot_training - Step 7855: {'lr': 0.00048186740673024614, 'samples': 1508352, 'steps': 7855, 'loss/train': 1.7381227612495422} 01/28/2022 22:55:01 - INFO - codeparrot_training - Step 7856: {'lr': 0.00048186128833366536, 'samples': 1508544, 'steps': 7856, 'loss/train': 1.425907462835312} 01/28/2022 22:55:06 - INFO - codeparrot_training - Step 7857: {'lr': 0.0004818551689438644, 'samples': 1508736, 'steps': 7857, 'loss/train': 1.6962416768074036} 01/28/2022 22:55:10 - INFO - codeparrot_training - Step 7858: {'lr': 0.00048184904856086953, 'samples': 1508928, 'steps': 7858, 'loss/train': 2.293864667415619} 01/28/2022 22:55:14 - INFO - codeparrot_training - Step 7859: {'lr': 0.0004818429271847069, 'samples': 1509120, 'steps': 7859, 'loss/train': 1.8066426515579224} 01/28/2022 22:55:18 - INFO - codeparrot_training - Step 7860: {'lr': 0.00048183680481540293, 'samples': 1509312, 'steps': 7860, 'loss/train': 2.5659345388412476} 01/28/2022 22:55:24 - INFO - codeparrot_training - Step 7861: {'lr': 0.0004818306814529836, 'samples': 1509504, 'steps': 7861, 'loss/train': 1.9416285753250122} 01/28/2022 22:55:28 - INFO - codeparrot_training - Step 7862: {'lr': 0.00048182455709747525, 'samples': 1509696, 'steps': 7862, 'loss/train': 1.9317662715911865} 01/28/2022 22:55:32 - INFO - codeparrot_training - Step 7863: {'lr': 0.0004818184317489041, 'samples': 1509888, 'steps': 7863, 'loss/train': 1.7620412707328796} 01/28/2022 22:55:37 - INFO - codeparrot_training - Step 7864: {'lr': 0.00048181230540729643, 'samples': 1510080, 'steps': 7864, 'loss/train': 1.1632209420204163} 01/28/2022 22:55:41 - INFO - codeparrot_training - Step 7865: {'lr': 0.00048180617807267844, 'samples': 1510272, 'steps': 7865, 'loss/train': 1.684711754322052} 01/28/2022 22:55:47 - INFO - codeparrot_training - Step 7866: {'lr': 0.0004818000497450764, 'samples': 1510464, 'steps': 7866, 'loss/train': 2.140132248401642} 01/28/2022 22:55:51 - INFO - codeparrot_training - Step 7867: {'lr': 0.00048179392042451655, 'samples': 1510656, 'steps': 7867, 'loss/train': 1.6070880889892578} 01/28/2022 22:55:55 - INFO - codeparrot_training - Step 7868: {'lr': 0.0004817877901110251, 'samples': 1510848, 'steps': 7868, 'loss/train': 2.0577451586723328} 01/28/2022 22:55:59 - INFO - codeparrot_training - Step 7869: {'lr': 0.00048178165880462845, 'samples': 1511040, 'steps': 7869, 'loss/train': 1.6020948886871338} 01/28/2022 22:56:04 - INFO - codeparrot_training - Step 7870: {'lr': 0.0004817755265053527, 'samples': 1511232, 'steps': 7870, 'loss/train': 3.0345135927200317} 01/28/2022 22:56:09 - INFO - codeparrot_training - Step 7871: {'lr': 0.0004817693932132242, 'samples': 1511424, 'steps': 7871, 'loss/train': 0.7431640774011612} 01/28/2022 22:56:13 - INFO - codeparrot_training - Step 7872: {'lr': 0.0004817632589282693, 'samples': 1511616, 'steps': 7872, 'loss/train': 0.6876773089170456} 01/28/2022 22:56:17 - INFO - codeparrot_training - Step 7873: {'lr': 0.00048175712365051407, 'samples': 1511808, 'steps': 7873, 'loss/train': 2.0585973858833313} 01/28/2022 22:56:22 - INFO - codeparrot_training - Step 7874: {'lr': 0.00048175098737998504, 'samples': 1512000, 'steps': 7874, 'loss/train': 2.115124225616455} 01/28/2022 22:56:26 - INFO - codeparrot_training - Step 7875: {'lr': 0.0004817448501167082, 'samples': 1512192, 'steps': 7875, 'loss/train': 0.8173881769180298} 01/28/2022 22:56:31 - INFO - codeparrot_training - Step 7876: {'lr': 0.0004817387118607102, 'samples': 1512384, 'steps': 7876, 'loss/train': 2.084643244743347} 01/28/2022 22:56:35 - INFO - codeparrot_training - Step 7877: {'lr': 0.00048173257261201695, 'samples': 1512576, 'steps': 7877, 'loss/train': 1.6637849807739258} 01/28/2022 22:56:40 - INFO - codeparrot_training - Step 7878: {'lr': 0.00048172643237065504, 'samples': 1512768, 'steps': 7878, 'loss/train': 0.2689330503344536} 01/28/2022 22:56:44 - INFO - codeparrot_training - Step 7879: {'lr': 0.00048172029113665075, 'samples': 1512960, 'steps': 7879, 'loss/train': 2.458013355731964} 01/28/2022 22:56:48 - INFO - codeparrot_training - Step 7880: {'lr': 0.0004817141489100302, 'samples': 1513152, 'steps': 7880, 'loss/train': 2.3483328223228455} 01/28/2022 22:56:54 - INFO - codeparrot_training - Step 7881: {'lr': 0.00048170800569081985, 'samples': 1513344, 'steps': 7881, 'loss/train': 1.8395490646362305} 01/28/2022 22:56:59 - INFO - codeparrot_training - Step 7882: {'lr': 0.000481701861479046, 'samples': 1513536, 'steps': 7882, 'loss/train': 1.6207918524742126} 01/28/2022 22:57:03 - INFO - codeparrot_training - Step 7883: {'lr': 0.000481695716274735, 'samples': 1513728, 'steps': 7883, 'loss/train': 1.1021147668361664} 01/28/2022 22:57:07 - INFO - codeparrot_training - Step 7884: {'lr': 0.000481689570077913, 'samples': 1513920, 'steps': 7884, 'loss/train': 1.218127191066742} 01/28/2022 22:57:11 - INFO - codeparrot_training - Step 7885: {'lr': 0.00048168342288860646, 'samples': 1514112, 'steps': 7885, 'loss/train': 1.5890368223190308} 01/28/2022 22:57:17 - INFO - codeparrot_training - Step 7886: {'lr': 0.00048167727470684176, 'samples': 1514304, 'steps': 7886, 'loss/train': 1.6367177367210388} 01/28/2022 22:57:21 - INFO - codeparrot_training - Step 7887: {'lr': 0.0004816711255326452, 'samples': 1514496, 'steps': 7887, 'loss/train': 2.500396728515625} 01/28/2022 22:57:25 - INFO - codeparrot_training - Step 7888: {'lr': 0.00048166497536604306, 'samples': 1514688, 'steps': 7888, 'loss/train': 1.9063512682914734} 01/28/2022 22:57:29 - INFO - codeparrot_training - Step 7889: {'lr': 0.00048165882420706175, 'samples': 1514880, 'steps': 7889, 'loss/train': 1.8097137808799744} 01/28/2022 22:57:34 - INFO - codeparrot_training - Step 7890: {'lr': 0.0004816526720557276, 'samples': 1515072, 'steps': 7890, 'loss/train': 1.9702547192573547} 01/28/2022 22:57:40 - INFO - codeparrot_training - Step 7891: {'lr': 0.0004816465189120669, 'samples': 1515264, 'steps': 7891, 'loss/train': 7.387704849243164} 01/28/2022 22:57:44 - INFO - codeparrot_training - Step 7892: {'lr': 0.00048164036477610616, 'samples': 1515456, 'steps': 7892, 'loss/train': 0.9233295321464539} 01/28/2022 22:57:48 - INFO - codeparrot_training - Step 7893: {'lr': 0.0004816342096478716, 'samples': 1515648, 'steps': 7893, 'loss/train': 1.1141942739486694} 01/28/2022 22:57:52 - INFO - codeparrot_training - Step 7894: {'lr': 0.00048162805352738966, 'samples': 1515840, 'steps': 7894, 'loss/train': 1.475207895040512} 01/28/2022 22:57:57 - INFO - codeparrot_training - Step 7895: {'lr': 0.0004816218964146867, 'samples': 1516032, 'steps': 7895, 'loss/train': 1.8180726170539856} 01/28/2022 22:58:02 - INFO - codeparrot_training - Step 7896: {'lr': 0.000481615738309789, 'samples': 1516224, 'steps': 7896, 'loss/train': 1.7286189794540405} 01/28/2022 22:58:06 - INFO - codeparrot_training - Step 7897: {'lr': 0.00048160957921272306, 'samples': 1516416, 'steps': 7897, 'loss/train': 1.9479986429214478} 01/28/2022 22:58:10 - INFO - codeparrot_training - Step 7898: {'lr': 0.00048160341912351523, 'samples': 1516608, 'steps': 7898, 'loss/train': 1.9840874075889587} 01/28/2022 22:58:15 - INFO - codeparrot_training - Step 7899: {'lr': 0.00048159725804219195, 'samples': 1516800, 'steps': 7899, 'loss/train': 2.634193181991577} 01/28/2022 22:58:19 - INFO - codeparrot_training - Step 7900: {'lr': 0.00048159109596877954, 'samples': 1516992, 'steps': 7900, 'loss/train': 1.2341729700565338} 01/28/2022 22:58:24 - INFO - codeparrot_training - Step 7901: {'lr': 0.00048158493290330443, 'samples': 1517184, 'steps': 7901, 'loss/train': 1.513707160949707} 01/28/2022 22:58:28 - INFO - codeparrot_training - Step 7902: {'lr': 0.00048157876884579294, 'samples': 1517376, 'steps': 7902, 'loss/train': 1.430705577135086} 01/28/2022 22:58:33 - INFO - codeparrot_training - Step 7903: {'lr': 0.00048157260379627154, 'samples': 1517568, 'steps': 7903, 'loss/train': 2.559544086456299} 01/28/2022 22:58:37 - INFO - codeparrot_training - Step 7904: {'lr': 0.0004815664377547667, 'samples': 1517760, 'steps': 7904, 'loss/train': 1.871995210647583} 01/28/2022 22:58:41 - INFO - codeparrot_training - Step 7905: {'lr': 0.0004815602707213047, 'samples': 1517952, 'steps': 7905, 'loss/train': 2.2661545872688293} 01/28/2022 22:58:46 - INFO - codeparrot_training - Step 7906: {'lr': 0.00048155410269591203, 'samples': 1518144, 'steps': 7906, 'loss/train': 1.1900784373283386} 01/28/2022 22:58:51 - INFO - codeparrot_training - Step 7907: {'lr': 0.00048154793367861514, 'samples': 1518336, 'steps': 7907, 'loss/train': 1.7525015473365784} 01/28/2022 22:58:55 - INFO - codeparrot_training - Step 7908: {'lr': 0.00048154176366944045, 'samples': 1518528, 'steps': 7908, 'loss/train': 1.3627042472362518} 01/28/2022 22:58:59 - INFO - codeparrot_training - Step 7909: {'lr': 0.0004815355926684144, 'samples': 1518720, 'steps': 7909, 'loss/train': 1.6340217590332031} 01/28/2022 22:59:03 - INFO - codeparrot_training - Step 7910: {'lr': 0.0004815294206755633, 'samples': 1518912, 'steps': 7910, 'loss/train': 0.6454538851976395} 01/28/2022 22:59:10 - INFO - codeparrot_training - Step 7911: {'lr': 0.0004815232476909137, 'samples': 1519104, 'steps': 7911, 'loss/train': 1.5570657849311829} 01/28/2022 22:59:14 - INFO - codeparrot_training - Step 7912: {'lr': 0.00048151707371449213, 'samples': 1519296, 'steps': 7912, 'loss/train': 2.442830801010132} 01/28/2022 22:59:18 - INFO - codeparrot_training - Step 7913: {'lr': 0.0004815108987463248, 'samples': 1519488, 'steps': 7913, 'loss/train': 1.2608304619789124} 01/28/2022 22:59:22 - INFO - codeparrot_training - Step 7914: {'lr': 0.00048150472278643834, 'samples': 1519680, 'steps': 7914, 'loss/train': 1.4535412788391113} 01/28/2022 22:59:28 - INFO - codeparrot_training - Step 7915: {'lr': 0.0004814985458348592, 'samples': 1519872, 'steps': 7915, 'loss/train': 0.38627882301807404} 01/28/2022 22:59:32 - INFO - codeparrot_training - Step 7916: {'lr': 0.00048149236789161374, 'samples': 1520064, 'steps': 7916, 'loss/train': 1.6791406273841858} 01/28/2022 22:59:36 - INFO - codeparrot_training - Step 7917: {'lr': 0.00048148618895672846, 'samples': 1520256, 'steps': 7917, 'loss/train': 1.898451805114746} 01/28/2022 22:59:40 - INFO - codeparrot_training - Step 7918: {'lr': 0.0004814800090302299, 'samples': 1520448, 'steps': 7918, 'loss/train': 1.7325409054756165} 01/28/2022 22:59:45 - INFO - codeparrot_training - Step 7919: {'lr': 0.00048147382811214445, 'samples': 1520640, 'steps': 7919, 'loss/train': 2.0812928080558777} 01/28/2022 22:59:50 - INFO - codeparrot_training - Step 7920: {'lr': 0.0004814676462024987, 'samples': 1520832, 'steps': 7920, 'loss/train': 1.536562979221344} 01/28/2022 22:59:54 - INFO - codeparrot_training - Step 7921: {'lr': 0.000481461463301319, 'samples': 1521024, 'steps': 7921, 'loss/train': 2.6792839765548706} 01/28/2022 22:59:59 - INFO - codeparrot_training - Step 7922: {'lr': 0.00048145527940863186, 'samples': 1521216, 'steps': 7922, 'loss/train': 2.056126892566681} 01/28/2022 23:00:03 - INFO - codeparrot_training - Step 7923: {'lr': 0.00048144909452446384, 'samples': 1521408, 'steps': 7923, 'loss/train': 1.0434521734714508} 01/28/2022 23:00:07 - INFO - codeparrot_training - Step 7924: {'lr': 0.00048144290864884145, 'samples': 1521600, 'steps': 7924, 'loss/train': 0.15391482040286064} 01/28/2022 23:00:13 - INFO - codeparrot_training - Step 7925: {'lr': 0.000481436721781791, 'samples': 1521792, 'steps': 7925, 'loss/train': 2.6682531237602234} 01/28/2022 23:00:18 - INFO - codeparrot_training - Step 7926: {'lr': 0.00048143053392333917, 'samples': 1521984, 'steps': 7926, 'loss/train': 2.0607547760009766} 01/28/2022 23:00:22 - INFO - codeparrot_training - Step 7927: {'lr': 0.00048142434507351245, 'samples': 1522176, 'steps': 7927, 'loss/train': 1.9089224338531494} 01/28/2022 23:00:26 - INFO - codeparrot_training - Step 7928: {'lr': 0.00048141815523233735, 'samples': 1522368, 'steps': 7928, 'loss/train': 1.5212799310684204} 01/28/2022 23:00:30 - INFO - codeparrot_training - Step 7929: {'lr': 0.00048141196439984026, 'samples': 1522560, 'steps': 7929, 'loss/train': 2.277049720287323} 01/28/2022 23:00:36 - INFO - codeparrot_training - Step 7930: {'lr': 0.0004814057725760479, 'samples': 1522752, 'steps': 7930, 'loss/train': 1.4522710740566254} 01/28/2022 23:00:40 - INFO - codeparrot_training - Step 7931: {'lr': 0.0004813995797609866, 'samples': 1522944, 'steps': 7931, 'loss/train': 1.5186420679092407} 01/28/2022 23:00:44 - INFO - codeparrot_training - Step 7932: {'lr': 0.000481393385954683, 'samples': 1523136, 'steps': 7932, 'loss/train': 1.0802622735500336} 01/28/2022 23:00:48 - INFO - codeparrot_training - Step 7933: {'lr': 0.00048138719115716367, 'samples': 1523328, 'steps': 7933, 'loss/train': 2.867285907268524} 01/28/2022 23:00:53 - INFO - codeparrot_training - Step 7934: {'lr': 0.00048138099536845503, 'samples': 1523520, 'steps': 7934, 'loss/train': 2.2081183791160583} 01/28/2022 23:00:58 - INFO - codeparrot_training - Step 7935: {'lr': 0.0004813747985885837, 'samples': 1523712, 'steps': 7935, 'loss/train': 2.450596868991852} 01/28/2022 23:01:02 - INFO - codeparrot_training - Step 7936: {'lr': 0.00048136860081757617, 'samples': 1523904, 'steps': 7936, 'loss/train': 1.6293627619743347} 01/28/2022 23:01:06 - INFO - codeparrot_training - Step 7937: {'lr': 0.00048136240205545907, 'samples': 1524096, 'steps': 7937, 'loss/train': 1.9373449087142944} 01/28/2022 23:01:11 - INFO - codeparrot_training - Step 7938: {'lr': 0.0004813562023022588, 'samples': 1524288, 'steps': 7938, 'loss/train': 1.3865605294704437} 01/28/2022 23:01:15 - INFO - codeparrot_training - Step 7939: {'lr': 0.00048135000155800217, 'samples': 1524480, 'steps': 7939, 'loss/train': 1.5539424419403076} 01/28/2022 23:01:22 - INFO - codeparrot_training - Step 7940: {'lr': 0.0004813437998227155, 'samples': 1524672, 'steps': 7940, 'loss/train': 2.564990222454071} 01/28/2022 23:01:26 - INFO - codeparrot_training - Step 7941: {'lr': 0.00048133759709642556, 'samples': 1524864, 'steps': 7941, 'loss/train': 1.9176549911499023} 01/28/2022 23:01:30 - INFO - codeparrot_training - Step 7942: {'lr': 0.00048133139337915866, 'samples': 1525056, 'steps': 7942, 'loss/train': 1.907092273235321} 01/28/2022 23:01:35 - INFO - codeparrot_training - Step 7943: {'lr': 0.00048132518867094167, 'samples': 1525248, 'steps': 7943, 'loss/train': 1.44759401679039} 01/28/2022 23:01:39 - INFO - codeparrot_training - Step 7944: {'lr': 0.00048131898297180085, 'samples': 1525440, 'steps': 7944, 'loss/train': 2.1818093061447144} 01/28/2022 23:01:44 - INFO - codeparrot_training - Step 7945: {'lr': 0.0004813127762817631, 'samples': 1525632, 'steps': 7945, 'loss/train': 1.2915104627609253} 01/28/2022 23:01:48 - INFO - codeparrot_training - Step 7946: {'lr': 0.00048130656860085485, 'samples': 1525824, 'steps': 7946, 'loss/train': 1.1471115052700043} 01/28/2022 23:01:53 - INFO - codeparrot_training - Step 7947: {'lr': 0.0004813003599291027, 'samples': 1526016, 'steps': 7947, 'loss/train': 1.8688897490501404} 01/28/2022 23:01:57 - INFO - codeparrot_training - Step 7948: {'lr': 0.0004812941502665332, 'samples': 1526208, 'steps': 7948, 'loss/train': 1.6374328136444092} 01/28/2022 23:02:01 - INFO - codeparrot_training - Step 7949: {'lr': 0.0004812879396131731, 'samples': 1526400, 'steps': 7949, 'loss/train': 2.2116880416870117} 01/28/2022 23:02:06 - INFO - codeparrot_training - Step 7950: {'lr': 0.0004812817279690488, 'samples': 1526592, 'steps': 7950, 'loss/train': 2.078616499900818} 01/28/2022 23:02:11 - INFO - codeparrot_training - Step 7951: {'lr': 0.00048127551533418714, 'samples': 1526784, 'steps': 7951, 'loss/train': 2.6341094970703125} 01/28/2022 23:02:15 - INFO - codeparrot_training - Step 7952: {'lr': 0.0004812693017086145, 'samples': 1526976, 'steps': 7952, 'loss/train': 2.317441999912262} 01/28/2022 23:02:19 - INFO - codeparrot_training - Step 7953: {'lr': 0.0004812630870923577, 'samples': 1527168, 'steps': 7953, 'loss/train': 1.8146535158157349} 01/28/2022 23:02:23 - INFO - codeparrot_training - Step 7954: {'lr': 0.00048125687148544316, 'samples': 1527360, 'steps': 7954, 'loss/train': 1.6802253127098083} 01/28/2022 23:02:30 - INFO - codeparrot_training - Step 7955: {'lr': 0.0004812506548878977, 'samples': 1527552, 'steps': 7955, 'loss/train': 2.3571953773498535} 01/28/2022 23:02:34 - INFO - codeparrot_training - Step 7956: {'lr': 0.0004812444372997479, 'samples': 1527744, 'steps': 7956, 'loss/train': 0.39221061766147614} 01/28/2022 23:02:39 - INFO - codeparrot_training - Step 7957: {'lr': 0.00048123821872102023, 'samples': 1527936, 'steps': 7957, 'loss/train': 1.907985806465149} 01/28/2022 23:02:43 - INFO - codeparrot_training - Step 7958: {'lr': 0.00048123199915174153, 'samples': 1528128, 'steps': 7958, 'loss/train': 1.6987498998641968} 01/28/2022 23:02:47 - INFO - codeparrot_training - Step 7959: {'lr': 0.0004812257785919384, 'samples': 1528320, 'steps': 7959, 'loss/train': 0.7608422935009003} 01/28/2022 23:02:52 - INFO - codeparrot_training - Step 7960: {'lr': 0.00048121955704163744, 'samples': 1528512, 'steps': 7960, 'loss/train': 1.7429214119911194} 01/28/2022 23:02:57 - INFO - codeparrot_training - Step 7961: {'lr': 0.00048121333450086524, 'samples': 1528704, 'steps': 7961, 'loss/train': 2.283876121044159} 01/28/2022 23:03:01 - INFO - codeparrot_training - Step 7962: {'lr': 0.00048120711096964866, 'samples': 1528896, 'steps': 7962, 'loss/train': 0.6499162763357162} 01/28/2022 23:03:05 - INFO - codeparrot_training - Step 7963: {'lr': 0.0004812008864480142, 'samples': 1529088, 'steps': 7963, 'loss/train': 1.5825753808021545} 01/28/2022 23:03:09 - INFO - codeparrot_training - Step 7964: {'lr': 0.0004811946609359885, 'samples': 1529280, 'steps': 7964, 'loss/train': 1.3800552785396576} 01/28/2022 23:03:15 - INFO - codeparrot_training - Step 7965: {'lr': 0.00048118843443359827, 'samples': 1529472, 'steps': 7965, 'loss/train': 0.940134733915329} 01/28/2022 23:03:19 - INFO - codeparrot_training - Step 7966: {'lr': 0.00048118220694087023, 'samples': 1529664, 'steps': 7966, 'loss/train': 1.7043684124946594} 01/28/2022 23:03:24 - INFO - codeparrot_training - Step 7967: {'lr': 0.00048117597845783106, 'samples': 1529856, 'steps': 7967, 'loss/train': 2.159859538078308} 01/28/2022 23:03:28 - INFO - codeparrot_training - Step 7968: {'lr': 0.0004811697489845074, 'samples': 1530048, 'steps': 7968, 'loss/train': 1.598741054534912} 01/28/2022 23:03:32 - INFO - codeparrot_training - Step 7969: {'lr': 0.0004811635185209259, 'samples': 1530240, 'steps': 7969, 'loss/train': 1.594975233078003} 01/28/2022 23:03:36 - INFO - codeparrot_training - Step 7970: {'lr': 0.0004811572870671133, 'samples': 1530432, 'steps': 7970, 'loss/train': 1.813740313053131} 01/28/2022 23:03:43 - INFO - codeparrot_training - Step 7971: {'lr': 0.0004811510546230963, 'samples': 1530624, 'steps': 7971, 'loss/train': 2.2452873587608337} 01/28/2022 23:03:47 - INFO - codeparrot_training - Step 7972: {'lr': 0.0004811448211889016, 'samples': 1530816, 'steps': 7972, 'loss/train': 1.862133800983429} 01/28/2022 23:03:52 - INFO - codeparrot_training - Step 7973: {'lr': 0.0004811385867645558, 'samples': 1531008, 'steps': 7973, 'loss/train': 1.681159257888794} 01/28/2022 23:03:56 - INFO - codeparrot_training - Step 7974: {'lr': 0.00048113235135008574, 'samples': 1531200, 'steps': 7974, 'loss/train': 1.1467814147472382} 01/28/2022 23:04:02 - INFO - codeparrot_training - Step 7975: {'lr': 0.0004811261149455181, 'samples': 1531392, 'steps': 7975, 'loss/train': 1.9608795046806335} 01/28/2022 23:04:06 - INFO - codeparrot_training - Step 7976: {'lr': 0.0004811198775508796, 'samples': 1531584, 'steps': 7976, 'loss/train': 1.3989355266094208} 01/28/2022 23:04:10 - INFO - codeparrot_training - Step 7977: {'lr': 0.0004811136391661969, 'samples': 1531776, 'steps': 7977, 'loss/train': 0.587819516658783} 01/28/2022 23:04:14 - INFO - codeparrot_training - Step 7978: {'lr': 0.0004811073997914967, 'samples': 1531968, 'steps': 7978, 'loss/train': 1.2841217815876007} 01/28/2022 23:04:19 - INFO - codeparrot_training - Step 7979: {'lr': 0.00048110115942680585, 'samples': 1532160, 'steps': 7979, 'loss/train': 2.4310653805732727} 01/28/2022 23:04:25 - INFO - codeparrot_training - Step 7980: {'lr': 0.000481094918072151, 'samples': 1532352, 'steps': 7980, 'loss/train': 1.1511639654636383} 01/28/2022 23:04:29 - INFO - codeparrot_training - Step 7981: {'lr': 0.0004810886757275589, 'samples': 1532544, 'steps': 7981, 'loss/train': 2.163039207458496} 01/28/2022 23:04:34 - INFO - codeparrot_training - Step 7982: {'lr': 0.0004810824323930563, 'samples': 1532736, 'steps': 7982, 'loss/train': 2.438766896724701} 01/28/2022 23:04:38 - INFO - codeparrot_training - Step 7983: {'lr': 0.00048107618806866994, 'samples': 1532928, 'steps': 7983, 'loss/train': 1.974665880203247} 01/28/2022 23:04:42 - INFO - codeparrot_training - Step 7984: {'lr': 0.0004810699427544265, 'samples': 1533120, 'steps': 7984, 'loss/train': 2.2091172337532043} 01/28/2022 23:04:47 - INFO - codeparrot_training - Step 7985: {'lr': 0.00048106369645035284, 'samples': 1533312, 'steps': 7985, 'loss/train': 0.8627391457557678} 01/28/2022 23:04:52 - INFO - codeparrot_training - Step 7986: {'lr': 0.0004810574491564757, 'samples': 1533504, 'steps': 7986, 'loss/train': 1.3290626406669617} 01/28/2022 23:04:56 - INFO - codeparrot_training - Step 7987: {'lr': 0.0004810512008728218, 'samples': 1533696, 'steps': 7987, 'loss/train': 1.8721845746040344} 01/28/2022 23:05:00 - INFO - codeparrot_training - Step 7988: {'lr': 0.00048104495159941794, 'samples': 1533888, 'steps': 7988, 'loss/train': 1.4774695336818695} 01/28/2022 23:05:05 - INFO - codeparrot_training - Step 7989: {'lr': 0.00048103870133629084, 'samples': 1534080, 'steps': 7989, 'loss/train': 1.3786727786064148} 01/28/2022 23:05:10 - INFO - codeparrot_training - Step 7990: {'lr': 0.00048103245008346735, 'samples': 1534272, 'steps': 7990, 'loss/train': 1.6376867294311523} 01/28/2022 23:05:14 - INFO - codeparrot_training - Step 7991: {'lr': 0.0004810261978409742, 'samples': 1534464, 'steps': 7991, 'loss/train': 2.520334303379059} 01/28/2022 23:05:18 - INFO - codeparrot_training - Step 7992: {'lr': 0.00048101994460883815, 'samples': 1534656, 'steps': 7992, 'loss/train': 2.815465807914734} 01/28/2022 23:05:23 - INFO - codeparrot_training - Step 7993: {'lr': 0.00048101369038708596, 'samples': 1534848, 'steps': 7993, 'loss/train': 1.0246513187885284} 01/28/2022 23:05:27 - INFO - codeparrot_training - Step 7994: {'lr': 0.0004810074351757446, 'samples': 1535040, 'steps': 7994, 'loss/train': 1.4288785457611084} 01/28/2022 23:05:33 - INFO - codeparrot_training - Step 7995: {'lr': 0.00048100117897484064, 'samples': 1535232, 'steps': 7995, 'loss/train': 2.317704498767853} 01/28/2022 23:05:38 - INFO - codeparrot_training - Step 7996: {'lr': 0.0004809949217844011, 'samples': 1535424, 'steps': 7996, 'loss/train': 2.752677798271179} 01/28/2022 23:05:42 - INFO - codeparrot_training - Step 7997: {'lr': 0.00048098866360445254, 'samples': 1535616, 'steps': 7997, 'loss/train': 2.232500910758972} 01/28/2022 23:05:46 - INFO - codeparrot_training - Step 7998: {'lr': 0.00048098240443502195, 'samples': 1535808, 'steps': 7998, 'loss/train': 2.1213247776031494} 01/28/2022 23:05:50 - INFO - codeparrot_training - Step 7999: {'lr': 0.000480976144276136, 'samples': 1536000, 'steps': 7999, 'loss/train': 1.8033165335655212} 01/28/2022 23:05:50 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/28/2022 23:06:23 - WARNING - huggingface_hub.repository - Several commits (4) will be pushed upstream. 01/28/2022 23:06:23 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/28/2022 23:08:12 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py ece0f24..e19ad09 expert-dust-2 -> expert-dust-2 01/28/2022 23:08:19 - INFO - codeparrot_training - Step 8000: {'lr': 0.0004809698831278217, 'samples': 1536192, 'steps': 8000, 'loss/train': 1.8277968764305115} 01/28/2022 23:08:23 - INFO - codeparrot_training - Step 8001: {'lr': 0.0004809636209901057, 'samples': 1536384, 'steps': 8001, 'loss/train': 2.774828016757965} 01/28/2022 23:08:28 - INFO - codeparrot_training - Step 8002: {'lr': 0.00048095735786301495, 'samples': 1536576, 'steps': 8002, 'loss/train': 1.5831871032714844} 01/28/2022 23:08:32 - INFO - codeparrot_training - Step 8003: {'lr': 0.00048095109374657617, 'samples': 1536768, 'steps': 8003, 'loss/train': 2.1385889053344727} 01/28/2022 23:08:36 - INFO - codeparrot_training - Step 8004: {'lr': 0.00048094482864081625, 'samples': 1536960, 'steps': 8004, 'loss/train': 3.3074666261672974} 01/28/2022 23:08:42 - INFO - codeparrot_training - Step 8005: {'lr': 0.00048093856254576196, 'samples': 1537152, 'steps': 8005, 'loss/train': 1.7264803647994995} 01/28/2022 23:08:46 - INFO - codeparrot_training - Step 8006: {'lr': 0.0004809322954614403, 'samples': 1537344, 'steps': 8006, 'loss/train': 2.0465553402900696} 01/28/2022 23:08:50 - INFO - codeparrot_training - Step 8007: {'lr': 0.00048092602738787795, 'samples': 1537536, 'steps': 8007, 'loss/train': 1.7547606825828552} 01/28/2022 23:08:55 - INFO - codeparrot_training - Step 8008: {'lr': 0.00048091975832510183, 'samples': 1537728, 'steps': 8008, 'loss/train': 1.735833764076233} 01/28/2022 23:08:59 - INFO - codeparrot_training - Step 8009: {'lr': 0.00048091348827313885, 'samples': 1537920, 'steps': 8009, 'loss/train': 1.8520917892456055} 01/28/2022 23:09:04 - INFO - codeparrot_training - Step 8010: {'lr': 0.0004809072172320157, 'samples': 1538112, 'steps': 8010, 'loss/train': 2.0175324082374573} 01/28/2022 23:09:08 - INFO - codeparrot_training - Step 8011: {'lr': 0.0004809009452017594, 'samples': 1538304, 'steps': 8011, 'loss/train': 1.3946013450622559} 01/28/2022 23:09:13 - INFO - codeparrot_training - Step 8012: {'lr': 0.00048089467218239687, 'samples': 1538496, 'steps': 8012, 'loss/train': 2.1314679980278015} 01/28/2022 23:09:17 - INFO - codeparrot_training - Step 8013: {'lr': 0.0004808883981739548, 'samples': 1538688, 'steps': 8013, 'loss/train': 2.114421308040619} 01/28/2022 23:09:21 - INFO - codeparrot_training - Step 8014: {'lr': 0.00048088212317646016, 'samples': 1538880, 'steps': 8014, 'loss/train': 1.942182719707489} 01/28/2022 23:09:28 - INFO - codeparrot_training - Step 8015: {'lr': 0.00048087584718993975, 'samples': 1539072, 'steps': 8015, 'loss/train': 1.621875822544098} 01/28/2022 23:09:32 - INFO - codeparrot_training - Step 8016: {'lr': 0.0004808695702144206, 'samples': 1539264, 'steps': 8016, 'loss/train': 1.975276529788971} 01/28/2022 23:09:36 - INFO - codeparrot_training - Step 8017: {'lr': 0.0004808632922499295, 'samples': 1539456, 'steps': 8017, 'loss/train': 2.08492112159729} 01/28/2022 23:09:40 - INFO - codeparrot_training - Step 8018: {'lr': 0.00048085701329649336, 'samples': 1539648, 'steps': 8018, 'loss/train': 2.107542335987091} 01/28/2022 23:09:45 - INFO - codeparrot_training - Step 8019: {'lr': 0.0004808507333541391, 'samples': 1539840, 'steps': 8019, 'loss/train': 2.0734965205192566} 01/28/2022 23:09:50 - INFO - codeparrot_training - Step 8020: {'lr': 0.00048084445242289355, 'samples': 1540032, 'steps': 8020, 'loss/train': 1.571608543395996} 01/28/2022 23:09:54 - INFO - codeparrot_training - Step 8021: {'lr': 0.0004808381705027837, 'samples': 1540224, 'steps': 8021, 'loss/train': 1.7808068990707397} 01/28/2022 23:09:59 - INFO - codeparrot_training - Step 8022: {'lr': 0.00048083188759383646, 'samples': 1540416, 'steps': 8022, 'loss/train': 1.0384466350078583} 01/28/2022 23:10:03 - INFO - codeparrot_training - Step 8023: {'lr': 0.00048082560369607863, 'samples': 1540608, 'steps': 8023, 'loss/train': 1.6748043894767761} 01/28/2022 23:10:07 - INFO - codeparrot_training - Step 8024: {'lr': 0.0004808193188095372, 'samples': 1540800, 'steps': 8024, 'loss/train': 2.2584367990493774} 01/28/2022 23:10:13 - INFO - codeparrot_training - Step 8025: {'lr': 0.00048081303293423923, 'samples': 1540992, 'steps': 8025, 'loss/train': 2.296036183834076} 01/28/2022 23:10:17 - INFO - codeparrot_training - Step 8026: {'lr': 0.0004808067460702115, 'samples': 1541184, 'steps': 8026, 'loss/train': 1.8155069947242737} 01/28/2022 23:10:21 - INFO - codeparrot_training - Step 8027: {'lr': 0.00048080045821748086, 'samples': 1541376, 'steps': 8027, 'loss/train': 1.997374713420868} 01/28/2022 23:10:25 - INFO - codeparrot_training - Step 8028: {'lr': 0.00048079416937607436, 'samples': 1541568, 'steps': 8028, 'loss/train': 1.490767389535904} 01/28/2022 23:10:30 - INFO - codeparrot_training - Step 8029: {'lr': 0.000480787879546019, 'samples': 1541760, 'steps': 8029, 'loss/train': 1.7440422177314758} 01/28/2022 23:10:36 - INFO - codeparrot_training - Step 8030: {'lr': 0.00048078158872734157, 'samples': 1541952, 'steps': 8030, 'loss/train': 2.6793887615203857} 01/28/2022 23:10:40 - INFO - codeparrot_training - Step 8031: {'lr': 0.0004807752969200691, 'samples': 1542144, 'steps': 8031, 'loss/train': 2.0582016706466675} 01/28/2022 23:10:45 - INFO - codeparrot_training - Step 8032: {'lr': 0.0004807690041242286, 'samples': 1542336, 'steps': 8032, 'loss/train': 1.171893149614334} 01/28/2022 23:10:49 - INFO - codeparrot_training - Step 8033: {'lr': 0.00048076271033984687, 'samples': 1542528, 'steps': 8033, 'loss/train': 1.6981517672538757} 01/28/2022 23:10:53 - INFO - codeparrot_training - Step 8034: {'lr': 0.00048075641556695107, 'samples': 1542720, 'steps': 8034, 'loss/train': 1.4557982683181763} 01/28/2022 23:10:58 - INFO - codeparrot_training - Step 8035: {'lr': 0.000480750119805568, 'samples': 1542912, 'steps': 8035, 'loss/train': 2.619280993938446} 01/28/2022 23:11:03 - INFO - codeparrot_training - Step 8036: {'lr': 0.0004807438230557247, 'samples': 1543104, 'steps': 8036, 'loss/train': 2.1495476961135864} 01/28/2022 23:11:07 - INFO - codeparrot_training - Step 8037: {'lr': 0.00048073752531744814, 'samples': 1543296, 'steps': 8037, 'loss/train': 1.2981513440608978} 01/28/2022 23:11:11 - INFO - codeparrot_training - Step 8038: {'lr': 0.0004807312265907653, 'samples': 1543488, 'steps': 8038, 'loss/train': 2.4210877418518066} 01/28/2022 23:11:15 - INFO - codeparrot_training - Step 8039: {'lr': 0.0004807249268757031, 'samples': 1543680, 'steps': 8039, 'loss/train': 1.4762594103813171} 01/28/2022 23:11:22 - INFO - codeparrot_training - Step 8040: {'lr': 0.00048071862617228854, 'samples': 1543872, 'steps': 8040, 'loss/train': 1.1597137749195099} 01/28/2022 23:11:26 - INFO - codeparrot_training - Step 8041: {'lr': 0.0004807123244805488, 'samples': 1544064, 'steps': 8041, 'loss/train': 2.1101328134536743} 01/28/2022 23:11:30 - INFO - codeparrot_training - Step 8042: {'lr': 0.0004807060218005106, 'samples': 1544256, 'steps': 8042, 'loss/train': 1.8825836777687073} 01/28/2022 23:11:34 - INFO - codeparrot_training - Step 8043: {'lr': 0.00048069971813220107, 'samples': 1544448, 'steps': 8043, 'loss/train': 1.4167212545871735} 01/28/2022 23:11:39 - INFO - codeparrot_training - Step 8044: {'lr': 0.0004806934134756472, 'samples': 1544640, 'steps': 8044, 'loss/train': 0.9640754163265228} 01/28/2022 23:11:44 - INFO - codeparrot_training - Step 8045: {'lr': 0.0004806871078308761, 'samples': 1544832, 'steps': 8045, 'loss/train': 2.4353479743003845} 01/28/2022 23:11:48 - INFO - codeparrot_training - Step 8046: {'lr': 0.0004806808011979146, 'samples': 1545024, 'steps': 8046, 'loss/train': 1.8154107928276062} 01/28/2022 23:11:52 - INFO - codeparrot_training - Step 8047: {'lr': 0.00048067449357678984, 'samples': 1545216, 'steps': 8047, 'loss/train': 2.274843692779541} 01/28/2022 23:11:57 - INFO - codeparrot_training - Step 8048: {'lr': 0.0004806681849675287, 'samples': 1545408, 'steps': 8048, 'loss/train': 1.9456116557121277} 01/28/2022 23:12:01 - INFO - codeparrot_training - Step 8049: {'lr': 0.00048066187537015837, 'samples': 1545600, 'steps': 8049, 'loss/train': 1.8161961436271667} 01/28/2022 23:12:06 - INFO - codeparrot_training - Step 8050: {'lr': 0.00048065556478470584, 'samples': 1545792, 'steps': 8050, 'loss/train': 2.343078911304474} 01/28/2022 23:12:10 - INFO - codeparrot_training - Step 8051: {'lr': 0.0004806492532111981, 'samples': 1545984, 'steps': 8051, 'loss/train': 1.757231891155243} 01/28/2022 23:12:15 - INFO - codeparrot_training - Step 8052: {'lr': 0.00048064294064966215, 'samples': 1546176, 'steps': 8052, 'loss/train': 1.690420389175415} 01/28/2022 23:12:19 - INFO - codeparrot_training - Step 8053: {'lr': 0.00048063662710012513, 'samples': 1546368, 'steps': 8053, 'loss/train': 1.6178237199783325} 01/28/2022 23:12:23 - INFO - codeparrot_training - Step 8054: {'lr': 0.000480630312562614, 'samples': 1546560, 'steps': 8054, 'loss/train': 1.0615192651748657} 01/28/2022 23:12:29 - INFO - codeparrot_training - Step 8055: {'lr': 0.0004806239970371558, 'samples': 1546752, 'steps': 8055, 'loss/train': 1.7742383480072021} 01/28/2022 23:12:33 - INFO - codeparrot_training - Step 8056: {'lr': 0.0004806176805237777, 'samples': 1546944, 'steps': 8056, 'loss/train': 1.7321115732192993} 01/28/2022 23:12:37 - INFO - codeparrot_training - Step 8057: {'lr': 0.0004806113630225066, 'samples': 1547136, 'steps': 8057, 'loss/train': 0.6597488969564438} 01/28/2022 23:12:41 - INFO - codeparrot_training - Step 8058: {'lr': 0.0004806050445333697, 'samples': 1547328, 'steps': 8058, 'loss/train': 2.090686798095703} 01/28/2022 23:12:46 - INFO - codeparrot_training - Step 8059: {'lr': 0.00048059872505639415, 'samples': 1547520, 'steps': 8059, 'loss/train': 1.8213179111480713} 01/28/2022 23:12:52 - INFO - codeparrot_training - Step 8060: {'lr': 0.0004805924045916067, 'samples': 1547712, 'steps': 8060, 'loss/train': 1.7285757064819336} 01/28/2022 23:12:56 - INFO - codeparrot_training - Step 8061: {'lr': 0.00048058608313903474, 'samples': 1547904, 'steps': 8061, 'loss/train': 1.8578183054924011} 01/28/2022 23:13:01 - INFO - codeparrot_training - Step 8062: {'lr': 0.0004805797606987051, 'samples': 1548096, 'steps': 8062, 'loss/train': 2.103694438934326} 01/28/2022 23:13:05 - INFO - codeparrot_training - Step 8063: {'lr': 0.0004805734372706451, 'samples': 1548288, 'steps': 8063, 'loss/train': 1.830561339855194} 01/28/2022 23:13:09 - INFO - codeparrot_training - Step 8064: {'lr': 0.0004805671128548816, 'samples': 1548480, 'steps': 8064, 'loss/train': 1.5916099548339844} 01/28/2022 23:13:14 - INFO - codeparrot_training - Step 8065: {'lr': 0.00048056078745144183, 'samples': 1548672, 'steps': 8065, 'loss/train': 1.258909910917282} 01/28/2022 23:13:19 - INFO - codeparrot_training - Step 8066: {'lr': 0.0004805544610603529, 'samples': 1548864, 'steps': 8066, 'loss/train': 1.854101836681366} 01/28/2022 23:13:23 - INFO - codeparrot_training - Step 8067: {'lr': 0.00048054813368164184, 'samples': 1549056, 'steps': 8067, 'loss/train': 2.2869930267333984} 01/28/2022 23:13:27 - INFO - codeparrot_training - Step 8068: {'lr': 0.00048054180531533576, 'samples': 1549248, 'steps': 8068, 'loss/train': 0.9893284142017365} 01/28/2022 23:13:31 - INFO - codeparrot_training - Step 8069: {'lr': 0.00048053547596146185, 'samples': 1549440, 'steps': 8069, 'loss/train': 2.2931222319602966} 01/28/2022 23:13:36 - INFO - codeparrot_training - Step 8070: {'lr': 0.0004805291456200471, 'samples': 1549632, 'steps': 8070, 'loss/train': 2.415253758430481} 01/28/2022 23:13:41 - INFO - codeparrot_training - Step 8071: {'lr': 0.0004805228142911188, 'samples': 1549824, 'steps': 8071, 'loss/train': 1.4850337207317352} 01/28/2022 23:13:45 - INFO - codeparrot_training - Step 8072: {'lr': 0.0004805164819747038, 'samples': 1550016, 'steps': 8072, 'loss/train': 2.003046691417694} 01/28/2022 23:13:49 - INFO - codeparrot_training - Step 8073: {'lr': 0.0004805101486708295, 'samples': 1550208, 'steps': 8073, 'loss/train': 1.1098718047142029} 01/28/2022 23:13:53 - INFO - codeparrot_training - Step 8074: {'lr': 0.0004805038143795229, 'samples': 1550400, 'steps': 8074, 'loss/train': 1.7936385869979858} 01/28/2022 23:14:00 - INFO - codeparrot_training - Step 8075: {'lr': 0.00048049747910081114, 'samples': 1550592, 'steps': 8075, 'loss/train': 1.8391649723052979} 01/28/2022 23:14:04 - INFO - codeparrot_training - Step 8076: {'lr': 0.0004804911428347214, 'samples': 1550784, 'steps': 8076, 'loss/train': 1.648226022720337} 01/28/2022 23:14:08 - INFO - codeparrot_training - Step 8077: {'lr': 0.0004804848055812807, 'samples': 1550976, 'steps': 8077, 'loss/train': 1.6887609958648682} 01/28/2022 23:14:13 - INFO - codeparrot_training - Step 8078: {'lr': 0.0004804784673405164, 'samples': 1551168, 'steps': 8078, 'loss/train': 1.1627211570739746} 01/28/2022 23:14:17 - INFO - codeparrot_training - Step 8079: {'lr': 0.00048047212811245545, 'samples': 1551360, 'steps': 8079, 'loss/train': 2.0902440547943115} 01/28/2022 23:14:22 - INFO - codeparrot_training - Step 8080: {'lr': 0.00048046578789712516, 'samples': 1551552, 'steps': 8080, 'loss/train': 1.3872016668319702} 01/28/2022 23:14:27 - INFO - codeparrot_training - Step 8081: {'lr': 0.0004804594466945525, 'samples': 1551744, 'steps': 8081, 'loss/train': 2.0500059127807617} 01/28/2022 23:14:31 - INFO - codeparrot_training - Step 8082: {'lr': 0.00048045310450476486, 'samples': 1551936, 'steps': 8082, 'loss/train': 1.2834526598453522} 01/28/2022 23:14:35 - INFO - codeparrot_training - Step 8083: {'lr': 0.0004804467613277893, 'samples': 1552128, 'steps': 8083, 'loss/train': 2.0212342143058777} 01/28/2022 23:14:41 - INFO - codeparrot_training - Step 8084: {'lr': 0.00048044041716365296, 'samples': 1552320, 'steps': 8084, 'loss/train': 2.118532955646515} 01/28/2022 23:14:46 - INFO - codeparrot_training - Step 8085: {'lr': 0.000480434072012383, 'samples': 1552512, 'steps': 8085, 'loss/train': 1.304971307516098} 01/28/2022 23:14:50 - INFO - codeparrot_training - Step 8086: {'lr': 0.0004804277258740067, 'samples': 1552704, 'steps': 8086, 'loss/train': 2.3580604791641235} 01/28/2022 23:14:54 - INFO - codeparrot_training - Step 8087: {'lr': 0.0004804213787485512, 'samples': 1552896, 'steps': 8087, 'loss/train': 2.2235047817230225} 01/28/2022 23:14:58 - INFO - codeparrot_training - Step 8088: {'lr': 0.00048041503063604366, 'samples': 1553088, 'steps': 8088, 'loss/train': 1.9019529819488525} 01/28/2022 23:15:03 - INFO - codeparrot_training - Step 8089: {'lr': 0.00048040868153651124, 'samples': 1553280, 'steps': 8089, 'loss/train': 0.6999814957380295} 01/28/2022 23:15:08 - INFO - codeparrot_training - Step 8090: {'lr': 0.00048040233144998123, 'samples': 1553472, 'steps': 8090, 'loss/train': 1.5442585945129395} 01/28/2022 23:15:13 - INFO - codeparrot_training - Step 8091: {'lr': 0.0004803959803764808, 'samples': 1553664, 'steps': 8091, 'loss/train': 1.5790055394172668} 01/28/2022 23:15:17 - INFO - codeparrot_training - Step 8092: {'lr': 0.0004803896283160372, 'samples': 1553856, 'steps': 8092, 'loss/train': 2.1142221093177795} 01/28/2022 23:15:21 - INFO - codeparrot_training - Step 8093: {'lr': 0.0004803832752686775, 'samples': 1554048, 'steps': 8093, 'loss/train': 2.2325525879859924} 01/28/2022 23:15:26 - INFO - codeparrot_training - Step 8094: {'lr': 0.00048037692123442904, 'samples': 1554240, 'steps': 8094, 'loss/train': 1.458351731300354} 01/28/2022 23:15:31 - INFO - codeparrot_training - Step 8095: {'lr': 0.000480370566213319, 'samples': 1554432, 'steps': 8095, 'loss/train': 1.5504661202430725} 01/28/2022 23:15:35 - INFO - codeparrot_training - Step 8096: {'lr': 0.00048036421020537464, 'samples': 1554624, 'steps': 8096, 'loss/train': 1.4189144968986511} 01/28/2022 23:15:39 - INFO - codeparrot_training - Step 8097: {'lr': 0.0004803578532106231, 'samples': 1554816, 'steps': 8097, 'loss/train': 2.1931899189949036} 01/28/2022 23:15:43 - INFO - codeparrot_training - Step 8098: {'lr': 0.00048035149522909174, 'samples': 1555008, 'steps': 8098, 'loss/train': 2.285154640674591} 01/28/2022 23:15:50 - INFO - codeparrot_training - Step 8099: {'lr': 0.0004803451362608076, 'samples': 1555200, 'steps': 8099, 'loss/train': 1.245823323726654} 01/28/2022 23:15:54 - INFO - codeparrot_training - Step 8100: {'lr': 0.00048033877630579815, 'samples': 1555392, 'steps': 8100, 'loss/train': 1.1915404200553894} 01/28/2022 23:15:58 - INFO - codeparrot_training - Step 8101: {'lr': 0.00048033241536409043, 'samples': 1555584, 'steps': 8101, 'loss/train': 2.36740243434906} 01/28/2022 23:16:02 - INFO - codeparrot_training - Step 8102: {'lr': 0.0004803260534357119, 'samples': 1555776, 'steps': 8102, 'loss/train': 2.0894984006881714} 01/28/2022 23:16:07 - INFO - codeparrot_training - Step 8103: {'lr': 0.00048031969052068956, 'samples': 1555968, 'steps': 8103, 'loss/train': 2.6643092036247253} 01/28/2022 23:16:11 - INFO - codeparrot_training - Step 8104: {'lr': 0.00048031332661905093, 'samples': 1556160, 'steps': 8104, 'loss/train': 1.8813759684562683} 01/28/2022 23:16:17 - INFO - codeparrot_training - Step 8105: {'lr': 0.000480306961730823, 'samples': 1556352, 'steps': 8105, 'loss/train': 7.872609615325928} 01/28/2022 23:16:21 - INFO - codeparrot_training - Step 8106: {'lr': 0.00048030059585603326, 'samples': 1556544, 'steps': 8106, 'loss/train': 0.4036218076944351} 01/28/2022 23:16:25 - INFO - codeparrot_training - Step 8107: {'lr': 0.0004802942289947089, 'samples': 1556736, 'steps': 8107, 'loss/train': 1.4369768500328064} 01/28/2022 23:16:30 - INFO - codeparrot_training - Step 8108: {'lr': 0.00048028786114687715, 'samples': 1556928, 'steps': 8108, 'loss/train': 1.9205294251441956} 01/28/2022 23:16:34 - INFO - codeparrot_training - Step 8109: {'lr': 0.0004802814923125654, 'samples': 1557120, 'steps': 8109, 'loss/train': 1.0200315713882446} 01/28/2022 23:16:40 - INFO - codeparrot_training - Step 8110: {'lr': 0.00048027512249180083, 'samples': 1557312, 'steps': 8110, 'loss/train': 2.1545923948287964} 01/28/2022 23:16:44 - INFO - codeparrot_training - Step 8111: {'lr': 0.0004802687516846107, 'samples': 1557504, 'steps': 8111, 'loss/train': 2.0268505811691284} 01/28/2022 23:16:48 - INFO - codeparrot_training - Step 8112: {'lr': 0.0004802623798910224, 'samples': 1557696, 'steps': 8112, 'loss/train': 1.0930474698543549} 01/28/2022 23:16:52 - INFO - codeparrot_training - Step 8113: {'lr': 0.00048025600711106323, 'samples': 1557888, 'steps': 8113, 'loss/train': 1.2581970691680908} 01/28/2022 23:16:57 - INFO - codeparrot_training - Step 8114: {'lr': 0.00048024963334476035, 'samples': 1558080, 'steps': 8114, 'loss/train': 2.2866722345352173} 01/28/2022 23:17:02 - INFO - codeparrot_training - Step 8115: {'lr': 0.00048024325859214123, 'samples': 1558272, 'steps': 8115, 'loss/train': 2.2666200399398804} 01/28/2022 23:17:06 - INFO - codeparrot_training - Step 8116: {'lr': 0.00048023688285323305, 'samples': 1558464, 'steps': 8116, 'loss/train': 2.327328622341156} 01/28/2022 23:17:10 - INFO - codeparrot_training - Step 8117: {'lr': 0.0004802305061280632, 'samples': 1558656, 'steps': 8117, 'loss/train': 1.126854658126831} 01/28/2022 23:17:15 - INFO - codeparrot_training - Step 8118: {'lr': 0.0004802241284166589, 'samples': 1558848, 'steps': 8118, 'loss/train': 1.7692574858665466} 01/28/2022 23:17:19 - INFO - codeparrot_training - Step 8119: {'lr': 0.00048021774971904765, 'samples': 1559040, 'steps': 8119, 'loss/train': 2.219208776950836} 01/28/2022 23:17:26 - INFO - codeparrot_training - Step 8120: {'lr': 0.0004802113700352566, 'samples': 1559232, 'steps': 8120, 'loss/train': 1.6895066499710083} 01/28/2022 23:17:31 - INFO - codeparrot_training - Step 8121: {'lr': 0.0004802049893653131, 'samples': 1559424, 'steps': 8121, 'loss/train': 1.4856242537498474} 01/28/2022 23:17:35 - INFO - codeparrot_training - Step 8122: {'lr': 0.0004801986077092446, 'samples': 1559616, 'steps': 8122, 'loss/train': 1.1710546016693115} 01/28/2022 23:17:39 - INFO - codeparrot_training - Step 8123: {'lr': 0.0004801922250670783, 'samples': 1559808, 'steps': 8123, 'loss/train': 2.1056408286094666} 01/28/2022 23:17:44 - INFO - codeparrot_training - Step 8124: {'lr': 0.0004801858414388416, 'samples': 1560000, 'steps': 8124, 'loss/train': 2.4010709524154663} 01/28/2022 23:17:49 - INFO - codeparrot_training - Step 8125: {'lr': 0.0004801794568245619, 'samples': 1560192, 'steps': 8125, 'loss/train': 1.9007611870765686} 01/28/2022 23:17:53 - INFO - codeparrot_training - Step 8126: {'lr': 0.00048017307122426653, 'samples': 1560384, 'steps': 8126, 'loss/train': 2.3092743158340454} 01/28/2022 23:17:57 - INFO - codeparrot_training - Step 8127: {'lr': 0.0004801666846379827, 'samples': 1560576, 'steps': 8127, 'loss/train': 1.8147348761558533} 01/28/2022 23:18:01 - INFO - codeparrot_training - Step 8128: {'lr': 0.00048016029706573793, 'samples': 1560768, 'steps': 8128, 'loss/train': 1.8351911902427673} 01/28/2022 23:18:09 - INFO - codeparrot_training - Step 8129: {'lr': 0.0004801539085075596, 'samples': 1560960, 'steps': 8129, 'loss/train': 1.9623661637306213} 01/28/2022 23:18:13 - INFO - codeparrot_training - Step 8130: {'lr': 0.0004801475189634749, 'samples': 1561152, 'steps': 8130, 'loss/train': 2.319947361946106} 01/28/2022 23:18:17 - INFO - codeparrot_training - Step 8131: {'lr': 0.0004801411284335114, 'samples': 1561344, 'steps': 8131, 'loss/train': 1.452459454536438} 01/28/2022 23:18:21 - INFO - codeparrot_training - Step 8132: {'lr': 0.0004801347369176963, 'samples': 1561536, 'steps': 8132, 'loss/train': 1.2853721380233765} 01/28/2022 23:18:26 - INFO - codeparrot_training - Step 8133: {'lr': 0.0004801283444160571, 'samples': 1561728, 'steps': 8133, 'loss/train': 1.528907060623169} 01/28/2022 23:18:31 - INFO - codeparrot_training - Step 8134: {'lr': 0.0004801219509286212, 'samples': 1561920, 'steps': 8134, 'loss/train': 2.245767295360565} 01/28/2022 23:18:35 - INFO - codeparrot_training - Step 8135: {'lr': 0.00048011555645541585, 'samples': 1562112, 'steps': 8135, 'loss/train': 1.5668628215789795} 01/28/2022 23:18:39 - INFO - codeparrot_training - Step 8136: {'lr': 0.00048010916099646854, 'samples': 1562304, 'steps': 8136, 'loss/train': 0.1726708821952343} 01/28/2022 23:18:43 - INFO - codeparrot_training - Step 8137: {'lr': 0.0004801027645518067, 'samples': 1562496, 'steps': 8137, 'loss/train': 1.4837707579135895} 01/28/2022 23:18:48 - INFO - codeparrot_training - Step 8138: {'lr': 0.00048009636712145764, 'samples': 1562688, 'steps': 8138, 'loss/train': 2.405760884284973} 01/28/2022 23:18:53 - INFO - codeparrot_training - Step 8139: {'lr': 0.00048008996870544887, 'samples': 1562880, 'steps': 8139, 'loss/train': 1.561888039112091} 01/28/2022 23:18:57 - INFO - codeparrot_training - Step 8140: {'lr': 0.0004800835693038076, 'samples': 1563072, 'steps': 8140, 'loss/train': 1.5546528697013855} 01/28/2022 23:19:01 - INFO - codeparrot_training - Step 8141: {'lr': 0.0004800771689165615, 'samples': 1563264, 'steps': 8141, 'loss/train': 1.1061692833900452} 01/28/2022 23:19:06 - INFO - codeparrot_training - Step 8142: {'lr': 0.00048007076754373785, 'samples': 1563456, 'steps': 8142, 'loss/train': 1.9645798802375793} 01/28/2022 23:19:10 - INFO - codeparrot_training - Step 8143: {'lr': 0.00048006436518536403, 'samples': 1563648, 'steps': 8143, 'loss/train': 0.8510580360889435} 01/28/2022 23:19:17 - INFO - codeparrot_training - Step 8144: {'lr': 0.0004800579618414676, 'samples': 1563840, 'steps': 8144, 'loss/train': 1.6023525595664978} 01/28/2022 23:19:21 - INFO - codeparrot_training - Step 8145: {'lr': 0.00048005155751207584, 'samples': 1564032, 'steps': 8145, 'loss/train': 1.7906604409217834} 01/28/2022 23:19:25 - INFO - codeparrot_training - Step 8146: {'lr': 0.0004800451521972163, 'samples': 1564224, 'steps': 8146, 'loss/train': 1.5987153053283691} 01/28/2022 23:19:29 - INFO - codeparrot_training - Step 8147: {'lr': 0.0004800387458969164, 'samples': 1564416, 'steps': 8147, 'loss/train': 2.3394895792007446} 01/28/2022 23:19:34 - INFO - codeparrot_training - Step 8148: {'lr': 0.00048003233861120356, 'samples': 1564608, 'steps': 8148, 'loss/train': 1.7759696245193481} 01/28/2022 23:19:39 - INFO - codeparrot_training - Step 8149: {'lr': 0.00048002593034010516, 'samples': 1564800, 'steps': 8149, 'loss/train': 1.3195476830005646} 01/28/2022 23:19:43 - INFO - codeparrot_training - Step 8150: {'lr': 0.00048001952108364876, 'samples': 1564992, 'steps': 8150, 'loss/train': 1.131368637084961} 01/28/2022 23:19:47 - INFO - codeparrot_training - Step 8151: {'lr': 0.00048001311084186173, 'samples': 1565184, 'steps': 8151, 'loss/train': 2.215786099433899} 01/28/2022 23:19:52 - INFO - codeparrot_training - Step 8152: {'lr': 0.0004800066996147716, 'samples': 1565376, 'steps': 8152, 'loss/train': 0.08640754222869873} 01/28/2022 23:19:56 - INFO - codeparrot_training - Step 8153: {'lr': 0.0004800002874024058, 'samples': 1565568, 'steps': 8153, 'loss/train': 2.034705698490143} 01/28/2022 23:20:02 - INFO - codeparrot_training - Step 8154: {'lr': 0.0004799938742047918, 'samples': 1565760, 'steps': 8154, 'loss/train': 2.569763481616974} 01/28/2022 23:20:07 - INFO - codeparrot_training - Step 8155: {'lr': 0.0004799874600219571, 'samples': 1565952, 'steps': 8155, 'loss/train': 1.1597465872764587} 01/28/2022 23:20:11 - INFO - codeparrot_training - Step 8156: {'lr': 0.00047998104485392915, 'samples': 1566144, 'steps': 8156, 'loss/train': 0.09717515856027603} 01/28/2022 23:20:15 - INFO - codeparrot_training - Step 8157: {'lr': 0.0004799746287007354, 'samples': 1566336, 'steps': 8157, 'loss/train': 1.1358219981193542} 01/28/2022 23:20:19 - INFO - codeparrot_training - Step 8158: {'lr': 0.00047996821156240333, 'samples': 1566528, 'steps': 8158, 'loss/train': 2.7654497623443604} 01/28/2022 23:20:25 - INFO - codeparrot_training - Step 8159: {'lr': 0.0004799617934389605, 'samples': 1566720, 'steps': 8159, 'loss/train': 2.059463918209076} 01/28/2022 23:20:29 - INFO - codeparrot_training - Step 8160: {'lr': 0.00047995537433043444, 'samples': 1566912, 'steps': 8160, 'loss/train': 2.604654014110565} 01/28/2022 23:20:33 - INFO - codeparrot_training - Step 8161: {'lr': 0.00047994895423685246, 'samples': 1567104, 'steps': 8161, 'loss/train': 0.9545430243015289} 01/28/2022 23:20:37 - INFO - codeparrot_training - Step 8162: {'lr': 0.0004799425331582423, 'samples': 1567296, 'steps': 8162, 'loss/train': 1.6036280393600464} 01/28/2022 23:20:42 - INFO - codeparrot_training - Step 8163: {'lr': 0.00047993611109463125, 'samples': 1567488, 'steps': 8163, 'loss/train': 2.3259699940681458} 01/28/2022 23:20:47 - INFO - codeparrot_training - Step 8164: {'lr': 0.00047992968804604693, 'samples': 1567680, 'steps': 8164, 'loss/train': 0.24094589799642563} 01/28/2022 23:20:51 - INFO - codeparrot_training - Step 8165: {'lr': 0.00047992326401251686, 'samples': 1567872, 'steps': 8165, 'loss/train': 1.9043508768081665} 01/28/2022 23:20:55 - INFO - codeparrot_training - Step 8166: {'lr': 0.0004799168389940685, 'samples': 1568064, 'steps': 8166, 'loss/train': 0.8821186423301697} 01/28/2022 23:21:00 - INFO - codeparrot_training - Step 8167: {'lr': 0.00047991041299072946, 'samples': 1568256, 'steps': 8167, 'loss/train': 2.9777864813804626} 01/28/2022 23:21:04 - INFO - codeparrot_training - Step 8168: {'lr': 0.00047990398600252713, 'samples': 1568448, 'steps': 8168, 'loss/train': 1.4612028300762177} 01/28/2022 23:21:09 - INFO - codeparrot_training - Step 8169: {'lr': 0.0004798975580294892, 'samples': 1568640, 'steps': 8169, 'loss/train': 1.5959027409553528} 01/28/2022 23:21:14 - INFO - codeparrot_training - Step 8170: {'lr': 0.0004798911290716431, 'samples': 1568832, 'steps': 8170, 'loss/train': 2.3044331073760986} 01/28/2022 23:21:18 - INFO - codeparrot_training - Step 8171: {'lr': 0.0004798846991290164, 'samples': 1569024, 'steps': 8171, 'loss/train': 0.08001427352428436} 01/28/2022 23:21:22 - INFO - codeparrot_training - Step 8172: {'lr': 0.0004798782682016367, 'samples': 1569216, 'steps': 8172, 'loss/train': 1.377126932144165} 01/28/2022 23:21:26 - INFO - codeparrot_training - Step 8173: {'lr': 0.0004798718362895315, 'samples': 1569408, 'steps': 8173, 'loss/train': 1.7058232426643372} 01/28/2022 23:21:33 - INFO - codeparrot_training - Step 8174: {'lr': 0.0004798654033927283, 'samples': 1569600, 'steps': 8174, 'loss/train': 1.1788830757141113} 01/28/2022 23:21:37 - INFO - codeparrot_training - Step 8175: {'lr': 0.00047985896951125464, 'samples': 1569792, 'steps': 8175, 'loss/train': 2.426045536994934} 01/28/2022 23:21:41 - INFO - codeparrot_training - Step 8176: {'lr': 0.00047985253464513823, 'samples': 1569984, 'steps': 8176, 'loss/train': 1.8254368901252747} 01/28/2022 23:21:45 - INFO - codeparrot_training - Step 8177: {'lr': 0.00047984609879440655, 'samples': 1570176, 'steps': 8177, 'loss/train': 1.555117428302765} 01/28/2022 23:21:50 - INFO - codeparrot_training - Step 8178: {'lr': 0.0004798396619590871, 'samples': 1570368, 'steps': 8178, 'loss/train': 1.9643216729164124} 01/28/2022 23:21:55 - INFO - codeparrot_training - Step 8179: {'lr': 0.0004798332241392076, 'samples': 1570560, 'steps': 8179, 'loss/train': 2.2876607179641724} 01/28/2022 23:21:59 - INFO - codeparrot_training - Step 8180: {'lr': 0.0004798267853347955, 'samples': 1570752, 'steps': 8180, 'loss/train': 2.6795631051063538} 01/28/2022 23:22:03 - INFO - codeparrot_training - Step 8181: {'lr': 0.00047982034554587837, 'samples': 1570944, 'steps': 8181, 'loss/train': 2.1408222913742065} 01/28/2022 23:22:08 - INFO - codeparrot_training - Step 8182: {'lr': 0.000479813904772484, 'samples': 1571136, 'steps': 8182, 'loss/train': 1.6766690611839294} 01/28/2022 23:22:12 - INFO - codeparrot_training - Step 8183: {'lr': 0.0004798074630146397, 'samples': 1571328, 'steps': 8183, 'loss/train': 6.006467342376709} 01/28/2022 23:22:18 - INFO - codeparrot_training - Step 8184: {'lr': 0.0004798010202723733, 'samples': 1571520, 'steps': 8184, 'loss/train': 0.4242751747369766} 01/28/2022 23:22:22 - INFO - codeparrot_training - Step 8185: {'lr': 0.00047979457654571223, 'samples': 1571712, 'steps': 8185, 'loss/train': 2.0959598422050476} 01/28/2022 23:22:26 - INFO - codeparrot_training - Step 8186: {'lr': 0.0004797881318346842, 'samples': 1571904, 'steps': 8186, 'loss/train': 1.9503439664840698} 01/28/2022 23:22:30 - INFO - codeparrot_training - Step 8187: {'lr': 0.00047978168613931684, 'samples': 1572096, 'steps': 8187, 'loss/train': 4.001049757003784} 01/28/2022 23:22:35 - INFO - codeparrot_training - Step 8188: {'lr': 0.0004797752394596376, 'samples': 1572288, 'steps': 8188, 'loss/train': 2.175572991371155} 01/28/2022 23:22:41 - INFO - codeparrot_training - Step 8189: {'lr': 0.0004797687917956742, 'samples': 1572480, 'steps': 8189, 'loss/train': 1.8452092409133911} 01/28/2022 23:22:45 - INFO - codeparrot_training - Step 8190: {'lr': 0.0004797623431474543, 'samples': 1572672, 'steps': 8190, 'loss/train': 1.6645324230194092} 01/28/2022 23:22:49 - INFO - codeparrot_training - Step 8191: {'lr': 0.0004797558935150055, 'samples': 1572864, 'steps': 8191, 'loss/train': 1.6943246126174927} 01/28/2022 23:22:54 - INFO - codeparrot_training - Step 8192: {'lr': 0.0004797494428983553, 'samples': 1573056, 'steps': 8192, 'loss/train': 1.6441020369529724} 01/28/2022 23:22:58 - INFO - codeparrot_training - Step 8193: {'lr': 0.0004797429912975316, 'samples': 1573248, 'steps': 8193, 'loss/train': 1.8040963411331177} 01/28/2022 23:23:03 - INFO - codeparrot_training - Step 8194: {'lr': 0.00047973653871256173, 'samples': 1573440, 'steps': 8194, 'loss/train': 2.1735137701034546} 01/28/2022 23:23:08 - INFO - codeparrot_training - Step 8195: {'lr': 0.00047973008514347353, 'samples': 1573632, 'steps': 8195, 'loss/train': 1.1857485473155975} 01/28/2022 23:23:12 - INFO - codeparrot_training - Step 8196: {'lr': 0.00047972363059029465, 'samples': 1573824, 'steps': 8196, 'loss/train': 1.2236518263816833} 01/28/2022 23:23:16 - INFO - codeparrot_training - Step 8197: {'lr': 0.0004797171750530526, 'samples': 1574016, 'steps': 8197, 'loss/train': 1.4419191777706146} 01/28/2022 23:23:20 - INFO - codeparrot_training - Step 8198: {'lr': 0.00047971071853177515, 'samples': 1574208, 'steps': 8198, 'loss/train': 2.3372331261634827} 01/28/2022 23:23:26 - INFO - codeparrot_training - Step 8199: {'lr': 0.0004797042610264899, 'samples': 1574400, 'steps': 8199, 'loss/train': 2.215328335762024} 01/28/2022 23:23:30 - INFO - codeparrot_training - Step 8200: {'lr': 0.0004796978025372246, 'samples': 1574592, 'steps': 8200, 'loss/train': 1.6027432680130005} 01/28/2022 23:23:34 - INFO - codeparrot_training - Step 8201: {'lr': 0.0004796913430640068, 'samples': 1574784, 'steps': 8201, 'loss/train': 1.316816121339798} 01/28/2022 23:23:38 - INFO - codeparrot_training - Step 8202: {'lr': 0.0004796848826068642, 'samples': 1574976, 'steps': 8202, 'loss/train': 2.0585081577301025} 01/28/2022 23:23:43 - INFO - codeparrot_training - Step 8203: {'lr': 0.00047967842116582453, 'samples': 1575168, 'steps': 8203, 'loss/train': 2.283383846282959} 01/28/2022 23:23:49 - INFO - codeparrot_training - Step 8204: {'lr': 0.00047967195874091547, 'samples': 1575360, 'steps': 8204, 'loss/train': 2.3724396228790283} 01/28/2022 23:23:53 - INFO - codeparrot_training - Step 8205: {'lr': 0.00047966549533216466, 'samples': 1575552, 'steps': 8205, 'loss/train': 2.1929216980934143} 01/28/2022 23:23:58 - INFO - codeparrot_training - Step 8206: {'lr': 0.00047965903093959974, 'samples': 1575744, 'steps': 8206, 'loss/train': 1.6324684023857117} 01/28/2022 23:24:02 - INFO - codeparrot_training - Step 8207: {'lr': 0.0004796525655632484, 'samples': 1575936, 'steps': 8207, 'loss/train': 1.5147616267204285} 01/28/2022 23:24:06 - INFO - codeparrot_training - Step 8208: {'lr': 0.0004796460992031385, 'samples': 1576128, 'steps': 8208, 'loss/train': 1.8568509221076965} 01/28/2022 23:24:11 - INFO - codeparrot_training - Step 8209: {'lr': 0.0004796396318592976, 'samples': 1576320, 'steps': 8209, 'loss/train': 1.8578656911849976} 01/28/2022 23:24:16 - INFO - codeparrot_training - Step 8210: {'lr': 0.00047963316353175344, 'samples': 1576512, 'steps': 8210, 'loss/train': 1.6886377930641174} 01/28/2022 23:24:20 - INFO - codeparrot_training - Step 8211: {'lr': 0.00047962669422053374, 'samples': 1576704, 'steps': 8211, 'loss/train': 2.010769486427307} 01/28/2022 23:24:24 - INFO - codeparrot_training - Step 8212: {'lr': 0.0004796202239256662, 'samples': 1576896, 'steps': 8212, 'loss/train': 1.5186302661895752} 01/28/2022 23:24:28 - INFO - codeparrot_training - Step 8213: {'lr': 0.0004796137526471785, 'samples': 1577088, 'steps': 8213, 'loss/train': 1.7465917468070984} 01/28/2022 23:24:34 - INFO - codeparrot_training - Step 8214: {'lr': 0.0004796072803850984, 'samples': 1577280, 'steps': 8214, 'loss/train': 0.5696220695972443} 01/28/2022 23:24:38 - INFO - codeparrot_training - Step 8215: {'lr': 0.00047960080713945364, 'samples': 1577472, 'steps': 8215, 'loss/train': 1.9160258173942566} 01/28/2022 23:24:42 - INFO - codeparrot_training - Step 8216: {'lr': 0.0004795943329102719, 'samples': 1577664, 'steps': 8216, 'loss/train': 2.4777292013168335} 01/28/2022 23:24:47 - INFO - codeparrot_training - Step 8217: {'lr': 0.00047958785769758094, 'samples': 1577856, 'steps': 8217, 'loss/train': 1.0215655267238617} 01/28/2022 23:24:51 - INFO - codeparrot_training - Step 8218: {'lr': 0.0004795813815014085, 'samples': 1578048, 'steps': 8218, 'loss/train': 1.0164026319980621} 01/28/2022 23:24:57 - INFO - codeparrot_training - Step 8219: {'lr': 0.0004795749043217824, 'samples': 1578240, 'steps': 8219, 'loss/train': 1.549621045589447} 01/28/2022 23:25:02 - INFO - codeparrot_training - Step 8220: {'lr': 0.0004795684261587302, 'samples': 1578432, 'steps': 8220, 'loss/train': 1.5746819972991943} 01/28/2022 23:25:06 - INFO - codeparrot_training - Step 8221: {'lr': 0.00047956194701227983, 'samples': 1578624, 'steps': 8221, 'loss/train': 1.5813533663749695} 01/28/2022 23:25:10 - INFO - codeparrot_training - Step 8222: {'lr': 0.000479555466882459, 'samples': 1578816, 'steps': 8222, 'loss/train': 2.6761786937713623} 01/28/2022 23:25:14 - INFO - codeparrot_training - Step 8223: {'lr': 0.00047954898576929534, 'samples': 1579008, 'steps': 8223, 'loss/train': 1.7375637888908386} 01/28/2022 23:25:19 - INFO - codeparrot_training - Step 8224: {'lr': 0.0004795425036728168, 'samples': 1579200, 'steps': 8224, 'loss/train': 1.3366458714008331} 01/28/2022 23:25:24 - INFO - codeparrot_training - Step 8225: {'lr': 0.000479536020593051, 'samples': 1579392, 'steps': 8225, 'loss/train': 1.9501360058784485} 01/28/2022 23:25:28 - INFO - codeparrot_training - Step 8226: {'lr': 0.0004795295365300258, 'samples': 1579584, 'steps': 8226, 'loss/train': 1.2832372784614563} 01/28/2022 23:25:33 - INFO - codeparrot_training - Step 8227: {'lr': 0.00047952305148376895, 'samples': 1579776, 'steps': 8227, 'loss/train': 1.537396788597107} 01/28/2022 23:25:37 - INFO - codeparrot_training - Step 8228: {'lr': 0.0004795165654543082, 'samples': 1579968, 'steps': 8228, 'loss/train': 0.9384343922138214} 01/28/2022 23:25:41 - INFO - codeparrot_training - Step 8229: {'lr': 0.0004795100784416714, 'samples': 1580160, 'steps': 8229, 'loss/train': 2.0285589694976807} 01/28/2022 23:25:47 - INFO - codeparrot_training - Step 8230: {'lr': 0.0004795035904458863, 'samples': 1580352, 'steps': 8230, 'loss/train': 2.3230297565460205} 01/28/2022 23:25:51 - INFO - codeparrot_training - Step 8231: {'lr': 0.00047949710146698066, 'samples': 1580544, 'steps': 8231, 'loss/train': 2.3113269209861755} 01/28/2022 23:25:55 - INFO - codeparrot_training - Step 8232: {'lr': 0.0004794906115049824, 'samples': 1580736, 'steps': 8232, 'loss/train': 2.1005024313926697} 01/28/2022 23:25:59 - INFO - codeparrot_training - Step 8233: {'lr': 0.00047948412055991916, 'samples': 1580928, 'steps': 8233, 'loss/train': 1.5171852707862854} 01/28/2022 23:26:04 - INFO - codeparrot_training - Step 8234: {'lr': 0.0004794776286318188, 'samples': 1581120, 'steps': 8234, 'loss/train': 1.8494405150413513} 01/28/2022 23:26:10 - INFO - codeparrot_training - Step 8235: {'lr': 0.0004794711357207092, 'samples': 1581312, 'steps': 8235, 'loss/train': 1.4154198467731476} 01/28/2022 23:26:14 - INFO - codeparrot_training - Step 8236: {'lr': 0.0004794646418266181, 'samples': 1581504, 'steps': 8236, 'loss/train': 1.6107301712036133} 01/28/2022 23:26:19 - INFO - codeparrot_training - Step 8237: {'lr': 0.0004794581469495733, 'samples': 1581696, 'steps': 8237, 'loss/train': 0.20512991398572922} 01/28/2022 23:26:23 - INFO - codeparrot_training - Step 8238: {'lr': 0.00047945165108960274, 'samples': 1581888, 'steps': 8238, 'loss/train': 2.1029938459396362} 01/28/2022 23:26:27 - INFO - codeparrot_training - Step 8239: {'lr': 0.0004794451542467341, 'samples': 1582080, 'steps': 8239, 'loss/train': 2.0196257829666138} 01/28/2022 23:26:33 - INFO - codeparrot_training - Step 8240: {'lr': 0.00047943865642099525, 'samples': 1582272, 'steps': 8240, 'loss/train': 1.5284752249717712} 01/28/2022 23:26:37 - INFO - codeparrot_training - Step 8241: {'lr': 0.0004794321576124141, 'samples': 1582464, 'steps': 8241, 'loss/train': 1.7936246395111084} 01/28/2022 23:26:41 - INFO - codeparrot_training - Step 8242: {'lr': 0.0004794256578210184, 'samples': 1582656, 'steps': 8242, 'loss/train': 1.4742061793804169} 01/28/2022 23:26:46 - INFO - codeparrot_training - Step 8243: {'lr': 0.0004794191570468361, 'samples': 1582848, 'steps': 8243, 'loss/train': 1.6424381732940674} 01/28/2022 23:26:52 - INFO - codeparrot_training - Step 8244: {'lr': 0.00047941265528989496, 'samples': 1583040, 'steps': 8244, 'loss/train': 1.2740954160690308} 01/28/2022 23:26:56 - INFO - codeparrot_training - Step 8245: {'lr': 0.0004794061525502229, 'samples': 1583232, 'steps': 8245, 'loss/train': 1.9658063650131226} 01/28/2022 23:27:00 - INFO - codeparrot_training - Step 8246: {'lr': 0.00047939964882784766, 'samples': 1583424, 'steps': 8246, 'loss/train': 1.506958544254303} 01/28/2022 23:27:05 - INFO - codeparrot_training - Step 8247: {'lr': 0.0004793931441227972, 'samples': 1583616, 'steps': 8247, 'loss/train': 1.8402854204177856} 01/28/2022 23:27:09 - INFO - codeparrot_training - Step 8248: {'lr': 0.00047938663843509927, 'samples': 1583808, 'steps': 8248, 'loss/train': 1.8982658386230469} 01/28/2022 23:27:14 - INFO - codeparrot_training - Step 8249: {'lr': 0.00047938013176478193, 'samples': 1584000, 'steps': 8249, 'loss/train': 1.945980191230774} 01/28/2022 23:27:18 - INFO - codeparrot_training - Step 8250: {'lr': 0.0004793736241118728, 'samples': 1584192, 'steps': 8250, 'loss/train': 1.548606276512146} 01/28/2022 23:27:23 - INFO - codeparrot_training - Step 8251: {'lr': 0.0004793671154764, 'samples': 1584384, 'steps': 8251, 'loss/train': 1.6413021683692932} 01/28/2022 23:27:27 - INFO - codeparrot_training - Step 8252: {'lr': 0.0004793606058583913, 'samples': 1584576, 'steps': 8252, 'loss/train': 1.9956350326538086} 01/28/2022 23:27:31 - INFO - codeparrot_training - Step 8253: {'lr': 0.0004793540952578746, 'samples': 1584768, 'steps': 8253, 'loss/train': 2.3505597710609436} 01/28/2022 23:27:37 - INFO - codeparrot_training - Step 8254: {'lr': 0.0004793475836748777, 'samples': 1584960, 'steps': 8254, 'loss/train': 2.518977999687195} 01/28/2022 23:27:41 - INFO - codeparrot_training - Step 8255: {'lr': 0.0004793410711094287, 'samples': 1585152, 'steps': 8255, 'loss/train': 1.7512089014053345} 01/28/2022 23:27:45 - INFO - codeparrot_training - Step 8256: {'lr': 0.00047933455756155534, 'samples': 1585344, 'steps': 8256, 'loss/train': 2.857607066631317} 01/28/2022 23:27:49 - INFO - codeparrot_training - Step 8257: {'lr': 0.00047932804303128557, 'samples': 1585536, 'steps': 8257, 'loss/train': 1.9699634313583374} 01/28/2022 23:27:54 - INFO - codeparrot_training - Step 8258: {'lr': 0.0004793215275186472, 'samples': 1585728, 'steps': 8258, 'loss/train': 1.5863104462623596} 01/28/2022 23:28:00 - INFO - codeparrot_training - Step 8259: {'lr': 0.0004793150110236684, 'samples': 1585920, 'steps': 8259, 'loss/train': 2.6847445964813232} 01/28/2022 23:28:04 - INFO - codeparrot_training - Step 8260: {'lr': 0.00047930849354637674, 'samples': 1586112, 'steps': 8260, 'loss/train': 2.4168448448181152} 01/28/2022 23:28:08 - INFO - codeparrot_training - Step 8261: {'lr': 0.00047930197508680027, 'samples': 1586304, 'steps': 8261, 'loss/train': 1.378115862607956} 01/28/2022 23:28:13 - INFO - codeparrot_training - Step 8262: {'lr': 0.00047929545564496715, 'samples': 1586496, 'steps': 8262, 'loss/train': 2.0817155241966248} 01/28/2022 23:28:17 - INFO - codeparrot_training - Step 8263: {'lr': 0.0004792889352209049, 'samples': 1586688, 'steps': 8263, 'loss/train': 1.3432399034500122} 01/28/2022 23:28:22 - INFO - codeparrot_training - Step 8264: {'lr': 0.00047928241381464177, 'samples': 1586880, 'steps': 8264, 'loss/train': 2.41122168302536} 01/28/2022 23:28:26 - INFO - codeparrot_training - Step 8265: {'lr': 0.00047927589142620556, 'samples': 1587072, 'steps': 8265, 'loss/train': 1.8314836621284485} 01/28/2022 23:28:31 - INFO - codeparrot_training - Step 8266: {'lr': 0.0004792693680556243, 'samples': 1587264, 'steps': 8266, 'loss/train': 1.794735610485077} 01/28/2022 23:28:35 - INFO - codeparrot_training - Step 8267: {'lr': 0.0004792628437029258, 'samples': 1587456, 'steps': 8267, 'loss/train': 1.5745237469673157} 01/28/2022 23:28:39 - INFO - codeparrot_training - Step 8268: {'lr': 0.0004792563183681381, 'samples': 1587648, 'steps': 8268, 'loss/train': 2.4342477321624756} 01/28/2022 23:28:44 - INFO - codeparrot_training - Step 8269: {'lr': 0.0004792497920512891, 'samples': 1587840, 'steps': 8269, 'loss/train': 1.6522677540779114} 01/28/2022 23:28:49 - INFO - codeparrot_training - Step 8270: {'lr': 0.00047924326475240676, 'samples': 1588032, 'steps': 8270, 'loss/train': 1.9838991165161133} 01/28/2022 23:28:53 - INFO - codeparrot_training - Step 8271: {'lr': 0.00047923673647151915, 'samples': 1588224, 'steps': 8271, 'loss/train': 1.3852111101150513} 01/28/2022 23:28:57 - INFO - codeparrot_training - Step 8272: {'lr': 0.00047923020720865413, 'samples': 1588416, 'steps': 8272, 'loss/train': 1.1496371626853943} 01/28/2022 23:29:01 - INFO - codeparrot_training - Step 8273: {'lr': 0.0004792236769638396, 'samples': 1588608, 'steps': 8273, 'loss/train': 1.750813901424408} 01/28/2022 23:29:07 - INFO - codeparrot_training - Step 8274: {'lr': 0.00047921714573710374, 'samples': 1588800, 'steps': 8274, 'loss/train': 1.571412742137909} 01/28/2022 23:29:11 - INFO - codeparrot_training - Step 8275: {'lr': 0.0004792106135284744, 'samples': 1588992, 'steps': 8275, 'loss/train': 1.1416810154914856} 01/28/2022 23:29:15 - INFO - codeparrot_training - Step 8276: {'lr': 0.00047920408033797954, 'samples': 1589184, 'steps': 8276, 'loss/train': 2.281437635421753} 01/28/2022 23:29:19 - INFO - codeparrot_training - Step 8277: {'lr': 0.00047919754616564716, 'samples': 1589376, 'steps': 8277, 'loss/train': 2.030203878879547} 01/28/2022 23:29:24 - INFO - codeparrot_training - Step 8278: {'lr': 0.0004791910110115053, 'samples': 1589568, 'steps': 8278, 'loss/train': 2.3535009026527405} 01/28/2022 23:29:30 - INFO - codeparrot_training - Step 8279: {'lr': 0.0004791844748755819, 'samples': 1589760, 'steps': 8279, 'loss/train': 0.991818755865097} 01/28/2022 23:29:34 - INFO - codeparrot_training - Step 8280: {'lr': 0.00047917793775790503, 'samples': 1589952, 'steps': 8280, 'loss/train': 1.7177061438560486} 01/28/2022 23:29:39 - INFO - codeparrot_training - Step 8281: {'lr': 0.00047917139965850266, 'samples': 1590144, 'steps': 8281, 'loss/train': 1.867152214050293} 01/28/2022 23:29:43 - INFO - codeparrot_training - Step 8282: {'lr': 0.0004791648605774027, 'samples': 1590336, 'steps': 8282, 'loss/train': 1.2748550176620483} 01/28/2022 23:29:47 - INFO - codeparrot_training - Step 8283: {'lr': 0.00047915832051463326, 'samples': 1590528, 'steps': 8283, 'loss/train': 1.4732052683830261} 01/28/2022 23:29:52 - INFO - codeparrot_training - Step 8284: {'lr': 0.0004791517794702224, 'samples': 1590720, 'steps': 8284, 'loss/train': 2.089179575443268} 01/28/2022 23:29:57 - INFO - codeparrot_training - Step 8285: {'lr': 0.00047914523744419803, 'samples': 1590912, 'steps': 8285, 'loss/train': 1.6179326176643372} 01/28/2022 23:30:01 - INFO - codeparrot_training - Step 8286: {'lr': 0.00047913869443658825, 'samples': 1591104, 'steps': 8286, 'loss/train': 1.97983056306839} 01/28/2022 23:30:06 - INFO - codeparrot_training - Step 8287: {'lr': 0.0004791321504474211, 'samples': 1591296, 'steps': 8287, 'loss/train': 2.5414613485336304} 01/28/2022 23:30:10 - INFO - codeparrot_training - Step 8288: {'lr': 0.00047912560547672453, 'samples': 1591488, 'steps': 8288, 'loss/train': 1.479904979467392} 01/28/2022 23:30:16 - INFO - codeparrot_training - Step 8289: {'lr': 0.0004791190595245266, 'samples': 1591680, 'steps': 8289, 'loss/train': 1.908293902873993} 01/28/2022 23:30:20 - INFO - codeparrot_training - Step 8290: {'lr': 0.0004791125125908554, 'samples': 1591872, 'steps': 8290, 'loss/train': 1.8471549153327942} 01/28/2022 23:30:24 - INFO - codeparrot_training - Step 8291: {'lr': 0.000479105964675739, 'samples': 1592064, 'steps': 8291, 'loss/train': 1.702105700969696} 01/28/2022 23:30:29 - INFO - codeparrot_training - Step 8292: {'lr': 0.0004790994157792053, 'samples': 1592256, 'steps': 8292, 'loss/train': 0.11588030308485031} 01/28/2022 23:30:33 - INFO - codeparrot_training - Step 8293: {'lr': 0.0004790928659012825, 'samples': 1592448, 'steps': 8293, 'loss/train': 1.4194262623786926} 01/28/2022 23:30:40 - INFO - codeparrot_training - Step 8294: {'lr': 0.00047908631504199855, 'samples': 1592640, 'steps': 8294, 'loss/train': 1.8892199993133545} 01/28/2022 23:30:44 - INFO - codeparrot_training - Step 8295: {'lr': 0.00047907976320138163, 'samples': 1592832, 'steps': 8295, 'loss/train': 0.7996648550033569} 01/28/2022 23:30:48 - INFO - codeparrot_training - Step 8296: {'lr': 0.00047907321037945973, 'samples': 1593024, 'steps': 8296, 'loss/train': 1.5666214227676392} 01/28/2022 23:30:52 - INFO - codeparrot_training - Step 8297: {'lr': 0.0004790666565762609, 'samples': 1593216, 'steps': 8297, 'loss/train': 1.6368343234062195} 01/28/2022 23:30:57 - INFO - codeparrot_training - Step 8298: {'lr': 0.0004790601017918134, 'samples': 1593408, 'steps': 8298, 'loss/train': 1.2306329011917114} 01/28/2022 23:31:02 - INFO - codeparrot_training - Step 8299: {'lr': 0.00047905354602614504, 'samples': 1593600, 'steps': 8299, 'loss/train': 2.1181867718696594} 01/28/2022 23:31:06 - INFO - codeparrot_training - Step 8300: {'lr': 0.00047904698927928404, 'samples': 1593792, 'steps': 8300, 'loss/train': 2.394898474216461} 01/28/2022 23:31:10 - INFO - codeparrot_training - Step 8301: {'lr': 0.0004790404315512584, 'samples': 1593984, 'steps': 8301, 'loss/train': 1.3394485116004944} 01/28/2022 23:31:15 - INFO - codeparrot_training - Step 8302: {'lr': 0.0004790338728420963, 'samples': 1594176, 'steps': 8302, 'loss/train': 0.18285422399640083} 01/28/2022 23:31:19 - INFO - codeparrot_training - Step 8303: {'lr': 0.0004790273131518259, 'samples': 1594368, 'steps': 8303, 'loss/train': 1.491758555173874} 01/28/2022 23:31:25 - INFO - codeparrot_training - Step 8304: {'lr': 0.00047902075248047515, 'samples': 1594560, 'steps': 8304, 'loss/train': 2.3759071826934814} 01/28/2022 23:31:30 - INFO - codeparrot_training - Step 8305: {'lr': 0.0004790141908280723, 'samples': 1594752, 'steps': 8305, 'loss/train': 2.694950759410858} 01/28/2022 23:31:34 - INFO - codeparrot_training - Step 8306: {'lr': 0.00047900762819464527, 'samples': 1594944, 'steps': 8306, 'loss/train': 7.3777148723602295} 01/28/2022 23:31:38 - INFO - codeparrot_training - Step 8307: {'lr': 0.0004790010645802223, 'samples': 1595136, 'steps': 8307, 'loss/train': 2.5076165199279785} 01/28/2022 23:31:42 - INFO - codeparrot_training - Step 8308: {'lr': 0.0004789944999848316, 'samples': 1595328, 'steps': 8308, 'loss/train': 1.9066795706748962} 01/28/2022 23:31:48 - INFO - codeparrot_training - Step 8309: {'lr': 0.00047898793440850104, 'samples': 1595520, 'steps': 8309, 'loss/train': 1.7593882083892822} 01/28/2022 23:31:52 - INFO - codeparrot_training - Step 8310: {'lr': 0.0004789813678512589, 'samples': 1595712, 'steps': 8310, 'loss/train': 0.6623484045267105} 01/28/2022 23:31:56 - INFO - codeparrot_training - Step 8311: {'lr': 0.0004789748003131333, 'samples': 1595904, 'steps': 8311, 'loss/train': 2.1717352867126465} 01/28/2022 23:32:00 - INFO - codeparrot_training - Step 8312: {'lr': 0.00047896823179415237, 'samples': 1596096, 'steps': 8312, 'loss/train': 2.1809052228927612} 01/28/2022 23:32:05 - INFO - codeparrot_training - Step 8313: {'lr': 0.00047896166229434423, 'samples': 1596288, 'steps': 8313, 'loss/train': 0.36402033269405365} 01/28/2022 23:32:10 - INFO - codeparrot_training - Step 8314: {'lr': 0.0004789550918137371, 'samples': 1596480, 'steps': 8314, 'loss/train': 1.6227757930755615} 01/28/2022 23:32:14 - INFO - codeparrot_training - Step 8315: {'lr': 0.000478948520352359, 'samples': 1596672, 'steps': 8315, 'loss/train': 1.7098124027252197} 01/28/2022 23:32:18 - INFO - codeparrot_training - Step 8316: {'lr': 0.00047894194791023813, 'samples': 1596864, 'steps': 8316, 'loss/train': 2.382127046585083} 01/28/2022 23:32:23 - INFO - codeparrot_training - Step 8317: {'lr': 0.0004789353744874027, 'samples': 1597056, 'steps': 8317, 'loss/train': 1.7619654536247253} 01/28/2022 23:32:27 - INFO - codeparrot_training - Step 8318: {'lr': 0.0004789288000838808, 'samples': 1597248, 'steps': 8318, 'loss/train': 1.16694974899292} 01/28/2022 23:32:33 - INFO - codeparrot_training - Step 8319: {'lr': 0.0004789222246997006, 'samples': 1597440, 'steps': 8319, 'loss/train': 2.0125595927238464} 01/28/2022 23:32:38 - INFO - codeparrot_training - Step 8320: {'lr': 0.00047891564833489034, 'samples': 1597632, 'steps': 8320, 'loss/train': 2.1167938113212585} 01/28/2022 23:32:42 - INFO - codeparrot_training - Step 8321: {'lr': 0.000478909070989478, 'samples': 1597824, 'steps': 8321, 'loss/train': 2.0563817024230957} 01/28/2022 23:32:46 - INFO - codeparrot_training - Step 8322: {'lr': 0.00047890249266349194, 'samples': 1598016, 'steps': 8322, 'loss/train': 1.7593225836753845} 01/28/2022 23:32:50 - INFO - codeparrot_training - Step 8323: {'lr': 0.0004788959133569604, 'samples': 1598208, 'steps': 8323, 'loss/train': 2.316305458545685} 01/28/2022 23:32:55 - INFO - codeparrot_training - Step 8324: {'lr': 0.00047888933306991136, 'samples': 1598400, 'steps': 8324, 'loss/train': 1.9069490432739258} 01/28/2022 23:33:00 - INFO - codeparrot_training - Step 8325: {'lr': 0.00047888275180237304, 'samples': 1598592, 'steps': 8325, 'loss/train': 1.749363899230957} 01/28/2022 23:33:04 - INFO - codeparrot_training - Step 8326: {'lr': 0.00047887616955437373, 'samples': 1598784, 'steps': 8326, 'loss/train': 1.9812728762626648} 01/28/2022 23:33:08 - INFO - codeparrot_training - Step 8327: {'lr': 0.0004788695863259416, 'samples': 1598976, 'steps': 8327, 'loss/train': 1.3936545252799988} 01/28/2022 23:33:13 - INFO - codeparrot_training - Step 8328: {'lr': 0.0004788630021171049, 'samples': 1599168, 'steps': 8328, 'loss/train': 2.2088754773139954} 01/28/2022 23:33:18 - INFO - codeparrot_training - Step 8329: {'lr': 0.0004788564169278917, 'samples': 1599360, 'steps': 8329, 'loss/train': 2.1809574365615845} 01/28/2022 23:33:22 - INFO - codeparrot_training - Step 8330: {'lr': 0.00047884983075833023, 'samples': 1599552, 'steps': 8330, 'loss/train': 1.935439109802246} 01/28/2022 23:33:26 - INFO - codeparrot_training - Step 8331: {'lr': 0.00047884324360844885, 'samples': 1599744, 'steps': 8331, 'loss/train': 1.6880355477333069} 01/28/2022 23:33:31 - INFO - codeparrot_training - Step 8332: {'lr': 0.0004788366554782756, 'samples': 1599936, 'steps': 8332, 'loss/train': 1.882067620754242} 01/28/2022 23:33:35 - INFO - codeparrot_training - Step 8333: {'lr': 0.00047883006636783887, 'samples': 1600128, 'steps': 8333, 'loss/train': 1.6858434677124023} 01/28/2022 23:33:40 - INFO - codeparrot_training - Step 8334: {'lr': 0.0004788234762771667, 'samples': 1600320, 'steps': 8334, 'loss/train': 2.1985334157943726} 01/28/2022 23:33:44 - INFO - codeparrot_training - Step 8335: {'lr': 0.0004788168852062875, 'samples': 1600512, 'steps': 8335, 'loss/train': 1.6883729696273804} 01/28/2022 23:33:48 - INFO - codeparrot_training - Step 8336: {'lr': 0.0004788102931552294, 'samples': 1600704, 'steps': 8336, 'loss/train': 2.04763126373291} 01/28/2022 23:33:53 - INFO - codeparrot_training - Step 8337: {'lr': 0.00047880370012402064, 'samples': 1600896, 'steps': 8337, 'loss/train': 2.4194889664649963} 01/28/2022 23:33:57 - INFO - codeparrot_training - Step 8338: {'lr': 0.0004787971061126895, 'samples': 1601088, 'steps': 8338, 'loss/train': 0.3153833895921707} 01/28/2022 23:34:03 - INFO - codeparrot_training - Step 8339: {'lr': 0.0004787905111212642, 'samples': 1601280, 'steps': 8339, 'loss/train': 1.7896637320518494} 01/28/2022 23:34:07 - INFO - codeparrot_training - Step 8340: {'lr': 0.00047878391514977306, 'samples': 1601472, 'steps': 8340, 'loss/train': 1.081070601940155} 01/28/2022 23:34:12 - INFO - codeparrot_training - Step 8341: {'lr': 0.0004787773181982442, 'samples': 1601664, 'steps': 8341, 'loss/train': 2.205226242542267} 01/28/2022 23:34:16 - INFO - codeparrot_training - Step 8342: {'lr': 0.0004787707202667059, 'samples': 1601856, 'steps': 8342, 'loss/train': 1.555022120475769} 01/28/2022 23:34:20 - INFO - codeparrot_training - Step 8343: {'lr': 0.00047876412135518655, 'samples': 1602048, 'steps': 8343, 'loss/train': 1.1652584373950958} 01/28/2022 23:34:26 - INFO - codeparrot_training - Step 8344: {'lr': 0.0004787575214637144, 'samples': 1602240, 'steps': 8344, 'loss/train': 1.9640700817108154} 01/28/2022 23:34:30 - INFO - codeparrot_training - Step 8345: {'lr': 0.00047875092059231756, 'samples': 1602432, 'steps': 8345, 'loss/train': 1.2616499662399292} 01/28/2022 23:34:34 - INFO - codeparrot_training - Step 8346: {'lr': 0.0004787443187410245, 'samples': 1602624, 'steps': 8346, 'loss/train': 1.95295250415802} 01/28/2022 23:34:38 - INFO - codeparrot_training - Step 8347: {'lr': 0.00047873771590986337, 'samples': 1602816, 'steps': 8347, 'loss/train': 1.4795432388782501} 01/28/2022 23:34:43 - INFO - codeparrot_training - Step 8348: {'lr': 0.00047873111209886245, 'samples': 1603008, 'steps': 8348, 'loss/train': 2.120443046092987} 01/28/2022 23:34:49 - INFO - codeparrot_training - Step 8349: {'lr': 0.00047872450730805015, 'samples': 1603200, 'steps': 8349, 'loss/train': 1.7403318285942078} 01/28/2022 23:34:53 - INFO - codeparrot_training - Step 8350: {'lr': 0.00047871790153745464, 'samples': 1603392, 'steps': 8350, 'loss/train': 2.4829747080802917} 01/28/2022 23:34:58 - INFO - codeparrot_training - Step 8351: {'lr': 0.0004787112947871043, 'samples': 1603584, 'steps': 8351, 'loss/train': 1.7125797271728516} 01/28/2022 23:35:02 - INFO - codeparrot_training - Step 8352: {'lr': 0.0004787046870570274, 'samples': 1603776, 'steps': 8352, 'loss/train': 1.91068696975708} 01/28/2022 23:35:06 - INFO - codeparrot_training - Step 8353: {'lr': 0.00047869807834725225, 'samples': 1603968, 'steps': 8353, 'loss/train': 1.7818937301635742} 01/28/2022 23:35:11 - INFO - codeparrot_training - Step 8354: {'lr': 0.0004786914686578071, 'samples': 1604160, 'steps': 8354, 'loss/train': 2.44059294462204} 01/28/2022 23:35:16 - INFO - codeparrot_training - Step 8355: {'lr': 0.00047868485798872044, 'samples': 1604352, 'steps': 8355, 'loss/train': 1.9464592337608337} 01/28/2022 23:35:20 - INFO - codeparrot_training - Step 8356: {'lr': 0.00047867824634002034, 'samples': 1604544, 'steps': 8356, 'loss/train': 0.8032460510730743} 01/28/2022 23:35:24 - INFO - codeparrot_training - Step 8357: {'lr': 0.0004786716337117353, 'samples': 1604736, 'steps': 8357, 'loss/train': 1.8397216200828552} 01/28/2022 23:35:28 - INFO - codeparrot_training - Step 8358: {'lr': 0.00047866502010389356, 'samples': 1604928, 'steps': 8358, 'loss/train': 2.1420278549194336} 01/28/2022 23:35:33 - INFO - codeparrot_training - Step 8359: {'lr': 0.00047865840551652343, 'samples': 1605120, 'steps': 8359, 'loss/train': 1.9651220440864563} 01/28/2022 23:35:38 - INFO - codeparrot_training - Step 8360: {'lr': 0.0004786517899496534, 'samples': 1605312, 'steps': 8360, 'loss/train': 1.1495666205883026} 01/28/2022 23:35:42 - INFO - codeparrot_training - Step 8361: {'lr': 0.0004786451734033117, 'samples': 1605504, 'steps': 8361, 'loss/train': 1.9448978304862976} 01/28/2022 23:35:46 - INFO - codeparrot_training - Step 8362: {'lr': 0.00047863855587752666, 'samples': 1605696, 'steps': 8362, 'loss/train': 1.5856195092201233} 01/28/2022 23:35:50 - INFO - codeparrot_training - Step 8363: {'lr': 0.0004786319373723266, 'samples': 1605888, 'steps': 8363, 'loss/train': 1.0556966364383698} 01/28/2022 23:35:57 - INFO - codeparrot_training - Step 8364: {'lr': 0.00047862531788774, 'samples': 1606080, 'steps': 8364, 'loss/train': 1.579420566558838} 01/28/2022 23:36:01 - INFO - codeparrot_training - Step 8365: {'lr': 0.00047861869742379503, 'samples': 1606272, 'steps': 8365, 'loss/train': 1.5446873903274536} 01/28/2022 23:36:05 - INFO - codeparrot_training - Step 8366: {'lr': 0.0004786120759805203, 'samples': 1606464, 'steps': 8366, 'loss/train': 1.1489618718624115} 01/28/2022 23:36:10 - INFO - codeparrot_training - Step 8367: {'lr': 0.0004786054535579439, 'samples': 1606656, 'steps': 8367, 'loss/train': 1.2359886467456818} 01/28/2022 23:36:14 - INFO - codeparrot_training - Step 8368: {'lr': 0.0004785988301560944, 'samples': 1606848, 'steps': 8368, 'loss/train': 2.2880626916885376} 01/28/2022 23:36:19 - INFO - codeparrot_training - Step 8369: {'lr': 0.0004785922057750001, 'samples': 1607040, 'steps': 8369, 'loss/train': 0.9159793853759766} 01/28/2022 23:36:23 - INFO - codeparrot_training - Step 8370: {'lr': 0.00047858558041468925, 'samples': 1607232, 'steps': 8370, 'loss/train': 1.5954376459121704} 01/28/2022 23:36:28 - INFO - codeparrot_training - Step 8371: {'lr': 0.0004785789540751905, 'samples': 1607424, 'steps': 8371, 'loss/train': 1.9775294065475464} 01/28/2022 23:36:32 - INFO - codeparrot_training - Step 8372: {'lr': 0.00047857232675653207, 'samples': 1607616, 'steps': 8372, 'loss/train': 1.1849581897258759} 01/28/2022 23:36:36 - INFO - codeparrot_training - Step 8373: {'lr': 0.0004785656984587423, 'samples': 1607808, 'steps': 8373, 'loss/train': 1.6922326683998108} 01/28/2022 23:36:43 - INFO - codeparrot_training - Step 8374: {'lr': 0.0004785590691818498, 'samples': 1608000, 'steps': 8374, 'loss/train': 2.312446653842926} 01/28/2022 23:36:48 - INFO - codeparrot_training - Step 8375: {'lr': 0.0004785524389258827, 'samples': 1608192, 'steps': 8375, 'loss/train': 1.6185507774353027} 01/28/2022 23:36:52 - INFO - codeparrot_training - Step 8376: {'lr': 0.0004785458076908695, 'samples': 1608384, 'steps': 8376, 'loss/train': 0.37733638286590576} 01/28/2022 23:36:56 - INFO - codeparrot_training - Step 8377: {'lr': 0.00047853917547683873, 'samples': 1608576, 'steps': 8377, 'loss/train': 1.8402249813079834} 01/28/2022 23:37:00 - INFO - codeparrot_training - Step 8378: {'lr': 0.00047853254228381864, 'samples': 1608768, 'steps': 8378, 'loss/train': 1.5733453631401062} 01/28/2022 23:37:06 - INFO - codeparrot_training - Step 8379: {'lr': 0.0004785259081118377, 'samples': 1608960, 'steps': 8379, 'loss/train': 1.5412498712539673} 01/28/2022 23:37:10 - INFO - codeparrot_training - Step 8380: {'lr': 0.0004785192729609244, 'samples': 1609152, 'steps': 8380, 'loss/train': 2.0584208965301514} 01/28/2022 23:37:14 - INFO - codeparrot_training - Step 8381: {'lr': 0.00047851263683110706, 'samples': 1609344, 'steps': 8381, 'loss/train': 1.731005072593689} 01/28/2022 23:37:19 - INFO - codeparrot_training - Step 8382: {'lr': 0.0004785059997224142, 'samples': 1609536, 'steps': 8382, 'loss/train': 1.7289699912071228} 01/28/2022 23:37:24 - INFO - codeparrot_training - Step 8383: {'lr': 0.0004784993616348741, 'samples': 1609728, 'steps': 8383, 'loss/train': 2.3300009965896606} 01/28/2022 23:37:28 - INFO - codeparrot_training - Step 8384: {'lr': 0.0004784927225685153, 'samples': 1609920, 'steps': 8384, 'loss/train': 2.13554710149765} 01/28/2022 23:37:32 - INFO - codeparrot_training - Step 8385: {'lr': 0.0004784860825233662, 'samples': 1610112, 'steps': 8385, 'loss/train': 1.6816365122795105} 01/28/2022 23:37:37 - INFO - codeparrot_training - Step 8386: {'lr': 0.00047847944149945545, 'samples': 1610304, 'steps': 8386, 'loss/train': 1.875330626964569} 01/28/2022 23:37:41 - INFO - codeparrot_training - Step 8387: {'lr': 0.00047847279949681117, 'samples': 1610496, 'steps': 8387, 'loss/train': 1.030475378036499} 01/28/2022 23:37:45 - INFO - codeparrot_training - Step 8388: {'lr': 0.000478466156515462, 'samples': 1610688, 'steps': 8388, 'loss/train': 0.5964917242527008} 01/28/2022 23:37:51 - INFO - codeparrot_training - Step 8389: {'lr': 0.0004784595125554364, 'samples': 1610880, 'steps': 8389, 'loss/train': 0.7475326359272003} 01/28/2022 23:37:55 - INFO - codeparrot_training - Step 8390: {'lr': 0.00047845286761676276, 'samples': 1611072, 'steps': 8390, 'loss/train': 3.887911319732666} 01/28/2022 23:37:59 - INFO - codeparrot_training - Step 8391: {'lr': 0.00047844622169946954, 'samples': 1611264, 'steps': 8391, 'loss/train': 1.753699243068695} 01/28/2022 23:38:04 - INFO - codeparrot_training - Step 8392: {'lr': 0.0004784395748035853, 'samples': 1611456, 'steps': 8392, 'loss/train': 1.8532281517982483} 01/28/2022 23:38:08 - INFO - codeparrot_training - Step 8393: {'lr': 0.0004784329269291384, 'samples': 1611648, 'steps': 8393, 'loss/train': 1.8878461718559265} 01/28/2022 23:38:14 - INFO - codeparrot_training - Step 8394: {'lr': 0.0004784262780761575, 'samples': 1611840, 'steps': 8394, 'loss/train': 1.1579212546348572} 01/28/2022 23:38:19 - INFO - codeparrot_training - Step 8395: {'lr': 0.00047841962824467086, 'samples': 1612032, 'steps': 8395, 'loss/train': 1.2470292448997498} 01/28/2022 23:38:23 - INFO - codeparrot_training - Step 8396: {'lr': 0.000478412977434707, 'samples': 1612224, 'steps': 8396, 'loss/train': 1.6459481120109558} 01/28/2022 23:38:27 - INFO - codeparrot_training - Step 8397: {'lr': 0.0004784063256462946, 'samples': 1612416, 'steps': 8397, 'loss/train': 1.4800085127353668} 01/28/2022 23:38:31 - INFO - codeparrot_training - Step 8398: {'lr': 0.00047839967287946196, 'samples': 1612608, 'steps': 8398, 'loss/train': 1.3968497514724731} 01/28/2022 23:38:37 - INFO - codeparrot_training - Step 8399: {'lr': 0.00047839301913423773, 'samples': 1612800, 'steps': 8399, 'loss/train': 1.673012137413025} 01/28/2022 23:38:41 - INFO - codeparrot_training - Step 8400: {'lr': 0.0004783863644106502, 'samples': 1612992, 'steps': 8400, 'loss/train': 1.1592565476894379} 01/28/2022 23:38:45 - INFO - codeparrot_training - Step 8401: {'lr': 0.0004783797087087281, 'samples': 1613184, 'steps': 8401, 'loss/train': 2.3275980949401855} 01/28/2022 23:38:49 - INFO - codeparrot_training - Step 8402: {'lr': 0.00047837305202849987, 'samples': 1613376, 'steps': 8402, 'loss/train': 2.099149525165558} 01/28/2022 23:38:54 - INFO - codeparrot_training - Step 8403: {'lr': 0.0004783663943699939, 'samples': 1613568, 'steps': 8403, 'loss/train': 1.561631977558136} 01/28/2022 23:38:59 - INFO - codeparrot_training - Step 8404: {'lr': 0.00047835973573323885, 'samples': 1613760, 'steps': 8404, 'loss/train': 1.0706333220005035} 01/28/2022 23:39:03 - INFO - codeparrot_training - Step 8405: {'lr': 0.00047835307611826327, 'samples': 1613952, 'steps': 8405, 'loss/train': 0.4830096960067749} 01/28/2022 23:39:08 - INFO - codeparrot_training - Step 8406: {'lr': 0.0004783464155250955, 'samples': 1614144, 'steps': 8406, 'loss/train': 2.7486737966537476} 01/28/2022 23:39:12 - INFO - codeparrot_training - Step 8407: {'lr': 0.00047833975395376426, 'samples': 1614336, 'steps': 8407, 'loss/train': 1.2147127389907837} 01/28/2022 23:39:16 - INFO - codeparrot_training - Step 8408: {'lr': 0.00047833309140429803, 'samples': 1614528, 'steps': 8408, 'loss/train': 1.7480033040046692} 01/28/2022 23:39:22 - INFO - codeparrot_training - Step 8409: {'lr': 0.00047832642787672537, 'samples': 1614720, 'steps': 8409, 'loss/train': 2.382943332195282} 01/28/2022 23:39:27 - INFO - codeparrot_training - Step 8410: {'lr': 0.00047831976337107474, 'samples': 1614912, 'steps': 8410, 'loss/train': 2.9340230226516724} 01/28/2022 23:39:31 - INFO - codeparrot_training - Step 8411: {'lr': 0.00047831309788737476, 'samples': 1615104, 'steps': 8411, 'loss/train': 0.8711531460285187} 01/28/2022 23:39:35 - INFO - codeparrot_training - Step 8412: {'lr': 0.000478306431425654, 'samples': 1615296, 'steps': 8412, 'loss/train': 1.798323154449463} 01/28/2022 23:39:39 - INFO - codeparrot_training - Step 8413: {'lr': 0.0004782997639859409, 'samples': 1615488, 'steps': 8413, 'loss/train': 2.2703161239624023} 01/28/2022 23:39:46 - INFO - codeparrot_training - Step 8414: {'lr': 0.00047829309556826415, 'samples': 1615680, 'steps': 8414, 'loss/train': 1.6566901803016663} 01/28/2022 23:39:50 - INFO - codeparrot_training - Step 8415: {'lr': 0.0004782864261726523, 'samples': 1615872, 'steps': 8415, 'loss/train': 1.9590155482292175} 01/28/2022 23:39:54 - INFO - codeparrot_training - Step 8416: {'lr': 0.0004782797557991339, 'samples': 1616064, 'steps': 8416, 'loss/train': 1.916592299938202} 01/28/2022 23:39:59 - INFO - codeparrot_training - Step 8417: {'lr': 0.00047827308444773746, 'samples': 1616256, 'steps': 8417, 'loss/train': 1.806049883365631} 01/28/2022 23:40:03 - INFO - codeparrot_training - Step 8418: {'lr': 0.00047826641211849165, 'samples': 1616448, 'steps': 8418, 'loss/train': 2.1721810698509216} 01/28/2022 23:40:07 - INFO - codeparrot_training - Step 8419: {'lr': 0.000478259738811425, 'samples': 1616640, 'steps': 8419, 'loss/train': 2.1018640995025635} 01/28/2022 23:40:13 - INFO - codeparrot_training - Step 8420: {'lr': 0.0004782530645265661, 'samples': 1616832, 'steps': 8420, 'loss/train': 0.8735218048095703} 01/28/2022 23:40:18 - INFO - codeparrot_training - Step 8421: {'lr': 0.00047824638926394355, 'samples': 1617024, 'steps': 8421, 'loss/train': 1.7116684913635254} 01/28/2022 23:40:22 - INFO - codeparrot_training - Step 8422: {'lr': 0.0004782397130235859, 'samples': 1617216, 'steps': 8422, 'loss/train': 1.6485503911972046} 01/28/2022 23:40:26 - INFO - codeparrot_training - Step 8423: {'lr': 0.0004782330358055219, 'samples': 1617408, 'steps': 8423, 'loss/train': 2.004319131374359} 01/28/2022 23:40:30 - INFO - codeparrot_training - Step 8424: {'lr': 0.00047822635760977995, 'samples': 1617600, 'steps': 8424, 'loss/train': 0.6104010343551636} 01/28/2022 23:40:36 - INFO - codeparrot_training - Step 8425: {'lr': 0.0004782196784363888, 'samples': 1617792, 'steps': 8425, 'loss/train': 1.6763663291931152} 01/28/2022 23:40:40 - INFO - codeparrot_training - Step 8426: {'lr': 0.000478212998285377, 'samples': 1617984, 'steps': 8426, 'loss/train': 1.5564182996749878} 01/28/2022 23:40:44 - INFO - codeparrot_training - Step 8427: {'lr': 0.0004782063171567732, 'samples': 1618176, 'steps': 8427, 'loss/train': 2.0926013588905334} 01/28/2022 23:40:48 - INFO - codeparrot_training - Step 8428: {'lr': 0.000478199635050606, 'samples': 1618368, 'steps': 8428, 'loss/train': 1.6663329005241394} 01/28/2022 23:40:53 - INFO - codeparrot_training - Step 8429: {'lr': 0.000478192951966904, 'samples': 1618560, 'steps': 8429, 'loss/train': 2.4476741552352905} 01/28/2022 23:40:59 - INFO - codeparrot_training - Step 8430: {'lr': 0.00047818626790569586, 'samples': 1618752, 'steps': 8430, 'loss/train': 1.937357246875763} 01/28/2022 23:41:03 - INFO - codeparrot_training - Step 8431: {'lr': 0.00047817958286701026, 'samples': 1618944, 'steps': 8431, 'loss/train': 1.4852543771266937} 01/28/2022 23:41:07 - INFO - codeparrot_training - Step 8432: {'lr': 0.00047817289685087575, 'samples': 1619136, 'steps': 8432, 'loss/train': 1.9135052561759949} 01/28/2022 23:41:12 - INFO - codeparrot_training - Step 8433: {'lr': 0.00047816620985732095, 'samples': 1619328, 'steps': 8433, 'loss/train': 2.035112500190735} 01/28/2022 23:41:16 - INFO - codeparrot_training - Step 8434: {'lr': 0.0004781595218863746, 'samples': 1619520, 'steps': 8434, 'loss/train': 1.6949608325958252} 01/28/2022 23:41:22 - INFO - codeparrot_training - Step 8435: {'lr': 0.00047815283293806533, 'samples': 1619712, 'steps': 8435, 'loss/train': 1.5810068249702454} 01/28/2022 23:41:27 - INFO - codeparrot_training - Step 8436: {'lr': 0.0004781461430124217, 'samples': 1619904, 'steps': 8436, 'loss/train': 2.0802072286605835} 01/28/2022 23:41:31 - INFO - codeparrot_training - Step 8437: {'lr': 0.0004781394521094725, 'samples': 1620096, 'steps': 8437, 'loss/train': 2.7012938261032104} 01/28/2022 23:41:35 - INFO - codeparrot_training - Step 8438: {'lr': 0.00047813276022924634, 'samples': 1620288, 'steps': 8438, 'loss/train': 1.6218053698539734} 01/28/2022 23:41:39 - INFO - codeparrot_training - Step 8439: {'lr': 0.0004781260673717718, 'samples': 1620480, 'steps': 8439, 'loss/train': 1.0832648277282715} 01/28/2022 23:41:45 - INFO - codeparrot_training - Step 8440: {'lr': 0.0004781193735370777, 'samples': 1620672, 'steps': 8440, 'loss/train': 1.249299019575119} 01/28/2022 23:41:49 - INFO - codeparrot_training - Step 8441: {'lr': 0.0004781126787251926, 'samples': 1620864, 'steps': 8441, 'loss/train': 1.8376554250717163} 01/28/2022 23:41:53 - INFO - codeparrot_training - Step 8442: {'lr': 0.0004781059829361453, 'samples': 1621056, 'steps': 8442, 'loss/train': 1.7073875069618225} 01/28/2022 23:41:58 - INFO - codeparrot_training - Step 8443: {'lr': 0.00047809928616996425, 'samples': 1621248, 'steps': 8443, 'loss/train': 1.835850477218628} 01/28/2022 23:42:02 - INFO - codeparrot_training - Step 8444: {'lr': 0.00047809258842667837, 'samples': 1621440, 'steps': 8444, 'loss/train': 1.1594673693180084} 01/28/2022 23:42:07 - INFO - codeparrot_training - Step 8445: {'lr': 0.00047808588970631627, 'samples': 1621632, 'steps': 8445, 'loss/train': 1.5557658076286316} 01/28/2022 23:42:11 - INFO - codeparrot_training - Step 8446: {'lr': 0.0004780791900089066, 'samples': 1621824, 'steps': 8446, 'loss/train': 1.4457016289234161} 01/28/2022 23:42:15 - INFO - codeparrot_training - Step 8447: {'lr': 0.0004780724893344782, 'samples': 1622016, 'steps': 8447, 'loss/train': 1.9329162240028381} 01/28/2022 23:42:20 - INFO - codeparrot_training - Step 8448: {'lr': 0.00047806578768305963, 'samples': 1622208, 'steps': 8448, 'loss/train': 2.1740421652793884} 01/28/2022 23:42:24 - INFO - codeparrot_training - Step 8449: {'lr': 0.00047805908505467963, 'samples': 1622400, 'steps': 8449, 'loss/train': 2.8106001019477844} 01/28/2022 23:42:31 - INFO - codeparrot_training - Step 8450: {'lr': 0.0004780523814493669, 'samples': 1622592, 'steps': 8450, 'loss/train': 2.254831552505493} 01/28/2022 23:42:35 - INFO - codeparrot_training - Step 8451: {'lr': 0.0004780456768671503, 'samples': 1622784, 'steps': 8451, 'loss/train': 1.1263148188591003} 01/28/2022 23:42:39 - INFO - codeparrot_training - Step 8452: {'lr': 0.0004780389713080583, 'samples': 1622976, 'steps': 8452, 'loss/train': 1.971414864063263} 01/28/2022 23:42:43 - INFO - codeparrot_training - Step 8453: {'lr': 0.0004780322647721198, 'samples': 1623168, 'steps': 8453, 'loss/train': 1.3919185996055603} 01/28/2022 23:42:48 - INFO - codeparrot_training - Step 8454: {'lr': 0.00047802555725936347, 'samples': 1623360, 'steps': 8454, 'loss/train': 2.2156893610954285} 01/28/2022 23:42:53 - INFO - codeparrot_training - Step 8455: {'lr': 0.00047801884876981813, 'samples': 1623552, 'steps': 8455, 'loss/train': 2.163508951663971} 01/28/2022 23:42:57 - INFO - codeparrot_training - Step 8456: {'lr': 0.0004780121393035124, 'samples': 1623744, 'steps': 8456, 'loss/train': 1.4961950182914734} 01/28/2022 23:43:02 - INFO - codeparrot_training - Step 8457: {'lr': 0.00047800542886047506, 'samples': 1623936, 'steps': 8457, 'loss/train': 1.7579233646392822} 01/28/2022 23:43:06 - INFO - codeparrot_training - Step 8458: {'lr': 0.00047799871744073485, 'samples': 1624128, 'steps': 8458, 'loss/train': 2.39155250787735} 01/28/2022 23:43:10 - INFO - codeparrot_training - Step 8459: {'lr': 0.00047799200504432054, 'samples': 1624320, 'steps': 8459, 'loss/train': 2.548298478126526} 01/28/2022 23:43:16 - INFO - codeparrot_training - Step 8460: {'lr': 0.0004779852916712609, 'samples': 1624512, 'steps': 8460, 'loss/train': 1.9428457617759705} 01/28/2022 23:43:20 - INFO - codeparrot_training - Step 8461: {'lr': 0.0004779785773215847, 'samples': 1624704, 'steps': 8461, 'loss/train': 1.5186556577682495} 01/28/2022 23:43:24 - INFO - codeparrot_training - Step 8462: {'lr': 0.00047797186199532055, 'samples': 1624896, 'steps': 8462, 'loss/train': 0.208734892308712} 01/28/2022 23:43:29 - INFO - codeparrot_training - Step 8463: {'lr': 0.0004779651456924974, 'samples': 1625088, 'steps': 8463, 'loss/train': 2.196469724178314} 01/28/2022 23:43:33 - INFO - codeparrot_training - Step 8464: {'lr': 0.00047795842841314394, 'samples': 1625280, 'steps': 8464, 'loss/train': 2.121881067752838} 01/28/2022 23:43:38 - INFO - codeparrot_training - Step 8465: {'lr': 0.000477951710157289, 'samples': 1625472, 'steps': 8465, 'loss/train': 1.6519496440887451} 01/28/2022 23:43:43 - INFO - codeparrot_training - Step 8466: {'lr': 0.00047794499092496123, 'samples': 1625664, 'steps': 8466, 'loss/train': 1.4932176768779755} 01/28/2022 23:43:47 - INFO - codeparrot_training - Step 8467: {'lr': 0.00047793827071618955, 'samples': 1625856, 'steps': 8467, 'loss/train': 1.9623128771781921} 01/28/2022 23:43:51 - INFO - codeparrot_training - Step 8468: {'lr': 0.0004779315495310027, 'samples': 1626048, 'steps': 8468, 'loss/train': 2.036266028881073} 01/28/2022 23:43:57 - INFO - codeparrot_training - Step 8469: {'lr': 0.00047792482736942955, 'samples': 1626240, 'steps': 8469, 'loss/train': 2.0425561666488647} 01/28/2022 23:44:02 - INFO - codeparrot_training - Step 8470: {'lr': 0.00047791810423149873, 'samples': 1626432, 'steps': 8470, 'loss/train': 1.566725492477417} 01/28/2022 23:44:06 - INFO - codeparrot_training - Step 8471: {'lr': 0.0004779113801172391, 'samples': 1626624, 'steps': 8471, 'loss/train': 1.7893651127815247} 01/28/2022 23:44:10 - INFO - codeparrot_training - Step 8472: {'lr': 0.0004779046550266795, 'samples': 1626816, 'steps': 8472, 'loss/train': 1.6925323605537415} 01/28/2022 23:44:15 - INFO - codeparrot_training - Step 8473: {'lr': 0.00047789792895984874, 'samples': 1627008, 'steps': 8473, 'loss/train': 1.7786570191383362} 01/28/2022 23:44:20 - INFO - codeparrot_training - Step 8474: {'lr': 0.0004778912019167756, 'samples': 1627200, 'steps': 8474, 'loss/train': 0.38726484775543213} 01/28/2022 23:44:24 - INFO - codeparrot_training - Step 8475: {'lr': 0.00047788447389748894, 'samples': 1627392, 'steps': 8475, 'loss/train': 1.835214614868164} 01/28/2022 23:44:28 - INFO - codeparrot_training - Step 8476: {'lr': 0.0004778777449020176, 'samples': 1627584, 'steps': 8476, 'loss/train': 1.7770925760269165} 01/28/2022 23:44:33 - INFO - codeparrot_training - Step 8477: {'lr': 0.0004778710149303903, 'samples': 1627776, 'steps': 8477, 'loss/train': 2.1550206542015076} 01/28/2022 23:44:37 - INFO - codeparrot_training - Step 8478: {'lr': 0.00047786428398263595, 'samples': 1627968, 'steps': 8478, 'loss/train': 0.7916440665721893} 01/28/2022 23:44:43 - INFO - codeparrot_training - Step 8479: {'lr': 0.00047785755205878333, 'samples': 1628160, 'steps': 8479, 'loss/train': 2.282312750816345} 01/28/2022 23:44:48 - INFO - codeparrot_training - Step 8480: {'lr': 0.0004778508191588613, 'samples': 1628352, 'steps': 8480, 'loss/train': 1.8270758986473083} 01/28/2022 23:44:52 - INFO - codeparrot_training - Step 8481: {'lr': 0.0004778440852828988, 'samples': 1628544, 'steps': 8481, 'loss/train': 0.9754917919635773} 01/28/2022 23:44:56 - INFO - codeparrot_training - Step 8482: {'lr': 0.00047783735043092446, 'samples': 1628736, 'steps': 8482, 'loss/train': 2.0899738669395447} 01/28/2022 23:45:00 - INFO - codeparrot_training - Step 8483: {'lr': 0.0004778306146029674, 'samples': 1628928, 'steps': 8483, 'loss/train': 1.8399680256843567} 01/28/2022 23:45:06 - INFO - codeparrot_training - Step 8484: {'lr': 0.0004778238777990562, 'samples': 1629120, 'steps': 8484, 'loss/train': 1.7027698159217834} 01/28/2022 23:45:10 - INFO - codeparrot_training - Step 8485: {'lr': 0.00047781714001921997, 'samples': 1629312, 'steps': 8485, 'loss/train': 2.3590282201766968} 01/28/2022 23:45:14 - INFO - codeparrot_training - Step 8486: {'lr': 0.00047781040126348734, 'samples': 1629504, 'steps': 8486, 'loss/train': 1.5724133849143982} 01/28/2022 23:45:19 - INFO - codeparrot_training - Step 8487: {'lr': 0.0004778036615318874, 'samples': 1629696, 'steps': 8487, 'loss/train': 2.044830322265625} 01/28/2022 23:45:23 - INFO - codeparrot_training - Step 8488: {'lr': 0.0004777969208244488, 'samples': 1629888, 'steps': 8488, 'loss/train': 2.2902188301086426} 01/28/2022 23:45:28 - INFO - codeparrot_training - Step 8489: {'lr': 0.0004777901791412006, 'samples': 1630080, 'steps': 8489, 'loss/train': 1.6515660881996155} 01/28/2022 23:45:32 - INFO - codeparrot_training - Step 8490: {'lr': 0.00047778343648217155, 'samples': 1630272, 'steps': 8490, 'loss/train': 1.8303735852241516} 01/28/2022 23:45:37 - INFO - codeparrot_training - Step 8491: {'lr': 0.00047777669284739064, 'samples': 1630464, 'steps': 8491, 'loss/train': 2.3527432680130005} 01/28/2022 23:45:41 - INFO - codeparrot_training - Step 8492: {'lr': 0.0004777699482368867, 'samples': 1630656, 'steps': 8492, 'loss/train': 2.0249383449554443} 01/28/2022 23:45:45 - INFO - codeparrot_training - Step 8493: {'lr': 0.0004777632026506886, 'samples': 1630848, 'steps': 8493, 'loss/train': 1.5152819752693176} 01/28/2022 23:45:52 - INFO - codeparrot_training - Step 8494: {'lr': 0.0004777564560888252, 'samples': 1631040, 'steps': 8494, 'loss/train': 2.2012377977371216} 01/28/2022 23:45:56 - INFO - codeparrot_training - Step 8495: {'lr': 0.0004777497085513256, 'samples': 1631232, 'steps': 8495, 'loss/train': 2.499211013317108} 01/28/2022 23:46:00 - INFO - codeparrot_training - Step 8496: {'lr': 0.0004777429600382185, 'samples': 1631424, 'steps': 8496, 'loss/train': 0.9980524778366089} 01/28/2022 23:46:04 - INFO - codeparrot_training - Step 8497: {'lr': 0.00047773621054953287, 'samples': 1631616, 'steps': 8497, 'loss/train': 1.7510653138160706} 01/28/2022 23:46:09 - INFO - codeparrot_training - Step 8498: {'lr': 0.0004777294600852976, 'samples': 1631808, 'steps': 8498, 'loss/train': 1.3055810630321503} 01/28/2022 23:46:14 - INFO - codeparrot_training - Step 8499: {'lr': 0.0004777227086455417, 'samples': 1632000, 'steps': 8499, 'loss/train': 1.3916918635368347} 01/28/2022 23:46:18 - INFO - codeparrot_training - Step 8500: {'lr': 0.000477715956230294, 'samples': 1632192, 'steps': 8500, 'loss/train': 1.6860310435295105} 01/28/2022 23:46:22 - INFO - codeparrot_training - Step 8501: {'lr': 0.0004777092028395834, 'samples': 1632384, 'steps': 8501, 'loss/train': 2.4123440980911255} 01/28/2022 23:46:27 - INFO - codeparrot_training - Step 8502: {'lr': 0.00047770244847343893, 'samples': 1632576, 'steps': 8502, 'loss/train': 2.5125104784965515} 01/28/2022 23:46:31 - INFO - codeparrot_training - Step 8503: {'lr': 0.0004776956931318895, 'samples': 1632768, 'steps': 8503, 'loss/train': 1.635115385055542} 01/28/2022 23:46:36 - INFO - codeparrot_training - Step 8504: {'lr': 0.00047768893681496397, 'samples': 1632960, 'steps': 8504, 'loss/train': 1.1004793345928192} 01/28/2022 23:46:41 - INFO - codeparrot_training - Step 8505: {'lr': 0.0004776821795226913, 'samples': 1633152, 'steps': 8505, 'loss/train': 2.135480761528015} 01/28/2022 23:46:45 - INFO - codeparrot_training - Step 8506: {'lr': 0.0004776754212551006, 'samples': 1633344, 'steps': 8506, 'loss/train': 2.0535934567451477} 01/28/2022 23:46:49 - INFO - codeparrot_training - Step 8507: {'lr': 0.0004776686620122206, 'samples': 1633536, 'steps': 8507, 'loss/train': 1.3382439315319061} 01/28/2022 23:46:53 - INFO - codeparrot_training - Step 8508: {'lr': 0.00047766190179408043, 'samples': 1633728, 'steps': 8508, 'loss/train': 2.2628434896469116} 01/28/2022 23:47:00 - INFO - codeparrot_training - Step 8509: {'lr': 0.00047765514060070887, 'samples': 1633920, 'steps': 8509, 'loss/train': 1.2264068126678467} 01/28/2022 23:47:04 - INFO - codeparrot_training - Step 8510: {'lr': 0.00047764837843213497, 'samples': 1634112, 'steps': 8510, 'loss/train': 2.0770137906074524} 01/28/2022 23:47:08 - INFO - codeparrot_training - Step 8511: {'lr': 0.0004776416152883878, 'samples': 1634304, 'steps': 8511, 'loss/train': 1.7878696918487549} 01/28/2022 23:47:12 - INFO - codeparrot_training - Step 8512: {'lr': 0.0004776348511694961, 'samples': 1634496, 'steps': 8512, 'loss/train': 0.9759897887706757} 01/28/2022 23:47:17 - INFO - codeparrot_training - Step 8513: {'lr': 0.0004776280860754891, 'samples': 1634688, 'steps': 8513, 'loss/train': 1.9171258807182312} 01/28/2022 23:47:22 - INFO - codeparrot_training - Step 8514: {'lr': 0.0004776213200063956, 'samples': 1634880, 'steps': 8514, 'loss/train': 0.44239693880081177} 01/28/2022 23:47:26 - INFO - codeparrot_training - Step 8515: {'lr': 0.00047761455296224464, 'samples': 1635072, 'steps': 8515, 'loss/train': 1.4687160551548004} 01/28/2022 23:47:31 - INFO - codeparrot_training - Step 8516: {'lr': 0.0004776077849430652, 'samples': 1635264, 'steps': 8516, 'loss/train': 1.5666364431381226} 01/28/2022 23:47:35 - INFO - codeparrot_training - Step 8517: {'lr': 0.00047760101594888633, 'samples': 1635456, 'steps': 8517, 'loss/train': 1.5744942426681519} 01/28/2022 23:47:39 - INFO - codeparrot_training - Step 8518: {'lr': 0.000477594245979737, 'samples': 1635648, 'steps': 8518, 'loss/train': 2.4899822473526} 01/28/2022 23:47:45 - INFO - codeparrot_training - Step 8519: {'lr': 0.0004775874750356461, 'samples': 1635840, 'steps': 8519, 'loss/train': 1.7767932415008545} 01/28/2022 23:47:49 - INFO - codeparrot_training - Step 8520: {'lr': 0.00047758070311664283, 'samples': 1636032, 'steps': 8520, 'loss/train': 1.7673518657684326} 01/28/2022 23:47:53 - INFO - codeparrot_training - Step 8521: {'lr': 0.000477573930222756, 'samples': 1636224, 'steps': 8521, 'loss/train': 1.1009739339351654} 01/28/2022 23:47:57 - INFO - codeparrot_training - Step 8522: {'lr': 0.0004775671563540147, 'samples': 1636416, 'steps': 8522, 'loss/train': 2.022703170776367} 01/28/2022 23:48:01 - INFO - codeparrot_training - Step 8523: {'lr': 0.000477560381510448, 'samples': 1636608, 'steps': 8523, 'loss/train': 2.0226927995681763} 01/28/2022 23:48:08 - INFO - codeparrot_training - Step 8524: {'lr': 0.00047755360569208495, 'samples': 1636800, 'steps': 8524, 'loss/train': 0.36171868443489075} 01/28/2022 23:48:12 - INFO - codeparrot_training - Step 8525: {'lr': 0.00047754682889895444, 'samples': 1636992, 'steps': 8525, 'loss/train': 1.6532047390937805} 01/28/2022 23:48:17 - INFO - codeparrot_training - Step 8526: {'lr': 0.00047754005113108557, 'samples': 1637184, 'steps': 8526, 'loss/train': 1.4939721822738647} 01/28/2022 23:48:21 - INFO - codeparrot_training - Step 8527: {'lr': 0.0004775332723885074, 'samples': 1637376, 'steps': 8527, 'loss/train': 1.9951704740524292} 01/28/2022 23:48:25 - INFO - codeparrot_training - Step 8528: {'lr': 0.00047752649267124894, 'samples': 1637568, 'steps': 8528, 'loss/train': 1.8812561631202698} 01/28/2022 23:48:31 - INFO - codeparrot_training - Step 8529: {'lr': 0.0004775197119793392, 'samples': 1637760, 'steps': 8529, 'loss/train': 1.780267596244812} 01/28/2022 23:48:35 - INFO - codeparrot_training - Step 8530: {'lr': 0.0004775129303128073, 'samples': 1637952, 'steps': 8530, 'loss/train': 2.265652298927307} 01/28/2022 23:48:39 - INFO - codeparrot_training - Step 8531: {'lr': 0.0004775061476716822, 'samples': 1638144, 'steps': 8531, 'loss/train': 1.1353626251220703} 01/28/2022 23:48:43 - INFO - codeparrot_training - Step 8532: {'lr': 0.000477499364055993, 'samples': 1638336, 'steps': 8532, 'loss/train': 1.7196316123008728} 01/28/2022 23:48:48 - INFO - codeparrot_training - Step 8533: {'lr': 0.00047749257946576887, 'samples': 1638528, 'steps': 8533, 'loss/train': 1.2777956128120422} 01/28/2022 23:48:53 - INFO - codeparrot_training - Step 8534: {'lr': 0.0004774857939010387, 'samples': 1638720, 'steps': 8534, 'loss/train': 1.788597822189331} 01/28/2022 23:48:57 - INFO - codeparrot_training - Step 8535: {'lr': 0.0004774790073618316, 'samples': 1638912, 'steps': 8535, 'loss/train': 0.9430022835731506} 01/28/2022 23:49:01 - INFO - codeparrot_training - Step 8536: {'lr': 0.00047747221984817666, 'samples': 1639104, 'steps': 8536, 'loss/train': 2.3155514001846313} 01/28/2022 23:49:06 - INFO - codeparrot_training - Step 8537: {'lr': 0.000477465431360103, 'samples': 1639296, 'steps': 8537, 'loss/train': 0.96853968501091} 01/28/2022 23:49:10 - INFO - codeparrot_training - Step 8538: {'lr': 0.00047745864189763964, 'samples': 1639488, 'steps': 8538, 'loss/train': 1.957313060760498} 01/28/2022 23:49:16 - INFO - codeparrot_training - Step 8539: {'lr': 0.0004774518514608157, 'samples': 1639680, 'steps': 8539, 'loss/train': 1.282094120979309} 01/28/2022 23:49:20 - INFO - codeparrot_training - Step 8540: {'lr': 0.00047744506004966024, 'samples': 1639872, 'steps': 8540, 'loss/train': 1.786019504070282} 01/28/2022 23:49:25 - INFO - codeparrot_training - Step 8541: {'lr': 0.0004774382676642024, 'samples': 1640064, 'steps': 8541, 'loss/train': 1.9947790503501892} 01/28/2022 23:49:29 - INFO - codeparrot_training - Step 8542: {'lr': 0.0004774314743044712, 'samples': 1640256, 'steps': 8542, 'loss/train': 0.8746558427810669} 01/28/2022 23:49:33 - INFO - codeparrot_training - Step 8543: {'lr': 0.00047742467997049576, 'samples': 1640448, 'steps': 8543, 'loss/train': 1.2809733152389526} 01/28/2022 23:49:39 - INFO - codeparrot_training - Step 8544: {'lr': 0.00047741788466230527, 'samples': 1640640, 'steps': 8544, 'loss/train': 2.3836312294006348} 01/28/2022 23:49:43 - INFO - codeparrot_training - Step 8545: {'lr': 0.00047741108837992877, 'samples': 1640832, 'steps': 8545, 'loss/train': 1.6132264137268066} 01/28/2022 23:49:48 - INFO - codeparrot_training - Step 8546: {'lr': 0.0004774042911233953, 'samples': 1641024, 'steps': 8546, 'loss/train': 2.189418375492096} 01/28/2022 23:49:52 - INFO - codeparrot_training - Step 8547: {'lr': 0.0004773974928927342, 'samples': 1641216, 'steps': 8547, 'loss/train': 2.0565746426582336} 01/28/2022 23:49:56 - INFO - codeparrot_training - Step 8548: {'lr': 0.00047739069368797426, 'samples': 1641408, 'steps': 8548, 'loss/train': 2.149860978126526} 01/28/2022 23:50:00 - INFO - codeparrot_training - Step 8549: {'lr': 0.0004773838935091449, 'samples': 1641600, 'steps': 8549, 'loss/train': 1.632987141609192} 01/28/2022 23:50:07 - INFO - codeparrot_training - Step 8550: {'lr': 0.00047737709235627515, 'samples': 1641792, 'steps': 8550, 'loss/train': 2.9891399145126343} 01/28/2022 23:50:11 - INFO - codeparrot_training - Step 8551: {'lr': 0.00047737029022939414, 'samples': 1641984, 'steps': 8551, 'loss/train': 1.8073149919509888} 01/28/2022 23:50:15 - INFO - codeparrot_training - Step 8552: {'lr': 0.00047736348712853094, 'samples': 1642176, 'steps': 8552, 'loss/train': 1.4273951947689056} 01/28/2022 23:50:19 - INFO - codeparrot_training - Step 8553: {'lr': 0.00047735668305371484, 'samples': 1642368, 'steps': 8553, 'loss/train': 0.8215029537677765} 01/28/2022 23:50:24 - INFO - codeparrot_training - Step 8554: {'lr': 0.0004773498780049749, 'samples': 1642560, 'steps': 8554, 'loss/train': 2.361010730266571} 01/28/2022 23:50:29 - INFO - codeparrot_training - Step 8555: {'lr': 0.00047734307198234015, 'samples': 1642752, 'steps': 8555, 'loss/train': 1.3331718742847443} 01/28/2022 23:50:33 - INFO - codeparrot_training - Step 8556: {'lr': 0.00047733626498584, 'samples': 1642944, 'steps': 8556, 'loss/train': 1.6810757517814636} 01/28/2022 23:50:37 - INFO - codeparrot_training - Step 8557: {'lr': 0.0004773294570155035, 'samples': 1643136, 'steps': 8557, 'loss/train': 1.4069682955741882} 01/28/2022 23:50:42 - INFO - codeparrot_training - Step 8558: {'lr': 0.0004773226480713596, 'samples': 1643328, 'steps': 8558, 'loss/train': 1.6890199184417725} 01/28/2022 23:50:46 - INFO - codeparrot_training - Step 8559: {'lr': 0.00047731583815343784, 'samples': 1643520, 'steps': 8559, 'loss/train': 1.3930284976959229} 01/28/2022 23:50:51 - INFO - codeparrot_training - Step 8560: {'lr': 0.00047730902726176715, 'samples': 1643712, 'steps': 8560, 'loss/train': 1.8335398435592651} 01/28/2022 23:50:55 - INFO - codeparrot_training - Step 8561: {'lr': 0.00047730221539637677, 'samples': 1643904, 'steps': 8561, 'loss/train': 2.5961714386940002} 01/28/2022 23:51:00 - INFO - codeparrot_training - Step 8562: {'lr': 0.00047729540255729585, 'samples': 1644096, 'steps': 8562, 'loss/train': 1.645579218864441} 01/28/2022 23:51:04 - INFO - codeparrot_training - Step 8563: {'lr': 0.0004772885887445536, 'samples': 1644288, 'steps': 8563, 'loss/train': 1.4136682748794556} 01/28/2022 23:51:08 - INFO - codeparrot_training - Step 8564: {'lr': 0.0004772817739581793, 'samples': 1644480, 'steps': 8564, 'loss/train': 1.8637479543685913} 01/28/2022 23:51:14 - INFO - codeparrot_training - Step 8565: {'lr': 0.000477274958198202, 'samples': 1644672, 'steps': 8565, 'loss/train': 1.5729210376739502} 01/28/2022 23:51:18 - INFO - codeparrot_training - Step 8566: {'lr': 0.0004772681414646509, 'samples': 1644864, 'steps': 8566, 'loss/train': 1.056260347366333} 01/28/2022 23:51:22 - INFO - codeparrot_training - Step 8567: {'lr': 0.00047726132375755525, 'samples': 1645056, 'steps': 8567, 'loss/train': 2.157909572124481} 01/28/2022 23:51:27 - INFO - codeparrot_training - Step 8568: {'lr': 0.00047725450507694433, 'samples': 1645248, 'steps': 8568, 'loss/train': 2.465226709842682} 01/28/2022 23:51:31 - INFO - codeparrot_training - Step 8569: {'lr': 0.00047724768542284726, 'samples': 1645440, 'steps': 8569, 'loss/train': 2.1700281500816345} 01/28/2022 23:51:37 - INFO - codeparrot_training - Step 8570: {'lr': 0.0004772408647952932, 'samples': 1645632, 'steps': 8570, 'loss/train': 2.8225284218788147} 01/28/2022 23:51:42 - INFO - codeparrot_training - Step 8571: {'lr': 0.0004772340431943114, 'samples': 1645824, 'steps': 8571, 'loss/train': 1.6879574060440063} 01/28/2022 23:51:46 - INFO - codeparrot_training - Step 8572: {'lr': 0.0004772272206199312, 'samples': 1646016, 'steps': 8572, 'loss/train': 2.67165344953537} 01/28/2022 23:51:50 - INFO - codeparrot_training - Step 8573: {'lr': 0.0004772203970721817, 'samples': 1646208, 'steps': 8573, 'loss/train': 1.9372501373291016} 01/28/2022 23:51:54 - INFO - codeparrot_training - Step 8574: {'lr': 0.0004772135725510922, 'samples': 1646400, 'steps': 8574, 'loss/train': 1.8047221899032593} 01/28/2022 23:52:00 - INFO - codeparrot_training - Step 8575: {'lr': 0.0004772067470566919, 'samples': 1646592, 'steps': 8575, 'loss/train': 1.4191653728485107} 01/28/2022 23:52:04 - INFO - codeparrot_training - Step 8576: {'lr': 0.00047719992058901006, 'samples': 1646784, 'steps': 8576, 'loss/train': 2.5249128341674805} 01/28/2022 23:52:08 - INFO - codeparrot_training - Step 8577: {'lr': 0.00047719309314807584, 'samples': 1646976, 'steps': 8577, 'loss/train': 1.736538290977478} 01/28/2022 23:52:12 - INFO - codeparrot_training - Step 8578: {'lr': 0.0004771862647339186, 'samples': 1647168, 'steps': 8578, 'loss/train': 1.6985791325569153} 01/28/2022 23:52:17 - INFO - codeparrot_training - Step 8579: {'lr': 0.0004771794353465675, 'samples': 1647360, 'steps': 8579, 'loss/train': 0.7105893641710281} 01/28/2022 23:52:22 - INFO - codeparrot_training - Step 8580: {'lr': 0.00047717260498605186, 'samples': 1647552, 'steps': 8580, 'loss/train': 0.7137841433286667} 01/28/2022 23:52:26 - INFO - codeparrot_training - Step 8581: {'lr': 0.0004771657736524009, 'samples': 1647744, 'steps': 8581, 'loss/train': 2.5139304399490356} 01/28/2022 23:52:30 - INFO - codeparrot_training - Step 8582: {'lr': 0.00047715894134564395, 'samples': 1647936, 'steps': 8582, 'loss/train': 1.3798456192016602} 01/28/2022 23:52:35 - INFO - codeparrot_training - Step 8583: {'lr': 0.0004771521080658102, 'samples': 1648128, 'steps': 8583, 'loss/train': 1.7414857149124146} 01/28/2022 23:52:39 - INFO - codeparrot_training - Step 8584: {'lr': 0.00047714527381292893, 'samples': 1648320, 'steps': 8584, 'loss/train': 1.346385419368744} 01/28/2022 23:52:45 - INFO - codeparrot_training - Step 8585: {'lr': 0.00047713843858702943, 'samples': 1648512, 'steps': 8585, 'loss/train': 0.8127720952033997} 01/28/2022 23:52:50 - INFO - codeparrot_training - Step 8586: {'lr': 0.000477131602388141, 'samples': 1648704, 'steps': 8586, 'loss/train': 1.5661060810089111} 01/28/2022 23:52:54 - INFO - codeparrot_training - Step 8587: {'lr': 0.00047712476521629294, 'samples': 1648896, 'steps': 8587, 'loss/train': 1.9332520365715027} 01/28/2022 23:52:58 - INFO - codeparrot_training - Step 8588: {'lr': 0.0004771179270715145, 'samples': 1649088, 'steps': 8588, 'loss/train': 1.89623361825943} 01/28/2022 23:53:02 - INFO - codeparrot_training - Step 8589: {'lr': 0.000477111087953835, 'samples': 1649280, 'steps': 8589, 'loss/train': 1.8358824849128723} 01/28/2022 23:53:09 - INFO - codeparrot_training - Step 8590: {'lr': 0.0004771042478632836, 'samples': 1649472, 'steps': 8590, 'loss/train': 1.8030495643615723} 01/28/2022 23:53:13 - INFO - codeparrot_training - Step 8591: {'lr': 0.0004770974067998898, 'samples': 1649664, 'steps': 8591, 'loss/train': 0.05420493520796299} 01/28/2022 23:53:17 - INFO - codeparrot_training - Step 8592: {'lr': 0.0004770905647636828, 'samples': 1649856, 'steps': 8592, 'loss/train': 1.5404364466667175} 01/28/2022 23:53:22 - INFO - codeparrot_training - Step 8593: {'lr': 0.00047708372175469193, 'samples': 1650048, 'steps': 8593, 'loss/train': 1.2742089629173279} 01/28/2022 23:53:26 - INFO - codeparrot_training - Step 8594: {'lr': 0.0004770768777729465, 'samples': 1650240, 'steps': 8594, 'loss/train': 1.9184015393257141} 01/28/2022 23:53:30 - INFO - codeparrot_training - Step 8595: {'lr': 0.0004770700328184758, 'samples': 1650432, 'steps': 8595, 'loss/train': 1.6467604637145996} 01/28/2022 23:53:37 - INFO - codeparrot_training - Step 8596: {'lr': 0.00047706318689130924, 'samples': 1650624, 'steps': 8596, 'loss/train': 1.4789958000183105} 01/28/2022 23:53:41 - INFO - codeparrot_training - Step 8597: {'lr': 0.0004770563399914761, 'samples': 1650816, 'steps': 8597, 'loss/train': 1.1213649809360504} 01/28/2022 23:53:45 - INFO - codeparrot_training - Step 8598: {'lr': 0.00047704949211900565, 'samples': 1651008, 'steps': 8598, 'loss/train': 2.3558841347694397} 01/28/2022 23:53:49 - INFO - codeparrot_training - Step 8599: {'lr': 0.0004770426432739273, 'samples': 1651200, 'steps': 8599, 'loss/train': 3.4263389110565186} 01/28/2022 23:53:54 - INFO - codeparrot_training - Step 8600: {'lr': 0.00047703579345627036, 'samples': 1651392, 'steps': 8600, 'loss/train': 2.415001630783081} 01/28/2022 23:53:59 - INFO - codeparrot_training - Step 8601: {'lr': 0.00047702894266606413, 'samples': 1651584, 'steps': 8601, 'loss/train': 1.6911917924880981} 01/28/2022 23:54:03 - INFO - codeparrot_training - Step 8602: {'lr': 0.00047702209090333804, 'samples': 1651776, 'steps': 8602, 'loss/train': 1.9348036050796509} 01/28/2022 23:54:07 - INFO - codeparrot_training - Step 8603: {'lr': 0.0004770152381681214, 'samples': 1651968, 'steps': 8603, 'loss/train': 1.6179994940757751} 01/28/2022 23:54:12 - INFO - codeparrot_training - Step 8604: {'lr': 0.0004770083844604435, 'samples': 1652160, 'steps': 8604, 'loss/train': 1.7106817960739136} 01/28/2022 23:54:16 - INFO - codeparrot_training - Step 8605: {'lr': 0.00047700152978033387, 'samples': 1652352, 'steps': 8605, 'loss/train': 1.9218392372131348} 01/28/2022 23:54:23 - INFO - codeparrot_training - Step 8606: {'lr': 0.0004769946741278217, 'samples': 1652544, 'steps': 8606, 'loss/train': 1.6420192122459412} 01/28/2022 23:54:28 - INFO - codeparrot_training - Step 8607: {'lr': 0.00047698781750293644, 'samples': 1652736, 'steps': 8607, 'loss/train': 0.6078401505947113} 01/28/2022 23:54:32 - INFO - codeparrot_training - Step 8608: {'lr': 0.00047698095990570744, 'samples': 1652928, 'steps': 8608, 'loss/train': 2.1221646666526794} 01/28/2022 23:54:36 - INFO - codeparrot_training - Step 8609: {'lr': 0.00047697410133616414, 'samples': 1653120, 'steps': 8609, 'loss/train': 1.7866061925888062} 01/28/2022 23:54:40 - INFO - codeparrot_training - Step 8610: {'lr': 0.0004769672417943358, 'samples': 1653312, 'steps': 8610, 'loss/train': 1.9247469305992126} 01/28/2022 23:54:46 - INFO - codeparrot_training - Step 8611: {'lr': 0.00047696038128025185, 'samples': 1653504, 'steps': 8611, 'loss/train': 1.8070172667503357} 01/28/2022 23:54:50 - INFO - codeparrot_training - Step 8612: {'lr': 0.00047695351979394173, 'samples': 1653696, 'steps': 8612, 'loss/train': 2.5719741582870483} 01/28/2022 23:54:54 - INFO - codeparrot_training - Step 8613: {'lr': 0.00047694665733543485, 'samples': 1653888, 'steps': 8613, 'loss/train': 1.929156482219696} 01/28/2022 23:54:58 - INFO - codeparrot_training - Step 8614: {'lr': 0.00047693979390476046, 'samples': 1654080, 'steps': 8614, 'loss/train': 2.3127790689468384} 01/28/2022 23:55:03 - INFO - codeparrot_training - Step 8615: {'lr': 0.00047693292950194813, 'samples': 1654272, 'steps': 8615, 'loss/train': 1.6037564277648926} 01/28/2022 23:55:08 - INFO - codeparrot_training - Step 8616: {'lr': 0.0004769260641270271, 'samples': 1654464, 'steps': 8616, 'loss/train': 3.0485962629318237} 01/28/2022 23:55:12 - INFO - codeparrot_training - Step 8617: {'lr': 0.0004769191977800269, 'samples': 1654656, 'steps': 8617, 'loss/train': 1.9416736364364624} 01/28/2022 23:55:17 - INFO - codeparrot_training - Step 8618: {'lr': 0.0004769123304609769, 'samples': 1654848, 'steps': 8618, 'loss/train': 0.9054560959339142} 01/28/2022 23:55:21 - INFO - codeparrot_training - Step 8619: {'lr': 0.0004769054621699066, 'samples': 1655040, 'steps': 8619, 'loss/train': 0.9333488345146179} 01/28/2022 23:55:25 - INFO - codeparrot_training - Step 8620: {'lr': 0.0004768985929068453, 'samples': 1655232, 'steps': 8620, 'loss/train': 2.0342021584510803} 01/28/2022 23:55:32 - INFO - codeparrot_training - Step 8621: {'lr': 0.0004768917226718225, 'samples': 1655424, 'steps': 8621, 'loss/train': 2.0486100912094116} 01/28/2022 23:55:37 - INFO - codeparrot_training - Step 8622: {'lr': 0.0004768848514648676, 'samples': 1655616, 'steps': 8622, 'loss/train': 2.0284118056297302} 01/28/2022 23:55:41 - INFO - codeparrot_training - Step 8623: {'lr': 0.0004768779792860101, 'samples': 1655808, 'steps': 8623, 'loss/train': 1.0541521310806274} 01/28/2022 23:55:45 - INFO - codeparrot_training - Step 8624: {'lr': 0.00047687110613527924, 'samples': 1656000, 'steps': 8624, 'loss/train': 2.7435944080352783} 01/28/2022 23:55:49 - INFO - codeparrot_training - Step 8625: {'lr': 0.0004768642320127047, 'samples': 1656192, 'steps': 8625, 'loss/train': 2.8600956201553345} 01/28/2022 23:55:55 - INFO - codeparrot_training - Step 8626: {'lr': 0.0004768573569183158, 'samples': 1656384, 'steps': 8626, 'loss/train': 1.646634042263031} 01/28/2022 23:55:59 - INFO - codeparrot_training - Step 8627: {'lr': 0.000476850480852142, 'samples': 1656576, 'steps': 8627, 'loss/train': 1.6080601215362549} 01/28/2022 23:56:03 - INFO - codeparrot_training - Step 8628: {'lr': 0.0004768436038142128, 'samples': 1656768, 'steps': 8628, 'loss/train': 2.38661527633667} 01/28/2022 23:56:07 - INFO - codeparrot_training - Step 8629: {'lr': 0.00047683672580455764, 'samples': 1656960, 'steps': 8629, 'loss/train': 2.0751675367355347} 01/28/2022 23:56:12 - INFO - codeparrot_training - Step 8630: {'lr': 0.00047682984682320597, 'samples': 1657152, 'steps': 8630, 'loss/train': 2.9605844020843506} 01/28/2022 23:56:17 - INFO - codeparrot_training - Step 8631: {'lr': 0.0004768229668701872, 'samples': 1657344, 'steps': 8631, 'loss/train': 0.8860856294631958} 01/28/2022 23:56:21 - INFO - codeparrot_training - Step 8632: {'lr': 0.00047681608594553093, 'samples': 1657536, 'steps': 8632, 'loss/train': 1.5601208209991455} 01/28/2022 23:56:25 - INFO - codeparrot_training - Step 8633: {'lr': 0.00047680920404926655, 'samples': 1657728, 'steps': 8633, 'loss/train': 3.1279306411743164} 01/28/2022 23:56:30 - INFO - codeparrot_training - Step 8634: {'lr': 0.0004768023211814236, 'samples': 1657920, 'steps': 8634, 'loss/train': 1.8861674666404724} 01/28/2022 23:56:34 - INFO - codeparrot_training - Step 8635: {'lr': 0.0004767954373420315, 'samples': 1658112, 'steps': 8635, 'loss/train': 2.3267499804496765} 01/28/2022 23:56:41 - INFO - codeparrot_training - Step 8636: {'lr': 0.0004767885525311197, 'samples': 1658304, 'steps': 8636, 'loss/train': 2.085799992084503} 01/28/2022 23:56:45 - INFO - codeparrot_training - Step 8637: {'lr': 0.00047678166674871783, 'samples': 1658496, 'steps': 8637, 'loss/train': 2.0182530283927917} 01/28/2022 23:56:49 - INFO - codeparrot_training - Step 8638: {'lr': 0.0004767747799948553, 'samples': 1658688, 'steps': 8638, 'loss/train': 1.662813663482666} 01/28/2022 23:56:54 - INFO - codeparrot_training - Step 8639: {'lr': 0.0004767678922695616, 'samples': 1658880, 'steps': 8639, 'loss/train': 1.6807531714439392} 01/28/2022 23:56:58 - INFO - codeparrot_training - Step 8640: {'lr': 0.0004767610035728662, 'samples': 1659072, 'steps': 8640, 'loss/train': 1.8675999641418457} 01/28/2022 23:57:03 - INFO - codeparrot_training - Step 8641: {'lr': 0.00047675411390479876, 'samples': 1659264, 'steps': 8641, 'loss/train': 1.9053109288215637} 01/28/2022 23:57:07 - INFO - codeparrot_training - Step 8642: {'lr': 0.0004767472232653887, 'samples': 1659456, 'steps': 8642, 'loss/train': 2.303641676902771} 01/28/2022 23:57:12 - INFO - codeparrot_training - Step 8643: {'lr': 0.00047674033165466545, 'samples': 1659648, 'steps': 8643, 'loss/train': 2.089065670967102} 01/28/2022 23:57:16 - INFO - codeparrot_training - Step 8644: {'lr': 0.0004767334390726588, 'samples': 1659840, 'steps': 8644, 'loss/train': 1.6438824534416199} 01/28/2022 23:57:20 - INFO - codeparrot_training - Step 8645: {'lr': 0.00047672654551939785, 'samples': 1660032, 'steps': 8645, 'loss/train': 1.6238908767700195} 01/28/2022 23:57:27 - INFO - codeparrot_training - Step 8646: {'lr': 0.00047671965099491256, 'samples': 1660224, 'steps': 8646, 'loss/train': 1.3598805963993073} 01/28/2022 23:57:31 - INFO - codeparrot_training - Step 8647: {'lr': 0.0004767127554992322, 'samples': 1660416, 'steps': 8647, 'loss/train': 1.8513818979263306} 01/28/2022 23:57:36 - INFO - codeparrot_training - Step 8648: {'lr': 0.0004767058590323864, 'samples': 1660608, 'steps': 8648, 'loss/train': 2.207431733608246} 01/28/2022 23:57:40 - INFO - codeparrot_training - Step 8649: {'lr': 0.00047669896159440464, 'samples': 1660800, 'steps': 8649, 'loss/train': 2.034680664539337} 01/28/2022 23:57:44 - INFO - codeparrot_training - Step 8650: {'lr': 0.00047669206318531654, 'samples': 1660992, 'steps': 8650, 'loss/train': 1.663465440273285} 01/28/2022 23:57:50 - INFO - codeparrot_training - Step 8651: {'lr': 0.00047668516380515165, 'samples': 1661184, 'steps': 8651, 'loss/train': 0.9582851529121399} 01/28/2022 23:57:54 - INFO - codeparrot_training - Step 8652: {'lr': 0.0004766782634539395, 'samples': 1661376, 'steps': 8652, 'loss/train': 2.100766360759735} 01/28/2022 23:57:58 - INFO - codeparrot_training - Step 8653: {'lr': 0.00047667136213170957, 'samples': 1661568, 'steps': 8653, 'loss/train': 1.9471229910850525} 01/28/2022 23:58:02 - INFO - codeparrot_training - Step 8654: {'lr': 0.00047666445983849163, 'samples': 1661760, 'steps': 8654, 'loss/train': 2.244483232498169} 01/28/2022 23:58:06 - INFO - codeparrot_training - Step 8655: {'lr': 0.000476657556574315, 'samples': 1661952, 'steps': 8655, 'loss/train': 2.1397789120674133} 01/28/2022 23:58:12 - INFO - codeparrot_training - Step 8656: {'lr': 0.00047665065233920946, 'samples': 1662144, 'steps': 8656, 'loss/train': 2.3500646352767944} 01/28/2022 23:58:16 - INFO - codeparrot_training - Step 8657: {'lr': 0.0004766437471332045, 'samples': 1662336, 'steps': 8657, 'loss/train': 2.2829039096832275} 01/28/2022 23:58:20 - INFO - codeparrot_training - Step 8658: {'lr': 0.0004766368409563296, 'samples': 1662528, 'steps': 8658, 'loss/train': 1.5669286251068115} 01/28/2022 23:58:25 - INFO - codeparrot_training - Step 8659: {'lr': 0.0004766299338086145, 'samples': 1662720, 'steps': 8659, 'loss/train': 1.7537629008293152} 01/28/2022 23:58:29 - INFO - codeparrot_training - Step 8660: {'lr': 0.0004766230256900887, 'samples': 1662912, 'steps': 8660, 'loss/train': 1.4032718539237976} 01/28/2022 23:58:34 - INFO - codeparrot_training - Step 8661: {'lr': 0.00047661611660078184, 'samples': 1663104, 'steps': 8661, 'loss/train': 2.4985011219978333} 01/28/2022 23:58:38 - INFO - codeparrot_training - Step 8662: {'lr': 0.0004766092065407235, 'samples': 1663296, 'steps': 8662, 'loss/train': 2.36860191822052} 01/28/2022 23:58:43 - INFO - codeparrot_training - Step 8663: {'lr': 0.0004766022955099433, 'samples': 1663488, 'steps': 8663, 'loss/train': 2.3284127712249756} 01/28/2022 23:58:47 - INFO - codeparrot_training - Step 8664: {'lr': 0.00047659538350847076, 'samples': 1663680, 'steps': 8664, 'loss/train': 1.4259631633758545} 01/28/2022 23:58:51 - INFO - codeparrot_training - Step 8665: {'lr': 0.00047658847053633555, 'samples': 1663872, 'steps': 8665, 'loss/train': 1.5340272188186646} 01/28/2022 23:58:59 - INFO - codeparrot_training - Step 8666: {'lr': 0.00047658155659356725, 'samples': 1664064, 'steps': 8666, 'loss/train': 2.051609516143799} 01/28/2022 23:59:03 - INFO - codeparrot_training - Step 8667: {'lr': 0.0004765746416801956, 'samples': 1664256, 'steps': 8667, 'loss/train': 1.9870237112045288} 01/28/2022 23:59:07 - INFO - codeparrot_training - Step 8668: {'lr': 0.0004765677257962501, 'samples': 1664448, 'steps': 8668, 'loss/train': 2.5644103288650513} 01/28/2022 23:59:12 - INFO - codeparrot_training - Step 8669: {'lr': 0.0004765608089417604, 'samples': 1664640, 'steps': 8669, 'loss/train': 2.2778225541114807} 01/28/2022 23:59:16 - INFO - codeparrot_training - Step 8670: {'lr': 0.0004765538911167562, 'samples': 1664832, 'steps': 8670, 'loss/train': 0.7458811104297638} 01/28/2022 23:59:21 - INFO - codeparrot_training - Step 8671: {'lr': 0.00047654697232126696, 'samples': 1665024, 'steps': 8671, 'loss/train': 2.121347665786743} 01/28/2022 23:59:25 - INFO - codeparrot_training - Step 8672: {'lr': 0.00047654005255532247, 'samples': 1665216, 'steps': 8672, 'loss/train': 2.0761678218841553} 01/28/2022 23:59:30 - INFO - codeparrot_training - Step 8673: {'lr': 0.0004765331318189523, 'samples': 1665408, 'steps': 8673, 'loss/train': 2.2579761743545532} 01/28/2022 23:59:34 - INFO - codeparrot_training - Step 8674: {'lr': 0.00047652621011218623, 'samples': 1665600, 'steps': 8674, 'loss/train': 1.3089662790298462} 01/28/2022 23:59:38 - INFO - codeparrot_training - Step 8675: {'lr': 0.0004765192874350537, 'samples': 1665792, 'steps': 8675, 'loss/train': 1.2550534307956696} 01/28/2022 23:59:43 - INFO - codeparrot_training - Step 8676: {'lr': 0.0004765123637875845, 'samples': 1665984, 'steps': 8676, 'loss/train': 2.0823728442192078} 01/28/2022 23:59:48 - INFO - codeparrot_training - Step 8677: {'lr': 0.00047650543916980827, 'samples': 1666176, 'steps': 8677, 'loss/train': 1.4547397792339325} 01/28/2022 23:59:52 - INFO - codeparrot_training - Step 8678: {'lr': 0.00047649851358175466, 'samples': 1666368, 'steps': 8678, 'loss/train': 2.186675012111664} 01/28/2022 23:59:56 - INFO - codeparrot_training - Step 8679: {'lr': 0.0004764915870234533, 'samples': 1666560, 'steps': 8679, 'loss/train': 2.4797117114067078} 01/29/2022 00:00:00 - INFO - codeparrot_training - Step 8680: {'lr': 0.000476484659494934, 'samples': 1666752, 'steps': 8680, 'loss/train': 1.3398757874965668} 01/29/2022 00:00:08 - INFO - codeparrot_training - Step 8681: {'lr': 0.0004764777309962263, 'samples': 1666944, 'steps': 8681, 'loss/train': 2.0947269201278687} 01/29/2022 00:00:12 - INFO - codeparrot_training - Step 8682: {'lr': 0.0004764708015273599, 'samples': 1667136, 'steps': 8682, 'loss/train': 0.8809542953968048} 01/29/2022 00:00:16 - INFO - codeparrot_training - Step 8683: {'lr': 0.0004764638710883644, 'samples': 1667328, 'steps': 8683, 'loss/train': 1.9852100014686584} 01/29/2022 00:00:20 - INFO - codeparrot_training - Step 8684: {'lr': 0.0004764569396792697, 'samples': 1667520, 'steps': 8684, 'loss/train': 3.37554931640625} 01/29/2022 00:00:25 - INFO - codeparrot_training - Step 8685: {'lr': 0.00047645000730010535, 'samples': 1667712, 'steps': 8685, 'loss/train': 2.3160899877548218} 01/29/2022 00:00:29 - INFO - codeparrot_training - Step 8686: {'lr': 0.00047644307395090107, 'samples': 1667904, 'steps': 8686, 'loss/train': 1.6138801574707031} 01/29/2022 00:00:34 - INFO - codeparrot_training - Step 8687: {'lr': 0.0004764361396316866, 'samples': 1668096, 'steps': 8687, 'loss/train': 1.1958121061325073} 01/29/2022 00:00:38 - INFO - codeparrot_training - Step 8688: {'lr': 0.0004764292043424916, 'samples': 1668288, 'steps': 8688, 'loss/train': 1.4787799715995789} 01/29/2022 00:00:42 - INFO - codeparrot_training - Step 8689: {'lr': 0.0004764222680833458, 'samples': 1668480, 'steps': 8689, 'loss/train': 2.2952622771263123} 01/29/2022 00:00:47 - INFO - codeparrot_training - Step 8690: {'lr': 0.0004764153308542788, 'samples': 1668672, 'steps': 8690, 'loss/train': 0.8666911125183105} 01/29/2022 00:00:51 - INFO - codeparrot_training - Step 8691: {'lr': 0.0004764083926553205, 'samples': 1668864, 'steps': 8691, 'loss/train': 2.2006022930145264} 01/29/2022 00:00:58 - INFO - codeparrot_training - Step 8692: {'lr': 0.00047640145348650057, 'samples': 1669056, 'steps': 8692, 'loss/train': 2.016057550907135} 01/29/2022 00:01:02 - INFO - codeparrot_training - Step 8693: {'lr': 0.0004763945133478486, 'samples': 1669248, 'steps': 8693, 'loss/train': 2.0274422764778137} 01/29/2022 00:01:06 - INFO - codeparrot_training - Step 8694: {'lr': 0.0004763875722393945, 'samples': 1669440, 'steps': 8694, 'loss/train': 2.405430257320404} 01/29/2022 00:01:10 - INFO - codeparrot_training - Step 8695: {'lr': 0.000476380630161168, 'samples': 1669632, 'steps': 8695, 'loss/train': 2.158643066883087} 01/29/2022 00:01:16 - INFO - codeparrot_training - Step 8696: {'lr': 0.00047637368711319863, 'samples': 1669824, 'steps': 8696, 'loss/train': 1.929850995540619} 01/29/2022 00:01:20 - INFO - codeparrot_training - Step 8697: {'lr': 0.00047636674309551626, 'samples': 1670016, 'steps': 8697, 'loss/train': 1.369842678308487} 01/29/2022 00:01:24 - INFO - codeparrot_training - Step 8698: {'lr': 0.0004763597981081507, 'samples': 1670208, 'steps': 8698, 'loss/train': 0.6169248819351196} 01/29/2022 00:01:28 - INFO - codeparrot_training - Step 8699: {'lr': 0.00047635285215113165, 'samples': 1670400, 'steps': 8699, 'loss/train': 2.2864381670951843} 01/29/2022 00:01:33 - INFO - codeparrot_training - Step 8700: {'lr': 0.0004763459052244888, 'samples': 1670592, 'steps': 8700, 'loss/train': 1.1710460186004639} 01/29/2022 00:01:38 - INFO - codeparrot_training - Step 8701: {'lr': 0.0004763389573282521, 'samples': 1670784, 'steps': 8701, 'loss/train': 1.309242993593216} 01/29/2022 00:01:42 - INFO - codeparrot_training - Step 8702: {'lr': 0.00047633200846245106, 'samples': 1670976, 'steps': 8702, 'loss/train': 0.9310757517814636} 01/29/2022 00:01:46 - INFO - codeparrot_training - Step 8703: {'lr': 0.0004763250586271156, 'samples': 1671168, 'steps': 8703, 'loss/train': 1.9924360513687134} 01/29/2022 00:01:51 - INFO - codeparrot_training - Step 8704: {'lr': 0.00047631810782227535, 'samples': 1671360, 'steps': 8704, 'loss/train': 0.979319840669632} 01/29/2022 00:01:55 - INFO - codeparrot_training - Step 8705: {'lr': 0.00047631115604796035, 'samples': 1671552, 'steps': 8705, 'loss/train': 2.395061194896698} 01/29/2022 00:02:02 - INFO - codeparrot_training - Step 8706: {'lr': 0.0004763042033042001, 'samples': 1671744, 'steps': 8706, 'loss/train': 1.8926101326942444} 01/29/2022 00:02:06 - INFO - codeparrot_training - Step 8707: {'lr': 0.0004762972495910246, 'samples': 1671936, 'steps': 8707, 'loss/train': 2.0762404203414917} 01/29/2022 00:02:11 - INFO - codeparrot_training - Step 8708: {'lr': 0.00047629029490846346, 'samples': 1672128, 'steps': 8708, 'loss/train': 1.8954212665557861} 01/29/2022 00:02:15 - INFO - codeparrot_training - Step 8709: {'lr': 0.0004762833392565466, 'samples': 1672320, 'steps': 8709, 'loss/train': 1.396007001399994} 01/29/2022 00:02:19 - INFO - codeparrot_training - Step 8710: {'lr': 0.00047627638263530374, 'samples': 1672512, 'steps': 8710, 'loss/train': 1.3651852011680603} 01/29/2022 00:02:24 - INFO - codeparrot_training - Step 8711: {'lr': 0.00047626942504476477, 'samples': 1672704, 'steps': 8711, 'loss/train': 1.6534423828125} 01/29/2022 00:02:29 - INFO - codeparrot_training - Step 8712: {'lr': 0.00047626246648495936, 'samples': 1672896, 'steps': 8712, 'loss/train': 3.944196581840515} 01/29/2022 00:02:33 - INFO - codeparrot_training - Step 8713: {'lr': 0.0004762555069559175, 'samples': 1673088, 'steps': 8713, 'loss/train': 1.7688603401184082} 01/29/2022 00:02:37 - INFO - codeparrot_training - Step 8714: {'lr': 0.00047624854645766875, 'samples': 1673280, 'steps': 8714, 'loss/train': 1.2135432064533234} 01/29/2022 00:02:41 - INFO - codeparrot_training - Step 8715: {'lr': 0.0004762415849902431, 'samples': 1673472, 'steps': 8715, 'loss/train': 1.873966097831726} 01/29/2022 00:02:47 - INFO - codeparrot_training - Step 8716: {'lr': 0.0004762346225536703, 'samples': 1673664, 'steps': 8716, 'loss/train': 0.9122543334960938} 01/29/2022 00:02:51 - INFO - codeparrot_training - Step 8717: {'lr': 0.0004762276591479804, 'samples': 1673856, 'steps': 8717, 'loss/train': 1.6956748366355896} 01/29/2022 00:02:55 - INFO - codeparrot_training - Step 8718: {'lr': 0.00047622069477320285, 'samples': 1674048, 'steps': 8718, 'loss/train': 1.5597424507141113} 01/29/2022 00:03:00 - INFO - codeparrot_training - Step 8719: {'lr': 0.0004762137294293678, 'samples': 1674240, 'steps': 8719, 'loss/train': 0.5600965172052383} 01/29/2022 00:03:04 - INFO - codeparrot_training - Step 8720: {'lr': 0.0004762067631165049, 'samples': 1674432, 'steps': 8720, 'loss/train': 1.7262886762619019} 01/29/2022 00:03:09 - INFO - codeparrot_training - Step 8721: {'lr': 0.0004761997958346441, 'samples': 1674624, 'steps': 8721, 'loss/train': 2.057145595550537} 01/29/2022 00:03:13 - INFO - codeparrot_training - Step 8722: {'lr': 0.00047619282758381513, 'samples': 1674816, 'steps': 8722, 'loss/train': 2.7292065024375916} 01/29/2022 00:03:18 - INFO - codeparrot_training - Step 8723: {'lr': 0.0004761858583640479, 'samples': 1675008, 'steps': 8723, 'loss/train': 2.37179034948349} 01/29/2022 00:03:22 - INFO - codeparrot_training - Step 8724: {'lr': 0.00047617888817537234, 'samples': 1675200, 'steps': 8724, 'loss/train': 1.8175397515296936} 01/29/2022 00:03:26 - INFO - codeparrot_training - Step 8725: {'lr': 0.00047617191701781824, 'samples': 1675392, 'steps': 8725, 'loss/train': 2.1591930985450745} 01/29/2022 00:03:33 - INFO - codeparrot_training - Step 8726: {'lr': 0.0004761649448914155, 'samples': 1675584, 'steps': 8726, 'loss/train': 1.9700823426246643} 01/29/2022 00:03:38 - INFO - codeparrot_training - Step 8727: {'lr': 0.0004761579717961939, 'samples': 1675776, 'steps': 8727, 'loss/train': 1.9735532999038696} 01/29/2022 00:03:42 - INFO - codeparrot_training - Step 8728: {'lr': 0.0004761509977321834, 'samples': 1675968, 'steps': 8728, 'loss/train': 1.9471290707588196} 01/29/2022 00:03:46 - INFO - codeparrot_training - Step 8729: {'lr': 0.0004761440226994138, 'samples': 1676160, 'steps': 8729, 'loss/train': 2.295344889163971} 01/29/2022 00:03:50 - INFO - codeparrot_training - Step 8730: {'lr': 0.000476137046697915, 'samples': 1676352, 'steps': 8730, 'loss/train': 1.8767237663269043} 01/29/2022 00:03:56 - INFO - codeparrot_training - Step 8731: {'lr': 0.0004761300697277169, 'samples': 1676544, 'steps': 8731, 'loss/train': 2.3638097047805786} 01/29/2022 00:04:00 - INFO - codeparrot_training - Step 8732: {'lr': 0.0004761230917888494, 'samples': 1676736, 'steps': 8732, 'loss/train': 1.6057484149932861} 01/29/2022 00:04:04 - INFO - codeparrot_training - Step 8733: {'lr': 0.00047611611288134236, 'samples': 1676928, 'steps': 8733, 'loss/train': 1.8470236659049988} 01/29/2022 00:04:09 - INFO - codeparrot_training - Step 8734: {'lr': 0.00047610913300522576, 'samples': 1677120, 'steps': 8734, 'loss/train': 2.2338473796844482} 01/29/2022 00:04:13 - INFO - codeparrot_training - Step 8735: {'lr': 0.00047610215216052946, 'samples': 1677312, 'steps': 8735, 'loss/train': 2.2561100721359253} 01/29/2022 00:04:20 - INFO - codeparrot_training - Step 8736: {'lr': 0.0004760951703472832, 'samples': 1677504, 'steps': 8736, 'loss/train': 1.8367019891738892} 01/29/2022 00:04:24 - INFO - codeparrot_training - Step 8737: {'lr': 0.0004760881875655171, 'samples': 1677696, 'steps': 8737, 'loss/train': 1.9490118026733398} 01/29/2022 00:04:28 - INFO - codeparrot_training - Step 8738: {'lr': 0.000476081203815261, 'samples': 1677888, 'steps': 8738, 'loss/train': 0.9679361879825592} 01/29/2022 00:04:33 - INFO - codeparrot_training - Step 8739: {'lr': 0.0004760742190965447, 'samples': 1678080, 'steps': 8739, 'loss/train': 2.0652195811271667} 01/29/2022 00:04:37 - INFO - codeparrot_training - Step 8740: {'lr': 0.0004760672334093984, 'samples': 1678272, 'steps': 8740, 'loss/train': 1.9898040890693665} 01/29/2022 00:04:42 - INFO - codeparrot_training - Step 8741: {'lr': 0.0004760602467538517, 'samples': 1678464, 'steps': 8741, 'loss/train': 2.221531391143799} 01/29/2022 00:04:47 - INFO - codeparrot_training - Step 8742: {'lr': 0.0004760532591299348, 'samples': 1678656, 'steps': 8742, 'loss/train': 1.2532351613044739} 01/29/2022 00:04:51 - INFO - codeparrot_training - Step 8743: {'lr': 0.00047604627053767754, 'samples': 1678848, 'steps': 8743, 'loss/train': 2.1543622612953186} 01/29/2022 00:04:55 - INFO - codeparrot_training - Step 8744: {'lr': 0.0004760392809771098, 'samples': 1679040, 'steps': 8744, 'loss/train': 1.4474655389785767} 01/29/2022 00:04:59 - INFO - codeparrot_training - Step 8745: {'lr': 0.00047603229044826146, 'samples': 1679232, 'steps': 8745, 'loss/train': 2.5656960010528564} 01/29/2022 00:05:05 - INFO - codeparrot_training - Step 8746: {'lr': 0.00047602529895116264, 'samples': 1679424, 'steps': 8746, 'loss/train': 1.4441919922828674} 01/29/2022 00:05:09 - INFO - codeparrot_training - Step 8747: {'lr': 0.0004760183064858432, 'samples': 1679616, 'steps': 8747, 'loss/train': 1.7442219257354736} 01/29/2022 00:05:13 - INFO - codeparrot_training - Step 8748: {'lr': 0.0004760113130523331, 'samples': 1679808, 'steps': 8748, 'loss/train': 3.1581262350082397} 01/29/2022 00:05:18 - INFO - codeparrot_training - Step 8749: {'lr': 0.0004760043186506624, 'samples': 1680000, 'steps': 8749, 'loss/train': 1.7738120555877686} 01/29/2022 00:05:22 - INFO - codeparrot_training - Step 8750: {'lr': 0.0004759973232808609, 'samples': 1680192, 'steps': 8750, 'loss/train': 1.546533465385437} 01/29/2022 00:05:29 - INFO - codeparrot_training - Step 8751: {'lr': 0.0004759903269429585, 'samples': 1680384, 'steps': 8751, 'loss/train': 0.3621523976325989} 01/29/2022 00:05:33 - INFO - codeparrot_training - Step 8752: {'lr': 0.00047598332963698543, 'samples': 1680576, 'steps': 8752, 'loss/train': 1.5677799582481384} 01/29/2022 00:05:37 - INFO - codeparrot_training - Step 8753: {'lr': 0.00047597633136297154, 'samples': 1680768, 'steps': 8753, 'loss/train': 0.6215721666812897} 01/29/2022 00:05:42 - INFO - codeparrot_training - Step 8754: {'lr': 0.0004759693321209467, 'samples': 1680960, 'steps': 8754, 'loss/train': 2.261747181415558} 01/29/2022 00:05:46 - INFO - codeparrot_training - Step 8755: {'lr': 0.00047596233191094114, 'samples': 1681152, 'steps': 8755, 'loss/train': 1.6278630495071411} 01/29/2022 00:05:51 - INFO - codeparrot_training - Step 8756: {'lr': 0.0004759553307329846, 'samples': 1681344, 'steps': 8756, 'loss/train': 2.0800082087516785} 01/29/2022 00:05:55 - INFO - codeparrot_training - Step 8757: {'lr': 0.00047594832858710725, 'samples': 1681536, 'steps': 8757, 'loss/train': 1.2640980184078217} 01/29/2022 00:06:00 - INFO - codeparrot_training - Step 8758: {'lr': 0.0004759413254733389, 'samples': 1681728, 'steps': 8758, 'loss/train': 2.079067826271057} 01/29/2022 00:06:04 - INFO - codeparrot_training - Step 8759: {'lr': 0.0004759343213917097, 'samples': 1681920, 'steps': 8759, 'loss/train': 1.7873647212982178} 01/29/2022 00:06:08 - INFO - codeparrot_training - Step 8760: {'lr': 0.0004759273163422496, 'samples': 1682112, 'steps': 8760, 'loss/train': 1.2825530469417572} 01/29/2022 00:06:15 - INFO - codeparrot_training - Step 8761: {'lr': 0.00047592031032498875, 'samples': 1682304, 'steps': 8761, 'loss/train': 1.205560326576233} 01/29/2022 00:06:19 - INFO - codeparrot_training - Step 8762: {'lr': 0.00047591330333995684, 'samples': 1682496, 'steps': 8762, 'loss/train': 1.5076484084129333} 01/29/2022 00:06:23 - INFO - codeparrot_training - Step 8763: {'lr': 0.0004759062953871842, 'samples': 1682688, 'steps': 8763, 'loss/train': 2.7452605962753296} 01/29/2022 00:06:28 - INFO - codeparrot_training - Step 8764: {'lr': 0.0004758992864667007, 'samples': 1682880, 'steps': 8764, 'loss/train': 1.92545485496521} 01/29/2022 00:06:32 - INFO - codeparrot_training - Step 8765: {'lr': 0.0004758922765785363, 'samples': 1683072, 'steps': 8765, 'loss/train': 2.2150830030441284} 01/29/2022 00:06:37 - INFO - codeparrot_training - Step 8766: {'lr': 0.00047588526572272117, 'samples': 1683264, 'steps': 8766, 'loss/train': 1.9623318314552307} 01/29/2022 00:06:41 - INFO - codeparrot_training - Step 8767: {'lr': 0.0004758782538992853, 'samples': 1683456, 'steps': 8767, 'loss/train': 1.3578039407730103} 01/29/2022 00:06:46 - INFO - codeparrot_training - Step 8768: {'lr': 0.00047587124110825874, 'samples': 1683648, 'steps': 8768, 'loss/train': 1.6790208220481873} 01/29/2022 00:06:50 - INFO - codeparrot_training - Step 8769: {'lr': 0.0004758642273496714, 'samples': 1683840, 'steps': 8769, 'loss/train': 2.576123535633087} 01/29/2022 00:06:54 - INFO - codeparrot_training - Step 8770: {'lr': 0.0004758572126235535, 'samples': 1684032, 'steps': 8770, 'loss/train': 0.5597745627164841} 01/29/2022 00:06:59 - INFO - codeparrot_training - Step 8771: {'lr': 0.0004758501969299351, 'samples': 1684224, 'steps': 8771, 'loss/train': 1.3221253752708435} 01/29/2022 00:07:04 - INFO - codeparrot_training - Step 8772: {'lr': 0.0004758431802688461, 'samples': 1684416, 'steps': 8772, 'loss/train': 2.2080768942832947} 01/29/2022 00:07:08 - INFO - codeparrot_training - Step 8773: {'lr': 0.00047583616264031657, 'samples': 1684608, 'steps': 8773, 'loss/train': 0.3103395402431488} 01/29/2022 00:07:12 - INFO - codeparrot_training - Step 8774: {'lr': 0.00047582914404437673, 'samples': 1684800, 'steps': 8774, 'loss/train': 1.9629963040351868} 01/29/2022 00:07:16 - INFO - codeparrot_training - Step 8775: {'lr': 0.00047582212448105647, 'samples': 1684992, 'steps': 8775, 'loss/train': 2.000032424926758} 01/29/2022 00:07:22 - INFO - codeparrot_training - Step 8776: {'lr': 0.000475815103950386, 'samples': 1685184, 'steps': 8776, 'loss/train': 1.9252193570137024} 01/29/2022 00:07:26 - INFO - codeparrot_training - Step 8777: {'lr': 0.00047580808245239526, 'samples': 1685376, 'steps': 8777, 'loss/train': 1.9618881940841675} 01/29/2022 00:07:30 - INFO - codeparrot_training - Step 8778: {'lr': 0.0004758010599871145, 'samples': 1685568, 'steps': 8778, 'loss/train': 1.5990706086158752} 01/29/2022 00:07:34 - INFO - codeparrot_training - Step 8779: {'lr': 0.0004757940365545736, 'samples': 1685760, 'steps': 8779, 'loss/train': 2.230811655521393} 01/29/2022 00:07:39 - INFO - codeparrot_training - Step 8780: {'lr': 0.0004757870121548028, 'samples': 1685952, 'steps': 8780, 'loss/train': 2.2144333720207214} 01/29/2022 00:07:46 - INFO - codeparrot_training - Step 8781: {'lr': 0.00047577998678783207, 'samples': 1686144, 'steps': 8781, 'loss/train': 1.8929146528244019} 01/29/2022 00:07:51 - INFO - codeparrot_training - Step 8782: {'lr': 0.0004757729604536917, 'samples': 1686336, 'steps': 8782, 'loss/train': 1.7531061172485352} 01/29/2022 00:07:55 - INFO - codeparrot_training - Step 8783: {'lr': 0.0004757659331524115, 'samples': 1686528, 'steps': 8783, 'loss/train': 1.7218599915504456} 01/29/2022 00:07:59 - INFO - codeparrot_training - Step 8784: {'lr': 0.00047575890488402183, 'samples': 1686720, 'steps': 8784, 'loss/train': 1.7140125632286072} 01/29/2022 00:08:03 - INFO - codeparrot_training - Step 8785: {'lr': 0.00047575187564855264, 'samples': 1686912, 'steps': 8785, 'loss/train': 1.57326078414917} 01/29/2022 00:08:08 - INFO - codeparrot_training - Step 8786: {'lr': 0.00047574484544603415, 'samples': 1687104, 'steps': 8786, 'loss/train': 1.7071167826652527} 01/29/2022 00:08:13 - INFO - codeparrot_training - Step 8787: {'lr': 0.00047573781427649644, 'samples': 1687296, 'steps': 8787, 'loss/train': 1.492128610610962} 01/29/2022 00:08:17 - INFO - codeparrot_training - Step 8788: {'lr': 0.00047573078213996954, 'samples': 1687488, 'steps': 8788, 'loss/train': 0.7898291945457458} 01/29/2022 00:08:21 - INFO - codeparrot_training - Step 8789: {'lr': 0.0004757237490364836, 'samples': 1687680, 'steps': 8789, 'loss/train': 2.3873342871665955} 01/29/2022 00:08:25 - INFO - codeparrot_training - Step 8790: {'lr': 0.00047571671496606893, 'samples': 1687872, 'steps': 8790, 'loss/train': 2.3833975195884705} 01/29/2022 00:08:31 - INFO - codeparrot_training - Step 8791: {'lr': 0.0004757096799287555, 'samples': 1688064, 'steps': 8791, 'loss/train': 1.811859369277954} 01/29/2022 00:08:35 - INFO - codeparrot_training - Step 8792: {'lr': 0.0004757026439245735, 'samples': 1688256, 'steps': 8792, 'loss/train': 2.3179982900619507} 01/29/2022 00:08:39 - INFO - codeparrot_training - Step 8793: {'lr': 0.00047569560695355295, 'samples': 1688448, 'steps': 8793, 'loss/train': 1.623557209968567} 01/29/2022 00:08:43 - INFO - codeparrot_training - Step 8794: {'lr': 0.0004756885690157241, 'samples': 1688640, 'steps': 8794, 'loss/train': 1.0462644696235657} 01/29/2022 00:08:48 - INFO - codeparrot_training - Step 8795: {'lr': 0.00047568153011111715, 'samples': 1688832, 'steps': 8795, 'loss/train': 1.9598702788352966} 01/29/2022 00:08:55 - INFO - codeparrot_training - Step 8796: {'lr': 0.00047567449023976213, 'samples': 1689024, 'steps': 8796, 'loss/train': 1.6298884749412537} 01/29/2022 00:08:59 - INFO - codeparrot_training - Step 8797: {'lr': 0.00047566744940168924, 'samples': 1689216, 'steps': 8797, 'loss/train': 1.7772504687309265} 01/29/2022 00:09:03 - INFO - codeparrot_training - Step 8798: {'lr': 0.0004756604075969287, 'samples': 1689408, 'steps': 8798, 'loss/train': 1.4522274434566498} 01/29/2022 00:09:07 - INFO - codeparrot_training - Step 8799: {'lr': 0.0004756533648255106, 'samples': 1689600, 'steps': 8799, 'loss/train': 1.2079349756240845} 01/29/2022 00:09:13 - INFO - codeparrot_training - Step 8800: {'lr': 0.0004756463210874652, 'samples': 1689792, 'steps': 8800, 'loss/train': 1.3300538063049316} 01/29/2022 00:09:17 - INFO - codeparrot_training - Step 8801: {'lr': 0.0004756392763828226, 'samples': 1689984, 'steps': 8801, 'loss/train': 1.485665202140808} 01/29/2022 00:09:21 - INFO - codeparrot_training - Step 8802: {'lr': 0.0004756322307116129, 'samples': 1690176, 'steps': 8802, 'loss/train': 1.87589693069458} 01/29/2022 00:09:25 - INFO - codeparrot_training - Step 8803: {'lr': 0.0004756251840738664, 'samples': 1690368, 'steps': 8803, 'loss/train': 1.5665629506111145} 01/29/2022 00:09:30 - INFO - codeparrot_training - Step 8804: {'lr': 0.00047561813646961325, 'samples': 1690560, 'steps': 8804, 'loss/train': 1.6793966889381409} 01/29/2022 00:09:37 - INFO - codeparrot_training - Step 8805: {'lr': 0.00047561108789888367, 'samples': 1690752, 'steps': 8805, 'loss/train': 1.7055307030677795} 01/29/2022 00:09:41 - INFO - codeparrot_training - Step 8806: {'lr': 0.0004756040383617078, 'samples': 1690944, 'steps': 8806, 'loss/train': 1.7726365327835083} 01/29/2022 00:09:45 - INFO - codeparrot_training - Step 8807: {'lr': 0.00047559698785811595, 'samples': 1691136, 'steps': 8807, 'loss/train': 1.0722298622131348} 01/29/2022 00:09:49 - INFO - codeparrot_training - Step 8808: {'lr': 0.0004755899363881382, 'samples': 1691328, 'steps': 8808, 'loss/train': 0.11983025819063187} 01/29/2022 00:09:54 - INFO - codeparrot_training - Step 8809: {'lr': 0.00047558288395180477, 'samples': 1691520, 'steps': 8809, 'loss/train': 1.4731637835502625} 01/29/2022 00:09:59 - INFO - codeparrot_training - Step 8810: {'lr': 0.0004755758305491459, 'samples': 1691712, 'steps': 8810, 'loss/train': 1.3191196024417877} 01/29/2022 00:10:03 - INFO - codeparrot_training - Step 8811: {'lr': 0.0004755687761801918, 'samples': 1691904, 'steps': 8811, 'loss/train': 1.4959874153137207} 01/29/2022 00:10:07 - INFO - codeparrot_training - Step 8812: {'lr': 0.00047556172084497274, 'samples': 1692096, 'steps': 8812, 'loss/train': 2.1222957372665405} 01/29/2022 00:10:12 - INFO - codeparrot_training - Step 8813: {'lr': 0.0004755546645435188, 'samples': 1692288, 'steps': 8813, 'loss/train': 1.237536460161209} 01/29/2022 00:10:16 - INFO - codeparrot_training - Step 8814: {'lr': 0.0004755476072758604, 'samples': 1692480, 'steps': 8814, 'loss/train': 0.8051057159900665} 01/29/2022 00:10:21 - INFO - codeparrot_training - Step 8815: {'lr': 0.0004755405490420276, 'samples': 1692672, 'steps': 8815, 'loss/train': 1.8230631351470947} 01/29/2022 00:10:25 - INFO - codeparrot_training - Step 8816: {'lr': 0.0004755334898420507, 'samples': 1692864, 'steps': 8816, 'loss/train': 2.161327600479126} 01/29/2022 00:10:30 - INFO - codeparrot_training - Step 8817: {'lr': 0.00047552642967596, 'samples': 1693056, 'steps': 8817, 'loss/train': 3.439960241317749} 01/29/2022 00:10:34 - INFO - codeparrot_training - Step 8818: {'lr': 0.00047551936854378564, 'samples': 1693248, 'steps': 8818, 'loss/train': 1.9045182466506958} 01/29/2022 00:10:38 - INFO - codeparrot_training - Step 8819: {'lr': 0.00047551230644555793, 'samples': 1693440, 'steps': 8819, 'loss/train': 2.2378724813461304} 01/29/2022 00:10:45 - INFO - codeparrot_training - Step 8820: {'lr': 0.00047550524338130706, 'samples': 1693632, 'steps': 8820, 'loss/train': 1.830689013004303} 01/29/2022 00:10:49 - INFO - codeparrot_training - Step 8821: {'lr': 0.00047549817935106344, 'samples': 1693824, 'steps': 8821, 'loss/train': 7.18413233757019} 01/29/2022 00:10:54 - INFO - codeparrot_training - Step 8822: {'lr': 0.00047549111435485716, 'samples': 1694016, 'steps': 8822, 'loss/train': 1.2030088305473328} 01/29/2022 00:10:58 - INFO - codeparrot_training - Step 8823: {'lr': 0.0004754840483927185, 'samples': 1694208, 'steps': 8823, 'loss/train': 1.6242932081222534} 01/29/2022 00:11:02 - INFO - codeparrot_training - Step 8824: {'lr': 0.0004754769814646779, 'samples': 1694400, 'steps': 8824, 'loss/train': 1.6519707441329956} 01/29/2022 00:11:06 - INFO - codeparrot_training - Step 8825: {'lr': 0.00047546991357076544, 'samples': 1694592, 'steps': 8825, 'loss/train': 2.1059144139289856} 01/29/2022 00:11:12 - INFO - codeparrot_training - Step 8826: {'lr': 0.00047546284471101143, 'samples': 1694784, 'steps': 8826, 'loss/train': 1.0268814861774445} 01/29/2022 00:11:16 - INFO - codeparrot_training - Step 8827: {'lr': 0.00047545577488544623, 'samples': 1694976, 'steps': 8827, 'loss/train': 2.1810248494148254} 01/29/2022 00:11:20 - INFO - codeparrot_training - Step 8828: {'lr': 0.0004754487040941001, 'samples': 1695168, 'steps': 8828, 'loss/train': 1.8196935653686523} 01/29/2022 00:11:25 - INFO - codeparrot_training - Step 8829: {'lr': 0.00047544163233700324, 'samples': 1695360, 'steps': 8829, 'loss/train': 1.1202350556850433} 01/29/2022 00:11:29 - INFO - codeparrot_training - Step 8830: {'lr': 0.00047543455961418605, 'samples': 1695552, 'steps': 8830, 'loss/train': 1.8576745390892029} 01/29/2022 00:11:34 - INFO - codeparrot_training - Step 8831: {'lr': 0.0004754274859256788, 'samples': 1695744, 'steps': 8831, 'loss/train': 1.2094101905822754} 01/29/2022 00:11:39 - INFO - codeparrot_training - Step 8832: {'lr': 0.0004754204112715118, 'samples': 1695936, 'steps': 8832, 'loss/train': 1.7890411019325256} 01/29/2022 00:11:43 - INFO - codeparrot_training - Step 8833: {'lr': 0.0004754133356517153, 'samples': 1696128, 'steps': 8833, 'loss/train': 2.000766634941101} 01/29/2022 00:11:47 - INFO - codeparrot_training - Step 8834: {'lr': 0.0004754062590663196, 'samples': 1696320, 'steps': 8834, 'loss/train': 2.137618124485016} 01/29/2022 00:11:51 - INFO - codeparrot_training - Step 8835: {'lr': 0.00047539918151535515, 'samples': 1696512, 'steps': 8835, 'loss/train': 1.758943498134613} 01/29/2022 00:11:57 - INFO - codeparrot_training - Step 8836: {'lr': 0.00047539210299885217, 'samples': 1696704, 'steps': 8836, 'loss/train': 2.1336450576782227} 01/29/2022 00:12:01 - INFO - codeparrot_training - Step 8837: {'lr': 0.00047538502351684097, 'samples': 1696896, 'steps': 8837, 'loss/train': 1.8265319466590881} 01/29/2022 00:12:05 - INFO - codeparrot_training - Step 8838: {'lr': 0.0004753779430693519, 'samples': 1697088, 'steps': 8838, 'loss/train': 2.16593599319458} 01/29/2022 00:12:10 - INFO - codeparrot_training - Step 8839: {'lr': 0.0004753708616564153, 'samples': 1697280, 'steps': 8839, 'loss/train': 1.286354899406433} 01/29/2022 00:12:14 - INFO - codeparrot_training - Step 8840: {'lr': 0.00047536377927806143, 'samples': 1697472, 'steps': 8840, 'loss/train': 2.037503957748413} 01/29/2022 00:12:20 - INFO - codeparrot_training - Step 8841: {'lr': 0.0004753566959343207, 'samples': 1697664, 'steps': 8841, 'loss/train': 1.4485865235328674} 01/29/2022 00:12:24 - INFO - codeparrot_training - Step 8842: {'lr': 0.0004753496116252235, 'samples': 1697856, 'steps': 8842, 'loss/train': 1.8694242238998413} 01/29/2022 00:12:28 - INFO - codeparrot_training - Step 8843: {'lr': 0.0004753425263508001, 'samples': 1698048, 'steps': 8843, 'loss/train': 2.200652003288269} 01/29/2022 00:12:33 - INFO - codeparrot_training - Step 8844: {'lr': 0.0004753354401110809, 'samples': 1698240, 'steps': 8844, 'loss/train': 1.1223079562187195} 01/29/2022 00:12:37 - INFO - codeparrot_training - Step 8845: {'lr': 0.00047532835290609623, 'samples': 1698432, 'steps': 8845, 'loss/train': 1.9928945302963257} 01/29/2022 00:12:42 - INFO - codeparrot_training - Step 8846: {'lr': 0.00047532126473587635, 'samples': 1698624, 'steps': 8846, 'loss/train': 1.2588201463222504} 01/29/2022 00:12:47 - INFO - codeparrot_training - Step 8847: {'lr': 0.0004753141756004518, 'samples': 1698816, 'steps': 8847, 'loss/train': 3.073520064353943} 01/29/2022 00:12:51 - INFO - codeparrot_training - Step 8848: {'lr': 0.00047530708549985287, 'samples': 1699008, 'steps': 8848, 'loss/train': 1.1489570438861847} 01/29/2022 00:12:55 - INFO - codeparrot_training - Step 8849: {'lr': 0.00047529999443410986, 'samples': 1699200, 'steps': 8849, 'loss/train': 1.9413810968399048} 01/29/2022 00:13:01 - INFO - codeparrot_training - Step 8850: {'lr': 0.0004752929024032533, 'samples': 1699392, 'steps': 8850, 'loss/train': 2.1380832195281982} 01/29/2022 00:13:05 - INFO - codeparrot_training - Step 8851: {'lr': 0.0004752858094073134, 'samples': 1699584, 'steps': 8851, 'loss/train': 2.48450767993927} 01/29/2022 00:13:09 - INFO - codeparrot_training - Step 8852: {'lr': 0.0004752787154463207, 'samples': 1699776, 'steps': 8852, 'loss/train': 0.5671792477369308} 01/29/2022 00:13:13 - INFO - codeparrot_training - Step 8853: {'lr': 0.0004752716205203055, 'samples': 1699968, 'steps': 8853, 'loss/train': 1.7307239770889282} 01/29/2022 00:13:18 - INFO - codeparrot_training - Step 8854: {'lr': 0.0004752645246292982, 'samples': 1700160, 'steps': 8854, 'loss/train': 1.0513864159584045} 01/29/2022 00:13:24 - INFO - codeparrot_training - Step 8855: {'lr': 0.0004752574277733292, 'samples': 1700352, 'steps': 8855, 'loss/train': 2.0805745124816895} 01/29/2022 00:13:28 - INFO - codeparrot_training - Step 8856: {'lr': 0.0004752503299524289, 'samples': 1700544, 'steps': 8856, 'loss/train': 1.6037800312042236} 01/29/2022 00:13:32 - INFO - codeparrot_training - Step 8857: {'lr': 0.0004752432311666277, 'samples': 1700736, 'steps': 8857, 'loss/train': 1.0598106980323792} 01/29/2022 00:13:37 - INFO - codeparrot_training - Step 8858: {'lr': 0.0004752361314159561, 'samples': 1700928, 'steps': 8858, 'loss/train': 1.710972011089325} 01/29/2022 00:13:41 - INFO - codeparrot_training - Step 8859: {'lr': 0.0004752290307004444, 'samples': 1701120, 'steps': 8859, 'loss/train': 1.047551840543747} 01/29/2022 00:13:46 - INFO - codeparrot_training - Step 8860: {'lr': 0.000475221929020123, 'samples': 1701312, 'steps': 8860, 'loss/train': 0.3189517557621002} 01/29/2022 00:13:51 - INFO - codeparrot_training - Step 8861: {'lr': 0.00047521482637502246, 'samples': 1701504, 'steps': 8861, 'loss/train': 0.2518366724252701} 01/29/2022 00:13:55 - INFO - codeparrot_training - Step 8862: {'lr': 0.00047520772276517297, 'samples': 1701696, 'steps': 8862, 'loss/train': 3.18711519241333} 01/29/2022 00:13:59 - INFO - codeparrot_training - Step 8863: {'lr': 0.0004752006181906052, 'samples': 1701888, 'steps': 8863, 'loss/train': 0.45283864438533783} 01/29/2022 00:14:03 - INFO - codeparrot_training - Step 8864: {'lr': 0.00047519351265134954, 'samples': 1702080, 'steps': 8864, 'loss/train': 1.0217683017253876} 01/29/2022 00:14:09 - INFO - codeparrot_training - Step 8865: {'lr': 0.0004751864061474364, 'samples': 1702272, 'steps': 8865, 'loss/train': 1.4548844397068024} 01/29/2022 00:14:14 - INFO - codeparrot_training - Step 8866: {'lr': 0.000475179298678896, 'samples': 1702464, 'steps': 8866, 'loss/train': 2.3633941411972046} 01/29/2022 00:14:18 - INFO - codeparrot_training - Step 8867: {'lr': 0.0004751721902457592, 'samples': 1702656, 'steps': 8867, 'loss/train': 1.8264520168304443} 01/29/2022 00:14:22 - INFO - codeparrot_training - Step 8868: {'lr': 0.0004751650808480561, 'samples': 1702848, 'steps': 8868, 'loss/train': 1.9430869817733765} 01/29/2022 00:14:26 - INFO - codeparrot_training - Step 8869: {'lr': 0.00047515797048581734, 'samples': 1703040, 'steps': 8869, 'loss/train': 1.8501057028770447} 01/29/2022 00:14:31 - INFO - codeparrot_training - Step 8870: {'lr': 0.00047515085915907334, 'samples': 1703232, 'steps': 8870, 'loss/train': 1.9409483671188354} 01/29/2022 00:14:36 - INFO - codeparrot_training - Step 8871: {'lr': 0.00047514374686785454, 'samples': 1703424, 'steps': 8871, 'loss/train': 2.2944170236587524} 01/29/2022 00:14:40 - INFO - codeparrot_training - Step 8872: {'lr': 0.00047513663361219144, 'samples': 1703616, 'steps': 8872, 'loss/train': 1.7094401121139526} 01/29/2022 00:14:44 - INFO - codeparrot_training - Step 8873: {'lr': 0.00047512951939211447, 'samples': 1703808, 'steps': 8873, 'loss/train': 2.2084615230560303} 01/29/2022 00:14:48 - INFO - codeparrot_training - Step 8874: {'lr': 0.0004751224042076542, 'samples': 1704000, 'steps': 8874, 'loss/train': 2.371273934841156} 01/29/2022 00:14:54 - INFO - codeparrot_training - Step 8875: {'lr': 0.0004751152880588409, 'samples': 1704192, 'steps': 8875, 'loss/train': 1.5205112099647522} 01/29/2022 00:14:58 - INFO - codeparrot_training - Step 8876: {'lr': 0.00047510817094570526, 'samples': 1704384, 'steps': 8876, 'loss/train': 1.7173828482627869} 01/29/2022 00:15:02 - INFO - codeparrot_training - Step 8877: {'lr': 0.0004751010528682777, 'samples': 1704576, 'steps': 8877, 'loss/train': 1.8158823251724243} 01/29/2022 00:15:07 - INFO - codeparrot_training - Step 8878: {'lr': 0.0004750939338265887, 'samples': 1704768, 'steps': 8878, 'loss/train': 1.9934032559394836} 01/29/2022 00:15:11 - INFO - codeparrot_training - Step 8879: {'lr': 0.0004750868138206688, 'samples': 1704960, 'steps': 8879, 'loss/train': 2.2388625741004944} 01/29/2022 00:15:17 - INFO - codeparrot_training - Step 8880: {'lr': 0.0004750796928505484, 'samples': 1705152, 'steps': 8880, 'loss/train': 1.7484202980995178} 01/29/2022 00:15:22 - INFO - codeparrot_training - Step 8881: {'lr': 0.0004750725709162581, 'samples': 1705344, 'steps': 8881, 'loss/train': 1.345548927783966} 01/29/2022 00:15:26 - INFO - codeparrot_training - Step 8882: {'lr': 0.00047506544801782834, 'samples': 1705536, 'steps': 8882, 'loss/train': 1.614531397819519} 01/29/2022 00:15:30 - INFO - codeparrot_training - Step 8883: {'lr': 0.00047505832415528973, 'samples': 1705728, 'steps': 8883, 'loss/train': 2.202703356742859} 01/29/2022 00:15:34 - INFO - codeparrot_training - Step 8884: {'lr': 0.0004750511993286727, 'samples': 1705920, 'steps': 8884, 'loss/train': 0.3767813444137573} 01/29/2022 00:15:40 - INFO - codeparrot_training - Step 8885: {'lr': 0.0004750440735380077, 'samples': 1706112, 'steps': 8885, 'loss/train': 1.8167994618415833} 01/29/2022 00:15:44 - INFO - codeparrot_training - Step 8886: {'lr': 0.00047503694678332543, 'samples': 1706304, 'steps': 8886, 'loss/train': 2.0722596645355225} 01/29/2022 00:15:48 - INFO - codeparrot_training - Step 8887: {'lr': 0.00047502981906465634, 'samples': 1706496, 'steps': 8887, 'loss/train': 1.8757867813110352} 01/29/2022 00:15:53 - INFO - codeparrot_training - Step 8888: {'lr': 0.000475022690382031, 'samples': 1706688, 'steps': 8888, 'loss/train': 1.3836701810359955} 01/29/2022 00:15:57 - INFO - codeparrot_training - Step 8889: {'lr': 0.0004750155607354799, 'samples': 1706880, 'steps': 8889, 'loss/train': 1.7845532298088074} 01/29/2022 00:16:01 - INFO - codeparrot_training - Step 8890: {'lr': 0.0004750084301250335, 'samples': 1707072, 'steps': 8890, 'loss/train': 0.9938734173774719} 01/29/2022 00:16:06 - INFO - codeparrot_training - Step 8891: {'lr': 0.0004750012985507225, 'samples': 1707264, 'steps': 8891, 'loss/train': 1.5380916595458984} 01/29/2022 00:16:11 - INFO - codeparrot_training - Step 8892: {'lr': 0.0004749941660125774, 'samples': 1707456, 'steps': 8892, 'loss/train': 2.8817542791366577} 01/29/2022 00:16:15 - INFO - codeparrot_training - Step 8893: {'lr': 0.0004749870325106287, 'samples': 1707648, 'steps': 8893, 'loss/train': 1.818344235420227} 01/29/2022 00:16:19 - INFO - codeparrot_training - Step 8894: {'lr': 0.00047497989804490693, 'samples': 1707840, 'steps': 8894, 'loss/train': 1.1531207263469696} 01/29/2022 00:16:23 - INFO - codeparrot_training - Step 8895: {'lr': 0.0004749727626154428, 'samples': 1708032, 'steps': 8895, 'loss/train': 1.445004254579544} 01/29/2022 00:16:29 - INFO - codeparrot_training - Step 8896: {'lr': 0.0004749656262222668, 'samples': 1708224, 'steps': 8896, 'loss/train': 2.0661978721618652} 01/29/2022 00:16:33 - INFO - codeparrot_training - Step 8897: {'lr': 0.0004749584888654095, 'samples': 1708416, 'steps': 8897, 'loss/train': 2.131526827812195} 01/29/2022 00:16:38 - INFO - codeparrot_training - Step 8898: {'lr': 0.0004749513505449014, 'samples': 1708608, 'steps': 8898, 'loss/train': 1.4280672669410706} 01/29/2022 00:16:42 - INFO - codeparrot_training - Step 8899: {'lr': 0.00047494421126077313, 'samples': 1708800, 'steps': 8899, 'loss/train': 1.0434986650943756} 01/29/2022 00:16:46 - INFO - codeparrot_training - Step 8900: {'lr': 0.0004749370710130554, 'samples': 1708992, 'steps': 8900, 'loss/train': 2.2776060104370117} 01/29/2022 00:16:52 - INFO - codeparrot_training - Step 8901: {'lr': 0.0004749299298017786, 'samples': 1709184, 'steps': 8901, 'loss/train': 2.0595794320106506} 01/29/2022 00:16:56 - INFO - codeparrot_training - Step 8902: {'lr': 0.00047492278762697337, 'samples': 1709376, 'steps': 8902, 'loss/train': 1.9869089126586914} 01/29/2022 00:17:01 - INFO - codeparrot_training - Step 8903: {'lr': 0.0004749156444886704, 'samples': 1709568, 'steps': 8903, 'loss/train': 1.6249203085899353} 01/29/2022 00:17:05 - INFO - codeparrot_training - Step 8904: {'lr': 0.0004749085003869003, 'samples': 1709760, 'steps': 8904, 'loss/train': 0.6984638124704361} 01/29/2022 00:17:09 - INFO - codeparrot_training - Step 8905: {'lr': 0.00047490135532169347, 'samples': 1709952, 'steps': 8905, 'loss/train': 2.1978930830955505} 01/29/2022 00:17:15 - INFO - codeparrot_training - Step 8906: {'lr': 0.0004748942092930807, 'samples': 1710144, 'steps': 8906, 'loss/train': 2.01222825050354} 01/29/2022 00:17:19 - INFO - codeparrot_training - Step 8907: {'lr': 0.00047488706230109257, 'samples': 1710336, 'steps': 8907, 'loss/train': 0.21036799997091293} 01/29/2022 00:17:23 - INFO - codeparrot_training - Step 8908: {'lr': 0.00047487991434575963, 'samples': 1710528, 'steps': 8908, 'loss/train': 1.313698947429657} 01/29/2022 00:17:28 - INFO - codeparrot_training - Step 8909: {'lr': 0.0004748727654271126, 'samples': 1710720, 'steps': 8909, 'loss/train': 1.5430394411087036} 01/29/2022 00:17:33 - INFO - codeparrot_training - Step 8910: {'lr': 0.000474865615545182, 'samples': 1710912, 'steps': 8910, 'loss/train': 1.6408326029777527} 01/29/2022 00:17:38 - INFO - codeparrot_training - Step 8911: {'lr': 0.0004748584646999985, 'samples': 1711104, 'steps': 8911, 'loss/train': 1.6760326623916626} 01/29/2022 00:17:42 - INFO - codeparrot_training - Step 8912: {'lr': 0.0004748513128915928, 'samples': 1711296, 'steps': 8912, 'loss/train': 1.3250713348388672} 01/29/2022 00:17:46 - INFO - codeparrot_training - Step 8913: {'lr': 0.0004748441601199954, 'samples': 1711488, 'steps': 8913, 'loss/train': 1.4923744797706604} 01/29/2022 00:17:50 - INFO - codeparrot_training - Step 8914: {'lr': 0.0004748370063852371, 'samples': 1711680, 'steps': 8914, 'loss/train': 1.451037883758545} 01/29/2022 00:17:56 - INFO - codeparrot_training - Step 8915: {'lr': 0.0004748298516873484, 'samples': 1711872, 'steps': 8915, 'loss/train': 1.9222320914268494} 01/29/2022 00:18:00 - INFO - codeparrot_training - Step 8916: {'lr': 0.00047482269602636, 'samples': 1712064, 'steps': 8916, 'loss/train': 1.5875152945518494} 01/29/2022 00:18:04 - INFO - codeparrot_training - Step 8917: {'lr': 0.00047481553940230257, 'samples': 1712256, 'steps': 8917, 'loss/train': 1.785503625869751} 01/29/2022 00:18:08 - INFO - codeparrot_training - Step 8918: {'lr': 0.0004748083818152067, 'samples': 1712448, 'steps': 8918, 'loss/train': 1.5873371958732605} 01/29/2022 00:18:13 - INFO - codeparrot_training - Step 8919: {'lr': 0.00047480122326510325, 'samples': 1712640, 'steps': 8919, 'loss/train': 1.941427767276764} 01/29/2022 00:18:18 - INFO - codeparrot_training - Step 8920: {'lr': 0.0004747940637520226, 'samples': 1712832, 'steps': 8920, 'loss/train': 2.3060744404792786} 01/29/2022 00:18:22 - INFO - codeparrot_training - Step 8921: {'lr': 0.0004747869032759956, 'samples': 1713024, 'steps': 8921, 'loss/train': 1.5511837005615234} 01/29/2022 00:18:26 - INFO - codeparrot_training - Step 8922: {'lr': 0.00047477974183705293, 'samples': 1713216, 'steps': 8922, 'loss/train': 1.0721950829029083} 01/29/2022 00:18:31 - INFO - codeparrot_training - Step 8923: {'lr': 0.0004747725794352252, 'samples': 1713408, 'steps': 8923, 'loss/train': 1.5898793935775757} 01/29/2022 00:18:35 - INFO - codeparrot_training - Step 8924: {'lr': 0.00047476541607054313, 'samples': 1713600, 'steps': 8924, 'loss/train': 1.5716437697410583} 01/29/2022 00:18:41 - INFO - codeparrot_training - Step 8925: {'lr': 0.0004747582517430373, 'samples': 1713792, 'steps': 8925, 'loss/train': 1.2819152176380157} 01/29/2022 00:18:45 - INFO - codeparrot_training - Step 8926: {'lr': 0.00047475108645273856, 'samples': 1713984, 'steps': 8926, 'loss/train': 1.2246852815151215} 01/29/2022 00:18:49 - INFO - codeparrot_training - Step 8927: {'lr': 0.00047474392019967754, 'samples': 1714176, 'steps': 8927, 'loss/train': 0.9956980347633362} 01/29/2022 00:18:54 - INFO - codeparrot_training - Step 8928: {'lr': 0.0004747367529838849, 'samples': 1714368, 'steps': 8928, 'loss/train': 1.5253615379333496} 01/29/2022 00:18:58 - INFO - codeparrot_training - Step 8929: {'lr': 0.0004747295848053914, 'samples': 1714560, 'steps': 8929, 'loss/train': 2.1445348262786865} 01/29/2022 00:19:03 - INFO - codeparrot_training - Step 8930: {'lr': 0.0004747224156642277, 'samples': 1714752, 'steps': 8930, 'loss/train': 2.4166781902313232} 01/29/2022 00:19:08 - INFO - codeparrot_training - Step 8931: {'lr': 0.00047471524556042454, 'samples': 1714944, 'steps': 8931, 'loss/train': 1.3901365399360657} 01/29/2022 00:19:12 - INFO - codeparrot_training - Step 8932: {'lr': 0.00047470807449401264, 'samples': 1715136, 'steps': 8932, 'loss/train': 2.4518423080444336} 01/29/2022 00:19:16 - INFO - codeparrot_training - Step 8933: {'lr': 0.0004747009024650227, 'samples': 1715328, 'steps': 8933, 'loss/train': 1.8514711260795593} 01/29/2022 00:19:20 - INFO - codeparrot_training - Step 8934: {'lr': 0.00047469372947348546, 'samples': 1715520, 'steps': 8934, 'loss/train': 1.9056853652000427} 01/29/2022 00:19:25 - INFO - codeparrot_training - Step 8935: {'lr': 0.0004746865555194315, 'samples': 1715712, 'steps': 8935, 'loss/train': 1.4439412951469421} 01/29/2022 00:19:30 - INFO - codeparrot_training - Step 8936: {'lr': 0.00047467938060289185, 'samples': 1715904, 'steps': 8936, 'loss/train': 2.2502177953720093} 01/29/2022 00:19:34 - INFO - codeparrot_training - Step 8937: {'lr': 0.00047467220472389694, 'samples': 1716096, 'steps': 8937, 'loss/train': 1.2151841819286346} 01/29/2022 00:19:38 - INFO - codeparrot_training - Step 8938: {'lr': 0.0004746650278824777, 'samples': 1716288, 'steps': 8938, 'loss/train': 1.7106306552886963} 01/29/2022 00:19:42 - INFO - codeparrot_training - Step 8939: {'lr': 0.00047465785007866487, 'samples': 1716480, 'steps': 8939, 'loss/train': 1.3804548382759094} 01/29/2022 00:19:48 - INFO - codeparrot_training - Step 8940: {'lr': 0.00047465067131248907, 'samples': 1716672, 'steps': 8940, 'loss/train': 1.485702931880951} 01/29/2022 00:19:52 - INFO - codeparrot_training - Step 8941: {'lr': 0.0004746434915839812, 'samples': 1716864, 'steps': 8941, 'loss/train': 1.6254027485847473} 01/29/2022 00:19:56 - INFO - codeparrot_training - Step 8942: {'lr': 0.00047463631089317195, 'samples': 1717056, 'steps': 8942, 'loss/train': 1.278743952512741} 01/29/2022 00:20:00 - INFO - codeparrot_training - Step 8943: {'lr': 0.000474629129240092, 'samples': 1717248, 'steps': 8943, 'loss/train': 1.6037734150886536} 01/29/2022 00:20:05 - INFO - codeparrot_training - Step 8944: {'lr': 0.0004746219466247722, 'samples': 1717440, 'steps': 8944, 'loss/train': 1.1273286938667297} 01/29/2022 00:20:10 - INFO - codeparrot_training - Step 8945: {'lr': 0.0004746147630472434, 'samples': 1717632, 'steps': 8945, 'loss/train': 1.8238282799720764} 01/29/2022 00:20:15 - INFO - codeparrot_training - Step 8946: {'lr': 0.00047460757850753614, 'samples': 1717824, 'steps': 8946, 'loss/train': 1.7665647268295288} 01/29/2022 00:20:19 - INFO - codeparrot_training - Step 8947: {'lr': 0.00047460039300568143, 'samples': 1718016, 'steps': 8947, 'loss/train': 1.693817675113678} 01/29/2022 00:20:23 - INFO - codeparrot_training - Step 8948: {'lr': 0.0004745932065417099, 'samples': 1718208, 'steps': 8948, 'loss/train': 1.6499165296554565} 01/29/2022 00:20:27 - INFO - codeparrot_training - Step 8949: {'lr': 0.00047458601911565246, 'samples': 1718400, 'steps': 8949, 'loss/train': 1.0078609585762024} 01/29/2022 00:20:33 - INFO - codeparrot_training - Step 8950: {'lr': 0.0004745788307275398, 'samples': 1718592, 'steps': 8950, 'loss/train': 2.196719527244568} 01/29/2022 00:20:37 - INFO - codeparrot_training - Step 8951: {'lr': 0.0004745716413774027, 'samples': 1718784, 'steps': 8951, 'loss/train': 2.00017386674881} 01/29/2022 00:20:41 - INFO - codeparrot_training - Step 8952: {'lr': 0.000474564451065272, 'samples': 1718976, 'steps': 8952, 'loss/train': 1.519524335861206} 01/29/2022 00:20:45 - INFO - codeparrot_training - Step 8953: {'lr': 0.00047455725979117855, 'samples': 1719168, 'steps': 8953, 'loss/train': 1.8535282015800476} 01/29/2022 00:20:50 - INFO - codeparrot_training - Step 8954: {'lr': 0.00047455006755515306, 'samples': 1719360, 'steps': 8954, 'loss/train': 0.6453346163034439} 01/29/2022 00:20:56 - INFO - codeparrot_training - Step 8955: {'lr': 0.00047454287435722643, 'samples': 1719552, 'steps': 8955, 'loss/train': 1.2843382358551025} 01/29/2022 00:21:00 - INFO - codeparrot_training - Step 8956: {'lr': 0.00047453568019742936, 'samples': 1719744, 'steps': 8956, 'loss/train': 2.016169846057892} 01/29/2022 00:21:05 - INFO - codeparrot_training - Step 8957: {'lr': 0.0004745284850757928, 'samples': 1719936, 'steps': 8957, 'loss/train': 1.8076931834220886} 01/29/2022 00:21:09 - INFO - codeparrot_training - Step 8958: {'lr': 0.00047452128899234746, 'samples': 1720128, 'steps': 8958, 'loss/train': 2.070251762866974} 01/29/2022 00:21:13 - INFO - codeparrot_training - Step 8959: {'lr': 0.0004745140919471243, 'samples': 1720320, 'steps': 8959, 'loss/train': 1.119606614112854} 01/29/2022 00:21:18 - INFO - codeparrot_training - Step 8960: {'lr': 0.0004745068939401539, 'samples': 1720512, 'steps': 8960, 'loss/train': 0.09280085004866123} 01/29/2022 00:21:23 - INFO - codeparrot_training - Step 8961: {'lr': 0.0004744996949714674, 'samples': 1720704, 'steps': 8961, 'loss/train': 1.7466904520988464} 01/29/2022 00:21:28 - INFO - codeparrot_training - Step 8962: {'lr': 0.0004744924950410954, 'samples': 1720896, 'steps': 8962, 'loss/train': 1.5212445259094238} 01/29/2022 00:21:32 - INFO - codeparrot_training - Step 8963: {'lr': 0.0004744852941490689, 'samples': 1721088, 'steps': 8963, 'loss/train': 1.7269161343574524} 01/29/2022 00:21:36 - INFO - codeparrot_training - Step 8964: {'lr': 0.0004744780922954186, 'samples': 1721280, 'steps': 8964, 'loss/train': 0.5225283801555634} 01/29/2022 00:21:40 - INFO - codeparrot_training - Step 8965: {'lr': 0.00047447088948017555, 'samples': 1721472, 'steps': 8965, 'loss/train': 1.7061760425567627} 01/29/2022 00:21:46 - INFO - codeparrot_training - Step 8966: {'lr': 0.0004744636857033704, 'samples': 1721664, 'steps': 8966, 'loss/train': 0.7811174690723419} 01/29/2022 00:21:50 - INFO - codeparrot_training - Step 8967: {'lr': 0.00047445648096503413, 'samples': 1721856, 'steps': 8967, 'loss/train': 1.2047859728336334} 01/29/2022 00:21:54 - INFO - codeparrot_training - Step 8968: {'lr': 0.00047444927526519757, 'samples': 1722048, 'steps': 8968, 'loss/train': 2.264284908771515} 01/29/2022 00:21:59 - INFO - codeparrot_training - Step 8969: {'lr': 0.00047444206860389155, 'samples': 1722240, 'steps': 8969, 'loss/train': 1.8737919330596924} 01/29/2022 00:22:03 - INFO - codeparrot_training - Step 8970: {'lr': 0.00047443486098114703, 'samples': 1722432, 'steps': 8970, 'loss/train': 2.5610942244529724} 01/29/2022 00:22:09 - INFO - codeparrot_training - Step 8971: {'lr': 0.0004744276523969948, 'samples': 1722624, 'steps': 8971, 'loss/train': 2.015753209590912} 01/29/2022 00:22:13 - INFO - codeparrot_training - Step 8972: {'lr': 0.0004744204428514658, 'samples': 1722816, 'steps': 8972, 'loss/train': 0.6708272695541382} 01/29/2022 00:22:18 - INFO - codeparrot_training - Step 8973: {'lr': 0.0004744132323445908, 'samples': 1723008, 'steps': 8973, 'loss/train': 1.230182558298111} 01/29/2022 00:22:22 - INFO - codeparrot_training - Step 8974: {'lr': 0.00047440602087640084, 'samples': 1723200, 'steps': 8974, 'loss/train': 2.688316583633423} 01/29/2022 00:22:27 - INFO - codeparrot_training - Step 8975: {'lr': 0.0004743988084469267, 'samples': 1723392, 'steps': 8975, 'loss/train': 2.377837836742401} 01/29/2022 00:22:32 - INFO - codeparrot_training - Step 8976: {'lr': 0.00047439159505619936, 'samples': 1723584, 'steps': 8976, 'loss/train': 2.0049986243247986} 01/29/2022 00:22:36 - INFO - codeparrot_training - Step 8977: {'lr': 0.0004743843807042497, 'samples': 1723776, 'steps': 8977, 'loss/train': 2.280084192752838} 01/29/2022 00:22:40 - INFO - codeparrot_training - Step 8978: {'lr': 0.0004743771653911086, 'samples': 1723968, 'steps': 8978, 'loss/train': 2.4175254106521606} 01/29/2022 00:22:44 - INFO - codeparrot_training - Step 8979: {'lr': 0.00047436994911680694, 'samples': 1724160, 'steps': 8979, 'loss/train': 1.7717489004135132} 01/29/2022 00:22:50 - INFO - codeparrot_training - Step 8980: {'lr': 0.0004743627318813757, 'samples': 1724352, 'steps': 8980, 'loss/train': 2.358184039592743} 01/29/2022 00:22:54 - INFO - codeparrot_training - Step 8981: {'lr': 0.00047435551368484567, 'samples': 1724544, 'steps': 8981, 'loss/train': 2.950459063053131} 01/29/2022 00:22:58 - INFO - codeparrot_training - Step 8982: {'lr': 0.00047434829452724795, 'samples': 1724736, 'steps': 8982, 'loss/train': 1.5545448660850525} 01/29/2022 00:23:02 - INFO - codeparrot_training - Step 8983: {'lr': 0.00047434107440861336, 'samples': 1724928, 'steps': 8983, 'loss/train': 1.9959191679954529} 01/29/2022 00:23:06 - INFO - codeparrot_training - Step 8984: {'lr': 0.0004743338533289728, 'samples': 1725120, 'steps': 8984, 'loss/train': 1.8698923587799072} 01/29/2022 00:23:13 - INFO - codeparrot_training - Step 8985: {'lr': 0.00047432663128835727, 'samples': 1725312, 'steps': 8985, 'loss/train': 1.7949918508529663} 01/29/2022 00:23:17 - INFO - codeparrot_training - Step 8986: {'lr': 0.0004743194082867977, 'samples': 1725504, 'steps': 8986, 'loss/train': 1.8081718683242798} 01/29/2022 00:23:21 - INFO - codeparrot_training - Step 8987: {'lr': 0.000474312184324325, 'samples': 1725696, 'steps': 8987, 'loss/train': 2.294632315635681} 01/29/2022 00:23:25 - INFO - codeparrot_training - Step 8988: {'lr': 0.0004743049594009701, 'samples': 1725888, 'steps': 8988, 'loss/train': 1.7149266600608826} 01/29/2022 00:23:30 - INFO - codeparrot_training - Step 8989: {'lr': 0.0004742977335167641, 'samples': 1726080, 'steps': 8989, 'loss/train': 1.3053812384605408} 01/29/2022 00:23:34 - INFO - codeparrot_training - Step 8990: {'lr': 0.0004742905066717377, 'samples': 1726272, 'steps': 8990, 'loss/train': 0.8518482148647308} 01/29/2022 00:23:39 - INFO - codeparrot_training - Step 8991: {'lr': 0.00047428327886592204, 'samples': 1726464, 'steps': 8991, 'loss/train': 1.9660577774047852} 01/29/2022 00:23:44 - INFO - codeparrot_training - Step 8992: {'lr': 0.00047427605009934805, 'samples': 1726656, 'steps': 8992, 'loss/train': 1.965314269065857} 01/29/2022 00:23:48 - INFO - codeparrot_training - Step 8993: {'lr': 0.00047426882037204663, 'samples': 1726848, 'steps': 8993, 'loss/train': 1.6179476380348206} 01/29/2022 00:23:52 - INFO - codeparrot_training - Step 8994: {'lr': 0.0004742615896840488, 'samples': 1727040, 'steps': 8994, 'loss/train': 1.9801486730575562} 01/29/2022 00:23:57 - INFO - codeparrot_training - Step 8995: {'lr': 0.00047425435803538554, 'samples': 1727232, 'steps': 8995, 'loss/train': 1.7462589740753174} 01/29/2022 00:24:02 - INFO - codeparrot_training - Step 8996: {'lr': 0.0004742471254260878, 'samples': 1727424, 'steps': 8996, 'loss/train': 2.4691008925437927} 01/29/2022 00:24:06 - INFO - codeparrot_training - Step 8997: {'lr': 0.00047423989185618666, 'samples': 1727616, 'steps': 8997, 'loss/train': 1.522091567516327} 01/29/2022 00:24:10 - INFO - codeparrot_training - Step 8998: {'lr': 0.00047423265732571295, 'samples': 1727808, 'steps': 8998, 'loss/train': 1.8721379041671753} 01/29/2022 00:24:14 - INFO - codeparrot_training - Step 8999: {'lr': 0.00047422542183469775, 'samples': 1728000, 'steps': 8999, 'loss/train': 2.0316402912139893} 01/29/2022 00:24:20 - INFO - codeparrot_training - Step 9000: {'lr': 0.0004742181853831721, 'samples': 1728192, 'steps': 9000, 'loss/train': 1.5273796319961548} 01/29/2022 00:24:25 - INFO - codeparrot_training - Step 9001: {'lr': 0.00047421094797116687, 'samples': 1728384, 'steps': 9001, 'loss/train': 2.6366100311279297} 01/29/2022 00:24:29 - INFO - codeparrot_training - Step 9002: {'lr': 0.00047420370959871315, 'samples': 1728576, 'steps': 9002, 'loss/train': 1.8127264380455017} 01/29/2022 00:24:33 - INFO - codeparrot_training - Step 9003: {'lr': 0.000474196470265842, 'samples': 1728768, 'steps': 9003, 'loss/train': 1.6343028545379639} 01/29/2022 00:24:37 - INFO - codeparrot_training - Step 9004: {'lr': 0.0004741892299725843, 'samples': 1728960, 'steps': 9004, 'loss/train': 1.6700883507728577} 01/29/2022 00:24:43 - INFO - codeparrot_training - Step 9005: {'lr': 0.0004741819887189711, 'samples': 1729152, 'steps': 9005, 'loss/train': 1.7710556387901306} 01/29/2022 00:24:47 - INFO - codeparrot_training - Step 9006: {'lr': 0.00047417474650503347, 'samples': 1729344, 'steps': 9006, 'loss/train': 1.2632234394550323} 01/29/2022 00:24:51 - INFO - codeparrot_training - Step 9007: {'lr': 0.00047416750333080244, 'samples': 1729536, 'steps': 9007, 'loss/train': 1.9305499792099} 01/29/2022 00:24:55 - INFO - codeparrot_training - Step 9008: {'lr': 0.000474160259196309, 'samples': 1729728, 'steps': 9008, 'loss/train': 1.4694810211658478} 01/29/2022 00:25:00 - INFO - codeparrot_training - Step 9009: {'lr': 0.00047415301410158416, 'samples': 1729920, 'steps': 9009, 'loss/train': 1.5341708064079285} 01/29/2022 00:25:05 - INFO - codeparrot_training - Step 9010: {'lr': 0.00047414576804665897, 'samples': 1730112, 'steps': 9010, 'loss/train': 1.459893822669983} 01/29/2022 00:25:09 - INFO - codeparrot_training - Step 9011: {'lr': 0.0004741385210315645, 'samples': 1730304, 'steps': 9011, 'loss/train': 1.9215111136436462} 01/29/2022 00:25:14 - INFO - codeparrot_training - Step 9012: {'lr': 0.0004741312730563318, 'samples': 1730496, 'steps': 9012, 'loss/train': 1.3990541696548462} 01/29/2022 00:25:18 - INFO - codeparrot_training - Step 9013: {'lr': 0.00047412402412099185, 'samples': 1730688, 'steps': 9013, 'loss/train': 1.8167295455932617} 01/29/2022 00:25:22 - INFO - codeparrot_training - Step 9014: {'lr': 0.00047411677422557586, 'samples': 1730880, 'steps': 9014, 'loss/train': 2.0193400382995605} 01/29/2022 00:25:28 - INFO - codeparrot_training - Step 9015: {'lr': 0.0004741095233701147, 'samples': 1731072, 'steps': 9015, 'loss/train': 2.1063496470451355} 01/29/2022 00:25:32 - INFO - codeparrot_training - Step 9016: {'lr': 0.00047410227155463946, 'samples': 1731264, 'steps': 9016, 'loss/train': 1.5116860270500183} 01/29/2022 00:25:36 - INFO - codeparrot_training - Step 9017: {'lr': 0.00047409501877918134, 'samples': 1731456, 'steps': 9017, 'loss/train': 1.6822602152824402} 01/29/2022 00:25:41 - INFO - codeparrot_training - Step 9018: {'lr': 0.00047408776504377127, 'samples': 1731648, 'steps': 9018, 'loss/train': 1.4372069835662842} 01/29/2022 00:25:45 - INFO - codeparrot_training - Step 9019: {'lr': 0.00047408051034844036, 'samples': 1731840, 'steps': 9019, 'loss/train': 1.663207232952118} 01/29/2022 00:25:50 - INFO - codeparrot_training - Step 9020: {'lr': 0.00047407325469321973, 'samples': 1732032, 'steps': 9020, 'loss/train': 2.271194636821747} 01/29/2022 00:25:55 - INFO - codeparrot_training - Step 9021: {'lr': 0.00047406599807814034, 'samples': 1732224, 'steps': 9021, 'loss/train': 1.0223395228385925} 01/29/2022 00:25:59 - INFO - codeparrot_training - Step 9022: {'lr': 0.00047405874050323346, 'samples': 1732416, 'steps': 9022, 'loss/train': 0.8626428544521332} 01/29/2022 00:26:03 - INFO - codeparrot_training - Step 9023: {'lr': 0.00047405148196853005, 'samples': 1732608, 'steps': 9023, 'loss/train': 1.7820094227790833} 01/29/2022 00:26:07 - INFO - codeparrot_training - Step 9024: {'lr': 0.0004740442224740612, 'samples': 1732800, 'steps': 9024, 'loss/train': 1.4607646465301514} 01/29/2022 00:26:13 - INFO - codeparrot_training - Step 9025: {'lr': 0.00047403696201985814, 'samples': 1732992, 'steps': 9025, 'loss/train': 2.0196295380592346} 01/29/2022 00:26:17 - INFO - codeparrot_training - Step 9026: {'lr': 0.0004740297006059517, 'samples': 1733184, 'steps': 9026, 'loss/train': 1.8399091958999634} 01/29/2022 00:26:21 - INFO - codeparrot_training - Step 9027: {'lr': 0.00047402243823237335, 'samples': 1733376, 'steps': 9027, 'loss/train': 0.4233810603618622} 01/29/2022 00:26:25 - INFO - codeparrot_training - Step 9028: {'lr': 0.0004740151748991539, 'samples': 1733568, 'steps': 9028, 'loss/train': 2.5047767758369446} 01/29/2022 00:26:30 - INFO - codeparrot_training - Step 9029: {'lr': 0.00047400791060632464, 'samples': 1733760, 'steps': 9029, 'loss/train': 1.369269222021103} 01/29/2022 00:26:36 - INFO - codeparrot_training - Step 9030: {'lr': 0.0004740006453539166, 'samples': 1733952, 'steps': 9030, 'loss/train': 2.394858419895172} 01/29/2022 00:26:40 - INFO - codeparrot_training - Step 9031: {'lr': 0.0004739933791419609, 'samples': 1734144, 'steps': 9031, 'loss/train': 1.9164068698883057} 01/29/2022 00:26:44 - INFO - codeparrot_training - Step 9032: {'lr': 0.0004739861119704887, 'samples': 1734336, 'steps': 9032, 'loss/train': 1.0405627191066742} 01/29/2022 00:26:48 - INFO - codeparrot_training - Step 9033: {'lr': 0.00047397884383953114, 'samples': 1734528, 'steps': 9033, 'loss/train': 1.717029333114624} 01/29/2022 00:26:53 - INFO - codeparrot_training - Step 9034: {'lr': 0.0004739715747491193, 'samples': 1734720, 'steps': 9034, 'loss/train': 0.7171564400196075} 01/29/2022 00:26:58 - INFO - codeparrot_training - Step 9035: {'lr': 0.00047396430469928436, 'samples': 1734912, 'steps': 9035, 'loss/train': 1.9788413643836975} 01/29/2022 00:27:02 - INFO - codeparrot_training - Step 9036: {'lr': 0.0004739570336900575, 'samples': 1735104, 'steps': 9036, 'loss/train': 2.012026011943817} 01/29/2022 00:27:06 - INFO - codeparrot_training - Step 9037: {'lr': 0.00047394976172146974, 'samples': 1735296, 'steps': 9037, 'loss/train': 1.9961573481559753} 01/29/2022 00:27:10 - INFO - codeparrot_training - Step 9038: {'lr': 0.0004739424887935524, 'samples': 1735488, 'steps': 9038, 'loss/train': 0.8543922901153564} 01/29/2022 00:27:15 - INFO - codeparrot_training - Step 9039: {'lr': 0.0004739352149063365, 'samples': 1735680, 'steps': 9039, 'loss/train': 1.968008816242218} 01/29/2022 00:27:21 - INFO - codeparrot_training - Step 9040: {'lr': 0.0004739279400598532, 'samples': 1735872, 'steps': 9040, 'loss/train': 1.9744908213615417} 01/29/2022 00:27:25 - INFO - codeparrot_training - Step 9041: {'lr': 0.0004739206642541338, 'samples': 1736064, 'steps': 9041, 'loss/train': 1.714539885520935} 01/29/2022 00:27:29 - INFO - codeparrot_training - Step 9042: {'lr': 0.0004739133874892093, 'samples': 1736256, 'steps': 9042, 'loss/train': 1.4468913674354553} 01/29/2022 00:27:33 - INFO - codeparrot_training - Step 9043: {'lr': 0.0004739061097651111, 'samples': 1736448, 'steps': 9043, 'loss/train': 2.098978042602539} 01/29/2022 00:27:38 - INFO - codeparrot_training - Step 9044: {'lr': 0.00047389883108187004, 'samples': 1736640, 'steps': 9044, 'loss/train': 1.3569864928722382} 01/29/2022 00:27:43 - INFO - codeparrot_training - Step 9045: {'lr': 0.0004738915514395176, 'samples': 1736832, 'steps': 9045, 'loss/train': 1.6922484040260315} 01/29/2022 00:27:47 - INFO - codeparrot_training - Step 9046: {'lr': 0.0004738842708380847, 'samples': 1737024, 'steps': 9046, 'loss/train': 1.2878260016441345} 01/29/2022 00:27:51 - INFO - codeparrot_training - Step 9047: {'lr': 0.0004738769892776028, 'samples': 1737216, 'steps': 9047, 'loss/train': 1.7181095480918884} 01/29/2022 00:27:55 - INFO - codeparrot_training - Step 9048: {'lr': 0.00047386970675810297, 'samples': 1737408, 'steps': 9048, 'loss/train': 1.0245516300201416} 01/29/2022 00:28:00 - INFO - codeparrot_training - Step 9049: {'lr': 0.00047386242327961635, 'samples': 1737600, 'steps': 9049, 'loss/train': 1.910882592201233} 01/29/2022 00:28:05 - INFO - codeparrot_training - Step 9050: {'lr': 0.0004738551388421742, 'samples': 1737792, 'steps': 9050, 'loss/train': 2.213915705680847} 01/29/2022 00:28:10 - INFO - codeparrot_training - Step 9051: {'lr': 0.00047384785344580784, 'samples': 1737984, 'steps': 9051, 'loss/train': 2.073532819747925} 01/29/2022 00:28:14 - INFO - codeparrot_training - Step 9052: {'lr': 0.00047384056709054824, 'samples': 1738176, 'steps': 9052, 'loss/train': 2.0055609941482544} 01/29/2022 00:28:18 - INFO - codeparrot_training - Step 9053: {'lr': 0.0004738332797764267, 'samples': 1738368, 'steps': 9053, 'loss/train': 2.285977363586426} 01/29/2022 00:28:24 - INFO - codeparrot_training - Step 9054: {'lr': 0.0004738259915034745, 'samples': 1738560, 'steps': 9054, 'loss/train': 1.6939862966537476} 01/29/2022 00:28:28 - INFO - codeparrot_training - Step 9055: {'lr': 0.00047381870227172285, 'samples': 1738752, 'steps': 9055, 'loss/train': 0.99030402302742} 01/29/2022 00:28:32 - INFO - codeparrot_training - Step 9056: {'lr': 0.0004738114120812029, 'samples': 1738944, 'steps': 9056, 'loss/train': 2.597422957420349} 01/29/2022 00:28:36 - INFO - codeparrot_training - Step 9057: {'lr': 0.000473804120931946, 'samples': 1739136, 'steps': 9057, 'loss/train': 1.7573168277740479} 01/29/2022 00:28:41 - INFO - codeparrot_training - Step 9058: {'lr': 0.0004737968288239832, 'samples': 1739328, 'steps': 9058, 'loss/train': 1.3731518983840942} 01/29/2022 00:28:47 - INFO - codeparrot_training - Step 9059: {'lr': 0.00047378953575734594, 'samples': 1739520, 'steps': 9059, 'loss/train': 2.4543278217315674} 01/29/2022 00:28:51 - INFO - codeparrot_training - Step 9060: {'lr': 0.0004737822417320654, 'samples': 1739712, 'steps': 9060, 'loss/train': 1.491113305091858} 01/29/2022 00:28:55 - INFO - codeparrot_training - Step 9061: {'lr': 0.00047377494674817275, 'samples': 1739904, 'steps': 9061, 'loss/train': 1.7257318496704102} 01/29/2022 00:29:00 - INFO - codeparrot_training - Step 9062: {'lr': 0.00047376765080569925, 'samples': 1740096, 'steps': 9062, 'loss/train': 2.37028831243515} 01/29/2022 00:29:04 - INFO - codeparrot_training - Step 9063: {'lr': 0.0004737603539046762, 'samples': 1740288, 'steps': 9063, 'loss/train': 1.9773842096328735} 01/29/2022 00:29:09 - INFO - codeparrot_training - Step 9064: {'lr': 0.0004737530560451349, 'samples': 1740480, 'steps': 9064, 'loss/train': 1.1166025400161743} 01/29/2022 00:29:13 - INFO - codeparrot_training - Step 9065: {'lr': 0.00047374575722710656, 'samples': 1740672, 'steps': 9065, 'loss/train': 1.8429597616195679} 01/29/2022 00:29:18 - INFO - codeparrot_training - Step 9066: {'lr': 0.0004737384574506224, 'samples': 1740864, 'steps': 9066, 'loss/train': 1.944546103477478} 01/29/2022 00:29:22 - INFO - codeparrot_training - Step 9067: {'lr': 0.0004737311567157137, 'samples': 1741056, 'steps': 9067, 'loss/train': 1.780111849308014} 01/29/2022 00:29:26 - INFO - codeparrot_training - Step 9068: {'lr': 0.00047372385502241176, 'samples': 1741248, 'steps': 9068, 'loss/train': 1.7093557119369507} 01/29/2022 00:29:31 - INFO - codeparrot_training - Step 9069: {'lr': 0.00047371655237074794, 'samples': 1741440, 'steps': 9069, 'loss/train': 1.8804463148117065} 01/29/2022 00:29:36 - INFO - codeparrot_training - Step 9070: {'lr': 0.0004737092487607534, 'samples': 1741632, 'steps': 9070, 'loss/train': 0.43390703201293945} 01/29/2022 00:29:40 - INFO - codeparrot_training - Step 9071: {'lr': 0.00047370194419245955, 'samples': 1741824, 'steps': 9071, 'loss/train': 1.6919741034507751} 01/29/2022 00:29:44 - INFO - codeparrot_training - Step 9072: {'lr': 0.00047369463866589755, 'samples': 1742016, 'steps': 9072, 'loss/train': 1.8644896745681763} 01/29/2022 00:29:48 - INFO - codeparrot_training - Step 9073: {'lr': 0.00047368733218109874, 'samples': 1742208, 'steps': 9073, 'loss/train': 1.8182523250579834} 01/29/2022 00:29:54 - INFO - codeparrot_training - Step 9074: {'lr': 0.00047368002473809447, 'samples': 1742400, 'steps': 9074, 'loss/train': 1.6522157192230225} 01/29/2022 00:29:59 - INFO - codeparrot_training - Step 9075: {'lr': 0.0004736727163369159, 'samples': 1742592, 'steps': 9075, 'loss/train': 2.445735812187195} 01/29/2022 00:30:03 - INFO - codeparrot_training - Step 9076: {'lr': 0.00047366540697759454, 'samples': 1742784, 'steps': 9076, 'loss/train': 2.354703426361084} 01/29/2022 00:30:07 - INFO - codeparrot_training - Step 9077: {'lr': 0.00047365809666016155, 'samples': 1742976, 'steps': 9077, 'loss/train': 0.586450606584549} 01/29/2022 00:30:11 - INFO - codeparrot_training - Step 9078: {'lr': 0.00047365078538464826, 'samples': 1743168, 'steps': 9078, 'loss/train': 1.6023826003074646} 01/29/2022 00:30:17 - INFO - codeparrot_training - Step 9079: {'lr': 0.0004736434731510861, 'samples': 1743360, 'steps': 9079, 'loss/train': 2.2399930357933044} 01/29/2022 00:30:21 - INFO - codeparrot_training - Step 9080: {'lr': 0.00047363615995950624, 'samples': 1743552, 'steps': 9080, 'loss/train': 1.2977181673049927} 01/29/2022 00:30:25 - INFO - codeparrot_training - Step 9081: {'lr': 0.0004736288458099401, 'samples': 1743744, 'steps': 9081, 'loss/train': 2.1770262122154236} 01/29/2022 00:30:30 - INFO - codeparrot_training - Step 9082: {'lr': 0.0004736215307024191, 'samples': 1743936, 'steps': 9082, 'loss/train': 1.9061638712882996} 01/29/2022 00:30:34 - INFO - codeparrot_training - Step 9083: {'lr': 0.0004736142146369744, 'samples': 1744128, 'steps': 9083, 'loss/train': 2.4274760484695435} 01/29/2022 00:30:40 - INFO - codeparrot_training - Step 9084: {'lr': 0.0004736068976136374, 'samples': 1744320, 'steps': 9084, 'loss/train': 2.1461223363876343} 01/29/2022 00:30:44 - INFO - codeparrot_training - Step 9085: {'lr': 0.00047359957963243943, 'samples': 1744512, 'steps': 9085, 'loss/train': 1.5628867149353027} 01/29/2022 00:30:48 - INFO - codeparrot_training - Step 9086: {'lr': 0.0004735922606934119, 'samples': 1744704, 'steps': 9086, 'loss/train': 1.6171472668647766} 01/29/2022 00:30:53 - INFO - codeparrot_training - Step 9087: {'lr': 0.0004735849407965861, 'samples': 1744896, 'steps': 9087, 'loss/train': 1.5456052422523499} 01/29/2022 00:30:57 - INFO - codeparrot_training - Step 9088: {'lr': 0.00047357761994199345, 'samples': 1745088, 'steps': 9088, 'loss/train': 2.0332286953926086} 01/29/2022 00:31:02 - INFO - codeparrot_training - Step 9089: {'lr': 0.00047357029812966525, 'samples': 1745280, 'steps': 9089, 'loss/train': 0.13962693884968758} 01/29/2022 00:31:06 - INFO - codeparrot_training - Step 9090: {'lr': 0.0004735629753596328, 'samples': 1745472, 'steps': 9090, 'loss/train': 1.8044849038124084} 01/29/2022 00:31:11 - INFO - codeparrot_training - Step 9091: {'lr': 0.00047355565163192763, 'samples': 1745664, 'steps': 9091, 'loss/train': 2.024993598461151} 01/29/2022 00:31:15 - INFO - codeparrot_training - Step 9092: {'lr': 0.00047354832694658104, 'samples': 1745856, 'steps': 9092, 'loss/train': 1.167715698480606} 01/29/2022 00:31:19 - INFO - codeparrot_training - Step 9093: {'lr': 0.00047354100130362443, 'samples': 1746048, 'steps': 9093, 'loss/train': 1.8007004857063293} 01/29/2022 00:31:24 - INFO - codeparrot_training - Step 9094: {'lr': 0.00047353367470308913, 'samples': 1746240, 'steps': 9094, 'loss/train': 1.6633154153823853} 01/29/2022 00:31:29 - INFO - codeparrot_training - Step 9095: {'lr': 0.0004735263471450065, 'samples': 1746432, 'steps': 9095, 'loss/train': 1.9684236645698547} 01/29/2022 00:31:33 - INFO - codeparrot_training - Step 9096: {'lr': 0.00047351901862940807, 'samples': 1746624, 'steps': 9096, 'loss/train': 0.9972791969776154} 01/29/2022 00:31:37 - INFO - codeparrot_training - Step 9097: {'lr': 0.000473511689156325, 'samples': 1746816, 'steps': 9097, 'loss/train': 1.8450114727020264} 01/29/2022 00:31:41 - INFO - codeparrot_training - Step 9098: {'lr': 0.0004735043587257889, 'samples': 1747008, 'steps': 9098, 'loss/train': 2.243839681148529} 01/29/2022 00:31:47 - INFO - codeparrot_training - Step 9099: {'lr': 0.00047349702733783113, 'samples': 1747200, 'steps': 9099, 'loss/train': 0.21277771890163422} 01/29/2022 00:31:52 - INFO - codeparrot_training - Step 9100: {'lr': 0.00047348969499248306, 'samples': 1747392, 'steps': 9100, 'loss/train': 1.4239500761032104} 01/29/2022 00:31:56 - INFO - codeparrot_training - Step 9101: {'lr': 0.0004734823616897761, 'samples': 1747584, 'steps': 9101, 'loss/train': 1.9856343269348145} 01/29/2022 00:32:00 - INFO - codeparrot_training - Step 9102: {'lr': 0.0004734750274297416, 'samples': 1747776, 'steps': 9102, 'loss/train': 0.5345341265201569} 01/29/2022 00:32:04 - INFO - codeparrot_training - Step 9103: {'lr': 0.0004734676922124111, 'samples': 1747968, 'steps': 9103, 'loss/train': 1.575616478919983} 01/29/2022 00:32:10 - INFO - codeparrot_training - Step 9104: {'lr': 0.00047346035603781597, 'samples': 1748160, 'steps': 9104, 'loss/train': 0.9906434118747711} 01/29/2022 00:32:14 - INFO - codeparrot_training - Step 9105: {'lr': 0.0004734530189059876, 'samples': 1748352, 'steps': 9105, 'loss/train': 1.873069167137146} 01/29/2022 00:32:18 - INFO - codeparrot_training - Step 9106: {'lr': 0.0004734456808169575, 'samples': 1748544, 'steps': 9106, 'loss/train': 1.3482929170131683} 01/29/2022 00:32:23 - INFO - codeparrot_training - Step 9107: {'lr': 0.00047343834177075695, 'samples': 1748736, 'steps': 9107, 'loss/train': 2.07370126247406} 01/29/2022 00:32:27 - INFO - codeparrot_training - Step 9108: {'lr': 0.0004734310017674176, 'samples': 1748928, 'steps': 9108, 'loss/train': 2.3167662620544434} 01/29/2022 00:32:32 - INFO - codeparrot_training - Step 9109: {'lr': 0.00047342366080697077, 'samples': 1749120, 'steps': 9109, 'loss/train': 2.6190810799598694} 01/29/2022 00:32:36 - INFO - codeparrot_training - Step 9110: {'lr': 0.00047341631888944794, 'samples': 1749312, 'steps': 9110, 'loss/train': 1.6951271295547485} 01/29/2022 00:32:41 - INFO - codeparrot_training - Step 9111: {'lr': 0.0004734089760148805, 'samples': 1749504, 'steps': 9111, 'loss/train': 2.336572051048279} 01/29/2022 00:32:45 - INFO - codeparrot_training - Step 9112: {'lr': 0.0004734016321832999, 'samples': 1749696, 'steps': 9112, 'loss/train': 1.4744910299777985} 01/29/2022 00:32:49 - INFO - codeparrot_training - Step 9113: {'lr': 0.0004733942873947377, 'samples': 1749888, 'steps': 9113, 'loss/train': 2.3771467208862305} 01/29/2022 00:32:55 - INFO - codeparrot_training - Step 9114: {'lr': 0.00047338694164922535, 'samples': 1750080, 'steps': 9114, 'loss/train': 1.3063930571079254} 01/29/2022 00:32:59 - INFO - codeparrot_training - Step 9115: {'lr': 0.0004733795949467942, 'samples': 1750272, 'steps': 9115, 'loss/train': 1.6264615058898926} 01/29/2022 00:33:03 - INFO - codeparrot_training - Step 9116: {'lr': 0.0004733722472874759, 'samples': 1750464, 'steps': 9116, 'loss/train': 1.5908979177474976} 01/29/2022 00:33:07 - INFO - codeparrot_training - Step 9117: {'lr': 0.0004733648986713017, 'samples': 1750656, 'steps': 9117, 'loss/train': 2.563513219356537} 01/29/2022 00:33:11 - INFO - codeparrot_training - Step 9118: {'lr': 0.00047335754909830327, 'samples': 1750848, 'steps': 9118, 'loss/train': 0.6034076660871506} 01/29/2022 00:33:18 - INFO - codeparrot_training - Step 9119: {'lr': 0.00047335019856851204, 'samples': 1751040, 'steps': 9119, 'loss/train': 2.968452215194702} 01/29/2022 00:33:22 - INFO - codeparrot_training - Step 9120: {'lr': 0.0004733428470819594, 'samples': 1751232, 'steps': 9120, 'loss/train': 1.2858002185821533} 01/29/2022 00:33:26 - INFO - codeparrot_training - Step 9121: {'lr': 0.000473335494638677, 'samples': 1751424, 'steps': 9121, 'loss/train': 1.0978128612041473} 01/29/2022 00:33:30 - INFO - codeparrot_training - Step 9122: {'lr': 0.00047332814123869616, 'samples': 1751616, 'steps': 9122, 'loss/train': 2.3364819288253784} 01/29/2022 00:33:35 - INFO - codeparrot_training - Step 9123: {'lr': 0.0004733207868820486, 'samples': 1751808, 'steps': 9123, 'loss/train': 1.6441399455070496} 01/29/2022 00:33:40 - INFO - codeparrot_training - Step 9124: {'lr': 0.0004733134315687656, 'samples': 1752000, 'steps': 9124, 'loss/train': 0.6640878170728683} 01/29/2022 00:33:45 - INFO - codeparrot_training - Step 9125: {'lr': 0.00047330607529887884, 'samples': 1752192, 'steps': 9125, 'loss/train': 0.07220271416008472} 01/29/2022 00:33:49 - INFO - codeparrot_training - Step 9126: {'lr': 0.00047329871807241976, 'samples': 1752384, 'steps': 9126, 'loss/train': 2.8685068488121033} 01/29/2022 00:33:53 - INFO - codeparrot_training - Step 9127: {'lr': 0.00047329135988941984, 'samples': 1752576, 'steps': 9127, 'loss/train': 1.9541775584220886} 01/29/2022 00:33:57 - INFO - codeparrot_training - Step 9128: {'lr': 0.00047328400074991064, 'samples': 1752768, 'steps': 9128, 'loss/train': 1.8743024468421936} 01/29/2022 00:34:03 - INFO - codeparrot_training - Step 9129: {'lr': 0.00047327664065392375, 'samples': 1752960, 'steps': 9129, 'loss/train': 2.2488672137260437} 01/29/2022 00:34:07 - INFO - codeparrot_training - Step 9130: {'lr': 0.0004732692796014905, 'samples': 1753152, 'steps': 9130, 'loss/train': 1.2727356255054474} 01/29/2022 00:34:11 - INFO - codeparrot_training - Step 9131: {'lr': 0.00047326191759264265, 'samples': 1753344, 'steps': 9131, 'loss/train': 2.159435749053955} 01/29/2022 00:34:15 - INFO - codeparrot_training - Step 9132: {'lr': 0.00047325455462741164, 'samples': 1753536, 'steps': 9132, 'loss/train': 1.5059775710105896} 01/29/2022 00:34:20 - INFO - codeparrot_training - Step 9133: {'lr': 0.00047324719070582894, 'samples': 1753728, 'steps': 9133, 'loss/train': 1.4830310940742493} 01/29/2022 00:34:26 - INFO - codeparrot_training - Step 9134: {'lr': 0.00047323982582792625, 'samples': 1753920, 'steps': 9134, 'loss/train': 1.6352198123931885} 01/29/2022 00:34:30 - INFO - codeparrot_training - Step 9135: {'lr': 0.00047323245999373497, 'samples': 1754112, 'steps': 9135, 'loss/train': 1.2777871191501617} 01/29/2022 00:34:34 - INFO - codeparrot_training - Step 9136: {'lr': 0.0004732250932032867, 'samples': 1754304, 'steps': 9136, 'loss/train': 2.3021791577339172} 01/29/2022 00:34:39 - INFO - codeparrot_training - Step 9137: {'lr': 0.0004732177254566131, 'samples': 1754496, 'steps': 9137, 'loss/train': 2.080966293811798} 01/29/2022 00:34:44 - INFO - codeparrot_training - Step 9138: {'lr': 0.0004732103567537456, 'samples': 1754688, 'steps': 9138, 'loss/train': 2.2410505414009094} 01/29/2022 00:34:48 - INFO - codeparrot_training - Step 9139: {'lr': 0.00047320298709471574, 'samples': 1754880, 'steps': 9139, 'loss/train': 1.93620765209198} 01/29/2022 00:34:52 - INFO - codeparrot_training - Step 9140: {'lr': 0.0004731956164795552, 'samples': 1755072, 'steps': 9140, 'loss/train': 2.0757755041122437} 01/29/2022 00:34:57 - INFO - codeparrot_training - Step 9141: {'lr': 0.0004731882449082956, 'samples': 1755264, 'steps': 9141, 'loss/train': 1.845704197883606} 01/29/2022 00:35:01 - INFO - codeparrot_training - Step 9142: {'lr': 0.0004731808723809683, 'samples': 1755456, 'steps': 9142, 'loss/train': 1.6592318415641785} 01/29/2022 00:35:07 - INFO - codeparrot_training - Step 9143: {'lr': 0.0004731734988976051, 'samples': 1755648, 'steps': 9143, 'loss/train': 0.8035163283348083} 01/29/2022 00:35:12 - INFO - codeparrot_training - Step 9144: {'lr': 0.00047316612445823746, 'samples': 1755840, 'steps': 9144, 'loss/train': 2.08003306388855} 01/29/2022 00:35:16 - INFO - codeparrot_training - Step 9145: {'lr': 0.000473158749062897, 'samples': 1756032, 'steps': 9145, 'loss/train': 0.7901073396205902} 01/29/2022 00:35:20 - INFO - codeparrot_training - Step 9146: {'lr': 0.00047315137271161537, 'samples': 1756224, 'steps': 9146, 'loss/train': 1.5298749804496765} 01/29/2022 00:35:24 - INFO - codeparrot_training - Step 9147: {'lr': 0.00047314399540442407, 'samples': 1756416, 'steps': 9147, 'loss/train': 1.3833662867546082} 01/29/2022 00:35:30 - INFO - codeparrot_training - Step 9148: {'lr': 0.00047313661714135476, 'samples': 1756608, 'steps': 9148, 'loss/train': 1.1431147456169128} 01/29/2022 00:35:34 - INFO - codeparrot_training - Step 9149: {'lr': 0.000473129237922439, 'samples': 1756800, 'steps': 9149, 'loss/train': 1.8677982687950134} 01/29/2022 00:35:38 - INFO - codeparrot_training - Step 9150: {'lr': 0.0004731218577477085, 'samples': 1756992, 'steps': 9150, 'loss/train': 1.6898367404937744} 01/29/2022 00:35:42 - INFO - codeparrot_training - Step 9151: {'lr': 0.0004731144766171948, 'samples': 1757184, 'steps': 9151, 'loss/train': 1.9860327243804932} 01/29/2022 00:35:47 - INFO - codeparrot_training - Step 9152: {'lr': 0.0004731070945309295, 'samples': 1757376, 'steps': 9152, 'loss/train': 1.8382059931755066} 01/29/2022 00:35:52 - INFO - codeparrot_training - Step 9153: {'lr': 0.00047309971148894425, 'samples': 1757568, 'steps': 9153, 'loss/train': 1.06614750623703} 01/29/2022 00:35:56 - INFO - codeparrot_training - Step 9154: {'lr': 0.00047309232749127074, 'samples': 1757760, 'steps': 9154, 'loss/train': 2.1482319831848145} 01/29/2022 00:36:00 - INFO - codeparrot_training - Step 9155: {'lr': 0.0004730849425379404, 'samples': 1757952, 'steps': 9155, 'loss/train': 2.388911247253418} 01/29/2022 00:36:05 - INFO - codeparrot_training - Step 9156: {'lr': 0.0004730775566289851, 'samples': 1758144, 'steps': 9156, 'loss/train': 1.0613180100917816} 01/29/2022 00:36:09 - INFO - codeparrot_training - Step 9157: {'lr': 0.0004730701697644364, 'samples': 1758336, 'steps': 9157, 'loss/train': 2.1121609210968018} 01/29/2022 00:36:15 - INFO - codeparrot_training - Step 9158: {'lr': 0.00047306278194432597, 'samples': 1758528, 'steps': 9158, 'loss/train': 2.7043156027793884} 01/29/2022 00:36:19 - INFO - codeparrot_training - Step 9159: {'lr': 0.0004730553931686853, 'samples': 1758720, 'steps': 9159, 'loss/train': 1.9765974283218384} 01/29/2022 00:36:24 - INFO - codeparrot_training - Step 9160: {'lr': 0.00047304800343754615, 'samples': 1758912, 'steps': 9160, 'loss/train': 2.4750452041625977} 01/29/2022 00:36:28 - INFO - codeparrot_training - Step 9161: {'lr': 0.00047304061275094025, 'samples': 1759104, 'steps': 9161, 'loss/train': 1.7882591485977173} 01/29/2022 00:36:32 - INFO - codeparrot_training - Step 9162: {'lr': 0.0004730332211088992, 'samples': 1759296, 'steps': 9162, 'loss/train': 1.9887328147888184} 01/29/2022 00:36:37 - INFO - codeparrot_training - Step 9163: {'lr': 0.0004730258285114546, 'samples': 1759488, 'steps': 9163, 'loss/train': 1.0902144312858582} 01/29/2022 00:36:41 - INFO - codeparrot_training - Step 9164: {'lr': 0.0004730184349586382, 'samples': 1759680, 'steps': 9164, 'loss/train': 2.844996213912964} 01/29/2022 00:36:46 - INFO - codeparrot_training - Step 9165: {'lr': 0.0004730110404504816, 'samples': 1759872, 'steps': 9165, 'loss/train': 1.6465251445770264} 01/29/2022 00:36:50 - INFO - codeparrot_training - Step 9166: {'lr': 0.00047300364498701654, 'samples': 1760064, 'steps': 9166, 'loss/train': 1.6248788237571716} 01/29/2022 00:36:54 - INFO - codeparrot_training - Step 9167: {'lr': 0.00047299624856827474, 'samples': 1760256, 'steps': 9167, 'loss/train': 2.480435371398926} 01/29/2022 00:37:00 - INFO - codeparrot_training - Step 9168: {'lr': 0.0004729888511942877, 'samples': 1760448, 'steps': 9168, 'loss/train': 2.0113797783851624} 01/29/2022 00:37:04 - INFO - codeparrot_training - Step 9169: {'lr': 0.0004729814528650873, 'samples': 1760640, 'steps': 9169, 'loss/train': 1.7659542560577393} 01/29/2022 00:37:08 - INFO - codeparrot_training - Step 9170: {'lr': 0.00047297405358070517, 'samples': 1760832, 'steps': 9170, 'loss/train': 1.9644368290901184} 01/29/2022 00:37:12 - INFO - codeparrot_training - Step 9171: {'lr': 0.00047296665334117295, 'samples': 1761024, 'steps': 9171, 'loss/train': 1.2640112042427063} 01/29/2022 00:37:17 - INFO - codeparrot_training - Step 9172: {'lr': 0.0004729592521465224, 'samples': 1761216, 'steps': 9172, 'loss/train': 1.7293973565101624} 01/29/2022 00:37:22 - INFO - codeparrot_training - Step 9173: {'lr': 0.00047295184999678524, 'samples': 1761408, 'steps': 9173, 'loss/train': 2.1278858184814453} 01/29/2022 00:37:26 - INFO - codeparrot_training - Step 9174: {'lr': 0.00047294444689199313, 'samples': 1761600, 'steps': 9174, 'loss/train': 2.4514934420585632} 01/29/2022 00:37:30 - INFO - codeparrot_training - Step 9175: {'lr': 0.0004729370428321778, 'samples': 1761792, 'steps': 9175, 'loss/train': 1.338348627090454} 01/29/2022 00:37:35 - INFO - codeparrot_training - Step 9176: {'lr': 0.000472929637817371, 'samples': 1761984, 'steps': 9176, 'loss/train': 1.8731927275657654} 01/29/2022 00:37:39 - INFO - codeparrot_training - Step 9177: {'lr': 0.0004729222318476044, 'samples': 1762176, 'steps': 9177, 'loss/train': 0.9213743805885315} 01/29/2022 00:37:45 - INFO - codeparrot_training - Step 9178: {'lr': 0.0004729148249229097, 'samples': 1762368, 'steps': 9178, 'loss/train': 1.9057372212409973} 01/29/2022 00:37:49 - INFO - codeparrot_training - Step 9179: {'lr': 0.0004729074170433187, 'samples': 1762560, 'steps': 9179, 'loss/train': 1.2083254158496857} 01/29/2022 00:37:53 - INFO - codeparrot_training - Step 9180: {'lr': 0.0004729000082088631, 'samples': 1762752, 'steps': 9180, 'loss/train': 1.0713548362255096} 01/29/2022 00:37:57 - INFO - codeparrot_training - Step 9181: {'lr': 0.0004728925984195748, 'samples': 1762944, 'steps': 9181, 'loss/train': 0.9069043099880219} 01/29/2022 00:38:02 - INFO - codeparrot_training - Step 9182: {'lr': 0.00047288518767548516, 'samples': 1763136, 'steps': 9182, 'loss/train': 1.5140864253044128} 01/29/2022 00:38:08 - INFO - codeparrot_training - Step 9183: {'lr': 0.0004728777759766263, 'samples': 1763328, 'steps': 9183, 'loss/train': 1.973143458366394} 01/29/2022 00:38:12 - INFO - codeparrot_training - Step 9184: {'lr': 0.00047287036332302967, 'samples': 1763520, 'steps': 9184, 'loss/train': 1.5618647933006287} 01/29/2022 00:38:16 - INFO - codeparrot_training - Step 9185: {'lr': 0.0004728629497147273, 'samples': 1763712, 'steps': 9185, 'loss/train': 1.4138829410076141} 01/29/2022 00:38:20 - INFO - codeparrot_training - Step 9186: {'lr': 0.00047285553515175077, 'samples': 1763904, 'steps': 9186, 'loss/train': 0.27308788150548935} 01/29/2022 00:38:25 - INFO - codeparrot_training - Step 9187: {'lr': 0.0004728481196341319, 'samples': 1764096, 'steps': 9187, 'loss/train': 0.8456195890903473} 01/29/2022 00:38:31 - INFO - codeparrot_training - Step 9188: {'lr': 0.0004728407031619025, 'samples': 1764288, 'steps': 9188, 'loss/train': 1.0389814674854279} 01/29/2022 00:38:35 - INFO - codeparrot_training - Step 9189: {'lr': 0.0004728332857350942, 'samples': 1764480, 'steps': 9189, 'loss/train': 1.5827288031578064} 01/29/2022 00:38:39 - INFO - codeparrot_training - Step 9190: {'lr': 0.00047282586735373887, 'samples': 1764672, 'steps': 9190, 'loss/train': 1.8741161227226257} 01/29/2022 00:38:43 - INFO - codeparrot_training - Step 9191: {'lr': 0.0004728184480178683, 'samples': 1764864, 'steps': 9191, 'loss/train': 1.7327109575271606} 01/29/2022 00:38:48 - INFO - codeparrot_training - Step 9192: {'lr': 0.00047281102772751425, 'samples': 1765056, 'steps': 9192, 'loss/train': 2.547459661960602} 01/29/2022 00:38:53 - INFO - codeparrot_training - Step 9193: {'lr': 0.0004728036064827086, 'samples': 1765248, 'steps': 9193, 'loss/train': 1.5323127508163452} 01/29/2022 00:38:58 - INFO - codeparrot_training - Step 9194: {'lr': 0.00047279618428348294, 'samples': 1765440, 'steps': 9194, 'loss/train': 1.776381254196167} 01/29/2022 00:39:02 - INFO - codeparrot_training - Step 9195: {'lr': 0.00047278876112986923, 'samples': 1765632, 'steps': 9195, 'loss/train': 2.122790515422821} 01/29/2022 00:39:06 - INFO - codeparrot_training - Step 9196: {'lr': 0.0004727813370218992, 'samples': 1765824, 'steps': 9196, 'loss/train': 1.7557817101478577} 01/29/2022 00:39:10 - INFO - codeparrot_training - Step 9197: {'lr': 0.00047277391195960463, 'samples': 1766016, 'steps': 9197, 'loss/train': 1.828341543674469} 01/29/2022 00:39:15 - INFO - codeparrot_training - Step 9198: {'lr': 0.00047276648594301733, 'samples': 1766208, 'steps': 9198, 'loss/train': 2.9691452980041504} 01/29/2022 00:39:20 - INFO - codeparrot_training - Step 9199: {'lr': 0.0004727590589721692, 'samples': 1766400, 'steps': 9199, 'loss/train': 2.4548470973968506} 01/29/2022 00:39:24 - INFO - codeparrot_training - Step 9200: {'lr': 0.00047275163104709196, 'samples': 1766592, 'steps': 9200, 'loss/train': 1.3719601929187775} 01/29/2022 00:39:28 - INFO - codeparrot_training - Step 9201: {'lr': 0.0004727442021678175, 'samples': 1766784, 'steps': 9201, 'loss/train': 2.353726029396057} 01/29/2022 00:39:32 - INFO - codeparrot_training - Step 9202: {'lr': 0.0004727367723343776, 'samples': 1766976, 'steps': 9202, 'loss/train': 0.2960490435361862} 01/29/2022 00:39:37 - INFO - codeparrot_training - Step 9203: {'lr': 0.0004727293415468041, 'samples': 1767168, 'steps': 9203, 'loss/train': 2.7354689240455627} 01/29/2022 00:39:43 - INFO - codeparrot_training - Step 9204: {'lr': 0.00047272190980512875, 'samples': 1767360, 'steps': 9204, 'loss/train': 1.808249294757843} 01/29/2022 00:39:47 - INFO - codeparrot_training - Step 9205: {'lr': 0.0004727144771093835, 'samples': 1767552, 'steps': 9205, 'loss/train': 2.973677694797516} 01/29/2022 00:39:51 - INFO - codeparrot_training - Step 9206: {'lr': 0.00047270704345960023, 'samples': 1767744, 'steps': 9206, 'loss/train': 1.816547691822052} 01/29/2022 00:39:55 - INFO - codeparrot_training - Step 9207: {'lr': 0.00047269960885581064, 'samples': 1767936, 'steps': 9207, 'loss/train': 2.252173125743866} 01/29/2022 00:40:00 - INFO - codeparrot_training - Step 9208: {'lr': 0.00047269217329804663, 'samples': 1768128, 'steps': 9208, 'loss/train': 2.430577039718628} 01/29/2022 00:40:05 - INFO - codeparrot_training - Step 9209: {'lr': 0.00047268473678634007, 'samples': 1768320, 'steps': 9209, 'loss/train': 2.1431697607040405} 01/29/2022 00:40:09 - INFO - codeparrot_training - Step 9210: {'lr': 0.00047267729932072284, 'samples': 1768512, 'steps': 9210, 'loss/train': 1.6573290824890137} 01/29/2022 00:40:13 - INFO - codeparrot_training - Step 9211: {'lr': 0.00047266986090122677, 'samples': 1768704, 'steps': 9211, 'loss/train': 1.8473426699638367} 01/29/2022 00:40:18 - INFO - codeparrot_training - Step 9212: {'lr': 0.0004726624215278836, 'samples': 1768896, 'steps': 9212, 'loss/train': 2.2602063417434692} 01/29/2022 00:40:22 - INFO - codeparrot_training - Step 9213: {'lr': 0.00047265498120072546, 'samples': 1769088, 'steps': 9213, 'loss/train': 1.4731852412223816} 01/29/2022 00:40:27 - INFO - codeparrot_training - Step 9214: {'lr': 0.00047264753991978404, 'samples': 1769280, 'steps': 9214, 'loss/train': 3.0883984565734863} 01/29/2022 00:40:31 - INFO - codeparrot_training - Step 9215: {'lr': 0.00047264009768509127, 'samples': 1769472, 'steps': 9215, 'loss/train': 1.5478785037994385} 01/29/2022 00:40:36 - INFO - codeparrot_training - Step 9216: {'lr': 0.000472632654496679, 'samples': 1769664, 'steps': 9216, 'loss/train': 2.1506820917129517} 01/29/2022 00:40:40 - INFO - codeparrot_training - Step 9217: {'lr': 0.00047262521035457914, 'samples': 1769856, 'steps': 9217, 'loss/train': 0.9972326159477234} 01/29/2022 00:40:44 - INFO - codeparrot_training - Step 9218: {'lr': 0.00047261776525882353, 'samples': 1770048, 'steps': 9218, 'loss/train': 1.797178030014038} 01/29/2022 00:40:49 - INFO - codeparrot_training - Step 9219: {'lr': 0.00047261031920944413, 'samples': 1770240, 'steps': 9219, 'loss/train': 2.3390859961509705} 01/29/2022 00:40:54 - INFO - codeparrot_training - Step 9220: {'lr': 0.0004726028722064728, 'samples': 1770432, 'steps': 9220, 'loss/train': 1.8034788966178894} 01/29/2022 00:40:58 - INFO - codeparrot_training - Step 9221: {'lr': 0.0004725954242499415, 'samples': 1770624, 'steps': 9221, 'loss/train': 2.2696362733840942} 01/29/2022 00:41:02 - INFO - codeparrot_training - Step 9222: {'lr': 0.00047258797533988205, 'samples': 1770816, 'steps': 9222, 'loss/train': 1.94929039478302} 01/29/2022 00:41:06 - INFO - codeparrot_training - Step 9223: {'lr': 0.00047258052547632636, 'samples': 1771008, 'steps': 9223, 'loss/train': 2.0253660678863525} 01/29/2022 00:41:13 - INFO - codeparrot_training - Step 9224: {'lr': 0.0004725730746593064, 'samples': 1771200, 'steps': 9224, 'loss/train': 1.8095530271530151} 01/29/2022 00:41:17 - INFO - codeparrot_training - Step 9225: {'lr': 0.0004725656228888541, 'samples': 1771392, 'steps': 9225, 'loss/train': 1.744529664516449} 01/29/2022 00:41:21 - INFO - codeparrot_training - Step 9226: {'lr': 0.0004725581701650014, 'samples': 1771584, 'steps': 9226, 'loss/train': 2.3015635013580322} 01/29/2022 00:41:25 - INFO - codeparrot_training - Step 9227: {'lr': 0.00047255071648778004, 'samples': 1771776, 'steps': 9227, 'loss/train': 2.6734079718589783} 01/29/2022 00:41:31 - INFO - codeparrot_training - Step 9228: {'lr': 0.00047254326185722207, 'samples': 1771968, 'steps': 9228, 'loss/train': 1.1862082779407501} 01/29/2022 00:41:35 - INFO - codeparrot_training - Step 9229: {'lr': 0.00047253580627335944, 'samples': 1772160, 'steps': 9229, 'loss/train': 1.2780067920684814} 01/29/2022 00:41:40 - INFO - codeparrot_training - Step 9230: {'lr': 0.00047252834973622414, 'samples': 1772352, 'steps': 9230, 'loss/train': 0.961852490901947} 01/29/2022 00:41:44 - INFO - codeparrot_training - Step 9231: {'lr': 0.00047252089224584804, 'samples': 1772544, 'steps': 9231, 'loss/train': 1.677070677280426} 01/29/2022 00:41:48 - INFO - codeparrot_training - Step 9232: {'lr': 0.0004725134338022631, 'samples': 1772736, 'steps': 9232, 'loss/train': 1.768613040447235} 01/29/2022 00:41:52 - INFO - codeparrot_training - Step 9233: {'lr': 0.00047250597440550124, 'samples': 1772928, 'steps': 9233, 'loss/train': 2.1631513237953186} 01/29/2022 00:41:59 - INFO - codeparrot_training - Step 9234: {'lr': 0.0004724985140555945, 'samples': 1773120, 'steps': 9234, 'loss/train': 1.9459986090660095} 01/29/2022 00:42:03 - INFO - codeparrot_training - Step 9235: {'lr': 0.0004724910527525748, 'samples': 1773312, 'steps': 9235, 'loss/train': 0.15963930264115334} 01/29/2022 00:42:07 - INFO - codeparrot_training - Step 9236: {'lr': 0.0004724835904964739, 'samples': 1773504, 'steps': 9236, 'loss/train': 1.8402323126792908} 01/29/2022 00:42:11 - INFO - codeparrot_training - Step 9237: {'lr': 0.00047247612728732407, 'samples': 1773696, 'steps': 9237, 'loss/train': 1.680489420890808} 01/29/2022 00:42:16 - INFO - codeparrot_training - Step 9238: {'lr': 0.0004724686631251572, 'samples': 1773888, 'steps': 9238, 'loss/train': 1.061968356370926} 01/29/2022 00:42:21 - INFO - codeparrot_training - Step 9239: {'lr': 0.00047246119801000507, 'samples': 1774080, 'steps': 9239, 'loss/train': 2.1220484375953674} 01/29/2022 00:42:25 - INFO - codeparrot_training - Step 9240: {'lr': 0.00047245373194189995, 'samples': 1774272, 'steps': 9240, 'loss/train': 0.09492328763008118} 01/29/2022 00:42:30 - INFO - codeparrot_training - Step 9241: {'lr': 0.0004724462649208736, 'samples': 1774464, 'steps': 9241, 'loss/train': 0.9882046580314636} 01/29/2022 00:42:34 - INFO - codeparrot_training - Step 9242: {'lr': 0.0004724387969469581, 'samples': 1774656, 'steps': 9242, 'loss/train': 1.2588430345058441} 01/29/2022 00:42:40 - INFO - codeparrot_training - Step 9243: {'lr': 0.00047243132802018544, 'samples': 1774848, 'steps': 9243, 'loss/train': 2.0144965052604675} 01/29/2022 00:42:44 - INFO - codeparrot_training - Step 9244: {'lr': 0.00047242385814058764, 'samples': 1775040, 'steps': 9244, 'loss/train': 1.0085070133209229} 01/29/2022 00:42:48 - INFO - codeparrot_training - Step 9245: {'lr': 0.0004724163873081966, 'samples': 1775232, 'steps': 9245, 'loss/train': 1.6586966514587402} 01/29/2022 00:42:52 - INFO - codeparrot_training - Step 9246: {'lr': 0.00047240891552304443, 'samples': 1775424, 'steps': 9246, 'loss/train': 1.4809105396270752} 01/29/2022 00:42:56 - INFO - codeparrot_training - Step 9247: {'lr': 0.0004724014427851631, 'samples': 1775616, 'steps': 9247, 'loss/train': 2.4878806471824646} 01/29/2022 00:43:01 - INFO - codeparrot_training - Step 9248: {'lr': 0.0004723939690945845, 'samples': 1775808, 'steps': 9248, 'loss/train': 2.223680555820465} 01/29/2022 00:43:07 - INFO - codeparrot_training - Step 9249: {'lr': 0.00047238649445134086, 'samples': 1776000, 'steps': 9249, 'loss/train': 1.6275250911712646} 01/29/2022 00:43:11 - INFO - codeparrot_training - Step 9250: {'lr': 0.00047237901885546405, 'samples': 1776192, 'steps': 9250, 'loss/train': 1.685858130455017} 01/29/2022 00:43:16 - INFO - codeparrot_training - Step 9251: {'lr': 0.00047237154230698607, 'samples': 1776384, 'steps': 9251, 'loss/train': 2.163008451461792} 01/29/2022 00:43:20 - INFO - codeparrot_training - Step 9252: {'lr': 0.0004723640648059391, 'samples': 1776576, 'steps': 9252, 'loss/train': 1.7005115747451782} 01/29/2022 00:43:25 - INFO - codeparrot_training - Step 9253: {'lr': 0.0004723565863523551, 'samples': 1776768, 'steps': 9253, 'loss/train': 2.8298057913780212} 01/29/2022 00:43:30 - INFO - codeparrot_training - Step 9254: {'lr': 0.0004723491069462661, 'samples': 1776960, 'steps': 9254, 'loss/train': 1.174233466386795} 01/29/2022 00:43:34 - INFO - codeparrot_training - Step 9255: {'lr': 0.00047234162658770407, 'samples': 1777152, 'steps': 9255, 'loss/train': 1.3519831001758575} 01/29/2022 00:43:38 - INFO - codeparrot_training - Step 9256: {'lr': 0.00047233414527670113, 'samples': 1777344, 'steps': 9256, 'loss/train': 1.8386013507843018} 01/29/2022 00:43:42 - INFO - codeparrot_training - Step 9257: {'lr': 0.0004723266630132893, 'samples': 1777536, 'steps': 9257, 'loss/train': 0.9863115549087524} 01/29/2022 00:43:47 - INFO - codeparrot_training - Step 9258: {'lr': 0.0004723191797975007, 'samples': 1777728, 'steps': 9258, 'loss/train': 1.8420875072479248} 01/29/2022 00:43:53 - INFO - codeparrot_training - Step 9259: {'lr': 0.00047231169562936726, 'samples': 1777920, 'steps': 9259, 'loss/train': 1.1267982423305511} 01/29/2022 00:43:57 - INFO - codeparrot_training - Step 9260: {'lr': 0.00047230421050892116, 'samples': 1778112, 'steps': 9260, 'loss/train': 1.1890801191329956} 01/29/2022 00:44:01 - INFO - codeparrot_training - Step 9261: {'lr': 0.00047229672443619433, 'samples': 1778304, 'steps': 9261, 'loss/train': 1.1272596716880798} 01/29/2022 00:44:06 - INFO - codeparrot_training - Step 9262: {'lr': 0.00047228923741121897, 'samples': 1778496, 'steps': 9262, 'loss/train': 1.3539657890796661} 01/29/2022 00:44:11 - INFO - codeparrot_training - Step 9263: {'lr': 0.0004722817494340271, 'samples': 1778688, 'steps': 9263, 'loss/train': 2.1361727714538574} 01/29/2022 00:44:15 - INFO - codeparrot_training - Step 9264: {'lr': 0.00047227426050465085, 'samples': 1778880, 'steps': 9264, 'loss/train': 2.048719346523285} 01/29/2022 00:44:20 - INFO - codeparrot_training - Step 9265: {'lr': 0.00047226677062312217, 'samples': 1779072, 'steps': 9265, 'loss/train': 1.5153241753578186} 01/29/2022 00:44:24 - INFO - codeparrot_training - Step 9266: {'lr': 0.00047225927978947327, 'samples': 1779264, 'steps': 9266, 'loss/train': 2.4958101511001587} 01/29/2022 00:44:28 - INFO - codeparrot_training - Step 9267: {'lr': 0.00047225178800373613, 'samples': 1779456, 'steps': 9267, 'loss/train': 1.0270746052265167} 01/29/2022 00:44:33 - INFO - codeparrot_training - Step 9268: {'lr': 0.00047224429526594296, 'samples': 1779648, 'steps': 9268, 'loss/train': 2.503147602081299} 01/29/2022 00:44:38 - INFO - codeparrot_training - Step 9269: {'lr': 0.0004722368015761258, 'samples': 1779840, 'steps': 9269, 'loss/train': 1.0583344995975494} 01/29/2022 00:44:42 - INFO - codeparrot_training - Step 9270: {'lr': 0.0004722293069343168, 'samples': 1780032, 'steps': 9270, 'loss/train': 1.2585438787937164} 01/29/2022 00:44:46 - INFO - codeparrot_training - Step 9271: {'lr': 0.00047222181134054785, 'samples': 1780224, 'steps': 9271, 'loss/train': 1.1876013278961182} 01/29/2022 00:44:50 - INFO - codeparrot_training - Step 9272: {'lr': 0.0004722143147948513, 'samples': 1780416, 'steps': 9272, 'loss/train': 2.1458680629730225} 01/29/2022 00:44:56 - INFO - codeparrot_training - Step 9273: {'lr': 0.0004722068172972593, 'samples': 1780608, 'steps': 9273, 'loss/train': 2.3364720940589905} 01/29/2022 00:45:00 - INFO - codeparrot_training - Step 9274: {'lr': 0.00047219931884780376, 'samples': 1780800, 'steps': 9274, 'loss/train': 1.2901414632797241} 01/29/2022 00:45:04 - INFO - codeparrot_training - Step 9275: {'lr': 0.0004721918194465169, 'samples': 1780992, 'steps': 9275, 'loss/train': 1.6875754594802856} 01/29/2022 00:45:08 - INFO - codeparrot_training - Step 9276: {'lr': 0.00047218431909343083, 'samples': 1781184, 'steps': 9276, 'loss/train': 1.8024850487709045} 01/29/2022 00:45:13 - INFO - codeparrot_training - Step 9277: {'lr': 0.0004721768177885777, 'samples': 1781376, 'steps': 9277, 'loss/train': 1.6051361560821533} 01/29/2022 00:45:18 - INFO - codeparrot_training - Step 9278: {'lr': 0.00047216931553198963, 'samples': 1781568, 'steps': 9278, 'loss/train': 1.6421340107917786} 01/29/2022 00:45:23 - INFO - codeparrot_training - Step 9279: {'lr': 0.0004721618123236987, 'samples': 1781760, 'steps': 9279, 'loss/train': 1.6488434672355652} 01/29/2022 00:45:27 - INFO - codeparrot_training - Step 9280: {'lr': 0.0004721543081637372, 'samples': 1781952, 'steps': 9280, 'loss/train': 2.197360575199127} 01/29/2022 00:45:31 - INFO - codeparrot_training - Step 9281: {'lr': 0.0004721468030521372, 'samples': 1782144, 'steps': 9281, 'loss/train': 2.0506240725517273} 01/29/2022 00:45:35 - INFO - codeparrot_training - Step 9282: {'lr': 0.0004721392969889308, 'samples': 1782336, 'steps': 9282, 'loss/train': 1.8737037777900696} 01/29/2022 00:45:42 - INFO - codeparrot_training - Step 9283: {'lr': 0.00047213178997415015, 'samples': 1782528, 'steps': 9283, 'loss/train': 7.655548095703125} 01/29/2022 00:45:46 - INFO - codeparrot_training - Step 9284: {'lr': 0.00047212428200782744, 'samples': 1782720, 'steps': 9284, 'loss/train': 2.188063681125641} 01/29/2022 00:45:51 - INFO - codeparrot_training - Step 9285: {'lr': 0.0004721167730899949, 'samples': 1782912, 'steps': 9285, 'loss/train': 1.6407605409622192} 01/29/2022 00:45:55 - INFO - codeparrot_training - Step 9286: {'lr': 0.0004721092632206846, 'samples': 1783104, 'steps': 9286, 'loss/train': 1.9725814461708069} 01/29/2022 00:45:59 - INFO - codeparrot_training - Step 9287: {'lr': 0.00047210175239992876, 'samples': 1783296, 'steps': 9287, 'loss/train': 1.0008582472801208} 01/29/2022 00:46:04 - INFO - codeparrot_training - Step 9288: {'lr': 0.0004720942406277595, 'samples': 1783488, 'steps': 9288, 'loss/train': 1.991754949092865} 01/29/2022 00:46:09 - INFO - codeparrot_training - Step 9289: {'lr': 0.0004720867279042091, 'samples': 1783680, 'steps': 9289, 'loss/train': 2.1617156267166138} 01/29/2022 00:46:13 - INFO - codeparrot_training - Step 9290: {'lr': 0.00047207921422930967, 'samples': 1783872, 'steps': 9290, 'loss/train': 1.0141316056251526} 01/29/2022 00:46:17 - INFO - codeparrot_training - Step 9291: {'lr': 0.00047207169960309335, 'samples': 1784064, 'steps': 9291, 'loss/train': 1.2709274590015411} 01/29/2022 00:46:21 - INFO - codeparrot_training - Step 9292: {'lr': 0.00047206418402559236, 'samples': 1784256, 'steps': 9292, 'loss/train': 2.1602699160575867} 01/29/2022 00:46:26 - INFO - codeparrot_training - Step 9293: {'lr': 0.000472056667496839, 'samples': 1784448, 'steps': 9293, 'loss/train': 2.6212636828422546} 01/29/2022 00:46:31 - INFO - codeparrot_training - Step 9294: {'lr': 0.0004720491500168654, 'samples': 1784640, 'steps': 9294, 'loss/train': 1.4734647274017334} 01/29/2022 00:46:35 - INFO - codeparrot_training - Step 9295: {'lr': 0.0004720416315857037, 'samples': 1784832, 'steps': 9295, 'loss/train': 1.9818509817123413} 01/29/2022 00:46:39 - INFO - codeparrot_training - Step 9296: {'lr': 0.00047203411220338615, 'samples': 1785024, 'steps': 9296, 'loss/train': 1.8184345364570618} 01/29/2022 00:46:43 - INFO - codeparrot_training - Step 9297: {'lr': 0.000472026591869945, 'samples': 1785216, 'steps': 9297, 'loss/train': 1.9924623370170593} 01/29/2022 00:46:50 - INFO - codeparrot_training - Step 9298: {'lr': 0.00047201907058541236, 'samples': 1785408, 'steps': 9298, 'loss/train': 1.177130788564682} 01/29/2022 00:46:54 - INFO - codeparrot_training - Step 9299: {'lr': 0.0004720115483498206, 'samples': 1785600, 'steps': 9299, 'loss/train': 1.5766090750694275} 01/29/2022 00:46:59 - INFO - codeparrot_training - Step 9300: {'lr': 0.00047200402516320186, 'samples': 1785792, 'steps': 9300, 'loss/train': 2.1490615010261536} 01/29/2022 00:47:03 - INFO - codeparrot_training - Step 9301: {'lr': 0.00047199650102558834, 'samples': 1785984, 'steps': 9301, 'loss/train': 2.6273730397224426} 01/29/2022 00:47:07 - INFO - codeparrot_training - Step 9302: {'lr': 0.0004719889759370123, 'samples': 1786176, 'steps': 9302, 'loss/train': 1.8308755159378052} 01/29/2022 00:47:12 - INFO - codeparrot_training - Step 9303: {'lr': 0.00047198144989750603, 'samples': 1786368, 'steps': 9303, 'loss/train': 0.6367986649274826} 01/29/2022 00:47:17 - INFO - codeparrot_training - Step 9304: {'lr': 0.00047197392290710164, 'samples': 1786560, 'steps': 9304, 'loss/train': 0.7833798229694366} 01/29/2022 00:47:21 - INFO - codeparrot_training - Step 9305: {'lr': 0.0004719663949658315, 'samples': 1786752, 'steps': 9305, 'loss/train': 1.0830393433570862} 01/29/2022 00:47:25 - INFO - codeparrot_training - Step 9306: {'lr': 0.00047195886607372773, 'samples': 1786944, 'steps': 9306, 'loss/train': 0.7134044319391251} 01/29/2022 00:47:31 - INFO - codeparrot_training - Step 9307: {'lr': 0.0004719513362308228, 'samples': 1787136, 'steps': 9307, 'loss/train': 1.6307533979415894} 01/29/2022 00:47:36 - INFO - codeparrot_training - Step 9308: {'lr': 0.0004719438054371487, 'samples': 1787328, 'steps': 9308, 'loss/train': 1.9173290133476257} 01/29/2022 00:47:40 - INFO - codeparrot_training - Step 9309: {'lr': 0.00047193627369273786, 'samples': 1787520, 'steps': 9309, 'loss/train': 1.7351163625717163} 01/29/2022 00:47:44 - INFO - codeparrot_training - Step 9310: {'lr': 0.00047192874099762246, 'samples': 1787712, 'steps': 9310, 'loss/train': 1.9514166712760925} 01/29/2022 00:47:49 - INFO - codeparrot_training - Step 9311: {'lr': 0.00047192120735183485, 'samples': 1787904, 'steps': 9311, 'loss/train': 2.0257864594459534} 01/29/2022 00:47:54 - INFO - codeparrot_training - Step 9312: {'lr': 0.0004719136727554072, 'samples': 1788096, 'steps': 9312, 'loss/train': 2.151766061782837} 01/29/2022 00:47:58 - INFO - codeparrot_training - Step 9313: {'lr': 0.0004719061372083719, 'samples': 1788288, 'steps': 9313, 'loss/train': 1.2183737754821777} 01/29/2022 00:48:02 - INFO - codeparrot_training - Step 9314: {'lr': 0.00047189860071076114, 'samples': 1788480, 'steps': 9314, 'loss/train': 1.611926257610321} 01/29/2022 00:48:07 - INFO - codeparrot_training - Step 9315: {'lr': 0.00047189106326260723, 'samples': 1788672, 'steps': 9315, 'loss/train': 1.0789880454540253} 01/29/2022 00:48:11 - INFO - codeparrot_training - Step 9316: {'lr': 0.0004718835248639425, 'samples': 1788864, 'steps': 9316, 'loss/train': 0.4686100333929062} 01/29/2022 00:48:15 - INFO - codeparrot_training - Step 9317: {'lr': 0.0004718759855147992, 'samples': 1789056, 'steps': 9317, 'loss/train': 1.6688597202301025} 01/29/2022 00:48:21 - INFO - codeparrot_training - Step 9318: {'lr': 0.00047186844521520955, 'samples': 1789248, 'steps': 9318, 'loss/train': 1.5599353909492493} 01/29/2022 00:48:25 - INFO - codeparrot_training - Step 9319: {'lr': 0.000471860903965206, 'samples': 1789440, 'steps': 9319, 'loss/train': 1.634299099445343} 01/29/2022 00:48:29 - INFO - codeparrot_training - Step 9320: {'lr': 0.00047185336176482084, 'samples': 1789632, 'steps': 9320, 'loss/train': 1.3432240784168243} 01/29/2022 00:48:33 - INFO - codeparrot_training - Step 9321: {'lr': 0.0004718458186140863, 'samples': 1789824, 'steps': 9321, 'loss/train': 1.4707454144954681} 01/29/2022 00:48:38 - INFO - codeparrot_training - Step 9322: {'lr': 0.0004718382745130346, 'samples': 1790016, 'steps': 9322, 'loss/train': 1.774458110332489} 01/29/2022 00:48:44 - INFO - codeparrot_training - Step 9323: {'lr': 0.0004718307294616983, 'samples': 1790208, 'steps': 9323, 'loss/train': 1.9877711534500122} 01/29/2022 00:48:48 - INFO - codeparrot_training - Step 9324: {'lr': 0.00047182318346010953, 'samples': 1790400, 'steps': 9324, 'loss/train': 1.5844674110412598} 01/29/2022 00:48:53 - INFO - codeparrot_training - Step 9325: {'lr': 0.0004718156365083007, 'samples': 1790592, 'steps': 9325, 'loss/train': 0.6234182864427567} 01/29/2022 00:48:57 - INFO - codeparrot_training - Step 9326: {'lr': 0.0004718080886063041, 'samples': 1790784, 'steps': 9326, 'loss/train': 0.9157371819019318} 01/29/2022 00:49:01 - INFO - codeparrot_training - Step 9327: {'lr': 0.00047180053975415216, 'samples': 1790976, 'steps': 9327, 'loss/train': 0.05183024704456329} 01/29/2022 00:49:07 - INFO - codeparrot_training - Step 9328: {'lr': 0.00047179298995187705, 'samples': 1791168, 'steps': 9328, 'loss/train': 1.6480981707572937} 01/29/2022 00:49:11 - INFO - codeparrot_training - Step 9329: {'lr': 0.00047178543919951124, 'samples': 1791360, 'steps': 9329, 'loss/train': 1.328737199306488} 01/29/2022 00:49:15 - INFO - codeparrot_training - Step 9330: {'lr': 0.000471777887497087, 'samples': 1791552, 'steps': 9330, 'loss/train': 1.2254671454429626} 01/29/2022 00:49:19 - INFO - codeparrot_training - Step 9331: {'lr': 0.0004717703348446367, 'samples': 1791744, 'steps': 9331, 'loss/train': 2.1027027368545532} 01/29/2022 00:49:24 - INFO - codeparrot_training - Step 9332: {'lr': 0.00047176278124219276, 'samples': 1791936, 'steps': 9332, 'loss/train': 2.35007107257843} 01/29/2022 00:49:30 - INFO - codeparrot_training - Step 9333: {'lr': 0.0004717552266897874, 'samples': 1792128, 'steps': 9333, 'loss/train': 2.3419126868247986} 01/29/2022 00:49:34 - INFO - codeparrot_training - Step 9334: {'lr': 0.0004717476711874532, 'samples': 1792320, 'steps': 9334, 'loss/train': 1.5335211753845215} 01/29/2022 00:49:39 - INFO - codeparrot_training - Step 9335: {'lr': 0.00047174011473522225, 'samples': 1792512, 'steps': 9335, 'loss/train': 1.946344256401062} 01/29/2022 00:49:43 - INFO - codeparrot_training - Step 9336: {'lr': 0.0004717325573331271, 'samples': 1792704, 'steps': 9336, 'loss/train': 1.6850756406784058} 01/29/2022 00:49:47 - INFO - codeparrot_training - Step 9337: {'lr': 0.00047172499898120014, 'samples': 1792896, 'steps': 9337, 'loss/train': 1.1056491136550903} 01/29/2022 00:49:53 - INFO - codeparrot_training - Step 9338: {'lr': 0.0004717174396794737, 'samples': 1793088, 'steps': 9338, 'loss/train': 1.8622063994407654} 01/29/2022 00:49:57 - INFO - codeparrot_training - Step 9339: {'lr': 0.00047170987942798004, 'samples': 1793280, 'steps': 9339, 'loss/train': 1.5909374356269836} 01/29/2022 00:50:01 - INFO - codeparrot_training - Step 9340: {'lr': 0.0004717023182267518, 'samples': 1793472, 'steps': 9340, 'loss/train': 1.401316523551941} 01/29/2022 00:50:05 - INFO - codeparrot_training - Step 9341: {'lr': 0.00047169475607582113, 'samples': 1793664, 'steps': 9341, 'loss/train': 1.7981611490249634} 01/29/2022 00:50:10 - INFO - codeparrot_training - Step 9342: {'lr': 0.00047168719297522053, 'samples': 1793856, 'steps': 9342, 'loss/train': 0.7844476997852325} 01/29/2022 00:50:15 - INFO - codeparrot_training - Step 9343: {'lr': 0.0004716796289249824, 'samples': 1794048, 'steps': 9343, 'loss/train': 1.0743144750595093} 01/29/2022 00:50:19 - INFO - codeparrot_training - Step 9344: {'lr': 0.0004716720639251392, 'samples': 1794240, 'steps': 9344, 'loss/train': 2.2706605195999146} 01/29/2022 00:50:23 - INFO - codeparrot_training - Step 9345: {'lr': 0.00047166449797572316, 'samples': 1794432, 'steps': 9345, 'loss/train': 1.3869922757148743} 01/29/2022 00:50:28 - INFO - codeparrot_training - Step 9346: {'lr': 0.0004716569310767668, 'samples': 1794624, 'steps': 9346, 'loss/train': 1.7862154841423035} 01/29/2022 00:50:32 - INFO - codeparrot_training - Step 9347: {'lr': 0.00047164936322830256, 'samples': 1794816, 'steps': 9347, 'loss/train': 0.6502305418252945} 01/29/2022 00:50:37 - INFO - codeparrot_training - Step 9348: {'lr': 0.0004716417944303628, 'samples': 1795008, 'steps': 9348, 'loss/train': 1.0511570870876312} 01/29/2022 00:50:41 - INFO - codeparrot_training - Step 9349: {'lr': 0.00047163422468298003, 'samples': 1795200, 'steps': 9349, 'loss/train': 1.6637864112854004} 01/29/2022 00:50:46 - INFO - codeparrot_training - Step 9350: {'lr': 0.00047162665398618666, 'samples': 1795392, 'steps': 9350, 'loss/train': 2.2500089406967163} 01/29/2022 00:50:50 - INFO - codeparrot_training - Step 9351: {'lr': 0.00047161908234001496, 'samples': 1795584, 'steps': 9351, 'loss/train': 1.171492874622345} 01/29/2022 00:50:54 - INFO - codeparrot_training - Step 9352: {'lr': 0.0004716115097444975, 'samples': 1795776, 'steps': 9352, 'loss/train': 1.76254802942276} 01/29/2022 00:51:01 - INFO - codeparrot_training - Step 9353: {'lr': 0.0004716039361996668, 'samples': 1795968, 'steps': 9353, 'loss/train': 1.584244966506958} 01/29/2022 00:51:05 - INFO - codeparrot_training - Step 9354: {'lr': 0.0004715963617055551, 'samples': 1796160, 'steps': 9354, 'loss/train': 1.2479124963283539} 01/29/2022 00:51:09 - INFO - codeparrot_training - Step 9355: {'lr': 0.00047158878626219505, 'samples': 1796352, 'steps': 9355, 'loss/train': 1.526560127735138} 01/29/2022 00:51:13 - INFO - codeparrot_training - Step 9356: {'lr': 0.00047158120986961897, 'samples': 1796544, 'steps': 9356, 'loss/train': 1.2376847863197327} 01/29/2022 00:51:17 - INFO - codeparrot_training - Step 9357: {'lr': 0.0004715736325278593, 'samples': 1796736, 'steps': 9357, 'loss/train': 2.3077717423439026} 01/29/2022 00:51:23 - INFO - codeparrot_training - Step 9358: {'lr': 0.0004715660542369485, 'samples': 1796928, 'steps': 9358, 'loss/train': 1.954277515411377} 01/29/2022 00:51:27 - INFO - codeparrot_training - Step 9359: {'lr': 0.0004715584749969192, 'samples': 1797120, 'steps': 9359, 'loss/train': 2.1303995847702026} 01/29/2022 00:51:31 - INFO - codeparrot_training - Step 9360: {'lr': 0.00047155089480780364, 'samples': 1797312, 'steps': 9360, 'loss/train': 1.339710921049118} 01/29/2022 00:51:36 - INFO - codeparrot_training - Step 9361: {'lr': 0.0004715433136696345, 'samples': 1797504, 'steps': 9361, 'loss/train': 1.6755069494247437} 01/29/2022 00:51:40 - INFO - codeparrot_training - Step 9362: {'lr': 0.0004715357315824441, 'samples': 1797696, 'steps': 9362, 'loss/train': 2.599196970462799} 01/29/2022 00:51:45 - INFO - codeparrot_training - Step 9363: {'lr': 0.00047152814854626494, 'samples': 1797888, 'steps': 9363, 'loss/train': 1.3868916034698486} 01/29/2022 00:51:49 - INFO - codeparrot_training - Step 9364: {'lr': 0.0004715205645611296, 'samples': 1798080, 'steps': 9364, 'loss/train': 2.4238239526748657} 01/29/2022 00:51:54 - INFO - codeparrot_training - Step 9365: {'lr': 0.00047151297962707054, 'samples': 1798272, 'steps': 9365, 'loss/train': 1.4559997916221619} 01/29/2022 00:51:58 - INFO - codeparrot_training - Step 9366: {'lr': 0.00047150539374412004, 'samples': 1798464, 'steps': 9366, 'loss/train': 2.5275142192840576} 01/29/2022 00:52:02 - INFO - codeparrot_training - Step 9367: {'lr': 0.0004714978069123109, 'samples': 1798656, 'steps': 9367, 'loss/train': 2.039728045463562} 01/29/2022 00:52:08 - INFO - codeparrot_training - Step 9368: {'lr': 0.00047149021913167545, 'samples': 1798848, 'steps': 9368, 'loss/train': 1.7944146394729614} 01/29/2022 00:52:12 - INFO - codeparrot_training - Step 9369: {'lr': 0.00047148263040224626, 'samples': 1799040, 'steps': 9369, 'loss/train': 1.689220905303955} 01/29/2022 00:52:17 - INFO - codeparrot_training - Step 9370: {'lr': 0.00047147504072405575, 'samples': 1799232, 'steps': 9370, 'loss/train': 0.32227541506290436} 01/29/2022 00:52:21 - INFO - codeparrot_training - Step 9371: {'lr': 0.0004714674500971366, 'samples': 1799424, 'steps': 9371, 'loss/train': 2.4542829394340515} 01/29/2022 00:52:26 - INFO - codeparrot_training - Step 9372: {'lr': 0.00047145985852152115, 'samples': 1799616, 'steps': 9372, 'loss/train': 1.553950309753418} 01/29/2022 00:52:31 - INFO - codeparrot_training - Step 9373: {'lr': 0.000471452265997242, 'samples': 1799808, 'steps': 9373, 'loss/train': 1.6613377332687378} 01/29/2022 00:52:35 - INFO - codeparrot_training - Step 9374: {'lr': 0.00047144467252433164, 'samples': 1800000, 'steps': 9374, 'loss/train': 1.9416847229003906} 01/29/2022 00:52:39 - INFO - codeparrot_training - Step 9375: {'lr': 0.00047143707810282266, 'samples': 1800192, 'steps': 9375, 'loss/train': 2.527619183063507} 01/29/2022 00:52:43 - INFO - codeparrot_training - Step 9376: {'lr': 0.0004714294827327475, 'samples': 1800384, 'steps': 9376, 'loss/train': 2.2277015447616577} 01/29/2022 00:52:50 - INFO - codeparrot_training - Step 9377: {'lr': 0.00047142188641413873, 'samples': 1800576, 'steps': 9377, 'loss/train': 1.2639042735099792} 01/29/2022 00:52:54 - INFO - codeparrot_training - Step 9378: {'lr': 0.000471414289147029, 'samples': 1800768, 'steps': 9378, 'loss/train': 2.278161585330963} 01/29/2022 00:52:58 - INFO - codeparrot_training - Step 9379: {'lr': 0.00047140669093145073, 'samples': 1800960, 'steps': 9379, 'loss/train': 2.0689424872398376} 01/29/2022 00:53:03 - INFO - codeparrot_training - Step 9380: {'lr': 0.00047139909176743643, 'samples': 1801152, 'steps': 9380, 'loss/train': 2.254423499107361} 01/29/2022 00:53:07 - INFO - codeparrot_training - Step 9381: {'lr': 0.0004713914916550188, 'samples': 1801344, 'steps': 9381, 'loss/train': 1.1702124774456024} 01/29/2022 00:53:12 - INFO - codeparrot_training - Step 9382: {'lr': 0.00047138389059423033, 'samples': 1801536, 'steps': 9382, 'loss/train': 1.6813539862632751} 01/29/2022 00:53:17 - INFO - codeparrot_training - Step 9383: {'lr': 0.0004713762885851035, 'samples': 1801728, 'steps': 9383, 'loss/train': 1.4335500597953796} 01/29/2022 00:53:21 - INFO - codeparrot_training - Step 9384: {'lr': 0.000471368685627671, 'samples': 1801920, 'steps': 9384, 'loss/train': 2.2783986926078796} 01/29/2022 00:53:25 - INFO - codeparrot_training - Step 9385: {'lr': 0.00047136108172196535, 'samples': 1802112, 'steps': 9385, 'loss/train': 1.117907166481018} 01/29/2022 00:53:29 - INFO - codeparrot_training - Step 9386: {'lr': 0.00047135347686801907, 'samples': 1802304, 'steps': 9386, 'loss/train': 0.6782258749008179} 01/29/2022 00:53:35 - INFO - codeparrot_training - Step 9387: {'lr': 0.0004713458710658648, 'samples': 1802496, 'steps': 9387, 'loss/train': 1.3625726401805878} 01/29/2022 00:53:39 - INFO - codeparrot_training - Step 9388: {'lr': 0.0004713382643155351, 'samples': 1802688, 'steps': 9388, 'loss/train': 2.1302502751350403} 01/29/2022 00:53:43 - INFO - codeparrot_training - Step 9389: {'lr': 0.00047133065661706254, 'samples': 1802880, 'steps': 9389, 'loss/train': 1.3093644976615906} 01/29/2022 00:53:47 - INFO - codeparrot_training - Step 9390: {'lr': 0.00047132304797047975, 'samples': 1803072, 'steps': 9390, 'loss/train': 1.3606964349746704} 01/29/2022 00:53:52 - INFO - codeparrot_training - Step 9391: {'lr': 0.00047131543837581935, 'samples': 1803264, 'steps': 9391, 'loss/train': 1.0497657358646393} 01/29/2022 00:53:57 - INFO - codeparrot_training - Step 9392: {'lr': 0.0004713078278331138, 'samples': 1803456, 'steps': 9392, 'loss/train': 2.835383713245392} 01/29/2022 00:54:01 - INFO - codeparrot_training - Step 9393: {'lr': 0.00047130021634239584, 'samples': 1803648, 'steps': 9393, 'loss/train': 1.0251881182193756} 01/29/2022 00:54:05 - INFO - codeparrot_training - Step 9394: {'lr': 0.000471292603903698, 'samples': 1803840, 'steps': 9394, 'loss/train': 1.482551246881485} 01/29/2022 00:54:10 - INFO - codeparrot_training - Step 9395: {'lr': 0.00047128499051705296, 'samples': 1804032, 'steps': 9395, 'loss/train': 1.139596939086914} 01/29/2022 00:54:14 - INFO - codeparrot_training - Step 9396: {'lr': 0.00047127737618249323, 'samples': 1804224, 'steps': 9396, 'loss/train': 1.1596744358539581} 01/29/2022 00:54:19 - INFO - codeparrot_training - Step 9397: {'lr': 0.00047126976090005153, 'samples': 1804416, 'steps': 9397, 'loss/train': 1.7180339097976685} 01/29/2022 00:54:23 - INFO - codeparrot_training - Step 9398: {'lr': 0.00047126214466976034, 'samples': 1804608, 'steps': 9398, 'loss/train': 2.547174632549286} 01/29/2022 00:54:28 - INFO - codeparrot_training - Step 9399: {'lr': 0.0004712545274916525, 'samples': 1804800, 'steps': 9399, 'loss/train': 1.9504621624946594} 01/29/2022 00:54:32 - INFO - codeparrot_training - Step 9400: {'lr': 0.00047124690936576046, 'samples': 1804992, 'steps': 9400, 'loss/train': 2.0409563183784485} 01/29/2022 00:54:36 - INFO - codeparrot_training - Step 9401: {'lr': 0.000471239290292117, 'samples': 1805184, 'steps': 9401, 'loss/train': 1.7452602982521057} 01/29/2022 00:54:42 - INFO - codeparrot_training - Step 9402: {'lr': 0.00047123167027075455, 'samples': 1805376, 'steps': 9402, 'loss/train': 2.0836039781570435} 01/29/2022 00:54:47 - INFO - codeparrot_training - Step 9403: {'lr': 0.0004712240493017059, 'samples': 1805568, 'steps': 9403, 'loss/train': 1.8404268622398376} 01/29/2022 00:54:51 - INFO - codeparrot_training - Step 9404: {'lr': 0.0004712164273850037, 'samples': 1805760, 'steps': 9404, 'loss/train': 1.942175030708313} 01/29/2022 00:54:55 - INFO - codeparrot_training - Step 9405: {'lr': 0.0004712088045206806, 'samples': 1805952, 'steps': 9405, 'loss/train': 2.150846779346466} 01/29/2022 00:54:59 - INFO - codeparrot_training - Step 9406: {'lr': 0.00047120118070876916, 'samples': 1806144, 'steps': 9406, 'loss/train': 1.3273176848888397} 01/29/2022 00:55:05 - INFO - codeparrot_training - Step 9407: {'lr': 0.0004711935559493021, 'samples': 1806336, 'steps': 9407, 'loss/train': 1.633010745048523} 01/29/2022 00:55:09 - INFO - codeparrot_training - Step 9408: {'lr': 0.00047118593024231216, 'samples': 1806528, 'steps': 9408, 'loss/train': 1.278823345899582} 01/29/2022 00:55:13 - INFO - codeparrot_training - Step 9409: {'lr': 0.00047117830358783184, 'samples': 1806720, 'steps': 9409, 'loss/train': 1.610081434249878} 01/29/2022 00:55:17 - INFO - codeparrot_training - Step 9410: {'lr': 0.0004711706759858939, 'samples': 1806912, 'steps': 9410, 'loss/train': 0.5310024619102478} 01/29/2022 00:55:22 - INFO - codeparrot_training - Step 9411: {'lr': 0.0004711630474365311, 'samples': 1807104, 'steps': 9411, 'loss/train': 1.5006453394889832} 01/29/2022 00:55:28 - INFO - codeparrot_training - Step 9412: {'lr': 0.000471155417939776, 'samples': 1807296, 'steps': 9412, 'loss/train': 1.297861933708191} 01/29/2022 00:55:32 - INFO - codeparrot_training - Step 9413: {'lr': 0.00047114778749566123, 'samples': 1807488, 'steps': 9413, 'loss/train': 1.4048365652561188} 01/29/2022 00:55:36 - INFO - codeparrot_training - Step 9414: {'lr': 0.00047114015610421966, 'samples': 1807680, 'steps': 9414, 'loss/train': 0.46803541481494904} 01/29/2022 00:55:41 - INFO - codeparrot_training - Step 9415: {'lr': 0.00047113252376548387, 'samples': 1807872, 'steps': 9415, 'loss/train': 1.759088158607483} 01/29/2022 00:55:45 - INFO - codeparrot_training - Step 9416: {'lr': 0.00047112489047948655, 'samples': 1808064, 'steps': 9416, 'loss/train': 1.6495773196220398} 01/29/2022 00:55:51 - INFO - codeparrot_training - Step 9417: {'lr': 0.0004711172562462604, 'samples': 1808256, 'steps': 9417, 'loss/train': 2.455265700817108} 01/29/2022 00:55:55 - INFO - codeparrot_training - Step 9418: {'lr': 0.0004711096210658381, 'samples': 1808448, 'steps': 9418, 'loss/train': 2.168998897075653} 01/29/2022 00:55:59 - INFO - codeparrot_training - Step 9419: {'lr': 0.0004711019849382525, 'samples': 1808640, 'steps': 9419, 'loss/train': 1.9656332731246948} 01/29/2022 00:56:03 - INFO - codeparrot_training - Step 9420: {'lr': 0.0004710943478635361, 'samples': 1808832, 'steps': 9420, 'loss/train': 1.3773847818374634} 01/29/2022 00:56:08 - INFO - codeparrot_training - Step 9421: {'lr': 0.00047108670984172176, 'samples': 1809024, 'steps': 9421, 'loss/train': 2.5360450744628906} 01/29/2022 00:56:13 - INFO - codeparrot_training - Step 9422: {'lr': 0.00047107907087284216, 'samples': 1809216, 'steps': 9422, 'loss/train': 1.807152271270752} 01/29/2022 00:56:17 - INFO - codeparrot_training - Step 9423: {'lr': 0.00047107143095693007, 'samples': 1809408, 'steps': 9423, 'loss/train': 1.797232747077942} 01/29/2022 00:56:21 - INFO - codeparrot_training - Step 9424: {'lr': 0.0004710637900940181, 'samples': 1809600, 'steps': 9424, 'loss/train': 2.2357251048088074} 01/29/2022 00:56:26 - INFO - codeparrot_training - Step 9425: {'lr': 0.00047105614828413906, 'samples': 1809792, 'steps': 9425, 'loss/train': 1.7981071472167969} 01/29/2022 00:56:30 - INFO - codeparrot_training - Step 9426: {'lr': 0.0004710485055273257, 'samples': 1809984, 'steps': 9426, 'loss/train': 1.3842253983020782} 01/29/2022 00:56:37 - INFO - codeparrot_training - Step 9427: {'lr': 0.00047104086182361073, 'samples': 1810176, 'steps': 9427, 'loss/train': 2.294833481311798} 01/29/2022 00:56:41 - INFO - codeparrot_training - Step 9428: {'lr': 0.00047103321717302684, 'samples': 1810368, 'steps': 9428, 'loss/train': 1.6925591826438904} 01/29/2022 00:56:45 - INFO - codeparrot_training - Step 9429: {'lr': 0.00047102557157560686, 'samples': 1810560, 'steps': 9429, 'loss/train': 1.8602394461631775} 01/29/2022 00:56:49 - INFO - codeparrot_training - Step 9430: {'lr': 0.00047101792503138353, 'samples': 1810752, 'steps': 9430, 'loss/train': 1.3655003607273102} 01/29/2022 00:56:54 - INFO - codeparrot_training - Step 9431: {'lr': 0.0004710102775403896, 'samples': 1810944, 'steps': 9431, 'loss/train': 1.1702842712402344} 01/29/2022 00:56:59 - INFO - codeparrot_training - Step 9432: {'lr': 0.00047100262910265787, 'samples': 1811136, 'steps': 9432, 'loss/train': 2.0362354516983032} 01/29/2022 00:57:03 - INFO - codeparrot_training - Step 9433: {'lr': 0.00047099497971822096, 'samples': 1811328, 'steps': 9433, 'loss/train': 1.2380524277687073} 01/29/2022 00:57:08 - INFO - codeparrot_training - Step 9434: {'lr': 0.00047098732938711174, 'samples': 1811520, 'steps': 9434, 'loss/train': 1.7872229218482971} 01/29/2022 00:57:12 - INFO - codeparrot_training - Step 9435: {'lr': 0.00047097967810936305, 'samples': 1811712, 'steps': 9435, 'loss/train': 0.14447945728898048} 01/29/2022 00:57:16 - INFO - codeparrot_training - Step 9436: {'lr': 0.00047097202588500747, 'samples': 1811904, 'steps': 9436, 'loss/train': 1.4426070749759674} 01/29/2022 00:57:23 - INFO - codeparrot_training - Step 9437: {'lr': 0.000470964372714078, 'samples': 1812096, 'steps': 9437, 'loss/train': 1.835712969303131} 01/29/2022 00:57:27 - INFO - codeparrot_training - Step 9438: {'lr': 0.00047095671859660726, 'samples': 1812288, 'steps': 9438, 'loss/train': 0.10396141558885574} 01/29/2022 00:57:31 - INFO - codeparrot_training - Step 9439: {'lr': 0.0004709490635326281, 'samples': 1812480, 'steps': 9439, 'loss/train': 7.33382248878479} 01/29/2022 00:57:36 - INFO - codeparrot_training - Step 9440: {'lr': 0.0004709414075221734, 'samples': 1812672, 'steps': 9440, 'loss/train': 1.7312330603599548} 01/29/2022 00:57:40 - INFO - codeparrot_training - Step 9441: {'lr': 0.00047093375056527577, 'samples': 1812864, 'steps': 9441, 'loss/train': 0.920101135969162} 01/29/2022 00:57:44 - INFO - codeparrot_training - Step 9442: {'lr': 0.0004709260926619682, 'samples': 1813056, 'steps': 9442, 'loss/train': 1.467010259628296} 01/29/2022 00:57:49 - INFO - codeparrot_training - Step 9443: {'lr': 0.00047091843381228326, 'samples': 1813248, 'steps': 9443, 'loss/train': 1.882117509841919} 01/29/2022 00:57:54 - INFO - codeparrot_training - Step 9444: {'lr': 0.000470910774016254, 'samples': 1813440, 'steps': 9444, 'loss/train': 1.7734538912773132} 01/29/2022 00:57:58 - INFO - codeparrot_training - Step 9445: {'lr': 0.0004709031132739131, 'samples': 1813632, 'steps': 9445, 'loss/train': 1.6789212226867676} 01/29/2022 00:58:02 - INFO - codeparrot_training - Step 9446: {'lr': 0.0004708954515852934, 'samples': 1813824, 'steps': 9446, 'loss/train': 1.6564906239509583} 01/29/2022 00:58:06 - INFO - codeparrot_training - Step 9447: {'lr': 0.00047088778895042774, 'samples': 1814016, 'steps': 9447, 'loss/train': 0.9830294251441956} 01/29/2022 00:58:12 - INFO - codeparrot_training - Step 9448: {'lr': 0.000470880125369349, 'samples': 1814208, 'steps': 9448, 'loss/train': 1.2278468906879425} 01/29/2022 00:58:16 - INFO - codeparrot_training - Step 9449: {'lr': 0.0004708724608420898, 'samples': 1814400, 'steps': 9449, 'loss/train': 1.7055519819259644} 01/29/2022 00:58:20 - INFO - codeparrot_training - Step 9450: {'lr': 0.0004708647953686832, 'samples': 1814592, 'steps': 9450, 'loss/train': 2.669167220592499} 01/29/2022 00:58:24 - INFO - codeparrot_training - Step 9451: {'lr': 0.000470857128949162, 'samples': 1814784, 'steps': 9451, 'loss/train': 0.7358705699443817} 01/29/2022 00:58:29 - INFO - codeparrot_training - Step 9452: {'lr': 0.0004708494615835589, 'samples': 1814976, 'steps': 9452, 'loss/train': 2.2485212087631226} 01/29/2022 00:58:34 - INFO - codeparrot_training - Step 9453: {'lr': 0.0004708417932719068, 'samples': 1815168, 'steps': 9453, 'loss/train': 1.8083088397979736} 01/29/2022 00:58:38 - INFO - codeparrot_training - Step 9454: {'lr': 0.0004708341240142387, 'samples': 1815360, 'steps': 9454, 'loss/train': 2.4610041975975037} 01/29/2022 00:58:42 - INFO - codeparrot_training - Step 9455: {'lr': 0.0004708264538105873, 'samples': 1815552, 'steps': 9455, 'loss/train': 1.3649060726165771} 01/29/2022 00:58:47 - INFO - codeparrot_training - Step 9456: {'lr': 0.0004708187826609854, 'samples': 1815744, 'steps': 9456, 'loss/train': 1.8872355222702026} 01/29/2022 00:58:51 - INFO - codeparrot_training - Step 9457: {'lr': 0.0004708111105654661, 'samples': 1815936, 'steps': 9457, 'loss/train': 1.7293294072151184} 01/29/2022 00:58:57 - INFO - codeparrot_training - Step 9458: {'lr': 0.000470803437524062, 'samples': 1816128, 'steps': 9458, 'loss/train': 1.6740585565567017} 01/29/2022 00:59:01 - INFO - codeparrot_training - Step 9459: {'lr': 0.00047079576353680614, 'samples': 1816320, 'steps': 9459, 'loss/train': 2.0264970660209656} 01/29/2022 00:59:06 - INFO - codeparrot_training - Step 9460: {'lr': 0.0004707880886037314, 'samples': 1816512, 'steps': 9460, 'loss/train': 2.1644756197929382} 01/29/2022 00:59:10 - INFO - codeparrot_training - Step 9461: {'lr': 0.00047078041272487046, 'samples': 1816704, 'steps': 9461, 'loss/train': 1.4961552321910858} 01/29/2022 00:59:14 - INFO - codeparrot_training - Step 9462: {'lr': 0.00047077273590025637, 'samples': 1816896, 'steps': 9462, 'loss/train': 1.2187874615192413} 01/29/2022 00:59:19 - INFO - codeparrot_training - Step 9463: {'lr': 0.00047076505812992204, 'samples': 1817088, 'steps': 9463, 'loss/train': 0.9934509694576263} 01/29/2022 00:59:24 - INFO - codeparrot_training - Step 9464: {'lr': 0.0004707573794139003, 'samples': 1817280, 'steps': 9464, 'loss/train': 2.487951636314392} 01/29/2022 00:59:28 - INFO - codeparrot_training - Step 9465: {'lr': 0.00047074969975222406, 'samples': 1817472, 'steps': 9465, 'loss/train': 0.9972691833972931} 01/29/2022 00:59:32 - INFO - codeparrot_training - Step 9466: {'lr': 0.0004707420191449261, 'samples': 1817664, 'steps': 9466, 'loss/train': 1.4644449949264526} 01/29/2022 00:59:36 - INFO - codeparrot_training - Step 9467: {'lr': 0.0004707343375920395, 'samples': 1817856, 'steps': 9467, 'loss/train': 2.4024170637130737} 01/29/2022 00:59:42 - INFO - codeparrot_training - Step 9468: {'lr': 0.0004707266550935971, 'samples': 1818048, 'steps': 9468, 'loss/train': 1.2553242444992065} 01/29/2022 00:59:46 - INFO - codeparrot_training - Step 9469: {'lr': 0.00047071897164963175, 'samples': 1818240, 'steps': 9469, 'loss/train': 1.6237719655036926} 01/29/2022 00:59:50 - INFO - codeparrot_training - Step 9470: {'lr': 0.00047071128726017643, 'samples': 1818432, 'steps': 9470, 'loss/train': 2.968183994293213} 01/29/2022 00:59:55 - INFO - codeparrot_training - Step 9471: {'lr': 0.0004707036019252641, 'samples': 1818624, 'steps': 9471, 'loss/train': 1.9304868578910828} 01/29/2022 00:59:59 - INFO - codeparrot_training - Step 9472: {'lr': 0.00047069591564492753, 'samples': 1818816, 'steps': 9472, 'loss/train': 2.1928768157958984} 01/29/2022 01:00:05 - INFO - codeparrot_training - Step 9473: {'lr': 0.00047068822841919976, 'samples': 1819008, 'steps': 9473, 'loss/train': 1.846702516078949} 01/29/2022 01:00:10 - INFO - codeparrot_training - Step 9474: {'lr': 0.0004706805402481137, 'samples': 1819200, 'steps': 9474, 'loss/train': 0.6807696372270584} 01/29/2022 01:00:14 - INFO - codeparrot_training - Step 9475: {'lr': 0.00047067285113170233, 'samples': 1819392, 'steps': 9475, 'loss/train': 2.2936588525772095} 01/29/2022 01:00:18 - INFO - codeparrot_training - Step 9476: {'lr': 0.0004706651610699985, 'samples': 1819584, 'steps': 9476, 'loss/train': 1.87196284532547} 01/29/2022 01:00:22 - INFO - codeparrot_training - Step 9477: {'lr': 0.0004706574700630352, 'samples': 1819776, 'steps': 9477, 'loss/train': 1.6070703864097595} 01/29/2022 01:00:28 - INFO - codeparrot_training - Step 9478: {'lr': 0.0004706497781108453, 'samples': 1819968, 'steps': 9478, 'loss/train': 1.749557375907898} 01/29/2022 01:00:32 - INFO - codeparrot_training - Step 9479: {'lr': 0.00047064208521346184, 'samples': 1820160, 'steps': 9479, 'loss/train': 1.5558757781982422} 01/29/2022 01:00:36 - INFO - codeparrot_training - Step 9480: {'lr': 0.0004706343913709178, 'samples': 1820352, 'steps': 9480, 'loss/train': 1.5916439294815063} 01/29/2022 01:00:40 - INFO - codeparrot_training - Step 9481: {'lr': 0.0004706266965832461, 'samples': 1820544, 'steps': 9481, 'loss/train': 2.496689021587372} 01/29/2022 01:00:45 - INFO - codeparrot_training - Step 9482: {'lr': 0.0004706190008504796, 'samples': 1820736, 'steps': 9482, 'loss/train': 1.9217721819877625} 01/29/2022 01:00:51 - INFO - codeparrot_training - Step 9483: {'lr': 0.00047061130417265143, 'samples': 1820928, 'steps': 9483, 'loss/train': 1.829582691192627} 01/29/2022 01:00:55 - INFO - codeparrot_training - Step 9484: {'lr': 0.0004706036065497944, 'samples': 1821120, 'steps': 9484, 'loss/train': 1.9590375423431396} 01/29/2022 01:01:00 - INFO - codeparrot_training - Step 9485: {'lr': 0.0004705959079819416, 'samples': 1821312, 'steps': 9485, 'loss/train': 1.8436471223831177} 01/29/2022 01:01:04 - INFO - codeparrot_training - Step 9486: {'lr': 0.0004705882084691261, 'samples': 1821504, 'steps': 9486, 'loss/train': 1.1333814561367035} 01/29/2022 01:01:08 - INFO - codeparrot_training - Step 9487: {'lr': 0.00047058050801138064, 'samples': 1821696, 'steps': 9487, 'loss/train': 1.6985412240028381} 01/29/2022 01:01:13 - INFO - codeparrot_training - Step 9488: {'lr': 0.00047057280660873835, 'samples': 1821888, 'steps': 9488, 'loss/train': 1.1067147552967072} 01/29/2022 01:01:18 - INFO - codeparrot_training - Step 9489: {'lr': 0.0004705651042612322, 'samples': 1822080, 'steps': 9489, 'loss/train': 2.347937822341919} 01/29/2022 01:01:22 - INFO - codeparrot_training - Step 9490: {'lr': 0.00047055740096889516, 'samples': 1822272, 'steps': 9490, 'loss/train': 1.084207445383072} 01/29/2022 01:01:26 - INFO - codeparrot_training - Step 9491: {'lr': 0.0004705496967317603, 'samples': 1822464, 'steps': 9491, 'loss/train': 1.6526498794555664} 01/29/2022 01:01:30 - INFO - codeparrot_training - Step 9492: {'lr': 0.0004705419915498605, 'samples': 1822656, 'steps': 9492, 'loss/train': 7.463773727416992} 01/29/2022 01:01:36 - INFO - codeparrot_training - Step 9493: {'lr': 0.0004705342854232288, 'samples': 1822848, 'steps': 9493, 'loss/train': 1.177017331123352} 01/29/2022 01:01:40 - INFO - codeparrot_training - Step 9494: {'lr': 0.00047052657835189836, 'samples': 1823040, 'steps': 9494, 'loss/train': 1.746666133403778} 01/29/2022 01:01:45 - INFO - codeparrot_training - Step 9495: {'lr': 0.00047051887033590205, 'samples': 1823232, 'steps': 9495, 'loss/train': 1.811070442199707} 01/29/2022 01:01:49 - INFO - codeparrot_training - Step 9496: {'lr': 0.00047051116137527296, 'samples': 1823424, 'steps': 9496, 'loss/train': 1.1769404411315918} 01/29/2022 01:01:53 - INFO - codeparrot_training - Step 9497: {'lr': 0.000470503451470044, 'samples': 1823616, 'steps': 9497, 'loss/train': 1.1170682609081268} 01/29/2022 01:01:57 - INFO - codeparrot_training - Step 9498: {'lr': 0.00047049574062024837, 'samples': 1823808, 'steps': 9498, 'loss/train': 1.553905427455902} 01/29/2022 01:02:05 - INFO - codeparrot_training - Step 9499: {'lr': 0.0004704880288259189, 'samples': 1824000, 'steps': 9499, 'loss/train': 1.7051544785499573} 01/29/2022 01:02:09 - INFO - codeparrot_training - Step 9500: {'lr': 0.00047048031608708875, 'samples': 1824192, 'steps': 9500, 'loss/train': 1.6711310148239136} 01/29/2022 01:02:13 - INFO - codeparrot_training - Step 9501: {'lr': 0.00047047260240379096, 'samples': 1824384, 'steps': 9501, 'loss/train': 1.2033387422561646} 01/29/2022 01:02:17 - INFO - codeparrot_training - Step 9502: {'lr': 0.00047046488777605853, 'samples': 1824576, 'steps': 9502, 'loss/train': 2.0625003576278687} 01/29/2022 01:02:22 - INFO - codeparrot_training - Step 9503: {'lr': 0.0004704571722039246, 'samples': 1824768, 'steps': 9503, 'loss/train': 0.9676428437232971} 01/29/2022 01:02:27 - INFO - codeparrot_training - Step 9504: {'lr': 0.00047044945568742205, 'samples': 1824960, 'steps': 9504, 'loss/train': 2.2906031012535095} 01/29/2022 01:02:31 - INFO - codeparrot_training - Step 9505: {'lr': 0.0004704417382265841, 'samples': 1825152, 'steps': 9505, 'loss/train': 1.4380751252174377} 01/29/2022 01:02:36 - INFO - codeparrot_training - Step 9506: {'lr': 0.0004704340198214437, 'samples': 1825344, 'steps': 9506, 'loss/train': 0.9425505101680756} 01/29/2022 01:02:40 - INFO - codeparrot_training - Step 9507: {'lr': 0.00047042630047203394, 'samples': 1825536, 'steps': 9507, 'loss/train': 1.0308730602264404} 01/29/2022 01:02:44 - INFO - codeparrot_training - Step 9508: {'lr': 0.0004704185801783879, 'samples': 1825728, 'steps': 9508, 'loss/train': 2.283954620361328} 01/29/2022 01:02:49 - INFO - codeparrot_training - Step 9509: {'lr': 0.0004704108589405387, 'samples': 1825920, 'steps': 9509, 'loss/train': 2.3461811542510986} 01/29/2022 01:02:54 - INFO - codeparrot_training - Step 9510: {'lr': 0.0004704031367585193, 'samples': 1826112, 'steps': 9510, 'loss/train': 1.3484361469745636} 01/29/2022 01:02:58 - INFO - codeparrot_training - Step 9511: {'lr': 0.0004703954136323629, 'samples': 1826304, 'steps': 9511, 'loss/train': 2.03324156999588} 01/29/2022 01:03:02 - INFO - codeparrot_training - Step 9512: {'lr': 0.0004703876895621025, 'samples': 1826496, 'steps': 9512, 'loss/train': 1.7590430974960327} 01/29/2022 01:03:06 - INFO - codeparrot_training - Step 9513: {'lr': 0.00047037996454777134, 'samples': 1826688, 'steps': 9513, 'loss/train': 1.2660524547100067} 01/29/2022 01:03:12 - INFO - codeparrot_training - Step 9514: {'lr': 0.00047037223858940224, 'samples': 1826880, 'steps': 9514, 'loss/train': 2.1735448837280273} 01/29/2022 01:03:16 - INFO - codeparrot_training - Step 9515: {'lr': 0.00047036451168702855, 'samples': 1827072, 'steps': 9515, 'loss/train': 1.460829645395279} 01/29/2022 01:03:21 - INFO - codeparrot_training - Step 9516: {'lr': 0.0004703567838406832, 'samples': 1827264, 'steps': 9516, 'loss/train': 1.8396313190460205} 01/29/2022 01:03:25 - INFO - codeparrot_training - Step 9517: {'lr': 0.00047034905505039936, 'samples': 1827456, 'steps': 9517, 'loss/train': 2.4390546083450317} 01/29/2022 01:03:29 - INFO - codeparrot_training - Step 9518: {'lr': 0.0004703413253162102, 'samples': 1827648, 'steps': 9518, 'loss/train': 0.8885741829872131} 01/29/2022 01:03:34 - INFO - codeparrot_training - Step 9519: {'lr': 0.00047033359463814875, 'samples': 1827840, 'steps': 9519, 'loss/train': 2.394376516342163} 01/29/2022 01:03:39 - INFO - codeparrot_training - Step 9520: {'lr': 0.00047032586301624804, 'samples': 1828032, 'steps': 9520, 'loss/train': 1.2939264178276062} 01/29/2022 01:03:43 - INFO - codeparrot_training - Step 9521: {'lr': 0.0004703181304505414, 'samples': 1828224, 'steps': 9521, 'loss/train': 1.9021275043487549} 01/29/2022 01:03:47 - INFO - codeparrot_training - Step 9522: {'lr': 0.0004703103969410618, 'samples': 1828416, 'steps': 9522, 'loss/train': 1.554284691810608} 01/29/2022 01:03:51 - INFO - codeparrot_training - Step 9523: {'lr': 0.0004703026624878425, 'samples': 1828608, 'steps': 9523, 'loss/train': 0.5832461267709732} 01/29/2022 01:03:56 - INFO - codeparrot_training - Step 9524: {'lr': 0.0004702949270909164, 'samples': 1828800, 'steps': 9524, 'loss/train': 1.2948206663131714} 01/29/2022 01:04:01 - INFO - codeparrot_training - Step 9525: {'lr': 0.0004702871907503169, 'samples': 1828992, 'steps': 9525, 'loss/train': 1.8408288359642029} 01/29/2022 01:04:05 - INFO - codeparrot_training - Step 9526: {'lr': 0.000470279453466077, 'samples': 1829184, 'steps': 9526, 'loss/train': 1.5947222113609314} 01/29/2022 01:04:09 - INFO - codeparrot_training - Step 9527: {'lr': 0.0004702717152382299, 'samples': 1829376, 'steps': 9527, 'loss/train': 2.0720871090888977} 01/29/2022 01:04:13 - INFO - codeparrot_training - Step 9528: {'lr': 0.0004702639760668086, 'samples': 1829568, 'steps': 9528, 'loss/train': 1.5243675112724304} 01/29/2022 01:04:20 - INFO - codeparrot_training - Step 9529: {'lr': 0.00047025623595184645, 'samples': 1829760, 'steps': 9529, 'loss/train': 2.313624680042267} 01/29/2022 01:04:24 - INFO - codeparrot_training - Step 9530: {'lr': 0.0004702484948933765, 'samples': 1829952, 'steps': 9530, 'loss/train': 1.3139685988426208} 01/29/2022 01:04:28 - INFO - codeparrot_training - Step 9531: {'lr': 0.000470240752891432, 'samples': 1830144, 'steps': 9531, 'loss/train': 1.4107933938503265} 01/29/2022 01:04:33 - INFO - codeparrot_training - Step 9532: {'lr': 0.000470233009946046, 'samples': 1830336, 'steps': 9532, 'loss/train': 0.9333900511264801} 01/29/2022 01:04:37 - INFO - codeparrot_training - Step 9533: {'lr': 0.0004702252660572517, 'samples': 1830528, 'steps': 9533, 'loss/train': 0.7539509832859039} 01/29/2022 01:04:42 - INFO - codeparrot_training - Step 9534: {'lr': 0.00047021752122508234, 'samples': 1830720, 'steps': 9534, 'loss/train': 1.2137131690979004} 01/29/2022 01:04:47 - INFO - codeparrot_training - Step 9535: {'lr': 0.000470209775449571, 'samples': 1830912, 'steps': 9535, 'loss/train': 3.535132884979248} 01/29/2022 01:04:51 - INFO - codeparrot_training - Step 9536: {'lr': 0.00047020202873075093, 'samples': 1831104, 'steps': 9536, 'loss/train': 3.066456913948059} 01/29/2022 01:04:55 - INFO - codeparrot_training - Step 9537: {'lr': 0.0004701942810686552, 'samples': 1831296, 'steps': 9537, 'loss/train': 1.862299919128418} 01/29/2022 01:04:59 - INFO - codeparrot_training - Step 9538: {'lr': 0.00047018653246331724, 'samples': 1831488, 'steps': 9538, 'loss/train': 2.1921608448028564} 01/29/2022 01:05:05 - INFO - codeparrot_training - Step 9539: {'lr': 0.00047017878291477, 'samples': 1831680, 'steps': 9539, 'loss/train': 1.9268487095832825} 01/29/2022 01:05:09 - INFO - codeparrot_training - Step 9540: {'lr': 0.0004701710324230468, 'samples': 1831872, 'steps': 9540, 'loss/train': 1.499432533979416} 01/29/2022 01:05:14 - INFO - codeparrot_training - Step 9541: {'lr': 0.00047016328098818086, 'samples': 1832064, 'steps': 9541, 'loss/train': 2.1580329537391663} 01/29/2022 01:05:18 - INFO - codeparrot_training - Step 9542: {'lr': 0.00047015552861020524, 'samples': 1832256, 'steps': 9542, 'loss/train': 1.0803640186786652} 01/29/2022 01:05:24 - INFO - codeparrot_training - Step 9543: {'lr': 0.00047014777528915327, 'samples': 1832448, 'steps': 9543, 'loss/train': 2.7000526785850525} 01/29/2022 01:05:28 - INFO - codeparrot_training - Step 9544: {'lr': 0.0004701400210250581, 'samples': 1832640, 'steps': 9544, 'loss/train': 1.7634199261665344} 01/29/2022 01:05:33 - INFO - codeparrot_training - Step 9545: {'lr': 0.00047013226581795305, 'samples': 1832832, 'steps': 9545, 'loss/train': 1.0401707589626312} 01/29/2022 01:05:37 - INFO - codeparrot_training - Step 9546: {'lr': 0.00047012450966787126, 'samples': 1833024, 'steps': 9546, 'loss/train': 1.8118799328804016} 01/29/2022 01:05:41 - INFO - codeparrot_training - Step 9547: {'lr': 0.000470116752574846, 'samples': 1833216, 'steps': 9547, 'loss/train': 1.4481662213802338} 01/29/2022 01:05:47 - INFO - codeparrot_training - Step 9548: {'lr': 0.0004701089945389104, 'samples': 1833408, 'steps': 9548, 'loss/train': 1.8191680312156677} 01/29/2022 01:05:51 - INFO - codeparrot_training - Step 9549: {'lr': 0.00047010123556009774, 'samples': 1833600, 'steps': 9549, 'loss/train': 1.869300663471222} 01/29/2022 01:05:55 - INFO - codeparrot_training - Step 9550: {'lr': 0.0004700934756384413, 'samples': 1833792, 'steps': 9550, 'loss/train': 1.498108685016632} 01/29/2022 01:05:59 - INFO - codeparrot_training - Step 9551: {'lr': 0.00047008571477397435, 'samples': 1833984, 'steps': 9551, 'loss/train': 1.2024531662464142} 01/29/2022 01:06:04 - INFO - codeparrot_training - Step 9552: {'lr': 0.00047007795296673006, 'samples': 1834176, 'steps': 9552, 'loss/train': 1.6947196125984192} 01/29/2022 01:06:09 - INFO - codeparrot_training - Step 9553: {'lr': 0.00047007019021674167, 'samples': 1834368, 'steps': 9553, 'loss/train': 2.3119030594825745} 01/29/2022 01:06:13 - INFO - codeparrot_training - Step 9554: {'lr': 0.0004700624265240425, 'samples': 1834560, 'steps': 9554, 'loss/train': 1.6426091194152832} 01/29/2022 01:06:17 - INFO - codeparrot_training - Step 9555: {'lr': 0.00047005466188866575, 'samples': 1834752, 'steps': 9555, 'loss/train': 1.5801016688346863} 01/29/2022 01:06:22 - INFO - codeparrot_training - Step 9556: {'lr': 0.00047004689631064474, 'samples': 1834944, 'steps': 9556, 'loss/train': 2.2503620982170105} 01/29/2022 01:06:26 - INFO - codeparrot_training - Step 9557: {'lr': 0.00047003912979001267, 'samples': 1835136, 'steps': 9557, 'loss/train': 1.671301245689392} 01/29/2022 01:06:32 - INFO - codeparrot_training - Step 9558: {'lr': 0.0004700313623268028, 'samples': 1835328, 'steps': 9558, 'loss/train': 0.40386343002319336} 01/29/2022 01:06:36 - INFO - codeparrot_training - Step 9559: {'lr': 0.00047002359392104854, 'samples': 1835520, 'steps': 9559, 'loss/train': 1.221272885799408} 01/29/2022 01:06:40 - INFO - codeparrot_training - Step 9560: {'lr': 0.000470015824572783, 'samples': 1835712, 'steps': 9560, 'loss/train': 2.6720014214515686} 01/29/2022 01:06:45 - INFO - codeparrot_training - Step 9561: {'lr': 0.00047000805428203953, 'samples': 1835904, 'steps': 9561, 'loss/train': 1.9177300930023193} 01/29/2022 01:06:49 - INFO - codeparrot_training - Step 9562: {'lr': 0.00047000028304885143, 'samples': 1836096, 'steps': 9562, 'loss/train': 2.0427074432373047} 01/29/2022 01:06:54 - INFO - codeparrot_training - Step 9563: {'lr': 0.00046999251087325204, 'samples': 1836288, 'steps': 9563, 'loss/train': 1.6737290024757385} 01/29/2022 01:06:58 - INFO - codeparrot_training - Step 9564: {'lr': 0.0004699847377552745, 'samples': 1836480, 'steps': 9564, 'loss/train': 2.428940176963806} 01/29/2022 01:07:03 - INFO - codeparrot_training - Step 9565: {'lr': 0.00046997696369495217, 'samples': 1836672, 'steps': 9565, 'loss/train': 1.6186455488204956} 01/29/2022 01:07:07 - INFO - codeparrot_training - Step 9566: {'lr': 0.00046996918869231843, 'samples': 1836864, 'steps': 9566, 'loss/train': 1.4094583690166473} 01/29/2022 01:07:11 - INFO - codeparrot_training - Step 9567: {'lr': 0.00046996141274740653, 'samples': 1837056, 'steps': 9567, 'loss/train': 1.2497654557228088} 01/29/2022 01:07:17 - INFO - codeparrot_training - Step 9568: {'lr': 0.00046995363586024977, 'samples': 1837248, 'steps': 9568, 'loss/train': 2.730341613292694} 01/29/2022 01:07:21 - INFO - codeparrot_training - Step 9569: {'lr': 0.0004699458580308815, 'samples': 1837440, 'steps': 9569, 'loss/train': 1.6501370072364807} 01/29/2022 01:07:25 - INFO - codeparrot_training - Step 9570: {'lr': 0.00046993807925933503, 'samples': 1837632, 'steps': 9570, 'loss/train': 1.992772400379181} 01/29/2022 01:07:30 - INFO - codeparrot_training - Step 9571: {'lr': 0.00046993029954564363, 'samples': 1837824, 'steps': 9571, 'loss/train': 1.5990614891052246} 01/29/2022 01:07:34 - INFO - codeparrot_training - Step 9572: {'lr': 0.0004699225188898407, 'samples': 1838016, 'steps': 9572, 'loss/train': 0.8068031966686249} 01/29/2022 01:07:40 - INFO - codeparrot_training - Step 9573: {'lr': 0.0004699147372919595, 'samples': 1838208, 'steps': 9573, 'loss/train': 1.5030779242515564} 01/29/2022 01:07:44 - INFO - codeparrot_training - Step 9574: {'lr': 0.00046990695475203337, 'samples': 1838400, 'steps': 9574, 'loss/train': 2.3961364030838013} 01/29/2022 01:07:48 - INFO - codeparrot_training - Step 9575: {'lr': 0.00046989917127009573, 'samples': 1838592, 'steps': 9575, 'loss/train': 2.1373761892318726} 01/29/2022 01:07:52 - INFO - codeparrot_training - Step 9576: {'lr': 0.0004698913868461798, 'samples': 1838784, 'steps': 9576, 'loss/train': 1.4180787205696106} 01/29/2022 01:07:57 - INFO - codeparrot_training - Step 9577: {'lr': 0.00046988360148031904, 'samples': 1838976, 'steps': 9577, 'loss/train': 1.41849884390831} 01/29/2022 01:08:02 - INFO - codeparrot_training - Step 9578: {'lr': 0.0004698758151725468, 'samples': 1839168, 'steps': 9578, 'loss/train': 1.5620452165603638} 01/29/2022 01:08:06 - INFO - codeparrot_training - Step 9579: {'lr': 0.0004698680279228963, 'samples': 1839360, 'steps': 9579, 'loss/train': 1.4320389926433563} 01/29/2022 01:08:10 - INFO - codeparrot_training - Step 9580: {'lr': 0.000469860239731401, 'samples': 1839552, 'steps': 9580, 'loss/train': 1.9286032319068909} 01/29/2022 01:08:15 - INFO - codeparrot_training - Step 9581: {'lr': 0.00046985245059809436, 'samples': 1839744, 'steps': 9581, 'loss/train': 1.95199853181839} 01/29/2022 01:08:19 - INFO - codeparrot_training - Step 9582: {'lr': 0.0004698446605230095, 'samples': 1839936, 'steps': 9582, 'loss/train': 2.0493741631507874} 01/29/2022 01:08:25 - INFO - codeparrot_training - Step 9583: {'lr': 0.00046983686950618, 'samples': 1840128, 'steps': 9583, 'loss/train': 1.966060996055603} 01/29/2022 01:08:29 - INFO - codeparrot_training - Step 9584: {'lr': 0.00046982907754763905, 'samples': 1840320, 'steps': 9584, 'loss/train': 1.2970342934131622} 01/29/2022 01:08:33 - INFO - codeparrot_training - Step 9585: {'lr': 0.00046982128464742026, 'samples': 1840512, 'steps': 9585, 'loss/train': 1.601638913154602} 01/29/2022 01:08:37 - INFO - codeparrot_training - Step 9586: {'lr': 0.0004698134908055568, 'samples': 1840704, 'steps': 9586, 'loss/train': 1.667349636554718} 01/29/2022 01:08:42 - INFO - codeparrot_training - Step 9587: {'lr': 0.00046980569602208215, 'samples': 1840896, 'steps': 9587, 'loss/train': 1.7288376688957214} 01/29/2022 01:08:48 - INFO - codeparrot_training - Step 9588: {'lr': 0.00046979790029702973, 'samples': 1841088, 'steps': 9588, 'loss/train': 1.972729504108429} 01/29/2022 01:08:52 - INFO - codeparrot_training - Step 9589: {'lr': 0.0004697901036304329, 'samples': 1841280, 'steps': 9589, 'loss/train': 2.448051631450653} 01/29/2022 01:08:56 - INFO - codeparrot_training - Step 9590: {'lr': 0.00046978230602232507, 'samples': 1841472, 'steps': 9590, 'loss/train': 0.9806842803955078} 01/29/2022 01:09:00 - INFO - codeparrot_training - Step 9591: {'lr': 0.00046977450747273956, 'samples': 1841664, 'steps': 9591, 'loss/train': 1.9131562113761902} 01/29/2022 01:09:05 - INFO - codeparrot_training - Step 9592: {'lr': 0.00046976670798171, 'samples': 1841856, 'steps': 9592, 'loss/train': 2.2307424545288086} 01/29/2022 01:09:10 - INFO - codeparrot_training - Step 9593: {'lr': 0.00046975890754926943, 'samples': 1842048, 'steps': 9593, 'loss/train': 2.0101452469825745} 01/29/2022 01:09:14 - INFO - codeparrot_training - Step 9594: {'lr': 0.0004697511061754516, 'samples': 1842240, 'steps': 9594, 'loss/train': 1.9503151774406433} 01/29/2022 01:09:18 - INFO - codeparrot_training - Step 9595: {'lr': 0.00046974330386028985, 'samples': 1842432, 'steps': 9595, 'loss/train': 2.044385075569153} 01/29/2022 01:09:23 - INFO - codeparrot_training - Step 9596: {'lr': 0.0004697355006038175, 'samples': 1842624, 'steps': 9596, 'loss/train': 1.1737596988677979} 01/29/2022 01:09:27 - INFO - codeparrot_training - Step 9597: {'lr': 0.00046972769640606804, 'samples': 1842816, 'steps': 9597, 'loss/train': 1.7438932657241821} 01/29/2022 01:09:32 - INFO - codeparrot_training - Step 9598: {'lr': 0.0004697198912670749, 'samples': 1843008, 'steps': 9598, 'loss/train': 1.1733096241950989} 01/29/2022 01:09:36 - INFO - codeparrot_training - Step 9599: {'lr': 0.0004697120851868715, 'samples': 1843200, 'steps': 9599, 'loss/train': 3.9273018836975098} 01/29/2022 01:09:41 - INFO - codeparrot_training - Step 9600: {'lr': 0.00046970427816549133, 'samples': 1843392, 'steps': 9600, 'loss/train': 2.143930435180664} 01/29/2022 01:09:45 - INFO - codeparrot_training - Step 9601: {'lr': 0.0004696964702029678, 'samples': 1843584, 'steps': 9601, 'loss/train': 1.0304186046123505} 01/29/2022 01:09:49 - INFO - codeparrot_training - Step 9602: {'lr': 0.00046968866129933436, 'samples': 1843776, 'steps': 9602, 'loss/train': 1.543471097946167} 01/29/2022 01:09:55 - INFO - codeparrot_training - Step 9603: {'lr': 0.0004696808514546244, 'samples': 1843968, 'steps': 9603, 'loss/train': 7.221921443939209} 01/29/2022 01:10:00 - INFO - codeparrot_training - Step 9604: {'lr': 0.0004696730406688715, 'samples': 1844160, 'steps': 9604, 'loss/train': 2.1220720410346985} 01/29/2022 01:10:04 - INFO - codeparrot_training - Step 9605: {'lr': 0.000469665228942109, 'samples': 1844352, 'steps': 9605, 'loss/train': 2.2865878343582153} 01/29/2022 01:10:08 - INFO - codeparrot_training - Step 9606: {'lr': 0.0004696574162743704, 'samples': 1844544, 'steps': 9606, 'loss/train': 1.5577313303947449} 01/29/2022 01:10:12 - INFO - codeparrot_training - Step 9607: {'lr': 0.00046964960266568926, 'samples': 1844736, 'steps': 9607, 'loss/train': 1.3426498174667358} 01/29/2022 01:10:18 - INFO - codeparrot_training - Step 9608: {'lr': 0.0004696417881160989, 'samples': 1844928, 'steps': 9608, 'loss/train': 2.0159060955047607} 01/29/2022 01:10:22 - INFO - codeparrot_training - Step 9609: {'lr': 0.0004696339726256328, 'samples': 1845120, 'steps': 9609, 'loss/train': 1.590851068496704} 01/29/2022 01:10:27 - INFO - codeparrot_training - Step 9610: {'lr': 0.00046962615619432457, 'samples': 1845312, 'steps': 9610, 'loss/train': 0.3315654695034027} 01/29/2022 01:10:31 - INFO - codeparrot_training - Step 9611: {'lr': 0.0004696183388222077, 'samples': 1845504, 'steps': 9611, 'loss/train': 1.5895466208457947} 01/29/2022 01:10:35 - INFO - codeparrot_training - Step 9612: {'lr': 0.0004696105205093155, 'samples': 1845696, 'steps': 9612, 'loss/train': 2.1326016783714294} 01/29/2022 01:10:41 - INFO - codeparrot_training - Step 9613: {'lr': 0.0004696027012556816, 'samples': 1845888, 'steps': 9613, 'loss/train': 0.9975491166114807} 01/29/2022 01:10:45 - INFO - codeparrot_training - Step 9614: {'lr': 0.00046959488106133944, 'samples': 1846080, 'steps': 9614, 'loss/train': 1.6877544522285461} 01/29/2022 01:10:49 - INFO - codeparrot_training - Step 9615: {'lr': 0.0004695870599263226, 'samples': 1846272, 'steps': 9615, 'loss/train': 2.952338218688965} 01/29/2022 01:10:54 - INFO - codeparrot_training - Step 9616: {'lr': 0.0004695792378506645, 'samples': 1846464, 'steps': 9616, 'loss/train': 1.5757144689559937} 01/29/2022 01:10:58 - INFO - codeparrot_training - Step 9617: {'lr': 0.00046957141483439856, 'samples': 1846656, 'steps': 9617, 'loss/train': 2.3053298592567444} 01/29/2022 01:11:03 - INFO - codeparrot_training - Step 9618: {'lr': 0.0004695635908775585, 'samples': 1846848, 'steps': 9618, 'loss/train': 0.5860447883605957} 01/29/2022 01:11:07 - INFO - codeparrot_training - Step 9619: {'lr': 0.0004695557659801778, 'samples': 1847040, 'steps': 9619, 'loss/train': 2.052815794944763} 01/29/2022 01:11:12 - INFO - codeparrot_training - Step 9620: {'lr': 0.0004695479401422898, 'samples': 1847232, 'steps': 9620, 'loss/train': 2.0494521260261536} 01/29/2022 01:11:16 - INFO - codeparrot_training - Step 9621: {'lr': 0.0004695401133639282, 'samples': 1847424, 'steps': 9621, 'loss/train': 1.1670379042625427} 01/29/2022 01:11:20 - INFO - codeparrot_training - Step 9622: {'lr': 0.0004695322856451264, 'samples': 1847616, 'steps': 9622, 'loss/train': 2.4032633900642395} 01/29/2022 01:11:25 - INFO - codeparrot_training - Step 9623: {'lr': 0.00046952445698591805, 'samples': 1847808, 'steps': 9623, 'loss/train': 0.9878394305706024} 01/29/2022 01:11:30 - INFO - codeparrot_training - Step 9624: {'lr': 0.0004695166273863367, 'samples': 1848000, 'steps': 9624, 'loss/train': 1.7101172804832458} 01/29/2022 01:11:34 - INFO - codeparrot_training - Step 9625: {'lr': 0.00046950879684641567, 'samples': 1848192, 'steps': 9625, 'loss/train': 2.4647406935691833} 01/29/2022 01:11:38 - INFO - codeparrot_training - Step 9626: {'lr': 0.00046950096536618876, 'samples': 1848384, 'steps': 9626, 'loss/train': 1.3459779918193817} 01/29/2022 01:11:42 - INFO - codeparrot_training - Step 9627: {'lr': 0.0004694931329456894, 'samples': 1848576, 'steps': 9627, 'loss/train': 0.9687640070915222} 01/29/2022 01:11:49 - INFO - codeparrot_training - Step 9628: {'lr': 0.0004694852995849511, 'samples': 1848768, 'steps': 9628, 'loss/train': 1.1475143730640411} 01/29/2022 01:11:53 - INFO - codeparrot_training - Step 9629: {'lr': 0.00046947746528400755, 'samples': 1848960, 'steps': 9629, 'loss/train': 1.3334544003009796} 01/29/2022 01:11:57 - INFO - codeparrot_training - Step 9630: {'lr': 0.00046946963004289223, 'samples': 1849152, 'steps': 9630, 'loss/train': 0.747911810874939} 01/29/2022 01:12:01 - INFO - codeparrot_training - Step 9631: {'lr': 0.0004694617938616386, 'samples': 1849344, 'steps': 9631, 'loss/train': 0.8688054978847504} 01/29/2022 01:12:05 - INFO - codeparrot_training - Step 9632: {'lr': 0.00046945395674028047, 'samples': 1849536, 'steps': 9632, 'loss/train': 4.5370484590530396} 01/29/2022 01:12:11 - INFO - codeparrot_training - Step 9633: {'lr': 0.0004694461186788512, 'samples': 1849728, 'steps': 9633, 'loss/train': 2.220086395740509} 01/29/2022 01:12:15 - INFO - codeparrot_training - Step 9634: {'lr': 0.0004694382796773844, 'samples': 1849920, 'steps': 9634, 'loss/train': 2.151846170425415} 01/29/2022 01:12:19 - INFO - codeparrot_training - Step 9635: {'lr': 0.0004694304397359137, 'samples': 1850112, 'steps': 9635, 'loss/train': 1.9580004215240479} 01/29/2022 01:12:24 - INFO - codeparrot_training - Step 9636: {'lr': 0.00046942259885447273, 'samples': 1850304, 'steps': 9636, 'loss/train': 1.8437672853469849} 01/29/2022 01:12:29 - INFO - codeparrot_training - Step 9637: {'lr': 0.000469414757033095, 'samples': 1850496, 'steps': 9637, 'loss/train': 1.9465879797935486} 01/29/2022 01:12:33 - INFO - codeparrot_training - Step 9638: {'lr': 0.00046940691427181414, 'samples': 1850688, 'steps': 9638, 'loss/train': 2.080633521080017} 01/29/2022 01:12:38 - INFO - codeparrot_training - Step 9639: {'lr': 0.00046939907057066374, 'samples': 1850880, 'steps': 9639, 'loss/train': 1.7749038934707642} 01/29/2022 01:12:42 - INFO - codeparrot_training - Step 9640: {'lr': 0.0004693912259296773, 'samples': 1851072, 'steps': 9640, 'loss/train': 1.5309154987335205} 01/29/2022 01:12:46 - INFO - codeparrot_training - Step 9641: {'lr': 0.0004693833803488886, 'samples': 1851264, 'steps': 9641, 'loss/train': 1.7835994362831116} 01/29/2022 01:12:51 - INFO - codeparrot_training - Step 9642: {'lr': 0.00046937553382833116, 'samples': 1851456, 'steps': 9642, 'loss/train': 1.7628863453865051} 01/29/2022 01:12:55 - INFO - codeparrot_training - Step 9643: {'lr': 0.00046936768636803857, 'samples': 1851648, 'steps': 9643, 'loss/train': 1.4445347785949707} 01/29/2022 01:13:00 - INFO - codeparrot_training - Step 9644: {'lr': 0.00046935983796804443, 'samples': 1851840, 'steps': 9644, 'loss/train': 1.3718561232089996} 01/29/2022 01:13:04 - INFO - codeparrot_training - Step 9645: {'lr': 0.00046935198862838246, 'samples': 1852032, 'steps': 9645, 'loss/train': 1.6837109327316284} 01/29/2022 01:13:08 - INFO - codeparrot_training - Step 9646: {'lr': 0.00046934413834908616, 'samples': 1852224, 'steps': 9646, 'loss/train': 2.082937180995941} 01/29/2022 01:15:25 - INFO - codeparrot_training - Step 9647: {'lr': 0.0004693362871301893, 'samples': 1852416, 'steps': 9647, 'loss/train': 2.034714102745056} 01/29/2022 01:15:29 - INFO - codeparrot_training - Step 9648: {'lr': 0.0004693284349717254, 'samples': 1852608, 'steps': 9648, 'loss/train': 2.057061195373535} 01/29/2022 01:15:33 - INFO - codeparrot_training - Step 9649: {'lr': 0.00046932058187372803, 'samples': 1852800, 'steps': 9649, 'loss/train': 1.7910794019699097} 01/29/2022 01:15:38 - INFO - codeparrot_training - Step 9650: {'lr': 0.00046931272783623106, 'samples': 1852992, 'steps': 9650, 'loss/train': 1.646960735321045} 01/29/2022 01:15:42 - INFO - codeparrot_training - Step 9651: {'lr': 0.00046930487285926797, 'samples': 1853184, 'steps': 9651, 'loss/train': 1.5344740748405457} 01/29/2022 01:15:47 - INFO - codeparrot_training - Step 9652: {'lr': 0.00046929701694287243, 'samples': 1853376, 'steps': 9652, 'loss/train': 1.2082862555980682} 01/29/2022 01:15:51 - INFO - codeparrot_training - Step 9653: {'lr': 0.0004692891600870781, 'samples': 1853568, 'steps': 9653, 'loss/train': 2.1247231364250183} 01/29/2022 01:15:56 - INFO - codeparrot_training - Step 9654: {'lr': 0.00046928130229191865, 'samples': 1853760, 'steps': 9654, 'loss/train': 1.8114972710609436} 01/29/2022 01:16:00 - INFO - codeparrot_training - Step 9655: {'lr': 0.00046927344355742774, 'samples': 1853952, 'steps': 9655, 'loss/train': 1.8089634776115417} 01/29/2022 01:16:04 - INFO - codeparrot_training - Step 9656: {'lr': 0.00046926558388363904, 'samples': 1854144, 'steps': 9656, 'loss/train': 2.259511113166809} 01/29/2022 01:16:10 - INFO - codeparrot_training - Step 9657: {'lr': 0.00046925772327058616, 'samples': 1854336, 'steps': 9657, 'loss/train': 1.353456974029541} 01/29/2022 01:16:14 - INFO - codeparrot_training - Step 9658: {'lr': 0.0004692498617183028, 'samples': 1854528, 'steps': 9658, 'loss/train': 1.2818047106266022} 01/29/2022 01:16:19 - INFO - codeparrot_training - Step 9659: {'lr': 0.0004692419992268227, 'samples': 1854720, 'steps': 9659, 'loss/train': 1.727177917957306} 01/29/2022 01:16:23 - INFO - codeparrot_training - Step 9660: {'lr': 0.00046923413579617944, 'samples': 1854912, 'steps': 9660, 'loss/train': 1.1926890313625336} 01/29/2022 01:16:27 - INFO - codeparrot_training - Step 9661: {'lr': 0.00046922627142640685, 'samples': 1855104, 'steps': 9661, 'loss/train': 1.0488232970237732} 01/29/2022 01:16:33 - INFO - codeparrot_training - Step 9662: {'lr': 0.00046921840611753845, 'samples': 1855296, 'steps': 9662, 'loss/train': 2.262908399105072} 01/29/2022 01:16:37 - INFO - codeparrot_training - Step 9663: {'lr': 0.000469210539869608, 'samples': 1855488, 'steps': 9663, 'loss/train': 1.1918239295482635} 01/29/2022 01:16:41 - INFO - codeparrot_training - Step 9664: {'lr': 0.0004692026726826493, 'samples': 1855680, 'steps': 9664, 'loss/train': 2.213357090950012} 01/29/2022 01:16:45 - INFO - codeparrot_training - Step 9665: {'lr': 0.0004691948045566958, 'samples': 1855872, 'steps': 9665, 'loss/train': 2.488498628139496} 01/29/2022 01:16:50 - INFO - codeparrot_training - Step 9666: {'lr': 0.0004691869354917815, 'samples': 1856064, 'steps': 9666, 'loss/train': 1.573985517024994} 01/29/2022 01:16:55 - INFO - codeparrot_training - Step 9667: {'lr': 0.0004691790654879399, 'samples': 1856256, 'steps': 9667, 'loss/train': 1.46282958984375} 01/29/2022 01:16:59 - INFO - codeparrot_training - Step 9668: {'lr': 0.00046917119454520487, 'samples': 1856448, 'steps': 9668, 'loss/train': 0.1398853361606598} 01/29/2022 01:17:03 - INFO - codeparrot_training - Step 9669: {'lr': 0.0004691633226636099, 'samples': 1856640, 'steps': 9669, 'loss/train': 1.9801536798477173} 01/29/2022 01:17:08 - INFO - codeparrot_training - Step 9670: {'lr': 0.0004691554498431889, 'samples': 1856832, 'steps': 9670, 'loss/train': 1.9953957796096802} 01/29/2022 01:17:12 - INFO - codeparrot_training - Step 9671: {'lr': 0.00046914757608397555, 'samples': 1857024, 'steps': 9671, 'loss/train': 2.3390203714370728} 01/29/2022 01:17:18 - INFO - codeparrot_training - Step 9672: {'lr': 0.00046913970138600357, 'samples': 1857216, 'steps': 9672, 'loss/train': 1.0483469367027283} 01/29/2022 01:17:22 - INFO - codeparrot_training - Step 9673: {'lr': 0.0004691318257493067, 'samples': 1857408, 'steps': 9673, 'loss/train': 1.7234503626823425} 01/29/2022 01:17:26 - INFO - codeparrot_training - Step 9674: {'lr': 0.00046912394917391866, 'samples': 1857600, 'steps': 9674, 'loss/train': 1.7290744185447693} 01/29/2022 01:17:31 - INFO - codeparrot_training - Step 9675: {'lr': 0.00046911607165987324, 'samples': 1857792, 'steps': 9675, 'loss/train': 1.5530431866645813} 01/29/2022 01:17:35 - INFO - codeparrot_training - Step 9676: {'lr': 0.0004691081932072041, 'samples': 1857984, 'steps': 9676, 'loss/train': 2.220620334148407} 01/29/2022 01:17:40 - INFO - codeparrot_training - Step 9677: {'lr': 0.0004691003138159451, 'samples': 1858176, 'steps': 9677, 'loss/train': 1.6644229888916016} 01/29/2022 01:17:44 - INFO - codeparrot_training - Step 9678: {'lr': 0.00046909243348612986, 'samples': 1858368, 'steps': 9678, 'loss/train': 1.2498251795768738} 01/29/2022 01:17:49 - INFO - codeparrot_training - Step 9679: {'lr': 0.0004690845522177922, 'samples': 1858560, 'steps': 9679, 'loss/train': 1.0261994004249573} 01/29/2022 01:17:53 - INFO - codeparrot_training - Step 9680: {'lr': 0.0004690766700109659, 'samples': 1858752, 'steps': 9680, 'loss/train': 1.3344587087631226} 01/29/2022 01:17:57 - INFO - codeparrot_training - Step 9681: {'lr': 0.0004690687868656847, 'samples': 1858944, 'steps': 9681, 'loss/train': 1.6408100724220276} 01/29/2022 01:18:04 - INFO - codeparrot_training - Step 9682: {'lr': 0.00046906090278198246, 'samples': 1859136, 'steps': 9682, 'loss/train': 1.6939215660095215} 01/29/2022 01:18:08 - INFO - codeparrot_training - Step 9683: {'lr': 0.00046905301775989277, 'samples': 1859328, 'steps': 9683, 'loss/train': 0.8884618878364563} 01/29/2022 01:18:13 - INFO - codeparrot_training - Step 9684: {'lr': 0.0004690451317994495, 'samples': 1859520, 'steps': 9684, 'loss/train': 1.4760176539421082} 01/29/2022 01:18:17 - INFO - codeparrot_training - Step 9685: {'lr': 0.00046903724490068654, 'samples': 1859712, 'steps': 9685, 'loss/train': 0.720037579536438} 01/29/2022 01:18:21 - INFO - codeparrot_training - Step 9686: {'lr': 0.00046902935706363754, 'samples': 1859904, 'steps': 9686, 'loss/train': 1.758859097957611} 01/29/2022 01:18:26 - INFO - codeparrot_training - Step 9687: {'lr': 0.0004690214682883363, 'samples': 1860096, 'steps': 9687, 'loss/train': 2.149196147918701} 01/29/2022 01:18:31 - INFO - codeparrot_training - Step 9688: {'lr': 0.00046901357857481664, 'samples': 1860288, 'steps': 9688, 'loss/train': 2.659943640232086} 01/29/2022 01:18:35 - INFO - codeparrot_training - Step 9689: {'lr': 0.0004690056879231124, 'samples': 1860480, 'steps': 9689, 'loss/train': 1.6668460965156555} 01/29/2022 01:18:39 - INFO - codeparrot_training - Step 9690: {'lr': 0.0004689977963332572, 'samples': 1860672, 'steps': 9690, 'loss/train': 2.0242722630500793} 01/29/2022 01:18:43 - INFO - codeparrot_training - Step 9691: {'lr': 0.0004689899038052852, 'samples': 1860864, 'steps': 9691, 'loss/train': 2.7598180174827576} 01/29/2022 01:18:48 - INFO - codeparrot_training - Step 9692: {'lr': 0.0004689820103392298, 'samples': 1861056, 'steps': 9692, 'loss/train': 1.6849719285964966} 01/29/2022 01:18:53 - INFO - codeparrot_training - Step 9693: {'lr': 0.0004689741159351251, 'samples': 1861248, 'steps': 9693, 'loss/train': 1.6866796016693115} 01/29/2022 01:18:57 - INFO - codeparrot_training - Step 9694: {'lr': 0.00046896622059300477, 'samples': 1861440, 'steps': 9694, 'loss/train': 1.807419240474701} 01/29/2022 01:19:02 - INFO - codeparrot_training - Step 9695: {'lr': 0.00046895832431290266, 'samples': 1861632, 'steps': 9695, 'loss/train': 1.4603020548820496} 01/29/2022 01:19:06 - INFO - codeparrot_training - Step 9696: {'lr': 0.0004689504270948527, 'samples': 1861824, 'steps': 9696, 'loss/train': 2.0549249053001404} 01/29/2022 01:19:10 - INFO - codeparrot_training - Step 9697: {'lr': 0.00046894252893888854, 'samples': 1862016, 'steps': 9697, 'loss/train': 1.5183470249176025} 01/29/2022 01:19:16 - INFO - codeparrot_training - Step 9698: {'lr': 0.0004689346298450442, 'samples': 1862208, 'steps': 9698, 'loss/train': 0.9426255226135254} 01/29/2022 01:19:20 - INFO - codeparrot_training - Step 9699: {'lr': 0.0004689267298133534, 'samples': 1862400, 'steps': 9699, 'loss/train': 1.6320045590400696} 01/29/2022 01:19:24 - INFO - codeparrot_training - Step 9700: {'lr': 0.00046891882884384997, 'samples': 1862592, 'steps': 9700, 'loss/train': 1.4761103689670563} 01/29/2022 01:19:28 - INFO - codeparrot_training - Step 9701: {'lr': 0.00046891092693656777, 'samples': 1862784, 'steps': 9701, 'loss/train': 0.6966546177864075} 01/29/2022 01:19:33 - INFO - codeparrot_training - Step 9702: {'lr': 0.0004689030240915407, 'samples': 1862976, 'steps': 9702, 'loss/train': 2.2843320965766907} 01/29/2022 01:19:37 - INFO - codeparrot_training - Step 9703: {'lr': 0.0004688951203088026, 'samples': 1863168, 'steps': 9703, 'loss/train': 7.098297357559204} 01/29/2022 01:19:43 - INFO - codeparrot_training - Step 9704: {'lr': 0.00046888721558838734, 'samples': 1863360, 'steps': 9704, 'loss/train': 1.3342247307300568} 01/29/2022 01:19:47 - INFO - codeparrot_training - Step 9705: {'lr': 0.0004688793099303287, 'samples': 1863552, 'steps': 9705, 'loss/train': 1.715222418308258} 01/29/2022 01:19:52 - INFO - codeparrot_training - Step 9706: {'lr': 0.0004688714033346606, 'samples': 1863744, 'steps': 9706, 'loss/train': 1.8904001712799072} 01/29/2022 01:19:56 - INFO - codeparrot_training - Step 9707: {'lr': 0.000468863495801417, 'samples': 1863936, 'steps': 9707, 'loss/train': 1.828139305114746} 01/29/2022 01:20:01 - INFO - codeparrot_training - Step 9708: {'lr': 0.00046885558733063157, 'samples': 1864128, 'steps': 9708, 'loss/train': 2.0095374584198} 01/29/2022 01:20:05 - INFO - codeparrot_training - Step 9709: {'lr': 0.00046884767792233827, 'samples': 1864320, 'steps': 9709, 'loss/train': 1.5546253323554993} 01/29/2022 01:20:09 - INFO - codeparrot_training - Step 9710: {'lr': 0.00046883976757657107, 'samples': 1864512, 'steps': 9710, 'loss/train': 2.5243438482284546} 01/29/2022 01:20:14 - INFO - codeparrot_training - Step 9711: {'lr': 0.00046883185629336386, 'samples': 1864704, 'steps': 9711, 'loss/train': 1.821349024772644} 01/29/2022 01:20:18 - INFO - codeparrot_training - Step 9712: {'lr': 0.0004688239440727504, 'samples': 1864896, 'steps': 9712, 'loss/train': 1.6845030784606934} 01/29/2022 01:20:24 - INFO - codeparrot_training - Step 9713: {'lr': 0.00046881603091476466, 'samples': 1865088, 'steps': 9713, 'loss/train': 0.8912531733512878} 01/29/2022 01:20:28 - INFO - codeparrot_training - Step 9714: {'lr': 0.0004688081168194405, 'samples': 1865280, 'steps': 9714, 'loss/train': 0.5207389891147614} 01/29/2022 01:20:33 - INFO - codeparrot_training - Step 9715: {'lr': 0.0004688002017868119, 'samples': 1865472, 'steps': 9715, 'loss/train': 1.8785253167152405} 01/29/2022 01:20:37 - INFO - codeparrot_training - Step 9716: {'lr': 0.0004687922858169126, 'samples': 1865664, 'steps': 9716, 'loss/train': 1.773241639137268} 01/29/2022 01:20:41 - INFO - codeparrot_training - Step 9717: {'lr': 0.0004687843689097767, 'samples': 1865856, 'steps': 9717, 'loss/train': 2.1754669547080994} 01/29/2022 01:20:45 - INFO - codeparrot_training - Step 9718: {'lr': 0.0004687764510654381, 'samples': 1866048, 'steps': 9718, 'loss/train': 1.8053396344184875} 01/29/2022 01:20:51 - INFO - codeparrot_training - Step 9719: {'lr': 0.0004687685322839306, 'samples': 1866240, 'steps': 9719, 'loss/train': 1.5213770270347595} 01/29/2022 01:20:56 - INFO - codeparrot_training - Step 9720: {'lr': 0.00046876061256528813, 'samples': 1866432, 'steps': 9720, 'loss/train': 1.6700393557548523} 01/29/2022 01:21:00 - INFO - codeparrot_training - Step 9721: {'lr': 0.00046875269190954465, 'samples': 1866624, 'steps': 9721, 'loss/train': 4.879400253295898} 01/29/2022 01:21:04 - INFO - codeparrot_training - Step 9722: {'lr': 0.00046874477031673417, 'samples': 1866816, 'steps': 9722, 'loss/train': 2.434549570083618} 01/29/2022 01:21:08 - INFO - codeparrot_training - Step 9723: {'lr': 0.00046873684778689053, 'samples': 1867008, 'steps': 9723, 'loss/train': 2.0805176496505737} 01/29/2022 01:21:14 - INFO - codeparrot_training - Step 9724: {'lr': 0.00046872892432004765, 'samples': 1867200, 'steps': 9724, 'loss/train': 1.8987810015678406} 01/29/2022 01:21:18 - INFO - codeparrot_training - Step 9725: {'lr': 0.00046872099991623954, 'samples': 1867392, 'steps': 9725, 'loss/train': 1.6098918914794922} 01/29/2022 01:21:22 - INFO - codeparrot_training - Step 9726: {'lr': 0.0004687130745755002, 'samples': 1867584, 'steps': 9726, 'loss/train': 1.7799514532089233} 01/29/2022 01:21:27 - INFO - codeparrot_training - Step 9727: {'lr': 0.0004687051482978634, 'samples': 1867776, 'steps': 9727, 'loss/train': 1.2879065573215485} 01/29/2022 01:21:31 - INFO - codeparrot_training - Step 9728: {'lr': 0.0004686972210833632, 'samples': 1867968, 'steps': 9728, 'loss/train': 2.0510222911834717} 01/29/2022 01:21:37 - INFO - codeparrot_training - Step 9729: {'lr': 0.00046868929293203355, 'samples': 1868160, 'steps': 9729, 'loss/train': 1.3882888555526733} 01/29/2022 01:21:41 - INFO - codeparrot_training - Step 9730: {'lr': 0.0004686813638439085, 'samples': 1868352, 'steps': 9730, 'loss/train': 2.0647348165512085} 01/29/2022 01:21:45 - INFO - codeparrot_training - Step 9731: {'lr': 0.00046867343381902185, 'samples': 1868544, 'steps': 9731, 'loss/train': 2.0532265305519104} 01/29/2022 01:21:49 - INFO - codeparrot_training - Step 9732: {'lr': 0.0004686655028574076, 'samples': 1868736, 'steps': 9732, 'loss/train': 2.4569685459136963} 01/29/2022 01:21:54 - INFO - codeparrot_training - Step 9733: {'lr': 0.0004686575709590998, 'samples': 1868928, 'steps': 9733, 'loss/train': 1.5346006751060486} 01/29/2022 01:21:59 - INFO - codeparrot_training - Step 9734: {'lr': 0.00046864963812413244, 'samples': 1869120, 'steps': 9734, 'loss/train': 2.1687650084495544} 01/29/2022 01:22:03 - INFO - codeparrot_training - Step 9735: {'lr': 0.00046864170435253946, 'samples': 1869312, 'steps': 9735, 'loss/train': 2.177955150604248} 01/29/2022 01:22:08 - INFO - codeparrot_training - Step 9736: {'lr': 0.0004686337696443548, 'samples': 1869504, 'steps': 9736, 'loss/train': 2.3539334535598755} 01/29/2022 01:22:12 - INFO - codeparrot_training - Step 9737: {'lr': 0.0004686258339996125, 'samples': 1869696, 'steps': 9737, 'loss/train': 2.4908734560012817} 01/29/2022 01:22:16 - INFO - codeparrot_training - Step 9738: {'lr': 0.0004686178974183466, 'samples': 1869888, 'steps': 9738, 'loss/train': 1.8888516426086426} 01/29/2022 01:22:21 - INFO - codeparrot_training - Step 9739: {'lr': 0.00046860995990059096, 'samples': 1870080, 'steps': 9739, 'loss/train': 1.1233705580234528} 01/29/2022 01:22:26 - INFO - codeparrot_training - Step 9740: {'lr': 0.00046860202144637976, 'samples': 1870272, 'steps': 9740, 'loss/train': 0.44939830899238586} 01/29/2022 01:22:30 - INFO - codeparrot_training - Step 9741: {'lr': 0.0004685940820557468, 'samples': 1870464, 'steps': 9741, 'loss/train': 0.7227788865566254} 01/29/2022 01:22:34 - INFO - codeparrot_training - Step 9742: {'lr': 0.0004685861417287263, 'samples': 1870656, 'steps': 9742, 'loss/train': 1.8168364763259888} 01/29/2022 01:22:38 - INFO - codeparrot_training - Step 9743: {'lr': 0.00046857820046535215, 'samples': 1870848, 'steps': 9743, 'loss/train': 1.389558881521225} 01/29/2022 01:22:47 - INFO - codeparrot_training - Step 9744: {'lr': 0.0004685702582656584, 'samples': 1871040, 'steps': 9744, 'loss/train': 1.9526170492172241} 01/29/2022 01:22:51 - INFO - codeparrot_training - Step 9745: {'lr': 0.0004685623151296791, 'samples': 1871232, 'steps': 9745, 'loss/train': 2.172942280769348} 01/29/2022 01:22:55 - INFO - codeparrot_training - Step 9746: {'lr': 0.0004685543710574482, 'samples': 1871424, 'steps': 9746, 'loss/train': 2.134368360042572} 01/29/2022 01:22:59 - INFO - codeparrot_training - Step 9747: {'lr': 0.00046854642604899976, 'samples': 1871616, 'steps': 9747, 'loss/train': 0.3543017506599426} 01/29/2022 01:23:04 - INFO - codeparrot_training - Step 9748: {'lr': 0.00046853848010436783, 'samples': 1871808, 'steps': 9748, 'loss/train': 1.4770959913730621} 01/29/2022 01:23:09 - INFO - codeparrot_training - Step 9749: {'lr': 0.00046853053322358653, 'samples': 1872000, 'steps': 9749, 'loss/train': 1.7871851921081543} 01/29/2022 01:23:13 - INFO - codeparrot_training - Step 9750: {'lr': 0.00046852258540668973, 'samples': 1872192, 'steps': 9750, 'loss/train': 2.183422029018402} 01/29/2022 01:23:17 - INFO - codeparrot_training - Step 9751: {'lr': 0.0004685146366537116, 'samples': 1872384, 'steps': 9751, 'loss/train': 1.4227757155895233} 01/29/2022 01:23:22 - INFO - codeparrot_training - Step 9752: {'lr': 0.00046850668696468614, 'samples': 1872576, 'steps': 9752, 'loss/train': 1.96302330493927} 01/29/2022 01:23:26 - INFO - codeparrot_training - Step 9753: {'lr': 0.0004684987363396474, 'samples': 1872768, 'steps': 9753, 'loss/train': 1.644569456577301} 01/29/2022 01:23:31 - INFO - codeparrot_training - Step 9754: {'lr': 0.0004684907847786295, 'samples': 1872960, 'steps': 9754, 'loss/train': 1.8175325989723206} 01/29/2022 01:23:35 - INFO - codeparrot_training - Step 9755: {'lr': 0.0004684828322816664, 'samples': 1873152, 'steps': 9755, 'loss/train': 1.616478145122528} 01/29/2022 01:23:40 - INFO - codeparrot_training - Step 9756: {'lr': 0.00046847487884879227, 'samples': 1873344, 'steps': 9756, 'loss/train': 2.213066339492798} 01/29/2022 01:23:44 - INFO - codeparrot_training - Step 9757: {'lr': 0.0004684669244800411, 'samples': 1873536, 'steps': 9757, 'loss/train': 1.8474072217941284} 01/29/2022 01:23:48 - INFO - codeparrot_training - Step 9758: {'lr': 0.00046845896917544703, 'samples': 1873728, 'steps': 9758, 'loss/train': 1.760368287563324} 01/29/2022 01:23:53 - INFO - codeparrot_training - Step 9759: {'lr': 0.00046845101293504403, 'samples': 1873920, 'steps': 9759, 'loss/train': 1.9760357737541199} 01/29/2022 01:23:58 - INFO - codeparrot_training - Step 9760: {'lr': 0.00046844305575886636, 'samples': 1874112, 'steps': 9760, 'loss/train': 1.8220483660697937} 01/29/2022 01:24:02 - INFO - codeparrot_training - Step 9761: {'lr': 0.00046843509764694794, 'samples': 1874304, 'steps': 9761, 'loss/train': 1.616989016532898} 01/29/2022 01:24:06 - INFO - codeparrot_training - Step 9762: {'lr': 0.0004684271385993229, 'samples': 1874496, 'steps': 9762, 'loss/train': 1.7950467467308044} 01/29/2022 01:24:10 - INFO - codeparrot_training - Step 9763: {'lr': 0.0004684191786160254, 'samples': 1874688, 'steps': 9763, 'loss/train': 1.7603141069412231} 01/29/2022 01:24:16 - INFO - codeparrot_training - Step 9764: {'lr': 0.0004684112176970895, 'samples': 1874880, 'steps': 9764, 'loss/train': 1.4921190440654755} 01/29/2022 01:24:21 - INFO - codeparrot_training - Step 9765: {'lr': 0.0004684032558425493, 'samples': 1875072, 'steps': 9765, 'loss/train': 1.5256755352020264} 01/29/2022 01:24:25 - INFO - codeparrot_training - Step 9766: {'lr': 0.00046839529305243885, 'samples': 1875264, 'steps': 9766, 'loss/train': 1.668056845664978} 01/29/2022 01:24:29 - INFO - codeparrot_training - Step 9767: {'lr': 0.00046838732932679236, 'samples': 1875456, 'steps': 9767, 'loss/train': 2.3340150117874146} 01/29/2022 01:24:33 - INFO - codeparrot_training - Step 9768: {'lr': 0.0004683793646656439, 'samples': 1875648, 'steps': 9768, 'loss/train': 1.969355821609497} 01/29/2022 01:24:39 - INFO - codeparrot_training - Step 9769: {'lr': 0.00046837139906902753, 'samples': 1875840, 'steps': 9769, 'loss/train': 1.6432669758796692} 01/29/2022 01:24:43 - INFO - codeparrot_training - Step 9770: {'lr': 0.00046836343253697744, 'samples': 1876032, 'steps': 9770, 'loss/train': 1.9594942331314087} 01/29/2022 01:24:47 - INFO - codeparrot_training - Step 9771: {'lr': 0.0004683554650695278, 'samples': 1876224, 'steps': 9771, 'loss/train': 1.6757846474647522} 01/29/2022 01:24:52 - INFO - codeparrot_training - Step 9772: {'lr': 0.0004683474966667127, 'samples': 1876416, 'steps': 9772, 'loss/train': 1.6863648891448975} 01/29/2022 01:24:56 - INFO - codeparrot_training - Step 9773: {'lr': 0.00046833952732856614, 'samples': 1876608, 'steps': 9773, 'loss/train': 2.2209736704826355} 01/29/2022 01:25:01 - INFO - codeparrot_training - Step 9774: {'lr': 0.00046833155705512246, 'samples': 1876800, 'steps': 9774, 'loss/train': 1.8259227275848389} 01/29/2022 01:25:05 - INFO - codeparrot_training - Step 9775: {'lr': 0.0004683235858464157, 'samples': 1876992, 'steps': 9775, 'loss/train': 1.8076820969581604} 01/29/2022 01:25:09 - INFO - codeparrot_training - Step 9776: {'lr': 0.0004683156137024801, 'samples': 1877184, 'steps': 9776, 'loss/train': 1.7927062511444092} 01/29/2022 01:25:14 - INFO - codeparrot_training - Step 9777: {'lr': 0.0004683076406233496, 'samples': 1877376, 'steps': 9777, 'loss/train': 2.31929612159729} 01/29/2022 01:25:18 - INFO - codeparrot_training - Step 9778: {'lr': 0.0004682996666090585, 'samples': 1877568, 'steps': 9778, 'loss/train': 2.2994787096977234} 01/29/2022 01:25:24 - INFO - codeparrot_training - Step 9779: {'lr': 0.00046829169165964104, 'samples': 1877760, 'steps': 9779, 'loss/train': 1.6763180494308472} 01/29/2022 01:25:28 - INFO - codeparrot_training - Step 9780: {'lr': 0.0004682837157751313, 'samples': 1877952, 'steps': 9780, 'loss/train': 2.1111525893211365} 01/29/2022 01:25:32 - INFO - codeparrot_training - Step 9781: {'lr': 0.00046827573895556334, 'samples': 1878144, 'steps': 9781, 'loss/train': 2.8244728446006775} 01/29/2022 01:25:36 - INFO - codeparrot_training - Step 9782: {'lr': 0.00046826776120097147, 'samples': 1878336, 'steps': 9782, 'loss/train': 1.9773026704788208} 01/29/2022 01:25:41 - INFO - codeparrot_training - Step 9783: {'lr': 0.0004682597825113898, 'samples': 1878528, 'steps': 9783, 'loss/train': 1.6807846426963806} 01/29/2022 01:25:46 - INFO - codeparrot_training - Step 9784: {'lr': 0.00046825180288685253, 'samples': 1878720, 'steps': 9784, 'loss/train': 1.642661690711975} 01/29/2022 01:25:50 - INFO - codeparrot_training - Step 9785: {'lr': 0.00046824382232739386, 'samples': 1878912, 'steps': 9785, 'loss/train': 2.2643997073173523} 01/29/2022 01:25:55 - INFO - codeparrot_training - Step 9786: {'lr': 0.00046823584083304794, 'samples': 1879104, 'steps': 9786, 'loss/train': 2.134947717189789} 01/29/2022 01:25:59 - INFO - codeparrot_training - Step 9787: {'lr': 0.00046822785840384897, 'samples': 1879296, 'steps': 9787, 'loss/train': 0.6268147230148315} 01/29/2022 01:26:03 - INFO - codeparrot_training - Step 9788: {'lr': 0.0004682198750398312, 'samples': 1879488, 'steps': 9788, 'loss/train': 2.501849949359894} 01/29/2022 01:26:09 - INFO - codeparrot_training - Step 9789: {'lr': 0.0004682118907410287, 'samples': 1879680, 'steps': 9789, 'loss/train': 1.2114023566246033} 01/29/2022 01:26:13 - INFO - codeparrot_training - Step 9790: {'lr': 0.00046820390550747585, 'samples': 1879872, 'steps': 9790, 'loss/train': 1.7028965950012207} 01/29/2022 01:26:18 - INFO - codeparrot_training - Step 9791: {'lr': 0.0004681959193392067, 'samples': 1880064, 'steps': 9791, 'loss/train': 2.130503475666046} 01/29/2022 01:26:22 - INFO - codeparrot_training - Step 9792: {'lr': 0.00046818793223625543, 'samples': 1880256, 'steps': 9792, 'loss/train': 1.99241441488266} 01/29/2022 01:26:26 - INFO - codeparrot_training - Step 9793: {'lr': 0.0004681799441986564, 'samples': 1880448, 'steps': 9793, 'loss/train': 1.362257033586502} 01/29/2022 01:26:31 - INFO - codeparrot_training - Step 9794: {'lr': 0.00046817195522644387, 'samples': 1880640, 'steps': 9794, 'loss/train': 1.0296496152877808} 01/29/2022 01:26:36 - INFO - codeparrot_training - Step 9795: {'lr': 0.00046816396531965186, 'samples': 1880832, 'steps': 9795, 'loss/train': 0.9851279854774475} 01/29/2022 01:26:40 - INFO - codeparrot_training - Step 9796: {'lr': 0.0004681559744783147, 'samples': 1881024, 'steps': 9796, 'loss/train': 1.4126262366771698} 01/29/2022 01:26:44 - INFO - codeparrot_training - Step 9797: {'lr': 0.00046814798270246663, 'samples': 1881216, 'steps': 9797, 'loss/train': 1.4602511823177338} 01/29/2022 01:26:48 - INFO - codeparrot_training - Step 9798: {'lr': 0.00046813998999214193, 'samples': 1881408, 'steps': 9798, 'loss/train': 2.0181230306625366} 01/29/2022 01:26:54 - INFO - codeparrot_training - Step 9799: {'lr': 0.0004681319963473747, 'samples': 1881600, 'steps': 9799, 'loss/train': 1.666883111000061} 01/29/2022 01:26:58 - INFO - codeparrot_training - Step 9800: {'lr': 0.0004681240017681993, 'samples': 1881792, 'steps': 9800, 'loss/train': 0.6280254274606705} 01/29/2022 01:27:02 - INFO - codeparrot_training - Step 9801: {'lr': 0.0004681160062546499, 'samples': 1881984, 'steps': 9801, 'loss/train': 0.0670415461063385} 01/29/2022 01:27:06 - INFO - codeparrot_training - Step 9802: {'lr': 0.00046810800980676083, 'samples': 1882176, 'steps': 9802, 'loss/train': 1.7316153645515442} 01/29/2022 01:27:11 - INFO - codeparrot_training - Step 9803: {'lr': 0.0004681000124245663, 'samples': 1882368, 'steps': 9803, 'loss/train': 2.02603018283844} 01/29/2022 01:27:17 - INFO - codeparrot_training - Step 9804: {'lr': 0.0004680920141081005, 'samples': 1882560, 'steps': 9804, 'loss/train': 1.6483606696128845} 01/29/2022 01:27:21 - INFO - codeparrot_training - Step 9805: {'lr': 0.00046808401485739793, 'samples': 1882752, 'steps': 9805, 'loss/train': 1.754873514175415} 01/29/2022 01:27:25 - INFO - codeparrot_training - Step 9806: {'lr': 0.00046807601467249255, 'samples': 1882944, 'steps': 9806, 'loss/train': 0.19909123331308365} 01/29/2022 01:27:30 - INFO - codeparrot_training - Step 9807: {'lr': 0.0004680680135534188, 'samples': 1883136, 'steps': 9807, 'loss/train': 0.7179824709892273} 01/29/2022 01:27:35 - INFO - codeparrot_training - Step 9808: {'lr': 0.00046806001150021095, 'samples': 1883328, 'steps': 9808, 'loss/train': 1.4083564281463623} 01/29/2022 01:27:39 - INFO - codeparrot_training - Step 9809: {'lr': 0.0004680520085129032, 'samples': 1883520, 'steps': 9809, 'loss/train': 2.91452693939209} 01/29/2022 01:27:44 - INFO - codeparrot_training - Step 9810: {'lr': 0.00046804400459152994, 'samples': 1883712, 'steps': 9810, 'loss/train': 1.7284017205238342} 01/29/2022 01:27:48 - INFO - codeparrot_training - Step 9811: {'lr': 0.0004680359997361254, 'samples': 1883904, 'steps': 9811, 'loss/train': 1.3370522260665894} 01/29/2022 01:27:52 - INFO - codeparrot_training - Step 9812: {'lr': 0.0004680279939467238, 'samples': 1884096, 'steps': 9812, 'loss/train': 1.060422420501709} 01/29/2022 01:27:57 - INFO - codeparrot_training - Step 9813: {'lr': 0.0004680199872233596, 'samples': 1884288, 'steps': 9813, 'loss/train': 2.2027611136436462} 01/29/2022 01:28:02 - INFO - codeparrot_training - Step 9814: {'lr': 0.00046801197956606693, 'samples': 1884480, 'steps': 9814, 'loss/train': 1.9827784895896912} 01/29/2022 01:28:06 - INFO - codeparrot_training - Step 9815: {'lr': 0.00046800397097488024, 'samples': 1884672, 'steps': 9815, 'loss/train': 2.1319960355758667} 01/29/2022 01:28:10 - INFO - codeparrot_training - Step 9816: {'lr': 0.0004679959614498337, 'samples': 1884864, 'steps': 9816, 'loss/train': 1.3279857337474823} 01/29/2022 01:28:14 - INFO - codeparrot_training - Step 9817: {'lr': 0.0004679879509909617, 'samples': 1885056, 'steps': 9817, 'loss/train': 1.3977899551391602} 01/29/2022 01:28:20 - INFO - codeparrot_training - Step 9818: {'lr': 0.00046797993959829857, 'samples': 1885248, 'steps': 9818, 'loss/train': 1.7221103310585022} 01/29/2022 01:28:24 - INFO - codeparrot_training - Step 9819: {'lr': 0.00046797192727187855, 'samples': 1885440, 'steps': 9819, 'loss/train': 0.2971556559205055} 01/29/2022 01:28:28 - INFO - codeparrot_training - Step 9820: {'lr': 0.000467963914011736, 'samples': 1885632, 'steps': 9820, 'loss/train': 2.021244764328003} 01/29/2022 01:28:32 - INFO - codeparrot_training - Step 9821: {'lr': 0.0004679558998179053, 'samples': 1885824, 'steps': 9821, 'loss/train': 2.015450656414032} 01/29/2022 01:28:37 - INFO - codeparrot_training - Step 9822: {'lr': 0.0004679478846904207, 'samples': 1886016, 'steps': 9822, 'loss/train': 1.9460231065750122} 01/29/2022 01:28:42 - INFO - codeparrot_training - Step 9823: {'lr': 0.00046793986862931654, 'samples': 1886208, 'steps': 9823, 'loss/train': 1.4506638050079346} 01/29/2022 01:28:47 - INFO - codeparrot_training - Step 9824: {'lr': 0.0004679318516346273, 'samples': 1886400, 'steps': 9824, 'loss/train': 1.5622769594192505} 01/29/2022 01:28:51 - INFO - codeparrot_training - Step 9825: {'lr': 0.00046792383370638705, 'samples': 1886592, 'steps': 9825, 'loss/train': 2.1696184873580933} 01/29/2022 01:28:55 - INFO - codeparrot_training - Step 9826: {'lr': 0.0004679158148446304, 'samples': 1886784, 'steps': 9826, 'loss/train': 2.3101311922073364} 01/29/2022 01:29:00 - INFO - codeparrot_training - Step 9827: {'lr': 0.00046790779504939155, 'samples': 1886976, 'steps': 9827, 'loss/train': 0.8237840831279755} 01/29/2022 01:29:05 - INFO - codeparrot_training - Step 9828: {'lr': 0.00046789977432070497, 'samples': 1887168, 'steps': 9828, 'loss/train': 1.7133252024650574} 01/29/2022 01:29:09 - INFO - codeparrot_training - Step 9829: {'lr': 0.00046789175265860483, 'samples': 1887360, 'steps': 9829, 'loss/train': 0.39605145156383514} 01/29/2022 01:29:13 - INFO - codeparrot_training - Step 9830: {'lr': 0.00046788373006312567, 'samples': 1887552, 'steps': 9830, 'loss/train': 1.7340640425682068} 01/29/2022 01:29:17 - INFO - codeparrot_training - Step 9831: {'lr': 0.0004678757065343019, 'samples': 1887744, 'steps': 9831, 'loss/train': 1.8939526677131653} 01/29/2022 01:29:22 - INFO - codeparrot_training - Step 9832: {'lr': 0.0004678676820721677, 'samples': 1887936, 'steps': 9832, 'loss/train': 1.240151971578598} 01/29/2022 01:29:28 - INFO - codeparrot_training - Step 9833: {'lr': 0.00046785965667675745, 'samples': 1888128, 'steps': 9833, 'loss/train': 2.15874320268631} 01/29/2022 01:29:32 - INFO - codeparrot_training - Step 9834: {'lr': 0.00046785163034810567, 'samples': 1888320, 'steps': 9834, 'loss/train': 1.0703206658363342} 01/29/2022 01:29:36 - INFO - codeparrot_training - Step 9835: {'lr': 0.00046784360308624675, 'samples': 1888512, 'steps': 9835, 'loss/train': 1.067740648984909} 01/29/2022 01:29:40 - INFO - codeparrot_training - Step 9836: {'lr': 0.0004678355748912149, 'samples': 1888704, 'steps': 9836, 'loss/train': 1.3081364929676056} 01/29/2022 01:29:45 - INFO - codeparrot_training - Step 9837: {'lr': 0.0004678275457630447, 'samples': 1888896, 'steps': 9837, 'loss/train': 1.3843064904212952} 01/29/2022 01:29:50 - INFO - codeparrot_training - Step 9838: {'lr': 0.0004678195157017704, 'samples': 1889088, 'steps': 9838, 'loss/train': 1.1687838435173035} 01/29/2022 01:29:54 - INFO - codeparrot_training - Step 9839: {'lr': 0.00046781148470742654, 'samples': 1889280, 'steps': 9839, 'loss/train': 1.6022427678108215} 01/29/2022 01:29:59 - INFO - codeparrot_training - Step 9840: {'lr': 0.0004678034527800474, 'samples': 1889472, 'steps': 9840, 'loss/train': 2.3600528240203857} 01/29/2022 01:30:03 - INFO - codeparrot_training - Step 9841: {'lr': 0.0004677954199196674, 'samples': 1889664, 'steps': 9841, 'loss/train': 0.5905073136091232} 01/29/2022 01:30:07 - INFO - codeparrot_training - Step 9842: {'lr': 0.00046778738612632097, 'samples': 1889856, 'steps': 9842, 'loss/train': 1.8292835354804993} 01/29/2022 01:30:12 - INFO - codeparrot_training - Step 9843: {'lr': 0.00046777935140004256, 'samples': 1890048, 'steps': 9843, 'loss/train': 2.0544952154159546} 01/29/2022 01:30:17 - INFO - codeparrot_training - Step 9844: {'lr': 0.00046777131574086663, 'samples': 1890240, 'steps': 9844, 'loss/train': 2.056985378265381} 01/29/2022 01:30:21 - INFO - codeparrot_training - Step 9845: {'lr': 0.0004677632791488274, 'samples': 1890432, 'steps': 9845, 'loss/train': 0.795783519744873} 01/29/2022 01:30:25 - INFO - codeparrot_training - Step 9846: {'lr': 0.00046775524162395954, 'samples': 1890624, 'steps': 9846, 'loss/train': 1.3512190282344818} 01/29/2022 01:30:29 - INFO - codeparrot_training - Step 9847: {'lr': 0.00046774720316629734, 'samples': 1890816, 'steps': 9847, 'loss/train': 2.633673369884491} 01/29/2022 01:30:35 - INFO - codeparrot_training - Step 9848: {'lr': 0.00046773916377587524, 'samples': 1891008, 'steps': 9848, 'loss/train': 1.328588604927063} 01/29/2022 01:30:40 - INFO - codeparrot_training - Step 9849: {'lr': 0.00046773112345272773, 'samples': 1891200, 'steps': 9849, 'loss/train': 1.2961955666542053} 01/29/2022 01:30:44 - INFO - codeparrot_training - Step 9850: {'lr': 0.0004677230821968892, 'samples': 1891392, 'steps': 9850, 'loss/train': 1.4931917488574982} 01/29/2022 01:30:48 - INFO - codeparrot_training - Step 9851: {'lr': 0.00046771504000839417, 'samples': 1891584, 'steps': 9851, 'loss/train': 2.4104247093200684} 01/29/2022 01:30:52 - INFO - codeparrot_training - Step 9852: {'lr': 0.0004677069968872769, 'samples': 1891776, 'steps': 9852, 'loss/train': 2.021940529346466} 01/29/2022 01:30:58 - INFO - codeparrot_training - Step 9853: {'lr': 0.0004676989528335721, 'samples': 1891968, 'steps': 9853, 'loss/train': 1.259043037891388} 01/29/2022 01:31:02 - INFO - codeparrot_training - Step 9854: {'lr': 0.0004676909078473142, 'samples': 1892160, 'steps': 9854, 'loss/train': 1.7933263778686523} 01/29/2022 01:31:06 - INFO - codeparrot_training - Step 9855: {'lr': 0.00046768286192853736, 'samples': 1892352, 'steps': 9855, 'loss/train': 2.0334836840629578} 01/29/2022 01:31:11 - INFO - codeparrot_training - Step 9856: {'lr': 0.00046767481507727646, 'samples': 1892544, 'steps': 9856, 'loss/train': 1.4960417747497559} 01/29/2022 01:31:17 - INFO - codeparrot_training - Step 9857: {'lr': 0.00046766676729356564, 'samples': 1892736, 'steps': 9857, 'loss/train': 1.954717218875885} 01/29/2022 01:31:21 - INFO - codeparrot_training - Step 9858: {'lr': 0.0004676587185774396, 'samples': 1892928, 'steps': 9858, 'loss/train': 1.8594483733177185} 01/29/2022 01:31:25 - INFO - codeparrot_training - Step 9859: {'lr': 0.00046765066892893266, 'samples': 1893120, 'steps': 9859, 'loss/train': 2.4203338623046875} 01/29/2022 01:31:29 - INFO - codeparrot_training - Step 9860: {'lr': 0.00046764261834807944, 'samples': 1893312, 'steps': 9860, 'loss/train': 2.1051512360572815} 01/29/2022 01:31:34 - INFO - codeparrot_training - Step 9861: {'lr': 0.0004676345668349142, 'samples': 1893504, 'steps': 9861, 'loss/train': 2.4483171701431274} 01/29/2022 01:31:39 - INFO - codeparrot_training - Step 9862: {'lr': 0.0004676265143894717, 'samples': 1893696, 'steps': 9862, 'loss/train': 0.6269695311784744} 01/29/2022 01:31:43 - INFO - codeparrot_training - Step 9863: {'lr': 0.0004676184610117863, 'samples': 1893888, 'steps': 9863, 'loss/train': 1.5357985496520996} 01/29/2022 01:31:47 - INFO - codeparrot_training - Step 9864: {'lr': 0.0004676104067018925, 'samples': 1894080, 'steps': 9864, 'loss/train': 1.7321455478668213} 01/29/2022 01:31:52 - INFO - codeparrot_training - Step 9865: {'lr': 0.0004676023514598249, 'samples': 1894272, 'steps': 9865, 'loss/train': 1.603420078754425} 01/29/2022 01:31:56 - INFO - codeparrot_training - Step 9866: {'lr': 0.0004675942952856178, 'samples': 1894464, 'steps': 9866, 'loss/train': 1.4249381124973297} 01/29/2022 01:32:01 - INFO - codeparrot_training - Step 9867: {'lr': 0.0004675862381793059, 'samples': 1894656, 'steps': 9867, 'loss/train': 2.4648284912109375} 01/29/2022 01:32:06 - INFO - codeparrot_training - Step 9868: {'lr': 0.0004675781801409236, 'samples': 1894848, 'steps': 9868, 'loss/train': 1.6102239489555359} 01/29/2022 01:32:10 - INFO - codeparrot_training - Step 9869: {'lr': 0.00046757012117050554, 'samples': 1895040, 'steps': 9869, 'loss/train': 1.402160793542862} 01/29/2022 01:32:14 - INFO - codeparrot_training - Step 9870: {'lr': 0.00046756206126808607, 'samples': 1895232, 'steps': 9870, 'loss/train': 2.0061439275741577} 01/29/2022 01:32:18 - INFO - codeparrot_training - Step 9871: {'lr': 0.0004675540004336999, 'samples': 1895424, 'steps': 9871, 'loss/train': 1.308531492948532} 01/29/2022 01:32:24 - INFO - codeparrot_training - Step 9872: {'lr': 0.00046754593866738144, 'samples': 1895616, 'steps': 9872, 'loss/train': 1.4023101925849915} 01/29/2022 01:32:28 - INFO - codeparrot_training - Step 9873: {'lr': 0.0004675378759691652, 'samples': 1895808, 'steps': 9873, 'loss/train': 2.114335834980011} 01/29/2022 01:32:32 - INFO - codeparrot_training - Step 9874: {'lr': 0.00046752981233908587, 'samples': 1896000, 'steps': 9874, 'loss/train': 1.714539349079132} 01/29/2022 01:32:36 - INFO - codeparrot_training - Step 9875: {'lr': 0.0004675217477771779, 'samples': 1896192, 'steps': 9875, 'loss/train': 1.4300369918346405} 01/29/2022 01:32:41 - INFO - codeparrot_training - Step 9876: {'lr': 0.0004675136822834758, 'samples': 1896384, 'steps': 9876, 'loss/train': 1.0258161127567291} 01/29/2022 01:32:47 - INFO - codeparrot_training - Step 9877: {'lr': 0.0004675056158580141, 'samples': 1896576, 'steps': 9877, 'loss/train': 1.6255831718444824} 01/29/2022 01:32:51 - INFO - codeparrot_training - Step 9878: {'lr': 0.0004674975485008275, 'samples': 1896768, 'steps': 9878, 'loss/train': 1.4851067662239075} 01/29/2022 01:32:55 - INFO - codeparrot_training - Step 9879: {'lr': 0.00046748948021195036, 'samples': 1896960, 'steps': 9879, 'loss/train': 1.418349713087082} 01/29/2022 01:32:59 - INFO - codeparrot_training - Step 9880: {'lr': 0.0004674814109914174, 'samples': 1897152, 'steps': 9880, 'loss/train': 1.8449535369873047} 01/29/2022 01:33:04 - INFO - codeparrot_training - Step 9881: {'lr': 0.00046747334083926316, 'samples': 1897344, 'steps': 9881, 'loss/train': 0.06420137733221054} 01/29/2022 01:33:09 - INFO - codeparrot_training - Step 9882: {'lr': 0.0004674652697555222, 'samples': 1897536, 'steps': 9882, 'loss/train': 2.004448413848877} 01/29/2022 01:33:13 - INFO - codeparrot_training - Step 9883: {'lr': 0.000467457197740229, 'samples': 1897728, 'steps': 9883, 'loss/train': 2.327472746372223} 01/29/2022 01:33:17 - INFO - codeparrot_training - Step 9884: {'lr': 0.00046744912479341826, 'samples': 1897920, 'steps': 9884, 'loss/train': 0.9653741419315338} 01/29/2022 01:33:22 - INFO - codeparrot_training - Step 9885: {'lr': 0.0004674410509151246, 'samples': 1898112, 'steps': 9885, 'loss/train': 1.9246931076049805} 01/29/2022 01:33:26 - INFO - codeparrot_training - Step 9886: {'lr': 0.0004674329761053824, 'samples': 1898304, 'steps': 9886, 'loss/train': 0.9119590222835541} 01/29/2022 01:33:31 - INFO - codeparrot_training - Step 9887: {'lr': 0.00046742490036422635, 'samples': 1898496, 'steps': 9887, 'loss/train': 1.7399539947509766} 01/29/2022 01:33:35 - INFO - codeparrot_training - Step 9888: {'lr': 0.00046741682369169115, 'samples': 1898688, 'steps': 9888, 'loss/train': 1.5037275552749634} 01/29/2022 01:33:40 - INFO - codeparrot_training - Step 9889: {'lr': 0.00046740874608781126, 'samples': 1898880, 'steps': 9889, 'loss/train': 1.9080183506011963} 01/29/2022 01:33:44 - INFO - codeparrot_training - Step 9890: {'lr': 0.0004674006675526214, 'samples': 1899072, 'steps': 9890, 'loss/train': 2.559327006340027} 01/29/2022 01:33:48 - INFO - codeparrot_training - Step 9891: {'lr': 0.00046739258808615607, 'samples': 1899264, 'steps': 9891, 'loss/train': 0.7903262972831726} 01/29/2022 01:33:54 - INFO - codeparrot_training - Step 9892: {'lr': 0.00046738450768845, 'samples': 1899456, 'steps': 9892, 'loss/train': 2.027642011642456} 01/29/2022 01:33:58 - INFO - codeparrot_training - Step 9893: {'lr': 0.0004673764263595376, 'samples': 1899648, 'steps': 9893, 'loss/train': 1.7447773218154907} 01/29/2022 01:34:02 - INFO - codeparrot_training - Step 9894: {'lr': 0.00046736834409945364, 'samples': 1899840, 'steps': 9894, 'loss/train': 1.0822137594223022} 01/29/2022 01:34:07 - INFO - codeparrot_training - Step 9895: {'lr': 0.0004673602609082328, 'samples': 1900032, 'steps': 9895, 'loss/train': 2.6166704893112183} 01/29/2022 01:34:11 - INFO - codeparrot_training - Step 9896: {'lr': 0.00046735217678590957, 'samples': 1900224, 'steps': 9896, 'loss/train': 2.1367433667182922} 01/29/2022 01:34:17 - INFO - codeparrot_training - Step 9897: {'lr': 0.0004673440917325186, 'samples': 1900416, 'steps': 9897, 'loss/train': 2.5662233233451843} 01/29/2022 01:34:21 - INFO - codeparrot_training - Step 9898: {'lr': 0.00046733600574809465, 'samples': 1900608, 'steps': 9898, 'loss/train': 1.338602900505066} 01/29/2022 01:34:25 - INFO - codeparrot_training - Step 9899: {'lr': 0.0004673279188326722, 'samples': 1900800, 'steps': 9899, 'loss/train': 1.7467287182807922} 01/29/2022 01:34:29 - INFO - codeparrot_training - Step 9900: {'lr': 0.00046731983098628597, 'samples': 1900992, 'steps': 9900, 'loss/train': 2.0499138236045837} 01/29/2022 01:34:34 - INFO - codeparrot_training - Step 9901: {'lr': 0.00046731174220897054, 'samples': 1901184, 'steps': 9901, 'loss/train': 1.4686086773872375} 01/29/2022 01:34:39 - INFO - codeparrot_training - Step 9902: {'lr': 0.0004673036525007607, 'samples': 1901376, 'steps': 9902, 'loss/train': 2.069655954837799} 01/29/2022 01:34:43 - INFO - codeparrot_training - Step 9903: {'lr': 0.0004672955618616909, 'samples': 1901568, 'steps': 9903, 'loss/train': 1.52781343460083} 01/29/2022 01:34:48 - INFO - codeparrot_training - Step 9904: {'lr': 0.00046728747029179594, 'samples': 1901760, 'steps': 9904, 'loss/train': 2.721159338951111} 01/29/2022 01:34:52 - INFO - codeparrot_training - Step 9905: {'lr': 0.00046727937779111054, 'samples': 1901952, 'steps': 9905, 'loss/train': 1.5916142463684082} 01/29/2022 01:34:56 - INFO - codeparrot_training - Step 9906: {'lr': 0.0004672712843596693, 'samples': 1902144, 'steps': 9906, 'loss/train': 2.180336058139801} 01/29/2022 01:35:02 - INFO - codeparrot_training - Step 9907: {'lr': 0.0004672631899975067, 'samples': 1902336, 'steps': 9907, 'loss/train': 1.6991794109344482} 01/29/2022 01:35:06 - INFO - codeparrot_training - Step 9908: {'lr': 0.0004672550947046577, 'samples': 1902528, 'steps': 9908, 'loss/train': 2.2215505242347717} 01/29/2022 01:35:11 - INFO - codeparrot_training - Step 9909: {'lr': 0.0004672469984811568, 'samples': 1902720, 'steps': 9909, 'loss/train': 1.8558146953582764} 01/29/2022 01:35:15 - INFO - codeparrot_training - Step 9910: {'lr': 0.00046723890132703886, 'samples': 1902912, 'steps': 9910, 'loss/train': 1.2212025225162506} 01/29/2022 01:35:19 - INFO - codeparrot_training - Step 9911: {'lr': 0.0004672308032423384, 'samples': 1903104, 'steps': 9911, 'loss/train': 1.0914042592048645} 01/29/2022 01:35:24 - INFO - codeparrot_training - Step 9912: {'lr': 0.0004672227042270901, 'samples': 1903296, 'steps': 9912, 'loss/train': 1.2423049807548523} 01/29/2022 01:35:29 - INFO - codeparrot_training - Step 9913: {'lr': 0.00046721460428132873, 'samples': 1903488, 'steps': 9913, 'loss/train': 1.857115924358368} 01/29/2022 01:35:33 - INFO - codeparrot_training - Step 9914: {'lr': 0.00046720650340508895, 'samples': 1903680, 'steps': 9914, 'loss/train': 2.4169304966926575} 01/29/2022 01:35:37 - INFO - codeparrot_training - Step 9915: {'lr': 0.00046719840159840557, 'samples': 1903872, 'steps': 9915, 'loss/train': 1.4196109771728516} 01/29/2022 01:35:41 - INFO - codeparrot_training - Step 9916: {'lr': 0.00046719029886131317, 'samples': 1904064, 'steps': 9916, 'loss/train': 1.801427185535431} 01/29/2022 01:35:47 - INFO - codeparrot_training - Step 9917: {'lr': 0.0004671821951938464, 'samples': 1904256, 'steps': 9917, 'loss/train': 0.8891903758049011} 01/29/2022 01:35:52 - INFO - codeparrot_training - Step 9918: {'lr': 0.0004671740905960401, 'samples': 1904448, 'steps': 9918, 'loss/train': 2.8290106058120728} 01/29/2022 01:35:56 - INFO - codeparrot_training - Step 9919: {'lr': 0.00046716598506792905, 'samples': 1904640, 'steps': 9919, 'loss/train': 2.278944432735443} 01/29/2022 01:36:00 - INFO - codeparrot_training - Step 9920: {'lr': 0.00046715787860954785, 'samples': 1904832, 'steps': 9920, 'loss/train': 1.6641658544540405} 01/29/2022 01:36:04 - INFO - codeparrot_training - Step 9921: {'lr': 0.0004671497712209312, 'samples': 1905024, 'steps': 9921, 'loss/train': 1.1046734154224396} 01/29/2022 01:36:10 - INFO - codeparrot_training - Step 9922: {'lr': 0.0004671416629021139, 'samples': 1905216, 'steps': 9922, 'loss/train': 1.6742010712623596} 01/29/2022 01:36:14 - INFO - codeparrot_training - Step 9923: {'lr': 0.0004671335536531307, 'samples': 1905408, 'steps': 9923, 'loss/train': 1.506627380847931} 01/29/2022 01:36:18 - INFO - codeparrot_training - Step 9924: {'lr': 0.00046712544347401623, 'samples': 1905600, 'steps': 9924, 'loss/train': 1.4152531921863556} 01/29/2022 01:36:22 - INFO - codeparrot_training - Step 9925: {'lr': 0.0004671173323648054, 'samples': 1905792, 'steps': 9925, 'loss/train': 1.1242337822914124} 01/29/2022 01:36:27 - INFO - codeparrot_training - Step 9926: {'lr': 0.00046710922032553283, 'samples': 1905984, 'steps': 9926, 'loss/train': 1.6694687604904175} 01/29/2022 01:36:32 - INFO - codeparrot_training - Step 9927: {'lr': 0.00046710110735623326, 'samples': 1906176, 'steps': 9927, 'loss/train': 1.4261307120323181} 01/29/2022 01:36:36 - INFO - codeparrot_training - Step 9928: {'lr': 0.00046709299345694156, 'samples': 1906368, 'steps': 9928, 'loss/train': 2.2305259108543396} 01/29/2022 01:36:40 - INFO - codeparrot_training - Step 9929: {'lr': 0.00046708487862769235, 'samples': 1906560, 'steps': 9929, 'loss/train': 2.2148884534835815} 01/29/2022 01:36:45 - INFO - codeparrot_training - Step 9930: {'lr': 0.0004670767628685204, 'samples': 1906752, 'steps': 9930, 'loss/train': 1.709048867225647} 01/29/2022 01:36:49 - INFO - codeparrot_training - Step 9931: {'lr': 0.00046706864617946064, 'samples': 1906944, 'steps': 9931, 'loss/train': 1.5844584703445435} 01/29/2022 01:36:54 - INFO - codeparrot_training - Step 9932: {'lr': 0.0004670605285605477, 'samples': 1907136, 'steps': 9932, 'loss/train': 2.0456458926200867} 01/29/2022 01:36:58 - INFO - codeparrot_training - Step 9933: {'lr': 0.0004670524100118163, 'samples': 1907328, 'steps': 9933, 'loss/train': 1.5994293093681335} 01/29/2022 01:37:03 - INFO - codeparrot_training - Step 9934: {'lr': 0.00046704429053330137, 'samples': 1907520, 'steps': 9934, 'loss/train': 1.2421969771385193} 01/29/2022 01:37:07 - INFO - codeparrot_training - Step 9935: {'lr': 0.00046703617012503764, 'samples': 1907712, 'steps': 9935, 'loss/train': 2.040019154548645} 01/29/2022 01:37:11 - INFO - codeparrot_training - Step 9936: {'lr': 0.00046702804878705987, 'samples': 1907904, 'steps': 9936, 'loss/train': 2.248599350452423} 01/29/2022 01:37:17 - INFO - codeparrot_training - Step 9937: {'lr': 0.00046701992651940275, 'samples': 1908096, 'steps': 9937, 'loss/train': 2.1610280871391296} 01/29/2022 01:37:22 - INFO - codeparrot_training - Step 9938: {'lr': 0.00046701180332210125, 'samples': 1908288, 'steps': 9938, 'loss/train': 1.5467262268066406} 01/29/2022 01:37:26 - INFO - codeparrot_training - Step 9939: {'lr': 0.0004670036791951901, 'samples': 1908480, 'steps': 9939, 'loss/train': 1.7213489413261414} 01/29/2022 01:37:30 - INFO - codeparrot_training - Step 9940: {'lr': 0.0004669955541387041, 'samples': 1908672, 'steps': 9940, 'loss/train': 1.9297624826431274} 01/29/2022 01:37:35 - INFO - codeparrot_training - Step 9941: {'lr': 0.000466987428152678, 'samples': 1908864, 'steps': 9941, 'loss/train': 2.548409879207611} 01/29/2022 01:37:40 - INFO - codeparrot_training - Step 9942: {'lr': 0.00046697930123714673, 'samples': 1909056, 'steps': 9942, 'loss/train': 1.6108113527297974} 01/29/2022 01:37:44 - INFO - codeparrot_training - Step 9943: {'lr': 0.000466971173392145, 'samples': 1909248, 'steps': 9943, 'loss/train': 1.5467010140419006} 01/29/2022 01:37:48 - INFO - codeparrot_training - Step 9944: {'lr': 0.0004669630446177077, 'samples': 1909440, 'steps': 9944, 'loss/train': 1.737030565738678} 01/29/2022 01:37:52 - INFO - codeparrot_training - Step 9945: {'lr': 0.00046695491491386955, 'samples': 1909632, 'steps': 9945, 'loss/train': 1.6655312776565552} 01/29/2022 01:37:58 - INFO - codeparrot_training - Step 9946: {'lr': 0.0004669467842806654, 'samples': 1909824, 'steps': 9946, 'loss/train': 1.6046923398971558} 01/29/2022 01:38:02 - INFO - codeparrot_training - Step 9947: {'lr': 0.00046693865271813016, 'samples': 1910016, 'steps': 9947, 'loss/train': 1.789564311504364} 01/29/2022 01:38:06 - INFO - codeparrot_training - Step 9948: {'lr': 0.0004669305202262987, 'samples': 1910208, 'steps': 9948, 'loss/train': 1.543137788772583} 01/29/2022 01:38:11 - INFO - codeparrot_training - Step 9949: {'lr': 0.00046692238680520564, 'samples': 1910400, 'steps': 9949, 'loss/train': 1.6994887590408325} 01/29/2022 01:38:15 - INFO - codeparrot_training - Step 9950: {'lr': 0.00046691425245488607, 'samples': 1910592, 'steps': 9950, 'loss/train': 1.2044717967510223} 01/29/2022 01:38:20 - INFO - codeparrot_training - Step 9951: {'lr': 0.0004669061171753746, 'samples': 1910784, 'steps': 9951, 'loss/train': 1.8275125622749329} 01/29/2022 01:38:25 - INFO - codeparrot_training - Step 9952: {'lr': 0.0004668979809667063, 'samples': 1910976, 'steps': 9952, 'loss/train': 1.1356662511825562} 01/29/2022 01:38:29 - INFO - codeparrot_training - Step 9953: {'lr': 0.0004668898438289159, 'samples': 1911168, 'steps': 9953, 'loss/train': 0.8673244714736938} 01/29/2022 01:38:33 - INFO - codeparrot_training - Step 9954: {'lr': 0.00046688170576203827, 'samples': 1911360, 'steps': 9954, 'loss/train': 1.5182093381881714} 01/29/2022 01:38:37 - INFO - codeparrot_training - Step 9955: {'lr': 0.00046687356676610825, 'samples': 1911552, 'steps': 9955, 'loss/train': 1.4415996372699738} 01/29/2022 01:38:43 - INFO - codeparrot_training - Step 9956: {'lr': 0.00046686542684116073, 'samples': 1911744, 'steps': 9956, 'loss/train': 2.040181338787079} 01/29/2022 01:38:47 - INFO - codeparrot_training - Step 9957: {'lr': 0.00046685728598723063, 'samples': 1911936, 'steps': 9957, 'loss/train': 2.499409854412079} 01/29/2022 01:38:51 - INFO - codeparrot_training - Step 9958: {'lr': 0.00046684914420435275, 'samples': 1912128, 'steps': 9958, 'loss/train': 2.405643939971924} 01/29/2022 01:38:56 - INFO - codeparrot_training - Step 9959: {'lr': 0.00046684100149256205, 'samples': 1912320, 'steps': 9959, 'loss/train': 1.831012487411499} 01/29/2022 01:39:00 - INFO - codeparrot_training - Step 9960: {'lr': 0.0004668328578518933, 'samples': 1912512, 'steps': 9960, 'loss/train': 1.7565289735794067} 01/29/2022 01:39:06 - INFO - codeparrot_training - Step 9961: {'lr': 0.0004668247132823814, 'samples': 1912704, 'steps': 9961, 'loss/train': 1.180438756942749} 01/29/2022 01:39:11 - INFO - codeparrot_training - Step 9962: {'lr': 0.00046681656778406136, 'samples': 1912896, 'steps': 9962, 'loss/train': 1.433887928724289} 01/29/2022 01:39:15 - INFO - codeparrot_training - Step 9963: {'lr': 0.000466808421356968, 'samples': 1913088, 'steps': 9963, 'loss/train': 1.4644307792186737} 01/29/2022 01:39:19 - INFO - codeparrot_training - Step 9964: {'lr': 0.00046680027400113614, 'samples': 1913280, 'steps': 9964, 'loss/train': 1.5737085342407227} 01/29/2022 01:39:23 - INFO - codeparrot_training - Step 9965: {'lr': 0.0004667921257166008, 'samples': 1913472, 'steps': 9965, 'loss/train': 2.152284264564514} 01/29/2022 01:39:29 - INFO - codeparrot_training - Step 9966: {'lr': 0.00046678397650339677, 'samples': 1913664, 'steps': 9966, 'loss/train': 1.1488637030124664} 01/29/2022 01:39:33 - INFO - codeparrot_training - Step 9967: {'lr': 0.00046677582636155904, 'samples': 1913856, 'steps': 9967, 'loss/train': 1.2733469009399414} 01/29/2022 01:39:37 - INFO - codeparrot_training - Step 9968: {'lr': 0.00046676767529112254, 'samples': 1914048, 'steps': 9968, 'loss/train': 1.36495703458786} 01/29/2022 01:39:41 - INFO - codeparrot_training - Step 9969: {'lr': 0.0004667595232921221, 'samples': 1914240, 'steps': 9969, 'loss/train': 1.1071585714817047} 01/29/2022 01:39:46 - INFO - codeparrot_training - Step 9970: {'lr': 0.00046675137036459273, 'samples': 1914432, 'steps': 9970, 'loss/train': 1.6680185794830322} 01/29/2022 01:39:51 - INFO - codeparrot_training - Step 9971: {'lr': 0.0004667432165085693, 'samples': 1914624, 'steps': 9971, 'loss/train': 1.953692078590393} 01/29/2022 01:39:55 - INFO - codeparrot_training - Step 9972: {'lr': 0.00046673506172408675, 'samples': 1914816, 'steps': 9972, 'loss/train': 1.301068514585495} 01/29/2022 01:39:59 - INFO - codeparrot_training - Step 9973: {'lr': 0.0004667269060111801, 'samples': 1915008, 'steps': 9973, 'loss/train': 1.5521276593208313} 01/29/2022 01:40:04 - INFO - codeparrot_training - Step 9974: {'lr': 0.0004667187493698841, 'samples': 1915200, 'steps': 9974, 'loss/train': 2.187178909778595} 01/29/2022 01:40:08 - INFO - codeparrot_training - Step 9975: {'lr': 0.00046671059180023377, 'samples': 1915392, 'steps': 9975, 'loss/train': 1.950165867805481} 01/29/2022 01:40:15 - INFO - codeparrot_training - Step 9976: {'lr': 0.0004667024333022642, 'samples': 1915584, 'steps': 9976, 'loss/train': 1.3632895946502686} 01/29/2022 01:40:19 - INFO - codeparrot_training - Step 9977: {'lr': 0.00046669427387601017, 'samples': 1915776, 'steps': 9977, 'loss/train': 2.175067663192749} 01/29/2022 01:40:24 - INFO - codeparrot_training - Step 9978: {'lr': 0.0004666861135215066, 'samples': 1915968, 'steps': 9978, 'loss/train': 1.4687270522117615} 01/29/2022 01:40:28 - INFO - codeparrot_training - Step 9979: {'lr': 0.0004666779522387886, 'samples': 1916160, 'steps': 9979, 'loss/train': 1.8690574765205383} 01/29/2022 01:40:32 - INFO - codeparrot_training - Step 9980: {'lr': 0.000466669790027891, 'samples': 1916352, 'steps': 9980, 'loss/train': 1.4040366411209106} 01/29/2022 01:40:37 - INFO - codeparrot_training - Step 9981: {'lr': 0.00046666162688884893, 'samples': 1916544, 'steps': 9981, 'loss/train': 1.243584305047989} 01/29/2022 01:40:42 - INFO - codeparrot_training - Step 9982: {'lr': 0.0004666534628216972, 'samples': 1916736, 'steps': 9982, 'loss/train': 1.448719561100006} 01/29/2022 01:40:46 - INFO - codeparrot_training - Step 9983: {'lr': 0.0004666452978264708, 'samples': 1916928, 'steps': 9983, 'loss/train': 1.6909217834472656} 01/29/2022 01:40:50 - INFO - codeparrot_training - Step 9984: {'lr': 0.0004666371319032047, 'samples': 1917120, 'steps': 9984, 'loss/train': 1.8013468980789185} 01/29/2022 01:40:56 - INFO - codeparrot_training - Step 9985: {'lr': 0.00046662896505193395, 'samples': 1917312, 'steps': 9985, 'loss/train': 1.8530933260917664} 01/29/2022 01:41:01 - INFO - codeparrot_training - Step 9986: {'lr': 0.00046662079727269356, 'samples': 1917504, 'steps': 9986, 'loss/train': 1.88957279920578} 01/29/2022 01:41:05 - INFO - codeparrot_training - Step 9987: {'lr': 0.0004666126285655184, 'samples': 1917696, 'steps': 9987, 'loss/train': 2.344163417816162} 01/29/2022 01:41:09 - INFO - codeparrot_training - Step 9988: {'lr': 0.0004666044589304436, 'samples': 1917888, 'steps': 9988, 'loss/train': 0.43918928503990173} 01/29/2022 01:41:13 - INFO - codeparrot_training - Step 9989: {'lr': 0.000466596288367504, 'samples': 1918080, 'steps': 9989, 'loss/train': 2.3905003666877747} 01/29/2022 01:41:19 - INFO - codeparrot_training - Step 9990: {'lr': 0.0004665881168767346, 'samples': 1918272, 'steps': 9990, 'loss/train': 1.5398701429367065} 01/29/2022 01:41:23 - INFO - codeparrot_training - Step 9991: {'lr': 0.00046657994445817064, 'samples': 1918464, 'steps': 9991, 'loss/train': 0.9140636622905731} 01/29/2022 01:41:27 - INFO - codeparrot_training - Step 9992: {'lr': 0.0004665717711118469, 'samples': 1918656, 'steps': 9992, 'loss/train': 1.4867514073848724} 01/29/2022 01:41:31 - INFO - codeparrot_training - Step 9993: {'lr': 0.00046656359683779845, 'samples': 1918848, 'steps': 9993, 'loss/train': 1.6643505692481995} 01/29/2022 01:41:36 - INFO - codeparrot_training - Step 9994: {'lr': 0.00046655542163606033, 'samples': 1919040, 'steps': 9994, 'loss/train': 0.8498890399932861} 01/29/2022 01:41:41 - INFO - codeparrot_training - Step 9995: {'lr': 0.0004665472455066675, 'samples': 1919232, 'steps': 9995, 'loss/train': 0.8319334387779236} 01/29/2022 01:41:45 - INFO - codeparrot_training - Step 9996: {'lr': 0.0004665390684496551, 'samples': 1919424, 'steps': 9996, 'loss/train': 2.45449036359787} 01/29/2022 01:41:49 - INFO - codeparrot_training - Step 9997: {'lr': 0.0004665308904650581, 'samples': 1919616, 'steps': 9997, 'loss/train': 1.726088047027588} 01/29/2022 01:41:54 - INFO - codeparrot_training - Step 9998: {'lr': 0.00046652271155291146, 'samples': 1919808, 'steps': 9998, 'loss/train': 1.6048874258995056} 01/29/2022 01:41:58 - INFO - codeparrot_training - Step 9999: {'lr': 0.0004665145317132503, 'samples': 1920000, 'steps': 9999, 'loss/train': 1.2912257015705109} 01/29/2022 01:41:58 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 01:42:31 - WARNING - huggingface_hub.repository - Several commits (5) will be pushed upstream. 01/29/2022 01:42:31 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 01:43:35 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py e19ad09..5069a3d expert-dust-2 -> expert-dust-2 01/29/2022 01:43:44 - INFO - codeparrot_training - Step 10000: {'lr': 0.00046650635094610973, 'samples': 1920192, 'steps': 10000, 'loss/train': 0.9473061561584473} 01/29/2022 01:43:48 - INFO - codeparrot_training - Step 10001: {'lr': 0.00046649816925152456, 'samples': 1920384, 'steps': 10001, 'loss/train': 1.238161325454712} 01/29/2022 01:43:53 - INFO - codeparrot_training - Step 10002: {'lr': 0.00046648998662953003, 'samples': 1920576, 'steps': 10002, 'loss/train': 2.3546106219291687} 01/29/2022 01:43:57 - INFO - codeparrot_training - Step 10003: {'lr': 0.00046648180308016116, 'samples': 1920768, 'steps': 10003, 'loss/train': 1.5486620664596558} 01/29/2022 01:44:01 - INFO - codeparrot_training - Step 10004: {'lr': 0.00046647361860345293, 'samples': 1920960, 'steps': 10004, 'loss/train': 1.8292316794395447} 01/29/2022 01:44:06 - INFO - codeparrot_training - Step 10005: {'lr': 0.00046646543319944057, 'samples': 1921152, 'steps': 10005, 'loss/train': 1.8547121286392212} 01/29/2022 01:44:11 - INFO - codeparrot_training - Step 10006: {'lr': 0.00046645724686815893, 'samples': 1921344, 'steps': 10006, 'loss/train': 1.7146250009536743} 01/29/2022 01:44:15 - INFO - codeparrot_training - Step 10007: {'lr': 0.00046644905960964325, 'samples': 1921536, 'steps': 10007, 'loss/train': 1.7744790315628052} 01/29/2022 01:44:19 - INFO - codeparrot_training - Step 10008: {'lr': 0.00046644087142392845, 'samples': 1921728, 'steps': 10008, 'loss/train': 1.676496684551239} 01/29/2022 01:44:23 - INFO - codeparrot_training - Step 10009: {'lr': 0.00046643268231104975, 'samples': 1921920, 'steps': 10009, 'loss/train': 1.2877540290355682} 01/29/2022 01:44:29 - INFO - codeparrot_training - Step 10010: {'lr': 0.00046642449227104213, 'samples': 1922112, 'steps': 10010, 'loss/train': 1.7743279337882996} 01/29/2022 01:44:33 - INFO - codeparrot_training - Step 10011: {'lr': 0.00046641630130394066, 'samples': 1922304, 'steps': 10011, 'loss/train': 1.5471746921539307} 01/29/2022 01:44:38 - INFO - codeparrot_training - Step 10012: {'lr': 0.0004664081094097805, 'samples': 1922496, 'steps': 10012, 'loss/train': 1.6109195351600647} 01/29/2022 01:44:42 - INFO - codeparrot_training - Step 10013: {'lr': 0.00046639991658859684, 'samples': 1922688, 'steps': 10013, 'loss/train': 2.009978234767914} 01/29/2022 01:44:46 - INFO - codeparrot_training - Step 10014: {'lr': 0.00046639172284042453, 'samples': 1922880, 'steps': 10014, 'loss/train': 1.246885746717453} 01/29/2022 01:44:51 - INFO - codeparrot_training - Step 10015: {'lr': 0.00046638352816529883, 'samples': 1923072, 'steps': 10015, 'loss/train': 1.3546336591243744} 01/29/2022 01:44:55 - INFO - codeparrot_training - Step 10016: {'lr': 0.00046637533256325476, 'samples': 1923264, 'steps': 10016, 'loss/train': 0.9458078742027283} 01/29/2022 01:45:00 - INFO - codeparrot_training - Step 10017: {'lr': 0.0004663671360343275, 'samples': 1923456, 'steps': 10017, 'loss/train': 1.866951048374176} 01/29/2022 01:45:04 - INFO - codeparrot_training - Step 10018: {'lr': 0.00046635893857855217, 'samples': 1923648, 'steps': 10018, 'loss/train': 1.9436956644058228} 01/29/2022 01:45:08 - INFO - codeparrot_training - Step 10019: {'lr': 0.0004663507401959638, 'samples': 1923840, 'steps': 10019, 'loss/train': 1.6378759145736694} 01/29/2022 01:45:15 - INFO - codeparrot_training - Step 10020: {'lr': 0.00046634254088659757, 'samples': 1924032, 'steps': 10020, 'loss/train': 2.1351377964019775} 01/29/2022 01:45:19 - INFO - codeparrot_training - Step 10021: {'lr': 0.00046633434065048855, 'samples': 1924224, 'steps': 10021, 'loss/train': 2.0643988251686096} 01/29/2022 01:45:23 - INFO - codeparrot_training - Step 10022: {'lr': 0.000466326139487672, 'samples': 1924416, 'steps': 10022, 'loss/train': 1.8094140887260437} 01/29/2022 01:45:27 - INFO - codeparrot_training - Step 10023: {'lr': 0.0004663179373981829, 'samples': 1924608, 'steps': 10023, 'loss/train': 1.439870685338974} 01/29/2022 01:45:32 - INFO - codeparrot_training - Step 10024: {'lr': 0.0004663097343820565, 'samples': 1924800, 'steps': 10024, 'loss/train': 2.0132378339767456} 01/29/2022 01:45:37 - INFO - codeparrot_training - Step 10025: {'lr': 0.00046630153043932784, 'samples': 1924992, 'steps': 10025, 'loss/train': 1.7374393343925476} 01/29/2022 01:45:41 - INFO - codeparrot_training - Step 10026: {'lr': 0.00046629332557003215, 'samples': 1925184, 'steps': 10026, 'loss/train': 2.0966098308563232} 01/29/2022 01:45:45 - INFO - codeparrot_training - Step 10027: {'lr': 0.00046628511977420443, 'samples': 1925376, 'steps': 10027, 'loss/train': 1.4120314121246338} 01/29/2022 01:45:50 - INFO - codeparrot_training - Step 10028: {'lr': 0.00046627691305188004, 'samples': 1925568, 'steps': 10028, 'loss/train': 2.308782935142517} 01/29/2022 01:45:54 - INFO - codeparrot_training - Step 10029: {'lr': 0.00046626870540309394, 'samples': 1925760, 'steps': 10029, 'loss/train': 2.562473773956299} 01/29/2022 01:46:01 - INFO - codeparrot_training - Step 10030: {'lr': 0.00046626049682788143, 'samples': 1925952, 'steps': 10030, 'loss/train': 2.788360297679901} 01/29/2022 01:46:05 - INFO - codeparrot_training - Step 10031: {'lr': 0.00046625228732627763, 'samples': 1926144, 'steps': 10031, 'loss/train': 2.7566927075386047} 01/29/2022 01:46:09 - INFO - codeparrot_training - Step 10032: {'lr': 0.00046624407689831773, 'samples': 1926336, 'steps': 10032, 'loss/train': 0.5930688679218292} 01/29/2022 01:46:13 - INFO - codeparrot_training - Step 10033: {'lr': 0.0004662358655440368, 'samples': 1926528, 'steps': 10033, 'loss/train': 1.848926067352295} 01/29/2022 01:46:18 - INFO - codeparrot_training - Step 10034: {'lr': 0.0004662276532634701, 'samples': 1926720, 'steps': 10034, 'loss/train': 1.3711004555225372} 01/29/2022 01:46:23 - INFO - codeparrot_training - Step 10035: {'lr': 0.0004662194400566528, 'samples': 1926912, 'steps': 10035, 'loss/train': 0.4746723175048828} 01/29/2022 01:46:27 - INFO - codeparrot_training - Step 10036: {'lr': 0.0004662112259236201, 'samples': 1927104, 'steps': 10036, 'loss/train': 1.0239878296852112} 01/29/2022 01:46:31 - INFO - codeparrot_training - Step 10037: {'lr': 0.00046620301086440713, 'samples': 1927296, 'steps': 10037, 'loss/train': 1.6438826322555542} 01/29/2022 01:46:36 - INFO - codeparrot_training - Step 10038: {'lr': 0.00046619479487904915, 'samples': 1927488, 'steps': 10038, 'loss/train': 1.592477023601532} 01/29/2022 01:46:40 - INFO - codeparrot_training - Step 10039: {'lr': 0.0004661865779675813, 'samples': 1927680, 'steps': 10039, 'loss/train': 1.6423092484474182} 01/29/2022 01:46:45 - INFO - codeparrot_training - Step 10040: {'lr': 0.0004661783601300388, 'samples': 1927872, 'steps': 10040, 'loss/train': 1.0886013507843018} 01/29/2022 01:46:49 - INFO - codeparrot_training - Step 10041: {'lr': 0.00046617014136645686, 'samples': 1928064, 'steps': 10041, 'loss/train': 2.451741635799408} 01/29/2022 01:46:54 - INFO - codeparrot_training - Step 10042: {'lr': 0.00046616192167687066, 'samples': 1928256, 'steps': 10042, 'loss/train': 0.17774800211191177} 01/29/2022 01:46:58 - INFO - codeparrot_training - Step 10043: {'lr': 0.00046615370106131536, 'samples': 1928448, 'steps': 10043, 'loss/train': 1.715687334537506} 01/29/2022 01:47:02 - INFO - codeparrot_training - Step 10044: {'lr': 0.00046614547951982636, 'samples': 1928640, 'steps': 10044, 'loss/train': 2.377831220626831} 01/29/2022 01:47:09 - INFO - codeparrot_training - Step 10045: {'lr': 0.00046613725705243873, 'samples': 1928832, 'steps': 10045, 'loss/train': 1.452502816915512} 01/29/2022 01:47:13 - INFO - codeparrot_training - Step 10046: {'lr': 0.0004661290336591877, 'samples': 1929024, 'steps': 10046, 'loss/train': 1.6298598647117615} 01/29/2022 01:47:18 - INFO - codeparrot_training - Step 10047: {'lr': 0.0004661208093401085, 'samples': 1929216, 'steps': 10047, 'loss/train': 1.7412450313568115} 01/29/2022 01:47:22 - INFO - codeparrot_training - Step 10048: {'lr': 0.0004661125840952364, 'samples': 1929408, 'steps': 10048, 'loss/train': 1.8512842655181885} 01/29/2022 01:47:26 - INFO - codeparrot_training - Step 10049: {'lr': 0.0004661043579246066, 'samples': 1929600, 'steps': 10049, 'loss/train': 1.5959844589233398} 01/29/2022 01:47:32 - INFO - codeparrot_training - Step 10050: {'lr': 0.00046609613082825436, 'samples': 1929792, 'steps': 10050, 'loss/train': 2.149889051914215} 01/29/2022 01:47:36 - INFO - codeparrot_training - Step 10051: {'lr': 0.00046608790280621494, 'samples': 1929984, 'steps': 10051, 'loss/train': 1.864513635635376} 01/29/2022 01:47:40 - INFO - codeparrot_training - Step 10052: {'lr': 0.0004660796738585235, 'samples': 1930176, 'steps': 10052, 'loss/train': 1.4654209613800049} 01/29/2022 01:47:44 - INFO - codeparrot_training - Step 10053: {'lr': 0.0004660714439852154, 'samples': 1930368, 'steps': 10053, 'loss/train': 2.2842568159103394} 01/29/2022 01:47:49 - INFO - codeparrot_training - Step 10054: {'lr': 0.0004660632131863258, 'samples': 1930560, 'steps': 10054, 'loss/train': 0.306283637881279} 01/29/2022 01:47:54 - INFO - codeparrot_training - Step 10055: {'lr': 0.0004660549814618901, 'samples': 1930752, 'steps': 10055, 'loss/train': 1.634218990802765} 01/29/2022 01:47:59 - INFO - codeparrot_training - Step 10056: {'lr': 0.00046604674881194335, 'samples': 1930944, 'steps': 10056, 'loss/train': 1.5157425999641418} 01/29/2022 01:48:03 - INFO - codeparrot_training - Step 10057: {'lr': 0.000466038515236521, 'samples': 1931136, 'steps': 10057, 'loss/train': 2.1594913601875305} 01/29/2022 01:48:07 - INFO - codeparrot_training - Step 10058: {'lr': 0.0004660302807356582, 'samples': 1931328, 'steps': 10058, 'loss/train': 1.8692198395729065} 01/29/2022 01:48:14 - INFO - codeparrot_training - Step 10059: {'lr': 0.0004660220453093903, 'samples': 1931520, 'steps': 10059, 'loss/train': 1.1258160173892975} 01/29/2022 01:48:18 - INFO - codeparrot_training - Step 10060: {'lr': 0.0004660138089577526, 'samples': 1931712, 'steps': 10060, 'loss/train': 1.876004934310913} 01/29/2022 01:48:22 - INFO - codeparrot_training - Step 10061: {'lr': 0.00046600557168078026, 'samples': 1931904, 'steps': 10061, 'loss/train': 2.614025115966797} 01/29/2022 01:48:26 - INFO - codeparrot_training - Step 10062: {'lr': 0.0004659973334785087, 'samples': 1932096, 'steps': 10062, 'loss/train': 1.1223741173744202} 01/29/2022 01:48:31 - INFO - codeparrot_training - Step 10063: {'lr': 0.00046598909435097315, 'samples': 1932288, 'steps': 10063, 'loss/train': 2.395276665687561} 01/29/2022 01:48:35 - INFO - codeparrot_training - Step 10064: {'lr': 0.0004659808542982088, 'samples': 1932480, 'steps': 10064, 'loss/train': 1.3296195566654205} 01/29/2022 01:48:40 - INFO - codeparrot_training - Step 10065: {'lr': 0.0004659726133202512, 'samples': 1932672, 'steps': 10065, 'loss/train': 1.902614951133728} 01/29/2022 01:48:44 - INFO - codeparrot_training - Step 10066: {'lr': 0.0004659643714171354, 'samples': 1932864, 'steps': 10066, 'loss/train': 1.997679054737091} 01/29/2022 01:48:49 - INFO - codeparrot_training - Step 10067: {'lr': 0.00046595612858889686, 'samples': 1933056, 'steps': 10067, 'loss/train': 1.3711969256401062} 01/29/2022 01:48:53 - INFO - codeparrot_training - Step 10068: {'lr': 0.00046594788483557084, 'samples': 1933248, 'steps': 10068, 'loss/train': 1.7007379531860352} 01/29/2022 01:48:57 - INFO - codeparrot_training - Step 10069: {'lr': 0.00046593964015719257, 'samples': 1933440, 'steps': 10069, 'loss/train': 1.4131675958633423} 01/29/2022 01:49:04 - INFO - codeparrot_training - Step 10070: {'lr': 0.0004659313945537975, 'samples': 1933632, 'steps': 10070, 'loss/train': 0.6016441583633423} 01/29/2022 01:49:08 - INFO - codeparrot_training - Step 10071: {'lr': 0.00046592314802542095, 'samples': 1933824, 'steps': 10071, 'loss/train': 1.4825356006622314} 01/29/2022 01:49:13 - INFO - codeparrot_training - Step 10072: {'lr': 0.0004659149005720982, 'samples': 1934016, 'steps': 10072, 'loss/train': 1.9098769426345825} 01/29/2022 01:49:17 - INFO - codeparrot_training - Step 10073: {'lr': 0.00046590665219386454, 'samples': 1934208, 'steps': 10073, 'loss/train': 2.2395359873771667} 01/29/2022 01:49:21 - INFO - codeparrot_training - Step 10074: {'lr': 0.0004658984028907553, 'samples': 1934400, 'steps': 10074, 'loss/train': 2.1628974080085754} 01/29/2022 01:49:27 - INFO - codeparrot_training - Step 10075: {'lr': 0.0004658901526628059, 'samples': 1934592, 'steps': 10075, 'loss/train': 1.3247108459472656} 01/29/2022 01:49:31 - INFO - codeparrot_training - Step 10076: {'lr': 0.00046588190151005163, 'samples': 1934784, 'steps': 10076, 'loss/train': 1.9790339469909668} 01/29/2022 01:49:35 - INFO - codeparrot_training - Step 10077: {'lr': 0.00046587364943252783, 'samples': 1934976, 'steps': 10077, 'loss/train': 2.184373140335083} 01/29/2022 01:49:39 - INFO - codeparrot_training - Step 10078: {'lr': 0.00046586539643026994, 'samples': 1935168, 'steps': 10078, 'loss/train': 1.9282214641571045} 01/29/2022 01:49:44 - INFO - codeparrot_training - Step 10079: {'lr': 0.0004658571425033131, 'samples': 1935360, 'steps': 10079, 'loss/train': 1.3327763378620148} 01/29/2022 01:49:49 - INFO - codeparrot_training - Step 10080: {'lr': 0.0004658488876516929, 'samples': 1935552, 'steps': 10080, 'loss/train': 1.8644043803215027} 01/29/2022 01:49:53 - INFO - codeparrot_training - Step 10081: {'lr': 0.0004658406318754446, 'samples': 1935744, 'steps': 10081, 'loss/train': 1.8561084866523743} 01/29/2022 01:49:57 - INFO - codeparrot_training - Step 10082: {'lr': 0.0004658323751746036, 'samples': 1935936, 'steps': 10082, 'loss/train': 1.2123920023441315} 01/29/2022 01:50:02 - INFO - codeparrot_training - Step 10083: {'lr': 0.00046582411754920517, 'samples': 1936128, 'steps': 10083, 'loss/train': 1.769654631614685} 01/29/2022 01:50:06 - INFO - codeparrot_training - Step 10084: {'lr': 0.0004658158589992848, 'samples': 1936320, 'steps': 10084, 'loss/train': 1.2982115149497986} 01/29/2022 01:50:11 - INFO - codeparrot_training - Step 10085: {'lr': 0.00046580759952487776, 'samples': 1936512, 'steps': 10085, 'loss/train': 3.171318769454956} 01/29/2022 01:50:16 - INFO - codeparrot_training - Step 10086: {'lr': 0.00046579933912601956, 'samples': 1936704, 'steps': 10086, 'loss/train': 1.6460092663764954} 01/29/2022 01:50:20 - INFO - codeparrot_training - Step 10087: {'lr': 0.00046579107780274543, 'samples': 1936896, 'steps': 10087, 'loss/train': 1.6831287145614624} 01/29/2022 01:50:25 - INFO - codeparrot_training - Step 10088: {'lr': 0.00046578281555509094, 'samples': 1937088, 'steps': 10088, 'loss/train': 1.8396878242492676} 01/29/2022 01:50:29 - INFO - codeparrot_training - Step 10089: {'lr': 0.0004657745523830914, 'samples': 1937280, 'steps': 10089, 'loss/train': 1.4506978690624237} 01/29/2022 01:50:37 - INFO - codeparrot_training - Step 10090: {'lr': 0.0004657662882867821, 'samples': 1937472, 'steps': 10090, 'loss/train': 1.1193044185638428} 01/29/2022 01:50:41 - INFO - codeparrot_training - Step 10091: {'lr': 0.0004657580232661985, 'samples': 1937664, 'steps': 10091, 'loss/train': 1.9980182647705078} 01/29/2022 01:50:45 - INFO - codeparrot_training - Step 10092: {'lr': 0.00046574975732137613, 'samples': 1937856, 'steps': 10092, 'loss/train': 1.83087819814682} 01/29/2022 01:50:49 - INFO - codeparrot_training - Step 10093: {'lr': 0.0004657414904523504, 'samples': 1938048, 'steps': 10093, 'loss/train': 0.8808763325214386} 01/29/2022 01:50:54 - INFO - codeparrot_training - Step 10094: {'lr': 0.0004657332226591565, 'samples': 1938240, 'steps': 10094, 'loss/train': 1.1642521619796753} 01/29/2022 01:50:58 - INFO - codeparrot_training - Step 10095: {'lr': 0.00046572495394183, 'samples': 1938432, 'steps': 10095, 'loss/train': 2.3081372380256653} 01/29/2022 01:51:03 - INFO - codeparrot_training - Step 10096: {'lr': 0.00046571668430040624, 'samples': 1938624, 'steps': 10096, 'loss/train': 2.495680868625641} 01/29/2022 01:51:08 - INFO - codeparrot_training - Step 10097: {'lr': 0.0004657084137349208, 'samples': 1938816, 'steps': 10097, 'loss/train': 1.179478257894516} 01/29/2022 01:51:12 - INFO - codeparrot_training - Step 10098: {'lr': 0.0004657001422454089, 'samples': 1939008, 'steps': 10098, 'loss/train': 1.7216035723686218} 01/29/2022 01:51:16 - INFO - codeparrot_training - Step 10099: {'lr': 0.0004656918698319062, 'samples': 1939200, 'steps': 10099, 'loss/train': 2.8020824790000916} 01/29/2022 01:51:20 - INFO - codeparrot_training - Step 10100: {'lr': 0.00046568359649444796, 'samples': 1939392, 'steps': 10100, 'loss/train': 2.3411765098571777} 01/29/2022 01:51:26 - INFO - codeparrot_training - Step 10101: {'lr': 0.0004656753222330697, 'samples': 1939584, 'steps': 10101, 'loss/train': 0.9064724743366241} 01/29/2022 01:51:30 - INFO - codeparrot_training - Step 10102: {'lr': 0.0004656670470478068, 'samples': 1939776, 'steps': 10102, 'loss/train': 2.064214825630188} 01/29/2022 01:51:34 - INFO - codeparrot_training - Step 10103: {'lr': 0.0004656587709386948, 'samples': 1939968, 'steps': 10103, 'loss/train': 1.7251977324485779} 01/29/2022 01:51:39 - INFO - codeparrot_training - Step 10104: {'lr': 0.00046565049390576906, 'samples': 1940160, 'steps': 10104, 'loss/train': 1.8851162195205688} 01/29/2022 01:51:46 - INFO - codeparrot_training - Step 10105: {'lr': 0.0004656422159490652, 'samples': 1940352, 'steps': 10105, 'loss/train': 0.8682186305522919} 01/29/2022 01:51:50 - INFO - codeparrot_training - Step 10106: {'lr': 0.00046563393706861847, 'samples': 1940544, 'steps': 10106, 'loss/train': 1.9099389910697937} 01/29/2022 01:51:54 - INFO - codeparrot_training - Step 10107: {'lr': 0.00046562565726446437, 'samples': 1940736, 'steps': 10107, 'loss/train': 1.6676902770996094} 01/29/2022 01:51:59 - INFO - codeparrot_training - Step 10108: {'lr': 0.0004656173765366385, 'samples': 1940928, 'steps': 10108, 'loss/train': 1.9144365191459656} 01/29/2022 01:52:03 - INFO - codeparrot_training - Step 10109: {'lr': 0.00046560909488517623, 'samples': 1941120, 'steps': 10109, 'loss/train': 0.18830875307321548} 01/29/2022 01:52:08 - INFO - codeparrot_training - Step 10110: {'lr': 0.0004656008123101131, 'samples': 1941312, 'steps': 10110, 'loss/train': 2.0798306465148926} 01/29/2022 01:52:12 - INFO - codeparrot_training - Step 10111: {'lr': 0.0004655925288114845, 'samples': 1941504, 'steps': 10111, 'loss/train': 1.311076819896698} 01/29/2022 01:52:17 - INFO - codeparrot_training - Step 10112: {'lr': 0.000465584244389326, 'samples': 1941696, 'steps': 10112, 'loss/train': 1.767151415348053} 01/29/2022 01:52:21 - INFO - codeparrot_training - Step 10113: {'lr': 0.000465575959043673, 'samples': 1941888, 'steps': 10113, 'loss/train': 1.71242094039917} 01/29/2022 01:52:25 - INFO - codeparrot_training - Step 10114: {'lr': 0.0004655676727745611, 'samples': 1942080, 'steps': 10114, 'loss/train': 2.0061964988708496} 01/29/2022 01:52:32 - INFO - codeparrot_training - Step 10115: {'lr': 0.0004655593855820257, 'samples': 1942272, 'steps': 10115, 'loss/train': 1.619692325592041} 01/29/2022 01:52:36 - INFO - codeparrot_training - Step 10116: {'lr': 0.00046555109746610244, 'samples': 1942464, 'steps': 10116, 'loss/train': 1.7891709208488464} 01/29/2022 01:52:41 - INFO - codeparrot_training - Step 10117: {'lr': 0.0004655428084268266, 'samples': 1942656, 'steps': 10117, 'loss/train': 0.9690996408462524} 01/29/2022 01:52:45 - INFO - codeparrot_training - Step 10118: {'lr': 0.00046553451846423387, 'samples': 1942848, 'steps': 10118, 'loss/train': 1.895133376121521} 01/29/2022 01:52:49 - INFO - codeparrot_training - Step 10119: {'lr': 0.0004655262275783597, 'samples': 1943040, 'steps': 10119, 'loss/train': 1.8821768760681152} 01/29/2022 01:52:54 - INFO - codeparrot_training - Step 10120: {'lr': 0.00046551793576923964, 'samples': 1943232, 'steps': 10120, 'loss/train': 2.0384373664855957} 01/29/2022 01:52:59 - INFO - codeparrot_training - Step 10121: {'lr': 0.0004655096430369091, 'samples': 1943424, 'steps': 10121, 'loss/train': 1.3667800426483154} 01/29/2022 01:53:03 - INFO - codeparrot_training - Step 10122: {'lr': 0.00046550134938140375, 'samples': 1943616, 'steps': 10122, 'loss/train': 1.749983310699463} 01/29/2022 01:53:07 - INFO - codeparrot_training - Step 10123: {'lr': 0.00046549305480275894, 'samples': 1943808, 'steps': 10123, 'loss/train': 1.6825684905052185} 01/29/2022 01:53:11 - INFO - codeparrot_training - Step 10124: {'lr': 0.0004654847593010104, 'samples': 1944000, 'steps': 10124, 'loss/train': 2.1488585472106934} 01/29/2022 01:53:17 - INFO - codeparrot_training - Step 10125: {'lr': 0.00046547646287619363, 'samples': 1944192, 'steps': 10125, 'loss/train': 1.4183043837547302} 01/29/2022 01:53:21 - INFO - codeparrot_training - Step 10126: {'lr': 0.00046546816552834404, 'samples': 1944384, 'steps': 10126, 'loss/train': 1.7077463865280151} 01/29/2022 01:53:25 - INFO - codeparrot_training - Step 10127: {'lr': 0.00046545986725749725, 'samples': 1944576, 'steps': 10127, 'loss/train': 1.298013836145401} 01/29/2022 01:53:29 - INFO - codeparrot_training - Step 10128: {'lr': 0.0004654515680636888, 'samples': 1944768, 'steps': 10128, 'loss/train': 2.285708248615265} 01/29/2022 01:53:34 - INFO - codeparrot_training - Step 10129: {'lr': 0.00046544326794695424, 'samples': 1944960, 'steps': 10129, 'loss/train': 0.8353098034858704} 01/29/2022 01:53:38 - INFO - codeparrot_training - Step 10130: {'lr': 0.00046543496690732914, 'samples': 1945152, 'steps': 10130, 'loss/train': 7.690306663513184} 01/29/2022 01:53:45 - INFO - codeparrot_training - Step 10131: {'lr': 0.0004654266649448491, 'samples': 1945344, 'steps': 10131, 'loss/train': 1.5591155290603638} 01/29/2022 01:53:49 - INFO - codeparrot_training - Step 10132: {'lr': 0.00046541836205954955, 'samples': 1945536, 'steps': 10132, 'loss/train': 2.00374835729599} 01/29/2022 01:53:54 - INFO - codeparrot_training - Step 10133: {'lr': 0.0004654100582514662, 'samples': 1945728, 'steps': 10133, 'loss/train': 1.6812701225280762} 01/29/2022 01:53:58 - INFO - codeparrot_training - Step 10134: {'lr': 0.0004654017535206345, 'samples': 1945920, 'steps': 10134, 'loss/train': 1.8142204284667969} 01/29/2022 01:54:04 - INFO - codeparrot_training - Step 10135: {'lr': 0.00046539344786709013, 'samples': 1946112, 'steps': 10135, 'loss/train': 1.3153761327266693} 01/29/2022 01:54:08 - INFO - codeparrot_training - Step 10136: {'lr': 0.0004653851412908686, 'samples': 1946304, 'steps': 10136, 'loss/train': 1.8384692072868347} 01/29/2022 01:54:12 - INFO - codeparrot_training - Step 10137: {'lr': 0.0004653768337920056, 'samples': 1946496, 'steps': 10137, 'loss/train': 1.555420696735382} 01/29/2022 01:54:16 - INFO - codeparrot_training - Step 10138: {'lr': 0.00046536852537053654, 'samples': 1946688, 'steps': 10138, 'loss/train': 1.8755941987037659} 01/29/2022 01:54:21 - INFO - codeparrot_training - Step 10139: {'lr': 0.00046536021602649715, 'samples': 1946880, 'steps': 10139, 'loss/train': 1.039050042629242} 01/29/2022 01:54:25 - INFO - codeparrot_training - Step 10140: {'lr': 0.0004653519057599229, 'samples': 1947072, 'steps': 10140, 'loss/train': 1.0783056914806366} 01/29/2022 01:54:30 - INFO - codeparrot_training - Step 10141: {'lr': 0.0004653435945708496, 'samples': 1947264, 'steps': 10141, 'loss/train': 2.573078155517578} 01/29/2022 01:54:34 - INFO - codeparrot_training - Step 10142: {'lr': 0.00046533528245931266, 'samples': 1947456, 'steps': 10142, 'loss/train': 2.277616024017334} 01/29/2022 01:54:39 - INFO - codeparrot_training - Step 10143: {'lr': 0.0004653269694253477, 'samples': 1947648, 'steps': 10143, 'loss/train': 2.2290483713150024} 01/29/2022 01:54:43 - INFO - codeparrot_training - Step 10144: {'lr': 0.00046531865546899044, 'samples': 1947840, 'steps': 10144, 'loss/train': 1.2815658152103424} 01/29/2022 01:54:47 - INFO - codeparrot_training - Step 10145: {'lr': 0.00046531034059027644, 'samples': 1948032, 'steps': 10145, 'loss/train': 0.08161896653473377} 01/29/2022 01:54:52 - INFO - codeparrot_training - Step 10146: {'lr': 0.0004653020247892412, 'samples': 1948224, 'steps': 10146, 'loss/train': 2.089769124984741} 01/29/2022 01:54:56 - INFO - codeparrot_training - Step 10147: {'lr': 0.0004652937080659206, 'samples': 1948416, 'steps': 10147, 'loss/train': 2.0837080478668213} 01/29/2022 01:55:01 - INFO - codeparrot_training - Step 10148: {'lr': 0.00046528539042035, 'samples': 1948608, 'steps': 10148, 'loss/train': 0.9614186882972717} 01/29/2022 01:55:05 - INFO - codeparrot_training - Step 10149: {'lr': 0.0004652770718525652, 'samples': 1948800, 'steps': 10149, 'loss/train': 1.1837078332901} 01/29/2022 01:55:09 - INFO - codeparrot_training - Step 10150: {'lr': 0.0004652687523626018, 'samples': 1948992, 'steps': 10150, 'loss/train': 1.7361323833465576} 01/29/2022 01:55:16 - INFO - codeparrot_training - Step 10151: {'lr': 0.0004652604319504954, 'samples': 1949184, 'steps': 10151, 'loss/train': 1.6110438108444214} 01/29/2022 01:55:20 - INFO - codeparrot_training - Step 10152: {'lr': 0.00046525211061628163, 'samples': 1949376, 'steps': 10152, 'loss/train': 1.7523635029792786} 01/29/2022 01:55:24 - INFO - codeparrot_training - Step 10153: {'lr': 0.0004652437883599962, 'samples': 1949568, 'steps': 10153, 'loss/train': 2.055122673511505} 01/29/2022 01:55:29 - INFO - codeparrot_training - Step 10154: {'lr': 0.0004652354651816747, 'samples': 1949760, 'steps': 10154, 'loss/train': 2.1387060284614563} 01/29/2022 01:55:34 - INFO - codeparrot_training - Step 10155: {'lr': 0.0004652271410813529, 'samples': 1949952, 'steps': 10155, 'loss/train': 2.4445740580558777} 01/29/2022 01:55:38 - INFO - codeparrot_training - Step 10156: {'lr': 0.0004652188160590663, 'samples': 1950144, 'steps': 10156, 'loss/train': 1.8389679193496704} 01/29/2022 01:55:42 - INFO - codeparrot_training - Step 10157: {'lr': 0.00046521049011485064, 'samples': 1950336, 'steps': 10157, 'loss/train': 2.6783464550971985} 01/29/2022 01:55:47 - INFO - codeparrot_training - Step 10158: {'lr': 0.0004652021632487415, 'samples': 1950528, 'steps': 10158, 'loss/train': 1.4002177119255066} 01/29/2022 01:55:51 - INFO - codeparrot_training - Step 10159: {'lr': 0.00046519383546077476, 'samples': 1950720, 'steps': 10159, 'loss/train': 0.957112580537796} 01/29/2022 01:55:57 - INFO - codeparrot_training - Step 10160: {'lr': 0.0004651855067509859, 'samples': 1950912, 'steps': 10160, 'loss/train': 1.1854931116104126} 01/29/2022 01:56:01 - INFO - codeparrot_training - Step 10161: {'lr': 0.00046517717711941066, 'samples': 1951104, 'steps': 10161, 'loss/train': 2.3499775528907776} 01/29/2022 01:56:06 - INFO - codeparrot_training - Step 10162: {'lr': 0.0004651688465660847, 'samples': 1951296, 'steps': 10162, 'loss/train': 2.151159703731537} 01/29/2022 01:56:10 - INFO - codeparrot_training - Step 10163: {'lr': 0.00046516051509104376, 'samples': 1951488, 'steps': 10163, 'loss/train': 2.014745593070984} 01/29/2022 01:56:14 - INFO - codeparrot_training - Step 10164: {'lr': 0.0004651521826943235, 'samples': 1951680, 'steps': 10164, 'loss/train': 2.886684536933899} 01/29/2022 01:56:20 - INFO - codeparrot_training - Step 10165: {'lr': 0.00046514384937595965, 'samples': 1951872, 'steps': 10165, 'loss/train': 1.0384769439697266} 01/29/2022 01:56:24 - INFO - codeparrot_training - Step 10166: {'lr': 0.0004651355151359878, 'samples': 1952064, 'steps': 10166, 'loss/train': 1.917663037776947} 01/29/2022 01:56:28 - INFO - codeparrot_training - Step 10167: {'lr': 0.0004651271799744437, 'samples': 1952256, 'steps': 10167, 'loss/train': 0.797090470790863} 01/29/2022 01:56:33 - INFO - codeparrot_training - Step 10168: {'lr': 0.0004651188438913631, 'samples': 1952448, 'steps': 10168, 'loss/train': 1.5568204522132874} 01/29/2022 01:56:37 - INFO - codeparrot_training - Step 10169: {'lr': 0.0004651105068867817, 'samples': 1952640, 'steps': 10169, 'loss/train': 1.791362464427948} 01/29/2022 01:56:42 - INFO - codeparrot_training - Step 10170: {'lr': 0.00046510216896073517, 'samples': 1952832, 'steps': 10170, 'loss/train': 2.3258866667747498} 01/29/2022 01:56:47 - INFO - codeparrot_training - Step 10171: {'lr': 0.00046509383011325925, 'samples': 1953024, 'steps': 10171, 'loss/train': 2.301881432533264} 01/29/2022 01:56:51 - INFO - codeparrot_training - Step 10172: {'lr': 0.0004650854903443896, 'samples': 1953216, 'steps': 10172, 'loss/train': 2.0276514887809753} 01/29/2022 01:56:55 - INFO - codeparrot_training - Step 10173: {'lr': 0.0004650771496541621, 'samples': 1953408, 'steps': 10173, 'loss/train': 2.5659558176994324} 01/29/2022 01:56:59 - INFO - codeparrot_training - Step 10174: {'lr': 0.0004650688080426123, 'samples': 1953600, 'steps': 10174, 'loss/train': 0.5799164324998856} 01/29/2022 01:57:06 - INFO - codeparrot_training - Step 10175: {'lr': 0.0004650604655097761, 'samples': 1953792, 'steps': 10175, 'loss/train': 1.346405804157257} 01/29/2022 01:57:10 - INFO - codeparrot_training - Step 10176: {'lr': 0.00046505212205568916, 'samples': 1953984, 'steps': 10176, 'loss/train': 1.539244830608368} 01/29/2022 01:57:15 - INFO - codeparrot_training - Step 10177: {'lr': 0.0004650437776803872, 'samples': 1954176, 'steps': 10177, 'loss/train': 1.9281034469604492} 01/29/2022 01:57:19 - INFO - codeparrot_training - Step 10178: {'lr': 0.00046503543238390595, 'samples': 1954368, 'steps': 10178, 'loss/train': 2.506875157356262} 01/29/2022 01:57:23 - INFO - codeparrot_training - Step 10179: {'lr': 0.0004650270861662812, 'samples': 1954560, 'steps': 10179, 'loss/train': 1.8842586278915405} 01/29/2022 01:57:27 - INFO - codeparrot_training - Step 10180: {'lr': 0.00046501873902754867, 'samples': 1954752, 'steps': 10180, 'loss/train': 2.1707645058631897} 01/29/2022 01:57:33 - INFO - codeparrot_training - Step 10181: {'lr': 0.00046501039096774415, 'samples': 1954944, 'steps': 10181, 'loss/train': 1.644096314907074} 01/29/2022 01:57:37 - INFO - codeparrot_training - Step 10182: {'lr': 0.00046500204198690343, 'samples': 1955136, 'steps': 10182, 'loss/train': 1.8046789169311523} 01/29/2022 01:57:41 - INFO - codeparrot_training - Step 10183: {'lr': 0.0004649936920850622, 'samples': 1955328, 'steps': 10183, 'loss/train': 4.110780715942383} 01/29/2022 01:57:46 - INFO - codeparrot_training - Step 10184: {'lr': 0.00046498534126225625, 'samples': 1955520, 'steps': 10184, 'loss/train': 1.186868280172348} 01/29/2022 01:57:50 - INFO - codeparrot_training - Step 10185: {'lr': 0.0004649769895185214, 'samples': 1955712, 'steps': 10185, 'loss/train': 3.096487283706665} 01/29/2022 01:57:55 - INFO - codeparrot_training - Step 10186: {'lr': 0.00046496863685389336, 'samples': 1955904, 'steps': 10186, 'loss/train': 1.903433918952942} 01/29/2022 01:57:59 - INFO - codeparrot_training - Step 10187: {'lr': 0.00046496028326840796, 'samples': 1956096, 'steps': 10187, 'loss/train': 1.3478753864765167} 01/29/2022 01:58:04 - INFO - codeparrot_training - Step 10188: {'lr': 0.000464951928762101, 'samples': 1956288, 'steps': 10188, 'loss/train': 1.879068911075592} 01/29/2022 01:58:08 - INFO - codeparrot_training - Step 10189: {'lr': 0.00046494357333500816, 'samples': 1956480, 'steps': 10189, 'loss/train': 1.8118262887001038} 01/29/2022 01:58:12 - INFO - codeparrot_training - Step 10190: {'lr': 0.00046493521698716536, 'samples': 1956672, 'steps': 10190, 'loss/train': 1.1367890238761902} 01/29/2022 01:58:18 - INFO - codeparrot_training - Step 10191: {'lr': 0.00046492685971860826, 'samples': 1956864, 'steps': 10191, 'loss/train': 2.0913782715797424} 01/29/2022 01:58:23 - INFO - codeparrot_training - Step 10192: {'lr': 0.00046491850152937276, 'samples': 1957056, 'steps': 10192, 'loss/train': 1.9677965641021729} 01/29/2022 01:58:27 - INFO - codeparrot_training - Step 10193: {'lr': 0.0004649101424194947, 'samples': 1957248, 'steps': 10193, 'loss/train': 2.7524730563163757} 01/29/2022 01:58:31 - INFO - codeparrot_training - Step 10194: {'lr': 0.0004649017823890098, 'samples': 1957440, 'steps': 10194, 'loss/train': 1.7170236110687256} 01/29/2022 01:58:35 - INFO - codeparrot_training - Step 10195: {'lr': 0.0004648934214379539, 'samples': 1957632, 'steps': 10195, 'loss/train': 1.5657563209533691} 01/29/2022 01:58:41 - INFO - codeparrot_training - Step 10196: {'lr': 0.00046488505956636286, 'samples': 1957824, 'steps': 10196, 'loss/train': 1.9420410990715027} 01/29/2022 01:58:45 - INFO - codeparrot_training - Step 10197: {'lr': 0.00046487669677427237, 'samples': 1958016, 'steps': 10197, 'loss/train': 2.545385420322418} 01/29/2022 01:58:49 - INFO - codeparrot_training - Step 10198: {'lr': 0.0004648683330617184, 'samples': 1958208, 'steps': 10198, 'loss/train': 1.1505011022090912} 01/29/2022 01:58:53 - INFO - codeparrot_training - Step 10199: {'lr': 0.00046485996842873676, 'samples': 1958400, 'steps': 10199, 'loss/train': 0.7740392982959747} 01/29/2022 01:58:58 - INFO - codeparrot_training - Step 10200: {'lr': 0.0004648516028753632, 'samples': 1958592, 'steps': 10200, 'loss/train': 2.3725647926330566} 01/29/2022 01:59:03 - INFO - codeparrot_training - Step 10201: {'lr': 0.00046484323640163356, 'samples': 1958784, 'steps': 10201, 'loss/train': 1.9355830550193787} 01/29/2022 01:59:08 - INFO - codeparrot_training - Step 10202: {'lr': 0.00046483486900758374, 'samples': 1958976, 'steps': 10202, 'loss/train': 2.1046282052993774} 01/29/2022 01:59:12 - INFO - codeparrot_training - Step 10203: {'lr': 0.0004648265006932496, 'samples': 1959168, 'steps': 10203, 'loss/train': 1.6122660040855408} 01/29/2022 01:59:16 - INFO - codeparrot_training - Step 10204: {'lr': 0.0004648181314586669, 'samples': 1959360, 'steps': 10204, 'loss/train': 1.4517356157302856} 01/29/2022 01:59:20 - INFO - codeparrot_training - Step 10205: {'lr': 0.00046480976130387156, 'samples': 1959552, 'steps': 10205, 'loss/train': 1.5824419856071472} 01/29/2022 01:59:27 - INFO - codeparrot_training - Step 10206: {'lr': 0.0004648013902288994, 'samples': 1959744, 'steps': 10206, 'loss/train': 1.2083078026771545} 01/29/2022 01:59:31 - INFO - codeparrot_training - Step 10207: {'lr': 0.0004647930182337863, 'samples': 1959936, 'steps': 10207, 'loss/train': 2.0530841946601868} 01/29/2022 01:59:35 - INFO - codeparrot_training - Step 10208: {'lr': 0.0004647846453185681, 'samples': 1960128, 'steps': 10208, 'loss/train': 1.9473904967308044} 01/29/2022 01:59:39 - INFO - codeparrot_training - Step 10209: {'lr': 0.0004647762714832807, 'samples': 1960320, 'steps': 10209, 'loss/train': 1.6014060974121094} 01/29/2022 01:59:44 - INFO - codeparrot_training - Step 10210: {'lr': 0.00046476789672795994, 'samples': 1960512, 'steps': 10210, 'loss/train': 1.7559555172920227} 01/29/2022 01:59:49 - INFO - codeparrot_training - Step 10211: {'lr': 0.00046475952105264176, 'samples': 1960704, 'steps': 10211, 'loss/train': 1.4732174277305603} 01/29/2022 01:59:53 - INFO - codeparrot_training - Step 10212: {'lr': 0.0004647511444573619, 'samples': 1960896, 'steps': 10212, 'loss/train': 1.9055870175361633} 01/29/2022 01:59:57 - INFO - codeparrot_training - Step 10213: {'lr': 0.00046474276694215635, 'samples': 1961088, 'steps': 10213, 'loss/train': 0.5670203268527985} 01/29/2022 02:00:02 - INFO - codeparrot_training - Step 10214: {'lr': 0.000464734388507061, 'samples': 1961280, 'steps': 10214, 'loss/train': 1.7040815949440002} 01/29/2022 02:00:06 - INFO - codeparrot_training - Step 10215: {'lr': 0.00046472600915211174, 'samples': 1961472, 'steps': 10215, 'loss/train': 1.2369346618652344} 01/29/2022 02:00:11 - INFO - codeparrot_training - Step 10216: {'lr': 0.00046471762887734437, 'samples': 1961664, 'steps': 10216, 'loss/train': 2.570968508720398} 01/29/2022 02:00:16 - INFO - codeparrot_training - Step 10217: {'lr': 0.0004647092476827949, 'samples': 1961856, 'steps': 10217, 'loss/train': 1.8915966153144836} 01/29/2022 02:00:20 - INFO - codeparrot_training - Step 10218: {'lr': 0.0004647008655684992, 'samples': 1962048, 'steps': 10218, 'loss/train': 2.3214662075042725} 01/29/2022 02:00:24 - INFO - codeparrot_training - Step 10219: {'lr': 0.00046469248253449316, 'samples': 1962240, 'steps': 10219, 'loss/train': 1.8622376918792725} 01/29/2022 02:00:28 - INFO - codeparrot_training - Step 10220: {'lr': 0.0004646840985808126, 'samples': 1962432, 'steps': 10220, 'loss/train': 1.9904247522354126} 01/29/2022 02:00:35 - INFO - codeparrot_training - Step 10221: {'lr': 0.00046467571370749366, 'samples': 1962624, 'steps': 10221, 'loss/train': 1.876541018486023} 01/29/2022 02:00:39 - INFO - codeparrot_training - Step 10222: {'lr': 0.0004646673279145721, 'samples': 1962816, 'steps': 10222, 'loss/train': 1.4235946834087372} 01/29/2022 02:00:43 - INFO - codeparrot_training - Step 10223: {'lr': 0.00046465894120208384, 'samples': 1963008, 'steps': 10223, 'loss/train': 0.6433491557836533} 01/29/2022 02:00:48 - INFO - codeparrot_training - Step 10224: {'lr': 0.00046465055357006494, 'samples': 1963200, 'steps': 10224, 'loss/train': 1.7199722528457642} 01/29/2022 02:00:52 - INFO - codeparrot_training - Step 10225: {'lr': 0.00046464216501855104, 'samples': 1963392, 'steps': 10225, 'loss/train': 2.2919331192970276} 01/29/2022 02:00:57 - INFO - codeparrot_training - Step 10226: {'lr': 0.0004646337755475784, 'samples': 1963584, 'steps': 10226, 'loss/train': 1.8751648664474487} 01/29/2022 02:01:01 - INFO - codeparrot_training - Step 10227: {'lr': 0.00046462538515718276, 'samples': 1963776, 'steps': 10227, 'loss/train': 2.5820928812026978} 01/29/2022 02:01:06 - INFO - codeparrot_training - Step 10228: {'lr': 0.0004646169938474002, 'samples': 1963968, 'steps': 10228, 'loss/train': 0.6457284986972809} 01/29/2022 02:01:10 - INFO - codeparrot_training - Step 10229: {'lr': 0.0004646086016182666, 'samples': 1964160, 'steps': 10229, 'loss/train': 0.922275960445404} 01/29/2022 02:01:14 - INFO - codeparrot_training - Step 10230: {'lr': 0.00046460020846981776, 'samples': 1964352, 'steps': 10230, 'loss/train': 1.8353117108345032} 01/29/2022 02:01:19 - INFO - codeparrot_training - Step 10231: {'lr': 0.00046459181440208986, 'samples': 1964544, 'steps': 10231, 'loss/train': 1.962851643562317} 01/29/2022 02:01:24 - INFO - codeparrot_training - Step 10232: {'lr': 0.0004645834194151187, 'samples': 1964736, 'steps': 10232, 'loss/train': 1.373026728630066} 01/29/2022 02:01:28 - INFO - codeparrot_training - Step 10233: {'lr': 0.00046457502350894046, 'samples': 1964928, 'steps': 10233, 'loss/train': 0.7221950590610504} 01/29/2022 02:01:32 - INFO - codeparrot_training - Step 10234: {'lr': 0.0004645666266835908, 'samples': 1965120, 'steps': 10234, 'loss/train': 1.9163507223129272} 01/29/2022 02:01:36 - INFO - codeparrot_training - Step 10235: {'lr': 0.0004645582289391059, 'samples': 1965312, 'steps': 10235, 'loss/train': 2.0049131512641907} 01/29/2022 02:01:43 - INFO - codeparrot_training - Step 10236: {'lr': 0.00046454983027552165, 'samples': 1965504, 'steps': 10236, 'loss/train': 1.6420354843139648} 01/29/2022 02:01:47 - INFO - codeparrot_training - Step 10237: {'lr': 0.0004645414306928741, 'samples': 1965696, 'steps': 10237, 'loss/train': 1.4250652492046356} 01/29/2022 02:01:51 - INFO - codeparrot_training - Step 10238: {'lr': 0.0004645330301911992, 'samples': 1965888, 'steps': 10238, 'loss/train': 1.9251192212104797} 01/29/2022 02:01:56 - INFO - codeparrot_training - Step 10239: {'lr': 0.0004645246287705329, 'samples': 1966080, 'steps': 10239, 'loss/train': 1.9506386518478394} 01/29/2022 02:02:01 - INFO - codeparrot_training - Step 10240: {'lr': 0.0004645162264309112, 'samples': 1966272, 'steps': 10240, 'loss/train': 0.8395531475543976} 01/29/2022 02:02:05 - INFO - codeparrot_training - Step 10241: {'lr': 0.0004645078231723701, 'samples': 1966464, 'steps': 10241, 'loss/train': 1.4434366822242737} 01/29/2022 02:02:09 - INFO - codeparrot_training - Step 10242: {'lr': 0.0004644994189949455, 'samples': 1966656, 'steps': 10242, 'loss/train': 1.7418403029441833} 01/29/2022 02:02:14 - INFO - codeparrot_training - Step 10243: {'lr': 0.00046449101389867364, 'samples': 1966848, 'steps': 10243, 'loss/train': 0.9129107594490051} 01/29/2022 02:02:18 - INFO - codeparrot_training - Step 10244: {'lr': 0.0004644826078835903, 'samples': 1967040, 'steps': 10244, 'loss/train': 2.094690442085266} 01/29/2022 02:02:24 - INFO - codeparrot_training - Step 10245: {'lr': 0.00046447420094973167, 'samples': 1967232, 'steps': 10245, 'loss/train': 0.8080131411552429} 01/29/2022 02:02:28 - INFO - codeparrot_training - Step 10246: {'lr': 0.0004644657930971336, 'samples': 1967424, 'steps': 10246, 'loss/train': 2.1317397952079773} 01/29/2022 02:02:33 - INFO - codeparrot_training - Step 10247: {'lr': 0.00046445738432583216, 'samples': 1967616, 'steps': 10247, 'loss/train': 1.5002660751342773} 01/29/2022 02:02:37 - INFO - codeparrot_training - Step 10248: {'lr': 0.00046444897463586345, 'samples': 1967808, 'steps': 10248, 'loss/train': 1.6706079840660095} 01/29/2022 02:02:41 - INFO - codeparrot_training - Step 10249: {'lr': 0.00046444056402726336, 'samples': 1968000, 'steps': 10249, 'loss/train': 1.3212739527225494} 01/29/2022 02:02:46 - INFO - codeparrot_training - Step 10250: {'lr': 0.00046443215250006805, 'samples': 1968192, 'steps': 10250, 'loss/train': 1.9551860690116882} 01/29/2022 02:02:51 - INFO - codeparrot_training - Step 10251: {'lr': 0.00046442374005431345, 'samples': 1968384, 'steps': 10251, 'loss/train': 2.034001350402832} 01/29/2022 02:02:55 - INFO - codeparrot_training - Step 10252: {'lr': 0.0004644153266900356, 'samples': 1968576, 'steps': 10252, 'loss/train': 1.8123494982719421} 01/29/2022 02:02:59 - INFO - codeparrot_training - Step 10253: {'lr': 0.0004644069124072706, 'samples': 1968768, 'steps': 10253, 'loss/train': 1.9388337135314941} 01/29/2022 02:03:03 - INFO - codeparrot_training - Step 10254: {'lr': 0.0004643984972060545, 'samples': 1968960, 'steps': 10254, 'loss/train': 1.3340267837047577} 01/29/2022 02:03:09 - INFO - codeparrot_training - Step 10255: {'lr': 0.00046439008108642335, 'samples': 1969152, 'steps': 10255, 'loss/train': 1.3191737830638885} 01/29/2022 02:03:13 - INFO - codeparrot_training - Step 10256: {'lr': 0.0004643816640484131, 'samples': 1969344, 'steps': 10256, 'loss/train': 1.2703944146633148} 01/29/2022 02:03:17 - INFO - codeparrot_training - Step 10257: {'lr': 0.0004643732460920599, 'samples': 1969536, 'steps': 10257, 'loss/train': 2.041809618473053} 01/29/2022 02:03:22 - INFO - codeparrot_training - Step 10258: {'lr': 0.00046436482721739976, 'samples': 1969728, 'steps': 10258, 'loss/train': 1.7417476773262024} 01/29/2022 02:03:26 - INFO - codeparrot_training - Step 10259: {'lr': 0.00046435640742446875, 'samples': 1969920, 'steps': 10259, 'loss/train': 2.021367073059082} 01/29/2022 02:03:30 - INFO - codeparrot_training - Step 10260: {'lr': 0.000464347986713303, 'samples': 1970112, 'steps': 10260, 'loss/train': 1.89224374294281} 01/29/2022 02:03:35 - INFO - codeparrot_training - Step 10261: {'lr': 0.00046433956508393855, 'samples': 1970304, 'steps': 10261, 'loss/train': 1.8632188439369202} 01/29/2022 02:03:40 - INFO - codeparrot_training - Step 10262: {'lr': 0.0004643311425364114, 'samples': 1970496, 'steps': 10262, 'loss/train': 1.5688505172729492} 01/29/2022 02:03:44 - INFO - codeparrot_training - Step 10263: {'lr': 0.0004643227190707577, 'samples': 1970688, 'steps': 10263, 'loss/train': 2.4274030923843384} 01/29/2022 02:03:48 - INFO - codeparrot_training - Step 10264: {'lr': 0.00046431429468701363, 'samples': 1970880, 'steps': 10264, 'loss/train': 1.488759309053421} 01/29/2022 02:03:53 - INFO - codeparrot_training - Step 10265: {'lr': 0.0004643058693852151, 'samples': 1971072, 'steps': 10265, 'loss/train': 2.1043179631233215} 01/29/2022 02:03:59 - INFO - codeparrot_training - Step 10266: {'lr': 0.0004642974431653983, 'samples': 1971264, 'steps': 10266, 'loss/train': 1.545377790927887} 01/29/2022 02:04:03 - INFO - codeparrot_training - Step 10267: {'lr': 0.00046428901602759933, 'samples': 1971456, 'steps': 10267, 'loss/train': 2.4428179264068604} 01/29/2022 02:04:07 - INFO - codeparrot_training - Step 10268: {'lr': 0.00046428058797185417, 'samples': 1971648, 'steps': 10268, 'loss/train': 1.8725641965866089} 01/29/2022 02:04:11 - INFO - codeparrot_training - Step 10269: {'lr': 0.0004642721589981991, 'samples': 1971840, 'steps': 10269, 'loss/train': 1.5402451157569885} 01/29/2022 02:04:16 - INFO - codeparrot_training - Step 10270: {'lr': 0.00046426372910667003, 'samples': 1972032, 'steps': 10270, 'loss/train': 1.631396234035492} 01/29/2022 02:04:21 - INFO - codeparrot_training - Step 10271: {'lr': 0.00046425529829730326, 'samples': 1972224, 'steps': 10271, 'loss/train': 1.8107013702392578} 01/29/2022 02:04:25 - INFO - codeparrot_training - Step 10272: {'lr': 0.0004642468665701348, 'samples': 1972416, 'steps': 10272, 'loss/train': 1.0152608156204224} 01/29/2022 02:04:30 - INFO - codeparrot_training - Step 10273: {'lr': 0.0004642384339252008, 'samples': 1972608, 'steps': 10273, 'loss/train': 2.307030200958252} 01/29/2022 02:04:34 - INFO - codeparrot_training - Step 10274: {'lr': 0.0004642300003625374, 'samples': 1972800, 'steps': 10274, 'loss/train': 1.0594737231731415} 01/29/2022 02:04:38 - INFO - codeparrot_training - Step 10275: {'lr': 0.0004642215658821807, 'samples': 1972992, 'steps': 10275, 'loss/train': 1.8283356428146362} 01/29/2022 02:04:43 - INFO - codeparrot_training - Step 10276: {'lr': 0.0004642131304841668, 'samples': 1973184, 'steps': 10276, 'loss/train': 1.9764642119407654} 01/29/2022 02:04:48 - INFO - codeparrot_training - Step 10277: {'lr': 0.00046420469416853197, 'samples': 1973376, 'steps': 10277, 'loss/train': 2.1226619482040405} 01/29/2022 02:04:52 - INFO - codeparrot_training - Step 10278: {'lr': 0.0004641962569353121, 'samples': 1973568, 'steps': 10278, 'loss/train': 1.2593618631362915} 01/29/2022 02:04:56 - INFO - codeparrot_training - Step 10279: {'lr': 0.0004641878187845436, 'samples': 1973760, 'steps': 10279, 'loss/train': 1.8663162589073181} 01/29/2022 02:05:00 - INFO - codeparrot_training - Step 10280: {'lr': 0.00046417937971626245, 'samples': 1973952, 'steps': 10280, 'loss/train': 1.5506705045700073} 01/29/2022 02:05:07 - INFO - codeparrot_training - Step 10281: {'lr': 0.00046417093973050486, 'samples': 1974144, 'steps': 10281, 'loss/train': 1.7515588402748108} 01/29/2022 02:05:11 - INFO - codeparrot_training - Step 10282: {'lr': 0.0004641624988273069, 'samples': 1974336, 'steps': 10282, 'loss/train': 1.2239179015159607} 01/29/2022 02:05:16 - INFO - codeparrot_training - Step 10283: {'lr': 0.0004641540570067049, 'samples': 1974528, 'steps': 10283, 'loss/train': 1.628600835800171} 01/29/2022 02:05:20 - INFO - codeparrot_training - Step 10284: {'lr': 0.0004641456142687348, 'samples': 1974720, 'steps': 10284, 'loss/train': 2.197891116142273} 01/29/2022 02:05:24 - INFO - codeparrot_training - Step 10285: {'lr': 0.000464137170613433, 'samples': 1974912, 'steps': 10285, 'loss/train': 1.5974555611610413} 01/29/2022 02:05:30 - INFO - codeparrot_training - Step 10286: {'lr': 0.00046412872604083554, 'samples': 1975104, 'steps': 10286, 'loss/train': 1.6320542693138123} 01/29/2022 02:05:34 - INFO - codeparrot_training - Step 10287: {'lr': 0.00046412028055097855, 'samples': 1975296, 'steps': 10287, 'loss/train': 1.8082714676856995} 01/29/2022 02:05:38 - INFO - codeparrot_training - Step 10288: {'lr': 0.00046411183414389834, 'samples': 1975488, 'steps': 10288, 'loss/train': 0.7530604898929596} 01/29/2022 02:05:42 - INFO - codeparrot_training - Step 10289: {'lr': 0.000464103386819631, 'samples': 1975680, 'steps': 10289, 'loss/train': 1.419023036956787} 01/29/2022 02:05:47 - INFO - codeparrot_training - Step 10290: {'lr': 0.00046409493857821273, 'samples': 1975872, 'steps': 10290, 'loss/train': 2.55893737077713} 01/29/2022 02:05:53 - INFO - codeparrot_training - Step 10291: {'lr': 0.00046408648941967975, 'samples': 1976064, 'steps': 10291, 'loss/train': 2.6620643734931946} 01/29/2022 02:05:57 - INFO - codeparrot_training - Step 10292: {'lr': 0.0004640780393440682, 'samples': 1976256, 'steps': 10292, 'loss/train': 1.9512407183647156} 01/29/2022 02:06:02 - INFO - codeparrot_training - Step 10293: {'lr': 0.0004640695883514143, 'samples': 1976448, 'steps': 10293, 'loss/train': 1.843548059463501} 01/29/2022 02:06:06 - INFO - codeparrot_training - Step 10294: {'lr': 0.0004640611364417543, 'samples': 1976640, 'steps': 10294, 'loss/train': 2.560545265674591} 01/29/2022 02:06:11 - INFO - codeparrot_training - Step 10295: {'lr': 0.0004640526836151243, 'samples': 1976832, 'steps': 10295, 'loss/train': 1.6445821523666382} 01/29/2022 02:06:15 - INFO - codeparrot_training - Step 10296: {'lr': 0.0004640442298715606, 'samples': 1977024, 'steps': 10296, 'loss/train': 1.8127241134643555} 01/29/2022 02:06:20 - INFO - codeparrot_training - Step 10297: {'lr': 0.0004640357752110994, 'samples': 1977216, 'steps': 10297, 'loss/train': 1.6562941074371338} 01/29/2022 02:06:24 - INFO - codeparrot_training - Step 10298: {'lr': 0.00046402731963377685, 'samples': 1977408, 'steps': 10298, 'loss/train': 1.7655453085899353} 01/29/2022 02:06:28 - INFO - codeparrot_training - Step 10299: {'lr': 0.0004640188631396293, 'samples': 1977600, 'steps': 10299, 'loss/train': 0.9088503420352936} 01/29/2022 02:06:33 - INFO - codeparrot_training - Step 10300: {'lr': 0.0004640104057286929, 'samples': 1977792, 'steps': 10300, 'loss/train': 1.7832884788513184} 01/29/2022 02:06:38 - INFO - codeparrot_training - Step 10301: {'lr': 0.0004640019474010038, 'samples': 1977984, 'steps': 10301, 'loss/train': 2.252906084060669} 01/29/2022 02:06:42 - INFO - codeparrot_training - Step 10302: {'lr': 0.00046399348815659837, 'samples': 1978176, 'steps': 10302, 'loss/train': 1.7271251678466797} 01/29/2022 02:06:46 - INFO - codeparrot_training - Step 10303: {'lr': 0.0004639850279955128, 'samples': 1978368, 'steps': 10303, 'loss/train': 0.9901270866394043} 01/29/2022 02:06:50 - INFO - codeparrot_training - Step 10304: {'lr': 0.0004639765669177833, 'samples': 1978560, 'steps': 10304, 'loss/train': 1.1583081185817719} 01/29/2022 02:06:57 - INFO - codeparrot_training - Step 10305: {'lr': 0.0004639681049234461, 'samples': 1978752, 'steps': 10305, 'loss/train': 1.7522476315498352} 01/29/2022 02:07:01 - INFO - codeparrot_training - Step 10306: {'lr': 0.0004639596420125375, 'samples': 1978944, 'steps': 10306, 'loss/train': 2.1118569374084473} 01/29/2022 02:07:05 - INFO - codeparrot_training - Step 10307: {'lr': 0.0004639511781850937, 'samples': 1979136, 'steps': 10307, 'loss/train': 1.2431582808494568} 01/29/2022 02:07:09 - INFO - codeparrot_training - Step 10308: {'lr': 0.000463942713441151, 'samples': 1979328, 'steps': 10308, 'loss/train': 1.66713148355484} 01/29/2022 02:07:14 - INFO - codeparrot_training - Step 10309: {'lr': 0.00046393424778074573, 'samples': 1979520, 'steps': 10309, 'loss/train': 2.0105745792388916} 01/29/2022 02:07:19 - INFO - codeparrot_training - Step 10310: {'lr': 0.000463925781203914, 'samples': 1979712, 'steps': 10310, 'loss/train': 1.2860544919967651} 01/29/2022 02:07:23 - INFO - codeparrot_training - Step 10311: {'lr': 0.00046391731371069224, 'samples': 1979904, 'steps': 10311, 'loss/train': 1.3863747417926788} 01/29/2022 02:07:28 - INFO - codeparrot_training - Step 10312: {'lr': 0.00046390884530111656, 'samples': 1980096, 'steps': 10312, 'loss/train': 2.176787853240967} 01/29/2022 02:07:32 - INFO - codeparrot_training - Step 10313: {'lr': 0.0004639003759752233, 'samples': 1980288, 'steps': 10313, 'loss/train': 1.8447686433792114} 01/29/2022 02:07:36 - INFO - codeparrot_training - Step 10314: {'lr': 0.00046389190573304875, 'samples': 1980480, 'steps': 10314, 'loss/train': 2.0807747840881348} 01/29/2022 02:07:41 - INFO - codeparrot_training - Step 10315: {'lr': 0.0004638834345746292, 'samples': 1980672, 'steps': 10315, 'loss/train': 2.3216707706451416} 01/29/2022 02:07:46 - INFO - codeparrot_training - Step 10316: {'lr': 0.00046387496250000095, 'samples': 1980864, 'steps': 10316, 'loss/train': 1.5773014426231384} 01/29/2022 02:07:50 - INFO - codeparrot_training - Step 10317: {'lr': 0.00046386648950920027, 'samples': 1981056, 'steps': 10317, 'loss/train': 0.9708190262317657} 01/29/2022 02:07:54 - INFO - codeparrot_training - Step 10318: {'lr': 0.0004638580156022635, 'samples': 1981248, 'steps': 10318, 'loss/train': 0.9396678507328033} 01/29/2022 02:07:58 - INFO - codeparrot_training - Step 10319: {'lr': 0.0004638495407792268, 'samples': 1981440, 'steps': 10319, 'loss/train': 1.428865760564804} 01/29/2022 02:08:04 - INFO - codeparrot_training - Step 10320: {'lr': 0.0004638410650401267, 'samples': 1981632, 'steps': 10320, 'loss/train': 1.6740147471427917} 01/29/2022 02:08:08 - INFO - codeparrot_training - Step 10321: {'lr': 0.0004638325883849993, 'samples': 1981824, 'steps': 10321, 'loss/train': 1.2966671884059906} 01/29/2022 02:08:12 - INFO - codeparrot_training - Step 10322: {'lr': 0.00046382411081388096, 'samples': 1982016, 'steps': 10322, 'loss/train': 1.921951711177826} 01/29/2022 02:08:16 - INFO - codeparrot_training - Step 10323: {'lr': 0.0004638156323268081, 'samples': 1982208, 'steps': 10323, 'loss/train': 1.8306536078453064} 01/29/2022 02:08:21 - INFO - codeparrot_training - Step 10324: {'lr': 0.00046380715292381695, 'samples': 1982400, 'steps': 10324, 'loss/train': 1.779377818107605} 01/29/2022 02:08:27 - INFO - codeparrot_training - Step 10325: {'lr': 0.0004637986726049438, 'samples': 1982592, 'steps': 10325, 'loss/train': 1.8141478300094604} 01/29/2022 02:08:31 - INFO - codeparrot_training - Step 10326: {'lr': 0.00046379019137022506, 'samples': 1982784, 'steps': 10326, 'loss/train': 2.0642539858818054} 01/29/2022 02:08:36 - INFO - codeparrot_training - Step 10327: {'lr': 0.000463781709219697, 'samples': 1982976, 'steps': 10327, 'loss/train': 1.213603287935257} 01/29/2022 02:08:40 - INFO - codeparrot_training - Step 10328: {'lr': 0.000463773226153396, 'samples': 1983168, 'steps': 10328, 'loss/train': 1.6722103357315063} 01/29/2022 02:08:44 - INFO - codeparrot_training - Step 10329: {'lr': 0.0004637647421713584, 'samples': 1983360, 'steps': 10329, 'loss/train': 2.0645259618759155} 01/29/2022 02:08:49 - INFO - codeparrot_training - Step 10330: {'lr': 0.0004637562572736205, 'samples': 1983552, 'steps': 10330, 'loss/train': 1.9142482280731201} 01/29/2022 02:08:54 - INFO - codeparrot_training - Step 10331: {'lr': 0.00046374777146021865, 'samples': 1983744, 'steps': 10331, 'loss/train': 2.4235677123069763} 01/29/2022 02:08:58 - INFO - codeparrot_training - Step 10332: {'lr': 0.00046373928473118927, 'samples': 1983936, 'steps': 10332, 'loss/train': 1.6544013619422913} 01/29/2022 02:09:02 - INFO - codeparrot_training - Step 10333: {'lr': 0.0004637307970865686, 'samples': 1984128, 'steps': 10333, 'loss/train': 1.455152839422226} 01/29/2022 02:09:06 - INFO - codeparrot_training - Step 10334: {'lr': 0.00046372230852639314, 'samples': 1984320, 'steps': 10334, 'loss/train': 1.3747150003910065} 01/29/2022 02:09:14 - INFO - codeparrot_training - Step 10335: {'lr': 0.0004637138190506991, 'samples': 1984512, 'steps': 10335, 'loss/train': 1.4546935558319092} 01/29/2022 02:09:18 - INFO - codeparrot_training - Step 10336: {'lr': 0.00046370532865952296, 'samples': 1984704, 'steps': 10336, 'loss/train': 3.456507682800293} 01/29/2022 02:09:22 - INFO - codeparrot_training - Step 10337: {'lr': 0.0004636968373529011, 'samples': 1984896, 'steps': 10337, 'loss/train': 1.196924775838852} 01/29/2022 02:09:27 - INFO - codeparrot_training - Step 10338: {'lr': 0.00046368834513086976, 'samples': 1985088, 'steps': 10338, 'loss/train': 1.5534946918487549} 01/29/2022 02:09:31 - INFO - codeparrot_training - Step 10339: {'lr': 0.00046367985199346546, 'samples': 1985280, 'steps': 10339, 'loss/train': 0.9436125755310059} 01/29/2022 02:09:36 - INFO - codeparrot_training - Step 10340: {'lr': 0.00046367135794072445, 'samples': 1985472, 'steps': 10340, 'loss/train': 1.9249151945114136} 01/29/2022 02:09:40 - INFO - codeparrot_training - Step 10341: {'lr': 0.00046366286297268327, 'samples': 1985664, 'steps': 10341, 'loss/train': 1.987684428691864} 01/29/2022 02:09:45 - INFO - codeparrot_training - Step 10342: {'lr': 0.0004636543670893782, 'samples': 1985856, 'steps': 10342, 'loss/train': 1.9488859176635742} 01/29/2022 02:09:49 - INFO - codeparrot_training - Step 10343: {'lr': 0.0004636458702908457, 'samples': 1986048, 'steps': 10343, 'loss/train': 1.9467805624008179} 01/29/2022 02:09:53 - INFO - codeparrot_training - Step 10344: {'lr': 0.0004636373725771221, 'samples': 1986240, 'steps': 10344, 'loss/train': 0.40585894882678986} 01/29/2022 02:09:59 - INFO - codeparrot_training - Step 10345: {'lr': 0.0004636288739482438, 'samples': 1986432, 'steps': 10345, 'loss/train': 0.9676865637302399} 01/29/2022 02:10:03 - INFO - codeparrot_training - Step 10346: {'lr': 0.0004636203744042473, 'samples': 1986624, 'steps': 10346, 'loss/train': 0.7036018073558807} 01/29/2022 02:10:07 - INFO - codeparrot_training - Step 10347: {'lr': 0.0004636118739451689, 'samples': 1986816, 'steps': 10347, 'loss/train': 2.5822806358337402} 01/29/2022 02:10:11 - INFO - codeparrot_training - Step 10348: {'lr': 0.0004636033725710451, 'samples': 1987008, 'steps': 10348, 'loss/train': 1.7044486999511719} 01/29/2022 02:10:16 - INFO - codeparrot_training - Step 10349: {'lr': 0.00046359487028191224, 'samples': 1987200, 'steps': 10349, 'loss/train': 1.7904925346374512} 01/29/2022 02:10:23 - INFO - codeparrot_training - Step 10350: {'lr': 0.0004635863670778068, 'samples': 1987392, 'steps': 10350, 'loss/train': 1.4993699491024017} 01/29/2022 02:10:27 - INFO - codeparrot_training - Step 10351: {'lr': 0.00046357786295876517, 'samples': 1987584, 'steps': 10351, 'loss/train': 2.04386830329895} 01/29/2022 02:10:31 - INFO - codeparrot_training - Step 10352: {'lr': 0.0004635693579248238, 'samples': 1987776, 'steps': 10352, 'loss/train': 1.404216080904007} 01/29/2022 02:10:35 - INFO - codeparrot_training - Step 10353: {'lr': 0.0004635608519760191, 'samples': 1987968, 'steps': 10353, 'loss/train': 2.4981783628463745} 01/29/2022 02:10:40 - INFO - codeparrot_training - Step 10354: {'lr': 0.00046355234511238756, 'samples': 1988160, 'steps': 10354, 'loss/train': 2.296134352684021} 01/29/2022 02:10:45 - INFO - codeparrot_training - Step 10355: {'lr': 0.00046354383733396553, 'samples': 1988352, 'steps': 10355, 'loss/train': 1.1272601187229156} 01/29/2022 02:10:49 - INFO - codeparrot_training - Step 10356: {'lr': 0.0004635353286407896, 'samples': 1988544, 'steps': 10356, 'loss/train': 1.7975739240646362} 01/29/2022 02:10:54 - INFO - codeparrot_training - Step 10357: {'lr': 0.00046352681903289605, 'samples': 1988736, 'steps': 10357, 'loss/train': 1.0629655122756958} 01/29/2022 02:10:58 - INFO - codeparrot_training - Step 10358: {'lr': 0.00046351830851032146, 'samples': 1988928, 'steps': 10358, 'loss/train': 1.4007860720157623} 01/29/2022 02:11:02 - INFO - codeparrot_training - Step 10359: {'lr': 0.00046350979707310226, 'samples': 1989120, 'steps': 10359, 'loss/train': 1.1456088423728943} 01/29/2022 02:11:08 - INFO - codeparrot_training - Step 10360: {'lr': 0.00046350128472127483, 'samples': 1989312, 'steps': 10360, 'loss/train': 2.174190938472748} 01/29/2022 02:11:12 - INFO - codeparrot_training - Step 10361: {'lr': 0.00046349277145487565, 'samples': 1989504, 'steps': 10361, 'loss/train': 2.8131390810012817} 01/29/2022 02:11:16 - INFO - codeparrot_training - Step 10362: {'lr': 0.00046348425727394126, 'samples': 1989696, 'steps': 10362, 'loss/train': 1.1495961248874664} 01/29/2022 02:11:21 - INFO - codeparrot_training - Step 10363: {'lr': 0.0004634757421785082, 'samples': 1989888, 'steps': 10363, 'loss/train': 1.4011945724487305} 01/29/2022 02:11:25 - INFO - codeparrot_training - Step 10364: {'lr': 0.0004634672261686127, 'samples': 1990080, 'steps': 10364, 'loss/train': 2.472922682762146} 01/29/2022 02:11:30 - INFO - codeparrot_training - Step 10365: {'lr': 0.0004634587092442915, 'samples': 1990272, 'steps': 10365, 'loss/train': 1.9496283531188965} 01/29/2022 02:11:34 - INFO - codeparrot_training - Step 10366: {'lr': 0.00046345019140558085, 'samples': 1990464, 'steps': 10366, 'loss/train': 1.2130300998687744} 01/29/2022 02:11:39 - INFO - codeparrot_training - Step 10367: {'lr': 0.0004634416726525175, 'samples': 1990656, 'steps': 10367, 'loss/train': 2.0814743041992188} 01/29/2022 02:11:43 - INFO - codeparrot_training - Step 10368: {'lr': 0.00046343315298513765, 'samples': 1990848, 'steps': 10368, 'loss/train': 1.6174618005752563} 01/29/2022 02:11:47 - INFO - codeparrot_training - Step 10369: {'lr': 0.0004634246324034781, 'samples': 1991040, 'steps': 10369, 'loss/train': 2.408037006855011} 01/29/2022 02:11:53 - INFO - codeparrot_training - Step 10370: {'lr': 0.0004634161109075751, 'samples': 1991232, 'steps': 10370, 'loss/train': 1.8478848338127136} 01/29/2022 02:11:57 - INFO - codeparrot_training - Step 10371: {'lr': 0.0004634075884974652, 'samples': 1991424, 'steps': 10371, 'loss/train': 2.0142678022384644} 01/29/2022 02:12:01 - INFO - codeparrot_training - Step 10372: {'lr': 0.00046339906517318507, 'samples': 1991616, 'steps': 10372, 'loss/train': 1.8933913707733154} 01/29/2022 02:12:05 - INFO - codeparrot_training - Step 10373: {'lr': 0.0004633905409347711, 'samples': 1991808, 'steps': 10373, 'loss/train': 1.4197501838207245} 01/29/2022 02:12:09 - INFO - codeparrot_training - Step 10374: {'lr': 0.00046338201578225975, 'samples': 1992000, 'steps': 10374, 'loss/train': 1.7709885835647583} 01/29/2022 02:12:15 - INFO - codeparrot_training - Step 10375: {'lr': 0.0004633734897156876, 'samples': 1992192, 'steps': 10375, 'loss/train': 1.6472956538200378} 01/29/2022 02:12:19 - INFO - codeparrot_training - Step 10376: {'lr': 0.0004633649627350912, 'samples': 1992384, 'steps': 10376, 'loss/train': 2.1560271978378296} 01/29/2022 02:12:23 - INFO - codeparrot_training - Step 10377: {'lr': 0.000463356434840507, 'samples': 1992576, 'steps': 10377, 'loss/train': 1.372110664844513} 01/29/2022 02:12:28 - INFO - codeparrot_training - Step 10378: {'lr': 0.0004633479060319717, 'samples': 1992768, 'steps': 10378, 'loss/train': 2.6294530034065247} 01/29/2022 02:12:32 - INFO - codeparrot_training - Step 10379: {'lr': 0.00046333937630952163, 'samples': 1992960, 'steps': 10379, 'loss/train': 2.5314184427261353} 01/29/2022 02:12:38 - INFO - codeparrot_training - Step 10380: {'lr': 0.00046333084567319344, 'samples': 1993152, 'steps': 10380, 'loss/train': 1.055513709783554} 01/29/2022 02:12:42 - INFO - codeparrot_training - Step 10381: {'lr': 0.0004633223141230236, 'samples': 1993344, 'steps': 10381, 'loss/train': 2.210267722606659} 01/29/2022 02:12:46 - INFO - codeparrot_training - Step 10382: {'lr': 0.0004633137816590488, 'samples': 1993536, 'steps': 10382, 'loss/train': 1.5198968052864075} 01/29/2022 02:12:51 - INFO - codeparrot_training - Step 10383: {'lr': 0.00046330524828130536, 'samples': 1993728, 'steps': 10383, 'loss/train': 1.7781917452812195} 01/29/2022 02:12:55 - INFO - codeparrot_training - Step 10384: {'lr': 0.00046329671398983007, 'samples': 1993920, 'steps': 10384, 'loss/train': 2.3882943391799927} 01/29/2022 02:13:00 - INFO - codeparrot_training - Step 10385: {'lr': 0.0004632881787846594, 'samples': 1994112, 'steps': 10385, 'loss/train': 1.239342749118805} 01/29/2022 02:13:04 - INFO - codeparrot_training - Step 10386: {'lr': 0.0004632796426658298, 'samples': 1994304, 'steps': 10386, 'loss/train': 1.976548433303833} 01/29/2022 02:13:08 - INFO - codeparrot_training - Step 10387: {'lr': 0.00046327110563337804, 'samples': 1994496, 'steps': 10387, 'loss/train': 2.1946130990982056} 01/29/2022 02:13:13 - INFO - codeparrot_training - Step 10388: {'lr': 0.00046326256768734053, 'samples': 1994688, 'steps': 10388, 'loss/train': 1.5021872520446777} 01/29/2022 02:13:17 - INFO - codeparrot_training - Step 10389: {'lr': 0.0004632540288277539, 'samples': 1994880, 'steps': 10389, 'loss/train': 1.7506712079048157} 01/29/2022 02:13:22 - INFO - codeparrot_training - Step 10390: {'lr': 0.0004632454890546547, 'samples': 1995072, 'steps': 10390, 'loss/train': 1.4730844795703888} 01/29/2022 02:13:27 - INFO - codeparrot_training - Step 10391: {'lr': 0.0004632369483680796, 'samples': 1995264, 'steps': 10391, 'loss/train': 1.1331894099712372} 01/29/2022 02:13:31 - INFO - codeparrot_training - Step 10392: {'lr': 0.0004632284067680651, 'samples': 1995456, 'steps': 10392, 'loss/train': 1.4295052886009216} 01/29/2022 02:13:35 - INFO - codeparrot_training - Step 10393: {'lr': 0.0004632198642546478, 'samples': 1995648, 'steps': 10393, 'loss/train': 1.4432195127010345} 01/29/2022 02:13:39 - INFO - codeparrot_training - Step 10394: {'lr': 0.0004632113208278643, 'samples': 1995840, 'steps': 10394, 'loss/train': 1.6039246916770935} 01/29/2022 02:13:45 - INFO - codeparrot_training - Step 10395: {'lr': 0.00046320277648775123, 'samples': 1996032, 'steps': 10395, 'loss/train': 2.0021769404411316} 01/29/2022 02:13:50 - INFO - codeparrot_training - Step 10396: {'lr': 0.0004631942312343452, 'samples': 1996224, 'steps': 10396, 'loss/train': 2.0506380200386047} 01/29/2022 02:13:54 - INFO - codeparrot_training - Step 10397: {'lr': 0.00046318568506768267, 'samples': 1996416, 'steps': 10397, 'loss/train': 1.4732894003391266} 01/29/2022 02:13:58 - INFO - codeparrot_training - Step 10398: {'lr': 0.0004631771379878005, 'samples': 1996608, 'steps': 10398, 'loss/train': 0.07829659804701805} 01/29/2022 02:14:02 - INFO - codeparrot_training - Step 10399: {'lr': 0.00046316858999473506, 'samples': 1996800, 'steps': 10399, 'loss/train': 2.005343735218048} 01/29/2022 02:14:08 - INFO - codeparrot_training - Step 10400: {'lr': 0.00046316004108852305, 'samples': 1996992, 'steps': 10400, 'loss/train': 2.1385855078697205} 01/29/2022 02:14:12 - INFO - codeparrot_training - Step 10401: {'lr': 0.0004631514912692012, 'samples': 1997184, 'steps': 10401, 'loss/train': 1.2496611177921295} 01/29/2022 02:14:16 - INFO - codeparrot_training - Step 10402: {'lr': 0.00046314294053680593, 'samples': 1997376, 'steps': 10402, 'loss/train': 2.552318751811981} 01/29/2022 02:14:20 - INFO - codeparrot_training - Step 10403: {'lr': 0.0004631343888913741, 'samples': 1997568, 'steps': 10403, 'loss/train': 1.0635836720466614} 01/29/2022 02:14:25 - INFO - codeparrot_training - Step 10404: {'lr': 0.00046312583633294213, 'samples': 1997760, 'steps': 10404, 'loss/train': 1.615422248840332} 01/29/2022 02:14:30 - INFO - codeparrot_training - Step 10405: {'lr': 0.0004631172828615469, 'samples': 1997952, 'steps': 10405, 'loss/train': 1.1322325766086578} 01/29/2022 02:14:34 - INFO - codeparrot_training - Step 10406: {'lr': 0.0004631087284772247, 'samples': 1998144, 'steps': 10406, 'loss/train': 1.6696429252624512} 01/29/2022 02:14:38 - INFO - codeparrot_training - Step 10407: {'lr': 0.0004631001731800125, 'samples': 1998336, 'steps': 10407, 'loss/train': 0.8825591504573822} 01/29/2022 02:14:42 - INFO - codeparrot_training - Step 10408: {'lr': 0.0004630916169699468, 'samples': 1998528, 'steps': 10408, 'loss/train': 2.6237568855285645} 01/29/2022 02:14:47 - INFO - codeparrot_training - Step 10409: {'lr': 0.00046308305984706435, 'samples': 1998720, 'steps': 10409, 'loss/train': 1.2536983788013458} 01/29/2022 02:14:53 - INFO - codeparrot_training - Step 10410: {'lr': 0.00046307450181140163, 'samples': 1998912, 'steps': 10410, 'loss/train': 1.6138161420822144} 01/29/2022 02:14:57 - INFO - codeparrot_training - Step 10411: {'lr': 0.00046306594286299544, 'samples': 1999104, 'steps': 10411, 'loss/train': 1.9073511958122253} 01/29/2022 02:15:02 - INFO - codeparrot_training - Step 10412: {'lr': 0.0004630573830018824, 'samples': 1999296, 'steps': 10412, 'loss/train': 1.9485588669776917} 01/29/2022 02:15:06 - INFO - codeparrot_training - Step 10413: {'lr': 0.00046304882222809917, 'samples': 1999488, 'steps': 10413, 'loss/train': 2.478643476963043} 01/29/2022 02:15:10 - INFO - codeparrot_training - Step 10414: {'lr': 0.0004630402605416825, 'samples': 1999680, 'steps': 10414, 'loss/train': 1.2862010300159454} 01/29/2022 02:15:16 - INFO - codeparrot_training - Step 10415: {'lr': 0.0004630316979426689, 'samples': 1999872, 'steps': 10415, 'loss/train': 1.9831482768058777} 01/29/2022 02:15:20 - INFO - codeparrot_training - Step 10416: {'lr': 0.00046302313443109523, 'samples': 2000064, 'steps': 10416, 'loss/train': 1.3861355781555176} 01/29/2022 02:15:24 - INFO - codeparrot_training - Step 10417: {'lr': 0.00046301457000699807, 'samples': 2000256, 'steps': 10417, 'loss/train': 1.9132928252220154} 01/29/2022 02:15:29 - INFO - codeparrot_training - Step 10418: {'lr': 0.0004630060046704141, 'samples': 2000448, 'steps': 10418, 'loss/train': 1.7030560970306396} 01/29/2022 02:15:34 - INFO - codeparrot_training - Step 10419: {'lr': 0.0004629974384213801, 'samples': 2000640, 'steps': 10419, 'loss/train': 1.4748719036579132} 01/29/2022 02:15:39 - INFO - codeparrot_training - Step 10420: {'lr': 0.0004629888712599327, 'samples': 2000832, 'steps': 10420, 'loss/train': 1.6077160835266113} 01/29/2022 02:15:43 - INFO - codeparrot_training - Step 10421: {'lr': 0.0004629803031861086, 'samples': 2001024, 'steps': 10421, 'loss/train': 0.4632362723350525} 01/29/2022 02:15:47 - INFO - codeparrot_training - Step 10422: {'lr': 0.0004629717341999445, 'samples': 2001216, 'steps': 10422, 'loss/train': 2.006037175655365} 01/29/2022 02:15:52 - INFO - codeparrot_training - Step 10423: {'lr': 0.0004629631643014771, 'samples': 2001408, 'steps': 10423, 'loss/train': 2.775108218193054} 01/29/2022 02:15:57 - INFO - codeparrot_training - Step 10424: {'lr': 0.00046295459349074316, 'samples': 2001600, 'steps': 10424, 'loss/train': 1.8099351525306702} 01/29/2022 02:16:01 - INFO - codeparrot_training - Step 10425: {'lr': 0.00046294602176777936, 'samples': 2001792, 'steps': 10425, 'loss/train': 1.7061039805412292} 01/29/2022 02:16:05 - INFO - codeparrot_training - Step 10426: {'lr': 0.0004629374491326224, 'samples': 2001984, 'steps': 10426, 'loss/train': 1.8333365321159363} 01/29/2022 02:16:10 - INFO - codeparrot_training - Step 10427: {'lr': 0.00046292887558530905, 'samples': 2002176, 'steps': 10427, 'loss/train': 2.1049768924713135} 01/29/2022 02:16:14 - INFO - codeparrot_training - Step 10428: {'lr': 0.000462920301125876, 'samples': 2002368, 'steps': 10428, 'loss/train': 1.555307686328888} 01/29/2022 02:16:19 - INFO - codeparrot_training - Step 10429: {'lr': 0.0004629117257543599, 'samples': 2002560, 'steps': 10429, 'loss/train': 1.6471232771873474} 01/29/2022 02:16:23 - INFO - codeparrot_training - Step 10430: {'lr': 0.0004629031494707977, 'samples': 2002752, 'steps': 10430, 'loss/train': 0.6949552148580551} 01/29/2022 02:16:28 - INFO - codeparrot_training - Step 10431: {'lr': 0.00046289457227522595, 'samples': 2002944, 'steps': 10431, 'loss/train': 2.7989078164100647} 01/29/2022 02:16:32 - INFO - codeparrot_training - Step 10432: {'lr': 0.0004628859941676815, 'samples': 2003136, 'steps': 10432, 'loss/train': 1.9233421683311462} 01/29/2022 02:16:36 - INFO - codeparrot_training - Step 10433: {'lr': 0.000462877415148201, 'samples': 2003328, 'steps': 10433, 'loss/train': 1.5787986516952515} 01/29/2022 02:16:42 - INFO - codeparrot_training - Step 10434: {'lr': 0.0004628688352168213, 'samples': 2003520, 'steps': 10434, 'loss/train': 2.204012632369995} 01/29/2022 02:16:46 - INFO - codeparrot_training - Step 10435: {'lr': 0.00046286025437357905, 'samples': 2003712, 'steps': 10435, 'loss/train': 1.77376127243042} 01/29/2022 02:16:50 - INFO - codeparrot_training - Step 10436: {'lr': 0.00046285167261851114, 'samples': 2003904, 'steps': 10436, 'loss/train': 0.801154375076294} 01/29/2022 02:16:54 - INFO - codeparrot_training - Step 10437: {'lr': 0.00046284308995165414, 'samples': 2004096, 'steps': 10437, 'loss/train': 1.2599983513355255} 01/29/2022 02:16:59 - INFO - codeparrot_training - Step 10438: {'lr': 0.00046283450637304497, 'samples': 2004288, 'steps': 10438, 'loss/train': 1.4943668246269226} 01/29/2022 02:17:03 - INFO - codeparrot_training - Step 10439: {'lr': 0.0004628259218827204, 'samples': 2004480, 'steps': 10439, 'loss/train': 2.34255588054657} 01/29/2022 02:17:09 - INFO - codeparrot_training - Step 10440: {'lr': 0.0004628173364807171, 'samples': 2004672, 'steps': 10440, 'loss/train': 2.0031752586364746} 01/29/2022 02:17:13 - INFO - codeparrot_training - Step 10441: {'lr': 0.00046280875016707195, 'samples': 2004864, 'steps': 10441, 'loss/train': 1.834102213382721} 01/29/2022 02:17:18 - INFO - codeparrot_training - Step 10442: {'lr': 0.0004628001629418217, 'samples': 2005056, 'steps': 10442, 'loss/train': 2.4021400809288025} 01/29/2022 02:17:22 - INFO - codeparrot_training - Step 10443: {'lr': 0.0004627915748050031, 'samples': 2005248, 'steps': 10443, 'loss/train': 1.7581563591957092} 01/29/2022 02:17:27 - INFO - codeparrot_training - Step 10444: {'lr': 0.000462782985756653, 'samples': 2005440, 'steps': 10444, 'loss/train': 1.6690723299980164} 01/29/2022 02:17:31 - INFO - codeparrot_training - Step 10445: {'lr': 0.0004627743957968081, 'samples': 2005632, 'steps': 10445, 'loss/train': 1.5540656447410583} 01/29/2022 02:17:36 - INFO - codeparrot_training - Step 10446: {'lr': 0.00046276580492550523, 'samples': 2005824, 'steps': 10446, 'loss/train': 1.7293340563774109} 01/29/2022 02:17:40 - INFO - codeparrot_training - Step 10447: {'lr': 0.0004627572131427813, 'samples': 2006016, 'steps': 10447, 'loss/train': 2.1670442819595337} 01/29/2022 02:17:44 - INFO - codeparrot_training - Step 10448: {'lr': 0.000462748620448673, 'samples': 2006208, 'steps': 10448, 'loss/train': 1.7648488283157349} 01/29/2022 02:17:48 - INFO - codeparrot_training - Step 10449: {'lr': 0.00046274002684321716, 'samples': 2006400, 'steps': 10449, 'loss/train': 2.1238174438476562} 01/29/2022 02:17:54 - INFO - codeparrot_training - Step 10450: {'lr': 0.00046273143232645054, 'samples': 2006592, 'steps': 10450, 'loss/train': 1.984116554260254} 01/29/2022 02:17:58 - INFO - codeparrot_training - Step 10451: {'lr': 0.0004627228368984101, 'samples': 2006784, 'steps': 10451, 'loss/train': 1.3784101009368896} 01/29/2022 02:18:02 - INFO - codeparrot_training - Step 10452: {'lr': 0.0004627142405591325, 'samples': 2006976, 'steps': 10452, 'loss/train': 1.2459063827991486} 01/29/2022 02:18:07 - INFO - codeparrot_training - Step 10453: {'lr': 0.00046270564330865466, 'samples': 2007168, 'steps': 10453, 'loss/train': 2.43729704618454} 01/29/2022 02:18:13 - INFO - codeparrot_training - Step 10454: {'lr': 0.0004626970451470134, 'samples': 2007360, 'steps': 10454, 'loss/train': 1.974733293056488} 01/29/2022 02:18:17 - INFO - codeparrot_training - Step 10455: {'lr': 0.0004626884460742455, 'samples': 2007552, 'steps': 10455, 'loss/train': 1.6557067036628723} 01/29/2022 02:18:21 - INFO - codeparrot_training - Step 10456: {'lr': 0.00046267984609038793, 'samples': 2007744, 'steps': 10456, 'loss/train': 1.7072839736938477} 01/29/2022 02:18:25 - INFO - codeparrot_training - Step 10457: {'lr': 0.0004626712451954773, 'samples': 2007936, 'steps': 10457, 'loss/train': 2.083284616470337} 01/29/2022 02:18:30 - INFO - codeparrot_training - Step 10458: {'lr': 0.0004626626433895507, 'samples': 2008128, 'steps': 10458, 'loss/train': 1.2306950390338898} 01/29/2022 02:18:35 - INFO - codeparrot_training - Step 10459: {'lr': 0.00046265404067264484, 'samples': 2008320, 'steps': 10459, 'loss/train': 1.3687025606632233} 01/29/2022 02:18:39 - INFO - codeparrot_training - Step 10460: {'lr': 0.00046264543704479654, 'samples': 2008512, 'steps': 10460, 'loss/train': 1.6769964694976807} 01/29/2022 02:18:43 - INFO - codeparrot_training - Step 10461: {'lr': 0.0004626368325060428, 'samples': 2008704, 'steps': 10461, 'loss/train': 1.7890827655792236} 01/29/2022 02:18:48 - INFO - codeparrot_training - Step 10462: {'lr': 0.00046262822705642025, 'samples': 2008896, 'steps': 10462, 'loss/train': 1.1654881238937378} 01/29/2022 02:18:52 - INFO - codeparrot_training - Step 10463: {'lr': 0.00046261962069596603, 'samples': 2009088, 'steps': 10463, 'loss/train': 1.7898820638656616} 01/29/2022 02:18:58 - INFO - codeparrot_training - Step 10464: {'lr': 0.0004626110134247168, 'samples': 2009280, 'steps': 10464, 'loss/train': 1.6845068335533142} 01/29/2022 02:19:02 - INFO - codeparrot_training - Step 10465: {'lr': 0.0004626024052427095, 'samples': 2009472, 'steps': 10465, 'loss/train': 2.6259060502052307} 01/29/2022 02:19:06 - INFO - codeparrot_training - Step 10466: {'lr': 0.00046259379614998103, 'samples': 2009664, 'steps': 10466, 'loss/train': 2.335039258003235} 01/29/2022 02:19:10 - INFO - codeparrot_training - Step 10467: {'lr': 0.00046258518614656827, 'samples': 2009856, 'steps': 10467, 'loss/train': 2.275171995162964} 01/29/2022 02:19:15 - INFO - codeparrot_training - Step 10468: {'lr': 0.0004625765752325081, 'samples': 2010048, 'steps': 10468, 'loss/train': 1.6959807872772217} 01/29/2022 02:19:20 - INFO - codeparrot_training - Step 10469: {'lr': 0.0004625679634078372, 'samples': 2010240, 'steps': 10469, 'loss/train': 0.7854411005973816} 01/29/2022 02:19:24 - INFO - codeparrot_training - Step 10470: {'lr': 0.0004625593506725928, 'samples': 2010432, 'steps': 10470, 'loss/train': 1.6876174807548523} 01/29/2022 02:19:29 - INFO - codeparrot_training - Step 10471: {'lr': 0.0004625507370268116, 'samples': 2010624, 'steps': 10471, 'loss/train': 2.516521990299225} 01/29/2022 02:19:33 - INFO - codeparrot_training - Step 10472: {'lr': 0.00046254212247053055, 'samples': 2010816, 'steps': 10472, 'loss/train': 1.2072528898715973} 01/29/2022 02:19:37 - INFO - codeparrot_training - Step 10473: {'lr': 0.00046253350700378655, 'samples': 2011008, 'steps': 10473, 'loss/train': 2.29680597782135} 01/29/2022 02:19:42 - INFO - codeparrot_training - Step 10474: {'lr': 0.0004625248906266165, 'samples': 2011200, 'steps': 10474, 'loss/train': 1.9576857089996338} 01/29/2022 02:19:47 - INFO - codeparrot_training - Step 10475: {'lr': 0.00046251627333905723, 'samples': 2011392, 'steps': 10475, 'loss/train': 1.6133944988250732} 01/29/2022 02:19:51 - INFO - codeparrot_training - Step 10476: {'lr': 0.0004625076551411458, 'samples': 2011584, 'steps': 10476, 'loss/train': 1.755099892616272} 01/29/2022 02:19:55 - INFO - codeparrot_training - Step 10477: {'lr': 0.000462499036032919, 'samples': 2011776, 'steps': 10477, 'loss/train': 2.4364133477211} 01/29/2022 02:19:59 - INFO - codeparrot_training - Step 10478: {'lr': 0.0004624904160144138, 'samples': 2011968, 'steps': 10478, 'loss/train': 1.9887292385101318} 01/29/2022 02:20:05 - INFO - codeparrot_training - Step 10479: {'lr': 0.00046248179508566716, 'samples': 2012160, 'steps': 10479, 'loss/train': 0.05933966115117073} 01/29/2022 02:20:10 - INFO - codeparrot_training - Step 10480: {'lr': 0.000462473173246716, 'samples': 2012352, 'steps': 10480, 'loss/train': 1.0021434724330902} 01/29/2022 02:20:14 - INFO - codeparrot_training - Step 10481: {'lr': 0.00046246455049759716, 'samples': 2012544, 'steps': 10481, 'loss/train': 2.2269351482391357} 01/29/2022 02:20:18 - INFO - codeparrot_training - Step 10482: {'lr': 0.00046245592683834773, 'samples': 2012736, 'steps': 10482, 'loss/train': 1.9856103658676147} 01/29/2022 02:20:22 - INFO - codeparrot_training - Step 10483: {'lr': 0.00046244730226900453, 'samples': 2012928, 'steps': 10483, 'loss/train': 2.1203116178512573} 01/29/2022 02:20:28 - INFO - codeparrot_training - Step 10484: {'lr': 0.00046243867678960463, 'samples': 2013120, 'steps': 10484, 'loss/train': 1.3930315375328064} 01/29/2022 02:20:32 - INFO - codeparrot_training - Step 10485: {'lr': 0.00046243005040018484, 'samples': 2013312, 'steps': 10485, 'loss/train': 2.1421619653701782} 01/29/2022 02:20:36 - INFO - codeparrot_training - Step 10486: {'lr': 0.0004624214231007821, 'samples': 2013504, 'steps': 10486, 'loss/train': 1.4104161858558655} 01/29/2022 02:20:40 - INFO - codeparrot_training - Step 10487: {'lr': 0.0004624127948914335, 'samples': 2013696, 'steps': 10487, 'loss/train': 0.6160946488380432} 01/29/2022 02:20:45 - INFO - codeparrot_training - Step 10488: {'lr': 0.0004624041657721759, 'samples': 2013888, 'steps': 10488, 'loss/train': 2.047080874443054} 01/29/2022 02:20:50 - INFO - codeparrot_training - Step 10489: {'lr': 0.0004623955357430464, 'samples': 2014080, 'steps': 10489, 'loss/train': 1.9038893580436707} 01/29/2022 02:20:54 - INFO - codeparrot_training - Step 10490: {'lr': 0.0004623869048040817, 'samples': 2014272, 'steps': 10490, 'loss/train': 1.7116990685462952} 01/29/2022 02:20:58 - INFO - codeparrot_training - Step 10491: {'lr': 0.0004623782729553191, 'samples': 2014464, 'steps': 10491, 'loss/train': 2.095885992050171} 01/29/2022 02:21:03 - INFO - codeparrot_training - Step 10492: {'lr': 0.00046236964019679533, 'samples': 2014656, 'steps': 10492, 'loss/train': 0.9931211471557617} 01/29/2022 02:21:07 - INFO - codeparrot_training - Step 10493: {'lr': 0.0004623610065285475, 'samples': 2014848, 'steps': 10493, 'loss/train': 1.6276824474334717} 01/29/2022 02:21:13 - INFO - codeparrot_training - Step 10494: {'lr': 0.00046235237195061253, 'samples': 2015040, 'steps': 10494, 'loss/train': 1.8396783471107483} 01/29/2022 02:21:17 - INFO - codeparrot_training - Step 10495: {'lr': 0.00046234373646302743, 'samples': 2015232, 'steps': 10495, 'loss/train': 2.076543867588043} 01/29/2022 02:21:21 - INFO - codeparrot_training - Step 10496: {'lr': 0.00046233510006582913, 'samples': 2015424, 'steps': 10496, 'loss/train': 1.7780644297599792} 01/29/2022 02:21:26 - INFO - codeparrot_training - Step 10497: {'lr': 0.00046232646275905475, 'samples': 2015616, 'steps': 10497, 'loss/train': 2.199299991130829} 01/29/2022 02:21:30 - INFO - codeparrot_training - Step 10498: {'lr': 0.00046231782454274117, 'samples': 2015808, 'steps': 10498, 'loss/train': 2.0793803930282593} 01/29/2022 02:21:35 - INFO - codeparrot_training - Step 10499: {'lr': 0.00046230918541692557, 'samples': 2016000, 'steps': 10499, 'loss/train': 1.924455463886261} 01/29/2022 02:21:39 - INFO - codeparrot_training - Step 10500: {'lr': 0.00046230054538164475, 'samples': 2016192, 'steps': 10500, 'loss/train': 1.4760064780712128} 01/29/2022 02:21:44 - INFO - codeparrot_training - Step 10501: {'lr': 0.0004622919044369358, 'samples': 2016384, 'steps': 10501, 'loss/train': 1.036214143037796} 01/29/2022 02:21:48 - INFO - codeparrot_training - Step 10502: {'lr': 0.00046228326258283576, 'samples': 2016576, 'steps': 10502, 'loss/train': 1.60471111536026} 01/29/2022 02:21:53 - INFO - codeparrot_training - Step 10503: {'lr': 0.0004622746198193816, 'samples': 2016768, 'steps': 10503, 'loss/train': 1.9483169317245483} 01/29/2022 02:21:57 - INFO - codeparrot_training - Step 10504: {'lr': 0.00046226597614661044, 'samples': 2016960, 'steps': 10504, 'loss/train': 0.8393184542655945} 01/29/2022 02:22:02 - INFO - codeparrot_training - Step 10505: {'lr': 0.00046225733156455916, 'samples': 2017152, 'steps': 10505, 'loss/train': 1.677347481250763} 01/29/2022 02:22:06 - INFO - codeparrot_training - Step 10506: {'lr': 0.00046224868607326494, 'samples': 2017344, 'steps': 10506, 'loss/train': 0.9656520187854767} 01/29/2022 02:22:10 - INFO - codeparrot_training - Step 10507: {'lr': 0.00046224003967276474, 'samples': 2017536, 'steps': 10507, 'loss/train': 1.704331398010254} 01/29/2022 02:22:16 - INFO - codeparrot_training - Step 10508: {'lr': 0.00046223139236309553, 'samples': 2017728, 'steps': 10508, 'loss/train': 1.45520880818367} 01/29/2022 02:22:20 - INFO - codeparrot_training - Step 10509: {'lr': 0.0004622227441442945, 'samples': 2017920, 'steps': 10509, 'loss/train': 1.4532374739646912} 01/29/2022 02:22:25 - INFO - codeparrot_training - Step 10510: {'lr': 0.00046221409501639863, 'samples': 2018112, 'steps': 10510, 'loss/train': 1.6910693049430847} 01/29/2022 02:22:29 - INFO - codeparrot_training - Step 10511: {'lr': 0.0004622054449794449, 'samples': 2018304, 'steps': 10511, 'loss/train': 1.9708836078643799} 01/29/2022 02:22:33 - INFO - codeparrot_training - Step 10512: {'lr': 0.0004621967940334705, 'samples': 2018496, 'steps': 10512, 'loss/train': 0.8846657574176788} 01/29/2022 02:22:39 - INFO - codeparrot_training - Step 10513: {'lr': 0.00046218814217851233, 'samples': 2018688, 'steps': 10513, 'loss/train': 1.5239213705062866} 01/29/2022 02:22:43 - INFO - codeparrot_training - Step 10514: {'lr': 0.0004621794894146076, 'samples': 2018880, 'steps': 10514, 'loss/train': 2.21094810962677} 01/29/2022 02:22:47 - INFO - codeparrot_training - Step 10515: {'lr': 0.0004621708357417933, 'samples': 2019072, 'steps': 10515, 'loss/train': 1.2871394455432892} 01/29/2022 02:22:51 - INFO - codeparrot_training - Step 10516: {'lr': 0.00046216218116010646, 'samples': 2019264, 'steps': 10516, 'loss/train': 1.7460370659828186} 01/29/2022 02:22:56 - INFO - codeparrot_training - Step 10517: {'lr': 0.00046215352566958423, 'samples': 2019456, 'steps': 10517, 'loss/train': 2.1532487869262695} 01/29/2022 02:23:01 - INFO - codeparrot_training - Step 10518: {'lr': 0.00046214486927026373, 'samples': 2019648, 'steps': 10518, 'loss/train': 1.093041479587555} 01/29/2022 02:23:05 - INFO - codeparrot_training - Step 10519: {'lr': 0.0004621362119621819, 'samples': 2019840, 'steps': 10519, 'loss/train': 2.753826856613159} 01/29/2022 02:23:09 - INFO - codeparrot_training - Step 10520: {'lr': 0.00046212755374537594, 'samples': 2020032, 'steps': 10520, 'loss/train': 1.829745590686798} 01/29/2022 02:23:14 - INFO - codeparrot_training - Step 10521: {'lr': 0.00046211889461988286, 'samples': 2020224, 'steps': 10521, 'loss/train': 1.4865564107894897} 01/29/2022 02:23:18 - INFO - codeparrot_training - Step 10522: {'lr': 0.0004621102345857399, 'samples': 2020416, 'steps': 10522, 'loss/train': 2.4513596892356873} 01/29/2022 02:23:23 - INFO - codeparrot_training - Step 10523: {'lr': 0.0004621015736429839, 'samples': 2020608, 'steps': 10523, 'loss/train': 2.138052463531494} 01/29/2022 02:23:29 - INFO - codeparrot_training - Step 10524: {'lr': 0.00046209291179165216, 'samples': 2020800, 'steps': 10524, 'loss/train': 2.3575488924980164} 01/29/2022 02:23:33 - INFO - codeparrot_training - Step 10525: {'lr': 0.0004620842490317817, 'samples': 2020992, 'steps': 10525, 'loss/train': 1.8883066177368164} 01/29/2022 02:23:38 - INFO - codeparrot_training - Step 10526: {'lr': 0.0004620755853634097, 'samples': 2021184, 'steps': 10526, 'loss/train': 0.8183105885982513} 01/29/2022 02:23:42 - INFO - codeparrot_training - Step 10527: {'lr': 0.00046206692078657325, 'samples': 2021376, 'steps': 10527, 'loss/train': 1.4288960695266724} 01/29/2022 02:23:46 - INFO - codeparrot_training - Step 10528: {'lr': 0.0004620582553013094, 'samples': 2021568, 'steps': 10528, 'loss/train': 0.17438143864274025} 01/29/2022 02:23:52 - INFO - codeparrot_training - Step 10529: {'lr': 0.00046204958890765536, 'samples': 2021760, 'steps': 10529, 'loss/train': 0.7521646320819855} 01/29/2022 02:23:56 - INFO - codeparrot_training - Step 10530: {'lr': 0.0004620409216056483, 'samples': 2021952, 'steps': 10530, 'loss/train': 1.882391095161438} 01/29/2022 02:24:00 - INFO - codeparrot_training - Step 10531: {'lr': 0.00046203225339532515, 'samples': 2022144, 'steps': 10531, 'loss/train': 1.4050602614879608} 01/29/2022 02:24:04 - INFO - codeparrot_training - Step 10532: {'lr': 0.00046202358427672313, 'samples': 2022336, 'steps': 10532, 'loss/train': 1.8752031326293945} 01/29/2022 02:24:09 - INFO - codeparrot_training - Step 10533: {'lr': 0.0004620149142498795, 'samples': 2022528, 'steps': 10533, 'loss/train': 2.27212393283844} 01/29/2022 02:24:14 - INFO - codeparrot_training - Step 10534: {'lr': 0.0004620062433148312, 'samples': 2022720, 'steps': 10534, 'loss/train': 0.8607363402843475} 01/29/2022 02:24:18 - INFO - codeparrot_training - Step 10535: {'lr': 0.00046199757147161554, 'samples': 2022912, 'steps': 10535, 'loss/train': 1.8948953747749329} 01/29/2022 02:24:23 - INFO - codeparrot_training - Step 10536: {'lr': 0.00046198889872026963, 'samples': 2023104, 'steps': 10536, 'loss/train': 2.1178545355796814} 01/29/2022 02:24:27 - INFO - codeparrot_training - Step 10537: {'lr': 0.0004619802250608305, 'samples': 2023296, 'steps': 10537, 'loss/train': 1.8635759353637695} 01/29/2022 02:24:31 - INFO - codeparrot_training - Step 10538: {'lr': 0.0004619715504933354, 'samples': 2023488, 'steps': 10538, 'loss/train': 1.412628561258316} 01/29/2022 02:24:36 - INFO - codeparrot_training - Step 10539: {'lr': 0.00046196287501782155, 'samples': 2023680, 'steps': 10539, 'loss/train': 1.9972962141036987} 01/29/2022 02:24:40 - INFO - codeparrot_training - Step 10540: {'lr': 0.00046195419863432604, 'samples': 2023872, 'steps': 10540, 'loss/train': 2.123002052307129} 01/29/2022 02:24:45 - INFO - codeparrot_training - Step 10541: {'lr': 0.000461945521342886, 'samples': 2024064, 'steps': 10541, 'loss/train': 1.2753423750400543} 01/29/2022 02:24:49 - INFO - codeparrot_training - Step 10542: {'lr': 0.0004619368431435387, 'samples': 2024256, 'steps': 10542, 'loss/train': 1.6891783475875854} 01/29/2022 02:24:53 - INFO - codeparrot_training - Step 10543: {'lr': 0.0004619281640363212, 'samples': 2024448, 'steps': 10543, 'loss/train': 2.0057349801063538} 01/29/2022 02:25:00 - INFO - codeparrot_training - Step 10544: {'lr': 0.0004619194840212708, 'samples': 2024640, 'steps': 10544, 'loss/train': 1.460323691368103} 01/29/2022 02:25:04 - INFO - codeparrot_training - Step 10545: {'lr': 0.00046191080309842457, 'samples': 2024832, 'steps': 10545, 'loss/train': 1.9816632270812988} 01/29/2022 02:25:08 - INFO - codeparrot_training - Step 10546: {'lr': 0.0004619021212678198, 'samples': 2025024, 'steps': 10546, 'loss/train': 1.8694038391113281} 01/29/2022 02:25:13 - INFO - codeparrot_training - Step 10547: {'lr': 0.0004618934385294936, 'samples': 2025216, 'steps': 10547, 'loss/train': 2.420705795288086} 01/29/2022 02:25:17 - INFO - codeparrot_training - Step 10548: {'lr': 0.0004618847548834833, 'samples': 2025408, 'steps': 10548, 'loss/train': 1.4392016530036926} 01/29/2022 02:25:22 - INFO - codeparrot_training - Step 10549: {'lr': 0.0004618760703298258, 'samples': 2025600, 'steps': 10549, 'loss/train': 2.095106363296509} 01/29/2022 02:25:26 - INFO - codeparrot_training - Step 10550: {'lr': 0.0004618673848685586, 'samples': 2025792, 'steps': 10550, 'loss/train': 1.7009847164154053} 01/29/2022 02:25:31 - INFO - codeparrot_training - Step 10551: {'lr': 0.00046185869849971884, 'samples': 2025984, 'steps': 10551, 'loss/train': 0.8155942261219025} 01/29/2022 02:25:35 - INFO - codeparrot_training - Step 10552: {'lr': 0.0004618500112233436, 'samples': 2026176, 'steps': 10552, 'loss/train': 1.8063050508499146} 01/29/2022 02:25:39 - INFO - codeparrot_training - Step 10553: {'lr': 0.0004618413230394702, 'samples': 2026368, 'steps': 10553, 'loss/train': 1.3875306844711304} 01/29/2022 02:25:45 - INFO - codeparrot_training - Step 10554: {'lr': 0.0004618326339481359, 'samples': 2026560, 'steps': 10554, 'loss/train': 2.4537894129753113} 01/29/2022 02:25:49 - INFO - codeparrot_training - Step 10555: {'lr': 0.00046182394394937774, 'samples': 2026752, 'steps': 10555, 'loss/train': 0.9998846054077148} 01/29/2022 02:25:53 - INFO - codeparrot_training - Step 10556: {'lr': 0.00046181525304323325, 'samples': 2026944, 'steps': 10556, 'loss/train': 2.083501696586609} 01/29/2022 02:25:58 - INFO - codeparrot_training - Step 10557: {'lr': 0.0004618065612297393, 'samples': 2027136, 'steps': 10557, 'loss/train': 2.230538785457611} 01/29/2022 02:26:02 - INFO - codeparrot_training - Step 10558: {'lr': 0.00046179786850893335, 'samples': 2027328, 'steps': 10558, 'loss/train': 2.6895015835762024} 01/29/2022 02:26:07 - INFO - codeparrot_training - Step 10559: {'lr': 0.0004617891748808526, 'samples': 2027520, 'steps': 10559, 'loss/train': 1.8408147096633911} 01/29/2022 02:26:11 - INFO - codeparrot_training - Step 10560: {'lr': 0.0004617804803455343, 'samples': 2027712, 'steps': 10560, 'loss/train': 2.6397095918655396} 01/29/2022 02:26:16 - INFO - codeparrot_training - Step 10561: {'lr': 0.0004617717849030156, 'samples': 2027904, 'steps': 10561, 'loss/train': 1.9840768575668335} 01/29/2022 02:26:20 - INFO - codeparrot_training - Step 10562: {'lr': 0.00046176308855333395, 'samples': 2028096, 'steps': 10562, 'loss/train': 1.5342156887054443} 01/29/2022 02:26:24 - INFO - codeparrot_training - Step 10563: {'lr': 0.00046175439129652636, 'samples': 2028288, 'steps': 10563, 'loss/train': 1.828351378440857} 01/29/2022 02:26:29 - INFO - codeparrot_training - Step 10564: {'lr': 0.0004617456931326302, 'samples': 2028480, 'steps': 10564, 'loss/train': 1.9525489211082458} 01/29/2022 02:26:34 - INFO - codeparrot_training - Step 10565: {'lr': 0.00046173699406168277, 'samples': 2028672, 'steps': 10565, 'loss/train': 2.1642411947250366} 01/29/2022 02:26:38 - INFO - codeparrot_training - Step 10566: {'lr': 0.00046172829408372125, 'samples': 2028864, 'steps': 10566, 'loss/train': 2.0029504895210266} 01/29/2022 02:26:42 - INFO - codeparrot_training - Step 10567: {'lr': 0.000461719593198783, 'samples': 2029056, 'steps': 10567, 'loss/train': 1.4527453780174255} 01/29/2022 02:26:46 - INFO - codeparrot_training - Step 10568: {'lr': 0.0004617108914069052, 'samples': 2029248, 'steps': 10568, 'loss/train': 1.9271119236946106} 01/29/2022 02:26:53 - INFO - codeparrot_training - Step 10569: {'lr': 0.00046170218870812517, 'samples': 2029440, 'steps': 10569, 'loss/train': 2.0841005444526672} 01/29/2022 02:26:57 - INFO - codeparrot_training - Step 10570: {'lr': 0.0004616934851024802, 'samples': 2029632, 'steps': 10570, 'loss/train': 1.1271197497844696} 01/29/2022 02:27:01 - INFO - codeparrot_training - Step 10571: {'lr': 0.00046168478059000753, 'samples': 2029824, 'steps': 10571, 'loss/train': 0.7219787836074829} 01/29/2022 02:27:05 - INFO - codeparrot_training - Step 10572: {'lr': 0.0004616760751707445, 'samples': 2030016, 'steps': 10572, 'loss/train': 2.032085359096527} 01/29/2022 02:27:09 - INFO - codeparrot_training - Step 10573: {'lr': 0.0004616673688447284, 'samples': 2030208, 'steps': 10573, 'loss/train': 1.4799162447452545} 01/29/2022 02:27:15 - INFO - codeparrot_training - Step 10574: {'lr': 0.0004616586616119964, 'samples': 2030400, 'steps': 10574, 'loss/train': 1.7534072399139404} 01/29/2022 02:27:19 - INFO - codeparrot_training - Step 10575: {'lr': 0.0004616499534725861, 'samples': 2030592, 'steps': 10575, 'loss/train': 1.5634766221046448} 01/29/2022 02:27:23 - INFO - codeparrot_training - Step 10576: {'lr': 0.00046164124442653445, 'samples': 2030784, 'steps': 10576, 'loss/train': 1.9624550342559814} 01/29/2022 02:27:27 - INFO - codeparrot_training - Step 10577: {'lr': 0.00046163253447387896, 'samples': 2030976, 'steps': 10577, 'loss/train': 1.6795796155929565} 01/29/2022 02:27:32 - INFO - codeparrot_training - Step 10578: {'lr': 0.0004616238236146569, 'samples': 2031168, 'steps': 10578, 'loss/train': 1.7713351249694824} 01/29/2022 02:27:38 - INFO - codeparrot_training - Step 10579: {'lr': 0.0004616151118489056, 'samples': 2031360, 'steps': 10579, 'loss/train': 0.8225702941417694} 01/29/2022 02:27:42 - INFO - codeparrot_training - Step 10580: {'lr': 0.0004616063991766623, 'samples': 2031552, 'steps': 10580, 'loss/train': 2.0306153297424316} 01/29/2022 02:27:46 - INFO - codeparrot_training - Step 10581: {'lr': 0.00046159768559796437, 'samples': 2031744, 'steps': 10581, 'loss/train': 1.9427055716514587} 01/29/2022 02:27:51 - INFO - codeparrot_training - Step 10582: {'lr': 0.0004615889711128492, 'samples': 2031936, 'steps': 10582, 'loss/train': 1.9076609015464783} 01/29/2022 02:27:55 - INFO - codeparrot_training - Step 10583: {'lr': 0.00046158025572135404, 'samples': 2032128, 'steps': 10583, 'loss/train': 1.3167943954467773} 01/29/2022 02:28:00 - INFO - codeparrot_training - Step 10584: {'lr': 0.00046157153942351625, 'samples': 2032320, 'steps': 10584, 'loss/train': 1.56724351644516} 01/29/2022 02:28:05 - INFO - codeparrot_training - Step 10585: {'lr': 0.0004615628222193732, 'samples': 2032512, 'steps': 10585, 'loss/train': 1.913419783115387} 01/29/2022 02:28:09 - INFO - codeparrot_training - Step 10586: {'lr': 0.00046155410410896215, 'samples': 2032704, 'steps': 10586, 'loss/train': 1.7636311054229736} 01/29/2022 02:28:13 - INFO - codeparrot_training - Step 10587: {'lr': 0.00046154538509232044, 'samples': 2032896, 'steps': 10587, 'loss/train': 2.206956624984741} 01/29/2022 02:28:17 - INFO - codeparrot_training - Step 10588: {'lr': 0.00046153666516948554, 'samples': 2033088, 'steps': 10588, 'loss/train': 0.9601352512836456} 01/29/2022 02:28:24 - INFO - codeparrot_training - Step 10589: {'lr': 0.0004615279443404948, 'samples': 2033280, 'steps': 10589, 'loss/train': 1.5423350930213928} 01/29/2022 02:28:28 - INFO - codeparrot_training - Step 10590: {'lr': 0.0004615192226053855, 'samples': 2033472, 'steps': 10590, 'loss/train': 2.245752274990082} 01/29/2022 02:28:32 - INFO - codeparrot_training - Step 10591: {'lr': 0.0004615104999641949, 'samples': 2033664, 'steps': 10591, 'loss/train': 1.4288842678070068} 01/29/2022 02:28:37 - INFO - codeparrot_training - Step 10592: {'lr': 0.0004615017764169606, 'samples': 2033856, 'steps': 10592, 'loss/train': 2.0371994376182556} 01/29/2022 02:28:41 - INFO - codeparrot_training - Step 10593: {'lr': 0.0004614930519637198, 'samples': 2034048, 'steps': 10593, 'loss/train': 1.8735982775688171} 01/29/2022 02:28:45 - INFO - codeparrot_training - Step 10594: {'lr': 0.0004614843266045099, 'samples': 2034240, 'steps': 10594, 'loss/train': 1.7012681365013123} 01/29/2022 02:28:50 - INFO - codeparrot_training - Step 10595: {'lr': 0.0004614756003393683, 'samples': 2034432, 'steps': 10595, 'loss/train': 2.717039108276367} 01/29/2022 02:28:55 - INFO - codeparrot_training - Step 10596: {'lr': 0.00046146687316833235, 'samples': 2034624, 'steps': 10596, 'loss/train': 1.9825621247291565} 01/29/2022 02:28:59 - INFO - codeparrot_training - Step 10597: {'lr': 0.00046145814509143955, 'samples': 2034816, 'steps': 10597, 'loss/train': 1.919028103351593} 01/29/2022 02:29:03 - INFO - codeparrot_training - Step 10598: {'lr': 0.0004614494161087271, 'samples': 2035008, 'steps': 10598, 'loss/train': 2.3201224207878113} 01/29/2022 02:29:07 - INFO - codeparrot_training - Step 10599: {'lr': 0.00046144068622023263, 'samples': 2035200, 'steps': 10599, 'loss/train': 1.652245581150055} 01/29/2022 02:29:13 - INFO - codeparrot_training - Step 10600: {'lr': 0.00046143195542599336, 'samples': 2035392, 'steps': 10600, 'loss/train': 0.9679967164993286} 01/29/2022 02:29:18 - INFO - codeparrot_training - Step 10601: {'lr': 0.00046142322372604667, 'samples': 2035584, 'steps': 10601, 'loss/train': 1.3334515392780304} 01/29/2022 02:29:22 - INFO - codeparrot_training - Step 10602: {'lr': 0.00046141449112043, 'samples': 2035776, 'steps': 10602, 'loss/train': 2.139526605606079} 01/29/2022 02:29:26 - INFO - codeparrot_training - Step 10603: {'lr': 0.0004614057576091809, 'samples': 2035968, 'steps': 10603, 'loss/train': 1.805783450603485} 01/29/2022 02:29:30 - INFO - codeparrot_training - Step 10604: {'lr': 0.00046139702319233656, 'samples': 2036160, 'steps': 10604, 'loss/train': 1.3359942734241486} 01/29/2022 02:29:36 - INFO - codeparrot_training - Step 10605: {'lr': 0.00046138828786993456, 'samples': 2036352, 'steps': 10605, 'loss/train': 0.6962241232395172} 01/29/2022 02:29:40 - INFO - codeparrot_training - Step 10606: {'lr': 0.0004613795516420122, 'samples': 2036544, 'steps': 10606, 'loss/train': 1.6845661997795105} 01/29/2022 02:29:44 - INFO - codeparrot_training - Step 10607: {'lr': 0.000461370814508607, 'samples': 2036736, 'steps': 10607, 'loss/train': 1.6013256311416626} 01/29/2022 02:29:48 - INFO - codeparrot_training - Step 10608: {'lr': 0.00046136207646975635, 'samples': 2036928, 'steps': 10608, 'loss/train': 0.9409006834030151} 01/29/2022 02:29:52 - INFO - codeparrot_training - Step 10609: {'lr': 0.0004613533375254977, 'samples': 2037120, 'steps': 10609, 'loss/train': 1.9122297763824463} 01/29/2022 02:29:58 - INFO - codeparrot_training - Step 10610: {'lr': 0.00046134459767586847, 'samples': 2037312, 'steps': 10610, 'loss/train': 1.7953122854232788} 01/29/2022 02:30:02 - INFO - codeparrot_training - Step 10611: {'lr': 0.00046133585692090603, 'samples': 2037504, 'steps': 10611, 'loss/train': 1.9076886177062988} 01/29/2022 02:30:06 - INFO - codeparrot_training - Step 10612: {'lr': 0.0004613271152606479, 'samples': 2037696, 'steps': 10612, 'loss/train': 2.030038297176361} 01/29/2022 02:30:11 - INFO - codeparrot_training - Step 10613: {'lr': 0.00046131837269513154, 'samples': 2037888, 'steps': 10613, 'loss/train': 1.3571937382221222} 01/29/2022 02:30:15 - INFO - codeparrot_training - Step 10614: {'lr': 0.00046130962922439435, 'samples': 2038080, 'steps': 10614, 'loss/train': 2.6800553798675537} 01/29/2022 02:30:22 - INFO - codeparrot_training - Step 10615: {'lr': 0.00046130088484847383, 'samples': 2038272, 'steps': 10615, 'loss/train': 0.7535739541053772} 01/29/2022 02:30:27 - INFO - codeparrot_training - Step 10616: {'lr': 0.0004612921395674074, 'samples': 2038464, 'steps': 10616, 'loss/train': 2.0328471064567566} 01/29/2022 02:30:31 - INFO - codeparrot_training - Step 10617: {'lr': 0.00046128339338123253, 'samples': 2038656, 'steps': 10617, 'loss/train': 1.5159276723861694} 01/29/2022 02:30:35 - INFO - codeparrot_training - Step 10618: {'lr': 0.0004612746462899867, 'samples': 2038848, 'steps': 10618, 'loss/train': 2.287253201007843} 01/29/2022 02:30:39 - INFO - codeparrot_training - Step 10619: {'lr': 0.00046126589829370736, 'samples': 2039040, 'steps': 10619, 'loss/train': 1.6742924451828003} 01/29/2022 02:30:45 - INFO - codeparrot_training - Step 10620: {'lr': 0.00046125714939243204, 'samples': 2039232, 'steps': 10620, 'loss/train': 1.7503082156181335} 01/29/2022 02:30:49 - INFO - codeparrot_training - Step 10621: {'lr': 0.00046124839958619815, 'samples': 2039424, 'steps': 10621, 'loss/train': 1.2887937426567078} 01/29/2022 02:30:54 - INFO - codeparrot_training - Step 10622: {'lr': 0.0004612396488750432, 'samples': 2039616, 'steps': 10622, 'loss/train': 1.770032823085785} 01/29/2022 02:30:58 - INFO - codeparrot_training - Step 10623: {'lr': 0.00046123089725900464, 'samples': 2039808, 'steps': 10623, 'loss/train': 1.4215929508209229} 01/29/2022 02:31:02 - INFO - codeparrot_training - Step 10624: {'lr': 0.00046122214473812005, 'samples': 2040000, 'steps': 10624, 'loss/train': 1.8699982166290283} 01/29/2022 02:31:08 - INFO - codeparrot_training - Step 10625: {'lr': 0.0004612133913124268, 'samples': 2040192, 'steps': 10625, 'loss/train': 1.925185739994049} 01/29/2022 02:31:13 - INFO - codeparrot_training - Step 10626: {'lr': 0.00046120463698196245, 'samples': 2040384, 'steps': 10626, 'loss/train': 2.320303201675415} 01/29/2022 02:31:17 - INFO - codeparrot_training - Step 10627: {'lr': 0.00046119588174676454, 'samples': 2040576, 'steps': 10627, 'loss/train': 0.9765987396240234} 01/29/2022 02:31:21 - INFO - codeparrot_training - Step 10628: {'lr': 0.0004611871256068705, 'samples': 2040768, 'steps': 10628, 'loss/train': 1.5219263434410095} 01/29/2022 02:31:25 - INFO - codeparrot_training - Step 10629: {'lr': 0.0004611783685623179, 'samples': 2040960, 'steps': 10629, 'loss/train': 1.9594895839691162} 01/29/2022 02:31:31 - INFO - codeparrot_training - Step 10630: {'lr': 0.00046116961061314424, 'samples': 2041152, 'steps': 10630, 'loss/train': 1.7113391160964966} 01/29/2022 02:31:35 - INFO - codeparrot_training - Step 10631: {'lr': 0.00046116085175938694, 'samples': 2041344, 'steps': 10631, 'loss/train': 1.453393667936325} 01/29/2022 02:31:39 - INFO - codeparrot_training - Step 10632: {'lr': 0.00046115209200108366, 'samples': 2041536, 'steps': 10632, 'loss/train': 1.2969297766685486} 01/29/2022 02:31:43 - INFO - codeparrot_training - Step 10633: {'lr': 0.00046114333133827194, 'samples': 2041728, 'steps': 10633, 'loss/train': 1.426218420267105} 01/29/2022 02:31:48 - INFO - codeparrot_training - Step 10634: {'lr': 0.0004611345697709891, 'samples': 2041920, 'steps': 10634, 'loss/train': 1.175944447517395} 01/29/2022 02:31:53 - INFO - codeparrot_training - Step 10635: {'lr': 0.0004611258072992729, 'samples': 2042112, 'steps': 10635, 'loss/train': 1.3672232329845428} 01/29/2022 02:31:57 - INFO - codeparrot_training - Step 10636: {'lr': 0.0004611170439231607, 'samples': 2042304, 'steps': 10636, 'loss/train': 1.6543786525726318} 01/29/2022 02:32:02 - INFO - codeparrot_training - Step 10637: {'lr': 0.0004611082796426902, 'samples': 2042496, 'steps': 10637, 'loss/train': 2.418352425098419} 01/29/2022 02:32:06 - INFO - codeparrot_training - Step 10638: {'lr': 0.00046109951445789883, 'samples': 2042688, 'steps': 10638, 'loss/train': 1.229158490896225} 01/29/2022 02:32:10 - INFO - codeparrot_training - Step 10639: {'lr': 0.00046109074836882415, 'samples': 2042880, 'steps': 10639, 'loss/train': 1.684144377708435} 01/29/2022 02:32:17 - INFO - codeparrot_training - Step 10640: {'lr': 0.00046108198137550377, 'samples': 2043072, 'steps': 10640, 'loss/train': 1.9657065868377686} 01/29/2022 02:32:21 - INFO - codeparrot_training - Step 10641: {'lr': 0.0004610732134779752, 'samples': 2043264, 'steps': 10641, 'loss/train': 1.776064932346344} 01/29/2022 02:32:26 - INFO - codeparrot_training - Step 10642: {'lr': 0.000461064444676276, 'samples': 2043456, 'steps': 10642, 'loss/train': 1.6549115180969238} 01/29/2022 02:32:30 - INFO - codeparrot_training - Step 10643: {'lr': 0.0004610556749704438, 'samples': 2043648, 'steps': 10643, 'loss/train': 2.529043436050415} 01/29/2022 02:32:34 - INFO - codeparrot_training - Step 10644: {'lr': 0.000461046904360516, 'samples': 2043840, 'steps': 10644, 'loss/train': 2.067208707332611} 01/29/2022 02:32:39 - INFO - codeparrot_training - Step 10645: {'lr': 0.0004610381328465303, 'samples': 2044032, 'steps': 10645, 'loss/train': 1.8209928274154663} 01/29/2022 02:32:44 - INFO - codeparrot_training - Step 10646: {'lr': 0.0004610293604285243, 'samples': 2044224, 'steps': 10646, 'loss/train': 1.6294480562210083} 01/29/2022 02:32:48 - INFO - codeparrot_training - Step 10647: {'lr': 0.0004610205871065355, 'samples': 2044416, 'steps': 10647, 'loss/train': 2.272886574268341} 01/29/2022 02:32:52 - INFO - codeparrot_training - Step 10648: {'lr': 0.0004610118128806016, 'samples': 2044608, 'steps': 10648, 'loss/train': 1.454947829246521} 01/29/2022 02:32:56 - INFO - codeparrot_training - Step 10649: {'lr': 0.0004610030377507599, 'samples': 2044800, 'steps': 10649, 'loss/train': 1.8599224090576172} 01/29/2022 02:33:03 - INFO - codeparrot_training - Step 10650: {'lr': 0.0004609942617170483, 'samples': 2044992, 'steps': 10650, 'loss/train': 1.4608780145645142} 01/29/2022 02:33:08 - INFO - codeparrot_training - Step 10651: {'lr': 0.0004609854847795043, 'samples': 2045184, 'steps': 10651, 'loss/train': 1.814218282699585} 01/29/2022 02:33:12 - INFO - codeparrot_training - Step 10652: {'lr': 0.0004609767069381655, 'samples': 2045376, 'steps': 10652, 'loss/train': 1.669449806213379} 01/29/2022 02:33:16 - INFO - codeparrot_training - Step 10653: {'lr': 0.00046096792819306945, 'samples': 2045568, 'steps': 10653, 'loss/train': 2.6135571599006653} 01/29/2022 02:33:20 - INFO - codeparrot_training - Step 10654: {'lr': 0.00046095914854425376, 'samples': 2045760, 'steps': 10654, 'loss/train': 1.8478744626045227} 01/29/2022 02:33:26 - INFO - codeparrot_training - Step 10655: {'lr': 0.00046095036799175606, 'samples': 2045952, 'steps': 10655, 'loss/train': 1.5592333674430847} 01/29/2022 02:33:30 - INFO - codeparrot_training - Step 10656: {'lr': 0.000460941586535614, 'samples': 2046144, 'steps': 10656, 'loss/train': 1.9041273593902588} 01/29/2022 02:33:34 - INFO - codeparrot_training - Step 10657: {'lr': 0.00046093280417586517, 'samples': 2046336, 'steps': 10657, 'loss/train': 1.6491562128067017} 01/29/2022 02:33:38 - INFO - codeparrot_training - Step 10658: {'lr': 0.0004609240209125472, 'samples': 2046528, 'steps': 10658, 'loss/train': 2.360618233680725} 01/29/2022 02:33:43 - INFO - codeparrot_training - Step 10659: {'lr': 0.00046091523674569765, 'samples': 2046720, 'steps': 10659, 'loss/train': 2.0949084162712097} 01/29/2022 02:33:48 - INFO - codeparrot_training - Step 10660: {'lr': 0.00046090645167535415, 'samples': 2046912, 'steps': 10660, 'loss/train': 2.1969552040100098} 01/29/2022 02:33:53 - INFO - codeparrot_training - Step 10661: {'lr': 0.00046089766570155447, 'samples': 2047104, 'steps': 10661, 'loss/train': 1.4882600605487823} 01/29/2022 02:33:57 - INFO - codeparrot_training - Step 10662: {'lr': 0.0004608888788243362, 'samples': 2047296, 'steps': 10662, 'loss/train': 1.5425954461097717} 01/29/2022 02:34:01 - INFO - codeparrot_training - Step 10663: {'lr': 0.00046088009104373683, 'samples': 2047488, 'steps': 10663, 'loss/train': 2.144900321960449} 01/29/2022 02:34:05 - INFO - codeparrot_training - Step 10664: {'lr': 0.0004608713023597941, 'samples': 2047680, 'steps': 10664, 'loss/train': 1.3801867067813873} 01/29/2022 02:34:11 - INFO - codeparrot_training - Step 10665: {'lr': 0.0004608625127725458, 'samples': 2047872, 'steps': 10665, 'loss/train': 1.4517011940479279} 01/29/2022 02:34:15 - INFO - codeparrot_training - Step 10666: {'lr': 0.0004608537222820294, 'samples': 2048064, 'steps': 10666, 'loss/train': 2.2849549055099487} 01/29/2022 02:34:19 - INFO - codeparrot_training - Step 10667: {'lr': 0.0004608449308882826, 'samples': 2048256, 'steps': 10667, 'loss/train': 2.1972362995147705} 01/29/2022 02:34:23 - INFO - codeparrot_training - Step 10668: {'lr': 0.000460836138591343, 'samples': 2048448, 'steps': 10668, 'loss/train': 2.3991916179656982} 01/29/2022 02:34:28 - INFO - codeparrot_training - Step 10669: {'lr': 0.0004608273453912484, 'samples': 2048640, 'steps': 10669, 'loss/train': 2.0182188749313354} 01/29/2022 02:34:35 - INFO - codeparrot_training - Step 10670: {'lr': 0.0004608185512880364, 'samples': 2048832, 'steps': 10670, 'loss/train': 0.19229324162006378} 01/29/2022 02:34:39 - INFO - codeparrot_training - Step 10671: {'lr': 0.0004608097562817446, 'samples': 2049024, 'steps': 10671, 'loss/train': 1.9911956191062927} 01/29/2022 02:34:43 - INFO - codeparrot_training - Step 10672: {'lr': 0.0004608009603724108, 'samples': 2049216, 'steps': 10672, 'loss/train': 1.7027347683906555} 01/29/2022 02:34:47 - INFO - codeparrot_training - Step 10673: {'lr': 0.0004607921635600726, 'samples': 2049408, 'steps': 10673, 'loss/train': 1.6724740862846375} 01/29/2022 02:34:52 - INFO - codeparrot_training - Step 10674: {'lr': 0.00046078336584476777, 'samples': 2049600, 'steps': 10674, 'loss/train': 1.4776278734207153} 01/29/2022 02:34:57 - INFO - codeparrot_training - Step 10675: {'lr': 0.00046077456722653387, 'samples': 2049792, 'steps': 10675, 'loss/train': 1.2813770771026611} 01/29/2022 02:35:01 - INFO - codeparrot_training - Step 10676: {'lr': 0.00046076576770540865, 'samples': 2049984, 'steps': 10676, 'loss/train': 1.20794677734375} 01/29/2022 02:35:06 - INFO - codeparrot_training - Step 10677: {'lr': 0.00046075696728142986, 'samples': 2050176, 'steps': 10677, 'loss/train': 1.4178874790668488} 01/29/2022 02:35:10 - INFO - codeparrot_training - Step 10678: {'lr': 0.0004607481659546351, 'samples': 2050368, 'steps': 10678, 'loss/train': 2.129621386528015} 01/29/2022 02:35:14 - INFO - codeparrot_training - Step 10679: {'lr': 0.0004607393637250621, 'samples': 2050560, 'steps': 10679, 'loss/train': 1.5379005074501038} 01/29/2022 02:35:21 - INFO - codeparrot_training - Step 10680: {'lr': 0.00046073056059274867, 'samples': 2050752, 'steps': 10680, 'loss/train': 1.3351032435894012} 01/29/2022 02:35:25 - INFO - codeparrot_training - Step 10681: {'lr': 0.0004607217565577323, 'samples': 2050944, 'steps': 10681, 'loss/train': 1.4266431033611298} 01/29/2022 02:35:29 - INFO - codeparrot_training - Step 10682: {'lr': 0.0004607129516200509, 'samples': 2051136, 'steps': 10682, 'loss/train': 1.758672058582306} 01/29/2022 02:35:33 - INFO - codeparrot_training - Step 10683: {'lr': 0.00046070414577974216, 'samples': 2051328, 'steps': 10683, 'loss/train': 1.7272957563400269} 01/29/2022 02:35:38 - INFO - codeparrot_training - Step 10684: {'lr': 0.00046069533903684374, 'samples': 2051520, 'steps': 10684, 'loss/train': 0.7189895063638687} 01/29/2022 02:35:42 - INFO - codeparrot_training - Step 10685: {'lr': 0.00046068653139139337, 'samples': 2051712, 'steps': 10685, 'loss/train': 0.25677473098039627} 01/29/2022 02:35:49 - INFO - codeparrot_training - Step 10686: {'lr': 0.0004606777228434288, 'samples': 2051904, 'steps': 10686, 'loss/train': 1.5487902760505676} 01/29/2022 02:35:53 - INFO - codeparrot_training - Step 10687: {'lr': 0.00046066891339298783, 'samples': 2052096, 'steps': 10687, 'loss/train': 2.3276862502098083} 01/29/2022 02:35:58 - INFO - codeparrot_training - Step 10688: {'lr': 0.0004606601030401081, 'samples': 2052288, 'steps': 10688, 'loss/train': 1.0804195404052734} 01/29/2022 02:36:02 - INFO - codeparrot_training - Step 10689: {'lr': 0.00046065129178482733, 'samples': 2052480, 'steps': 10689, 'loss/train': 1.729480504989624} 01/29/2022 02:36:06 - INFO - codeparrot_training - Step 10690: {'lr': 0.0004606424796271834, 'samples': 2052672, 'steps': 10690, 'loss/train': 2.0665993094444275} 01/29/2022 02:36:11 - INFO - codeparrot_training - Step 10691: {'lr': 0.0004606336665672139, 'samples': 2052864, 'steps': 10691, 'loss/train': 1.7770941853523254} 01/29/2022 02:36:16 - INFO - codeparrot_training - Step 10692: {'lr': 0.00046062485260495666, 'samples': 2053056, 'steps': 10692, 'loss/train': 1.4524624943733215} 01/29/2022 02:36:20 - INFO - codeparrot_training - Step 10693: {'lr': 0.00046061603774044945, 'samples': 2053248, 'steps': 10693, 'loss/train': 1.989290177822113} 01/29/2022 02:36:24 - INFO - codeparrot_training - Step 10694: {'lr': 0.00046060722197373, 'samples': 2053440, 'steps': 10694, 'loss/train': 1.7680363655090332} 01/29/2022 02:36:28 - INFO - codeparrot_training - Step 10695: {'lr': 0.0004605984053048361, 'samples': 2053632, 'steps': 10695, 'loss/train': 2.307180941104889} 01/29/2022 02:36:36 - INFO - codeparrot_training - Step 10696: {'lr': 0.0004605895877338055, 'samples': 2053824, 'steps': 10696, 'loss/train': 1.4169869720935822} 01/29/2022 02:36:40 - INFO - codeparrot_training - Step 10697: {'lr': 0.000460580769260676, 'samples': 2054016, 'steps': 10697, 'loss/train': 1.1043691635131836} 01/29/2022 02:36:45 - INFO - codeparrot_training - Step 10698: {'lr': 0.0004605719498854853, 'samples': 2054208, 'steps': 10698, 'loss/train': 2.099345862865448} 01/29/2022 02:36:49 - INFO - codeparrot_training - Step 10699: {'lr': 0.0004605631296082713, 'samples': 2054400, 'steps': 10699, 'loss/train': 2.1227513551712036} 01/29/2022 02:36:53 - INFO - codeparrot_training - Step 10700: {'lr': 0.0004605543084290716, 'samples': 2054592, 'steps': 10700, 'loss/train': 2.130711078643799} 01/29/2022 02:36:57 - INFO - codeparrot_training - Step 10701: {'lr': 0.00046054548634792426, 'samples': 2054784, 'steps': 10701, 'loss/train': 1.4194128513336182} 01/29/2022 02:37:02 - INFO - codeparrot_training - Step 10702: {'lr': 0.0004605366633648668, 'samples': 2054976, 'steps': 10702, 'loss/train': 2.415556311607361} 01/29/2022 02:37:07 - INFO - codeparrot_training - Step 10703: {'lr': 0.00046052783947993713, 'samples': 2055168, 'steps': 10703, 'loss/train': 1.9843744039535522} 01/29/2022 02:37:11 - INFO - codeparrot_training - Step 10704: {'lr': 0.0004605190146931731, 'samples': 2055360, 'steps': 10704, 'loss/train': 2.1184317469596863} 01/29/2022 02:37:15 - INFO - codeparrot_training - Step 10705: {'lr': 0.0004605101890046124, 'samples': 2055552, 'steps': 10705, 'loss/train': 2.1005122661590576} 01/29/2022 02:37:20 - INFO - codeparrot_training - Step 10706: {'lr': 0.00046050136241429295, 'samples': 2055744, 'steps': 10706, 'loss/train': 2.461723029613495} 01/29/2022 02:37:25 - INFO - codeparrot_training - Step 10707: {'lr': 0.0004604925349222525, 'samples': 2055936, 'steps': 10707, 'loss/train': 2.9612422585487366} 01/29/2022 02:37:29 - INFO - codeparrot_training - Step 10708: {'lr': 0.00046048370652852885, 'samples': 2056128, 'steps': 10708, 'loss/train': 1.3435405790805817} 01/29/2022 02:37:33 - INFO - codeparrot_training - Step 10709: {'lr': 0.00046047487723315986, 'samples': 2056320, 'steps': 10709, 'loss/train': 2.3466108441352844} 01/29/2022 02:37:37 - INFO - codeparrot_training - Step 10710: {'lr': 0.0004604660470361832, 'samples': 2056512, 'steps': 10710, 'loss/train': 2.047964036464691} 01/29/2022 02:37:45 - INFO - codeparrot_training - Step 10711: {'lr': 0.000460457215937637, 'samples': 2056704, 'steps': 10711, 'loss/train': 1.5312527418136597} 01/29/2022 02:37:49 - INFO - codeparrot_training - Step 10712: {'lr': 0.00046044838393755885, 'samples': 2056896, 'steps': 10712, 'loss/train': 1.8207868337631226} 01/29/2022 02:37:53 - INFO - codeparrot_training - Step 10713: {'lr': 0.0004604395510359867, 'samples': 2057088, 'steps': 10713, 'loss/train': 2.1752570271492004} 01/29/2022 02:37:57 - INFO - codeparrot_training - Step 10714: {'lr': 0.0004604307172329582, 'samples': 2057280, 'steps': 10714, 'loss/train': 1.773135244846344} 01/29/2022 02:38:02 - INFO - codeparrot_training - Step 10715: {'lr': 0.0004604218825285114, 'samples': 2057472, 'steps': 10715, 'loss/train': 2.2241912484169006} 01/29/2022 02:38:07 - INFO - codeparrot_training - Step 10716: {'lr': 0.00046041304692268407, 'samples': 2057664, 'steps': 10716, 'loss/train': 1.8226202130317688} 01/29/2022 02:38:11 - INFO - codeparrot_training - Step 10717: {'lr': 0.00046040421041551404, 'samples': 2057856, 'steps': 10717, 'loss/train': 1.6285656094551086} 01/29/2022 02:38:15 - INFO - codeparrot_training - Step 10718: {'lr': 0.00046039537300703926, 'samples': 2058048, 'steps': 10718, 'loss/train': 1.6096984148025513} 01/29/2022 02:38:20 - INFO - codeparrot_training - Step 10719: {'lr': 0.00046038653469729747, 'samples': 2058240, 'steps': 10719, 'loss/train': 2.2347428798675537} 01/29/2022 02:38:24 - INFO - codeparrot_training - Step 10720: {'lr': 0.00046037769548632656, 'samples': 2058432, 'steps': 10720, 'loss/train': 1.8966025114059448} 01/29/2022 02:38:29 - INFO - codeparrot_training - Step 10721: {'lr': 0.0004603688553741644, 'samples': 2058624, 'steps': 10721, 'loss/train': 2.2543952465057373} 01/29/2022 02:38:33 - INFO - codeparrot_training - Step 10722: {'lr': 0.0004603600143608488, 'samples': 2058816, 'steps': 10722, 'loss/train': 1.9309237003326416} 01/29/2022 02:38:38 - INFO - codeparrot_training - Step 10723: {'lr': 0.00046035117244641783, 'samples': 2059008, 'steps': 10723, 'loss/train': 1.7039417624473572} 01/29/2022 02:38:42 - INFO - codeparrot_training - Step 10724: {'lr': 0.0004603423296309092, 'samples': 2059200, 'steps': 10724, 'loss/train': 1.6023947596549988} 01/29/2022 02:38:46 - INFO - codeparrot_training - Step 10725: {'lr': 0.0004603334859143608, 'samples': 2059392, 'steps': 10725, 'loss/train': 2.217836916446686} 01/29/2022 02:38:52 - INFO - codeparrot_training - Step 10726: {'lr': 0.0004603246412968105, 'samples': 2059584, 'steps': 10726, 'loss/train': 1.3240654170513153} 01/29/2022 02:38:56 - INFO - codeparrot_training - Step 10727: {'lr': 0.00046031579577829616, 'samples': 2059776, 'steps': 10727, 'loss/train': 1.8544353246688843} 01/29/2022 02:39:00 - INFO - codeparrot_training - Step 10728: {'lr': 0.00046030694935885586, 'samples': 2059968, 'steps': 10728, 'loss/train': 1.8219408988952637} 01/29/2022 02:39:04 - INFO - codeparrot_training - Step 10729: {'lr': 0.00046029810203852736, 'samples': 2060160, 'steps': 10729, 'loss/train': 0.4963051825761795} 01/29/2022 02:39:08 - INFO - codeparrot_training - Step 10730: {'lr': 0.00046028925381734855, 'samples': 2060352, 'steps': 10730, 'loss/train': 1.4289253056049347} 01/29/2022 02:39:13 - INFO - codeparrot_training - Step 10731: {'lr': 0.00046028040469535734, 'samples': 2060544, 'steps': 10731, 'loss/train': 2.619897186756134} 01/29/2022 02:39:20 - INFO - codeparrot_training - Step 10732: {'lr': 0.00046027155467259166, 'samples': 2060736, 'steps': 10732, 'loss/train': 0.7352414578199387} 01/29/2022 02:39:24 - INFO - codeparrot_training - Step 10733: {'lr': 0.00046026270374908935, 'samples': 2060928, 'steps': 10733, 'loss/train': 2.1662017107009888} 01/29/2022 02:39:28 - INFO - codeparrot_training - Step 10734: {'lr': 0.0004602538519248884, 'samples': 2061120, 'steps': 10734, 'loss/train': 1.5667294263839722} 01/29/2022 02:39:32 - INFO - codeparrot_training - Step 10735: {'lr': 0.00046024499920002676, 'samples': 2061312, 'steps': 10735, 'loss/train': 1.5815134048461914} 01/29/2022 02:39:37 - INFO - codeparrot_training - Step 10736: {'lr': 0.0004602361455745423, 'samples': 2061504, 'steps': 10736, 'loss/train': 3.533870816230774} 01/29/2022 02:39:42 - INFO - codeparrot_training - Step 10737: {'lr': 0.00046022729104847293, 'samples': 2061696, 'steps': 10737, 'loss/train': 1.6078280210494995} 01/29/2022 02:39:47 - INFO - codeparrot_training - Step 10738: {'lr': 0.0004602184356218566, 'samples': 2061888, 'steps': 10738, 'loss/train': 2.626078426837921} 01/29/2022 02:39:51 - INFO - codeparrot_training - Step 10739: {'lr': 0.0004602095792947312, 'samples': 2062080, 'steps': 10739, 'loss/train': 0.7855964004993439} 01/29/2022 02:39:55 - INFO - codeparrot_training - Step 10740: {'lr': 0.00046020072206713484, 'samples': 2062272, 'steps': 10740, 'loss/train': 1.5321303606033325} 01/29/2022 02:40:02 - INFO - codeparrot_training - Step 10741: {'lr': 0.0004601918639391052, 'samples': 2062464, 'steps': 10741, 'loss/train': 1.7986950874328613} 01/29/2022 02:40:07 - INFO - codeparrot_training - Step 10742: {'lr': 0.0004601830049106804, 'samples': 2062656, 'steps': 10742, 'loss/train': 1.6266451478004456} 01/29/2022 02:40:11 - INFO - codeparrot_training - Step 10743: {'lr': 0.0004601741449818984, 'samples': 2062848, 'steps': 10743, 'loss/train': 0.9730447232723236} 01/29/2022 02:40:15 - INFO - codeparrot_training - Step 10744: {'lr': 0.000460165284152797, 'samples': 2063040, 'steps': 10744, 'loss/train': 1.626600444316864} 01/29/2022 02:40:19 - INFO - codeparrot_training - Step 10745: {'lr': 0.0004601564224234143, 'samples': 2063232, 'steps': 10745, 'loss/train': 2.107704520225525} 01/29/2022 02:40:24 - INFO - codeparrot_training - Step 10746: {'lr': 0.00046014755979378825, 'samples': 2063424, 'steps': 10746, 'loss/train': 2.2230942249298096} 01/29/2022 02:40:29 - INFO - codeparrot_training - Step 10747: {'lr': 0.0004601386962639568, 'samples': 2063616, 'steps': 10747, 'loss/train': 1.5625927448272705} 01/29/2022 02:40:33 - INFO - codeparrot_training - Step 10748: {'lr': 0.0004601298318339578, 'samples': 2063808, 'steps': 10748, 'loss/train': 1.3531129360198975} 01/29/2022 02:40:37 - INFO - codeparrot_training - Step 10749: {'lr': 0.0004601209665038294, 'samples': 2064000, 'steps': 10749, 'loss/train': 1.619209885597229} 01/29/2022 02:40:41 - INFO - codeparrot_training - Step 10750: {'lr': 0.0004601121002736095, 'samples': 2064192, 'steps': 10750, 'loss/train': 1.117049664258957} 01/29/2022 02:40:46 - INFO - codeparrot_training - Step 10751: {'lr': 0.0004601032331433361, 'samples': 2064384, 'steps': 10751, 'loss/train': 1.8197972774505615} 01/29/2022 02:40:51 - INFO - codeparrot_training - Step 10752: {'lr': 0.00046009436511304714, 'samples': 2064576, 'steps': 10752, 'loss/train': 1.8501736521720886} 01/29/2022 02:40:55 - INFO - codeparrot_training - Step 10753: {'lr': 0.0004600854961827806, 'samples': 2064768, 'steps': 10753, 'loss/train': 1.1538449227809906} 01/29/2022 02:40:59 - INFO - codeparrot_training - Step 10754: {'lr': 0.00046007662635257453, 'samples': 2064960, 'steps': 10754, 'loss/train': 1.441622257232666} 01/29/2022 02:41:04 - INFO - codeparrot_training - Step 10755: {'lr': 0.0004600677556224669, 'samples': 2065152, 'steps': 10755, 'loss/train': 2.3218220472335815} 01/29/2022 02:41:08 - INFO - codeparrot_training - Step 10756: {'lr': 0.00046005888399249575, 'samples': 2065344, 'steps': 10756, 'loss/train': 1.9373867511749268} 01/29/2022 02:41:15 - INFO - codeparrot_training - Step 10757: {'lr': 0.000460050011462699, 'samples': 2065536, 'steps': 10757, 'loss/train': 2.2260314226150513} 01/29/2022 02:41:19 - INFO - codeparrot_training - Step 10758: {'lr': 0.0004600411380331146, 'samples': 2065728, 'steps': 10758, 'loss/train': 0.26648763567209244} 01/29/2022 02:41:24 - INFO - codeparrot_training - Step 10759: {'lr': 0.0004600322637037808, 'samples': 2065920, 'steps': 10759, 'loss/train': 1.8725073337554932} 01/29/2022 02:41:28 - INFO - codeparrot_training - Step 10760: {'lr': 0.00046002338847473545, 'samples': 2066112, 'steps': 10760, 'loss/train': 1.5974063873291016} 01/29/2022 02:41:32 - INFO - codeparrot_training - Step 10761: {'lr': 0.00046001451234601665, 'samples': 2066304, 'steps': 10761, 'loss/train': 2.071327328681946} 01/29/2022 02:41:38 - INFO - codeparrot_training - Step 10762: {'lr': 0.0004600056353176623, 'samples': 2066496, 'steps': 10762, 'loss/train': 1.5526435375213623} 01/29/2022 02:41:42 - INFO - codeparrot_training - Step 10763: {'lr': 0.00045999675738971047, 'samples': 2066688, 'steps': 10763, 'loss/train': 1.1431218981742859} 01/29/2022 02:41:46 - INFO - codeparrot_training - Step 10764: {'lr': 0.00045998787856219925, 'samples': 2066880, 'steps': 10764, 'loss/train': 1.4787788093090057} 01/29/2022 02:41:50 - INFO - codeparrot_training - Step 10765: {'lr': 0.0004599789988351666, 'samples': 2067072, 'steps': 10765, 'loss/train': 0.6735800206661224} 01/29/2022 02:41:55 - INFO - codeparrot_training - Step 10766: {'lr': 0.0004599701182086506, 'samples': 2067264, 'steps': 10766, 'loss/train': 2.393151104450226} 01/29/2022 02:42:00 - INFO - codeparrot_training - Step 10767: {'lr': 0.0004599612366826893, 'samples': 2067456, 'steps': 10767, 'loss/train': 1.371230274438858} 01/29/2022 02:42:04 - INFO - codeparrot_training - Step 10768: {'lr': 0.00045995235425732076, 'samples': 2067648, 'steps': 10768, 'loss/train': 1.5121389627456665} 01/29/2022 02:42:09 - INFO - codeparrot_training - Step 10769: {'lr': 0.00045994347093258295, 'samples': 2067840, 'steps': 10769, 'loss/train': 1.211263507604599} 01/29/2022 02:42:13 - INFO - codeparrot_training - Step 10770: {'lr': 0.00045993458670851397, 'samples': 2068032, 'steps': 10770, 'loss/train': 1.6499719619750977} 01/29/2022 02:42:17 - INFO - codeparrot_training - Step 10771: {'lr': 0.0004599257015851519, 'samples': 2068224, 'steps': 10771, 'loss/train': 1.6332907676696777} 01/29/2022 02:42:24 - INFO - codeparrot_training - Step 10772: {'lr': 0.0004599168155625348, 'samples': 2068416, 'steps': 10772, 'loss/train': 1.7747463583946228} 01/29/2022 02:42:28 - INFO - codeparrot_training - Step 10773: {'lr': 0.00045990792864070075, 'samples': 2068608, 'steps': 10773, 'loss/train': 4.085143446922302} 01/29/2022 02:42:33 - INFO - codeparrot_training - Step 10774: {'lr': 0.0004598990408196878, 'samples': 2068800, 'steps': 10774, 'loss/train': 1.0413206219673157} 01/29/2022 02:42:37 - INFO - codeparrot_training - Step 10775: {'lr': 0.00045989015209953394, 'samples': 2068992, 'steps': 10775, 'loss/train': 1.0764258205890656} 01/29/2022 02:42:41 - INFO - codeparrot_training - Step 10776: {'lr': 0.00045988126248027735, 'samples': 2069184, 'steps': 10776, 'loss/train': 1.2206481993198395} 01/29/2022 02:42:46 - INFO - codeparrot_training - Step 10777: {'lr': 0.00045987237196195603, 'samples': 2069376, 'steps': 10777, 'loss/train': 2.029295325279236} 01/29/2022 02:42:51 - INFO - codeparrot_training - Step 10778: {'lr': 0.00045986348054460815, 'samples': 2069568, 'steps': 10778, 'loss/train': 1.8465809226036072} 01/29/2022 02:42:55 - INFO - codeparrot_training - Step 10779: {'lr': 0.00045985458822827175, 'samples': 2069760, 'steps': 10779, 'loss/train': 2.0536201000213623} 01/29/2022 02:42:59 - INFO - codeparrot_training - Step 10780: {'lr': 0.0004598456950129849, 'samples': 2069952, 'steps': 10780, 'loss/train': 1.3193787038326263} 01/29/2022 02:43:03 - INFO - codeparrot_training - Step 10781: {'lr': 0.00045983680089878575, 'samples': 2070144, 'steps': 10781, 'loss/train': 1.037681668996811} 01/29/2022 02:43:09 - INFO - codeparrot_training - Step 10782: {'lr': 0.0004598279058857124, 'samples': 2070336, 'steps': 10782, 'loss/train': 1.1070585250854492} 01/29/2022 02:43:13 - INFO - codeparrot_training - Step 10783: {'lr': 0.00045981900997380296, 'samples': 2070528, 'steps': 10783, 'loss/train': 1.0226801633834839} 01/29/2022 02:43:17 - INFO - codeparrot_training - Step 10784: {'lr': 0.0004598101131630954, 'samples': 2070720, 'steps': 10784, 'loss/train': 1.4426125288009644} 01/29/2022 02:43:21 - INFO - codeparrot_training - Step 10785: {'lr': 0.00045980121545362805, 'samples': 2070912, 'steps': 10785, 'loss/train': 1.0271371006965637} 01/29/2022 02:43:26 - INFO - codeparrot_training - Step 10786: {'lr': 0.0004597923168454389, 'samples': 2071104, 'steps': 10786, 'loss/train': 1.8959199786186218} 01/29/2022 02:43:31 - INFO - codeparrot_training - Step 10787: {'lr': 0.000459783417338566, 'samples': 2071296, 'steps': 10787, 'loss/train': 1.5118085145950317} 01/29/2022 02:43:35 - INFO - codeparrot_training - Step 10788: {'lr': 0.0004597745169330476, 'samples': 2071488, 'steps': 10788, 'loss/train': 0.9018083810806274} 01/29/2022 02:43:39 - INFO - codeparrot_training - Step 10789: {'lr': 0.0004597656156289217, 'samples': 2071680, 'steps': 10789, 'loss/train': 2.069471061229706} 01/29/2022 02:43:44 - INFO - codeparrot_training - Step 10790: {'lr': 0.0004597567134262266, 'samples': 2071872, 'steps': 10790, 'loss/train': 2.434960663318634} 01/29/2022 02:43:48 - INFO - codeparrot_training - Step 10791: {'lr': 0.00045974781032500034, 'samples': 2072064, 'steps': 10791, 'loss/train': 1.2616863548755646} 01/29/2022 02:43:55 - INFO - codeparrot_training - Step 10792: {'lr': 0.00045973890632528106, 'samples': 2072256, 'steps': 10792, 'loss/train': 2.624614655971527} 01/29/2022 02:43:59 - INFO - codeparrot_training - Step 10793: {'lr': 0.00045973000142710696, 'samples': 2072448, 'steps': 10793, 'loss/train': 2.266907215118408} 01/29/2022 02:44:03 - INFO - codeparrot_training - Step 10794: {'lr': 0.000459721095630516, 'samples': 2072640, 'steps': 10794, 'loss/train': 2.01785945892334} 01/29/2022 02:44:08 - INFO - codeparrot_training - Step 10795: {'lr': 0.00045971218893554655, 'samples': 2072832, 'steps': 10795, 'loss/train': 1.9001805782318115} 01/29/2022 02:44:12 - INFO - codeparrot_training - Step 10796: {'lr': 0.0004597032813422367, 'samples': 2073024, 'steps': 10796, 'loss/train': 1.8406047821044922} 01/29/2022 02:44:17 - INFO - codeparrot_training - Step 10797: {'lr': 0.00045969437285062453, 'samples': 2073216, 'steps': 10797, 'loss/train': 1.621059000492096} 01/29/2022 02:44:21 - INFO - codeparrot_training - Step 10798: {'lr': 0.00045968546346074823, 'samples': 2073408, 'steps': 10798, 'loss/train': 2.0285369753837585} 01/29/2022 02:44:26 - INFO - codeparrot_training - Step 10799: {'lr': 0.000459676553172646, 'samples': 2073600, 'steps': 10799, 'loss/train': 2.1153650879859924} 01/29/2022 02:44:30 - INFO - codeparrot_training - Step 10800: {'lr': 0.00045966764198635603, 'samples': 2073792, 'steps': 10800, 'loss/train': 1.620656669139862} 01/29/2022 02:44:34 - INFO - codeparrot_training - Step 10801: {'lr': 0.0004596587299019164, 'samples': 2073984, 'steps': 10801, 'loss/train': 1.9496975541114807} 01/29/2022 02:44:41 - INFO - codeparrot_training - Step 10802: {'lr': 0.0004596498169193654, 'samples': 2074176, 'steps': 10802, 'loss/train': 1.846284806728363} 01/29/2022 02:44:45 - INFO - codeparrot_training - Step 10803: {'lr': 0.00045964090303874115, 'samples': 2074368, 'steps': 10803, 'loss/train': 1.4127413928508759} 01/29/2022 02:44:50 - INFO - codeparrot_training - Step 10804: {'lr': 0.0004596319882600818, 'samples': 2074560, 'steps': 10804, 'loss/train': 1.953410267829895} 01/29/2022 02:44:54 - INFO - codeparrot_training - Step 10805: {'lr': 0.00045962307258342564, 'samples': 2074752, 'steps': 10805, 'loss/train': 1.839213788509369} 01/29/2022 02:44:58 - INFO - codeparrot_training - Step 10806: {'lr': 0.00045961415600881075, 'samples': 2074944, 'steps': 10806, 'loss/train': 2.891886055469513} 01/29/2022 02:45:03 - INFO - codeparrot_training - Step 10807: {'lr': 0.0004596052385362754, 'samples': 2075136, 'steps': 10807, 'loss/train': 2.1437491178512573} 01/29/2022 02:45:08 - INFO - codeparrot_training - Step 10808: {'lr': 0.00045959632016585774, 'samples': 2075328, 'steps': 10808, 'loss/train': 1.5538533926010132} 01/29/2022 02:45:12 - INFO - codeparrot_training - Step 10809: {'lr': 0.00045958740089759606, 'samples': 2075520, 'steps': 10809, 'loss/train': 2.409689784049988} 01/29/2022 02:45:16 - INFO - codeparrot_training - Step 10810: {'lr': 0.0004595784807315284, 'samples': 2075712, 'steps': 10810, 'loss/train': 1.2921512424945831} 01/29/2022 02:45:20 - INFO - codeparrot_training - Step 10811: {'lr': 0.0004595695596676932, 'samples': 2075904, 'steps': 10811, 'loss/train': 3.1741586923599243} 01/29/2022 02:45:26 - INFO - codeparrot_training - Step 10812: {'lr': 0.00045956063770612843, 'samples': 2076096, 'steps': 10812, 'loss/train': 1.965694785118103} 01/29/2022 02:45:30 - INFO - codeparrot_training - Step 10813: {'lr': 0.00045955171484687255, 'samples': 2076288, 'steps': 10813, 'loss/train': 1.920142114162445} 01/29/2022 02:45:34 - INFO - codeparrot_training - Step 10814: {'lr': 0.0004595427910899636, 'samples': 2076480, 'steps': 10814, 'loss/train': 2.3330817818641663} 01/29/2022 02:45:38 - INFO - codeparrot_training - Step 10815: {'lr': 0.00045953386643543987, 'samples': 2076672, 'steps': 10815, 'loss/train': 2.88501513004303} 01/29/2022 02:45:43 - INFO - codeparrot_training - Step 10816: {'lr': 0.0004595249408833396, 'samples': 2076864, 'steps': 10816, 'loss/train': 0.9811631441116333} 01/29/2022 02:45:50 - INFO - codeparrot_training - Step 10817: {'lr': 0.00045951601443370107, 'samples': 2077056, 'steps': 10817, 'loss/train': 1.951640009880066} 01/29/2022 02:45:54 - INFO - codeparrot_training - Step 10818: {'lr': 0.00045950708708656236, 'samples': 2077248, 'steps': 10818, 'loss/train': 1.7401301264762878} 01/29/2022 02:45:58 - INFO - codeparrot_training - Step 10819: {'lr': 0.0004594981588419619, 'samples': 2077440, 'steps': 10819, 'loss/train': 2.233912467956543} 01/29/2022 02:46:02 - INFO - codeparrot_training - Step 10820: {'lr': 0.00045948922969993777, 'samples': 2077632, 'steps': 10820, 'loss/train': 2.0676509141921997} 01/29/2022 02:46:06 - INFO - codeparrot_training - Step 10821: {'lr': 0.00045948029966052834, 'samples': 2077824, 'steps': 10821, 'loss/train': 1.4082961678504944} 01/29/2022 02:46:12 - INFO - codeparrot_training - Step 10822: {'lr': 0.0004594713687237718, 'samples': 2078016, 'steps': 10822, 'loss/train': 1.597428023815155} 01/29/2022 02:46:16 - INFO - codeparrot_training - Step 10823: {'lr': 0.00045946243688970643, 'samples': 2078208, 'steps': 10823, 'loss/train': 0.20070601254701614} 01/29/2022 02:46:20 - INFO - codeparrot_training - Step 10824: {'lr': 0.00045945350415837056, 'samples': 2078400, 'steps': 10824, 'loss/train': 1.7285099029541016} 01/29/2022 02:46:24 - INFO - codeparrot_training - Step 10825: {'lr': 0.00045944457052980237, 'samples': 2078592, 'steps': 10825, 'loss/train': 0.9612627625465393} 01/29/2022 02:46:29 - INFO - codeparrot_training - Step 10826: {'lr': 0.0004594356360040401, 'samples': 2078784, 'steps': 10826, 'loss/train': 1.9975739121437073} 01/29/2022 02:46:36 - INFO - codeparrot_training - Step 10827: {'lr': 0.0004594267005811221, 'samples': 2078976, 'steps': 10827, 'loss/train': 2.190916657447815} 01/29/2022 02:46:40 - INFO - codeparrot_training - Step 10828: {'lr': 0.0004594177642610866, 'samples': 2079168, 'steps': 10828, 'loss/train': 1.70473712682724} 01/29/2022 02:46:44 - INFO - codeparrot_training - Step 10829: {'lr': 0.0004594088270439719, 'samples': 2079360, 'steps': 10829, 'loss/train': 2.1010292172431946} 01/29/2022 02:46:48 - INFO - codeparrot_training - Step 10830: {'lr': 0.00045939988892981624, 'samples': 2079552, 'steps': 10830, 'loss/train': 1.8817999362945557} 01/29/2022 02:46:53 - INFO - codeparrot_training - Step 10831: {'lr': 0.00045939094991865806, 'samples': 2079744, 'steps': 10831, 'loss/train': 2.647814691066742} 01/29/2022 02:46:58 - INFO - codeparrot_training - Step 10832: {'lr': 0.00045938201001053546, 'samples': 2079936, 'steps': 10832, 'loss/train': 2.1577611565589905} 01/29/2022 02:47:02 - INFO - codeparrot_training - Step 10833: {'lr': 0.00045937306920548684, 'samples': 2080128, 'steps': 10833, 'loss/train': 4.308091878890991} 01/29/2022 02:47:06 - INFO - codeparrot_training - Step 10834: {'lr': 0.0004593641275035504, 'samples': 2080320, 'steps': 10834, 'loss/train': 1.8301609754562378} 01/29/2022 02:47:11 - INFO - codeparrot_training - Step 10835: {'lr': 0.00045935518490476456, 'samples': 2080512, 'steps': 10835, 'loss/train': 2.2186737656593323} 01/29/2022 02:47:15 - INFO - codeparrot_training - Step 10836: {'lr': 0.00045934624140916763, 'samples': 2080704, 'steps': 10836, 'loss/train': 1.6024510860443115} 01/29/2022 02:47:21 - INFO - codeparrot_training - Step 10837: {'lr': 0.0004593372970167978, 'samples': 2080896, 'steps': 10837, 'loss/train': 1.6799147129058838} 01/29/2022 02:47:25 - INFO - codeparrot_training - Step 10838: {'lr': 0.0004593283517276936, 'samples': 2081088, 'steps': 10838, 'loss/train': 1.793018639087677} 01/29/2022 02:47:29 - INFO - codeparrot_training - Step 10839: {'lr': 0.0004593194055418931, 'samples': 2081280, 'steps': 10839, 'loss/train': 2.2254011034965515} 01/29/2022 02:47:33 - INFO - codeparrot_training - Step 10840: {'lr': 0.00045931045845943474, 'samples': 2081472, 'steps': 10840, 'loss/train': 2.1777098178863525} 01/29/2022 02:47:38 - INFO - codeparrot_training - Step 10841: {'lr': 0.00045930151048035684, 'samples': 2081664, 'steps': 10841, 'loss/train': 2.7179022431373596} 01/29/2022 02:47:43 - INFO - codeparrot_training - Step 10842: {'lr': 0.0004592925616046978, 'samples': 2081856, 'steps': 10842, 'loss/train': 1.8927059769630432} 01/29/2022 02:47:47 - INFO - codeparrot_training - Step 10843: {'lr': 0.0004592836118324958, 'samples': 2082048, 'steps': 10843, 'loss/train': 2.0604498982429504} 01/29/2022 02:47:51 - INFO - codeparrot_training - Step 10844: {'lr': 0.0004592746611637893, 'samples': 2082240, 'steps': 10844, 'loss/train': 1.5603125095367432} 01/29/2022 02:47:56 - INFO - codeparrot_training - Step 10845: {'lr': 0.00045926570959861656, 'samples': 2082432, 'steps': 10845, 'loss/train': 1.5439388751983643} 01/29/2022 02:48:00 - INFO - codeparrot_training - Step 10846: {'lr': 0.000459256757137016, 'samples': 2082624, 'steps': 10846, 'loss/train': 1.8630316257476807} 01/29/2022 02:48:06 - INFO - codeparrot_training - Step 10847: {'lr': 0.00045924780377902595, 'samples': 2082816, 'steps': 10847, 'loss/train': 1.693344533443451} 01/29/2022 02:48:10 - INFO - codeparrot_training - Step 10848: {'lr': 0.00045923884952468475, 'samples': 2083008, 'steps': 10848, 'loss/train': 1.953079640865326} 01/29/2022 02:48:14 - INFO - codeparrot_training - Step 10849: {'lr': 0.00045922989437403074, 'samples': 2083200, 'steps': 10849, 'loss/train': 2.400382339954376} 01/29/2022 02:48:19 - INFO - codeparrot_training - Step 10850: {'lr': 0.0004592209383271023, 'samples': 2083392, 'steps': 10850, 'loss/train': 1.3357498347759247} 01/29/2022 02:48:23 - INFO - codeparrot_training - Step 10851: {'lr': 0.0004592119813839378, 'samples': 2083584, 'steps': 10851, 'loss/train': 2.8223127722740173} 01/29/2022 02:48:28 - INFO - codeparrot_training - Step 10852: {'lr': 0.0004592030235445757, 'samples': 2083776, 'steps': 10852, 'loss/train': 2.318661153316498} 01/29/2022 02:48:32 - INFO - codeparrot_training - Step 10853: {'lr': 0.00045919406480905413, 'samples': 2083968, 'steps': 10853, 'loss/train': 2.080644428730011} 01/29/2022 02:48:37 - INFO - codeparrot_training - Step 10854: {'lr': 0.0004591851051774117, 'samples': 2084160, 'steps': 10854, 'loss/train': 1.4615268409252167} 01/29/2022 02:48:41 - INFO - codeparrot_training - Step 10855: {'lr': 0.00045917614464968665, 'samples': 2084352, 'steps': 10855, 'loss/train': 2.3466542959213257} 01/29/2022 02:48:45 - INFO - codeparrot_training - Step 10856: {'lr': 0.0004591671832259174, 'samples': 2084544, 'steps': 10856, 'loss/train': 1.2829773724079132} 01/29/2022 02:48:51 - INFO - codeparrot_training - Step 10857: {'lr': 0.00045915822090614243, 'samples': 2084736, 'steps': 10857, 'loss/train': 0.25973352044820786} 01/29/2022 02:48:55 - INFO - codeparrot_training - Step 10858: {'lr': 0.00045914925769040006, 'samples': 2084928, 'steps': 10858, 'loss/train': 2.0290303230285645} 01/29/2022 02:48:59 - INFO - codeparrot_training - Step 10859: {'lr': 0.0004591402935787287, 'samples': 2085120, 'steps': 10859, 'loss/train': 1.8933874368667603} 01/29/2022 02:49:03 - INFO - codeparrot_training - Step 10860: {'lr': 0.00045913132857116663, 'samples': 2085312, 'steps': 10860, 'loss/train': 1.7667015194892883} 01/29/2022 02:49:08 - INFO - codeparrot_training - Step 10861: {'lr': 0.00045912236266775245, 'samples': 2085504, 'steps': 10861, 'loss/train': 2.398171842098236} 01/29/2022 02:49:14 - INFO - codeparrot_training - Step 10862: {'lr': 0.0004591133958685244, 'samples': 2085696, 'steps': 10862, 'loss/train': 1.6572409272193909} 01/29/2022 02:49:18 - INFO - codeparrot_training - Step 10863: {'lr': 0.00045910442817352095, 'samples': 2085888, 'steps': 10863, 'loss/train': 1.5378824472427368} 01/29/2022 02:49:22 - INFO - codeparrot_training - Step 10864: {'lr': 0.0004590954595827806, 'samples': 2086080, 'steps': 10864, 'loss/train': 1.11887589097023} 01/29/2022 02:49:26 - INFO - codeparrot_training - Step 10865: {'lr': 0.00045908649009634165, 'samples': 2086272, 'steps': 10865, 'loss/train': 1.6532519459724426} 01/29/2022 02:49:31 - INFO - codeparrot_training - Step 10866: {'lr': 0.0004590775197142426, 'samples': 2086464, 'steps': 10866, 'loss/train': 1.3237550854682922} 01/29/2022 02:49:36 - INFO - codeparrot_training - Step 10867: {'lr': 0.0004590685484365218, 'samples': 2086656, 'steps': 10867, 'loss/train': 1.6895867586135864} 01/29/2022 02:49:40 - INFO - codeparrot_training - Step 10868: {'lr': 0.00045905957626321775, 'samples': 2086848, 'steps': 10868, 'loss/train': 1.903009057044983} 01/29/2022 02:49:44 - INFO - codeparrot_training - Step 10869: {'lr': 0.0004590506031943689, 'samples': 2087040, 'steps': 10869, 'loss/train': 1.456180214881897} 01/29/2022 02:49:49 - INFO - codeparrot_training - Step 10870: {'lr': 0.00045904162923001356, 'samples': 2087232, 'steps': 10870, 'loss/train': 1.8255334496498108} 01/29/2022 02:49:53 - INFO - codeparrot_training - Step 10871: {'lr': 0.00045903265437019036, 'samples': 2087424, 'steps': 10871, 'loss/train': 2.3717170357704163} 01/29/2022 02:49:59 - INFO - codeparrot_training - Step 10872: {'lr': 0.00045902367861493754, 'samples': 2087616, 'steps': 10872, 'loss/train': 1.3186876773834229} 01/29/2022 02:50:03 - INFO - codeparrot_training - Step 10873: {'lr': 0.00045901470196429376, 'samples': 2087808, 'steps': 10873, 'loss/train': 1.1427774131298065} 01/29/2022 02:50:08 - INFO - codeparrot_training - Step 10874: {'lr': 0.0004590057244182972, 'samples': 2088000, 'steps': 10874, 'loss/train': 2.6973043084144592} 01/29/2022 02:50:12 - INFO - codeparrot_training - Step 10875: {'lr': 0.0004589967459769867, 'samples': 2088192, 'steps': 10875, 'loss/train': 2.0799888968467712} 01/29/2022 02:50:16 - INFO - codeparrot_training - Step 10876: {'lr': 0.00045898776664040036, 'samples': 2088384, 'steps': 10876, 'loss/train': 2.447737455368042} 01/29/2022 02:50:21 - INFO - codeparrot_training - Step 10877: {'lr': 0.00045897878640857684, 'samples': 2088576, 'steps': 10877, 'loss/train': 1.976021647453308} 01/29/2022 02:50:25 - INFO - codeparrot_training - Step 10878: {'lr': 0.00045896980528155454, 'samples': 2088768, 'steps': 10878, 'loss/train': 3.723253011703491} 01/29/2022 02:50:30 - INFO - codeparrot_training - Step 10879: {'lr': 0.0004589608232593719, 'samples': 2088960, 'steps': 10879, 'loss/train': 1.5632806420326233} 01/29/2022 02:50:34 - INFO - codeparrot_training - Step 10880: {'lr': 0.0004589518403420676, 'samples': 2089152, 'steps': 10880, 'loss/train': 2.239621102809906} 01/29/2022 02:50:38 - INFO - codeparrot_training - Step 10881: {'lr': 0.0004589428565296798, 'samples': 2089344, 'steps': 10881, 'loss/train': 2.0494027733802795} 01/29/2022 02:50:43 - INFO - codeparrot_training - Step 10882: {'lr': 0.0004589338718222473, 'samples': 2089536, 'steps': 10882, 'loss/train': 2.5835021138191223} 01/29/2022 02:50:48 - INFO - codeparrot_training - Step 10883: {'lr': 0.0004589248862198083, 'samples': 2089728, 'steps': 10883, 'loss/train': 1.853062391281128} 01/29/2022 02:50:52 - INFO - codeparrot_training - Step 10884: {'lr': 0.0004589158997224015, 'samples': 2089920, 'steps': 10884, 'loss/train': 2.5314674377441406} 01/29/2022 02:50:56 - INFO - codeparrot_training - Step 10885: {'lr': 0.0004589069123300653, 'samples': 2090112, 'steps': 10885, 'loss/train': 1.5083019733428955} 01/29/2022 02:51:00 - INFO - codeparrot_training - Step 10886: {'lr': 0.0004588979240428383, 'samples': 2090304, 'steps': 10886, 'loss/train': 2.1940993666648865} 01/29/2022 02:51:06 - INFO - codeparrot_training - Step 10887: {'lr': 0.00045888893486075875, 'samples': 2090496, 'steps': 10887, 'loss/train': 2.0167717337608337} 01/29/2022 02:51:11 - INFO - codeparrot_training - Step 10888: {'lr': 0.0004588799447838655, 'samples': 2090688, 'steps': 10888, 'loss/train': 0.16052399575710297} 01/29/2022 02:51:15 - INFO - codeparrot_training - Step 10889: {'lr': 0.0004588709538121968, 'samples': 2090880, 'steps': 10889, 'loss/train': 1.8594800233840942} 01/29/2022 02:51:19 - INFO - codeparrot_training - Step 10890: {'lr': 0.00045886196194579133, 'samples': 2091072, 'steps': 10890, 'loss/train': 1.5327653288841248} 01/29/2022 02:51:23 - INFO - codeparrot_training - Step 10891: {'lr': 0.00045885296918468746, 'samples': 2091264, 'steps': 10891, 'loss/train': 1.8189380764961243} 01/29/2022 02:51:29 - INFO - codeparrot_training - Step 10892: {'lr': 0.0004588439755289238, 'samples': 2091456, 'steps': 10892, 'loss/train': 3.0448265075683594} 01/29/2022 02:51:33 - INFO - codeparrot_training - Step 10893: {'lr': 0.00045883498097853894, 'samples': 2091648, 'steps': 10893, 'loss/train': 3.1840689182281494} 01/29/2022 02:51:37 - INFO - codeparrot_training - Step 10894: {'lr': 0.00045882598553357125, 'samples': 2091840, 'steps': 10894, 'loss/train': 1.609287142753601} 01/29/2022 02:51:42 - INFO - codeparrot_training - Step 10895: {'lr': 0.00045881698919405937, 'samples': 2092032, 'steps': 10895, 'loss/train': 1.8767035603523254} 01/29/2022 02:51:46 - INFO - codeparrot_training - Step 10896: {'lr': 0.00045880799196004187, 'samples': 2092224, 'steps': 10896, 'loss/train': 2.449673295021057} 01/29/2022 02:51:51 - INFO - codeparrot_training - Step 10897: {'lr': 0.00045879899383155715, 'samples': 2092416, 'steps': 10897, 'loss/train': 2.1252776384353638} 01/29/2022 02:51:55 - INFO - codeparrot_training - Step 10898: {'lr': 0.00045878999480864386, 'samples': 2092608, 'steps': 10898, 'loss/train': 1.8448360562324524} 01/29/2022 02:52:00 - INFO - codeparrot_training - Step 10899: {'lr': 0.0004587809948913406, 'samples': 2092800, 'steps': 10899, 'loss/train': 2.2460235357284546} 01/29/2022 02:52:04 - INFO - codeparrot_training - Step 10900: {'lr': 0.00045877199407968577, 'samples': 2092992, 'steps': 10900, 'loss/train': 1.1173878014087677} 01/29/2022 02:52:08 - INFO - codeparrot_training - Step 10901: {'lr': 0.00045876299237371807, 'samples': 2093184, 'steps': 10901, 'loss/train': 1.9532471895217896} 01/29/2022 02:52:13 - INFO - codeparrot_training - Step 10902: {'lr': 0.00045875398977347596, 'samples': 2093376, 'steps': 10902, 'loss/train': 1.8231140971183777} 01/29/2022 02:52:18 - INFO - codeparrot_training - Step 10903: {'lr': 0.00045874498627899806, 'samples': 2093568, 'steps': 10903, 'loss/train': 1.7026968598365784} 01/29/2022 02:52:22 - INFO - codeparrot_training - Step 10904: {'lr': 0.00045873598189032295, 'samples': 2093760, 'steps': 10904, 'loss/train': 1.8984893560409546} 01/29/2022 02:52:26 - INFO - codeparrot_training - Step 10905: {'lr': 0.0004587269766074891, 'samples': 2093952, 'steps': 10905, 'loss/train': 1.736166536808014} 01/29/2022 02:52:30 - INFO - codeparrot_training - Step 10906: {'lr': 0.0004587179704305353, 'samples': 2094144, 'steps': 10906, 'loss/train': 0.07502766139805317} 01/29/2022 02:52:36 - INFO - codeparrot_training - Step 10907: {'lr': 0.00045870896335949987, 'samples': 2094336, 'steps': 10907, 'loss/train': 1.565970003604889} 01/29/2022 02:52:40 - INFO - codeparrot_training - Step 10908: {'lr': 0.00045869995539442153, 'samples': 2094528, 'steps': 10908, 'loss/train': 1.5962349772453308} 01/29/2022 02:52:45 - INFO - codeparrot_training - Step 10909: {'lr': 0.0004586909465353388, 'samples': 2094720, 'steps': 10909, 'loss/train': 1.3883053958415985} 01/29/2022 02:52:49 - INFO - codeparrot_training - Step 10910: {'lr': 0.0004586819367822904, 'samples': 2094912, 'steps': 10910, 'loss/train': 0.5290099382400513} 01/29/2022 02:52:53 - INFO - codeparrot_training - Step 10911: {'lr': 0.00045867292613531484, 'samples': 2095104, 'steps': 10911, 'loss/train': 2.2542566657066345} 01/29/2022 02:52:59 - INFO - codeparrot_training - Step 10912: {'lr': 0.0004586639145944508, 'samples': 2095296, 'steps': 10912, 'loss/train': 2.698245584964752} 01/29/2022 02:53:03 - INFO - codeparrot_training - Step 10913: {'lr': 0.0004586549021597367, 'samples': 2095488, 'steps': 10913, 'loss/train': 1.524863362312317} 01/29/2022 02:53:07 - INFO - codeparrot_training - Step 10914: {'lr': 0.00045864588883121125, 'samples': 2095680, 'steps': 10914, 'loss/train': 1.983938455581665} 01/29/2022 02:53:11 - INFO - codeparrot_training - Step 10915: {'lr': 0.00045863687460891313, 'samples': 2095872, 'steps': 10915, 'loss/train': 1.506875216960907} 01/29/2022 02:53:16 - INFO - codeparrot_training - Step 10916: {'lr': 0.0004586278594928808, 'samples': 2096064, 'steps': 10916, 'loss/train': 2.164627969264984} 01/29/2022 02:53:22 - INFO - codeparrot_training - Step 10917: {'lr': 0.0004586188434831531, 'samples': 2096256, 'steps': 10917, 'loss/train': 1.8503451347351074} 01/29/2022 02:53:26 - INFO - codeparrot_training - Step 10918: {'lr': 0.00045860982657976835, 'samples': 2096448, 'steps': 10918, 'loss/train': 2.187465190887451} 01/29/2022 02:53:30 - INFO - codeparrot_training - Step 10919: {'lr': 0.00045860080878276546, 'samples': 2096640, 'steps': 10919, 'loss/train': 1.5774518251419067} 01/29/2022 02:53:34 - INFO - codeparrot_training - Step 10920: {'lr': 0.0004585917900921829, 'samples': 2096832, 'steps': 10920, 'loss/train': 2.348115563392639} 01/29/2022 02:53:40 - INFO - codeparrot_training - Step 10921: {'lr': 0.0004585827705080594, 'samples': 2097024, 'steps': 10921, 'loss/train': 1.8549554944038391} 01/29/2022 02:53:44 - INFO - codeparrot_training - Step 10922: {'lr': 0.0004585737500304335, 'samples': 2097216, 'steps': 10922, 'loss/train': 1.889409363269806} 01/29/2022 02:53:48 - INFO - codeparrot_training - Step 10923: {'lr': 0.0004585647286593439, 'samples': 2097408, 'steps': 10923, 'loss/train': 1.4398806989192963} 01/29/2022 02:53:52 - INFO - codeparrot_training - Step 10924: {'lr': 0.0004585557063948292, 'samples': 2097600, 'steps': 10924, 'loss/train': 1.9374668598175049} 01/29/2022 02:53:57 - INFO - codeparrot_training - Step 10925: {'lr': 0.00045854668323692813, 'samples': 2097792, 'steps': 10925, 'loss/train': 2.0653345584869385} 01/29/2022 02:54:02 - INFO - codeparrot_training - Step 10926: {'lr': 0.00045853765918567926, 'samples': 2097984, 'steps': 10926, 'loss/train': 2.0357800126075745} 01/29/2022 02:54:06 - INFO - codeparrot_training - Step 10927: {'lr': 0.00045852863424112125, 'samples': 2098176, 'steps': 10927, 'loss/train': 2.729155719280243} 01/29/2022 02:54:10 - INFO - codeparrot_training - Step 10928: {'lr': 0.0004585196084032928, 'samples': 2098368, 'steps': 10928, 'loss/train': 2.3322152495384216} 01/29/2022 02:54:15 - INFO - codeparrot_training - Step 10929: {'lr': 0.0004585105816722326, 'samples': 2098560, 'steps': 10929, 'loss/train': 1.4977059066295624} 01/29/2022 02:54:19 - INFO - codeparrot_training - Step 10930: {'lr': 0.0004585015540479792, 'samples': 2098752, 'steps': 10930, 'loss/train': 2.5830270051956177} 01/29/2022 02:54:23 - INFO - codeparrot_training - Step 10931: {'lr': 0.00045849252553057144, 'samples': 2098944, 'steps': 10931, 'loss/train': 2.2371767163276672} 01/29/2022 02:54:29 - INFO - codeparrot_training - Step 10932: {'lr': 0.00045848349612004786, 'samples': 2099136, 'steps': 10932, 'loss/train': 1.7786146402359009} 01/29/2022 02:54:34 - INFO - codeparrot_training - Step 10933: {'lr': 0.0004584744658164472, 'samples': 2099328, 'steps': 10933, 'loss/train': 1.1464000940322876} 01/29/2022 02:54:38 - INFO - codeparrot_training - Step 10934: {'lr': 0.00045846543461980805, 'samples': 2099520, 'steps': 10934, 'loss/train': 1.432134211063385} 01/29/2022 02:54:42 - INFO - codeparrot_training - Step 10935: {'lr': 0.0004584564025301693, 'samples': 2099712, 'steps': 10935, 'loss/train': 2.2034268379211426} 01/29/2022 02:54:46 - INFO - codeparrot_training - Step 10936: {'lr': 0.00045844736954756937, 'samples': 2099904, 'steps': 10936, 'loss/train': 1.5790003538131714} 01/29/2022 02:54:52 - INFO - codeparrot_training - Step 10937: {'lr': 0.0004584383356720472, 'samples': 2100096, 'steps': 10937, 'loss/train': 1.5918050408363342} 01/29/2022 02:54:56 - INFO - codeparrot_training - Step 10938: {'lr': 0.0004584293009036414, 'samples': 2100288, 'steps': 10938, 'loss/train': 1.642221987247467} 01/29/2022 02:55:00 - INFO - codeparrot_training - Step 10939: {'lr': 0.0004584202652423906, 'samples': 2100480, 'steps': 10939, 'loss/train': 1.66189706325531} 01/29/2022 02:55:04 - INFO - codeparrot_training - Step 10940: {'lr': 0.0004584112286883336, 'samples': 2100672, 'steps': 10940, 'loss/train': 2.5362587571144104} 01/29/2022 02:55:10 - INFO - codeparrot_training - Step 10941: {'lr': 0.00045840219124150907, 'samples': 2100864, 'steps': 10941, 'loss/train': 2.1862916350364685} 01/29/2022 02:55:14 - INFO - codeparrot_training - Step 10942: {'lr': 0.0004583931529019557, 'samples': 2101056, 'steps': 10942, 'loss/train': 1.9355442523956299} 01/29/2022 02:55:18 - INFO - codeparrot_training - Step 10943: {'lr': 0.00045838411366971225, 'samples': 2101248, 'steps': 10943, 'loss/train': 1.1224725544452667} 01/29/2022 02:55:22 - INFO - codeparrot_training - Step 10944: {'lr': 0.00045837507354481744, 'samples': 2101440, 'steps': 10944, 'loss/train': 1.5591466426849365} 01/29/2022 02:55:27 - INFO - codeparrot_training - Step 10945: {'lr': 0.00045836603252731004, 'samples': 2101632, 'steps': 10945, 'loss/train': 1.6861605048179626} 01/29/2022 02:55:33 - INFO - codeparrot_training - Step 10946: {'lr': 0.0004583569906172286, 'samples': 2101824, 'steps': 10946, 'loss/train': 0.9380249083042145} 01/29/2022 02:55:37 - INFO - codeparrot_training - Step 10947: {'lr': 0.000458347947814612, 'samples': 2102016, 'steps': 10947, 'loss/train': 1.6907672882080078} 01/29/2022 02:55:41 - INFO - codeparrot_training - Step 10948: {'lr': 0.00045833890411949897, 'samples': 2102208, 'steps': 10948, 'loss/train': 1.8918070793151855} 01/29/2022 02:55:46 - INFO - codeparrot_training - Step 10949: {'lr': 0.0004583298595319283, 'samples': 2102400, 'steps': 10949, 'loss/train': 1.7877609729766846} 01/29/2022 02:55:50 - INFO - codeparrot_training - Step 10950: {'lr': 0.0004583208140519386, 'samples': 2102592, 'steps': 10950, 'loss/train': 1.8552078008651733} 01/29/2022 02:55:55 - INFO - codeparrot_training - Step 10951: {'lr': 0.00045831176767956866, 'samples': 2102784, 'steps': 10951, 'loss/train': 1.8946090936660767} 01/29/2022 02:56:00 - INFO - codeparrot_training - Step 10952: {'lr': 0.0004583027204148573, 'samples': 2102976, 'steps': 10952, 'loss/train': 1.685386598110199} 01/29/2022 02:56:04 - INFO - codeparrot_training - Step 10953: {'lr': 0.00045829367225784317, 'samples': 2103168, 'steps': 10953, 'loss/train': 1.397079885005951} 01/29/2022 02:56:08 - INFO - codeparrot_training - Step 10954: {'lr': 0.0004582846232085651, 'samples': 2103360, 'steps': 10954, 'loss/train': 0.6440107673406601} 01/29/2022 02:56:12 - INFO - codeparrot_training - Step 10955: {'lr': 0.0004582755732670619, 'samples': 2103552, 'steps': 10955, 'loss/train': 0.8695743978023529} 01/29/2022 02:56:17 - INFO - codeparrot_training - Step 10956: {'lr': 0.00045826652243337226, 'samples': 2103744, 'steps': 10956, 'loss/train': 1.3139175474643707} 01/29/2022 02:56:22 - INFO - codeparrot_training - Step 10957: {'lr': 0.0004582574707075349, 'samples': 2103936, 'steps': 10957, 'loss/train': 2.1546608805656433} 01/29/2022 02:56:26 - INFO - codeparrot_training - Step 10958: {'lr': 0.00045824841808958874, 'samples': 2104128, 'steps': 10958, 'loss/train': 1.792801022529602} 01/29/2022 02:56:31 - INFO - codeparrot_training - Step 10959: {'lr': 0.0004582393645795725, 'samples': 2104320, 'steps': 10959, 'loss/train': 0.95656818151474} 01/29/2022 02:56:35 - INFO - codeparrot_training - Step 10960: {'lr': 0.00045823031017752484, 'samples': 2104512, 'steps': 10960, 'loss/train': 2.1616535782814026} 01/29/2022 02:56:39 - INFO - codeparrot_training - Step 10961: {'lr': 0.00045822125488348474, 'samples': 2104704, 'steps': 10961, 'loss/train': 1.7157851457595825} 01/29/2022 02:56:45 - INFO - codeparrot_training - Step 10962: {'lr': 0.00045821219869749086, 'samples': 2104896, 'steps': 10962, 'loss/train': 2.7450098991394043} 01/29/2022 02:56:49 - INFO - codeparrot_training - Step 10963: {'lr': 0.00045820314161958207, 'samples': 2105088, 'steps': 10963, 'loss/train': 2.3833563923835754} 01/29/2022 02:56:54 - INFO - codeparrot_training - Step 10964: {'lr': 0.00045819408364979714, 'samples': 2105280, 'steps': 10964, 'loss/train': 1.546555995941162} 01/29/2022 02:56:58 - INFO - codeparrot_training - Step 10965: {'lr': 0.0004581850247881749, 'samples': 2105472, 'steps': 10965, 'loss/train': 2.2087910771369934} 01/29/2022 02:57:02 - INFO - codeparrot_training - Step 10966: {'lr': 0.000458175965034754, 'samples': 2105664, 'steps': 10966, 'loss/train': 1.5812425017356873} 01/29/2022 02:57:07 - INFO - codeparrot_training - Step 10967: {'lr': 0.0004581669043895734, 'samples': 2105856, 'steps': 10967, 'loss/train': 1.4856103956699371} 01/29/2022 02:57:12 - INFO - codeparrot_training - Step 10968: {'lr': 0.000458157842852672, 'samples': 2106048, 'steps': 10968, 'loss/train': 1.846884548664093} 01/29/2022 02:57:16 - INFO - codeparrot_training - Step 10969: {'lr': 0.0004581487804240884, 'samples': 2106240, 'steps': 10969, 'loss/train': 1.6262576580047607} 01/29/2022 02:57:20 - INFO - codeparrot_training - Step 10970: {'lr': 0.00045813971710386147, 'samples': 2106432, 'steps': 10970, 'loss/train': 1.9616628885269165} 01/29/2022 02:57:24 - INFO - codeparrot_training - Step 10971: {'lr': 0.0004581306528920302, 'samples': 2106624, 'steps': 10971, 'loss/train': 1.4652848839759827} 01/29/2022 02:57:30 - INFO - codeparrot_training - Step 10972: {'lr': 0.0004581215877886332, 'samples': 2106816, 'steps': 10972, 'loss/train': 1.6252613067626953} 01/29/2022 02:57:34 - INFO - codeparrot_training - Step 10973: {'lr': 0.0004581125217937095, 'samples': 2107008, 'steps': 10973, 'loss/train': 1.6935423016548157} 01/29/2022 02:57:38 - INFO - codeparrot_training - Step 10974: {'lr': 0.00045810345490729777, 'samples': 2107200, 'steps': 10974, 'loss/train': 1.7051206827163696} 01/29/2022 02:57:43 - INFO - codeparrot_training - Step 10975: {'lr': 0.00045809438712943694, 'samples': 2107392, 'steps': 10975, 'loss/train': 1.1879660189151764} 01/29/2022 02:57:47 - INFO - codeparrot_training - Step 10976: {'lr': 0.0004580853184601659, 'samples': 2107584, 'steps': 10976, 'loss/train': 1.6195515990257263} 01/29/2022 02:57:53 - INFO - codeparrot_training - Step 10977: {'lr': 0.00045807624889952336, 'samples': 2107776, 'steps': 10977, 'loss/train': 1.1813163757324219} 01/29/2022 02:57:57 - INFO - codeparrot_training - Step 10978: {'lr': 0.0004580671784475482, 'samples': 2107968, 'steps': 10978, 'loss/train': 1.7786852717399597} 01/29/2022 02:58:01 - INFO - codeparrot_training - Step 10979: {'lr': 0.0004580581071042794, 'samples': 2108160, 'steps': 10979, 'loss/train': 0.668301522731781} 01/29/2022 02:58:06 - INFO - codeparrot_training - Step 10980: {'lr': 0.00045804903486975566, 'samples': 2108352, 'steps': 10980, 'loss/train': 2.2543416023254395} 01/29/2022 02:58:10 - INFO - codeparrot_training - Step 10981: {'lr': 0.00045803996174401595, 'samples': 2108544, 'steps': 10981, 'loss/train': 2.0631385445594788} 01/29/2022 02:58:15 - INFO - codeparrot_training - Step 10982: {'lr': 0.00045803088772709914, 'samples': 2108736, 'steps': 10982, 'loss/train': 0.6436307430267334} 01/29/2022 02:58:19 - INFO - codeparrot_training - Step 10983: {'lr': 0.00045802181281904403, 'samples': 2108928, 'steps': 10983, 'loss/train': 1.7163219451904297} 01/29/2022 02:58:24 - INFO - codeparrot_training - Step 10984: {'lr': 0.00045801273701988955, 'samples': 2109120, 'steps': 10984, 'loss/train': 1.095946490764618} 01/29/2022 02:58:28 - INFO - codeparrot_training - Step 10985: {'lr': 0.0004580036603296746, 'samples': 2109312, 'steps': 10985, 'loss/train': 1.778371810913086} 01/29/2022 02:58:32 - INFO - codeparrot_training - Step 10986: {'lr': 0.00045799458274843786, 'samples': 2109504, 'steps': 10986, 'loss/train': 1.611534833908081} 01/29/2022 02:58:37 - INFO - codeparrot_training - Step 10987: {'lr': 0.0004579855042762185, 'samples': 2109696, 'steps': 10987, 'loss/train': 1.2278645038604736} 01/29/2022 02:58:42 - INFO - codeparrot_training - Step 10988: {'lr': 0.00045797642491305523, 'samples': 2109888, 'steps': 10988, 'loss/train': 1.4883065521717072} 01/29/2022 02:58:46 - INFO - codeparrot_training - Step 10989: {'lr': 0.00045796734465898705, 'samples': 2110080, 'steps': 10989, 'loss/train': 1.456531673669815} 01/29/2022 02:58:50 - INFO - codeparrot_training - Step 10990: {'lr': 0.00045795826351405276, 'samples': 2110272, 'steps': 10990, 'loss/train': 2.7866962552070618} 01/29/2022 02:58:54 - INFO - codeparrot_training - Step 10991: {'lr': 0.00045794918147829135, 'samples': 2110464, 'steps': 10991, 'loss/train': 1.8091980814933777} 01/29/2022 02:59:00 - INFO - codeparrot_training - Step 10992: {'lr': 0.00045794009855174163, 'samples': 2110656, 'steps': 10992, 'loss/train': 2.0230634808540344} 01/29/2022 02:59:05 - INFO - codeparrot_training - Step 10993: {'lr': 0.0004579310147344425, 'samples': 2110848, 'steps': 10993, 'loss/train': 3.170414686203003} 01/29/2022 02:59:09 - INFO - codeparrot_training - Step 10994: {'lr': 0.000457921930026433, 'samples': 2111040, 'steps': 10994, 'loss/train': 2.276571214199066} 01/29/2022 02:59:13 - INFO - codeparrot_training - Step 10995: {'lr': 0.00045791284442775205, 'samples': 2111232, 'steps': 10995, 'loss/train': 1.81891930103302} 01/29/2022 02:59:17 - INFO - codeparrot_training - Step 10996: {'lr': 0.0004579037579384384, 'samples': 2111424, 'steps': 10996, 'loss/train': 2.042871415615082} 01/29/2022 02:59:23 - INFO - codeparrot_training - Step 10997: {'lr': 0.00045789467055853104, 'samples': 2111616, 'steps': 10997, 'loss/train': 1.2409818470478058} 01/29/2022 02:59:27 - INFO - codeparrot_training - Step 10998: {'lr': 0.000457885582288069, 'samples': 2111808, 'steps': 10998, 'loss/train': 1.9054216146469116} 01/29/2022 02:59:31 - INFO - codeparrot_training - Step 10999: {'lr': 0.0004578764931270911, 'samples': 2112000, 'steps': 10999, 'loss/train': 0.9979706704616547} 01/29/2022 02:59:35 - INFO - codeparrot_training - Step 11000: {'lr': 0.00045786740307563633, 'samples': 2112192, 'steps': 11000, 'loss/train': 1.7420616745948792} 01/29/2022 02:59:40 - INFO - codeparrot_training - Step 11001: {'lr': 0.0004578583121337436, 'samples': 2112384, 'steps': 11001, 'loss/train': 1.691471815109253} 01/29/2022 02:59:46 - INFO - codeparrot_training - Step 11002: {'lr': 0.0004578492203014518, 'samples': 2112576, 'steps': 11002, 'loss/train': 0.8078907430171967} 01/29/2022 02:59:50 - INFO - codeparrot_training - Step 11003: {'lr': 0.00045784012757880006, 'samples': 2112768, 'steps': 11003, 'loss/train': 2.7723485827445984} 01/29/2022 02:59:54 - INFO - codeparrot_training - Step 11004: {'lr': 0.00045783103396582713, 'samples': 2112960, 'steps': 11004, 'loss/train': 1.2128231227397919} 01/29/2022 02:59:58 - INFO - codeparrot_training - Step 11005: {'lr': 0.0004578219394625721, 'samples': 2113152, 'steps': 11005, 'loss/train': 1.5595348477363586} 01/29/2022 03:00:03 - INFO - codeparrot_training - Step 11006: {'lr': 0.0004578128440690738, 'samples': 2113344, 'steps': 11006, 'loss/train': 1.8578519225120544} 01/29/2022 03:00:08 - INFO - codeparrot_training - Step 11007: {'lr': 0.00045780374778537134, 'samples': 2113536, 'steps': 11007, 'loss/train': 1.2620545327663422} 01/29/2022 03:00:12 - INFO - codeparrot_training - Step 11008: {'lr': 0.00045779465061150356, 'samples': 2113728, 'steps': 11008, 'loss/train': 1.762932300567627} 01/29/2022 03:00:16 - INFO - codeparrot_training - Step 11009: {'lr': 0.0004577855525475095, 'samples': 2113920, 'steps': 11009, 'loss/train': 1.5230228304862976} 01/29/2022 03:00:21 - INFO - codeparrot_training - Step 11010: {'lr': 0.0004577764535934281, 'samples': 2114112, 'steps': 11010, 'loss/train': 1.3759371042251587} 01/29/2022 03:00:25 - INFO - codeparrot_training - Step 11011: {'lr': 0.00045776735374929834, 'samples': 2114304, 'steps': 11011, 'loss/train': 1.3715993463993073} 01/29/2022 03:00:30 - INFO - codeparrot_training - Step 11012: {'lr': 0.00045775825301515923, 'samples': 2114496, 'steps': 11012, 'loss/train': 1.5326639413833618} 01/29/2022 03:00:34 - INFO - codeparrot_training - Step 11013: {'lr': 0.00045774915139104973, 'samples': 2114688, 'steps': 11013, 'loss/train': 1.4119288623332977} 01/29/2022 03:00:39 - INFO - codeparrot_training - Step 11014: {'lr': 0.0004577400488770088, 'samples': 2114880, 'steps': 11014, 'loss/train': 1.7609708905220032} 01/29/2022 03:00:43 - INFO - codeparrot_training - Step 11015: {'lr': 0.0004577309454730755, 'samples': 2115072, 'steps': 11015, 'loss/train': 2.0021313428878784} 01/29/2022 03:00:47 - INFO - codeparrot_training - Step 11016: {'lr': 0.00045772184117928884, 'samples': 2115264, 'steps': 11016, 'loss/train': 2.245270550251007} 01/29/2022 03:00:52 - INFO - codeparrot_training - Step 11017: {'lr': 0.00045771273599568767, 'samples': 2115456, 'steps': 11017, 'loss/train': 1.9079339504241943} 01/29/2022 03:00:57 - INFO - codeparrot_training - Step 11018: {'lr': 0.0004577036299223112, 'samples': 2115648, 'steps': 11018, 'loss/train': 1.9230127930641174} 01/29/2022 03:01:01 - INFO - codeparrot_training - Step 11019: {'lr': 0.0004576945229591982, 'samples': 2115840, 'steps': 11019, 'loss/train': 2.0322879552841187} 01/29/2022 03:01:05 - INFO - codeparrot_training - Step 11020: {'lr': 0.0004576854151063879, 'samples': 2116032, 'steps': 11020, 'loss/train': 1.3894959390163422} 01/29/2022 03:01:09 - INFO - codeparrot_training - Step 11021: {'lr': 0.0004576763063639192, 'samples': 2116224, 'steps': 11021, 'loss/train': 1.7131404876708984} 01/29/2022 03:01:15 - INFO - codeparrot_training - Step 11022: {'lr': 0.0004576671967318312, 'samples': 2116416, 'steps': 11022, 'loss/train': 1.4372143149375916} 01/29/2022 03:01:20 - INFO - codeparrot_training - Step 11023: {'lr': 0.0004576580862101628, 'samples': 2116608, 'steps': 11023, 'loss/train': 1.8508397340774536} 01/29/2022 03:01:24 - INFO - codeparrot_training - Step 11024: {'lr': 0.00045764897479895315, 'samples': 2116800, 'steps': 11024, 'loss/train': 1.8156257271766663} 01/29/2022 03:01:28 - INFO - codeparrot_training - Step 11025: {'lr': 0.00045763986249824126, 'samples': 2116992, 'steps': 11025, 'loss/train': 1.5034731030464172} 01/29/2022 03:01:32 - INFO - codeparrot_training - Step 11026: {'lr': 0.00045763074930806606, 'samples': 2117184, 'steps': 11026, 'loss/train': 2.4158318638801575} 01/29/2022 03:01:38 - INFO - codeparrot_training - Step 11027: {'lr': 0.0004576216352284667, 'samples': 2117376, 'steps': 11027, 'loss/train': 1.347691923379898} 01/29/2022 03:01:42 - INFO - codeparrot_training - Step 11028: {'lr': 0.0004576125202594822, 'samples': 2117568, 'steps': 11028, 'loss/train': 1.3419849872589111} 01/29/2022 03:01:46 - INFO - codeparrot_training - Step 11029: {'lr': 0.0004576034044011515, 'samples': 2117760, 'steps': 11029, 'loss/train': 1.0433460474014282} 01/29/2022 03:01:50 - INFO - codeparrot_training - Step 11030: {'lr': 0.00045759428765351377, 'samples': 2117952, 'steps': 11030, 'loss/train': 1.1405117511749268} 01/29/2022 03:01:55 - INFO - codeparrot_training - Step 11031: {'lr': 0.0004575851700166081, 'samples': 2118144, 'steps': 11031, 'loss/train': 1.2479178607463837} 01/29/2022 03:01:59 - INFO - codeparrot_training - Step 11032: {'lr': 0.0004575760514904734, 'samples': 2118336, 'steps': 11032, 'loss/train': 1.948680818080902} 01/29/2022 03:02:05 - INFO - codeparrot_training - Step 11033: {'lr': 0.0004575669320751489, 'samples': 2118528, 'steps': 11033, 'loss/train': 2.714608132839203} 01/29/2022 03:02:09 - INFO - codeparrot_training - Step 11034: {'lr': 0.00045755781177067345, 'samples': 2118720, 'steps': 11034, 'loss/train': 1.7472234964370728} 01/29/2022 03:02:13 - INFO - codeparrot_training - Step 11035: {'lr': 0.00045754869057708635, 'samples': 2118912, 'steps': 11035, 'loss/train': 1.7045010924339294} 01/29/2022 03:02:17 - INFO - codeparrot_training - Step 11036: {'lr': 0.00045753956849442647, 'samples': 2119104, 'steps': 11036, 'loss/train': 1.795670986175537} 01/29/2022 03:02:21 - INFO - codeparrot_training - Step 11037: {'lr': 0.00045753044552273306, 'samples': 2119296, 'steps': 11037, 'loss/train': 0.7239887416362762} 01/29/2022 03:02:27 - INFO - codeparrot_training - Step 11038: {'lr': 0.0004575213216620451, 'samples': 2119488, 'steps': 11038, 'loss/train': 1.6339808106422424} 01/29/2022 03:02:32 - INFO - codeparrot_training - Step 11039: {'lr': 0.0004575121969124016, 'samples': 2119680, 'steps': 11039, 'loss/train': 2.454552948474884} 01/29/2022 03:02:36 - INFO - codeparrot_training - Step 11040: {'lr': 0.00045750307127384186, 'samples': 2119872, 'steps': 11040, 'loss/train': 0.9548596143722534} 01/29/2022 03:02:40 - INFO - codeparrot_training - Step 11041: {'lr': 0.0004574939447464048, 'samples': 2120064, 'steps': 11041, 'loss/train': 2.150659918785095} 01/29/2022 03:02:45 - INFO - codeparrot_training - Step 11042: {'lr': 0.0004574848173301296, 'samples': 2120256, 'steps': 11042, 'loss/train': 1.4605387151241302} 01/29/2022 03:02:50 - INFO - codeparrot_training - Step 11043: {'lr': 0.0004574756890250553, 'samples': 2120448, 'steps': 11043, 'loss/train': 1.8018018007278442} 01/29/2022 03:02:54 - INFO - codeparrot_training - Step 11044: {'lr': 0.00045746655983122105, 'samples': 2120640, 'steps': 11044, 'loss/train': 1.0819123685359955} 01/29/2022 03:02:58 - INFO - codeparrot_training - Step 11045: {'lr': 0.0004574574297486659, 'samples': 2120832, 'steps': 11045, 'loss/train': 2.316555619239807} 01/29/2022 03:03:03 - INFO - codeparrot_training - Step 11046: {'lr': 0.00045744829877742907, 'samples': 2121024, 'steps': 11046, 'loss/train': 1.559184193611145} 01/29/2022 03:03:07 - INFO - codeparrot_training - Step 11047: {'lr': 0.0004574391669175495, 'samples': 2121216, 'steps': 11047, 'loss/train': 1.767267644405365} 01/29/2022 03:03:13 - INFO - codeparrot_training - Step 11048: {'lr': 0.0004574300341690665, 'samples': 2121408, 'steps': 11048, 'loss/train': 1.817686378955841} 01/29/2022 03:03:17 - INFO - codeparrot_training - Step 11049: {'lr': 0.000457420900532019, 'samples': 2121600, 'steps': 11049, 'loss/train': 1.3055469989776611} 01/29/2022 03:03:21 - INFO - codeparrot_training - Step 11050: {'lr': 0.0004574117660064463, 'samples': 2121792, 'steps': 11050, 'loss/train': 1.474993497133255} 01/29/2022 03:03:26 - INFO - codeparrot_training - Step 11051: {'lr': 0.0004574026305923875, 'samples': 2121984, 'steps': 11051, 'loss/train': 1.2967308461666107} 01/29/2022 03:03:30 - INFO - codeparrot_training - Step 11052: {'lr': 0.0004573934942898816, 'samples': 2122176, 'steps': 11052, 'loss/train': 1.424060583114624} 01/29/2022 03:03:36 - INFO - codeparrot_training - Step 11053: {'lr': 0.0004573843570989679, 'samples': 2122368, 'steps': 11053, 'loss/train': 2.135120451450348} 01/29/2022 03:03:40 - INFO - codeparrot_training - Step 11054: {'lr': 0.00045737521901968535, 'samples': 2122560, 'steps': 11054, 'loss/train': 2.283739149570465} 01/29/2022 03:03:44 - INFO - codeparrot_training - Step 11055: {'lr': 0.00045736608005207327, 'samples': 2122752, 'steps': 11055, 'loss/train': 2.0605672001838684} 01/29/2022 03:03:48 - INFO - codeparrot_training - Step 11056: {'lr': 0.0004573569401961708, 'samples': 2122944, 'steps': 11056, 'loss/train': 2.1752246618270874} 01/29/2022 03:03:53 - INFO - codeparrot_training - Step 11057: {'lr': 0.000457347799452017, 'samples': 2123136, 'steps': 11057, 'loss/train': 2.5185218453407288} 01/29/2022 03:03:58 - INFO - codeparrot_training - Step 11058: {'lr': 0.000457338657819651, 'samples': 2123328, 'steps': 11058, 'loss/train': 1.477600246667862} 01/29/2022 03:04:02 - INFO - codeparrot_training - Step 11059: {'lr': 0.00045732951529911216, 'samples': 2123520, 'steps': 11059, 'loss/train': 2.004671037197113} 01/29/2022 03:04:07 - INFO - codeparrot_training - Step 11060: {'lr': 0.0004573203718904394, 'samples': 2123712, 'steps': 11060, 'loss/train': 2.0980786085128784} 01/29/2022 03:04:11 - INFO - codeparrot_training - Step 11061: {'lr': 0.00045731122759367206, 'samples': 2123904, 'steps': 11061, 'loss/train': 1.8463459610939026} 01/29/2022 03:04:15 - INFO - codeparrot_training - Step 11062: {'lr': 0.00045730208240884926, 'samples': 2124096, 'steps': 11062, 'loss/train': 1.5288789868354797} 01/29/2022 03:04:21 - INFO - codeparrot_training - Step 11063: {'lr': 0.0004572929363360101, 'samples': 2124288, 'steps': 11063, 'loss/train': 1.2591485381126404} 01/29/2022 03:04:26 - INFO - codeparrot_training - Step 11064: {'lr': 0.0004572837893751939, 'samples': 2124480, 'steps': 11064, 'loss/train': 2.1494665145874023} 01/29/2022 03:04:30 - INFO - codeparrot_training - Step 11065: {'lr': 0.0004572746415264397, 'samples': 2124672, 'steps': 11065, 'loss/train': 1.1431968212127686} 01/29/2022 03:04:34 - INFO - codeparrot_training - Step 11066: {'lr': 0.0004572654927897868, 'samples': 2124864, 'steps': 11066, 'loss/train': 0.938539981842041} 01/29/2022 03:04:38 - INFO - codeparrot_training - Step 11067: {'lr': 0.0004572563431652743, 'samples': 2125056, 'steps': 11067, 'loss/train': 1.748281180858612} 01/29/2022 03:04:44 - INFO - codeparrot_training - Step 11068: {'lr': 0.00045724719265294143, 'samples': 2125248, 'steps': 11068, 'loss/train': 0.12429691851139069} 01/29/2022 03:04:48 - INFO - codeparrot_training - Step 11069: {'lr': 0.00045723804125282744, 'samples': 2125440, 'steps': 11069, 'loss/train': 1.9447996616363525} 01/29/2022 03:04:52 - INFO - codeparrot_training - Step 11070: {'lr': 0.0004572288889649715, 'samples': 2125632, 'steps': 11070, 'loss/train': 2.129128396511078} 01/29/2022 03:04:56 - INFO - codeparrot_training - Step 11071: {'lr': 0.00045721973578941277, 'samples': 2125824, 'steps': 11071, 'loss/train': 1.9370833039283752} 01/29/2022 03:05:01 - INFO - codeparrot_training - Step 11072: {'lr': 0.00045721058172619043, 'samples': 2126016, 'steps': 11072, 'loss/train': 1.0392116904258728} 01/29/2022 03:05:06 - INFO - codeparrot_training - Step 11073: {'lr': 0.00045720142677534387, 'samples': 2126208, 'steps': 11073, 'loss/train': 1.874682605266571} 01/29/2022 03:05:10 - INFO - codeparrot_training - Step 11074: {'lr': 0.00045719227093691216, 'samples': 2126400, 'steps': 11074, 'loss/train': 2.2069281935691833} 01/29/2022 03:05:14 - INFO - codeparrot_training - Step 11075: {'lr': 0.0004571831142109345, 'samples': 2126592, 'steps': 11075, 'loss/train': 2.547774374485016} 01/29/2022 03:05:19 - INFO - codeparrot_training - Step 11076: {'lr': 0.0004571739565974502, 'samples': 2126784, 'steps': 11076, 'loss/train': 2.1151458621025085} 01/29/2022 03:05:23 - INFO - codeparrot_training - Step 11077: {'lr': 0.0004571647980964985, 'samples': 2126976, 'steps': 11077, 'loss/train': 2.155127227306366} 01/29/2022 03:05:29 - INFO - codeparrot_training - Step 11078: {'lr': 0.0004571556387081185, 'samples': 2127168, 'steps': 11078, 'loss/train': 1.9998984932899475} 01/29/2022 03:05:33 - INFO - codeparrot_training - Step 11079: {'lr': 0.0004571464784323496, 'samples': 2127360, 'steps': 11079, 'loss/train': 1.6897868514060974} 01/29/2022 03:05:37 - INFO - codeparrot_training - Step 11080: {'lr': 0.0004571373172692309, 'samples': 2127552, 'steps': 11080, 'loss/train': 1.5816265940666199} 01/29/2022 03:05:42 - INFO - codeparrot_training - Step 11081: {'lr': 0.0004571281552188018, 'samples': 2127744, 'steps': 11081, 'loss/train': 2.3633183240890503} 01/29/2022 03:05:46 - INFO - codeparrot_training - Step 11082: {'lr': 0.0004571189922811013, 'samples': 2127936, 'steps': 11082, 'loss/train': 2.036399245262146} 01/29/2022 03:05:51 - INFO - codeparrot_training - Step 11083: {'lr': 0.00045710982845616893, 'samples': 2128128, 'steps': 11083, 'loss/train': 2.104658782482147} 01/29/2022 03:05:56 - INFO - codeparrot_training - Step 11084: {'lr': 0.0004571006637440438, 'samples': 2128320, 'steps': 11084, 'loss/train': 0.46983468532562256} 01/29/2022 03:06:00 - INFO - codeparrot_training - Step 11085: {'lr': 0.00045709149814476515, 'samples': 2128512, 'steps': 11085, 'loss/train': 1.7565901279449463} 01/29/2022 03:06:04 - INFO - codeparrot_training - Step 11086: {'lr': 0.0004570823316583723, 'samples': 2128704, 'steps': 11086, 'loss/train': 1.151732325553894} 01/29/2022 03:06:08 - INFO - codeparrot_training - Step 11087: {'lr': 0.00045707316428490453, 'samples': 2128896, 'steps': 11087, 'loss/train': 0.6338405013084412} 01/29/2022 03:06:14 - INFO - codeparrot_training - Step 11088: {'lr': 0.0004570639960244011, 'samples': 2129088, 'steps': 11088, 'loss/train': 2.286378264427185} 01/29/2022 03:06:18 - INFO - codeparrot_training - Step 11089: {'lr': 0.00045705482687690113, 'samples': 2129280, 'steps': 11089, 'loss/train': 1.9694337844848633} 01/29/2022 03:06:22 - INFO - codeparrot_training - Step 11090: {'lr': 0.00045704565684244415, 'samples': 2129472, 'steps': 11090, 'loss/train': 1.7689697742462158} 01/29/2022 03:06:27 - INFO - codeparrot_training - Step 11091: {'lr': 0.0004570364859210693, 'samples': 2129664, 'steps': 11091, 'loss/train': 1.2256093919277191} 01/29/2022 03:06:33 - INFO - codeparrot_training - Step 11092: {'lr': 0.0004570273141128158, 'samples': 2129856, 'steps': 11092, 'loss/train': 1.2145010232925415} 01/29/2022 03:06:37 - INFO - codeparrot_training - Step 11093: {'lr': 0.00045701814141772313, 'samples': 2130048, 'steps': 11093, 'loss/train': 1.7645317912101746} 01/29/2022 03:06:41 - INFO - codeparrot_training - Step 11094: {'lr': 0.0004570089678358305, 'samples': 2130240, 'steps': 11094, 'loss/train': 1.748621642589569} 01/29/2022 03:06:45 - INFO - codeparrot_training - Step 11095: {'lr': 0.000456999793367177, 'samples': 2130432, 'steps': 11095, 'loss/train': 2.0703041553497314} 01/29/2022 03:06:50 - INFO - codeparrot_training - Step 11096: {'lr': 0.0004569906180118023, 'samples': 2130624, 'steps': 11096, 'loss/train': 1.6868993639945984} 01/29/2022 03:06:55 - INFO - codeparrot_training - Step 11097: {'lr': 0.0004569814417697454, 'samples': 2130816, 'steps': 11097, 'loss/train': 1.7215631604194641} 01/29/2022 03:06:59 - INFO - codeparrot_training - Step 11098: {'lr': 0.0004569722646410458, 'samples': 2131008, 'steps': 11098, 'loss/train': 1.1454501450061798} 01/29/2022 03:07:03 - INFO - codeparrot_training - Step 11099: {'lr': 0.0004569630866257428, 'samples': 2131200, 'steps': 11099, 'loss/train': 1.2590970396995544} 01/29/2022 03:07:08 - INFO - codeparrot_training - Step 11100: {'lr': 0.00045695390772387557, 'samples': 2131392, 'steps': 11100, 'loss/train': 1.393472671508789} 01/29/2022 03:07:12 - INFO - codeparrot_training - Step 11101: {'lr': 0.00045694472793548346, 'samples': 2131584, 'steps': 11101, 'loss/train': 1.7046160697937012} 01/29/2022 03:07:17 - INFO - codeparrot_training - Step 11102: {'lr': 0.0004569355472606059, 'samples': 2131776, 'steps': 11102, 'loss/train': 1.0267439782619476} 01/29/2022 03:07:21 - INFO - codeparrot_training - Step 11103: {'lr': 0.0004569263656992822, 'samples': 2131968, 'steps': 11103, 'loss/train': 1.2468776106834412} 01/29/2022 03:07:26 - INFO - codeparrot_training - Step 11104: {'lr': 0.0004569171832515516, 'samples': 2132160, 'steps': 11104, 'loss/train': 1.6800663471221924} 01/29/2022 03:07:30 - INFO - codeparrot_training - Step 11105: {'lr': 0.0004569079999174536, 'samples': 2132352, 'steps': 11105, 'loss/train': 1.523986279964447} 01/29/2022 03:07:34 - INFO - codeparrot_training - Step 11106: {'lr': 0.0004568988156970273, 'samples': 2132544, 'steps': 11106, 'loss/train': 2.0551246404647827} 01/29/2022 03:07:40 - INFO - codeparrot_training - Step 11107: {'lr': 0.00045688963059031226, 'samples': 2132736, 'steps': 11107, 'loss/train': 1.8725046515464783} 01/29/2022 03:07:44 - INFO - codeparrot_training - Step 11108: {'lr': 0.00045688044459734766, 'samples': 2132928, 'steps': 11108, 'loss/train': 1.6863348484039307} 01/29/2022 03:07:48 - INFO - codeparrot_training - Step 11109: {'lr': 0.00045687125771817294, 'samples': 2133120, 'steps': 11109, 'loss/train': 1.6603753566741943} 01/29/2022 03:07:53 - INFO - codeparrot_training - Step 11110: {'lr': 0.00045686206995282754, 'samples': 2133312, 'steps': 11110, 'loss/train': 1.933146357536316} 01/29/2022 03:07:57 - INFO - codeparrot_training - Step 11111: {'lr': 0.00045685288130135063, 'samples': 2133504, 'steps': 11111, 'loss/train': 1.4153797030448914} 01/29/2022 03:08:02 - INFO - codeparrot_training - Step 11112: {'lr': 0.00045684369176378164, 'samples': 2133696, 'steps': 11112, 'loss/train': 1.2751400470733643} 01/29/2022 03:08:06 - INFO - codeparrot_training - Step 11113: {'lr': 0.00045683450134016, 'samples': 2133888, 'steps': 11113, 'loss/train': 2.502483308315277} 01/29/2022 03:08:11 - INFO - codeparrot_training - Step 11114: {'lr': 0.0004568253100305251, 'samples': 2134080, 'steps': 11114, 'loss/train': 1.4454304575920105} 01/29/2022 03:08:15 - INFO - codeparrot_training - Step 11115: {'lr': 0.0004568161178349161, 'samples': 2134272, 'steps': 11115, 'loss/train': 1.1408504247665405} 01/29/2022 03:08:19 - INFO - codeparrot_training - Step 11116: {'lr': 0.0004568069247533726, 'samples': 2134464, 'steps': 11116, 'loss/train': 0.12114929035305977} 01/29/2022 03:08:25 - INFO - codeparrot_training - Step 11117: {'lr': 0.0004567977307859339, 'samples': 2134656, 'steps': 11117, 'loss/train': 2.117275357246399} 01/29/2022 03:08:29 - INFO - codeparrot_training - Step 11118: {'lr': 0.0004567885359326394, 'samples': 2134848, 'steps': 11118, 'loss/train': 1.7779346108436584} 01/29/2022 03:08:34 - INFO - codeparrot_training - Step 11119: {'lr': 0.00045677934019352844, 'samples': 2135040, 'steps': 11119, 'loss/train': 2.1144150495529175} 01/29/2022 03:08:38 - INFO - codeparrot_training - Step 11120: {'lr': 0.00045677014356864043, 'samples': 2135232, 'steps': 11120, 'loss/train': 1.6604654788970947} 01/29/2022 03:08:42 - INFO - codeparrot_training - Step 11121: {'lr': 0.00045676094605801487, 'samples': 2135424, 'steps': 11121, 'loss/train': 1.1804867684841156} 01/29/2022 03:08:47 - INFO - codeparrot_training - Step 11122: {'lr': 0.00045675174766169105, 'samples': 2135616, 'steps': 11122, 'loss/train': 1.7497807145118713} 01/29/2022 03:08:52 - INFO - codeparrot_training - Step 11123: {'lr': 0.0004567425483797083, 'samples': 2135808, 'steps': 11123, 'loss/train': 1.6133385300636292} 01/29/2022 03:08:56 - INFO - codeparrot_training - Step 11124: {'lr': 0.0004567333482121062, 'samples': 2136000, 'steps': 11124, 'loss/train': 1.7726579904556274} 01/29/2022 03:09:00 - INFO - codeparrot_training - Step 11125: {'lr': 0.0004567241471589241, 'samples': 2136192, 'steps': 11125, 'loss/train': 1.2557094097137451} 01/29/2022 03:09:04 - INFO - codeparrot_training - Step 11126: {'lr': 0.0004567149452202013, 'samples': 2136384, 'steps': 11126, 'loss/train': 1.4891510903835297} 01/29/2022 03:09:10 - INFO - codeparrot_training - Step 11127: {'lr': 0.0004567057423959774, 'samples': 2136576, 'steps': 11127, 'loss/train': 1.7819965481758118} 01/29/2022 03:09:14 - INFO - codeparrot_training - Step 11128: {'lr': 0.0004566965386862917, 'samples': 2136768, 'steps': 11128, 'loss/train': 1.811652660369873} 01/29/2022 03:09:18 - INFO - codeparrot_training - Step 11129: {'lr': 0.0004566873340911837, 'samples': 2136960, 'steps': 11129, 'loss/train': 2.215536117553711} 01/29/2022 03:09:23 - INFO - codeparrot_training - Step 11130: {'lr': 0.00045667812861069275, 'samples': 2137152, 'steps': 11130, 'loss/train': 1.9181308150291443} 01/29/2022 03:09:27 - INFO - codeparrot_training - Step 11131: {'lr': 0.00045666892224485836, 'samples': 2137344, 'steps': 11131, 'loss/train': 1.522739052772522} 01/29/2022 03:09:31 - INFO - codeparrot_training - Step 11132: {'lr': 0.0004566597149937199, 'samples': 2137536, 'steps': 11132, 'loss/train': 1.057170957326889} 01/29/2022 03:09:37 - INFO - codeparrot_training - Step 11133: {'lr': 0.0004566505068573168, 'samples': 2137728, 'steps': 11133, 'loss/train': 1.7892610430717468} 01/29/2022 03:09:41 - INFO - codeparrot_training - Step 11134: {'lr': 0.00045664129783568866, 'samples': 2137920, 'steps': 11134, 'loss/train': 1.3491998612880707} 01/29/2022 03:09:45 - INFO - codeparrot_training - Step 11135: {'lr': 0.00045663208792887474, 'samples': 2138112, 'steps': 11135, 'loss/train': 1.9424035549163818} 01/29/2022 03:09:49 - INFO - codeparrot_training - Step 11136: {'lr': 0.0004566228771369146, 'samples': 2138304, 'steps': 11136, 'loss/train': 2.0339836478233337} 01/29/2022 03:09:54 - INFO - codeparrot_training - Step 11137: {'lr': 0.00045661366545984763, 'samples': 2138496, 'steps': 11137, 'loss/train': 1.5895441174507141} 01/29/2022 03:09:59 - INFO - codeparrot_training - Step 11138: {'lr': 0.00045660445289771336, 'samples': 2138688, 'steps': 11138, 'loss/train': 0.13281706720590591} 01/29/2022 03:10:04 - INFO - codeparrot_training - Step 11139: {'lr': 0.00045659523945055114, 'samples': 2138880, 'steps': 11139, 'loss/train': 1.152145117521286} 01/29/2022 03:10:08 - INFO - codeparrot_training - Step 11140: {'lr': 0.0004565860251184006, 'samples': 2139072, 'steps': 11140, 'loss/train': 1.3606326878070831} 01/29/2022 03:10:12 - INFO - codeparrot_training - Step 11141: {'lr': 0.0004565768099013011, 'samples': 2139264, 'steps': 11141, 'loss/train': 1.930014431476593} 01/29/2022 03:10:16 - INFO - codeparrot_training - Step 11142: {'lr': 0.00045656759379929213, 'samples': 2139456, 'steps': 11142, 'loss/train': 2.282250702381134} 01/29/2022 03:10:22 - INFO - codeparrot_training - Step 11143: {'lr': 0.0004565583768124132, 'samples': 2139648, 'steps': 11143, 'loss/train': 1.4044064283370972} 01/29/2022 03:10:26 - INFO - codeparrot_training - Step 11144: {'lr': 0.0004565491589407038, 'samples': 2139840, 'steps': 11144, 'loss/train': 1.340008020401001} 01/29/2022 03:10:30 - INFO - codeparrot_training - Step 11145: {'lr': 0.0004565399401842034, 'samples': 2140032, 'steps': 11145, 'loss/train': 1.336439698934555} 01/29/2022 03:10:35 - INFO - codeparrot_training - Step 11146: {'lr': 0.0004565307205429514, 'samples': 2140224, 'steps': 11146, 'loss/train': 1.1353472471237183} 01/29/2022 03:10:39 - INFO - codeparrot_training - Step 11147: {'lr': 0.00045652150001698744, 'samples': 2140416, 'steps': 11147, 'loss/train': 2.016536235809326} 01/29/2022 03:10:44 - INFO - codeparrot_training - Step 11148: {'lr': 0.00045651227860635094, 'samples': 2140608, 'steps': 11148, 'loss/train': 1.508321464061737} 01/29/2022 03:10:49 - INFO - codeparrot_training - Step 11149: {'lr': 0.00045650305631108137, 'samples': 2140800, 'steps': 11149, 'loss/train': 1.7040234804153442} 01/29/2022 03:10:53 - INFO - codeparrot_training - Step 11150: {'lr': 0.0004564938331312183, 'samples': 2140992, 'steps': 11150, 'loss/train': 1.817097544670105} 01/29/2022 03:10:57 - INFO - codeparrot_training - Step 11151: {'lr': 0.00045648460906680123, 'samples': 2141184, 'steps': 11151, 'loss/train': 1.7602390050888062} 01/29/2022 03:11:03 - INFO - codeparrot_training - Step 11152: {'lr': 0.00045647538411786964, 'samples': 2141376, 'steps': 11152, 'loss/train': 1.9547064900398254} 01/29/2022 03:11:08 - INFO - codeparrot_training - Step 11153: {'lr': 0.00045646615828446316, 'samples': 2141568, 'steps': 11153, 'loss/train': 1.5332595705986023} 01/29/2022 03:11:12 - INFO - codeparrot_training - Step 11154: {'lr': 0.00045645693156662104, 'samples': 2141760, 'steps': 11154, 'loss/train': 2.085973083972931} 01/29/2022 03:11:16 - INFO - codeparrot_training - Step 11155: {'lr': 0.0004564477039643831, 'samples': 2141952, 'steps': 11155, 'loss/train': 2.1560028791427612} 01/29/2022 03:11:21 - INFO - codeparrot_training - Step 11156: {'lr': 0.0004564384754777888, 'samples': 2142144, 'steps': 11156, 'loss/train': 2.194208085536957} 01/29/2022 03:11:26 - INFO - codeparrot_training - Step 11157: {'lr': 0.0004564292461068775, 'samples': 2142336, 'steps': 11157, 'loss/train': 0.7575746476650238} 01/29/2022 03:11:30 - INFO - codeparrot_training - Step 11158: {'lr': 0.00045642001585168896, 'samples': 2142528, 'steps': 11158, 'loss/train': 1.6484102010726929} 01/29/2022 03:11:34 - INFO - codeparrot_training - Step 11159: {'lr': 0.0004564107847122626, 'samples': 2142720, 'steps': 11159, 'loss/train': 2.287982940673828} 01/29/2022 03:11:39 - INFO - codeparrot_training - Step 11160: {'lr': 0.0004564015526886379, 'samples': 2142912, 'steps': 11160, 'loss/train': 1.73765230178833} 01/29/2022 03:11:43 - INFO - codeparrot_training - Step 11161: {'lr': 0.0004563923197808546, 'samples': 2143104, 'steps': 11161, 'loss/train': 0.48386116325855255} 01/29/2022 03:11:48 - INFO - codeparrot_training - Step 11162: {'lr': 0.00045638308598895205, 'samples': 2143296, 'steps': 11162, 'loss/train': 0.6383148729801178} 01/29/2022 03:11:52 - INFO - codeparrot_training - Step 11163: {'lr': 0.0004563738513129699, 'samples': 2143488, 'steps': 11163, 'loss/train': 0.9186467528343201} 01/29/2022 03:11:57 - INFO - codeparrot_training - Step 11164: {'lr': 0.0004563646157529477, 'samples': 2143680, 'steps': 11164, 'loss/train': 1.5494300723075867} 01/29/2022 03:12:01 - INFO - codeparrot_training - Step 11165: {'lr': 0.0004563553793089251, 'samples': 2143872, 'steps': 11165, 'loss/train': 0.5763495415449142} 01/29/2022 03:12:05 - INFO - codeparrot_training - Step 11166: {'lr': 0.00045634614198094154, 'samples': 2144064, 'steps': 11166, 'loss/train': 1.5352238416671753} 01/29/2022 03:12:11 - INFO - codeparrot_training - Step 11167: {'lr': 0.0004563369037690366, 'samples': 2144256, 'steps': 11167, 'loss/train': 2.292550027370453} 01/29/2022 03:12:15 - INFO - codeparrot_training - Step 11168: {'lr': 0.00045632766467324995, 'samples': 2144448, 'steps': 11168, 'loss/train': 1.8703635334968567} 01/29/2022 03:12:19 - INFO - codeparrot_training - Step 11169: {'lr': 0.00045631842469362103, 'samples': 2144640, 'steps': 11169, 'loss/train': 1.2928529977798462} 01/29/2022 03:12:24 - INFO - codeparrot_training - Step 11170: {'lr': 0.00045630918383018947, 'samples': 2144832, 'steps': 11170, 'loss/train': 0.5305692851543427} 01/29/2022 03:12:28 - INFO - codeparrot_training - Step 11171: {'lr': 0.00045629994208299496, 'samples': 2145024, 'steps': 11171, 'loss/train': 1.8851512670516968} 01/29/2022 03:12:33 - INFO - codeparrot_training - Step 11172: {'lr': 0.0004562906994520769, 'samples': 2145216, 'steps': 11172, 'loss/train': 2.0268187522888184} 01/29/2022 03:12:37 - INFO - codeparrot_training - Step 11173: {'lr': 0.0004562814559374751, 'samples': 2145408, 'steps': 11173, 'loss/train': 1.5824423432350159} 01/29/2022 03:12:42 - INFO - codeparrot_training - Step 11174: {'lr': 0.000456272211539229, 'samples': 2145600, 'steps': 11174, 'loss/train': 0.7721772193908691} 01/29/2022 03:12:46 - INFO - codeparrot_training - Step 11175: {'lr': 0.00045626296625737823, 'samples': 2145792, 'steps': 11175, 'loss/train': 1.4476226270198822} 01/29/2022 03:12:50 - INFO - codeparrot_training - Step 11176: {'lr': 0.0004562537200919624, 'samples': 2145984, 'steps': 11176, 'loss/train': 1.9713029265403748} 01/29/2022 03:12:56 - INFO - codeparrot_training - Step 11177: {'lr': 0.00045624447304302117, 'samples': 2146176, 'steps': 11177, 'loss/train': 2.3299525380134583} 01/29/2022 03:13:01 - INFO - codeparrot_training - Step 11178: {'lr': 0.00045623522511059405, 'samples': 2146368, 'steps': 11178, 'loss/train': 2.1456053853034973} 01/29/2022 03:13:05 - INFO - codeparrot_training - Step 11179: {'lr': 0.00045622597629472073, 'samples': 2146560, 'steps': 11179, 'loss/train': 3.5822081565856934} 01/29/2022 03:13:09 - INFO - codeparrot_training - Step 11180: {'lr': 0.0004562167265954409, 'samples': 2146752, 'steps': 11180, 'loss/train': 2.1465858221054077} 01/29/2022 03:13:13 - INFO - codeparrot_training - Step 11181: {'lr': 0.000456207476012794, 'samples': 2146944, 'steps': 11181, 'loss/train': 1.8461764454841614} 01/29/2022 03:13:19 - INFO - codeparrot_training - Step 11182: {'lr': 0.0004561982245468198, 'samples': 2147136, 'steps': 11182, 'loss/train': 1.3520475625991821} 01/29/2022 03:13:23 - INFO - codeparrot_training - Step 11183: {'lr': 0.0004561889721975578, 'samples': 2147328, 'steps': 11183, 'loss/train': 1.600852131843567} 01/29/2022 03:13:27 - INFO - codeparrot_training - Step 11184: {'lr': 0.0004561797189650478, 'samples': 2147520, 'steps': 11184, 'loss/train': 2.1536117792129517} 01/29/2022 03:13:31 - INFO - codeparrot_training - Step 11185: {'lr': 0.0004561704648493293, 'samples': 2147712, 'steps': 11185, 'loss/train': 2.234613597393036} 01/29/2022 03:13:36 - INFO - codeparrot_training - Step 11186: {'lr': 0.00045616120985044205, 'samples': 2147904, 'steps': 11186, 'loss/train': 2.0682923197746277} 01/29/2022 03:13:41 - INFO - codeparrot_training - Step 11187: {'lr': 0.0004561519539684256, 'samples': 2148096, 'steps': 11187, 'loss/train': 1.7383296489715576} 01/29/2022 03:13:45 - INFO - codeparrot_training - Step 11188: {'lr': 0.00045614269720331964, 'samples': 2148288, 'steps': 11188, 'loss/train': 1.7797476053237915} 01/29/2022 03:13:49 - INFO - codeparrot_training - Step 11189: {'lr': 0.00045613343955516386, 'samples': 2148480, 'steps': 11189, 'loss/train': 2.296999990940094} 01/29/2022 03:13:54 - INFO - codeparrot_training - Step 11190: {'lr': 0.00045612418102399785, 'samples': 2148672, 'steps': 11190, 'loss/train': 1.7317328453063965} 01/29/2022 03:13:58 - INFO - codeparrot_training - Step 11191: {'lr': 0.00045611492160986127, 'samples': 2148864, 'steps': 11191, 'loss/train': 1.8519614338874817} 01/29/2022 03:14:03 - INFO - codeparrot_training - Step 11192: {'lr': 0.00045610566131279386, 'samples': 2149056, 'steps': 11192, 'loss/train': 1.8222330808639526} 01/29/2022 03:14:08 - INFO - codeparrot_training - Step 11193: {'lr': 0.00045609640013283525, 'samples': 2149248, 'steps': 11193, 'loss/train': 2.0169387459754944} 01/29/2022 03:14:12 - INFO - codeparrot_training - Step 11194: {'lr': 0.00045608713807002507, 'samples': 2149440, 'steps': 11194, 'loss/train': 1.2898186147212982} 01/29/2022 03:14:16 - INFO - codeparrot_training - Step 11195: {'lr': 0.000456077875124403, 'samples': 2149632, 'steps': 11195, 'loss/train': 1.1366158425807953} 01/29/2022 03:14:20 - INFO - codeparrot_training - Step 11196: {'lr': 0.00045606861129600883, 'samples': 2149824, 'steps': 11196, 'loss/train': 1.6596277356147766} 01/29/2022 03:14:27 - INFO - codeparrot_training - Step 11197: {'lr': 0.00045605934658488214, 'samples': 2150016, 'steps': 11197, 'loss/train': 2.415482997894287} 01/29/2022 03:14:31 - INFO - codeparrot_training - Step 11198: {'lr': 0.0004560500809910626, 'samples': 2150208, 'steps': 11198, 'loss/train': 1.7361186146736145} 01/29/2022 03:14:35 - INFO - codeparrot_training - Step 11199: {'lr': 0.00045604081451459, 'samples': 2150400, 'steps': 11199, 'loss/train': 2.021891176700592} 01/29/2022 03:14:39 - INFO - codeparrot_training - Step 11200: {'lr': 0.0004560315471555039, 'samples': 2150592, 'steps': 11200, 'loss/train': 1.6323593258857727} 01/29/2022 03:14:44 - INFO - codeparrot_training - Step 11201: {'lr': 0.00045602227891384416, 'samples': 2150784, 'steps': 11201, 'loss/train': 1.464230865240097} 01/29/2022 03:14:49 - INFO - codeparrot_training - Step 11202: {'lr': 0.00045601300978965033, 'samples': 2150976, 'steps': 11202, 'loss/train': 0.6984920650720596} 01/29/2022 03:14:53 - INFO - codeparrot_training - Step 11203: {'lr': 0.00045600373978296223, 'samples': 2151168, 'steps': 11203, 'loss/train': 1.4695006012916565} 01/29/2022 03:14:57 - INFO - codeparrot_training - Step 11204: {'lr': 0.0004559944688938195, 'samples': 2151360, 'steps': 11204, 'loss/train': 1.2833400070667267} 01/29/2022 03:15:02 - INFO - codeparrot_training - Step 11205: {'lr': 0.0004559851971222618, 'samples': 2151552, 'steps': 11205, 'loss/train': 1.7192916870117188} 01/29/2022 03:15:06 - INFO - codeparrot_training - Step 11206: {'lr': 0.00045597592446832905, 'samples': 2151744, 'steps': 11206, 'loss/train': 1.6006845831871033} 01/29/2022 03:15:11 - INFO - codeparrot_training - Step 11207: {'lr': 0.0004559666509320608, 'samples': 2151936, 'steps': 11207, 'loss/train': 2.249397039413452} 01/29/2022 03:15:15 - INFO - codeparrot_training - Step 11208: {'lr': 0.0004559573765134968, 'samples': 2152128, 'steps': 11208, 'loss/train': 1.3004521429538727} 01/29/2022 03:15:19 - INFO - codeparrot_training - Step 11209: {'lr': 0.0004559481012126768, 'samples': 2152320, 'steps': 11209, 'loss/train': 1.94656902551651} 01/29/2022 03:15:24 - INFO - codeparrot_training - Step 11210: {'lr': 0.00045593882502964055, 'samples': 2152512, 'steps': 11210, 'loss/train': 2.272025227546692} 01/29/2022 03:15:28 - INFO - codeparrot_training - Step 11211: {'lr': 0.00045592954796442784, 'samples': 2152704, 'steps': 11211, 'loss/train': 1.9431601166725159} 01/29/2022 03:15:34 - INFO - codeparrot_training - Step 11212: {'lr': 0.0004559202700170782, 'samples': 2152896, 'steps': 11212, 'loss/train': 1.1363132894039154} 01/29/2022 03:15:38 - INFO - codeparrot_training - Step 11213: {'lr': 0.00045591099118763156, 'samples': 2153088, 'steps': 11213, 'loss/train': 1.3616143763065338} 01/29/2022 03:15:42 - INFO - codeparrot_training - Step 11214: {'lr': 0.0004559017114761276, 'samples': 2153280, 'steps': 11214, 'loss/train': 1.9101607203483582} 01/29/2022 03:15:47 - INFO - codeparrot_training - Step 11215: {'lr': 0.00045589243088260613, 'samples': 2153472, 'steps': 11215, 'loss/train': 1.9273847937583923} 01/29/2022 03:15:51 - INFO - codeparrot_training - Step 11216: {'lr': 0.00045588314940710683, 'samples': 2153664, 'steps': 11216, 'loss/train': 1.964722216129303} 01/29/2022 03:15:56 - INFO - codeparrot_training - Step 11217: {'lr': 0.00045587386704966956, 'samples': 2153856, 'steps': 11217, 'loss/train': 2.597435474395752} 01/29/2022 03:16:00 - INFO - codeparrot_training - Step 11218: {'lr': 0.00045586458381033395, 'samples': 2154048, 'steps': 11218, 'loss/train': 1.7260234951972961} 01/29/2022 03:16:05 - INFO - codeparrot_training - Step 11219: {'lr': 0.00045585529968913984, 'samples': 2154240, 'steps': 11219, 'loss/train': 1.730189859867096} 01/29/2022 03:16:09 - INFO - codeparrot_training - Step 11220: {'lr': 0.00045584601468612703, 'samples': 2154432, 'steps': 11220, 'loss/train': 0.9455739855766296} 01/29/2022 03:16:13 - INFO - codeparrot_training - Step 11221: {'lr': 0.0004558367288013352, 'samples': 2154624, 'steps': 11221, 'loss/train': 1.695722758769989} 01/29/2022 03:16:19 - INFO - codeparrot_training - Step 11222: {'lr': 0.00045582744203480417, 'samples': 2154816, 'steps': 11222, 'loss/train': 2.5327184200286865} 01/29/2022 03:16:23 - INFO - codeparrot_training - Step 11223: {'lr': 0.0004558181543865738, 'samples': 2155008, 'steps': 11223, 'loss/train': 1.3096154630184174} 01/29/2022 03:16:28 - INFO - codeparrot_training - Step 11224: {'lr': 0.0004558088658566838, 'samples': 2155200, 'steps': 11224, 'loss/train': 2.2733681201934814} 01/29/2022 03:16:32 - INFO - codeparrot_training - Step 11225: {'lr': 0.000455799576445174, 'samples': 2155392, 'steps': 11225, 'loss/train': 1.5137836933135986} 01/29/2022 03:16:36 - INFO - codeparrot_training - Step 11226: {'lr': 0.00045579028615208404, 'samples': 2155584, 'steps': 11226, 'loss/train': 1.9725553393363953} 01/29/2022 03:16:41 - INFO - codeparrot_training - Step 11227: {'lr': 0.000455780994977454, 'samples': 2155776, 'steps': 11227, 'loss/train': 1.141493797302246} 01/29/2022 03:16:46 - INFO - codeparrot_training - Step 11228: {'lr': 0.0004557717029213234, 'samples': 2155968, 'steps': 11228, 'loss/train': 1.334500640630722} 01/29/2022 03:16:50 - INFO - codeparrot_training - Step 11229: {'lr': 0.00045576240998373226, 'samples': 2156160, 'steps': 11229, 'loss/train': 1.8222035765647888} 01/29/2022 03:16:54 - INFO - codeparrot_training - Step 11230: {'lr': 0.00045575311616472024, 'samples': 2156352, 'steps': 11230, 'loss/train': 1.8376978039741516} 01/29/2022 03:16:58 - INFO - codeparrot_training - Step 11231: {'lr': 0.0004557438214643272, 'samples': 2156544, 'steps': 11231, 'loss/train': 1.95940500497818} 01/29/2022 03:17:04 - INFO - codeparrot_training - Step 11232: {'lr': 0.00045573452588259296, 'samples': 2156736, 'steps': 11232, 'loss/train': 1.6279689073562622} 01/29/2022 03:17:08 - INFO - codeparrot_training - Step 11233: {'lr': 0.0004557252294195573, 'samples': 2156928, 'steps': 11233, 'loss/train': 1.7091761827468872} 01/29/2022 03:17:12 - INFO - codeparrot_training - Step 11234: {'lr': 0.00045571593207526016, 'samples': 2157120, 'steps': 11234, 'loss/train': 1.7877221703529358} 01/29/2022 03:17:16 - INFO - codeparrot_training - Step 11235: {'lr': 0.00045570663384974125, 'samples': 2157312, 'steps': 11235, 'loss/train': 1.8327890038490295} 01/29/2022 03:17:21 - INFO - codeparrot_training - Step 11236: {'lr': 0.00045569733474304044, 'samples': 2157504, 'steps': 11236, 'loss/train': 0.9236770570278168} 01/29/2022 03:17:27 - INFO - codeparrot_training - Step 11237: {'lr': 0.0004556880347551976, 'samples': 2157696, 'steps': 11237, 'loss/train': 1.3751150965690613} 01/29/2022 03:17:31 - INFO - codeparrot_training - Step 11238: {'lr': 0.0004556787338862525, 'samples': 2157888, 'steps': 11238, 'loss/train': 1.9428945779800415} 01/29/2022 03:17:35 - INFO - codeparrot_training - Step 11239: {'lr': 0.000455669432136245, 'samples': 2158080, 'steps': 11239, 'loss/train': 1.5355947017669678} 01/29/2022 03:17:40 - INFO - codeparrot_training - Step 11240: {'lr': 0.00045566012950521497, 'samples': 2158272, 'steps': 11240, 'loss/train': 2.339399993419647} 01/29/2022 03:17:44 - INFO - codeparrot_training - Step 11241: {'lr': 0.0004556508259932022, 'samples': 2158464, 'steps': 11241, 'loss/train': 1.9926947951316833} 01/29/2022 03:17:49 - INFO - codeparrot_training - Step 11242: {'lr': 0.0004556415216002467, 'samples': 2158656, 'steps': 11242, 'loss/train': 2.475155532360077} 01/29/2022 03:17:53 - INFO - codeparrot_training - Step 11243: {'lr': 0.0004556322163263882, 'samples': 2158848, 'steps': 11243, 'loss/train': 2.07952880859375} 01/29/2022 03:17:58 - INFO - codeparrot_training - Step 11244: {'lr': 0.00045562291017166653, 'samples': 2159040, 'steps': 11244, 'loss/train': 1.436934471130371} 01/29/2022 03:18:02 - INFO - codeparrot_training - Step 11245: {'lr': 0.0004556136031361216, 'samples': 2159232, 'steps': 11245, 'loss/train': 1.9036306142807007} 01/29/2022 03:18:06 - INFO - codeparrot_training - Step 11246: {'lr': 0.0004556042952197933, 'samples': 2159424, 'steps': 11246, 'loss/train': 1.9103341698646545} 01/29/2022 03:18:12 - INFO - codeparrot_training - Step 11247: {'lr': 0.00045559498642272153, 'samples': 2159616, 'steps': 11247, 'loss/train': 1.589619219303131} 01/29/2022 03:18:16 - INFO - codeparrot_training - Step 11248: {'lr': 0.0004555856767449461, 'samples': 2159808, 'steps': 11248, 'loss/train': 1.4708066582679749} 01/29/2022 03:18:20 - INFO - codeparrot_training - Step 11249: {'lr': 0.00045557636618650686, 'samples': 2160000, 'steps': 11249, 'loss/train': 1.1541210114955902} 01/29/2022 03:18:24 - INFO - codeparrot_training - Step 11250: {'lr': 0.00045556705474744376, 'samples': 2160192, 'steps': 11250, 'loss/train': 1.2048068046569824} 01/29/2022 03:18:29 - INFO - codeparrot_training - Step 11251: {'lr': 0.00045555774242779675, 'samples': 2160384, 'steps': 11251, 'loss/train': 1.3117685616016388} 01/29/2022 03:18:34 - INFO - codeparrot_training - Step 11252: {'lr': 0.0004555484292276055, 'samples': 2160576, 'steps': 11252, 'loss/train': 1.934251070022583} 01/29/2022 03:18:38 - INFO - codeparrot_training - Step 11253: {'lr': 0.0004555391151469102, 'samples': 2160768, 'steps': 11253, 'loss/train': 0.7993076741695404} 01/29/2022 03:18:42 - INFO - codeparrot_training - Step 11254: {'lr': 0.00045552980018575054, 'samples': 2160960, 'steps': 11254, 'loss/train': 1.8182575106620789} 01/29/2022 03:18:47 - INFO - codeparrot_training - Step 11255: {'lr': 0.0004555204843441665, 'samples': 2161152, 'steps': 11255, 'loss/train': 1.8764623403549194} 01/29/2022 03:18:51 - INFO - codeparrot_training - Step 11256: {'lr': 0.0004555111676221979, 'samples': 2161344, 'steps': 11256, 'loss/train': 2.4003764390945435} 01/29/2022 03:18:57 - INFO - codeparrot_training - Step 11257: {'lr': 0.00045550185001988475, 'samples': 2161536, 'steps': 11257, 'loss/train': 1.6466043591499329} 01/29/2022 03:19:01 - INFO - codeparrot_training - Step 11258: {'lr': 0.00045549253153726694, 'samples': 2161728, 'steps': 11258, 'loss/train': 1.8304429650306702} 01/29/2022 03:19:05 - INFO - codeparrot_training - Step 11259: {'lr': 0.00045548321217438436, 'samples': 2161920, 'steps': 11259, 'loss/train': 1.9394670724868774} 01/29/2022 03:19:09 - INFO - codeparrot_training - Step 11260: {'lr': 0.00045547389193127696, 'samples': 2162112, 'steps': 11260, 'loss/train': 1.5179770588874817} 01/29/2022 03:19:14 - INFO - codeparrot_training - Step 11261: {'lr': 0.00045546457080798463, 'samples': 2162304, 'steps': 11261, 'loss/train': 1.0015426576137543} 01/29/2022 03:19:19 - INFO - codeparrot_training - Step 11262: {'lr': 0.00045545524880454734, 'samples': 2162496, 'steps': 11262, 'loss/train': 1.7629973888397217} 01/29/2022 03:19:24 - INFO - codeparrot_training - Step 11263: {'lr': 0.000455445925921005, 'samples': 2162688, 'steps': 11263, 'loss/train': 1.636474370956421} 01/29/2022 03:19:28 - INFO - codeparrot_training - Step 11264: {'lr': 0.00045543660215739755, 'samples': 2162880, 'steps': 11264, 'loss/train': 1.6943492889404297} 01/29/2022 03:19:32 - INFO - codeparrot_training - Step 11265: {'lr': 0.00045542727751376495, 'samples': 2163072, 'steps': 11265, 'loss/train': 1.9543291926383972} 01/29/2022 03:19:36 - INFO - codeparrot_training - Step 11266: {'lr': 0.00045541795199014715, 'samples': 2163264, 'steps': 11266, 'loss/train': 2.4096944332122803} 01/29/2022 03:19:41 - INFO - codeparrot_training - Step 11267: {'lr': 0.00045540862558658403, 'samples': 2163456, 'steps': 11267, 'loss/train': 2.0566611886024475} 01/29/2022 03:19:46 - INFO - codeparrot_training - Step 11268: {'lr': 0.00045539929830311555, 'samples': 2163648, 'steps': 11268, 'loss/train': 2.3882963061332703} 01/29/2022 03:19:50 - INFO - codeparrot_training - Step 11269: {'lr': 0.00045538997013978166, 'samples': 2163840, 'steps': 11269, 'loss/train': 1.6680375337600708} 01/29/2022 03:19:55 - INFO - codeparrot_training - Step 11270: {'lr': 0.0004553806410966225, 'samples': 2164032, 'steps': 11270, 'loss/train': 1.038785308599472} 01/29/2022 03:19:59 - INFO - codeparrot_training - Step 11271: {'lr': 0.0004553713111736778, 'samples': 2164224, 'steps': 11271, 'loss/train': 1.468270093202591} 01/29/2022 03:20:06 - INFO - codeparrot_training - Step 11272: {'lr': 0.0004553619803709876, 'samples': 2164416, 'steps': 11272, 'loss/train': 2.763234257698059} 01/29/2022 03:20:10 - INFO - codeparrot_training - Step 11273: {'lr': 0.00045535264868859195, 'samples': 2164608, 'steps': 11273, 'loss/train': 2.0633077025413513} 01/29/2022 03:20:14 - INFO - codeparrot_training - Step 11274: {'lr': 0.0004553433161265307, 'samples': 2164800, 'steps': 11274, 'loss/train': 2.0322993993759155} 01/29/2022 03:20:18 - INFO - codeparrot_training - Step 11275: {'lr': 0.00045533398268484396, 'samples': 2164992, 'steps': 11275, 'loss/train': 1.9518375992774963} 01/29/2022 03:20:23 - INFO - codeparrot_training - Step 11276: {'lr': 0.00045532464836357155, 'samples': 2165184, 'steps': 11276, 'loss/train': 1.831929624080658} 01/29/2022 03:20:29 - INFO - codeparrot_training - Step 11277: {'lr': 0.0004553153131627536, 'samples': 2165376, 'steps': 11277, 'loss/train': 0.9800271391868591} 01/29/2022 03:20:33 - INFO - codeparrot_training - Step 11278: {'lr': 0.00045530597708243, 'samples': 2165568, 'steps': 11278, 'loss/train': 1.7009364366531372} 01/29/2022 03:20:37 - INFO - codeparrot_training - Step 11279: {'lr': 0.0004552966401226408, 'samples': 2165760, 'steps': 11279, 'loss/train': 1.5604825615882874} 01/29/2022 03:20:42 - INFO - codeparrot_training - Step 11280: {'lr': 0.000455287302283426, 'samples': 2165952, 'steps': 11280, 'loss/train': 1.7550618052482605} 01/29/2022 03:20:46 - INFO - codeparrot_training - Step 11281: {'lr': 0.00045527796356482566, 'samples': 2166144, 'steps': 11281, 'loss/train': 1.7725077867507935} 01/29/2022 03:20:50 - INFO - codeparrot_training - Step 11282: {'lr': 0.00045526862396687957, 'samples': 2166336, 'steps': 11282, 'loss/train': 1.3597427308559418} 01/29/2022 03:20:54 - INFO - codeparrot_training - Step 11283: {'lr': 0.000455259283489628, 'samples': 2166528, 'steps': 11283, 'loss/train': 1.6805207133293152} 01/29/2022 03:21:01 - INFO - codeparrot_training - Step 11284: {'lr': 0.0004552499421331107, 'samples': 2166720, 'steps': 11284, 'loss/train': 1.7934218645095825} 01/29/2022 03:21:05 - INFO - codeparrot_training - Step 11285: {'lr': 0.0004552405998973679, 'samples': 2166912, 'steps': 11285, 'loss/train': 1.663662850856781} 01/29/2022 03:21:09 - INFO - codeparrot_training - Step 11286: {'lr': 0.0004552312567824395, 'samples': 2167104, 'steps': 11286, 'loss/train': 2.2846133708953857} 01/29/2022 03:21:13 - INFO - codeparrot_training - Step 11287: {'lr': 0.00045522191278836563, 'samples': 2167296, 'steps': 11287, 'loss/train': 1.764720618724823} 01/29/2022 03:21:18 - INFO - codeparrot_training - Step 11288: {'lr': 0.00045521256791518616, 'samples': 2167488, 'steps': 11288, 'loss/train': 1.872796654701233} 01/29/2022 03:21:23 - INFO - codeparrot_training - Step 11289: {'lr': 0.0004552032221629413, 'samples': 2167680, 'steps': 11289, 'loss/train': 1.7345525622367859} 01/29/2022 03:21:27 - INFO - codeparrot_training - Step 11290: {'lr': 0.000455193875531671, 'samples': 2167872, 'steps': 11290, 'loss/train': 1.3471530675888062} 01/29/2022 03:21:32 - INFO - codeparrot_training - Step 11291: {'lr': 0.00045518452802141524, 'samples': 2168064, 'steps': 11291, 'loss/train': 2.514611542224884} 01/29/2022 03:21:36 - INFO - codeparrot_training - Step 11292: {'lr': 0.0004551751796322141, 'samples': 2168256, 'steps': 11292, 'loss/train': 1.7097258567810059} 01/29/2022 03:21:40 - INFO - codeparrot_training - Step 11293: {'lr': 0.00045516583036410777, 'samples': 2168448, 'steps': 11293, 'loss/train': 1.5690032243728638} 01/29/2022 03:21:45 - INFO - codeparrot_training - Step 11294: {'lr': 0.00045515648021713604, 'samples': 2168640, 'steps': 11294, 'loss/train': 1.7336894273757935} 01/29/2022 03:21:50 - INFO - codeparrot_training - Step 11295: {'lr': 0.0004551471291913391, 'samples': 2168832, 'steps': 11295, 'loss/train': 1.7987582087516785} 01/29/2022 03:21:54 - INFO - codeparrot_training - Step 11296: {'lr': 0.00045513777728675703, 'samples': 2169024, 'steps': 11296, 'loss/train': 1.823686122894287} 01/29/2022 03:21:58 - INFO - codeparrot_training - Step 11297: {'lr': 0.0004551284245034298, 'samples': 2169216, 'steps': 11297, 'loss/train': 1.4051300883293152} 01/29/2022 03:22:02 - INFO - codeparrot_training - Step 11298: {'lr': 0.00045511907084139767, 'samples': 2169408, 'steps': 11298, 'loss/train': 1.6432183384895325} 01/29/2022 03:22:09 - INFO - codeparrot_training - Step 11299: {'lr': 0.0004551097163007005, 'samples': 2169600, 'steps': 11299, 'loss/train': 2.1310424208641052} 01/29/2022 03:22:13 - INFO - codeparrot_training - Step 11300: {'lr': 0.0004551003608813784, 'samples': 2169792, 'steps': 11300, 'loss/train': 2.2309266328811646} 01/29/2022 03:22:18 - INFO - codeparrot_training - Step 11301: {'lr': 0.00045509100458347154, 'samples': 2169984, 'steps': 11301, 'loss/train': 1.1757294237613678} 01/29/2022 03:22:22 - INFO - codeparrot_training - Step 11302: {'lr': 0.0004550816474070199, 'samples': 2170176, 'steps': 11302, 'loss/train': 1.8675840497016907} 01/29/2022 03:22:26 - INFO - codeparrot_training - Step 11303: {'lr': 0.0004550722893520636, 'samples': 2170368, 'steps': 11303, 'loss/train': 2.152157485485077} 01/29/2022 03:22:31 - INFO - codeparrot_training - Step 11304: {'lr': 0.0004550629304186428, 'samples': 2170560, 'steps': 11304, 'loss/train': 1.960834801197052} 01/29/2022 03:22:36 - INFO - codeparrot_training - Step 11305: {'lr': 0.0004550535706067974, 'samples': 2170752, 'steps': 11305, 'loss/train': 2.2956170439720154} 01/29/2022 03:22:40 - INFO - codeparrot_training - Step 11306: {'lr': 0.0004550442099165677, 'samples': 2170944, 'steps': 11306, 'loss/train': 1.6658797860145569} 01/29/2022 03:22:44 - INFO - codeparrot_training - Step 11307: {'lr': 0.0004550348483479937, 'samples': 2171136, 'steps': 11307, 'loss/train': 1.5259147882461548} 01/29/2022 03:22:48 - INFO - codeparrot_training - Step 11308: {'lr': 0.00045502548590111553, 'samples': 2171328, 'steps': 11308, 'loss/train': 1.8829237818717957} 01/29/2022 03:22:55 - INFO - codeparrot_training - Step 11309: {'lr': 0.0004550161225759732, 'samples': 2171520, 'steps': 11309, 'loss/train': 1.925114929676056} 01/29/2022 03:23:00 - INFO - codeparrot_training - Step 11310: {'lr': 0.000455006758372607, 'samples': 2171712, 'steps': 11310, 'loss/train': 1.5578051805496216} 01/29/2022 03:23:04 - INFO - codeparrot_training - Step 11311: {'lr': 0.00045499739329105696, 'samples': 2171904, 'steps': 11311, 'loss/train': 2.3262981176376343} 01/29/2022 03:23:08 - INFO - codeparrot_training - Step 11312: {'lr': 0.00045498802733136306, 'samples': 2172096, 'steps': 11312, 'loss/train': 0.9412428438663483} 01/29/2022 03:23:12 - INFO - codeparrot_training - Step 11313: {'lr': 0.00045497866049356564, 'samples': 2172288, 'steps': 11313, 'loss/train': 0.9432720243930817} 01/29/2022 03:23:17 - INFO - codeparrot_training - Step 11314: {'lr': 0.0004549692927777047, 'samples': 2172480, 'steps': 11314, 'loss/train': 2.1608017086982727} 01/29/2022 03:23:22 - INFO - codeparrot_training - Step 11315: {'lr': 0.00045495992418382035, 'samples': 2172672, 'steps': 11315, 'loss/train': 1.7738286852836609} 01/29/2022 03:23:26 - INFO - codeparrot_training - Step 11316: {'lr': 0.0004549505547119529, 'samples': 2172864, 'steps': 11316, 'loss/train': 2.3483662605285645} 01/29/2022 03:23:30 - INFO - codeparrot_training - Step 11317: {'lr': 0.00045494118436214225, 'samples': 2173056, 'steps': 11317, 'loss/train': 1.9808063507080078} 01/29/2022 03:23:35 - INFO - codeparrot_training - Step 11318: {'lr': 0.00045493181313442866, 'samples': 2173248, 'steps': 11318, 'loss/train': 2.1641446352005005} 01/29/2022 03:23:39 - INFO - codeparrot_training - Step 11319: {'lr': 0.00045492244102885224, 'samples': 2173440, 'steps': 11319, 'loss/train': 0.9424185454845428} 01/29/2022 03:23:44 - INFO - codeparrot_training - Step 11320: {'lr': 0.00045491306804545316, 'samples': 2173632, 'steps': 11320, 'loss/train': 2.007743775844574} 01/29/2022 03:23:49 - INFO - codeparrot_training - Step 11321: {'lr': 0.0004549036941842716, 'samples': 2173824, 'steps': 11321, 'loss/train': 0.9973467886447906} 01/29/2022 03:23:53 - INFO - codeparrot_training - Step 11322: {'lr': 0.0004548943194453476, 'samples': 2174016, 'steps': 11322, 'loss/train': 1.5474902987480164} 01/29/2022 03:23:57 - INFO - codeparrot_training - Step 11323: {'lr': 0.0004548849438287214, 'samples': 2174208, 'steps': 11323, 'loss/train': 2.000426173210144} 01/29/2022 03:24:01 - INFO - codeparrot_training - Step 11324: {'lr': 0.00045487556733443327, 'samples': 2174400, 'steps': 11324, 'loss/train': 2.5012702345848083} 01/29/2022 03:24:08 - INFO - codeparrot_training - Step 11325: {'lr': 0.00045486618996252315, 'samples': 2174592, 'steps': 11325, 'loss/train': 2.6773592233657837} 01/29/2022 03:24:12 - INFO - codeparrot_training - Step 11326: {'lr': 0.0004548568117130314, 'samples': 2174784, 'steps': 11326, 'loss/train': 1.355118066072464} 01/29/2022 03:24:16 - INFO - codeparrot_training - Step 11327: {'lr': 0.00045484743258599803, 'samples': 2174976, 'steps': 11327, 'loss/train': 0.7552887797355652} 01/29/2022 03:24:20 - INFO - codeparrot_training - Step 11328: {'lr': 0.0004548380525814634, 'samples': 2175168, 'steps': 11328, 'loss/train': 2.014601469039917} 01/29/2022 03:24:25 - INFO - codeparrot_training - Step 11329: {'lr': 0.0004548286716994676, 'samples': 2175360, 'steps': 11329, 'loss/train': 1.688047707080841} 01/29/2022 03:24:30 - INFO - codeparrot_training - Step 11330: {'lr': 0.0004548192899400507, 'samples': 2175552, 'steps': 11330, 'loss/train': 1.767384946346283} 01/29/2022 03:24:34 - INFO - codeparrot_training - Step 11331: {'lr': 0.0004548099073032531, 'samples': 2175744, 'steps': 11331, 'loss/train': 1.4861834049224854} 01/29/2022 03:24:39 - INFO - codeparrot_training - Step 11332: {'lr': 0.00045480052378911483, 'samples': 2175936, 'steps': 11332, 'loss/train': 2.0139922499656677} 01/29/2022 03:24:43 - INFO - codeparrot_training - Step 11333: {'lr': 0.0004547911393976762, 'samples': 2176128, 'steps': 11333, 'loss/train': 1.6935705542564392} 01/29/2022 03:24:47 - INFO - codeparrot_training - Step 11334: {'lr': 0.00045478175412897733, 'samples': 2176320, 'steps': 11334, 'loss/train': 1.852156162261963} 01/29/2022 03:24:53 - INFO - codeparrot_training - Step 11335: {'lr': 0.00045477236798305846, 'samples': 2176512, 'steps': 11335, 'loss/train': 1.6299203038215637} 01/29/2022 03:24:58 - INFO - codeparrot_training - Step 11336: {'lr': 0.00045476298095995985, 'samples': 2176704, 'steps': 11336, 'loss/train': 2.066567838191986} 01/29/2022 03:25:02 - INFO - codeparrot_training - Step 11337: {'lr': 0.0004547535930597215, 'samples': 2176896, 'steps': 11337, 'loss/train': 1.6429166793823242} 01/29/2022 03:25:06 - INFO - codeparrot_training - Step 11338: {'lr': 0.0004547442042823839, 'samples': 2177088, 'steps': 11338, 'loss/train': 1.5951610207557678} 01/29/2022 03:25:10 - INFO - codeparrot_training - Step 11339: {'lr': 0.0004547348146279871, 'samples': 2177280, 'steps': 11339, 'loss/train': 1.5383663177490234} 01/29/2022 03:25:16 - INFO - codeparrot_training - Step 11340: {'lr': 0.00045472542409657135, 'samples': 2177472, 'steps': 11340, 'loss/train': 1.1758827567100525} 01/29/2022 03:25:20 - INFO - codeparrot_training - Step 11341: {'lr': 0.00045471603268817696, 'samples': 2177664, 'steps': 11341, 'loss/train': 1.1881441175937653} 01/29/2022 03:25:24 - INFO - codeparrot_training - Step 11342: {'lr': 0.000454706640402844, 'samples': 2177856, 'steps': 11342, 'loss/train': 1.295925110578537} 01/29/2022 03:25:29 - INFO - codeparrot_training - Step 11343: {'lr': 0.00045469724724061286, 'samples': 2178048, 'steps': 11343, 'loss/train': 1.2973826229572296} 01/29/2022 03:25:33 - INFO - codeparrot_training - Step 11344: {'lr': 0.0004546878532015236, 'samples': 2178240, 'steps': 11344, 'loss/train': 1.6721341609954834} 01/29/2022 03:25:38 - INFO - codeparrot_training - Step 11345: {'lr': 0.00045467845828561673, 'samples': 2178432, 'steps': 11345, 'loss/train': 1.847069799900055} 01/29/2022 03:25:42 - INFO - codeparrot_training - Step 11346: {'lr': 0.0004546690624929322, 'samples': 2178624, 'steps': 11346, 'loss/train': 2.4157601594924927} 01/29/2022 03:25:47 - INFO - codeparrot_training - Step 11347: {'lr': 0.0004546596658235105, 'samples': 2178816, 'steps': 11347, 'loss/train': 2.05565482378006} 01/29/2022 03:25:51 - INFO - codeparrot_training - Step 11348: {'lr': 0.00045465026827739175, 'samples': 2179008, 'steps': 11348, 'loss/train': 2.3552677631378174} 01/29/2022 03:25:55 - INFO - codeparrot_training - Step 11349: {'lr': 0.00045464086985461615, 'samples': 2179200, 'steps': 11349, 'loss/train': 1.0852684378623962} 01/29/2022 03:26:01 - INFO - codeparrot_training - Step 11350: {'lr': 0.0004546314705552241, 'samples': 2179392, 'steps': 11350, 'loss/train': 2.447982430458069} 01/29/2022 03:26:05 - INFO - codeparrot_training - Step 11351: {'lr': 0.00045462207037925593, 'samples': 2179584, 'steps': 11351, 'loss/train': 0.8217381834983826} 01/29/2022 03:26:09 - INFO - codeparrot_training - Step 11352: {'lr': 0.0004546126693267516, 'samples': 2179776, 'steps': 11352, 'loss/train': 1.837282955646515} 01/29/2022 03:26:13 - INFO - codeparrot_training - Step 11353: {'lr': 0.0004546032673977517, 'samples': 2179968, 'steps': 11353, 'loss/train': 1.0400521159172058} 01/29/2022 03:26:17 - INFO - codeparrot_training - Step 11354: {'lr': 0.0004545938645922963, 'samples': 2180160, 'steps': 11354, 'loss/train': 1.609347939491272} 01/29/2022 03:26:24 - INFO - codeparrot_training - Step 11355: {'lr': 0.0004545844609104258, 'samples': 2180352, 'steps': 11355, 'loss/train': 1.5254432559013367} 01/29/2022 03:26:28 - INFO - codeparrot_training - Step 11356: {'lr': 0.0004545750563521804, 'samples': 2180544, 'steps': 11356, 'loss/train': 2.2938643097877502} 01/29/2022 03:26:32 - INFO - codeparrot_training - Step 11357: {'lr': 0.0004545656509176004, 'samples': 2180736, 'steps': 11357, 'loss/train': 2.0017136335372925} 01/29/2022 03:26:37 - INFO - codeparrot_training - Step 11358: {'lr': 0.0004545562446067261, 'samples': 2180928, 'steps': 11358, 'loss/train': 1.5695376992225647} 01/29/2022 03:26:41 - INFO - codeparrot_training - Step 11359: {'lr': 0.00045454683741959787, 'samples': 2181120, 'steps': 11359, 'loss/train': 1.481699287891388} 01/29/2022 03:26:46 - INFO - codeparrot_training - Step 11360: {'lr': 0.0004545374293562559, 'samples': 2181312, 'steps': 11360, 'loss/train': 1.535456120967865} 01/29/2022 03:26:50 - INFO - codeparrot_training - Step 11361: {'lr': 0.00045452802041674045, 'samples': 2181504, 'steps': 11361, 'loss/train': 1.9196044206619263} 01/29/2022 03:26:55 - INFO - codeparrot_training - Step 11362: {'lr': 0.000454518610601092, 'samples': 2181696, 'steps': 11362, 'loss/train': 1.9841909408569336} 01/29/2022 03:26:59 - INFO - codeparrot_training - Step 11363: {'lr': 0.0004545091999093508, 'samples': 2181888, 'steps': 11363, 'loss/train': 1.8652262091636658} 01/29/2022 03:27:03 - INFO - codeparrot_training - Step 11364: {'lr': 0.00045449978834155705, 'samples': 2182080, 'steps': 11364, 'loss/train': 1.1421687304973602} 01/29/2022 03:27:09 - INFO - codeparrot_training - Step 11365: {'lr': 0.00045449037589775123, 'samples': 2182272, 'steps': 11365, 'loss/train': 0.4598300904035568} 01/29/2022 03:27:14 - INFO - codeparrot_training - Step 11366: {'lr': 0.00045448096257797344, 'samples': 2182464, 'steps': 11366, 'loss/train': 1.2948718965053558} 01/29/2022 03:27:18 - INFO - codeparrot_training - Step 11367: {'lr': 0.0004544715483822642, 'samples': 2182656, 'steps': 11367, 'loss/train': 1.8842563033103943} 01/29/2022 03:27:22 - INFO - codeparrot_training - Step 11368: {'lr': 0.00045446213331066376, 'samples': 2182848, 'steps': 11368, 'loss/train': 7.306076288223267} 01/29/2022 03:27:26 - INFO - codeparrot_training - Step 11369: {'lr': 0.0004544527173632125, 'samples': 2183040, 'steps': 11369, 'loss/train': 3.415545701980591} 01/29/2022 03:27:31 - INFO - codeparrot_training - Step 11370: {'lr': 0.00045444330053995074, 'samples': 2183232, 'steps': 11370, 'loss/train': 1.7480807304382324} 01/29/2022 03:27:37 - INFO - codeparrot_training - Step 11371: {'lr': 0.00045443388284091877, 'samples': 2183424, 'steps': 11371, 'loss/train': 1.9104477167129517} 01/29/2022 03:27:41 - INFO - codeparrot_training - Step 11372: {'lr': 0.0004544244642661569, 'samples': 2183616, 'steps': 11372, 'loss/train': 1.7178947925567627} 01/29/2022 03:27:46 - INFO - codeparrot_training - Step 11373: {'lr': 0.0004544150448157056, 'samples': 2183808, 'steps': 11373, 'loss/train': 1.5509905815124512} 01/29/2022 03:27:50 - INFO - codeparrot_training - Step 11374: {'lr': 0.0004544056244896052, 'samples': 2184000, 'steps': 11374, 'loss/train': 1.7648245096206665} 01/29/2022 03:27:54 - INFO - codeparrot_training - Step 11375: {'lr': 0.00045439620328789593, 'samples': 2184192, 'steps': 11375, 'loss/train': 1.9207819104194641} 01/29/2022 03:27:59 - INFO - codeparrot_training - Step 11376: {'lr': 0.00045438678121061826, 'samples': 2184384, 'steps': 11376, 'loss/train': 2.030518412590027} 01/29/2022 03:28:04 - INFO - codeparrot_training - Step 11377: {'lr': 0.0004543773582578125, 'samples': 2184576, 'steps': 11377, 'loss/train': 1.8508012890815735} 01/29/2022 03:28:08 - INFO - codeparrot_training - Step 11378: {'lr': 0.00045436793442951907, 'samples': 2184768, 'steps': 11378, 'loss/train': 1.7445750832557678} 01/29/2022 03:28:12 - INFO - codeparrot_training - Step 11379: {'lr': 0.0004543585097257783, 'samples': 2184960, 'steps': 11379, 'loss/train': 2.5341773629188538} 01/29/2022 03:28:16 - INFO - codeparrot_training - Step 11380: {'lr': 0.0004543490841466306, 'samples': 2185152, 'steps': 11380, 'loss/train': 1.1516372859477997} 01/29/2022 03:28:22 - INFO - codeparrot_training - Step 11381: {'lr': 0.00045433965769211616, 'samples': 2185344, 'steps': 11381, 'loss/train': 1.6903191804885864} 01/29/2022 03:28:26 - INFO - codeparrot_training - Step 11382: {'lr': 0.00045433023036227566, 'samples': 2185536, 'steps': 11382, 'loss/train': 0.1803523153066635} 01/29/2022 03:28:30 - INFO - codeparrot_training - Step 11383: {'lr': 0.00045432080215714927, 'samples': 2185728, 'steps': 11383, 'loss/train': 2.3882614374160767} 01/29/2022 03:28:34 - INFO - codeparrot_training - Step 11384: {'lr': 0.00045431137307677753, 'samples': 2185920, 'steps': 11384, 'loss/train': 1.7923892140388489} 01/29/2022 03:28:39 - INFO - codeparrot_training - Step 11385: {'lr': 0.00045430194312120066, 'samples': 2186112, 'steps': 11385, 'loss/train': 0.42937391996383667} 01/29/2022 03:28:45 - INFO - codeparrot_training - Step 11386: {'lr': 0.0004542925122904591, 'samples': 2186304, 'steps': 11386, 'loss/train': 2.077559173107147} 01/29/2022 03:28:49 - INFO - codeparrot_training - Step 11387: {'lr': 0.00045428308058459335, 'samples': 2186496, 'steps': 11387, 'loss/train': 1.2358076870441437} 01/29/2022 03:28:53 - INFO - codeparrot_training - Step 11388: {'lr': 0.00045427364800364374, 'samples': 2186688, 'steps': 11388, 'loss/train': 1.9826631546020508} 01/29/2022 03:28:58 - INFO - codeparrot_training - Step 11389: {'lr': 0.00045426421454765065, 'samples': 2186880, 'steps': 11389, 'loss/train': 0.9385304152965546} 01/29/2022 03:29:02 - INFO - codeparrot_training - Step 11390: {'lr': 0.0004542547802166546, 'samples': 2187072, 'steps': 11390, 'loss/train': 1.2455726265907288} 01/29/2022 03:29:07 - INFO - codeparrot_training - Step 11391: {'lr': 0.00045424534501069594, 'samples': 2187264, 'steps': 11391, 'loss/train': 2.424434244632721} 01/29/2022 03:29:11 - INFO - codeparrot_training - Step 11392: {'lr': 0.00045423590892981503, 'samples': 2187456, 'steps': 11392, 'loss/train': 0.5971190929412842} 01/29/2022 03:29:16 - INFO - codeparrot_training - Step 11393: {'lr': 0.0004542264719740523, 'samples': 2187648, 'steps': 11393, 'loss/train': 1.1757019758224487} 01/29/2022 03:29:20 - INFO - codeparrot_training - Step 11394: {'lr': 0.0004542170341434483, 'samples': 2187840, 'steps': 11394, 'loss/train': 1.4677752256393433} 01/29/2022 03:29:24 - INFO - codeparrot_training - Step 11395: {'lr': 0.00045420759543804326, 'samples': 2188032, 'steps': 11395, 'loss/train': 2.7424641251564026} 01/29/2022 03:29:30 - INFO - codeparrot_training - Step 11396: {'lr': 0.0004541981558578778, 'samples': 2188224, 'steps': 11396, 'loss/train': 1.8530784845352173} 01/29/2022 03:29:35 - INFO - codeparrot_training - Step 11397: {'lr': 0.0004541887154029922, 'samples': 2188416, 'steps': 11397, 'loss/train': 1.796799659729004} 01/29/2022 03:29:39 - INFO - codeparrot_training - Step 11398: {'lr': 0.0004541792740734271, 'samples': 2188608, 'steps': 11398, 'loss/train': 1.0952838957309723} 01/29/2022 03:29:43 - INFO - codeparrot_training - Step 11399: {'lr': 0.0004541698318692228, 'samples': 2188800, 'steps': 11399, 'loss/train': 1.8725194931030273} 01/29/2022 03:29:47 - INFO - codeparrot_training - Step 11400: {'lr': 0.0004541603887904198, 'samples': 2188992, 'steps': 11400, 'loss/train': 1.9990721940994263} 01/29/2022 03:29:53 - INFO - codeparrot_training - Step 11401: {'lr': 0.0004541509448370584, 'samples': 2189184, 'steps': 11401, 'loss/train': 1.7015424370765686} 01/29/2022 03:29:57 - INFO - codeparrot_training - Step 11402: {'lr': 0.00045414150000917927, 'samples': 2189376, 'steps': 11402, 'loss/train': 2.276605725288391} 01/29/2022 03:30:02 - INFO - codeparrot_training - Step 11403: {'lr': 0.0004541320543068227, 'samples': 2189568, 'steps': 11403, 'loss/train': 1.7834181189537048} 01/29/2022 03:30:06 - INFO - codeparrot_training - Step 11404: {'lr': 0.00045412260773002933, 'samples': 2189760, 'steps': 11404, 'loss/train': 1.5749248266220093} 01/29/2022 03:30:11 - INFO - codeparrot_training - Step 11405: {'lr': 0.0004541131602788395, 'samples': 2189952, 'steps': 11405, 'loss/train': 1.8786394000053406} 01/29/2022 03:30:15 - INFO - codeparrot_training - Step 11406: {'lr': 0.00045410371195329365, 'samples': 2190144, 'steps': 11406, 'loss/train': 2.484268605709076} 01/29/2022 03:30:20 - INFO - codeparrot_training - Step 11407: {'lr': 0.00045409426275343234, 'samples': 2190336, 'steps': 11407, 'loss/train': 1.692126452922821} 01/29/2022 03:30:24 - INFO - codeparrot_training - Step 11408: {'lr': 0.00045408481267929604, 'samples': 2190528, 'steps': 11408, 'loss/train': 1.8069695234298706} 01/29/2022 03:30:28 - INFO - codeparrot_training - Step 11409: {'lr': 0.0004540753617309251, 'samples': 2190720, 'steps': 11409, 'loss/train': 1.9999140501022339} 01/29/2022 03:30:33 - INFO - codeparrot_training - Step 11410: {'lr': 0.0004540659099083602, 'samples': 2190912, 'steps': 11410, 'loss/train': 1.7709513902664185} 01/29/2022 03:30:38 - INFO - codeparrot_training - Step 11411: {'lr': 0.0004540564572116418, 'samples': 2191104, 'steps': 11411, 'loss/train': 1.9789652824401855} 01/29/2022 03:30:42 - INFO - codeparrot_training - Step 11412: {'lr': 0.0004540470036408102, 'samples': 2191296, 'steps': 11412, 'loss/train': 0.7466091960668564} 01/29/2022 03:30:46 - INFO - codeparrot_training - Step 11413: {'lr': 0.0004540375491959061, 'samples': 2191488, 'steps': 11413, 'loss/train': 1.6501631140708923} 01/29/2022 03:30:50 - INFO - codeparrot_training - Step 11414: {'lr': 0.00045402809387697, 'samples': 2191680, 'steps': 11414, 'loss/train': 1.8465242385864258} 01/29/2022 03:30:57 - INFO - codeparrot_training - Step 11415: {'lr': 0.00045401863768404217, 'samples': 2191872, 'steps': 11415, 'loss/train': 1.5292512774467468} 01/29/2022 03:31:01 - INFO - codeparrot_training - Step 11416: {'lr': 0.0004540091806171634, 'samples': 2192064, 'steps': 11416, 'loss/train': 1.3949403762817383} 01/29/2022 03:31:05 - INFO - codeparrot_training - Step 11417: {'lr': 0.000453999722676374, 'samples': 2192256, 'steps': 11417, 'loss/train': 2.035800039768219} 01/29/2022 03:31:09 - INFO - codeparrot_training - Step 11418: {'lr': 0.0004539902638617146, 'samples': 2192448, 'steps': 11418, 'loss/train': 2.008361577987671} 01/29/2022 03:31:14 - INFO - codeparrot_training - Step 11419: {'lr': 0.0004539808041732257, 'samples': 2192640, 'steps': 11419, 'loss/train': 2.045955777168274} 01/29/2022 03:31:19 - INFO - codeparrot_training - Step 11420: {'lr': 0.0004539713436109478, 'samples': 2192832, 'steps': 11420, 'loss/train': 1.5888762474060059} 01/29/2022 03:31:23 - INFO - codeparrot_training - Step 11421: {'lr': 0.00045396188217492145, 'samples': 2193024, 'steps': 11421, 'loss/train': 1.5132127404212952} 01/29/2022 03:31:28 - INFO - codeparrot_training - Step 11422: {'lr': 0.00045395241986518714, 'samples': 2193216, 'steps': 11422, 'loss/train': 1.6973376274108887} 01/29/2022 03:31:32 - INFO - codeparrot_training - Step 11423: {'lr': 0.0004539429566817854, 'samples': 2193408, 'steps': 11423, 'loss/train': 1.5436484813690186} 01/29/2022 03:31:36 - INFO - codeparrot_training - Step 11424: {'lr': 0.00045393349262475686, 'samples': 2193600, 'steps': 11424, 'loss/train': 1.686450719833374} 01/29/2022 03:31:41 - INFO - codeparrot_training - Step 11425: {'lr': 0.000453924027694142, 'samples': 2193792, 'steps': 11425, 'loss/train': 0.648909866809845} 01/29/2022 03:31:46 - INFO - codeparrot_training - Step 11426: {'lr': 0.00045391456188998124, 'samples': 2193984, 'steps': 11426, 'loss/train': 1.6327605843544006} 01/29/2022 03:31:50 - INFO - codeparrot_training - Step 11427: {'lr': 0.00045390509521231535, 'samples': 2194176, 'steps': 11427, 'loss/train': 1.8694923520088196} 01/29/2022 03:31:54 - INFO - codeparrot_training - Step 11428: {'lr': 0.00045389562766118475, 'samples': 2194368, 'steps': 11428, 'loss/train': 1.548404574394226} 01/29/2022 03:31:58 - INFO - codeparrot_training - Step 11429: {'lr': 0.00045388615923663004, 'samples': 2194560, 'steps': 11429, 'loss/train': 2.012193202972412} 01/29/2022 03:32:05 - INFO - codeparrot_training - Step 11430: {'lr': 0.0004538766899386917, 'samples': 2194752, 'steps': 11430, 'loss/train': 1.5377760529518127} 01/29/2022 03:32:09 - INFO - codeparrot_training - Step 11431: {'lr': 0.00045386721976741043, 'samples': 2194944, 'steps': 11431, 'loss/train': 1.7390713691711426} 01/29/2022 03:32:13 - INFO - codeparrot_training - Step 11432: {'lr': 0.0004538577487228267, 'samples': 2195136, 'steps': 11432, 'loss/train': 1.6728903651237488} 01/29/2022 03:32:17 - INFO - codeparrot_training - Step 11433: {'lr': 0.0004538482768049811, 'samples': 2195328, 'steps': 11433, 'loss/train': 1.980208396911621} 01/29/2022 03:32:22 - INFO - codeparrot_training - Step 11434: {'lr': 0.00045383880401391423, 'samples': 2195520, 'steps': 11434, 'loss/train': 3.5447927713394165} 01/29/2022 03:32:27 - INFO - codeparrot_training - Step 11435: {'lr': 0.00045382933034966667, 'samples': 2195712, 'steps': 11435, 'loss/train': 2.077974557876587} 01/29/2022 03:32:31 - INFO - codeparrot_training - Step 11436: {'lr': 0.0004538198558122789, 'samples': 2195904, 'steps': 11436, 'loss/train': 1.164035975933075} 01/29/2022 03:32:36 - INFO - codeparrot_training - Step 11437: {'lr': 0.0004538103804017917, 'samples': 2196096, 'steps': 11437, 'loss/train': 2.0168066024780273} 01/29/2022 03:32:40 - INFO - codeparrot_training - Step 11438: {'lr': 0.00045380090411824547, 'samples': 2196288, 'steps': 11438, 'loss/train': 1.643555223941803} 01/29/2022 03:32:44 - INFO - codeparrot_training - Step 11439: {'lr': 0.0004537914269616809, 'samples': 2196480, 'steps': 11439, 'loss/train': 1.4354747235774994} 01/29/2022 03:32:50 - INFO - codeparrot_training - Step 11440: {'lr': 0.00045378194893213854, 'samples': 2196672, 'steps': 11440, 'loss/train': 1.720002293586731} 01/29/2022 03:32:54 - INFO - codeparrot_training - Step 11441: {'lr': 0.00045377247002965904, 'samples': 2196864, 'steps': 11441, 'loss/train': 2.9259263277053833} 01/29/2022 03:32:59 - INFO - codeparrot_training - Step 11442: {'lr': 0.000453762990254283, 'samples': 2197056, 'steps': 11442, 'loss/train': 1.5572848320007324} 01/29/2022 03:33:03 - INFO - codeparrot_training - Step 11443: {'lr': 0.000453753509606051, 'samples': 2197248, 'steps': 11443, 'loss/train': 1.9700242280960083} 01/29/2022 03:33:07 - INFO - codeparrot_training - Step 11444: {'lr': 0.0004537440280850037, 'samples': 2197440, 'steps': 11444, 'loss/train': 1.7622835636138916} 01/29/2022 03:33:13 - INFO - codeparrot_training - Step 11445: {'lr': 0.00045373454569118166, 'samples': 2197632, 'steps': 11445, 'loss/train': 1.7247092127799988} 01/29/2022 03:33:17 - INFO - codeparrot_training - Step 11446: {'lr': 0.0004537250624246255, 'samples': 2197824, 'steps': 11446, 'loss/train': 7.409226894378662} 01/29/2022 03:33:21 - INFO - codeparrot_training - Step 11447: {'lr': 0.00045371557828537585, 'samples': 2198016, 'steps': 11447, 'loss/train': 2.151469588279724} 01/29/2022 03:33:25 - INFO - codeparrot_training - Step 11448: {'lr': 0.0004537060932734733, 'samples': 2198208, 'steps': 11448, 'loss/train': 2.003377676010132} 01/29/2022 03:33:30 - INFO - codeparrot_training - Step 11449: {'lr': 0.0004536966073889587, 'samples': 2198400, 'steps': 11449, 'loss/train': 1.0452354848384857} 01/29/2022 03:33:35 - INFO - codeparrot_training - Step 11450: {'lr': 0.00045368712063187237, 'samples': 2198592, 'steps': 11450, 'loss/train': 1.1897867918014526} 01/29/2022 03:33:39 - INFO - codeparrot_training - Step 11451: {'lr': 0.0004536776330022552, 'samples': 2198784, 'steps': 11451, 'loss/train': 1.7599119544029236} 01/29/2022 03:33:43 - INFO - codeparrot_training - Step 11452: {'lr': 0.0004536681445001476, 'samples': 2198976, 'steps': 11452, 'loss/train': 2.4139214158058167} 01/29/2022 03:33:47 - INFO - codeparrot_training - Step 11453: {'lr': 0.0004536586551255904, 'samples': 2199168, 'steps': 11453, 'loss/train': 1.8559445142745972} 01/29/2022 03:33:52 - INFO - codeparrot_training - Step 11454: {'lr': 0.0004536491648786242, 'samples': 2199360, 'steps': 11454, 'loss/train': 1.5321730971336365} 01/29/2022 03:33:59 - INFO - codeparrot_training - Step 11455: {'lr': 0.0004536396737592896, 'samples': 2199552, 'steps': 11455, 'loss/train': 1.290383219718933} 01/29/2022 03:34:03 - INFO - codeparrot_training - Step 11456: {'lr': 0.0004536301817676274, 'samples': 2199744, 'steps': 11456, 'loss/train': 1.592208981513977} 01/29/2022 03:34:07 - INFO - codeparrot_training - Step 11457: {'lr': 0.00045362068890367804, 'samples': 2199936, 'steps': 11457, 'loss/train': 1.6014404296875} 01/29/2022 03:34:12 - INFO - codeparrot_training - Step 11458: {'lr': 0.0004536111951674824, 'samples': 2200128, 'steps': 11458, 'loss/train': 2.279249668121338} 01/29/2022 03:34:16 - INFO - codeparrot_training - Step 11459: {'lr': 0.000453601700559081, 'samples': 2200320, 'steps': 11459, 'loss/train': 1.0863814651966095} 01/29/2022 03:34:21 - INFO - codeparrot_training - Step 11460: {'lr': 0.00045359220507851456, 'samples': 2200512, 'steps': 11460, 'loss/train': 2.0359033942222595} 01/29/2022 03:34:25 - INFO - codeparrot_training - Step 11461: {'lr': 0.0004535827087258238, 'samples': 2200704, 'steps': 11461, 'loss/train': 2.270588994026184} 01/29/2022 03:34:30 - INFO - codeparrot_training - Step 11462: {'lr': 0.00045357321150104934, 'samples': 2200896, 'steps': 11462, 'loss/train': 1.7276569604873657} 01/29/2022 03:34:34 - INFO - codeparrot_training - Step 11463: {'lr': 0.0004535637134042319, 'samples': 2201088, 'steps': 11463, 'loss/train': 1.9515286087989807} 01/29/2022 03:34:38 - INFO - codeparrot_training - Step 11464: {'lr': 0.00045355421443541214, 'samples': 2201280, 'steps': 11464, 'loss/train': 2.3486584424972534} 01/29/2022 03:34:43 - INFO - codeparrot_training - Step 11465: {'lr': 0.00045354471459463076, 'samples': 2201472, 'steps': 11465, 'loss/train': 1.2879086136817932} 01/29/2022 03:34:48 - INFO - codeparrot_training - Step 11466: {'lr': 0.0004535352138819284, 'samples': 2201664, 'steps': 11466, 'loss/train': 2.529560923576355} 01/29/2022 03:34:52 - INFO - codeparrot_training - Step 11467: {'lr': 0.0004535257122973459, 'samples': 2201856, 'steps': 11467, 'loss/train': 2.130414605140686} 01/29/2022 03:34:56 - INFO - codeparrot_training - Step 11468: {'lr': 0.0004535162098409238, 'samples': 2202048, 'steps': 11468, 'loss/train': 1.8788418173789978} 01/29/2022 03:35:02 - INFO - codeparrot_training - Step 11469: {'lr': 0.000453506706512703, 'samples': 2202240, 'steps': 11469, 'loss/train': 2.348231613636017} 01/29/2022 03:35:06 - INFO - codeparrot_training - Step 11470: {'lr': 0.00045349720231272395, 'samples': 2202432, 'steps': 11470, 'loss/train': 1.7391913533210754} 01/29/2022 03:35:10 - INFO - codeparrot_training - Step 11471: {'lr': 0.0004534876972410276, 'samples': 2202624, 'steps': 11471, 'loss/train': 1.4599612355232239} 01/29/2022 03:35:15 - INFO - codeparrot_training - Step 11472: {'lr': 0.0004534781912976545, 'samples': 2202816, 'steps': 11472, 'loss/train': 1.4326516091823578} 01/29/2022 03:35:19 - INFO - codeparrot_training - Step 11473: {'lr': 0.00045346868448264553, 'samples': 2203008, 'steps': 11473, 'loss/train': 1.8686646223068237} 01/29/2022 03:35:23 - INFO - codeparrot_training - Step 11474: {'lr': 0.00045345917679604126, 'samples': 2203200, 'steps': 11474, 'loss/train': 0.6861636936664581} 01/29/2022 03:35:29 - INFO - codeparrot_training - Step 11475: {'lr': 0.0004534496682378825, 'samples': 2203392, 'steps': 11475, 'loss/train': 1.302570641040802} 01/29/2022 03:35:33 - INFO - codeparrot_training - Step 11476: {'lr': 0.00045344015880821, 'samples': 2203584, 'steps': 11476, 'loss/train': 2.4197423458099365} 01/29/2022 03:35:38 - INFO - codeparrot_training - Step 11477: {'lr': 0.0004534306485070644, 'samples': 2203776, 'steps': 11477, 'loss/train': 0.7627531886100769} 01/29/2022 03:35:42 - INFO - codeparrot_training - Step 11478: {'lr': 0.0004534211373344864, 'samples': 2203968, 'steps': 11478, 'loss/train': 2.508784532546997} 01/29/2022 03:35:46 - INFO - codeparrot_training - Step 11479: {'lr': 0.00045341162529051704, 'samples': 2204160, 'steps': 11479, 'loss/train': 1.573932409286499} 01/29/2022 03:35:52 - INFO - codeparrot_training - Step 11480: {'lr': 0.0004534021123751968, 'samples': 2204352, 'steps': 11480, 'loss/train': 2.3280876874923706} 01/29/2022 03:35:56 - INFO - codeparrot_training - Step 11481: {'lr': 0.0004533925985885664, 'samples': 2204544, 'steps': 11481, 'loss/train': 2.010678470134735} 01/29/2022 03:36:00 - INFO - codeparrot_training - Step 11482: {'lr': 0.00045338308393066685, 'samples': 2204736, 'steps': 11482, 'loss/train': 1.7387441396713257} 01/29/2022 03:36:04 - INFO - codeparrot_training - Step 11483: {'lr': 0.00045337356840153864, 'samples': 2204928, 'steps': 11483, 'loss/train': 2.0939016938209534} 01/29/2022 03:36:09 - INFO - codeparrot_training - Step 11484: {'lr': 0.00045336405200122266, 'samples': 2205120, 'steps': 11484, 'loss/train': 2.637580454349518} 01/29/2022 03:36:15 - INFO - codeparrot_training - Step 11485: {'lr': 0.0004533545347297597, 'samples': 2205312, 'steps': 11485, 'loss/train': 1.937848448753357} 01/29/2022 03:36:19 - INFO - codeparrot_training - Step 11486: {'lr': 0.0004533450165871904, 'samples': 2205504, 'steps': 11486, 'loss/train': 1.7821692824363708} 01/29/2022 03:36:24 - INFO - codeparrot_training - Step 11487: {'lr': 0.00045333549757355573, 'samples': 2205696, 'steps': 11487, 'loss/train': 1.811506748199463} 01/29/2022 03:36:28 - INFO - codeparrot_training - Step 11488: {'lr': 0.0004533259776888963, 'samples': 2205888, 'steps': 11488, 'loss/train': 2.076960325241089} 01/29/2022 03:36:32 - INFO - codeparrot_training - Step 11489: {'lr': 0.00045331645693325295, 'samples': 2206080, 'steps': 11489, 'loss/train': 1.9608601927757263} 01/29/2022 03:36:36 - INFO - codeparrot_training - Step 11490: {'lr': 0.0004533069353066664, 'samples': 2206272, 'steps': 11490, 'loss/train': 1.3781518936157227} 01/29/2022 03:36:42 - INFO - codeparrot_training - Step 11491: {'lr': 0.0004532974128091776, 'samples': 2206464, 'steps': 11491, 'loss/train': 1.8466862440109253} 01/29/2022 03:36:46 - INFO - codeparrot_training - Step 11492: {'lr': 0.00045328788944082717, 'samples': 2206656, 'steps': 11492, 'loss/train': 2.169873833656311} 01/29/2022 03:36:50 - INFO - codeparrot_training - Step 11493: {'lr': 0.000453278365201656, 'samples': 2206848, 'steps': 11493, 'loss/train': 1.5087441802024841} 01/29/2022 03:36:55 - INFO - codeparrot_training - Step 11494: {'lr': 0.00045326884009170486, 'samples': 2207040, 'steps': 11494, 'loss/train': 1.9258734583854675} 01/29/2022 03:36:59 - INFO - codeparrot_training - Step 11495: {'lr': 0.0004532593141110145, 'samples': 2207232, 'steps': 11495, 'loss/train': 1.7472782135009766} 01/29/2022 03:37:04 - INFO - codeparrot_training - Step 11496: {'lr': 0.00045324978725962584, 'samples': 2207424, 'steps': 11496, 'loss/train': 1.3268840610980988} 01/29/2022 03:37:08 - INFO - codeparrot_training - Step 11497: {'lr': 0.0004532402595375796, 'samples': 2207616, 'steps': 11497, 'loss/train': 2.8671140670776367} 01/29/2022 03:37:13 - INFO - codeparrot_training - Step 11498: {'lr': 0.0004532307309449167, 'samples': 2207808, 'steps': 11498, 'loss/train': 1.5895453691482544} 01/29/2022 03:37:17 - INFO - codeparrot_training - Step 11499: {'lr': 0.00045322120148167777, 'samples': 2208000, 'steps': 11499, 'loss/train': 2.0679863691329956} 01/29/2022 03:37:21 - INFO - codeparrot_training - Step 11500: {'lr': 0.0004532116711479038, 'samples': 2208192, 'steps': 11500, 'loss/train': 1.5288469791412354} 01/29/2022 03:37:27 - INFO - codeparrot_training - Step 11501: {'lr': 0.00045320213994363555, 'samples': 2208384, 'steps': 11501, 'loss/train': 1.7633119225502014} 01/29/2022 03:37:31 - INFO - codeparrot_training - Step 11502: {'lr': 0.00045319260786891394, 'samples': 2208576, 'steps': 11502, 'loss/train': 1.8000606894493103} 01/29/2022 03:37:36 - INFO - codeparrot_training - Step 11503: {'lr': 0.0004531830749237796, 'samples': 2208768, 'steps': 11503, 'loss/train': 1.4430102109909058} 01/29/2022 03:37:40 - INFO - codeparrot_training - Step 11504: {'lr': 0.00045317354110827344, 'samples': 2208960, 'steps': 11504, 'loss/train': 1.238688737154007} 01/29/2022 03:37:44 - INFO - codeparrot_training - Step 11505: {'lr': 0.0004531640064224365, 'samples': 2209152, 'steps': 11505, 'loss/train': 2.4798404574394226} 01/29/2022 03:37:50 - INFO - codeparrot_training - Step 11506: {'lr': 0.00045315447086630937, 'samples': 2209344, 'steps': 11506, 'loss/train': 1.1338105201721191} 01/29/2022 03:37:54 - INFO - codeparrot_training - Step 11507: {'lr': 0.000453144934439933, 'samples': 2209536, 'steps': 11507, 'loss/train': 1.7574904561042786} 01/29/2022 03:37:58 - INFO - codeparrot_training - Step 11508: {'lr': 0.0004531353971433483, 'samples': 2209728, 'steps': 11508, 'loss/train': 0.8318012952804565} 01/29/2022 03:38:02 - INFO - codeparrot_training - Step 11509: {'lr': 0.000453125858976596, 'samples': 2209920, 'steps': 11509, 'loss/train': 1.4080781936645508} 01/29/2022 03:38:07 - INFO - codeparrot_training - Step 11510: {'lr': 0.000453116319939717, 'samples': 2210112, 'steps': 11510, 'loss/train': 1.7367138862609863} 01/29/2022 03:38:12 - INFO - codeparrot_training - Step 11511: {'lr': 0.0004531067800327523, 'samples': 2210304, 'steps': 11511, 'loss/train': 2.2904250025749207} 01/29/2022 03:38:17 - INFO - codeparrot_training - Step 11512: {'lr': 0.0004530972392557425, 'samples': 2210496, 'steps': 11512, 'loss/train': 2.27286297082901} 01/29/2022 03:38:21 - INFO - codeparrot_training - Step 11513: {'lr': 0.0004530876976087288, 'samples': 2210688, 'steps': 11513, 'loss/train': 2.0705541372299194} 01/29/2022 03:38:25 - INFO - codeparrot_training - Step 11514: {'lr': 0.00045307815509175177, 'samples': 2210880, 'steps': 11514, 'loss/train': 1.8019723892211914} 01/29/2022 03:38:30 - INFO - codeparrot_training - Step 11515: {'lr': 0.00045306861170485235, 'samples': 2211072, 'steps': 11515, 'loss/train': 2.1954848170280457} 01/29/2022 03:38:35 - INFO - codeparrot_training - Step 11516: {'lr': 0.00045305906744807156, 'samples': 2211264, 'steps': 11516, 'loss/train': 2.398768365383148} 01/29/2022 03:38:39 - INFO - codeparrot_training - Step 11517: {'lr': 0.0004530495223214502, 'samples': 2211456, 'steps': 11517, 'loss/train': 2.4173410534858704} 01/29/2022 03:38:43 - INFO - codeparrot_training - Step 11518: {'lr': 0.00045303997632502915, 'samples': 2211648, 'steps': 11518, 'loss/train': 1.1913221776485443} 01/29/2022 03:38:48 - INFO - codeparrot_training - Step 11519: {'lr': 0.00045303042945884933, 'samples': 2211840, 'steps': 11519, 'loss/train': 1.1198786795139313} 01/29/2022 03:38:52 - INFO - codeparrot_training - Step 11520: {'lr': 0.0004530208817229516, 'samples': 2212032, 'steps': 11520, 'loss/train': 1.8884380459785461} 01/29/2022 03:38:58 - INFO - codeparrot_training - Step 11521: {'lr': 0.00045301133311737685, 'samples': 2212224, 'steps': 11521, 'loss/train': 1.7276698350906372} 01/29/2022 03:39:02 - INFO - codeparrot_training - Step 11522: {'lr': 0.00045300178364216605, 'samples': 2212416, 'steps': 11522, 'loss/train': 1.695273756980896} 01/29/2022 03:39:06 - INFO - codeparrot_training - Step 11523: {'lr': 0.00045299223329736004, 'samples': 2212608, 'steps': 11523, 'loss/train': 1.841098666191101} 01/29/2022 03:39:10 - INFO - codeparrot_training - Step 11524: {'lr': 0.00045298268208299983, 'samples': 2212800, 'steps': 11524, 'loss/train': 1.091820627450943} 01/29/2022 03:39:14 - INFO - codeparrot_training - Step 11525: {'lr': 0.0004529731299991262, 'samples': 2212992, 'steps': 11525, 'loss/train': 0.27011990547180176} 01/29/2022 03:39:20 - INFO - codeparrot_training - Step 11526: {'lr': 0.00045296357704578016, 'samples': 2213184, 'steps': 11526, 'loss/train': 1.1928907334804535} 01/29/2022 03:39:24 - INFO - codeparrot_training - Step 11527: {'lr': 0.0004529540232230026, 'samples': 2213376, 'steps': 11527, 'loss/train': 2.055636942386627} 01/29/2022 03:39:28 - INFO - codeparrot_training - Step 11528: {'lr': 0.00045294446853083446, 'samples': 2213568, 'steps': 11528, 'loss/train': 1.8869585394859314} 01/29/2022 03:39:33 - INFO - codeparrot_training - Step 11529: {'lr': 0.0004529349129693166, 'samples': 2213760, 'steps': 11529, 'loss/train': 0.8147422671318054} 01/29/2022 03:39:37 - INFO - codeparrot_training - Step 11530: {'lr': 0.0004529253565384901, 'samples': 2213952, 'steps': 11530, 'loss/train': 0.5644413828849792} 01/29/2022 03:39:43 - INFO - codeparrot_training - Step 11531: {'lr': 0.00045291579923839576, 'samples': 2214144, 'steps': 11531, 'loss/train': 1.9052406549453735} 01/29/2022 03:39:47 - INFO - codeparrot_training - Step 11532: {'lr': 0.0004529062410690745, 'samples': 2214336, 'steps': 11532, 'loss/train': 1.3981282711029053} 01/29/2022 03:39:51 - INFO - codeparrot_training - Step 11533: {'lr': 0.00045289668203056743, 'samples': 2214528, 'steps': 11533, 'loss/train': 2.05683034658432} 01/29/2022 03:39:56 - INFO - codeparrot_training - Step 11534: {'lr': 0.00045288712212291537, 'samples': 2214720, 'steps': 11534, 'loss/train': 1.0307190120220184} 01/29/2022 03:40:00 - INFO - codeparrot_training - Step 11535: {'lr': 0.0004528775613461593, 'samples': 2214912, 'steps': 11535, 'loss/train': 0.6906991750001907} 01/29/2022 03:40:05 - INFO - codeparrot_training - Step 11536: {'lr': 0.0004528679997003403, 'samples': 2215104, 'steps': 11536, 'loss/train': 1.8677878975868225} 01/29/2022 03:40:09 - INFO - codeparrot_training - Step 11537: {'lr': 0.000452858437185499, 'samples': 2215296, 'steps': 11537, 'loss/train': 2.5113807320594788} 01/29/2022 03:40:14 - INFO - codeparrot_training - Step 11538: {'lr': 0.00045284887380167674, 'samples': 2215488, 'steps': 11538, 'loss/train': 1.384688526391983} 01/29/2022 03:40:18 - INFO - codeparrot_training - Step 11539: {'lr': 0.0004528393095489142, 'samples': 2215680, 'steps': 11539, 'loss/train': 2.0553991198539734} 01/29/2022 03:40:22 - INFO - codeparrot_training - Step 11540: {'lr': 0.0004528297444272525, 'samples': 2215872, 'steps': 11540, 'loss/train': 1.9386489987373352} 01/29/2022 03:40:27 - INFO - codeparrot_training - Step 11541: {'lr': 0.0004528201784367326, 'samples': 2216064, 'steps': 11541, 'loss/train': 1.9510772824287415} 01/29/2022 03:40:32 - INFO - codeparrot_training - Step 11542: {'lr': 0.00045281061157739544, 'samples': 2216256, 'steps': 11542, 'loss/train': 1.6556913256645203} 01/29/2022 03:40:36 - INFO - codeparrot_training - Step 11543: {'lr': 0.000452801043849282, 'samples': 2216448, 'steps': 11543, 'loss/train': 1.6003734469413757} 01/29/2022 03:40:40 - INFO - codeparrot_training - Step 11544: {'lr': 0.00045279147525243335, 'samples': 2216640, 'steps': 11544, 'loss/train': 1.241206169128418} 01/29/2022 03:40:44 - INFO - codeparrot_training - Step 11545: {'lr': 0.0004527819057868904, 'samples': 2216832, 'steps': 11545, 'loss/train': 0.257930226624012} 01/29/2022 03:40:50 - INFO - codeparrot_training - Step 11546: {'lr': 0.00045277233545269415, 'samples': 2217024, 'steps': 11546, 'loss/train': 1.9809819459915161} 01/29/2022 03:40:55 - INFO - codeparrot_training - Step 11547: {'lr': 0.00045276276424988554, 'samples': 2217216, 'steps': 11547, 'loss/train': 1.46727454662323} 01/29/2022 03:40:59 - INFO - codeparrot_training - Step 11548: {'lr': 0.0004527531921785057, 'samples': 2217408, 'steps': 11548, 'loss/train': 1.4627668261528015} 01/29/2022 03:41:03 - INFO - codeparrot_training - Step 11549: {'lr': 0.00045274361923859554, 'samples': 2217600, 'steps': 11549, 'loss/train': 1.6217828392982483} 01/29/2022 03:41:08 - INFO - codeparrot_training - Step 11550: {'lr': 0.0004527340454301961, 'samples': 2217792, 'steps': 11550, 'loss/train': 1.868097960948944} 01/29/2022 03:41:13 - INFO - codeparrot_training - Step 11551: {'lr': 0.0004527244707533483, 'samples': 2217984, 'steps': 11551, 'loss/train': 2.1133185625076294} 01/29/2022 03:41:17 - INFO - codeparrot_training - Step 11552: {'lr': 0.00045271489520809337, 'samples': 2218176, 'steps': 11552, 'loss/train': 2.354474902153015} 01/29/2022 03:41:21 - INFO - codeparrot_training - Step 11553: {'lr': 0.0004527053187944722, 'samples': 2218368, 'steps': 11553, 'loss/train': 1.5276870131492615} 01/29/2022 03:41:25 - INFO - codeparrot_training - Step 11554: {'lr': 0.00045269574151252567, 'samples': 2218560, 'steps': 11554, 'loss/train': 2.4152047634124756} 01/29/2022 03:41:32 - INFO - codeparrot_training - Step 11555: {'lr': 0.00045268616336229504, 'samples': 2218752, 'steps': 11555, 'loss/train': 1.8997766375541687} 01/29/2022 03:41:36 - INFO - codeparrot_training - Step 11556: {'lr': 0.0004526765843438213, 'samples': 2218944, 'steps': 11556, 'loss/train': 1.8911840915679932} 01/29/2022 03:41:40 - INFO - codeparrot_training - Step 11557: {'lr': 0.0004526670044571454, 'samples': 2219136, 'steps': 11557, 'loss/train': 0.5170216262340546} 01/29/2022 03:41:44 - INFO - codeparrot_training - Step 11558: {'lr': 0.00045265742370230835, 'samples': 2219328, 'steps': 11558, 'loss/train': 1.6941171884536743} 01/29/2022 03:41:49 - INFO - codeparrot_training - Step 11559: {'lr': 0.00045264784207935127, 'samples': 2219520, 'steps': 11559, 'loss/train': 1.6535139083862305} 01/29/2022 03:41:53 - INFO - codeparrot_training - Step 11560: {'lr': 0.0004526382595883152, 'samples': 2219712, 'steps': 11560, 'loss/train': 2.3828371167182922} 01/29/2022 03:41:59 - INFO - codeparrot_training - Step 11561: {'lr': 0.0004526286762292411, 'samples': 2219904, 'steps': 11561, 'loss/train': 1.5368338823318481} 01/29/2022 03:42:03 - INFO - codeparrot_training - Step 11562: {'lr': 0.00045261909200217023, 'samples': 2220096, 'steps': 11562, 'loss/train': 0.7625618577003479} 01/29/2022 03:42:07 - INFO - codeparrot_training - Step 11563: {'lr': 0.0004526095069071434, 'samples': 2220288, 'steps': 11563, 'loss/train': 1.8571668863296509} 01/29/2022 03:42:12 - INFO - codeparrot_training - Step 11564: {'lr': 0.0004525999209442018, 'samples': 2220480, 'steps': 11564, 'loss/train': 1.9137813448905945} 01/29/2022 03:42:16 - INFO - codeparrot_training - Step 11565: {'lr': 0.0004525903341133865, 'samples': 2220672, 'steps': 11565, 'loss/train': 1.6563110947608948} 01/29/2022 03:42:20 - INFO - codeparrot_training - Step 11566: {'lr': 0.0004525807464147385, 'samples': 2220864, 'steps': 11566, 'loss/train': 1.46744504570961} 01/29/2022 03:42:25 - INFO - codeparrot_training - Step 11567: {'lr': 0.00045257115784829897, 'samples': 2221056, 'steps': 11567, 'loss/train': 1.7691991925239563} 01/29/2022 03:42:30 - INFO - codeparrot_training - Step 11568: {'lr': 0.00045256156841410884, 'samples': 2221248, 'steps': 11568, 'loss/train': 2.1527522206306458} 01/29/2022 03:42:34 - INFO - codeparrot_training - Step 11569: {'lr': 0.0004525519781122093, 'samples': 2221440, 'steps': 11569, 'loss/train': 1.5633392930030823} 01/29/2022 03:42:38 - INFO - codeparrot_training - Step 11570: {'lr': 0.00045254238694264145, 'samples': 2221632, 'steps': 11570, 'loss/train': 1.8884208798408508} 01/29/2022 03:42:42 - INFO - codeparrot_training - Step 11571: {'lr': 0.00045253279490544627, 'samples': 2221824, 'steps': 11571, 'loss/train': 2.661360025405884} 01/29/2022 03:42:48 - INFO - codeparrot_training - Step 11572: {'lr': 0.0004525232020006649, 'samples': 2222016, 'steps': 11572, 'loss/train': 1.722945749759674} 01/29/2022 03:42:52 - INFO - codeparrot_training - Step 11573: {'lr': 0.00045251360822833855, 'samples': 2222208, 'steps': 11573, 'loss/train': 1.041934221982956} 01/29/2022 03:42:56 - INFO - codeparrot_training - Step 11574: {'lr': 0.00045250401358850814, 'samples': 2222400, 'steps': 11574, 'loss/train': 1.493066668510437} 01/29/2022 03:43:00 - INFO - codeparrot_training - Step 11575: {'lr': 0.00045249441808121484, 'samples': 2222592, 'steps': 11575, 'loss/train': 0.41400258243083954} 01/29/2022 03:43:05 - INFO - codeparrot_training - Step 11576: {'lr': 0.0004524848217064997, 'samples': 2222784, 'steps': 11576, 'loss/train': 1.7531720995903015} 01/29/2022 03:43:11 - INFO - codeparrot_training - Step 11577: {'lr': 0.0004524752244644039, 'samples': 2222976, 'steps': 11577, 'loss/train': 1.6396799683570862} 01/29/2022 03:43:15 - INFO - codeparrot_training - Step 11578: {'lr': 0.0004524656263549686, 'samples': 2223168, 'steps': 11578, 'loss/train': 2.051666736602783} 01/29/2022 03:43:19 - INFO - codeparrot_training - Step 11579: {'lr': 0.0004524560273782348, 'samples': 2223360, 'steps': 11579, 'loss/train': 0.8038114607334137} 01/29/2022 03:43:23 - INFO - codeparrot_training - Step 11580: {'lr': 0.00045244642753424364, 'samples': 2223552, 'steps': 11580, 'loss/train': 2.149272859096527} 01/29/2022 03:43:28 - INFO - codeparrot_training - Step 11581: {'lr': 0.0004524368268230363, 'samples': 2223744, 'steps': 11581, 'loss/train': 2.062819540500641} 01/29/2022 03:43:33 - INFO - codeparrot_training - Step 11582: {'lr': 0.00045242722524465386, 'samples': 2223936, 'steps': 11582, 'loss/train': 1.4595241248607635} 01/29/2022 03:43:37 - INFO - codeparrot_training - Step 11583: {'lr': 0.00045241762279913745, 'samples': 2224128, 'steps': 11583, 'loss/train': 1.4206110835075378} 01/29/2022 03:43:42 - INFO - codeparrot_training - Step 11584: {'lr': 0.0004524080194865283, 'samples': 2224320, 'steps': 11584, 'loss/train': 0.4645995944738388} 01/29/2022 03:43:46 - INFO - codeparrot_training - Step 11585: {'lr': 0.00045239841530686736, 'samples': 2224512, 'steps': 11585, 'loss/train': 1.8633236289024353} 01/29/2022 03:43:50 - INFO - codeparrot_training - Step 11586: {'lr': 0.000452388810260196, 'samples': 2224704, 'steps': 11586, 'loss/train': 1.1472111940383911} 01/29/2022 03:43:55 - INFO - codeparrot_training - Step 11587: {'lr': 0.0004523792043465551, 'samples': 2224896, 'steps': 11587, 'loss/train': 1.1976699829101562} 01/29/2022 03:43:59 - INFO - codeparrot_training - Step 11588: {'lr': 0.00045236959756598605, 'samples': 2225088, 'steps': 11588, 'loss/train': 1.6688445210456848} 01/29/2022 03:44:04 - INFO - codeparrot_training - Step 11589: {'lr': 0.0004523599899185299, 'samples': 2225280, 'steps': 11589, 'loss/train': 1.4039941728115082} 01/29/2022 03:44:08 - INFO - codeparrot_training - Step 11590: {'lr': 0.0004523503814042277, 'samples': 2225472, 'steps': 11590, 'loss/train': 1.3673027157783508} 01/29/2022 03:44:12 - INFO - codeparrot_training - Step 11591: {'lr': 0.00045234077202312086, 'samples': 2225664, 'steps': 11591, 'loss/train': 1.0380185544490814} 01/29/2022 03:44:18 - INFO - codeparrot_training - Step 11592: {'lr': 0.00045233116177525036, 'samples': 2225856, 'steps': 11592, 'loss/train': 2.0191449522972107} 01/29/2022 03:44:23 - INFO - codeparrot_training - Step 11593: {'lr': 0.00045232155066065737, 'samples': 2226048, 'steps': 11593, 'loss/train': 1.2876225113868713} 01/29/2022 03:44:27 - INFO - codeparrot_training - Step 11594: {'lr': 0.00045231193867938314, 'samples': 2226240, 'steps': 11594, 'loss/train': 1.8945468664169312} 01/29/2022 03:44:31 - INFO - codeparrot_training - Step 11595: {'lr': 0.0004523023258314688, 'samples': 2226432, 'steps': 11595, 'loss/train': 1.4390924870967865} 01/29/2022 03:44:37 - INFO - codeparrot_training - Step 11596: {'lr': 0.00045229271211695554, 'samples': 2226624, 'steps': 11596, 'loss/train': 1.5654398202896118} 01/29/2022 03:44:41 - INFO - codeparrot_training - Step 11597: {'lr': 0.00045228309753588447, 'samples': 2226816, 'steps': 11597, 'loss/train': 1.8892434239387512} 01/29/2022 03:44:45 - INFO - codeparrot_training - Step 11598: {'lr': 0.0004522734820882969, 'samples': 2227008, 'steps': 11598, 'loss/train': 1.2479358315467834} 01/29/2022 03:44:49 - INFO - codeparrot_training - Step 11599: {'lr': 0.00045226386577423394, 'samples': 2227200, 'steps': 11599, 'loss/train': 0.5949200391769409} 01/29/2022 03:44:54 - INFO - codeparrot_training - Step 11600: {'lr': 0.0004522542485937369, 'samples': 2227392, 'steps': 11600, 'loss/train': 1.616719365119934} 01/29/2022 03:44:58 - INFO - codeparrot_training - Step 11601: {'lr': 0.0004522446305468468, 'samples': 2227584, 'steps': 11601, 'loss/train': 1.601546823978424} 01/29/2022 03:45:04 - INFO - codeparrot_training - Step 11602: {'lr': 0.00045223501163360494, 'samples': 2227776, 'steps': 11602, 'loss/train': 1.815596580505371} 01/29/2022 03:45:08 - INFO - codeparrot_training - Step 11603: {'lr': 0.0004522253918540524, 'samples': 2227968, 'steps': 11603, 'loss/train': 1.347379982471466} 01/29/2022 03:45:13 - INFO - codeparrot_training - Step 11604: {'lr': 0.00045221577120823064, 'samples': 2228160, 'steps': 11604, 'loss/train': 2.0959120988845825} 01/29/2022 03:45:17 - INFO - codeparrot_training - Step 11605: {'lr': 0.00045220614969618066, 'samples': 2228352, 'steps': 11605, 'loss/train': 1.6772843599319458} 01/29/2022 03:45:21 - INFO - codeparrot_training - Step 11606: {'lr': 0.0004521965273179438, 'samples': 2228544, 'steps': 11606, 'loss/train': 1.5348699688911438} 01/29/2022 03:45:26 - INFO - codeparrot_training - Step 11607: {'lr': 0.00045218690407356117, 'samples': 2228736, 'steps': 11607, 'loss/train': 1.7249731421470642} 01/29/2022 03:45:31 - INFO - codeparrot_training - Step 11608: {'lr': 0.00045217727996307405, 'samples': 2228928, 'steps': 11608, 'loss/train': 1.6836007833480835} 01/29/2022 03:45:35 - INFO - codeparrot_training - Step 11609: {'lr': 0.0004521676549865237, 'samples': 2229120, 'steps': 11609, 'loss/train': 1.3726352155208588} 01/29/2022 03:45:39 - INFO - codeparrot_training - Step 11610: {'lr': 0.0004521580291439513, 'samples': 2229312, 'steps': 11610, 'loss/train': 0.9582113921642303} 01/29/2022 03:45:43 - INFO - codeparrot_training - Step 11611: {'lr': 0.00045214840243539803, 'samples': 2229504, 'steps': 11611, 'loss/train': 2.1285964250564575} 01/29/2022 03:45:49 - INFO - codeparrot_training - Step 11612: {'lr': 0.00045213877486090524, 'samples': 2229696, 'steps': 11612, 'loss/train': 1.8092120289802551} 01/29/2022 03:45:53 - INFO - codeparrot_training - Step 11613: {'lr': 0.0004521291464205141, 'samples': 2229888, 'steps': 11613, 'loss/train': 1.755541741847992} 01/29/2022 03:45:57 - INFO - codeparrot_training - Step 11614: {'lr': 0.0004521195171142659, 'samples': 2230080, 'steps': 11614, 'loss/train': 2.124791979789734} 01/29/2022 03:46:02 - INFO - codeparrot_training - Step 11615: {'lr': 0.0004521098869422019, 'samples': 2230272, 'steps': 11615, 'loss/train': 2.8610660433769226} 01/29/2022 03:46:08 - INFO - codeparrot_training - Step 11616: {'lr': 0.00045210025590436333, 'samples': 2230464, 'steps': 11616, 'loss/train': 3.287370800971985} 01/29/2022 03:46:12 - INFO - codeparrot_training - Step 11617: {'lr': 0.00045209062400079135, 'samples': 2230656, 'steps': 11617, 'loss/train': 1.826950192451477} 01/29/2022 03:46:16 - INFO - codeparrot_training - Step 11618: {'lr': 0.00045208099123152735, 'samples': 2230848, 'steps': 11618, 'loss/train': 2.000694751739502} 01/29/2022 03:46:20 - INFO - codeparrot_training - Step 11619: {'lr': 0.00045207135759661255, 'samples': 2231040, 'steps': 11619, 'loss/train': 1.8390271067619324} 01/29/2022 03:46:25 - INFO - codeparrot_training - Step 11620: {'lr': 0.0004520617230960883, 'samples': 2231232, 'steps': 11620, 'loss/train': 1.8134998083114624} 01/29/2022 03:46:30 - INFO - codeparrot_training - Step 11621: {'lr': 0.0004520520877299957, 'samples': 2231424, 'steps': 11621, 'loss/train': 2.2335509061813354} 01/29/2022 03:46:34 - INFO - codeparrot_training - Step 11622: {'lr': 0.00045204245149837606, 'samples': 2231616, 'steps': 11622, 'loss/train': 2.0407794713974} 01/29/2022 03:46:38 - INFO - codeparrot_training - Step 11623: {'lr': 0.00045203281440127087, 'samples': 2231808, 'steps': 11623, 'loss/train': 2.009180724620819} 01/29/2022 03:46:43 - INFO - codeparrot_training - Step 11624: {'lr': 0.00045202317643872113, 'samples': 2232000, 'steps': 11624, 'loss/train': 1.8293103575706482} 01/29/2022 03:46:47 - INFO - codeparrot_training - Step 11625: {'lr': 0.0004520135376107683, 'samples': 2232192, 'steps': 11625, 'loss/train': 2.0312072038650513} 01/29/2022 03:46:52 - INFO - codeparrot_training - Step 11626: {'lr': 0.00045200389791745364, 'samples': 2232384, 'steps': 11626, 'loss/train': 1.7986080050468445} 01/29/2022 03:46:57 - INFO - codeparrot_training - Step 11627: {'lr': 0.0004519942573588184, 'samples': 2232576, 'steps': 11627, 'loss/train': 1.341327577829361} 01/29/2022 03:47:01 - INFO - codeparrot_training - Step 11628: {'lr': 0.00045198461593490394, 'samples': 2232768, 'steps': 11628, 'loss/train': 2.2959278225898743} 01/29/2022 03:47:05 - INFO - codeparrot_training - Step 11629: {'lr': 0.0004519749736457515, 'samples': 2232960, 'steps': 11629, 'loss/train': 0.6822169572114944} 01/29/2022 03:47:09 - INFO - codeparrot_training - Step 11630: {'lr': 0.00045196533049140234, 'samples': 2233152, 'steps': 11630, 'loss/train': 1.916013479232788} 01/29/2022 03:47:15 - INFO - codeparrot_training - Step 11631: {'lr': 0.0004519556864718979, 'samples': 2233344, 'steps': 11631, 'loss/train': 2.334122121334076} 01/29/2022 03:47:19 - INFO - codeparrot_training - Step 11632: {'lr': 0.00045194604158727936, 'samples': 2233536, 'steps': 11632, 'loss/train': 1.8442962169647217} 01/29/2022 03:47:23 - INFO - codeparrot_training - Step 11633: {'lr': 0.0004519363958375882, 'samples': 2233728, 'steps': 11633, 'loss/train': 2.121142029762268} 01/29/2022 03:47:27 - INFO - codeparrot_training - Step 11634: {'lr': 0.00045192674922286556, 'samples': 2233920, 'steps': 11634, 'loss/train': 2.2375452518463135} 01/29/2022 03:47:32 - INFO - codeparrot_training - Step 11635: {'lr': 0.00045191710174315294, 'samples': 2234112, 'steps': 11635, 'loss/train': 2.175260066986084} 01/29/2022 03:47:37 - INFO - codeparrot_training - Step 11636: {'lr': 0.0004519074533984915, 'samples': 2234304, 'steps': 11636, 'loss/train': 1.5828561186790466} 01/29/2022 03:47:42 - INFO - codeparrot_training - Step 11637: {'lr': 0.0004518978041889227, 'samples': 2234496, 'steps': 11637, 'loss/train': 1.8344187140464783} 01/29/2022 03:47:46 - INFO - codeparrot_training - Step 11638: {'lr': 0.00045188815411448767, 'samples': 2234688, 'steps': 11638, 'loss/train': 1.629212200641632} 01/29/2022 03:47:50 - INFO - codeparrot_training - Step 11639: {'lr': 0.00045187850317522806, 'samples': 2234880, 'steps': 11639, 'loss/train': 1.5754020810127258} 01/29/2022 03:47:55 - INFO - codeparrot_training - Step 11640: {'lr': 0.00045186885137118494, 'samples': 2235072, 'steps': 11640, 'loss/train': 1.2553174495697021} 01/29/2022 03:48:00 - INFO - codeparrot_training - Step 11641: {'lr': 0.0004518591987023999, 'samples': 2235264, 'steps': 11641, 'loss/train': 1.9083808064460754} 01/29/2022 03:48:04 - INFO - codeparrot_training - Step 11642: {'lr': 0.000451849545168914, 'samples': 2235456, 'steps': 11642, 'loss/train': 1.9370334148406982} 01/29/2022 03:48:08 - INFO - codeparrot_training - Step 11643: {'lr': 0.00045183989077076883, 'samples': 2235648, 'steps': 11643, 'loss/train': 1.807543694972992} 01/29/2022 03:48:12 - INFO - codeparrot_training - Step 11644: {'lr': 0.00045183023550800564, 'samples': 2235840, 'steps': 11644, 'loss/train': 1.05147984623909} 01/29/2022 03:48:17 - INFO - codeparrot_training - Step 11645: {'lr': 0.0004518205793806658, 'samples': 2236032, 'steps': 11645, 'loss/train': 1.7187675833702087} 01/29/2022 03:48:23 - INFO - codeparrot_training - Step 11646: {'lr': 0.0004518109223887907, 'samples': 2236224, 'steps': 11646, 'loss/train': 1.7353856563568115} 01/29/2022 03:48:27 - INFO - codeparrot_training - Step 11647: {'lr': 0.0004518012645324217, 'samples': 2236416, 'steps': 11647, 'loss/train': 1.289982408285141} 01/29/2022 03:48:31 - INFO - codeparrot_training - Step 11648: {'lr': 0.00045179160581160005, 'samples': 2236608, 'steps': 11648, 'loss/train': 2.4607195258140564} 01/29/2022 03:48:35 - INFO - codeparrot_training - Step 11649: {'lr': 0.0004517819462263674, 'samples': 2236800, 'steps': 11649, 'loss/train': 1.6603290438652039} 01/29/2022 03:48:40 - INFO - codeparrot_training - Step 11650: {'lr': 0.0004517722857767649, 'samples': 2236992, 'steps': 11650, 'loss/train': 2.012375056743622} 01/29/2022 03:48:45 - INFO - codeparrot_training - Step 11651: {'lr': 0.0004517626244628339, 'samples': 2237184, 'steps': 11651, 'loss/train': 1.2570271790027618} 01/29/2022 03:48:49 - INFO - codeparrot_training - Step 11652: {'lr': 0.000451752962284616, 'samples': 2237376, 'steps': 11652, 'loss/train': 2.2056110501289368} 01/29/2022 03:48:53 - INFO - codeparrot_training - Step 11653: {'lr': 0.0004517432992421524, 'samples': 2237568, 'steps': 11653, 'loss/train': 2.248883306980133} 01/29/2022 03:48:58 - INFO - codeparrot_training - Step 11654: {'lr': 0.00045173363533548464, 'samples': 2237760, 'steps': 11654, 'loss/train': 1.7722352743148804} 01/29/2022 03:49:02 - INFO - codeparrot_training - Step 11655: {'lr': 0.00045172397056465405, 'samples': 2237952, 'steps': 11655, 'loss/train': 2.0896312594413757} 01/29/2022 03:49:07 - INFO - codeparrot_training - Step 11656: {'lr': 0.000451714304929702, 'samples': 2238144, 'steps': 11656, 'loss/train': 1.776763379573822} 01/29/2022 03:49:12 - INFO - codeparrot_training - Step 11657: {'lr': 0.0004517046384306699, 'samples': 2238336, 'steps': 11657, 'loss/train': 1.6245907545089722} 01/29/2022 03:49:16 - INFO - codeparrot_training - Step 11658: {'lr': 0.00045169497106759915, 'samples': 2238528, 'steps': 11658, 'loss/train': 0.2881436124444008} 01/29/2022 03:49:20 - INFO - codeparrot_training - Step 11659: {'lr': 0.0004516853028405312, 'samples': 2238720, 'steps': 11659, 'loss/train': 1.9458657503128052} 01/29/2022 03:49:24 - INFO - codeparrot_training - Step 11660: {'lr': 0.0004516756337495075, 'samples': 2238912, 'steps': 11660, 'loss/train': 1.7080069184303284} 01/29/2022 03:49:30 - INFO - codeparrot_training - Step 11661: {'lr': 0.00045166596379456935, 'samples': 2239104, 'steps': 11661, 'loss/train': 1.5387911796569824} 01/29/2022 03:49:35 - INFO - codeparrot_training - Step 11662: {'lr': 0.0004516562929757584, 'samples': 2239296, 'steps': 11662, 'loss/train': 0.9563243687152863} 01/29/2022 03:49:39 - INFO - codeparrot_training - Step 11663: {'lr': 0.0004516466212931158, 'samples': 2239488, 'steps': 11663, 'loss/train': 1.5598967671394348} 01/29/2022 03:49:43 - INFO - codeparrot_training - Step 11664: {'lr': 0.00045163694874668316, 'samples': 2239680, 'steps': 11664, 'loss/train': 2.2632628083229065} 01/29/2022 03:49:47 - INFO - codeparrot_training - Step 11665: {'lr': 0.0004516272753365018, 'samples': 2239872, 'steps': 11665, 'loss/train': 1.5191293358802795} 01/29/2022 03:49:53 - INFO - codeparrot_training - Step 11666: {'lr': 0.0004516176010626132, 'samples': 2240064, 'steps': 11666, 'loss/train': 1.396106779575348} 01/29/2022 03:49:57 - INFO - codeparrot_training - Step 11667: {'lr': 0.00045160792592505893, 'samples': 2240256, 'steps': 11667, 'loss/train': 2.5991177558898926} 01/29/2022 03:50:01 - INFO - codeparrot_training - Step 11668: {'lr': 0.0004515982499238802, 'samples': 2240448, 'steps': 11668, 'loss/train': 2.2428946495056152} 01/29/2022 03:50:06 - INFO - codeparrot_training - Step 11669: {'lr': 0.0004515885730591187, 'samples': 2240640, 'steps': 11669, 'loss/train': 1.5167414546012878} 01/29/2022 03:50:10 - INFO - codeparrot_training - Step 11670: {'lr': 0.0004515788953308156, 'samples': 2240832, 'steps': 11670, 'loss/train': 1.792053759098053} 01/29/2022 03:50:15 - INFO - codeparrot_training - Step 11671: {'lr': 0.00045156921673901267, 'samples': 2241024, 'steps': 11671, 'loss/train': 1.7223637104034424} 01/29/2022 03:50:19 - INFO - codeparrot_training - Step 11672: {'lr': 0.0004515595372837512, 'samples': 2241216, 'steps': 11672, 'loss/train': 2.8566774129867554} 01/29/2022 03:50:23 - INFO - codeparrot_training - Step 11673: {'lr': 0.00045154985696507267, 'samples': 2241408, 'steps': 11673, 'loss/train': 2.7782020568847656} 01/29/2022 03:50:28 - INFO - codeparrot_training - Step 11674: {'lr': 0.0004515401757830185, 'samples': 2241600, 'steps': 11674, 'loss/train': 1.424354374408722} 01/29/2022 03:50:32 - INFO - codeparrot_training - Step 11675: {'lr': 0.0004515304937376302, 'samples': 2241792, 'steps': 11675, 'loss/train': 0.21856462955474854} 01/29/2022 03:50:38 - INFO - codeparrot_training - Step 11676: {'lr': 0.00045152081082894935, 'samples': 2241984, 'steps': 11676, 'loss/train': 1.7401978969573975} 01/29/2022 03:50:42 - INFO - codeparrot_training - Step 11677: {'lr': 0.00045151112705701723, 'samples': 2242176, 'steps': 11677, 'loss/train': 2.2711211442947388} 01/29/2022 03:50:46 - INFO - codeparrot_training - Step 11678: {'lr': 0.00045150144242187554, 'samples': 2242368, 'steps': 11678, 'loss/train': 1.3939976692199707} 01/29/2022 03:50:51 - INFO - codeparrot_training - Step 11679: {'lr': 0.0004514917569235656, 'samples': 2242560, 'steps': 11679, 'loss/train': 2.6870094537734985} 01/29/2022 03:50:55 - INFO - codeparrot_training - Step 11680: {'lr': 0.00045148207056212896, 'samples': 2242752, 'steps': 11680, 'loss/train': 1.9541296362876892} 01/29/2022 03:51:00 - INFO - codeparrot_training - Step 11681: {'lr': 0.0004514723833376071, 'samples': 2242944, 'steps': 11681, 'loss/train': 1.4943962395191193} 01/29/2022 03:51:04 - INFO - codeparrot_training - Step 11682: {'lr': 0.00045146269525004153, 'samples': 2243136, 'steps': 11682, 'loss/train': 1.3963565826416016} 01/29/2022 03:51:09 - INFO - codeparrot_training - Step 11683: {'lr': 0.00045145300629947374, 'samples': 2243328, 'steps': 11683, 'loss/train': 1.3792707324028015} 01/29/2022 03:51:13 - INFO - codeparrot_training - Step 11684: {'lr': 0.0004514433164859453, 'samples': 2243520, 'steps': 11684, 'loss/train': 1.4764477908611298} 01/29/2022 03:51:17 - INFO - codeparrot_training - Step 11685: {'lr': 0.00045143362580949754, 'samples': 2243712, 'steps': 11685, 'loss/train': 2.7569430470466614} 01/29/2022 03:51:23 - INFO - codeparrot_training - Step 11686: {'lr': 0.00045142393427017214, 'samples': 2243904, 'steps': 11686, 'loss/train': 0.4210474044084549} 01/29/2022 03:51:27 - INFO - codeparrot_training - Step 11687: {'lr': 0.0004514142418680106, 'samples': 2244096, 'steps': 11687, 'loss/train': 1.5017101764678955} 01/29/2022 03:51:31 - INFO - codeparrot_training - Step 11688: {'lr': 0.00045140454860305435, 'samples': 2244288, 'steps': 11688, 'loss/train': 2.39946448802948} 01/29/2022 03:51:35 - INFO - codeparrot_training - Step 11689: {'lr': 0.000451394854475345, 'samples': 2244480, 'steps': 11689, 'loss/train': 1.3913597166538239} 01/29/2022 03:51:40 - INFO - codeparrot_training - Step 11690: {'lr': 0.0004513851594849241, 'samples': 2244672, 'steps': 11690, 'loss/train': 1.9571685791015625} 01/29/2022 03:51:45 - INFO - codeparrot_training - Step 11691: {'lr': 0.000451375463631833, 'samples': 2244864, 'steps': 11691, 'loss/train': 1.754383385181427} 01/29/2022 03:51:49 - INFO - codeparrot_training - Step 11692: {'lr': 0.0004513657669161134, 'samples': 2245056, 'steps': 11692, 'loss/train': 2.1099992394447327} 01/29/2022 03:51:53 - INFO - codeparrot_training - Step 11693: {'lr': 0.0004513560693378068, 'samples': 2245248, 'steps': 11693, 'loss/train': 1.4864959716796875} 01/29/2022 03:51:58 - INFO - codeparrot_training - Step 11694: {'lr': 0.00045134637089695484, 'samples': 2245440, 'steps': 11694, 'loss/train': 1.367801159620285} 01/29/2022 03:52:02 - INFO - codeparrot_training - Step 11695: {'lr': 0.0004513366715935988, 'samples': 2245632, 'steps': 11695, 'loss/train': 1.8091877102851868} 01/29/2022 03:52:08 - INFO - codeparrot_training - Step 11696: {'lr': 0.00045132697142778044, 'samples': 2245824, 'steps': 11696, 'loss/train': 1.6879899501800537} 01/29/2022 03:52:12 - INFO - codeparrot_training - Step 11697: {'lr': 0.00045131727039954137, 'samples': 2246016, 'steps': 11697, 'loss/train': 1.8613951206207275} 01/29/2022 03:52:16 - INFO - codeparrot_training - Step 11698: {'lr': 0.00045130756850892296, 'samples': 2246208, 'steps': 11698, 'loss/train': 1.4788760840892792} 01/29/2022 03:52:21 - INFO - codeparrot_training - Step 11699: {'lr': 0.00045129786575596683, 'samples': 2246400, 'steps': 11699, 'loss/train': 1.75091153383255} 01/29/2022 03:52:25 - INFO - codeparrot_training - Step 11700: {'lr': 0.00045128816214071453, 'samples': 2246592, 'steps': 11700, 'loss/train': 1.715316653251648} 01/29/2022 03:52:30 - INFO - codeparrot_training - Step 11701: {'lr': 0.00045127845766320773, 'samples': 2246784, 'steps': 11701, 'loss/train': 0.5538788437843323} 01/29/2022 03:52:34 - INFO - codeparrot_training - Step 11702: {'lr': 0.0004512687523234879, 'samples': 2246976, 'steps': 11702, 'loss/train': 0.7013111561536789} 01/29/2022 03:52:39 - INFO - codeparrot_training - Step 11703: {'lr': 0.0004512590461215967, 'samples': 2247168, 'steps': 11703, 'loss/train': 1.0400696396827698} 01/29/2022 03:52:43 - INFO - codeparrot_training - Step 11704: {'lr': 0.0004512493390575756, 'samples': 2247360, 'steps': 11704, 'loss/train': 1.6858996152877808} 01/29/2022 03:52:47 - INFO - codeparrot_training - Step 11705: {'lr': 0.0004512396311314662, 'samples': 2247552, 'steps': 11705, 'loss/train': 2.3113653659820557} 01/29/2022 03:52:53 - INFO - codeparrot_training - Step 11706: {'lr': 0.00045122992234331017, 'samples': 2247744, 'steps': 11706, 'loss/train': 0.7203619927167892} 01/29/2022 03:52:58 - INFO - codeparrot_training - Step 11707: {'lr': 0.00045122021269314907, 'samples': 2247936, 'steps': 11707, 'loss/train': 1.0785252749919891} 01/29/2022 03:53:02 - INFO - codeparrot_training - Step 11708: {'lr': 0.0004512105021810244, 'samples': 2248128, 'steps': 11708, 'loss/train': 2.4668864607810974} 01/29/2022 03:53:06 - INFO - codeparrot_training - Step 11709: {'lr': 0.0004512007908069779, 'samples': 2248320, 'steps': 11709, 'loss/train': 1.2142994105815887} 01/29/2022 03:53:10 - INFO - codeparrot_training - Step 11710: {'lr': 0.0004511910785710511, 'samples': 2248512, 'steps': 11710, 'loss/train': 6.365785360336304} 01/29/2022 03:53:16 - INFO - codeparrot_training - Step 11711: {'lr': 0.0004511813654732856, 'samples': 2248704, 'steps': 11711, 'loss/train': 2.332670509815216} 01/29/2022 03:53:20 - INFO - codeparrot_training - Step 11712: {'lr': 0.00045117165151372296, 'samples': 2248896, 'steps': 11712, 'loss/train': 2.003362476825714} 01/29/2022 03:53:24 - INFO - codeparrot_training - Step 11713: {'lr': 0.0004511619366924049, 'samples': 2249088, 'steps': 11713, 'loss/train': 1.6359483003616333} 01/29/2022 03:53:29 - INFO - codeparrot_training - Step 11714: {'lr': 0.00045115222100937293, 'samples': 2249280, 'steps': 11714, 'loss/train': 2.1699026226997375} 01/29/2022 03:53:33 - INFO - codeparrot_training - Step 11715: {'lr': 0.00045114250446466874, 'samples': 2249472, 'steps': 11715, 'loss/train': 2.0642917156219482} 01/29/2022 03:53:38 - INFO - codeparrot_training - Step 11716: {'lr': 0.00045113278705833396, 'samples': 2249664, 'steps': 11716, 'loss/train': 1.9573035836219788} 01/29/2022 03:53:43 - INFO - codeparrot_training - Step 11717: {'lr': 0.00045112306879041016, 'samples': 2249856, 'steps': 11717, 'loss/train': 1.9966163635253906} 01/29/2022 03:53:47 - INFO - codeparrot_training - Step 11718: {'lr': 0.000451113349660939, 'samples': 2250048, 'steps': 11718, 'loss/train': 0.9369330704212189} 01/29/2022 03:53:51 - INFO - codeparrot_training - Step 11719: {'lr': 0.0004511036296699621, 'samples': 2250240, 'steps': 11719, 'loss/train': 1.2158656418323517} 01/29/2022 03:53:55 - INFO - codeparrot_training - Step 11720: {'lr': 0.0004510939088175211, 'samples': 2250432, 'steps': 11720, 'loss/train': 1.8264357447624207} 01/29/2022 03:54:00 - INFO - codeparrot_training - Step 11721: {'lr': 0.00045108418710365774, 'samples': 2250624, 'steps': 11721, 'loss/train': 2.314378023147583} 01/29/2022 03:54:06 - INFO - codeparrot_training - Step 11722: {'lr': 0.0004510744645284135, 'samples': 2250816, 'steps': 11722, 'loss/train': 0.8971821963787079} 01/29/2022 03:54:10 - INFO - codeparrot_training - Step 11723: {'lr': 0.00045106474109183004, 'samples': 2251008, 'steps': 11723, 'loss/train': 2.40278559923172} 01/29/2022 03:54:14 - INFO - codeparrot_training - Step 11724: {'lr': 0.00045105501679394916, 'samples': 2251200, 'steps': 11724, 'loss/train': 2.4256487488746643} 01/29/2022 03:54:19 - INFO - codeparrot_training - Step 11725: {'lr': 0.00045104529163481245, 'samples': 2251392, 'steps': 11725, 'loss/train': 1.5345067977905273} 01/29/2022 03:54:23 - INFO - codeparrot_training - Step 11726: {'lr': 0.0004510355656144615, 'samples': 2251584, 'steps': 11726, 'loss/train': 1.7594595551490784} 01/29/2022 03:54:29 - INFO - codeparrot_training - Step 11727: {'lr': 0.000451025838732938, 'samples': 2251776, 'steps': 11727, 'loss/train': 2.019841432571411} 01/29/2022 03:54:33 - INFO - codeparrot_training - Step 11728: {'lr': 0.0004510161109902837, 'samples': 2251968, 'steps': 11728, 'loss/train': 1.7595469951629639} 01/29/2022 03:54:37 - INFO - codeparrot_training - Step 11729: {'lr': 0.00045100638238654013, 'samples': 2252160, 'steps': 11729, 'loss/train': 2.097174882888794} 01/29/2022 03:54:41 - INFO - codeparrot_training - Step 11730: {'lr': 0.00045099665292174917, 'samples': 2252352, 'steps': 11730, 'loss/train': 0.7048204690217972} 01/29/2022 03:54:46 - INFO - codeparrot_training - Step 11731: {'lr': 0.00045098692259595233, 'samples': 2252544, 'steps': 11731, 'loss/train': 1.3885349035263062} 01/29/2022 03:54:52 - INFO - codeparrot_training - Step 11732: {'lr': 0.00045097719140919126, 'samples': 2252736, 'steps': 11732, 'loss/train': 1.9385270476341248} 01/29/2022 03:54:56 - INFO - codeparrot_training - Step 11733: {'lr': 0.00045096745936150774, 'samples': 2252928, 'steps': 11733, 'loss/train': 1.5837008357048035} 01/29/2022 03:55:00 - INFO - codeparrot_training - Step 11734: {'lr': 0.00045095772645294347, 'samples': 2253120, 'steps': 11734, 'loss/train': 2.564228653907776} 01/29/2022 03:55:04 - INFO - codeparrot_training - Step 11735: {'lr': 0.00045094799268354007, 'samples': 2253312, 'steps': 11735, 'loss/train': 2.64538711309433} 01/29/2022 03:55:09 - INFO - codeparrot_training - Step 11736: {'lr': 0.00045093825805333934, 'samples': 2253504, 'steps': 11736, 'loss/train': 2.130959451198578} 01/29/2022 03:55:14 - INFO - codeparrot_training - Step 11737: {'lr': 0.0004509285225623829, 'samples': 2253696, 'steps': 11737, 'loss/train': 2.595681309700012} 01/29/2022 03:55:18 - INFO - codeparrot_training - Step 11738: {'lr': 0.0004509187862107125, 'samples': 2253888, 'steps': 11738, 'loss/train': 2.3941842913627625} 01/29/2022 03:55:22 - INFO - codeparrot_training - Step 11739: {'lr': 0.0004509090489983697, 'samples': 2254080, 'steps': 11739, 'loss/train': 1.172730803489685} 01/29/2022 03:55:26 - INFO - codeparrot_training - Step 11740: {'lr': 0.0004508993109253964, 'samples': 2254272, 'steps': 11740, 'loss/train': 2.076712667942047} 01/29/2022 03:55:31 - INFO - codeparrot_training - Step 11741: {'lr': 0.00045088957199183427, 'samples': 2254464, 'steps': 11741, 'loss/train': 1.8353835940361023} 01/29/2022 03:55:36 - INFO - codeparrot_training - Step 11742: {'lr': 0.000450879832197725, 'samples': 2254656, 'steps': 11742, 'loss/train': 1.1018444895744324} 01/29/2022 03:55:40 - INFO - codeparrot_training - Step 11743: {'lr': 0.0004508700915431103, 'samples': 2254848, 'steps': 11743, 'loss/train': 1.8858949542045593} 01/29/2022 03:55:44 - INFO - codeparrot_training - Step 11744: {'lr': 0.0004508603500280319, 'samples': 2255040, 'steps': 11744, 'loss/train': 0.7805453538894653} 01/29/2022 03:55:49 - INFO - codeparrot_training - Step 11745: {'lr': 0.00045085060765253157, 'samples': 2255232, 'steps': 11745, 'loss/train': 3.5079957246780396} 01/29/2022 03:55:53 - INFO - codeparrot_training - Step 11746: {'lr': 0.00045084086441665093, 'samples': 2255424, 'steps': 11746, 'loss/train': 2.7722275257110596} 01/29/2022 03:55:58 - INFO - codeparrot_training - Step 11747: {'lr': 0.00045083112032043196, 'samples': 2255616, 'steps': 11747, 'loss/train': 1.9661893844604492} 01/29/2022 03:56:02 - INFO - codeparrot_training - Step 11748: {'lr': 0.0004508213753639161, 'samples': 2255808, 'steps': 11748, 'loss/train': 1.9458292722702026} 01/29/2022 03:56:07 - INFO - codeparrot_training - Step 11749: {'lr': 0.0004508116295471453, 'samples': 2256000, 'steps': 11749, 'loss/train': 1.7582497000694275} 01/29/2022 03:56:11 - INFO - codeparrot_training - Step 11750: {'lr': 0.0004508018828701612, 'samples': 2256192, 'steps': 11750, 'loss/train': 1.4000887870788574} 01/29/2022 03:56:15 - INFO - codeparrot_training - Step 11751: {'lr': 0.0004507921353330057, 'samples': 2256384, 'steps': 11751, 'loss/train': 1.75747811794281} 01/29/2022 03:56:22 - INFO - codeparrot_training - Step 11752: {'lr': 0.0004507823869357204, 'samples': 2256576, 'steps': 11752, 'loss/train': 2.526452422142029} 01/29/2022 03:56:26 - INFO - codeparrot_training - Step 11753: {'lr': 0.00045077263767834703, 'samples': 2256768, 'steps': 11753, 'loss/train': 1.1046659052371979} 01/29/2022 03:56:30 - INFO - codeparrot_training - Step 11754: {'lr': 0.00045076288756092754, 'samples': 2256960, 'steps': 11754, 'loss/train': 3.1031992435455322} 01/29/2022 03:56:35 - INFO - codeparrot_training - Step 11755: {'lr': 0.0004507531365835035, 'samples': 2257152, 'steps': 11755, 'loss/train': 1.9837932586669922} 01/29/2022 03:56:39 - INFO - codeparrot_training - Step 11756: {'lr': 0.00045074338474611683, 'samples': 2257344, 'steps': 11756, 'loss/train': 1.734089434146881} 01/29/2022 03:56:44 - INFO - codeparrot_training - Step 11757: {'lr': 0.00045073363204880916, 'samples': 2257536, 'steps': 11757, 'loss/train': 2.782091796398163} 01/29/2022 03:56:49 - INFO - codeparrot_training - Step 11758: {'lr': 0.0004507238784916224, 'samples': 2257728, 'steps': 11758, 'loss/train': 2.7166369557380676} 01/29/2022 03:56:53 - INFO - codeparrot_training - Step 11759: {'lr': 0.0004507141240745983, 'samples': 2257920, 'steps': 11759, 'loss/train': 0.9493867456912994} 01/29/2022 03:56:57 - INFO - codeparrot_training - Step 11760: {'lr': 0.0004507043687977787, 'samples': 2258112, 'steps': 11760, 'loss/train': 1.3749045431613922} 01/29/2022 03:57:01 - INFO - codeparrot_training - Step 11761: {'lr': 0.00045069461266120515, 'samples': 2258304, 'steps': 11761, 'loss/train': 2.0700716972351074} 01/29/2022 03:57:07 - INFO - codeparrot_training - Step 11762: {'lr': 0.0004506848556649197, 'samples': 2258496, 'steps': 11762, 'loss/train': 2.420838475227356} 01/29/2022 03:57:11 - INFO - codeparrot_training - Step 11763: {'lr': 0.0004506750978089641, 'samples': 2258688, 'steps': 11763, 'loss/train': 3.6305043697357178} 01/29/2022 03:57:15 - INFO - codeparrot_training - Step 11764: {'lr': 0.00045066533909338005, 'samples': 2258880, 'steps': 11764, 'loss/train': 1.676388680934906} 01/29/2022 03:57:19 - INFO - codeparrot_training - Step 11765: {'lr': 0.00045065557951820935, 'samples': 2259072, 'steps': 11765, 'loss/train': 1.9165374040603638} 01/29/2022 03:57:24 - INFO - codeparrot_training - Step 11766: {'lr': 0.0004506458190834939, 'samples': 2259264, 'steps': 11766, 'loss/train': 0.12093453481793404} 01/29/2022 03:57:30 - INFO - codeparrot_training - Step 11767: {'lr': 0.0004506360577892755, 'samples': 2259456, 'steps': 11767, 'loss/train': 1.7929562330245972} 01/29/2022 03:57:34 - INFO - codeparrot_training - Step 11768: {'lr': 0.00045062629563559595, 'samples': 2259648, 'steps': 11768, 'loss/train': 2.5676121711730957} 01/29/2022 03:57:38 - INFO - codeparrot_training - Step 11769: {'lr': 0.00045061653262249703, 'samples': 2259840, 'steps': 11769, 'loss/train': 1.8060945868492126} 01/29/2022 03:57:42 - INFO - codeparrot_training - Step 11770: {'lr': 0.0004506067687500206, 'samples': 2260032, 'steps': 11770, 'loss/train': 2.1647624373435974} 01/29/2022 03:57:47 - INFO - codeparrot_training - Step 11771: {'lr': 0.00045059700401820846, 'samples': 2260224, 'steps': 11771, 'loss/train': 1.250896006822586} 01/29/2022 03:57:52 - INFO - codeparrot_training - Step 11772: {'lr': 0.00045058723842710246, 'samples': 2260416, 'steps': 11772, 'loss/train': 1.772875428199768} 01/29/2022 03:57:56 - INFO - codeparrot_training - Step 11773: {'lr': 0.0004505774719767444, 'samples': 2260608, 'steps': 11773, 'loss/train': 1.343925029039383} 01/29/2022 03:58:00 - INFO - codeparrot_training - Step 11774: {'lr': 0.0004505677046671761, 'samples': 2260800, 'steps': 11774, 'loss/train': 1.801077961921692} 01/29/2022 03:58:05 - INFO - codeparrot_training - Step 11775: {'lr': 0.0004505579364984396, 'samples': 2260992, 'steps': 11775, 'loss/train': 1.8194561004638672} 01/29/2022 03:58:09 - INFO - codeparrot_training - Step 11776: {'lr': 0.0004505481674705764, 'samples': 2261184, 'steps': 11776, 'loss/train': 1.4632312953472137} 01/29/2022 03:58:15 - INFO - codeparrot_training - Step 11777: {'lr': 0.0004505383975836286, 'samples': 2261376, 'steps': 11777, 'loss/train': 1.1278747022151947} 01/29/2022 03:58:19 - INFO - codeparrot_training - Step 11778: {'lr': 0.00045052862683763806, 'samples': 2261568, 'steps': 11778, 'loss/train': 1.2084819674491882} 01/29/2022 03:58:23 - INFO - codeparrot_training - Step 11779: {'lr': 0.0004505188552326465, 'samples': 2261760, 'steps': 11779, 'loss/train': 2.188933253288269} 01/29/2022 03:58:27 - INFO - codeparrot_training - Step 11780: {'lr': 0.00045050908276869585, 'samples': 2261952, 'steps': 11780, 'loss/train': 1.9368554949760437} 01/29/2022 03:58:32 - INFO - codeparrot_training - Step 11781: {'lr': 0.00045049930944582783, 'samples': 2262144, 'steps': 11781, 'loss/train': 1.3596437573432922} 01/29/2022 03:58:37 - INFO - codeparrot_training - Step 11782: {'lr': 0.0004504895352640846, 'samples': 2262336, 'steps': 11782, 'loss/train': 2.1987537145614624} 01/29/2022 03:58:41 - INFO - codeparrot_training - Step 11783: {'lr': 0.0004504797602235078, 'samples': 2262528, 'steps': 11783, 'loss/train': 2.0970903038978577} 01/29/2022 03:58:46 - INFO - codeparrot_training - Step 11784: {'lr': 0.0004504699843241394, 'samples': 2262720, 'steps': 11784, 'loss/train': 1.68572598695755} 01/29/2022 03:58:50 - INFO - codeparrot_training - Step 11785: {'lr': 0.0004504602075660212, 'samples': 2262912, 'steps': 11785, 'loss/train': 1.8866232633590698} 01/29/2022 03:58:54 - INFO - codeparrot_training - Step 11786: {'lr': 0.00045045042994919514, 'samples': 2263104, 'steps': 11786, 'loss/train': 7.508693218231201} 01/29/2022 03:58:59 - INFO - codeparrot_training - Step 11787: {'lr': 0.00045044065147370303, 'samples': 2263296, 'steps': 11787, 'loss/train': 2.0704010725021362} 01/29/2022 03:59:03 - INFO - codeparrot_training - Step 11788: {'lr': 0.0004504308721395869, 'samples': 2263488, 'steps': 11788, 'loss/train': 1.6575788855552673} 01/29/2022 03:59:08 - INFO - codeparrot_training - Step 11789: {'lr': 0.0004504210919468886, 'samples': 2263680, 'steps': 11789, 'loss/train': 3.1379538774490356} 01/29/2022 03:59:12 - INFO - codeparrot_training - Step 11790: {'lr': 0.0004504113108956499, 'samples': 2263872, 'steps': 11790, 'loss/train': 0.9860571026802063} 01/29/2022 03:59:16 - INFO - codeparrot_training - Step 11791: {'lr': 0.0004504015289859128, 'samples': 2264064, 'steps': 11791, 'loss/train': 1.6328580379486084} 01/29/2022 03:59:22 - INFO - codeparrot_training - Step 11792: {'lr': 0.00045039174621771915, 'samples': 2264256, 'steps': 11792, 'loss/train': 1.8610535860061646} 01/29/2022 03:59:26 - INFO - codeparrot_training - Step 11793: {'lr': 0.0004503819625911109, 'samples': 2264448, 'steps': 11793, 'loss/train': 1.5651358366012573} 01/29/2022 03:59:31 - INFO - codeparrot_training - Step 11794: {'lr': 0.00045037217810613004, 'samples': 2264640, 'steps': 11794, 'loss/train': 1.8906819820404053} 01/29/2022 03:59:35 - INFO - codeparrot_training - Step 11795: {'lr': 0.0004503623927628183, 'samples': 2264832, 'steps': 11795, 'loss/train': 1.7131009697914124} 01/29/2022 03:59:39 - INFO - codeparrot_training - Step 11796: {'lr': 0.0004503526065612177, 'samples': 2265024, 'steps': 11796, 'loss/train': 1.3402626514434814} 01/29/2022 03:59:45 - INFO - codeparrot_training - Step 11797: {'lr': 0.0004503428195013702, 'samples': 2265216, 'steps': 11797, 'loss/train': 2.0392520427703857} 01/29/2022 03:59:49 - INFO - codeparrot_training - Step 11798: {'lr': 0.00045033303158331764, 'samples': 2265408, 'steps': 11798, 'loss/train': 1.9144084453582764} 01/29/2022 03:59:53 - INFO - codeparrot_training - Step 11799: {'lr': 0.00045032324280710204, 'samples': 2265600, 'steps': 11799, 'loss/train': 1.3380374014377594} 01/29/2022 03:59:58 - INFO - codeparrot_training - Step 11800: {'lr': 0.0004503134531727652, 'samples': 2265792, 'steps': 11800, 'loss/train': 1.6082380414009094} 01/29/2022 04:00:02 - INFO - codeparrot_training - Step 11801: {'lr': 0.00045030366268034917, 'samples': 2265984, 'steps': 11801, 'loss/train': 0.8966116905212402} 01/29/2022 04:00:07 - INFO - codeparrot_training - Step 11802: {'lr': 0.00045029387132989587, 'samples': 2266176, 'steps': 11802, 'loss/train': 1.9790571928024292} 01/29/2022 04:00:12 - INFO - codeparrot_training - Step 11803: {'lr': 0.0004502840791214472, 'samples': 2266368, 'steps': 11803, 'loss/train': 2.8253148794174194} 01/29/2022 04:00:16 - INFO - codeparrot_training - Step 11804: {'lr': 0.00045027428605504507, 'samples': 2266560, 'steps': 11804, 'loss/train': 1.557717204093933} 01/29/2022 04:00:20 - INFO - codeparrot_training - Step 11805: {'lr': 0.00045026449213073154, 'samples': 2266752, 'steps': 11805, 'loss/train': 1.4013668596744537} 01/29/2022 04:00:25 - INFO - codeparrot_training - Step 11806: {'lr': 0.00045025469734854856, 'samples': 2266944, 'steps': 11806, 'loss/train': 1.9595131874084473} 01/29/2022 04:00:30 - INFO - codeparrot_training - Step 11807: {'lr': 0.00045024490170853806, 'samples': 2267136, 'steps': 11807, 'loss/train': 2.29588919878006} 01/29/2022 04:00:34 - INFO - codeparrot_training - Step 11808: {'lr': 0.000450235105210742, 'samples': 2267328, 'steps': 11808, 'loss/train': 1.4341498017311096} 01/29/2022 04:00:38 - INFO - codeparrot_training - Step 11809: {'lr': 0.0004502253078552022, 'samples': 2267520, 'steps': 11809, 'loss/train': 0.2868035137653351} 01/29/2022 04:00:42 - INFO - codeparrot_training - Step 11810: {'lr': 0.00045021550964196086, 'samples': 2267712, 'steps': 11810, 'loss/train': 2.016146242618561} 01/29/2022 04:00:48 - INFO - codeparrot_training - Step 11811: {'lr': 0.0004502057105710598, 'samples': 2267904, 'steps': 11811, 'loss/train': 2.021438241004944} 01/29/2022 04:00:53 - INFO - codeparrot_training - Step 11812: {'lr': 0.00045019591064254105, 'samples': 2268096, 'steps': 11812, 'loss/train': 1.9001176357269287} 01/29/2022 04:00:57 - INFO - codeparrot_training - Step 11813: {'lr': 0.00045018610985644663, 'samples': 2268288, 'steps': 11813, 'loss/train': 0.8909291625022888} 01/29/2022 04:01:01 - INFO - codeparrot_training - Step 11814: {'lr': 0.00045017630821281854, 'samples': 2268480, 'steps': 11814, 'loss/train': 0.1740126684308052} 01/29/2022 04:01:05 - INFO - codeparrot_training - Step 11815: {'lr': 0.0004501665057116986, 'samples': 2268672, 'steps': 11815, 'loss/train': 1.886644721031189} 01/29/2022 04:01:11 - INFO - codeparrot_training - Step 11816: {'lr': 0.00045015670235312895, 'samples': 2268864, 'steps': 11816, 'loss/train': 1.4490754008293152} 01/29/2022 04:01:15 - INFO - codeparrot_training - Step 11817: {'lr': 0.00045014689813715147, 'samples': 2269056, 'steps': 11817, 'loss/train': 1.6531078219413757} 01/29/2022 04:01:19 - INFO - codeparrot_training - Step 11818: {'lr': 0.00045013709306380837, 'samples': 2269248, 'steps': 11818, 'loss/train': 0.3029879406094551} 01/29/2022 04:01:23 - INFO - codeparrot_training - Step 11819: {'lr': 0.00045012728713314146, 'samples': 2269440, 'steps': 11819, 'loss/train': 1.0745114386081696} 01/29/2022 04:01:28 - INFO - codeparrot_training - Step 11820: {'lr': 0.00045011748034519275, 'samples': 2269632, 'steps': 11820, 'loss/train': 2.2095268964767456} 01/29/2022 04:01:34 - INFO - codeparrot_training - Step 11821: {'lr': 0.00045010767270000436, 'samples': 2269824, 'steps': 11821, 'loss/train': 0.7620357871055603} 01/29/2022 04:01:38 - INFO - codeparrot_training - Step 11822: {'lr': 0.00045009786419761825, 'samples': 2270016, 'steps': 11822, 'loss/train': 1.5830944776535034} 01/29/2022 04:01:42 - INFO - codeparrot_training - Step 11823: {'lr': 0.00045008805483807637, 'samples': 2270208, 'steps': 11823, 'loss/train': 1.8736674785614014} 01/29/2022 04:01:47 - INFO - codeparrot_training - Step 11824: {'lr': 0.0004500782446214208, 'samples': 2270400, 'steps': 11824, 'loss/train': 1.458779364824295} 01/29/2022 04:01:51 - INFO - codeparrot_training - Step 11825: {'lr': 0.00045006843354769354, 'samples': 2270592, 'steps': 11825, 'loss/train': 1.5662963390350342} 01/29/2022 04:01:56 - INFO - codeparrot_training - Step 11826: {'lr': 0.0004500586216169367, 'samples': 2270784, 'steps': 11826, 'loss/train': 1.712278962135315} 01/29/2022 04:02:00 - INFO - codeparrot_training - Step 11827: {'lr': 0.0004500488088291923, 'samples': 2270976, 'steps': 11827, 'loss/train': 1.963681697845459} 01/29/2022 04:02:05 - INFO - codeparrot_training - Step 11828: {'lr': 0.0004500389951845022, 'samples': 2271168, 'steps': 11828, 'loss/train': 1.3562531769275665} 01/29/2022 04:02:09 - INFO - codeparrot_training - Step 11829: {'lr': 0.00045002918068290864, 'samples': 2271360, 'steps': 11829, 'loss/train': 1.6613563299179077} 01/29/2022 04:02:13 - INFO - codeparrot_training - Step 11830: {'lr': 0.00045001936532445354, 'samples': 2271552, 'steps': 11830, 'loss/train': 1.9282811880111694} 01/29/2022 04:02:18 - INFO - codeparrot_training - Step 11831: {'lr': 0.000450009549109179, 'samples': 2271744, 'steps': 11831, 'loss/train': 2.0874513387680054} 01/29/2022 04:02:23 - INFO - codeparrot_training - Step 11832: {'lr': 0.0004499997320371271, 'samples': 2271936, 'steps': 11832, 'loss/train': 1.507251799106598} 01/29/2022 04:02:27 - INFO - codeparrot_training - Step 11833: {'lr': 0.0004499899141083399, 'samples': 2272128, 'steps': 11833, 'loss/train': 1.8090078234672546} 01/29/2022 04:02:31 - INFO - codeparrot_training - Step 11834: {'lr': 0.0004499800953228593, 'samples': 2272320, 'steps': 11834, 'loss/train': 2.176843285560608} 01/29/2022 04:02:35 - INFO - codeparrot_training - Step 11835: {'lr': 0.00044997027568072754, 'samples': 2272512, 'steps': 11835, 'loss/train': 2.5679020285606384} 01/29/2022 04:02:41 - INFO - codeparrot_training - Step 11836: {'lr': 0.00044996045518198657, 'samples': 2272704, 'steps': 11836, 'loss/train': 2.4285938143730164} 01/29/2022 04:02:46 - INFO - codeparrot_training - Step 11837: {'lr': 0.00044995063382667855, 'samples': 2272896, 'steps': 11837, 'loss/train': 1.4732081294059753} 01/29/2022 04:02:50 - INFO - codeparrot_training - Step 11838: {'lr': 0.0004499408116148455, 'samples': 2273088, 'steps': 11838, 'loss/train': 1.256505399942398} 01/29/2022 04:02:54 - INFO - codeparrot_training - Step 11839: {'lr': 0.00044993098854652954, 'samples': 2273280, 'steps': 11839, 'loss/train': 2.2231770157814026} 01/29/2022 04:02:58 - INFO - codeparrot_training - Step 11840: {'lr': 0.0004499211646217727, 'samples': 2273472, 'steps': 11840, 'loss/train': 1.3928905427455902} 01/29/2022 04:03:04 - INFO - codeparrot_training - Step 11841: {'lr': 0.000449911339840617, 'samples': 2273664, 'steps': 11841, 'loss/train': 2.023532509803772} 01/29/2022 04:03:08 - INFO - codeparrot_training - Step 11842: {'lr': 0.00044990151420310463, 'samples': 2273856, 'steps': 11842, 'loss/train': 2.1572880148887634} 01/29/2022 04:03:12 - INFO - codeparrot_training - Step 11843: {'lr': 0.0004498916877092776, 'samples': 2274048, 'steps': 11843, 'loss/train': 1.0245280265808105} 01/29/2022 04:03:16 - INFO - codeparrot_training - Step 11844: {'lr': 0.00044988186035917817, 'samples': 2274240, 'steps': 11844, 'loss/train': 1.572982907295227} 01/29/2022 04:03:21 - INFO - codeparrot_training - Step 11845: {'lr': 0.00044987203215284823, 'samples': 2274432, 'steps': 11845, 'loss/train': 1.53653222322464} 01/29/2022 04:03:26 - INFO - codeparrot_training - Step 11846: {'lr': 0.00044986220309033, 'samples': 2274624, 'steps': 11846, 'loss/train': 0.978980541229248} 01/29/2022 04:03:30 - INFO - codeparrot_training - Step 11847: {'lr': 0.00044985237317166554, 'samples': 2274816, 'steps': 11847, 'loss/train': 1.615548849105835} 01/29/2022 04:03:34 - INFO - codeparrot_training - Step 11848: {'lr': 0.00044984254239689703, 'samples': 2275008, 'steps': 11848, 'loss/train': 2.3795828819274902} 01/29/2022 04:03:39 - INFO - codeparrot_training - Step 11849: {'lr': 0.00044983271076606644, 'samples': 2275200, 'steps': 11849, 'loss/train': 1.4239554405212402} 01/29/2022 04:03:43 - INFO - codeparrot_training - Step 11850: {'lr': 0.000449822878279216, 'samples': 2275392, 'steps': 11850, 'loss/train': 1.9543254375457764} 01/29/2022 04:03:49 - INFO - codeparrot_training - Step 11851: {'lr': 0.00044981304493638786, 'samples': 2275584, 'steps': 11851, 'loss/train': 1.5867474675178528} 01/29/2022 04:03:53 - INFO - codeparrot_training - Step 11852: {'lr': 0.00044980321073762405, 'samples': 2275776, 'steps': 11852, 'loss/train': 1.9409083127975464} 01/29/2022 04:03:57 - INFO - codeparrot_training - Step 11853: {'lr': 0.0004497933756829667, 'samples': 2275968, 'steps': 11853, 'loss/train': 1.3755674958229065} 01/29/2022 04:04:01 - INFO - codeparrot_training - Step 11854: {'lr': 0.000449783539772458, 'samples': 2276160, 'steps': 11854, 'loss/train': 0.9929640591144562} 01/29/2022 04:04:06 - INFO - codeparrot_training - Step 11855: {'lr': 0.00044977370300614, 'samples': 2276352, 'steps': 11855, 'loss/train': 1.8404315114021301} 01/29/2022 04:04:11 - INFO - codeparrot_training - Step 11856: {'lr': 0.00044976386538405494, 'samples': 2276544, 'steps': 11856, 'loss/train': 2.58205783367157} 01/29/2022 04:04:16 - INFO - codeparrot_training - Step 11857: {'lr': 0.0004497540269062449, 'samples': 2276736, 'steps': 11857, 'loss/train': 1.7062876224517822} 01/29/2022 04:04:20 - INFO - codeparrot_training - Step 11858: {'lr': 0.00044974418757275206, 'samples': 2276928, 'steps': 11858, 'loss/train': 1.2111964523792267} 01/29/2022 04:04:24 - INFO - codeparrot_training - Step 11859: {'lr': 0.00044973434738361853, 'samples': 2277120, 'steps': 11859, 'loss/train': 2.8969935178756714} 01/29/2022 04:04:28 - INFO - codeparrot_training - Step 11860: {'lr': 0.0004497245063388865, 'samples': 2277312, 'steps': 11860, 'loss/train': 1.853253722190857} 01/29/2022 04:04:34 - INFO - codeparrot_training - Step 11861: {'lr': 0.0004497146644385981, 'samples': 2277504, 'steps': 11861, 'loss/train': 2.0628259778022766} 01/29/2022 04:04:38 - INFO - codeparrot_training - Step 11862: {'lr': 0.00044970482168279547, 'samples': 2277696, 'steps': 11862, 'loss/train': 1.6971302032470703} 01/29/2022 04:04:42 - INFO - codeparrot_training - Step 11863: {'lr': 0.0004496949780715208, 'samples': 2277888, 'steps': 11863, 'loss/train': 1.4003652334213257} 01/29/2022 04:04:46 - INFO - codeparrot_training - Step 11864: {'lr': 0.00044968513360481624, 'samples': 2278080, 'steps': 11864, 'loss/train': 2.0748599767684937} 01/29/2022 04:04:51 - INFO - codeparrot_training - Step 11865: {'lr': 0.000449675288282724, 'samples': 2278272, 'steps': 11865, 'loss/train': 0.6318256258964539} 01/29/2022 04:04:57 - INFO - codeparrot_training - Step 11866: {'lr': 0.0004496654421052862, 'samples': 2278464, 'steps': 11866, 'loss/train': 2.4398964643478394} 01/29/2022 04:05:01 - INFO - codeparrot_training - Step 11867: {'lr': 0.00044965559507254504, 'samples': 2278656, 'steps': 11867, 'loss/train': 1.594784438610077} 01/29/2022 04:05:05 - INFO - codeparrot_training - Step 11868: {'lr': 0.0004496457471845428, 'samples': 2278848, 'steps': 11868, 'loss/train': 1.942247450351715} 01/29/2022 04:05:09 - INFO - codeparrot_training - Step 11869: {'lr': 0.0004496358984413215, 'samples': 2279040, 'steps': 11869, 'loss/train': 1.2968876659870148} 01/29/2022 04:05:14 - INFO - codeparrot_training - Step 11870: {'lr': 0.0004496260488429234, 'samples': 2279232, 'steps': 11870, 'loss/train': 2.1712766289711} 01/29/2022 04:05:19 - INFO - codeparrot_training - Step 11871: {'lr': 0.0004496161983893907, 'samples': 2279424, 'steps': 11871, 'loss/train': 1.852479636669159} 01/29/2022 04:05:23 - INFO - codeparrot_training - Step 11872: {'lr': 0.0004496063470807656, 'samples': 2279616, 'steps': 11872, 'loss/train': 0.19213740527629852} 01/29/2022 04:05:27 - INFO - codeparrot_training - Step 11873: {'lr': 0.0004495964949170903, 'samples': 2279808, 'steps': 11873, 'loss/train': 1.7759474515914917} 01/29/2022 04:05:32 - INFO - codeparrot_training - Step 11874: {'lr': 0.000449586641898407, 'samples': 2280000, 'steps': 11874, 'loss/train': 1.6171594262123108} 01/29/2022 04:05:36 - INFO - codeparrot_training - Step 11875: {'lr': 0.0004495767880247579, 'samples': 2280192, 'steps': 11875, 'loss/train': 2.0840924978256226} 01/29/2022 04:05:41 - INFO - codeparrot_training - Step 11876: {'lr': 0.0004495669332961852, 'samples': 2280384, 'steps': 11876, 'loss/train': 2.2858795523643494} 01/29/2022 04:05:45 - INFO - codeparrot_training - Step 11877: {'lr': 0.0004495570777127311, 'samples': 2280576, 'steps': 11877, 'loss/train': 1.7387139201164246} 01/29/2022 04:05:50 - INFO - codeparrot_training - Step 11878: {'lr': 0.00044954722127443786, 'samples': 2280768, 'steps': 11878, 'loss/train': 1.5241438150405884} 01/29/2022 04:05:54 - INFO - codeparrot_training - Step 11879: {'lr': 0.0004495373639813477, 'samples': 2280960, 'steps': 11879, 'loss/train': 1.9106608629226685} 01/29/2022 04:05:58 - INFO - codeparrot_training - Step 11880: {'lr': 0.00044952750583350287, 'samples': 2281152, 'steps': 11880, 'loss/train': 0.8754987716674805} 01/29/2022 04:06:04 - INFO - codeparrot_training - Step 11881: {'lr': 0.00044951764683094555, 'samples': 2281344, 'steps': 11881, 'loss/train': 2.421107769012451} 01/29/2022 04:06:09 - INFO - codeparrot_training - Step 11882: {'lr': 0.000449507786973718, 'samples': 2281536, 'steps': 11882, 'loss/train': 1.6292869448661804} 01/29/2022 04:06:13 - INFO - codeparrot_training - Step 11883: {'lr': 0.0004494979262618624, 'samples': 2281728, 'steps': 11883, 'loss/train': 2.179417848587036} 01/29/2022 04:06:17 - INFO - codeparrot_training - Step 11884: {'lr': 0.00044948806469542095, 'samples': 2281920, 'steps': 11884, 'loss/train': 1.7759943008422852} 01/29/2022 04:06:21 - INFO - codeparrot_training - Step 11885: {'lr': 0.0004494782022744361, 'samples': 2282112, 'steps': 11885, 'loss/train': 2.4799987077713013} 01/29/2022 04:06:27 - INFO - codeparrot_training - Step 11886: {'lr': 0.0004494683389989499, 'samples': 2282304, 'steps': 11886, 'loss/train': 2.0633845925331116} 01/29/2022 04:06:31 - INFO - codeparrot_training - Step 11887: {'lr': 0.0004494584748690047, 'samples': 2282496, 'steps': 11887, 'loss/train': 2.6970133781433105} 01/29/2022 04:06:35 - INFO - codeparrot_training - Step 11888: {'lr': 0.00044944860988464276, 'samples': 2282688, 'steps': 11888, 'loss/train': 2.0808656215667725} 01/29/2022 04:06:40 - INFO - codeparrot_training - Step 11889: {'lr': 0.0004494387440459063, 'samples': 2282880, 'steps': 11889, 'loss/train': 2.812170624732971} 01/29/2022 04:06:44 - INFO - codeparrot_training - Step 11890: {'lr': 0.00044942887735283755, 'samples': 2283072, 'steps': 11890, 'loss/train': 2.110699474811554} 01/29/2022 04:06:49 - INFO - codeparrot_training - Step 11891: {'lr': 0.00044941900980547886, 'samples': 2283264, 'steps': 11891, 'loss/train': 1.7362861633300781} 01/29/2022 04:06:53 - INFO - codeparrot_training - Step 11892: {'lr': 0.00044940914140387245, 'samples': 2283456, 'steps': 11892, 'loss/train': 0.8635789453983307} 01/29/2022 04:06:58 - INFO - codeparrot_training - Step 11893: {'lr': 0.00044939927214806055, 'samples': 2283648, 'steps': 11893, 'loss/train': 2.0590036511421204} 01/29/2022 04:07:02 - INFO - codeparrot_training - Step 11894: {'lr': 0.0004493894020380855, 'samples': 2283840, 'steps': 11894, 'loss/train': 1.1141965091228485} 01/29/2022 04:07:06 - INFO - codeparrot_training - Step 11895: {'lr': 0.0004493795310739896, 'samples': 2284032, 'steps': 11895, 'loss/train': 1.6181157231330872} 01/29/2022 04:07:12 - INFO - codeparrot_training - Step 11896: {'lr': 0.00044936965925581506, 'samples': 2284224, 'steps': 11896, 'loss/train': 2.3293408155441284} 01/29/2022 04:07:16 - INFO - codeparrot_training - Step 11897: {'lr': 0.0004493597865836042, 'samples': 2284416, 'steps': 11897, 'loss/train': 2.7277904748916626} 01/29/2022 04:07:21 - INFO - codeparrot_training - Step 11898: {'lr': 0.00044934991305739936, 'samples': 2284608, 'steps': 11898, 'loss/train': 1.4362866282463074} 01/29/2022 04:07:25 - INFO - codeparrot_training - Step 11899: {'lr': 0.00044934003867724284, 'samples': 2284800, 'steps': 11899, 'loss/train': 1.794341504573822} 01/29/2022 04:07:29 - INFO - codeparrot_training - Step 11900: {'lr': 0.0004493301634431768, 'samples': 2284992, 'steps': 11900, 'loss/train': 1.5101755857467651} 01/29/2022 04:07:35 - INFO - codeparrot_training - Step 11901: {'lr': 0.00044932028735524367, 'samples': 2285184, 'steps': 11901, 'loss/train': 1.1753505170345306} 01/29/2022 04:07:39 - INFO - codeparrot_training - Step 11902: {'lr': 0.0004493104104134857, 'samples': 2285376, 'steps': 11902, 'loss/train': 0.372331403195858} 01/29/2022 04:07:43 - INFO - codeparrot_training - Step 11903: {'lr': 0.0004493005326179452, 'samples': 2285568, 'steps': 11903, 'loss/train': 2.0194929242134094} 01/29/2022 04:07:47 - INFO - codeparrot_training - Step 11904: {'lr': 0.00044929065396866457, 'samples': 2285760, 'steps': 11904, 'loss/train': 1.442219227552414} 01/29/2022 04:07:52 - INFO - codeparrot_training - Step 11905: {'lr': 0.00044928077446568606, 'samples': 2285952, 'steps': 11905, 'loss/train': 2.0896652340888977} 01/29/2022 04:07:57 - INFO - codeparrot_training - Step 11906: {'lr': 0.000449270894109052, 'samples': 2286144, 'steps': 11906, 'loss/train': 2.26868337392807} 01/29/2022 04:08:02 - INFO - codeparrot_training - Step 11907: {'lr': 0.0004492610128988046, 'samples': 2286336, 'steps': 11907, 'loss/train': 1.9593903422355652} 01/29/2022 04:08:06 - INFO - codeparrot_training - Step 11908: {'lr': 0.00044925113083498636, 'samples': 2286528, 'steps': 11908, 'loss/train': 0.8927529752254486} 01/29/2022 04:08:10 - INFO - codeparrot_training - Step 11909: {'lr': 0.00044924124791763956, 'samples': 2286720, 'steps': 11909, 'loss/train': 2.196645498275757} 01/29/2022 04:08:14 - INFO - codeparrot_training - Step 11910: {'lr': 0.0004492313641468065, 'samples': 2286912, 'steps': 11910, 'loss/train': 2.9239606261253357} 01/29/2022 04:08:20 - INFO - codeparrot_training - Step 11911: {'lr': 0.00044922147952252957, 'samples': 2287104, 'steps': 11911, 'loss/train': 1.729137361049652} 01/29/2022 04:08:24 - INFO - codeparrot_training - Step 11912: {'lr': 0.000449211594044851, 'samples': 2287296, 'steps': 11912, 'loss/train': 1.0402190387248993} 01/29/2022 04:08:28 - INFO - codeparrot_training - Step 11913: {'lr': 0.0004492017077138133, 'samples': 2287488, 'steps': 11913, 'loss/train': 1.6229551434516907} 01/29/2022 04:08:32 - INFO - codeparrot_training - Step 11914: {'lr': 0.00044919182052945866, 'samples': 2287680, 'steps': 11914, 'loss/train': 1.5309342741966248} 01/29/2022 04:08:37 - INFO - codeparrot_training - Step 11915: {'lr': 0.00044918193249182957, 'samples': 2287872, 'steps': 11915, 'loss/train': 1.6342200636863708} 01/29/2022 04:08:42 - INFO - codeparrot_training - Step 11916: {'lr': 0.0004491720436009683, 'samples': 2288064, 'steps': 11916, 'loss/train': 1.3016336560249329} 01/29/2022 04:08:46 - INFO - codeparrot_training - Step 11917: {'lr': 0.0004491621538569173, 'samples': 2288256, 'steps': 11917, 'loss/train': 2.273019254207611} 01/29/2022 04:08:50 - INFO - codeparrot_training - Step 11918: {'lr': 0.0004491522632597188, 'samples': 2288448, 'steps': 11918, 'loss/train': 1.8318099975585938} 01/29/2022 04:08:55 - INFO - codeparrot_training - Step 11919: {'lr': 0.0004491423718094153, 'samples': 2288640, 'steps': 11919, 'loss/train': 1.5899714827537537} 01/29/2022 04:08:59 - INFO - codeparrot_training - Step 11920: {'lr': 0.00044913247950604905, 'samples': 2288832, 'steps': 11920, 'loss/train': 2.3122302889823914} 01/29/2022 04:09:05 - INFO - codeparrot_training - Step 11921: {'lr': 0.0004491225863496625, 'samples': 2289024, 'steps': 11921, 'loss/train': 1.0913836061954498} 01/29/2022 04:09:09 - INFO - codeparrot_training - Step 11922: {'lr': 0.0004491126923402981, 'samples': 2289216, 'steps': 11922, 'loss/train': 0.13209909573197365} 01/29/2022 04:09:13 - INFO - codeparrot_training - Step 11923: {'lr': 0.0004491027974779981, 'samples': 2289408, 'steps': 11923, 'loss/train': 2.37447988986969} 01/29/2022 04:09:17 - INFO - codeparrot_training - Step 11924: {'lr': 0.00044909290176280495, 'samples': 2289600, 'steps': 11924, 'loss/train': 1.8194153308868408} 01/29/2022 04:09:22 - INFO - codeparrot_training - Step 11925: {'lr': 0.000449083005194761, 'samples': 2289792, 'steps': 11925, 'loss/train': 1.0832642912864685} 01/29/2022 04:09:28 - INFO - codeparrot_training - Step 11926: {'lr': 0.0004490731077739087, 'samples': 2289984, 'steps': 11926, 'loss/train': 1.6318788528442383} 01/29/2022 04:09:32 - INFO - codeparrot_training - Step 11927: {'lr': 0.0004490632095002904, 'samples': 2290176, 'steps': 11927, 'loss/train': 1.8588870763778687} 01/29/2022 04:09:36 - INFO - codeparrot_training - Step 11928: {'lr': 0.00044905331037394853, 'samples': 2290368, 'steps': 11928, 'loss/train': 7.548701763153076} 01/29/2022 04:09:40 - INFO - codeparrot_training - Step 11929: {'lr': 0.00044904341039492544, 'samples': 2290560, 'steps': 11929, 'loss/train': 2.598328471183777} 01/29/2022 04:09:45 - INFO - codeparrot_training - Step 11930: {'lr': 0.00044903350956326365, 'samples': 2290752, 'steps': 11930, 'loss/train': 2.1141334176063538} 01/29/2022 04:09:49 - INFO - codeparrot_training - Step 11931: {'lr': 0.0004490236078790055, 'samples': 2290944, 'steps': 11931, 'loss/train': 1.7166550755500793} 01/29/2022 04:09:54 - INFO - codeparrot_training - Step 11932: {'lr': 0.0004490137053421934, 'samples': 2291136, 'steps': 11932, 'loss/train': 1.772803544998169} 01/29/2022 04:09:58 - INFO - codeparrot_training - Step 11933: {'lr': 0.00044900380195286974, 'samples': 2291328, 'steps': 11933, 'loss/train': 2.1991358399391174} 01/29/2022 04:10:02 - INFO - codeparrot_training - Step 11934: {'lr': 0.00044899389771107704, 'samples': 2291520, 'steps': 11934, 'loss/train': 1.7559127807617188} 01/29/2022 04:10:07 - INFO - codeparrot_training - Step 11935: {'lr': 0.00044898399261685765, 'samples': 2291712, 'steps': 11935, 'loss/train': 0.8002734482288361} 01/29/2022 04:10:11 - INFO - codeparrot_training - Step 11936: {'lr': 0.00044897408667025397, 'samples': 2291904, 'steps': 11936, 'loss/train': 2.838518500328064} 01/29/2022 04:10:17 - INFO - codeparrot_training - Step 11937: {'lr': 0.00044896417987130854, 'samples': 2292096, 'steps': 11937, 'loss/train': 1.4808247983455658} 01/29/2022 04:10:21 - INFO - codeparrot_training - Step 11938: {'lr': 0.0004489542722200637, 'samples': 2292288, 'steps': 11938, 'loss/train': 1.3631645143032074} 01/29/2022 04:10:25 - INFO - codeparrot_training - Step 11939: {'lr': 0.000448944363716562, 'samples': 2292480, 'steps': 11939, 'loss/train': 1.9381644129753113} 01/29/2022 04:10:30 - INFO - codeparrot_training - Step 11940: {'lr': 0.0004489344543608458, 'samples': 2292672, 'steps': 11940, 'loss/train': 1.7145304083824158} 01/29/2022 04:10:34 - INFO - codeparrot_training - Step 11941: {'lr': 0.00044892454415295746, 'samples': 2292864, 'steps': 11941, 'loss/train': 1.486046701669693} 01/29/2022 04:10:39 - INFO - codeparrot_training - Step 11942: {'lr': 0.0004489146330929397, 'samples': 2293056, 'steps': 11942, 'loss/train': 2.4439775347709656} 01/29/2022 04:10:43 - INFO - codeparrot_training - Step 11943: {'lr': 0.0004489047211808347, 'samples': 2293248, 'steps': 11943, 'loss/train': 1.0496889352798462} 01/29/2022 04:10:48 - INFO - codeparrot_training - Step 11944: {'lr': 0.0004488948084166851, 'samples': 2293440, 'steps': 11944, 'loss/train': 2.1428298354148865} 01/29/2022 04:10:52 - INFO - codeparrot_training - Step 11945: {'lr': 0.00044888489480053324, 'samples': 2293632, 'steps': 11945, 'loss/train': 1.6831272840499878} 01/29/2022 04:10:56 - INFO - codeparrot_training - Step 11946: {'lr': 0.00044887498033242167, 'samples': 2293824, 'steps': 11946, 'loss/train': 1.7277878522872925} 01/29/2022 04:11:02 - INFO - codeparrot_training - Step 11947: {'lr': 0.0004488650650123929, 'samples': 2294016, 'steps': 11947, 'loss/train': 2.607155978679657} 01/29/2022 04:11:07 - INFO - codeparrot_training - Step 11948: {'lr': 0.00044885514884048926, 'samples': 2294208, 'steps': 11948, 'loss/train': 0.8032273650169373} 01/29/2022 04:11:11 - INFO - codeparrot_training - Step 11949: {'lr': 0.0004488452318167533, 'samples': 2294400, 'steps': 11949, 'loss/train': 1.0431544482707977} 01/29/2022 04:11:15 - INFO - codeparrot_training - Step 11950: {'lr': 0.00044883531394122753, 'samples': 2294592, 'steps': 11950, 'loss/train': 1.8261934518814087} 01/29/2022 04:11:19 - INFO - codeparrot_training - Step 11951: {'lr': 0.00044882539521395436, 'samples': 2294784, 'steps': 11951, 'loss/train': 1.8348519802093506} 01/29/2022 04:11:25 - INFO - codeparrot_training - Step 11952: {'lr': 0.0004488154756349764, 'samples': 2294976, 'steps': 11952, 'loss/train': 1.6538659930229187} 01/29/2022 04:11:29 - INFO - codeparrot_training - Step 11953: {'lr': 0.0004488055552043361, 'samples': 2295168, 'steps': 11953, 'loss/train': 1.702840805053711} 01/29/2022 04:11:33 - INFO - codeparrot_training - Step 11954: {'lr': 0.0004487956339220759, 'samples': 2295360, 'steps': 11954, 'loss/train': 1.1011039316654205} 01/29/2022 04:11:38 - INFO - codeparrot_training - Step 11955: {'lr': 0.00044878571178823826, 'samples': 2295552, 'steps': 11955, 'loss/train': 2.035111963748932} 01/29/2022 04:11:43 - INFO - codeparrot_training - Step 11956: {'lr': 0.00044877578880286585, 'samples': 2295744, 'steps': 11956, 'loss/train': 1.319482684135437} 01/29/2022 04:11:47 - INFO - codeparrot_training - Step 11957: {'lr': 0.000448765864966001, 'samples': 2295936, 'steps': 11957, 'loss/train': 1.6231448650360107} 01/29/2022 04:11:51 - INFO - codeparrot_training - Step 11958: {'lr': 0.00044875594027768634, 'samples': 2296128, 'steps': 11958, 'loss/train': 1.1715380251407623} 01/29/2022 04:11:56 - INFO - codeparrot_training - Step 11959: {'lr': 0.00044874601473796435, 'samples': 2296320, 'steps': 11959, 'loss/train': 2.6627374291419983} 01/29/2022 04:12:00 - INFO - codeparrot_training - Step 11960: {'lr': 0.00044873608834687754, 'samples': 2296512, 'steps': 11960, 'loss/train': 0.6774130314588547} 01/29/2022 04:12:06 - INFO - codeparrot_training - Step 11961: {'lr': 0.0004487261611044684, 'samples': 2296704, 'steps': 11961, 'loss/train': 2.352743983268738} 01/29/2022 04:12:10 - INFO - codeparrot_training - Step 11962: {'lr': 0.0004487162330107795, 'samples': 2296896, 'steps': 11962, 'loss/train': 1.1631629168987274} 01/29/2022 04:12:14 - INFO - codeparrot_training - Step 11963: {'lr': 0.0004487063040658534, 'samples': 2297088, 'steps': 11963, 'loss/train': 2.0400379300117493} 01/29/2022 04:12:18 - INFO - codeparrot_training - Step 11964: {'lr': 0.00044869637426973256, 'samples': 2297280, 'steps': 11964, 'loss/train': 1.2219336032867432} 01/29/2022 04:12:23 - INFO - codeparrot_training - Step 11965: {'lr': 0.0004486864436224595, 'samples': 2297472, 'steps': 11965, 'loss/train': 2.10578316450119} 01/29/2022 04:12:28 - INFO - codeparrot_training - Step 11966: {'lr': 0.0004486765121240769, 'samples': 2297664, 'steps': 11966, 'loss/train': 2.237364113330841} 01/29/2022 04:12:32 - INFO - codeparrot_training - Step 11967: {'lr': 0.0004486665797746271, 'samples': 2297856, 'steps': 11967, 'loss/train': 1.9944276809692383} 01/29/2022 04:12:36 - INFO - codeparrot_training - Step 11968: {'lr': 0.00044865664657415286, 'samples': 2298048, 'steps': 11968, 'loss/train': 0.8573837578296661} 01/29/2022 04:12:41 - INFO - codeparrot_training - Step 11969: {'lr': 0.00044864671252269663, 'samples': 2298240, 'steps': 11969, 'loss/train': 1.794024109840393} 01/29/2022 04:12:45 - INFO - codeparrot_training - Step 11970: {'lr': 0.00044863677762030087, 'samples': 2298432, 'steps': 11970, 'loss/train': 2.4545406103134155} 01/29/2022 04:12:50 - INFO - codeparrot_training - Step 11971: {'lr': 0.0004486268418670083, 'samples': 2298624, 'steps': 11971, 'loss/train': 2.113283693790436} 01/29/2022 04:12:54 - INFO - codeparrot_training - Step 11972: {'lr': 0.00044861690526286135, 'samples': 2298816, 'steps': 11972, 'loss/train': 0.9672141373157501} 01/29/2022 04:12:59 - INFO - codeparrot_training - Step 11973: {'lr': 0.00044860696780790266, 'samples': 2299008, 'steps': 11973, 'loss/train': 2.3527640104293823} 01/29/2022 04:13:03 - INFO - codeparrot_training - Step 11974: {'lr': 0.00044859702950217486, 'samples': 2299200, 'steps': 11974, 'loss/train': 0.9738459885120392} 01/29/2022 04:13:09 - INFO - codeparrot_training - Step 11975: {'lr': 0.00044858709034572035, 'samples': 2299392, 'steps': 11975, 'loss/train': 1.2135678827762604} 01/29/2022 04:13:13 - INFO - codeparrot_training - Step 11976: {'lr': 0.00044857715033858183, 'samples': 2299584, 'steps': 11976, 'loss/train': 1.584027349948883} 01/29/2022 04:13:17 - INFO - codeparrot_training - Step 11977: {'lr': 0.0004485672094808019, 'samples': 2299776, 'steps': 11977, 'loss/train': 0.25081392377614975} 01/29/2022 04:13:22 - INFO - codeparrot_training - Step 11978: {'lr': 0.0004485572677724231, 'samples': 2299968, 'steps': 11978, 'loss/train': 1.4716313481330872} 01/29/2022 04:13:26 - INFO - codeparrot_training - Step 11979: {'lr': 0.00044854732521348796, 'samples': 2300160, 'steps': 11979, 'loss/train': 1.0210805833339691} 01/29/2022 04:13:31 - INFO - codeparrot_training - Step 11980: {'lr': 0.0004485373818040391, 'samples': 2300352, 'steps': 11980, 'loss/train': 1.800119161605835} 01/29/2022 04:13:35 - INFO - codeparrot_training - Step 11981: {'lr': 0.00044852743754411915, 'samples': 2300544, 'steps': 11981, 'loss/train': 1.9518594145774841} 01/29/2022 04:13:40 - INFO - codeparrot_training - Step 11982: {'lr': 0.00044851749243377085, 'samples': 2300736, 'steps': 11982, 'loss/train': 1.901557445526123} 01/29/2022 04:13:44 - INFO - codeparrot_training - Step 11983: {'lr': 0.0004485075464730365, 'samples': 2300928, 'steps': 11983, 'loss/train': 2.802658796310425} 01/29/2022 04:13:48 - INFO - codeparrot_training - Step 11984: {'lr': 0.0004484975996619589, 'samples': 2301120, 'steps': 11984, 'loss/train': 1.918872892856598} 01/29/2022 04:13:54 - INFO - codeparrot_training - Step 11985: {'lr': 0.0004484876520005805, 'samples': 2301312, 'steps': 11985, 'loss/train': 1.7654984593391418} 01/29/2022 04:13:58 - INFO - codeparrot_training - Step 11986: {'lr': 0.0004484777034889441, 'samples': 2301504, 'steps': 11986, 'loss/train': 1.3624009788036346} 01/29/2022 04:14:03 - INFO - codeparrot_training - Step 11987: {'lr': 0.0004484677541270923, 'samples': 2301696, 'steps': 11987, 'loss/train': 1.5862471461296082} 01/29/2022 04:14:07 - INFO - codeparrot_training - Step 11988: {'lr': 0.00044845780391506763, 'samples': 2301888, 'steps': 11988, 'loss/train': 2.284824013710022} 01/29/2022 04:14:11 - INFO - codeparrot_training - Step 11989: {'lr': 0.0004484478528529128, 'samples': 2302080, 'steps': 11989, 'loss/train': 1.9272387027740479} 01/29/2022 04:14:16 - INFO - codeparrot_training - Step 11990: {'lr': 0.00044843790094067026, 'samples': 2302272, 'steps': 11990, 'loss/train': 1.9956060647964478} 01/29/2022 04:14:21 - INFO - codeparrot_training - Step 11991: {'lr': 0.00044842794817838286, 'samples': 2302464, 'steps': 11991, 'loss/train': 1.501246690750122} 01/29/2022 04:14:25 - INFO - codeparrot_training - Step 11992: {'lr': 0.0004484179945660931, 'samples': 2302656, 'steps': 11992, 'loss/train': 1.4883244335651398} 01/29/2022 04:14:29 - INFO - codeparrot_training - Step 11993: {'lr': 0.00044840804010384366, 'samples': 2302848, 'steps': 11993, 'loss/train': 1.5636361241340637} 01/29/2022 04:14:33 - INFO - codeparrot_training - Step 11994: {'lr': 0.00044839808479167723, 'samples': 2303040, 'steps': 11994, 'loss/train': 1.6065140962600708} 01/29/2022 04:14:39 - INFO - codeparrot_training - Step 11995: {'lr': 0.00044838812862963627, 'samples': 2303232, 'steps': 11995, 'loss/train': 2.061547636985779} 01/29/2022 04:14:43 - INFO - codeparrot_training - Step 11996: {'lr': 0.00044837817161776366, 'samples': 2303424, 'steps': 11996, 'loss/train': 0.9350040256977081} 01/29/2022 04:14:47 - INFO - codeparrot_training - Step 11997: {'lr': 0.00044836821375610194, 'samples': 2303616, 'steps': 11997, 'loss/train': 2.0862486362457275} 01/29/2022 04:14:51 - INFO - codeparrot_training - Step 11998: {'lr': 0.0004483582550446938, 'samples': 2303808, 'steps': 11998, 'loss/train': 1.6624356508255005} 01/29/2022 04:14:56 - INFO - codeparrot_training - Step 11999: {'lr': 0.0004483482954835819, 'samples': 2304000, 'steps': 11999, 'loss/train': 1.9538152813911438} 01/29/2022 04:14:56 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 04:15:28 - WARNING - huggingface_hub.repository - Several commits (6) will be pushed upstream. 01/29/2022 04:15:28 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 04:16:36 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 5069a3d..63b9456 expert-dust-2 -> expert-dust-2 01/29/2022 04:16:44 - INFO - codeparrot_training - Step 12000: {'lr': 0.0004483383350728088, 'samples': 2304192, 'steps': 12000, 'loss/train': 2.0816749334335327} 01/29/2022 04:16:49 - INFO - codeparrot_training - Step 12001: {'lr': 0.00044832837381241733, 'samples': 2304384, 'steps': 12001, 'loss/train': 0.96905717253685} 01/29/2022 04:16:53 - INFO - codeparrot_training - Step 12002: {'lr': 0.00044831841170245003, 'samples': 2304576, 'steps': 12002, 'loss/train': 1.7177585363388062} 01/29/2022 04:16:57 - INFO - codeparrot_training - Step 12003: {'lr': 0.0004483084487429496, 'samples': 2304768, 'steps': 12003, 'loss/train': 2.0082204937934875} 01/29/2022 04:17:02 - INFO - codeparrot_training - Step 12004: {'lr': 0.00044829848493395884, 'samples': 2304960, 'steps': 12004, 'loss/train': 1.0123384594917297} 01/29/2022 04:17:07 - INFO - codeparrot_training - Step 12005: {'lr': 0.00044828852027552023, 'samples': 2305152, 'steps': 12005, 'loss/train': 1.719336748123169} 01/29/2022 04:17:11 - INFO - codeparrot_training - Step 12006: {'lr': 0.00044827855476767665, 'samples': 2305344, 'steps': 12006, 'loss/train': 1.6771320104599} 01/29/2022 04:17:15 - INFO - codeparrot_training - Step 12007: {'lr': 0.00044826858841047067, 'samples': 2305536, 'steps': 12007, 'loss/train': 1.6261943578720093} 01/29/2022 04:17:19 - INFO - codeparrot_training - Step 12008: {'lr': 0.00044825862120394504, 'samples': 2305728, 'steps': 12008, 'loss/train': 1.939858853816986} 01/29/2022 04:17:25 - INFO - codeparrot_training - Step 12009: {'lr': 0.00044824865314814245, 'samples': 2305920, 'steps': 12009, 'loss/train': 1.8150234818458557} 01/29/2022 04:17:29 - INFO - codeparrot_training - Step 12010: {'lr': 0.00044823868424310553, 'samples': 2306112, 'steps': 12010, 'loss/train': 1.5280031561851501} 01/29/2022 04:17:33 - INFO - codeparrot_training - Step 12011: {'lr': 0.00044822871448887703, 'samples': 2306304, 'steps': 12011, 'loss/train': 1.796686828136444} 01/29/2022 04:17:37 - INFO - codeparrot_training - Step 12012: {'lr': 0.0004482187438854997, 'samples': 2306496, 'steps': 12012, 'loss/train': 1.8065477013587952} 01/29/2022 04:17:42 - INFO - codeparrot_training - Step 12013: {'lr': 0.00044820877243301617, 'samples': 2306688, 'steps': 12013, 'loss/train': 2.506659686565399} 01/29/2022 04:17:48 - INFO - codeparrot_training - Step 12014: {'lr': 0.00044819880013146924, 'samples': 2306880, 'steps': 12014, 'loss/train': 6.0698840618133545} 01/29/2022 04:17:52 - INFO - codeparrot_training - Step 12015: {'lr': 0.0004481888269809016, 'samples': 2307072, 'steps': 12015, 'loss/train': 2.090603291988373} 01/29/2022 04:17:56 - INFO - codeparrot_training - Step 12016: {'lr': 0.0004481788529813559, 'samples': 2307264, 'steps': 12016, 'loss/train': 1.767250120639801} 01/29/2022 04:18:00 - INFO - codeparrot_training - Step 12017: {'lr': 0.00044816887813287494, 'samples': 2307456, 'steps': 12017, 'loss/train': 1.456406593322754} 01/29/2022 04:18:05 - INFO - codeparrot_training - Step 12018: {'lr': 0.0004481589024355014, 'samples': 2307648, 'steps': 12018, 'loss/train': 1.2046127915382385} 01/29/2022 04:18:10 - INFO - codeparrot_training - Step 12019: {'lr': 0.00044814892588927816, 'samples': 2307840, 'steps': 12019, 'loss/train': 4.389662504196167} 01/29/2022 04:18:14 - INFO - codeparrot_training - Step 12020: {'lr': 0.00044813894849424777, 'samples': 2308032, 'steps': 12020, 'loss/train': 1.5759426355361938} 01/29/2022 04:18:18 - INFO - codeparrot_training - Step 12021: {'lr': 0.00044812897025045295, 'samples': 2308224, 'steps': 12021, 'loss/train': 1.8256902694702148} 01/29/2022 04:18:23 - INFO - codeparrot_training - Step 12022: {'lr': 0.00044811899115793666, 'samples': 2308416, 'steps': 12022, 'loss/train': 2.4643152952194214} 01/29/2022 04:18:27 - INFO - codeparrot_training - Step 12023: {'lr': 0.0004481090112167415, 'samples': 2308608, 'steps': 12023, 'loss/train': 2.1554941534996033} 01/29/2022 04:18:32 - INFO - codeparrot_training - Step 12024: {'lr': 0.0004480990304269102, 'samples': 2308800, 'steps': 12024, 'loss/train': 1.733257234096527} 01/29/2022 04:18:36 - INFO - codeparrot_training - Step 12025: {'lr': 0.00044808904878848555, 'samples': 2308992, 'steps': 12025, 'loss/train': 1.80075341463089} 01/29/2022 04:18:41 - INFO - codeparrot_training - Step 12026: {'lr': 0.00044807906630151033, 'samples': 2309184, 'steps': 12026, 'loss/train': 2.28225839138031} 01/29/2022 04:18:45 - INFO - codeparrot_training - Step 12027: {'lr': 0.00044806908296602733, 'samples': 2309376, 'steps': 12027, 'loss/train': 1.4631629884243011} 01/29/2022 04:18:49 - INFO - codeparrot_training - Step 12028: {'lr': 0.0004480590987820793, 'samples': 2309568, 'steps': 12028, 'loss/train': 2.2951993346214294} 01/29/2022 04:18:55 - INFO - codeparrot_training - Step 12029: {'lr': 0.00044804911374970893, 'samples': 2309760, 'steps': 12029, 'loss/train': 0.115275789052248} 01/29/2022 04:19:00 - INFO - codeparrot_training - Step 12030: {'lr': 0.000448039127868959, 'samples': 2309952, 'steps': 12030, 'loss/train': 1.919222116470337} 01/29/2022 04:19:04 - INFO - codeparrot_training - Step 12031: {'lr': 0.0004480291411398724, 'samples': 2310144, 'steps': 12031, 'loss/train': 2.1655193567276} 01/29/2022 04:19:08 - INFO - codeparrot_training - Step 12032: {'lr': 0.0004480191535624918, 'samples': 2310336, 'steps': 12032, 'loss/train': 1.460883378982544} 01/29/2022 04:19:13 - INFO - codeparrot_training - Step 12033: {'lr': 0.00044800916513686, 'samples': 2310528, 'steps': 12033, 'loss/train': 1.2252150177955627} 01/29/2022 04:19:18 - INFO - codeparrot_training - Step 12034: {'lr': 0.00044799917586301987, 'samples': 2310720, 'steps': 12034, 'loss/train': 0.9711203277111053} 01/29/2022 04:19:22 - INFO - codeparrot_training - Step 12035: {'lr': 0.00044798918574101413, 'samples': 2310912, 'steps': 12035, 'loss/train': 1.5555285215377808} 01/29/2022 04:19:26 - INFO - codeparrot_training - Step 12036: {'lr': 0.00044797919477088555, 'samples': 2311104, 'steps': 12036, 'loss/train': 1.3828530013561249} 01/29/2022 04:19:30 - INFO - codeparrot_training - Step 12037: {'lr': 0.00044796920295267696, 'samples': 2311296, 'steps': 12037, 'loss/train': 1.402132898569107} 01/29/2022 04:19:36 - INFO - codeparrot_training - Step 12038: {'lr': 0.0004479592102864313, 'samples': 2311488, 'steps': 12038, 'loss/train': 0.794544517993927} 01/29/2022 04:19:40 - INFO - codeparrot_training - Step 12039: {'lr': 0.0004479492167721911, 'samples': 2311680, 'steps': 12039, 'loss/train': 3.4393383264541626} 01/29/2022 04:19:44 - INFO - codeparrot_training - Step 12040: {'lr': 0.0004479392224099993, 'samples': 2311872, 'steps': 12040, 'loss/train': 1.396554708480835} 01/29/2022 04:19:48 - INFO - codeparrot_training - Step 12041: {'lr': 0.00044792922719989883, 'samples': 2312064, 'steps': 12041, 'loss/train': 1.705244779586792} 01/29/2022 04:19:53 - INFO - codeparrot_training - Step 12042: {'lr': 0.00044791923114193233, 'samples': 2312256, 'steps': 12042, 'loss/train': 2.272386074066162} 01/29/2022 04:19:59 - INFO - codeparrot_training - Step 12043: {'lr': 0.0004479092342361427, 'samples': 2312448, 'steps': 12043, 'loss/train': 1.7554423213005066} 01/29/2022 04:20:03 - INFO - codeparrot_training - Step 12044: {'lr': 0.0004478992364825728, 'samples': 2312640, 'steps': 12044, 'loss/train': 0.9704506695270538} 01/29/2022 04:20:07 - INFO - codeparrot_training - Step 12045: {'lr': 0.00044788923788126534, 'samples': 2312832, 'steps': 12045, 'loss/train': 1.8715344071388245} 01/29/2022 04:20:11 - INFO - codeparrot_training - Step 12046: {'lr': 0.00044787923843226323, 'samples': 2313024, 'steps': 12046, 'loss/train': 0.11802545562386513} 01/29/2022 04:20:16 - INFO - codeparrot_training - Step 12047: {'lr': 0.0004478692381356093, 'samples': 2313216, 'steps': 12047, 'loss/train': 2.371762275695801} 01/29/2022 04:20:21 - INFO - codeparrot_training - Step 12048: {'lr': 0.00044785923699134646, 'samples': 2313408, 'steps': 12048, 'loss/train': 1.5957112312316895} 01/29/2022 04:20:25 - INFO - codeparrot_training - Step 12049: {'lr': 0.0004478492349995174, 'samples': 2313600, 'steps': 12049, 'loss/train': 1.880310595035553} 01/29/2022 04:20:29 - INFO - codeparrot_training - Step 12050: {'lr': 0.00044783923216016507, 'samples': 2313792, 'steps': 12050, 'loss/train': 1.666087567806244} 01/29/2022 04:20:34 - INFO - codeparrot_training - Step 12051: {'lr': 0.0004478292284733323, 'samples': 2313984, 'steps': 12051, 'loss/train': 1.470834732055664} 01/29/2022 04:20:38 - INFO - codeparrot_training - Step 12052: {'lr': 0.00044781922393906186, 'samples': 2314176, 'steps': 12052, 'loss/train': 2.032983720302582} 01/29/2022 04:20:44 - INFO - codeparrot_training - Step 12053: {'lr': 0.00044780921855739676, 'samples': 2314368, 'steps': 12053, 'loss/train': 2.0970945954322815} 01/29/2022 04:20:48 - INFO - codeparrot_training - Step 12054: {'lr': 0.00044779921232837973, 'samples': 2314560, 'steps': 12054, 'loss/train': 1.736240029335022} 01/29/2022 04:20:52 - INFO - codeparrot_training - Step 12055: {'lr': 0.0004477892052520537, 'samples': 2314752, 'steps': 12055, 'loss/train': 1.5638240575790405} 01/29/2022 04:20:57 - INFO - codeparrot_training - Step 12056: {'lr': 0.0004477791973284616, 'samples': 2314944, 'steps': 12056, 'loss/train': 2.001521944999695} 01/29/2022 04:21:01 - INFO - codeparrot_training - Step 12057: {'lr': 0.00044776918855764616, 'samples': 2315136, 'steps': 12057, 'loss/train': 2.404481291770935} 01/29/2022 04:21:06 - INFO - codeparrot_training - Step 12058: {'lr': 0.00044775917893965025, 'samples': 2315328, 'steps': 12058, 'loss/train': 2.104567229747772} 01/29/2022 04:21:10 - INFO - codeparrot_training - Step 12059: {'lr': 0.00044774916847451683, 'samples': 2315520, 'steps': 12059, 'loss/train': 2.1350337266921997} 01/29/2022 04:21:14 - INFO - codeparrot_training - Step 12060: {'lr': 0.0004477391571622889, 'samples': 2315712, 'steps': 12060, 'loss/train': 2.3497856855392456} 01/29/2022 04:21:19 - INFO - codeparrot_training - Step 12061: {'lr': 0.00044772914500300907, 'samples': 2315904, 'steps': 12061, 'loss/train': 1.2904564440250397} 01/29/2022 04:21:24 - INFO - codeparrot_training - Step 12062: {'lr': 0.0004477191319967204, 'samples': 2316096, 'steps': 12062, 'loss/train': 1.381403625011444} 01/29/2022 04:21:28 - INFO - codeparrot_training - Step 12063: {'lr': 0.0004477091181434658, 'samples': 2316288, 'steps': 12063, 'loss/train': 1.4569500982761383} 01/29/2022 04:21:32 - INFO - codeparrot_training - Step 12064: {'lr': 0.00044769910344328803, 'samples': 2316480, 'steps': 12064, 'loss/train': 1.7745028138160706} 01/29/2022 04:21:37 - INFO - codeparrot_training - Step 12065: {'lr': 0.00044768908789623015, 'samples': 2316672, 'steps': 12065, 'loss/train': 1.5777856707572937} 01/29/2022 04:21:41 - INFO - codeparrot_training - Step 12066: {'lr': 0.00044767907150233496, 'samples': 2316864, 'steps': 12066, 'loss/train': 1.4092941284179688} 01/29/2022 04:21:47 - INFO - codeparrot_training - Step 12067: {'lr': 0.0004476690542616454, 'samples': 2317056, 'steps': 12067, 'loss/train': 1.0254608988761902} 01/29/2022 04:21:51 - INFO - codeparrot_training - Step 12068: {'lr': 0.00044765903617420436, 'samples': 2317248, 'steps': 12068, 'loss/train': 0.4437830150127411} 01/29/2022 04:21:55 - INFO - codeparrot_training - Step 12069: {'lr': 0.0004476490172400548, 'samples': 2317440, 'steps': 12069, 'loss/train': 2.0240662693977356} 01/29/2022 04:22:00 - INFO - codeparrot_training - Step 12070: {'lr': 0.00044763899745923965, 'samples': 2317632, 'steps': 12070, 'loss/train': 2.2272677421569824} 01/29/2022 04:22:04 - INFO - codeparrot_training - Step 12071: {'lr': 0.0004476289768318017, 'samples': 2317824, 'steps': 12071, 'loss/train': 1.3382343649864197} 01/29/2022 04:22:09 - INFO - codeparrot_training - Step 12072: {'lr': 0.00044761895535778404, 'samples': 2318016, 'steps': 12072, 'loss/train': 1.9367948770523071} 01/29/2022 04:22:13 - INFO - codeparrot_training - Step 12073: {'lr': 0.0004476089330372295, 'samples': 2318208, 'steps': 12073, 'loss/train': 1.9495723843574524} 01/29/2022 04:22:18 - INFO - codeparrot_training - Step 12074: {'lr': 0.00044759890987018105, 'samples': 2318400, 'steps': 12074, 'loss/train': 1.8456305265426636} 01/29/2022 04:22:22 - INFO - codeparrot_training - Step 12075: {'lr': 0.0004475888858566816, 'samples': 2318592, 'steps': 12075, 'loss/train': 0.06250034272670746} 01/29/2022 04:22:26 - INFO - codeparrot_training - Step 12076: {'lr': 0.00044757886099677416, 'samples': 2318784, 'steps': 12076, 'loss/train': 1.1668464839458466} 01/29/2022 04:22:31 - INFO - codeparrot_training - Step 12077: {'lr': 0.0004475688352905015, 'samples': 2318976, 'steps': 12077, 'loss/train': 2.0099313855171204} 01/29/2022 04:22:36 - INFO - codeparrot_training - Step 12078: {'lr': 0.00044755880873790675, 'samples': 2319168, 'steps': 12078, 'loss/train': 2.9538838863372803} 01/29/2022 04:22:40 - INFO - codeparrot_training - Step 12079: {'lr': 0.00044754878133903284, 'samples': 2319360, 'steps': 12079, 'loss/train': 0.5498796701431274} 01/29/2022 04:22:44 - INFO - codeparrot_training - Step 12080: {'lr': 0.0004475387530939226, 'samples': 2319552, 'steps': 12080, 'loss/train': 1.5394003987312317} 01/29/2022 04:22:48 - INFO - codeparrot_training - Step 12081: {'lr': 0.00044752872400261913, 'samples': 2319744, 'steps': 12081, 'loss/train': 0.3845197856426239} 01/29/2022 04:22:54 - INFO - codeparrot_training - Step 12082: {'lr': 0.0004475186940651653, 'samples': 2319936, 'steps': 12082, 'loss/train': 1.75567227602005} 01/29/2022 04:22:58 - INFO - codeparrot_training - Step 12083: {'lr': 0.0004475086632816041, 'samples': 2320128, 'steps': 12083, 'loss/train': 1.3719391822814941} 01/29/2022 04:23:02 - INFO - codeparrot_training - Step 12084: {'lr': 0.00044749863165197845, 'samples': 2320320, 'steps': 12084, 'loss/train': 1.9289708733558655} 01/29/2022 04:23:06 - INFO - codeparrot_training - Step 12085: {'lr': 0.00044748859917633144, 'samples': 2320512, 'steps': 12085, 'loss/train': 1.5361082553863525} 01/29/2022 04:23:12 - INFO - codeparrot_training - Step 12086: {'lr': 0.00044747856585470604, 'samples': 2320704, 'steps': 12086, 'loss/train': 2.2034338116645813} 01/29/2022 04:23:16 - INFO - codeparrot_training - Step 12087: {'lr': 0.00044746853168714507, 'samples': 2320896, 'steps': 12087, 'loss/train': 0.9686049520969391} 01/29/2022 04:23:21 - INFO - codeparrot_training - Step 12088: {'lr': 0.0004474584966736917, 'samples': 2321088, 'steps': 12088, 'loss/train': 2.0960810780525208} 01/29/2022 04:23:25 - INFO - codeparrot_training - Step 12089: {'lr': 0.00044744846081438874, 'samples': 2321280, 'steps': 12089, 'loss/train': 1.6753313541412354} 01/29/2022 04:23:29 - INFO - codeparrot_training - Step 12090: {'lr': 0.0004474384241092793, 'samples': 2321472, 'steps': 12090, 'loss/train': 1.6092765927314758} 01/29/2022 04:23:34 - INFO - codeparrot_training - Step 12091: {'lr': 0.00044742838655840636, 'samples': 2321664, 'steps': 12091, 'loss/train': 1.7252885699272156} 01/29/2022 04:23:39 - INFO - codeparrot_training - Step 12092: {'lr': 0.0004474183481618129, 'samples': 2321856, 'steps': 12092, 'loss/train': 1.3390056788921356} 01/29/2022 04:23:43 - INFO - codeparrot_training - Step 12093: {'lr': 0.00044740830891954196, 'samples': 2322048, 'steps': 12093, 'loss/train': 0.0717593114823103} 01/29/2022 04:23:47 - INFO - codeparrot_training - Step 12094: {'lr': 0.0004473982688316365, 'samples': 2322240, 'steps': 12094, 'loss/train': 1.88483065366745} 01/29/2022 04:23:51 - INFO - codeparrot_training - Step 12095: {'lr': 0.0004473882278981395, 'samples': 2322432, 'steps': 12095, 'loss/train': 1.7035017013549805} 01/29/2022 04:23:57 - INFO - codeparrot_training - Step 12096: {'lr': 0.000447378186119094, 'samples': 2322624, 'steps': 12096, 'loss/train': 1.9388671517372131} 01/29/2022 04:24:01 - INFO - codeparrot_training - Step 12097: {'lr': 0.00044736814349454303, 'samples': 2322816, 'steps': 12097, 'loss/train': 1.584329903125763} 01/29/2022 04:24:05 - INFO - codeparrot_training - Step 12098: {'lr': 0.0004473581000245296, 'samples': 2323008, 'steps': 12098, 'loss/train': 2.9480981826782227} 01/29/2022 04:24:09 - INFO - codeparrot_training - Step 12099: {'lr': 0.00044734805570909676, 'samples': 2323200, 'steps': 12099, 'loss/train': 0.6535285860300064} 01/29/2022 04:24:14 - INFO - codeparrot_training - Step 12100: {'lr': 0.0004473380105482875, 'samples': 2323392, 'steps': 12100, 'loss/train': 2.0229467153549194} 01/29/2022 04:24:19 - INFO - codeparrot_training - Step 12101: {'lr': 0.0004473279645421449, 'samples': 2323584, 'steps': 12101, 'loss/train': 1.083926796913147} 01/29/2022 04:24:24 - INFO - codeparrot_training - Step 12102: {'lr': 0.00044731791769071197, 'samples': 2323776, 'steps': 12102, 'loss/train': 1.668496549129486} 01/29/2022 04:24:28 - INFO - codeparrot_training - Step 12103: {'lr': 0.00044730786999403166, 'samples': 2323968, 'steps': 12103, 'loss/train': 1.9048047065734863} 01/29/2022 04:24:32 - INFO - codeparrot_training - Step 12104: {'lr': 0.00044729782145214717, 'samples': 2324160, 'steps': 12104, 'loss/train': 0.06679349206387997} 01/29/2022 04:24:38 - INFO - codeparrot_training - Step 12105: {'lr': 0.0004472877720651014, 'samples': 2324352, 'steps': 12105, 'loss/train': 1.8855400085449219} 01/29/2022 04:24:42 - INFO - codeparrot_training - Step 12106: {'lr': 0.0004472777218329375, 'samples': 2324544, 'steps': 12106, 'loss/train': 2.5367977023124695} 01/29/2022 04:24:46 - INFO - codeparrot_training - Step 12107: {'lr': 0.00044726767075569843, 'samples': 2324736, 'steps': 12107, 'loss/train': 1.4864815771579742} 01/29/2022 04:24:50 - INFO - codeparrot_training - Step 12108: {'lr': 0.0004472576188334273, 'samples': 2324928, 'steps': 12108, 'loss/train': 1.1327757239341736} 01/29/2022 04:24:55 - INFO - codeparrot_training - Step 12109: {'lr': 0.00044724756606616726, 'samples': 2325120, 'steps': 12109, 'loss/train': 0.22010140120983124} 01/29/2022 04:25:00 - INFO - codeparrot_training - Step 12110: {'lr': 0.00044723751245396117, 'samples': 2325312, 'steps': 12110, 'loss/train': 1.823180615901947} 01/29/2022 04:25:05 - INFO - codeparrot_training - Step 12111: {'lr': 0.00044722745799685227, 'samples': 2325504, 'steps': 12111, 'loss/train': 2.1249234080314636} 01/29/2022 04:25:09 - INFO - codeparrot_training - Step 12112: {'lr': 0.00044721740269488354, 'samples': 2325696, 'steps': 12112, 'loss/train': 1.8204100728034973} 01/29/2022 04:25:13 - INFO - codeparrot_training - Step 12113: {'lr': 0.0004472073465480981, 'samples': 2325888, 'steps': 12113, 'loss/train': 1.8628866076469421} 01/29/2022 04:25:17 - INFO - codeparrot_training - Step 12114: {'lr': 0.000447197289556539, 'samples': 2326080, 'steps': 12114, 'loss/train': 1.7802295088768005} 01/29/2022 04:25:23 - INFO - codeparrot_training - Step 12115: {'lr': 0.0004471872317202493, 'samples': 2326272, 'steps': 12115, 'loss/train': 1.7734692692756653} 01/29/2022 04:25:27 - INFO - codeparrot_training - Step 12116: {'lr': 0.0004471771730392722, 'samples': 2326464, 'steps': 12116, 'loss/train': 0.8065314888954163} 01/29/2022 04:25:31 - INFO - codeparrot_training - Step 12117: {'lr': 0.00044716711351365057, 'samples': 2326656, 'steps': 12117, 'loss/train': 1.6651344895362854} 01/29/2022 04:25:35 - INFO - codeparrot_training - Step 12118: {'lr': 0.00044715705314342776, 'samples': 2326848, 'steps': 12118, 'loss/train': 1.4290278553962708} 01/29/2022 04:25:40 - INFO - codeparrot_training - Step 12119: {'lr': 0.0004471469919286467, 'samples': 2327040, 'steps': 12119, 'loss/train': 2.0502381920814514} 01/29/2022 04:25:45 - INFO - codeparrot_training - Step 12120: {'lr': 0.0004471369298693505, 'samples': 2327232, 'steps': 12120, 'loss/train': 1.9849700331687927} 01/29/2022 04:25:49 - INFO - codeparrot_training - Step 12121: {'lr': 0.0004471268669655822, 'samples': 2327424, 'steps': 12121, 'loss/train': 1.6153265833854675} 01/29/2022 04:25:53 - INFO - codeparrot_training - Step 12122: {'lr': 0.0004471168032173852, 'samples': 2327616, 'steps': 12122, 'loss/train': 1.5792376399040222} 01/29/2022 04:25:57 - INFO - codeparrot_training - Step 12123: {'lr': 0.0004471067386248023, 'samples': 2327808, 'steps': 12123, 'loss/train': 1.5185195803642273} 01/29/2022 04:26:02 - INFO - codeparrot_training - Step 12124: {'lr': 0.0004470966731878767, 'samples': 2328000, 'steps': 12124, 'loss/train': 1.9692880511283875} 01/29/2022 04:26:08 - INFO - codeparrot_training - Step 12125: {'lr': 0.0004470866069066516, 'samples': 2328192, 'steps': 12125, 'loss/train': 0.9383422136306763} 01/29/2022 04:26:12 - INFO - codeparrot_training - Step 12126: {'lr': 0.00044707653978117004, 'samples': 2328384, 'steps': 12126, 'loss/train': 1.9976829886436462} 01/29/2022 04:26:16 - INFO - codeparrot_training - Step 12127: {'lr': 0.00044706647181147507, 'samples': 2328576, 'steps': 12127, 'loss/train': 1.6995429396629333} 01/29/2022 04:26:20 - INFO - codeparrot_training - Step 12128: {'lr': 0.00044705640299761004, 'samples': 2328768, 'steps': 12128, 'loss/train': 0.8417412042617798} 01/29/2022 04:26:25 - INFO - codeparrot_training - Step 12129: {'lr': 0.0004470463333396179, 'samples': 2328960, 'steps': 12129, 'loss/train': 1.410253643989563} 01/29/2022 04:26:30 - INFO - codeparrot_training - Step 12130: {'lr': 0.0004470362628375418, 'samples': 2329152, 'steps': 12130, 'loss/train': 1.7066954970359802} 01/29/2022 04:26:34 - INFO - codeparrot_training - Step 12131: {'lr': 0.000447026191491425, 'samples': 2329344, 'steps': 12131, 'loss/train': 2.228016972541809} 01/29/2022 04:26:38 - INFO - codeparrot_training - Step 12132: {'lr': 0.0004470161193013105, 'samples': 2329536, 'steps': 12132, 'loss/train': 1.5174502730369568} 01/29/2022 04:26:43 - INFO - codeparrot_training - Step 12133: {'lr': 0.0004470060462672415, 'samples': 2329728, 'steps': 12133, 'loss/train': 1.492167055606842} 01/29/2022 04:26:48 - INFO - codeparrot_training - Step 12134: {'lr': 0.0004469959723892612, 'samples': 2329920, 'steps': 12134, 'loss/train': 2.066657066345215} 01/29/2022 04:26:53 - INFO - codeparrot_training - Step 12135: {'lr': 0.0004469858976674126, 'samples': 2330112, 'steps': 12135, 'loss/train': 2.5521790981292725} 01/29/2022 04:26:57 - INFO - codeparrot_training - Step 12136: {'lr': 0.000446975822101739, 'samples': 2330304, 'steps': 12136, 'loss/train': 0.16567141190171242} 01/29/2022 04:27:01 - INFO - codeparrot_training - Step 12137: {'lr': 0.00044696574569228365, 'samples': 2330496, 'steps': 12137, 'loss/train': 2.4788069128990173} 01/29/2022 04:27:06 - INFO - codeparrot_training - Step 12138: {'lr': 0.00044695566843908947, 'samples': 2330688, 'steps': 12138, 'loss/train': 1.303817331790924} 01/29/2022 04:27:11 - INFO - codeparrot_training - Step 12139: {'lr': 0.0004469455903421998, 'samples': 2330880, 'steps': 12139, 'loss/train': 1.9133142828941345} 01/29/2022 04:27:15 - INFO - codeparrot_training - Step 12140: {'lr': 0.0004469355114016577, 'samples': 2331072, 'steps': 12140, 'loss/train': 2.83357697725296} 01/29/2022 04:27:19 - INFO - codeparrot_training - Step 12141: {'lr': 0.0004469254316175065, 'samples': 2331264, 'steps': 12141, 'loss/train': 2.944678544998169} 01/29/2022 04:27:23 - INFO - codeparrot_training - Step 12142: {'lr': 0.0004469153509897892, 'samples': 2331456, 'steps': 12142, 'loss/train': 1.2424110174179077} 01/29/2022 04:27:28 - INFO - codeparrot_training - Step 12143: {'lr': 0.00044690526951854907, 'samples': 2331648, 'steps': 12143, 'loss/train': 1.8677929043769836} 01/29/2022 04:27:33 - INFO - codeparrot_training - Step 12144: {'lr': 0.0004468951872038293, 'samples': 2331840, 'steps': 12144, 'loss/train': 1.962890625} 01/29/2022 04:27:37 - INFO - codeparrot_training - Step 12145: {'lr': 0.00044688510404567307, 'samples': 2332032, 'steps': 12145, 'loss/train': 1.8149478435516357} 01/29/2022 04:27:42 - INFO - codeparrot_training - Step 12146: {'lr': 0.0004468750200441236, 'samples': 2332224, 'steps': 12146, 'loss/train': 1.739150583744049} 01/29/2022 04:27:46 - INFO - codeparrot_training - Step 12147: {'lr': 0.00044686493519922405, 'samples': 2332416, 'steps': 12147, 'loss/train': 1.7361919283866882} 01/29/2022 04:27:50 - INFO - codeparrot_training - Step 12148: {'lr': 0.00044685484951101763, 'samples': 2332608, 'steps': 12148, 'loss/train': 1.98518568277359} 01/29/2022 04:27:55 - INFO - codeparrot_training - Step 12149: {'lr': 0.0004468447629795475, 'samples': 2332800, 'steps': 12149, 'loss/train': 1.3555321097373962} 01/29/2022 04:28:00 - INFO - codeparrot_training - Step 12150: {'lr': 0.00044683467560485696, 'samples': 2332992, 'steps': 12150, 'loss/train': 1.9140323996543884} 01/29/2022 04:28:04 - INFO - codeparrot_training - Step 12151: {'lr': 0.00044682458738698916, 'samples': 2333184, 'steps': 12151, 'loss/train': 1.169518768787384} 01/29/2022 04:28:08 - INFO - codeparrot_training - Step 12152: {'lr': 0.0004468144983259873, 'samples': 2333376, 'steps': 12152, 'loss/train': 2.0194973945617676} 01/29/2022 04:28:12 - INFO - codeparrot_training - Step 12153: {'lr': 0.00044680440842189464, 'samples': 2333568, 'steps': 12153, 'loss/train': 1.1532044112682343} 01/29/2022 04:28:19 - INFO - codeparrot_training - Step 12154: {'lr': 0.0004467943176747544, 'samples': 2333760, 'steps': 12154, 'loss/train': 1.833475112915039} 01/29/2022 04:28:23 - INFO - codeparrot_training - Step 12155: {'lr': 0.0004467842260846098, 'samples': 2333952, 'steps': 12155, 'loss/train': 1.9384133219718933} 01/29/2022 04:28:27 - INFO - codeparrot_training - Step 12156: {'lr': 0.00044677413365150397, 'samples': 2334144, 'steps': 12156, 'loss/train': 1.0114289224147797} 01/29/2022 04:28:31 - INFO - codeparrot_training - Step 12157: {'lr': 0.00044676404037548035, 'samples': 2334336, 'steps': 12157, 'loss/train': 1.3623916804790497} 01/29/2022 04:28:36 - INFO - codeparrot_training - Step 12158: {'lr': 0.0004467539462565821, 'samples': 2334528, 'steps': 12158, 'loss/train': 1.956975281238556} 01/29/2022 04:28:41 - INFO - codeparrot_training - Step 12159: {'lr': 0.0004467438512948523, 'samples': 2334720, 'steps': 12159, 'loss/train': 1.9958211779594421} 01/29/2022 04:28:45 - INFO - codeparrot_training - Step 12160: {'lr': 0.00044673375549033435, 'samples': 2334912, 'steps': 12160, 'loss/train': 1.6213755011558533} 01/29/2022 04:28:49 - INFO - codeparrot_training - Step 12161: {'lr': 0.0004467236588430714, 'samples': 2335104, 'steps': 12161, 'loss/train': 1.4915800094604492} 01/29/2022 04:28:54 - INFO - codeparrot_training - Step 12162: {'lr': 0.00044671356135310685, 'samples': 2335296, 'steps': 12162, 'loss/train': 1.577202022075653} 01/29/2022 04:28:58 - INFO - codeparrot_training - Step 12163: {'lr': 0.0004467034630204839, 'samples': 2335488, 'steps': 12163, 'loss/train': 2.352143347263336} 01/29/2022 04:29:03 - INFO - codeparrot_training - Step 12164: {'lr': 0.0004466933638452457, 'samples': 2335680, 'steps': 12164, 'loss/train': 1.7505578398704529} 01/29/2022 04:29:08 - INFO - codeparrot_training - Step 12165: {'lr': 0.0004466832638274356, 'samples': 2335872, 'steps': 12165, 'loss/train': 0.6905261278152466} 01/29/2022 04:29:12 - INFO - codeparrot_training - Step 12166: {'lr': 0.0004466731629670969, 'samples': 2336064, 'steps': 12166, 'loss/train': 1.9774351716041565} 01/29/2022 04:29:16 - INFO - codeparrot_training - Step 12167: {'lr': 0.00044666306126427276, 'samples': 2336256, 'steps': 12167, 'loss/train': 1.5618590712547302} 01/29/2022 04:29:20 - INFO - codeparrot_training - Step 12168: {'lr': 0.00044665295871900655, 'samples': 2336448, 'steps': 12168, 'loss/train': 1.828683614730835} 01/29/2022 04:29:26 - INFO - codeparrot_training - Step 12169: {'lr': 0.0004466428553313415, 'samples': 2336640, 'steps': 12169, 'loss/train': 1.1480599343776703} 01/29/2022 04:29:31 - INFO - codeparrot_training - Step 12170: {'lr': 0.0004466327511013208, 'samples': 2336832, 'steps': 12170, 'loss/train': 1.6662815809249878} 01/29/2022 04:29:35 - INFO - codeparrot_training - Step 12171: {'lr': 0.00044662264602898794, 'samples': 2337024, 'steps': 12171, 'loss/train': 1.1730671525001526} 01/29/2022 04:29:39 - INFO - codeparrot_training - Step 12172: {'lr': 0.00044661254011438614, 'samples': 2337216, 'steps': 12172, 'loss/train': 1.6144307255744934} 01/29/2022 04:29:43 - INFO - codeparrot_training - Step 12173: {'lr': 0.00044660243335755854, 'samples': 2337408, 'steps': 12173, 'loss/train': 1.5288995504379272} 01/29/2022 04:29:49 - INFO - codeparrot_training - Step 12174: {'lr': 0.00044659232575854866, 'samples': 2337600, 'steps': 12174, 'loss/train': 0.9281249642372131} 01/29/2022 04:29:53 - INFO - codeparrot_training - Step 12175: {'lr': 0.00044658221731739954, 'samples': 2337792, 'steps': 12175, 'loss/train': 1.9980647563934326} 01/29/2022 04:29:57 - INFO - codeparrot_training - Step 12176: {'lr': 0.0004465721080341547, 'samples': 2337984, 'steps': 12176, 'loss/train': 2.018482267856598} 01/29/2022 04:30:01 - INFO - codeparrot_training - Step 12177: {'lr': 0.00044656199790885743, 'samples': 2338176, 'steps': 12177, 'loss/train': 1.26471608877182} 01/29/2022 04:30:06 - INFO - codeparrot_training - Step 12178: {'lr': 0.0004465518869415509, 'samples': 2338368, 'steps': 12178, 'loss/train': 1.5259262323379517} 01/29/2022 04:30:11 - INFO - codeparrot_training - Step 12179: {'lr': 0.0004465417751322785, 'samples': 2338560, 'steps': 12179, 'loss/train': 2.1742372512817383} 01/29/2022 04:30:16 - INFO - codeparrot_training - Step 12180: {'lr': 0.00044653166248108357, 'samples': 2338752, 'steps': 12180, 'loss/train': 1.422169178724289} 01/29/2022 04:30:20 - INFO - codeparrot_training - Step 12181: {'lr': 0.00044652154898800937, 'samples': 2338944, 'steps': 12181, 'loss/train': 1.8934024572372437} 01/29/2022 04:30:24 - INFO - codeparrot_training - Step 12182: {'lr': 0.0004465114346530993, 'samples': 2339136, 'steps': 12182, 'loss/train': 2.3474584221839905} 01/29/2022 04:30:28 - INFO - codeparrot_training - Step 12183: {'lr': 0.0004465013194763966, 'samples': 2339328, 'steps': 12183, 'loss/train': 0.0707087516784668} 01/29/2022 04:30:34 - INFO - codeparrot_training - Step 12184: {'lr': 0.0004464912034579447, 'samples': 2339520, 'steps': 12184, 'loss/train': 0.9263564944267273} 01/29/2022 04:30:38 - INFO - codeparrot_training - Step 12185: {'lr': 0.00044648108659778687, 'samples': 2339712, 'steps': 12185, 'loss/train': 1.7533705830574036} 01/29/2022 04:30:42 - INFO - codeparrot_training - Step 12186: {'lr': 0.0004464709688959664, 'samples': 2339904, 'steps': 12186, 'loss/train': 2.2518430352211} 01/29/2022 04:30:46 - INFO - codeparrot_training - Step 12187: {'lr': 0.0004464608503525267, 'samples': 2340096, 'steps': 12187, 'loss/train': 1.1317309141159058} 01/29/2022 04:30:51 - INFO - codeparrot_training - Step 12188: {'lr': 0.0004464507309675111, 'samples': 2340288, 'steps': 12188, 'loss/train': 1.7305957674980164} 01/29/2022 04:30:56 - INFO - codeparrot_training - Step 12189: {'lr': 0.000446440610740963, 'samples': 2340480, 'steps': 12189, 'loss/train': 2.285313069820404} 01/29/2022 04:31:00 - INFO - codeparrot_training - Step 12190: {'lr': 0.0004464304896729257, 'samples': 2340672, 'steps': 12190, 'loss/train': 1.3491752743721008} 01/29/2022 04:31:05 - INFO - codeparrot_training - Step 12191: {'lr': 0.0004464203677634424, 'samples': 2340864, 'steps': 12191, 'loss/train': 2.320017099380493} 01/29/2022 04:31:09 - INFO - codeparrot_training - Step 12192: {'lr': 0.0004464102450125568, 'samples': 2341056, 'steps': 12192, 'loss/train': 1.437995821237564} 01/29/2022 04:31:15 - INFO - codeparrot_training - Step 12193: {'lr': 0.00044640012142031196, 'samples': 2341248, 'steps': 12193, 'loss/train': 1.579757273197174} 01/29/2022 04:31:19 - INFO - codeparrot_training - Step 12194: {'lr': 0.0004463899969867514, 'samples': 2341440, 'steps': 12194, 'loss/train': 2.004788875579834} 01/29/2022 04:31:24 - INFO - codeparrot_training - Step 12195: {'lr': 0.0004463798717119185, 'samples': 2341632, 'steps': 12195, 'loss/train': 1.8106847405433655} 01/29/2022 04:31:28 - INFO - codeparrot_training - Step 12196: {'lr': 0.00044636974559585655, 'samples': 2341824, 'steps': 12196, 'loss/train': 2.928601026535034} 01/29/2022 04:31:32 - INFO - codeparrot_training - Step 12197: {'lr': 0.00044635961863860894, 'samples': 2342016, 'steps': 12197, 'loss/train': 0.07849875837564468} 01/29/2022 04:31:37 - INFO - codeparrot_training - Step 12198: {'lr': 0.00044634949084021913, 'samples': 2342208, 'steps': 12198, 'loss/train': 1.0419245660305023} 01/29/2022 04:31:42 - INFO - codeparrot_training - Step 12199: {'lr': 0.0004463393622007305, 'samples': 2342400, 'steps': 12199, 'loss/train': 1.0912413597106934} 01/29/2022 04:31:46 - INFO - codeparrot_training - Step 12200: {'lr': 0.0004463292327201862, 'samples': 2342592, 'steps': 12200, 'loss/train': 1.8001019954681396} 01/29/2022 04:31:50 - INFO - codeparrot_training - Step 12201: {'lr': 0.0004463191023986299, 'samples': 2342784, 'steps': 12201, 'loss/train': 0.47257277369499207} 01/29/2022 04:31:54 - INFO - codeparrot_training - Step 12202: {'lr': 0.00044630897123610497, 'samples': 2342976, 'steps': 12202, 'loss/train': 0.8826001882553101} 01/29/2022 04:32:00 - INFO - codeparrot_training - Step 12203: {'lr': 0.0004462988392326547, 'samples': 2343168, 'steps': 12203, 'loss/train': 1.326718121767044} 01/29/2022 04:32:04 - INFO - codeparrot_training - Step 12204: {'lr': 0.00044628870638832254, 'samples': 2343360, 'steps': 12204, 'loss/train': 1.6888380646705627} 01/29/2022 04:32:08 - INFO - codeparrot_training - Step 12205: {'lr': 0.00044627857270315187, 'samples': 2343552, 'steps': 12205, 'loss/train': 1.6480770707130432} 01/29/2022 04:32:12 - INFO - codeparrot_training - Step 12206: {'lr': 0.00044626843817718615, 'samples': 2343744, 'steps': 12206, 'loss/train': 1.777711272239685} 01/29/2022 04:32:17 - INFO - codeparrot_training - Step 12207: {'lr': 0.00044625830281046875, 'samples': 2343936, 'steps': 12207, 'loss/train': 1.32471764087677} 01/29/2022 04:32:23 - INFO - codeparrot_training - Step 12208: {'lr': 0.0004462481666030431, 'samples': 2344128, 'steps': 12208, 'loss/train': 1.3939442932605743} 01/29/2022 04:32:27 - INFO - codeparrot_training - Step 12209: {'lr': 0.0004462380295549526, 'samples': 2344320, 'steps': 12209, 'loss/train': 1.2418769896030426} 01/29/2022 04:32:31 - INFO - codeparrot_training - Step 12210: {'lr': 0.0004462278916662407, 'samples': 2344512, 'steps': 12210, 'loss/train': 1.851535677909851} 01/29/2022 04:32:36 - INFO - codeparrot_training - Step 12211: {'lr': 0.00044621775293695085, 'samples': 2344704, 'steps': 12211, 'loss/train': 1.5475571751594543} 01/29/2022 04:32:40 - INFO - codeparrot_training - Step 12212: {'lr': 0.00044620761336712646, 'samples': 2344896, 'steps': 12212, 'loss/train': 1.7895948886871338} 01/29/2022 04:32:45 - INFO - codeparrot_training - Step 12213: {'lr': 0.0004461974729568109, 'samples': 2345088, 'steps': 12213, 'loss/train': 1.5134101510047913} 01/29/2022 04:32:49 - INFO - codeparrot_training - Step 12214: {'lr': 0.0004461873317060477, 'samples': 2345280, 'steps': 12214, 'loss/train': 3.7466357946395874} 01/29/2022 04:32:54 - INFO - codeparrot_training - Step 12215: {'lr': 0.00044617718961488024, 'samples': 2345472, 'steps': 12215, 'loss/train': 1.4613256752490997} 01/29/2022 04:32:58 - INFO - codeparrot_training - Step 12216: {'lr': 0.000446167046683352, 'samples': 2345664, 'steps': 12216, 'loss/train': 1.7510663866996765} 01/29/2022 04:33:02 - INFO - codeparrot_training - Step 12217: {'lr': 0.0004461569029115065, 'samples': 2345856, 'steps': 12217, 'loss/train': 1.8598136901855469} 01/29/2022 04:33:08 - INFO - codeparrot_training - Step 12218: {'lr': 0.000446146758299387, 'samples': 2346048, 'steps': 12218, 'loss/train': 1.8181596994400024} 01/29/2022 04:33:12 - INFO - codeparrot_training - Step 12219: {'lr': 0.0004461366128470371, 'samples': 2346240, 'steps': 12219, 'loss/train': 5.797663450241089} 01/29/2022 04:33:17 - INFO - codeparrot_training - Step 12220: {'lr': 0.0004461264665545003, 'samples': 2346432, 'steps': 12220, 'loss/train': 2.208126425743103} 01/29/2022 04:33:21 - INFO - codeparrot_training - Step 12221: {'lr': 0.00044611631942182, 'samples': 2346624, 'steps': 12221, 'loss/train': 2.8876540660858154} 01/29/2022 04:33:25 - INFO - codeparrot_training - Step 12222: {'lr': 0.0004461061714490395, 'samples': 2346816, 'steps': 12222, 'loss/train': 1.556031346321106} 01/29/2022 04:33:30 - INFO - codeparrot_training - Step 12223: {'lr': 0.0004460960226362026, 'samples': 2347008, 'steps': 12223, 'loss/train': 2.4447221159934998} 01/29/2022 04:33:34 - INFO - codeparrot_training - Step 12224: {'lr': 0.0004460858729833525, 'samples': 2347200, 'steps': 12224, 'loss/train': 0.774033933877945} 01/29/2022 04:33:39 - INFO - codeparrot_training - Step 12225: {'lr': 0.00044607572249053283, 'samples': 2347392, 'steps': 12225, 'loss/train': 5.023462414741516} 01/29/2022 04:33:43 - INFO - codeparrot_training - Step 12226: {'lr': 0.0004460655711577871, 'samples': 2347584, 'steps': 12226, 'loss/train': 1.9238753914833069} 01/29/2022 04:33:47 - INFO - codeparrot_training - Step 12227: {'lr': 0.00044605541898515863, 'samples': 2347776, 'steps': 12227, 'loss/train': 1.6857320666313171} 01/29/2022 04:33:52 - INFO - codeparrot_training - Step 12228: {'lr': 0.00044604526597269103, 'samples': 2347968, 'steps': 12228, 'loss/train': 1.1733031868934631} 01/29/2022 04:33:57 - INFO - codeparrot_training - Step 12229: {'lr': 0.0004460351121204277, 'samples': 2348160, 'steps': 12229, 'loss/train': 0.849522739648819} 01/29/2022 04:34:01 - INFO - codeparrot_training - Step 12230: {'lr': 0.00044602495742841226, 'samples': 2348352, 'steps': 12230, 'loss/train': 3.7145851850509644} 01/29/2022 04:34:05 - INFO - codeparrot_training - Step 12231: {'lr': 0.00044601480189668816, 'samples': 2348544, 'steps': 12231, 'loss/train': 1.980012595653534} 01/29/2022 04:34:09 - INFO - codeparrot_training - Step 12232: {'lr': 0.00044600464552529886, 'samples': 2348736, 'steps': 12232, 'loss/train': 1.9669665098190308} 01/29/2022 04:34:15 - INFO - codeparrot_training - Step 12233: {'lr': 0.0004459944883142879, 'samples': 2348928, 'steps': 12233, 'loss/train': 1.013300746679306} 01/29/2022 04:34:19 - INFO - codeparrot_training - Step 12234: {'lr': 0.0004459843302636988, 'samples': 2349120, 'steps': 12234, 'loss/train': 0.7321380972862244} 01/29/2022 04:34:24 - INFO - codeparrot_training - Step 12235: {'lr': 0.000445974171373575, 'samples': 2349312, 'steps': 12235, 'loss/train': 0.524231493473053} 01/29/2022 04:34:28 - INFO - codeparrot_training - Step 12236: {'lr': 0.0004459640116439602, 'samples': 2349504, 'steps': 12236, 'loss/train': 0.8725821375846863} 01/29/2022 04:34:32 - INFO - codeparrot_training - Step 12237: {'lr': 0.0004459538510748977, 'samples': 2349696, 'steps': 12237, 'loss/train': 2.073067009449005} 01/29/2022 04:34:38 - INFO - codeparrot_training - Step 12238: {'lr': 0.0004459436896664312, 'samples': 2349888, 'steps': 12238, 'loss/train': 1.592560887336731} 01/29/2022 04:34:42 - INFO - codeparrot_training - Step 12239: {'lr': 0.00044593352741860404, 'samples': 2350080, 'steps': 12239, 'loss/train': 1.6965224146842957} 01/29/2022 04:34:46 - INFO - codeparrot_training - Step 12240: {'lr': 0.00044592336433145995, 'samples': 2350272, 'steps': 12240, 'loss/train': 1.6927741169929504} 01/29/2022 04:34:51 - INFO - codeparrot_training - Step 12241: {'lr': 0.00044591320040504237, 'samples': 2350464, 'steps': 12241, 'loss/train': 2.336155593395233} 01/29/2022 04:34:55 - INFO - codeparrot_training - Step 12242: {'lr': 0.00044590303563939485, 'samples': 2350656, 'steps': 12242, 'loss/train': 0.7076707631349564} 01/29/2022 04:35:00 - INFO - codeparrot_training - Step 12243: {'lr': 0.0004458928700345609, 'samples': 2350848, 'steps': 12243, 'loss/train': 1.5440742373466492} 01/29/2022 04:35:05 - INFO - codeparrot_training - Step 12244: {'lr': 0.00044588270359058416, 'samples': 2351040, 'steps': 12244, 'loss/train': 2.1461387872695923} 01/29/2022 04:35:09 - INFO - codeparrot_training - Step 12245: {'lr': 0.000445872536307508, 'samples': 2351232, 'steps': 12245, 'loss/train': 1.4258633852005005} 01/29/2022 04:35:13 - INFO - codeparrot_training - Step 12246: {'lr': 0.0004458623681853762, 'samples': 2351424, 'steps': 12246, 'loss/train': 1.545150876045227} 01/29/2022 04:35:18 - INFO - codeparrot_training - Step 12247: {'lr': 0.0004458521992242322, 'samples': 2351616, 'steps': 12247, 'loss/train': 0.5657627731561661} 01/29/2022 04:35:23 - INFO - codeparrot_training - Step 12248: {'lr': 0.00044584202942411956, 'samples': 2351808, 'steps': 12248, 'loss/train': 1.473226010799408} 01/29/2022 04:35:27 - INFO - codeparrot_training - Step 12249: {'lr': 0.00044583185878508183, 'samples': 2352000, 'steps': 12249, 'loss/train': 1.0804018378257751} 01/29/2022 04:35:31 - INFO - codeparrot_training - Step 12250: {'lr': 0.0004458216873071626, 'samples': 2352192, 'steps': 12250, 'loss/train': 1.306284874677658} 01/29/2022 04:35:35 - INFO - codeparrot_training - Step 12251: {'lr': 0.00044581151499040547, 'samples': 2352384, 'steps': 12251, 'loss/train': 2.673767566680908} 01/29/2022 04:35:42 - INFO - codeparrot_training - Step 12252: {'lr': 0.000445801341834854, 'samples': 2352576, 'steps': 12252, 'loss/train': 1.9693593978881836} 01/29/2022 04:35:46 - INFO - codeparrot_training - Step 12253: {'lr': 0.0004457911678405517, 'samples': 2352768, 'steps': 12253, 'loss/train': 1.5457883477210999} 01/29/2022 04:35:50 - INFO - codeparrot_training - Step 12254: {'lr': 0.0004457809930075422, 'samples': 2352960, 'steps': 12254, 'loss/train': 7.359470844268799} 01/29/2022 04:35:54 - INFO - codeparrot_training - Step 12255: {'lr': 0.0004457708173358691, 'samples': 2353152, 'steps': 12255, 'loss/train': 1.9617093801498413} 01/29/2022 04:35:59 - INFO - codeparrot_training - Step 12256: {'lr': 0.00044576064082557605, 'samples': 2353344, 'steps': 12256, 'loss/train': 2.0154502987861633} 01/29/2022 04:36:04 - INFO - codeparrot_training - Step 12257: {'lr': 0.0004457504634767066, 'samples': 2353536, 'steps': 12257, 'loss/train': 2.0825706124305725} 01/29/2022 04:36:08 - INFO - codeparrot_training - Step 12258: {'lr': 0.0004457402852893042, 'samples': 2353728, 'steps': 12258, 'loss/train': 1.8670225739479065} 01/29/2022 04:36:12 - INFO - codeparrot_training - Step 12259: {'lr': 0.0004457301062634126, 'samples': 2353920, 'steps': 12259, 'loss/train': 1.9891195893287659} 01/29/2022 04:36:17 - INFO - codeparrot_training - Step 12260: {'lr': 0.0004457199263990754, 'samples': 2354112, 'steps': 12260, 'loss/train': 1.955865740776062} 01/29/2022 04:36:21 - INFO - codeparrot_training - Step 12261: {'lr': 0.0004457097456963362, 'samples': 2354304, 'steps': 12261, 'loss/train': 1.46025949716568} 01/29/2022 04:36:27 - INFO - codeparrot_training - Step 12262: {'lr': 0.0004456995641552386, 'samples': 2354496, 'steps': 12262, 'loss/train': 0.9616927206516266} 01/29/2022 04:36:31 - INFO - codeparrot_training - Step 12263: {'lr': 0.0004456893817758262, 'samples': 2354688, 'steps': 12263, 'loss/train': 1.5633707642555237} 01/29/2022 04:36:35 - INFO - codeparrot_training - Step 12264: {'lr': 0.00044567919855814257, 'samples': 2354880, 'steps': 12264, 'loss/train': 1.619301438331604} 01/29/2022 04:36:39 - INFO - codeparrot_training - Step 12265: {'lr': 0.0004456690145022314, 'samples': 2355072, 'steps': 12265, 'loss/train': 1.3987350761890411} 01/29/2022 04:36:44 - INFO - codeparrot_training - Step 12266: {'lr': 0.0004456588296081364, 'samples': 2355264, 'steps': 12266, 'loss/train': 1.3398469984531403} 01/29/2022 04:36:49 - INFO - codeparrot_training - Step 12267: {'lr': 0.000445648643875901, 'samples': 2355456, 'steps': 12267, 'loss/train': 1.8931387066841125} 01/29/2022 04:36:53 - INFO - codeparrot_training - Step 12268: {'lr': 0.000445638457305569, 'samples': 2355648, 'steps': 12268, 'loss/train': 2.3295541405677795} 01/29/2022 04:36:57 - INFO - codeparrot_training - Step 12269: {'lr': 0.00044562826989718397, 'samples': 2355840, 'steps': 12269, 'loss/train': 1.8650413155555725} 01/29/2022 04:37:02 - INFO - codeparrot_training - Step 12270: {'lr': 0.00044561808165078954, 'samples': 2356032, 'steps': 12270, 'loss/train': 1.5982296466827393} 01/29/2022 04:37:06 - INFO - codeparrot_training - Step 12271: {'lr': 0.0004456078925664293, 'samples': 2356224, 'steps': 12271, 'loss/train': 1.7742964625358582} 01/29/2022 04:37:11 - INFO - codeparrot_training - Step 12272: {'lr': 0.000445597702644147, 'samples': 2356416, 'steps': 12272, 'loss/train': 1.4682349562644958} 01/29/2022 04:37:15 - INFO - codeparrot_training - Step 12273: {'lr': 0.0004455875118839863, 'samples': 2356608, 'steps': 12273, 'loss/train': 1.4697888493537903} 01/29/2022 04:37:20 - INFO - codeparrot_training - Step 12274: {'lr': 0.00044557732028599077, 'samples': 2356800, 'steps': 12274, 'loss/train': 1.2547979056835175} 01/29/2022 04:37:24 - INFO - codeparrot_training - Step 12275: {'lr': 0.0004455671278502041, 'samples': 2356992, 'steps': 12275, 'loss/train': 1.6165273189544678} 01/29/2022 04:37:28 - INFO - codeparrot_training - Step 12276: {'lr': 0.00044555693457667, 'samples': 2357184, 'steps': 12276, 'loss/train': 1.0275557935237885} 01/29/2022 04:37:34 - INFO - codeparrot_training - Step 12277: {'lr': 0.000445546740465432, 'samples': 2357376, 'steps': 12277, 'loss/train': 1.578379511833191} 01/29/2022 04:37:38 - INFO - codeparrot_training - Step 12278: {'lr': 0.00044553654551653387, 'samples': 2357568, 'steps': 12278, 'loss/train': 1.3666019439697266} 01/29/2022 04:37:43 - INFO - codeparrot_training - Step 12279: {'lr': 0.0004455263497300194, 'samples': 2357760, 'steps': 12279, 'loss/train': 1.1799644529819489} 01/29/2022 04:37:47 - INFO - codeparrot_training - Step 12280: {'lr': 0.000445516153105932, 'samples': 2357952, 'steps': 12280, 'loss/train': 2.0393094420433044} 01/29/2022 04:37:51 - INFO - codeparrot_training - Step 12281: {'lr': 0.0004455059556443155, 'samples': 2358144, 'steps': 12281, 'loss/train': 2.404369533061981} 01/29/2022 04:37:57 - INFO - codeparrot_training - Step 12282: {'lr': 0.0004454957573452136, 'samples': 2358336, 'steps': 12282, 'loss/train': 1.6574222445487976} 01/29/2022 04:38:01 - INFO - codeparrot_training - Step 12283: {'lr': 0.0004454855582086699, 'samples': 2358528, 'steps': 12283, 'loss/train': 1.5458058714866638} 01/29/2022 04:38:05 - INFO - codeparrot_training - Step 12284: {'lr': 0.0004454753582347282, 'samples': 2358720, 'steps': 12284, 'loss/train': 2.1017505526542664} 01/29/2022 04:38:09 - INFO - codeparrot_training - Step 12285: {'lr': 0.00044546515742343207, 'samples': 2358912, 'steps': 12285, 'loss/train': 2.127759039402008} 01/29/2022 04:38:14 - INFO - codeparrot_training - Step 12286: {'lr': 0.00044545495577482535, 'samples': 2359104, 'steps': 12286, 'loss/train': 2.1030174493789673} 01/29/2022 04:38:19 - INFO - codeparrot_training - Step 12287: {'lr': 0.00044544475328895164, 'samples': 2359296, 'steps': 12287, 'loss/train': 1.3339869976043701} 01/29/2022 04:38:24 - INFO - codeparrot_training - Step 12288: {'lr': 0.00044543454996585463, 'samples': 2359488, 'steps': 12288, 'loss/train': 1.112989068031311} 01/29/2022 04:38:28 - INFO - codeparrot_training - Step 12289: {'lr': 0.0004454243458055781, 'samples': 2359680, 'steps': 12289, 'loss/train': 2.3055755496025085} 01/29/2022 04:38:32 - INFO - codeparrot_training - Step 12290: {'lr': 0.00044541414080816573, 'samples': 2359872, 'steps': 12290, 'loss/train': 1.8561471104621887} 01/29/2022 04:38:36 - INFO - codeparrot_training - Step 12291: {'lr': 0.00044540393497366124, 'samples': 2360064, 'steps': 12291, 'loss/train': 1.6540557146072388} 01/29/2022 04:38:42 - INFO - codeparrot_training - Step 12292: {'lr': 0.00044539372830210833, 'samples': 2360256, 'steps': 12292, 'loss/train': 2.239734470844269} 01/29/2022 04:38:46 - INFO - codeparrot_training - Step 12293: {'lr': 0.0004453835207935507, 'samples': 2360448, 'steps': 12293, 'loss/train': 1.5197898745536804} 01/29/2022 04:38:50 - INFO - codeparrot_training - Step 12294: {'lr': 0.0004453733124480321, 'samples': 2360640, 'steps': 12294, 'loss/train': 1.4231794774532318} 01/29/2022 04:38:54 - INFO - codeparrot_training - Step 12295: {'lr': 0.0004453631032655964, 'samples': 2360832, 'steps': 12295, 'loss/train': 1.3597119748592377} 01/29/2022 04:38:59 - INFO - codeparrot_training - Step 12296: {'lr': 0.00044535289324628704, 'samples': 2361024, 'steps': 12296, 'loss/train': 2.2332642674446106} 01/29/2022 04:39:05 - INFO - codeparrot_training - Step 12297: {'lr': 0.00044534268239014796, 'samples': 2361216, 'steps': 12297, 'loss/train': 2.6376526951789856} 01/29/2022 04:39:09 - INFO - codeparrot_training - Step 12298: {'lr': 0.00044533247069722295, 'samples': 2361408, 'steps': 12298, 'loss/train': 1.5222984552383423} 01/29/2022 04:39:13 - INFO - codeparrot_training - Step 12299: {'lr': 0.0004453222581675556, 'samples': 2361600, 'steps': 12299, 'loss/train': 1.880542516708374} 01/29/2022 04:39:17 - INFO - codeparrot_training - Step 12300: {'lr': 0.0004453120448011897, 'samples': 2361792, 'steps': 12300, 'loss/train': 1.9287487864494324} 01/29/2022 04:39:22 - INFO - codeparrot_training - Step 12301: {'lr': 0.00044530183059816896, 'samples': 2361984, 'steps': 12301, 'loss/train': 1.536081612110138} 01/29/2022 04:39:27 - INFO - codeparrot_training - Step 12302: {'lr': 0.00044529161555853725, 'samples': 2362176, 'steps': 12302, 'loss/train': 1.4695925116539001} 01/29/2022 04:39:31 - INFO - codeparrot_training - Step 12303: {'lr': 0.0004452813996823383, 'samples': 2362368, 'steps': 12303, 'loss/train': 1.434014081954956} 01/29/2022 04:39:35 - INFO - codeparrot_training - Step 12304: {'lr': 0.00044527118296961576, 'samples': 2362560, 'steps': 12304, 'loss/train': 1.307908147573471} 01/29/2022 04:39:40 - INFO - codeparrot_training - Step 12305: {'lr': 0.0004452609654204136, 'samples': 2362752, 'steps': 12305, 'loss/train': 1.8624899983406067} 01/29/2022 04:39:44 - INFO - codeparrot_training - Step 12306: {'lr': 0.0004452507470347754, 'samples': 2362944, 'steps': 12306, 'loss/train': 1.573940634727478} 01/29/2022 04:39:50 - INFO - codeparrot_training - Step 12307: {'lr': 0.00044524052781274497, 'samples': 2363136, 'steps': 12307, 'loss/train': 1.2799174189567566} 01/29/2022 04:39:54 - INFO - codeparrot_training - Step 12308: {'lr': 0.00044523030775436617, 'samples': 2363328, 'steps': 12308, 'loss/train': 0.7295094877481461} 01/29/2022 04:39:58 - INFO - codeparrot_training - Step 12309: {'lr': 0.0004452200868596827, 'samples': 2363520, 'steps': 12309, 'loss/train': 2.173540771007538} 01/29/2022 04:40:03 - INFO - codeparrot_training - Step 12310: {'lr': 0.0004452098651287384, 'samples': 2363712, 'steps': 12310, 'loss/train': 2.1678358912467957} 01/29/2022 04:40:07 - INFO - codeparrot_training - Step 12311: {'lr': 0.000445199642561577, 'samples': 2363904, 'steps': 12311, 'loss/train': 1.383168876171112} 01/29/2022 04:40:12 - INFO - codeparrot_training - Step 12312: {'lr': 0.0004451894191582423, 'samples': 2364096, 'steps': 12312, 'loss/train': 5.8623268604278564} 01/29/2022 04:40:16 - INFO - codeparrot_training - Step 12313: {'lr': 0.0004451791949187781, 'samples': 2364288, 'steps': 12313, 'loss/train': 1.5710631608963013} 01/29/2022 04:40:21 - INFO - codeparrot_training - Step 12314: {'lr': 0.0004451689698432282, 'samples': 2364480, 'steps': 12314, 'loss/train': 1.6478254795074463} 01/29/2022 04:40:25 - INFO - codeparrot_training - Step 12315: {'lr': 0.0004451587439316365, 'samples': 2364672, 'steps': 12315, 'loss/train': 1.6062315702438354} 01/29/2022 04:40:29 - INFO - codeparrot_training - Step 12316: {'lr': 0.0004451485171840466, 'samples': 2364864, 'steps': 12316, 'loss/train': 1.9940147995948792} 01/29/2022 04:40:35 - INFO - codeparrot_training - Step 12317: {'lr': 0.0004451382896005024, 'samples': 2365056, 'steps': 12317, 'loss/train': 1.4597174227237701} 01/29/2022 04:40:39 - INFO - codeparrot_training - Step 12318: {'lr': 0.00044512806118104784, 'samples': 2365248, 'steps': 12318, 'loss/train': 0.0784084964543581} 01/29/2022 04:40:43 - INFO - codeparrot_training - Step 12319: {'lr': 0.0004451178319257265, 'samples': 2365440, 'steps': 12319, 'loss/train': 1.6599873304367065} 01/29/2022 04:40:47 - INFO - codeparrot_training - Step 12320: {'lr': 0.0004451076018345824, 'samples': 2365632, 'steps': 12320, 'loss/train': 1.631901741027832} 01/29/2022 04:41:00 - INFO - codeparrot_training - Step 12321: {'lr': 0.00044509737090765933, 'samples': 2365824, 'steps': 12321, 'loss/train': 1.168037474155426} 01/29/2022 04:41:05 - INFO - codeparrot_training - Step 12322: {'lr': 0.00044508713914500107, 'samples': 2366016, 'steps': 12322, 'loss/train': 1.7654802203178406} 01/29/2022 04:41:09 - INFO - codeparrot_training - Step 12323: {'lr': 0.0004450769065466514, 'samples': 2366208, 'steps': 12323, 'loss/train': 1.2140488922595978} 01/29/2022 04:41:13 - INFO - codeparrot_training - Step 12324: {'lr': 0.0004450666731126542, 'samples': 2366400, 'steps': 12324, 'loss/train': 1.8287624716758728} 01/29/2022 04:41:17 - INFO - codeparrot_training - Step 12325: {'lr': 0.0004450564388430533, 'samples': 2366592, 'steps': 12325, 'loss/train': 1.825484275817871} 01/29/2022 04:41:22 - INFO - codeparrot_training - Step 12326: {'lr': 0.0004450462037378926, 'samples': 2366784, 'steps': 12326, 'loss/train': 1.7109630703926086} 01/29/2022 04:41:27 - INFO - codeparrot_training - Step 12327: {'lr': 0.0004450359677972159, 'samples': 2366976, 'steps': 12327, 'loss/train': 1.7192476987838745} 01/29/2022 04:41:31 - INFO - codeparrot_training - Step 12328: {'lr': 0.000445025731021067, 'samples': 2367168, 'steps': 12328, 'loss/train': 2.0363227128982544} 01/29/2022 04:41:36 - INFO - codeparrot_training - Step 12329: {'lr': 0.0004450154934094898, 'samples': 2367360, 'steps': 12329, 'loss/train': 1.948930263519287} 01/29/2022 04:41:40 - INFO - codeparrot_training - Step 12330: {'lr': 0.0004450052549625282, 'samples': 2367552, 'steps': 12330, 'loss/train': 1.0406874418258667} 01/29/2022 04:41:44 - INFO - codeparrot_training - Step 12331: {'lr': 0.000444995015680226, 'samples': 2367744, 'steps': 12331, 'loss/train': 1.878428041934967} 01/29/2022 04:41:49 - INFO - codeparrot_training - Step 12332: {'lr': 0.0004449847755626271, 'samples': 2367936, 'steps': 12332, 'loss/train': 1.8120762705802917} 01/29/2022 04:41:54 - INFO - codeparrot_training - Step 12333: {'lr': 0.00044497453460977523, 'samples': 2368128, 'steps': 12333, 'loss/train': 1.7973387837409973} 01/29/2022 04:41:58 - INFO - codeparrot_training - Step 12334: {'lr': 0.0004449642928217144, 'samples': 2368320, 'steps': 12334, 'loss/train': 1.935584843158722} 01/29/2022 04:42:02 - INFO - codeparrot_training - Step 12335: {'lr': 0.0004449540501984885, 'samples': 2368512, 'steps': 12335, 'loss/train': 0.7615935802459717} 01/29/2022 04:42:06 - INFO - codeparrot_training - Step 12336: {'lr': 0.0004449438067401413, 'samples': 2368704, 'steps': 12336, 'loss/train': 1.1219237744808197} 01/29/2022 04:42:13 - INFO - codeparrot_training - Step 12337: {'lr': 0.0004449335624467168, 'samples': 2368896, 'steps': 12337, 'loss/train': 1.6954039335250854} 01/29/2022 04:42:17 - INFO - codeparrot_training - Step 12338: {'lr': 0.00044492331731825875, 'samples': 2369088, 'steps': 12338, 'loss/train': 1.52471923828125} 01/29/2022 04:42:21 - INFO - codeparrot_training - Step 12339: {'lr': 0.0004449130713548111, 'samples': 2369280, 'steps': 12339, 'loss/train': 1.958014726638794} 01/29/2022 04:42:25 - INFO - codeparrot_training - Step 12340: {'lr': 0.00044490282455641783, 'samples': 2369472, 'steps': 12340, 'loss/train': 1.519099473953247} 01/29/2022 04:42:30 - INFO - codeparrot_training - Step 12341: {'lr': 0.0004448925769231227, 'samples': 2369664, 'steps': 12341, 'loss/train': 1.3988266289234161} 01/29/2022 04:42:35 - INFO - codeparrot_training - Step 12342: {'lr': 0.0004448823284549696, 'samples': 2369856, 'steps': 12342, 'loss/train': 2.126345694065094} 01/29/2022 04:42:39 - INFO - codeparrot_training - Step 12343: {'lr': 0.00044487207915200257, 'samples': 2370048, 'steps': 12343, 'loss/train': 2.03450345993042} 01/29/2022 04:42:44 - INFO - codeparrot_training - Step 12344: {'lr': 0.0004448618290142654, 'samples': 2370240, 'steps': 12344, 'loss/train': 2.6063486337661743} 01/29/2022 04:42:48 - INFO - codeparrot_training - Step 12345: {'lr': 0.000444851578041802, 'samples': 2370432, 'steps': 12345, 'loss/train': 2.083236336708069} 01/29/2022 04:42:52 - INFO - codeparrot_training - Step 12346: {'lr': 0.00044484132623465633, 'samples': 2370624, 'steps': 12346, 'loss/train': 1.8291730284690857} 01/29/2022 04:42:58 - INFO - codeparrot_training - Step 12347: {'lr': 0.0004448310735928723, 'samples': 2370816, 'steps': 12347, 'loss/train': 3.1150084733963013} 01/29/2022 04:43:02 - INFO - codeparrot_training - Step 12348: {'lr': 0.0004448208201164938, 'samples': 2371008, 'steps': 12348, 'loss/train': 1.5284666419029236} 01/29/2022 04:43:07 - INFO - codeparrot_training - Step 12349: {'lr': 0.0004448105658055648, 'samples': 2371200, 'steps': 12349, 'loss/train': 2.2427839636802673} 01/29/2022 04:43:11 - INFO - codeparrot_training - Step 12350: {'lr': 0.00044480031066012916, 'samples': 2371392, 'steps': 12350, 'loss/train': 0.7063575088977814} 01/29/2022 04:43:15 - INFO - codeparrot_training - Step 12351: {'lr': 0.00044479005468023086, 'samples': 2371584, 'steps': 12351, 'loss/train': 2.110847532749176} 01/29/2022 04:43:20 - INFO - codeparrot_training - Step 12352: {'lr': 0.0004447797978659138, 'samples': 2371776, 'steps': 12352, 'loss/train': 1.6702287197113037} 01/29/2022 04:43:25 - INFO - codeparrot_training - Step 12353: {'lr': 0.000444769540217222, 'samples': 2371968, 'steps': 12353, 'loss/train': 1.4511285424232483} 01/29/2022 04:43:29 - INFO - codeparrot_training - Step 12354: {'lr': 0.0004447592817341993, 'samples': 2372160, 'steps': 12354, 'loss/train': 1.849837303161621} 01/29/2022 04:43:33 - INFO - codeparrot_training - Step 12355: {'lr': 0.0004447490224168896, 'samples': 2372352, 'steps': 12355, 'loss/train': 0.7829861640930176} 01/29/2022 04:43:37 - INFO - codeparrot_training - Step 12356: {'lr': 0.00044473876226533703, 'samples': 2372544, 'steps': 12356, 'loss/train': 1.6197974681854248} 01/29/2022 04:43:42 - INFO - codeparrot_training - Step 12357: {'lr': 0.0004447285012795854, 'samples': 2372736, 'steps': 12357, 'loss/train': 1.6591848134994507} 01/29/2022 04:43:47 - INFO - codeparrot_training - Step 12358: {'lr': 0.0004447182394596788, 'samples': 2372928, 'steps': 12358, 'loss/train': 1.8456791639328003} 01/29/2022 04:43:51 - INFO - codeparrot_training - Step 12359: {'lr': 0.000444707976805661, 'samples': 2373120, 'steps': 12359, 'loss/train': 2.6074419021606445} 01/29/2022 04:43:55 - INFO - codeparrot_training - Step 12360: {'lr': 0.00044469771331757604, 'samples': 2373312, 'steps': 12360, 'loss/train': 1.4589949250221252} 01/29/2022 04:44:00 - INFO - codeparrot_training - Step 12361: {'lr': 0.00044468744899546785, 'samples': 2373504, 'steps': 12361, 'loss/train': 1.4685499370098114} 01/29/2022 04:44:05 - INFO - codeparrot_training - Step 12362: {'lr': 0.0004446771838393806, 'samples': 2373696, 'steps': 12362, 'loss/train': 1.6765143871307373} 01/29/2022 04:44:09 - INFO - codeparrot_training - Step 12363: {'lr': 0.00044466691784935796, 'samples': 2373888, 'steps': 12363, 'loss/train': 1.515254259109497} 01/29/2022 04:44:13 - INFO - codeparrot_training - Step 12364: {'lr': 0.00044465665102544415, 'samples': 2374080, 'steps': 12364, 'loss/train': 1.7532999515533447} 01/29/2022 04:44:18 - INFO - codeparrot_training - Step 12365: {'lr': 0.000444646383367683, 'samples': 2374272, 'steps': 12365, 'loss/train': 1.800626814365387} 01/29/2022 04:44:22 - INFO - codeparrot_training - Step 12366: {'lr': 0.00044463611487611864, 'samples': 2374464, 'steps': 12366, 'loss/train': 1.7188661098480225} 01/29/2022 04:44:28 - INFO - codeparrot_training - Step 12367: {'lr': 0.0004446258455507949, 'samples': 2374656, 'steps': 12367, 'loss/train': 1.2535572052001953} 01/29/2022 04:44:32 - INFO - codeparrot_training - Step 12368: {'lr': 0.00044461557539175587, 'samples': 2374848, 'steps': 12368, 'loss/train': 1.7026290893554688} 01/29/2022 04:44:37 - INFO - codeparrot_training - Step 12369: {'lr': 0.0004446053043990455, 'samples': 2375040, 'steps': 12369, 'loss/train': 1.1380680799484253} 01/29/2022 04:44:41 - INFO - codeparrot_training - Step 12370: {'lr': 0.00044459503257270776, 'samples': 2375232, 'steps': 12370, 'loss/train': 2.3138099312782288} 01/29/2022 04:44:45 - INFO - codeparrot_training - Step 12371: {'lr': 0.0004445847599127868, 'samples': 2375424, 'steps': 12371, 'loss/train': 1.786744236946106} 01/29/2022 04:44:50 - INFO - codeparrot_training - Step 12372: {'lr': 0.0004445744864193264, 'samples': 2375616, 'steps': 12372, 'loss/train': 2.0965545773506165} 01/29/2022 04:44:54 - INFO - codeparrot_training - Step 12373: {'lr': 0.00044456421209237073, 'samples': 2375808, 'steps': 12373, 'loss/train': 2.2403817772865295} 01/29/2022 04:44:59 - INFO - codeparrot_training - Step 12374: {'lr': 0.00044455393693196375, 'samples': 2376000, 'steps': 12374, 'loss/train': 1.9109011888504028} 01/29/2022 04:45:03 - INFO - codeparrot_training - Step 12375: {'lr': 0.00044454366093814947, 'samples': 2376192, 'steps': 12375, 'loss/train': 1.0999355614185333} 01/29/2022 04:45:07 - INFO - codeparrot_training - Step 12376: {'lr': 0.0004445333841109719, 'samples': 2376384, 'steps': 12376, 'loss/train': 2.3573641777038574} 01/29/2022 04:45:13 - INFO - codeparrot_training - Step 12377: {'lr': 0.0004445231064504751, 'samples': 2376576, 'steps': 12377, 'loss/train': 1.8146767616271973} 01/29/2022 04:45:17 - INFO - codeparrot_training - Step 12378: {'lr': 0.00044451282795670313, 'samples': 2376768, 'steps': 12378, 'loss/train': 1.385749876499176} 01/29/2022 04:45:21 - INFO - codeparrot_training - Step 12379: {'lr': 0.0004445025486297, 'samples': 2376960, 'steps': 12379, 'loss/train': 2.2892205119132996} 01/29/2022 04:45:26 - INFO - codeparrot_training - Step 12380: {'lr': 0.00044449226846950964, 'samples': 2377152, 'steps': 12380, 'loss/train': 1.835700273513794} 01/29/2022 04:45:30 - INFO - codeparrot_training - Step 12381: {'lr': 0.0004444819874761762, 'samples': 2377344, 'steps': 12381, 'loss/train': 1.1106928288936615} 01/29/2022 04:45:36 - INFO - codeparrot_training - Step 12382: {'lr': 0.0004444717056497436, 'samples': 2377536, 'steps': 12382, 'loss/train': 1.8935669660568237} 01/29/2022 04:45:40 - INFO - codeparrot_training - Step 12383: {'lr': 0.00044446142299025605, 'samples': 2377728, 'steps': 12383, 'loss/train': 0.7887769639492035} 01/29/2022 04:45:44 - INFO - codeparrot_training - Step 12384: {'lr': 0.0004444511394977575, 'samples': 2377920, 'steps': 12384, 'loss/train': 1.6197332739830017} 01/29/2022 04:45:48 - INFO - codeparrot_training - Step 12385: {'lr': 0.0004444408551722919, 'samples': 2378112, 'steps': 12385, 'loss/train': 1.4828008711338043} 01/29/2022 04:45:54 - INFO - codeparrot_training - Step 12386: {'lr': 0.00044443057001390354, 'samples': 2378304, 'steps': 12386, 'loss/train': 3.656924843788147} 01/29/2022 04:45:58 - INFO - codeparrot_training - Step 12387: {'lr': 0.00044442028402263636, 'samples': 2378496, 'steps': 12387, 'loss/train': 1.3863859176635742} 01/29/2022 04:46:02 - INFO - codeparrot_training - Step 12388: {'lr': 0.00044440999719853435, 'samples': 2378688, 'steps': 12388, 'loss/train': 1.5530741214752197} 01/29/2022 04:46:06 - INFO - codeparrot_training - Step 12389: {'lr': 0.0004443997095416417, 'samples': 2378880, 'steps': 12389, 'loss/train': 1.4467660188674927} 01/29/2022 04:46:11 - INFO - codeparrot_training - Step 12390: {'lr': 0.0004443894210520024, 'samples': 2379072, 'steps': 12390, 'loss/train': 1.810744285583496} 01/29/2022 04:46:16 - INFO - codeparrot_training - Step 12391: {'lr': 0.0004443791317296606, 'samples': 2379264, 'steps': 12391, 'loss/train': 1.68643319606781} 01/29/2022 04:46:21 - INFO - codeparrot_training - Step 12392: {'lr': 0.0004443688415746602, 'samples': 2379456, 'steps': 12392, 'loss/train': 1.5045064687728882} 01/29/2022 04:46:25 - INFO - codeparrot_training - Step 12393: {'lr': 0.0004443585505870456, 'samples': 2379648, 'steps': 12393, 'loss/train': 1.2212610840797424} 01/29/2022 04:46:29 - INFO - codeparrot_training - Step 12394: {'lr': 0.0004443482587668605, 'samples': 2379840, 'steps': 12394, 'loss/train': 2.250904619693756} 01/29/2022 04:46:33 - INFO - codeparrot_training - Step 12395: {'lr': 0.00044433796611414924, 'samples': 2380032, 'steps': 12395, 'loss/train': 1.0763529539108276} 01/29/2022 04:46:39 - INFO - codeparrot_training - Step 12396: {'lr': 0.0004443276726289558, 'samples': 2380224, 'steps': 12396, 'loss/train': 1.4984001517295837} 01/29/2022 04:46:43 - INFO - codeparrot_training - Step 12397: {'lr': 0.00044431737831132433, 'samples': 2380416, 'steps': 12397, 'loss/train': 1.486568570137024} 01/29/2022 04:46:47 - INFO - codeparrot_training - Step 12398: {'lr': 0.000444307083161299, 'samples': 2380608, 'steps': 12398, 'loss/train': 1.7619155645370483} 01/29/2022 04:46:52 - INFO - codeparrot_training - Step 12399: {'lr': 0.00044429678717892366, 'samples': 2380800, 'steps': 12399, 'loss/train': 2.4944971203804016} 01/29/2022 04:46:56 - INFO - codeparrot_training - Step 12400: {'lr': 0.0004442864903642427, 'samples': 2380992, 'steps': 12400, 'loss/train': 1.675423800945282} 01/29/2022 04:47:01 - INFO - codeparrot_training - Step 12401: {'lr': 0.00044427619271730014, 'samples': 2381184, 'steps': 12401, 'loss/train': 1.914786458015442} 01/29/2022 04:47:05 - INFO - codeparrot_training - Step 12402: {'lr': 0.00044426589423814003, 'samples': 2381376, 'steps': 12402, 'loss/train': 0.8129977583885193} 01/29/2022 04:47:10 - INFO - codeparrot_training - Step 12403: {'lr': 0.00044425559492680645, 'samples': 2381568, 'steps': 12403, 'loss/train': 1.1120071113109589} 01/29/2022 04:47:14 - INFO - codeparrot_training - Step 12404: {'lr': 0.00044424529478334364, 'samples': 2381760, 'steps': 12404, 'loss/train': 1.7157674431800842} 01/29/2022 04:47:18 - INFO - codeparrot_training - Step 12405: {'lr': 0.00044423499380779566, 'samples': 2381952, 'steps': 12405, 'loss/train': 1.215859204530716} 01/29/2022 04:47:24 - INFO - codeparrot_training - Step 12406: {'lr': 0.00044422469200020666, 'samples': 2382144, 'steps': 12406, 'loss/train': 1.8118404150009155} 01/29/2022 04:47:28 - INFO - codeparrot_training - Step 12407: {'lr': 0.0004442143893606207, 'samples': 2382336, 'steps': 12407, 'loss/train': 1.501489520072937} 01/29/2022 04:47:33 - INFO - codeparrot_training - Step 12408: {'lr': 0.000444204085889082, 'samples': 2382528, 'steps': 12408, 'loss/train': 1.643680214881897} 01/29/2022 04:47:37 - INFO - codeparrot_training - Step 12409: {'lr': 0.00044419378158563465, 'samples': 2382720, 'steps': 12409, 'loss/train': 1.3365052342414856} 01/29/2022 04:47:41 - INFO - codeparrot_training - Step 12410: {'lr': 0.0004441834764503228, 'samples': 2382912, 'steps': 12410, 'loss/train': 0.0793614350259304} 01/29/2022 04:47:47 - INFO - codeparrot_training - Step 12411: {'lr': 0.0004441731704831906, 'samples': 2383104, 'steps': 12411, 'loss/train': 1.5926873087882996} 01/29/2022 04:47:51 - INFO - codeparrot_training - Step 12412: {'lr': 0.0004441628636842822, 'samples': 2383296, 'steps': 12412, 'loss/train': 1.95592600107193} 01/29/2022 04:47:55 - INFO - codeparrot_training - Step 12413: {'lr': 0.0004441525560536418, 'samples': 2383488, 'steps': 12413, 'loss/train': 1.6740800142288208} 01/29/2022 04:47:59 - INFO - codeparrot_training - Step 12414: {'lr': 0.0004441422475913134, 'samples': 2383680, 'steps': 12414, 'loss/train': 1.042345404624939} 01/29/2022 04:48:04 - INFO - codeparrot_training - Step 12415: {'lr': 0.0004441319382973413, 'samples': 2383872, 'steps': 12415, 'loss/train': 0.9186925292015076} 01/29/2022 04:48:09 - INFO - codeparrot_training - Step 12416: {'lr': 0.00044412162817176966, 'samples': 2384064, 'steps': 12416, 'loss/train': 1.334037870168686} 01/29/2022 04:48:14 - INFO - codeparrot_training - Step 12417: {'lr': 0.0004441113172146426, 'samples': 2384256, 'steps': 12417, 'loss/train': 1.3600384891033173} 01/29/2022 04:48:18 - INFO - codeparrot_training - Step 12418: {'lr': 0.00044410100542600423, 'samples': 2384448, 'steps': 12418, 'loss/train': 1.6046208143234253} 01/29/2022 04:48:22 - INFO - codeparrot_training - Step 12419: {'lr': 0.00044409069280589887, 'samples': 2384640, 'steps': 12419, 'loss/train': 2.1483165621757507} 01/29/2022 04:48:26 - INFO - codeparrot_training - Step 12420: {'lr': 0.0004440803793543705, 'samples': 2384832, 'steps': 12420, 'loss/train': 1.1207968890666962} 01/29/2022 04:48:32 - INFO - codeparrot_training - Step 12421: {'lr': 0.00044407006507146354, 'samples': 2385024, 'steps': 12421, 'loss/train': 2.092801094055176} 01/29/2022 04:48:36 - INFO - codeparrot_training - Step 12422: {'lr': 0.000444059749957222, 'samples': 2385216, 'steps': 12422, 'loss/train': 2.10828298330307} 01/29/2022 04:48:40 - INFO - codeparrot_training - Step 12423: {'lr': 0.00044404943401169005, 'samples': 2385408, 'steps': 12423, 'loss/train': 1.2171521186828613} 01/29/2022 04:48:45 - INFO - codeparrot_training - Step 12424: {'lr': 0.00044403911723491196, 'samples': 2385600, 'steps': 12424, 'loss/train': 1.49852254986763} 01/29/2022 04:48:50 - INFO - codeparrot_training - Step 12425: {'lr': 0.000444028799626932, 'samples': 2385792, 'steps': 12425, 'loss/train': 1.93895423412323} 01/29/2022 04:48:54 - INFO - codeparrot_training - Step 12426: {'lr': 0.0004440184811877942, 'samples': 2385984, 'steps': 12426, 'loss/train': 1.6675336360931396} 01/29/2022 04:48:59 - INFO - codeparrot_training - Step 12427: {'lr': 0.0004440081619175428, 'samples': 2386176, 'steps': 12427, 'loss/train': 0.04093051515519619} 01/29/2022 04:49:03 - INFO - codeparrot_training - Step 12428: {'lr': 0.00044399784181622216, 'samples': 2386368, 'steps': 12428, 'loss/train': 1.4791464507579803} 01/29/2022 04:49:07 - INFO - codeparrot_training - Step 12429: {'lr': 0.0004439875208838763, 'samples': 2386560, 'steps': 12429, 'loss/train': 1.8929398655891418} 01/29/2022 04:49:12 - INFO - codeparrot_training - Step 12430: {'lr': 0.00044397719912054944, 'samples': 2386752, 'steps': 12430, 'loss/train': 1.4071728587150574} 01/29/2022 04:49:17 - INFO - codeparrot_training - Step 12431: {'lr': 0.00044396687652628586, 'samples': 2386944, 'steps': 12431, 'loss/train': 1.2378130853176117} 01/29/2022 04:49:21 - INFO - codeparrot_training - Step 12432: {'lr': 0.00044395655310112985, 'samples': 2387136, 'steps': 12432, 'loss/train': 1.203721135854721} 01/29/2022 04:49:25 - INFO - codeparrot_training - Step 12433: {'lr': 0.00044394622884512554, 'samples': 2387328, 'steps': 12433, 'loss/train': 1.708770990371704} 01/29/2022 04:49:29 - INFO - codeparrot_training - Step 12434: {'lr': 0.00044393590375831716, 'samples': 2387520, 'steps': 12434, 'loss/train': 1.676420509815216} 01/29/2022 04:49:35 - INFO - codeparrot_training - Step 12435: {'lr': 0.00044392557784074895, 'samples': 2387712, 'steps': 12435, 'loss/train': 2.215408980846405} 01/29/2022 04:49:40 - INFO - codeparrot_training - Step 12436: {'lr': 0.0004439152510924651, 'samples': 2387904, 'steps': 12436, 'loss/train': 1.6420916318893433} 01/29/2022 04:49:44 - INFO - codeparrot_training - Step 12437: {'lr': 0.0004439049235135099, 'samples': 2388096, 'steps': 12437, 'loss/train': 1.2179828882217407} 01/29/2022 04:49:48 - INFO - codeparrot_training - Step 12438: {'lr': 0.0004438945951039276, 'samples': 2388288, 'steps': 12438, 'loss/train': 2.219411551952362} 01/29/2022 04:49:52 - INFO - codeparrot_training - Step 12439: {'lr': 0.0004438842658637624, 'samples': 2388480, 'steps': 12439, 'loss/train': 2.3246853947639465} 01/29/2022 04:49:58 - INFO - codeparrot_training - Step 12440: {'lr': 0.0004438739357930586, 'samples': 2388672, 'steps': 12440, 'loss/train': 1.2254855632781982} 01/29/2022 04:50:02 - INFO - codeparrot_training - Step 12441: {'lr': 0.00044386360489186047, 'samples': 2388864, 'steps': 12441, 'loss/train': 1.7067355513572693} 01/29/2022 04:50:06 - INFO - codeparrot_training - Step 12442: {'lr': 0.00044385327316021214, 'samples': 2389056, 'steps': 12442, 'loss/train': 1.8688788414001465} 01/29/2022 04:50:10 - INFO - codeparrot_training - Step 12443: {'lr': 0.000443842940598158, 'samples': 2389248, 'steps': 12443, 'loss/train': 1.3827377557754517} 01/29/2022 04:50:15 - INFO - codeparrot_training - Step 12444: {'lr': 0.00044383260720574214, 'samples': 2389440, 'steps': 12444, 'loss/train': 1.465412199497223} 01/29/2022 04:50:20 - INFO - codeparrot_training - Step 12445: {'lr': 0.00044382227298300905, 'samples': 2389632, 'steps': 12445, 'loss/train': 1.6308832168579102} 01/29/2022 04:50:24 - INFO - codeparrot_training - Step 12446: {'lr': 0.0004438119379300028, 'samples': 2389824, 'steps': 12446, 'loss/train': 2.191139817237854} 01/29/2022 04:50:28 - INFO - codeparrot_training - Step 12447: {'lr': 0.00044380160204676787, 'samples': 2390016, 'steps': 12447, 'loss/train': 2.816121518611908} 01/29/2022 04:50:33 - INFO - codeparrot_training - Step 12448: {'lr': 0.00044379126533334836, 'samples': 2390208, 'steps': 12448, 'loss/train': 1.4625333845615387} 01/29/2022 04:50:37 - INFO - codeparrot_training - Step 12449: {'lr': 0.00044378092778978864, 'samples': 2390400, 'steps': 12449, 'loss/train': 1.5738442540168762} 01/29/2022 04:50:43 - INFO - codeparrot_training - Step 12450: {'lr': 0.00044377058941613283, 'samples': 2390592, 'steps': 12450, 'loss/train': 1.5647111535072327} 01/29/2022 04:50:47 - INFO - codeparrot_training - Step 12451: {'lr': 0.0004437602502124255, 'samples': 2390784, 'steps': 12451, 'loss/train': 1.1362501680850983} 01/29/2022 04:50:51 - INFO - codeparrot_training - Step 12452: {'lr': 0.0004437499101787107, 'samples': 2390976, 'steps': 12452, 'loss/train': 1.252355307340622} 01/29/2022 04:50:56 - INFO - codeparrot_training - Step 12453: {'lr': 0.0004437395693150328, 'samples': 2391168, 'steps': 12453, 'loss/train': 1.8990565538406372} 01/29/2022 04:51:00 - INFO - codeparrot_training - Step 12454: {'lr': 0.0004437292276214361, 'samples': 2391360, 'steps': 12454, 'loss/train': 2.9771291613578796} 01/29/2022 04:51:05 - INFO - codeparrot_training - Step 12455: {'lr': 0.000443718885097965, 'samples': 2391552, 'steps': 12455, 'loss/train': 2.0056525468826294} 01/29/2022 04:51:09 - INFO - codeparrot_training - Step 12456: {'lr': 0.0004437085417446636, 'samples': 2391744, 'steps': 12456, 'loss/train': 1.5319241881370544} 01/29/2022 04:51:14 - INFO - codeparrot_training - Step 12457: {'lr': 0.0004436981975615764, 'samples': 2391936, 'steps': 12457, 'loss/train': 1.3117224276065826} 01/29/2022 04:51:18 - INFO - codeparrot_training - Step 12458: {'lr': 0.00044368785254874754, 'samples': 2392128, 'steps': 12458, 'loss/train': 1.8933674097061157} 01/29/2022 04:51:22 - INFO - codeparrot_training - Step 12459: {'lr': 0.00044367750670622143, 'samples': 2392320, 'steps': 12459, 'loss/train': 1.899684727191925} 01/29/2022 04:51:28 - INFO - codeparrot_training - Step 12460: {'lr': 0.0004436671600340424, 'samples': 2392512, 'steps': 12460, 'loss/train': 0.9585011601448059} 01/29/2022 04:51:32 - INFO - codeparrot_training - Step 12461: {'lr': 0.00044365681253225476, 'samples': 2392704, 'steps': 12461, 'loss/train': 1.6172906756401062} 01/29/2022 04:51:36 - INFO - codeparrot_training - Step 12462: {'lr': 0.0004436464642009029, 'samples': 2392896, 'steps': 12462, 'loss/train': 1.5675907731056213} 01/29/2022 04:51:41 - INFO - codeparrot_training - Step 12463: {'lr': 0.00044363611504003096, 'samples': 2393088, 'steps': 12463, 'loss/train': 1.1826061606407166} 01/29/2022 04:51:45 - INFO - codeparrot_training - Step 12464: {'lr': 0.00044362576504968344, 'samples': 2393280, 'steps': 12464, 'loss/train': 1.1823298931121826} 01/29/2022 04:51:50 - INFO - codeparrot_training - Step 12465: {'lr': 0.0004436154142299046, 'samples': 2393472, 'steps': 12465, 'loss/train': 0.9710347652435303} 01/29/2022 04:51:55 - INFO - codeparrot_training - Step 12466: {'lr': 0.00044360506258073884, 'samples': 2393664, 'steps': 12466, 'loss/train': 1.6779304146766663} 01/29/2022 04:51:59 - INFO - codeparrot_training - Step 12467: {'lr': 0.0004435947101022305, 'samples': 2393856, 'steps': 12467, 'loss/train': 1.6593355536460876} 01/29/2022 04:52:03 - INFO - codeparrot_training - Step 12468: {'lr': 0.0004435843567944239, 'samples': 2394048, 'steps': 12468, 'loss/train': 1.4363941848278046} 01/29/2022 04:52:07 - INFO - codeparrot_training - Step 12469: {'lr': 0.0004435740026573633, 'samples': 2394240, 'steps': 12469, 'loss/train': 3.324295163154602} 01/29/2022 04:52:13 - INFO - codeparrot_training - Step 12470: {'lr': 0.0004435636476910932, 'samples': 2394432, 'steps': 12470, 'loss/train': 1.6372708082199097} 01/29/2022 04:52:17 - INFO - codeparrot_training - Step 12471: {'lr': 0.00044355329189565783, 'samples': 2394624, 'steps': 12471, 'loss/train': 1.7748363018035889} 01/29/2022 04:52:22 - INFO - codeparrot_training - Step 12472: {'lr': 0.00044354293527110167, 'samples': 2394816, 'steps': 12472, 'loss/train': 1.9122726917266846} 01/29/2022 04:52:26 - INFO - codeparrot_training - Step 12473: {'lr': 0.000443532577817469, 'samples': 2395008, 'steps': 12473, 'loss/train': 2.1901477575302124} 01/29/2022 04:52:30 - INFO - codeparrot_training - Step 12474: {'lr': 0.0004435222195348043, 'samples': 2395200, 'steps': 12474, 'loss/train': 1.7917181253433228} 01/29/2022 04:52:36 - INFO - codeparrot_training - Step 12475: {'lr': 0.00044351186042315184, 'samples': 2395392, 'steps': 12475, 'loss/train': 1.7025774121284485} 01/29/2022 04:52:41 - INFO - codeparrot_training - Step 12476: {'lr': 0.000443501500482556, 'samples': 2395584, 'steps': 12476, 'loss/train': 0.6423642486333847} 01/29/2022 04:52:45 - INFO - codeparrot_training - Step 12477: {'lr': 0.0004434911397130612, 'samples': 2395776, 'steps': 12477, 'loss/train': 0.7414898872375488} 01/29/2022 04:52:49 - INFO - codeparrot_training - Step 12478: {'lr': 0.0004434807781147117, 'samples': 2395968, 'steps': 12478, 'loss/train': 1.8815365433692932} 01/29/2022 04:52:53 - INFO - codeparrot_training - Step 12479: {'lr': 0.0004434704156875521, 'samples': 2396160, 'steps': 12479, 'loss/train': 1.274631679058075} 01/29/2022 04:52:58 - INFO - codeparrot_training - Step 12480: {'lr': 0.00044346005243162654, 'samples': 2396352, 'steps': 12480, 'loss/train': 1.8607168793678284} 01/29/2022 04:53:03 - INFO - codeparrot_training - Step 12481: {'lr': 0.0004434496883469796, 'samples': 2396544, 'steps': 12481, 'loss/train': 2.022799551486969} 01/29/2022 04:53:07 - INFO - codeparrot_training - Step 12482: {'lr': 0.0004434393234336557, 'samples': 2396736, 'steps': 12482, 'loss/train': 1.203377902507782} 01/29/2022 04:53:11 - INFO - codeparrot_training - Step 12483: {'lr': 0.0004434289576916991, 'samples': 2396928, 'steps': 12483, 'loss/train': 1.0936646461486816} 01/29/2022 04:53:16 - INFO - codeparrot_training - Step 12484: {'lr': 0.00044341859112115425, 'samples': 2397120, 'steps': 12484, 'loss/train': 1.3204021453857422} 01/29/2022 04:53:21 - INFO - codeparrot_training - Step 12485: {'lr': 0.00044340822372206557, 'samples': 2397312, 'steps': 12485, 'loss/train': 1.8226141333580017} 01/29/2022 04:53:25 - INFO - codeparrot_training - Step 12486: {'lr': 0.00044339785549447756, 'samples': 2397504, 'steps': 12486, 'loss/train': 1.1794877350330353} 01/29/2022 04:53:29 - INFO - codeparrot_training - Step 12487: {'lr': 0.00044338748643843446, 'samples': 2397696, 'steps': 12487, 'loss/train': 1.68450129032135} 01/29/2022 04:53:33 - INFO - codeparrot_training - Step 12488: {'lr': 0.00044337711655398083, 'samples': 2397888, 'steps': 12488, 'loss/train': 1.3526991605758667} 01/29/2022 04:53:39 - INFO - codeparrot_training - Step 12489: {'lr': 0.00044336674584116096, 'samples': 2398080, 'steps': 12489, 'loss/train': 1.5408759713172913} 01/29/2022 04:53:43 - INFO - codeparrot_training - Step 12490: {'lr': 0.0004433563743000195, 'samples': 2398272, 'steps': 12490, 'loss/train': 2.168072283267975} 01/29/2022 04:53:47 - INFO - codeparrot_training - Step 12491: {'lr': 0.0004433460019306006, 'samples': 2398464, 'steps': 12491, 'loss/train': 1.1509525179862976} 01/29/2022 04:53:51 - INFO - codeparrot_training - Step 12492: {'lr': 0.00044333562873294884, 'samples': 2398656, 'steps': 12492, 'loss/train': 1.7387203574180603} 01/29/2022 04:53:56 - INFO - codeparrot_training - Step 12493: {'lr': 0.00044332525470710865, 'samples': 2398848, 'steps': 12493, 'loss/train': 1.8986307978630066} 01/29/2022 04:54:02 - INFO - codeparrot_training - Step 12494: {'lr': 0.0004433148798531245, 'samples': 2399040, 'steps': 12494, 'loss/train': 1.1339418590068817} 01/29/2022 04:54:06 - INFO - codeparrot_training - Step 12495: {'lr': 0.0004433045041710407, 'samples': 2399232, 'steps': 12495, 'loss/train': 0.9315836727619171} 01/29/2022 04:54:10 - INFO - codeparrot_training - Step 12496: {'lr': 0.0004432941276609018, 'samples': 2399424, 'steps': 12496, 'loss/train': 1.49957737326622} 01/29/2022 04:54:14 - INFO - codeparrot_training - Step 12497: {'lr': 0.00044328375032275227, 'samples': 2399616, 'steps': 12497, 'loss/train': 0.5052544176578522} 01/29/2022 04:54:19 - INFO - codeparrot_training - Step 12498: {'lr': 0.00044327337215663656, 'samples': 2399808, 'steps': 12498, 'loss/train': 1.0551702082157135} 01/29/2022 04:54:24 - INFO - codeparrot_training - Step 12499: {'lr': 0.000443262993162599, 'samples': 2400000, 'steps': 12499, 'loss/train': 1.3977015316486359} 01/29/2022 04:54:28 - INFO - codeparrot_training - Step 12500: {'lr': 0.0004432526133406842, 'samples': 2400192, 'steps': 12500, 'loss/train': 1.6387977004051208} 01/29/2022 04:54:32 - INFO - codeparrot_training - Step 12501: {'lr': 0.00044324223269093666, 'samples': 2400384, 'steps': 12501, 'loss/train': 1.5692986249923706} 01/29/2022 04:54:37 - INFO - codeparrot_training - Step 12502: {'lr': 0.00044323185121340064, 'samples': 2400576, 'steps': 12502, 'loss/train': 1.9912222623825073} 01/29/2022 04:54:41 - INFO - codeparrot_training - Step 12503: {'lr': 0.00044322146890812076, 'samples': 2400768, 'steps': 12503, 'loss/train': 3.5526126623153687} 01/29/2022 04:54:47 - INFO - codeparrot_training - Step 12504: {'lr': 0.0004432110857751415, 'samples': 2400960, 'steps': 12504, 'loss/train': 1.4400695264339447} 01/29/2022 04:54:51 - INFO - codeparrot_training - Step 12505: {'lr': 0.0004432007018145072, 'samples': 2401152, 'steps': 12505, 'loss/train': 2.7553396224975586} 01/29/2022 04:54:55 - INFO - codeparrot_training - Step 12506: {'lr': 0.00044319031702626255, 'samples': 2401344, 'steps': 12506, 'loss/train': 1.240113615989685} 01/29/2022 04:55:00 - INFO - codeparrot_training - Step 12507: {'lr': 0.0004431799314104519, 'samples': 2401536, 'steps': 12507, 'loss/train': 1.9809136390686035} 01/29/2022 04:55:04 - INFO - codeparrot_training - Step 12508: {'lr': 0.0004431695449671197, 'samples': 2401728, 'steps': 12508, 'loss/train': 2.158774495124817} 01/29/2022 04:55:09 - INFO - codeparrot_training - Step 12509: {'lr': 0.00044315915769631054, 'samples': 2401920, 'steps': 12509, 'loss/train': 2.318394184112549} 01/29/2022 04:55:14 - INFO - codeparrot_training - Step 12510: {'lr': 0.0004431487695980689, 'samples': 2402112, 'steps': 12510, 'loss/train': 1.73695707321167} 01/29/2022 04:55:18 - INFO - codeparrot_training - Step 12511: {'lr': 0.0004431383806724393, 'samples': 2402304, 'steps': 12511, 'loss/train': 1.672675609588623} 01/29/2022 04:55:22 - INFO - codeparrot_training - Step 12512: {'lr': 0.0004431279909194661, 'samples': 2402496, 'steps': 12512, 'loss/train': 1.4075648188591003} 01/29/2022 04:55:26 - INFO - codeparrot_training - Step 12513: {'lr': 0.000443117600339194, 'samples': 2402688, 'steps': 12513, 'loss/train': 2.351353883743286} 01/29/2022 04:55:32 - INFO - codeparrot_training - Step 12514: {'lr': 0.0004431072089316674, 'samples': 2402880, 'steps': 12514, 'loss/train': 1.6808767318725586} 01/29/2022 04:55:36 - INFO - codeparrot_training - Step 12515: {'lr': 0.0004430968166969308, 'samples': 2403072, 'steps': 12515, 'loss/train': 1.3808543980121613} 01/29/2022 04:55:41 - INFO - codeparrot_training - Step 12516: {'lr': 0.00044308642363502884, 'samples': 2403264, 'steps': 12516, 'loss/train': 1.8384969234466553} 01/29/2022 04:55:45 - INFO - codeparrot_training - Step 12517: {'lr': 0.00044307602974600594, 'samples': 2403456, 'steps': 12517, 'loss/train': 2.509050965309143} 01/29/2022 04:55:49 - INFO - codeparrot_training - Step 12518: {'lr': 0.00044306563502990656, 'samples': 2403648, 'steps': 12518, 'loss/train': 1.7296092510223389} 01/29/2022 04:55:55 - INFO - codeparrot_training - Step 12519: {'lr': 0.0004430552394867753, 'samples': 2403840, 'steps': 12519, 'loss/train': 2.7202067971229553} 01/29/2022 04:55:59 - INFO - codeparrot_training - Step 12520: {'lr': 0.0004430448431166567, 'samples': 2404032, 'steps': 12520, 'loss/train': 1.985876441001892} 01/29/2022 04:56:03 - INFO - codeparrot_training - Step 12521: {'lr': 0.00044303444591959533, 'samples': 2404224, 'steps': 12521, 'loss/train': 0.9865850508213043} 01/29/2022 04:56:08 - INFO - codeparrot_training - Step 12522: {'lr': 0.00044302404789563573, 'samples': 2404416, 'steps': 12522, 'loss/train': 1.6370569467544556} 01/29/2022 04:56:12 - INFO - codeparrot_training - Step 12523: {'lr': 0.0004430136490448223, 'samples': 2404608, 'steps': 12523, 'loss/train': 0.5498399287462234} 01/29/2022 04:56:18 - INFO - codeparrot_training - Step 12524: {'lr': 0.0004430032493671998, 'samples': 2404800, 'steps': 12524, 'loss/train': 1.9336915612220764} 01/29/2022 04:56:22 - INFO - codeparrot_training - Step 12525: {'lr': 0.0004429928488628126, 'samples': 2404992, 'steps': 12525, 'loss/train': 1.4406372606754303} 01/29/2022 04:56:26 - INFO - codeparrot_training - Step 12526: {'lr': 0.00044298244753170535, 'samples': 2405184, 'steps': 12526, 'loss/train': 2.610859751701355} 01/29/2022 04:56:30 - INFO - codeparrot_training - Step 12527: {'lr': 0.00044297204537392253, 'samples': 2405376, 'steps': 12527, 'loss/train': 1.6525080800056458} 01/29/2022 04:56:36 - INFO - codeparrot_training - Step 12528: {'lr': 0.00044296164238950874, 'samples': 2405568, 'steps': 12528, 'loss/train': 1.884831726551056} 01/29/2022 04:56:40 - INFO - codeparrot_training - Step 12529: {'lr': 0.0004429512385785086, 'samples': 2405760, 'steps': 12529, 'loss/train': 2.238109588623047} 01/29/2022 04:56:44 - INFO - codeparrot_training - Step 12530: {'lr': 0.0004429408339409666, 'samples': 2405952, 'steps': 12530, 'loss/train': 1.815506637096405} 01/29/2022 04:56:49 - INFO - codeparrot_training - Step 12531: {'lr': 0.00044293042847692735, 'samples': 2406144, 'steps': 12531, 'loss/train': 7.4273364543914795} 01/29/2022 04:56:53 - INFO - codeparrot_training - Step 12532: {'lr': 0.00044292002218643533, 'samples': 2406336, 'steps': 12532, 'loss/train': 2.68412321805954} 01/29/2022 04:56:59 - INFO - codeparrot_training - Step 12533: {'lr': 0.00044290961506953525, 'samples': 2406528, 'steps': 12533, 'loss/train': 1.5378227233886719} 01/29/2022 04:57:03 - INFO - codeparrot_training - Step 12534: {'lr': 0.0004428992071262716, 'samples': 2406720, 'steps': 12534, 'loss/train': 0.8159152865409851} 01/29/2022 04:57:07 - INFO - codeparrot_training - Step 12535: {'lr': 0.00044288879835668903, 'samples': 2406912, 'steps': 12535, 'loss/train': 2.440918207168579} 01/29/2022 04:57:12 - INFO - codeparrot_training - Step 12536: {'lr': 0.0004428783887608321, 'samples': 2407104, 'steps': 12536, 'loss/train': 0.6433504521846771} 01/29/2022 04:57:16 - INFO - codeparrot_training - Step 12537: {'lr': 0.0004428679783387454, 'samples': 2407296, 'steps': 12537, 'loss/train': 1.88468599319458} 01/29/2022 04:57:21 - INFO - codeparrot_training - Step 12538: {'lr': 0.00044285756709047354, 'samples': 2407488, 'steps': 12538, 'loss/train': 1.5922227501869202} 01/29/2022 04:57:26 - INFO - codeparrot_training - Step 12539: {'lr': 0.0004428471550160611, 'samples': 2407680, 'steps': 12539, 'loss/train': 2.205190122127533} 01/29/2022 04:57:30 - INFO - codeparrot_training - Step 12540: {'lr': 0.00044283674211555266, 'samples': 2407872, 'steps': 12540, 'loss/train': 1.3236960768699646} 01/29/2022 04:57:34 - INFO - codeparrot_training - Step 12541: {'lr': 0.0004428263283889928, 'samples': 2408064, 'steps': 12541, 'loss/train': 2.3902904391288757} 01/29/2022 04:57:38 - INFO - codeparrot_training - Step 12542: {'lr': 0.0004428159138364263, 'samples': 2408256, 'steps': 12542, 'loss/train': 1.1542347371578217} 01/29/2022 04:57:43 - INFO - codeparrot_training - Step 12543: {'lr': 0.0004428054984578975, 'samples': 2408448, 'steps': 12543, 'loss/train': 1.4450943768024445} 01/29/2022 04:57:49 - INFO - codeparrot_training - Step 12544: {'lr': 0.0004427950822534513, 'samples': 2408640, 'steps': 12544, 'loss/train': 2.262570798397064} 01/29/2022 04:57:53 - INFO - codeparrot_training - Step 12545: {'lr': 0.0004427846652231321, 'samples': 2408832, 'steps': 12545, 'loss/train': 1.376072108745575} 01/29/2022 04:57:57 - INFO - codeparrot_training - Step 12546: {'lr': 0.0004427742473669847, 'samples': 2409024, 'steps': 12546, 'loss/train': 2.0690767765045166} 01/29/2022 04:58:01 - INFO - codeparrot_training - Step 12547: {'lr': 0.00044276382868505356, 'samples': 2409216, 'steps': 12547, 'loss/train': 0.8263890445232391} 01/29/2022 04:58:06 - INFO - codeparrot_training - Step 12548: {'lr': 0.0004427534091773834, 'samples': 2409408, 'steps': 12548, 'loss/train': 1.8779590129852295} 01/29/2022 04:58:12 - INFO - codeparrot_training - Step 12549: {'lr': 0.00044274298884401886, 'samples': 2409600, 'steps': 12549, 'loss/train': 1.1084376275539398} 01/29/2022 04:58:16 - INFO - codeparrot_training - Step 12550: {'lr': 0.0004427325676850045, 'samples': 2409792, 'steps': 12550, 'loss/train': 0.5546620041131973} 01/29/2022 04:58:20 - INFO - codeparrot_training - Step 12551: {'lr': 0.00044272214570038513, 'samples': 2409984, 'steps': 12551, 'loss/train': 2.3378909826278687} 01/29/2022 04:58:24 - INFO - codeparrot_training - Step 12552: {'lr': 0.00044271172289020525, 'samples': 2410176, 'steps': 12552, 'loss/train': 1.5767224431037903} 01/29/2022 04:58:29 - INFO - codeparrot_training - Step 12553: {'lr': 0.00044270129925450945, 'samples': 2410368, 'steps': 12553, 'loss/train': 1.7132842540740967} 01/29/2022 04:58:34 - INFO - codeparrot_training - Step 12554: {'lr': 0.00044269087479334256, 'samples': 2410560, 'steps': 12554, 'loss/train': 1.4880360960960388} 01/29/2022 04:58:38 - INFO - codeparrot_training - Step 12555: {'lr': 0.00044268044950674913, 'samples': 2410752, 'steps': 12555, 'loss/train': 1.5689508318901062} 01/29/2022 04:58:42 - INFO - codeparrot_training - Step 12556: {'lr': 0.0004426700233947738, 'samples': 2410944, 'steps': 12556, 'loss/train': 1.9208090901374817} 01/29/2022 04:58:47 - INFO - codeparrot_training - Step 12557: {'lr': 0.00044265959645746136, 'samples': 2411136, 'steps': 12557, 'loss/train': 1.9780515432357788} 01/29/2022 04:58:51 - INFO - codeparrot_training - Step 12558: {'lr': 0.0004426491686948563, 'samples': 2411328, 'steps': 12558, 'loss/train': 1.1309807896614075} 01/29/2022 04:58:56 - INFO - codeparrot_training - Step 12559: {'lr': 0.00044263874010700343, 'samples': 2411520, 'steps': 12559, 'loss/train': 2.46095073223114} 01/29/2022 04:59:00 - INFO - codeparrot_training - Step 12560: {'lr': 0.0004426283106939473, 'samples': 2411712, 'steps': 12560, 'loss/train': 2.131766974925995} 01/29/2022 04:59:05 - INFO - codeparrot_training - Step 12561: {'lr': 0.0004426178804557327, 'samples': 2411904, 'steps': 12561, 'loss/train': 1.8924177289009094} 01/29/2022 04:59:09 - INFO - codeparrot_training - Step 12562: {'lr': 0.0004426074493924043, 'samples': 2412096, 'steps': 12562, 'loss/train': 1.7619304060935974} 01/29/2022 04:59:13 - INFO - codeparrot_training - Step 12563: {'lr': 0.00044259701750400674, 'samples': 2412288, 'steps': 12563, 'loss/train': 1.4763538241386414} 01/29/2022 04:59:19 - INFO - codeparrot_training - Step 12564: {'lr': 0.00044258658479058463, 'samples': 2412480, 'steps': 12564, 'loss/train': 2.0075506567955017} 01/29/2022 04:59:23 - INFO - codeparrot_training - Step 12565: {'lr': 0.00044257615125218273, 'samples': 2412672, 'steps': 12565, 'loss/train': 2.656714081764221} 01/29/2022 04:59:28 - INFO - codeparrot_training - Step 12566: {'lr': 0.00044256571688884583, 'samples': 2412864, 'steps': 12566, 'loss/train': 1.3389569520950317} 01/29/2022 04:59:32 - INFO - codeparrot_training - Step 12567: {'lr': 0.00044255528170061853, 'samples': 2413056, 'steps': 12567, 'loss/train': 1.6102285981178284} 01/29/2022 04:59:36 - INFO - codeparrot_training - Step 12568: {'lr': 0.00044254484568754556, 'samples': 2413248, 'steps': 12568, 'loss/train': 2.9007174968719482} 01/29/2022 04:59:41 - INFO - codeparrot_training - Step 12569: {'lr': 0.0004425344088496716, 'samples': 2413440, 'steps': 12569, 'loss/train': 1.9966063499450684} 01/29/2022 04:59:45 - INFO - codeparrot_training - Step 12570: {'lr': 0.00044252397118704133, 'samples': 2413632, 'steps': 12570, 'loss/train': 1.1435751914978027} 01/29/2022 04:59:50 - INFO - codeparrot_training - Step 12571: {'lr': 0.0004425135326996995, 'samples': 2413824, 'steps': 12571, 'loss/train': 1.664912760257721} 01/29/2022 04:59:54 - INFO - codeparrot_training - Step 12572: {'lr': 0.0004425030933876909, 'samples': 2414016, 'steps': 12572, 'loss/train': 1.9139962792396545} 01/29/2022 04:59:58 - INFO - codeparrot_training - Step 12573: {'lr': 0.00044249265325106013, 'samples': 2414208, 'steps': 12573, 'loss/train': 2.163596034049988} 01/29/2022 05:00:04 - INFO - codeparrot_training - Step 12574: {'lr': 0.000442482212289852, 'samples': 2414400, 'steps': 12574, 'loss/train': 2.158854067325592} 01/29/2022 05:00:08 - INFO - codeparrot_training - Step 12575: {'lr': 0.00044247177050411114, 'samples': 2414592, 'steps': 12575, 'loss/train': 2.1053486466407776} 01/29/2022 05:00:12 - INFO - codeparrot_training - Step 12576: {'lr': 0.00044246132789388235, 'samples': 2414784, 'steps': 12576, 'loss/train': 2.029995918273926} 01/29/2022 05:00:17 - INFO - codeparrot_training - Step 12577: {'lr': 0.00044245088445921035, 'samples': 2414976, 'steps': 12577, 'loss/train': 0.7215382754802704} 01/29/2022 05:00:22 - INFO - codeparrot_training - Step 12578: {'lr': 0.00044244044020013985, 'samples': 2415168, 'steps': 12578, 'loss/train': 1.1869924664497375} 01/29/2022 05:00:27 - INFO - codeparrot_training - Step 12579: {'lr': 0.0004424299951167156, 'samples': 2415360, 'steps': 12579, 'loss/train': 1.8837944269180298} 01/29/2022 05:00:31 - INFO - codeparrot_training - Step 12580: {'lr': 0.0004424195492089824, 'samples': 2415552, 'steps': 12580, 'loss/train': 2.1669201850891113} 01/29/2022 05:00:35 - INFO - codeparrot_training - Step 12581: {'lr': 0.0004424091024769849, 'samples': 2415744, 'steps': 12581, 'loss/train': 2.056535303592682} 01/29/2022 05:00:39 - INFO - codeparrot_training - Step 12582: {'lr': 0.00044239865492076794, 'samples': 2415936, 'steps': 12582, 'loss/train': 1.4175308346748352} 01/29/2022 05:00:45 - INFO - codeparrot_training - Step 12583: {'lr': 0.0004423882065403762, 'samples': 2416128, 'steps': 12583, 'loss/train': 1.1001624763011932} 01/29/2022 05:00:49 - INFO - codeparrot_training - Step 12584: {'lr': 0.0004423777573358545, 'samples': 2416320, 'steps': 12584, 'loss/train': 1.9811350107192993} 01/29/2022 05:00:53 - INFO - codeparrot_training - Step 12585: {'lr': 0.0004423673073072476, 'samples': 2416512, 'steps': 12585, 'loss/train': 1.9940840005874634} 01/29/2022 05:00:57 - INFO - codeparrot_training - Step 12586: {'lr': 0.0004423568564546002, 'samples': 2416704, 'steps': 12586, 'loss/train': 1.0407328605651855} 01/29/2022 05:01:02 - INFO - codeparrot_training - Step 12587: {'lr': 0.00044234640477795707, 'samples': 2416896, 'steps': 12587, 'loss/train': 0.061590252444148064} 01/29/2022 05:01:07 - INFO - codeparrot_training - Step 12588: {'lr': 0.0004423359522773631, 'samples': 2417088, 'steps': 12588, 'loss/train': 1.6563776135444641} 01/29/2022 05:01:11 - INFO - codeparrot_training - Step 12589: {'lr': 0.00044232549895286294, 'samples': 2417280, 'steps': 12589, 'loss/train': 1.1891431510448456} 01/29/2022 05:01:15 - INFO - codeparrot_training - Step 12590: {'lr': 0.00044231504480450145, 'samples': 2417472, 'steps': 12590, 'loss/train': 1.1960814893245697} 01/29/2022 05:01:20 - INFO - codeparrot_training - Step 12591: {'lr': 0.0004423045898323233, 'samples': 2417664, 'steps': 12591, 'loss/train': 1.9993488192558289} 01/29/2022 05:01:24 - INFO - codeparrot_training - Step 12592: {'lr': 0.0004422941340363734, 'samples': 2417856, 'steps': 12592, 'loss/train': 1.6519880890846252} 01/29/2022 05:01:32 - INFO - codeparrot_training - Step 12593: {'lr': 0.0004422836774166965, 'samples': 2418048, 'steps': 12593, 'loss/train': 2.2167270183563232} 01/29/2022 05:01:36 - INFO - codeparrot_training - Step 12594: {'lr': 0.00044227321997333737, 'samples': 2418240, 'steps': 12594, 'loss/train': 1.546593725681305} 01/29/2022 05:01:41 - INFO - codeparrot_training - Step 12595: {'lr': 0.0004422627617063408, 'samples': 2418432, 'steps': 12595, 'loss/train': 1.6839682459831238} 01/29/2022 05:01:45 - INFO - codeparrot_training - Step 12596: {'lr': 0.00044225230261575165, 'samples': 2418624, 'steps': 12596, 'loss/train': 1.6135076880455017} 01/29/2022 05:01:49 - INFO - codeparrot_training - Step 12597: {'lr': 0.00044224184270161466, 'samples': 2418816, 'steps': 12597, 'loss/train': 1.415757179260254} 01/29/2022 05:01:54 - INFO - codeparrot_training - Step 12598: {'lr': 0.0004422313819639747, 'samples': 2419008, 'steps': 12598, 'loss/train': 1.7575036883354187} 01/29/2022 05:01:59 - INFO - codeparrot_training - Step 12599: {'lr': 0.0004422209204028765, 'samples': 2419200, 'steps': 12599, 'loss/train': 1.613481044769287} 01/29/2022 05:02:03 - INFO - codeparrot_training - Step 12600: {'lr': 0.0004422104580183649, 'samples': 2419392, 'steps': 12600, 'loss/train': 1.7118310332298279} 01/29/2022 05:02:07 - INFO - codeparrot_training - Step 12601: {'lr': 0.0004421999948104848, 'samples': 2419584, 'steps': 12601, 'loss/train': 1.9450836181640625} 01/29/2022 05:02:11 - INFO - codeparrot_training - Step 12602: {'lr': 0.00044218953077928083, 'samples': 2419776, 'steps': 12602, 'loss/train': 2.0174397826194763} 01/29/2022 05:02:17 - INFO - codeparrot_training - Step 12603: {'lr': 0.000442179065924798, 'samples': 2419968, 'steps': 12603, 'loss/train': 1.4973329901695251} 01/29/2022 05:02:21 - INFO - codeparrot_training - Step 12604: {'lr': 0.0004421686002470811, 'samples': 2420160, 'steps': 12604, 'loss/train': 1.6935580372810364} 01/29/2022 05:02:25 - INFO - codeparrot_training - Step 12605: {'lr': 0.0004421581337461749, 'samples': 2420352, 'steps': 12605, 'loss/train': 0.4677473455667496} 01/29/2022 05:02:29 - INFO - codeparrot_training - Step 12606: {'lr': 0.00044214766642212435, 'samples': 2420544, 'steps': 12606, 'loss/train': 1.8961102366447449} 01/29/2022 05:02:34 - INFO - codeparrot_training - Step 12607: {'lr': 0.00044213719827497413, 'samples': 2420736, 'steps': 12607, 'loss/train': 0.9462493658065796} 01/29/2022 05:02:42 - INFO - codeparrot_training - Step 12608: {'lr': 0.0004421267293047692, 'samples': 2420928, 'steps': 12608, 'loss/train': 1.1715770959854126} 01/29/2022 05:02:46 - INFO - codeparrot_training - Step 12609: {'lr': 0.00044211625951155433, 'samples': 2421120, 'steps': 12609, 'loss/train': 1.5310730338096619} 01/29/2022 05:02:50 - INFO - codeparrot_training - Step 12610: {'lr': 0.00044210578889537446, 'samples': 2421312, 'steps': 12610, 'loss/train': 0.9628008306026459} 01/29/2022 05:02:54 - INFO - codeparrot_training - Step 12611: {'lr': 0.0004420953174562743, 'samples': 2421504, 'steps': 12611, 'loss/train': 1.3402446806430817} 01/29/2022 05:02:59 - INFO - codeparrot_training - Step 12612: {'lr': 0.0004420848451942989, 'samples': 2421696, 'steps': 12612, 'loss/train': 1.997071087360382} 01/29/2022 05:03:04 - INFO - codeparrot_training - Step 12613: {'lr': 0.000442074372109493, 'samples': 2421888, 'steps': 12613, 'loss/train': 0.7827358245849609} 01/29/2022 05:03:08 - INFO - codeparrot_training - Step 12614: {'lr': 0.0004420638982019014, 'samples': 2422080, 'steps': 12614, 'loss/train': 1.9506275653839111} 01/29/2022 05:03:12 - INFO - codeparrot_training - Step 12615: {'lr': 0.0004420534234715691, 'samples': 2422272, 'steps': 12615, 'loss/train': 2.0711896419525146} 01/29/2022 05:03:17 - INFO - codeparrot_training - Step 12616: {'lr': 0.00044204294791854094, 'samples': 2422464, 'steps': 12616, 'loss/train': 1.4947412610054016} 01/29/2022 05:03:21 - INFO - codeparrot_training - Step 12617: {'lr': 0.00044203247154286175, 'samples': 2422656, 'steps': 12617, 'loss/train': 2.3182965517044067} 01/29/2022 05:03:28 - INFO - codeparrot_training - Step 12618: {'lr': 0.0004420219943445765, 'samples': 2422848, 'steps': 12618, 'loss/train': 2.0369160175323486} 01/29/2022 05:03:32 - INFO - codeparrot_training - Step 12619: {'lr': 0.0004420115163237299, 'samples': 2423040, 'steps': 12619, 'loss/train': 1.6558177471160889} 01/29/2022 05:03:36 - INFO - codeparrot_training - Step 12620: {'lr': 0.000442001037480367, 'samples': 2423232, 'steps': 12620, 'loss/train': 1.98753422498703} 01/29/2022 05:03:41 - INFO - codeparrot_training - Step 12621: {'lr': 0.0004419905578145326, 'samples': 2423424, 'steps': 12621, 'loss/train': 2.0271116495132446} 01/29/2022 05:03:45 - INFO - codeparrot_training - Step 12622: {'lr': 0.00044198007732627155, 'samples': 2423616, 'steps': 12622, 'loss/train': 1.6228762865066528} 01/29/2022 05:03:50 - INFO - codeparrot_training - Step 12623: {'lr': 0.00044196959601562884, 'samples': 2423808, 'steps': 12623, 'loss/train': 2.1614155769348145} 01/29/2022 05:03:54 - INFO - codeparrot_training - Step 12624: {'lr': 0.0004419591138826494, 'samples': 2424000, 'steps': 12624, 'loss/train': 1.9775063395500183} 01/29/2022 05:03:58 - INFO - codeparrot_training - Step 12625: {'lr': 0.000441948630927378, 'samples': 2424192, 'steps': 12625, 'loss/train': 1.938023328781128} 01/29/2022 05:04:03 - INFO - codeparrot_training - Step 12626: {'lr': 0.0004419381471498597, 'samples': 2424384, 'steps': 12626, 'loss/train': 1.5775349736213684} 01/29/2022 05:04:07 - INFO - codeparrot_training - Step 12627: {'lr': 0.00044192766255013926, 'samples': 2424576, 'steps': 12627, 'loss/train': 1.4892894923686981} 01/29/2022 05:04:12 - INFO - codeparrot_training - Step 12628: {'lr': 0.0004419171771282616, 'samples': 2424768, 'steps': 12628, 'loss/train': 2.4222185611724854} 01/29/2022 05:04:17 - INFO - codeparrot_training - Step 12629: {'lr': 0.0004419066908842718, 'samples': 2424960, 'steps': 12629, 'loss/train': 1.7887839674949646} 01/29/2022 05:04:21 - INFO - codeparrot_training - Step 12630: {'lr': 0.0004418962038182146, 'samples': 2425152, 'steps': 12630, 'loss/train': 1.9461023211479187} 01/29/2022 05:04:25 - INFO - codeparrot_training - Step 12631: {'lr': 0.00044188571593013504, 'samples': 2425344, 'steps': 12631, 'loss/train': 1.5005711317062378} 01/29/2022 05:04:29 - INFO - codeparrot_training - Step 12632: {'lr': 0.000441875227220078, 'samples': 2425536, 'steps': 12632, 'loss/train': 1.424033761024475} 01/29/2022 05:04:37 - INFO - codeparrot_training - Step 12633: {'lr': 0.00044186473768808844, 'samples': 2425728, 'steps': 12633, 'loss/train': 0.09278112463653088} 01/29/2022 05:04:41 - INFO - codeparrot_training - Step 12634: {'lr': 0.0004418542473342112, 'samples': 2425920, 'steps': 12634, 'loss/train': 1.7632412910461426} 01/29/2022 05:04:45 - INFO - codeparrot_training - Step 12635: {'lr': 0.0004418437561584914, 'samples': 2426112, 'steps': 12635, 'loss/train': 2.0080862045288086} 01/29/2022 05:04:50 - INFO - codeparrot_training - Step 12636: {'lr': 0.00044183326416097373, 'samples': 2426304, 'steps': 12636, 'loss/train': 2.328071415424347} 01/29/2022 05:04:55 - INFO - codeparrot_training - Step 12637: {'lr': 0.0004418227713417033, 'samples': 2426496, 'steps': 12637, 'loss/train': 1.6838917136192322} 01/29/2022 05:04:59 - INFO - codeparrot_training - Step 12638: {'lr': 0.0004418122777007251, 'samples': 2426688, 'steps': 12638, 'loss/train': 2.9536076188087463} 01/29/2022 05:05:04 - INFO - codeparrot_training - Step 12639: {'lr': 0.00044180178323808395, 'samples': 2426880, 'steps': 12639, 'loss/train': 1.8260904550552368} 01/29/2022 05:05:08 - INFO - codeparrot_training - Step 12640: {'lr': 0.00044179128795382493, 'samples': 2427072, 'steps': 12640, 'loss/train': 1.7327088117599487} 01/29/2022 05:05:12 - INFO - codeparrot_training - Step 12641: {'lr': 0.00044178079184799284, 'samples': 2427264, 'steps': 12641, 'loss/train': 1.7099382877349854} 01/29/2022 05:05:16 - INFO - codeparrot_training - Step 12642: {'lr': 0.0004417702949206328, 'samples': 2427456, 'steps': 12642, 'loss/train': 1.6526711583137512} 01/29/2022 05:05:22 - INFO - codeparrot_training - Step 12643: {'lr': 0.0004417597971717897, 'samples': 2427648, 'steps': 12643, 'loss/train': 1.7400612831115723} 01/29/2022 05:05:26 - INFO - codeparrot_training - Step 12644: {'lr': 0.0004417492986015085, 'samples': 2427840, 'steps': 12644, 'loss/train': 1.4412421882152557} 01/29/2022 05:05:30 - INFO - codeparrot_training - Step 12645: {'lr': 0.00044173879920983417, 'samples': 2428032, 'steps': 12645, 'loss/train': 0.5153398364782333} 01/29/2022 05:05:34 - INFO - codeparrot_training - Step 12646: {'lr': 0.00044172829899681175, 'samples': 2428224, 'steps': 12646, 'loss/train': 1.0119255781173706} 01/29/2022 05:05:39 - INFO - codeparrot_training - Step 12647: {'lr': 0.00044171779796248623, 'samples': 2428416, 'steps': 12647, 'loss/train': 2.602136492729187} 01/29/2022 05:05:46 - INFO - codeparrot_training - Step 12648: {'lr': 0.0004417072961069024, 'samples': 2428608, 'steps': 12648, 'loss/train': 1.2084876894950867} 01/29/2022 05:05:50 - INFO - codeparrot_training - Step 12649: {'lr': 0.0004416967934301055, 'samples': 2428800, 'steps': 12649, 'loss/train': 1.5774884819984436} 01/29/2022 05:05:54 - INFO - codeparrot_training - Step 12650: {'lr': 0.00044168628993214036, 'samples': 2428992, 'steps': 12650, 'loss/train': 1.9136683344841003} 01/29/2022 05:05:58 - INFO - codeparrot_training - Step 12651: {'lr': 0.0004416757856130521, 'samples': 2429184, 'steps': 12651, 'loss/train': 0.715473935008049} 01/29/2022 05:06:03 - INFO - codeparrot_training - Step 12652: {'lr': 0.0004416652804728855, 'samples': 2429376, 'steps': 12652, 'loss/train': 1.4473559260368347} 01/29/2022 05:06:08 - INFO - codeparrot_training - Step 12653: {'lr': 0.0004416547745116858, 'samples': 2429568, 'steps': 12653, 'loss/train': 1.9304122924804688} 01/29/2022 05:06:12 - INFO - codeparrot_training - Step 12654: {'lr': 0.00044164426772949785, 'samples': 2429760, 'steps': 12654, 'loss/train': 1.5167550444602966} 01/29/2022 05:06:17 - INFO - codeparrot_training - Step 12655: {'lr': 0.0004416337601263667, 'samples': 2429952, 'steps': 12655, 'loss/train': 2.0048933029174805} 01/29/2022 05:06:21 - INFO - codeparrot_training - Step 12656: {'lr': 0.00044162325170233745, 'samples': 2430144, 'steps': 12656, 'loss/train': 0.9891262650489807} 01/29/2022 05:06:27 - INFO - codeparrot_training - Step 12657: {'lr': 0.00044161274245745497, 'samples': 2430336, 'steps': 12657, 'loss/train': 1.8136430382728577} 01/29/2022 05:06:31 - INFO - codeparrot_training - Step 12658: {'lr': 0.00044160223239176445, 'samples': 2430528, 'steps': 12658, 'loss/train': 2.129329025745392} 01/29/2022 05:06:35 - INFO - codeparrot_training - Step 12659: {'lr': 0.0004415917215053107, 'samples': 2430720, 'steps': 12659, 'loss/train': 1.8865495920181274} 01/29/2022 05:06:39 - INFO - codeparrot_training - Step 12660: {'lr': 0.00044158120979813885, 'samples': 2430912, 'steps': 12660, 'loss/train': 0.37255827337503433} 01/29/2022 05:06:44 - INFO - codeparrot_training - Step 12661: {'lr': 0.000441570697270294, 'samples': 2431104, 'steps': 12661, 'loss/train': 1.2988338768482208} 01/29/2022 05:06:48 - INFO - codeparrot_training - Step 12662: {'lr': 0.00044156018392182105, 'samples': 2431296, 'steps': 12662, 'loss/train': 1.7060921788215637} 01/29/2022 05:06:55 - INFO - codeparrot_training - Step 12663: {'lr': 0.00044154966975276514, 'samples': 2431488, 'steps': 12663, 'loss/train': 0.3906639665365219} 01/29/2022 05:06:59 - INFO - codeparrot_training - Step 12664: {'lr': 0.00044153915476317126, 'samples': 2431680, 'steps': 12664, 'loss/train': 1.383568435907364} 01/29/2022 05:07:04 - INFO - codeparrot_training - Step 12665: {'lr': 0.00044152863895308446, 'samples': 2431872, 'steps': 12665, 'loss/train': 0.7050128281116486} 01/29/2022 05:07:08 - INFO - codeparrot_training - Step 12666: {'lr': 0.0004415181223225497, 'samples': 2432064, 'steps': 12666, 'loss/train': 2.2480849027633667} 01/29/2022 05:07:12 - INFO - codeparrot_training - Step 12667: {'lr': 0.0004415076048716122, 'samples': 2432256, 'steps': 12667, 'loss/train': 1.0198883414268494} 01/29/2022 05:07:18 - INFO - codeparrot_training - Step 12668: {'lr': 0.00044149708660031704, 'samples': 2432448, 'steps': 12668, 'loss/train': 1.9742276072502136} 01/29/2022 05:07:22 - INFO - codeparrot_training - Step 12669: {'lr': 0.000441486567508709, 'samples': 2432640, 'steps': 12669, 'loss/train': 1.6656190752983093} 01/29/2022 05:07:26 - INFO - codeparrot_training - Step 12670: {'lr': 0.0004414760475968334, 'samples': 2432832, 'steps': 12670, 'loss/train': 1.669978380203247} 01/29/2022 05:07:30 - INFO - codeparrot_training - Step 12671: {'lr': 0.0004414655268647352, 'samples': 2433024, 'steps': 12671, 'loss/train': 1.4325159788131714} 01/29/2022 05:07:35 - INFO - codeparrot_training - Step 12672: {'lr': 0.0004414550053124594, 'samples': 2433216, 'steps': 12672, 'loss/train': 2.089853346347809} 01/29/2022 05:07:40 - INFO - codeparrot_training - Step 12673: {'lr': 0.0004414444829400512, 'samples': 2433408, 'steps': 12673, 'loss/train': 1.2283404171466827} 01/29/2022 05:07:44 - INFO - codeparrot_training - Step 12674: {'lr': 0.00044143395974755565, 'samples': 2433600, 'steps': 12674, 'loss/train': 0.1775280497968197} 01/29/2022 05:07:48 - INFO - codeparrot_training - Step 12675: {'lr': 0.00044142343573501787, 'samples': 2433792, 'steps': 12675, 'loss/train': 1.3244779407978058} 01/29/2022 05:07:53 - INFO - codeparrot_training - Step 12676: {'lr': 0.0004414129109024827, 'samples': 2433984, 'steps': 12676, 'loss/train': 1.9868114590644836} 01/29/2022 05:07:57 - INFO - codeparrot_training - Step 12677: {'lr': 0.00044140238524999556, 'samples': 2434176, 'steps': 12677, 'loss/train': 1.8035261034965515} 01/29/2022 05:08:04 - INFO - codeparrot_training - Step 12678: {'lr': 0.0004413918587776013, 'samples': 2434368, 'steps': 12678, 'loss/train': 0.6958416402339935} 01/29/2022 05:08:08 - INFO - codeparrot_training - Step 12679: {'lr': 0.0004413813314853451, 'samples': 2434560, 'steps': 12679, 'loss/train': 1.9704583883285522} 01/29/2022 05:08:13 - INFO - codeparrot_training - Step 12680: {'lr': 0.00044137080337327205, 'samples': 2434752, 'steps': 12680, 'loss/train': 2.267236053943634} 01/29/2022 05:08:17 - INFO - codeparrot_training - Step 12681: {'lr': 0.00044136027444142723, 'samples': 2434944, 'steps': 12681, 'loss/train': 1.2856233716011047} 01/29/2022 05:08:22 - INFO - codeparrot_training - Step 12682: {'lr': 0.0004413497446898558, 'samples': 2435136, 'steps': 12682, 'loss/train': 1.9663242101669312} 01/29/2022 05:08:26 - INFO - codeparrot_training - Step 12683: {'lr': 0.0004413392141186028, 'samples': 2435328, 'steps': 12683, 'loss/train': 1.8186612725257874} 01/29/2022 05:08:31 - INFO - codeparrot_training - Step 12684: {'lr': 0.00044132868272771334, 'samples': 2435520, 'steps': 12684, 'loss/train': 2.248606503009796} 01/29/2022 05:08:35 - INFO - codeparrot_training - Step 12685: {'lr': 0.0004413181505172326, 'samples': 2435712, 'steps': 12685, 'loss/train': 1.4657447040081024} 01/29/2022 05:08:39 - INFO - codeparrot_training - Step 12686: {'lr': 0.0004413076174872056, 'samples': 2435904, 'steps': 12686, 'loss/train': 1.5099586844444275} 01/29/2022 05:08:46 - INFO - codeparrot_training - Step 12687: {'lr': 0.0004412970836376776, 'samples': 2436096, 'steps': 12687, 'loss/train': 1.9614660143852234} 01/29/2022 05:08:51 - INFO - codeparrot_training - Step 12688: {'lr': 0.00044128654896869357, 'samples': 2436288, 'steps': 12688, 'loss/train': 1.303460955619812} 01/29/2022 05:08:55 - INFO - codeparrot_training - Step 12689: {'lr': 0.00044127601348029874, 'samples': 2436480, 'steps': 12689, 'loss/train': 0.6923206150531769} 01/29/2022 05:08:59 - INFO - codeparrot_training - Step 12690: {'lr': 0.0004412654771725382, 'samples': 2436672, 'steps': 12690, 'loss/train': 1.3294703364372253} 01/29/2022 05:09:03 - INFO - codeparrot_training - Step 12691: {'lr': 0.00044125494004545703, 'samples': 2436864, 'steps': 12691, 'loss/train': 1.660435438156128} 01/29/2022 05:09:09 - INFO - codeparrot_training - Step 12692: {'lr': 0.0004412444020991004, 'samples': 2437056, 'steps': 12692, 'loss/train': 1.4426465928554535} 01/29/2022 05:09:13 - INFO - codeparrot_training - Step 12693: {'lr': 0.00044123386333351364, 'samples': 2437248, 'steps': 12693, 'loss/train': 1.8588895797729492} 01/29/2022 05:09:17 - INFO - codeparrot_training - Step 12694: {'lr': 0.00044122332374874166, 'samples': 2437440, 'steps': 12694, 'loss/train': 2.0919467210769653} 01/29/2022 05:09:21 - INFO - codeparrot_training - Step 12695: {'lr': 0.0004412127833448296, 'samples': 2437632, 'steps': 12695, 'loss/train': 1.4895223081111908} 01/29/2022 05:09:26 - INFO - codeparrot_training - Step 12696: {'lr': 0.00044120224212182283, 'samples': 2437824, 'steps': 12696, 'loss/train': 1.6551184058189392} 01/29/2022 05:09:31 - INFO - codeparrot_training - Step 12697: {'lr': 0.0004411917000797663, 'samples': 2438016, 'steps': 12697, 'loss/train': 1.4386522471904755} 01/29/2022 05:09:35 - INFO - codeparrot_training - Step 12698: {'lr': 0.0004411811572187052, 'samples': 2438208, 'steps': 12698, 'loss/train': 2.1388383507728577} 01/29/2022 05:09:39 - INFO - codeparrot_training - Step 12699: {'lr': 0.0004411706135386847, 'samples': 2438400, 'steps': 12699, 'loss/train': 2.1975608468055725} 01/29/2022 05:09:44 - INFO - codeparrot_training - Step 12700: {'lr': 0.0004411600690397501, 'samples': 2438592, 'steps': 12700, 'loss/train': 1.0127954185009003} 01/29/2022 05:09:48 - INFO - codeparrot_training - Step 12701: {'lr': 0.0004411495237219464, 'samples': 2438784, 'steps': 12701, 'loss/train': 2.865829110145569} 01/29/2022 05:09:55 - INFO - codeparrot_training - Step 12702: {'lr': 0.00044113897758531884, 'samples': 2438976, 'steps': 12702, 'loss/train': 0.9863681495189667} 01/29/2022 05:09:59 - INFO - codeparrot_training - Step 12703: {'lr': 0.00044112843062991264, 'samples': 2439168, 'steps': 12703, 'loss/train': 1.818746030330658} 01/29/2022 05:10:03 - INFO - codeparrot_training - Step 12704: {'lr': 0.0004411178828557729, 'samples': 2439360, 'steps': 12704, 'loss/train': 1.5352388620376587} 01/29/2022 05:10:08 - INFO - codeparrot_training - Step 12705: {'lr': 0.00044110733426294484, 'samples': 2439552, 'steps': 12705, 'loss/train': 2.277554512023926} 01/29/2022 05:10:12 - INFO - codeparrot_training - Step 12706: {'lr': 0.00044109678485147367, 'samples': 2439744, 'steps': 12706, 'loss/train': 1.8442078828811646} 01/29/2022 05:10:17 - INFO - codeparrot_training - Step 12707: {'lr': 0.00044108623462140454, 'samples': 2439936, 'steps': 12707, 'loss/train': 1.8193913698196411} 01/29/2022 05:10:21 - INFO - codeparrot_training - Step 12708: {'lr': 0.0004410756835727826, 'samples': 2440128, 'steps': 12708, 'loss/train': 1.491545855998993} 01/29/2022 05:10:26 - INFO - codeparrot_training - Step 12709: {'lr': 0.0004410651317056532, 'samples': 2440320, 'steps': 12709, 'loss/train': 1.9551061391830444} 01/29/2022 05:10:30 - INFO - codeparrot_training - Step 12710: {'lr': 0.0004410545790200614, 'samples': 2440512, 'steps': 12710, 'loss/train': 2.2520915865898132} 01/29/2022 05:10:34 - INFO - codeparrot_training - Step 12711: {'lr': 0.00044104402551605246, 'samples': 2440704, 'steps': 12711, 'loss/train': 1.8951671719551086} 01/29/2022 05:10:39 - INFO - codeparrot_training - Step 12712: {'lr': 0.00044103347119367155, 'samples': 2440896, 'steps': 12712, 'loss/train': 0.8394147455692291} 01/29/2022 05:10:44 - INFO - codeparrot_training - Step 12713: {'lr': 0.0004410229160529639, 'samples': 2441088, 'steps': 12713, 'loss/train': 0.8065033257007599} 01/29/2022 05:10:48 - INFO - codeparrot_training - Step 12714: {'lr': 0.0004410123600939747, 'samples': 2441280, 'steps': 12714, 'loss/train': 1.8774788975715637} 01/29/2022 05:10:52 - INFO - codeparrot_training - Step 12715: {'lr': 0.00044100180331674933, 'samples': 2441472, 'steps': 12715, 'loss/train': 2.481971561908722} 01/29/2022 05:10:56 - INFO - codeparrot_training - Step 12716: {'lr': 0.00044099124572133283, 'samples': 2441664, 'steps': 12716, 'loss/train': 1.9471845030784607} 01/29/2022 05:11:02 - INFO - codeparrot_training - Step 12717: {'lr': 0.0004409806873077704, 'samples': 2441856, 'steps': 12717, 'loss/train': 1.7676181197166443} 01/29/2022 05:11:06 - INFO - codeparrot_training - Step 12718: {'lr': 0.0004409701280761075, 'samples': 2442048, 'steps': 12718, 'loss/train': 1.759041666984558} 01/29/2022 05:11:10 - INFO - codeparrot_training - Step 12719: {'lr': 0.0004409595680263891, 'samples': 2442240, 'steps': 12719, 'loss/train': 2.5526402592658997} 01/29/2022 05:11:14 - INFO - codeparrot_training - Step 12720: {'lr': 0.0004409490071586606, 'samples': 2442432, 'steps': 12720, 'loss/train': 0.4119468033313751} 01/29/2022 05:11:19 - INFO - codeparrot_training - Step 12721: {'lr': 0.00044093844547296715, 'samples': 2442624, 'steps': 12721, 'loss/train': 0.9553551077842712} 01/29/2022 05:11:26 - INFO - codeparrot_training - Step 12722: {'lr': 0.000440927882969354, 'samples': 2442816, 'steps': 12722, 'loss/train': 1.4334447383880615} 01/29/2022 05:11:30 - INFO - codeparrot_training - Step 12723: {'lr': 0.0004409173196478665, 'samples': 2443008, 'steps': 12723, 'loss/train': 2.011676073074341} 01/29/2022 05:11:34 - INFO - codeparrot_training - Step 12724: {'lr': 0.00044090675550854973, 'samples': 2443200, 'steps': 12724, 'loss/train': 1.6317581534385681} 01/29/2022 05:11:39 - INFO - codeparrot_training - Step 12725: {'lr': 0.00044089619055144916, 'samples': 2443392, 'steps': 12725, 'loss/train': 1.6669172644615173} 01/29/2022 05:11:43 - INFO - codeparrot_training - Step 12726: {'lr': 0.0004408856247766098, 'samples': 2443584, 'steps': 12726, 'loss/train': 1.1767698526382446} 01/29/2022 05:11:47 - INFO - codeparrot_training - Step 12727: {'lr': 0.00044087505818407715, 'samples': 2443776, 'steps': 12727, 'loss/train': 1.7317762970924377} 01/29/2022 05:11:53 - INFO - codeparrot_training - Step 12728: {'lr': 0.00044086449077389636, 'samples': 2443968, 'steps': 12728, 'loss/train': 2.0336095690727234} 01/29/2022 05:11:57 - INFO - codeparrot_training - Step 12729: {'lr': 0.0004408539225461126, 'samples': 2444160, 'steps': 12729, 'loss/train': 1.6728134751319885} 01/29/2022 05:12:01 - INFO - codeparrot_training - Step 12730: {'lr': 0.0004408433535007713, 'samples': 2444352, 'steps': 12730, 'loss/train': 1.4579365253448486} 01/29/2022 05:12:05 - INFO - codeparrot_training - Step 12731: {'lr': 0.0004408327836379177, 'samples': 2444544, 'steps': 12731, 'loss/train': 2.1353947520256042} 01/29/2022 05:12:11 - INFO - codeparrot_training - Step 12732: {'lr': 0.0004408222129575969, 'samples': 2444736, 'steps': 12732, 'loss/train': 0.8459131121635437} 01/29/2022 05:12:15 - INFO - codeparrot_training - Step 12733: {'lr': 0.0004408116414598545, 'samples': 2444928, 'steps': 12733, 'loss/train': 2.5016820430755615} 01/29/2022 05:12:19 - INFO - codeparrot_training - Step 12734: {'lr': 0.0004408010691447356, 'samples': 2445120, 'steps': 12734, 'loss/train': 1.9830000400543213} 01/29/2022 05:12:23 - INFO - codeparrot_training - Step 12735: {'lr': 0.00044079049601228543, 'samples': 2445312, 'steps': 12735, 'loss/train': 1.7136086225509644} 01/29/2022 05:12:28 - INFO - codeparrot_training - Step 12736: {'lr': 0.00044077992206254934, 'samples': 2445504, 'steps': 12736, 'loss/train': 1.96657133102417} 01/29/2022 05:12:35 - INFO - codeparrot_training - Step 12737: {'lr': 0.0004407693472955727, 'samples': 2445696, 'steps': 12737, 'loss/train': 2.3129947185516357} 01/29/2022 05:12:39 - INFO - codeparrot_training - Step 12738: {'lr': 0.00044075877171140075, 'samples': 2445888, 'steps': 12738, 'loss/train': 2.2722004652023315} 01/29/2022 05:12:43 - INFO - codeparrot_training - Step 12739: {'lr': 0.00044074819531007885, 'samples': 2446080, 'steps': 12739, 'loss/train': 0.15025253221392632} 01/29/2022 05:12:47 - INFO - codeparrot_training - Step 12740: {'lr': 0.0004407376180916522, 'samples': 2446272, 'steps': 12740, 'loss/train': 1.3647170662879944} 01/29/2022 05:12:52 - INFO - codeparrot_training - Step 12741: {'lr': 0.00044072704005616614, 'samples': 2446464, 'steps': 12741, 'loss/train': 1.9606693983078003} 01/29/2022 05:12:57 - INFO - codeparrot_training - Step 12742: {'lr': 0.00044071646120366604, 'samples': 2446656, 'steps': 12742, 'loss/train': 2.0925614833831787} 01/29/2022 05:13:01 - INFO - codeparrot_training - Step 12743: {'lr': 0.00044070588153419715, 'samples': 2446848, 'steps': 12743, 'loss/train': 1.9030762910842896} 01/29/2022 05:13:05 - INFO - codeparrot_training - Step 12744: {'lr': 0.00044069530104780486, 'samples': 2447040, 'steps': 12744, 'loss/train': 2.0480421781539917} 01/29/2022 05:13:10 - INFO - codeparrot_training - Step 12745: {'lr': 0.00044068471974453437, 'samples': 2447232, 'steps': 12745, 'loss/train': 2.233704149723053} 01/29/2022 05:13:14 - INFO - codeparrot_training - Step 12746: {'lr': 0.0004406741376244312, 'samples': 2447424, 'steps': 12746, 'loss/train': 1.9561462998390198} 01/29/2022 05:13:22 - INFO - codeparrot_training - Step 12747: {'lr': 0.00044066355468754047, 'samples': 2447616, 'steps': 12747, 'loss/train': 1.150499939918518} 01/29/2022 05:13:26 - INFO - codeparrot_training - Step 12748: {'lr': 0.00044065297093390764, 'samples': 2447808, 'steps': 12748, 'loss/train': 1.445552945137024} 01/29/2022 05:13:30 - INFO - codeparrot_training - Step 12749: {'lr': 0.0004406423863635781, 'samples': 2448000, 'steps': 12749, 'loss/train': 1.8823294043540955} 01/29/2022 05:13:35 - INFO - codeparrot_training - Step 12750: {'lr': 0.00044063180097659704, 'samples': 2448192, 'steps': 12750, 'loss/train': 2.3707852363586426} 01/29/2022 05:13:39 - INFO - codeparrot_training - Step 12751: {'lr': 0.00044062121477300985, 'samples': 2448384, 'steps': 12751, 'loss/train': 1.2573887407779694} 01/29/2022 05:13:44 - INFO - codeparrot_training - Step 12752: {'lr': 0.000440610627752862, 'samples': 2448576, 'steps': 12752, 'loss/train': 1.5080206990242004} 01/29/2022 05:13:48 - INFO - codeparrot_training - Step 12753: {'lr': 0.0004406000399161987, 'samples': 2448768, 'steps': 12753, 'loss/train': 2.052558660507202} 01/29/2022 05:13:53 - INFO - codeparrot_training - Step 12754: {'lr': 0.00044058945126306535, 'samples': 2448960, 'steps': 12754, 'loss/train': 5.193202614784241} 01/29/2022 05:13:57 - INFO - codeparrot_training - Step 12755: {'lr': 0.0004405788617935073, 'samples': 2449152, 'steps': 12755, 'loss/train': 1.8356852531433105} 01/29/2022 05:14:01 - INFO - codeparrot_training - Step 12756: {'lr': 0.0004405682715075699, 'samples': 2449344, 'steps': 12756, 'loss/train': 0.6322875469923019} 01/29/2022 05:14:06 - INFO - codeparrot_training - Step 12757: {'lr': 0.0004405576804052985, 'samples': 2449536, 'steps': 12757, 'loss/train': 1.7244361639022827} 01/29/2022 05:14:10 - INFO - codeparrot_training - Step 12758: {'lr': 0.0004405470884867386, 'samples': 2449728, 'steps': 12758, 'loss/train': 2.075754225254059} 01/29/2022 05:14:15 - INFO - codeparrot_training - Step 12759: {'lr': 0.00044053649575193543, 'samples': 2449920, 'steps': 12759, 'loss/train': 0.967012882232666} 01/29/2022 05:14:19 - INFO - codeparrot_training - Step 12760: {'lr': 0.00044052590220093445, 'samples': 2450112, 'steps': 12760, 'loss/train': 0.3451038524508476} 01/29/2022 05:14:23 - INFO - codeparrot_training - Step 12761: {'lr': 0.00044051530783378103, 'samples': 2450304, 'steps': 12761, 'loss/train': 1.4458031058311462} 01/29/2022 05:14:30 - INFO - codeparrot_training - Step 12762: {'lr': 0.0004405047126505204, 'samples': 2450496, 'steps': 12762, 'loss/train': 1.8601976037025452} 01/29/2022 05:14:34 - INFO - codeparrot_training - Step 12763: {'lr': 0.0004404941166511982, 'samples': 2450688, 'steps': 12763, 'loss/train': 1.6797171235084534} 01/29/2022 05:14:39 - INFO - codeparrot_training - Step 12764: {'lr': 0.00044048351983585966, 'samples': 2450880, 'steps': 12764, 'loss/train': 1.950163722038269} 01/29/2022 05:14:43 - INFO - codeparrot_training - Step 12765: {'lr': 0.00044047292220455016, 'samples': 2451072, 'steps': 12765, 'loss/train': 1.6445692777633667} 01/29/2022 05:14:47 - INFO - codeparrot_training - Step 12766: {'lr': 0.0004404623237573152, 'samples': 2451264, 'steps': 12766, 'loss/train': 1.6649145483970642} 01/29/2022 05:14:53 - INFO - codeparrot_training - Step 12767: {'lr': 0.00044045172449420005, 'samples': 2451456, 'steps': 12767, 'loss/train': 2.311026871204376} 01/29/2022 05:14:57 - INFO - codeparrot_training - Step 12768: {'lr': 0.00044044112441525026, 'samples': 2451648, 'steps': 12768, 'loss/train': 1.614140510559082} 01/29/2022 05:15:01 - INFO - codeparrot_training - Step 12769: {'lr': 0.0004404305235205112, 'samples': 2451840, 'steps': 12769, 'loss/train': 1.1773547530174255} 01/29/2022 05:15:05 - INFO - codeparrot_training - Step 12770: {'lr': 0.0004404199218100281, 'samples': 2452032, 'steps': 12770, 'loss/train': 1.0589320063591003} 01/29/2022 05:15:10 - INFO - codeparrot_training - Step 12771: {'lr': 0.00044040931928384665, 'samples': 2452224, 'steps': 12771, 'loss/train': 0.9789971709251404} 01/29/2022 05:15:15 - INFO - codeparrot_training - Step 12772: {'lr': 0.0004403987159420121, 'samples': 2452416, 'steps': 12772, 'loss/train': 1.0980230569839478} 01/29/2022 05:15:19 - INFO - codeparrot_training - Step 12773: {'lr': 0.0004403881117845699, 'samples': 2452608, 'steps': 12773, 'loss/train': 1.0606946647167206} 01/29/2022 05:15:24 - INFO - codeparrot_training - Step 12774: {'lr': 0.00044037750681156547, 'samples': 2452800, 'steps': 12774, 'loss/train': 1.4732279777526855} 01/29/2022 05:15:28 - INFO - codeparrot_training - Step 12775: {'lr': 0.0004403669010230443, 'samples': 2452992, 'steps': 12775, 'loss/train': 1.3100104629993439} 01/29/2022 05:15:32 - INFO - codeparrot_training - Step 12776: {'lr': 0.00044035629441905173, 'samples': 2453184, 'steps': 12776, 'loss/train': 1.259375900030136} 01/29/2022 05:15:37 - INFO - codeparrot_training - Step 12777: {'lr': 0.0004403456869996333, 'samples': 2453376, 'steps': 12777, 'loss/train': 1.8579847812652588} 01/29/2022 05:15:41 - INFO - codeparrot_training - Step 12778: {'lr': 0.0004403350787648343, 'samples': 2453568, 'steps': 12778, 'loss/train': 1.7371625304222107} 01/29/2022 05:15:46 - INFO - codeparrot_training - Step 12779: {'lr': 0.0004403244697147003, 'samples': 2453760, 'steps': 12779, 'loss/train': 0.12399648874998093} 01/29/2022 05:15:50 - INFO - codeparrot_training - Step 12780: {'lr': 0.00044031385984927675, 'samples': 2453952, 'steps': 12780, 'loss/train': 1.8133785724639893} 01/29/2022 05:15:54 - INFO - codeparrot_training - Step 12781: {'lr': 0.000440303249168609, 'samples': 2454144, 'steps': 12781, 'loss/train': 1.8810975551605225} 01/29/2022 05:16:01 - INFO - codeparrot_training - Step 12782: {'lr': 0.0004402926376727425, 'samples': 2454336, 'steps': 12782, 'loss/train': 1.359175980091095} 01/29/2022 05:16:06 - INFO - codeparrot_training - Step 12783: {'lr': 0.0004402820253617229, 'samples': 2454528, 'steps': 12783, 'loss/train': 1.7063035368919373} 01/29/2022 05:16:10 - INFO - codeparrot_training - Step 12784: {'lr': 0.0004402714122355955, 'samples': 2454720, 'steps': 12784, 'loss/train': 1.6151788830757141} 01/29/2022 05:16:14 - INFO - codeparrot_training - Step 12785: {'lr': 0.00044026079829440567, 'samples': 2454912, 'steps': 12785, 'loss/train': 1.8653456568717957} 01/29/2022 05:16:18 - INFO - codeparrot_training - Step 12786: {'lr': 0.0004402501835381991, 'samples': 2455104, 'steps': 12786, 'loss/train': 1.3108440935611725} 01/29/2022 05:16:24 - INFO - codeparrot_training - Step 12787: {'lr': 0.00044023956796702116, 'samples': 2455296, 'steps': 12787, 'loss/train': 2.3834139704704285} 01/29/2022 05:16:28 - INFO - codeparrot_training - Step 12788: {'lr': 0.0004402289515809172, 'samples': 2455488, 'steps': 12788, 'loss/train': 1.8589852452278137} 01/29/2022 05:16:32 - INFO - codeparrot_training - Step 12789: {'lr': 0.00044021833437993296, 'samples': 2455680, 'steps': 12789, 'loss/train': 2.5922024846076965} 01/29/2022 05:16:37 - INFO - codeparrot_training - Step 12790: {'lr': 0.0004402077163641137, 'samples': 2455872, 'steps': 12790, 'loss/train': 1.5013271570205688} 01/29/2022 05:16:41 - INFO - codeparrot_training - Step 12791: {'lr': 0.000440197097533505, 'samples': 2456064, 'steps': 12791, 'loss/train': 1.5363873839378357} 01/29/2022 05:16:48 - INFO - codeparrot_training - Step 12792: {'lr': 0.00044018647788815235, 'samples': 2456256, 'steps': 12792, 'loss/train': 1.1879550218582153} 01/29/2022 05:16:52 - INFO - codeparrot_training - Step 12793: {'lr': 0.00044017585742810124, 'samples': 2456448, 'steps': 12793, 'loss/train': 1.0080689191818237} 01/29/2022 05:16:57 - INFO - codeparrot_training - Step 12794: {'lr': 0.0004401652361533971, 'samples': 2456640, 'steps': 12794, 'loss/train': 2.0906853675842285} 01/29/2022 05:17:01 - INFO - codeparrot_training - Step 12795: {'lr': 0.00044015461406408544, 'samples': 2456832, 'steps': 12795, 'loss/train': 1.911319613456726} 01/29/2022 05:17:05 - INFO - codeparrot_training - Step 12796: {'lr': 0.00044014399116021184, 'samples': 2457024, 'steps': 12796, 'loss/train': 1.761464238166809} 01/29/2022 05:17:11 - INFO - codeparrot_training - Step 12797: {'lr': 0.00044013336744182176, 'samples': 2457216, 'steps': 12797, 'loss/train': 1.8234713673591614} 01/29/2022 05:17:15 - INFO - codeparrot_training - Step 12798: {'lr': 0.0004401227429089607, 'samples': 2457408, 'steps': 12798, 'loss/train': 1.5808430314064026} 01/29/2022 05:17:19 - INFO - codeparrot_training - Step 12799: {'lr': 0.00044011211756167425, 'samples': 2457600, 'steps': 12799, 'loss/train': 2.262382686138153} 01/29/2022 05:17:23 - INFO - codeparrot_training - Step 12800: {'lr': 0.0004401014914000078, 'samples': 2457792, 'steps': 12800, 'loss/train': 1.3939415216445923} 01/29/2022 05:17:28 - INFO - codeparrot_training - Step 12801: {'lr': 0.00044009086442400684, 'samples': 2457984, 'steps': 12801, 'loss/train': 0.18209557235240936} 01/29/2022 05:17:33 - INFO - codeparrot_training - Step 12802: {'lr': 0.0004400802366337171, 'samples': 2458176, 'steps': 12802, 'loss/train': 1.7358539700508118} 01/29/2022 05:17:37 - INFO - codeparrot_training - Step 12803: {'lr': 0.00044006960802918393, 'samples': 2458368, 'steps': 12803, 'loss/train': 1.7693622708320618} 01/29/2022 05:17:41 - INFO - codeparrot_training - Step 12804: {'lr': 0.0004400589786104529, 'samples': 2458560, 'steps': 12804, 'loss/train': 1.4138893783092499} 01/29/2022 05:17:46 - INFO - codeparrot_training - Step 12805: {'lr': 0.0004400483483775696, 'samples': 2458752, 'steps': 12805, 'loss/train': 2.3015055656433105} 01/29/2022 05:17:50 - INFO - codeparrot_training - Step 12806: {'lr': 0.00044003771733057943, 'samples': 2458944, 'steps': 12806, 'loss/train': 1.4505667984485626} 01/29/2022 05:17:56 - INFO - codeparrot_training - Step 12807: {'lr': 0.0004400270854695281, 'samples': 2459136, 'steps': 12807, 'loss/train': 0.8233286440372467} 01/29/2022 05:18:00 - INFO - codeparrot_training - Step 12808: {'lr': 0.0004400164527944611, 'samples': 2459328, 'steps': 12808, 'loss/train': 0.6352570205926895} 01/29/2022 05:18:04 - INFO - codeparrot_training - Step 12809: {'lr': 0.0004400058193054239, 'samples': 2459520, 'steps': 12809, 'loss/train': 1.7343460321426392} 01/29/2022 05:18:09 - INFO - codeparrot_training - Step 12810: {'lr': 0.0004399951850024621, 'samples': 2459712, 'steps': 12810, 'loss/train': 1.2363418936729431} 01/29/2022 05:18:13 - INFO - codeparrot_training - Step 12811: {'lr': 0.0004399845498856213, 'samples': 2459904, 'steps': 12811, 'loss/train': 1.3014071881771088} 01/29/2022 05:18:18 - INFO - codeparrot_training - Step 12812: {'lr': 0.000439973913954947, 'samples': 2460096, 'steps': 12812, 'loss/train': 1.7780431509017944} 01/29/2022 05:18:23 - INFO - codeparrot_training - Step 12813: {'lr': 0.0004399632772104848, 'samples': 2460288, 'steps': 12813, 'loss/train': 2.5385618805885315} 01/29/2022 05:18:27 - INFO - codeparrot_training - Step 12814: {'lr': 0.00043995263965228016, 'samples': 2460480, 'steps': 12814, 'loss/train': 0.8797463178634644} 01/29/2022 05:18:31 - INFO - codeparrot_training - Step 12815: {'lr': 0.00043994200128037877, 'samples': 2460672, 'steps': 12815, 'loss/train': 1.0306484699249268} 01/29/2022 05:18:35 - INFO - codeparrot_training - Step 12816: {'lr': 0.0004399313620948262, 'samples': 2460864, 'steps': 12816, 'loss/train': 1.8510403633117676} 01/29/2022 05:18:40 - INFO - codeparrot_training - Step 12817: {'lr': 0.00043992072209566793, 'samples': 2461056, 'steps': 12817, 'loss/train': 2.3746862411499023} 01/29/2022 05:18:45 - INFO - codeparrot_training - Step 12818: {'lr': 0.0004399100812829496, 'samples': 2461248, 'steps': 12818, 'loss/train': 1.1694561839103699} 01/29/2022 05:18:50 - INFO - codeparrot_training - Step 12819: {'lr': 0.00043989943965671685, 'samples': 2461440, 'steps': 12819, 'loss/train': 1.6783454418182373} 01/29/2022 05:18:54 - INFO - codeparrot_training - Step 12820: {'lr': 0.00043988879721701515, 'samples': 2461632, 'steps': 12820, 'loss/train': 1.8028650283813477} 01/29/2022 05:18:58 - INFO - codeparrot_training - Step 12821: {'lr': 0.0004398781539638901, 'samples': 2461824, 'steps': 12821, 'loss/train': 1.786424696445465} 01/29/2022 05:19:02 - INFO - codeparrot_training - Step 12822: {'lr': 0.00043986750989738737, 'samples': 2462016, 'steps': 12822, 'loss/train': 2.1693121790885925} 01/29/2022 05:19:08 - INFO - codeparrot_training - Step 12823: {'lr': 0.0004398568650175525, 'samples': 2462208, 'steps': 12823, 'loss/train': 1.5343453288078308} 01/29/2022 05:19:12 - INFO - codeparrot_training - Step 12824: {'lr': 0.00043984621932443115, 'samples': 2462400, 'steps': 12824, 'loss/train': 1.9312137365341187} 01/29/2022 05:19:16 - INFO - codeparrot_training - Step 12825: {'lr': 0.0004398355728180689, 'samples': 2462592, 'steps': 12825, 'loss/train': 1.5247923731803894} 01/29/2022 05:19:20 - INFO - codeparrot_training - Step 12826: {'lr': 0.0004398249254985113, 'samples': 2462784, 'steps': 12826, 'loss/train': 2.3229868412017822} 01/29/2022 05:19:25 - INFO - codeparrot_training - Step 12827: {'lr': 0.00043981427736580395, 'samples': 2462976, 'steps': 12827, 'loss/train': 2.0270209908485413} 01/29/2022 05:19:30 - INFO - codeparrot_training - Step 12828: {'lr': 0.00043980362841999253, 'samples': 2463168, 'steps': 12828, 'loss/train': 0.9276581704616547} 01/29/2022 05:19:34 - INFO - codeparrot_training - Step 12829: {'lr': 0.0004397929786611227, 'samples': 2463360, 'steps': 12829, 'loss/train': 1.701474666595459} 01/29/2022 05:19:39 - INFO - codeparrot_training - Step 12830: {'lr': 0.00043978232808923996, 'samples': 2463552, 'steps': 12830, 'loss/train': 1.735746145248413} 01/29/2022 05:19:43 - INFO - codeparrot_training - Step 12831: {'lr': 0.00043977167670439, 'samples': 2463744, 'steps': 12831, 'loss/train': 1.6250710487365723} 01/29/2022 05:19:47 - INFO - codeparrot_training - Step 12832: {'lr': 0.0004397610245066184, 'samples': 2463936, 'steps': 12832, 'loss/train': 2.079562246799469} 01/29/2022 05:19:52 - INFO - codeparrot_training - Step 12833: {'lr': 0.00043975037149597085, 'samples': 2464128, 'steps': 12833, 'loss/train': 1.2603830695152283} 01/29/2022 05:19:56 - INFO - codeparrot_training - Step 12834: {'lr': 0.00043973971767249297, 'samples': 2464320, 'steps': 12834, 'loss/train': 0.461931511759758} 01/29/2022 05:20:01 - INFO - codeparrot_training - Step 12835: {'lr': 0.0004397290630362304, 'samples': 2464512, 'steps': 12835, 'loss/train': 1.0081153213977814} 01/29/2022 05:20:05 - INFO - codeparrot_training - Step 12836: {'lr': 0.0004397184075872288, 'samples': 2464704, 'steps': 12836, 'loss/train': 0.642036572098732} 01/29/2022 05:20:09 - INFO - codeparrot_training - Step 12837: {'lr': 0.00043970775132553375, 'samples': 2464896, 'steps': 12837, 'loss/train': 0.9043226838111877} 01/29/2022 05:20:15 - INFO - codeparrot_training - Step 12838: {'lr': 0.00043969709425119085, 'samples': 2465088, 'steps': 12838, 'loss/train': 1.4068498313426971} 01/29/2022 05:20:19 - INFO - codeparrot_training - Step 12839: {'lr': 0.000439686436364246, 'samples': 2465280, 'steps': 12839, 'loss/train': 1.3583035469055176} 01/29/2022 05:20:23 - INFO - codeparrot_training - Step 12840: {'lr': 0.00043967577766474455, 'samples': 2465472, 'steps': 12840, 'loss/train': 1.819860577583313} 01/29/2022 05:20:28 - INFO - codeparrot_training - Step 12841: {'lr': 0.00043966511815273233, 'samples': 2465664, 'steps': 12841, 'loss/train': 1.4793394804000854} 01/29/2022 05:20:32 - INFO - codeparrot_training - Step 12842: {'lr': 0.00043965445782825495, 'samples': 2465856, 'steps': 12842, 'loss/train': 1.5403060913085938} 01/29/2022 05:20:38 - INFO - codeparrot_training - Step 12843: {'lr': 0.00043964379669135815, 'samples': 2466048, 'steps': 12843, 'loss/train': 1.5791094303131104} 01/29/2022 05:20:42 - INFO - codeparrot_training - Step 12844: {'lr': 0.00043963313474208753, 'samples': 2466240, 'steps': 12844, 'loss/train': 1.9080972075462341} 01/29/2022 05:20:46 - INFO - codeparrot_training - Step 12845: {'lr': 0.0004396224719804888, 'samples': 2466432, 'steps': 12845, 'loss/train': 1.7312917113304138} 01/29/2022 05:20:50 - INFO - codeparrot_training - Step 12846: {'lr': 0.0004396118084066075, 'samples': 2466624, 'steps': 12846, 'loss/train': 1.7552717328071594} 01/29/2022 05:20:56 - INFO - codeparrot_training - Step 12847: {'lr': 0.00043960114402048957, 'samples': 2466816, 'steps': 12847, 'loss/train': 1.7503511309623718} 01/29/2022 05:21:00 - INFO - codeparrot_training - Step 12848: {'lr': 0.0004395904788221805, 'samples': 2467008, 'steps': 12848, 'loss/train': 1.77873694896698} 01/29/2022 05:21:04 - INFO - codeparrot_training - Step 12849: {'lr': 0.00043957981281172597, 'samples': 2467200, 'steps': 12849, 'loss/train': 1.914694368839264} 01/29/2022 05:21:08 - INFO - codeparrot_training - Step 12850: {'lr': 0.00043956914598917177, 'samples': 2467392, 'steps': 12850, 'loss/train': 1.1050477623939514} 01/29/2022 05:21:13 - INFO - codeparrot_training - Step 12851: {'lr': 0.00043955847835456353, 'samples': 2467584, 'steps': 12851, 'loss/train': 0.6074629873037338} 01/29/2022 05:21:17 - INFO - codeparrot_training - Step 12852: {'lr': 0.00043954780990794695, 'samples': 2467776, 'steps': 12852, 'loss/train': 0.6613724827766418} 01/29/2022 05:21:23 - INFO - codeparrot_training - Step 12853: {'lr': 0.0004395371406493677, 'samples': 2467968, 'steps': 12853, 'loss/train': 1.0041609406471252} 01/29/2022 05:21:27 - INFO - codeparrot_training - Step 12854: {'lr': 0.0004395264705788716, 'samples': 2468160, 'steps': 12854, 'loss/train': 2.3023568987846375} 01/29/2022 05:21:31 - INFO - codeparrot_training - Step 12855: {'lr': 0.00043951579969650424, 'samples': 2468352, 'steps': 12855, 'loss/train': 1.585341453552246} 01/29/2022 05:21:36 - INFO - codeparrot_training - Step 12856: {'lr': 0.00043950512800231136, 'samples': 2468544, 'steps': 12856, 'loss/train': 2.294011652469635} 01/29/2022 05:21:40 - INFO - codeparrot_training - Step 12857: {'lr': 0.0004394944554963387, 'samples': 2468736, 'steps': 12857, 'loss/train': 2.357441246509552} 01/29/2022 05:21:45 - INFO - codeparrot_training - Step 12858: {'lr': 0.000439483782178632, 'samples': 2468928, 'steps': 12858, 'loss/train': 1.9966887831687927} 01/29/2022 05:21:49 - INFO - codeparrot_training - Step 12859: {'lr': 0.0004394731080492369, 'samples': 2469120, 'steps': 12859, 'loss/train': 1.7292348146438599} 01/29/2022 05:21:53 - INFO - codeparrot_training - Step 12860: {'lr': 0.0004394624331081992, 'samples': 2469312, 'steps': 12860, 'loss/train': 1.659792423248291} 01/29/2022 05:21:58 - INFO - codeparrot_training - Step 12861: {'lr': 0.00043945175735556454, 'samples': 2469504, 'steps': 12861, 'loss/train': 1.659350574016571} 01/29/2022 05:22:02 - INFO - codeparrot_training - Step 12862: {'lr': 0.0004394410807913788, 'samples': 2469696, 'steps': 12862, 'loss/train': 1.0988935232162476} 01/29/2022 05:22:08 - INFO - codeparrot_training - Step 12863: {'lr': 0.0004394304034156875, 'samples': 2469888, 'steps': 12863, 'loss/train': 0.12750555574893951} 01/29/2022 05:22:12 - INFO - codeparrot_training - Step 12864: {'lr': 0.00043941972522853665, 'samples': 2470080, 'steps': 12864, 'loss/train': 7.430654525756836} 01/29/2022 05:22:16 - INFO - codeparrot_training - Step 12865: {'lr': 0.00043940904622997176, 'samples': 2470272, 'steps': 12865, 'loss/train': 1.785643458366394} 01/29/2022 05:22:20 - INFO - codeparrot_training - Step 12866: {'lr': 0.00043939836642003865, 'samples': 2470464, 'steps': 12866, 'loss/train': 1.427980363368988} 01/29/2022 05:22:25 - INFO - codeparrot_training - Step 12867: {'lr': 0.0004393876857987831, 'samples': 2470656, 'steps': 12867, 'loss/train': 2.7295698523521423} 01/29/2022 05:22:30 - INFO - codeparrot_training - Step 12868: {'lr': 0.0004393770043662508, 'samples': 2470848, 'steps': 12868, 'loss/train': 0.521094337105751} 01/29/2022 05:22:34 - INFO - codeparrot_training - Step 12869: {'lr': 0.0004393663221224876, 'samples': 2471040, 'steps': 12869, 'loss/train': 1.9501522779464722} 01/29/2022 05:22:38 - INFO - codeparrot_training - Step 12870: {'lr': 0.00043935563906753923, 'samples': 2471232, 'steps': 12870, 'loss/train': 2.116305649280548} 01/29/2022 05:22:43 - INFO - codeparrot_training - Step 12871: {'lr': 0.0004393449552014514, 'samples': 2471424, 'steps': 12871, 'loss/train': 1.4569865763187408} 01/29/2022 05:22:47 - INFO - codeparrot_training - Step 12872: {'lr': 0.00043933427052426986, 'samples': 2471616, 'steps': 12872, 'loss/train': 1.56919527053833} 01/29/2022 05:22:52 - INFO - codeparrot_training - Step 12873: {'lr': 0.00043932358503604054, 'samples': 2471808, 'steps': 12873, 'loss/train': 1.7953269481658936} 01/29/2022 05:22:57 - INFO - codeparrot_training - Step 12874: {'lr': 0.000439312898736809, 'samples': 2472000, 'steps': 12874, 'loss/train': 0.5533168762922287} 01/29/2022 05:23:01 - INFO - codeparrot_training - Step 12875: {'lr': 0.00043930221162662115, 'samples': 2472192, 'steps': 12875, 'loss/train': 1.6240224838256836} 01/29/2022 05:23:05 - INFO - codeparrot_training - Step 12876: {'lr': 0.0004392915237055227, 'samples': 2472384, 'steps': 12876, 'loss/train': 1.648563802242279} 01/29/2022 05:23:11 - INFO - codeparrot_training - Step 12877: {'lr': 0.00043928083497355954, 'samples': 2472576, 'steps': 12877, 'loss/train': 2.737821936607361} 01/29/2022 05:23:16 - INFO - codeparrot_training - Step 12878: {'lr': 0.0004392701454307773, 'samples': 2472768, 'steps': 12878, 'loss/train': 1.1448459327220917} 01/29/2022 05:23:20 - INFO - codeparrot_training - Step 12879: {'lr': 0.00043925945507722195, 'samples': 2472960, 'steps': 12879, 'loss/train': 0.9246217310428619} 01/29/2022 05:23:24 - INFO - codeparrot_training - Step 12880: {'lr': 0.0004392487639129391, 'samples': 2473152, 'steps': 12880, 'loss/train': 1.5727066397666931} 01/29/2022 05:23:29 - INFO - codeparrot_training - Step 12881: {'lr': 0.0004392380719379747, 'samples': 2473344, 'steps': 12881, 'loss/train': 1.014132410287857} 01/29/2022 05:23:34 - INFO - codeparrot_training - Step 12882: {'lr': 0.0004392273791523744, 'samples': 2473536, 'steps': 12882, 'loss/train': 1.6371143460273743} 01/29/2022 05:23:38 - INFO - codeparrot_training - Step 12883: {'lr': 0.0004392166855561842, 'samples': 2473728, 'steps': 12883, 'loss/train': 1.7523667216300964} 01/29/2022 05:23:42 - INFO - codeparrot_training - Step 12884: {'lr': 0.0004392059911494498, 'samples': 2473920, 'steps': 12884, 'loss/train': 1.972130835056305} 01/29/2022 05:23:47 - INFO - codeparrot_training - Step 12885: {'lr': 0.00043919529593221696, 'samples': 2474112, 'steps': 12885, 'loss/train': 1.8712769150733948} 01/29/2022 05:23:51 - INFO - codeparrot_training - Step 12886: {'lr': 0.00043918459990453156, 'samples': 2474304, 'steps': 12886, 'loss/train': 0.9470495581626892} 01/29/2022 05:23:57 - INFO - codeparrot_training - Step 12887: {'lr': 0.00043917390306643945, 'samples': 2474496, 'steps': 12887, 'loss/train': 1.220715343952179} 01/29/2022 05:24:01 - INFO - codeparrot_training - Step 12888: {'lr': 0.0004391632054179864, 'samples': 2474688, 'steps': 12888, 'loss/train': 1.328229546546936} 01/29/2022 05:24:05 - INFO - codeparrot_training - Step 12889: {'lr': 0.00043915250695921815, 'samples': 2474880, 'steps': 12889, 'loss/train': 1.911810278892517} 01/29/2022 05:24:10 - INFO - codeparrot_training - Step 12890: {'lr': 0.00043914180769018073, 'samples': 2475072, 'steps': 12890, 'loss/train': 2.380773961544037} 01/29/2022 05:24:14 - INFO - codeparrot_training - Step 12891: {'lr': 0.0004391311076109198, 'samples': 2475264, 'steps': 12891, 'loss/train': 0.7239826619625092} 01/29/2022 05:24:19 - INFO - codeparrot_training - Step 12892: {'lr': 0.00043912040672148135, 'samples': 2475456, 'steps': 12892, 'loss/train': 0.5625948160886765} 01/29/2022 05:24:23 - INFO - codeparrot_training - Step 12893: {'lr': 0.00043910970502191105, 'samples': 2475648, 'steps': 12893, 'loss/train': 1.6311684250831604} 01/29/2022 05:24:28 - INFO - codeparrot_training - Step 12894: {'lr': 0.00043909900251225476, 'samples': 2475840, 'steps': 12894, 'loss/train': 2.745983362197876} 01/29/2022 05:24:32 - INFO - codeparrot_training - Step 12895: {'lr': 0.00043908829919255855, 'samples': 2476032, 'steps': 12895, 'loss/train': 1.2991851568222046} 01/29/2022 05:24:36 - INFO - codeparrot_training - Step 12896: {'lr': 0.00043907759506286797, 'samples': 2476224, 'steps': 12896, 'loss/train': 1.8087027668952942} 01/29/2022 05:24:40 - INFO - codeparrot_training - Step 12897: {'lr': 0.0004390668901232291, 'samples': 2476416, 'steps': 12897, 'loss/train': 0.952428549528122} 01/29/2022 05:24:46 - INFO - codeparrot_training - Step 12898: {'lr': 0.00043905618437368766, 'samples': 2476608, 'steps': 12898, 'loss/train': 1.4090906381607056} 01/29/2022 05:24:50 - INFO - codeparrot_training - Step 12899: {'lr': 0.0004390454778142896, 'samples': 2476800, 'steps': 12899, 'loss/train': 1.5206878781318665} 01/29/2022 05:24:54 - INFO - codeparrot_training - Step 12900: {'lr': 0.00043903477044508066, 'samples': 2476992, 'steps': 12900, 'loss/train': 1.7108597159385681} 01/29/2022 05:24:59 - INFO - codeparrot_training - Step 12901: {'lr': 0.0004390240622661069, 'samples': 2477184, 'steps': 12901, 'loss/train': 2.0932517051696777} 01/29/2022 05:25:03 - INFO - codeparrot_training - Step 12902: {'lr': 0.000439013353277414, 'samples': 2477376, 'steps': 12902, 'loss/train': 2.4237492084503174} 01/29/2022 05:25:09 - INFO - codeparrot_training - Step 12903: {'lr': 0.00043900264347904796, 'samples': 2477568, 'steps': 12903, 'loss/train': 1.872925579547882} 01/29/2022 05:25:13 - INFO - codeparrot_training - Step 12904: {'lr': 0.00043899193287105456, 'samples': 2477760, 'steps': 12904, 'loss/train': 1.70545095205307} 01/29/2022 05:25:17 - INFO - codeparrot_training - Step 12905: {'lr': 0.0004389812214534798, 'samples': 2477952, 'steps': 12905, 'loss/train': 1.6259490251541138} 01/29/2022 05:25:21 - INFO - codeparrot_training - Step 12906: {'lr': 0.00043897050922636947, 'samples': 2478144, 'steps': 12906, 'loss/train': 1.5756073594093323} 01/29/2022 05:25:27 - INFO - codeparrot_training - Step 12907: {'lr': 0.00043895979618976944, 'samples': 2478336, 'steps': 12907, 'loss/train': 1.740752935409546} 01/29/2022 05:25:32 - INFO - codeparrot_training - Step 12908: {'lr': 0.00043894908234372564, 'samples': 2478528, 'steps': 12908, 'loss/train': 2.0073012113571167} 01/29/2022 05:25:36 - INFO - codeparrot_training - Step 12909: {'lr': 0.00043893836768828405, 'samples': 2478720, 'steps': 12909, 'loss/train': 2.011568605899811} 01/29/2022 05:25:40 - INFO - codeparrot_training - Step 12910: {'lr': 0.0004389276522234904, 'samples': 2478912, 'steps': 12910, 'loss/train': 1.3836204707622528} 01/29/2022 05:25:44 - INFO - codeparrot_training - Step 12911: {'lr': 0.00043891693594939077, 'samples': 2479104, 'steps': 12911, 'loss/train': 1.9428677558898926} 01/29/2022 05:25:49 - INFO - codeparrot_training - Step 12912: {'lr': 0.0004389062188660309, 'samples': 2479296, 'steps': 12912, 'loss/train': 1.8986833691596985} 01/29/2022 05:25:54 - INFO - codeparrot_training - Step 12913: {'lr': 0.00043889550097345675, 'samples': 2479488, 'steps': 12913, 'loss/train': 1.7464808821678162} 01/29/2022 05:25:58 - INFO - codeparrot_training - Step 12914: {'lr': 0.0004388847822717144, 'samples': 2479680, 'steps': 12914, 'loss/train': 1.5749309062957764} 01/29/2022 05:26:02 - INFO - codeparrot_training - Step 12915: {'lr': 0.0004388740627608495, 'samples': 2479872, 'steps': 12915, 'loss/train': 1.616429328918457} 01/29/2022 05:26:06 - INFO - codeparrot_training - Step 12916: {'lr': 0.0004388633424409081, 'samples': 2480064, 'steps': 12916, 'loss/train': 1.8283211588859558} 01/29/2022 05:26:12 - INFO - codeparrot_training - Step 12917: {'lr': 0.0004388526213119361, 'samples': 2480256, 'steps': 12917, 'loss/train': 1.759376585483551} 01/29/2022 05:26:16 - INFO - codeparrot_training - Step 12918: {'lr': 0.00043884189937397946, 'samples': 2480448, 'steps': 12918, 'loss/train': 2.059068202972412} 01/29/2022 05:26:20 - INFO - codeparrot_training - Step 12919: {'lr': 0.00043883117662708404, 'samples': 2480640, 'steps': 12919, 'loss/train': 1.0941623747348785} 01/29/2022 05:26:25 - INFO - codeparrot_training - Step 12920: {'lr': 0.0004388204530712959, 'samples': 2480832, 'steps': 12920, 'loss/train': 1.9190958738327026} 01/29/2022 05:26:29 - INFO - codeparrot_training - Step 12921: {'lr': 0.00043880972870666084, 'samples': 2481024, 'steps': 12921, 'loss/train': 1.732791244983673} 01/29/2022 05:26:35 - INFO - codeparrot_training - Step 12922: {'lr': 0.0004387990035332249, 'samples': 2481216, 'steps': 12922, 'loss/train': 1.6215873956680298} 01/29/2022 05:26:39 - INFO - codeparrot_training - Step 12923: {'lr': 0.00043878827755103404, 'samples': 2481408, 'steps': 12923, 'loss/train': 2.5581981539726257} 01/29/2022 05:26:43 - INFO - codeparrot_training - Step 12924: {'lr': 0.00043877755076013406, 'samples': 2481600, 'steps': 12924, 'loss/train': 2.5906927585601807} 01/29/2022 05:26:47 - INFO - codeparrot_training - Step 12925: {'lr': 0.00043876682316057095, 'samples': 2481792, 'steps': 12925, 'loss/train': 2.385948657989502} 01/29/2022 05:26:52 - INFO - codeparrot_training - Step 12926: {'lr': 0.0004387560947523908, 'samples': 2481984, 'steps': 12926, 'loss/train': 1.0260069072246552} 01/29/2022 05:26:57 - INFO - codeparrot_training - Step 12927: {'lr': 0.0004387453655356394, 'samples': 2482176, 'steps': 12927, 'loss/train': 1.4835203289985657} 01/29/2022 05:27:01 - INFO - codeparrot_training - Step 12928: {'lr': 0.00043873463551036284, 'samples': 2482368, 'steps': 12928, 'loss/train': 1.9926631450653076} 01/29/2022 05:27:06 - INFO - codeparrot_training - Step 12929: {'lr': 0.000438723904676607, 'samples': 2482560, 'steps': 12929, 'loss/train': 1.7171155214309692} 01/29/2022 05:27:10 - INFO - codeparrot_training - Step 12930: {'lr': 0.0004387131730344179, 'samples': 2482752, 'steps': 12930, 'loss/train': 1.7018687725067139} 01/29/2022 05:27:14 - INFO - codeparrot_training - Step 12931: {'lr': 0.00043870244058384145, 'samples': 2482944, 'steps': 12931, 'loss/train': 1.534247875213623} 01/29/2022 05:27:20 - INFO - codeparrot_training - Step 12932: {'lr': 0.0004386917073249237, 'samples': 2483136, 'steps': 12932, 'loss/train': 1.4357050359249115} 01/29/2022 05:27:25 - INFO - codeparrot_training - Step 12933: {'lr': 0.00043868097325771064, 'samples': 2483328, 'steps': 12933, 'loss/train': 2.0787559747695923} 01/29/2022 05:27:29 - INFO - codeparrot_training - Step 12934: {'lr': 0.0004386702383822482, 'samples': 2483520, 'steps': 12934, 'loss/train': 1.4824654161930084} 01/29/2022 05:27:33 - INFO - codeparrot_training - Step 12935: {'lr': 0.00043865950269858224, 'samples': 2483712, 'steps': 12935, 'loss/train': 2.0698726773262024} 01/29/2022 05:27:37 - INFO - codeparrot_training - Step 12936: {'lr': 0.000438648766206759, 'samples': 2483904, 'steps': 12936, 'loss/train': 2.0256353616714478} 01/29/2022 05:27:42 - INFO - codeparrot_training - Step 12937: {'lr': 0.0004386380289068243, 'samples': 2484096, 'steps': 12937, 'loss/train': 0.06996365077793598} 01/29/2022 05:27:47 - INFO - codeparrot_training - Step 12938: {'lr': 0.0004386272907988242, 'samples': 2484288, 'steps': 12938, 'loss/train': 1.742783546447754} 01/29/2022 05:27:51 - INFO - codeparrot_training - Step 12939: {'lr': 0.0004386165518828047, 'samples': 2484480, 'steps': 12939, 'loss/train': 1.265261024236679} 01/29/2022 05:27:55 - INFO - codeparrot_training - Step 12940: {'lr': 0.0004386058121588117, 'samples': 2484672, 'steps': 12940, 'loss/train': 2.163534164428711} 01/29/2022 05:28:00 - INFO - codeparrot_training - Step 12941: {'lr': 0.0004385950716268914, 'samples': 2484864, 'steps': 12941, 'loss/train': 1.0129565298557281} 01/29/2022 05:28:05 - INFO - codeparrot_training - Step 12942: {'lr': 0.0004385843302870896, 'samples': 2485056, 'steps': 12942, 'loss/train': 1.1182453036308289} 01/29/2022 05:28:09 - INFO - codeparrot_training - Step 12943: {'lr': 0.0004385735881394525, 'samples': 2485248, 'steps': 12943, 'loss/train': 1.353615939617157} 01/29/2022 05:28:13 - INFO - codeparrot_training - Step 12944: {'lr': 0.00043856284518402594, 'samples': 2485440, 'steps': 12944, 'loss/train': 1.6306425333023071} 01/29/2022 05:28:18 - INFO - codeparrot_training - Step 12945: {'lr': 0.00043855210142085613, 'samples': 2485632, 'steps': 12945, 'loss/train': 1.7615314722061157} 01/29/2022 05:28:22 - INFO - codeparrot_training - Step 12946: {'lr': 0.00043854135684998893, 'samples': 2485824, 'steps': 12946, 'loss/train': 2.0957570672035217} 01/29/2022 05:28:28 - INFO - codeparrot_training - Step 12947: {'lr': 0.0004385306114714704, 'samples': 2486016, 'steps': 12947, 'loss/train': 0.7317834198474884} 01/29/2022 05:28:32 - INFO - codeparrot_training - Step 12948: {'lr': 0.0004385198652853466, 'samples': 2486208, 'steps': 12948, 'loss/train': 1.2829736173152924} 01/29/2022 05:28:36 - INFO - codeparrot_training - Step 12949: {'lr': 0.00043850911829166364, 'samples': 2486400, 'steps': 12949, 'loss/train': 1.1232690811157227} 01/29/2022 05:28:41 - INFO - codeparrot_training - Step 12950: {'lr': 0.00043849837049046735, 'samples': 2486592, 'steps': 12950, 'loss/train': 1.891406536102295} 01/29/2022 05:28:45 - INFO - codeparrot_training - Step 12951: {'lr': 0.000438487621881804, 'samples': 2486784, 'steps': 12951, 'loss/train': 1.1876004338264465} 01/29/2022 05:28:51 - INFO - codeparrot_training - Step 12952: {'lr': 0.00043847687246571955, 'samples': 2486976, 'steps': 12952, 'loss/train': 1.39870485663414} 01/29/2022 05:28:55 - INFO - codeparrot_training - Step 12953: {'lr': 0.0004384661222422599, 'samples': 2487168, 'steps': 12953, 'loss/train': 1.443879872560501} 01/29/2022 05:28:59 - INFO - codeparrot_training - Step 12954: {'lr': 0.00043845537121147126, 'samples': 2487360, 'steps': 12954, 'loss/train': 0.46540166437625885} 01/29/2022 05:29:04 - INFO - codeparrot_training - Step 12955: {'lr': 0.00043844461937339976, 'samples': 2487552, 'steps': 12955, 'loss/train': 1.208257108926773} 01/29/2022 05:29:08 - INFO - codeparrot_training - Step 12956: {'lr': 0.00043843386672809127, 'samples': 2487744, 'steps': 12956, 'loss/train': 1.38129723072052} 01/29/2022 05:29:12 - INFO - codeparrot_training - Step 12957: {'lr': 0.00043842311327559194, 'samples': 2487936, 'steps': 12957, 'loss/train': 1.8249843120574951} 01/29/2022 05:29:18 - INFO - codeparrot_training - Step 12958: {'lr': 0.0004384123590159478, 'samples': 2488128, 'steps': 12958, 'loss/train': 1.4840508699417114} 01/29/2022 05:29:22 - INFO - codeparrot_training - Step 12959: {'lr': 0.000438401603949205, 'samples': 2488320, 'steps': 12959, 'loss/train': 0.5969503819942474} 01/29/2022 05:29:26 - INFO - codeparrot_training - Step 12960: {'lr': 0.0004383908480754095, 'samples': 2488512, 'steps': 12960, 'loss/train': 1.6024695038795471} 01/29/2022 05:29:30 - INFO - codeparrot_training - Step 12961: {'lr': 0.0004383800913946074, 'samples': 2488704, 'steps': 12961, 'loss/train': 2.0671216249465942} 01/29/2022 05:29:35 - INFO - codeparrot_training - Step 12962: {'lr': 0.00043836933390684486, 'samples': 2488896, 'steps': 12962, 'loss/train': 0.3530043885111809} 01/29/2022 05:29:40 - INFO - codeparrot_training - Step 12963: {'lr': 0.0004383585756121679, 'samples': 2489088, 'steps': 12963, 'loss/train': 1.0706360936164856} 01/29/2022 05:29:44 - INFO - codeparrot_training - Step 12964: {'lr': 0.00043834781651062263, 'samples': 2489280, 'steps': 12964, 'loss/train': 1.0999100804328918} 01/29/2022 05:29:48 - INFO - codeparrot_training - Step 12965: {'lr': 0.00043833705660225507, 'samples': 2489472, 'steps': 12965, 'loss/train': 1.6071208119392395} 01/29/2022 05:29:53 - INFO - codeparrot_training - Step 12966: {'lr': 0.0004383262958871114, 'samples': 2489664, 'steps': 12966, 'loss/train': 3.6105741262435913} 01/29/2022 05:29:57 - INFO - codeparrot_training - Step 12967: {'lr': 0.0004383155343652377, 'samples': 2489856, 'steps': 12967, 'loss/train': 1.891556203365326} 01/29/2022 05:30:03 - INFO - codeparrot_training - Step 12968: {'lr': 0.00043830477203668, 'samples': 2490048, 'steps': 12968, 'loss/train': 3.024282217025757} 01/29/2022 05:30:07 - INFO - codeparrot_training - Step 12969: {'lr': 0.00043829400890148446, 'samples': 2490240, 'steps': 12969, 'loss/train': 1.3367811441421509} 01/29/2022 05:30:11 - INFO - codeparrot_training - Step 12970: {'lr': 0.0004382832449596972, 'samples': 2490432, 'steps': 12970, 'loss/train': 2.1319207549095154} 01/29/2022 05:30:16 - INFO - codeparrot_training - Step 12971: {'lr': 0.0004382724802113643, 'samples': 2490624, 'steps': 12971, 'loss/train': 1.974151074886322} 01/29/2022 05:30:20 - INFO - codeparrot_training - Step 12972: {'lr': 0.0004382617146565319, 'samples': 2490816, 'steps': 12972, 'loss/train': 1.4646756649017334} 01/29/2022 05:30:25 - INFO - codeparrot_training - Step 12973: {'lr': 0.00043825094829524604, 'samples': 2491008, 'steps': 12973, 'loss/train': 1.299886554479599} 01/29/2022 05:30:30 - INFO - codeparrot_training - Step 12974: {'lr': 0.0004382401811275529, 'samples': 2491200, 'steps': 12974, 'loss/train': 1.8107596635818481} 01/29/2022 05:30:34 - INFO - codeparrot_training - Step 12975: {'lr': 0.0004382294131534986, 'samples': 2491392, 'steps': 12975, 'loss/train': 1.9447846412658691} 01/29/2022 05:30:38 - INFO - codeparrot_training - Step 12976: {'lr': 0.00043821864437312933, 'samples': 2491584, 'steps': 12976, 'loss/train': 1.8752279877662659} 01/29/2022 05:30:42 - INFO - codeparrot_training - Step 12977: {'lr': 0.00043820787478649105, 'samples': 2491776, 'steps': 12977, 'loss/train': 1.859725534915924} 01/29/2022 05:30:48 - INFO - codeparrot_training - Step 12978: {'lr': 0.00043819710439363, 'samples': 2491968, 'steps': 12978, 'loss/train': 2.4799144864082336} 01/29/2022 05:30:52 - INFO - codeparrot_training - Step 12979: {'lr': 0.00043818633319459244, 'samples': 2492160, 'steps': 12979, 'loss/train': 1.7338528633117676} 01/29/2022 05:30:56 - INFO - codeparrot_training - Step 12980: {'lr': 0.00043817556118942426, 'samples': 2492352, 'steps': 12980, 'loss/train': 1.5468412041664124} 01/29/2022 05:31:00 - INFO - codeparrot_training - Step 12981: {'lr': 0.00043816478837817183, 'samples': 2492544, 'steps': 12981, 'loss/train': 3.1006211042404175} 01/29/2022 05:31:05 - INFO - codeparrot_training - Step 12982: {'lr': 0.0004381540147608811, 'samples': 2492736, 'steps': 12982, 'loss/train': 2.0035223364830017} 01/29/2022 05:31:11 - INFO - codeparrot_training - Step 12983: {'lr': 0.00043814324033759834, 'samples': 2492928, 'steps': 12983, 'loss/train': 1.6350395679473877} 01/29/2022 05:31:15 - INFO - codeparrot_training - Step 12984: {'lr': 0.0004381324651083697, 'samples': 2493120, 'steps': 12984, 'loss/train': 1.7434197664260864} 01/29/2022 05:31:19 - INFO - codeparrot_training - Step 12985: {'lr': 0.00043812168907324137, 'samples': 2493312, 'steps': 12985, 'loss/train': 1.9488485455513} 01/29/2022 05:31:23 - INFO - codeparrot_training - Step 12986: {'lr': 0.0004381109122322594, 'samples': 2493504, 'steps': 12986, 'loss/train': 2.2148770093917847} 01/29/2022 05:31:28 - INFO - codeparrot_training - Step 12987: {'lr': 0.00043810013458547007, 'samples': 2493696, 'steps': 12987, 'loss/train': 1.396326631307602} 01/29/2022 05:31:33 - INFO - codeparrot_training - Step 12988: {'lr': 0.00043808935613291934, 'samples': 2493888, 'steps': 12988, 'loss/train': 1.7412963509559631} 01/29/2022 05:31:37 - INFO - codeparrot_training - Step 12989: {'lr': 0.0004380785768746537, 'samples': 2494080, 'steps': 12989, 'loss/train': 1.5515491962432861} 01/29/2022 05:31:42 - INFO - codeparrot_training - Step 12990: {'lr': 0.00043806779681071907, 'samples': 2494272, 'steps': 12990, 'loss/train': 1.2723155915737152} 01/29/2022 05:31:46 - INFO - codeparrot_training - Step 12991: {'lr': 0.00043805701594116175, 'samples': 2494464, 'steps': 12991, 'loss/train': 2.0996480584144592} 01/29/2022 05:31:50 - INFO - codeparrot_training - Step 12992: {'lr': 0.00043804623426602784, 'samples': 2494656, 'steps': 12992, 'loss/train': 2.5178028345108032} 01/29/2022 05:31:56 - INFO - codeparrot_training - Step 12993: {'lr': 0.00043803545178536365, 'samples': 2494848, 'steps': 12993, 'loss/train': 0.8757471442222595} 01/29/2022 05:32:00 - INFO - codeparrot_training - Step 12994: {'lr': 0.00043802466849921526, 'samples': 2495040, 'steps': 12994, 'loss/train': 2.1554262042045593} 01/29/2022 05:32:04 - INFO - codeparrot_training - Step 12995: {'lr': 0.0004380138844076289, 'samples': 2495232, 'steps': 12995, 'loss/train': 1.6584972739219666} 01/29/2022 05:32:09 - INFO - codeparrot_training - Step 12996: {'lr': 0.00043800309951065076, 'samples': 2495424, 'steps': 12996, 'loss/train': 1.9471798539161682} 01/29/2022 05:32:13 - INFO - codeparrot_training - Step 12997: {'lr': 0.000437992313808327, 'samples': 2495616, 'steps': 12997, 'loss/train': 0.9807632267475128} 01/29/2022 05:32:18 - INFO - codeparrot_training - Step 12998: {'lr': 0.0004379815273007039, 'samples': 2495808, 'steps': 12998, 'loss/train': 2.3379077911376953} 01/29/2022 05:32:22 - INFO - codeparrot_training - Step 12999: {'lr': 0.0004379707399878276, 'samples': 2496000, 'steps': 12999, 'loss/train': 1.5460837483406067} 01/29/2022 05:32:27 - INFO - codeparrot_training - Step 13000: {'lr': 0.00043795995186974435, 'samples': 2496192, 'steps': 13000, 'loss/train': 2.351653218269348} 01/29/2022 05:32:31 - INFO - codeparrot_training - Step 13001: {'lr': 0.0004379491629465004, 'samples': 2496384, 'steps': 13001, 'loss/train': 0.9831352829933167} 01/29/2022 05:32:35 - INFO - codeparrot_training - Step 13002: {'lr': 0.00043793837321814185, 'samples': 2496576, 'steps': 13002, 'loss/train': 2.020948648452759} 01/29/2022 05:32:40 - INFO - codeparrot_training - Step 13003: {'lr': 0.000437927582684715, 'samples': 2496768, 'steps': 13003, 'loss/train': 1.897554337978363} 01/29/2022 05:32:45 - INFO - codeparrot_training - Step 13004: {'lr': 0.0004379167913462661, 'samples': 2496960, 'steps': 13004, 'loss/train': 1.5257502794265747} 01/29/2022 05:32:49 - INFO - codeparrot_training - Step 13005: {'lr': 0.0004379059992028412, 'samples': 2497152, 'steps': 13005, 'loss/train': 1.9804641008377075} 01/29/2022 05:32:53 - INFO - codeparrot_training - Step 13006: {'lr': 0.00043789520625448685, 'samples': 2497344, 'steps': 13006, 'loss/train': 0.6112655997276306} 01/29/2022 05:32:58 - INFO - codeparrot_training - Step 13007: {'lr': 0.000437884412501249, 'samples': 2497536, 'steps': 13007, 'loss/train': 0.25235798209905624} 01/29/2022 05:33:03 - INFO - codeparrot_training - Step 13008: {'lr': 0.00043787361794317403, 'samples': 2497728, 'steps': 13008, 'loss/train': 1.502852976322174} 01/29/2022 05:33:07 - INFO - codeparrot_training - Step 13009: {'lr': 0.0004378628225803081, 'samples': 2497920, 'steps': 13009, 'loss/train': 0.5894414931535721} 01/29/2022 05:33:11 - INFO - codeparrot_training - Step 13010: {'lr': 0.0004378520264126975, 'samples': 2498112, 'steps': 13010, 'loss/train': 2.139915704727173} 01/29/2022 05:33:16 - INFO - codeparrot_training - Step 13011: {'lr': 0.0004378412294403885, 'samples': 2498304, 'steps': 13011, 'loss/train': 1.74457186460495} 01/29/2022 05:33:20 - INFO - codeparrot_training - Step 13012: {'lr': 0.0004378304316634273, 'samples': 2498496, 'steps': 13012, 'loss/train': 0.9771181046962738} 01/29/2022 05:33:26 - INFO - codeparrot_training - Step 13013: {'lr': 0.0004378196330818602, 'samples': 2498688, 'steps': 13013, 'loss/train': 2.328189432621002} 01/29/2022 05:33:31 - INFO - codeparrot_training - Step 13014: {'lr': 0.00043780883369573336, 'samples': 2498880, 'steps': 13014, 'loss/train': 2.0527644753456116} 01/29/2022 05:33:35 - INFO - codeparrot_training - Step 13015: {'lr': 0.00043779803350509316, 'samples': 2499072, 'steps': 13015, 'loss/train': 1.4538741409778595} 01/29/2022 05:33:39 - INFO - codeparrot_training - Step 13016: {'lr': 0.0004377872325099858, 'samples': 2499264, 'steps': 13016, 'loss/train': 1.5298057794570923} 01/29/2022 05:33:44 - INFO - codeparrot_training - Step 13017: {'lr': 0.0004377764307104576, 'samples': 2499456, 'steps': 13017, 'loss/train': 0.6866713464260101} 01/29/2022 05:33:49 - INFO - codeparrot_training - Step 13018: {'lr': 0.00043776562810655473, 'samples': 2499648, 'steps': 13018, 'loss/train': 2.2802236676216125} 01/29/2022 05:33:53 - INFO - codeparrot_training - Step 13019: {'lr': 0.0004377548246983236, 'samples': 2499840, 'steps': 13019, 'loss/train': 2.0050780177116394} 01/29/2022 05:33:57 - INFO - codeparrot_training - Step 13020: {'lr': 0.0004377440204858104, 'samples': 2500032, 'steps': 13020, 'loss/train': 0.6506450325250626} 01/29/2022 05:34:02 - INFO - codeparrot_training - Step 13021: {'lr': 0.0004377332154690614, 'samples': 2500224, 'steps': 13021, 'loss/train': 1.841631531715393} 01/29/2022 05:34:06 - INFO - codeparrot_training - Step 13022: {'lr': 0.0004377224096481229, 'samples': 2500416, 'steps': 13022, 'loss/train': 2.0582610368728638} 01/29/2022 05:34:11 - INFO - codeparrot_training - Step 13023: {'lr': 0.0004377116030230413, 'samples': 2500608, 'steps': 13023, 'loss/train': 1.928077518939972} 01/29/2022 05:34:15 - INFO - codeparrot_training - Step 13024: {'lr': 0.0004377007955938628, 'samples': 2500800, 'steps': 13024, 'loss/train': 1.798496961593628} 01/29/2022 05:34:20 - INFO - codeparrot_training - Step 13025: {'lr': 0.0004376899873606336, 'samples': 2500992, 'steps': 13025, 'loss/train': 1.9179210662841797} 01/29/2022 05:34:24 - INFO - codeparrot_training - Step 13026: {'lr': 0.0004376791783234001, 'samples': 2501184, 'steps': 13026, 'loss/train': 2.0095311999320984} 01/29/2022 05:34:28 - INFO - codeparrot_training - Step 13027: {'lr': 0.0004376683684822086, 'samples': 2501376, 'steps': 13027, 'loss/train': 1.6027103662490845} 01/29/2022 05:34:34 - INFO - codeparrot_training - Step 13028: {'lr': 0.0004376575578371055, 'samples': 2501568, 'steps': 13028, 'loss/train': 1.792811930179596} 01/29/2022 05:34:38 - INFO - codeparrot_training - Step 13029: {'lr': 0.0004376467463881369, 'samples': 2501760, 'steps': 13029, 'loss/train': 2.0119518041610718} 01/29/2022 05:34:43 - INFO - codeparrot_training - Step 13030: {'lr': 0.0004376359341353492, 'samples': 2501952, 'steps': 13030, 'loss/train': 1.39010488986969} 01/29/2022 05:34:47 - INFO - codeparrot_training - Step 13031: {'lr': 0.00043762512107878884, 'samples': 2502144, 'steps': 13031, 'loss/train': 2.253875255584717} 01/29/2022 05:34:51 - INFO - codeparrot_training - Step 13032: {'lr': 0.00043761430721850206, 'samples': 2502336, 'steps': 13032, 'loss/train': 1.2039504647254944} 01/29/2022 05:34:56 - INFO - codeparrot_training - Step 13033: {'lr': 0.0004376034925545351, 'samples': 2502528, 'steps': 13033, 'loss/train': 1.9867773056030273} 01/29/2022 05:35:01 - INFO - codeparrot_training - Step 13034: {'lr': 0.0004375926770869343, 'samples': 2502720, 'steps': 13034, 'loss/train': 2.3981913328170776} 01/29/2022 05:35:05 - INFO - codeparrot_training - Step 13035: {'lr': 0.00043758186081574614, 'samples': 2502912, 'steps': 13035, 'loss/train': 0.45952239632606506} 01/29/2022 05:35:09 - INFO - codeparrot_training - Step 13036: {'lr': 0.00043757104374101677, 'samples': 2503104, 'steps': 13036, 'loss/train': 1.8566338419914246} 01/29/2022 05:35:13 - INFO - codeparrot_training - Step 13037: {'lr': 0.00043756022586279264, 'samples': 2503296, 'steps': 13037, 'loss/train': 0.830915093421936} 01/29/2022 05:35:19 - INFO - codeparrot_training - Step 13038: {'lr': 0.00043754940718112, 'samples': 2503488, 'steps': 13038, 'loss/train': 1.0003852844238281} 01/29/2022 05:35:24 - INFO - codeparrot_training - Step 13039: {'lr': 0.0004375385876960454, 'samples': 2503680, 'steps': 13039, 'loss/train': 2.271778643131256} 01/29/2022 05:35:28 - INFO - codeparrot_training - Step 13040: {'lr': 0.0004375277674076149, 'samples': 2503872, 'steps': 13040, 'loss/train': 1.9277476072311401} 01/29/2022 05:35:32 - INFO - codeparrot_training - Step 13041: {'lr': 0.00043751694631587504, 'samples': 2504064, 'steps': 13041, 'loss/train': 2.1388145685195923} 01/29/2022 05:35:36 - INFO - codeparrot_training - Step 13042: {'lr': 0.00043750612442087215, 'samples': 2504256, 'steps': 13042, 'loss/train': 1.3576493561267853} 01/29/2022 05:35:42 - INFO - codeparrot_training - Step 13043: {'lr': 0.0004374953017226525, 'samples': 2504448, 'steps': 13043, 'loss/train': 1.7837828993797302} 01/29/2022 05:35:46 - INFO - codeparrot_training - Step 13044: {'lr': 0.0004374844782212626, 'samples': 2504640, 'steps': 13044, 'loss/train': 2.395980477333069} 01/29/2022 05:35:50 - INFO - codeparrot_training - Step 13045: {'lr': 0.0004374736539167487, 'samples': 2504832, 'steps': 13045, 'loss/train': 1.2809646427631378} 01/29/2022 05:35:55 - INFO - codeparrot_training - Step 13046: {'lr': 0.0004374628288091571, 'samples': 2505024, 'steps': 13046, 'loss/train': 0.6828116923570633} 01/29/2022 05:35:59 - INFO - codeparrot_training - Step 13047: {'lr': 0.0004374520028985344, 'samples': 2505216, 'steps': 13047, 'loss/train': 1.6075192093849182} 01/29/2022 05:36:04 - INFO - codeparrot_training - Step 13048: {'lr': 0.0004374411761849268, 'samples': 2505408, 'steps': 13048, 'loss/train': 2.437318503856659} 01/29/2022 05:36:08 - INFO - codeparrot_training - Step 13049: {'lr': 0.0004374303486683807, 'samples': 2505600, 'steps': 13049, 'loss/train': 0.6258412152528763} 01/29/2022 05:36:13 - INFO - codeparrot_training - Step 13050: {'lr': 0.0004374195203489425, 'samples': 2505792, 'steps': 13050, 'loss/train': 1.316127598285675} 01/29/2022 05:36:17 - INFO - codeparrot_training - Step 13051: {'lr': 0.0004374086912266586, 'samples': 2505984, 'steps': 13051, 'loss/train': 2.1805298924446106} 01/29/2022 05:36:21 - INFO - codeparrot_training - Step 13052: {'lr': 0.0004373978613015753, 'samples': 2506176, 'steps': 13052, 'loss/train': 1.7382504343986511} 01/29/2022 05:36:27 - INFO - codeparrot_training - Step 13053: {'lr': 0.0004373870305737392, 'samples': 2506368, 'steps': 13053, 'loss/train': 1.8975107073783875} 01/29/2022 05:36:31 - INFO - codeparrot_training - Step 13054: {'lr': 0.00043737619904319654, 'samples': 2506560, 'steps': 13054, 'loss/train': 1.5273789167404175} 01/29/2022 05:36:36 - INFO - codeparrot_training - Step 13055: {'lr': 0.0004373653667099937, 'samples': 2506752, 'steps': 13055, 'loss/train': 1.6835993528366089} 01/29/2022 05:36:40 - INFO - codeparrot_training - Step 13056: {'lr': 0.00043735453357417707, 'samples': 2506944, 'steps': 13056, 'loss/train': 2.249004900455475} 01/29/2022 05:36:44 - INFO - codeparrot_training - Step 13057: {'lr': 0.00043734369963579323, 'samples': 2507136, 'steps': 13057, 'loss/train': 2.0130441784858704} 01/29/2022 05:36:49 - INFO - codeparrot_training - Step 13058: {'lr': 0.0004373328648948884, 'samples': 2507328, 'steps': 13058, 'loss/train': 1.075301617383957} 01/29/2022 05:36:54 - INFO - codeparrot_training - Step 13059: {'lr': 0.0004373220293515091, 'samples': 2507520, 'steps': 13059, 'loss/train': 1.897719383239746} 01/29/2022 05:36:58 - INFO - codeparrot_training - Step 13060: {'lr': 0.00043731119300570166, 'samples': 2507712, 'steps': 13060, 'loss/train': 1.8411943316459656} 01/29/2022 05:37:02 - INFO - codeparrot_training - Step 13061: {'lr': 0.0004373003558575126, 'samples': 2507904, 'steps': 13061, 'loss/train': 2.0014278888702393} 01/29/2022 05:37:06 - INFO - codeparrot_training - Step 13062: {'lr': 0.00043728951790698823, 'samples': 2508096, 'steps': 13062, 'loss/train': 2.023819863796234} 01/29/2022 05:37:12 - INFO - codeparrot_training - Step 13063: {'lr': 0.00043727867915417505, 'samples': 2508288, 'steps': 13063, 'loss/train': 1.1814191937446594} 01/29/2022 05:37:16 - INFO - codeparrot_training - Step 13064: {'lr': 0.00043726783959911953, 'samples': 2508480, 'steps': 13064, 'loss/train': 0.7646805346012115} 01/29/2022 05:37:21 - INFO - codeparrot_training - Step 13065: {'lr': 0.00043725699924186803, 'samples': 2508672, 'steps': 13065, 'loss/train': 1.9173688888549805} 01/29/2022 05:37:25 - INFO - codeparrot_training - Step 13066: {'lr': 0.00043724615808246695, 'samples': 2508864, 'steps': 13066, 'loss/train': 1.2043429613113403} 01/29/2022 05:37:29 - INFO - codeparrot_training - Step 13067: {'lr': 0.0004372353161209628, 'samples': 2509056, 'steps': 13067, 'loss/train': 2.2138266563415527} 01/29/2022 05:37:34 - INFO - codeparrot_training - Step 13068: {'lr': 0.000437224473357402, 'samples': 2509248, 'steps': 13068, 'loss/train': 1.8217104077339172} 01/29/2022 05:37:39 - INFO - codeparrot_training - Step 13069: {'lr': 0.0004372136297918311, 'samples': 2509440, 'steps': 13069, 'loss/train': 1.6239157319068909} 01/29/2022 05:37:43 - INFO - codeparrot_training - Step 13070: {'lr': 0.0004372027854242964, 'samples': 2509632, 'steps': 13070, 'loss/train': 2.328347146511078} 01/29/2022 05:37:47 - INFO - codeparrot_training - Step 13071: {'lr': 0.0004371919402548444, 'samples': 2509824, 'steps': 13071, 'loss/train': 2.081751823425293} 01/29/2022 05:37:51 - INFO - codeparrot_training - Step 13072: {'lr': 0.00043718109428352156, 'samples': 2510016, 'steps': 13072, 'loss/train': 1.7362266182899475} 01/29/2022 05:37:57 - INFO - codeparrot_training - Step 13073: {'lr': 0.00043717024751037436, 'samples': 2510208, 'steps': 13073, 'loss/train': 1.3518016934394836} 01/29/2022 05:38:02 - INFO - codeparrot_training - Step 13074: {'lr': 0.0004371593999354493, 'samples': 2510400, 'steps': 13074, 'loss/train': 1.185422122478485} 01/29/2022 05:38:06 - INFO - codeparrot_training - Step 13075: {'lr': 0.0004371485515587927, 'samples': 2510592, 'steps': 13075, 'loss/train': 1.2423159778118134} 01/29/2022 05:38:10 - INFO - codeparrot_training - Step 13076: {'lr': 0.0004371377023804512, 'samples': 2510784, 'steps': 13076, 'loss/train': 1.671271562576294} 01/29/2022 05:38:14 - INFO - codeparrot_training - Step 13077: {'lr': 0.00043712685240047125, 'samples': 2510976, 'steps': 13077, 'loss/train': 1.6936776638031006} 01/29/2022 05:38:20 - INFO - codeparrot_training - Step 13078: {'lr': 0.00043711600161889917, 'samples': 2511168, 'steps': 13078, 'loss/train': 0.3076658472418785} 01/29/2022 05:38:24 - INFO - codeparrot_training - Step 13079: {'lr': 0.0004371051500357816, 'samples': 2511360, 'steps': 13079, 'loss/train': 0.8947497010231018} 01/29/2022 05:38:29 - INFO - codeparrot_training - Step 13080: {'lr': 0.000437094297651165, 'samples': 2511552, 'steps': 13080, 'loss/train': 0.8053799271583557} 01/29/2022 05:38:33 - INFO - codeparrot_training - Step 13081: {'lr': 0.00043708344446509586, 'samples': 2511744, 'steps': 13081, 'loss/train': 1.2129490971565247} 01/29/2022 05:38:37 - INFO - codeparrot_training - Step 13082: {'lr': 0.0004370725904776206, 'samples': 2511936, 'steps': 13082, 'loss/train': 1.633419692516327} 01/29/2022 05:38:42 - INFO - codeparrot_training - Step 13083: {'lr': 0.0004370617356887858, 'samples': 2512128, 'steps': 13083, 'loss/train': 0.6016107648611069} 01/29/2022 05:38:47 - INFO - codeparrot_training - Step 13084: {'lr': 0.00043705088009863793, 'samples': 2512320, 'steps': 13084, 'loss/train': 1.8295596241950989} 01/29/2022 05:38:51 - INFO - codeparrot_training - Step 13085: {'lr': 0.0004370400237072234, 'samples': 2512512, 'steps': 13085, 'loss/train': 1.5207785367965698} 01/29/2022 05:38:55 - INFO - codeparrot_training - Step 13086: {'lr': 0.0004370291665145889, 'samples': 2512704, 'steps': 13086, 'loss/train': 1.4781053066253662} 01/29/2022 05:38:59 - INFO - codeparrot_training - Step 13087: {'lr': 0.00043701830852078076, 'samples': 2512896, 'steps': 13087, 'loss/train': 1.3622003495693207} 01/29/2022 05:39:05 - INFO - codeparrot_training - Step 13088: {'lr': 0.0004370074497258456, 'samples': 2513088, 'steps': 13088, 'loss/train': 1.8143336176872253} 01/29/2022 05:39:10 - INFO - codeparrot_training - Step 13089: {'lr': 0.00043699659012983, 'samples': 2513280, 'steps': 13089, 'loss/train': 1.5323572754859924} 01/29/2022 05:39:14 - INFO - codeparrot_training - Step 13090: {'lr': 0.00043698572973278026, 'samples': 2513472, 'steps': 13090, 'loss/train': 1.233353465795517} 01/29/2022 05:39:18 - INFO - codeparrot_training - Step 13091: {'lr': 0.0004369748685347431, 'samples': 2513664, 'steps': 13091, 'loss/train': 2.0912452340126038} 01/29/2022 05:39:22 - INFO - codeparrot_training - Step 13092: {'lr': 0.00043696400653576496, 'samples': 2513856, 'steps': 13092, 'loss/train': 1.9289411902427673} 01/29/2022 05:39:28 - INFO - codeparrot_training - Step 13093: {'lr': 0.00043695314373589234, 'samples': 2514048, 'steps': 13093, 'loss/train': 1.28072789311409} 01/29/2022 05:39:32 - INFO - codeparrot_training - Step 13094: {'lr': 0.00043694228013517185, 'samples': 2514240, 'steps': 13094, 'loss/train': 2.2056578993797302} 01/29/2022 05:39:36 - INFO - codeparrot_training - Step 13095: {'lr': 0.00043693141573365003, 'samples': 2514432, 'steps': 13095, 'loss/train': 1.4330829977989197} 01/29/2022 05:39:40 - INFO - codeparrot_training - Step 13096: {'lr': 0.0004369205505313733, 'samples': 2514624, 'steps': 13096, 'loss/train': 0.8998064696788788} 01/29/2022 05:39:45 - INFO - codeparrot_training - Step 13097: {'lr': 0.0004369096845283883, 'samples': 2514816, 'steps': 13097, 'loss/train': 1.6600529551506042} 01/29/2022 05:39:51 - INFO - codeparrot_training - Step 13098: {'lr': 0.0004368988177247416, 'samples': 2515008, 'steps': 13098, 'loss/train': 1.8541236519813538} 01/29/2022 05:39:55 - INFO - codeparrot_training - Step 13099: {'lr': 0.00043688795012047975, 'samples': 2515200, 'steps': 13099, 'loss/train': 1.5766233801841736} 01/29/2022 05:39:59 - INFO - codeparrot_training - Step 13100: {'lr': 0.00043687708171564923, 'samples': 2515392, 'steps': 13100, 'loss/train': 1.4939066469669342} 01/29/2022 05:40:03 - INFO - codeparrot_training - Step 13101: {'lr': 0.0004368662125102966, 'samples': 2515584, 'steps': 13101, 'loss/train': 2.0700899362564087} 01/29/2022 05:40:08 - INFO - codeparrot_training - Step 13102: {'lr': 0.00043685534250446846, 'samples': 2515776, 'steps': 13102, 'loss/train': 1.1621447503566742} 01/29/2022 05:40:13 - INFO - codeparrot_training - Step 13103: {'lr': 0.0004368444716982114, 'samples': 2515968, 'steps': 13103, 'loss/train': 2.2013726234436035} 01/29/2022 05:40:18 - INFO - codeparrot_training - Step 13104: {'lr': 0.0004368336000915719, 'samples': 2516160, 'steps': 13104, 'loss/train': 1.3577578961849213} 01/29/2022 05:40:22 - INFO - codeparrot_training - Step 13105: {'lr': 0.0004368227276845966, 'samples': 2516352, 'steps': 13105, 'loss/train': 2.2510133385658264} 01/29/2022 05:40:26 - INFO - codeparrot_training - Step 13106: {'lr': 0.0004368118544773321, 'samples': 2516544, 'steps': 13106, 'loss/train': 1.9667088389396667} 01/29/2022 05:40:30 - INFO - codeparrot_training - Step 13107: {'lr': 0.00043680098046982495, 'samples': 2516736, 'steps': 13107, 'loss/train': 1.741389513015747} 01/29/2022 05:40:35 - INFO - codeparrot_training - Step 13108: {'lr': 0.00043679010566212163, 'samples': 2516928, 'steps': 13108, 'loss/train': 1.8332594633102417} 01/29/2022 05:40:41 - INFO - codeparrot_training - Step 13109: {'lr': 0.0004367792300542689, 'samples': 2517120, 'steps': 13109, 'loss/train': 1.520705223083496} 01/29/2022 05:40:45 - INFO - codeparrot_training - Step 13110: {'lr': 0.00043676835364631316, 'samples': 2517312, 'steps': 13110, 'loss/train': 1.9380571246147156} 01/29/2022 05:40:49 - INFO - codeparrot_training - Step 13111: {'lr': 0.00043675747643830116, 'samples': 2517504, 'steps': 13111, 'loss/train': 1.448473870754242} 01/29/2022 05:40:53 - INFO - codeparrot_training - Step 13112: {'lr': 0.0004367465984302794, 'samples': 2517696, 'steps': 13112, 'loss/train': 1.438607007265091} 01/29/2022 05:40:58 - INFO - codeparrot_training - Step 13113: {'lr': 0.0004367357196222946, 'samples': 2517888, 'steps': 13113, 'loss/train': 1.3474333584308624} 01/29/2022 05:41:03 - INFO - codeparrot_training - Step 13114: {'lr': 0.00043672484001439316, 'samples': 2518080, 'steps': 13114, 'loss/train': 1.7193394303321838} 01/29/2022 05:41:08 - INFO - codeparrot_training - Step 13115: {'lr': 0.00043671395960662184, 'samples': 2518272, 'steps': 13115, 'loss/train': 0.8537970185279846} 01/29/2022 05:41:12 - INFO - codeparrot_training - Step 13116: {'lr': 0.0004367030783990272, 'samples': 2518464, 'steps': 13116, 'loss/train': 1.878079354763031} 01/29/2022 05:41:16 - INFO - codeparrot_training - Step 13117: {'lr': 0.0004366921963916559, 'samples': 2518656, 'steps': 13117, 'loss/train': 2.1541441082954407} 01/29/2022 05:41:20 - INFO - codeparrot_training - Step 13118: {'lr': 0.0004366813135845545, 'samples': 2518848, 'steps': 13118, 'loss/train': 1.0892657339572906} 01/29/2022 05:41:26 - INFO - codeparrot_training - Step 13119: {'lr': 0.00043667042997776965, 'samples': 2519040, 'steps': 13119, 'loss/train': 1.8796337842941284} 01/29/2022 05:41:30 - INFO - codeparrot_training - Step 13120: {'lr': 0.00043665954557134786, 'samples': 2519232, 'steps': 13120, 'loss/train': 1.4121294915676117} 01/29/2022 05:41:34 - INFO - codeparrot_training - Step 13121: {'lr': 0.0004366486603653359, 'samples': 2519424, 'steps': 13121, 'loss/train': 1.8369038701057434} 01/29/2022 05:41:39 - INFO - codeparrot_training - Step 13122: {'lr': 0.00043663777435978037, 'samples': 2519616, 'steps': 13122, 'loss/train': 1.5863664150238037} 01/29/2022 05:41:43 - INFO - codeparrot_training - Step 13123: {'lr': 0.0004366268875547278, 'samples': 2519808, 'steps': 13123, 'loss/train': 0.6524147987365723} 01/29/2022 05:41:48 - INFO - codeparrot_training - Step 13124: {'lr': 0.000436615999950225, 'samples': 2520000, 'steps': 13124, 'loss/train': 1.5731810331344604} 01/29/2022 05:41:52 - INFO - codeparrot_training - Step 13125: {'lr': 0.0004366051115463184, 'samples': 2520192, 'steps': 13125, 'loss/train': 1.9857820272445679} 01/29/2022 05:41:57 - INFO - codeparrot_training - Step 13126: {'lr': 0.0004365942223430549, 'samples': 2520384, 'steps': 13126, 'loss/train': 1.4861505925655365} 01/29/2022 05:42:01 - INFO - codeparrot_training - Step 13127: {'lr': 0.0004365833323404809, 'samples': 2520576, 'steps': 13127, 'loss/train': 1.4840426445007324} 01/29/2022 05:42:05 - INFO - codeparrot_training - Step 13128: {'lr': 0.0004365724415386432, 'samples': 2520768, 'steps': 13128, 'loss/train': 1.957548201084137} 01/29/2022 05:42:11 - INFO - codeparrot_training - Step 13129: {'lr': 0.0004365615499375884, 'samples': 2520960, 'steps': 13129, 'loss/train': 0.3828115314245224} 01/29/2022 05:42:15 - INFO - codeparrot_training - Step 13130: {'lr': 0.0004365506575373631, 'samples': 2521152, 'steps': 13130, 'loss/train': 1.8292443752288818} 01/29/2022 05:42:20 - INFO - codeparrot_training - Step 13131: {'lr': 0.0004365397643380141, 'samples': 2521344, 'steps': 13131, 'loss/train': 0.32390105724334717} 01/29/2022 05:42:24 - INFO - codeparrot_training - Step 13132: {'lr': 0.000436528870339588, 'samples': 2521536, 'steps': 13132, 'loss/train': 1.739885151386261} 01/29/2022 05:42:28 - INFO - codeparrot_training - Step 13133: {'lr': 0.0004365179755421314, 'samples': 2521728, 'steps': 13133, 'loss/train': 1.1433919966220856} 01/29/2022 05:42:33 - INFO - codeparrot_training - Step 13134: {'lr': 0.00043650707994569095, 'samples': 2521920, 'steps': 13134, 'loss/train': 2.039856433868408} 01/29/2022 05:42:38 - INFO - codeparrot_training - Step 13135: {'lr': 0.0004364961835503135, 'samples': 2522112, 'steps': 13135, 'loss/train': 1.1098886132240295} 01/29/2022 05:42:42 - INFO - codeparrot_training - Step 13136: {'lr': 0.00043648528635604556, 'samples': 2522304, 'steps': 13136, 'loss/train': 1.5333897471427917} 01/29/2022 05:42:46 - INFO - codeparrot_training - Step 13137: {'lr': 0.00043647438836293383, 'samples': 2522496, 'steps': 13137, 'loss/train': 1.3783250749111176} 01/29/2022 05:42:51 - INFO - codeparrot_training - Step 13138: {'lr': 0.0004364634895710251, 'samples': 2522688, 'steps': 13138, 'loss/train': 1.9339508414268494} 01/29/2022 05:42:56 - INFO - codeparrot_training - Step 13139: {'lr': 0.000436452589980366, 'samples': 2522880, 'steps': 13139, 'loss/train': 1.4477554857730865} 01/29/2022 05:43:00 - INFO - codeparrot_training - Step 13140: {'lr': 0.00043644168959100315, 'samples': 2523072, 'steps': 13140, 'loss/train': 1.3679928481578827} 01/29/2022 05:43:04 - INFO - codeparrot_training - Step 13141: {'lr': 0.0004364307884029834, 'samples': 2523264, 'steps': 13141, 'loss/train': 1.9248824715614319} 01/29/2022 05:43:09 - INFO - codeparrot_training - Step 13142: {'lr': 0.0004364198864163533, 'samples': 2523456, 'steps': 13142, 'loss/train': 1.8773936033248901} 01/29/2022 05:43:13 - INFO - codeparrot_training - Step 13143: {'lr': 0.00043640898363115954, 'samples': 2523648, 'steps': 13143, 'loss/train': 0.9614924490451813} 01/29/2022 05:43:19 - INFO - codeparrot_training - Step 13144: {'lr': 0.000436398080047449, 'samples': 2523840, 'steps': 13144, 'loss/train': 1.3546881079673767} 01/29/2022 05:43:23 - INFO - codeparrot_training - Step 13145: {'lr': 0.0004363871756652682, 'samples': 2524032, 'steps': 13145, 'loss/train': 1.0135522484779358} 01/29/2022 05:43:27 - INFO - codeparrot_training - Step 13146: {'lr': 0.00043637627048466395, 'samples': 2524224, 'steps': 13146, 'loss/train': 1.9728357195854187} 01/29/2022 05:43:31 - INFO - codeparrot_training - Step 13147: {'lr': 0.00043636536450568293, 'samples': 2524416, 'steps': 13147, 'loss/train': 1.732046663761139} 01/29/2022 05:43:36 - INFO - codeparrot_training - Step 13148: {'lr': 0.0004363544577283718, 'samples': 2524608, 'steps': 13148, 'loss/train': 2.1052724719047546} 01/29/2022 05:43:41 - INFO - codeparrot_training - Step 13149: {'lr': 0.00043634355015277745, 'samples': 2524800, 'steps': 13149, 'loss/train': 1.2877047657966614} 01/29/2022 05:43:45 - INFO - codeparrot_training - Step 13150: {'lr': 0.0004363326417789465, 'samples': 2524992, 'steps': 13150, 'loss/train': 1.8565465807914734} 01/29/2022 05:43:49 - INFO - codeparrot_training - Step 13151: {'lr': 0.0004363217326069256, 'samples': 2525184, 'steps': 13151, 'loss/train': 2.339028060436249} 01/29/2022 05:43:54 - INFO - codeparrot_training - Step 13152: {'lr': 0.0004363108226367616, 'samples': 2525376, 'steps': 13152, 'loss/train': 0.666097104549408} 01/29/2022 05:43:59 - INFO - codeparrot_training - Step 13153: {'lr': 0.0004362999118685012, 'samples': 2525568, 'steps': 13153, 'loss/train': 1.351633608341217} 01/29/2022 05:44:03 - INFO - codeparrot_training - Step 13154: {'lr': 0.0004362890003021911, 'samples': 2525760, 'steps': 13154, 'loss/train': 0.9790256917476654} 01/29/2022 05:44:07 - INFO - codeparrot_training - Step 13155: {'lr': 0.00043627808793787813, 'samples': 2525952, 'steps': 13155, 'loss/train': 1.6890453100204468} 01/29/2022 05:44:12 - INFO - codeparrot_training - Step 13156: {'lr': 0.00043626717477560897, 'samples': 2526144, 'steps': 13156, 'loss/train': 1.537218153476715} 01/29/2022 05:44:16 - INFO - codeparrot_training - Step 13157: {'lr': 0.00043625626081543033, 'samples': 2526336, 'steps': 13157, 'loss/train': 1.3493433594703674} 01/29/2022 05:44:22 - INFO - codeparrot_training - Step 13158: {'lr': 0.0004362453460573891, 'samples': 2526528, 'steps': 13158, 'loss/train': 1.0280087292194366} 01/29/2022 05:44:26 - INFO - codeparrot_training - Step 13159: {'lr': 0.0004362344305015319, 'samples': 2526720, 'steps': 13159, 'loss/train': 1.4829039573669434} 01/29/2022 05:44:30 - INFO - codeparrot_training - Step 13160: {'lr': 0.0004362235141479055, 'samples': 2526912, 'steps': 13160, 'loss/train': 2.3341028094291687} 01/29/2022 05:44:35 - INFO - codeparrot_training - Step 13161: {'lr': 0.00043621259699655674, 'samples': 2527104, 'steps': 13161, 'loss/train': 1.0910650491714478} 01/29/2022 05:44:39 - INFO - codeparrot_training - Step 13162: {'lr': 0.0004362016790475324, 'samples': 2527296, 'steps': 13162, 'loss/train': 1.955718219280243} 01/29/2022 05:44:43 - INFO - codeparrot_training - Step 13163: {'lr': 0.0004361907603008791, 'samples': 2527488, 'steps': 13163, 'loss/train': 1.6882041692733765} 01/29/2022 05:44:49 - INFO - codeparrot_training - Step 13164: {'lr': 0.00043617984075664375, 'samples': 2527680, 'steps': 13164, 'loss/train': 7.327325820922852} 01/29/2022 05:44:53 - INFO - codeparrot_training - Step 13165: {'lr': 0.000436168920414873, 'samples': 2527872, 'steps': 13165, 'loss/train': 1.381412297487259} 01/29/2022 05:44:57 - INFO - codeparrot_training - Step 13166: {'lr': 0.0004361579992756138, 'samples': 2528064, 'steps': 13166, 'loss/train': 1.4219078421592712} 01/29/2022 05:45:02 - INFO - codeparrot_training - Step 13167: {'lr': 0.00043614707733891285, 'samples': 2528256, 'steps': 13167, 'loss/train': 1.3526045680046082} 01/29/2022 05:45:06 - INFO - codeparrot_training - Step 13168: {'lr': 0.00043613615460481685, 'samples': 2528448, 'steps': 13168, 'loss/train': 2.304851710796356} 01/29/2022 05:45:12 - INFO - codeparrot_training - Step 13169: {'lr': 0.0004361252310733728, 'samples': 2528640, 'steps': 13169, 'loss/train': 1.7886949181556702} 01/29/2022 05:45:16 - INFO - codeparrot_training - Step 13170: {'lr': 0.0004361143067446273, 'samples': 2528832, 'steps': 13170, 'loss/train': 1.4791942834854126} 01/29/2022 05:45:20 - INFO - codeparrot_training - Step 13171: {'lr': 0.00043610338161862713, 'samples': 2529024, 'steps': 13171, 'loss/train': 0.9406485557556152} 01/29/2022 05:45:25 - INFO - codeparrot_training - Step 13172: {'lr': 0.00043609245569541924, 'samples': 2529216, 'steps': 13172, 'loss/train': 1.7785474061965942} 01/29/2022 05:45:29 - INFO - codeparrot_training - Step 13173: {'lr': 0.0004360815289750503, 'samples': 2529408, 'steps': 13173, 'loss/train': 2.2774881720542908} 01/29/2022 05:45:34 - INFO - codeparrot_training - Step 13174: {'lr': 0.0004360706014575672, 'samples': 2529600, 'steps': 13174, 'loss/train': 1.2486292719841003} 01/29/2022 05:45:38 - INFO - codeparrot_training - Step 13175: {'lr': 0.00043605967314301673, 'samples': 2529792, 'steps': 13175, 'loss/train': 0.8580906093120575} 01/29/2022 05:45:43 - INFO - codeparrot_training - Step 13176: {'lr': 0.0004360487440314458, 'samples': 2529984, 'steps': 13176, 'loss/train': 1.5761237740516663} 01/29/2022 05:45:47 - INFO - codeparrot_training - Step 13177: {'lr': 0.000436037814122901, 'samples': 2530176, 'steps': 13177, 'loss/train': 1.635887324810028} 01/29/2022 05:45:51 - INFO - codeparrot_training - Step 13178: {'lr': 0.0004360268834174294, 'samples': 2530368, 'steps': 13178, 'loss/train': 2.1579789519309998} 01/29/2022 05:45:56 - INFO - codeparrot_training - Step 13179: {'lr': 0.00043601595191507757, 'samples': 2530560, 'steps': 13179, 'loss/train': 2.441765606403351} 01/29/2022 05:46:01 - INFO - codeparrot_training - Step 13180: {'lr': 0.0004360050196158925, 'samples': 2530752, 'steps': 13180, 'loss/train': 1.9916544556617737} 01/29/2022 05:46:05 - INFO - codeparrot_training - Step 13181: {'lr': 0.000435994086519921, 'samples': 2530944, 'steps': 13181, 'loss/train': 2.0214326977729797} 01/29/2022 05:46:09 - INFO - codeparrot_training - Step 13182: {'lr': 0.00043598315262720995, 'samples': 2531136, 'steps': 13182, 'loss/train': 1.6305225491523743} 01/29/2022 05:46:13 - INFO - codeparrot_training - Step 13183: {'lr': 0.00043597221793780606, 'samples': 2531328, 'steps': 13183, 'loss/train': 1.936576008796692} 01/29/2022 05:46:19 - INFO - codeparrot_training - Step 13184: {'lr': 0.0004359612824517563, 'samples': 2531520, 'steps': 13184, 'loss/train': 1.5924702286720276} 01/29/2022 05:46:24 - INFO - codeparrot_training - Step 13185: {'lr': 0.0004359503461691074, 'samples': 2531712, 'steps': 13185, 'loss/train': 1.8952538967132568} 01/29/2022 05:46:28 - INFO - codeparrot_training - Step 13186: {'lr': 0.00043593940908990625, 'samples': 2531904, 'steps': 13186, 'loss/train': 1.1726801991462708} 01/29/2022 05:46:32 - INFO - codeparrot_training - Step 13187: {'lr': 0.00043592847121419974, 'samples': 2532096, 'steps': 13187, 'loss/train': 2.2711726427078247} 01/29/2022 05:46:36 - INFO - codeparrot_training - Step 13188: {'lr': 0.00043591753254203474, 'samples': 2532288, 'steps': 13188, 'loss/train': 1.8549578189849854} 01/29/2022 05:46:42 - INFO - codeparrot_training - Step 13189: {'lr': 0.00043590659307345803, 'samples': 2532480, 'steps': 13189, 'loss/train': 1.6086031794548035} 01/29/2022 05:46:46 - INFO - codeparrot_training - Step 13190: {'lr': 0.0004358956528085165, 'samples': 2532672, 'steps': 13190, 'loss/train': 1.882176160812378} 01/29/2022 05:46:51 - INFO - codeparrot_training - Step 13191: {'lr': 0.0004358847117472571, 'samples': 2532864, 'steps': 13191, 'loss/train': 2.354579508304596} 01/29/2022 05:46:55 - INFO - codeparrot_training - Step 13192: {'lr': 0.00043587376988972655, 'samples': 2533056, 'steps': 13192, 'loss/train': 2.43038409948349} 01/29/2022 05:46:59 - INFO - codeparrot_training - Step 13193: {'lr': 0.0004358628272359718, 'samples': 2533248, 'steps': 13193, 'loss/train': 1.8929357528686523} 01/29/2022 05:47:03 - INFO - codeparrot_training - Step 13194: {'lr': 0.0004358518837860397, 'samples': 2533440, 'steps': 13194, 'loss/train': 0.8482474386692047} 01/29/2022 05:47:09 - INFO - codeparrot_training - Step 13195: {'lr': 0.0004358409395399772, 'samples': 2533632, 'steps': 13195, 'loss/train': 1.9371598362922668} 01/29/2022 05:47:13 - INFO - codeparrot_training - Step 13196: {'lr': 0.00043582999449783103, 'samples': 2533824, 'steps': 13196, 'loss/train': 1.836475431919098} 01/29/2022 05:47:17 - INFO - codeparrot_training - Step 13197: {'lr': 0.00043581904865964825, 'samples': 2534016, 'steps': 13197, 'loss/train': 1.816279649734497} 01/29/2022 05:47:21 - INFO - codeparrot_training - Step 13198: {'lr': 0.0004358081020254756, 'samples': 2534208, 'steps': 13198, 'loss/train': 2.3531153798103333} 01/29/2022 05:47:26 - INFO - codeparrot_training - Step 13199: {'lr': 0.0004357971545953601, 'samples': 2534400, 'steps': 13199, 'loss/train': 1.8281936645507812} 01/29/2022 05:47:31 - INFO - codeparrot_training - Step 13200: {'lr': 0.00043578620636934855, 'samples': 2534592, 'steps': 13200, 'loss/train': 1.806220293045044} 01/29/2022 05:47:35 - INFO - codeparrot_training - Step 13201: {'lr': 0.0004357752573474879, 'samples': 2534784, 'steps': 13201, 'loss/train': 1.5962448120117188} 01/29/2022 05:47:39 - INFO - codeparrot_training - Step 13202: {'lr': 0.0004357643075298251, 'samples': 2534976, 'steps': 13202, 'loss/train': 2.4208452701568604} 01/29/2022 05:47:44 - INFO - codeparrot_training - Step 13203: {'lr': 0.00043575335691640695, 'samples': 2535168, 'steps': 13203, 'loss/train': 2.112578809261322} 01/29/2022 05:47:48 - INFO - codeparrot_training - Step 13204: {'lr': 0.0004357424055072804, 'samples': 2535360, 'steps': 13204, 'loss/train': 1.5842576622962952} 01/29/2022 05:47:54 - INFO - codeparrot_training - Step 13205: {'lr': 0.0004357314533024923, 'samples': 2535552, 'steps': 13205, 'loss/train': 1.6534216403961182} 01/29/2022 05:47:58 - INFO - codeparrot_training - Step 13206: {'lr': 0.0004357205003020897, 'samples': 2535744, 'steps': 13206, 'loss/train': 1.3650809526443481} 01/29/2022 05:48:03 - INFO - codeparrot_training - Step 13207: {'lr': 0.00043570954650611944, 'samples': 2535936, 'steps': 13207, 'loss/train': 1.3413663804531097} 01/29/2022 05:48:07 - INFO - codeparrot_training - Step 13208: {'lr': 0.00043569859191462847, 'samples': 2536128, 'steps': 13208, 'loss/train': 0.8910572826862335} 01/29/2022 05:48:11 - INFO - codeparrot_training - Step 13209: {'lr': 0.0004356876365276636, 'samples': 2536320, 'steps': 13209, 'loss/train': 7.445134878158569} 01/29/2022 05:48:16 - INFO - codeparrot_training - Step 13210: {'lr': 0.00043567668034527195, 'samples': 2536512, 'steps': 13210, 'loss/train': 0.8932599127292633} 01/29/2022 05:48:21 - INFO - codeparrot_training - Step 13211: {'lr': 0.0004356657233675004, 'samples': 2536704, 'steps': 13211, 'loss/train': 2.0157471299171448} 01/29/2022 05:48:25 - INFO - codeparrot_training - Step 13212: {'lr': 0.00043565476559439577, 'samples': 2536896, 'steps': 13212, 'loss/train': 2.1863377690315247} 01/29/2022 05:48:29 - INFO - codeparrot_training - Step 13213: {'lr': 0.0004356438070260051, 'samples': 2537088, 'steps': 13213, 'loss/train': 0.24518763273954391} 01/29/2022 05:48:33 - INFO - codeparrot_training - Step 13214: {'lr': 0.00043563284766237533, 'samples': 2537280, 'steps': 13214, 'loss/train': 1.7553802728652954} 01/29/2022 05:48:40 - INFO - codeparrot_training - Step 13215: {'lr': 0.00043562188750355336, 'samples': 2537472, 'steps': 13215, 'loss/train': 1.936525046825409} 01/29/2022 05:48:44 - INFO - codeparrot_training - Step 13216: {'lr': 0.0004356109265495861, 'samples': 2537664, 'steps': 13216, 'loss/train': 2.7999227643013} 01/29/2022 05:48:48 - INFO - codeparrot_training - Step 13217: {'lr': 0.00043559996480052067, 'samples': 2537856, 'steps': 13217, 'loss/train': 0.7937850952148438} 01/29/2022 05:48:52 - INFO - codeparrot_training - Step 13218: {'lr': 0.0004355890022564039, 'samples': 2538048, 'steps': 13218, 'loss/train': 1.6324342489242554} 01/29/2022 05:48:57 - INFO - codeparrot_training - Step 13219: {'lr': 0.00043557803891728275, 'samples': 2538240, 'steps': 13219, 'loss/train': 2.006324887275696} 01/29/2022 05:49:02 - INFO - codeparrot_training - Step 13220: {'lr': 0.00043556707478320425, 'samples': 2538432, 'steps': 13220, 'loss/train': 0.8986963927745819} 01/29/2022 05:49:07 - INFO - codeparrot_training - Step 13221: {'lr': 0.00043555610985421527, 'samples': 2538624, 'steps': 13221, 'loss/train': 2.0161399841308594} 01/29/2022 05:49:11 - INFO - codeparrot_training - Step 13222: {'lr': 0.0004355451441303629, 'samples': 2538816, 'steps': 13222, 'loss/train': 0.5859481394290924} 01/29/2022 05:49:15 - INFO - codeparrot_training - Step 13223: {'lr': 0.000435534177611694, 'samples': 2539008, 'steps': 13223, 'loss/train': 1.7839586734771729} 01/29/2022 05:49:19 - INFO - codeparrot_training - Step 13224: {'lr': 0.0004355232102982556, 'samples': 2539200, 'steps': 13224, 'loss/train': 2.17781800031662} 01/29/2022 05:49:25 - INFO - codeparrot_training - Step 13225: {'lr': 0.00043551224219009473, 'samples': 2539392, 'steps': 13225, 'loss/train': 2.2299731969833374} 01/29/2022 05:49:29 - INFO - codeparrot_training - Step 13226: {'lr': 0.0004355012732872583, 'samples': 2539584, 'steps': 13226, 'loss/train': 1.7548075318336487} 01/29/2022 05:49:33 - INFO - codeparrot_training - Step 13227: {'lr': 0.00043549030358979324, 'samples': 2539776, 'steps': 13227, 'loss/train': 1.8972787857055664} 01/29/2022 05:49:37 - INFO - codeparrot_training - Step 13228: {'lr': 0.0004354793330977467, 'samples': 2539968, 'steps': 13228, 'loss/train': 1.9090924859046936} 01/29/2022 05:49:42 - INFO - codeparrot_training - Step 13229: {'lr': 0.00043546836181116555, 'samples': 2540160, 'steps': 13229, 'loss/train': 2.249807059764862} 01/29/2022 05:49:47 - INFO - codeparrot_training - Step 13230: {'lr': 0.0004354573897300969, 'samples': 2540352, 'steps': 13230, 'loss/train': 1.8572931289672852} 01/29/2022 05:49:52 - INFO - codeparrot_training - Step 13231: {'lr': 0.0004354464168545876, 'samples': 2540544, 'steps': 13231, 'loss/train': 0.7623015940189362} 01/29/2022 05:49:56 - INFO - codeparrot_training - Step 13232: {'lr': 0.0004354354431846848, 'samples': 2540736, 'steps': 13232, 'loss/train': 1.056787133216858} 01/29/2022 05:50:00 - INFO - codeparrot_training - Step 13233: {'lr': 0.0004354244687204354, 'samples': 2540928, 'steps': 13233, 'loss/train': 1.80745929479599} 01/29/2022 05:50:04 - INFO - codeparrot_training - Step 13234: {'lr': 0.00043541349346188653, 'samples': 2541120, 'steps': 13234, 'loss/train': 2.22723126411438} 01/29/2022 05:50:10 - INFO - codeparrot_training - Step 13235: {'lr': 0.000435402517409085, 'samples': 2541312, 'steps': 13235, 'loss/train': 1.8618614673614502} 01/29/2022 05:50:14 - INFO - codeparrot_training - Step 13236: {'lr': 0.0004353915405620781, 'samples': 2541504, 'steps': 13236, 'loss/train': 1.7394996285438538} 01/29/2022 05:50:18 - INFO - codeparrot_training - Step 13237: {'lr': 0.0004353805629209126, 'samples': 2541696, 'steps': 13237, 'loss/train': 2.186763882637024} 01/29/2022 05:50:22 - INFO - codeparrot_training - Step 13238: {'lr': 0.0004353695844856357, 'samples': 2541888, 'steps': 13238, 'loss/train': 1.6593714952468872} 01/29/2022 05:50:27 - INFO - codeparrot_training - Step 13239: {'lr': 0.00043535860525629436, 'samples': 2542080, 'steps': 13239, 'loss/train': 1.5037137866020203} 01/29/2022 05:50:32 - INFO - codeparrot_training - Step 13240: {'lr': 0.00043534762523293557, 'samples': 2542272, 'steps': 13240, 'loss/train': 1.819876492023468} 01/29/2022 05:50:37 - INFO - codeparrot_training - Step 13241: {'lr': 0.00043533664441560636, 'samples': 2542464, 'steps': 13241, 'loss/train': 1.302780658006668} 01/29/2022 05:50:41 - INFO - codeparrot_training - Step 13242: {'lr': 0.0004353256628043539, 'samples': 2542656, 'steps': 13242, 'loss/train': 1.6870651245117188} 01/29/2022 05:50:45 - INFO - codeparrot_training - Step 13243: {'lr': 0.00043531468039922515, 'samples': 2542848, 'steps': 13243, 'loss/train': 1.3878247439861298} 01/29/2022 05:50:49 - INFO - codeparrot_training - Step 13244: {'lr': 0.0004353036972002671, 'samples': 2543040, 'steps': 13244, 'loss/train': 0.47267644107341766} 01/29/2022 05:50:54 - INFO - codeparrot_training - Step 13245: {'lr': 0.0004352927132075269, 'samples': 2543232, 'steps': 13245, 'loss/train': 1.766625165939331} 01/29/2022 05:50:59 - INFO - codeparrot_training - Step 13246: {'lr': 0.00043528172842105154, 'samples': 2543424, 'steps': 13246, 'loss/train': 2.2671759724617004} 01/29/2022 05:51:04 - INFO - codeparrot_training - Step 13247: {'lr': 0.00043527074284088806, 'samples': 2543616, 'steps': 13247, 'loss/train': 1.5519261360168457} 01/29/2022 05:51:08 - INFO - codeparrot_training - Step 13248: {'lr': 0.0004352597564670836, 'samples': 2543808, 'steps': 13248, 'loss/train': 1.7292605638504028} 01/29/2022 05:51:12 - INFO - codeparrot_training - Step 13249: {'lr': 0.00043524876929968516, 'samples': 2544000, 'steps': 13249, 'loss/train': 1.475553274154663} 01/29/2022 05:51:20 - INFO - codeparrot_training - Step 13250: {'lr': 0.0004352377813387398, 'samples': 2544192, 'steps': 13250, 'loss/train': 2.956510841846466} 01/29/2022 05:51:24 - INFO - codeparrot_training - Step 13251: {'lr': 0.0004352267925842946, 'samples': 2544384, 'steps': 13251, 'loss/train': 3.102517604827881} 01/29/2022 05:51:28 - INFO - codeparrot_training - Step 13252: {'lr': 0.00043521580303639663, 'samples': 2544576, 'steps': 13252, 'loss/train': 1.811089038848877} 01/29/2022 05:51:32 - INFO - codeparrot_training - Step 13253: {'lr': 0.000435204812695093, 'samples': 2544768, 'steps': 13253, 'loss/train': 1.741011142730713} 01/29/2022 05:51:37 - INFO - codeparrot_training - Step 13254: {'lr': 0.00043519382156043075, 'samples': 2544960, 'steps': 13254, 'loss/train': 1.1412127017974854} 01/29/2022 05:51:42 - INFO - codeparrot_training - Step 13255: {'lr': 0.0004351828296324569, 'samples': 2545152, 'steps': 13255, 'loss/train': 1.6596754789352417} 01/29/2022 05:51:46 - INFO - codeparrot_training - Step 13256: {'lr': 0.00043517183691121875, 'samples': 2545344, 'steps': 13256, 'loss/train': 1.497141569852829} 01/29/2022 05:51:50 - INFO - codeparrot_training - Step 13257: {'lr': 0.00043516084339676316, 'samples': 2545536, 'steps': 13257, 'loss/train': 1.4136725664138794} 01/29/2022 05:51:55 - INFO - codeparrot_training - Step 13258: {'lr': 0.00043514984908913734, 'samples': 2545728, 'steps': 13258, 'loss/train': 1.325924277305603} 01/29/2022 05:51:59 - INFO - codeparrot_training - Step 13259: {'lr': 0.0004351388539883883, 'samples': 2545920, 'steps': 13259, 'loss/train': 1.619542121887207} 01/29/2022 05:52:04 - INFO - codeparrot_training - Step 13260: {'lr': 0.00043512785809456323, 'samples': 2546112, 'steps': 13260, 'loss/train': 1.4355667233467102} 01/29/2022 05:52:09 - INFO - codeparrot_training - Step 13261: {'lr': 0.00043511686140770925, 'samples': 2546304, 'steps': 13261, 'loss/train': 1.8720817565917969} 01/29/2022 05:52:13 - INFO - codeparrot_training - Step 13262: {'lr': 0.0004351058639278734, 'samples': 2546496, 'steps': 13262, 'loss/train': 2.0232932567596436} 01/29/2022 05:52:17 - INFO - codeparrot_training - Step 13263: {'lr': 0.0004350948656551028, 'samples': 2546688, 'steps': 13263, 'loss/train': 1.3607611656188965} 01/29/2022 05:52:21 - INFO - codeparrot_training - Step 13264: {'lr': 0.0004350838665894445, 'samples': 2546880, 'steps': 13264, 'loss/train': 1.3879238963127136} 01/29/2022 05:52:28 - INFO - codeparrot_training - Step 13265: {'lr': 0.0004350728667309458, 'samples': 2547072, 'steps': 13265, 'loss/train': 2.456774890422821} 01/29/2022 05:52:33 - INFO - codeparrot_training - Step 13266: {'lr': 0.0004350618660796536, 'samples': 2547264, 'steps': 13266, 'loss/train': 2.1613662242889404} 01/29/2022 05:52:37 - INFO - codeparrot_training - Step 13267: {'lr': 0.0004350508646356152, 'samples': 2547456, 'steps': 13267, 'loss/train': 2.1321685910224915} 01/29/2022 05:52:41 - INFO - codeparrot_training - Step 13268: {'lr': 0.00043503986239887765, 'samples': 2547648, 'steps': 13268, 'loss/train': 2.0965331196784973} 01/29/2022 05:52:45 - INFO - codeparrot_training - Step 13269: {'lr': 0.0004350288593694881, 'samples': 2547840, 'steps': 13269, 'loss/train': 1.87388277053833} 01/29/2022 05:52:51 - INFO - codeparrot_training - Step 13270: {'lr': 0.00043501785554749363, 'samples': 2548032, 'steps': 13270, 'loss/train': 1.0091410875320435} 01/29/2022 05:52:55 - INFO - codeparrot_training - Step 13271: {'lr': 0.00043500685093294145, 'samples': 2548224, 'steps': 13271, 'loss/train': 1.6871716976165771} 01/29/2022 05:52:59 - INFO - codeparrot_training - Step 13272: {'lr': 0.0004349958455258786, 'samples': 2548416, 'steps': 13272, 'loss/train': 1.5996687412261963} 01/29/2022 05:53:03 - INFO - codeparrot_training - Step 13273: {'lr': 0.00043498483932635237, 'samples': 2548608, 'steps': 13273, 'loss/train': 1.8500958681106567} 01/29/2022 05:53:08 - INFO - codeparrot_training - Step 13274: {'lr': 0.0004349738323344098, 'samples': 2548800, 'steps': 13274, 'loss/train': 1.101547122001648} 01/29/2022 05:53:15 - INFO - codeparrot_training - Step 13275: {'lr': 0.00043496282455009807, 'samples': 2548992, 'steps': 13275, 'loss/train': 2.163508951663971} 01/29/2022 05:53:19 - INFO - codeparrot_training - Step 13276: {'lr': 0.00043495181597346435, 'samples': 2549184, 'steps': 13276, 'loss/train': 2.068496525287628} 01/29/2022 05:53:23 - INFO - codeparrot_training - Step 13277: {'lr': 0.0004349408066045557, 'samples': 2549376, 'steps': 13277, 'loss/train': 1.1130215227603912} 01/29/2022 05:53:28 - INFO - codeparrot_training - Step 13278: {'lr': 0.00043492979644341943, 'samples': 2549568, 'steps': 13278, 'loss/train': 2.0930845141410828} 01/29/2022 05:53:32 - INFO - codeparrot_training - Step 13279: {'lr': 0.0004349187854901026, 'samples': 2549760, 'steps': 13279, 'loss/train': 1.455047070980072} 01/29/2022 05:53:37 - INFO - codeparrot_training - Step 13280: {'lr': 0.00043490777374465244, 'samples': 2549952, 'steps': 13280, 'loss/train': 1.8061927556991577} 01/29/2022 05:53:41 - INFO - codeparrot_training - Step 13281: {'lr': 0.0004348967612071161, 'samples': 2550144, 'steps': 13281, 'loss/train': 1.4441678524017334} 01/29/2022 05:53:46 - INFO - codeparrot_training - Step 13282: {'lr': 0.0004348857478775407, 'samples': 2550336, 'steps': 13282, 'loss/train': 7.985876798629761} 01/29/2022 05:53:50 - INFO - codeparrot_training - Step 13283: {'lr': 0.00043487473375597354, 'samples': 2550528, 'steps': 13283, 'loss/train': 1.4502280354499817} 01/29/2022 05:53:54 - INFO - codeparrot_training - Step 13284: {'lr': 0.00043486371884246164, 'samples': 2550720, 'steps': 13284, 'loss/train': 2.176802694797516} 01/29/2022 05:54:00 - INFO - codeparrot_training - Step 13285: {'lr': 0.0004348527031370523, 'samples': 2550912, 'steps': 13285, 'loss/train': 5.557783842086792} 01/29/2022 05:54:04 - INFO - codeparrot_training - Step 13286: {'lr': 0.00043484168663979265, 'samples': 2551104, 'steps': 13286, 'loss/train': 2.5904277563095093} 01/29/2022 05:54:08 - INFO - codeparrot_training - Step 13287: {'lr': 0.00043483066935073, 'samples': 2551296, 'steps': 13287, 'loss/train': 1.7175752520561218} 01/29/2022 05:54:12 - INFO - codeparrot_training - Step 13288: {'lr': 0.0004348196512699114, 'samples': 2551488, 'steps': 13288, 'loss/train': 1.6412063241004944} 01/29/2022 05:54:17 - INFO - codeparrot_training - Step 13289: {'lr': 0.00043480863239738404, 'samples': 2551680, 'steps': 13289, 'loss/train': 0.9779422581195831} 01/29/2022 05:54:24 - INFO - codeparrot_training - Step 13290: {'lr': 0.0004347976127331953, 'samples': 2551872, 'steps': 13290, 'loss/train': 1.956332266330719} 01/29/2022 05:54:28 - INFO - codeparrot_training - Step 13291: {'lr': 0.00043478659227739216, 'samples': 2552064, 'steps': 13291, 'loss/train': 1.6271472573280334} 01/29/2022 05:54:32 - INFO - codeparrot_training - Step 13292: {'lr': 0.00043477557103002197, 'samples': 2552256, 'steps': 13292, 'loss/train': 1.6660839915275574} 01/29/2022 05:54:36 - INFO - codeparrot_training - Step 13293: {'lr': 0.00043476454899113193, 'samples': 2552448, 'steps': 13293, 'loss/train': 2.338370203971863} 01/29/2022 05:54:41 - INFO - codeparrot_training - Step 13294: {'lr': 0.00043475352616076927, 'samples': 2552640, 'steps': 13294, 'loss/train': 1.7031276226043701} 01/29/2022 05:54:46 - INFO - codeparrot_training - Step 13295: {'lr': 0.0004347425025389811, 'samples': 2552832, 'steps': 13295, 'loss/train': 1.8960038423538208} 01/29/2022 05:54:50 - INFO - codeparrot_training - Step 13296: {'lr': 0.0004347314781258147, 'samples': 2553024, 'steps': 13296, 'loss/train': 1.2781618237495422} 01/29/2022 05:54:54 - INFO - codeparrot_training - Step 13297: {'lr': 0.00043472045292131735, 'samples': 2553216, 'steps': 13297, 'loss/train': 2.149798572063446} 01/29/2022 05:54:59 - INFO - codeparrot_training - Step 13298: {'lr': 0.0004347094269255362, 'samples': 2553408, 'steps': 13298, 'loss/train': 2.482848644256592} 01/29/2022 05:55:03 - INFO - codeparrot_training - Step 13299: {'lr': 0.0004346984001385186, 'samples': 2553600, 'steps': 13299, 'loss/train': 1.3951779305934906} 01/29/2022 05:55:08 - INFO - codeparrot_training - Step 13300: {'lr': 0.00043468737256031155, 'samples': 2553792, 'steps': 13300, 'loss/train': 1.5529398322105408} 01/29/2022 05:55:13 - INFO - codeparrot_training - Step 13301: {'lr': 0.00043467634419096257, 'samples': 2553984, 'steps': 13301, 'loss/train': 2.6743903756141663} 01/29/2022 05:55:17 - INFO - codeparrot_training - Step 13302: {'lr': 0.00043466531503051875, 'samples': 2554176, 'steps': 13302, 'loss/train': 2.034672796726227} 01/29/2022 05:55:21 - INFO - codeparrot_training - Step 13303: {'lr': 0.0004346542850790273, 'samples': 2554368, 'steps': 13303, 'loss/train': 1.3803758025169373} 01/29/2022 05:55:25 - INFO - codeparrot_training - Step 13304: {'lr': 0.00043464325433653563, 'samples': 2554560, 'steps': 13304, 'loss/train': 1.7495437860488892} 01/29/2022 05:55:32 - INFO - codeparrot_training - Step 13305: {'lr': 0.00043463222280309076, 'samples': 2554752, 'steps': 13305, 'loss/train': 1.2153604924678802} 01/29/2022 05:55:36 - INFO - codeparrot_training - Step 13306: {'lr': 0.00043462119047874015, 'samples': 2554944, 'steps': 13306, 'loss/train': 2.051835000514984} 01/29/2022 05:55:41 - INFO - codeparrot_training - Step 13307: {'lr': 0.000434610157363531, 'samples': 2555136, 'steps': 13307, 'loss/train': 1.242196798324585} 01/29/2022 05:55:45 - INFO - codeparrot_training - Step 13308: {'lr': 0.0004345991234575105, 'samples': 2555328, 'steps': 13308, 'loss/train': 1.1569359004497528} 01/29/2022 05:55:49 - INFO - codeparrot_training - Step 13309: {'lr': 0.00043458808876072595, 'samples': 2555520, 'steps': 13309, 'loss/train': 1.1922306418418884} 01/29/2022 05:55:55 - INFO - codeparrot_training - Step 13310: {'lr': 0.0004345770532732247, 'samples': 2555712, 'steps': 13310, 'loss/train': 1.2354009747505188} 01/29/2022 05:55:59 - INFO - codeparrot_training - Step 13311: {'lr': 0.00043456601699505407, 'samples': 2555904, 'steps': 13311, 'loss/train': 1.6187919974327087} 01/29/2022 05:56:03 - INFO - codeparrot_training - Step 13312: {'lr': 0.00043455497992626104, 'samples': 2556096, 'steps': 13312, 'loss/train': 1.7180925607681274} 01/29/2022 05:56:07 - INFO - codeparrot_training - Step 13313: {'lr': 0.0004345439420668932, 'samples': 2556288, 'steps': 13313, 'loss/train': 1.897522509098053} 01/29/2022 05:56:12 - INFO - codeparrot_training - Step 13314: {'lr': 0.0004345329034169977, 'samples': 2556480, 'steps': 13314, 'loss/train': 2.053542137145996} 01/29/2022 05:56:17 - INFO - codeparrot_training - Step 13315: {'lr': 0.00043452186397662174, 'samples': 2556672, 'steps': 13315, 'loss/train': 1.7537986636161804} 01/29/2022 05:56:21 - INFO - codeparrot_training - Step 13316: {'lr': 0.0004345108237458128, 'samples': 2556864, 'steps': 13316, 'loss/train': 2.0988643169403076} 01/29/2022 05:56:25 - INFO - codeparrot_training - Step 13317: {'lr': 0.00043449978272461806, 'samples': 2557056, 'steps': 13317, 'loss/train': 1.6278730630874634} 01/29/2022 05:56:30 - INFO - codeparrot_training - Step 13318: {'lr': 0.0004344887409130848, 'samples': 2557248, 'steps': 13318, 'loss/train': 2.22104412317276} 01/29/2022 05:56:35 - INFO - codeparrot_training - Step 13319: {'lr': 0.0004344776983112604, 'samples': 2557440, 'steps': 13319, 'loss/train': 2.522648334503174} 01/29/2022 05:56:39 - INFO - codeparrot_training - Step 13320: {'lr': 0.0004344666549191921, 'samples': 2557632, 'steps': 13320, 'loss/train': 1.425373613834381} 01/29/2022 05:56:43 - INFO - codeparrot_training - Step 13321: {'lr': 0.0004344556107369272, 'samples': 2557824, 'steps': 13321, 'loss/train': 1.4528351426124573} 01/29/2022 05:56:48 - INFO - codeparrot_training - Step 13322: {'lr': 0.00043444456576451307, 'samples': 2558016, 'steps': 13322, 'loss/train': 1.8381493091583252} 01/29/2022 05:56:52 - INFO - codeparrot_training - Step 13323: {'lr': 0.000434433520001997, 'samples': 2558208, 'steps': 13323, 'loss/train': 2.87593013048172} 01/29/2022 05:56:59 - INFO - codeparrot_training - Step 13324: {'lr': 0.0004344224734494263, 'samples': 2558400, 'steps': 13324, 'loss/train': 1.164175271987915} 01/29/2022 05:57:03 - INFO - codeparrot_training - Step 13325: {'lr': 0.00043441142610684826, 'samples': 2558592, 'steps': 13325, 'loss/train': 0.8681517541408539} 01/29/2022 05:57:07 - INFO - codeparrot_training - Step 13326: {'lr': 0.0004344003779743102, 'samples': 2558784, 'steps': 13326, 'loss/train': 1.522430419921875} 01/29/2022 05:57:12 - INFO - codeparrot_training - Step 13327: {'lr': 0.0004343893290518595, 'samples': 2558976, 'steps': 13327, 'loss/train': 1.7679226398468018} 01/29/2022 05:57:16 - INFO - codeparrot_training - Step 13328: {'lr': 0.0004343782793395435, 'samples': 2559168, 'steps': 13328, 'loss/train': 2.488399565219879} 01/29/2022 05:57:20 - INFO - codeparrot_training - Step 13329: {'lr': 0.00043436722883740943, 'samples': 2559360, 'steps': 13329, 'loss/train': 1.5164510607719421} 01/29/2022 05:57:25 - INFO - codeparrot_training - Step 13330: {'lr': 0.0004343561775455047, 'samples': 2559552, 'steps': 13330, 'loss/train': 1.6299249529838562} 01/29/2022 05:57:30 - INFO - codeparrot_training - Step 13331: {'lr': 0.00043434512546387674, 'samples': 2559744, 'steps': 13331, 'loss/train': 3.2200008630752563} 01/29/2022 05:57:34 - INFO - codeparrot_training - Step 13332: {'lr': 0.0004343340725925727, 'samples': 2559936, 'steps': 13332, 'loss/train': 0.26820628345012665} 01/29/2022 05:57:38 - INFO - codeparrot_training - Step 13333: {'lr': 0.0004343230189316401, 'samples': 2560128, 'steps': 13333, 'loss/train': 2.202392578125} 01/29/2022 05:57:43 - INFO - codeparrot_training - Step 13334: {'lr': 0.00043431196448112615, 'samples': 2560320, 'steps': 13334, 'loss/train': 0.8400595486164093} 01/29/2022 05:57:50 - INFO - codeparrot_training - Step 13335: {'lr': 0.0004343009092410783, 'samples': 2560512, 'steps': 13335, 'loss/train': 2.3448997735977173} 01/29/2022 05:57:54 - INFO - codeparrot_training - Step 13336: {'lr': 0.0004342898532115439, 'samples': 2560704, 'steps': 13336, 'loss/train': 2.032275974750519} 01/29/2022 05:57:58 - INFO - codeparrot_training - Step 13337: {'lr': 0.00043427879639257024, 'samples': 2560896, 'steps': 13337, 'loss/train': 1.7656851410865784} 01/29/2022 05:58:03 - INFO - codeparrot_training - Step 13338: {'lr': 0.0004342677387842048, 'samples': 2561088, 'steps': 13338, 'loss/train': 2.516857624053955} 01/29/2022 05:58:08 - INFO - codeparrot_training - Step 13339: {'lr': 0.0004342566803864948, 'samples': 2561280, 'steps': 13339, 'loss/train': 1.9558305144309998} 01/29/2022 05:58:12 - INFO - codeparrot_training - Step 13340: {'lr': 0.0004342456211994877, 'samples': 2561472, 'steps': 13340, 'loss/train': 5.01799464225769} 01/29/2022 05:58:17 - INFO - codeparrot_training - Step 13341: {'lr': 0.0004342345612232309, 'samples': 2561664, 'steps': 13341, 'loss/train': 2.070551812648773} 01/29/2022 05:58:21 - INFO - codeparrot_training - Step 13342: {'lr': 0.0004342235004577717, 'samples': 2561856, 'steps': 13342, 'loss/train': 1.8110716938972473} 01/29/2022 05:58:25 - INFO - codeparrot_training - Step 13343: {'lr': 0.00043421243890315753, 'samples': 2562048, 'steps': 13343, 'loss/train': 0.8002532422542572} 01/29/2022 05:58:29 - INFO - codeparrot_training - Step 13344: {'lr': 0.0004342013765594358, 'samples': 2562240, 'steps': 13344, 'loss/train': 1.2190118730068207} 01/29/2022 05:58:34 - INFO - codeparrot_training - Step 13345: {'lr': 0.0004341903134266538, 'samples': 2562432, 'steps': 13345, 'loss/train': 1.2463487684726715} 01/29/2022 05:58:39 - INFO - codeparrot_training - Step 13346: {'lr': 0.0004341792495048591, 'samples': 2562624, 'steps': 13346, 'loss/train': 1.2656783759593964} 01/29/2022 05:58:43 - INFO - codeparrot_training - Step 13347: {'lr': 0.00043416818479409894, 'samples': 2562816, 'steps': 13347, 'loss/train': 1.3764779269695282} 01/29/2022 05:58:47 - INFO - codeparrot_training - Step 13348: {'lr': 0.0004341571192944207, 'samples': 2563008, 'steps': 13348, 'loss/train': 1.721831202507019} 01/29/2022 05:58:51 - INFO - codeparrot_training - Step 13349: {'lr': 0.00043414605300587183, 'samples': 2563200, 'steps': 13349, 'loss/train': 0.7698646187782288} 01/29/2022 05:58:58 - INFO - codeparrot_training - Step 13350: {'lr': 0.0004341349859284998, 'samples': 2563392, 'steps': 13350, 'loss/train': 1.1449421346187592} 01/29/2022 05:59:03 - INFO - codeparrot_training - Step 13351: {'lr': 0.0004341239180623519, 'samples': 2563584, 'steps': 13351, 'loss/train': 1.1276456415653229} 01/29/2022 05:59:07 - INFO - codeparrot_training - Step 13352: {'lr': 0.0004341128494074756, 'samples': 2563776, 'steps': 13352, 'loss/train': 2.069171905517578} 01/29/2022 05:59:11 - INFO - codeparrot_training - Step 13353: {'lr': 0.00043410177996391837, 'samples': 2563968, 'steps': 13353, 'loss/train': 1.5709192156791687} 01/29/2022 05:59:16 - INFO - codeparrot_training - Step 13354: {'lr': 0.00043409070973172753, 'samples': 2564160, 'steps': 13354, 'loss/train': 2.1824585795402527} 01/29/2022 05:59:21 - INFO - codeparrot_training - Step 13355: {'lr': 0.0004340796387109506, 'samples': 2564352, 'steps': 13355, 'loss/train': 2.3634291887283325} 01/29/2022 05:59:25 - INFO - codeparrot_training - Step 13356: {'lr': 0.00043406856690163487, 'samples': 2564544, 'steps': 13356, 'loss/train': 1.575720191001892} 01/29/2022 05:59:29 - INFO - codeparrot_training - Step 13357: {'lr': 0.0004340574943038279, 'samples': 2564736, 'steps': 13357, 'loss/train': 1.415948063135147} 01/29/2022 05:59:33 - INFO - codeparrot_training - Step 13358: {'lr': 0.00043404642091757705, 'samples': 2564928, 'steps': 13358, 'loss/train': 1.4314433634281158} 01/29/2022 05:59:39 - INFO - codeparrot_training - Step 13359: {'lr': 0.0004340353467429299, 'samples': 2565120, 'steps': 13359, 'loss/train': 1.739876389503479} 01/29/2022 05:59:43 - INFO - codeparrot_training - Step 13360: {'lr': 0.00043402427177993366, 'samples': 2565312, 'steps': 13360, 'loss/train': 1.6327187418937683} 01/29/2022 05:59:47 - INFO - codeparrot_training - Step 13361: {'lr': 0.00043401319602863584, 'samples': 2565504, 'steps': 13361, 'loss/train': 1.6885666251182556} 01/29/2022 05:59:52 - INFO - codeparrot_training - Step 13362: {'lr': 0.0004340021194890839, 'samples': 2565696, 'steps': 13362, 'loss/train': 2.1894861459732056} 01/29/2022 05:59:56 - INFO - codeparrot_training - Step 13363: {'lr': 0.0004339910421613253, 'samples': 2565888, 'steps': 13363, 'loss/train': 1.3293505311012268} 01/29/2022 06:00:01 - INFO - codeparrot_training - Step 13364: {'lr': 0.0004339799640454076, 'samples': 2566080, 'steps': 13364, 'loss/train': 2.188887298107147} 01/29/2022 06:00:06 - INFO - codeparrot_training - Step 13365: {'lr': 0.0004339688851413781, 'samples': 2566272, 'steps': 13365, 'loss/train': 3.4914876222610474} 01/29/2022 06:00:10 - INFO - codeparrot_training - Step 13366: {'lr': 0.0004339578054492843, 'samples': 2566464, 'steps': 13366, 'loss/train': 2.308052122592926} 01/29/2022 06:00:14 - INFO - codeparrot_training - Step 13367: {'lr': 0.0004339467249691737, 'samples': 2566656, 'steps': 13367, 'loss/train': 1.418888658285141} 01/29/2022 06:00:18 - INFO - codeparrot_training - Step 13368: {'lr': 0.0004339356437010937, 'samples': 2566848, 'steps': 13368, 'loss/train': 1.6789146065711975} 01/29/2022 06:00:25 - INFO - codeparrot_training - Step 13369: {'lr': 0.00043392456164509185, 'samples': 2567040, 'steps': 13369, 'loss/train': 2.3254001140594482} 01/29/2022 06:00:30 - INFO - codeparrot_training - Step 13370: {'lr': 0.00043391347880121554, 'samples': 2567232, 'steps': 13370, 'loss/train': 2.0309091210365295} 01/29/2022 06:00:34 - INFO - codeparrot_training - Step 13371: {'lr': 0.00043390239516951235, 'samples': 2567424, 'steps': 13371, 'loss/train': 1.381440818309784} 01/29/2022 06:00:38 - INFO - codeparrot_training - Step 13372: {'lr': 0.0004338913107500297, 'samples': 2567616, 'steps': 13372, 'loss/train': 2.413585066795349} 01/29/2022 06:00:42 - INFO - codeparrot_training - Step 13373: {'lr': 0.00043388022554281504, 'samples': 2567808, 'steps': 13373, 'loss/train': 1.3982647061347961} 01/29/2022 06:00:48 - INFO - codeparrot_training - Step 13374: {'lr': 0.00043386913954791584, 'samples': 2568000, 'steps': 13374, 'loss/train': 1.0687474608421326} 01/29/2022 06:00:52 - INFO - codeparrot_training - Step 13375: {'lr': 0.0004338580527653797, 'samples': 2568192, 'steps': 13375, 'loss/train': 1.6216732263565063} 01/29/2022 06:00:56 - INFO - codeparrot_training - Step 13376: {'lr': 0.000433846965195254, 'samples': 2568384, 'steps': 13376, 'loss/train': 1.5185324549674988} 01/29/2022 06:01:00 - INFO - codeparrot_training - Step 13377: {'lr': 0.0004338358768375863, 'samples': 2568576, 'steps': 13377, 'loss/train': 1.9253574013710022} 01/29/2022 06:01:04 - INFO - codeparrot_training - Step 13378: {'lr': 0.000433824787692424, 'samples': 2568768, 'steps': 13378, 'loss/train': 1.4330487549304962} 01/29/2022 06:01:11 - INFO - codeparrot_training - Step 13379: {'lr': 0.0004338136977598148, 'samples': 2568960, 'steps': 13379, 'loss/train': 1.436346709728241} 01/29/2022 06:01:16 - INFO - codeparrot_training - Step 13380: {'lr': 0.000433802607039806, 'samples': 2569152, 'steps': 13380, 'loss/train': 1.464594304561615} 01/29/2022 06:01:20 - INFO - codeparrot_training - Step 13381: {'lr': 0.00043379151553244523, 'samples': 2569344, 'steps': 13381, 'loss/train': 2.4089693427085876} 01/29/2022 06:01:24 - INFO - codeparrot_training - Step 13382: {'lr': 0.00043378042323778, 'samples': 2569536, 'steps': 13382, 'loss/train': 2.312623143196106} 01/29/2022 06:01:28 - INFO - codeparrot_training - Step 13383: {'lr': 0.00043376933015585776, 'samples': 2569728, 'steps': 13383, 'loss/train': 1.8470301032066345} 01/29/2022 06:01:34 - INFO - codeparrot_training - Step 13384: {'lr': 0.000433758236286726, 'samples': 2569920, 'steps': 13384, 'loss/train': 1.6067944765090942} 01/29/2022 06:01:38 - INFO - codeparrot_training - Step 13385: {'lr': 0.0004337471416304324, 'samples': 2570112, 'steps': 13385, 'loss/train': 1.6286600232124329} 01/29/2022 06:01:42 - INFO - codeparrot_training - Step 13386: {'lr': 0.00043373604618702436, 'samples': 2570304, 'steps': 13386, 'loss/train': 1.5243008136749268} 01/29/2022 06:01:46 - INFO - codeparrot_training - Step 13387: {'lr': 0.00043372494995654943, 'samples': 2570496, 'steps': 13387, 'loss/train': 1.8581503629684448} 01/29/2022 06:01:51 - INFO - codeparrot_training - Step 13388: {'lr': 0.00043371385293905517, 'samples': 2570688, 'steps': 13388, 'loss/train': 1.6518666744232178} 01/29/2022 06:01:56 - INFO - codeparrot_training - Step 13389: {'lr': 0.0004337027551345891, 'samples': 2570880, 'steps': 13389, 'loss/train': 1.303390771150589} 01/29/2022 06:02:00 - INFO - codeparrot_training - Step 13390: {'lr': 0.0004336916565431987, 'samples': 2571072, 'steps': 13390, 'loss/train': 1.2434244453907013} 01/29/2022 06:02:04 - INFO - codeparrot_training - Step 13391: {'lr': 0.0004336805571649316, 'samples': 2571264, 'steps': 13391, 'loss/train': 2.185740888118744} 01/29/2022 06:02:09 - INFO - codeparrot_training - Step 13392: {'lr': 0.0004336694569998354, 'samples': 2571456, 'steps': 13392, 'loss/train': 1.7196308970451355} 01/29/2022 06:02:13 - INFO - codeparrot_training - Step 13393: {'lr': 0.00043365835604795746, 'samples': 2571648, 'steps': 13393, 'loss/train': 1.2907069623470306} 01/29/2022 06:02:20 - INFO - codeparrot_training - Step 13394: {'lr': 0.0004336472543093455, 'samples': 2571840, 'steps': 13394, 'loss/train': 1.331289678812027} 01/29/2022 06:02:24 - INFO - codeparrot_training - Step 13395: {'lr': 0.000433636151784047, 'samples': 2572032, 'steps': 13395, 'loss/train': 2.085209548473358} 01/29/2022 06:02:28 - INFO - codeparrot_training - Step 13396: {'lr': 0.00043362504847210956, 'samples': 2572224, 'steps': 13396, 'loss/train': 1.2931822836399078} 01/29/2022 06:02:33 - INFO - codeparrot_training - Step 13397: {'lr': 0.0004336139443735807, 'samples': 2572416, 'steps': 13397, 'loss/train': 2.059756636619568} 01/29/2022 06:02:37 - INFO - codeparrot_training - Step 13398: {'lr': 0.000433602839488508, 'samples': 2572608, 'steps': 13398, 'loss/train': 1.1624458730220795} 01/29/2022 06:02:42 - INFO - codeparrot_training - Step 13399: {'lr': 0.00043359173381693906, 'samples': 2572800, 'steps': 13399, 'loss/train': 1.4983878135681152} 01/29/2022 06:02:46 - INFO - codeparrot_training - Step 13400: {'lr': 0.0004335806273589214, 'samples': 2572992, 'steps': 13400, 'loss/train': 1.6892364621162415} 01/29/2022 06:02:51 - INFO - codeparrot_training - Step 13401: {'lr': 0.00043356952011450265, 'samples': 2573184, 'steps': 13401, 'loss/train': 2.1278207302093506} 01/29/2022 06:02:55 - INFO - codeparrot_training - Step 13402: {'lr': 0.0004335584120837304, 'samples': 2573376, 'steps': 13402, 'loss/train': 0.8964771330356598} 01/29/2022 06:02:59 - INFO - codeparrot_training - Step 13403: {'lr': 0.0004335473032666521, 'samples': 2573568, 'steps': 13403, 'loss/train': 1.6632254719734192} 01/29/2022 06:03:07 - INFO - codeparrot_training - Step 13404: {'lr': 0.00043353619366331546, 'samples': 2573760, 'steps': 13404, 'loss/train': 0.7194614410400391} 01/29/2022 06:03:11 - INFO - codeparrot_training - Step 13405: {'lr': 0.0004335250832737681, 'samples': 2573952, 'steps': 13405, 'loss/train': 2.458836793899536} 01/29/2022 06:03:15 - INFO - codeparrot_training - Step 13406: {'lr': 0.00043351397209805755, 'samples': 2574144, 'steps': 13406, 'loss/train': 1.8921185731887817} 01/29/2022 06:03:19 - INFO - codeparrot_training - Step 13407: {'lr': 0.0004335028601362314, 'samples': 2574336, 'steps': 13407, 'loss/train': 1.6001649498939514} 01/29/2022 06:03:24 - INFO - codeparrot_training - Step 13408: {'lr': 0.0004334917473883373, 'samples': 2574528, 'steps': 13408, 'loss/train': 1.8024703860282898} 01/29/2022 06:03:29 - INFO - codeparrot_training - Step 13409: {'lr': 0.0004334806338544227, 'samples': 2574720, 'steps': 13409, 'loss/train': 2.1265307664871216} 01/29/2022 06:03:33 - INFO - codeparrot_training - Step 13410: {'lr': 0.0004334695195345355, 'samples': 2574912, 'steps': 13410, 'loss/train': 2.194754183292389} 01/29/2022 06:03:37 - INFO - codeparrot_training - Step 13411: {'lr': 0.000433458404428723, 'samples': 2575104, 'steps': 13411, 'loss/train': 1.7375510931015015} 01/29/2022 06:03:42 - INFO - codeparrot_training - Step 13412: {'lr': 0.00043344728853703297, 'samples': 2575296, 'steps': 13412, 'loss/train': 1.94101220369339} 01/29/2022 06:03:46 - INFO - codeparrot_training - Step 13413: {'lr': 0.00043343617185951305, 'samples': 2575488, 'steps': 13413, 'loss/train': 1.4047406315803528} 01/29/2022 06:03:51 - INFO - codeparrot_training - Step 13414: {'lr': 0.0004334250543962108, 'samples': 2575680, 'steps': 13414, 'loss/train': 1.7194390296936035} 01/29/2022 06:03:55 - INFO - codeparrot_training - Step 13415: {'lr': 0.00043341393614717384, 'samples': 2575872, 'steps': 13415, 'loss/train': 1.067373275756836} 01/29/2022 06:04:00 - INFO - codeparrot_training - Step 13416: {'lr': 0.0004334028171124499, 'samples': 2576064, 'steps': 13416, 'loss/train': 2.1289742588996887} 01/29/2022 06:04:04 - INFO - codeparrot_training - Step 13417: {'lr': 0.0004333916972920864, 'samples': 2576256, 'steps': 13417, 'loss/train': 1.9166988730430603} 01/29/2022 06:04:08 - INFO - codeparrot_training - Step 13418: {'lr': 0.00043338057668613117, 'samples': 2576448, 'steps': 13418, 'loss/train': 1.3155131936073303} 01/29/2022 06:04:14 - INFO - codeparrot_training - Step 13419: {'lr': 0.00043336945529463177, 'samples': 2576640, 'steps': 13419, 'loss/train': 1.8582252860069275} 01/29/2022 06:04:18 - INFO - codeparrot_training - Step 13420: {'lr': 0.00043335833311763597, 'samples': 2576832, 'steps': 13420, 'loss/train': 2.0896185636520386} 01/29/2022 06:04:22 - INFO - codeparrot_training - Step 13421: {'lr': 0.00043334721015519115, 'samples': 2577024, 'steps': 13421, 'loss/train': 1.7799904346466064} 01/29/2022 06:04:26 - INFO - codeparrot_training - Step 13422: {'lr': 0.00043333608640734513, 'samples': 2577216, 'steps': 13422, 'loss/train': 1.6240689754486084} 01/29/2022 06:04:31 - INFO - codeparrot_training - Step 13423: {'lr': 0.0004333249618741455, 'samples': 2577408, 'steps': 13423, 'loss/train': 1.5186867713928223} 01/29/2022 06:04:37 - INFO - codeparrot_training - Step 13424: {'lr': 0.00043331383655564003, 'samples': 2577600, 'steps': 13424, 'loss/train': 1.347746729850769} 01/29/2022 06:04:41 - INFO - codeparrot_training - Step 13425: {'lr': 0.0004333027104518762, 'samples': 2577792, 'steps': 13425, 'loss/train': 1.7476714253425598} 01/29/2022 06:04:45 - INFO - codeparrot_training - Step 13426: {'lr': 0.00043329158356290187, 'samples': 2577984, 'steps': 13426, 'loss/train': 2.0219748616218567} 01/29/2022 06:04:49 - INFO - codeparrot_training - Step 13427: {'lr': 0.00043328045588876454, 'samples': 2578176, 'steps': 13427, 'loss/train': 3.0124175548553467} 01/29/2022 06:04:54 - INFO - codeparrot_training - Step 13428: {'lr': 0.0004332693274295119, 'samples': 2578368, 'steps': 13428, 'loss/train': 1.5210460424423218} 01/29/2022 06:05:00 - INFO - codeparrot_training - Step 13429: {'lr': 0.0004332581981851917, 'samples': 2578560, 'steps': 13429, 'loss/train': 1.4307059347629547} 01/29/2022 06:05:04 - INFO - codeparrot_training - Step 13430: {'lr': 0.00043324706815585156, 'samples': 2578752, 'steps': 13430, 'loss/train': 1.9695634245872498} 01/29/2022 06:05:09 - INFO - codeparrot_training - Step 13431: {'lr': 0.00043323593734153915, 'samples': 2578944, 'steps': 13431, 'loss/train': 1.935002088546753} 01/29/2022 06:05:13 - INFO - codeparrot_training - Step 13432: {'lr': 0.00043322480574230215, 'samples': 2579136, 'steps': 13432, 'loss/train': 1.6659804582595825} 01/29/2022 06:05:17 - INFO - codeparrot_training - Step 13433: {'lr': 0.00043321367335818833, 'samples': 2579328, 'steps': 13433, 'loss/train': 1.1773436665534973} 01/29/2022 06:05:21 - INFO - codeparrot_training - Step 13434: {'lr': 0.0004332025401892453, 'samples': 2579520, 'steps': 13434, 'loss/train': 2.045417904853821} 01/29/2022 06:05:26 - INFO - codeparrot_training - Step 13435: {'lr': 0.00043319140623552073, 'samples': 2579712, 'steps': 13435, 'loss/train': 1.5127032995224} 01/29/2022 06:05:31 - INFO - codeparrot_training - Step 13436: {'lr': 0.0004331802714970624, 'samples': 2579904, 'steps': 13436, 'loss/train': 2.94064199924469} 01/29/2022 06:05:35 - INFO - codeparrot_training - Step 13437: {'lr': 0.00043316913597391785, 'samples': 2580096, 'steps': 13437, 'loss/train': 1.4379460215568542} 01/29/2022 06:05:39 - INFO - codeparrot_training - Step 13438: {'lr': 0.00043315799966613496, 'samples': 2580288, 'steps': 13438, 'loss/train': 0.15324795618653297} 01/29/2022 06:05:43 - INFO - codeparrot_training - Step 13439: {'lr': 0.00043314686257376136, 'samples': 2580480, 'steps': 13439, 'loss/train': 1.8208526372909546} 01/29/2022 06:05:49 - INFO - codeparrot_training - Step 13440: {'lr': 0.0004331357246968447, 'samples': 2580672, 'steps': 13440, 'loss/train': 1.5997624397277832} 01/29/2022 06:05:54 - INFO - codeparrot_training - Step 13441: {'lr': 0.0004331245860354328, 'samples': 2580864, 'steps': 13441, 'loss/train': 1.55361807346344} 01/29/2022 06:05:58 - INFO - codeparrot_training - Step 13442: {'lr': 0.0004331134465895733, 'samples': 2581056, 'steps': 13442, 'loss/train': 1.195289522409439} 01/29/2022 06:06:02 - INFO - codeparrot_training - Step 13443: {'lr': 0.00043310230635931394, 'samples': 2581248, 'steps': 13443, 'loss/train': 1.7672715783119202} 01/29/2022 06:06:06 - INFO - codeparrot_training - Step 13444: {'lr': 0.0004330911653447024, 'samples': 2581440, 'steps': 13444, 'loss/train': 1.323103129863739} 01/29/2022 06:06:12 - INFO - codeparrot_training - Step 13445: {'lr': 0.0004330800235457866, 'samples': 2581632, 'steps': 13445, 'loss/train': 1.8142719268798828} 01/29/2022 06:06:16 - INFO - codeparrot_training - Step 13446: {'lr': 0.00043306888096261394, 'samples': 2581824, 'steps': 13446, 'loss/train': 0.6032689958810806} 01/29/2022 06:06:20 - INFO - codeparrot_training - Step 13447: {'lr': 0.0004330577375952324, 'samples': 2582016, 'steps': 13447, 'loss/train': 1.8730738162994385} 01/29/2022 06:06:24 - INFO - codeparrot_training - Step 13448: {'lr': 0.0004330465934436896, 'samples': 2582208, 'steps': 13448, 'loss/train': 2.1040669083595276} 01/29/2022 06:06:29 - INFO - codeparrot_training - Step 13449: {'lr': 0.0004330354485080334, 'samples': 2582400, 'steps': 13449, 'loss/train': 7.317395210266113} 01/29/2022 06:06:35 - INFO - codeparrot_training - Step 13450: {'lr': 0.0004330243027883114, 'samples': 2582592, 'steps': 13450, 'loss/train': 1.3808358907699585} 01/29/2022 06:06:39 - INFO - codeparrot_training - Step 13451: {'lr': 0.0004330131562845714, 'samples': 2582784, 'steps': 13451, 'loss/train': 1.5401079654693604} 01/29/2022 06:06:43 - INFO - codeparrot_training - Step 13452: {'lr': 0.00043300200899686113, 'samples': 2582976, 'steps': 13452, 'loss/train': 1.2005772292613983} 01/29/2022 06:06:48 - INFO - codeparrot_training - Step 13453: {'lr': 0.0004329908609252284, 'samples': 2583168, 'steps': 13453, 'loss/train': 1.294623702764511} 01/29/2022 06:06:52 - INFO - codeparrot_training - Step 13454: {'lr': 0.00043297971206972095, 'samples': 2583360, 'steps': 13454, 'loss/train': 1.725761353969574} 01/29/2022 06:06:56 - INFO - codeparrot_training - Step 13455: {'lr': 0.0004329685624303865, 'samples': 2583552, 'steps': 13455, 'loss/train': 2.247318148612976} 01/29/2022 06:07:02 - INFO - codeparrot_training - Step 13456: {'lr': 0.0004329574120072728, 'samples': 2583744, 'steps': 13456, 'loss/train': 1.841443955898285} 01/29/2022 06:07:06 - INFO - codeparrot_training - Step 13457: {'lr': 0.00043294626080042767, 'samples': 2583936, 'steps': 13457, 'loss/train': 1.8257696628570557} 01/29/2022 06:07:10 - INFO - codeparrot_training - Step 13458: {'lr': 0.0004329351088098988, 'samples': 2584128, 'steps': 13458, 'loss/train': 2.2761247158050537} 01/29/2022 06:07:15 - INFO - codeparrot_training - Step 13459: {'lr': 0.0004329239560357341, 'samples': 2584320, 'steps': 13459, 'loss/train': 2.1938758492469788} 01/29/2022 06:07:20 - INFO - codeparrot_training - Step 13460: {'lr': 0.0004329128024779812, 'samples': 2584512, 'steps': 13460, 'loss/train': 1.5645613074302673} 01/29/2022 06:07:24 - INFO - codeparrot_training - Step 13461: {'lr': 0.00043290164813668795, 'samples': 2584704, 'steps': 13461, 'loss/train': 2.1877214312553406} 01/29/2022 06:07:28 - INFO - codeparrot_training - Step 13462: {'lr': 0.0004328904930119021, 'samples': 2584896, 'steps': 13462, 'loss/train': 1.7435038089752197} 01/29/2022 06:07:33 - INFO - codeparrot_training - Step 13463: {'lr': 0.0004328793371036714, 'samples': 2585088, 'steps': 13463, 'loss/train': 1.7626590728759766} 01/29/2022 06:07:37 - INFO - codeparrot_training - Step 13464: {'lr': 0.0004328681804120438, 'samples': 2585280, 'steps': 13464, 'loss/train': 1.5823693871498108} 01/29/2022 06:07:43 - INFO - codeparrot_training - Step 13465: {'lr': 0.000432857022937067, 'samples': 2585472, 'steps': 13465, 'loss/train': 2.349624752998352} 01/29/2022 06:07:47 - INFO - codeparrot_training - Step 13466: {'lr': 0.00043284586467878865, 'samples': 2585664, 'steps': 13466, 'loss/train': 1.9168020486831665} 01/29/2022 06:07:52 - INFO - codeparrot_training - Step 13467: {'lr': 0.0004328347056372568, 'samples': 2585856, 'steps': 13467, 'loss/train': 3.225254774093628} 01/29/2022 06:07:56 - INFO - codeparrot_training - Step 13468: {'lr': 0.00043282354581251903, 'samples': 2586048, 'steps': 13468, 'loss/train': 1.8328925371170044} 01/29/2022 06:08:00 - INFO - codeparrot_training - Step 13469: {'lr': 0.0004328123852046233, 'samples': 2586240, 'steps': 13469, 'loss/train': 2.175078570842743} 01/29/2022 06:08:05 - INFO - codeparrot_training - Step 13470: {'lr': 0.0004328012238136173, 'samples': 2586432, 'steps': 13470, 'loss/train': 1.8343266248703003} 01/29/2022 06:08:10 - INFO - codeparrot_training - Step 13471: {'lr': 0.000432790061639549, 'samples': 2586624, 'steps': 13471, 'loss/train': 1.5816328525543213} 01/29/2022 06:08:14 - INFO - codeparrot_training - Step 13472: {'lr': 0.00043277889868246605, 'samples': 2586816, 'steps': 13472, 'loss/train': 1.5078045129776} 01/29/2022 06:08:18 - INFO - codeparrot_training - Step 13473: {'lr': 0.0004327677349424164, 'samples': 2587008, 'steps': 13473, 'loss/train': 1.8954359292984009} 01/29/2022 06:08:22 - INFO - codeparrot_training - Step 13474: {'lr': 0.0004327565704194477, 'samples': 2587200, 'steps': 13474, 'loss/train': 1.848997950553894} 01/29/2022 06:08:27 - INFO - codeparrot_training - Step 13475: {'lr': 0.0004327454051136079, 'samples': 2587392, 'steps': 13475, 'loss/train': 1.4690422117710114} 01/29/2022 06:08:32 - INFO - codeparrot_training - Step 13476: {'lr': 0.0004327342390249449, 'samples': 2587584, 'steps': 13476, 'loss/train': 1.7195504307746887} 01/29/2022 06:08:36 - INFO - codeparrot_training - Step 13477: {'lr': 0.00043272307215350635, 'samples': 2587776, 'steps': 13477, 'loss/train': 1.8272838592529297} 01/29/2022 06:08:40 - INFO - codeparrot_training - Step 13478: {'lr': 0.0004327119044993403, 'samples': 2587968, 'steps': 13478, 'loss/train': 2.0143961906433105} 01/29/2022 06:08:44 - INFO - codeparrot_training - Step 13479: {'lr': 0.0004327007360624944, 'samples': 2588160, 'steps': 13479, 'loss/train': 2.121809184551239} 01/29/2022 06:08:50 - INFO - codeparrot_training - Step 13480: {'lr': 0.0004326895668430165, 'samples': 2588352, 'steps': 13480, 'loss/train': 1.938064992427826} 01/29/2022 06:08:54 - INFO - codeparrot_training - Step 13481: {'lr': 0.0004326783968409546, 'samples': 2588544, 'steps': 13481, 'loss/train': 1.6725304126739502} 01/29/2022 06:08:58 - INFO - codeparrot_training - Step 13482: {'lr': 0.00043266722605635644, 'samples': 2588736, 'steps': 13482, 'loss/train': 1.8717557787895203} 01/29/2022 06:09:02 - INFO - codeparrot_training - Step 13483: {'lr': 0.0004326560544892699, 'samples': 2588928, 'steps': 13483, 'loss/train': 1.4960118234157562} 01/29/2022 06:09:07 - INFO - codeparrot_training - Step 13484: {'lr': 0.00043264488213974275, 'samples': 2589120, 'steps': 13484, 'loss/train': 0.9554436206817627} 01/29/2022 06:09:13 - INFO - codeparrot_training - Step 13485: {'lr': 0.00043263370900782297, 'samples': 2589312, 'steps': 13485, 'loss/train': 1.4541750848293304} 01/29/2022 06:09:17 - INFO - codeparrot_training - Step 13486: {'lr': 0.0004326225350935583, 'samples': 2589504, 'steps': 13486, 'loss/train': 1.98480623960495} 01/29/2022 06:09:21 - INFO - codeparrot_training - Step 13487: {'lr': 0.00043261136039699676, 'samples': 2589696, 'steps': 13487, 'loss/train': 0.8502643704414368} 01/29/2022 06:09:26 - INFO - codeparrot_training - Step 13488: {'lr': 0.0004326001849181862, 'samples': 2589888, 'steps': 13488, 'loss/train': 0.9912834763526917} 01/29/2022 06:09:30 - INFO - codeparrot_training - Step 13489: {'lr': 0.0004325890086571743, 'samples': 2590080, 'steps': 13489, 'loss/train': 0.34227872639894485} 01/29/2022 06:09:35 - INFO - codeparrot_training - Step 13490: {'lr': 0.00043257783161400917, 'samples': 2590272, 'steps': 13490, 'loss/train': 1.69527268409729} 01/29/2022 06:09:39 - INFO - codeparrot_training - Step 13491: {'lr': 0.0004325666537887385, 'samples': 2590464, 'steps': 13491, 'loss/train': 1.0391930937767029} 01/29/2022 06:09:44 - INFO - codeparrot_training - Step 13492: {'lr': 0.00043255547518141033, 'samples': 2590656, 'steps': 13492, 'loss/train': 2.36600661277771} 01/29/2022 06:09:48 - INFO - codeparrot_training - Step 13493: {'lr': 0.0004325442957920724, 'samples': 2590848, 'steps': 13493, 'loss/train': 1.7008284330368042} 01/29/2022 06:09:52 - INFO - codeparrot_training - Step 13494: {'lr': 0.0004325331156207727, 'samples': 2591040, 'steps': 13494, 'loss/train': 1.5375943779945374} 01/29/2022 06:09:58 - INFO - codeparrot_training - Step 13495: {'lr': 0.00043252193466755906, 'samples': 2591232, 'steps': 13495, 'loss/train': 2.182510793209076} 01/29/2022 06:10:02 - INFO - codeparrot_training - Step 13496: {'lr': 0.0004325107529324795, 'samples': 2591424, 'steps': 13496, 'loss/train': 1.3888937830924988} 01/29/2022 06:10:07 - INFO - codeparrot_training - Step 13497: {'lr': 0.0004324995704155817, 'samples': 2591616, 'steps': 13497, 'loss/train': 1.303226888179779} 01/29/2022 06:10:11 - INFO - codeparrot_training - Step 13498: {'lr': 0.0004324883871169138, 'samples': 2591808, 'steps': 13498, 'loss/train': 1.36032235622406} 01/29/2022 06:10:16 - INFO - codeparrot_training - Step 13499: {'lr': 0.00043247720303652353, 'samples': 2592000, 'steps': 13499, 'loss/train': 1.126018077135086} 01/29/2022 06:10:20 - INFO - codeparrot_training - Step 13500: {'lr': 0.0004324660181744589, 'samples': 2592192, 'steps': 13500, 'loss/train': 2.710191249847412} 01/29/2022 06:10:25 - INFO - codeparrot_training - Step 13501: {'lr': 0.00043245483253076777, 'samples': 2592384, 'steps': 13501, 'loss/train': 1.8080928325653076} 01/29/2022 06:10:29 - INFO - codeparrot_training - Step 13502: {'lr': 0.0004324436461054981, 'samples': 2592576, 'steps': 13502, 'loss/train': 1.7511863708496094} 01/29/2022 06:10:33 - INFO - codeparrot_training - Step 13503: {'lr': 0.00043243245889869775, 'samples': 2592768, 'steps': 13503, 'loss/train': 2.2789045572280884} 01/29/2022 06:10:38 - INFO - codeparrot_training - Step 13504: {'lr': 0.0004324212709104147, 'samples': 2592960, 'steps': 13504, 'loss/train': 2.407231092453003} 01/29/2022 06:10:43 - INFO - codeparrot_training - Step 13505: {'lr': 0.0004324100821406969, 'samples': 2593152, 'steps': 13505, 'loss/train': 1.7705631852149963} 01/29/2022 06:10:47 - INFO - codeparrot_training - Step 13506: {'lr': 0.00043239889258959215, 'samples': 2593344, 'steps': 13506, 'loss/train': 1.3696051239967346} 01/29/2022 06:10:51 - INFO - codeparrot_training - Step 13507: {'lr': 0.00043238770225714854, 'samples': 2593536, 'steps': 13507, 'loss/train': 1.7051417827606201} 01/29/2022 06:10:55 - INFO - codeparrot_training - Step 13508: {'lr': 0.00043237651114341383, 'samples': 2593728, 'steps': 13508, 'loss/train': 1.4025695621967316} 01/29/2022 06:11:01 - INFO - codeparrot_training - Step 13509: {'lr': 0.0004323653192484361, 'samples': 2593920, 'steps': 13509, 'loss/train': 2.0961434841156006} 01/29/2022 06:11:05 - INFO - codeparrot_training - Step 13510: {'lr': 0.0004323541265722633, 'samples': 2594112, 'steps': 13510, 'loss/train': 1.7324817180633545} 01/29/2022 06:11:10 - INFO - codeparrot_training - Step 13511: {'lr': 0.0004323429331149432, 'samples': 2594304, 'steps': 13511, 'loss/train': 1.9700767993927002} 01/29/2022 06:11:14 - INFO - codeparrot_training - Step 13512: {'lr': 0.000432331738876524, 'samples': 2594496, 'steps': 13512, 'loss/train': 1.7256453037261963} 01/29/2022 06:11:18 - INFO - codeparrot_training - Step 13513: {'lr': 0.00043232054385705345, 'samples': 2594688, 'steps': 13513, 'loss/train': 0.5854919403791428} 01/29/2022 06:11:23 - INFO - codeparrot_training - Step 13514: {'lr': 0.0004323093480565796, 'samples': 2594880, 'steps': 13514, 'loss/train': 2.2740649580955505} 01/29/2022 06:11:28 - INFO - codeparrot_training - Step 13515: {'lr': 0.0004322981514751504, 'samples': 2595072, 'steps': 13515, 'loss/train': 1.4587741792201996} 01/29/2022 06:11:32 - INFO - codeparrot_training - Step 13516: {'lr': 0.0004322869541128138, 'samples': 2595264, 'steps': 13516, 'loss/train': 1.2585226893424988} 01/29/2022 06:11:36 - INFO - codeparrot_training - Step 13517: {'lr': 0.00043227575596961783, 'samples': 2595456, 'steps': 13517, 'loss/train': 0.2049030214548111} 01/29/2022 06:11:40 - INFO - codeparrot_training - Step 13518: {'lr': 0.00043226455704561034, 'samples': 2595648, 'steps': 13518, 'loss/train': 1.3150463998317719} 01/29/2022 06:11:46 - INFO - codeparrot_training - Step 13519: {'lr': 0.0004322533573408394, 'samples': 2595840, 'steps': 13519, 'loss/train': 0.7882022559642792} 01/29/2022 06:11:51 - INFO - codeparrot_training - Step 13520: {'lr': 0.00043224215685535287, 'samples': 2596032, 'steps': 13520, 'loss/train': 0.31803736835718155} 01/29/2022 06:11:55 - INFO - codeparrot_training - Step 13521: {'lr': 0.0004322309555891989, 'samples': 2596224, 'steps': 13521, 'loss/train': 1.695698082447052} 01/29/2022 06:11:59 - INFO - codeparrot_training - Step 13522: {'lr': 0.00043221975354242536, 'samples': 2596416, 'steps': 13522, 'loss/train': 1.1789030134677887} 01/29/2022 06:12:03 - INFO - codeparrot_training - Step 13523: {'lr': 0.0004322085507150802, 'samples': 2596608, 'steps': 13523, 'loss/train': 1.400271713733673} 01/29/2022 06:12:09 - INFO - codeparrot_training - Step 13524: {'lr': 0.00043219734710721146, 'samples': 2596800, 'steps': 13524, 'loss/train': 1.371074616909027} 01/29/2022 06:12:13 - INFO - codeparrot_training - Step 13525: {'lr': 0.00043218614271886725, 'samples': 2596992, 'steps': 13525, 'loss/train': 1.8910725116729736} 01/29/2022 06:12:17 - INFO - codeparrot_training - Step 13526: {'lr': 0.0004321749375500954, 'samples': 2597184, 'steps': 13526, 'loss/train': 1.5574691891670227} 01/29/2022 06:12:21 - INFO - codeparrot_training - Step 13527: {'lr': 0.0004321637316009439, 'samples': 2597376, 'steps': 13527, 'loss/train': 2.7561798691749573} 01/29/2022 06:12:26 - INFO - codeparrot_training - Step 13528: {'lr': 0.00043215252487146096, 'samples': 2597568, 'steps': 13528, 'loss/train': 1.3916254341602325} 01/29/2022 06:12:31 - INFO - codeparrot_training - Step 13529: {'lr': 0.0004321413173616943, 'samples': 2597760, 'steps': 13529, 'loss/train': 2.2654314637184143} 01/29/2022 06:12:35 - INFO - codeparrot_training - Step 13530: {'lr': 0.00043213010907169213, 'samples': 2597952, 'steps': 13530, 'loss/train': 1.5493465662002563} 01/29/2022 06:12:39 - INFO - codeparrot_training - Step 13531: {'lr': 0.00043211890000150247, 'samples': 2598144, 'steps': 13531, 'loss/train': 1.6437529921531677} 01/29/2022 06:12:44 - INFO - codeparrot_training - Step 13532: {'lr': 0.0004321076901511731, 'samples': 2598336, 'steps': 13532, 'loss/train': 2.0402913093566895} 01/29/2022 06:12:48 - INFO - codeparrot_training - Step 13533: {'lr': 0.00043209647952075235, 'samples': 2598528, 'steps': 13533, 'loss/train': 1.6395122408866882} 01/29/2022 06:12:53 - INFO - codeparrot_training - Step 13534: {'lr': 0.00043208526811028806, 'samples': 2598720, 'steps': 13534, 'loss/train': 2.0522201657295227} 01/29/2022 06:12:57 - INFO - codeparrot_training - Step 13535: {'lr': 0.00043207405591982835, 'samples': 2598912, 'steps': 13535, 'loss/train': 1.8481587767601013} 01/29/2022 06:13:02 - INFO - codeparrot_training - Step 13536: {'lr': 0.0004320628429494212, 'samples': 2599104, 'steps': 13536, 'loss/train': 1.1885110437870026} 01/29/2022 06:13:06 - INFO - codeparrot_training - Step 13537: {'lr': 0.00043205162919911455, 'samples': 2599296, 'steps': 13537, 'loss/train': 1.379797250032425} 01/29/2022 06:13:10 - INFO - codeparrot_training - Step 13538: {'lr': 0.0004320404146689566, 'samples': 2599488, 'steps': 13538, 'loss/train': 1.7951679825782776} 01/29/2022 06:13:15 - INFO - codeparrot_training - Step 13539: {'lr': 0.0004320291993589953, 'samples': 2599680, 'steps': 13539, 'loss/train': 2.0083236694335938} 01/29/2022 06:13:20 - INFO - codeparrot_training - Step 13540: {'lr': 0.0004320179832692787, 'samples': 2599872, 'steps': 13540, 'loss/train': 2.5087791681289673} 01/29/2022 06:13:24 - INFO - codeparrot_training - Step 13541: {'lr': 0.0004320067663998549, 'samples': 2600064, 'steps': 13541, 'loss/train': 0.8344263732433319} 01/29/2022 06:13:28 - INFO - codeparrot_training - Step 13542: {'lr': 0.00043199554875077183, 'samples': 2600256, 'steps': 13542, 'loss/train': 0.28624773770570755} 01/29/2022 06:13:32 - INFO - codeparrot_training - Step 13543: {'lr': 0.00043198433032207774, 'samples': 2600448, 'steps': 13543, 'loss/train': 1.850756049156189} 01/29/2022 06:13:38 - INFO - codeparrot_training - Step 13544: {'lr': 0.00043197311111382045, 'samples': 2600640, 'steps': 13544, 'loss/train': 1.7786337733268738} 01/29/2022 06:13:43 - INFO - codeparrot_training - Step 13545: {'lr': 0.0004319618911260482, 'samples': 2600832, 'steps': 13545, 'loss/train': 1.784429132938385} 01/29/2022 06:13:47 - INFO - codeparrot_training - Step 13546: {'lr': 0.0004319506703588089, 'samples': 2601024, 'steps': 13546, 'loss/train': 1.6882216930389404} 01/29/2022 06:13:51 - INFO - codeparrot_training - Step 13547: {'lr': 0.00043193944881215075, 'samples': 2601216, 'steps': 13547, 'loss/train': 1.5433248281478882} 01/29/2022 06:13:55 - INFO - codeparrot_training - Step 13548: {'lr': 0.00043192822648612184, 'samples': 2601408, 'steps': 13548, 'loss/train': 0.8608026802539825} 01/29/2022 06:14:01 - INFO - codeparrot_training - Step 13549: {'lr': 0.0004319170033807701, 'samples': 2601600, 'steps': 13549, 'loss/train': 1.934524118900299} 01/29/2022 06:14:05 - INFO - codeparrot_training - Step 13550: {'lr': 0.00043190577949614375, 'samples': 2601792, 'steps': 13550, 'loss/train': 0.04593568481504917} 01/29/2022 06:14:09 - INFO - codeparrot_training - Step 13551: {'lr': 0.00043189455483229073, 'samples': 2601984, 'steps': 13551, 'loss/train': 1.5704535841941833} 01/29/2022 06:14:14 - INFO - codeparrot_training - Step 13552: {'lr': 0.00043188332938925923, 'samples': 2602176, 'steps': 13552, 'loss/train': 1.1831722855567932} 01/29/2022 06:14:18 - INFO - codeparrot_training - Step 13553: {'lr': 0.0004318721031670973, 'samples': 2602368, 'steps': 13553, 'loss/train': 1.0118553042411804} 01/29/2022 06:14:24 - INFO - codeparrot_training - Step 13554: {'lr': 0.00043186087616585303, 'samples': 2602560, 'steps': 13554, 'loss/train': 1.2040721476078033} 01/29/2022 06:14:28 - INFO - codeparrot_training - Step 13555: {'lr': 0.0004318496483855745, 'samples': 2602752, 'steps': 13555, 'loss/train': 1.2692490220069885} 01/29/2022 06:14:33 - INFO - codeparrot_training - Step 13556: {'lr': 0.0004318384198263099, 'samples': 2602944, 'steps': 13556, 'loss/train': 1.2660733759403229} 01/29/2022 06:14:37 - INFO - codeparrot_training - Step 13557: {'lr': 0.00043182719048810714, 'samples': 2603136, 'steps': 13557, 'loss/train': 1.4490635991096497} 01/29/2022 06:14:41 - INFO - codeparrot_training - Step 13558: {'lr': 0.00043181596037101443, 'samples': 2603328, 'steps': 13558, 'loss/train': 2.5860167741775513} 01/29/2022 06:14:47 - INFO - codeparrot_training - Step 13559: {'lr': 0.00043180472947508, 'samples': 2603520, 'steps': 13559, 'loss/train': 1.2692659199237823} 01/29/2022 06:14:51 - INFO - codeparrot_training - Step 13560: {'lr': 0.0004317934978003517, 'samples': 2603712, 'steps': 13560, 'loss/train': 1.9695791602134705} 01/29/2022 06:14:55 - INFO - codeparrot_training - Step 13561: {'lr': 0.0004317822653468778, 'samples': 2603904, 'steps': 13561, 'loss/train': 1.7104055285453796} 01/29/2022 06:14:59 - INFO - codeparrot_training - Step 13562: {'lr': 0.00043177103211470647, 'samples': 2604096, 'steps': 13562, 'loss/train': 2.1291292905807495} 01/29/2022 06:15:04 - INFO - codeparrot_training - Step 13563: {'lr': 0.00043175979810388575, 'samples': 2604288, 'steps': 13563, 'loss/train': 1.8893383741378784} 01/29/2022 06:15:09 - INFO - codeparrot_training - Step 13564: {'lr': 0.0004317485633144638, 'samples': 2604480, 'steps': 13564, 'loss/train': 1.5092012286186218} 01/29/2022 06:15:13 - INFO - codeparrot_training - Step 13565: {'lr': 0.0004317373277464886, 'samples': 2604672, 'steps': 13565, 'loss/train': 1.04832261800766} 01/29/2022 06:15:17 - INFO - codeparrot_training - Step 13566: {'lr': 0.0004317260914000085, 'samples': 2604864, 'steps': 13566, 'loss/train': 1.7781193256378174} 01/29/2022 06:15:22 - INFO - codeparrot_training - Step 13567: {'lr': 0.00043171485427507145, 'samples': 2605056, 'steps': 13567, 'loss/train': 1.8803719282150269} 01/29/2022 06:15:26 - INFO - codeparrot_training - Step 13568: {'lr': 0.0004317036163717257, 'samples': 2605248, 'steps': 13568, 'loss/train': 1.235567718744278} 01/29/2022 06:15:32 - INFO - codeparrot_training - Step 13569: {'lr': 0.00043169237769001936, 'samples': 2605440, 'steps': 13569, 'loss/train': 1.6825881600379944} 01/29/2022 06:15:36 - INFO - codeparrot_training - Step 13570: {'lr': 0.0004316811382300006, 'samples': 2605632, 'steps': 13570, 'loss/train': 1.9253783226013184} 01/29/2022 06:15:40 - INFO - codeparrot_training - Step 13571: {'lr': 0.0004316698979917175, 'samples': 2605824, 'steps': 13571, 'loss/train': 1.2269757986068726} 01/29/2022 06:15:44 - INFO - codeparrot_training - Step 13572: {'lr': 0.0004316586569752182, 'samples': 2606016, 'steps': 13572, 'loss/train': 1.4354599714279175} 01/29/2022 06:15:50 - INFO - codeparrot_training - Step 13573: {'lr': 0.00043164741518055097, 'samples': 2606208, 'steps': 13573, 'loss/train': 1.3871847689151764} 01/29/2022 06:15:54 - INFO - codeparrot_training - Step 13574: {'lr': 0.0004316361726077639, 'samples': 2606400, 'steps': 13574, 'loss/train': 1.9572923183441162} 01/29/2022 06:15:58 - INFO - codeparrot_training - Step 13575: {'lr': 0.0004316249292569051, 'samples': 2606592, 'steps': 13575, 'loss/train': 1.6516188383102417} 01/29/2022 06:16:03 - INFO - codeparrot_training - Step 13576: {'lr': 0.0004316136851280228, 'samples': 2606784, 'steps': 13576, 'loss/train': 2.052129328250885} 01/29/2022 06:16:07 - INFO - codeparrot_training - Step 13577: {'lr': 0.00043160244022116514, 'samples': 2606976, 'steps': 13577, 'loss/train': 1.6195971965789795} 01/29/2022 06:16:13 - INFO - codeparrot_training - Step 13578: {'lr': 0.0004315911945363802, 'samples': 2607168, 'steps': 13578, 'loss/train': 1.3266708254814148} 01/29/2022 06:16:17 - INFO - codeparrot_training - Step 13579: {'lr': 0.00043157994807371634, 'samples': 2607360, 'steps': 13579, 'loss/train': 1.2467440366744995} 01/29/2022 06:16:22 - INFO - codeparrot_training - Step 13580: {'lr': 0.00043156870083322166, 'samples': 2607552, 'steps': 13580, 'loss/train': 1.6857159733772278} 01/29/2022 06:16:26 - INFO - codeparrot_training - Step 13581: {'lr': 0.0004315574528149443, 'samples': 2607744, 'steps': 13581, 'loss/train': 0.4871618002653122} 01/29/2022 06:16:30 - INFO - codeparrot_training - Step 13582: {'lr': 0.00043154620401893244, 'samples': 2607936, 'steps': 13582, 'loss/train': 2.06255042552948} 01/29/2022 06:16:34 - INFO - codeparrot_training - Step 13583: {'lr': 0.0004315349544452343, 'samples': 2608128, 'steps': 13583, 'loss/train': 0.9880801141262054} 01/29/2022 06:16:40 - INFO - codeparrot_training - Step 13584: {'lr': 0.00043152370409389794, 'samples': 2608320, 'steps': 13584, 'loss/train': 2.155823528766632} 01/29/2022 06:16:44 - INFO - codeparrot_training - Step 13585: {'lr': 0.00043151245296497184, 'samples': 2608512, 'steps': 13585, 'loss/train': 3.0235469341278076} 01/29/2022 06:16:48 - INFO - codeparrot_training - Step 13586: {'lr': 0.000431501201058504, 'samples': 2608704, 'steps': 13586, 'loss/train': 1.638911247253418} 01/29/2022 06:16:52 - INFO - codeparrot_training - Step 13587: {'lr': 0.0004314899483745426, 'samples': 2608896, 'steps': 13587, 'loss/train': 1.8207646608352661} 01/29/2022 06:16:57 - INFO - codeparrot_training - Step 13588: {'lr': 0.0004314786949131359, 'samples': 2609088, 'steps': 13588, 'loss/train': 1.6279048919677734} 01/29/2022 06:17:02 - INFO - codeparrot_training - Step 13589: {'lr': 0.0004314674406743321, 'samples': 2609280, 'steps': 13589, 'loss/train': 3.299721837043762} 01/29/2022 06:17:06 - INFO - codeparrot_training - Step 13590: {'lr': 0.00043145618565817946, 'samples': 2609472, 'steps': 13590, 'loss/train': 1.3830616772174835} 01/29/2022 06:17:10 - INFO - codeparrot_training - Step 13591: {'lr': 0.00043144492986472603, 'samples': 2609664, 'steps': 13591, 'loss/train': 2.374583065509796} 01/29/2022 06:17:15 - INFO - codeparrot_training - Step 13592: {'lr': 0.0004314336732940202, 'samples': 2609856, 'steps': 13592, 'loss/train': 1.739046335220337} 01/29/2022 06:17:19 - INFO - codeparrot_training - Step 13593: {'lr': 0.0004314224159461102, 'samples': 2610048, 'steps': 13593, 'loss/train': 1.0279978215694427} 01/29/2022 06:17:24 - INFO - codeparrot_training - Step 13594: {'lr': 0.0004314111578210441, 'samples': 2610240, 'steps': 13594, 'loss/train': 1.9301950335502625} 01/29/2022 06:17:28 - INFO - codeparrot_training - Step 13595: {'lr': 0.0004313998989188702, 'samples': 2610432, 'steps': 13595, 'loss/train': 2.1172966361045837} 01/29/2022 06:17:33 - INFO - codeparrot_training - Step 13596: {'lr': 0.00043138863923963664, 'samples': 2610624, 'steps': 13596, 'loss/train': 2.225733697414398} 01/29/2022 06:17:37 - INFO - codeparrot_training - Step 13597: {'lr': 0.0004313773787833919, 'samples': 2610816, 'steps': 13597, 'loss/train': 1.4937330186367035} 01/29/2022 06:17:41 - INFO - codeparrot_training - Step 13598: {'lr': 0.0004313661175501841, 'samples': 2611008, 'steps': 13598, 'loss/train': 1.767982006072998} 01/29/2022 06:17:47 - INFO - codeparrot_training - Step 13599: {'lr': 0.00043135485554006127, 'samples': 2611200, 'steps': 13599, 'loss/train': 2.194265127182007} 01/29/2022 06:17:51 - INFO - codeparrot_training - Step 13600: {'lr': 0.0004313435927530719, 'samples': 2611392, 'steps': 13600, 'loss/train': 2.1584386825561523} 01/29/2022 06:17:55 - INFO - codeparrot_training - Step 13601: {'lr': 0.00043133232918926426, 'samples': 2611584, 'steps': 13601, 'loss/train': 1.5534408688545227} 01/29/2022 06:18:00 - INFO - codeparrot_training - Step 13602: {'lr': 0.0004313210648486864, 'samples': 2611776, 'steps': 13602, 'loss/train': 1.471606582403183} 01/29/2022 06:18:04 - INFO - codeparrot_training - Step 13603: {'lr': 0.00043130979973138664, 'samples': 2611968, 'steps': 13603, 'loss/train': 0.7013121396303177} 01/29/2022 06:18:10 - INFO - codeparrot_training - Step 13604: {'lr': 0.00043129853383741334, 'samples': 2612160, 'steps': 13604, 'loss/train': 1.8838765025138855} 01/29/2022 06:18:14 - INFO - codeparrot_training - Step 13605: {'lr': 0.00043128726716681464, 'samples': 2612352, 'steps': 13605, 'loss/train': 2.090446650981903} 01/29/2022 06:18:18 - INFO - codeparrot_training - Step 13606: {'lr': 0.0004312759997196389, 'samples': 2612544, 'steps': 13606, 'loss/train': 1.3652314245700836} 01/29/2022 06:18:22 - INFO - codeparrot_training - Step 13607: {'lr': 0.00043126473149593424, 'samples': 2612736, 'steps': 13607, 'loss/train': 1.008107990026474} 01/29/2022 06:18:28 - INFO - codeparrot_training - Step 13608: {'lr': 0.00043125346249574915, 'samples': 2612928, 'steps': 13608, 'loss/train': 2.018501579761505} 01/29/2022 06:18:32 - INFO - codeparrot_training - Step 13609: {'lr': 0.0004312421927191318, 'samples': 2613120, 'steps': 13609, 'loss/train': 1.9809603095054626} 01/29/2022 06:18:36 - INFO - codeparrot_training - Step 13610: {'lr': 0.00043123092216613035, 'samples': 2613312, 'steps': 13610, 'loss/train': 2.089411497116089} 01/29/2022 06:18:40 - INFO - codeparrot_training - Step 13611: {'lr': 0.0004312196508367932, 'samples': 2613504, 'steps': 13611, 'loss/train': 1.4908486604690552} 01/29/2022 06:18:45 - INFO - codeparrot_training - Step 13612: {'lr': 0.0004312083787311686, 'samples': 2613696, 'steps': 13612, 'loss/train': 1.7357538342475891} 01/29/2022 06:18:50 - INFO - codeparrot_training - Step 13613: {'lr': 0.0004311971058493049, 'samples': 2613888, 'steps': 13613, 'loss/train': 1.2983563542366028} 01/29/2022 06:18:55 - INFO - codeparrot_training - Step 13614: {'lr': 0.0004311858321912503, 'samples': 2614080, 'steps': 13614, 'loss/train': 1.4988605976104736} 01/29/2022 06:18:59 - INFO - codeparrot_training - Step 13615: {'lr': 0.0004311745577570531, 'samples': 2614272, 'steps': 13615, 'loss/train': 0.9347771108150482} 01/29/2022 06:19:03 - INFO - codeparrot_training - Step 13616: {'lr': 0.0004311632825467617, 'samples': 2614464, 'steps': 13616, 'loss/train': 1.233414351940155} 01/29/2022 06:19:07 - INFO - codeparrot_training - Step 13617: {'lr': 0.00043115200656042426, 'samples': 2614656, 'steps': 13617, 'loss/train': 1.1805562376976013} 01/29/2022 06:19:13 - INFO - codeparrot_training - Step 13618: {'lr': 0.00043114072979808914, 'samples': 2614848, 'steps': 13618, 'loss/train': 1.5285728573799133} 01/29/2022 06:19:17 - INFO - codeparrot_training - Step 13619: {'lr': 0.00043112945225980473, 'samples': 2615040, 'steps': 13619, 'loss/train': 1.691996455192566} 01/29/2022 06:19:21 - INFO - codeparrot_training - Step 13620: {'lr': 0.00043111817394561917, 'samples': 2615232, 'steps': 13620, 'loss/train': 1.6251266598701477} 01/29/2022 06:19:26 - INFO - codeparrot_training - Step 13621: {'lr': 0.0004311068948555809, 'samples': 2615424, 'steps': 13621, 'loss/train': 1.950099527835846} 01/29/2022 06:19:30 - INFO - codeparrot_training - Step 13622: {'lr': 0.0004310956149897382, 'samples': 2615616, 'steps': 13622, 'loss/train': 2.039683520793915} 01/29/2022 06:19:36 - INFO - codeparrot_training - Step 13623: {'lr': 0.00043108433434813943, 'samples': 2615808, 'steps': 13623, 'loss/train': 1.8007603883743286} 01/29/2022 06:19:40 - INFO - codeparrot_training - Step 13624: {'lr': 0.00043107305293083276, 'samples': 2616000, 'steps': 13624, 'loss/train': 1.365618735551834} 01/29/2022 06:19:45 - INFO - codeparrot_training - Step 13625: {'lr': 0.0004310617707378668, 'samples': 2616192, 'steps': 13625, 'loss/train': 1.3689753413200378} 01/29/2022 06:19:49 - INFO - codeparrot_training - Step 13626: {'lr': 0.0004310504877692896, 'samples': 2616384, 'steps': 13626, 'loss/train': 1.2335574924945831} 01/29/2022 06:19:53 - INFO - codeparrot_training - Step 13627: {'lr': 0.00043103920402514956, 'samples': 2616576, 'steps': 13627, 'loss/train': 1.2364375591278076} 01/29/2022 06:19:57 - INFO - codeparrot_training - Step 13628: {'lr': 0.00043102791950549513, 'samples': 2616768, 'steps': 13628, 'loss/train': 1.3045687079429626} 01/29/2022 06:20:03 - INFO - codeparrot_training - Step 13629: {'lr': 0.00043101663421037453, 'samples': 2616960, 'steps': 13629, 'loss/train': 0.7305966317653656} 01/29/2022 06:20:07 - INFO - codeparrot_training - Step 13630: {'lr': 0.00043100534813983617, 'samples': 2617152, 'steps': 13630, 'loss/train': 1.2638610899448395} 01/29/2022 06:20:11 - INFO - codeparrot_training - Step 13631: {'lr': 0.00043099406129392835, 'samples': 2617344, 'steps': 13631, 'loss/train': 2.1995957493782043} 01/29/2022 06:20:16 - INFO - codeparrot_training - Step 13632: {'lr': 0.00043098277367269953, 'samples': 2617536, 'steps': 13632, 'loss/train': 1.1011139452457428} 01/29/2022 06:20:20 - INFO - codeparrot_training - Step 13633: {'lr': 0.0004309714852761979, 'samples': 2617728, 'steps': 13633, 'loss/train': 2.0358084440231323} 01/29/2022 06:20:26 - INFO - codeparrot_training - Step 13634: {'lr': 0.0004309601961044719, 'samples': 2617920, 'steps': 13634, 'loss/train': 1.6475780010223389} 01/29/2022 06:20:30 - INFO - codeparrot_training - Step 13635: {'lr': 0.0004309489061575699, 'samples': 2618112, 'steps': 13635, 'loss/train': 1.466475248336792} 01/29/2022 06:20:34 - INFO - codeparrot_training - Step 13636: {'lr': 0.0004309376154355402, 'samples': 2618304, 'steps': 13636, 'loss/train': 2.748559355735779} 01/29/2022 06:20:38 - INFO - codeparrot_training - Step 13637: {'lr': 0.00043092632393843124, 'samples': 2618496, 'steps': 13637, 'loss/train': 0.49553096294403076} 01/29/2022 06:20:42 - INFO - codeparrot_training - Step 13638: {'lr': 0.00043091503166629136, 'samples': 2618688, 'steps': 13638, 'loss/train': 1.3202498853206635} 01/29/2022 06:20:48 - INFO - codeparrot_training - Step 13639: {'lr': 0.000430903738619169, 'samples': 2618880, 'steps': 13639, 'loss/train': 2.114658772945404} 01/29/2022 06:20:52 - INFO - codeparrot_training - Step 13640: {'lr': 0.00043089244479711233, 'samples': 2619072, 'steps': 13640, 'loss/train': 1.077944666147232} 01/29/2022 06:20:56 - INFO - codeparrot_training - Step 13641: {'lr': 0.00043088115020016994, 'samples': 2619264, 'steps': 13641, 'loss/train': 2.3300063610076904} 01/29/2022 06:21:01 - INFO - codeparrot_training - Step 13642: {'lr': 0.00043086985482839016, 'samples': 2619456, 'steps': 13642, 'loss/train': 1.7049513459205627} 01/29/2022 06:21:05 - INFO - codeparrot_training - Step 13643: {'lr': 0.00043085855868182135, 'samples': 2619648, 'steps': 13643, 'loss/train': 2.0980051159858704} 01/29/2022 06:21:11 - INFO - codeparrot_training - Step 13644: {'lr': 0.0004308472617605118, 'samples': 2619840, 'steps': 13644, 'loss/train': 3.414746403694153} 01/29/2022 06:21:15 - INFO - codeparrot_training - Step 13645: {'lr': 0.00043083596406451015, 'samples': 2620032, 'steps': 13645, 'loss/train': 1.4889006614685059} 01/29/2022 06:21:20 - INFO - codeparrot_training - Step 13646: {'lr': 0.0004308246655938646, 'samples': 2620224, 'steps': 13646, 'loss/train': 0.9087861478328705} 01/29/2022 06:21:24 - INFO - codeparrot_training - Step 13647: {'lr': 0.0004308133663486236, 'samples': 2620416, 'steps': 13647, 'loss/train': 0.2674355283379555} 01/29/2022 06:21:29 - INFO - codeparrot_training - Step 13648: {'lr': 0.00043080206632883553, 'samples': 2620608, 'steps': 13648, 'loss/train': 0.10571680963039398} 01/29/2022 06:21:34 - INFO - codeparrot_training - Step 13649: {'lr': 0.0004307907655345488, 'samples': 2620800, 'steps': 13649, 'loss/train': 2.35772305727005} 01/29/2022 06:21:38 - INFO - codeparrot_training - Step 13650: {'lr': 0.0004307794639658119, 'samples': 2620992, 'steps': 13650, 'loss/train': 1.260413646697998} 01/29/2022 06:21:42 - INFO - codeparrot_training - Step 13651: {'lr': 0.0004307681616226732, 'samples': 2621184, 'steps': 13651, 'loss/train': 1.9304205179214478} 01/29/2022 06:21:46 - INFO - codeparrot_training - Step 13652: {'lr': 0.000430756858505181, 'samples': 2621376, 'steps': 13652, 'loss/train': 1.6068068146705627} 01/29/2022 06:21:52 - INFO - codeparrot_training - Step 13653: {'lr': 0.0004307455546133838, 'samples': 2621568, 'steps': 13653, 'loss/train': 1.5248916149139404} 01/29/2022 06:21:56 - INFO - codeparrot_training - Step 13654: {'lr': 0.00043073424994733014, 'samples': 2621760, 'steps': 13654, 'loss/train': 0.9964984953403473} 01/29/2022 06:22:00 - INFO - codeparrot_training - Step 13655: {'lr': 0.0004307229445070683, 'samples': 2621952, 'steps': 13655, 'loss/train': 2.410698115825653} 01/29/2022 06:22:04 - INFO - codeparrot_training - Step 13656: {'lr': 0.0004307116382926468, 'samples': 2622144, 'steps': 13656, 'loss/train': 1.9658253192901611} 01/29/2022 06:22:09 - INFO - codeparrot_training - Step 13657: {'lr': 0.0004307003313041139, 'samples': 2622336, 'steps': 13657, 'loss/train': 2.30033540725708} 01/29/2022 06:22:15 - INFO - codeparrot_training - Step 13658: {'lr': 0.0004306890235415183, 'samples': 2622528, 'steps': 13658, 'loss/train': 1.800135612487793} 01/29/2022 06:22:19 - INFO - codeparrot_training - Step 13659: {'lr': 0.0004306777150049082, 'samples': 2622720, 'steps': 13659, 'loss/train': 2.502466320991516} 01/29/2022 06:22:23 - INFO - codeparrot_training - Step 13660: {'lr': 0.0004306664056943322, 'samples': 2622912, 'steps': 13660, 'loss/train': 1.393627256155014} 01/29/2022 06:22:27 - INFO - codeparrot_training - Step 13661: {'lr': 0.0004306550956098386, 'samples': 2623104, 'steps': 13661, 'loss/train': 1.0716945827007294} 01/29/2022 06:22:32 - INFO - codeparrot_training - Step 13662: {'lr': 0.000430643784751476, 'samples': 2623296, 'steps': 13662, 'loss/train': 1.5514485239982605} 01/29/2022 06:22:37 - INFO - codeparrot_training - Step 13663: {'lr': 0.0004306324731192929, 'samples': 2623488, 'steps': 13663, 'loss/train': 2.286593019962311} 01/29/2022 06:22:41 - INFO - codeparrot_training - Step 13664: {'lr': 0.00043062116071333745, 'samples': 2623680, 'steps': 13664, 'loss/train': 0.22659782320261002} 01/29/2022 06:22:46 - INFO - codeparrot_training - Step 13665: {'lr': 0.0004306098475336584, 'samples': 2623872, 'steps': 13665, 'loss/train': 1.7812179923057556} 01/29/2022 06:22:50 - INFO - codeparrot_training - Step 13666: {'lr': 0.0004305985335803041, 'samples': 2624064, 'steps': 13666, 'loss/train': 2.299623727798462} 01/29/2022 06:22:54 - INFO - codeparrot_training - Step 13667: {'lr': 0.000430587218853323, 'samples': 2624256, 'steps': 13667, 'loss/train': 0.4851318597793579} 01/29/2022 06:23:00 - INFO - codeparrot_training - Step 13668: {'lr': 0.0004305759033527636, 'samples': 2624448, 'steps': 13668, 'loss/train': 1.7568050622940063} 01/29/2022 06:23:04 - INFO - codeparrot_training - Step 13669: {'lr': 0.0004305645870786744, 'samples': 2624640, 'steps': 13669, 'loss/train': 1.296150952577591} 01/29/2022 06:23:09 - INFO - codeparrot_training - Step 13670: {'lr': 0.00043055327003110384, 'samples': 2624832, 'steps': 13670, 'loss/train': 1.8159543871879578} 01/29/2022 06:23:13 - INFO - codeparrot_training - Step 13671: {'lr': 0.00043054195221010037, 'samples': 2625024, 'steps': 13671, 'loss/train': 1.6416041851043701} 01/29/2022 06:23:17 - INFO - codeparrot_training - Step 13672: {'lr': 0.00043053063361571256, 'samples': 2625216, 'steps': 13672, 'loss/train': 1.8470125794410706} 01/29/2022 06:23:22 - INFO - codeparrot_training - Step 13673: {'lr': 0.0004305193142479888, 'samples': 2625408, 'steps': 13673, 'loss/train': 1.1163547933101654} 01/29/2022 06:23:26 - INFO - codeparrot_training - Step 13674: {'lr': 0.0004305079941069776, 'samples': 2625600, 'steps': 13674, 'loss/train': 1.5328956842422485} 01/29/2022 06:23:31 - INFO - codeparrot_training - Step 13675: {'lr': 0.0004304966731927276, 'samples': 2625792, 'steps': 13675, 'loss/train': 1.5216928124427795} 01/29/2022 06:23:35 - INFO - codeparrot_training - Step 13676: {'lr': 0.000430485351505287, 'samples': 2625984, 'steps': 13676, 'loss/train': 1.5687920451164246} 01/29/2022 06:23:39 - INFO - codeparrot_training - Step 13677: {'lr': 0.00043047402904470455, 'samples': 2626176, 'steps': 13677, 'loss/train': 1.495640605688095} 01/29/2022 06:23:44 - INFO - codeparrot_training - Step 13678: {'lr': 0.00043046270581102865, 'samples': 2626368, 'steps': 13678, 'loss/train': 1.7564873099327087} 01/29/2022 06:23:49 - INFO - codeparrot_training - Step 13679: {'lr': 0.00043045138180430783, 'samples': 2626560, 'steps': 13679, 'loss/train': 1.360159009695053} 01/29/2022 06:23:53 - INFO - codeparrot_training - Step 13680: {'lr': 0.00043044005702459054, 'samples': 2626752, 'steps': 13680, 'loss/train': 2.01127552986145} 01/29/2022 06:23:57 - INFO - codeparrot_training - Step 13681: {'lr': 0.0004304287314719254, 'samples': 2626944, 'steps': 13681, 'loss/train': 0.9712519347667694} 01/29/2022 06:24:01 - INFO - codeparrot_training - Step 13682: {'lr': 0.00043041740514636085, 'samples': 2627136, 'steps': 13682, 'loss/train': 0.665398433804512} 01/29/2022 06:24:07 - INFO - codeparrot_training - Step 13683: {'lr': 0.0004304060780479454, 'samples': 2627328, 'steps': 13683, 'loss/train': 2.5037536025047302} 01/29/2022 06:24:12 - INFO - codeparrot_training - Step 13684: {'lr': 0.0004303947501767276, 'samples': 2627520, 'steps': 13684, 'loss/train': 2.1740834712982178} 01/29/2022 06:24:16 - INFO - codeparrot_training - Step 13685: {'lr': 0.0004303834215327561, 'samples': 2627712, 'steps': 13685, 'loss/train': 1.4932619333267212} 01/29/2022 06:24:20 - INFO - codeparrot_training - Step 13686: {'lr': 0.00043037209211607913, 'samples': 2627904, 'steps': 13686, 'loss/train': 1.0393441021442413} 01/29/2022 06:24:24 - INFO - codeparrot_training - Step 13687: {'lr': 0.00043036076192674546, 'samples': 2628096, 'steps': 13687, 'loss/train': 1.8189391493797302} 01/29/2022 06:24:30 - INFO - codeparrot_training - Step 13688: {'lr': 0.00043034943096480357, 'samples': 2628288, 'steps': 13688, 'loss/train': 1.8728045225143433} 01/29/2022 06:24:34 - INFO - codeparrot_training - Step 13689: {'lr': 0.000430338099230302, 'samples': 2628480, 'steps': 13689, 'loss/train': 1.9349627494812012} 01/29/2022 06:24:38 - INFO - codeparrot_training - Step 13690: {'lr': 0.00043032676672328916, 'samples': 2628672, 'steps': 13690, 'loss/train': 1.8920105695724487} 01/29/2022 06:24:42 - INFO - codeparrot_training - Step 13691: {'lr': 0.00043031543344381384, 'samples': 2628864, 'steps': 13691, 'loss/train': 2.0133050680160522} 01/29/2022 06:24:48 - INFO - codeparrot_training - Step 13692: {'lr': 0.0004303040993919244, 'samples': 2629056, 'steps': 13692, 'loss/train': 1.6315466165542603} 01/29/2022 06:24:52 - INFO - codeparrot_training - Step 13693: {'lr': 0.00043029276456766946, 'samples': 2629248, 'steps': 13693, 'loss/train': 1.3789785504341125} 01/29/2022 06:24:56 - INFO - codeparrot_training - Step 13694: {'lr': 0.00043028142897109754, 'samples': 2629440, 'steps': 13694, 'loss/train': 1.747191846370697} 01/29/2022 06:25:00 - INFO - codeparrot_training - Step 13695: {'lr': 0.0004302700926022573, 'samples': 2629632, 'steps': 13695, 'loss/train': 2.7561038732528687} 01/29/2022 06:25:05 - INFO - codeparrot_training - Step 13696: {'lr': 0.0004302587554611972, 'samples': 2629824, 'steps': 13696, 'loss/train': 1.540584683418274} 01/29/2022 06:25:10 - INFO - codeparrot_training - Step 13697: {'lr': 0.0004302474175479658, 'samples': 2630016, 'steps': 13697, 'loss/train': 1.8361271023750305} 01/29/2022 06:25:14 - INFO - codeparrot_training - Step 13698: {'lr': 0.0004302360788626117, 'samples': 2630208, 'steps': 13698, 'loss/train': 1.3616795539855957} 01/29/2022 06:25:18 - INFO - codeparrot_training - Step 13699: {'lr': 0.00043022473940518345, 'samples': 2630400, 'steps': 13699, 'loss/train': 1.4462610483169556} 01/29/2022 06:25:23 - INFO - codeparrot_training - Step 13700: {'lr': 0.0004302133991757297, 'samples': 2630592, 'steps': 13700, 'loss/train': 1.4248495995998383} 01/29/2022 06:25:27 - INFO - codeparrot_training - Step 13701: {'lr': 0.00043020205817429895, 'samples': 2630784, 'steps': 13701, 'loss/train': 1.4512164294719696} 01/29/2022 06:25:33 - INFO - codeparrot_training - Step 13702: {'lr': 0.0004301907164009398, 'samples': 2630976, 'steps': 13702, 'loss/train': 2.306535243988037} 01/29/2022 06:25:37 - INFO - codeparrot_training - Step 13703: {'lr': 0.00043017937385570083, 'samples': 2631168, 'steps': 13703, 'loss/train': 0.8838938176631927} 01/29/2022 06:25:41 - INFO - codeparrot_training - Step 13704: {'lr': 0.00043016803053863063, 'samples': 2631360, 'steps': 13704, 'loss/train': 2.006199896335602} 01/29/2022 06:25:46 - INFO - codeparrot_training - Step 13705: {'lr': 0.00043015668644977783, 'samples': 2631552, 'steps': 13705, 'loss/train': 1.4981686770915985} 01/29/2022 06:25:50 - INFO - codeparrot_training - Step 13706: {'lr': 0.000430145341589191, 'samples': 2631744, 'steps': 13706, 'loss/train': 1.9198195338249207} 01/29/2022 06:25:55 - INFO - codeparrot_training - Step 13707: {'lr': 0.0004301339959569187, 'samples': 2631936, 'steps': 13707, 'loss/train': 1.8990431427955627} 01/29/2022 06:25:59 - INFO - codeparrot_training - Step 13708: {'lr': 0.00043012264955300954, 'samples': 2632128, 'steps': 13708, 'loss/train': 2.842590808868408} 01/29/2022 06:26:04 - INFO - codeparrot_training - Step 13709: {'lr': 0.0004301113023775122, 'samples': 2632320, 'steps': 13709, 'loss/train': 1.2966397404670715} 01/29/2022 06:26:08 - INFO - codeparrot_training - Step 13710: {'lr': 0.00043009995443047517, 'samples': 2632512, 'steps': 13710, 'loss/train': 1.6479061245918274} 01/29/2022 06:26:12 - INFO - codeparrot_training - Step 13711: {'lr': 0.0004300886057119472, 'samples': 2632704, 'steps': 13711, 'loss/train': 1.1428575217723846} 01/29/2022 06:26:17 - INFO - codeparrot_training - Step 13712: {'lr': 0.00043007725622197675, 'samples': 2632896, 'steps': 13712, 'loss/train': 1.7657434344291687} 01/29/2022 06:26:21 - INFO - codeparrot_training - Step 13713: {'lr': 0.00043006590596061256, 'samples': 2633088, 'steps': 13713, 'loss/train': 1.3301533162593842} 01/29/2022 06:26:26 - INFO - codeparrot_training - Step 13714: {'lr': 0.0004300545549279032, 'samples': 2633280, 'steps': 13714, 'loss/train': 1.6671225428581238} 01/29/2022 06:26:30 - INFO - codeparrot_training - Step 13715: {'lr': 0.0004300432031238973, 'samples': 2633472, 'steps': 13715, 'loss/train': 1.6573466062545776} 01/29/2022 06:26:34 - INFO - codeparrot_training - Step 13716: {'lr': 0.00043003185054864344, 'samples': 2633664, 'steps': 13716, 'loss/train': 1.6309835314750671} 01/29/2022 06:26:40 - INFO - codeparrot_training - Step 13717: {'lr': 0.0004300204972021903, 'samples': 2633856, 'steps': 13717, 'loss/train': 2.7892057299613953} 01/29/2022 06:26:44 - INFO - codeparrot_training - Step 13718: {'lr': 0.00043000914308458663, 'samples': 2634048, 'steps': 13718, 'loss/train': 1.7785897850990295} 01/29/2022 06:26:49 - INFO - codeparrot_training - Step 13719: {'lr': 0.0004299977881958808, 'samples': 2634240, 'steps': 13719, 'loss/train': 0.7684366106987} 01/29/2022 06:26:53 - INFO - codeparrot_training - Step 13720: {'lr': 0.0004299864325361217, 'samples': 2634432, 'steps': 13720, 'loss/train': 1.7737703919410706} 01/29/2022 06:26:57 - INFO - codeparrot_training - Step 13721: {'lr': 0.00042997507610535783, 'samples': 2634624, 'steps': 13721, 'loss/train': 2.4103403091430664} 01/29/2022 06:27:03 - INFO - codeparrot_training - Step 13722: {'lr': 0.00042996371890363796, 'samples': 2634816, 'steps': 13722, 'loss/train': 1.9592067003250122} 01/29/2022 06:27:07 - INFO - codeparrot_training - Step 13723: {'lr': 0.00042995236093101055, 'samples': 2635008, 'steps': 13723, 'loss/train': 1.030463844537735} 01/29/2022 06:27:11 - INFO - codeparrot_training - Step 13724: {'lr': 0.0004299410021875244, 'samples': 2635200, 'steps': 13724, 'loss/train': 2.1454601883888245} 01/29/2022 06:27:15 - INFO - codeparrot_training - Step 13725: {'lr': 0.00042992964267322823, 'samples': 2635392, 'steps': 13725, 'loss/train': 1.406396359205246} 01/29/2022 06:27:20 - INFO - codeparrot_training - Step 13726: {'lr': 0.00042991828238817046, 'samples': 2635584, 'steps': 13726, 'loss/train': 2.199042499065399} 01/29/2022 06:27:26 - INFO - codeparrot_training - Step 13727: {'lr': 0.0004299069213324, 'samples': 2635776, 'steps': 13727, 'loss/train': 0.9230546057224274} 01/29/2022 06:27:30 - INFO - codeparrot_training - Step 13728: {'lr': 0.0004298955595059654, 'samples': 2635968, 'steps': 13728, 'loss/train': 1.9899192452430725} 01/29/2022 06:27:34 - INFO - codeparrot_training - Step 13729: {'lr': 0.00042988419690891534, 'samples': 2636160, 'steps': 13729, 'loss/train': 2.010991930961609} 01/29/2022 06:27:38 - INFO - codeparrot_training - Step 13730: {'lr': 0.00042987283354129846, 'samples': 2636352, 'steps': 13730, 'loss/train': 0.9695610702037811} 01/29/2022 06:27:43 - INFO - codeparrot_training - Step 13731: {'lr': 0.0004298614694031635, 'samples': 2636544, 'steps': 13731, 'loss/train': 2.06003201007843} 01/29/2022 06:27:48 - INFO - codeparrot_training - Step 13732: {'lr': 0.0004298501044945591, 'samples': 2636736, 'steps': 13732, 'loss/train': 1.0946700274944305} 01/29/2022 06:27:52 - INFO - codeparrot_training - Step 13733: {'lr': 0.000429838738815534, 'samples': 2636928, 'steps': 13733, 'loss/train': 1.5376644730567932} 01/29/2022 06:27:56 - INFO - codeparrot_training - Step 13734: {'lr': 0.00042982737236613687, 'samples': 2637120, 'steps': 13734, 'loss/train': 1.450358659029007} 01/29/2022 06:28:01 - INFO - codeparrot_training - Step 13735: {'lr': 0.00042981600514641635, 'samples': 2637312, 'steps': 13735, 'loss/train': 1.943388819694519} 01/29/2022 06:28:05 - INFO - codeparrot_training - Step 13736: {'lr': 0.00042980463715642115, 'samples': 2637504, 'steps': 13736, 'loss/train': 1.933775782585144} 01/29/2022 06:28:10 - INFO - codeparrot_training - Step 13737: {'lr': 0.0004297932683962, 'samples': 2637696, 'steps': 13737, 'loss/train': 1.6405588388442993} 01/29/2022 06:28:14 - INFO - codeparrot_training - Step 13738: {'lr': 0.00042978189886580157, 'samples': 2637888, 'steps': 13738, 'loss/train': 1.3971214592456818} 01/29/2022 06:28:19 - INFO - codeparrot_training - Step 13739: {'lr': 0.00042977052856527456, 'samples': 2638080, 'steps': 13739, 'loss/train': 1.9602203965187073} 01/29/2022 06:28:23 - INFO - codeparrot_training - Step 13740: {'lr': 0.00042975915749466763, 'samples': 2638272, 'steps': 13740, 'loss/train': 2.193275213241577} 01/29/2022 06:28:27 - INFO - codeparrot_training - Step 13741: {'lr': 0.0004297477856540296, 'samples': 2638464, 'steps': 13741, 'loss/train': 1.261234313249588} 01/29/2022 06:28:33 - INFO - codeparrot_training - Step 13742: {'lr': 0.00042973641304340916, 'samples': 2638656, 'steps': 13742, 'loss/train': 1.7109671831130981} 01/29/2022 06:28:38 - INFO - codeparrot_training - Step 13743: {'lr': 0.00042972503966285503, 'samples': 2638848, 'steps': 13743, 'loss/train': 1.113318532705307} 01/29/2022 06:28:42 - INFO - codeparrot_training - Step 13744: {'lr': 0.00042971366551241587, 'samples': 2639040, 'steps': 13744, 'loss/train': 0.8633158206939697} 01/29/2022 06:28:46 - INFO - codeparrot_training - Step 13745: {'lr': 0.00042970229059214037, 'samples': 2639232, 'steps': 13745, 'loss/train': 2.1952820420265198} 01/29/2022 06:28:50 - INFO - codeparrot_training - Step 13746: {'lr': 0.0004296909149020774, 'samples': 2639424, 'steps': 13746, 'loss/train': 2.4027925729751587} 01/29/2022 06:28:56 - INFO - codeparrot_training - Step 13747: {'lr': 0.0004296795384422756, 'samples': 2639616, 'steps': 13747, 'loss/train': 1.7723209261894226} 01/29/2022 06:29:00 - INFO - codeparrot_training - Step 13748: {'lr': 0.00042966816121278365, 'samples': 2639808, 'steps': 13748, 'loss/train': 1.5737625360488892} 01/29/2022 06:29:04 - INFO - codeparrot_training - Step 13749: {'lr': 0.00042965678321365045, 'samples': 2640000, 'steps': 13749, 'loss/train': 0.8763398230075836} 01/29/2022 06:29:08 - INFO - codeparrot_training - Step 13750: {'lr': 0.00042964540444492453, 'samples': 2640192, 'steps': 13750, 'loss/train': 1.399617075920105} 01/29/2022 06:29:12 - INFO - codeparrot_training - Step 13751: {'lr': 0.00042963402490665484, 'samples': 2640384, 'steps': 13751, 'loss/train': 1.706677258014679} 01/29/2022 06:29:18 - INFO - codeparrot_training - Step 13752: {'lr': 0.0004296226445988899, 'samples': 2640576, 'steps': 13752, 'loss/train': 2.115329146385193} 01/29/2022 06:29:22 - INFO - codeparrot_training - Step 13753: {'lr': 0.0004296112635216787, 'samples': 2640768, 'steps': 13753, 'loss/train': 0.7609190046787262} 01/29/2022 06:29:26 - INFO - codeparrot_training - Step 13754: {'lr': 0.00042959988167506983, 'samples': 2640960, 'steps': 13754, 'loss/train': 1.606011986732483} 01/29/2022 06:29:30 - INFO - codeparrot_training - Step 13755: {'lr': 0.00042958849905911213, 'samples': 2641152, 'steps': 13755, 'loss/train': 1.5673359632492065} 01/29/2022 06:29:36 - INFO - codeparrot_training - Step 13756: {'lr': 0.0004295771156738543, 'samples': 2641344, 'steps': 13756, 'loss/train': 1.7026222944259644} 01/29/2022 06:29:40 - INFO - codeparrot_training - Step 13757: {'lr': 0.00042956573151934507, 'samples': 2641536, 'steps': 13757, 'loss/train': 1.3401290774345398} 01/29/2022 06:29:44 - INFO - codeparrot_training - Step 13758: {'lr': 0.00042955434659563334, 'samples': 2641728, 'steps': 13758, 'loss/train': 2.0404345393180847} 01/29/2022 06:29:48 - INFO - codeparrot_training - Step 13759: {'lr': 0.00042954296090276777, 'samples': 2641920, 'steps': 13759, 'loss/train': 1.431463211774826} 01/29/2022 06:29:53 - INFO - codeparrot_training - Step 13760: {'lr': 0.0004295315744407972, 'samples': 2642112, 'steps': 13760, 'loss/train': 1.7007343769073486} 01/29/2022 06:29:59 - INFO - codeparrot_training - Step 13761: {'lr': 0.0004295201872097704, 'samples': 2642304, 'steps': 13761, 'loss/train': 2.0008013248443604} 01/29/2022 06:30:03 - INFO - codeparrot_training - Step 13762: {'lr': 0.0004295087992097361, 'samples': 2642496, 'steps': 13762, 'loss/train': 1.9778430461883545} 01/29/2022 06:30:08 - INFO - codeparrot_training - Step 13763: {'lr': 0.00042949741044074306, 'samples': 2642688, 'steps': 13763, 'loss/train': 1.4320195019245148} 01/29/2022 06:30:12 - INFO - codeparrot_training - Step 13764: {'lr': 0.00042948602090284014, 'samples': 2642880, 'steps': 13764, 'loss/train': 0.7547801434993744} 01/29/2022 06:30:16 - INFO - codeparrot_training - Step 13765: {'lr': 0.00042947463059607606, 'samples': 2643072, 'steps': 13765, 'loss/train': 1.5034700632095337} 01/29/2022 06:30:20 - INFO - codeparrot_training - Step 13766: {'lr': 0.0004294632395204997, 'samples': 2643264, 'steps': 13766, 'loss/train': 0.10973840206861496} 01/29/2022 06:30:26 - INFO - codeparrot_training - Step 13767: {'lr': 0.0004294518476761598, 'samples': 2643456, 'steps': 13767, 'loss/train': 1.935423195362091} 01/29/2022 06:30:30 - INFO - codeparrot_training - Step 13768: {'lr': 0.00042944045506310515, 'samples': 2643648, 'steps': 13768, 'loss/train': 1.8659580945968628} 01/29/2022 06:30:34 - INFO - codeparrot_training - Step 13769: {'lr': 0.0004294290616813846, 'samples': 2643840, 'steps': 13769, 'loss/train': 1.8304633498191833} 01/29/2022 06:30:38 - INFO - codeparrot_training - Step 13770: {'lr': 0.00042941766753104696, 'samples': 2644032, 'steps': 13770, 'loss/train': 1.4904951453208923} 01/29/2022 06:30:43 - INFO - codeparrot_training - Step 13771: {'lr': 0.00042940627261214094, 'samples': 2644224, 'steps': 13771, 'loss/train': 1.4741347432136536} 01/29/2022 06:30:49 - INFO - codeparrot_training - Step 13772: {'lr': 0.00042939487692471534, 'samples': 2644416, 'steps': 13772, 'loss/train': 0.5678623616695404} 01/29/2022 06:30:53 - INFO - codeparrot_training - Step 13773: {'lr': 0.0004293834804688192, 'samples': 2644608, 'steps': 13773, 'loss/train': 1.8308270573616028} 01/29/2022 06:30:57 - INFO - codeparrot_training - Step 13774: {'lr': 0.00042937208324450116, 'samples': 2644800, 'steps': 13774, 'loss/train': 1.482191652059555} 01/29/2022 06:31:01 - INFO - codeparrot_training - Step 13775: {'lr': 0.00042936068525181004, 'samples': 2644992, 'steps': 13775, 'loss/train': 0.054904064163565636} 01/29/2022 06:31:06 - INFO - codeparrot_training - Step 13776: {'lr': 0.00042934928649079467, 'samples': 2645184, 'steps': 13776, 'loss/train': 7.399700403213501} 01/29/2022 06:31:11 - INFO - codeparrot_training - Step 13777: {'lr': 0.0004293378869615039, 'samples': 2645376, 'steps': 13777, 'loss/train': 0.9565658569335938} 01/29/2022 06:31:15 - INFO - codeparrot_training - Step 13778: {'lr': 0.00042932648666398667, 'samples': 2645568, 'steps': 13778, 'loss/train': 2.0283243656158447} 01/29/2022 06:31:20 - INFO - codeparrot_training - Step 13779: {'lr': 0.0004293150855982916, 'samples': 2645760, 'steps': 13779, 'loss/train': 1.8588488101959229} 01/29/2022 06:31:24 - INFO - codeparrot_training - Step 13780: {'lr': 0.0004293036837644677, 'samples': 2645952, 'steps': 13780, 'loss/train': 2.0661814212799072} 01/29/2022 06:31:28 - INFO - codeparrot_training - Step 13781: {'lr': 0.0004292922811625637, 'samples': 2646144, 'steps': 13781, 'loss/train': 1.6030787229537964} 01/29/2022 06:31:34 - INFO - codeparrot_training - Step 13782: {'lr': 0.0004292808777926286, 'samples': 2646336, 'steps': 13782, 'loss/train': 1.2596519887447357} 01/29/2022 06:31:39 - INFO - codeparrot_training - Step 13783: {'lr': 0.0004292694736547111, 'samples': 2646528, 'steps': 13783, 'loss/train': 1.7288913130760193} 01/29/2022 06:31:43 - INFO - codeparrot_training - Step 13784: {'lr': 0.0004292580687488601, 'samples': 2646720, 'steps': 13784, 'loss/train': 1.3235026895999908} 01/29/2022 06:31:47 - INFO - codeparrot_training - Step 13785: {'lr': 0.00042924666307512437, 'samples': 2646912, 'steps': 13785, 'loss/train': 2.0386651754379272} 01/29/2022 06:31:51 - INFO - codeparrot_training - Step 13786: {'lr': 0.000429235256633553, 'samples': 2647104, 'steps': 13786, 'loss/train': 1.041328489780426} 01/29/2022 06:31:55 - INFO - codeparrot_training - Step 13787: {'lr': 0.0004292238494241946, 'samples': 2647296, 'steps': 13787, 'loss/train': 1.6812281012535095} 01/29/2022 06:32:01 - INFO - codeparrot_training - Step 13788: {'lr': 0.00042921244144709817, 'samples': 2647488, 'steps': 13788, 'loss/train': 1.8244736194610596} 01/29/2022 06:32:06 - INFO - codeparrot_training - Step 13789: {'lr': 0.0004292010327023125, 'samples': 2647680, 'steps': 13789, 'loss/train': 8.053145170211792} 01/29/2022 06:32:10 - INFO - codeparrot_training - Step 13790: {'lr': 0.00042918962318988664, 'samples': 2647872, 'steps': 13790, 'loss/train': 1.930548369884491} 01/29/2022 06:32:14 - INFO - codeparrot_training - Step 13791: {'lr': 0.00042917821290986926, 'samples': 2648064, 'steps': 13791, 'loss/train': 1.6707619428634644} 01/29/2022 06:32:18 - INFO - codeparrot_training - Step 13792: {'lr': 0.0004291668018623093, 'samples': 2648256, 'steps': 13792, 'loss/train': 2.0988117456436157} 01/29/2022 06:32:24 - INFO - codeparrot_training - Step 13793: {'lr': 0.00042915539004725564, 'samples': 2648448, 'steps': 13793, 'loss/train': 0.43657006323337555} 01/29/2022 06:32:28 - INFO - codeparrot_training - Step 13794: {'lr': 0.0004291439774647572, 'samples': 2648640, 'steps': 13794, 'loss/train': 0.176686093211174} 01/29/2022 06:32:32 - INFO - codeparrot_training - Step 13795: {'lr': 0.00042913256411486277, 'samples': 2648832, 'steps': 13795, 'loss/train': 1.4630799293518066} 01/29/2022 06:32:37 - INFO - codeparrot_training - Step 13796: {'lr': 0.0004291211499976214, 'samples': 2649024, 'steps': 13796, 'loss/train': 1.5902429223060608} 01/29/2022 06:32:41 - INFO - codeparrot_training - Step 13797: {'lr': 0.00042910973511308195, 'samples': 2649216, 'steps': 13797, 'loss/train': 1.5976447463035583} 01/29/2022 06:32:47 - INFO - codeparrot_training - Step 13798: {'lr': 0.0004290983194612932, 'samples': 2649408, 'steps': 13798, 'loss/train': 1.5548474192619324} 01/29/2022 06:32:51 - INFO - codeparrot_training - Step 13799: {'lr': 0.00042908690304230415, 'samples': 2649600, 'steps': 13799, 'loss/train': 2.422869086265564} 01/29/2022 06:32:55 - INFO - codeparrot_training - Step 13800: {'lr': 0.00042907548585616363, 'samples': 2649792, 'steps': 13800, 'loss/train': 1.907843291759491} 01/29/2022 06:33:00 - INFO - codeparrot_training - Step 13801: {'lr': 0.00042906406790292053, 'samples': 2649984, 'steps': 13801, 'loss/train': 1.5765491724014282} 01/29/2022 06:33:04 - INFO - codeparrot_training - Step 13802: {'lr': 0.00042905264918262386, 'samples': 2650176, 'steps': 13802, 'loss/train': 2.4178066849708557} 01/29/2022 06:33:09 - INFO - codeparrot_training - Step 13803: {'lr': 0.00042904122969532256, 'samples': 2650368, 'steps': 13803, 'loss/train': 2.2238963842391968} 01/29/2022 06:33:13 - INFO - codeparrot_training - Step 13804: {'lr': 0.0004290298094410655, 'samples': 2650560, 'steps': 13804, 'loss/train': 1.7713614106178284} 01/29/2022 06:33:18 - INFO - codeparrot_training - Step 13805: {'lr': 0.0004290183884199015, 'samples': 2650752, 'steps': 13805, 'loss/train': 1.113345891237259} 01/29/2022 06:33:22 - INFO - codeparrot_training - Step 13806: {'lr': 0.00042900696663187963, 'samples': 2650944, 'steps': 13806, 'loss/train': 1.7025707960128784} 01/29/2022 06:33:26 - INFO - codeparrot_training - Step 13807: {'lr': 0.00042899554407704876, 'samples': 2651136, 'steps': 13807, 'loss/train': 1.8094473481178284} 01/29/2022 06:33:32 - INFO - codeparrot_training - Step 13808: {'lr': 0.0004289841207554578, 'samples': 2651328, 'steps': 13808, 'loss/train': 1.8368107080459595} 01/29/2022 06:33:36 - INFO - codeparrot_training - Step 13809: {'lr': 0.0004289726966671557, 'samples': 2651520, 'steps': 13809, 'loss/train': 0.3419109284877777} 01/29/2022 06:33:41 - INFO - codeparrot_training - Step 13810: {'lr': 0.00042896127181219135, 'samples': 2651712, 'steps': 13810, 'loss/train': 0.995702862739563} 01/29/2022 06:33:45 - INFO - codeparrot_training - Step 13811: {'lr': 0.0004289498461906138, 'samples': 2651904, 'steps': 13811, 'loss/train': 1.1781824827194214} 01/29/2022 06:33:49 - INFO - codeparrot_training - Step 13812: {'lr': 0.00042893841980247194, 'samples': 2652096, 'steps': 13812, 'loss/train': 1.6449739336967468} 01/29/2022 06:33:54 - INFO - codeparrot_training - Step 13813: {'lr': 0.00042892699264781463, 'samples': 2652288, 'steps': 13813, 'loss/train': 1.9275296330451965} 01/29/2022 06:33:59 - INFO - codeparrot_training - Step 13814: {'lr': 0.000428915564726691, 'samples': 2652480, 'steps': 13814, 'loss/train': 1.2843952775001526} 01/29/2022 06:34:03 - INFO - codeparrot_training - Step 13815: {'lr': 0.0004289041360391499, 'samples': 2652672, 'steps': 13815, 'loss/train': 1.7705857157707214} 01/29/2022 06:34:07 - INFO - codeparrot_training - Step 13816: {'lr': 0.0004288927065852402, 'samples': 2652864, 'steps': 13816, 'loss/train': 2.1629664301872253} 01/29/2022 06:34:11 - INFO - codeparrot_training - Step 13817: {'lr': 0.000428881276365011, 'samples': 2653056, 'steps': 13817, 'loss/train': 1.659337341785431} 01/29/2022 06:34:17 - INFO - codeparrot_training - Step 13818: {'lr': 0.00042886984537851124, 'samples': 2653248, 'steps': 13818, 'loss/train': 1.990621268749237} 01/29/2022 06:34:22 - INFO - codeparrot_training - Step 13819: {'lr': 0.0004288584136257898, 'samples': 2653440, 'steps': 13819, 'loss/train': 1.7552799582481384} 01/29/2022 06:34:26 - INFO - codeparrot_training - Step 13820: {'lr': 0.00042884698110689574, 'samples': 2653632, 'steps': 13820, 'loss/train': 1.0054684281349182} 01/29/2022 06:34:30 - INFO - codeparrot_training - Step 13821: {'lr': 0.000428835547821878, 'samples': 2653824, 'steps': 13821, 'loss/train': 1.5776740908622742} 01/29/2022 06:34:34 - INFO - codeparrot_training - Step 13822: {'lr': 0.00042882411377078556, 'samples': 2654016, 'steps': 13822, 'loss/train': 1.8298451900482178} 01/29/2022 06:34:40 - INFO - codeparrot_training - Step 13823: {'lr': 0.00042881267895366736, 'samples': 2654208, 'steps': 13823, 'loss/train': 1.8889244198799133} 01/29/2022 06:34:44 - INFO - codeparrot_training - Step 13824: {'lr': 0.00042880124337057253, 'samples': 2654400, 'steps': 13824, 'loss/train': 1.6768004894256592} 01/29/2022 06:34:48 - INFO - codeparrot_training - Step 13825: {'lr': 0.00042878980702154985, 'samples': 2654592, 'steps': 13825, 'loss/train': 1.2896912097930908} 01/29/2022 06:34:52 - INFO - codeparrot_training - Step 13826: {'lr': 0.00042877836990664844, 'samples': 2654784, 'steps': 13826, 'loss/train': 1.5767971873283386} 01/29/2022 06:34:57 - INFO - codeparrot_training - Step 13827: {'lr': 0.00042876693202591724, 'samples': 2654976, 'steps': 13827, 'loss/train': 2.0463595390319824} 01/29/2022 06:35:02 - INFO - codeparrot_training - Step 13828: {'lr': 0.0004287554933794053, 'samples': 2655168, 'steps': 13828, 'loss/train': 1.6585137248039246} 01/29/2022 06:35:06 - INFO - codeparrot_training - Step 13829: {'lr': 0.0004287440539671616, 'samples': 2655360, 'steps': 13829, 'loss/train': 2.393657147884369} 01/29/2022 06:35:11 - INFO - codeparrot_training - Step 13830: {'lr': 0.0004287326137892351, 'samples': 2655552, 'steps': 13830, 'loss/train': 0.6920550763607025} 01/29/2022 06:35:15 - INFO - codeparrot_training - Step 13831: {'lr': 0.00042872117284567486, 'samples': 2655744, 'steps': 13831, 'loss/train': 2.3531221747398376} 01/29/2022 06:35:19 - INFO - codeparrot_training - Step 13832: {'lr': 0.0004287097311365299, 'samples': 2655936, 'steps': 13832, 'loss/train': 0.7249685525894165} 01/29/2022 06:35:25 - INFO - codeparrot_training - Step 13833: {'lr': 0.0004286982886618491, 'samples': 2656128, 'steps': 13833, 'loss/train': 1.7302080988883972} 01/29/2022 06:35:30 - INFO - codeparrot_training - Step 13834: {'lr': 0.0004286868454216816, 'samples': 2656320, 'steps': 13834, 'loss/train': 1.27950918674469} 01/29/2022 06:35:34 - INFO - codeparrot_training - Step 13835: {'lr': 0.00042867540141607643, 'samples': 2656512, 'steps': 13835, 'loss/train': 2.285217583179474} 01/29/2022 06:35:38 - INFO - codeparrot_training - Step 13836: {'lr': 0.0004286639566450826, 'samples': 2656704, 'steps': 13836, 'loss/train': 1.5234886407852173} 01/29/2022 06:35:43 - INFO - codeparrot_training - Step 13837: {'lr': 0.00042865251110874903, 'samples': 2656896, 'steps': 13837, 'loss/train': 1.558485746383667} 01/29/2022 06:35:48 - INFO - codeparrot_training - Step 13838: {'lr': 0.00042864106480712495, 'samples': 2657088, 'steps': 13838, 'loss/train': 1.500166118144989} 01/29/2022 06:35:52 - INFO - codeparrot_training - Step 13839: {'lr': 0.00042862961774025915, 'samples': 2657280, 'steps': 13839, 'loss/train': 1.970747172832489} 01/29/2022 06:35:56 - INFO - codeparrot_training - Step 13840: {'lr': 0.00042861816990820087, 'samples': 2657472, 'steps': 13840, 'loss/train': 2.160472333431244} 01/29/2022 06:36:00 - INFO - codeparrot_training - Step 13841: {'lr': 0.00042860672131099904, 'samples': 2657664, 'steps': 13841, 'loss/train': 2.7138365507125854} 01/29/2022 06:36:06 - INFO - codeparrot_training - Step 13842: {'lr': 0.00042859527194870275, 'samples': 2657856, 'steps': 13842, 'loss/train': 1.4932209849357605} 01/29/2022 06:36:11 - INFO - codeparrot_training - Step 13843: {'lr': 0.000428583821821361, 'samples': 2658048, 'steps': 13843, 'loss/train': 2.0975117683410645} 01/29/2022 06:36:15 - INFO - codeparrot_training - Step 13844: {'lr': 0.00042857237092902285, 'samples': 2658240, 'steps': 13844, 'loss/train': 2.0918326377868652} 01/29/2022 06:36:19 - INFO - codeparrot_training - Step 13845: {'lr': 0.0004285609192717374, 'samples': 2658432, 'steps': 13845, 'loss/train': 2.123161196708679} 01/29/2022 06:36:24 - INFO - codeparrot_training - Step 13846: {'lr': 0.00042854946684955366, 'samples': 2658624, 'steps': 13846, 'loss/train': 1.8253929018974304} 01/29/2022 06:36:29 - INFO - codeparrot_training - Step 13847: {'lr': 0.00042853801366252067, 'samples': 2658816, 'steps': 13847, 'loss/train': 1.7710676193237305} 01/29/2022 06:36:33 - INFO - codeparrot_training - Step 13848: {'lr': 0.00042852655971068756, 'samples': 2659008, 'steps': 13848, 'loss/train': 1.2080220580101013} 01/29/2022 06:36:38 - INFO - codeparrot_training - Step 13849: {'lr': 0.0004285151049941033, 'samples': 2659200, 'steps': 13849, 'loss/train': 1.6558865904808044} 01/29/2022 06:36:42 - INFO - codeparrot_training - Step 13850: {'lr': 0.00042850364951281707, 'samples': 2659392, 'steps': 13850, 'loss/train': 0.7705931067466736} 01/29/2022 06:36:46 - INFO - codeparrot_training - Step 13851: {'lr': 0.00042849219326687786, 'samples': 2659584, 'steps': 13851, 'loss/train': 2.5486063957214355} 01/29/2022 06:36:51 - INFO - codeparrot_training - Step 13852: {'lr': 0.0004284807362563348, 'samples': 2659776, 'steps': 13852, 'loss/train': 0.7199935466051102} 01/29/2022 06:36:56 - INFO - codeparrot_training - Step 13853: {'lr': 0.00042846927848123694, 'samples': 2659968, 'steps': 13853, 'loss/train': 0.48522788286209106} 01/29/2022 06:37:00 - INFO - codeparrot_training - Step 13854: {'lr': 0.00042845781994163334, 'samples': 2660160, 'steps': 13854, 'loss/train': 1.3741343021392822} 01/29/2022 06:37:04 - INFO - codeparrot_training - Step 13855: {'lr': 0.00042844636063757316, 'samples': 2660352, 'steps': 13855, 'loss/train': 2.4469608664512634} 01/29/2022 06:37:08 - INFO - codeparrot_training - Step 13856: {'lr': 0.00042843490056910534, 'samples': 2660544, 'steps': 13856, 'loss/train': 2.5675129294395447} 01/29/2022 06:37:14 - INFO - codeparrot_training - Step 13857: {'lr': 0.0004284234397362791, 'samples': 2660736, 'steps': 13857, 'loss/train': 2.1211429238319397} 01/29/2022 06:37:18 - INFO - codeparrot_training - Step 13858: {'lr': 0.0004284119781391436, 'samples': 2660928, 'steps': 13858, 'loss/train': 1.5940502285957336} 01/29/2022 06:37:22 - INFO - codeparrot_training - Step 13859: {'lr': 0.00042840051577774766, 'samples': 2661120, 'steps': 13859, 'loss/train': 4.1393526792526245} 01/29/2022 06:37:27 - INFO - codeparrot_training - Step 13860: {'lr': 0.00042838905265214067, 'samples': 2661312, 'steps': 13860, 'loss/train': 1.127545326948166} 01/29/2022 06:37:31 - INFO - codeparrot_training - Step 13861: {'lr': 0.0004283775887623716, 'samples': 2661504, 'steps': 13861, 'loss/train': 1.0484192669391632} 01/29/2022 06:37:37 - INFO - codeparrot_training - Step 13862: {'lr': 0.0004283661241084896, 'samples': 2661696, 'steps': 13862, 'loss/train': 1.8948270678520203} 01/29/2022 06:37:41 - INFO - codeparrot_training - Step 13863: {'lr': 0.0004283546586905437, 'samples': 2661888, 'steps': 13863, 'loss/train': 1.922925889492035} 01/29/2022 06:37:45 - INFO - codeparrot_training - Step 13864: {'lr': 0.00042834319250858316, 'samples': 2662080, 'steps': 13864, 'loss/train': 2.367581605911255} 01/29/2022 06:37:50 - INFO - codeparrot_training - Step 13865: {'lr': 0.000428331725562657, 'samples': 2662272, 'steps': 13865, 'loss/train': 1.4985913932323456} 01/29/2022 06:37:54 - INFO - codeparrot_training - Step 13866: {'lr': 0.0004283202578528143, 'samples': 2662464, 'steps': 13866, 'loss/train': 1.2176638841629028} 01/29/2022 06:37:59 - INFO - codeparrot_training - Step 13867: {'lr': 0.00042830878937910426, 'samples': 2662656, 'steps': 13867, 'loss/train': 1.9522536993026733} 01/29/2022 06:38:03 - INFO - codeparrot_training - Step 13868: {'lr': 0.000428297320141576, 'samples': 2662848, 'steps': 13868, 'loss/train': 1.1397868394851685} 01/29/2022 06:38:08 - INFO - codeparrot_training - Step 13869: {'lr': 0.00042828585014027863, 'samples': 2663040, 'steps': 13869, 'loss/train': 1.4475368857383728} 01/29/2022 06:38:12 - INFO - codeparrot_training - Step 13870: {'lr': 0.0004282743793752613, 'samples': 2663232, 'steps': 13870, 'loss/train': 2.3095155358314514} 01/29/2022 06:38:16 - INFO - codeparrot_training - Step 13871: {'lr': 0.0004282629078465732, 'samples': 2663424, 'steps': 13871, 'loss/train': 1.425329178571701} 01/29/2022 06:38:21 - INFO - codeparrot_training - Step 13872: {'lr': 0.0004282514355542633, 'samples': 2663616, 'steps': 13872, 'loss/train': 1.4369869530200958} 01/29/2022 06:38:25 - INFO - codeparrot_training - Step 13873: {'lr': 0.0004282399624983808, 'samples': 2663808, 'steps': 13873, 'loss/train': 1.4175506830215454} 01/29/2022 06:38:30 - INFO - codeparrot_training - Step 13874: {'lr': 0.000428228488678975, 'samples': 2664000, 'steps': 13874, 'loss/train': 1.4045261442661285} 01/29/2022 06:38:34 - INFO - codeparrot_training - Step 13875: {'lr': 0.000428217014096095, 'samples': 2664192, 'steps': 13875, 'loss/train': 1.9449938535690308} 01/29/2022 06:38:38 - INFO - codeparrot_training - Step 13876: {'lr': 0.00042820553874978987, 'samples': 2664384, 'steps': 13876, 'loss/train': 1.5041726231575012} 01/29/2022 06:38:44 - INFO - codeparrot_training - Step 13877: {'lr': 0.0004281940626401087, 'samples': 2664576, 'steps': 13877, 'loss/train': 1.6722108721733093} 01/29/2022 06:38:48 - INFO - codeparrot_training - Step 13878: {'lr': 0.0004281825857671008, 'samples': 2664768, 'steps': 13878, 'loss/train': 0.8999199271202087} 01/29/2022 06:38:53 - INFO - codeparrot_training - Step 13879: {'lr': 0.00042817110813081526, 'samples': 2664960, 'steps': 13879, 'loss/train': 1.5491634607315063} 01/29/2022 06:38:57 - INFO - codeparrot_training - Step 13880: {'lr': 0.00042815962973130134, 'samples': 2665152, 'steps': 13880, 'loss/train': 1.9125938415527344} 01/29/2022 06:39:01 - INFO - codeparrot_training - Step 13881: {'lr': 0.00042814815056860814, 'samples': 2665344, 'steps': 13881, 'loss/train': 1.1536671817302704} 01/29/2022 06:39:07 - INFO - codeparrot_training - Step 13882: {'lr': 0.0004281366706427848, 'samples': 2665536, 'steps': 13882, 'loss/train': 1.8896058797836304} 01/29/2022 06:39:11 - INFO - codeparrot_training - Step 13883: {'lr': 0.0004281251899538805, 'samples': 2665728, 'steps': 13883, 'loss/train': 1.0235669016838074} 01/29/2022 06:39:15 - INFO - codeparrot_training - Step 13884: {'lr': 0.0004281137085019445, 'samples': 2665920, 'steps': 13884, 'loss/train': 1.4503541886806488} 01/29/2022 06:39:20 - INFO - codeparrot_training - Step 13885: {'lr': 0.0004281022262870259, 'samples': 2666112, 'steps': 13885, 'loss/train': 0.08343134075403214} 01/29/2022 06:39:26 - INFO - codeparrot_training - Step 13886: {'lr': 0.00042809074330917387, 'samples': 2666304, 'steps': 13886, 'loss/train': 0.6983247846364975} 01/29/2022 06:39:30 - INFO - codeparrot_training - Step 13887: {'lr': 0.00042807925956843775, 'samples': 2666496, 'steps': 13887, 'loss/train': 0.4446938931941986} 01/29/2022 06:39:34 - INFO - codeparrot_training - Step 13888: {'lr': 0.0004280677750648665, 'samples': 2666688, 'steps': 13888, 'loss/train': 1.0824604332447052} 01/29/2022 06:39:38 - INFO - codeparrot_training - Step 13889: {'lr': 0.0004280562897985095, 'samples': 2666880, 'steps': 13889, 'loss/train': 1.1111344993114471} 01/29/2022 06:39:43 - INFO - codeparrot_training - Step 13890: {'lr': 0.00042804480376941597, 'samples': 2667072, 'steps': 13890, 'loss/train': 4.034359574317932} 01/29/2022 06:39:48 - INFO - codeparrot_training - Step 13891: {'lr': 0.0004280333169776349, 'samples': 2667264, 'steps': 13891, 'loss/train': 1.2311806976795197} 01/29/2022 06:39:52 - INFO - codeparrot_training - Step 13892: {'lr': 0.00042802182942321576, 'samples': 2667456, 'steps': 13892, 'loss/train': 1.07962104678154} 01/29/2022 06:39:56 - INFO - codeparrot_training - Step 13893: {'lr': 0.00042801034110620756, 'samples': 2667648, 'steps': 13893, 'loss/train': 1.857422411441803} 01/29/2022 06:40:01 - INFO - codeparrot_training - Step 13894: {'lr': 0.00042799885202665964, 'samples': 2667840, 'steps': 13894, 'loss/train': 1.8777113556861877} 01/29/2022 06:40:05 - INFO - codeparrot_training - Step 13895: {'lr': 0.0004279873621846211, 'samples': 2668032, 'steps': 13895, 'loss/train': 1.9995521306991577} 01/29/2022 06:40:10 - INFO - codeparrot_training - Step 13896: {'lr': 0.0004279758715801412, 'samples': 2668224, 'steps': 13896, 'loss/train': 2.397905945777893} 01/29/2022 06:40:14 - INFO - codeparrot_training - Step 13897: {'lr': 0.0004279643802132692, 'samples': 2668416, 'steps': 13897, 'loss/train': 1.5147185325622559} 01/29/2022 06:40:19 - INFO - codeparrot_training - Step 13898: {'lr': 0.0004279528880840544, 'samples': 2668608, 'steps': 13898, 'loss/train': 1.6235116124153137} 01/29/2022 06:40:23 - INFO - codeparrot_training - Step 13899: {'lr': 0.00042794139519254583, 'samples': 2668800, 'steps': 13899, 'loss/train': 1.7083210945129395} 01/29/2022 06:40:27 - INFO - codeparrot_training - Step 13900: {'lr': 0.00042792990153879285, 'samples': 2668992, 'steps': 13900, 'loss/train': 2.1591022610664368} 01/29/2022 06:40:34 - INFO - codeparrot_training - Step 13901: {'lr': 0.00042791840712284466, 'samples': 2669184, 'steps': 13901, 'loss/train': 2.2778454422950745} 01/29/2022 06:40:38 - INFO - codeparrot_training - Step 13902: {'lr': 0.0004279069119447505, 'samples': 2669376, 'steps': 13902, 'loss/train': 0.7745148539543152} 01/29/2022 06:40:42 - INFO - codeparrot_training - Step 13903: {'lr': 0.0004278954160045597, 'samples': 2669568, 'steps': 13903, 'loss/train': 1.0945559442043304} 01/29/2022 06:40:46 - INFO - codeparrot_training - Step 13904: {'lr': 0.0004278839193023214, 'samples': 2669760, 'steps': 13904, 'loss/train': 2.005390763282776} 01/29/2022 06:40:51 - INFO - codeparrot_training - Step 13905: {'lr': 0.00042787242183808485, 'samples': 2669952, 'steps': 13905, 'loss/train': 0.3064424693584442} 01/29/2022 06:40:56 - INFO - codeparrot_training - Step 13906: {'lr': 0.00042786092361189927, 'samples': 2670144, 'steps': 13906, 'loss/train': 2.454463005065918} 01/29/2022 06:41:00 - INFO - codeparrot_training - Step 13907: {'lr': 0.00042784942462381403, 'samples': 2670336, 'steps': 13907, 'loss/train': 1.4241194128990173} 01/29/2022 06:41:04 - INFO - codeparrot_training - Step 13908: {'lr': 0.0004278379248738783, 'samples': 2670528, 'steps': 13908, 'loss/train': 0.9492053389549255} 01/29/2022 06:41:09 - INFO - codeparrot_training - Step 13909: {'lr': 0.00042782642436214137, 'samples': 2670720, 'steps': 13909, 'loss/train': 1.737643539905548} 01/29/2022 06:41:13 - INFO - codeparrot_training - Step 13910: {'lr': 0.00042781492308865255, 'samples': 2670912, 'steps': 13910, 'loss/train': 1.2320946156978607} 01/29/2022 06:41:18 - INFO - codeparrot_training - Step 13911: {'lr': 0.000427803421053461, 'samples': 2671104, 'steps': 13911, 'loss/train': 0.401675283908844} 01/29/2022 06:41:23 - INFO - codeparrot_training - Step 13912: {'lr': 0.0004277919182566161, 'samples': 2671296, 'steps': 13912, 'loss/train': 1.1430652141571045} 01/29/2022 06:41:27 - INFO - codeparrot_training - Step 13913: {'lr': 0.0004277804146981671, 'samples': 2671488, 'steps': 13913, 'loss/train': 1.354055106639862} 01/29/2022 06:41:31 - INFO - codeparrot_training - Step 13914: {'lr': 0.00042776891037816324, 'samples': 2671680, 'steps': 13914, 'loss/train': 1.9557958245277405} 01/29/2022 06:41:35 - INFO - codeparrot_training - Step 13915: {'lr': 0.00042775740529665373, 'samples': 2671872, 'steps': 13915, 'loss/train': 0.9709673523902893} 01/29/2022 06:41:41 - INFO - codeparrot_training - Step 13916: {'lr': 0.000427745899453688, 'samples': 2672064, 'steps': 13916, 'loss/train': 2.18353271484375} 01/29/2022 06:41:45 - INFO - codeparrot_training - Step 13917: {'lr': 0.0004277343928493153, 'samples': 2672256, 'steps': 13917, 'loss/train': 2.4946049451828003} 01/29/2022 06:41:49 - INFO - codeparrot_training - Step 13918: {'lr': 0.0004277228854835849, 'samples': 2672448, 'steps': 13918, 'loss/train': 1.3622478246688843} 01/29/2022 06:41:53 - INFO - codeparrot_training - Step 13919: {'lr': 0.0004277113773565461, 'samples': 2672640, 'steps': 13919, 'loss/train': 1.788859248161316} 01/29/2022 06:41:58 - INFO - codeparrot_training - Step 13920: {'lr': 0.00042769986846824813, 'samples': 2672832, 'steps': 13920, 'loss/train': 1.3343457877635956} 01/29/2022 06:42:04 - INFO - codeparrot_training - Step 13921: {'lr': 0.00042768835881874036, 'samples': 2673024, 'steps': 13921, 'loss/train': 2.0974475741386414} 01/29/2022 06:42:08 - INFO - codeparrot_training - Step 13922: {'lr': 0.00042767684840807214, 'samples': 2673216, 'steps': 13922, 'loss/train': 2.442525029182434} 01/29/2022 06:42:12 - INFO - codeparrot_training - Step 13923: {'lr': 0.00042766533723629264, 'samples': 2673408, 'steps': 13923, 'loss/train': 1.6808579564094543} 01/29/2022 06:42:16 - INFO - codeparrot_training - Step 13924: {'lr': 0.0004276538253034513, 'samples': 2673600, 'steps': 13924, 'loss/train': 0.3728507459163666} 01/29/2022 06:42:20 - INFO - codeparrot_training - Step 13925: {'lr': 0.0004276423126095974, 'samples': 2673792, 'steps': 13925, 'loss/train': 1.8197817206382751} 01/29/2022 06:42:26 - INFO - codeparrot_training - Step 13926: {'lr': 0.0004276307991547802, 'samples': 2673984, 'steps': 13926, 'loss/train': 1.233088731765747} 01/29/2022 06:42:30 - INFO - codeparrot_training - Step 13927: {'lr': 0.0004276192849390491, 'samples': 2674176, 'steps': 13927, 'loss/train': 1.3816435933113098} 01/29/2022 06:42:34 - INFO - codeparrot_training - Step 13928: {'lr': 0.0004276077699624534, 'samples': 2674368, 'steps': 13928, 'loss/train': 1.6738636493682861} 01/29/2022 06:42:38 - INFO - codeparrot_training - Step 13929: {'lr': 0.00042759625422504236, 'samples': 2674560, 'steps': 13929, 'loss/train': 1.9856300354003906} 01/29/2022 06:42:43 - INFO - codeparrot_training - Step 13930: {'lr': 0.00042758473772686533, 'samples': 2674752, 'steps': 13930, 'loss/train': 1.7657536268234253} 01/29/2022 06:42:48 - INFO - codeparrot_training - Step 13931: {'lr': 0.0004275732204679718, 'samples': 2674944, 'steps': 13931, 'loss/train': 1.679295837879181} 01/29/2022 06:42:52 - INFO - codeparrot_training - Step 13932: {'lr': 0.0004275617024484109, 'samples': 2675136, 'steps': 13932, 'loss/train': 1.623340129852295} 01/29/2022 06:42:57 - INFO - codeparrot_training - Step 13933: {'lr': 0.000427550183668232, 'samples': 2675328, 'steps': 13933, 'loss/train': 0.8277965784072876} 01/29/2022 06:43:01 - INFO - codeparrot_training - Step 13934: {'lr': 0.00042753866412748455, 'samples': 2675520, 'steps': 13934, 'loss/train': 1.7584657073020935} 01/29/2022 06:43:05 - INFO - codeparrot_training - Step 13935: {'lr': 0.00042752714382621784, 'samples': 2675712, 'steps': 13935, 'loss/train': 1.7482856512069702} 01/29/2022 06:43:11 - INFO - codeparrot_training - Step 13936: {'lr': 0.0004275156227644812, 'samples': 2675904, 'steps': 13936, 'loss/train': 1.6380223631858826} 01/29/2022 06:43:15 - INFO - codeparrot_training - Step 13937: {'lr': 0.00042750410094232394, 'samples': 2676096, 'steps': 13937, 'loss/train': 1.662246286869049} 01/29/2022 06:43:19 - INFO - codeparrot_training - Step 13938: {'lr': 0.0004274925783597956, 'samples': 2676288, 'steps': 13938, 'loss/train': 1.219499409198761} 01/29/2022 06:43:24 - INFO - codeparrot_training - Step 13939: {'lr': 0.0004274810550169453, 'samples': 2676480, 'steps': 13939, 'loss/train': 2.144312560558319} 01/29/2022 06:43:28 - INFO - codeparrot_training - Step 13940: {'lr': 0.00042746953091382254, 'samples': 2676672, 'steps': 13940, 'loss/train': 2.0512468814849854} 01/29/2022 06:43:35 - INFO - codeparrot_training - Step 13941: {'lr': 0.00042745800605047677, 'samples': 2676864, 'steps': 13941, 'loss/train': 1.7188398241996765} 01/29/2022 06:43:40 - INFO - codeparrot_training - Step 13942: {'lr': 0.00042744648042695717, 'samples': 2677056, 'steps': 13942, 'loss/train': 1.4251260459423065} 01/29/2022 06:43:44 - INFO - codeparrot_training - Step 13943: {'lr': 0.0004274349540433132, 'samples': 2677248, 'steps': 13943, 'loss/train': 1.8387685418128967} 01/29/2022 06:43:48 - INFO - codeparrot_training - Step 13944: {'lr': 0.00042742342689959425, 'samples': 2677440, 'steps': 13944, 'loss/train': 6.8143415451049805} 01/29/2022 06:43:52 - INFO - codeparrot_training - Step 13945: {'lr': 0.00042741189899584965, 'samples': 2677632, 'steps': 13945, 'loss/train': 1.2249946296215057} 01/29/2022 06:43:58 - INFO - codeparrot_training - Step 13946: {'lr': 0.00042740037033212877, 'samples': 2677824, 'steps': 13946, 'loss/train': 1.6012571454048157} 01/29/2022 06:44:02 - INFO - codeparrot_training - Step 13947: {'lr': 0.0004273888409084811, 'samples': 2678016, 'steps': 13947, 'loss/train': 1.6507907509803772} 01/29/2022 06:44:06 - INFO - codeparrot_training - Step 13948: {'lr': 0.0004273773107249559, 'samples': 2678208, 'steps': 13948, 'loss/train': 1.1889106929302216} 01/29/2022 06:44:11 - INFO - codeparrot_training - Step 13949: {'lr': 0.0004273657797816027, 'samples': 2678400, 'steps': 13949, 'loss/train': 0.2923434153199196} 01/29/2022 06:44:15 - INFO - codeparrot_training - Step 13950: {'lr': 0.0004273542480784708, 'samples': 2678592, 'steps': 13950, 'loss/train': 1.7759968042373657} 01/29/2022 06:44:22 - INFO - codeparrot_training - Step 13951: {'lr': 0.00042734271561560956, 'samples': 2678784, 'steps': 13951, 'loss/train': 2.1233906149864197} 01/29/2022 06:44:26 - INFO - codeparrot_training - Step 13952: {'lr': 0.00042733118239306845, 'samples': 2678976, 'steps': 13952, 'loss/train': 1.6124455332756042} 01/29/2022 06:44:30 - INFO - codeparrot_training - Step 13953: {'lr': 0.0004273196484108969, 'samples': 2679168, 'steps': 13953, 'loss/train': 1.9506588578224182} 01/29/2022 06:44:35 - INFO - codeparrot_training - Step 13954: {'lr': 0.00042730811366914435, 'samples': 2679360, 'steps': 13954, 'loss/train': 2.223779797554016} 01/29/2022 06:44:39 - INFO - codeparrot_training - Step 13955: {'lr': 0.0004272965781678601, 'samples': 2679552, 'steps': 13955, 'loss/train': 1.5854816436767578} 01/29/2022 06:44:44 - INFO - codeparrot_training - Step 13956: {'lr': 0.0004272850419070935, 'samples': 2679744, 'steps': 13956, 'loss/train': 1.0478631556034088} 01/29/2022 06:44:48 - INFO - codeparrot_training - Step 13957: {'lr': 0.00042727350488689416, 'samples': 2679936, 'steps': 13957, 'loss/train': 1.4337507784366608} 01/29/2022 06:44:53 - INFO - codeparrot_training - Step 13958: {'lr': 0.00042726196710731135, 'samples': 2680128, 'steps': 13958, 'loss/train': 0.986218124628067} 01/29/2022 06:44:57 - INFO - codeparrot_training - Step 13959: {'lr': 0.0004272504285683947, 'samples': 2680320, 'steps': 13959, 'loss/train': 1.5973058938980103} 01/29/2022 06:45:01 - INFO - codeparrot_training - Step 13960: {'lr': 0.0004272388892701934, 'samples': 2680512, 'steps': 13960, 'loss/train': 2.2290351390838623} 01/29/2022 06:45:06 - INFO - codeparrot_training - Step 13961: {'lr': 0.000427227349212757, 'samples': 2680704, 'steps': 13961, 'loss/train': 2.3811437487602234} 01/29/2022 06:45:11 - INFO - codeparrot_training - Step 13962: {'lr': 0.0004272158083961348, 'samples': 2680896, 'steps': 13962, 'loss/train': 1.830545425415039} 01/29/2022 06:45:15 - INFO - codeparrot_training - Step 13963: {'lr': 0.0004272042668203765, 'samples': 2681088, 'steps': 13963, 'loss/train': 1.398682951927185} 01/29/2022 06:45:19 - INFO - codeparrot_training - Step 13964: {'lr': 0.00042719272448553137, 'samples': 2681280, 'steps': 13964, 'loss/train': 2.417587637901306} 01/29/2022 06:45:23 - INFO - codeparrot_training - Step 13965: {'lr': 0.00042718118139164883, 'samples': 2681472, 'steps': 13965, 'loss/train': 1.1522625982761383} 01/29/2022 06:45:31 - INFO - codeparrot_training - Step 13966: {'lr': 0.00042716963753877836, 'samples': 2681664, 'steps': 13966, 'loss/train': 2.35971862077713} 01/29/2022 06:45:35 - INFO - codeparrot_training - Step 13967: {'lr': 0.0004271580929269695, 'samples': 2681856, 'steps': 13967, 'loss/train': 3.3612388372421265} 01/29/2022 06:45:39 - INFO - codeparrot_training - Step 13968: {'lr': 0.0004271465475562716, 'samples': 2682048, 'steps': 13968, 'loss/train': 1.4412042796611786} 01/29/2022 06:45:43 - INFO - codeparrot_training - Step 13969: {'lr': 0.00042713500142673404, 'samples': 2682240, 'steps': 13969, 'loss/train': 1.5895999073982239} 01/29/2022 06:45:48 - INFO - codeparrot_training - Step 13970: {'lr': 0.00042712345453840644, 'samples': 2682432, 'steps': 13970, 'loss/train': 1.6072758436203003} 01/29/2022 06:45:53 - INFO - codeparrot_training - Step 13971: {'lr': 0.00042711190689133827, 'samples': 2682624, 'steps': 13971, 'loss/train': 0.7396117597818375} 01/29/2022 06:45:57 - INFO - codeparrot_training - Step 13972: {'lr': 0.0004271003584855788, 'samples': 2682816, 'steps': 13972, 'loss/train': 1.9291670322418213} 01/29/2022 06:46:01 - INFO - codeparrot_training - Step 13973: {'lr': 0.0004270888093211778, 'samples': 2683008, 'steps': 13973, 'loss/train': 0.8546175062656403} 01/29/2022 06:46:06 - INFO - codeparrot_training - Step 13974: {'lr': 0.0004270772593981844, 'samples': 2683200, 'steps': 13974, 'loss/train': 1.679021179676056} 01/29/2022 06:46:10 - INFO - codeparrot_training - Step 13975: {'lr': 0.0004270657087166484, 'samples': 2683392, 'steps': 13975, 'loss/train': 2.1620165705680847} 01/29/2022 06:46:15 - INFO - codeparrot_training - Step 13976: {'lr': 0.000427054157276619, 'samples': 2683584, 'steps': 13976, 'loss/train': 2.1041248440742493} 01/29/2022 06:46:20 - INFO - codeparrot_training - Step 13977: {'lr': 0.0004270426050781458, 'samples': 2683776, 'steps': 13977, 'loss/train': 1.5583410859107971} 01/29/2022 06:46:24 - INFO - codeparrot_training - Step 13978: {'lr': 0.00042703105212127846, 'samples': 2683968, 'steps': 13978, 'loss/train': 0.26416316628456116} 01/29/2022 06:46:28 - INFO - codeparrot_training - Step 13979: {'lr': 0.0004270194984060662, 'samples': 2684160, 'steps': 13979, 'loss/train': 1.0128165185451508} 01/29/2022 06:46:32 - INFO - codeparrot_training - Step 13980: {'lr': 0.0004270079439325586, 'samples': 2684352, 'steps': 13980, 'loss/train': 1.7293363809585571} 01/29/2022 06:46:39 - INFO - codeparrot_training - Step 13981: {'lr': 0.0004269963887008053, 'samples': 2684544, 'steps': 13981, 'loss/train': 1.6636215448379517} 01/29/2022 06:46:44 - INFO - codeparrot_training - Step 13982: {'lr': 0.00042698483271085555, 'samples': 2684736, 'steps': 13982, 'loss/train': 1.2671833634376526} 01/29/2022 06:46:48 - INFO - codeparrot_training - Step 13983: {'lr': 0.0004269732759627589, 'samples': 2684928, 'steps': 13983, 'loss/train': 1.0052492022514343} 01/29/2022 06:46:52 - INFO - codeparrot_training - Step 13984: {'lr': 0.0004269617184565651, 'samples': 2685120, 'steps': 13984, 'loss/train': 2.2687541842460632} 01/29/2022 06:46:56 - INFO - codeparrot_training - Step 13985: {'lr': 0.00042695016019232343, 'samples': 2685312, 'steps': 13985, 'loss/train': 1.663730263710022} 01/29/2022 06:47:02 - INFO - codeparrot_training - Step 13986: {'lr': 0.0004269386011700834, 'samples': 2685504, 'steps': 13986, 'loss/train': 1.680025041103363} 01/29/2022 06:47:06 - INFO - codeparrot_training - Step 13987: {'lr': 0.00042692704138989467, 'samples': 2685696, 'steps': 13987, 'loss/train': 1.9605572819709778} 01/29/2022 06:47:10 - INFO - codeparrot_training - Step 13988: {'lr': 0.00042691548085180666, 'samples': 2685888, 'steps': 13988, 'loss/train': 1.881409764289856} 01/29/2022 06:47:15 - INFO - codeparrot_training - Step 13989: {'lr': 0.00042690391955586886, 'samples': 2686080, 'steps': 13989, 'loss/train': 1.729099988937378} 01/29/2022 06:47:20 - INFO - codeparrot_training - Step 13990: {'lr': 0.00042689235750213093, 'samples': 2686272, 'steps': 13990, 'loss/train': 1.8692232370376587} 01/29/2022 06:47:24 - INFO - codeparrot_training - Step 13991: {'lr': 0.0004268807946906422, 'samples': 2686464, 'steps': 13991, 'loss/train': 1.7243199348449707} 01/29/2022 06:47:29 - INFO - codeparrot_training - Step 13992: {'lr': 0.0004268692311214524, 'samples': 2686656, 'steps': 13992, 'loss/train': 1.4205359816551208} 01/29/2022 06:47:33 - INFO - codeparrot_training - Step 13993: {'lr': 0.00042685766679461095, 'samples': 2686848, 'steps': 13993, 'loss/train': 0.5105420798063278} 01/29/2022 06:47:37 - INFO - codeparrot_training - Step 13994: {'lr': 0.0004268461017101674, 'samples': 2687040, 'steps': 13994, 'loss/train': 2.2138343453407288} 01/29/2022 06:47:44 - INFO - codeparrot_training - Step 13995: {'lr': 0.00042683453586817136, 'samples': 2687232, 'steps': 13995, 'loss/train': 1.3615910410881042} 01/29/2022 06:47:49 - INFO - codeparrot_training - Step 13996: {'lr': 0.00042682296926867226, 'samples': 2687424, 'steps': 13996, 'loss/train': 1.6779361367225647} 01/29/2022 06:47:53 - INFO - codeparrot_training - Step 13997: {'lr': 0.0004268114019117197, 'samples': 2687616, 'steps': 13997, 'loss/train': 1.2415257096290588} 01/29/2022 06:47:57 - INFO - codeparrot_training - Step 13998: {'lr': 0.00042679983379736324, 'samples': 2687808, 'steps': 13998, 'loss/train': 1.56243234872818} 01/29/2022 06:48:01 - INFO - codeparrot_training - Step 13999: {'lr': 0.0004267882649256525, 'samples': 2688000, 'steps': 13999, 'loss/train': 1.2767246961593628} 01/29/2022 06:48:01 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 06:48:34 - WARNING - huggingface_hub.repository - Several commits (7) will be pushed upstream. 01/29/2022 06:48:34 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 06:49:43 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 63b9456..6a0d43b expert-dust-2 -> expert-dust-2 01/29/2022 06:49:51 - INFO - codeparrot_training - Step 14000: {'lr': 0.00042677669529663686, 'samples': 2688192, 'steps': 14000, 'loss/train': 2.1169806718826294} 01/29/2022 06:49:56 - INFO - codeparrot_training - Step 14001: {'lr': 0.0004267651249103661, 'samples': 2688384, 'steps': 14001, 'loss/train': 1.836431086063385} 01/29/2022 06:50:00 - INFO - codeparrot_training - Step 14002: {'lr': 0.00042675355376688964, 'samples': 2688576, 'steps': 14002, 'loss/train': 0.845357358455658} 01/29/2022 06:50:04 - INFO - codeparrot_training - Step 14003: {'lr': 0.000426741981866257, 'samples': 2688768, 'steps': 14003, 'loss/train': 1.6069564819335938} 01/29/2022 06:50:09 - INFO - codeparrot_training - Step 14004: {'lr': 0.00042673040920851793, 'samples': 2688960, 'steps': 14004, 'loss/train': 1.5618410110473633} 01/29/2022 06:50:13 - INFO - codeparrot_training - Step 14005: {'lr': 0.00042671883579372186, 'samples': 2689152, 'steps': 14005, 'loss/train': 2.4762521982192993} 01/29/2022 06:50:17 - INFO - codeparrot_training - Step 14006: {'lr': 0.00042670726162191843, 'samples': 2689344, 'steps': 14006, 'loss/train': 1.2904294431209564} 01/29/2022 06:50:22 - INFO - codeparrot_training - Step 14007: {'lr': 0.0004266956866931572, 'samples': 2689536, 'steps': 14007, 'loss/train': 1.917520523071289} 01/29/2022 06:50:27 - INFO - codeparrot_training - Step 14008: {'lr': 0.0004266841110074878, 'samples': 2689728, 'steps': 14008, 'loss/train': 2.0412177443504333} 01/29/2022 06:50:31 - INFO - codeparrot_training - Step 14009: {'lr': 0.0004266725345649597, 'samples': 2689920, 'steps': 14009, 'loss/train': 2.062922716140747} 01/29/2022 06:50:35 - INFO - codeparrot_training - Step 14010: {'lr': 0.0004266609573656226, 'samples': 2690112, 'steps': 14010, 'loss/train': 1.83206444978714} 01/29/2022 06:50:39 - INFO - codeparrot_training - Step 14011: {'lr': 0.000426649379409526, 'samples': 2690304, 'steps': 14011, 'loss/train': 2.543639302253723} 01/29/2022 06:50:46 - INFO - codeparrot_training - Step 14012: {'lr': 0.00042663780069671965, 'samples': 2690496, 'steps': 14012, 'loss/train': 2.5149096250534058} 01/29/2022 06:50:50 - INFO - codeparrot_training - Step 14013: {'lr': 0.000426626221227253, 'samples': 2690688, 'steps': 14013, 'loss/train': 2.0621105432510376} 01/29/2022 06:50:55 - INFO - codeparrot_training - Step 14014: {'lr': 0.00042661464100117566, 'samples': 2690880, 'steps': 14014, 'loss/train': 2.250849723815918} 01/29/2022 06:50:59 - INFO - codeparrot_training - Step 14015: {'lr': 0.00042660306001853735, 'samples': 2691072, 'steps': 14015, 'loss/train': 1.6557201147079468} 01/29/2022 06:51:04 - INFO - codeparrot_training - Step 14016: {'lr': 0.0004265914782793875, 'samples': 2691264, 'steps': 14016, 'loss/train': 1.4448899924755096} 01/29/2022 06:51:08 - INFO - codeparrot_training - Step 14017: {'lr': 0.000426579895783776, 'samples': 2691456, 'steps': 14017, 'loss/train': 1.7316282391548157} 01/29/2022 06:51:13 - INFO - codeparrot_training - Step 14018: {'lr': 0.0004265683125317521, 'samples': 2691648, 'steps': 14018, 'loss/train': 1.7778371572494507} 01/29/2022 06:51:17 - INFO - codeparrot_training - Step 14019: {'lr': 0.0004265567285233658, 'samples': 2691840, 'steps': 14019, 'loss/train': 2.024523675441742} 01/29/2022 06:51:21 - INFO - codeparrot_training - Step 14020: {'lr': 0.0004265451437586664, 'samples': 2692032, 'steps': 14020, 'loss/train': 2.714150369167328} 01/29/2022 06:51:28 - INFO - codeparrot_training - Step 14021: {'lr': 0.0004265335582377038, 'samples': 2692224, 'steps': 14021, 'loss/train': 1.8789412379264832} 01/29/2022 06:51:33 - INFO - codeparrot_training - Step 14022: {'lr': 0.0004265219719605273, 'samples': 2692416, 'steps': 14022, 'loss/train': 1.329720675945282} 01/29/2022 06:51:37 - INFO - codeparrot_training - Step 14023: {'lr': 0.0004265103849271869, 'samples': 2692608, 'steps': 14023, 'loss/train': 1.684112012386322} 01/29/2022 06:51:41 - INFO - codeparrot_training - Step 14024: {'lr': 0.000426498797137732, 'samples': 2692800, 'steps': 14024, 'loss/train': 2.1809452772140503} 01/29/2022 06:51:45 - INFO - codeparrot_training - Step 14025: {'lr': 0.0004264872085922122, 'samples': 2692992, 'steps': 14025, 'loss/train': 1.5497771501541138} 01/29/2022 06:51:50 - INFO - codeparrot_training - Step 14026: {'lr': 0.0004264756192906774, 'samples': 2693184, 'steps': 14026, 'loss/train': 1.856495976448059} 01/29/2022 06:51:55 - INFO - codeparrot_training - Step 14027: {'lr': 0.000426464029233177, 'samples': 2693376, 'steps': 14027, 'loss/train': 0.9734286367893219} 01/29/2022 06:51:59 - INFO - codeparrot_training - Step 14028: {'lr': 0.0004264524384197608, 'samples': 2693568, 'steps': 14028, 'loss/train': 1.8519535660743713} 01/29/2022 06:52:03 - INFO - codeparrot_training - Step 14029: {'lr': 0.0004264408468504783, 'samples': 2693760, 'steps': 14029, 'loss/train': 1.808992087841034} 01/29/2022 06:52:08 - INFO - codeparrot_training - Step 14030: {'lr': 0.00042642925452537927, 'samples': 2693952, 'steps': 14030, 'loss/train': 2.0513293147087097} 01/29/2022 06:52:12 - INFO - codeparrot_training - Step 14031: {'lr': 0.0004264176614445133, 'samples': 2694144, 'steps': 14031, 'loss/train': 1.6724052429199219} 01/29/2022 06:52:17 - INFO - codeparrot_training - Step 14032: {'lr': 0.0004264060676079302, 'samples': 2694336, 'steps': 14032, 'loss/train': 1.8992391228675842} 01/29/2022 06:52:22 - INFO - codeparrot_training - Step 14033: {'lr': 0.00042639447301567944, 'samples': 2694528, 'steps': 14033, 'loss/train': 2.3017271161079407} 01/29/2022 06:52:26 - INFO - codeparrot_training - Step 14034: {'lr': 0.0004263828776678108, 'samples': 2694720, 'steps': 14034, 'loss/train': 2.613349735736847} 01/29/2022 06:52:30 - INFO - codeparrot_training - Step 14035: {'lr': 0.00042637128156437385, 'samples': 2694912, 'steps': 14035, 'loss/train': 2.0985469222068787} 01/29/2022 06:52:34 - INFO - codeparrot_training - Step 14036: {'lr': 0.0004263596847054184, 'samples': 2695104, 'steps': 14036, 'loss/train': 1.5820140838623047} 01/29/2022 06:52:42 - INFO - codeparrot_training - Step 14037: {'lr': 0.00042634808709099403, 'samples': 2695296, 'steps': 14037, 'loss/train': 2.1593695878982544} 01/29/2022 06:52:46 - INFO - codeparrot_training - Step 14038: {'lr': 0.0004263364887211505, 'samples': 2695488, 'steps': 14038, 'loss/train': 1.2901479005813599} 01/29/2022 06:52:50 - INFO - codeparrot_training - Step 14039: {'lr': 0.0004263248895959374, 'samples': 2695680, 'steps': 14039, 'loss/train': 2.405539870262146} 01/29/2022 06:52:54 - INFO - codeparrot_training - Step 14040: {'lr': 0.0004263132897154044, 'samples': 2695872, 'steps': 14040, 'loss/train': 0.49504758417606354} 01/29/2022 06:53:00 - INFO - codeparrot_training - Step 14041: {'lr': 0.0004263016890796014, 'samples': 2696064, 'steps': 14041, 'loss/train': 1.4246743619441986} 01/29/2022 06:53:04 - INFO - codeparrot_training - Step 14042: {'lr': 0.0004262900876885778, 'samples': 2696256, 'steps': 14042, 'loss/train': 0.4990568608045578} 01/29/2022 06:53:09 - INFO - codeparrot_training - Step 14043: {'lr': 0.0004262784855423836, 'samples': 2696448, 'steps': 14043, 'loss/train': 7.342336893081665} 01/29/2022 06:53:13 - INFO - codeparrot_training - Step 14044: {'lr': 0.00042626688264106816, 'samples': 2696640, 'steps': 14044, 'loss/train': 2.1736426949501038} 01/29/2022 06:53:17 - INFO - codeparrot_training - Step 14045: {'lr': 0.00042625527898468155, 'samples': 2696832, 'steps': 14045, 'loss/train': 1.6477206945419312} 01/29/2022 06:53:21 - INFO - codeparrot_training - Step 14046: {'lr': 0.0004262436745732732, 'samples': 2697024, 'steps': 14046, 'loss/train': 1.7633471488952637} 01/29/2022 06:53:27 - INFO - codeparrot_training - Step 14047: {'lr': 0.00042623206940689285, 'samples': 2697216, 'steps': 14047, 'loss/train': 1.6829102039337158} 01/29/2022 06:53:31 - INFO - codeparrot_training - Step 14048: {'lr': 0.00042622046348559034, 'samples': 2697408, 'steps': 14048, 'loss/train': 1.6667364835739136} 01/29/2022 06:53:35 - INFO - codeparrot_training - Step 14049: {'lr': 0.0004262088568094153, 'samples': 2697600, 'steps': 14049, 'loss/train': 2.525112569332123} 01/29/2022 06:53:39 - INFO - codeparrot_training - Step 14050: {'lr': 0.0004261972493784175, 'samples': 2697792, 'steps': 14050, 'loss/train': 0.5889054983854294} 01/29/2022 06:53:44 - INFO - codeparrot_training - Step 14051: {'lr': 0.0004261856411926467, 'samples': 2697984, 'steps': 14051, 'loss/train': 1.4122178256511688} 01/29/2022 06:53:51 - INFO - codeparrot_training - Step 14052: {'lr': 0.0004261740322521525, 'samples': 2698176, 'steps': 14052, 'loss/train': 2.2008363604545593} 01/29/2022 06:53:55 - INFO - codeparrot_training - Step 14053: {'lr': 0.00042616242255698463, 'samples': 2698368, 'steps': 14053, 'loss/train': 1.883333444595337} 01/29/2022 06:54:00 - INFO - codeparrot_training - Step 14054: {'lr': 0.0004261508121071929, 'samples': 2698560, 'steps': 14054, 'loss/train': 1.673233687877655} 01/29/2022 06:54:04 - INFO - codeparrot_training - Step 14055: {'lr': 0.00042613920090282706, 'samples': 2698752, 'steps': 14055, 'loss/train': 1.112282395362854} 01/29/2022 06:54:10 - INFO - codeparrot_training - Step 14056: {'lr': 0.0004261275889439368, 'samples': 2698944, 'steps': 14056, 'loss/train': 0.9557541310787201} 01/29/2022 06:54:14 - INFO - codeparrot_training - Step 14057: {'lr': 0.0004261159762305719, 'samples': 2699136, 'steps': 14057, 'loss/train': 1.6814446449279785} 01/29/2022 06:54:18 - INFO - codeparrot_training - Step 14058: {'lr': 0.00042610436276278196, 'samples': 2699328, 'steps': 14058, 'loss/train': 1.947301983833313} 01/29/2022 06:54:23 - INFO - codeparrot_training - Step 14059: {'lr': 0.00042609274854061695, 'samples': 2699520, 'steps': 14059, 'loss/train': 2.1075514554977417} 01/29/2022 06:54:27 - INFO - codeparrot_training - Step 14060: {'lr': 0.0004260811335641266, 'samples': 2699712, 'steps': 14060, 'loss/train': 1.6866005659103394} 01/29/2022 06:54:31 - INFO - codeparrot_training - Step 14061: {'lr': 0.00042606951783336045, 'samples': 2699904, 'steps': 14061, 'loss/train': 1.7970868349075317} 01/29/2022 06:54:37 - INFO - codeparrot_training - Step 14062: {'lr': 0.0004260579013483684, 'samples': 2700096, 'steps': 14062, 'loss/train': 1.644244372844696} 01/29/2022 06:54:41 - INFO - codeparrot_training - Step 14063: {'lr': 0.0004260462841092003, 'samples': 2700288, 'steps': 14063, 'loss/train': 1.5741847157478333} 01/29/2022 06:54:45 - INFO - codeparrot_training - Step 14064: {'lr': 0.00042603466611590575, 'samples': 2700480, 'steps': 14064, 'loss/train': 1.977423906326294} 01/29/2022 06:54:49 - INFO - codeparrot_training - Step 14065: {'lr': 0.00042602304736853464, 'samples': 2700672, 'steps': 14065, 'loss/train': 1.8599048852920532} 01/29/2022 06:54:54 - INFO - codeparrot_training - Step 14066: {'lr': 0.00042601142786713664, 'samples': 2700864, 'steps': 14066, 'loss/train': 3.28290331363678} 01/29/2022 06:54:58 - INFO - codeparrot_training - Step 14067: {'lr': 0.0004259998076117616, 'samples': 2701056, 'steps': 14067, 'loss/train': 2.068389058113098} 01/29/2022 06:55:05 - INFO - codeparrot_training - Step 14068: {'lr': 0.00042598818660245926, 'samples': 2701248, 'steps': 14068, 'loss/train': 1.659691572189331} 01/29/2022 06:55:09 - INFO - codeparrot_training - Step 14069: {'lr': 0.00042597656483927936, 'samples': 2701440, 'steps': 14069, 'loss/train': 0.4879726320505142} 01/29/2022 06:55:14 - INFO - codeparrot_training - Step 14070: {'lr': 0.0004259649423222718, 'samples': 2701632, 'steps': 14070, 'loss/train': 1.7999029755592346} 01/29/2022 06:55:18 - INFO - codeparrot_training - Step 14071: {'lr': 0.0004259533190514863, 'samples': 2701824, 'steps': 14071, 'loss/train': 2.4615560173988342} 01/29/2022 06:55:23 - INFO - codeparrot_training - Step 14072: {'lr': 0.00042594169502697265, 'samples': 2702016, 'steps': 14072, 'loss/train': 2.09393048286438} 01/29/2022 06:55:28 - INFO - codeparrot_training - Step 14073: {'lr': 0.0004259300702487806, 'samples': 2702208, 'steps': 14073, 'loss/train': 1.7787063717842102} 01/29/2022 06:55:32 - INFO - codeparrot_training - Step 14074: {'lr': 0.00042591844471696005, 'samples': 2702400, 'steps': 14074, 'loss/train': 1.4904887080192566} 01/29/2022 06:55:36 - INFO - codeparrot_training - Step 14075: {'lr': 0.00042590681843156073, 'samples': 2702592, 'steps': 14075, 'loss/train': 1.0375717878341675} 01/29/2022 06:55:40 - INFO - codeparrot_training - Step 14076: {'lr': 0.00042589519139263246, 'samples': 2702784, 'steps': 14076, 'loss/train': 2.512045383453369} 01/29/2022 06:55:46 - INFO - codeparrot_training - Step 14077: {'lr': 0.0004258835636002251, 'samples': 2702976, 'steps': 14077, 'loss/train': 2.192805826663971} 01/29/2022 06:55:50 - INFO - codeparrot_training - Step 14078: {'lr': 0.0004258719350543883, 'samples': 2703168, 'steps': 14078, 'loss/train': 1.040602058172226} 01/29/2022 06:55:54 - INFO - codeparrot_training - Step 14079: {'lr': 0.00042586030575517196, 'samples': 2703360, 'steps': 14079, 'loss/train': 1.8828288316726685} 01/29/2022 06:55:58 - INFO - codeparrot_training - Step 14080: {'lr': 0.00042584867570262595, 'samples': 2703552, 'steps': 14080, 'loss/train': 2.1292696595191956} 01/29/2022 06:56:03 - INFO - codeparrot_training - Step 14081: {'lr': 0.00042583704489680007, 'samples': 2703744, 'steps': 14081, 'loss/train': 2.079941689968109} 01/29/2022 06:56:10 - INFO - codeparrot_training - Step 14082: {'lr': 0.00042582541333774414, 'samples': 2703936, 'steps': 14082, 'loss/train': 1.4402447640895844} 01/29/2022 06:56:14 - INFO - codeparrot_training - Step 14083: {'lr': 0.0004258137810255079, 'samples': 2704128, 'steps': 14083, 'loss/train': 1.5756098628044128} 01/29/2022 06:56:18 - INFO - codeparrot_training - Step 14084: {'lr': 0.0004258021479601414, 'samples': 2704320, 'steps': 14084, 'loss/train': 2.6888524889945984} 01/29/2022 06:56:22 - INFO - codeparrot_training - Step 14085: {'lr': 0.00042579051414169417, 'samples': 2704512, 'steps': 14085, 'loss/train': 1.030411720275879} 01/29/2022 06:56:27 - INFO - codeparrot_training - Step 14086: {'lr': 0.0004257788795702162, 'samples': 2704704, 'steps': 14086, 'loss/train': 2.10004198551178} 01/29/2022 06:56:32 - INFO - codeparrot_training - Step 14087: {'lr': 0.0004257672442457574, 'samples': 2704896, 'steps': 14087, 'loss/train': 1.6206852793693542} 01/29/2022 06:56:36 - INFO - codeparrot_training - Step 14088: {'lr': 0.00042575560816836755, 'samples': 2705088, 'steps': 14088, 'loss/train': 2.2403279542922974} 01/29/2022 06:56:40 - INFO - codeparrot_training - Step 14089: {'lr': 0.00042574397133809646, 'samples': 2705280, 'steps': 14089, 'loss/train': 2.36732017993927} 01/29/2022 06:56:45 - INFO - codeparrot_training - Step 14090: {'lr': 0.000425732333754994, 'samples': 2705472, 'steps': 14090, 'loss/train': 1.608970284461975} 01/29/2022 06:56:49 - INFO - codeparrot_training - Step 14091: {'lr': 0.00042572069541911, 'samples': 2705664, 'steps': 14091, 'loss/train': 1.5642285346984863} 01/29/2022 06:56:54 - INFO - codeparrot_training - Step 14092: {'lr': 0.0004257090563304943, 'samples': 2705856, 'steps': 14092, 'loss/train': 1.9042262434959412} 01/29/2022 06:56:58 - INFO - codeparrot_training - Step 14093: {'lr': 0.0004256974164891969, 'samples': 2706048, 'steps': 14093, 'loss/train': 1.8361680507659912} 01/29/2022 06:57:03 - INFO - codeparrot_training - Step 14094: {'lr': 0.00042568577589526744, 'samples': 2706240, 'steps': 14094, 'loss/train': 1.5460293889045715} 01/29/2022 06:57:07 - INFO - codeparrot_training - Step 14095: {'lr': 0.00042567413454875605, 'samples': 2706432, 'steps': 14095, 'loss/train': 1.3389974534511566} 01/29/2022 06:57:11 - INFO - codeparrot_training - Step 14096: {'lr': 0.00042566249244971235, 'samples': 2706624, 'steps': 14096, 'loss/train': 1.3748991787433624} 01/29/2022 06:57:18 - INFO - codeparrot_training - Step 14097: {'lr': 0.0004256508495981863, 'samples': 2706816, 'steps': 14097, 'loss/train': 1.7849467992782593} 01/29/2022 06:57:23 - INFO - codeparrot_training - Step 14098: {'lr': 0.00042563920599422776, 'samples': 2707008, 'steps': 14098, 'loss/train': 1.1576943397521973} 01/29/2022 06:57:27 - INFO - codeparrot_training - Step 14099: {'lr': 0.00042562756163788673, 'samples': 2707200, 'steps': 14099, 'loss/train': 2.092700242996216} 01/29/2022 06:57:31 - INFO - codeparrot_training - Step 14100: {'lr': 0.00042561591652921294, 'samples': 2707392, 'steps': 14100, 'loss/train': 2.0541849732398987} 01/29/2022 06:57:35 - INFO - codeparrot_training - Step 14101: {'lr': 0.00042560427066825636, 'samples': 2707584, 'steps': 14101, 'loss/train': 2.0057618021965027} 01/29/2022 06:57:41 - INFO - codeparrot_training - Step 14102: {'lr': 0.0004255926240550668, 'samples': 2707776, 'steps': 14102, 'loss/train': 1.147643119096756} 01/29/2022 06:57:45 - INFO - codeparrot_training - Step 14103: {'lr': 0.0004255809766896942, 'samples': 2707968, 'steps': 14103, 'loss/train': 2.2244905829429626} 01/29/2022 06:57:49 - INFO - codeparrot_training - Step 14104: {'lr': 0.00042556932857218855, 'samples': 2708160, 'steps': 14104, 'loss/train': 2.064887523651123} 01/29/2022 06:57:53 - INFO - codeparrot_training - Step 14105: {'lr': 0.0004255576797025995, 'samples': 2708352, 'steps': 14105, 'loss/train': 1.799609363079071} 01/29/2022 06:57:58 - INFO - codeparrot_training - Step 14106: {'lr': 0.0004255460300809772, 'samples': 2708544, 'steps': 14106, 'loss/train': 2.153924345970154} 01/29/2022 06:58:04 - INFO - codeparrot_training - Step 14107: {'lr': 0.00042553437970737143, 'samples': 2708736, 'steps': 14107, 'loss/train': 2.618758499622345} 01/29/2022 06:58:09 - INFO - codeparrot_training - Step 14108: {'lr': 0.00042552272858183203, 'samples': 2708928, 'steps': 14108, 'loss/train': 1.4411762058734894} 01/29/2022 06:58:13 - INFO - codeparrot_training - Step 14109: {'lr': 0.0004255110767044091, 'samples': 2709120, 'steps': 14109, 'loss/train': 2.0211979150772095} 01/29/2022 06:58:17 - INFO - codeparrot_training - Step 14110: {'lr': 0.0004254994240751524, 'samples': 2709312, 'steps': 14110, 'loss/train': 0.9262122809886932} 01/29/2022 06:58:21 - INFO - codeparrot_training - Step 14111: {'lr': 0.00042548777069411194, 'samples': 2709504, 'steps': 14111, 'loss/train': 0.6807265877723694} 01/29/2022 06:58:27 - INFO - codeparrot_training - Step 14112: {'lr': 0.0004254761165613375, 'samples': 2709696, 'steps': 14112, 'loss/train': 1.3376303315162659} 01/29/2022 06:58:31 - INFO - codeparrot_training - Step 14113: {'lr': 0.00042546446167687914, 'samples': 2709888, 'steps': 14113, 'loss/train': 1.515503704547882} 01/29/2022 06:58:35 - INFO - codeparrot_training - Step 14114: {'lr': 0.00042545280604078673, 'samples': 2710080, 'steps': 14114, 'loss/train': 1.3604789972305298} 01/29/2022 06:58:40 - INFO - codeparrot_training - Step 14115: {'lr': 0.0004254411496531103, 'samples': 2710272, 'steps': 14115, 'loss/train': 1.9585157632827759} 01/29/2022 06:58:44 - INFO - codeparrot_training - Step 14116: {'lr': 0.0004254294925138996, 'samples': 2710464, 'steps': 14116, 'loss/train': 1.8706591129302979} 01/29/2022 06:58:49 - INFO - codeparrot_training - Step 14117: {'lr': 0.00042541783462320473, 'samples': 2710656, 'steps': 14117, 'loss/train': 2.333977997303009} 01/29/2022 06:58:53 - INFO - codeparrot_training - Step 14118: {'lr': 0.00042540617598107544, 'samples': 2710848, 'steps': 14118, 'loss/train': 1.9544286131858826} 01/29/2022 06:58:58 - INFO - codeparrot_training - Step 14119: {'lr': 0.00042539451658756195, 'samples': 2711040, 'steps': 14119, 'loss/train': 1.436524897813797} 01/29/2022 06:59:02 - INFO - codeparrot_training - Step 14120: {'lr': 0.000425382856442714, 'samples': 2711232, 'steps': 14120, 'loss/train': 1.2643276154994965} 01/29/2022 06:59:06 - INFO - codeparrot_training - Step 14121: {'lr': 0.0004253711955465815, 'samples': 2711424, 'steps': 14121, 'loss/train': 1.1351717412471771} 01/29/2022 06:59:11 - INFO - codeparrot_training - Step 14122: {'lr': 0.00042535953389921454, 'samples': 2711616, 'steps': 14122, 'loss/train': 2.0668383836746216} 01/29/2022 06:59:16 - INFO - codeparrot_training - Step 14123: {'lr': 0.000425347871500663, 'samples': 2711808, 'steps': 14123, 'loss/train': 1.0407064855098724} 01/29/2022 06:59:20 - INFO - codeparrot_training - Step 14124: {'lr': 0.0004253362083509769, 'samples': 2712000, 'steps': 14124, 'loss/train': 2.288865029811859} 01/29/2022 06:59:24 - INFO - codeparrot_training - Step 14125: {'lr': 0.0004253245444502061, 'samples': 2712192, 'steps': 14125, 'loss/train': 1.6580927968025208} 01/29/2022 06:59:28 - INFO - codeparrot_training - Step 14126: {'lr': 0.00042531287979840065, 'samples': 2712384, 'steps': 14126, 'loss/train': 2.1414830088615417} 01/29/2022 06:59:36 - INFO - codeparrot_training - Step 14127: {'lr': 0.0004253012143956105, 'samples': 2712576, 'steps': 14127, 'loss/train': 0.6233404576778412} 01/29/2022 06:59:40 - INFO - codeparrot_training - Step 14128: {'lr': 0.0004252895482418856, 'samples': 2712768, 'steps': 14128, 'loss/train': 2.0842537879943848} 01/29/2022 06:59:44 - INFO - codeparrot_training - Step 14129: {'lr': 0.00042527788133727595, 'samples': 2712960, 'steps': 14129, 'loss/train': 1.594681441783905} 01/29/2022 06:59:48 - INFO - codeparrot_training - Step 14130: {'lr': 0.0004252662136818315, 'samples': 2713152, 'steps': 14130, 'loss/train': 0.5353028923273087} 01/29/2022 06:59:53 - INFO - codeparrot_training - Step 14131: {'lr': 0.00042525454527560225, 'samples': 2713344, 'steps': 14131, 'loss/train': 2.5177250504493713} 01/29/2022 06:59:58 - INFO - codeparrot_training - Step 14132: {'lr': 0.0004252428761186382, 'samples': 2713536, 'steps': 14132, 'loss/train': 2.6257204413414} 01/29/2022 07:00:02 - INFO - codeparrot_training - Step 14133: {'lr': 0.00042523120621098924, 'samples': 2713728, 'steps': 14133, 'loss/train': 1.5632851123809814} 01/29/2022 07:00:06 - INFO - codeparrot_training - Step 14134: {'lr': 0.0004252195355527055, 'samples': 2713920, 'steps': 14134, 'loss/train': 1.971772849559784} 01/29/2022 07:00:11 - INFO - codeparrot_training - Step 14135: {'lr': 0.0004252078641438369, 'samples': 2714112, 'steps': 14135, 'loss/train': 2.4208556413650513} 01/29/2022 07:00:15 - INFO - codeparrot_training - Step 14136: {'lr': 0.00042519619198443337, 'samples': 2714304, 'steps': 14136, 'loss/train': 2.1377116441726685} 01/29/2022 07:00:20 - INFO - codeparrot_training - Step 14137: {'lr': 0.0004251845190745451, 'samples': 2714496, 'steps': 14137, 'loss/train': 1.9605764150619507} 01/29/2022 07:00:25 - INFO - codeparrot_training - Step 14138: {'lr': 0.00042517284541422195, 'samples': 2714688, 'steps': 14138, 'loss/train': 1.4420650005340576} 01/29/2022 07:00:29 - INFO - codeparrot_training - Step 14139: {'lr': 0.00042516117100351394, 'samples': 2714880, 'steps': 14139, 'loss/train': 1.8975053429603577} 01/29/2022 07:00:33 - INFO - codeparrot_training - Step 14140: {'lr': 0.0004251494958424711, 'samples': 2715072, 'steps': 14140, 'loss/train': 2.212775945663452} 01/29/2022 07:00:37 - INFO - codeparrot_training - Step 14141: {'lr': 0.0004251378199311434, 'samples': 2715264, 'steps': 14141, 'loss/train': 1.777789056301117} 01/29/2022 07:00:44 - INFO - codeparrot_training - Step 14142: {'lr': 0.0004251261432695809, 'samples': 2715456, 'steps': 14142, 'loss/train': 1.7369860410690308} 01/29/2022 07:00:49 - INFO - codeparrot_training - Step 14143: {'lr': 0.00042511446585783363, 'samples': 2715648, 'steps': 14143, 'loss/train': 1.7869589924812317} 01/29/2022 07:00:53 - INFO - codeparrot_training - Step 14144: {'lr': 0.0004251027876959516, 'samples': 2715840, 'steps': 14144, 'loss/train': 1.6570785641670227} 01/29/2022 07:00:57 - INFO - codeparrot_training - Step 14145: {'lr': 0.0004250911087839848, 'samples': 2716032, 'steps': 14145, 'loss/train': 1.7279917001724243} 01/29/2022 07:01:01 - INFO - codeparrot_training - Step 14146: {'lr': 0.0004250794291219833, 'samples': 2716224, 'steps': 14146, 'loss/train': 1.2545511424541473} 01/29/2022 07:01:07 - INFO - codeparrot_training - Step 14147: {'lr': 0.00042506774870999716, 'samples': 2716416, 'steps': 14147, 'loss/train': 1.783622145652771} 01/29/2022 07:01:11 - INFO - codeparrot_training - Step 14148: {'lr': 0.00042505606754807634, 'samples': 2716608, 'steps': 14148, 'loss/train': 1.9465527534484863} 01/29/2022 07:01:15 - INFO - codeparrot_training - Step 14149: {'lr': 0.00042504438563627093, 'samples': 2716800, 'steps': 14149, 'loss/train': 1.8254917860031128} 01/29/2022 07:01:20 - INFO - codeparrot_training - Step 14150: {'lr': 0.0004250327029746309, 'samples': 2716992, 'steps': 14150, 'loss/train': 1.4239225387573242} 01/29/2022 07:01:24 - INFO - codeparrot_training - Step 14151: {'lr': 0.0004250210195632064, 'samples': 2717184, 'steps': 14151, 'loss/train': 1.4137829840183258} 01/29/2022 07:01:30 - INFO - codeparrot_training - Step 14152: {'lr': 0.00042500933540204745, 'samples': 2717376, 'steps': 14152, 'loss/train': 0.9884549081325531} 01/29/2022 07:01:34 - INFO - codeparrot_training - Step 14153: {'lr': 0.00042499765049120396, 'samples': 2717568, 'steps': 14153, 'loss/train': 2.2235249876976013} 01/29/2022 07:01:38 - INFO - codeparrot_training - Step 14154: {'lr': 0.0004249859648307263, 'samples': 2717760, 'steps': 14154, 'loss/train': 1.621131420135498} 01/29/2022 07:01:43 - INFO - codeparrot_training - Step 14155: {'lr': 0.0004249742784206642, 'samples': 2717952, 'steps': 14155, 'loss/train': 1.646874725818634} 01/29/2022 07:01:47 - INFO - codeparrot_training - Step 14156: {'lr': 0.00042496259126106786, 'samples': 2718144, 'steps': 14156, 'loss/train': 2.2478561997413635} 01/29/2022 07:01:52 - INFO - codeparrot_training - Step 14157: {'lr': 0.00042495090335198735, 'samples': 2718336, 'steps': 14157, 'loss/train': 1.5520548820495605} 01/29/2022 07:01:56 - INFO - codeparrot_training - Step 14158: {'lr': 0.0004249392146934726, 'samples': 2718528, 'steps': 14158, 'loss/train': 1.575212001800537} 01/29/2022 07:02:01 - INFO - codeparrot_training - Step 14159: {'lr': 0.000424927525285574, 'samples': 2718720, 'steps': 14159, 'loss/train': 1.5866180062294006} 01/29/2022 07:02:05 - INFO - codeparrot_training - Step 14160: {'lr': 0.00042491583512834137, 'samples': 2718912, 'steps': 14160, 'loss/train': 2.0038368701934814} 01/29/2022 07:02:09 - INFO - codeparrot_training - Step 14161: {'lr': 0.00042490414422182484, 'samples': 2719104, 'steps': 14161, 'loss/train': 2.359514594078064} 01/29/2022 07:02:14 - INFO - codeparrot_training - Step 14162: {'lr': 0.00042489245256607447, 'samples': 2719296, 'steps': 14162, 'loss/train': 0.8875291049480438} 01/29/2022 07:02:19 - INFO - codeparrot_training - Step 14163: {'lr': 0.0004248807601611404, 'samples': 2719488, 'steps': 14163, 'loss/train': 1.6662034392356873} 01/29/2022 07:02:23 - INFO - codeparrot_training - Step 14164: {'lr': 0.0004248690670070726, 'samples': 2719680, 'steps': 14164, 'loss/train': 1.3840234279632568} 01/29/2022 07:02:27 - INFO - codeparrot_training - Step 14165: {'lr': 0.00042485737310392135, 'samples': 2719872, 'steps': 14165, 'loss/train': 1.8776739835739136} 01/29/2022 07:02:31 - INFO - codeparrot_training - Step 14166: {'lr': 0.0004248456784517366, 'samples': 2720064, 'steps': 14166, 'loss/train': 1.4214701056480408} 01/29/2022 07:02:37 - INFO - codeparrot_training - Step 14167: {'lr': 0.00042483398305056847, 'samples': 2720256, 'steps': 14167, 'loss/train': 1.5598211288452148} 01/29/2022 07:02:42 - INFO - codeparrot_training - Step 14168: {'lr': 0.0004248222869004671, 'samples': 2720448, 'steps': 14168, 'loss/train': 2.057329773902893} 01/29/2022 07:02:46 - INFO - codeparrot_training - Step 14169: {'lr': 0.00042481059000148253, 'samples': 2720640, 'steps': 14169, 'loss/train': 1.274716168642044} 01/29/2022 07:02:50 - INFO - codeparrot_training - Step 14170: {'lr': 0.00042479889235366486, 'samples': 2720832, 'steps': 14170, 'loss/train': 1.611600637435913} 01/29/2022 07:02:55 - INFO - codeparrot_training - Step 14171: {'lr': 0.0004247871939570643, 'samples': 2721024, 'steps': 14171, 'loss/train': 1.3940664231777191} 01/29/2022 07:03:00 - INFO - codeparrot_training - Step 14172: {'lr': 0.00042477549481173093, 'samples': 2721216, 'steps': 14172, 'loss/train': 1.8574756979942322} 01/29/2022 07:03:04 - INFO - codeparrot_training - Step 14173: {'lr': 0.00042476379491771475, 'samples': 2721408, 'steps': 14173, 'loss/train': 1.7594838738441467} 01/29/2022 07:03:08 - INFO - codeparrot_training - Step 14174: {'lr': 0.00042475209427506614, 'samples': 2721600, 'steps': 14174, 'loss/train': 2.095754384994507} 01/29/2022 07:03:12 - INFO - codeparrot_training - Step 14175: {'lr': 0.00042474039288383484, 'samples': 2721792, 'steps': 14175, 'loss/train': 3.63670814037323} 01/29/2022 07:03:18 - INFO - codeparrot_training - Step 14176: {'lr': 0.0004247286907440713, 'samples': 2721984, 'steps': 14176, 'loss/train': 2.2964420914649963} 01/29/2022 07:03:22 - INFO - codeparrot_training - Step 14177: {'lr': 0.00042471698785582546, 'samples': 2722176, 'steps': 14177, 'loss/train': 1.8006359338760376} 01/29/2022 07:03:27 - INFO - codeparrot_training - Step 14178: {'lr': 0.00042470528421914767, 'samples': 2722368, 'steps': 14178, 'loss/train': 2.1566432118415833} 01/29/2022 07:03:31 - INFO - codeparrot_training - Step 14179: {'lr': 0.0004246935798340877, 'samples': 2722560, 'steps': 14179, 'loss/train': 1.7954617738723755} 01/29/2022 07:03:35 - INFO - codeparrot_training - Step 14180: {'lr': 0.0004246818747006961, 'samples': 2722752, 'steps': 14180, 'loss/train': 0.6667154878377914} 01/29/2022 07:03:39 - INFO - codeparrot_training - Step 14181: {'lr': 0.0004246701688190227, 'samples': 2722944, 'steps': 14181, 'loss/train': 1.5585730075836182} 01/29/2022 07:03:45 - INFO - codeparrot_training - Step 14182: {'lr': 0.0004246584621891179, 'samples': 2723136, 'steps': 14182, 'loss/train': 2.281168520450592} 01/29/2022 07:03:49 - INFO - codeparrot_training - Step 14183: {'lr': 0.00042464675481103154, 'samples': 2723328, 'steps': 14183, 'loss/train': 2.7013328075408936} 01/29/2022 07:03:53 - INFO - codeparrot_training - Step 14184: {'lr': 0.00042463504668481403, 'samples': 2723520, 'steps': 14184, 'loss/train': 1.7574008703231812} 01/29/2022 07:03:57 - INFO - codeparrot_training - Step 14185: {'lr': 0.00042462333781051535, 'samples': 2723712, 'steps': 14185, 'loss/train': 1.2568769752979279} 01/29/2022 07:04:02 - INFO - codeparrot_training - Step 14186: {'lr': 0.00042461162818818585, 'samples': 2723904, 'steps': 14186, 'loss/train': 2.000056028366089} 01/29/2022 07:04:06 - INFO - codeparrot_training - Step 14187: {'lr': 0.0004245999178178755, 'samples': 2724096, 'steps': 14187, 'loss/train': 1.640231430530548} 01/29/2022 07:04:12 - INFO - codeparrot_training - Step 14188: {'lr': 0.0004245882066996346, 'samples': 2724288, 'steps': 14188, 'loss/train': 1.7075157165527344} 01/29/2022 07:04:16 - INFO - codeparrot_training - Step 14189: {'lr': 0.0004245764948335132, 'samples': 2724480, 'steps': 14189, 'loss/train': 1.6965996623039246} 01/29/2022 07:04:21 - INFO - codeparrot_training - Step 14190: {'lr': 0.0004245647822195616, 'samples': 2724672, 'steps': 14190, 'loss/train': 1.271152138710022} 01/29/2022 07:04:25 - INFO - codeparrot_training - Step 14191: {'lr': 0.00042455306885782985, 'samples': 2724864, 'steps': 14191, 'loss/train': 1.6756466031074524} 01/29/2022 07:04:30 - INFO - codeparrot_training - Step 14192: {'lr': 0.00042454135474836817, 'samples': 2725056, 'steps': 14192, 'loss/train': 1.3592062890529633} 01/29/2022 07:04:35 - INFO - codeparrot_training - Step 14193: {'lr': 0.00042452963989122685, 'samples': 2725248, 'steps': 14193, 'loss/train': 1.8394437432289124} 01/29/2022 07:04:39 - INFO - codeparrot_training - Step 14194: {'lr': 0.00042451792428645587, 'samples': 2725440, 'steps': 14194, 'loss/train': 2.1627895832061768} 01/29/2022 07:04:43 - INFO - codeparrot_training - Step 14195: {'lr': 0.0004245062079341055, 'samples': 2725632, 'steps': 14195, 'loss/train': 2.1608190536499023} 01/29/2022 07:04:47 - INFO - codeparrot_training - Step 14196: {'lr': 0.000424494490834226, 'samples': 2725824, 'steps': 14196, 'loss/train': 2.37788325548172} 01/29/2022 07:04:52 - INFO - codeparrot_training - Step 14197: {'lr': 0.0004244827729868675, 'samples': 2726016, 'steps': 14197, 'loss/train': 1.0551278293132782} 01/29/2022 07:04:58 - INFO - codeparrot_training - Step 14198: {'lr': 0.00042447105439208024, 'samples': 2726208, 'steps': 14198, 'loss/train': 2.0224252939224243} 01/29/2022 07:05:02 - INFO - codeparrot_training - Step 14199: {'lr': 0.0004244593350499143, 'samples': 2726400, 'steps': 14199, 'loss/train': 1.918438732624054} 01/29/2022 07:05:06 - INFO - codeparrot_training - Step 14200: {'lr': 0.00042444761496042004, 'samples': 2726592, 'steps': 14200, 'loss/train': 0.9171465039253235} 01/29/2022 07:05:11 - INFO - codeparrot_training - Step 14201: {'lr': 0.0004244358941236476, 'samples': 2726784, 'steps': 14201, 'loss/train': 1.4991580545902252} 01/29/2022 07:05:16 - INFO - codeparrot_training - Step 14202: {'lr': 0.00042442417253964713, 'samples': 2726976, 'steps': 14202, 'loss/train': 0.8783180415630341} 01/29/2022 07:05:20 - INFO - codeparrot_training - Step 14203: {'lr': 0.00042441245020846885, 'samples': 2727168, 'steps': 14203, 'loss/train': 2.0106319785118103} 01/29/2022 07:05:24 - INFO - codeparrot_training - Step 14204: {'lr': 0.00042440072713016317, 'samples': 2727360, 'steps': 14204, 'loss/train': 1.1678602695465088} 01/29/2022 07:05:29 - INFO - codeparrot_training - Step 14205: {'lr': 0.00042438900330478, 'samples': 2727552, 'steps': 14205, 'loss/train': 1.4657484591007233} 01/29/2022 07:05:33 - INFO - codeparrot_training - Step 14206: {'lr': 0.00042437727873236974, 'samples': 2727744, 'steps': 14206, 'loss/train': 1.5799114108085632} 01/29/2022 07:05:37 - INFO - codeparrot_training - Step 14207: {'lr': 0.00042436555341298266, 'samples': 2727936, 'steps': 14207, 'loss/train': 1.685329556465149} 01/29/2022 07:05:43 - INFO - codeparrot_training - Step 14208: {'lr': 0.0004243538273466689, 'samples': 2728128, 'steps': 14208, 'loss/train': 1.6729372143745422} 01/29/2022 07:05:47 - INFO - codeparrot_training - Step 14209: {'lr': 0.00042434210053347865, 'samples': 2728320, 'steps': 14209, 'loss/train': 0.8094899654388428} 01/29/2022 07:05:51 - INFO - codeparrot_training - Step 14210: {'lr': 0.0004243303729734622, 'samples': 2728512, 'steps': 14210, 'loss/train': 0.6064700782299042} 01/29/2022 07:05:55 - INFO - codeparrot_training - Step 14211: {'lr': 0.0004243186446666699, 'samples': 2728704, 'steps': 14211, 'loss/train': 2.157345950603485} 01/29/2022 07:06:00 - INFO - codeparrot_training - Step 14212: {'lr': 0.00042430691561315176, 'samples': 2728896, 'steps': 14212, 'loss/train': 1.289964884519577} 01/29/2022 07:06:06 - INFO - codeparrot_training - Step 14213: {'lr': 0.0004242951858129582, 'samples': 2729088, 'steps': 14213, 'loss/train': 1.8448898792266846} 01/29/2022 07:06:10 - INFO - codeparrot_training - Step 14214: {'lr': 0.0004242834552661394, 'samples': 2729280, 'steps': 14214, 'loss/train': 1.3235544562339783} 01/29/2022 07:06:15 - INFO - codeparrot_training - Step 14215: {'lr': 0.0004242717239727456, 'samples': 2729472, 'steps': 14215, 'loss/train': 1.8447561264038086} 01/29/2022 07:06:19 - INFO - codeparrot_training - Step 14216: {'lr': 0.00042425999193282713, 'samples': 2729664, 'steps': 14216, 'loss/train': 0.5155043005943298} 01/29/2022 07:06:23 - INFO - codeparrot_training - Step 14217: {'lr': 0.0004242482591464342, 'samples': 2729856, 'steps': 14217, 'loss/train': 2.8352563977241516} 01/29/2022 07:06:28 - INFO - codeparrot_training - Step 14218: {'lr': 0.0004242365256136169, 'samples': 2730048, 'steps': 14218, 'loss/train': 2.0526461005210876} 01/29/2022 07:06:33 - INFO - codeparrot_training - Step 14219: {'lr': 0.00042422479133442573, 'samples': 2730240, 'steps': 14219, 'loss/train': 1.8443450331687927} 01/29/2022 07:06:37 - INFO - codeparrot_training - Step 14220: {'lr': 0.00042421305630891093, 'samples': 2730432, 'steps': 14220, 'loss/train': 2.466835856437683} 01/29/2022 07:06:41 - INFO - codeparrot_training - Step 14221: {'lr': 0.0004242013205371227, 'samples': 2730624, 'steps': 14221, 'loss/train': 1.9229466319084167} 01/29/2022 07:06:45 - INFO - codeparrot_training - Step 14222: {'lr': 0.00042418958401911134, 'samples': 2730816, 'steps': 14222, 'loss/train': 1.5450972318649292} 01/29/2022 07:06:51 - INFO - codeparrot_training - Step 14223: {'lr': 0.000424177846754927, 'samples': 2731008, 'steps': 14223, 'loss/train': 1.8939642906188965} 01/29/2022 07:06:55 - INFO - codeparrot_training - Step 14224: {'lr': 0.0004241661087446202, 'samples': 2731200, 'steps': 14224, 'loss/train': 1.6586183309555054} 01/29/2022 07:06:59 - INFO - codeparrot_training - Step 14225: {'lr': 0.00042415436998824105, 'samples': 2731392, 'steps': 14225, 'loss/train': 1.5962843298912048} 01/29/2022 07:07:04 - INFO - codeparrot_training - Step 14226: {'lr': 0.0004241426304858399, 'samples': 2731584, 'steps': 14226, 'loss/train': 1.644051432609558} 01/29/2022 07:07:08 - INFO - codeparrot_training - Step 14227: {'lr': 0.00042413089023746696, 'samples': 2731776, 'steps': 14227, 'loss/train': 1.574592411518097} 01/29/2022 07:07:14 - INFO - codeparrot_training - Step 14228: {'lr': 0.00042411914924317265, 'samples': 2731968, 'steps': 14228, 'loss/train': 0.772014856338501} 01/29/2022 07:07:18 - INFO - codeparrot_training - Step 14229: {'lr': 0.00042410740750300715, 'samples': 2732160, 'steps': 14229, 'loss/train': 1.5859651565551758} 01/29/2022 07:07:22 - INFO - codeparrot_training - Step 14230: {'lr': 0.0004240956650170208, 'samples': 2732352, 'steps': 14230, 'loss/train': 1.6008709073066711} 01/29/2022 07:07:26 - INFO - codeparrot_training - Step 14231: {'lr': 0.00042408392178526396, 'samples': 2732544, 'steps': 14231, 'loss/train': 1.1752960681915283} 01/29/2022 07:07:31 - INFO - codeparrot_training - Step 14232: {'lr': 0.0004240721778077868, 'samples': 2732736, 'steps': 14232, 'loss/train': 1.6792986989021301} 01/29/2022 07:07:36 - INFO - codeparrot_training - Step 14233: {'lr': 0.0004240604330846397, 'samples': 2732928, 'steps': 14233, 'loss/train': 3.18259334564209} 01/29/2022 07:07:41 - INFO - codeparrot_training - Step 14234: {'lr': 0.000424048687615873, 'samples': 2733120, 'steps': 14234, 'loss/train': 2.008500337600708} 01/29/2022 07:07:45 - INFO - codeparrot_training - Step 14235: {'lr': 0.00042403694140153705, 'samples': 2733312, 'steps': 14235, 'loss/train': 2.334032356739044} 01/29/2022 07:07:49 - INFO - codeparrot_training - Step 14236: {'lr': 0.00042402519444168207, 'samples': 2733504, 'steps': 14236, 'loss/train': 1.8310582637786865} 01/29/2022 07:07:53 - INFO - codeparrot_training - Step 14237: {'lr': 0.00042401344673635846, 'samples': 2733696, 'steps': 14237, 'loss/train': 1.654456079006195} 01/29/2022 07:07:59 - INFO - codeparrot_training - Step 14238: {'lr': 0.00042400169828561636, 'samples': 2733888, 'steps': 14238, 'loss/train': 3.096288800239563} 01/29/2022 07:08:03 - INFO - codeparrot_training - Step 14239: {'lr': 0.0004239899490895063, 'samples': 2734080, 'steps': 14239, 'loss/train': 1.7988072037696838} 01/29/2022 07:08:07 - INFO - codeparrot_training - Step 14240: {'lr': 0.00042397819914807855, 'samples': 2734272, 'steps': 14240, 'loss/train': 2.2546529173851013} 01/29/2022 07:08:12 - INFO - codeparrot_training - Step 14241: {'lr': 0.00042396644846138355, 'samples': 2734464, 'steps': 14241, 'loss/train': 1.8067976832389832} 01/29/2022 07:08:16 - INFO - codeparrot_training - Step 14242: {'lr': 0.00042395469702947135, 'samples': 2734656, 'steps': 14242, 'loss/train': 2.4843040108680725} 01/29/2022 07:08:22 - INFO - codeparrot_training - Step 14243: {'lr': 0.0004239429448523925, 'samples': 2734848, 'steps': 14243, 'loss/train': 1.2728606164455414} 01/29/2022 07:08:26 - INFO - codeparrot_training - Step 14244: {'lr': 0.00042393119193019743, 'samples': 2735040, 'steps': 14244, 'loss/train': 2.00191068649292} 01/29/2022 07:08:30 - INFO - codeparrot_training - Step 14245: {'lr': 0.00042391943826293623, 'samples': 2735232, 'steps': 14245, 'loss/train': 1.4110826253890991} 01/29/2022 07:08:34 - INFO - codeparrot_training - Step 14246: {'lr': 0.0004239076838506595, 'samples': 2735424, 'steps': 14246, 'loss/train': 1.0712120532989502} 01/29/2022 07:08:39 - INFO - codeparrot_training - Step 14247: {'lr': 0.0004238959286934174, 'samples': 2735616, 'steps': 14247, 'loss/train': 0.3718765676021576} 01/29/2022 07:08:44 - INFO - codeparrot_training - Step 14248: {'lr': 0.0004238841727912603, 'samples': 2735808, 'steps': 14248, 'loss/train': 1.49701988697052} 01/29/2022 07:08:48 - INFO - codeparrot_training - Step 14249: {'lr': 0.00042387241614423875, 'samples': 2736000, 'steps': 14249, 'loss/train': 1.3157945573329926} 01/29/2022 07:08:52 - INFO - codeparrot_training - Step 14250: {'lr': 0.0004238606587524029, 'samples': 2736192, 'steps': 14250, 'loss/train': 1.8084089756011963} 01/29/2022 07:08:57 - INFO - codeparrot_training - Step 14251: {'lr': 0.0004238489006158033, 'samples': 2736384, 'steps': 14251, 'loss/train': 1.2503765523433685} 01/29/2022 07:09:01 - INFO - codeparrot_training - Step 14252: {'lr': 0.00042383714173449007, 'samples': 2736576, 'steps': 14252, 'loss/train': 2.001819670200348} 01/29/2022 07:09:06 - INFO - codeparrot_training - Step 14253: {'lr': 0.0004238253821085138, 'samples': 2736768, 'steps': 14253, 'loss/train': 1.5434104800224304} 01/29/2022 07:09:10 - INFO - codeparrot_training - Step 14254: {'lr': 0.00042381362173792475, 'samples': 2736960, 'steps': 14254, 'loss/train': 2.167966604232788} 01/29/2022 07:09:15 - INFO - codeparrot_training - Step 14255: {'lr': 0.00042380186062277337, 'samples': 2737152, 'steps': 14255, 'loss/train': 1.1182852685451508} 01/29/2022 07:09:19 - INFO - codeparrot_training - Step 14256: {'lr': 0.00042379009876311, 'samples': 2737344, 'steps': 14256, 'loss/train': 2.1626939177513123} 01/29/2022 07:09:23 - INFO - codeparrot_training - Step 14257: {'lr': 0.00042377833615898496, 'samples': 2737536, 'steps': 14257, 'loss/train': 1.7642937898635864} 01/29/2022 07:09:29 - INFO - codeparrot_training - Step 14258: {'lr': 0.0004237665728104488, 'samples': 2737728, 'steps': 14258, 'loss/train': 1.7216999530792236} 01/29/2022 07:09:33 - INFO - codeparrot_training - Step 14259: {'lr': 0.0004237548087175518, 'samples': 2737920, 'steps': 14259, 'loss/train': 2.2383170127868652} 01/29/2022 07:09:38 - INFO - codeparrot_training - Step 14260: {'lr': 0.00042374304388034437, 'samples': 2738112, 'steps': 14260, 'loss/train': 1.5217559337615967} 01/29/2022 07:09:42 - INFO - codeparrot_training - Step 14261: {'lr': 0.00042373127829887694, 'samples': 2738304, 'steps': 14261, 'loss/train': 2.088662803173065} 01/29/2022 07:09:46 - INFO - codeparrot_training - Step 14262: {'lr': 0.0004237195119731998, 'samples': 2738496, 'steps': 14262, 'loss/train': 1.2743845582008362} 01/29/2022 07:09:52 - INFO - codeparrot_training - Step 14263: {'lr': 0.0004237077449033635, 'samples': 2738688, 'steps': 14263, 'loss/train': 1.7369851469993591} 01/29/2022 07:09:56 - INFO - codeparrot_training - Step 14264: {'lr': 0.0004236959770894183, 'samples': 2738880, 'steps': 14264, 'loss/train': 1.9072734117507935} 01/29/2022 07:10:00 - INFO - codeparrot_training - Step 14265: {'lr': 0.0004236842085314148, 'samples': 2739072, 'steps': 14265, 'loss/train': 2.459077477455139} 01/29/2022 07:10:04 - INFO - codeparrot_training - Step 14266: {'lr': 0.0004236724392294032, 'samples': 2739264, 'steps': 14266, 'loss/train': 1.6937236189842224} 01/29/2022 07:10:09 - INFO - codeparrot_training - Step 14267: {'lr': 0.0004236606691834341, 'samples': 2739456, 'steps': 14267, 'loss/train': 1.8029612302780151} 01/29/2022 07:10:15 - INFO - codeparrot_training - Step 14268: {'lr': 0.0004236488983935578, 'samples': 2739648, 'steps': 14268, 'loss/train': 0.7015127688646317} 01/29/2022 07:10:20 - INFO - codeparrot_training - Step 14269: {'lr': 0.0004236371268598248, 'samples': 2739840, 'steps': 14269, 'loss/train': 1.7181814312934875} 01/29/2022 07:10:24 - INFO - codeparrot_training - Step 14270: {'lr': 0.0004236253545822855, 'samples': 2740032, 'steps': 14270, 'loss/train': 1.777155876159668} 01/29/2022 07:10:28 - INFO - codeparrot_training - Step 14271: {'lr': 0.00042361358156099016, 'samples': 2740224, 'steps': 14271, 'loss/train': 1.6533563733100891} 01/29/2022 07:10:32 - INFO - codeparrot_training - Step 14272: {'lr': 0.0004236018077959895, 'samples': 2740416, 'steps': 14272, 'loss/train': 1.9363861083984375} 01/29/2022 07:10:37 - INFO - codeparrot_training - Step 14273: {'lr': 0.00042359003328733383, 'samples': 2740608, 'steps': 14273, 'loss/train': 0.8921083509922028} 01/29/2022 07:10:42 - INFO - codeparrot_training - Step 14274: {'lr': 0.0004235782580350734, 'samples': 2740800, 'steps': 14274, 'loss/train': 0.8943741917610168} 01/29/2022 07:10:47 - INFO - codeparrot_training - Step 14275: {'lr': 0.0004235664820392591, 'samples': 2740992, 'steps': 14275, 'loss/train': 1.7950674891471863} 01/29/2022 07:10:52 - INFO - codeparrot_training - Step 14276: {'lr': 0.0004235547052999409, 'samples': 2741184, 'steps': 14276, 'loss/train': 1.9203336238861084} 01/29/2022 07:10:56 - INFO - codeparrot_training - Step 14277: {'lr': 0.0004235429278171695, 'samples': 2741376, 'steps': 14277, 'loss/train': 1.9601395726203918} 01/29/2022 07:11:00 - INFO - codeparrot_training - Step 14278: {'lr': 0.00042353114959099535, 'samples': 2741568, 'steps': 14278, 'loss/train': 1.5396963357925415} 01/29/2022 07:11:05 - INFO - codeparrot_training - Step 14279: {'lr': 0.0004235193706214688, 'samples': 2741760, 'steps': 14279, 'loss/train': 1.1706012189388275} 01/29/2022 07:11:10 - INFO - codeparrot_training - Step 14280: {'lr': 0.00042350759090864043, 'samples': 2741952, 'steps': 14280, 'loss/train': 1.8276786804199219} 01/29/2022 07:11:14 - INFO - codeparrot_training - Step 14281: {'lr': 0.00042349581045256055, 'samples': 2742144, 'steps': 14281, 'loss/train': 1.5572900176048279} 01/29/2022 07:11:18 - INFO - codeparrot_training - Step 14282: {'lr': 0.00042348402925327977, 'samples': 2742336, 'steps': 14282, 'loss/train': 2.1048197150230408} 01/29/2022 07:11:22 - INFO - codeparrot_training - Step 14283: {'lr': 0.00042347224731084854, 'samples': 2742528, 'steps': 14283, 'loss/train': 1.6870547533035278} 01/29/2022 07:11:28 - INFO - codeparrot_training - Step 14284: {'lr': 0.0004234604646253172, 'samples': 2742720, 'steps': 14284, 'loss/train': 1.7567662596702576} 01/29/2022 07:11:32 - INFO - codeparrot_training - Step 14285: {'lr': 0.0004234486811967364, 'samples': 2742912, 'steps': 14285, 'loss/train': 1.4209433197975159} 01/29/2022 07:11:36 - INFO - codeparrot_training - Step 14286: {'lr': 0.00042343689702515643, 'samples': 2743104, 'steps': 14286, 'loss/train': 1.4674612283706665} 01/29/2022 07:11:41 - INFO - codeparrot_training - Step 14287: {'lr': 0.0004234251121106279, 'samples': 2743296, 'steps': 14287, 'loss/train': 1.5796127915382385} 01/29/2022 07:11:45 - INFO - codeparrot_training - Step 14288: {'lr': 0.00042341332645320126, 'samples': 2743488, 'steps': 14288, 'loss/train': 1.38860884308815} 01/29/2022 07:11:51 - INFO - codeparrot_training - Step 14289: {'lr': 0.000423401540052927, 'samples': 2743680, 'steps': 14289, 'loss/train': 1.8064441680908203} 01/29/2022 07:11:55 - INFO - codeparrot_training - Step 14290: {'lr': 0.0004233897529098556, 'samples': 2743872, 'steps': 14290, 'loss/train': 3.018296241760254} 01/29/2022 07:11:59 - INFO - codeparrot_training - Step 14291: {'lr': 0.0004233779650240376, 'samples': 2744064, 'steps': 14291, 'loss/train': 2.0475398898124695} 01/29/2022 07:12:03 - INFO - codeparrot_training - Step 14292: {'lr': 0.00042336617639552335, 'samples': 2744256, 'steps': 14292, 'loss/train': 2.173059403896332} 01/29/2022 07:12:08 - INFO - codeparrot_training - Step 14293: {'lr': 0.00042335438702436354, 'samples': 2744448, 'steps': 14293, 'loss/train': 1.6503832340240479} 01/29/2022 07:12:13 - INFO - codeparrot_training - Step 14294: {'lr': 0.0004233425969106085, 'samples': 2744640, 'steps': 14294, 'loss/train': 1.8559802770614624} 01/29/2022 07:12:17 - INFO - codeparrot_training - Step 14295: {'lr': 0.00042333080605430883, 'samples': 2744832, 'steps': 14295, 'loss/train': 2.2452281713485718} 01/29/2022 07:12:21 - INFO - codeparrot_training - Step 14296: {'lr': 0.00042331901445551514, 'samples': 2745024, 'steps': 14296, 'loss/train': 1.3240739107131958} 01/29/2022 07:12:26 - INFO - codeparrot_training - Step 14297: {'lr': 0.00042330722211427775, 'samples': 2745216, 'steps': 14297, 'loss/train': 1.9770047664642334} 01/29/2022 07:12:30 - INFO - codeparrot_training - Step 14298: {'lr': 0.00042329542903064724, 'samples': 2745408, 'steps': 14298, 'loss/train': 1.0813098549842834} 01/29/2022 07:12:35 - INFO - codeparrot_training - Step 14299: {'lr': 0.00042328363520467417, 'samples': 2745600, 'steps': 14299, 'loss/train': 1.46525439620018} 01/29/2022 07:12:40 - INFO - codeparrot_training - Step 14300: {'lr': 0.000423271840636409, 'samples': 2745792, 'steps': 14300, 'loss/train': 1.6018951535224915} 01/29/2022 07:12:44 - INFO - codeparrot_training - Step 14301: {'lr': 0.0004232600453259023, 'samples': 2745984, 'steps': 14301, 'loss/train': 1.4147399961948395} 01/29/2022 07:12:48 - INFO - codeparrot_training - Step 14302: {'lr': 0.0004232482492732046, 'samples': 2746176, 'steps': 14302, 'loss/train': 0.3804209679365158} 01/29/2022 07:12:52 - INFO - codeparrot_training - Step 14303: {'lr': 0.00042323645247836636, 'samples': 2746368, 'steps': 14303, 'loss/train': 1.8565176129341125} 01/29/2022 07:12:59 - INFO - codeparrot_training - Step 14304: {'lr': 0.00042322465494143814, 'samples': 2746560, 'steps': 14304, 'loss/train': 2.3914827704429626} 01/29/2022 07:13:03 - INFO - codeparrot_training - Step 14305: {'lr': 0.00042321285666247063, 'samples': 2746752, 'steps': 14305, 'loss/train': 1.4432352483272552} 01/29/2022 07:13:07 - INFO - codeparrot_training - Step 14306: {'lr': 0.0004232010576415141, 'samples': 2746944, 'steps': 14306, 'loss/train': 1.9119515419006348} 01/29/2022 07:13:11 - INFO - codeparrot_training - Step 14307: {'lr': 0.00042318925787861937, 'samples': 2747136, 'steps': 14307, 'loss/train': 1.2352802753448486} 01/29/2022 07:13:17 - INFO - codeparrot_training - Step 14308: {'lr': 0.0004231774573738367, 'samples': 2747328, 'steps': 14308, 'loss/train': 1.4658061265945435} 01/29/2022 07:13:21 - INFO - codeparrot_training - Step 14309: {'lr': 0.000423165656127217, 'samples': 2747520, 'steps': 14309, 'loss/train': 2.097342073917389} 01/29/2022 07:13:25 - INFO - codeparrot_training - Step 14310: {'lr': 0.00042315385413881047, 'samples': 2747712, 'steps': 14310, 'loss/train': 1.1189163029193878} 01/29/2022 07:13:29 - INFO - codeparrot_training - Step 14311: {'lr': 0.00042314205140866785, 'samples': 2747904, 'steps': 14311, 'loss/train': 1.2000610828399658} 01/29/2022 07:13:34 - INFO - codeparrot_training - Step 14312: {'lr': 0.00042313024793683965, 'samples': 2748096, 'steps': 14312, 'loss/train': 2.08345627784729} 01/29/2022 07:13:39 - INFO - codeparrot_training - Step 14313: {'lr': 0.0004231184437233765, 'samples': 2748288, 'steps': 14313, 'loss/train': 1.4659142196178436} 01/29/2022 07:13:43 - INFO - codeparrot_training - Step 14314: {'lr': 0.0004231066387683288, 'samples': 2748480, 'steps': 14314, 'loss/train': 1.4320198595523834} 01/29/2022 07:13:47 - INFO - codeparrot_training - Step 14315: {'lr': 0.0004230948330717472, 'samples': 2748672, 'steps': 14315, 'loss/train': 1.2393445372581482} 01/29/2022 07:13:51 - INFO - codeparrot_training - Step 14316: {'lr': 0.0004230830266336825, 'samples': 2748864, 'steps': 14316, 'loss/train': 1.004084587097168} 01/29/2022 07:13:56 - INFO - codeparrot_training - Step 14317: {'lr': 0.00042307121945418493, 'samples': 2749056, 'steps': 14317, 'loss/train': 2.81844699382782} 01/29/2022 07:14:02 - INFO - codeparrot_training - Step 14318: {'lr': 0.00042305941153330525, 'samples': 2749248, 'steps': 14318, 'loss/train': 0.892785519361496} 01/29/2022 07:14:06 - INFO - codeparrot_training - Step 14319: {'lr': 0.00042304760287109394, 'samples': 2749440, 'steps': 14319, 'loss/train': 1.3810853362083435} 01/29/2022 07:14:10 - INFO - codeparrot_training - Step 14320: {'lr': 0.0004230357934676017, 'samples': 2749632, 'steps': 14320, 'loss/train': 1.7995662689208984} 01/29/2022 07:14:14 - INFO - codeparrot_training - Step 14321: {'lr': 0.00042302398332287903, 'samples': 2749824, 'steps': 14321, 'loss/train': 1.0258069932460785} 01/29/2022 07:14:19 - INFO - codeparrot_training - Step 14322: {'lr': 0.00042301217243697665, 'samples': 2750016, 'steps': 14322, 'loss/train': 1.4613102972507477} 01/29/2022 07:14:24 - INFO - codeparrot_training - Step 14323: {'lr': 0.00042300036080994495, 'samples': 2750208, 'steps': 14323, 'loss/train': 1.720184326171875} 01/29/2022 07:14:28 - INFO - codeparrot_training - Step 14324: {'lr': 0.00042298854844183476, 'samples': 2750400, 'steps': 14324, 'loss/train': 1.8412511944770813} 01/29/2022 07:14:32 - INFO - codeparrot_training - Step 14325: {'lr': 0.0004229767353326964, 'samples': 2750592, 'steps': 14325, 'loss/train': 1.7835599184036255} 01/29/2022 07:14:37 - INFO - codeparrot_training - Step 14326: {'lr': 0.0004229649214825808, 'samples': 2750784, 'steps': 14326, 'loss/train': 1.8215657472610474} 01/29/2022 07:14:41 - INFO - codeparrot_training - Step 14327: {'lr': 0.0004229531068915383, 'samples': 2750976, 'steps': 14327, 'loss/train': 2.052380919456482} 01/29/2022 07:14:47 - INFO - codeparrot_training - Step 14328: {'lr': 0.0004229412915596196, 'samples': 2751168, 'steps': 14328, 'loss/train': 1.57797771692276} 01/29/2022 07:14:51 - INFO - codeparrot_training - Step 14329: {'lr': 0.0004229294754868754, 'samples': 2751360, 'steps': 14329, 'loss/train': 2.5059694647789} 01/29/2022 07:14:55 - INFO - codeparrot_training - Step 14330: {'lr': 0.0004229176586733562, 'samples': 2751552, 'steps': 14330, 'loss/train': 1.361911118030548} 01/29/2022 07:15:00 - INFO - codeparrot_training - Step 14331: {'lr': 0.0004229058411191126, 'samples': 2751744, 'steps': 14331, 'loss/train': 1.313982993364334} 01/29/2022 07:15:04 - INFO - codeparrot_training - Step 14332: {'lr': 0.0004228940228241953, 'samples': 2751936, 'steps': 14332, 'loss/train': 2.1710879802703857} 01/29/2022 07:15:09 - INFO - codeparrot_training - Step 14333: {'lr': 0.0004228822037886549, 'samples': 2752128, 'steps': 14333, 'loss/train': 1.9900705218315125} 01/29/2022 07:15:14 - INFO - codeparrot_training - Step 14334: {'lr': 0.00042287038401254214, 'samples': 2752320, 'steps': 14334, 'loss/train': 2.3977949023246765} 01/29/2022 07:15:18 - INFO - codeparrot_training - Step 14335: {'lr': 0.00042285856349590746, 'samples': 2752512, 'steps': 14335, 'loss/train': 0.24034836888313293} 01/29/2022 07:15:22 - INFO - codeparrot_training - Step 14336: {'lr': 0.0004228467422388016, 'samples': 2752704, 'steps': 14336, 'loss/train': 1.4434238076210022} 01/29/2022 07:15:26 - INFO - codeparrot_training - Step 14337: {'lr': 0.00042283492024127524, 'samples': 2752896, 'steps': 14337, 'loss/train': 1.1958298087120056} 01/29/2022 07:15:32 - INFO - codeparrot_training - Step 14338: {'lr': 0.00042282309750337887, 'samples': 2753088, 'steps': 14338, 'loss/train': 0.9441312253475189} 01/29/2022 07:15:36 - INFO - codeparrot_training - Step 14339: {'lr': 0.0004228112740251632, 'samples': 2753280, 'steps': 14339, 'loss/train': 0.9966102540493011} 01/29/2022 07:15:40 - INFO - codeparrot_training - Step 14340: {'lr': 0.00042279944980667906, 'samples': 2753472, 'steps': 14340, 'loss/train': 1.3779836297035217} 01/29/2022 07:15:44 - INFO - codeparrot_training - Step 14341: {'lr': 0.00042278762484797684, 'samples': 2753664, 'steps': 14341, 'loss/train': 2.163648247718811} 01/29/2022 07:15:49 - INFO - codeparrot_training - Step 14342: {'lr': 0.0004227757991491073, 'samples': 2753856, 'steps': 14342, 'loss/train': 1.8241453170776367} 01/29/2022 07:15:54 - INFO - codeparrot_training - Step 14343: {'lr': 0.0004227639727101211, 'samples': 2754048, 'steps': 14343, 'loss/train': 1.3577722907066345} 01/29/2022 07:15:58 - INFO - codeparrot_training - Step 14344: {'lr': 0.0004227521455310689, 'samples': 2754240, 'steps': 14344, 'loss/train': 1.22686967253685} 01/29/2022 07:16:03 - INFO - codeparrot_training - Step 14345: {'lr': 0.0004227403176120014, 'samples': 2754432, 'steps': 14345, 'loss/train': 1.4091423153877258} 01/29/2022 07:16:07 - INFO - codeparrot_training - Step 14346: {'lr': 0.00042272848895296924, 'samples': 2754624, 'steps': 14346, 'loss/train': 2.0275885462760925} 01/29/2022 07:16:11 - INFO - codeparrot_training - Step 14347: {'lr': 0.000422716659554023, 'samples': 2754816, 'steps': 14347, 'loss/train': 2.33362877368927} 01/29/2022 07:16:17 - INFO - codeparrot_training - Step 14348: {'lr': 0.00042270482941521347, 'samples': 2755008, 'steps': 14348, 'loss/train': 1.4905627369880676} 01/29/2022 07:16:21 - INFO - codeparrot_training - Step 14349: {'lr': 0.0004226929985365913, 'samples': 2755200, 'steps': 14349, 'loss/train': 2.0265550017356873} 01/29/2022 07:16:26 - INFO - codeparrot_training - Step 14350: {'lr': 0.00042268116691820723, 'samples': 2755392, 'steps': 14350, 'loss/train': 1.4970893561840057} 01/29/2022 07:16:30 - INFO - codeparrot_training - Step 14351: {'lr': 0.00042266933456011174, 'samples': 2755584, 'steps': 14351, 'loss/train': 2.0151625871658325} 01/29/2022 07:16:34 - INFO - codeparrot_training - Step 14352: {'lr': 0.0004226575014623557, 'samples': 2755776, 'steps': 14352, 'loss/train': 1.7040050625801086} 01/29/2022 07:16:40 - INFO - codeparrot_training - Step 14353: {'lr': 0.0004226456676249898, 'samples': 2755968, 'steps': 14353, 'loss/train': 2.217286169528961} 01/29/2022 07:16:44 - INFO - codeparrot_training - Step 14354: {'lr': 0.0004226338330480646, 'samples': 2756160, 'steps': 14354, 'loss/train': 2.083034098148346} 01/29/2022 07:16:48 - INFO - codeparrot_training - Step 14355: {'lr': 0.00042262199773163096, 'samples': 2756352, 'steps': 14355, 'loss/train': 1.9397369027137756} 01/29/2022 07:16:52 - INFO - codeparrot_training - Step 14356: {'lr': 0.00042261016167573944, 'samples': 2756544, 'steps': 14356, 'loss/train': 1.9854878783226013} 01/29/2022 07:16:57 - INFO - codeparrot_training - Step 14357: {'lr': 0.0004225983248804408, 'samples': 2756736, 'steps': 14357, 'loss/train': 1.7676769495010376} 01/29/2022 07:17:02 - INFO - codeparrot_training - Step 14358: {'lr': 0.0004225864873457858, 'samples': 2756928, 'steps': 14358, 'loss/train': 1.5571426749229431} 01/29/2022 07:17:06 - INFO - codeparrot_training - Step 14359: {'lr': 0.0004225746490718251, 'samples': 2757120, 'steps': 14359, 'loss/train': 1.423046350479126} 01/29/2022 07:17:10 - INFO - codeparrot_training - Step 14360: {'lr': 0.0004225628100586093, 'samples': 2757312, 'steps': 14360, 'loss/train': 0.801643967628479} 01/29/2022 07:17:15 - INFO - codeparrot_training - Step 14361: {'lr': 0.0004225509703061893, 'samples': 2757504, 'steps': 14361, 'loss/train': 1.2467625439167023} 01/29/2022 07:17:19 - INFO - codeparrot_training - Step 14362: {'lr': 0.0004225391298146157, 'samples': 2757696, 'steps': 14362, 'loss/train': 2.1683555245399475} 01/29/2022 07:17:25 - INFO - codeparrot_training - Step 14363: {'lr': 0.0004225272885839392, 'samples': 2757888, 'steps': 14363, 'loss/train': 1.453921526670456} 01/29/2022 07:17:29 - INFO - codeparrot_training - Step 14364: {'lr': 0.0004225154466142107, 'samples': 2758080, 'steps': 14364, 'loss/train': 1.6437739133834839} 01/29/2022 07:17:33 - INFO - codeparrot_training - Step 14365: {'lr': 0.0004225036039054807, 'samples': 2758272, 'steps': 14365, 'loss/train': 2.3820444345474243} 01/29/2022 07:17:38 - INFO - codeparrot_training - Step 14366: {'lr': 0.00042249176045780013, 'samples': 2758464, 'steps': 14366, 'loss/train': 1.7551589012145996} 01/29/2022 07:17:42 - INFO - codeparrot_training - Step 14367: {'lr': 0.0004224799162712195, 'samples': 2758656, 'steps': 14367, 'loss/train': 1.7280131578445435} 01/29/2022 07:17:47 - INFO - codeparrot_training - Step 14368: {'lr': 0.0004224680713457898, 'samples': 2758848, 'steps': 14368, 'loss/train': 1.4786121547222137} 01/29/2022 07:17:51 - INFO - codeparrot_training - Step 14369: {'lr': 0.00042245622568156164, 'samples': 2759040, 'steps': 14369, 'loss/train': 1.0129806697368622} 01/29/2022 07:17:56 - INFO - codeparrot_training - Step 14370: {'lr': 0.0004224443792785857, 'samples': 2759232, 'steps': 14370, 'loss/train': 1.7741641402244568} 01/29/2022 07:18:00 - INFO - codeparrot_training - Step 14371: {'lr': 0.0004224325321369128, 'samples': 2759424, 'steps': 14371, 'loss/train': 2.1572166681289673} 01/29/2022 07:18:04 - INFO - codeparrot_training - Step 14372: {'lr': 0.0004224206842565937, 'samples': 2759616, 'steps': 14372, 'loss/train': 1.8513574004173279} 01/29/2022 07:18:10 - INFO - codeparrot_training - Step 14373: {'lr': 0.00042240883563767916, 'samples': 2759808, 'steps': 14373, 'loss/train': 1.6986742615699768} 01/29/2022 07:18:14 - INFO - codeparrot_training - Step 14374: {'lr': 0.00042239698628021994, 'samples': 2760000, 'steps': 14374, 'loss/train': 2.409805476665497} 01/29/2022 07:18:19 - INFO - codeparrot_training - Step 14375: {'lr': 0.0004223851361842668, 'samples': 2760192, 'steps': 14375, 'loss/train': 1.435655415058136} 01/29/2022 07:18:23 - INFO - codeparrot_training - Step 14376: {'lr': 0.00042237328534987034, 'samples': 2760384, 'steps': 14376, 'loss/train': 2.2102476954460144} 01/29/2022 07:18:27 - INFO - codeparrot_training - Step 14377: {'lr': 0.0004223614337770816, 'samples': 2760576, 'steps': 14377, 'loss/train': 1.8306180238723755} 01/29/2022 07:18:33 - INFO - codeparrot_training - Step 14378: {'lr': 0.0004223495814659511, 'samples': 2760768, 'steps': 14378, 'loss/train': 1.4979347884654999} 01/29/2022 07:18:37 - INFO - codeparrot_training - Step 14379: {'lr': 0.00042233772841652974, 'samples': 2760960, 'steps': 14379, 'loss/train': 1.996829330921173} 01/29/2022 07:18:41 - INFO - codeparrot_training - Step 14380: {'lr': 0.00042232587462886833, 'samples': 2761152, 'steps': 14380, 'loss/train': 0.810723602771759} 01/29/2022 07:18:45 - INFO - codeparrot_training - Step 14381: {'lr': 0.0004223140201030176, 'samples': 2761344, 'steps': 14381, 'loss/train': 2.184474527835846} 01/29/2022 07:18:50 - INFO - codeparrot_training - Step 14382: {'lr': 0.0004223021648390283, 'samples': 2761536, 'steps': 14382, 'loss/train': 0.9697263836860657} 01/29/2022 07:18:55 - INFO - codeparrot_training - Step 14383: {'lr': 0.0004222903088369512, 'samples': 2761728, 'steps': 14383, 'loss/train': 1.000659316778183} 01/29/2022 07:18:59 - INFO - codeparrot_training - Step 14384: {'lr': 0.0004222784520968371, 'samples': 2761920, 'steps': 14384, 'loss/train': 1.1710796356201172} 01/29/2022 07:19:03 - INFO - codeparrot_training - Step 14385: {'lr': 0.000422266594618737, 'samples': 2762112, 'steps': 14385, 'loss/train': 2.322173595428467} 01/29/2022 07:19:08 - INFO - codeparrot_training - Step 14386: {'lr': 0.0004222547364027013, 'samples': 2762304, 'steps': 14386, 'loss/train': 1.790786325931549} 01/29/2022 07:19:12 - INFO - codeparrot_training - Step 14387: {'lr': 0.0004222428774487811, 'samples': 2762496, 'steps': 14387, 'loss/train': 1.9528437852859497} 01/29/2022 07:19:18 - INFO - codeparrot_training - Step 14388: {'lr': 0.00042223101775702704, 'samples': 2762688, 'steps': 14388, 'loss/train': 1.9958410263061523} 01/29/2022 07:19:22 - INFO - codeparrot_training - Step 14389: {'lr': 0.00042221915732749006, 'samples': 2762880, 'steps': 14389, 'loss/train': 1.8250300884246826} 01/29/2022 07:19:27 - INFO - codeparrot_training - Step 14390: {'lr': 0.0004222072961602209, 'samples': 2763072, 'steps': 14390, 'loss/train': 1.6691701412200928} 01/29/2022 07:19:31 - INFO - codeparrot_training - Step 14391: {'lr': 0.0004221954342552703, 'samples': 2763264, 'steps': 14391, 'loss/train': 2.116393268108368} 01/29/2022 07:19:35 - INFO - codeparrot_training - Step 14392: {'lr': 0.00042218357161268917, 'samples': 2763456, 'steps': 14392, 'loss/train': 2.348782002925873} 01/29/2022 07:19:41 - INFO - codeparrot_training - Step 14393: {'lr': 0.0004221717082325283, 'samples': 2763648, 'steps': 14393, 'loss/train': 2.1860745549201965} 01/29/2022 07:19:45 - INFO - codeparrot_training - Step 14394: {'lr': 0.00042215984411483854, 'samples': 2763840, 'steps': 14394, 'loss/train': 0.8961333632469177} 01/29/2022 07:19:49 - INFO - codeparrot_training - Step 14395: {'lr': 0.00042214797925967064, 'samples': 2764032, 'steps': 14395, 'loss/train': 1.5802448987960815} 01/29/2022 07:19:54 - INFO - codeparrot_training - Step 14396: {'lr': 0.00042213611366707547, 'samples': 2764224, 'steps': 14396, 'loss/train': 0.7040022164583206} 01/29/2022 07:19:58 - INFO - codeparrot_training - Step 14397: {'lr': 0.0004221242473371038, 'samples': 2764416, 'steps': 14397, 'loss/train': 1.7688531875610352} 01/29/2022 07:20:02 - INFO - codeparrot_training - Step 14398: {'lr': 0.00042211238026980657, 'samples': 2764608, 'steps': 14398, 'loss/train': 1.5764340162277222} 01/29/2022 07:20:08 - INFO - codeparrot_training - Step 14399: {'lr': 0.0004221005124652345, 'samples': 2764800, 'steps': 14399, 'loss/train': 1.1862452030181885} 01/29/2022 07:20:13 - INFO - codeparrot_training - Step 14400: {'lr': 0.0004220886439234385, 'samples': 2764992, 'steps': 14400, 'loss/train': 2.057471215724945} 01/29/2022 07:20:17 - INFO - codeparrot_training - Step 14401: {'lr': 0.0004220767746444694, 'samples': 2765184, 'steps': 14401, 'loss/train': 2.1638342142105103} 01/29/2022 07:20:21 - INFO - codeparrot_training - Step 14402: {'lr': 0.0004220649046283781, 'samples': 2765376, 'steps': 14402, 'loss/train': 1.8508697748184204} 01/29/2022 07:20:25 - INFO - codeparrot_training - Step 14403: {'lr': 0.00042205303387521533, 'samples': 2765568, 'steps': 14403, 'loss/train': 1.8794764280319214} 01/29/2022 07:20:31 - INFO - codeparrot_training - Step 14404: {'lr': 0.00042204116238503197, 'samples': 2765760, 'steps': 14404, 'loss/train': 1.471622496843338} 01/29/2022 07:20:35 - INFO - codeparrot_training - Step 14405: {'lr': 0.00042202929015787893, 'samples': 2765952, 'steps': 14405, 'loss/train': 1.708060383796692} 01/29/2022 07:20:39 - INFO - codeparrot_training - Step 14406: {'lr': 0.000422017417193807, 'samples': 2766144, 'steps': 14406, 'loss/train': 1.2927195131778717} 01/29/2022 07:20:44 - INFO - codeparrot_training - Step 14407: {'lr': 0.0004220055434928671, 'samples': 2766336, 'steps': 14407, 'loss/train': 2.202977478504181} 01/29/2022 07:20:49 - INFO - codeparrot_training - Step 14408: {'lr': 0.0004219936690551101, 'samples': 2766528, 'steps': 14408, 'loss/train': 0.08689874410629272} 01/29/2022 07:20:53 - INFO - codeparrot_training - Step 14409: {'lr': 0.0004219817938805869, 'samples': 2766720, 'steps': 14409, 'loss/train': 1.3344499468803406} 01/29/2022 07:20:57 - INFO - codeparrot_training - Step 14410: {'lr': 0.0004219699179693481, 'samples': 2766912, 'steps': 14410, 'loss/train': 1.9238837957382202} 01/29/2022 07:21:02 - INFO - codeparrot_training - Step 14411: {'lr': 0.000421958041321445, 'samples': 2767104, 'steps': 14411, 'loss/train': 2.1014091968536377} 01/29/2022 07:21:06 - INFO - codeparrot_training - Step 14412: {'lr': 0.0004219461639369281, 'samples': 2767296, 'steps': 14412, 'loss/train': 1.4405300617218018} 01/29/2022 07:21:10 - INFO - codeparrot_training - Step 14413: {'lr': 0.0004219342858158485, 'samples': 2767488, 'steps': 14413, 'loss/train': 1.5440706610679626} 01/29/2022 07:21:16 - INFO - codeparrot_training - Step 14414: {'lr': 0.000421922406958257, 'samples': 2767680, 'steps': 14414, 'loss/train': 1.997670829296112} 01/29/2022 07:21:20 - INFO - codeparrot_training - Step 14415: {'lr': 0.00042191052736420445, 'samples': 2767872, 'steps': 14415, 'loss/train': 0.9177044034004211} 01/29/2022 07:21:24 - INFO - codeparrot_training - Step 14416: {'lr': 0.0004218986470337419, 'samples': 2768064, 'steps': 14416, 'loss/train': 1.8185436129570007} 01/29/2022 07:21:29 - INFO - codeparrot_training - Step 14417: {'lr': 0.00042188676596692, 'samples': 2768256, 'steps': 14417, 'loss/train': 1.7179092764854431} 01/29/2022 07:21:33 - INFO - codeparrot_training - Step 14418: {'lr': 0.0004218748841637899, 'samples': 2768448, 'steps': 14418, 'loss/train': 1.621084749698639} 01/29/2022 07:21:39 - INFO - codeparrot_training - Step 14419: {'lr': 0.0004218630016244023, 'samples': 2768640, 'steps': 14419, 'loss/train': 1.9629555344581604} 01/29/2022 07:21:43 - INFO - codeparrot_training - Step 14420: {'lr': 0.0004218511183488082, 'samples': 2768832, 'steps': 14420, 'loss/train': 1.843888521194458} 01/29/2022 07:21:48 - INFO - codeparrot_training - Step 14421: {'lr': 0.0004218392343370584, 'samples': 2769024, 'steps': 14421, 'loss/train': 1.8765011429786682} 01/29/2022 07:21:52 - INFO - codeparrot_training - Step 14422: {'lr': 0.000421827349589204, 'samples': 2769216, 'steps': 14422, 'loss/train': 1.2889766693115234} 01/29/2022 07:21:56 - INFO - codeparrot_training - Step 14423: {'lr': 0.0004218154641052957, 'samples': 2769408, 'steps': 14423, 'loss/train': 1.9257003664970398} 01/29/2022 07:22:02 - INFO - codeparrot_training - Step 14424: {'lr': 0.0004218035778853846, 'samples': 2769600, 'steps': 14424, 'loss/train': 0.8826457858085632} 01/29/2022 07:22:06 - INFO - codeparrot_training - Step 14425: {'lr': 0.0004217916909295215, 'samples': 2769792, 'steps': 14425, 'loss/train': 2.4449055790901184} 01/29/2022 07:22:10 - INFO - codeparrot_training - Step 14426: {'lr': 0.00042177980323775734, 'samples': 2769984, 'steps': 14426, 'loss/train': 1.3763867318630219} 01/29/2022 07:22:14 - INFO - codeparrot_training - Step 14427: {'lr': 0.00042176791481014303, 'samples': 2770176, 'steps': 14427, 'loss/train': 0.9478782713413239} 01/29/2022 07:22:19 - INFO - codeparrot_training - Step 14428: {'lr': 0.0004217560256467295, 'samples': 2770368, 'steps': 14428, 'loss/train': 1.5429121255874634} 01/29/2022 07:22:24 - INFO - codeparrot_training - Step 14429: {'lr': 0.00042174413574756775, 'samples': 2770560, 'steps': 14429, 'loss/train': 2.401980936527252} 01/29/2022 07:22:28 - INFO - codeparrot_training - Step 14430: {'lr': 0.0004217322451127086, 'samples': 2770752, 'steps': 14430, 'loss/train': 2.223499059677124} 01/29/2022 07:22:32 - INFO - codeparrot_training - Step 14431: {'lr': 0.00042172035374220306, 'samples': 2770944, 'steps': 14431, 'loss/train': 2.471871256828308} 01/29/2022 07:22:37 - INFO - codeparrot_training - Step 14432: {'lr': 0.0004217084616361021, 'samples': 2771136, 'steps': 14432, 'loss/train': 1.1851838529109955} 01/29/2022 07:22:41 - INFO - codeparrot_training - Step 14433: {'lr': 0.00042169656879445657, 'samples': 2771328, 'steps': 14433, 'loss/train': 2.0381643176078796} 01/29/2022 07:22:47 - INFO - codeparrot_training - Step 14434: {'lr': 0.00042168467521731747, 'samples': 2771520, 'steps': 14434, 'loss/train': 1.5401846766471863} 01/29/2022 07:22:51 - INFO - codeparrot_training - Step 14435: {'lr': 0.00042167278090473573, 'samples': 2771712, 'steps': 14435, 'loss/train': 2.085393011569977} 01/29/2022 07:22:55 - INFO - codeparrot_training - Step 14436: {'lr': 0.0004216608858567623, 'samples': 2771904, 'steps': 14436, 'loss/train': 1.0268925726413727} 01/29/2022 07:23:00 - INFO - codeparrot_training - Step 14437: {'lr': 0.00042164899007344814, 'samples': 2772096, 'steps': 14437, 'loss/train': 1.3928945660591125} 01/29/2022 07:23:04 - INFO - codeparrot_training - Step 14438: {'lr': 0.00042163709355484425, 'samples': 2772288, 'steps': 14438, 'loss/train': 1.966795563697815} 01/29/2022 07:23:09 - INFO - codeparrot_training - Step 14439: {'lr': 0.0004216251963010015, 'samples': 2772480, 'steps': 14439, 'loss/train': 1.9099738597869873} 01/29/2022 07:23:13 - INFO - codeparrot_training - Step 14440: {'lr': 0.0004216132983119709, 'samples': 2772672, 'steps': 14440, 'loss/train': 1.7588909268379211} 01/29/2022 07:23:18 - INFO - codeparrot_training - Step 14441: {'lr': 0.00042160139958780346, 'samples': 2772864, 'steps': 14441, 'loss/train': 0.7383115589618683} 01/29/2022 07:23:22 - INFO - codeparrot_training - Step 14442: {'lr': 0.0004215895001285501, 'samples': 2773056, 'steps': 14442, 'loss/train': 1.7345121502876282} 01/29/2022 07:23:26 - INFO - codeparrot_training - Step 14443: {'lr': 0.0004215775999342618, 'samples': 2773248, 'steps': 14443, 'loss/train': 1.9027436971664429} 01/29/2022 07:23:32 - INFO - codeparrot_training - Step 14444: {'lr': 0.0004215656990049896, 'samples': 2773440, 'steps': 14444, 'loss/train': 1.725503146648407} 01/29/2022 07:23:36 - INFO - codeparrot_training - Step 14445: {'lr': 0.0004215537973407844, 'samples': 2773632, 'steps': 14445, 'loss/train': 0.5684448927640915} 01/29/2022 07:23:41 - INFO - codeparrot_training - Step 14446: {'lr': 0.0004215418949416972, 'samples': 2773824, 'steps': 14446, 'loss/train': 2.26283061504364} 01/29/2022 07:23:45 - INFO - codeparrot_training - Step 14447: {'lr': 0.00042152999180777894, 'samples': 2774016, 'steps': 14447, 'loss/train': 2.045747995376587} 01/29/2022 07:23:49 - INFO - codeparrot_training - Step 14448: {'lr': 0.0004215180879390807, 'samples': 2774208, 'steps': 14448, 'loss/train': 2.089721381664276} 01/29/2022 07:23:54 - INFO - codeparrot_training - Step 14449: {'lr': 0.0004215061833356535, 'samples': 2774400, 'steps': 14449, 'loss/train': 2.2905004620552063} 01/29/2022 07:23:59 - INFO - codeparrot_training - Step 14450: {'lr': 0.00042149427799754817, 'samples': 2774592, 'steps': 14450, 'loss/train': 3.005881905555725} 01/29/2022 07:24:03 - INFO - codeparrot_training - Step 14451: {'lr': 0.00042148237192481586, 'samples': 2774784, 'steps': 14451, 'loss/train': 3.12182879447937} 01/29/2022 07:24:07 - INFO - codeparrot_training - Step 14452: {'lr': 0.0004214704651175075, 'samples': 2774976, 'steps': 14452, 'loss/train': 1.8554871082305908} 01/29/2022 07:24:11 - INFO - codeparrot_training - Step 14453: {'lr': 0.0004214585575756742, 'samples': 2775168, 'steps': 14453, 'loss/train': 1.7997490167617798} 01/29/2022 07:24:17 - INFO - codeparrot_training - Step 14454: {'lr': 0.0004214466492993668, 'samples': 2775360, 'steps': 14454, 'loss/train': 1.8918582201004028} 01/29/2022 07:24:21 - INFO - codeparrot_training - Step 14455: {'lr': 0.00042143474028863637, 'samples': 2775552, 'steps': 14455, 'loss/train': 1.8126785159111023} 01/29/2022 07:24:25 - INFO - codeparrot_training - Step 14456: {'lr': 0.000421422830543534, 'samples': 2775744, 'steps': 14456, 'loss/train': 1.6535566449165344} 01/29/2022 07:24:30 - INFO - codeparrot_training - Step 14457: {'lr': 0.0004214109200641106, 'samples': 2775936, 'steps': 14457, 'loss/train': 1.1999429762363434} 01/29/2022 07:24:34 - INFO - codeparrot_training - Step 14458: {'lr': 0.00042139900885041734, 'samples': 2776128, 'steps': 14458, 'loss/train': 1.4701765179634094} 01/29/2022 07:24:40 - INFO - codeparrot_training - Step 14459: {'lr': 0.00042138709690250507, 'samples': 2776320, 'steps': 14459, 'loss/train': 1.2078534364700317} 01/29/2022 07:24:45 - INFO - codeparrot_training - Step 14460: {'lr': 0.0004213751842204249, 'samples': 2776512, 'steps': 14460, 'loss/train': 2.553554356098175} 01/29/2022 07:24:49 - INFO - codeparrot_training - Step 14461: {'lr': 0.00042136327080422785, 'samples': 2776704, 'steps': 14461, 'loss/train': 1.7477468848228455} 01/29/2022 07:24:53 - INFO - codeparrot_training - Step 14462: {'lr': 0.0004213513566539651, 'samples': 2776896, 'steps': 14462, 'loss/train': 2.066974103450775} 01/29/2022 07:24:57 - INFO - codeparrot_training - Step 14463: {'lr': 0.0004213394417696874, 'samples': 2777088, 'steps': 14463, 'loss/train': 1.8301105499267578} 01/29/2022 07:25:02 - INFO - codeparrot_training - Step 14464: {'lr': 0.00042132752615144597, 'samples': 2777280, 'steps': 14464, 'loss/train': 2.6353374123573303} 01/29/2022 07:25:07 - INFO - codeparrot_training - Step 14465: {'lr': 0.00042131560979929186, 'samples': 2777472, 'steps': 14465, 'loss/train': 2.0910340547561646} 01/29/2022 07:25:11 - INFO - codeparrot_training - Step 14466: {'lr': 0.00042130369271327605, 'samples': 2777664, 'steps': 14466, 'loss/train': 2.0229833722114563} 01/29/2022 07:25:15 - INFO - codeparrot_training - Step 14467: {'lr': 0.0004212917748934496, 'samples': 2777856, 'steps': 14467, 'loss/train': 1.414706826210022} 01/29/2022 07:25:20 - INFO - codeparrot_training - Step 14468: {'lr': 0.00042127985633986365, 'samples': 2778048, 'steps': 14468, 'loss/train': 1.334990680217743} 01/29/2022 07:25:24 - INFO - codeparrot_training - Step 14469: {'lr': 0.00042126793705256913, 'samples': 2778240, 'steps': 14469, 'loss/train': 2.2811797857284546} 01/29/2022 07:25:29 - INFO - codeparrot_training - Step 14470: {'lr': 0.00042125601703161706, 'samples': 2778432, 'steps': 14470, 'loss/train': 1.9299076795578003} 01/29/2022 07:25:34 - INFO - codeparrot_training - Step 14471: {'lr': 0.00042124409627705873, 'samples': 2778624, 'steps': 14471, 'loss/train': 2.271094501018524} 01/29/2022 07:25:38 - INFO - codeparrot_training - Step 14472: {'lr': 0.00042123217478894504, 'samples': 2778816, 'steps': 14472, 'loss/train': 1.0703908503055573} 01/29/2022 07:25:42 - INFO - codeparrot_training - Step 14473: {'lr': 0.0004212202525673271, 'samples': 2779008, 'steps': 14473, 'loss/train': 2.3345671892166138} 01/29/2022 07:25:47 - INFO - codeparrot_training - Step 14474: {'lr': 0.00042120832961225585, 'samples': 2779200, 'steps': 14474, 'loss/train': 1.6503118872642517} 01/29/2022 07:25:52 - INFO - codeparrot_training - Step 14475: {'lr': 0.00042119640592378263, 'samples': 2779392, 'steps': 14475, 'loss/train': 1.582588255405426} 01/29/2022 07:25:56 - INFO - codeparrot_training - Step 14476: {'lr': 0.00042118448150195827, 'samples': 2779584, 'steps': 14476, 'loss/train': 1.7892801761627197} 01/29/2022 07:26:00 - INFO - codeparrot_training - Step 14477: {'lr': 0.000421172556346834, 'samples': 2779776, 'steps': 14477, 'loss/train': 2.7032257318496704} 01/29/2022 07:26:04 - INFO - codeparrot_training - Step 14478: {'lr': 0.00042116063045846073, 'samples': 2779968, 'steps': 14478, 'loss/train': 0.8708471059799194} 01/29/2022 07:26:10 - INFO - codeparrot_training - Step 14479: {'lr': 0.00042114870383688985, 'samples': 2780160, 'steps': 14479, 'loss/train': 0.18720265477895737} 01/29/2022 07:26:15 - INFO - codeparrot_training - Step 14480: {'lr': 0.0004211367764821722, 'samples': 2780352, 'steps': 14480, 'loss/train': 1.8962393403053284} 01/29/2022 07:26:19 - INFO - codeparrot_training - Step 14481: {'lr': 0.00042112484839435893, 'samples': 2780544, 'steps': 14481, 'loss/train': 1.8334636688232422} 01/29/2022 07:26:23 - INFO - codeparrot_training - Step 14482: {'lr': 0.00042111291957350113, 'samples': 2780736, 'steps': 14482, 'loss/train': 1.1702342927455902} 01/29/2022 07:26:27 - INFO - codeparrot_training - Step 14483: {'lr': 0.00042110099001964996, 'samples': 2780928, 'steps': 14483, 'loss/train': 2.3151764273643494} 01/29/2022 07:26:33 - INFO - codeparrot_training - Step 14484: {'lr': 0.0004210890597328564, 'samples': 2781120, 'steps': 14484, 'loss/train': 1.9982850551605225} 01/29/2022 07:26:37 - INFO - codeparrot_training - Step 14485: {'lr': 0.0004210771287131717, 'samples': 2781312, 'steps': 14485, 'loss/train': 1.4847143590450287} 01/29/2022 07:26:41 - INFO - codeparrot_training - Step 14486: {'lr': 0.00042106519696064694, 'samples': 2781504, 'steps': 14486, 'loss/train': 2.03131502866745} 01/29/2022 07:26:46 - INFO - codeparrot_training - Step 14487: {'lr': 0.0004210532644753331, 'samples': 2781696, 'steps': 14487, 'loss/train': 1.7109069228172302} 01/29/2022 07:26:50 - INFO - codeparrot_training - Step 14488: {'lr': 0.00042104133125728146, 'samples': 2781888, 'steps': 14488, 'loss/train': 0.6296261698007584} 01/29/2022 07:26:56 - INFO - codeparrot_training - Step 14489: {'lr': 0.00042102939730654304, 'samples': 2782080, 'steps': 14489, 'loss/train': 1.353477269411087} 01/29/2022 07:27:00 - INFO - codeparrot_training - Step 14490: {'lr': 0.000421017462623169, 'samples': 2782272, 'steps': 14490, 'loss/train': 0.9455248117446899} 01/29/2022 07:27:04 - INFO - codeparrot_training - Step 14491: {'lr': 0.0004210055272072104, 'samples': 2782464, 'steps': 14491, 'loss/train': 1.5317323207855225} 01/29/2022 07:27:09 - INFO - codeparrot_training - Step 14492: {'lr': 0.00042099359105871856, 'samples': 2782656, 'steps': 14492, 'loss/train': 2.5333616137504578} 01/29/2022 07:27:13 - INFO - codeparrot_training - Step 14493: {'lr': 0.0004209816541777444, 'samples': 2782848, 'steps': 14493, 'loss/train': 0.7564753890037537} 01/29/2022 07:27:18 - INFO - codeparrot_training - Step 14494: {'lr': 0.0004209697165643391, 'samples': 2783040, 'steps': 14494, 'loss/train': 1.8638384342193604} 01/29/2022 07:27:22 - INFO - codeparrot_training - Step 14495: {'lr': 0.0004209577782185538, 'samples': 2783232, 'steps': 14495, 'loss/train': 1.2176214158535004} 01/29/2022 07:27:27 - INFO - codeparrot_training - Step 14496: {'lr': 0.0004209458391404397, 'samples': 2783424, 'steps': 14496, 'loss/train': 1.9734951853752136} 01/29/2022 07:27:31 - INFO - codeparrot_training - Step 14497: {'lr': 0.0004209338993300479, 'samples': 2783616, 'steps': 14497, 'loss/train': 1.8056280612945557} 01/29/2022 07:27:35 - INFO - codeparrot_training - Step 14498: {'lr': 0.00042092195878742954, 'samples': 2783808, 'steps': 14498, 'loss/train': 1.5617870092391968} 01/29/2022 07:27:39 - INFO - codeparrot_training - Step 14499: {'lr': 0.0004209100175126358, 'samples': 2784000, 'steps': 14499, 'loss/train': 1.6604345440864563} 01/29/2022 07:27:45 - INFO - codeparrot_training - Step 14500: {'lr': 0.0004208980755057178, 'samples': 2784192, 'steps': 14500, 'loss/train': 1.5573588609695435} 01/29/2022 07:27:49 - INFO - codeparrot_training - Step 14501: {'lr': 0.0004208861327667268, 'samples': 2784384, 'steps': 14501, 'loss/train': 1.9321839809417725} 01/29/2022 07:27:54 - INFO - codeparrot_training - Step 14502: {'lr': 0.00042087418929571377, 'samples': 2784576, 'steps': 14502, 'loss/train': 1.9565525650978088} 01/29/2022 07:27:58 - INFO - codeparrot_training - Step 14503: {'lr': 0.00042086224509272995, 'samples': 2784768, 'steps': 14503, 'loss/train': 2.150606632232666} 01/29/2022 07:28:02 - INFO - codeparrot_training - Step 14504: {'lr': 0.0004208503001578266, 'samples': 2784960, 'steps': 14504, 'loss/train': 1.2201190888881683} 01/29/2022 07:28:08 - INFO - codeparrot_training - Step 14505: {'lr': 0.00042083835449105477, 'samples': 2785152, 'steps': 14505, 'loss/train': 1.557945191860199} 01/29/2022 07:28:13 - INFO - codeparrot_training - Step 14506: {'lr': 0.00042082640809246576, 'samples': 2785344, 'steps': 14506, 'loss/train': 2.0154842734336853} 01/29/2022 07:28:17 - INFO - codeparrot_training - Step 14507: {'lr': 0.0004208144609621106, 'samples': 2785536, 'steps': 14507, 'loss/train': 2.359750986099243} 01/29/2022 07:28:21 - INFO - codeparrot_training - Step 14508: {'lr': 0.0004208025131000405, 'samples': 2785728, 'steps': 14508, 'loss/train': 1.7493817806243896} 01/29/2022 07:28:25 - INFO - codeparrot_training - Step 14509: {'lr': 0.0004207905645063067, 'samples': 2785920, 'steps': 14509, 'loss/train': 3.7052736282348633} 01/29/2022 07:28:31 - INFO - codeparrot_training - Step 14510: {'lr': 0.00042077861518096033, 'samples': 2786112, 'steps': 14510, 'loss/train': 1.6920163035392761} 01/29/2022 07:28:35 - INFO - codeparrot_training - Step 14511: {'lr': 0.0004207666651240526, 'samples': 2786304, 'steps': 14511, 'loss/train': 1.2043116688728333} 01/29/2022 07:28:39 - INFO - codeparrot_training - Step 14512: {'lr': 0.0004207547143356347, 'samples': 2786496, 'steps': 14512, 'loss/train': 1.7158997654914856} 01/29/2022 07:28:43 - INFO - codeparrot_training - Step 14513: {'lr': 0.00042074276281575787, 'samples': 2786688, 'steps': 14513, 'loss/train': 1.61361962556839} 01/29/2022 07:28:48 - INFO - codeparrot_training - Step 14514: {'lr': 0.00042073081056447325, 'samples': 2786880, 'steps': 14514, 'loss/train': 1.6901006698608398} 01/29/2022 07:28:53 - INFO - codeparrot_training - Step 14515: {'lr': 0.00042071885758183204, 'samples': 2787072, 'steps': 14515, 'loss/train': 2.0431739687919617} 01/29/2022 07:28:57 - INFO - codeparrot_training - Step 14516: {'lr': 0.00042070690386788545, 'samples': 2787264, 'steps': 14516, 'loss/train': 0.84835284948349} 01/29/2022 07:29:01 - INFO - codeparrot_training - Step 14517: {'lr': 0.0004206949494226847, 'samples': 2787456, 'steps': 14517, 'loss/train': 0.6445474326610565} 01/29/2022 07:29:05 - INFO - codeparrot_training - Step 14518: {'lr': 0.000420682994246281, 'samples': 2787648, 'steps': 14518, 'loss/train': 1.3481811583042145} 01/29/2022 07:29:10 - INFO - codeparrot_training - Step 14519: {'lr': 0.00042067103833872554, 'samples': 2787840, 'steps': 14519, 'loss/train': 0.6300614476203918} 01/29/2022 07:29:16 - INFO - codeparrot_training - Step 14520: {'lr': 0.0004206590817000695, 'samples': 2788032, 'steps': 14520, 'loss/train': 1.7990416288375854} 01/29/2022 07:29:20 - INFO - codeparrot_training - Step 14521: {'lr': 0.0004206471243303642, 'samples': 2788224, 'steps': 14521, 'loss/train': 2.2437087893486023} 01/29/2022 07:29:24 - INFO - codeparrot_training - Step 14522: {'lr': 0.0004206351662296608, 'samples': 2788416, 'steps': 14522, 'loss/train': 1.1347453594207764} 01/29/2022 07:29:29 - INFO - codeparrot_training - Step 14523: {'lr': 0.0004206232073980105, 'samples': 2788608, 'steps': 14523, 'loss/train': 1.3150812685489655} 01/29/2022 07:29:34 - INFO - codeparrot_training - Step 14524: {'lr': 0.00042061124783546454, 'samples': 2788800, 'steps': 14524, 'loss/train': 1.524727463722229} 01/29/2022 07:29:38 - INFO - codeparrot_training - Step 14525: {'lr': 0.0004205992875420742, 'samples': 2788992, 'steps': 14525, 'loss/train': 1.6000065207481384} 01/29/2022 07:29:43 - INFO - codeparrot_training - Step 14526: {'lr': 0.0004205873265178907, 'samples': 2789184, 'steps': 14526, 'loss/train': 1.3574143052101135} 01/29/2022 07:29:47 - INFO - codeparrot_training - Step 14527: {'lr': 0.0004205753647629653, 'samples': 2789376, 'steps': 14527, 'loss/train': 2.5441283583641052} 01/29/2022 07:29:51 - INFO - codeparrot_training - Step 14528: {'lr': 0.0004205634022773491, 'samples': 2789568, 'steps': 14528, 'loss/train': 1.8022534847259521} 01/29/2022 07:29:55 - INFO - codeparrot_training - Step 14529: {'lr': 0.0004205514390610935, 'samples': 2789760, 'steps': 14529, 'loss/train': 1.247961848974228} 01/29/2022 07:30:01 - INFO - codeparrot_training - Step 14530: {'lr': 0.00042053947511424975, 'samples': 2789952, 'steps': 14530, 'loss/train': 1.529762864112854} 01/29/2022 07:30:05 - INFO - codeparrot_training - Step 14531: {'lr': 0.00042052751043686895, 'samples': 2790144, 'steps': 14531, 'loss/train': 1.5092960000038147} 01/29/2022 07:30:09 - INFO - codeparrot_training - Step 14532: {'lr': 0.00042051554502900245, 'samples': 2790336, 'steps': 14532, 'loss/train': 1.6491490602493286} 01/29/2022 07:30:14 - INFO - codeparrot_training - Step 14533: {'lr': 0.0004205035788907015, 'samples': 2790528, 'steps': 14533, 'loss/train': 0.9834522306919098} 01/29/2022 07:30:19 - INFO - codeparrot_training - Step 14534: {'lr': 0.0004204916120220174, 'samples': 2790720, 'steps': 14534, 'loss/train': 2.297532320022583} 01/29/2022 07:30:23 - INFO - codeparrot_training - Step 14535: {'lr': 0.00042047964442300137, 'samples': 2790912, 'steps': 14535, 'loss/train': 1.34956955909729} 01/29/2022 07:30:28 - INFO - codeparrot_training - Step 14536: {'lr': 0.0004204676760937046, 'samples': 2791104, 'steps': 14536, 'loss/train': 1.7398833632469177} 01/29/2022 07:30:32 - INFO - codeparrot_training - Step 14537: {'lr': 0.00042045570703417857, 'samples': 2791296, 'steps': 14537, 'loss/train': 1.6283987760543823} 01/29/2022 07:30:36 - INFO - codeparrot_training - Step 14538: {'lr': 0.00042044373724447434, 'samples': 2791488, 'steps': 14538, 'loss/train': 1.4288078248500824} 01/29/2022 07:30:41 - INFO - codeparrot_training - Step 14539: {'lr': 0.0004204317667246432, 'samples': 2791680, 'steps': 14539, 'loss/train': 1.4381072223186493} 01/29/2022 07:30:47 - INFO - codeparrot_training - Step 14540: {'lr': 0.00042041979547473665, 'samples': 2791872, 'steps': 14540, 'loss/train': 1.2457567155361176} 01/29/2022 07:30:52 - INFO - codeparrot_training - Step 14541: {'lr': 0.0004204078234948057, 'samples': 2792064, 'steps': 14541, 'loss/train': 1.9157531261444092} 01/29/2022 07:30:56 - INFO - codeparrot_training - Step 14542: {'lr': 0.00042039585078490173, 'samples': 2792256, 'steps': 14542, 'loss/train': 1.9193170666694641} 01/29/2022 07:31:00 - INFO - codeparrot_training - Step 14543: {'lr': 0.000420383877345076, 'samples': 2792448, 'steps': 14543, 'loss/train': 1.9541226625442505} 01/29/2022 07:31:04 - INFO - codeparrot_training - Step 14544: {'lr': 0.00042037190317538, 'samples': 2792640, 'steps': 14544, 'loss/train': 1.0173795819282532} 01/29/2022 07:31:10 - INFO - codeparrot_training - Step 14545: {'lr': 0.00042035992827586474, 'samples': 2792832, 'steps': 14545, 'loss/train': 0.5195764303207397} 01/29/2022 07:31:14 - INFO - codeparrot_training - Step 14546: {'lr': 0.00042034795264658163, 'samples': 2793024, 'steps': 14546, 'loss/train': 1.6899443864822388} 01/29/2022 07:31:18 - INFO - codeparrot_training - Step 14547: {'lr': 0.00042033597628758206, 'samples': 2793216, 'steps': 14547, 'loss/train': 0.7192261219024658} 01/29/2022 07:31:23 - INFO - codeparrot_training - Step 14548: {'lr': 0.00042032399919891724, 'samples': 2793408, 'steps': 14548, 'loss/train': 0.3194011375308037} 01/29/2022 07:31:27 - INFO - codeparrot_training - Step 14549: {'lr': 0.0004203120213806385, 'samples': 2793600, 'steps': 14549, 'loss/train': 1.199869841337204} 01/29/2022 07:31:33 - INFO - codeparrot_training - Step 14550: {'lr': 0.0004203000428327971, 'samples': 2793792, 'steps': 14550, 'loss/train': 1.59242445230484} 01/29/2022 07:31:37 - INFO - codeparrot_training - Step 14551: {'lr': 0.00042028806355544443, 'samples': 2793984, 'steps': 14551, 'loss/train': 2.2658718824386597} 01/29/2022 07:31:41 - INFO - codeparrot_training - Step 14552: {'lr': 0.0004202760835486317, 'samples': 2794176, 'steps': 14552, 'loss/train': 0.2548123374581337} 01/29/2022 07:31:46 - INFO - codeparrot_training - Step 14553: {'lr': 0.00042026410281241033, 'samples': 2794368, 'steps': 14553, 'loss/train': 1.7108582854270935} 01/29/2022 07:31:50 - INFO - codeparrot_training - Step 14554: {'lr': 0.00042025212134683165, 'samples': 2794560, 'steps': 14554, 'loss/train': 1.6059443950653076} 01/29/2022 07:31:55 - INFO - codeparrot_training - Step 14555: {'lr': 0.0004202401391519469, 'samples': 2794752, 'steps': 14555, 'loss/train': 1.8725786805152893} 01/29/2022 07:31:59 - INFO - codeparrot_training - Step 14556: {'lr': 0.0004202281562278075, 'samples': 2794944, 'steps': 14556, 'loss/train': 1.8244423270225525} 01/29/2022 07:32:04 - INFO - codeparrot_training - Step 14557: {'lr': 0.0004202161725744647, 'samples': 2795136, 'steps': 14557, 'loss/train': 2.208681106567383} 01/29/2022 07:32:08 - INFO - codeparrot_training - Step 14558: {'lr': 0.0004202041881919699, 'samples': 2795328, 'steps': 14558, 'loss/train': 0.6147661954164505} 01/29/2022 07:32:12 - INFO - codeparrot_training - Step 14559: {'lr': 0.0004201922030803743, 'samples': 2795520, 'steps': 14559, 'loss/train': 2.457047939300537} 01/29/2022 07:32:18 - INFO - codeparrot_training - Step 14560: {'lr': 0.0004201802172397295, 'samples': 2795712, 'steps': 14560, 'loss/train': 1.6951945424079895} 01/29/2022 07:32:22 - INFO - codeparrot_training - Step 14561: {'lr': 0.0004201682306700866, 'samples': 2795904, 'steps': 14561, 'loss/train': 0.2579018622636795} 01/29/2022 07:32:26 - INFO - codeparrot_training - Step 14562: {'lr': 0.00042015624337149703, 'samples': 2796096, 'steps': 14562, 'loss/train': 1.2316477596759796} 01/29/2022 07:32:30 - INFO - codeparrot_training - Step 14563: {'lr': 0.0004201442553440121, 'samples': 2796288, 'steps': 14563, 'loss/train': 0.789521187543869} 01/29/2022 07:32:36 - INFO - codeparrot_training - Step 14564: {'lr': 0.00042013226658768333, 'samples': 2796480, 'steps': 14564, 'loss/train': 1.5324586629867554} 01/29/2022 07:32:40 - INFO - codeparrot_training - Step 14565: {'lr': 0.0004201202771025618, 'samples': 2796672, 'steps': 14565, 'loss/train': 1.3721378445625305} 01/29/2022 07:32:45 - INFO - codeparrot_training - Step 14566: {'lr': 0.0004201082868886992, 'samples': 2796864, 'steps': 14566, 'loss/train': 1.6852607131004333} 01/29/2022 07:32:49 - INFO - codeparrot_training - Step 14567: {'lr': 0.00042009629594614656, 'samples': 2797056, 'steps': 14567, 'loss/train': 2.5912283062934875} 01/29/2022 07:32:53 - INFO - codeparrot_training - Step 14568: {'lr': 0.0004200843042749555, 'samples': 2797248, 'steps': 14568, 'loss/train': 1.94737046957016} 01/29/2022 07:32:58 - INFO - codeparrot_training - Step 14569: {'lr': 0.0004200723118751772, 'samples': 2797440, 'steps': 14569, 'loss/train': 1.1588698625564575} 01/29/2022 07:33:03 - INFO - codeparrot_training - Step 14570: {'lr': 0.00042006031874686315, 'samples': 2797632, 'steps': 14570, 'loss/train': 2.112196683883667} 01/29/2022 07:33:07 - INFO - codeparrot_training - Step 14571: {'lr': 0.00042004832489006474, 'samples': 2797824, 'steps': 14571, 'loss/train': 2.183388590812683} 01/29/2022 07:33:11 - INFO - codeparrot_training - Step 14572: {'lr': 0.0004200363303048332, 'samples': 2798016, 'steps': 14572, 'loss/train': 1.5028687119483948} 01/29/2022 07:33:16 - INFO - codeparrot_training - Step 14573: {'lr': 0.00042002433499122016, 'samples': 2798208, 'steps': 14573, 'loss/train': 1.7206217050552368} 01/29/2022 07:33:22 - INFO - codeparrot_training - Step 14574: {'lr': 0.00042001233894927684, 'samples': 2798400, 'steps': 14574, 'loss/train': 2.224058747291565} 01/29/2022 07:33:26 - INFO - codeparrot_training - Step 14575: {'lr': 0.0004200003421790546, 'samples': 2798592, 'steps': 14575, 'loss/train': 2.3225501775741577} 01/29/2022 07:33:30 - INFO - codeparrot_training - Step 14576: {'lr': 0.0004199883446806048, 'samples': 2798784, 'steps': 14576, 'loss/train': 1.7892619371414185} 01/29/2022 07:33:34 - INFO - codeparrot_training - Step 14577: {'lr': 0.00041997634645397897, 'samples': 2798976, 'steps': 14577, 'loss/train': 2.125301778316498} 01/29/2022 07:33:39 - INFO - codeparrot_training - Step 14578: {'lr': 0.0004199643474992285, 'samples': 2799168, 'steps': 14578, 'loss/train': 0.8241084516048431} 01/29/2022 07:33:44 - INFO - codeparrot_training - Step 14579: {'lr': 0.00041995234781640466, 'samples': 2799360, 'steps': 14579, 'loss/train': 1.6846738457679749} 01/29/2022 07:33:48 - INFO - codeparrot_training - Step 14580: {'lr': 0.00041994034740555896, 'samples': 2799552, 'steps': 14580, 'loss/train': 1.8457682132720947} 01/29/2022 07:33:53 - INFO - codeparrot_training - Step 14581: {'lr': 0.00041992834626674273, 'samples': 2799744, 'steps': 14581, 'loss/train': 1.5464458465576172} 01/29/2022 07:33:57 - INFO - codeparrot_training - Step 14582: {'lr': 0.0004199163444000075, 'samples': 2799936, 'steps': 14582, 'loss/train': 0.635113924741745} 01/29/2022 07:34:01 - INFO - codeparrot_training - Step 14583: {'lr': 0.00041990434180540453, 'samples': 2800128, 'steps': 14583, 'loss/train': 1.9578734636306763} 01/29/2022 07:34:06 - INFO - codeparrot_training - Step 14584: {'lr': 0.00041989233848298534, 'samples': 2800320, 'steps': 14584, 'loss/train': 1.8895987272262573} 01/29/2022 07:34:11 - INFO - codeparrot_training - Step 14585: {'lr': 0.00041988033443280136, 'samples': 2800512, 'steps': 14585, 'loss/train': 2.2803186178207397} 01/29/2022 07:34:15 - INFO - codeparrot_training - Step 14586: {'lr': 0.00041986832965490396, 'samples': 2800704, 'steps': 14586, 'loss/train': 1.7194240093231201} 01/29/2022 07:34:19 - INFO - codeparrot_training - Step 14587: {'lr': 0.0004198563241493445, 'samples': 2800896, 'steps': 14587, 'loss/train': 2.1906575560569763} 01/29/2022 07:34:23 - INFO - codeparrot_training - Step 14588: {'lr': 0.00041984431791617456, 'samples': 2801088, 'steps': 14588, 'loss/train': 0.6385345011949539} 01/29/2022 07:34:31 - INFO - codeparrot_training - Step 14589: {'lr': 0.00041983231095544545, 'samples': 2801280, 'steps': 14589, 'loss/train': 1.9744606018066406} 01/29/2022 07:34:35 - INFO - codeparrot_training - Step 14590: {'lr': 0.00041982030326720866, 'samples': 2801472, 'steps': 14590, 'loss/train': 2.23634934425354} 01/29/2022 07:34:39 - INFO - codeparrot_training - Step 14591: {'lr': 0.00041980829485151563, 'samples': 2801664, 'steps': 14591, 'loss/train': 1.6111652255058289} 01/29/2022 07:34:44 - INFO - codeparrot_training - Step 14592: {'lr': 0.00041979628570841776, 'samples': 2801856, 'steps': 14592, 'loss/train': 2.0182803869247437} 01/29/2022 07:34:48 - INFO - codeparrot_training - Step 14593: {'lr': 0.00041978427583796654, 'samples': 2802048, 'steps': 14593, 'loss/train': 0.8850443959236145} 01/29/2022 07:34:53 - INFO - codeparrot_training - Step 14594: {'lr': 0.00041977226524021337, 'samples': 2802240, 'steps': 14594, 'loss/train': 1.8767670392990112} 01/29/2022 07:34:57 - INFO - codeparrot_training - Step 14595: {'lr': 0.0004197602539152098, 'samples': 2802432, 'steps': 14595, 'loss/train': 1.5370686650276184} 01/29/2022 07:35:02 - INFO - codeparrot_training - Step 14596: {'lr': 0.00041974824186300706, 'samples': 2802624, 'steps': 14596, 'loss/train': 1.6400495767593384} 01/29/2022 07:35:06 - INFO - codeparrot_training - Step 14597: {'lr': 0.0004197362290836569, 'samples': 2802816, 'steps': 14597, 'loss/train': 1.5525126457214355} 01/29/2022 07:35:10 - INFO - codeparrot_training - Step 14598: {'lr': 0.00041972421557721055, 'samples': 2803008, 'steps': 14598, 'loss/train': 1.1191623508930206} 01/29/2022 07:35:17 - INFO - codeparrot_training - Step 14599: {'lr': 0.00041971220134371957, 'samples': 2803200, 'steps': 14599, 'loss/train': 1.3893853425979614} 01/29/2022 07:35:22 - INFO - codeparrot_training - Step 14600: {'lr': 0.00041970018638323546, 'samples': 2803392, 'steps': 14600, 'loss/train': 1.6352371573448181} 01/29/2022 07:35:26 - INFO - codeparrot_training - Step 14601: {'lr': 0.0004196881706958096, 'samples': 2803584, 'steps': 14601, 'loss/train': 1.7283992171287537} 01/29/2022 07:35:30 - INFO - codeparrot_training - Step 14602: {'lr': 0.00041967615428149346, 'samples': 2803776, 'steps': 14602, 'loss/train': 1.598533272743225} 01/29/2022 07:35:34 - INFO - codeparrot_training - Step 14603: {'lr': 0.0004196641371403386, 'samples': 2803968, 'steps': 14603, 'loss/train': 2.080232262611389} 01/29/2022 07:35:40 - INFO - codeparrot_training - Step 14604: {'lr': 0.00041965211927239644, 'samples': 2804160, 'steps': 14604, 'loss/train': 2.1582512855529785} 01/29/2022 07:35:44 - INFO - codeparrot_training - Step 14605: {'lr': 0.0004196401006777185, 'samples': 2804352, 'steps': 14605, 'loss/train': 1.2556570172309875} 01/29/2022 07:35:48 - INFO - codeparrot_training - Step 14606: {'lr': 0.00041962808135635624, 'samples': 2804544, 'steps': 14606, 'loss/train': 1.070598542690277} 01/29/2022 07:35:52 - INFO - codeparrot_training - Step 14607: {'lr': 0.00041961606130836105, 'samples': 2804736, 'steps': 14607, 'loss/train': 2.1192256808280945} 01/29/2022 07:35:57 - INFO - codeparrot_training - Step 14608: {'lr': 0.0004196040405337845, 'samples': 2804928, 'steps': 14608, 'loss/train': 1.4337134957313538} 01/29/2022 07:36:02 - INFO - codeparrot_training - Step 14609: {'lr': 0.0004195920190326782, 'samples': 2805120, 'steps': 14609, 'loss/train': 1.8143699169158936} 01/29/2022 07:36:06 - INFO - codeparrot_training - Step 14610: {'lr': 0.0004195799968050935, 'samples': 2805312, 'steps': 14610, 'loss/train': 1.935801386833191} 01/29/2022 07:36:11 - INFO - codeparrot_training - Step 14611: {'lr': 0.000419567973851082, 'samples': 2805504, 'steps': 14611, 'loss/train': 0.5590234994888306} 01/29/2022 07:36:15 - INFO - codeparrot_training - Step 14612: {'lr': 0.0004195559501706951, 'samples': 2805696, 'steps': 14612, 'loss/train': 1.3749047219753265} 01/29/2022 07:36:19 - INFO - codeparrot_training - Step 14613: {'lr': 0.00041954392576398433, 'samples': 2805888, 'steps': 14613, 'loss/train': 1.6970547437667847} 01/29/2022 07:36:26 - INFO - codeparrot_training - Step 14614: {'lr': 0.0004195319006310012, 'samples': 2806080, 'steps': 14614, 'loss/train': 2.0781404972076416} 01/29/2022 07:36:30 - INFO - codeparrot_training - Step 14615: {'lr': 0.0004195198747717973, 'samples': 2806272, 'steps': 14615, 'loss/train': 1.9426288604736328} 01/29/2022 07:36:35 - INFO - codeparrot_training - Step 14616: {'lr': 0.00041950784818642404, 'samples': 2806464, 'steps': 14616, 'loss/train': 1.8719174265861511} 01/29/2022 07:36:39 - INFO - codeparrot_training - Step 14617: {'lr': 0.000419495820874933, 'samples': 2806656, 'steps': 14617, 'loss/train': 1.7556384801864624} 01/29/2022 07:36:43 - INFO - codeparrot_training - Step 14618: {'lr': 0.0004194837928373757, 'samples': 2806848, 'steps': 14618, 'loss/train': 0.8839792907238007} 01/29/2022 07:36:48 - INFO - codeparrot_training - Step 14619: {'lr': 0.0004194717640738036, 'samples': 2807040, 'steps': 14619, 'loss/train': 1.9098783731460571} 01/29/2022 07:36:53 - INFO - codeparrot_training - Step 14620: {'lr': 0.0004194597345842683, 'samples': 2807232, 'steps': 14620, 'loss/train': 2.1255460381507874} 01/29/2022 07:36:57 - INFO - codeparrot_training - Step 14621: {'lr': 0.00041944770436882134, 'samples': 2807424, 'steps': 14621, 'loss/train': 1.0711853206157684} 01/29/2022 07:37:01 - INFO - codeparrot_training - Step 14622: {'lr': 0.00041943567342751423, 'samples': 2807616, 'steps': 14622, 'loss/train': 1.4763576686382294} 01/29/2022 07:37:05 - INFO - codeparrot_training - Step 14623: {'lr': 0.0004194236417603985, 'samples': 2807808, 'steps': 14623, 'loss/train': 2.053570568561554} 01/29/2022 07:37:11 - INFO - codeparrot_training - Step 14624: {'lr': 0.0004194116093675256, 'samples': 2808000, 'steps': 14624, 'loss/train': 1.3271527290344238} 01/29/2022 07:37:15 - INFO - codeparrot_training - Step 14625: {'lr': 0.0004193995762489472, 'samples': 2808192, 'steps': 14625, 'loss/train': 0.8372401893138885} 01/29/2022 07:37:19 - INFO - codeparrot_training - Step 14626: {'lr': 0.0004193875424047148, 'samples': 2808384, 'steps': 14626, 'loss/train': 0.767348438501358} 01/29/2022 07:37:23 - INFO - codeparrot_training - Step 14627: {'lr': 0.00041937550783488, 'samples': 2808576, 'steps': 14627, 'loss/train': 0.10987803339958191} 01/29/2022 07:37:28 - INFO - codeparrot_training - Step 14628: {'lr': 0.00041936347253949426, 'samples': 2808768, 'steps': 14628, 'loss/train': 1.8783091306686401} 01/29/2022 07:37:35 - INFO - codeparrot_training - Step 14629: {'lr': 0.00041935143651860917, 'samples': 2808960, 'steps': 14629, 'loss/train': 2.0904820561408997} 01/29/2022 07:37:39 - INFO - codeparrot_training - Step 14630: {'lr': 0.0004193393997722764, 'samples': 2809152, 'steps': 14630, 'loss/train': 1.527367115020752} 01/29/2022 07:37:43 - INFO - codeparrot_training - Step 14631: {'lr': 0.00041932736230054725, 'samples': 2809344, 'steps': 14631, 'loss/train': 1.1943849921226501} 01/29/2022 07:37:47 - INFO - codeparrot_training - Step 14632: {'lr': 0.0004193153241034736, 'samples': 2809536, 'steps': 14632, 'loss/train': 1.369998425245285} 01/29/2022 07:37:52 - INFO - codeparrot_training - Step 14633: {'lr': 0.00041930328518110675, 'samples': 2809728, 'steps': 14633, 'loss/train': 0.9665107727050781} 01/29/2022 07:37:57 - INFO - codeparrot_training - Step 14634: {'lr': 0.0004192912455334985, 'samples': 2809920, 'steps': 14634, 'loss/train': 2.3331229090690613} 01/29/2022 07:38:01 - INFO - codeparrot_training - Step 14635: {'lr': 0.0004192792051607002, 'samples': 2810112, 'steps': 14635, 'loss/train': 0.23352138698101044} 01/29/2022 07:38:05 - INFO - codeparrot_training - Step 14636: {'lr': 0.00041926716406276367, 'samples': 2810304, 'steps': 14636, 'loss/train': 0.6825500428676605} 01/29/2022 07:38:10 - INFO - codeparrot_training - Step 14637: {'lr': 0.0004192551222397402, 'samples': 2810496, 'steps': 14637, 'loss/train': 1.3700804114341736} 01/29/2022 07:38:14 - INFO - codeparrot_training - Step 14638: {'lr': 0.0004192430796916816, 'samples': 2810688, 'steps': 14638, 'loss/train': 1.1645970940589905} 01/29/2022 07:38:21 - INFO - codeparrot_training - Step 14639: {'lr': 0.0004192310364186394, 'samples': 2810880, 'steps': 14639, 'loss/train': 1.8817657828330994} 01/29/2022 07:38:25 - INFO - codeparrot_training - Step 14640: {'lr': 0.0004192189924206652, 'samples': 2811072, 'steps': 14640, 'loss/train': 1.665916621685028} 01/29/2022 07:38:29 - INFO - codeparrot_training - Step 14641: {'lr': 0.0004192069476978105, 'samples': 2811264, 'steps': 14641, 'loss/train': 1.8858795762062073} 01/29/2022 07:38:33 - INFO - codeparrot_training - Step 14642: {'lr': 0.000419194902250127, 'samples': 2811456, 'steps': 14642, 'loss/train': 1.3279864490032196} 01/29/2022 07:38:38 - INFO - codeparrot_training - Step 14643: {'lr': 0.0004191828560776663, 'samples': 2811648, 'steps': 14643, 'loss/train': 0.7437730729579926} 01/29/2022 07:38:43 - INFO - codeparrot_training - Step 14644: {'lr': 0.00041917080918047996, 'samples': 2811840, 'steps': 14644, 'loss/train': 1.0728041231632233} 01/29/2022 07:38:48 - INFO - codeparrot_training - Step 14645: {'lr': 0.00041915876155861954, 'samples': 2812032, 'steps': 14645, 'loss/train': 2.898564040660858} 01/29/2022 07:38:52 - INFO - codeparrot_training - Step 14646: {'lr': 0.0004191467132121367, 'samples': 2812224, 'steps': 14646, 'loss/train': 0.040412239730358124} 01/29/2022 07:38:56 - INFO - codeparrot_training - Step 14647: {'lr': 0.00041913466414108315, 'samples': 2812416, 'steps': 14647, 'loss/train': 1.7324556112289429} 01/29/2022 07:39:00 - INFO - codeparrot_training - Step 14648: {'lr': 0.0004191226143455103, 'samples': 2812608, 'steps': 14648, 'loss/train': 2.567655086517334} 01/29/2022 07:39:06 - INFO - codeparrot_training - Step 14649: {'lr': 0.00041911056382546997, 'samples': 2812800, 'steps': 14649, 'loss/train': 1.8096202611923218} 01/29/2022 07:39:10 - INFO - codeparrot_training - Step 14650: {'lr': 0.00041909851258101357, 'samples': 2812992, 'steps': 14650, 'loss/train': 2.0645905137062073} 01/29/2022 07:39:15 - INFO - codeparrot_training - Step 14651: {'lr': 0.0004190864606121929, 'samples': 2813184, 'steps': 14651, 'loss/train': 1.836733102798462} 01/29/2022 07:39:19 - INFO - codeparrot_training - Step 14652: {'lr': 0.0004190744079190595, 'samples': 2813376, 'steps': 14652, 'loss/train': 2.1375603675842285} 01/29/2022 07:39:24 - INFO - codeparrot_training - Step 14653: {'lr': 0.0004190623545016651, 'samples': 2813568, 'steps': 14653, 'loss/train': 1.5396214127540588} 01/29/2022 07:39:28 - INFO - codeparrot_training - Step 14654: {'lr': 0.00041905030036006106, 'samples': 2813760, 'steps': 14654, 'loss/train': 0.9718827903270721} 01/29/2022 07:39:33 - INFO - codeparrot_training - Step 14655: {'lr': 0.00041903824549429936, 'samples': 2813952, 'steps': 14655, 'loss/train': 2.4884321093559265} 01/29/2022 07:39:37 - INFO - codeparrot_training - Step 14656: {'lr': 0.00041902618990443156, 'samples': 2814144, 'steps': 14656, 'loss/train': 1.8059211373329163} 01/29/2022 07:39:41 - INFO - codeparrot_training - Step 14657: {'lr': 0.0004190141335905091, 'samples': 2814336, 'steps': 14657, 'loss/train': 0.8698606789112091} 01/29/2022 07:39:48 - INFO - codeparrot_training - Step 14658: {'lr': 0.0004190020765525838, 'samples': 2814528, 'steps': 14658, 'loss/train': 2.0184386372566223} 01/29/2022 07:39:53 - INFO - codeparrot_training - Step 14659: {'lr': 0.0004189900187907073, 'samples': 2814720, 'steps': 14659, 'loss/train': 1.746178686618805} 01/29/2022 07:39:57 - INFO - codeparrot_training - Step 14660: {'lr': 0.0004189779603049312, 'samples': 2814912, 'steps': 14660, 'loss/train': 1.8160216212272644} 01/29/2022 07:40:01 - INFO - codeparrot_training - Step 14661: {'lr': 0.00041896590109530713, 'samples': 2815104, 'steps': 14661, 'loss/train': 1.7563700079917908} 01/29/2022 07:40:05 - INFO - codeparrot_training - Step 14662: {'lr': 0.00041895384116188685, 'samples': 2815296, 'steps': 14662, 'loss/train': 2.1772196888923645} 01/29/2022 07:40:11 - INFO - codeparrot_training - Step 14663: {'lr': 0.000418941780504722, 'samples': 2815488, 'steps': 14663, 'loss/train': 2.3637789487838745} 01/29/2022 07:40:15 - INFO - codeparrot_training - Step 14664: {'lr': 0.00041892971912386415, 'samples': 2815680, 'steps': 14664, 'loss/train': 1.5390849709510803} 01/29/2022 07:40:19 - INFO - codeparrot_training - Step 14665: {'lr': 0.000418917657019365, 'samples': 2815872, 'steps': 14665, 'loss/train': 2.0491570830345154} 01/29/2022 07:40:23 - INFO - codeparrot_training - Step 14666: {'lr': 0.0004189055941912763, 'samples': 2816064, 'steps': 14666, 'loss/train': 3.2310376167297363} 01/29/2022 07:40:28 - INFO - codeparrot_training - Step 14667: {'lr': 0.0004188935306396496, 'samples': 2816256, 'steps': 14667, 'loss/train': 1.5629498362541199} 01/29/2022 07:40:33 - INFO - codeparrot_training - Step 14668: {'lr': 0.00041888146636453674, 'samples': 2816448, 'steps': 14668, 'loss/train': 1.9288848638534546} 01/29/2022 07:40:37 - INFO - codeparrot_training - Step 14669: {'lr': 0.0004188694013659892, 'samples': 2816640, 'steps': 14669, 'loss/train': 1.6100462079048157} 01/29/2022 07:40:41 - INFO - codeparrot_training - Step 14670: {'lr': 0.0004188573356440588, 'samples': 2816832, 'steps': 14670, 'loss/train': 1.154600590467453} 01/29/2022 07:40:46 - INFO - codeparrot_training - Step 14671: {'lr': 0.0004188452691987973, 'samples': 2817024, 'steps': 14671, 'loss/train': 1.9099713563919067} 01/29/2022 07:40:50 - INFO - codeparrot_training - Step 14672: {'lr': 0.0004188332020302561, 'samples': 2817216, 'steps': 14672, 'loss/train': 2.0441490411758423} 01/29/2022 07:40:57 - INFO - codeparrot_training - Step 14673: {'lr': 0.0004188211341384872, 'samples': 2817408, 'steps': 14673, 'loss/train': 1.5887591242790222} 01/29/2022 07:41:01 - INFO - codeparrot_training - Step 14674: {'lr': 0.0004188090655235421, 'samples': 2817600, 'steps': 14674, 'loss/train': 0.9292658865451813} 01/29/2022 07:41:06 - INFO - codeparrot_training - Step 14675: {'lr': 0.00041879699618547263, 'samples': 2817792, 'steps': 14675, 'loss/train': 1.9377737045288086} 01/29/2022 07:41:10 - INFO - codeparrot_training - Step 14676: {'lr': 0.0004187849261243304, 'samples': 2817984, 'steps': 14676, 'loss/train': 1.1661002039909363} 01/29/2022 07:41:14 - INFO - codeparrot_training - Step 14677: {'lr': 0.0004187728553401671, 'samples': 2818176, 'steps': 14677, 'loss/train': 1.1482776403427124} 01/29/2022 07:41:19 - INFO - codeparrot_training - Step 14678: {'lr': 0.0004187607838330345, 'samples': 2818368, 'steps': 14678, 'loss/train': 1.0984405875205994} 01/29/2022 07:41:24 - INFO - codeparrot_training - Step 14679: {'lr': 0.0004187487116029843, 'samples': 2818560, 'steps': 14679, 'loss/train': 0.6392176151275635} 01/29/2022 07:41:28 - INFO - codeparrot_training - Step 14680: {'lr': 0.0004187366386500683, 'samples': 2818752, 'steps': 14680, 'loss/train': 2.210580289363861} 01/29/2022 07:41:32 - INFO - codeparrot_training - Step 14681: {'lr': 0.00041872456497433797, 'samples': 2818944, 'steps': 14681, 'loss/train': 2.5347731709480286} 01/29/2022 07:41:36 - INFO - codeparrot_training - Step 14682: {'lr': 0.00041871249057584526, 'samples': 2819136, 'steps': 14682, 'loss/train': 1.851693570613861} 01/29/2022 07:41:44 - INFO - codeparrot_training - Step 14683: {'lr': 0.00041870041545464176, 'samples': 2819328, 'steps': 14683, 'loss/train': 2.58297336101532} 01/29/2022 07:41:48 - INFO - codeparrot_training - Step 14684: {'lr': 0.00041868833961077935, 'samples': 2819520, 'steps': 14684, 'loss/train': 1.2033694088459015} 01/29/2022 07:41:52 - INFO - codeparrot_training - Step 14685: {'lr': 0.0004186762630443096, 'samples': 2819712, 'steps': 14685, 'loss/train': 0.912581741809845} 01/29/2022 07:41:56 - INFO - codeparrot_training - Step 14686: {'lr': 0.0004186641857552842, 'samples': 2819904, 'steps': 14686, 'loss/train': 1.98110032081604} 01/29/2022 07:42:01 - INFO - codeparrot_training - Step 14687: {'lr': 0.0004186521077437551, 'samples': 2820096, 'steps': 14687, 'loss/train': 1.3444613814353943} 01/29/2022 07:42:06 - INFO - codeparrot_training - Step 14688: {'lr': 0.00041864002900977393, 'samples': 2820288, 'steps': 14688, 'loss/train': 0.6049502044916153} 01/29/2022 07:42:10 - INFO - codeparrot_training - Step 14689: {'lr': 0.0004186279495533923, 'samples': 2820480, 'steps': 14689, 'loss/train': 1.6209678053855896} 01/29/2022 07:42:14 - INFO - codeparrot_training - Step 14690: {'lr': 0.0004186158693746622, 'samples': 2820672, 'steps': 14690, 'loss/train': 1.8880231976509094} 01/29/2022 07:42:19 - INFO - codeparrot_training - Step 14691: {'lr': 0.0004186037884736352, 'samples': 2820864, 'steps': 14691, 'loss/train': 1.597578227519989} 01/29/2022 07:42:23 - INFO - codeparrot_training - Step 14692: {'lr': 0.0004185917068503632, 'samples': 2821056, 'steps': 14692, 'loss/train': 2.1217914819717407} 01/29/2022 07:42:28 - INFO - codeparrot_training - Step 14693: {'lr': 0.00041857962450489786, 'samples': 2821248, 'steps': 14693, 'loss/train': 1.9438812732696533} 01/29/2022 07:42:32 - INFO - codeparrot_training - Step 14694: {'lr': 0.0004185675414372908, 'samples': 2821440, 'steps': 14694, 'loss/train': 1.7536336183547974} 01/29/2022 07:42:37 - INFO - codeparrot_training - Step 14695: {'lr': 0.000418555457647594, 'samples': 2821632, 'steps': 14695, 'loss/train': 1.4564698040485382} 01/29/2022 07:42:41 - INFO - codeparrot_training - Step 14696: {'lr': 0.00041854337313585913, 'samples': 2821824, 'steps': 14696, 'loss/train': 1.6861361861228943} 01/29/2022 07:42:45 - INFO - codeparrot_training - Step 14697: {'lr': 0.00041853128790213804, 'samples': 2822016, 'steps': 14697, 'loss/train': 1.6375452876091003} 01/29/2022 07:42:52 - INFO - codeparrot_training - Step 14698: {'lr': 0.0004185192019464823, 'samples': 2822208, 'steps': 14698, 'loss/train': 1.6235507726669312} 01/29/2022 07:42:56 - INFO - codeparrot_training - Step 14699: {'lr': 0.0004185071152689439, 'samples': 2822400, 'steps': 14699, 'loss/train': 1.3871772587299347} 01/29/2022 07:43:01 - INFO - codeparrot_training - Step 14700: {'lr': 0.0004184950278695745, 'samples': 2822592, 'steps': 14700, 'loss/train': 0.1820424422621727} 01/29/2022 07:43:05 - INFO - codeparrot_training - Step 14701: {'lr': 0.0004184829397484259, 'samples': 2822784, 'steps': 14701, 'loss/train': 0.7987244725227356} 01/29/2022 07:43:09 - INFO - codeparrot_training - Step 14702: {'lr': 0.00041847085090554985, 'samples': 2822976, 'steps': 14702, 'loss/train': 1.601298451423645} 01/29/2022 07:43:15 - INFO - codeparrot_training - Step 14703: {'lr': 0.00041845876134099825, 'samples': 2823168, 'steps': 14703, 'loss/train': 1.6572638154029846} 01/29/2022 07:43:19 - INFO - codeparrot_training - Step 14704: {'lr': 0.0004184466710548227, 'samples': 2823360, 'steps': 14704, 'loss/train': 1.5541555881500244} 01/29/2022 07:43:23 - INFO - codeparrot_training - Step 14705: {'lr': 0.0004184345800470752, 'samples': 2823552, 'steps': 14705, 'loss/train': 1.1823585033416748} 01/29/2022 07:43:28 - INFO - codeparrot_training - Step 14706: {'lr': 0.00041842248831780736, 'samples': 2823744, 'steps': 14706, 'loss/train': 1.6729926466941833} 01/29/2022 07:43:32 - INFO - codeparrot_training - Step 14707: {'lr': 0.0004184103958670712, 'samples': 2823936, 'steps': 14707, 'loss/train': 1.3266091346740723} 01/29/2022 07:43:37 - INFO - codeparrot_training - Step 14708: {'lr': 0.00041839830269491823, 'samples': 2824128, 'steps': 14708, 'loss/train': 1.0259166061878204} 01/29/2022 07:43:42 - INFO - codeparrot_training - Step 14709: {'lr': 0.00041838620880140046, 'samples': 2824320, 'steps': 14709, 'loss/train': 2.1620482206344604} 01/29/2022 07:43:46 - INFO - codeparrot_training - Step 14710: {'lr': 0.00041837411418656965, 'samples': 2824512, 'steps': 14710, 'loss/train': 1.0678355991840363} 01/29/2022 07:43:50 - INFO - codeparrot_training - Step 14711: {'lr': 0.0004183620188504776, 'samples': 2824704, 'steps': 14711, 'loss/train': 1.2722895741462708} 01/29/2022 07:43:55 - INFO - codeparrot_training - Step 14712: {'lr': 0.0004183499227931761, 'samples': 2824896, 'steps': 14712, 'loss/train': 0.9293153285980225} 01/29/2022 07:44:00 - INFO - codeparrot_training - Step 14713: {'lr': 0.00041833782601471704, 'samples': 2825088, 'steps': 14713, 'loss/train': 0.7300698012113571} 01/29/2022 07:44:04 - INFO - codeparrot_training - Step 14714: {'lr': 0.0004183257285151521, 'samples': 2825280, 'steps': 14714, 'loss/train': 1.5837487578392029} 01/29/2022 07:44:08 - INFO - codeparrot_training - Step 14715: {'lr': 0.00041831363029453327, 'samples': 2825472, 'steps': 14715, 'loss/train': 1.2859849333763123} 01/29/2022 07:44:12 - INFO - codeparrot_training - Step 14716: {'lr': 0.0004183015313529123, 'samples': 2825664, 'steps': 14716, 'loss/train': 1.5965432524681091} 01/29/2022 07:44:17 - INFO - codeparrot_training - Step 14717: {'lr': 0.00041828943169034094, 'samples': 2825856, 'steps': 14717, 'loss/train': 1.6724300980567932} 01/29/2022 07:44:23 - INFO - codeparrot_training - Step 14718: {'lr': 0.0004182773313068711, 'samples': 2826048, 'steps': 14718, 'loss/train': 1.1053301095962524} 01/29/2022 07:44:27 - INFO - codeparrot_training - Step 14719: {'lr': 0.00041826523020255463, 'samples': 2826240, 'steps': 14719, 'loss/train': 1.8172320127487183} 01/29/2022 07:44:32 - INFO - codeparrot_training - Step 14720: {'lr': 0.00041825312837744333, 'samples': 2826432, 'steps': 14720, 'loss/train': 1.908564269542694} 01/29/2022 07:44:36 - INFO - codeparrot_training - Step 14721: {'lr': 0.00041824102583158906, 'samples': 2826624, 'steps': 14721, 'loss/train': 2.0385260581970215} 01/29/2022 07:44:40 - INFO - codeparrot_training - Step 14722: {'lr': 0.0004182289225650437, 'samples': 2826816, 'steps': 14722, 'loss/train': 0.806147038936615} 01/29/2022 07:44:46 - INFO - codeparrot_training - Step 14723: {'lr': 0.00041821681857785904, 'samples': 2827008, 'steps': 14723, 'loss/train': 1.7630333304405212} 01/29/2022 07:44:50 - INFO - codeparrot_training - Step 14724: {'lr': 0.0004182047138700869, 'samples': 2827200, 'steps': 14724, 'loss/train': 2.038862407207489} 01/29/2022 07:44:54 - INFO - codeparrot_training - Step 14725: {'lr': 0.0004181926084417792, 'samples': 2827392, 'steps': 14725, 'loss/train': 1.0700731873512268} 01/29/2022 07:44:58 - INFO - codeparrot_training - Step 14726: {'lr': 0.0004181805022929878, 'samples': 2827584, 'steps': 14726, 'loss/train': 2.048722207546234} 01/29/2022 07:45:02 - INFO - codeparrot_training - Step 14727: {'lr': 0.0004181683954237645, 'samples': 2827776, 'steps': 14727, 'loss/train': 1.336771935224533} 01/29/2022 07:45:10 - INFO - codeparrot_training - Step 14728: {'lr': 0.00041815628783416117, 'samples': 2827968, 'steps': 14728, 'loss/train': 2.6362457871437073} 01/29/2022 07:45:14 - INFO - codeparrot_training - Step 14729: {'lr': 0.00041814417952422975, 'samples': 2828160, 'steps': 14729, 'loss/train': 1.4447035789489746} 01/29/2022 07:45:18 - INFO - codeparrot_training - Step 14730: {'lr': 0.000418132070494022, 'samples': 2828352, 'steps': 14730, 'loss/train': 1.5862573385238647} 01/29/2022 07:45:22 - INFO - codeparrot_training - Step 14731: {'lr': 0.00041811996074358993, 'samples': 2828544, 'steps': 14731, 'loss/train': 2.2655938267707825} 01/29/2022 07:45:27 - INFO - codeparrot_training - Step 14732: {'lr': 0.00041810785027298524, 'samples': 2828736, 'steps': 14732, 'loss/train': 1.358866810798645} 01/29/2022 07:45:32 - INFO - codeparrot_training - Step 14733: {'lr': 0.00041809573908225997, 'samples': 2828928, 'steps': 14733, 'loss/train': 1.8087208271026611} 01/29/2022 07:45:36 - INFO - codeparrot_training - Step 14734: {'lr': 0.00041808362717146594, 'samples': 2829120, 'steps': 14734, 'loss/train': 1.3954329192638397} 01/29/2022 07:45:41 - INFO - codeparrot_training - Step 14735: {'lr': 0.00041807151454065493, 'samples': 2829312, 'steps': 14735, 'loss/train': 2.238184154033661} 01/29/2022 07:45:45 - INFO - codeparrot_training - Step 14736: {'lr': 0.00041805940118987904, 'samples': 2829504, 'steps': 14736, 'loss/train': 2.132441461086273} 01/29/2022 07:45:49 - INFO - codeparrot_training - Step 14737: {'lr': 0.0004180472871191899, 'samples': 2829696, 'steps': 14737, 'loss/train': 1.6447902917861938} 01/29/2022 07:45:54 - INFO - codeparrot_training - Step 14738: {'lr': 0.0004180351723286396, 'samples': 2829888, 'steps': 14738, 'loss/train': 0.7094964534044266} 01/29/2022 07:45:58 - INFO - codeparrot_training - Step 14739: {'lr': 0.00041802305681828007, 'samples': 2830080, 'steps': 14739, 'loss/train': 0.9619274139404297} 01/29/2022 07:46:03 - INFO - codeparrot_training - Step 14740: {'lr': 0.00041801094058816304, 'samples': 2830272, 'steps': 14740, 'loss/train': 0.47173959016799927} 01/29/2022 07:46:07 - INFO - codeparrot_training - Step 14741: {'lr': 0.0004179988236383405, 'samples': 2830464, 'steps': 14741, 'loss/train': 1.601915180683136} 01/29/2022 07:46:11 - INFO - codeparrot_training - Step 14742: {'lr': 0.00041798670596886433, 'samples': 2830656, 'steps': 14742, 'loss/train': 1.459823191165924} 01/29/2022 07:46:19 - INFO - codeparrot_training - Step 14743: {'lr': 0.00041797458757978647, 'samples': 2830848, 'steps': 14743, 'loss/train': 2.0570987462997437} 01/29/2022 07:46:23 - INFO - codeparrot_training - Step 14744: {'lr': 0.0004179624684711588, 'samples': 2831040, 'steps': 14744, 'loss/train': 1.0176912546157837} 01/29/2022 07:46:28 - INFO - codeparrot_training - Step 14745: {'lr': 0.0004179503486430333, 'samples': 2831232, 'steps': 14745, 'loss/train': 1.5634276270866394} 01/29/2022 07:46:32 - INFO - codeparrot_training - Step 14746: {'lr': 0.00041793822809546176, 'samples': 2831424, 'steps': 14746, 'loss/train': 1.863638699054718} 01/29/2022 07:46:36 - INFO - codeparrot_training - Step 14747: {'lr': 0.0004179261068284963, 'samples': 2831616, 'steps': 14747, 'loss/train': 1.5127102732658386} 01/29/2022 07:46:40 - INFO - codeparrot_training - Step 14748: {'lr': 0.00041791398484218855, 'samples': 2831808, 'steps': 14748, 'loss/train': 1.8199957609176636} 01/29/2022 07:46:45 - INFO - codeparrot_training - Step 14749: {'lr': 0.0004179018621365908, 'samples': 2832000, 'steps': 14749, 'loss/train': 1.7811564803123474} 01/29/2022 07:46:50 - INFO - codeparrot_training - Step 14750: {'lr': 0.00041788973871175465, 'samples': 2832192, 'steps': 14750, 'loss/train': 1.6166333556175232} 01/29/2022 07:46:54 - INFO - codeparrot_training - Step 14751: {'lr': 0.00041787761456773214, 'samples': 2832384, 'steps': 14751, 'loss/train': 1.9107112884521484} 01/29/2022 07:46:58 - INFO - codeparrot_training - Step 14752: {'lr': 0.00041786548970457535, 'samples': 2832576, 'steps': 14752, 'loss/train': 1.7162248492240906} 01/29/2022 07:47:03 - INFO - codeparrot_training - Step 14753: {'lr': 0.000417853364122336, 'samples': 2832768, 'steps': 14753, 'loss/train': 1.576162576675415} 01/29/2022 07:47:07 - INFO - codeparrot_training - Step 14754: {'lr': 0.0004178412378210662, 'samples': 2832960, 'steps': 14754, 'loss/train': 2.500679612159729} 01/29/2022 07:47:14 - INFO - codeparrot_training - Step 14755: {'lr': 0.0004178291108008179, 'samples': 2833152, 'steps': 14755, 'loss/train': 2.0506873726844788} 01/29/2022 07:47:18 - INFO - codeparrot_training - Step 14756: {'lr': 0.00041781698306164283, 'samples': 2833344, 'steps': 14756, 'loss/train': 1.7090578079223633} 01/29/2022 07:47:22 - INFO - codeparrot_training - Step 14757: {'lr': 0.0004178048546035932, 'samples': 2833536, 'steps': 14757, 'loss/train': 1.3097282946109772} 01/29/2022 07:47:27 - INFO - codeparrot_training - Step 14758: {'lr': 0.00041779272542672086, 'samples': 2833728, 'steps': 14758, 'loss/train': 1.579956293106079} 01/29/2022 07:47:31 - INFO - codeparrot_training - Step 14759: {'lr': 0.00041778059553107766, 'samples': 2833920, 'steps': 14759, 'loss/train': 1.9343758821487427} 01/29/2022 07:47:37 - INFO - codeparrot_training - Step 14760: {'lr': 0.00041776846491671575, 'samples': 2834112, 'steps': 14760, 'loss/train': 1.671204686164856} 01/29/2022 07:47:41 - INFO - codeparrot_training - Step 14761: {'lr': 0.000417756333583687, 'samples': 2834304, 'steps': 14761, 'loss/train': 1.114368975162506} 01/29/2022 07:47:45 - INFO - codeparrot_training - Step 14762: {'lr': 0.0004177442015320434, 'samples': 2834496, 'steps': 14762, 'loss/train': 1.0609111189842224} 01/29/2022 07:47:49 - INFO - codeparrot_training - Step 14763: {'lr': 0.0004177320687618369, 'samples': 2834688, 'steps': 14763, 'loss/train': 3.3522716760635376} 01/29/2022 07:47:54 - INFO - codeparrot_training - Step 14764: {'lr': 0.0004177199352731194, 'samples': 2834880, 'steps': 14764, 'loss/train': 2.6006094217300415} 01/29/2022 07:47:59 - INFO - codeparrot_training - Step 14765: {'lr': 0.0004177078010659431, 'samples': 2835072, 'steps': 14765, 'loss/train': 1.6263624429702759} 01/29/2022 07:48:03 - INFO - codeparrot_training - Step 14766: {'lr': 0.0004176956661403597, 'samples': 2835264, 'steps': 14766, 'loss/train': 1.7673150300979614} 01/29/2022 07:48:07 - INFO - codeparrot_training - Step 14767: {'lr': 0.0004176835304964214, 'samples': 2835456, 'steps': 14767, 'loss/train': 1.3966688811779022} 01/29/2022 07:48:12 - INFO - codeparrot_training - Step 14768: {'lr': 0.00041767139413418, 'samples': 2835648, 'steps': 14768, 'loss/train': 1.9985541701316833} 01/29/2022 07:48:16 - INFO - codeparrot_training - Step 14769: {'lr': 0.00041765925705368766, 'samples': 2835840, 'steps': 14769, 'loss/train': 1.7398134469985962} 01/29/2022 07:48:21 - INFO - codeparrot_training - Step 14770: {'lr': 0.00041764711925499633, 'samples': 2836032, 'steps': 14770, 'loss/train': 1.7964758276939392} 01/29/2022 07:48:26 - INFO - codeparrot_training - Step 14771: {'lr': 0.0004176349807381579, 'samples': 2836224, 'steps': 14771, 'loss/train': 0.7279864400625229} 01/29/2022 07:48:30 - INFO - codeparrot_training - Step 14772: {'lr': 0.0004176228415032245, 'samples': 2836416, 'steps': 14772, 'loss/train': 2.2362005710601807} 01/29/2022 07:48:34 - INFO - codeparrot_training - Step 14773: {'lr': 0.000417610701550248, 'samples': 2836608, 'steps': 14773, 'loss/train': 1.8463668823242188} 01/29/2022 07:48:38 - INFO - codeparrot_training - Step 14774: {'lr': 0.0004175985608792806, 'samples': 2836800, 'steps': 14774, 'loss/train': 1.6563343405723572} 01/29/2022 07:48:45 - INFO - codeparrot_training - Step 14775: {'lr': 0.00041758641949037414, 'samples': 2836992, 'steps': 14775, 'loss/train': 2.1203479170799255} 01/29/2022 07:48:50 - INFO - codeparrot_training - Step 14776: {'lr': 0.00041757427738358066, 'samples': 2837184, 'steps': 14776, 'loss/train': 2.3304664492607117} 01/29/2022 07:48:54 - INFO - codeparrot_training - Step 14777: {'lr': 0.00041756213455895215, 'samples': 2837376, 'steps': 14777, 'loss/train': 0.7210203409194946} 01/29/2022 07:48:58 - INFO - codeparrot_training - Step 14778: {'lr': 0.00041754999101654066, 'samples': 2837568, 'steps': 14778, 'loss/train': 2.006070077419281} 01/29/2022 07:49:02 - INFO - codeparrot_training - Step 14779: {'lr': 0.0004175378467563983, 'samples': 2837760, 'steps': 14779, 'loss/train': 2.251125454902649} 01/29/2022 07:49:08 - INFO - codeparrot_training - Step 14780: {'lr': 0.00041752570177857695, 'samples': 2837952, 'steps': 14780, 'loss/train': 1.1003661453723907} 01/29/2022 07:49:12 - INFO - codeparrot_training - Step 14781: {'lr': 0.0004175135560831287, 'samples': 2838144, 'steps': 14781, 'loss/train': 1.8192538619041443} 01/29/2022 07:49:16 - INFO - codeparrot_training - Step 14782: {'lr': 0.00041750140967010554, 'samples': 2838336, 'steps': 14782, 'loss/train': 2.870938003063202} 01/29/2022 07:49:20 - INFO - codeparrot_training - Step 14783: {'lr': 0.00041748926253955954, 'samples': 2838528, 'steps': 14783, 'loss/train': 1.3208567798137665} 01/29/2022 07:49:25 - INFO - codeparrot_training - Step 14784: {'lr': 0.0004174771146915427, 'samples': 2838720, 'steps': 14784, 'loss/train': 1.6993002891540527} 01/29/2022 07:49:30 - INFO - codeparrot_training - Step 14785: {'lr': 0.00041746496612610705, 'samples': 2838912, 'steps': 14785, 'loss/train': 1.5110995173454285} 01/29/2022 07:49:34 - INFO - codeparrot_training - Step 14786: {'lr': 0.00041745281684330476, 'samples': 2839104, 'steps': 14786, 'loss/train': 1.8616318702697754} 01/29/2022 07:49:39 - INFO - codeparrot_training - Step 14787: {'lr': 0.0004174406668431877, 'samples': 2839296, 'steps': 14787, 'loss/train': 2.548315644264221} 01/29/2022 07:49:43 - INFO - codeparrot_training - Step 14788: {'lr': 0.000417428516125808, 'samples': 2839488, 'steps': 14788, 'loss/train': 1.8561106324195862} 01/29/2022 07:49:47 - INFO - codeparrot_training - Step 14789: {'lr': 0.0004174163646912178, 'samples': 2839680, 'steps': 14789, 'loss/train': 1.7047539353370667} 01/29/2022 07:49:54 - INFO - codeparrot_training - Step 14790: {'lr': 0.0004174042125394689, 'samples': 2839872, 'steps': 14790, 'loss/train': 1.8985549807548523} 01/29/2022 07:49:58 - INFO - codeparrot_training - Step 14791: {'lr': 0.00041739205967061366, 'samples': 2840064, 'steps': 14791, 'loss/train': 1.339759111404419} 01/29/2022 07:50:02 - INFO - codeparrot_training - Step 14792: {'lr': 0.0004173799060847039, 'samples': 2840256, 'steps': 14792, 'loss/train': 2.180040121078491} 01/29/2022 07:50:07 - INFO - codeparrot_training - Step 14793: {'lr': 0.00041736775178179174, 'samples': 2840448, 'steps': 14793, 'loss/train': 2.086351454257965} 01/29/2022 07:50:11 - INFO - codeparrot_training - Step 14794: {'lr': 0.0004173555967619294, 'samples': 2840640, 'steps': 14794, 'loss/train': 1.4851714968681335} 01/29/2022 07:50:17 - INFO - codeparrot_training - Step 14795: {'lr': 0.00041734344102516873, 'samples': 2840832, 'steps': 14795, 'loss/train': 0.8326549530029297} 01/29/2022 07:50:21 - INFO - codeparrot_training - Step 14796: {'lr': 0.0004173312845715619, 'samples': 2841024, 'steps': 14796, 'loss/train': 1.8279476165771484} 01/29/2022 07:50:25 - INFO - codeparrot_training - Step 14797: {'lr': 0.000417319127401161, 'samples': 2841216, 'steps': 14797, 'loss/train': 1.796562910079956} 01/29/2022 07:50:29 - INFO - codeparrot_training - Step 14798: {'lr': 0.00041730696951401816, 'samples': 2841408, 'steps': 14798, 'loss/train': 1.8053303360939026} 01/29/2022 07:50:34 - INFO - codeparrot_training - Step 14799: {'lr': 0.00041729481091018527, 'samples': 2841600, 'steps': 14799, 'loss/train': 1.726984977722168} 01/29/2022 07:50:40 - INFO - codeparrot_training - Step 14800: {'lr': 0.0004172826515897146, 'samples': 2841792, 'steps': 14800, 'loss/train': 1.4042425453662872} 01/29/2022 07:50:45 - INFO - codeparrot_training - Step 14801: {'lr': 0.0004172704915526581, 'samples': 2841984, 'steps': 14801, 'loss/train': 1.4817472100257874} 01/29/2022 07:50:49 - INFO - codeparrot_training - Step 14802: {'lr': 0.000417258330799068, 'samples': 2842176, 'steps': 14802, 'loss/train': 1.6901110410690308} 01/29/2022 07:50:53 - INFO - codeparrot_training - Step 14803: {'lr': 0.00041724616932899627, 'samples': 2842368, 'steps': 14803, 'loss/train': 1.4317803382873535} 01/29/2022 07:50:57 - INFO - codeparrot_training - Step 14804: {'lr': 0.0004172340071424951, 'samples': 2842560, 'steps': 14804, 'loss/train': 1.4258102774620056} 01/29/2022 07:51:03 - INFO - codeparrot_training - Step 14805: {'lr': 0.0004172218442396165, 'samples': 2842752, 'steps': 14805, 'loss/train': 1.98375803232193} 01/29/2022 07:51:07 - INFO - codeparrot_training - Step 14806: {'lr': 0.00041720968062041266, 'samples': 2842944, 'steps': 14806, 'loss/train': 1.5688258409500122} 01/29/2022 07:51:11 - INFO - codeparrot_training - Step 14807: {'lr': 0.0004171975162849356, 'samples': 2843136, 'steps': 14807, 'loss/train': 1.9588977098464966} 01/29/2022 07:51:16 - INFO - codeparrot_training - Step 14808: {'lr': 0.0004171853512332375, 'samples': 2843328, 'steps': 14808, 'loss/train': 0.3006189689040184} 01/29/2022 07:51:20 - INFO - codeparrot_training - Step 14809: {'lr': 0.00041717318546537045, 'samples': 2843520, 'steps': 14809, 'loss/train': 1.2891583442687988} 01/29/2022 07:51:25 - INFO - codeparrot_training - Step 14810: {'lr': 0.0004171610189813866, 'samples': 2843712, 'steps': 14810, 'loss/train': 0.24274975061416626} 01/29/2022 07:51:30 - INFO - codeparrot_training - Step 14811: {'lr': 0.000417148851781338, 'samples': 2843904, 'steps': 14811, 'loss/train': 1.6081420183181763} 01/29/2022 07:51:34 - INFO - codeparrot_training - Step 14812: {'lr': 0.0004171366838652767, 'samples': 2844096, 'steps': 14812, 'loss/train': 1.2327404916286469} 01/29/2022 07:51:38 - INFO - codeparrot_training - Step 14813: {'lr': 0.000417124515233255, 'samples': 2844288, 'steps': 14813, 'loss/train': 2.92358136177063} 01/29/2022 07:51:42 - INFO - codeparrot_training - Step 14814: {'lr': 0.00041711234588532497, 'samples': 2844480, 'steps': 14814, 'loss/train': 0.76579749584198} 01/29/2022 07:51:50 - INFO - codeparrot_training - Step 14815: {'lr': 0.0004171001758215387, 'samples': 2844672, 'steps': 14815, 'loss/train': 2.192947804927826} 01/29/2022 07:51:54 - INFO - codeparrot_training - Step 14816: {'lr': 0.0004170880050419483, 'samples': 2844864, 'steps': 14816, 'loss/train': 1.1268714666366577} 01/29/2022 07:51:58 - INFO - codeparrot_training - Step 14817: {'lr': 0.00041707583354660597, 'samples': 2845056, 'steps': 14817, 'loss/train': 1.1897330582141876} 01/29/2022 07:52:02 - INFO - codeparrot_training - Step 14818: {'lr': 0.0004170636613355638, 'samples': 2845248, 'steps': 14818, 'loss/train': 2.27647465467453} 01/29/2022 07:52:07 - INFO - codeparrot_training - Step 14819: {'lr': 0.000417051488408874, 'samples': 2845440, 'steps': 14819, 'loss/train': 1.4263178408145905} 01/29/2022 07:52:12 - INFO - codeparrot_training - Step 14820: {'lr': 0.00041703931476658857, 'samples': 2845632, 'steps': 14820, 'loss/train': 1.5707974433898926} 01/29/2022 07:52:16 - INFO - codeparrot_training - Step 14821: {'lr': 0.0004170271404087598, 'samples': 2845824, 'steps': 14821, 'loss/train': 2.286352515220642} 01/29/2022 07:52:20 - INFO - codeparrot_training - Step 14822: {'lr': 0.0004170149653354398, 'samples': 2846016, 'steps': 14822, 'loss/train': 2.1002864241600037} 01/29/2022 07:52:25 - INFO - codeparrot_training - Step 14823: {'lr': 0.0004170027895466807, 'samples': 2846208, 'steps': 14823, 'loss/train': 1.480638027191162} 01/29/2022 07:52:29 - INFO - codeparrot_training - Step 14824: {'lr': 0.00041699061304253476, 'samples': 2846400, 'steps': 14824, 'loss/train': 1.5941689610481262} 01/29/2022 07:52:34 - INFO - codeparrot_training - Step 14825: {'lr': 0.00041697843582305406, 'samples': 2846592, 'steps': 14825, 'loss/train': 1.9341180324554443} 01/29/2022 07:52:39 - INFO - codeparrot_training - Step 14826: {'lr': 0.0004169662578882907, 'samples': 2846784, 'steps': 14826, 'loss/train': 1.7857342958450317} 01/29/2022 07:52:43 - INFO - codeparrot_training - Step 14827: {'lr': 0.0004169540792382969, 'samples': 2846976, 'steps': 14827, 'loss/train': 0.9316413402557373} 01/29/2022 07:52:47 - INFO - codeparrot_training - Step 14828: {'lr': 0.0004169418998731249, 'samples': 2847168, 'steps': 14828, 'loss/train': 1.6560800671577454} 01/29/2022 07:52:51 - INFO - codeparrot_training - Step 14829: {'lr': 0.0004169297197928268, 'samples': 2847360, 'steps': 14829, 'loss/train': 1.5038017630577087} 01/29/2022 07:52:57 - INFO - codeparrot_training - Step 14830: {'lr': 0.0004169175389974548, 'samples': 2847552, 'steps': 14830, 'loss/train': 1.6532251238822937} 01/29/2022 07:53:01 - INFO - codeparrot_training - Step 14831: {'lr': 0.0004169053574870609, 'samples': 2847744, 'steps': 14831, 'loss/train': 1.6681841611862183} 01/29/2022 07:53:05 - INFO - codeparrot_training - Step 14832: {'lr': 0.0004168931752616977, 'samples': 2847936, 'steps': 14832, 'loss/train': 1.6803204417228699} 01/29/2022 07:53:10 - INFO - codeparrot_training - Step 14833: {'lr': 0.00041688099232141694, 'samples': 2848128, 'steps': 14833, 'loss/train': 1.0744663774967194} 01/29/2022 07:53:14 - INFO - codeparrot_training - Step 14834: {'lr': 0.0004168688086662711, 'samples': 2848320, 'steps': 14834, 'loss/train': 1.8327096104621887} 01/29/2022 07:53:21 - INFO - codeparrot_training - Step 14835: {'lr': 0.0004168566242963122, 'samples': 2848512, 'steps': 14835, 'loss/train': 0.9760490655899048} 01/29/2022 07:53:25 - INFO - codeparrot_training - Step 14836: {'lr': 0.00041684443921159253, 'samples': 2848704, 'steps': 14836, 'loss/train': 7.319260597229004} 01/29/2022 07:53:30 - INFO - codeparrot_training - Step 14837: {'lr': 0.00041683225341216426, 'samples': 2848896, 'steps': 14837, 'loss/train': 2.0358227491378784} 01/29/2022 07:53:34 - INFO - codeparrot_training - Step 14838: {'lr': 0.0004168200668980796, 'samples': 2849088, 'steps': 14838, 'loss/train': 2.266633629798889} 01/29/2022 07:53:38 - INFO - codeparrot_training - Step 14839: {'lr': 0.0004168078796693908, 'samples': 2849280, 'steps': 14839, 'loss/train': 2.0545740723609924} 01/29/2022 07:53:44 - INFO - codeparrot_training - Step 14840: {'lr': 0.00041679569172614996, 'samples': 2849472, 'steps': 14840, 'loss/train': 1.5501911044120789} 01/29/2022 07:53:48 - INFO - codeparrot_training - Step 14841: {'lr': 0.0004167835030684093, 'samples': 2849664, 'steps': 14841, 'loss/train': 1.5721212029457092} 01/29/2022 07:53:52 - INFO - codeparrot_training - Step 14842: {'lr': 0.0004167713136962211, 'samples': 2849856, 'steps': 14842, 'loss/train': 1.812760591506958} 01/29/2022 07:53:56 - INFO - codeparrot_training - Step 14843: {'lr': 0.00041675912360963766, 'samples': 2850048, 'steps': 14843, 'loss/train': 1.8590962886810303} 01/29/2022 07:54:00 - INFO - codeparrot_training - Step 14844: {'lr': 0.0004167469328087109, 'samples': 2850240, 'steps': 14844, 'loss/train': 1.6664648652076721} 01/29/2022 07:54:07 - INFO - codeparrot_training - Step 14845: {'lr': 0.0004167347412934933, 'samples': 2850432, 'steps': 14845, 'loss/train': 1.948605716228485} 01/29/2022 07:54:12 - INFO - codeparrot_training - Step 14846: {'lr': 0.00041672254906403703, 'samples': 2850624, 'steps': 14846, 'loss/train': 1.4612247347831726} 01/29/2022 07:54:16 - INFO - codeparrot_training - Step 14847: {'lr': 0.00041671035612039434, 'samples': 2850816, 'steps': 14847, 'loss/train': 1.3501315712928772} 01/29/2022 07:54:20 - INFO - codeparrot_training - Step 14848: {'lr': 0.0004166981624626174, 'samples': 2851008, 'steps': 14848, 'loss/train': 1.960882544517517} 01/29/2022 07:54:25 - INFO - codeparrot_training - Step 14849: {'lr': 0.00041668596809075835, 'samples': 2851200, 'steps': 14849, 'loss/train': 0.41799795627593994} 01/29/2022 07:54:30 - INFO - codeparrot_training - Step 14850: {'lr': 0.0004166737730048697, 'samples': 2851392, 'steps': 14850, 'loss/train': 1.9465314745903015} 01/29/2022 07:54:34 - INFO - codeparrot_training - Step 14851: {'lr': 0.00041666157720500344, 'samples': 2851584, 'steps': 14851, 'loss/train': 1.8584636449813843} 01/29/2022 07:54:38 - INFO - codeparrot_training - Step 14852: {'lr': 0.00041664938069121195, 'samples': 2851776, 'steps': 14852, 'loss/train': 2.025085151195526} 01/29/2022 07:54:43 - INFO - codeparrot_training - Step 14853: {'lr': 0.0004166371834635474, 'samples': 2851968, 'steps': 14853, 'loss/train': 0.7742372453212738} 01/29/2022 07:54:47 - INFO - codeparrot_training - Step 14854: {'lr': 0.00041662498552206206, 'samples': 2852160, 'steps': 14854, 'loss/train': 2.387975335121155} 01/29/2022 07:54:52 - INFO - codeparrot_training - Step 14855: {'lr': 0.00041661278686680827, 'samples': 2852352, 'steps': 14855, 'loss/train': 1.511150300502777} 01/29/2022 07:54:57 - INFO - codeparrot_training - Step 14856: {'lr': 0.00041660058749783813, 'samples': 2852544, 'steps': 14856, 'loss/train': 1.5350212454795837} 01/29/2022 07:55:01 - INFO - codeparrot_training - Step 14857: {'lr': 0.000416588387415204, 'samples': 2852736, 'steps': 14857, 'loss/train': 0.8023554682731628} 01/29/2022 07:55:05 - INFO - codeparrot_training - Step 14858: {'lr': 0.0004165761866189581, 'samples': 2852928, 'steps': 14858, 'loss/train': 0.5165295302867889} 01/29/2022 07:55:09 - INFO - codeparrot_training - Step 14859: {'lr': 0.00041656398510915273, 'samples': 2853120, 'steps': 14859, 'loss/train': 2.2441526055336} 01/29/2022 07:55:16 - INFO - codeparrot_training - Step 14860: {'lr': 0.00041655178288584006, 'samples': 2853312, 'steps': 14860, 'loss/train': 2.410227656364441} 01/29/2022 07:55:21 - INFO - codeparrot_training - Step 14861: {'lr': 0.00041653957994907255, 'samples': 2853504, 'steps': 14861, 'loss/train': 2.1271519660949707} 01/29/2022 07:55:25 - INFO - codeparrot_training - Step 14862: {'lr': 0.0004165273762989023, 'samples': 2853696, 'steps': 14862, 'loss/train': 1.5011100769042969} 01/29/2022 07:55:29 - INFO - codeparrot_training - Step 14863: {'lr': 0.0004165151719353817, 'samples': 2853888, 'steps': 14863, 'loss/train': 1.9992613792419434} 01/29/2022 07:55:33 - INFO - codeparrot_training - Step 14864: {'lr': 0.0004165029668585629, 'samples': 2854080, 'steps': 14864, 'loss/train': 2.718128263950348} 01/29/2022 07:55:39 - INFO - codeparrot_training - Step 14865: {'lr': 0.00041649076106849836, 'samples': 2854272, 'steps': 14865, 'loss/train': 2.5828335285186768} 01/29/2022 07:55:43 - INFO - codeparrot_training - Step 14866: {'lr': 0.0004164785545652402, 'samples': 2854464, 'steps': 14866, 'loss/train': 1.608683466911316} 01/29/2022 07:55:47 - INFO - codeparrot_training - Step 14867: {'lr': 0.0004164663473488408, 'samples': 2854656, 'steps': 14867, 'loss/train': 1.0036963820457458} 01/29/2022 07:55:51 - INFO - codeparrot_training - Step 14868: {'lr': 0.0004164541394193524, 'samples': 2854848, 'steps': 14868, 'loss/train': 1.5187286138534546} 01/29/2022 07:55:56 - INFO - codeparrot_training - Step 14869: {'lr': 0.00041644193077682734, 'samples': 2855040, 'steps': 14869, 'loss/train': 1.6477369666099548} 01/29/2022 07:56:01 - INFO - codeparrot_training - Step 14870: {'lr': 0.0004164297214213179, 'samples': 2855232, 'steps': 14870, 'loss/train': 1.8005071878433228} 01/29/2022 07:56:05 - INFO - codeparrot_training - Step 14871: {'lr': 0.0004164175113528763, 'samples': 2855424, 'steps': 14871, 'loss/train': 1.6190716624259949} 01/29/2022 07:56:09 - INFO - codeparrot_training - Step 14872: {'lr': 0.000416405300571555, 'samples': 2855616, 'steps': 14872, 'loss/train': 1.3655348718166351} 01/29/2022 07:56:14 - INFO - codeparrot_training - Step 14873: {'lr': 0.00041639308907740624, 'samples': 2855808, 'steps': 14873, 'loss/train': 1.9876880049705505} 01/29/2022 07:56:18 - INFO - codeparrot_training - Step 14874: {'lr': 0.0004163808768704823, 'samples': 2856000, 'steps': 14874, 'loss/train': 1.958113968372345} 01/29/2022 07:56:25 - INFO - codeparrot_training - Step 14875: {'lr': 0.0004163686639508356, 'samples': 2856192, 'steps': 14875, 'loss/train': 1.3001383244991302} 01/29/2022 07:56:29 - INFO - codeparrot_training - Step 14876: {'lr': 0.00041635645031851826, 'samples': 2856384, 'steps': 14876, 'loss/train': 2.6113228797912598} 01/29/2022 07:56:34 - INFO - codeparrot_training - Step 14877: {'lr': 0.0004163442359735827, 'samples': 2856576, 'steps': 14877, 'loss/train': 1.2086198329925537} 01/29/2022 07:56:38 - INFO - codeparrot_training - Step 14878: {'lr': 0.00041633202091608136, 'samples': 2856768, 'steps': 14878, 'loss/train': 1.8418230414390564} 01/29/2022 07:56:42 - INFO - codeparrot_training - Step 14879: {'lr': 0.00041631980514606636, 'samples': 2856960, 'steps': 14879, 'loss/train': 1.63366037607193} 01/29/2022 07:56:47 - INFO - codeparrot_training - Step 14880: {'lr': 0.0004163075886635902, 'samples': 2857152, 'steps': 14880, 'loss/train': 0.8683609664440155} 01/29/2022 07:56:52 - INFO - codeparrot_training - Step 14881: {'lr': 0.0004162953714687051, 'samples': 2857344, 'steps': 14881, 'loss/train': 0.8577122390270233} 01/29/2022 07:56:56 - INFO - codeparrot_training - Step 14882: {'lr': 0.0004162831535614635, 'samples': 2857536, 'steps': 14882, 'loss/train': 1.7579167485237122} 01/29/2022 07:57:00 - INFO - codeparrot_training - Step 14883: {'lr': 0.0004162709349419176, 'samples': 2857728, 'steps': 14883, 'loss/train': 1.0761348009109497} 01/29/2022 07:57:05 - INFO - codeparrot_training - Step 14884: {'lr': 0.0004162587156101198, 'samples': 2857920, 'steps': 14884, 'loss/train': 1.7477563619613647} 01/29/2022 07:57:10 - INFO - codeparrot_training - Step 14885: {'lr': 0.0004162464955661225, 'samples': 2858112, 'steps': 14885, 'loss/train': 1.085214525461197} 01/29/2022 07:57:14 - INFO - codeparrot_training - Step 14886: {'lr': 0.000416234274809978, 'samples': 2858304, 'steps': 14886, 'loss/train': 1.2939151525497437} 01/29/2022 07:57:18 - INFO - codeparrot_training - Step 14887: {'lr': 0.00041622205334173863, 'samples': 2858496, 'steps': 14887, 'loss/train': 2.072055995464325} 01/29/2022 07:57:22 - INFO - codeparrot_training - Step 14888: {'lr': 0.00041620983116145673, 'samples': 2858688, 'steps': 14888, 'loss/train': 1.5099883675575256} 01/29/2022 07:57:30 - INFO - codeparrot_training - Step 14889: {'lr': 0.00041619760826918474, 'samples': 2858880, 'steps': 14889, 'loss/train': 1.4549678564071655} 01/29/2022 07:57:34 - INFO - codeparrot_training - Step 14890: {'lr': 0.00041618538466497496, 'samples': 2859072, 'steps': 14890, 'loss/train': 2.192798674106598} 01/29/2022 07:57:38 - INFO - codeparrot_training - Step 14891: {'lr': 0.00041617316034887983, 'samples': 2859264, 'steps': 14891, 'loss/train': 1.4382958710193634} 01/29/2022 07:57:42 - INFO - codeparrot_training - Step 14892: {'lr': 0.00041616093532095155, 'samples': 2859456, 'steps': 14892, 'loss/train': 1.6172057390213013} 01/29/2022 07:57:47 - INFO - codeparrot_training - Step 14893: {'lr': 0.00041614870958124264, 'samples': 2859648, 'steps': 14893, 'loss/train': 1.8731127977371216} 01/29/2022 07:57:51 - INFO - codeparrot_training - Step 14894: {'lr': 0.00041613648312980537, 'samples': 2859840, 'steps': 14894, 'loss/train': 1.687482476234436} 01/29/2022 07:57:56 - INFO - codeparrot_training - Step 14895: {'lr': 0.00041612425596669215, 'samples': 2860032, 'steps': 14895, 'loss/train': 0.8422976732254028} 01/29/2022 07:58:00 - INFO - codeparrot_training - Step 14896: {'lr': 0.0004161120280919554, 'samples': 2860224, 'steps': 14896, 'loss/train': 1.3270120024681091} 01/29/2022 07:58:05 - INFO - codeparrot_training - Step 14897: {'lr': 0.00041609979950564747, 'samples': 2860416, 'steps': 14897, 'loss/train': 1.33772474527359} 01/29/2022 07:58:09 - INFO - codeparrot_training - Step 14898: {'lr': 0.00041608757020782073, 'samples': 2860608, 'steps': 14898, 'loss/train': 0.479444682598114} 01/29/2022 07:58:13 - INFO - codeparrot_training - Step 14899: {'lr': 0.0004160753401985276, 'samples': 2860800, 'steps': 14899, 'loss/train': 1.2268057465553284} 01/29/2022 07:58:19 - INFO - codeparrot_training - Step 14900: {'lr': 0.00041606310947782046, 'samples': 2860992, 'steps': 14900, 'loss/train': 1.6575393676757812} 01/29/2022 07:58:23 - INFO - codeparrot_training - Step 14901: {'lr': 0.00041605087804575167, 'samples': 2861184, 'steps': 14901, 'loss/train': 1.4198284149169922} 01/29/2022 07:58:27 - INFO - codeparrot_training - Step 14902: {'lr': 0.0004160386459023736, 'samples': 2861376, 'steps': 14902, 'loss/train': 2.662415564060211} 01/29/2022 07:58:32 - INFO - codeparrot_training - Step 14903: {'lr': 0.00041602641304773876, 'samples': 2861568, 'steps': 14903, 'loss/train': 1.2079688608646393} 01/29/2022 07:58:39 - INFO - codeparrot_training - Step 14904: {'lr': 0.0004160141794818995, 'samples': 2861760, 'steps': 14904, 'loss/train': 1.8073493242263794} 01/29/2022 07:58:43 - INFO - codeparrot_training - Step 14905: {'lr': 0.00041600194520490815, 'samples': 2861952, 'steps': 14905, 'loss/train': 1.4401820003986359} 01/29/2022 07:58:47 - INFO - codeparrot_training - Step 14906: {'lr': 0.0004159897102168172, 'samples': 2862144, 'steps': 14906, 'loss/train': 1.9710479378700256} 01/29/2022 07:58:51 - INFO - codeparrot_training - Step 14907: {'lr': 0.00041597747451767905, 'samples': 2862336, 'steps': 14907, 'loss/train': 2.0474583506584167} 01/29/2022 07:58:56 - INFO - codeparrot_training - Step 14908: {'lr': 0.00041596523810754607, 'samples': 2862528, 'steps': 14908, 'loss/train': 0.5999112725257874} 01/29/2022 07:59:01 - INFO - codeparrot_training - Step 14909: {'lr': 0.0004159530009864707, 'samples': 2862720, 'steps': 14909, 'loss/train': 1.801630675792694} 01/29/2022 07:59:05 - INFO - codeparrot_training - Step 14910: {'lr': 0.0004159407631545054, 'samples': 2862912, 'steps': 14910, 'loss/train': 1.4088412821292877} 01/29/2022 07:59:09 - INFO - codeparrot_training - Step 14911: {'lr': 0.0004159285246117026, 'samples': 2863104, 'steps': 14911, 'loss/train': 0.9082679450511932} 01/29/2022 07:59:14 - INFO - codeparrot_training - Step 14912: {'lr': 0.00041591628535811464, 'samples': 2863296, 'steps': 14912, 'loss/train': 1.7773164510726929} 01/29/2022 07:59:18 - INFO - codeparrot_training - Step 14913: {'lr': 0.000415904045393794, 'samples': 2863488, 'steps': 14913, 'loss/train': 2.0590656995773315} 01/29/2022 07:59:23 - INFO - codeparrot_training - Step 14914: {'lr': 0.0004158918047187931, 'samples': 2863680, 'steps': 14914, 'loss/train': 1.8360686302185059} 01/29/2022 07:59:28 - INFO - codeparrot_training - Step 14915: {'lr': 0.0004158795633331645, 'samples': 2863872, 'steps': 14915, 'loss/train': 2.096566379070282} 01/29/2022 07:59:32 - INFO - codeparrot_training - Step 14916: {'lr': 0.00041586732123696037, 'samples': 2864064, 'steps': 14916, 'loss/train': 0.6402906328439713} 01/29/2022 07:59:36 - INFO - codeparrot_training - Step 14917: {'lr': 0.0004158550784302334, 'samples': 2864256, 'steps': 14917, 'loss/train': 2.0901641249656677} 01/29/2022 07:59:40 - INFO - codeparrot_training - Step 14918: {'lr': 0.0004158428349130359, 'samples': 2864448, 'steps': 14918, 'loss/train': 2.1740745306015015} 01/29/2022 07:59:47 - INFO - codeparrot_training - Step 14919: {'lr': 0.00041583059068542034, 'samples': 2864640, 'steps': 14919, 'loss/train': 1.9350594878196716} 01/29/2022 07:59:52 - INFO - codeparrot_training - Step 14920: {'lr': 0.0004158183457474392, 'samples': 2864832, 'steps': 14920, 'loss/train': 1.7555705308914185} 01/29/2022 07:59:56 - INFO - codeparrot_training - Step 14921: {'lr': 0.00041580610009914486, 'samples': 2865024, 'steps': 14921, 'loss/train': 1.8407208323478699} 01/29/2022 08:00:00 - INFO - codeparrot_training - Step 14922: {'lr': 0.00041579385374058996, 'samples': 2865216, 'steps': 14922, 'loss/train': 1.3112214803695679} 01/29/2022 08:00:04 - INFO - codeparrot_training - Step 14923: {'lr': 0.00041578160667182676, 'samples': 2865408, 'steps': 14923, 'loss/train': 1.3217875957489014} 01/29/2022 08:00:10 - INFO - codeparrot_training - Step 14924: {'lr': 0.00041576935889290777, 'samples': 2865600, 'steps': 14924, 'loss/train': 2.1585150361061096} 01/29/2022 08:00:14 - INFO - codeparrot_training - Step 14925: {'lr': 0.0004157571104038856, 'samples': 2865792, 'steps': 14925, 'loss/train': 2.0371259450912476} 01/29/2022 08:00:18 - INFO - codeparrot_training - Step 14926: {'lr': 0.00041574486120481255, 'samples': 2865984, 'steps': 14926, 'loss/train': 1.8123366236686707} 01/29/2022 08:00:23 - INFO - codeparrot_training - Step 14927: {'lr': 0.0004157326112957411, 'samples': 2866176, 'steps': 14927, 'loss/train': 2.1809415221214294} 01/29/2022 08:00:27 - INFO - codeparrot_training - Step 14928: {'lr': 0.0004157203606767238, 'samples': 2866368, 'steps': 14928, 'loss/train': 2.2045156359672546} 01/29/2022 08:00:34 - INFO - codeparrot_training - Step 14929: {'lr': 0.0004157081093478131, 'samples': 2866560, 'steps': 14929, 'loss/train': 0.7402604520320892} 01/29/2022 08:00:38 - INFO - codeparrot_training - Step 14930: {'lr': 0.00041569585730906147, 'samples': 2866752, 'steps': 14930, 'loss/train': 1.277196228504181} 01/29/2022 08:00:43 - INFO - codeparrot_training - Step 14931: {'lr': 0.0004156836045605214, 'samples': 2866944, 'steps': 14931, 'loss/train': 1.6429105997085571} 01/29/2022 08:00:47 - INFO - codeparrot_training - Step 14932: {'lr': 0.0004156713511022454, 'samples': 2867136, 'steps': 14932, 'loss/train': 0.7030593752861023} 01/29/2022 08:00:51 - INFO - codeparrot_training - Step 14933: {'lr': 0.00041565909693428593, 'samples': 2867328, 'steps': 14933, 'loss/train': 1.790054976940155} 01/29/2022 08:00:56 - INFO - codeparrot_training - Step 14934: {'lr': 0.00041564684205669546, 'samples': 2867520, 'steps': 14934, 'loss/train': 1.5456233024597168} 01/29/2022 08:01:01 - INFO - codeparrot_training - Step 14935: {'lr': 0.00041563458646952655, 'samples': 2867712, 'steps': 14935, 'loss/train': 1.4700077176094055} 01/29/2022 08:01:05 - INFO - codeparrot_training - Step 14936: {'lr': 0.0004156223301728316, 'samples': 2867904, 'steps': 14936, 'loss/train': 1.4032713174819946} 01/29/2022 08:01:09 - INFO - codeparrot_training - Step 14937: {'lr': 0.00041561007316666333, 'samples': 2868096, 'steps': 14937, 'loss/train': 2.011631190776825} 01/29/2022 08:01:13 - INFO - codeparrot_training - Step 14938: {'lr': 0.00041559781545107393, 'samples': 2868288, 'steps': 14938, 'loss/train': 1.642297625541687} 01/29/2022 08:01:19 - INFO - codeparrot_training - Step 14939: {'lr': 0.00041558555702611615, 'samples': 2868480, 'steps': 14939, 'loss/train': 0.8575319051742554} 01/29/2022 08:01:23 - INFO - codeparrot_training - Step 14940: {'lr': 0.0004155732978918424, 'samples': 2868672, 'steps': 14940, 'loss/train': 1.4190123081207275} 01/29/2022 08:01:27 - INFO - codeparrot_training - Step 14941: {'lr': 0.00041556103804830523, 'samples': 2868864, 'steps': 14941, 'loss/train': 1.9762113690376282} 01/29/2022 08:01:32 - INFO - codeparrot_training - Step 14942: {'lr': 0.0004155487774955572, 'samples': 2869056, 'steps': 14942, 'loss/train': 3.1588858366012573} 01/29/2022 08:01:36 - INFO - codeparrot_training - Step 14943: {'lr': 0.00041553651623365076, 'samples': 2869248, 'steps': 14943, 'loss/train': 2.0172051787376404} 01/29/2022 08:01:41 - INFO - codeparrot_training - Step 14944: {'lr': 0.00041552425426263836, 'samples': 2869440, 'steps': 14944, 'loss/train': 1.7314685583114624} 01/29/2022 08:01:45 - INFO - codeparrot_training - Step 14945: {'lr': 0.00041551199158257264, 'samples': 2869632, 'steps': 14945, 'loss/train': 1.9947289824485779} 01/29/2022 08:01:50 - INFO - codeparrot_training - Step 14946: {'lr': 0.00041549972819350615, 'samples': 2869824, 'steps': 14946, 'loss/train': 1.5389100909233093} 01/29/2022 08:01:54 - INFO - codeparrot_training - Step 14947: {'lr': 0.00041548746409549134, 'samples': 2870016, 'steps': 14947, 'loss/train': 2.2326385974884033} 01/29/2022 08:01:58 - INFO - codeparrot_training - Step 14948: {'lr': 0.0004154751992885808, 'samples': 2870208, 'steps': 14948, 'loss/train': 1.8559754490852356} 01/29/2022 08:02:05 - INFO - codeparrot_training - Step 14949: {'lr': 0.0004154629337728271, 'samples': 2870400, 'steps': 14949, 'loss/train': 1.3367234766483307} 01/29/2022 08:02:09 - INFO - codeparrot_training - Step 14950: {'lr': 0.00041545066754828264, 'samples': 2870592, 'steps': 14950, 'loss/train': 1.9013076424598694} 01/29/2022 08:02:14 - INFO - codeparrot_training - Step 14951: {'lr': 0.00041543840061500007, 'samples': 2870784, 'steps': 14951, 'loss/train': 1.5653942227363586} 01/29/2022 08:02:18 - INFO - codeparrot_training - Step 14952: {'lr': 0.000415426132973032, 'samples': 2870976, 'steps': 14952, 'loss/train': 1.9525832533836365} 01/29/2022 08:02:22 - INFO - codeparrot_training - Step 14953: {'lr': 0.0004154138646224308, 'samples': 2871168, 'steps': 14953, 'loss/train': 1.7787664532661438} 01/29/2022 08:02:28 - INFO - codeparrot_training - Step 14954: {'lr': 0.0004154015955632492, 'samples': 2871360, 'steps': 14954, 'loss/train': 0.6162500828504562} 01/29/2022 08:02:32 - INFO - codeparrot_training - Step 14955: {'lr': 0.0004153893257955397, 'samples': 2871552, 'steps': 14955, 'loss/train': 1.8206185698509216} 01/29/2022 08:02:36 - INFO - codeparrot_training - Step 14956: {'lr': 0.00041537705531935476, 'samples': 2871744, 'steps': 14956, 'loss/train': 1.5472996830940247} 01/29/2022 08:02:41 - INFO - codeparrot_training - Step 14957: {'lr': 0.0004153647841347471, 'samples': 2871936, 'steps': 14957, 'loss/train': 1.7817703485488892} 01/29/2022 08:02:45 - INFO - codeparrot_training - Step 14958: {'lr': 0.0004153525122417692, 'samples': 2872128, 'steps': 14958, 'loss/train': 1.0234524607658386} 01/29/2022 08:02:50 - INFO - codeparrot_training - Step 14959: {'lr': 0.00041534023964047363, 'samples': 2872320, 'steps': 14959, 'loss/train': 1.572096347808838} 01/29/2022 08:02:54 - INFO - codeparrot_training - Step 14960: {'lr': 0.00041532796633091297, 'samples': 2872512, 'steps': 14960, 'loss/train': 1.8984814882278442} 01/29/2022 08:02:59 - INFO - codeparrot_training - Step 14961: {'lr': 0.0004153156923131398, 'samples': 2872704, 'steps': 14961, 'loss/train': 1.401386171579361} 01/29/2022 08:03:03 - INFO - codeparrot_training - Step 14962: {'lr': 0.0004153034175872067, 'samples': 2872896, 'steps': 14962, 'loss/train': 1.2972718477249146} 01/29/2022 08:03:07 - INFO - codeparrot_training - Step 14963: {'lr': 0.00041529114215316633, 'samples': 2873088, 'steps': 14963, 'loss/train': 1.0496560335159302} 01/29/2022 08:03:14 - INFO - codeparrot_training - Step 14964: {'lr': 0.0004152788660110711, 'samples': 2873280, 'steps': 14964, 'loss/train': 2.6297664642333984} 01/29/2022 08:03:18 - INFO - codeparrot_training - Step 14965: {'lr': 0.0004152665891609737, 'samples': 2873472, 'steps': 14965, 'loss/train': 2.007663309574127} 01/29/2022 08:03:23 - INFO - codeparrot_training - Step 14966: {'lr': 0.0004152543116029267, 'samples': 2873664, 'steps': 14966, 'loss/train': 1.9815006852149963} 01/29/2022 08:03:27 - INFO - codeparrot_training - Step 14967: {'lr': 0.0004152420333369827, 'samples': 2873856, 'steps': 14967, 'loss/train': 0.9560388922691345} 01/29/2022 08:03:31 - INFO - codeparrot_training - Step 14968: {'lr': 0.00041522975436319445, 'samples': 2874048, 'steps': 14968, 'loss/train': 2.059500753879547} 01/29/2022 08:03:37 - INFO - codeparrot_training - Step 14969: {'lr': 0.00041521747468161417, 'samples': 2874240, 'steps': 14969, 'loss/train': 1.4737579822540283} 01/29/2022 08:03:41 - INFO - codeparrot_training - Step 14970: {'lr': 0.00041520519429229485, 'samples': 2874432, 'steps': 14970, 'loss/train': 1.9572311639785767} 01/29/2022 08:03:45 - INFO - codeparrot_training - Step 14971: {'lr': 0.00041519291319528886, 'samples': 2874624, 'steps': 14971, 'loss/train': 1.5994879603385925} 01/29/2022 08:03:50 - INFO - codeparrot_training - Step 14972: {'lr': 0.00041518063139064893, 'samples': 2874816, 'steps': 14972, 'loss/train': 1.9500035047531128} 01/29/2022 08:03:54 - INFO - codeparrot_training - Step 14973: {'lr': 0.0004151683488784276, 'samples': 2875008, 'steps': 14973, 'loss/train': 1.0596278607845306} 01/29/2022 08:04:01 - INFO - codeparrot_training - Step 14974: {'lr': 0.00041515606565867746, 'samples': 2875200, 'steps': 14974, 'loss/train': 1.9110385179519653} 01/29/2022 08:04:05 - INFO - codeparrot_training - Step 14975: {'lr': 0.0004151437817314513, 'samples': 2875392, 'steps': 14975, 'loss/train': 2.046882748603821} 01/29/2022 08:04:09 - INFO - codeparrot_training - Step 14976: {'lr': 0.00041513149709680155, 'samples': 2875584, 'steps': 14976, 'loss/train': 0.9745549857616425} 01/29/2022 08:04:13 - INFO - codeparrot_training - Step 14977: {'lr': 0.00041511921175478085, 'samples': 2875776, 'steps': 14977, 'loss/train': 2.6296573877334595} 01/29/2022 08:04:18 - INFO - codeparrot_training - Step 14978: {'lr': 0.0004151069257054419, 'samples': 2875968, 'steps': 14978, 'loss/train': 1.3379172384738922} 01/29/2022 08:04:23 - INFO - codeparrot_training - Step 14979: {'lr': 0.0004150946389488374, 'samples': 2876160, 'steps': 14979, 'loss/train': 1.9195098280906677} 01/29/2022 08:04:27 - INFO - codeparrot_training - Step 14980: {'lr': 0.0004150823514850198, 'samples': 2876352, 'steps': 14980, 'loss/train': 2.0515159964561462} 01/29/2022 08:04:31 - INFO - codeparrot_training - Step 14981: {'lr': 0.00041507006331404186, 'samples': 2876544, 'steps': 14981, 'loss/train': 2.0382946729660034} 01/29/2022 08:04:36 - INFO - codeparrot_training - Step 14982: {'lr': 0.00041505777443595615, 'samples': 2876736, 'steps': 14982, 'loss/train': 1.783940076828003} 01/29/2022 08:04:40 - INFO - codeparrot_training - Step 14983: {'lr': 0.0004150454848508154, 'samples': 2876928, 'steps': 14983, 'loss/train': 1.3992882370948792} 01/29/2022 08:04:45 - INFO - codeparrot_training - Step 14984: {'lr': 0.00041503319455867216, 'samples': 2877120, 'steps': 14984, 'loss/train': 1.0208209455013275} 01/29/2022 08:04:50 - INFO - codeparrot_training - Step 14985: {'lr': 0.0004150209035595791, 'samples': 2877312, 'steps': 14985, 'loss/train': 2.9092556834220886} 01/29/2022 08:04:54 - INFO - codeparrot_training - Step 14986: {'lr': 0.000415008611853589, 'samples': 2877504, 'steps': 14986, 'loss/train': 1.1989895403385162} 01/29/2022 08:04:58 - INFO - codeparrot_training - Step 14987: {'lr': 0.0004149963194407543, 'samples': 2877696, 'steps': 14987, 'loss/train': 1.5882670283317566} 01/29/2022 08:05:02 - INFO - codeparrot_training - Step 14988: {'lr': 0.00041498402632112776, 'samples': 2877888, 'steps': 14988, 'loss/train': 1.683288037776947} 01/29/2022 08:05:09 - INFO - codeparrot_training - Step 14989: {'lr': 0.00041497173249476204, 'samples': 2878080, 'steps': 14989, 'loss/train': 1.4154548943042755} 01/29/2022 08:05:14 - INFO - codeparrot_training - Step 14990: {'lr': 0.0004149594379617099, 'samples': 2878272, 'steps': 14990, 'loss/train': 0.5165659636259079} 01/29/2022 08:05:18 - INFO - codeparrot_training - Step 14991: {'lr': 0.00041494714272202385, 'samples': 2878464, 'steps': 14991, 'loss/train': 1.6988039016723633} 01/29/2022 08:05:22 - INFO - codeparrot_training - Step 14992: {'lr': 0.00041493484677575655, 'samples': 2878656, 'steps': 14992, 'loss/train': 1.7177460193634033} 01/29/2022 08:05:26 - INFO - codeparrot_training - Step 14993: {'lr': 0.00041492255012296077, 'samples': 2878848, 'steps': 14993, 'loss/train': 1.7337734699249268} 01/29/2022 08:05:32 - INFO - codeparrot_training - Step 14994: {'lr': 0.0004149102527636892, 'samples': 2879040, 'steps': 14994, 'loss/train': 2.2196186184883118} 01/29/2022 08:05:36 - INFO - codeparrot_training - Step 14995: {'lr': 0.0004148979546979944, 'samples': 2879232, 'steps': 14995, 'loss/train': 2.523157060146332} 01/29/2022 08:05:40 - INFO - codeparrot_training - Step 14996: {'lr': 0.00041488565592592917, 'samples': 2879424, 'steps': 14996, 'loss/train': 1.3133066296577454} 01/29/2022 08:05:45 - INFO - codeparrot_training - Step 14997: {'lr': 0.0004148733564475462, 'samples': 2879616, 'steps': 14997, 'loss/train': 7.572167873382568} 01/29/2022 08:05:49 - INFO - codeparrot_training - Step 14998: {'lr': 0.000414861056262898, 'samples': 2879808, 'steps': 14998, 'loss/train': 1.7404543161392212} 01/29/2022 08:05:54 - INFO - codeparrot_training - Step 14999: {'lr': 0.0004148487553720375, 'samples': 2880000, 'steps': 14999, 'loss/train': 2.595106065273285} 01/29/2022 08:05:58 - INFO - codeparrot_training - Step 15000: {'lr': 0.0004148364537750172, 'samples': 2880192, 'steps': 15000, 'loss/train': 1.9403398633003235} 01/29/2022 08:06:03 - INFO - codeparrot_training - Step 15001: {'lr': 0.0004148241514718899, 'samples': 2880384, 'steps': 15001, 'loss/train': 2.0255518555641174} 01/29/2022 08:06:07 - INFO - codeparrot_training - Step 15002: {'lr': 0.00041481184846270836, 'samples': 2880576, 'steps': 15002, 'loss/train': 1.7625580430030823} 01/29/2022 08:06:11 - INFO - codeparrot_training - Step 15003: {'lr': 0.00041479954474752507, 'samples': 2880768, 'steps': 15003, 'loss/train': 1.6319534182548523} 01/29/2022 08:06:17 - INFO - codeparrot_training - Step 15004: {'lr': 0.0004147872403263929, 'samples': 2880960, 'steps': 15004, 'loss/train': 0.8735050857067108} 01/29/2022 08:06:21 - INFO - codeparrot_training - Step 15005: {'lr': 0.0004147749351993645, 'samples': 2881152, 'steps': 15005, 'loss/train': 1.1051542460918427} 01/29/2022 08:06:25 - INFO - codeparrot_training - Step 15006: {'lr': 0.0004147626293664926, 'samples': 2881344, 'steps': 15006, 'loss/train': 1.8613064289093018} 01/29/2022 08:06:30 - INFO - codeparrot_training - Step 15007: {'lr': 0.00041475032282783, 'samples': 2881536, 'steps': 15007, 'loss/train': 2.3475467562675476} 01/29/2022 08:06:34 - INFO - codeparrot_training - Step 15008: {'lr': 0.0004147380155834293, 'samples': 2881728, 'steps': 15008, 'loss/train': 1.695883333683014} 01/29/2022 08:06:40 - INFO - codeparrot_training - Step 15009: {'lr': 0.00041472570763334316, 'samples': 2881920, 'steps': 15009, 'loss/train': 1.2801354825496674} 01/29/2022 08:06:44 - INFO - codeparrot_training - Step 15010: {'lr': 0.00041471339897762447, 'samples': 2882112, 'steps': 15010, 'loss/train': 1.817686915397644} 01/29/2022 08:06:49 - INFO - codeparrot_training - Step 15011: {'lr': 0.0004147010896163259, 'samples': 2882304, 'steps': 15011, 'loss/train': 1.581150233745575} 01/29/2022 08:06:53 - INFO - codeparrot_training - Step 15012: {'lr': 0.00041468877954950006, 'samples': 2882496, 'steps': 15012, 'loss/train': 2.2386866211891174} 01/29/2022 08:06:57 - INFO - codeparrot_training - Step 15013: {'lr': 0.0004146764687771999, 'samples': 2882688, 'steps': 15013, 'loss/train': 0.974184662103653} 01/29/2022 08:07:03 - INFO - codeparrot_training - Step 15014: {'lr': 0.00041466415729947794, 'samples': 2882880, 'steps': 15014, 'loss/train': 2.176405370235443} 01/29/2022 08:07:07 - INFO - codeparrot_training - Step 15015: {'lr': 0.0004146518451163871, 'samples': 2883072, 'steps': 15015, 'loss/train': 0.9285376667976379} 01/29/2022 08:07:11 - INFO - codeparrot_training - Step 15016: {'lr': 0.00041463953222798, 'samples': 2883264, 'steps': 15016, 'loss/train': 1.7482125163078308} 01/29/2022 08:07:15 - INFO - codeparrot_training - Step 15017: {'lr': 0.00041462721863430943, 'samples': 2883456, 'steps': 15017, 'loss/train': 1.9642433524131775} 01/29/2022 08:07:20 - INFO - codeparrot_training - Step 15018: {'lr': 0.0004146149043354281, 'samples': 2883648, 'steps': 15018, 'loss/train': 1.6770338416099548} 01/29/2022 08:07:27 - INFO - codeparrot_training - Step 15019: {'lr': 0.0004146025893313888, 'samples': 2883840, 'steps': 15019, 'loss/train': 1.8631110191345215} 01/29/2022 08:07:31 - INFO - codeparrot_training - Step 15020: {'lr': 0.00041459027362224433, 'samples': 2884032, 'steps': 15020, 'loss/train': 1.7773681282997131} 01/29/2022 08:07:35 - INFO - codeparrot_training - Step 15021: {'lr': 0.0004145779572080473, 'samples': 2884224, 'steps': 15021, 'loss/train': 1.3856760263442993} 01/29/2022 08:07:39 - INFO - codeparrot_training - Step 15022: {'lr': 0.0004145656400888506, 'samples': 2884416, 'steps': 15022, 'loss/train': 1.4010607302188873} 01/29/2022 08:07:44 - INFO - codeparrot_training - Step 15023: {'lr': 0.000414553322264707, 'samples': 2884608, 'steps': 15023, 'loss/train': 1.0947321653366089} 01/29/2022 08:07:49 - INFO - codeparrot_training - Step 15024: {'lr': 0.00041454100373566915, 'samples': 2884800, 'steps': 15024, 'loss/train': 1.8277616500854492} 01/29/2022 08:07:53 - INFO - codeparrot_training - Step 15025: {'lr': 0.00041452868450178994, 'samples': 2884992, 'steps': 15025, 'loss/train': 1.351940631866455} 01/29/2022 08:07:57 - INFO - codeparrot_training - Step 15026: {'lr': 0.00041451636456312207, 'samples': 2885184, 'steps': 15026, 'loss/train': 1.8429728150367737} 01/29/2022 08:08:02 - INFO - codeparrot_training - Step 15027: {'lr': 0.0004145040439197183, 'samples': 2885376, 'steps': 15027, 'loss/train': 1.5902343392372131} 01/29/2022 08:08:06 - INFO - codeparrot_training - Step 15028: {'lr': 0.00041449172257163156, 'samples': 2885568, 'steps': 15028, 'loss/train': 1.745762050151825} 01/29/2022 08:08:11 - INFO - codeparrot_training - Step 15029: {'lr': 0.00041447940051891435, 'samples': 2885760, 'steps': 15029, 'loss/train': 2.024137258529663} 01/29/2022 08:08:15 - INFO - codeparrot_training - Step 15030: {'lr': 0.00041446707776161975, 'samples': 2885952, 'steps': 15030, 'loss/train': 0.9848880171775818} 01/29/2022 08:08:20 - INFO - codeparrot_training - Step 15031: {'lr': 0.00041445475429980033, 'samples': 2886144, 'steps': 15031, 'loss/train': 1.9651796221733093} 01/29/2022 08:08:24 - INFO - codeparrot_training - Step 15032: {'lr': 0.000414442430133509, 'samples': 2886336, 'steps': 15032, 'loss/train': 1.942741334438324} 01/29/2022 08:08:28 - INFO - codeparrot_training - Step 15033: {'lr': 0.0004144301052627985, 'samples': 2886528, 'steps': 15033, 'loss/train': 1.8168441653251648} 01/29/2022 08:08:35 - INFO - codeparrot_training - Step 15034: {'lr': 0.00041441777968772165, 'samples': 2886720, 'steps': 15034, 'loss/train': 2.4031657576560974} 01/29/2022 08:08:39 - INFO - codeparrot_training - Step 15035: {'lr': 0.00041440545340833124, 'samples': 2886912, 'steps': 15035, 'loss/train': 1.7597562074661255} 01/29/2022 08:08:44 - INFO - codeparrot_training - Step 15036: {'lr': 0.00041439312642468007, 'samples': 2887104, 'steps': 15036, 'loss/train': 1.9711387753486633} 01/29/2022 08:08:48 - INFO - codeparrot_training - Step 15037: {'lr': 0.000414380798736821, 'samples': 2887296, 'steps': 15037, 'loss/train': 0.9770673215389252} 01/29/2022 08:08:52 - INFO - codeparrot_training - Step 15038: {'lr': 0.0004143684703448067, 'samples': 2887488, 'steps': 15038, 'loss/train': 1.602137804031372} 01/29/2022 08:08:57 - INFO - codeparrot_training - Step 15039: {'lr': 0.0004143561412486901, 'samples': 2887680, 'steps': 15039, 'loss/train': 0.9636953473091125} 01/29/2022 08:09:02 - INFO - codeparrot_training - Step 15040: {'lr': 0.00041434381144852395, 'samples': 2887872, 'steps': 15040, 'loss/train': 1.8525913953781128} 01/29/2022 08:09:06 - INFO - codeparrot_training - Step 15041: {'lr': 0.00041433148094436115, 'samples': 2888064, 'steps': 15041, 'loss/train': 1.1056961417198181} 01/29/2022 08:09:10 - INFO - codeparrot_training - Step 15042: {'lr': 0.0004143191497362544, 'samples': 2888256, 'steps': 15042, 'loss/train': 0.6761987060308456} 01/29/2022 08:09:14 - INFO - codeparrot_training - Step 15043: {'lr': 0.0004143068178242566, 'samples': 2888448, 'steps': 15043, 'loss/train': 1.4440027177333832} 01/29/2022 08:09:20 - INFO - codeparrot_training - Step 15044: {'lr': 0.00041429448520842064, 'samples': 2888640, 'steps': 15044, 'loss/train': 2.0349878668785095} 01/29/2022 08:09:24 - INFO - codeparrot_training - Step 15045: {'lr': 0.00041428215188879926, 'samples': 2888832, 'steps': 15045, 'loss/train': 1.791248381137848} 01/29/2022 08:09:28 - INFO - codeparrot_training - Step 15046: {'lr': 0.0004142698178654453, 'samples': 2889024, 'steps': 15046, 'loss/train': 1.713328778743744} 01/29/2022 08:09:33 - INFO - codeparrot_training - Step 15047: {'lr': 0.0004142574831384115, 'samples': 2889216, 'steps': 15047, 'loss/train': 1.1289810240268707} 01/29/2022 08:09:37 - INFO - codeparrot_training - Step 15048: {'lr': 0.0004142451477077509, 'samples': 2889408, 'steps': 15048, 'loss/train': 2.367025852203369} 01/29/2022 08:09:41 - INFO - codeparrot_training - Step 15049: {'lr': 0.00041423281157351624, 'samples': 2889600, 'steps': 15049, 'loss/train': 6.412234783172607} 01/29/2022 08:09:49 - INFO - codeparrot_training - Step 15050: {'lr': 0.00041422047473576033, 'samples': 2889792, 'steps': 15050, 'loss/train': 0.6972000449895859} 01/29/2022 08:09:53 - INFO - codeparrot_training - Step 15051: {'lr': 0.0004142081371945361, 'samples': 2889984, 'steps': 15051, 'loss/train': 1.7079216241836548} 01/29/2022 08:09:57 - INFO - codeparrot_training - Step 15052: {'lr': 0.00041419579894989633, 'samples': 2890176, 'steps': 15052, 'loss/train': 1.7545332312583923} 01/29/2022 08:10:01 - INFO - codeparrot_training - Step 15053: {'lr': 0.0004141834600018939, 'samples': 2890368, 'steps': 15053, 'loss/train': 1.216050624847412} 01/29/2022 08:10:06 - INFO - codeparrot_training - Step 15054: {'lr': 0.00041417112035058157, 'samples': 2890560, 'steps': 15054, 'loss/train': 1.5378566980361938} 01/29/2022 08:10:11 - INFO - codeparrot_training - Step 15055: {'lr': 0.00041415877999601236, 'samples': 2890752, 'steps': 15055, 'loss/train': 1.8195657134056091} 01/29/2022 08:10:15 - INFO - codeparrot_training - Step 15056: {'lr': 0.0004141464389382391, 'samples': 2890944, 'steps': 15056, 'loss/train': 2.2793367505073547} 01/29/2022 08:10:20 - INFO - codeparrot_training - Step 15057: {'lr': 0.0004141340971773147, 'samples': 2891136, 'steps': 15057, 'loss/train': 1.0538970530033112} 01/29/2022 08:10:24 - INFO - codeparrot_training - Step 15058: {'lr': 0.00041412175471329174, 'samples': 2891328, 'steps': 15058, 'loss/train': 1.912961483001709} 01/29/2022 08:10:28 - INFO - codeparrot_training - Step 15059: {'lr': 0.0004141094115462234, 'samples': 2891520, 'steps': 15059, 'loss/train': 1.6907238364219666} 01/29/2022 08:10:33 - INFO - codeparrot_training - Step 15060: {'lr': 0.00041409706767616246, 'samples': 2891712, 'steps': 15060, 'loss/train': 1.6314915418624878} 01/29/2022 08:10:38 - INFO - codeparrot_training - Step 15061: {'lr': 0.0004140847231031618, 'samples': 2891904, 'steps': 15061, 'loss/train': 1.1236285865306854} 01/29/2022 08:10:42 - INFO - codeparrot_training - Step 15062: {'lr': 0.00041407237782727427, 'samples': 2892096, 'steps': 15062, 'loss/train': 1.986975073814392} 01/29/2022 08:10:46 - INFO - codeparrot_training - Step 15063: {'lr': 0.0004140600318485527, 'samples': 2892288, 'steps': 15063, 'loss/train': 0.846555233001709} 01/29/2022 08:10:51 - INFO - codeparrot_training - Step 15064: {'lr': 0.0004140476851670502, 'samples': 2892480, 'steps': 15064, 'loss/train': 1.7122732400894165} 01/29/2022 08:10:58 - INFO - codeparrot_training - Step 15065: {'lr': 0.00041403533778281934, 'samples': 2892672, 'steps': 15065, 'loss/train': 1.6720243692398071} 01/29/2022 08:11:02 - INFO - codeparrot_training - Step 15066: {'lr': 0.0004140229896959132, 'samples': 2892864, 'steps': 15066, 'loss/train': 2.696657717227936} 01/29/2022 08:11:06 - INFO - codeparrot_training - Step 15067: {'lr': 0.00041401064090638474, 'samples': 2893056, 'steps': 15067, 'loss/train': 1.9335576295852661} 01/29/2022 08:11:10 - INFO - codeparrot_training - Step 15068: {'lr': 0.0004139982914142868, 'samples': 2893248, 'steps': 15068, 'loss/train': 1.698594868183136} 01/29/2022 08:11:15 - INFO - codeparrot_training - Step 15069: {'lr': 0.00041398594121967215, 'samples': 2893440, 'steps': 15069, 'loss/train': 2.3123783469200134} 01/29/2022 08:11:20 - INFO - codeparrot_training - Step 15070: {'lr': 0.0004139735903225939, 'samples': 2893632, 'steps': 15070, 'loss/train': 1.8869338631629944} 01/29/2022 08:11:24 - INFO - codeparrot_training - Step 15071: {'lr': 0.0004139612387231048, 'samples': 2893824, 'steps': 15071, 'loss/train': 0.8810751736164093} 01/29/2022 08:11:29 - INFO - codeparrot_training - Step 15072: {'lr': 0.0004139488864212578, 'samples': 2894016, 'steps': 15072, 'loss/train': 1.401540756225586} 01/29/2022 08:11:33 - INFO - codeparrot_training - Step 15073: {'lr': 0.0004139365334171059, 'samples': 2894208, 'steps': 15073, 'loss/train': 2.6108949780464172} 01/29/2022 08:11:37 - INFO - codeparrot_training - Step 15074: {'lr': 0.0004139241797107019, 'samples': 2894400, 'steps': 15074, 'loss/train': 1.863554298877716} 01/29/2022 08:11:42 - INFO - codeparrot_training - Step 15075: {'lr': 0.00041391182530209873, 'samples': 2894592, 'steps': 15075, 'loss/train': 1.157733142375946} 01/29/2022 08:11:47 - INFO - codeparrot_training - Step 15076: {'lr': 0.0004138994701913494, 'samples': 2894784, 'steps': 15076, 'loss/train': 1.4093922972679138} 01/29/2022 08:11:51 - INFO - codeparrot_training - Step 15077: {'lr': 0.00041388711437850676, 'samples': 2894976, 'steps': 15077, 'loss/train': 1.6825520396232605} 01/29/2022 08:11:55 - INFO - codeparrot_training - Step 15078: {'lr': 0.00041387475786362386, 'samples': 2895168, 'steps': 15078, 'loss/train': 0.6657692492008209} 01/29/2022 08:11:59 - INFO - codeparrot_training - Step 15079: {'lr': 0.0004138624006467534, 'samples': 2895360, 'steps': 15079, 'loss/train': 0.7845547199249268} 01/29/2022 08:12:06 - INFO - codeparrot_training - Step 15080: {'lr': 0.00041385004272794846, 'samples': 2895552, 'steps': 15080, 'loss/train': 1.5567240715026855} 01/29/2022 08:12:11 - INFO - codeparrot_training - Step 15081: {'lr': 0.00041383768410726207, 'samples': 2895744, 'steps': 15081, 'loss/train': 1.6225064992904663} 01/29/2022 08:12:15 - INFO - codeparrot_training - Step 15082: {'lr': 0.000413825324784747, 'samples': 2895936, 'steps': 15082, 'loss/train': 1.4735843539237976} 01/29/2022 08:12:19 - INFO - codeparrot_training - Step 15083: {'lr': 0.00041381296476045626, 'samples': 2896128, 'steps': 15083, 'loss/train': 1.7891099452972412} 01/29/2022 08:12:23 - INFO - codeparrot_training - Step 15084: {'lr': 0.0004138006040344428, 'samples': 2896320, 'steps': 15084, 'loss/train': 1.7426426410675049} 01/29/2022 08:12:29 - INFO - codeparrot_training - Step 15085: {'lr': 0.0004137882426067595, 'samples': 2896512, 'steps': 15085, 'loss/train': 1.4833364188671112} 01/29/2022 08:12:33 - INFO - codeparrot_training - Step 15086: {'lr': 0.0004137758804774594, 'samples': 2896704, 'steps': 15086, 'loss/train': 1.627194106578827} 01/29/2022 08:12:37 - INFO - codeparrot_training - Step 15087: {'lr': 0.0004137635176465955, 'samples': 2896896, 'steps': 15087, 'loss/train': 2.7885721921920776} 01/29/2022 08:12:41 - INFO - codeparrot_training - Step 15088: {'lr': 0.00041375115411422064, 'samples': 2897088, 'steps': 15088, 'loss/train': 1.586511254310608} 01/29/2022 08:12:46 - INFO - codeparrot_training - Step 15089: {'lr': 0.0004137387898803878, 'samples': 2897280, 'steps': 15089, 'loss/train': 1.1611791551113129} 01/29/2022 08:12:51 - INFO - codeparrot_training - Step 15090: {'lr': 0.0004137264249451501, 'samples': 2897472, 'steps': 15090, 'loss/train': 1.9005162119865417} 01/29/2022 08:12:56 - INFO - codeparrot_training - Step 15091: {'lr': 0.00041371405930856026, 'samples': 2897664, 'steps': 15091, 'loss/train': 1.8270662426948547} 01/29/2022 08:13:00 - INFO - codeparrot_training - Step 15092: {'lr': 0.00041370169297067145, 'samples': 2897856, 'steps': 15092, 'loss/train': 1.529507339000702} 01/29/2022 08:13:04 - INFO - codeparrot_training - Step 15093: {'lr': 0.0004136893259315365, 'samples': 2898048, 'steps': 15093, 'loss/train': 1.343118667602539} 01/29/2022 08:13:11 - INFO - codeparrot_training - Step 15094: {'lr': 0.00041367695819120854, 'samples': 2898240, 'steps': 15094, 'loss/train': 1.4697912633419037} 01/29/2022 08:13:16 - INFO - codeparrot_training - Step 15095: {'lr': 0.0004136645897497404, 'samples': 2898432, 'steps': 15095, 'loss/train': 1.1833214163780212} 01/29/2022 08:13:20 - INFO - codeparrot_training - Step 15096: {'lr': 0.0004136522206071852, 'samples': 2898624, 'steps': 15096, 'loss/train': 1.806803047657013} 01/29/2022 08:13:24 - INFO - codeparrot_training - Step 15097: {'lr': 0.0004136398507635958, 'samples': 2898816, 'steps': 15097, 'loss/train': 1.4344451129436493} 01/29/2022 08:13:28 - INFO - codeparrot_training - Step 15098: {'lr': 0.00041362748021902526, 'samples': 2899008, 'steps': 15098, 'loss/train': 1.4482572376728058} 01/29/2022 08:13:34 - INFO - codeparrot_training - Step 15099: {'lr': 0.0004136151089735265, 'samples': 2899200, 'steps': 15099, 'loss/train': 1.3729945421218872} 01/29/2022 08:13:38 - INFO - codeparrot_training - Step 15100: {'lr': 0.00041360273702715263, 'samples': 2899392, 'steps': 15100, 'loss/train': 1.5081088542938232} 01/29/2022 08:13:42 - INFO - codeparrot_training - Step 15101: {'lr': 0.0004135903643799566, 'samples': 2899584, 'steps': 15101, 'loss/train': 1.408628225326538} 01/29/2022 08:13:46 - INFO - codeparrot_training - Step 15102: {'lr': 0.00041357799103199127, 'samples': 2899776, 'steps': 15102, 'loss/train': 0.7718151211738586} 01/29/2022 08:13:51 - INFO - codeparrot_training - Step 15103: {'lr': 0.00041356561698330984, 'samples': 2899968, 'steps': 15103, 'loss/train': 2.003062605857849} 01/29/2022 08:13:58 - INFO - codeparrot_training - Step 15104: {'lr': 0.0004135532422339653, 'samples': 2900160, 'steps': 15104, 'loss/train': 1.766345500946045} 01/29/2022 08:14:02 - INFO - codeparrot_training - Step 15105: {'lr': 0.00041354086678401056, 'samples': 2900352, 'steps': 15105, 'loss/train': 1.1169481873512268} 01/29/2022 08:14:06 - INFO - codeparrot_training - Step 15106: {'lr': 0.00041352849063349865, 'samples': 2900544, 'steps': 15106, 'loss/train': 1.9412382245063782} 01/29/2022 08:14:10 - INFO - codeparrot_training - Step 15107: {'lr': 0.0004135161137824827, 'samples': 2900736, 'steps': 15107, 'loss/train': 1.4961302876472473} 01/29/2022 08:14:15 - INFO - codeparrot_training - Step 15108: {'lr': 0.0004135037362310155, 'samples': 2900928, 'steps': 15108, 'loss/train': 1.3744585812091827} 01/29/2022 08:14:20 - INFO - codeparrot_training - Step 15109: {'lr': 0.0004134913579791503, 'samples': 2901120, 'steps': 15109, 'loss/train': 1.7331576347351074} 01/29/2022 08:14:24 - INFO - codeparrot_training - Step 15110: {'lr': 0.00041347897902694003, 'samples': 2901312, 'steps': 15110, 'loss/train': 1.7765836715698242} 01/29/2022 08:14:28 - INFO - codeparrot_training - Step 15111: {'lr': 0.00041346659937443775, 'samples': 2901504, 'steps': 15111, 'loss/train': 1.8897237181663513} 01/29/2022 08:14:33 - INFO - codeparrot_training - Step 15112: {'lr': 0.00041345421902169645, 'samples': 2901696, 'steps': 15112, 'loss/train': 1.6234098672866821} 01/29/2022 08:14:37 - INFO - codeparrot_training - Step 15113: {'lr': 0.0004134418379687691, 'samples': 2901888, 'steps': 15113, 'loss/train': 2.386593282222748} 01/29/2022 08:14:42 - INFO - codeparrot_training - Step 15114: {'lr': 0.0004134294562157089, 'samples': 2902080, 'steps': 15114, 'loss/train': 2.2017163038253784} 01/29/2022 08:14:47 - INFO - codeparrot_training - Step 15115: {'lr': 0.00041341707376256877, 'samples': 2902272, 'steps': 15115, 'loss/train': 0.4178610295057297} 01/29/2022 08:14:51 - INFO - codeparrot_training - Step 15116: {'lr': 0.00041340469060940183, 'samples': 2902464, 'steps': 15116, 'loss/train': 3.3779579401016235} 01/29/2022 08:14:55 - INFO - codeparrot_training - Step 15117: {'lr': 0.0004133923067562611, 'samples': 2902656, 'steps': 15117, 'loss/train': 1.134184330701828} 01/29/2022 08:14:59 - INFO - codeparrot_training - Step 15118: {'lr': 0.0004133799222031995, 'samples': 2902848, 'steps': 15118, 'loss/train': 0.8985936641693115} 01/29/2022 08:15:05 - INFO - codeparrot_training - Step 15119: {'lr': 0.0004133675369502703, 'samples': 2903040, 'steps': 15119, 'loss/train': 2.0892282128334045} 01/29/2022 08:15:09 - INFO - codeparrot_training - Step 15120: {'lr': 0.0004133551509975264, 'samples': 2903232, 'steps': 15120, 'loss/train': 1.9944536089897156} 01/29/2022 08:15:13 - INFO - codeparrot_training - Step 15121: {'lr': 0.0004133427643450209, 'samples': 2903424, 'steps': 15121, 'loss/train': 1.727641224861145} 01/29/2022 08:15:17 - INFO - codeparrot_training - Step 15122: {'lr': 0.0004133303769928068, 'samples': 2903616, 'steps': 15122, 'loss/train': 1.8204787373542786} 01/29/2022 08:15:22 - INFO - codeparrot_training - Step 15123: {'lr': 0.00041331798894093735, 'samples': 2903808, 'steps': 15123, 'loss/train': 1.4235865473747253} 01/29/2022 08:15:29 - INFO - codeparrot_training - Step 15124: {'lr': 0.0004133056001894655, 'samples': 2904000, 'steps': 15124, 'loss/train': 1.5448059439659119} 01/29/2022 08:15:33 - INFO - codeparrot_training - Step 15125: {'lr': 0.0004132932107384442, 'samples': 2904192, 'steps': 15125, 'loss/train': 1.5973691940307617} 01/29/2022 08:15:37 - INFO - codeparrot_training - Step 15126: {'lr': 0.0004132808205879267, 'samples': 2904384, 'steps': 15126, 'loss/train': 2.092495858669281} 01/29/2022 08:15:42 - INFO - codeparrot_training - Step 15127: {'lr': 0.000413268429737966, 'samples': 2904576, 'steps': 15127, 'loss/train': 1.5769910216331482} 01/29/2022 08:15:46 - INFO - codeparrot_training - Step 15128: {'lr': 0.00041325603818861517, 'samples': 2904768, 'steps': 15128, 'loss/train': 0.5945752859115601} 01/29/2022 08:15:51 - INFO - codeparrot_training - Step 15129: {'lr': 0.00041324364593992735, 'samples': 2904960, 'steps': 15129, 'loss/train': 2.1045196652412415} 01/29/2022 08:15:56 - INFO - codeparrot_training - Step 15130: {'lr': 0.00041323125299195563, 'samples': 2905152, 'steps': 15130, 'loss/train': 0.5386669188737869} 01/29/2022 08:16:00 - INFO - codeparrot_training - Step 15131: {'lr': 0.000413218859344753, 'samples': 2905344, 'steps': 15131, 'loss/train': 2.026195764541626} 01/29/2022 08:16:04 - INFO - codeparrot_training - Step 15132: {'lr': 0.00041320646499837254, 'samples': 2905536, 'steps': 15132, 'loss/train': 1.6976138949394226} 01/29/2022 08:16:08 - INFO - codeparrot_training - Step 15133: {'lr': 0.00041319406995286753, 'samples': 2905728, 'steps': 15133, 'loss/train': 1.8078707456588745} 01/29/2022 08:16:14 - INFO - codeparrot_training - Step 15134: {'lr': 0.0004131816742082909, 'samples': 2905920, 'steps': 15134, 'loss/train': 1.4771286249160767} 01/29/2022 08:16:18 - INFO - codeparrot_training - Step 15135: {'lr': 0.00041316927776469575, 'samples': 2906112, 'steps': 15135, 'loss/train': 1.431641936302185} 01/29/2022 08:16:22 - INFO - codeparrot_training - Step 15136: {'lr': 0.00041315688062213524, 'samples': 2906304, 'steps': 15136, 'loss/train': 1.226797878742218} 01/29/2022 08:16:26 - INFO - codeparrot_training - Step 15137: {'lr': 0.0004131444827806625, 'samples': 2906496, 'steps': 15137, 'loss/train': 1.8453279733657837} 01/29/2022 08:16:31 - INFO - codeparrot_training - Step 15138: {'lr': 0.00041313208424033056, 'samples': 2906688, 'steps': 15138, 'loss/train': 1.4142328798770905} 01/29/2022 08:16:38 - INFO - codeparrot_training - Step 15139: {'lr': 0.0004131196850011926, 'samples': 2906880, 'steps': 15139, 'loss/train': 1.2811496257781982} 01/29/2022 08:16:42 - INFO - codeparrot_training - Step 15140: {'lr': 0.0004131072850633017, 'samples': 2907072, 'steps': 15140, 'loss/train': 1.5575401782989502} 01/29/2022 08:16:46 - INFO - codeparrot_training - Step 15141: {'lr': 0.00041309488442671093, 'samples': 2907264, 'steps': 15141, 'loss/train': 1.6074773669242859} 01/29/2022 08:16:50 - INFO - codeparrot_training - Step 15142: {'lr': 0.00041308248309147356, 'samples': 2907456, 'steps': 15142, 'loss/train': 1.9576318860054016} 01/29/2022 08:16:55 - INFO - codeparrot_training - Step 15143: {'lr': 0.00041307008105764256, 'samples': 2907648, 'steps': 15143, 'loss/train': 1.017648607492447} 01/29/2022 08:17:00 - INFO - codeparrot_training - Step 15144: {'lr': 0.0004130576783252712, 'samples': 2907840, 'steps': 15144, 'loss/train': 1.4414893984794617} 01/29/2022 08:17:04 - INFO - codeparrot_training - Step 15145: {'lr': 0.00041304527489441237, 'samples': 2908032, 'steps': 15145, 'loss/train': 1.857582449913025} 01/29/2022 08:17:08 - INFO - codeparrot_training - Step 15146: {'lr': 0.0004130328707651195, 'samples': 2908224, 'steps': 15146, 'loss/train': 0.9011490941047668} 01/29/2022 08:17:13 - INFO - codeparrot_training - Step 15147: {'lr': 0.00041302046593744547, 'samples': 2908416, 'steps': 15147, 'loss/train': 1.7484194040298462} 01/29/2022 08:17:17 - INFO - codeparrot_training - Step 15148: {'lr': 0.00041300806041144356, 'samples': 2908608, 'steps': 15148, 'loss/train': 1.871580719947815} 01/29/2022 08:17:24 - INFO - codeparrot_training - Step 15149: {'lr': 0.0004129956541871669, 'samples': 2908800, 'steps': 15149, 'loss/train': 1.2926664054393768} 01/29/2022 08:17:28 - INFO - codeparrot_training - Step 15150: {'lr': 0.00041298324726466855, 'samples': 2908992, 'steps': 15150, 'loss/train': 2.0263127088546753} 01/29/2022 08:17:32 - INFO - codeparrot_training - Step 15151: {'lr': 0.0004129708396440018, 'samples': 2909184, 'steps': 15151, 'loss/train': 1.9328851103782654} 01/29/2022 08:17:37 - INFO - codeparrot_training - Step 15152: {'lr': 0.00041295843132521973, 'samples': 2909376, 'steps': 15152, 'loss/train': 1.5150275230407715} 01/29/2022 08:17:41 - INFO - codeparrot_training - Step 15153: {'lr': 0.0004129460223083754, 'samples': 2909568, 'steps': 15153, 'loss/train': 1.8311099410057068} 01/29/2022 08:17:46 - INFO - codeparrot_training - Step 15154: {'lr': 0.0004129336125935221, 'samples': 2909760, 'steps': 15154, 'loss/train': 1.7330914735794067} 01/29/2022 08:17:50 - INFO - codeparrot_training - Step 15155: {'lr': 0.000412921202180713, 'samples': 2909952, 'steps': 15155, 'loss/train': 1.5591992139816284} 01/29/2022 08:17:55 - INFO - codeparrot_training - Step 15156: {'lr': 0.00041290879107000114, 'samples': 2910144, 'steps': 15156, 'loss/train': 0.9354685842990875} 01/29/2022 08:17:59 - INFO - codeparrot_training - Step 15157: {'lr': 0.00041289637926143974, 'samples': 2910336, 'steps': 15157, 'loss/train': 1.97055584192276} 01/29/2022 08:18:03 - INFO - codeparrot_training - Step 15158: {'lr': 0.000412883966755082, 'samples': 2910528, 'steps': 15158, 'loss/train': 1.7676154375076294} 01/29/2022 08:18:09 - INFO - codeparrot_training - Step 15159: {'lr': 0.000412871553550981, 'samples': 2910720, 'steps': 15159, 'loss/train': 2.1380467414855957} 01/29/2022 08:18:13 - INFO - codeparrot_training - Step 15160: {'lr': 0.00041285913964919006, 'samples': 2910912, 'steps': 15160, 'loss/train': 1.8467057347297668} 01/29/2022 08:18:17 - INFO - codeparrot_training - Step 15161: {'lr': 0.0004128467250497623, 'samples': 2911104, 'steps': 15161, 'loss/train': 2.2535635828971863} 01/29/2022 08:18:21 - INFO - codeparrot_training - Step 15162: {'lr': 0.00041283430975275085, 'samples': 2911296, 'steps': 15162, 'loss/train': 0.5748158097267151} 01/29/2022 08:18:26 - INFO - codeparrot_training - Step 15163: {'lr': 0.0004128218937582089, 'samples': 2911488, 'steps': 15163, 'loss/train': 0.6813349574804306} 01/29/2022 08:18:32 - INFO - codeparrot_training - Step 15164: {'lr': 0.00041280947706618965, 'samples': 2911680, 'steps': 15164, 'loss/train': 1.341524362564087} 01/29/2022 08:18:36 - INFO - codeparrot_training - Step 15165: {'lr': 0.00041279705967674636, 'samples': 2911872, 'steps': 15165, 'loss/train': 2.091796875} 01/29/2022 08:18:40 - INFO - codeparrot_training - Step 15166: {'lr': 0.00041278464158993214, 'samples': 2912064, 'steps': 15166, 'loss/train': 1.993895173072815} 01/29/2022 08:18:45 - INFO - codeparrot_training - Step 15167: {'lr': 0.0004127722228058002, 'samples': 2912256, 'steps': 15167, 'loss/train': 1.3012103140354156} 01/29/2022 08:18:49 - INFO - codeparrot_training - Step 15168: {'lr': 0.0004127598033244037, 'samples': 2912448, 'steps': 15168, 'loss/train': 2.2747535705566406} 01/29/2022 08:18:54 - INFO - codeparrot_training - Step 15169: {'lr': 0.0004127473831457959, 'samples': 2912640, 'steps': 15169, 'loss/train': 1.309122383594513} 01/29/2022 08:18:58 - INFO - codeparrot_training - Step 15170: {'lr': 0.00041273496227003004, 'samples': 2912832, 'steps': 15170, 'loss/train': 1.2520784139633179} 01/29/2022 08:19:03 - INFO - codeparrot_training - Step 15171: {'lr': 0.0004127225406971592, 'samples': 2913024, 'steps': 15171, 'loss/train': 1.8160341382026672} 01/29/2022 08:19:07 - INFO - codeparrot_training - Step 15172: {'lr': 0.00041271011842723676, 'samples': 2913216, 'steps': 15172, 'loss/train': 2.5868207216262817} 01/29/2022 08:19:11 - INFO - codeparrot_training - Step 15173: {'lr': 0.00041269769546031576, 'samples': 2913408, 'steps': 15173, 'loss/train': 1.193785160779953} 01/29/2022 08:19:16 - INFO - codeparrot_training - Step 15174: {'lr': 0.0004126852717964495, 'samples': 2913600, 'steps': 15174, 'loss/train': 0.6200904697179794} 01/29/2022 08:19:21 - INFO - codeparrot_training - Step 15175: {'lr': 0.0004126728474356912, 'samples': 2913792, 'steps': 15175, 'loss/train': 1.8251041173934937} 01/29/2022 08:19:25 - INFO - codeparrot_training - Step 15176: {'lr': 0.0004126604223780941, 'samples': 2913984, 'steps': 15176, 'loss/train': 1.746650218963623} 01/29/2022 08:19:29 - INFO - codeparrot_training - Step 15177: {'lr': 0.00041264799662371144, 'samples': 2914176, 'steps': 15177, 'loss/train': 1.0056418776512146} 01/29/2022 08:19:33 - INFO - codeparrot_training - Step 15178: {'lr': 0.0004126355701725963, 'samples': 2914368, 'steps': 15178, 'loss/train': 1.497155249118805} 01/29/2022 08:19:39 - INFO - codeparrot_training - Step 15179: {'lr': 0.00041262314302480216, 'samples': 2914560, 'steps': 15179, 'loss/train': 2.154334545135498} 01/29/2022 08:19:43 - INFO - codeparrot_training - Step 15180: {'lr': 0.000412610715180382, 'samples': 2914752, 'steps': 15180, 'loss/train': 2.24158376455307} 01/29/2022 08:19:47 - INFO - codeparrot_training - Step 15181: {'lr': 0.0004125982866393892, 'samples': 2914944, 'steps': 15181, 'loss/train': 1.9275251626968384} 01/29/2022 08:19:51 - INFO - codeparrot_training - Step 15182: {'lr': 0.0004125858574018769, 'samples': 2915136, 'steps': 15182, 'loss/train': 1.87395840883255} 01/29/2022 08:19:56 - INFO - codeparrot_training - Step 15183: {'lr': 0.0004125734274678986, 'samples': 2915328, 'steps': 15183, 'loss/train': 2.1786460876464844} 01/29/2022 08:20:03 - INFO - codeparrot_training - Step 15184: {'lr': 0.0004125609968375072, 'samples': 2915520, 'steps': 15184, 'loss/train': 1.6768015623092651} 01/29/2022 08:20:07 - INFO - codeparrot_training - Step 15185: {'lr': 0.00041254856551075616, 'samples': 2915712, 'steps': 15185, 'loss/train': 2.275212049484253} 01/29/2022 08:20:11 - INFO - codeparrot_training - Step 15186: {'lr': 0.0004125361334876987, 'samples': 2915904, 'steps': 15186, 'loss/train': 1.9542558789253235} 01/29/2022 08:20:16 - INFO - codeparrot_training - Step 15187: {'lr': 0.000412523700768388, 'samples': 2916096, 'steps': 15187, 'loss/train': 2.416799247264862} 01/29/2022 08:20:20 - INFO - codeparrot_training - Step 15188: {'lr': 0.0004125112673528775, 'samples': 2916288, 'steps': 15188, 'loss/train': 1.9682777523994446} 01/29/2022 08:20:25 - INFO - codeparrot_training - Step 15189: {'lr': 0.0004124988332412202, 'samples': 2916480, 'steps': 15189, 'loss/train': 1.0462448000907898} 01/29/2022 08:20:29 - INFO - codeparrot_training - Step 15190: {'lr': 0.00041248639843346953, 'samples': 2916672, 'steps': 15190, 'loss/train': 1.692095696926117} 01/29/2022 08:20:34 - INFO - codeparrot_training - Step 15191: {'lr': 0.0004124739629296787, 'samples': 2916864, 'steps': 15191, 'loss/train': 1.2927670776844025} 01/29/2022 08:20:38 - INFO - codeparrot_training - Step 15192: {'lr': 0.00041246152672990105, 'samples': 2917056, 'steps': 15192, 'loss/train': 1.239982545375824} 01/29/2022 08:20:42 - INFO - codeparrot_training - Step 15193: {'lr': 0.00041244908983418985, 'samples': 2917248, 'steps': 15193, 'loss/train': 1.5925519466400146} 01/29/2022 08:20:49 - INFO - codeparrot_training - Step 15194: {'lr': 0.0004124366522425982, 'samples': 2917440, 'steps': 15194, 'loss/train': 1.9282933473587036} 01/29/2022 08:20:54 - INFO - codeparrot_training - Step 15195: {'lr': 0.0004124242139551796, 'samples': 2917632, 'steps': 15195, 'loss/train': 1.6419267654418945} 01/29/2022 08:20:58 - INFO - codeparrot_training - Step 15196: {'lr': 0.00041241177497198725, 'samples': 2917824, 'steps': 15196, 'loss/train': 1.2443050146102905} 01/29/2022 08:21:02 - INFO - codeparrot_training - Step 15197: {'lr': 0.00041239933529307437, 'samples': 2918016, 'steps': 15197, 'loss/train': 1.4238861501216888} 01/29/2022 08:21:06 - INFO - codeparrot_training - Step 15198: {'lr': 0.00041238689491849434, 'samples': 2918208, 'steps': 15198, 'loss/train': 0.20501554012298584} 01/29/2022 08:21:12 - INFO - codeparrot_training - Step 15199: {'lr': 0.00041237445384830043, 'samples': 2918400, 'steps': 15199, 'loss/train': 1.6712032556533813} 01/29/2022 08:21:16 - INFO - codeparrot_training - Step 15200: {'lr': 0.0004123620120825459, 'samples': 2918592, 'steps': 15200, 'loss/train': 1.5452263355255127} 01/29/2022 08:21:20 - INFO - codeparrot_training - Step 15201: {'lr': 0.0004123495696212841, 'samples': 2918784, 'steps': 15201, 'loss/train': 1.6578864455223083} 01/29/2022 08:21:24 - INFO - codeparrot_training - Step 15202: {'lr': 0.00041233712646456823, 'samples': 2918976, 'steps': 15202, 'loss/train': 2.0576496720314026} 01/29/2022 08:21:29 - INFO - codeparrot_training - Step 15203: {'lr': 0.0004123246826124517, 'samples': 2919168, 'steps': 15203, 'loss/train': 1.941560447216034} 01/29/2022 08:21:34 - INFO - codeparrot_training - Step 15204: {'lr': 0.00041231223806498777, 'samples': 2919360, 'steps': 15204, 'loss/train': 1.5055382251739502} 01/29/2022 08:21:38 - INFO - codeparrot_training - Step 15205: {'lr': 0.0004122997928222298, 'samples': 2919552, 'steps': 15205, 'loss/train': 1.637494683265686} 01/29/2022 08:21:42 - INFO - codeparrot_training - Step 15206: {'lr': 0.000412287346884231, 'samples': 2919744, 'steps': 15206, 'loss/train': 0.7907828092575073} 01/29/2022 08:21:47 - INFO - codeparrot_training - Step 15207: {'lr': 0.00041227490025104474, 'samples': 2919936, 'steps': 15207, 'loss/train': 1.4168875515460968} 01/29/2022 08:21:51 - INFO - codeparrot_training - Step 15208: {'lr': 0.00041226245292272433, 'samples': 2920128, 'steps': 15208, 'loss/train': 1.8990494012832642} 01/29/2022 08:21:59 - INFO - codeparrot_training - Step 15209: {'lr': 0.00041225000489932315, 'samples': 2920320, 'steps': 15209, 'loss/train': 1.4599458575248718} 01/29/2022 08:22:03 - INFO - codeparrot_training - Step 15210: {'lr': 0.00041223755618089445, 'samples': 2920512, 'steps': 15210, 'loss/train': 1.5318288803100586} 01/29/2022 08:22:07 - INFO - codeparrot_training - Step 15211: {'lr': 0.0004122251067674915, 'samples': 2920704, 'steps': 15211, 'loss/train': 1.5231199264526367} 01/29/2022 08:22:11 - INFO - codeparrot_training - Step 15212: {'lr': 0.00041221265665916776, 'samples': 2920896, 'steps': 15212, 'loss/train': 0.21180767565965652} 01/29/2022 08:22:16 - INFO - codeparrot_training - Step 15213: {'lr': 0.0004122002058559765, 'samples': 2921088, 'steps': 15213, 'loss/train': 1.0553897023200989} 01/29/2022 08:22:20 - INFO - codeparrot_training - Step 15214: {'lr': 0.00041218775435797106, 'samples': 2921280, 'steps': 15214, 'loss/train': 1.9887396097183228} 01/29/2022 08:22:25 - INFO - codeparrot_training - Step 15215: {'lr': 0.0004121753021652048, 'samples': 2921472, 'steps': 15215, 'loss/train': 1.6127732992172241} 01/29/2022 08:22:30 - INFO - codeparrot_training - Step 15216: {'lr': 0.0004121628492777311, 'samples': 2921664, 'steps': 15216, 'loss/train': 0.388149619102478} 01/29/2022 08:22:34 - INFO - codeparrot_training - Step 15217: {'lr': 0.0004121503956956031, 'samples': 2921856, 'steps': 15217, 'loss/train': 1.1470326483249664} 01/29/2022 08:22:38 - INFO - codeparrot_training - Step 15218: {'lr': 0.0004121379414188744, 'samples': 2922048, 'steps': 15218, 'loss/train': 1.3231312930583954} 01/29/2022 08:22:42 - INFO - codeparrot_training - Step 15219: {'lr': 0.0004121254864475982, 'samples': 2922240, 'steps': 15219, 'loss/train': 1.893519937992096} 01/29/2022 08:22:48 - INFO - codeparrot_training - Step 15220: {'lr': 0.0004121130307818279, 'samples': 2922432, 'steps': 15220, 'loss/train': 2.538579046726227} 01/29/2022 08:22:52 - INFO - codeparrot_training - Step 15221: {'lr': 0.00041210057442161687, 'samples': 2922624, 'steps': 15221, 'loss/train': 0.9470820128917694} 01/29/2022 08:22:56 - INFO - codeparrot_training - Step 15222: {'lr': 0.0004120881173670184, 'samples': 2922816, 'steps': 15222, 'loss/train': 1.5097439289093018} 01/29/2022 08:23:00 - INFO - codeparrot_training - Step 15223: {'lr': 0.000412075659618086, 'samples': 2923008, 'steps': 15223, 'loss/train': 2.0377694964408875} 01/29/2022 08:23:04 - INFO - codeparrot_training - Step 15224: {'lr': 0.0004120632011748728, 'samples': 2923200, 'steps': 15224, 'loss/train': 1.4389970004558563} 01/29/2022 08:23:10 - INFO - codeparrot_training - Step 15225: {'lr': 0.00041205074203743244, 'samples': 2923392, 'steps': 15225, 'loss/train': 1.5978124737739563} 01/29/2022 08:23:14 - INFO - codeparrot_training - Step 15226: {'lr': 0.00041203828220581805, 'samples': 2923584, 'steps': 15226, 'loss/train': 1.6960362195968628} 01/29/2022 08:23:18 - INFO - codeparrot_training - Step 15227: {'lr': 0.00041202582168008324, 'samples': 2923776, 'steps': 15227, 'loss/train': 1.5006796717643738} 01/29/2022 08:23:23 - INFO - codeparrot_training - Step 15228: {'lr': 0.00041201336046028117, 'samples': 2923968, 'steps': 15228, 'loss/train': 1.8769043684005737} 01/29/2022 08:23:27 - INFO - codeparrot_training - Step 15229: {'lr': 0.0004120008985464654, 'samples': 2924160, 'steps': 15229, 'loss/train': 1.4348432421684265} 01/29/2022 08:23:34 - INFO - codeparrot_training - Step 15230: {'lr': 0.0004119884359386891, 'samples': 2924352, 'steps': 15230, 'loss/train': 1.665434181690216} 01/29/2022 08:23:38 - INFO - codeparrot_training - Step 15231: {'lr': 0.0004119759726370058, 'samples': 2924544, 'steps': 15231, 'loss/train': 1.6084919571876526} 01/29/2022 08:23:42 - INFO - codeparrot_training - Step 15232: {'lr': 0.0004119635086414689, 'samples': 2924736, 'steps': 15232, 'loss/train': 0.712407723069191} 01/29/2022 08:23:47 - INFO - codeparrot_training - Step 15233: {'lr': 0.0004119510439521318, 'samples': 2924928, 'steps': 15233, 'loss/train': 1.7936153411865234} 01/29/2022 08:23:51 - INFO - codeparrot_training - Step 15234: {'lr': 0.0004119385785690478, 'samples': 2925120, 'steps': 15234, 'loss/train': 1.8187077641487122} 01/29/2022 08:23:56 - INFO - codeparrot_training - Step 15235: {'lr': 0.0004119261124922703, 'samples': 2925312, 'steps': 15235, 'loss/train': 2.145869493484497} 01/29/2022 08:24:00 - INFO - codeparrot_training - Step 15236: {'lr': 0.00041191364572185286, 'samples': 2925504, 'steps': 15236, 'loss/train': 2.111529529094696} 01/29/2022 08:24:05 - INFO - codeparrot_training - Step 15237: {'lr': 0.0004119011782578487, 'samples': 2925696, 'steps': 15237, 'loss/train': 1.5937442779541016} 01/29/2022 08:24:09 - INFO - codeparrot_training - Step 15238: {'lr': 0.00041188871010031135, 'samples': 2925888, 'steps': 15238, 'loss/train': 2.341589391231537} 01/29/2022 08:24:13 - INFO - codeparrot_training - Step 15239: {'lr': 0.0004118762412492941, 'samples': 2926080, 'steps': 15239, 'loss/train': 1.3399047553539276} 01/29/2022 08:24:18 - INFO - codeparrot_training - Step 15240: {'lr': 0.00041186377170485057, 'samples': 2926272, 'steps': 15240, 'loss/train': 1.5497767925262451} 01/29/2022 08:24:23 - INFO - codeparrot_training - Step 15241: {'lr': 0.00041185130146703387, 'samples': 2926464, 'steps': 15241, 'loss/train': 0.7225509434938431} 01/29/2022 08:24:27 - INFO - codeparrot_training - Step 15242: {'lr': 0.0004118388305358977, 'samples': 2926656, 'steps': 15242, 'loss/train': 1.4872126579284668} 01/29/2022 08:24:31 - INFO - codeparrot_training - Step 15243: {'lr': 0.0004118263589114953, 'samples': 2926848, 'steps': 15243, 'loss/train': 1.5146681070327759} 01/29/2022 08:24:35 - INFO - codeparrot_training - Step 15244: {'lr': 0.00041181388659388026, 'samples': 2927040, 'steps': 15244, 'loss/train': 2.1830785274505615} 01/29/2022 08:24:43 - INFO - codeparrot_training - Step 15245: {'lr': 0.00041180141358310586, 'samples': 2927232, 'steps': 15245, 'loss/train': 1.7768341898918152} 01/29/2022 08:24:47 - INFO - codeparrot_training - Step 15246: {'lr': 0.00041178893987922556, 'samples': 2927424, 'steps': 15246, 'loss/train': 1.7384342551231384} 01/29/2022 08:24:51 - INFO - codeparrot_training - Step 15247: {'lr': 0.0004117764654822929, 'samples': 2927616, 'steps': 15247, 'loss/train': 1.2193244397640228} 01/29/2022 08:24:56 - INFO - codeparrot_training - Step 15248: {'lr': 0.0004117639903923611, 'samples': 2927808, 'steps': 15248, 'loss/train': 2.2396881580352783} 01/29/2022 08:25:00 - INFO - codeparrot_training - Step 15249: {'lr': 0.0004117515146094838, 'samples': 2928000, 'steps': 15249, 'loss/train': 2.828302502632141} 01/29/2022 08:25:05 - INFO - codeparrot_training - Step 15250: {'lr': 0.0004117390381337144, 'samples': 2928192, 'steps': 15250, 'loss/train': 1.8149113655090332} 01/29/2022 08:25:09 - INFO - codeparrot_training - Step 15251: {'lr': 0.00041172656096510624, 'samples': 2928384, 'steps': 15251, 'loss/train': 1.0710758864879608} 01/29/2022 08:25:13 - INFO - codeparrot_training - Step 15252: {'lr': 0.0004117140831037129, 'samples': 2928576, 'steps': 15252, 'loss/train': 2.4006593227386475} 01/29/2022 08:25:18 - INFO - codeparrot_training - Step 15253: {'lr': 0.00041170160454958785, 'samples': 2928768, 'steps': 15253, 'loss/train': 1.5841596722602844} 01/29/2022 08:25:22 - INFO - codeparrot_training - Step 15254: {'lr': 0.00041168912530278434, 'samples': 2928960, 'steps': 15254, 'loss/train': 1.8690118789672852} 01/29/2022 08:25:28 - INFO - codeparrot_training - Step 15255: {'lr': 0.00041167664536335605, 'samples': 2929152, 'steps': 15255, 'loss/train': 1.609805166721344} 01/29/2022 08:25:32 - INFO - codeparrot_training - Step 15256: {'lr': 0.0004116641647313563, 'samples': 2929344, 'steps': 15256, 'loss/train': 1.4375178515911102} 01/29/2022 08:25:36 - INFO - codeparrot_training - Step 15257: {'lr': 0.00041165168340683857, 'samples': 2929536, 'steps': 15257, 'loss/train': 1.6104936003684998} 01/29/2022 08:25:40 - INFO - codeparrot_training - Step 15258: {'lr': 0.0004116392013898564, 'samples': 2929728, 'steps': 15258, 'loss/train': 1.479010909795761} 01/29/2022 08:25:45 - INFO - codeparrot_training - Step 15259: {'lr': 0.0004116267186804632, 'samples': 2929920, 'steps': 15259, 'loss/train': 1.5902377367019653} 01/29/2022 08:25:52 - INFO - codeparrot_training - Step 15260: {'lr': 0.0004116142352787125, 'samples': 2930112, 'steps': 15260, 'loss/train': 1.8174558877944946} 01/29/2022 08:25:56 - INFO - codeparrot_training - Step 15261: {'lr': 0.0004116017511846577, 'samples': 2930304, 'steps': 15261, 'loss/train': 1.1601885259151459} 01/29/2022 08:26:00 - INFO - codeparrot_training - Step 15262: {'lr': 0.00041158926639835234, 'samples': 2930496, 'steps': 15262, 'loss/train': 1.8222021460533142} 01/29/2022 08:26:05 - INFO - codeparrot_training - Step 15263: {'lr': 0.00041157678091984987, 'samples': 2930688, 'steps': 15263, 'loss/train': 1.8692291378974915} 01/29/2022 08:26:09 - INFO - codeparrot_training - Step 15264: {'lr': 0.0004115642947492038, 'samples': 2930880, 'steps': 15264, 'loss/train': 1.131101131439209} 01/29/2022 08:26:14 - INFO - codeparrot_training - Step 15265: {'lr': 0.0004115518078864675, 'samples': 2931072, 'steps': 15265, 'loss/train': 1.6278803944587708} 01/29/2022 08:26:19 - INFO - codeparrot_training - Step 15266: {'lr': 0.0004115393203316946, 'samples': 2931264, 'steps': 15266, 'loss/train': 1.5019238591194153} 01/29/2022 08:26:23 - INFO - codeparrot_training - Step 15267: {'lr': 0.00041152683208493855, 'samples': 2931456, 'steps': 15267, 'loss/train': 1.637876272201538} 01/29/2022 08:26:27 - INFO - codeparrot_training - Step 15268: {'lr': 0.0004115143431462529, 'samples': 2931648, 'steps': 15268, 'loss/train': 1.5432516932487488} 01/29/2022 08:26:31 - INFO - codeparrot_training - Step 15269: {'lr': 0.000411501853515691, 'samples': 2931840, 'steps': 15269, 'loss/train': 1.788755714893341} 01/29/2022 08:26:36 - INFO - codeparrot_training - Step 15270: {'lr': 0.00041148936319330656, 'samples': 2932032, 'steps': 15270, 'loss/train': 1.251273661851883} 01/29/2022 08:26:41 - INFO - codeparrot_training - Step 15271: {'lr': 0.0004114768721791529, 'samples': 2932224, 'steps': 15271, 'loss/train': 1.030684232711792} 01/29/2022 08:26:45 - INFO - codeparrot_training - Step 15272: {'lr': 0.00041146438047328347, 'samples': 2932416, 'steps': 15272, 'loss/train': 2.5470592975616455} 01/29/2022 08:26:49 - INFO - codeparrot_training - Step 15273: {'lr': 0.00041145188807575206, 'samples': 2932608, 'steps': 15273, 'loss/train': 2.367966949939728} 01/29/2022 08:26:54 - INFO - codeparrot_training - Step 15274: {'lr': 0.000411439394986612, 'samples': 2932800, 'steps': 15274, 'loss/train': 1.2911600768566132} 01/29/2022 08:27:01 - INFO - codeparrot_training - Step 15275: {'lr': 0.00041142690120591686, 'samples': 2932992, 'steps': 15275, 'loss/train': 1.952061116695404} 01/29/2022 08:27:05 - INFO - codeparrot_training - Step 15276: {'lr': 0.0004114144067337201, 'samples': 2933184, 'steps': 15276, 'loss/train': 1.7754866480827332} 01/29/2022 08:27:09 - INFO - codeparrot_training - Step 15277: {'lr': 0.0004114019115700752, 'samples': 2933376, 'steps': 15277, 'loss/train': 2.4849997758865356} 01/29/2022 08:27:14 - INFO - codeparrot_training - Step 15278: {'lr': 0.00041138941571503587, 'samples': 2933568, 'steps': 15278, 'loss/train': 2.254154920578003} 01/29/2022 08:27:19 - INFO - codeparrot_training - Step 15279: {'lr': 0.0004113769191686555, 'samples': 2933760, 'steps': 15279, 'loss/train': 1.3550479710102081} 01/29/2022 08:27:23 - INFO - codeparrot_training - Step 15280: {'lr': 0.00041136442193098765, 'samples': 2933952, 'steps': 15280, 'loss/train': 1.5571593046188354} 01/29/2022 08:27:27 - INFO - codeparrot_training - Step 15281: {'lr': 0.00041135192400208585, 'samples': 2934144, 'steps': 15281, 'loss/train': 2.152525842189789} 01/29/2022 08:27:32 - INFO - codeparrot_training - Step 15282: {'lr': 0.00041133942538200364, 'samples': 2934336, 'steps': 15282, 'loss/train': 0.2021542489528656} 01/29/2022 08:27:36 - INFO - codeparrot_training - Step 15283: {'lr': 0.0004113269260707946, 'samples': 2934528, 'steps': 15283, 'loss/train': 2.249054253101349} 01/29/2022 08:27:43 - INFO - codeparrot_training - Step 15284: {'lr': 0.0004113144260685122, 'samples': 2934720, 'steps': 15284, 'loss/train': 1.5502529740333557} 01/29/2022 08:27:47 - INFO - codeparrot_training - Step 15285: {'lr': 0.00041130192537521, 'samples': 2934912, 'steps': 15285, 'loss/train': 1.214540809392929} 01/29/2022 08:27:52 - INFO - codeparrot_training - Step 15286: {'lr': 0.0004112894239909416, 'samples': 2935104, 'steps': 15286, 'loss/train': 1.9395713210105896} 01/29/2022 08:27:56 - INFO - codeparrot_training - Step 15287: {'lr': 0.0004112769219157605, 'samples': 2935296, 'steps': 15287, 'loss/train': 2.1861966848373413} 01/29/2022 08:28:00 - INFO - codeparrot_training - Step 15288: {'lr': 0.00041126441914972036, 'samples': 2935488, 'steps': 15288, 'loss/train': 2.393306851387024} 01/29/2022 08:28:05 - INFO - codeparrot_training - Step 15289: {'lr': 0.00041125191569287456, 'samples': 2935680, 'steps': 15289, 'loss/train': 1.5689519047737122} 01/29/2022 08:28:09 - INFO - codeparrot_training - Step 15290: {'lr': 0.0004112394115452768, 'samples': 2935872, 'steps': 15290, 'loss/train': 1.3165604174137115} 01/29/2022 08:28:14 - INFO - codeparrot_training - Step 15291: {'lr': 0.00041122690670698054, 'samples': 2936064, 'steps': 15291, 'loss/train': 1.3100583851337433} 01/29/2022 08:28:18 - INFO - codeparrot_training - Step 15292: {'lr': 0.0004112144011780395, 'samples': 2936256, 'steps': 15292, 'loss/train': 1.4052416682243347} 01/29/2022 08:28:22 - INFO - codeparrot_training - Step 15293: {'lr': 0.00041120189495850713, 'samples': 2936448, 'steps': 15293, 'loss/train': 1.7541486024856567} 01/29/2022 08:28:27 - INFO - codeparrot_training - Step 15294: {'lr': 0.000411189388048437, 'samples': 2936640, 'steps': 15294, 'loss/train': 1.0467639863491058} 01/29/2022 08:28:32 - INFO - codeparrot_training - Step 15295: {'lr': 0.0004111768804478827, 'samples': 2936832, 'steps': 15295, 'loss/train': 2.011999726295471} 01/29/2022 08:28:36 - INFO - codeparrot_training - Step 15296: {'lr': 0.00041116437215689785, 'samples': 2937024, 'steps': 15296, 'loss/train': 1.4650488495826721} 01/29/2022 08:28:40 - INFO - codeparrot_training - Step 15297: {'lr': 0.000411151863175536, 'samples': 2937216, 'steps': 15297, 'loss/train': 1.2194892168045044} 01/29/2022 08:28:44 - INFO - codeparrot_training - Step 15298: {'lr': 0.00041113935350385074, 'samples': 2937408, 'steps': 15298, 'loss/train': 1.3220123648643494} 01/29/2022 08:28:50 - INFO - codeparrot_training - Step 15299: {'lr': 0.0004111268431418957, 'samples': 2937600, 'steps': 15299, 'loss/train': 1.1745732128620148} 01/29/2022 08:28:54 - INFO - codeparrot_training - Step 15300: {'lr': 0.0004111143320897244, 'samples': 2937792, 'steps': 15300, 'loss/train': 2.022272050380707} 01/29/2022 08:28:58 - INFO - codeparrot_training - Step 15301: {'lr': 0.0004111018203473904, 'samples': 2937984, 'steps': 15301, 'loss/train': 1.5996118783950806} 01/29/2022 08:29:03 - INFO - codeparrot_training - Step 15302: {'lr': 0.0004110893079149474, 'samples': 2938176, 'steps': 15302, 'loss/train': 1.2783848941326141} 01/29/2022 08:29:07 - INFO - codeparrot_training - Step 15303: {'lr': 0.000411076794792449, 'samples': 2938368, 'steps': 15303, 'loss/train': 0.3700369969010353} 01/29/2022 08:29:14 - INFO - codeparrot_training - Step 15304: {'lr': 0.0004110642809799487, 'samples': 2938560, 'steps': 15304, 'loss/train': 1.5327340364456177} 01/29/2022 08:29:18 - INFO - codeparrot_training - Step 15305: {'lr': 0.0004110517664775002, 'samples': 2938752, 'steps': 15305, 'loss/train': 1.4720250964164734} 01/29/2022 08:29:22 - INFO - codeparrot_training - Step 15306: {'lr': 0.00041103925128515705, 'samples': 2938944, 'steps': 15306, 'loss/train': 0.9702134728431702} 01/29/2022 08:29:27 - INFO - codeparrot_training - Step 15307: {'lr': 0.0004110267354029729, 'samples': 2939136, 'steps': 15307, 'loss/train': 0.7887586355209351} 01/29/2022 08:29:31 - INFO - codeparrot_training - Step 15308: {'lr': 0.0004110142188310013, 'samples': 2939328, 'steps': 15308, 'loss/train': 1.3245441913604736} 01/29/2022 08:29:36 - INFO - codeparrot_training - Step 15309: {'lr': 0.00041100170156929596, 'samples': 2939520, 'steps': 15309, 'loss/train': 1.7393916249275208} 01/29/2022 08:29:40 - INFO - codeparrot_training - Step 15310: {'lr': 0.0004109891836179105, 'samples': 2939712, 'steps': 15310, 'loss/train': 1.624392807483673} 01/29/2022 08:29:45 - INFO - codeparrot_training - Step 15311: {'lr': 0.0004109766649768984, 'samples': 2939904, 'steps': 15311, 'loss/train': 1.5868688821792603} 01/29/2022 08:29:49 - INFO - codeparrot_training - Step 15312: {'lr': 0.00041096414564631347, 'samples': 2940096, 'steps': 15312, 'loss/train': 1.8810635805130005} 01/29/2022 08:29:53 - INFO - codeparrot_training - Step 15313: {'lr': 0.00041095162562620915, 'samples': 2940288, 'steps': 15313, 'loss/train': 1.3824619352817535} 01/29/2022 08:29:59 - INFO - codeparrot_training - Step 15314: {'lr': 0.00041093910491663926, 'samples': 2940480, 'steps': 15314, 'loss/train': 1.8872854113578796} 01/29/2022 08:30:03 - INFO - codeparrot_training - Step 15315: {'lr': 0.0004109265835176573, 'samples': 2940672, 'steps': 15315, 'loss/train': 2.4581133127212524} 01/29/2022 08:30:07 - INFO - codeparrot_training - Step 15316: {'lr': 0.00041091406142931705, 'samples': 2940864, 'steps': 15316, 'loss/train': 1.5707994103431702} 01/29/2022 08:30:11 - INFO - codeparrot_training - Step 15317: {'lr': 0.00041090153865167196, 'samples': 2941056, 'steps': 15317, 'loss/train': 0.6789873540401459} 01/29/2022 08:30:15 - INFO - codeparrot_training - Step 15318: {'lr': 0.0004108890151847758, 'samples': 2941248, 'steps': 15318, 'loss/train': 1.2910837233066559} 01/29/2022 08:30:23 - INFO - codeparrot_training - Step 15319: {'lr': 0.0004108764910286822, 'samples': 2941440, 'steps': 15319, 'loss/train': 1.1907437145709991} 01/29/2022 08:30:27 - INFO - codeparrot_training - Step 15320: {'lr': 0.00041086396618344475, 'samples': 2941632, 'steps': 15320, 'loss/train': 1.623778223991394} 01/29/2022 08:30:31 - INFO - codeparrot_training - Step 15321: {'lr': 0.0004108514406491172, 'samples': 2941824, 'steps': 15321, 'loss/train': 0.9402571320533752} 01/29/2022 08:30:35 - INFO - codeparrot_training - Step 15322: {'lr': 0.0004108389144257531, 'samples': 2942016, 'steps': 15322, 'loss/train': 0.636827141046524} 01/29/2022 08:30:40 - INFO - codeparrot_training - Step 15323: {'lr': 0.0004108263875134062, 'samples': 2942208, 'steps': 15323, 'loss/train': 1.6730080246925354} 01/29/2022 08:30:45 - INFO - codeparrot_training - Step 15324: {'lr': 0.0004108138599121301, 'samples': 2942400, 'steps': 15324, 'loss/train': 1.6189962029457092} 01/29/2022 08:30:49 - INFO - codeparrot_training - Step 15325: {'lr': 0.00041080133162197855, 'samples': 2942592, 'steps': 15325, 'loss/train': 2.819095015525818} 01/29/2022 08:30:54 - INFO - codeparrot_training - Step 15326: {'lr': 0.0004107888026430051, 'samples': 2942784, 'steps': 15326, 'loss/train': 1.220901221036911} 01/29/2022 08:30:58 - INFO - codeparrot_training - Step 15327: {'lr': 0.0004107762729752635, 'samples': 2942976, 'steps': 15327, 'loss/train': 1.2236826717853546} 01/29/2022 08:31:02 - INFO - codeparrot_training - Step 15328: {'lr': 0.00041076374261880735, 'samples': 2943168, 'steps': 15328, 'loss/train': 0.5461159497499466} 01/29/2022 08:31:09 - INFO - codeparrot_training - Step 15329: {'lr': 0.0004107512115736904, 'samples': 2943360, 'steps': 15329, 'loss/train': 2.063129246234894} 01/29/2022 08:31:14 - INFO - codeparrot_training - Step 15330: {'lr': 0.0004107386798399664, 'samples': 2943552, 'steps': 15330, 'loss/train': 2.3139184713363647} 01/29/2022 08:31:18 - INFO - codeparrot_training - Step 15331: {'lr': 0.00041072614741768877, 'samples': 2943744, 'steps': 15331, 'loss/train': 1.955963909626007} 01/29/2022 08:31:22 - INFO - codeparrot_training - Step 15332: {'lr': 0.00041071361430691143, 'samples': 2943936, 'steps': 15332, 'loss/train': 2.0098718404769897} 01/29/2022 08:31:26 - INFO - codeparrot_training - Step 15333: {'lr': 0.00041070108050768805, 'samples': 2944128, 'steps': 15333, 'loss/train': 1.1483801901340485} 01/29/2022 08:31:32 - INFO - codeparrot_training - Step 15334: {'lr': 0.00041068854602007224, 'samples': 2944320, 'steps': 15334, 'loss/train': 1.3531901836395264} 01/29/2022 08:31:36 - INFO - codeparrot_training - Step 15335: {'lr': 0.0004106760108441177, 'samples': 2944512, 'steps': 15335, 'loss/train': 1.4855899214744568} 01/29/2022 08:31:40 - INFO - codeparrot_training - Step 15336: {'lr': 0.0004106634749798782, 'samples': 2944704, 'steps': 15336, 'loss/train': 1.7924121022224426} 01/29/2022 08:31:44 - INFO - codeparrot_training - Step 15337: {'lr': 0.0004106509384274073, 'samples': 2944896, 'steps': 15337, 'loss/train': 1.884955108165741} 01/29/2022 08:31:49 - INFO - codeparrot_training - Step 15338: {'lr': 0.0004106384011867589, 'samples': 2945088, 'steps': 15338, 'loss/train': 1.8159116506576538} 01/29/2022 08:31:54 - INFO - codeparrot_training - Step 15339: {'lr': 0.00041062586325798654, 'samples': 2945280, 'steps': 15339, 'loss/train': 0.7650327086448669} 01/29/2022 08:31:58 - INFO - codeparrot_training - Step 15340: {'lr': 0.000410613324641144, 'samples': 2945472, 'steps': 15340, 'loss/train': 0.6602612882852554} 01/29/2022 08:32:02 - INFO - codeparrot_training - Step 15341: {'lr': 0.000410600785336285, 'samples': 2945664, 'steps': 15341, 'loss/train': 1.8722571730613708} 01/29/2022 08:32:07 - INFO - codeparrot_training - Step 15342: {'lr': 0.0004105882453434632, 'samples': 2945856, 'steps': 15342, 'loss/train': 2.1032389998435974} 01/29/2022 08:32:11 - INFO - codeparrot_training - Step 15343: {'lr': 0.0004105757046627323, 'samples': 2946048, 'steps': 15343, 'loss/train': 1.7843091487884521} 01/29/2022 08:32:19 - INFO - codeparrot_training - Step 15344: {'lr': 0.00041056316329414613, 'samples': 2946240, 'steps': 15344, 'loss/train': 1.6143223643302917} 01/29/2022 08:32:23 - INFO - codeparrot_training - Step 15345: {'lr': 0.0004105506212377583, 'samples': 2946432, 'steps': 15345, 'loss/train': 1.6731337308883667} 01/29/2022 08:32:27 - INFO - codeparrot_training - Step 15346: {'lr': 0.0004105380784936227, 'samples': 2946624, 'steps': 15346, 'loss/train': 1.2519668340682983} 01/29/2022 08:32:31 - INFO - codeparrot_training - Step 15347: {'lr': 0.0004105255350617928, 'samples': 2946816, 'steps': 15347, 'loss/train': 1.7845573425292969} 01/29/2022 08:32:36 - INFO - codeparrot_training - Step 15348: {'lr': 0.0004105129909423226, 'samples': 2947008, 'steps': 15348, 'loss/train': 0.3167693316936493} 01/29/2022 08:32:41 - INFO - codeparrot_training - Step 15349: {'lr': 0.0004105004461352657, 'samples': 2947200, 'steps': 15349, 'loss/train': 2.040864944458008} 01/29/2022 08:32:45 - INFO - codeparrot_training - Step 15350: {'lr': 0.00041048790064067577, 'samples': 2947392, 'steps': 15350, 'loss/train': 1.9263399839401245} 01/29/2022 08:32:49 - INFO - codeparrot_training - Step 15351: {'lr': 0.0004104753544586067, 'samples': 2947584, 'steps': 15351, 'loss/train': 2.149373173713684} 01/29/2022 08:32:54 - INFO - codeparrot_training - Step 15352: {'lr': 0.0004104628075891121, 'samples': 2947776, 'steps': 15352, 'loss/train': 1.5879315733909607} 01/29/2022 08:32:58 - INFO - codeparrot_training - Step 15353: {'lr': 0.00041045026003224593, 'samples': 2947968, 'steps': 15353, 'loss/train': 1.6160028576850891} 01/29/2022 08:33:03 - INFO - codeparrot_training - Step 15354: {'lr': 0.00041043771178806164, 'samples': 2948160, 'steps': 15354, 'loss/train': 1.4156005382537842} 01/29/2022 08:33:07 - INFO - codeparrot_training - Step 15355: {'lr': 0.00041042516285661325, 'samples': 2948352, 'steps': 15355, 'loss/train': 1.0238318145275116} 01/29/2022 08:33:12 - INFO - codeparrot_training - Step 15356: {'lr': 0.00041041261323795437, 'samples': 2948544, 'steps': 15356, 'loss/train': 3.490318536758423} 01/29/2022 08:33:16 - INFO - codeparrot_training - Step 15357: {'lr': 0.00041040006293213883, 'samples': 2948736, 'steps': 15357, 'loss/train': 1.8137549757957458} 01/29/2022 08:33:20 - INFO - codeparrot_training - Step 15358: {'lr': 0.0004103875119392203, 'samples': 2948928, 'steps': 15358, 'loss/train': 1.6686891317367554} 01/29/2022 08:33:25 - INFO - codeparrot_training - Step 15359: {'lr': 0.00041037496025925256, 'samples': 2949120, 'steps': 15359, 'loss/train': 1.8576632738113403} 01/29/2022 08:33:30 - INFO - codeparrot_training - Step 15360: {'lr': 0.0004103624078922895, 'samples': 2949312, 'steps': 15360, 'loss/train': 1.951171338558197} 01/29/2022 08:33:34 - INFO - codeparrot_training - Step 15361: {'lr': 0.0004103498548383847, 'samples': 2949504, 'steps': 15361, 'loss/train': 1.174238920211792} 01/29/2022 08:33:38 - INFO - codeparrot_training - Step 15362: {'lr': 0.00041033730109759216, 'samples': 2949696, 'steps': 15362, 'loss/train': 1.4406657814979553} 01/29/2022 08:33:42 - INFO - codeparrot_training - Step 15363: {'lr': 0.00041032474666996544, 'samples': 2949888, 'steps': 15363, 'loss/train': 1.435492068529129} 01/29/2022 08:33:49 - INFO - codeparrot_training - Step 15364: {'lr': 0.0004103121915555585, 'samples': 2950080, 'steps': 15364, 'loss/train': 2.4006773829460144} 01/29/2022 08:33:54 - INFO - codeparrot_training - Step 15365: {'lr': 0.00041029963575442494, 'samples': 2950272, 'steps': 15365, 'loss/train': 1.3469891846179962} 01/29/2022 08:33:58 - INFO - codeparrot_training - Step 15366: {'lr': 0.0004102870792666187, 'samples': 2950464, 'steps': 15366, 'loss/train': 1.3226274847984314} 01/29/2022 08:34:02 - INFO - codeparrot_training - Step 15367: {'lr': 0.0004102745220921935, 'samples': 2950656, 'steps': 15367, 'loss/train': 1.032450020313263} 01/29/2022 08:34:06 - INFO - codeparrot_training - Step 15368: {'lr': 0.0004102619642312031, 'samples': 2950848, 'steps': 15368, 'loss/train': 1.2640847861766815} 01/29/2022 08:34:12 - INFO - codeparrot_training - Step 15369: {'lr': 0.0004102494056837014, 'samples': 2951040, 'steps': 15369, 'loss/train': 0.8208642303943634} 01/29/2022 08:34:16 - INFO - codeparrot_training - Step 15370: {'lr': 0.00041023684644974213, 'samples': 2951232, 'steps': 15370, 'loss/train': 1.5897061228752136} 01/29/2022 08:34:20 - INFO - codeparrot_training - Step 15371: {'lr': 0.00041022428652937905, 'samples': 2951424, 'steps': 15371, 'loss/train': 1.6851691603660583} 01/29/2022 08:34:25 - INFO - codeparrot_training - Step 15372: {'lr': 0.000410211725922666, 'samples': 2951616, 'steps': 15372, 'loss/train': 2.1033833026885986} 01/29/2022 08:34:29 - INFO - codeparrot_training - Step 15373: {'lr': 0.00041019916462965684, 'samples': 2951808, 'steps': 15373, 'loss/train': 1.3419556617736816} 01/29/2022 08:34:36 - INFO - codeparrot_training - Step 15374: {'lr': 0.0004101866026504053, 'samples': 2952000, 'steps': 15374, 'loss/train': 1.3225377202033997} 01/29/2022 08:34:40 - INFO - codeparrot_training - Step 15375: {'lr': 0.00041017403998496523, 'samples': 2952192, 'steps': 15375, 'loss/train': 1.230318546295166} 01/29/2022 08:34:44 - INFO - codeparrot_training - Step 15376: {'lr': 0.0004101614766333904, 'samples': 2952384, 'steps': 15376, 'loss/train': 1.1400379836559296} 01/29/2022 08:34:49 - INFO - codeparrot_training - Step 15377: {'lr': 0.0004101489125957347, 'samples': 2952576, 'steps': 15377, 'loss/train': 1.8290455341339111} 01/29/2022 08:34:53 - INFO - codeparrot_training - Step 15378: {'lr': 0.0004101363478720519, 'samples': 2952768, 'steps': 15378, 'loss/train': 2.1048683524131775} 01/29/2022 08:34:58 - INFO - codeparrot_training - Step 15379: {'lr': 0.0004101237824623958, 'samples': 2952960, 'steps': 15379, 'loss/train': 0.7828798592090607} 01/29/2022 08:35:02 - INFO - codeparrot_training - Step 15380: {'lr': 0.00041011121636682024, 'samples': 2953152, 'steps': 15380, 'loss/train': 1.1936440765857697} 01/29/2022 08:35:07 - INFO - codeparrot_training - Step 15381: {'lr': 0.0004100986495853791, 'samples': 2953344, 'steps': 15381, 'loss/train': 1.6827728748321533} 01/29/2022 08:35:11 - INFO - codeparrot_training - Step 15382: {'lr': 0.00041008608211812625, 'samples': 2953536, 'steps': 15382, 'loss/train': 1.6905911564826965} 01/29/2022 08:35:15 - INFO - codeparrot_training - Step 15383: {'lr': 0.00041007351396511537, 'samples': 2953728, 'steps': 15383, 'loss/train': 2.1410431265830994} 01/29/2022 08:35:20 - INFO - codeparrot_training - Step 15384: {'lr': 0.00041006094512640044, 'samples': 2953920, 'steps': 15384, 'loss/train': 0.9778372943401337} 01/29/2022 08:35:25 - INFO - codeparrot_training - Step 15385: {'lr': 0.00041004837560203525, 'samples': 2954112, 'steps': 15385, 'loss/train': 1.987266182899475} 01/29/2022 08:35:29 - INFO - codeparrot_training - Step 15386: {'lr': 0.0004100358053920736, 'samples': 2954304, 'steps': 15386, 'loss/train': 2.449130594730377} 01/29/2022 08:35:33 - INFO - codeparrot_training - Step 15387: {'lr': 0.00041002323449656943, 'samples': 2954496, 'steps': 15387, 'loss/train': 2.0006949305534363} 01/29/2022 08:35:37 - INFO - codeparrot_training - Step 15388: {'lr': 0.00041001066291557653, 'samples': 2954688, 'steps': 15388, 'loss/train': 1.6950727701187134} 01/29/2022 08:35:44 - INFO - codeparrot_training - Step 15389: {'lr': 0.0004099980906491487, 'samples': 2954880, 'steps': 15389, 'loss/train': 0.2468755915760994} 01/29/2022 08:35:49 - INFO - codeparrot_training - Step 15390: {'lr': 0.0004099855176973399, 'samples': 2955072, 'steps': 15390, 'loss/train': 1.0579480826854706} 01/29/2022 08:35:53 - INFO - codeparrot_training - Step 15391: {'lr': 0.0004099729440602039, 'samples': 2955264, 'steps': 15391, 'loss/train': 1.486814260482788} 01/29/2022 08:35:57 - INFO - codeparrot_training - Step 15392: {'lr': 0.0004099603697377946, 'samples': 2955456, 'steps': 15392, 'loss/train': 1.9538261890411377} 01/29/2022 08:36:01 - INFO - codeparrot_training - Step 15393: {'lr': 0.000409947794730166, 'samples': 2955648, 'steps': 15393, 'loss/train': 1.4606100618839264} 01/29/2022 08:36:07 - INFO - codeparrot_training - Step 15394: {'lr': 0.0004099352190373716, 'samples': 2955840, 'steps': 15394, 'loss/train': 2.3461888432502747} 01/29/2022 08:36:11 - INFO - codeparrot_training - Step 15395: {'lr': 0.0004099226426594657, 'samples': 2956032, 'steps': 15395, 'loss/train': 2.135383665561676} 01/29/2022 08:36:15 - INFO - codeparrot_training - Step 15396: {'lr': 0.0004099100655965019, 'samples': 2956224, 'steps': 15396, 'loss/train': 2.302512288093567} 01/29/2022 08:36:20 - INFO - codeparrot_training - Step 15397: {'lr': 0.0004098974878485342, 'samples': 2956416, 'steps': 15397, 'loss/train': 1.6408370733261108} 01/29/2022 08:36:24 - INFO - codeparrot_training - Step 15398: {'lr': 0.0004098849094156164, 'samples': 2956608, 'steps': 15398, 'loss/train': 0.43669863045215607} 01/29/2022 08:36:31 - INFO - codeparrot_training - Step 15399: {'lr': 0.0004098723302978025, 'samples': 2956800, 'steps': 15399, 'loss/train': 1.572043240070343} 01/29/2022 08:36:35 - INFO - codeparrot_training - Step 15400: {'lr': 0.00040985975049514617, 'samples': 2956992, 'steps': 15400, 'loss/train': 1.282497078180313} 01/29/2022 08:36:39 - INFO - codeparrot_training - Step 15401: {'lr': 0.00040984717000770157, 'samples': 2957184, 'steps': 15401, 'loss/train': 1.323259323835373} 01/29/2022 08:36:44 - INFO - codeparrot_training - Step 15402: {'lr': 0.00040983458883552237, 'samples': 2957376, 'steps': 15402, 'loss/train': 0.6949165016412735} 01/29/2022 08:36:48 - INFO - codeparrot_training - Step 15403: {'lr': 0.00040982200697866256, 'samples': 2957568, 'steps': 15403, 'loss/train': 0.5321265310049057} 01/29/2022 08:36:54 - INFO - codeparrot_training - Step 15404: {'lr': 0.00040980942443717596, 'samples': 2957760, 'steps': 15404, 'loss/train': 1.7723248600959778} 01/29/2022 08:36:58 - INFO - codeparrot_training - Step 15405: {'lr': 0.0004097968412111166, 'samples': 2957952, 'steps': 15405, 'loss/train': 2.5496528148651123} 01/29/2022 08:37:02 - INFO - codeparrot_training - Step 15406: {'lr': 0.0004097842573005383, 'samples': 2958144, 'steps': 15406, 'loss/train': 1.6192429661750793} 01/29/2022 08:37:06 - INFO - codeparrot_training - Step 15407: {'lr': 0.000409771672705495, 'samples': 2958336, 'steps': 15407, 'loss/train': 1.6542379260063171} 01/29/2022 08:37:11 - INFO - codeparrot_training - Step 15408: {'lr': 0.0004097590874260405, 'samples': 2958528, 'steps': 15408, 'loss/train': 0.9979343712329865} 01/29/2022 08:37:16 - INFO - codeparrot_training - Step 15409: {'lr': 0.0004097465014622289, 'samples': 2958720, 'steps': 15409, 'loss/train': 1.5639792680740356} 01/29/2022 08:37:20 - INFO - codeparrot_training - Step 15410: {'lr': 0.00040973391481411396, 'samples': 2958912, 'steps': 15410, 'loss/train': 1.6345760822296143} 01/29/2022 08:37:25 - INFO - codeparrot_training - Step 15411: {'lr': 0.00040972132748174966, 'samples': 2959104, 'steps': 15411, 'loss/train': 1.5475573539733887} 01/29/2022 08:37:29 - INFO - codeparrot_training - Step 15412: {'lr': 0.00040970873946518993, 'samples': 2959296, 'steps': 15412, 'loss/train': 0.980728805065155} 01/29/2022 08:37:34 - INFO - codeparrot_training - Step 15413: {'lr': 0.00040969615076448865, 'samples': 2959488, 'steps': 15413, 'loss/train': 1.8013531565666199} 01/29/2022 08:37:39 - INFO - codeparrot_training - Step 15414: {'lr': 0.0004096835613796998, 'samples': 2959680, 'steps': 15414, 'loss/train': 1.2773078083992004} 01/29/2022 08:37:43 - INFO - codeparrot_training - Step 15415: {'lr': 0.00040967097131087727, 'samples': 2959872, 'steps': 15415, 'loss/train': 1.2530645728111267} 01/29/2022 08:37:47 - INFO - codeparrot_training - Step 15416: {'lr': 0.00040965838055807493, 'samples': 2960064, 'steps': 15416, 'loss/train': 1.5899083614349365} 01/29/2022 08:37:51 - INFO - codeparrot_training - Step 15417: {'lr': 0.00040964578912134687, 'samples': 2960256, 'steps': 15417, 'loss/train': 0.866601973772049} 01/29/2022 08:37:59 - INFO - codeparrot_training - Step 15418: {'lr': 0.00040963319700074684, 'samples': 2960448, 'steps': 15418, 'loss/train': 3.1060802936553955} 01/29/2022 08:38:03 - INFO - codeparrot_training - Step 15419: {'lr': 0.00040962060419632906, 'samples': 2960640, 'steps': 15419, 'loss/train': 1.949350655078888} 01/29/2022 08:38:07 - INFO - codeparrot_training - Step 15420: {'lr': 0.00040960801070814715, 'samples': 2960832, 'steps': 15420, 'loss/train': 1.3234752416610718} 01/29/2022 08:38:11 - INFO - codeparrot_training - Step 15421: {'lr': 0.00040959541653625526, 'samples': 2961024, 'steps': 15421, 'loss/train': 2.2260478734970093} 01/29/2022 08:38:16 - INFO - codeparrot_training - Step 15422: {'lr': 0.0004095828216807073, 'samples': 2961216, 'steps': 15422, 'loss/train': 1.5583266019821167} 01/29/2022 08:38:21 - INFO - codeparrot_training - Step 15423: {'lr': 0.00040957022614155714, 'samples': 2961408, 'steps': 15423, 'loss/train': 2.2539551854133606} 01/29/2022 08:38:25 - INFO - codeparrot_training - Step 15424: {'lr': 0.0004095576299188589, 'samples': 2961600, 'steps': 15424, 'loss/train': 2.350996971130371} 01/29/2022 08:38:29 - INFO - codeparrot_training - Step 15425: {'lr': 0.0004095450330126663, 'samples': 2961792, 'steps': 15425, 'loss/train': 1.2624067068099976} 01/29/2022 08:38:34 - INFO - codeparrot_training - Step 15426: {'lr': 0.0004095324354230335, 'samples': 2961984, 'steps': 15426, 'loss/train': 2.5116376876831055} 01/29/2022 08:38:38 - INFO - codeparrot_training - Step 15427: {'lr': 0.0004095198371500145, 'samples': 2962176, 'steps': 15427, 'loss/train': 1.9128334522247314} 01/29/2022 08:38:43 - INFO - codeparrot_training - Step 15428: {'lr': 0.00040950723819366307, 'samples': 2962368, 'steps': 15428, 'loss/train': 1.661276400089264} 01/29/2022 08:38:48 - INFO - codeparrot_training - Step 15429: {'lr': 0.00040949463855403326, 'samples': 2962560, 'steps': 15429, 'loss/train': 1.512509286403656} 01/29/2022 08:38:52 - INFO - codeparrot_training - Step 15430: {'lr': 0.00040948203823117915, 'samples': 2962752, 'steps': 15430, 'loss/train': 1.2765282690525055} 01/29/2022 08:38:57 - INFO - codeparrot_training - Step 15431: {'lr': 0.00040946943722515455, 'samples': 2962944, 'steps': 15431, 'loss/train': 1.679790437221527} 01/29/2022 08:39:01 - INFO - codeparrot_training - Step 15432: {'lr': 0.0004094568355360135, 'samples': 2963136, 'steps': 15432, 'loss/train': 0.8619175851345062} 01/29/2022 08:39:05 - INFO - codeparrot_training - Step 15433: {'lr': 0.00040944423316381006, 'samples': 2963328, 'steps': 15433, 'loss/train': 1.7664377689361572} 01/29/2022 08:39:12 - INFO - codeparrot_training - Step 15434: {'lr': 0.0004094316301085982, 'samples': 2963520, 'steps': 15434, 'loss/train': 0.9328490495681763} 01/29/2022 08:39:16 - INFO - codeparrot_training - Step 15435: {'lr': 0.00040941902637043183, 'samples': 2963712, 'steps': 15435, 'loss/train': 2.1331087946891785} 01/29/2022 08:39:20 - INFO - codeparrot_training - Step 15436: {'lr': 0.00040940642194936495, 'samples': 2963904, 'steps': 15436, 'loss/train': 1.2866670191287994} 01/29/2022 08:39:25 - INFO - codeparrot_training - Step 15437: {'lr': 0.0004093938168454515, 'samples': 2964096, 'steps': 15437, 'loss/train': 1.7617746591567993} 01/29/2022 08:39:29 - INFO - codeparrot_training - Step 15438: {'lr': 0.00040938121105874573, 'samples': 2964288, 'steps': 15438, 'loss/train': 1.66194087266922} 01/29/2022 08:39:35 - INFO - codeparrot_training - Step 15439: {'lr': 0.0004093686045893013, 'samples': 2964480, 'steps': 15439, 'loss/train': 1.8829373717308044} 01/29/2022 08:39:39 - INFO - codeparrot_training - Step 15440: {'lr': 0.00040935599743717243, 'samples': 2964672, 'steps': 15440, 'loss/train': 1.9425449967384338} 01/29/2022 08:39:43 - INFO - codeparrot_training - Step 15441: {'lr': 0.00040934338960241305, 'samples': 2964864, 'steps': 15441, 'loss/train': 2.5912182927131653} 01/29/2022 08:39:47 - INFO - codeparrot_training - Step 15442: {'lr': 0.00040933078108507727, 'samples': 2965056, 'steps': 15442, 'loss/train': 3.192289710044861} 01/29/2022 08:39:54 - INFO - codeparrot_training - Step 15443: {'lr': 0.00040931817188521894, 'samples': 2965248, 'steps': 15443, 'loss/train': 0.9504202902317047} 01/29/2022 08:39:58 - INFO - codeparrot_training - Step 15444: {'lr': 0.00040930556200289214, 'samples': 2965440, 'steps': 15444, 'loss/train': 1.2454623878002167} 01/29/2022 08:40:03 - INFO - codeparrot_training - Step 15445: {'lr': 0.00040929295143815093, 'samples': 2965632, 'steps': 15445, 'loss/train': 0.9283797740936279} 01/29/2022 08:40:07 - INFO - codeparrot_training - Step 15446: {'lr': 0.0004092803401910493, 'samples': 2965824, 'steps': 15446, 'loss/train': 0.6395350992679596} 01/29/2022 08:40:11 - INFO - codeparrot_training - Step 15447: {'lr': 0.00040926772826164126, 'samples': 2966016, 'steps': 15447, 'loss/train': 2.016853451728821} 01/29/2022 08:40:17 - INFO - codeparrot_training - Step 15448: {'lr': 0.0004092551156499809, 'samples': 2966208, 'steps': 15448, 'loss/train': 2.29189270734787} 01/29/2022 08:40:21 - INFO - codeparrot_training - Step 15449: {'lr': 0.000409242502356122, 'samples': 2966400, 'steps': 15449, 'loss/train': 1.4455103874206543} 01/29/2022 08:40:25 - INFO - codeparrot_training - Step 15450: {'lr': 0.000409229888380119, 'samples': 2966592, 'steps': 15450, 'loss/train': 1.0626650154590607} 01/29/2022 08:40:30 - INFO - codeparrot_training - Step 15451: {'lr': 0.00040921727372202565, 'samples': 2966784, 'steps': 15451, 'loss/train': 1.341075360774994} 01/29/2022 08:40:34 - INFO - codeparrot_training - Step 15452: {'lr': 0.000409204658381896, 'samples': 2966976, 'steps': 15452, 'loss/train': 2.7533708810806274} 01/29/2022 08:40:39 - INFO - codeparrot_training - Step 15453: {'lr': 0.00040919204235978425, 'samples': 2967168, 'steps': 15453, 'loss/train': 1.6698174476623535} 01/29/2022 08:40:43 - INFO - codeparrot_training - Step 15454: {'lr': 0.0004091794256557443, 'samples': 2967360, 'steps': 15454, 'loss/train': 1.8401357531547546} 01/29/2022 08:40:48 - INFO - codeparrot_training - Step 15455: {'lr': 0.00040916680826983017, 'samples': 2967552, 'steps': 15455, 'loss/train': 1.1278643310070038} 01/29/2022 08:40:52 - INFO - codeparrot_training - Step 15456: {'lr': 0.00040915419020209605, 'samples': 2967744, 'steps': 15456, 'loss/train': 3.355304718017578} 01/29/2022 08:40:56 - INFO - codeparrot_training - Step 15457: {'lr': 0.0004091415714525959, 'samples': 2967936, 'steps': 15457, 'loss/train': 1.3957541584968567} 01/29/2022 08:41:03 - INFO - codeparrot_training - Step 15458: {'lr': 0.0004091289520213838, 'samples': 2968128, 'steps': 15458, 'loss/train': 2.177289068698883} 01/29/2022 08:41:07 - INFO - codeparrot_training - Step 15459: {'lr': 0.0004091163319085137, 'samples': 2968320, 'steps': 15459, 'loss/train': 2.1686933040618896} 01/29/2022 08:41:12 - INFO - codeparrot_training - Step 15460: {'lr': 0.0004091037111140399, 'samples': 2968512, 'steps': 15460, 'loss/train': 1.7054198384284973} 01/29/2022 08:41:16 - INFO - codeparrot_training - Step 15461: {'lr': 0.00040909108963801624, 'samples': 2968704, 'steps': 15461, 'loss/train': 1.9887481927871704} 01/29/2022 08:41:20 - INFO - codeparrot_training - Step 15462: {'lr': 0.0004090784674804969, 'samples': 2968896, 'steps': 15462, 'loss/train': 1.5241751074790955} 01/29/2022 08:41:25 - INFO - codeparrot_training - Step 15463: {'lr': 0.0004090658446415359, 'samples': 2969088, 'steps': 15463, 'loss/train': 2.26199072599411} 01/29/2022 08:41:30 - INFO - codeparrot_training - Step 15464: {'lr': 0.0004090532211211874, 'samples': 2969280, 'steps': 15464, 'loss/train': 1.589154303073883} 01/29/2022 08:41:34 - INFO - codeparrot_training - Step 15465: {'lr': 0.0004090405969195053, 'samples': 2969472, 'steps': 15465, 'loss/train': 1.3242074847221375} 01/29/2022 08:41:38 - INFO - codeparrot_training - Step 15466: {'lr': 0.0004090279720365438, 'samples': 2969664, 'steps': 15466, 'loss/train': 1.4339764416217804} 01/29/2022 08:41:43 - INFO - codeparrot_training - Step 15467: {'lr': 0.00040901534647235703, 'samples': 2969856, 'steps': 15467, 'loss/train': 3.0806912183761597} 01/29/2022 08:41:48 - INFO - codeparrot_training - Step 15468: {'lr': 0.00040900272022699897, 'samples': 2970048, 'steps': 15468, 'loss/train': 1.1594054102897644} 01/29/2022 08:41:52 - INFO - codeparrot_training - Step 15469: {'lr': 0.00040899009330052375, 'samples': 2970240, 'steps': 15469, 'loss/train': 1.4095110297203064} 01/29/2022 08:41:56 - INFO - codeparrot_training - Step 15470: {'lr': 0.00040897746569298546, 'samples': 2970432, 'steps': 15470, 'loss/train': 1.839836061000824} 01/29/2022 08:42:01 - INFO - codeparrot_training - Step 15471: {'lr': 0.0004089648374044382, 'samples': 2970624, 'steps': 15471, 'loss/train': 0.7342821210622787} 01/29/2022 08:42:05 - INFO - codeparrot_training - Step 15472: {'lr': 0.000408952208434936, 'samples': 2970816, 'steps': 15472, 'loss/train': 1.3422144949436188} 01/29/2022 08:42:10 - INFO - codeparrot_training - Step 15473: {'lr': 0.00040893957878453314, 'samples': 2971008, 'steps': 15473, 'loss/train': 1.1508747339248657} 01/29/2022 08:42:14 - INFO - codeparrot_training - Step 15474: {'lr': 0.0004089269484532834, 'samples': 2971200, 'steps': 15474, 'loss/train': 2.1163968443870544} 01/29/2022 08:42:18 - INFO - codeparrot_training - Step 15475: {'lr': 0.00040891431744124123, 'samples': 2971392, 'steps': 15475, 'loss/train': 1.1306232511997223} 01/29/2022 08:42:23 - INFO - codeparrot_training - Step 15476: {'lr': 0.00040890168574846055, 'samples': 2971584, 'steps': 15476, 'loss/train': 1.773357331752777} 01/29/2022 08:42:27 - INFO - codeparrot_training - Step 15477: {'lr': 0.0004088890533749955, 'samples': 2971776, 'steps': 15477, 'loss/train': 2.156071186065674} 01/29/2022 08:42:33 - INFO - codeparrot_training - Step 15478: {'lr': 0.0004088764203209002, 'samples': 2971968, 'steps': 15478, 'loss/train': 2.9400141835212708} 01/29/2022 08:42:38 - INFO - codeparrot_training - Step 15479: {'lr': 0.0004088637865862287, 'samples': 2972160, 'steps': 15479, 'loss/train': 1.5281381607055664} 01/29/2022 08:42:42 - INFO - codeparrot_training - Step 15480: {'lr': 0.0004088511521710352, 'samples': 2972352, 'steps': 15480, 'loss/train': 1.7017706036567688} 01/29/2022 08:42:46 - INFO - codeparrot_training - Step 15481: {'lr': 0.0004088385170753739, 'samples': 2972544, 'steps': 15481, 'loss/train': 1.3485442399978638} 01/29/2022 08:42:50 - INFO - codeparrot_training - Step 15482: {'lr': 0.00040882588129929876, 'samples': 2972736, 'steps': 15482, 'loss/train': 1.4264526665210724} 01/29/2022 08:42:56 - INFO - codeparrot_training - Step 15483: {'lr': 0.000408813244842864, 'samples': 2972928, 'steps': 15483, 'loss/train': 0.06673642247915268} 01/29/2022 08:43:00 - INFO - codeparrot_training - Step 15484: {'lr': 0.0004088006077061237, 'samples': 2973120, 'steps': 15484, 'loss/train': 2.21416836977005} 01/29/2022 08:43:04 - INFO - codeparrot_training - Step 15485: {'lr': 0.00040878796988913204, 'samples': 2973312, 'steps': 15485, 'loss/train': 0.6175428181886673} 01/29/2022 08:43:08 - INFO - codeparrot_training - Step 15486: {'lr': 0.00040877533139194313, 'samples': 2973504, 'steps': 15486, 'loss/train': 1.20223268866539} 01/29/2022 08:43:13 - INFO - codeparrot_training - Step 15487: {'lr': 0.00040876269221461117, 'samples': 2973696, 'steps': 15487, 'loss/train': 1.6846375465393066} 01/29/2022 08:43:17 - INFO - codeparrot_training - Step 15488: {'lr': 0.0004087500523571902, 'samples': 2973888, 'steps': 15488, 'loss/train': 1.3561195135116577} 01/29/2022 08:43:22 - INFO - codeparrot_training - Step 15489: {'lr': 0.0004087374118197344, 'samples': 2974080, 'steps': 15489, 'loss/train': 1.9062670469284058} 01/29/2022 08:43:27 - INFO - codeparrot_training - Step 15490: {'lr': 0.00040872477060229797, 'samples': 2974272, 'steps': 15490, 'loss/train': 2.048014998435974} 01/29/2022 08:43:31 - INFO - codeparrot_training - Step 15491: {'lr': 0.00040871212870493504, 'samples': 2974464, 'steps': 15491, 'loss/train': 1.8905104994773865} 01/29/2022 08:43:35 - INFO - codeparrot_training - Step 15492: {'lr': 0.0004086994861276996, 'samples': 2974656, 'steps': 15492, 'loss/train': 2.733356237411499} 01/29/2022 08:43:39 - INFO - codeparrot_training - Step 15493: {'lr': 0.00040868684287064617, 'samples': 2974848, 'steps': 15493, 'loss/train': 1.7940409183502197} 01/29/2022 08:43:46 - INFO - codeparrot_training - Step 15494: {'lr': 0.0004086741989338285, 'samples': 2975040, 'steps': 15494, 'loss/train': 0.8614713549613953} 01/29/2022 08:43:50 - INFO - codeparrot_training - Step 15495: {'lr': 0.0004086615543173011, 'samples': 2975232, 'steps': 15495, 'loss/train': 1.793753743171692} 01/29/2022 08:43:55 - INFO - codeparrot_training - Step 15496: {'lr': 0.0004086489090211178, 'samples': 2975424, 'steps': 15496, 'loss/train': 1.3300332427024841} 01/29/2022 08:43:59 - INFO - codeparrot_training - Step 15497: {'lr': 0.00040863626304533316, 'samples': 2975616, 'steps': 15497, 'loss/train': 1.6138465404510498} 01/29/2022 08:44:03 - INFO - codeparrot_training - Step 15498: {'lr': 0.000408623616390001, 'samples': 2975808, 'steps': 15498, 'loss/train': 0.5747692286968231} 01/29/2022 08:44:09 - INFO - codeparrot_training - Step 15499: {'lr': 0.00040861096905517574, 'samples': 2976000, 'steps': 15499, 'loss/train': 1.1361396610736847} 01/29/2022 08:44:13 - INFO - codeparrot_training - Step 15500: {'lr': 0.0004085983210409114, 'samples': 2976192, 'steps': 15500, 'loss/train': 1.8723324537277222} 01/29/2022 08:44:17 - INFO - codeparrot_training - Step 15501: {'lr': 0.00040858567234726217, 'samples': 2976384, 'steps': 15501, 'loss/train': 2.3082666993141174} 01/29/2022 08:44:22 - INFO - codeparrot_training - Step 15502: {'lr': 0.00040857302297428233, 'samples': 2976576, 'steps': 15502, 'loss/train': 2.2635439038276672} 01/29/2022 08:44:29 - INFO - codeparrot_training - Step 15503: {'lr': 0.000408560372922026, 'samples': 2976768, 'steps': 15503, 'loss/train': 1.6806528568267822} 01/29/2022 08:44:33 - INFO - codeparrot_training - Step 15504: {'lr': 0.00040854772219054737, 'samples': 2976960, 'steps': 15504, 'loss/train': 1.5643306374549866} 01/29/2022 08:44:37 - INFO - codeparrot_training - Step 15505: {'lr': 0.00040853507077990073, 'samples': 2977152, 'steps': 15505, 'loss/train': 1.7380737662315369} 01/29/2022 08:44:41 - INFO - codeparrot_training - Step 15506: {'lr': 0.00040852241869014004, 'samples': 2977344, 'steps': 15506, 'loss/train': 1.9474210739135742} 01/29/2022 08:44:46 - INFO - codeparrot_training - Step 15507: {'lr': 0.00040850976592131974, 'samples': 2977536, 'steps': 15507, 'loss/train': 0.8696832060813904} 01/29/2022 08:44:51 - INFO - codeparrot_training - Step 15508: {'lr': 0.0004084971124734939, 'samples': 2977728, 'steps': 15508, 'loss/train': 1.7797858715057373} 01/29/2022 08:44:55 - INFO - codeparrot_training - Step 15509: {'lr': 0.0004084844583467168, 'samples': 2977920, 'steps': 15509, 'loss/train': 1.8263754844665527} 01/29/2022 08:44:59 - INFO - codeparrot_training - Step 15510: {'lr': 0.00040847180354104256, 'samples': 2978112, 'steps': 15510, 'loss/train': 1.9529528617858887} 01/29/2022 08:45:04 - INFO - codeparrot_training - Step 15511: {'lr': 0.00040845914805652544, 'samples': 2978304, 'steps': 15511, 'loss/train': 1.5512773990631104} 01/29/2022 08:45:08 - INFO - codeparrot_training - Step 15512: {'lr': 0.0004084464918932197, 'samples': 2978496, 'steps': 15512, 'loss/train': 1.2130476236343384} 01/29/2022 08:45:13 - INFO - codeparrot_training - Step 15513: {'lr': 0.0004084338350511795, 'samples': 2978688, 'steps': 15513, 'loss/train': 3.5196722745895386} 01/29/2022 08:45:17 - INFO - codeparrot_training - Step 15514: {'lr': 0.00040842117753045893, 'samples': 2978880, 'steps': 15514, 'loss/train': 2.0302155017852783} 01/29/2022 08:45:22 - INFO - codeparrot_training - Step 15515: {'lr': 0.0004084085193311124, 'samples': 2979072, 'steps': 15515, 'loss/train': 1.891713559627533} 01/29/2022 08:45:26 - INFO - codeparrot_training - Step 15516: {'lr': 0.0004083958604531941, 'samples': 2979264, 'steps': 15516, 'loss/train': 1.784426987171173} 01/29/2022 08:45:30 - INFO - codeparrot_training - Step 15517: {'lr': 0.0004083832008967583, 'samples': 2979456, 'steps': 15517, 'loss/train': 0.8925789892673492} 01/29/2022 08:45:37 - INFO - codeparrot_training - Step 15518: {'lr': 0.00040837054066185906, 'samples': 2979648, 'steps': 15518, 'loss/train': 1.4026641547679901} 01/29/2022 08:45:41 - INFO - codeparrot_training - Step 15519: {'lr': 0.0004083578797485508, 'samples': 2979840, 'steps': 15519, 'loss/train': 2.900633454322815} 01/29/2022 08:45:46 - INFO - codeparrot_training - Step 15520: {'lr': 0.00040834521815688753, 'samples': 2980032, 'steps': 15520, 'loss/train': 1.4200959205627441} 01/29/2022 08:45:50 - INFO - codeparrot_training - Step 15521: {'lr': 0.00040833255588692375, 'samples': 2980224, 'steps': 15521, 'loss/train': 1.2177327275276184} 01/29/2022 08:45:54 - INFO - codeparrot_training - Step 15522: {'lr': 0.0004083198929387135, 'samples': 2980416, 'steps': 15522, 'loss/train': 1.5308443307876587} 01/29/2022 08:46:00 - INFO - codeparrot_training - Step 15523: {'lr': 0.0004083072293123111, 'samples': 2980608, 'steps': 15523, 'loss/train': 0.8218154311180115} 01/29/2022 08:46:04 - INFO - codeparrot_training - Step 15524: {'lr': 0.00040829456500777084, 'samples': 2980800, 'steps': 15524, 'loss/train': 1.728347897529602} 01/29/2022 08:46:08 - INFO - codeparrot_training - Step 15525: {'lr': 0.00040828190002514694, 'samples': 2980992, 'steps': 15525, 'loss/train': 0.7599301636219025} 01/29/2022 08:46:12 - INFO - codeparrot_training - Step 15526: {'lr': 0.0004082692343644936, 'samples': 2981184, 'steps': 15526, 'loss/train': 1.951661467552185} 01/29/2022 08:46:17 - INFO - codeparrot_training - Step 15527: {'lr': 0.00040825656802586513, 'samples': 2981376, 'steps': 15527, 'loss/train': 1.6698076128959656} 01/29/2022 08:46:22 - INFO - codeparrot_training - Step 15528: {'lr': 0.00040824390100931585, 'samples': 2981568, 'steps': 15528, 'loss/train': 1.4228705763816833} 01/29/2022 08:46:26 - INFO - codeparrot_training - Step 15529: {'lr': 0.00040823123331489985, 'samples': 2981760, 'steps': 15529, 'loss/train': 0.6872095316648483} 01/29/2022 08:46:31 - INFO - codeparrot_training - Step 15530: {'lr': 0.0004082185649426715, 'samples': 2981952, 'steps': 15530, 'loss/train': 0.9863774478435516} 01/29/2022 08:46:35 - INFO - codeparrot_training - Step 15531: {'lr': 0.0004082058958926851, 'samples': 2982144, 'steps': 15531, 'loss/train': 1.086602121591568} 01/29/2022 08:46:39 - INFO - codeparrot_training - Step 15532: {'lr': 0.0004081932261649949, 'samples': 2982336, 'steps': 15532, 'loss/train': 1.7089081406593323} 01/29/2022 08:46:44 - INFO - codeparrot_training - Step 15533: {'lr': 0.00040818055575965505, 'samples': 2982528, 'steps': 15533, 'loss/train': 1.8927149176597595} 01/29/2022 08:46:49 - INFO - codeparrot_training - Step 15534: {'lr': 0.0004081678846767199, 'samples': 2982720, 'steps': 15534, 'loss/train': 0.7872477471828461} 01/29/2022 08:46:53 - INFO - codeparrot_training - Step 15535: {'lr': 0.00040815521291624393, 'samples': 2982912, 'steps': 15535, 'loss/train': 0.8275935351848602} 01/29/2022 08:46:57 - INFO - codeparrot_training - Step 15536: {'lr': 0.0004081425404782811, 'samples': 2983104, 'steps': 15536, 'loss/train': 1.0777945518493652} 01/29/2022 08:47:02 - INFO - codeparrot_training - Step 15537: {'lr': 0.0004081298673628859, 'samples': 2983296, 'steps': 15537, 'loss/train': 1.99870365858078} 01/29/2022 08:47:09 - INFO - codeparrot_training - Step 15538: {'lr': 0.00040811719357011257, 'samples': 2983488, 'steps': 15538, 'loss/train': 0.6599339693784714} 01/29/2022 08:47:13 - INFO - codeparrot_training - Step 15539: {'lr': 0.00040810451910001537, 'samples': 2983680, 'steps': 15539, 'loss/train': 1.5966444611549377} 01/29/2022 08:47:17 - INFO - codeparrot_training - Step 15540: {'lr': 0.00040809184395264867, 'samples': 2983872, 'steps': 15540, 'loss/train': 1.3918440341949463} 01/29/2022 08:47:22 - INFO - codeparrot_training - Step 15541: {'lr': 0.0004080791681280667, 'samples': 2984064, 'steps': 15541, 'loss/train': 2.830319344997406} 01/29/2022 08:47:26 - INFO - codeparrot_training - Step 15542: {'lr': 0.00040806649162632364, 'samples': 2984256, 'steps': 15542, 'loss/train': 0.5347171425819397} 01/29/2022 08:47:31 - INFO - codeparrot_training - Step 15543: {'lr': 0.000408053814447474, 'samples': 2984448, 'steps': 15543, 'loss/train': 2.1530563831329346} 01/29/2022 08:47:35 - INFO - codeparrot_training - Step 15544: {'lr': 0.00040804113659157203, 'samples': 2984640, 'steps': 15544, 'loss/train': 1.7080208659172058} 01/29/2022 08:47:40 - INFO - codeparrot_training - Step 15545: {'lr': 0.00040802845805867205, 'samples': 2984832, 'steps': 15545, 'loss/train': 1.5290751457214355} 01/29/2022 08:47:44 - INFO - codeparrot_training - Step 15546: {'lr': 0.0004080157788488282, 'samples': 2985024, 'steps': 15546, 'loss/train': 1.96509450674057} 01/29/2022 08:47:48 - INFO - codeparrot_training - Step 15547: {'lr': 0.0004080030989620951, 'samples': 2985216, 'steps': 15547, 'loss/train': 1.5227383375167847} 01/29/2022 08:47:55 - INFO - codeparrot_training - Step 15548: {'lr': 0.0004079904183985268, 'samples': 2985408, 'steps': 15548, 'loss/train': 1.9040104150772095} 01/29/2022 08:48:00 - INFO - codeparrot_training - Step 15549: {'lr': 0.0004079777371581777, 'samples': 2985600, 'steps': 15549, 'loss/train': 0.46413879096508026} 01/29/2022 08:48:04 - INFO - codeparrot_training - Step 15550: {'lr': 0.00040796505524110215, 'samples': 2985792, 'steps': 15550, 'loss/train': 1.993338167667389} 01/29/2022 08:48:08 - INFO - codeparrot_training - Step 15551: {'lr': 0.00040795237264735454, 'samples': 2985984, 'steps': 15551, 'loss/train': 1.2871539294719696} 01/29/2022 08:48:12 - INFO - codeparrot_training - Step 15552: {'lr': 0.00040793968937698905, 'samples': 2986176, 'steps': 15552, 'loss/train': 1.911103069782257} 01/29/2022 08:48:19 - INFO - codeparrot_training - Step 15553: {'lr': 0.00040792700543006014, 'samples': 2986368, 'steps': 15553, 'loss/train': 7.842292785644531} 01/29/2022 08:48:23 - INFO - codeparrot_training - Step 15554: {'lr': 0.000407914320806622, 'samples': 2986560, 'steps': 15554, 'loss/train': 1.7788395881652832} 01/29/2022 08:48:27 - INFO - codeparrot_training - Step 15555: {'lr': 0.0004079016355067291, 'samples': 2986752, 'steps': 15555, 'loss/train': 1.8027682900428772} 01/29/2022 08:48:31 - INFO - codeparrot_training - Step 15556: {'lr': 0.0004078889495304357, 'samples': 2986944, 'steps': 15556, 'loss/train': 1.2526581287384033} 01/29/2022 08:48:36 - INFO - codeparrot_training - Step 15557: {'lr': 0.00040787626287779624, 'samples': 2987136, 'steps': 15557, 'loss/train': 0.8909316658973694} 01/29/2022 08:48:40 - INFO - codeparrot_training - Step 15558: {'lr': 0.0004078635755488649, 'samples': 2987328, 'steps': 15558, 'loss/train': 1.923949420452118} 01/29/2022 08:48:45 - INFO - codeparrot_training - Step 15559: {'lr': 0.00040785088754369627, 'samples': 2987520, 'steps': 15559, 'loss/train': 1.5804702043533325} 01/29/2022 08:48:49 - INFO - codeparrot_training - Step 15560: {'lr': 0.00040783819886234445, 'samples': 2987712, 'steps': 15560, 'loss/train': 1.6723657250404358} 01/29/2022 08:48:54 - INFO - codeparrot_training - Step 15561: {'lr': 0.000407825509504864, 'samples': 2987904, 'steps': 15561, 'loss/train': 1.946934700012207} 01/29/2022 08:48:58 - INFO - codeparrot_training - Step 15562: {'lr': 0.00040781281947130897, 'samples': 2988096, 'steps': 15562, 'loss/train': 2.01655375957489} 01/29/2022 08:49:02 - INFO - codeparrot_training - Step 15563: {'lr': 0.0004078001287617342, 'samples': 2988288, 'steps': 15563, 'loss/train': 2.201747953891754} 01/29/2022 08:49:10 - INFO - codeparrot_training - Step 15564: {'lr': 0.0004077874373761936, 'samples': 2988480, 'steps': 15564, 'loss/train': 1.9940922260284424} 01/29/2022 08:49:14 - INFO - codeparrot_training - Step 15565: {'lr': 0.0004077747453147418, 'samples': 2988672, 'steps': 15565, 'loss/train': 2.697249948978424} 01/29/2022 08:49:18 - INFO - codeparrot_training - Step 15566: {'lr': 0.0004077620525774331, 'samples': 2988864, 'steps': 15566, 'loss/train': 2.550661325454712} 01/29/2022 08:49:22 - INFO - codeparrot_training - Step 15567: {'lr': 0.0004077493591643219, 'samples': 2989056, 'steps': 15567, 'loss/train': 1.6567697525024414} 01/29/2022 08:49:27 - INFO - codeparrot_training - Step 15568: {'lr': 0.00040773666507546244, 'samples': 2989248, 'steps': 15568, 'loss/train': 1.9674865007400513} 01/29/2022 08:49:32 - INFO - codeparrot_training - Step 15569: {'lr': 0.00040772397031090923, 'samples': 2989440, 'steps': 15569, 'loss/train': 4.34122359752655} 01/29/2022 08:49:36 - INFO - codeparrot_training - Step 15570: {'lr': 0.0004077112748707166, 'samples': 2989632, 'steps': 15570, 'loss/train': 0.954957515001297} 01/29/2022 08:49:40 - INFO - codeparrot_training - Step 15571: {'lr': 0.000407698578754939, 'samples': 2989824, 'steps': 15571, 'loss/train': 1.8677911162376404} 01/29/2022 08:49:45 - INFO - codeparrot_training - Step 15572: {'lr': 0.0004076858819636307, 'samples': 2990016, 'steps': 15572, 'loss/train': 1.5498111248016357} 01/29/2022 08:49:49 - INFO - codeparrot_training - Step 15573: {'lr': 0.0004076731844968462, 'samples': 2990208, 'steps': 15573, 'loss/train': 2.1383233666419983} 01/29/2022 08:49:56 - INFO - codeparrot_training - Step 15574: {'lr': 0.00040766048635463984, 'samples': 2990400, 'steps': 15574, 'loss/train': 1.6912214756011963} 01/29/2022 08:50:00 - INFO - codeparrot_training - Step 15575: {'lr': 0.000407647787537066, 'samples': 2990592, 'steps': 15575, 'loss/train': 1.5515769124031067} 01/29/2022 08:50:04 - INFO - codeparrot_training - Step 15576: {'lr': 0.00040763508804417904, 'samples': 2990784, 'steps': 15576, 'loss/train': 1.7243912816047668} 01/29/2022 08:50:09 - INFO - codeparrot_training - Step 15577: {'lr': 0.0004076223878760335, 'samples': 2990976, 'steps': 15577, 'loss/train': 1.5415635108947754} 01/29/2022 08:50:13 - INFO - codeparrot_training - Step 15578: {'lr': 0.0004076096870326837, 'samples': 2991168, 'steps': 15578, 'loss/train': 1.850174903869629} 01/29/2022 08:50:18 - INFO - codeparrot_training - Step 15579: {'lr': 0.000407596985514184, 'samples': 2991360, 'steps': 15579, 'loss/train': 1.3745046257972717} 01/29/2022 08:50:22 - INFO - codeparrot_training - Step 15580: {'lr': 0.00040758428332058895, 'samples': 2991552, 'steps': 15580, 'loss/train': 1.7276421189308167} 01/29/2022 08:50:27 - INFO - codeparrot_training - Step 15581: {'lr': 0.00040757158045195274, 'samples': 2991744, 'steps': 15581, 'loss/train': 1.8433810472488403} 01/29/2022 08:50:31 - INFO - codeparrot_training - Step 15582: {'lr': 0.00040755887690833005, 'samples': 2991936, 'steps': 15582, 'loss/train': 1.778451919555664} 01/29/2022 08:50:35 - INFO - codeparrot_training - Step 15583: {'lr': 0.00040754617268977503, 'samples': 2992128, 'steps': 15583, 'loss/train': 1.1699650883674622} 01/29/2022 08:50:40 - INFO - codeparrot_training - Step 15584: {'lr': 0.0004075334677963423, 'samples': 2992320, 'steps': 15584, 'loss/train': 1.856697678565979} 01/29/2022 08:50:45 - INFO - codeparrot_training - Step 15585: {'lr': 0.00040752076222808623, 'samples': 2992512, 'steps': 15585, 'loss/train': 1.9370512962341309} 01/29/2022 08:50:49 - INFO - codeparrot_training - Step 15586: {'lr': 0.00040750805598506115, 'samples': 2992704, 'steps': 15586, 'loss/train': 1.1433367431163788} 01/29/2022 08:50:53 - INFO - codeparrot_training - Step 15587: {'lr': 0.00040749534906732167, 'samples': 2992896, 'steps': 15587, 'loss/train': 1.7219220399856567} 01/29/2022 08:50:57 - INFO - codeparrot_training - Step 15588: {'lr': 0.0004074826414749221, 'samples': 2993088, 'steps': 15588, 'loss/train': 2.112025201320648} 01/29/2022 08:51:03 - INFO - codeparrot_training - Step 15589: {'lr': 0.00040746993320791685, 'samples': 2993280, 'steps': 15589, 'loss/train': 2.364961624145508} 01/29/2022 08:51:07 - INFO - codeparrot_training - Step 15590: {'lr': 0.00040745722426636043, 'samples': 2993472, 'steps': 15590, 'loss/train': 1.5646242499351501} 01/29/2022 08:51:11 - INFO - codeparrot_training - Step 15591: {'lr': 0.0004074445146503073, 'samples': 2993664, 'steps': 15591, 'loss/train': 2.477022707462311} 01/29/2022 08:51:16 - INFO - codeparrot_training - Step 15592: {'lr': 0.00040743180435981187, 'samples': 2993856, 'steps': 15592, 'loss/train': 1.188499242067337} 01/29/2022 08:51:20 - INFO - codeparrot_training - Step 15593: {'lr': 0.0004074190933949286, 'samples': 2994048, 'steps': 15593, 'loss/train': 1.1514906585216522} 01/29/2022 08:51:27 - INFO - codeparrot_training - Step 15594: {'lr': 0.00040740638175571175, 'samples': 2994240, 'steps': 15594, 'loss/train': 2.3395564556121826} 01/29/2022 08:51:31 - INFO - codeparrot_training - Step 15595: {'lr': 0.0004073936694422161, 'samples': 2994432, 'steps': 15595, 'loss/train': 0.7581578493118286} 01/29/2022 08:51:35 - INFO - codeparrot_training - Step 15596: {'lr': 0.0004073809564544959, 'samples': 2994624, 'steps': 15596, 'loss/train': 1.902231752872467} 01/29/2022 08:51:40 - INFO - codeparrot_training - Step 15597: {'lr': 0.0004073682427926057, 'samples': 2994816, 'steps': 15597, 'loss/train': 1.034072756767273} 01/29/2022 08:51:44 - INFO - codeparrot_training - Step 15598: {'lr': 0.00040735552845659986, 'samples': 2995008, 'steps': 15598, 'loss/train': 1.1489141285419464} 01/29/2022 08:51:49 - INFO - codeparrot_training - Step 15599: {'lr': 0.00040734281344653294, 'samples': 2995200, 'steps': 15599, 'loss/train': 1.7568976879119873} 01/29/2022 08:51:53 - INFO - codeparrot_training - Step 15600: {'lr': 0.0004073300977624594, 'samples': 2995392, 'steps': 15600, 'loss/train': 2.1434163451194763} 01/29/2022 08:51:58 - INFO - codeparrot_training - Step 15601: {'lr': 0.0004073173814044336, 'samples': 2995584, 'steps': 15601, 'loss/train': 0.46418139338493347} 01/29/2022 08:52:02 - INFO - codeparrot_training - Step 15602: {'lr': 0.0004073046643725101, 'samples': 2995776, 'steps': 15602, 'loss/train': 1.7931076884269714} 01/29/2022 08:52:06 - INFO - codeparrot_training - Step 15603: {'lr': 0.0004072919466667434, 'samples': 2995968, 'steps': 15603, 'loss/train': 0.3386364206671715} 01/29/2022 08:52:11 - INFO - codeparrot_training - Step 15604: {'lr': 0.000407279228287188, 'samples': 2996160, 'steps': 15604, 'loss/train': 1.270950436592102} 01/29/2022 08:52:16 - INFO - codeparrot_training - Step 15605: {'lr': 0.00040726650923389825, 'samples': 2996352, 'steps': 15605, 'loss/train': 1.3433426320552826} 01/29/2022 08:52:20 - INFO - codeparrot_training - Step 15606: {'lr': 0.00040725378950692874, 'samples': 2996544, 'steps': 15606, 'loss/train': 1.6416188478469849} 01/29/2022 08:52:24 - INFO - codeparrot_training - Step 15607: {'lr': 0.0004072410691063339, 'samples': 2996736, 'steps': 15607, 'loss/train': 1.1912034451961517} 01/29/2022 08:52:28 - INFO - codeparrot_training - Step 15608: {'lr': 0.00040722834803216834, 'samples': 2996928, 'steps': 15608, 'loss/train': 0.687039390206337} 01/29/2022 08:52:36 - INFO - codeparrot_training - Step 15609: {'lr': 0.0004072156262844864, 'samples': 2997120, 'steps': 15609, 'loss/train': 1.964522659778595} 01/29/2022 08:52:40 - INFO - codeparrot_training - Step 15610: {'lr': 0.0004072029038633426, 'samples': 2997312, 'steps': 15610, 'loss/train': 1.9642505049705505} 01/29/2022 08:52:44 - INFO - codeparrot_training - Step 15611: {'lr': 0.0004071901807687915, 'samples': 2997504, 'steps': 15611, 'loss/train': 1.248433381319046} 01/29/2022 08:52:49 - INFO - codeparrot_training - Step 15612: {'lr': 0.0004071774570008876, 'samples': 2997696, 'steps': 15612, 'loss/train': 2.4371067881584167} 01/29/2022 08:52:53 - INFO - codeparrot_training - Step 15613: {'lr': 0.00040716473255968534, 'samples': 2997888, 'steps': 15613, 'loss/train': 1.5374447107315063} 01/29/2022 08:52:58 - INFO - codeparrot_training - Step 15614: {'lr': 0.0004071520074452393, 'samples': 2998080, 'steps': 15614, 'loss/train': 2.086022973060608} 01/29/2022 08:53:03 - INFO - codeparrot_training - Step 15615: {'lr': 0.000407139281657604, 'samples': 2998272, 'steps': 15615, 'loss/train': 1.4504790902137756} 01/29/2022 08:53:07 - INFO - codeparrot_training - Step 15616: {'lr': 0.0004071265551968338, 'samples': 2998464, 'steps': 15616, 'loss/train': 1.8190587759017944} 01/29/2022 08:53:11 - INFO - codeparrot_training - Step 15617: {'lr': 0.0004071138280629835, 'samples': 2998656, 'steps': 15617, 'loss/train': 1.609880805015564} 01/29/2022 08:53:18 - INFO - codeparrot_training - Step 15618: {'lr': 0.00040710110025610733, 'samples': 2998848, 'steps': 15618, 'loss/train': 1.7183491587638855} 01/29/2022 08:53:22 - INFO - codeparrot_training - Step 15619: {'lr': 0.00040708837177626, 'samples': 2999040, 'steps': 15619, 'loss/train': 1.8836297392845154} 01/29/2022 08:53:27 - INFO - codeparrot_training - Step 15620: {'lr': 0.00040707564262349594, 'samples': 2999232, 'steps': 15620, 'loss/train': 1.4534865617752075} 01/29/2022 08:53:31 - INFO - codeparrot_training - Step 15621: {'lr': 0.00040706291279786965, 'samples': 2999424, 'steps': 15621, 'loss/train': 1.7988973259925842} 01/29/2022 08:53:35 - INFO - codeparrot_training - Step 15622: {'lr': 0.0004070501822994358, 'samples': 2999616, 'steps': 15622, 'loss/train': 2.4344265460968018} 01/29/2022 08:53:40 - INFO - codeparrot_training - Step 15623: {'lr': 0.00040703745112824876, 'samples': 2999808, 'steps': 15623, 'loss/train': 1.8340832591056824} 01/29/2022 08:53:45 - INFO - codeparrot_training - Step 15624: {'lr': 0.00040702471928436316, 'samples': 3000000, 'steps': 15624, 'loss/train': 1.1986874341964722} 01/29/2022 08:53:49 - INFO - codeparrot_training - Step 15625: {'lr': 0.00040701198676783355, 'samples': 3000192, 'steps': 15625, 'loss/train': 1.8628880381584167} 01/29/2022 08:53:53 - INFO - codeparrot_training - Step 15626: {'lr': 0.00040699925357871446, 'samples': 3000384, 'steps': 15626, 'loss/train': 1.3762772977352142} 01/29/2022 08:53:57 - INFO - codeparrot_training - Step 15627: {'lr': 0.00040698651971706037, 'samples': 3000576, 'steps': 15627, 'loss/train': 1.4639402031898499} 01/29/2022 08:54:03 - INFO - codeparrot_training - Step 15628: {'lr': 0.00040697378518292593, 'samples': 3000768, 'steps': 15628, 'loss/train': 1.513400673866272} 01/29/2022 08:54:08 - INFO - codeparrot_training - Step 15629: {'lr': 0.0004069610499763656, 'samples': 3000960, 'steps': 15629, 'loss/train': 2.1630685329437256} 01/29/2022 08:54:12 - INFO - codeparrot_training - Step 15630: {'lr': 0.00040694831409743406, 'samples': 3001152, 'steps': 15630, 'loss/train': 2.2417749166488647} 01/29/2022 08:54:16 - INFO - codeparrot_training - Step 15631: {'lr': 0.00040693557754618566, 'samples': 3001344, 'steps': 15631, 'loss/train': 0.7694900929927826} 01/29/2022 08:54:20 - INFO - codeparrot_training - Step 15632: {'lr': 0.00040692284032267515, 'samples': 3001536, 'steps': 15632, 'loss/train': 1.1958398222923279} 01/29/2022 08:54:24 - INFO - codeparrot_training - Step 15633: {'lr': 0.00040691010242695696, 'samples': 3001728, 'steps': 15633, 'loss/train': 1.7964815497398376} 01/29/2022 08:54:32 - INFO - codeparrot_training - Step 15634: {'lr': 0.00040689736385908574, 'samples': 3001920, 'steps': 15634, 'loss/train': 2.4392253756523132} 01/29/2022 08:54:36 - INFO - codeparrot_training - Step 15635: {'lr': 0.0004068846246191161, 'samples': 3002112, 'steps': 15635, 'loss/train': 2.1245477199554443} 01/29/2022 08:54:40 - INFO - codeparrot_training - Step 15636: {'lr': 0.00040687188470710245, 'samples': 3002304, 'steps': 15636, 'loss/train': 2.5137773752212524} 01/29/2022 08:54:44 - INFO - codeparrot_training - Step 15637: {'lr': 0.00040685914412309955, 'samples': 3002496, 'steps': 15637, 'loss/train': 2.842802345752716} 01/29/2022 08:54:49 - INFO - codeparrot_training - Step 15638: {'lr': 0.0004068464028671618, 'samples': 3002688, 'steps': 15638, 'loss/train': 2.0633127093315125} 01/29/2022 08:54:55 - INFO - codeparrot_training - Step 15639: {'lr': 0.00040683366093934394, 'samples': 3002880, 'steps': 15639, 'loss/train': 1.8270097374916077} 01/29/2022 08:54:59 - INFO - codeparrot_training - Step 15640: {'lr': 0.0004068209183397004, 'samples': 3003072, 'steps': 15640, 'loss/train': 0.9712969064712524} 01/29/2022 08:55:04 - INFO - codeparrot_training - Step 15641: {'lr': 0.0004068081750682859, 'samples': 3003264, 'steps': 15641, 'loss/train': 1.831731140613556} 01/29/2022 08:55:08 - INFO - codeparrot_training - Step 15642: {'lr': 0.00040679543112515494, 'samples': 3003456, 'steps': 15642, 'loss/train': 1.8359280824661255} 01/29/2022 08:55:12 - INFO - codeparrot_training - Step 15643: {'lr': 0.00040678268651036213, 'samples': 3003648, 'steps': 15643, 'loss/train': 1.9364696145057678} 01/29/2022 08:55:16 - INFO - codeparrot_training - Step 15644: {'lr': 0.0004067699412239622, 'samples': 3003840, 'steps': 15644, 'loss/train': 2.0501561164855957} 01/29/2022 08:55:22 - INFO - codeparrot_training - Step 15645: {'lr': 0.00040675719526600947, 'samples': 3004032, 'steps': 15645, 'loss/train': 0.6917399168014526} 01/29/2022 08:55:26 - INFO - codeparrot_training - Step 15646: {'lr': 0.0004067444486365587, 'samples': 3004224, 'steps': 15646, 'loss/train': 1.8368732929229736} 01/29/2022 08:55:31 - INFO - codeparrot_training - Step 15647: {'lr': 0.00040673170133566453, 'samples': 3004416, 'steps': 15647, 'loss/train': 0.7152887731790543} 01/29/2022 08:55:35 - INFO - codeparrot_training - Step 15648: {'lr': 0.0004067189533633815, 'samples': 3004608, 'steps': 15648, 'loss/train': 1.1684433817863464} 01/29/2022 08:55:39 - INFO - codeparrot_training - Step 15649: {'lr': 0.00040670620471976426, 'samples': 3004800, 'steps': 15649, 'loss/train': 0.9164820313453674} 01/29/2022 08:55:44 - INFO - codeparrot_training - Step 15650: {'lr': 0.0004066934554048674, 'samples': 3004992, 'steps': 15650, 'loss/train': 2.131416141986847} 01/29/2022 08:55:49 - INFO - codeparrot_training - Step 15651: {'lr': 0.00040668070541874553, 'samples': 3005184, 'steps': 15651, 'loss/train': 2.080080986022949} 01/29/2022 08:55:53 - INFO - codeparrot_training - Step 15652: {'lr': 0.00040666795476145326, 'samples': 3005376, 'steps': 15652, 'loss/train': 1.4612444043159485} 01/29/2022 08:55:57 - INFO - codeparrot_training - Step 15653: {'lr': 0.00040665520343304516, 'samples': 3005568, 'steps': 15653, 'loss/train': 1.410999834537506} 01/29/2022 08:56:01 - INFO - codeparrot_training - Step 15654: {'lr': 0.00040664245143357604, 'samples': 3005760, 'steps': 15654, 'loss/train': 1.2054231762886047} 01/29/2022 08:56:08 - INFO - codeparrot_training - Step 15655: {'lr': 0.0004066296987631003, 'samples': 3005952, 'steps': 15655, 'loss/train': 2.3030113577842712} 01/29/2022 08:56:13 - INFO - codeparrot_training - Step 15656: {'lr': 0.0004066169454216727, 'samples': 3006144, 'steps': 15656, 'loss/train': 1.8425729870796204} 01/29/2022 08:56:17 - INFO - codeparrot_training - Step 15657: {'lr': 0.00040660419140934787, 'samples': 3006336, 'steps': 15657, 'loss/train': 1.1294539868831635} 01/29/2022 08:56:21 - INFO - codeparrot_training - Step 15658: {'lr': 0.0004065914367261804, 'samples': 3006528, 'steps': 15658, 'loss/train': 0.07340705394744873} 01/29/2022 08:56:25 - INFO - codeparrot_training - Step 15659: {'lr': 0.00040657868137222486, 'samples': 3006720, 'steps': 15659, 'loss/train': 1.5437747240066528} 01/29/2022 08:56:31 - INFO - codeparrot_training - Step 15660: {'lr': 0.000406565925347536, 'samples': 3006912, 'steps': 15660, 'loss/train': 1.4725343585014343} 01/29/2022 08:56:35 - INFO - codeparrot_training - Step 15661: {'lr': 0.0004065531686521685, 'samples': 3007104, 'steps': 15661, 'loss/train': 1.6720399260520935} 01/29/2022 08:56:39 - INFO - codeparrot_training - Step 15662: {'lr': 0.00040654041128617693, 'samples': 3007296, 'steps': 15662, 'loss/train': 3.1301125288009644} 01/29/2022 08:56:43 - INFO - codeparrot_training - Step 15663: {'lr': 0.0004065276532496158, 'samples': 3007488, 'steps': 15663, 'loss/train': 0.8795398771762848} 01/29/2022 08:56:48 - INFO - codeparrot_training - Step 15664: {'lr': 0.0004065148945425401, 'samples': 3007680, 'steps': 15664, 'loss/train': 2.1717939376831055} 01/29/2022 08:56:53 - INFO - codeparrot_training - Step 15665: {'lr': 0.0004065021351650042, 'samples': 3007872, 'steps': 15665, 'loss/train': 1.437207967042923} 01/29/2022 08:56:57 - INFO - codeparrot_training - Step 15666: {'lr': 0.00040648937511706285, 'samples': 3008064, 'steps': 15666, 'loss/train': 1.2205211520195007} 01/29/2022 08:57:01 - INFO - codeparrot_training - Step 15667: {'lr': 0.0004064766143987707, 'samples': 3008256, 'steps': 15667, 'loss/train': 0.6660687625408173} 01/29/2022 08:57:06 - INFO - codeparrot_training - Step 15668: {'lr': 0.00040646385301018243, 'samples': 3008448, 'steps': 15668, 'loss/train': 1.604384958744049} 01/29/2022 08:57:10 - INFO - codeparrot_training - Step 15669: {'lr': 0.0004064510909513527, 'samples': 3008640, 'steps': 15669, 'loss/train': 1.712737798690796} 01/29/2022 08:57:17 - INFO - codeparrot_training - Step 15670: {'lr': 0.00040643832822233615, 'samples': 3008832, 'steps': 15670, 'loss/train': 2.0656723380088806} 01/29/2022 08:57:21 - INFO - codeparrot_training - Step 15671: {'lr': 0.0004064255648231875, 'samples': 3009024, 'steps': 15671, 'loss/train': 1.7048320770263672} 01/29/2022 08:57:25 - INFO - codeparrot_training - Step 15672: {'lr': 0.00040641280075396144, 'samples': 3009216, 'steps': 15672, 'loss/train': 1.9046815037727356} 01/29/2022 08:57:30 - INFO - codeparrot_training - Step 15673: {'lr': 0.00040640003601471255, 'samples': 3009408, 'steps': 15673, 'loss/train': 7.476911544799805} 01/29/2022 08:57:34 - INFO - codeparrot_training - Step 15674: {'lr': 0.00040638727060549556, 'samples': 3009600, 'steps': 15674, 'loss/train': 1.848035216331482} 01/29/2022 08:57:39 - INFO - codeparrot_training - Step 15675: {'lr': 0.00040637450452636517, 'samples': 3009792, 'steps': 15675, 'loss/train': 2.207838535308838} 01/29/2022 08:57:43 - INFO - codeparrot_training - Step 15676: {'lr': 0.00040636173777737613, 'samples': 3009984, 'steps': 15676, 'loss/train': 1.5015875101089478} 01/29/2022 08:57:48 - INFO - codeparrot_training - Step 15677: {'lr': 0.000406348970358583, 'samples': 3010176, 'steps': 15677, 'loss/train': 0.5550481081008911} 01/29/2022 08:57:52 - INFO - codeparrot_training - Step 15678: {'lr': 0.00040633620227004054, 'samples': 3010368, 'steps': 15678, 'loss/train': 2.238046646118164} 01/29/2022 08:57:56 - INFO - codeparrot_training - Step 15679: {'lr': 0.0004063234335118033, 'samples': 3010560, 'steps': 15679, 'loss/train': 1.8426030278205872} 01/29/2022 08:58:03 - INFO - codeparrot_training - Step 15680: {'lr': 0.00040631066408392636, 'samples': 3010752, 'steps': 15680, 'loss/train': 1.8671488165855408} 01/29/2022 08:58:08 - INFO - codeparrot_training - Step 15681: {'lr': 0.000406297893986464, 'samples': 3010944, 'steps': 15681, 'loss/train': 1.386085957288742} 01/29/2022 08:58:12 - INFO - codeparrot_training - Step 15682: {'lr': 0.0004062851232194711, 'samples': 3011136, 'steps': 15682, 'loss/train': 1.6793639659881592} 01/29/2022 08:58:16 - INFO - codeparrot_training - Step 15683: {'lr': 0.00040627235178300236, 'samples': 3011328, 'steps': 15683, 'loss/train': 0.8637796640396118} 01/29/2022 08:58:20 - INFO - codeparrot_training - Step 15684: {'lr': 0.0004062595796771126, 'samples': 3011520, 'steps': 15684, 'loss/train': 2.720530092716217} 01/29/2022 08:58:26 - INFO - codeparrot_training - Step 15685: {'lr': 0.0004062468069018563, 'samples': 3011712, 'steps': 15685, 'loss/train': 1.3980046212673187} 01/29/2022 08:58:30 - INFO - codeparrot_training - Step 15686: {'lr': 0.0004062340334572883, 'samples': 3011904, 'steps': 15686, 'loss/train': 1.4173736572265625} 01/29/2022 08:58:34 - INFO - codeparrot_training - Step 15687: {'lr': 0.0004062212593434634, 'samples': 3012096, 'steps': 15687, 'loss/train': 1.903229534626007} 01/29/2022 08:58:38 - INFO - codeparrot_training - Step 15688: {'lr': 0.0004062084845604361, 'samples': 3012288, 'steps': 15688, 'loss/train': 2.079732298851013} 01/29/2022 08:58:43 - INFO - codeparrot_training - Step 15689: {'lr': 0.00040619570910826135, 'samples': 3012480, 'steps': 15689, 'loss/train': 2.137880802154541} 01/29/2022 08:58:48 - INFO - codeparrot_training - Step 15690: {'lr': 0.0004061829329869937, 'samples': 3012672, 'steps': 15690, 'loss/train': 1.4721179008483887} 01/29/2022 08:58:52 - INFO - codeparrot_training - Step 15691: {'lr': 0.0004061701561966881, 'samples': 3012864, 'steps': 15691, 'loss/train': 1.4713338911533356} 01/29/2022 08:58:56 - INFO - codeparrot_training - Step 15692: {'lr': 0.000406157378737399, 'samples': 3013056, 'steps': 15692, 'loss/train': 1.4817469418048859} 01/29/2022 08:59:01 - INFO - codeparrot_training - Step 15693: {'lr': 0.00040614460060918136, 'samples': 3013248, 'steps': 15693, 'loss/train': 3.6101202964782715} 01/29/2022 08:59:05 - INFO - codeparrot_training - Step 15694: {'lr': 0.0004061318218120898, 'samples': 3013440, 'steps': 15694, 'loss/train': 2.0195000767707825} 01/29/2022 08:59:12 - INFO - codeparrot_training - Step 15695: {'lr': 0.000406119042346179, 'samples': 3013632, 'steps': 15695, 'loss/train': 1.710060954093933} 01/29/2022 08:59:16 - INFO - codeparrot_training - Step 15696: {'lr': 0.0004061062622115039, 'samples': 3013824, 'steps': 15696, 'loss/train': 0.8577764332294464} 01/29/2022 08:59:20 - INFO - codeparrot_training - Step 15697: {'lr': 0.0004060934814081192, 'samples': 3014016, 'steps': 15697, 'loss/train': 1.7961816787719727} 01/29/2022 08:59:25 - INFO - codeparrot_training - Step 15698: {'lr': 0.00040608069993607954, 'samples': 3014208, 'steps': 15698, 'loss/train': 1.368751734495163} 01/29/2022 08:59:29 - INFO - codeparrot_training - Step 15699: {'lr': 0.00040606791779543966, 'samples': 3014400, 'steps': 15699, 'loss/train': 1.199414223432541} 01/29/2022 08:59:34 - INFO - codeparrot_training - Step 15700: {'lr': 0.00040605513498625443, 'samples': 3014592, 'steps': 15700, 'loss/train': 1.4876404702663422} 01/29/2022 08:59:38 - INFO - codeparrot_training - Step 15701: {'lr': 0.00040604235150857855, 'samples': 3014784, 'steps': 15701, 'loss/train': 1.5231510400772095} 01/29/2022 08:59:43 - INFO - codeparrot_training - Step 15702: {'lr': 0.00040602956736246677, 'samples': 3014976, 'steps': 15702, 'loss/train': 1.9359861016273499} 01/29/2022 08:59:47 - INFO - codeparrot_training - Step 15703: {'lr': 0.00040601678254797394, 'samples': 3015168, 'steps': 15703, 'loss/train': 1.2914350926876068} 01/29/2022 08:59:51 - INFO - codeparrot_training - Step 15704: {'lr': 0.00040600399706515466, 'samples': 3015360, 'steps': 15704, 'loss/train': 0.8871455490589142} 01/29/2022 08:59:58 - INFO - codeparrot_training - Step 15705: {'lr': 0.0004059912109140638, 'samples': 3015552, 'steps': 15705, 'loss/train': 0.9753232598304749} 01/29/2022 09:00:02 - INFO - codeparrot_training - Step 15706: {'lr': 0.00040597842409475615, 'samples': 3015744, 'steps': 15706, 'loss/train': 1.9990264177322388} 01/29/2022 09:00:07 - INFO - codeparrot_training - Step 15707: {'lr': 0.00040596563660728646, 'samples': 3015936, 'steps': 15707, 'loss/train': 1.7630985975265503} 01/29/2022 09:00:11 - INFO - codeparrot_training - Step 15708: {'lr': 0.00040595284845170956, 'samples': 3016128, 'steps': 15708, 'loss/train': 1.6127547025680542} 01/29/2022 09:00:15 - INFO - codeparrot_training - Step 15709: {'lr': 0.0004059400596280801, 'samples': 3016320, 'steps': 15709, 'loss/train': 1.7774314284324646} 01/29/2022 09:00:20 - INFO - codeparrot_training - Step 15710: {'lr': 0.00040592727013645297, 'samples': 3016512, 'steps': 15710, 'loss/train': 2.3504660725593567} 01/29/2022 09:00:24 - INFO - codeparrot_training - Step 15711: {'lr': 0.0004059144799768829, 'samples': 3016704, 'steps': 15711, 'loss/train': 1.7261558175086975} 01/29/2022 09:00:29 - INFO - codeparrot_training - Step 15712: {'lr': 0.00040590168914942477, 'samples': 3016896, 'steps': 15712, 'loss/train': 1.5060269236564636} 01/29/2022 09:00:33 - INFO - codeparrot_training - Step 15713: {'lr': 0.0004058888976541333, 'samples': 3017088, 'steps': 15713, 'loss/train': 1.8296765685081482} 01/29/2022 09:00:37 - INFO - codeparrot_training - Step 15714: {'lr': 0.00040587610549106326, 'samples': 3017280, 'steps': 15714, 'loss/train': 1.3421682715415955} 01/29/2022 09:00:42 - INFO - codeparrot_training - Step 15715: {'lr': 0.00040586331266026943, 'samples': 3017472, 'steps': 15715, 'loss/train': 1.7708394527435303} 01/29/2022 09:00:47 - INFO - codeparrot_training - Step 15716: {'lr': 0.0004058505191618067, 'samples': 3017664, 'steps': 15716, 'loss/train': 1.805825650691986} 01/29/2022 09:00:51 - INFO - codeparrot_training - Step 15717: {'lr': 0.0004058377249957299, 'samples': 3017856, 'steps': 15717, 'loss/train': 1.7845492959022522} 01/29/2022 09:00:55 - INFO - codeparrot_training - Step 15718: {'lr': 0.0004058249301620937, 'samples': 3018048, 'steps': 15718, 'loss/train': 2.315739691257477} 01/29/2022 09:00:59 - INFO - codeparrot_training - Step 15719: {'lr': 0.00040581213466095304, 'samples': 3018240, 'steps': 15719, 'loss/train': 1.5544057488441467} 01/29/2022 09:01:05 - INFO - codeparrot_training - Step 15720: {'lr': 0.0004057993384923626, 'samples': 3018432, 'steps': 15720, 'loss/train': 0.6605349630117416} 01/29/2022 09:01:09 - INFO - codeparrot_training - Step 15721: {'lr': 0.0004057865416563773, 'samples': 3018624, 'steps': 15721, 'loss/train': 1.3978388607501984} 01/29/2022 09:01:13 - INFO - codeparrot_training - Step 15722: {'lr': 0.0004057737441530519, 'samples': 3018816, 'steps': 15722, 'loss/train': 0.7731878757476807} 01/29/2022 09:01:18 - INFO - codeparrot_training - Step 15723: {'lr': 0.0004057609459824412, 'samples': 3019008, 'steps': 15723, 'loss/train': 1.5609213709831238} 01/29/2022 09:01:22 - INFO - codeparrot_training - Step 15724: {'lr': 0.00040574814714460015, 'samples': 3019200, 'steps': 15724, 'loss/train': 7.462007761001587} 01/29/2022 09:01:29 - INFO - codeparrot_training - Step 15725: {'lr': 0.0004057353476395835, 'samples': 3019392, 'steps': 15725, 'loss/train': 1.6506147980690002} 01/29/2022 09:01:33 - INFO - codeparrot_training - Step 15726: {'lr': 0.00040572254746744607, 'samples': 3019584, 'steps': 15726, 'loss/train': 1.0056433081626892} 01/29/2022 09:01:38 - INFO - codeparrot_training - Step 15727: {'lr': 0.00040570974662824266, 'samples': 3019776, 'steps': 15727, 'loss/train': 1.673384964466095} 01/29/2022 09:01:42 - INFO - codeparrot_training - Step 15728: {'lr': 0.00040569694512202815, 'samples': 3019968, 'steps': 15728, 'loss/train': 1.4155804216861725} 01/29/2022 09:01:46 - INFO - codeparrot_training - Step 15729: {'lr': 0.00040568414294885736, 'samples': 3020160, 'steps': 15729, 'loss/train': 2.149828612804413} 01/29/2022 09:01:52 - INFO - codeparrot_training - Step 15730: {'lr': 0.00040567134010878513, 'samples': 3020352, 'steps': 15730, 'loss/train': 1.6744801998138428} 01/29/2022 09:01:56 - INFO - codeparrot_training - Step 15731: {'lr': 0.00040565853660186633, 'samples': 3020544, 'steps': 15731, 'loss/train': 1.2306550741195679} 01/29/2022 09:02:00 - INFO - codeparrot_training - Step 15732: {'lr': 0.0004056457324281557, 'samples': 3020736, 'steps': 15732, 'loss/train': 1.5887060165405273} 01/29/2022 09:02:04 - INFO - codeparrot_training - Step 15733: {'lr': 0.0004056329275877083, 'samples': 3020928, 'steps': 15733, 'loss/train': 1.8689979314804077} 01/29/2022 09:02:09 - INFO - codeparrot_training - Step 15734: {'lr': 0.00040562012208057886, 'samples': 3021120, 'steps': 15734, 'loss/train': 0.9916386902332306} 01/29/2022 09:02:14 - INFO - codeparrot_training - Step 15735: {'lr': 0.0004056073159068222, 'samples': 3021312, 'steps': 15735, 'loss/train': 1.0254566073417664} 01/29/2022 09:02:18 - INFO - codeparrot_training - Step 15736: {'lr': 0.0004055945090664931, 'samples': 3021504, 'steps': 15736, 'loss/train': 2.7794846892356873} 01/29/2022 09:02:22 - INFO - codeparrot_training - Step 15737: {'lr': 0.0004055817015596467, 'samples': 3021696, 'steps': 15737, 'loss/train': 2.1830846071243286} 01/29/2022 09:02:27 - INFO - codeparrot_training - Step 15738: {'lr': 0.00040556889338633754, 'samples': 3021888, 'steps': 15738, 'loss/train': 1.6054102778434753} 01/29/2022 09:02:31 - INFO - codeparrot_training - Step 15739: {'lr': 0.00040555608454662074, 'samples': 3022080, 'steps': 15739, 'loss/train': 2.357646882534027} 01/29/2022 09:02:38 - INFO - codeparrot_training - Step 15740: {'lr': 0.00040554327504055106, 'samples': 3022272, 'steps': 15740, 'loss/train': 1.8675840497016907} 01/29/2022 09:02:42 - INFO - codeparrot_training - Step 15741: {'lr': 0.00040553046486818336, 'samples': 3022464, 'steps': 15741, 'loss/train': 0.6522845327854156} 01/29/2022 09:02:46 - INFO - codeparrot_training - Step 15742: {'lr': 0.0004055176540295725, 'samples': 3022656, 'steps': 15742, 'loss/train': 1.5771911144256592} 01/29/2022 09:02:50 - INFO - codeparrot_training - Step 15743: {'lr': 0.00040550484252477347, 'samples': 3022848, 'steps': 15743, 'loss/train': 1.8101357817649841} 01/29/2022 09:02:55 - INFO - codeparrot_training - Step 15744: {'lr': 0.00040549203035384105, 'samples': 3023040, 'steps': 15744, 'loss/train': 1.325219839811325} 01/29/2022 09:03:00 - INFO - codeparrot_training - Step 15745: {'lr': 0.0004054792175168301, 'samples': 3023232, 'steps': 15745, 'loss/train': 1.9359957575798035} 01/29/2022 09:03:04 - INFO - codeparrot_training - Step 15746: {'lr': 0.00040546640401379556, 'samples': 3023424, 'steps': 15746, 'loss/train': 1.1835662126541138} 01/29/2022 09:03:08 - INFO - codeparrot_training - Step 15747: {'lr': 0.0004054535898447924, 'samples': 3023616, 'steps': 15747, 'loss/train': 1.2798188030719757} 01/29/2022 09:03:13 - INFO - codeparrot_training - Step 15748: {'lr': 0.0004054407750098753, 'samples': 3023808, 'steps': 15748, 'loss/train': 1.6714532375335693} 01/29/2022 09:03:17 - INFO - codeparrot_training - Step 15749: {'lr': 0.0004054279595090994, 'samples': 3024000, 'steps': 15749, 'loss/train': 1.7100366353988647} 01/29/2022 09:03:24 - INFO - codeparrot_training - Step 15750: {'lr': 0.0004054151433425194, 'samples': 3024192, 'steps': 15750, 'loss/train': 1.664452850818634} 01/29/2022 09:03:28 - INFO - codeparrot_training - Step 15751: {'lr': 0.00040540232651019027, 'samples': 3024384, 'steps': 15751, 'loss/train': 0.9632386565208435} 01/29/2022 09:03:33 - INFO - codeparrot_training - Step 15752: {'lr': 0.0004053895090121669, 'samples': 3024576, 'steps': 15752, 'loss/train': 1.7355610728263855} 01/29/2022 09:03:37 - INFO - codeparrot_training - Step 15753: {'lr': 0.00040537669084850426, 'samples': 3024768, 'steps': 15753, 'loss/train': 0.8011109232902527} 01/29/2022 09:03:41 - INFO - codeparrot_training - Step 15754: {'lr': 0.0004053638720192572, 'samples': 3024960, 'steps': 15754, 'loss/train': 1.5677509903907776} 01/29/2022 09:03:46 - INFO - codeparrot_training - Step 15755: {'lr': 0.00040535105252448067, 'samples': 3025152, 'steps': 15755, 'loss/train': 1.9668812155723572} 01/29/2022 09:03:50 - INFO - codeparrot_training - Step 15756: {'lr': 0.0004053382323642295, 'samples': 3025344, 'steps': 15756, 'loss/train': 1.8563318252563477} 01/29/2022 09:03:55 - INFO - codeparrot_training - Step 15757: {'lr': 0.0004053254115385587, 'samples': 3025536, 'steps': 15757, 'loss/train': 1.7669822573661804} 01/29/2022 09:03:59 - INFO - codeparrot_training - Step 15758: {'lr': 0.00040531259004752317, 'samples': 3025728, 'steps': 15758, 'loss/train': 2.069235384464264} 01/29/2022 09:04:03 - INFO - codeparrot_training - Step 15759: {'lr': 0.00040529976789117786, 'samples': 3025920, 'steps': 15759, 'loss/train': 1.6292635202407837} 01/29/2022 09:04:08 - INFO - codeparrot_training - Step 15760: {'lr': 0.0004052869450695776, 'samples': 3026112, 'steps': 15760, 'loss/train': 1.5067350268363953} 01/29/2022 09:04:13 - INFO - codeparrot_training - Step 15761: {'lr': 0.00040527412158277744, 'samples': 3026304, 'steps': 15761, 'loss/train': 1.4118070900440216} 01/29/2022 09:04:17 - INFO - codeparrot_training - Step 15762: {'lr': 0.00040526129743083216, 'samples': 3026496, 'steps': 15762, 'loss/train': 1.5558403730392456} 01/29/2022 09:04:21 - INFO - codeparrot_training - Step 15763: {'lr': 0.0004052484726137968, 'samples': 3026688, 'steps': 15763, 'loss/train': 1.5674877762794495} 01/29/2022 09:04:25 - INFO - codeparrot_training - Step 15764: {'lr': 0.00040523564713172634, 'samples': 3026880, 'steps': 15764, 'loss/train': 1.478917121887207} 01/29/2022 09:04:31 - INFO - codeparrot_training - Step 15765: {'lr': 0.0004052228209846756, 'samples': 3027072, 'steps': 15765, 'loss/train': 1.6952805519104004} 01/29/2022 09:04:35 - INFO - codeparrot_training - Step 15766: {'lr': 0.0004052099941726996, 'samples': 3027264, 'steps': 15766, 'loss/train': 2.1091954708099365} 01/29/2022 09:04:39 - INFO - codeparrot_training - Step 15767: {'lr': 0.0004051971666958533, 'samples': 3027456, 'steps': 15767, 'loss/train': 1.71734619140625} 01/29/2022 09:04:43 - INFO - codeparrot_training - Step 15768: {'lr': 0.0004051843385541916, 'samples': 3027648, 'steps': 15768, 'loss/train': 1.894408106803894} 01/29/2022 09:04:48 - INFO - codeparrot_training - Step 15769: {'lr': 0.00040517150974776945, 'samples': 3027840, 'steps': 15769, 'loss/train': 0.061209987848997116} 01/29/2022 09:04:55 - INFO - codeparrot_training - Step 15770: {'lr': 0.00040515868027664185, 'samples': 3028032, 'steps': 15770, 'loss/train': 1.2787964344024658} 01/29/2022 09:04:59 - INFO - codeparrot_training - Step 15771: {'lr': 0.00040514585014086367, 'samples': 3028224, 'steps': 15771, 'loss/train': 1.5956924557685852} 01/29/2022 09:05:04 - INFO - codeparrot_training - Step 15772: {'lr': 0.00040513301934049005, 'samples': 3028416, 'steps': 15772, 'loss/train': 0.6870658099651337} 01/29/2022 09:05:08 - INFO - codeparrot_training - Step 15773: {'lr': 0.00040512018787557574, 'samples': 3028608, 'steps': 15773, 'loss/train': 1.7962895035743713} 01/29/2022 09:05:12 - INFO - codeparrot_training - Step 15774: {'lr': 0.0004051073557461759, 'samples': 3028800, 'steps': 15774, 'loss/train': 2.4208931922912598} 01/29/2022 09:05:18 - INFO - codeparrot_training - Step 15775: {'lr': 0.00040509452295234527, 'samples': 3028992, 'steps': 15775, 'loss/train': 1.2850223779678345} 01/29/2022 09:05:22 - INFO - codeparrot_training - Step 15776: {'lr': 0.00040508168949413904, 'samples': 3029184, 'steps': 15776, 'loss/train': 0.810541570186615} 01/29/2022 09:05:26 - INFO - codeparrot_training - Step 15777: {'lr': 0.0004050688553716121, 'samples': 3029376, 'steps': 15777, 'loss/train': 1.9892250895500183} 01/29/2022 09:05:30 - INFO - codeparrot_training - Step 15778: {'lr': 0.0004050560205848194, 'samples': 3029568, 'steps': 15778, 'loss/train': 0.9986806511878967} 01/29/2022 09:05:35 - INFO - codeparrot_training - Step 15779: {'lr': 0.0004050431851338159, 'samples': 3029760, 'steps': 15779, 'loss/train': 1.7800235152244568} 01/29/2022 09:05:40 - INFO - codeparrot_training - Step 15780: {'lr': 0.00040503034901865666, 'samples': 3029952, 'steps': 15780, 'loss/train': 2.0461732149124146} 01/29/2022 09:05:44 - INFO - codeparrot_training - Step 15781: {'lr': 0.00040501751223939665, 'samples': 3030144, 'steps': 15781, 'loss/train': 1.772668719291687} 01/29/2022 09:05:49 - INFO - codeparrot_training - Step 15782: {'lr': 0.00040500467479609084, 'samples': 3030336, 'steps': 15782, 'loss/train': 1.596316158771515} 01/29/2022 09:05:53 - INFO - codeparrot_training - Step 15783: {'lr': 0.00040499183668879415, 'samples': 3030528, 'steps': 15783, 'loss/train': 0.6524758189916611} 01/29/2022 09:05:57 - INFO - codeparrot_training - Step 15784: {'lr': 0.0004049789979175617, 'samples': 3030720, 'steps': 15784, 'loss/train': 2.29544734954834} 01/29/2022 09:06:04 - INFO - codeparrot_training - Step 15785: {'lr': 0.00040496615848244845, 'samples': 3030912, 'steps': 15785, 'loss/train': 1.1273763477802277} 01/29/2022 09:06:08 - INFO - codeparrot_training - Step 15786: {'lr': 0.00040495331838350933, 'samples': 3031104, 'steps': 15786, 'loss/train': 2.0256566405296326} 01/29/2022 09:06:13 - INFO - codeparrot_training - Step 15787: {'lr': 0.00040494047762079953, 'samples': 3031296, 'steps': 15787, 'loss/train': 1.768447995185852} 01/29/2022 09:06:17 - INFO - codeparrot_training - Step 15788: {'lr': 0.0004049276361943738, 'samples': 3031488, 'steps': 15788, 'loss/train': 1.6783152222633362} 01/29/2022 09:06:21 - INFO - codeparrot_training - Step 15789: {'lr': 0.00040491479410428735, 'samples': 3031680, 'steps': 15789, 'loss/train': 1.580599844455719} 01/29/2022 09:06:27 - INFO - codeparrot_training - Step 15790: {'lr': 0.00040490195135059503, 'samples': 3031872, 'steps': 15790, 'loss/train': 1.7793912291526794} 01/29/2022 09:06:31 - INFO - codeparrot_training - Step 15791: {'lr': 0.000404889107933352, 'samples': 3032064, 'steps': 15791, 'loss/train': 0.19555049389600754} 01/29/2022 09:06:35 - INFO - codeparrot_training - Step 15792: {'lr': 0.0004048762638526132, 'samples': 3032256, 'steps': 15792, 'loss/train': 2.103805661201477} 01/29/2022 09:06:40 - INFO - codeparrot_training - Step 15793: {'lr': 0.0004048634191084336, 'samples': 3032448, 'steps': 15793, 'loss/train': 1.7867628335952759} 01/29/2022 09:06:45 - INFO - codeparrot_training - Step 15794: {'lr': 0.0004048505737008684, 'samples': 3032640, 'steps': 15794, 'loss/train': 1.9601765871047974} 01/29/2022 09:06:49 - INFO - codeparrot_training - Step 15795: {'lr': 0.0004048377276299724, 'samples': 3032832, 'steps': 15795, 'loss/train': 2.402946889400482} 01/29/2022 09:06:53 - INFO - codeparrot_training - Step 15796: {'lr': 0.00040482488089580083, 'samples': 3033024, 'steps': 15796, 'loss/train': 1.6950682997703552} 01/29/2022 09:06:57 - INFO - codeparrot_training - Step 15797: {'lr': 0.00040481203349840864, 'samples': 3033216, 'steps': 15797, 'loss/train': 1.1089743375778198} 01/29/2022 09:07:02 - INFO - codeparrot_training - Step 15798: {'lr': 0.0004047991854378508, 'samples': 3033408, 'steps': 15798, 'loss/train': 1.4121431708335876} 01/29/2022 09:07:09 - INFO - codeparrot_training - Step 15799: {'lr': 0.00040478633671418244, 'samples': 3033600, 'steps': 15799, 'loss/train': 2.2737273573875427} 01/29/2022 09:07:13 - INFO - codeparrot_training - Step 15800: {'lr': 0.00040477348732745853, 'samples': 3033792, 'steps': 15800, 'loss/train': 1.0030933320522308} 01/29/2022 09:07:17 - INFO - codeparrot_training - Step 15801: {'lr': 0.00040476063727773416, 'samples': 3033984, 'steps': 15801, 'loss/train': 1.9875630140304565} 01/29/2022 09:07:22 - INFO - codeparrot_training - Step 15802: {'lr': 0.0004047477865650644, 'samples': 3034176, 'steps': 15802, 'loss/train': 1.4014144241809845} 01/29/2022 09:07:26 - INFO - codeparrot_training - Step 15803: {'lr': 0.00040473493518950414, 'samples': 3034368, 'steps': 15803, 'loss/train': 1.5253722667694092} 01/29/2022 09:07:31 - INFO - codeparrot_training - Step 15804: {'lr': 0.00040472208315110866, 'samples': 3034560, 'steps': 15804, 'loss/train': 2.172639012336731} 01/29/2022 09:07:35 - INFO - codeparrot_training - Step 15805: {'lr': 0.0004047092304499329, 'samples': 3034752, 'steps': 15805, 'loss/train': 2.0445209741592407} 01/29/2022 09:07:40 - INFO - codeparrot_training - Step 15806: {'lr': 0.0004046963770860319, 'samples': 3034944, 'steps': 15806, 'loss/train': 0.18795908242464066} 01/29/2022 09:07:44 - INFO - codeparrot_training - Step 15807: {'lr': 0.0004046835230594608, 'samples': 3035136, 'steps': 15807, 'loss/train': 1.663870096206665} 01/29/2022 09:07:48 - INFO - codeparrot_training - Step 15808: {'lr': 0.0004046706683702744, 'samples': 3035328, 'steps': 15808, 'loss/train': 2.2189314365386963} 01/29/2022 09:07:55 - INFO - codeparrot_training - Step 15809: {'lr': 0.0004046578130185282, 'samples': 3035520, 'steps': 15809, 'loss/train': 1.7762550115585327} 01/29/2022 09:07:59 - INFO - codeparrot_training - Step 15810: {'lr': 0.00040464495700427694, 'samples': 3035712, 'steps': 15810, 'loss/train': 2.297776937484741} 01/29/2022 09:08:04 - INFO - codeparrot_training - Step 15811: {'lr': 0.0004046321003275759, 'samples': 3035904, 'steps': 15811, 'loss/train': 1.6604282855987549} 01/29/2022 09:08:08 - INFO - codeparrot_training - Step 15812: {'lr': 0.00040461924298847987, 'samples': 3036096, 'steps': 15812, 'loss/train': 3.3297961950302124} 01/29/2022 09:08:12 - INFO - codeparrot_training - Step 15813: {'lr': 0.0004046063849870442, 'samples': 3036288, 'steps': 15813, 'loss/train': 1.8810387253761292} 01/29/2022 09:08:17 - INFO - codeparrot_training - Step 15814: {'lr': 0.00040459352632332387, 'samples': 3036480, 'steps': 15814, 'loss/train': 1.4276057481765747} 01/29/2022 09:08:22 - INFO - codeparrot_training - Step 15815: {'lr': 0.0004045806669973739, 'samples': 3036672, 'steps': 15815, 'loss/train': 1.1482541263103485} 01/29/2022 09:08:26 - INFO - codeparrot_training - Step 15816: {'lr': 0.00040456780700924956, 'samples': 3036864, 'steps': 15816, 'loss/train': 2.279689371585846} 01/29/2022 09:08:30 - INFO - codeparrot_training - Step 15817: {'lr': 0.0004045549463590057, 'samples': 3037056, 'steps': 15817, 'loss/train': 1.3574218153953552} 01/29/2022 09:08:34 - INFO - codeparrot_training - Step 15818: {'lr': 0.0004045420850466975, 'samples': 3037248, 'steps': 15818, 'loss/train': 1.6250578165054321} 01/29/2022 09:08:40 - INFO - codeparrot_training - Step 15819: {'lr': 0.00040452922307238016, 'samples': 3037440, 'steps': 15819, 'loss/train': 1.4973648190498352} 01/29/2022 09:08:44 - INFO - codeparrot_training - Step 15820: {'lr': 0.00040451636043610875, 'samples': 3037632, 'steps': 15820, 'loss/train': 1.5950582027435303} 01/29/2022 09:08:48 - INFO - codeparrot_training - Step 15821: {'lr': 0.0004045034971379382, 'samples': 3037824, 'steps': 15821, 'loss/train': 1.936671495437622} 01/29/2022 09:08:52 - INFO - codeparrot_training - Step 15822: {'lr': 0.0004044906331779238, 'samples': 3038016, 'steps': 15822, 'loss/train': 1.2875267565250397} 01/29/2022 09:08:57 - INFO - codeparrot_training - Step 15823: {'lr': 0.00040447776855612053, 'samples': 3038208, 'steps': 15823, 'loss/train': 1.761625349521637} 01/29/2022 09:09:02 - INFO - codeparrot_training - Step 15824: {'lr': 0.0004044649032725836, 'samples': 3038400, 'steps': 15824, 'loss/train': 1.7579591274261475} 01/29/2022 09:09:06 - INFO - codeparrot_training - Step 15825: {'lr': 0.000404452037327368, 'samples': 3038592, 'steps': 15825, 'loss/train': 1.7739606499671936} 01/29/2022 09:09:10 - INFO - codeparrot_training - Step 15826: {'lr': 0.00040443917072052906, 'samples': 3038784, 'steps': 15826, 'loss/train': 1.153180181980133} 01/29/2022 09:09:15 - INFO - codeparrot_training - Step 15827: {'lr': 0.0004044263034521216, 'samples': 3038976, 'steps': 15827, 'loss/train': 1.080981284379959} 01/29/2022 09:09:19 - INFO - codeparrot_training - Step 15828: {'lr': 0.000404413435522201, 'samples': 3039168, 'steps': 15828, 'loss/train': 1.2978836596012115} 01/29/2022 09:09:26 - INFO - codeparrot_training - Step 15829: {'lr': 0.00040440056693082224, 'samples': 3039360, 'steps': 15829, 'loss/train': 1.746304214000702} 01/29/2022 09:09:30 - INFO - codeparrot_training - Step 15830: {'lr': 0.0004043876976780404, 'samples': 3039552, 'steps': 15830, 'loss/train': 2.2692131996154785} 01/29/2022 09:09:34 - INFO - codeparrot_training - Step 15831: {'lr': 0.0004043748277639108, 'samples': 3039744, 'steps': 15831, 'loss/train': 1.6723934412002563} 01/29/2022 09:09:39 - INFO - codeparrot_training - Step 15832: {'lr': 0.0004043619571884884, 'samples': 3039936, 'steps': 15832, 'loss/train': 1.9715943932533264} 01/29/2022 09:09:43 - INFO - codeparrot_training - Step 15833: {'lr': 0.0004043490859518284, 'samples': 3040128, 'steps': 15833, 'loss/train': 1.9268794655799866} 01/29/2022 09:09:48 - INFO - codeparrot_training - Step 15834: {'lr': 0.0004043362140539859, 'samples': 3040320, 'steps': 15834, 'loss/train': 0.8784551024436951} 01/29/2022 09:09:53 - INFO - codeparrot_training - Step 15835: {'lr': 0.00040432334149501613, 'samples': 3040512, 'steps': 15835, 'loss/train': 1.8549097180366516} 01/29/2022 09:09:57 - INFO - codeparrot_training - Step 15836: {'lr': 0.00040431046827497415, 'samples': 3040704, 'steps': 15836, 'loss/train': 6.818218231201172} 01/29/2022 09:10:01 - INFO - codeparrot_training - Step 15837: {'lr': 0.00040429759439391513, 'samples': 3040896, 'steps': 15837, 'loss/train': 1.055581659078598} 01/29/2022 09:10:05 - INFO - codeparrot_training - Step 15838: {'lr': 0.00040428471985189416, 'samples': 3041088, 'steps': 15838, 'loss/train': 2.2454981803894043} 01/29/2022 09:10:10 - INFO - codeparrot_training - Step 15839: {'lr': 0.0004042718446489665, 'samples': 3041280, 'steps': 15839, 'loss/train': 1.7505533695220947} 01/29/2022 09:10:15 - INFO - codeparrot_training - Step 15840: {'lr': 0.0004042589687851872, 'samples': 3041472, 'steps': 15840, 'loss/train': 1.056126058101654} 01/29/2022 09:10:19 - INFO - codeparrot_training - Step 15841: {'lr': 0.00040424609226061146, 'samples': 3041664, 'steps': 15841, 'loss/train': 1.8463559746742249} 01/29/2022 09:10:23 - INFO - codeparrot_training - Step 15842: {'lr': 0.0004042332150752944, 'samples': 3041856, 'steps': 15842, 'loss/train': 1.1920777559280396} 01/29/2022 09:10:27 - INFO - codeparrot_training - Step 15843: {'lr': 0.0004042203372292913, 'samples': 3042048, 'steps': 15843, 'loss/train': 0.9490391314029694} 01/29/2022 09:10:32 - INFO - codeparrot_training - Step 15844: {'lr': 0.00040420745872265726, 'samples': 3042240, 'steps': 15844, 'loss/train': 1.1476582288742065} 01/29/2022 09:10:39 - INFO - codeparrot_training - Step 15845: {'lr': 0.0004041945795554474, 'samples': 3042432, 'steps': 15845, 'loss/train': 2.488074839115143} 01/29/2022 09:10:43 - INFO - codeparrot_training - Step 15846: {'lr': 0.0004041816997277169, 'samples': 3042624, 'steps': 15846, 'loss/train': 1.6462962627410889} 01/29/2022 09:10:48 - INFO - codeparrot_training - Step 15847: {'lr': 0.000404168819239521, 'samples': 3042816, 'steps': 15847, 'loss/train': 1.5162352323532104} 01/29/2022 09:10:52 - INFO - codeparrot_training - Step 15848: {'lr': 0.0004041559380909148, 'samples': 3043008, 'steps': 15848, 'loss/train': 2.1325437426567078} 01/29/2022 09:10:57 - INFO - codeparrot_training - Step 15849: {'lr': 0.00040414305628195347, 'samples': 3043200, 'steps': 15849, 'loss/train': 0.9461259841918945} 01/29/2022 09:11:01 - INFO - codeparrot_training - Step 15850: {'lr': 0.00040413017381269237, 'samples': 3043392, 'steps': 15850, 'loss/train': 1.8992659449577332} 01/29/2022 09:11:06 - INFO - codeparrot_training - Step 15851: {'lr': 0.00040411729068318635, 'samples': 3043584, 'steps': 15851, 'loss/train': 1.0788329243659973} 01/29/2022 09:11:10 - INFO - codeparrot_training - Step 15852: {'lr': 0.0004041044068934909, 'samples': 3043776, 'steps': 15852, 'loss/train': 2.1555776596069336} 01/29/2022 09:11:14 - INFO - codeparrot_training - Step 15853: {'lr': 0.00040409152244366117, 'samples': 3043968, 'steps': 15853, 'loss/train': 0.577135294675827} 01/29/2022 09:11:21 - INFO - codeparrot_training - Step 15854: {'lr': 0.00040407863733375217, 'samples': 3044160, 'steps': 15854, 'loss/train': 1.803304374217987} 01/29/2022 09:11:26 - INFO - codeparrot_training - Step 15855: {'lr': 0.0004040657515638193, 'samples': 3044352, 'steps': 15855, 'loss/train': 1.978739619255066} 01/29/2022 09:11:30 - INFO - codeparrot_training - Step 15856: {'lr': 0.0004040528651339176, 'samples': 3044544, 'steps': 15856, 'loss/train': 1.6699329614639282} 01/29/2022 09:11:34 - INFO - codeparrot_training - Step 15857: {'lr': 0.00040403997804410244, 'samples': 3044736, 'steps': 15857, 'loss/train': 1.4849051535129547} 01/29/2022 09:11:38 - INFO - codeparrot_training - Step 15858: {'lr': 0.00040402709029442883, 'samples': 3044928, 'steps': 15858, 'loss/train': 1.5744212865829468} 01/29/2022 09:11:44 - INFO - codeparrot_training - Step 15859: {'lr': 0.0004040142018849521, 'samples': 3045120, 'steps': 15859, 'loss/train': 1.5051194429397583} 01/29/2022 09:11:48 - INFO - codeparrot_training - Step 15860: {'lr': 0.0004040013128157275, 'samples': 3045312, 'steps': 15860, 'loss/train': 2.198719024658203} 01/29/2022 09:11:52 - INFO - codeparrot_training - Step 15861: {'lr': 0.0004039884230868101, 'samples': 3045504, 'steps': 15861, 'loss/train': 1.8887379169464111} 01/29/2022 09:11:56 - INFO - codeparrot_training - Step 15862: {'lr': 0.0004039755326982552, 'samples': 3045696, 'steps': 15862, 'loss/train': 2.2645248770713806} 01/29/2022 09:12:01 - INFO - codeparrot_training - Step 15863: {'lr': 0.000403962641650118, 'samples': 3045888, 'steps': 15863, 'loss/train': 1.7501438856124878} 01/29/2022 09:12:06 - INFO - codeparrot_training - Step 15864: {'lr': 0.0004039497499424538, 'samples': 3046080, 'steps': 15864, 'loss/train': 1.3590266704559326} 01/29/2022 09:12:10 - INFO - codeparrot_training - Step 15865: {'lr': 0.00040393685757531776, 'samples': 3046272, 'steps': 15865, 'loss/train': 1.9106172323226929} 01/29/2022 09:12:15 - INFO - codeparrot_training - Step 15866: {'lr': 0.000403923964548765, 'samples': 3046464, 'steps': 15866, 'loss/train': 1.7808115482330322} 01/29/2022 09:12:19 - INFO - codeparrot_training - Step 15867: {'lr': 0.0004039110708628509, 'samples': 3046656, 'steps': 15867, 'loss/train': 1.7304258942604065} 01/29/2022 09:12:23 - INFO - codeparrot_training - Step 15868: {'lr': 0.00040389817651763073, 'samples': 3046848, 'steps': 15868, 'loss/train': 1.4321047067642212} 01/29/2022 09:12:30 - INFO - codeparrot_training - Step 15869: {'lr': 0.0004038852815131595, 'samples': 3047040, 'steps': 15869, 'loss/train': 1.8956401348114014} 01/29/2022 09:12:34 - INFO - codeparrot_training - Step 15870: {'lr': 0.0004038723858494927, 'samples': 3047232, 'steps': 15870, 'loss/train': 2.5252110958099365} 01/29/2022 09:12:39 - INFO - codeparrot_training - Step 15871: {'lr': 0.00040385948952668537, 'samples': 3047424, 'steps': 15871, 'loss/train': 1.2814320623874664} 01/29/2022 09:12:43 - INFO - codeparrot_training - Step 15872: {'lr': 0.0004038465925447929, 'samples': 3047616, 'steps': 15872, 'loss/train': 1.6413381099700928} 01/29/2022 09:12:47 - INFO - codeparrot_training - Step 15873: {'lr': 0.00040383369490387043, 'samples': 3047808, 'steps': 15873, 'loss/train': 2.1323888897895813} 01/29/2022 09:12:52 - INFO - codeparrot_training - Step 15874: {'lr': 0.0004038207966039733, 'samples': 3048000, 'steps': 15874, 'loss/train': 1.3808780014514923} 01/29/2022 09:12:57 - INFO - codeparrot_training - Step 15875: {'lr': 0.00040380789764515667, 'samples': 3048192, 'steps': 15875, 'loss/train': 1.2690866589546204} 01/29/2022 09:13:01 - INFO - codeparrot_training - Step 15876: {'lr': 0.0004037949980274759, 'samples': 3048384, 'steps': 15876, 'loss/train': 2.07533723115921} 01/29/2022 09:13:05 - INFO - codeparrot_training - Step 15877: {'lr': 0.0004037820977509862, 'samples': 3048576, 'steps': 15877, 'loss/train': 1.4377294778823853} 01/29/2022 09:13:09 - INFO - codeparrot_training - Step 15878: {'lr': 0.00040376919681574285, 'samples': 3048768, 'steps': 15878, 'loss/train': 1.663480281829834} 01/29/2022 09:13:15 - INFO - codeparrot_training - Step 15879: {'lr': 0.000403756295221801, 'samples': 3048960, 'steps': 15879, 'loss/train': 2.247421681880951} 01/29/2022 09:13:19 - INFO - codeparrot_training - Step 15880: {'lr': 0.00040374339296921606, 'samples': 3049152, 'steps': 15880, 'loss/train': 1.764414668083191} 01/29/2022 09:13:23 - INFO - codeparrot_training - Step 15881: {'lr': 0.00040373049005804323, 'samples': 3049344, 'steps': 15881, 'loss/train': 0.9876252114772797} 01/29/2022 09:13:27 - INFO - codeparrot_training - Step 15882: {'lr': 0.00040371758648833776, 'samples': 3049536, 'steps': 15882, 'loss/train': 0.8226064145565033} 01/29/2022 09:13:32 - INFO - codeparrot_training - Step 15883: {'lr': 0.00040370468226015507, 'samples': 3049728, 'steps': 15883, 'loss/train': 1.8273603916168213} 01/29/2022 09:13:37 - INFO - codeparrot_training - Step 15884: {'lr': 0.0004036917773735502, 'samples': 3049920, 'steps': 15884, 'loss/train': 1.7788172364234924} 01/29/2022 09:13:41 - INFO - codeparrot_training - Step 15885: {'lr': 0.00040367887182857866, 'samples': 3050112, 'steps': 15885, 'loss/train': 1.8757460117340088} 01/29/2022 09:13:45 - INFO - codeparrot_training - Step 15886: {'lr': 0.00040366596562529554, 'samples': 3050304, 'steps': 15886, 'loss/train': 1.7038162350654602} 01/29/2022 09:13:50 - INFO - codeparrot_training - Step 15887: {'lr': 0.00040365305876375636, 'samples': 3050496, 'steps': 15887, 'loss/train': 2.2904725670814514} 01/29/2022 09:13:54 - INFO - codeparrot_training - Step 15888: {'lr': 0.0004036401512440161, 'samples': 3050688, 'steps': 15888, 'loss/train': 1.6507158279418945} 01/29/2022 09:14:01 - INFO - codeparrot_training - Step 15889: {'lr': 0.0004036272430661303, 'samples': 3050880, 'steps': 15889, 'loss/train': 1.0095433294773102} 01/29/2022 09:14:05 - INFO - codeparrot_training - Step 15890: {'lr': 0.0004036143342301542, 'samples': 3051072, 'steps': 15890, 'loss/train': 1.814734697341919} 01/29/2022 09:14:09 - INFO - codeparrot_training - Step 15891: {'lr': 0.000403601424736143, 'samples': 3051264, 'steps': 15891, 'loss/train': 1.7150108814239502} 01/29/2022 09:14:14 - INFO - codeparrot_training - Step 15892: {'lr': 0.0004035885145841521, 'samples': 3051456, 'steps': 15892, 'loss/train': 1.1775349080562592} 01/29/2022 09:14:18 - INFO - codeparrot_training - Step 15893: {'lr': 0.00040357560377423675, 'samples': 3051648, 'steps': 15893, 'loss/train': 2.022306740283966} 01/29/2022 09:14:23 - INFO - codeparrot_training - Step 15894: {'lr': 0.0004035626923064524, 'samples': 3051840, 'steps': 15894, 'loss/train': 1.2451132535934448} 01/29/2022 09:14:27 - INFO - codeparrot_training - Step 15895: {'lr': 0.00040354978018085407, 'samples': 3052032, 'steps': 15895, 'loss/train': 1.4094392359256744} 01/29/2022 09:14:32 - INFO - codeparrot_training - Step 15896: {'lr': 0.00040353686739749733, 'samples': 3052224, 'steps': 15896, 'loss/train': 1.5618980526924133} 01/29/2022 09:14:36 - INFO - codeparrot_training - Step 15897: {'lr': 0.00040352395395643737, 'samples': 3052416, 'steps': 15897, 'loss/train': 1.2400891184806824} 01/29/2022 09:14:40 - INFO - codeparrot_training - Step 15898: {'lr': 0.00040351103985772964, 'samples': 3052608, 'steps': 15898, 'loss/train': 1.550605058670044} 01/29/2022 09:14:45 - INFO - codeparrot_training - Step 15899: {'lr': 0.00040349812510142923, 'samples': 3052800, 'steps': 15899, 'loss/train': 2.258391559123993} 01/29/2022 09:14:50 - INFO - codeparrot_training - Step 15900: {'lr': 0.0004034852096875916, 'samples': 3052992, 'steps': 15900, 'loss/train': 1.8108022212982178} 01/29/2022 09:14:54 - INFO - codeparrot_training - Step 15901: {'lr': 0.0004034722936162721, 'samples': 3053184, 'steps': 15901, 'loss/train': 1.897644817829132} 01/29/2022 09:14:58 - INFO - codeparrot_training - Step 15902: {'lr': 0.00040345937688752607, 'samples': 3053376, 'steps': 15902, 'loss/train': 1.0020193755626678} 01/29/2022 09:15:05 - INFO - codeparrot_training - Step 15903: {'lr': 0.0004034464595014088, 'samples': 3053568, 'steps': 15903, 'loss/train': 1.8780730962753296} 01/29/2022 09:15:10 - INFO - codeparrot_training - Step 15904: {'lr': 0.00040343354145797554, 'samples': 3053760, 'steps': 15904, 'loss/train': 1.9632834792137146} 01/29/2022 09:15:14 - INFO - codeparrot_training - Step 15905: {'lr': 0.0004034206227572818, 'samples': 3053952, 'steps': 15905, 'loss/train': 1.6144931316375732} 01/29/2022 09:15:18 - INFO - codeparrot_training - Step 15906: {'lr': 0.0004034077033993828, 'samples': 3054144, 'steps': 15906, 'loss/train': 1.920293390750885} 01/29/2022 09:15:22 - INFO - codeparrot_training - Step 15907: {'lr': 0.00040339478338433386, 'samples': 3054336, 'steps': 15907, 'loss/train': 2.4408363103866577} 01/29/2022 09:15:28 - INFO - codeparrot_training - Step 15908: {'lr': 0.0004033818627121904, 'samples': 3054528, 'steps': 15908, 'loss/train': 6.082031965255737} 01/29/2022 09:15:32 - INFO - codeparrot_training - Step 15909: {'lr': 0.00040336894138300777, 'samples': 3054720, 'steps': 15909, 'loss/train': 2.239634335041046} 01/29/2022 09:15:36 - INFO - codeparrot_training - Step 15910: {'lr': 0.0004033560193968413, 'samples': 3054912, 'steps': 15910, 'loss/train': 1.4528560638427734} 01/29/2022 09:15:40 - INFO - codeparrot_training - Step 15911: {'lr': 0.00040334309675374636, 'samples': 3055104, 'steps': 15911, 'loss/train': 1.9939761757850647} 01/29/2022 09:15:45 - INFO - codeparrot_training - Step 15912: {'lr': 0.0004033301734537782, 'samples': 3055296, 'steps': 15912, 'loss/train': 1.233426958322525} 01/29/2022 09:15:51 - INFO - codeparrot_training - Step 15913: {'lr': 0.0004033172494969923, 'samples': 3055488, 'steps': 15913, 'loss/train': 1.0851809084415436} 01/29/2022 09:15:55 - INFO - codeparrot_training - Step 15914: {'lr': 0.000403304324883444, 'samples': 3055680, 'steps': 15914, 'loss/train': 1.478596419095993} 01/29/2022 09:15:59 - INFO - codeparrot_training - Step 15915: {'lr': 0.00040329139961318863, 'samples': 3055872, 'steps': 15915, 'loss/train': 1.499066412448883} 01/29/2022 09:16:03 - INFO - codeparrot_training - Step 15916: {'lr': 0.00040327847368628163, 'samples': 3056064, 'steps': 15916, 'loss/train': 1.1353044211864471} 01/29/2022 09:16:08 - INFO - codeparrot_training - Step 15917: {'lr': 0.0004032655471027783, 'samples': 3056256, 'steps': 15917, 'loss/train': 1.0672600865364075} 01/29/2022 09:16:15 - INFO - codeparrot_training - Step 15918: {'lr': 0.000403252619862734, 'samples': 3056448, 'steps': 15918, 'loss/train': 1.3684524893760681} 01/29/2022 09:16:20 - INFO - codeparrot_training - Step 15919: {'lr': 0.0004032396919662041, 'samples': 3056640, 'steps': 15919, 'loss/train': 2.015312612056732} 01/29/2022 09:16:24 - INFO - codeparrot_training - Step 15920: {'lr': 0.00040322676341324415, 'samples': 3056832, 'steps': 15920, 'loss/train': 2.3362671732902527} 01/29/2022 09:16:28 - INFO - codeparrot_training - Step 15921: {'lr': 0.0004032138342039093, 'samples': 3057024, 'steps': 15921, 'loss/train': 1.831884741783142} 01/29/2022 09:16:32 - INFO - codeparrot_training - Step 15922: {'lr': 0.0004032009043382551, 'samples': 3057216, 'steps': 15922, 'loss/train': 1.7697032690048218} 01/29/2022 09:16:38 - INFO - codeparrot_training - Step 15923: {'lr': 0.0004031879738163368, 'samples': 3057408, 'steps': 15923, 'loss/train': 1.7934610247612} 01/29/2022 09:16:42 - INFO - codeparrot_training - Step 15924: {'lr': 0.00040317504263820994, 'samples': 3057600, 'steps': 15924, 'loss/train': 1.9819074869155884} 01/29/2022 09:16:46 - INFO - codeparrot_training - Step 15925: {'lr': 0.0004031621108039298, 'samples': 3057792, 'steps': 15925, 'loss/train': 1.118232786655426} 01/29/2022 09:16:50 - INFO - codeparrot_training - Step 15926: {'lr': 0.0004031491783135518, 'samples': 3057984, 'steps': 15926, 'loss/train': 1.6305320262908936} 01/29/2022 09:16:55 - INFO - codeparrot_training - Step 15927: {'lr': 0.0004031362451671314, 'samples': 3058176, 'steps': 15927, 'loss/train': 1.7720662951469421} 01/29/2022 09:17:02 - INFO - codeparrot_training - Step 15928: {'lr': 0.00040312331136472385, 'samples': 3058368, 'steps': 15928, 'loss/train': 0.7668691277503967} 01/29/2022 09:17:06 - INFO - codeparrot_training - Step 15929: {'lr': 0.00040311037690638477, 'samples': 3058560, 'steps': 15929, 'loss/train': 1.651951253414154} 01/29/2022 09:17:10 - INFO - codeparrot_training - Step 15930: {'lr': 0.00040309744179216936, 'samples': 3058752, 'steps': 15930, 'loss/train': 2.064578890800476} 01/29/2022 09:17:14 - INFO - codeparrot_training - Step 15931: {'lr': 0.0004030845060221332, 'samples': 3058944, 'steps': 15931, 'loss/train': 2.334549307823181} 01/29/2022 09:17:19 - INFO - codeparrot_training - Step 15932: {'lr': 0.00040307156959633154, 'samples': 3059136, 'steps': 15932, 'loss/train': 1.5490214824676514} 01/29/2022 09:17:24 - INFO - codeparrot_training - Step 15933: {'lr': 0.00040305863251482, 'samples': 3059328, 'steps': 15933, 'loss/train': 2.2018391489982605} 01/29/2022 09:17:28 - INFO - codeparrot_training - Step 15934: {'lr': 0.00040304569477765375, 'samples': 3059520, 'steps': 15934, 'loss/train': 1.399107813835144} 01/29/2022 09:17:33 - INFO - codeparrot_training - Step 15935: {'lr': 0.0004030327563848885, 'samples': 3059712, 'steps': 15935, 'loss/train': 1.9266818761825562} 01/29/2022 09:17:37 - INFO - codeparrot_training - Step 15936: {'lr': 0.00040301981733657934, 'samples': 3059904, 'steps': 15936, 'loss/train': 1.9460538625717163} 01/29/2022 09:17:41 - INFO - codeparrot_training - Step 15937: {'lr': 0.00040300687763278196, 'samples': 3060096, 'steps': 15937, 'loss/train': 2.2309876084327698} 01/29/2022 09:17:46 - INFO - codeparrot_training - Step 15938: {'lr': 0.0004029939372735517, 'samples': 3060288, 'steps': 15938, 'loss/train': 1.8152457475662231} 01/29/2022 09:17:51 - INFO - codeparrot_training - Step 15939: {'lr': 0.000402980996258944, 'samples': 3060480, 'steps': 15939, 'loss/train': 1.214030385017395} 01/29/2022 09:17:55 - INFO - codeparrot_training - Step 15940: {'lr': 0.00040296805458901427, 'samples': 3060672, 'steps': 15940, 'loss/train': 1.8160538077354431} 01/29/2022 09:17:59 - INFO - codeparrot_training - Step 15941: {'lr': 0.0004029551122638179, 'samples': 3060864, 'steps': 15941, 'loss/train': 1.931471049785614} 01/29/2022 09:18:03 - INFO - codeparrot_training - Step 15942: {'lr': 0.0004029421692834105, 'samples': 3061056, 'steps': 15942, 'loss/train': 1.9393250942230225} 01/29/2022 09:18:10 - INFO - codeparrot_training - Step 15943: {'lr': 0.0004029292256478474, 'samples': 3061248, 'steps': 15943, 'loss/train': 1.256534367799759} 01/29/2022 09:18:15 - INFO - codeparrot_training - Step 15944: {'lr': 0.00040291628135718404, 'samples': 3061440, 'steps': 15944, 'loss/train': 1.6917039155960083} 01/29/2022 09:18:19 - INFO - codeparrot_training - Step 15945: {'lr': 0.0004029033364114759, 'samples': 3061632, 'steps': 15945, 'loss/train': 1.9685039520263672} 01/29/2022 09:18:23 - INFO - codeparrot_training - Step 15946: {'lr': 0.00040289039081077837, 'samples': 3061824, 'steps': 15946, 'loss/train': 1.4056544601917267} 01/29/2022 09:18:27 - INFO - codeparrot_training - Step 15947: {'lr': 0.00040287744455514703, 'samples': 3062016, 'steps': 15947, 'loss/train': 1.733673870563507} 01/29/2022 09:18:33 - INFO - codeparrot_training - Step 15948: {'lr': 0.00040286449764463715, 'samples': 3062208, 'steps': 15948, 'loss/train': 1.7259222865104675} 01/29/2022 09:18:37 - INFO - codeparrot_training - Step 15949: {'lr': 0.0004028515500793044, 'samples': 3062400, 'steps': 15949, 'loss/train': 1.993053674697876} 01/29/2022 09:18:41 - INFO - codeparrot_training - Step 15950: {'lr': 0.0004028386018592041, 'samples': 3062592, 'steps': 15950, 'loss/train': 1.5868787169456482} 01/29/2022 09:18:46 - INFO - codeparrot_training - Step 15951: {'lr': 0.0004028256529843918, 'samples': 3062784, 'steps': 15951, 'loss/train': 1.8761349320411682} 01/29/2022 09:18:50 - INFO - codeparrot_training - Step 15952: {'lr': 0.00040281270345492295, 'samples': 3062976, 'steps': 15952, 'loss/train': 1.497162401676178} 01/29/2022 09:18:57 - INFO - codeparrot_training - Step 15953: {'lr': 0.00040279975327085294, 'samples': 3063168, 'steps': 15953, 'loss/train': 2.5623920559883118} 01/29/2022 09:19:01 - INFO - codeparrot_training - Step 15954: {'lr': 0.00040278680243223733, 'samples': 3063360, 'steps': 15954, 'loss/train': 1.9662208557128906} 01/29/2022 09:19:06 - INFO - codeparrot_training - Step 15955: {'lr': 0.00040277385093913154, 'samples': 3063552, 'steps': 15955, 'loss/train': 1.5210131406784058} 01/29/2022 09:19:10 - INFO - codeparrot_training - Step 15956: {'lr': 0.0004027608987915912, 'samples': 3063744, 'steps': 15956, 'loss/train': 2.215580463409424} 01/29/2022 09:19:14 - INFO - codeparrot_training - Step 15957: {'lr': 0.0004027479459896716, 'samples': 3063936, 'steps': 15957, 'loss/train': 2.320435345172882} 01/29/2022 09:19:19 - INFO - codeparrot_training - Step 15958: {'lr': 0.0004027349925334282, 'samples': 3064128, 'steps': 15958, 'loss/train': 1.4497727751731873} 01/29/2022 09:19:24 - INFO - codeparrot_training - Step 15959: {'lr': 0.00040272203842291676, 'samples': 3064320, 'steps': 15959, 'loss/train': 1.5297231674194336} 01/29/2022 09:19:28 - INFO - codeparrot_training - Step 15960: {'lr': 0.00040270908365819247, 'samples': 3064512, 'steps': 15960, 'loss/train': 1.655392348766327} 01/29/2022 09:19:32 - INFO - codeparrot_training - Step 15961: {'lr': 0.000402696128239311, 'samples': 3064704, 'steps': 15961, 'loss/train': 1.4419383108615875} 01/29/2022 09:19:36 - INFO - codeparrot_training - Step 15962: {'lr': 0.00040268317216632783, 'samples': 3064896, 'steps': 15962, 'loss/train': 1.3145779967308044} 01/29/2022 09:19:42 - INFO - codeparrot_training - Step 15963: {'lr': 0.0004026702154392984, 'samples': 3065088, 'steps': 15963, 'loss/train': 1.7280864715576172} 01/29/2022 09:19:46 - INFO - codeparrot_training - Step 15964: {'lr': 0.0004026572580582783, 'samples': 3065280, 'steps': 15964, 'loss/train': 1.339078813791275} 01/29/2022 09:19:50 - INFO - codeparrot_training - Step 15965: {'lr': 0.000402644300023323, 'samples': 3065472, 'steps': 15965, 'loss/train': 2.087270200252533} 01/29/2022 09:19:54 - INFO - codeparrot_training - Step 15966: {'lr': 0.0004026313413344879, 'samples': 3065664, 'steps': 15966, 'loss/train': 1.6591699719429016} 01/29/2022 09:19:59 - INFO - codeparrot_training - Step 15967: {'lr': 0.0004026183819918286, 'samples': 3065856, 'steps': 15967, 'loss/train': 1.7632965445518494} 01/29/2022 09:20:04 - INFO - codeparrot_training - Step 15968: {'lr': 0.00040260542199540064, 'samples': 3066048, 'steps': 15968, 'loss/train': 1.68978613615036} 01/29/2022 09:20:08 - INFO - codeparrot_training - Step 15969: {'lr': 0.00040259246134525953, 'samples': 3066240, 'steps': 15969, 'loss/train': 1.8833173513412476} 01/29/2022 09:20:12 - INFO - codeparrot_training - Step 15970: {'lr': 0.0004025795000414608, 'samples': 3066432, 'steps': 15970, 'loss/train': 1.7858020663261414} 01/29/2022 09:20:17 - INFO - codeparrot_training - Step 15971: {'lr': 0.0004025665380840599, 'samples': 3066624, 'steps': 15971, 'loss/train': 1.810580849647522} 01/29/2022 09:20:21 - INFO - codeparrot_training - Step 15972: {'lr': 0.00040255357547311235, 'samples': 3066816, 'steps': 15972, 'loss/train': 1.0957676768302917} 01/29/2022 09:20:28 - INFO - codeparrot_training - Step 15973: {'lr': 0.0004025406122086738, 'samples': 3067008, 'steps': 15973, 'loss/train': 1.6720499396324158} 01/29/2022 09:20:32 - INFO - codeparrot_training - Step 15974: {'lr': 0.0004025276482907996, 'samples': 3067200, 'steps': 15974, 'loss/train': 1.978411853313446} 01/29/2022 09:20:37 - INFO - codeparrot_training - Step 15975: {'lr': 0.0004025146837195455, 'samples': 3067392, 'steps': 15975, 'loss/train': 1.7392423152923584} 01/29/2022 09:20:41 - INFO - codeparrot_training - Step 15976: {'lr': 0.00040250171849496685, 'samples': 3067584, 'steps': 15976, 'loss/train': 1.2616603374481201} 01/29/2022 09:20:45 - INFO - codeparrot_training - Step 15977: {'lr': 0.0004024887526171193, 'samples': 3067776, 'steps': 15977, 'loss/train': 2.6636427640914917} 01/29/2022 09:20:50 - INFO - codeparrot_training - Step 15978: {'lr': 0.0004024757860860584, 'samples': 3067968, 'steps': 15978, 'loss/train': 2.125788152217865} 01/29/2022 09:20:55 - INFO - codeparrot_training - Step 15979: {'lr': 0.00040246281890183954, 'samples': 3068160, 'steps': 15979, 'loss/train': 1.666952133178711} 01/29/2022 09:20:59 - INFO - codeparrot_training - Step 15980: {'lr': 0.0004024498510645185, 'samples': 3068352, 'steps': 15980, 'loss/train': 1.7384732365608215} 01/29/2022 09:21:03 - INFO - codeparrot_training - Step 15981: {'lr': 0.00040243688257415064, 'samples': 3068544, 'steps': 15981, 'loss/train': 1.91730397939682} 01/29/2022 09:21:08 - INFO - codeparrot_training - Step 15982: {'lr': 0.00040242391343079157, 'samples': 3068736, 'steps': 15982, 'loss/train': 2.7130554914474487} 01/29/2022 09:21:13 - INFO - codeparrot_training - Step 15983: {'lr': 0.00040241094363449684, 'samples': 3068928, 'steps': 15983, 'loss/train': 1.1779786348342896} 01/29/2022 09:21:18 - INFO - codeparrot_training - Step 15984: {'lr': 0.000402397973185322, 'samples': 3069120, 'steps': 15984, 'loss/train': 0.8823498487472534} 01/29/2022 09:21:22 - INFO - codeparrot_training - Step 15985: {'lr': 0.0004023850020833227, 'samples': 3069312, 'steps': 15985, 'loss/train': 1.8409371972084045} 01/29/2022 09:21:26 - INFO - codeparrot_training - Step 15986: {'lr': 0.00040237203032855446, 'samples': 3069504, 'steps': 15986, 'loss/train': 1.470118135213852} 01/29/2022 09:21:30 - INFO - codeparrot_training - Step 15987: {'lr': 0.00040235905792107275, 'samples': 3069696, 'steps': 15987, 'loss/train': 1.4168379306793213} 01/29/2022 09:21:35 - INFO - codeparrot_training - Step 15988: {'lr': 0.00040234608486093326, 'samples': 3069888, 'steps': 15988, 'loss/train': 2.1111574172973633} 01/29/2022 09:21:41 - INFO - codeparrot_training - Step 15989: {'lr': 0.00040233311114819156, 'samples': 3070080, 'steps': 15989, 'loss/train': 1.2105537056922913} 01/29/2022 09:21:45 - INFO - codeparrot_training - Step 15990: {'lr': 0.00040232013678290316, 'samples': 3070272, 'steps': 15990, 'loss/train': 1.3847141861915588} 01/29/2022 09:21:49 - INFO - codeparrot_training - Step 15991: {'lr': 0.0004023071617651236, 'samples': 3070464, 'steps': 15991, 'loss/train': 1.277459979057312} 01/29/2022 09:21:54 - INFO - codeparrot_training - Step 15992: {'lr': 0.0004022941860949085, 'samples': 3070656, 'steps': 15992, 'loss/train': 1.7672118544578552} 01/29/2022 09:21:58 - INFO - codeparrot_training - Step 15993: {'lr': 0.00040228120977231355, 'samples': 3070848, 'steps': 15993, 'loss/train': 1.622049629688263} 01/29/2022 09:22:03 - INFO - codeparrot_training - Step 15994: {'lr': 0.00040226823279739427, 'samples': 3071040, 'steps': 15994, 'loss/train': 1.9929011464118958} 01/29/2022 09:22:08 - INFO - codeparrot_training - Step 15995: {'lr': 0.00040225525517020616, 'samples': 3071232, 'steps': 15995, 'loss/train': 1.6743324995040894} 01/29/2022 09:22:12 - INFO - codeparrot_training - Step 15996: {'lr': 0.0004022422768908049, 'samples': 3071424, 'steps': 15996, 'loss/train': 1.28694087266922} 01/29/2022 09:22:16 - INFO - codeparrot_training - Step 15997: {'lr': 0.00040222929795924613, 'samples': 3071616, 'steps': 15997, 'loss/train': 1.711976408958435} 01/29/2022 09:22:20 - INFO - codeparrot_training - Step 15998: {'lr': 0.0004022163183755853, 'samples': 3071808, 'steps': 15998, 'loss/train': 1.1591658890247345} 01/29/2022 09:22:27 - INFO - codeparrot_training - Step 15999: {'lr': 0.0004022033381398781, 'samples': 3072000, 'steps': 15999, 'loss/train': 2.6808194518089294} 01/29/2022 09:22:27 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 09:23:01 - WARNING - huggingface_hub.repository - Several commits (8) will be pushed upstream. 01/29/2022 09:23:01 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 09:24:21 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 6a0d43b..24f4fe8 expert-dust-2 -> expert-dust-2 01/29/2022 09:24:27 - INFO - codeparrot_training - Step 16000: {'lr': 0.0004021903572521802, 'samples': 3072192, 'steps': 16000, 'loss/train': 0.7151697725057602} 01/29/2022 09:24:31 - INFO - codeparrot_training - Step 16001: {'lr': 0.0004021773757125471, 'samples': 3072384, 'steps': 16001, 'loss/train': 2.498311400413513} 01/29/2022 09:24:36 - INFO - codeparrot_training - Step 16002: {'lr': 0.0004021643935210344, 'samples': 3072576, 'steps': 16002, 'loss/train': 2.1144672632217407} 01/29/2022 09:24:40 - INFO - codeparrot_training - Step 16003: {'lr': 0.0004021514106776978, 'samples': 3072768, 'steps': 16003, 'loss/train': 1.67921644449234} 01/29/2022 09:24:46 - INFO - codeparrot_training - Step 16004: {'lr': 0.00040213842718259287, 'samples': 3072960, 'steps': 16004, 'loss/train': 3.7155075073242188} 01/29/2022 09:24:50 - INFO - codeparrot_training - Step 16005: {'lr': 0.00040212544303577525, 'samples': 3073152, 'steps': 16005, 'loss/train': 1.8821927905082703} 01/29/2022 09:24:54 - INFO - codeparrot_training - Step 16006: {'lr': 0.00040211245823730047, 'samples': 3073344, 'steps': 16006, 'loss/train': 1.6106205582618713} 01/29/2022 09:24:58 - INFO - codeparrot_training - Step 16007: {'lr': 0.00040209947278722425, 'samples': 3073536, 'steps': 16007, 'loss/train': 1.7612839937210083} 01/29/2022 09:25:03 - INFO - codeparrot_training - Step 16008: {'lr': 0.0004020864866856022, 'samples': 3073728, 'steps': 16008, 'loss/train': 1.3063985109329224} 01/29/2022 09:25:08 - INFO - codeparrot_training - Step 16009: {'lr': 0.0004020734999324899, 'samples': 3073920, 'steps': 16009, 'loss/train': 0.11523842811584473} 01/29/2022 09:25:12 - INFO - codeparrot_training - Step 16010: {'lr': 0.0004020605125279431, 'samples': 3074112, 'steps': 16010, 'loss/train': 0.7590537071228027} 01/29/2022 09:25:16 - INFO - codeparrot_training - Step 16011: {'lr': 0.0004020475244720173, 'samples': 3074304, 'steps': 16011, 'loss/train': 1.663329005241394} 01/29/2022 09:25:21 - INFO - codeparrot_training - Step 16012: {'lr': 0.0004020345357647681, 'samples': 3074496, 'steps': 16012, 'loss/train': 1.6949028968811035} 01/29/2022 09:25:25 - INFO - codeparrot_training - Step 16013: {'lr': 0.0004020215464062513, 'samples': 3074688, 'steps': 16013, 'loss/train': 2.9074294567108154} 01/29/2022 09:25:32 - INFO - codeparrot_training - Step 16014: {'lr': 0.0004020085563965226, 'samples': 3074880, 'steps': 16014, 'loss/train': 1.4916483163833618} 01/29/2022 09:25:36 - INFO - codeparrot_training - Step 16015: {'lr': 0.00040199556573563736, 'samples': 3075072, 'steps': 16015, 'loss/train': 1.6826840043067932} 01/29/2022 09:25:41 - INFO - codeparrot_training - Step 16016: {'lr': 0.0004019825744236514, 'samples': 3075264, 'steps': 16016, 'loss/train': 1.7819677591323853} 01/29/2022 09:25:45 - INFO - codeparrot_training - Step 16017: {'lr': 0.00040196958246062033, 'samples': 3075456, 'steps': 16017, 'loss/train': 1.6597768664360046} 01/29/2022 09:25:49 - INFO - codeparrot_training - Step 16018: {'lr': 0.00040195658984659987, 'samples': 3075648, 'steps': 16018, 'loss/train': 1.9836852550506592} 01/29/2022 09:25:53 - INFO - codeparrot_training - Step 16019: {'lr': 0.0004019435965816456, 'samples': 3075840, 'steps': 16019, 'loss/train': 1.8603547811508179} 01/29/2022 09:25:59 - INFO - codeparrot_training - Step 16020: {'lr': 0.0004019306026658132, 'samples': 3076032, 'steps': 16020, 'loss/train': 1.4741830229759216} 01/29/2022 09:26:03 - INFO - codeparrot_training - Step 16021: {'lr': 0.00040191760809915833, 'samples': 3076224, 'steps': 16021, 'loss/train': 2.0461501479148865} 01/29/2022 09:26:07 - INFO - codeparrot_training - Step 16022: {'lr': 0.00040190461288173675, 'samples': 3076416, 'steps': 16022, 'loss/train': 1.5811328887939453} 01/29/2022 09:26:12 - INFO - codeparrot_training - Step 16023: {'lr': 0.000401891617013604, 'samples': 3076608, 'steps': 16023, 'loss/train': 2.342420518398285} 01/29/2022 09:26:17 - INFO - codeparrot_training - Step 16024: {'lr': 0.00040187862049481573, 'samples': 3076800, 'steps': 16024, 'loss/train': 1.3272047638893127} 01/29/2022 09:26:21 - INFO - codeparrot_training - Step 16025: {'lr': 0.00040186562332542773, 'samples': 3076992, 'steps': 16025, 'loss/train': 1.9637991786003113} 01/29/2022 09:26:25 - INFO - codeparrot_training - Step 16026: {'lr': 0.0004018526255054956, 'samples': 3077184, 'steps': 16026, 'loss/train': 1.6400068402290344} 01/29/2022 09:26:30 - INFO - codeparrot_training - Step 16027: {'lr': 0.00040183962703507515, 'samples': 3077376, 'steps': 16027, 'loss/train': 1.7337464690208435} 01/29/2022 09:26:34 - INFO - codeparrot_training - Step 16028: {'lr': 0.00040182662791422185, 'samples': 3077568, 'steps': 16028, 'loss/train': 1.7533255219459534} 01/29/2022 09:26:40 - INFO - codeparrot_training - Step 16029: {'lr': 0.0004018136281429915, 'samples': 3077760, 'steps': 16029, 'loss/train': 3.5224757194519043} 01/29/2022 09:26:44 - INFO - codeparrot_training - Step 16030: {'lr': 0.0004018006277214398, 'samples': 3077952, 'steps': 16030, 'loss/train': 1.7645387649536133} 01/29/2022 09:26:49 - INFO - codeparrot_training - Step 16031: {'lr': 0.00040178762664962235, 'samples': 3078144, 'steps': 16031, 'loss/train': 0.8732951581478119} 01/29/2022 09:26:53 - INFO - codeparrot_training - Step 16032: {'lr': 0.000401774624927595, 'samples': 3078336, 'steps': 16032, 'loss/train': 1.415795624256134} 01/29/2022 09:26:57 - INFO - codeparrot_training - Step 16033: {'lr': 0.00040176162255541325, 'samples': 3078528, 'steps': 16033, 'loss/train': 1.6387327909469604} 01/29/2022 09:27:03 - INFO - codeparrot_training - Step 16034: {'lr': 0.00040174861953313297, 'samples': 3078720, 'steps': 16034, 'loss/train': 2.009262442588806} 01/29/2022 09:27:07 - INFO - codeparrot_training - Step 16035: {'lr': 0.00040173561586080974, 'samples': 3078912, 'steps': 16035, 'loss/train': 1.5819234251976013} 01/29/2022 09:27:11 - INFO - codeparrot_training - Step 16036: {'lr': 0.0004017226115384994, 'samples': 3079104, 'steps': 16036, 'loss/train': 0.8649418652057648} 01/29/2022 09:27:15 - INFO - codeparrot_training - Step 16037: {'lr': 0.00040170960656625744, 'samples': 3079296, 'steps': 16037, 'loss/train': 1.6062474846839905} 01/29/2022 09:27:20 - INFO - codeparrot_training - Step 16038: {'lr': 0.00040169660094413977, 'samples': 3079488, 'steps': 16038, 'loss/train': 1.263925015926361} 01/29/2022 09:27:25 - INFO - codeparrot_training - Step 16039: {'lr': 0.00040168359467220206, 'samples': 3079680, 'steps': 16039, 'loss/train': 1.30784672498703} 01/29/2022 09:27:29 - INFO - codeparrot_training - Step 16040: {'lr': 0.00040167058775049993, 'samples': 3079872, 'steps': 16040, 'loss/train': 6.535450458526611} 01/29/2022 09:27:34 - INFO - codeparrot_training - Step 16041: {'lr': 0.0004016575801790892, 'samples': 3080064, 'steps': 16041, 'loss/train': 1.6979899406433105} 01/29/2022 09:27:38 - INFO - codeparrot_training - Step 16042: {'lr': 0.0004016445719580256, 'samples': 3080256, 'steps': 16042, 'loss/train': 1.2836476564407349} 01/29/2022 09:27:42 - INFO - codeparrot_training - Step 16043: {'lr': 0.0004016315630873647, 'samples': 3080448, 'steps': 16043, 'loss/train': 1.2211732864379883} 01/29/2022 09:27:46 - INFO - codeparrot_training - Step 16044: {'lr': 0.00040161855356716245, 'samples': 3080640, 'steps': 16044, 'loss/train': 1.7170974612236023} 01/29/2022 09:27:52 - INFO - codeparrot_training - Step 16045: {'lr': 0.00040160554339747434, 'samples': 3080832, 'steps': 16045, 'loss/train': 1.922326683998108} 01/29/2022 09:27:56 - INFO - codeparrot_training - Step 16046: {'lr': 0.00040159253257835624, 'samples': 3081024, 'steps': 16046, 'loss/train': 1.4981196820735931} 01/29/2022 09:28:00 - INFO - codeparrot_training - Step 16047: {'lr': 0.00040157952110986397, 'samples': 3081216, 'steps': 16047, 'loss/train': 2.529182732105255} 01/29/2022 09:28:04 - INFO - codeparrot_training - Step 16048: {'lr': 0.00040156650899205305, 'samples': 3081408, 'steps': 16048, 'loss/train': 1.236330270767212} 01/29/2022 09:28:09 - INFO - codeparrot_training - Step 16049: {'lr': 0.00040155349622497937, 'samples': 3081600, 'steps': 16049, 'loss/train': 1.6333376169204712} 01/29/2022 09:28:16 - INFO - codeparrot_training - Step 16050: {'lr': 0.0004015404828086987, 'samples': 3081792, 'steps': 16050, 'loss/train': 1.7191194891929626} 01/29/2022 09:28:20 - INFO - codeparrot_training - Step 16051: {'lr': 0.0004015274687432667, 'samples': 3081984, 'steps': 16051, 'loss/train': 1.988234281539917} 01/29/2022 09:28:24 - INFO - codeparrot_training - Step 16052: {'lr': 0.0004015144540287391, 'samples': 3082176, 'steps': 16052, 'loss/train': 2.0188552737236023} 01/29/2022 09:28:28 - INFO - codeparrot_training - Step 16053: {'lr': 0.00040150143866517164, 'samples': 3082368, 'steps': 16053, 'loss/train': 2.6098535656929016} 01/29/2022 09:28:33 - INFO - codeparrot_training - Step 16054: {'lr': 0.0004014884226526202, 'samples': 3082560, 'steps': 16054, 'loss/train': 1.1091047823429108} 01/29/2022 09:28:38 - INFO - codeparrot_training - Step 16055: {'lr': 0.0004014754059911405, 'samples': 3082752, 'steps': 16055, 'loss/train': 1.2211859822273254} 01/29/2022 09:28:42 - INFO - codeparrot_training - Step 16056: {'lr': 0.0004014623886807882, 'samples': 3082944, 'steps': 16056, 'loss/train': 1.4241421222686768} 01/29/2022 09:28:46 - INFO - codeparrot_training - Step 16057: {'lr': 0.0004014493707216191, 'samples': 3083136, 'steps': 16057, 'loss/train': 1.6696722507476807} 01/29/2022 09:28:51 - INFO - codeparrot_training - Step 16058: {'lr': 0.00040143635211368903, 'samples': 3083328, 'steps': 16058, 'loss/train': 1.433230698108673} 01/29/2022 09:28:55 - INFO - codeparrot_training - Step 16059: {'lr': 0.0004014233328570537, 'samples': 3083520, 'steps': 16059, 'loss/train': 2.3994417786598206} 01/29/2022 09:29:02 - INFO - codeparrot_training - Step 16060: {'lr': 0.0004014103129517689, 'samples': 3083712, 'steps': 16060, 'loss/train': 1.2569689750671387} 01/29/2022 09:29:06 - INFO - codeparrot_training - Step 16061: {'lr': 0.00040139729239789036, 'samples': 3083904, 'steps': 16061, 'loss/train': 1.605882704257965} 01/29/2022 09:29:11 - INFO - codeparrot_training - Step 16062: {'lr': 0.0004013842711954739, 'samples': 3084096, 'steps': 16062, 'loss/train': 1.7664059400558472} 01/29/2022 09:29:15 - INFO - codeparrot_training - Step 16063: {'lr': 0.0004013712493445753, 'samples': 3084288, 'steps': 16063, 'loss/train': 1.7942132949829102} 01/29/2022 09:29:20 - INFO - codeparrot_training - Step 16064: {'lr': 0.00040135822684525036, 'samples': 3084480, 'steps': 16064, 'loss/train': 1.5873827934265137} 01/29/2022 09:29:24 - INFO - codeparrot_training - Step 16065: {'lr': 0.0004013452036975548, 'samples': 3084672, 'steps': 16065, 'loss/train': 2.067289888858795} 01/29/2022 09:29:29 - INFO - codeparrot_training - Step 16066: {'lr': 0.0004013321799015445, 'samples': 3084864, 'steps': 16066, 'loss/train': 1.528454840183258} 01/29/2022 09:29:33 - INFO - codeparrot_training - Step 16067: {'lr': 0.00040131915545727517, 'samples': 3085056, 'steps': 16067, 'loss/train': 1.1266776323318481} 01/29/2022 09:29:37 - INFO - codeparrot_training - Step 16068: {'lr': 0.00040130613036480265, 'samples': 3085248, 'steps': 16068, 'loss/train': 1.712391972541809} 01/29/2022 09:29:43 - INFO - codeparrot_training - Step 16069: {'lr': 0.0004012931046241827, 'samples': 3085440, 'steps': 16069, 'loss/train': 1.8348873853683472} 01/29/2022 09:29:47 - INFO - codeparrot_training - Step 16070: {'lr': 0.00040128007823547106, 'samples': 3085632, 'steps': 16070, 'loss/train': 1.3453968465328217} 01/29/2022 09:29:51 - INFO - codeparrot_training - Step 16071: {'lr': 0.00040126705119872367, 'samples': 3085824, 'steps': 16071, 'loss/train': 1.8421266674995422} 01/29/2022 09:29:55 - INFO - codeparrot_training - Step 16072: {'lr': 0.00040125402351399623, 'samples': 3086016, 'steps': 16072, 'loss/train': 1.0995557606220245} 01/29/2022 09:30:00 - INFO - codeparrot_training - Step 16073: {'lr': 0.0004012409951813446, 'samples': 3086208, 'steps': 16073, 'loss/train': 1.687647521495819} 01/29/2022 09:30:07 - INFO - codeparrot_training - Step 16074: {'lr': 0.0004012279662008246, 'samples': 3086400, 'steps': 16074, 'loss/train': 2.3401442170143127} 01/29/2022 09:30:11 - INFO - codeparrot_training - Step 16075: {'lr': 0.000401214936572492, 'samples': 3086592, 'steps': 16075, 'loss/train': 1.1085634231567383} 01/29/2022 09:30:15 - INFO - codeparrot_training - Step 16076: {'lr': 0.0004012019062964026, 'samples': 3086784, 'steps': 16076, 'loss/train': 1.053724765777588} 01/29/2022 09:30:20 - INFO - codeparrot_training - Step 16077: {'lr': 0.0004011888753726123, 'samples': 3086976, 'steps': 16077, 'loss/train': 1.9222918152809143} 01/29/2022 09:30:24 - INFO - codeparrot_training - Step 16078: {'lr': 0.00040117584380117675, 'samples': 3087168, 'steps': 16078, 'loss/train': 1.5883591175079346} 01/29/2022 09:30:29 - INFO - codeparrot_training - Step 16079: {'lr': 0.000401162811582152, 'samples': 3087360, 'steps': 16079, 'loss/train': 1.390828549861908} 01/29/2022 09:30:34 - INFO - codeparrot_training - Step 16080: {'lr': 0.00040114977871559375, 'samples': 3087552, 'steps': 16080, 'loss/train': 2.071602165699005} 01/29/2022 09:30:38 - INFO - codeparrot_training - Step 16081: {'lr': 0.0004011367452015578, 'samples': 3087744, 'steps': 16081, 'loss/train': 1.6311905980110168} 01/29/2022 09:30:42 - INFO - codeparrot_training - Step 16082: {'lr': 0.00040112371104010004, 'samples': 3087936, 'steps': 16082, 'loss/train': 1.9674893617630005} 01/29/2022 09:30:46 - INFO - codeparrot_training - Step 16083: {'lr': 0.00040111067623127626, 'samples': 3088128, 'steps': 16083, 'loss/train': 1.337884783744812} 01/29/2022 09:30:53 - INFO - codeparrot_training - Step 16084: {'lr': 0.0004010976407751424, 'samples': 3088320, 'steps': 16084, 'loss/train': 1.7399552464485168} 01/29/2022 09:30:58 - INFO - codeparrot_training - Step 16085: {'lr': 0.00040108460467175425, 'samples': 3088512, 'steps': 16085, 'loss/train': 1.6039357781410217} 01/29/2022 09:31:02 - INFO - codeparrot_training - Step 16086: {'lr': 0.00040107156792116753, 'samples': 3088704, 'steps': 16086, 'loss/train': 1.8542683124542236} 01/29/2022 09:31:06 - INFO - codeparrot_training - Step 16087: {'lr': 0.0004010585305234382, 'samples': 3088896, 'steps': 16087, 'loss/train': 2.1007116436958313} 01/29/2022 09:31:10 - INFO - codeparrot_training - Step 16088: {'lr': 0.00040104549247862217, 'samples': 3089088, 'steps': 16088, 'loss/train': 0.6850930899381638} 01/29/2022 09:31:16 - INFO - codeparrot_training - Step 16089: {'lr': 0.0004010324537867751, 'samples': 3089280, 'steps': 16089, 'loss/train': 1.5547435283660889} 01/29/2022 09:31:20 - INFO - codeparrot_training - Step 16090: {'lr': 0.000401019414447953, 'samples': 3089472, 'steps': 16090, 'loss/train': 1.7412164211273193} 01/29/2022 09:31:24 - INFO - codeparrot_training - Step 16091: {'lr': 0.0004010063744622117, 'samples': 3089664, 'steps': 16091, 'loss/train': 1.7443440556526184} 01/29/2022 09:31:28 - INFO - codeparrot_training - Step 16092: {'lr': 0.00040099333382960707, 'samples': 3089856, 'steps': 16092, 'loss/train': 1.540110468864441} 01/29/2022 09:31:33 - INFO - codeparrot_training - Step 16093: {'lr': 0.00040098029255019484, 'samples': 3090048, 'steps': 16093, 'loss/train': 1.98995840549469} 01/29/2022 09:31:38 - INFO - codeparrot_training - Step 16094: {'lr': 0.0004009672506240311, 'samples': 3090240, 'steps': 16094, 'loss/train': 1.1036542654037476} 01/29/2022 09:31:42 - INFO - codeparrot_training - Step 16095: {'lr': 0.00040095420805117153, 'samples': 3090432, 'steps': 16095, 'loss/train': 1.6944671273231506} 01/29/2022 09:31:46 - INFO - codeparrot_training - Step 16096: {'lr': 0.0004009411648316721, 'samples': 3090624, 'steps': 16096, 'loss/train': 2.668065905570984} 01/29/2022 09:31:51 - INFO - codeparrot_training - Step 16097: {'lr': 0.0004009281209655886, 'samples': 3090816, 'steps': 16097, 'loss/train': 1.8709961771965027} 01/29/2022 09:31:55 - INFO - codeparrot_training - Step 16098: {'lr': 0.000400915076452977, 'samples': 3091008, 'steps': 16098, 'loss/train': 2.049090564250946} 01/29/2022 09:32:00 - INFO - codeparrot_training - Step 16099: {'lr': 0.0004009020312938931, 'samples': 3091200, 'steps': 16099, 'loss/train': 1.937486708164215} 01/29/2022 09:32:05 - INFO - codeparrot_training - Step 16100: {'lr': 0.0004008889854883929, 'samples': 3091392, 'steps': 16100, 'loss/train': 1.0199356377124786} 01/29/2022 09:32:09 - INFO - codeparrot_training - Step 16101: {'lr': 0.0004008759390365321, 'samples': 3091584, 'steps': 16101, 'loss/train': 1.5860542058944702} 01/29/2022 09:32:13 - INFO - codeparrot_training - Step 16102: {'lr': 0.00040086289193836674, 'samples': 3091776, 'steps': 16102, 'loss/train': 4.094605207443237} 01/29/2022 09:32:17 - INFO - codeparrot_training - Step 16103: {'lr': 0.00040084984419395264, 'samples': 3091968, 'steps': 16103, 'loss/train': 1.56448495388031} 01/29/2022 09:32:25 - INFO - codeparrot_training - Step 16104: {'lr': 0.00040083679580334565, 'samples': 3092160, 'steps': 16104, 'loss/train': 1.246147245168686} 01/29/2022 09:32:29 - INFO - codeparrot_training - Step 16105: {'lr': 0.00040082374676660176, 'samples': 3092352, 'steps': 16105, 'loss/train': 2.124193847179413} 01/29/2022 09:32:33 - INFO - codeparrot_training - Step 16106: {'lr': 0.00040081069708377686, 'samples': 3092544, 'steps': 16106, 'loss/train': 1.616718828678131} 01/29/2022 09:32:37 - INFO - codeparrot_training - Step 16107: {'lr': 0.0004007976467549268, 'samples': 3092736, 'steps': 16107, 'loss/train': 1.4747883081436157} 01/29/2022 09:32:42 - INFO - codeparrot_training - Step 16108: {'lr': 0.0004007845957801075, 'samples': 3092928, 'steps': 16108, 'loss/train': 2.279394507408142} 01/29/2022 09:32:47 - INFO - codeparrot_training - Step 16109: {'lr': 0.0004007715441593749, 'samples': 3093120, 'steps': 16109, 'loss/train': 1.7296788096427917} 01/29/2022 09:32:51 - INFO - codeparrot_training - Step 16110: {'lr': 0.0004007584918927849, 'samples': 3093312, 'steps': 16110, 'loss/train': 3.098236083984375} 01/29/2022 09:32:55 - INFO - codeparrot_training - Step 16111: {'lr': 0.0004007454389803933, 'samples': 3093504, 'steps': 16111, 'loss/train': 1.5479249954223633} 01/29/2022 09:33:00 - INFO - codeparrot_training - Step 16112: {'lr': 0.00040073238542225623, 'samples': 3093696, 'steps': 16112, 'loss/train': 1.4556279480457306} 01/29/2022 09:33:04 - INFO - codeparrot_training - Step 16113: {'lr': 0.00040071933121842943, 'samples': 3093888, 'steps': 16113, 'loss/train': 1.6622266173362732} 01/29/2022 09:33:09 - INFO - codeparrot_training - Step 16114: {'lr': 0.00040070627636896886, 'samples': 3094080, 'steps': 16114, 'loss/train': 1.3132501244544983} 01/29/2022 09:33:13 - INFO - codeparrot_training - Step 16115: {'lr': 0.0004006932208739304, 'samples': 3094272, 'steps': 16115, 'loss/train': 1.8295143842697144} 01/29/2022 09:33:18 - INFO - codeparrot_training - Step 16116: {'lr': 0.0004006801647333701, 'samples': 3094464, 'steps': 16116, 'loss/train': 1.8344839811325073} 01/29/2022 09:33:22 - INFO - codeparrot_training - Step 16117: {'lr': 0.0004006671079473438, 'samples': 3094656, 'steps': 16117, 'loss/train': 1.730781376361847} 01/29/2022 09:33:26 - INFO - codeparrot_training - Step 16118: {'lr': 0.00040065405051590745, 'samples': 3094848, 'steps': 16118, 'loss/train': 2.1651358008384705} 01/29/2022 09:33:33 - INFO - codeparrot_training - Step 16119: {'lr': 0.000400640992439117, 'samples': 3095040, 'steps': 16119, 'loss/train': 1.1407520771026611} 01/29/2022 09:33:37 - INFO - codeparrot_training - Step 16120: {'lr': 0.0004006279337170283, 'samples': 3095232, 'steps': 16120, 'loss/train': 1.6346497535705566} 01/29/2022 09:33:42 - INFO - codeparrot_training - Step 16121: {'lr': 0.00040061487434969744, 'samples': 3095424, 'steps': 16121, 'loss/train': 1.3453713655471802} 01/29/2022 09:33:46 - INFO - codeparrot_training - Step 16122: {'lr': 0.00040060181433718037, 'samples': 3095616, 'steps': 16122, 'loss/train': 2.071579098701477} 01/29/2022 09:33:50 - INFO - codeparrot_training - Step 16123: {'lr': 0.00040058875367953285, 'samples': 3095808, 'steps': 16123, 'loss/train': 1.6583905220031738} 01/29/2022 09:33:55 - INFO - codeparrot_training - Step 16124: {'lr': 0.0004005756923768109, 'samples': 3096000, 'steps': 16124, 'loss/train': 1.8183256387710571} 01/29/2022 09:34:00 - INFO - codeparrot_training - Step 16125: {'lr': 0.0004005626304290705, 'samples': 3096192, 'steps': 16125, 'loss/train': 1.36117422580719} 01/29/2022 09:34:04 - INFO - codeparrot_training - Step 16126: {'lr': 0.00040054956783636765, 'samples': 3096384, 'steps': 16126, 'loss/train': 2.749576985836029} 01/29/2022 09:34:08 - INFO - codeparrot_training - Step 16127: {'lr': 0.00040053650459875823, 'samples': 3096576, 'steps': 16127, 'loss/train': 1.3259802460670471} 01/29/2022 09:34:12 - INFO - codeparrot_training - Step 16128: {'lr': 0.0004005234407162982, 'samples': 3096768, 'steps': 16128, 'loss/train': 1.8268484473228455} 01/29/2022 09:34:18 - INFO - codeparrot_training - Step 16129: {'lr': 0.00040051037618904365, 'samples': 3096960, 'steps': 16129, 'loss/train': 1.269326537847519} 01/29/2022 09:34:22 - INFO - codeparrot_training - Step 16130: {'lr': 0.0004004973110170503, 'samples': 3097152, 'steps': 16130, 'loss/train': 0.5224051773548126} 01/29/2022 09:34:26 - INFO - codeparrot_training - Step 16131: {'lr': 0.0004004842452003743, 'samples': 3097344, 'steps': 16131, 'loss/train': 0.49524687230587006} 01/29/2022 09:34:31 - INFO - codeparrot_training - Step 16132: {'lr': 0.0004004711787390716, 'samples': 3097536, 'steps': 16132, 'loss/train': 1.5038049817085266} 01/29/2022 09:34:35 - INFO - codeparrot_training - Step 16133: {'lr': 0.0004004581116331981, 'samples': 3097728, 'steps': 16133, 'loss/train': 2.026257812976837} 01/29/2022 09:34:41 - INFO - codeparrot_training - Step 16134: {'lr': 0.00040044504388280996, 'samples': 3097920, 'steps': 16134, 'loss/train': 0.9817058444023132} 01/29/2022 09:34:45 - INFO - codeparrot_training - Step 16135: {'lr': 0.00040043197548796295, 'samples': 3098112, 'steps': 16135, 'loss/train': 1.1383106410503387} 01/29/2022 09:34:50 - INFO - codeparrot_training - Step 16136: {'lr': 0.0004004189064487131, 'samples': 3098304, 'steps': 16136, 'loss/train': 1.410020112991333} 01/29/2022 09:34:54 - INFO - codeparrot_training - Step 16137: {'lr': 0.00040040583676511645, 'samples': 3098496, 'steps': 16137, 'loss/train': 0.9637947678565979} 01/29/2022 09:34:58 - INFO - codeparrot_training - Step 16138: {'lr': 0.0004003927664372289, 'samples': 3098688, 'steps': 16138, 'loss/train': 1.2356614172458649} 01/29/2022 09:35:04 - INFO - codeparrot_training - Step 16139: {'lr': 0.00040037969546510653, 'samples': 3098880, 'steps': 16139, 'loss/train': 1.7057674527168274} 01/29/2022 09:35:08 - INFO - codeparrot_training - Step 16140: {'lr': 0.0004003666238488053, 'samples': 3099072, 'steps': 16140, 'loss/train': 2.4422571659088135} 01/29/2022 09:35:12 - INFO - codeparrot_training - Step 16141: {'lr': 0.00040035355158838114, 'samples': 3099264, 'steps': 16141, 'loss/train': 1.7026941776275635} 01/29/2022 09:35:16 - INFO - codeparrot_training - Step 16142: {'lr': 0.0004003404786838902, 'samples': 3099456, 'steps': 16142, 'loss/train': 1.9203298687934875} 01/29/2022 09:35:24 - INFO - codeparrot_training - Step 16143: {'lr': 0.0004003274051353884, 'samples': 3099648, 'steps': 16143, 'loss/train': 1.7094197273254395} 01/29/2022 09:35:28 - INFO - codeparrot_training - Step 16144: {'lr': 0.00040031433094293167, 'samples': 3099840, 'steps': 16144, 'loss/train': 2.1770090460777283} 01/29/2022 09:35:32 - INFO - codeparrot_training - Step 16145: {'lr': 0.0004003012561065761, 'samples': 3100032, 'steps': 16145, 'loss/train': 1.6952181458473206} 01/29/2022 09:35:36 - INFO - codeparrot_training - Step 16146: {'lr': 0.0004002881806263776, 'samples': 3100224, 'steps': 16146, 'loss/train': 1.7377334833145142} 01/29/2022 09:35:41 - INFO - codeparrot_training - Step 16147: {'lr': 0.0004002751045023924, 'samples': 3100416, 'steps': 16147, 'loss/train': 2.070653736591339} 01/29/2022 09:35:46 - INFO - codeparrot_training - Step 16148: {'lr': 0.00040026202773467623, 'samples': 3100608, 'steps': 16148, 'loss/train': 1.6832361817359924} 01/29/2022 09:35:50 - INFO - codeparrot_training - Step 16149: {'lr': 0.00040024895032328536, 'samples': 3100800, 'steps': 16149, 'loss/train': 1.5993992686271667} 01/29/2022 09:35:54 - INFO - codeparrot_training - Step 16150: {'lr': 0.0004002358722682756, 'samples': 3100992, 'steps': 16150, 'loss/train': 2.0043137669563293} 01/29/2022 09:35:59 - INFO - codeparrot_training - Step 16151: {'lr': 0.00040022279356970316, 'samples': 3101184, 'steps': 16151, 'loss/train': 0.09481430053710938} 01/29/2022 09:36:03 - INFO - codeparrot_training - Step 16152: {'lr': 0.0004002097142276239, 'samples': 3101376, 'steps': 16152, 'loss/train': 1.5783158540725708} 01/29/2022 09:36:07 - INFO - codeparrot_training - Step 16153: {'lr': 0.00040019663424209397, 'samples': 3101568, 'steps': 16153, 'loss/train': 1.8199287056922913} 01/29/2022 09:36:13 - INFO - codeparrot_training - Step 16154: {'lr': 0.0004001835536131693, 'samples': 3101760, 'steps': 16154, 'loss/train': 2.26240074634552} 01/29/2022 09:36:17 - INFO - codeparrot_training - Step 16155: {'lr': 0.00040017047234090596, 'samples': 3101952, 'steps': 16155, 'loss/train': 1.8857894539833069} 01/29/2022 09:36:21 - INFO - codeparrot_training - Step 16156: {'lr': 0.00040015739042536, 'samples': 3102144, 'steps': 16156, 'loss/train': 1.5845400094985962} 01/29/2022 09:36:26 - INFO - codeparrot_training - Step 16157: {'lr': 0.00040014430786658754, 'samples': 3102336, 'steps': 16157, 'loss/train': 1.7740582823753357} 01/29/2022 09:36:33 - INFO - codeparrot_training - Step 16158: {'lr': 0.0004001312246646446, 'samples': 3102528, 'steps': 16158, 'loss/train': 1.69147789478302} 01/29/2022 09:36:37 - INFO - codeparrot_training - Step 16159: {'lr': 0.000400118140819587, 'samples': 3102720, 'steps': 16159, 'loss/train': 1.866086483001709} 01/29/2022 09:36:41 - INFO - codeparrot_training - Step 16160: {'lr': 0.00040010505633147106, 'samples': 3102912, 'steps': 16160, 'loss/train': 2.3034598231315613} 01/29/2022 09:36:45 - INFO - codeparrot_training - Step 16161: {'lr': 0.0004000919712003526, 'samples': 3103104, 'steps': 16161, 'loss/train': 1.3062476813793182} 01/29/2022 09:36:50 - INFO - codeparrot_training - Step 16162: {'lr': 0.0004000788854262879, 'samples': 3103296, 'steps': 16162, 'loss/train': 1.7223830223083496} 01/29/2022 09:36:56 - INFO - codeparrot_training - Step 16163: {'lr': 0.00040006579900933294, 'samples': 3103488, 'steps': 16163, 'loss/train': 0.7954505681991577} 01/29/2022 09:37:00 - INFO - codeparrot_training - Step 16164: {'lr': 0.00040005271194954367, 'samples': 3103680, 'steps': 16164, 'loss/train': 0.10501837357878685} 01/29/2022 09:37:05 - INFO - codeparrot_training - Step 16165: {'lr': 0.00040003962424697625, 'samples': 3103872, 'steps': 16165, 'loss/train': 1.8490199446678162} 01/29/2022 09:37:09 - INFO - codeparrot_training - Step 16166: {'lr': 0.0004000265359016867, 'samples': 3104064, 'steps': 16166, 'loss/train': 1.84483140707016} 01/29/2022 09:37:13 - INFO - codeparrot_training - Step 16167: {'lr': 0.0004000134469137312, 'samples': 3104256, 'steps': 16167, 'loss/train': 2.5605658292770386} 01/29/2022 09:37:17 - INFO - codeparrot_training - Step 16168: {'lr': 0.00040000035728316564, 'samples': 3104448, 'steps': 16168, 'loss/train': 1.7968518733978271} 01/29/2022 09:37:22 - INFO - codeparrot_training - Step 16169: {'lr': 0.0003999872670100462, 'samples': 3104640, 'steps': 16169, 'loss/train': 0.6761555224657059} 01/29/2022 09:37:27 - INFO - codeparrot_training - Step 16170: {'lr': 0.000399974176094429, 'samples': 3104832, 'steps': 16170, 'loss/train': 1.8659078478813171} 01/29/2022 09:37:31 - INFO - codeparrot_training - Step 16171: {'lr': 0.00039996108453637, 'samples': 3105024, 'steps': 16171, 'loss/train': 1.8292361497879028} 01/29/2022 09:37:36 - INFO - codeparrot_training - Step 16172: {'lr': 0.0003999479923359253, 'samples': 3105216, 'steps': 16172, 'loss/train': 1.46623894572258} 01/29/2022 09:37:40 - INFO - codeparrot_training - Step 16173: {'lr': 0.00039993489949315103, 'samples': 3105408, 'steps': 16173, 'loss/train': 1.0936353206634521} 01/29/2022 09:37:44 - INFO - codeparrot_training - Step 16174: {'lr': 0.0003999218060081032, 'samples': 3105600, 'steps': 16174, 'loss/train': 2.28893905878067} 01/29/2022 09:37:51 - INFO - codeparrot_training - Step 16175: {'lr': 0.0003999087118808381, 'samples': 3105792, 'steps': 16175, 'loss/train': 1.920710027217865} 01/29/2022 09:37:55 - INFO - codeparrot_training - Step 16176: {'lr': 0.0003998956171114116, 'samples': 3105984, 'steps': 16176, 'loss/train': 1.403317004442215} 01/29/2022 09:38:00 - INFO - codeparrot_training - Step 16177: {'lr': 0.0003998825216998799, 'samples': 3106176, 'steps': 16177, 'loss/train': 1.9959474205970764} 01/29/2022 09:38:04 - INFO - codeparrot_training - Step 16178: {'lr': 0.00039986942564629904, 'samples': 3106368, 'steps': 16178, 'loss/train': 1.5769945979118347} 01/29/2022 09:38:08 - INFO - codeparrot_training - Step 16179: {'lr': 0.0003998563289507251, 'samples': 3106560, 'steps': 16179, 'loss/train': 1.8931934237480164} 01/29/2022 09:38:13 - INFO - codeparrot_training - Step 16180: {'lr': 0.0003998432316132143, 'samples': 3106752, 'steps': 16180, 'loss/train': 2.06840443611145} 01/29/2022 09:38:18 - INFO - codeparrot_training - Step 16181: {'lr': 0.0003998301336338227, 'samples': 3106944, 'steps': 16181, 'loss/train': 1.3182663023471832} 01/29/2022 09:38:22 - INFO - codeparrot_training - Step 16182: {'lr': 0.0003998170350126064, 'samples': 3107136, 'steps': 16182, 'loss/train': 1.619469702243805} 01/29/2022 09:38:26 - INFO - codeparrot_training - Step 16183: {'lr': 0.0003998039357496214, 'samples': 3107328, 'steps': 16183, 'loss/train': 1.468217521905899} 01/29/2022 09:38:30 - INFO - codeparrot_training - Step 16184: {'lr': 0.000399790835844924, 'samples': 3107520, 'steps': 16184, 'loss/train': 1.6761909127235413} 01/29/2022 09:38:36 - INFO - codeparrot_training - Step 16185: {'lr': 0.00039977773529857016, 'samples': 3107712, 'steps': 16185, 'loss/train': 2.1891270875930786} 01/29/2022 09:38:40 - INFO - codeparrot_training - Step 16186: {'lr': 0.00039976463411061606, 'samples': 3107904, 'steps': 16186, 'loss/train': 1.6098412871360779} 01/29/2022 09:38:44 - INFO - codeparrot_training - Step 16187: {'lr': 0.00039975153228111784, 'samples': 3108096, 'steps': 16187, 'loss/train': 0.44619230926036835} 01/29/2022 09:38:49 - INFO - codeparrot_training - Step 16188: {'lr': 0.0003997384298101316, 'samples': 3108288, 'steps': 16188, 'loss/train': 1.726869821548462} 01/29/2022 09:38:53 - INFO - codeparrot_training - Step 16189: {'lr': 0.0003997253266977135, 'samples': 3108480, 'steps': 16189, 'loss/train': 1.3324012756347656} 01/29/2022 09:39:00 - INFO - codeparrot_training - Step 16190: {'lr': 0.0003997122229439196, 'samples': 3108672, 'steps': 16190, 'loss/train': 1.6147754788398743} 01/29/2022 09:39:04 - INFO - codeparrot_training - Step 16191: {'lr': 0.00039969911854880613, 'samples': 3108864, 'steps': 16191, 'loss/train': 2.127252995967865} 01/29/2022 09:39:08 - INFO - codeparrot_training - Step 16192: {'lr': 0.0003996860135124292, 'samples': 3109056, 'steps': 16192, 'loss/train': 2.2448254823684692} 01/29/2022 09:39:13 - INFO - codeparrot_training - Step 16193: {'lr': 0.00039967290783484485, 'samples': 3109248, 'steps': 16193, 'loss/train': 1.9249089360237122} 01/29/2022 09:39:18 - INFO - codeparrot_training - Step 16194: {'lr': 0.00039965980151610925, 'samples': 3109440, 'steps': 16194, 'loss/train': 1.457798570394516} 01/29/2022 09:39:22 - INFO - codeparrot_training - Step 16195: {'lr': 0.0003996466945562787, 'samples': 3109632, 'steps': 16195, 'loss/train': 2.4471945762634277} 01/29/2022 09:39:26 - INFO - codeparrot_training - Step 16196: {'lr': 0.00039963358695540907, 'samples': 3109824, 'steps': 16196, 'loss/train': 1.6852995157241821} 01/29/2022 09:39:31 - INFO - codeparrot_training - Step 16197: {'lr': 0.00039962047871355686, 'samples': 3110016, 'steps': 16197, 'loss/train': 1.3776808083057404} 01/29/2022 09:39:35 - INFO - codeparrot_training - Step 16198: {'lr': 0.00039960736983077783, 'samples': 3110208, 'steps': 16198, 'loss/train': 1.3817788660526276} 01/29/2022 09:39:40 - INFO - codeparrot_training - Step 16199: {'lr': 0.0003995942603071285, 'samples': 3110400, 'steps': 16199, 'loss/train': 0.7536833882331848} 01/29/2022 09:39:44 - INFO - codeparrot_training - Step 16200: {'lr': 0.0003995811501426648, 'samples': 3110592, 'steps': 16200, 'loss/train': 2.0316529870033264} 01/29/2022 09:39:49 - INFO - codeparrot_training - Step 16201: {'lr': 0.0003995680393374429, 'samples': 3110784, 'steps': 16201, 'loss/train': 1.330168604850769} 01/29/2022 09:39:53 - INFO - codeparrot_training - Step 16202: {'lr': 0.00039955492789151904, 'samples': 3110976, 'steps': 16202, 'loss/train': 1.6979745626449585} 01/29/2022 09:39:57 - INFO - codeparrot_training - Step 16203: {'lr': 0.0003995418158049494, 'samples': 3111168, 'steps': 16203, 'loss/train': 3.6201045513153076} 01/29/2022 09:40:04 - INFO - codeparrot_training - Step 16204: {'lr': 0.0003995287030777901, 'samples': 3111360, 'steps': 16204, 'loss/train': 1.7923502326011658} 01/29/2022 09:40:09 - INFO - codeparrot_training - Step 16205: {'lr': 0.0003995155897100973, 'samples': 3111552, 'steps': 16205, 'loss/train': 1.0921432971954346} 01/29/2022 09:40:13 - INFO - codeparrot_training - Step 16206: {'lr': 0.0003995024757019272, 'samples': 3111744, 'steps': 16206, 'loss/train': 0.27980484813451767} 01/29/2022 09:40:17 - INFO - codeparrot_training - Step 16207: {'lr': 0.00039948936105333593, 'samples': 3111936, 'steps': 16207, 'loss/train': 1.4878664016723633} 01/29/2022 09:40:21 - INFO - codeparrot_training - Step 16208: {'lr': 0.0003994762457643797, 'samples': 3112128, 'steps': 16208, 'loss/train': 2.7731813192367554} 01/29/2022 09:40:27 - INFO - codeparrot_training - Step 16209: {'lr': 0.0003994631298351148, 'samples': 3112320, 'steps': 16209, 'loss/train': 1.7307171821594238} 01/29/2022 09:40:31 - INFO - codeparrot_training - Step 16210: {'lr': 0.0003994500132655972, 'samples': 3112512, 'steps': 16210, 'loss/train': 1.8521692156791687} 01/29/2022 09:40:35 - INFO - codeparrot_training - Step 16211: {'lr': 0.0003994368960558832, 'samples': 3112704, 'steps': 16211, 'loss/train': 1.8425604701042175} 01/29/2022 09:40:40 - INFO - codeparrot_training - Step 16212: {'lr': 0.0003994237782060291, 'samples': 3112896, 'steps': 16212, 'loss/train': 1.7812215685844421} 01/29/2022 09:40:44 - INFO - codeparrot_training - Step 16213: {'lr': 0.00039941065971609084, 'samples': 3113088, 'steps': 16213, 'loss/train': 1.5674500465393066} 01/29/2022 09:40:51 - INFO - codeparrot_training - Step 16214: {'lr': 0.00039939754058612487, 'samples': 3113280, 'steps': 16214, 'loss/train': 2.5283771753311157} 01/29/2022 09:40:55 - INFO - codeparrot_training - Step 16215: {'lr': 0.0003993844208161872, 'samples': 3113472, 'steps': 16215, 'loss/train': 1.6947988271713257} 01/29/2022 09:40:59 - INFO - codeparrot_training - Step 16216: {'lr': 0.0003993713004063341, 'samples': 3113664, 'steps': 16216, 'loss/train': 2.049388289451599} 01/29/2022 09:41:04 - INFO - codeparrot_training - Step 16217: {'lr': 0.0003993581793566219, 'samples': 3113856, 'steps': 16217, 'loss/train': 1.1307084560394287} 01/29/2022 09:41:08 - INFO - codeparrot_training - Step 16218: {'lr': 0.00039934505766710656, 'samples': 3114048, 'steps': 16218, 'loss/train': 0.31189052760601044} 01/29/2022 09:41:13 - INFO - codeparrot_training - Step 16219: {'lr': 0.0003993319353378445, 'samples': 3114240, 'steps': 16219, 'loss/train': 1.8119770288467407} 01/29/2022 09:41:18 - INFO - codeparrot_training - Step 16220: {'lr': 0.0003993188123688918, 'samples': 3114432, 'steps': 16220, 'loss/train': 1.9173381328582764} 01/29/2022 09:41:22 - INFO - codeparrot_training - Step 16221: {'lr': 0.00039930568876030473, 'samples': 3114624, 'steps': 16221, 'loss/train': 1.952261745929718} 01/29/2022 09:41:26 - INFO - codeparrot_training - Step 16222: {'lr': 0.0003992925645121395, 'samples': 3114816, 'steps': 16222, 'loss/train': 1.4496101438999176} 01/29/2022 09:41:30 - INFO - codeparrot_training - Step 16223: {'lr': 0.00039927943962445234, 'samples': 3115008, 'steps': 16223, 'loss/train': 2.7664130330085754} 01/29/2022 09:41:36 - INFO - codeparrot_training - Step 16224: {'lr': 0.0003992663140972994, 'samples': 3115200, 'steps': 16224, 'loss/train': 1.3043758571147919} 01/29/2022 09:41:40 - INFO - codeparrot_training - Step 16225: {'lr': 0.0003992531879307371, 'samples': 3115392, 'steps': 16225, 'loss/train': 1.9210544228553772} 01/29/2022 09:41:44 - INFO - codeparrot_training - Step 16226: {'lr': 0.0003992400611248214, 'samples': 3115584, 'steps': 16226, 'loss/train': 1.7209686040878296} 01/29/2022 09:41:48 - INFO - codeparrot_training - Step 16227: {'lr': 0.0003992269336796087, 'samples': 3115776, 'steps': 16227, 'loss/train': 1.4247992634773254} 01/29/2022 09:41:53 - INFO - codeparrot_training - Step 16228: {'lr': 0.0003992138055951552, 'samples': 3115968, 'steps': 16228, 'loss/train': 0.99423748254776} 01/29/2022 09:41:58 - INFO - codeparrot_training - Step 16229: {'lr': 0.00039920067687151717, 'samples': 3116160, 'steps': 16229, 'loss/train': 1.9007075428962708} 01/29/2022 09:42:02 - INFO - codeparrot_training - Step 16230: {'lr': 0.0003991875475087508, 'samples': 3116352, 'steps': 16230, 'loss/train': 1.4716215133666992} 01/29/2022 09:42:07 - INFO - codeparrot_training - Step 16231: {'lr': 0.00039917441750691237, 'samples': 3116544, 'steps': 16231, 'loss/train': 1.3694702088832855} 01/29/2022 09:42:11 - INFO - codeparrot_training - Step 16232: {'lr': 0.0003991612868660581, 'samples': 3116736, 'steps': 16232, 'loss/train': 0.2747119143605232} 01/29/2022 09:42:15 - INFO - codeparrot_training - Step 16233: {'lr': 0.0003991481555862442, 'samples': 3116928, 'steps': 16233, 'loss/train': 2.363905370235443} 01/29/2022 09:42:22 - INFO - codeparrot_training - Step 16234: {'lr': 0.00039913502366752704, 'samples': 3117120, 'steps': 16234, 'loss/train': 2.456750750541687} 01/29/2022 09:42:26 - INFO - codeparrot_training - Step 16235: {'lr': 0.0003991218911099627, 'samples': 3117312, 'steps': 16235, 'loss/train': 1.5015411972999573} 01/29/2022 09:42:31 - INFO - codeparrot_training - Step 16236: {'lr': 0.0003991087579136076, 'samples': 3117504, 'steps': 16236, 'loss/train': 1.3899801671504974} 01/29/2022 09:42:35 - INFO - codeparrot_training - Step 16237: {'lr': 0.00039909562407851784, 'samples': 3117696, 'steps': 16237, 'loss/train': 1.7059409022331238} 01/29/2022 09:42:39 - INFO - codeparrot_training - Step 16238: {'lr': 0.0003990824896047498, 'samples': 3117888, 'steps': 16238, 'loss/train': 1.0497426688671112} 01/29/2022 09:42:45 - INFO - codeparrot_training - Step 16239: {'lr': 0.00039906935449235983, 'samples': 3118080, 'steps': 16239, 'loss/train': 2.091832995414734} 01/29/2022 09:42:50 - INFO - codeparrot_training - Step 16240: {'lr': 0.00039905621874140396, 'samples': 3118272, 'steps': 16240, 'loss/train': 1.740374743938446} 01/29/2022 09:42:54 - INFO - codeparrot_training - Step 16241: {'lr': 0.00039904308235193866, 'samples': 3118464, 'steps': 16241, 'loss/train': 1.731377899646759} 01/29/2022 09:42:58 - INFO - codeparrot_training - Step 16242: {'lr': 0.00039902994532402004, 'samples': 3118656, 'steps': 16242, 'loss/train': 7.253089427947998} 01/29/2022 09:43:02 - INFO - codeparrot_training - Step 16243: {'lr': 0.0003990168076577045, 'samples': 3118848, 'steps': 16243, 'loss/train': 1.8813502192497253} 01/29/2022 09:43:07 - INFO - codeparrot_training - Step 16244: {'lr': 0.00039900366935304824, 'samples': 3119040, 'steps': 16244, 'loss/train': 2.3676279187202454} 01/29/2022 09:43:12 - INFO - codeparrot_training - Step 16245: {'lr': 0.00039899053041010765, 'samples': 3119232, 'steps': 16245, 'loss/train': 2.4627934098243713} 01/29/2022 09:43:16 - INFO - codeparrot_training - Step 16246: {'lr': 0.00039897739082893883, 'samples': 3119424, 'steps': 16246, 'loss/train': 1.9652729630470276} 01/29/2022 09:43:20 - INFO - codeparrot_training - Step 16247: {'lr': 0.0003989642506095983, 'samples': 3119616, 'steps': 16247, 'loss/train': 1.301195114850998} 01/29/2022 09:43:25 - INFO - codeparrot_training - Step 16248: {'lr': 0.0003989511097521421, 'samples': 3119808, 'steps': 16248, 'loss/train': 2.088002622127533} 01/29/2022 09:43:29 - INFO - codeparrot_training - Step 16249: {'lr': 0.00039893796825662676, 'samples': 3120000, 'steps': 16249, 'loss/train': 1.8796699047088623} 01/29/2022 09:43:36 - INFO - codeparrot_training - Step 16250: {'lr': 0.0003989248261231084, 'samples': 3120192, 'steps': 16250, 'loss/train': 1.8696736693382263} 01/29/2022 09:43:40 - INFO - codeparrot_training - Step 16251: {'lr': 0.0003989116833516433, 'samples': 3120384, 'steps': 16251, 'loss/train': 1.9957789778709412} 01/29/2022 09:43:45 - INFO - codeparrot_training - Step 16252: {'lr': 0.000398898539942288, 'samples': 3120576, 'steps': 16252, 'loss/train': 1.827740728855133} 01/29/2022 09:43:49 - INFO - codeparrot_training - Step 16253: {'lr': 0.0003988853958950984, 'samples': 3120768, 'steps': 16253, 'loss/train': 1.9081335067749023} 01/29/2022 09:43:53 - INFO - codeparrot_training - Step 16254: {'lr': 0.00039887225121013124, 'samples': 3120960, 'steps': 16254, 'loss/train': 1.312264233827591} 01/29/2022 09:43:59 - INFO - codeparrot_training - Step 16255: {'lr': 0.0003988591058874426, 'samples': 3121152, 'steps': 16255, 'loss/train': 1.0677239298820496} 01/29/2022 09:44:03 - INFO - codeparrot_training - Step 16256: {'lr': 0.00039884595992708877, 'samples': 3121344, 'steps': 16256, 'loss/train': 1.869579792022705} 01/29/2022 09:44:07 - INFO - codeparrot_training - Step 16257: {'lr': 0.0003988328133291261, 'samples': 3121536, 'steps': 16257, 'loss/train': 1.7883403301239014} 01/29/2022 09:44:11 - INFO - codeparrot_training - Step 16258: {'lr': 0.000398819666093611, 'samples': 3121728, 'steps': 16258, 'loss/train': 1.6690120697021484} 01/29/2022 09:44:16 - INFO - codeparrot_training - Step 16259: {'lr': 0.0003988065182205996, 'samples': 3121920, 'steps': 16259, 'loss/train': 1.5701948404312134} 01/29/2022 09:44:23 - INFO - codeparrot_training - Step 16260: {'lr': 0.0003987933697101484, 'samples': 3122112, 'steps': 16260, 'loss/train': 0.9566218256950378} 01/29/2022 09:44:27 - INFO - codeparrot_training - Step 16261: {'lr': 0.0003987802205623136, 'samples': 3122304, 'steps': 16261, 'loss/train': 3.448452115058899} 01/29/2022 09:44:31 - INFO - codeparrot_training - Step 16262: {'lr': 0.0003987670707771516, 'samples': 3122496, 'steps': 16262, 'loss/train': 2.4280993938446045} 01/29/2022 09:44:36 - INFO - codeparrot_training - Step 16263: {'lr': 0.0003987539203547187, 'samples': 3122688, 'steps': 16263, 'loss/train': 1.6855709552764893} 01/29/2022 09:44:40 - INFO - codeparrot_training - Step 16264: {'lr': 0.00039874076929507124, 'samples': 3122880, 'steps': 16264, 'loss/train': 1.9477724432945251} 01/29/2022 09:44:45 - INFO - codeparrot_training - Step 16265: {'lr': 0.0003987276175982656, 'samples': 3123072, 'steps': 16265, 'loss/train': 1.7188307046890259} 01/29/2022 09:44:50 - INFO - codeparrot_training - Step 16266: {'lr': 0.00039871446526435806, 'samples': 3123264, 'steps': 16266, 'loss/train': 2.1779648065567017} 01/29/2022 09:44:54 - INFO - codeparrot_training - Step 16267: {'lr': 0.00039870131229340495, 'samples': 3123456, 'steps': 16267, 'loss/train': 1.7002699971199036} 01/29/2022 09:44:58 - INFO - codeparrot_training - Step 16268: {'lr': 0.00039868815868546257, 'samples': 3123648, 'steps': 16268, 'loss/train': 1.862796664237976} 01/29/2022 09:45:02 - INFO - codeparrot_training - Step 16269: {'lr': 0.00039867500444058747, 'samples': 3123840, 'steps': 16269, 'loss/train': 1.3558046221733093} 01/29/2022 09:45:07 - INFO - codeparrot_training - Step 16270: {'lr': 0.0003986618495588358, 'samples': 3124032, 'steps': 16270, 'loss/train': 2.1729401350021362} 01/29/2022 09:45:12 - INFO - codeparrot_training - Step 16271: {'lr': 0.00039864869404026394, 'samples': 3124224, 'steps': 16271, 'loss/train': 1.0924484431743622} 01/29/2022 09:45:16 - INFO - codeparrot_training - Step 16272: {'lr': 0.0003986355378849283, 'samples': 3124416, 'steps': 16272, 'loss/train': 1.9176453351974487} 01/29/2022 09:45:20 - INFO - codeparrot_training - Step 16273: {'lr': 0.00039862238109288523, 'samples': 3124608, 'steps': 16273, 'loss/train': 1.252457320690155} 01/29/2022 09:45:25 - INFO - codeparrot_training - Step 16274: {'lr': 0.0003986092236641911, 'samples': 3124800, 'steps': 16274, 'loss/train': 2.872820198535919} 01/29/2022 09:45:29 - INFO - codeparrot_training - Step 16275: {'lr': 0.00039859606559890215, 'samples': 3124992, 'steps': 16275, 'loss/train': 1.8557106256484985} 01/29/2022 09:45:36 - INFO - codeparrot_training - Step 16276: {'lr': 0.0003985829068970749, 'samples': 3125184, 'steps': 16276, 'loss/train': 2.345091998577118} 01/29/2022 09:45:40 - INFO - codeparrot_training - Step 16277: {'lr': 0.00039856974755876563, 'samples': 3125376, 'steps': 16277, 'loss/train': 1.4312195777893066} 01/29/2022 09:45:45 - INFO - codeparrot_training - Step 16278: {'lr': 0.0003985565875840308, 'samples': 3125568, 'steps': 16278, 'loss/train': 2.200916290283203} 01/29/2022 09:45:49 - INFO - codeparrot_training - Step 16279: {'lr': 0.0003985434269729267, 'samples': 3125760, 'steps': 16279, 'loss/train': 1.1695991456508636} 01/29/2022 09:45:54 - INFO - codeparrot_training - Step 16280: {'lr': 0.00039853026572550965, 'samples': 3125952, 'steps': 16280, 'loss/train': 1.158160775899887} 01/29/2022 09:45:58 - INFO - codeparrot_training - Step 16281: {'lr': 0.00039851710384183615, 'samples': 3126144, 'steps': 16281, 'loss/train': 1.8997589349746704} 01/29/2022 09:46:03 - INFO - codeparrot_training - Step 16282: {'lr': 0.0003985039413219626, 'samples': 3126336, 'steps': 16282, 'loss/train': 1.7653360962867737} 01/29/2022 09:46:07 - INFO - codeparrot_training - Step 16283: {'lr': 0.0003984907781659452, 'samples': 3126528, 'steps': 16283, 'loss/train': 0.5914157330989838} 01/29/2022 09:46:11 - INFO - codeparrot_training - Step 16284: {'lr': 0.00039847761437384054, 'samples': 3126720, 'steps': 16284, 'loss/train': 0.8723741769790649} 01/29/2022 09:46:16 - INFO - codeparrot_training - Step 16285: {'lr': 0.0003984644499457049, 'samples': 3126912, 'steps': 16285, 'loss/train': 1.8800505995750427} 01/29/2022 09:46:21 - INFO - codeparrot_training - Step 16286: {'lr': 0.0003984512848815948, 'samples': 3127104, 'steps': 16286, 'loss/train': 2.4759585857391357} 01/29/2022 09:46:25 - INFO - codeparrot_training - Step 16287: {'lr': 0.00039843811918156635, 'samples': 3127296, 'steps': 16287, 'loss/train': 0.640467569231987} 01/29/2022 09:46:29 - INFO - codeparrot_training - Step 16288: {'lr': 0.0003984249528456762, 'samples': 3127488, 'steps': 16288, 'loss/train': 1.4254584610462189} 01/29/2022 09:46:33 - INFO - codeparrot_training - Step 16289: {'lr': 0.00039841178587398074, 'samples': 3127680, 'steps': 16289, 'loss/train': 1.5701680183410645} 01/29/2022 09:46:40 - INFO - codeparrot_training - Step 16290: {'lr': 0.0003983986182665362, 'samples': 3127872, 'steps': 16290, 'loss/train': 1.2650945484638214} 01/29/2022 09:46:45 - INFO - codeparrot_training - Step 16291: {'lr': 0.00039838545002339926, 'samples': 3128064, 'steps': 16291, 'loss/train': 1.5039857625961304} 01/29/2022 09:46:49 - INFO - codeparrot_training - Step 16292: {'lr': 0.0003983722811446261, 'samples': 3128256, 'steps': 16292, 'loss/train': 1.1525125801563263} 01/29/2022 09:46:53 - INFO - codeparrot_training - Step 16293: {'lr': 0.00039835911163027315, 'samples': 3128448, 'steps': 16293, 'loss/train': 0.7531093060970306} 01/29/2022 09:46:57 - INFO - codeparrot_training - Step 16294: {'lr': 0.00039834594148039693, 'samples': 3128640, 'steps': 16294, 'loss/train': 1.9175657629966736} 01/29/2022 09:47:03 - INFO - codeparrot_training - Step 16295: {'lr': 0.0003983327706950538, 'samples': 3128832, 'steps': 16295, 'loss/train': 1.2597221732139587} 01/29/2022 09:47:07 - INFO - codeparrot_training - Step 16296: {'lr': 0.00039831959927430017, 'samples': 3129024, 'steps': 16296, 'loss/train': 1.7902595400810242} 01/29/2022 09:47:11 - INFO - codeparrot_training - Step 16297: {'lr': 0.00039830642721819254, 'samples': 3129216, 'steps': 16297, 'loss/train': 1.1990575790405273} 01/29/2022 09:47:15 - INFO - codeparrot_training - Step 16298: {'lr': 0.0003982932545267872, 'samples': 3129408, 'steps': 16298, 'loss/train': 1.8972011804580688} 01/29/2022 09:47:20 - INFO - codeparrot_training - Step 16299: {'lr': 0.00039828008120014057, 'samples': 3129600, 'steps': 16299, 'loss/train': 1.8557916283607483} 01/29/2022 09:47:25 - INFO - codeparrot_training - Step 16300: {'lr': 0.00039826690723830926, 'samples': 3129792, 'steps': 16300, 'loss/train': 1.5236093401908875} 01/29/2022 09:47:29 - INFO - codeparrot_training - Step 16301: {'lr': 0.00039825373264134955, 'samples': 3129984, 'steps': 16301, 'loss/train': 1.7551685571670532} 01/29/2022 09:47:33 - INFO - codeparrot_training - Step 16302: {'lr': 0.00039824055740931804, 'samples': 3130176, 'steps': 16302, 'loss/train': 1.1893487870693207} 01/29/2022 09:47:38 - INFO - codeparrot_training - Step 16303: {'lr': 0.0003982273815422709, 'samples': 3130368, 'steps': 16303, 'loss/train': 1.9443642497062683} 01/29/2022 09:47:42 - INFO - codeparrot_training - Step 16304: {'lr': 0.00039821420504026486, 'samples': 3130560, 'steps': 16304, 'loss/train': 0.9580758512020111} 01/29/2022 09:47:49 - INFO - codeparrot_training - Step 16305: {'lr': 0.0003982010279033561, 'samples': 3130752, 'steps': 16305, 'loss/train': 1.5156505107879639} 01/29/2022 09:47:53 - INFO - codeparrot_training - Step 16306: {'lr': 0.0003981878501316013, 'samples': 3130944, 'steps': 16306, 'loss/train': 1.347068041563034} 01/29/2022 09:47:58 - INFO - codeparrot_training - Step 16307: {'lr': 0.0003981746717250567, 'samples': 3131136, 'steps': 16307, 'loss/train': 1.684946358203888} 01/29/2022 09:48:02 - INFO - codeparrot_training - Step 16308: {'lr': 0.000398161492683779, 'samples': 3131328, 'steps': 16308, 'loss/train': 2.133822977542877} 01/29/2022 09:48:06 - INFO - codeparrot_training - Step 16309: {'lr': 0.0003981483130078244, 'samples': 3131520, 'steps': 16309, 'loss/train': 2.0452761054039} 01/29/2022 09:48:12 - INFO - codeparrot_training - Step 16310: {'lr': 0.0003981351326972495, 'samples': 3131712, 'steps': 16310, 'loss/train': 1.6492858529090881} 01/29/2022 09:48:16 - INFO - codeparrot_training - Step 16311: {'lr': 0.00039812195175211075, 'samples': 3131904, 'steps': 16311, 'loss/train': 1.1057851016521454} 01/29/2022 09:48:20 - INFO - codeparrot_training - Step 16312: {'lr': 0.0003981087701724645, 'samples': 3132096, 'steps': 16312, 'loss/train': 1.3972765803337097} 01/29/2022 09:48:24 - INFO - codeparrot_training - Step 16313: {'lr': 0.00039809558795836743, 'samples': 3132288, 'steps': 16313, 'loss/train': 0.5865895003080368} 01/29/2022 09:48:30 - INFO - codeparrot_training - Step 16314: {'lr': 0.00039808240510987584, 'samples': 3132480, 'steps': 16314, 'loss/train': 0.7674595713615417} 01/29/2022 09:48:34 - INFO - codeparrot_training - Step 16315: {'lr': 0.0003980692216270462, 'samples': 3132672, 'steps': 16315, 'loss/train': 2.014768660068512} 01/29/2022 09:48:38 - INFO - codeparrot_training - Step 16316: {'lr': 0.00039805603750993514, 'samples': 3132864, 'steps': 16316, 'loss/train': 0.9983388483524323} 01/29/2022 09:48:43 - INFO - codeparrot_training - Step 16317: {'lr': 0.0003980428527585989, 'samples': 3133056, 'steps': 16317, 'loss/train': 2.106541335582733} 01/29/2022 09:48:47 - INFO - codeparrot_training - Step 16318: {'lr': 0.0003980296673730942, 'samples': 3133248, 'steps': 16318, 'loss/train': 1.4748781621456146} 01/29/2022 09:48:54 - INFO - codeparrot_training - Step 16319: {'lr': 0.0003980164813534773, 'samples': 3133440, 'steps': 16319, 'loss/train': 1.2324441075325012} 01/29/2022 09:48:58 - INFO - codeparrot_training - Step 16320: {'lr': 0.0003980032946998049, 'samples': 3133632, 'steps': 16320, 'loss/train': 0.9695978164672852} 01/29/2022 09:49:03 - INFO - codeparrot_training - Step 16321: {'lr': 0.00039799010741213336, 'samples': 3133824, 'steps': 16321, 'loss/train': 1.6770880222320557} 01/29/2022 09:49:07 - INFO - codeparrot_training - Step 16322: {'lr': 0.0003979769194905192, 'samples': 3134016, 'steps': 16322, 'loss/train': 2.369347929954529} 01/29/2022 09:49:11 - INFO - codeparrot_training - Step 16323: {'lr': 0.0003979637309350188, 'samples': 3134208, 'steps': 16323, 'loss/train': 0.8970301151275635} 01/29/2022 09:49:17 - INFO - codeparrot_training - Step 16324: {'lr': 0.0003979505417456889, 'samples': 3134400, 'steps': 16324, 'loss/train': 1.761531114578247} 01/29/2022 09:49:21 - INFO - codeparrot_training - Step 16325: {'lr': 0.00039793735192258575, 'samples': 3134592, 'steps': 16325, 'loss/train': 1.725495457649231} 01/29/2022 09:49:25 - INFO - codeparrot_training - Step 16326: {'lr': 0.000397924161465766, 'samples': 3134784, 'steps': 16326, 'loss/train': 1.8646613359451294} 01/29/2022 09:49:29 - INFO - codeparrot_training - Step 16327: {'lr': 0.0003979109703752861, 'samples': 3134976, 'steps': 16327, 'loss/train': 1.217332273721695} 01/29/2022 09:49:34 - INFO - codeparrot_training - Step 16328: {'lr': 0.00039789777865120257, 'samples': 3135168, 'steps': 16328, 'loss/train': 1.5781821012496948} 01/29/2022 09:49:39 - INFO - codeparrot_training - Step 16329: {'lr': 0.00039788458629357195, 'samples': 3135360, 'steps': 16329, 'loss/train': 1.3786216378211975} 01/29/2022 09:49:43 - INFO - codeparrot_training - Step 16330: {'lr': 0.0003978713933024507, 'samples': 3135552, 'steps': 16330, 'loss/train': 1.5119547843933105} 01/29/2022 09:49:47 - INFO - codeparrot_training - Step 16331: {'lr': 0.0003978581996778954, 'samples': 3135744, 'steps': 16331, 'loss/train': 0.933406412601471} 01/29/2022 09:49:52 - INFO - codeparrot_training - Step 16332: {'lr': 0.0003978450054199625, 'samples': 3135936, 'steps': 16332, 'loss/train': 1.038929432630539} 01/29/2022 09:49:56 - INFO - codeparrot_training - Step 16333: {'lr': 0.0003978318105287085, 'samples': 3136128, 'steps': 16333, 'loss/train': 2.116549015045166} 01/29/2022 09:50:03 - INFO - codeparrot_training - Step 16334: {'lr': 0.00039781861500419, 'samples': 3136320, 'steps': 16334, 'loss/train': 1.6934954524040222} 01/29/2022 09:50:07 - INFO - codeparrot_training - Step 16335: {'lr': 0.00039780541884646347, 'samples': 3136512, 'steps': 16335, 'loss/train': 2.0979875922203064} 01/29/2022 09:50:12 - INFO - codeparrot_training - Step 16336: {'lr': 0.0003977922220555855, 'samples': 3136704, 'steps': 16336, 'loss/train': 1.4073685705661774} 01/29/2022 09:50:16 - INFO - codeparrot_training - Step 16337: {'lr': 0.0003977790246316125, 'samples': 3136896, 'steps': 16337, 'loss/train': 2.00515615940094} 01/29/2022 09:50:20 - INFO - codeparrot_training - Step 16338: {'lr': 0.00039776582657460115, 'samples': 3137088, 'steps': 16338, 'loss/train': 2.2368680834770203} 01/29/2022 09:50:25 - INFO - codeparrot_training - Step 16339: {'lr': 0.000397752627884608, 'samples': 3137280, 'steps': 16339, 'loss/train': 2.0939692854881287} 01/29/2022 09:50:30 - INFO - codeparrot_training - Step 16340: {'lr': 0.0003977394285616893, 'samples': 3137472, 'steps': 16340, 'loss/train': 1.4812975823879242} 01/29/2022 09:50:34 - INFO - codeparrot_training - Step 16341: {'lr': 0.000397726228605902, 'samples': 3137664, 'steps': 16341, 'loss/train': 1.6938008666038513} 01/29/2022 09:50:38 - INFO - codeparrot_training - Step 16342: {'lr': 0.00039771302801730235, 'samples': 3137856, 'steps': 16342, 'loss/train': 1.9590970873832703} 01/29/2022 09:50:42 - INFO - codeparrot_training - Step 16343: {'lr': 0.00039769982679594703, 'samples': 3138048, 'steps': 16343, 'loss/train': 1.121490865945816} 01/29/2022 09:50:49 - INFO - codeparrot_training - Step 16344: {'lr': 0.0003976866249418925, 'samples': 3138240, 'steps': 16344, 'loss/train': 1.6049850583076477} 01/29/2022 09:50:54 - INFO - codeparrot_training - Step 16345: {'lr': 0.0003976734224551954, 'samples': 3138432, 'steps': 16345, 'loss/train': 3.0333473682403564} 01/29/2022 09:50:58 - INFO - codeparrot_training - Step 16346: {'lr': 0.0003976602193359122, 'samples': 3138624, 'steps': 16346, 'loss/train': 1.2441650927066803} 01/29/2022 09:51:02 - INFO - codeparrot_training - Step 16347: {'lr': 0.00039764701558409955, 'samples': 3138816, 'steps': 16347, 'loss/train': 0.927304744720459} 01/29/2022 09:51:06 - INFO - codeparrot_training - Step 16348: {'lr': 0.000397633811199814, 'samples': 3139008, 'steps': 16348, 'loss/train': 1.7366023063659668} 01/29/2022 09:51:12 - INFO - codeparrot_training - Step 16349: {'lr': 0.000397620606183112, 'samples': 3139200, 'steps': 16349, 'loss/train': 2.060905694961548} 01/29/2022 09:51:16 - INFO - codeparrot_training - Step 16350: {'lr': 0.00039760740053405033, 'samples': 3139392, 'steps': 16350, 'loss/train': 1.7449838519096375} 01/29/2022 09:51:20 - INFO - codeparrot_training - Step 16351: {'lr': 0.00039759419425268526, 'samples': 3139584, 'steps': 16351, 'loss/train': 2.218151092529297} 01/29/2022 09:51:25 - INFO - codeparrot_training - Step 16352: {'lr': 0.00039758098733907364, 'samples': 3139776, 'steps': 16352, 'loss/train': 1.7253767251968384} 01/29/2022 09:51:29 - INFO - codeparrot_training - Step 16353: {'lr': 0.00039756777979327193, 'samples': 3139968, 'steps': 16353, 'loss/train': 1.7297862768173218} 01/29/2022 09:51:34 - INFO - codeparrot_training - Step 16354: {'lr': 0.0003975545716153367, 'samples': 3140160, 'steps': 16354, 'loss/train': 2.1718329191207886} 01/29/2022 09:51:39 - INFO - codeparrot_training - Step 16355: {'lr': 0.0003975413628053245, 'samples': 3140352, 'steps': 16355, 'loss/train': 1.8037458658218384} 01/29/2022 09:51:43 - INFO - codeparrot_training - Step 16356: {'lr': 0.000397528153363292, 'samples': 3140544, 'steps': 16356, 'loss/train': 0.7934592962265015} 01/29/2022 09:51:47 - INFO - codeparrot_training - Step 16357: {'lr': 0.00039751494328929565, 'samples': 3140736, 'steps': 16357, 'loss/train': 0.7551371455192566} 01/29/2022 09:51:51 - INFO - codeparrot_training - Step 16358: {'lr': 0.00039750173258339225, 'samples': 3140928, 'steps': 16358, 'loss/train': 1.3484479486942291} 01/29/2022 09:51:57 - INFO - codeparrot_training - Step 16359: {'lr': 0.00039748852124563816, 'samples': 3141120, 'steps': 16359, 'loss/train': 2.2374009490013123} 01/29/2022 09:52:01 - INFO - codeparrot_training - Step 16360: {'lr': 0.0003974753092760901, 'samples': 3141312, 'steps': 16360, 'loss/train': 2.2668678760528564} 01/29/2022 09:52:05 - INFO - codeparrot_training - Step 16361: {'lr': 0.00039746209667480473, 'samples': 3141504, 'steps': 16361, 'loss/train': 1.5869886875152588} 01/29/2022 09:52:09 - INFO - codeparrot_training - Step 16362: {'lr': 0.00039744888344183846, 'samples': 3141696, 'steps': 16362, 'loss/train': 2.0645615458488464} 01/29/2022 09:52:14 - INFO - codeparrot_training - Step 16363: {'lr': 0.00039743566957724805, 'samples': 3141888, 'steps': 16363, 'loss/train': 0.9442607760429382} 01/29/2022 09:52:21 - INFO - codeparrot_training - Step 16364: {'lr': 0.00039742245508109, 'samples': 3142080, 'steps': 16364, 'loss/train': 0.9744172990322113} 01/29/2022 09:52:25 - INFO - codeparrot_training - Step 16365: {'lr': 0.000397409239953421, 'samples': 3142272, 'steps': 16365, 'loss/train': 2.4782609939575195} 01/29/2022 09:52:29 - INFO - codeparrot_training - Step 16366: {'lr': 0.00039739602419429755, 'samples': 3142464, 'steps': 16366, 'loss/train': 0.2035776749253273} 01/29/2022 09:52:34 - INFO - codeparrot_training - Step 16367: {'lr': 0.00039738280780377645, 'samples': 3142656, 'steps': 16367, 'loss/train': 1.3602831065654755} 01/29/2022 09:52:38 - INFO - codeparrot_training - Step 16368: {'lr': 0.0003973695907819141, 'samples': 3142848, 'steps': 16368, 'loss/train': 1.3473874926567078} 01/29/2022 09:52:43 - INFO - codeparrot_training - Step 16369: {'lr': 0.0003973563731287673, 'samples': 3143040, 'steps': 16369, 'loss/train': 2.00489741563797} 01/29/2022 09:52:48 - INFO - codeparrot_training - Step 16370: {'lr': 0.00039734315484439255, 'samples': 3143232, 'steps': 16370, 'loss/train': 1.5064133405685425} 01/29/2022 09:52:52 - INFO - codeparrot_training - Step 16371: {'lr': 0.0003973299359288465, 'samples': 3143424, 'steps': 16371, 'loss/train': 1.5436581373214722} 01/29/2022 09:52:56 - INFO - codeparrot_training - Step 16372: {'lr': 0.0003973167163821858, 'samples': 3143616, 'steps': 16372, 'loss/train': 1.9766154885292053} 01/29/2022 09:53:00 - INFO - codeparrot_training - Step 16373: {'lr': 0.0003973034962044671, 'samples': 3143808, 'steps': 16373, 'loss/train': 0.8840398192405701} 01/29/2022 09:53:06 - INFO - codeparrot_training - Step 16374: {'lr': 0.00039729027539574696, 'samples': 3144000, 'steps': 16374, 'loss/train': 1.2827720940113068} 01/29/2022 09:53:10 - INFO - codeparrot_training - Step 16375: {'lr': 0.00039727705395608203, 'samples': 3144192, 'steps': 16375, 'loss/train': 1.3793991208076477} 01/29/2022 09:53:15 - INFO - codeparrot_training - Step 16376: {'lr': 0.00039726383188552907, 'samples': 3144384, 'steps': 16376, 'loss/train': 1.5535885691642761} 01/29/2022 09:53:19 - INFO - codeparrot_training - Step 16377: {'lr': 0.00039725060918414446, 'samples': 3144576, 'steps': 16377, 'loss/train': 1.8770653009414673} 01/29/2022 09:53:23 - INFO - codeparrot_training - Step 16378: {'lr': 0.0003972373858519851, 'samples': 3144768, 'steps': 16378, 'loss/train': 1.7982718348503113} 01/29/2022 09:53:27 - INFO - codeparrot_training - Step 16379: {'lr': 0.00039722416188910754, 'samples': 3144960, 'steps': 16379, 'loss/train': 1.3183542788028717} 01/29/2022 09:53:34 - INFO - codeparrot_training - Step 16380: {'lr': 0.00039721093729556836, 'samples': 3145152, 'steps': 16380, 'loss/train': 1.7124980092048645} 01/29/2022 09:53:39 - INFO - codeparrot_training - Step 16381: {'lr': 0.0003971977120714243, 'samples': 3145344, 'steps': 16381, 'loss/train': 0.3951737880706787} 01/29/2022 09:53:43 - INFO - codeparrot_training - Step 16382: {'lr': 0.000397184486216732, 'samples': 3145536, 'steps': 16382, 'loss/train': 2.0021581649780273} 01/29/2022 09:53:47 - INFO - codeparrot_training - Step 16383: {'lr': 0.0003971712597315481, 'samples': 3145728, 'steps': 16383, 'loss/train': 1.3746788799762726} 01/29/2022 09:53:53 - INFO - codeparrot_training - Step 16384: {'lr': 0.0003971580326159292, 'samples': 3145920, 'steps': 16384, 'loss/train': 1.9047197699546814} 01/29/2022 09:53:57 - INFO - codeparrot_training - Step 16385: {'lr': 0.0003971448048699321, 'samples': 3146112, 'steps': 16385, 'loss/train': 2.2658790349960327} 01/29/2022 09:54:01 - INFO - codeparrot_training - Step 16386: {'lr': 0.00039713157649361327, 'samples': 3146304, 'steps': 16386, 'loss/train': 2.3800007700920105} 01/29/2022 09:54:05 - INFO - codeparrot_training - Step 16387: {'lr': 0.00039711834748702956, 'samples': 3146496, 'steps': 16387, 'loss/train': 1.779450237751007} 01/29/2022 09:54:10 - INFO - codeparrot_training - Step 16388: {'lr': 0.0003971051178502375, 'samples': 3146688, 'steps': 16388, 'loss/train': 1.4233751893043518} 01/29/2022 09:54:17 - INFO - codeparrot_training - Step 16389: {'lr': 0.00039709188758329394, 'samples': 3146880, 'steps': 16389, 'loss/train': 2.174984335899353} 01/29/2022 09:54:21 - INFO - codeparrot_training - Step 16390: {'lr': 0.0003970786566862553, 'samples': 3147072, 'steps': 16390, 'loss/train': 0.37751179933547974} 01/29/2022 09:54:25 - INFO - codeparrot_training - Step 16391: {'lr': 0.00039706542515917853, 'samples': 3147264, 'steps': 16391, 'loss/train': 1.7714864015579224} 01/29/2022 09:54:29 - INFO - codeparrot_training - Step 16392: {'lr': 0.00039705219300212015, 'samples': 3147456, 'steps': 16392, 'loss/train': 2.200640559196472} 01/29/2022 09:54:34 - INFO - codeparrot_training - Step 16393: {'lr': 0.00039703896021513684, 'samples': 3147648, 'steps': 16393, 'loss/train': 2.050222635269165} 01/29/2022 09:54:39 - INFO - codeparrot_training - Step 16394: {'lr': 0.0003970257267982853, 'samples': 3147840, 'steps': 16394, 'loss/train': 1.029740810394287} 01/29/2022 09:54:43 - INFO - codeparrot_training - Step 16395: {'lr': 0.0003970124927516222, 'samples': 3148032, 'steps': 16395, 'loss/train': 1.5577141642570496} 01/29/2022 09:54:48 - INFO - codeparrot_training - Step 16396: {'lr': 0.0003969992580752043, 'samples': 3148224, 'steps': 16396, 'loss/train': 1.7650485634803772} 01/29/2022 09:54:52 - INFO - codeparrot_training - Step 16397: {'lr': 0.00039698602276908826, 'samples': 3148416, 'steps': 16397, 'loss/train': 1.551403284072876} 01/29/2022 09:54:56 - INFO - codeparrot_training - Step 16398: {'lr': 0.0003969727868333308, 'samples': 3148608, 'steps': 16398, 'loss/train': 1.3216236233711243} 01/29/2022 09:55:01 - INFO - codeparrot_training - Step 16399: {'lr': 0.00039695955026798857, 'samples': 3148800, 'steps': 16399, 'loss/train': 1.3609507083892822} 01/29/2022 09:55:06 - INFO - codeparrot_training - Step 16400: {'lr': 0.0003969463130731183, 'samples': 3148992, 'steps': 16400, 'loss/train': 1.7786228656768799} 01/29/2022 09:55:10 - INFO - codeparrot_training - Step 16401: {'lr': 0.00039693307524877664, 'samples': 3149184, 'steps': 16401, 'loss/train': 1.3740902245044708} 01/29/2022 09:55:14 - INFO - codeparrot_training - Step 16402: {'lr': 0.0003969198367950204, 'samples': 3149376, 'steps': 16402, 'loss/train': 2.089896082878113} 01/29/2022 09:55:18 - INFO - codeparrot_training - Step 16403: {'lr': 0.00039690659771190616, 'samples': 3149568, 'steps': 16403, 'loss/train': 1.2402033805847168} 01/29/2022 09:55:26 - INFO - codeparrot_training - Step 16404: {'lr': 0.0003968933579994908, 'samples': 3149760, 'steps': 16404, 'loss/train': 2.1777491569519043} 01/29/2022 09:55:30 - INFO - codeparrot_training - Step 16405: {'lr': 0.0003968801176578309, 'samples': 3149952, 'steps': 16405, 'loss/train': 1.5437244772911072} 01/29/2022 09:55:34 - INFO - codeparrot_training - Step 16406: {'lr': 0.00039686687668698316, 'samples': 3150144, 'steps': 16406, 'loss/train': 1.7856098413467407} 01/29/2022 09:55:38 - INFO - codeparrot_training - Step 16407: {'lr': 0.00039685363508700443, 'samples': 3150336, 'steps': 16407, 'loss/train': 1.4051093459129333} 01/29/2022 09:55:43 - INFO - codeparrot_training - Step 16408: {'lr': 0.00039684039285795133, 'samples': 3150528, 'steps': 16408, 'loss/train': 1.751251459121704} 01/29/2022 09:55:48 - INFO - codeparrot_training - Step 16409: {'lr': 0.0003968271499998806, 'samples': 3150720, 'steps': 16409, 'loss/train': 2.0423524975776672} 01/29/2022 09:55:52 - INFO - codeparrot_training - Step 16410: {'lr': 0.000396813906512849, 'samples': 3150912, 'steps': 16410, 'loss/train': 1.5255727171897888} 01/29/2022 09:55:57 - INFO - codeparrot_training - Step 16411: {'lr': 0.00039680066239691325, 'samples': 3151104, 'steps': 16411, 'loss/train': 1.9380286931991577} 01/29/2022 09:56:01 - INFO - codeparrot_training - Step 16412: {'lr': 0.00039678741765213006, 'samples': 3151296, 'steps': 16412, 'loss/train': 1.8211802244186401} 01/29/2022 09:56:05 - INFO - codeparrot_training - Step 16413: {'lr': 0.00039677417227855624, 'samples': 3151488, 'steps': 16413, 'loss/train': 2.0522437691688538} 01/29/2022 09:56:10 - INFO - codeparrot_training - Step 16414: {'lr': 0.0003967609262762484, 'samples': 3151680, 'steps': 16414, 'loss/train': 2.047342836856842} 01/29/2022 09:56:14 - INFO - codeparrot_training - Step 16415: {'lr': 0.0003967476796452634, 'samples': 3151872, 'steps': 16415, 'loss/train': 1.7888309955596924} 01/29/2022 09:56:19 - INFO - codeparrot_training - Step 16416: {'lr': 0.00039673443238565786, 'samples': 3152064, 'steps': 16416, 'loss/train': 1.5976662039756775} 01/29/2022 09:56:23 - INFO - codeparrot_training - Step 16417: {'lr': 0.0003967211844974887, 'samples': 3152256, 'steps': 16417, 'loss/train': 1.5732141137123108} 01/29/2022 09:56:27 - INFO - codeparrot_training - Step 16418: {'lr': 0.0003967079359808125, 'samples': 3152448, 'steps': 16418, 'loss/train': 1.7355190515518188} 01/29/2022 09:56:32 - INFO - codeparrot_training - Step 16419: {'lr': 0.0003966946868356861, 'samples': 3152640, 'steps': 16419, 'loss/train': 1.9788272380828857} 01/29/2022 09:56:37 - INFO - codeparrot_training - Step 16420: {'lr': 0.0003966814370621663, 'samples': 3152832, 'steps': 16420, 'loss/train': 1.758451223373413} 01/29/2022 09:56:41 - INFO - codeparrot_training - Step 16421: {'lr': 0.00039666818666030974, 'samples': 3153024, 'steps': 16421, 'loss/train': 1.3756913244724274} 01/29/2022 09:56:45 - INFO - codeparrot_training - Step 16422: {'lr': 0.0003966549356301733, 'samples': 3153216, 'steps': 16422, 'loss/train': 1.3728422820568085} 01/29/2022 09:56:49 - INFO - codeparrot_training - Step 16423: {'lr': 0.0003966416839718136, 'samples': 3153408, 'steps': 16423, 'loss/train': 1.5697186589241028} 01/29/2022 09:56:57 - INFO - codeparrot_training - Step 16424: {'lr': 0.00039662843168528756, 'samples': 3153600, 'steps': 16424, 'loss/train': 1.9612671732902527} 01/29/2022 09:57:01 - INFO - codeparrot_training - Step 16425: {'lr': 0.00039661517877065183, 'samples': 3153792, 'steps': 16425, 'loss/train': 1.7815638184547424} 01/29/2022 09:57:05 - INFO - codeparrot_training - Step 16426: {'lr': 0.0003966019252279633, 'samples': 3153984, 'steps': 16426, 'loss/train': 1.3490484952926636} 01/29/2022 09:57:09 - INFO - codeparrot_training - Step 16427: {'lr': 0.00039658867105727856, 'samples': 3154176, 'steps': 16427, 'loss/train': 2.122195065021515} 01/29/2022 09:57:13 - INFO - codeparrot_training - Step 16428: {'lr': 0.0003965754162586547, 'samples': 3154368, 'steps': 16428, 'loss/train': 1.7959937453269958} 01/29/2022 09:57:19 - INFO - codeparrot_training - Step 16429: {'lr': 0.0003965621608321481, 'samples': 3154560, 'steps': 16429, 'loss/train': 1.3365381360054016} 01/29/2022 09:57:23 - INFO - codeparrot_training - Step 16430: {'lr': 0.0003965489047778158, 'samples': 3154752, 'steps': 16430, 'loss/train': 1.4114321172237396} 01/29/2022 09:57:27 - INFO - codeparrot_training - Step 16431: {'lr': 0.0003965356480957145, 'samples': 3154944, 'steps': 16431, 'loss/train': 2.792866587638855} 01/29/2022 09:57:32 - INFO - codeparrot_training - Step 16432: {'lr': 0.0003965223907859011, 'samples': 3155136, 'steps': 16432, 'loss/train': 1.4973228871822357} 01/29/2022 09:57:36 - INFO - codeparrot_training - Step 16433: {'lr': 0.00039650913284843225, 'samples': 3155328, 'steps': 16433, 'loss/train': 2.0862075090408325} 01/29/2022 09:57:43 - INFO - codeparrot_training - Step 16434: {'lr': 0.00039649587428336474, 'samples': 3155520, 'steps': 16434, 'loss/train': 2.2427626848220825} 01/29/2022 09:57:47 - INFO - codeparrot_training - Step 16435: {'lr': 0.00039648261509075554, 'samples': 3155712, 'steps': 16435, 'loss/train': 1.4425588846206665} 01/29/2022 09:57:52 - INFO - codeparrot_training - Step 16436: {'lr': 0.00039646935527066124, 'samples': 3155904, 'steps': 16436, 'loss/train': 1.6806288957595825} 01/29/2022 09:57:56 - INFO - codeparrot_training - Step 16437: {'lr': 0.0003964560948231388, 'samples': 3156096, 'steps': 16437, 'loss/train': 1.6289759874343872} 01/29/2022 09:58:00 - INFO - codeparrot_training - Step 16438: {'lr': 0.0003964428337482449, 'samples': 3156288, 'steps': 16438, 'loss/train': 1.5462504029273987} 01/29/2022 09:58:05 - INFO - codeparrot_training - Step 16439: {'lr': 0.00039642957204603647, 'samples': 3156480, 'steps': 16439, 'loss/train': 1.3623686134815216} 01/29/2022 09:58:10 - INFO - codeparrot_training - Step 16440: {'lr': 0.0003964163097165702, 'samples': 3156672, 'steps': 16440, 'loss/train': 1.3475196361541748} 01/29/2022 09:58:14 - INFO - codeparrot_training - Step 16441: {'lr': 0.0003964030467599029, 'samples': 3156864, 'steps': 16441, 'loss/train': 0.41176633536815643} 01/29/2022 09:58:18 - INFO - codeparrot_training - Step 16442: {'lr': 0.00039638978317609155, 'samples': 3157056, 'steps': 16442, 'loss/train': 1.4232304394245148} 01/29/2022 09:58:22 - INFO - codeparrot_training - Step 16443: {'lr': 0.0003963765189651928, 'samples': 3157248, 'steps': 16443, 'loss/train': 1.2272733449935913} 01/29/2022 09:58:28 - INFO - codeparrot_training - Step 16444: {'lr': 0.0003963632541272635, 'samples': 3157440, 'steps': 16444, 'loss/train': 0.8639549911022186} 01/29/2022 09:58:32 - INFO - codeparrot_training - Step 16445: {'lr': 0.00039634998866236047, 'samples': 3157632, 'steps': 16445, 'loss/train': 1.5964876413345337} 01/29/2022 09:58:36 - INFO - codeparrot_training - Step 16446: {'lr': 0.0003963367225705406, 'samples': 3157824, 'steps': 16446, 'loss/train': 1.925788164138794} 01/29/2022 09:58:41 - INFO - codeparrot_training - Step 16447: {'lr': 0.0003963234558518607, 'samples': 3158016, 'steps': 16447, 'loss/train': 2.277436137199402} 01/29/2022 09:58:45 - INFO - codeparrot_training - Step 16448: {'lr': 0.0003963101885063776, 'samples': 3158208, 'steps': 16448, 'loss/train': 1.3049277663230896} 01/29/2022 09:58:52 - INFO - codeparrot_training - Step 16449: {'lr': 0.000396296920534148, 'samples': 3158400, 'steps': 16449, 'loss/train': 1.8481927514076233} 01/29/2022 09:58:56 - INFO - codeparrot_training - Step 16450: {'lr': 0.000396283651935229, 'samples': 3158592, 'steps': 16450, 'loss/train': 1.9058502316474915} 01/29/2022 09:59:01 - INFO - codeparrot_training - Step 16451: {'lr': 0.0003962703827096771, 'samples': 3158784, 'steps': 16451, 'loss/train': 1.1450316309928894} 01/29/2022 09:59:05 - INFO - codeparrot_training - Step 16452: {'lr': 0.00039625711285754943, 'samples': 3158976, 'steps': 16452, 'loss/train': 1.4364323616027832} 01/29/2022 09:59:09 - INFO - codeparrot_training - Step 16453: {'lr': 0.00039624384237890275, 'samples': 3159168, 'steps': 16453, 'loss/train': 2.452010214328766} 01/29/2022 09:59:14 - INFO - codeparrot_training - Step 16454: {'lr': 0.00039623057127379386, 'samples': 3159360, 'steps': 16454, 'loss/train': 1.6423770189285278} 01/29/2022 09:59:19 - INFO - codeparrot_training - Step 16455: {'lr': 0.0003962172995422796, 'samples': 3159552, 'steps': 16455, 'loss/train': 1.817514717578888} 01/29/2022 09:59:23 - INFO - codeparrot_training - Step 16456: {'lr': 0.00039620402718441687, 'samples': 3159744, 'steps': 16456, 'loss/train': 1.7328983545303345} 01/29/2022 09:59:27 - INFO - codeparrot_training - Step 16457: {'lr': 0.0003961907542002626, 'samples': 3159936, 'steps': 16457, 'loss/train': 0.9740546643733978} 01/29/2022 09:59:31 - INFO - codeparrot_training - Step 16458: {'lr': 0.00039617748058987345, 'samples': 3160128, 'steps': 16458, 'loss/train': 1.0239468812942505} 01/29/2022 09:59:39 - INFO - codeparrot_training - Step 16459: {'lr': 0.0003961642063533065, 'samples': 3160320, 'steps': 16459, 'loss/train': 1.9137312769889832} 01/29/2022 09:59:43 - INFO - codeparrot_training - Step 16460: {'lr': 0.0003961509314906184, 'samples': 3160512, 'steps': 16460, 'loss/train': 1.0648790895938873} 01/29/2022 09:59:47 - INFO - codeparrot_training - Step 16461: {'lr': 0.0003961376560018662, 'samples': 3160704, 'steps': 16461, 'loss/train': 1.6658434867858887} 01/29/2022 09:59:51 - INFO - codeparrot_training - Step 16462: {'lr': 0.0003961243798871066, 'samples': 3160896, 'steps': 16462, 'loss/train': 1.6636117100715637} 01/29/2022 09:59:56 - INFO - codeparrot_training - Step 16463: {'lr': 0.00039611110314639663, 'samples': 3161088, 'steps': 16463, 'loss/train': 1.855680227279663} 01/29/2022 10:00:01 - INFO - codeparrot_training - Step 16464: {'lr': 0.00039609782577979306, 'samples': 3161280, 'steps': 16464, 'loss/train': 5.065540194511414} 01/29/2022 10:00:05 - INFO - codeparrot_training - Step 16465: {'lr': 0.0003960845477873528, 'samples': 3161472, 'steps': 16465, 'loss/train': 1.6591735482215881} 01/29/2022 10:00:09 - INFO - codeparrot_training - Step 16466: {'lr': 0.00039607126916913274, 'samples': 3161664, 'steps': 16466, 'loss/train': 2.5877553820610046} 01/29/2022 10:00:13 - INFO - codeparrot_training - Step 16467: {'lr': 0.00039605798992518973, 'samples': 3161856, 'steps': 16467, 'loss/train': 2.316097676753998} 01/29/2022 10:00:18 - INFO - codeparrot_training - Step 16468: {'lr': 0.00039604471005558065, 'samples': 3162048, 'steps': 16468, 'loss/train': 1.1394854485988617} 01/29/2022 10:00:23 - INFO - codeparrot_training - Step 16469: {'lr': 0.0003960314295603624, 'samples': 3162240, 'steps': 16469, 'loss/train': 1.2369599640369415} 01/29/2022 10:00:28 - INFO - codeparrot_training - Step 16470: {'lr': 0.00039601814843959193, 'samples': 3162432, 'steps': 16470, 'loss/train': 1.355243593454361} 01/29/2022 10:00:32 - INFO - codeparrot_training - Step 16471: {'lr': 0.00039600486669332603, 'samples': 3162624, 'steps': 16471, 'loss/train': 1.101194590330124} 01/29/2022 10:00:36 - INFO - codeparrot_training - Step 16472: {'lr': 0.00039599158432162163, 'samples': 3162816, 'steps': 16472, 'loss/train': 0.8149831295013428} 01/29/2022 10:00:40 - INFO - codeparrot_training - Step 16473: {'lr': 0.0003959783013245357, 'samples': 3163008, 'steps': 16473, 'loss/train': 2.1018898487091064} 01/29/2022 10:00:46 - INFO - codeparrot_training - Step 16474: {'lr': 0.000395965017702125, 'samples': 3163200, 'steps': 16474, 'loss/train': 1.987441599369049} 01/29/2022 10:00:50 - INFO - codeparrot_training - Step 16475: {'lr': 0.00039595173345444656, 'samples': 3163392, 'steps': 16475, 'loss/train': 1.763448178768158} 01/29/2022 10:00:55 - INFO - codeparrot_training - Step 16476: {'lr': 0.0003959384485815573, 'samples': 3163584, 'steps': 16476, 'loss/train': 2.3337857723236084} 01/29/2022 10:00:59 - INFO - codeparrot_training - Step 16477: {'lr': 0.000395925163083514, 'samples': 3163776, 'steps': 16477, 'loss/train': 1.561941683292389} 01/29/2022 10:01:06 - INFO - codeparrot_training - Step 16478: {'lr': 0.00039591187696037366, 'samples': 3163968, 'steps': 16478, 'loss/train': 1.9361302256584167} 01/29/2022 10:01:11 - INFO - codeparrot_training - Step 16479: {'lr': 0.0003958985902121931, 'samples': 3164160, 'steps': 16479, 'loss/train': 1.763413667678833} 01/29/2022 10:01:15 - INFO - codeparrot_training - Step 16480: {'lr': 0.00039588530283902936, 'samples': 3164352, 'steps': 16480, 'loss/train': 2.557704985141754} 01/29/2022 10:01:19 - INFO - codeparrot_training - Step 16481: {'lr': 0.00039587201484093937, 'samples': 3164544, 'steps': 16481, 'loss/train': 1.2966244518756866} 01/29/2022 10:01:23 - INFO - codeparrot_training - Step 16482: {'lr': 0.0003958587262179799, 'samples': 3164736, 'steps': 16482, 'loss/train': 2.064624845981598} 01/29/2022 10:01:29 - INFO - codeparrot_training - Step 16483: {'lr': 0.00039584543697020804, 'samples': 3164928, 'steps': 16483, 'loss/train': 1.2570989727973938} 01/29/2022 10:01:33 - INFO - codeparrot_training - Step 16484: {'lr': 0.00039583214709768054, 'samples': 3165120, 'steps': 16484, 'loss/train': 1.9883456826210022} 01/29/2022 10:01:37 - INFO - codeparrot_training - Step 16485: {'lr': 0.00039581885660045445, 'samples': 3165312, 'steps': 16485, 'loss/train': 1.5234371423721313} 01/29/2022 10:01:41 - INFO - codeparrot_training - Step 16486: {'lr': 0.0003958055654785867, 'samples': 3165504, 'steps': 16486, 'loss/train': 1.7291793823242188} 01/29/2022 10:01:46 - INFO - codeparrot_training - Step 16487: {'lr': 0.0003957922737321343, 'samples': 3165696, 'steps': 16487, 'loss/train': 1.8120052814483643} 01/29/2022 10:01:51 - INFO - codeparrot_training - Step 16488: {'lr': 0.00039577898136115397, 'samples': 3165888, 'steps': 16488, 'loss/train': 1.5673542022705078} 01/29/2022 10:01:55 - INFO - codeparrot_training - Step 16489: {'lr': 0.00039576568836570283, 'samples': 3166080, 'steps': 16489, 'loss/train': 0.3780474364757538} 01/29/2022 10:02:00 - INFO - codeparrot_training - Step 16490: {'lr': 0.0003957523947458377, 'samples': 3166272, 'steps': 16490, 'loss/train': 2.5790035128593445} 01/29/2022 10:02:04 - INFO - codeparrot_training - Step 16491: {'lr': 0.00039573910050161564, 'samples': 3166464, 'steps': 16491, 'loss/train': 1.0481556951999664} 01/29/2022 10:02:08 - INFO - codeparrot_training - Step 16492: {'lr': 0.0003957258056330936, 'samples': 3166656, 'steps': 16492, 'loss/train': 1.8490567803382874} 01/29/2022 10:02:15 - INFO - codeparrot_training - Step 16493: {'lr': 0.00039571251014032847, 'samples': 3166848, 'steps': 16493, 'loss/train': 2.0104088187217712} 01/29/2022 10:02:20 - INFO - codeparrot_training - Step 16494: {'lr': 0.00039569921402337715, 'samples': 3167040, 'steps': 16494, 'loss/train': 1.5497469305992126} 01/29/2022 10:02:24 - INFO - codeparrot_training - Step 16495: {'lr': 0.00039568591728229667, 'samples': 3167232, 'steps': 16495, 'loss/train': 1.5829605460166931} 01/29/2022 10:02:28 - INFO - codeparrot_training - Step 16496: {'lr': 0.00039567261991714406, 'samples': 3167424, 'steps': 16496, 'loss/train': 1.5383226871490479} 01/29/2022 10:02:32 - INFO - codeparrot_training - Step 16497: {'lr': 0.0003956593219279761, 'samples': 3167616, 'steps': 16497, 'loss/train': 1.0172449350357056} 01/29/2022 10:02:37 - INFO - codeparrot_training - Step 16498: {'lr': 0.00039564602331484993, 'samples': 3167808, 'steps': 16498, 'loss/train': 2.1465872526168823} 01/29/2022 10:02:42 - INFO - codeparrot_training - Step 16499: {'lr': 0.0003956327240778224, 'samples': 3168000, 'steps': 16499, 'loss/train': 1.5393855571746826} 01/29/2022 10:02:46 - INFO - codeparrot_training - Step 16500: {'lr': 0.00039561942421695057, 'samples': 3168192, 'steps': 16500, 'loss/train': 1.859079122543335} 01/29/2022 10:02:50 - INFO - codeparrot_training - Step 16501: {'lr': 0.00039560612373229135, 'samples': 3168384, 'steps': 16501, 'loss/train': 1.6054020524024963} 01/29/2022 10:02:54 - INFO - codeparrot_training - Step 16502: {'lr': 0.0003955928226239017, 'samples': 3168576, 'steps': 16502, 'loss/train': 2.4443060159683228} 01/29/2022 10:03:00 - INFO - codeparrot_training - Step 16503: {'lr': 0.00039557952089183863, 'samples': 3168768, 'steps': 16503, 'loss/train': 0.9892701208591461} 01/29/2022 10:03:04 - INFO - codeparrot_training - Step 16504: {'lr': 0.00039556621853615914, 'samples': 3168960, 'steps': 16504, 'loss/train': 2.1637154817581177} 01/29/2022 10:03:08 - INFO - codeparrot_training - Step 16505: {'lr': 0.0003955529155569202, 'samples': 3169152, 'steps': 16505, 'loss/train': 1.604027509689331} 01/29/2022 10:03:12 - INFO - codeparrot_training - Step 16506: {'lr': 0.0003955396119541788, 'samples': 3169344, 'steps': 16506, 'loss/train': 2.123780608177185} 01/29/2022 10:03:17 - INFO - codeparrot_training - Step 16507: {'lr': 0.00039552630772799185, 'samples': 3169536, 'steps': 16507, 'loss/train': 2.000720500946045} 01/29/2022 10:03:23 - INFO - codeparrot_training - Step 16508: {'lr': 0.0003955130028784165, 'samples': 3169728, 'steps': 16508, 'loss/train': 2.117712378501892} 01/29/2022 10:03:28 - INFO - codeparrot_training - Step 16509: {'lr': 0.00039549969740550954, 'samples': 3169920, 'steps': 16509, 'loss/train': 1.4271519184112549} 01/29/2022 10:03:32 - INFO - codeparrot_training - Step 16510: {'lr': 0.00039548639130932816, 'samples': 3170112, 'steps': 16510, 'loss/train': 1.5825544595718384} 01/29/2022 10:03:36 - INFO - codeparrot_training - Step 16511: {'lr': 0.00039547308458992927, 'samples': 3170304, 'steps': 16511, 'loss/train': 1.8014490008354187} 01/29/2022 10:03:41 - INFO - codeparrot_training - Step 16512: {'lr': 0.00039545977724736984, 'samples': 3170496, 'steps': 16512, 'loss/train': 2.417867124080658} 01/29/2022 10:03:46 - INFO - codeparrot_training - Step 16513: {'lr': 0.00039544646928170695, 'samples': 3170688, 'steps': 16513, 'loss/train': 1.9197882413864136} 01/29/2022 10:03:50 - INFO - codeparrot_training - Step 16514: {'lr': 0.0003954331606929976, 'samples': 3170880, 'steps': 16514, 'loss/train': 2.461312472820282} 01/29/2022 10:03:54 - INFO - codeparrot_training - Step 16515: {'lr': 0.00039541985148129865, 'samples': 3171072, 'steps': 16515, 'loss/train': 1.9011712074279785} 01/29/2022 10:03:59 - INFO - codeparrot_training - Step 16516: {'lr': 0.00039540654164666735, 'samples': 3171264, 'steps': 16516, 'loss/train': 1.7663451433181763} 01/29/2022 10:04:03 - INFO - codeparrot_training - Step 16517: {'lr': 0.00039539323118916055, 'samples': 3171456, 'steps': 16517, 'loss/train': 0.7631203830242157} 01/29/2022 10:04:08 - INFO - codeparrot_training - Step 16518: {'lr': 0.0003953799201088353, 'samples': 3171648, 'steps': 16518, 'loss/train': 1.4464370012283325} 01/29/2022 10:04:12 - INFO - codeparrot_training - Step 16519: {'lr': 0.00039536660840574866, 'samples': 3171840, 'steps': 16519, 'loss/train': 1.644777774810791} 01/29/2022 10:04:17 - INFO - codeparrot_training - Step 16520: {'lr': 0.0003953532960799577, 'samples': 3172032, 'steps': 16520, 'loss/train': 1.9035694599151611} 01/29/2022 10:04:21 - INFO - codeparrot_training - Step 16521: {'lr': 0.00039533998313151926, 'samples': 3172224, 'steps': 16521, 'loss/train': 1.8700028657913208} 01/29/2022 10:04:25 - INFO - codeparrot_training - Step 16522: {'lr': 0.0003953266695604906, 'samples': 3172416, 'steps': 16522, 'loss/train': 1.8114774227142334} 01/29/2022 10:04:31 - INFO - codeparrot_training - Step 16523: {'lr': 0.0003953133553669285, 'samples': 3172608, 'steps': 16523, 'loss/train': 1.2196729481220245} 01/29/2022 10:04:35 - INFO - codeparrot_training - Step 16524: {'lr': 0.0003953000405508902, 'samples': 3172800, 'steps': 16524, 'loss/train': 0.4316704720258713} 01/29/2022 10:04:39 - INFO - codeparrot_training - Step 16525: {'lr': 0.00039528672511243256, 'samples': 3172992, 'steps': 16525, 'loss/train': 1.453224241733551} 01/29/2022 10:04:43 - INFO - codeparrot_training - Step 16526: {'lr': 0.0003952734090516129, 'samples': 3173184, 'steps': 16526, 'loss/train': 1.7400317788124084} 01/29/2022 10:04:48 - INFO - codeparrot_training - Step 16527: {'lr': 0.000395260092368488, 'samples': 3173376, 'steps': 16527, 'loss/train': 1.9956103563308716} 01/29/2022 10:04:55 - INFO - codeparrot_training - Step 16528: {'lr': 0.000395246775063115, 'samples': 3173568, 'steps': 16528, 'loss/train': 1.9447535276412964} 01/29/2022 10:04:59 - INFO - codeparrot_training - Step 16529: {'lr': 0.0003952334571355509, 'samples': 3173760, 'steps': 16529, 'loss/train': 2.058306634426117} 01/29/2022 10:05:03 - INFO - codeparrot_training - Step 16530: {'lr': 0.0003952201385858528, 'samples': 3173952, 'steps': 16530, 'loss/train': 1.808964192867279} 01/29/2022 10:05:08 - INFO - codeparrot_training - Step 16531: {'lr': 0.00039520681941407777, 'samples': 3174144, 'steps': 16531, 'loss/train': 1.4054943323135376} 01/29/2022 10:05:12 - INFO - codeparrot_training - Step 16532: {'lr': 0.00039519349962028276, 'samples': 3174336, 'steps': 16532, 'loss/train': 0.9324934780597687} 01/29/2022 10:05:17 - INFO - codeparrot_training - Step 16533: {'lr': 0.000395180179204525, 'samples': 3174528, 'steps': 16533, 'loss/train': 2.7161863446235657} 01/29/2022 10:05:21 - INFO - codeparrot_training - Step 16534: {'lr': 0.0003951668581668614, 'samples': 3174720, 'steps': 16534, 'loss/train': 1.7732786536216736} 01/29/2022 10:05:26 - INFO - codeparrot_training - Step 16535: {'lr': 0.0003951535365073491, 'samples': 3174912, 'steps': 16535, 'loss/train': 0.6042108535766602} 01/29/2022 10:05:30 - INFO - codeparrot_training - Step 16536: {'lr': 0.00039514021422604515, 'samples': 3175104, 'steps': 16536, 'loss/train': 2.0263620615005493} 01/29/2022 10:05:34 - INFO - codeparrot_training - Step 16537: {'lr': 0.0003951268913230066, 'samples': 3175296, 'steps': 16537, 'loss/train': 2.435311496257782} 01/29/2022 10:05:41 - INFO - codeparrot_training - Step 16538: {'lr': 0.0003951135677982904, 'samples': 3175488, 'steps': 16538, 'loss/train': 2.3793030381202698} 01/29/2022 10:05:45 - INFO - codeparrot_training - Step 16539: {'lr': 0.000395100243651954, 'samples': 3175680, 'steps': 16539, 'loss/train': 2.1967413425445557} 01/29/2022 10:05:50 - INFO - codeparrot_training - Step 16540: {'lr': 0.00039508691888405403, 'samples': 3175872, 'steps': 16540, 'loss/train': 1.25507453083992} 01/29/2022 10:05:54 - INFO - codeparrot_training - Step 16541: {'lr': 0.0003950735934946478, 'samples': 3176064, 'steps': 16541, 'loss/train': 2.059330701828003} 01/29/2022 10:05:58 - INFO - codeparrot_training - Step 16542: {'lr': 0.0003950602674837924, 'samples': 3176256, 'steps': 16542, 'loss/train': 0.7932861149311066} 01/29/2022 10:06:03 - INFO - codeparrot_training - Step 16543: {'lr': 0.0003950469408515449, 'samples': 3176448, 'steps': 16543, 'loss/train': 1.2289756536483765} 01/29/2022 10:06:08 - INFO - codeparrot_training - Step 16544: {'lr': 0.00039503361359796235, 'samples': 3176640, 'steps': 16544, 'loss/train': 0.27532660961151123} 01/29/2022 10:06:12 - INFO - codeparrot_training - Step 16545: {'lr': 0.00039502028572310186, 'samples': 3176832, 'steps': 16545, 'loss/train': 2.1534531712532043} 01/29/2022 10:06:16 - INFO - codeparrot_training - Step 16546: {'lr': 0.0003950069572270205, 'samples': 3177024, 'steps': 16546, 'loss/train': 2.117771029472351} 01/29/2022 10:06:21 - INFO - codeparrot_training - Step 16547: {'lr': 0.00039499362810977535, 'samples': 3177216, 'steps': 16547, 'loss/train': 2.163719058036804} 01/29/2022 10:06:26 - INFO - codeparrot_training - Step 16548: {'lr': 0.00039498029837142356, 'samples': 3177408, 'steps': 16548, 'loss/train': 1.089830070734024} 01/29/2022 10:06:30 - INFO - codeparrot_training - Step 16549: {'lr': 0.0003949669680120223, 'samples': 3177600, 'steps': 16549, 'loss/train': 1.9007141590118408} 01/29/2022 10:06:34 - INFO - codeparrot_training - Step 16550: {'lr': 0.00039495363703162843, 'samples': 3177792, 'steps': 16550, 'loss/train': 0.9742453694343567} 01/29/2022 10:06:39 - INFO - codeparrot_training - Step 16551: {'lr': 0.00039494030543029925, 'samples': 3177984, 'steps': 16551, 'loss/train': 1.668297529220581} 01/29/2022 10:06:43 - INFO - codeparrot_training - Step 16552: {'lr': 0.0003949269732080919, 'samples': 3178176, 'steps': 16552, 'loss/train': 1.5024408102035522} 01/29/2022 10:06:49 - INFO - codeparrot_training - Step 16553: {'lr': 0.0003949136403650633, 'samples': 3178368, 'steps': 16553, 'loss/train': 1.6356546878814697} 01/29/2022 10:06:54 - INFO - codeparrot_training - Step 16554: {'lr': 0.0003949003069012708, 'samples': 3178560, 'steps': 16554, 'loss/train': 1.7356928586959839} 01/29/2022 10:06:58 - INFO - codeparrot_training - Step 16555: {'lr': 0.0003948869728167713, 'samples': 3178752, 'steps': 16555, 'loss/train': 1.6565762758255005} 01/29/2022 10:07:02 - INFO - codeparrot_training - Step 16556: {'lr': 0.0003948736381116221, 'samples': 3178944, 'steps': 16556, 'loss/train': 0.11675149574875832} 01/29/2022 10:07:06 - INFO - codeparrot_training - Step 16557: {'lr': 0.0003948603027858802, 'samples': 3179136, 'steps': 16557, 'loss/train': 2.3331159353256226} 01/29/2022 10:07:12 - INFO - codeparrot_training - Step 16558: {'lr': 0.00039484696683960276, 'samples': 3179328, 'steps': 16558, 'loss/train': 1.1762832999229431} 01/29/2022 10:07:16 - INFO - codeparrot_training - Step 16559: {'lr': 0.0003948336302728469, 'samples': 3179520, 'steps': 16559, 'loss/train': 1.884304404258728} 01/29/2022 10:07:20 - INFO - codeparrot_training - Step 16560: {'lr': 0.0003948202930856697, 'samples': 3179712, 'steps': 16560, 'loss/train': 1.4349595606327057} 01/29/2022 10:07:25 - INFO - codeparrot_training - Step 16561: {'lr': 0.0003948069552781285, 'samples': 3179904, 'steps': 16561, 'loss/train': 1.7169976830482483} 01/29/2022 10:07:29 - INFO - codeparrot_training - Step 16562: {'lr': 0.00039479361685028016, 'samples': 3180096, 'steps': 16562, 'loss/train': 1.859145998954773} 01/29/2022 10:07:36 - INFO - codeparrot_training - Step 16563: {'lr': 0.00039478027780218193, 'samples': 3180288, 'steps': 16563, 'loss/train': 1.9528185725212097} 01/29/2022 10:07:40 - INFO - codeparrot_training - Step 16564: {'lr': 0.00039476693813389105, 'samples': 3180480, 'steps': 16564, 'loss/train': 1.8139736652374268} 01/29/2022 10:07:45 - INFO - codeparrot_training - Step 16565: {'lr': 0.0003947535978454645, 'samples': 3180672, 'steps': 16565, 'loss/train': 0.9553848803043365} 01/29/2022 10:07:49 - INFO - codeparrot_training - Step 16566: {'lr': 0.0003947402569369596, 'samples': 3180864, 'steps': 16566, 'loss/train': 1.6173568367958069} 01/29/2022 10:07:53 - INFO - codeparrot_training - Step 16567: {'lr': 0.0003947269154084333, 'samples': 3181056, 'steps': 16567, 'loss/train': 2.63725608587265} 01/29/2022 10:07:59 - INFO - codeparrot_training - Step 16568: {'lr': 0.0003947135732599428, 'samples': 3181248, 'steps': 16568, 'loss/train': 1.9539572596549988} 01/29/2022 10:08:03 - INFO - codeparrot_training - Step 16569: {'lr': 0.00039470023049154544, 'samples': 3181440, 'steps': 16569, 'loss/train': 1.701082170009613} 01/29/2022 10:08:07 - INFO - codeparrot_training - Step 16570: {'lr': 0.00039468688710329826, 'samples': 3181632, 'steps': 16570, 'loss/train': 2.3136096596717834} 01/29/2022 10:08:11 - INFO - codeparrot_training - Step 16571: {'lr': 0.0003946735430952583, 'samples': 3181824, 'steps': 16571, 'loss/train': 1.7828637957572937} 01/29/2022 10:08:16 - INFO - codeparrot_training - Step 16572: {'lr': 0.0003946601984674828, 'samples': 3182016, 'steps': 16572, 'loss/train': 1.7603829503059387} 01/29/2022 10:08:21 - INFO - codeparrot_training - Step 16573: {'lr': 0.00039464685322002904, 'samples': 3182208, 'steps': 16573, 'loss/train': 1.965904176235199} 01/29/2022 10:08:26 - INFO - codeparrot_training - Step 16574: {'lr': 0.000394633507352954, 'samples': 3182400, 'steps': 16574, 'loss/train': 2.2079665660858154} 01/29/2022 10:08:30 - INFO - codeparrot_training - Step 16575: {'lr': 0.00039462016086631505, 'samples': 3182592, 'steps': 16575, 'loss/train': 1.992587149143219} 01/29/2022 10:08:34 - INFO - codeparrot_training - Step 16576: {'lr': 0.00039460681376016915, 'samples': 3182784, 'steps': 16576, 'loss/train': 1.8369011878967285} 01/29/2022 10:08:38 - INFO - codeparrot_training - Step 16577: {'lr': 0.0003945934660345736, 'samples': 3182976, 'steps': 16577, 'loss/train': 2.288991093635559} 01/29/2022 10:08:44 - INFO - codeparrot_training - Step 16578: {'lr': 0.00039458011768958557, 'samples': 3183168, 'steps': 16578, 'loss/train': 0.5220033377408981} 01/29/2022 10:08:48 - INFO - codeparrot_training - Step 16579: {'lr': 0.00039456676872526227, 'samples': 3183360, 'steps': 16579, 'loss/train': 1.408366173505783} 01/29/2022 10:08:53 - INFO - codeparrot_training - Step 16580: {'lr': 0.00039455341914166074, 'samples': 3183552, 'steps': 16580, 'loss/train': 2.543523609638214} 01/29/2022 10:08:57 - INFO - codeparrot_training - Step 16581: {'lr': 0.0003945400689388384, 'samples': 3183744, 'steps': 16581, 'loss/train': 1.77419775724411} 01/29/2022 10:09:01 - INFO - codeparrot_training - Step 16582: {'lr': 0.00039452671811685214, 'samples': 3183936, 'steps': 16582, 'loss/train': 1.1506903767585754} 01/29/2022 10:09:09 - INFO - codeparrot_training - Step 16583: {'lr': 0.00039451336667575945, 'samples': 3184128, 'steps': 16583, 'loss/train': 2.18042653799057} 01/29/2022 10:09:13 - INFO - codeparrot_training - Step 16584: {'lr': 0.0003945000146156173, 'samples': 3184320, 'steps': 16584, 'loss/train': 2.902724325656891} 01/29/2022 10:09:17 - INFO - codeparrot_training - Step 16585: {'lr': 0.00039448666193648305, 'samples': 3184512, 'steps': 16585, 'loss/train': 2.5255526304244995} 01/29/2022 10:09:21 - INFO - codeparrot_training - Step 16586: {'lr': 0.0003944733086384137, 'samples': 3184704, 'steps': 16586, 'loss/train': 1.3919043838977814} 01/29/2022 10:09:26 - INFO - codeparrot_training - Step 16587: {'lr': 0.00039445995472146665, 'samples': 3184896, 'steps': 16587, 'loss/train': 1.2453790605068207} 01/29/2022 10:09:31 - INFO - codeparrot_training - Step 16588: {'lr': 0.000394446600185699, 'samples': 3185088, 'steps': 16588, 'loss/train': 1.3477823138237} 01/29/2022 10:09:35 - INFO - codeparrot_training - Step 16589: {'lr': 0.000394433245031168, 'samples': 3185280, 'steps': 16589, 'loss/train': 1.5747767686843872} 01/29/2022 10:09:40 - INFO - codeparrot_training - Step 16590: {'lr': 0.0003944198892579309, 'samples': 3185472, 'steps': 16590, 'loss/train': 1.9511622190475464} 01/29/2022 10:09:44 - INFO - codeparrot_training - Step 16591: {'lr': 0.0003944065328660447, 'samples': 3185664, 'steps': 16591, 'loss/train': 1.0717131793498993} 01/29/2022 10:09:48 - INFO - codeparrot_training - Step 16592: {'lr': 0.0003943931758555669, 'samples': 3185856, 'steps': 16592, 'loss/train': 1.728873074054718} 01/29/2022 10:09:54 - INFO - codeparrot_training - Step 16593: {'lr': 0.00039437981822655453, 'samples': 3186048, 'steps': 16593, 'loss/train': 1.8469734191894531} 01/29/2022 10:09:58 - INFO - codeparrot_training - Step 16594: {'lr': 0.00039436645997906487, 'samples': 3186240, 'steps': 16594, 'loss/train': 2.2799545526504517} 01/29/2022 10:10:02 - INFO - codeparrot_training - Step 16595: {'lr': 0.00039435310111315513, 'samples': 3186432, 'steps': 16595, 'loss/train': 1.6332462430000305} 01/29/2022 10:10:07 - INFO - codeparrot_training - Step 16596: {'lr': 0.00039433974162888266, 'samples': 3186624, 'steps': 16596, 'loss/train': 1.9357166290283203} 01/29/2022 10:10:11 - INFO - codeparrot_training - Step 16597: {'lr': 0.0003943263815263044, 'samples': 3186816, 'steps': 16597, 'loss/train': 1.3057205379009247} 01/29/2022 10:10:18 - INFO - codeparrot_training - Step 16598: {'lr': 0.0003943130208054778, 'samples': 3187008, 'steps': 16598, 'loss/train': 1.9226428270339966} 01/29/2022 10:10:22 - INFO - codeparrot_training - Step 16599: {'lr': 0.0003942996594664601, 'samples': 3187200, 'steps': 16599, 'loss/train': 1.7951129078865051} 01/29/2022 10:10:27 - INFO - codeparrot_training - Step 16600: {'lr': 0.00039428629750930846, 'samples': 3187392, 'steps': 16600, 'loss/train': 2.404941201210022} 01/29/2022 10:10:31 - INFO - codeparrot_training - Step 16601: {'lr': 0.0003942729349340801, 'samples': 3187584, 'steps': 16601, 'loss/train': 2.111468017101288} 01/29/2022 10:10:35 - INFO - codeparrot_training - Step 16602: {'lr': 0.00039425957174083224, 'samples': 3187776, 'steps': 16602, 'loss/train': 1.0949051678180695} 01/29/2022 10:10:41 - INFO - codeparrot_training - Step 16603: {'lr': 0.0003942462079296223, 'samples': 3187968, 'steps': 16603, 'loss/train': 1.112159103155136} 01/29/2022 10:10:45 - INFO - codeparrot_training - Step 16604: {'lr': 0.00039423284350050735, 'samples': 3188160, 'steps': 16604, 'loss/train': 1.7977532744407654} 01/29/2022 10:10:49 - INFO - codeparrot_training - Step 16605: {'lr': 0.00039421947845354476, 'samples': 3188352, 'steps': 16605, 'loss/train': 1.422737717628479} 01/29/2022 10:10:54 - INFO - codeparrot_training - Step 16606: {'lr': 0.0003942061127887916, 'samples': 3188544, 'steps': 16606, 'loss/train': 2.2532694339752197} 01/29/2022 10:10:58 - INFO - codeparrot_training - Step 16607: {'lr': 0.00039419274650630536, 'samples': 3188736, 'steps': 16607, 'loss/train': 1.6283848285675049} 01/29/2022 10:11:04 - INFO - codeparrot_training - Step 16608: {'lr': 0.00039417937960614316, 'samples': 3188928, 'steps': 16608, 'loss/train': 0.11215128377079964} 01/29/2022 10:11:08 - INFO - codeparrot_training - Step 16609: {'lr': 0.0003941660120883622, 'samples': 3189120, 'steps': 16609, 'loss/train': 3.7304495573043823} 01/29/2022 10:11:13 - INFO - codeparrot_training - Step 16610: {'lr': 0.0003941526439530199, 'samples': 3189312, 'steps': 16610, 'loss/train': 1.6247733235359192} 01/29/2022 10:11:17 - INFO - codeparrot_training - Step 16611: {'lr': 0.00039413927520017347, 'samples': 3189504, 'steps': 16611, 'loss/train': 1.9555363655090332} 01/29/2022 10:11:21 - INFO - codeparrot_training - Step 16612: {'lr': 0.00039412590582988007, 'samples': 3189696, 'steps': 16612, 'loss/train': 2.0832241773605347} 01/29/2022 10:11:25 - INFO - codeparrot_training - Step 16613: {'lr': 0.00039411253584219707, 'samples': 3189888, 'steps': 16613, 'loss/train': 1.4995031654834747} 01/29/2022 10:11:31 - INFO - codeparrot_training - Step 16614: {'lr': 0.0003940991652371818, 'samples': 3190080, 'steps': 16614, 'loss/train': 1.8601688146591187} 01/29/2022 10:11:35 - INFO - codeparrot_training - Step 16615: {'lr': 0.0003940857940148914, 'samples': 3190272, 'steps': 16615, 'loss/train': 1.57475745677948} 01/29/2022 10:11:39 - INFO - codeparrot_training - Step 16616: {'lr': 0.00039407242217538317, 'samples': 3190464, 'steps': 16616, 'loss/train': 1.6258803606033325} 01/29/2022 10:11:44 - INFO - codeparrot_training - Step 16617: {'lr': 0.00039405904971871454, 'samples': 3190656, 'steps': 16617, 'loss/train': 1.538643479347229} 01/29/2022 10:11:49 - INFO - codeparrot_training - Step 16618: {'lr': 0.00039404567664494264, 'samples': 3190848, 'steps': 16618, 'loss/train': 1.56458580493927} 01/29/2022 10:11:53 - INFO - codeparrot_training - Step 16619: {'lr': 0.0003940323029541248, 'samples': 3191040, 'steps': 16619, 'loss/train': 1.3805764317512512} 01/29/2022 10:11:58 - INFO - codeparrot_training - Step 16620: {'lr': 0.00039401892864631826, 'samples': 3191232, 'steps': 16620, 'loss/train': 1.7339335083961487} 01/29/2022 10:12:02 - INFO - codeparrot_training - Step 16621: {'lr': 0.0003940055537215804, 'samples': 3191424, 'steps': 16621, 'loss/train': 0.7792381346225739} 01/29/2022 10:12:06 - INFO - codeparrot_training - Step 16622: {'lr': 0.0003939921781799685, 'samples': 3191616, 'steps': 16622, 'loss/train': 0.8387044072151184} 01/29/2022 10:12:13 - INFO - codeparrot_training - Step 16623: {'lr': 0.0003939788020215398, 'samples': 3191808, 'steps': 16623, 'loss/train': 2.362669050693512} 01/29/2022 10:12:18 - INFO - codeparrot_training - Step 16624: {'lr': 0.0003939654252463517, 'samples': 3192000, 'steps': 16624, 'loss/train': 1.9692372679710388} 01/29/2022 10:12:22 - INFO - codeparrot_training - Step 16625: {'lr': 0.00039395204785446137, 'samples': 3192192, 'steps': 16625, 'loss/train': 1.6718437671661377} 01/29/2022 10:12:26 - INFO - codeparrot_training - Step 16626: {'lr': 0.00039393866984592616, 'samples': 3192384, 'steps': 16626, 'loss/train': 1.8463810086250305} 01/29/2022 10:12:31 - INFO - codeparrot_training - Step 16627: {'lr': 0.00039392529122080343, 'samples': 3192576, 'steps': 16627, 'loss/train': 2.2890599370002747} 01/29/2022 10:12:35 - INFO - codeparrot_training - Step 16628: {'lr': 0.0003939119119791504, 'samples': 3192768, 'steps': 16628, 'loss/train': 1.0076735615730286} 01/29/2022 10:12:40 - INFO - codeparrot_training - Step 16629: {'lr': 0.0003938985321210245, 'samples': 3192960, 'steps': 16629, 'loss/train': 1.3705705404281616} 01/29/2022 10:12:44 - INFO - codeparrot_training - Step 16630: {'lr': 0.00039388515164648293, 'samples': 3193152, 'steps': 16630, 'loss/train': 2.1753979325294495} 01/29/2022 10:12:49 - INFO - codeparrot_training - Step 16631: {'lr': 0.0003938717705555831, 'samples': 3193344, 'steps': 16631, 'loss/train': 2.2613309025764465} 01/29/2022 10:12:53 - INFO - codeparrot_training - Step 16632: {'lr': 0.0003938583888483823, 'samples': 3193536, 'steps': 16632, 'loss/train': 1.157452404499054} 01/29/2022 10:12:57 - INFO - codeparrot_training - Step 16633: {'lr': 0.0003938450065249378, 'samples': 3193728, 'steps': 16633, 'loss/train': 1.687616765499115} 01/29/2022 10:13:03 - INFO - codeparrot_training - Step 16634: {'lr': 0.00039383162358530696, 'samples': 3193920, 'steps': 16634, 'loss/train': 0.37210940569639206} 01/29/2022 10:13:07 - INFO - codeparrot_training - Step 16635: {'lr': 0.0003938182400295471, 'samples': 3194112, 'steps': 16635, 'loss/train': 1.8253231644630432} 01/29/2022 10:13:11 - INFO - codeparrot_training - Step 16636: {'lr': 0.00039380485585771563, 'samples': 3194304, 'steps': 16636, 'loss/train': 1.4416822493076324} 01/29/2022 10:13:15 - INFO - codeparrot_training - Step 16637: {'lr': 0.00039379147106986985, 'samples': 3194496, 'steps': 16637, 'loss/train': 0.6297416836023331} 01/29/2022 10:13:20 - INFO - codeparrot_training - Step 16638: {'lr': 0.00039377808566606697, 'samples': 3194688, 'steps': 16638, 'loss/train': 2.061771869659424} 01/29/2022 10:13:25 - INFO - codeparrot_training - Step 16639: {'lr': 0.00039376469964636445, 'samples': 3194880, 'steps': 16639, 'loss/train': 1.2772228717803955} 01/29/2022 10:13:29 - INFO - codeparrot_training - Step 16640: {'lr': 0.0003937513130108197, 'samples': 3195072, 'steps': 16640, 'loss/train': 1.961366593837738} 01/29/2022 10:13:33 - INFO - codeparrot_training - Step 16641: {'lr': 0.00039373792575948986, 'samples': 3195264, 'steps': 16641, 'loss/train': 1.3248401284217834} 01/29/2022 10:13:38 - INFO - codeparrot_training - Step 16642: {'lr': 0.00039372453789243245, 'samples': 3195456, 'steps': 16642, 'loss/train': 0.8833397626876831} 01/29/2022 10:13:42 - INFO - codeparrot_training - Step 16643: {'lr': 0.0003937111494097047, 'samples': 3195648, 'steps': 16643, 'loss/train': 0.9323001801967621} 01/29/2022 10:13:50 - INFO - codeparrot_training - Step 16644: {'lr': 0.0003936977603113641, 'samples': 3195840, 'steps': 16644, 'loss/train': 2.3796000480651855} 01/29/2022 10:13:54 - INFO - codeparrot_training - Step 16645: {'lr': 0.00039368437059746785, 'samples': 3196032, 'steps': 16645, 'loss/train': 1.8085706233978271} 01/29/2022 10:13:59 - INFO - codeparrot_training - Step 16646: {'lr': 0.0003936709802680734, 'samples': 3196224, 'steps': 16646, 'loss/train': 0.21686531603336334} 01/29/2022 10:14:03 - INFO - codeparrot_training - Step 16647: {'lr': 0.0003936575893232381, 'samples': 3196416, 'steps': 16647, 'loss/train': 1.4323959052562714} 01/29/2022 10:14:07 - INFO - codeparrot_training - Step 16648: {'lr': 0.0003936441977630193, 'samples': 3196608, 'steps': 16648, 'loss/train': 2.1559683680534363} 01/29/2022 10:14:13 - INFO - codeparrot_training - Step 16649: {'lr': 0.0003936308055874744, 'samples': 3196800, 'steps': 16649, 'loss/train': 2.3571996688842773} 01/29/2022 10:14:17 - INFO - codeparrot_training - Step 16650: {'lr': 0.00039361741279666065, 'samples': 3196992, 'steps': 16650, 'loss/train': 1.875488519668579} 01/29/2022 10:14:21 - INFO - codeparrot_training - Step 16651: {'lr': 0.0003936040193906356, 'samples': 3197184, 'steps': 16651, 'loss/train': 1.6205608248710632} 01/29/2022 10:14:25 - INFO - codeparrot_training - Step 16652: {'lr': 0.00039359062536945645, 'samples': 3197376, 'steps': 16652, 'loss/train': 1.8562704920768738} 01/29/2022 10:14:30 - INFO - codeparrot_training - Step 16653: {'lr': 0.00039357723073318076, 'samples': 3197568, 'steps': 16653, 'loss/train': 1.4853870570659637} 01/29/2022 10:14:37 - INFO - codeparrot_training - Step 16654: {'lr': 0.0003935638354818657, 'samples': 3197760, 'steps': 16654, 'loss/train': 1.7972787022590637} 01/29/2022 10:14:42 - INFO - codeparrot_training - Step 16655: {'lr': 0.0003935504396155688, 'samples': 3197952, 'steps': 16655, 'loss/train': 0.99278524518013} 01/29/2022 10:14:46 - INFO - codeparrot_training - Step 16656: {'lr': 0.00039353704313434745, 'samples': 3198144, 'steps': 16656, 'loss/train': 0.986509770154953} 01/29/2022 10:14:50 - INFO - codeparrot_training - Step 16657: {'lr': 0.000393523646038259, 'samples': 3198336, 'steps': 16657, 'loss/train': 1.9850846529006958} 01/29/2022 10:14:54 - INFO - codeparrot_training - Step 16658: {'lr': 0.0003935102483273607, 'samples': 3198528, 'steps': 16658, 'loss/train': 2.616917073726654} 01/29/2022 10:14:59 - INFO - codeparrot_training - Step 16659: {'lr': 0.0003934968500017101, 'samples': 3198720, 'steps': 16659, 'loss/train': 1.0127062797546387} 01/29/2022 10:15:04 - INFO - codeparrot_training - Step 16660: {'lr': 0.0003934834510613646, 'samples': 3198912, 'steps': 16660, 'loss/train': 0.7800741791725159} 01/29/2022 10:15:08 - INFO - codeparrot_training - Step 16661: {'lr': 0.00039347005150638156, 'samples': 3199104, 'steps': 16661, 'loss/train': 1.665528953075409} 01/29/2022 10:15:12 - INFO - codeparrot_training - Step 16662: {'lr': 0.0003934566513368183, 'samples': 3199296, 'steps': 16662, 'loss/train': 1.0847549736499786} 01/29/2022 10:15:17 - INFO - codeparrot_training - Step 16663: {'lr': 0.00039344325055273236, 'samples': 3199488, 'steps': 16663, 'loss/train': 1.7811350226402283} 01/29/2022 10:15:21 - INFO - codeparrot_training - Step 16664: {'lr': 0.0003934298491541811, 'samples': 3199680, 'steps': 16664, 'loss/train': 2.0610255002975464} 01/29/2022 10:15:26 - INFO - codeparrot_training - Step 16665: {'lr': 0.00039341644714122195, 'samples': 3199872, 'steps': 16665, 'loss/train': 1.2430238127708435} 01/29/2022 10:15:30 - INFO - codeparrot_training - Step 16666: {'lr': 0.00039340304451391216, 'samples': 3200064, 'steps': 16666, 'loss/train': 1.9129709601402283} 01/29/2022 10:15:35 - INFO - codeparrot_training - Step 16667: {'lr': 0.00039338964127230935, 'samples': 3200256, 'steps': 16667, 'loss/train': 2.0587892532348633} 01/29/2022 10:15:39 - INFO - codeparrot_training - Step 16668: {'lr': 0.00039337623741647084, 'samples': 3200448, 'steps': 16668, 'loss/train': 1.8350493907928467} 01/29/2022 10:15:43 - INFO - codeparrot_training - Step 16669: {'lr': 0.000393362832946454, 'samples': 3200640, 'steps': 16669, 'loss/train': 1.822721242904663} 01/29/2022 10:15:50 - INFO - codeparrot_training - Step 16670: {'lr': 0.0003933494278623164, 'samples': 3200832, 'steps': 16670, 'loss/train': 2.1449881196022034} 01/29/2022 10:15:54 - INFO - codeparrot_training - Step 16671: {'lr': 0.0003933360221641153, 'samples': 3201024, 'steps': 16671, 'loss/train': 0.9677494168281555} 01/29/2022 10:15:59 - INFO - codeparrot_training - Step 16672: {'lr': 0.0003933226158519082, 'samples': 3201216, 'steps': 16672, 'loss/train': 1.0756748020648956} 01/29/2022 10:16:03 - INFO - codeparrot_training - Step 16673: {'lr': 0.0003933092089257525, 'samples': 3201408, 'steps': 16673, 'loss/train': 1.0392647981643677} 01/29/2022 10:16:07 - INFO - codeparrot_training - Step 16674: {'lr': 0.0003932958013857057, 'samples': 3201600, 'steps': 16674, 'loss/train': 2.1864014267921448} 01/29/2022 10:16:13 - INFO - codeparrot_training - Step 16675: {'lr': 0.0003932823932318252, 'samples': 3201792, 'steps': 16675, 'loss/train': 1.3847404718399048} 01/29/2022 10:16:17 - INFO - codeparrot_training - Step 16676: {'lr': 0.0003932689844641684, 'samples': 3201984, 'steps': 16676, 'loss/train': 2.5158419609069824} 01/29/2022 10:16:21 - INFO - codeparrot_training - Step 16677: {'lr': 0.00039325557508279276, 'samples': 3202176, 'steps': 16677, 'loss/train': 1.4521050453186035} 01/29/2022 10:16:25 - INFO - codeparrot_training - Step 16678: {'lr': 0.00039324216508775567, 'samples': 3202368, 'steps': 16678, 'loss/train': 1.597252607345581} 01/29/2022 10:16:30 - INFO - codeparrot_training - Step 16679: {'lr': 0.0003932287544791148, 'samples': 3202560, 'steps': 16679, 'loss/train': 1.0847378969192505} 01/29/2022 10:16:35 - INFO - codeparrot_training - Step 16680: {'lr': 0.00039321534325692726, 'samples': 3202752, 'steps': 16680, 'loss/train': 1.8572800755500793} 01/29/2022 10:16:39 - INFO - codeparrot_training - Step 16681: {'lr': 0.0003932019314212507, 'samples': 3202944, 'steps': 16681, 'loss/train': 1.4195149540901184} 01/29/2022 10:16:43 - INFO - codeparrot_training - Step 16682: {'lr': 0.0003931885189721426, 'samples': 3203136, 'steps': 16682, 'loss/train': 1.2949495017528534} 01/29/2022 10:16:47 - INFO - codeparrot_training - Step 16683: {'lr': 0.00039317510590966033, 'samples': 3203328, 'steps': 16683, 'loss/train': 1.3210918307304382} 01/29/2022 10:16:52 - INFO - codeparrot_training - Step 16684: {'lr': 0.0003931616922338613, 'samples': 3203520, 'steps': 16684, 'loss/train': 2.5739235877990723} 01/29/2022 10:16:59 - INFO - codeparrot_training - Step 16685: {'lr': 0.00039314827794480314, 'samples': 3203712, 'steps': 16685, 'loss/train': 1.33837428689003} 01/29/2022 10:17:03 - INFO - codeparrot_training - Step 16686: {'lr': 0.00039313486304254315, 'samples': 3203904, 'steps': 16686, 'loss/train': 1.4516814351081848} 01/29/2022 10:17:07 - INFO - codeparrot_training - Step 16687: {'lr': 0.00039312144752713885, 'samples': 3204096, 'steps': 16687, 'loss/train': 0.5313814580440521} 01/29/2022 10:17:12 - INFO - codeparrot_training - Step 16688: {'lr': 0.00039310803139864777, 'samples': 3204288, 'steps': 16688, 'loss/train': 1.7648212909698486} 01/29/2022 10:17:16 - INFO - codeparrot_training - Step 16689: {'lr': 0.00039309461465712725, 'samples': 3204480, 'steps': 16689, 'loss/train': 1.423804521560669} 01/29/2022 10:17:21 - INFO - codeparrot_training - Step 16690: {'lr': 0.00039308119730263494, 'samples': 3204672, 'steps': 16690, 'loss/train': 6.924987316131592} 01/29/2022 10:17:25 - INFO - codeparrot_training - Step 16691: {'lr': 0.00039306777933522806, 'samples': 3204864, 'steps': 16691, 'loss/train': 1.7045121788978577} 01/29/2022 10:17:30 - INFO - codeparrot_training - Step 16692: {'lr': 0.00039305436075496436, 'samples': 3205056, 'steps': 16692, 'loss/train': 1.4338832795619965} 01/29/2022 10:17:34 - INFO - codeparrot_training - Step 16693: {'lr': 0.0003930409415619012, 'samples': 3205248, 'steps': 16693, 'loss/train': 1.9428003430366516} 01/29/2022 10:17:38 - INFO - codeparrot_training - Step 16694: {'lr': 0.000393027521756096, 'samples': 3205440, 'steps': 16694, 'loss/train': 1.4055815041065216} 01/29/2022 10:17:45 - INFO - codeparrot_training - Step 16695: {'lr': 0.0003930141013376064, 'samples': 3205632, 'steps': 16695, 'loss/train': 1.1097658574581146} 01/29/2022 10:17:50 - INFO - codeparrot_training - Step 16696: {'lr': 0.00039300068030648976, 'samples': 3205824, 'steps': 16696, 'loss/train': 1.530295193195343} 01/29/2022 10:17:54 - INFO - codeparrot_training - Step 16697: {'lr': 0.0003929872586628036, 'samples': 3206016, 'steps': 16697, 'loss/train': 1.0138428211212158} 01/29/2022 10:17:58 - INFO - codeparrot_training - Step 16698: {'lr': 0.00039297383640660545, 'samples': 3206208, 'steps': 16698, 'loss/train': 2.6336479783058167} 01/29/2022 10:18:02 - INFO - codeparrot_training - Step 16699: {'lr': 0.0003929604135379528, 'samples': 3206400, 'steps': 16699, 'loss/train': 1.8164436221122742} 01/29/2022 10:18:08 - INFO - codeparrot_training - Step 16700: {'lr': 0.000392946990056903, 'samples': 3206592, 'steps': 16700, 'loss/train': 1.428510457277298} 01/29/2022 10:18:12 - INFO - codeparrot_training - Step 16701: {'lr': 0.0003929335659635139, 'samples': 3206784, 'steps': 16701, 'loss/train': 1.3749555945396423} 01/29/2022 10:18:16 - INFO - codeparrot_training - Step 16702: {'lr': 0.00039292014125784266, 'samples': 3206976, 'steps': 16702, 'loss/train': 2.1551334857940674} 01/29/2022 10:18:20 - INFO - codeparrot_training - Step 16703: {'lr': 0.00039290671593994697, 'samples': 3207168, 'steps': 16703, 'loss/train': 1.7282347083091736} 01/29/2022 10:18:25 - INFO - codeparrot_training - Step 16704: {'lr': 0.0003928932900098842, 'samples': 3207360, 'steps': 16704, 'loss/train': 1.7163630723953247} 01/29/2022 10:18:30 - INFO - codeparrot_training - Step 16705: {'lr': 0.00039287986346771205, 'samples': 3207552, 'steps': 16705, 'loss/train': 1.829207718372345} 01/29/2022 10:18:34 - INFO - codeparrot_training - Step 16706: {'lr': 0.0003928664363134879, 'samples': 3207744, 'steps': 16706, 'loss/train': 2.440627098083496} 01/29/2022 10:18:39 - INFO - codeparrot_training - Step 16707: {'lr': 0.00039285300854726926, 'samples': 3207936, 'steps': 16707, 'loss/train': 1.6729533076286316} 01/29/2022 10:18:43 - INFO - codeparrot_training - Step 16708: {'lr': 0.00039283958016911373, 'samples': 3208128, 'steps': 16708, 'loss/train': 1.7243367433547974} 01/29/2022 10:18:47 - INFO - codeparrot_training - Step 16709: {'lr': 0.00039282615117907884, 'samples': 3208320, 'steps': 16709, 'loss/train': 1.2922056019306183} 01/29/2022 10:18:53 - INFO - codeparrot_training - Step 16710: {'lr': 0.00039281272157722205, 'samples': 3208512, 'steps': 16710, 'loss/train': 2.1174795627593994} 01/29/2022 10:18:57 - INFO - codeparrot_training - Step 16711: {'lr': 0.0003927992913636008, 'samples': 3208704, 'steps': 16711, 'loss/train': 1.5526944994926453} 01/29/2022 10:19:01 - INFO - codeparrot_training - Step 16712: {'lr': 0.0003927858605382728, 'samples': 3208896, 'steps': 16712, 'loss/train': 1.655833125114441} 01/29/2022 10:19:05 - INFO - codeparrot_training - Step 16713: {'lr': 0.0003927724291012955, 'samples': 3209088, 'steps': 16713, 'loss/train': 2.166074216365814} 01/29/2022 10:19:10 - INFO - codeparrot_training - Step 16714: {'lr': 0.00039275899705272656, 'samples': 3209280, 'steps': 16714, 'loss/train': 1.4728783965110779} 01/29/2022 10:19:17 - INFO - codeparrot_training - Step 16715: {'lr': 0.00039274556439262325, 'samples': 3209472, 'steps': 16715, 'loss/train': 1.1795284152030945} 01/29/2022 10:19:21 - INFO - codeparrot_training - Step 16716: {'lr': 0.0003927321311210434, 'samples': 3209664, 'steps': 16716, 'loss/train': 1.3332219421863556} 01/29/2022 10:19:25 - INFO - codeparrot_training - Step 16717: {'lr': 0.00039271869723804434, 'samples': 3209856, 'steps': 16717, 'loss/train': 1.9755585193634033} 01/29/2022 10:19:29 - INFO - codeparrot_training - Step 16718: {'lr': 0.0003927052627436837, 'samples': 3210048, 'steps': 16718, 'loss/train': 1.4920679032802582} 01/29/2022 10:19:34 - INFO - codeparrot_training - Step 16719: {'lr': 0.000392691827638019, 'samples': 3210240, 'steps': 16719, 'loss/train': 1.4181400537490845} 01/29/2022 10:19:39 - INFO - codeparrot_training - Step 16720: {'lr': 0.000392678391921108, 'samples': 3210432, 'steps': 16720, 'loss/train': 1.7286717295646667} 01/29/2022 10:19:43 - INFO - codeparrot_training - Step 16721: {'lr': 0.00039266495559300786, 'samples': 3210624, 'steps': 16721, 'loss/train': 1.857630729675293} 01/29/2022 10:19:48 - INFO - codeparrot_training - Step 16722: {'lr': 0.00039265151865377644, 'samples': 3210816, 'steps': 16722, 'loss/train': 2.4338239431381226} 01/29/2022 10:19:52 - INFO - codeparrot_training - Step 16723: {'lr': 0.0003926380811034712, 'samples': 3211008, 'steps': 16723, 'loss/train': 1.858870267868042} 01/29/2022 10:19:56 - INFO - codeparrot_training - Step 16724: {'lr': 0.0003926246429421497, 'samples': 3211200, 'steps': 16724, 'loss/train': 0.8128192126750946} 01/29/2022 10:20:01 - INFO - codeparrot_training - Step 16725: {'lr': 0.0003926112041698696, 'samples': 3211392, 'steps': 16725, 'loss/train': 1.3447832465171814} 01/29/2022 10:20:06 - INFO - codeparrot_training - Step 16726: {'lr': 0.0003925977647866883, 'samples': 3211584, 'steps': 16726, 'loss/train': 1.6682960987091064} 01/29/2022 10:20:10 - INFO - codeparrot_training - Step 16727: {'lr': 0.0003925843247926635, 'samples': 3211776, 'steps': 16727, 'loss/train': 1.7838658690452576} 01/29/2022 10:20:14 - INFO - codeparrot_training - Step 16728: {'lr': 0.00039257088418785267, 'samples': 3211968, 'steps': 16728, 'loss/train': 2.6350929737091064} 01/29/2022 10:20:18 - INFO - codeparrot_training - Step 16729: {'lr': 0.00039255744297231354, 'samples': 3212160, 'steps': 16729, 'loss/train': 2.0924445390701294} 01/29/2022 10:20:25 - INFO - codeparrot_training - Step 16730: {'lr': 0.0003925440011461035, 'samples': 3212352, 'steps': 16730, 'loss/train': 1.6348124742507935} 01/29/2022 10:20:30 - INFO - codeparrot_training - Step 16731: {'lr': 0.0003925305587092802, 'samples': 3212544, 'steps': 16731, 'loss/train': 3.580124258995056} 01/29/2022 10:20:34 - INFO - codeparrot_training - Step 16732: {'lr': 0.00039251711566190133, 'samples': 3212736, 'steps': 16732, 'loss/train': 1.1283283531665802} 01/29/2022 10:20:38 - INFO - codeparrot_training - Step 16733: {'lr': 0.0003925036720040244, 'samples': 3212928, 'steps': 16733, 'loss/train': 2.404061436653137} 01/29/2022 10:20:42 - INFO - codeparrot_training - Step 16734: {'lr': 0.000392490227735707, 'samples': 3213120, 'steps': 16734, 'loss/train': 0.4268980175256729} 01/29/2022 10:20:48 - INFO - codeparrot_training - Step 16735: {'lr': 0.0003924767828570066, 'samples': 3213312, 'steps': 16735, 'loss/train': 1.5311803221702576} 01/29/2022 10:20:52 - INFO - codeparrot_training - Step 16736: {'lr': 0.00039246333736798095, 'samples': 3213504, 'steps': 16736, 'loss/train': 1.190729945898056} 01/29/2022 10:20:56 - INFO - codeparrot_training - Step 16737: {'lr': 0.00039244989126868755, 'samples': 3213696, 'steps': 16737, 'loss/train': 1.1265859007835388} 01/29/2022 10:21:00 - INFO - codeparrot_training - Step 16738: {'lr': 0.0003924364445591842, 'samples': 3213888, 'steps': 16738, 'loss/train': 1.8900291323661804} 01/29/2022 10:21:05 - INFO - codeparrot_training - Step 16739: {'lr': 0.0003924229972395282, 'samples': 3214080, 'steps': 16739, 'loss/train': 1.593159556388855} 01/29/2022 10:21:12 - INFO - codeparrot_training - Step 16740: {'lr': 0.00039240954930977744, 'samples': 3214272, 'steps': 16740, 'loss/train': 2.2298049330711365} 01/29/2022 10:21:16 - INFO - codeparrot_training - Step 16741: {'lr': 0.0003923961007699893, 'samples': 3214464, 'steps': 16741, 'loss/train': 2.050552010536194} 01/29/2022 10:21:20 - INFO - codeparrot_training - Step 16742: {'lr': 0.00039238265162022147, 'samples': 3214656, 'steps': 16742, 'loss/train': 1.368489146232605} 01/29/2022 10:21:24 - INFO - codeparrot_training - Step 16743: {'lr': 0.0003923692018605316, 'samples': 3214848, 'steps': 16743, 'loss/train': 1.0882146656513214} 01/29/2022 10:21:29 - INFO - codeparrot_training - Step 16744: {'lr': 0.0003923557514909773, 'samples': 3215040, 'steps': 16744, 'loss/train': 1.8256132006645203} 01/29/2022 10:21:34 - INFO - codeparrot_training - Step 16745: {'lr': 0.00039234230051161614, 'samples': 3215232, 'steps': 16745, 'loss/train': 1.1631099879741669} 01/29/2022 10:21:38 - INFO - codeparrot_training - Step 16746: {'lr': 0.00039232884892250575, 'samples': 3215424, 'steps': 16746, 'loss/train': 1.597777783870697} 01/29/2022 10:21:43 - INFO - codeparrot_training - Step 16747: {'lr': 0.00039231539672370376, 'samples': 3215616, 'steps': 16747, 'loss/train': 1.4560542404651642} 01/29/2022 10:21:47 - INFO - codeparrot_training - Step 16748: {'lr': 0.00039230194391526784, 'samples': 3215808, 'steps': 16748, 'loss/train': 0.45832282304763794} 01/29/2022 10:21:51 - INFO - codeparrot_training - Step 16749: {'lr': 0.0003922884904972556, 'samples': 3216000, 'steps': 16749, 'loss/train': 1.8821879625320435} 01/29/2022 10:21:56 - INFO - codeparrot_training - Step 16750: {'lr': 0.0003922750364697246, 'samples': 3216192, 'steps': 16750, 'loss/train': 1.3213786482810974} 01/29/2022 10:22:00 - INFO - codeparrot_training - Step 16751: {'lr': 0.0003922615818327325, 'samples': 3216384, 'steps': 16751, 'loss/train': 1.812800645828247} 01/29/2022 10:22:05 - INFO - codeparrot_training - Step 16752: {'lr': 0.000392248126586337, 'samples': 3216576, 'steps': 16752, 'loss/train': 1.0227026045322418} 01/29/2022 10:22:09 - INFO - codeparrot_training - Step 16753: {'lr': 0.0003922346707305957, 'samples': 3216768, 'steps': 16753, 'loss/train': 0.8907147645950317} 01/29/2022 10:22:13 - INFO - codeparrot_training - Step 16754: {'lr': 0.00039222121426556617, 'samples': 3216960, 'steps': 16754, 'loss/train': 1.4376622438430786} 01/29/2022 10:22:20 - INFO - codeparrot_training - Step 16755: {'lr': 0.0003922077571913062, 'samples': 3217152, 'steps': 16755, 'loss/train': 2.0117504596710205} 01/29/2022 10:22:25 - INFO - codeparrot_training - Step 16756: {'lr': 0.00039219429950787326, 'samples': 3217344, 'steps': 16756, 'loss/train': 1.190393328666687} 01/29/2022 10:22:29 - INFO - codeparrot_training - Step 16757: {'lr': 0.0003921808412153252, 'samples': 3217536, 'steps': 16757, 'loss/train': 2.286807060241699} 01/29/2022 10:22:33 - INFO - codeparrot_training - Step 16758: {'lr': 0.0003921673823137195, 'samples': 3217728, 'steps': 16758, 'loss/train': 2.056823194026947} 01/29/2022 10:22:37 - INFO - codeparrot_training - Step 16759: {'lr': 0.00039215392280311383, 'samples': 3217920, 'steps': 16759, 'loss/train': 1.1708970665931702} 01/29/2022 10:22:43 - INFO - codeparrot_training - Step 16760: {'lr': 0.000392140462683566, 'samples': 3218112, 'steps': 16760, 'loss/train': 2.254982829093933} 01/29/2022 10:22:47 - INFO - codeparrot_training - Step 16761: {'lr': 0.0003921270019551335, 'samples': 3218304, 'steps': 16761, 'loss/train': 1.2602201700210571} 01/29/2022 10:22:52 - INFO - codeparrot_training - Step 16762: {'lr': 0.00039211354061787407, 'samples': 3218496, 'steps': 16762, 'loss/train': 1.1845095455646515} 01/29/2022 10:22:56 - INFO - codeparrot_training - Step 16763: {'lr': 0.0003921000786718454, 'samples': 3218688, 'steps': 16763, 'loss/train': 0.9131118357181549} 01/29/2022 10:23:00 - INFO - codeparrot_training - Step 16764: {'lr': 0.0003920866161171051, 'samples': 3218880, 'steps': 16764, 'loss/train': 1.8787261247634888} 01/29/2022 10:23:06 - INFO - codeparrot_training - Step 16765: {'lr': 0.0003920731529537108, 'samples': 3219072, 'steps': 16765, 'loss/train': 1.9762558937072754} 01/29/2022 10:23:10 - INFO - codeparrot_training - Step 16766: {'lr': 0.00039205968918172026, 'samples': 3219264, 'steps': 16766, 'loss/train': 7.23966908454895} 01/29/2022 10:23:14 - INFO - codeparrot_training - Step 16767: {'lr': 0.00039204622480119107, 'samples': 3219456, 'steps': 16767, 'loss/train': 1.4298252761363983} 01/29/2022 10:23:18 - INFO - codeparrot_training - Step 16768: {'lr': 0.000392032759812181, 'samples': 3219648, 'steps': 16768, 'loss/train': 1.4308929741382599} 01/29/2022 10:23:23 - INFO - codeparrot_training - Step 16769: {'lr': 0.0003920192942147477, 'samples': 3219840, 'steps': 16769, 'loss/train': 1.9013716578483582} 01/29/2022 10:23:28 - INFO - codeparrot_training - Step 16770: {'lr': 0.00039200582800894885, 'samples': 3220032, 'steps': 16770, 'loss/train': 2.1282872557640076} 01/29/2022 10:23:32 - INFO - codeparrot_training - Step 16771: {'lr': 0.00039199236119484207, 'samples': 3220224, 'steps': 16771, 'loss/train': 1.3442273139953613} 01/29/2022 10:23:36 - INFO - codeparrot_training - Step 16772: {'lr': 0.0003919788937724852, 'samples': 3220416, 'steps': 16772, 'loss/train': 1.2134239375591278} 01/29/2022 10:23:41 - INFO - codeparrot_training - Step 16773: {'lr': 0.0003919654257419357, 'samples': 3220608, 'steps': 16773, 'loss/train': 1.8026313185691833} 01/29/2022 10:23:45 - INFO - codeparrot_training - Step 16774: {'lr': 0.0003919519571032515, 'samples': 3220800, 'steps': 16774, 'loss/train': 0.9512287974357605} 01/29/2022 10:23:52 - INFO - codeparrot_training - Step 16775: {'lr': 0.00039193848785649016, 'samples': 3220992, 'steps': 16775, 'loss/train': 1.614664614200592} 01/29/2022 10:23:56 - INFO - codeparrot_training - Step 16776: {'lr': 0.0003919250180017094, 'samples': 3221184, 'steps': 16776, 'loss/train': 2.0624531507492065} 01/29/2022 10:24:00 - INFO - codeparrot_training - Step 16777: {'lr': 0.00039191154753896696, 'samples': 3221376, 'steps': 16777, 'loss/train': 2.0077972412109375} 01/29/2022 10:24:05 - INFO - codeparrot_training - Step 16778: {'lr': 0.00039189807646832045, 'samples': 3221568, 'steps': 16778, 'loss/train': 1.1488702297210693} 01/29/2022 10:24:09 - INFO - codeparrot_training - Step 16779: {'lr': 0.0003918846047898277, 'samples': 3221760, 'steps': 16779, 'loss/train': 2.527200758457184} 01/29/2022 10:24:13 - INFO - codeparrot_training - Step 16780: {'lr': 0.00039187113250354635, 'samples': 3221952, 'steps': 16780, 'loss/train': 5.076360583305359} 01/29/2022 10:24:19 - INFO - codeparrot_training - Step 16781: {'lr': 0.00039185765960953405, 'samples': 3222144, 'steps': 16781, 'loss/train': 1.4458185732364655} 01/29/2022 10:24:23 - INFO - codeparrot_training - Step 16782: {'lr': 0.0003918441861078486, 'samples': 3222336, 'steps': 16782, 'loss/train': 2.0739499926567078} 01/29/2022 10:24:27 - INFO - codeparrot_training - Step 16783: {'lr': 0.0003918307119985477, 'samples': 3222528, 'steps': 16783, 'loss/train': 2.1993638277053833} 01/29/2022 10:24:32 - INFO - codeparrot_training - Step 16784: {'lr': 0.0003918172372816892, 'samples': 3222720, 'steps': 16784, 'loss/train': 1.2256551682949066} 01/29/2022 10:24:39 - INFO - codeparrot_training - Step 16785: {'lr': 0.0003918037619573305, 'samples': 3222912, 'steps': 16785, 'loss/train': 2.1855329275131226} 01/29/2022 10:24:43 - INFO - codeparrot_training - Step 16786: {'lr': 0.0003917902860255296, 'samples': 3223104, 'steps': 16786, 'loss/train': 2.130736291408539} 01/29/2022 10:24:48 - INFO - codeparrot_training - Step 16787: {'lr': 0.0003917768094863441, 'samples': 3223296, 'steps': 16787, 'loss/train': 1.0908532440662384} 01/29/2022 10:24:52 - INFO - codeparrot_training - Step 16788: {'lr': 0.00039176333233983187, 'samples': 3223488, 'steps': 16788, 'loss/train': 1.3883661925792694} 01/29/2022 10:24:56 - INFO - codeparrot_training - Step 16789: {'lr': 0.0003917498545860504, 'samples': 3223680, 'steps': 16789, 'loss/train': 1.2565718293190002} 01/29/2022 10:25:02 - INFO - codeparrot_training - Step 16790: {'lr': 0.0003917363762250576, 'samples': 3223872, 'steps': 16790, 'loss/train': 1.8349823355674744} 01/29/2022 10:25:06 - INFO - codeparrot_training - Step 16791: {'lr': 0.00039172289725691124, 'samples': 3224064, 'steps': 16791, 'loss/train': 2.082786798477173} 01/29/2022 10:25:10 - INFO - codeparrot_training - Step 16792: {'lr': 0.000391709417681669, 'samples': 3224256, 'steps': 16792, 'loss/train': 1.4368951320648193} 01/29/2022 10:25:14 - INFO - codeparrot_training - Step 16793: {'lr': 0.0003916959374993885, 'samples': 3224448, 'steps': 16793, 'loss/train': 0.8271666169166565} 01/29/2022 10:25:19 - INFO - codeparrot_training - Step 16794: {'lr': 0.0003916824567101277, 'samples': 3224640, 'steps': 16794, 'loss/train': 1.7285521030426025} 01/29/2022 10:25:23 - INFO - codeparrot_training - Step 16795: {'lr': 0.0003916689753139442, 'samples': 3224832, 'steps': 16795, 'loss/train': 1.7151201367378235} 01/29/2022 10:25:29 - INFO - codeparrot_training - Step 16796: {'lr': 0.0003916554933108958, 'samples': 3225024, 'steps': 16796, 'loss/train': 0.42726248502731323} 01/29/2022 10:25:33 - INFO - codeparrot_training - Step 16797: {'lr': 0.0003916420107010402, 'samples': 3225216, 'steps': 16797, 'loss/train': 0.798198401927948} 01/29/2022 10:25:37 - INFO - codeparrot_training - Step 16798: {'lr': 0.0003916285274844353, 'samples': 3225408, 'steps': 16798, 'loss/train': 1.3174801468849182} 01/29/2022 10:25:41 - INFO - codeparrot_training - Step 16799: {'lr': 0.0003916150436611387, 'samples': 3225600, 'steps': 16799, 'loss/train': 0.6889029443264008} 01/29/2022 10:25:46 - INFO - codeparrot_training - Step 16800: {'lr': 0.0003916015592312082, 'samples': 3225792, 'steps': 16800, 'loss/train': 2.8873722553253174} 01/29/2022 10:25:53 - INFO - codeparrot_training - Step 16801: {'lr': 0.00039158807419470166, 'samples': 3225984, 'steps': 16801, 'loss/train': 1.9231270551681519} 01/29/2022 10:25:57 - INFO - codeparrot_training - Step 16802: {'lr': 0.0003915745885516767, 'samples': 3226176, 'steps': 16802, 'loss/train': 2.3995789289474487} 01/29/2022 10:26:01 - INFO - codeparrot_training - Step 16803: {'lr': 0.0003915611023021912, 'samples': 3226368, 'steps': 16803, 'loss/train': 1.881272792816162} 01/29/2022 10:26:06 - INFO - codeparrot_training - Step 16804: {'lr': 0.00039154761544630287, 'samples': 3226560, 'steps': 16804, 'loss/train': 1.1365398466587067} 01/29/2022 10:26:10 - INFO - codeparrot_training - Step 16805: {'lr': 0.0003915341279840695, 'samples': 3226752, 'steps': 16805, 'loss/train': 1.3418397009372711} 01/29/2022 10:26:15 - INFO - codeparrot_training - Step 16806: {'lr': 0.00039152063991554885, 'samples': 3226944, 'steps': 16806, 'loss/train': 2.0765921473503113} 01/29/2022 10:26:20 - INFO - codeparrot_training - Step 16807: {'lr': 0.0003915071512407987, 'samples': 3227136, 'steps': 16807, 'loss/train': 1.7030991911888123} 01/29/2022 10:26:24 - INFO - codeparrot_training - Step 16808: {'lr': 0.0003914936619598769, 'samples': 3227328, 'steps': 16808, 'loss/train': 1.9241044521331787} 01/29/2022 10:26:28 - INFO - codeparrot_training - Step 16809: {'lr': 0.00039148017207284115, 'samples': 3227520, 'steps': 16809, 'loss/train': 1.4261896312236786} 01/29/2022 10:26:35 - INFO - codeparrot_training - Step 16810: {'lr': 0.0003914666815797493, 'samples': 3227712, 'steps': 16810, 'loss/train': 1.3895412683486938} 01/29/2022 10:26:39 - INFO - codeparrot_training - Step 16811: {'lr': 0.00039145319048065907, 'samples': 3227904, 'steps': 16811, 'loss/train': 1.6077625751495361} 01/29/2022 10:26:44 - INFO - codeparrot_training - Step 16812: {'lr': 0.00039143969877562833, 'samples': 3228096, 'steps': 16812, 'loss/train': 2.2827617526054382} 01/29/2022 10:26:48 - INFO - codeparrot_training - Step 16813: {'lr': 0.00039142620646471485, 'samples': 3228288, 'steps': 16813, 'loss/train': 0.5022507905960083} 01/29/2022 10:26:52 - INFO - codeparrot_training - Step 16814: {'lr': 0.00039141271354797635, 'samples': 3228480, 'steps': 16814, 'loss/train': 1.446347415447235} 01/29/2022 10:26:57 - INFO - codeparrot_training - Step 16815: {'lr': 0.0003913992200254707, 'samples': 3228672, 'steps': 16815, 'loss/train': 1.4697057008743286} 01/29/2022 10:27:02 - INFO - codeparrot_training - Step 16816: {'lr': 0.0003913857258972557, 'samples': 3228864, 'steps': 16816, 'loss/train': 1.559062421321869} 01/29/2022 10:27:06 - INFO - codeparrot_training - Step 16817: {'lr': 0.0003913722311633892, 'samples': 3229056, 'steps': 16817, 'loss/train': 1.520251750946045} 01/29/2022 10:27:10 - INFO - codeparrot_training - Step 16818: {'lr': 0.0003913587358239288, 'samples': 3229248, 'steps': 16818, 'loss/train': 1.2830083966255188} 01/29/2022 10:27:14 - INFO - codeparrot_training - Step 16819: {'lr': 0.0003913452398789326, 'samples': 3229440, 'steps': 16819, 'loss/train': 1.5846190452575684} 01/29/2022 10:27:20 - INFO - codeparrot_training - Step 16820: {'lr': 0.0003913317433284582, 'samples': 3229632, 'steps': 16820, 'loss/train': 1.6875311136245728} 01/29/2022 10:27:24 - INFO - codeparrot_training - Step 16821: {'lr': 0.00039131824617256354, 'samples': 3229824, 'steps': 16821, 'loss/train': 2.1663435101509094} 01/29/2022 10:27:28 - INFO - codeparrot_training - Step 16822: {'lr': 0.0003913047484113064, 'samples': 3230016, 'steps': 16822, 'loss/train': 0.6495552510023117} 01/29/2022 10:27:33 - INFO - codeparrot_training - Step 16823: {'lr': 0.0003912912500447445, 'samples': 3230208, 'steps': 16823, 'loss/train': 1.6469973921775818} 01/29/2022 10:27:37 - INFO - codeparrot_training - Step 16824: {'lr': 0.0003912777510729358, 'samples': 3230400, 'steps': 16824, 'loss/train': 1.4791227579116821} 01/29/2022 10:27:41 - INFO - codeparrot_training - Step 16825: {'lr': 0.0003912642514959381, 'samples': 3230592, 'steps': 16825, 'loss/train': 7.302022218704224} 01/29/2022 10:27:46 - INFO - codeparrot_training - Step 16826: {'lr': 0.00039125075131380923, 'samples': 3230784, 'steps': 16826, 'loss/train': 1.176868736743927} 01/29/2022 10:27:51 - INFO - codeparrot_training - Step 16827: {'lr': 0.00039123725052660696, 'samples': 3230976, 'steps': 16827, 'loss/train': 0.4770878702402115} 01/29/2022 10:27:55 - INFO - codeparrot_training - Step 16828: {'lr': 0.00039122374913438913, 'samples': 3231168, 'steps': 16828, 'loss/train': 1.2421563863754272} 01/29/2022 10:27:59 - INFO - codeparrot_training - Step 16829: {'lr': 0.00039121024713721365, 'samples': 3231360, 'steps': 16829, 'loss/train': 1.383282333612442} 01/29/2022 10:28:03 - INFO - codeparrot_training - Step 16830: {'lr': 0.0003911967445351382, 'samples': 3231552, 'steps': 16830, 'loss/train': 2.11133873462677} 01/29/2022 10:28:11 - INFO - codeparrot_training - Step 16831: {'lr': 0.00039118324132822083, 'samples': 3231744, 'steps': 16831, 'loss/train': 3.59945011138916} 01/29/2022 10:28:15 - INFO - codeparrot_training - Step 16832: {'lr': 0.0003911697375165193, 'samples': 3231936, 'steps': 16832, 'loss/train': 1.2727417051792145} 01/29/2022 10:28:20 - INFO - codeparrot_training - Step 16833: {'lr': 0.00039115623310009135, 'samples': 3232128, 'steps': 16833, 'loss/train': 1.9685204029083252} 01/29/2022 10:28:24 - INFO - codeparrot_training - Step 16834: {'lr': 0.00039114272807899496, 'samples': 3232320, 'steps': 16834, 'loss/train': 1.3831627070903778} 01/29/2022 10:28:28 - INFO - codeparrot_training - Step 16835: {'lr': 0.000391129222453288, 'samples': 3232512, 'steps': 16835, 'loss/train': 1.8371365070343018} 01/29/2022 10:28:33 - INFO - codeparrot_training - Step 16836: {'lr': 0.00039111571622302824, 'samples': 3232704, 'steps': 16836, 'loss/train': 1.4813019633293152} 01/29/2022 10:28:38 - INFO - codeparrot_training - Step 16837: {'lr': 0.0003911022093882736, 'samples': 3232896, 'steps': 16837, 'loss/train': 2.0372405648231506} 01/29/2022 10:28:42 - INFO - codeparrot_training - Step 16838: {'lr': 0.00039108870194908175, 'samples': 3233088, 'steps': 16838, 'loss/train': 2.149670362472534} 01/29/2022 10:28:46 - INFO - codeparrot_training - Step 16839: {'lr': 0.00039107519390551085, 'samples': 3233280, 'steps': 16839, 'loss/train': 1.4493505954742432} 01/29/2022 10:28:51 - INFO - codeparrot_training - Step 16840: {'lr': 0.00039106168525761855, 'samples': 3233472, 'steps': 16840, 'loss/train': 4.017752766609192} 01/29/2022 10:28:56 - INFO - codeparrot_training - Step 16841: {'lr': 0.00039104817600546277, 'samples': 3233664, 'steps': 16841, 'loss/train': 0.5524229407310486} 01/29/2022 10:29:00 - INFO - codeparrot_training - Step 16842: {'lr': 0.00039103466614910144, 'samples': 3233856, 'steps': 16842, 'loss/train': 1.4236180186271667} 01/29/2022 10:29:05 - INFO - codeparrot_training - Step 16843: {'lr': 0.0003910211556885923, 'samples': 3234048, 'steps': 16843, 'loss/train': 1.8760691285133362} 01/29/2022 10:29:09 - INFO - codeparrot_training - Step 16844: {'lr': 0.0003910076446239934, 'samples': 3234240, 'steps': 16844, 'loss/train': 2.2812328934669495} 01/29/2022 10:29:13 - INFO - codeparrot_training - Step 16845: {'lr': 0.00039099413295536246, 'samples': 3234432, 'steps': 16845, 'loss/train': 1.0943586230278015} 01/29/2022 10:29:20 - INFO - codeparrot_training - Step 16846: {'lr': 0.0003909806206827575, 'samples': 3234624, 'steps': 16846, 'loss/train': 2.3673110604286194} 01/29/2022 10:29:24 - INFO - codeparrot_training - Step 16847: {'lr': 0.00039096710780623625, 'samples': 3234816, 'steps': 16847, 'loss/train': 0.7997471988201141} 01/29/2022 10:29:29 - INFO - codeparrot_training - Step 16848: {'lr': 0.0003909535943258567, 'samples': 3235008, 'steps': 16848, 'loss/train': 1.3141856789588928} 01/29/2022 10:29:33 - INFO - codeparrot_training - Step 16849: {'lr': 0.0003909400802416767, 'samples': 3235200, 'steps': 16849, 'loss/train': 1.5106279850006104} 01/29/2022 10:29:37 - INFO - codeparrot_training - Step 16850: {'lr': 0.00039092656555375416, 'samples': 3235392, 'steps': 16850, 'loss/train': 2.182944595813751} 01/29/2022 10:29:43 - INFO - codeparrot_training - Step 16851: {'lr': 0.00039091305026214704, 'samples': 3235584, 'steps': 16851, 'loss/train': 2.0701661109924316} 01/29/2022 10:29:47 - INFO - codeparrot_training - Step 16852: {'lr': 0.0003908995343669131, 'samples': 3235776, 'steps': 16852, 'loss/train': 1.142134040594101} 01/29/2022 10:29:51 - INFO - codeparrot_training - Step 16853: {'lr': 0.0003908860178681102, 'samples': 3235968, 'steps': 16853, 'loss/train': 2.054054081439972} 01/29/2022 10:29:55 - INFO - codeparrot_training - Step 16854: {'lr': 0.0003908725007657964, 'samples': 3236160, 'steps': 16854, 'loss/train': 0.9485130608081818} 01/29/2022 10:30:00 - INFO - codeparrot_training - Step 16855: {'lr': 0.0003908589830600296, 'samples': 3236352, 'steps': 16855, 'loss/train': 1.2122511863708496} 01/29/2022 10:30:07 - INFO - codeparrot_training - Step 16856: {'lr': 0.0003908454647508676, 'samples': 3236544, 'steps': 16856, 'loss/train': 2.193904459476471} 01/29/2022 10:30:11 - INFO - codeparrot_training - Step 16857: {'lr': 0.00039083194583836836, 'samples': 3236736, 'steps': 16857, 'loss/train': 2.106600344181061} 01/29/2022 10:30:15 - INFO - codeparrot_training - Step 16858: {'lr': 0.0003908184263225898, 'samples': 3236928, 'steps': 16858, 'loss/train': 2.053637444972992} 01/29/2022 10:30:20 - INFO - codeparrot_training - Step 16859: {'lr': 0.0003908049062035898, 'samples': 3237120, 'steps': 16859, 'loss/train': 0.5545950829982758} 01/29/2022 10:30:24 - INFO - codeparrot_training - Step 16860: {'lr': 0.0003907913854814262, 'samples': 3237312, 'steps': 16860, 'loss/train': 1.6358912587165833} 01/29/2022 10:30:29 - INFO - codeparrot_training - Step 16861: {'lr': 0.00039077786415615714, 'samples': 3237504, 'steps': 16861, 'loss/train': 0.5158216059207916} 01/29/2022 10:30:33 - INFO - codeparrot_training - Step 16862: {'lr': 0.0003907643422278404, 'samples': 3237696, 'steps': 16862, 'loss/train': 2.0487354397773743} 01/29/2022 10:30:38 - INFO - codeparrot_training - Step 16863: {'lr': 0.00039075081969653383, 'samples': 3237888, 'steps': 16863, 'loss/train': 1.6445510387420654} 01/29/2022 10:30:42 - INFO - codeparrot_training - Step 16864: {'lr': 0.0003907372965622955, 'samples': 3238080, 'steps': 16864, 'loss/train': 2.1177163124084473} 01/29/2022 10:30:46 - INFO - codeparrot_training - Step 16865: {'lr': 0.0003907237728251833, 'samples': 3238272, 'steps': 16865, 'loss/train': 2.0712738633155823} 01/29/2022 10:30:51 - INFO - codeparrot_training - Step 16866: {'lr': 0.0003907102484852551, 'samples': 3238464, 'steps': 16866, 'loss/train': 1.4736966490745544} 01/29/2022 10:30:56 - INFO - codeparrot_training - Step 16867: {'lr': 0.0003906967235425689, 'samples': 3238656, 'steps': 16867, 'loss/train': 1.9972797632217407} 01/29/2022 10:31:00 - INFO - codeparrot_training - Step 16868: {'lr': 0.0003906831979971826, 'samples': 3238848, 'steps': 16868, 'loss/train': 2.120745062828064} 01/29/2022 10:31:04 - INFO - codeparrot_training - Step 16869: {'lr': 0.0003906696718491541, 'samples': 3239040, 'steps': 16869, 'loss/train': 1.172189176082611} 01/29/2022 10:31:08 - INFO - codeparrot_training - Step 16870: {'lr': 0.0003906561450985415, 'samples': 3239232, 'steps': 16870, 'loss/train': 0.08693616650998592} 01/29/2022 10:31:15 - INFO - codeparrot_training - Step 16871: {'lr': 0.00039064261774540254, 'samples': 3239424, 'steps': 16871, 'loss/train': 1.5757373571395874} 01/29/2022 10:31:20 - INFO - codeparrot_training - Step 16872: {'lr': 0.0003906290897897953, 'samples': 3239616, 'steps': 16872, 'loss/train': 1.9148274064064026} 01/29/2022 10:31:24 - INFO - codeparrot_training - Step 16873: {'lr': 0.00039061556123177777, 'samples': 3239808, 'steps': 16873, 'loss/train': 1.1914802491664886} 01/29/2022 10:31:28 - INFO - codeparrot_training - Step 16874: {'lr': 0.00039060203207140774, 'samples': 3240000, 'steps': 16874, 'loss/train': 2.1685529351234436} 01/29/2022 10:31:32 - INFO - codeparrot_training - Step 16875: {'lr': 0.0003905885023087433, 'samples': 3240192, 'steps': 16875, 'loss/train': 1.3816192746162415} 01/29/2022 10:31:38 - INFO - codeparrot_training - Step 16876: {'lr': 0.0003905749719438423, 'samples': 3240384, 'steps': 16876, 'loss/train': 0.773820698261261} 01/29/2022 10:31:42 - INFO - codeparrot_training - Step 16877: {'lr': 0.00039056144097676285, 'samples': 3240576, 'steps': 16877, 'loss/train': 2.3745937943458557} 01/29/2022 10:31:46 - INFO - codeparrot_training - Step 16878: {'lr': 0.0003905479094075627, 'samples': 3240768, 'steps': 16878, 'loss/train': 1.5710645914077759} 01/29/2022 10:31:50 - INFO - codeparrot_training - Step 16879: {'lr': 0.00039053437723630003, 'samples': 3240960, 'steps': 16879, 'loss/train': 1.7557395100593567} 01/29/2022 10:31:55 - INFO - codeparrot_training - Step 16880: {'lr': 0.00039052084446303264, 'samples': 3241152, 'steps': 16880, 'loss/train': 1.384920448064804} 01/29/2022 10:32:00 - INFO - codeparrot_training - Step 16881: {'lr': 0.0003905073110878186, 'samples': 3241344, 'steps': 16881, 'loss/train': 1.4886843860149384} 01/29/2022 10:32:04 - INFO - codeparrot_training - Step 16882: {'lr': 0.00039049377711071595, 'samples': 3241536, 'steps': 16882, 'loss/train': 1.1429522037506104} 01/29/2022 10:32:09 - INFO - codeparrot_training - Step 16883: {'lr': 0.00039048024253178243, 'samples': 3241728, 'steps': 16883, 'loss/train': 2.698486089706421} 01/29/2022 10:32:13 - INFO - codeparrot_training - Step 16884: {'lr': 0.00039046670735107627, 'samples': 3241920, 'steps': 16884, 'loss/train': 0.38007232546806335} 01/29/2022 10:32:17 - INFO - codeparrot_training - Step 16885: {'lr': 0.00039045317156865525, 'samples': 3242112, 'steps': 16885, 'loss/train': 1.526931881904602} 01/29/2022 10:32:24 - INFO - codeparrot_training - Step 16886: {'lr': 0.0003904396351845775, 'samples': 3242304, 'steps': 16886, 'loss/train': 2.3115177154541016} 01/29/2022 10:32:28 - INFO - codeparrot_training - Step 16887: {'lr': 0.00039042609819890087, 'samples': 3242496, 'steps': 16887, 'loss/train': 2.3251453042030334} 01/29/2022 10:32:33 - INFO - codeparrot_training - Step 16888: {'lr': 0.0003904125606116835, 'samples': 3242688, 'steps': 16888, 'loss/train': 0.5003494620323181} 01/29/2022 10:32:37 - INFO - codeparrot_training - Step 16889: {'lr': 0.0003903990224229833, 'samples': 3242880, 'steps': 16889, 'loss/train': 1.6927978992462158} 01/29/2022 10:32:41 - INFO - codeparrot_training - Step 16890: {'lr': 0.00039038548363285825, 'samples': 3243072, 'steps': 16890, 'loss/train': 1.4198605120182037} 01/29/2022 10:32:47 - INFO - codeparrot_training - Step 16891: {'lr': 0.00039037194424136634, 'samples': 3243264, 'steps': 16891, 'loss/train': 1.9560100436210632} 01/29/2022 10:32:51 - INFO - codeparrot_training - Step 16892: {'lr': 0.0003903584042485656, 'samples': 3243456, 'steps': 16892, 'loss/train': 1.095026582479477} 01/29/2022 10:32:55 - INFO - codeparrot_training - Step 16893: {'lr': 0.00039034486365451405, 'samples': 3243648, 'steps': 16893, 'loss/train': 1.3450468182563782} 01/29/2022 10:32:59 - INFO - codeparrot_training - Step 16894: {'lr': 0.00039033132245926974, 'samples': 3243840, 'steps': 16894, 'loss/train': 1.2156391739845276} 01/29/2022 10:33:04 - INFO - codeparrot_training - Step 16895: {'lr': 0.0003903177806628905, 'samples': 3244032, 'steps': 16895, 'loss/train': 1.9629474878311157} 01/29/2022 10:33:09 - INFO - codeparrot_training - Step 16896: {'lr': 0.00039030423826543446, 'samples': 3244224, 'steps': 16896, 'loss/train': 2.298588752746582} 01/29/2022 10:33:13 - INFO - codeparrot_training - Step 16897: {'lr': 0.0003902906952669596, 'samples': 3244416, 'steps': 16897, 'loss/train': 1.5079792141914368} 01/29/2022 10:33:17 - INFO - codeparrot_training - Step 16898: {'lr': 0.000390277151667524, 'samples': 3244608, 'steps': 16898, 'loss/train': 1.889365553855896} 01/29/2022 10:33:22 - INFO - codeparrot_training - Step 16899: {'lr': 0.0003902636074671856, 'samples': 3244800, 'steps': 16899, 'loss/train': 2.518427073955536} 01/29/2022 10:33:26 - INFO - codeparrot_training - Step 16900: {'lr': 0.0003902500626660025, 'samples': 3244992, 'steps': 16900, 'loss/train': 1.3101682662963867} 01/29/2022 10:33:31 - INFO - codeparrot_training - Step 16901: {'lr': 0.00039023651726403263, 'samples': 3245184, 'steps': 16901, 'loss/train': 1.7445237636566162} 01/29/2022 10:33:36 - INFO - codeparrot_training - Step 16902: {'lr': 0.00039022297126133397, 'samples': 3245376, 'steps': 16902, 'loss/train': 1.2098912000656128} 01/29/2022 10:33:40 - INFO - codeparrot_training - Step 16903: {'lr': 0.0003902094246579647, 'samples': 3245568, 'steps': 16903, 'loss/train': 1.5819200277328491} 01/29/2022 10:33:44 - INFO - codeparrot_training - Step 16904: {'lr': 0.00039019587745398276, 'samples': 3245760, 'steps': 16904, 'loss/train': 1.9720044136047363} 01/29/2022 10:33:48 - INFO - codeparrot_training - Step 16905: {'lr': 0.00039018232964944623, 'samples': 3245952, 'steps': 16905, 'loss/train': 1.9717631936073303} 01/29/2022 10:33:55 - INFO - codeparrot_training - Step 16906: {'lr': 0.0003901687812444131, 'samples': 3246144, 'steps': 16906, 'loss/train': 2.2453290224075317} 01/29/2022 10:34:00 - INFO - codeparrot_training - Step 16907: {'lr': 0.0003901552322389414, 'samples': 3246336, 'steps': 16907, 'loss/train': 2.35040420293808} 01/29/2022 10:34:04 - INFO - codeparrot_training - Step 16908: {'lr': 0.00039014168263308926, 'samples': 3246528, 'steps': 16908, 'loss/train': 1.2921063601970673} 01/29/2022 10:34:08 - INFO - codeparrot_training - Step 16909: {'lr': 0.00039012813242691454, 'samples': 3246720, 'steps': 16909, 'loss/train': 1.0171800255775452} 01/29/2022 10:34:12 - INFO - codeparrot_training - Step 16910: {'lr': 0.00039011458162047547, 'samples': 3246912, 'steps': 16910, 'loss/train': 2.1007373929023743} 01/29/2022 10:34:18 - INFO - codeparrot_training - Step 16911: {'lr': 0.00039010103021383, 'samples': 3247104, 'steps': 16911, 'loss/train': 2.132027506828308} 01/29/2022 10:34:22 - INFO - codeparrot_training - Step 16912: {'lr': 0.00039008747820703615, 'samples': 3247296, 'steps': 16912, 'loss/train': 0.9262654781341553} 01/29/2022 10:34:26 - INFO - codeparrot_training - Step 16913: {'lr': 0.0003900739256001521, 'samples': 3247488, 'steps': 16913, 'loss/train': 0.9576310515403748} 01/29/2022 10:34:30 - INFO - codeparrot_training - Step 16914: {'lr': 0.00039006037239323584, 'samples': 3247680, 'steps': 16914, 'loss/train': 1.8025948405265808} 01/29/2022 10:34:35 - INFO - codeparrot_training - Step 16915: {'lr': 0.00039004681858634537, 'samples': 3247872, 'steps': 16915, 'loss/train': 1.504253089427948} 01/29/2022 10:34:42 - INFO - codeparrot_training - Step 16916: {'lr': 0.0003900332641795388, 'samples': 3248064, 'steps': 16916, 'loss/train': 1.6010727882385254} 01/29/2022 10:34:46 - INFO - codeparrot_training - Step 16917: {'lr': 0.0003900197091728742, 'samples': 3248256, 'steps': 16917, 'loss/train': 1.7002810835838318} 01/29/2022 10:34:50 - INFO - codeparrot_training - Step 16918: {'lr': 0.0003900061535664097, 'samples': 3248448, 'steps': 16918, 'loss/train': 0.923406332731247} 01/29/2022 10:34:54 - INFO - codeparrot_training - Step 16919: {'lr': 0.0003899925973602032, 'samples': 3248640, 'steps': 16919, 'loss/train': 1.877615511417389} 01/29/2022 10:34:59 - INFO - codeparrot_training - Step 16920: {'lr': 0.0003899790405543129, 'samples': 3248832, 'steps': 16920, 'loss/train': 1.6260380744934082} 01/29/2022 10:35:04 - INFO - codeparrot_training - Step 16921: {'lr': 0.0003899654831487969, 'samples': 3249024, 'steps': 16921, 'loss/train': 1.7675861120224} 01/29/2022 10:35:08 - INFO - codeparrot_training - Step 16922: {'lr': 0.0003899519251437131, 'samples': 3249216, 'steps': 16922, 'loss/train': 1.3335223495960236} 01/29/2022 10:35:13 - INFO - codeparrot_training - Step 16923: {'lr': 0.00038993836653911974, 'samples': 3249408, 'steps': 16923, 'loss/train': 1.1609817445278168} 01/29/2022 10:35:17 - INFO - codeparrot_training - Step 16924: {'lr': 0.00038992480733507487, 'samples': 3249600, 'steps': 16924, 'loss/train': 1.4903208911418915} 01/29/2022 10:35:21 - INFO - codeparrot_training - Step 16925: {'lr': 0.0003899112475316365, 'samples': 3249792, 'steps': 16925, 'loss/train': 1.5406590700149536} 01/29/2022 10:35:26 - INFO - codeparrot_training - Step 16926: {'lr': 0.00038989768712886287, 'samples': 3249984, 'steps': 16926, 'loss/train': 1.2196505069732666} 01/29/2022 10:35:31 - INFO - codeparrot_training - Step 16927: {'lr': 0.0003898841261268119, 'samples': 3250176, 'steps': 16927, 'loss/train': 1.2672123312950134} 01/29/2022 10:35:35 - INFO - codeparrot_training - Step 16928: {'lr': 0.00038987056452554177, 'samples': 3250368, 'steps': 16928, 'loss/train': 1.6310505867004395} 01/29/2022 10:35:39 - INFO - codeparrot_training - Step 16929: {'lr': 0.00038985700232511055, 'samples': 3250560, 'steps': 16929, 'loss/train': 1.3364865481853485} 01/29/2022 10:35:43 - INFO - codeparrot_training - Step 16930: {'lr': 0.0003898434395255763, 'samples': 3250752, 'steps': 16930, 'loss/train': 2.0975895524024963} 01/29/2022 10:35:51 - INFO - codeparrot_training - Step 16931: {'lr': 0.0003898298761269973, 'samples': 3250944, 'steps': 16931, 'loss/train': 0.530364140868187} 01/29/2022 10:35:55 - INFO - codeparrot_training - Step 16932: {'lr': 0.0003898163121294314, 'samples': 3251136, 'steps': 16932, 'loss/train': 1.9303949475288391} 01/29/2022 10:35:59 - INFO - codeparrot_training - Step 16933: {'lr': 0.0003898027475329368, 'samples': 3251328, 'steps': 16933, 'loss/train': 1.9173890948295593} 01/29/2022 10:36:03 - INFO - codeparrot_training - Step 16934: {'lr': 0.00038978918233757167, 'samples': 3251520, 'steps': 16934, 'loss/train': 1.395615577697754} 01/29/2022 10:36:08 - INFO - codeparrot_training - Step 16935: {'lr': 0.000389775616543394, 'samples': 3251712, 'steps': 16935, 'loss/train': 0.22820591926574707} 01/29/2022 10:36:13 - INFO - codeparrot_training - Step 16936: {'lr': 0.00038976205015046206, 'samples': 3251904, 'steps': 16936, 'loss/train': 1.4508285820484161} 01/29/2022 10:36:17 - INFO - codeparrot_training - Step 16937: {'lr': 0.00038974848315883383, 'samples': 3252096, 'steps': 16937, 'loss/train': 2.5387948751449585} 01/29/2022 10:36:21 - INFO - codeparrot_training - Step 16938: {'lr': 0.00038973491556856755, 'samples': 3252288, 'steps': 16938, 'loss/train': 0.8217852115631104} 01/29/2022 10:36:26 - INFO - codeparrot_training - Step 16939: {'lr': 0.0003897213473797212, 'samples': 3252480, 'steps': 16939, 'loss/train': 1.6834226846694946} 01/29/2022 10:36:30 - INFO - codeparrot_training - Step 16940: {'lr': 0.0003897077785923529, 'samples': 3252672, 'steps': 16940, 'loss/train': 1.6542617082595825} 01/29/2022 10:36:37 - INFO - codeparrot_training - Step 16941: {'lr': 0.0003896942092065209, 'samples': 3252864, 'steps': 16941, 'loss/train': 2.2599082589149475} 01/29/2022 10:36:41 - INFO - codeparrot_training - Step 16942: {'lr': 0.0003896806392222833, 'samples': 3253056, 'steps': 16942, 'loss/train': 1.6453984379768372} 01/29/2022 10:36:45 - INFO - codeparrot_training - Step 16943: {'lr': 0.00038966706863969815, 'samples': 3253248, 'steps': 16943, 'loss/train': 1.954940915107727} 01/29/2022 10:36:50 - INFO - codeparrot_training - Step 16944: {'lr': 0.00038965349745882365, 'samples': 3253440, 'steps': 16944, 'loss/train': 1.100074052810669} 01/29/2022 10:36:54 - INFO - codeparrot_training - Step 16945: {'lr': 0.00038963992567971794, 'samples': 3253632, 'steps': 16945, 'loss/train': 0.4488758146762848} 01/29/2022 10:36:59 - INFO - codeparrot_training - Step 16946: {'lr': 0.0003896263533024391, 'samples': 3253824, 'steps': 16946, 'loss/train': 1.7938418984413147} 01/29/2022 10:37:03 - INFO - codeparrot_training - Step 16947: {'lr': 0.0003896127803270453, 'samples': 3254016, 'steps': 16947, 'loss/train': 2.3004550337791443} 01/29/2022 10:37:08 - INFO - codeparrot_training - Step 16948: {'lr': 0.0003895992067535946, 'samples': 3254208, 'steps': 16948, 'loss/train': 2.856915593147278} 01/29/2022 10:37:12 - INFO - codeparrot_training - Step 16949: {'lr': 0.0003895856325821454, 'samples': 3254400, 'steps': 16949, 'loss/train': 1.68781578540802} 01/29/2022 10:37:16 - INFO - codeparrot_training - Step 16950: {'lr': 0.00038957205781275554, 'samples': 3254592, 'steps': 16950, 'loss/train': 1.4263645112514496} 01/29/2022 10:37:22 - INFO - codeparrot_training - Step 16951: {'lr': 0.00038955848244548333, 'samples': 3254784, 'steps': 16951, 'loss/train': 0.9986209273338318} 01/29/2022 10:37:26 - INFO - codeparrot_training - Step 16952: {'lr': 0.00038954490648038687, 'samples': 3254976, 'steps': 16952, 'loss/train': 1.3674576580524445} 01/29/2022 10:37:30 - INFO - codeparrot_training - Step 16953: {'lr': 0.0003895313299175244, 'samples': 3255168, 'steps': 16953, 'loss/train': 2.1667327880859375} 01/29/2022 10:37:34 - INFO - codeparrot_training - Step 16954: {'lr': 0.000389517752756954, 'samples': 3255360, 'steps': 16954, 'loss/train': 0.3191409409046173} 01/29/2022 10:37:39 - INFO - codeparrot_training - Step 16955: {'lr': 0.0003895041749987338, 'samples': 3255552, 'steps': 16955, 'loss/train': 1.1602347493171692} 01/29/2022 10:37:43 - INFO - codeparrot_training - Step 16956: {'lr': 0.00038949059664292207, 'samples': 3255744, 'steps': 16956, 'loss/train': 1.1860148012638092} 01/29/2022 10:37:48 - INFO - codeparrot_training - Step 16957: {'lr': 0.0003894770176895769, 'samples': 3255936, 'steps': 16957, 'loss/train': 0.6989329308271408} 01/29/2022 10:37:52 - INFO - codeparrot_training - Step 16958: {'lr': 0.0003894634381387565, 'samples': 3256128, 'steps': 16958, 'loss/train': 1.865104079246521} 01/29/2022 10:37:57 - INFO - codeparrot_training - Step 16959: {'lr': 0.00038944985799051896, 'samples': 3256320, 'steps': 16959, 'loss/train': 1.199327677488327} 01/29/2022 10:38:01 - INFO - codeparrot_training - Step 16960: {'lr': 0.0003894362772449226, 'samples': 3256512, 'steps': 16960, 'loss/train': 1.9214344024658203} 01/29/2022 10:38:05 - INFO - codeparrot_training - Step 16961: {'lr': 0.0003894226959020254, 'samples': 3256704, 'steps': 16961, 'loss/train': 1.10103839635849} 01/29/2022 10:38:12 - INFO - codeparrot_training - Step 16962: {'lr': 0.00038940911396188573, 'samples': 3256896, 'steps': 16962, 'loss/train': 2.1269994378089905} 01/29/2022 10:38:17 - INFO - codeparrot_training - Step 16963: {'lr': 0.0003893955314245616, 'samples': 3257088, 'steps': 16963, 'loss/train': 1.8476784825325012} 01/29/2022 10:38:21 - INFO - codeparrot_training - Step 16964: {'lr': 0.0003893819482901113, 'samples': 3257280, 'steps': 16964, 'loss/train': 1.95977783203125} 01/29/2022 10:38:25 - INFO - codeparrot_training - Step 16965: {'lr': 0.000389368364558593, 'samples': 3257472, 'steps': 16965, 'loss/train': 1.5598037838935852} 01/29/2022 10:38:29 - INFO - codeparrot_training - Step 16966: {'lr': 0.00038935478023006487, 'samples': 3257664, 'steps': 16966, 'loss/train': 1.4113229513168335} 01/29/2022 10:38:34 - INFO - codeparrot_training - Step 16967: {'lr': 0.0003893411953045852, 'samples': 3257856, 'steps': 16967, 'loss/train': 1.498142123222351} 01/29/2022 10:38:39 - INFO - codeparrot_training - Step 16968: {'lr': 0.000389327609782212, 'samples': 3258048, 'steps': 16968, 'loss/train': 1.8317508101463318} 01/29/2022 10:38:43 - INFO - codeparrot_training - Step 16969: {'lr': 0.0003893140236630036, 'samples': 3258240, 'steps': 16969, 'loss/train': 1.7274999618530273} 01/29/2022 10:38:47 - INFO - codeparrot_training - Step 16970: {'lr': 0.0003893004369470181, 'samples': 3258432, 'steps': 16970, 'loss/train': 2.1565096378326416} 01/29/2022 10:38:51 - INFO - codeparrot_training - Step 16971: {'lr': 0.00038928684963431383, 'samples': 3258624, 'steps': 16971, 'loss/train': 1.4377966225147247} 01/29/2022 10:38:57 - INFO - codeparrot_training - Step 16972: {'lr': 0.00038927326172494894, 'samples': 3258816, 'steps': 16972, 'loss/train': 1.2625859677791595} 01/29/2022 10:39:01 - INFO - codeparrot_training - Step 16973: {'lr': 0.0003892596732189816, 'samples': 3259008, 'steps': 16973, 'loss/train': 2.820018947124481} 01/29/2022 10:39:05 - INFO - codeparrot_training - Step 16974: {'lr': 0.00038924608411647, 'samples': 3259200, 'steps': 16974, 'loss/train': 1.6788209080696106} 01/29/2022 10:39:09 - INFO - codeparrot_training - Step 16975: {'lr': 0.00038923249441747245, 'samples': 3259392, 'steps': 16975, 'loss/train': 1.7066486477851868} 01/29/2022 10:39:14 - INFO - codeparrot_training - Step 16976: {'lr': 0.000389218904122047, 'samples': 3259584, 'steps': 16976, 'loss/train': 1.4383661448955536} 01/29/2022 10:39:21 - INFO - codeparrot_training - Step 16977: {'lr': 0.00038920531323025206, 'samples': 3259776, 'steps': 16977, 'loss/train': 1.3712151646614075} 01/29/2022 10:39:25 - INFO - codeparrot_training - Step 16978: {'lr': 0.0003891917217421458, 'samples': 3259968, 'steps': 16978, 'loss/train': 2.15572053194046} 01/29/2022 10:39:29 - INFO - codeparrot_training - Step 16979: {'lr': 0.00038917812965778625, 'samples': 3260160, 'steps': 16979, 'loss/train': 0.8222602307796478} 01/29/2022 10:39:34 - INFO - codeparrot_training - Step 16980: {'lr': 0.00038916453697723194, 'samples': 3260352, 'steps': 16980, 'loss/train': 1.3681604862213135} 01/29/2022 10:39:38 - INFO - codeparrot_training - Step 16981: {'lr': 0.00038915094370054083, 'samples': 3260544, 'steps': 16981, 'loss/train': 1.7265561819076538} 01/29/2022 10:39:44 - INFO - codeparrot_training - Step 16982: {'lr': 0.00038913734982777136, 'samples': 3260736, 'steps': 16982, 'loss/train': 1.0276813209056854} 01/29/2022 10:39:48 - INFO - codeparrot_training - Step 16983: {'lr': 0.0003891237553589816, 'samples': 3260928, 'steps': 16983, 'loss/train': 1.6327714920043945} 01/29/2022 10:39:52 - INFO - codeparrot_training - Step 16984: {'lr': 0.00038911016029422984, 'samples': 3261120, 'steps': 16984, 'loss/train': 0.7537458837032318} 01/29/2022 10:39:56 - INFO - codeparrot_training - Step 16985: {'lr': 0.0003890965646335744, 'samples': 3261312, 'steps': 16985, 'loss/train': 1.378136157989502} 01/29/2022 10:40:03 - INFO - codeparrot_training - Step 16986: {'lr': 0.0003890829683770734, 'samples': 3261504, 'steps': 16986, 'loss/train': 1.2897769510746002} 01/29/2022 10:40:08 - INFO - codeparrot_training - Step 16987: {'lr': 0.0003890693715247851, 'samples': 3261696, 'steps': 16987, 'loss/train': 1.8441957235336304} 01/29/2022 10:40:12 - INFO - codeparrot_training - Step 16988: {'lr': 0.0003890557740767678, 'samples': 3261888, 'steps': 16988, 'loss/train': 1.5633094310760498} 01/29/2022 10:40:16 - INFO - codeparrot_training - Step 16989: {'lr': 0.0003890421760330798, 'samples': 3262080, 'steps': 16989, 'loss/train': 1.408099740743637} 01/29/2022 10:40:20 - INFO - codeparrot_training - Step 16990: {'lr': 0.0003890285773937792, 'samples': 3262272, 'steps': 16990, 'loss/train': 1.0532081723213196} 01/29/2022 10:40:26 - INFO - codeparrot_training - Step 16991: {'lr': 0.0003890149781589243, 'samples': 3262464, 'steps': 16991, 'loss/train': 1.785430669784546} 01/29/2022 10:40:30 - INFO - codeparrot_training - Step 16992: {'lr': 0.0003890013783285733, 'samples': 3262656, 'steps': 16992, 'loss/train': 1.8718560934066772} 01/29/2022 10:40:34 - INFO - codeparrot_training - Step 16993: {'lr': 0.00038898777790278465, 'samples': 3262848, 'steps': 16993, 'loss/train': 0.9435406029224396} 01/29/2022 10:40:38 - INFO - codeparrot_training - Step 16994: {'lr': 0.00038897417688161644, 'samples': 3263040, 'steps': 16994, 'loss/train': 1.4047889113426208} 01/29/2022 10:40:43 - INFO - codeparrot_training - Step 16995: {'lr': 0.0003889605752651271, 'samples': 3263232, 'steps': 16995, 'loss/train': 2.1811829209327698} 01/29/2022 10:40:48 - INFO - codeparrot_training - Step 16996: {'lr': 0.0003889469730533746, 'samples': 3263424, 'steps': 16996, 'loss/train': 1.1752880215644836} 01/29/2022 10:40:53 - INFO - codeparrot_training - Step 16997: {'lr': 0.0003889333702464175, 'samples': 3263616, 'steps': 16997, 'loss/train': 2.0239904522895813} 01/29/2022 10:40:57 - INFO - codeparrot_training - Step 16998: {'lr': 0.00038891976684431395, 'samples': 3263808, 'steps': 16998, 'loss/train': 1.4607873558998108} 01/29/2022 10:41:01 - INFO - codeparrot_training - Step 16999: {'lr': 0.0003889061628471222, 'samples': 3264000, 'steps': 16999, 'loss/train': 1.6753495931625366} 01/29/2022 10:41:05 - INFO - codeparrot_training - Step 17000: {'lr': 0.00038889255825490053, 'samples': 3264192, 'steps': 17000, 'loss/train': 1.8445531725883484} 01/29/2022 10:41:09 - INFO - codeparrot_training - Step 17001: {'lr': 0.0003888789530677073, 'samples': 3264384, 'steps': 17001, 'loss/train': 2.016255497932434} 01/29/2022 10:41:15 - INFO - codeparrot_training - Step 17002: {'lr': 0.00038886534728560073, 'samples': 3264576, 'steps': 17002, 'loss/train': 1.7544350624084473} 01/29/2022 10:41:19 - INFO - codeparrot_training - Step 17003: {'lr': 0.0003888517409086391, 'samples': 3264768, 'steps': 17003, 'loss/train': 1.734588861465454} 01/29/2022 10:41:23 - INFO - codeparrot_training - Step 17004: {'lr': 0.0003888381339368807, 'samples': 3264960, 'steps': 17004, 'loss/train': 2.2083443999290466} 01/29/2022 10:41:28 - INFO - codeparrot_training - Step 17005: {'lr': 0.00038882452637038377, 'samples': 3265152, 'steps': 17005, 'loss/train': 1.971737265586853} 01/29/2022 10:41:32 - INFO - codeparrot_training - Step 17006: {'lr': 0.00038881091820920676, 'samples': 3265344, 'steps': 17006, 'loss/train': 1.984837532043457} 01/29/2022 10:41:39 - INFO - codeparrot_training - Step 17007: {'lr': 0.00038879730945340775, 'samples': 3265536, 'steps': 17007, 'loss/train': 1.7094889283180237} 01/29/2022 10:41:43 - INFO - codeparrot_training - Step 17008: {'lr': 0.0003887837001030452, 'samples': 3265728, 'steps': 17008, 'loss/train': 1.2798450887203217} 01/29/2022 10:41:48 - INFO - codeparrot_training - Step 17009: {'lr': 0.00038877009015817734, 'samples': 3265920, 'steps': 17009, 'loss/train': 0.8648926913738251} 01/29/2022 10:41:52 - INFO - codeparrot_training - Step 17010: {'lr': 0.0003887564796188625, 'samples': 3266112, 'steps': 17010, 'loss/train': 2.4699520468711853} 01/29/2022 10:41:56 - INFO - codeparrot_training - Step 17011: {'lr': 0.0003887428684851589, 'samples': 3266304, 'steps': 17011, 'loss/train': 1.1986410319805145} 01/29/2022 10:42:01 - INFO - codeparrot_training - Step 17012: {'lr': 0.00038872925675712493, 'samples': 3266496, 'steps': 17012, 'loss/train': 2.2809805870056152} 01/29/2022 10:42:06 - INFO - codeparrot_training - Step 17013: {'lr': 0.00038871564443481886, 'samples': 3266688, 'steps': 17013, 'loss/train': 1.46668741106987} 01/29/2022 10:42:10 - INFO - codeparrot_training - Step 17014: {'lr': 0.0003887020315182991, 'samples': 3266880, 'steps': 17014, 'loss/train': 1.5152361989021301} 01/29/2022 10:42:14 - INFO - codeparrot_training - Step 17015: {'lr': 0.0003886884180076238, 'samples': 3267072, 'steps': 17015, 'loss/train': 1.4418732225894928} 01/29/2022 10:42:18 - INFO - codeparrot_training - Step 17016: {'lr': 0.0003886748039028514, 'samples': 3267264, 'steps': 17016, 'loss/train': 1.9649439454078674} 01/29/2022 10:42:24 - INFO - codeparrot_training - Step 17017: {'lr': 0.00038866118920404013, 'samples': 3267456, 'steps': 17017, 'loss/train': 1.7717199325561523} 01/29/2022 10:42:28 - INFO - codeparrot_training - Step 17018: {'lr': 0.0003886475739112484, 'samples': 3267648, 'steps': 17018, 'loss/train': 1.5782938599586487} 01/29/2022 10:42:32 - INFO - codeparrot_training - Step 17019: {'lr': 0.0003886339580245344, 'samples': 3267840, 'steps': 17019, 'loss/train': 2.0382909178733826} 01/29/2022 10:42:36 - INFO - codeparrot_training - Step 17020: {'lr': 0.00038862034154395664, 'samples': 3268032, 'steps': 17020, 'loss/train': 0.5836121588945389} 01/29/2022 10:42:41 - INFO - codeparrot_training - Step 17021: {'lr': 0.00038860672446957336, 'samples': 3268224, 'steps': 17021, 'loss/train': 1.5100160837173462} 01/29/2022 10:42:48 - INFO - codeparrot_training - Step 17022: {'lr': 0.00038859310680144276, 'samples': 3268416, 'steps': 17022, 'loss/train': 1.4003418982028961} 01/29/2022 10:42:52 - INFO - codeparrot_training - Step 17023: {'lr': 0.0003885794885396234, 'samples': 3268608, 'steps': 17023, 'loss/train': 1.183697909116745} 01/29/2022 10:42:56 - INFO - codeparrot_training - Step 17024: {'lr': 0.00038856586968417353, 'samples': 3268800, 'steps': 17024, 'loss/train': 1.3994328081607819} 01/29/2022 10:43:00 - INFO - codeparrot_training - Step 17025: {'lr': 0.0003885522502351514, 'samples': 3268992, 'steps': 17025, 'loss/train': 1.0920743644237518} 01/29/2022 10:43:05 - INFO - codeparrot_training - Step 17026: {'lr': 0.0003885386301926155, 'samples': 3269184, 'steps': 17026, 'loss/train': 1.729607105255127} 01/29/2022 10:43:10 - INFO - codeparrot_training - Step 17027: {'lr': 0.00038852500955662407, 'samples': 3269376, 'steps': 17027, 'loss/train': 0.685339629650116} 01/29/2022 10:43:14 - INFO - codeparrot_training - Step 17028: {'lr': 0.0003885113883272355, 'samples': 3269568, 'steps': 17028, 'loss/train': 2.1666696667671204} 01/29/2022 10:43:18 - INFO - codeparrot_training - Step 17029: {'lr': 0.0003884977665045081, 'samples': 3269760, 'steps': 17029, 'loss/train': 1.063511073589325} 01/29/2022 10:43:23 - INFO - codeparrot_training - Step 17030: {'lr': 0.0003884841440885003, 'samples': 3269952, 'steps': 17030, 'loss/train': 1.9176045656204224} 01/29/2022 10:43:27 - INFO - codeparrot_training - Step 17031: {'lr': 0.0003884705210792703, 'samples': 3270144, 'steps': 17031, 'loss/train': 2.027297079563141} 01/29/2022 10:43:35 - INFO - codeparrot_training - Step 17032: {'lr': 0.00038845689747687664, 'samples': 3270336, 'steps': 17032, 'loss/train': 1.4867211878299713} 01/29/2022 10:43:39 - INFO - codeparrot_training - Step 17033: {'lr': 0.0003884432732813776, 'samples': 3270528, 'steps': 17033, 'loss/train': 1.9254317879676819} 01/29/2022 10:43:43 - INFO - codeparrot_training - Step 17034: {'lr': 0.00038842964849283146, 'samples': 3270720, 'steps': 17034, 'loss/train': 1.6611778736114502} 01/29/2022 10:43:47 - INFO - codeparrot_training - Step 17035: {'lr': 0.0003884160231112968, 'samples': 3270912, 'steps': 17035, 'loss/train': 1.8683028817176819} 01/29/2022 10:43:52 - INFO - codeparrot_training - Step 17036: {'lr': 0.00038840239713683165, 'samples': 3271104, 'steps': 17036, 'loss/train': 2.2577818036079407} 01/29/2022 10:43:57 - INFO - codeparrot_training - Step 17037: {'lr': 0.00038838877056949475, 'samples': 3271296, 'steps': 17037, 'loss/train': 2.041619896888733} 01/29/2022 10:44:02 - INFO - codeparrot_training - Step 17038: {'lr': 0.00038837514340934424, 'samples': 3271488, 'steps': 17038, 'loss/train': 1.7643722891807556} 01/29/2022 10:44:06 - INFO - codeparrot_training - Step 17039: {'lr': 0.0003883615156564385, 'samples': 3271680, 'steps': 17039, 'loss/train': 2.7068424224853516} 01/29/2022 10:44:10 - INFO - codeparrot_training - Step 17040: {'lr': 0.000388347887310836, 'samples': 3271872, 'steps': 17040, 'loss/train': 1.644143521785736} 01/29/2022 10:44:14 - INFO - codeparrot_training - Step 17041: {'lr': 0.0003883342583725952, 'samples': 3272064, 'steps': 17041, 'loss/train': 1.786518394947052} 01/29/2022 10:44:20 - INFO - codeparrot_training - Step 17042: {'lr': 0.0003883206288417742, 'samples': 3272256, 'steps': 17042, 'loss/train': 1.7815364599227905} 01/29/2022 10:44:24 - INFO - codeparrot_training - Step 17043: {'lr': 0.0003883069987184316, 'samples': 3272448, 'steps': 17043, 'loss/train': 1.901073932647705} 01/29/2022 10:44:28 - INFO - codeparrot_training - Step 17044: {'lr': 0.0003882933680026257, 'samples': 3272640, 'steps': 17044, 'loss/train': 2.0647534132003784} 01/29/2022 10:44:32 - INFO - codeparrot_training - Step 17045: {'lr': 0.000388279736694415, 'samples': 3272832, 'steps': 17045, 'loss/train': 1.5754947066307068} 01/29/2022 10:44:37 - INFO - codeparrot_training - Step 17046: {'lr': 0.00038826610479385774, 'samples': 3273024, 'steps': 17046, 'loss/train': 1.0447481274604797} 01/29/2022 10:44:44 - INFO - codeparrot_training - Step 17047: {'lr': 0.00038825247230101244, 'samples': 3273216, 'steps': 17047, 'loss/train': 2.002388119697571} 01/29/2022 10:44:48 - INFO - codeparrot_training - Step 17048: {'lr': 0.0003882388392159375, 'samples': 3273408, 'steps': 17048, 'loss/train': 0.90829136967659} 01/29/2022 10:44:52 - INFO - codeparrot_training - Step 17049: {'lr': 0.0003882252055386912, 'samples': 3273600, 'steps': 17049, 'loss/train': 1.7133808135986328} 01/29/2022 10:44:57 - INFO - codeparrot_training - Step 17050: {'lr': 0.00038821157126933204, 'samples': 3273792, 'steps': 17050, 'loss/train': 1.6922038793563843} 01/29/2022 10:45:01 - INFO - codeparrot_training - Step 17051: {'lr': 0.00038819793640791834, 'samples': 3273984, 'steps': 17051, 'loss/train': 1.5073299407958984} 01/29/2022 10:45:06 - INFO - codeparrot_training - Step 17052: {'lr': 0.0003881843009545086, 'samples': 3274176, 'steps': 17052, 'loss/train': 1.1262216567993164} 01/29/2022 10:45:11 - INFO - codeparrot_training - Step 17053: {'lr': 0.0003881706649091612, 'samples': 3274368, 'steps': 17053, 'loss/train': 2.0338107347488403} 01/29/2022 10:45:15 - INFO - codeparrot_training - Step 17054: {'lr': 0.0003881570282719346, 'samples': 3274560, 'steps': 17054, 'loss/train': 1.5922465324401855} 01/29/2022 10:45:19 - INFO - codeparrot_training - Step 17055: {'lr': 0.00038814339104288706, 'samples': 3274752, 'steps': 17055, 'loss/train': 2.028695046901703} 01/29/2022 10:45:24 - INFO - codeparrot_training - Step 17056: {'lr': 0.00038812975322207713, 'samples': 3274944, 'steps': 17056, 'loss/train': 1.5385844707489014} 01/29/2022 10:45:29 - INFO - codeparrot_training - Step 17057: {'lr': 0.0003881161148095632, 'samples': 3275136, 'steps': 17057, 'loss/train': 1.9096519947052002} 01/29/2022 10:45:33 - INFO - codeparrot_training - Step 17058: {'lr': 0.0003881024758054037, 'samples': 3275328, 'steps': 17058, 'loss/train': 1.8184032440185547} 01/29/2022 10:45:38 - INFO - codeparrot_training - Step 17059: {'lr': 0.00038808883620965705, 'samples': 3275520, 'steps': 17059, 'loss/train': 0.9890677034854889} 01/29/2022 10:45:42 - INFO - codeparrot_training - Step 17060: {'lr': 0.00038807519602238174, 'samples': 3275712, 'steps': 17060, 'loss/train': 5.871440291404724} 01/29/2022 10:45:46 - INFO - codeparrot_training - Step 17061: {'lr': 0.00038806155524363594, 'samples': 3275904, 'steps': 17061, 'loss/train': 1.3420734107494354} 01/29/2022 10:45:53 - INFO - codeparrot_training - Step 17062: {'lr': 0.00038804791387347844, 'samples': 3276096, 'steps': 17062, 'loss/train': 1.5198719501495361} 01/29/2022 10:45:57 - INFO - codeparrot_training - Step 17063: {'lr': 0.0003880342719119675, 'samples': 3276288, 'steps': 17063, 'loss/train': 2.5478153228759766} 01/29/2022 10:46:02 - INFO - codeparrot_training - Step 17064: {'lr': 0.0003880206293591615, 'samples': 3276480, 'steps': 17064, 'loss/train': 0.8977796137332916} 01/29/2022 10:46:06 - INFO - codeparrot_training - Step 17065: {'lr': 0.000388006986215119, 'samples': 3276672, 'steps': 17065, 'loss/train': 1.5273898243904114} 01/29/2022 10:46:10 - INFO - codeparrot_training - Step 17066: {'lr': 0.0003879933424798984, 'samples': 3276864, 'steps': 17066, 'loss/train': 1.438490241765976} 01/29/2022 10:46:16 - INFO - codeparrot_training - Step 17067: {'lr': 0.0003879796981535582, 'samples': 3277056, 'steps': 17067, 'loss/train': 1.6188783645629883} 01/29/2022 10:46:20 - INFO - codeparrot_training - Step 17068: {'lr': 0.00038796605323615664, 'samples': 3277248, 'steps': 17068, 'loss/train': 2.037006676197052} 01/29/2022 10:46:24 - INFO - codeparrot_training - Step 17069: {'lr': 0.00038795240772775244, 'samples': 3277440, 'steps': 17069, 'loss/train': 1.9682471752166748} 01/29/2022 10:46:28 - INFO - codeparrot_training - Step 17070: {'lr': 0.0003879387616284038, 'samples': 3277632, 'steps': 17070, 'loss/train': 1.3872532546520233} 01/29/2022 10:46:33 - INFO - codeparrot_training - Step 17071: {'lr': 0.0003879251149381694, 'samples': 3277824, 'steps': 17071, 'loss/train': 1.6037406921386719} 01/29/2022 10:46:38 - INFO - codeparrot_training - Step 17072: {'lr': 0.0003879114676571076, 'samples': 3278016, 'steps': 17072, 'loss/train': 1.2712989449501038} 01/29/2022 10:46:42 - INFO - codeparrot_training - Step 17073: {'lr': 0.00038789781978527683, 'samples': 3278208, 'steps': 17073, 'loss/train': 1.1852496564388275} 01/29/2022 10:46:46 - INFO - codeparrot_training - Step 17074: {'lr': 0.0003878841713227356, 'samples': 3278400, 'steps': 17074, 'loss/train': 2.2409257292747498} 01/29/2022 10:46:51 - INFO - codeparrot_training - Step 17075: {'lr': 0.00038787052226954235, 'samples': 3278592, 'steps': 17075, 'loss/train': 2.1591535806655884} 01/29/2022 10:46:55 - INFO - codeparrot_training - Step 17076: {'lr': 0.0003878568726257556, 'samples': 3278784, 'steps': 17076, 'loss/train': 1.7655499577522278} 01/29/2022 10:47:00 - INFO - codeparrot_training - Step 17077: {'lr': 0.0003878432223914338, 'samples': 3278976, 'steps': 17077, 'loss/train': 0.8191031813621521} 01/29/2022 10:47:04 - INFO - codeparrot_training - Step 17078: {'lr': 0.00038782957156663535, 'samples': 3279168, 'steps': 17078, 'loss/train': 2.5404621362686157} 01/29/2022 10:47:09 - INFO - codeparrot_training - Step 17079: {'lr': 0.0003878159201514188, 'samples': 3279360, 'steps': 17079, 'loss/train': 1.8873814344406128} 01/29/2022 10:47:13 - INFO - codeparrot_training - Step 17080: {'lr': 0.00038780226814584263, 'samples': 3279552, 'steps': 17080, 'loss/train': 2.28753125667572} 01/29/2022 10:47:17 - INFO - codeparrot_training - Step 17081: {'lr': 0.00038778861554996524, 'samples': 3279744, 'steps': 17081, 'loss/train': 1.1080789268016815} 01/29/2022 10:47:24 - INFO - codeparrot_training - Step 17082: {'lr': 0.00038777496236384526, 'samples': 3279936, 'steps': 17082, 'loss/train': 2.3166988492012024} 01/29/2022 10:47:28 - INFO - codeparrot_training - Step 17083: {'lr': 0.000387761308587541, 'samples': 3280128, 'steps': 17083, 'loss/train': 0.7903331816196442} 01/29/2022 10:47:33 - INFO - codeparrot_training - Step 17084: {'lr': 0.0003877476542211111, 'samples': 3280320, 'steps': 17084, 'loss/train': 2.1834331154823303} 01/29/2022 10:47:37 - INFO - codeparrot_training - Step 17085: {'lr': 0.00038773399926461395, 'samples': 3280512, 'steps': 17085, 'loss/train': 1.3150516748428345} 01/29/2022 10:47:41 - INFO - codeparrot_training - Step 17086: {'lr': 0.0003877203437181081, 'samples': 3280704, 'steps': 17086, 'loss/train': 1.997374713420868} 01/29/2022 10:47:46 - INFO - codeparrot_training - Step 17087: {'lr': 0.0003877066875816521, 'samples': 3280896, 'steps': 17087, 'loss/train': 1.1167733073234558} 01/29/2022 10:47:51 - INFO - codeparrot_training - Step 17088: {'lr': 0.00038769303085530425, 'samples': 3281088, 'steps': 17088, 'loss/train': 2.4921804070472717} 01/29/2022 10:47:55 - INFO - codeparrot_training - Step 17089: {'lr': 0.0003876793735391233, 'samples': 3281280, 'steps': 17089, 'loss/train': 1.7331669330596924} 01/29/2022 10:47:59 - INFO - codeparrot_training - Step 17090: {'lr': 0.00038766571563316756, 'samples': 3281472, 'steps': 17090, 'loss/train': 1.5019603371620178} 01/29/2022 10:48:03 - INFO - codeparrot_training - Step 17091: {'lr': 0.00038765205713749563, 'samples': 3281664, 'steps': 17091, 'loss/train': 3.363991141319275} 01/29/2022 10:48:10 - INFO - codeparrot_training - Step 17092: {'lr': 0.0003876383980521659, 'samples': 3281856, 'steps': 17092, 'loss/train': 0.7303884476423264} 01/29/2022 10:48:14 - INFO - codeparrot_training - Step 17093: {'lr': 0.0003876247383772371, 'samples': 3282048, 'steps': 17093, 'loss/train': 2.0719894766807556} 01/29/2022 10:48:19 - INFO - codeparrot_training - Step 17094: {'lr': 0.00038761107811276756, 'samples': 3282240, 'steps': 17094, 'loss/train': 1.819458246231079} 01/29/2022 10:48:23 - INFO - codeparrot_training - Step 17095: {'lr': 0.00038759741725881593, 'samples': 3282432, 'steps': 17095, 'loss/train': 1.2460893988609314} 01/29/2022 10:48:27 - INFO - codeparrot_training - Step 17096: {'lr': 0.0003875837558154406, 'samples': 3282624, 'steps': 17096, 'loss/train': 1.6035343408584595} 01/29/2022 10:48:32 - INFO - codeparrot_training - Step 17097: {'lr': 0.00038757009378270014, 'samples': 3282816, 'steps': 17097, 'loss/train': 1.4052945971488953} 01/29/2022 10:48:37 - INFO - codeparrot_training - Step 17098: {'lr': 0.0003875564311606531, 'samples': 3283008, 'steps': 17098, 'loss/train': 0.8526719212532043} 01/29/2022 10:48:41 - INFO - codeparrot_training - Step 17099: {'lr': 0.000387542767949358, 'samples': 3283200, 'steps': 17099, 'loss/train': 1.4912743270397186} 01/29/2022 10:48:45 - INFO - codeparrot_training - Step 17100: {'lr': 0.0003875291041488734, 'samples': 3283392, 'steps': 17100, 'loss/train': 1.7128865718841553} 01/29/2022 10:48:49 - INFO - codeparrot_training - Step 17101: {'lr': 0.00038751543975925766, 'samples': 3283584, 'steps': 17101, 'loss/train': 0.24675826728343964} 01/29/2022 10:48:55 - INFO - codeparrot_training - Step 17102: {'lr': 0.00038750177478056956, 'samples': 3283776, 'steps': 17102, 'loss/train': 1.6545828580856323} 01/29/2022 10:48:59 - INFO - codeparrot_training - Step 17103: {'lr': 0.0003874881092128675, 'samples': 3283968, 'steps': 17103, 'loss/train': 1.818679690361023} 01/29/2022 10:49:03 - INFO - codeparrot_training - Step 17104: {'lr': 0.00038747444305621, 'samples': 3284160, 'steps': 17104, 'loss/train': 1.2209846377372742} 01/29/2022 10:49:08 - INFO - codeparrot_training - Step 17105: {'lr': 0.0003874607763106556, 'samples': 3284352, 'steps': 17105, 'loss/train': 1.977083444595337} 01/29/2022 10:49:12 - INFO - codeparrot_training - Step 17106: {'lr': 0.00038744710897626293, 'samples': 3284544, 'steps': 17106, 'loss/train': 1.6710017323493958} 01/29/2022 10:49:16 - INFO - codeparrot_training - Step 17107: {'lr': 0.00038743344105309055, 'samples': 3284736, 'steps': 17107, 'loss/train': 1.7748277187347412} 01/29/2022 10:49:23 - INFO - codeparrot_training - Step 17108: {'lr': 0.0003874197725411969, 'samples': 3284928, 'steps': 17108, 'loss/train': 0.9195365309715271} 01/29/2022 10:49:28 - INFO - codeparrot_training - Step 17109: {'lr': 0.0003874061034406405, 'samples': 3285120, 'steps': 17109, 'loss/train': 0.8107149302959442} 01/29/2022 10:49:32 - INFO - codeparrot_training - Step 17110: {'lr': 0.00038739243375148, 'samples': 3285312, 'steps': 17110, 'loss/train': 1.6681666374206543} 01/29/2022 10:49:36 - INFO - codeparrot_training - Step 17111: {'lr': 0.0003873787634737741, 'samples': 3285504, 'steps': 17111, 'loss/train': 1.9375514388084412} 01/29/2022 10:49:40 - INFO - codeparrot_training - Step 17112: {'lr': 0.00038736509260758103, 'samples': 3285696, 'steps': 17112, 'loss/train': 0.8816976249217987} 01/29/2022 10:49:46 - INFO - codeparrot_training - Step 17113: {'lr': 0.00038735142115295965, 'samples': 3285888, 'steps': 17113, 'loss/train': 2.1950742602348328} 01/29/2022 10:49:50 - INFO - codeparrot_training - Step 17114: {'lr': 0.00038733774910996825, 'samples': 3286080, 'steps': 17114, 'loss/train': 1.7400580644607544} 01/29/2022 10:49:54 - INFO - codeparrot_training - Step 17115: {'lr': 0.00038732407647866567, 'samples': 3286272, 'steps': 17115, 'loss/train': 0.9920718669891357} 01/29/2022 10:49:58 - INFO - codeparrot_training - Step 17116: {'lr': 0.00038731040325911027, 'samples': 3286464, 'steps': 17116, 'loss/train': 1.4244860708713531} 01/29/2022 10:50:03 - INFO - codeparrot_training - Step 17117: {'lr': 0.0003872967294513608, 'samples': 3286656, 'steps': 17117, 'loss/train': 1.6492230892181396} 01/29/2022 10:50:10 - INFO - codeparrot_training - Step 17118: {'lr': 0.0003872830550554757, 'samples': 3286848, 'steps': 17118, 'loss/train': 1.8601900935173035} 01/29/2022 10:50:14 - INFO - codeparrot_training - Step 17119: {'lr': 0.0003872693800715135, 'samples': 3287040, 'steps': 17119, 'loss/train': 0.5346035063266754} 01/29/2022 10:50:18 - INFO - codeparrot_training - Step 17120: {'lr': 0.00038725570449953296, 'samples': 3287232, 'steps': 17120, 'loss/train': 2.9868069291114807} 01/29/2022 10:50:23 - INFO - codeparrot_training - Step 17121: {'lr': 0.00038724202833959254, 'samples': 3287424, 'steps': 17121, 'loss/train': 2.3080819845199585} 01/29/2022 10:50:27 - INFO - codeparrot_training - Step 17122: {'lr': 0.00038722835159175087, 'samples': 3287616, 'steps': 17122, 'loss/train': 1.943572461605072} 01/29/2022 10:50:32 - INFO - codeparrot_training - Step 17123: {'lr': 0.00038721467425606644, 'samples': 3287808, 'steps': 17123, 'loss/train': 1.5796129703521729} 01/29/2022 10:50:36 - INFO - codeparrot_training - Step 17124: {'lr': 0.000387200996332598, 'samples': 3288000, 'steps': 17124, 'loss/train': 1.8383821249008179} 01/29/2022 10:50:41 - INFO - codeparrot_training - Step 17125: {'lr': 0.000387187317821404, 'samples': 3288192, 'steps': 17125, 'loss/train': 1.5927008986473083} 01/29/2022 10:50:45 - INFO - codeparrot_training - Step 17126: {'lr': 0.0003871736387225431, 'samples': 3288384, 'steps': 17126, 'loss/train': 1.7301594614982605} 01/29/2022 10:50:49 - INFO - codeparrot_training - Step 17127: {'lr': 0.0003871599590360739, 'samples': 3288576, 'steps': 17127, 'loss/train': 1.4085074365139008} 01/29/2022 10:50:55 - INFO - codeparrot_training - Step 17128: {'lr': 0.000387146278762055, 'samples': 3288768, 'steps': 17128, 'loss/train': 2.6631075739860535} 01/29/2022 10:50:59 - INFO - codeparrot_training - Step 17129: {'lr': 0.000387132597900545, 'samples': 3288960, 'steps': 17129, 'loss/train': 0.10307550430297852} 01/29/2022 10:51:03 - INFO - codeparrot_training - Step 17130: {'lr': 0.0003871189164516025, 'samples': 3289152, 'steps': 17130, 'loss/train': 2.152691602706909} 01/29/2022 10:51:07 - INFO - codeparrot_training - Step 17131: {'lr': 0.000387105234415286, 'samples': 3289344, 'steps': 17131, 'loss/train': 1.9099527597427368} 01/29/2022 10:51:12 - INFO - codeparrot_training - Step 17132: {'lr': 0.00038709155179165436, 'samples': 3289536, 'steps': 17132, 'loss/train': 2.039638102054596} 01/29/2022 10:51:17 - INFO - codeparrot_training - Step 17133: {'lr': 0.000387077868580766, 'samples': 3289728, 'steps': 17133, 'loss/train': 1.9252803325653076} 01/29/2022 10:51:21 - INFO - codeparrot_training - Step 17134: {'lr': 0.00038706418478267945, 'samples': 3289920, 'steps': 17134, 'loss/train': 0.44855618476867676} 01/29/2022 10:51:26 - INFO - codeparrot_training - Step 17135: {'lr': 0.0003870505003974536, 'samples': 3290112, 'steps': 17135, 'loss/train': 1.0733457505702972} 01/29/2022 10:51:30 - INFO - codeparrot_training - Step 17136: {'lr': 0.0003870368154251469, 'samples': 3290304, 'steps': 17136, 'loss/train': 2.4469603300094604} 01/29/2022 10:51:34 - INFO - codeparrot_training - Step 17137: {'lr': 0.000387023129865818, 'samples': 3290496, 'steps': 17137, 'loss/train': 2.5818784832954407} 01/29/2022 10:51:41 - INFO - codeparrot_training - Step 17138: {'lr': 0.00038700944371952543, 'samples': 3290688, 'steps': 17138, 'loss/train': 1.8246118426322937} 01/29/2022 10:51:46 - INFO - codeparrot_training - Step 17139: {'lr': 0.00038699575698632806, 'samples': 3290880, 'steps': 17139, 'loss/train': 1.1473965346813202} 01/29/2022 10:51:50 - INFO - codeparrot_training - Step 17140: {'lr': 0.00038698206966628426, 'samples': 3291072, 'steps': 17140, 'loss/train': 1.670935034751892} 01/29/2022 10:51:54 - INFO - codeparrot_training - Step 17141: {'lr': 0.00038696838175945284, 'samples': 3291264, 'steps': 17141, 'loss/train': 1.390147626399994} 01/29/2022 10:51:58 - INFO - codeparrot_training - Step 17142: {'lr': 0.0003869546932658923, 'samples': 3291456, 'steps': 17142, 'loss/train': 2.2817586064338684} 01/29/2022 10:52:04 - INFO - codeparrot_training - Step 17143: {'lr': 0.0003869410041856614, 'samples': 3291648, 'steps': 17143, 'loss/train': 2.1082035899162292} 01/29/2022 10:52:09 - INFO - codeparrot_training - Step 17144: {'lr': 0.0003869273145188186, 'samples': 3291840, 'steps': 17144, 'loss/train': 1.5444438457489014} 01/29/2022 10:52:13 - INFO - codeparrot_training - Step 17145: {'lr': 0.00038691362426542273, 'samples': 3292032, 'steps': 17145, 'loss/train': 1.797669768333435} 01/29/2022 10:52:17 - INFO - codeparrot_training - Step 17146: {'lr': 0.0003868999334255324, 'samples': 3292224, 'steps': 17146, 'loss/train': 1.5128547549247742} 01/29/2022 10:52:21 - INFO - codeparrot_training - Step 17147: {'lr': 0.00038688624199920623, 'samples': 3292416, 'steps': 17147, 'loss/train': 1.0591136813163757} 01/29/2022 10:52:26 - INFO - codeparrot_training - Step 17148: {'lr': 0.0003868725499865029, 'samples': 3292608, 'steps': 17148, 'loss/train': 1.537438452243805} 01/29/2022 10:52:31 - INFO - codeparrot_training - Step 17149: {'lr': 0.00038685885738748096, 'samples': 3292800, 'steps': 17149, 'loss/train': 1.9974777102470398} 01/29/2022 10:52:35 - INFO - codeparrot_training - Step 17150: {'lr': 0.0003868451642021992, 'samples': 3292992, 'steps': 17150, 'loss/train': 1.5383716821670532} 01/29/2022 10:52:39 - INFO - codeparrot_training - Step 17151: {'lr': 0.0003868314704307161, 'samples': 3293184, 'steps': 17151, 'loss/train': 1.7187319993972778} 01/29/2022 10:52:44 - INFO - codeparrot_training - Step 17152: {'lr': 0.0003868177760730905, 'samples': 3293376, 'steps': 17152, 'loss/train': 1.9189471006393433} 01/29/2022 10:52:48 - INFO - codeparrot_training - Step 17153: {'lr': 0.00038680408112938097, 'samples': 3293568, 'steps': 17153, 'loss/train': 2.230590283870697} 01/29/2022 10:52:55 - INFO - codeparrot_training - Step 17154: {'lr': 0.00038679038559964626, 'samples': 3293760, 'steps': 17154, 'loss/train': 2.161348521709442} 01/29/2022 10:52:59 - INFO - codeparrot_training - Step 17155: {'lr': 0.0003867766894839449, 'samples': 3293952, 'steps': 17155, 'loss/train': 1.886978566646576} 01/29/2022 10:53:04 - INFO - codeparrot_training - Step 17156: {'lr': 0.0003867629927823357, 'samples': 3294144, 'steps': 17156, 'loss/train': 1.418520838022232} 01/29/2022 10:53:08 - INFO - codeparrot_training - Step 17157: {'lr': 0.00038674929549487714, 'samples': 3294336, 'steps': 17157, 'loss/train': 1.9518520832061768} 01/29/2022 10:53:12 - INFO - codeparrot_training - Step 17158: {'lr': 0.00038673559762162816, 'samples': 3294528, 'steps': 17158, 'loss/train': 1.4684103727340698} 01/29/2022 10:53:17 - INFO - codeparrot_training - Step 17159: {'lr': 0.0003867218991626472, 'samples': 3294720, 'steps': 17159, 'loss/train': 0.13888023421168327} 01/29/2022 10:53:22 - INFO - codeparrot_training - Step 17160: {'lr': 0.0003867082001179931, 'samples': 3294912, 'steps': 17160, 'loss/train': 1.569188117980957} 01/29/2022 10:53:26 - INFO - codeparrot_training - Step 17161: {'lr': 0.0003866945004877245, 'samples': 3295104, 'steps': 17161, 'loss/train': 1.4606604874134064} 01/29/2022 10:53:30 - INFO - codeparrot_training - Step 17162: {'lr': 0.0003866808002719, 'samples': 3295296, 'steps': 17162, 'loss/train': 2.027415454387665} 01/29/2022 10:53:34 - INFO - codeparrot_training - Step 17163: {'lr': 0.00038666709947057836, 'samples': 3295488, 'steps': 17163, 'loss/train': 1.6564425230026245} 01/29/2022 10:53:41 - INFO - codeparrot_training - Step 17164: {'lr': 0.0003866533980838183, 'samples': 3295680, 'steps': 17164, 'loss/train': 1.5406238436698914} 01/29/2022 10:53:45 - INFO - codeparrot_training - Step 17165: {'lr': 0.0003866396961116785, 'samples': 3295872, 'steps': 17165, 'loss/train': 1.1223236918449402} 01/29/2022 10:53:50 - INFO - codeparrot_training - Step 17166: {'lr': 0.00038662599355421756, 'samples': 3296064, 'steps': 17166, 'loss/train': 2.017394006252289} 01/29/2022 10:53:54 - INFO - codeparrot_training - Step 17167: {'lr': 0.00038661229041149427, 'samples': 3296256, 'steps': 17167, 'loss/train': 1.7608122825622559} 01/29/2022 10:53:58 - INFO - codeparrot_training - Step 17168: {'lr': 0.0003865985866835673, 'samples': 3296448, 'steps': 17168, 'loss/train': 1.8586961030960083} 01/29/2022 10:54:03 - INFO - codeparrot_training - Step 17169: {'lr': 0.0003865848823704954, 'samples': 3296640, 'steps': 17169, 'loss/train': 0.5087234079837799} 01/29/2022 10:54:08 - INFO - codeparrot_training - Step 17170: {'lr': 0.00038657117747233717, 'samples': 3296832, 'steps': 17170, 'loss/train': 1.011224091053009} 01/29/2022 10:54:12 - INFO - codeparrot_training - Step 17171: {'lr': 0.00038655747198915137, 'samples': 3297024, 'steps': 17171, 'loss/train': 1.451595515012741} 01/29/2022 10:54:16 - INFO - codeparrot_training - Step 17172: {'lr': 0.0003865437659209968, 'samples': 3297216, 'steps': 17172, 'loss/train': 1.553127408027649} 01/29/2022 10:54:20 - INFO - codeparrot_training - Step 17173: {'lr': 0.00038653005926793203, 'samples': 3297408, 'steps': 17173, 'loss/train': 1.8474587202072144} 01/29/2022 10:54:26 - INFO - codeparrot_training - Step 17174: {'lr': 0.0003865163520300159, 'samples': 3297600, 'steps': 17174, 'loss/train': 1.7103915810585022} 01/29/2022 10:54:30 - INFO - codeparrot_training - Step 17175: {'lr': 0.00038650264420730707, 'samples': 3297792, 'steps': 17175, 'loss/train': 1.4958268404006958} 01/29/2022 10:54:34 - INFO - codeparrot_training - Step 17176: {'lr': 0.00038648893579986424, 'samples': 3297984, 'steps': 17176, 'loss/train': 2.1992964148521423} 01/29/2022 10:54:38 - INFO - codeparrot_training - Step 17177: {'lr': 0.00038647522680774603, 'samples': 3298176, 'steps': 17177, 'loss/train': 2.107395887374878} 01/29/2022 10:54:43 - INFO - codeparrot_training - Step 17178: {'lr': 0.0003864615172310115, 'samples': 3298368, 'steps': 17178, 'loss/train': 2.2947757244110107} 01/29/2022 10:54:50 - INFO - codeparrot_training - Step 17179: {'lr': 0.000386447807069719, 'samples': 3298560, 'steps': 17179, 'loss/train': 1.614132821559906} 01/29/2022 10:54:54 - INFO - codeparrot_training - Step 17180: {'lr': 0.0003864340963239275, 'samples': 3298752, 'steps': 17180, 'loss/train': 1.0520194172859192} 01/29/2022 10:54:59 - INFO - codeparrot_training - Step 17181: {'lr': 0.00038642038499369556, 'samples': 3298944, 'steps': 17181, 'loss/train': 1.8109670877456665} 01/29/2022 10:55:03 - INFO - codeparrot_training - Step 17182: {'lr': 0.0003864066730790821, 'samples': 3299136, 'steps': 17182, 'loss/train': 1.1362866461277008} 01/29/2022 10:55:07 - INFO - codeparrot_training - Step 17183: {'lr': 0.00038639296058014575, 'samples': 3299328, 'steps': 17183, 'loss/train': 2.080322027206421} 01/29/2022 10:55:12 - INFO - codeparrot_training - Step 17184: {'lr': 0.0003863792474969453, 'samples': 3299520, 'steps': 17184, 'loss/train': 2.0197209119796753} 01/29/2022 10:55:17 - INFO - codeparrot_training - Step 17185: {'lr': 0.00038636553382953944, 'samples': 3299712, 'steps': 17185, 'loss/train': 1.6774978637695312} 01/29/2022 10:55:21 - INFO - codeparrot_training - Step 17186: {'lr': 0.00038635181957798686, 'samples': 3299904, 'steps': 17186, 'loss/train': 1.867211401462555} 01/29/2022 10:55:25 - INFO - codeparrot_training - Step 17187: {'lr': 0.00038633810474234643, 'samples': 3300096, 'steps': 17187, 'loss/train': 1.384442299604416} 01/29/2022 10:55:29 - INFO - codeparrot_training - Step 17188: {'lr': 0.00038632438932267686, 'samples': 3300288, 'steps': 17188, 'loss/train': 1.7663826942443848} 01/29/2022 10:55:36 - INFO - codeparrot_training - Step 17189: {'lr': 0.0003863106733190369, 'samples': 3300480, 'steps': 17189, 'loss/train': 1.975433886051178} 01/29/2022 10:55:41 - INFO - codeparrot_training - Step 17190: {'lr': 0.0003862969567314852, 'samples': 3300672, 'steps': 17190, 'loss/train': 2.136876404285431} 01/29/2022 10:55:45 - INFO - codeparrot_training - Step 17191: {'lr': 0.0003862832395600808, 'samples': 3300864, 'steps': 17191, 'loss/train': 1.262523204088211} 01/29/2022 10:55:49 - INFO - codeparrot_training - Step 17192: {'lr': 0.0003862695218048822, 'samples': 3301056, 'steps': 17192, 'loss/train': 2.2433377504348755} 01/29/2022 10:55:53 - INFO - codeparrot_training - Step 17193: {'lr': 0.00038625580346594824, 'samples': 3301248, 'steps': 17193, 'loss/train': 1.7115052342414856} 01/29/2022 10:55:59 - INFO - codeparrot_training - Step 17194: {'lr': 0.00038624208454333763, 'samples': 3301440, 'steps': 17194, 'loss/train': 1.5932528972625732} 01/29/2022 10:56:03 - INFO - codeparrot_training - Step 17195: {'lr': 0.00038622836503710917, 'samples': 3301632, 'steps': 17195, 'loss/train': 1.4251309633255005} 01/29/2022 10:56:08 - INFO - codeparrot_training - Step 17196: {'lr': 0.00038621464494732174, 'samples': 3301824, 'steps': 17196, 'loss/train': 0.30925512313842773} 01/29/2022 10:56:12 - INFO - codeparrot_training - Step 17197: {'lr': 0.00038620092427403395, 'samples': 3302016, 'steps': 17197, 'loss/train': 7.498128890991211} 01/29/2022 10:56:16 - INFO - codeparrot_training - Step 17198: {'lr': 0.0003861872030173047, 'samples': 3302208, 'steps': 17198, 'loss/train': 1.6583184599876404} 01/29/2022 10:56:22 - INFO - codeparrot_training - Step 17199: {'lr': 0.0003861734811771928, 'samples': 3302400, 'steps': 17199, 'loss/train': 1.9600905776023865} 01/29/2022 10:56:26 - INFO - codeparrot_training - Step 17200: {'lr': 0.00038615975875375683, 'samples': 3302592, 'steps': 17200, 'loss/train': 6.40573525428772} 01/29/2022 10:56:30 - INFO - codeparrot_training - Step 17201: {'lr': 0.0003861460357470556, 'samples': 3302784, 'steps': 17201, 'loss/train': 7.295032739639282} 01/29/2022 10:56:35 - INFO - codeparrot_training - Step 17202: {'lr': 0.0003861323121571482, 'samples': 3302976, 'steps': 17202, 'loss/train': 1.4776736497879028} 01/29/2022 10:56:39 - INFO - codeparrot_training - Step 17203: {'lr': 0.0003861185879840931, 'samples': 3303168, 'steps': 17203, 'loss/train': 2.2994157671928406} 01/29/2022 10:56:43 - INFO - codeparrot_training - Step 17204: {'lr': 0.00038610486322794915, 'samples': 3303360, 'steps': 17204, 'loss/train': 2.315956950187683} 01/29/2022 10:56:50 - INFO - codeparrot_training - Step 17205: {'lr': 0.0003860911378887752, 'samples': 3303552, 'steps': 17205, 'loss/train': 1.9142176508903503} 01/29/2022 10:56:55 - INFO - codeparrot_training - Step 17206: {'lr': 0.00038607741196663005, 'samples': 3303744, 'steps': 17206, 'loss/train': 1.7936650514602661} 01/29/2022 10:56:59 - INFO - codeparrot_training - Step 17207: {'lr': 0.0003860636854615725, 'samples': 3303936, 'steps': 17207, 'loss/train': 1.1395200490951538} 01/29/2022 10:57:03 - INFO - codeparrot_training - Step 17208: {'lr': 0.0003860499583736613, 'samples': 3304128, 'steps': 17208, 'loss/train': 1.4375137388706207} 01/29/2022 10:57:07 - INFO - codeparrot_training - Step 17209: {'lr': 0.00038603623070295536, 'samples': 3304320, 'steps': 17209, 'loss/train': 2.527778148651123} 01/29/2022 10:57:13 - INFO - codeparrot_training - Step 17210: {'lr': 0.0003860225024495133, 'samples': 3304512, 'steps': 17210, 'loss/train': 1.8341163396835327} 01/29/2022 10:57:17 - INFO - codeparrot_training - Step 17211: {'lr': 0.000386008773613394, 'samples': 3304704, 'steps': 17211, 'loss/train': 2.121522009372711} 01/29/2022 10:57:21 - INFO - codeparrot_training - Step 17212: {'lr': 0.0003859950441946564, 'samples': 3304896, 'steps': 17212, 'loss/train': 1.5507450699806213} 01/29/2022 10:57:25 - INFO - codeparrot_training - Step 17213: {'lr': 0.0003859813141933592, 'samples': 3305088, 'steps': 17213, 'loss/train': 1.8627356886863708} 01/29/2022 10:57:30 - INFO - codeparrot_training - Step 17214: {'lr': 0.0003859675836095612, 'samples': 3305280, 'steps': 17214, 'loss/train': 1.632836401462555} 01/29/2022 10:57:34 - INFO - codeparrot_training - Step 17215: {'lr': 0.00038595385244332125, 'samples': 3305472, 'steps': 17215, 'loss/train': 2.3823922276496887} 01/29/2022 10:57:40 - INFO - codeparrot_training - Step 17216: {'lr': 0.00038594012069469814, 'samples': 3305664, 'steps': 17216, 'loss/train': 1.792051076889038} 01/29/2022 10:57:44 - INFO - codeparrot_training - Step 17217: {'lr': 0.00038592638836375075, 'samples': 3305856, 'steps': 17217, 'loss/train': 1.085476577281952} 01/29/2022 10:57:48 - INFO - codeparrot_training - Step 17218: {'lr': 0.0003859126554505379, 'samples': 3306048, 'steps': 17218, 'loss/train': 0.7837775945663452} 01/29/2022 10:57:52 - INFO - codeparrot_training - Step 17219: {'lr': 0.00038589892195511834, 'samples': 3306240, 'steps': 17219, 'loss/train': 2.116000771522522} 01/29/2022 10:57:57 - INFO - codeparrot_training - Step 17220: {'lr': 0.00038588518787755096, 'samples': 3306432, 'steps': 17220, 'loss/train': 1.7239184975624084} 01/29/2022 10:58:02 - INFO - codeparrot_training - Step 17221: {'lr': 0.00038587145321789456, 'samples': 3306624, 'steps': 17221, 'loss/train': 1.7740543484687805} 01/29/2022 10:58:07 - INFO - codeparrot_training - Step 17222: {'lr': 0.00038585771797620803, 'samples': 3306816, 'steps': 17222, 'loss/train': 2.657557725906372} 01/29/2022 10:58:11 - INFO - codeparrot_training - Step 17223: {'lr': 0.00038584398215255023, 'samples': 3307008, 'steps': 17223, 'loss/train': 1.8574978709220886} 01/29/2022 10:58:15 - INFO - codeparrot_training - Step 17224: {'lr': 0.0003858302457469799, 'samples': 3307200, 'steps': 17224, 'loss/train': 1.3537523746490479} 01/29/2022 10:58:22 - INFO - codeparrot_training - Step 17225: {'lr': 0.0003858165087595559, 'samples': 3307392, 'steps': 17225, 'loss/train': 1.0350916385650635} 01/29/2022 10:58:26 - INFO - codeparrot_training - Step 17226: {'lr': 0.00038580277119033715, 'samples': 3307584, 'steps': 17226, 'loss/train': 1.843604028224945} 01/29/2022 10:58:31 - INFO - codeparrot_training - Step 17227: {'lr': 0.0003857890330393824, 'samples': 3307776, 'steps': 17227, 'loss/train': 2.0885653495788574} 01/29/2022 10:58:35 - INFO - codeparrot_training - Step 17228: {'lr': 0.0003857752943067506, 'samples': 3307968, 'steps': 17228, 'loss/train': 1.2859049141407013} 01/29/2022 10:58:39 - INFO - codeparrot_training - Step 17229: {'lr': 0.00038576155499250056, 'samples': 3308160, 'steps': 17229, 'loss/train': 0.9710030257701874} 01/29/2022 10:58:45 - INFO - codeparrot_training - Step 17230: {'lr': 0.000385747815096691, 'samples': 3308352, 'steps': 17230, 'loss/train': 1.4031226336956024} 01/29/2022 10:58:49 - INFO - codeparrot_training - Step 17231: {'lr': 0.00038573407461938103, 'samples': 3308544, 'steps': 17231, 'loss/train': 1.616248905658722} 01/29/2022 10:58:53 - INFO - codeparrot_training - Step 17232: {'lr': 0.0003857203335606294, 'samples': 3308736, 'steps': 17232, 'loss/train': 1.9803269505500793} 01/29/2022 10:58:57 - INFO - codeparrot_training - Step 17233: {'lr': 0.00038570659192049497, 'samples': 3308928, 'steps': 17233, 'loss/train': 2.541278064250946} 01/29/2022 10:59:02 - INFO - codeparrot_training - Step 17234: {'lr': 0.0003856928496990364, 'samples': 3309120, 'steps': 17234, 'loss/train': 1.2501046657562256} 01/29/2022 10:59:06 - INFO - codeparrot_training - Step 17235: {'lr': 0.000385679106896313, 'samples': 3309312, 'steps': 17235, 'loss/train': 1.4253760278224945} 01/29/2022 10:59:13 - INFO - codeparrot_training - Step 17236: {'lr': 0.0003856653635123832, 'samples': 3309504, 'steps': 17236, 'loss/train': 1.5097777247428894} 01/29/2022 10:59:18 - INFO - codeparrot_training - Step 17237: {'lr': 0.0003856516195473062, 'samples': 3309696, 'steps': 17237, 'loss/train': 1.399737685918808} 01/29/2022 10:59:22 - INFO - codeparrot_training - Step 17238: {'lr': 0.0003856378750011407, 'samples': 3309888, 'steps': 17238, 'loss/train': 1.4928744435310364} 01/29/2022 10:59:26 - INFO - codeparrot_training - Step 17239: {'lr': 0.0003856241298739456, 'samples': 3310080, 'steps': 17239, 'loss/train': 1.8066505193710327} 01/29/2022 10:59:30 - INFO - codeparrot_training - Step 17240: {'lr': 0.0003856103841657797, 'samples': 3310272, 'steps': 17240, 'loss/train': 1.5199917554855347} 01/29/2022 10:59:36 - INFO - codeparrot_training - Step 17241: {'lr': 0.0003855966378767021, 'samples': 3310464, 'steps': 17241, 'loss/train': 2.8911126852035522} 01/29/2022 10:59:40 - INFO - codeparrot_training - Step 17242: {'lr': 0.00038558289100677144, 'samples': 3310656, 'steps': 17242, 'loss/train': 2.120400130748749} 01/29/2022 10:59:44 - INFO - codeparrot_training - Step 17243: {'lr': 0.00038556914355604676, 'samples': 3310848, 'steps': 17243, 'loss/train': 0.8288732171058655} 01/29/2022 10:59:49 - INFO - codeparrot_training - Step 17244: {'lr': 0.0003855553955245871, 'samples': 3311040, 'steps': 17244, 'loss/train': 0.19920607656240463} 01/29/2022 10:59:53 - INFO - codeparrot_training - Step 17245: {'lr': 0.00038554164691245095, 'samples': 3311232, 'steps': 17245, 'loss/train': 1.5965972542762756} 01/29/2022 11:00:00 - INFO - codeparrot_training - Step 17246: {'lr': 0.00038552789771969755, 'samples': 3311424, 'steps': 17246, 'loss/train': 2.874321162700653} 01/29/2022 11:00:04 - INFO - codeparrot_training - Step 17247: {'lr': 0.00038551414794638555, 'samples': 3311616, 'steps': 17247, 'loss/train': 2.3317673206329346} 01/29/2022 11:00:08 - INFO - codeparrot_training - Step 17248: {'lr': 0.00038550039759257404, 'samples': 3311808, 'steps': 17248, 'loss/train': 1.7154813408851624} 01/29/2022 11:00:13 - INFO - codeparrot_training - Step 17249: {'lr': 0.0003854866466583219, 'samples': 3312000, 'steps': 17249, 'loss/train': 2.0277822017669678} 01/29/2022 11:00:17 - INFO - codeparrot_training - Step 17250: {'lr': 0.00038547289514368795, 'samples': 3312192, 'steps': 17250, 'loss/train': 2.252538800239563} 01/29/2022 11:00:22 - INFO - codeparrot_training - Step 17251: {'lr': 0.00038545914304873117, 'samples': 3312384, 'steps': 17251, 'loss/train': 1.0102063715457916} 01/29/2022 11:00:26 - INFO - codeparrot_training - Step 17252: {'lr': 0.00038544539037351037, 'samples': 3312576, 'steps': 17252, 'loss/train': 1.5689758658409119} 01/29/2022 11:00:31 - INFO - codeparrot_training - Step 17253: {'lr': 0.00038543163711808457, 'samples': 3312768, 'steps': 17253, 'loss/train': 1.6558109521865845} 01/29/2022 11:00:35 - INFO - codeparrot_training - Step 17254: {'lr': 0.0003854178832825126, 'samples': 3312960, 'steps': 17254, 'loss/train': 2.055568277835846} 01/29/2022 11:00:39 - INFO - codeparrot_training - Step 17255: {'lr': 0.0003854041288668534, 'samples': 3313152, 'steps': 17255, 'loss/train': 1.0411302745342255} 01/29/2022 11:00:45 - INFO - codeparrot_training - Step 17256: {'lr': 0.00038539037387116595, 'samples': 3313344, 'steps': 17256, 'loss/train': 0.8076386153697968} 01/29/2022 11:00:49 - INFO - codeparrot_training - Step 17257: {'lr': 0.0003853766182955092, 'samples': 3313536, 'steps': 17257, 'loss/train': 2.2827356457710266} 01/29/2022 11:00:53 - INFO - codeparrot_training - Step 17258: {'lr': 0.0003853628621399419, 'samples': 3313728, 'steps': 17258, 'loss/train': 1.7332624197006226} 01/29/2022 11:00:57 - INFO - codeparrot_training - Step 17259: {'lr': 0.00038534910540452305, 'samples': 3313920, 'steps': 17259, 'loss/train': 1.752720057964325} 01/29/2022 11:01:02 - INFO - codeparrot_training - Step 17260: {'lr': 0.0003853353480893117, 'samples': 3314112, 'steps': 17260, 'loss/train': 1.333802729845047} 01/29/2022 11:01:09 - INFO - codeparrot_training - Step 17261: {'lr': 0.0003853215901943667, 'samples': 3314304, 'steps': 17261, 'loss/train': 1.689863383769989} 01/29/2022 11:01:14 - INFO - codeparrot_training - Step 17262: {'lr': 0.00038530783171974694, 'samples': 3314496, 'steps': 17262, 'loss/train': 1.197862833738327} 01/29/2022 11:01:18 - INFO - codeparrot_training - Step 17263: {'lr': 0.0003852940726655114, 'samples': 3314688, 'steps': 17263, 'loss/train': 1.8171905279159546} 01/29/2022 11:01:22 - INFO - codeparrot_training - Step 17264: {'lr': 0.000385280313031719, 'samples': 3314880, 'steps': 17264, 'loss/train': 1.8484516739845276} 01/29/2022 11:01:26 - INFO - codeparrot_training - Step 17265: {'lr': 0.0003852665528184287, 'samples': 3315072, 'steps': 17265, 'loss/train': 0.40166643261909485} 01/29/2022 11:01:31 - INFO - codeparrot_training - Step 17266: {'lr': 0.0003852527920256994, 'samples': 3315264, 'steps': 17266, 'loss/train': 1.488066405057907} 01/29/2022 11:01:36 - INFO - codeparrot_training - Step 17267: {'lr': 0.00038523903065359013, 'samples': 3315456, 'steps': 17267, 'loss/train': 2.0682642459869385} 01/29/2022 11:01:40 - INFO - codeparrot_training - Step 17268: {'lr': 0.0003852252687021598, 'samples': 3315648, 'steps': 17268, 'loss/train': 1.8681879043579102} 01/29/2022 11:01:44 - INFO - codeparrot_training - Step 17269: {'lr': 0.00038521150617146737, 'samples': 3315840, 'steps': 17269, 'loss/train': 2.0458715558052063} 01/29/2022 11:01:49 - INFO - codeparrot_training - Step 17270: {'lr': 0.00038519774306157174, 'samples': 3316032, 'steps': 17270, 'loss/train': 1.5371257066726685} 01/29/2022 11:01:53 - INFO - codeparrot_training - Step 17271: {'lr': 0.00038518397937253195, 'samples': 3316224, 'steps': 17271, 'loss/train': 1.833054006099701} 01/29/2022 11:02:00 - INFO - codeparrot_training - Step 17272: {'lr': 0.00038517021510440694, 'samples': 3316416, 'steps': 17272, 'loss/train': 2.100705921649933} 01/29/2022 11:02:04 - INFO - codeparrot_training - Step 17273: {'lr': 0.0003851564502572556, 'samples': 3316608, 'steps': 17273, 'loss/train': 1.6058633923530579} 01/29/2022 11:02:08 - INFO - codeparrot_training - Step 17274: {'lr': 0.00038514268483113694, 'samples': 3316800, 'steps': 17274, 'loss/train': 1.7213864922523499} 01/29/2022 11:02:13 - INFO - codeparrot_training - Step 17275: {'lr': 0.00038512891882610997, 'samples': 3316992, 'steps': 17275, 'loss/train': 1.069853514432907} 01/29/2022 11:02:17 - INFO - codeparrot_training - Step 17276: {'lr': 0.0003851151522422336, 'samples': 3317184, 'steps': 17276, 'loss/train': 1.6992251873016357} 01/29/2022 11:02:22 - INFO - codeparrot_training - Step 17277: {'lr': 0.0003851013850795668, 'samples': 3317376, 'steps': 17277, 'loss/train': 1.6569333672523499} 01/29/2022 11:02:26 - INFO - codeparrot_training - Step 17278: {'lr': 0.00038508761733816864, 'samples': 3317568, 'steps': 17278, 'loss/train': 2.4539222717285156} 01/29/2022 11:02:31 - INFO - codeparrot_training - Step 17279: {'lr': 0.00038507384901809795, 'samples': 3317760, 'steps': 17279, 'loss/train': 1.5949612855911255} 01/29/2022 11:02:35 - INFO - codeparrot_training - Step 17280: {'lr': 0.00038506008011941376, 'samples': 3317952, 'steps': 17280, 'loss/train': 1.8627639412879944} 01/29/2022 11:02:39 - INFO - codeparrot_training - Step 17281: {'lr': 0.0003850463106421751, 'samples': 3318144, 'steps': 17281, 'loss/train': 0.8367638289928436} 01/29/2022 11:02:45 - INFO - codeparrot_training - Step 17282: {'lr': 0.000385032540586441, 'samples': 3318336, 'steps': 17282, 'loss/train': 2.198942184448242} 01/29/2022 11:02:49 - INFO - codeparrot_training - Step 17283: {'lr': 0.00038501876995227023, 'samples': 3318528, 'steps': 17283, 'loss/train': 0.412501260638237} 01/29/2022 11:02:53 - INFO - codeparrot_training - Step 17284: {'lr': 0.00038500499873972204, 'samples': 3318720, 'steps': 17284, 'loss/train': 1.4798129796981812} 01/29/2022 11:02:57 - INFO - codeparrot_training - Step 17285: {'lr': 0.0003849912269488552, 'samples': 3318912, 'steps': 17285, 'loss/train': 2.099704921245575} 01/29/2022 11:03:03 - INFO - codeparrot_training - Step 17286: {'lr': 0.000384977454579729, 'samples': 3319104, 'steps': 17286, 'loss/train': 1.0868985056877136} 01/29/2022 11:03:07 - INFO - codeparrot_training - Step 17287: {'lr': 0.00038496368163240215, 'samples': 3319296, 'steps': 17287, 'loss/train': 1.5059664845466614} 01/29/2022 11:03:12 - INFO - codeparrot_training - Step 17288: {'lr': 0.00038494990810693366, 'samples': 3319488, 'steps': 17288, 'loss/train': 1.0394772291183472} 01/29/2022 11:03:16 - INFO - codeparrot_training - Step 17289: {'lr': 0.00038493613400338267, 'samples': 3319680, 'steps': 17289, 'loss/train': 1.7585890889167786} 01/29/2022 11:03:20 - INFO - codeparrot_training - Step 17290: {'lr': 0.0003849223593218082, 'samples': 3319872, 'steps': 17290, 'loss/train': 1.838178277015686} 01/29/2022 11:03:24 - INFO - codeparrot_training - Step 17291: {'lr': 0.00038490858406226903, 'samples': 3320064, 'steps': 17291, 'loss/train': 1.7093886137008667} 01/29/2022 11:03:32 - INFO - codeparrot_training - Step 17292: {'lr': 0.00038489480822482446, 'samples': 3320256, 'steps': 17292, 'loss/train': 1.4837048649787903} 01/29/2022 11:03:36 - INFO - codeparrot_training - Step 17293: {'lr': 0.00038488103180953326, 'samples': 3320448, 'steps': 17293, 'loss/train': 2.857185959815979} 01/29/2022 11:03:40 - INFO - codeparrot_training - Step 17294: {'lr': 0.00038486725481645467, 'samples': 3320640, 'steps': 17294, 'loss/train': 1.9199450612068176} 01/29/2022 11:03:44 - INFO - codeparrot_training - Step 17295: {'lr': 0.00038485347724564746, 'samples': 3320832, 'steps': 17295, 'loss/train': 1.5616990327835083} 01/29/2022 11:03:48 - INFO - codeparrot_training - Step 17296: {'lr': 0.0003848396990971709, 'samples': 3321024, 'steps': 17296, 'loss/train': 0.9765254259109497} 01/29/2022 11:03:54 - INFO - codeparrot_training - Step 17297: {'lr': 0.00038482592037108375, 'samples': 3321216, 'steps': 17297, 'loss/train': 1.9118456840515137} 01/29/2022 11:03:58 - INFO - codeparrot_training - Step 17298: {'lr': 0.0003848121410674453, 'samples': 3321408, 'steps': 17298, 'loss/train': 0.6689186543226242} 01/29/2022 11:04:02 - INFO - codeparrot_training - Step 17299: {'lr': 0.0003847983611863144, 'samples': 3321600, 'steps': 17299, 'loss/train': 1.4836830496788025} 01/29/2022 11:04:06 - INFO - codeparrot_training - Step 17300: {'lr': 0.0003847845807277501, 'samples': 3321792, 'steps': 17300, 'loss/train': 1.6864982843399048} 01/29/2022 11:04:11 - INFO - codeparrot_training - Step 17301: {'lr': 0.00038477079969181146, 'samples': 3321984, 'steps': 17301, 'loss/train': 1.2162141501903534} 01/29/2022 11:04:16 - INFO - codeparrot_training - Step 17302: {'lr': 0.00038475701807855753, 'samples': 3322176, 'steps': 17302, 'loss/train': 1.8167062997817993} 01/29/2022 11:04:20 - INFO - codeparrot_training - Step 17303: {'lr': 0.00038474323588804727, 'samples': 3322368, 'steps': 17303, 'loss/train': 1.1347159445285797} 01/29/2022 11:04:25 - INFO - codeparrot_training - Step 17304: {'lr': 0.0003847294531203398, 'samples': 3322560, 'steps': 17304, 'loss/train': 1.3041397333145142} 01/29/2022 11:04:29 - INFO - codeparrot_training - Step 17305: {'lr': 0.0003847156697754942, 'samples': 3322752, 'steps': 17305, 'loss/train': 1.5251237154006958} 01/29/2022 11:04:33 - INFO - codeparrot_training - Step 17306: {'lr': 0.00038470188585356936, 'samples': 3322944, 'steps': 17306, 'loss/train': 1.9165837168693542} 01/29/2022 11:04:40 - INFO - codeparrot_training - Step 17307: {'lr': 0.00038468810135462445, 'samples': 3323136, 'steps': 17307, 'loss/train': 0.9213853776454926} 01/29/2022 11:04:44 - INFO - codeparrot_training - Step 17308: {'lr': 0.00038467431627871844, 'samples': 3323328, 'steps': 17308, 'loss/train': 1.0223100185394287} 01/29/2022 11:04:49 - INFO - codeparrot_training - Step 17309: {'lr': 0.0003846605306259105, 'samples': 3323520, 'steps': 17309, 'loss/train': 1.7144613862037659} 01/29/2022 11:04:53 - INFO - codeparrot_training - Step 17310: {'lr': 0.0003846467443962596, 'samples': 3323712, 'steps': 17310, 'loss/train': 0.8298778831958771} 01/29/2022 11:04:57 - INFO - codeparrot_training - Step 17311: {'lr': 0.00038463295758982475, 'samples': 3323904, 'steps': 17311, 'loss/train': 2.008354425430298} 01/29/2022 11:05:03 - INFO - codeparrot_training - Step 17312: {'lr': 0.00038461917020666506, 'samples': 3324096, 'steps': 17312, 'loss/train': 2.063431978225708} 01/29/2022 11:05:07 - INFO - codeparrot_training - Step 17313: {'lr': 0.0003846053822468396, 'samples': 3324288, 'steps': 17313, 'loss/train': 1.5036942958831787} 01/29/2022 11:05:11 - INFO - codeparrot_training - Step 17314: {'lr': 0.00038459159371040743, 'samples': 3324480, 'steps': 17314, 'loss/train': 1.393176645040512} 01/29/2022 11:05:15 - INFO - codeparrot_training - Step 17315: {'lr': 0.0003845778045974276, 'samples': 3324672, 'steps': 17315, 'loss/train': 2.1398173570632935} 01/29/2022 11:05:20 - INFO - codeparrot_training - Step 17316: {'lr': 0.0003845640149079592, 'samples': 3324864, 'steps': 17316, 'loss/train': 0.8267620503902435} 01/29/2022 11:05:25 - INFO - codeparrot_training - Step 17317: {'lr': 0.0003845502246420613, 'samples': 3325056, 'steps': 17317, 'loss/train': 1.2690302431583405} 01/29/2022 11:05:29 - INFO - codeparrot_training - Step 17318: {'lr': 0.00038453643379979295, 'samples': 3325248, 'steps': 17318, 'loss/train': 2.0814287066459656} 01/29/2022 11:05:33 - INFO - codeparrot_training - Step 17319: {'lr': 0.00038452264238121326, 'samples': 3325440, 'steps': 17319, 'loss/train': 1.3578351438045502} 01/29/2022 11:05:38 - INFO - codeparrot_training - Step 17320: {'lr': 0.0003845088503863813, 'samples': 3325632, 'steps': 17320, 'loss/train': 1.1987173855304718} 01/29/2022 11:05:42 - INFO - codeparrot_training - Step 17321: {'lr': 0.0003844950578153561, 'samples': 3325824, 'steps': 17321, 'loss/train': 2.1142369508743286} 01/29/2022 11:05:49 - INFO - codeparrot_training - Step 17322: {'lr': 0.00038448126466819675, 'samples': 3326016, 'steps': 17322, 'loss/train': 1.2410096526145935} 01/29/2022 11:05:53 - INFO - codeparrot_training - Step 17323: {'lr': 0.00038446747094496243, 'samples': 3326208, 'steps': 17323, 'loss/train': 1.500884771347046} 01/29/2022 11:05:58 - INFO - codeparrot_training - Step 17324: {'lr': 0.00038445367664571216, 'samples': 3326400, 'steps': 17324, 'loss/train': 1.206505000591278} 01/29/2022 11:06:02 - INFO - codeparrot_training - Step 17325: {'lr': 0.000384439881770505, 'samples': 3326592, 'steps': 17325, 'loss/train': 1.4785232841968536} 01/29/2022 11:06:06 - INFO - codeparrot_training - Step 17326: {'lr': 0.0003844260863194001, 'samples': 3326784, 'steps': 17326, 'loss/train': 1.7918504476547241} 01/29/2022 11:06:12 - INFO - codeparrot_training - Step 17327: {'lr': 0.0003844122902924565, 'samples': 3326976, 'steps': 17327, 'loss/train': 1.7724148035049438} 01/29/2022 11:06:16 - INFO - codeparrot_training - Step 17328: {'lr': 0.00038439849368973334, 'samples': 3327168, 'steps': 17328, 'loss/train': 2.683814227581024} 01/29/2022 11:06:20 - INFO - codeparrot_training - Step 17329: {'lr': 0.0003843846965112897, 'samples': 3327360, 'steps': 17329, 'loss/train': 0.920055627822876} 01/29/2022 11:06:24 - INFO - codeparrot_training - Step 17330: {'lr': 0.0003843708987571847, 'samples': 3327552, 'steps': 17330, 'loss/train': 2.167359173297882} 01/29/2022 11:06:29 - INFO - codeparrot_training - Step 17331: {'lr': 0.0003843571004274775, 'samples': 3327744, 'steps': 17331, 'loss/train': 2.0206130146980286} 01/29/2022 11:06:36 - INFO - codeparrot_training - Step 17332: {'lr': 0.0003843433015222271, 'samples': 3327936, 'steps': 17332, 'loss/train': 1.8252466320991516} 01/29/2022 11:06:40 - INFO - codeparrot_training - Step 17333: {'lr': 0.0003843295020414926, 'samples': 3328128, 'steps': 17333, 'loss/train': 0.17046506702899933} 01/29/2022 11:06:45 - INFO - codeparrot_training - Step 17334: {'lr': 0.0003843157019853332, 'samples': 3328320, 'steps': 17334, 'loss/train': 0.7261201590299606} 01/29/2022 11:06:49 - INFO - codeparrot_training - Step 17335: {'lr': 0.00038430190135380803, 'samples': 3328512, 'steps': 17335, 'loss/train': 1.569157898426056} 01/29/2022 11:06:53 - INFO - codeparrot_training - Step 17336: {'lr': 0.00038428810014697615, 'samples': 3328704, 'steps': 17336, 'loss/train': 2.5282092690467834} 01/29/2022 11:06:58 - INFO - codeparrot_training - Step 17337: {'lr': 0.00038427429836489663, 'samples': 3328896, 'steps': 17337, 'loss/train': 1.8995051980018616} 01/29/2022 11:07:03 - INFO - codeparrot_training - Step 17338: {'lr': 0.00038426049600762867, 'samples': 3329088, 'steps': 17338, 'loss/train': 2.1235944628715515} 01/29/2022 11:07:07 - INFO - codeparrot_training - Step 17339: {'lr': 0.00038424669307523135, 'samples': 3329280, 'steps': 17339, 'loss/train': 1.3178859651088715} 01/29/2022 11:07:11 - INFO - codeparrot_training - Step 17340: {'lr': 0.00038423288956776394, 'samples': 3329472, 'steps': 17340, 'loss/train': 1.1104039549827576} 01/29/2022 11:07:15 - INFO - codeparrot_training - Step 17341: {'lr': 0.00038421908548528534, 'samples': 3329664, 'steps': 17341, 'loss/train': 0.8163414895534515} 01/29/2022 11:07:21 - INFO - codeparrot_training - Step 17342: {'lr': 0.0003842052808278549, 'samples': 3329856, 'steps': 17342, 'loss/train': 1.8154524564743042} 01/29/2022 11:07:25 - INFO - codeparrot_training - Step 17343: {'lr': 0.0003841914755955315, 'samples': 3330048, 'steps': 17343, 'loss/train': 1.5661944150924683} 01/29/2022 11:07:29 - INFO - codeparrot_training - Step 17344: {'lr': 0.00038417766978837453, 'samples': 3330240, 'steps': 17344, 'loss/train': 1.816237986087799} 01/29/2022 11:07:34 - INFO - codeparrot_training - Step 17345: {'lr': 0.00038416386340644305, 'samples': 3330432, 'steps': 17345, 'loss/train': 1.6448634266853333} 01/29/2022 11:07:38 - INFO - codeparrot_training - Step 17346: {'lr': 0.00038415005644979616, 'samples': 3330624, 'steps': 17346, 'loss/train': 1.6914303302764893} 01/29/2022 11:07:44 - INFO - codeparrot_training - Step 17347: {'lr': 0.00038413624891849295, 'samples': 3330816, 'steps': 17347, 'loss/train': 2.1245539784431458} 01/29/2022 11:07:48 - INFO - codeparrot_training - Step 17348: {'lr': 0.00038412244081259273, 'samples': 3331008, 'steps': 17348, 'loss/train': 1.8596652746200562} 01/29/2022 11:07:52 - INFO - codeparrot_training - Step 17349: {'lr': 0.00038410863213215454, 'samples': 3331200, 'steps': 17349, 'loss/train': 1.3818313479423523} 01/29/2022 11:07:56 - INFO - codeparrot_training - Step 17350: {'lr': 0.0003840948228772376, 'samples': 3331392, 'steps': 17350, 'loss/train': 1.0008570849895477} 01/29/2022 11:08:01 - INFO - codeparrot_training - Step 17351: {'lr': 0.00038408101304790096, 'samples': 3331584, 'steps': 17351, 'loss/train': 1.9579038619995117} 01/29/2022 11:08:08 - INFO - codeparrot_training - Step 17352: {'lr': 0.0003840672026442038, 'samples': 3331776, 'steps': 17352, 'loss/train': 1.2951360940933228} 01/29/2022 11:08:12 - INFO - codeparrot_training - Step 17353: {'lr': 0.0003840533916662054, 'samples': 3331968, 'steps': 17353, 'loss/train': 0.6714740842580795} 01/29/2022 11:08:16 - INFO - codeparrot_training - Step 17354: {'lr': 0.00038403958011396476, 'samples': 3332160, 'steps': 17354, 'loss/train': 2.0202906131744385} 01/29/2022 11:08:20 - INFO - codeparrot_training - Step 17355: {'lr': 0.0003840257679875412, 'samples': 3332352, 'steps': 17355, 'loss/train': 1.977771520614624} 01/29/2022 11:08:25 - INFO - codeparrot_training - Step 17356: {'lr': 0.00038401195528699374, 'samples': 3332544, 'steps': 17356, 'loss/train': 1.098795622587204} 01/29/2022 11:08:30 - INFO - codeparrot_training - Step 17357: {'lr': 0.0003839981420123817, 'samples': 3332736, 'steps': 17357, 'loss/train': 1.677553653717041} 01/29/2022 11:08:34 - INFO - codeparrot_training - Step 17358: {'lr': 0.00038398432816376404, 'samples': 3332928, 'steps': 17358, 'loss/train': 1.452240228652954} 01/29/2022 11:08:38 - INFO - codeparrot_training - Step 17359: {'lr': 0.00038397051374120016, 'samples': 3333120, 'steps': 17359, 'loss/train': 1.845356047153473} 01/29/2022 11:08:43 - INFO - codeparrot_training - Step 17360: {'lr': 0.00038395669874474915, 'samples': 3333312, 'steps': 17360, 'loss/train': 1.7777538299560547} 01/29/2022 11:08:47 - INFO - codeparrot_training - Step 17361: {'lr': 0.0003839428831744702, 'samples': 3333504, 'steps': 17361, 'loss/train': 1.8834335803985596} 01/29/2022 11:08:53 - INFO - codeparrot_training - Step 17362: {'lr': 0.0003839290670304224, 'samples': 3333696, 'steps': 17362, 'loss/train': 2.5547114610671997} 01/29/2022 11:08:58 - INFO - codeparrot_training - Step 17363: {'lr': 0.00038391525031266494, 'samples': 3333888, 'steps': 17363, 'loss/train': 0.8410661816596985} 01/29/2022 11:09:02 - INFO - codeparrot_training - Step 17364: {'lr': 0.0003839014330212572, 'samples': 3334080, 'steps': 17364, 'loss/train': 2.551266074180603} 01/29/2022 11:09:06 - INFO - codeparrot_training - Step 17365: {'lr': 0.00038388761515625815, 'samples': 3334272, 'steps': 17365, 'loss/train': 2.053507089614868} 01/29/2022 11:09:10 - INFO - codeparrot_training - Step 17366: {'lr': 0.0003838737967177271, 'samples': 3334464, 'steps': 17366, 'loss/train': 1.3290648758411407} 01/29/2022 11:09:15 - INFO - codeparrot_training - Step 17367: {'lr': 0.00038385997770572336, 'samples': 3334656, 'steps': 17367, 'loss/train': 3.087788701057434} 01/29/2022 11:09:22 - INFO - codeparrot_training - Step 17368: {'lr': 0.0003838461581203058, 'samples': 3334848, 'steps': 17368, 'loss/train': 1.4824360013008118} 01/29/2022 11:09:26 - INFO - codeparrot_training - Step 17369: {'lr': 0.00038383233796153383, 'samples': 3335040, 'steps': 17369, 'loss/train': 0.9873517155647278} 01/29/2022 11:09:30 - INFO - codeparrot_training - Step 17370: {'lr': 0.00038381851722946663, 'samples': 3335232, 'steps': 17370, 'loss/train': 1.9972123503684998} 01/29/2022 11:09:34 - INFO - codeparrot_training - Step 17371: {'lr': 0.00038380469592416347, 'samples': 3335424, 'steps': 17371, 'loss/train': 1.6507514119148254} 01/29/2022 11:09:39 - INFO - codeparrot_training - Step 17372: {'lr': 0.00038379087404568333, 'samples': 3335616, 'steps': 17372, 'loss/train': 1.7238966822624207} 01/29/2022 11:09:44 - INFO - codeparrot_training - Step 17373: {'lr': 0.0003837770515940857, 'samples': 3335808, 'steps': 17373, 'loss/train': 1.2163143754005432} 01/29/2022 11:09:48 - INFO - codeparrot_training - Step 17374: {'lr': 0.0003837632285694296, 'samples': 3336000, 'steps': 17374, 'loss/train': 1.3615264892578125} 01/29/2022 11:09:53 - INFO - codeparrot_training - Step 17375: {'lr': 0.00038374940497177434, 'samples': 3336192, 'steps': 17375, 'loss/train': 2.421557664871216} 01/29/2022 11:09:57 - INFO - codeparrot_training - Step 17376: {'lr': 0.000383735580801179, 'samples': 3336384, 'steps': 17376, 'loss/train': 1.9491975903511047} 01/29/2022 11:10:01 - INFO - codeparrot_training - Step 17377: {'lr': 0.00038372175605770305, 'samples': 3336576, 'steps': 17377, 'loss/train': 1.3652218580245972} 01/29/2022 11:10:08 - INFO - codeparrot_training - Step 17378: {'lr': 0.00038370793074140545, 'samples': 3336768, 'steps': 17378, 'loss/train': 1.7333707809448242} 01/29/2022 11:10:13 - INFO - codeparrot_training - Step 17379: {'lr': 0.00038369410485234557, 'samples': 3336960, 'steps': 17379, 'loss/train': 1.4319947361946106} 01/29/2022 11:10:17 - INFO - codeparrot_training - Step 17380: {'lr': 0.0003836802783905826, 'samples': 3337152, 'steps': 17380, 'loss/train': 2.34954571723938} 01/29/2022 11:10:21 - INFO - codeparrot_training - Step 17381: {'lr': 0.0003836664513561758, 'samples': 3337344, 'steps': 17381, 'loss/train': 1.3466216325759888} 01/29/2022 11:10:25 - INFO - codeparrot_training - Step 17382: {'lr': 0.0003836526237491843, 'samples': 3337536, 'steps': 17382, 'loss/train': 0.7096752673387527} 01/29/2022 11:10:31 - INFO - codeparrot_training - Step 17383: {'lr': 0.0003836387955696674, 'samples': 3337728, 'steps': 17383, 'loss/train': 1.8865297436714172} 01/29/2022 11:10:35 - INFO - codeparrot_training - Step 17384: {'lr': 0.00038362496681768434, 'samples': 3337920, 'steps': 17384, 'loss/train': 2.2765971422195435} 01/29/2022 11:10:39 - INFO - codeparrot_training - Step 17385: {'lr': 0.00038361113749329443, 'samples': 3338112, 'steps': 17385, 'loss/train': 1.5017067790031433} 01/29/2022 11:10:43 - INFO - codeparrot_training - Step 17386: {'lr': 0.00038359730759655674, 'samples': 3338304, 'steps': 17386, 'loss/train': 2.3486157059669495} 01/29/2022 11:10:48 - INFO - codeparrot_training - Step 17387: {'lr': 0.00038358347712753063, 'samples': 3338496, 'steps': 17387, 'loss/train': 1.2275963723659515} 01/29/2022 11:10:53 - INFO - codeparrot_training - Step 17388: {'lr': 0.0003835696460862753, 'samples': 3338688, 'steps': 17388, 'loss/train': 2.023470461368561} 01/29/2022 11:10:57 - INFO - codeparrot_training - Step 17389: {'lr': 0.00038355581447285005, 'samples': 3338880, 'steps': 17389, 'loss/train': 2.163162410259247} 01/29/2022 11:11:01 - INFO - codeparrot_training - Step 17390: {'lr': 0.00038354198228731414, 'samples': 3339072, 'steps': 17390, 'loss/train': 1.8556053042411804} 01/29/2022 11:11:06 - INFO - codeparrot_training - Step 17391: {'lr': 0.0003835281495297267, 'samples': 3339264, 'steps': 17391, 'loss/train': 1.7246105074882507} 01/29/2022 11:11:10 - INFO - codeparrot_training - Step 17392: {'lr': 0.0003835143162001472, 'samples': 3339456, 'steps': 17392, 'loss/train': 1.5103606581687927} 01/29/2022 11:11:17 - INFO - codeparrot_training - Step 17393: {'lr': 0.0003835004822986346, 'samples': 3339648, 'steps': 17393, 'loss/train': 1.0682519674301147} 01/29/2022 11:11:21 - INFO - codeparrot_training - Step 17394: {'lr': 0.00038348664782524846, 'samples': 3339840, 'steps': 17394, 'loss/train': 2.5401281118392944} 01/29/2022 11:11:25 - INFO - codeparrot_training - Step 17395: {'lr': 0.00038347281278004774, 'samples': 3340032, 'steps': 17395, 'loss/train': 1.4841105043888092} 01/29/2022 11:11:30 - INFO - codeparrot_training - Step 17396: {'lr': 0.0003834589771630921, 'samples': 3340224, 'steps': 17396, 'loss/train': 2.084425449371338} 01/29/2022 11:11:34 - INFO - codeparrot_training - Step 17397: {'lr': 0.0003834451409744404, 'samples': 3340416, 'steps': 17397, 'loss/train': 2.2099993228912354} 01/29/2022 11:11:39 - INFO - codeparrot_training - Step 17398: {'lr': 0.0003834313042141522, 'samples': 3340608, 'steps': 17398, 'loss/train': 1.7619928121566772} 01/29/2022 11:11:43 - INFO - codeparrot_training - Step 17399: {'lr': 0.0003834174668822865, 'samples': 3340800, 'steps': 17399, 'loss/train': 1.770680844783783} 01/29/2022 11:11:48 - INFO - codeparrot_training - Step 17400: {'lr': 0.0003834036289789029, 'samples': 3340992, 'steps': 17400, 'loss/train': 1.3709571361541748} 01/29/2022 11:11:52 - INFO - codeparrot_training - Step 17401: {'lr': 0.0003833897905040604, 'samples': 3341184, 'steps': 17401, 'loss/train': 1.7961452007293701} 01/29/2022 11:11:56 - INFO - codeparrot_training - Step 17402: {'lr': 0.00038337595145781844, 'samples': 3341376, 'steps': 17402, 'loss/train': 1.77364182472229} 01/29/2022 11:12:02 - INFO - codeparrot_training - Step 17403: {'lr': 0.00038336211184023634, 'samples': 3341568, 'steps': 17403, 'loss/train': 0.6753454059362411} 01/29/2022 11:12:06 - INFO - codeparrot_training - Step 17404: {'lr': 0.0003833482716513732, 'samples': 3341760, 'steps': 17404, 'loss/train': 1.0501681566238403} 01/29/2022 11:12:10 - INFO - codeparrot_training - Step 17405: {'lr': 0.0003833344308912885, 'samples': 3341952, 'steps': 17405, 'loss/train': 1.9733623266220093} 01/29/2022 11:12:15 - INFO - codeparrot_training - Step 17406: {'lr': 0.00038332058956004134, 'samples': 3342144, 'steps': 17406, 'loss/train': 1.7546408772468567} 01/29/2022 11:12:19 - INFO - codeparrot_training - Step 17407: {'lr': 0.0003833067476576911, 'samples': 3342336, 'steps': 17407, 'loss/train': 1.819142460823059} 01/29/2022 11:12:24 - INFO - codeparrot_training - Step 17408: {'lr': 0.0003832929051842972, 'samples': 3342528, 'steps': 17408, 'loss/train': 1.890596866607666} 01/29/2022 11:12:28 - INFO - codeparrot_training - Step 17409: {'lr': 0.0003832790621399187, 'samples': 3342720, 'steps': 17409, 'loss/train': 1.44879949092865} 01/29/2022 11:12:32 - INFO - codeparrot_training - Step 17410: {'lr': 0.00038326521852461505, 'samples': 3342912, 'steps': 17410, 'loss/train': 2.3454546332359314} 01/29/2022 11:12:37 - INFO - codeparrot_training - Step 17411: {'lr': 0.0003832513743384456, 'samples': 3343104, 'steps': 17411, 'loss/train': 1.9065855145454407} 01/29/2022 11:12:41 - INFO - codeparrot_training - Step 17412: {'lr': 0.0003832375295814695, 'samples': 3343296, 'steps': 17412, 'loss/train': 2.0820520520210266} 01/29/2022 11:12:48 - INFO - codeparrot_training - Step 17413: {'lr': 0.0003832236842537461, 'samples': 3343488, 'steps': 17413, 'loss/train': 2.109977602958679} 01/29/2022 11:12:53 - INFO - codeparrot_training - Step 17414: {'lr': 0.0003832098383553347, 'samples': 3343680, 'steps': 17414, 'loss/train': 1.7225253582000732} 01/29/2022 11:12:57 - INFO - codeparrot_training - Step 17415: {'lr': 0.00038319599188629485, 'samples': 3343872, 'steps': 17415, 'loss/train': 1.215823084115982} 01/29/2022 11:13:01 - INFO - codeparrot_training - Step 17416: {'lr': 0.00038318214484668557, 'samples': 3344064, 'steps': 17416, 'loss/train': 1.1320513486862183} 01/29/2022 11:13:05 - INFO - codeparrot_training - Step 17417: {'lr': 0.0003831682972365662, 'samples': 3344256, 'steps': 17417, 'loss/train': 1.0140132308006287} 01/29/2022 11:13:11 - INFO - codeparrot_training - Step 17418: {'lr': 0.0003831544490559962, 'samples': 3344448, 'steps': 17418, 'loss/train': 1.944778561592102} 01/29/2022 11:13:15 - INFO - codeparrot_training - Step 17419: {'lr': 0.00038314060030503476, 'samples': 3344640, 'steps': 17419, 'loss/train': 0.9598752558231354} 01/29/2022 11:13:19 - INFO - codeparrot_training - Step 17420: {'lr': 0.00038312675098374136, 'samples': 3344832, 'steps': 17420, 'loss/train': 1.3469472527503967} 01/29/2022 11:13:24 - INFO - codeparrot_training - Step 17421: {'lr': 0.0003831129010921751, 'samples': 3345024, 'steps': 17421, 'loss/train': 1.4930753409862518} 01/29/2022 11:13:28 - INFO - codeparrot_training - Step 17422: {'lr': 0.0003830990506303956, 'samples': 3345216, 'steps': 17422, 'loss/train': 1.531470537185669} 01/29/2022 11:13:35 - INFO - codeparrot_training - Step 17423: {'lr': 0.0003830851995984619, 'samples': 3345408, 'steps': 17423, 'loss/train': 2.0096311569213867} 01/29/2022 11:13:40 - INFO - codeparrot_training - Step 17424: {'lr': 0.0003830713479964335, 'samples': 3345600, 'steps': 17424, 'loss/train': 1.6990259885787964} 01/29/2022 11:13:44 - INFO - codeparrot_training - Step 17425: {'lr': 0.0003830574958243697, 'samples': 3345792, 'steps': 17425, 'loss/train': 2.0558305978775024} 01/29/2022 11:13:48 - INFO - codeparrot_training - Step 17426: {'lr': 0.00038304364308232986, 'samples': 3345984, 'steps': 17426, 'loss/train': 2.0504903197288513} 01/29/2022 11:13:52 - INFO - codeparrot_training - Step 17427: {'lr': 0.0003830297897703733, 'samples': 3346176, 'steps': 17427, 'loss/train': 0.8618507981300354} 01/29/2022 11:13:58 - INFO - codeparrot_training - Step 17428: {'lr': 0.0003830159358885593, 'samples': 3346368, 'steps': 17428, 'loss/train': 3.2181812524795532} 01/29/2022 11:14:02 - INFO - codeparrot_training - Step 17429: {'lr': 0.00038300208143694737, 'samples': 3346560, 'steps': 17429, 'loss/train': 0.9249376058578491} 01/29/2022 11:14:06 - INFO - codeparrot_training - Step 17430: {'lr': 0.00038298822641559673, 'samples': 3346752, 'steps': 17430, 'loss/train': 1.5338433980941772} 01/29/2022 11:14:10 - INFO - codeparrot_training - Step 17431: {'lr': 0.0003829743708245667, 'samples': 3346944, 'steps': 17431, 'loss/train': 0.09528439119458199} 01/29/2022 11:14:15 - INFO - codeparrot_training - Step 17432: {'lr': 0.0003829605146639167, 'samples': 3347136, 'steps': 17432, 'loss/train': 2.1030526757240295} 01/29/2022 11:14:19 - INFO - codeparrot_training - Step 17433: {'lr': 0.0003829466579337061, 'samples': 3347328, 'steps': 17433, 'loss/train': 0.7805485725402832} 01/29/2022 11:14:24 - INFO - codeparrot_training - Step 17434: {'lr': 0.00038293280063399427, 'samples': 3347520, 'steps': 17434, 'loss/train': 0.9474246203899384} 01/29/2022 11:14:29 - INFO - codeparrot_training - Step 17435: {'lr': 0.00038291894276484053, 'samples': 3347712, 'steps': 17435, 'loss/train': 0.8313876092433929} 01/29/2022 11:14:33 - INFO - codeparrot_training - Step 17436: {'lr': 0.0003829050843263041, 'samples': 3347904, 'steps': 17436, 'loss/train': 2.0053610801696777} 01/29/2022 11:14:37 - INFO - codeparrot_training - Step 17437: {'lr': 0.0003828912253184446, 'samples': 3348096, 'steps': 17437, 'loss/train': 0.6983019411563873} 01/29/2022 11:14:41 - INFO - codeparrot_training - Step 17438: {'lr': 0.0003828773657413213, 'samples': 3348288, 'steps': 17438, 'loss/train': 2.208992063999176} 01/29/2022 11:14:48 - INFO - codeparrot_training - Step 17439: {'lr': 0.0003828635055949935, 'samples': 3348480, 'steps': 17439, 'loss/train': 1.1750013828277588} 01/29/2022 11:14:53 - INFO - codeparrot_training - Step 17440: {'lr': 0.0003828496448795207, 'samples': 3348672, 'steps': 17440, 'loss/train': 2.9827374815940857} 01/29/2022 11:14:57 - INFO - codeparrot_training - Step 17441: {'lr': 0.0003828357835949622, 'samples': 3348864, 'steps': 17441, 'loss/train': 1.5980197191238403} 01/29/2022 11:15:01 - INFO - codeparrot_training - Step 17442: {'lr': 0.00038282192174137744, 'samples': 3349056, 'steps': 17442, 'loss/train': 1.729647159576416} 01/29/2022 11:15:05 - INFO - codeparrot_training - Step 17443: {'lr': 0.00038280805931882557, 'samples': 3349248, 'steps': 17443, 'loss/train': 1.7964704632759094} 01/29/2022 11:15:11 - INFO - codeparrot_training - Step 17444: {'lr': 0.0003827941963273663, 'samples': 3349440, 'steps': 17444, 'loss/train': 2.012898802757263} 01/29/2022 11:15:15 - INFO - codeparrot_training - Step 17445: {'lr': 0.00038278033276705875, 'samples': 3349632, 'steps': 17445, 'loss/train': 1.2494190037250519} 01/29/2022 11:15:19 - INFO - codeparrot_training - Step 17446: {'lr': 0.0003827664686379625, 'samples': 3349824, 'steps': 17446, 'loss/train': 0.5374457985162735} 01/29/2022 11:15:23 - INFO - codeparrot_training - Step 17447: {'lr': 0.00038275260394013676, 'samples': 3350016, 'steps': 17447, 'loss/train': 1.7585334777832031} 01/29/2022 11:15:28 - INFO - codeparrot_training - Step 17448: {'lr': 0.0003827387386736411, 'samples': 3350208, 'steps': 17448, 'loss/train': 2.200250744819641} 01/29/2022 11:15:33 - INFO - codeparrot_training - Step 17449: {'lr': 0.0003827248728385349, 'samples': 3350400, 'steps': 17449, 'loss/train': 4.074919581413269} 01/29/2022 11:15:37 - INFO - codeparrot_training - Step 17450: {'lr': 0.0003827110064348773, 'samples': 3350592, 'steps': 17450, 'loss/train': 1.0208214819431305} 01/29/2022 11:15:41 - INFO - codeparrot_training - Step 17451: {'lr': 0.000382697139462728, 'samples': 3350784, 'steps': 17451, 'loss/train': 1.8550618886947632} 01/29/2022 11:15:46 - INFO - codeparrot_training - Step 17452: {'lr': 0.00038268327192214635, 'samples': 3350976, 'steps': 17452, 'loss/train': 1.6671949625015259} 01/29/2022 11:15:50 - INFO - codeparrot_training - Step 17453: {'lr': 0.0003826694038131916, 'samples': 3351168, 'steps': 17453, 'loss/train': 0.30637530237436295} 01/29/2022 11:15:55 - INFO - codeparrot_training - Step 17454: {'lr': 0.00038265553513592334, 'samples': 3351360, 'steps': 17454, 'loss/train': 1.5196856260299683} 01/29/2022 11:15:59 - INFO - codeparrot_training - Step 17455: {'lr': 0.00038264166589040084, 'samples': 3351552, 'steps': 17455, 'loss/train': 0.8878085017204285} 01/29/2022 11:16:04 - INFO - codeparrot_training - Step 17456: {'lr': 0.00038262779607668354, 'samples': 3351744, 'steps': 17456, 'loss/train': 1.9073029160499573} 01/29/2022 11:16:08 - INFO - codeparrot_training - Step 17457: {'lr': 0.00038261392569483087, 'samples': 3351936, 'steps': 17457, 'loss/train': 1.3921692073345184} 01/29/2022 11:16:12 - INFO - codeparrot_training - Step 17458: {'lr': 0.0003826000547449023, 'samples': 3352128, 'steps': 17458, 'loss/train': 1.8040223121643066} 01/29/2022 11:16:19 - INFO - codeparrot_training - Step 17459: {'lr': 0.0003825861832269571, 'samples': 3352320, 'steps': 17459, 'loss/train': 1.607316792011261} 01/29/2022 11:16:24 - INFO - codeparrot_training - Step 17460: {'lr': 0.00038257231114105495, 'samples': 3352512, 'steps': 17460, 'loss/train': 0.8815893530845642} 01/29/2022 11:16:28 - INFO - codeparrot_training - Step 17461: {'lr': 0.00038255843848725504, 'samples': 3352704, 'steps': 17461, 'loss/train': 1.8836637139320374} 01/29/2022 11:16:32 - INFO - codeparrot_training - Step 17462: {'lr': 0.0003825445652656169, 'samples': 3352896, 'steps': 17462, 'loss/train': 1.464695781469345} 01/29/2022 11:16:36 - INFO - codeparrot_training - Step 17463: {'lr': 0.00038253069147619977, 'samples': 3353088, 'steps': 17463, 'loss/train': 1.891726791858673} 01/29/2022 11:16:42 - INFO - codeparrot_training - Step 17464: {'lr': 0.00038251681711906345, 'samples': 3353280, 'steps': 17464, 'loss/train': 1.8139470219612122} 01/29/2022 11:16:46 - INFO - codeparrot_training - Step 17465: {'lr': 0.00038250294219426706, 'samples': 3353472, 'steps': 17465, 'loss/train': 1.0786367654800415} 01/29/2022 11:16:50 - INFO - codeparrot_training - Step 17466: {'lr': 0.00038248906670187017, 'samples': 3353664, 'steps': 17466, 'loss/train': 2.6586933732032776} 01/29/2022 11:16:54 - INFO - codeparrot_training - Step 17467: {'lr': 0.00038247519064193216, 'samples': 3353856, 'steps': 17467, 'loss/train': 2.085359573364258} 01/29/2022 11:16:59 - INFO - codeparrot_training - Step 17468: {'lr': 0.0003824613140145125, 'samples': 3354048, 'steps': 17468, 'loss/train': 1.4588617086410522} 01/29/2022 11:17:06 - INFO - codeparrot_training - Step 17469: {'lr': 0.00038244743681967066, 'samples': 3354240, 'steps': 17469, 'loss/train': 2.21147757768631} 01/29/2022 11:17:10 - INFO - codeparrot_training - Step 17470: {'lr': 0.000382433559057466, 'samples': 3354432, 'steps': 17470, 'loss/train': 1.7279242873191833} 01/29/2022 11:17:14 - INFO - codeparrot_training - Step 17471: {'lr': 0.00038241968072795805, 'samples': 3354624, 'steps': 17471, 'loss/train': 1.3244170546531677} 01/29/2022 11:17:18 - INFO - codeparrot_training - Step 17472: {'lr': 0.00038240580183120624, 'samples': 3354816, 'steps': 17472, 'loss/train': 0.1971646249294281} 01/29/2022 11:17:23 - INFO - codeparrot_training - Step 17473: {'lr': 0.0003823919223672701, 'samples': 3355008, 'steps': 17473, 'loss/train': 1.128189504146576} 01/29/2022 11:17:28 - INFO - codeparrot_training - Step 17474: {'lr': 0.00038237804233620887, 'samples': 3355200, 'steps': 17474, 'loss/train': 1.4313903450965881} 01/29/2022 11:17:32 - INFO - codeparrot_training - Step 17475: {'lr': 0.0003823641617380823, 'samples': 3355392, 'steps': 17475, 'loss/train': 1.5417537689208984} 01/29/2022 11:17:36 - INFO - codeparrot_training - Step 17476: {'lr': 0.00038235028057294953, 'samples': 3355584, 'steps': 17476, 'loss/train': 2.3351333141326904} 01/29/2022 11:17:41 - INFO - codeparrot_training - Step 17477: {'lr': 0.0003823363988408703, 'samples': 3355776, 'steps': 17477, 'loss/train': 1.7604385614395142} 01/29/2022 11:17:45 - INFO - codeparrot_training - Step 17478: {'lr': 0.00038232251654190386, 'samples': 3355968, 'steps': 17478, 'loss/train': 1.58789724111557} 01/29/2022 11:17:50 - INFO - codeparrot_training - Step 17479: {'lr': 0.0003823086336761099, 'samples': 3356160, 'steps': 17479, 'loss/train': 2.423803925514221} 01/29/2022 11:17:54 - INFO - codeparrot_training - Step 17480: {'lr': 0.00038229475024354766, 'samples': 3356352, 'steps': 17480, 'loss/train': 2.055462598800659} 01/29/2022 11:17:59 - INFO - codeparrot_training - Step 17481: {'lr': 0.00038228086624427675, 'samples': 3356544, 'steps': 17481, 'loss/train': 1.4254381656646729} 01/29/2022 11:18:03 - INFO - codeparrot_training - Step 17482: {'lr': 0.0003822669816783566, 'samples': 3356736, 'steps': 17482, 'loss/train': 1.1837694346904755} 01/29/2022 11:18:07 - INFO - codeparrot_training - Step 17483: {'lr': 0.0003822530965458467, 'samples': 3356928, 'steps': 17483, 'loss/train': 1.3762933015823364} 01/29/2022 11:18:14 - INFO - codeparrot_training - Step 17484: {'lr': 0.0003822392108468066, 'samples': 3357120, 'steps': 17484, 'loss/train': 1.5708544850349426} 01/29/2022 11:18:19 - INFO - codeparrot_training - Step 17485: {'lr': 0.00038222532458129563, 'samples': 3357312, 'steps': 17485, 'loss/train': 2.1244558095932007} 01/29/2022 11:18:23 - INFO - codeparrot_training - Step 17486: {'lr': 0.0003822114377493734, 'samples': 3357504, 'steps': 17486, 'loss/train': 1.3040986955165863} 01/29/2022 11:18:27 - INFO - codeparrot_training - Step 17487: {'lr': 0.0003821975503510993, 'samples': 3357696, 'steps': 17487, 'loss/train': 0.6690208911895752} 01/29/2022 11:18:33 - INFO - codeparrot_training - Step 17488: {'lr': 0.0003821836623865329, 'samples': 3357888, 'steps': 17488, 'loss/train': 0.8735794723033905} 01/29/2022 11:18:37 - INFO - codeparrot_training - Step 17489: {'lr': 0.0003821697738557337, 'samples': 3358080, 'steps': 17489, 'loss/train': 1.357588291168213} 01/29/2022 11:18:41 - INFO - codeparrot_training - Step 17490: {'lr': 0.00038215588475876117, 'samples': 3358272, 'steps': 17490, 'loss/train': 1.9520722031593323} 01/29/2022 11:18:45 - INFO - codeparrot_training - Step 17491: {'lr': 0.0003821419950956747, 'samples': 3358464, 'steps': 17491, 'loss/train': 1.5339218974113464} 01/29/2022 11:18:50 - INFO - codeparrot_training - Step 17492: {'lr': 0.00038212810486653394, 'samples': 3358656, 'steps': 17492, 'loss/train': 0.2873241752386093} 01/29/2022 11:18:54 - INFO - codeparrot_training - Step 17493: {'lr': 0.0003821142140713983, 'samples': 3358848, 'steps': 17493, 'loss/train': 1.231101393699646} 01/29/2022 11:19:01 - INFO - codeparrot_training - Step 17494: {'lr': 0.0003821003227103274, 'samples': 3359040, 'steps': 17494, 'loss/train': 1.4643441438674927} 01/29/2022 11:19:05 - INFO - codeparrot_training - Step 17495: {'lr': 0.00038208643078338055, 'samples': 3359232, 'steps': 17495, 'loss/train': 1.1840828955173492} 01/29/2022 11:19:09 - INFO - codeparrot_training - Step 17496: {'lr': 0.0003820725382906175, 'samples': 3359424, 'steps': 17496, 'loss/train': 1.8158061504364014} 01/29/2022 11:19:14 - INFO - codeparrot_training - Step 17497: {'lr': 0.0003820586452320975, 'samples': 3359616, 'steps': 17497, 'loss/train': 1.9240710139274597} 01/29/2022 11:19:18 - INFO - codeparrot_training - Step 17498: {'lr': 0.0003820447516078803, 'samples': 3359808, 'steps': 17498, 'loss/train': 1.460592359304428} 01/29/2022 11:19:23 - INFO - codeparrot_training - Step 17499: {'lr': 0.0003820308574180253, 'samples': 3360000, 'steps': 17499, 'loss/train': 1.564569890499115} 01/29/2022 11:19:28 - INFO - codeparrot_training - Step 17500: {'lr': 0.000382016962662592, 'samples': 3360192, 'steps': 17500, 'loss/train': 2.0434656143188477} 01/29/2022 11:19:32 - INFO - codeparrot_training - Step 17501: {'lr': 0.0003820030673416399, 'samples': 3360384, 'steps': 17501, 'loss/train': 2.424527406692505} 01/29/2022 11:19:36 - INFO - codeparrot_training - Step 17502: {'lr': 0.0003819891714552287, 'samples': 3360576, 'steps': 17502, 'loss/train': 1.9205777049064636} 01/29/2022 11:19:40 - INFO - codeparrot_training - Step 17503: {'lr': 0.00038197527500341777, 'samples': 3360768, 'steps': 17503, 'loss/train': 2.2325302362442017} 01/29/2022 11:19:46 - INFO - codeparrot_training - Step 17504: {'lr': 0.00038196137798626663, 'samples': 3360960, 'steps': 17504, 'loss/train': 1.9534213542938232} 01/29/2022 11:19:50 - INFO - codeparrot_training - Step 17505: {'lr': 0.00038194748040383487, 'samples': 3361152, 'steps': 17505, 'loss/train': 2.606542468070984} 01/29/2022 11:19:54 - INFO - codeparrot_training - Step 17506: {'lr': 0.00038193358225618195, 'samples': 3361344, 'steps': 17506, 'loss/train': 1.7159156799316406} 01/29/2022 11:19:58 - INFO - codeparrot_training - Step 17507: {'lr': 0.0003819196835433675, 'samples': 3361536, 'steps': 17507, 'loss/train': 1.4366327226161957} 01/29/2022 11:20:03 - INFO - codeparrot_training - Step 17508: {'lr': 0.000381905784265451, 'samples': 3361728, 'steps': 17508, 'loss/train': 2.7367916107177734} 01/29/2022 11:20:08 - INFO - codeparrot_training - Step 17509: {'lr': 0.000381891884422492, 'samples': 3361920, 'steps': 17509, 'loss/train': 1.8920245170593262} 01/29/2022 11:20:12 - INFO - codeparrot_training - Step 17510: {'lr': 0.0003818779840145501, 'samples': 3362112, 'steps': 17510, 'loss/train': 1.9407007098197937} 01/29/2022 11:20:17 - INFO - codeparrot_training - Step 17511: {'lr': 0.00038186408304168474, 'samples': 3362304, 'steps': 17511, 'loss/train': 0.18493368476629257} 01/29/2022 11:20:21 - INFO - codeparrot_training - Step 17512: {'lr': 0.00038185018150395557, 'samples': 3362496, 'steps': 17512, 'loss/train': 1.6433565616607666} 01/29/2022 11:20:25 - INFO - codeparrot_training - Step 17513: {'lr': 0.000381836279401422, 'samples': 3362688, 'steps': 17513, 'loss/train': 1.0022537112236023} 01/29/2022 11:20:32 - INFO - codeparrot_training - Step 17514: {'lr': 0.00038182237673414375, 'samples': 3362880, 'steps': 17514, 'loss/train': 2.8584614396095276} 01/29/2022 11:20:36 - INFO - codeparrot_training - Step 17515: {'lr': 0.0003818084735021803, 'samples': 3363072, 'steps': 17515, 'loss/train': 1.8361566066741943} 01/29/2022 11:20:41 - INFO - codeparrot_training - Step 17516: {'lr': 0.00038179456970559116, 'samples': 3363264, 'steps': 17516, 'loss/train': 0.2956957519054413} 01/29/2022 11:20:45 - INFO - codeparrot_training - Step 17517: {'lr': 0.00038178066534443587, 'samples': 3363456, 'steps': 17517, 'loss/train': 1.8443264365196228} 01/29/2022 11:20:49 - INFO - codeparrot_training - Step 17518: {'lr': 0.00038176676041877424, 'samples': 3363648, 'steps': 17518, 'loss/train': 1.1212809383869171} 01/29/2022 11:20:54 - INFO - codeparrot_training - Step 17519: {'lr': 0.0003817528549286655, 'samples': 3363840, 'steps': 17519, 'loss/train': 2.4442076683044434} 01/29/2022 11:20:59 - INFO - codeparrot_training - Step 17520: {'lr': 0.00038173894887416946, 'samples': 3364032, 'steps': 17520, 'loss/train': 1.258743703365326} 01/29/2022 11:21:03 - INFO - codeparrot_training - Step 17521: {'lr': 0.0003817250422553455, 'samples': 3364224, 'steps': 17521, 'loss/train': 1.7207955121994019} 01/29/2022 11:21:07 - INFO - codeparrot_training - Step 17522: {'lr': 0.0003817111350722533, 'samples': 3364416, 'steps': 17522, 'loss/train': 1.6335124969482422} 01/29/2022 11:21:11 - INFO - codeparrot_training - Step 17523: {'lr': 0.0003816972273249525, 'samples': 3364608, 'steps': 17523, 'loss/train': 1.8629334568977356} 01/29/2022 11:21:17 - INFO - codeparrot_training - Step 17524: {'lr': 0.00038168331901350253, 'samples': 3364800, 'steps': 17524, 'loss/train': 2.356639087200165} 01/29/2022 11:21:21 - INFO - codeparrot_training - Step 17525: {'lr': 0.0003816694101379631, 'samples': 3364992, 'steps': 17525, 'loss/train': 1.728572130203247} 01/29/2022 11:21:25 - INFO - codeparrot_training - Step 17526: {'lr': 0.0003816555006983936, 'samples': 3365184, 'steps': 17526, 'loss/train': 2.4855018854141235} 01/29/2022 11:21:30 - INFO - codeparrot_training - Step 17527: {'lr': 0.0003816415906948538, 'samples': 3365376, 'steps': 17527, 'loss/train': 1.3352742791175842} 01/29/2022 11:21:34 - INFO - codeparrot_training - Step 17528: {'lr': 0.00038162768012740323, 'samples': 3365568, 'steps': 17528, 'loss/train': 1.8906058073043823} 01/29/2022 11:21:41 - INFO - codeparrot_training - Step 17529: {'lr': 0.00038161376899610154, 'samples': 3365760, 'steps': 17529, 'loss/train': 1.3058178126811981} 01/29/2022 11:21:45 - INFO - codeparrot_training - Step 17530: {'lr': 0.0003815998573010082, 'samples': 3365952, 'steps': 17530, 'loss/train': 1.4749209880828857} 01/29/2022 11:21:49 - INFO - codeparrot_training - Step 17531: {'lr': 0.0003815859450421829, 'samples': 3366144, 'steps': 17531, 'loss/train': 2.0898399353027344} 01/29/2022 11:21:54 - INFO - codeparrot_training - Step 17532: {'lr': 0.00038157203221968514, 'samples': 3366336, 'steps': 17532, 'loss/train': 0.8854266107082367} 01/29/2022 11:21:58 - INFO - codeparrot_training - Step 17533: {'lr': 0.00038155811883357454, 'samples': 3366528, 'steps': 17533, 'loss/train': 1.7514908909797668} 01/29/2022 11:22:03 - INFO - codeparrot_training - Step 17534: {'lr': 0.0003815442048839108, 'samples': 3366720, 'steps': 17534, 'loss/train': 1.318922370672226} 01/29/2022 11:22:07 - INFO - codeparrot_training - Step 17535: {'lr': 0.0003815302903707534, 'samples': 3366912, 'steps': 17535, 'loss/train': 1.5486509799957275} 01/29/2022 11:22:12 - INFO - codeparrot_training - Step 17536: {'lr': 0.0003815163752941621, 'samples': 3367104, 'steps': 17536, 'loss/train': 1.7825220823287964} 01/29/2022 11:22:16 - INFO - codeparrot_training - Step 17537: {'lr': 0.00038150245965419636, 'samples': 3367296, 'steps': 17537, 'loss/train': 1.4314967393875122} 01/29/2022 11:22:20 - INFO - codeparrot_training - Step 17538: {'lr': 0.0003814885434509158, 'samples': 3367488, 'steps': 17538, 'loss/train': 2.2431371212005615} 01/29/2022 11:22:27 - INFO - codeparrot_training - Step 17539: {'lr': 0.0003814746266843801, 'samples': 3367680, 'steps': 17539, 'loss/train': 1.3589124083518982} 01/29/2022 11:22:31 - INFO - codeparrot_training - Step 17540: {'lr': 0.0003814607093546489, 'samples': 3367872, 'steps': 17540, 'loss/train': 1.6304644346237183} 01/29/2022 11:22:36 - INFO - codeparrot_training - Step 17541: {'lr': 0.00038144679146178166, 'samples': 3368064, 'steps': 17541, 'loss/train': 1.9637115597724915} 01/29/2022 11:22:40 - INFO - codeparrot_training - Step 17542: {'lr': 0.00038143287300583816, 'samples': 3368256, 'steps': 17542, 'loss/train': 0.8294963836669922} 01/29/2022 11:22:44 - INFO - codeparrot_training - Step 17543: {'lr': 0.00038141895398687806, 'samples': 3368448, 'steps': 17543, 'loss/train': 2.2209935188293457} 01/29/2022 11:22:50 - INFO - codeparrot_training - Step 17544: {'lr': 0.0003814050344049608, 'samples': 3368640, 'steps': 17544, 'loss/train': 2.494365155696869} 01/29/2022 11:22:54 - INFO - codeparrot_training - Step 17545: {'lr': 0.00038139111426014607, 'samples': 3368832, 'steps': 17545, 'loss/train': 1.8198092579841614} 01/29/2022 11:22:58 - INFO - codeparrot_training - Step 17546: {'lr': 0.00038137719355249355, 'samples': 3369024, 'steps': 17546, 'loss/train': 1.2068096995353699} 01/29/2022 11:23:02 - INFO - codeparrot_training - Step 17547: {'lr': 0.00038136327228206285, 'samples': 3369216, 'steps': 17547, 'loss/train': 0.4820363223552704} 01/29/2022 11:23:07 - INFO - codeparrot_training - Step 17548: {'lr': 0.0003813493504489136, 'samples': 3369408, 'steps': 17548, 'loss/train': 1.9257667064666748} 01/29/2022 11:23:12 - INFO - codeparrot_training - Step 17549: {'lr': 0.0003813354280531055, 'samples': 3369600, 'steps': 17549, 'loss/train': 0.7461222857236862} 01/29/2022 11:23:16 - INFO - codeparrot_training - Step 17550: {'lr': 0.00038132150509469806, 'samples': 3369792, 'steps': 17550, 'loss/train': 0.9359439611434937} 01/29/2022 11:23:20 - INFO - codeparrot_training - Step 17551: {'lr': 0.000381307581573751, 'samples': 3369984, 'steps': 17551, 'loss/train': 1.755483090877533} 01/29/2022 11:23:25 - INFO - codeparrot_training - Step 17552: {'lr': 0.00038129365749032395, 'samples': 3370176, 'steps': 17552, 'loss/train': 0.7258264124393463} 01/29/2022 11:23:29 - INFO - codeparrot_training - Step 17553: {'lr': 0.0003812797328444766, 'samples': 3370368, 'steps': 17553, 'loss/train': 1.1861618757247925} 01/29/2022 11:23:36 - INFO - codeparrot_training - Step 17554: {'lr': 0.0003812658076362685, 'samples': 3370560, 'steps': 17554, 'loss/train': 1.113831102848053} 01/29/2022 11:23:40 - INFO - codeparrot_training - Step 17555: {'lr': 0.00038125188186575944, 'samples': 3370752, 'steps': 17555, 'loss/train': 2.1691015362739563} 01/29/2022 11:23:45 - INFO - codeparrot_training - Step 17556: {'lr': 0.00038123795553300893, 'samples': 3370944, 'steps': 17556, 'loss/train': 1.254064679145813} 01/29/2022 11:23:49 - INFO - codeparrot_training - Step 17557: {'lr': 0.0003812240286380767, 'samples': 3371136, 'steps': 17557, 'loss/train': 1.296345055103302} 01/29/2022 11:23:53 - INFO - codeparrot_training - Step 17558: {'lr': 0.0003812101011810224, 'samples': 3371328, 'steps': 17558, 'loss/train': 1.4404080212116241} 01/29/2022 11:23:58 - INFO - codeparrot_training - Step 17559: {'lr': 0.0003811961731619057, 'samples': 3371520, 'steps': 17559, 'loss/train': 1.0555340051651} 01/29/2022 11:24:03 - INFO - codeparrot_training - Step 17560: {'lr': 0.0003811822445807863, 'samples': 3371712, 'steps': 17560, 'loss/train': 1.6344947218894958} 01/29/2022 11:24:07 - INFO - codeparrot_training - Step 17561: {'lr': 0.00038116831543772377, 'samples': 3371904, 'steps': 17561, 'loss/train': 2.3309220671653748} 01/29/2022 11:24:11 - INFO - codeparrot_training - Step 17562: {'lr': 0.00038115438573277784, 'samples': 3372096, 'steps': 17562, 'loss/train': 1.3839837312698364} 01/29/2022 11:24:15 - INFO - codeparrot_training - Step 17563: {'lr': 0.0003811404554660082, 'samples': 3372288, 'steps': 17563, 'loss/train': 1.8742371797561646} 01/29/2022 11:24:21 - INFO - codeparrot_training - Step 17564: {'lr': 0.00038112652463747444, 'samples': 3372480, 'steps': 17564, 'loss/train': 1.4661870896816254} 01/29/2022 11:24:25 - INFO - codeparrot_training - Step 17565: {'lr': 0.00038111259324723624, 'samples': 3372672, 'steps': 17565, 'loss/train': 1.491075038909912} 01/29/2022 11:24:29 - INFO - codeparrot_training - Step 17566: {'lr': 0.0003810986612953534, 'samples': 3372864, 'steps': 17566, 'loss/train': 1.5378878116607666} 01/29/2022 11:24:33 - INFO - codeparrot_training - Step 17567: {'lr': 0.0003810847287818855, 'samples': 3373056, 'steps': 17567, 'loss/train': 1.3138062357902527} 01/29/2022 11:24:38 - INFO - codeparrot_training - Step 17568: {'lr': 0.0003810707957068923, 'samples': 3373248, 'steps': 17568, 'loss/train': 1.8258883953094482} 01/29/2022 11:24:43 - INFO - codeparrot_training - Step 17569: {'lr': 0.0003810568620704334, 'samples': 3373440, 'steps': 17569, 'loss/train': 2.480765640735626} 01/29/2022 11:24:47 - INFO - codeparrot_training - Step 17570: {'lr': 0.00038104292787256844, 'samples': 3373632, 'steps': 17570, 'loss/train': 1.398542046546936} 01/29/2022 11:24:52 - INFO - codeparrot_training - Step 17571: {'lr': 0.0003810289931133573, 'samples': 3373824, 'steps': 17571, 'loss/train': 0.993572473526001} 01/29/2022 11:24:56 - INFO - codeparrot_training - Step 17572: {'lr': 0.0003810150577928595, 'samples': 3374016, 'steps': 17572, 'loss/train': 1.3959875106811523} 01/29/2022 11:25:00 - INFO - codeparrot_training - Step 17573: {'lr': 0.0003810011219111348, 'samples': 3374208, 'steps': 17573, 'loss/train': 1.813166320323944} 01/29/2022 11:25:08 - INFO - codeparrot_training - Step 17574: {'lr': 0.00038098718546824287, 'samples': 3374400, 'steps': 17574, 'loss/train': 0.6806096434593201} 01/29/2022 11:25:12 - INFO - codeparrot_training - Step 17575: {'lr': 0.00038097324846424354, 'samples': 3374592, 'steps': 17575, 'loss/train': 1.7437103390693665} 01/29/2022 11:25:16 - INFO - codeparrot_training - Step 17576: {'lr': 0.0003809593108991962, 'samples': 3374784, 'steps': 17576, 'loss/train': 1.5834268927574158} 01/29/2022 11:25:21 - INFO - codeparrot_training - Step 17577: {'lr': 0.0003809453727731609, 'samples': 3374976, 'steps': 17577, 'loss/train': 2.044576942920685} 01/29/2022 11:25:25 - INFO - codeparrot_training - Step 17578: {'lr': 0.00038093143408619726, 'samples': 3375168, 'steps': 17578, 'loss/train': 2.010015606880188} 01/29/2022 11:25:30 - INFO - codeparrot_training - Step 17579: {'lr': 0.0003809174948383648, 'samples': 3375360, 'steps': 17579, 'loss/train': 1.3075015246868134} 01/29/2022 11:25:35 - INFO - codeparrot_training - Step 17580: {'lr': 0.0003809035550297234, 'samples': 3375552, 'steps': 17580, 'loss/train': 1.9742758870124817} 01/29/2022 11:25:39 - INFO - codeparrot_training - Step 17581: {'lr': 0.00038088961466033276, 'samples': 3375744, 'steps': 17581, 'loss/train': 1.3943225741386414} 01/29/2022 11:25:43 - INFO - codeparrot_training - Step 17582: {'lr': 0.00038087567373025255, 'samples': 3375936, 'steps': 17582, 'loss/train': 1.6056033968925476} 01/29/2022 11:25:47 - INFO - codeparrot_training - Step 17583: {'lr': 0.0003808617322395425, 'samples': 3376128, 'steps': 17583, 'loss/train': 1.833800196647644} 01/29/2022 11:25:54 - INFO - codeparrot_training - Step 17584: {'lr': 0.00038084779018826245, 'samples': 3376320, 'steps': 17584, 'loss/train': 0.4114777743816376} 01/29/2022 11:25:58 - INFO - codeparrot_training - Step 17585: {'lr': 0.00038083384757647186, 'samples': 3376512, 'steps': 17585, 'loss/train': 1.6660611033439636} 01/29/2022 11:26:03 - INFO - codeparrot_training - Step 17586: {'lr': 0.0003808199044042308, 'samples': 3376704, 'steps': 17586, 'loss/train': 0.7799055576324463} 01/29/2022 11:26:07 - INFO - codeparrot_training - Step 17587: {'lr': 0.00038080596067159865, 'samples': 3376896, 'steps': 17587, 'loss/train': 1.7723861932754517} 01/29/2022 11:26:11 - INFO - codeparrot_training - Step 17588: {'lr': 0.0003807920163786353, 'samples': 3377088, 'steps': 17588, 'loss/train': 0.6972278505563736} 01/29/2022 11:26:17 - INFO - codeparrot_training - Step 17589: {'lr': 0.0003807780715254006, 'samples': 3377280, 'steps': 17589, 'loss/train': 1.6132235527038574} 01/29/2022 11:26:21 - INFO - codeparrot_training - Step 17590: {'lr': 0.000380764126111954, 'samples': 3377472, 'steps': 17590, 'loss/train': 1.9102535247802734} 01/29/2022 11:26:25 - INFO - codeparrot_training - Step 17591: {'lr': 0.0003807501801383555, 'samples': 3377664, 'steps': 17591, 'loss/train': 1.6183072328567505} 01/29/2022 11:26:29 - INFO - codeparrot_training - Step 17592: {'lr': 0.0003807362336046648, 'samples': 3377856, 'steps': 17592, 'loss/train': 1.8124324679374695} 01/29/2022 11:26:34 - INFO - codeparrot_training - Step 17593: {'lr': 0.00038072228651094155, 'samples': 3378048, 'steps': 17593, 'loss/train': 1.4820899963378906} 01/29/2022 11:26:39 - INFO - codeparrot_training - Step 17594: {'lr': 0.0003807083388572455, 'samples': 3378240, 'steps': 17594, 'loss/train': 1.7423166632652283} 01/29/2022 11:26:43 - INFO - codeparrot_training - Step 17595: {'lr': 0.0003806943906436364, 'samples': 3378432, 'steps': 17595, 'loss/train': 1.4614554047584534} 01/29/2022 11:26:48 - INFO - codeparrot_training - Step 17596: {'lr': 0.0003806804418701741, 'samples': 3378624, 'steps': 17596, 'loss/train': 1.5625203251838684} 01/29/2022 11:26:52 - INFO - codeparrot_training - Step 17597: {'lr': 0.0003806664925369183, 'samples': 3378816, 'steps': 17597, 'loss/train': 1.5194551348686218} 01/29/2022 11:26:56 - INFO - codeparrot_training - Step 17598: {'lr': 0.0003806525426439287, 'samples': 3379008, 'steps': 17598, 'loss/train': 1.3228540420532227} 01/29/2022 11:27:03 - INFO - codeparrot_training - Step 17599: {'lr': 0.00038063859219126514, 'samples': 3379200, 'steps': 17599, 'loss/train': 2.109266459941864} 01/29/2022 11:27:07 - INFO - codeparrot_training - Step 17600: {'lr': 0.0003806246411789872, 'samples': 3379392, 'steps': 17600, 'loss/train': 1.845055639743805} 01/29/2022 11:27:12 - INFO - codeparrot_training - Step 17601: {'lr': 0.00038061068960715494, 'samples': 3379584, 'steps': 17601, 'loss/train': 2.0185768604278564} 01/29/2022 11:27:16 - INFO - codeparrot_training - Step 17602: {'lr': 0.00038059673747582783, 'samples': 3379776, 'steps': 17602, 'loss/train': 1.640307605266571} 01/29/2022 11:27:20 - INFO - codeparrot_training - Step 17603: {'lr': 0.00038058278478506584, 'samples': 3379968, 'steps': 17603, 'loss/train': 2.78970730304718} 01/29/2022 11:27:26 - INFO - codeparrot_training - Step 17604: {'lr': 0.0003805688315349286, 'samples': 3380160, 'steps': 17604, 'loss/train': 2.049752712249756} 01/29/2022 11:27:30 - INFO - codeparrot_training - Step 17605: {'lr': 0.00038055487772547603, 'samples': 3380352, 'steps': 17605, 'loss/train': 0.8530494868755341} 01/29/2022 11:27:34 - INFO - codeparrot_training - Step 17606: {'lr': 0.00038054092335676774, 'samples': 3380544, 'steps': 17606, 'loss/train': 1.62990403175354} 01/29/2022 11:27:39 - INFO - codeparrot_training - Step 17607: {'lr': 0.00038052696842886364, 'samples': 3380736, 'steps': 17607, 'loss/train': 0.6512189358472824} 01/29/2022 11:27:43 - INFO - codeparrot_training - Step 17608: {'lr': 0.0003805130129418235, 'samples': 3380928, 'steps': 17608, 'loss/train': 0.2693111523985863} 01/29/2022 11:27:49 - INFO - codeparrot_training - Step 17609: {'lr': 0.00038049905689570697, 'samples': 3381120, 'steps': 17609, 'loss/train': 2.0984700322151184} 01/29/2022 11:27:53 - INFO - codeparrot_training - Step 17610: {'lr': 0.00038048510029057393, 'samples': 3381312, 'steps': 17610, 'loss/train': 1.511958360671997} 01/29/2022 11:27:57 - INFO - codeparrot_training - Step 17611: {'lr': 0.00038047114312648414, 'samples': 3381504, 'steps': 17611, 'loss/train': 1.5327145457267761} 01/29/2022 11:28:01 - INFO - codeparrot_training - Step 17612: {'lr': 0.0003804571854034975, 'samples': 3381696, 'steps': 17612, 'loss/train': 1.9099947810173035} 01/29/2022 11:28:06 - INFO - codeparrot_training - Step 17613: {'lr': 0.0003804432271216736, 'samples': 3381888, 'steps': 17613, 'loss/train': 1.2879390120506287} 01/29/2022 11:28:13 - INFO - codeparrot_training - Step 17614: {'lr': 0.0003804292682810724, 'samples': 3382080, 'steps': 17614, 'loss/train': 2.0028435587882996} 01/29/2022 11:28:17 - INFO - codeparrot_training - Step 17615: {'lr': 0.00038041530888175356, 'samples': 3382272, 'steps': 17615, 'loss/train': 1.4851513803005219} 01/29/2022 11:28:21 - INFO - codeparrot_training - Step 17616: {'lr': 0.00038040134892377696, 'samples': 3382464, 'steps': 17616, 'loss/train': 0.8715249001979828} 01/29/2022 11:28:25 - INFO - codeparrot_training - Step 17617: {'lr': 0.00038038738840720244, 'samples': 3382656, 'steps': 17617, 'loss/train': 1.783355176448822} 01/29/2022 11:28:30 - INFO - codeparrot_training - Step 17618: {'lr': 0.0003803734273320897, 'samples': 3382848, 'steps': 17618, 'loss/train': 1.5126602053642273} 01/29/2022 11:28:35 - INFO - codeparrot_training - Step 17619: {'lr': 0.0003803594656984986, 'samples': 3383040, 'steps': 17619, 'loss/train': 1.3833812177181244} 01/29/2022 11:28:40 - INFO - codeparrot_training - Step 17620: {'lr': 0.000380345503506489, 'samples': 3383232, 'steps': 17620, 'loss/train': 1.4647624790668488} 01/29/2022 11:28:44 - INFO - codeparrot_training - Step 17621: {'lr': 0.00038033154075612063, 'samples': 3383424, 'steps': 17621, 'loss/train': 1.8665145635604858} 01/29/2022 11:28:48 - INFO - codeparrot_training - Step 17622: {'lr': 0.00038031757744745327, 'samples': 3383616, 'steps': 17622, 'loss/train': 2.101449429988861} 01/29/2022 11:28:52 - INFO - codeparrot_training - Step 17623: {'lr': 0.0003803036135805469, 'samples': 3383808, 'steps': 17623, 'loss/train': 1.4377171397209167} 01/29/2022 11:28:58 - INFO - codeparrot_training - Step 17624: {'lr': 0.00038028964915546107, 'samples': 3384000, 'steps': 17624, 'loss/train': 1.5672125816345215} 01/29/2022 11:29:02 - INFO - codeparrot_training - Step 17625: {'lr': 0.00038027568417225586, 'samples': 3384192, 'steps': 17625, 'loss/train': 1.9499963521957397} 01/29/2022 11:29:06 - INFO - codeparrot_training - Step 17626: {'lr': 0.00038026171863099093, 'samples': 3384384, 'steps': 17626, 'loss/train': 1.1355502903461456} 01/29/2022 11:29:10 - INFO - codeparrot_training - Step 17627: {'lr': 0.0003802477525317263, 'samples': 3384576, 'steps': 17627, 'loss/train': 1.7707768678665161} 01/29/2022 11:29:15 - INFO - codeparrot_training - Step 17628: {'lr': 0.00038023378587452144, 'samples': 3384768, 'steps': 17628, 'loss/train': 1.6894075870513916} 01/29/2022 11:29:20 - INFO - codeparrot_training - Step 17629: {'lr': 0.0003802198186594366, 'samples': 3384960, 'steps': 17629, 'loss/train': 2.6403860449790955} 01/29/2022 11:29:24 - INFO - codeparrot_training - Step 17630: {'lr': 0.00038020585088653126, 'samples': 3385152, 'steps': 17630, 'loss/train': 1.6930516362190247} 01/29/2022 11:29:28 - INFO - codeparrot_training - Step 17631: {'lr': 0.00038019188255586546, 'samples': 3385344, 'steps': 17631, 'loss/train': 2.58327716588974} 01/29/2022 11:29:33 - INFO - codeparrot_training - Step 17632: {'lr': 0.00038017791366749896, 'samples': 3385536, 'steps': 17632, 'loss/train': 1.5052217245101929} 01/29/2022 11:29:37 - INFO - codeparrot_training - Step 17633: {'lr': 0.0003801639442214916, 'samples': 3385728, 'steps': 17633, 'loss/train': 1.5974138975143433} 01/29/2022 11:29:45 - INFO - codeparrot_training - Step 17634: {'lr': 0.0003801499742179033, 'samples': 3385920, 'steps': 17634, 'loss/train': 1.8386959433555603} 01/29/2022 11:29:49 - INFO - codeparrot_training - Step 17635: {'lr': 0.0003801360036567938, 'samples': 3386112, 'steps': 17635, 'loss/train': 1.8956524729728699} 01/29/2022 11:29:53 - INFO - codeparrot_training - Step 17636: {'lr': 0.000380122032538223, 'samples': 3386304, 'steps': 17636, 'loss/train': 1.1651215553283691} 01/29/2022 11:29:57 - INFO - codeparrot_training - Step 17637: {'lr': 0.0003801080608622507, 'samples': 3386496, 'steps': 17637, 'loss/train': 1.4177897572517395} 01/29/2022 11:30:01 - INFO - codeparrot_training - Step 17638: {'lr': 0.0003800940886289368, 'samples': 3386688, 'steps': 17638, 'loss/train': 2.530653655529022} 01/29/2022 11:30:07 - INFO - codeparrot_training - Step 17639: {'lr': 0.0003800801158383411, 'samples': 3386880, 'steps': 17639, 'loss/train': 1.6739501953125} 01/29/2022 11:30:11 - INFO - codeparrot_training - Step 17640: {'lr': 0.00038006614249052353, 'samples': 3387072, 'steps': 17640, 'loss/train': 1.8574118614196777} 01/29/2022 11:30:16 - INFO - codeparrot_training - Step 17641: {'lr': 0.0003800521685855439, 'samples': 3387264, 'steps': 17641, 'loss/train': 2.0841211080551147} 01/29/2022 11:30:20 - INFO - codeparrot_training - Step 17642: {'lr': 0.000380038194123462, 'samples': 3387456, 'steps': 17642, 'loss/train': 1.5945813059806824} 01/29/2022 11:30:24 - INFO - codeparrot_training - Step 17643: {'lr': 0.0003800242191043379, 'samples': 3387648, 'steps': 17643, 'loss/train': 1.5577957034111023} 01/29/2022 11:30:31 - INFO - codeparrot_training - Step 17644: {'lr': 0.00038001024352823123, 'samples': 3387840, 'steps': 17644, 'loss/train': 2.0011850595474243} 01/29/2022 11:30:35 - INFO - codeparrot_training - Step 17645: {'lr': 0.00037999626739520197, 'samples': 3388032, 'steps': 17645, 'loss/train': 3.0331302881240845} 01/29/2022 11:30:40 - INFO - codeparrot_training - Step 17646: {'lr': 0.00037998229070531, 'samples': 3388224, 'steps': 17646, 'loss/train': 1.43987175822258} 01/29/2022 11:30:44 - INFO - codeparrot_training - Step 17647: {'lr': 0.0003799683134586152, 'samples': 3388416, 'steps': 17647, 'loss/train': 1.9024270176887512} 01/29/2022 11:30:49 - INFO - codeparrot_training - Step 17648: {'lr': 0.0003799543356551773, 'samples': 3388608, 'steps': 17648, 'loss/train': 1.3922624588012695} 01/29/2022 11:30:54 - INFO - codeparrot_training - Step 17649: {'lr': 0.0003799403572950565, 'samples': 3388800, 'steps': 17649, 'loss/train': 3.461510181427002} 01/29/2022 11:30:58 - INFO - codeparrot_training - Step 17650: {'lr': 0.00037992637837831235, 'samples': 3388992, 'steps': 17650, 'loss/train': 1.0696598589420319} 01/29/2022 11:31:02 - INFO - codeparrot_training - Step 17651: {'lr': 0.00037991239890500483, 'samples': 3389184, 'steps': 17651, 'loss/train': 2.147655487060547} 01/29/2022 11:31:06 - INFO - codeparrot_training - Step 17652: {'lr': 0.00037989841887519385, 'samples': 3389376, 'steps': 17652, 'loss/train': 1.2667794227600098} 01/29/2022 11:31:12 - INFO - codeparrot_training - Step 17653: {'lr': 0.00037988443828893936, 'samples': 3389568, 'steps': 17653, 'loss/train': 1.8973565697669983} 01/29/2022 11:31:17 - INFO - codeparrot_training - Step 17654: {'lr': 0.0003798704571463011, 'samples': 3389760, 'steps': 17654, 'loss/train': 0.20303960144519806} 01/29/2022 11:31:21 - INFO - codeparrot_training - Step 17655: {'lr': 0.00037985647544733903, 'samples': 3389952, 'steps': 17655, 'loss/train': 1.188598483800888} 01/29/2022 11:31:25 - INFO - codeparrot_training - Step 17656: {'lr': 0.0003798424931921131, 'samples': 3390144, 'steps': 17656, 'loss/train': 2.2400758266448975} 01/29/2022 11:31:29 - INFO - codeparrot_training - Step 17657: {'lr': 0.0003798285103806831, 'samples': 3390336, 'steps': 17657, 'loss/train': 1.9271939992904663} 01/29/2022 11:31:34 - INFO - codeparrot_training - Step 17658: {'lr': 0.0003798145270131091, 'samples': 3390528, 'steps': 17658, 'loss/train': 1.7040342092514038} 01/29/2022 11:31:41 - INFO - codeparrot_training - Step 17659: {'lr': 0.00037980054308945076, 'samples': 3390720, 'steps': 17659, 'loss/train': 2.0413740277290344} 01/29/2022 11:31:45 - INFO - codeparrot_training - Step 17660: {'lr': 0.00037978655860976826, 'samples': 3390912, 'steps': 17660, 'loss/train': 2.523838520050049} 01/29/2022 11:31:49 - INFO - codeparrot_training - Step 17661: {'lr': 0.0003797725735741212, 'samples': 3391104, 'steps': 17661, 'loss/train': 2.035212993621826} 01/29/2022 11:31:54 - INFO - codeparrot_training - Step 17662: {'lr': 0.0003797585879825698, 'samples': 3391296, 'steps': 17662, 'loss/train': 1.8075710535049438} 01/29/2022 11:31:58 - INFO - codeparrot_training - Step 17663: {'lr': 0.00037974460183517366, 'samples': 3391488, 'steps': 17663, 'loss/train': 1.627257227897644} 01/29/2022 11:32:03 - INFO - codeparrot_training - Step 17664: {'lr': 0.0003797306151319929, 'samples': 3391680, 'steps': 17664, 'loss/train': 1.8872737884521484} 01/29/2022 11:32:07 - INFO - codeparrot_training - Step 17665: {'lr': 0.00037971662787308734, 'samples': 3391872, 'steps': 17665, 'loss/train': 2.947884500026703} 01/29/2022 11:32:12 - INFO - codeparrot_training - Step 17666: {'lr': 0.00037970264005851703, 'samples': 3392064, 'steps': 17666, 'loss/train': 2.4276856184005737} 01/29/2022 11:32:16 - INFO - codeparrot_training - Step 17667: {'lr': 0.0003796886516883418, 'samples': 3392256, 'steps': 17667, 'loss/train': 2.2185405492782593} 01/29/2022 11:32:20 - INFO - codeparrot_training - Step 17668: {'lr': 0.0003796746627626214, 'samples': 3392448, 'steps': 17668, 'loss/train': 1.8105615377426147} 01/29/2022 11:32:28 - INFO - codeparrot_training - Step 17669: {'lr': 0.00037966067328141606, 'samples': 3392640, 'steps': 17669, 'loss/train': 1.036440521478653} 01/29/2022 11:32:33 - INFO - codeparrot_training - Step 17670: {'lr': 0.0003796466832447856, 'samples': 3392832, 'steps': 17670, 'loss/train': 1.7083913683891296} 01/29/2022 11:32:37 - INFO - codeparrot_training - Step 17671: {'lr': 0.00037963269265278986, 'samples': 3393024, 'steps': 17671, 'loss/train': 1.571995496749878} 01/29/2022 11:32:41 - INFO - codeparrot_training - Step 17672: {'lr': 0.0003796187015054888, 'samples': 3393216, 'steps': 17672, 'loss/train': 1.4964545667171478} 01/29/2022 11:32:45 - INFO - codeparrot_training - Step 17673: {'lr': 0.0003796047098029424, 'samples': 3393408, 'steps': 17673, 'loss/train': 1.6095832586288452} 01/29/2022 11:32:50 - INFO - codeparrot_training - Step 17674: {'lr': 0.0003795907175452106, 'samples': 3393600, 'steps': 17674, 'loss/train': 1.921058714389801} 01/29/2022 11:32:55 - INFO - codeparrot_training - Step 17675: {'lr': 0.0003795767247323533, 'samples': 3393792, 'steps': 17675, 'loss/train': 2.0774710178375244} 01/29/2022 11:32:59 - INFO - codeparrot_training - Step 17676: {'lr': 0.00037956273136443056, 'samples': 3393984, 'steps': 17676, 'loss/train': 0.6160942912101746} 01/29/2022 11:33:04 - INFO - codeparrot_training - Step 17677: {'lr': 0.000379548737441502, 'samples': 3394176, 'steps': 17677, 'loss/train': 1.7585846185684204} 01/29/2022 11:33:08 - INFO - codeparrot_training - Step 17678: {'lr': 0.00037953474296362796, 'samples': 3394368, 'steps': 17678, 'loss/train': 1.2684491872787476} 01/29/2022 11:33:12 - INFO - codeparrot_training - Step 17679: {'lr': 0.0003795207479308681, 'samples': 3394560, 'steps': 17679, 'loss/train': 2.904358685016632} 01/29/2022 11:33:16 - INFO - codeparrot_training - Step 17680: {'lr': 0.00037950675234328256, 'samples': 3394752, 'steps': 17680, 'loss/train': 2.131645917892456} 01/29/2022 11:33:22 - INFO - codeparrot_training - Step 17681: {'lr': 0.00037949275620093124, 'samples': 3394944, 'steps': 17681, 'loss/train': 1.4825413227081299} 01/29/2022 11:33:27 - INFO - codeparrot_training - Step 17682: {'lr': 0.000379478759503874, 'samples': 3395136, 'steps': 17682, 'loss/train': 0.4065762162208557} 01/29/2022 11:33:31 - INFO - codeparrot_training - Step 17683: {'lr': 0.00037946476225217087, 'samples': 3395328, 'steps': 17683, 'loss/train': 0.7758090198040009} 01/29/2022 11:33:36 - INFO - codeparrot_training - Step 17684: {'lr': 0.0003794507644458819, 'samples': 3395520, 'steps': 17684, 'loss/train': 1.792350947856903} 01/29/2022 11:33:43 - INFO - codeparrot_training - Step 17685: {'lr': 0.00037943676608506683, 'samples': 3395712, 'steps': 17685, 'loss/train': 0.45575858652591705} 01/29/2022 11:33:48 - INFO - codeparrot_training - Step 17686: {'lr': 0.00037942276716978584, 'samples': 3395904, 'steps': 17686, 'loss/train': 1.2885256111621857} 01/29/2022 11:33:52 - INFO - codeparrot_training - Step 17687: {'lr': 0.0003794087677000988, 'samples': 3396096, 'steps': 17687, 'loss/train': 1.5541360974311829} 01/29/2022 11:33:56 - INFO - codeparrot_training - Step 17688: {'lr': 0.0003793947676760657, 'samples': 3396288, 'steps': 17688, 'loss/train': 1.554330825805664} 01/29/2022 11:34:00 - INFO - codeparrot_training - Step 17689: {'lr': 0.00037938076709774645, 'samples': 3396480, 'steps': 17689, 'loss/train': 1.408969223499298} 01/29/2022 11:34:05 - INFO - codeparrot_training - Step 17690: {'lr': 0.0003793667659652011, 'samples': 3396672, 'steps': 17690, 'loss/train': 1.955627202987671} 01/29/2022 11:34:10 - INFO - codeparrot_training - Step 17691: {'lr': 0.0003793527642784896, 'samples': 3396864, 'steps': 17691, 'loss/train': 1.756999433040619} 01/29/2022 11:34:15 - INFO - codeparrot_training - Step 17692: {'lr': 0.0003793387620376719, 'samples': 3397056, 'steps': 17692, 'loss/train': 2.078055202960968} 01/29/2022 11:34:19 - INFO - codeparrot_training - Step 17693: {'lr': 0.0003793247592428081, 'samples': 3397248, 'steps': 17693, 'loss/train': 1.8091356754302979} 01/29/2022 11:34:23 - INFO - codeparrot_training - Step 17694: {'lr': 0.00037931075589395805, 'samples': 3397440, 'steps': 17694, 'loss/train': 1.0251497626304626} 01/29/2022 11:34:29 - INFO - codeparrot_training - Step 17695: {'lr': 0.00037929675199118183, 'samples': 3397632, 'steps': 17695, 'loss/train': 1.186520129442215} 01/29/2022 11:34:33 - INFO - codeparrot_training - Step 17696: {'lr': 0.0003792827475345393, 'samples': 3397824, 'steps': 17696, 'loss/train': 1.7453480958938599} 01/29/2022 11:34:37 - INFO - codeparrot_training - Step 17697: {'lr': 0.0003792687425240906, 'samples': 3398016, 'steps': 17697, 'loss/train': 0.8568270206451416} 01/29/2022 11:34:42 - INFO - codeparrot_training - Step 17698: {'lr': 0.0003792547369598956, 'samples': 3398208, 'steps': 17698, 'loss/train': 1.3458905518054962} 01/29/2022 11:34:46 - INFO - codeparrot_training - Step 17699: {'lr': 0.0003792407308420144, 'samples': 3398400, 'steps': 17699, 'loss/train': 1.0633753538131714} 01/29/2022 11:34:50 - INFO - codeparrot_training - Step 17700: {'lr': 0.00037922672417050685, 'samples': 3398592, 'steps': 17700, 'loss/train': 1.860334038734436} 01/29/2022 11:34:56 - INFO - codeparrot_training - Step 17701: {'lr': 0.00037921271694543317, 'samples': 3398784, 'steps': 17701, 'loss/train': 1.671833574771881} 01/29/2022 11:35:00 - INFO - codeparrot_training - Step 17702: {'lr': 0.0003791987091668532, 'samples': 3398976, 'steps': 17702, 'loss/train': 2.1921345591545105} 01/29/2022 11:35:04 - INFO - codeparrot_training - Step 17703: {'lr': 0.00037918470083482693, 'samples': 3399168, 'steps': 17703, 'loss/train': 0.9292149245738983} 01/29/2022 11:35:08 - INFO - codeparrot_training - Step 17704: {'lr': 0.0003791706919494145, 'samples': 3399360, 'steps': 17704, 'loss/train': 1.511088252067566} 01/29/2022 11:35:15 - INFO - codeparrot_training - Step 17705: {'lr': 0.0003791566825106758, 'samples': 3399552, 'steps': 17705, 'loss/train': 0.6814961135387421} 01/29/2022 11:35:20 - INFO - codeparrot_training - Step 17706: {'lr': 0.0003791426725186709, 'samples': 3399744, 'steps': 17706, 'loss/train': 1.7272326350212097} 01/29/2022 11:35:24 - INFO - codeparrot_training - Step 17707: {'lr': 0.0003791286619734597, 'samples': 3399936, 'steps': 17707, 'loss/train': 1.7970430254936218} 01/29/2022 11:35:28 - INFO - codeparrot_training - Step 17708: {'lr': 0.0003791146508751025, 'samples': 3400128, 'steps': 17708, 'loss/train': 1.0267236828804016} 01/29/2022 11:35:32 - INFO - codeparrot_training - Step 17709: {'lr': 0.00037910063922365903, 'samples': 3400320, 'steps': 17709, 'loss/train': 2.1769267916679382} 01/29/2022 11:35:38 - INFO - codeparrot_training - Step 17710: {'lr': 0.00037908662701918944, 'samples': 3400512, 'steps': 17710, 'loss/train': 1.431027352809906} 01/29/2022 11:35:42 - INFO - codeparrot_training - Step 17711: {'lr': 0.00037907261426175365, 'samples': 3400704, 'steps': 17711, 'loss/train': 2.0392672419548035} 01/29/2022 11:35:47 - INFO - codeparrot_training - Step 17712: {'lr': 0.0003790586009514119, 'samples': 3400896, 'steps': 17712, 'loss/train': 0.6640901863574982} 01/29/2022 11:35:51 - INFO - codeparrot_training - Step 17713: {'lr': 0.000379044587088224, 'samples': 3401088, 'steps': 17713, 'loss/train': 1.865285575389862} 01/29/2022 11:35:55 - INFO - codeparrot_training - Step 17714: {'lr': 0.0003790305726722501, 'samples': 3401280, 'steps': 17714, 'loss/train': 1.3228795230388641} 01/29/2022 11:36:02 - INFO - codeparrot_training - Step 17715: {'lr': 0.00037901655770355015, 'samples': 3401472, 'steps': 17715, 'loss/train': 2.1079087257385254} 01/29/2022 11:36:06 - INFO - codeparrot_training - Step 17716: {'lr': 0.0003790025421821843, 'samples': 3401664, 'steps': 17716, 'loss/train': 2.7039220333099365} 01/29/2022 11:36:11 - INFO - codeparrot_training - Step 17717: {'lr': 0.0003789885261082124, 'samples': 3401856, 'steps': 17717, 'loss/train': 2.2112966179847717} 01/29/2022 11:36:15 - INFO - codeparrot_training - Step 17718: {'lr': 0.00037897450948169476, 'samples': 3402048, 'steps': 17718, 'loss/train': 1.6655488014221191} 01/29/2022 11:36:19 - INFO - codeparrot_training - Step 17719: {'lr': 0.0003789604923026912, 'samples': 3402240, 'steps': 17719, 'loss/train': 1.6883541941642761} 01/29/2022 11:36:24 - INFO - codeparrot_training - Step 17720: {'lr': 0.00037894647457126186, 'samples': 3402432, 'steps': 17720, 'loss/train': 1.9755545854568481} 01/29/2022 11:36:29 - INFO - codeparrot_training - Step 17721: {'lr': 0.0003789324562874668, 'samples': 3402624, 'steps': 17721, 'loss/train': 1.6884716749191284} 01/29/2022 11:36:33 - INFO - codeparrot_training - Step 17722: {'lr': 0.000378918437451366, 'samples': 3402816, 'steps': 17722, 'loss/train': 0.21666497737169266} 01/29/2022 11:36:37 - INFO - codeparrot_training - Step 17723: {'lr': 0.00037890441806301954, 'samples': 3403008, 'steps': 17723, 'loss/train': 1.6255229115486145} 01/29/2022 11:36:41 - INFO - codeparrot_training - Step 17724: {'lr': 0.0003788903981224875, 'samples': 3403200, 'steps': 17724, 'loss/train': 1.3891080021858215} 01/29/2022 11:36:47 - INFO - codeparrot_training - Step 17725: {'lr': 0.00037887637762982996, 'samples': 3403392, 'steps': 17725, 'loss/train': 1.761842429637909} 01/29/2022 11:36:51 - INFO - codeparrot_training - Step 17726: {'lr': 0.0003788623565851068, 'samples': 3403584, 'steps': 17726, 'loss/train': 1.6414082050323486} 01/29/2022 11:36:55 - INFO - codeparrot_training - Step 17727: {'lr': 0.00037884833498837833, 'samples': 3403776, 'steps': 17727, 'loss/train': 2.7450989484786987} 01/29/2022 11:36:59 - INFO - codeparrot_training - Step 17728: {'lr': 0.00037883431283970454, 'samples': 3403968, 'steps': 17728, 'loss/train': 2.1335724592208862} 01/29/2022 11:37:04 - INFO - codeparrot_training - Step 17729: {'lr': 0.00037882029013914544, 'samples': 3404160, 'steps': 17729, 'loss/train': 1.378300666809082} 01/29/2022 11:37:11 - INFO - codeparrot_training - Step 17730: {'lr': 0.0003788062668867611, 'samples': 3404352, 'steps': 17730, 'loss/train': 1.394225835800171} 01/29/2022 11:37:15 - INFO - codeparrot_training - Step 17731: {'lr': 0.00037879224308261163, 'samples': 3404544, 'steps': 17731, 'loss/train': 0.9893307387828827} 01/29/2022 11:37:19 - INFO - codeparrot_training - Step 17732: {'lr': 0.00037877821872675705, 'samples': 3404736, 'steps': 17732, 'loss/train': 1.7018246054649353} 01/29/2022 11:37:23 - INFO - codeparrot_training - Step 17733: {'lr': 0.0003787641938192575, 'samples': 3404928, 'steps': 17733, 'loss/train': 1.6084343791007996} 01/29/2022 11:37:28 - INFO - codeparrot_training - Step 17734: {'lr': 0.00037875016836017304, 'samples': 3405120, 'steps': 17734, 'loss/train': 2.1594929695129395} 01/29/2022 11:37:33 - INFO - codeparrot_training - Step 17735: {'lr': 0.0003787361423495637, 'samples': 3405312, 'steps': 17735, 'loss/train': 2.1168733835220337} 01/29/2022 11:37:37 - INFO - codeparrot_training - Step 17736: {'lr': 0.0003787221157874897, 'samples': 3405504, 'steps': 17736, 'loss/train': 2.1354235410690308} 01/29/2022 11:37:41 - INFO - codeparrot_training - Step 17737: {'lr': 0.00037870808867401085, 'samples': 3405696, 'steps': 17737, 'loss/train': 1.8965526223182678} 01/29/2022 11:37:46 - INFO - codeparrot_training - Step 17738: {'lr': 0.00037869406100918756, 'samples': 3405888, 'steps': 17738, 'loss/train': 1.6471754908561707} 01/29/2022 11:37:50 - INFO - codeparrot_training - Step 17739: {'lr': 0.0003786800327930797, 'samples': 3406080, 'steps': 17739, 'loss/train': 1.281560093164444} 01/29/2022 11:37:55 - INFO - codeparrot_training - Step 17740: {'lr': 0.0003786660040257475, 'samples': 3406272, 'steps': 17740, 'loss/train': 1.7054584622383118} 01/29/2022 11:37:59 - INFO - codeparrot_training - Step 17741: {'lr': 0.00037865197470725103, 'samples': 3406464, 'steps': 17741, 'loss/train': 1.406152456998825} 01/29/2022 11:38:04 - INFO - codeparrot_training - Step 17742: {'lr': 0.0003786379448376503, 'samples': 3406656, 'steps': 17742, 'loss/train': 1.2180178463459015} 01/29/2022 11:38:08 - INFO - codeparrot_training - Step 17743: {'lr': 0.0003786239144170055, 'samples': 3406848, 'steps': 17743, 'loss/train': 1.6292238235473633} 01/29/2022 11:38:12 - INFO - codeparrot_training - Step 17744: {'lr': 0.0003786098834453766, 'samples': 3407040, 'steps': 17744, 'loss/train': 2.6406564116477966} 01/29/2022 11:38:19 - INFO - codeparrot_training - Step 17745: {'lr': 0.00037859585192282386, 'samples': 3407232, 'steps': 17745, 'loss/train': 1.6541771292686462} 01/29/2022 11:38:24 - INFO - codeparrot_training - Step 17746: {'lr': 0.00037858181984940734, 'samples': 3407424, 'steps': 17746, 'loss/train': 1.5316223502159119} 01/29/2022 11:38:28 - INFO - codeparrot_training - Step 17747: {'lr': 0.0003785677872251871, 'samples': 3407616, 'steps': 17747, 'loss/train': 1.781548798084259} 01/29/2022 11:38:32 - INFO - codeparrot_training - Step 17748: {'lr': 0.0003785537540502233, 'samples': 3407808, 'steps': 17748, 'loss/train': 1.3350982367992401} 01/29/2022 11:38:36 - INFO - codeparrot_training - Step 17749: {'lr': 0.0003785397203245761, 'samples': 3408000, 'steps': 17749, 'loss/train': 1.4226578772068024} 01/29/2022 11:38:42 - INFO - codeparrot_training - Step 17750: {'lr': 0.0003785256860483054, 'samples': 3408192, 'steps': 17750, 'loss/train': 0.5973874479532242} 01/29/2022 11:38:46 - INFO - codeparrot_training - Step 17751: {'lr': 0.0003785116512214716, 'samples': 3408384, 'steps': 17751, 'loss/train': 1.5414544343948364} 01/29/2022 11:38:50 - INFO - codeparrot_training - Step 17752: {'lr': 0.0003784976158441347, 'samples': 3408576, 'steps': 17752, 'loss/train': 1.387190043926239} 01/29/2022 11:38:55 - INFO - codeparrot_training - Step 17753: {'lr': 0.0003784835799163547, 'samples': 3408768, 'steps': 17753, 'loss/train': 1.5656654834747314} 01/29/2022 11:38:59 - INFO - codeparrot_training - Step 17754: {'lr': 0.00037846954343819195, 'samples': 3408960, 'steps': 17754, 'loss/train': 1.06986004114151} 01/29/2022 11:39:04 - INFO - codeparrot_training - Step 17755: {'lr': 0.00037845550640970636, 'samples': 3409152, 'steps': 17755, 'loss/train': 2.105341136455536} 01/29/2022 11:39:08 - INFO - codeparrot_training - Step 17756: {'lr': 0.0003784414688309583, 'samples': 3409344, 'steps': 17756, 'loss/train': 2.1651533246040344} 01/29/2022 11:39:13 - INFO - codeparrot_training - Step 17757: {'lr': 0.00037842743070200767, 'samples': 3409536, 'steps': 17757, 'loss/train': 1.7543517351150513} 01/29/2022 11:39:17 - INFO - codeparrot_training - Step 17758: {'lr': 0.0003784133920229148, 'samples': 3409728, 'steps': 17758, 'loss/train': 1.5111560225486755} 01/29/2022 11:39:21 - INFO - codeparrot_training - Step 17759: {'lr': 0.0003783993527937397, 'samples': 3409920, 'steps': 17759, 'loss/train': 1.5231564044952393} 01/29/2022 11:39:28 - INFO - codeparrot_training - Step 17760: {'lr': 0.0003783853130145425, 'samples': 3410112, 'steps': 17760, 'loss/train': 0.8127701282501221} 01/29/2022 11:39:32 - INFO - codeparrot_training - Step 17761: {'lr': 0.0003783712726853835, 'samples': 3410304, 'steps': 17761, 'loss/train': 1.4749175906181335} 01/29/2022 11:39:37 - INFO - codeparrot_training - Step 17762: {'lr': 0.00037835723180632263, 'samples': 3410496, 'steps': 17762, 'loss/train': 0.917858362197876} 01/29/2022 11:39:41 - INFO - codeparrot_training - Step 17763: {'lr': 0.00037834319037742016, 'samples': 3410688, 'steps': 17763, 'loss/train': 1.6232416033744812} 01/29/2022 11:39:45 - INFO - codeparrot_training - Step 17764: {'lr': 0.00037832914839873623, 'samples': 3410880, 'steps': 17764, 'loss/train': 1.203791856765747} 01/29/2022 11:39:51 - INFO - codeparrot_training - Step 17765: {'lr': 0.0003783151058703309, 'samples': 3411072, 'steps': 17765, 'loss/train': 1.7228025197982788} 01/29/2022 11:39:55 - INFO - codeparrot_training - Step 17766: {'lr': 0.0003783010627922645, 'samples': 3411264, 'steps': 17766, 'loss/train': 0.7496535927057266} 01/29/2022 11:39:59 - INFO - codeparrot_training - Step 17767: {'lr': 0.0003782870191645971, 'samples': 3411456, 'steps': 17767, 'loss/train': 1.589672327041626} 01/29/2022 11:40:03 - INFO - codeparrot_training - Step 17768: {'lr': 0.0003782729749873887, 'samples': 3411648, 'steps': 17768, 'loss/train': 1.2837603986263275} 01/29/2022 11:40:07 - INFO - codeparrot_training - Step 17769: {'lr': 0.00037825893026069977, 'samples': 3411840, 'steps': 17769, 'loss/train': 2.1456416845321655} 01/29/2022 11:40:13 - INFO - codeparrot_training - Step 17770: {'lr': 0.0003782448849845902, 'samples': 3412032, 'steps': 17770, 'loss/train': 1.5683627128601074} 01/29/2022 11:40:17 - INFO - codeparrot_training - Step 17771: {'lr': 0.0003782308391591203, 'samples': 3412224, 'steps': 17771, 'loss/train': 1.7627559900283813} 01/29/2022 11:40:22 - INFO - codeparrot_training - Step 17772: {'lr': 0.00037821679278435017, 'samples': 3412416, 'steps': 17772, 'loss/train': 1.5670339465141296} 01/29/2022 11:40:26 - INFO - codeparrot_training - Step 17773: {'lr': 0.0003782027458603401, 'samples': 3412608, 'steps': 17773, 'loss/train': 1.3577392101287842} 01/29/2022 11:40:30 - INFO - codeparrot_training - Step 17774: {'lr': 0.0003781886983871501, 'samples': 3412800, 'steps': 17774, 'loss/train': 0.9608800113201141} 01/29/2022 11:40:37 - INFO - codeparrot_training - Step 17775: {'lr': 0.00037817465036484043, 'samples': 3412992, 'steps': 17775, 'loss/train': 1.7903831005096436} 01/29/2022 11:40:41 - INFO - codeparrot_training - Step 17776: {'lr': 0.0003781606017934713, 'samples': 3413184, 'steps': 17776, 'loss/train': 0.9895058870315552} 01/29/2022 11:40:46 - INFO - codeparrot_training - Step 17777: {'lr': 0.0003781465526731028, 'samples': 3413376, 'steps': 17777, 'loss/train': 2.1936228275299072} 01/29/2022 11:40:50 - INFO - codeparrot_training - Step 17778: {'lr': 0.0003781325030037952, 'samples': 3413568, 'steps': 17778, 'loss/train': 2.08761066198349} 01/29/2022 11:40:54 - INFO - codeparrot_training - Step 17779: {'lr': 0.00037811845278560864, 'samples': 3413760, 'steps': 17779, 'loss/train': 1.6830037236213684} 01/29/2022 11:41:00 - INFO - codeparrot_training - Step 17780: {'lr': 0.0003781044020186033, 'samples': 3413952, 'steps': 17780, 'loss/train': 2.3580440282821655} 01/29/2022 11:41:04 - INFO - codeparrot_training - Step 17781: {'lr': 0.0003780903507028393, 'samples': 3414144, 'steps': 17781, 'loss/train': 1.4483700692653656} 01/29/2022 11:41:08 - INFO - codeparrot_training - Step 17782: {'lr': 0.00037807629883837703, 'samples': 3414336, 'steps': 17782, 'loss/train': 1.1916337609291077} 01/29/2022 11:41:13 - INFO - codeparrot_training - Step 17783: {'lr': 0.00037806224642527653, 'samples': 3414528, 'steps': 17783, 'loss/train': 1.4962374866008759} 01/29/2022 11:41:17 - INFO - codeparrot_training - Step 17784: {'lr': 0.000378048193463598, 'samples': 3414720, 'steps': 17784, 'loss/train': 1.822056770324707} 01/29/2022 11:41:22 - INFO - codeparrot_training - Step 17785: {'lr': 0.0003780341399534017, 'samples': 3414912, 'steps': 17785, 'loss/train': 1.9872068166732788} 01/29/2022 11:41:26 - INFO - codeparrot_training - Step 17786: {'lr': 0.00037802008589474777, 'samples': 3415104, 'steps': 17786, 'loss/train': 0.7964824140071869} 01/29/2022 11:41:31 - INFO - codeparrot_training - Step 17787: {'lr': 0.0003780060312876965, 'samples': 3415296, 'steps': 17787, 'loss/train': 1.8380454182624817} 01/29/2022 11:41:35 - INFO - codeparrot_training - Step 17788: {'lr': 0.00037799197613230795, 'samples': 3415488, 'steps': 17788, 'loss/train': 1.050055056810379} 01/29/2022 11:41:39 - INFO - codeparrot_training - Step 17789: {'lr': 0.00037797792042864247, 'samples': 3415680, 'steps': 17789, 'loss/train': 1.8282049298286438} 01/29/2022 11:41:46 - INFO - codeparrot_training - Step 17790: {'lr': 0.0003779638641767602, 'samples': 3415872, 'steps': 17790, 'loss/train': 1.5565304160118103} 01/29/2022 11:41:50 - INFO - codeparrot_training - Step 17791: {'lr': 0.0003779498073767214, 'samples': 3416064, 'steps': 17791, 'loss/train': 1.7745675444602966} 01/29/2022 11:41:55 - INFO - codeparrot_training - Step 17792: {'lr': 0.00037793575002858625, 'samples': 3416256, 'steps': 17792, 'loss/train': 1.3542865812778473} 01/29/2022 11:41:59 - INFO - codeparrot_training - Step 17793: {'lr': 0.00037792169213241494, 'samples': 3416448, 'steps': 17793, 'loss/train': 0.4121933877468109} 01/29/2022 11:42:03 - INFO - codeparrot_training - Step 17794: {'lr': 0.00037790763368826774, 'samples': 3416640, 'steps': 17794, 'loss/train': 2.0331480503082275} 01/29/2022 11:42:08 - INFO - codeparrot_training - Step 17795: {'lr': 0.00037789357469620487, 'samples': 3416832, 'steps': 17795, 'loss/train': 2.009753465652466} 01/29/2022 11:42:13 - INFO - codeparrot_training - Step 17796: {'lr': 0.0003778795151562865, 'samples': 3417024, 'steps': 17796, 'loss/train': 1.3121416568756104} 01/29/2022 11:42:17 - INFO - codeparrot_training - Step 17797: {'lr': 0.00037786545506857295, 'samples': 3417216, 'steps': 17797, 'loss/train': 1.2565578818321228} 01/29/2022 11:42:21 - INFO - codeparrot_training - Step 17798: {'lr': 0.0003778513944331243, 'samples': 3417408, 'steps': 17798, 'loss/train': 2.2388636469841003} 01/29/2022 11:42:25 - INFO - codeparrot_training - Step 17799: {'lr': 0.0003778373332500009, 'samples': 3417600, 'steps': 17799, 'loss/train': 0.9644931256771088} 01/29/2022 11:42:33 - INFO - codeparrot_training - Step 17800: {'lr': 0.00037782327151926297, 'samples': 3417792, 'steps': 17800, 'loss/train': 1.0443956851959229} 01/29/2022 11:42:37 - INFO - codeparrot_training - Step 17801: {'lr': 0.00037780920924097085, 'samples': 3417984, 'steps': 17801, 'loss/train': 1.7778216004371643} 01/29/2022 11:42:41 - INFO - codeparrot_training - Step 17802: {'lr': 0.00037779514641518455, 'samples': 3418176, 'steps': 17802, 'loss/train': 1.3607203960418701} 01/29/2022 11:42:45 - INFO - codeparrot_training - Step 17803: {'lr': 0.0003777810830419644, 'samples': 3418368, 'steps': 17803, 'loss/train': 2.244379162788391} 01/29/2022 11:42:50 - INFO - codeparrot_training - Step 17804: {'lr': 0.00037776701912137066, 'samples': 3418560, 'steps': 17804, 'loss/train': 1.2192912697792053} 01/29/2022 11:42:55 - INFO - codeparrot_training - Step 17805: {'lr': 0.00037775295465346373, 'samples': 3418752, 'steps': 17805, 'loss/train': 1.3357660174369812} 01/29/2022 11:42:59 - INFO - codeparrot_training - Step 17806: {'lr': 0.0003777388896383035, 'samples': 3418944, 'steps': 17806, 'loss/train': 2.8255181908607483} 01/29/2022 11:43:03 - INFO - codeparrot_training - Step 17807: {'lr': 0.00037772482407595056, 'samples': 3419136, 'steps': 17807, 'loss/train': 1.3246745467185974} 01/29/2022 11:43:08 - INFO - codeparrot_training - Step 17808: {'lr': 0.000377710757966465, 'samples': 3419328, 'steps': 17808, 'loss/train': 1.9233353734016418} 01/29/2022 11:43:12 - INFO - codeparrot_training - Step 17809: {'lr': 0.0003776966913099071, 'samples': 3419520, 'steps': 17809, 'loss/train': 1.3222984671592712} 01/29/2022 11:43:17 - INFO - codeparrot_training - Step 17810: {'lr': 0.00037768262410633715, 'samples': 3419712, 'steps': 17810, 'loss/train': 2.159421443939209} 01/29/2022 11:43:22 - INFO - codeparrot_training - Step 17811: {'lr': 0.0003776685563558153, 'samples': 3419904, 'steps': 17811, 'loss/train': 1.9051601886749268} 01/29/2022 11:43:26 - INFO - codeparrot_training - Step 17812: {'lr': 0.00037765448805840196, 'samples': 3420096, 'steps': 17812, 'loss/train': 1.884469985961914} 01/29/2022 11:43:30 - INFO - codeparrot_training - Step 17813: {'lr': 0.00037764041921415736, 'samples': 3420288, 'steps': 17813, 'loss/train': 1.3773360550403595} 01/29/2022 11:43:34 - INFO - codeparrot_training - Step 17814: {'lr': 0.00037762634982314164, 'samples': 3420480, 'steps': 17814, 'loss/train': 1.569515883922577} 01/29/2022 11:43:40 - INFO - codeparrot_training - Step 17815: {'lr': 0.00037761227988541523, 'samples': 3420672, 'steps': 17815, 'loss/train': 1.6910539269447327} 01/29/2022 11:43:44 - INFO - codeparrot_training - Step 17816: {'lr': 0.00037759820940103827, 'samples': 3420864, 'steps': 17816, 'loss/train': 1.9775959253311157} 01/29/2022 11:43:48 - INFO - codeparrot_training - Step 17817: {'lr': 0.00037758413837007124, 'samples': 3421056, 'steps': 17817, 'loss/train': 2.1117008328437805} 01/29/2022 11:43:53 - INFO - codeparrot_training - Step 17818: {'lr': 0.0003775700667925741, 'samples': 3421248, 'steps': 17818, 'loss/train': 1.9447814226150513} 01/29/2022 11:43:57 - INFO - codeparrot_training - Step 17819: {'lr': 0.0003775559946686075, 'samples': 3421440, 'steps': 17819, 'loss/train': 0.8702967166900635} 01/29/2022 11:44:04 - INFO - codeparrot_training - Step 17820: {'lr': 0.00037754192199823135, 'samples': 3421632, 'steps': 17820, 'loss/train': 1.460905909538269} 01/29/2022 11:44:08 - INFO - codeparrot_training - Step 17821: {'lr': 0.00037752784878150613, 'samples': 3421824, 'steps': 17821, 'loss/train': 2.140490770339966} 01/29/2022 11:44:13 - INFO - codeparrot_training - Step 17822: {'lr': 0.00037751377501849215, 'samples': 3422016, 'steps': 17822, 'loss/train': 1.4389355778694153} 01/29/2022 11:44:17 - INFO - codeparrot_training - Step 17823: {'lr': 0.0003774997007092496, 'samples': 3422208, 'steps': 17823, 'loss/train': 1.4123019576072693} 01/29/2022 11:44:21 - INFO - codeparrot_training - Step 17824: {'lr': 0.00037748562585383886, 'samples': 3422400, 'steps': 17824, 'loss/train': 1.0260034203529358} 01/29/2022 11:44:27 - INFO - codeparrot_training - Step 17825: {'lr': 0.00037747155045232016, 'samples': 3422592, 'steps': 17825, 'loss/train': 1.330059438943863} 01/29/2022 11:44:31 - INFO - codeparrot_training - Step 17826: {'lr': 0.0003774574745047539, 'samples': 3422784, 'steps': 17826, 'loss/train': 1.2776820659637451} 01/29/2022 11:44:35 - INFO - codeparrot_training - Step 17827: {'lr': 0.0003774433980112001, 'samples': 3422976, 'steps': 17827, 'loss/train': 1.9816757440567017} 01/29/2022 11:44:39 - INFO - codeparrot_training - Step 17828: {'lr': 0.00037742932097171945, 'samples': 3423168, 'steps': 17828, 'loss/train': 1.5442028045654297} 01/29/2022 11:44:45 - INFO - codeparrot_training - Step 17829: {'lr': 0.0003774152433863719, 'samples': 3423360, 'steps': 17829, 'loss/train': 2.1547855138778687} 01/29/2022 11:44:49 - INFO - codeparrot_training - Step 17830: {'lr': 0.000377401165255218, 'samples': 3423552, 'steps': 17830, 'loss/train': 1.022608458995819} 01/29/2022 11:44:53 - INFO - codeparrot_training - Step 17831: {'lr': 0.0003773870865783179, 'samples': 3423744, 'steps': 17831, 'loss/train': 1.8086446523666382} 01/29/2022 11:44:58 - INFO - codeparrot_training - Step 17832: {'lr': 0.00037737300735573204, 'samples': 3423936, 'steps': 17832, 'loss/train': 1.82222181558609} 01/29/2022 11:45:02 - INFO - codeparrot_training - Step 17833: {'lr': 0.00037735892758752063, 'samples': 3424128, 'steps': 17833, 'loss/train': 1.0903638303279877} 01/29/2022 11:45:09 - INFO - codeparrot_training - Step 17834: {'lr': 0.000377344847273744, 'samples': 3424320, 'steps': 17834, 'loss/train': 1.966995120048523} 01/29/2022 11:45:13 - INFO - codeparrot_training - Step 17835: {'lr': 0.0003773307664144625, 'samples': 3424512, 'steps': 17835, 'loss/train': 1.800465166568756} 01/29/2022 11:45:17 - INFO - codeparrot_training - Step 17836: {'lr': 0.00037731668500973637, 'samples': 3424704, 'steps': 17836, 'loss/train': 1.4379624724388123} 01/29/2022 11:45:22 - INFO - codeparrot_training - Step 17837: {'lr': 0.00037730260305962604, 'samples': 3424896, 'steps': 17837, 'loss/train': 1.329147756099701} 01/29/2022 11:45:26 - INFO - codeparrot_training - Step 17838: {'lr': 0.00037728852056419183, 'samples': 3425088, 'steps': 17838, 'loss/train': 1.4836018681526184} 01/29/2022 11:45:31 - INFO - codeparrot_training - Step 17839: {'lr': 0.000377274437523494, 'samples': 3425280, 'steps': 17839, 'loss/train': 1.083869308233261} 01/29/2022 11:45:35 - INFO - codeparrot_training - Step 17840: {'lr': 0.00037726035393759286, 'samples': 3425472, 'steps': 17840, 'loss/train': 1.709087312221527} 01/29/2022 11:45:40 - INFO - codeparrot_training - Step 17841: {'lr': 0.00037724626980654877, 'samples': 3425664, 'steps': 17841, 'loss/train': 2.048208475112915} 01/29/2022 11:45:44 - INFO - codeparrot_training - Step 17842: {'lr': 0.00037723218513042203, 'samples': 3425856, 'steps': 17842, 'loss/train': 1.5763855576515198} 01/29/2022 11:45:48 - INFO - codeparrot_training - Step 17843: {'lr': 0.0003772180999092731, 'samples': 3426048, 'steps': 17843, 'loss/train': 1.610806167125702} 01/29/2022 11:45:55 - INFO - codeparrot_training - Step 17844: {'lr': 0.00037720401414316213, 'samples': 3426240, 'steps': 17844, 'loss/train': 2.104403078556061} 01/29/2022 11:46:00 - INFO - codeparrot_training - Step 17845: {'lr': 0.00037718992783214965, 'samples': 3426432, 'steps': 17845, 'loss/train': 1.778703510761261} 01/29/2022 11:46:04 - INFO - codeparrot_training - Step 17846: {'lr': 0.0003771758409762958, 'samples': 3426624, 'steps': 17846, 'loss/train': 2.101648986339569} 01/29/2022 11:46:08 - INFO - codeparrot_training - Step 17847: {'lr': 0.0003771617535756611, 'samples': 3426816, 'steps': 17847, 'loss/train': 2.3830504417419434} 01/29/2022 11:46:12 - INFO - codeparrot_training - Step 17848: {'lr': 0.00037714766563030585, 'samples': 3427008, 'steps': 17848, 'loss/train': 0.9164147973060608} 01/29/2022 11:46:18 - INFO - codeparrot_training - Step 17849: {'lr': 0.00037713357714029035, 'samples': 3427200, 'steps': 17849, 'loss/train': 1.9778555631637573} 01/29/2022 11:46:22 - INFO - codeparrot_training - Step 17850: {'lr': 0.000377119488105675, 'samples': 3427392, 'steps': 17850, 'loss/train': 1.6409913897514343} 01/29/2022 11:46:26 - INFO - codeparrot_training - Step 17851: {'lr': 0.00037710539852652003, 'samples': 3427584, 'steps': 17851, 'loss/train': 1.3223820626735687} 01/29/2022 11:46:31 - INFO - codeparrot_training - Step 17852: {'lr': 0.00037709130840288605, 'samples': 3427776, 'steps': 17852, 'loss/train': 1.534074604511261} 01/29/2022 11:46:35 - INFO - codeparrot_training - Step 17853: {'lr': 0.0003770772177348331, 'samples': 3427968, 'steps': 17853, 'loss/train': 1.1905927956104279} 01/29/2022 11:46:40 - INFO - codeparrot_training - Step 17854: {'lr': 0.0003770631265224218, 'samples': 3428160, 'steps': 17854, 'loss/train': 1.1035901606082916} 01/29/2022 11:46:44 - INFO - codeparrot_training - Step 17855: {'lr': 0.0003770490347657124, 'samples': 3428352, 'steps': 17855, 'loss/train': 1.7657411098480225} 01/29/2022 11:46:49 - INFO - codeparrot_training - Step 17856: {'lr': 0.00037703494246476524, 'samples': 3428544, 'steps': 17856, 'loss/train': 1.3065650761127472} 01/29/2022 11:46:53 - INFO - codeparrot_training - Step 17857: {'lr': 0.00037702084961964075, 'samples': 3428736, 'steps': 17857, 'loss/train': 1.581479251384735} 01/29/2022 11:46:57 - INFO - codeparrot_training - Step 17858: {'lr': 0.00037700675623039925, 'samples': 3428928, 'steps': 17858, 'loss/train': 1.6627787947654724} 01/29/2022 11:47:04 - INFO - codeparrot_training - Step 17859: {'lr': 0.00037699266229710115, 'samples': 3429120, 'steps': 17859, 'loss/train': 0.6386693269014359} 01/29/2022 11:47:09 - INFO - codeparrot_training - Step 17860: {'lr': 0.0003769785678198068, 'samples': 3429312, 'steps': 17860, 'loss/train': 0.964388519525528} 01/29/2022 11:47:13 - INFO - codeparrot_training - Step 17861: {'lr': 0.0003769644727985766, 'samples': 3429504, 'steps': 17861, 'loss/train': 0.9118866920471191} 01/29/2022 11:47:17 - INFO - codeparrot_training - Step 17862: {'lr': 0.00037695037723347094, 'samples': 3429696, 'steps': 17862, 'loss/train': 1.6071118712425232} 01/29/2022 11:47:21 - INFO - codeparrot_training - Step 17863: {'lr': 0.00037693628112455015, 'samples': 3429888, 'steps': 17863, 'loss/train': 0.2967844605445862} 01/29/2022 11:47:27 - INFO - codeparrot_training - Step 17864: {'lr': 0.0003769221844718746, 'samples': 3430080, 'steps': 17864, 'loss/train': 0.8278945684432983} 01/29/2022 11:47:31 - INFO - codeparrot_training - Step 17865: {'lr': 0.00037690808727550477, 'samples': 3430272, 'steps': 17865, 'loss/train': 1.6116206645965576} 01/29/2022 11:47:35 - INFO - codeparrot_training - Step 17866: {'lr': 0.0003768939895355009, 'samples': 3430464, 'steps': 17866, 'loss/train': 1.2521502077579498} 01/29/2022 11:47:39 - INFO - codeparrot_training - Step 17867: {'lr': 0.0003768798912519236, 'samples': 3430656, 'steps': 17867, 'loss/train': 1.8923301100730896} 01/29/2022 11:47:44 - INFO - codeparrot_training - Step 17868: {'lr': 0.0003768657924248331, 'samples': 3430848, 'steps': 17868, 'loss/train': 1.3599186837673187} 01/29/2022 11:47:49 - INFO - codeparrot_training - Step 17869: {'lr': 0.0003768516930542898, 'samples': 3431040, 'steps': 17869, 'loss/train': 1.748081088066101} 01/29/2022 11:47:53 - INFO - codeparrot_training - Step 17870: {'lr': 0.00037683759314035414, 'samples': 3431232, 'steps': 17870, 'loss/train': 1.7497336864471436} 01/29/2022 11:47:57 - INFO - codeparrot_training - Step 17871: {'lr': 0.0003768234926830865, 'samples': 3431424, 'steps': 17871, 'loss/train': 2.144351899623871} 01/29/2022 11:48:02 - INFO - codeparrot_training - Step 17872: {'lr': 0.0003768093916825473, 'samples': 3431616, 'steps': 17872, 'loss/train': 1.8932636976242065} 01/29/2022 11:48:06 - INFO - codeparrot_training - Step 17873: {'lr': 0.00037679529013879686, 'samples': 3431808, 'steps': 17873, 'loss/train': 2.1445952653884888} 01/29/2022 11:48:12 - INFO - codeparrot_training - Step 17874: {'lr': 0.00037678118805189575, 'samples': 3432000, 'steps': 17874, 'loss/train': 1.3501280844211578} 01/29/2022 11:48:16 - INFO - codeparrot_training - Step 17875: {'lr': 0.0003767670854219043, 'samples': 3432192, 'steps': 17875, 'loss/train': 1.3966467082500458} 01/29/2022 11:48:20 - INFO - codeparrot_training - Step 17876: {'lr': 0.00037675298224888287, 'samples': 3432384, 'steps': 17876, 'loss/train': 0.9419212639331818} 01/29/2022 11:48:24 - INFO - codeparrot_training - Step 17877: {'lr': 0.0003767388785328919, 'samples': 3432576, 'steps': 17877, 'loss/train': 1.84236341714859} 01/29/2022 11:48:29 - INFO - codeparrot_training - Step 17878: {'lr': 0.0003767247742739918, 'samples': 3432768, 'steps': 17878, 'loss/train': 1.307982087135315} 01/29/2022 11:48:36 - INFO - codeparrot_training - Step 17879: {'lr': 0.0003767106694722431, 'samples': 3432960, 'steps': 17879, 'loss/train': 1.3917059898376465} 01/29/2022 11:48:40 - INFO - codeparrot_training - Step 17880: {'lr': 0.000376696564127706, 'samples': 3433152, 'steps': 17880, 'loss/train': 0.8303562104701996} 01/29/2022 11:48:44 - INFO - codeparrot_training - Step 17881: {'lr': 0.0003766824582404411, 'samples': 3433344, 'steps': 17881, 'loss/train': 1.8257920145988464} 01/29/2022 11:48:48 - INFO - codeparrot_training - Step 17882: {'lr': 0.00037666835181050887, 'samples': 3433536, 'steps': 17882, 'loss/train': 1.9615421891212463} 01/29/2022 11:48:53 - INFO - codeparrot_training - Step 17883: {'lr': 0.0003766542448379695, 'samples': 3433728, 'steps': 17883, 'loss/train': 1.6141524910926819} 01/29/2022 11:48:58 - INFO - codeparrot_training - Step 17884: {'lr': 0.0003766401373228836, 'samples': 3433920, 'steps': 17884, 'loss/train': 0.911241352558136} 01/29/2022 11:49:03 - INFO - codeparrot_training - Step 17885: {'lr': 0.00037662602926531166, 'samples': 3434112, 'steps': 17885, 'loss/train': 1.5968840718269348} 01/29/2022 11:49:07 - INFO - codeparrot_training - Step 17886: {'lr': 0.0003766119206653139, 'samples': 3434304, 'steps': 17886, 'loss/train': 1.417489618062973} 01/29/2022 11:49:11 - INFO - codeparrot_training - Step 17887: {'lr': 0.00037659781152295094, 'samples': 3434496, 'steps': 17887, 'loss/train': 1.0851528346538544} 01/29/2022 11:49:15 - INFO - codeparrot_training - Step 17888: {'lr': 0.0003765837018382831, 'samples': 3434688, 'steps': 17888, 'loss/train': 1.3738586604595184} 01/29/2022 11:49:22 - INFO - codeparrot_training - Step 17889: {'lr': 0.00037656959161137094, 'samples': 3434880, 'steps': 17889, 'loss/train': 1.5356045365333557} 01/29/2022 11:49:26 - INFO - codeparrot_training - Step 17890: {'lr': 0.00037655548084227484, 'samples': 3435072, 'steps': 17890, 'loss/train': 0.6393213719129562} 01/29/2022 11:49:31 - INFO - codeparrot_training - Step 17891: {'lr': 0.0003765413695310552, 'samples': 3435264, 'steps': 17891, 'loss/train': 1.5733734369277954} 01/29/2022 11:49:35 - INFO - codeparrot_training - Step 17892: {'lr': 0.00037652725767777255, 'samples': 3435456, 'steps': 17892, 'loss/train': 1.6641933917999268} 01/29/2022 11:49:39 - INFO - codeparrot_training - Step 17893: {'lr': 0.00037651314528248724, 'samples': 3435648, 'steps': 17893, 'loss/train': 2.5400195717811584} 01/29/2022 11:49:44 - INFO - codeparrot_training - Step 17894: {'lr': 0.00037649903234525996, 'samples': 3435840, 'steps': 17894, 'loss/train': 2.0635902285575867} 01/29/2022 11:49:48 - INFO - codeparrot_training - Step 17895: {'lr': 0.00037648491886615077, 'samples': 3436032, 'steps': 17895, 'loss/train': 1.4373500347137451} 01/29/2022 11:49:53 - INFO - codeparrot_training - Step 17896: {'lr': 0.0003764708048452205, 'samples': 3436224, 'steps': 17896, 'loss/train': 2.0334836840629578} 01/29/2022 11:49:57 - INFO - codeparrot_training - Step 17897: {'lr': 0.0003764566902825294, 'samples': 3436416, 'steps': 17897, 'loss/train': 1.389478325843811} 01/29/2022 11:50:01 - INFO - codeparrot_training - Step 17898: {'lr': 0.0003764425751781381, 'samples': 3436608, 'steps': 17898, 'loss/train': 1.2955220639705658} 01/29/2022 11:50:09 - INFO - codeparrot_training - Step 17899: {'lr': 0.0003764284595321068, 'samples': 3436800, 'steps': 17899, 'loss/train': 1.5145967602729797} 01/29/2022 11:50:14 - INFO - codeparrot_training - Step 17900: {'lr': 0.0003764143433444962, 'samples': 3436992, 'steps': 17900, 'loss/train': 0.972434788942337} 01/29/2022 11:50:18 - INFO - codeparrot_training - Step 17901: {'lr': 0.00037640022661536665, 'samples': 3437184, 'steps': 17901, 'loss/train': 2.3261847496032715} 01/29/2022 11:50:22 - INFO - codeparrot_training - Step 17902: {'lr': 0.0003763861093447787, 'samples': 3437376, 'steps': 17902, 'loss/train': 1.203780859708786} 01/29/2022 11:50:27 - INFO - codeparrot_training - Step 17903: {'lr': 0.0003763719915327928, 'samples': 3437568, 'steps': 17903, 'loss/train': 1.7315629720687866} 01/29/2022 11:50:32 - INFO - codeparrot_training - Step 17904: {'lr': 0.00037635787317946945, 'samples': 3437760, 'steps': 17904, 'loss/train': 1.4553228914737701} 01/29/2022 11:50:36 - INFO - codeparrot_training - Step 17905: {'lr': 0.000376343754284869, 'samples': 3437952, 'steps': 17905, 'loss/train': 1.8560189008712769} 01/29/2022 11:50:40 - INFO - codeparrot_training - Step 17906: {'lr': 0.00037632963484905213, 'samples': 3438144, 'steps': 17906, 'loss/train': 1.626166820526123} 01/29/2022 11:50:44 - INFO - codeparrot_training - Step 17907: {'lr': 0.0003763155148720791, 'samples': 3438336, 'steps': 17907, 'loss/train': 1.0951216220855713} 01/29/2022 11:50:51 - INFO - codeparrot_training - Step 17908: {'lr': 0.00037630139435401055, 'samples': 3438528, 'steps': 17908, 'loss/train': 1.6756101250648499} 01/29/2022 11:50:55 - INFO - codeparrot_training - Step 17909: {'lr': 0.000376287273294907, 'samples': 3438720, 'steps': 17909, 'loss/train': 1.5036810636520386} 01/29/2022 11:51:00 - INFO - codeparrot_training - Step 17910: {'lr': 0.0003762731516948288, 'samples': 3438912, 'steps': 17910, 'loss/train': 2.141830086708069} 01/29/2022 11:51:04 - INFO - codeparrot_training - Step 17911: {'lr': 0.00037625902955383664, 'samples': 3439104, 'steps': 17911, 'loss/train': 0.8431402444839478} 01/29/2022 11:51:08 - INFO - codeparrot_training - Step 17912: {'lr': 0.0003762449068719907, 'samples': 3439296, 'steps': 17912, 'loss/train': 0.8285300731658936} 01/29/2022 11:51:13 - INFO - codeparrot_training - Step 17913: {'lr': 0.0003762307836493518, 'samples': 3439488, 'steps': 17913, 'loss/train': 1.9023528099060059} 01/29/2022 11:51:18 - INFO - codeparrot_training - Step 17914: {'lr': 0.00037621665988598024, 'samples': 3439680, 'steps': 17914, 'loss/train': 1.0067820847034454} 01/29/2022 11:51:22 - INFO - codeparrot_training - Step 17915: {'lr': 0.0003762025355819366, 'samples': 3439872, 'steps': 17915, 'loss/train': 1.572079360485077} 01/29/2022 11:51:26 - INFO - codeparrot_training - Step 17916: {'lr': 0.0003761884107372814, 'samples': 3440064, 'steps': 17916, 'loss/train': 2.2505013942718506} 01/29/2022 11:51:30 - INFO - codeparrot_training - Step 17917: {'lr': 0.0003761742853520751, 'samples': 3440256, 'steps': 17917, 'loss/train': 1.5555020570755005} 01/29/2022 11:51:36 - INFO - codeparrot_training - Step 17918: {'lr': 0.00037616015942637824, 'samples': 3440448, 'steps': 17918, 'loss/train': 1.467326134443283} 01/29/2022 11:51:40 - INFO - codeparrot_training - Step 17919: {'lr': 0.0003761460329602513, 'samples': 3440640, 'steps': 17919, 'loss/train': 1.562170386314392} 01/29/2022 11:51:44 - INFO - codeparrot_training - Step 17920: {'lr': 0.0003761319059537548, 'samples': 3440832, 'steps': 17920, 'loss/train': 0.8960031867027283} 01/29/2022 11:51:48 - INFO - codeparrot_training - Step 17921: {'lr': 0.0003761177784069493, 'samples': 3441024, 'steps': 17921, 'loss/train': 1.6342021822929382} 01/29/2022 11:51:53 - INFO - codeparrot_training - Step 17922: {'lr': 0.00037610365031989524, 'samples': 3441216, 'steps': 17922, 'loss/train': 2.0107340812683105} 01/29/2022 11:52:00 - INFO - codeparrot_training - Step 17923: {'lr': 0.0003760895216926532, 'samples': 3441408, 'steps': 17923, 'loss/train': 1.4979778826236725} 01/29/2022 11:52:04 - INFO - codeparrot_training - Step 17924: {'lr': 0.0003760753925252838, 'samples': 3441600, 'steps': 17924, 'loss/train': 2.274380922317505} 01/29/2022 11:52:08 - INFO - codeparrot_training - Step 17925: {'lr': 0.00037606126281784725, 'samples': 3441792, 'steps': 17925, 'loss/train': 0.8172633647918701} 01/29/2022 11:52:13 - INFO - codeparrot_training - Step 17926: {'lr': 0.0003760471325704045, 'samples': 3441984, 'steps': 17926, 'loss/train': 0.63948954641819} 01/29/2022 11:52:17 - INFO - codeparrot_training - Step 17927: {'lr': 0.0003760330017830157, 'samples': 3442176, 'steps': 17927, 'loss/train': 2.3333889842033386} 01/29/2022 11:52:22 - INFO - codeparrot_training - Step 17928: {'lr': 0.00037601887045574155, 'samples': 3442368, 'steps': 17928, 'loss/train': 1.6911256313323975} 01/29/2022 11:52:26 - INFO - codeparrot_training - Step 17929: {'lr': 0.0003760047385886426, 'samples': 3442560, 'steps': 17929, 'loss/train': 0.97004234790802} 01/29/2022 11:52:31 - INFO - codeparrot_training - Step 17930: {'lr': 0.0003759906061817794, 'samples': 3442752, 'steps': 17930, 'loss/train': 0.6602307558059692} 01/29/2022 11:52:35 - INFO - codeparrot_training - Step 17931: {'lr': 0.00037597647323521234, 'samples': 3442944, 'steps': 17931, 'loss/train': 0.9723016619682312} 01/29/2022 11:52:39 - INFO - codeparrot_training - Step 17932: {'lr': 0.0003759623397490022, 'samples': 3443136, 'steps': 17932, 'loss/train': 1.442619949579239} 01/29/2022 11:52:46 - INFO - codeparrot_training - Step 17933: {'lr': 0.00037594820572320933, 'samples': 3443328, 'steps': 17933, 'loss/train': 2.0018792152404785} 01/29/2022 11:52:51 - INFO - codeparrot_training - Step 17934: {'lr': 0.0003759340711578944, 'samples': 3443520, 'steps': 17934, 'loss/train': 3.6022896766662598} 01/29/2022 11:52:55 - INFO - codeparrot_training - Step 17935: {'lr': 0.0003759199360531178, 'samples': 3443712, 'steps': 17935, 'loss/train': 1.7619468569755554} 01/29/2022 11:52:59 - INFO - codeparrot_training - Step 17936: {'lr': 0.00037590580040894024, 'samples': 3443904, 'steps': 17936, 'loss/train': 2.977206587791443} 01/29/2022 11:53:03 - INFO - codeparrot_training - Step 17937: {'lr': 0.0003758916642254222, 'samples': 3444096, 'steps': 17937, 'loss/train': 1.395327776670456} 01/29/2022 11:53:09 - INFO - codeparrot_training - Step 17938: {'lr': 0.00037587752750262426, 'samples': 3444288, 'steps': 17938, 'loss/train': 2.1576701402664185} 01/29/2022 11:53:13 - INFO - codeparrot_training - Step 17939: {'lr': 0.00037586339024060696, 'samples': 3444480, 'steps': 17939, 'loss/train': 4.089133143424988} 01/29/2022 11:53:17 - INFO - codeparrot_training - Step 17940: {'lr': 0.0003758492524394308, 'samples': 3444672, 'steps': 17940, 'loss/train': 1.7054802775382996} 01/29/2022 11:53:21 - INFO - codeparrot_training - Step 17941: {'lr': 0.0003758351140991565, 'samples': 3444864, 'steps': 17941, 'loss/train': 7.380005836486816} 01/29/2022 11:53:26 - INFO - codeparrot_training - Step 17942: {'lr': 0.0003758209752198444, 'samples': 3445056, 'steps': 17942, 'loss/train': 1.1312094926834106} 01/29/2022 11:53:31 - INFO - codeparrot_training - Step 17943: {'lr': 0.0003758068358015553, 'samples': 3445248, 'steps': 17943, 'loss/train': 1.2997902631759644} 01/29/2022 11:53:35 - INFO - codeparrot_training - Step 17944: {'lr': 0.0003757926958443496, 'samples': 3445440, 'steps': 17944, 'loss/train': 2.1535815596580505} 01/29/2022 11:53:40 - INFO - codeparrot_training - Step 17945: {'lr': 0.000375778555348288, 'samples': 3445632, 'steps': 17945, 'loss/train': 4.031954884529114} 01/29/2022 11:53:44 - INFO - codeparrot_training - Step 17946: {'lr': 0.000375764414313431, 'samples': 3445824, 'steps': 17946, 'loss/train': 1.437666267156601} 01/29/2022 11:53:48 - INFO - codeparrot_training - Step 17947: {'lr': 0.0003757502727398391, 'samples': 3446016, 'steps': 17947, 'loss/train': 1.1062148809432983} 01/29/2022 11:53:55 - INFO - codeparrot_training - Step 17948: {'lr': 0.00037573613062757304, 'samples': 3446208, 'steps': 17948, 'loss/train': 1.6611145734786987} 01/29/2022 11:54:00 - INFO - codeparrot_training - Step 17949: {'lr': 0.0003757219879766933, 'samples': 3446400, 'steps': 17949, 'loss/train': 1.2762086391448975} 01/29/2022 11:54:04 - INFO - codeparrot_training - Step 17950: {'lr': 0.00037570784478726057, 'samples': 3446592, 'steps': 17950, 'loss/train': 7.554613351821899} 01/29/2022 11:54:08 - INFO - codeparrot_training - Step 17951: {'lr': 0.00037569370105933523, 'samples': 3446784, 'steps': 17951, 'loss/train': 1.778895914554596} 01/29/2022 11:54:13 - INFO - codeparrot_training - Step 17952: {'lr': 0.00037567955679297806, 'samples': 3446976, 'steps': 17952, 'loss/train': 1.6590970158576965} 01/29/2022 11:54:18 - INFO - codeparrot_training - Step 17953: {'lr': 0.0003756654119882496, 'samples': 3447168, 'steps': 17953, 'loss/train': 2.8731173872947693} 01/29/2022 11:54:22 - INFO - codeparrot_training - Step 17954: {'lr': 0.0003756512666452103, 'samples': 3447360, 'steps': 17954, 'loss/train': 1.2181152999401093} 01/29/2022 11:54:26 - INFO - codeparrot_training - Step 17955: {'lr': 0.0003756371207639209, 'samples': 3447552, 'steps': 17955, 'loss/train': 1.4685286581516266} 01/29/2022 11:54:31 - INFO - codeparrot_training - Step 17956: {'lr': 0.00037562297434444203, 'samples': 3447744, 'steps': 17956, 'loss/train': 1.6467506289482117} 01/29/2022 11:54:35 - INFO - codeparrot_training - Step 17957: {'lr': 0.0003756088273868342, 'samples': 3447936, 'steps': 17957, 'loss/train': 2.189185380935669} 01/29/2022 11:54:40 - INFO - codeparrot_training - Step 17958: {'lr': 0.00037559467989115806, 'samples': 3448128, 'steps': 17958, 'loss/train': 1.6271308064460754} 01/29/2022 11:54:44 - INFO - codeparrot_training - Step 17959: {'lr': 0.00037558053185747416, 'samples': 3448320, 'steps': 17959, 'loss/train': 1.130033165216446} 01/29/2022 11:54:49 - INFO - codeparrot_training - Step 17960: {'lr': 0.00037556638328584314, 'samples': 3448512, 'steps': 17960, 'loss/train': 1.5559471249580383} 01/29/2022 11:54:53 - INFO - codeparrot_training - Step 17961: {'lr': 0.00037555223417632565, 'samples': 3448704, 'steps': 17961, 'loss/train': 1.6516247391700745} 01/29/2022 11:54:57 - INFO - codeparrot_training - Step 17962: {'lr': 0.0003755380845289822, 'samples': 3448896, 'steps': 17962, 'loss/train': 1.7382334470748901} 01/29/2022 11:55:02 - INFO - codeparrot_training - Step 17963: {'lr': 0.0003755239343438735, 'samples': 3449088, 'steps': 17963, 'loss/train': 2.035792887210846} 01/29/2022 11:55:07 - INFO - codeparrot_training - Step 17964: {'lr': 0.00037550978362106, 'samples': 3449280, 'steps': 17964, 'loss/train': 1.6256182193756104} 01/29/2022 11:55:11 - INFO - codeparrot_training - Step 17965: {'lr': 0.0003754956323606026, 'samples': 3449472, 'steps': 17965, 'loss/train': 0.7530994713306427} 01/29/2022 11:55:15 - INFO - codeparrot_training - Step 17966: {'lr': 0.0003754814805625617, 'samples': 3449664, 'steps': 17966, 'loss/train': 1.7876331210136414} 01/29/2022 11:55:19 - INFO - codeparrot_training - Step 17967: {'lr': 0.00037546732822699803, 'samples': 3449856, 'steps': 17967, 'loss/train': 1.6319426894187927} 01/29/2022 11:55:26 - INFO - codeparrot_training - Step 17968: {'lr': 0.0003754531753539721, 'samples': 3450048, 'steps': 17968, 'loss/train': 1.825276494026184} 01/29/2022 11:55:31 - INFO - codeparrot_training - Step 17969: {'lr': 0.0003754390219435446, 'samples': 3450240, 'steps': 17969, 'loss/train': 1.5341341495513916} 01/29/2022 11:55:35 - INFO - codeparrot_training - Step 17970: {'lr': 0.00037542486799577624, 'samples': 3450432, 'steps': 17970, 'loss/train': 1.2224070131778717} 01/29/2022 11:55:39 - INFO - codeparrot_training - Step 17971: {'lr': 0.00037541071351072746, 'samples': 3450624, 'steps': 17971, 'loss/train': 1.3731146156787872} 01/29/2022 11:55:43 - INFO - codeparrot_training - Step 17972: {'lr': 0.0003753965584884591, 'samples': 3450816, 'steps': 17972, 'loss/train': 1.7320606112480164} 01/29/2022 11:55:49 - INFO - codeparrot_training - Step 17973: {'lr': 0.00037538240292903167, 'samples': 3451008, 'steps': 17973, 'loss/train': 1.2125519514083862} 01/29/2022 11:55:53 - INFO - codeparrot_training - Step 17974: {'lr': 0.0003753682468325059, 'samples': 3451200, 'steps': 17974, 'loss/train': 1.7776554822921753} 01/29/2022 11:55:57 - INFO - codeparrot_training - Step 17975: {'lr': 0.0003753540901989422, 'samples': 3451392, 'steps': 17975, 'loss/train': 1.4396408200263977} 01/29/2022 11:56:02 - INFO - codeparrot_training - Step 17976: {'lr': 0.00037533993302840153, 'samples': 3451584, 'steps': 17976, 'loss/train': 1.281578779220581} 01/29/2022 11:56:06 - INFO - codeparrot_training - Step 17977: {'lr': 0.00037532577532094436, 'samples': 3451776, 'steps': 17977, 'loss/train': 1.7082856893539429} 01/29/2022 11:56:12 - INFO - codeparrot_training - Step 17978: {'lr': 0.00037531161707663136, 'samples': 3451968, 'steps': 17978, 'loss/train': 1.9674893617630005} 01/29/2022 11:56:16 - INFO - codeparrot_training - Step 17979: {'lr': 0.0003752974582955232, 'samples': 3452160, 'steps': 17979, 'loss/train': 2.353934347629547} 01/29/2022 11:56:21 - INFO - codeparrot_training - Step 17980: {'lr': 0.0003752832989776804, 'samples': 3452352, 'steps': 17980, 'loss/train': 1.6705557703971863} 01/29/2022 11:56:25 - INFO - codeparrot_training - Step 17981: {'lr': 0.0003752691391231639, 'samples': 3452544, 'steps': 17981, 'loss/train': 1.1496459245681763} 01/29/2022 11:56:29 - INFO - codeparrot_training - Step 17982: {'lr': 0.00037525497873203405, 'samples': 3452736, 'steps': 17982, 'loss/train': 1.665442407131195} 01/29/2022 11:56:35 - INFO - codeparrot_training - Step 17983: {'lr': 0.0003752408178043518, 'samples': 3452928, 'steps': 17983, 'loss/train': 1.427359163761139} 01/29/2022 11:56:39 - INFO - codeparrot_training - Step 17984: {'lr': 0.0003752266563401775, 'samples': 3453120, 'steps': 17984, 'loss/train': 4.274573922157288} 01/29/2022 11:56:43 - INFO - codeparrot_training - Step 17985: {'lr': 0.00037521249433957203, 'samples': 3453312, 'steps': 17985, 'loss/train': 1.780468225479126} 01/29/2022 11:56:47 - INFO - codeparrot_training - Step 17986: {'lr': 0.000375198331802596, 'samples': 3453504, 'steps': 17986, 'loss/train': 1.5254042744636536} 01/29/2022 11:56:52 - INFO - codeparrot_training - Step 17987: {'lr': 0.00037518416872931007, 'samples': 3453696, 'steps': 17987, 'loss/train': 1.188851237297058} 01/29/2022 11:56:57 - INFO - codeparrot_training - Step 17988: {'lr': 0.00037517000511977486, 'samples': 3453888, 'steps': 17988, 'loss/train': 2.10402113199234} 01/29/2022 11:57:01 - INFO - codeparrot_training - Step 17989: {'lr': 0.00037515584097405115, 'samples': 3454080, 'steps': 17989, 'loss/train': 1.1970133781433105} 01/29/2022 11:57:05 - INFO - codeparrot_training - Step 17990: {'lr': 0.00037514167629219955, 'samples': 3454272, 'steps': 17990, 'loss/train': 1.668408215045929} 01/29/2022 11:57:10 - INFO - codeparrot_training - Step 17991: {'lr': 0.0003751275110742807, 'samples': 3454464, 'steps': 17991, 'loss/train': 2.29311740398407} 01/29/2022 11:57:14 - INFO - codeparrot_training - Step 17992: {'lr': 0.00037511334532035537, 'samples': 3454656, 'steps': 17992, 'loss/train': 0.9371518492698669} 01/29/2022 11:57:21 - INFO - codeparrot_training - Step 17993: {'lr': 0.00037509917903048417, 'samples': 3454848, 'steps': 17993, 'loss/train': 1.1806014776229858} 01/29/2022 11:57:25 - INFO - codeparrot_training - Step 17994: {'lr': 0.00037508501220472783, 'samples': 3455040, 'steps': 17994, 'loss/train': 2.794136703014374} 01/29/2022 11:57:30 - INFO - codeparrot_training - Step 17995: {'lr': 0.000375070844843147, 'samples': 3455232, 'steps': 17995, 'loss/train': 2.6259997487068176} 01/29/2022 11:57:34 - INFO - codeparrot_training - Step 17996: {'lr': 0.00037505667694580244, 'samples': 3455424, 'steps': 17996, 'loss/train': 1.8776578903198242} 01/29/2022 11:57:38 - INFO - codeparrot_training - Step 17997: {'lr': 0.00037504250851275466, 'samples': 3455616, 'steps': 17997, 'loss/train': 2.7002139687538147} 01/29/2022 11:57:44 - INFO - codeparrot_training - Step 17998: {'lr': 0.0003750283395440647, 'samples': 3455808, 'steps': 17998, 'loss/train': 1.966883361339569} 01/29/2022 11:57:48 - INFO - codeparrot_training - Step 17999: {'lr': 0.0003750141700397928, 'samples': 3456000, 'steps': 17999, 'loss/train': 1.3923524022102356} 01/29/2022 11:57:48 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 11:58:21 - WARNING - huggingface_hub.repository - Several commits (9) will be pushed upstream. 01/29/2022 11:58:21 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 11:59:25 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 24f4fe8..a5ccdc4 expert-dust-2 -> expert-dust-2 01/29/2022 11:59:31 - INFO - codeparrot_training - Step 18000: {'lr': 0.000375, 'samples': 3456192, 'steps': 18000, 'loss/train': 1.793394148349762} 01/29/2022 11:59:35 - INFO - codeparrot_training - Step 18001: {'lr': 0.0003749858294247469, 'samples': 3456384, 'steps': 18001, 'loss/train': 0.841869056224823} 01/29/2022 11:59:40 - INFO - codeparrot_training - Step 18002: {'lr': 0.0003749716583140942, 'samples': 3456576, 'steps': 18002, 'loss/train': 1.5484989881515503} 01/29/2022 11:59:45 - INFO - codeparrot_training - Step 18003: {'lr': 0.00037495748666810255, 'samples': 3456768, 'steps': 18003, 'loss/train': 1.726917028427124} 01/29/2022 11:59:49 - INFO - codeparrot_training - Step 18004: {'lr': 0.0003749433144868327, 'samples': 3456960, 'steps': 18004, 'loss/train': 1.1275345981121063} 01/29/2022 11:59:53 - INFO - codeparrot_training - Step 18005: {'lr': 0.00037492914177034544, 'samples': 3457152, 'steps': 18005, 'loss/train': 2.1269550919532776} 01/29/2022 11:59:58 - INFO - codeparrot_training - Step 18006: {'lr': 0.00037491496851870134, 'samples': 3457344, 'steps': 18006, 'loss/train': 1.5078458189964294} 01/29/2022 12:00:02 - INFO - codeparrot_training - Step 18007: {'lr': 0.0003749007947319612, 'samples': 3457536, 'steps': 18007, 'loss/train': 0.8527906537055969} 01/29/2022 12:00:09 - INFO - codeparrot_training - Step 18008: {'lr': 0.00037488662041018574, 'samples': 3457728, 'steps': 18008, 'loss/train': 3.0683562755584717} 01/29/2022 12:00:14 - INFO - codeparrot_training - Step 18009: {'lr': 0.0003748724455534356, 'samples': 3457920, 'steps': 18009, 'loss/train': 1.512012004852295} 01/29/2022 12:00:18 - INFO - codeparrot_training - Step 18010: {'lr': 0.0003748582701617716, 'samples': 3458112, 'steps': 18010, 'loss/train': 1.701117753982544} 01/29/2022 12:00:22 - INFO - codeparrot_training - Step 18011: {'lr': 0.00037484409423525446, 'samples': 3458304, 'steps': 18011, 'loss/train': 2.06421822309494} 01/29/2022 12:00:26 - INFO - codeparrot_training - Step 18012: {'lr': 0.00037482991777394476, 'samples': 3458496, 'steps': 18012, 'loss/train': 1.3796198666095734} 01/29/2022 12:00:32 - INFO - codeparrot_training - Step 18013: {'lr': 0.00037481574077790334, 'samples': 3458688, 'steps': 18013, 'loss/train': 1.4650324881076813} 01/29/2022 12:00:36 - INFO - codeparrot_training - Step 18014: {'lr': 0.00037480156324719093, 'samples': 3458880, 'steps': 18014, 'loss/train': 0.9260669052600861} 01/29/2022 12:00:40 - INFO - codeparrot_training - Step 18015: {'lr': 0.00037478738518186835, 'samples': 3459072, 'steps': 18015, 'loss/train': 3.0185916423797607} 01/29/2022 12:00:45 - INFO - codeparrot_training - Step 18016: {'lr': 0.00037477320658199615, 'samples': 3459264, 'steps': 18016, 'loss/train': 1.9335461854934692} 01/29/2022 12:00:52 - INFO - codeparrot_training - Step 18017: {'lr': 0.0003747590274476351, 'samples': 3459456, 'steps': 18017, 'loss/train': 1.3069671392440796} 01/29/2022 12:00:56 - INFO - codeparrot_training - Step 18018: {'lr': 0.0003747448477788461, 'samples': 3459648, 'steps': 18018, 'loss/train': 1.0038188695907593} 01/29/2022 12:01:00 - INFO - codeparrot_training - Step 18019: {'lr': 0.0003747306675756898, 'samples': 3459840, 'steps': 18019, 'loss/train': 1.2594912350177765} 01/29/2022 12:01:04 - INFO - codeparrot_training - Step 18020: {'lr': 0.00037471648683822683, 'samples': 3460032, 'steps': 18020, 'loss/train': 1.8110178709030151} 01/29/2022 12:01:09 - INFO - codeparrot_training - Step 18021: {'lr': 0.00037470230556651814, 'samples': 3460224, 'steps': 18021, 'loss/train': 1.555586814880371} 01/29/2022 12:01:13 - INFO - codeparrot_training - Step 18022: {'lr': 0.00037468812376062423, 'samples': 3460416, 'steps': 18022, 'loss/train': 0.7639601826667786} 01/29/2022 12:01:18 - INFO - codeparrot_training - Step 18023: {'lr': 0.00037467394142060614, 'samples': 3460608, 'steps': 18023, 'loss/train': 1.5207540392875671} 01/29/2022 12:01:23 - INFO - codeparrot_training - Step 18024: {'lr': 0.0003746597585465243, 'samples': 3460800, 'steps': 18024, 'loss/train': 1.6756977438926697} 01/29/2022 12:01:27 - INFO - codeparrot_training - Step 18025: {'lr': 0.00037464557513843975, 'samples': 3460992, 'steps': 18025, 'loss/train': 1.7122070789337158} 01/29/2022 12:01:31 - INFO - codeparrot_training - Step 18026: {'lr': 0.0003746313911964132, 'samples': 3461184, 'steps': 18026, 'loss/train': 1.314896821975708} 01/29/2022 12:01:36 - INFO - codeparrot_training - Step 18027: {'lr': 0.00037461720672050524, 'samples': 3461376, 'steps': 18027, 'loss/train': 1.0187567174434662} 01/29/2022 12:01:41 - INFO - codeparrot_training - Step 18028: {'lr': 0.0003746030217107768, 'samples': 3461568, 'steps': 18028, 'loss/train': 1.5505174398422241} 01/29/2022 12:01:45 - INFO - codeparrot_training - Step 18029: {'lr': 0.0003745888361672885, 'samples': 3461760, 'steps': 18029, 'loss/train': 1.426064282655716} 01/29/2022 12:01:49 - INFO - codeparrot_training - Step 18030: {'lr': 0.00037457465009010124, 'samples': 3461952, 'steps': 18030, 'loss/train': 1.7900235056877136} 01/29/2022 12:01:54 - INFO - codeparrot_training - Step 18031: {'lr': 0.00037456046347927576, 'samples': 3462144, 'steps': 18031, 'loss/train': 1.9910162687301636} 01/29/2022 12:01:58 - INFO - codeparrot_training - Step 18032: {'lr': 0.0003745462763348727, 'samples': 3462336, 'steps': 18032, 'loss/train': 2.1277515292167664} 01/29/2022 12:02:03 - INFO - codeparrot_training - Step 18033: {'lr': 0.00037453208865695305, 'samples': 3462528, 'steps': 18033, 'loss/train': 1.5016295313835144} 01/29/2022 12:02:08 - INFO - codeparrot_training - Step 18034: {'lr': 0.0003745179004455774, 'samples': 3462720, 'steps': 18034, 'loss/train': 1.6938207149505615} 01/29/2022 12:02:12 - INFO - codeparrot_training - Step 18035: {'lr': 0.00037450371170080673, 'samples': 3462912, 'steps': 18035, 'loss/train': 1.547167718410492} 01/29/2022 12:02:16 - INFO - codeparrot_training - Step 18036: {'lr': 0.00037448952242270155, 'samples': 3463104, 'steps': 18036, 'loss/train': 0.9955905675888062} 01/29/2022 12:02:20 - INFO - codeparrot_training - Step 18037: {'lr': 0.0003744753326113229, 'samples': 3463296, 'steps': 18037, 'loss/train': 1.8608790636062622} 01/29/2022 12:02:27 - INFO - codeparrot_training - Step 18038: {'lr': 0.00037446114226673136, 'samples': 3463488, 'steps': 18038, 'loss/train': 1.6425259709358215} 01/29/2022 12:02:32 - INFO - codeparrot_training - Step 18039: {'lr': 0.00037444695138898784, 'samples': 3463680, 'steps': 18039, 'loss/train': 0.5161550492048264} 01/29/2022 12:02:36 - INFO - codeparrot_training - Step 18040: {'lr': 0.00037443275997815306, 'samples': 3463872, 'steps': 18040, 'loss/train': 3.516466498374939} 01/29/2022 12:02:40 - INFO - codeparrot_training - Step 18041: {'lr': 0.00037441856803428785, 'samples': 3464064, 'steps': 18041, 'loss/train': 1.566408097743988} 01/29/2022 12:02:46 - INFO - codeparrot_training - Step 18042: {'lr': 0.0003744043755574531, 'samples': 3464256, 'steps': 18042, 'loss/train': 1.5955350995063782} 01/29/2022 12:02:50 - INFO - codeparrot_training - Step 18043: {'lr': 0.0003743901825477094, 'samples': 3464448, 'steps': 18043, 'loss/train': 1.9419023394584656} 01/29/2022 12:02:54 - INFO - codeparrot_training - Step 18044: {'lr': 0.0003743759890051177, 'samples': 3464640, 'steps': 18044, 'loss/train': 1.4915449619293213} 01/29/2022 12:02:58 - INFO - codeparrot_training - Step 18045: {'lr': 0.00037436179492973876, 'samples': 3464832, 'steps': 18045, 'loss/train': 1.2144988775253296} 01/29/2022 12:03:03 - INFO - codeparrot_training - Step 18046: {'lr': 0.0003743476003216334, 'samples': 3465024, 'steps': 18046, 'loss/train': 1.3111326098442078} 01/29/2022 12:03:08 - INFO - codeparrot_training - Step 18047: {'lr': 0.00037433340518086235, 'samples': 3465216, 'steps': 18047, 'loss/train': 1.5028805136680603} 01/29/2022 12:03:12 - INFO - codeparrot_training - Step 18048: {'lr': 0.0003743192095074866, 'samples': 3465408, 'steps': 18048, 'loss/train': 1.1007929742336273} 01/29/2022 12:03:16 - INFO - codeparrot_training - Step 18049: {'lr': 0.0003743050133015666, 'samples': 3465600, 'steps': 18049, 'loss/train': 1.6350257992744446} 01/29/2022 12:03:21 - INFO - codeparrot_training - Step 18050: {'lr': 0.0003742908165631636, 'samples': 3465792, 'steps': 18050, 'loss/train': 1.6567139625549316} 01/29/2022 12:03:25 - INFO - codeparrot_training - Step 18051: {'lr': 0.00037427661929233816, 'samples': 3465984, 'steps': 18051, 'loss/train': 0.4555405229330063} 01/29/2022 12:03:32 - INFO - codeparrot_training - Step 18052: {'lr': 0.00037426242148915113, 'samples': 3466176, 'steps': 18052, 'loss/train': 1.8896899223327637} 01/29/2022 12:03:36 - INFO - codeparrot_training - Step 18053: {'lr': 0.0003742482231536633, 'samples': 3466368, 'steps': 18053, 'loss/train': 0.8353548645973206} 01/29/2022 12:03:41 - INFO - codeparrot_training - Step 18054: {'lr': 0.00037423402428593555, 'samples': 3466560, 'steps': 18054, 'loss/train': 1.4258164465427399} 01/29/2022 12:03:45 - INFO - codeparrot_training - Step 18055: {'lr': 0.00037421982488602875, 'samples': 3466752, 'steps': 18055, 'loss/train': 0.5234869569540024} 01/29/2022 12:03:49 - INFO - codeparrot_training - Step 18056: {'lr': 0.0003742056249540036, 'samples': 3466944, 'steps': 18056, 'loss/train': 1.4328564405441284} 01/29/2022 12:03:54 - INFO - codeparrot_training - Step 18057: {'lr': 0.00037419142448992103, 'samples': 3467136, 'steps': 18057, 'loss/train': 0.8078154623508453} 01/29/2022 12:03:59 - INFO - codeparrot_training - Step 18058: {'lr': 0.0003741772234938418, 'samples': 3467328, 'steps': 18058, 'loss/train': 1.921965479850769} 01/29/2022 12:04:03 - INFO - codeparrot_training - Step 18059: {'lr': 0.00037416302196582684, 'samples': 3467520, 'steps': 18059, 'loss/train': 1.4438258707523346} 01/29/2022 12:04:07 - INFO - codeparrot_training - Step 18060: {'lr': 0.00037414881990593686, 'samples': 3467712, 'steps': 18060, 'loss/train': 1.241078495979309} 01/29/2022 12:04:11 - INFO - codeparrot_training - Step 18061: {'lr': 0.0003741346173142328, 'samples': 3467904, 'steps': 18061, 'loss/train': 2.0896684527397156} 01/29/2022 12:04:19 - INFO - codeparrot_training - Step 18062: {'lr': 0.0003741204141907754, 'samples': 3468096, 'steps': 18062, 'loss/train': 0.9914067685604095} 01/29/2022 12:04:23 - INFO - codeparrot_training - Step 18063: {'lr': 0.00037410621053562563, 'samples': 3468288, 'steps': 18063, 'loss/train': 0.8902843594551086} 01/29/2022 12:04:27 - INFO - codeparrot_training - Step 18064: {'lr': 0.0003740920063488442, 'samples': 3468480, 'steps': 18064, 'loss/train': 1.9591943621635437} 01/29/2022 12:04:32 - INFO - codeparrot_training - Step 18065: {'lr': 0.0003740778016304921, 'samples': 3468672, 'steps': 18065, 'loss/train': 1.2138808071613312} 01/29/2022 12:04:36 - INFO - codeparrot_training - Step 18066: {'lr': 0.00037406359638063005, 'samples': 3468864, 'steps': 18066, 'loss/train': 1.0772926211357117} 01/29/2022 12:04:41 - INFO - codeparrot_training - Step 18067: {'lr': 0.00037404939059931884, 'samples': 3469056, 'steps': 18067, 'loss/train': 0.8410990834236145} 01/29/2022 12:04:46 - INFO - codeparrot_training - Step 18068: {'lr': 0.0003740351842866196, 'samples': 3469248, 'steps': 18068, 'loss/train': 1.4320593774318695} 01/29/2022 12:04:50 - INFO - codeparrot_training - Step 18069: {'lr': 0.0003740209774425929, 'samples': 3469440, 'steps': 18069, 'loss/train': 1.837456226348877} 01/29/2022 12:04:54 - INFO - codeparrot_training - Step 18070: {'lr': 0.00037400677006729977, 'samples': 3469632, 'steps': 18070, 'loss/train': 0.9644785523414612} 01/29/2022 12:04:58 - INFO - codeparrot_training - Step 18071: {'lr': 0.000373992562160801, 'samples': 3469824, 'steps': 18071, 'loss/train': 1.8920888900756836} 01/29/2022 12:05:04 - INFO - codeparrot_training - Step 18072: {'lr': 0.0003739783537231575, 'samples': 3470016, 'steps': 18072, 'loss/train': 1.7793524265289307} 01/29/2022 12:05:08 - INFO - codeparrot_training - Step 18073: {'lr': 0.0003739641447544301, 'samples': 3470208, 'steps': 18073, 'loss/train': 1.659468412399292} 01/29/2022 12:05:12 - INFO - codeparrot_training - Step 18074: {'lr': 0.0003739499352546796, 'samples': 3470400, 'steps': 18074, 'loss/train': 1.6198149919509888} 01/29/2022 12:05:16 - INFO - codeparrot_training - Step 18075: {'lr': 0.00037393572522396697, 'samples': 3470592, 'steps': 18075, 'loss/train': 1.8882377743721008} 01/29/2022 12:05:21 - INFO - codeparrot_training - Step 18076: {'lr': 0.00037392151466235303, 'samples': 3470784, 'steps': 18076, 'loss/train': 1.1290701627731323} 01/29/2022 12:05:26 - INFO - codeparrot_training - Step 18077: {'lr': 0.00037390730356989864, 'samples': 3470976, 'steps': 18077, 'loss/train': 1.7199917435646057} 01/29/2022 12:05:30 - INFO - codeparrot_training - Step 18078: {'lr': 0.0003738930919466648, 'samples': 3471168, 'steps': 18078, 'loss/train': 2.3367732167243958} 01/29/2022 12:05:34 - INFO - codeparrot_training - Step 18079: {'lr': 0.0003738788797927122, 'samples': 3471360, 'steps': 18079, 'loss/train': 0.1023317389190197} 01/29/2022 12:05:39 - INFO - codeparrot_training - Step 18080: {'lr': 0.0003738646671081019, 'samples': 3471552, 'steps': 18080, 'loss/train': 0.967819333076477} 01/29/2022 12:05:43 - INFO - codeparrot_training - Step 18081: {'lr': 0.0003738504538928947, 'samples': 3471744, 'steps': 18081, 'loss/train': 1.5513916611671448} 01/29/2022 12:05:50 - INFO - codeparrot_training - Step 18082: {'lr': 0.0003738362401471514, 'samples': 3471936, 'steps': 18082, 'loss/train': 1.748563528060913} 01/29/2022 12:05:54 - INFO - codeparrot_training - Step 18083: {'lr': 0.00037382202587093305, 'samples': 3472128, 'steps': 18083, 'loss/train': 0.530794233083725} 01/29/2022 12:05:59 - INFO - codeparrot_training - Step 18084: {'lr': 0.0003738078110643004, 'samples': 3472320, 'steps': 18084, 'loss/train': 0.8698614835739136} 01/29/2022 12:06:03 - INFO - codeparrot_training - Step 18085: {'lr': 0.0003737935957273144, 'samples': 3472512, 'steps': 18085, 'loss/train': 1.6312817931175232} 01/29/2022 12:06:07 - INFO - codeparrot_training - Step 18086: {'lr': 0.000373779379860036, 'samples': 3472704, 'steps': 18086, 'loss/train': 1.5755317211151123} 01/29/2022 12:06:13 - INFO - codeparrot_training - Step 18087: {'lr': 0.000373765163462526, 'samples': 3472896, 'steps': 18087, 'loss/train': 2.0636138319969177} 01/29/2022 12:06:17 - INFO - codeparrot_training - Step 18088: {'lr': 0.0003737509465348453, 'samples': 3473088, 'steps': 18088, 'loss/train': 1.1821095049381256} 01/29/2022 12:06:21 - INFO - codeparrot_training - Step 18089: {'lr': 0.0003737367290770549, 'samples': 3473280, 'steps': 18089, 'loss/train': 3.3519448041915894} 01/29/2022 12:06:25 - INFO - codeparrot_training - Step 18090: {'lr': 0.0003737225110892156, 'samples': 3473472, 'steps': 18090, 'loss/train': 0.2817564681172371} 01/29/2022 12:06:30 - INFO - codeparrot_training - Step 18091: {'lr': 0.0003737082925713884, 'samples': 3473664, 'steps': 18091, 'loss/train': 1.6428183317184448} 01/29/2022 12:06:35 - INFO - codeparrot_training - Step 18092: {'lr': 0.00037369407352363417, 'samples': 3473856, 'steps': 18092, 'loss/train': 1.1679150760173798} 01/29/2022 12:06:39 - INFO - codeparrot_training - Step 18093: {'lr': 0.0003736798539460138, 'samples': 3474048, 'steps': 18093, 'loss/train': 1.2685106992721558} 01/29/2022 12:06:43 - INFO - codeparrot_training - Step 18094: {'lr': 0.00037366563383858814, 'samples': 3474240, 'steps': 18094, 'loss/train': 1.6147010922431946} 01/29/2022 12:06:48 - INFO - codeparrot_training - Step 18095: {'lr': 0.0003736514132014182, 'samples': 3474432, 'steps': 18095, 'loss/train': 1.0596114099025726} 01/29/2022 12:06:52 - INFO - codeparrot_training - Step 18096: {'lr': 0.0003736371920345649, 'samples': 3474624, 'steps': 18096, 'loss/train': 0.6768450736999512} 01/29/2022 12:06:59 - INFO - codeparrot_training - Step 18097: {'lr': 0.0003736229703380891, 'samples': 3474816, 'steps': 18097, 'loss/train': 0.9334922432899475} 01/29/2022 12:07:03 - INFO - codeparrot_training - Step 18098: {'lr': 0.0003736087481120518, 'samples': 3475008, 'steps': 18098, 'loss/train': 1.44498673081398} 01/29/2022 12:07:08 - INFO - codeparrot_training - Step 18099: {'lr': 0.0003735945253565138, 'samples': 3475200, 'steps': 18099, 'loss/train': 0.904050350189209} 01/29/2022 12:07:12 - INFO - codeparrot_training - Step 18100: {'lr': 0.00037358030207153617, 'samples': 3475392, 'steps': 18100, 'loss/train': 1.4110100269317627} 01/29/2022 12:07:16 - INFO - codeparrot_training - Step 18101: {'lr': 0.0003735660782571797, 'samples': 3475584, 'steps': 18101, 'loss/train': 0.9100617170333862} 01/29/2022 12:07:22 - INFO - codeparrot_training - Step 18102: {'lr': 0.00037355185391350546, 'samples': 3475776, 'steps': 18102, 'loss/train': 3.158251404762268} 01/29/2022 12:07:26 - INFO - codeparrot_training - Step 18103: {'lr': 0.00037353762904057425, 'samples': 3475968, 'steps': 18103, 'loss/train': 1.4920680820941925} 01/29/2022 12:07:30 - INFO - codeparrot_training - Step 18104: {'lr': 0.00037352340363844706, 'samples': 3476160, 'steps': 18104, 'loss/train': 1.3626091182231903} 01/29/2022 12:07:34 - INFO - codeparrot_training - Step 18105: {'lr': 0.0003735091777071849, 'samples': 3476352, 'steps': 18105, 'loss/train': 0.07319319806993008} 01/29/2022 12:07:39 - INFO - codeparrot_training - Step 18106: {'lr': 0.0003734949512468486, 'samples': 3476544, 'steps': 18106, 'loss/train': 2.229703187942505} 01/29/2022 12:07:46 - INFO - codeparrot_training - Step 18107: {'lr': 0.0003734807242574991, 'samples': 3476736, 'steps': 18107, 'loss/train': 2.3444554209709167} 01/29/2022 12:07:50 - INFO - codeparrot_training - Step 18108: {'lr': 0.0003734664967391975, 'samples': 3476928, 'steps': 18108, 'loss/train': 1.4465903341770172} 01/29/2022 12:07:54 - INFO - codeparrot_training - Step 18109: {'lr': 0.0003734522686920045, 'samples': 3477120, 'steps': 18109, 'loss/train': 1.916029930114746} 01/29/2022 12:07:58 - INFO - codeparrot_training - Step 18110: {'lr': 0.00037343804011598125, 'samples': 3477312, 'steps': 18110, 'loss/train': 1.43662628531456} 01/29/2022 12:08:03 - INFO - codeparrot_training - Step 18111: {'lr': 0.0003734238110111886, 'samples': 3477504, 'steps': 18111, 'loss/train': 0.12639067322015762} 01/29/2022 12:08:08 - INFO - codeparrot_training - Step 18112: {'lr': 0.0003734095813776876, 'samples': 3477696, 'steps': 18112, 'loss/train': 1.5245162844657898} 01/29/2022 12:08:12 - INFO - codeparrot_training - Step 18113: {'lr': 0.00037339535121553907, 'samples': 3477888, 'steps': 18113, 'loss/train': 1.4679155051708221} 01/29/2022 12:08:17 - INFO - codeparrot_training - Step 18114: {'lr': 0.0003733811205248041, 'samples': 3478080, 'steps': 18114, 'loss/train': 2.0123135447502136} 01/29/2022 12:08:21 - INFO - codeparrot_training - Step 18115: {'lr': 0.0003733668893055435, 'samples': 3478272, 'steps': 18115, 'loss/train': 1.3687938451766968} 01/29/2022 12:08:25 - INFO - codeparrot_training - Step 18116: {'lr': 0.00037335265755781844, 'samples': 3478464, 'steps': 18116, 'loss/train': 1.048015147447586} 01/29/2022 12:08:30 - INFO - codeparrot_training - Step 18117: {'lr': 0.0003733384252816897, 'samples': 3478656, 'steps': 18117, 'loss/train': 1.9848995804786682} 01/29/2022 12:08:35 - INFO - codeparrot_training - Step 18118: {'lr': 0.0003733241924772183, 'samples': 3478848, 'steps': 18118, 'loss/train': 1.381471574306488} 01/29/2022 12:08:39 - INFO - codeparrot_training - Step 18119: {'lr': 0.0003733099591444652, 'samples': 3479040, 'steps': 18119, 'loss/train': 1.4891998171806335} 01/29/2022 12:08:43 - INFO - codeparrot_training - Step 18120: {'lr': 0.0003732957252834914, 'samples': 3479232, 'steps': 18120, 'loss/train': 1.5895015597343445} 01/29/2022 12:08:47 - INFO - codeparrot_training - Step 18121: {'lr': 0.0003732814908943579, 'samples': 3479424, 'steps': 18121, 'loss/train': 2.4317089319229126} 01/29/2022 12:08:54 - INFO - codeparrot_training - Step 18122: {'lr': 0.0003732672559771256, 'samples': 3479616, 'steps': 18122, 'loss/train': 0.7688592374324799} 01/29/2022 12:08:59 - INFO - codeparrot_training - Step 18123: {'lr': 0.00037325302053185555, 'samples': 3479808, 'steps': 18123, 'loss/train': 0.8788305222988129} 01/29/2022 12:09:03 - INFO - codeparrot_training - Step 18124: {'lr': 0.0003732387845586087, 'samples': 3480000, 'steps': 18124, 'loss/train': 1.8913503885269165} 01/29/2022 12:09:07 - INFO - codeparrot_training - Step 18125: {'lr': 0.00037322454805744607, 'samples': 3480192, 'steps': 18125, 'loss/train': 1.8966466784477234} 01/29/2022 12:09:11 - INFO - codeparrot_training - Step 18126: {'lr': 0.0003732103110284285, 'samples': 3480384, 'steps': 18126, 'loss/train': 2.153110384941101} 01/29/2022 12:09:17 - INFO - codeparrot_training - Step 18127: {'lr': 0.00037319607347161715, 'samples': 3480576, 'steps': 18127, 'loss/train': 1.6672868728637695} 01/29/2022 12:09:21 - INFO - codeparrot_training - Step 18128: {'lr': 0.0003731818353870729, 'samples': 3480768, 'steps': 18128, 'loss/train': 1.7708510756492615} 01/29/2022 12:09:26 - INFO - codeparrot_training - Step 18129: {'lr': 0.0003731675967748568, 'samples': 3480960, 'steps': 18129, 'loss/train': 1.814273178577423} 01/29/2022 12:09:30 - INFO - codeparrot_training - Step 18130: {'lr': 0.0003731533576350298, 'samples': 3481152, 'steps': 18130, 'loss/train': 1.900387465953827} 01/29/2022 12:09:37 - INFO - codeparrot_training - Step 18131: {'lr': 0.000373139117967653, 'samples': 3481344, 'steps': 18131, 'loss/train': 2.0244909524917603} 01/29/2022 12:09:41 - INFO - codeparrot_training - Step 18132: {'lr': 0.00037312487777278725, 'samples': 3481536, 'steps': 18132, 'loss/train': 1.3670471906661987} 01/29/2022 12:09:46 - INFO - codeparrot_training - Step 18133: {'lr': 0.00037311063705049364, 'samples': 3481728, 'steps': 18133, 'loss/train': 1.3017703592777252} 01/29/2022 12:09:50 - INFO - codeparrot_training - Step 18134: {'lr': 0.00037309639580083314, 'samples': 3481920, 'steps': 18134, 'loss/train': 1.2872646152973175} 01/29/2022 12:09:54 - INFO - codeparrot_training - Step 18135: {'lr': 0.0003730821540238668, 'samples': 3482112, 'steps': 18135, 'loss/train': 3.621247172355652} 01/29/2022 12:09:59 - INFO - codeparrot_training - Step 18136: {'lr': 0.0003730679117196556, 'samples': 3482304, 'steps': 18136, 'loss/train': 1.4316925406455994} 01/29/2022 12:10:04 - INFO - codeparrot_training - Step 18137: {'lr': 0.0003730536688882605, 'samples': 3482496, 'steps': 18137, 'loss/train': 1.4876698851585388} 01/29/2022 12:10:08 - INFO - codeparrot_training - Step 18138: {'lr': 0.00037303942552974255, 'samples': 3482688, 'steps': 18138, 'loss/train': 1.368165671825409} 01/29/2022 12:10:12 - INFO - codeparrot_training - Step 18139: {'lr': 0.0003730251816441628, 'samples': 3482880, 'steps': 18139, 'loss/train': 0.6820598244667053} 01/29/2022 12:10:16 - INFO - codeparrot_training - Step 18140: {'lr': 0.00037301093723158223, 'samples': 3483072, 'steps': 18140, 'loss/train': 1.4739657640457153} 01/29/2022 12:10:22 - INFO - codeparrot_training - Step 18141: {'lr': 0.0003729966922920619, 'samples': 3483264, 'steps': 18141, 'loss/train': 0.7262347787618637} 01/29/2022 12:10:26 - INFO - codeparrot_training - Step 18142: {'lr': 0.00037298244682566273, 'samples': 3483456, 'steps': 18142, 'loss/train': 1.5670064091682434} 01/29/2022 12:10:30 - INFO - codeparrot_training - Step 18143: {'lr': 0.0003729682008324459, 'samples': 3483648, 'steps': 18143, 'loss/train': 1.7371754050254822} 01/29/2022 12:10:34 - INFO - codeparrot_training - Step 18144: {'lr': 0.00037295395431247223, 'samples': 3483840, 'steps': 18144, 'loss/train': 1.4927595555782318} 01/29/2022 12:10:39 - INFO - codeparrot_training - Step 18145: {'lr': 0.000372939707265803, 'samples': 3484032, 'steps': 18145, 'loss/train': 1.6821280717849731} 01/29/2022 12:10:45 - INFO - codeparrot_training - Step 18146: {'lr': 0.000372925459692499, 'samples': 3484224, 'steps': 18146, 'loss/train': 0.9717731773853302} 01/29/2022 12:10:49 - INFO - codeparrot_training - Step 18147: {'lr': 0.0003729112115926214, 'samples': 3484416, 'steps': 18147, 'loss/train': 1.7993813753128052} 01/29/2022 12:10:54 - INFO - codeparrot_training - Step 18148: {'lr': 0.00037289696296623117, 'samples': 3484608, 'steps': 18148, 'loss/train': 1.7579233646392822} 01/29/2022 12:10:58 - INFO - codeparrot_training - Step 18149: {'lr': 0.00037288271381338936, 'samples': 3484800, 'steps': 18149, 'loss/train': 1.6905092597007751} 01/29/2022 12:11:02 - INFO - codeparrot_training - Step 18150: {'lr': 0.0003728684641341571, 'samples': 3484992, 'steps': 18150, 'loss/train': 1.4699674844741821} 01/29/2022 12:11:06 - INFO - codeparrot_training - Step 18151: {'lr': 0.00037285421392859526, 'samples': 3485184, 'steps': 18151, 'loss/train': 2.8678566813468933} 01/29/2022 12:11:14 - INFO - codeparrot_training - Step 18152: {'lr': 0.00037283996319676505, 'samples': 3485376, 'steps': 18152, 'loss/train': 1.4940086603164673} 01/29/2022 12:11:18 - INFO - codeparrot_training - Step 18153: {'lr': 0.0003728257119387274, 'samples': 3485568, 'steps': 18153, 'loss/train': 1.382049947977066} 01/29/2022 12:11:22 - INFO - codeparrot_training - Step 18154: {'lr': 0.0003728114601545435, 'samples': 3485760, 'steps': 18154, 'loss/train': 2.3727352023124695} 01/29/2022 12:11:26 - INFO - codeparrot_training - Step 18155: {'lr': 0.0003727972078442742, 'samples': 3485952, 'steps': 18155, 'loss/train': 0.7645070850849152} 01/29/2022 12:11:31 - INFO - codeparrot_training - Step 18156: {'lr': 0.0003727829550079806, 'samples': 3486144, 'steps': 18156, 'loss/train': 2.5953208208084106} 01/29/2022 12:11:36 - INFO - codeparrot_training - Step 18157: {'lr': 0.00037276870164572394, 'samples': 3486336, 'steps': 18157, 'loss/train': 0.9899666011333466} 01/29/2022 12:11:40 - INFO - codeparrot_training - Step 18158: {'lr': 0.00037275444775756506, 'samples': 3486528, 'steps': 18158, 'loss/train': 1.8496114611625671} 01/29/2022 12:11:45 - INFO - codeparrot_training - Step 18159: {'lr': 0.00037274019334356516, 'samples': 3486720, 'steps': 18159, 'loss/train': 2.1979758739471436} 01/29/2022 12:11:49 - INFO - codeparrot_training - Step 18160: {'lr': 0.0003727259384037852, 'samples': 3486912, 'steps': 18160, 'loss/train': 1.88671875} 01/29/2022 12:11:53 - INFO - codeparrot_training - Step 18161: {'lr': 0.0003727116829382864, 'samples': 3487104, 'steps': 18161, 'loss/train': 1.770254373550415} 01/29/2022 12:11:58 - INFO - codeparrot_training - Step 18162: {'lr': 0.00037269742694712963, 'samples': 3487296, 'steps': 18162, 'loss/train': 1.5175741910934448} 01/29/2022 12:12:03 - INFO - codeparrot_training - Step 18163: {'lr': 0.000372683170430376, 'samples': 3487488, 'steps': 18163, 'loss/train': 1.098404198884964} 01/29/2022 12:12:07 - INFO - codeparrot_training - Step 18164: {'lr': 0.00037266891338808667, 'samples': 3487680, 'steps': 18164, 'loss/train': 1.3034772276878357} 01/29/2022 12:12:11 - INFO - codeparrot_training - Step 18165: {'lr': 0.0003726546558203227, 'samples': 3487872, 'steps': 18165, 'loss/train': 1.7180407047271729} 01/29/2022 12:12:15 - INFO - codeparrot_training - Step 18166: {'lr': 0.0003726403977271451, 'samples': 3488064, 'steps': 18166, 'loss/train': 0.8251974284648895} 01/29/2022 12:12:22 - INFO - codeparrot_training - Step 18167: {'lr': 0.00037262613910861485, 'samples': 3488256, 'steps': 18167, 'loss/train': 1.4770501255989075} 01/29/2022 12:12:27 - INFO - codeparrot_training - Step 18168: {'lr': 0.0003726118799647933, 'samples': 3488448, 'steps': 18168, 'loss/train': 2.657713294029236} 01/29/2022 12:12:31 - INFO - codeparrot_training - Step 18169: {'lr': 0.0003725976202957414, 'samples': 3488640, 'steps': 18169, 'loss/train': 1.630905568599701} 01/29/2022 12:12:35 - INFO - codeparrot_training - Step 18170: {'lr': 0.0003725833601015202, 'samples': 3488832, 'steps': 18170, 'loss/train': 2.0646459460258484} 01/29/2022 12:12:40 - INFO - codeparrot_training - Step 18171: {'lr': 0.0003725690993821908, 'samples': 3489024, 'steps': 18171, 'loss/train': 1.8594734072685242} 01/29/2022 12:12:45 - INFO - codeparrot_training - Step 18172: {'lr': 0.00037255483813781434, 'samples': 3489216, 'steps': 18172, 'loss/train': 0.3176276385784149} 01/29/2022 12:12:49 - INFO - codeparrot_training - Step 18173: {'lr': 0.00037254057636845186, 'samples': 3489408, 'steps': 18173, 'loss/train': 1.6008312106132507} 01/29/2022 12:12:53 - INFO - codeparrot_training - Step 18174: {'lr': 0.0003725263140741644, 'samples': 3489600, 'steps': 18174, 'loss/train': 1.2843563854694366} 01/29/2022 12:12:58 - INFO - codeparrot_training - Step 18175: {'lr': 0.0003725120512550131, 'samples': 3489792, 'steps': 18175, 'loss/train': 1.9102912545204163} 01/29/2022 12:13:02 - INFO - codeparrot_training - Step 18176: {'lr': 0.00037249778791105916, 'samples': 3489984, 'steps': 18176, 'loss/train': 1.1488133668899536} 01/29/2022 12:13:07 - INFO - codeparrot_training - Step 18177: {'lr': 0.0003724835240423636, 'samples': 3490176, 'steps': 18177, 'loss/train': 2.078878104686737} 01/29/2022 12:13:12 - INFO - codeparrot_training - Step 18178: {'lr': 0.0003724692596489874, 'samples': 3490368, 'steps': 18178, 'loss/train': 1.5872377753257751} 01/29/2022 12:13:16 - INFO - codeparrot_training - Step 18179: {'lr': 0.0003724549947309919, 'samples': 3490560, 'steps': 18179, 'loss/train': 0.8312597572803497} 01/29/2022 12:13:21 - INFO - codeparrot_training - Step 18180: {'lr': 0.00037244072928843805, 'samples': 3490752, 'steps': 18180, 'loss/train': 1.9412906169891357} 01/29/2022 12:13:25 - INFO - codeparrot_training - Step 18181: {'lr': 0.000372426463321387, 'samples': 3490944, 'steps': 18181, 'loss/train': 0.09054830856621265} 01/29/2022 12:13:32 - INFO - codeparrot_training - Step 18182: {'lr': 0.0003724121968298998, 'samples': 3491136, 'steps': 18182, 'loss/train': 0.12125251069664955} 01/29/2022 12:13:37 - INFO - codeparrot_training - Step 18183: {'lr': 0.0003723979298140377, 'samples': 3491328, 'steps': 18183, 'loss/train': 1.5740200281143188} 01/29/2022 12:13:41 - INFO - codeparrot_training - Step 18184: {'lr': 0.0003723836622738617, 'samples': 3491520, 'steps': 18184, 'loss/train': 1.6058639287948608} 01/29/2022 12:13:45 - INFO - codeparrot_training - Step 18185: {'lr': 0.00037236939420943294, 'samples': 3491712, 'steps': 18185, 'loss/train': 1.5751785635948181} 01/29/2022 12:13:49 - INFO - codeparrot_training - Step 18186: {'lr': 0.0003723551256208125, 'samples': 3491904, 'steps': 18186, 'loss/train': 1.8734684586524963} 01/29/2022 12:13:55 - INFO - codeparrot_training - Step 18187: {'lr': 0.0003723408565080616, 'samples': 3492096, 'steps': 18187, 'loss/train': 1.4998326301574707} 01/29/2022 12:13:59 - INFO - codeparrot_training - Step 18188: {'lr': 0.00037232658687124135, 'samples': 3492288, 'steps': 18188, 'loss/train': 1.2200101017951965} 01/29/2022 12:14:03 - INFO - codeparrot_training - Step 18189: {'lr': 0.0003723123167104128, 'samples': 3492480, 'steps': 18189, 'loss/train': 2.2466315031051636} 01/29/2022 12:14:07 - INFO - codeparrot_training - Step 18190: {'lr': 0.0003722980460256372, 'samples': 3492672, 'steps': 18190, 'loss/train': 1.2689035534858704} 01/29/2022 12:14:12 - INFO - codeparrot_training - Step 18191: {'lr': 0.00037228377481697555, 'samples': 3492864, 'steps': 18191, 'loss/train': 1.9470434188842773} 01/29/2022 12:14:19 - INFO - codeparrot_training - Step 18192: {'lr': 0.0003722695030844891, 'samples': 3493056, 'steps': 18192, 'loss/train': 1.53695547580719} 01/29/2022 12:14:23 - INFO - codeparrot_training - Step 18193: {'lr': 0.0003722552308282388, 'samples': 3493248, 'steps': 18193, 'loss/train': 1.8880008459091187} 01/29/2022 12:14:27 - INFO - codeparrot_training - Step 18194: {'lr': 0.000372240958048286, 'samples': 3493440, 'steps': 18194, 'loss/train': 2.050142526626587} 01/29/2022 12:14:32 - INFO - codeparrot_training - Step 18195: {'lr': 0.0003722266847446918, 'samples': 3493632, 'steps': 18195, 'loss/train': 1.96506929397583} 01/29/2022 12:14:36 - INFO - codeparrot_training - Step 18196: {'lr': 0.00037221241091751716, 'samples': 3493824, 'steps': 18196, 'loss/train': 1.2854923903942108} 01/29/2022 12:14:41 - INFO - codeparrot_training - Step 18197: {'lr': 0.00037219813656682346, 'samples': 3494016, 'steps': 18197, 'loss/train': 0.7245981395244598} 01/29/2022 12:14:45 - INFO - codeparrot_training - Step 18198: {'lr': 0.0003721838616926717, 'samples': 3494208, 'steps': 18198, 'loss/train': 1.8219521641731262} 01/29/2022 12:14:50 - INFO - codeparrot_training - Step 18199: {'lr': 0.0003721695862951231, 'samples': 3494400, 'steps': 18199, 'loss/train': 2.230021834373474} 01/29/2022 12:14:54 - INFO - codeparrot_training - Step 18200: {'lr': 0.0003721553103742388, 'samples': 3494592, 'steps': 18200, 'loss/train': 2.2362077236175537} 01/29/2022 12:14:58 - INFO - codeparrot_training - Step 18201: {'lr': 0.0003721410339300799, 'samples': 3494784, 'steps': 18201, 'loss/train': 0.6577657163143158} 01/29/2022 12:15:03 - INFO - codeparrot_training - Step 18202: {'lr': 0.0003721267569627076, 'samples': 3494976, 'steps': 18202, 'loss/train': 0.881548672914505} 01/29/2022 12:15:08 - INFO - codeparrot_training - Step 18203: {'lr': 0.00037211247947218306, 'samples': 3495168, 'steps': 18203, 'loss/train': 1.5494298934936523} 01/29/2022 12:15:12 - INFO - codeparrot_training - Step 18204: {'lr': 0.0003720982014585674, 'samples': 3495360, 'steps': 18204, 'loss/train': 1.7211468815803528} 01/29/2022 12:15:16 - INFO - codeparrot_training - Step 18205: {'lr': 0.0003720839229219218, 'samples': 3495552, 'steps': 18205, 'loss/train': 1.6194811463356018} 01/29/2022 12:15:20 - INFO - codeparrot_training - Step 18206: {'lr': 0.00037206964386230754, 'samples': 3495744, 'steps': 18206, 'loss/train': 0.5254446119070053} 01/29/2022 12:15:26 - INFO - codeparrot_training - Step 18207: {'lr': 0.00037205536427978563, 'samples': 3495936, 'steps': 18207, 'loss/train': 2.3024033904075623} 01/29/2022 12:15:30 - INFO - codeparrot_training - Step 18208: {'lr': 0.0003720410841744173, 'samples': 3496128, 'steps': 18208, 'loss/train': 1.7739434838294983} 01/29/2022 12:15:34 - INFO - codeparrot_training - Step 18209: {'lr': 0.0003720268035462637, 'samples': 3496320, 'steps': 18209, 'loss/train': 1.5063877701759338} 01/29/2022 12:15:39 - INFO - codeparrot_training - Step 18210: {'lr': 0.000372012522395386, 'samples': 3496512, 'steps': 18210, 'loss/train': 1.549130916595459} 01/29/2022 12:15:46 - INFO - codeparrot_training - Step 18211: {'lr': 0.00037199824072184546, 'samples': 3496704, 'steps': 18211, 'loss/train': 1.9521533846855164} 01/29/2022 12:15:50 - INFO - codeparrot_training - Step 18212: {'lr': 0.0003719839585257032, 'samples': 3496896, 'steps': 18212, 'loss/train': 2.6537917256355286} 01/29/2022 12:15:54 - INFO - codeparrot_training - Step 18213: {'lr': 0.00037196967580702036, 'samples': 3497088, 'steps': 18213, 'loss/train': 2.4821950793266296} 01/29/2022 12:15:59 - INFO - codeparrot_training - Step 18214: {'lr': 0.0003719553925658581, 'samples': 3497280, 'steps': 18214, 'loss/train': 1.282963514328003} 01/29/2022 12:16:03 - INFO - codeparrot_training - Step 18215: {'lr': 0.00037194110880227777, 'samples': 3497472, 'steps': 18215, 'loss/train': 1.380917340517044} 01/29/2022 12:16:08 - INFO - codeparrot_training - Step 18216: {'lr': 0.0003719268245163404, 'samples': 3497664, 'steps': 18216, 'loss/train': 1.2568900287151337} 01/29/2022 12:16:13 - INFO - codeparrot_training - Step 18217: {'lr': 0.0003719125397081072, 'samples': 3497856, 'steps': 18217, 'loss/train': 2.5132647156715393} 01/29/2022 12:16:17 - INFO - codeparrot_training - Step 18218: {'lr': 0.00037189825437763946, 'samples': 3498048, 'steps': 18218, 'loss/train': 1.6881022453308105} 01/29/2022 12:16:21 - INFO - codeparrot_training - Step 18219: {'lr': 0.0003718839685249983, 'samples': 3498240, 'steps': 18219, 'loss/train': 0.6754095554351807} 01/29/2022 12:16:25 - INFO - codeparrot_training - Step 18220: {'lr': 0.0003718696821502449, 'samples': 3498432, 'steps': 18220, 'loss/train': 1.5338360667228699} 01/29/2022 12:16:31 - INFO - codeparrot_training - Step 18221: {'lr': 0.0003718553952534405, 'samples': 3498624, 'steps': 18221, 'loss/train': 1.8201643824577332} 01/29/2022 12:16:35 - INFO - codeparrot_training - Step 18222: {'lr': 0.0003718411078346462, 'samples': 3498816, 'steps': 18222, 'loss/train': 1.8550028800964355} 01/29/2022 12:16:39 - INFO - codeparrot_training - Step 18223: {'lr': 0.0003718268198939234, 'samples': 3499008, 'steps': 18223, 'loss/train': 1.2721031606197357} 01/29/2022 12:16:44 - INFO - codeparrot_training - Step 18224: {'lr': 0.0003718125314313331, 'samples': 3499200, 'steps': 18224, 'loss/train': 2.2223671674728394} 01/29/2022 12:16:48 - INFO - codeparrot_training - Step 18225: {'lr': 0.0003717982424469366, 'samples': 3499392, 'steps': 18225, 'loss/train': 1.7056583762168884} 01/29/2022 12:16:55 - INFO - codeparrot_training - Step 18226: {'lr': 0.00037178395294079516, 'samples': 3499584, 'steps': 18226, 'loss/train': 1.5957409143447876} 01/29/2022 12:16:59 - INFO - codeparrot_training - Step 18227: {'lr': 0.0003717696629129699, 'samples': 3499776, 'steps': 18227, 'loss/train': 1.5475532412528992} 01/29/2022 12:17:04 - INFO - codeparrot_training - Step 18228: {'lr': 0.00037175537236352205, 'samples': 3499968, 'steps': 18228, 'loss/train': 1.0317682921886444} 01/29/2022 12:17:08 - INFO - codeparrot_training - Step 18229: {'lr': 0.0003717410812925129, 'samples': 3500160, 'steps': 18229, 'loss/train': 1.5918611884117126} 01/29/2022 12:17:12 - INFO - codeparrot_training - Step 18230: {'lr': 0.0003717267897000036, 'samples': 3500352, 'steps': 18230, 'loss/train': 1.6402875781059265} 01/29/2022 12:17:17 - INFO - codeparrot_training - Step 18231: {'lr': 0.0003717124975860554, 'samples': 3500544, 'steps': 18231, 'loss/train': 1.5625049471855164} 01/29/2022 12:17:22 - INFO - codeparrot_training - Step 18232: {'lr': 0.00037169820495072935, 'samples': 3500736, 'steps': 18232, 'loss/train': 1.502197265625} 01/29/2022 12:17:26 - INFO - codeparrot_training - Step 18233: {'lr': 0.000371683911794087, 'samples': 3500928, 'steps': 18233, 'loss/train': 1.472390055656433} 01/29/2022 12:17:30 - INFO - codeparrot_training - Step 18234: {'lr': 0.0003716696181161894, 'samples': 3501120, 'steps': 18234, 'loss/train': 1.2115479111671448} 01/29/2022 12:17:34 - INFO - codeparrot_training - Step 18235: {'lr': 0.00037165532391709777, 'samples': 3501312, 'steps': 18235, 'loss/train': 1.1803240478038788} 01/29/2022 12:17:41 - INFO - codeparrot_training - Step 18236: {'lr': 0.00037164102919687335, 'samples': 3501504, 'steps': 18236, 'loss/train': 1.1061519384384155} 01/29/2022 12:17:46 - INFO - codeparrot_training - Step 18237: {'lr': 0.00037162673395557737, 'samples': 3501696, 'steps': 18237, 'loss/train': 1.95254784822464} 01/29/2022 12:17:50 - INFO - codeparrot_training - Step 18238: {'lr': 0.0003716124381932711, 'samples': 3501888, 'steps': 18238, 'loss/train': 0.4846368581056595} 01/29/2022 12:17:54 - INFO - codeparrot_training - Step 18239: {'lr': 0.00037159814191001586, 'samples': 3502080, 'steps': 18239, 'loss/train': 1.0130386054515839} 01/29/2022 12:17:58 - INFO - codeparrot_training - Step 18240: {'lr': 0.00037158384510587263, 'samples': 3502272, 'steps': 18240, 'loss/train': 1.6292715668678284} 01/29/2022 12:18:04 - INFO - codeparrot_training - Step 18241: {'lr': 0.0003715695477809029, 'samples': 3502464, 'steps': 18241, 'loss/train': 1.9575623273849487} 01/29/2022 12:18:08 - INFO - codeparrot_training - Step 18242: {'lr': 0.0003715552499351678, 'samples': 3502656, 'steps': 18242, 'loss/train': 1.4046303033828735} 01/29/2022 12:18:12 - INFO - codeparrot_training - Step 18243: {'lr': 0.0003715409515687286, 'samples': 3502848, 'steps': 18243, 'loss/train': 1.3760444819927216} 01/29/2022 12:18:16 - INFO - codeparrot_training - Step 18244: {'lr': 0.00037152665268164664, 'samples': 3503040, 'steps': 18244, 'loss/train': 1.371396392583847} 01/29/2022 12:18:21 - INFO - codeparrot_training - Step 18245: {'lr': 0.00037151235327398304, 'samples': 3503232, 'steps': 18245, 'loss/train': 1.8382186889648438} 01/29/2022 12:18:26 - INFO - codeparrot_training - Step 18246: {'lr': 0.000371498053345799, 'samples': 3503424, 'steps': 18246, 'loss/train': 2.0951982736587524} 01/29/2022 12:18:30 - INFO - codeparrot_training - Step 18247: {'lr': 0.000371483752897156, 'samples': 3503616, 'steps': 18247, 'loss/train': 1.4541707932949066} 01/29/2022 12:18:35 - INFO - codeparrot_training - Step 18248: {'lr': 0.00037146945192811513, 'samples': 3503808, 'steps': 18248, 'loss/train': 1.5691673755645752} 01/29/2022 12:18:39 - INFO - codeparrot_training - Step 18249: {'lr': 0.0003714551504387378, 'samples': 3504000, 'steps': 18249, 'loss/train': 0.6012602895498276} 01/29/2022 12:18:43 - INFO - codeparrot_training - Step 18250: {'lr': 0.000371440848429085, 'samples': 3504192, 'steps': 18250, 'loss/train': 1.7056716084480286} 01/29/2022 12:18:50 - INFO - codeparrot_training - Step 18251: {'lr': 0.0003714265458992183, 'samples': 3504384, 'steps': 18251, 'loss/train': 1.5410667657852173} 01/29/2022 12:18:55 - INFO - codeparrot_training - Step 18252: {'lr': 0.00037141224284919876, 'samples': 3504576, 'steps': 18252, 'loss/train': 0.6303787082433701} 01/29/2022 12:18:59 - INFO - codeparrot_training - Step 18253: {'lr': 0.0003713979392790878, 'samples': 3504768, 'steps': 18253, 'loss/train': 1.5827937126159668} 01/29/2022 12:19:03 - INFO - codeparrot_training - Step 18254: {'lr': 0.0003713836351889465, 'samples': 3504960, 'steps': 18254, 'loss/train': 1.949292004108429} 01/29/2022 12:19:09 - INFO - codeparrot_training - Step 18255: {'lr': 0.00037136933057883636, 'samples': 3505152, 'steps': 18255, 'loss/train': 1.2214660048484802} 01/29/2022 12:19:13 - INFO - codeparrot_training - Step 18256: {'lr': 0.0003713550254488185, 'samples': 3505344, 'steps': 18256, 'loss/train': 1.9032899737358093} 01/29/2022 12:19:17 - INFO - codeparrot_training - Step 18257: {'lr': 0.0003713407197989543, 'samples': 3505536, 'steps': 18257, 'loss/train': 2.0072972774505615} 01/29/2022 12:19:22 - INFO - codeparrot_training - Step 18258: {'lr': 0.0003713264136293049, 'samples': 3505728, 'steps': 18258, 'loss/train': 0.12111549451947212} 01/29/2022 12:19:26 - INFO - codeparrot_training - Step 18259: {'lr': 0.0003713121069399317, 'samples': 3505920, 'steps': 18259, 'loss/train': 1.3028965294361115} 01/29/2022 12:19:30 - INFO - codeparrot_training - Step 18260: {'lr': 0.00037129779973089596, 'samples': 3506112, 'steps': 18260, 'loss/train': 0.05034978501498699} 01/29/2022 12:19:35 - INFO - codeparrot_training - Step 18261: {'lr': 0.00037128349200225895, 'samples': 3506304, 'steps': 18261, 'loss/train': 1.1887177526950836} 01/29/2022 12:19:40 - INFO - codeparrot_training - Step 18262: {'lr': 0.000371269183754082, 'samples': 3506496, 'steps': 18262, 'loss/train': 1.8084896206855774} 01/29/2022 12:19:44 - INFO - codeparrot_training - Step 18263: {'lr': 0.00037125487498642636, 'samples': 3506688, 'steps': 18263, 'loss/train': 1.4208240509033203} 01/29/2022 12:19:48 - INFO - codeparrot_training - Step 18264: {'lr': 0.00037124056569935336, 'samples': 3506880, 'steps': 18264, 'loss/train': 2.043797492980957} 01/29/2022 12:19:52 - INFO - codeparrot_training - Step 18265: {'lr': 0.00037122625589292425, 'samples': 3507072, 'steps': 18265, 'loss/train': 1.695846676826477} 01/29/2022 12:19:58 - INFO - codeparrot_training - Step 18266: {'lr': 0.0003712119455672004, 'samples': 3507264, 'steps': 18266, 'loss/train': 1.3006528615951538} 01/29/2022 12:20:02 - INFO - codeparrot_training - Step 18267: {'lr': 0.000371197634722243, 'samples': 3507456, 'steps': 18267, 'loss/train': 1.3656741678714752} 01/29/2022 12:20:06 - INFO - codeparrot_training - Step 18268: {'lr': 0.0003711833233581134, 'samples': 3507648, 'steps': 18268, 'loss/train': 3.0787943601608276} 01/29/2022 12:20:11 - INFO - codeparrot_training - Step 18269: {'lr': 0.000371169011474873, 'samples': 3507840, 'steps': 18269, 'loss/train': 0.3005756735801697} 01/29/2022 12:20:15 - INFO - codeparrot_training - Step 18270: {'lr': 0.00037115469907258303, 'samples': 3508032, 'steps': 18270, 'loss/train': 2.7394269704818726} 01/29/2022 12:20:22 - INFO - codeparrot_training - Step 18271: {'lr': 0.0003711403861513047, 'samples': 3508224, 'steps': 18271, 'loss/train': 1.0526898801326752} 01/29/2022 12:20:27 - INFO - codeparrot_training - Step 18272: {'lr': 0.0003711260727110995, 'samples': 3508416, 'steps': 18272, 'loss/train': 1.553400456905365} 01/29/2022 12:20:31 - INFO - codeparrot_training - Step 18273: {'lr': 0.00037111175875202863, 'samples': 3508608, 'steps': 18273, 'loss/train': 1.173907309770584} 01/29/2022 12:20:35 - INFO - codeparrot_training - Step 18274: {'lr': 0.00037109744427415346, 'samples': 3508800, 'steps': 18274, 'loss/train': 0.31969068199396133} 01/29/2022 12:20:39 - INFO - codeparrot_training - Step 18275: {'lr': 0.0003710831292775353, 'samples': 3508992, 'steps': 18275, 'loss/train': 1.4322770833969116} 01/29/2022 12:20:45 - INFO - codeparrot_training - Step 18276: {'lr': 0.00037106881376223544, 'samples': 3509184, 'steps': 18276, 'loss/train': 2.176482617855072} 01/29/2022 12:20:49 - INFO - codeparrot_training - Step 18277: {'lr': 0.00037105449772831527, 'samples': 3509376, 'steps': 18277, 'loss/train': 1.1494027376174927} 01/29/2022 12:20:53 - INFO - codeparrot_training - Step 18278: {'lr': 0.00037104018117583605, 'samples': 3509568, 'steps': 18278, 'loss/train': 1.9819970726966858} 01/29/2022 12:20:57 - INFO - codeparrot_training - Step 18279: {'lr': 0.00037102586410485915, 'samples': 3509760, 'steps': 18279, 'loss/train': 2.328076958656311} 01/29/2022 12:21:02 - INFO - codeparrot_training - Step 18280: {'lr': 0.00037101154651544583, 'samples': 3509952, 'steps': 18280, 'loss/train': 1.926895022392273} 01/29/2022 12:21:09 - INFO - codeparrot_training - Step 18281: {'lr': 0.0003709972284076575, 'samples': 3510144, 'steps': 18281, 'loss/train': 1.9904024004936218} 01/29/2022 12:21:13 - INFO - codeparrot_training - Step 18282: {'lr': 0.0003709829097815555, 'samples': 3510336, 'steps': 18282, 'loss/train': 1.5206315517425537} 01/29/2022 12:21:18 - INFO - codeparrot_training - Step 18283: {'lr': 0.00037096859063720104, 'samples': 3510528, 'steps': 18283, 'loss/train': 2.0740106105804443} 01/29/2022 12:21:22 - INFO - codeparrot_training - Step 18284: {'lr': 0.00037095427097465564, 'samples': 3510720, 'steps': 18284, 'loss/train': 0.5349076688289642} 01/29/2022 12:21:26 - INFO - codeparrot_training - Step 18285: {'lr': 0.0003709399507939805, 'samples': 3510912, 'steps': 18285, 'loss/train': 1.9882253408432007} 01/29/2022 12:21:32 - INFO - codeparrot_training - Step 18286: {'lr': 0.00037092563009523703, 'samples': 3511104, 'steps': 18286, 'loss/train': 1.46456578373909} 01/29/2022 12:21:36 - INFO - codeparrot_training - Step 18287: {'lr': 0.0003709113088784865, 'samples': 3511296, 'steps': 18287, 'loss/train': 1.3061458468437195} 01/29/2022 12:21:40 - INFO - codeparrot_training - Step 18288: {'lr': 0.0003708969871437904, 'samples': 3511488, 'steps': 18288, 'loss/train': 0.9212410748004913} 01/29/2022 12:21:44 - INFO - codeparrot_training - Step 18289: {'lr': 0.00037088266489120996, 'samples': 3511680, 'steps': 18289, 'loss/train': 1.6068665385246277} 01/29/2022 12:21:49 - INFO - codeparrot_training - Step 18290: {'lr': 0.0003708683421208066, 'samples': 3511872, 'steps': 18290, 'loss/train': 1.6905990242958069} 01/29/2022 12:21:54 - INFO - codeparrot_training - Step 18291: {'lr': 0.0003708540188326416, 'samples': 3512064, 'steps': 18291, 'loss/train': 2.193987786769867} 01/29/2022 12:21:58 - INFO - codeparrot_training - Step 18292: {'lr': 0.0003708396950267764, 'samples': 3512256, 'steps': 18292, 'loss/train': 2.1340933442115784} 01/29/2022 12:22:03 - INFO - codeparrot_training - Step 18293: {'lr': 0.00037082537070327225, 'samples': 3512448, 'steps': 18293, 'loss/train': 1.4171446859836578} 01/29/2022 12:22:07 - INFO - codeparrot_training - Step 18294: {'lr': 0.0003708110458621906, 'samples': 3512640, 'steps': 18294, 'loss/train': 1.5949932932853699} 01/29/2022 12:22:14 - INFO - codeparrot_training - Step 18295: {'lr': 0.00037079672050359283, 'samples': 3512832, 'steps': 18295, 'loss/train': 1.394292265176773} 01/29/2022 12:22:18 - INFO - codeparrot_training - Step 18296: {'lr': 0.00037078239462754023, 'samples': 3513024, 'steps': 18296, 'loss/train': 1.276871144771576} 01/29/2022 12:22:22 - INFO - codeparrot_training - Step 18297: {'lr': 0.00037076806823409426, 'samples': 3513216, 'steps': 18297, 'loss/train': 1.1297978460788727} 01/29/2022 12:22:27 - INFO - codeparrot_training - Step 18298: {'lr': 0.00037075374132331613, 'samples': 3513408, 'steps': 18298, 'loss/train': 1.9511182308197021} 01/29/2022 12:22:31 - INFO - codeparrot_training - Step 18299: {'lr': 0.0003707394138952674, 'samples': 3513600, 'steps': 18299, 'loss/train': 0.9236361086368561} 01/29/2022 12:22:37 - INFO - codeparrot_training - Step 18300: {'lr': 0.0003707250859500093, 'samples': 3513792, 'steps': 18300, 'loss/train': 2.4674872756004333} 01/29/2022 12:22:41 - INFO - codeparrot_training - Step 18301: {'lr': 0.00037071075748760336, 'samples': 3513984, 'steps': 18301, 'loss/train': 0.7580151557922363} 01/29/2022 12:22:46 - INFO - codeparrot_training - Step 18302: {'lr': 0.0003706964285081108, 'samples': 3514176, 'steps': 18302, 'loss/train': 1.7052749991416931} 01/29/2022 12:22:50 - INFO - codeparrot_training - Step 18303: {'lr': 0.0003706820990115931, 'samples': 3514368, 'steps': 18303, 'loss/train': 2.1549190878868103} 01/29/2022 12:22:54 - INFO - codeparrot_training - Step 18304: {'lr': 0.00037066776899811153, 'samples': 3514560, 'steps': 18304, 'loss/train': 1.72999906539917} 01/29/2022 12:22:58 - INFO - codeparrot_training - Step 18305: {'lr': 0.00037065343846772765, 'samples': 3514752, 'steps': 18305, 'loss/train': 1.7706098556518555} 01/29/2022 12:23:04 - INFO - codeparrot_training - Step 18306: {'lr': 0.0003706391074205027, 'samples': 3514944, 'steps': 18306, 'loss/train': 1.0015349686145782} 01/29/2022 12:23:08 - INFO - codeparrot_training - Step 18307: {'lr': 0.00037062477585649814, 'samples': 3515136, 'steps': 18307, 'loss/train': 2.5578768253326416} 01/29/2022 12:23:12 - INFO - codeparrot_training - Step 18308: {'lr': 0.00037061044377577535, 'samples': 3515328, 'steps': 18308, 'loss/train': 1.813682198524475} 01/29/2022 12:23:16 - INFO - codeparrot_training - Step 18309: {'lr': 0.00037059611117839565, 'samples': 3515520, 'steps': 18309, 'loss/train': 2.116473376750946} 01/29/2022 12:23:21 - INFO - codeparrot_training - Step 18310: {'lr': 0.0003705817780644206, 'samples': 3515712, 'steps': 18310, 'loss/train': 1.8480039238929749} 01/29/2022 12:23:26 - INFO - codeparrot_training - Step 18311: {'lr': 0.0003705674444339114, 'samples': 3515904, 'steps': 18311, 'loss/train': 1.327487200498581} 01/29/2022 12:23:30 - INFO - codeparrot_training - Step 18312: {'lr': 0.0003705531102869297, 'samples': 3516096, 'steps': 18312, 'loss/train': 3.4338237047195435} 01/29/2022 12:23:34 - INFO - codeparrot_training - Step 18313: {'lr': 0.0003705387756235366, 'samples': 3516288, 'steps': 18313, 'loss/train': 3.389279007911682} 01/29/2022 12:23:39 - INFO - codeparrot_training - Step 18314: {'lr': 0.00037052444044379375, 'samples': 3516480, 'steps': 18314, 'loss/train': 1.32213494181633} 01/29/2022 12:23:43 - INFO - codeparrot_training - Step 18315: {'lr': 0.00037051010474776244, 'samples': 3516672, 'steps': 18315, 'loss/train': 1.2114620804786682} 01/29/2022 12:23:50 - INFO - codeparrot_training - Step 18316: {'lr': 0.0003704957685355041, 'samples': 3516864, 'steps': 18316, 'loss/train': 0.6661014407873154} 01/29/2022 12:23:54 - INFO - codeparrot_training - Step 18317: {'lr': 0.00037048143180708014, 'samples': 3517056, 'steps': 18317, 'loss/train': 1.8724830150604248} 01/29/2022 12:23:59 - INFO - codeparrot_training - Step 18318: {'lr': 0.000370467094562552, 'samples': 3517248, 'steps': 18318, 'loss/train': 1.6845706701278687} 01/29/2022 12:24:03 - INFO - codeparrot_training - Step 18319: {'lr': 0.0003704527568019811, 'samples': 3517440, 'steps': 18319, 'loss/train': 1.424507975578308} 01/29/2022 12:24:07 - INFO - codeparrot_training - Step 18320: {'lr': 0.0003704384185254288, 'samples': 3517632, 'steps': 18320, 'loss/train': 1.627564251422882} 01/29/2022 12:24:12 - INFO - codeparrot_training - Step 18321: {'lr': 0.0003704240797329566, 'samples': 3517824, 'steps': 18321, 'loss/train': 1.2212259471416473} 01/29/2022 12:24:17 - INFO - codeparrot_training - Step 18322: {'lr': 0.00037040974042462584, 'samples': 3518016, 'steps': 18322, 'loss/train': 1.4269566535949707} 01/29/2022 12:24:21 - INFO - codeparrot_training - Step 18323: {'lr': 0.000370395400600498, 'samples': 3518208, 'steps': 18323, 'loss/train': 1.9644150137901306} 01/29/2022 12:24:25 - INFO - codeparrot_training - Step 18324: {'lr': 0.00037038106026063457, 'samples': 3518400, 'steps': 18324, 'loss/train': 0.4337739944458008} 01/29/2022 12:24:29 - INFO - codeparrot_training - Step 18325: {'lr': 0.0003703667194050968, 'samples': 3518592, 'steps': 18325, 'loss/train': 2.2764403223991394} 01/29/2022 12:24:37 - INFO - codeparrot_training - Step 18326: {'lr': 0.0003703523780339463, 'samples': 3518784, 'steps': 18326, 'loss/train': 1.473079651594162} 01/29/2022 12:24:41 - INFO - codeparrot_training - Step 18327: {'lr': 0.0003703380361472444, 'samples': 3518976, 'steps': 18327, 'loss/train': 1.811319351196289} 01/29/2022 12:24:45 - INFO - codeparrot_training - Step 18328: {'lr': 0.00037032369374505255, 'samples': 3519168, 'steps': 18328, 'loss/train': 2.97759085893631} 01/29/2022 12:24:49 - INFO - codeparrot_training - Step 18329: {'lr': 0.0003703093508274322, 'samples': 3519360, 'steps': 18329, 'loss/train': 1.4853261709213257} 01/29/2022 12:24:54 - INFO - codeparrot_training - Step 18330: {'lr': 0.0003702950073944448, 'samples': 3519552, 'steps': 18330, 'loss/train': 1.9374173283576965} 01/29/2022 12:24:59 - INFO - codeparrot_training - Step 18331: {'lr': 0.00037028066344615176, 'samples': 3519744, 'steps': 18331, 'loss/train': 0.9489447176456451} 01/29/2022 12:25:03 - INFO - codeparrot_training - Step 18332: {'lr': 0.0003702663189826146, 'samples': 3519936, 'steps': 18332, 'loss/train': 0.10172895714640617} 01/29/2022 12:25:08 - INFO - codeparrot_training - Step 18333: {'lr': 0.00037025197400389467, 'samples': 3520128, 'steps': 18333, 'loss/train': 1.466053158044815} 01/29/2022 12:25:12 - INFO - codeparrot_training - Step 18334: {'lr': 0.0003702376285100535, 'samples': 3520320, 'steps': 18334, 'loss/train': 0.7701416015625} 01/29/2022 12:25:16 - INFO - codeparrot_training - Step 18335: {'lr': 0.00037022328250115244, 'samples': 3520512, 'steps': 18335, 'loss/train': 1.2118055820465088} 01/29/2022 12:25:22 - INFO - codeparrot_training - Step 18336: {'lr': 0.00037020893597725313, 'samples': 3520704, 'steps': 18336, 'loss/train': 2.1275083422660828} 01/29/2022 12:25:26 - INFO - codeparrot_training - Step 18337: {'lr': 0.0003701945889384168, 'samples': 3520896, 'steps': 18337, 'loss/train': 1.6779981851577759} 01/29/2022 12:25:30 - INFO - codeparrot_training - Step 18338: {'lr': 0.00037018024138470515, 'samples': 3521088, 'steps': 18338, 'loss/train': 1.955142080783844} 01/29/2022 12:25:34 - INFO - codeparrot_training - Step 18339: {'lr': 0.0003701658933161794, 'samples': 3521280, 'steps': 18339, 'loss/train': 1.0982564091682434} 01/29/2022 12:25:39 - INFO - codeparrot_training - Step 18340: {'lr': 0.00037015154473290113, 'samples': 3521472, 'steps': 18340, 'loss/train': 0.679490715265274} 01/29/2022 12:25:46 - INFO - codeparrot_training - Step 18341: {'lr': 0.0003701371956349318, 'samples': 3521664, 'steps': 18341, 'loss/train': 2.018187403678894} 01/29/2022 12:25:50 - INFO - codeparrot_training - Step 18342: {'lr': 0.00037012284602233294, 'samples': 3521856, 'steps': 18342, 'loss/train': 1.9919228553771973} 01/29/2022 12:25:54 - INFO - codeparrot_training - Step 18343: {'lr': 0.0003701084958951659, 'samples': 3522048, 'steps': 18343, 'loss/train': 1.4960542023181915} 01/29/2022 12:25:58 - INFO - codeparrot_training - Step 18344: {'lr': 0.0003700941452534922, 'samples': 3522240, 'steps': 18344, 'loss/train': 2.110656201839447} 01/29/2022 12:26:03 - INFO - codeparrot_training - Step 18345: {'lr': 0.00037007979409737324, 'samples': 3522432, 'steps': 18345, 'loss/train': 0.9172049760818481} 01/29/2022 12:26:08 - INFO - codeparrot_training - Step 18346: {'lr': 0.0003700654424268707, 'samples': 3522624, 'steps': 18346, 'loss/train': 1.9705599546432495} 01/29/2022 12:26:12 - INFO - codeparrot_training - Step 18347: {'lr': 0.00037005109024204586, 'samples': 3522816, 'steps': 18347, 'loss/train': 2.2481226325035095} 01/29/2022 12:26:17 - INFO - codeparrot_training - Step 18348: {'lr': 0.00037003673754296026, 'samples': 3523008, 'steps': 18348, 'loss/train': 1.987036943435669} 01/29/2022 12:26:21 - INFO - codeparrot_training - Step 18349: {'lr': 0.00037002238432967547, 'samples': 3523200, 'steps': 18349, 'loss/train': 1.3251473307609558} 01/29/2022 12:26:25 - INFO - codeparrot_training - Step 18350: {'lr': 0.0003700080306022528, 'samples': 3523392, 'steps': 18350, 'loss/train': 1.5499985218048096} 01/29/2022 12:26:31 - INFO - codeparrot_training - Step 18351: {'lr': 0.00036999367636075386, 'samples': 3523584, 'steps': 18351, 'loss/train': 1.9864944219589233} 01/29/2022 12:26:35 - INFO - codeparrot_training - Step 18352: {'lr': 0.00036997932160524015, 'samples': 3523776, 'steps': 18352, 'loss/train': 1.962226688861847} 01/29/2022 12:26:39 - INFO - codeparrot_training - Step 18353: {'lr': 0.00036996496633577314, 'samples': 3523968, 'steps': 18353, 'loss/train': 1.216504544019699} 01/29/2022 12:26:43 - INFO - codeparrot_training - Step 18354: {'lr': 0.00036995061055241426, 'samples': 3524160, 'steps': 18354, 'loss/train': 1.6335768699645996} 01/29/2022 12:26:51 - INFO - codeparrot_training - Step 18355: {'lr': 0.000369936254255225, 'samples': 3524352, 'steps': 18355, 'loss/train': 0.8663639724254608} 01/29/2022 12:26:55 - INFO - codeparrot_training - Step 18356: {'lr': 0.000369921897444267, 'samples': 3524544, 'steps': 18356, 'loss/train': 1.9794026613235474} 01/29/2022 12:26:59 - INFO - codeparrot_training - Step 18357: {'lr': 0.00036990754011960165, 'samples': 3524736, 'steps': 18357, 'loss/train': 1.278918743133545} 01/29/2022 12:27:04 - INFO - codeparrot_training - Step 18358: {'lr': 0.0003698931822812905, 'samples': 3524928, 'steps': 18358, 'loss/train': 0.9955911934375763} 01/29/2022 12:27:08 - INFO - codeparrot_training - Step 18359: {'lr': 0.000369878823929395, 'samples': 3525120, 'steps': 18359, 'loss/train': 1.760909378528595} 01/29/2022 12:27:13 - INFO - codeparrot_training - Step 18360: {'lr': 0.00036986446506397666, 'samples': 3525312, 'steps': 18360, 'loss/train': 1.6410788297653198} 01/29/2022 12:27:18 - INFO - codeparrot_training - Step 18361: {'lr': 0.00036985010568509703, 'samples': 3525504, 'steps': 18361, 'loss/train': 1.8103240728378296} 01/29/2022 12:27:22 - INFO - codeparrot_training - Step 18362: {'lr': 0.00036983574579281764, 'samples': 3525696, 'steps': 18362, 'loss/train': 1.368035763502121} 01/29/2022 12:27:26 - INFO - codeparrot_training - Step 18363: {'lr': 0.0003698213853871999, 'samples': 3525888, 'steps': 18363, 'loss/train': 1.651326298713684} 01/29/2022 12:27:30 - INFO - codeparrot_training - Step 18364: {'lr': 0.00036980702446830547, 'samples': 3526080, 'steps': 18364, 'loss/train': 0.8011318445205688} 01/29/2022 12:27:35 - INFO - codeparrot_training - Step 18365: {'lr': 0.0003697926630361957, 'samples': 3526272, 'steps': 18365, 'loss/train': 0.9165473878383636} 01/29/2022 12:27:42 - INFO - codeparrot_training - Step 18366: {'lr': 0.00036977830109093227, 'samples': 3526464, 'steps': 18366, 'loss/train': 1.5957648754119873} 01/29/2022 12:27:46 - INFO - codeparrot_training - Step 18367: {'lr': 0.0003697639386325766, 'samples': 3526656, 'steps': 18367, 'loss/train': 1.8903667330741882} 01/29/2022 12:27:50 - INFO - codeparrot_training - Step 18368: {'lr': 0.00036974957566119027, 'samples': 3526848, 'steps': 18368, 'loss/train': 3.0409791469573975} 01/29/2022 12:27:55 - INFO - codeparrot_training - Step 18369: {'lr': 0.00036973521217683475, 'samples': 3527040, 'steps': 18369, 'loss/train': 1.6545941233634949} 01/29/2022 12:28:00 - INFO - codeparrot_training - Step 18370: {'lr': 0.00036972084817957164, 'samples': 3527232, 'steps': 18370, 'loss/train': 0.8430377840995789} 01/29/2022 12:28:04 - INFO - codeparrot_training - Step 18371: {'lr': 0.0003697064836694624, 'samples': 3527424, 'steps': 18371, 'loss/train': 1.7714788913726807} 01/29/2022 12:28:09 - INFO - codeparrot_training - Step 18372: {'lr': 0.0003696921186465686, 'samples': 3527616, 'steps': 18372, 'loss/train': 1.102092057466507} 01/29/2022 12:28:13 - INFO - codeparrot_training - Step 18373: {'lr': 0.00036967775311095186, 'samples': 3527808, 'steps': 18373, 'loss/train': 1.6713777780532837} 01/29/2022 12:28:17 - INFO - codeparrot_training - Step 18374: {'lr': 0.00036966338706267347, 'samples': 3528000, 'steps': 18374, 'loss/train': 1.725452721118927} 01/29/2022 12:28:22 - INFO - codeparrot_training - Step 18375: {'lr': 0.0003696490205017953, 'samples': 3528192, 'steps': 18375, 'loss/train': 0.8352937996387482} 01/29/2022 12:28:27 - INFO - codeparrot_training - Step 18376: {'lr': 0.00036963465342837855, 'samples': 3528384, 'steps': 18376, 'loss/train': 0.6875453442335129} 01/29/2022 12:28:31 - INFO - codeparrot_training - Step 18377: {'lr': 0.000369620285842485, 'samples': 3528576, 'steps': 18377, 'loss/train': 1.5055214166641235} 01/29/2022 12:28:35 - INFO - codeparrot_training - Step 18378: {'lr': 0.00036960591774417613, 'samples': 3528768, 'steps': 18378, 'loss/train': 1.4013667702674866} 01/29/2022 12:28:39 - INFO - codeparrot_training - Step 18379: {'lr': 0.00036959154913351357, 'samples': 3528960, 'steps': 18379, 'loss/train': 1.0053694546222687} 01/29/2022 12:28:45 - INFO - codeparrot_training - Step 18380: {'lr': 0.0003695771800105586, 'samples': 3529152, 'steps': 18380, 'loss/train': 2.2845134139060974} 01/29/2022 12:28:49 - INFO - codeparrot_training - Step 18381: {'lr': 0.00036956281037537307, 'samples': 3529344, 'steps': 18381, 'loss/train': 0.9361863434314728} 01/29/2022 12:28:54 - INFO - codeparrot_training - Step 18382: {'lr': 0.00036954844022801846, 'samples': 3529536, 'steps': 18382, 'loss/train': 1.849583387374878} 01/29/2022 12:28:58 - INFO - codeparrot_training - Step 18383: {'lr': 0.00036953406956855624, 'samples': 3529728, 'steps': 18383, 'loss/train': 1.7630615830421448} 01/29/2022 12:29:02 - INFO - codeparrot_training - Step 18384: {'lr': 0.0003695196983970481, 'samples': 3529920, 'steps': 18384, 'loss/train': 1.630388081073761} 01/29/2022 12:29:09 - INFO - codeparrot_training - Step 18385: {'lr': 0.0003695053267135554, 'samples': 3530112, 'steps': 18385, 'loss/train': 1.5136774778366089} 01/29/2022 12:29:14 - INFO - codeparrot_training - Step 18386: {'lr': 0.00036949095451813997, 'samples': 3530304, 'steps': 18386, 'loss/train': 0.07966944761574268} 01/29/2022 12:29:18 - INFO - codeparrot_training - Step 18387: {'lr': 0.0003694765818108631, 'samples': 3530496, 'steps': 18387, 'loss/train': 0.9349797070026398} 01/29/2022 12:29:22 - INFO - codeparrot_training - Step 18388: {'lr': 0.00036946220859178656, 'samples': 3530688, 'steps': 18388, 'loss/train': 2.125426769256592} 01/29/2022 12:29:26 - INFO - codeparrot_training - Step 18389: {'lr': 0.0003694478348609718, 'samples': 3530880, 'steps': 18389, 'loss/train': 1.500010371208191} 01/29/2022 12:29:31 - INFO - codeparrot_training - Step 18390: {'lr': 0.00036943346061848054, 'samples': 3531072, 'steps': 18390, 'loss/train': 0.3707476481795311} 01/29/2022 12:29:36 - INFO - codeparrot_training - Step 18391: {'lr': 0.00036941908586437416, 'samples': 3531264, 'steps': 18391, 'loss/train': 1.3005796372890472} 01/29/2022 12:29:40 - INFO - codeparrot_training - Step 18392: {'lr': 0.0003694047105987144, 'samples': 3531456, 'steps': 18392, 'loss/train': 1.6797850728034973} 01/29/2022 12:29:45 - INFO - codeparrot_training - Step 18393: {'lr': 0.00036939033482156277, 'samples': 3531648, 'steps': 18393, 'loss/train': 1.6549513936042786} 01/29/2022 12:29:49 - INFO - codeparrot_training - Step 18394: {'lr': 0.00036937595853298076, 'samples': 3531840, 'steps': 18394, 'loss/train': 2.4506202936172485} 01/29/2022 12:29:53 - INFO - codeparrot_training - Step 18395: {'lr': 0.0003693615817330302, 'samples': 3532032, 'steps': 18395, 'loss/train': 1.4271597862243652} 01/29/2022 12:29:58 - INFO - codeparrot_training - Step 18396: {'lr': 0.00036934720442177244, 'samples': 3532224, 'steps': 18396, 'loss/train': 1.3922104239463806} 01/29/2022 12:30:03 - INFO - codeparrot_training - Step 18397: {'lr': 0.0003693328265992692, 'samples': 3532416, 'steps': 18397, 'loss/train': 6.6460912227630615} 01/29/2022 12:30:07 - INFO - codeparrot_training - Step 18398: {'lr': 0.000369318448265582, 'samples': 3532608, 'steps': 18398, 'loss/train': 1.3145935535430908} 01/29/2022 12:30:11 - INFO - codeparrot_training - Step 18399: {'lr': 0.00036930406942077245, 'samples': 3532800, 'steps': 18399, 'loss/train': 1.0318608283996582} 01/29/2022 12:30:15 - INFO - codeparrot_training - Step 18400: {'lr': 0.0003692896900649021, 'samples': 3532992, 'steps': 18400, 'loss/train': 0.8430815041065216} 01/29/2022 12:30:23 - INFO - codeparrot_training - Step 18401: {'lr': 0.0003692753101980327, 'samples': 3533184, 'steps': 18401, 'loss/train': 1.595080018043518} 01/29/2022 12:30:27 - INFO - codeparrot_training - Step 18402: {'lr': 0.00036926092982022564, 'samples': 3533376, 'steps': 18402, 'loss/train': 1.3125067949295044} 01/29/2022 12:30:31 - INFO - codeparrot_training - Step 18403: {'lr': 0.0003692465489315427, 'samples': 3533568, 'steps': 18403, 'loss/train': 1.2397415041923523} 01/29/2022 12:30:35 - INFO - codeparrot_training - Step 18404: {'lr': 0.00036923216753204536, 'samples': 3533760, 'steps': 18404, 'loss/train': 1.3800381124019623} 01/29/2022 12:30:40 - INFO - codeparrot_training - Step 18405: {'lr': 0.0003692177856217953, 'samples': 3533952, 'steps': 18405, 'loss/train': 1.576728880405426} 01/29/2022 12:30:45 - INFO - codeparrot_training - Step 18406: {'lr': 0.00036920340320085413, 'samples': 3534144, 'steps': 18406, 'loss/train': 1.5958512425422668} 01/29/2022 12:30:49 - INFO - codeparrot_training - Step 18407: {'lr': 0.00036918902026928334, 'samples': 3534336, 'steps': 18407, 'loss/train': 1.5001872181892395} 01/29/2022 12:30:54 - INFO - codeparrot_training - Step 18408: {'lr': 0.00036917463682714473, 'samples': 3534528, 'steps': 18408, 'loss/train': 1.7610121965408325} 01/29/2022 12:30:58 - INFO - codeparrot_training - Step 18409: {'lr': 0.00036916025287449976, 'samples': 3534720, 'steps': 18409, 'loss/train': 2.02873295545578} 01/29/2022 12:31:05 - INFO - codeparrot_training - Step 18410: {'lr': 0.0003691458684114102, 'samples': 3534912, 'steps': 18410, 'loss/train': 0.51890529692173} 01/29/2022 12:31:09 - INFO - codeparrot_training - Step 18411: {'lr': 0.00036913148343793744, 'samples': 3535104, 'steps': 18411, 'loss/train': 0.7545211315155029} 01/29/2022 12:31:13 - INFO - codeparrot_training - Step 18412: {'lr': 0.00036911709795414336, 'samples': 3535296, 'steps': 18412, 'loss/train': 0.11373452469706535} 01/29/2022 12:31:18 - INFO - codeparrot_training - Step 18413: {'lr': 0.00036910271196008936, 'samples': 3535488, 'steps': 18413, 'loss/train': 0.6676560491323471} 01/29/2022 12:31:22 - INFO - codeparrot_training - Step 18414: {'lr': 0.0003690883254558372, 'samples': 3535680, 'steps': 18414, 'loss/train': 1.1542191803455353} 01/29/2022 12:31:27 - INFO - codeparrot_training - Step 18415: {'lr': 0.0003690739384414485, 'samples': 3535872, 'steps': 18415, 'loss/train': 1.0550419092178345} 01/29/2022 12:31:32 - INFO - codeparrot_training - Step 18416: {'lr': 0.0003690595509169848, 'samples': 3536064, 'steps': 18416, 'loss/train': 1.9516375064849854} 01/29/2022 12:31:36 - INFO - codeparrot_training - Step 18417: {'lr': 0.00036904516288250786, 'samples': 3536256, 'steps': 18417, 'loss/train': 1.3834426403045654} 01/29/2022 12:31:40 - INFO - codeparrot_training - Step 18418: {'lr': 0.0003690307743380791, 'samples': 3536448, 'steps': 18418, 'loss/train': 2.353658080101013} 01/29/2022 12:31:45 - INFO - codeparrot_training - Step 18419: {'lr': 0.00036901638528376047, 'samples': 3536640, 'steps': 18419, 'loss/train': 0.29326146841049194} 01/29/2022 12:31:50 - INFO - codeparrot_training - Step 18420: {'lr': 0.00036900199571961336, 'samples': 3536832, 'steps': 18420, 'loss/train': 1.4427541494369507} 01/29/2022 12:31:54 - INFO - codeparrot_training - Step 18421: {'lr': 0.0003689876056456995, 'samples': 3537024, 'steps': 18421, 'loss/train': 1.5277524590492249} 01/29/2022 12:31:58 - INFO - codeparrot_training - Step 18422: {'lr': 0.0003689732150620805, 'samples': 3537216, 'steps': 18422, 'loss/train': 1.516697645187378} 01/29/2022 12:32:03 - INFO - codeparrot_training - Step 18423: {'lr': 0.00036895882396881805, 'samples': 3537408, 'steps': 18423, 'loss/train': 2.1264413595199585} 01/29/2022 12:32:07 - INFO - codeparrot_training - Step 18424: {'lr': 0.0003689444323659737, 'samples': 3537600, 'steps': 18424, 'loss/train': 1.7150609493255615} 01/29/2022 12:32:12 - INFO - codeparrot_training - Step 18425: {'lr': 0.00036893004025360926, 'samples': 3537792, 'steps': 18425, 'loss/train': 0.12706198543310165} 01/29/2022 12:32:16 - INFO - codeparrot_training - Step 18426: {'lr': 0.0003689156476317862, 'samples': 3537984, 'steps': 18426, 'loss/train': 0.7343739867210388} 01/29/2022 12:32:21 - INFO - codeparrot_training - Step 18427: {'lr': 0.0003689012545005664, 'samples': 3538176, 'steps': 18427, 'loss/train': 1.6808493733406067} 01/29/2022 12:32:25 - INFO - codeparrot_training - Step 18428: {'lr': 0.0003688868608600113, 'samples': 3538368, 'steps': 18428, 'loss/train': 1.935016393661499} 01/29/2022 12:32:29 - INFO - codeparrot_training - Step 18429: {'lr': 0.0003688724667101826, 'samples': 3538560, 'steps': 18429, 'loss/train': 1.0168588757514954} 01/29/2022 12:32:36 - INFO - codeparrot_training - Step 18430: {'lr': 0.0003688580720511421, 'samples': 3538752, 'steps': 18430, 'loss/train': 1.5823500752449036} 01/29/2022 12:32:41 - INFO - codeparrot_training - Step 18431: {'lr': 0.0003688436768829512, 'samples': 3538944, 'steps': 18431, 'loss/train': 1.8480836749076843} 01/29/2022 12:32:45 - INFO - codeparrot_training - Step 18432: {'lr': 0.0003688292812056719, 'samples': 3539136, 'steps': 18432, 'loss/train': 0.3081633523106575} 01/29/2022 12:32:49 - INFO - codeparrot_training - Step 18433: {'lr': 0.00036881488501936554, 'samples': 3539328, 'steps': 18433, 'loss/train': 1.7460980415344238} 01/29/2022 12:32:53 - INFO - codeparrot_training - Step 18434: {'lr': 0.00036880048832409407, 'samples': 3539520, 'steps': 18434, 'loss/train': 0.2013304978609085} 01/29/2022 12:32:59 - INFO - codeparrot_training - Step 18435: {'lr': 0.000368786091119919, 'samples': 3539712, 'steps': 18435, 'loss/train': 2.0349369049072266} 01/29/2022 12:33:03 - INFO - codeparrot_training - Step 18436: {'lr': 0.00036877169340690204, 'samples': 3539904, 'steps': 18436, 'loss/train': 1.5865311026573181} 01/29/2022 12:33:07 - INFO - codeparrot_training - Step 18437: {'lr': 0.0003687572951851048, 'samples': 3540096, 'steps': 18437, 'loss/train': 0.999147891998291} 01/29/2022 12:33:11 - INFO - codeparrot_training - Step 18438: {'lr': 0.0003687428964545891, 'samples': 3540288, 'steps': 18438, 'loss/train': 1.780926525592804} 01/29/2022 12:33:16 - INFO - codeparrot_training - Step 18439: {'lr': 0.00036872849721541643, 'samples': 3540480, 'steps': 18439, 'loss/train': 1.5352200865745544} 01/29/2022 12:33:21 - INFO - codeparrot_training - Step 18440: {'lr': 0.0003687140974676486, 'samples': 3540672, 'steps': 18440, 'loss/train': 1.8672861456871033} 01/29/2022 12:33:25 - INFO - codeparrot_training - Step 18441: {'lr': 0.00036869969721134736, 'samples': 3540864, 'steps': 18441, 'loss/train': 0.9585631191730499} 01/29/2022 12:33:29 - INFO - codeparrot_training - Step 18442: {'lr': 0.0003686852964465742, 'samples': 3541056, 'steps': 18442, 'loss/train': 1.2732227146625519} 01/29/2022 12:33:34 - INFO - codeparrot_training - Step 18443: {'lr': 0.000368670895173391, 'samples': 3541248, 'steps': 18443, 'loss/train': 1.3238774836063385} 01/29/2022 12:33:38 - INFO - codeparrot_training - Step 18444: {'lr': 0.00036865649339185935, 'samples': 3541440, 'steps': 18444, 'loss/train': 1.8863078355789185} 01/29/2022 12:33:45 - INFO - codeparrot_training - Step 18445: {'lr': 0.000368642091102041, 'samples': 3541632, 'steps': 18445, 'loss/train': 1.8715057969093323} 01/29/2022 12:33:49 - INFO - codeparrot_training - Step 18446: {'lr': 0.0003686276883039975, 'samples': 3541824, 'steps': 18446, 'loss/train': 7.340464353561401} 01/29/2022 12:33:54 - INFO - codeparrot_training - Step 18447: {'lr': 0.0003686132849977908, 'samples': 3542016, 'steps': 18447, 'loss/train': 0.8358563482761383} 01/29/2022 12:33:58 - INFO - codeparrot_training - Step 18448: {'lr': 0.0003685988811834823, 'samples': 3542208, 'steps': 18448, 'loss/train': 1.9787943363189697} 01/29/2022 12:34:02 - INFO - codeparrot_training - Step 18449: {'lr': 0.00036858447686113395, 'samples': 3542400, 'steps': 18449, 'loss/train': 2.1037906408309937} 01/29/2022 12:34:07 - INFO - codeparrot_training - Step 18450: {'lr': 0.0003685700720308073, 'samples': 3542592, 'steps': 18450, 'loss/train': 1.5992730259895325} 01/29/2022 12:34:12 - INFO - codeparrot_training - Step 18451: {'lr': 0.0003685556666925641, 'samples': 3542784, 'steps': 18451, 'loss/train': 2.4897340536117554} 01/29/2022 12:34:16 - INFO - codeparrot_training - Step 18452: {'lr': 0.0003685412608464661, 'samples': 3542976, 'steps': 18452, 'loss/train': 1.780957281589508} 01/29/2022 12:34:20 - INFO - codeparrot_training - Step 18453: {'lr': 0.00036852685449257505, 'samples': 3543168, 'steps': 18453, 'loss/train': 1.2816991209983826} 01/29/2022 12:34:24 - INFO - codeparrot_training - Step 18454: {'lr': 0.00036851244763095247, 'samples': 3543360, 'steps': 18454, 'loss/train': 2.530522048473358} 01/29/2022 12:34:32 - INFO - codeparrot_training - Step 18455: {'lr': 0.0003684980402616603, 'samples': 3543552, 'steps': 18455, 'loss/train': 1.325286090373993} 01/29/2022 12:34:36 - INFO - codeparrot_training - Step 18456: {'lr': 0.0003684836323847601, 'samples': 3543744, 'steps': 18456, 'loss/train': 1.2771605551242828} 01/29/2022 12:34:40 - INFO - codeparrot_training - Step 18457: {'lr': 0.0003684692240003137, 'samples': 3543936, 'steps': 18457, 'loss/train': 0.34299013763666153} 01/29/2022 12:34:44 - INFO - codeparrot_training - Step 18458: {'lr': 0.00036845481510838264, 'samples': 3544128, 'steps': 18458, 'loss/train': 2.250565230846405} 01/29/2022 12:34:49 - INFO - codeparrot_training - Step 18459: {'lr': 0.00036844040570902886, 'samples': 3544320, 'steps': 18459, 'loss/train': 1.4109798073768616} 01/29/2022 12:34:54 - INFO - codeparrot_training - Step 18460: {'lr': 0.00036842599580231395, 'samples': 3544512, 'steps': 18460, 'loss/train': 0.8816058039665222} 01/29/2022 12:34:58 - INFO - codeparrot_training - Step 18461: {'lr': 0.0003684115853882997, 'samples': 3544704, 'steps': 18461, 'loss/train': 2.1395788192749023} 01/29/2022 12:35:03 - INFO - codeparrot_training - Step 18462: {'lr': 0.00036839717446704787, 'samples': 3544896, 'steps': 18462, 'loss/train': 1.3391646444797516} 01/29/2022 12:35:07 - INFO - codeparrot_training - Step 18463: {'lr': 0.00036838276303862, 'samples': 3545088, 'steps': 18463, 'loss/train': 1.220382571220398} 01/29/2022 12:35:11 - INFO - codeparrot_training - Step 18464: {'lr': 0.00036836835110307803, 'samples': 3545280, 'steps': 18464, 'loss/train': 1.5703060626983643} 01/29/2022 12:35:16 - INFO - codeparrot_training - Step 18465: {'lr': 0.0003683539386604837, 'samples': 3545472, 'steps': 18465, 'loss/train': 1.9028708338737488} 01/29/2022 12:35:20 - INFO - codeparrot_training - Step 18466: {'lr': 0.00036833952571089856, 'samples': 3545664, 'steps': 18466, 'loss/train': 1.3546797037124634} 01/29/2022 12:35:25 - INFO - codeparrot_training - Step 18467: {'lr': 0.0003683251122543846, 'samples': 3545856, 'steps': 18467, 'loss/train': 1.7408697009086609} 01/29/2022 12:35:29 - INFO - codeparrot_training - Step 18468: {'lr': 0.0003683106982910033, 'samples': 3546048, 'steps': 18468, 'loss/train': 1.6705225110054016} 01/29/2022 12:35:33 - INFO - codeparrot_training - Step 18469: {'lr': 0.0003682962838208166, 'samples': 3546240, 'steps': 18469, 'loss/train': 1.6668999195098877} 01/29/2022 12:35:40 - INFO - codeparrot_training - Step 18470: {'lr': 0.0003682818688438862, 'samples': 3546432, 'steps': 18470, 'loss/train': 2.1890611052513123} 01/29/2022 12:35:45 - INFO - codeparrot_training - Step 18471: {'lr': 0.00036826745336027383, 'samples': 3546624, 'steps': 18471, 'loss/train': 1.7394611835479736} 01/29/2022 12:35:49 - INFO - codeparrot_training - Step 18472: {'lr': 0.0003682530373700412, 'samples': 3546816, 'steps': 18472, 'loss/train': 1.966459572315216} 01/29/2022 12:35:53 - INFO - codeparrot_training - Step 18473: {'lr': 0.00036823862087325017, 'samples': 3547008, 'steps': 18473, 'loss/train': 1.7252227663993835} 01/29/2022 12:35:57 - INFO - codeparrot_training - Step 18474: {'lr': 0.00036822420386996237, 'samples': 3547200, 'steps': 18474, 'loss/train': 0.3287992626428604} 01/29/2022 12:36:03 - INFO - codeparrot_training - Step 18475: {'lr': 0.0003682097863602397, 'samples': 3547392, 'steps': 18475, 'loss/train': 1.3335233330726624} 01/29/2022 12:36:07 - INFO - codeparrot_training - Step 18476: {'lr': 0.00036819536834414374, 'samples': 3547584, 'steps': 18476, 'loss/train': 1.3640932738780975} 01/29/2022 12:36:11 - INFO - codeparrot_training - Step 18477: {'lr': 0.0003681809498217364, 'samples': 3547776, 'steps': 18477, 'loss/train': 1.627009391784668} 01/29/2022 12:36:15 - INFO - codeparrot_training - Step 18478: {'lr': 0.0003681665307930794, 'samples': 3547968, 'steps': 18478, 'loss/train': 2.390514314174652} 01/29/2022 12:36:20 - INFO - codeparrot_training - Step 18479: {'lr': 0.0003681521112582345, 'samples': 3548160, 'steps': 18479, 'loss/train': 1.6940467357635498} 01/29/2022 12:36:25 - INFO - codeparrot_training - Step 18480: {'lr': 0.00036813769121726354, 'samples': 3548352, 'steps': 18480, 'loss/train': 1.0661333799362183} 01/29/2022 12:36:29 - INFO - codeparrot_training - Step 18481: {'lr': 0.00036812327067022813, 'samples': 3548544, 'steps': 18481, 'loss/train': 1.614500641822815} 01/29/2022 12:36:34 - INFO - codeparrot_training - Step 18482: {'lr': 0.00036810884961719015, 'samples': 3548736, 'steps': 18482, 'loss/train': 1.7876036167144775} 01/29/2022 12:36:38 - INFO - codeparrot_training - Step 18483: {'lr': 0.0003680944280582114, 'samples': 3548928, 'steps': 18483, 'loss/train': 1.4252581894397736} 01/29/2022 12:36:42 - INFO - codeparrot_training - Step 18484: {'lr': 0.0003680800059933536, 'samples': 3549120, 'steps': 18484, 'loss/train': 0.7346000969409943} 01/29/2022 12:36:48 - INFO - codeparrot_training - Step 18485: {'lr': 0.00036806558342267854, 'samples': 3549312, 'steps': 18485, 'loss/train': 1.1456231474876404} 01/29/2022 12:36:52 - INFO - codeparrot_training - Step 18486: {'lr': 0.0003680511603462481, 'samples': 3549504, 'steps': 18486, 'loss/train': 1.1341567039489746} 01/29/2022 12:36:56 - INFO - codeparrot_training - Step 18487: {'lr': 0.00036803673676412386, 'samples': 3549696, 'steps': 18487, 'loss/train': 1.2321845591068268} 01/29/2022 12:37:00 - INFO - codeparrot_training - Step 18488: {'lr': 0.00036802231267636773, 'samples': 3549888, 'steps': 18488, 'loss/train': 1.5295066237449646} 01/29/2022 12:37:05 - INFO - codeparrot_training - Step 18489: {'lr': 0.0003680078880830415, 'samples': 3550080, 'steps': 18489, 'loss/train': 1.6309831738471985} 01/29/2022 12:37:12 - INFO - codeparrot_training - Step 18490: {'lr': 0.000367993462984207, 'samples': 3550272, 'steps': 18490, 'loss/train': 0.7819494903087616} 01/29/2022 12:37:16 - INFO - codeparrot_training - Step 18491: {'lr': 0.0003679790373799259, 'samples': 3550464, 'steps': 18491, 'loss/train': 1.6502800583839417} 01/29/2022 12:37:20 - INFO - codeparrot_training - Step 18492: {'lr': 0.0003679646112702601, 'samples': 3550656, 'steps': 18492, 'loss/train': 1.2341411411762238} 01/29/2022 12:37:24 - INFO - codeparrot_training - Step 18493: {'lr': 0.0003679501846552714, 'samples': 3550848, 'steps': 18493, 'loss/train': 2.6842410564422607} 01/29/2022 12:37:29 - INFO - codeparrot_training - Step 18494: {'lr': 0.00036793575753502153, 'samples': 3551040, 'steps': 18494, 'loss/train': 1.5082679986953735} 01/29/2022 12:37:34 - INFO - codeparrot_training - Step 18495: {'lr': 0.0003679213299095723, 'samples': 3551232, 'steps': 18495, 'loss/train': 0.6030578166246414} 01/29/2022 12:37:38 - INFO - codeparrot_training - Step 18496: {'lr': 0.00036790690177898556, 'samples': 3551424, 'steps': 18496, 'loss/train': 1.9623075127601624} 01/29/2022 12:37:43 - INFO - codeparrot_training - Step 18497: {'lr': 0.00036789247314332306, 'samples': 3551616, 'steps': 18497, 'loss/train': 1.7321373224258423} 01/29/2022 12:37:47 - INFO - codeparrot_training - Step 18498: {'lr': 0.00036787804400264666, 'samples': 3551808, 'steps': 18498, 'loss/train': 1.2711803913116455} 01/29/2022 12:37:51 - INFO - codeparrot_training - Step 18499: {'lr': 0.00036786361435701823, 'samples': 3552000, 'steps': 18499, 'loss/train': 2.4585235118865967} 01/29/2022 12:37:58 - INFO - codeparrot_training - Step 18500: {'lr': 0.0003678491842064995, 'samples': 3552192, 'steps': 18500, 'loss/train': 1.0203268826007843} 01/29/2022 12:38:03 - INFO - codeparrot_training - Step 18501: {'lr': 0.00036783475355115213, 'samples': 3552384, 'steps': 18501, 'loss/train': 1.4355229139328003} 01/29/2022 12:38:07 - INFO - codeparrot_training - Step 18502: {'lr': 0.0003678203223910382, 'samples': 3552576, 'steps': 18502, 'loss/train': 1.534355878829956} 01/29/2022 12:38:11 - INFO - codeparrot_training - Step 18503: {'lr': 0.0003678058907262194, 'samples': 3552768, 'steps': 18503, 'loss/train': 1.7417045831680298} 01/29/2022 12:38:15 - INFO - codeparrot_training - Step 18504: {'lr': 0.00036779145855675763, 'samples': 3552960, 'steps': 18504, 'loss/train': 0.823778361082077} 01/29/2022 12:38:20 - INFO - codeparrot_training - Step 18505: {'lr': 0.00036777702588271455, 'samples': 3553152, 'steps': 18505, 'loss/train': 1.6647031903266907} 01/29/2022 12:38:25 - INFO - codeparrot_training - Step 18506: {'lr': 0.0003677625927041522, 'samples': 3553344, 'steps': 18506, 'loss/train': 1.1762559413909912} 01/29/2022 12:38:29 - INFO - codeparrot_training - Step 18507: {'lr': 0.0003677481590211322, 'samples': 3553536, 'steps': 18507, 'loss/train': 1.6643226742744446} 01/29/2022 12:38:33 - INFO - codeparrot_training - Step 18508: {'lr': 0.0003677337248337165, 'samples': 3553728, 'steps': 18508, 'loss/train': 1.5685969591140747} 01/29/2022 12:38:37 - INFO - codeparrot_training - Step 18509: {'lr': 0.0003677192901419669, 'samples': 3553920, 'steps': 18509, 'loss/train': 2.3907805681228638} 01/29/2022 12:38:43 - INFO - codeparrot_training - Step 18510: {'lr': 0.0003677048549459453, 'samples': 3554112, 'steps': 18510, 'loss/train': 0.9755196869373322} 01/29/2022 12:38:47 - INFO - codeparrot_training - Step 18511: {'lr': 0.00036769041924571345, 'samples': 3554304, 'steps': 18511, 'loss/train': 1.7264683842658997} 01/29/2022 12:38:52 - INFO - codeparrot_training - Step 18512: {'lr': 0.0003676759830413332, 'samples': 3554496, 'steps': 18512, 'loss/train': 1.6809906363487244} 01/29/2022 12:38:56 - INFO - codeparrot_training - Step 18513: {'lr': 0.00036766154633286635, 'samples': 3554688, 'steps': 18513, 'loss/train': 1.3616614937782288} 01/29/2022 12:39:03 - INFO - codeparrot_training - Step 18514: {'lr': 0.00036764710912037487, 'samples': 3554880, 'steps': 18514, 'loss/train': 0.19830049574375153} 01/29/2022 12:39:07 - INFO - codeparrot_training - Step 18515: {'lr': 0.00036763267140392053, 'samples': 3555072, 'steps': 18515, 'loss/train': 1.982647955417633} 01/29/2022 12:39:12 - INFO - codeparrot_training - Step 18516: {'lr': 0.0003676182331835651, 'samples': 3555264, 'steps': 18516, 'loss/train': 2.1556947827339172} 01/29/2022 12:39:16 - INFO - codeparrot_training - Step 18517: {'lr': 0.00036760379445937067, 'samples': 3555456, 'steps': 18517, 'loss/train': 1.093934565782547} 01/29/2022 12:39:20 - INFO - codeparrot_training - Step 18518: {'lr': 0.0003675893552313988, 'samples': 3555648, 'steps': 18518, 'loss/train': 1.8020027875900269} 01/29/2022 12:39:24 - INFO - codeparrot_training - Step 18519: {'lr': 0.0003675749154997115, 'samples': 3555840, 'steps': 18519, 'loss/train': 1.5400083661079407} 01/29/2022 12:39:30 - INFO - codeparrot_training - Step 18520: {'lr': 0.00036756047526437057, 'samples': 3556032, 'steps': 18520, 'loss/train': 1.2136101722717285} 01/29/2022 12:39:34 - INFO - codeparrot_training - Step 18521: {'lr': 0.00036754603452543796, 'samples': 3556224, 'steps': 18521, 'loss/train': 2.2601028084754944} 01/29/2022 12:39:38 - INFO - codeparrot_training - Step 18522: {'lr': 0.00036753159328297536, 'samples': 3556416, 'steps': 18522, 'loss/train': 1.2368578612804413} 01/29/2022 12:39:42 - INFO - codeparrot_training - Step 18523: {'lr': 0.00036751715153704483, 'samples': 3556608, 'steps': 18523, 'loss/train': 1.2911649942398071} 01/29/2022 12:39:47 - INFO - codeparrot_training - Step 18524: {'lr': 0.0003675027092877081, 'samples': 3556800, 'steps': 18524, 'loss/train': 1.4072974920272827} 01/29/2022 12:39:52 - INFO - codeparrot_training - Step 18525: {'lr': 0.0003674882665350271, 'samples': 3556992, 'steps': 18525, 'loss/train': 1.6381609439849854} 01/29/2022 12:39:56 - INFO - codeparrot_training - Step 18526: {'lr': 0.0003674738232790636, 'samples': 3557184, 'steps': 18526, 'loss/train': 1.9753469824790955} 01/29/2022 12:40:00 - INFO - codeparrot_training - Step 18527: {'lr': 0.0003674593795198796, 'samples': 3557376, 'steps': 18527, 'loss/train': 1.987178385257721} 01/29/2022 12:40:05 - INFO - codeparrot_training - Step 18528: {'lr': 0.00036744493525753697, 'samples': 3557568, 'steps': 18528, 'loss/train': 1.9508507251739502} 01/29/2022 12:40:12 - INFO - codeparrot_training - Step 18529: {'lr': 0.00036743049049209743, 'samples': 3557760, 'steps': 18529, 'loss/train': 1.350919246673584} 01/29/2022 12:40:16 - INFO - codeparrot_training - Step 18530: {'lr': 0.00036741604522362304, 'samples': 3557952, 'steps': 18530, 'loss/train': 1.6911841034889221} 01/29/2022 12:40:20 - INFO - codeparrot_training - Step 18531: {'lr': 0.00036740159945217556, 'samples': 3558144, 'steps': 18531, 'loss/train': 2.5203636288642883} 01/29/2022 12:40:25 - INFO - codeparrot_training - Step 18532: {'lr': 0.0003673871531778169, 'samples': 3558336, 'steps': 18532, 'loss/train': 0.7373560667037964} 01/29/2022 12:40:29 - INFO - codeparrot_training - Step 18533: {'lr': 0.00036737270640060894, 'samples': 3558528, 'steps': 18533, 'loss/train': 1.072446584701538} 01/29/2022 12:40:34 - INFO - codeparrot_training - Step 18534: {'lr': 0.0003673582591206136, 'samples': 3558720, 'steps': 18534, 'loss/train': 1.8134994506835938} 01/29/2022 12:40:38 - INFO - codeparrot_training - Step 18535: {'lr': 0.00036734381133789277, 'samples': 3558912, 'steps': 18535, 'loss/train': 1.4183776080608368} 01/29/2022 12:40:43 - INFO - codeparrot_training - Step 18536: {'lr': 0.00036732936305250826, 'samples': 3559104, 'steps': 18536, 'loss/train': 1.8974096775054932} 01/29/2022 12:40:47 - INFO - codeparrot_training - Step 18537: {'lr': 0.00036731491426452204, 'samples': 3559296, 'steps': 18537, 'loss/train': 1.5647735595703125} 01/29/2022 12:40:51 - INFO - codeparrot_training - Step 18538: {'lr': 0.00036730046497399587, 'samples': 3559488, 'steps': 18538, 'loss/train': 2.1492862701416016} 01/29/2022 12:40:56 - INFO - codeparrot_training - Step 18539: {'lr': 0.0003672860151809919, 'samples': 3559680, 'steps': 18539, 'loss/train': 2.0175440311431885} 01/29/2022 12:41:01 - INFO - codeparrot_training - Step 18540: {'lr': 0.0003672715648855718, 'samples': 3559872, 'steps': 18540, 'loss/train': 1.587472915649414} 01/29/2022 12:41:05 - INFO - codeparrot_training - Step 18541: {'lr': 0.00036725711408779765, 'samples': 3560064, 'steps': 18541, 'loss/train': 2.020105004310608} 01/29/2022 12:41:09 - INFO - codeparrot_training - Step 18542: {'lr': 0.0003672426627877312, 'samples': 3560256, 'steps': 18542, 'loss/train': 1.0497146844863892} 01/29/2022 12:41:13 - INFO - codeparrot_training - Step 18543: {'lr': 0.0003672282109854344, 'samples': 3560448, 'steps': 18543, 'loss/train': 1.9084571599960327} 01/29/2022 12:41:21 - INFO - codeparrot_training - Step 18544: {'lr': 0.00036721375868096925, 'samples': 3560640, 'steps': 18544, 'loss/train': 1.6371710300445557} 01/29/2022 12:41:25 - INFO - codeparrot_training - Step 18545: {'lr': 0.00036719930587439744, 'samples': 3560832, 'steps': 18545, 'loss/train': 1.7894182205200195} 01/29/2022 12:41:29 - INFO - codeparrot_training - Step 18546: {'lr': 0.00036718485256578116, 'samples': 3561024, 'steps': 18546, 'loss/train': 1.6968610882759094} 01/29/2022 12:41:34 - INFO - codeparrot_training - Step 18547: {'lr': 0.00036717039875518203, 'samples': 3561216, 'steps': 18547, 'loss/train': 1.2137132585048676} 01/29/2022 12:41:38 - INFO - codeparrot_training - Step 18548: {'lr': 0.00036715594444266224, 'samples': 3561408, 'steps': 18548, 'loss/train': 2.0984973907470703} 01/29/2022 12:41:43 - INFO - codeparrot_training - Step 18549: {'lr': 0.00036714148962828353, 'samples': 3561600, 'steps': 18549, 'loss/train': 1.338249385356903} 01/29/2022 12:41:47 - INFO - codeparrot_training - Step 18550: {'lr': 0.0003671270343121079, 'samples': 3561792, 'steps': 18550, 'loss/train': 1.0917499959468842} 01/29/2022 12:41:52 - INFO - codeparrot_training - Step 18551: {'lr': 0.0003671125784941972, 'samples': 3561984, 'steps': 18551, 'loss/train': 1.027806669473648} 01/29/2022 12:41:56 - INFO - codeparrot_training - Step 18552: {'lr': 0.00036709812217461347, 'samples': 3562176, 'steps': 18552, 'loss/train': 0.07465476170182228} 01/29/2022 12:42:00 - INFO - codeparrot_training - Step 18553: {'lr': 0.0003670836653534185, 'samples': 3562368, 'steps': 18553, 'loss/train': 1.1601490080356598} 01/29/2022 12:42:06 - INFO - codeparrot_training - Step 18554: {'lr': 0.0003670692080306743, 'samples': 3562560, 'steps': 18554, 'loss/train': 1.125421553850174} 01/29/2022 12:42:10 - INFO - codeparrot_training - Step 18555: {'lr': 0.0003670547502064429, 'samples': 3562752, 'steps': 18555, 'loss/train': 1.3335337042808533} 01/29/2022 12:42:14 - INFO - codeparrot_training - Step 18556: {'lr': 0.000367040291880786, 'samples': 3562944, 'steps': 18556, 'loss/train': 0.9236810803413391} 01/29/2022 12:42:19 - INFO - codeparrot_training - Step 18557: {'lr': 0.0003670258330537656, 'samples': 3563136, 'steps': 18557, 'loss/train': 0.5916202515363693} 01/29/2022 12:42:23 - INFO - codeparrot_training - Step 18558: {'lr': 0.0003670113737254438, 'samples': 3563328, 'steps': 18558, 'loss/train': 2.164103329181671} 01/29/2022 12:42:30 - INFO - codeparrot_training - Step 18559: {'lr': 0.0003669969138958824, 'samples': 3563520, 'steps': 18559, 'loss/train': 1.8295773267745972} 01/29/2022 12:42:34 - INFO - codeparrot_training - Step 18560: {'lr': 0.00036698245356514336, 'samples': 3563712, 'steps': 18560, 'loss/train': 0.6395140886306763} 01/29/2022 12:42:39 - INFO - codeparrot_training - Step 18561: {'lr': 0.00036696799273328864, 'samples': 3563904, 'steps': 18561, 'loss/train': 1.5202040076255798} 01/29/2022 12:42:43 - INFO - codeparrot_training - Step 18562: {'lr': 0.0003669535314003802, 'samples': 3564096, 'steps': 18562, 'loss/train': 1.0152559876441956} 01/29/2022 12:42:47 - INFO - codeparrot_training - Step 18563: {'lr': 0.00036693906956647996, 'samples': 3564288, 'steps': 18563, 'loss/train': 1.2651605308055878} 01/29/2022 12:42:52 - INFO - codeparrot_training - Step 18564: {'lr': 0.0003669246072316498, 'samples': 3564480, 'steps': 18564, 'loss/train': 1.532278060913086} 01/29/2022 12:42:57 - INFO - codeparrot_training - Step 18565: {'lr': 0.00036691014439595187, 'samples': 3564672, 'steps': 18565, 'loss/train': 1.7134239077568054} 01/29/2022 12:43:01 - INFO - codeparrot_training - Step 18566: {'lr': 0.00036689568105944794, 'samples': 3564864, 'steps': 18566, 'loss/train': 1.8174245953559875} 01/29/2022 12:43:05 - INFO - codeparrot_training - Step 18567: {'lr': 0.0003668812172222001, 'samples': 3565056, 'steps': 18567, 'loss/train': 1.4164190590381622} 01/29/2022 12:43:09 - INFO - codeparrot_training - Step 18568: {'lr': 0.0003668667528842702, 'samples': 3565248, 'steps': 18568, 'loss/train': 1.867656648159027} 01/29/2022 12:43:15 - INFO - codeparrot_training - Step 18569: {'lr': 0.0003668522880457202, 'samples': 3565440, 'steps': 18569, 'loss/train': 1.6867561340332031} 01/29/2022 12:43:19 - INFO - codeparrot_training - Step 18570: {'lr': 0.0003668378227066121, 'samples': 3565632, 'steps': 18570, 'loss/train': 1.343284159898758} 01/29/2022 12:43:24 - INFO - codeparrot_training - Step 18571: {'lr': 0.00036682335686700796, 'samples': 3565824, 'steps': 18571, 'loss/train': 0.9731751680374146} 01/29/2022 12:43:28 - INFO - codeparrot_training - Step 18572: {'lr': 0.00036680889052696954, 'samples': 3566016, 'steps': 18572, 'loss/train': 0.8356671631336212} 01/29/2022 12:43:32 - INFO - codeparrot_training - Step 18573: {'lr': 0.00036679442368655897, 'samples': 3566208, 'steps': 18573, 'loss/train': 1.119455873966217} 01/29/2022 12:43:37 - INFO - codeparrot_training - Step 18574: {'lr': 0.00036677995634583815, 'samples': 3566400, 'steps': 18574, 'loss/train': 1.6120262145996094} 01/29/2022 12:43:42 - INFO - codeparrot_training - Step 18575: {'lr': 0.0003667654885048691, 'samples': 3566592, 'steps': 18575, 'loss/train': 0.7396845817565918} 01/29/2022 12:43:46 - INFO - codeparrot_training - Step 18576: {'lr': 0.00036675102016371386, 'samples': 3566784, 'steps': 18576, 'loss/train': 1.160598635673523} 01/29/2022 12:43:50 - INFO - codeparrot_training - Step 18577: {'lr': 0.0003667365513224342, 'samples': 3566976, 'steps': 18577, 'loss/train': 1.7574507594108582} 01/29/2022 12:43:54 - INFO - codeparrot_training - Step 18578: {'lr': 0.0003667220819810923, 'samples': 3567168, 'steps': 18578, 'loss/train': 1.64760160446167} 01/29/2022 12:44:02 - INFO - codeparrot_training - Step 18579: {'lr': 0.00036670761213975, 'samples': 3567360, 'steps': 18579, 'loss/train': 1.9698268175125122} 01/29/2022 12:44:06 - INFO - codeparrot_training - Step 18580: {'lr': 0.0003666931417984694, 'samples': 3567552, 'steps': 18580, 'loss/train': 2.0080182552337646} 01/29/2022 12:44:10 - INFO - codeparrot_training - Step 18581: {'lr': 0.00036667867095731244, 'samples': 3567744, 'steps': 18581, 'loss/train': 0.11605871468782425} 01/29/2022 12:44:15 - INFO - codeparrot_training - Step 18582: {'lr': 0.0003666641996163411, 'samples': 3567936, 'steps': 18582, 'loss/train': 1.1315969824790955} 01/29/2022 12:44:19 - INFO - codeparrot_training - Step 18583: {'lr': 0.0003666497277756173, 'samples': 3568128, 'steps': 18583, 'loss/train': 1.9399028420448303} 01/29/2022 12:44:24 - INFO - codeparrot_training - Step 18584: {'lr': 0.0003666352554352032, 'samples': 3568320, 'steps': 18584, 'loss/train': 1.5269211530685425} 01/29/2022 12:44:28 - INFO - codeparrot_training - Step 18585: {'lr': 0.0003666207825951606, 'samples': 3568512, 'steps': 18585, 'loss/train': 1.2025273740291595} 01/29/2022 12:44:32 - INFO - codeparrot_training - Step 18586: {'lr': 0.00036660630925555173, 'samples': 3568704, 'steps': 18586, 'loss/train': 1.186234563589096} 01/29/2022 12:44:37 - INFO - codeparrot_training - Step 18587: {'lr': 0.0003665918354164384, 'samples': 3568896, 'steps': 18587, 'loss/train': 1.8086113929748535} 01/29/2022 12:44:42 - INFO - codeparrot_training - Step 18588: {'lr': 0.00036657736107788264, 'samples': 3569088, 'steps': 18588, 'loss/train': 1.3696953356266022} 01/29/2022 12:44:46 - INFO - codeparrot_training - Step 18589: {'lr': 0.00036656288623994647, 'samples': 3569280, 'steps': 18589, 'loss/train': 1.0348294973373413} 01/29/2022 12:44:51 - INFO - codeparrot_training - Step 18590: {'lr': 0.000366548410902692, 'samples': 3569472, 'steps': 18590, 'loss/train': 1.1697555184364319} 01/29/2022 12:44:55 - INFO - codeparrot_training - Step 18591: {'lr': 0.00036653393506618106, 'samples': 3569664, 'steps': 18591, 'loss/train': 0.05027663893997669} 01/29/2022 12:44:59 - INFO - codeparrot_training - Step 18592: {'lr': 0.00036651945873047574, 'samples': 3569856, 'steps': 18592, 'loss/train': 0.7075382620096207} 01/29/2022 12:45:06 - INFO - codeparrot_training - Step 18593: {'lr': 0.0003665049818956381, 'samples': 3570048, 'steps': 18593, 'loss/train': 1.844730019569397} 01/29/2022 12:45:11 - INFO - codeparrot_training - Step 18594: {'lr': 0.0003664905045617301, 'samples': 3570240, 'steps': 18594, 'loss/train': 0.4205291122198105} 01/29/2022 12:45:15 - INFO - codeparrot_training - Step 18595: {'lr': 0.0003664760267288138, 'samples': 3570432, 'steps': 18595, 'loss/train': 1.8586726784706116} 01/29/2022 12:45:19 - INFO - codeparrot_training - Step 18596: {'lr': 0.0003664615483969511, 'samples': 3570624, 'steps': 18596, 'loss/train': 0.9880940616130829} 01/29/2022 12:45:23 - INFO - codeparrot_training - Step 18597: {'lr': 0.0003664470695662042, 'samples': 3570816, 'steps': 18597, 'loss/train': 2.46082466840744} 01/29/2022 12:45:28 - INFO - codeparrot_training - Step 18598: {'lr': 0.000366432590236635, 'samples': 3571008, 'steps': 18598, 'loss/train': 1.7149452567100525} 01/29/2022 12:45:33 - INFO - codeparrot_training - Step 18599: {'lr': 0.0003664181104083055, 'samples': 3571200, 'steps': 18599, 'loss/train': 1.2062064707279205} 01/29/2022 12:45:37 - INFO - codeparrot_training - Step 18600: {'lr': 0.00036640363008127785, 'samples': 3571392, 'steps': 18600, 'loss/train': 1.7725583910942078} 01/29/2022 12:45:41 - INFO - codeparrot_training - Step 18601: {'lr': 0.000366389149255614, 'samples': 3571584, 'steps': 18601, 'loss/train': 1.5571660995483398} 01/29/2022 12:45:46 - INFO - codeparrot_training - Step 18602: {'lr': 0.00036637466793137605, 'samples': 3571776, 'steps': 18602, 'loss/train': 2.0654300451278687} 01/29/2022 12:45:50 - INFO - codeparrot_training - Step 18603: {'lr': 0.0003663601861086259, 'samples': 3571968, 'steps': 18603, 'loss/train': 1.1103185713291168} 01/29/2022 12:45:57 - INFO - codeparrot_training - Step 18604: {'lr': 0.00036634570378742565, 'samples': 3572160, 'steps': 18604, 'loss/train': 0.5116106271743774} 01/29/2022 12:46:02 - INFO - codeparrot_training - Step 18605: {'lr': 0.00036633122096783736, 'samples': 3572352, 'steps': 18605, 'loss/train': 1.790239691734314} 01/29/2022 12:46:06 - INFO - codeparrot_training - Step 18606: {'lr': 0.00036631673764992307, 'samples': 3572544, 'steps': 18606, 'loss/train': 1.6982824802398682} 01/29/2022 12:46:10 - INFO - codeparrot_training - Step 18607: {'lr': 0.00036630225383374476, 'samples': 3572736, 'steps': 18607, 'loss/train': 1.4028820395469666} 01/29/2022 12:46:15 - INFO - codeparrot_training - Step 18608: {'lr': 0.0003662877695193646, 'samples': 3572928, 'steps': 18608, 'loss/train': 0.6111644357442856} 01/29/2022 12:46:20 - INFO - codeparrot_training - Step 18609: {'lr': 0.0003662732847068445, 'samples': 3573120, 'steps': 18609, 'loss/train': 2.1049389839172363} 01/29/2022 12:46:24 - INFO - codeparrot_training - Step 18610: {'lr': 0.00036625879939624663, 'samples': 3573312, 'steps': 18610, 'loss/train': 1.416559249162674} 01/29/2022 12:46:28 - INFO - codeparrot_training - Step 18611: {'lr': 0.000366244313587633, 'samples': 3573504, 'steps': 18611, 'loss/train': 1.789162516593933} 01/29/2022 12:46:32 - INFO - codeparrot_training - Step 18612: {'lr': 0.0003662298272810655, 'samples': 3573696, 'steps': 18612, 'loss/train': 1.5365692377090454} 01/29/2022 12:46:37 - INFO - codeparrot_training - Step 18613: {'lr': 0.00036621534047660647, 'samples': 3573888, 'steps': 18613, 'loss/train': 2.118475377559662} 01/29/2022 12:46:42 - INFO - codeparrot_training - Step 18614: {'lr': 0.00036620085317431777, 'samples': 3574080, 'steps': 18614, 'loss/train': 1.9140663743019104} 01/29/2022 12:46:46 - INFO - codeparrot_training - Step 18615: {'lr': 0.0003661863653742615, 'samples': 3574272, 'steps': 18615, 'loss/train': 1.643444001674652} 01/29/2022 12:46:51 - INFO - codeparrot_training - Step 18616: {'lr': 0.0003661718770764998, 'samples': 3574464, 'steps': 18616, 'loss/train': 1.9421117305755615} 01/29/2022 12:46:55 - INFO - codeparrot_training - Step 18617: {'lr': 0.00036615738828109465, 'samples': 3574656, 'steps': 18617, 'loss/train': 1.5884113311767578} 01/29/2022 12:47:02 - INFO - codeparrot_training - Step 18618: {'lr': 0.00036614289898810804, 'samples': 3574848, 'steps': 18618, 'loss/train': 1.8724141716957092} 01/29/2022 12:47:06 - INFO - codeparrot_training - Step 18619: {'lr': 0.00036612840919760225, 'samples': 3575040, 'steps': 18619, 'loss/train': 2.2428104281425476} 01/29/2022 12:47:11 - INFO - codeparrot_training - Step 18620: {'lr': 0.00036611391890963913, 'samples': 3575232, 'steps': 18620, 'loss/train': 2.0830569863319397} 01/29/2022 12:47:15 - INFO - codeparrot_training - Step 18621: {'lr': 0.00036609942812428087, 'samples': 3575424, 'steps': 18621, 'loss/train': 1.9396812915802002} 01/29/2022 12:47:19 - INFO - codeparrot_training - Step 18622: {'lr': 0.00036608493684158963, 'samples': 3575616, 'steps': 18622, 'loss/train': 1.7496723532676697} 01/29/2022 12:47:25 - INFO - codeparrot_training - Step 18623: {'lr': 0.0003660704450616272, 'samples': 3575808, 'steps': 18623, 'loss/train': 1.6651362776756287} 01/29/2022 12:47:29 - INFO - codeparrot_training - Step 18624: {'lr': 0.00036605595278445605, 'samples': 3576000, 'steps': 18624, 'loss/train': 2.6713133454322815} 01/29/2022 12:47:33 - INFO - codeparrot_training - Step 18625: {'lr': 0.0003660414600101379, 'samples': 3576192, 'steps': 18625, 'loss/train': 1.4377528131008148} 01/29/2022 12:47:37 - INFO - codeparrot_training - Step 18626: {'lr': 0.00036602696673873505, 'samples': 3576384, 'steps': 18626, 'loss/train': 1.1481982469558716} 01/29/2022 12:47:42 - INFO - codeparrot_training - Step 18627: {'lr': 0.00036601247297030943, 'samples': 3576576, 'steps': 18627, 'loss/train': 1.873545527458191} 01/29/2022 12:47:49 - INFO - codeparrot_training - Step 18628: {'lr': 0.00036599797870492327, 'samples': 3576768, 'steps': 18628, 'loss/train': 3.606541156768799} 01/29/2022 12:47:53 - INFO - codeparrot_training - Step 18629: {'lr': 0.0003659834839426387, 'samples': 3576960, 'steps': 18629, 'loss/train': 1.9199808239936829} 01/29/2022 12:47:57 - INFO - codeparrot_training - Step 18630: {'lr': 0.0003659689886835176, 'samples': 3577152, 'steps': 18630, 'loss/train': 1.352348953485489} 01/29/2022 12:48:02 - INFO - codeparrot_training - Step 18631: {'lr': 0.00036595449292762215, 'samples': 3577344, 'steps': 18631, 'loss/train': 1.96770840883255} 01/29/2022 12:48:06 - INFO - codeparrot_training - Step 18632: {'lr': 0.00036593999667501457, 'samples': 3577536, 'steps': 18632, 'loss/train': 2.0988965034484863} 01/29/2022 12:48:11 - INFO - codeparrot_training - Step 18633: {'lr': 0.0003659254999257568, 'samples': 3577728, 'steps': 18633, 'loss/train': 1.6772915124893188} 01/29/2022 12:48:16 - INFO - codeparrot_training - Step 18634: {'lr': 0.000365911002679911, 'samples': 3577920, 'steps': 18634, 'loss/train': 1.6984840035438538} 01/29/2022 12:48:20 - INFO - codeparrot_training - Step 18635: {'lr': 0.00036589650493753937, 'samples': 3578112, 'steps': 18635, 'loss/train': 1.5136560201644897} 01/29/2022 12:48:24 - INFO - codeparrot_training - Step 18636: {'lr': 0.00036588200669870376, 'samples': 3578304, 'steps': 18636, 'loss/train': 0.055389756336808205} 01/29/2022 12:48:28 - INFO - codeparrot_training - Step 18637: {'lr': 0.0003658675079634665, 'samples': 3578496, 'steps': 18637, 'loss/train': 1.776803970336914} 01/29/2022 12:48:34 - INFO - codeparrot_training - Step 18638: {'lr': 0.0003658530087318896, 'samples': 3578688, 'steps': 18638, 'loss/train': 0.6893539577722549} 01/29/2022 12:48:38 - INFO - codeparrot_training - Step 18639: {'lr': 0.00036583850900403527, 'samples': 3578880, 'steps': 18639, 'loss/train': 2.822615325450897} 01/29/2022 12:48:42 - INFO - codeparrot_training - Step 18640: {'lr': 0.00036582400877996547, 'samples': 3579072, 'steps': 18640, 'loss/train': 1.8111995458602905} 01/29/2022 12:48:46 - INFO - codeparrot_training - Step 18641: {'lr': 0.0003658095080597424, 'samples': 3579264, 'steps': 18641, 'loss/train': 2.1719196438789368} 01/29/2022 12:48:51 - INFO - codeparrot_training - Step 18642: {'lr': 0.0003657950068434282, 'samples': 3579456, 'steps': 18642, 'loss/train': 1.4523714780807495} 01/29/2022 12:48:56 - INFO - codeparrot_training - Step 18643: {'lr': 0.000365780505131085, 'samples': 3579648, 'steps': 18643, 'loss/train': 1.6798782348632812} 01/29/2022 12:49:00 - INFO - codeparrot_training - Step 18644: {'lr': 0.00036576600292277477, 'samples': 3579840, 'steps': 18644, 'loss/train': 1.6879363059997559} 01/29/2022 12:49:04 - INFO - codeparrot_training - Step 18645: {'lr': 0.00036575150021855987, 'samples': 3580032, 'steps': 18645, 'loss/train': 1.7203522324562073} 01/29/2022 12:49:09 - INFO - codeparrot_training - Step 18646: {'lr': 0.00036573699701850223, 'samples': 3580224, 'steps': 18646, 'loss/train': 2.2857969403266907} 01/29/2022 12:49:13 - INFO - codeparrot_training - Step 18647: {'lr': 0.000365722493322664, 'samples': 3580416, 'steps': 18647, 'loss/train': 2.045991361141205} 01/29/2022 12:49:20 - INFO - codeparrot_training - Step 18648: {'lr': 0.0003657079891311075, 'samples': 3580608, 'steps': 18648, 'loss/train': 1.9060696363449097} 01/29/2022 12:49:25 - INFO - codeparrot_training - Step 18649: {'lr': 0.00036569348444389456, 'samples': 3580800, 'steps': 18649, 'loss/train': 2.086331784725189} 01/29/2022 12:49:29 - INFO - codeparrot_training - Step 18650: {'lr': 0.00036567897926108756, 'samples': 3580992, 'steps': 18650, 'loss/train': 1.8128119111061096} 01/29/2022 12:49:33 - INFO - codeparrot_training - Step 18651: {'lr': 0.00036566447358274846, 'samples': 3581184, 'steps': 18651, 'loss/train': 1.5841459035873413} 01/29/2022 12:49:37 - INFO - codeparrot_training - Step 18652: {'lr': 0.0003656499674089396, 'samples': 3581376, 'steps': 18652, 'loss/train': 1.9997987151145935} 01/29/2022 12:49:43 - INFO - codeparrot_training - Step 18653: {'lr': 0.0003656354607397229, 'samples': 3581568, 'steps': 18653, 'loss/train': 0.5018536895513535} 01/29/2022 12:49:47 - INFO - codeparrot_training - Step 18654: {'lr': 0.00036562095357516066, 'samples': 3581760, 'steps': 18654, 'loss/train': 1.4621324837207794} 01/29/2022 12:49:52 - INFO - codeparrot_training - Step 18655: {'lr': 0.00036560644591531496, 'samples': 3581952, 'steps': 18655, 'loss/train': 1.8376811742782593} 01/29/2022 12:49:56 - INFO - codeparrot_training - Step 18656: {'lr': 0.00036559193776024794, 'samples': 3582144, 'steps': 18656, 'loss/train': 1.8787960410118103} 01/29/2022 12:50:00 - INFO - codeparrot_training - Step 18657: {'lr': 0.0003655774291100218, 'samples': 3582336, 'steps': 18657, 'loss/train': 2.0735008120536804} 01/29/2022 12:50:04 - INFO - codeparrot_training - Step 18658: {'lr': 0.0003655629199646986, 'samples': 3582528, 'steps': 18658, 'loss/train': 1.7820691466331482} 01/29/2022 12:50:10 - INFO - codeparrot_training - Step 18659: {'lr': 0.00036554841032434063, 'samples': 3582720, 'steps': 18659, 'loss/train': 1.600608229637146} 01/29/2022 12:50:14 - INFO - codeparrot_training - Step 18660: {'lr': 0.00036553390018900984, 'samples': 3582912, 'steps': 18660, 'loss/train': 1.057717502117157} 01/29/2022 12:50:18 - INFO - codeparrot_training - Step 18661: {'lr': 0.0003655193895587686, 'samples': 3583104, 'steps': 18661, 'loss/train': 2.4849854707717896} 01/29/2022 12:50:23 - INFO - codeparrot_training - Step 18662: {'lr': 0.000365504878433679, 'samples': 3583296, 'steps': 18662, 'loss/train': 1.5461263060569763} 01/29/2022 12:50:30 - INFO - codeparrot_training - Step 18663: {'lr': 0.00036549036681380307, 'samples': 3583488, 'steps': 18663, 'loss/train': 0.7609294652938843} 01/29/2022 12:50:34 - INFO - codeparrot_training - Step 18664: {'lr': 0.00036547585469920316, 'samples': 3583680, 'steps': 18664, 'loss/train': 0.3629555180668831} 01/29/2022 12:50:38 - INFO - codeparrot_training - Step 18665: {'lr': 0.00036546134208994137, 'samples': 3583872, 'steps': 18665, 'loss/train': 1.8876492977142334} 01/29/2022 12:50:43 - INFO - codeparrot_training - Step 18666: {'lr': 0.00036544682898607977, 'samples': 3584064, 'steps': 18666, 'loss/train': 1.8054310083389282} 01/29/2022 12:50:47 - INFO - codeparrot_training - Step 18667: {'lr': 0.00036543231538768066, 'samples': 3584256, 'steps': 18667, 'loss/train': 1.5529814958572388} 01/29/2022 12:50:52 - INFO - codeparrot_training - Step 18668: {'lr': 0.00036541780129480616, 'samples': 3584448, 'steps': 18668, 'loss/train': 1.9448368549346924} 01/29/2022 12:50:56 - INFO - codeparrot_training - Step 18669: {'lr': 0.0003654032867075185, 'samples': 3584640, 'steps': 18669, 'loss/train': 1.5240068435668945} 01/29/2022 12:51:01 - INFO - codeparrot_training - Step 18670: {'lr': 0.00036538877162587975, 'samples': 3584832, 'steps': 18670, 'loss/train': 1.555258870124817} 01/29/2022 12:51:05 - INFO - codeparrot_training - Step 18671: {'lr': 0.00036537425604995214, 'samples': 3585024, 'steps': 18671, 'loss/train': 1.8609952926635742} 01/29/2022 12:51:09 - INFO - codeparrot_training - Step 18672: {'lr': 0.00036535973997979787, 'samples': 3585216, 'steps': 18672, 'loss/train': 2.369134783744812} 01/29/2022 12:51:14 - INFO - codeparrot_training - Step 18673: {'lr': 0.0003653452234154791, 'samples': 3585408, 'steps': 18673, 'loss/train': 1.1366987228393555} 01/29/2022 12:51:19 - INFO - codeparrot_training - Step 18674: {'lr': 0.000365330706357058, 'samples': 3585600, 'steps': 18674, 'loss/train': 2.06714129447937} 01/29/2022 12:51:23 - INFO - codeparrot_training - Step 18675: {'lr': 0.0003653161888045968, 'samples': 3585792, 'steps': 18675, 'loss/train': 1.7753732800483704} 01/29/2022 12:51:27 - INFO - codeparrot_training - Step 18676: {'lr': 0.0003653016707581577, 'samples': 3585984, 'steps': 18676, 'loss/train': 2.149220108985901} 01/29/2022 12:51:31 - INFO - codeparrot_training - Step 18677: {'lr': 0.00036528715221780276, 'samples': 3586176, 'steps': 18677, 'loss/train': 1.0409964323043823} 01/29/2022 12:51:39 - INFO - codeparrot_training - Step 18678: {'lr': 0.0003652726331835944, 'samples': 3586368, 'steps': 18678, 'loss/train': 1.6158633828163147} 01/29/2022 12:51:43 - INFO - codeparrot_training - Step 18679: {'lr': 0.00036525811365559457, 'samples': 3586560, 'steps': 18679, 'loss/train': 1.4489710628986359} 01/29/2022 12:51:48 - INFO - codeparrot_training - Step 18680: {'lr': 0.0003652435936338656, 'samples': 3586752, 'steps': 18680, 'loss/train': 0.8924150168895721} 01/29/2022 12:51:52 - INFO - codeparrot_training - Step 18681: {'lr': 0.0003652290731184697, 'samples': 3586944, 'steps': 18681, 'loss/train': 1.6034325957298279} 01/29/2022 12:51:56 - INFO - codeparrot_training - Step 18682: {'lr': 0.000365214552109469, 'samples': 3587136, 'steps': 18682, 'loss/train': 1.7981071472167969} 01/29/2022 12:52:01 - INFO - codeparrot_training - Step 18683: {'lr': 0.0003652000306069258, 'samples': 3587328, 'steps': 18683, 'loss/train': 1.7556933760643005} 01/29/2022 12:52:06 - INFO - codeparrot_training - Step 18684: {'lr': 0.00036518550861090217, 'samples': 3587520, 'steps': 18684, 'loss/train': 1.634325385093689} 01/29/2022 12:52:10 - INFO - codeparrot_training - Step 18685: {'lr': 0.0003651709861214605, 'samples': 3587712, 'steps': 18685, 'loss/train': 1.2179193198680878} 01/29/2022 12:52:14 - INFO - codeparrot_training - Step 18686: {'lr': 0.0003651564631386628, 'samples': 3587904, 'steps': 18686, 'loss/train': 1.880557358264923} 01/29/2022 12:52:18 - INFO - codeparrot_training - Step 18687: {'lr': 0.0003651419396625714, 'samples': 3588096, 'steps': 18687, 'loss/train': 2.852054715156555} 01/29/2022 12:52:24 - INFO - codeparrot_training - Step 18688: {'lr': 0.0003651274156932485, 'samples': 3588288, 'steps': 18688, 'loss/train': 1.5318070650100708} 01/29/2022 12:52:28 - INFO - codeparrot_training - Step 18689: {'lr': 0.00036511289123075636, 'samples': 3588480, 'steps': 18689, 'loss/train': 1.7302865982055664} 01/29/2022 12:52:32 - INFO - codeparrot_training - Step 18690: {'lr': 0.0003650983662751571, 'samples': 3588672, 'steps': 18690, 'loss/train': 1.2562245726585388} 01/29/2022 12:52:36 - INFO - codeparrot_training - Step 18691: {'lr': 0.000365083840826513, 'samples': 3588864, 'steps': 18691, 'loss/train': 1.6308492422103882} 01/29/2022 12:52:41 - INFO - codeparrot_training - Step 18692: {'lr': 0.00036506931488488627, 'samples': 3589056, 'steps': 18692, 'loss/train': 1.6853617429733276} 01/29/2022 12:52:48 - INFO - codeparrot_training - Step 18693: {'lr': 0.0003650547884503391, 'samples': 3589248, 'steps': 18693, 'loss/train': 1.557268738746643} 01/29/2022 12:52:52 - INFO - codeparrot_training - Step 18694: {'lr': 0.0003650402615229338, 'samples': 3589440, 'steps': 18694, 'loss/train': 1.677379310131073} 01/29/2022 12:52:57 - INFO - codeparrot_training - Step 18695: {'lr': 0.00036502573410273243, 'samples': 3589632, 'steps': 18695, 'loss/train': 1.171916663646698} 01/29/2022 12:53:01 - INFO - codeparrot_training - Step 18696: {'lr': 0.0003650112061897975, 'samples': 3589824, 'steps': 18696, 'loss/train': 1.8450567126274109} 01/29/2022 12:53:05 - INFO - codeparrot_training - Step 18697: {'lr': 0.000364996677784191, 'samples': 3590016, 'steps': 18697, 'loss/train': 0.853326290845871} 01/29/2022 12:53:11 - INFO - codeparrot_training - Step 18698: {'lr': 0.00036498214888597524, 'samples': 3590208, 'steps': 18698, 'loss/train': 1.5603116154670715} 01/29/2022 12:53:15 - INFO - codeparrot_training - Step 18699: {'lr': 0.0003649676194952125, 'samples': 3590400, 'steps': 18699, 'loss/train': 1.7290053963661194} 01/29/2022 12:53:19 - INFO - codeparrot_training - Step 18700: {'lr': 0.000364953089611965, 'samples': 3590592, 'steps': 18700, 'loss/train': 1.182478666305542} 01/29/2022 12:53:24 - INFO - codeparrot_training - Step 18701: {'lr': 0.00036493855923629495, 'samples': 3590784, 'steps': 18701, 'loss/train': 1.1883784532546997} 01/29/2022 12:53:28 - INFO - codeparrot_training - Step 18702: {'lr': 0.00036492402836826453, 'samples': 3590976, 'steps': 18702, 'loss/train': 0.7216736823320389} 01/29/2022 12:53:33 - INFO - codeparrot_training - Step 18703: {'lr': 0.0003649094970079362, 'samples': 3591168, 'steps': 18703, 'loss/train': 1.9464651346206665} 01/29/2022 12:53:37 - INFO - codeparrot_training - Step 18704: {'lr': 0.000364894965155372, 'samples': 3591360, 'steps': 18704, 'loss/train': 1.4146501421928406} 01/29/2022 12:53:42 - INFO - codeparrot_training - Step 18705: {'lr': 0.0003648804328106342, 'samples': 3591552, 'steps': 18705, 'loss/train': 1.7312461137771606} 01/29/2022 12:53:46 - INFO - codeparrot_training - Step 18706: {'lr': 0.0003648658999737852, 'samples': 3591744, 'steps': 18706, 'loss/train': 1.9372714161872864} 01/29/2022 12:53:50 - INFO - codeparrot_training - Step 18707: {'lr': 0.0003648513666448871, 'samples': 3591936, 'steps': 18707, 'loss/train': 1.7814179062843323} 01/29/2022 12:53:58 - INFO - codeparrot_training - Step 18708: {'lr': 0.0003648368328240022, 'samples': 3592128, 'steps': 18708, 'loss/train': 0.6286721080541611} 01/29/2022 12:54:02 - INFO - codeparrot_training - Step 18709: {'lr': 0.00036482229851119287, 'samples': 3592320, 'steps': 18709, 'loss/train': 1.1059933304786682} 01/29/2022 12:54:06 - INFO - codeparrot_training - Step 18710: {'lr': 0.0003648077637065212, 'samples': 3592512, 'steps': 18710, 'loss/train': 2.1714711785316467} 01/29/2022 12:54:11 - INFO - codeparrot_training - Step 18711: {'lr': 0.00036479322841004953, 'samples': 3592704, 'steps': 18711, 'loss/train': 6.40164041519165} 01/29/2022 12:54:15 - INFO - codeparrot_training - Step 18712: {'lr': 0.0003647786926218401, 'samples': 3592896, 'steps': 18712, 'loss/train': 1.8112589120864868} 01/29/2022 12:54:21 - INFO - codeparrot_training - Step 18713: {'lr': 0.00036476415634195523, 'samples': 3593088, 'steps': 18713, 'loss/train': 1.8041004538536072} 01/29/2022 12:54:25 - INFO - codeparrot_training - Step 18714: {'lr': 0.00036474961957045715, 'samples': 3593280, 'steps': 18714, 'loss/train': 0.20810526609420776} 01/29/2022 12:54:29 - INFO - codeparrot_training - Step 18715: {'lr': 0.00036473508230740816, 'samples': 3593472, 'steps': 18715, 'loss/train': 0.3600281774997711} 01/29/2022 12:54:33 - INFO - codeparrot_training - Step 18716: {'lr': 0.00036472054455287053, 'samples': 3593664, 'steps': 18716, 'loss/train': 2.053857922554016} 01/29/2022 12:54:38 - INFO - codeparrot_training - Step 18717: {'lr': 0.0003647060063069064, 'samples': 3593856, 'steps': 18717, 'loss/train': 2.4095755219459534} 01/29/2022 12:54:43 - INFO - codeparrot_training - Step 18718: {'lr': 0.0003646914675695783, 'samples': 3594048, 'steps': 18718, 'loss/train': 2.5709540247917175} 01/29/2022 12:54:47 - INFO - codeparrot_training - Step 18719: {'lr': 0.0003646769283409483, 'samples': 3594240, 'steps': 18719, 'loss/train': 1.77484792470932} 01/29/2022 12:54:51 - INFO - codeparrot_training - Step 18720: {'lr': 0.0003646623886210788, 'samples': 3594432, 'steps': 18720, 'loss/train': 1.606357455253601} 01/29/2022 12:54:56 - INFO - codeparrot_training - Step 18721: {'lr': 0.00036464784841003196, 'samples': 3594624, 'steps': 18721, 'loss/train': 1.801158607006073} 01/29/2022 12:55:00 - INFO - codeparrot_training - Step 18722: {'lr': 0.0003646333077078702, 'samples': 3594816, 'steps': 18722, 'loss/train': 1.3337555229663849} 01/29/2022 12:55:07 - INFO - codeparrot_training - Step 18723: {'lr': 0.0003646187665146557, 'samples': 3595008, 'steps': 18723, 'loss/train': 2.071053206920624} 01/29/2022 12:55:11 - INFO - codeparrot_training - Step 18724: {'lr': 0.00036460422483045084, 'samples': 3595200, 'steps': 18724, 'loss/train': 1.525860071182251} 01/29/2022 12:55:15 - INFO - codeparrot_training - Step 18725: {'lr': 0.0003645896826553178, 'samples': 3595392, 'steps': 18725, 'loss/train': 1.339848518371582} 01/29/2022 12:55:20 - INFO - codeparrot_training - Step 18726: {'lr': 0.0003645751399893191, 'samples': 3595584, 'steps': 18726, 'loss/train': 1.4633881151676178} 01/29/2022 12:55:24 - INFO - codeparrot_training - Step 18727: {'lr': 0.0003645605968325167, 'samples': 3595776, 'steps': 18727, 'loss/train': 1.4059519171714783} 01/29/2022 12:55:29 - INFO - codeparrot_training - Step 18728: {'lr': 0.00036454605318497323, 'samples': 3595968, 'steps': 18728, 'loss/train': 2.035488545894623} 01/29/2022 12:55:34 - INFO - codeparrot_training - Step 18729: {'lr': 0.00036453150904675074, 'samples': 3596160, 'steps': 18729, 'loss/train': 2.020914852619171} 01/29/2022 12:55:38 - INFO - codeparrot_training - Step 18730: {'lr': 0.00036451696441791164, 'samples': 3596352, 'steps': 18730, 'loss/train': 1.5732387900352478} 01/29/2022 12:55:42 - INFO - codeparrot_training - Step 18731: {'lr': 0.0003645024192985183, 'samples': 3596544, 'steps': 18731, 'loss/train': 1.3006778955459595} 01/29/2022 12:55:46 - INFO - codeparrot_training - Step 18732: {'lr': 0.0003644878736886329, 'samples': 3596736, 'steps': 18732, 'loss/train': 2.283536374568939} 01/29/2022 12:55:54 - INFO - codeparrot_training - Step 18733: {'lr': 0.0003644733275883179, 'samples': 3596928, 'steps': 18733, 'loss/train': 1.8078904151916504} 01/29/2022 12:55:58 - INFO - codeparrot_training - Step 18734: {'lr': 0.00036445878099763534, 'samples': 3597120, 'steps': 18734, 'loss/train': 2.021205425262451} 01/29/2022 12:56:02 - INFO - codeparrot_training - Step 18735: {'lr': 0.00036444423391664783, 'samples': 3597312, 'steps': 18735, 'loss/train': 1.3842618763446808} 01/29/2022 12:56:06 - INFO - codeparrot_training - Step 18736: {'lr': 0.0003644296863454175, 'samples': 3597504, 'steps': 18736, 'loss/train': 2.1629297733306885} 01/29/2022 12:56:11 - INFO - codeparrot_training - Step 18737: {'lr': 0.0003644151382840068, 'samples': 3597696, 'steps': 18737, 'loss/train': 1.3590403497219086} 01/29/2022 12:56:16 - INFO - codeparrot_training - Step 18738: {'lr': 0.00036440058973247793, 'samples': 3597888, 'steps': 18738, 'loss/train': 0.7404333204030991} 01/29/2022 12:56:20 - INFO - codeparrot_training - Step 18739: {'lr': 0.0003643860406908933, 'samples': 3598080, 'steps': 18739, 'loss/train': 1.3107792735099792} 01/29/2022 12:56:24 - INFO - codeparrot_training - Step 18740: {'lr': 0.0003643714911593151, 'samples': 3598272, 'steps': 18740, 'loss/train': 0.7999853789806366} 01/29/2022 12:56:28 - INFO - codeparrot_training - Step 18741: {'lr': 0.00036435694113780585, 'samples': 3598464, 'steps': 18741, 'loss/train': 0.853234201669693} 01/29/2022 12:56:33 - INFO - codeparrot_training - Step 18742: {'lr': 0.0003643423906264277, 'samples': 3598656, 'steps': 18742, 'loss/train': 1.3464604318141937} 01/29/2022 12:56:38 - INFO - codeparrot_training - Step 18743: {'lr': 0.0003643278396252431, 'samples': 3598848, 'steps': 18743, 'loss/train': 1.4937594830989838} 01/29/2022 12:56:43 - INFO - codeparrot_training - Step 18744: {'lr': 0.0003643132881343144, 'samples': 3599040, 'steps': 18744, 'loss/train': 1.380234181880951} 01/29/2022 12:56:47 - INFO - codeparrot_training - Step 18745: {'lr': 0.00036429873615370374, 'samples': 3599232, 'steps': 18745, 'loss/train': 1.6750166416168213} 01/29/2022 12:56:51 - INFO - codeparrot_training - Step 18746: {'lr': 0.00036428418368347363, 'samples': 3599424, 'steps': 18746, 'loss/train': 1.556464433670044} 01/29/2022 12:56:55 - INFO - codeparrot_training - Step 18747: {'lr': 0.0003642696307236864, 'samples': 3599616, 'steps': 18747, 'loss/train': 0.7349376082420349} 01/29/2022 12:57:02 - INFO - codeparrot_training - Step 18748: {'lr': 0.0003642550772744044, 'samples': 3599808, 'steps': 18748, 'loss/train': 1.3048003613948822} 01/29/2022 12:57:07 - INFO - codeparrot_training - Step 18749: {'lr': 0.0003642405233356898, 'samples': 3600000, 'steps': 18749, 'loss/train': 1.504185676574707} 01/29/2022 12:57:11 - INFO - codeparrot_training - Step 18750: {'lr': 0.00036422596890760517, 'samples': 3600192, 'steps': 18750, 'loss/train': 2.2901682257652283} 01/29/2022 12:57:15 - INFO - codeparrot_training - Step 18751: {'lr': 0.00036421141399021274, 'samples': 3600384, 'steps': 18751, 'loss/train': 0.1376195065677166} 01/29/2022 12:57:19 - INFO - codeparrot_training - Step 18752: {'lr': 0.00036419685858357485, 'samples': 3600576, 'steps': 18752, 'loss/train': 2.287505865097046} 01/29/2022 12:57:25 - INFO - codeparrot_training - Step 18753: {'lr': 0.00036418230268775393, 'samples': 3600768, 'steps': 18753, 'loss/train': 1.0784751176834106} 01/29/2022 12:57:29 - INFO - codeparrot_training - Step 18754: {'lr': 0.0003641677463028123, 'samples': 3600960, 'steps': 18754, 'loss/train': 1.2982756197452545} 01/29/2022 12:57:33 - INFO - codeparrot_training - Step 18755: {'lr': 0.0003641531894288122, 'samples': 3601152, 'steps': 18755, 'loss/train': 1.480896234512329} 01/29/2022 12:57:38 - INFO - codeparrot_training - Step 18756: {'lr': 0.0003641386320658161, 'samples': 3601344, 'steps': 18756, 'loss/train': 1.3294234871864319} 01/29/2022 12:57:42 - INFO - codeparrot_training - Step 18757: {'lr': 0.00036412407421388646, 'samples': 3601536, 'steps': 18757, 'loss/train': 1.7139846682548523} 01/29/2022 12:57:47 - INFO - codeparrot_training - Step 18758: {'lr': 0.00036410951587308545, 'samples': 3601728, 'steps': 18758, 'loss/train': 1.4932655096054077} 01/29/2022 12:57:52 - INFO - codeparrot_training - Step 18759: {'lr': 0.00036409495704347553, 'samples': 3601920, 'steps': 18759, 'loss/train': 1.9205800294876099} 01/29/2022 12:57:56 - INFO - codeparrot_training - Step 18760: {'lr': 0.000364080397725119, 'samples': 3602112, 'steps': 18760, 'loss/train': 1.063752830028534} 01/29/2022 12:58:00 - INFO - codeparrot_training - Step 18761: {'lr': 0.00036406583791807824, 'samples': 3602304, 'steps': 18761, 'loss/train': 2.018700063228607} 01/29/2022 12:58:06 - INFO - codeparrot_training - Step 18762: {'lr': 0.0003640512776224157, 'samples': 3602496, 'steps': 18762, 'loss/train': 2.1599273085594177} 01/29/2022 12:58:10 - INFO - codeparrot_training - Step 18763: {'lr': 0.0003640367168381937, 'samples': 3602688, 'steps': 18763, 'loss/train': 1.5168058276176453} 01/29/2022 12:58:14 - INFO - codeparrot_training - Step 18764: {'lr': 0.0003640221555654747, 'samples': 3602880, 'steps': 18764, 'loss/train': 1.6747369766235352} 01/29/2022 12:58:18 - INFO - codeparrot_training - Step 18765: {'lr': 0.00036400759380432083, 'samples': 3603072, 'steps': 18765, 'loss/train': 1.549962043762207} 01/29/2022 12:58:23 - INFO - codeparrot_training - Step 18766: {'lr': 0.00036399303155479476, 'samples': 3603264, 'steps': 18766, 'loss/train': 1.7996007800102234} 01/29/2022 12:58:30 - INFO - codeparrot_training - Step 18767: {'lr': 0.00036397846881695866, 'samples': 3603456, 'steps': 18767, 'loss/train': 1.7633585929870605} 01/29/2022 12:58:34 - INFO - codeparrot_training - Step 18768: {'lr': 0.00036396390559087505, 'samples': 3603648, 'steps': 18768, 'loss/train': 1.6559926271438599} 01/29/2022 12:58:38 - INFO - codeparrot_training - Step 18769: {'lr': 0.0003639493418766062, 'samples': 3603840, 'steps': 18769, 'loss/train': 1.5028225779533386} 01/29/2022 12:58:42 - INFO - codeparrot_training - Step 18770: {'lr': 0.0003639347776742146, 'samples': 3604032, 'steps': 18770, 'loss/train': 2.0127135515213013} 01/29/2022 12:58:47 - INFO - codeparrot_training - Step 18771: {'lr': 0.00036392021298376257, 'samples': 3604224, 'steps': 18771, 'loss/train': 1.415427714586258} 01/29/2022 12:58:52 - INFO - codeparrot_training - Step 18772: {'lr': 0.0003639056478053125, 'samples': 3604416, 'steps': 18772, 'loss/train': 1.7980901598930359} 01/29/2022 12:58:56 - INFO - codeparrot_training - Step 18773: {'lr': 0.0003638910821389268, 'samples': 3604608, 'steps': 18773, 'loss/train': 1.2051030099391937} 01/29/2022 12:59:00 - INFO - codeparrot_training - Step 18774: {'lr': 0.0003638765159846679, 'samples': 3604800, 'steps': 18774, 'loss/train': 1.4746043980121613} 01/29/2022 12:59:05 - INFO - codeparrot_training - Step 18775: {'lr': 0.0003638619493425982, 'samples': 3604992, 'steps': 18775, 'loss/train': 1.5147619843482971} 01/29/2022 12:59:09 - INFO - codeparrot_training - Step 18776: {'lr': 0.00036384738221278, 'samples': 3605184, 'steps': 18776, 'loss/train': 2.2832086086273193} 01/29/2022 12:59:16 - INFO - codeparrot_training - Step 18777: {'lr': 0.0003638328145952758, 'samples': 3605376, 'steps': 18777, 'loss/train': 1.9759510159492493} 01/29/2022 12:59:20 - INFO - codeparrot_training - Step 18778: {'lr': 0.0003638182464901479, 'samples': 3605568, 'steps': 18778, 'loss/train': 1.7735726237297058} 01/29/2022 12:59:24 - INFO - codeparrot_training - Step 18779: {'lr': 0.00036380367789745887, 'samples': 3605760, 'steps': 18779, 'loss/train': 1.9765863418579102} 01/29/2022 12:59:29 - INFO - codeparrot_training - Step 18780: {'lr': 0.00036378910881727095, 'samples': 3605952, 'steps': 18780, 'loss/train': 1.7733877301216125} 01/29/2022 12:59:33 - INFO - codeparrot_training - Step 18781: {'lr': 0.00036377453924964665, 'samples': 3606144, 'steps': 18781, 'loss/train': 0.4212532639503479} 01/29/2022 12:59:38 - INFO - codeparrot_training - Step 18782: {'lr': 0.0003637599691946484, 'samples': 3606336, 'steps': 18782, 'loss/train': 1.19785675406456} 01/29/2022 12:59:42 - INFO - codeparrot_training - Step 18783: {'lr': 0.00036374539865233847, 'samples': 3606528, 'steps': 18783, 'loss/train': 1.2913153767585754} 01/29/2022 12:59:47 - INFO - codeparrot_training - Step 18784: {'lr': 0.00036373082762277943, 'samples': 3606720, 'steps': 18784, 'loss/train': 2.0619667768478394} 01/29/2022 12:59:51 - INFO - codeparrot_training - Step 18785: {'lr': 0.00036371625610603366, 'samples': 3606912, 'steps': 18785, 'loss/train': 1.481520652770996} 01/29/2022 12:59:55 - INFO - codeparrot_training - Step 18786: {'lr': 0.0003637016841021635, 'samples': 3607104, 'steps': 18786, 'loss/train': 1.6758023500442505} 01/29/2022 13:00:00 - INFO - codeparrot_training - Step 18787: {'lr': 0.00036368711161123145, 'samples': 3607296, 'steps': 18787, 'loss/train': 2.240450084209442} 01/29/2022 13:00:05 - INFO - codeparrot_training - Step 18788: {'lr': 0.0003636725386332999, 'samples': 3607488, 'steps': 18788, 'loss/train': 2.0388253927230835} 01/29/2022 13:00:09 - INFO - codeparrot_training - Step 18789: {'lr': 0.0003636579651684313, 'samples': 3607680, 'steps': 18789, 'loss/train': 2.1943466663360596} 01/29/2022 13:00:13 - INFO - codeparrot_training - Step 18790: {'lr': 0.0003636433912166881, 'samples': 3607872, 'steps': 18790, 'loss/train': 1.145098239183426} 01/29/2022 13:00:17 - INFO - codeparrot_training - Step 18791: {'lr': 0.00036362881677813266, 'samples': 3608064, 'steps': 18791, 'loss/train': 1.4641976058483124} 01/29/2022 13:00:25 - INFO - codeparrot_training - Step 18792: {'lr': 0.00036361424185282743, 'samples': 3608256, 'steps': 18792, 'loss/train': 1.5567678809165955} 01/29/2022 13:00:29 - INFO - codeparrot_training - Step 18793: {'lr': 0.00036359966644083485, 'samples': 3608448, 'steps': 18793, 'loss/train': 1.2069793045520782} 01/29/2022 13:00:33 - INFO - codeparrot_training - Step 18794: {'lr': 0.0003635850905422175, 'samples': 3608640, 'steps': 18794, 'loss/train': 1.1432298123836517} 01/29/2022 13:00:37 - INFO - codeparrot_training - Step 18795: {'lr': 0.00036357051415703755, 'samples': 3608832, 'steps': 18795, 'loss/train': 0.9562841355800629} 01/29/2022 13:00:42 - INFO - codeparrot_training - Step 18796: {'lr': 0.00036355593728535767, 'samples': 3609024, 'steps': 18796, 'loss/train': 2.2667779326438904} 01/29/2022 13:00:47 - INFO - codeparrot_training - Step 18797: {'lr': 0.00036354135992724013, 'samples': 3609216, 'steps': 18797, 'loss/train': 0.33310288935899734} 01/29/2022 13:00:52 - INFO - codeparrot_training - Step 18798: {'lr': 0.0003635267820827476, 'samples': 3609408, 'steps': 18798, 'loss/train': 1.789575219154358} 01/29/2022 13:00:56 - INFO - codeparrot_training - Step 18799: {'lr': 0.0003635122037519422, 'samples': 3609600, 'steps': 18799, 'loss/train': 1.2936218082904816} 01/29/2022 13:01:00 - INFO - codeparrot_training - Step 18800: {'lr': 0.00036349762493488667, 'samples': 3609792, 'steps': 18800, 'loss/train': 1.4442803263664246} 01/29/2022 13:01:04 - INFO - codeparrot_training - Step 18801: {'lr': 0.0003634830456316434, 'samples': 3609984, 'steps': 18801, 'loss/train': 1.5475299954414368} 01/29/2022 13:01:11 - INFO - codeparrot_training - Step 18802: {'lr': 0.00036346846584227473, 'samples': 3610176, 'steps': 18802, 'loss/train': 1.4341148436069489} 01/29/2022 13:01:15 - INFO - codeparrot_training - Step 18803: {'lr': 0.0003634538855668432, 'samples': 3610368, 'steps': 18803, 'loss/train': 0.9992045760154724} 01/29/2022 13:01:20 - INFO - codeparrot_training - Step 18804: {'lr': 0.00036343930480541123, 'samples': 3610560, 'steps': 18804, 'loss/train': 1.4918714761734009} 01/29/2022 13:01:24 - INFO - codeparrot_training - Step 18805: {'lr': 0.0003634247235580413, 'samples': 3610752, 'steps': 18805, 'loss/train': 2.5570027828216553} 01/29/2022 13:01:28 - INFO - codeparrot_training - Step 18806: {'lr': 0.0003634101418247959, 'samples': 3610944, 'steps': 18806, 'loss/train': 0.9795911908149719} 01/29/2022 13:01:34 - INFO - codeparrot_training - Step 18807: {'lr': 0.0003633955596057374, 'samples': 3611136, 'steps': 18807, 'loss/train': 1.2458516657352448} 01/29/2022 13:01:38 - INFO - codeparrot_training - Step 18808: {'lr': 0.00036338097690092843, 'samples': 3611328, 'steps': 18808, 'loss/train': 1.812943160533905} 01/29/2022 13:01:42 - INFO - codeparrot_training - Step 18809: {'lr': 0.0003633663937104313, 'samples': 3611520, 'steps': 18809, 'loss/train': 0.46776795387268066} 01/29/2022 13:01:46 - INFO - codeparrot_training - Step 18810: {'lr': 0.0003633518100343085, 'samples': 3611712, 'steps': 18810, 'loss/train': 1.3189018070697784} 01/29/2022 13:01:51 - INFO - codeparrot_training - Step 18811: {'lr': 0.0003633372258726226, 'samples': 3611904, 'steps': 18811, 'loss/train': 0.482195645570755} 01/29/2022 13:01:56 - INFO - codeparrot_training - Step 18812: {'lr': 0.00036332264122543594, 'samples': 3612096, 'steps': 18812, 'loss/train': 2.0041980743408203} 01/29/2022 13:02:00 - INFO - codeparrot_training - Step 18813: {'lr': 0.0003633080560928111, 'samples': 3612288, 'steps': 18813, 'loss/train': 1.394830048084259} 01/29/2022 13:02:04 - INFO - codeparrot_training - Step 18814: {'lr': 0.0003632934704748106, 'samples': 3612480, 'steps': 18814, 'loss/train': 1.4253646731376648} 01/29/2022 13:02:09 - INFO - codeparrot_training - Step 18815: {'lr': 0.00036327888437149674, 'samples': 3612672, 'steps': 18815, 'loss/train': 1.0837603211402893} 01/29/2022 13:02:13 - INFO - codeparrot_training - Step 18816: {'lr': 0.00036326429778293223, 'samples': 3612864, 'steps': 18816, 'loss/train': 1.79723060131073} 01/29/2022 13:02:20 - INFO - codeparrot_training - Step 18817: {'lr': 0.00036324971070917934, 'samples': 3613056, 'steps': 18817, 'loss/train': 1.0986398756504059} 01/29/2022 13:02:25 - INFO - codeparrot_training - Step 18818: {'lr': 0.00036323512315030067, 'samples': 3613248, 'steps': 18818, 'loss/train': 0.796992838382721} 01/29/2022 13:02:29 - INFO - codeparrot_training - Step 18819: {'lr': 0.0003632205351063587, 'samples': 3613440, 'steps': 18819, 'loss/train': 1.544754445552826} 01/29/2022 13:02:33 - INFO - codeparrot_training - Step 18820: {'lr': 0.0003632059465774159, 'samples': 3613632, 'steps': 18820, 'loss/train': 2.3184673190116882} 01/29/2022 13:02:37 - INFO - codeparrot_training - Step 18821: {'lr': 0.0003631913575635348, 'samples': 3613824, 'steps': 18821, 'loss/train': 1.3588120937347412} 01/29/2022 13:02:43 - INFO - codeparrot_training - Step 18822: {'lr': 0.00036317676806477784, 'samples': 3614016, 'steps': 18822, 'loss/train': 1.9357243180274963} 01/29/2022 13:02:47 - INFO - codeparrot_training - Step 18823: {'lr': 0.0003631621780812075, 'samples': 3614208, 'steps': 18823, 'loss/train': 2.1397050619125366} 01/29/2022 13:02:51 - INFO - codeparrot_training - Step 18824: {'lr': 0.00036314758761288643, 'samples': 3614400, 'steps': 18824, 'loss/train': 2.053815186023712} 01/29/2022 13:02:56 - INFO - codeparrot_training - Step 18825: {'lr': 0.0003631329966598769, 'samples': 3614592, 'steps': 18825, 'loss/train': 1.6951489448547363} 01/29/2022 13:03:00 - INFO - codeparrot_training - Step 18826: {'lr': 0.0003631184052222416, 'samples': 3614784, 'steps': 18826, 'loss/train': 1.1982732117176056} 01/29/2022 13:03:05 - INFO - codeparrot_training - Step 18827: {'lr': 0.00036310381330004296, 'samples': 3614976, 'steps': 18827, 'loss/train': 1.0155166983604431} 01/29/2022 13:03:09 - INFO - codeparrot_training - Step 18828: {'lr': 0.0003630892208933435, 'samples': 3615168, 'steps': 18828, 'loss/train': 1.4294448494911194} 01/29/2022 13:03:14 - INFO - codeparrot_training - Step 18829: {'lr': 0.00036307462800220575, 'samples': 3615360, 'steps': 18829, 'loss/train': 1.0535415709018707} 01/29/2022 13:03:18 - INFO - codeparrot_training - Step 18830: {'lr': 0.0003630600346266922, 'samples': 3615552, 'steps': 18830, 'loss/train': 0.960901290178299} 01/29/2022 13:03:22 - INFO - codeparrot_training - Step 18831: {'lr': 0.00036304544076686527, 'samples': 3615744, 'steps': 18831, 'loss/train': 1.8144900798797607} 01/29/2022 13:03:27 - INFO - codeparrot_training - Step 18832: {'lr': 0.0003630308464227877, 'samples': 3615936, 'steps': 18832, 'loss/train': 1.1855605244636536} 01/29/2022 13:03:31 - INFO - codeparrot_training - Step 18833: {'lr': 0.0003630162515945218, 'samples': 3616128, 'steps': 18833, 'loss/train': 0.31805428862571716} 01/29/2022 13:03:36 - INFO - codeparrot_training - Step 18834: {'lr': 0.00036300165628213015, 'samples': 3616320, 'steps': 18834, 'loss/train': 1.9095953106880188} 01/29/2022 13:03:40 - INFO - codeparrot_training - Step 18835: {'lr': 0.0003629870604856754, 'samples': 3616512, 'steps': 18835, 'loss/train': 2.1730032563209534} 01/29/2022 13:03:44 - INFO - codeparrot_training - Step 18836: {'lr': 0.0003629724642052198, 'samples': 3616704, 'steps': 18836, 'loss/train': 2.29136198759079} 01/29/2022 13:03:52 - INFO - codeparrot_training - Step 18837: {'lr': 0.00036295786744082616, 'samples': 3616896, 'steps': 18837, 'loss/train': 1.6186999082565308} 01/29/2022 13:03:56 - INFO - codeparrot_training - Step 18838: {'lr': 0.0003629432701925568, 'samples': 3617088, 'steps': 18838, 'loss/train': 0.916742205619812} 01/29/2022 13:04:00 - INFO - codeparrot_training - Step 18839: {'lr': 0.0003629286724604744, 'samples': 3617280, 'steps': 18839, 'loss/train': 1.5189651846885681} 01/29/2022 13:04:04 - INFO - codeparrot_training - Step 18840: {'lr': 0.0003629140742446414, 'samples': 3617472, 'steps': 18840, 'loss/train': 1.7847899794578552} 01/29/2022 13:04:09 - INFO - codeparrot_training - Step 18841: {'lr': 0.00036289947554512034, 'samples': 3617664, 'steps': 18841, 'loss/train': 1.6998265385627747} 01/29/2022 13:04:14 - INFO - codeparrot_training - Step 18842: {'lr': 0.0003628848763619738, 'samples': 3617856, 'steps': 18842, 'loss/train': 1.6256802678108215} 01/29/2022 13:04:18 - INFO - codeparrot_training - Step 18843: {'lr': 0.0003628702766952643, 'samples': 3618048, 'steps': 18843, 'loss/train': 1.8198076486587524} 01/29/2022 13:04:23 - INFO - codeparrot_training - Step 18844: {'lr': 0.00036285567654505433, 'samples': 3618240, 'steps': 18844, 'loss/train': 1.9973516464233398} 01/29/2022 13:04:27 - INFO - codeparrot_training - Step 18845: {'lr': 0.00036284107591140653, 'samples': 3618432, 'steps': 18845, 'loss/train': 1.5361804962158203} 01/29/2022 13:04:34 - INFO - codeparrot_training - Step 18846: {'lr': 0.0003628264747943834, 'samples': 3618624, 'steps': 18846, 'loss/train': 1.9246591329574585} 01/29/2022 13:04:39 - INFO - codeparrot_training - Step 18847: {'lr': 0.0003628118731940475, 'samples': 3618816, 'steps': 18847, 'loss/train': 0.9112773835659027} 01/29/2022 13:04:43 - INFO - codeparrot_training - Step 18848: {'lr': 0.00036279727111046127, 'samples': 3619008, 'steps': 18848, 'loss/train': 2.4143117666244507} 01/29/2022 13:04:47 - INFO - codeparrot_training - Step 18849: {'lr': 0.0003627826685436874, 'samples': 3619200, 'steps': 18849, 'loss/train': 1.3006608188152313} 01/29/2022 13:04:51 - INFO - codeparrot_training - Step 18850: {'lr': 0.00036276806549378836, 'samples': 3619392, 'steps': 18850, 'loss/train': 1.7037312984466553} 01/29/2022 13:04:57 - INFO - codeparrot_training - Step 18851: {'lr': 0.0003627534619608268, 'samples': 3619584, 'steps': 18851, 'loss/train': 1.5812859535217285} 01/29/2022 13:05:01 - INFO - codeparrot_training - Step 18852: {'lr': 0.00036273885794486514, 'samples': 3619776, 'steps': 18852, 'loss/train': 1.7801012992858887} 01/29/2022 13:05:05 - INFO - codeparrot_training - Step 18853: {'lr': 0.00036272425344596607, 'samples': 3619968, 'steps': 18853, 'loss/train': 1.7941746711730957} 01/29/2022 13:05:09 - INFO - codeparrot_training - Step 18854: {'lr': 0.000362709648464192, 'samples': 3620160, 'steps': 18854, 'loss/train': 0.9573946595191956} 01/29/2022 13:05:14 - INFO - codeparrot_training - Step 18855: {'lr': 0.00036269504299960573, 'samples': 3620352, 'steps': 18855, 'loss/train': 1.449877291917801} 01/29/2022 13:05:19 - INFO - codeparrot_training - Step 18856: {'lr': 0.00036268043705226953, 'samples': 3620544, 'steps': 18856, 'loss/train': 1.475682020187378} 01/29/2022 13:05:23 - INFO - codeparrot_training - Step 18857: {'lr': 0.00036266583062224625, 'samples': 3620736, 'steps': 18857, 'loss/train': 1.641392469406128} 01/29/2022 13:05:27 - INFO - codeparrot_training - Step 18858: {'lr': 0.0003626512237095982, 'samples': 3620928, 'steps': 18858, 'loss/train': 0.5724369138479233} 01/29/2022 13:05:32 - INFO - codeparrot_training - Step 18859: {'lr': 0.00036263661631438814, 'samples': 3621120, 'steps': 18859, 'loss/train': 1.0783506631851196} 01/29/2022 13:05:36 - INFO - codeparrot_training - Step 18860: {'lr': 0.0003626220084366786, 'samples': 3621312, 'steps': 18860, 'loss/train': 1.5686402320861816} 01/29/2022 13:05:43 - INFO - codeparrot_training - Step 18861: {'lr': 0.00036260740007653216, 'samples': 3621504, 'steps': 18861, 'loss/train': 1.8916903138160706} 01/29/2022 13:05:47 - INFO - codeparrot_training - Step 18862: {'lr': 0.00036259279123401127, 'samples': 3621696, 'steps': 18862, 'loss/train': 1.5539345741271973} 01/29/2022 13:05:52 - INFO - codeparrot_training - Step 18863: {'lr': 0.0003625781819091787, 'samples': 3621888, 'steps': 18863, 'loss/train': 1.9387698769569397} 01/29/2022 13:05:56 - INFO - codeparrot_training - Step 18864: {'lr': 0.0003625635721020969, 'samples': 3622080, 'steps': 18864, 'loss/train': 1.0637215375900269} 01/29/2022 13:06:00 - INFO - codeparrot_training - Step 18865: {'lr': 0.00036254896181282846, 'samples': 3622272, 'steps': 18865, 'loss/train': 1.8814911246299744} 01/29/2022 13:06:05 - INFO - codeparrot_training - Step 18866: {'lr': 0.0003625343510414362, 'samples': 3622464, 'steps': 18866, 'loss/train': 1.6990883946418762} 01/29/2022 13:06:10 - INFO - codeparrot_training - Step 18867: {'lr': 0.0003625197397879823, 'samples': 3622656, 'steps': 18867, 'loss/train': 1.8027557730674744} 01/29/2022 13:06:14 - INFO - codeparrot_training - Step 18868: {'lr': 0.0003625051280525297, 'samples': 3622848, 'steps': 18868, 'loss/train': 1.5799407362937927} 01/29/2022 13:06:18 - INFO - codeparrot_training - Step 18869: {'lr': 0.00036249051583514075, 'samples': 3623040, 'steps': 18869, 'loss/train': 2.317506730556488} 01/29/2022 13:06:22 - INFO - codeparrot_training - Step 18870: {'lr': 0.00036247590313587827, 'samples': 3623232, 'steps': 18870, 'loss/train': 1.6453077793121338} 01/29/2022 13:06:27 - INFO - codeparrot_training - Step 18871: {'lr': 0.0003624612899548046, 'samples': 3623424, 'steps': 18871, 'loss/train': 1.6273632645606995} 01/29/2022 13:06:32 - INFO - codeparrot_training - Step 18872: {'lr': 0.0003624466762919826, 'samples': 3623616, 'steps': 18872, 'loss/train': 1.9479278326034546} 01/29/2022 13:06:36 - INFO - codeparrot_training - Step 18873: {'lr': 0.00036243206214747466, 'samples': 3623808, 'steps': 18873, 'loss/train': 2.250178277492523} 01/29/2022 13:06:40 - INFO - codeparrot_training - Step 18874: {'lr': 0.00036241744752134347, 'samples': 3624000, 'steps': 18874, 'loss/train': 1.9786909818649292} 01/29/2022 13:06:44 - INFO - codeparrot_training - Step 18875: {'lr': 0.0003624028324136517, 'samples': 3624192, 'steps': 18875, 'loss/train': 0.48438186943531036} 01/29/2022 13:06:52 - INFO - codeparrot_training - Step 18876: {'lr': 0.00036238821682446176, 'samples': 3624384, 'steps': 18876, 'loss/train': 1.8979681134223938} 01/29/2022 13:06:56 - INFO - codeparrot_training - Step 18877: {'lr': 0.0003623736007538365, 'samples': 3624576, 'steps': 18877, 'loss/train': 1.5703493356704712} 01/29/2022 13:07:00 - INFO - codeparrot_training - Step 18878: {'lr': 0.00036235898420183837, 'samples': 3624768, 'steps': 18878, 'loss/train': 0.9158120155334473} 01/29/2022 13:07:04 - INFO - codeparrot_training - Step 18879: {'lr': 0.0003623443671685301, 'samples': 3624960, 'steps': 18879, 'loss/train': 0.33702249079942703} 01/29/2022 13:07:09 - INFO - codeparrot_training - Step 18880: {'lr': 0.00036232974965397414, 'samples': 3625152, 'steps': 18880, 'loss/train': 1.2326836287975311} 01/29/2022 13:07:14 - INFO - codeparrot_training - Step 18881: {'lr': 0.00036231513165823324, 'samples': 3625344, 'steps': 18881, 'loss/train': 0.9185325801372528} 01/29/2022 13:07:18 - INFO - codeparrot_training - Step 18882: {'lr': 0.00036230051318137, 'samples': 3625536, 'steps': 18882, 'loss/train': 2.107521414756775} 01/29/2022 13:07:23 - INFO - codeparrot_training - Step 18883: {'lr': 0.00036228589422344703, 'samples': 3625728, 'steps': 18883, 'loss/train': 0.08126913942396641} 01/29/2022 13:07:27 - INFO - codeparrot_training - Step 18884: {'lr': 0.0003622712747845269, 'samples': 3625920, 'steps': 18884, 'loss/train': 1.7925117015838623} 01/29/2022 13:07:31 - INFO - codeparrot_training - Step 18885: {'lr': 0.0003622566548646723, 'samples': 3626112, 'steps': 18885, 'loss/train': 1.451992928981781} 01/29/2022 13:07:37 - INFO - codeparrot_training - Step 18886: {'lr': 0.00036224203446394584, 'samples': 3626304, 'steps': 18886, 'loss/train': 1.4601687490940094} 01/29/2022 13:07:41 - INFO - codeparrot_training - Step 18887: {'lr': 0.00036222741358241014, 'samples': 3626496, 'steps': 18887, 'loss/train': 1.4935105741024017} 01/29/2022 13:07:45 - INFO - codeparrot_training - Step 18888: {'lr': 0.0003622127922201278, 'samples': 3626688, 'steps': 18888, 'loss/train': 0.9174356460571289} 01/29/2022 13:07:50 - INFO - codeparrot_training - Step 18889: {'lr': 0.0003621981703771616, 'samples': 3626880, 'steps': 18889, 'loss/train': 1.1771468818187714} 01/29/2022 13:07:54 - INFO - codeparrot_training - Step 18890: {'lr': 0.0003621835480535739, 'samples': 3627072, 'steps': 18890, 'loss/train': 2.0752190351486206} 01/29/2022 13:07:59 - INFO - codeparrot_training - Step 18891: {'lr': 0.00036216892524942764, 'samples': 3627264, 'steps': 18891, 'loss/train': 1.659877896308899} 01/29/2022 13:08:03 - INFO - codeparrot_training - Step 18892: {'lr': 0.00036215430196478526, 'samples': 3627456, 'steps': 18892, 'loss/train': 2.1779011487960815} 01/29/2022 13:08:08 - INFO - codeparrot_training - Step 18893: {'lr': 0.0003621396781997095, 'samples': 3627648, 'steps': 18893, 'loss/train': 1.573481261730194} 01/29/2022 13:08:12 - INFO - codeparrot_training - Step 18894: {'lr': 0.00036212505395426297, 'samples': 3627840, 'steps': 18894, 'loss/train': 1.5893195271492004} 01/29/2022 13:08:16 - INFO - codeparrot_training - Step 18895: {'lr': 0.0003621104292285082, 'samples': 3628032, 'steps': 18895, 'loss/train': 0.6581803858280182} 01/29/2022 13:08:23 - INFO - codeparrot_training - Step 18896: {'lr': 0.0003620958040225081, 'samples': 3628224, 'steps': 18896, 'loss/train': 1.6807993054389954} 01/29/2022 13:08:28 - INFO - codeparrot_training - Step 18897: {'lr': 0.0003620811783363251, 'samples': 3628416, 'steps': 18897, 'loss/train': 1.9344987273216248} 01/29/2022 13:08:32 - INFO - codeparrot_training - Step 18898: {'lr': 0.0003620665521700219, 'samples': 3628608, 'steps': 18898, 'loss/train': 1.6161929368972778} 01/29/2022 13:08:36 - INFO - codeparrot_training - Step 18899: {'lr': 0.00036205192552366124, 'samples': 3628800, 'steps': 18899, 'loss/train': 1.4587308168411255} 01/29/2022 13:08:40 - INFO - codeparrot_training - Step 18900: {'lr': 0.0003620372983973057, 'samples': 3628992, 'steps': 18900, 'loss/train': 1.999110996723175} 01/29/2022 13:08:46 - INFO - codeparrot_training - Step 18901: {'lr': 0.00036202267079101793, 'samples': 3629184, 'steps': 18901, 'loss/train': 1.646933376789093} 01/29/2022 13:08:50 - INFO - codeparrot_training - Step 18902: {'lr': 0.0003620080427048605, 'samples': 3629376, 'steps': 18902, 'loss/train': 1.8508039712905884} 01/29/2022 13:08:54 - INFO - codeparrot_training - Step 18903: {'lr': 0.00036199341413889637, 'samples': 3629568, 'steps': 18903, 'loss/train': 0.9350711703300476} 01/29/2022 13:08:58 - INFO - codeparrot_training - Step 18904: {'lr': 0.00036197878509318794, 'samples': 3629760, 'steps': 18904, 'loss/train': 1.657585322856903} 01/29/2022 13:09:03 - INFO - codeparrot_training - Step 18905: {'lr': 0.00036196415556779795, 'samples': 3629952, 'steps': 18905, 'loss/train': 1.4584467709064484} 01/29/2022 13:09:10 - INFO - codeparrot_training - Step 18906: {'lr': 0.00036194952556278906, 'samples': 3630144, 'steps': 18906, 'loss/train': 1.650327980518341} 01/29/2022 13:09:14 - INFO - codeparrot_training - Step 18907: {'lr': 0.00036193489507822395, 'samples': 3630336, 'steps': 18907, 'loss/train': 1.4402938485145569} 01/29/2022 13:09:18 - INFO - codeparrot_training - Step 18908: {'lr': 0.0003619202641141652, 'samples': 3630528, 'steps': 18908, 'loss/train': 2.3435633182525635} 01/29/2022 13:09:22 - INFO - codeparrot_training - Step 18909: {'lr': 0.00036190563267067576, 'samples': 3630720, 'steps': 18909, 'loss/train': 1.9682487845420837} 01/29/2022 13:09:27 - INFO - codeparrot_training - Step 18910: {'lr': 0.000361891000747818, 'samples': 3630912, 'steps': 18910, 'loss/train': 0.8943352997303009} 01/29/2022 13:09:32 - INFO - codeparrot_training - Step 18911: {'lr': 0.0003618763683456548, 'samples': 3631104, 'steps': 18911, 'loss/train': 1.061264991760254} 01/29/2022 13:09:36 - INFO - codeparrot_training - Step 18912: {'lr': 0.00036186173546424874, 'samples': 3631296, 'steps': 18912, 'loss/train': 1.490627646446228} 01/29/2022 13:09:40 - INFO - codeparrot_training - Step 18913: {'lr': 0.00036184710210366243, 'samples': 3631488, 'steps': 18913, 'loss/train': 1.6474583745002747} 01/29/2022 13:09:45 - INFO - codeparrot_training - Step 18914: {'lr': 0.0003618324682639588, 'samples': 3631680, 'steps': 18914, 'loss/train': 1.6868825554847717} 01/29/2022 13:09:49 - INFO - codeparrot_training - Step 18915: {'lr': 0.0003618178339452003, 'samples': 3631872, 'steps': 18915, 'loss/train': 1.2939606606960297} 01/29/2022 13:09:54 - INFO - codeparrot_training - Step 18916: {'lr': 0.00036180319914744973, 'samples': 3632064, 'steps': 18916, 'loss/train': 1.6147436499595642} 01/29/2022 13:09:59 - INFO - codeparrot_training - Step 18917: {'lr': 0.00036178856387076967, 'samples': 3632256, 'steps': 18917, 'loss/train': 1.2200028598308563} 01/29/2022 13:10:03 - INFO - codeparrot_training - Step 18918: {'lr': 0.00036177392811522304, 'samples': 3632448, 'steps': 18918, 'loss/train': 1.7024561762809753} 01/29/2022 13:10:07 - INFO - codeparrot_training - Step 18919: {'lr': 0.0003617592918808723, 'samples': 3632640, 'steps': 18919, 'loss/train': 1.7989835143089294} 01/29/2022 13:10:11 - INFO - codeparrot_training - Step 18920: {'lr': 0.0003617446551677803, 'samples': 3632832, 'steps': 18920, 'loss/train': 1.4121868014335632} 01/29/2022 13:10:20 - INFO - codeparrot_training - Step 18921: {'lr': 0.0003617300179760096, 'samples': 3633024, 'steps': 18921, 'loss/train': 1.6601407527923584} 01/29/2022 13:10:24 - INFO - codeparrot_training - Step 18922: {'lr': 0.0003617153803056231, 'samples': 3633216, 'steps': 18922, 'loss/train': 2.355049788951874} 01/29/2022 13:10:28 - INFO - codeparrot_training - Step 18923: {'lr': 0.0003617007421566833, 'samples': 3633408, 'steps': 18923, 'loss/train': 1.9899158477783203} 01/29/2022 13:10:32 - INFO - codeparrot_training - Step 18924: {'lr': 0.00036168610352925294, 'samples': 3633600, 'steps': 18924, 'loss/train': 1.7067201733589172} 01/29/2022 13:10:37 - INFO - codeparrot_training - Step 18925: {'lr': 0.00036167146442339486, 'samples': 3633792, 'steps': 18925, 'loss/train': 1.5584515929222107} 01/29/2022 13:10:42 - INFO - codeparrot_training - Step 18926: {'lr': 0.00036165682483917165, 'samples': 3633984, 'steps': 18926, 'loss/train': 1.7382397055625916} 01/29/2022 13:10:46 - INFO - codeparrot_training - Step 18927: {'lr': 0.0003616421847766461, 'samples': 3634176, 'steps': 18927, 'loss/train': 1.1458315551280975} 01/29/2022 13:10:50 - INFO - codeparrot_training - Step 18928: {'lr': 0.00036162754423588086, 'samples': 3634368, 'steps': 18928, 'loss/train': 1.398253619670868} 01/29/2022 13:10:55 - INFO - codeparrot_training - Step 18929: {'lr': 0.0003616129032169387, 'samples': 3634560, 'steps': 18929, 'loss/train': 1.194619596004486} 01/29/2022 13:10:59 - INFO - codeparrot_training - Step 18930: {'lr': 0.00036159826171988223, 'samples': 3634752, 'steps': 18930, 'loss/train': 1.632926881313324} 01/29/2022 13:11:04 - INFO - codeparrot_training - Step 18931: {'lr': 0.00036158361974477434, 'samples': 3634944, 'steps': 18931, 'loss/train': 1.131531000137329} 01/29/2022 13:11:09 - INFO - codeparrot_training - Step 18932: {'lr': 0.0003615689772916776, 'samples': 3635136, 'steps': 18932, 'loss/train': 1.7076283693313599} 01/29/2022 13:11:13 - INFO - codeparrot_training - Step 18933: {'lr': 0.00036155433436065477, 'samples': 3635328, 'steps': 18933, 'loss/train': 1.277990698814392} 01/29/2022 13:11:17 - INFO - codeparrot_training - Step 18934: {'lr': 0.0003615396909517686, 'samples': 3635520, 'steps': 18934, 'loss/train': 0.7242351919412613} 01/29/2022 13:11:21 - INFO - codeparrot_training - Step 18935: {'lr': 0.00036152504706508183, 'samples': 3635712, 'steps': 18935, 'loss/train': 0.19249577075242996} 01/29/2022 13:11:27 - INFO - codeparrot_training - Step 18936: {'lr': 0.00036151040270065716, 'samples': 3635904, 'steps': 18936, 'loss/train': 1.8067061305046082} 01/29/2022 13:11:31 - INFO - codeparrot_training - Step 18937: {'lr': 0.00036149575785855736, 'samples': 3636096, 'steps': 18937, 'loss/train': 1.9897091388702393} 01/29/2022 13:11:35 - INFO - codeparrot_training - Step 18938: {'lr': 0.00036148111253884514, 'samples': 3636288, 'steps': 18938, 'loss/train': 1.9050398468971252} 01/29/2022 13:11:39 - INFO - codeparrot_training - Step 18939: {'lr': 0.00036146646674158317, 'samples': 3636480, 'steps': 18939, 'loss/train': 3.372174382209778} 01/29/2022 13:11:44 - INFO - codeparrot_training - Step 18940: {'lr': 0.00036145182046683433, 'samples': 3636672, 'steps': 18940, 'loss/train': 2.173302412033081} 01/29/2022 13:11:51 - INFO - codeparrot_training - Step 18941: {'lr': 0.0003614371737146612, 'samples': 3636864, 'steps': 18941, 'loss/train': 1.9686534404754639} 01/29/2022 13:11:55 - INFO - codeparrot_training - Step 18942: {'lr': 0.0003614225264851266, 'samples': 3637056, 'steps': 18942, 'loss/train': 1.54408460855484} 01/29/2022 13:12:00 - INFO - codeparrot_training - Step 18943: {'lr': 0.00036140787877829336, 'samples': 3637248, 'steps': 18943, 'loss/train': 2.05509352684021} 01/29/2022 13:12:04 - INFO - codeparrot_training - Step 18944: {'lr': 0.0003613932305942241, 'samples': 3637440, 'steps': 18944, 'loss/train': 1.9332972764968872} 01/29/2022 13:12:08 - INFO - codeparrot_training - Step 18945: {'lr': 0.0003613785819329816, 'samples': 3637632, 'steps': 18945, 'loss/train': 2.2312493920326233} 01/29/2022 13:12:14 - INFO - codeparrot_training - Step 18946: {'lr': 0.00036136393279462866, 'samples': 3637824, 'steps': 18946, 'loss/train': 1.5807301998138428} 01/29/2022 13:12:18 - INFO - codeparrot_training - Step 18947: {'lr': 0.00036134928317922796, 'samples': 3638016, 'steps': 18947, 'loss/train': 1.1976524591445923} 01/29/2022 13:12:22 - INFO - codeparrot_training - Step 18948: {'lr': 0.0003613346330868423, 'samples': 3638208, 'steps': 18948, 'loss/train': 2.7916789054870605} 01/29/2022 13:12:26 - INFO - codeparrot_training - Step 18949: {'lr': 0.0003613199825175344, 'samples': 3638400, 'steps': 18949, 'loss/train': 2.0958303809165955} 01/29/2022 13:12:31 - INFO - codeparrot_training - Step 18950: {'lr': 0.0003613053314713671, 'samples': 3638592, 'steps': 18950, 'loss/train': 1.4847599565982819} 01/29/2022 13:12:38 - INFO - codeparrot_training - Step 18951: {'lr': 0.00036129067994840303, 'samples': 3638784, 'steps': 18951, 'loss/train': 1.6513359546661377} 01/29/2022 13:12:42 - INFO - codeparrot_training - Step 18952: {'lr': 0.0003612760279487051, 'samples': 3638976, 'steps': 18952, 'loss/train': 2.0124279856681824} 01/29/2022 13:12:46 - INFO - codeparrot_training - Step 18953: {'lr': 0.00036126137547233593, 'samples': 3639168, 'steps': 18953, 'loss/train': 1.404537320137024} 01/29/2022 13:12:51 - INFO - codeparrot_training - Step 18954: {'lr': 0.00036124672251935843, 'samples': 3639360, 'steps': 18954, 'loss/train': 1.3548588752746582} 01/29/2022 13:12:55 - INFO - codeparrot_training - Step 18955: {'lr': 0.00036123206908983527, 'samples': 3639552, 'steps': 18955, 'loss/train': 2.085644245147705} 01/29/2022 13:13:00 - INFO - codeparrot_training - Step 18956: {'lr': 0.00036121741518382915, 'samples': 3639744, 'steps': 18956, 'loss/train': 1.9675909280776978} 01/29/2022 13:13:04 - INFO - codeparrot_training - Step 18957: {'lr': 0.0003612027608014031, 'samples': 3639936, 'steps': 18957, 'loss/train': 1.8563477396965027} 01/29/2022 13:13:09 - INFO - codeparrot_training - Step 18958: {'lr': 0.0003611881059426196, 'samples': 3640128, 'steps': 18958, 'loss/train': 0.9168482422828674} 01/29/2022 13:13:13 - INFO - codeparrot_training - Step 18959: {'lr': 0.00036117345060754163, 'samples': 3640320, 'steps': 18959, 'loss/train': 1.6638441681861877} 01/29/2022 13:13:17 - INFO - codeparrot_training - Step 18960: {'lr': 0.00036115879479623185, 'samples': 3640512, 'steps': 18960, 'loss/train': 1.9477078914642334} 01/29/2022 13:13:22 - INFO - codeparrot_training - Step 18961: {'lr': 0.00036114413850875317, 'samples': 3640704, 'steps': 18961, 'loss/train': 1.7387055158615112} 01/29/2022 13:13:27 - INFO - codeparrot_training - Step 18962: {'lr': 0.00036112948174516824, 'samples': 3640896, 'steps': 18962, 'loss/train': 1.6721979975700378} 01/29/2022 13:13:31 - INFO - codeparrot_training - Step 18963: {'lr': 0.0003611148245055399, 'samples': 3641088, 'steps': 18963, 'loss/train': 1.6837783455848694} 01/29/2022 13:13:35 - INFO - codeparrot_training - Step 18964: {'lr': 0.000361100166789931, 'samples': 3641280, 'steps': 18964, 'loss/train': 1.3598139882087708} 01/29/2022 13:13:40 - INFO - codeparrot_training - Step 18965: {'lr': 0.0003610855085984041, 'samples': 3641472, 'steps': 18965, 'loss/train': 2.1683927178382874} 01/29/2022 13:13:47 - INFO - codeparrot_training - Step 18966: {'lr': 0.0003610708499310223, 'samples': 3641664, 'steps': 18966, 'loss/train': 1.3279868066310883} 01/29/2022 13:13:51 - INFO - codeparrot_training - Step 18967: {'lr': 0.0003610561907878482, 'samples': 3641856, 'steps': 18967, 'loss/train': 1.875109612941742} 01/29/2022 13:13:55 - INFO - codeparrot_training - Step 18968: {'lr': 0.00036104153116894465, 'samples': 3642048, 'steps': 18968, 'loss/train': 1.318249672651291} 01/29/2022 13:14:00 - INFO - codeparrot_training - Step 18969: {'lr': 0.00036102687107437444, 'samples': 3642240, 'steps': 18969, 'loss/train': 7.6681787967681885} 01/29/2022 13:14:04 - INFO - codeparrot_training - Step 18970: {'lr': 0.0003610122105042004, 'samples': 3642432, 'steps': 18970, 'loss/train': 1.6843736171722412} 01/29/2022 13:14:08 - INFO - codeparrot_training - Step 18971: {'lr': 0.0003609975494584852, 'samples': 3642624, 'steps': 18971, 'loss/train': 1.6110672354698181} 01/29/2022 13:14:14 - INFO - codeparrot_training - Step 18972: {'lr': 0.0003609828879372918, 'samples': 3642816, 'steps': 18972, 'loss/train': 1.7926071882247925} 01/29/2022 13:14:18 - INFO - codeparrot_training - Step 18973: {'lr': 0.000360968225940683, 'samples': 3643008, 'steps': 18973, 'loss/train': 0.8176266252994537} 01/29/2022 13:14:22 - INFO - codeparrot_training - Step 18974: {'lr': 0.00036095356346872156, 'samples': 3643200, 'steps': 18974, 'loss/train': 1.7165510058403015} 01/29/2022 13:14:26 - INFO - codeparrot_training - Step 18975: {'lr': 0.0003609389005214702, 'samples': 3643392, 'steps': 18975, 'loss/train': 1.739109456539154} 01/29/2022 13:14:34 - INFO - codeparrot_training - Step 18976: {'lr': 0.0003609242370989919, 'samples': 3643584, 'steps': 18976, 'loss/train': 2.2313575744628906} 01/29/2022 13:14:38 - INFO - codeparrot_training - Step 18977: {'lr': 0.0003609095732013493, 'samples': 3643776, 'steps': 18977, 'loss/train': 1.7967894673347473} 01/29/2022 13:14:42 - INFO - codeparrot_training - Step 18978: {'lr': 0.0003608949088286054, 'samples': 3643968, 'steps': 18978, 'loss/train': 3.888216733932495} 01/29/2022 13:14:46 - INFO - codeparrot_training - Step 18979: {'lr': 0.00036088024398082285, 'samples': 3644160, 'steps': 18979, 'loss/train': 1.9993354082107544} 01/29/2022 13:14:51 - INFO - codeparrot_training - Step 18980: {'lr': 0.00036086557865806464, 'samples': 3644352, 'steps': 18980, 'loss/train': 1.3970926702022552} 01/29/2022 13:14:55 - INFO - codeparrot_training - Step 18981: {'lr': 0.0003608509128603934, 'samples': 3644544, 'steps': 18981, 'loss/train': 1.8184139728546143} 01/29/2022 13:15:00 - INFO - codeparrot_training - Step 18982: {'lr': 0.0003608362465878721, 'samples': 3644736, 'steps': 18982, 'loss/train': 1.8919764161109924} 01/29/2022 13:15:05 - INFO - codeparrot_training - Step 18983: {'lr': 0.0003608215798405635, 'samples': 3644928, 'steps': 18983, 'loss/train': 2.7011629343032837} 01/29/2022 13:15:09 - INFO - codeparrot_training - Step 18984: {'lr': 0.0003608069126185304, 'samples': 3645120, 'steps': 18984, 'loss/train': 1.7924028038978577} 01/29/2022 13:15:13 - INFO - codeparrot_training - Step 18985: {'lr': 0.0003607922449218357, 'samples': 3645312, 'steps': 18985, 'loss/train': 1.6772119402885437} 01/29/2022 13:15:18 - INFO - codeparrot_training - Step 18986: {'lr': 0.00036077757675054224, 'samples': 3645504, 'steps': 18986, 'loss/train': 1.389106661081314} 01/29/2022 13:15:23 - INFO - codeparrot_training - Step 18987: {'lr': 0.0003607629081047128, 'samples': 3645696, 'steps': 18987, 'loss/train': 1.326643019914627} 01/29/2022 13:15:27 - INFO - codeparrot_training - Step 18988: {'lr': 0.0003607482389844102, 'samples': 3645888, 'steps': 18988, 'loss/train': 3.8683419227600098} 01/29/2022 13:15:31 - INFO - codeparrot_training - Step 18989: {'lr': 0.0003607335693896973, 'samples': 3646080, 'steps': 18989, 'loss/train': 1.904746949672699} 01/29/2022 13:15:35 - INFO - codeparrot_training - Step 18990: {'lr': 0.00036071889932063706, 'samples': 3646272, 'steps': 18990, 'loss/train': 2.417538285255432} 01/29/2022 13:15:43 - INFO - codeparrot_training - Step 18991: {'lr': 0.0003607042287772921, 'samples': 3646464, 'steps': 18991, 'loss/train': 1.2414553463459015} 01/29/2022 13:15:47 - INFO - codeparrot_training - Step 18992: {'lr': 0.0003606895577597254, 'samples': 3646656, 'steps': 18992, 'loss/train': 2.070372998714447} 01/29/2022 13:15:52 - INFO - codeparrot_training - Step 18993: {'lr': 0.0003606748862679998, 'samples': 3646848, 'steps': 18993, 'loss/train': 1.515671968460083} 01/29/2022 13:15:56 - INFO - codeparrot_training - Step 18994: {'lr': 0.00036066021430217817, 'samples': 3647040, 'steps': 18994, 'loss/train': 1.5182937383651733} 01/29/2022 13:16:00 - INFO - codeparrot_training - Step 18995: {'lr': 0.00036064554186232323, 'samples': 3647232, 'steps': 18995, 'loss/train': 0.7784636914730072} 01/29/2022 13:16:06 - INFO - codeparrot_training - Step 18996: {'lr': 0.0003606308689484979, 'samples': 3647424, 'steps': 18996, 'loss/train': 0.19328630715608597} 01/29/2022 13:16:10 - INFO - codeparrot_training - Step 18997: {'lr': 0.0003606161955607652, 'samples': 3647616, 'steps': 18997, 'loss/train': 1.6340777277946472} 01/29/2022 13:16:14 - INFO - codeparrot_training - Step 18998: {'lr': 0.0003606015216991877, 'samples': 3647808, 'steps': 18998, 'loss/train': 0.9566640257835388} 01/29/2022 13:16:18 - INFO - codeparrot_training - Step 18999: {'lr': 0.0003605868473638285, 'samples': 3648000, 'steps': 18999, 'loss/train': 1.7524378895759583} 01/29/2022 13:16:23 - INFO - codeparrot_training - Step 19000: {'lr': 0.00036057217255475036, 'samples': 3648192, 'steps': 19000, 'loss/train': 1.7772658467292786} 01/29/2022 13:16:28 - INFO - codeparrot_training - Step 19001: {'lr': 0.0003605574972720161, 'samples': 3648384, 'steps': 19001, 'loss/train': 1.9185935854911804} 01/29/2022 13:16:33 - INFO - codeparrot_training - Step 19002: {'lr': 0.00036054282151568864, 'samples': 3648576, 'steps': 19002, 'loss/train': 1.7660865783691406} 01/29/2022 13:16:37 - INFO - codeparrot_training - Step 19003: {'lr': 0.0003605281452858308, 'samples': 3648768, 'steps': 19003, 'loss/train': 1.7050434350967407} 01/29/2022 13:16:41 - INFO - codeparrot_training - Step 19004: {'lr': 0.00036051346858250556, 'samples': 3648960, 'steps': 19004, 'loss/train': 0.5999943315982819} 01/29/2022 13:16:45 - INFO - codeparrot_training - Step 19005: {'lr': 0.00036049879140577566, 'samples': 3649152, 'steps': 19005, 'loss/train': 0.04460104275494814} 01/29/2022 13:16:51 - INFO - codeparrot_training - Step 19006: {'lr': 0.000360484113755704, 'samples': 3649344, 'steps': 19006, 'loss/train': 0.45505383610725403} 01/29/2022 13:16:55 - INFO - codeparrot_training - Step 19007: {'lr': 0.0003604694356323536, 'samples': 3649536, 'steps': 19007, 'loss/train': 1.7480045557022095} 01/29/2022 13:16:59 - INFO - codeparrot_training - Step 19008: {'lr': 0.00036045475703578705, 'samples': 3649728, 'steps': 19008, 'loss/train': 1.8235908150672913} 01/29/2022 13:17:03 - INFO - codeparrot_training - Step 19009: {'lr': 0.0003604400779660674, 'samples': 3649920, 'steps': 19009, 'loss/train': 1.863411784172058} 01/29/2022 13:17:08 - INFO - codeparrot_training - Step 19010: {'lr': 0.00036042539842325765, 'samples': 3650112, 'steps': 19010, 'loss/train': 0.5287927687168121} 01/29/2022 13:17:15 - INFO - codeparrot_training - Step 19011: {'lr': 0.0003604107184074205, 'samples': 3650304, 'steps': 19011, 'loss/train': 1.3359088897705078} 01/29/2022 13:17:19 - INFO - codeparrot_training - Step 19012: {'lr': 0.0003603960379186189, 'samples': 3650496, 'steps': 19012, 'loss/train': 2.374765634536743} 01/29/2022 13:17:24 - INFO - codeparrot_training - Step 19013: {'lr': 0.0003603813569569157, 'samples': 3650688, 'steps': 19013, 'loss/train': 2.0061762928962708} 01/29/2022 13:17:28 - INFO - codeparrot_training - Step 19014: {'lr': 0.0003603666755223739, 'samples': 3650880, 'steps': 19014, 'loss/train': 1.0023930072784424} 01/29/2022 13:17:32 - INFO - codeparrot_training - Step 19015: {'lr': 0.0003603519936150562, 'samples': 3651072, 'steps': 19015, 'loss/train': 1.6430287957191467} 01/29/2022 13:17:37 - INFO - codeparrot_training - Step 19016: {'lr': 0.00036033731123502567, 'samples': 3651264, 'steps': 19016, 'loss/train': 0.9431582093238831} 01/29/2022 13:17:42 - INFO - codeparrot_training - Step 19017: {'lr': 0.00036032262838234507, 'samples': 3651456, 'steps': 19017, 'loss/train': 1.3721626996994019} 01/29/2022 13:17:46 - INFO - codeparrot_training - Step 19018: {'lr': 0.0003603079450570775, 'samples': 3651648, 'steps': 19018, 'loss/train': 0.31214739382267} 01/29/2022 13:17:50 - INFO - codeparrot_training - Step 19019: {'lr': 0.00036029326125928556, 'samples': 3651840, 'steps': 19019, 'loss/train': 1.341545730829239} 01/29/2022 13:17:55 - INFO - codeparrot_training - Step 19020: {'lr': 0.00036027857698903235, 'samples': 3652032, 'steps': 19020, 'loss/train': 2.2257930636405945} 01/29/2022 13:18:02 - INFO - codeparrot_training - Step 19021: {'lr': 0.00036026389224638077, 'samples': 3652224, 'steps': 19021, 'loss/train': 2.424652397632599} 01/29/2022 13:18:06 - INFO - codeparrot_training - Step 19022: {'lr': 0.00036024920703139375, 'samples': 3652416, 'steps': 19022, 'loss/train': 1.3961570262908936} 01/29/2022 13:18:10 - INFO - codeparrot_training - Step 19023: {'lr': 0.00036023452134413406, 'samples': 3652608, 'steps': 19023, 'loss/train': 1.871220052242279} 01/29/2022 13:18:14 - INFO - codeparrot_training - Step 19024: {'lr': 0.00036021983518466465, 'samples': 3652800, 'steps': 19024, 'loss/train': 1.3462152779102325} 01/29/2022 13:18:19 - INFO - codeparrot_training - Step 19025: {'lr': 0.00036020514855304855, 'samples': 3652992, 'steps': 19025, 'loss/train': 1.6023804545402527} 01/29/2022 13:18:24 - INFO - codeparrot_training - Step 19026: {'lr': 0.00036019046144934855, 'samples': 3653184, 'steps': 19026, 'loss/train': 1.9061381220817566} 01/29/2022 13:18:28 - INFO - codeparrot_training - Step 19027: {'lr': 0.0003601757738736275, 'samples': 3653376, 'steps': 19027, 'loss/train': 1.425366371870041} 01/29/2022 13:18:32 - INFO - codeparrot_training - Step 19028: {'lr': 0.0003601610858259485, 'samples': 3653568, 'steps': 19028, 'loss/train': 2.2720736861228943} 01/29/2022 13:18:37 - INFO - codeparrot_training - Step 19029: {'lr': 0.0003601463973063745, 'samples': 3653760, 'steps': 19029, 'loss/train': 1.8868734240531921} 01/29/2022 13:18:41 - INFO - codeparrot_training - Step 19030: {'lr': 0.0003601317083149682, 'samples': 3653952, 'steps': 19030, 'loss/train': 1.562028408050537} 01/29/2022 13:18:46 - INFO - codeparrot_training - Step 19031: {'lr': 0.0003601170188517926, 'samples': 3654144, 'steps': 19031, 'loss/train': 1.4520439803600311} 01/29/2022 13:18:50 - INFO - codeparrot_training - Step 19032: {'lr': 0.0003601023289169107, 'samples': 3654336, 'steps': 19032, 'loss/train': 1.1619180142879486} 01/29/2022 13:18:55 - INFO - codeparrot_training - Step 19033: {'lr': 0.00036008763851038534, 'samples': 3654528, 'steps': 19033, 'loss/train': 2.3290518522262573} 01/29/2022 13:18:59 - INFO - codeparrot_training - Step 19034: {'lr': 0.0003600729476322796, 'samples': 3654720, 'steps': 19034, 'loss/train': 1.6529611945152283} 01/29/2022 13:19:03 - INFO - codeparrot_training - Step 19035: {'lr': 0.0003600582562826562, 'samples': 3654912, 'steps': 19035, 'loss/train': 2.1910688281059265} 01/29/2022 13:19:10 - INFO - codeparrot_training - Step 19036: {'lr': 0.0003600435644615783, 'samples': 3655104, 'steps': 19036, 'loss/train': 2.1134272813796997} 01/29/2022 13:19:15 - INFO - codeparrot_training - Step 19037: {'lr': 0.0003600288721691085, 'samples': 3655296, 'steps': 19037, 'loss/train': 1.7941215634346008} 01/29/2022 13:19:19 - INFO - codeparrot_training - Step 19038: {'lr': 0.0003600141794053102, 'samples': 3655488, 'steps': 19038, 'loss/train': 1.5966922044754028} 01/29/2022 13:19:23 - INFO - codeparrot_training - Step 19039: {'lr': 0.00035999948617024594, 'samples': 3655680, 'steps': 19039, 'loss/train': 1.3435332477092743} 01/29/2022 13:19:27 - INFO - codeparrot_training - Step 19040: {'lr': 0.0003599847924639788, 'samples': 3655872, 'steps': 19040, 'loss/train': 1.2173956632614136} 01/29/2022 13:19:33 - INFO - codeparrot_training - Step 19041: {'lr': 0.0003599700982865718, 'samples': 3656064, 'steps': 19041, 'loss/train': 1.3257758617401123} 01/29/2022 13:19:37 - INFO - codeparrot_training - Step 19042: {'lr': 0.00035995540363808776, 'samples': 3656256, 'steps': 19042, 'loss/train': 2.762475550174713} 01/29/2022 13:19:41 - INFO - codeparrot_training - Step 19043: {'lr': 0.0003599407085185897, 'samples': 3656448, 'steps': 19043, 'loss/train': 1.3533879518508911} 01/29/2022 13:19:46 - INFO - codeparrot_training - Step 19044: {'lr': 0.00035992601292814065, 'samples': 3656640, 'steps': 19044, 'loss/train': 1.8102060556411743} 01/29/2022 13:19:50 - INFO - codeparrot_training - Step 19045: {'lr': 0.00035991131686680333, 'samples': 3656832, 'steps': 19045, 'loss/train': 0.3686137944459915} 01/29/2022 13:19:56 - INFO - codeparrot_training - Step 19046: {'lr': 0.00035989662033464087, 'samples': 3657024, 'steps': 19046, 'loss/train': 0.7070975303649902} 01/29/2022 13:20:00 - INFO - codeparrot_training - Step 19047: {'lr': 0.0003598819233317162, 'samples': 3657216, 'steps': 19047, 'loss/train': 1.1997852623462677} 01/29/2022 13:20:04 - INFO - codeparrot_training - Step 19048: {'lr': 0.0003598672258580923, 'samples': 3657408, 'steps': 19048, 'loss/train': 2.3307997584342957} 01/29/2022 13:20:09 - INFO - codeparrot_training - Step 19049: {'lr': 0.00035985252791383203, 'samples': 3657600, 'steps': 19049, 'loss/train': 1.299210548400879} 01/29/2022 13:20:13 - INFO - codeparrot_training - Step 19050: {'lr': 0.0003598378294989984, 'samples': 3657792, 'steps': 19050, 'loss/train': 0.9360286295413971} 01/29/2022 13:20:17 - INFO - codeparrot_training - Step 19051: {'lr': 0.00035982313061365444, 'samples': 3657984, 'steps': 19051, 'loss/train': 2.406458079814911} 01/29/2022 13:20:22 - INFO - codeparrot_training - Step 19052: {'lr': 0.00035980843125786306, 'samples': 3658176, 'steps': 19052, 'loss/train': 1.543322503566742} 01/29/2022 13:20:27 - INFO - codeparrot_training - Step 19053: {'lr': 0.00035979373143168726, 'samples': 3658368, 'steps': 19053, 'loss/train': 1.689965844154358} 01/29/2022 13:20:31 - INFO - codeparrot_training - Step 19054: {'lr': 0.0003597790311351898, 'samples': 3658560, 'steps': 19054, 'loss/train': 1.5114546418190002} 01/29/2022 13:20:35 - INFO - codeparrot_training - Step 19055: {'lr': 0.00035976433036843405, 'samples': 3658752, 'steps': 19055, 'loss/train': 1.4845130145549774} 01/29/2022 13:20:40 - INFO - codeparrot_training - Step 19056: {'lr': 0.00035974962913148267, 'samples': 3658944, 'steps': 19056, 'loss/train': 2.017932415008545} 01/29/2022 13:20:47 - INFO - codeparrot_training - Step 19057: {'lr': 0.0003597349274243987, 'samples': 3659136, 'steps': 19057, 'loss/train': 2.070568263530731} 01/29/2022 13:20:51 - INFO - codeparrot_training - Step 19058: {'lr': 0.0003597202252472452, 'samples': 3659328, 'steps': 19058, 'loss/train': 1.6614925861358643} 01/29/2022 13:20:55 - INFO - codeparrot_training - Step 19059: {'lr': 0.00035970552260008505, 'samples': 3659520, 'steps': 19059, 'loss/train': 2.5152822732925415} 01/29/2022 13:20:59 - INFO - codeparrot_training - Step 19060: {'lr': 0.00035969081948298133, 'samples': 3659712, 'steps': 19060, 'loss/train': 3.4131839275360107} 01/29/2022 13:21:04 - INFO - codeparrot_training - Step 19061: {'lr': 0.0003596761158959969, 'samples': 3659904, 'steps': 19061, 'loss/train': 1.9495636224746704} 01/29/2022 13:21:10 - INFO - codeparrot_training - Step 19062: {'lr': 0.00035966141183919485, 'samples': 3660096, 'steps': 19062, 'loss/train': 1.7530794739723206} 01/29/2022 13:21:14 - INFO - codeparrot_training - Step 19063: {'lr': 0.0003596467073126382, 'samples': 3660288, 'steps': 19063, 'loss/train': 1.9410991072654724} 01/29/2022 13:21:18 - INFO - codeparrot_training - Step 19064: {'lr': 0.0003596320023163898, 'samples': 3660480, 'steps': 19064, 'loss/train': 2.406691074371338} 01/29/2022 13:21:22 - INFO - codeparrot_training - Step 19065: {'lr': 0.0003596172968505127, 'samples': 3660672, 'steps': 19065, 'loss/train': 1.0030098259449005} 01/29/2022 13:21:27 - INFO - codeparrot_training - Step 19066: {'lr': 0.0003596025909150699, 'samples': 3660864, 'steps': 19066, 'loss/train': 7.1089768409729} 01/29/2022 13:21:34 - INFO - codeparrot_training - Step 19067: {'lr': 0.00035958788451012446, 'samples': 3661056, 'steps': 19067, 'loss/train': 1.9420375227928162} 01/29/2022 13:21:38 - INFO - codeparrot_training - Step 19068: {'lr': 0.0003595731776357393, 'samples': 3661248, 'steps': 19068, 'loss/train': 0.5874467343091965} 01/29/2022 13:21:42 - INFO - codeparrot_training - Step 19069: {'lr': 0.00035955847029197746, 'samples': 3661440, 'steps': 19069, 'loss/train': 0.8998392820358276} 01/29/2022 13:21:46 - INFO - codeparrot_training - Step 19070: {'lr': 0.0003595437624789019, 'samples': 3661632, 'steps': 19070, 'loss/train': 0.7388012856245041} 01/29/2022 13:21:51 - INFO - codeparrot_training - Step 19071: {'lr': 0.00035952905419657565, 'samples': 3661824, 'steps': 19071, 'loss/train': 1.1948910355567932} 01/29/2022 13:21:56 - INFO - codeparrot_training - Step 19072: {'lr': 0.0003595143454450617, 'samples': 3662016, 'steps': 19072, 'loss/train': 1.4379514753818512} 01/29/2022 13:22:00 - INFO - codeparrot_training - Step 19073: {'lr': 0.0003594996362244231, 'samples': 3662208, 'steps': 19073, 'loss/train': 1.9028926491737366} 01/29/2022 13:22:04 - INFO - codeparrot_training - Step 19074: {'lr': 0.0003594849265347228, 'samples': 3662400, 'steps': 19074, 'loss/train': 2.7854912281036377} 01/29/2022 13:22:09 - INFO - codeparrot_training - Step 19075: {'lr': 0.00035947021637602384, 'samples': 3662592, 'steps': 19075, 'loss/train': 0.528982624411583} 01/29/2022 13:22:13 - INFO - codeparrot_training - Step 19076: {'lr': 0.0003594555057483892, 'samples': 3662784, 'steps': 19076, 'loss/train': 1.326968640089035} 01/29/2022 13:22:18 - INFO - codeparrot_training - Step 19077: {'lr': 0.00035944079465188194, 'samples': 3662976, 'steps': 19077, 'loss/train': 1.7312564849853516} 01/29/2022 13:22:22 - INFO - codeparrot_training - Step 19078: {'lr': 0.0003594260830865651, 'samples': 3663168, 'steps': 19078, 'loss/train': 1.8594428300857544} 01/29/2022 13:22:27 - INFO - codeparrot_training - Step 19079: {'lr': 0.00035941137105250173, 'samples': 3663360, 'steps': 19079, 'loss/train': 0.6714165508747101} 01/29/2022 13:22:31 - INFO - codeparrot_training - Step 19080: {'lr': 0.00035939665854975466, 'samples': 3663552, 'steps': 19080, 'loss/train': 2.591692864894867} 01/29/2022 13:22:35 - INFO - codeparrot_training - Step 19081: {'lr': 0.0003593819455783871, 'samples': 3663744, 'steps': 19081, 'loss/train': 1.3171046376228333} 01/29/2022 13:22:42 - INFO - codeparrot_training - Step 19082: {'lr': 0.000359367232138462, 'samples': 3663936, 'steps': 19082, 'loss/train': 1.952938199043274} 01/29/2022 13:22:47 - INFO - codeparrot_training - Step 19083: {'lr': 0.00035935251823004244, 'samples': 3664128, 'steps': 19083, 'loss/train': 2.0334028601646423} 01/29/2022 13:22:51 - INFO - codeparrot_training - Step 19084: {'lr': 0.00035933780385319136, 'samples': 3664320, 'steps': 19084, 'loss/train': 0.8537875413894653} 01/29/2022 13:22:55 - INFO - codeparrot_training - Step 19085: {'lr': 0.0003593230890079719, 'samples': 3664512, 'steps': 19085, 'loss/train': 1.6471717357635498} 01/29/2022 13:22:59 - INFO - codeparrot_training - Step 19086: {'lr': 0.0003593083736944471, 'samples': 3664704, 'steps': 19086, 'loss/train': 1.8163753151893616} 01/29/2022 13:23:05 - INFO - codeparrot_training - Step 19087: {'lr': 0.00035929365791267974, 'samples': 3664896, 'steps': 19087, 'loss/train': 1.8281711339950562} 01/29/2022 13:23:09 - INFO - codeparrot_training - Step 19088: {'lr': 0.00035927894166273323, 'samples': 3665088, 'steps': 19088, 'loss/train': 1.4539900124073029} 01/29/2022 13:23:14 - INFO - codeparrot_training - Step 19089: {'lr': 0.00035926422494467035, 'samples': 3665280, 'steps': 19089, 'loss/train': 1.9387802481651306} 01/29/2022 13:23:18 - INFO - codeparrot_training - Step 19090: {'lr': 0.0003592495077585543, 'samples': 3665472, 'steps': 19090, 'loss/train': 0.8585978150367737} 01/29/2022 13:23:22 - INFO - codeparrot_training - Step 19091: {'lr': 0.000359234790104448, 'samples': 3665664, 'steps': 19091, 'loss/train': 1.4362089335918427} 01/29/2022 13:23:29 - INFO - codeparrot_training - Step 19092: {'lr': 0.0003592200719824146, 'samples': 3665856, 'steps': 19092, 'loss/train': 1.7634265422821045} 01/29/2022 13:23:33 - INFO - codeparrot_training - Step 19093: {'lr': 0.000359205353392517, 'samples': 3666048, 'steps': 19093, 'loss/train': 1.3872312605381012} 01/29/2022 13:23:38 - INFO - codeparrot_training - Step 19094: {'lr': 0.00035919063433481835, 'samples': 3666240, 'steps': 19094, 'loss/train': 1.4883447289466858} 01/29/2022 13:23:42 - INFO - codeparrot_training - Step 19095: {'lr': 0.0003591759148093818, 'samples': 3666432, 'steps': 19095, 'loss/train': 2.197709619998932} 01/29/2022 13:23:46 - INFO - codeparrot_training - Step 19096: {'lr': 0.0003591611948162702, 'samples': 3666624, 'steps': 19096, 'loss/train': 2.1395931243896484} 01/29/2022 13:23:51 - INFO - codeparrot_training - Step 19097: {'lr': 0.0003591464743555467, 'samples': 3666816, 'steps': 19097, 'loss/train': 1.045850157737732} 01/29/2022 13:23:56 - INFO - codeparrot_training - Step 19098: {'lr': 0.0003591317534272744, 'samples': 3667008, 'steps': 19098, 'loss/train': 0.9835448563098907} 01/29/2022 13:24:00 - INFO - codeparrot_training - Step 19099: {'lr': 0.0003591170320315164, 'samples': 3667200, 'steps': 19099, 'loss/train': 0.7302280962467194} 01/29/2022 13:24:04 - INFO - codeparrot_training - Step 19100: {'lr': 0.0003591023101683355, 'samples': 3667392, 'steps': 19100, 'loss/train': 1.5930300951004028} 01/29/2022 13:24:08 - INFO - codeparrot_training - Step 19101: {'lr': 0.00035908758783779506, 'samples': 3667584, 'steps': 19101, 'loss/train': 1.7965587973594666} 01/29/2022 13:24:14 - INFO - codeparrot_training - Step 19102: {'lr': 0.000359072865039958, 'samples': 3667776, 'steps': 19102, 'loss/train': 1.3352348506450653} 01/29/2022 13:24:18 - INFO - codeparrot_training - Step 19103: {'lr': 0.0003590581417748875, 'samples': 3667968, 'steps': 19103, 'loss/train': 2.5523895621299744} 01/29/2022 13:24:22 - INFO - codeparrot_training - Step 19104: {'lr': 0.0003590434180426465, 'samples': 3668160, 'steps': 19104, 'loss/train': 1.083931803703308} 01/29/2022 13:24:26 - INFO - codeparrot_training - Step 19105: {'lr': 0.00035902869384329803, 'samples': 3668352, 'steps': 19105, 'loss/train': 1.5809701681137085} 01/29/2022 13:24:31 - INFO - codeparrot_training - Step 19106: {'lr': 0.0003590139691769054, 'samples': 3668544, 'steps': 19106, 'loss/train': 1.8378338813781738} 01/29/2022 13:24:38 - INFO - codeparrot_training - Step 19107: {'lr': 0.00035899924404353146, 'samples': 3668736, 'steps': 19107, 'loss/train': 1.856191635131836} 01/29/2022 13:24:42 - INFO - codeparrot_training - Step 19108: {'lr': 0.00035898451844323937, 'samples': 3668928, 'steps': 19108, 'loss/train': 1.5934613943099976} 01/29/2022 13:24:46 - INFO - codeparrot_training - Step 19109: {'lr': 0.0003589697923760923, 'samples': 3669120, 'steps': 19109, 'loss/train': 1.837687075138092} 01/29/2022 13:24:51 - INFO - codeparrot_training - Step 19110: {'lr': 0.00035895506584215307, 'samples': 3669312, 'steps': 19110, 'loss/train': 1.88239985704422} 01/29/2022 13:24:55 - INFO - codeparrot_training - Step 19111: {'lr': 0.0003589403388414851, 'samples': 3669504, 'steps': 19111, 'loss/train': 2.0785083174705505} 01/29/2022 13:25:00 - INFO - codeparrot_training - Step 19112: {'lr': 0.0003589256113741513, 'samples': 3669696, 'steps': 19112, 'loss/train': 1.8513007164001465} 01/29/2022 13:25:05 - INFO - codeparrot_training - Step 19113: {'lr': 0.00035891088344021464, 'samples': 3669888, 'steps': 19113, 'loss/train': 0.8411562144756317} 01/29/2022 13:25:09 - INFO - codeparrot_training - Step 19114: {'lr': 0.00035889615503973847, 'samples': 3670080, 'steps': 19114, 'loss/train': 1.5741167664527893} 01/29/2022 13:25:13 - INFO - codeparrot_training - Step 19115: {'lr': 0.00035888142617278567, 'samples': 3670272, 'steps': 19115, 'loss/train': 0.8620348870754242} 01/29/2022 13:25:17 - INFO - codeparrot_training - Step 19116: {'lr': 0.00035886669683941945, 'samples': 3670464, 'steps': 19116, 'loss/train': 2.7125006318092346} 01/29/2022 13:25:22 - INFO - codeparrot_training - Step 19117: {'lr': 0.00035885196703970286, 'samples': 3670656, 'steps': 19117, 'loss/train': 1.8651981353759766} 01/29/2022 13:25:27 - INFO - codeparrot_training - Step 19118: {'lr': 0.00035883723677369893, 'samples': 3670848, 'steps': 19118, 'loss/train': 1.412830263376236} 01/29/2022 13:25:31 - INFO - codeparrot_training - Step 19119: {'lr': 0.000358822506041471, 'samples': 3671040, 'steps': 19119, 'loss/train': 1.340617150068283} 01/29/2022 13:25:35 - INFO - codeparrot_training - Step 19120: {'lr': 0.0003588077748430819, 'samples': 3671232, 'steps': 19120, 'loss/train': 2.0291082859039307} 01/29/2022 13:25:39 - INFO - codeparrot_training - Step 19121: {'lr': 0.00035879304317859486, 'samples': 3671424, 'steps': 19121, 'loss/train': 1.7855111360549927} 01/29/2022 13:25:45 - INFO - codeparrot_training - Step 19122: {'lr': 0.00035877831104807296, 'samples': 3671616, 'steps': 19122, 'loss/train': 1.4710339307785034} 01/29/2022 13:25:49 - INFO - codeparrot_training - Step 19123: {'lr': 0.00035876357845157934, 'samples': 3671808, 'steps': 19123, 'loss/train': 1.6232927441596985} 01/29/2022 13:25:53 - INFO - codeparrot_training - Step 19124: {'lr': 0.00035874884538917705, 'samples': 3672000, 'steps': 19124, 'loss/train': 0.7494579702615738} 01/29/2022 13:25:58 - INFO - codeparrot_training - Step 19125: {'lr': 0.0003587341118609293, 'samples': 3672192, 'steps': 19125, 'loss/train': 0.9818254709243774} 01/29/2022 13:26:02 - INFO - codeparrot_training - Step 19126: {'lr': 0.00035871937786689914, 'samples': 3672384, 'steps': 19126, 'loss/train': 1.4432436525821686} 01/29/2022 13:26:09 - INFO - codeparrot_training - Step 19127: {'lr': 0.00035870464340714966, 'samples': 3672576, 'steps': 19127, 'loss/train': 1.4813455045223236} 01/29/2022 13:26:13 - INFO - codeparrot_training - Step 19128: {'lr': 0.00035868990848174397, 'samples': 3672768, 'steps': 19128, 'loss/train': 1.6238468885421753} 01/29/2022 13:26:18 - INFO - codeparrot_training - Step 19129: {'lr': 0.00035867517309074527, 'samples': 3672960, 'steps': 19129, 'loss/train': 1.7421156764030457} 01/29/2022 13:26:22 - INFO - codeparrot_training - Step 19130: {'lr': 0.0003586604372342166, 'samples': 3673152, 'steps': 19130, 'loss/train': 1.8182922005653381} 01/29/2022 13:26:27 - INFO - codeparrot_training - Step 19131: {'lr': 0.0003586457009122211, 'samples': 3673344, 'steps': 19131, 'loss/train': 1.927520513534546} 01/29/2022 13:26:31 - INFO - codeparrot_training - Step 19132: {'lr': 0.000358630964124822, 'samples': 3673536, 'steps': 19132, 'loss/train': 2.0738001465797424} 01/29/2022 13:26:36 - INFO - codeparrot_training - Step 19133: {'lr': 0.0003586162268720823, 'samples': 3673728, 'steps': 19133, 'loss/train': 1.6744399666786194} 01/29/2022 13:26:40 - INFO - codeparrot_training - Step 19134: {'lr': 0.00035860148915406513, 'samples': 3673920, 'steps': 19134, 'loss/train': 1.9089596271514893} 01/29/2022 13:26:44 - INFO - codeparrot_training - Step 19135: {'lr': 0.00035858675097083373, 'samples': 3674112, 'steps': 19135, 'loss/train': 1.7071872353553772} 01/29/2022 13:26:50 - INFO - codeparrot_training - Step 19136: {'lr': 0.0003585720123224512, 'samples': 3674304, 'steps': 19136, 'loss/train': 1.9685041308403015} 01/29/2022 13:26:54 - INFO - codeparrot_training - Step 19137: {'lr': 0.0003585572732089806, 'samples': 3674496, 'steps': 19137, 'loss/train': 1.4253321290016174} 01/29/2022 13:26:58 - INFO - codeparrot_training - Step 19138: {'lr': 0.00035854253363048507, 'samples': 3674688, 'steps': 19138, 'loss/train': 0.9194163680076599} 01/29/2022 13:27:02 - INFO - codeparrot_training - Step 19139: {'lr': 0.00035852779358702783, 'samples': 3674880, 'steps': 19139, 'loss/train': 0.6903900504112244} 01/29/2022 13:27:07 - INFO - codeparrot_training - Step 19140: {'lr': 0.00035851305307867197, 'samples': 3675072, 'steps': 19140, 'loss/train': 1.7371063828468323} 01/29/2022 13:27:14 - INFO - codeparrot_training - Step 19141: {'lr': 0.0003584983121054807, 'samples': 3675264, 'steps': 19141, 'loss/train': 7.234664440155029} 01/29/2022 13:27:18 - INFO - codeparrot_training - Step 19142: {'lr': 0.00035848357066751703, 'samples': 3675456, 'steps': 19142, 'loss/train': 1.8997242450714111} 01/29/2022 13:27:22 - INFO - codeparrot_training - Step 19143: {'lr': 0.0003584688287648442, 'samples': 3675648, 'steps': 19143, 'loss/train': 1.1944343447685242} 01/29/2022 13:27:27 - INFO - codeparrot_training - Step 19144: {'lr': 0.00035845408639752544, 'samples': 3675840, 'steps': 19144, 'loss/train': 1.0585733950138092} 01/29/2022 13:27:31 - INFO - codeparrot_training - Step 19145: {'lr': 0.00035843934356562376, 'samples': 3676032, 'steps': 19145, 'loss/train': 0.10427992790937424} 01/29/2022 13:27:36 - INFO - codeparrot_training - Step 19146: {'lr': 0.00035842460026920233, 'samples': 3676224, 'steps': 19146, 'loss/train': 1.384695678949356} 01/29/2022 13:27:41 - INFO - codeparrot_training - Step 19147: {'lr': 0.00035840985650832435, 'samples': 3676416, 'steps': 19147, 'loss/train': 2.2876471281051636} 01/29/2022 13:27:45 - INFO - codeparrot_training - Step 19148: {'lr': 0.000358395112283053, 'samples': 3676608, 'steps': 19148, 'loss/train': 1.7060700058937073} 01/29/2022 13:27:49 - INFO - codeparrot_training - Step 19149: {'lr': 0.00035838036759345144, 'samples': 3676800, 'steps': 19149, 'loss/train': 1.3352810740470886} 01/29/2022 13:27:53 - INFO - codeparrot_training - Step 19150: {'lr': 0.0003583656224395827, 'samples': 3676992, 'steps': 19150, 'loss/train': 1.2977584898471832} 01/29/2022 13:28:00 - INFO - codeparrot_training - Step 19151: {'lr': 0.00035835087682151016, 'samples': 3677184, 'steps': 19151, 'loss/train': 1.2550313472747803} 01/29/2022 13:28:05 - INFO - codeparrot_training - Step 19152: {'lr': 0.00035833613073929684, 'samples': 3677376, 'steps': 19152, 'loss/train': 1.3357459902763367} 01/29/2022 13:28:09 - INFO - codeparrot_training - Step 19153: {'lr': 0.00035832138419300585, 'samples': 3677568, 'steps': 19153, 'loss/train': 1.7927806377410889} 01/29/2022 13:28:13 - INFO - codeparrot_training - Step 19154: {'lr': 0.00035830663718270056, 'samples': 3677760, 'steps': 19154, 'loss/train': 2.1347034573554993} 01/29/2022 13:28:17 - INFO - codeparrot_training - Step 19155: {'lr': 0.00035829188970844397, 'samples': 3677952, 'steps': 19155, 'loss/train': 1.1853492558002472} 01/29/2022 13:28:23 - INFO - codeparrot_training - Step 19156: {'lr': 0.0003582771417702993, 'samples': 3678144, 'steps': 19156, 'loss/train': 0.7113336771726608} 01/29/2022 13:28:27 - INFO - codeparrot_training - Step 19157: {'lr': 0.0003582623933683297, 'samples': 3678336, 'steps': 19157, 'loss/train': 1.6353476643562317} 01/29/2022 13:28:32 - INFO - codeparrot_training - Step 19158: {'lr': 0.0003582476445025985, 'samples': 3678528, 'steps': 19158, 'loss/train': 1.6448442935943604} 01/29/2022 13:28:36 - INFO - codeparrot_training - Step 19159: {'lr': 0.00035823289517316866, 'samples': 3678720, 'steps': 19159, 'loss/train': 1.8056654334068298} 01/29/2022 13:28:40 - INFO - codeparrot_training - Step 19160: {'lr': 0.00035821814538010356, 'samples': 3678912, 'steps': 19160, 'loss/train': 1.707993507385254} 01/29/2022 13:28:44 - INFO - codeparrot_training - Step 19161: {'lr': 0.00035820339512346614, 'samples': 3679104, 'steps': 19161, 'loss/train': 1.0773520767688751} 01/29/2022 13:28:50 - INFO - codeparrot_training - Step 19162: {'lr': 0.0003581886444033199, 'samples': 3679296, 'steps': 19162, 'loss/train': 1.695008397102356} 01/29/2022 13:28:54 - INFO - codeparrot_training - Step 19163: {'lr': 0.00035817389321972777, 'samples': 3679488, 'steps': 19163, 'loss/train': 2.294372320175171} 01/29/2022 13:28:58 - INFO - codeparrot_training - Step 19164: {'lr': 0.000358159141572753, 'samples': 3679680, 'steps': 19164, 'loss/train': 3.4453139305114746} 01/29/2022 13:29:02 - INFO - codeparrot_training - Step 19165: {'lr': 0.0003581443894624589, 'samples': 3679872, 'steps': 19165, 'loss/train': 1.7671675086021423} 01/29/2022 13:29:07 - INFO - codeparrot_training - Step 19166: {'lr': 0.0003581296368889085, 'samples': 3680064, 'steps': 19166, 'loss/train': 2.478075385093689} 01/29/2022 13:29:14 - INFO - codeparrot_training - Step 19167: {'lr': 0.0003581148838521651, 'samples': 3680256, 'steps': 19167, 'loss/train': 1.5676581859588623} 01/29/2022 13:29:18 - INFO - codeparrot_training - Step 19168: {'lr': 0.00035810013035229187, 'samples': 3680448, 'steps': 19168, 'loss/train': 0.7548801004886627} 01/29/2022 13:29:22 - INFO - codeparrot_training - Step 19169: {'lr': 0.000358085376389352, 'samples': 3680640, 'steps': 19169, 'loss/train': 1.2793725728988647} 01/29/2022 13:29:27 - INFO - codeparrot_training - Step 19170: {'lr': 0.0003580706219634087, 'samples': 3680832, 'steps': 19170, 'loss/train': 0.9828908443450928} 01/29/2022 13:29:32 - INFO - codeparrot_training - Step 19171: {'lr': 0.0003580558670745252, 'samples': 3681024, 'steps': 19171, 'loss/train': 1.885801076889038} 01/29/2022 13:29:37 - INFO - codeparrot_training - Step 19172: {'lr': 0.00035804111172276464, 'samples': 3681216, 'steps': 19172, 'loss/train': 1.678754210472107} 01/29/2022 13:29:41 - INFO - codeparrot_training - Step 19173: {'lr': 0.00035802635590819035, 'samples': 3681408, 'steps': 19173, 'loss/train': 0.04207469243556261} 01/29/2022 13:29:45 - INFO - codeparrot_training - Step 19174: {'lr': 0.00035801159963086535, 'samples': 3681600, 'steps': 19174, 'loss/train': 1.7900744676589966} 01/29/2022 13:29:49 - INFO - codeparrot_training - Step 19175: {'lr': 0.000357996842890853, 'samples': 3681792, 'steps': 19175, 'loss/train': 1.3701259195804596} 01/29/2022 13:29:55 - INFO - codeparrot_training - Step 19176: {'lr': 0.00035798208568821647, 'samples': 3681984, 'steps': 19176, 'loss/train': 1.092117726802826} 01/29/2022 13:29:59 - INFO - codeparrot_training - Step 19177: {'lr': 0.00035796732802301895, 'samples': 3682176, 'steps': 19177, 'loss/train': 2.2769984006881714} 01/29/2022 13:30:03 - INFO - codeparrot_training - Step 19178: {'lr': 0.00035795256989532367, 'samples': 3682368, 'steps': 19178, 'loss/train': 2.5427345037460327} 01/29/2022 13:30:08 - INFO - codeparrot_training - Step 19179: {'lr': 0.0003579378113051939, 'samples': 3682560, 'steps': 19179, 'loss/train': 1.680856168270111} 01/29/2022 13:30:12 - INFO - codeparrot_training - Step 19180: {'lr': 0.0003579230522526928, 'samples': 3682752, 'steps': 19180, 'loss/train': 1.948284387588501} 01/29/2022 13:30:16 - INFO - codeparrot_training - Step 19181: {'lr': 0.00035790829273788356, 'samples': 3682944, 'steps': 19181, 'loss/train': 1.5862866640090942} 01/29/2022 13:30:24 - INFO - codeparrot_training - Step 19182: {'lr': 0.0003578935327608295, 'samples': 3683136, 'steps': 19182, 'loss/train': 1.7088117599487305} 01/29/2022 13:30:28 - INFO - codeparrot_training - Step 19183: {'lr': 0.00035787877232159384, 'samples': 3683328, 'steps': 19183, 'loss/train': 1.3487604260444641} 01/29/2022 13:30:32 - INFO - codeparrot_training - Step 19184: {'lr': 0.00035786401142023975, 'samples': 3683520, 'steps': 19184, 'loss/train': 1.986409306526184} 01/29/2022 13:30:36 - INFO - codeparrot_training - Step 19185: {'lr': 0.0003578492500568304, 'samples': 3683712, 'steps': 19185, 'loss/train': 1.1798272132873535} 01/29/2022 13:30:42 - INFO - codeparrot_training - Step 19186: {'lr': 0.00035783448823142926, 'samples': 3683904, 'steps': 19186, 'loss/train': 0.8462817370891571} 01/29/2022 13:30:46 - INFO - codeparrot_training - Step 19187: {'lr': 0.00035781972594409937, 'samples': 3684096, 'steps': 19187, 'loss/train': 1.960910439491272} 01/29/2022 13:30:50 - INFO - codeparrot_training - Step 19188: {'lr': 0.000357804963194904, 'samples': 3684288, 'steps': 19188, 'loss/train': 1.2366439998149872} 01/29/2022 13:30:54 - INFO - codeparrot_training - Step 19189: {'lr': 0.00035779019998390636, 'samples': 3684480, 'steps': 19189, 'loss/train': 2.42198646068573} 01/29/2022 13:30:59 - INFO - codeparrot_training - Step 19190: {'lr': 0.00035777543631116977, 'samples': 3684672, 'steps': 19190, 'loss/train': 1.2709892392158508} 01/29/2022 13:31:04 - INFO - codeparrot_training - Step 19191: {'lr': 0.00035776067217675744, 'samples': 3684864, 'steps': 19191, 'loss/train': 1.7666298151016235} 01/29/2022 13:31:08 - INFO - codeparrot_training - Step 19192: {'lr': 0.0003577459075807326, 'samples': 3685056, 'steps': 19192, 'loss/train': 0.9597534835338593} 01/29/2022 13:31:13 - INFO - codeparrot_training - Step 19193: {'lr': 0.00035773114252315844, 'samples': 3685248, 'steps': 19193, 'loss/train': 1.9381586909294128} 01/29/2022 13:31:17 - INFO - codeparrot_training - Step 19194: {'lr': 0.0003577163770040984, 'samples': 3685440, 'steps': 19194, 'loss/train': 1.1506942212581635} 01/29/2022 13:31:21 - INFO - codeparrot_training - Step 19195: {'lr': 0.00035770161102361553, 'samples': 3685632, 'steps': 19195, 'loss/train': 1.7325226664543152} 01/29/2022 13:31:28 - INFO - codeparrot_training - Step 19196: {'lr': 0.0003576868445817732, 'samples': 3685824, 'steps': 19196, 'loss/train': 0.8932693898677826} 01/29/2022 13:31:32 - INFO - codeparrot_training - Step 19197: {'lr': 0.00035767207767863453, 'samples': 3686016, 'steps': 19197, 'loss/train': 1.3711411356925964} 01/29/2022 13:31:36 - INFO - codeparrot_training - Step 19198: {'lr': 0.00035765731031426294, 'samples': 3686208, 'steps': 19198, 'loss/train': 1.0911141335964203} 01/29/2022 13:31:40 - INFO - codeparrot_training - Step 19199: {'lr': 0.0003576425424887216, 'samples': 3686400, 'steps': 19199, 'loss/train': 1.9010708928108215} 01/29/2022 13:31:45 - INFO - codeparrot_training - Step 19200: {'lr': 0.0003576277742020738, 'samples': 3686592, 'steps': 19200, 'loss/train': 0.7320184707641602} 01/29/2022 13:31:50 - INFO - codeparrot_training - Step 19201: {'lr': 0.0003576130054543828, 'samples': 3686784, 'steps': 19201, 'loss/train': 1.8097001910209656} 01/29/2022 13:31:54 - INFO - codeparrot_training - Step 19202: {'lr': 0.00035759823624571184, 'samples': 3686976, 'steps': 19202, 'loss/train': 0.7084873616695404} 01/29/2022 13:31:58 - INFO - codeparrot_training - Step 19203: {'lr': 0.00035758346657612417, 'samples': 3687168, 'steps': 19203, 'loss/train': 2.1385143399238586} 01/29/2022 13:32:03 - INFO - codeparrot_training - Step 19204: {'lr': 0.00035756869644568314, 'samples': 3687360, 'steps': 19204, 'loss/train': 1.0585517585277557} 01/29/2022 13:32:07 - INFO - codeparrot_training - Step 19205: {'lr': 0.00035755392585445195, 'samples': 3687552, 'steps': 19205, 'loss/train': 2.0039656162261963} 01/29/2022 13:32:12 - INFO - codeparrot_training - Step 19206: {'lr': 0.0003575391548024939, 'samples': 3687744, 'steps': 19206, 'loss/train': 1.1226259768009186} 01/29/2022 13:32:17 - INFO - codeparrot_training - Step 19207: {'lr': 0.00035752438328987224, 'samples': 3687936, 'steps': 19207, 'loss/train': 0.8216875791549683} 01/29/2022 13:32:21 - INFO - codeparrot_training - Step 19208: {'lr': 0.00035750961131665034, 'samples': 3688128, 'steps': 19208, 'loss/train': 1.384520173072815} 01/29/2022 13:32:25 - INFO - codeparrot_training - Step 19209: {'lr': 0.0003574948388828913, 'samples': 3688320, 'steps': 19209, 'loss/train': 1.9042561054229736} 01/29/2022 13:32:29 - INFO - codeparrot_training - Step 19210: {'lr': 0.0003574800659886586, 'samples': 3688512, 'steps': 19210, 'loss/train': 2.3454397916793823} 01/29/2022 13:32:36 - INFO - codeparrot_training - Step 19211: {'lr': 0.0003574652926340153, 'samples': 3688704, 'steps': 19211, 'loss/train': 1.020180344581604} 01/29/2022 13:32:40 - INFO - codeparrot_training - Step 19212: {'lr': 0.0003574505188190249, 'samples': 3688896, 'steps': 19212, 'loss/train': 1.8579484820365906} 01/29/2022 13:32:45 - INFO - codeparrot_training - Step 19213: {'lr': 0.0003574357445437506, 'samples': 3689088, 'steps': 19213, 'loss/train': 2.024747908115387} 01/29/2022 13:32:49 - INFO - codeparrot_training - Step 19214: {'lr': 0.0003574209698082556, 'samples': 3689280, 'steps': 19214, 'loss/train': 7.9760448932647705} 01/29/2022 13:32:53 - INFO - codeparrot_training - Step 19215: {'lr': 0.0003574061946126034, 'samples': 3689472, 'steps': 19215, 'loss/train': 1.4452555775642395} 01/29/2022 13:32:58 - INFO - codeparrot_training - Step 19216: {'lr': 0.0003573914189568571, 'samples': 3689664, 'steps': 19216, 'loss/train': 0.5891256630420685} 01/29/2022 13:33:02 - INFO - codeparrot_training - Step 19217: {'lr': 0.0003573766428410801, 'samples': 3689856, 'steps': 19217, 'loss/train': 2.176286816596985} 01/29/2022 13:33:07 - INFO - codeparrot_training - Step 19218: {'lr': 0.00035736186626533556, 'samples': 3690048, 'steps': 19218, 'loss/train': 1.997312843799591} 01/29/2022 13:33:11 - INFO - codeparrot_training - Step 19219: {'lr': 0.00035734708922968695, 'samples': 3690240, 'steps': 19219, 'loss/train': 1.5430026054382324} 01/29/2022 13:33:15 - INFO - codeparrot_training - Step 19220: {'lr': 0.0003573323117341975, 'samples': 3690432, 'steps': 19220, 'loss/train': 1.5133920907974243} 01/29/2022 13:33:22 - INFO - codeparrot_training - Step 19221: {'lr': 0.0003573175337789305, 'samples': 3690624, 'steps': 19221, 'loss/train': 1.2105044424533844} 01/29/2022 13:33:26 - INFO - codeparrot_training - Step 19222: {'lr': 0.0003573027553639492, 'samples': 3690816, 'steps': 19222, 'loss/train': 1.9259485602378845} 01/29/2022 13:33:31 - INFO - codeparrot_training - Step 19223: {'lr': 0.0003572879764893171, 'samples': 3691008, 'steps': 19223, 'loss/train': 1.4545269012451172} 01/29/2022 13:33:35 - INFO - codeparrot_training - Step 19224: {'lr': 0.0003572731971550973, 'samples': 3691200, 'steps': 19224, 'loss/train': 1.1434401869773865} 01/29/2022 13:33:39 - INFO - codeparrot_training - Step 19225: {'lr': 0.0003572584173613532, 'samples': 3691392, 'steps': 19225, 'loss/train': 1.6175792813301086} 01/29/2022 13:33:44 - INFO - codeparrot_training - Step 19226: {'lr': 0.00035724363710814807, 'samples': 3691584, 'steps': 19226, 'loss/train': 2.093824088573456} 01/29/2022 13:33:49 - INFO - codeparrot_training - Step 19227: {'lr': 0.00035722885639554526, 'samples': 3691776, 'steps': 19227, 'loss/train': 0.39717337489128113} 01/29/2022 13:33:53 - INFO - codeparrot_training - Step 19228: {'lr': 0.0003572140752236081, 'samples': 3691968, 'steps': 19228, 'loss/train': 0.5762393921613693} 01/29/2022 13:33:57 - INFO - codeparrot_training - Step 19229: {'lr': 0.0003571992935923999, 'samples': 3692160, 'steps': 19229, 'loss/train': 1.4902958571910858} 01/29/2022 13:34:01 - INFO - codeparrot_training - Step 19230: {'lr': 0.0003571845115019839, 'samples': 3692352, 'steps': 19230, 'loss/train': 1.7547014951705933} 01/29/2022 13:34:07 - INFO - codeparrot_training - Step 19231: {'lr': 0.0003571697289524235, 'samples': 3692544, 'steps': 19231, 'loss/train': 1.673136591911316} 01/29/2022 13:34:11 - INFO - codeparrot_training - Step 19232: {'lr': 0.0003571549459437821, 'samples': 3692736, 'steps': 19232, 'loss/train': 1.1024494171142578} 01/29/2022 13:34:16 - INFO - codeparrot_training - Step 19233: {'lr': 0.0003571401624761229, 'samples': 3692928, 'steps': 19233, 'loss/train': 1.391087919473648} 01/29/2022 13:34:20 - INFO - codeparrot_training - Step 19234: {'lr': 0.0003571253785495092, 'samples': 3693120, 'steps': 19234, 'loss/train': 1.9254305362701416} 01/29/2022 13:34:24 - INFO - codeparrot_training - Step 19235: {'lr': 0.00035711059416400446, 'samples': 3693312, 'steps': 19235, 'loss/train': 1.542740821838379} 01/29/2022 13:34:32 - INFO - codeparrot_training - Step 19236: {'lr': 0.0003570958093196719, 'samples': 3693504, 'steps': 19236, 'loss/train': 1.1490901708602905} 01/29/2022 13:34:36 - INFO - codeparrot_training - Step 19237: {'lr': 0.00035708102401657495, 'samples': 3693696, 'steps': 19237, 'loss/train': 1.7277556657791138} 01/29/2022 13:34:40 - INFO - codeparrot_training - Step 19238: {'lr': 0.00035706623825477687, 'samples': 3693888, 'steps': 19238, 'loss/train': 1.4623018205165863} 01/29/2022 13:34:45 - INFO - codeparrot_training - Step 19239: {'lr': 0.000357051452034341, 'samples': 3694080, 'steps': 19239, 'loss/train': 2.1460978388786316} 01/29/2022 13:34:49 - INFO - codeparrot_training - Step 19240: {'lr': 0.00035703666535533076, 'samples': 3694272, 'steps': 19240, 'loss/train': 1.9008185863494873} 01/29/2022 13:34:54 - INFO - codeparrot_training - Step 19241: {'lr': 0.0003570218782178094, 'samples': 3694464, 'steps': 19241, 'loss/train': 2.4921846985816956} 01/29/2022 13:34:59 - INFO - codeparrot_training - Step 19242: {'lr': 0.0003570070906218403, 'samples': 3694656, 'steps': 19242, 'loss/train': 1.7346253395080566} 01/29/2022 13:35:03 - INFO - codeparrot_training - Step 19243: {'lr': 0.00035699230256748684, 'samples': 3694848, 'steps': 19243, 'loss/train': 1.1288199126720428} 01/29/2022 13:35:07 - INFO - codeparrot_training - Step 19244: {'lr': 0.0003569775140548122, 'samples': 3695040, 'steps': 19244, 'loss/train': 1.4968595802783966} 01/29/2022 13:35:11 - INFO - codeparrot_training - Step 19245: {'lr': 0.00035696272508388, 'samples': 3695232, 'steps': 19245, 'loss/train': 0.6848379671573639} 01/29/2022 13:35:19 - INFO - codeparrot_training - Step 19246: {'lr': 0.00035694793565475337, 'samples': 3695424, 'steps': 19246, 'loss/train': 1.31462824344635} 01/29/2022 13:35:23 - INFO - codeparrot_training - Step 19247: {'lr': 0.0003569331457674958, 'samples': 3695616, 'steps': 19247, 'loss/train': 1.3087421357631683} 01/29/2022 13:35:27 - INFO - codeparrot_training - Step 19248: {'lr': 0.00035691835542217055, 'samples': 3695808, 'steps': 19248, 'loss/train': 1.3146426379680634} 01/29/2022 13:35:31 - INFO - codeparrot_training - Step 19249: {'lr': 0.00035690356461884104, 'samples': 3696000, 'steps': 19249, 'loss/train': 1.66413152217865} 01/29/2022 13:35:36 - INFO - codeparrot_training - Step 19250: {'lr': 0.0003568887733575705, 'samples': 3696192, 'steps': 19250, 'loss/train': 1.9856561422348022} 01/29/2022 13:35:41 - INFO - codeparrot_training - Step 19251: {'lr': 0.0003568739816384225, 'samples': 3696384, 'steps': 19251, 'loss/train': 1.942230999469757} 01/29/2022 13:35:45 - INFO - codeparrot_training - Step 19252: {'lr': 0.00035685918946146036, 'samples': 3696576, 'steps': 19252, 'loss/train': 2.129975974559784} 01/29/2022 13:35:49 - INFO - codeparrot_training - Step 19253: {'lr': 0.00035684439682674723, 'samples': 3696768, 'steps': 19253, 'loss/train': 1.8710902333259583} 01/29/2022 13:35:54 - INFO - codeparrot_training - Step 19254: {'lr': 0.00035682960373434677, 'samples': 3696960, 'steps': 19254, 'loss/train': 1.42221137881279} 01/29/2022 13:35:58 - INFO - codeparrot_training - Step 19255: {'lr': 0.0003568148101843221, 'samples': 3697152, 'steps': 19255, 'loss/train': 1.9691407084465027} 01/29/2022 13:36:03 - INFO - codeparrot_training - Step 19256: {'lr': 0.0003568000161767368, 'samples': 3697344, 'steps': 19256, 'loss/train': 1.9023810625076294} 01/29/2022 13:36:07 - INFO - codeparrot_training - Step 19257: {'lr': 0.00035678522171165406, 'samples': 3697536, 'steps': 19257, 'loss/train': 1.8395074009895325} 01/29/2022 13:36:12 - INFO - codeparrot_training - Step 19258: {'lr': 0.0003567704267891374, 'samples': 3697728, 'steps': 19258, 'loss/train': 1.0057803690433502} 01/29/2022 13:36:16 - INFO - codeparrot_training - Step 19259: {'lr': 0.00035675563140925, 'samples': 3697920, 'steps': 19259, 'loss/train': 1.5698003768920898} 01/29/2022 13:36:20 - INFO - codeparrot_training - Step 19260: {'lr': 0.00035674083557205553, 'samples': 3698112, 'steps': 19260, 'loss/train': 2.31926429271698} 01/29/2022 13:36:28 - INFO - codeparrot_training - Step 19261: {'lr': 0.00035672603927761716, 'samples': 3698304, 'steps': 19261, 'loss/train': 1.6613608002662659} 01/29/2022 13:36:32 - INFO - codeparrot_training - Step 19262: {'lr': 0.0003567112425259984, 'samples': 3698496, 'steps': 19262, 'loss/train': 1.7592909336090088} 01/29/2022 13:36:36 - INFO - codeparrot_training - Step 19263: {'lr': 0.00035669644531726244, 'samples': 3698688, 'steps': 19263, 'loss/train': 0.9310351610183716} 01/29/2022 13:36:40 - INFO - codeparrot_training - Step 19264: {'lr': 0.00035668164765147284, 'samples': 3698880, 'steps': 19264, 'loss/train': 1.6920138001441956} 01/29/2022 13:36:45 - INFO - codeparrot_training - Step 19265: {'lr': 0.00035666684952869295, 'samples': 3699072, 'steps': 19265, 'loss/train': 1.0444784760475159} 01/29/2022 13:36:50 - INFO - codeparrot_training - Step 19266: {'lr': 0.00035665205094898613, 'samples': 3699264, 'steps': 19266, 'loss/train': 1.8601064085960388} 01/29/2022 13:36:54 - INFO - codeparrot_training - Step 19267: {'lr': 0.0003566372519124158, 'samples': 3699456, 'steps': 19267, 'loss/train': 1.8334482908248901} 01/29/2022 13:36:58 - INFO - codeparrot_training - Step 19268: {'lr': 0.00035662245241904533, 'samples': 3699648, 'steps': 19268, 'loss/train': 1.6811134815216064} 01/29/2022 13:37:03 - INFO - codeparrot_training - Step 19269: {'lr': 0.00035660765246893814, 'samples': 3699840, 'steps': 19269, 'loss/train': 1.8557490706443787} 01/29/2022 13:37:07 - INFO - codeparrot_training - Step 19270: {'lr': 0.00035659285206215757, 'samples': 3700032, 'steps': 19270, 'loss/train': 1.9926695823669434} 01/29/2022 13:37:12 - INFO - codeparrot_training - Step 19271: {'lr': 0.0003565780511987672, 'samples': 3700224, 'steps': 19271, 'loss/train': 1.1517523527145386} 01/29/2022 13:37:17 - INFO - codeparrot_training - Step 19272: {'lr': 0.00035656324987883015, 'samples': 3700416, 'steps': 19272, 'loss/train': 1.3806808590888977} 01/29/2022 13:37:21 - INFO - codeparrot_training - Step 19273: {'lr': 0.00035654844810241004, 'samples': 3700608, 'steps': 19273, 'loss/train': 2.1577606201171875} 01/29/2022 13:37:25 - INFO - codeparrot_training - Step 19274: {'lr': 0.0003565336458695702, 'samples': 3700800, 'steps': 19274, 'loss/train': 1.4848219156265259} 01/29/2022 13:37:29 - INFO - codeparrot_training - Step 19275: {'lr': 0.0003565188431803741, 'samples': 3700992, 'steps': 19275, 'loss/train': 1.3708254396915436} 01/29/2022 13:37:37 - INFO - codeparrot_training - Step 19276: {'lr': 0.0003565040400348851, 'samples': 3701184, 'steps': 19276, 'loss/train': 1.5268263816833496} 01/29/2022 13:37:41 - INFO - codeparrot_training - Step 19277: {'lr': 0.0003564892364331665, 'samples': 3701376, 'steps': 19277, 'loss/train': 1.0698886513710022} 01/29/2022 13:37:45 - INFO - codeparrot_training - Step 19278: {'lr': 0.0003564744323752819, 'samples': 3701568, 'steps': 19278, 'loss/train': 1.7077456712722778} 01/29/2022 13:37:50 - INFO - codeparrot_training - Step 19279: {'lr': 0.00035645962786129464, 'samples': 3701760, 'steps': 19279, 'loss/train': 1.61700439453125} 01/29/2022 13:37:54 - INFO - codeparrot_training - Step 19280: {'lr': 0.0003564448228912682, 'samples': 3701952, 'steps': 19280, 'loss/train': 2.370968520641327} 01/29/2022 13:37:59 - INFO - codeparrot_training - Step 19281: {'lr': 0.00035643001746526586, 'samples': 3702144, 'steps': 19281, 'loss/train': 1.7856831550598145} 01/29/2022 13:38:04 - INFO - codeparrot_training - Step 19282: {'lr': 0.0003564152115833511, 'samples': 3702336, 'steps': 19282, 'loss/train': 2.1468159556388855} 01/29/2022 13:38:08 - INFO - codeparrot_training - Step 19283: {'lr': 0.00035640040524558746, 'samples': 3702528, 'steps': 19283, 'loss/train': 2.326727271080017} 01/29/2022 13:38:12 - INFO - codeparrot_training - Step 19284: {'lr': 0.00035638559845203817, 'samples': 3702720, 'steps': 19284, 'loss/train': 1.8577830791473389} 01/29/2022 13:38:16 - INFO - codeparrot_training - Step 19285: {'lr': 0.00035637079120276683, 'samples': 3702912, 'steps': 19285, 'loss/train': 1.3111810684204102} 01/29/2022 13:38:22 - INFO - codeparrot_training - Step 19286: {'lr': 0.00035635598349783676, 'samples': 3703104, 'steps': 19286, 'loss/train': 2.109703838825226} 01/29/2022 13:38:26 - INFO - codeparrot_training - Step 19287: {'lr': 0.0003563411753373115, 'samples': 3703296, 'steps': 19287, 'loss/train': 2.2120776772499084} 01/29/2022 13:38:30 - INFO - codeparrot_training - Step 19288: {'lr': 0.0003563263667212543, 'samples': 3703488, 'steps': 19288, 'loss/train': 0.11800005286931992} 01/29/2022 13:38:35 - INFO - codeparrot_training - Step 19289: {'lr': 0.0003563115576497288, 'samples': 3703680, 'steps': 19289, 'loss/train': 1.860452950000763} 01/29/2022 13:38:39 - INFO - codeparrot_training - Step 19290: {'lr': 0.0003562967481227982, 'samples': 3703872, 'steps': 19290, 'loss/train': 0.8543533086776733} 01/29/2022 13:38:46 - INFO - codeparrot_training - Step 19291: {'lr': 0.0003562819381405262, 'samples': 3704064, 'steps': 19291, 'loss/train': 1.9657502174377441} 01/29/2022 13:38:51 - INFO - codeparrot_training - Step 19292: {'lr': 0.0003562671277029761, 'samples': 3704256, 'steps': 19292, 'loss/train': 1.7908161878585815} 01/29/2022 13:38:55 - INFO - codeparrot_training - Step 19293: {'lr': 0.0003562523168102114, 'samples': 3704448, 'steps': 19293, 'loss/train': 1.2823776304721832} 01/29/2022 13:38:59 - INFO - codeparrot_training - Step 19294: {'lr': 0.00035623750546229547, 'samples': 3704640, 'steps': 19294, 'loss/train': 1.7922567129135132} 01/29/2022 13:39:03 - INFO - codeparrot_training - Step 19295: {'lr': 0.0003562226936592919, 'samples': 3704832, 'steps': 19295, 'loss/train': 1.828393042087555} 01/29/2022 13:39:09 - INFO - codeparrot_training - Step 19296: {'lr': 0.0003562078814012639, 'samples': 3705024, 'steps': 19296, 'loss/train': 1.4118152260780334} 01/29/2022 13:39:13 - INFO - codeparrot_training - Step 19297: {'lr': 0.00035619306868827516, 'samples': 3705216, 'steps': 19297, 'loss/train': 1.311986356973648} 01/29/2022 13:39:17 - INFO - codeparrot_training - Step 19298: {'lr': 0.00035617825552038894, 'samples': 3705408, 'steps': 19298, 'loss/train': 1.1642891764640808} 01/29/2022 13:39:21 - INFO - codeparrot_training - Step 19299: {'lr': 0.00035616344189766885, 'samples': 3705600, 'steps': 19299, 'loss/train': 2.0450152158737183} 01/29/2022 13:39:26 - INFO - codeparrot_training - Step 19300: {'lr': 0.0003561486278201783, 'samples': 3705792, 'steps': 19300, 'loss/train': 1.3505418598651886} 01/29/2022 13:39:31 - INFO - codeparrot_training - Step 19301: {'lr': 0.00035613381328798065, 'samples': 3705984, 'steps': 19301, 'loss/train': 1.696957290172577} 01/29/2022 13:39:35 - INFO - codeparrot_training - Step 19302: {'lr': 0.0003561189983011396, 'samples': 3706176, 'steps': 19302, 'loss/train': 1.3571658432483673} 01/29/2022 13:39:40 - INFO - codeparrot_training - Step 19303: {'lr': 0.00035610418285971835, 'samples': 3706368, 'steps': 19303, 'loss/train': 1.767609715461731} 01/29/2022 13:39:44 - INFO - codeparrot_training - Step 19304: {'lr': 0.00035608936696378046, 'samples': 3706560, 'steps': 19304, 'loss/train': 2.2328937649726868} 01/29/2022 13:39:48 - INFO - codeparrot_training - Step 19305: {'lr': 0.00035607455061338947, 'samples': 3706752, 'steps': 19305, 'loss/train': 2.107507824897766} 01/29/2022 13:39:55 - INFO - codeparrot_training - Step 19306: {'lr': 0.0003560597338086088, 'samples': 3706944, 'steps': 19306, 'loss/train': 2.0700804591178894} 01/29/2022 13:40:00 - INFO - codeparrot_training - Step 19307: {'lr': 0.0003560449165495018, 'samples': 3707136, 'steps': 19307, 'loss/train': 1.2972187399864197} 01/29/2022 13:40:04 - INFO - codeparrot_training - Step 19308: {'lr': 0.0003560300988361321, 'samples': 3707328, 'steps': 19308, 'loss/train': 0.29542820155620575} 01/29/2022 13:40:08 - INFO - codeparrot_training - Step 19309: {'lr': 0.00035601528066856315, 'samples': 3707520, 'steps': 19309, 'loss/train': 2.1591036915779114} 01/29/2022 13:40:12 - INFO - codeparrot_training - Step 19310: {'lr': 0.00035600046204685844, 'samples': 3707712, 'steps': 19310, 'loss/train': 0.9085870385169983} 01/29/2022 13:40:18 - INFO - codeparrot_training - Step 19311: {'lr': 0.00035598564297108134, 'samples': 3707904, 'steps': 19311, 'loss/train': 1.3352764248847961} 01/29/2022 13:40:22 - INFO - codeparrot_training - Step 19312: {'lr': 0.0003559708234412954, 'samples': 3708096, 'steps': 19312, 'loss/train': 1.1895466446876526} 01/29/2022 13:40:26 - INFO - codeparrot_training - Step 19313: {'lr': 0.00035595600345756414, 'samples': 3708288, 'steps': 19313, 'loss/train': 1.4241807460784912} 01/29/2022 13:40:31 - INFO - codeparrot_training - Step 19314: {'lr': 0.00035594118301995095, 'samples': 3708480, 'steps': 19314, 'loss/train': 0.9037580788135529} 01/29/2022 13:40:35 - INFO - codeparrot_training - Step 19315: {'lr': 0.00035592636212851945, 'samples': 3708672, 'steps': 19315, 'loss/train': 1.9752067923545837} 01/29/2022 13:40:41 - INFO - codeparrot_training - Step 19316: {'lr': 0.000355911540783333, 'samples': 3708864, 'steps': 19316, 'loss/train': 1.5568808913230896} 01/29/2022 13:40:45 - INFO - codeparrot_training - Step 19317: {'lr': 0.00035589671898445517, 'samples': 3709056, 'steps': 19317, 'loss/train': 1.9113410711288452} 01/29/2022 13:40:50 - INFO - codeparrot_training - Step 19318: {'lr': 0.0003558818967319494, 'samples': 3709248, 'steps': 19318, 'loss/train': 1.41365647315979} 01/29/2022 13:40:54 - INFO - codeparrot_training - Step 19319: {'lr': 0.0003558670740258792, 'samples': 3709440, 'steps': 19319, 'loss/train': 2.0681442618370056} 01/29/2022 13:40:58 - INFO - codeparrot_training - Step 19320: {'lr': 0.00035585225086630807, 'samples': 3709632, 'steps': 19320, 'loss/train': 1.7366800904273987} 01/29/2022 13:41:03 - INFO - codeparrot_training - Step 19321: {'lr': 0.00035583742725329954, 'samples': 3709824, 'steps': 19321, 'loss/train': 1.7318909168243408} 01/29/2022 13:41:07 - INFO - codeparrot_training - Step 19322: {'lr': 0.0003558226031869171, 'samples': 3710016, 'steps': 19322, 'loss/train': 1.3606842756271362} 01/29/2022 13:41:13 - INFO - codeparrot_training - Step 19323: {'lr': 0.00035580777866722415, 'samples': 3710208, 'steps': 19323, 'loss/train': 2.124649465084076} 01/29/2022 13:41:18 - INFO - codeparrot_training - Step 19324: {'lr': 0.00035579295369428425, 'samples': 3710400, 'steps': 19324, 'loss/train': 0.9372937381267548} 01/29/2022 13:41:22 - INFO - codeparrot_training - Step 19325: {'lr': 0.000355778128268161, 'samples': 3710592, 'steps': 19325, 'loss/train': 1.4259324967861176} 01/29/2022 13:41:26 - INFO - codeparrot_training - Step 19326: {'lr': 0.0003557633023889179, 'samples': 3710784, 'steps': 19326, 'loss/train': 2.076359510421753} 01/29/2022 13:41:30 - INFO - codeparrot_training - Step 19327: {'lr': 0.0003557484760566183, 'samples': 3710976, 'steps': 19327, 'loss/train': 1.405491828918457} 01/29/2022 13:41:36 - INFO - codeparrot_training - Step 19328: {'lr': 0.0003557336492713258, 'samples': 3711168, 'steps': 19328, 'loss/train': 1.9286068081855774} 01/29/2022 13:41:41 - INFO - codeparrot_training - Step 19329: {'lr': 0.000355718822033104, 'samples': 3711360, 'steps': 19329, 'loss/train': 2.0601757764816284} 01/29/2022 13:41:45 - INFO - codeparrot_training - Step 19330: {'lr': 0.0003557039943420163, 'samples': 3711552, 'steps': 19330, 'loss/train': 1.3102186024188995} 01/29/2022 13:41:49 - INFO - codeparrot_training - Step 19331: {'lr': 0.00035568916619812624, 'samples': 3711744, 'steps': 19331, 'loss/train': 1.2405036985874176} 01/29/2022 13:41:53 - INFO - codeparrot_training - Step 19332: {'lr': 0.00035567433760149737, 'samples': 3711936, 'steps': 19332, 'loss/train': 1.8251160979270935} 01/29/2022 13:42:00 - INFO - codeparrot_training - Step 19333: {'lr': 0.0003556595085521931, 'samples': 3712128, 'steps': 19333, 'loss/train': 1.403119057416916} 01/29/2022 13:42:05 - INFO - codeparrot_training - Step 19334: {'lr': 0.0003556446790502772, 'samples': 3712320, 'steps': 19334, 'loss/train': 2.090827524662018} 01/29/2022 13:42:09 - INFO - codeparrot_training - Step 19335: {'lr': 0.00035562984909581297, 'samples': 3712512, 'steps': 19335, 'loss/train': 2.109629988670349} 01/29/2022 13:42:13 - INFO - codeparrot_training - Step 19336: {'lr': 0.0003556150186888639, 'samples': 3712704, 'steps': 19336, 'loss/train': 2.5612149238586426} 01/29/2022 13:42:17 - INFO - codeparrot_training - Step 19337: {'lr': 0.00035560018782949384, 'samples': 3712896, 'steps': 19337, 'loss/train': 0.7052338421344757} 01/29/2022 13:42:23 - INFO - codeparrot_training - Step 19338: {'lr': 0.0003555853565177659, 'samples': 3713088, 'steps': 19338, 'loss/train': 1.527634084224701} 01/29/2022 13:42:27 - INFO - codeparrot_training - Step 19339: {'lr': 0.00035557052475374397, 'samples': 3713280, 'steps': 19339, 'loss/train': 1.6220794916152954} 01/29/2022 13:42:31 - INFO - codeparrot_training - Step 19340: {'lr': 0.00035555569253749135, 'samples': 3713472, 'steps': 19340, 'loss/train': 2.183756947517395} 01/29/2022 13:42:36 - INFO - codeparrot_training - Step 19341: {'lr': 0.0003555408598690718, 'samples': 3713664, 'steps': 19341, 'loss/train': 1.3899992108345032} 01/29/2022 13:42:41 - INFO - codeparrot_training - Step 19342: {'lr': 0.0003555260267485485, 'samples': 3713856, 'steps': 19342, 'loss/train': 1.6968523263931274} 01/29/2022 13:42:45 - INFO - codeparrot_training - Step 19343: {'lr': 0.00035551119317598533, 'samples': 3714048, 'steps': 19343, 'loss/train': 2.1894920468330383} 01/29/2022 13:42:50 - INFO - codeparrot_training - Step 19344: {'lr': 0.00035549635915144574, 'samples': 3714240, 'steps': 19344, 'loss/train': 1.5634589195251465} 01/29/2022 13:42:54 - INFO - codeparrot_training - Step 19345: {'lr': 0.0003554815246749932, 'samples': 3714432, 'steps': 19345, 'loss/train': 0.26373372226953506} 01/29/2022 13:42:58 - INFO - codeparrot_training - Step 19346: {'lr': 0.00035546668974669127, 'samples': 3714624, 'steps': 19346, 'loss/train': 1.7100175023078918} 01/29/2022 13:43:02 - INFO - codeparrot_training - Step 19347: {'lr': 0.00035545185436660357, 'samples': 3714816, 'steps': 19347, 'loss/train': 2.1255136728286743} 01/29/2022 13:43:08 - INFO - codeparrot_training - Step 19348: {'lr': 0.00035543701853479366, 'samples': 3715008, 'steps': 19348, 'loss/train': 0.8012173175811768} 01/29/2022 13:43:12 - INFO - codeparrot_training - Step 19349: {'lr': 0.00035542218225132497, 'samples': 3715200, 'steps': 19349, 'loss/train': 1.5481020212173462} 01/29/2022 13:43:16 - INFO - codeparrot_training - Step 19350: {'lr': 0.00035540734551626113, 'samples': 3715392, 'steps': 19350, 'loss/train': 1.3155416250228882} 01/29/2022 13:43:21 - INFO - codeparrot_training - Step 19351: {'lr': 0.00035539250832966574, 'samples': 3715584, 'steps': 19351, 'loss/train': 1.7126977443695068} 01/29/2022 13:43:28 - INFO - codeparrot_training - Step 19352: {'lr': 0.00035537767069160234, 'samples': 3715776, 'steps': 19352, 'loss/train': 1.134744107723236} 01/29/2022 13:43:32 - INFO - codeparrot_training - Step 19353: {'lr': 0.00035536283260213434, 'samples': 3715968, 'steps': 19353, 'loss/train': 1.9037554264068604} 01/29/2022 13:43:37 - INFO - codeparrot_training - Step 19354: {'lr': 0.0003553479940613255, 'samples': 3716160, 'steps': 19354, 'loss/train': 0.6456656903028488} 01/29/2022 13:43:41 - INFO - codeparrot_training - Step 19355: {'lr': 0.00035533315506923924, 'samples': 3716352, 'steps': 19355, 'loss/train': 1.4372904896736145} 01/29/2022 13:43:45 - INFO - codeparrot_training - Step 19356: {'lr': 0.0003553183156259393, 'samples': 3716544, 'steps': 19356, 'loss/train': 1.4618771374225616} 01/29/2022 13:43:51 - INFO - codeparrot_training - Step 19357: {'lr': 0.00035530347573148904, 'samples': 3716736, 'steps': 19357, 'loss/train': 1.7988474369049072} 01/29/2022 13:43:55 - INFO - codeparrot_training - Step 19358: {'lr': 0.0003552886353859522, 'samples': 3716928, 'steps': 19358, 'loss/train': 1.2287420332431793} 01/29/2022 13:43:59 - INFO - codeparrot_training - Step 19359: {'lr': 0.00035527379458939225, 'samples': 3717120, 'steps': 19359, 'loss/train': 1.3816721141338348} 01/29/2022 13:44:04 - INFO - codeparrot_training - Step 19360: {'lr': 0.00035525895334187274, 'samples': 3717312, 'steps': 19360, 'loss/train': 1.214848279953003} 01/29/2022 13:44:08 - INFO - codeparrot_training - Step 19361: {'lr': 0.0003552441116434574, 'samples': 3717504, 'steps': 19361, 'loss/train': 2.0817970633506775} 01/29/2022 13:44:12 - INFO - codeparrot_training - Step 19362: {'lr': 0.0003552292694942096, 'samples': 3717696, 'steps': 19362, 'loss/train': 0.23057153820991516} 01/29/2022 13:44:17 - INFO - codeparrot_training - Step 19363: {'lr': 0.0003552144268941931, 'samples': 3717888, 'steps': 19363, 'loss/train': 0.7876706421375275} 01/29/2022 13:44:22 - INFO - codeparrot_training - Step 19364: {'lr': 0.00035519958384347134, 'samples': 3718080, 'steps': 19364, 'loss/train': 1.5334550142288208} 01/29/2022 13:44:26 - INFO - codeparrot_training - Step 19365: {'lr': 0.000355184740342108, 'samples': 3718272, 'steps': 19365, 'loss/train': 2.7046676874160767} 01/29/2022 13:44:30 - INFO - codeparrot_training - Step 19366: {'lr': 0.00035516989639016664, 'samples': 3718464, 'steps': 19366, 'loss/train': 1.9476765990257263} 01/29/2022 13:44:34 - INFO - codeparrot_training - Step 19367: {'lr': 0.00035515505198771086, 'samples': 3718656, 'steps': 19367, 'loss/train': 2.058043956756592} 01/29/2022 13:44:41 - INFO - codeparrot_training - Step 19368: {'lr': 0.0003551402071348042, 'samples': 3718848, 'steps': 19368, 'loss/train': 1.7699398398399353} 01/29/2022 13:44:46 - INFO - codeparrot_training - Step 19369: {'lr': 0.0003551253618315103, 'samples': 3719040, 'steps': 19369, 'loss/train': 1.7578523755073547} 01/29/2022 13:44:50 - INFO - codeparrot_training - Step 19370: {'lr': 0.0003551105160778927, 'samples': 3719232, 'steps': 19370, 'loss/train': 0.9237118363380432} 01/29/2022 13:44:54 - INFO - codeparrot_training - Step 19371: {'lr': 0.000355095669874015, 'samples': 3719424, 'steps': 19371, 'loss/train': 2.881007194519043} 01/29/2022 13:44:58 - INFO - codeparrot_training - Step 19372: {'lr': 0.00035508082321994097, 'samples': 3719616, 'steps': 19372, 'loss/train': 1.4174251556396484} 01/29/2022 13:45:04 - INFO - codeparrot_training - Step 19373: {'lr': 0.00035506597611573387, 'samples': 3719808, 'steps': 19373, 'loss/train': 1.471727192401886} 01/29/2022 13:45:08 - INFO - codeparrot_training - Step 19374: {'lr': 0.0003550511285614576, 'samples': 3720000, 'steps': 19374, 'loss/train': 1.7877835035324097} 01/29/2022 13:45:12 - INFO - codeparrot_training - Step 19375: {'lr': 0.0003550362805571756, 'samples': 3720192, 'steps': 19375, 'loss/train': 1.680222988128662} 01/29/2022 13:45:16 - INFO - codeparrot_training - Step 19376: {'lr': 0.00035502143210295163, 'samples': 3720384, 'steps': 19376, 'loss/train': 1.674788475036621} 01/29/2022 13:45:21 - INFO - codeparrot_training - Step 19377: {'lr': 0.000355006583198849, 'samples': 3720576, 'steps': 19377, 'loss/train': 1.992154061794281} 01/29/2022 13:45:28 - INFO - codeparrot_training - Step 19378: {'lr': 0.00035499173384493174, 'samples': 3720768, 'steps': 19378, 'loss/train': 1.6372137665748596} 01/29/2022 13:45:32 - INFO - codeparrot_training - Step 19379: {'lr': 0.00035497688404126306, 'samples': 3720960, 'steps': 19379, 'loss/train': 2.202315866947174} 01/29/2022 13:45:37 - INFO - codeparrot_training - Step 19380: {'lr': 0.00035496203378790683, 'samples': 3721152, 'steps': 19380, 'loss/train': 1.9696261882781982} 01/29/2022 13:45:41 - INFO - codeparrot_training - Step 19381: {'lr': 0.0003549471830849265, 'samples': 3721344, 'steps': 19381, 'loss/train': 1.954457938671112} 01/29/2022 13:45:46 - INFO - codeparrot_training - Step 19382: {'lr': 0.00035493233193238584, 'samples': 3721536, 'steps': 19382, 'loss/train': 0.9396941363811493} 01/29/2022 13:45:50 - INFO - codeparrot_training - Step 19383: {'lr': 0.00035491748033034836, 'samples': 3721728, 'steps': 19383, 'loss/train': 1.1355358958244324} 01/29/2022 13:45:55 - INFO - codeparrot_training - Step 19384: {'lr': 0.00035490262827887764, 'samples': 3721920, 'steps': 19384, 'loss/train': 2.603632092475891} 01/29/2022 13:45:59 - INFO - codeparrot_training - Step 19385: {'lr': 0.0003548877757780375, 'samples': 3722112, 'steps': 19385, 'loss/train': 1.6643811464309692} 01/29/2022 13:46:03 - INFO - codeparrot_training - Step 19386: {'lr': 0.00035487292282789136, 'samples': 3722304, 'steps': 19386, 'loss/train': 1.9192959666252136} 01/29/2022 13:46:08 - INFO - codeparrot_training - Step 19387: {'lr': 0.000354858069428503, 'samples': 3722496, 'steps': 19387, 'loss/train': 1.5997576117515564} 01/29/2022 13:46:13 - INFO - codeparrot_training - Step 19388: {'lr': 0.0003548432155799358, 'samples': 3722688, 'steps': 19388, 'loss/train': 1.7787086963653564} 01/29/2022 13:46:17 - INFO - codeparrot_training - Step 19389: {'lr': 0.0003548283612822537, 'samples': 3722880, 'steps': 19389, 'loss/train': 1.7992352843284607} 01/29/2022 13:46:21 - INFO - codeparrot_training - Step 19390: {'lr': 0.0003548135065355201, 'samples': 3723072, 'steps': 19390, 'loss/train': 0.2576488181948662} 01/29/2022 13:46:25 - INFO - codeparrot_training - Step 19391: {'lr': 0.0003547986513397988, 'samples': 3723264, 'steps': 19391, 'loss/train': 2.141083538532257} 01/29/2022 13:46:33 - INFO - codeparrot_training - Step 19392: {'lr': 0.0003547837956951533, 'samples': 3723456, 'steps': 19392, 'loss/train': 1.795125961303711} 01/29/2022 13:46:37 - INFO - codeparrot_training - Step 19393: {'lr': 0.00035476893960164734, 'samples': 3723648, 'steps': 19393, 'loss/train': 1.8483738899230957} 01/29/2022 13:46:41 - INFO - codeparrot_training - Step 19394: {'lr': 0.00035475408305934444, 'samples': 3723840, 'steps': 19394, 'loss/train': 1.6809527277946472} 01/29/2022 13:46:45 - INFO - codeparrot_training - Step 19395: {'lr': 0.0003547392260683084, 'samples': 3724032, 'steps': 19395, 'loss/train': 1.80326646566391} 01/29/2022 13:46:50 - INFO - codeparrot_training - Step 19396: {'lr': 0.0003547243686286027, 'samples': 3724224, 'steps': 19396, 'loss/train': 3.2037320137023926} 01/29/2022 13:46:56 - INFO - codeparrot_training - Step 19397: {'lr': 0.000354709510740291, 'samples': 3724416, 'steps': 19397, 'loss/train': 2.3395907878875732} 01/29/2022 13:47:00 - INFO - codeparrot_training - Step 19398: {'lr': 0.0003546946524034371, 'samples': 3724608, 'steps': 19398, 'loss/train': 2.3769335746765137} 01/29/2022 13:47:04 - INFO - codeparrot_training - Step 19399: {'lr': 0.00035467979361810455, 'samples': 3724800, 'steps': 19399, 'loss/train': 2.237196922302246} 01/29/2022 13:47:08 - INFO - codeparrot_training - Step 19400: {'lr': 0.00035466493438435703, 'samples': 3724992, 'steps': 19400, 'loss/train': 1.913466453552246} 01/29/2022 13:47:13 - INFO - codeparrot_training - Step 19401: {'lr': 0.00035465007470225813, 'samples': 3725184, 'steps': 19401, 'loss/train': 2.1927616596221924} 01/29/2022 13:47:17 - INFO - codeparrot_training - Step 19402: {'lr': 0.0003546352145718715, 'samples': 3725376, 'steps': 19402, 'loss/train': 1.193751722574234} 01/29/2022 13:47:22 - INFO - codeparrot_training - Step 19403: {'lr': 0.0003546203539932609, 'samples': 3725568, 'steps': 19403, 'loss/train': 1.2580593824386597} 01/29/2022 13:47:26 - INFO - codeparrot_training - Step 19404: {'lr': 0.0003546054929664899, 'samples': 3725760, 'steps': 19404, 'loss/train': 1.8426806330680847} 01/29/2022 13:47:31 - INFO - codeparrot_training - Step 19405: {'lr': 0.0003545906314916222, 'samples': 3725952, 'steps': 19405, 'loss/train': 2.292962908744812} 01/29/2022 13:47:35 - INFO - codeparrot_training - Step 19406: {'lr': 0.00035457576956872145, 'samples': 3726144, 'steps': 19406, 'loss/train': 2.4563063979148865} 01/29/2022 13:47:39 - INFO - codeparrot_training - Step 19407: {'lr': 0.00035456090719785126, 'samples': 3726336, 'steps': 19407, 'loss/train': 0.9129968583583832} 01/29/2022 13:47:44 - INFO - codeparrot_training - Step 19408: {'lr': 0.00035454604437907536, 'samples': 3726528, 'steps': 19408, 'loss/train': 1.7886977791786194} 01/29/2022 13:47:49 - INFO - codeparrot_training - Step 19409: {'lr': 0.0003545311811124574, 'samples': 3726720, 'steps': 19409, 'loss/train': 1.0276727378368378} 01/29/2022 13:47:53 - INFO - codeparrot_training - Step 19410: {'lr': 0.0003545163173980611, 'samples': 3726912, 'steps': 19410, 'loss/train': 0.5992851108312607} 01/29/2022 13:47:57 - INFO - codeparrot_training - Step 19411: {'lr': 0.0003545014532359501, 'samples': 3727104, 'steps': 19411, 'loss/train': 1.8265594840049744} 01/29/2022 13:48:02 - INFO - codeparrot_training - Step 19412: {'lr': 0.000354486588626188, 'samples': 3727296, 'steps': 19412, 'loss/train': 1.9748257398605347} 01/29/2022 13:48:09 - INFO - codeparrot_training - Step 19413: {'lr': 0.0003544717235688385, 'samples': 3727488, 'steps': 19413, 'loss/train': 1.2573156952857971} 01/29/2022 13:48:13 - INFO - codeparrot_training - Step 19414: {'lr': 0.00035445685806396543, 'samples': 3727680, 'steps': 19414, 'loss/train': 2.3846726417541504} 01/29/2022 13:48:17 - INFO - codeparrot_training - Step 19415: {'lr': 0.00035444199211163226, 'samples': 3727872, 'steps': 19415, 'loss/train': 1.3412871658802032} 01/29/2022 13:48:22 - INFO - codeparrot_training - Step 19416: {'lr': 0.0003544271257119028, 'samples': 3728064, 'steps': 19416, 'loss/train': 1.7819963693618774} 01/29/2022 13:48:26 - INFO - codeparrot_training - Step 19417: {'lr': 0.00035441225886484066, 'samples': 3728256, 'steps': 19417, 'loss/train': 2.5657854080200195} 01/29/2022 13:48:32 - INFO - codeparrot_training - Step 19418: {'lr': 0.0003543973915705095, 'samples': 3728448, 'steps': 19418, 'loss/train': 2.0778547525405884} 01/29/2022 13:48:36 - INFO - codeparrot_training - Step 19419: {'lr': 0.0003543825238289732, 'samples': 3728640, 'steps': 19419, 'loss/train': 2.973832368850708} 01/29/2022 13:48:40 - INFO - codeparrot_training - Step 19420: {'lr': 0.0003543676556402952, 'samples': 3728832, 'steps': 19420, 'loss/train': 0.0903626773506403} 01/29/2022 13:48:45 - INFO - codeparrot_training - Step 19421: {'lr': 0.00035435278700453934, 'samples': 3729024, 'steps': 19421, 'loss/train': 1.5863229632377625} 01/29/2022 13:48:49 - INFO - codeparrot_training - Step 19422: {'lr': 0.0003543379179217693, 'samples': 3729216, 'steps': 19422, 'loss/train': 7.239912271499634} 01/29/2022 13:48:56 - INFO - codeparrot_training - Step 19423: {'lr': 0.0003543230483920487, 'samples': 3729408, 'steps': 19423, 'loss/train': 1.2824660539627075} 01/29/2022 13:49:00 - INFO - codeparrot_training - Step 19424: {'lr': 0.0003543081784154414, 'samples': 3729600, 'steps': 19424, 'loss/train': 0.7865186333656311} 01/29/2022 13:49:04 - INFO - codeparrot_training - Step 19425: {'lr': 0.00035429330799201085, 'samples': 3729792, 'steps': 19425, 'loss/train': 2.378458321094513} 01/29/2022 13:49:09 - INFO - codeparrot_training - Step 19426: {'lr': 0.00035427843712182097, 'samples': 3729984, 'steps': 19426, 'loss/train': 0.98880735039711} 01/29/2022 13:49:13 - INFO - codeparrot_training - Step 19427: {'lr': 0.0003542635658049353, 'samples': 3730176, 'steps': 19427, 'loss/train': 1.2088256478309631} 01/29/2022 13:49:18 - INFO - codeparrot_training - Step 19428: {'lr': 0.0003542486940414177, 'samples': 3730368, 'steps': 19428, 'loss/train': 2.0833160877227783} 01/29/2022 13:49:22 - INFO - codeparrot_training - Step 19429: {'lr': 0.00035423382183133174, 'samples': 3730560, 'steps': 19429, 'loss/train': 1.7698091268539429} 01/29/2022 13:49:27 - INFO - codeparrot_training - Step 19430: {'lr': 0.00035421894917474125, 'samples': 3730752, 'steps': 19430, 'loss/train': 1.4944253861904144} 01/29/2022 13:49:31 - INFO - codeparrot_training - Step 19431: {'lr': 0.0003542040760717098, 'samples': 3730944, 'steps': 19431, 'loss/train': 1.739732265472412} 01/29/2022 13:49:35 - INFO - codeparrot_training - Step 19432: {'lr': 0.0003541892025223012, 'samples': 3731136, 'steps': 19432, 'loss/train': 1.420136958360672} 01/29/2022 13:49:41 - INFO - codeparrot_training - Step 19433: {'lr': 0.00035417432852657916, 'samples': 3731328, 'steps': 19433, 'loss/train': 2.20688796043396} 01/29/2022 13:49:45 - INFO - codeparrot_training - Step 19434: {'lr': 0.00035415945408460737, 'samples': 3731520, 'steps': 19434, 'loss/train': 1.6722416281700134} 01/29/2022 13:49:49 - INFO - codeparrot_training - Step 19435: {'lr': 0.0003541445791964496, 'samples': 3731712, 'steps': 19435, 'loss/train': 1.4715738594532013} 01/29/2022 13:49:53 - INFO - codeparrot_training - Step 19436: {'lr': 0.0003541297038621694, 'samples': 3731904, 'steps': 19436, 'loss/train': 2.1680826544761658} 01/29/2022 13:49:58 - INFO - codeparrot_training - Step 19437: {'lr': 0.0003541148280818307, 'samples': 3732096, 'steps': 19437, 'loss/train': 2.4934911131858826} 01/29/2022 13:50:06 - INFO - codeparrot_training - Step 19438: {'lr': 0.00035409995185549717, 'samples': 3732288, 'steps': 19438, 'loss/train': 1.6800488233566284} 01/29/2022 13:50:10 - INFO - codeparrot_training - Step 19439: {'lr': 0.00035408507518323244, 'samples': 3732480, 'steps': 19439, 'loss/train': 1.5106651782989502} 01/29/2022 13:50:14 - INFO - codeparrot_training - Step 19440: {'lr': 0.00035407019806510035, 'samples': 3732672, 'steps': 19440, 'loss/train': 3.785479187965393} 01/29/2022 13:50:19 - INFO - codeparrot_training - Step 19441: {'lr': 0.0003540553205011645, 'samples': 3732864, 'steps': 19441, 'loss/train': 1.7353291511535645} 01/29/2022 13:50:23 - INFO - codeparrot_training - Step 19442: {'lr': 0.00035404044249148873, 'samples': 3733056, 'steps': 19442, 'loss/train': 1.4693103432655334} 01/29/2022 13:50:27 - INFO - codeparrot_training - Step 19443: {'lr': 0.0003540255640361368, 'samples': 3733248, 'steps': 19443, 'loss/train': 2.1677250266075134} 01/29/2022 13:50:33 - INFO - codeparrot_training - Step 19444: {'lr': 0.0003540106851351723, 'samples': 3733440, 'steps': 19444, 'loss/train': 1.919721007347107} 01/29/2022 13:50:37 - INFO - codeparrot_training - Step 19445: {'lr': 0.00035399580578865907, 'samples': 3733632, 'steps': 19445, 'loss/train': 1.376562774181366} 01/29/2022 13:50:41 - INFO - codeparrot_training - Step 19446: {'lr': 0.00035398092599666086, 'samples': 3733824, 'steps': 19446, 'loss/train': 1.6293708086013794} 01/29/2022 13:50:45 - INFO - codeparrot_training - Step 19447: {'lr': 0.00035396604575924133, 'samples': 3734016, 'steps': 19447, 'loss/train': 2.5913155674934387} 01/29/2022 13:50:50 - INFO - codeparrot_training - Step 19448: {'lr': 0.00035395116507646435, 'samples': 3734208, 'steps': 19448, 'loss/train': 1.7563676834106445} 01/29/2022 13:50:55 - INFO - codeparrot_training - Step 19449: {'lr': 0.00035393628394839356, 'samples': 3734400, 'steps': 19449, 'loss/train': 2.173669159412384} 01/29/2022 13:50:59 - INFO - codeparrot_training - Step 19450: {'lr': 0.00035392140237509276, 'samples': 3734592, 'steps': 19450, 'loss/train': 2.195496439933777} 01/29/2022 13:51:03 - INFO - codeparrot_training - Step 19451: {'lr': 0.0003539065203566256, 'samples': 3734784, 'steps': 19451, 'loss/train': 1.5667673349380493} 01/29/2022 13:51:08 - INFO - codeparrot_training - Step 19452: {'lr': 0.0003538916378930559, 'samples': 3734976, 'steps': 19452, 'loss/train': 1.9388174414634705} 01/29/2022 13:51:12 - INFO - codeparrot_training - Step 19453: {'lr': 0.0003538767549844475, 'samples': 3735168, 'steps': 19453, 'loss/train': 1.6762695908546448} 01/29/2022 13:51:17 - INFO - codeparrot_training - Step 19454: {'lr': 0.000353861871630864, 'samples': 3735360, 'steps': 19454, 'loss/train': 1.8814929127693176} 01/29/2022 13:51:22 - INFO - codeparrot_training - Step 19455: {'lr': 0.00035384698783236923, 'samples': 3735552, 'steps': 19455, 'loss/train': 1.7577164769172668} 01/29/2022 13:51:26 - INFO - codeparrot_training - Step 19456: {'lr': 0.00035383210358902695, 'samples': 3735744, 'steps': 19456, 'loss/train': 1.5839633345603943} 01/29/2022 13:51:30 - INFO - codeparrot_training - Step 19457: {'lr': 0.0003538172189009009, 'samples': 3735936, 'steps': 19457, 'loss/train': 2.394657254219055} 01/29/2022 13:51:35 - INFO - codeparrot_training - Step 19458: {'lr': 0.00035380233376805487, 'samples': 3736128, 'steps': 19458, 'loss/train': 1.3597136735916138} 01/29/2022 13:51:42 - INFO - codeparrot_training - Step 19459: {'lr': 0.00035378744819055264, 'samples': 3736320, 'steps': 19459, 'loss/train': 1.5673354268074036} 01/29/2022 13:51:46 - INFO - codeparrot_training - Step 19460: {'lr': 0.0003537725621684578, 'samples': 3736512, 'steps': 19460, 'loss/train': 1.6952446103096008} 01/29/2022 13:51:50 - INFO - codeparrot_training - Step 19461: {'lr': 0.0003537576757018344, 'samples': 3736704, 'steps': 19461, 'loss/train': 1.9990239143371582} 01/29/2022 13:51:55 - INFO - codeparrot_training - Step 19462: {'lr': 0.0003537427887907459, 'samples': 3736896, 'steps': 19462, 'loss/train': 2.709523558616638} 01/29/2022 13:51:59 - INFO - codeparrot_training - Step 19463: {'lr': 0.0003537279014352565, 'samples': 3737088, 'steps': 19463, 'loss/train': 1.7900345921516418} 01/29/2022 13:52:04 - INFO - codeparrot_training - Step 19464: {'lr': 0.00035371301363542945, 'samples': 3737280, 'steps': 19464, 'loss/train': 2.46576726436615} 01/29/2022 13:52:08 - INFO - codeparrot_training - Step 19465: {'lr': 0.00035369812539132894, 'samples': 3737472, 'steps': 19465, 'loss/train': 1.5680267214775085} 01/29/2022 13:52:13 - INFO - codeparrot_training - Step 19466: {'lr': 0.0003536832367030185, 'samples': 3737664, 'steps': 19466, 'loss/train': 1.630452275276184} 01/29/2022 13:52:17 - INFO - codeparrot_training - Step 19467: {'lr': 0.00035366834757056203, 'samples': 3737856, 'steps': 19467, 'loss/train': 1.847682237625122} 01/29/2022 13:52:21 - INFO - codeparrot_training - Step 19468: {'lr': 0.0003536534579940233, 'samples': 3738048, 'steps': 19468, 'loss/train': 1.864047646522522} 01/29/2022 13:52:29 - INFO - codeparrot_training - Step 19469: {'lr': 0.0003536385679734659, 'samples': 3738240, 'steps': 19469, 'loss/train': 1.6869807243347168} 01/29/2022 13:52:33 - INFO - codeparrot_training - Step 19470: {'lr': 0.000353623677508954, 'samples': 3738432, 'steps': 19470, 'loss/train': 1.159453958272934} 01/29/2022 13:52:37 - INFO - codeparrot_training - Step 19471: {'lr': 0.00035360878660055107, 'samples': 3738624, 'steps': 19471, 'loss/train': 1.711617887020111} 01/29/2022 13:52:41 - INFO - codeparrot_training - Step 19472: {'lr': 0.0003535938952483211, 'samples': 3738816, 'steps': 19472, 'loss/train': 1.557694673538208} 01/29/2022 13:52:46 - INFO - codeparrot_training - Step 19473: {'lr': 0.00035357900345232764, 'samples': 3739008, 'steps': 19473, 'loss/train': 2.2966407537460327} 01/29/2022 13:52:52 - INFO - codeparrot_training - Step 19474: {'lr': 0.0003535641112126347, 'samples': 3739200, 'steps': 19474, 'loss/train': 1.4037455320358276} 01/29/2022 13:52:56 - INFO - codeparrot_training - Step 19475: {'lr': 0.00035354921852930596, 'samples': 3739392, 'steps': 19475, 'loss/train': 1.5435470938682556} 01/29/2022 13:53:00 - INFO - codeparrot_training - Step 19476: {'lr': 0.0003535343254024053, 'samples': 3739584, 'steps': 19476, 'loss/train': 1.512457251548767} 01/29/2022 13:53:04 - INFO - codeparrot_training - Step 19477: {'lr': 0.00035351943183199643, 'samples': 3739776, 'steps': 19477, 'loss/train': 1.8418217897415161} 01/29/2022 13:53:08 - INFO - codeparrot_training - Step 19478: {'lr': 0.0003535045378181432, 'samples': 3739968, 'steps': 19478, 'loss/train': 1.2724326252937317} 01/29/2022 13:53:14 - INFO - codeparrot_training - Step 19479: {'lr': 0.0003534896433609093, 'samples': 3740160, 'steps': 19479, 'loss/train': 0.9708538055419922} 01/29/2022 13:53:18 - INFO - codeparrot_training - Step 19480: {'lr': 0.0003534747484603587, 'samples': 3740352, 'steps': 19480, 'loss/train': 2.4773143529891968} 01/29/2022 13:53:22 - INFO - codeparrot_training - Step 19481: {'lr': 0.00035345985311655513, 'samples': 3740544, 'steps': 19481, 'loss/train': 1.8531169295310974} 01/29/2022 13:53:27 - INFO - codeparrot_training - Step 19482: {'lr': 0.0003534449573295624, 'samples': 3740736, 'steps': 19482, 'loss/train': 0.5209571868181229} 01/29/2022 13:53:31 - INFO - codeparrot_training - Step 19483: {'lr': 0.0003534300610994444, 'samples': 3740928, 'steps': 19483, 'loss/train': 1.5321469902992249} 01/29/2022 13:53:37 - INFO - codeparrot_training - Step 19484: {'lr': 0.00035341516442626475, 'samples': 3741120, 'steps': 19484, 'loss/train': 2.0868038535118103} 01/29/2022 13:53:42 - INFO - codeparrot_training - Step 19485: {'lr': 0.00035340026731008745, 'samples': 3741312, 'steps': 19485, 'loss/train': 1.280583769083023} 01/29/2022 13:53:46 - INFO - codeparrot_training - Step 19486: {'lr': 0.0003533853697509762, 'samples': 3741504, 'steps': 19486, 'loss/train': 1.291856288909912} 01/29/2022 13:53:50 - INFO - codeparrot_training - Step 19487: {'lr': 0.0003533704717489949, 'samples': 3741696, 'steps': 19487, 'loss/train': 2.454797923564911} 01/29/2022 13:53:54 - INFO - codeparrot_training - Step 19488: {'lr': 0.00035335557330420724, 'samples': 3741888, 'steps': 19488, 'loss/train': 0.6178584694862366} 01/29/2022 13:54:00 - INFO - codeparrot_training - Step 19489: {'lr': 0.0003533406744166772, 'samples': 3742080, 'steps': 19489, 'loss/train': 0.793362557888031} 01/29/2022 13:54:04 - INFO - codeparrot_training - Step 19490: {'lr': 0.00035332577508646846, 'samples': 3742272, 'steps': 19490, 'loss/train': 2.011304497718811} 01/29/2022 13:54:08 - INFO - codeparrot_training - Step 19491: {'lr': 0.000353310875313645, 'samples': 3742464, 'steps': 19491, 'loss/train': 1.840882658958435} 01/29/2022 13:54:12 - INFO - codeparrot_training - Step 19492: {'lr': 0.00035329597509827046, 'samples': 3742656, 'steps': 19492, 'loss/train': 1.760693371295929} 01/29/2022 13:54:17 - INFO - codeparrot_training - Step 19493: {'lr': 0.00035328107444040875, 'samples': 3742848, 'steps': 19493, 'loss/train': 1.657431185245514} 01/29/2022 13:54:22 - INFO - codeparrot_training - Step 19494: {'lr': 0.0003532661733401238, 'samples': 3743040, 'steps': 19494, 'loss/train': 1.3900751173496246} 01/29/2022 13:54:26 - INFO - codeparrot_training - Step 19495: {'lr': 0.0003532512717974793, 'samples': 3743232, 'steps': 19495, 'loss/train': 0.9286145567893982} 01/29/2022 13:54:30 - INFO - codeparrot_training - Step 19496: {'lr': 0.00035323636981253914, 'samples': 3743424, 'steps': 19496, 'loss/train': 2.0184507966041565} 01/29/2022 13:54:34 - INFO - codeparrot_training - Step 19497: {'lr': 0.00035322146738536714, 'samples': 3743616, 'steps': 19497, 'loss/train': 1.5808299779891968} 01/29/2022 13:54:39 - INFO - codeparrot_training - Step 19498: {'lr': 0.00035320656451602725, 'samples': 3743808, 'steps': 19498, 'loss/train': 2.684774100780487} 01/29/2022 13:54:46 - INFO - codeparrot_training - Step 19499: {'lr': 0.00035319166120458315, 'samples': 3744000, 'steps': 19499, 'loss/train': 2.4118316173553467} 01/29/2022 13:54:50 - INFO - codeparrot_training - Step 19500: {'lr': 0.00035317675745109866, 'samples': 3744192, 'steps': 19500, 'loss/train': 1.6627668142318726} 01/29/2022 13:54:54 - INFO - codeparrot_training - Step 19501: {'lr': 0.0003531618532556378, 'samples': 3744384, 'steps': 19501, 'loss/train': 1.4586199522018433} 01/29/2022 13:54:59 - INFO - codeparrot_training - Step 19502: {'lr': 0.00035314694861826427, 'samples': 3744576, 'steps': 19502, 'loss/train': 1.9487754106521606} 01/29/2022 13:55:03 - INFO - codeparrot_training - Step 19503: {'lr': 0.00035313204353904203, 'samples': 3744768, 'steps': 19503, 'loss/train': 1.8045464158058167} 01/29/2022 13:55:08 - INFO - codeparrot_training - Step 19504: {'lr': 0.0003531171380180348, 'samples': 3744960, 'steps': 19504, 'loss/train': 1.2990846633911133} 01/29/2022 13:55:13 - INFO - codeparrot_training - Step 19505: {'lr': 0.00035310223205530655, 'samples': 3745152, 'steps': 19505, 'loss/train': 1.790789008140564} 01/29/2022 13:55:17 - INFO - codeparrot_training - Step 19506: {'lr': 0.00035308732565092097, 'samples': 3745344, 'steps': 19506, 'loss/train': 1.737772285938263} 01/29/2022 13:55:21 - INFO - codeparrot_training - Step 19507: {'lr': 0.0003530724188049421, 'samples': 3745536, 'steps': 19507, 'loss/train': 2.950739800930023} 01/29/2022 13:55:25 - INFO - codeparrot_training - Step 19508: {'lr': 0.0003530575115174337, 'samples': 3745728, 'steps': 19508, 'loss/train': 1.5188419818878174} 01/29/2022 13:55:30 - INFO - codeparrot_training - Step 19509: {'lr': 0.0003530426037884597, 'samples': 3745920, 'steps': 19509, 'loss/train': 1.4093458950519562} 01/29/2022 13:55:37 - INFO - codeparrot_training - Step 19510: {'lr': 0.00035302769561808387, 'samples': 3746112, 'steps': 19510, 'loss/train': 0.32400885969400406} 01/29/2022 13:55:41 - INFO - codeparrot_training - Step 19511: {'lr': 0.00035301278700637005, 'samples': 3746304, 'steps': 19511, 'loss/train': 2.321252167224884} 01/29/2022 13:55:46 - INFO - codeparrot_training - Step 19512: {'lr': 0.0003529978779533822, 'samples': 3746496, 'steps': 19512, 'loss/train': 1.572445571422577} 01/29/2022 13:55:50 - INFO - codeparrot_training - Step 19513: {'lr': 0.0003529829684591842, 'samples': 3746688, 'steps': 19513, 'loss/train': 1.700684666633606} 01/29/2022 13:55:54 - INFO - codeparrot_training - Step 19514: {'lr': 0.0003529680585238398, 'samples': 3746880, 'steps': 19514, 'loss/train': 1.3998777866363525} 01/29/2022 13:56:00 - INFO - codeparrot_training - Step 19515: {'lr': 0.000352953148147413, 'samples': 3747072, 'steps': 19515, 'loss/train': 0.9174230396747589} 01/29/2022 13:56:04 - INFO - codeparrot_training - Step 19516: {'lr': 0.0003529382373299675, 'samples': 3747264, 'steps': 19516, 'loss/train': 2.1975476145744324} 01/29/2022 13:56:08 - INFO - codeparrot_training - Step 19517: {'lr': 0.0003529233260715673, 'samples': 3747456, 'steps': 19517, 'loss/train': 1.5417950749397278} 01/29/2022 13:56:12 - INFO - codeparrot_training - Step 19518: {'lr': 0.00035290841437227627, 'samples': 3747648, 'steps': 19518, 'loss/train': 2.0201252102851868} 01/29/2022 13:56:17 - INFO - codeparrot_training - Step 19519: {'lr': 0.00035289350223215827, 'samples': 3747840, 'steps': 19519, 'loss/train': 2.1016008853912354} 01/29/2022 13:56:22 - INFO - codeparrot_training - Step 19520: {'lr': 0.00035287858965127723, 'samples': 3748032, 'steps': 19520, 'loss/train': 1.269083708524704} 01/29/2022 13:56:26 - INFO - codeparrot_training - Step 19521: {'lr': 0.0003528636766296969, 'samples': 3748224, 'steps': 19521, 'loss/train': 1.7651302814483643} 01/29/2022 13:56:30 - INFO - codeparrot_training - Step 19522: {'lr': 0.0003528487631674813, 'samples': 3748416, 'steps': 19522, 'loss/train': 1.8210936784744263} 01/29/2022 13:56:35 - INFO - codeparrot_training - Step 19523: {'lr': 0.00035283384926469426, 'samples': 3748608, 'steps': 19523, 'loss/train': 1.4903614819049835} 01/29/2022 13:56:39 - INFO - codeparrot_training - Step 19524: {'lr': 0.00035281893492139965, 'samples': 3748800, 'steps': 19524, 'loss/train': 1.9984921216964722} 01/29/2022 13:56:44 - INFO - codeparrot_training - Step 19525: {'lr': 0.0003528040201376613, 'samples': 3748992, 'steps': 19525, 'loss/train': 1.4449845850467682} 01/29/2022 13:56:49 - INFO - codeparrot_training - Step 19526: {'lr': 0.00035278910491354335, 'samples': 3749184, 'steps': 19526, 'loss/train': 0.8089700639247894} 01/29/2022 13:56:53 - INFO - codeparrot_training - Step 19527: {'lr': 0.0003527741892491093, 'samples': 3749376, 'steps': 19527, 'loss/train': 2.0091476440429688} 01/29/2022 13:56:57 - INFO - codeparrot_training - Step 19528: {'lr': 0.00035275927314442344, 'samples': 3749568, 'steps': 19528, 'loss/train': 1.4588657319545746} 01/29/2022 13:57:01 - INFO - codeparrot_training - Step 19529: {'lr': 0.00035274435659954936, 'samples': 3749760, 'steps': 19529, 'loss/train': 1.7259677052497864} 01/29/2022 13:57:09 - INFO - codeparrot_training - Step 19530: {'lr': 0.00035272943961455106, 'samples': 3749952, 'steps': 19530, 'loss/train': 1.4744846820831299} 01/29/2022 13:57:13 - INFO - codeparrot_training - Step 19531: {'lr': 0.00035271452218949256, 'samples': 3750144, 'steps': 19531, 'loss/train': 2.182523488998413} 01/29/2022 13:57:18 - INFO - codeparrot_training - Step 19532: {'lr': 0.0003526996043244376, 'samples': 3750336, 'steps': 19532, 'loss/train': 2.1006322503089905} 01/29/2022 13:57:22 - INFO - codeparrot_training - Step 19533: {'lr': 0.0003526846860194502, 'samples': 3750528, 'steps': 19533, 'loss/train': 2.132924973964691} 01/29/2022 13:57:27 - INFO - codeparrot_training - Step 19534: {'lr': 0.0003526697672745942, 'samples': 3750720, 'steps': 19534, 'loss/train': 0.8952598571777344} 01/29/2022 13:57:32 - INFO - codeparrot_training - Step 19535: {'lr': 0.0003526548480899335, 'samples': 3750912, 'steps': 19535, 'loss/train': 1.4049686193466187} 01/29/2022 13:57:36 - INFO - codeparrot_training - Step 19536: {'lr': 0.000352639928465532, 'samples': 3751104, 'steps': 19536, 'loss/train': 1.954500138759613} 01/29/2022 13:57:40 - INFO - codeparrot_training - Step 19537: {'lr': 0.00035262500840145365, 'samples': 3751296, 'steps': 19537, 'loss/train': 1.0037781894207} 01/29/2022 13:57:44 - INFO - codeparrot_training - Step 19538: {'lr': 0.00035261008789776237, 'samples': 3751488, 'steps': 19538, 'loss/train': 1.6709182262420654} 01/29/2022 13:57:52 - INFO - codeparrot_training - Step 19539: {'lr': 0.0003525951669545221, 'samples': 3751680, 'steps': 19539, 'loss/train': 1.6458663940429688} 01/29/2022 13:57:56 - INFO - codeparrot_training - Step 19540: {'lr': 0.00035258024557179664, 'samples': 3751872, 'steps': 19540, 'loss/train': 1.3110726177692413} 01/29/2022 13:58:00 - INFO - codeparrot_training - Step 19541: {'lr': 0.00035256532374965, 'samples': 3752064, 'steps': 19541, 'loss/train': 2.868515968322754} 01/29/2022 13:58:04 - INFO - codeparrot_training - Step 19542: {'lr': 0.0003525504014881461, 'samples': 3752256, 'steps': 19542, 'loss/train': 1.891094148159027} 01/29/2022 13:58:08 - INFO - codeparrot_training - Step 19543: {'lr': 0.0003525354787873488, 'samples': 3752448, 'steps': 19543, 'loss/train': 1.2481439709663391} 01/29/2022 13:58:14 - INFO - codeparrot_training - Step 19544: {'lr': 0.0003525205556473221, 'samples': 3752640, 'steps': 19544, 'loss/train': 1.9434349536895752} 01/29/2022 13:58:18 - INFO - codeparrot_training - Step 19545: {'lr': 0.0003525056320681299, 'samples': 3752832, 'steps': 19545, 'loss/train': 0.9212867617607117} 01/29/2022 13:58:22 - INFO - codeparrot_training - Step 19546: {'lr': 0.00035249070804983616, 'samples': 3753024, 'steps': 19546, 'loss/train': 1.4880885779857635} 01/29/2022 13:58:26 - INFO - codeparrot_training - Step 19547: {'lr': 0.00035247578359250473, 'samples': 3753216, 'steps': 19547, 'loss/train': 1.6198936700820923} 01/29/2022 13:58:31 - INFO - codeparrot_training - Step 19548: {'lr': 0.0003524608586961996, 'samples': 3753408, 'steps': 19548, 'loss/train': 1.8888539671897888} 01/29/2022 13:58:36 - INFO - codeparrot_training - Step 19549: {'lr': 0.00035244593336098464, 'samples': 3753600, 'steps': 19549, 'loss/train': 1.8572811484336853} 01/29/2022 13:58:40 - INFO - codeparrot_training - Step 19550: {'lr': 0.0003524310075869239, 'samples': 3753792, 'steps': 19550, 'loss/train': 1.7526002526283264} 01/29/2022 13:58:45 - INFO - codeparrot_training - Step 19551: {'lr': 0.0003524160813740812, 'samples': 3753984, 'steps': 19551, 'loss/train': 0.5165838003158569} 01/29/2022 13:58:49 - INFO - codeparrot_training - Step 19552: {'lr': 0.00035240115472252056, 'samples': 3754176, 'steps': 19552, 'loss/train': 1.802254557609558} 01/29/2022 13:58:53 - INFO - codeparrot_training - Step 19553: {'lr': 0.00035238622763230574, 'samples': 3754368, 'steps': 19553, 'loss/train': 1.5928376913070679} 01/29/2022 13:59:00 - INFO - codeparrot_training - Step 19554: {'lr': 0.000352371300103501, 'samples': 3754560, 'steps': 19554, 'loss/train': 1.9613651633262634} 01/29/2022 13:59:05 - INFO - codeparrot_training - Step 19555: {'lr': 0.0003523563721361701, 'samples': 3754752, 'steps': 19555, 'loss/train': 2.2395138144493103} 01/29/2022 13:59:09 - INFO - codeparrot_training - Step 19556: {'lr': 0.0003523414437303769, 'samples': 3754944, 'steps': 19556, 'loss/train': 1.961573302745819} 01/29/2022 13:59:13 - INFO - codeparrot_training - Step 19557: {'lr': 0.00035232651488618564, 'samples': 3755136, 'steps': 19557, 'loss/train': 1.7777307629585266} 01/29/2022 13:59:17 - INFO - codeparrot_training - Step 19558: {'lr': 0.00035231158560365987, 'samples': 3755328, 'steps': 19558, 'loss/train': 1.2774946689605713} 01/29/2022 13:59:23 - INFO - codeparrot_training - Step 19559: {'lr': 0.0003522966558828639, 'samples': 3755520, 'steps': 19559, 'loss/train': 1.59798002243042} 01/29/2022 13:59:27 - INFO - codeparrot_training - Step 19560: {'lr': 0.00035228172572386146, 'samples': 3755712, 'steps': 19560, 'loss/train': 1.5673606395721436} 01/29/2022 13:59:31 - INFO - codeparrot_training - Step 19561: {'lr': 0.00035226679512671664, 'samples': 3755904, 'steps': 19561, 'loss/train': 2.5425111651420593} 01/29/2022 13:59:35 - INFO - codeparrot_training - Step 19562: {'lr': 0.00035225186409149333, 'samples': 3756096, 'steps': 19562, 'loss/train': 0.9245305359363556} 01/29/2022 13:59:40 - INFO - codeparrot_training - Step 19563: {'lr': 0.00035223693261825554, 'samples': 3756288, 'steps': 19563, 'loss/train': 2.053345799446106} 01/29/2022 13:59:45 - INFO - codeparrot_training - Step 19564: {'lr': 0.0003522220007070671, 'samples': 3756480, 'steps': 19564, 'loss/train': 1.1428393721580505} 01/29/2022 13:59:49 - INFO - codeparrot_training - Step 19565: {'lr': 0.0003522070683579921, 'samples': 3756672, 'steps': 19565, 'loss/train': 0.6288470327854156} 01/29/2022 13:59:54 - INFO - codeparrot_training - Step 19566: {'lr': 0.00035219213557109453, 'samples': 3756864, 'steps': 19566, 'loss/train': 3.0380584001541138} 01/29/2022 13:59:58 - INFO - codeparrot_training - Step 19567: {'lr': 0.00035217720234643823, 'samples': 3757056, 'steps': 19567, 'loss/train': 2.2456308603286743} 01/29/2022 14:00:02 - INFO - codeparrot_training - Step 19568: {'lr': 0.0003521622686840873, 'samples': 3757248, 'steps': 19568, 'loss/train': 1.2071158289909363} 01/29/2022 14:00:10 - INFO - codeparrot_training - Step 19569: {'lr': 0.00035214733458410557, 'samples': 3757440, 'steps': 19569, 'loss/train': 2.0399372577667236} 01/29/2022 14:00:14 - INFO - codeparrot_training - Step 19570: {'lr': 0.00035213240004655714, 'samples': 3757632, 'steps': 19570, 'loss/train': 2.1984211206436157} 01/29/2022 14:00:18 - INFO - codeparrot_training - Step 19571: {'lr': 0.000352117465071506, 'samples': 3757824, 'steps': 19571, 'loss/train': 1.5947125554084778} 01/29/2022 14:00:23 - INFO - codeparrot_training - Step 19572: {'lr': 0.000352102529659016, 'samples': 3758016, 'steps': 19572, 'loss/train': 2.2019042372703552} 01/29/2022 14:00:27 - INFO - codeparrot_training - Step 19573: {'lr': 0.00035208759380915116, 'samples': 3758208, 'steps': 19573, 'loss/train': 1.3847911655902863} 01/29/2022 14:00:32 - INFO - codeparrot_training - Step 19574: {'lr': 0.0003520726575219756, 'samples': 3758400, 'steps': 19574, 'loss/train': 1.9298349022865295} 01/29/2022 14:00:37 - INFO - codeparrot_training - Step 19575: {'lr': 0.00035205772079755307, 'samples': 3758592, 'steps': 19575, 'loss/train': 1.870249629020691} 01/29/2022 14:00:41 - INFO - codeparrot_training - Step 19576: {'lr': 0.0003520427836359477, 'samples': 3758784, 'steps': 19576, 'loss/train': 1.6755763292312622} 01/29/2022 14:00:45 - INFO - codeparrot_training - Step 19577: {'lr': 0.00035202784603722344, 'samples': 3758976, 'steps': 19577, 'loss/train': 2.231200397014618} 01/29/2022 14:00:49 - INFO - codeparrot_training - Step 19578: {'lr': 0.00035201290800144423, 'samples': 3759168, 'steps': 19578, 'loss/train': 1.1307263374328613} 01/29/2022 14:00:57 - INFO - codeparrot_training - Step 19579: {'lr': 0.00035199796952867425, 'samples': 3759360, 'steps': 19579, 'loss/train': 1.7772932052612305} 01/29/2022 14:01:01 - INFO - codeparrot_training - Step 19580: {'lr': 0.0003519830306189773, 'samples': 3759552, 'steps': 19580, 'loss/train': 1.2517442107200623} 01/29/2022 14:01:05 - INFO - codeparrot_training - Step 19581: {'lr': 0.0003519680912724174, 'samples': 3759744, 'steps': 19581, 'loss/train': 2.104509115219116} 01/29/2022 14:01:09 - INFO - codeparrot_training - Step 19582: {'lr': 0.0003519531514890586, 'samples': 3759936, 'steps': 19582, 'loss/train': 2.2618594765663147} 01/29/2022 14:01:14 - INFO - codeparrot_training - Step 19583: {'lr': 0.00035193821126896493, 'samples': 3760128, 'steps': 19583, 'loss/train': 1.5676639080047607} 01/29/2022 14:01:19 - INFO - codeparrot_training - Step 19584: {'lr': 0.0003519232706122002, 'samples': 3760320, 'steps': 19584, 'loss/train': 1.856880784034729} 01/29/2022 14:01:23 - INFO - codeparrot_training - Step 19585: {'lr': 0.0003519083295188287, 'samples': 3760512, 'steps': 19585, 'loss/train': 2.044880211353302} 01/29/2022 14:01:28 - INFO - codeparrot_training - Step 19586: {'lr': 0.0003518933879889142, 'samples': 3760704, 'steps': 19586, 'loss/train': 0.6441984325647354} 01/29/2022 14:01:32 - INFO - codeparrot_training - Step 19587: {'lr': 0.00035187844602252075, 'samples': 3760896, 'steps': 19587, 'loss/train': 2.462242841720581} 01/29/2022 14:01:36 - INFO - codeparrot_training - Step 19588: {'lr': 0.0003518635036197124, 'samples': 3761088, 'steps': 19588, 'loss/train': 4.564357280731201} 01/29/2022 14:01:42 - INFO - codeparrot_training - Step 19589: {'lr': 0.00035184856078055316, 'samples': 3761280, 'steps': 19589, 'loss/train': 2.262763559818268} 01/29/2022 14:01:46 - INFO - codeparrot_training - Step 19590: {'lr': 0.000351833617505107, 'samples': 3761472, 'steps': 19590, 'loss/train': 1.2851955592632294} 01/29/2022 14:01:50 - INFO - codeparrot_training - Step 19591: {'lr': 0.000351818673793438, 'samples': 3761664, 'steps': 19591, 'loss/train': 2.461495578289032} 01/29/2022 14:01:54 - INFO - codeparrot_training - Step 19592: {'lr': 0.00035180372964561013, 'samples': 3761856, 'steps': 19592, 'loss/train': 1.5108845829963684} 01/29/2022 14:01:59 - INFO - codeparrot_training - Step 19593: {'lr': 0.00035178878506168733, 'samples': 3762048, 'steps': 19593, 'loss/train': 0.6586971580982208} 01/29/2022 14:02:04 - INFO - codeparrot_training - Step 19594: {'lr': 0.0003517738400417338, 'samples': 3762240, 'steps': 19594, 'loss/train': 1.4403846859931946} 01/29/2022 14:02:08 - INFO - codeparrot_training - Step 19595: {'lr': 0.0003517588945858134, 'samples': 3762432, 'steps': 19595, 'loss/train': 1.6501041054725647} 01/29/2022 14:02:12 - INFO - codeparrot_training - Step 19596: {'lr': 0.00035174394869399024, 'samples': 3762624, 'steps': 19596, 'loss/train': 2.035598337650299} 01/29/2022 14:02:17 - INFO - codeparrot_training - Step 19597: {'lr': 0.0003517290023663283, 'samples': 3762816, 'steps': 19597, 'loss/train': 1.8477931022644043} 01/29/2022 14:02:21 - INFO - codeparrot_training - Step 19598: {'lr': 0.0003517140556028916, 'samples': 3763008, 'steps': 19598, 'loss/train': 1.9907875657081604} 01/29/2022 14:02:28 - INFO - codeparrot_training - Step 19599: {'lr': 0.0003516991084037442, 'samples': 3763200, 'steps': 19599, 'loss/train': 1.4489312767982483} 01/29/2022 14:02:32 - INFO - codeparrot_training - Step 19600: {'lr': 0.0003516841607689501, 'samples': 3763392, 'steps': 19600, 'loss/train': 1.477241188287735} 01/29/2022 14:02:36 - INFO - codeparrot_training - Step 19601: {'lr': 0.0003516692126985733, 'samples': 3763584, 'steps': 19601, 'loss/train': 2.0068593621253967} 01/29/2022 14:02:41 - INFO - codeparrot_training - Step 19602: {'lr': 0.00035165426419267795, 'samples': 3763776, 'steps': 19602, 'loss/train': 1.024934560060501} 01/29/2022 14:02:45 - INFO - codeparrot_training - Step 19603: {'lr': 0.00035163931525132797, 'samples': 3763968, 'steps': 19603, 'loss/train': 1.5859326124191284} 01/29/2022 14:02:50 - INFO - codeparrot_training - Step 19604: {'lr': 0.00035162436587458744, 'samples': 3764160, 'steps': 19604, 'loss/train': 1.321341723203659} 01/29/2022 14:02:55 - INFO - codeparrot_training - Step 19605: {'lr': 0.00035160941606252044, 'samples': 3764352, 'steps': 19605, 'loss/train': 1.6899343729019165} 01/29/2022 14:02:59 - INFO - codeparrot_training - Step 19606: {'lr': 0.00035159446581519094, 'samples': 3764544, 'steps': 19606, 'loss/train': 1.0988199412822723} 01/29/2022 14:03:03 - INFO - codeparrot_training - Step 19607: {'lr': 0.000351579515132663, 'samples': 3764736, 'steps': 19607, 'loss/train': 1.3990175127983093} 01/29/2022 14:03:08 - INFO - codeparrot_training - Step 19608: {'lr': 0.0003515645640150006, 'samples': 3764928, 'steps': 19608, 'loss/train': 1.5430710911750793} 01/29/2022 14:03:13 - INFO - codeparrot_training - Step 19609: {'lr': 0.000351549612462268, 'samples': 3765120, 'steps': 19609, 'loss/train': 1.920520842075348} 01/29/2022 14:03:17 - INFO - codeparrot_training - Step 19610: {'lr': 0.00035153466047452904, 'samples': 3765312, 'steps': 19610, 'loss/train': 1.442579448223114} 01/29/2022 14:03:21 - INFO - codeparrot_training - Step 19611: {'lr': 0.00035151970805184785, 'samples': 3765504, 'steps': 19611, 'loss/train': 1.3297543823719025} 01/29/2022 14:03:26 - INFO - codeparrot_training - Step 19612: {'lr': 0.00035150475519428844, 'samples': 3765696, 'steps': 19612, 'loss/train': 2.0274494290351868} 01/29/2022 14:03:30 - INFO - codeparrot_training - Step 19613: {'lr': 0.00035148980190191485, 'samples': 3765888, 'steps': 19613, 'loss/train': 1.9307114481925964} 01/29/2022 14:03:37 - INFO - codeparrot_training - Step 19614: {'lr': 0.00035147484817479126, 'samples': 3766080, 'steps': 19614, 'loss/train': 1.6345120668411255} 01/29/2022 14:03:41 - INFO - codeparrot_training - Step 19615: {'lr': 0.00035145989401298163, 'samples': 3766272, 'steps': 19615, 'loss/train': 1.7351179718971252} 01/29/2022 14:03:46 - INFO - codeparrot_training - Step 19616: {'lr': 0.00035144493941655, 'samples': 3766464, 'steps': 19616, 'loss/train': 1.900309145450592} 01/29/2022 14:03:50 - INFO - codeparrot_training - Step 19617: {'lr': 0.0003514299843855605, 'samples': 3766656, 'steps': 19617, 'loss/train': 2.0559099912643433} 01/29/2022 14:03:54 - INFO - codeparrot_training - Step 19618: {'lr': 0.0003514150289200771, 'samples': 3766848, 'steps': 19618, 'loss/train': 1.8528833985328674} 01/29/2022 14:03:59 - INFO - codeparrot_training - Step 19619: {'lr': 0.00035140007302016395, 'samples': 3767040, 'steps': 19619, 'loss/train': 1.444758027791977} 01/29/2022 14:04:04 - INFO - codeparrot_training - Step 19620: {'lr': 0.0003513851166858851, 'samples': 3767232, 'steps': 19620, 'loss/train': 1.6924585103988647} 01/29/2022 14:04:08 - INFO - codeparrot_training - Step 19621: {'lr': 0.0003513701599173046, 'samples': 3767424, 'steps': 19621, 'loss/train': 2.0144256949424744} 01/29/2022 14:04:12 - INFO - codeparrot_training - Step 19622: {'lr': 0.0003513552027144865, 'samples': 3767616, 'steps': 19622, 'loss/train': 0.7530358135700226} 01/29/2022 14:04:16 - INFO - codeparrot_training - Step 19623: {'lr': 0.00035134024507749487, 'samples': 3767808, 'steps': 19623, 'loss/train': 1.4464199244976044} 01/29/2022 14:04:24 - INFO - codeparrot_training - Step 19624: {'lr': 0.0003513252870063939, 'samples': 3768000, 'steps': 19624, 'loss/train': 0.1477404423058033} 01/29/2022 14:04:28 - INFO - codeparrot_training - Step 19625: {'lr': 0.00035131032850124744, 'samples': 3768192, 'steps': 19625, 'loss/train': 1.6065157055854797} 01/29/2022 14:04:32 - INFO - codeparrot_training - Step 19626: {'lr': 0.0003512953695621198, 'samples': 3768384, 'steps': 19626, 'loss/train': 1.60177481174469} 01/29/2022 14:04:36 - INFO - codeparrot_training - Step 19627: {'lr': 0.00035128041018907496, 'samples': 3768576, 'steps': 19627, 'loss/train': 1.4664024710655212} 01/29/2022 14:04:41 - INFO - codeparrot_training - Step 19628: {'lr': 0.000351265450382177, 'samples': 3768768, 'steps': 19628, 'loss/train': 1.5226516127586365} 01/29/2022 14:04:46 - INFO - codeparrot_training - Step 19629: {'lr': 0.00035125049014148995, 'samples': 3768960, 'steps': 19629, 'loss/train': 1.6171470880508423} 01/29/2022 14:04:50 - INFO - codeparrot_training - Step 19630: {'lr': 0.000351235529467078, 'samples': 3769152, 'steps': 19630, 'loss/train': 0.9177548289299011} 01/29/2022 14:04:54 - INFO - codeparrot_training - Step 19631: {'lr': 0.00035122056835900517, 'samples': 3769344, 'steps': 19631, 'loss/train': 0.7838274836540222} 01/29/2022 14:04:59 - INFO - codeparrot_training - Step 19632: {'lr': 0.0003512056068173356, 'samples': 3769536, 'steps': 19632, 'loss/train': 1.3559203147888184} 01/29/2022 14:05:03 - INFO - codeparrot_training - Step 19633: {'lr': 0.00035119064484213333, 'samples': 3769728, 'steps': 19633, 'loss/train': 1.542752981185913} 01/29/2022 14:05:08 - INFO - codeparrot_training - Step 19634: {'lr': 0.0003511756824334625, 'samples': 3769920, 'steps': 19634, 'loss/train': 5.942113280296326} 01/29/2022 14:05:12 - INFO - codeparrot_training - Step 19635: {'lr': 0.0003511607195913872, 'samples': 3770112, 'steps': 19635, 'loss/train': 0.9757825434207916} 01/29/2022 14:05:17 - INFO - codeparrot_training - Step 19636: {'lr': 0.00035114575631597136, 'samples': 3770304, 'steps': 19636, 'loss/train': 1.5466904640197754} 01/29/2022 14:05:21 - INFO - codeparrot_training - Step 19637: {'lr': 0.0003511307926072793, 'samples': 3770496, 'steps': 19637, 'loss/train': 1.3602662980556488} 01/29/2022 14:05:25 - INFO - codeparrot_training - Step 19638: {'lr': 0.00035111582846537507, 'samples': 3770688, 'steps': 19638, 'loss/train': 1.5489512085914612} 01/29/2022 14:05:32 - INFO - codeparrot_training - Step 19639: {'lr': 0.00035110086389032264, 'samples': 3770880, 'steps': 19639, 'loss/train': 1.9342172741889954} 01/29/2022 14:05:36 - INFO - codeparrot_training - Step 19640: {'lr': 0.0003510858988821863, 'samples': 3771072, 'steps': 19640, 'loss/train': 2.037616789340973} 01/29/2022 14:05:41 - INFO - codeparrot_training - Step 19641: {'lr': 0.00035107093344103, 'samples': 3771264, 'steps': 19641, 'loss/train': 1.6552891731262207} 01/29/2022 14:05:45 - INFO - codeparrot_training - Step 19642: {'lr': 0.000351055967566918, 'samples': 3771456, 'steps': 19642, 'loss/train': 2.434575140476227} 01/29/2022 14:05:49 - INFO - codeparrot_training - Step 19643: {'lr': 0.0003510410012599142, 'samples': 3771648, 'steps': 19643, 'loss/train': 1.7529621720314026} 01/29/2022 14:05:55 - INFO - codeparrot_training - Step 19644: {'lr': 0.00035102603452008297, 'samples': 3771840, 'steps': 19644, 'loss/train': 1.9989604353904724} 01/29/2022 14:05:59 - INFO - codeparrot_training - Step 19645: {'lr': 0.00035101106734748824, 'samples': 3772032, 'steps': 19645, 'loss/train': 0.5493624061346054} 01/29/2022 14:06:03 - INFO - codeparrot_training - Step 19646: {'lr': 0.00035099609974219417, 'samples': 3772224, 'steps': 19646, 'loss/train': 1.8737175464630127} 01/29/2022 14:06:07 - INFO - codeparrot_training - Step 19647: {'lr': 0.00035098113170426484, 'samples': 3772416, 'steps': 19647, 'loss/train': 1.8729409575462341} 01/29/2022 14:06:12 - INFO - codeparrot_training - Step 19648: {'lr': 0.0003509661632337645, 'samples': 3772608, 'steps': 19648, 'loss/train': 1.638586163520813} 01/29/2022 14:06:17 - INFO - codeparrot_training - Step 19649: {'lr': 0.00035095119433075706, 'samples': 3772800, 'steps': 19649, 'loss/train': 1.0321412980556488} 01/29/2022 14:06:21 - INFO - codeparrot_training - Step 19650: {'lr': 0.00035093622499530677, 'samples': 3772992, 'steps': 19650, 'loss/train': 2.141265571117401} 01/29/2022 14:06:25 - INFO - codeparrot_training - Step 19651: {'lr': 0.0003509212552274778, 'samples': 3773184, 'steps': 19651, 'loss/train': 2.349947512149811} 01/29/2022 14:06:30 - INFO - codeparrot_training - Step 19652: {'lr': 0.0003509062850273342, 'samples': 3773376, 'steps': 19652, 'loss/train': 2.5071138739585876} 01/29/2022 14:06:34 - INFO - codeparrot_training - Step 19653: {'lr': 0.00035089131439494013, 'samples': 3773568, 'steps': 19653, 'loss/train': 1.833016276359558} 01/29/2022 14:06:39 - INFO - codeparrot_training - Step 19654: {'lr': 0.00035087634333035966, 'samples': 3773760, 'steps': 19654, 'loss/train': 2.0072457790374756} 01/29/2022 14:06:43 - INFO - codeparrot_training - Step 19655: {'lr': 0.00035086137183365707, 'samples': 3773952, 'steps': 19655, 'loss/train': 1.8549497723579407} 01/29/2022 14:06:48 - INFO - codeparrot_training - Step 19656: {'lr': 0.0003508463999048963, 'samples': 3774144, 'steps': 19656, 'loss/train': 1.7396131753921509} 01/29/2022 14:06:52 - INFO - codeparrot_training - Step 19657: {'lr': 0.0003508314275441416, 'samples': 3774336, 'steps': 19657, 'loss/train': 1.1820440590381622} 01/29/2022 14:06:56 - INFO - codeparrot_training - Step 19658: {'lr': 0.0003508164547514571, 'samples': 3774528, 'steps': 19658, 'loss/train': 2.3902164101600647} 01/29/2022 14:07:03 - INFO - codeparrot_training - Step 19659: {'lr': 0.00035080148152690687, 'samples': 3774720, 'steps': 19659, 'loss/train': 2.453500807285309} 01/29/2022 14:07:07 - INFO - codeparrot_training - Step 19660: {'lr': 0.0003507865078705551, 'samples': 3774912, 'steps': 19660, 'loss/train': 2.155257225036621} 01/29/2022 14:07:12 - INFO - codeparrot_training - Step 19661: {'lr': 0.00035077153378246603, 'samples': 3775104, 'steps': 19661, 'loss/train': 2.5479875206947327} 01/29/2022 14:07:16 - INFO - codeparrot_training - Step 19662: {'lr': 0.0003507565592627036, 'samples': 3775296, 'steps': 19662, 'loss/train': 0.3170679062604904} 01/29/2022 14:07:20 - INFO - codeparrot_training - Step 19663: {'lr': 0.0003507415843113321, 'samples': 3775488, 'steps': 19663, 'loss/train': 1.320829063653946} 01/29/2022 14:07:25 - INFO - codeparrot_training - Step 19664: {'lr': 0.00035072660892841566, 'samples': 3775680, 'steps': 19664, 'loss/train': 1.0738238990306854} 01/29/2022 14:07:30 - INFO - codeparrot_training - Step 19665: {'lr': 0.0003507116331140184, 'samples': 3775872, 'steps': 19665, 'loss/train': 1.5602524280548096} 01/29/2022 14:07:34 - INFO - codeparrot_training - Step 19666: {'lr': 0.00035069665686820453, 'samples': 3776064, 'steps': 19666, 'loss/train': 2.254408121109009} 01/29/2022 14:07:38 - INFO - codeparrot_training - Step 19667: {'lr': 0.0003506816801910381, 'samples': 3776256, 'steps': 19667, 'loss/train': 1.4906377494335175} 01/29/2022 14:07:42 - INFO - codeparrot_training - Step 19668: {'lr': 0.00035066670308258333, 'samples': 3776448, 'steps': 19668, 'loss/train': 1.939783751964569} 01/29/2022 14:07:50 - INFO - codeparrot_training - Step 19669: {'lr': 0.00035065172554290435, 'samples': 3776640, 'steps': 19669, 'loss/train': 1.5260143876075745} 01/29/2022 14:07:54 - INFO - codeparrot_training - Step 19670: {'lr': 0.0003506367475720654, 'samples': 3776832, 'steps': 19670, 'loss/train': 1.8460993766784668} 01/29/2022 14:07:58 - INFO - codeparrot_training - Step 19671: {'lr': 0.0003506217691701305, 'samples': 3777024, 'steps': 19671, 'loss/train': 0.638056755065918} 01/29/2022 14:08:03 - INFO - codeparrot_training - Step 19672: {'lr': 0.000350606790337164, 'samples': 3777216, 'steps': 19672, 'loss/train': 1.8427913188934326} 01/29/2022 14:08:07 - INFO - codeparrot_training - Step 19673: {'lr': 0.00035059181107322977, 'samples': 3777408, 'steps': 19673, 'loss/train': 2.1983261704444885} 01/29/2022 14:08:12 - INFO - codeparrot_training - Step 19674: {'lr': 0.00035057683137839236, 'samples': 3777600, 'steps': 19674, 'loss/train': 1.5611692070960999} 01/29/2022 14:08:17 - INFO - codeparrot_training - Step 19675: {'lr': 0.00035056185125271566, 'samples': 3777792, 'steps': 19675, 'loss/train': 1.0163801908493042} 01/29/2022 14:08:21 - INFO - codeparrot_training - Step 19676: {'lr': 0.0003505468706962639, 'samples': 3777984, 'steps': 19676, 'loss/train': 1.4691177606582642} 01/29/2022 14:08:25 - INFO - codeparrot_training - Step 19677: {'lr': 0.0003505318897091013, 'samples': 3778176, 'steps': 19677, 'loss/train': 1.5172242522239685} 01/29/2022 14:08:29 - INFO - codeparrot_training - Step 19678: {'lr': 0.000350516908291292, 'samples': 3778368, 'steps': 19678, 'loss/train': 1.5670557618141174} 01/29/2022 14:08:35 - INFO - codeparrot_training - Step 19679: {'lr': 0.00035050192644290023, 'samples': 3778560, 'steps': 19679, 'loss/train': 1.0323463082313538} 01/29/2022 14:08:39 - INFO - codeparrot_training - Step 19680: {'lr': 0.00035048694416399005, 'samples': 3778752, 'steps': 19680, 'loss/train': 1.4877442717552185} 01/29/2022 14:08:43 - INFO - codeparrot_training - Step 19681: {'lr': 0.0003504719614546258, 'samples': 3778944, 'steps': 19681, 'loss/train': 1.1043822169303894} 01/29/2022 14:08:47 - INFO - codeparrot_training - Step 19682: {'lr': 0.00035045697831487146, 'samples': 3779136, 'steps': 19682, 'loss/train': 1.0860323309898376} 01/29/2022 14:08:52 - INFO - codeparrot_training - Step 19683: {'lr': 0.00035044199474479137, 'samples': 3779328, 'steps': 19683, 'loss/train': 1.689866065979004} 01/29/2022 14:08:59 - INFO - codeparrot_training - Step 19684: {'lr': 0.00035042701074444965, 'samples': 3779520, 'steps': 19684, 'loss/train': 1.9691861271858215} 01/29/2022 14:09:03 - INFO - codeparrot_training - Step 19685: {'lr': 0.00035041202631391056, 'samples': 3779712, 'steps': 19685, 'loss/train': 2.34195077419281} 01/29/2022 14:09:07 - INFO - codeparrot_training - Step 19686: {'lr': 0.0003503970414532382, 'samples': 3779904, 'steps': 19686, 'loss/train': 1.8718124628067017} 01/29/2022 14:09:12 - INFO - codeparrot_training - Step 19687: {'lr': 0.00035038205616249674, 'samples': 3780096, 'steps': 19687, 'loss/train': 0.7481386810541153} 01/29/2022 14:09:16 - INFO - codeparrot_training - Step 19688: {'lr': 0.00035036707044175055, 'samples': 3780288, 'steps': 19688, 'loss/train': 0.9949378967285156} 01/29/2022 14:09:21 - INFO - codeparrot_training - Step 19689: {'lr': 0.00035035208429106356, 'samples': 3780480, 'steps': 19689, 'loss/train': 1.9216840267181396} 01/29/2022 14:09:25 - INFO - codeparrot_training - Step 19690: {'lr': 0.0003503370977105002, 'samples': 3780672, 'steps': 19690, 'loss/train': 1.0327304899692535} 01/29/2022 14:09:30 - INFO - codeparrot_training - Step 19691: {'lr': 0.00035032211070012455, 'samples': 3780864, 'steps': 19691, 'loss/train': 1.9824092388153076} 01/29/2022 14:09:34 - INFO - codeparrot_training - Step 19692: {'lr': 0.00035030712326000084, 'samples': 3781056, 'steps': 19692, 'loss/train': 2.190586030483246} 01/29/2022 14:09:38 - INFO - codeparrot_training - Step 19693: {'lr': 0.00035029213539019324, 'samples': 3781248, 'steps': 19693, 'loss/train': 1.880137324333191} 01/29/2022 14:09:44 - INFO - codeparrot_training - Step 19694: {'lr': 0.000350277147090766, 'samples': 3781440, 'steps': 19694, 'loss/train': 1.6619719862937927} 01/29/2022 14:09:48 - INFO - codeparrot_training - Step 19695: {'lr': 0.0003502621583617833, 'samples': 3781632, 'steps': 19695, 'loss/train': 1.5852715373039246} 01/29/2022 14:09:52 - INFO - codeparrot_training - Step 19696: {'lr': 0.00035024716920330933, 'samples': 3781824, 'steps': 19696, 'loss/train': 1.832072675228119} 01/29/2022 14:09:56 - INFO - codeparrot_training - Step 19697: {'lr': 0.0003502321796154084, 'samples': 3782016, 'steps': 19697, 'loss/train': 1.6952313780784607} 01/29/2022 14:10:01 - INFO - codeparrot_training - Step 19698: {'lr': 0.00035021718959814453, 'samples': 3782208, 'steps': 19698, 'loss/train': 1.9832067489624023} 01/29/2022 14:10:09 - INFO - codeparrot_training - Step 19699: {'lr': 0.0003502021991515821, 'samples': 3782400, 'steps': 19699, 'loss/train': 2.1510306000709534} 01/29/2022 14:10:13 - INFO - codeparrot_training - Step 19700: {'lr': 0.0003501872082757852, 'samples': 3782592, 'steps': 19700, 'loss/train': 2.2429862022399902} 01/29/2022 14:10:17 - INFO - codeparrot_training - Step 19701: {'lr': 0.00035017221697081826, 'samples': 3782784, 'steps': 19701, 'loss/train': 0.63504758477211} 01/29/2022 14:10:21 - INFO - codeparrot_training - Step 19702: {'lr': 0.0003501572252367452, 'samples': 3782976, 'steps': 19702, 'loss/train': 1.792928695678711} 01/29/2022 14:10:26 - INFO - codeparrot_training - Step 19703: {'lr': 0.00035014223307363045, 'samples': 3783168, 'steps': 19703, 'loss/train': 2.1419926285743713} 01/29/2022 14:10:31 - INFO - codeparrot_training - Step 19704: {'lr': 0.0003501272404815382, 'samples': 3783360, 'steps': 19704, 'loss/train': 1.6062076091766357} 01/29/2022 14:10:35 - INFO - codeparrot_training - Step 19705: {'lr': 0.0003501122474605326, 'samples': 3783552, 'steps': 19705, 'loss/train': 2.02508944272995} 01/29/2022 14:10:40 - INFO - codeparrot_training - Step 19706: {'lr': 0.00035009725401067795, 'samples': 3783744, 'steps': 19706, 'loss/train': 1.706768810749054} 01/29/2022 14:10:44 - INFO - codeparrot_training - Step 19707: {'lr': 0.00035008226013203845, 'samples': 3783936, 'steps': 19707, 'loss/train': 1.3412852883338928} 01/29/2022 14:10:48 - INFO - codeparrot_training - Step 19708: {'lr': 0.0003500672658246783, 'samples': 3784128, 'steps': 19708, 'loss/train': 1.7725866436958313} 01/29/2022 14:10:53 - INFO - codeparrot_training - Step 19709: {'lr': 0.0003500522710886618, 'samples': 3784320, 'steps': 19709, 'loss/train': 1.266708254814148} 01/29/2022 14:10:58 - INFO - codeparrot_training - Step 19710: {'lr': 0.0003500372759240531, 'samples': 3784512, 'steps': 19710, 'loss/train': 1.968526303768158} 01/29/2022 14:11:02 - INFO - codeparrot_training - Step 19711: {'lr': 0.00035002228033091643, 'samples': 3784704, 'steps': 19711, 'loss/train': 1.702824890613556} 01/29/2022 14:11:06 - INFO - codeparrot_training - Step 19712: {'lr': 0.00035000728430931616, 'samples': 3784896, 'steps': 19712, 'loss/train': 1.7813449501991272} 01/29/2022 14:11:10 - INFO - codeparrot_training - Step 19713: {'lr': 0.00034999228785931644, 'samples': 3785088, 'steps': 19713, 'loss/train': 1.4843980371952057} 01/29/2022 14:11:17 - INFO - codeparrot_training - Step 19714: {'lr': 0.0003499772909809815, 'samples': 3785280, 'steps': 19714, 'loss/train': 1.678901731967926} 01/29/2022 14:11:22 - INFO - codeparrot_training - Step 19715: {'lr': 0.0003499622936743756, 'samples': 3785472, 'steps': 19715, 'loss/train': 1.3774671256542206} 01/29/2022 14:11:26 - INFO - codeparrot_training - Step 19716: {'lr': 0.0003499472959395629, 'samples': 3785664, 'steps': 19716, 'loss/train': 1.3773882687091827} 01/29/2022 14:11:30 - INFO - codeparrot_training - Step 19717: {'lr': 0.00034993229777660785, 'samples': 3785856, 'steps': 19717, 'loss/train': 1.3572555184364319} 01/29/2022 14:11:34 - INFO - codeparrot_training - Step 19718: {'lr': 0.0003499172991855744, 'samples': 3786048, 'steps': 19718, 'loss/train': 1.7865843772888184} 01/29/2022 14:11:40 - INFO - codeparrot_training - Step 19719: {'lr': 0.00034990230016652713, 'samples': 3786240, 'steps': 19719, 'loss/train': 0.6274407505989075} 01/29/2022 14:11:44 - INFO - codeparrot_training - Step 19720: {'lr': 0.00034988730071953, 'samples': 3786432, 'steps': 19720, 'loss/train': 2.1880613565444946} 01/29/2022 14:11:48 - INFO - codeparrot_training - Step 19721: {'lr': 0.0003498723008446475, 'samples': 3786624, 'steps': 19721, 'loss/train': 2.022126317024231} 01/29/2022 14:11:53 - INFO - codeparrot_training - Step 19722: {'lr': 0.0003498573005419437, 'samples': 3786816, 'steps': 19722, 'loss/train': 2.326438844203949} 01/29/2022 14:11:57 - INFO - codeparrot_training - Step 19723: {'lr': 0.000349842299811483, 'samples': 3787008, 'steps': 19723, 'loss/train': 1.209514081478119} 01/29/2022 14:12:02 - INFO - codeparrot_training - Step 19724: {'lr': 0.00034982729865332953, 'samples': 3787200, 'steps': 19724, 'loss/train': 1.792485237121582} 01/29/2022 14:12:06 - INFO - codeparrot_training - Step 19725: {'lr': 0.00034981229706754755, 'samples': 3787392, 'steps': 19725, 'loss/train': 2.4344456791877747} 01/29/2022 14:12:11 - INFO - codeparrot_training - Step 19726: {'lr': 0.0003497972950542015, 'samples': 3787584, 'steps': 19726, 'loss/train': 1.4639666676521301} 01/29/2022 14:12:15 - INFO - codeparrot_training - Step 19727: {'lr': 0.0003497822926133555, 'samples': 3787776, 'steps': 19727, 'loss/train': 2.5094606280326843} 01/29/2022 14:12:19 - INFO - codeparrot_training - Step 19728: {'lr': 0.0003497672897450738, 'samples': 3787968, 'steps': 19728, 'loss/train': 2.11323881149292} 01/29/2022 14:12:26 - INFO - codeparrot_training - Step 19729: {'lr': 0.0003497522864494208, 'samples': 3788160, 'steps': 19729, 'loss/train': 1.4131785035133362} 01/29/2022 14:12:30 - INFO - codeparrot_training - Step 19730: {'lr': 0.0003497372827264606, 'samples': 3788352, 'steps': 19730, 'loss/train': 1.2618277072906494} 01/29/2022 14:12:35 - INFO - codeparrot_training - Step 19731: {'lr': 0.0003497222785762576, 'samples': 3788544, 'steps': 19731, 'loss/train': 1.4416040182113647} 01/29/2022 14:12:39 - INFO - codeparrot_training - Step 19732: {'lr': 0.000349707273998876, 'samples': 3788736, 'steps': 19732, 'loss/train': 1.728092908859253} 01/29/2022 14:12:43 - INFO - codeparrot_training - Step 19733: {'lr': 0.00034969226899438013, 'samples': 3788928, 'steps': 19733, 'loss/train': 2.0034152269363403} 01/29/2022 14:12:48 - INFO - codeparrot_training - Step 19734: {'lr': 0.00034967726356283416, 'samples': 3789120, 'steps': 19734, 'loss/train': 1.7299422025680542} 01/29/2022 14:12:53 - INFO - codeparrot_training - Step 19735: {'lr': 0.00034966225770430244, 'samples': 3789312, 'steps': 19735, 'loss/train': 1.6099339127540588} 01/29/2022 14:12:57 - INFO - codeparrot_training - Step 19736: {'lr': 0.00034964725141884936, 'samples': 3789504, 'steps': 19736, 'loss/train': 1.876524031162262} 01/29/2022 14:13:01 - INFO - codeparrot_training - Step 19737: {'lr': 0.000349632244706539, 'samples': 3789696, 'steps': 19737, 'loss/train': 1.48909592628479} 01/29/2022 14:13:05 - INFO - codeparrot_training - Step 19738: {'lr': 0.0003496172375674358, 'samples': 3789888, 'steps': 19738, 'loss/train': 1.7747041583061218} 01/29/2022 14:13:10 - INFO - codeparrot_training - Step 19739: {'lr': 0.0003496022300016039, 'samples': 3790080, 'steps': 19739, 'loss/train': 1.3411322236061096} 01/29/2022 14:13:15 - INFO - codeparrot_training - Step 19740: {'lr': 0.00034958722200910777, 'samples': 3790272, 'steps': 19740, 'loss/train': 1.6153925657272339} 01/29/2022 14:13:19 - INFO - codeparrot_training - Step 19741: {'lr': 0.00034957221359001154, 'samples': 3790464, 'steps': 19741, 'loss/train': 0.8570278286933899} 01/29/2022 14:13:23 - INFO - codeparrot_training - Step 19742: {'lr': 0.0003495572047443796, 'samples': 3790656, 'steps': 19742, 'loss/train': 1.112043857574463} 01/29/2022 14:13:27 - INFO - codeparrot_training - Step 19743: {'lr': 0.0003495421954722762, 'samples': 3790848, 'steps': 19743, 'loss/train': 2.697856843471527} 01/29/2022 14:13:34 - INFO - codeparrot_training - Step 19744: {'lr': 0.0003495271857737657, 'samples': 3791040, 'steps': 19744, 'loss/train': 1.1792753040790558} 01/29/2022 14:13:38 - INFO - codeparrot_training - Step 19745: {'lr': 0.00034951217564891226, 'samples': 3791232, 'steps': 19745, 'loss/train': 2.1194894313812256} 01/29/2022 14:13:42 - INFO - codeparrot_training - Step 19746: {'lr': 0.0003494971650977802, 'samples': 3791424, 'steps': 19746, 'loss/train': 1.548491656780243} 01/29/2022 14:13:46 - INFO - codeparrot_training - Step 19747: {'lr': 0.00034948215412043405, 'samples': 3791616, 'steps': 19747, 'loss/train': 1.5777938961982727} 01/29/2022 14:13:52 - INFO - codeparrot_training - Step 19748: {'lr': 0.00034946714271693783, 'samples': 3791808, 'steps': 19748, 'loss/train': 1.7935116291046143} 01/29/2022 14:13:56 - INFO - codeparrot_training - Step 19749: {'lr': 0.00034945213088735595, 'samples': 3792000, 'steps': 19749, 'loss/train': 1.9397994875907898} 01/29/2022 14:14:00 - INFO - codeparrot_training - Step 19750: {'lr': 0.00034943711863175277, 'samples': 3792192, 'steps': 19750, 'loss/train': 0.8046764731407166} 01/29/2022 14:14:05 - INFO - codeparrot_training - Step 19751: {'lr': 0.0003494221059501925, 'samples': 3792384, 'steps': 19751, 'loss/train': 1.9519981741905212} 01/29/2022 14:14:09 - INFO - codeparrot_training - Step 19752: {'lr': 0.0003494070928427395, 'samples': 3792576, 'steps': 19752, 'loss/train': 0.1673419252038002} 01/29/2022 14:14:15 - INFO - codeparrot_training - Step 19753: {'lr': 0.00034939207930945816, 'samples': 3792768, 'steps': 19753, 'loss/train': 1.7891263961791992} 01/29/2022 14:14:20 - INFO - codeparrot_training - Step 19754: {'lr': 0.00034937706535041263, 'samples': 3792960, 'steps': 19754, 'loss/train': 2.340876281261444} 01/29/2022 14:14:24 - INFO - codeparrot_training - Step 19755: {'lr': 0.0003493620509656674, 'samples': 3793152, 'steps': 19755, 'loss/train': 1.2606218755245209} 01/29/2022 14:14:28 - INFO - codeparrot_training - Step 19756: {'lr': 0.00034934703615528657, 'samples': 3793344, 'steps': 19756, 'loss/train': 1.4011223316192627} 01/29/2022 14:14:32 - INFO - codeparrot_training - Step 19757: {'lr': 0.0003493320209193347, 'samples': 3793536, 'steps': 19757, 'loss/train': 1.8545182943344116} 01/29/2022 14:14:38 - INFO - codeparrot_training - Step 19758: {'lr': 0.0003493170052578759, 'samples': 3793728, 'steps': 19758, 'loss/train': 1.5312845706939697} 01/29/2022 14:14:42 - INFO - codeparrot_training - Step 19759: {'lr': 0.00034930198917097467, 'samples': 3793920, 'steps': 19759, 'loss/train': 7.437383651733398} 01/29/2022 14:14:46 - INFO - codeparrot_training - Step 19760: {'lr': 0.00034928697265869515, 'samples': 3794112, 'steps': 19760, 'loss/train': 1.2777235507965088} 01/29/2022 14:14:51 - INFO - codeparrot_training - Step 19761: {'lr': 0.0003492719557211018, 'samples': 3794304, 'steps': 19761, 'loss/train': 1.4565538465976715} 01/29/2022 14:14:55 - INFO - codeparrot_training - Step 19762: {'lr': 0.0003492569383582589, 'samples': 3794496, 'steps': 19762, 'loss/train': 0.5189431607723236} 01/29/2022 14:14:59 - INFO - codeparrot_training - Step 19763: {'lr': 0.0003492419205702309, 'samples': 3794688, 'steps': 19763, 'loss/train': 1.2124149799346924} 01/29/2022 14:15:05 - INFO - codeparrot_training - Step 19764: {'lr': 0.00034922690235708197, 'samples': 3794880, 'steps': 19764, 'loss/train': 1.5172428488731384} 01/29/2022 14:15:09 - INFO - codeparrot_training - Step 19765: {'lr': 0.0003492118837188765, 'samples': 3795072, 'steps': 19765, 'loss/train': 2.0008116960525513} 01/29/2022 14:15:13 - INFO - codeparrot_training - Step 19766: {'lr': 0.0003491968646556788, 'samples': 3795264, 'steps': 19766, 'loss/train': 1.1429085731506348} 01/29/2022 14:15:17 - INFO - codeparrot_training - Step 19767: {'lr': 0.00034918184516755324, 'samples': 3795456, 'steps': 19767, 'loss/train': 0.8826964795589447} 01/29/2022 14:15:22 - INFO - codeparrot_training - Step 19768: {'lr': 0.00034916682525456416, 'samples': 3795648, 'steps': 19768, 'loss/train': 6.847780466079712} 01/29/2022 14:15:28 - INFO - codeparrot_training - Step 19769: {'lr': 0.00034915180491677583, 'samples': 3795840, 'steps': 19769, 'loss/train': 1.028709590435028} 01/29/2022 14:15:32 - INFO - codeparrot_training - Step 19770: {'lr': 0.00034913678415425276, 'samples': 3796032, 'steps': 19770, 'loss/train': 0.7736815810203552} 01/29/2022 14:15:36 - INFO - codeparrot_training - Step 19771: {'lr': 0.00034912176296705903, 'samples': 3796224, 'steps': 19771, 'loss/train': 2.0578147172927856} 01/29/2022 14:15:41 - INFO - codeparrot_training - Step 19772: {'lr': 0.00034910674135525926, 'samples': 3796416, 'steps': 19772, 'loss/train': 1.5663599967956543} 01/29/2022 14:15:45 - INFO - codeparrot_training - Step 19773: {'lr': 0.0003490917193189177, 'samples': 3796608, 'steps': 19773, 'loss/train': 4.390951037406921} 01/29/2022 14:15:51 - INFO - codeparrot_training - Step 19774: {'lr': 0.00034907669685809855, 'samples': 3796800, 'steps': 19774, 'loss/train': 1.921850860118866} 01/29/2022 14:15:55 - INFO - codeparrot_training - Step 19775: {'lr': 0.0003490616739728664, 'samples': 3796992, 'steps': 19775, 'loss/train': 1.1326653957366943} 01/29/2022 14:15:59 - INFO - codeparrot_training - Step 19776: {'lr': 0.00034904665066328545, 'samples': 3797184, 'steps': 19776, 'loss/train': 1.024182915687561} 01/29/2022 14:16:03 - INFO - codeparrot_training - Step 19777: {'lr': 0.00034903162692942013, 'samples': 3797376, 'steps': 19777, 'loss/train': 1.0490966141223907} 01/29/2022 14:16:08 - INFO - codeparrot_training - Step 19778: {'lr': 0.0003490166027713348, 'samples': 3797568, 'steps': 19778, 'loss/train': 1.2744774520397186} 01/29/2022 14:16:13 - INFO - codeparrot_training - Step 19779: {'lr': 0.0003490015781890937, 'samples': 3797760, 'steps': 19779, 'loss/train': 1.5149382948875427} 01/29/2022 14:16:17 - INFO - codeparrot_training - Step 19780: {'lr': 0.00034898655318276134, 'samples': 3797952, 'steps': 19780, 'loss/train': 1.7454552054405212} 01/29/2022 14:16:22 - INFO - codeparrot_training - Step 19781: {'lr': 0.00034897152775240206, 'samples': 3798144, 'steps': 19781, 'loss/train': 1.6790614128112793} 01/29/2022 14:16:26 - INFO - codeparrot_training - Step 19782: {'lr': 0.00034895650189808005, 'samples': 3798336, 'steps': 19782, 'loss/train': 2.2933799028396606} 01/29/2022 14:16:30 - INFO - codeparrot_training - Step 19783: {'lr': 0.0003489414756198598, 'samples': 3798528, 'steps': 19783, 'loss/train': 1.748346984386444} 01/29/2022 14:16:35 - INFO - codeparrot_training - Step 19784: {'lr': 0.0003489264489178058, 'samples': 3798720, 'steps': 19784, 'loss/train': 0.8759416043758392} 01/29/2022 14:16:40 - INFO - codeparrot_training - Step 19785: {'lr': 0.0003489114217919823, 'samples': 3798912, 'steps': 19785, 'loss/train': 1.6516976952552795} 01/29/2022 14:16:44 - INFO - codeparrot_training - Step 19786: {'lr': 0.00034889639424245364, 'samples': 3799104, 'steps': 19786, 'loss/train': 1.7406084537506104} 01/29/2022 14:16:48 - INFO - codeparrot_training - Step 19787: {'lr': 0.00034888136626928427, 'samples': 3799296, 'steps': 19787, 'loss/train': 1.8739387392997742} 01/29/2022 14:16:52 - INFO - codeparrot_training - Step 19788: {'lr': 0.00034886633787253846, 'samples': 3799488, 'steps': 19788, 'loss/train': 1.1490823924541473} 01/29/2022 14:17:00 - INFO - codeparrot_training - Step 19789: {'lr': 0.00034885130905228063, 'samples': 3799680, 'steps': 19789, 'loss/train': 1.2922384142875671} 01/29/2022 14:17:04 - INFO - codeparrot_training - Step 19790: {'lr': 0.00034883627980857523, 'samples': 3799872, 'steps': 19790, 'loss/train': 1.353898286819458} 01/29/2022 14:17:08 - INFO - codeparrot_training - Step 19791: {'lr': 0.0003488212501414866, 'samples': 3800064, 'steps': 19791, 'loss/train': 1.7522440552711487} 01/29/2022 14:17:12 - INFO - codeparrot_training - Step 19792: {'lr': 0.00034880622005107915, 'samples': 3800256, 'steps': 19792, 'loss/train': 1.4582071602344513} 01/29/2022 14:17:17 - INFO - codeparrot_training - Step 19793: {'lr': 0.00034879118953741716, 'samples': 3800448, 'steps': 19793, 'loss/train': 1.8131874203681946} 01/29/2022 14:17:22 - INFO - codeparrot_training - Step 19794: {'lr': 0.0003487761586005651, 'samples': 3800640, 'steps': 19794, 'loss/train': 1.5801542401313782} 01/29/2022 14:17:26 - INFO - codeparrot_training - Step 19795: {'lr': 0.0003487611272405873, 'samples': 3800832, 'steps': 19795, 'loss/train': 2.409130811691284} 01/29/2022 14:17:30 - INFO - codeparrot_training - Step 19796: {'lr': 0.00034874609545754826, 'samples': 3801024, 'steps': 19796, 'loss/train': 1.649191975593567} 01/29/2022 14:17:35 - INFO - codeparrot_training - Step 19797: {'lr': 0.00034873106325151234, 'samples': 3801216, 'steps': 19797, 'loss/train': 1.7152056097984314} 01/29/2022 14:17:39 - INFO - codeparrot_training - Step 19798: {'lr': 0.0003487160306225438, 'samples': 3801408, 'steps': 19798, 'loss/train': 0.9925260543823242} 01/29/2022 14:17:46 - INFO - codeparrot_training - Step 19799: {'lr': 0.00034870099757070716, 'samples': 3801600, 'steps': 19799, 'loss/train': 2.2739052772521973} 01/29/2022 14:17:51 - INFO - codeparrot_training - Step 19800: {'lr': 0.0003486859640960668, 'samples': 3801792, 'steps': 19800, 'loss/train': 1.9082112908363342} 01/29/2022 14:17:55 - INFO - codeparrot_training - Step 19801: {'lr': 0.0003486709301986871, 'samples': 3801984, 'steps': 19801, 'loss/train': 2.308757185935974} 01/29/2022 14:17:59 - INFO - codeparrot_training - Step 19802: {'lr': 0.00034865589587863247, 'samples': 3802176, 'steps': 19802, 'loss/train': 1.5565851330757141} 01/29/2022 14:18:03 - INFO - codeparrot_training - Step 19803: {'lr': 0.0003486408611359673, 'samples': 3802368, 'steps': 19803, 'loss/train': 0.7234418392181396} 01/29/2022 14:18:09 - INFO - codeparrot_training - Step 19804: {'lr': 0.00034862582597075607, 'samples': 3802560, 'steps': 19804, 'loss/train': 0.9227169156074524} 01/29/2022 14:18:13 - INFO - codeparrot_training - Step 19805: {'lr': 0.00034861079038306305, 'samples': 3802752, 'steps': 19805, 'loss/train': 1.6006398797035217} 01/29/2022 14:18:17 - INFO - codeparrot_training - Step 19806: {'lr': 0.00034859575437295277, 'samples': 3802944, 'steps': 19806, 'loss/train': 2.0407683849334717} 01/29/2022 14:18:22 - INFO - codeparrot_training - Step 19807: {'lr': 0.00034858071794048953, 'samples': 3803136, 'steps': 19807, 'loss/train': 1.5317568182945251} 01/29/2022 14:18:27 - INFO - codeparrot_training - Step 19808: {'lr': 0.0003485656810857378, 'samples': 3803328, 'steps': 19808, 'loss/train': 0.46536679565906525} 01/29/2022 14:18:31 - INFO - codeparrot_training - Step 19809: {'lr': 0.00034855064380876193, 'samples': 3803520, 'steps': 19809, 'loss/train': 2.2627605199813843} 01/29/2022 14:18:36 - INFO - codeparrot_training - Step 19810: {'lr': 0.00034853560610962654, 'samples': 3803712, 'steps': 19810, 'loss/train': 1.8315449953079224} 01/29/2022 14:18:40 - INFO - codeparrot_training - Step 19811: {'lr': 0.0003485205679883958, 'samples': 3803904, 'steps': 19811, 'loss/train': 1.7730242013931274} 01/29/2022 14:18:44 - INFO - codeparrot_training - Step 19812: {'lr': 0.00034850552944513426, 'samples': 3804096, 'steps': 19812, 'loss/train': 1.6249321103096008} 01/29/2022 14:18:51 - INFO - codeparrot_training - Step 19813: {'lr': 0.00034849049047990633, 'samples': 3804288, 'steps': 19813, 'loss/train': 1.6453623175621033} 01/29/2022 14:18:56 - INFO - codeparrot_training - Step 19814: {'lr': 0.0003484754510927764, 'samples': 3804480, 'steps': 19814, 'loss/train': 1.0813652873039246} 01/29/2022 14:19:00 - INFO - codeparrot_training - Step 19815: {'lr': 0.00034846041128380886, 'samples': 3804672, 'steps': 19815, 'loss/train': 1.7471283674240112} 01/29/2022 14:19:04 - INFO - codeparrot_training - Step 19816: {'lr': 0.0003484453710530682, 'samples': 3804864, 'steps': 19816, 'loss/train': 1.274608165025711} 01/29/2022 14:19:08 - INFO - codeparrot_training - Step 19817: {'lr': 0.0003484303304006189, 'samples': 3805056, 'steps': 19817, 'loss/train': 1.1971831619739532} 01/29/2022 14:19:14 - INFO - codeparrot_training - Step 19818: {'lr': 0.0003484152893265253, 'samples': 3805248, 'steps': 19818, 'loss/train': 1.5759419202804565} 01/29/2022 14:19:18 - INFO - codeparrot_training - Step 19819: {'lr': 0.00034840024783085177, 'samples': 3805440, 'steps': 19819, 'loss/train': 1.3070612847805023} 01/29/2022 14:19:22 - INFO - codeparrot_training - Step 19820: {'lr': 0.00034838520591366285, 'samples': 3805632, 'steps': 19820, 'loss/train': 1.3472236096858978} 01/29/2022 14:19:26 - INFO - codeparrot_training - Step 19821: {'lr': 0.00034837016357502297, 'samples': 3805824, 'steps': 19821, 'loss/train': 2.3193057775497437} 01/29/2022 14:19:31 - INFO - codeparrot_training - Step 19822: {'lr': 0.0003483551208149965, 'samples': 3806016, 'steps': 19822, 'loss/train': 1.166859358549118} 01/29/2022 14:19:36 - INFO - codeparrot_training - Step 19823: {'lr': 0.00034834007763364803, 'samples': 3806208, 'steps': 19823, 'loss/train': 2.481230914592743} 01/29/2022 14:19:40 - INFO - codeparrot_training - Step 19824: {'lr': 0.00034832503403104184, 'samples': 3806400, 'steps': 19824, 'loss/train': 1.5599092841148376} 01/29/2022 14:19:44 - INFO - codeparrot_training - Step 19825: {'lr': 0.00034830999000724246, 'samples': 3806592, 'steps': 19825, 'loss/train': 1.6990495920181274} 01/29/2022 14:19:49 - INFO - codeparrot_training - Step 19826: {'lr': 0.00034829494556231423, 'samples': 3806784, 'steps': 19826, 'loss/train': 1.843087077140808} 01/29/2022 14:19:53 - INFO - codeparrot_training - Step 19827: {'lr': 0.00034827990069632173, 'samples': 3806976, 'steps': 19827, 'loss/train': 1.98023521900177} 01/29/2022 14:20:00 - INFO - codeparrot_training - Step 19828: {'lr': 0.0003482648554093293, 'samples': 3807168, 'steps': 19828, 'loss/train': 0.794867992401123} 01/29/2022 14:20:04 - INFO - codeparrot_training - Step 19829: {'lr': 0.0003482498097014015, 'samples': 3807360, 'steps': 19829, 'loss/train': 1.6497362852096558} 01/29/2022 14:20:09 - INFO - codeparrot_training - Step 19830: {'lr': 0.0003482347635726026, 'samples': 3807552, 'steps': 19830, 'loss/train': 1.9548872709274292} 01/29/2022 14:20:13 - INFO - codeparrot_training - Step 19831: {'lr': 0.00034821971702299716, 'samples': 3807744, 'steps': 19831, 'loss/train': 1.785404920578003} 01/29/2022 14:20:17 - INFO - codeparrot_training - Step 19832: {'lr': 0.0003482046700526498, 'samples': 3807936, 'steps': 19832, 'loss/train': 1.636880099773407} 01/29/2022 14:20:22 - INFO - codeparrot_training - Step 19833: {'lr': 0.0003481896226616246, 'samples': 3808128, 'steps': 19833, 'loss/train': 1.9316301941871643} 01/29/2022 14:20:27 - INFO - codeparrot_training - Step 19834: {'lr': 0.00034817457484998644, 'samples': 3808320, 'steps': 19834, 'loss/train': 1.7153601050376892} 01/29/2022 14:20:31 - INFO - codeparrot_training - Step 19835: {'lr': 0.00034815952661779946, 'samples': 3808512, 'steps': 19835, 'loss/train': 1.8885546326637268} 01/29/2022 14:20:35 - INFO - codeparrot_training - Step 19836: {'lr': 0.00034814447796512824, 'samples': 3808704, 'steps': 19836, 'loss/train': 1.6702181696891785} 01/29/2022 14:20:39 - INFO - codeparrot_training - Step 19837: {'lr': 0.0003481294288920373, 'samples': 3808896, 'steps': 19837, 'loss/train': 1.8149269223213196} 01/29/2022 14:20:45 - INFO - codeparrot_training - Step 19838: {'lr': 0.00034811437939859106, 'samples': 3809088, 'steps': 19838, 'loss/train': 1.5805286765098572} 01/29/2022 14:20:49 - INFO - codeparrot_training - Step 19839: {'lr': 0.0003480993294848539, 'samples': 3809280, 'steps': 19839, 'loss/train': 1.7009682655334473} 01/29/2022 14:20:53 - INFO - codeparrot_training - Step 19840: {'lr': 0.00034808427915089036, 'samples': 3809472, 'steps': 19840, 'loss/train': 1.5617056488990784} 01/29/2022 14:20:58 - INFO - codeparrot_training - Step 19841: {'lr': 0.000348069228396765, 'samples': 3809664, 'steps': 19841, 'loss/train': 1.367524802684784} 01/29/2022 14:21:02 - INFO - codeparrot_training - Step 19842: {'lr': 0.00034805417722254213, 'samples': 3809856, 'steps': 19842, 'loss/train': 1.4482136964797974} 01/29/2022 14:21:09 - INFO - codeparrot_training - Step 19843: {'lr': 0.00034803912562828633, 'samples': 3810048, 'steps': 19843, 'loss/train': 1.0547080636024475} 01/29/2022 14:21:13 - INFO - codeparrot_training - Step 19844: {'lr': 0.000348024073614062, 'samples': 3810240, 'steps': 19844, 'loss/train': 1.4657354056835175} 01/29/2022 14:21:18 - INFO - codeparrot_training - Step 19845: {'lr': 0.0003480090211799337, 'samples': 3810432, 'steps': 19845, 'loss/train': 0.3764060139656067} 01/29/2022 14:21:22 - INFO - codeparrot_training - Step 19846: {'lr': 0.0003479939683259659, 'samples': 3810624, 'steps': 19846, 'loss/train': 1.7224706411361694} 01/29/2022 14:21:26 - INFO - codeparrot_training - Step 19847: {'lr': 0.000347978915052223, 'samples': 3810816, 'steps': 19847, 'loss/train': 1.6035627722740173} 01/29/2022 14:21:31 - INFO - codeparrot_training - Step 19848: {'lr': 0.0003479638613587696, 'samples': 3811008, 'steps': 19848, 'loss/train': 2.2710530161857605} 01/29/2022 14:21:36 - INFO - codeparrot_training - Step 19849: {'lr': 0.0003479488072456701, 'samples': 3811200, 'steps': 19849, 'loss/train': 1.7309265732765198} 01/29/2022 14:21:40 - INFO - codeparrot_training - Step 19850: {'lr': 0.000347933752712989, 'samples': 3811392, 'steps': 19850, 'loss/train': 7.385162115097046} 01/29/2022 14:21:44 - INFO - codeparrot_training - Step 19851: {'lr': 0.00034791869776079084, 'samples': 3811584, 'steps': 19851, 'loss/train': 1.1382853388786316} 01/29/2022 14:21:48 - INFO - codeparrot_training - Step 19852: {'lr': 0.00034790364238914003, 'samples': 3811776, 'steps': 19852, 'loss/train': 1.9685683250427246} 01/29/2022 14:21:53 - INFO - codeparrot_training - Step 19853: {'lr': 0.0003478885865981011, 'samples': 3811968, 'steps': 19853, 'loss/train': 1.9657409191131592} 01/29/2022 14:21:58 - INFO - codeparrot_training - Step 19854: {'lr': 0.0003478735303877386, 'samples': 3812160, 'steps': 19854, 'loss/train': 2.826220214366913} 01/29/2022 14:22:02 - INFO - codeparrot_training - Step 19855: {'lr': 0.0003478584737581169, 'samples': 3812352, 'steps': 19855, 'loss/train': 0.7709982991218567} 01/29/2022 14:22:06 - INFO - codeparrot_training - Step 19856: {'lr': 0.00034784341670930066, 'samples': 3812544, 'steps': 19856, 'loss/train': 0.9648662209510803} 01/29/2022 14:22:11 - INFO - codeparrot_training - Step 19857: {'lr': 0.00034782835924135417, 'samples': 3812736, 'steps': 19857, 'loss/train': 1.6739838123321533} 01/29/2022 14:22:18 - INFO - codeparrot_training - Step 19858: {'lr': 0.0003478133013543422, 'samples': 3812928, 'steps': 19858, 'loss/train': 1.8691980242729187} 01/29/2022 14:22:22 - INFO - codeparrot_training - Step 19859: {'lr': 0.000347798243048329, 'samples': 3813120, 'steps': 19859, 'loss/train': 2.084520399570465} 01/29/2022 14:22:27 - INFO - codeparrot_training - Step 19860: {'lr': 0.00034778318432337926, 'samples': 3813312, 'steps': 19860, 'loss/train': 1.5093498229980469} 01/29/2022 14:22:31 - INFO - codeparrot_training - Step 19861: {'lr': 0.0003477681251795573, 'samples': 3813504, 'steps': 19861, 'loss/train': 1.056684672832489} 01/29/2022 14:22:35 - INFO - codeparrot_training - Step 19862: {'lr': 0.0003477530656169278, 'samples': 3813696, 'steps': 19862, 'loss/train': 1.090873807668686} 01/29/2022 14:22:40 - INFO - codeparrot_training - Step 19863: {'lr': 0.00034773800563555517, 'samples': 3813888, 'steps': 19863, 'loss/train': 1.2027782499790192} 01/29/2022 14:22:45 - INFO - codeparrot_training - Step 19864: {'lr': 0.0003477229452355041, 'samples': 3814080, 'steps': 19864, 'loss/train': 1.595878779888153} 01/29/2022 14:22:49 - INFO - codeparrot_training - Step 19865: {'lr': 0.00034770788441683875, 'samples': 3814272, 'steps': 19865, 'loss/train': 1.8192477822303772} 01/29/2022 14:22:53 - INFO - codeparrot_training - Step 19866: {'lr': 0.00034769282317962405, 'samples': 3814464, 'steps': 19866, 'loss/train': 1.2971927225589752} 01/29/2022 14:22:57 - INFO - codeparrot_training - Step 19867: {'lr': 0.00034767776152392417, 'samples': 3814656, 'steps': 19867, 'loss/train': 2.361308455467224} 01/29/2022 14:23:03 - INFO - codeparrot_training - Step 19868: {'lr': 0.0003476626994498038, 'samples': 3814848, 'steps': 19868, 'loss/train': 1.6397425532341003} 01/29/2022 14:23:07 - INFO - codeparrot_training - Step 19869: {'lr': 0.00034764763695732746, 'samples': 3815040, 'steps': 19869, 'loss/train': 1.6785916686058044} 01/29/2022 14:23:11 - INFO - codeparrot_training - Step 19870: {'lr': 0.0003476325740465597, 'samples': 3815232, 'steps': 19870, 'loss/train': 1.8425313234329224} 01/29/2022 14:23:15 - INFO - codeparrot_training - Step 19871: {'lr': 0.0003476175107175649, 'samples': 3815424, 'steps': 19871, 'loss/train': 1.5046595335006714} 01/29/2022 14:23:20 - INFO - codeparrot_training - Step 19872: {'lr': 0.00034760244697040776, 'samples': 3815616, 'steps': 19872, 'loss/train': 1.8550429344177246} 01/29/2022 14:23:26 - INFO - codeparrot_training - Step 19873: {'lr': 0.00034758738280515265, 'samples': 3815808, 'steps': 19873, 'loss/train': 2.011393904685974} 01/29/2022 14:23:31 - INFO - codeparrot_training - Step 19874: {'lr': 0.00034757231822186426, 'samples': 3816000, 'steps': 19874, 'loss/train': 1.8400919437408447} 01/29/2022 14:23:35 - INFO - codeparrot_training - Step 19875: {'lr': 0.00034755725322060705, 'samples': 3816192, 'steps': 19875, 'loss/train': 1.2496461868286133} 01/29/2022 14:23:39 - INFO - codeparrot_training - Step 19876: {'lr': 0.00034754218780144546, 'samples': 3816384, 'steps': 19876, 'loss/train': 1.4729155898094177} 01/29/2022 14:23:43 - INFO - codeparrot_training - Step 19877: {'lr': 0.00034752712196444417, 'samples': 3816576, 'steps': 19877, 'loss/train': 3.1340367794036865} 01/29/2022 14:23:49 - INFO - codeparrot_training - Step 19878: {'lr': 0.00034751205570966764, 'samples': 3816768, 'steps': 19878, 'loss/train': 1.9158115983009338} 01/29/2022 14:23:53 - INFO - codeparrot_training - Step 19879: {'lr': 0.0003474969890371805, 'samples': 3816960, 'steps': 19879, 'loss/train': 1.8159837126731873} 01/29/2022 14:23:57 - INFO - codeparrot_training - Step 19880: {'lr': 0.0003474819219470471, 'samples': 3817152, 'steps': 19880, 'loss/train': 1.2966741621494293} 01/29/2022 14:24:02 - INFO - codeparrot_training - Step 19881: {'lr': 0.0003474668544393321, 'samples': 3817344, 'steps': 19881, 'loss/train': 0.9819155931472778} 01/29/2022 14:24:06 - INFO - codeparrot_training - Step 19882: {'lr': 0.00034745178651410014, 'samples': 3817536, 'steps': 19882, 'loss/train': 1.4983792304992676} 01/29/2022 14:24:11 - INFO - codeparrot_training - Step 19883: {'lr': 0.0003474367181714156, 'samples': 3817728, 'steps': 19883, 'loss/train': 1.6764573454856873} 01/29/2022 14:24:15 - INFO - codeparrot_training - Step 19884: {'lr': 0.0003474216494113431, 'samples': 3817920, 'steps': 19884, 'loss/train': 0.8942650258541107} 01/29/2022 14:24:20 - INFO - codeparrot_training - Step 19885: {'lr': 0.00034740658023394723, 'samples': 3818112, 'steps': 19885, 'loss/train': 1.686253845691681} 01/29/2022 14:24:24 - INFO - codeparrot_training - Step 19886: {'lr': 0.0003473915106392925, 'samples': 3818304, 'steps': 19886, 'loss/train': 2.4046667218208313} 01/29/2022 14:24:28 - INFO - codeparrot_training - Step 19887: {'lr': 0.00034737644062744343, 'samples': 3818496, 'steps': 19887, 'loss/train': 2.338243246078491} 01/29/2022 14:24:36 - INFO - codeparrot_training - Step 19888: {'lr': 0.0003473613701984646, 'samples': 3818688, 'steps': 19888, 'loss/train': 1.1308327317237854} 01/29/2022 14:24:40 - INFO - codeparrot_training - Step 19889: {'lr': 0.0003473462993524206, 'samples': 3818880, 'steps': 19889, 'loss/train': 1.6103596687316895} 01/29/2022 14:24:44 - INFO - codeparrot_training - Step 19890: {'lr': 0.000347331228089376, 'samples': 3819072, 'steps': 19890, 'loss/train': 1.113207846879959} 01/29/2022 14:24:49 - INFO - codeparrot_training - Step 19891: {'lr': 0.0003473161564093953, 'samples': 3819264, 'steps': 19891, 'loss/train': 1.9750801920890808} 01/29/2022 14:24:53 - INFO - codeparrot_training - Step 19892: {'lr': 0.0003473010843125431, 'samples': 3819456, 'steps': 19892, 'loss/train': 1.6419458985328674} 01/29/2022 14:24:58 - INFO - codeparrot_training - Step 19893: {'lr': 0.00034728601179888395, 'samples': 3819648, 'steps': 19893, 'loss/train': 0.7975393831729889} 01/29/2022 14:25:02 - INFO - codeparrot_training - Step 19894: {'lr': 0.00034727093886848236, 'samples': 3819840, 'steps': 19894, 'loss/train': 3.2060272693634033} 01/29/2022 14:25:07 - INFO - codeparrot_training - Step 19895: {'lr': 0.000347255865521403, 'samples': 3820032, 'steps': 19895, 'loss/train': 4.066907644271851} 01/29/2022 14:25:11 - INFO - codeparrot_training - Step 19896: {'lr': 0.0003472407917577104, 'samples': 3820224, 'steps': 19896, 'loss/train': 1.732602059841156} 01/29/2022 14:25:15 - INFO - codeparrot_training - Step 19897: {'lr': 0.0003472257175774691, 'samples': 3820416, 'steps': 19897, 'loss/train': 1.8369308710098267} 01/29/2022 14:25:23 - INFO - codeparrot_training - Step 19898: {'lr': 0.00034721064298074366, 'samples': 3820608, 'steps': 19898, 'loss/train': 4.56077778339386} 01/29/2022 14:25:28 - INFO - codeparrot_training - Step 19899: {'lr': 0.0003471955679675988, 'samples': 3820800, 'steps': 19899, 'loss/train': 1.45445117354393} 01/29/2022 14:25:32 - INFO - codeparrot_training - Step 19900: {'lr': 0.0003471804925380989, 'samples': 3820992, 'steps': 19900, 'loss/train': 1.7058148384094238} 01/29/2022 14:25:36 - INFO - codeparrot_training - Step 19901: {'lr': 0.0003471654166923087, 'samples': 3821184, 'steps': 19901, 'loss/train': 1.6645422577857971} 01/29/2022 14:25:40 - INFO - codeparrot_training - Step 19902: {'lr': 0.00034715034043029263, 'samples': 3821376, 'steps': 19902, 'loss/train': 0.7476664334535599} 01/29/2022 14:25:45 - INFO - codeparrot_training - Step 19903: {'lr': 0.00034713526375211546, 'samples': 3821568, 'steps': 19903, 'loss/train': 0.9350361227989197} 01/29/2022 14:25:50 - INFO - codeparrot_training - Step 19904: {'lr': 0.00034712018665784155, 'samples': 3821760, 'steps': 19904, 'loss/train': 1.3651572167873383} 01/29/2022 14:25:54 - INFO - codeparrot_training - Step 19905: {'lr': 0.00034710510914753563, 'samples': 3821952, 'steps': 19905, 'loss/train': 1.0748846232891083} 01/29/2022 14:25:59 - INFO - codeparrot_training - Step 19906: {'lr': 0.00034709003122126227, 'samples': 3822144, 'steps': 19906, 'loss/train': 1.994054138660431} 01/29/2022 14:26:03 - INFO - codeparrot_training - Step 19907: {'lr': 0.000347074952879086, 'samples': 3822336, 'steps': 19907, 'loss/train': 1.7318127751350403} 01/29/2022 14:26:07 - INFO - codeparrot_training - Step 19908: {'lr': 0.0003470598741210715, 'samples': 3822528, 'steps': 19908, 'loss/train': 1.764779806137085} 01/29/2022 14:26:13 - INFO - codeparrot_training - Step 19909: {'lr': 0.00034704479494728337, 'samples': 3822720, 'steps': 19909, 'loss/train': 2.1693670749664307} 01/29/2022 14:26:17 - INFO - codeparrot_training - Step 19910: {'lr': 0.00034702971535778614, 'samples': 3822912, 'steps': 19910, 'loss/train': 1.7673897743225098} 01/29/2022 14:26:21 - INFO - codeparrot_training - Step 19911: {'lr': 0.00034701463535264434, 'samples': 3823104, 'steps': 19911, 'loss/train': 1.1912703216075897} 01/29/2022 14:26:25 - INFO - codeparrot_training - Step 19912: {'lr': 0.0003469995549319227, 'samples': 3823296, 'steps': 19912, 'loss/train': 2.6301464438438416} 01/29/2022 14:26:33 - INFO - codeparrot_training - Step 19913: {'lr': 0.0003469844740956858, 'samples': 3823488, 'steps': 19913, 'loss/train': 0.46722784638404846} 01/29/2022 14:26:37 - INFO - codeparrot_training - Step 19914: {'lr': 0.0003469693928439982, 'samples': 3823680, 'steps': 19914, 'loss/train': 0.593633383512497} 01/29/2022 14:26:41 - INFO - codeparrot_training - Step 19915: {'lr': 0.00034695431117692446, 'samples': 3823872, 'steps': 19915, 'loss/train': 2.0860360264778137} 01/29/2022 14:26:45 - INFO - codeparrot_training - Step 19916: {'lr': 0.0003469392290945292, 'samples': 3824064, 'steps': 19916, 'loss/train': 1.608172059059143} 01/29/2022 14:26:50 - INFO - codeparrot_training - Step 19917: {'lr': 0.00034692414659687714, 'samples': 3824256, 'steps': 19917, 'loss/train': 1.942573070526123} 01/29/2022 14:26:55 - INFO - codeparrot_training - Step 19918: {'lr': 0.0003469090636840328, 'samples': 3824448, 'steps': 19918, 'loss/train': 1.7452152371406555} 01/29/2022 14:26:59 - INFO - codeparrot_training - Step 19919: {'lr': 0.0003468939803560608, 'samples': 3824640, 'steps': 19919, 'loss/train': 2.1630476117134094} 01/29/2022 14:27:04 - INFO - codeparrot_training - Step 19920: {'lr': 0.00034687889661302575, 'samples': 3824832, 'steps': 19920, 'loss/train': 3.287796378135681} 01/29/2022 14:27:08 - INFO - codeparrot_training - Step 19921: {'lr': 0.0003468638124549923, 'samples': 3825024, 'steps': 19921, 'loss/train': 1.5578044652938843} 01/29/2022 14:27:12 - INFO - codeparrot_training - Step 19922: {'lr': 0.00034684872788202497, 'samples': 3825216, 'steps': 19922, 'loss/train': 1.4943364262580872} 01/29/2022 14:27:19 - INFO - codeparrot_training - Step 19923: {'lr': 0.0003468336428941885, 'samples': 3825408, 'steps': 19923, 'loss/train': 1.6500687003135681} 01/29/2022 14:27:24 - INFO - codeparrot_training - Step 19924: {'lr': 0.00034681855749154743, 'samples': 3825600, 'steps': 19924, 'loss/train': 1.2760613858699799} 01/29/2022 14:27:28 - INFO - codeparrot_training - Step 19925: {'lr': 0.00034680347167416643, 'samples': 3825792, 'steps': 19925, 'loss/train': 2.0168983340263367} 01/29/2022 14:27:32 - INFO - codeparrot_training - Step 19926: {'lr': 0.00034678838544211003, 'samples': 3825984, 'steps': 19926, 'loss/train': 2.1354546546936035} 01/29/2022 14:27:36 - INFO - codeparrot_training - Step 19927: {'lr': 0.000346773298795443, 'samples': 3826176, 'steps': 19927, 'loss/train': 1.5801830291748047} 01/29/2022 14:27:42 - INFO - codeparrot_training - Step 19928: {'lr': 0.00034675821173422983, 'samples': 3826368, 'steps': 19928, 'loss/train': 1.5084391236305237} 01/29/2022 14:27:46 - INFO - codeparrot_training - Step 19929: {'lr': 0.0003467431242585352, 'samples': 3826560, 'steps': 19929, 'loss/train': 1.4028106033802032} 01/29/2022 14:27:50 - INFO - codeparrot_training - Step 19930: {'lr': 0.0003467280363684238, 'samples': 3826752, 'steps': 19930, 'loss/train': 0.48328667879104614} 01/29/2022 14:27:54 - INFO - codeparrot_training - Step 19931: {'lr': 0.0003467129480639601, 'samples': 3826944, 'steps': 19931, 'loss/train': 1.9911219477653503} 01/29/2022 14:27:59 - INFO - codeparrot_training - Step 19932: {'lr': 0.000346697859345209, 'samples': 3827136, 'steps': 19932, 'loss/train': 1.2557665407657623} 01/29/2022 14:28:04 - INFO - codeparrot_training - Step 19933: {'lr': 0.00034668277021223493, 'samples': 3827328, 'steps': 19933, 'loss/train': 1.890133023262024} 01/29/2022 14:28:08 - INFO - codeparrot_training - Step 19934: {'lr': 0.0003466676806651025, 'samples': 3827520, 'steps': 19934, 'loss/train': 2.054341971874237} 01/29/2022 14:28:12 - INFO - codeparrot_training - Step 19935: {'lr': 0.0003466525907038765, 'samples': 3827712, 'steps': 19935, 'loss/train': 1.6177391409873962} 01/29/2022 14:28:17 - INFO - codeparrot_training - Step 19936: {'lr': 0.0003466375003286214, 'samples': 3827904, 'steps': 19936, 'loss/train': 1.0954255163669586} 01/29/2022 14:28:21 - INFO - codeparrot_training - Step 19937: {'lr': 0.00034662240953940205, 'samples': 3828096, 'steps': 19937, 'loss/train': 1.7784624695777893} 01/29/2022 14:28:26 - INFO - codeparrot_training - Step 19938: {'lr': 0.0003466073183362829, 'samples': 3828288, 'steps': 19938, 'loss/train': 1.5079877972602844} 01/29/2022 14:28:30 - INFO - codeparrot_training - Step 19939: {'lr': 0.00034659222671932865, 'samples': 3828480, 'steps': 19939, 'loss/train': 1.28348046541214} 01/29/2022 14:28:35 - INFO - codeparrot_training - Step 19940: {'lr': 0.000346577134688604, 'samples': 3828672, 'steps': 19940, 'loss/train': 2.201064348220825} 01/29/2022 14:28:39 - INFO - codeparrot_training - Step 19941: {'lr': 0.0003465620422441737, 'samples': 3828864, 'steps': 19941, 'loss/train': 2.67813903093338} 01/29/2022 14:28:43 - INFO - codeparrot_training - Step 19942: {'lr': 0.00034654694938610205, 'samples': 3829056, 'steps': 19942, 'loss/train': 2.111135244369507} 01/29/2022 14:28:51 - INFO - codeparrot_training - Step 19943: {'lr': 0.00034653185611445403, 'samples': 3829248, 'steps': 19943, 'loss/train': 1.6094774007797241} 01/29/2022 14:28:55 - INFO - codeparrot_training - Step 19944: {'lr': 0.0003465167624292942, 'samples': 3829440, 'steps': 19944, 'loss/train': 1.3651962876319885} 01/29/2022 14:28:59 - INFO - codeparrot_training - Step 19945: {'lr': 0.0003465016683306872, 'samples': 3829632, 'steps': 19945, 'loss/train': 2.038383901119232} 01/29/2022 14:29:03 - INFO - codeparrot_training - Step 19946: {'lr': 0.0003464865738186977, 'samples': 3829824, 'steps': 19946, 'loss/train': 1.5532368421554565} 01/29/2022 14:29:07 - INFO - codeparrot_training - Step 19947: {'lr': 0.0003464714788933904, 'samples': 3830016, 'steps': 19947, 'loss/train': 0.892951101064682} 01/29/2022 14:29:13 - INFO - codeparrot_training - Step 19948: {'lr': 0.0003464563835548298, 'samples': 3830208, 'steps': 19948, 'loss/train': 2.2537230849266052} 01/29/2022 14:29:17 - INFO - codeparrot_training - Step 19949: {'lr': 0.0003464412878030808, 'samples': 3830400, 'steps': 19949, 'loss/train': 1.8132789731025696} 01/29/2022 14:29:21 - INFO - codeparrot_training - Step 19950: {'lr': 0.0003464261916382079, 'samples': 3830592, 'steps': 19950, 'loss/train': 2.224495232105255} 01/29/2022 14:29:26 - INFO - codeparrot_training - Step 19951: {'lr': 0.0003464110950602758, 'samples': 3830784, 'steps': 19951, 'loss/train': 1.9621624946594238} 01/29/2022 14:29:30 - INFO - codeparrot_training - Step 19952: {'lr': 0.00034639599806934917, 'samples': 3830976, 'steps': 19952, 'loss/train': 1.5201650261878967} 01/29/2022 14:29:35 - INFO - codeparrot_training - Step 19953: {'lr': 0.0003463809006654927, 'samples': 3831168, 'steps': 19953, 'loss/train': 0.6196822375059128} 01/29/2022 14:29:39 - INFO - codeparrot_training - Step 19954: {'lr': 0.0003463658028487711, 'samples': 3831360, 'steps': 19954, 'loss/train': 2.035664677619934} 01/29/2022 14:29:44 - INFO - codeparrot_training - Step 19955: {'lr': 0.0003463507046192489, 'samples': 3831552, 'steps': 19955, 'loss/train': 1.5688592791557312} 01/29/2022 14:29:48 - INFO - codeparrot_training - Step 19956: {'lr': 0.000346335605976991, 'samples': 3831744, 'steps': 19956, 'loss/train': 1.753345549106598} 01/29/2022 14:29:52 - INFO - codeparrot_training - Step 19957: {'lr': 0.00034632050692206184, 'samples': 3831936, 'steps': 19957, 'loss/train': 1.7933902144432068} 01/29/2022 14:29:59 - INFO - codeparrot_training - Step 19958: {'lr': 0.0003463054074545263, 'samples': 3832128, 'steps': 19958, 'loss/train': 1.2326876521110535} 01/29/2022 14:30:04 - INFO - codeparrot_training - Step 19959: {'lr': 0.00034629030757444895, 'samples': 3832320, 'steps': 19959, 'loss/train': 4.130524635314941} 01/29/2022 14:30:08 - INFO - codeparrot_training - Step 19960: {'lr': 0.00034627520728189456, 'samples': 3832512, 'steps': 19960, 'loss/train': 1.626019299030304} 01/29/2022 14:30:12 - INFO - codeparrot_training - Step 19961: {'lr': 0.00034626010657692766, 'samples': 3832704, 'steps': 19961, 'loss/train': 1.4293895959854126} 01/29/2022 14:30:16 - INFO - codeparrot_training - Step 19962: {'lr': 0.0003462450054596131, 'samples': 3832896, 'steps': 19962, 'loss/train': 1.7640570402145386} 01/29/2022 14:30:22 - INFO - codeparrot_training - Step 19963: {'lr': 0.0003462299039300154, 'samples': 3833088, 'steps': 19963, 'loss/train': 1.843812346458435} 01/29/2022 14:30:26 - INFO - codeparrot_training - Step 19964: {'lr': 0.0003462148019881994, 'samples': 3833280, 'steps': 19964, 'loss/train': 1.7086785435676575} 01/29/2022 14:30:30 - INFO - codeparrot_training - Step 19965: {'lr': 0.0003461996996342298, 'samples': 3833472, 'steps': 19965, 'loss/train': 1.9650426506996155} 01/29/2022 14:30:35 - INFO - codeparrot_training - Step 19966: {'lr': 0.00034618459686817113, 'samples': 3833664, 'steps': 19966, 'loss/train': 1.5044565796852112} 01/29/2022 14:30:39 - INFO - codeparrot_training - Step 19967: {'lr': 0.0003461694936900883, 'samples': 3833856, 'steps': 19967, 'loss/train': 1.5175361037254333} 01/29/2022 14:30:46 - INFO - codeparrot_training - Step 19968: {'lr': 0.0003461543901000458, 'samples': 3834048, 'steps': 19968, 'loss/train': 1.7948278784751892} 01/29/2022 14:30:50 - INFO - codeparrot_training - Step 19969: {'lr': 0.00034613928609810845, 'samples': 3834240, 'steps': 19969, 'loss/train': 1.4453086853027344} 01/29/2022 14:30:54 - INFO - codeparrot_training - Step 19970: {'lr': 0.0003461241816843409, 'samples': 3834432, 'steps': 19970, 'loss/train': 1.7310937643051147} 01/29/2022 14:30:59 - INFO - codeparrot_training - Step 19971: {'lr': 0.00034610907685880794, 'samples': 3834624, 'steps': 19971, 'loss/train': 7.641424655914307} 01/29/2022 14:31:03 - INFO - codeparrot_training - Step 19972: {'lr': 0.00034609397162157417, 'samples': 3834816, 'steps': 19972, 'loss/train': 1.2687291204929352} 01/29/2022 14:31:08 - INFO - codeparrot_training - Step 19973: {'lr': 0.0003460788659727044, 'samples': 3835008, 'steps': 19973, 'loss/train': 1.560583233833313} 01/29/2022 14:31:12 - INFO - codeparrot_training - Step 19974: {'lr': 0.0003460637599122632, 'samples': 3835200, 'steps': 19974, 'loss/train': 1.3242722153663635} 01/29/2022 14:31:17 - INFO - codeparrot_training - Step 19975: {'lr': 0.0003460486534403154, 'samples': 3835392, 'steps': 19975, 'loss/train': 1.8274168968200684} 01/29/2022 14:31:21 - INFO - codeparrot_training - Step 19976: {'lr': 0.0003460335465569256, 'samples': 3835584, 'steps': 19976, 'loss/train': 2.0555736422538757} 01/29/2022 14:31:25 - INFO - codeparrot_training - Step 19977: {'lr': 0.0003460184392621587, 'samples': 3835776, 'steps': 19977, 'loss/train': 2.4475278854370117} 01/29/2022 14:31:31 - INFO - codeparrot_training - Step 19978: {'lr': 0.0003460033315560792, 'samples': 3835968, 'steps': 19978, 'loss/train': 1.6971244812011719} 01/29/2022 14:31:35 - INFO - codeparrot_training - Step 19979: {'lr': 0.00034598822343875197, 'samples': 3836160, 'steps': 19979, 'loss/train': 1.6380870938301086} 01/29/2022 14:31:39 - INFO - codeparrot_training - Step 19980: {'lr': 0.0003459731149102417, 'samples': 3836352, 'steps': 19980, 'loss/train': 1.2615710198879242} 01/29/2022 14:31:43 - INFO - codeparrot_training - Step 19981: {'lr': 0.000345958005970613, 'samples': 3836544, 'steps': 19981, 'loss/train': 1.4810679852962494} 01/29/2022 14:31:48 - INFO - codeparrot_training - Step 19982: {'lr': 0.0003459428966199307, 'samples': 3836736, 'steps': 19982, 'loss/train': 1.7996129393577576} 01/29/2022 14:31:55 - INFO - codeparrot_training - Step 19983: {'lr': 0.0003459277868582595, 'samples': 3836928, 'steps': 19983, 'loss/train': 1.7412285804748535} 01/29/2022 14:31:59 - INFO - codeparrot_training - Step 19984: {'lr': 0.00034591267668566413, 'samples': 3837120, 'steps': 19984, 'loss/train': 1.5591740012168884} 01/29/2022 14:32:03 - INFO - codeparrot_training - Step 19985: {'lr': 0.00034589756610220923, 'samples': 3837312, 'steps': 19985, 'loss/train': 0.9495113790035248} 01/29/2022 14:32:08 - INFO - codeparrot_training - Step 19986: {'lr': 0.0003458824551079597, 'samples': 3837504, 'steps': 19986, 'loss/train': 1.9260481595993042} 01/29/2022 14:32:12 - INFO - codeparrot_training - Step 19987: {'lr': 0.00034586734370298017, 'samples': 3837696, 'steps': 19987, 'loss/train': 1.7354246377944946} 01/29/2022 14:32:18 - INFO - codeparrot_training - Step 19988: {'lr': 0.00034585223188733535, 'samples': 3837888, 'steps': 19988, 'loss/train': 2.014822840690613} 01/29/2022 14:32:22 - INFO - codeparrot_training - Step 19989: {'lr': 0.00034583711966109005, 'samples': 3838080, 'steps': 19989, 'loss/train': 1.6913952827453613} 01/29/2022 14:32:26 - INFO - codeparrot_training - Step 19990: {'lr': 0.0003458220070243089, 'samples': 3838272, 'steps': 19990, 'loss/train': 1.6448739767074585} 01/29/2022 14:32:30 - INFO - codeparrot_training - Step 19991: {'lr': 0.0003458068939770567, 'samples': 3838464, 'steps': 19991, 'loss/train': 2.260426640510559} 01/29/2022 14:32:35 - INFO - codeparrot_training - Step 19992: {'lr': 0.00034579178051939827, 'samples': 3838656, 'steps': 19992, 'loss/train': 1.696556031703949} 01/29/2022 14:32:40 - INFO - codeparrot_training - Step 19993: {'lr': 0.00034577666665139815, 'samples': 3838848, 'steps': 19993, 'loss/train': 1.7164812684059143} 01/29/2022 14:32:44 - INFO - codeparrot_training - Step 19994: {'lr': 0.0003457615523731213, 'samples': 3839040, 'steps': 19994, 'loss/train': 2.1028931736946106} 01/29/2022 14:32:49 - INFO - codeparrot_training - Step 19995: {'lr': 0.00034574643768463237, 'samples': 3839232, 'steps': 19995, 'loss/train': 1.5937997102737427} 01/29/2022 14:32:53 - INFO - codeparrot_training - Step 19996: {'lr': 0.00034573132258599606, 'samples': 3839424, 'steps': 19996, 'loss/train': 1.5879653692245483} 01/29/2022 14:32:57 - INFO - codeparrot_training - Step 19997: {'lr': 0.00034571620707727713, 'samples': 3839616, 'steps': 19997, 'loss/train': 1.4173873364925385} 01/29/2022 14:33:03 - INFO - codeparrot_training - Step 19998: {'lr': 0.0003457010911585404, 'samples': 3839808, 'steps': 19998, 'loss/train': 1.615257740020752} 01/29/2022 14:33:07 - INFO - codeparrot_training - Step 19999: {'lr': 0.00034568597482985067, 'samples': 3840000, 'steps': 19999, 'loss/train': 1.7240061163902283} 01/29/2022 14:33:07 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 14:33:40 - WARNING - huggingface_hub.repository - Several commits (10) will be pushed upstream. 01/29/2022 14:33:40 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 14:34:47 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py a5ccdc4..5707af2 expert-dust-2 -> expert-dust-2 01/29/2022 14:34:53 - INFO - codeparrot_training - Step 20000: {'lr': 0.0003456708580912725, 'samples': 3840192, 'steps': 20000, 'loss/train': 1.5932216048240662} 01/29/2022 14:34:57 - INFO - codeparrot_training - Step 20001: {'lr': 0.00034565574094287075, 'samples': 3840384, 'steps': 20001, 'loss/train': 0.6899607628583908} 01/29/2022 14:35:02 - INFO - codeparrot_training - Step 20002: {'lr': 0.0003456406233847102, 'samples': 3840576, 'steps': 20002, 'loss/train': 0.738361045718193} 01/29/2022 14:35:10 - INFO - codeparrot_training - Step 20003: {'lr': 0.00034562550541685557, 'samples': 3840768, 'steps': 20003, 'loss/train': 1.544768750667572} 01/29/2022 14:35:14 - INFO - codeparrot_training - Step 20004: {'lr': 0.0003456103870393717, 'samples': 3840960, 'steps': 20004, 'loss/train': 1.6897265911102295} 01/29/2022 14:35:18 - INFO - codeparrot_training - Step 20005: {'lr': 0.0003455952682523232, 'samples': 3841152, 'steps': 20005, 'loss/train': 2.024812638759613} 01/29/2022 14:35:23 - INFO - codeparrot_training - Step 20006: {'lr': 0.00034558014905577506, 'samples': 3841344, 'steps': 20006, 'loss/train': 1.1149585247039795} 01/29/2022 14:35:27 - INFO - codeparrot_training - Step 20007: {'lr': 0.00034556502944979177, 'samples': 3841536, 'steps': 20007, 'loss/train': 1.6402711272239685} 01/29/2022 14:35:32 - INFO - codeparrot_training - Step 20008: {'lr': 0.0003455499094344383, 'samples': 3841728, 'steps': 20008, 'loss/train': 2.434077501296997} 01/29/2022 14:35:36 - INFO - codeparrot_training - Step 20009: {'lr': 0.00034553478900977943, 'samples': 3841920, 'steps': 20009, 'loss/train': 0.8659319579601288} 01/29/2022 14:35:41 - INFO - codeparrot_training - Step 20010: {'lr': 0.0003455196681758798, 'samples': 3842112, 'steps': 20010, 'loss/train': 2.160567820072174} 01/29/2022 14:35:45 - INFO - codeparrot_training - Step 20011: {'lr': 0.00034550454693280417, 'samples': 3842304, 'steps': 20011, 'loss/train': 1.2308279871940613} 01/29/2022 14:35:49 - INFO - codeparrot_training - Step 20012: {'lr': 0.0003454894252806175, 'samples': 3842496, 'steps': 20012, 'loss/train': 2.2019230127334595} 01/29/2022 14:35:57 - INFO - codeparrot_training - Step 20013: {'lr': 0.0003454743032193844, 'samples': 3842688, 'steps': 20013, 'loss/train': 2.1922924518585205} 01/29/2022 14:36:02 - INFO - codeparrot_training - Step 20014: {'lr': 0.00034545918074916965, 'samples': 3842880, 'steps': 20014, 'loss/train': 1.323004961013794} 01/29/2022 14:36:06 - INFO - codeparrot_training - Step 20015: {'lr': 0.00034544405787003817, 'samples': 3843072, 'steps': 20015, 'loss/train': 1.6215190887451172} 01/29/2022 14:36:10 - INFO - codeparrot_training - Step 20016: {'lr': 0.0003454289345820546, 'samples': 3843264, 'steps': 20016, 'loss/train': 1.295650988817215} 01/29/2022 14:36:14 - INFO - codeparrot_training - Step 20017: {'lr': 0.00034541381088528376, 'samples': 3843456, 'steps': 20017, 'loss/train': 1.3166475892066956} 01/29/2022 14:36:19 - INFO - codeparrot_training - Step 20018: {'lr': 0.00034539868677979055, 'samples': 3843648, 'steps': 20018, 'loss/train': 0.9329046607017517} 01/29/2022 14:36:24 - INFO - codeparrot_training - Step 20019: {'lr': 0.0003453835622656396, 'samples': 3843840, 'steps': 20019, 'loss/train': 1.5942893028259277} 01/29/2022 14:36:29 - INFO - codeparrot_training - Step 20020: {'lr': 0.0003453684373428957, 'samples': 3844032, 'steps': 20020, 'loss/train': 1.4464179575443268} 01/29/2022 14:36:33 - INFO - codeparrot_training - Step 20021: {'lr': 0.0003453533120116238, 'samples': 3844224, 'steps': 20021, 'loss/train': 2.2602803707122803} 01/29/2022 14:36:37 - INFO - codeparrot_training - Step 20022: {'lr': 0.0003453381862718886, 'samples': 3844416, 'steps': 20022, 'loss/train': 2.034035861492157} 01/29/2022 14:36:41 - INFO - codeparrot_training - Step 20023: {'lr': 0.00034532306012375474, 'samples': 3844608, 'steps': 20023, 'loss/train': 0.14405613765120506} 01/29/2022 14:36:47 - INFO - codeparrot_training - Step 20024: {'lr': 0.00034530793356728727, 'samples': 3844800, 'steps': 20024, 'loss/train': 1.5499262809753418} 01/29/2022 14:36:51 - INFO - codeparrot_training - Step 20025: {'lr': 0.00034529280660255084, 'samples': 3844992, 'steps': 20025, 'loss/train': 1.2833970487117767} 01/29/2022 14:36:55 - INFO - codeparrot_training - Step 20026: {'lr': 0.00034527767922961034, 'samples': 3845184, 'steps': 20026, 'loss/train': 1.583337128162384} 01/29/2022 14:36:59 - INFO - codeparrot_training - Step 20027: {'lr': 0.0003452625514485305, 'samples': 3845376, 'steps': 20027, 'loss/train': 2.075325071811676} 01/29/2022 14:37:07 - INFO - codeparrot_training - Step 20028: {'lr': 0.0003452474232593761, 'samples': 3845568, 'steps': 20028, 'loss/train': 2.1944079995155334} 01/29/2022 14:37:11 - INFO - codeparrot_training - Step 20029: {'lr': 0.00034523229466221195, 'samples': 3845760, 'steps': 20029, 'loss/train': 0.6824075728654861} 01/29/2022 14:37:16 - INFO - codeparrot_training - Step 20030: {'lr': 0.00034521716565710293, 'samples': 3845952, 'steps': 20030, 'loss/train': 1.4223999381065369} 01/29/2022 14:37:20 - INFO - codeparrot_training - Step 20031: {'lr': 0.00034520203624411385, 'samples': 3846144, 'steps': 20031, 'loss/train': 1.327891767024994} 01/29/2022 14:37:24 - INFO - codeparrot_training - Step 20032: {'lr': 0.0003451869064233094, 'samples': 3846336, 'steps': 20032, 'loss/train': 1.3210590183734894} 01/29/2022 14:37:28 - INFO - codeparrot_training - Step 20033: {'lr': 0.0003451717761947545, 'samples': 3846528, 'steps': 20033, 'loss/train': 0.6761310249567032} 01/29/2022 14:37:34 - INFO - codeparrot_training - Step 20034: {'lr': 0.0003451566455585139, 'samples': 3846720, 'steps': 20034, 'loss/train': 2.457672894001007} 01/29/2022 14:37:38 - INFO - codeparrot_training - Step 20035: {'lr': 0.00034514151451465254, 'samples': 3846912, 'steps': 20035, 'loss/train': 2.1825268864631653} 01/29/2022 14:37:42 - INFO - codeparrot_training - Step 20036: {'lr': 0.00034512638306323506, 'samples': 3847104, 'steps': 20036, 'loss/train': 2.1176796555519104} 01/29/2022 14:37:47 - INFO - codeparrot_training - Step 20037: {'lr': 0.0003451112512043264, 'samples': 3847296, 'steps': 20037, 'loss/train': 0.7753494679927826} 01/29/2022 14:37:51 - INFO - codeparrot_training - Step 20038: {'lr': 0.0003450961189379913, 'samples': 3847488, 'steps': 20038, 'loss/train': 1.13225457072258} 01/29/2022 14:37:56 - INFO - codeparrot_training - Step 20039: {'lr': 0.0003450809862642947, 'samples': 3847680, 'steps': 20039, 'loss/train': 1.837877869606018} 01/29/2022 14:38:00 - INFO - codeparrot_training - Step 20040: {'lr': 0.0003450658531833013, 'samples': 3847872, 'steps': 20040, 'loss/train': 1.465055376291275} 01/29/2022 14:38:05 - INFO - codeparrot_training - Step 20041: {'lr': 0.00034505071969507595, 'samples': 3848064, 'steps': 20041, 'loss/train': 2.067894458770752} 01/29/2022 14:38:09 - INFO - codeparrot_training - Step 20042: {'lr': 0.0003450355857996835, 'samples': 3848256, 'steps': 20042, 'loss/train': 1.8883525729179382} 01/29/2022 14:38:13 - INFO - codeparrot_training - Step 20043: {'lr': 0.0003450204514971888, 'samples': 3848448, 'steps': 20043, 'loss/train': 1.4919598996639252} 01/29/2022 14:38:21 - INFO - codeparrot_training - Step 20044: {'lr': 0.0003450053167876566, 'samples': 3848640, 'steps': 20044, 'loss/train': 1.3093605637550354} 01/29/2022 14:38:25 - INFO - codeparrot_training - Step 20045: {'lr': 0.0003449901816711519, 'samples': 3848832, 'steps': 20045, 'loss/train': 2.472912311553955} 01/29/2022 14:38:29 - INFO - codeparrot_training - Step 20046: {'lr': 0.00034497504614773935, 'samples': 3849024, 'steps': 20046, 'loss/train': 1.6081700921058655} 01/29/2022 14:38:33 - INFO - codeparrot_training - Step 20047: {'lr': 0.0003449599102174839, 'samples': 3849216, 'steps': 20047, 'loss/train': 1.8284838795661926} 01/29/2022 14:38:38 - INFO - codeparrot_training - Step 20048: {'lr': 0.0003449447738804503, 'samples': 3849408, 'steps': 20048, 'loss/train': 2.7133124470710754} 01/29/2022 14:38:43 - INFO - codeparrot_training - Step 20049: {'lr': 0.0003449296371367034, 'samples': 3849600, 'steps': 20049, 'loss/train': 1.4144226908683777} 01/29/2022 14:38:47 - INFO - codeparrot_training - Step 20050: {'lr': 0.0003449144999863082, 'samples': 3849792, 'steps': 20050, 'loss/train': 2.9545633792877197} 01/29/2022 14:38:51 - INFO - codeparrot_training - Step 20051: {'lr': 0.00034489936242932935, 'samples': 3849984, 'steps': 20051, 'loss/train': 1.9186345338821411} 01/29/2022 14:38:55 - INFO - codeparrot_training - Step 20052: {'lr': 0.00034488422446583177, 'samples': 3850176, 'steps': 20052, 'loss/train': 1.6749610304832458} 01/29/2022 14:39:00 - INFO - codeparrot_training - Step 20053: {'lr': 0.0003448690860958803, 'samples': 3850368, 'steps': 20053, 'loss/train': 1.5612353682518005} 01/29/2022 14:39:05 - INFO - codeparrot_training - Step 20054: {'lr': 0.00034485394731953976, 'samples': 3850560, 'steps': 20054, 'loss/train': 1.422300785779953} 01/29/2022 14:39:10 - INFO - codeparrot_training - Step 20055: {'lr': 0.00034483880813687505, 'samples': 3850752, 'steps': 20055, 'loss/train': 0.9697716236114502} 01/29/2022 14:39:14 - INFO - codeparrot_training - Step 20056: {'lr': 0.0003448236685479511, 'samples': 3850944, 'steps': 20056, 'loss/train': 0.13715573027729988} 01/29/2022 14:39:18 - INFO - codeparrot_training - Step 20057: {'lr': 0.0003448085285528326, 'samples': 3851136, 'steps': 20057, 'loss/train': 1.5211524367332458} 01/29/2022 14:39:22 - INFO - codeparrot_training - Step 20058: {'lr': 0.00034479338815158447, 'samples': 3851328, 'steps': 20058, 'loss/train': 1.6388179063796997} 01/29/2022 14:39:27 - INFO - codeparrot_training - Step 20059: {'lr': 0.0003447782473442715, 'samples': 3851520, 'steps': 20059, 'loss/train': 1.7031176090240479} 01/29/2022 14:39:33 - INFO - codeparrot_training - Step 20060: {'lr': 0.00034476310613095867, 'samples': 3851712, 'steps': 20060, 'loss/train': 1.19339981675148} 01/29/2022 14:39:37 - INFO - codeparrot_training - Step 20061: {'lr': 0.00034474796451171075, 'samples': 3851904, 'steps': 20061, 'loss/train': 1.984813928604126} 01/29/2022 14:39:41 - INFO - codeparrot_training - Step 20062: {'lr': 0.00034473282248659266, 'samples': 3852096, 'steps': 20062, 'loss/train': 1.9042355418205261} 01/29/2022 14:39:46 - INFO - codeparrot_training - Step 20063: {'lr': 0.00034471768005566925, 'samples': 3852288, 'steps': 20063, 'loss/train': 1.6902102828025818} 01/29/2022 14:39:50 - INFO - codeparrot_training - Step 20064: {'lr': 0.00034470253721900535, 'samples': 3852480, 'steps': 20064, 'loss/train': 1.495532512664795} 01/29/2022 14:39:55 - INFO - codeparrot_training - Step 20065: {'lr': 0.0003446873939766659, 'samples': 3852672, 'steps': 20065, 'loss/train': 1.8598018884658813} 01/29/2022 14:40:00 - INFO - codeparrot_training - Step 20066: {'lr': 0.0003446722503287157, 'samples': 3852864, 'steps': 20066, 'loss/train': 1.193986415863037} 01/29/2022 14:40:04 - INFO - codeparrot_training - Step 20067: {'lr': 0.0003446571062752196, 'samples': 3853056, 'steps': 20067, 'loss/train': 1.8381017446517944} 01/29/2022 14:40:08 - INFO - codeparrot_training - Step 20068: {'lr': 0.0003446419618162425, 'samples': 3853248, 'steps': 20068, 'loss/train': 2.1863513588905334} 01/29/2022 14:40:14 - INFO - codeparrot_training - Step 20069: {'lr': 0.0003446268169518494, 'samples': 3853440, 'steps': 20069, 'loss/train': 1.5939579606056213} 01/29/2022 14:40:18 - INFO - codeparrot_training - Step 20070: {'lr': 0.00034461167168210494, 'samples': 3853632, 'steps': 20070, 'loss/train': 1.4262259304523468} 01/29/2022 14:40:22 - INFO - codeparrot_training - Step 20071: {'lr': 0.00034459652600707423, 'samples': 3853824, 'steps': 20071, 'loss/train': 1.5966092348098755} 01/29/2022 14:40:26 - INFO - codeparrot_training - Step 20072: {'lr': 0.000344581379926822, 'samples': 3854016, 'steps': 20072, 'loss/train': 2.0192341804504395} 01/29/2022 14:40:31 - INFO - codeparrot_training - Step 20073: {'lr': 0.0003445662334414131, 'samples': 3854208, 'steps': 20073, 'loss/train': 2.3481908440589905} 01/29/2022 14:40:37 - INFO - codeparrot_training - Step 20074: {'lr': 0.00034455108655091256, 'samples': 3854400, 'steps': 20074, 'loss/train': 1.992734670639038} 01/29/2022 14:40:42 - INFO - codeparrot_training - Step 20075: {'lr': 0.00034453593925538515, 'samples': 3854592, 'steps': 20075, 'loss/train': 1.523255467414856} 01/29/2022 14:40:46 - INFO - codeparrot_training - Step 20076: {'lr': 0.00034452079155489586, 'samples': 3854784, 'steps': 20076, 'loss/train': 1.7714492082595825} 01/29/2022 14:40:50 - INFO - codeparrot_training - Step 20077: {'lr': 0.00034450564344950944, 'samples': 3854976, 'steps': 20077, 'loss/train': 1.4731105864048004} 01/29/2022 14:40:54 - INFO - codeparrot_training - Step 20078: {'lr': 0.00034449049493929086, 'samples': 3855168, 'steps': 20078, 'loss/train': 1.4247803092002869} 01/29/2022 14:41:00 - INFO - codeparrot_training - Step 20079: {'lr': 0.00034447534602430503, 'samples': 3855360, 'steps': 20079, 'loss/train': 1.2580898702144623} 01/29/2022 14:41:04 - INFO - codeparrot_training - Step 20080: {'lr': 0.00034446019670461683, 'samples': 3855552, 'steps': 20080, 'loss/train': 0.8280675709247589} 01/29/2022 14:41:08 - INFO - codeparrot_training - Step 20081: {'lr': 0.0003444450469802911, 'samples': 3855744, 'steps': 20081, 'loss/train': 1.8747518062591553} 01/29/2022 14:41:12 - INFO - codeparrot_training - Step 20082: {'lr': 0.0003444298968513928, 'samples': 3855936, 'steps': 20082, 'loss/train': 2.3729594349861145} 01/29/2022 14:41:17 - INFO - codeparrot_training - Step 20083: {'lr': 0.0003444147463179868, 'samples': 3856128, 'steps': 20083, 'loss/train': 0.2278796285390854} 01/29/2022 14:43:33 - INFO - codeparrot_training - Step 20084: {'lr': 0.00034439959538013805, 'samples': 3856320, 'steps': 20084, 'loss/train': 1.809497058391571} 01/29/2022 14:43:38 - INFO - codeparrot_training - Step 20085: {'lr': 0.00034438444403791135, 'samples': 3856512, 'steps': 20085, 'loss/train': 1.2593252062797546} 01/29/2022 14:43:42 - INFO - codeparrot_training - Step 20086: {'lr': 0.00034436929229137163, 'samples': 3856704, 'steps': 20086, 'loss/train': 1.8411124348640442} 01/29/2022 14:43:46 - INFO - codeparrot_training - Step 20087: {'lr': 0.00034435414014058393, 'samples': 3856896, 'steps': 20087, 'loss/train': 1.8039600849151611} 01/29/2022 14:43:51 - INFO - codeparrot_training - Step 20088: {'lr': 0.000344338987585613, 'samples': 3857088, 'steps': 20088, 'loss/train': 1.836947500705719} 01/29/2022 14:43:56 - INFO - codeparrot_training - Step 20089: {'lr': 0.0003443238346265238, 'samples': 3857280, 'steps': 20089, 'loss/train': 0.1768769659101963} 01/29/2022 14:44:00 - INFO - codeparrot_training - Step 20090: {'lr': 0.0003443086812633812, 'samples': 3857472, 'steps': 20090, 'loss/train': 1.5348261594772339} 01/29/2022 14:44:04 - INFO - codeparrot_training - Step 20091: {'lr': 0.00034429352749625026, 'samples': 3857664, 'steps': 20091, 'loss/train': 0.2980248034000397} 01/29/2022 14:44:09 - INFO - codeparrot_training - Step 20092: {'lr': 0.00034427837332519573, 'samples': 3857856, 'steps': 20092, 'loss/train': 0.9335902333259583} 01/29/2022 14:44:13 - INFO - codeparrot_training - Step 20093: {'lr': 0.0003442632187502826, 'samples': 3858048, 'steps': 20093, 'loss/train': 2.0462188124656677} 01/29/2022 14:44:18 - INFO - codeparrot_training - Step 20094: {'lr': 0.00034424806377157576, 'samples': 3858240, 'steps': 20094, 'loss/train': 1.6135373711585999} 01/29/2022 14:44:22 - INFO - codeparrot_training - Step 20095: {'lr': 0.0003442329083891402, 'samples': 3858432, 'steps': 20095, 'loss/train': 1.9122095704078674} 01/29/2022 14:44:27 - INFO - codeparrot_training - Step 20096: {'lr': 0.00034421775260304067, 'samples': 3858624, 'steps': 20096, 'loss/train': 2.196041464805603} 01/29/2022 14:44:31 - INFO - codeparrot_training - Step 20097: {'lr': 0.0003442025964133422, 'samples': 3858816, 'steps': 20097, 'loss/train': 1.5499917268753052} 01/29/2022 14:44:35 - INFO - codeparrot_training - Step 20098: {'lr': 0.0003441874398201099, 'samples': 3859008, 'steps': 20098, 'loss/train': 0.6708481907844543} 01/29/2022 14:44:41 - INFO - codeparrot_training - Step 20099: {'lr': 0.00034417228282340837, 'samples': 3859200, 'steps': 20099, 'loss/train': 1.1958065629005432} 01/29/2022 14:44:45 - INFO - codeparrot_training - Step 20100: {'lr': 0.0003441571254233027, 'samples': 3859392, 'steps': 20100, 'loss/train': 1.5445623993873596} 01/29/2022 14:44:49 - INFO - codeparrot_training - Step 20101: {'lr': 0.00034414196761985784, 'samples': 3859584, 'steps': 20101, 'loss/train': 0.8824001848697662} 01/29/2022 14:44:54 - INFO - codeparrot_training - Step 20102: {'lr': 0.00034412680941313866, 'samples': 3859776, 'steps': 20102, 'loss/train': 1.316443383693695} 01/29/2022 14:44:58 - INFO - codeparrot_training - Step 20103: {'lr': 0.00034411165080321007, 'samples': 3859968, 'steps': 20103, 'loss/train': 0.157295610755682} 01/29/2022 14:45:05 - INFO - codeparrot_training - Step 20104: {'lr': 0.00034409649179013716, 'samples': 3860160, 'steps': 20104, 'loss/train': 1.7967220544815063} 01/29/2022 14:45:10 - INFO - codeparrot_training - Step 20105: {'lr': 0.00034408133237398466, 'samples': 3860352, 'steps': 20105, 'loss/train': 1.8475586771965027} 01/29/2022 14:45:14 - INFO - codeparrot_training - Step 20106: {'lr': 0.0003440661725548176, 'samples': 3860544, 'steps': 20106, 'loss/train': 1.802587866783142} 01/29/2022 14:45:18 - INFO - codeparrot_training - Step 20107: {'lr': 0.00034405101233270105, 'samples': 3860736, 'steps': 20107, 'loss/train': 2.018379271030426} 01/29/2022 14:45:22 - INFO - codeparrot_training - Step 20108: {'lr': 0.0003440358517076997, 'samples': 3860928, 'steps': 20108, 'loss/train': 1.7816893458366394} 01/29/2022 14:45:28 - INFO - codeparrot_training - Step 20109: {'lr': 0.00034402069067987874, 'samples': 3861120, 'steps': 20109, 'loss/train': 1.053110808134079} 01/29/2022 14:45:32 - INFO - codeparrot_training - Step 20110: {'lr': 0.0003440055292493029, 'samples': 3861312, 'steps': 20110, 'loss/train': 1.5384095907211304} 01/29/2022 14:45:36 - INFO - codeparrot_training - Step 20111: {'lr': 0.0003439903674160373, 'samples': 3861504, 'steps': 20111, 'loss/train': 2.3601831793785095} 01/29/2022 14:45:41 - INFO - codeparrot_training - Step 20112: {'lr': 0.0003439752051801467, 'samples': 3861696, 'steps': 20112, 'loss/train': 1.0173123478889465} 01/29/2022 14:45:45 - INFO - codeparrot_training - Step 20113: {'lr': 0.0003439600425416963, 'samples': 3861888, 'steps': 20113, 'loss/train': 1.4353854060173035} 01/29/2022 14:45:51 - INFO - codeparrot_training - Step 20114: {'lr': 0.00034394487950075076, 'samples': 3862080, 'steps': 20114, 'loss/train': 1.779433786869049} 01/29/2022 14:45:55 - INFO - codeparrot_training - Step 20115: {'lr': 0.0003439297160573753, 'samples': 3862272, 'steps': 20115, 'loss/train': 1.7051278352737427} 01/29/2022 14:46:00 - INFO - codeparrot_training - Step 20116: {'lr': 0.0003439145522116347, 'samples': 3862464, 'steps': 20116, 'loss/train': 1.8785675168037415} 01/29/2022 14:46:04 - INFO - codeparrot_training - Step 20117: {'lr': 0.000343899387963594, 'samples': 3862656, 'steps': 20117, 'loss/train': 1.6129131317138672} 01/29/2022 14:46:08 - INFO - codeparrot_training - Step 20118: {'lr': 0.00034388422331331817, 'samples': 3862848, 'steps': 20118, 'loss/train': 1.7744182348251343} 01/29/2022 14:46:12 - INFO - codeparrot_training - Step 20119: {'lr': 0.0003438690582608721, 'samples': 3863040, 'steps': 20119, 'loss/train': 1.898685872554779} 01/29/2022 14:46:19 - INFO - codeparrot_training - Step 20120: {'lr': 0.00034385389280632077, 'samples': 3863232, 'steps': 20120, 'loss/train': 1.598275065422058} 01/29/2022 14:46:24 - INFO - codeparrot_training - Step 20121: {'lr': 0.00034383872694972916, 'samples': 3863424, 'steps': 20121, 'loss/train': 1.402668535709381} 01/29/2022 14:46:28 - INFO - codeparrot_training - Step 20122: {'lr': 0.0003438235606911623, 'samples': 3863616, 'steps': 20122, 'loss/train': 1.378655344247818} 01/29/2022 14:46:32 - INFO - codeparrot_training - Step 20123: {'lr': 0.0003438083940306851, 'samples': 3863808, 'steps': 20123, 'loss/train': 1.0715658366680145} 01/29/2022 14:46:36 - INFO - codeparrot_training - Step 20124: {'lr': 0.0003437932269683625, 'samples': 3864000, 'steps': 20124, 'loss/train': 2.246076464653015} 01/29/2022 14:46:42 - INFO - codeparrot_training - Step 20125: {'lr': 0.0003437780595042595, 'samples': 3864192, 'steps': 20125, 'loss/train': 0.8066728413105011} 01/29/2022 14:46:46 - INFO - codeparrot_training - Step 20126: {'lr': 0.0003437628916384411, 'samples': 3864384, 'steps': 20126, 'loss/train': 1.3690118193626404} 01/29/2022 14:46:50 - INFO - codeparrot_training - Step 20127: {'lr': 0.0003437477233709722, 'samples': 3864576, 'steps': 20127, 'loss/train': 0.8243601322174072} 01/29/2022 14:46:54 - INFO - codeparrot_training - Step 20128: {'lr': 0.0003437325547019179, 'samples': 3864768, 'steps': 20128, 'loss/train': 1.3611111044883728} 01/29/2022 14:46:59 - INFO - codeparrot_training - Step 20129: {'lr': 0.000343717385631343, 'samples': 3864960, 'steps': 20129, 'loss/train': 1.4382675290107727} 01/29/2022 14:47:04 - INFO - codeparrot_training - Step 20130: {'lr': 0.00034370221615931265, 'samples': 3865152, 'steps': 20130, 'loss/train': 1.9148186445236206} 01/29/2022 14:47:08 - INFO - codeparrot_training - Step 20131: {'lr': 0.0003436870462858917, 'samples': 3865344, 'steps': 20131, 'loss/train': 1.8971797227859497} 01/29/2022 14:47:13 - INFO - codeparrot_training - Step 20132: {'lr': 0.0003436718760111452, 'samples': 3865536, 'steps': 20132, 'loss/train': 1.4772975146770477} 01/29/2022 14:47:17 - INFO - codeparrot_training - Step 20133: {'lr': 0.00034365670533513813, 'samples': 3865728, 'steps': 20133, 'loss/train': 1.5250146389007568} 01/29/2022 14:47:21 - INFO - codeparrot_training - Step 20134: {'lr': 0.00034364153425793547, 'samples': 3865920, 'steps': 20134, 'loss/train': 1.2085027992725372} 01/29/2022 14:47:28 - INFO - codeparrot_training - Step 20135: {'lr': 0.0003436263627796023, 'samples': 3866112, 'steps': 20135, 'loss/train': 1.960325539112091} 01/29/2022 14:47:33 - INFO - codeparrot_training - Step 20136: {'lr': 0.00034361119090020343, 'samples': 3866304, 'steps': 20136, 'loss/train': 1.842449426651001} 01/29/2022 14:47:37 - INFO - codeparrot_training - Step 20137: {'lr': 0.000343596018619804, 'samples': 3866496, 'steps': 20137, 'loss/train': 1.7517271041870117} 01/29/2022 14:47:41 - INFO - codeparrot_training - Step 20138: {'lr': 0.00034358084593846886, 'samples': 3866688, 'steps': 20138, 'loss/train': 1.9166706204414368} 01/29/2022 14:47:45 - INFO - codeparrot_training - Step 20139: {'lr': 0.00034356567285626316, 'samples': 3866880, 'steps': 20139, 'loss/train': 1.2655912935733795} 01/29/2022 14:47:51 - INFO - codeparrot_training - Step 20140: {'lr': 0.0003435504993732518, 'samples': 3867072, 'steps': 20140, 'loss/train': 1.7599865198135376} 01/29/2022 14:47:55 - INFO - codeparrot_training - Step 20141: {'lr': 0.00034353532548949984, 'samples': 3867264, 'steps': 20141, 'loss/train': 1.4167943894863129} 01/29/2022 14:47:59 - INFO - codeparrot_training - Step 20142: {'lr': 0.0003435201512050722, 'samples': 3867456, 'steps': 20142, 'loss/train': 1.5408039093017578} 01/29/2022 14:48:03 - INFO - codeparrot_training - Step 20143: {'lr': 0.000343504976520034, 'samples': 3867648, 'steps': 20143, 'loss/train': 1.8595726490020752} 01/29/2022 14:48:08 - INFO - codeparrot_training - Step 20144: {'lr': 0.0003434898014344501, 'samples': 3867840, 'steps': 20144, 'loss/train': 2.076372742652893} 01/29/2022 14:48:15 - INFO - codeparrot_training - Step 20145: {'lr': 0.00034347462594838565, 'samples': 3868032, 'steps': 20145, 'loss/train': 1.1456752717494965} 01/29/2022 14:48:19 - INFO - codeparrot_training - Step 20146: {'lr': 0.0003434594500619055, 'samples': 3868224, 'steps': 20146, 'loss/train': 0.9970659613609314} 01/29/2022 14:48:23 - INFO - codeparrot_training - Step 20147: {'lr': 0.00034344427377507475, 'samples': 3868416, 'steps': 20147, 'loss/train': 1.766141653060913} 01/29/2022 14:48:28 - INFO - codeparrot_training - Step 20148: {'lr': 0.00034342909708795846, 'samples': 3868608, 'steps': 20148, 'loss/train': 3.1116875410079956} 01/29/2022 14:48:32 - INFO - codeparrot_training - Step 20149: {'lr': 0.0003434139200006216, 'samples': 3868800, 'steps': 20149, 'loss/train': 1.6213417053222656} 01/29/2022 14:48:38 - INFO - codeparrot_training - Step 20150: {'lr': 0.0003433987425131291, 'samples': 3868992, 'steps': 20150, 'loss/train': 0.6705380827188492} 01/29/2022 14:48:42 - INFO - codeparrot_training - Step 20151: {'lr': 0.0003433835646255461, 'samples': 3869184, 'steps': 20151, 'loss/train': 0.6663770377635956} 01/29/2022 14:48:46 - INFO - codeparrot_training - Step 20152: {'lr': 0.0003433683863379375, 'samples': 3869376, 'steps': 20152, 'loss/train': 1.7140772938728333} 01/29/2022 14:48:51 - INFO - codeparrot_training - Step 20153: {'lr': 0.00034335320765036843, 'samples': 3869568, 'steps': 20153, 'loss/train': 1.6093711853027344} 01/29/2022 14:48:55 - INFO - codeparrot_training - Step 20154: {'lr': 0.0003433380285629039, 'samples': 3869760, 'steps': 20154, 'loss/train': 1.6208800077438354} 01/29/2022 14:48:59 - INFO - codeparrot_training - Step 20155: {'lr': 0.0003433228490756088, 'samples': 3869952, 'steps': 20155, 'loss/train': 1.2851395010948181} 01/29/2022 14:49:04 - INFO - codeparrot_training - Step 20156: {'lr': 0.00034330766918854827, 'samples': 3870144, 'steps': 20156, 'loss/train': 1.8220553398132324} 01/29/2022 14:49:09 - INFO - codeparrot_training - Step 20157: {'lr': 0.0003432924889017873, 'samples': 3870336, 'steps': 20157, 'loss/train': 1.9040733575820923} 01/29/2022 14:49:13 - INFO - codeparrot_training - Step 20158: {'lr': 0.000343277308215391, 'samples': 3870528, 'steps': 20158, 'loss/train': 0.7572324872016907} 01/29/2022 14:49:17 - INFO - codeparrot_training - Step 20159: {'lr': 0.0003432621271294243, 'samples': 3870720, 'steps': 20159, 'loss/train': 1.5827202200889587} 01/29/2022 14:49:21 - INFO - codeparrot_training - Step 20160: {'lr': 0.00034324694564395226, 'samples': 3870912, 'steps': 20160, 'loss/train': 1.9285009503364563} 01/29/2022 14:49:27 - INFO - codeparrot_training - Step 20161: {'lr': 0.0003432317637590399, 'samples': 3871104, 'steps': 20161, 'loss/train': 1.39308562874794} 01/29/2022 14:49:31 - INFO - codeparrot_training - Step 20162: {'lr': 0.0003432165814747523, 'samples': 3871296, 'steps': 20162, 'loss/train': 1.4054026901721954} 01/29/2022 14:49:35 - INFO - codeparrot_training - Step 20163: {'lr': 0.0003432013987911544, 'samples': 3871488, 'steps': 20163, 'loss/train': 1.3039517998695374} 01/29/2022 14:49:39 - INFO - codeparrot_training - Step 20164: {'lr': 0.0003431862157083114, 'samples': 3871680, 'steps': 20164, 'loss/train': 1.5070322155952454} 01/29/2022 14:49:44 - INFO - codeparrot_training - Step 20165: {'lr': 0.0003431710322262882, 'samples': 3871872, 'steps': 20165, 'loss/train': 0.46209418773651123} 01/29/2022 14:49:51 - INFO - codeparrot_training - Step 20166: {'lr': 0.0003431558483451498, 'samples': 3872064, 'steps': 20166, 'loss/train': 1.75504070520401} 01/29/2022 14:49:55 - INFO - codeparrot_training - Step 20167: {'lr': 0.00034314066406496146, 'samples': 3872256, 'steps': 20167, 'loss/train': 0.5171753168106079} 01/29/2022 14:49:59 - INFO - codeparrot_training - Step 20168: {'lr': 0.00034312547938578796, 'samples': 3872448, 'steps': 20168, 'loss/train': 0.9004482328891754} 01/29/2022 14:50:03 - INFO - codeparrot_training - Step 20169: {'lr': 0.0003431102943076946, 'samples': 3872640, 'steps': 20169, 'loss/train': 0.6404382884502411} 01/29/2022 14:50:08 - INFO - codeparrot_training - Step 20170: {'lr': 0.00034309510883074625, 'samples': 3872832, 'steps': 20170, 'loss/train': 1.8279551267623901} 01/29/2022 14:50:13 - INFO - codeparrot_training - Step 20171: {'lr': 0.0003430799229550079, 'samples': 3873024, 'steps': 20171, 'loss/train': 2.5164161324501038} 01/29/2022 14:50:17 - INFO - codeparrot_training - Step 20172: {'lr': 0.0003430647366805449, 'samples': 3873216, 'steps': 20172, 'loss/train': 2.803004264831543} 01/29/2022 14:50:22 - INFO - codeparrot_training - Step 20173: {'lr': 0.000343049550007422, 'samples': 3873408, 'steps': 20173, 'loss/train': 1.5209051370620728} 01/29/2022 14:50:26 - INFO - codeparrot_training - Step 20174: {'lr': 0.0003430343629357044, 'samples': 3873600, 'steps': 20174, 'loss/train': 1.2292855381965637} 01/29/2022 14:50:31 - INFO - codeparrot_training - Step 20175: {'lr': 0.0003430191754654572, 'samples': 3873792, 'steps': 20175, 'loss/train': 1.8968185186386108} 01/29/2022 14:50:36 - INFO - codeparrot_training - Step 20176: {'lr': 0.0003430039875967454, 'samples': 3873984, 'steps': 20176, 'loss/train': 2.1927096247673035} 01/29/2022 14:50:40 - INFO - codeparrot_training - Step 20177: {'lr': 0.00034298879932963397, 'samples': 3874176, 'steps': 20177, 'loss/train': 1.3722452223300934} 01/29/2022 14:50:44 - INFO - codeparrot_training - Step 20178: {'lr': 0.0003429736106641881, 'samples': 3874368, 'steps': 20178, 'loss/train': 1.4646147787570953} 01/29/2022 14:50:48 - INFO - codeparrot_training - Step 20179: {'lr': 0.0003429584216004728, 'samples': 3874560, 'steps': 20179, 'loss/train': 1.3869905769824982} 01/29/2022 14:50:55 - INFO - codeparrot_training - Step 20180: {'lr': 0.0003429432321385531, 'samples': 3874752, 'steps': 20180, 'loss/train': 1.8176754713058472} 01/29/2022 14:51:00 - INFO - codeparrot_training - Step 20181: {'lr': 0.00034292804227849407, 'samples': 3874944, 'steps': 20181, 'loss/train': 1.3470846712589264} 01/29/2022 14:51:04 - INFO - codeparrot_training - Step 20182: {'lr': 0.0003429128520203608, 'samples': 3875136, 'steps': 20182, 'loss/train': 1.396623283624649} 01/29/2022 14:51:08 - INFO - codeparrot_training - Step 20183: {'lr': 0.00034289766136421854, 'samples': 3875328, 'steps': 20183, 'loss/train': 1.7885704636573792} 01/29/2022 14:51:12 - INFO - codeparrot_training - Step 20184: {'lr': 0.000342882470310132, 'samples': 3875520, 'steps': 20184, 'loss/train': 1.5133752822875977} 01/29/2022 14:51:18 - INFO - codeparrot_training - Step 20185: {'lr': 0.0003428672788581666, 'samples': 3875712, 'steps': 20185, 'loss/train': 1.3819350600242615} 01/29/2022 14:51:23 - INFO - codeparrot_training - Step 20186: {'lr': 0.0003428520870083872, 'samples': 3875904, 'steps': 20186, 'loss/train': 2.225423276424408} 01/29/2022 14:51:27 - INFO - codeparrot_training - Step 20187: {'lr': 0.000342836894760859, 'samples': 3876096, 'steps': 20187, 'loss/train': 1.5695244669914246} 01/29/2022 14:51:31 - INFO - codeparrot_training - Step 20188: {'lr': 0.00034282170211564697, 'samples': 3876288, 'steps': 20188, 'loss/train': 1.7944058775901794} 01/29/2022 14:51:35 - INFO - codeparrot_training - Step 20189: {'lr': 0.0003428065090728163, 'samples': 3876480, 'steps': 20189, 'loss/train': 2.5365915298461914} 01/29/2022 14:51:40 - INFO - codeparrot_training - Step 20190: {'lr': 0.0003427913156324319, 'samples': 3876672, 'steps': 20190, 'loss/train': 0.6410530507564545} 01/29/2022 14:51:47 - INFO - codeparrot_training - Step 20191: {'lr': 0.00034277612179455907, 'samples': 3876864, 'steps': 20191, 'loss/train': 2.1465813517570496} 01/29/2022 14:51:51 - INFO - codeparrot_training - Step 20192: {'lr': 0.00034276092755926275, 'samples': 3877056, 'steps': 20192, 'loss/train': 1.3814224898815155} 01/29/2022 14:51:55 - INFO - codeparrot_training - Step 20193: {'lr': 0.0003427457329266081, 'samples': 3877248, 'steps': 20193, 'loss/train': 1.8003657460212708} 01/29/2022 14:51:59 - INFO - codeparrot_training - Step 20194: {'lr': 0.0003427305378966601, 'samples': 3877440, 'steps': 20194, 'loss/train': 0.9633331596851349} 01/29/2022 14:52:04 - INFO - codeparrot_training - Step 20195: {'lr': 0.00034271534246948403, 'samples': 3877632, 'steps': 20195, 'loss/train': 1.0341107547283173} 01/29/2022 14:52:09 - INFO - codeparrot_training - Step 20196: {'lr': 0.0003427001466451448, 'samples': 3877824, 'steps': 20196, 'loss/train': 0.9920843839645386} 01/29/2022 14:52:13 - INFO - codeparrot_training - Step 20197: {'lr': 0.00034268495042370767, 'samples': 3878016, 'steps': 20197, 'loss/train': 0.8779277801513672} 01/29/2022 14:52:17 - INFO - codeparrot_training - Step 20198: {'lr': 0.00034266975380523756, 'samples': 3878208, 'steps': 20198, 'loss/train': 1.177339643239975} 01/29/2022 14:52:22 - INFO - codeparrot_training - Step 20199: {'lr': 0.00034265455678979967, 'samples': 3878400, 'steps': 20199, 'loss/train': 1.940559446811676} 01/29/2022 14:52:26 - INFO - codeparrot_training - Step 20200: {'lr': 0.0003426393593774591, 'samples': 3878592, 'steps': 20200, 'loss/train': 1.9352204203605652} 01/29/2022 14:52:31 - INFO - codeparrot_training - Step 20201: {'lr': 0.0003426241615682809, 'samples': 3878784, 'steps': 20201, 'loss/train': 1.8082484006881714} 01/29/2022 14:52:35 - INFO - codeparrot_training - Step 20202: {'lr': 0.0003426089633623302, 'samples': 3878976, 'steps': 20202, 'loss/train': 1.9255594611167908} 01/29/2022 14:52:40 - INFO - codeparrot_training - Step 20203: {'lr': 0.0003425937647596721, 'samples': 3879168, 'steps': 20203, 'loss/train': 1.7149291634559631} 01/29/2022 14:52:44 - INFO - codeparrot_training - Step 20204: {'lr': 0.0003425785657603718, 'samples': 3879360, 'steps': 20204, 'loss/train': 0.5063153654336929} 01/29/2022 14:52:48 - INFO - codeparrot_training - Step 20205: {'lr': 0.0003425633663644942, 'samples': 3879552, 'steps': 20205, 'loss/train': 1.7759916186332703} 01/29/2022 14:52:55 - INFO - codeparrot_training - Step 20206: {'lr': 0.00034254816657210455, 'samples': 3879744, 'steps': 20206, 'loss/train': 1.9452940821647644} 01/29/2022 14:52:59 - INFO - codeparrot_training - Step 20207: {'lr': 0.00034253296638326805, 'samples': 3879936, 'steps': 20207, 'loss/train': 2.2454829812049866} 01/29/2022 14:53:04 - INFO - codeparrot_training - Step 20208: {'lr': 0.0003425177657980496, 'samples': 3880128, 'steps': 20208, 'loss/train': 1.454706072807312} 01/29/2022 14:53:08 - INFO - codeparrot_training - Step 20209: {'lr': 0.0003425025648165145, 'samples': 3880320, 'steps': 20209, 'loss/train': 0.09608327597379684} 01/29/2022 14:53:12 - INFO - codeparrot_training - Step 20210: {'lr': 0.00034248736343872767, 'samples': 3880512, 'steps': 20210, 'loss/train': 1.1553959548473358} 01/29/2022 14:53:18 - INFO - codeparrot_training - Step 20211: {'lr': 0.0003424721616647544, 'samples': 3880704, 'steps': 20211, 'loss/train': 2.395691156387329} 01/29/2022 14:53:22 - INFO - codeparrot_training - Step 20212: {'lr': 0.00034245695949465977, 'samples': 3880896, 'steps': 20212, 'loss/train': 2.1652450561523438} 01/29/2022 14:53:26 - INFO - codeparrot_training - Step 20213: {'lr': 0.00034244175692850894, 'samples': 3881088, 'steps': 20213, 'loss/train': 1.8616236448287964} 01/29/2022 14:53:30 - INFO - codeparrot_training - Step 20214: {'lr': 0.00034242655396636687, 'samples': 3881280, 'steps': 20214, 'loss/train': 1.3511251509189606} 01/29/2022 14:53:35 - INFO - codeparrot_training - Step 20215: {'lr': 0.0003424113506082989, 'samples': 3881472, 'steps': 20215, 'loss/train': 0.6315525323152542} 01/29/2022 14:53:40 - INFO - codeparrot_training - Step 20216: {'lr': 0.00034239614685436994, 'samples': 3881664, 'steps': 20216, 'loss/train': 1.5084174871444702} 01/29/2022 14:53:45 - INFO - codeparrot_training - Step 20217: {'lr': 0.00034238094270464523, 'samples': 3881856, 'steps': 20217, 'loss/train': 1.8950550556182861} 01/29/2022 14:53:49 - INFO - codeparrot_training - Step 20218: {'lr': 0.00034236573815918993, 'samples': 3882048, 'steps': 20218, 'loss/train': 1.6301168203353882} 01/29/2022 14:53:53 - INFO - codeparrot_training - Step 20219: {'lr': 0.00034235053321806915, 'samples': 3882240, 'steps': 20219, 'loss/train': 0.8721860647201538} 01/29/2022 14:53:57 - INFO - codeparrot_training - Step 20220: {'lr': 0.00034233532788134803, 'samples': 3882432, 'steps': 20220, 'loss/train': 1.1619457304477692} 01/29/2022 14:54:03 - INFO - codeparrot_training - Step 20221: {'lr': 0.0003423201221490916, 'samples': 3882624, 'steps': 20221, 'loss/train': 2.00116628408432} 01/29/2022 14:54:07 - INFO - codeparrot_training - Step 20222: {'lr': 0.00034230491602136513, 'samples': 3882816, 'steps': 20222, 'loss/train': 1.4085693061351776} 01/29/2022 14:54:11 - INFO - codeparrot_training - Step 20223: {'lr': 0.0003422897094982337, 'samples': 3883008, 'steps': 20223, 'loss/train': 2.3569064140319824} 01/29/2022 14:54:16 - INFO - codeparrot_training - Step 20224: {'lr': 0.0003422745025797626, 'samples': 3883200, 'steps': 20224, 'loss/train': 0.6640467345714569} 01/29/2022 14:54:20 - INFO - codeparrot_training - Step 20225: {'lr': 0.00034225929526601664, 'samples': 3883392, 'steps': 20225, 'loss/train': 1.6291138529777527} 01/29/2022 14:54:26 - INFO - codeparrot_training - Step 20226: {'lr': 0.0003422440875570612, 'samples': 3883584, 'steps': 20226, 'loss/train': 1.2989445626735687} 01/29/2022 14:54:31 - INFO - codeparrot_training - Step 20227: {'lr': 0.00034222887945296144, 'samples': 3883776, 'steps': 20227, 'loss/train': 0.914666086435318} 01/29/2022 14:54:35 - INFO - codeparrot_training - Step 20228: {'lr': 0.0003422136709537824, 'samples': 3883968, 'steps': 20228, 'loss/train': 1.551994800567627} 01/29/2022 14:54:39 - INFO - codeparrot_training - Step 20229: {'lr': 0.00034219846205958926, 'samples': 3884160, 'steps': 20229, 'loss/train': 1.7309321165084839} 01/29/2022 14:54:43 - INFO - codeparrot_training - Step 20230: {'lr': 0.0003421832527704471, 'samples': 3884352, 'steps': 20230, 'loss/train': 2.0439469814300537} 01/29/2022 14:54:49 - INFO - codeparrot_training - Step 20231: {'lr': 0.0003421680430864214, 'samples': 3884544, 'steps': 20231, 'loss/train': 1.0870542526245117} 01/29/2022 14:54:53 - INFO - codeparrot_training - Step 20232: {'lr': 0.0003421528330075769, 'samples': 3884736, 'steps': 20232, 'loss/train': 2.4030057191848755} 01/29/2022 14:54:57 - INFO - codeparrot_training - Step 20233: {'lr': 0.00034213762253397896, 'samples': 3884928, 'steps': 20233, 'loss/train': 1.5720697045326233} 01/29/2022 14:55:01 - INFO - codeparrot_training - Step 20234: {'lr': 0.0003421224116656927, 'samples': 3885120, 'steps': 20234, 'loss/train': 1.9227195382118225} 01/29/2022 14:55:06 - INFO - codeparrot_training - Step 20235: {'lr': 0.0003421072004027833, 'samples': 3885312, 'steps': 20235, 'loss/train': 1.3529155254364014} 01/29/2022 14:55:11 - INFO - codeparrot_training - Step 20236: {'lr': 0.00034209198874531586, 'samples': 3885504, 'steps': 20236, 'loss/train': 1.3175978064537048} 01/29/2022 14:55:15 - INFO - codeparrot_training - Step 20237: {'lr': 0.00034207677669335565, 'samples': 3885696, 'steps': 20237, 'loss/train': 1.518036425113678} 01/29/2022 14:55:20 - INFO - codeparrot_training - Step 20238: {'lr': 0.0003420615642469678, 'samples': 3885888, 'steps': 20238, 'loss/train': 2.290502607822418} 01/29/2022 14:55:24 - INFO - codeparrot_training - Step 20239: {'lr': 0.00034204635140621726, 'samples': 3886080, 'steps': 20239, 'loss/train': 1.327254742383957} 01/29/2022 14:55:28 - INFO - codeparrot_training - Step 20240: {'lr': 0.00034203113817116957, 'samples': 3886272, 'steps': 20240, 'loss/train': 0.7432199120521545} 01/29/2022 14:55:35 - INFO - codeparrot_training - Step 20241: {'lr': 0.0003420159245418896, 'samples': 3886464, 'steps': 20241, 'loss/train': 2.3497318625450134} 01/29/2022 14:55:40 - INFO - codeparrot_training - Step 20242: {'lr': 0.0003420007105184426, 'samples': 3886656, 'steps': 20242, 'loss/train': 1.9220917224884033} 01/29/2022 14:55:44 - INFO - codeparrot_training - Step 20243: {'lr': 0.0003419854961008938, 'samples': 3886848, 'steps': 20243, 'loss/train': 1.283949762582779} 01/29/2022 14:55:48 - INFO - codeparrot_training - Step 20244: {'lr': 0.0003419702812893084, 'samples': 3887040, 'steps': 20244, 'loss/train': 1.0023605525493622} 01/29/2022 14:55:52 - INFO - codeparrot_training - Step 20245: {'lr': 0.0003419550660837515, 'samples': 3887232, 'steps': 20245, 'loss/train': 2.1281375885009766} 01/29/2022 14:55:58 - INFO - codeparrot_training - Step 20246: {'lr': 0.0003419398504842883, 'samples': 3887424, 'steps': 20246, 'loss/train': 1.4497916400432587} 01/29/2022 14:56:02 - INFO - codeparrot_training - Step 20247: {'lr': 0.00034192463449098386, 'samples': 3887616, 'steps': 20247, 'loss/train': 1.9439932107925415} 01/29/2022 14:56:06 - INFO - codeparrot_training - Step 20248: {'lr': 0.00034190941810390365, 'samples': 3887808, 'steps': 20248, 'loss/train': 1.536878228187561} 01/29/2022 14:56:10 - INFO - codeparrot_training - Step 20249: {'lr': 0.00034189420132311256, 'samples': 3888000, 'steps': 20249, 'loss/train': 1.9175230264663696} 01/29/2022 14:56:15 - INFO - codeparrot_training - Step 20250: {'lr': 0.0003418789841486759, 'samples': 3888192, 'steps': 20250, 'loss/train': 2.02079701423645} 01/29/2022 14:56:22 - INFO - codeparrot_training - Step 20251: {'lr': 0.0003418637665806589, 'samples': 3888384, 'steps': 20251, 'loss/train': 2.2766536474227905} 01/29/2022 14:56:26 - INFO - codeparrot_training - Step 20252: {'lr': 0.0003418485486191267, 'samples': 3888576, 'steps': 20252, 'loss/train': 1.4928991198539734} 01/29/2022 14:56:30 - INFO - codeparrot_training - Step 20253: {'lr': 0.0003418333302641444, 'samples': 3888768, 'steps': 20253, 'loss/train': 2.113327145576477} 01/29/2022 14:56:34 - INFO - codeparrot_training - Step 20254: {'lr': 0.0003418181115157774, 'samples': 3888960, 'steps': 20254, 'loss/train': 2.676075518131256} 01/29/2022 14:56:39 - INFO - codeparrot_training - Step 20255: {'lr': 0.00034180289237409063, 'samples': 3889152, 'steps': 20255, 'loss/train': 1.9080480337142944} 01/29/2022 14:56:44 - INFO - codeparrot_training - Step 20256: {'lr': 0.00034178767283914944, 'samples': 3889344, 'steps': 20256, 'loss/train': 1.87123042345047} 01/29/2022 14:56:48 - INFO - codeparrot_training - Step 20257: {'lr': 0.000341772452911019, 'samples': 3889536, 'steps': 20257, 'loss/train': 2.1276084780693054} 01/29/2022 14:56:53 - INFO - codeparrot_training - Step 20258: {'lr': 0.0003417572325897646, 'samples': 3889728, 'steps': 20258, 'loss/train': 1.3970601260662079} 01/29/2022 14:56:57 - INFO - codeparrot_training - Step 20259: {'lr': 0.00034174201187545133, 'samples': 3889920, 'steps': 20259, 'loss/train': 1.4650062024593353} 01/29/2022 14:57:01 - INFO - codeparrot_training - Step 20260: {'lr': 0.00034172679076814437, 'samples': 3890112, 'steps': 20260, 'loss/train': 1.1701010763645172} 01/29/2022 14:57:06 - INFO - codeparrot_training - Step 20261: {'lr': 0.00034171156926790904, 'samples': 3890304, 'steps': 20261, 'loss/train': 2.2573299407958984} 01/29/2022 14:57:10 - INFO - codeparrot_training - Step 20262: {'lr': 0.00034169634737481034, 'samples': 3890496, 'steps': 20262, 'loss/train': 1.2232737243175507} 01/29/2022 14:57:15 - INFO - codeparrot_training - Step 20263: {'lr': 0.0003416811250889137, 'samples': 3890688, 'steps': 20263, 'loss/train': 1.592911183834076} 01/29/2022 14:57:19 - INFO - codeparrot_training - Step 20264: {'lr': 0.00034166590241028425, 'samples': 3890880, 'steps': 20264, 'loss/train': 1.9228271842002869} 01/29/2022 14:57:23 - INFO - codeparrot_training - Step 20265: {'lr': 0.0003416506793389871, 'samples': 3891072, 'steps': 20265, 'loss/train': 1.0099842846393585} 01/29/2022 14:57:30 - INFO - codeparrot_training - Step 20266: {'lr': 0.0003416354558750876, 'samples': 3891264, 'steps': 20266, 'loss/train': 2.090367615222931} 01/29/2022 14:57:35 - INFO - codeparrot_training - Step 20267: {'lr': 0.0003416202320186508, 'samples': 3891456, 'steps': 20267, 'loss/train': 1.8091199398040771} 01/29/2022 14:57:39 - INFO - codeparrot_training - Step 20268: {'lr': 0.0003416050077697422, 'samples': 3891648, 'steps': 20268, 'loss/train': 1.8624234795570374} 01/29/2022 14:57:43 - INFO - codeparrot_training - Step 20269: {'lr': 0.0003415897831284267, 'samples': 3891840, 'steps': 20269, 'loss/train': 2.0738450288772583} 01/29/2022 14:57:47 - INFO - codeparrot_training - Step 20270: {'lr': 0.0003415745580947697, 'samples': 3892032, 'steps': 20270, 'loss/train': 2.1260252594947815} 01/29/2022 14:57:53 - INFO - codeparrot_training - Step 20271: {'lr': 0.0003415593326688364, 'samples': 3892224, 'steps': 20271, 'loss/train': 1.638635516166687} 01/29/2022 14:57:57 - INFO - codeparrot_training - Step 20272: {'lr': 0.0003415441068506919, 'samples': 3892416, 'steps': 20272, 'loss/train': 0.7271547317504883} 01/29/2022 14:58:01 - INFO - codeparrot_training - Step 20273: {'lr': 0.0003415288806404016, 'samples': 3892608, 'steps': 20273, 'loss/train': 0.9647633135318756} 01/29/2022 14:58:06 - INFO - codeparrot_training - Step 20274: {'lr': 0.00034151365403803065, 'samples': 3892800, 'steps': 20274, 'loss/train': 1.447198748588562} 01/29/2022 14:58:10 - INFO - codeparrot_training - Step 20275: {'lr': 0.0003414984270436442, 'samples': 3892992, 'steps': 20275, 'loss/train': 0.9622863829135895} 01/29/2022 14:58:16 - INFO - codeparrot_training - Step 20276: {'lr': 0.00034148319965730757, 'samples': 3893184, 'steps': 20276, 'loss/train': 1.6975994110107422} 01/29/2022 14:58:20 - INFO - codeparrot_training - Step 20277: {'lr': 0.000341467971879086, 'samples': 3893376, 'steps': 20277, 'loss/train': 1.1993383169174194} 01/29/2022 14:58:25 - INFO - codeparrot_training - Step 20278: {'lr': 0.0003414527437090446, 'samples': 3893568, 'steps': 20278, 'loss/train': 1.2681516408920288} 01/29/2022 14:58:29 - INFO - codeparrot_training - Step 20279: {'lr': 0.00034143751514724874, 'samples': 3893760, 'steps': 20279, 'loss/train': 1.4550232887268066} 01/29/2022 14:58:33 - INFO - codeparrot_training - Step 20280: {'lr': 0.0003414222861937636, 'samples': 3893952, 'steps': 20280, 'loss/train': 1.8193180561065674} 01/29/2022 14:58:37 - INFO - codeparrot_training - Step 20281: {'lr': 0.00034140705684865437, 'samples': 3894144, 'steps': 20281, 'loss/train': 1.729939341545105} 01/29/2022 14:58:44 - INFO - codeparrot_training - Step 20282: {'lr': 0.0003413918271119864, 'samples': 3894336, 'steps': 20282, 'loss/train': 1.5059775710105896} 01/29/2022 14:58:49 - INFO - codeparrot_training - Step 20283: {'lr': 0.00034137659698382485, 'samples': 3894528, 'steps': 20283, 'loss/train': 2.1794444918632507} 01/29/2022 14:58:53 - INFO - codeparrot_training - Step 20284: {'lr': 0.0003413613664642349, 'samples': 3894720, 'steps': 20284, 'loss/train': 1.3043635189533234} 01/29/2022 14:58:57 - INFO - codeparrot_training - Step 20285: {'lr': 0.00034134613555328195, 'samples': 3894912, 'steps': 20285, 'loss/train': 1.752720594406128} 01/29/2022 14:59:01 - INFO - codeparrot_training - Step 20286: {'lr': 0.00034133090425103114, 'samples': 3895104, 'steps': 20286, 'loss/train': 1.9857271313667297} 01/29/2022 14:59:07 - INFO - codeparrot_training - Step 20287: {'lr': 0.00034131567255754776, 'samples': 3895296, 'steps': 20287, 'loss/train': 1.858414649963379} 01/29/2022 14:59:11 - INFO - codeparrot_training - Step 20288: {'lr': 0.000341300440472897, 'samples': 3895488, 'steps': 20288, 'loss/train': 0.9452903866767883} 01/29/2022 14:59:15 - INFO - codeparrot_training - Step 20289: {'lr': 0.0003412852079971441, 'samples': 3895680, 'steps': 20289, 'loss/train': 0.5946394354104996} 01/29/2022 14:59:19 - INFO - codeparrot_training - Step 20290: {'lr': 0.0003412699751303544, 'samples': 3895872, 'steps': 20290, 'loss/train': 1.8937674164772034} 01/29/2022 14:59:24 - INFO - codeparrot_training - Step 20291: {'lr': 0.00034125474187259307, 'samples': 3896064, 'steps': 20291, 'loss/train': 1.72637540102005} 01/29/2022 14:59:29 - INFO - codeparrot_training - Step 20292: {'lr': 0.0003412395082239255, 'samples': 3896256, 'steps': 20292, 'loss/train': 0.7661410868167877} 01/29/2022 14:59:33 - INFO - codeparrot_training - Step 20293: {'lr': 0.0003412242741844168, 'samples': 3896448, 'steps': 20293, 'loss/train': 1.9728332161903381} 01/29/2022 14:59:37 - INFO - codeparrot_training - Step 20294: {'lr': 0.0003412090397541323, 'samples': 3896640, 'steps': 20294, 'loss/train': 1.1735967993736267} 01/29/2022 14:59:42 - INFO - codeparrot_training - Step 20295: {'lr': 0.0003411938049331372, 'samples': 3896832, 'steps': 20295, 'loss/train': 1.0972967147827148} 01/29/2022 14:59:46 - INFO - codeparrot_training - Step 20296: {'lr': 0.0003411785697214968, 'samples': 3897024, 'steps': 20296, 'loss/train': 1.328143984079361} 01/29/2022 14:59:53 - INFO - codeparrot_training - Step 20297: {'lr': 0.00034116333411927637, 'samples': 3897216, 'steps': 20297, 'loss/train': 0.8541306853294373} 01/29/2022 14:59:57 - INFO - codeparrot_training - Step 20298: {'lr': 0.0003411480981265411, 'samples': 3897408, 'steps': 20298, 'loss/train': 1.7542451620101929} 01/29/2022 15:00:02 - INFO - codeparrot_training - Step 20299: {'lr': 0.0003411328617433564, 'samples': 3897600, 'steps': 20299, 'loss/train': 1.6225116848945618} 01/29/2022 15:00:06 - INFO - codeparrot_training - Step 20300: {'lr': 0.0003411176249697875, 'samples': 3897792, 'steps': 20300, 'loss/train': 1.5602408051490784} 01/29/2022 15:00:11 - INFO - codeparrot_training - Step 20301: {'lr': 0.0003411023878058995, 'samples': 3897984, 'steps': 20301, 'loss/train': 1.892921268939972} 01/29/2022 15:00:15 - INFO - codeparrot_training - Step 20302: {'lr': 0.0003410871502517579, 'samples': 3898176, 'steps': 20302, 'loss/train': 0.5798148661851883} 01/29/2022 15:00:20 - INFO - codeparrot_training - Step 20303: {'lr': 0.00034107191230742776, 'samples': 3898368, 'steps': 20303, 'loss/train': 1.922489047050476} 01/29/2022 15:00:24 - INFO - codeparrot_training - Step 20304: {'lr': 0.00034105667397297455, 'samples': 3898560, 'steps': 20304, 'loss/train': 2.190313160419464} 01/29/2022 15:00:28 - INFO - codeparrot_training - Step 20305: {'lr': 0.0003410414352484635, 'samples': 3898752, 'steps': 20305, 'loss/train': 1.6520102620124817} 01/29/2022 15:00:33 - INFO - codeparrot_training - Step 20306: {'lr': 0.00034102619613395974, 'samples': 3898944, 'steps': 20306, 'loss/train': 0.9862254559993744} 01/29/2022 15:00:38 - INFO - codeparrot_training - Step 20307: {'lr': 0.00034101095662952873, 'samples': 3899136, 'steps': 20307, 'loss/train': 1.8974343538284302} 01/29/2022 15:00:42 - INFO - codeparrot_training - Step 20308: {'lr': 0.00034099571673523564, 'samples': 3899328, 'steps': 20308, 'loss/train': 1.1622766256332397} 01/29/2022 15:00:46 - INFO - codeparrot_training - Step 20309: {'lr': 0.0003409804764511459, 'samples': 3899520, 'steps': 20309, 'loss/train': 1.6445417404174805} 01/29/2022 15:00:50 - INFO - codeparrot_training - Step 20310: {'lr': 0.00034096523577732457, 'samples': 3899712, 'steps': 20310, 'loss/train': 1.5256378054618835} 01/29/2022 15:00:58 - INFO - codeparrot_training - Step 20311: {'lr': 0.00034094999471383713, 'samples': 3899904, 'steps': 20311, 'loss/train': 1.5400564670562744} 01/29/2022 15:01:02 - INFO - codeparrot_training - Step 20312: {'lr': 0.00034093475326074874, 'samples': 3900096, 'steps': 20312, 'loss/train': 1.635098934173584} 01/29/2022 15:01:06 - INFO - codeparrot_training - Step 20313: {'lr': 0.00034091951141812483, 'samples': 3900288, 'steps': 20313, 'loss/train': 1.9037883281707764} 01/29/2022 15:01:10 - INFO - codeparrot_training - Step 20314: {'lr': 0.00034090426918603045, 'samples': 3900480, 'steps': 20314, 'loss/train': 2.2608938813209534} 01/29/2022 15:01:15 - INFO - codeparrot_training - Step 20315: {'lr': 0.00034088902656453116, 'samples': 3900672, 'steps': 20315, 'loss/train': 0.44232138991355896} 01/29/2022 15:01:20 - INFO - codeparrot_training - Step 20316: {'lr': 0.00034087378355369215, 'samples': 3900864, 'steps': 20316, 'loss/train': 2.1430492401123047} 01/29/2022 15:01:24 - INFO - codeparrot_training - Step 20317: {'lr': 0.00034085854015357864, 'samples': 3901056, 'steps': 20317, 'loss/train': 0.4315869212150574} 01/29/2022 15:01:28 - INFO - codeparrot_training - Step 20318: {'lr': 0.000340843296364256, 'samples': 3901248, 'steps': 20318, 'loss/train': 2.0481619834899902} 01/29/2022 15:01:33 - INFO - codeparrot_training - Step 20319: {'lr': 0.00034082805218578954, 'samples': 3901440, 'steps': 20319, 'loss/train': 0.8643024265766144} 01/29/2022 15:01:37 - INFO - codeparrot_training - Step 20320: {'lr': 0.00034081280761824465, 'samples': 3901632, 'steps': 20320, 'loss/train': 1.7074878215789795} 01/29/2022 15:01:44 - INFO - codeparrot_training - Step 20321: {'lr': 0.0003407975626616864, 'samples': 3901824, 'steps': 20321, 'loss/train': 1.7395302057266235} 01/29/2022 15:01:48 - INFO - codeparrot_training - Step 20322: {'lr': 0.00034078231731618025, 'samples': 3902016, 'steps': 20322, 'loss/train': 1.9136971235275269} 01/29/2022 15:01:53 - INFO - codeparrot_training - Step 20323: {'lr': 0.00034076707158179145, 'samples': 3902208, 'steps': 20323, 'loss/train': 1.3440645039081573} 01/29/2022 15:01:57 - INFO - codeparrot_training - Step 20324: {'lr': 0.0003407518254585854, 'samples': 3902400, 'steps': 20324, 'loss/train': 1.468199461698532} 01/29/2022 15:02:01 - INFO - codeparrot_training - Step 20325: {'lr': 0.0003407365789466273, 'samples': 3902592, 'steps': 20325, 'loss/train': 1.740078628063202} 01/29/2022 15:02:06 - INFO - codeparrot_training - Step 20326: {'lr': 0.00034072133204598255, 'samples': 3902784, 'steps': 20326, 'loss/train': 1.1815299689769745} 01/29/2022 15:02:11 - INFO - codeparrot_training - Step 20327: {'lr': 0.00034070608475671643, 'samples': 3902976, 'steps': 20327, 'loss/train': 1.597914218902588} 01/29/2022 15:02:15 - INFO - codeparrot_training - Step 20328: {'lr': 0.0003406908370788942, 'samples': 3903168, 'steps': 20328, 'loss/train': 0.9041950106620789} 01/29/2022 15:02:19 - INFO - codeparrot_training - Step 20329: {'lr': 0.0003406755890125813, 'samples': 3903360, 'steps': 20329, 'loss/train': 1.2680225372314453} 01/29/2022 15:02:23 - INFO - codeparrot_training - Step 20330: {'lr': 0.00034066034055784284, 'samples': 3903552, 'steps': 20330, 'loss/train': 1.4818372428417206} 01/29/2022 15:02:29 - INFO - codeparrot_training - Step 20331: {'lr': 0.00034064509171474435, 'samples': 3903744, 'steps': 20331, 'loss/train': 1.9432284235954285} 01/29/2022 15:02:33 - INFO - codeparrot_training - Step 20332: {'lr': 0.00034062984248335105, 'samples': 3903936, 'steps': 20332, 'loss/train': 0.6578197628259659} 01/29/2022 15:02:37 - INFO - codeparrot_training - Step 20333: {'lr': 0.0003406145928637283, 'samples': 3904128, 'steps': 20333, 'loss/train': 1.1447053849697113} 01/29/2022 15:02:41 - INFO - codeparrot_training - Step 20334: {'lr': 0.0003405993428559414, 'samples': 3904320, 'steps': 20334, 'loss/train': 1.9597527980804443} 01/29/2022 15:02:46 - INFO - codeparrot_training - Step 20335: {'lr': 0.0003405840924600557, 'samples': 3904512, 'steps': 20335, 'loss/train': 1.8636755347251892} 01/29/2022 15:02:51 - INFO - codeparrot_training - Step 20336: {'lr': 0.0003405688416761364, 'samples': 3904704, 'steps': 20336, 'loss/train': 1.6132065653800964} 01/29/2022 15:02:55 - INFO - codeparrot_training - Step 20337: {'lr': 0.0003405535905042491, 'samples': 3904896, 'steps': 20337, 'loss/train': 1.3030094504356384} 01/29/2022 15:03:00 - INFO - codeparrot_training - Step 20338: {'lr': 0.0003405383389444589, 'samples': 3905088, 'steps': 20338, 'loss/train': 1.6324135065078735} 01/29/2022 15:03:04 - INFO - codeparrot_training - Step 20339: {'lr': 0.00034052308699683114, 'samples': 3905280, 'steps': 20339, 'loss/train': 1.9027905464172363} 01/29/2022 15:03:08 - INFO - codeparrot_training - Step 20340: {'lr': 0.0003405078346614313, 'samples': 3905472, 'steps': 20340, 'loss/train': 1.1827641427516937} 01/29/2022 15:03:15 - INFO - codeparrot_training - Step 20341: {'lr': 0.00034049258193832464, 'samples': 3905664, 'steps': 20341, 'loss/train': 0.76374551653862} 01/29/2022 15:03:19 - INFO - codeparrot_training - Step 20342: {'lr': 0.00034047732882757655, 'samples': 3905856, 'steps': 20342, 'loss/train': 1.938787043094635} 01/29/2022 15:03:24 - INFO - codeparrot_training - Step 20343: {'lr': 0.00034046207532925215, 'samples': 3906048, 'steps': 20343, 'loss/train': 2.2264530658721924} 01/29/2022 15:03:28 - INFO - codeparrot_training - Step 20344: {'lr': 0.0003404468214434171, 'samples': 3906240, 'steps': 20344, 'loss/train': 1.9421120882034302} 01/29/2022 15:03:32 - INFO - codeparrot_training - Step 20345: {'lr': 0.00034043156717013647, 'samples': 3906432, 'steps': 20345, 'loss/train': 1.552709698677063} 01/29/2022 15:03:37 - INFO - codeparrot_training - Step 20346: {'lr': 0.0003404163125094758, 'samples': 3906624, 'steps': 20346, 'loss/train': 0.7834568023681641} 01/29/2022 15:03:42 - INFO - codeparrot_training - Step 20347: {'lr': 0.0003404010574615003, 'samples': 3906816, 'steps': 20347, 'loss/train': 1.307450294494629} 01/29/2022 15:03:46 - INFO - codeparrot_training - Step 20348: {'lr': 0.00034038580202627543, 'samples': 3907008, 'steps': 20348, 'loss/train': 0.666166752576828} 01/29/2022 15:03:50 - INFO - codeparrot_training - Step 20349: {'lr': 0.0003403705462038665, 'samples': 3907200, 'steps': 20349, 'loss/train': 2.299770176410675} 01/29/2022 15:03:54 - INFO - codeparrot_training - Step 20350: {'lr': 0.0003403552899943388, 'samples': 3907392, 'steps': 20350, 'loss/train': 1.4068491160869598} 01/29/2022 15:04:00 - INFO - codeparrot_training - Step 20351: {'lr': 0.0003403400333977577, 'samples': 3907584, 'steps': 20351, 'loss/train': 0.6515597552061081} 01/29/2022 15:04:04 - INFO - codeparrot_training - Step 20352: {'lr': 0.00034032477641418856, 'samples': 3907776, 'steps': 20352, 'loss/train': 1.6208542585372925} 01/29/2022 15:04:08 - INFO - codeparrot_training - Step 20353: {'lr': 0.00034030951904369687, 'samples': 3907968, 'steps': 20353, 'loss/train': 2.238842189311981} 01/29/2022 15:04:13 - INFO - codeparrot_training - Step 20354: {'lr': 0.0003402942612863478, 'samples': 3908160, 'steps': 20354, 'loss/train': 0.6882151365280151} 01/29/2022 15:04:17 - INFO - codeparrot_training - Step 20355: {'lr': 0.00034027900314220684, 'samples': 3908352, 'steps': 20355, 'loss/train': 2.3142807483673096} 01/29/2022 15:04:24 - INFO - codeparrot_training - Step 20356: {'lr': 0.00034026374461133927, 'samples': 3908544, 'steps': 20356, 'loss/train': 1.7482938766479492} 01/29/2022 15:04:28 - INFO - codeparrot_training - Step 20357: {'lr': 0.0003402484856938105, 'samples': 3908736, 'steps': 20357, 'loss/train': 1.7020105719566345} 01/29/2022 15:04:33 - INFO - codeparrot_training - Step 20358: {'lr': 0.00034023322638968587, 'samples': 3908928, 'steps': 20358, 'loss/train': 1.7691720128059387} 01/29/2022 15:04:37 - INFO - codeparrot_training - Step 20359: {'lr': 0.0003402179666990307, 'samples': 3909120, 'steps': 20359, 'loss/train': 1.8766433000564575} 01/29/2022 15:04:41 - INFO - codeparrot_training - Step 20360: {'lr': 0.00034020270662191046, 'samples': 3909312, 'steps': 20360, 'loss/train': 1.0336976051330566} 01/29/2022 15:04:47 - INFO - codeparrot_training - Step 20361: {'lr': 0.0003401874461583905, 'samples': 3909504, 'steps': 20361, 'loss/train': 2.000019371509552} 01/29/2022 15:04:51 - INFO - codeparrot_training - Step 20362: {'lr': 0.00034017218530853606, 'samples': 3909696, 'steps': 20362, 'loss/train': 1.3867172598838806} 01/29/2022 15:04:55 - INFO - codeparrot_training - Step 20363: {'lr': 0.0003401569240724126, 'samples': 3909888, 'steps': 20363, 'loss/train': 0.9696155190467834} 01/29/2022 15:05:00 - INFO - codeparrot_training - Step 20364: {'lr': 0.0003401416624500856, 'samples': 3910080, 'steps': 20364, 'loss/train': 2.00413978099823} 01/29/2022 15:05:04 - INFO - codeparrot_training - Step 20365: {'lr': 0.0003401264004416203, 'samples': 3910272, 'steps': 20365, 'loss/train': 0.630110040307045} 01/29/2022 15:05:11 - INFO - codeparrot_training - Step 20366: {'lr': 0.00034011113804708216, 'samples': 3910464, 'steps': 20366, 'loss/train': 0.8796539604663849} 01/29/2022 15:05:15 - INFO - codeparrot_training - Step 20367: {'lr': 0.0003400958752665365, 'samples': 3910656, 'steps': 20367, 'loss/train': 1.8138951659202576} 01/29/2022 15:05:19 - INFO - codeparrot_training - Step 20368: {'lr': 0.0003400806121000487, 'samples': 3910848, 'steps': 20368, 'loss/train': 0.9961678683757782} 01/29/2022 15:05:23 - INFO - codeparrot_training - Step 20369: {'lr': 0.00034006534854768417, 'samples': 3911040, 'steps': 20369, 'loss/train': 1.4566734731197357} 01/29/2022 15:05:28 - INFO - codeparrot_training - Step 20370: {'lr': 0.00034005008460950825, 'samples': 3911232, 'steps': 20370, 'loss/train': 2.1120116114616394} 01/29/2022 15:05:33 - INFO - codeparrot_training - Step 20371: {'lr': 0.00034003482028558644, 'samples': 3911424, 'steps': 20371, 'loss/train': 1.8311918377876282} 01/29/2022 15:05:37 - INFO - codeparrot_training - Step 20372: {'lr': 0.000340019555575984, 'samples': 3911616, 'steps': 20372, 'loss/train': 0.9728458821773529} 01/29/2022 15:05:42 - INFO - codeparrot_training - Step 20373: {'lr': 0.00034000429048076637, 'samples': 3911808, 'steps': 20373, 'loss/train': 1.8213728070259094} 01/29/2022 15:05:46 - INFO - codeparrot_training - Step 20374: {'lr': 0.00033998902499999894, 'samples': 3912000, 'steps': 20374, 'loss/train': 1.6026166677474976} 01/29/2022 15:05:50 - INFO - codeparrot_training - Step 20375: {'lr': 0.00033997375913374705, 'samples': 3912192, 'steps': 20375, 'loss/train': 1.4176054000854492} 01/29/2022 15:05:55 - INFO - codeparrot_training - Step 20376: {'lr': 0.0003399584928820762, 'samples': 3912384, 'steps': 20376, 'loss/train': 1.2622166275978088} 01/29/2022 15:06:00 - INFO - codeparrot_training - Step 20377: {'lr': 0.00033994322624505177, 'samples': 3912576, 'steps': 20377, 'loss/train': 1.8392488360404968} 01/29/2022 15:06:04 - INFO - codeparrot_training - Step 20378: {'lr': 0.000339927959222739, 'samples': 3912768, 'steps': 20378, 'loss/train': 0.8237197995185852} 01/29/2022 15:06:08 - INFO - codeparrot_training - Step 20379: {'lr': 0.0003399126918152036, 'samples': 3912960, 'steps': 20379, 'loss/train': 1.0856419801712036} 01/29/2022 15:06:12 - INFO - codeparrot_training - Step 20380: {'lr': 0.00033989742402251055, 'samples': 3913152, 'steps': 20380, 'loss/train': 2.000274896621704} 01/29/2022 15:06:18 - INFO - codeparrot_training - Step 20381: {'lr': 0.00033988215584472564, 'samples': 3913344, 'steps': 20381, 'loss/train': 0.6187106966972351} 01/29/2022 15:06:22 - INFO - codeparrot_training - Step 20382: {'lr': 0.00033986688728191404, 'samples': 3913536, 'steps': 20382, 'loss/train': 1.7667658925056458} 01/29/2022 15:06:26 - INFO - codeparrot_training - Step 20383: {'lr': 0.0003398516183341413, 'samples': 3913728, 'steps': 20383, 'loss/train': 1.4779493808746338} 01/29/2022 15:06:30 - INFO - codeparrot_training - Step 20384: {'lr': 0.0003398363490014727, 'samples': 3913920, 'steps': 20384, 'loss/train': 1.5188205242156982} 01/29/2022 15:06:35 - INFO - codeparrot_training - Step 20385: {'lr': 0.0003398210792839738, 'samples': 3914112, 'steps': 20385, 'loss/train': 1.6562329530715942} 01/29/2022 15:06:42 - INFO - codeparrot_training - Step 20386: {'lr': 0.0003398058091817098, 'samples': 3914304, 'steps': 20386, 'loss/train': 1.1987663805484772} 01/29/2022 15:06:46 - INFO - codeparrot_training - Step 20387: {'lr': 0.0003397905386947462, 'samples': 3914496, 'steps': 20387, 'loss/train': 1.577255666255951} 01/29/2022 15:06:50 - INFO - codeparrot_training - Step 20388: {'lr': 0.00033977526782314854, 'samples': 3914688, 'steps': 20388, 'loss/train': 1.5485039949417114} 01/29/2022 15:06:54 - INFO - codeparrot_training - Step 20389: {'lr': 0.00033975999656698206, 'samples': 3914880, 'steps': 20389, 'loss/train': 1.2126111388206482} 01/29/2022 15:06:59 - INFO - codeparrot_training - Step 20390: {'lr': 0.00033974472492631234, 'samples': 3915072, 'steps': 20390, 'loss/train': 1.6842470169067383} 01/29/2022 15:07:04 - INFO - codeparrot_training - Step 20391: {'lr': 0.0003397294529012047, 'samples': 3915264, 'steps': 20391, 'loss/train': 1.5756627917289734} 01/29/2022 15:07:08 - INFO - codeparrot_training - Step 20392: {'lr': 0.0003397141804917246, 'samples': 3915456, 'steps': 20392, 'loss/train': 1.3454349339008331} 01/29/2022 15:07:13 - INFO - codeparrot_training - Step 20393: {'lr': 0.00033969890769793736, 'samples': 3915648, 'steps': 20393, 'loss/train': 1.4069823324680328} 01/29/2022 15:07:17 - INFO - codeparrot_training - Step 20394: {'lr': 0.0003396836345199086, 'samples': 3915840, 'steps': 20394, 'loss/train': 4.022717714309692} 01/29/2022 15:07:21 - INFO - codeparrot_training - Step 20395: {'lr': 0.0003396683609577035, 'samples': 3916032, 'steps': 20395, 'loss/train': 1.4989724457263947} 01/29/2022 15:07:26 - INFO - codeparrot_training - Step 20396: {'lr': 0.0003396530870113877, 'samples': 3916224, 'steps': 20396, 'loss/train': 1.4080390334129333} 01/29/2022 15:07:30 - INFO - codeparrot_training - Step 20397: {'lr': 0.0003396378126810264, 'samples': 3916416, 'steps': 20397, 'loss/train': 0.13992251828312874} 01/29/2022 15:07:35 - INFO - codeparrot_training - Step 20398: {'lr': 0.0003396225379666854, 'samples': 3916608, 'steps': 20398, 'loss/train': 1.5347918272018433} 01/29/2022 15:07:39 - INFO - codeparrot_training - Step 20399: {'lr': 0.00033960726286842973, 'samples': 3916800, 'steps': 20399, 'loss/train': 0.667443037033081} 01/29/2022 15:07:43 - INFO - codeparrot_training - Step 20400: {'lr': 0.00033959198738632503, 'samples': 3916992, 'steps': 20400, 'loss/train': 1.1779886484146118} 01/29/2022 15:07:51 - INFO - codeparrot_training - Step 20401: {'lr': 0.00033957671152043677, 'samples': 3917184, 'steps': 20401, 'loss/train': 1.3283991515636444} 01/29/2022 15:07:55 - INFO - codeparrot_training - Step 20402: {'lr': 0.0003395614352708303, 'samples': 3917376, 'steps': 20402, 'loss/train': 1.3082592487335205} 01/29/2022 15:07:59 - INFO - codeparrot_training - Step 20403: {'lr': 0.00033954615863757105, 'samples': 3917568, 'steps': 20403, 'loss/train': 1.2958797812461853} 01/29/2022 15:08:03 - INFO - codeparrot_training - Step 20404: {'lr': 0.0003395308816207245, 'samples': 3917760, 'steps': 20404, 'loss/train': 0.9502822458744049} 01/29/2022 15:08:08 - INFO - codeparrot_training - Step 20405: {'lr': 0.0003395156042203561, 'samples': 3917952, 'steps': 20405, 'loss/train': 2.0363945960998535} 01/29/2022 15:08:13 - INFO - codeparrot_training - Step 20406: {'lr': 0.0003395003264365313, 'samples': 3918144, 'steps': 20406, 'loss/train': 2.8302783966064453} 01/29/2022 15:08:18 - INFO - codeparrot_training - Step 20407: {'lr': 0.00033948504826931546, 'samples': 3918336, 'steps': 20407, 'loss/train': 1.610321044921875} 01/29/2022 15:08:22 - INFO - codeparrot_training - Step 20408: {'lr': 0.0003394697697187741, 'samples': 3918528, 'steps': 20408, 'loss/train': 1.6429458260536194} 01/29/2022 15:08:26 - INFO - codeparrot_training - Step 20409: {'lr': 0.0003394544907849727, 'samples': 3918720, 'steps': 20409, 'loss/train': 1.7962573170661926} 01/29/2022 15:08:33 - INFO - codeparrot_training - Step 20410: {'lr': 0.0003394392114679766, 'samples': 3918912, 'steps': 20410, 'loss/train': 1.799841284751892} 01/29/2022 15:08:37 - INFO - codeparrot_training - Step 20411: {'lr': 0.00033942393176785134, 'samples': 3919104, 'steps': 20411, 'loss/train': 2.3849746584892273} 01/29/2022 15:08:42 - INFO - codeparrot_training - Step 20412: {'lr': 0.00033940865168466237, 'samples': 3919296, 'steps': 20412, 'loss/train': 1.221556305885315} 01/29/2022 15:08:46 - INFO - codeparrot_training - Step 20413: {'lr': 0.00033939337121847513, 'samples': 3919488, 'steps': 20413, 'loss/train': 1.7912178039550781} 01/29/2022 15:08:50 - INFO - codeparrot_training - Step 20414: {'lr': 0.00033937809036935505, 'samples': 3919680, 'steps': 20414, 'loss/train': 1.1370640397071838} 01/29/2022 15:08:54 - INFO - codeparrot_training - Step 20415: {'lr': 0.0003393628091373677, 'samples': 3919872, 'steps': 20415, 'loss/train': 1.2955875098705292} 01/29/2022 15:09:00 - INFO - codeparrot_training - Step 20416: {'lr': 0.00033934752752257834, 'samples': 3920064, 'steps': 20416, 'loss/train': 1.5902286171913147} 01/29/2022 15:09:04 - INFO - codeparrot_training - Step 20417: {'lr': 0.00033933224552505257, 'samples': 3920256, 'steps': 20417, 'loss/train': 1.5762426853179932} 01/29/2022 15:09:08 - INFO - codeparrot_training - Step 20418: {'lr': 0.0003393169631448559, 'samples': 3920448, 'steps': 20418, 'loss/train': 1.7752953171730042} 01/29/2022 15:09:12 - INFO - codeparrot_training - Step 20419: {'lr': 0.0003393016803820536, 'samples': 3920640, 'steps': 20419, 'loss/train': 1.5281996726989746} 01/29/2022 15:09:17 - INFO - codeparrot_training - Step 20420: {'lr': 0.0003392863972367114, 'samples': 3920832, 'steps': 20420, 'loss/train': 1.5355321168899536} 01/29/2022 15:09:22 - INFO - codeparrot_training - Step 20421: {'lr': 0.0003392711137088945, 'samples': 3921024, 'steps': 20421, 'loss/train': 4.56645405292511} 01/29/2022 15:09:26 - INFO - codeparrot_training - Step 20422: {'lr': 0.00033925582979866853, 'samples': 3921216, 'steps': 20422, 'loss/train': 1.2765905857086182} 01/29/2022 15:09:30 - INFO - codeparrot_training - Step 20423: {'lr': 0.00033924054550609894, 'samples': 3921408, 'steps': 20423, 'loss/train': 1.3490334749221802} 01/29/2022 15:09:35 - INFO - codeparrot_training - Step 20424: {'lr': 0.00033922526083125115, 'samples': 3921600, 'steps': 20424, 'loss/train': 1.0341689586639404} 01/29/2022 15:09:39 - INFO - codeparrot_training - Step 20425: {'lr': 0.00033920997577419076, 'samples': 3921792, 'steps': 20425, 'loss/train': 2.021753489971161} 01/29/2022 15:09:46 - INFO - codeparrot_training - Step 20426: {'lr': 0.00033919469033498304, 'samples': 3921984, 'steps': 20426, 'loss/train': 2.006833076477051} 01/29/2022 15:09:51 - INFO - codeparrot_training - Step 20427: {'lr': 0.0003391794045136937, 'samples': 3922176, 'steps': 20427, 'loss/train': 0.9698481559753418} 01/29/2022 15:09:55 - INFO - codeparrot_training - Step 20428: {'lr': 0.00033916411831038805, 'samples': 3922368, 'steps': 20428, 'loss/train': 1.6321361660957336} 01/29/2022 15:09:59 - INFO - codeparrot_training - Step 20429: {'lr': 0.0003391488317251317, 'samples': 3922560, 'steps': 20429, 'loss/train': 1.3058980107307434} 01/29/2022 15:10:03 - INFO - codeparrot_training - Step 20430: {'lr': 0.00033913354475798995, 'samples': 3922752, 'steps': 20430, 'loss/train': 1.791288435459137} 01/29/2022 15:10:09 - INFO - codeparrot_training - Step 20431: {'lr': 0.0003391182574090285, 'samples': 3922944, 'steps': 20431, 'loss/train': 1.403247445821762} 01/29/2022 15:10:13 - INFO - codeparrot_training - Step 20432: {'lr': 0.00033910296967831267, 'samples': 3923136, 'steps': 20432, 'loss/train': 2.161141276359558} 01/29/2022 15:10:17 - INFO - codeparrot_training - Step 20433: {'lr': 0.00033908768156590806, 'samples': 3923328, 'steps': 20433, 'loss/train': 2.307257115840912} 01/29/2022 15:10:22 - INFO - codeparrot_training - Step 20434: {'lr': 0.0003390723930718801, 'samples': 3923520, 'steps': 20434, 'loss/train': 1.2903171479701996} 01/29/2022 15:10:27 - INFO - codeparrot_training - Step 20435: {'lr': 0.0003390571041962943, 'samples': 3923712, 'steps': 20435, 'loss/train': 1.0070922374725342} 01/29/2022 15:10:31 - INFO - codeparrot_training - Step 20436: {'lr': 0.0003390418149392161, 'samples': 3923904, 'steps': 20436, 'loss/train': 1.5938807129859924} 01/29/2022 15:10:36 - INFO - codeparrot_training - Step 20437: {'lr': 0.000339026525300711, 'samples': 3924096, 'steps': 20437, 'loss/train': 1.4415831863880157} 01/29/2022 15:10:40 - INFO - codeparrot_training - Step 20438: {'lr': 0.0003390112352808447, 'samples': 3924288, 'steps': 20438, 'loss/train': 1.105229526758194} 01/29/2022 15:10:44 - INFO - codeparrot_training - Step 20439: {'lr': 0.00033899594487968234, 'samples': 3924480, 'steps': 20439, 'loss/train': 1.6759251952171326} 01/29/2022 15:10:48 - INFO - codeparrot_training - Step 20440: {'lr': 0.00033898065409728974, 'samples': 3924672, 'steps': 20440, 'loss/train': 2.0443305373191833} 01/29/2022 15:10:54 - INFO - codeparrot_training - Step 20441: {'lr': 0.0003389653629337322, 'samples': 3924864, 'steps': 20441, 'loss/train': 1.358858048915863} 01/29/2022 15:10:58 - INFO - codeparrot_training - Step 20442: {'lr': 0.00033895007138907534, 'samples': 3925056, 'steps': 20442, 'loss/train': 1.2433043718338013} 01/29/2022 15:11:02 - INFO - codeparrot_training - Step 20443: {'lr': 0.00033893477946338456, 'samples': 3925248, 'steps': 20443, 'loss/train': 2.3116332292556763} 01/29/2022 15:11:07 - INFO - codeparrot_training - Step 20444: {'lr': 0.0003389194871567255, 'samples': 3925440, 'steps': 20444, 'loss/train': 1.6873818039894104} 01/29/2022 15:11:14 - INFO - codeparrot_training - Step 20445: {'lr': 0.00033890419446916354, 'samples': 3925632, 'steps': 20445, 'loss/train': 0.9072606861591339} 01/29/2022 15:11:18 - INFO - codeparrot_training - Step 20446: {'lr': 0.00033888890140076433, 'samples': 3925824, 'steps': 20446, 'loss/train': 0.9757212102413177} 01/29/2022 15:11:22 - INFO - codeparrot_training - Step 20447: {'lr': 0.00033887360795159315, 'samples': 3926016, 'steps': 20447, 'loss/train': 1.5347650051116943} 01/29/2022 15:11:27 - INFO - codeparrot_training - Step 20448: {'lr': 0.00033885831412171577, 'samples': 3926208, 'steps': 20448, 'loss/train': 1.7547889351844788} 01/29/2022 15:11:31 - INFO - codeparrot_training - Step 20449: {'lr': 0.00033884301991119755, 'samples': 3926400, 'steps': 20449, 'loss/train': 2.09001624584198} 01/29/2022 15:11:36 - INFO - codeparrot_training - Step 20450: {'lr': 0.00033882772532010404, 'samples': 3926592, 'steps': 20450, 'loss/train': 1.0333138704299927} 01/29/2022 15:11:40 - INFO - codeparrot_training - Step 20451: {'lr': 0.0003388124303485007, 'samples': 3926784, 'steps': 20451, 'loss/train': 1.9513052701950073} 01/29/2022 15:11:45 - INFO - codeparrot_training - Step 20452: {'lr': 0.0003387971349964532, 'samples': 3926976, 'steps': 20452, 'loss/train': 1.4730294942855835} 01/29/2022 15:11:49 - INFO - codeparrot_training - Step 20453: {'lr': 0.00033878183926402695, 'samples': 3927168, 'steps': 20453, 'loss/train': 1.3161222338676453} 01/29/2022 15:11:53 - INFO - codeparrot_training - Step 20454: {'lr': 0.00033876654315128746, 'samples': 3927360, 'steps': 20454, 'loss/train': 1.5554405450820923} 01/29/2022 15:12:00 - INFO - codeparrot_training - Step 20455: {'lr': 0.0003387512466583004, 'samples': 3927552, 'steps': 20455, 'loss/train': 1.8055971264839172} 01/29/2022 15:12:04 - INFO - codeparrot_training - Step 20456: {'lr': 0.0003387359497851311, 'samples': 3927744, 'steps': 20456, 'loss/train': 1.5592812895774841} 01/29/2022 15:12:09 - INFO - codeparrot_training - Step 20457: {'lr': 0.0003387206525318451, 'samples': 3927936, 'steps': 20457, 'loss/train': 1.9145334362983704} 01/29/2022 15:12:13 - INFO - codeparrot_training - Step 20458: {'lr': 0.00033870535489850806, 'samples': 3928128, 'steps': 20458, 'loss/train': 1.67143052816391} 01/29/2022 15:12:17 - INFO - codeparrot_training - Step 20459: {'lr': 0.00033869005688518547, 'samples': 3928320, 'steps': 20459, 'loss/train': 1.1910241842269897} 01/29/2022 15:12:23 - INFO - codeparrot_training - Step 20460: {'lr': 0.0003386747584919428, 'samples': 3928512, 'steps': 20460, 'loss/train': 0.8905946016311646} 01/29/2022 15:12:27 - INFO - codeparrot_training - Step 20461: {'lr': 0.0003386594597188456, 'samples': 3928704, 'steps': 20461, 'loss/train': 1.3734668791294098} 01/29/2022 15:12:31 - INFO - codeparrot_training - Step 20462: {'lr': 0.0003386441605659595, 'samples': 3928896, 'steps': 20462, 'loss/train': 1.7607250213623047} 01/29/2022 15:12:35 - INFO - codeparrot_training - Step 20463: {'lr': 0.00033862886103334987, 'samples': 3929088, 'steps': 20463, 'loss/train': 1.6092056035995483} 01/29/2022 15:12:40 - INFO - codeparrot_training - Step 20464: {'lr': 0.0003386135611210824, 'samples': 3929280, 'steps': 20464, 'loss/train': 1.6509758234024048} 01/29/2022 15:12:45 - INFO - codeparrot_training - Step 20465: {'lr': 0.0003385982608292226, 'samples': 3929472, 'steps': 20465, 'loss/train': 1.2273750007152557} 01/29/2022 15:12:49 - INFO - codeparrot_training - Step 20466: {'lr': 0.000338582960157836, 'samples': 3929664, 'steps': 20466, 'loss/train': 1.6206609606742859} 01/29/2022 15:12:53 - INFO - codeparrot_training - Step 20467: {'lr': 0.000338567659106988, 'samples': 3929856, 'steps': 20467, 'loss/train': 1.4280180037021637} 01/29/2022 15:12:58 - INFO - codeparrot_training - Step 20468: {'lr': 0.0003385523576767444, 'samples': 3930048, 'steps': 20468, 'loss/train': 1.3080410063266754} 01/29/2022 15:13:02 - INFO - codeparrot_training - Step 20469: {'lr': 0.00033853705586717056, 'samples': 3930240, 'steps': 20469, 'loss/train': 1.8385698795318604} 01/29/2022 15:13:09 - INFO - codeparrot_training - Step 20470: {'lr': 0.0003385217536783321, 'samples': 3930432, 'steps': 20470, 'loss/train': 1.5111638903617859} 01/29/2022 15:13:13 - INFO - codeparrot_training - Step 20471: {'lr': 0.00033850645111029455, 'samples': 3930624, 'steps': 20471, 'loss/train': 1.8867030143737793} 01/29/2022 15:13:18 - INFO - codeparrot_training - Step 20472: {'lr': 0.00033849114816312343, 'samples': 3930816, 'steps': 20472, 'loss/train': 1.4382957816123962} 01/29/2022 15:13:22 - INFO - codeparrot_training - Step 20473: {'lr': 0.0003384758448368844, 'samples': 3931008, 'steps': 20473, 'loss/train': 1.7911641597747803} 01/29/2022 15:13:26 - INFO - codeparrot_training - Step 20474: {'lr': 0.00033846054113164287, 'samples': 3931200, 'steps': 20474, 'loss/train': 1.2322266697883606} 01/29/2022 15:13:31 - INFO - codeparrot_training - Step 20475: {'lr': 0.00033844523704746455, 'samples': 3931392, 'steps': 20475, 'loss/train': 0.7555853426456451} 01/29/2022 15:13:36 - INFO - codeparrot_training - Step 20476: {'lr': 0.00033842993258441487, 'samples': 3931584, 'steps': 20476, 'loss/train': 0.4983438402414322} 01/29/2022 15:13:40 - INFO - codeparrot_training - Step 20477: {'lr': 0.0003384146277425595, 'samples': 3931776, 'steps': 20477, 'loss/train': 1.912131428718567} 01/29/2022 15:13:44 - INFO - codeparrot_training - Step 20478: {'lr': 0.0003383993225219638, 'samples': 3931968, 'steps': 20478, 'loss/train': 1.4238241016864777} 01/29/2022 15:13:48 - INFO - codeparrot_training - Step 20479: {'lr': 0.0003383840169226936, 'samples': 3932160, 'steps': 20479, 'loss/train': 0.46603797376155853} 01/29/2022 15:13:54 - INFO - codeparrot_training - Step 20480: {'lr': 0.00033836871094481433, 'samples': 3932352, 'steps': 20480, 'loss/train': 1.9818634986877441} 01/29/2022 15:13:58 - INFO - codeparrot_training - Step 20481: {'lr': 0.00033835340458839155, 'samples': 3932544, 'steps': 20481, 'loss/train': 1.2626044750213623} 01/29/2022 15:14:02 - INFO - codeparrot_training - Step 20482: {'lr': 0.0003383380978534908, 'samples': 3932736, 'steps': 20482, 'loss/train': 1.6386621594429016} 01/29/2022 15:14:06 - INFO - codeparrot_training - Step 20483: {'lr': 0.0003383227907401777, 'samples': 3932928, 'steps': 20483, 'loss/train': 1.4892034828662872} 01/29/2022 15:14:11 - INFO - codeparrot_training - Step 20484: {'lr': 0.0003383074832485178, 'samples': 3933120, 'steps': 20484, 'loss/train': 1.997819423675537} 01/29/2022 15:14:18 - INFO - codeparrot_training - Step 20485: {'lr': 0.0003382921753785767, 'samples': 3933312, 'steps': 20485, 'loss/train': 1.1110649406909943} 01/29/2022 15:14:22 - INFO - codeparrot_training - Step 20486: {'lr': 0.0003382768671304199, 'samples': 3933504, 'steps': 20486, 'loss/train': 0.929896742105484} 01/29/2022 15:14:26 - INFO - codeparrot_training - Step 20487: {'lr': 0.0003382615585041131, 'samples': 3933696, 'steps': 20487, 'loss/train': 1.6542654633522034} 01/29/2022 15:14:31 - INFO - codeparrot_training - Step 20488: {'lr': 0.0003382462494997219, 'samples': 3933888, 'steps': 20488, 'loss/train': 1.3711117208003998} 01/29/2022 15:14:35 - INFO - codeparrot_training - Step 20489: {'lr': 0.00033823094011731165, 'samples': 3934080, 'steps': 20489, 'loss/train': 1.6129882335662842} 01/29/2022 15:14:40 - INFO - codeparrot_training - Step 20490: {'lr': 0.0003382156303569481, 'samples': 3934272, 'steps': 20490, 'loss/train': 1.3370000123977661} 01/29/2022 15:14:44 - INFO - codeparrot_training - Step 20491: {'lr': 0.0003382003202186968, 'samples': 3934464, 'steps': 20491, 'loss/train': 1.6871253848075867} 01/29/2022 15:14:49 - INFO - codeparrot_training - Step 20492: {'lr': 0.0003381850097026234, 'samples': 3934656, 'steps': 20492, 'loss/train': 1.6736973524093628} 01/29/2022 15:14:53 - INFO - codeparrot_training - Step 20493: {'lr': 0.0003381696988087933, 'samples': 3934848, 'steps': 20493, 'loss/train': 1.659108817577362} 01/29/2022 15:14:57 - INFO - codeparrot_training - Step 20494: {'lr': 0.00033815438753727236, 'samples': 3935040, 'steps': 20494, 'loss/train': 1.8286703824996948} 01/29/2022 15:15:03 - INFO - codeparrot_training - Step 20495: {'lr': 0.00033813907588812586, 'samples': 3935232, 'steps': 20495, 'loss/train': 1.4521292746067047} 01/29/2022 15:15:07 - INFO - codeparrot_training - Step 20496: {'lr': 0.0003381237638614196, 'samples': 3935424, 'steps': 20496, 'loss/train': 1.2925730645656586} 01/29/2022 15:15:11 - INFO - codeparrot_training - Step 20497: {'lr': 0.00033810845145721915, 'samples': 3935616, 'steps': 20497, 'loss/train': 1.555936574935913} 01/29/2022 15:15:15 - INFO - codeparrot_training - Step 20498: {'lr': 0.0003380931386755901, 'samples': 3935808, 'steps': 20498, 'loss/train': 1.6733425855636597} 01/29/2022 15:15:20 - INFO - codeparrot_training - Step 20499: {'lr': 0.00033807782551659795, 'samples': 3936000, 'steps': 20499, 'loss/train': 1.7780914306640625} 01/29/2022 15:15:27 - INFO - codeparrot_training - Step 20500: {'lr': 0.0003380625119803084, 'samples': 3936192, 'steps': 20500, 'loss/train': 1.8635931015014648} 01/29/2022 15:15:31 - INFO - codeparrot_training - Step 20501: {'lr': 0.00033804719806678694, 'samples': 3936384, 'steps': 20501, 'loss/train': 1.1444000601768494} 01/29/2022 15:15:35 - INFO - codeparrot_training - Step 20502: {'lr': 0.00033803188377609937, 'samples': 3936576, 'steps': 20502, 'loss/train': 1.6903476119041443} 01/29/2022 15:15:39 - INFO - codeparrot_training - Step 20503: {'lr': 0.00033801656910831116, 'samples': 3936768, 'steps': 20503, 'loss/train': 1.490405559539795} 01/29/2022 15:15:44 - INFO - codeparrot_training - Step 20504: {'lr': 0.0003380012540634878, 'samples': 3936960, 'steps': 20504, 'loss/train': 1.7666537761688232} 01/29/2022 15:15:49 - INFO - codeparrot_training - Step 20505: {'lr': 0.0003379859386416951, 'samples': 3937152, 'steps': 20505, 'loss/train': 1.267581045627594} 01/29/2022 15:15:53 - INFO - codeparrot_training - Step 20506: {'lr': 0.00033797062284299847, 'samples': 3937344, 'steps': 20506, 'loss/train': 2.053695559501648} 01/29/2022 15:15:57 - INFO - codeparrot_training - Step 20507: {'lr': 0.0003379553066674637, 'samples': 3937536, 'steps': 20507, 'loss/train': 1.7428926229476929} 01/29/2022 15:16:02 - INFO - codeparrot_training - Step 20508: {'lr': 0.0003379399901151563, 'samples': 3937728, 'steps': 20508, 'loss/train': 0.9708525538444519} 01/29/2022 15:16:06 - INFO - codeparrot_training - Step 20509: {'lr': 0.00033792467318614195, 'samples': 3937920, 'steps': 20509, 'loss/train': 0.8379393517971039} 01/29/2022 15:16:11 - INFO - codeparrot_training - Step 20510: {'lr': 0.0003379093558804862, 'samples': 3938112, 'steps': 20510, 'loss/train': 2.276368260383606} 01/29/2022 15:16:15 - INFO - codeparrot_training - Step 20511: {'lr': 0.00033789403819825466, 'samples': 3938304, 'steps': 20511, 'loss/train': 1.7483618259429932} 01/29/2022 15:16:20 - INFO - codeparrot_training - Step 20512: {'lr': 0.00033787872013951297, 'samples': 3938496, 'steps': 20512, 'loss/train': 1.2899705171585083} 01/29/2022 15:16:24 - INFO - codeparrot_training - Step 20513: {'lr': 0.0003378634017043268, 'samples': 3938688, 'steps': 20513, 'loss/train': 1.083937257528305} 01/29/2022 15:16:28 - INFO - codeparrot_training - Step 20514: {'lr': 0.0003378480828927617, 'samples': 3938880, 'steps': 20514, 'loss/train': 1.1171649098396301} 01/29/2022 15:16:35 - INFO - codeparrot_training - Step 20515: {'lr': 0.00033783276370488323, 'samples': 3939072, 'steps': 20515, 'loss/train': 1.416550487279892} 01/29/2022 15:16:39 - INFO - codeparrot_training - Step 20516: {'lr': 0.00033781744414075723, 'samples': 3939264, 'steps': 20516, 'loss/train': 1.8396663665771484} 01/29/2022 15:16:44 - INFO - codeparrot_training - Step 20517: {'lr': 0.00033780212420044903, 'samples': 3939456, 'steps': 20517, 'loss/train': 1.3662015795707703} 01/29/2022 15:16:48 - INFO - codeparrot_training - Step 20518: {'lr': 0.0003377868038840245, 'samples': 3939648, 'steps': 20518, 'loss/train': 1.9617095589637756} 01/29/2022 15:16:52 - INFO - codeparrot_training - Step 20519: {'lr': 0.00033777148319154917, 'samples': 3939840, 'steps': 20519, 'loss/train': 1.4035409688949585} 01/29/2022 15:16:58 - INFO - codeparrot_training - Step 20520: {'lr': 0.0003377561621230887, 'samples': 3940032, 'steps': 20520, 'loss/train': 2.214169442653656} 01/29/2022 15:17:02 - INFO - codeparrot_training - Step 20521: {'lr': 0.0003377408406787086, 'samples': 3940224, 'steps': 20521, 'loss/train': 0.7588511109352112} 01/29/2022 15:17:06 - INFO - codeparrot_training - Step 20522: {'lr': 0.00033772551885847467, 'samples': 3940416, 'steps': 20522, 'loss/train': 0.7075484097003937} 01/29/2022 15:17:10 - INFO - codeparrot_training - Step 20523: {'lr': 0.0003377101966624525, 'samples': 3940608, 'steps': 20523, 'loss/train': 0.6067754477262497} 01/29/2022 15:17:16 - INFO - codeparrot_training - Step 20524: {'lr': 0.0003376948740907077, 'samples': 3940800, 'steps': 20524, 'loss/train': 1.4171505868434906} 01/29/2022 15:17:20 - INFO - codeparrot_training - Step 20525: {'lr': 0.0003376795511433058, 'samples': 3940992, 'steps': 20525, 'loss/train': 1.376609355211258} 01/29/2022 15:17:24 - INFO - codeparrot_training - Step 20526: {'lr': 0.00033766422782031263, 'samples': 3941184, 'steps': 20526, 'loss/train': 7.240141868591309} 01/29/2022 15:17:29 - INFO - codeparrot_training - Step 20527: {'lr': 0.0003376489041217937, 'samples': 3941376, 'steps': 20527, 'loss/train': 0.9608463048934937} 01/29/2022 15:17:33 - INFO - codeparrot_training - Step 20528: {'lr': 0.00033763358004781474, 'samples': 3941568, 'steps': 20528, 'loss/train': 1.626526415348053} 01/29/2022 15:17:37 - INFO - codeparrot_training - Step 20529: {'lr': 0.00033761825559844137, 'samples': 3941760, 'steps': 20529, 'loss/train': 1.5798692107200623} 01/29/2022 15:17:44 - INFO - codeparrot_training - Step 20530: {'lr': 0.00033760293077373917, 'samples': 3941952, 'steps': 20530, 'loss/train': 0.6749232709407806} 01/29/2022 15:17:49 - INFO - codeparrot_training - Step 20531: {'lr': 0.0003375876055737739, 'samples': 3942144, 'steps': 20531, 'loss/train': 1.3485791981220245} 01/29/2022 15:17:53 - INFO - codeparrot_training - Step 20532: {'lr': 0.000337572279998611, 'samples': 3942336, 'steps': 20532, 'loss/train': 1.8513550758361816} 01/29/2022 15:17:57 - INFO - codeparrot_training - Step 20533: {'lr': 0.00033755695404831635, 'samples': 3942528, 'steps': 20533, 'loss/train': 1.1616955697536469} 01/29/2022 15:18:01 - INFO - codeparrot_training - Step 20534: {'lr': 0.00033754162772295555, 'samples': 3942720, 'steps': 20534, 'loss/train': 1.3635700643062592} 01/29/2022 15:18:06 - INFO - codeparrot_training - Step 20535: {'lr': 0.0003375263010225941, 'samples': 3942912, 'steps': 20535, 'loss/train': 2.0634909868240356} 01/29/2022 15:18:11 - INFO - codeparrot_training - Step 20536: {'lr': 0.00033751097394729793, 'samples': 3943104, 'steps': 20536, 'loss/train': 1.1993198096752167} 01/29/2022 15:18:15 - INFO - codeparrot_training - Step 20537: {'lr': 0.0003374956464971324, 'samples': 3943296, 'steps': 20537, 'loss/train': 1.8540778756141663} 01/29/2022 15:18:19 - INFO - codeparrot_training - Step 20538: {'lr': 0.00033748031867216334, 'samples': 3943488, 'steps': 20538, 'loss/train': 1.635798454284668} 01/29/2022 15:18:23 - INFO - codeparrot_training - Step 20539: {'lr': 0.0003374649904724564, 'samples': 3943680, 'steps': 20539, 'loss/train': 1.9518998265266418} 01/29/2022 15:18:31 - INFO - codeparrot_training - Step 20540: {'lr': 0.0003374496618980772, 'samples': 3943872, 'steps': 20540, 'loss/train': 1.3010281920433044} 01/29/2022 15:18:36 - INFO - codeparrot_training - Step 20541: {'lr': 0.0003374343329490914, 'samples': 3944064, 'steps': 20541, 'loss/train': 1.6688826084136963} 01/29/2022 15:18:40 - INFO - codeparrot_training - Step 20542: {'lr': 0.00033741900362556474, 'samples': 3944256, 'steps': 20542, 'loss/train': 0.65839684009552} 01/29/2022 15:18:44 - INFO - codeparrot_training - Step 20543: {'lr': 0.00033740367392756274, 'samples': 3944448, 'steps': 20543, 'loss/train': 2.0947458744049072} 01/29/2022 15:18:49 - INFO - codeparrot_training - Step 20544: {'lr': 0.0003373883438551512, 'samples': 3944640, 'steps': 20544, 'loss/train': 1.5796993374824524} 01/29/2022 15:18:53 - INFO - codeparrot_training - Step 20545: {'lr': 0.0003373730134083958, 'samples': 3944832, 'steps': 20545, 'loss/train': 0.6397930830717087} 01/29/2022 15:18:58 - INFO - codeparrot_training - Step 20546: {'lr': 0.00033735768258736206, 'samples': 3945024, 'steps': 20546, 'loss/train': 1.7627427577972412} 01/29/2022 15:19:02 - INFO - codeparrot_training - Step 20547: {'lr': 0.0003373423513921158, 'samples': 3945216, 'steps': 20547, 'loss/train': 2.5777010321617126} 01/29/2022 15:19:07 - INFO - codeparrot_training - Step 20548: {'lr': 0.00033732701982272257, 'samples': 3945408, 'steps': 20548, 'loss/train': 2.264759659767151} 01/29/2022 15:19:11 - INFO - codeparrot_training - Step 20549: {'lr': 0.0003373116878792482, 'samples': 3945600, 'steps': 20549, 'loss/train': 2.3162428736686707} 01/29/2022 15:19:15 - INFO - codeparrot_training - Step 20550: {'lr': 0.00033729635556175826, 'samples': 3945792, 'steps': 20550, 'loss/train': 0.7157769799232483} 01/29/2022 15:19:20 - INFO - codeparrot_training - Step 20551: {'lr': 0.0003372810228703184, 'samples': 3945984, 'steps': 20551, 'loss/train': 1.6435250043869019} 01/29/2022 15:19:25 - INFO - codeparrot_training - Step 20552: {'lr': 0.0003372656898049944, 'samples': 3946176, 'steps': 20552, 'loss/train': 1.3252581059932709} 01/29/2022 15:19:29 - INFO - codeparrot_training - Step 20553: {'lr': 0.00033725035636585196, 'samples': 3946368, 'steps': 20553, 'loss/train': 1.72898930311203} 01/29/2022 15:19:33 - INFO - codeparrot_training - Step 20554: {'lr': 0.0003372350225529566, 'samples': 3946560, 'steps': 20554, 'loss/train': 1.7589948177337646} 01/29/2022 15:19:37 - INFO - codeparrot_training - Step 20555: {'lr': 0.00033721968836637414, 'samples': 3946752, 'steps': 20555, 'loss/train': 1.2462790310382843} 01/29/2022 15:19:45 - INFO - codeparrot_training - Step 20556: {'lr': 0.0003372043538061702, 'samples': 3946944, 'steps': 20556, 'loss/train': 1.1181211173534393} 01/29/2022 15:19:49 - INFO - codeparrot_training - Step 20557: {'lr': 0.00033718901887241047, 'samples': 3947136, 'steps': 20557, 'loss/train': 1.8199523091316223} 01/29/2022 15:19:54 - INFO - codeparrot_training - Step 20558: {'lr': 0.00033717368356516077, 'samples': 3947328, 'steps': 20558, 'loss/train': 1.0211577415466309} 01/29/2022 15:19:58 - INFO - codeparrot_training - Step 20559: {'lr': 0.0003371583478844866, 'samples': 3947520, 'steps': 20559, 'loss/train': 1.9077025651931763} 01/29/2022 15:20:02 - INFO - codeparrot_training - Step 20560: {'lr': 0.0003371430118304538, 'samples': 3947712, 'steps': 20560, 'loss/train': 1.71349436044693} 01/29/2022 15:20:07 - INFO - codeparrot_training - Step 20561: {'lr': 0.000337127675403128, 'samples': 3947904, 'steps': 20561, 'loss/train': 1.6976362466812134} 01/29/2022 15:20:12 - INFO - codeparrot_training - Step 20562: {'lr': 0.00033711233860257494, 'samples': 3948096, 'steps': 20562, 'loss/train': 1.1829875707626343} 01/29/2022 15:20:16 - INFO - codeparrot_training - Step 20563: {'lr': 0.00033709700142886025, 'samples': 3948288, 'steps': 20563, 'loss/train': 1.9711382389068604} 01/29/2022 15:20:20 - INFO - codeparrot_training - Step 20564: {'lr': 0.00033708166388204963, 'samples': 3948480, 'steps': 20564, 'loss/train': 1.6200742721557617} 01/29/2022 15:20:24 - INFO - codeparrot_training - Step 20565: {'lr': 0.00033706632596220885, 'samples': 3948672, 'steps': 20565, 'loss/train': 1.1276845335960388} 01/29/2022 15:20:32 - INFO - codeparrot_training - Step 20566: {'lr': 0.00033705098766940354, 'samples': 3948864, 'steps': 20566, 'loss/train': 2.07281631231308} 01/29/2022 15:20:36 - INFO - codeparrot_training - Step 20567: {'lr': 0.00033703564900369943, 'samples': 3949056, 'steps': 20567, 'loss/train': 1.064938724040985} 01/29/2022 15:20:40 - INFO - codeparrot_training - Step 20568: {'lr': 0.0003370203099651623, 'samples': 3949248, 'steps': 20568, 'loss/train': 1.5933061838150024} 01/29/2022 15:20:44 - INFO - codeparrot_training - Step 20569: {'lr': 0.0003370049705538578, 'samples': 3949440, 'steps': 20569, 'loss/train': 1.248652070760727} 01/29/2022 15:20:49 - INFO - codeparrot_training - Step 20570: {'lr': 0.00033698963076985155, 'samples': 3949632, 'steps': 20570, 'loss/train': 1.7562979459762573} 01/29/2022 15:20:54 - INFO - codeparrot_training - Step 20571: {'lr': 0.00033697429061320936, 'samples': 3949824, 'steps': 20571, 'loss/train': 1.6223652362823486} 01/29/2022 15:20:58 - INFO - codeparrot_training - Step 20572: {'lr': 0.00033695895008399704, 'samples': 3950016, 'steps': 20572, 'loss/train': 1.942706286907196} 01/29/2022 15:21:02 - INFO - codeparrot_training - Step 20573: {'lr': 0.00033694360918228006, 'samples': 3950208, 'steps': 20573, 'loss/train': 1.9354462623596191} 01/29/2022 15:21:07 - INFO - codeparrot_training - Step 20574: {'lr': 0.0003369282679081243, 'samples': 3950400, 'steps': 20574, 'loss/train': 1.473976582288742} 01/29/2022 15:21:11 - INFO - codeparrot_training - Step 20575: {'lr': 0.00033691292626159556, 'samples': 3950592, 'steps': 20575, 'loss/train': 1.212308406829834} 01/29/2022 15:21:16 - INFO - codeparrot_training - Step 20576: {'lr': 0.0003368975842427592, 'samples': 3950784, 'steps': 20576, 'loss/train': 1.323205053806305} 01/29/2022 15:21:20 - INFO - codeparrot_training - Step 20577: {'lr': 0.00033688224185168136, 'samples': 3950976, 'steps': 20577, 'loss/train': 1.2278985679149628} 01/29/2022 15:21:25 - INFO - codeparrot_training - Step 20578: {'lr': 0.0003368668990884275, 'samples': 3951168, 'steps': 20578, 'loss/train': 0.6229681670665741} 01/29/2022 15:21:29 - INFO - codeparrot_training - Step 20579: {'lr': 0.0003368515559530635, 'samples': 3951360, 'steps': 20579, 'loss/train': 0.5936249792575836} 01/29/2022 15:21:36 - INFO - codeparrot_training - Step 20580: {'lr': 0.00033683621244565506, 'samples': 3951552, 'steps': 20580, 'loss/train': 2.130909562110901} 01/29/2022 15:21:40 - INFO - codeparrot_training - Step 20581: {'lr': 0.00033682086856626773, 'samples': 3951744, 'steps': 20581, 'loss/train': 1.2404204607009888} 01/29/2022 15:21:45 - INFO - codeparrot_training - Step 20582: {'lr': 0.00033680552431496744, 'samples': 3951936, 'steps': 20582, 'loss/train': 1.6533622741699219} 01/29/2022 15:21:49 - INFO - codeparrot_training - Step 20583: {'lr': 0.0003367901796918198, 'samples': 3952128, 'steps': 20583, 'loss/train': 2.011413037776947} 01/29/2022 15:21:53 - INFO - codeparrot_training - Step 20584: {'lr': 0.0003367748346968907, 'samples': 3952320, 'steps': 20584, 'loss/train': 1.7454607486724854} 01/29/2022 15:21:58 - INFO - codeparrot_training - Step 20585: {'lr': 0.0003367594893302457, 'samples': 3952512, 'steps': 20585, 'loss/train': 1.9856334328651428} 01/29/2022 15:22:03 - INFO - codeparrot_training - Step 20586: {'lr': 0.00033674414359195055, 'samples': 3952704, 'steps': 20586, 'loss/train': 1.4828296601772308} 01/29/2022 15:22:07 - INFO - codeparrot_training - Step 20587: {'lr': 0.00033672879748207105, 'samples': 3952896, 'steps': 20587, 'loss/train': 1.559693992137909} 01/29/2022 15:22:11 - INFO - codeparrot_training - Step 20588: {'lr': 0.000336713451000673, 'samples': 3953088, 'steps': 20588, 'loss/train': 1.3191765546798706} 01/29/2022 15:22:15 - INFO - codeparrot_training - Step 20589: {'lr': 0.000336698104147822, 'samples': 3953280, 'steps': 20589, 'loss/train': 1.5470529198646545} 01/29/2022 15:22:23 - INFO - codeparrot_training - Step 20590: {'lr': 0.0003366827569235838, 'samples': 3953472, 'steps': 20590, 'loss/train': 0.050518931820988655} 01/29/2022 15:22:27 - INFO - codeparrot_training - Step 20591: {'lr': 0.00033666740932802424, 'samples': 3953664, 'steps': 20591, 'loss/train': 1.9328924417495728} 01/29/2022 15:22:31 - INFO - codeparrot_training - Step 20592: {'lr': 0.0003366520613612091, 'samples': 3953856, 'steps': 20592, 'loss/train': 2.440435767173767} 01/29/2022 15:22:35 - INFO - codeparrot_training - Step 20593: {'lr': 0.00033663671302320387, 'samples': 3954048, 'steps': 20593, 'loss/train': 1.336210459470749} 01/29/2022 15:22:40 - INFO - codeparrot_training - Step 20594: {'lr': 0.0003366213643140745, 'samples': 3954240, 'steps': 20594, 'loss/train': 1.2348790168762207} 01/29/2022 15:22:45 - INFO - codeparrot_training - Step 20595: {'lr': 0.0003366060152338869, 'samples': 3954432, 'steps': 20595, 'loss/train': 2.2825092673301697} 01/29/2022 15:22:49 - INFO - codeparrot_training - Step 20596: {'lr': 0.0003365906657827065, 'samples': 3954624, 'steps': 20596, 'loss/train': 0.9645757377147675} 01/29/2022 15:22:53 - INFO - codeparrot_training - Step 20597: {'lr': 0.00033657531596059914, 'samples': 3954816, 'steps': 20597, 'loss/train': 0.9242624938488007} 01/29/2022 15:22:58 - INFO - codeparrot_training - Step 20598: {'lr': 0.00033655996576763064, 'samples': 3955008, 'steps': 20598, 'loss/train': 1.55363130569458} 01/29/2022 15:23:02 - INFO - codeparrot_training - Step 20599: {'lr': 0.00033654461520386683, 'samples': 3955200, 'steps': 20599, 'loss/train': 2.233937680721283} 01/29/2022 15:23:07 - INFO - codeparrot_training - Step 20600: {'lr': 0.0003365292642693733, 'samples': 3955392, 'steps': 20600, 'loss/train': 1.234091877937317} 01/29/2022 15:23:12 - INFO - codeparrot_training - Step 20601: {'lr': 0.00033651391296421587, 'samples': 3955584, 'steps': 20601, 'loss/train': 2.082865297794342} 01/29/2022 15:23:16 - INFO - codeparrot_training - Step 20602: {'lr': 0.0003364985612884603, 'samples': 3955776, 'steps': 20602, 'loss/train': 0.15940463170409203} 01/29/2022 15:23:20 - INFO - codeparrot_training - Step 20603: {'lr': 0.0003364832092421724, 'samples': 3955968, 'steps': 20603, 'loss/train': 1.5200965404510498} 01/29/2022 15:23:24 - INFO - codeparrot_training - Step 20604: {'lr': 0.0003364678568254178, 'samples': 3956160, 'steps': 20604, 'loss/train': 1.4951978623867035} 01/29/2022 15:23:30 - INFO - codeparrot_training - Step 20605: {'lr': 0.00033645250403826246, 'samples': 3956352, 'steps': 20605, 'loss/train': 2.0227349996566772} 01/29/2022 15:23:34 - INFO - codeparrot_training - Step 20606: {'lr': 0.00033643715088077205, 'samples': 3956544, 'steps': 20606, 'loss/train': 2.0213990807533264} 01/29/2022 15:23:38 - INFO - codeparrot_training - Step 20607: {'lr': 0.0003364217973530123, 'samples': 3956736, 'steps': 20607, 'loss/train': 2.160372734069824} 01/29/2022 15:23:42 - INFO - codeparrot_training - Step 20608: {'lr': 0.00033640644345504897, 'samples': 3956928, 'steps': 20608, 'loss/train': 1.838454008102417} 01/29/2022 15:23:47 - INFO - codeparrot_training - Step 20609: {'lr': 0.0003363910891869479, 'samples': 3957120, 'steps': 20609, 'loss/train': 1.6748880743980408} 01/29/2022 15:23:54 - INFO - codeparrot_training - Step 20610: {'lr': 0.00033637573454877486, 'samples': 3957312, 'steps': 20610, 'loss/train': 0.48586830496788025} 01/29/2022 15:23:58 - INFO - codeparrot_training - Step 20611: {'lr': 0.0003363603795405956, 'samples': 3957504, 'steps': 20611, 'loss/train': 0.9645303189754486} 01/29/2022 15:24:02 - INFO - codeparrot_training - Step 20612: {'lr': 0.0003363450241624759, 'samples': 3957696, 'steps': 20612, 'loss/train': 0.6221351623535156} 01/29/2022 15:24:07 - INFO - codeparrot_training - Step 20613: {'lr': 0.00033632966841448144, 'samples': 3957888, 'steps': 20613, 'loss/train': 0.8747757375240326} 01/29/2022 15:24:11 - INFO - codeparrot_training - Step 20614: {'lr': 0.0003363143122966782, 'samples': 3958080, 'steps': 20614, 'loss/train': 1.3894371092319489} 01/29/2022 15:24:16 - INFO - codeparrot_training - Step 20615: {'lr': 0.00033629895580913183, 'samples': 3958272, 'steps': 20615, 'loss/train': 1.5148723125457764} 01/29/2022 15:24:20 - INFO - codeparrot_training - Step 20616: {'lr': 0.00033628359895190814, 'samples': 3958464, 'steps': 20616, 'loss/train': 1.5219172239303589} 01/29/2022 15:24:25 - INFO - codeparrot_training - Step 20617: {'lr': 0.0003362682417250729, 'samples': 3958656, 'steps': 20617, 'loss/train': 1.701085388660431} 01/29/2022 15:24:29 - INFO - codeparrot_training - Step 20618: {'lr': 0.00033625288412869185, 'samples': 3958848, 'steps': 20618, 'loss/train': 0.524105116724968} 01/29/2022 15:24:33 - INFO - codeparrot_training - Step 20619: {'lr': 0.00033623752616283094, 'samples': 3959040, 'steps': 20619, 'loss/train': 1.3596049547195435} 01/29/2022 15:24:38 - INFO - codeparrot_training - Step 20620: {'lr': 0.0003362221678275558, 'samples': 3959232, 'steps': 20620, 'loss/train': 2.2634661197662354} 01/29/2022 15:24:42 - INFO - codeparrot_training - Step 20621: {'lr': 0.00033620680912293226, 'samples': 3959424, 'steps': 20621, 'loss/train': 1.6007566452026367} 01/29/2022 15:24:47 - INFO - codeparrot_training - Step 20622: {'lr': 0.0003361914500490261, 'samples': 3959616, 'steps': 20622, 'loss/train': 1.320447564125061} 01/29/2022 15:24:51 - INFO - codeparrot_training - Step 20623: {'lr': 0.0003361760906059032, 'samples': 3959808, 'steps': 20623, 'loss/train': 1.6897813081741333} 01/29/2022 15:24:58 - INFO - codeparrot_training - Step 20624: {'lr': 0.00033616073079362923, 'samples': 3960000, 'steps': 20624, 'loss/train': 1.9293978810310364} 01/29/2022 15:25:03 - INFO - codeparrot_training - Step 20625: {'lr': 0.00033614537061227014, 'samples': 3960192, 'steps': 20625, 'loss/train': 1.5580884218215942} 01/29/2022 15:25:07 - INFO - codeparrot_training - Step 20626: {'lr': 0.00033613001006189156, 'samples': 3960384, 'steps': 20626, 'loss/train': 1.7297199368476868} 01/29/2022 15:25:11 - INFO - codeparrot_training - Step 20627: {'lr': 0.0003361146491425594, 'samples': 3960576, 'steps': 20627, 'loss/train': 0.677249863743782} 01/29/2022 15:25:15 - INFO - codeparrot_training - Step 20628: {'lr': 0.0003360992878543394, 'samples': 3960768, 'steps': 20628, 'loss/train': 0.9314627051353455} 01/29/2022 15:25:21 - INFO - codeparrot_training - Step 20629: {'lr': 0.00033608392619729745, 'samples': 3960960, 'steps': 20629, 'loss/train': 2.1138349771499634} 01/29/2022 15:25:25 - INFO - codeparrot_training - Step 20630: {'lr': 0.0003360685641714992, 'samples': 3961152, 'steps': 20630, 'loss/train': 1.8548203110694885} 01/29/2022 15:25:29 - INFO - codeparrot_training - Step 20631: {'lr': 0.0003360532017770106, 'samples': 3961344, 'steps': 20631, 'loss/train': 1.25352081656456} 01/29/2022 15:25:33 - INFO - codeparrot_training - Step 20632: {'lr': 0.00033603783901389744, 'samples': 3961536, 'steps': 20632, 'loss/train': 2.1644797325134277} 01/29/2022 15:25:38 - INFO - codeparrot_training - Step 20633: {'lr': 0.00033602247588222545, 'samples': 3961728, 'steps': 20633, 'loss/train': 1.688739538192749} 01/29/2022 15:25:45 - INFO - codeparrot_training - Step 20634: {'lr': 0.00033600711238206056, 'samples': 3961920, 'steps': 20634, 'loss/train': 1.3663673400878906} 01/29/2022 15:25:49 - INFO - codeparrot_training - Step 20635: {'lr': 0.00033599174851346844, 'samples': 3962112, 'steps': 20635, 'loss/train': 1.5412566661834717} 01/29/2022 15:25:53 - INFO - codeparrot_training - Step 20636: {'lr': 0.000335976384276515, 'samples': 3962304, 'steps': 20636, 'loss/train': 1.922441840171814} 01/29/2022 15:25:57 - INFO - codeparrot_training - Step 20637: {'lr': 0.000335961019671266, 'samples': 3962496, 'steps': 20637, 'loss/train': 1.6026606559753418} 01/29/2022 15:26:02 - INFO - codeparrot_training - Step 20638: {'lr': 0.00033594565469778734, 'samples': 3962688, 'steps': 20638, 'loss/train': 1.3747539818286896} 01/29/2022 15:26:07 - INFO - codeparrot_training - Step 20639: {'lr': 0.00033593028935614475, 'samples': 3962880, 'steps': 20639, 'loss/train': 1.7050738334655762} 01/29/2022 15:26:11 - INFO - codeparrot_training - Step 20640: {'lr': 0.0003359149236464041, 'samples': 3963072, 'steps': 20640, 'loss/train': 1.4253871142864227} 01/29/2022 15:26:16 - INFO - codeparrot_training - Step 20641: {'lr': 0.0003358995575686311, 'samples': 3963264, 'steps': 20641, 'loss/train': 1.4648299813270569} 01/29/2022 15:26:20 - INFO - codeparrot_training - Step 20642: {'lr': 0.00033588419112289176, 'samples': 3963456, 'steps': 20642, 'loss/train': 1.5113024711608887} 01/29/2022 15:26:24 - INFO - codeparrot_training - Step 20643: {'lr': 0.00033586882430925184, 'samples': 3963648, 'steps': 20643, 'loss/train': 1.2443740367889404} 01/29/2022 15:26:29 - INFO - codeparrot_training - Step 20644: {'lr': 0.0003358534571277771, 'samples': 3963840, 'steps': 20644, 'loss/train': 2.092617094516754} 01/29/2022 15:26:34 - INFO - codeparrot_training - Step 20645: {'lr': 0.0003358380895785335, 'samples': 3964032, 'steps': 20645, 'loss/train': 0.02487914264202118} 01/29/2022 15:26:38 - INFO - codeparrot_training - Step 20646: {'lr': 0.00033582272166158666, 'samples': 3964224, 'steps': 20646, 'loss/train': 1.7763079404830933} 01/29/2022 15:26:42 - INFO - codeparrot_training - Step 20647: {'lr': 0.00033580735337700266, 'samples': 3964416, 'steps': 20647, 'loss/train': 1.4006497263908386} 01/29/2022 15:26:46 - INFO - codeparrot_training - Step 20648: {'lr': 0.00033579198472484707, 'samples': 3964608, 'steps': 20648, 'loss/train': 1.6353837847709656} 01/29/2022 15:26:54 - INFO - codeparrot_training - Step 20649: {'lr': 0.0003357766157051859, 'samples': 3964800, 'steps': 20649, 'loss/train': 1.6166986227035522} 01/29/2022 15:26:58 - INFO - codeparrot_training - Step 20650: {'lr': 0.000335761246318085, 'samples': 3964992, 'steps': 20650, 'loss/train': 1.5684648156166077} 01/29/2022 15:27:02 - INFO - codeparrot_training - Step 20651: {'lr': 0.0003357458765636101, 'samples': 3965184, 'steps': 20651, 'loss/train': 2.065155029296875} 01/29/2022 15:27:06 - INFO - codeparrot_training - Step 20652: {'lr': 0.00033573050644182713, 'samples': 3965376, 'steps': 20652, 'loss/train': 1.264516443014145} 01/29/2022 15:27:10 - INFO - codeparrot_training - Step 20653: {'lr': 0.00033571513595280185, 'samples': 3965568, 'steps': 20653, 'loss/train': 1.285666286945343} 01/29/2022 15:27:16 - INFO - codeparrot_training - Step 20654: {'lr': 0.0003356997650966002, 'samples': 3965760, 'steps': 20654, 'loss/train': 2.1707873940467834} 01/29/2022 15:27:20 - INFO - codeparrot_training - Step 20655: {'lr': 0.0003356843938732879, 'samples': 3965952, 'steps': 20655, 'loss/train': 1.4746550917625427} 01/29/2022 15:27:24 - INFO - codeparrot_training - Step 20656: {'lr': 0.0003356690222829309, 'samples': 3966144, 'steps': 20656, 'loss/train': 1.7867962718009949} 01/29/2022 15:27:29 - INFO - codeparrot_training - Step 20657: {'lr': 0.00033565365032559496, 'samples': 3966336, 'steps': 20657, 'loss/train': 2.169263184070587} 01/29/2022 15:27:33 - INFO - codeparrot_training - Step 20658: {'lr': 0.00033563827800134604, 'samples': 3966528, 'steps': 20658, 'loss/train': 1.1588896214962006} 01/29/2022 15:27:38 - INFO - codeparrot_training - Step 20659: {'lr': 0.00033562290531025, 'samples': 3966720, 'steps': 20659, 'loss/train': 1.2660705149173737} 01/29/2022 15:27:42 - INFO - codeparrot_training - Step 20660: {'lr': 0.0003356075322523725, 'samples': 3966912, 'steps': 20660, 'loss/train': 0.7864570319652557} 01/29/2022 15:27:47 - INFO - codeparrot_training - Step 20661: {'lr': 0.00033559215882777955, 'samples': 3967104, 'steps': 20661, 'loss/train': 2.316074252128601} 01/29/2022 15:27:51 - INFO - codeparrot_training - Step 20662: {'lr': 0.000335576785036537, 'samples': 3967296, 'steps': 20662, 'loss/train': 1.1009271740913391} 01/29/2022 15:27:55 - INFO - codeparrot_training - Step 20663: {'lr': 0.0003355614108787106, 'samples': 3967488, 'steps': 20663, 'loss/train': 1.4852557182312012} 01/29/2022 15:28:00 - INFO - codeparrot_training - Step 20664: {'lr': 0.00033554603635436645, 'samples': 3967680, 'steps': 20664, 'loss/train': 0.9702913463115692} 01/29/2022 15:28:05 - INFO - codeparrot_training - Step 20665: {'lr': 0.00033553066146357, 'samples': 3967872, 'steps': 20665, 'loss/train': 1.2300085723400116} 01/29/2022 15:28:09 - INFO - codeparrot_training - Step 20666: {'lr': 0.0003355152862063875, 'samples': 3968064, 'steps': 20666, 'loss/train': 1.3212489187717438} 01/29/2022 15:28:13 - INFO - codeparrot_training - Step 20667: {'lr': 0.0003354999105828847, 'samples': 3968256, 'steps': 20667, 'loss/train': 1.9712485671043396} 01/29/2022 15:28:17 - INFO - codeparrot_training - Step 20668: {'lr': 0.0003354845345931274, 'samples': 3968448, 'steps': 20668, 'loss/train': 1.8012328147888184} 01/29/2022 15:28:25 - INFO - codeparrot_training - Step 20669: {'lr': 0.0003354691582371815, 'samples': 3968640, 'steps': 20669, 'loss/train': 1.0455325841903687} 01/29/2022 15:28:29 - INFO - codeparrot_training - Step 20670: {'lr': 0.00033545378151511287, 'samples': 3968832, 'steps': 20670, 'loss/train': 1.3675392866134644} 01/29/2022 15:28:34 - INFO - codeparrot_training - Step 20671: {'lr': 0.00033543840442698737, 'samples': 3969024, 'steps': 20671, 'loss/train': 0.10084109008312225} 01/29/2022 15:28:38 - INFO - codeparrot_training - Step 20672: {'lr': 0.0003354230269728709, 'samples': 3969216, 'steps': 20672, 'loss/train': 0.4533005654811859} 01/29/2022 15:28:42 - INFO - codeparrot_training - Step 20673: {'lr': 0.0003354076491528292, 'samples': 3969408, 'steps': 20673, 'loss/train': 1.0238211750984192} 01/29/2022 15:28:48 - INFO - codeparrot_training - Step 20674: {'lr': 0.00033539227096692837, 'samples': 3969600, 'steps': 20674, 'loss/train': 1.4723173677921295} 01/29/2022 15:28:52 - INFO - codeparrot_training - Step 20675: {'lr': 0.00033537689241523407, 'samples': 3969792, 'steps': 20675, 'loss/train': 2.199397087097168} 01/29/2022 15:28:56 - INFO - codeparrot_training - Step 20676: {'lr': 0.00033536151349781236, 'samples': 3969984, 'steps': 20676, 'loss/train': 2.8803001642227173} 01/29/2022 15:29:00 - INFO - codeparrot_training - Step 20677: {'lr': 0.000335346134214729, 'samples': 3970176, 'steps': 20677, 'loss/train': 1.589141607284546} 01/29/2022 15:29:05 - INFO - codeparrot_training - Step 20678: {'lr': 0.0003353307545660499, 'samples': 3970368, 'steps': 20678, 'loss/train': 1.8927977085113525} 01/29/2022 15:29:10 - INFO - codeparrot_training - Step 20679: {'lr': 0.0003353153745518409, 'samples': 3970560, 'steps': 20679, 'loss/train': 1.6576032042503357} 01/29/2022 15:29:14 - INFO - codeparrot_training - Step 20680: {'lr': 0.000335299994172168, 'samples': 3970752, 'steps': 20680, 'loss/train': 0.7685587406158447} 01/29/2022 15:29:18 - INFO - codeparrot_training - Step 20681: {'lr': 0.0003352846134270969, 'samples': 3970944, 'steps': 20681, 'loss/train': 1.1717989146709442} 01/29/2022 15:29:22 - INFO - codeparrot_training - Step 20682: {'lr': 0.0003352692323166938, 'samples': 3971136, 'steps': 20682, 'loss/train': 1.8088831901550293} 01/29/2022 15:29:30 - INFO - codeparrot_training - Step 20683: {'lr': 0.0003352538508410242, 'samples': 3971328, 'steps': 20683, 'loss/train': 1.533752202987671} 01/29/2022 15:29:34 - INFO - codeparrot_training - Step 20684: {'lr': 0.00033523846900015427, 'samples': 3971520, 'steps': 20684, 'loss/train': 0.9627051651477814} 01/29/2022 15:29:38 - INFO - codeparrot_training - Step 20685: {'lr': 0.0003352230867941497, 'samples': 3971712, 'steps': 20685, 'loss/train': 1.0651317536830902} 01/29/2022 15:29:42 - INFO - codeparrot_training - Step 20686: {'lr': 0.0003352077042230766, 'samples': 3971904, 'steps': 20686, 'loss/train': 2.0810242295265198} 01/29/2022 15:29:47 - INFO - codeparrot_training - Step 20687: {'lr': 0.0003351923212870007, 'samples': 3972096, 'steps': 20687, 'loss/train': 1.4019235968589783} 01/29/2022 15:29:52 - INFO - codeparrot_training - Step 20688: {'lr': 0.000335176937985988, 'samples': 3972288, 'steps': 20688, 'loss/train': 1.8699992895126343} 01/29/2022 15:29:56 - INFO - codeparrot_training - Step 20689: {'lr': 0.0003351615543201042, 'samples': 3972480, 'steps': 20689, 'loss/train': 1.5509242415428162} 01/29/2022 15:30:01 - INFO - codeparrot_training - Step 20690: {'lr': 0.00033514617028941547, 'samples': 3972672, 'steps': 20690, 'loss/train': 1.5711806416511536} 01/29/2022 15:30:05 - INFO - codeparrot_training - Step 20691: {'lr': 0.0003351307858939875, 'samples': 3972864, 'steps': 20691, 'loss/train': 2.199850022792816} 01/29/2022 15:30:09 - INFO - codeparrot_training - Step 20692: {'lr': 0.0003351154011338864, 'samples': 3973056, 'steps': 20692, 'loss/train': 0.7842033505439758} 01/29/2022 15:30:16 - INFO - codeparrot_training - Step 20693: {'lr': 0.00033510001600917783, 'samples': 3973248, 'steps': 20693, 'loss/train': 1.3485438823699951} 01/29/2022 15:30:20 - INFO - codeparrot_training - Step 20694: {'lr': 0.00033508463051992786, 'samples': 3973440, 'steps': 20694, 'loss/train': 1.210010826587677} 01/29/2022 15:30:25 - INFO - codeparrot_training - Step 20695: {'lr': 0.00033506924466620235, 'samples': 3973632, 'steps': 20695, 'loss/train': 2.2278708815574646} 01/29/2022 15:30:29 - INFO - codeparrot_training - Step 20696: {'lr': 0.0003350538584480672, 'samples': 3973824, 'steps': 20696, 'loss/train': 1.3840405941009521} 01/29/2022 15:30:33 - INFO - codeparrot_training - Step 20697: {'lr': 0.0003350384718655884, 'samples': 3974016, 'steps': 20697, 'loss/train': 1.7496973872184753} 01/29/2022 15:30:39 - INFO - codeparrot_training - Step 20698: {'lr': 0.0003350230849188317, 'samples': 3974208, 'steps': 20698, 'loss/train': 2.257767677307129} 01/29/2022 15:30:43 - INFO - codeparrot_training - Step 20699: {'lr': 0.00033500769760786314, 'samples': 3974400, 'steps': 20699, 'loss/train': 1.6711267232894897} 01/29/2022 15:30:47 - INFO - codeparrot_training - Step 20700: {'lr': 0.0003349923099327485, 'samples': 3974592, 'steps': 20700, 'loss/train': 2.042680263519287} 01/29/2022 15:30:51 - INFO - codeparrot_training - Step 20701: {'lr': 0.000334976921893554, 'samples': 3974784, 'steps': 20701, 'loss/train': 0.5528790503740311} 01/29/2022 15:30:56 - INFO - codeparrot_training - Step 20702: {'lr': 0.0003349615334903452, 'samples': 3974976, 'steps': 20702, 'loss/train': 1.349012017250061} 01/29/2022 15:31:01 - INFO - codeparrot_training - Step 20703: {'lr': 0.00033494614472318816, 'samples': 3975168, 'steps': 20703, 'loss/train': 1.7043916583061218} 01/29/2022 15:31:06 - INFO - codeparrot_training - Step 20704: {'lr': 0.00033493075559214885, 'samples': 3975360, 'steps': 20704, 'loss/train': 2.855801582336426} 01/29/2022 15:31:10 - INFO - codeparrot_training - Step 20705: {'lr': 0.00033491536609729313, 'samples': 3975552, 'steps': 20705, 'loss/train': 2.275004804134369} 01/29/2022 15:31:14 - INFO - codeparrot_training - Step 20706: {'lr': 0.000334899976238687, 'samples': 3975744, 'steps': 20706, 'loss/train': 0.9977985620498657} 01/29/2022 15:31:18 - INFO - codeparrot_training - Step 20707: {'lr': 0.00033488458601639624, 'samples': 3975936, 'steps': 20707, 'loss/train': 1.2597219049930573} 01/29/2022 15:31:25 - INFO - codeparrot_training - Step 20708: {'lr': 0.000334869195430487, 'samples': 3976128, 'steps': 20708, 'loss/train': 1.3876125812530518} 01/29/2022 15:31:30 - INFO - codeparrot_training - Step 20709: {'lr': 0.00033485380448102496, 'samples': 3976320, 'steps': 20709, 'loss/train': 1.7161086201667786} 01/29/2022 15:31:34 - INFO - codeparrot_training - Step 20710: {'lr': 0.0003348384131680762, 'samples': 3976512, 'steps': 20710, 'loss/train': 1.5981978178024292} 01/29/2022 15:31:38 - INFO - codeparrot_training - Step 20711: {'lr': 0.0003348230214917066, 'samples': 3976704, 'steps': 20711, 'loss/train': 0.9947510361671448} 01/29/2022 15:31:44 - INFO - codeparrot_training - Step 20712: {'lr': 0.0003348076294519822, 'samples': 3976896, 'steps': 20712, 'loss/train': 0.9111236929893494} 01/29/2022 15:31:49 - INFO - codeparrot_training - Step 20713: {'lr': 0.0003347922370489687, 'samples': 3977088, 'steps': 20713, 'loss/train': 1.3525833785533905} 01/29/2022 15:31:53 - INFO - codeparrot_training - Step 20714: {'lr': 0.00033477684428273233, 'samples': 3977280, 'steps': 20714, 'loss/train': 1.1689449548721313} 01/29/2022 15:31:58 - INFO - codeparrot_training - Step 20715: {'lr': 0.0003347614511533388, 'samples': 3977472, 'steps': 20715, 'loss/train': 1.6495174169540405} 01/29/2022 15:32:02 - INFO - codeparrot_training - Step 20716: {'lr': 0.0003347460576608541, 'samples': 3977664, 'steps': 20716, 'loss/train': 1.9044798016548157} 01/29/2022 15:32:06 - INFO - codeparrot_training - Step 20717: {'lr': 0.00033473066380534423, 'samples': 3977856, 'steps': 20717, 'loss/train': 1.4335607886314392} 01/29/2022 15:32:13 - INFO - codeparrot_training - Step 20718: {'lr': 0.00033471526958687514, 'samples': 3978048, 'steps': 20718, 'loss/train': 1.1188367307186127} 01/29/2022 15:32:17 - INFO - codeparrot_training - Step 20719: {'lr': 0.0003346998750055127, 'samples': 3978240, 'steps': 20719, 'loss/train': 1.6998054385185242} 01/29/2022 15:32:22 - INFO - codeparrot_training - Step 20720: {'lr': 0.0003346844800613229, 'samples': 3978432, 'steps': 20720, 'loss/train': 1.496210664510727} 01/29/2022 15:32:26 - INFO - codeparrot_training - Step 20721: {'lr': 0.0003346690847543717, 'samples': 3978624, 'steps': 20721, 'loss/train': 7.364120721817017} 01/29/2022 15:32:30 - INFO - codeparrot_training - Step 20722: {'lr': 0.00033465368908472496, 'samples': 3978816, 'steps': 20722, 'loss/train': 2.3030269145965576} 01/29/2022 15:32:35 - INFO - codeparrot_training - Step 20723: {'lr': 0.00033463829305244874, 'samples': 3979008, 'steps': 20723, 'loss/train': 1.6213780045509338} 01/29/2022 15:32:40 - INFO - codeparrot_training - Step 20724: {'lr': 0.0003346228966576089, 'samples': 3979200, 'steps': 20724, 'loss/train': 1.9527198672294617} 01/29/2022 15:32:44 - INFO - codeparrot_training - Step 20725: {'lr': 0.0003346074999002715, 'samples': 3979392, 'steps': 20725, 'loss/train': 1.8476529121398926} 01/29/2022 15:32:48 - INFO - codeparrot_training - Step 20726: {'lr': 0.0003345921027805024, 'samples': 3979584, 'steps': 20726, 'loss/train': 1.1755662560462952} 01/29/2022 15:32:52 - INFO - codeparrot_training - Step 20727: {'lr': 0.00033457670529836756, 'samples': 3979776, 'steps': 20727, 'loss/train': 0.6994704455137253} 01/29/2022 15:32:58 - INFO - codeparrot_training - Step 20728: {'lr': 0.0003345613074539331, 'samples': 3979968, 'steps': 20728, 'loss/train': 1.8496786952018738} 01/29/2022 15:33:02 - INFO - codeparrot_training - Step 20729: {'lr': 0.00033454590924726467, 'samples': 3980160, 'steps': 20729, 'loss/train': 1.707481563091278} 01/29/2022 15:33:06 - INFO - codeparrot_training - Step 20730: {'lr': 0.0003345305106784286, 'samples': 3980352, 'steps': 20730, 'loss/train': 1.5915974378585815} 01/29/2022 15:33:10 - INFO - codeparrot_training - Step 20731: {'lr': 0.00033451511174749057, 'samples': 3980544, 'steps': 20731, 'loss/train': 0.9873283803462982} 01/29/2022 15:33:15 - INFO - codeparrot_training - Step 20732: {'lr': 0.0003344997124545166, 'samples': 3980736, 'steps': 20732, 'loss/train': 0.024102581664919853} 01/29/2022 15:33:20 - INFO - codeparrot_training - Step 20733: {'lr': 0.0003344843127995728, 'samples': 3980928, 'steps': 20733, 'loss/train': 1.4944276213645935} 01/29/2022 15:33:25 - INFO - codeparrot_training - Step 20734: {'lr': 0.00033446891278272493, 'samples': 3981120, 'steps': 20734, 'loss/train': 1.156048983335495} 01/29/2022 15:33:29 - INFO - codeparrot_training - Step 20735: {'lr': 0.0003344535124040391, 'samples': 3981312, 'steps': 20735, 'loss/train': 1.7106256484985352} 01/29/2022 15:33:33 - INFO - codeparrot_training - Step 20736: {'lr': 0.0003344381116635812, 'samples': 3981504, 'steps': 20736, 'loss/train': 0.22635946422815323} 01/29/2022 15:33:37 - INFO - codeparrot_training - Step 20737: {'lr': 0.0003344227105614173, 'samples': 3981696, 'steps': 20737, 'loss/train': 1.9549328684806824} 01/29/2022 15:33:45 - INFO - codeparrot_training - Step 20738: {'lr': 0.0003344073090976132, 'samples': 3981888, 'steps': 20738, 'loss/train': 1.9829775094985962} 01/29/2022 15:33:49 - INFO - codeparrot_training - Step 20739: {'lr': 0.00033439190727223517, 'samples': 3982080, 'steps': 20739, 'loss/train': 2.8201326727867126} 01/29/2022 15:33:53 - INFO - codeparrot_training - Step 20740: {'lr': 0.00033437650508534887, 'samples': 3982272, 'steps': 20740, 'loss/train': 1.4784728586673737} 01/29/2022 15:33:57 - INFO - codeparrot_training - Step 20741: {'lr': 0.0003343611025370205, 'samples': 3982464, 'steps': 20741, 'loss/train': 1.3386459052562714} 01/29/2022 15:34:02 - INFO - codeparrot_training - Step 20742: {'lr': 0.00033434569962731593, 'samples': 3982656, 'steps': 20742, 'loss/train': 2.163412392139435} 01/29/2022 15:34:07 - INFO - codeparrot_training - Step 20743: {'lr': 0.0003343302963563012, 'samples': 3982848, 'steps': 20743, 'loss/train': 0.9729949235916138} 01/29/2022 15:34:11 - INFO - codeparrot_training - Step 20744: {'lr': 0.00033431489272404215, 'samples': 3983040, 'steps': 20744, 'loss/train': 2.193366050720215} 01/29/2022 15:34:15 - INFO - codeparrot_training - Step 20745: {'lr': 0.00033429948873060496, 'samples': 3983232, 'steps': 20745, 'loss/train': 1.4614900946617126} 01/29/2022 15:34:20 - INFO - codeparrot_training - Step 20746: {'lr': 0.0003342840843760555, 'samples': 3983424, 'steps': 20746, 'loss/train': 1.1534090638160706} 01/29/2022 15:34:24 - INFO - codeparrot_training - Step 20747: {'lr': 0.00033426867966045984, 'samples': 3983616, 'steps': 20747, 'loss/train': 1.65896737575531} 01/29/2022 15:34:29 - INFO - codeparrot_training - Step 20748: {'lr': 0.00033425327458388375, 'samples': 3983808, 'steps': 20748, 'loss/train': 2.4810630083084106} 01/29/2022 15:34:34 - INFO - codeparrot_training - Step 20749: {'lr': 0.0003342378691463936, 'samples': 3984000, 'steps': 20749, 'loss/train': 0.8940116465091705} 01/29/2022 15:34:38 - INFO - codeparrot_training - Step 20750: {'lr': 0.00033422246334805503, 'samples': 3984192, 'steps': 20750, 'loss/train': 0.03701603785157204} 01/29/2022 15:34:42 - INFO - codeparrot_training - Step 20751: {'lr': 0.0003342070571889342, 'samples': 3984384, 'steps': 20751, 'loss/train': 2.7757657170295715} 01/29/2022 15:34:49 - INFO - codeparrot_training - Step 20752: {'lr': 0.00033419165066909707, 'samples': 3984576, 'steps': 20752, 'loss/train': 1.3682125210762024} 01/29/2022 15:34:53 - INFO - codeparrot_training - Step 20753: {'lr': 0.0003341762437886097, 'samples': 3984768, 'steps': 20753, 'loss/train': 1.9323154091835022} 01/29/2022 15:34:58 - INFO - codeparrot_training - Step 20754: {'lr': 0.0003341608365475379, 'samples': 3984960, 'steps': 20754, 'loss/train': 0.5639694929122925} 01/29/2022 15:35:02 - INFO - codeparrot_training - Step 20755: {'lr': 0.00033414542894594793, 'samples': 3985152, 'steps': 20755, 'loss/train': 1.3723271191120148} 01/29/2022 15:35:06 - INFO - codeparrot_training - Step 20756: {'lr': 0.00033413002098390567, 'samples': 3985344, 'steps': 20756, 'loss/train': 1.041151374578476} 01/29/2022 15:35:11 - INFO - codeparrot_training - Step 20757: {'lr': 0.00033411461266147705, 'samples': 3985536, 'steps': 20757, 'loss/train': 2.10954612493515} 01/29/2022 15:35:16 - INFO - codeparrot_training - Step 20758: {'lr': 0.00033409920397872814, 'samples': 3985728, 'steps': 20758, 'loss/train': 0.9892973005771637} 01/29/2022 15:35:20 - INFO - codeparrot_training - Step 20759: {'lr': 0.00033408379493572493, 'samples': 3985920, 'steps': 20759, 'loss/train': 1.3197703063488007} 01/29/2022 15:35:24 - INFO - codeparrot_training - Step 20760: {'lr': 0.0003340683855325335, 'samples': 3986112, 'steps': 20760, 'loss/train': 1.724420964717865} 01/29/2022 15:35:29 - INFO - codeparrot_training - Step 20761: {'lr': 0.00033405297576921976, 'samples': 3986304, 'steps': 20761, 'loss/train': 2.236099362373352} 01/29/2022 15:35:36 - INFO - codeparrot_training - Step 20762: {'lr': 0.00033403756564584974, 'samples': 3986496, 'steps': 20762, 'loss/train': 1.69395250082016} 01/29/2022 15:35:40 - INFO - codeparrot_training - Step 20763: {'lr': 0.0003340221551624896, 'samples': 3986688, 'steps': 20763, 'loss/train': 1.017003446817398} 01/29/2022 15:35:44 - INFO - codeparrot_training - Step 20764: {'lr': 0.0003340067443192051, 'samples': 3986880, 'steps': 20764, 'loss/train': 0.35336876660585403} 01/29/2022 15:35:49 - INFO - codeparrot_training - Step 20765: {'lr': 0.0003339913331160624, 'samples': 3987072, 'steps': 20765, 'loss/train': 0.30203888565301895} 01/29/2022 15:35:53 - INFO - codeparrot_training - Step 20766: {'lr': 0.0003339759215531275, 'samples': 3987264, 'steps': 20766, 'loss/train': 1.0381840467453003} 01/29/2022 15:35:58 - INFO - codeparrot_training - Step 20767: {'lr': 0.0003339605096304664, 'samples': 3987456, 'steps': 20767, 'loss/train': 1.9184051156044006} 01/29/2022 15:36:02 - INFO - codeparrot_training - Step 20768: {'lr': 0.00033394509734814516, 'samples': 3987648, 'steps': 20768, 'loss/train': 1.5343753695487976} 01/29/2022 15:36:07 - INFO - codeparrot_training - Step 20769: {'lr': 0.00033392968470622987, 'samples': 3987840, 'steps': 20769, 'loss/train': 1.6760458946228027} 01/29/2022 15:36:11 - INFO - codeparrot_training - Step 20770: {'lr': 0.0003339142717047863, 'samples': 3988032, 'steps': 20770, 'loss/train': 1.6416215300559998} 01/29/2022 15:36:15 - INFO - codeparrot_training - Step 20771: {'lr': 0.0003338988583438808, 'samples': 3988224, 'steps': 20771, 'loss/train': 2.168592631816864} 01/29/2022 15:36:20 - INFO - codeparrot_training - Step 20772: {'lr': 0.0003338834446235791, 'samples': 3988416, 'steps': 20772, 'loss/train': 1.463373452425003} 01/29/2022 15:36:25 - INFO - codeparrot_training - Step 20773: {'lr': 0.00033386803054394744, 'samples': 3988608, 'steps': 20773, 'loss/train': 1.1393744051456451} 01/29/2022 15:36:29 - INFO - codeparrot_training - Step 20774: {'lr': 0.0003338526161050517, 'samples': 3988800, 'steps': 20774, 'loss/train': 1.8662574291229248} 01/29/2022 15:36:33 - INFO - codeparrot_training - Step 20775: {'lr': 0.00033383720130695794, 'samples': 3988992, 'steps': 20775, 'loss/train': 0.9698847234249115} 01/29/2022 15:36:37 - INFO - codeparrot_training - Step 20776: {'lr': 0.0003338217861497324, 'samples': 3989184, 'steps': 20776, 'loss/train': 1.4524587392807007} 01/29/2022 15:36:44 - INFO - codeparrot_training - Step 20777: {'lr': 0.0003338063706334408, 'samples': 3989376, 'steps': 20777, 'loss/train': 1.2064872086048126} 01/29/2022 15:36:49 - INFO - codeparrot_training - Step 20778: {'lr': 0.00033379095475814937, 'samples': 3989568, 'steps': 20778, 'loss/train': 1.0620570480823517} 01/29/2022 15:36:53 - INFO - codeparrot_training - Step 20779: {'lr': 0.00033377553852392404, 'samples': 3989760, 'steps': 20779, 'loss/train': 1.2934266328811646} 01/29/2022 15:36:57 - INFO - codeparrot_training - Step 20780: {'lr': 0.000333760121930831, 'samples': 3989952, 'steps': 20780, 'loss/train': 1.536778450012207} 01/29/2022 15:37:01 - INFO - codeparrot_training - Step 20781: {'lr': 0.00033374470497893614, 'samples': 3990144, 'steps': 20781, 'loss/train': 0.25684694945812225} 01/29/2022 15:37:07 - INFO - codeparrot_training - Step 20782: {'lr': 0.0003337292876683056, 'samples': 3990336, 'steps': 20782, 'loss/train': 1.7025303840637207} 01/29/2022 15:37:11 - INFO - codeparrot_training - Step 20783: {'lr': 0.0003337138699990053, 'samples': 3990528, 'steps': 20783, 'loss/train': 1.242076724767685} 01/29/2022 15:37:15 - INFO - codeparrot_training - Step 20784: {'lr': 0.00033369845197110144, 'samples': 3990720, 'steps': 20784, 'loss/train': 2.0841994285583496} 01/29/2022 15:37:19 - INFO - codeparrot_training - Step 20785: {'lr': 0.00033368303358465994, 'samples': 3990912, 'steps': 20785, 'loss/train': 1.5980215072631836} 01/29/2022 15:37:24 - INFO - codeparrot_training - Step 20786: {'lr': 0.00033366761483974693, 'samples': 3991104, 'steps': 20786, 'loss/train': 1.368498533964157} 01/29/2022 15:37:29 - INFO - codeparrot_training - Step 20787: {'lr': 0.0003336521957364284, 'samples': 3991296, 'steps': 20787, 'loss/train': 1.128170907497406} 01/29/2022 15:37:33 - INFO - codeparrot_training - Step 20788: {'lr': 0.0003336367762747704, 'samples': 3991488, 'steps': 20788, 'loss/train': 1.5224854946136475} 01/29/2022 15:37:37 - INFO - codeparrot_training - Step 20789: {'lr': 0.0003336213564548391, 'samples': 3991680, 'steps': 20789, 'loss/train': 0.8008773028850555} 01/29/2022 15:37:41 - INFO - codeparrot_training - Step 20790: {'lr': 0.0003336059362767004, 'samples': 3991872, 'steps': 20790, 'loss/train': 2.1459937691688538} 01/29/2022 15:37:46 - INFO - codeparrot_training - Step 20791: {'lr': 0.0003335905157404204, 'samples': 3992064, 'steps': 20791, 'loss/train': 0.9612685739994049} 01/29/2022 15:37:51 - INFO - codeparrot_training - Step 20792: {'lr': 0.0003335750948460652, 'samples': 3992256, 'steps': 20792, 'loss/train': 2.1800195574760437} 01/29/2022 15:37:56 - INFO - codeparrot_training - Step 20793: {'lr': 0.0003335596735937009, 'samples': 3992448, 'steps': 20793, 'loss/train': 3.372136116027832} 01/29/2022 15:38:00 - INFO - codeparrot_training - Step 20794: {'lr': 0.0003335442519833933, 'samples': 3992640, 'steps': 20794, 'loss/train': 1.2938776016235352} 01/29/2022 15:38:04 - INFO - codeparrot_training - Step 20795: {'lr': 0.00033352883001520884, 'samples': 3992832, 'steps': 20795, 'loss/train': 1.4494560062885284} 01/29/2022 15:38:08 - INFO - codeparrot_training - Step 20796: {'lr': 0.0003335134076892133, 'samples': 3993024, 'steps': 20796, 'loss/train': 1.6833711862564087} 01/29/2022 15:38:15 - INFO - codeparrot_training - Step 20797: {'lr': 0.0003334979850054729, 'samples': 3993216, 'steps': 20797, 'loss/train': 1.9853074550628662} 01/29/2022 15:38:20 - INFO - codeparrot_training - Step 20798: {'lr': 0.0003334825619640536, 'samples': 3993408, 'steps': 20798, 'loss/train': 1.3641487061977386} 01/29/2022 15:38:24 - INFO - codeparrot_training - Step 20799: {'lr': 0.0003334671385650215, 'samples': 3993600, 'steps': 20799, 'loss/train': 1.0631844699382782} 01/29/2022 15:38:28 - INFO - codeparrot_training - Step 20800: {'lr': 0.0003334517148084427, 'samples': 3993792, 'steps': 20800, 'loss/train': 2.0943835973739624} 01/29/2022 15:38:34 - INFO - codeparrot_training - Step 20801: {'lr': 0.00033343629069438333, 'samples': 3993984, 'steps': 20801, 'loss/train': 1.5599707961082458} 01/29/2022 15:38:38 - INFO - codeparrot_training - Step 20802: {'lr': 0.0003334208662229093, 'samples': 3994176, 'steps': 20802, 'loss/train': 1.3107604086399078} 01/29/2022 15:38:42 - INFO - codeparrot_training - Step 20803: {'lr': 0.0003334054413940868, 'samples': 3994368, 'steps': 20803, 'loss/train': 1.4725946187973022} 01/29/2022 15:38:47 - INFO - codeparrot_training - Step 20804: {'lr': 0.0003333900162079818, 'samples': 3994560, 'steps': 20804, 'loss/train': 1.4075176119804382} 01/29/2022 15:38:51 - INFO - codeparrot_training - Step 20805: {'lr': 0.00033337459066466057, 'samples': 3994752, 'steps': 20805, 'loss/train': 1.8063027262687683} 01/29/2022 15:38:55 - INFO - codeparrot_training - Step 20806: {'lr': 0.000333359164764189, 'samples': 3994944, 'steps': 20806, 'loss/train': 1.1738193333148956} 01/29/2022 15:39:02 - INFO - codeparrot_training - Step 20807: {'lr': 0.00033334373850663323, 'samples': 3995136, 'steps': 20807, 'loss/train': 1.6813432574272156} 01/29/2022 15:39:06 - INFO - codeparrot_training - Step 20808: {'lr': 0.00033332831189205936, 'samples': 3995328, 'steps': 20808, 'loss/train': 1.6133742928504944} 01/29/2022 15:39:11 - INFO - codeparrot_training - Step 20809: {'lr': 0.00033331288492053344, 'samples': 3995520, 'steps': 20809, 'loss/train': 0.42559415102005005} 01/29/2022 15:39:15 - INFO - codeparrot_training - Step 20810: {'lr': 0.0003332974575921217, 'samples': 3995712, 'steps': 20810, 'loss/train': 0.5386562794446945} 01/29/2022 15:39:19 - INFO - codeparrot_training - Step 20811: {'lr': 0.00033328202990688996, 'samples': 3995904, 'steps': 20811, 'loss/train': 1.9301045536994934} 01/29/2022 15:39:24 - INFO - codeparrot_training - Step 20812: {'lr': 0.0003332666018649044, 'samples': 3996096, 'steps': 20812, 'loss/train': 0.034463522024452686} 01/29/2022 15:39:29 - INFO - codeparrot_training - Step 20813: {'lr': 0.00033325117346623135, 'samples': 3996288, 'steps': 20813, 'loss/train': 0.5683047473430634} 01/29/2022 15:39:33 - INFO - codeparrot_training - Step 20814: {'lr': 0.00033323574471093656, 'samples': 3996480, 'steps': 20814, 'loss/train': 1.957407832145691} 01/29/2022 15:39:37 - INFO - codeparrot_training - Step 20815: {'lr': 0.0003332203155990863, 'samples': 3996672, 'steps': 20815, 'loss/train': 1.508158028125763} 01/29/2022 15:39:41 - INFO - codeparrot_training - Step 20816: {'lr': 0.00033320488613074666, 'samples': 3996864, 'steps': 20816, 'loss/train': 1.2576942443847656} 01/29/2022 15:39:47 - INFO - codeparrot_training - Step 20817: {'lr': 0.00033318945630598373, 'samples': 3997056, 'steps': 20817, 'loss/train': 1.9489281177520752} 01/29/2022 15:39:51 - INFO - codeparrot_training - Step 20818: {'lr': 0.00033317402612486355, 'samples': 3997248, 'steps': 20818, 'loss/train': 0.15650837123394012} 01/29/2022 15:39:55 - INFO - codeparrot_training - Step 20819: {'lr': 0.00033315859558745225, 'samples': 3997440, 'steps': 20819, 'loss/train': 2.0657594203948975} 01/29/2022 15:40:00 - INFO - codeparrot_training - Step 20820: {'lr': 0.0003331431646938159, 'samples': 3997632, 'steps': 20820, 'loss/train': 1.2078742682933807} 01/29/2022 15:40:04 - INFO - codeparrot_training - Step 20821: {'lr': 0.00033312773344402075, 'samples': 3997824, 'steps': 20821, 'loss/train': 1.406617283821106} 01/29/2022 15:40:11 - INFO - codeparrot_training - Step 20822: {'lr': 0.00033311230183813266, 'samples': 3998016, 'steps': 20822, 'loss/train': 1.5953078269958496} 01/29/2022 15:40:15 - INFO - codeparrot_training - Step 20823: {'lr': 0.0003330968698762179, 'samples': 3998208, 'steps': 20823, 'loss/train': 1.237924039363861} 01/29/2022 15:40:19 - INFO - codeparrot_training - Step 20824: {'lr': 0.0003330814375583426, 'samples': 3998400, 'steps': 20824, 'loss/train': 1.681788146495819} 01/29/2022 15:40:23 - INFO - codeparrot_training - Step 20825: {'lr': 0.00033306600488457264, 'samples': 3998592, 'steps': 20825, 'loss/train': 1.68085116147995} 01/29/2022 15:40:29 - INFO - codeparrot_training - Step 20826: {'lr': 0.00033305057185497444, 'samples': 3998784, 'steps': 20826, 'loss/train': 0.9188005328178406} 01/29/2022 15:40:33 - INFO - codeparrot_training - Step 20827: {'lr': 0.0003330351384696139, 'samples': 3998976, 'steps': 20827, 'loss/train': 1.6390541195869446} 01/29/2022 15:40:37 - INFO - codeparrot_training - Step 20828: {'lr': 0.00033301970472855724, 'samples': 3999168, 'steps': 20828, 'loss/train': 0.9289453625679016} 01/29/2022 15:40:41 - INFO - codeparrot_training - Step 20829: {'lr': 0.0003330042706318705, 'samples': 3999360, 'steps': 20829, 'loss/train': 1.609765112400055} 01/29/2022 15:40:46 - INFO - codeparrot_training - Step 20830: {'lr': 0.00033298883617961984, 'samples': 3999552, 'steps': 20830, 'loss/train': 1.8826496601104736} 01/29/2022 15:40:51 - INFO - codeparrot_training - Step 20831: {'lr': 0.0003329734013718713, 'samples': 3999744, 'steps': 20831, 'loss/train': 1.170198529958725} 01/29/2022 15:40:55 - INFO - codeparrot_training - Step 20832: {'lr': 0.0003329579662086911, 'samples': 3999936, 'steps': 20832, 'loss/train': 1.6639577150344849} 01/29/2022 15:40:59 - INFO - codeparrot_training - Step 20833: {'lr': 0.00033294253069014534, 'samples': 4000128, 'steps': 20833, 'loss/train': 0.9064138233661652} 01/29/2022 15:41:04 - INFO - codeparrot_training - Step 20834: {'lr': 0.0003329270948163001, 'samples': 4000320, 'steps': 20834, 'loss/train': 1.6404411792755127} 01/29/2022 15:41:08 - INFO - codeparrot_training - Step 20835: {'lr': 0.0003329116585872215, 'samples': 4000512, 'steps': 20835, 'loss/train': 1.4690199494361877} 01/29/2022 15:41:15 - INFO - codeparrot_training - Step 20836: {'lr': 0.00033289622200297563, 'samples': 4000704, 'steps': 20836, 'loss/train': 1.3139523267745972} 01/29/2022 15:41:19 - INFO - codeparrot_training - Step 20837: {'lr': 0.0003328807850636287, 'samples': 4000896, 'steps': 20837, 'loss/train': 0.8657792508602142} 01/29/2022 15:41:24 - INFO - codeparrot_training - Step 20838: {'lr': 0.0003328653477692469, 'samples': 4001088, 'steps': 20838, 'loss/train': 1.0660114288330078} 01/29/2022 15:41:28 - INFO - codeparrot_training - Step 20839: {'lr': 0.0003328499101198962, 'samples': 4001280, 'steps': 20839, 'loss/train': 1.6138415336608887} 01/29/2022 15:41:32 - INFO - codeparrot_training - Step 20840: {'lr': 0.0003328344721156427, 'samples': 4001472, 'steps': 20840, 'loss/train': 1.5601034760475159} 01/29/2022 15:41:38 - INFO - codeparrot_training - Step 20841: {'lr': 0.00033281903375655277, 'samples': 4001664, 'steps': 20841, 'loss/train': 2.1535622477531433} 01/29/2022 15:41:42 - INFO - codeparrot_training - Step 20842: {'lr': 0.0003328035950426923, 'samples': 4001856, 'steps': 20842, 'loss/train': 1.790719985961914} 01/29/2022 15:41:46 - INFO - codeparrot_training - Step 20843: {'lr': 0.0003327881559741276, 'samples': 4002048, 'steps': 20843, 'loss/train': 1.3517180979251862} 01/29/2022 15:41:50 - INFO - codeparrot_training - Step 20844: {'lr': 0.00033277271655092467, 'samples': 4002240, 'steps': 20844, 'loss/train': 1.5982841849327087} 01/29/2022 15:41:54 - INFO - codeparrot_training - Step 20845: {'lr': 0.0003327572767731497, 'samples': 4002432, 'steps': 20845, 'loss/train': 2.380273997783661} 01/29/2022 15:42:00 - INFO - codeparrot_training - Step 20846: {'lr': 0.0003327418366408689, 'samples': 4002624, 'steps': 20846, 'loss/train': 1.1196769773960114} 01/29/2022 15:42:04 - INFO - codeparrot_training - Step 20847: {'lr': 0.0003327263961541483, 'samples': 4002816, 'steps': 20847, 'loss/train': 2.0257195830345154} 01/29/2022 15:42:08 - INFO - codeparrot_training - Step 20848: {'lr': 0.0003327109553130541, 'samples': 4003008, 'steps': 20848, 'loss/train': 1.8389074802398682} 01/29/2022 15:42:13 - INFO - codeparrot_training - Step 20849: {'lr': 0.0003326955141176524, 'samples': 4003200, 'steps': 20849, 'loss/train': 1.300293356180191} 01/29/2022 15:42:17 - INFO - codeparrot_training - Step 20850: {'lr': 0.0003326800725680094, 'samples': 4003392, 'steps': 20850, 'loss/train': 1.65634685754776} 01/29/2022 15:42:23 - INFO - codeparrot_training - Step 20851: {'lr': 0.0003326646306641912, 'samples': 4003584, 'steps': 20851, 'loss/train': 1.355427324771881} 01/29/2022 15:42:28 - INFO - codeparrot_training - Step 20852: {'lr': 0.000332649188406264, 'samples': 4003776, 'steps': 20852, 'loss/train': 1.2414889633655548} 01/29/2022 15:42:32 - INFO - codeparrot_training - Step 20853: {'lr': 0.0003326337457942939, 'samples': 4003968, 'steps': 20853, 'loss/train': 1.4909141063690186} 01/29/2022 15:42:36 - INFO - codeparrot_training - Step 20854: {'lr': 0.00033261830282834716, 'samples': 4004160, 'steps': 20854, 'loss/train': 1.6817666888237} 01/29/2022 15:42:40 - INFO - codeparrot_training - Step 20855: {'lr': 0.00033260285950848965, 'samples': 4004352, 'steps': 20855, 'loss/train': 1.6323711276054382} 01/29/2022 15:42:46 - INFO - codeparrot_training - Step 20856: {'lr': 0.0003325874158347879, 'samples': 4004544, 'steps': 20856, 'loss/train': 1.4387262761592865} 01/29/2022 15:42:50 - INFO - codeparrot_training - Step 20857: {'lr': 0.0003325719718073078, 'samples': 4004736, 'steps': 20857, 'loss/train': 1.4511819183826447} 01/29/2022 15:42:54 - INFO - codeparrot_training - Step 20858: {'lr': 0.00033255652742611566, 'samples': 4004928, 'steps': 20858, 'loss/train': 1.6656469702720642} 01/29/2022 15:42:58 - INFO - codeparrot_training - Step 20859: {'lr': 0.0003325410826912775, 'samples': 4005120, 'steps': 20859, 'loss/train': 2.259753942489624} 01/29/2022 15:43:03 - INFO - codeparrot_training - Step 20860: {'lr': 0.0003325256376028595, 'samples': 4005312, 'steps': 20860, 'loss/train': 1.9785827994346619} 01/29/2022 15:43:08 - INFO - codeparrot_training - Step 20861: {'lr': 0.000332510192160928, 'samples': 4005504, 'steps': 20861, 'loss/train': 1.6207834482192993} 01/29/2022 15:43:12 - INFO - codeparrot_training - Step 20862: {'lr': 0.000332494746365549, 'samples': 4005696, 'steps': 20862, 'loss/train': 1.3372573256492615} 01/29/2022 15:43:17 - INFO - codeparrot_training - Step 20863: {'lr': 0.00033247930021678866, 'samples': 4005888, 'steps': 20863, 'loss/train': 2.006789803504944} 01/29/2022 15:43:21 - INFO - codeparrot_training - Step 20864: {'lr': 0.0003324638537147132, 'samples': 4006080, 'steps': 20864, 'loss/train': 2.220345675945282} 01/29/2022 15:43:25 - INFO - codeparrot_training - Step 20865: {'lr': 0.00033244840685938884, 'samples': 4006272, 'steps': 20865, 'loss/train': 1.746385931968689} 01/29/2022 15:43:32 - INFO - codeparrot_training - Step 20866: {'lr': 0.0003324329596508816, 'samples': 4006464, 'steps': 20866, 'loss/train': 2.2771897315979004} 01/29/2022 15:43:37 - INFO - codeparrot_training - Step 20867: {'lr': 0.0003324175120892579, 'samples': 4006656, 'steps': 20867, 'loss/train': 1.8453410267829895} 01/29/2022 15:43:41 - INFO - codeparrot_training - Step 20868: {'lr': 0.00033240206417458354, 'samples': 4006848, 'steps': 20868, 'loss/train': 0.12633991241455078} 01/29/2022 15:43:45 - INFO - codeparrot_training - Step 20869: {'lr': 0.00033238661590692496, 'samples': 4007040, 'steps': 20869, 'loss/train': 1.6960228085517883} 01/29/2022 15:43:49 - INFO - codeparrot_training - Step 20870: {'lr': 0.00033237116728634833, 'samples': 4007232, 'steps': 20870, 'loss/train': 1.1391291618347168} 01/29/2022 15:43:55 - INFO - codeparrot_training - Step 20871: {'lr': 0.0003323557183129197, 'samples': 4007424, 'steps': 20871, 'loss/train': 1.189409226179123} 01/29/2022 15:43:59 - INFO - codeparrot_training - Step 20872: {'lr': 0.0003323402689867054, 'samples': 4007616, 'steps': 20872, 'loss/train': 1.2369532585144043} 01/29/2022 15:44:03 - INFO - codeparrot_training - Step 20873: {'lr': 0.0003323248193077715, 'samples': 4007808, 'steps': 20873, 'loss/train': 1.2295579612255096} 01/29/2022 15:44:07 - INFO - codeparrot_training - Step 20874: {'lr': 0.0003323093692761842, 'samples': 4008000, 'steps': 20874, 'loss/train': 1.642782211303711} 01/29/2022 15:44:12 - INFO - codeparrot_training - Step 20875: {'lr': 0.00033229391889200974, 'samples': 4008192, 'steps': 20875, 'loss/train': 1.653010368347168} 01/29/2022 15:44:17 - INFO - codeparrot_training - Step 20876: {'lr': 0.00033227846815531424, 'samples': 4008384, 'steps': 20876, 'loss/train': 1.7453659772872925} 01/29/2022 15:44:22 - INFO - codeparrot_training - Step 20877: {'lr': 0.0003322630170661639, 'samples': 4008576, 'steps': 20877, 'loss/train': 2.629130244255066} 01/29/2022 15:44:26 - INFO - codeparrot_training - Step 20878: {'lr': 0.0003322475656246249, 'samples': 4008768, 'steps': 20878, 'loss/train': 0.9745735824108124} 01/29/2022 15:44:30 - INFO - codeparrot_training - Step 20879: {'lr': 0.0003322321138307635, 'samples': 4008960, 'steps': 20879, 'loss/train': 1.098114252090454} 01/29/2022 15:44:34 - INFO - codeparrot_training - Step 20880: {'lr': 0.0003322166616846458, 'samples': 4009152, 'steps': 20880, 'loss/train': 1.1067626774311066} 01/29/2022 15:44:41 - INFO - codeparrot_training - Step 20881: {'lr': 0.0003322012091863381, 'samples': 4009344, 'steps': 20881, 'loss/train': 0.44870536029338837} 01/29/2022 15:44:46 - INFO - codeparrot_training - Step 20882: {'lr': 0.0003321857563359064, 'samples': 4009536, 'steps': 20882, 'loss/train': 1.3485171496868134} 01/29/2022 15:44:50 - INFO - codeparrot_training - Step 20883: {'lr': 0.00033217030313341704, 'samples': 4009728, 'steps': 20883, 'loss/train': 1.390221118927002} 01/29/2022 15:44:54 - INFO - codeparrot_training - Step 20884: {'lr': 0.00033215484957893626, 'samples': 4009920, 'steps': 20884, 'loss/train': 1.8087111711502075} 01/29/2022 15:44:59 - INFO - codeparrot_training - Step 20885: {'lr': 0.0003321393956725302, 'samples': 4010112, 'steps': 20885, 'loss/train': 1.6779475808143616} 01/29/2022 15:45:04 - INFO - codeparrot_training - Step 20886: {'lr': 0.00033212394141426493, 'samples': 4010304, 'steps': 20886, 'loss/train': 0.13270994648337364} 01/29/2022 15:45:08 - INFO - codeparrot_training - Step 20887: {'lr': 0.00033210848680420693, 'samples': 4010496, 'steps': 20887, 'loss/train': 1.9635990858078003} 01/29/2022 15:45:12 - INFO - codeparrot_training - Step 20888: {'lr': 0.00033209303184242214, 'samples': 4010688, 'steps': 20888, 'loss/train': 1.9148350954055786} 01/29/2022 15:45:16 - INFO - codeparrot_training - Step 20889: {'lr': 0.0003320775765289769, 'samples': 4010880, 'steps': 20889, 'loss/train': 2.0560296177864075} 01/29/2022 15:45:24 - INFO - codeparrot_training - Step 20890: {'lr': 0.0003320621208639374, 'samples': 4011072, 'steps': 20890, 'loss/train': 0.81401726603508} 01/29/2022 15:45:28 - INFO - codeparrot_training - Step 20891: {'lr': 0.00033204666484736977, 'samples': 4011264, 'steps': 20891, 'loss/train': 1.770305871963501} 01/29/2022 15:45:32 - INFO - codeparrot_training - Step 20892: {'lr': 0.0003320312084793404, 'samples': 4011456, 'steps': 20892, 'loss/train': 2.50945508480072} 01/29/2022 15:45:36 - INFO - codeparrot_training - Step 20893: {'lr': 0.0003320157517599153, 'samples': 4011648, 'steps': 20893, 'loss/train': 1.8209365010261536} 01/29/2022 15:45:41 - INFO - codeparrot_training - Step 20894: {'lr': 0.00033200029468916076, 'samples': 4011840, 'steps': 20894, 'loss/train': 1.9046174883842468} 01/29/2022 15:45:46 - INFO - codeparrot_training - Step 20895: {'lr': 0.00033198483726714294, 'samples': 4012032, 'steps': 20895, 'loss/train': 2.259221076965332} 01/29/2022 15:45:50 - INFO - codeparrot_training - Step 20896: {'lr': 0.00033196937949392824, 'samples': 4012224, 'steps': 20896, 'loss/train': 1.6157573461532593} 01/29/2022 15:45:54 - INFO - codeparrot_training - Step 20897: {'lr': 0.00033195392136958264, 'samples': 4012416, 'steps': 20897, 'loss/train': 1.6267524361610413} 01/29/2022 15:45:59 - INFO - codeparrot_training - Step 20898: {'lr': 0.00033193846289417253, 'samples': 4012608, 'steps': 20898, 'loss/train': 1.2953223288059235} 01/29/2022 15:46:03 - INFO - codeparrot_training - Step 20899: {'lr': 0.00033192300406776406, 'samples': 4012800, 'steps': 20899, 'loss/train': 0.9579076766967773} 01/29/2022 15:46:08 - INFO - codeparrot_training - Step 20900: {'lr': 0.0003319075448904234, 'samples': 4012992, 'steps': 20900, 'loss/train': 1.4023267328739166} 01/29/2022 15:46:12 - INFO - codeparrot_training - Step 20901: {'lr': 0.00033189208536221683, 'samples': 4013184, 'steps': 20901, 'loss/train': 1.55168616771698} 01/29/2022 15:46:17 - INFO - codeparrot_training - Step 20902: {'lr': 0.00033187662548321063, 'samples': 4013376, 'steps': 20902, 'loss/train': 2.0719116926193237} 01/29/2022 15:46:21 - INFO - codeparrot_training - Step 20903: {'lr': 0.00033186116525347093, 'samples': 4013568, 'steps': 20903, 'loss/train': 2.0495585203170776} 01/29/2022 15:46:25 - INFO - codeparrot_training - Step 20904: {'lr': 0.00033184570467306403, 'samples': 4013760, 'steps': 20904, 'loss/train': 1.242009848356247} 01/29/2022 15:46:30 - INFO - codeparrot_training - Step 20905: {'lr': 0.0003318302437420561, 'samples': 4013952, 'steps': 20905, 'loss/train': 1.8586950302124023} 01/29/2022 15:46:35 - INFO - codeparrot_training - Step 20906: {'lr': 0.0003318147824605133, 'samples': 4014144, 'steps': 20906, 'loss/train': 1.1040890514850616} 01/29/2022 15:46:39 - INFO - codeparrot_training - Step 20907: {'lr': 0.0003317993208285021, 'samples': 4014336, 'steps': 20907, 'loss/train': 1.9563013315200806} 01/29/2022 15:46:43 - INFO - codeparrot_training - Step 20908: {'lr': 0.0003317838588460884, 'samples': 4014528, 'steps': 20908, 'loss/train': 0.9322928488254547} 01/29/2022 15:46:48 - INFO - codeparrot_training - Step 20909: {'lr': 0.0003317683965133388, 'samples': 4014720, 'steps': 20909, 'loss/train': 1.731247365474701} 01/29/2022 15:46:55 - INFO - codeparrot_training - Step 20910: {'lr': 0.0003317529338303192, 'samples': 4014912, 'steps': 20910, 'loss/train': 2.0266900062561035} 01/29/2022 15:47:00 - INFO - codeparrot_training - Step 20911: {'lr': 0.00033173747079709616, 'samples': 4015104, 'steps': 20911, 'loss/train': 1.338807463645935} 01/29/2022 15:47:04 - INFO - codeparrot_training - Step 20912: {'lr': 0.0003317220074137356, 'samples': 4015296, 'steps': 20912, 'loss/train': 1.6525583267211914} 01/29/2022 15:47:08 - INFO - codeparrot_training - Step 20913: {'lr': 0.000331706543680304, 'samples': 4015488, 'steps': 20913, 'loss/train': 2.292983829975128} 01/29/2022 15:47:12 - INFO - codeparrot_training - Step 20914: {'lr': 0.0003316910795968675, 'samples': 4015680, 'steps': 20914, 'loss/train': 1.5925884246826172} 01/29/2022 15:47:18 - INFO - codeparrot_training - Step 20915: {'lr': 0.00033167561516349233, 'samples': 4015872, 'steps': 20915, 'loss/train': 1.1877743303775787} 01/29/2022 15:47:22 - INFO - codeparrot_training - Step 20916: {'lr': 0.0003316601503802448, 'samples': 4016064, 'steps': 20916, 'loss/train': 0.8577270805835724} 01/29/2022 15:47:26 - INFO - codeparrot_training - Step 20917: {'lr': 0.00033164468524719105, 'samples': 4016256, 'steps': 20917, 'loss/train': 1.8879602551460266} 01/29/2022 15:47:30 - INFO - codeparrot_training - Step 20918: {'lr': 0.00033162921976439744, 'samples': 4016448, 'steps': 20918, 'loss/train': 1.1660050749778748} 01/29/2022 15:47:35 - INFO - codeparrot_training - Step 20919: {'lr': 0.00033161375393193015, 'samples': 4016640, 'steps': 20919, 'loss/train': 1.6761555075645447} 01/29/2022 15:47:40 - INFO - codeparrot_training - Step 20920: {'lr': 0.00033159828774985547, 'samples': 4016832, 'steps': 20920, 'loss/train': 1.7591866850852966} 01/29/2022 15:47:44 - INFO - codeparrot_training - Step 20921: {'lr': 0.0003315828212182396, 'samples': 4017024, 'steps': 20921, 'loss/train': 1.0612701773643494} 01/29/2022 15:47:49 - INFO - codeparrot_training - Step 20922: {'lr': 0.00033156735433714893, 'samples': 4017216, 'steps': 20922, 'loss/train': 2.019614338874817} 01/29/2022 15:47:53 - INFO - codeparrot_training - Step 20923: {'lr': 0.00033155188710664945, 'samples': 4017408, 'steps': 20923, 'loss/train': 1.3282370567321777} 01/29/2022 15:47:57 - INFO - codeparrot_training - Step 20924: {'lr': 0.00033153641952680767, 'samples': 4017600, 'steps': 20924, 'loss/train': 1.7200194597244263} 01/29/2022 15:48:04 - INFO - codeparrot_training - Step 20925: {'lr': 0.0003315209515976898, 'samples': 4017792, 'steps': 20925, 'loss/train': 1.7955068349838257} 01/29/2022 15:48:08 - INFO - codeparrot_training - Step 20926: {'lr': 0.000331505483319362, 'samples': 4017984, 'steps': 20926, 'loss/train': 0.5948418974876404} 01/29/2022 15:48:13 - INFO - codeparrot_training - Step 20927: {'lr': 0.0003314900146918906, 'samples': 4018176, 'steps': 20927, 'loss/train': 1.4099409878253937} 01/29/2022 15:48:17 - INFO - codeparrot_training - Step 20928: {'lr': 0.0003314745457153419, 'samples': 4018368, 'steps': 20928, 'loss/train': 0.8801292479038239} 01/29/2022 15:48:21 - INFO - codeparrot_training - Step 20929: {'lr': 0.00033145907638978207, 'samples': 4018560, 'steps': 20929, 'loss/train': 2.0382365584373474} 01/29/2022 15:48:26 - INFO - codeparrot_training - Step 20930: {'lr': 0.00033144360671527747, 'samples': 4018752, 'steps': 20930, 'loss/train': 1.2808323204517365} 01/29/2022 15:48:31 - INFO - codeparrot_training - Step 20931: {'lr': 0.0003314281366918943, 'samples': 4018944, 'steps': 20931, 'loss/train': 0.9074069559574127} 01/29/2022 15:48:35 - INFO - codeparrot_training - Step 20932: {'lr': 0.0003314126663196988, 'samples': 4019136, 'steps': 20932, 'loss/train': 0.07225823029875755} 01/29/2022 15:48:39 - INFO - codeparrot_training - Step 20933: {'lr': 0.0003313971955987573, 'samples': 4019328, 'steps': 20933, 'loss/train': 1.844655454158783} 01/29/2022 15:48:45 - INFO - codeparrot_training - Step 20934: {'lr': 0.0003313817245291361, 'samples': 4019520, 'steps': 20934, 'loss/train': 1.2185944318771362} 01/29/2022 15:48:49 - INFO - codeparrot_training - Step 20935: {'lr': 0.0003313662531109014, 'samples': 4019712, 'steps': 20935, 'loss/train': 1.335074543952942} 01/29/2022 15:48:53 - INFO - codeparrot_training - Step 20936: {'lr': 0.00033135078134411956, 'samples': 4019904, 'steps': 20936, 'loss/train': 1.766428828239441} 01/29/2022 15:48:57 - INFO - codeparrot_training - Step 20937: {'lr': 0.0003313353092288568, 'samples': 4020096, 'steps': 20937, 'loss/train': 1.5590538382530212} 01/29/2022 15:49:02 - INFO - codeparrot_training - Step 20938: {'lr': 0.00033131983676517934, 'samples': 4020288, 'steps': 20938, 'loss/train': 2.089913070201874} 01/29/2022 15:49:06 - INFO - codeparrot_training - Step 20939: {'lr': 0.0003313043639531536, 'samples': 4020480, 'steps': 20939, 'loss/train': 0.09191923029720783} 01/29/2022 15:49:13 - INFO - codeparrot_training - Step 20940: {'lr': 0.00033128889079284574, 'samples': 4020672, 'steps': 20940, 'loss/train': 1.6117556691169739} 01/29/2022 15:49:17 - INFO - codeparrot_training - Step 20941: {'lr': 0.0003312734172843221, 'samples': 4020864, 'steps': 20941, 'loss/train': 1.8441506624221802} 01/29/2022 15:49:22 - INFO - codeparrot_training - Step 20942: {'lr': 0.0003312579434276489, 'samples': 4021056, 'steps': 20942, 'loss/train': 1.4550850689411163} 01/29/2022 15:49:26 - INFO - codeparrot_training - Step 20943: {'lr': 0.0003312424692228925, 'samples': 4021248, 'steps': 20943, 'loss/train': 0.6181416660547256} 01/29/2022 15:49:30 - INFO - codeparrot_training - Step 20944: {'lr': 0.0003312269946701191, 'samples': 4021440, 'steps': 20944, 'loss/train': 1.5689256191253662} 01/29/2022 15:49:36 - INFO - codeparrot_training - Step 20945: {'lr': 0.0003312115197693951, 'samples': 4021632, 'steps': 20945, 'loss/train': 1.5066536664962769} 01/29/2022 15:49:40 - INFO - codeparrot_training - Step 20946: {'lr': 0.00033119604452078676, 'samples': 4021824, 'steps': 20946, 'loss/train': 2.1621996760368347} 01/29/2022 15:49:44 - INFO - codeparrot_training - Step 20947: {'lr': 0.00033118056892436035, 'samples': 4022016, 'steps': 20947, 'loss/train': 2.138200521469116} 01/29/2022 15:49:49 - INFO - codeparrot_training - Step 20948: {'lr': 0.00033116509298018217, 'samples': 4022208, 'steps': 20948, 'loss/train': 2.430663228034973} 01/29/2022 15:49:56 - INFO - codeparrot_training - Step 20949: {'lr': 0.00033114961668831845, 'samples': 4022400, 'steps': 20949, 'loss/train': 1.9784578084945679} 01/29/2022 15:50:00 - INFO - codeparrot_training - Step 20950: {'lr': 0.00033113414004883556, 'samples': 4022592, 'steps': 20950, 'loss/train': 1.4494662880897522} 01/29/2022 15:50:04 - INFO - codeparrot_training - Step 20951: {'lr': 0.0003311186630617998, 'samples': 4022784, 'steps': 20951, 'loss/train': 1.5792240500450134} 01/29/2022 15:50:09 - INFO - codeparrot_training - Step 20952: {'lr': 0.00033110318572727743, 'samples': 4022976, 'steps': 20952, 'loss/train': 1.3941369652748108} 01/29/2022 15:50:13 - INFO - codeparrot_training - Step 20953: {'lr': 0.0003310877080453348, 'samples': 4023168, 'steps': 20953, 'loss/train': 1.3077583014965057} 01/29/2022 15:50:18 - INFO - codeparrot_training - Step 20954: {'lr': 0.00033107223001603814, 'samples': 4023360, 'steps': 20954, 'loss/train': 1.3541394174098969} 01/29/2022 15:50:23 - INFO - codeparrot_training - Step 20955: {'lr': 0.00033105675163945373, 'samples': 4023552, 'steps': 20955, 'loss/train': 2.1653298139572144} 01/29/2022 15:50:27 - INFO - codeparrot_training - Step 20956: {'lr': 0.000331041272915648, 'samples': 4023744, 'steps': 20956, 'loss/train': 1.6444076299667358} 01/29/2022 15:50:31 - INFO - codeparrot_training - Step 20957: {'lr': 0.00033102579384468723, 'samples': 4023936, 'steps': 20957, 'loss/train': 2.003636956214905} 01/29/2022 15:50:35 - INFO - codeparrot_training - Step 20958: {'lr': 0.0003310103144266376, 'samples': 4024128, 'steps': 20958, 'loss/train': 1.8218069672584534} 01/29/2022 15:50:41 - INFO - codeparrot_training - Step 20959: {'lr': 0.00033099483466156554, 'samples': 4024320, 'steps': 20959, 'loss/train': 1.6950541734695435} 01/29/2022 15:50:45 - INFO - codeparrot_training - Step 20960: {'lr': 0.00033097935454953737, 'samples': 4024512, 'steps': 20960, 'loss/train': 1.8212515711784363} 01/29/2022 15:50:49 - INFO - codeparrot_training - Step 20961: {'lr': 0.00033096387409061937, 'samples': 4024704, 'steps': 20961, 'loss/train': 0.4681478440761566} 01/29/2022 15:50:53 - INFO - codeparrot_training - Step 20962: {'lr': 0.00033094839328487777, 'samples': 4024896, 'steps': 20962, 'loss/train': 1.4663191437721252} 01/29/2022 15:50:58 - INFO - codeparrot_training - Step 20963: {'lr': 0.000330932912132379, 'samples': 4025088, 'steps': 20963, 'loss/train': 1.6601682901382446} 01/29/2022 15:51:06 - INFO - codeparrot_training - Step 20964: {'lr': 0.0003309174306331893, 'samples': 4025280, 'steps': 20964, 'loss/train': 1.7584757208824158} 01/29/2022 15:51:10 - INFO - codeparrot_training - Step 20965: {'lr': 0.00033090194878737504, 'samples': 4025472, 'steps': 20965, 'loss/train': 1.9127084612846375} 01/29/2022 15:51:14 - INFO - codeparrot_training - Step 20966: {'lr': 0.0003308864665950025, 'samples': 4025664, 'steps': 20966, 'loss/train': 1.724746584892273} 01/29/2022 15:51:18 - INFO - codeparrot_training - Step 20967: {'lr': 0.0003308709840561381, 'samples': 4025856, 'steps': 20967, 'loss/train': 7.433789491653442} 01/29/2022 15:51:23 - INFO - codeparrot_training - Step 20968: {'lr': 0.00033085550117084795, 'samples': 4026048, 'steps': 20968, 'loss/train': 1.8165687918663025} 01/29/2022 15:51:27 - INFO - codeparrot_training - Step 20969: {'lr': 0.0003308400179391986, 'samples': 4026240, 'steps': 20969, 'loss/train': 1.6678531765937805} 01/29/2022 15:51:32 - INFO - codeparrot_training - Step 20970: {'lr': 0.00033082453436125627, 'samples': 4026432, 'steps': 20970, 'loss/train': 2.895739495754242} 01/29/2022 15:51:36 - INFO - codeparrot_training - Step 20971: {'lr': 0.00033080905043708734, 'samples': 4026624, 'steps': 20971, 'loss/train': 1.7630876898765564} 01/29/2022 15:51:41 - INFO - codeparrot_training - Step 20972: {'lr': 0.000330793566166758, 'samples': 4026816, 'steps': 20972, 'loss/train': 1.0898487567901611} 01/29/2022 15:51:45 - INFO - codeparrot_training - Step 20973: {'lr': 0.00033077808155033473, 'samples': 4027008, 'steps': 20973, 'loss/train': 1.5588998794555664} 01/29/2022 15:51:49 - INFO - codeparrot_training - Step 20974: {'lr': 0.0003307625965878838, 'samples': 4027200, 'steps': 20974, 'loss/train': 1.3078726530075073} 01/29/2022 15:51:55 - INFO - codeparrot_training - Step 20975: {'lr': 0.00033074711127947153, 'samples': 4027392, 'steps': 20975, 'loss/train': 1.5970024466514587} 01/29/2022 15:51:59 - INFO - codeparrot_training - Step 20976: {'lr': 0.0003307316256251644, 'samples': 4027584, 'steps': 20976, 'loss/train': 1.586527168750763} 01/29/2022 15:52:03 - INFO - codeparrot_training - Step 20977: {'lr': 0.0003307161396250285, 'samples': 4027776, 'steps': 20977, 'loss/train': 1.7016844153404236} 01/29/2022 15:52:07 - INFO - codeparrot_training - Step 20978: {'lr': 0.00033070065327913035, 'samples': 4027968, 'steps': 20978, 'loss/train': 1.0239018201828003} 01/29/2022 15:52:12 - INFO - codeparrot_training - Step 20979: {'lr': 0.00033068516658753624, 'samples': 4028160, 'steps': 20979, 'loss/train': 0.042930071242153645} 01/29/2022 15:52:19 - INFO - codeparrot_training - Step 20980: {'lr': 0.00033066967955031236, 'samples': 4028352, 'steps': 20980, 'loss/train': 2.152346670627594} 01/29/2022 15:52:23 - INFO - codeparrot_training - Step 20981: {'lr': 0.0003306541921675253, 'samples': 4028544, 'steps': 20981, 'loss/train': 2.4123982787132263} 01/29/2022 15:52:27 - INFO - codeparrot_training - Step 20982: {'lr': 0.0003306387044392413, 'samples': 4028736, 'steps': 20982, 'loss/train': 1.051643192768097} 01/29/2022 15:52:32 - INFO - codeparrot_training - Step 20983: {'lr': 0.0003306232163655267, 'samples': 4028928, 'steps': 20983, 'loss/train': 2.022999107837677} 01/29/2022 15:52:36 - INFO - codeparrot_training - Step 20984: {'lr': 0.00033060772794644776, 'samples': 4029120, 'steps': 20984, 'loss/train': 1.7808045744895935} 01/29/2022 15:52:41 - INFO - codeparrot_training - Step 20985: {'lr': 0.000330592239182071, 'samples': 4029312, 'steps': 20985, 'loss/train': 1.6138604879379272} 01/29/2022 15:52:45 - INFO - codeparrot_training - Step 20986: {'lr': 0.0003305767500724626, 'samples': 4029504, 'steps': 20986, 'loss/train': 1.062985360622406} 01/29/2022 15:52:50 - INFO - codeparrot_training - Step 20987: {'lr': 0.00033056126061768905, 'samples': 4029696, 'steps': 20987, 'loss/train': 2.1511539816856384} 01/29/2022 15:52:54 - INFO - codeparrot_training - Step 20988: {'lr': 0.00033054577081781654, 'samples': 4029888, 'steps': 20988, 'loss/train': 1.0105661451816559} 01/29/2022 15:52:59 - INFO - codeparrot_training - Step 20989: {'lr': 0.00033053028067291166, 'samples': 4030080, 'steps': 20989, 'loss/train': 1.2675653994083405} 01/29/2022 15:53:03 - INFO - codeparrot_training - Step 20990: {'lr': 0.00033051479018304054, 'samples': 4030272, 'steps': 20990, 'loss/train': 1.857532560825348} 01/29/2022 15:53:08 - INFO - codeparrot_training - Step 20991: {'lr': 0.0003304992993482697, 'samples': 4030464, 'steps': 20991, 'loss/train': 1.9148340225219727} 01/29/2022 15:53:12 - INFO - codeparrot_training - Step 20992: {'lr': 0.0003304838081686653, 'samples': 4030656, 'steps': 20992, 'loss/train': 1.7590672373771667} 01/29/2022 15:53:16 - INFO - codeparrot_training - Step 20993: {'lr': 0.0003304683166442939, 'samples': 4030848, 'steps': 20993, 'loss/train': 1.673109233379364} 01/29/2022 15:53:23 - INFO - codeparrot_training - Step 20994: {'lr': 0.0003304528247752218, 'samples': 4031040, 'steps': 20994, 'loss/train': 1.6011019349098206} 01/29/2022 15:53:28 - INFO - codeparrot_training - Step 20995: {'lr': 0.0003304373325615153, 'samples': 4031232, 'steps': 20995, 'loss/train': 1.4661990702152252} 01/29/2022 15:53:32 - INFO - codeparrot_training - Step 20996: {'lr': 0.00033042184000324086, 'samples': 4031424, 'steps': 20996, 'loss/train': 2.0293652415275574} 01/29/2022 15:53:36 - INFO - codeparrot_training - Step 20997: {'lr': 0.00033040634710046474, 'samples': 4031616, 'steps': 20997, 'loss/train': 0.451700359582901} 01/29/2022 15:53:40 - INFO - codeparrot_training - Step 20998: {'lr': 0.0003303908538532534, 'samples': 4031808, 'steps': 20998, 'loss/train': 1.655495524406433} 01/29/2022 15:53:45 - INFO - codeparrot_training - Step 20999: {'lr': 0.00033037536026167313, 'samples': 4032000, 'steps': 20999, 'loss/train': 0.9433300495147705} 01/29/2022 15:53:50 - INFO - codeparrot_training - Step 21000: {'lr': 0.0003303598663257904, 'samples': 4032192, 'steps': 21000, 'loss/train': 1.4179742932319641} 01/29/2022 15:53:54 - INFO - codeparrot_training - Step 21001: {'lr': 0.00033034437204567145, 'samples': 4032384, 'steps': 21001, 'loss/train': 1.6752455234527588} 01/29/2022 15:53:58 - INFO - codeparrot_training - Step 21002: {'lr': 0.00033032887742138285, 'samples': 4032576, 'steps': 21002, 'loss/train': 1.6102598905563354} 01/29/2022 15:54:03 - INFO - codeparrot_training - Step 21003: {'lr': 0.0003303133824529907, 'samples': 4032768, 'steps': 21003, 'loss/train': 1.84110689163208} 01/29/2022 15:54:08 - INFO - codeparrot_training - Step 21004: {'lr': 0.00033029788714056165, 'samples': 4032960, 'steps': 21004, 'loss/train': 1.782547116279602} 01/29/2022 15:54:12 - INFO - codeparrot_training - Step 21005: {'lr': 0.0003302823914841618, 'samples': 4033152, 'steps': 21005, 'loss/train': 1.1592061221599579} 01/29/2022 15:54:16 - INFO - codeparrot_training - Step 21006: {'lr': 0.00033026689548385776, 'samples': 4033344, 'steps': 21006, 'loss/train': 0.9627821445465088} 01/29/2022 15:54:21 - INFO - codeparrot_training - Step 21007: {'lr': 0.00033025139913971585, 'samples': 4033536, 'steps': 21007, 'loss/train': 1.9324305653572083} 01/29/2022 15:54:25 - INFO - codeparrot_training - Step 21008: {'lr': 0.00033023590245180237, 'samples': 4033728, 'steps': 21008, 'loss/train': 1.9544665217399597} 01/29/2022 15:54:32 - INFO - codeparrot_training - Step 21009: {'lr': 0.00033022040542018385, 'samples': 4033920, 'steps': 21009, 'loss/train': 1.8153612613677979} 01/29/2022 15:54:36 - INFO - codeparrot_training - Step 21010: {'lr': 0.0003302049080449265, 'samples': 4034112, 'steps': 21010, 'loss/train': 1.9157375693321228} 01/29/2022 15:54:40 - INFO - codeparrot_training - Step 21011: {'lr': 0.0003301894103260968, 'samples': 4034304, 'steps': 21011, 'loss/train': 1.3599590063095093} 01/29/2022 15:54:45 - INFO - codeparrot_training - Step 21012: {'lr': 0.0003301739122637611, 'samples': 4034496, 'steps': 21012, 'loss/train': 2.1889089345932007} 01/29/2022 15:54:49 - INFO - codeparrot_training - Step 21013: {'lr': 0.00033015841385798596, 'samples': 4034688, 'steps': 21013, 'loss/train': 0.7260135412216187} 01/29/2022 15:54:54 - INFO - codeparrot_training - Step 21014: {'lr': 0.00033014291510883746, 'samples': 4034880, 'steps': 21014, 'loss/train': 2.4038604497909546} 01/29/2022 15:54:58 - INFO - codeparrot_training - Step 21015: {'lr': 0.0003301274160163823, 'samples': 4035072, 'steps': 21015, 'loss/train': 1.9903696775436401} 01/29/2022 15:55:03 - INFO - codeparrot_training - Step 21016: {'lr': 0.00033011191658068663, 'samples': 4035264, 'steps': 21016, 'loss/train': 0.9203288555145264} 01/29/2022 15:55:07 - INFO - codeparrot_training - Step 21017: {'lr': 0.00033009641680181696, 'samples': 4035456, 'steps': 21017, 'loss/train': 1.344110906124115} 01/29/2022 15:55:11 - INFO - codeparrot_training - Step 21018: {'lr': 0.00033008091667983974, 'samples': 4035648, 'steps': 21018, 'loss/train': 1.456246554851532} 01/29/2022 15:55:18 - INFO - codeparrot_training - Step 21019: {'lr': 0.0003300654162148213, 'samples': 4035840, 'steps': 21019, 'loss/train': 1.4409200549125671} 01/29/2022 15:55:22 - INFO - codeparrot_training - Step 21020: {'lr': 0.00033004991540682793, 'samples': 4036032, 'steps': 21020, 'loss/train': 1.7648025155067444} 01/29/2022 15:55:27 - INFO - codeparrot_training - Step 21021: {'lr': 0.00033003441425592627, 'samples': 4036224, 'steps': 21021, 'loss/train': 1.691142976284027} 01/29/2022 15:55:31 - INFO - codeparrot_training - Step 21022: {'lr': 0.00033001891276218247, 'samples': 4036416, 'steps': 21022, 'loss/train': 1.2991627156734467} 01/29/2022 15:55:35 - INFO - codeparrot_training - Step 21023: {'lr': 0.0003300034109256632, 'samples': 4036608, 'steps': 21023, 'loss/train': 2.11356782913208} 01/29/2022 15:55:41 - INFO - codeparrot_training - Step 21024: {'lr': 0.00032998790874643456, 'samples': 4036800, 'steps': 21024, 'loss/train': 1.1612597107887268} 01/29/2022 15:55:45 - INFO - codeparrot_training - Step 21025: {'lr': 0.00032997240622456326, 'samples': 4036992, 'steps': 21025, 'loss/train': 1.9228363037109375} 01/29/2022 15:55:49 - INFO - codeparrot_training - Step 21026: {'lr': 0.0003299569033601155, 'samples': 4037184, 'steps': 21026, 'loss/train': 1.7005038857460022} 01/29/2022 15:55:53 - INFO - codeparrot_training - Step 21027: {'lr': 0.0003299414001531578, 'samples': 4037376, 'steps': 21027, 'loss/train': 1.572537124156952} 01/29/2022 15:55:59 - INFO - codeparrot_training - Step 21028: {'lr': 0.0003299258966037565, 'samples': 4037568, 'steps': 21028, 'loss/train': 1.7763156294822693} 01/29/2022 15:56:03 - INFO - codeparrot_training - Step 21029: {'lr': 0.000329910392711978, 'samples': 4037760, 'steps': 21029, 'loss/train': 1.8178262114524841} 01/29/2022 15:56:07 - INFO - codeparrot_training - Step 21030: {'lr': 0.0003298948884778887, 'samples': 4037952, 'steps': 21030, 'loss/train': 1.903870403766632} 01/29/2022 15:56:11 - INFO - codeparrot_training - Step 21031: {'lr': 0.00032987938390155523, 'samples': 4038144, 'steps': 21031, 'loss/train': 1.768293857574463} 01/29/2022 15:56:16 - INFO - codeparrot_training - Step 21032: {'lr': 0.00032986387898304375, 'samples': 4038336, 'steps': 21032, 'loss/train': 1.8135986924171448} 01/29/2022 15:56:21 - INFO - codeparrot_training - Step 21033: {'lr': 0.00032984837372242084, 'samples': 4038528, 'steps': 21033, 'loss/train': 1.570193588733673} 01/29/2022 15:56:25 - INFO - codeparrot_training - Step 21034: {'lr': 0.0003298328681197528, 'samples': 4038720, 'steps': 21034, 'loss/train': 0.8292525708675385} 01/29/2022 15:56:29 - INFO - codeparrot_training - Step 21035: {'lr': 0.00032981736217510603, 'samples': 4038912, 'steps': 21035, 'loss/train': 2.1203123331069946} 01/29/2022 15:56:34 - INFO - codeparrot_training - Step 21036: {'lr': 0.0003298018558885471, 'samples': 4039104, 'steps': 21036, 'loss/train': 1.0516285300254822} 01/29/2022 15:56:38 - INFO - codeparrot_training - Step 21037: {'lr': 0.0003297863492601424, 'samples': 4039296, 'steps': 21037, 'loss/train': 1.5502071976661682} 01/29/2022 15:56:45 - INFO - codeparrot_training - Step 21038: {'lr': 0.0003297708422899582, 'samples': 4039488, 'steps': 21038, 'loss/train': 1.8705832958221436} 01/29/2022 15:56:49 - INFO - codeparrot_training - Step 21039: {'lr': 0.0003297553349780612, 'samples': 4039680, 'steps': 21039, 'loss/train': 0.046587067656219006} 01/29/2022 15:56:53 - INFO - codeparrot_training - Step 21040: {'lr': 0.0003297398273245175, 'samples': 4039872, 'steps': 21040, 'loss/train': 1.468140721321106} 01/29/2022 15:56:58 - INFO - codeparrot_training - Step 21041: {'lr': 0.0003297243193293938, 'samples': 4040064, 'steps': 21041, 'loss/train': 0.45200005173683167} 01/29/2022 15:57:02 - INFO - codeparrot_training - Step 21042: {'lr': 0.0003297088109927564, 'samples': 4040256, 'steps': 21042, 'loss/train': 1.0230132937431335} 01/29/2022 15:57:07 - INFO - codeparrot_training - Step 21043: {'lr': 0.00032969330231467177, 'samples': 4040448, 'steps': 21043, 'loss/train': 1.3176935613155365} 01/29/2022 15:57:11 - INFO - codeparrot_training - Step 21044: {'lr': 0.0003296777932952064, 'samples': 4040640, 'steps': 21044, 'loss/train': 0.6215450316667557} 01/29/2022 15:57:16 - INFO - codeparrot_training - Step 21045: {'lr': 0.0003296622839344265, 'samples': 4040832, 'steps': 21045, 'loss/train': 1.809429109096527} 01/29/2022 15:57:20 - INFO - codeparrot_training - Step 21046: {'lr': 0.00032964677423239885, 'samples': 4041024, 'steps': 21046, 'loss/train': 2.1652779579162598} 01/29/2022 15:57:24 - INFO - codeparrot_training - Step 21047: {'lr': 0.0003296312641891896, 'samples': 4041216, 'steps': 21047, 'loss/train': 1.0849498808383942} 01/29/2022 15:57:29 - INFO - codeparrot_training - Step 21048: {'lr': 0.0003296157538048654, 'samples': 4041408, 'steps': 21048, 'loss/train': 1.2680024206638336} 01/29/2022 15:57:34 - INFO - codeparrot_training - Step 21049: {'lr': 0.0003296002430794925, 'samples': 4041600, 'steps': 21049, 'loss/train': 1.3498704135417938} 01/29/2022 15:57:38 - INFO - codeparrot_training - Step 21050: {'lr': 0.00032958473201313745, 'samples': 4041792, 'steps': 21050, 'loss/train': 1.9133974313735962} 01/29/2022 15:57:42 - INFO - codeparrot_training - Step 21051: {'lr': 0.0003295692206058667, 'samples': 4041984, 'steps': 21051, 'loss/train': 1.0260525941848755} 01/29/2022 15:57:46 - INFO - codeparrot_training - Step 21052: {'lr': 0.00032955370885774665, 'samples': 4042176, 'steps': 21052, 'loss/train': 1.8365947008132935} 01/29/2022 15:57:54 - INFO - codeparrot_training - Step 21053: {'lr': 0.0003295381967688438, 'samples': 4042368, 'steps': 21053, 'loss/train': 1.5538026094436646} 01/29/2022 15:57:58 - INFO - codeparrot_training - Step 21054: {'lr': 0.0003295226843392245, 'samples': 4042560, 'steps': 21054, 'loss/train': 0.9739428162574768} 01/29/2022 15:58:02 - INFO - codeparrot_training - Step 21055: {'lr': 0.0003295071715689554, 'samples': 4042752, 'steps': 21055, 'loss/train': 0.6956293433904648} 01/29/2022 15:58:06 - INFO - codeparrot_training - Step 21056: {'lr': 0.0003294916584581027, 'samples': 4042944, 'steps': 21056, 'loss/train': 1.643361210823059} 01/29/2022 15:58:11 - INFO - codeparrot_training - Step 21057: {'lr': 0.00032947614500673306, 'samples': 4043136, 'steps': 21057, 'loss/train': 2.4297671914100647} 01/29/2022 15:58:16 - INFO - codeparrot_training - Step 21058: {'lr': 0.0003294606312149128, 'samples': 4043328, 'steps': 21058, 'loss/train': 1.9491307139396667} 01/29/2022 15:58:21 - INFO - codeparrot_training - Step 21059: {'lr': 0.00032944511708270853, 'samples': 4043520, 'steps': 21059, 'loss/train': 2.201681613922119} 01/29/2022 15:58:25 - INFO - codeparrot_training - Step 21060: {'lr': 0.00032942960261018653, 'samples': 4043712, 'steps': 21060, 'loss/train': 1.3024949133396149} 01/29/2022 15:58:29 - INFO - codeparrot_training - Step 21061: {'lr': 0.0003294140877974133, 'samples': 4043904, 'steps': 21061, 'loss/train': 1.6371434926986694} 01/29/2022 15:58:33 - INFO - codeparrot_training - Step 21062: {'lr': 0.00032939857264445535, 'samples': 4044096, 'steps': 21062, 'loss/train': 0.7300317138433456} 01/29/2022 15:58:40 - INFO - codeparrot_training - Step 21063: {'lr': 0.00032938305715137916, 'samples': 4044288, 'steps': 21063, 'loss/train': 1.5116407871246338} 01/29/2022 15:58:45 - INFO - codeparrot_training - Step 21064: {'lr': 0.00032936754131825113, 'samples': 4044480, 'steps': 21064, 'loss/train': 1.7129729390144348} 01/29/2022 15:58:49 - INFO - codeparrot_training - Step 21065: {'lr': 0.0003293520251451378, 'samples': 4044672, 'steps': 21065, 'loss/train': 0.9751740396022797} 01/29/2022 15:58:53 - INFO - codeparrot_training - Step 21066: {'lr': 0.00032933650863210557, 'samples': 4044864, 'steps': 21066, 'loss/train': 1.7153908610343933} 01/29/2022 15:58:57 - INFO - codeparrot_training - Step 21067: {'lr': 0.00032932099177922095, 'samples': 4045056, 'steps': 21067, 'loss/train': 1.7728772163391113} 01/29/2022 15:59:03 - INFO - codeparrot_training - Step 21068: {'lr': 0.00032930547458655035, 'samples': 4045248, 'steps': 21068, 'loss/train': 1.7877910137176514} 01/29/2022 15:59:07 - INFO - codeparrot_training - Step 21069: {'lr': 0.0003292899570541603, 'samples': 4045440, 'steps': 21069, 'loss/train': 1.3848290741443634} 01/29/2022 15:59:12 - INFO - codeparrot_training - Step 21070: {'lr': 0.00032927443918211717, 'samples': 4045632, 'steps': 21070, 'loss/train': 1.3535930514335632} 01/29/2022 15:59:16 - INFO - codeparrot_training - Step 21071: {'lr': 0.0003292589209704876, 'samples': 4045824, 'steps': 21071, 'loss/train': 0.6645129919052124} 01/29/2022 15:59:20 - INFO - codeparrot_training - Step 21072: {'lr': 0.000329243402419338, 'samples': 4046016, 'steps': 21072, 'loss/train': 0.5153068006038666} 01/29/2022 15:59:26 - INFO - codeparrot_training - Step 21073: {'lr': 0.00032922788352873474, 'samples': 4046208, 'steps': 21073, 'loss/train': 1.4877130687236786} 01/29/2022 15:59:30 - INFO - codeparrot_training - Step 21074: {'lr': 0.00032921236429874444, 'samples': 4046400, 'steps': 21074, 'loss/train': 1.2659651041030884} 01/29/2022 15:59:34 - INFO - codeparrot_training - Step 21075: {'lr': 0.0003291968447294335, 'samples': 4046592, 'steps': 21075, 'loss/train': 1.6376323699951172} 01/29/2022 15:59:39 - INFO - codeparrot_training - Step 21076: {'lr': 0.0003291813248208685, 'samples': 4046784, 'steps': 21076, 'loss/train': 1.1473042666912079} 01/29/2022 15:59:46 - INFO - codeparrot_training - Step 21077: {'lr': 0.00032916580457311573, 'samples': 4046976, 'steps': 21077, 'loss/train': 0.5242272466421127} 01/29/2022 15:59:50 - INFO - codeparrot_training - Step 21078: {'lr': 0.00032915028398624186, 'samples': 4047168, 'steps': 21078, 'loss/train': 1.323585569858551} 01/29/2022 15:59:54 - INFO - codeparrot_training - Step 21079: {'lr': 0.0003291347630603133, 'samples': 4047360, 'steps': 21079, 'loss/train': 0.929543137550354} 01/29/2022 15:59:58 - INFO - codeparrot_training - Step 21080: {'lr': 0.00032911924179539653, 'samples': 4047552, 'steps': 21080, 'loss/train': 0.8446750938892365} 01/29/2022 16:00:03 - INFO - codeparrot_training - Step 21081: {'lr': 0.0003291037201915581, 'samples': 4047744, 'steps': 21081, 'loss/train': 1.5219755172729492} 01/29/2022 16:00:08 - INFO - codeparrot_training - Step 21082: {'lr': 0.0003290881982488644, 'samples': 4047936, 'steps': 21082, 'loss/train': 0.6172434389591217} 01/29/2022 16:00:12 - INFO - codeparrot_training - Step 21083: {'lr': 0.000329072675967382, 'samples': 4048128, 'steps': 21083, 'loss/train': 1.5039417743682861} 01/29/2022 16:00:17 - INFO - codeparrot_training - Step 21084: {'lr': 0.00032905715334717735, 'samples': 4048320, 'steps': 21084, 'loss/train': 1.533058226108551} 01/29/2022 16:00:21 - INFO - codeparrot_training - Step 21085: {'lr': 0.000329041630388317, 'samples': 4048512, 'steps': 21085, 'loss/train': 2.3101821541786194} 01/29/2022 16:00:25 - INFO - codeparrot_training - Step 21086: {'lr': 0.00032902610709086727, 'samples': 4048704, 'steps': 21086, 'loss/train': 0.42404817044734955} 01/29/2022 16:00:30 - INFO - codeparrot_training - Step 21087: {'lr': 0.00032901058345489494, 'samples': 4048896, 'steps': 21087, 'loss/train': 1.733474314212799} 01/29/2022 16:00:35 - INFO - codeparrot_training - Step 21088: {'lr': 0.00032899505948046624, 'samples': 4049088, 'steps': 21088, 'loss/train': 1.245892345905304} 01/29/2022 16:00:39 - INFO - codeparrot_training - Step 21089: {'lr': 0.0003289795351676479, 'samples': 4049280, 'steps': 21089, 'loss/train': 1.357800453901291} 01/29/2022 16:00:43 - INFO - codeparrot_training - Step 21090: {'lr': 0.0003289640105165063, 'samples': 4049472, 'steps': 21090, 'loss/train': 1.677438497543335} 01/29/2022 16:00:47 - INFO - codeparrot_training - Step 21091: {'lr': 0.0003289484855271078, 'samples': 4049664, 'steps': 21091, 'loss/train': 1.9407965540885925} 01/29/2022 16:00:54 - INFO - codeparrot_training - Step 21092: {'lr': 0.0003289329601995192, 'samples': 4049856, 'steps': 21092, 'loss/train': 1.5903371572494507} 01/29/2022 16:00:59 - INFO - codeparrot_training - Step 21093: {'lr': 0.00032891743453380685, 'samples': 4050048, 'steps': 21093, 'loss/train': 1.2615842521190643} 01/29/2022 16:01:03 - INFO - codeparrot_training - Step 21094: {'lr': 0.00032890190853003727, 'samples': 4050240, 'steps': 21094, 'loss/train': 1.2445261180400848} 01/29/2022 16:01:07 - INFO - codeparrot_training - Step 21095: {'lr': 0.0003288863821882769, 'samples': 4050432, 'steps': 21095, 'loss/train': 0.9333060085773468} 01/29/2022 16:01:11 - INFO - codeparrot_training - Step 21096: {'lr': 0.0003288708555085924, 'samples': 4050624, 'steps': 21096, 'loss/train': 1.8421223759651184} 01/29/2022 16:01:17 - INFO - codeparrot_training - Step 21097: {'lr': 0.00032885532849105014, 'samples': 4050816, 'steps': 21097, 'loss/train': 1.4639641642570496} 01/29/2022 16:01:21 - INFO - codeparrot_training - Step 21098: {'lr': 0.0003288398011357168, 'samples': 4051008, 'steps': 21098, 'loss/train': 2.0838798880577087} 01/29/2022 16:01:25 - INFO - codeparrot_training - Step 21099: {'lr': 0.0003288242734426586, 'samples': 4051200, 'steps': 21099, 'loss/train': 1.4685582518577576} 01/29/2022 16:01:29 - INFO - codeparrot_training - Step 21100: {'lr': 0.00032880874541194244, 'samples': 4051392, 'steps': 21100, 'loss/train': 1.687257707118988} 01/29/2022 16:01:34 - INFO - codeparrot_training - Step 21101: {'lr': 0.0003287932170436345, 'samples': 4051584, 'steps': 21101, 'loss/train': 2.842062771320343} 01/29/2022 16:01:39 - INFO - codeparrot_training - Step 21102: {'lr': 0.00032877768833780146, 'samples': 4051776, 'steps': 21102, 'loss/train': 0.8884884417057037} 01/29/2022 16:01:43 - INFO - codeparrot_training - Step 21103: {'lr': 0.0003287621592945099, 'samples': 4051968, 'steps': 21103, 'loss/train': 0.7986539304256439} 01/29/2022 16:01:47 - INFO - codeparrot_training - Step 21104: {'lr': 0.0003287466299138262, 'samples': 4052160, 'steps': 21104, 'loss/train': 1.8965204358100891} 01/29/2022 16:01:52 - INFO - codeparrot_training - Step 21105: {'lr': 0.000328731100195817, 'samples': 4052352, 'steps': 21105, 'loss/train': 2.030494451522827} 01/29/2022 16:01:56 - INFO - codeparrot_training - Step 21106: {'lr': 0.00032871557014054864, 'samples': 4052544, 'steps': 21106, 'loss/train': 1.4105859696865082} 01/29/2022 16:02:01 - INFO - codeparrot_training - Step 21107: {'lr': 0.00032870003974808787, 'samples': 4052736, 'steps': 21107, 'loss/train': 7.304655075073242} 01/29/2022 16:02:06 - INFO - codeparrot_training - Step 21108: {'lr': 0.0003286845090185011, 'samples': 4052928, 'steps': 21108, 'loss/train': 1.451763242483139} 01/29/2022 16:02:10 - INFO - codeparrot_training - Step 21109: {'lr': 0.0003286689779518549, 'samples': 4053120, 'steps': 21109, 'loss/train': 1.4594423174858093} 01/29/2022 16:02:14 - INFO - codeparrot_training - Step 21110: {'lr': 0.00032865344654821576, 'samples': 4053312, 'steps': 21110, 'loss/train': 1.5045238137245178} 01/29/2022 16:02:18 - INFO - codeparrot_training - Step 21111: {'lr': 0.0003286379148076502, 'samples': 4053504, 'steps': 21111, 'loss/train': 1.5218417644500732} 01/29/2022 16:02:26 - INFO - codeparrot_training - Step 21112: {'lr': 0.00032862238273022483, 'samples': 4053696, 'steps': 21112, 'loss/train': 1.516460359096527} 01/29/2022 16:02:30 - INFO - codeparrot_training - Step 21113: {'lr': 0.0003286068503160061, 'samples': 4053888, 'steps': 21113, 'loss/train': 2.291634678840637} 01/29/2022 16:02:34 - INFO - codeparrot_training - Step 21114: {'lr': 0.0003285913175650606, 'samples': 4054080, 'steps': 21114, 'loss/train': 1.3441151082515717} 01/29/2022 16:02:38 - INFO - codeparrot_training - Step 21115: {'lr': 0.00032857578447745484, 'samples': 4054272, 'steps': 21115, 'loss/train': 1.4368802905082703} 01/29/2022 16:02:43 - INFO - codeparrot_training - Step 21116: {'lr': 0.00032856025105325537, 'samples': 4054464, 'steps': 21116, 'loss/train': 1.4613354206085205} 01/29/2022 16:02:48 - INFO - codeparrot_training - Step 21117: {'lr': 0.00032854471729252876, 'samples': 4054656, 'steps': 21117, 'loss/train': 1.578022062778473} 01/29/2022 16:02:52 - INFO - codeparrot_training - Step 21118: {'lr': 0.00032852918319534153, 'samples': 4054848, 'steps': 21118, 'loss/train': 0.34143950790166855} 01/29/2022 16:02:57 - INFO - codeparrot_training - Step 21119: {'lr': 0.00032851364876176014, 'samples': 4055040, 'steps': 21119, 'loss/train': 1.3641219735145569} 01/29/2022 16:03:01 - INFO - codeparrot_training - Step 21120: {'lr': 0.00032849811399185126, 'samples': 4055232, 'steps': 21120, 'loss/train': 0.5306697338819504} 01/29/2022 16:03:05 - INFO - codeparrot_training - Step 21121: {'lr': 0.0003284825788856814, 'samples': 4055424, 'steps': 21121, 'loss/train': 1.3367848098278046} 01/29/2022 16:03:12 - INFO - codeparrot_training - Step 21122: {'lr': 0.00032846704344331707, 'samples': 4055616, 'steps': 21122, 'loss/train': 1.2970980405807495} 01/29/2022 16:03:16 - INFO - codeparrot_training - Step 21123: {'lr': 0.00032845150766482484, 'samples': 4055808, 'steps': 21123, 'loss/train': 1.0045409202575684} 01/29/2022 16:03:20 - INFO - codeparrot_training - Step 21124: {'lr': 0.0003284359715502713, 'samples': 4056000, 'steps': 21124, 'loss/train': 2.2005234360694885} 01/29/2022 16:03:25 - INFO - codeparrot_training - Step 21125: {'lr': 0.0003284204350997229, 'samples': 4056192, 'steps': 21125, 'loss/train': 0.954658716917038} 01/29/2022 16:03:29 - INFO - codeparrot_training - Step 21126: {'lr': 0.0003284048983132463, 'samples': 4056384, 'steps': 21126, 'loss/train': 1.6743543148040771} 01/29/2022 16:03:34 - INFO - codeparrot_training - Step 21127: {'lr': 0.000328389361190908, 'samples': 4056576, 'steps': 21127, 'loss/train': 2.459975302219391} 01/29/2022 16:03:38 - INFO - codeparrot_training - Step 21128: {'lr': 0.0003283738237327745, 'samples': 4056768, 'steps': 21128, 'loss/train': 1.3099952638149261} 01/29/2022 16:03:43 - INFO - codeparrot_training - Step 21129: {'lr': 0.0003283582859389125, 'samples': 4056960, 'steps': 21129, 'loss/train': 1.1118730902671814} 01/29/2022 16:03:47 - INFO - codeparrot_training - Step 21130: {'lr': 0.0003283427478093885, 'samples': 4057152, 'steps': 21130, 'loss/train': 1.606732964515686} 01/29/2022 16:03:51 - INFO - codeparrot_training - Step 21131: {'lr': 0.0003283272093442691, 'samples': 4057344, 'steps': 21131, 'loss/train': 1.6964426636695862} 01/29/2022 16:03:57 - INFO - codeparrot_training - Step 21132: {'lr': 0.00032831167054362065, 'samples': 4057536, 'steps': 21132, 'loss/train': 2.0252389311790466} 01/29/2022 16:04:01 - INFO - codeparrot_training - Step 21133: {'lr': 0.00032829613140751004, 'samples': 4057728, 'steps': 21133, 'loss/train': 1.8567159175872803} 01/29/2022 16:04:05 - INFO - codeparrot_training - Step 21134: {'lr': 0.0003282805919360035, 'samples': 4057920, 'steps': 21134, 'loss/train': 2.174225091934204} 01/29/2022 16:04:09 - INFO - codeparrot_training - Step 21135: {'lr': 0.0003282650521291679, 'samples': 4058112, 'steps': 21135, 'loss/train': 1.4085129797458649} 01/29/2022 16:04:17 - INFO - codeparrot_training - Step 21136: {'lr': 0.0003282495119870695, 'samples': 4058304, 'steps': 21136, 'loss/train': 1.8923735618591309} 01/29/2022 16:04:21 - INFO - codeparrot_training - Step 21137: {'lr': 0.00032823397150977523, 'samples': 4058496, 'steps': 21137, 'loss/train': 1.448425143957138} 01/29/2022 16:04:25 - INFO - codeparrot_training - Step 21138: {'lr': 0.00032821843069735134, 'samples': 4058688, 'steps': 21138, 'loss/train': 1.8757511973381042} 01/29/2022 16:04:29 - INFO - codeparrot_training - Step 21139: {'lr': 0.00032820288954986453, 'samples': 4058880, 'steps': 21139, 'loss/train': 2.0295220613479614} 01/29/2022 16:04:34 - INFO - codeparrot_training - Step 21140: {'lr': 0.00032818734806738147, 'samples': 4059072, 'steps': 21140, 'loss/train': 1.5387568473815918} 01/29/2022 16:04:38 - INFO - codeparrot_training - Step 21141: {'lr': 0.0003281718062499686, 'samples': 4059264, 'steps': 21141, 'loss/train': 2.071555495262146} 01/29/2022 16:04:44 - INFO - codeparrot_training - Step 21142: {'lr': 0.0003281562640976925, 'samples': 4059456, 'steps': 21142, 'loss/train': 1.4780239462852478} 01/29/2022 16:04:48 - INFO - codeparrot_training - Step 21143: {'lr': 0.0003281407216106198, 'samples': 4059648, 'steps': 21143, 'loss/train': 1.9379494786262512} 01/29/2022 16:04:52 - INFO - codeparrot_training - Step 21144: {'lr': 0.0003281251787888171, 'samples': 4059840, 'steps': 21144, 'loss/train': 1.42442786693573} 01/29/2022 16:04:56 - INFO - codeparrot_training - Step 21145: {'lr': 0.00032810963563235083, 'samples': 4060032, 'steps': 21145, 'loss/train': 2.000829041004181} 01/29/2022 16:05:01 - INFO - codeparrot_training - Step 21146: {'lr': 0.00032809409214128784, 'samples': 4060224, 'steps': 21146, 'loss/train': 1.561347484588623} 01/29/2022 16:05:09 - INFO - codeparrot_training - Step 21147: {'lr': 0.0003280785483156944, 'samples': 4060416, 'steps': 21147, 'loss/train': 1.8228759169578552} 01/29/2022 16:05:13 - INFO - codeparrot_training - Step 21148: {'lr': 0.0003280630041556374, 'samples': 4060608, 'steps': 21148, 'loss/train': 1.823890507221222} 01/29/2022 16:05:17 - INFO - codeparrot_training - Step 21149: {'lr': 0.0003280474596611832, 'samples': 4060800, 'steps': 21149, 'loss/train': 1.1627540588378906} 01/29/2022 16:05:21 - INFO - codeparrot_training - Step 21150: {'lr': 0.0003280319148323985, 'samples': 4060992, 'steps': 21150, 'loss/train': 1.7814282774925232} 01/29/2022 16:05:26 - INFO - codeparrot_training - Step 21151: {'lr': 0.00032801636966934975, 'samples': 4061184, 'steps': 21151, 'loss/train': 1.271148920059204} 01/29/2022 16:05:31 - INFO - codeparrot_training - Step 21152: {'lr': 0.0003280008241721038, 'samples': 4061376, 'steps': 21152, 'loss/train': 2.6472997069358826} 01/29/2022 16:05:35 - INFO - codeparrot_training - Step 21153: {'lr': 0.0003279852783407269, 'samples': 4061568, 'steps': 21153, 'loss/train': 1.864529013633728} 01/29/2022 16:05:39 - INFO - codeparrot_training - Step 21154: {'lr': 0.00032796973217528595, 'samples': 4061760, 'steps': 21154, 'loss/train': 0.9200921058654785} 01/29/2022 16:05:44 - INFO - codeparrot_training - Step 21155: {'lr': 0.00032795418567584746, 'samples': 4061952, 'steps': 21155, 'loss/train': 1.5615895986557007} 01/29/2022 16:05:48 - INFO - codeparrot_training - Step 21156: {'lr': 0.00032793863884247794, 'samples': 4062144, 'steps': 21156, 'loss/train': 2.074945628643036} 01/29/2022 16:05:53 - INFO - codeparrot_training - Step 21157: {'lr': 0.00032792309167524404, 'samples': 4062336, 'steps': 21157, 'loss/train': 1.9370868802070618} 01/29/2022 16:05:57 - INFO - codeparrot_training - Step 21158: {'lr': 0.00032790754417421237, 'samples': 4062528, 'steps': 21158, 'loss/train': 1.9839053750038147} 01/29/2022 16:06:02 - INFO - codeparrot_training - Step 21159: {'lr': 0.00032789199633944955, 'samples': 4062720, 'steps': 21159, 'loss/train': 1.9942070245742798} 01/29/2022 16:06:06 - INFO - codeparrot_training - Step 21160: {'lr': 0.00032787644817102205, 'samples': 4062912, 'steps': 21160, 'loss/train': 1.877913773059845} 01/29/2022 16:06:10 - INFO - codeparrot_training - Step 21161: {'lr': 0.00032786089966899666, 'samples': 4063104, 'steps': 21161, 'loss/train': 0.6455883532762527} 01/29/2022 16:06:15 - INFO - codeparrot_training - Step 21162: {'lr': 0.0003278453508334399, 'samples': 4063296, 'steps': 21162, 'loss/train': 1.7539565563201904} 01/29/2022 16:06:20 - INFO - codeparrot_training - Step 21163: {'lr': 0.00032782980166441836, 'samples': 4063488, 'steps': 21163, 'loss/train': 1.8310164213180542} 01/29/2022 16:06:24 - INFO - codeparrot_training - Step 21164: {'lr': 0.00032781425216199864, 'samples': 4063680, 'steps': 21164, 'loss/train': 1.707143247127533} 01/29/2022 16:06:28 - INFO - codeparrot_training - Step 21165: {'lr': 0.00032779870232624737, 'samples': 4063872, 'steps': 21165, 'loss/train': 1.6706160306930542} 01/29/2022 16:06:33 - INFO - codeparrot_training - Step 21166: {'lr': 0.0003277831521572312, 'samples': 4064064, 'steps': 21166, 'loss/train': 0.5484946668148041} 01/29/2022 16:06:40 - INFO - codeparrot_training - Step 21167: {'lr': 0.00032776760165501663, 'samples': 4064256, 'steps': 21167, 'loss/train': 0.15316829457879066} 01/29/2022 16:06:44 - INFO - codeparrot_training - Step 21168: {'lr': 0.00032775205081967047, 'samples': 4064448, 'steps': 21168, 'loss/train': 0.9945698082447052} 01/29/2022 16:06:49 - INFO - codeparrot_training - Step 21169: {'lr': 0.00032773649965125914, 'samples': 4064640, 'steps': 21169, 'loss/train': 1.7751173973083496} 01/29/2022 16:06:53 - INFO - codeparrot_training - Step 21170: {'lr': 0.0003277209481498493, 'samples': 4064832, 'steps': 21170, 'loss/train': 1.3721177279949188} 01/29/2022 16:06:57 - INFO - codeparrot_training - Step 21171: {'lr': 0.00032770539631550767, 'samples': 4065024, 'steps': 21171, 'loss/train': 1.1855559647083282} 01/29/2022 16:07:03 - INFO - codeparrot_training - Step 21172: {'lr': 0.0003276898441483008, 'samples': 4065216, 'steps': 21172, 'loss/train': 2.010818123817444} 01/29/2022 16:07:07 - INFO - codeparrot_training - Step 21173: {'lr': 0.0003276742916482952, 'samples': 4065408, 'steps': 21173, 'loss/train': 1.9791417717933655} 01/29/2022 16:07:11 - INFO - codeparrot_training - Step 21174: {'lr': 0.00032765873881555765, 'samples': 4065600, 'steps': 21174, 'loss/train': 0.8966810703277588} 01/29/2022 16:07:15 - INFO - codeparrot_training - Step 21175: {'lr': 0.0003276431856501548, 'samples': 4065792, 'steps': 21175, 'loss/train': 1.2069447040557861} 01/29/2022 16:07:21 - INFO - codeparrot_training - Step 21176: {'lr': 0.0003276276321521531, 'samples': 4065984, 'steps': 21176, 'loss/train': 1.5877802968025208} 01/29/2022 16:07:25 - INFO - codeparrot_training - Step 21177: {'lr': 0.00032761207832161934, 'samples': 4066176, 'steps': 21177, 'loss/train': 1.6964013576507568} 01/29/2022 16:07:29 - INFO - codeparrot_training - Step 21178: {'lr': 0.00032759652415862003, 'samples': 4066368, 'steps': 21178, 'loss/train': 1.2949093580245972} 01/29/2022 16:07:33 - INFO - codeparrot_training - Step 21179: {'lr': 0.00032758096966322185, 'samples': 4066560, 'steps': 21179, 'loss/train': 1.3327010571956635} 01/29/2022 16:07:38 - INFO - codeparrot_training - Step 21180: {'lr': 0.00032756541483549146, 'samples': 4066752, 'steps': 21180, 'loss/train': 2.2171048521995544} 01/29/2022 16:07:45 - INFO - codeparrot_training - Step 21181: {'lr': 0.00032754985967549545, 'samples': 4066944, 'steps': 21181, 'loss/train': 1.628333330154419} 01/29/2022 16:07:49 - INFO - codeparrot_training - Step 21182: {'lr': 0.0003275343041833005, 'samples': 4067136, 'steps': 21182, 'loss/train': 0.5994250774383545} 01/29/2022 16:07:53 - INFO - codeparrot_training - Step 21183: {'lr': 0.00032751874835897316, 'samples': 4067328, 'steps': 21183, 'loss/train': 1.2531218826770782} 01/29/2022 16:07:58 - INFO - codeparrot_training - Step 21184: {'lr': 0.0003275031922025801, 'samples': 4067520, 'steps': 21184, 'loss/train': 0.7898932993412018} 01/29/2022 16:08:02 - INFO - codeparrot_training - Step 21185: {'lr': 0.00032748763571418805, 'samples': 4067712, 'steps': 21185, 'loss/train': 0.6734169870615005} 01/29/2022 16:08:07 - INFO - codeparrot_training - Step 21186: {'lr': 0.00032747207889386354, 'samples': 4067904, 'steps': 21186, 'loss/train': 1.25256809592247} 01/29/2022 16:08:12 - INFO - codeparrot_training - Step 21187: {'lr': 0.00032745652174167325, 'samples': 4068096, 'steps': 21187, 'loss/train': 1.3888893127441406} 01/29/2022 16:08:16 - INFO - codeparrot_training - Step 21188: {'lr': 0.00032744096425768376, 'samples': 4068288, 'steps': 21188, 'loss/train': 2.046785831451416} 01/29/2022 16:08:20 - INFO - codeparrot_training - Step 21189: {'lr': 0.00032742540644196185, 'samples': 4068480, 'steps': 21189, 'loss/train': 1.9390078783035278} 01/29/2022 16:08:24 - INFO - codeparrot_training - Step 21190: {'lr': 0.00032740984829457404, 'samples': 4068672, 'steps': 21190, 'loss/train': 0.7170006483793259} 01/29/2022 16:08:31 - INFO - codeparrot_training - Step 21191: {'lr': 0.00032739428981558706, 'samples': 4068864, 'steps': 21191, 'loss/train': 1.439355343580246} 01/29/2022 16:08:36 - INFO - codeparrot_training - Step 21192: {'lr': 0.0003273787310050675, 'samples': 4069056, 'steps': 21192, 'loss/train': 2.3440834879875183} 01/29/2022 16:08:40 - INFO - codeparrot_training - Step 21193: {'lr': 0.00032736317186308193, 'samples': 4069248, 'steps': 21193, 'loss/train': 2.0745128989219666} 01/29/2022 16:08:45 - INFO - codeparrot_training - Step 21194: {'lr': 0.00032734761238969724, 'samples': 4069440, 'steps': 21194, 'loss/train': 1.0349149703979492} 01/29/2022 16:08:49 - INFO - codeparrot_training - Step 21195: {'lr': 0.00032733205258497994, 'samples': 4069632, 'steps': 21195, 'loss/train': 1.0254266560077667} 01/29/2022 16:08:54 - INFO - codeparrot_training - Step 21196: {'lr': 0.0003273164924489966, 'samples': 4069824, 'steps': 21196, 'loss/train': 1.573561191558838} 01/29/2022 16:08:58 - INFO - codeparrot_training - Step 21197: {'lr': 0.000327300931981814, 'samples': 4070016, 'steps': 21197, 'loss/train': 1.1022193729877472} 01/29/2022 16:09:03 - INFO - codeparrot_training - Step 21198: {'lr': 0.0003272853711834987, 'samples': 4070208, 'steps': 21198, 'loss/train': 0.8199686408042908} 01/29/2022 16:09:07 - INFO - codeparrot_training - Step 21199: {'lr': 0.00032726981005411747, 'samples': 4070400, 'steps': 21199, 'loss/train': 1.7957999110221863} 01/29/2022 16:09:11 - INFO - codeparrot_training - Step 21200: {'lr': 0.00032725424859373687, 'samples': 4070592, 'steps': 21200, 'loss/train': 1.471916913986206} 01/29/2022 16:09:17 - INFO - codeparrot_training - Step 21201: {'lr': 0.0003272386868024236, 'samples': 4070784, 'steps': 21201, 'loss/train': 1.7681937217712402} 01/29/2022 16:09:21 - INFO - codeparrot_training - Step 21202: {'lr': 0.00032722312468024434, 'samples': 4070976, 'steps': 21202, 'loss/train': 1.179858237504959} 01/29/2022 16:09:26 - INFO - codeparrot_training - Step 21203: {'lr': 0.00032720756222726576, 'samples': 4071168, 'steps': 21203, 'loss/train': 1.6280057430267334} 01/29/2022 16:09:30 - INFO - codeparrot_training - Step 21204: {'lr': 0.0003271919994435545, 'samples': 4071360, 'steps': 21204, 'loss/train': 0.9892523288726807} 01/29/2022 16:09:34 - INFO - codeparrot_training - Step 21205: {'lr': 0.0003271764363291772, 'samples': 4071552, 'steps': 21205, 'loss/train': 1.6876723766326904} 01/29/2022 16:09:41 - INFO - codeparrot_training - Step 21206: {'lr': 0.00032716087288420047, 'samples': 4071744, 'steps': 21206, 'loss/train': 2.053990602493286} 01/29/2022 16:09:45 - INFO - codeparrot_training - Step 21207: {'lr': 0.0003271453091086912, 'samples': 4071936, 'steps': 21207, 'loss/train': 0.9875030815601349} 01/29/2022 16:09:50 - INFO - codeparrot_training - Step 21208: {'lr': 0.0003271297450027159, 'samples': 4072128, 'steps': 21208, 'loss/train': 1.7888685464859009} 01/29/2022 16:09:54 - INFO - codeparrot_training - Step 21209: {'lr': 0.0003271141805663412, 'samples': 4072320, 'steps': 21209, 'loss/train': 2.0186886191368103} 01/29/2022 16:09:58 - INFO - codeparrot_training - Step 21210: {'lr': 0.00032709861579963384, 'samples': 4072512, 'steps': 21210, 'loss/train': 2.066882014274597} 01/29/2022 16:10:03 - INFO - codeparrot_training - Step 21211: {'lr': 0.00032708305070266053, 'samples': 4072704, 'steps': 21211, 'loss/train': 1.8232916593551636} 01/29/2022 16:10:08 - INFO - codeparrot_training - Step 21212: {'lr': 0.00032706748527548793, 'samples': 4072896, 'steps': 21212, 'loss/train': 1.6019783020019531} 01/29/2022 16:10:12 - INFO - codeparrot_training - Step 21213: {'lr': 0.0003270519195181826, 'samples': 4073088, 'steps': 21213, 'loss/train': 1.2767611742019653} 01/29/2022 16:10:16 - INFO - codeparrot_training - Step 21214: {'lr': 0.00032703635343081137, 'samples': 4073280, 'steps': 21214, 'loss/train': 1.3241440057754517} 01/29/2022 16:10:20 - INFO - codeparrot_training - Step 21215: {'lr': 0.0003270207870134408, 'samples': 4073472, 'steps': 21215, 'loss/train': 1.2654865980148315} 01/29/2022 16:10:26 - INFO - codeparrot_training - Step 21216: {'lr': 0.00032700522026613785, 'samples': 4073664, 'steps': 21216, 'loss/train': 1.534800946712494} 01/29/2022 16:10:30 - INFO - codeparrot_training - Step 21217: {'lr': 0.00032698965318896876, 'samples': 4073856, 'steps': 21217, 'loss/train': 1.2151935696601868} 01/29/2022 16:10:34 - INFO - codeparrot_training - Step 21218: {'lr': 0.00032697408578200054, 'samples': 4074048, 'steps': 21218, 'loss/train': 1.4041151404380798} 01/29/2022 16:10:38 - INFO - codeparrot_training - Step 21219: {'lr': 0.00032695851804529977, 'samples': 4074240, 'steps': 21219, 'loss/train': 1.2761786878108978} 01/29/2022 16:10:43 - INFO - codeparrot_training - Step 21220: {'lr': 0.0003269429499789331, 'samples': 4074432, 'steps': 21220, 'loss/train': 1.859476089477539} 01/29/2022 16:10:49 - INFO - codeparrot_training - Step 21221: {'lr': 0.0003269273815829673, 'samples': 4074624, 'steps': 21221, 'loss/train': 0.7173726707696915} 01/29/2022 16:10:53 - INFO - codeparrot_training - Step 21222: {'lr': 0.00032691181285746904, 'samples': 4074816, 'steps': 21222, 'loss/train': 1.9804447889328003} 01/29/2022 16:10:58 - INFO - codeparrot_training - Step 21223: {'lr': 0.000326896243802505, 'samples': 4075008, 'steps': 21223, 'loss/train': 1.4922093451023102} 01/29/2022 16:11:02 - INFO - codeparrot_training - Step 21224: {'lr': 0.0003268806744181418, 'samples': 4075200, 'steps': 21224, 'loss/train': 1.7136915922164917} 01/29/2022 16:11:10 - INFO - codeparrot_training - Step 21225: {'lr': 0.0003268651047044463, 'samples': 4075392, 'steps': 21225, 'loss/train': 1.4842160940170288} 01/29/2022 16:11:14 - INFO - codeparrot_training - Step 21226: {'lr': 0.00032684953466148505, 'samples': 4075584, 'steps': 21226, 'loss/train': 1.8658404350280762} 01/29/2022 16:11:18 - INFO - codeparrot_training - Step 21227: {'lr': 0.00032683396428932483, 'samples': 4075776, 'steps': 21227, 'loss/train': 0.5144465267658234} 01/29/2022 16:11:22 - INFO - codeparrot_training - Step 21228: {'lr': 0.0003268183935880322, 'samples': 4075968, 'steps': 21228, 'loss/train': 1.9309231638908386} 01/29/2022 16:11:27 - INFO - codeparrot_training - Step 21229: {'lr': 0.0003268028225576741, 'samples': 4076160, 'steps': 21229, 'loss/train': 1.2645375430583954} 01/29/2022 16:11:32 - INFO - codeparrot_training - Step 21230: {'lr': 0.00032678725119831696, 'samples': 4076352, 'steps': 21230, 'loss/train': 1.6293017864227295} 01/29/2022 16:11:36 - INFO - codeparrot_training - Step 21231: {'lr': 0.0003267716795100278, 'samples': 4076544, 'steps': 21231, 'loss/train': 1.2582186162471771} 01/29/2022 16:11:40 - INFO - codeparrot_training - Step 21232: {'lr': 0.000326756107492873, 'samples': 4076736, 'steps': 21232, 'loss/train': 0.6697174608707428} 01/29/2022 16:11:45 - INFO - codeparrot_training - Step 21233: {'lr': 0.00032674053514691946, 'samples': 4076928, 'steps': 21233, 'loss/train': 1.630894660949707} 01/29/2022 16:11:49 - INFO - codeparrot_training - Step 21234: {'lr': 0.00032672496247223377, 'samples': 4077120, 'steps': 21234, 'loss/train': 1.9763419032096863} 01/29/2022 16:11:54 - INFO - codeparrot_training - Step 21235: {'lr': 0.00032670938946888277, 'samples': 4077312, 'steps': 21235, 'loss/train': 2.098225235939026} 01/29/2022 16:11:59 - INFO - codeparrot_training - Step 21236: {'lr': 0.00032669381613693307, 'samples': 4077504, 'steps': 21236, 'loss/train': 1.6552614569664001} 01/29/2022 16:12:03 - INFO - codeparrot_training - Step 21237: {'lr': 0.0003266782424764515, 'samples': 4077696, 'steps': 21237, 'loss/train': 1.2016628086566925} 01/29/2022 16:12:07 - INFO - codeparrot_training - Step 21238: {'lr': 0.0003266626684875046, 'samples': 4077888, 'steps': 21238, 'loss/train': 3.618349313735962} 01/29/2022 16:12:11 - INFO - codeparrot_training - Step 21239: {'lr': 0.0003266470941701592, 'samples': 4078080, 'steps': 21239, 'loss/train': 2.0109845995903015} 01/29/2022 16:12:19 - INFO - codeparrot_training - Step 21240: {'lr': 0.00032663151952448194, 'samples': 4078272, 'steps': 21240, 'loss/train': 1.483216255903244} 01/29/2022 16:12:23 - INFO - codeparrot_training - Step 21241: {'lr': 0.00032661594455053966, 'samples': 4078464, 'steps': 21241, 'loss/train': 0.7467908710241318} 01/29/2022 16:12:27 - INFO - codeparrot_training - Step 21242: {'lr': 0.0003266003692483989, 'samples': 4078656, 'steps': 21242, 'loss/train': 1.9025457501411438} 01/29/2022 16:12:31 - INFO - codeparrot_training - Step 21243: {'lr': 0.0003265847936181266, 'samples': 4078848, 'steps': 21243, 'loss/train': 2.4967607259750366} 01/29/2022 16:12:36 - INFO - codeparrot_training - Step 21244: {'lr': 0.0003265692176597893, 'samples': 4079040, 'steps': 21244, 'loss/train': 1.1065069735050201} 01/29/2022 16:12:41 - INFO - codeparrot_training - Step 21245: {'lr': 0.0003265536413734538, 'samples': 4079232, 'steps': 21245, 'loss/train': 3.0632418394088745} 01/29/2022 16:12:45 - INFO - codeparrot_training - Step 21246: {'lr': 0.0003265380647591868, 'samples': 4079424, 'steps': 21246, 'loss/train': 1.9384936094284058} 01/29/2022 16:12:49 - INFO - codeparrot_training - Step 21247: {'lr': 0.000326522487817055, 'samples': 4079616, 'steps': 21247, 'loss/train': 1.380155235528946} 01/29/2022 16:12:54 - INFO - codeparrot_training - Step 21248: {'lr': 0.00032650691054712523, 'samples': 4079808, 'steps': 21248, 'loss/train': 2.00444233417511} 01/29/2022 16:12:58 - INFO - codeparrot_training - Step 21249: {'lr': 0.0003264913329494641, 'samples': 4080000, 'steps': 21249, 'loss/train': 1.856681764125824} 01/29/2022 16:13:03 - INFO - codeparrot_training - Step 21250: {'lr': 0.0003264757550241384, 'samples': 4080192, 'steps': 21250, 'loss/train': 1.456993728876114} 01/29/2022 16:13:07 - INFO - codeparrot_training - Step 21251: {'lr': 0.00032646017677121484, 'samples': 4080384, 'steps': 21251, 'loss/train': 1.9205068945884705} 01/29/2022 16:13:12 - INFO - codeparrot_training - Step 21252: {'lr': 0.00032644459819076017, 'samples': 4080576, 'steps': 21252, 'loss/train': 2.4953269958496094} 01/29/2022 16:13:16 - INFO - codeparrot_training - Step 21253: {'lr': 0.00032642901928284115, 'samples': 4080768, 'steps': 21253, 'loss/train': 1.4688560664653778} 01/29/2022 16:13:20 - INFO - codeparrot_training - Step 21254: {'lr': 0.0003264134400475244, 'samples': 4080960, 'steps': 21254, 'loss/train': 1.609529435634613} 01/29/2022 16:13:27 - INFO - codeparrot_training - Step 21255: {'lr': 0.00032639786048487685, 'samples': 4081152, 'steps': 21255, 'loss/train': 1.6041235327720642} 01/29/2022 16:13:32 - INFO - codeparrot_training - Step 21256: {'lr': 0.000326382280594965, 'samples': 4081344, 'steps': 21256, 'loss/train': 3.0289549827575684} 01/29/2022 16:13:36 - INFO - codeparrot_training - Step 21257: {'lr': 0.00032636670037785583, 'samples': 4081536, 'steps': 21257, 'loss/train': 2.2679505944252014} 01/29/2022 16:13:40 - INFO - codeparrot_training - Step 21258: {'lr': 0.00032635111983361586, 'samples': 4081728, 'steps': 21258, 'loss/train': 1.4544684290885925} 01/29/2022 16:13:44 - INFO - codeparrot_training - Step 21259: {'lr': 0.0003263355389623119, 'samples': 4081920, 'steps': 21259, 'loss/train': 1.0377416610717773} 01/29/2022 16:13:50 - INFO - codeparrot_training - Step 21260: {'lr': 0.0003263199577640109, 'samples': 4082112, 'steps': 21260, 'loss/train': 1.5730133056640625} 01/29/2022 16:13:54 - INFO - codeparrot_training - Step 21261: {'lr': 0.00032630437623877936, 'samples': 4082304, 'steps': 21261, 'loss/train': 2.7873669862747192} 01/29/2022 16:13:58 - INFO - codeparrot_training - Step 21262: {'lr': 0.00032628879438668414, 'samples': 4082496, 'steps': 21262, 'loss/train': 1.6867529153823853} 01/29/2022 16:14:02 - INFO - codeparrot_training - Step 21263: {'lr': 0.00032627321220779184, 'samples': 4082688, 'steps': 21263, 'loss/train': 1.5460689067840576} 01/29/2022 16:14:07 - INFO - codeparrot_training - Step 21264: {'lr': 0.0003262576297021694, 'samples': 4082880, 'steps': 21264, 'loss/train': 1.7153116464614868} 01/29/2022 16:14:14 - INFO - codeparrot_training - Step 21265: {'lr': 0.00032624204686988343, 'samples': 4083072, 'steps': 21265, 'loss/train': 2.1497538685798645} 01/29/2022 16:14:18 - INFO - codeparrot_training - Step 21266: {'lr': 0.00032622646371100085, 'samples': 4083264, 'steps': 21266, 'loss/train': 1.419395238161087} 01/29/2022 16:14:22 - INFO - codeparrot_training - Step 21267: {'lr': 0.00032621088022558823, 'samples': 4083456, 'steps': 21267, 'loss/train': 2.115899920463562} 01/29/2022 16:14:27 - INFO - codeparrot_training - Step 21268: {'lr': 0.00032619529641371236, 'samples': 4083648, 'steps': 21268, 'loss/train': 1.7992649674415588} 01/29/2022 16:14:31 - INFO - codeparrot_training - Step 21269: {'lr': 0.0003261797122754401, 'samples': 4083840, 'steps': 21269, 'loss/train': 1.736428141593933} 01/29/2022 16:14:36 - INFO - codeparrot_training - Step 21270: {'lr': 0.00032616412781083813, 'samples': 4084032, 'steps': 21270, 'loss/train': 1.5450891852378845} 01/29/2022 16:14:41 - INFO - codeparrot_training - Step 21271: {'lr': 0.00032614854301997323, 'samples': 4084224, 'steps': 21271, 'loss/train': 2.331652343273163} 01/29/2022 16:14:45 - INFO - codeparrot_training - Step 21272: {'lr': 0.0003261329579029121, 'samples': 4084416, 'steps': 21272, 'loss/train': 1.8409114480018616} 01/29/2022 16:14:49 - INFO - codeparrot_training - Step 21273: {'lr': 0.00032611737245972155, 'samples': 4084608, 'steps': 21273, 'loss/train': 2.4883169531822205} 01/29/2022 16:14:53 - INFO - codeparrot_training - Step 21274: {'lr': 0.0003261017866904684, 'samples': 4084800, 'steps': 21274, 'loss/train': 1.58664071559906} 01/29/2022 16:14:59 - INFO - codeparrot_training - Step 21275: {'lr': 0.0003260862005952193, 'samples': 4084992, 'steps': 21275, 'loss/train': 1.3832688331604004} 01/29/2022 16:15:03 - INFO - codeparrot_training - Step 21276: {'lr': 0.00032607061417404113, 'samples': 4085184, 'steps': 21276, 'loss/train': 2.043902277946472} 01/29/2022 16:15:07 - INFO - codeparrot_training - Step 21277: {'lr': 0.0003260550274270007, 'samples': 4085376, 'steps': 21277, 'loss/train': 1.7405176162719727} 01/29/2022 16:15:12 - INFO - codeparrot_training - Step 21278: {'lr': 0.0003260394403541645, 'samples': 4085568, 'steps': 21278, 'loss/train': 0.20800381153821945} 01/29/2022 16:15:16 - INFO - codeparrot_training - Step 21279: {'lr': 0.00032602385295559953, 'samples': 4085760, 'steps': 21279, 'loss/train': 1.8625062704086304} 01/29/2022 16:15:22 - INFO - codeparrot_training - Step 21280: {'lr': 0.00032600826523137254, 'samples': 4085952, 'steps': 21280, 'loss/train': 0.9764403998851776} 01/29/2022 16:15:26 - INFO - codeparrot_training - Step 21281: {'lr': 0.00032599267718155036, 'samples': 4086144, 'steps': 21281, 'loss/train': 1.5406871438026428} 01/29/2022 16:15:30 - INFO - codeparrot_training - Step 21282: {'lr': 0.0003259770888061995, 'samples': 4086336, 'steps': 21282, 'loss/train': 1.772436261177063} 01/29/2022 16:15:34 - INFO - codeparrot_training - Step 21283: {'lr': 0.0003259615001053871, 'samples': 4086528, 'steps': 21283, 'loss/train': 1.8421627879142761} 01/29/2022 16:15:39 - INFO - codeparrot_training - Step 21284: {'lr': 0.0003259459110791797, 'samples': 4086720, 'steps': 21284, 'loss/train': 1.6550769209861755} 01/29/2022 16:15:46 - INFO - codeparrot_training - Step 21285: {'lr': 0.0003259303217276441, 'samples': 4086912, 'steps': 21285, 'loss/train': 1.643521785736084} 01/29/2022 16:15:50 - INFO - codeparrot_training - Step 21286: {'lr': 0.0003259147320508471, 'samples': 4087104, 'steps': 21286, 'loss/train': 1.865044355392456} 01/29/2022 16:15:55 - INFO - codeparrot_training - Step 21287: {'lr': 0.0003258991420488555, 'samples': 4087296, 'steps': 21287, 'loss/train': 2.255763351917267} 01/29/2022 16:15:59 - INFO - codeparrot_training - Step 21288: {'lr': 0.0003258835517217361, 'samples': 4087488, 'steps': 21288, 'loss/train': 1.6051849722862244} 01/29/2022 16:16:04 - INFO - codeparrot_training - Step 21289: {'lr': 0.0003258679610695556, 'samples': 4087680, 'steps': 21289, 'loss/train': 2.135962128639221} 01/29/2022 16:16:09 - INFO - codeparrot_training - Step 21290: {'lr': 0.000325852370092381, 'samples': 4087872, 'steps': 21290, 'loss/train': 2.394187331199646} 01/29/2022 16:16:13 - INFO - codeparrot_training - Step 21291: {'lr': 0.00032583677879027877, 'samples': 4088064, 'steps': 21291, 'loss/train': 1.81588876247406} 01/29/2022 16:16:17 - INFO - codeparrot_training - Step 21292: {'lr': 0.00032582118716331594, 'samples': 4088256, 'steps': 21292, 'loss/train': 1.7176986336708069} 01/29/2022 16:16:22 - INFO - codeparrot_training - Step 21293: {'lr': 0.00032580559521155916, 'samples': 4088448, 'steps': 21293, 'loss/train': 1.6786365509033203} 01/29/2022 16:16:26 - INFO - codeparrot_training - Step 21294: {'lr': 0.00032579000293507537, 'samples': 4088640, 'steps': 21294, 'loss/train': 1.8271111249923706} 01/29/2022 16:16:31 - INFO - codeparrot_training - Step 21295: {'lr': 0.0003257744103339312, 'samples': 4088832, 'steps': 21295, 'loss/train': 1.9345554113388062} 01/29/2022 16:16:36 - INFO - codeparrot_training - Step 21296: {'lr': 0.00032575881740819353, 'samples': 4089024, 'steps': 21296, 'loss/train': 1.87902170419693} 01/29/2022 16:16:40 - INFO - codeparrot_training - Step 21297: {'lr': 0.0003257432241579291, 'samples': 4089216, 'steps': 21297, 'loss/train': 1.7686627507209778} 01/29/2022 16:16:44 - INFO - codeparrot_training - Step 21298: {'lr': 0.00032572763058320484, 'samples': 4089408, 'steps': 21298, 'loss/train': 1.200745850801468} 01/29/2022 16:16:48 - INFO - codeparrot_training - Step 21299: {'lr': 0.00032571203668408744, 'samples': 4089600, 'steps': 21299, 'loss/train': 1.506136178970337} 01/29/2022 16:16:56 - INFO - codeparrot_training - Step 21300: {'lr': 0.0003256964424606437, 'samples': 4089792, 'steps': 21300, 'loss/train': 0.45515888929367065} 01/29/2022 16:17:00 - INFO - codeparrot_training - Step 21301: {'lr': 0.0003256808479129404, 'samples': 4089984, 'steps': 21301, 'loss/train': 1.0094176232814789} 01/29/2022 16:17:04 - INFO - codeparrot_training - Step 21302: {'lr': 0.0003256652530410444, 'samples': 4090176, 'steps': 21302, 'loss/train': 2.5749022364616394} 01/29/2022 16:17:08 - INFO - codeparrot_training - Step 21303: {'lr': 0.00032564965784502255, 'samples': 4090368, 'steps': 21303, 'loss/train': 1.3498204350471497} 01/29/2022 16:17:12 - INFO - codeparrot_training - Step 21304: {'lr': 0.0003256340623249415, 'samples': 4090560, 'steps': 21304, 'loss/train': 2.682082235813141} 01/29/2022 16:17:18 - INFO - codeparrot_training - Step 21305: {'lr': 0.00032561846648086816, 'samples': 4090752, 'steps': 21305, 'loss/train': 1.6945096850395203} 01/29/2022 16:17:22 - INFO - codeparrot_training - Step 21306: {'lr': 0.0003256028703128693, 'samples': 4090944, 'steps': 21306, 'loss/train': 1.7163337469100952} 01/29/2022 16:17:26 - INFO - codeparrot_training - Step 21307: {'lr': 0.0003255872738210118, 'samples': 4091136, 'steps': 21307, 'loss/train': 1.9161824584007263} 01/29/2022 16:17:31 - INFO - codeparrot_training - Step 21308: {'lr': 0.0003255716770053624, 'samples': 4091328, 'steps': 21308, 'loss/train': 1.404287338256836} 01/29/2022 16:17:35 - INFO - codeparrot_training - Step 21309: {'lr': 0.0003255560798659879, 'samples': 4091520, 'steps': 21309, 'loss/train': 0.623766615986824} 01/29/2022 16:17:42 - INFO - codeparrot_training - Step 21310: {'lr': 0.0003255404824029552, 'samples': 4091712, 'steps': 21310, 'loss/train': 2.649457633495331} 01/29/2022 16:17:47 - INFO - codeparrot_training - Step 21311: {'lr': 0.00032552488461633103, 'samples': 4091904, 'steps': 21311, 'loss/train': 1.6968857645988464} 01/29/2022 16:17:51 - INFO - codeparrot_training - Step 21312: {'lr': 0.00032550928650618225, 'samples': 4092096, 'steps': 21312, 'loss/train': 2.7511462569236755} 01/29/2022 16:17:55 - INFO - codeparrot_training - Step 21313: {'lr': 0.0003254936880725757, 'samples': 4092288, 'steps': 21313, 'loss/train': 2.239277958869934} 01/29/2022 16:17:59 - INFO - codeparrot_training - Step 21314: {'lr': 0.0003254780893155781, 'samples': 4092480, 'steps': 21314, 'loss/train': 2.0013129115104675} 01/29/2022 16:18:05 - INFO - codeparrot_training - Step 21315: {'lr': 0.00032546249023525636, 'samples': 4092672, 'steps': 21315, 'loss/train': 1.6697033643722534} 01/29/2022 16:18:09 - INFO - codeparrot_training - Step 21316: {'lr': 0.00032544689083167727, 'samples': 4092864, 'steps': 21316, 'loss/train': 1.7999056577682495} 01/29/2022 16:18:13 - INFO - codeparrot_training - Step 21317: {'lr': 0.00032543129110490764, 'samples': 4093056, 'steps': 21317, 'loss/train': 1.527424693107605} 01/29/2022 16:18:17 - INFO - codeparrot_training - Step 21318: {'lr': 0.00032541569105501433, 'samples': 4093248, 'steps': 21318, 'loss/train': 1.5211613774299622} 01/29/2022 16:18:22 - INFO - codeparrot_training - Step 21319: {'lr': 0.0003254000906820642, 'samples': 4093440, 'steps': 21319, 'loss/train': 1.4109130203723907} 01/29/2022 16:18:27 - INFO - codeparrot_training - Step 21320: {'lr': 0.0003253844899861239, 'samples': 4093632, 'steps': 21320, 'loss/train': 1.9668912291526794} 01/29/2022 16:18:31 - INFO - codeparrot_training - Step 21321: {'lr': 0.00032536888896726046, 'samples': 4093824, 'steps': 21321, 'loss/train': 0.8641357719898224} 01/29/2022 16:18:35 - INFO - codeparrot_training - Step 21322: {'lr': 0.00032535328762554064, 'samples': 4094016, 'steps': 21322, 'loss/train': 2.24845290184021} 01/29/2022 16:18:40 - INFO - codeparrot_training - Step 21323: {'lr': 0.00032533768596103123, 'samples': 4094208, 'steps': 21323, 'loss/train': 1.443838119506836} 01/29/2022 16:18:44 - INFO - codeparrot_training - Step 21324: {'lr': 0.0003253220839737991, 'samples': 4094400, 'steps': 21324, 'loss/train': 1.7982171177864075} 01/29/2022 16:18:51 - INFO - codeparrot_training - Step 21325: {'lr': 0.00032530648166391115, 'samples': 4094592, 'steps': 21325, 'loss/train': 1.7446029782295227} 01/29/2022 16:18:55 - INFO - codeparrot_training - Step 21326: {'lr': 0.00032529087903143407, 'samples': 4094784, 'steps': 21326, 'loss/train': 2.3090014457702637} 01/29/2022 16:18:59 - INFO - codeparrot_training - Step 21327: {'lr': 0.00032527527607643475, 'samples': 4094976, 'steps': 21327, 'loss/train': 1.6037260293960571} 01/29/2022 16:19:03 - INFO - codeparrot_training - Step 21328: {'lr': 0.00032525967279898015, 'samples': 4095168, 'steps': 21328, 'loss/train': 1.9234054684638977} 01/29/2022 16:19:08 - INFO - codeparrot_training - Step 21329: {'lr': 0.0003252440691991369, 'samples': 4095360, 'steps': 21329, 'loss/train': 1.1596773862838745} 01/29/2022 16:19:13 - INFO - codeparrot_training - Step 21330: {'lr': 0.000325228465276972, 'samples': 4095552, 'steps': 21330, 'loss/train': 3.2502490282058716} 01/29/2022 16:19:18 - INFO - codeparrot_training - Step 21331: {'lr': 0.00032521286103255224, 'samples': 4095744, 'steps': 21331, 'loss/train': 1.8466964364051819} 01/29/2022 16:19:22 - INFO - codeparrot_training - Step 21332: {'lr': 0.0003251972564659445, 'samples': 4095936, 'steps': 21332, 'loss/train': 1.4877787828445435} 01/29/2022 16:19:26 - INFO - codeparrot_training - Step 21333: {'lr': 0.00032518165157721554, 'samples': 4096128, 'steps': 21333, 'loss/train': 1.5264753699302673} 01/29/2022 16:19:31 - INFO - codeparrot_training - Step 21334: {'lr': 0.00032516604636643234, 'samples': 4096320, 'steps': 21334, 'loss/train': 2.010649859905243} 01/29/2022 16:19:36 - INFO - codeparrot_training - Step 21335: {'lr': 0.00032515044083366153, 'samples': 4096512, 'steps': 21335, 'loss/train': 2.503872513771057} 01/29/2022 16:19:40 - INFO - codeparrot_training - Step 21336: {'lr': 0.0003251348349789702, 'samples': 4096704, 'steps': 21336, 'loss/train': 2.106466591358185} 01/29/2022 16:19:44 - INFO - codeparrot_training - Step 21337: {'lr': 0.00032511922880242505, 'samples': 4096896, 'steps': 21337, 'loss/train': 2.0116966366767883} 01/29/2022 16:19:48 - INFO - codeparrot_training - Step 21338: {'lr': 0.00032510362230409295, 'samples': 4097088, 'steps': 21338, 'loss/train': 1.389072060585022} 01/29/2022 16:19:54 - INFO - codeparrot_training - Step 21339: {'lr': 0.0003250880154840408, 'samples': 4097280, 'steps': 21339, 'loss/train': 1.0529972612857819} 01/29/2022 16:19:58 - INFO - codeparrot_training - Step 21340: {'lr': 0.0003250724083423355, 'samples': 4097472, 'steps': 21340, 'loss/train': 0.6684716641902924} 01/29/2022 16:20:02 - INFO - codeparrot_training - Step 21341: {'lr': 0.00032505680087904375, 'samples': 4097664, 'steps': 21341, 'loss/train': 1.1481914520263672} 01/29/2022 16:20:06 - INFO - codeparrot_training - Step 21342: {'lr': 0.0003250411930942326, 'samples': 4097856, 'steps': 21342, 'loss/train': 1.0113745629787445} 01/29/2022 16:20:11 - INFO - codeparrot_training - Step 21343: {'lr': 0.00032502558498796876, 'samples': 4098048, 'steps': 21343, 'loss/train': 1.8014282584190369} 01/29/2022 16:20:18 - INFO - codeparrot_training - Step 21344: {'lr': 0.00032500997656031907, 'samples': 4098240, 'steps': 21344, 'loss/train': 1.8896387815475464} 01/29/2022 16:20:22 - INFO - codeparrot_training - Step 21345: {'lr': 0.0003249943678113505, 'samples': 4098432, 'steps': 21345, 'loss/train': 0.5616430342197418} 01/29/2022 16:20:27 - INFO - codeparrot_training - Step 21346: {'lr': 0.00032497875874112995, 'samples': 4098624, 'steps': 21346, 'loss/train': 2.6124390363693237} 01/29/2022 16:20:31 - INFO - codeparrot_training - Step 21347: {'lr': 0.0003249631493497241, 'samples': 4098816, 'steps': 21347, 'loss/train': 1.6549142003059387} 01/29/2022 16:20:35 - INFO - codeparrot_training - Step 21348: {'lr': 0.0003249475396371999, 'samples': 4099008, 'steps': 21348, 'loss/train': 2.1245368123054504} 01/29/2022 16:20:40 - INFO - codeparrot_training - Step 21349: {'lr': 0.00032493192960362437, 'samples': 4099200, 'steps': 21349, 'loss/train': 1.2734472155570984} 01/29/2022 16:20:45 - INFO - codeparrot_training - Step 21350: {'lr': 0.00032491631924906416, 'samples': 4099392, 'steps': 21350, 'loss/train': 1.7758451700210571} 01/29/2022 16:20:49 - INFO - codeparrot_training - Step 21351: {'lr': 0.0003249007085735863, 'samples': 4099584, 'steps': 21351, 'loss/train': 1.2801771461963654} 01/29/2022 16:20:53 - INFO - codeparrot_training - Step 21352: {'lr': 0.0003248850975772575, 'samples': 4099776, 'steps': 21352, 'loss/train': 1.6539614796638489} 01/29/2022 16:20:57 - INFO - codeparrot_training - Step 21353: {'lr': 0.00032486948626014476, 'samples': 4099968, 'steps': 21353, 'loss/train': 1.3015233278274536} 01/29/2022 16:21:05 - INFO - codeparrot_training - Step 21354: {'lr': 0.00032485387462231484, 'samples': 4100160, 'steps': 21354, 'loss/train': 1.986818790435791} 01/29/2022 16:21:09 - INFO - codeparrot_training - Step 21355: {'lr': 0.0003248382626638348, 'samples': 4100352, 'steps': 21355, 'loss/train': 2.6929669976234436} 01/29/2022 16:21:13 - INFO - codeparrot_training - Step 21356: {'lr': 0.0003248226503847714, 'samples': 4100544, 'steps': 21356, 'loss/train': 1.4608545005321503} 01/29/2022 16:21:17 - INFO - codeparrot_training - Step 21357: {'lr': 0.00032480703778519146, 'samples': 4100736, 'steps': 21357, 'loss/train': 2.3198275566101074} 01/29/2022 16:21:22 - INFO - codeparrot_training - Step 21358: {'lr': 0.00032479142486516193, 'samples': 4100928, 'steps': 21358, 'loss/train': 2.026347041130066} 01/29/2022 16:21:27 - INFO - codeparrot_training - Step 21359: {'lr': 0.00032477581162474974, 'samples': 4101120, 'steps': 21359, 'loss/train': 1.4135409593582153} 01/29/2022 16:21:31 - INFO - codeparrot_training - Step 21360: {'lr': 0.0003247601980640217, 'samples': 4101312, 'steps': 21360, 'loss/train': 1.4448538720607758} 01/29/2022 16:21:35 - INFO - codeparrot_training - Step 21361: {'lr': 0.0003247445841830446, 'samples': 4101504, 'steps': 21361, 'loss/train': 1.8870216608047485} 01/29/2022 16:21:40 - INFO - codeparrot_training - Step 21362: {'lr': 0.0003247289699818856, 'samples': 4101696, 'steps': 21362, 'loss/train': 1.0722309350967407} 01/29/2022 16:21:44 - INFO - codeparrot_training - Step 21363: {'lr': 0.0003247133554606113, 'samples': 4101888, 'steps': 21363, 'loss/train': 1.71693617105484} 01/29/2022 16:21:49 - INFO - codeparrot_training - Step 21364: {'lr': 0.0003246977406192888, 'samples': 4102080, 'steps': 21364, 'loss/train': 0.5811547189950943} 01/29/2022 16:21:53 - INFO - codeparrot_training - Step 21365: {'lr': 0.00032468212545798484, 'samples': 4102272, 'steps': 21365, 'loss/train': 1.3477471768856049} 01/29/2022 16:21:58 - INFO - codeparrot_training - Step 21366: {'lr': 0.0003246665099767664, 'samples': 4102464, 'steps': 21366, 'loss/train': 1.7991503477096558} 01/29/2022 16:22:02 - INFO - codeparrot_training - Step 21367: {'lr': 0.0003246508941757004, 'samples': 4102656, 'steps': 21367, 'loss/train': 1.6478728652000427} 01/29/2022 16:22:06 - INFO - codeparrot_training - Step 21368: {'lr': 0.0003246352780548536, 'samples': 4102848, 'steps': 21368, 'loss/train': 1.40075021982193} 01/29/2022 16:22:13 - INFO - codeparrot_training - Step 21369: {'lr': 0.0003246196616142929, 'samples': 4103040, 'steps': 21369, 'loss/train': 0.7184461355209351} 01/29/2022 16:22:17 - INFO - codeparrot_training - Step 21370: {'lr': 0.0003246040448540854, 'samples': 4103232, 'steps': 21370, 'loss/train': 1.7508655786514282} 01/29/2022 16:22:22 - INFO - codeparrot_training - Step 21371: {'lr': 0.00032458842777429776, 'samples': 4103424, 'steps': 21371, 'loss/train': 1.9958050847053528} 01/29/2022 16:22:26 - INFO - codeparrot_training - Step 21372: {'lr': 0.00032457281037499706, 'samples': 4103616, 'steps': 21372, 'loss/train': 1.5087352395057678} 01/29/2022 16:22:30 - INFO - codeparrot_training - Step 21373: {'lr': 0.0003245571926562501, 'samples': 4103808, 'steps': 21373, 'loss/train': 1.2303811311721802} 01/29/2022 16:22:35 - INFO - codeparrot_training - Step 21374: {'lr': 0.0003245415746181237, 'samples': 4104000, 'steps': 21374, 'loss/train': 2.3281479477882385} 01/29/2022 16:22:40 - INFO - codeparrot_training - Step 21375: {'lr': 0.00032452595626068504, 'samples': 4104192, 'steps': 21375, 'loss/train': 1.472451388835907} 01/29/2022 16:22:44 - INFO - codeparrot_training - Step 21376: {'lr': 0.0003245103375840007, 'samples': 4104384, 'steps': 21376, 'loss/train': 1.422315537929535} 01/29/2022 16:22:48 - INFO - codeparrot_training - Step 21377: {'lr': 0.0003244947185881378, 'samples': 4104576, 'steps': 21377, 'loss/train': 2.019866466522217} 01/29/2022 16:22:52 - INFO - codeparrot_training - Step 21378: {'lr': 0.00032447909927316317, 'samples': 4104768, 'steps': 21378, 'loss/train': 1.8694446086883545} 01/29/2022 16:22:58 - INFO - codeparrot_training - Step 21379: {'lr': 0.00032446347963914376, 'samples': 4104960, 'steps': 21379, 'loss/train': 1.0709823668003082} 01/29/2022 16:23:02 - INFO - codeparrot_training - Step 21380: {'lr': 0.0003244478596861464, 'samples': 4105152, 'steps': 21380, 'loss/train': 1.4108982682228088} 01/29/2022 16:23:06 - INFO - codeparrot_training - Step 21381: {'lr': 0.000324432239414238, 'samples': 4105344, 'steps': 21381, 'loss/train': 1.0688588619232178} 01/29/2022 16:23:10 - INFO - codeparrot_training - Step 21382: {'lr': 0.0003244166188234856, 'samples': 4105536, 'steps': 21382, 'loss/train': 0.6948254406452179} 01/29/2022 16:23:17 - INFO - codeparrot_training - Step 21383: {'lr': 0.000324400997913956, 'samples': 4105728, 'steps': 21383, 'loss/train': 1.5123406648635864} 01/29/2022 16:23:22 - INFO - codeparrot_training - Step 21384: {'lr': 0.0003243853766857162, 'samples': 4105920, 'steps': 21384, 'loss/train': 1.0747803747653961} 01/29/2022 16:23:26 - INFO - codeparrot_training - Step 21385: {'lr': 0.000324369755138833, 'samples': 4106112, 'steps': 21385, 'loss/train': 1.5405950546264648} 01/29/2022 16:23:30 - INFO - codeparrot_training - Step 21386: {'lr': 0.0003243541332733734, 'samples': 4106304, 'steps': 21386, 'loss/train': 1.9160329699516296} 01/29/2022 16:23:34 - INFO - codeparrot_training - Step 21387: {'lr': 0.00032433851108940433, 'samples': 4106496, 'steps': 21387, 'loss/train': 0.5556090474128723} 01/29/2022 16:23:40 - INFO - codeparrot_training - Step 21388: {'lr': 0.0003243228885869927, 'samples': 4106688, 'steps': 21388, 'loss/train': 1.885901391506195} 01/29/2022 16:23:44 - INFO - codeparrot_training - Step 21389: {'lr': 0.0003243072657662054, 'samples': 4106880, 'steps': 21389, 'loss/train': 2.3754372596740723} 01/29/2022 16:23:48 - INFO - codeparrot_training - Step 21390: {'lr': 0.00032429164262710934, 'samples': 4107072, 'steps': 21390, 'loss/train': 0.05633929744362831} 01/29/2022 16:23:53 - INFO - codeparrot_training - Step 21391: {'lr': 0.0003242760191697714, 'samples': 4107264, 'steps': 21391, 'loss/train': 1.1967202126979828} 01/29/2022 16:23:57 - INFO - codeparrot_training - Step 21392: {'lr': 0.0003242603953942587, 'samples': 4107456, 'steps': 21392, 'loss/train': 1.2972411811351776} 01/29/2022 16:24:02 - INFO - codeparrot_training - Step 21393: {'lr': 0.00032424477130063806, 'samples': 4107648, 'steps': 21393, 'loss/train': 1.5769566893577576} 01/29/2022 16:24:06 - INFO - codeparrot_training - Step 21394: {'lr': 0.0003242291468889763, 'samples': 4107840, 'steps': 21394, 'loss/train': 1.2232670187950134} 01/29/2022 16:24:11 - INFO - codeparrot_training - Step 21395: {'lr': 0.0003242135221593405, 'samples': 4108032, 'steps': 21395, 'loss/train': 1.0863144099712372} 01/29/2022 16:24:15 - INFO - codeparrot_training - Step 21396: {'lr': 0.0003241978971117976, 'samples': 4108224, 'steps': 21396, 'loss/train': 1.5255310535430908} 01/29/2022 16:24:19 - INFO - codeparrot_training - Step 21397: {'lr': 0.0003241822717464144, 'samples': 4108416, 'steps': 21397, 'loss/train': 1.914526641368866} 01/29/2022 16:24:26 - INFO - codeparrot_training - Step 21398: {'lr': 0.0003241666460632579, 'samples': 4108608, 'steps': 21398, 'loss/train': 1.6157037019729614} 01/29/2022 16:24:31 - INFO - codeparrot_training - Step 21399: {'lr': 0.00032415102006239506, 'samples': 4108800, 'steps': 21399, 'loss/train': 1.7980520725250244} 01/29/2022 16:24:35 - INFO - codeparrot_training - Step 21400: {'lr': 0.0003241353937438927, 'samples': 4108992, 'steps': 21400, 'loss/train': 1.4414768815040588} 01/29/2022 16:24:39 - INFO - codeparrot_training - Step 21401: {'lr': 0.000324119767107818, 'samples': 4109184, 'steps': 21401, 'loss/train': 1.0084426403045654} 01/29/2022 16:24:43 - INFO - codeparrot_training - Step 21402: {'lr': 0.0003241041401542377, 'samples': 4109376, 'steps': 21402, 'loss/train': 1.7544010877609253} 01/29/2022 16:24:49 - INFO - codeparrot_training - Step 21403: {'lr': 0.0003240885128832188, 'samples': 4109568, 'steps': 21403, 'loss/train': 1.6077577471733093} 01/29/2022 16:24:53 - INFO - codeparrot_training - Step 21404: {'lr': 0.0003240728852948281, 'samples': 4109760, 'steps': 21404, 'loss/train': 1.2326406240463257} 01/29/2022 16:24:57 - INFO - codeparrot_training - Step 21405: {'lr': 0.00032405725738913284, 'samples': 4109952, 'steps': 21405, 'loss/train': 1.9303720593452454} 01/29/2022 16:25:01 - INFO - codeparrot_training - Step 21406: {'lr': 0.0003240416291661998, 'samples': 4110144, 'steps': 21406, 'loss/train': 1.8361369371414185} 01/29/2022 16:25:06 - INFO - codeparrot_training - Step 21407: {'lr': 0.0003240260006260959, 'samples': 4110336, 'steps': 21407, 'loss/train': 1.8576586246490479} 01/29/2022 16:25:11 - INFO - codeparrot_training - Step 21408: {'lr': 0.0003240103717688881, 'samples': 4110528, 'steps': 21408, 'loss/train': 0.06335282698273659} 01/29/2022 16:25:15 - INFO - codeparrot_training - Step 21409: {'lr': 0.00032399474259464336, 'samples': 4110720, 'steps': 21409, 'loss/train': 1.5992335081100464} 01/29/2022 16:25:20 - INFO - codeparrot_training - Step 21410: {'lr': 0.0003239791131034287, 'samples': 4110912, 'steps': 21410, 'loss/train': 1.6112449765205383} 01/29/2022 16:25:24 - INFO - codeparrot_training - Step 21411: {'lr': 0.00032396348329531097, 'samples': 4111104, 'steps': 21411, 'loss/train': 1.1893548667430878} 01/29/2022 16:25:28 - INFO - codeparrot_training - Step 21412: {'lr': 0.0003239478531703571, 'samples': 4111296, 'steps': 21412, 'loss/train': 0.946617990732193} 01/29/2022 16:25:35 - INFO - codeparrot_training - Step 21413: {'lr': 0.0003239322227286343, 'samples': 4111488, 'steps': 21413, 'loss/train': 1.807542622089386} 01/29/2022 16:25:39 - INFO - codeparrot_training - Step 21414: {'lr': 0.0003239165919702092, 'samples': 4111680, 'steps': 21414, 'loss/train': 1.7610695958137512} 01/29/2022 16:25:44 - INFO - codeparrot_training - Step 21415: {'lr': 0.0003239009608951489, 'samples': 4111872, 'steps': 21415, 'loss/train': 1.64895361661911} 01/29/2022 16:25:48 - INFO - codeparrot_training - Step 21416: {'lr': 0.0003238853295035203, 'samples': 4112064, 'steps': 21416, 'loss/train': 1.0766386091709137} 01/29/2022 16:25:52 - INFO - codeparrot_training - Step 21417: {'lr': 0.0003238696977953905, 'samples': 4112256, 'steps': 21417, 'loss/train': 0.693704143166542} 01/29/2022 16:25:57 - INFO - codeparrot_training - Step 21418: {'lr': 0.0003238540657708263, 'samples': 4112448, 'steps': 21418, 'loss/train': 1.6927154660224915} 01/29/2022 16:26:02 - INFO - codeparrot_training - Step 21419: {'lr': 0.0003238384334298948, 'samples': 4112640, 'steps': 21419, 'loss/train': 1.6143254041671753} 01/29/2022 16:26:06 - INFO - codeparrot_training - Step 21420: {'lr': 0.0003238228007726628, 'samples': 4112832, 'steps': 21420, 'loss/train': 1.6385196447372437} 01/29/2022 16:26:10 - INFO - codeparrot_training - Step 21421: {'lr': 0.00032380716779919745, 'samples': 4113024, 'steps': 21421, 'loss/train': 1.6275955438613892} 01/29/2022 16:26:14 - INFO - codeparrot_training - Step 21422: {'lr': 0.00032379153450956555, 'samples': 4113216, 'steps': 21422, 'loss/train': 1.8384223580360413} 01/29/2022 16:26:20 - INFO - codeparrot_training - Step 21423: {'lr': 0.0003237759009038342, 'samples': 4113408, 'steps': 21423, 'loss/train': 1.3817853927612305} 01/29/2022 16:26:24 - INFO - codeparrot_training - Step 21424: {'lr': 0.0003237602669820703, 'samples': 4113600, 'steps': 21424, 'loss/train': 2.0086169242858887} 01/29/2022 16:26:28 - INFO - codeparrot_training - Step 21425: {'lr': 0.00032374463274434097, 'samples': 4113792, 'steps': 21425, 'loss/train': 1.784329891204834} 01/29/2022 16:26:32 - INFO - codeparrot_training - Step 21426: {'lr': 0.0003237289981907129, 'samples': 4113984, 'steps': 21426, 'loss/train': 1.8286004662513733} 01/29/2022 16:26:37 - INFO - codeparrot_training - Step 21427: {'lr': 0.00032371336332125323, 'samples': 4114176, 'steps': 21427, 'loss/train': 1.46690833568573} 01/29/2022 16:26:44 - INFO - codeparrot_training - Step 21428: {'lr': 0.0003236977281360289, 'samples': 4114368, 'steps': 21428, 'loss/train': 1.4152776002883911} 01/29/2022 16:26:48 - INFO - codeparrot_training - Step 21429: {'lr': 0.00032368209263510694, 'samples': 4114560, 'steps': 21429, 'loss/train': 1.8617717027664185} 01/29/2022 16:26:52 - INFO - codeparrot_training - Step 21430: {'lr': 0.00032366645681855435, 'samples': 4114752, 'steps': 21430, 'loss/train': 1.4479272365570068} 01/29/2022 16:26:56 - INFO - codeparrot_training - Step 21431: {'lr': 0.0003236508206864379, 'samples': 4114944, 'steps': 21431, 'loss/train': 1.5281794667243958} 01/29/2022 16:27:01 - INFO - codeparrot_training - Step 21432: {'lr': 0.0003236351842388249, 'samples': 4115136, 'steps': 21432, 'loss/train': 1.868825912475586} 01/29/2022 16:27:06 - INFO - codeparrot_training - Step 21433: {'lr': 0.00032361954747578203, 'samples': 4115328, 'steps': 21433, 'loss/train': 1.9782485961914062} 01/29/2022 16:27:10 - INFO - codeparrot_training - Step 21434: {'lr': 0.00032360391039737646, 'samples': 4115520, 'steps': 21434, 'loss/train': 1.3847224116325378} 01/29/2022 16:27:15 - INFO - codeparrot_training - Step 21435: {'lr': 0.00032358827300367504, 'samples': 4115712, 'steps': 21435, 'loss/train': 1.0670125186443329} 01/29/2022 16:27:19 - INFO - codeparrot_training - Step 21436: {'lr': 0.0003235726352947449, 'samples': 4115904, 'steps': 21436, 'loss/train': 1.751111626625061} 01/29/2022 16:27:23 - INFO - codeparrot_training - Step 21437: {'lr': 0.0003235569972706529, 'samples': 4116096, 'steps': 21437, 'loss/train': 2.0019482374191284} 01/29/2022 16:27:30 - INFO - codeparrot_training - Step 21438: {'lr': 0.00032354135893146614, 'samples': 4116288, 'steps': 21438, 'loss/train': 1.5327345728874207} 01/29/2022 16:27:35 - INFO - codeparrot_training - Step 21439: {'lr': 0.0003235257202772515, 'samples': 4116480, 'steps': 21439, 'loss/train': 1.6024693250656128} 01/29/2022 16:27:39 - INFO - codeparrot_training - Step 21440: {'lr': 0.000323510081308076, 'samples': 4116672, 'steps': 21440, 'loss/train': 1.2045646011829376} 01/29/2022 16:27:43 - INFO - codeparrot_training - Step 21441: {'lr': 0.00032349444202400666, 'samples': 4116864, 'steps': 21441, 'loss/train': 1.8200435042381287} 01/29/2022 16:27:47 - INFO - codeparrot_training - Step 21442: {'lr': 0.0003234788024251105, 'samples': 4117056, 'steps': 21442, 'loss/train': 2.0016610622406006} 01/29/2022 16:27:53 - INFO - codeparrot_training - Step 21443: {'lr': 0.00032346316251145445, 'samples': 4117248, 'steps': 21443, 'loss/train': 1.9649899005889893} 01/29/2022 16:27:57 - INFO - codeparrot_training - Step 21444: {'lr': 0.0003234475222831056, 'samples': 4117440, 'steps': 21444, 'loss/train': 1.996222972869873} 01/29/2022 16:28:01 - INFO - codeparrot_training - Step 21445: {'lr': 0.0003234318817401309, 'samples': 4117632, 'steps': 21445, 'loss/train': 0.6008615344762802} 01/29/2022 16:28:05 - INFO - codeparrot_training - Step 21446: {'lr': 0.00032341624088259727, 'samples': 4117824, 'steps': 21446, 'loss/train': 1.429191380739212} 01/29/2022 16:28:10 - INFO - codeparrot_training - Step 21447: {'lr': 0.0003234005997105718, 'samples': 4118016, 'steps': 21447, 'loss/train': 1.7067634463310242} 01/29/2022 16:28:15 - INFO - codeparrot_training - Step 21448: {'lr': 0.0003233849582241214, 'samples': 4118208, 'steps': 21448, 'loss/train': 1.4728859961032867} 01/29/2022 16:28:19 - INFO - codeparrot_training - Step 21449: {'lr': 0.0003233693164233132, 'samples': 4118400, 'steps': 21449, 'loss/train': 1.691517412662506} 01/29/2022 16:28:23 - INFO - codeparrot_training - Step 21450: {'lr': 0.00032335367430821416, 'samples': 4118592, 'steps': 21450, 'loss/train': 1.4900348782539368} 01/29/2022 16:28:28 - INFO - codeparrot_training - Step 21451: {'lr': 0.00032333803187889135, 'samples': 4118784, 'steps': 21451, 'loss/train': 0.8325344324111938} 01/29/2022 16:28:32 - INFO - codeparrot_training - Step 21452: {'lr': 0.0003233223891354116, 'samples': 4118976, 'steps': 21452, 'loss/train': 1.3065834045410156} 01/29/2022 16:28:37 - INFO - codeparrot_training - Step 21453: {'lr': 0.000323306746077842, 'samples': 4119168, 'steps': 21453, 'loss/train': 0.784011572599411} 01/29/2022 16:28:41 - INFO - codeparrot_training - Step 21454: {'lr': 0.00032329110270624956, 'samples': 4119360, 'steps': 21454, 'loss/train': 0.7759304344654083} 01/29/2022 16:28:46 - INFO - codeparrot_training - Step 21455: {'lr': 0.00032327545902070137, 'samples': 4119552, 'steps': 21455, 'loss/train': 1.6490651965141296} 01/29/2022 16:28:50 - INFO - codeparrot_training - Step 21456: {'lr': 0.00032325981502126435, 'samples': 4119744, 'steps': 21456, 'loss/train': 2.20273357629776} 01/29/2022 16:28:54 - INFO - codeparrot_training - Step 21457: {'lr': 0.0003232441707080056, 'samples': 4119936, 'steps': 21457, 'loss/train': 1.9138042330741882} 01/29/2022 16:29:02 - INFO - codeparrot_training - Step 21458: {'lr': 0.00032322852608099203, 'samples': 4120128, 'steps': 21458, 'loss/train': 1.8285738229751587} 01/29/2022 16:29:06 - INFO - codeparrot_training - Step 21459: {'lr': 0.00032321288114029074, 'samples': 4120320, 'steps': 21459, 'loss/train': 1.7413558959960938} 01/29/2022 16:29:10 - INFO - codeparrot_training - Step 21460: {'lr': 0.00032319723588596875, 'samples': 4120512, 'steps': 21460, 'loss/train': 1.925997018814087} 01/29/2022 16:29:14 - INFO - codeparrot_training - Step 21461: {'lr': 0.00032318159031809293, 'samples': 4120704, 'steps': 21461, 'loss/train': 1.0272433161735535} 01/29/2022 16:29:19 - INFO - codeparrot_training - Step 21462: {'lr': 0.00032316594443673047, 'samples': 4120896, 'steps': 21462, 'loss/train': 3.277416229248047} 01/29/2022 16:29:24 - INFO - codeparrot_training - Step 21463: {'lr': 0.0003231502982419483, 'samples': 4121088, 'steps': 21463, 'loss/train': 1.435203641653061} 01/29/2022 16:29:28 - INFO - codeparrot_training - Step 21464: {'lr': 0.0003231346517338135, 'samples': 4121280, 'steps': 21464, 'loss/train': 1.2918032705783844} 01/29/2022 16:29:32 - INFO - codeparrot_training - Step 21465: {'lr': 0.0003231190049123931, 'samples': 4121472, 'steps': 21465, 'loss/train': 1.5339131355285645} 01/29/2022 16:29:37 - INFO - codeparrot_training - Step 21466: {'lr': 0.00032310335777775413, 'samples': 4121664, 'steps': 21466, 'loss/train': 2.1032556295394897} 01/29/2022 16:29:41 - INFO - codeparrot_training - Step 21467: {'lr': 0.00032308771032996353, 'samples': 4121856, 'steps': 21467, 'loss/train': 1.461120218038559} 01/29/2022 16:29:46 - INFO - codeparrot_training - Step 21468: {'lr': 0.0003230720625690884, 'samples': 4122048, 'steps': 21468, 'loss/train': 1.7283784747123718} 01/29/2022 16:29:50 - INFO - codeparrot_training - Step 21469: {'lr': 0.0003230564144951958, 'samples': 4122240, 'steps': 21469, 'loss/train': 1.0158703923225403} 01/29/2022 16:29:55 - INFO - codeparrot_training - Step 21470: {'lr': 0.00032304076610835267, 'samples': 4122432, 'steps': 21470, 'loss/train': 1.0639493465423584} 01/29/2022 16:29:59 - INFO - codeparrot_training - Step 21471: {'lr': 0.0003230251174086261, 'samples': 4122624, 'steps': 21471, 'loss/train': 1.3948386311531067} 01/29/2022 16:30:03 - INFO - codeparrot_training - Step 21472: {'lr': 0.0003230094683960831, 'samples': 4122816, 'steps': 21472, 'loss/train': 2.0831896662712097} 01/29/2022 16:30:10 - INFO - codeparrot_training - Step 21473: {'lr': 0.0003229938190707908, 'samples': 4123008, 'steps': 21473, 'loss/train': 1.9384796619415283} 01/29/2022 16:30:15 - INFO - codeparrot_training - Step 21474: {'lr': 0.00032297816943281605, 'samples': 4123200, 'steps': 21474, 'loss/train': 1.7792959213256836} 01/29/2022 16:30:19 - INFO - codeparrot_training - Step 21475: {'lr': 0.00032296251948222605, 'samples': 4123392, 'steps': 21475, 'loss/train': 0.6867115348577499} 01/29/2022 16:30:23 - INFO - codeparrot_training - Step 21476: {'lr': 0.0003229468692190878, 'samples': 4123584, 'steps': 21476, 'loss/train': 1.7780399322509766} 01/29/2022 16:30:27 - INFO - codeparrot_training - Step 21477: {'lr': 0.00032293121864346823, 'samples': 4123776, 'steps': 21477, 'loss/train': 1.5147700309753418} 01/29/2022 16:30:33 - INFO - codeparrot_training - Step 21478: {'lr': 0.00032291556775543463, 'samples': 4123968, 'steps': 21478, 'loss/train': 1.4182853400707245} 01/29/2022 16:30:37 - INFO - codeparrot_training - Step 21479: {'lr': 0.0003228999165550537, 'samples': 4124160, 'steps': 21479, 'loss/train': 1.7379325032234192} 01/29/2022 16:30:41 - INFO - codeparrot_training - Step 21480: {'lr': 0.0003228842650423929, 'samples': 4124352, 'steps': 21480, 'loss/train': 1.8495397567749023} 01/29/2022 16:30:45 - INFO - codeparrot_training - Step 21481: {'lr': 0.0003228686132175189, 'samples': 4124544, 'steps': 21481, 'loss/train': 1.31673663854599} 01/29/2022 16:30:50 - INFO - codeparrot_training - Step 21482: {'lr': 0.0003228529610804989, 'samples': 4124736, 'steps': 21482, 'loss/train': 2.20517098903656} 01/29/2022 16:30:57 - INFO - codeparrot_training - Step 21483: {'lr': 0.00032283730863140003, 'samples': 4124928, 'steps': 21483, 'loss/train': 1.6145655512809753} 01/29/2022 16:31:01 - INFO - codeparrot_training - Step 21484: {'lr': 0.0003228216558702892, 'samples': 4125120, 'steps': 21484, 'loss/train': 1.7834871411323547} 01/29/2022 16:31:05 - INFO - codeparrot_training - Step 21485: {'lr': 0.00032280600279723355, 'samples': 4125312, 'steps': 21485, 'loss/train': 0.09911185875535011} 01/29/2022 16:31:09 - INFO - codeparrot_training - Step 21486: {'lr': 0.00032279034941230014, 'samples': 4125504, 'steps': 21486, 'loss/train': 1.5547199249267578} 01/29/2022 16:31:14 - INFO - codeparrot_training - Step 21487: {'lr': 0.00032277469571555587, 'samples': 4125696, 'steps': 21487, 'loss/train': 1.6607580184936523} 01/29/2022 16:31:19 - INFO - codeparrot_training - Step 21488: {'lr': 0.0003227590417070679, 'samples': 4125888, 'steps': 21488, 'loss/train': 1.7451483607292175} 01/29/2022 16:31:23 - INFO - codeparrot_training - Step 21489: {'lr': 0.00032274338738690344, 'samples': 4126080, 'steps': 21489, 'loss/train': 1.3138876855373383} 01/29/2022 16:31:27 - INFO - codeparrot_training - Step 21490: {'lr': 0.00032272773275512933, 'samples': 4126272, 'steps': 21490, 'loss/train': 1.2832450568675995} 01/29/2022 16:31:32 - INFO - codeparrot_training - Step 21491: {'lr': 0.0003227120778118127, 'samples': 4126464, 'steps': 21491, 'loss/train': 0.6221323013305664} 01/29/2022 16:31:36 - INFO - codeparrot_training - Step 21492: {'lr': 0.00032269642255702065, 'samples': 4126656, 'steps': 21492, 'loss/train': 0.5143368244171143} 01/29/2022 16:31:42 - INFO - codeparrot_training - Step 21493: {'lr': 0.00032268076699082024, 'samples': 4126848, 'steps': 21493, 'loss/train': 1.7972105741500854} 01/29/2022 16:31:46 - INFO - codeparrot_training - Step 21494: {'lr': 0.0003226651111132784, 'samples': 4127040, 'steps': 21494, 'loss/train': 1.7118337154388428} 01/29/2022 16:31:50 - INFO - codeparrot_training - Step 21495: {'lr': 0.0003226494549244624, 'samples': 4127232, 'steps': 21495, 'loss/train': 0.9519549608230591} 01/29/2022 16:31:55 - INFO - codeparrot_training - Step 21496: {'lr': 0.00032263379842443915, 'samples': 4127424, 'steps': 21496, 'loss/train': 1.8175755143165588} 01/29/2022 16:31:59 - INFO - codeparrot_training - Step 21497: {'lr': 0.0003226181416132758, 'samples': 4127616, 'steps': 21497, 'loss/train': 1.9865802526474} 01/29/2022 16:32:03 - INFO - codeparrot_training - Step 21498: {'lr': 0.00032260248449103937, 'samples': 4127808, 'steps': 21498, 'loss/train': 1.6297295093536377} 01/29/2022 16:32:10 - INFO - codeparrot_training - Step 21499: {'lr': 0.00032258682705779695, 'samples': 4128000, 'steps': 21499, 'loss/train': 2.157749891281128} 01/29/2022 16:32:14 - INFO - codeparrot_training - Step 21500: {'lr': 0.00032257116931361555, 'samples': 4128192, 'steps': 21500, 'loss/train': 1.6212771534919739} 01/29/2022 16:32:19 - INFO - codeparrot_training - Step 21501: {'lr': 0.0003225555112585624, 'samples': 4128384, 'steps': 21501, 'loss/train': 2.166091561317444} 01/29/2022 16:32:23 - INFO - codeparrot_training - Step 21502: {'lr': 0.0003225398528927045, 'samples': 4128576, 'steps': 21502, 'loss/train': 1.6351638436317444} 01/29/2022 16:32:27 - INFO - codeparrot_training - Step 21503: {'lr': 0.00032252419421610883, 'samples': 4128768, 'steps': 21503, 'loss/train': 2.0157074332237244} 01/29/2022 16:32:32 - INFO - codeparrot_training - Step 21504: {'lr': 0.0003225085352288426, 'samples': 4128960, 'steps': 21504, 'loss/train': 1.7170904874801636} 01/29/2022 16:32:37 - INFO - codeparrot_training - Step 21505: {'lr': 0.00032249287593097274, 'samples': 4129152, 'steps': 21505, 'loss/train': 2.183380365371704} 01/29/2022 16:32:41 - INFO - codeparrot_training - Step 21506: {'lr': 0.00032247721632256657, 'samples': 4129344, 'steps': 21506, 'loss/train': 1.2339611649513245} 01/29/2022 16:32:45 - INFO - codeparrot_training - Step 21507: {'lr': 0.0003224615564036908, 'samples': 4129536, 'steps': 21507, 'loss/train': 1.9897438287734985} 01/29/2022 16:32:49 - INFO - codeparrot_training - Step 21508: {'lr': 0.00032244589617441287, 'samples': 4129728, 'steps': 21508, 'loss/train': 1.5992170572280884} 01/29/2022 16:32:55 - INFO - codeparrot_training - Step 21509: {'lr': 0.0003224302356347997, 'samples': 4129920, 'steps': 21509, 'loss/train': 2.080373704433441} 01/29/2022 16:32:59 - INFO - codeparrot_training - Step 21510: {'lr': 0.0003224145747849185, 'samples': 4130112, 'steps': 21510, 'loss/train': 1.2903198301792145} 01/29/2022 16:33:03 - INFO - codeparrot_training - Step 21511: {'lr': 0.0003223989136248361, 'samples': 4130304, 'steps': 21511, 'loss/train': 1.3153952658176422} 01/29/2022 16:33:07 - INFO - codeparrot_training - Step 21512: {'lr': 0.0003223832521546198, 'samples': 4130496, 'steps': 21512, 'loss/train': 1.1970336735248566} 01/29/2022 16:33:12 - INFO - codeparrot_training - Step 21513: {'lr': 0.0003223675903743366, 'samples': 4130688, 'steps': 21513, 'loss/train': 0.6549136787652969} 01/29/2022 16:33:17 - INFO - codeparrot_training - Step 21514: {'lr': 0.0003223519282840537, 'samples': 4130880, 'steps': 21514, 'loss/train': 1.9357255697250366} 01/29/2022 16:33:21 - INFO - codeparrot_training - Step 21515: {'lr': 0.00032233626588383806, 'samples': 4131072, 'steps': 21515, 'loss/train': 0.9292050004005432} 01/29/2022 16:33:25 - INFO - codeparrot_training - Step 21516: {'lr': 0.00032232060317375684, 'samples': 4131264, 'steps': 21516, 'loss/train': 1.595432996749878} 01/29/2022 16:33:29 - INFO - codeparrot_training - Step 21517: {'lr': 0.00032230494015387715, 'samples': 4131456, 'steps': 21517, 'loss/train': 2.556818425655365} 01/29/2022 16:33:37 - INFO - codeparrot_training - Step 21518: {'lr': 0.000322289276824266, 'samples': 4131648, 'steps': 21518, 'loss/train': 0.31152769178152084} 01/29/2022 16:33:41 - INFO - codeparrot_training - Step 21519: {'lr': 0.0003222736131849906, 'samples': 4131840, 'steps': 21519, 'loss/train': 2.1659536957740784} 01/29/2022 16:33:46 - INFO - codeparrot_training - Step 21520: {'lr': 0.0003222579492361179, 'samples': 4132032, 'steps': 21520, 'loss/train': 1.9744389653205872} 01/29/2022 16:33:50 - INFO - codeparrot_training - Step 21521: {'lr': 0.0003222422849777152, 'samples': 4132224, 'steps': 21521, 'loss/train': 1.6978754997253418} 01/29/2022 16:33:54 - INFO - codeparrot_training - Step 21522: {'lr': 0.0003222266204098494, 'samples': 4132416, 'steps': 21522, 'loss/train': 0.34921201318502426} 01/29/2022 16:34:00 - INFO - codeparrot_training - Step 21523: {'lr': 0.0003222109555325877, 'samples': 4132608, 'steps': 21523, 'loss/train': 0.4103868752717972} 01/29/2022 16:34:04 - INFO - codeparrot_training - Step 21524: {'lr': 0.00032219529034599725, 'samples': 4132800, 'steps': 21524, 'loss/train': 1.9095659852027893} 01/29/2022 16:34:08 - INFO - codeparrot_training - Step 21525: {'lr': 0.00032217962485014506, 'samples': 4132992, 'steps': 21525, 'loss/train': 1.4768745303153992} 01/29/2022 16:34:12 - INFO - codeparrot_training - Step 21526: {'lr': 0.0003221639590450983, 'samples': 4133184, 'steps': 21526, 'loss/train': 1.8614246249198914} 01/29/2022 16:34:17 - INFO - codeparrot_training - Step 21527: {'lr': 0.00032214829293092406, 'samples': 4133376, 'steps': 21527, 'loss/train': 0.528283417224884} 01/29/2022 16:34:24 - INFO - codeparrot_training - Step 21528: {'lr': 0.0003221326265076894, 'samples': 4133568, 'steps': 21528, 'loss/train': 2.1594809889793396} 01/29/2022 16:34:28 - INFO - codeparrot_training - Step 21529: {'lr': 0.00032211695977546153, 'samples': 4133760, 'steps': 21529, 'loss/train': 1.9909794330596924} 01/29/2022 16:34:32 - INFO - codeparrot_training - Step 21530: {'lr': 0.0003221012927343075, 'samples': 4133952, 'steps': 21530, 'loss/train': 1.8013553023338318} 01/29/2022 16:34:36 - INFO - codeparrot_training - Step 21531: {'lr': 0.0003220856253842944, 'samples': 4134144, 'steps': 21531, 'loss/train': 1.5486277341842651} 01/29/2022 16:34:41 - INFO - codeparrot_training - Step 21532: {'lr': 0.00032206995772548943, 'samples': 4134336, 'steps': 21532, 'loss/train': 2.3462730646133423} 01/29/2022 16:34:46 - INFO - codeparrot_training - Step 21533: {'lr': 0.00032205428975795955, 'samples': 4134528, 'steps': 21533, 'loss/train': 1.9634159803390503} 01/29/2022 16:34:50 - INFO - codeparrot_training - Step 21534: {'lr': 0.000322038621481772, 'samples': 4134720, 'steps': 21534, 'loss/train': 0.9084018766880035} 01/29/2022 16:34:54 - INFO - codeparrot_training - Step 21535: {'lr': 0.0003220229528969939, 'samples': 4134912, 'steps': 21535, 'loss/train': 1.754385530948639} 01/29/2022 16:34:59 - INFO - codeparrot_training - Step 21536: {'lr': 0.00032200728400369233, 'samples': 4135104, 'steps': 21536, 'loss/train': 1.5636584758758545} 01/29/2022 16:35:03 - INFO - codeparrot_training - Step 21537: {'lr': 0.0003219916148019344, 'samples': 4135296, 'steps': 21537, 'loss/train': 2.02745658159256} 01/29/2022 16:35:08 - INFO - codeparrot_training - Step 21538: {'lr': 0.0003219759452917872, 'samples': 4135488, 'steps': 21538, 'loss/train': 1.6636914610862732} 01/29/2022 16:35:12 - INFO - codeparrot_training - Step 21539: {'lr': 0.000321960275473318, 'samples': 4135680, 'steps': 21539, 'loss/train': 0.5629440397024155} 01/29/2022 16:35:17 - INFO - codeparrot_training - Step 21540: {'lr': 0.0003219446053465938, 'samples': 4135872, 'steps': 21540, 'loss/train': 1.5080233812332153} 01/29/2022 16:35:21 - INFO - codeparrot_training - Step 21541: {'lr': 0.0003219289349116818, 'samples': 4136064, 'steps': 21541, 'loss/train': 0.9030154645442963} 01/29/2022 16:35:25 - INFO - codeparrot_training - Step 21542: {'lr': 0.0003219132641686491, 'samples': 4136256, 'steps': 21542, 'loss/train': 1.3157023787498474} 01/29/2022 16:35:32 - INFO - codeparrot_training - Step 21543: {'lr': 0.0003218975931175627, 'samples': 4136448, 'steps': 21543, 'loss/train': 1.3100630342960358} 01/29/2022 16:35:37 - INFO - codeparrot_training - Step 21544: {'lr': 0.0003218819217584899, 'samples': 4136640, 'steps': 21544, 'loss/train': 1.3927524983882904} 01/29/2022 16:35:41 - INFO - codeparrot_training - Step 21545: {'lr': 0.0003218662500914977, 'samples': 4136832, 'steps': 21545, 'loss/train': 2.0643467903137207} 01/29/2022 16:35:45 - INFO - codeparrot_training - Step 21546: {'lr': 0.0003218505781166534, 'samples': 4137024, 'steps': 21546, 'loss/train': 0.8049803674221039} 01/29/2022 16:35:49 - INFO - codeparrot_training - Step 21547: {'lr': 0.000321834905834024, 'samples': 4137216, 'steps': 21547, 'loss/train': 1.788675606250763} 01/29/2022 16:35:55 - INFO - codeparrot_training - Step 21548: {'lr': 0.00032181923324367675, 'samples': 4137408, 'steps': 21548, 'loss/train': 1.0890695750713348} 01/29/2022 16:35:59 - INFO - codeparrot_training - Step 21549: {'lr': 0.0003218035603456786, 'samples': 4137600, 'steps': 21549, 'loss/train': 1.489492803812027} 01/29/2022 16:36:03 - INFO - codeparrot_training - Step 21550: {'lr': 0.00032178788714009687, 'samples': 4137792, 'steps': 21550, 'loss/train': 1.4241848587989807} 01/29/2022 16:36:08 - INFO - codeparrot_training - Step 21551: {'lr': 0.00032177221362699853, 'samples': 4137984, 'steps': 21551, 'loss/train': 1.442140281200409} 01/29/2022 16:36:12 - INFO - codeparrot_training - Step 21552: {'lr': 0.0003217565398064509, 'samples': 4138176, 'steps': 21552, 'loss/train': 0.9041439592838287} 01/29/2022 16:36:17 - INFO - codeparrot_training - Step 21553: {'lr': 0.000321740865678521, 'samples': 4138368, 'steps': 21553, 'loss/train': 1.2496611177921295} 01/29/2022 16:36:22 - INFO - codeparrot_training - Step 21554: {'lr': 0.00032172519124327607, 'samples': 4138560, 'steps': 21554, 'loss/train': 1.800967276096344} 01/29/2022 16:36:26 - INFO - codeparrot_training - Step 21555: {'lr': 0.00032170951650078316, 'samples': 4138752, 'steps': 21555, 'loss/train': 2.0443641543388367} 01/29/2022 16:36:30 - INFO - codeparrot_training - Step 21556: {'lr': 0.0003216938414511095, 'samples': 4138944, 'steps': 21556, 'loss/train': 0.07303005829453468} 01/29/2022 16:36:34 - INFO - codeparrot_training - Step 21557: {'lr': 0.0003216781660943221, 'samples': 4139136, 'steps': 21557, 'loss/train': 0.5929491966962814} 01/29/2022 16:36:41 - INFO - codeparrot_training - Step 21558: {'lr': 0.0003216624904304882, 'samples': 4139328, 'steps': 21558, 'loss/train': 2.344435930252075} 01/29/2022 16:36:46 - INFO - codeparrot_training - Step 21559: {'lr': 0.000321646814459675, 'samples': 4139520, 'steps': 21559, 'loss/train': 2.68536114692688} 01/29/2022 16:36:50 - INFO - codeparrot_training - Step 21560: {'lr': 0.0003216311381819496, 'samples': 4139712, 'steps': 21560, 'loss/train': 1.6553632020950317} 01/29/2022 16:36:54 - INFO - codeparrot_training - Step 21561: {'lr': 0.00032161546159737917, 'samples': 4139904, 'steps': 21561, 'loss/train': 1.7094349265098572} 01/29/2022 16:36:58 - INFO - codeparrot_training - Step 21562: {'lr': 0.0003215997847060307, 'samples': 4140096, 'steps': 21562, 'loss/train': 2.2703145146369934} 01/29/2022 16:37:04 - INFO - codeparrot_training - Step 21563: {'lr': 0.00032158410750797163, 'samples': 4140288, 'steps': 21563, 'loss/train': 0.23333825916051865} 01/29/2022 16:37:08 - INFO - codeparrot_training - Step 21564: {'lr': 0.000321568430003269, 'samples': 4140480, 'steps': 21564, 'loss/train': 1.433699369430542} 01/29/2022 16:37:12 - INFO - codeparrot_training - Step 21565: {'lr': 0.00032155275219198986, 'samples': 4140672, 'steps': 21565, 'loss/train': 1.3845269680023193} 01/29/2022 16:37:17 - INFO - codeparrot_training - Step 21566: {'lr': 0.0003215370740742014, 'samples': 4140864, 'steps': 21566, 'loss/train': 1.2066005766391754} 01/29/2022 16:37:21 - INFO - codeparrot_training - Step 21567: {'lr': 0.00032152139564997097, 'samples': 4141056, 'steps': 21567, 'loss/train': 0.03010717313736677} 01/29/2022 16:37:26 - INFO - codeparrot_training - Step 21568: {'lr': 0.0003215057169193655, 'samples': 4141248, 'steps': 21568, 'loss/train': 1.6521202325820923} 01/29/2022 16:37:30 - INFO - codeparrot_training - Step 21569: {'lr': 0.00032149003788245223, 'samples': 4141440, 'steps': 21569, 'loss/train': 1.7991853952407837} 01/29/2022 16:37:35 - INFO - codeparrot_training - Step 21570: {'lr': 0.0003214743585392984, 'samples': 4141632, 'steps': 21570, 'loss/train': 1.9928301572799683} 01/29/2022 16:37:39 - INFO - codeparrot_training - Step 21571: {'lr': 0.0003214586788899711, 'samples': 4141824, 'steps': 21571, 'loss/train': 2.0250880122184753} 01/29/2022 16:37:43 - INFO - codeparrot_training - Step 21572: {'lr': 0.00032144299893453743, 'samples': 4142016, 'steps': 21572, 'loss/train': 1.1777637898921967} 01/29/2022 16:37:50 - INFO - codeparrot_training - Step 21573: {'lr': 0.00032142731867306466, 'samples': 4142208, 'steps': 21573, 'loss/train': 1.5899255275726318} 01/29/2022 16:37:55 - INFO - codeparrot_training - Step 21574: {'lr': 0.00032141163810562, 'samples': 4142400, 'steps': 21574, 'loss/train': 1.4607300460338593} 01/29/2022 16:37:59 - INFO - codeparrot_training - Step 21575: {'lr': 0.00032139595723227054, 'samples': 4142592, 'steps': 21575, 'loss/train': 1.689505934715271} 01/29/2022 16:38:03 - INFO - codeparrot_training - Step 21576: {'lr': 0.0003213802760530835, 'samples': 4142784, 'steps': 21576, 'loss/train': 1.0810178518295288} 01/29/2022 16:38:07 - INFO - codeparrot_training - Step 21577: {'lr': 0.000321364594568126, 'samples': 4142976, 'steps': 21577, 'loss/train': 1.5603477358818054} 01/29/2022 16:38:13 - INFO - codeparrot_training - Step 21578: {'lr': 0.00032134891277746527, 'samples': 4143168, 'steps': 21578, 'loss/train': 1.1335854828357697} 01/29/2022 16:38:17 - INFO - codeparrot_training - Step 21579: {'lr': 0.0003213332306811684, 'samples': 4143360, 'steps': 21579, 'loss/train': 1.9130002856254578} 01/29/2022 16:38:21 - INFO - codeparrot_training - Step 21580: {'lr': 0.0003213175482793026, 'samples': 4143552, 'steps': 21580, 'loss/train': 0.2278280183672905} 01/29/2022 16:38:25 - INFO - codeparrot_training - Step 21581: {'lr': 0.00032130186557193506, 'samples': 4143744, 'steps': 21581, 'loss/train': 1.2457708418369293} 01/29/2022 16:38:31 - INFO - codeparrot_training - Step 21582: {'lr': 0.0003212861825591331, 'samples': 4143936, 'steps': 21582, 'loss/train': 1.3969421982765198} 01/29/2022 16:38:35 - INFO - codeparrot_training - Step 21583: {'lr': 0.00032127049924096364, 'samples': 4144128, 'steps': 21583, 'loss/train': 2.591210424900055} 01/29/2022 16:38:39 - INFO - codeparrot_training - Step 21584: {'lr': 0.00032125481561749405, 'samples': 4144320, 'steps': 21584, 'loss/train': 1.492270588874817} 01/29/2022 16:38:44 - INFO - codeparrot_training - Step 21585: {'lr': 0.00032123913168879146, 'samples': 4144512, 'steps': 21585, 'loss/train': 2.8121767044067383} 01/29/2022 16:38:48 - INFO - codeparrot_training - Step 21586: {'lr': 0.00032122344745492303, 'samples': 4144704, 'steps': 21586, 'loss/train': 1.109397679567337} 01/29/2022 16:38:55 - INFO - codeparrot_training - Step 21587: {'lr': 0.00032120776291595594, 'samples': 4144896, 'steps': 21587, 'loss/train': 1.5596997141838074} 01/29/2022 16:38:59 - INFO - codeparrot_training - Step 21588: {'lr': 0.00032119207807195747, 'samples': 4145088, 'steps': 21588, 'loss/train': 1.8228551745414734} 01/29/2022 16:39:03 - INFO - codeparrot_training - Step 21589: {'lr': 0.0003211763929229947, 'samples': 4145280, 'steps': 21589, 'loss/train': 0.8057121634483337} 01/29/2022 16:39:08 - INFO - codeparrot_training - Step 21590: {'lr': 0.00032116070746913484, 'samples': 4145472, 'steps': 21590, 'loss/train': 1.6382612586021423} 01/29/2022 16:39:12 - INFO - codeparrot_training - Step 21591: {'lr': 0.0003211450217104452, 'samples': 4145664, 'steps': 21591, 'loss/train': 1.5078953504562378} 01/29/2022 16:39:17 - INFO - codeparrot_training - Step 21592: {'lr': 0.00032112933564699275, 'samples': 4145856, 'steps': 21592, 'loss/train': 2.132845401763916} 01/29/2022 16:39:22 - INFO - codeparrot_training - Step 21593: {'lr': 0.0003211136492788449, 'samples': 4146048, 'steps': 21593, 'loss/train': 1.7453742027282715} 01/29/2022 16:39:26 - INFO - codeparrot_training - Step 21594: {'lr': 0.0003210979626060687, 'samples': 4146240, 'steps': 21594, 'loss/train': 1.4613568782806396} 01/29/2022 16:39:30 - INFO - codeparrot_training - Step 21595: {'lr': 0.00032108227562873147, 'samples': 4146432, 'steps': 21595, 'loss/train': 1.4512999355793} 01/29/2022 16:39:34 - INFO - codeparrot_training - Step 21596: {'lr': 0.0003210665883469003, 'samples': 4146624, 'steps': 21596, 'loss/train': 0.0798215176910162} 01/29/2022 16:39:40 - INFO - codeparrot_training - Step 21597: {'lr': 0.0003210509007606424, 'samples': 4146816, 'steps': 21597, 'loss/train': 1.71167653799057} 01/29/2022 16:39:44 - INFO - codeparrot_training - Step 21598: {'lr': 0.00032103521287002505, 'samples': 4147008, 'steps': 21598, 'loss/train': 1.593159019947052} 01/29/2022 16:39:48 - INFO - codeparrot_training - Step 21599: {'lr': 0.0003210195246751154, 'samples': 4147200, 'steps': 21599, 'loss/train': 1.5682770609855652} 01/29/2022 16:39:52 - INFO - codeparrot_training - Step 21600: {'lr': 0.0003210038361759807, 'samples': 4147392, 'steps': 21600, 'loss/train': 1.656806766986847} 01/29/2022 16:39:57 - INFO - codeparrot_training - Step 21601: {'lr': 0.000320988147372688, 'samples': 4147584, 'steps': 21601, 'loss/train': 1.9480562210083008} 01/29/2022 16:40:04 - INFO - codeparrot_training - Step 21602: {'lr': 0.00032097245826530476, 'samples': 4147776, 'steps': 21602, 'loss/train': 1.1482498347759247} 01/29/2022 16:40:08 - INFO - codeparrot_training - Step 21603: {'lr': 0.00032095676885389793, 'samples': 4147968, 'steps': 21603, 'loss/train': 2.282365143299103} 01/29/2022 16:40:12 - INFO - codeparrot_training - Step 21604: {'lr': 0.00032094107913853485, 'samples': 4148160, 'steps': 21604, 'loss/train': 1.053196907043457} 01/29/2022 16:40:17 - INFO - codeparrot_training - Step 21605: {'lr': 0.00032092538911928276, 'samples': 4148352, 'steps': 21605, 'loss/train': 1.4439847469329834} 01/29/2022 16:40:21 - INFO - codeparrot_training - Step 21606: {'lr': 0.00032090969879620886, 'samples': 4148544, 'steps': 21606, 'loss/train': 0.20305777341127396} 01/29/2022 16:40:26 - INFO - codeparrot_training - Step 21607: {'lr': 0.00032089400816938016, 'samples': 4148736, 'steps': 21607, 'loss/train': 1.7635923027992249} 01/29/2022 16:40:30 - INFO - codeparrot_training - Step 21608: {'lr': 0.0003208783172388642, 'samples': 4148928, 'steps': 21608, 'loss/train': 2.0154372453689575} 01/29/2022 16:40:35 - INFO - codeparrot_training - Step 21609: {'lr': 0.000320862626004728, 'samples': 4149120, 'steps': 21609, 'loss/train': 1.1454711556434631} 01/29/2022 16:40:39 - INFO - codeparrot_training - Step 21610: {'lr': 0.00032084693446703875, 'samples': 4149312, 'steps': 21610, 'loss/train': 1.5647023916244507} 01/29/2022 16:40:43 - INFO - codeparrot_training - Step 21611: {'lr': 0.00032083124262586384, 'samples': 4149504, 'steps': 21611, 'loss/train': 1.553775429725647} 01/29/2022 16:40:49 - INFO - codeparrot_training - Step 21612: {'lr': 0.0003208155504812703, 'samples': 4149696, 'steps': 21612, 'loss/train': 1.3515832722187042} 01/29/2022 16:40:54 - INFO - codeparrot_training - Step 21613: {'lr': 0.00032079985803332546, 'samples': 4149888, 'steps': 21613, 'loss/train': 2.0639476776123047} 01/29/2022 16:40:58 - INFO - codeparrot_training - Step 21614: {'lr': 0.0003207841652820964, 'samples': 4150080, 'steps': 21614, 'loss/train': 1.9457103610038757} 01/29/2022 16:41:02 - INFO - codeparrot_training - Step 21615: {'lr': 0.0003207684722276506, 'samples': 4150272, 'steps': 21615, 'loss/train': 1.2299189865589142} 01/29/2022 16:41:07 - INFO - codeparrot_training - Step 21616: {'lr': 0.00032075277887005503, 'samples': 4150464, 'steps': 21616, 'loss/train': 1.6755284070968628} 01/29/2022 16:41:12 - INFO - codeparrot_training - Step 21617: {'lr': 0.0003207370852093771, 'samples': 4150656, 'steps': 21617, 'loss/train': 1.4029744863510132} 01/29/2022 16:41:16 - INFO - codeparrot_training - Step 21618: {'lr': 0.00032072139124568396, 'samples': 4150848, 'steps': 21618, 'loss/train': 0.589465856552124} 01/29/2022 16:41:20 - INFO - codeparrot_training - Step 21619: {'lr': 0.0003207056969790428, 'samples': 4151040, 'steps': 21619, 'loss/train': 1.0325101017951965} 01/29/2022 16:41:25 - INFO - codeparrot_training - Step 21620: {'lr': 0.0003206900024095208, 'samples': 4151232, 'steps': 21620, 'loss/train': 1.688573956489563} 01/29/2022 16:41:29 - INFO - codeparrot_training - Step 21621: {'lr': 0.0003206743075371854, 'samples': 4151424, 'steps': 21621, 'loss/train': 1.533450186252594} 01/29/2022 16:41:34 - INFO - codeparrot_training - Step 21622: {'lr': 0.0003206586123621037, 'samples': 4151616, 'steps': 21622, 'loss/train': 1.2325274348258972} 01/29/2022 16:41:38 - INFO - codeparrot_training - Step 21623: {'lr': 0.00032064291688434286, 'samples': 4151808, 'steps': 21623, 'loss/train': 1.8069384098052979} 01/29/2022 16:41:43 - INFO - codeparrot_training - Step 21624: {'lr': 0.00032062722110397034, 'samples': 4152000, 'steps': 21624, 'loss/train': 2.378345489501953} 01/29/2022 16:41:47 - INFO - codeparrot_training - Step 21625: {'lr': 0.0003206115250210531, 'samples': 4152192, 'steps': 21625, 'loss/train': 1.4951856136322021} 01/29/2022 16:41:51 - INFO - codeparrot_training - Step 21626: {'lr': 0.00032059582863565864, 'samples': 4152384, 'steps': 21626, 'loss/train': 0.941668689250946} 01/29/2022 16:41:57 - INFO - codeparrot_training - Step 21627: {'lr': 0.0003205801319478539, 'samples': 4152576, 'steps': 21627, 'loss/train': 1.5209380388259888} 01/29/2022 16:42:01 - INFO - codeparrot_training - Step 21628: {'lr': 0.00032056443495770637, 'samples': 4152768, 'steps': 21628, 'loss/train': 1.9341793656349182} 01/29/2022 16:42:06 - INFO - codeparrot_training - Step 21629: {'lr': 0.0003205487376652833, 'samples': 4152960, 'steps': 21629, 'loss/train': 2.590539336204529} 01/29/2022 16:42:10 - INFO - codeparrot_training - Step 21630: {'lr': 0.0003205330400706517, 'samples': 4153152, 'steps': 21630, 'loss/train': 1.7471920251846313} 01/29/2022 16:42:14 - INFO - codeparrot_training - Step 21631: {'lr': 0.000320517342173879, 'samples': 4153344, 'steps': 21631, 'loss/train': 1.761817753314972} 01/29/2022 16:42:18 - INFO - codeparrot_training - Step 21632: {'lr': 0.0003205016439750323, 'samples': 4153536, 'steps': 21632, 'loss/train': 1.6830110549926758} 01/29/2022 16:42:25 - INFO - codeparrot_training - Step 21633: {'lr': 0.00032048594547417916, 'samples': 4153728, 'steps': 21633, 'loss/train': 0.9457539618015289} 01/29/2022 16:42:30 - INFO - codeparrot_training - Step 21634: {'lr': 0.00032047024667138644, 'samples': 4153920, 'steps': 21634, 'loss/train': 0.8563356399536133} 01/29/2022 16:42:34 - INFO - codeparrot_training - Step 21635: {'lr': 0.00032045454756672164, 'samples': 4154112, 'steps': 21635, 'loss/train': 1.4272909462451935} 01/29/2022 16:42:38 - INFO - codeparrot_training - Step 21636: {'lr': 0.00032043884816025187, 'samples': 4154304, 'steps': 21636, 'loss/train': 2.016688048839569} 01/29/2022 16:42:43 - INFO - codeparrot_training - Step 21637: {'lr': 0.0003204231484520445, 'samples': 4154496, 'steps': 21637, 'loss/train': 1.9672536849975586} 01/29/2022 16:42:48 - INFO - codeparrot_training - Step 21638: {'lr': 0.0003204074484421667, 'samples': 4154688, 'steps': 21638, 'loss/train': 3.082314133644104} 01/29/2022 16:42:52 - INFO - codeparrot_training - Step 21639: {'lr': 0.0003203917481306857, 'samples': 4154880, 'steps': 21639, 'loss/train': 2.1821420788764954} 01/29/2022 16:42:57 - INFO - codeparrot_training - Step 21640: {'lr': 0.0003203760475176689, 'samples': 4155072, 'steps': 21640, 'loss/train': 1.4574891328811646} 01/29/2022 16:43:01 - INFO - codeparrot_training - Step 21641: {'lr': 0.00032036034660318344, 'samples': 4155264, 'steps': 21641, 'loss/train': 1.7432259321212769} 01/29/2022 16:43:05 - INFO - codeparrot_training - Step 21642: {'lr': 0.00032034464538729647, 'samples': 4155456, 'steps': 21642, 'loss/train': 2.147330582141876} 01/29/2022 16:43:10 - INFO - codeparrot_training - Step 21643: {'lr': 0.0003203289438700755, 'samples': 4155648, 'steps': 21643, 'loss/train': 1.6227259039878845} 01/29/2022 16:43:15 - INFO - codeparrot_training - Step 21644: {'lr': 0.0003203132420515876, 'samples': 4155840, 'steps': 21644, 'loss/train': 2.252621591091156} 01/29/2022 16:43:19 - INFO - codeparrot_training - Step 21645: {'lr': 0.0003202975399319002, 'samples': 4156032, 'steps': 21645, 'loss/train': 0.9174507558345795} 01/29/2022 16:43:23 - INFO - codeparrot_training - Step 21646: {'lr': 0.00032028183751108035, 'samples': 4156224, 'steps': 21646, 'loss/train': 1.969442903995514} 01/29/2022 16:43:27 - INFO - codeparrot_training - Step 21647: {'lr': 0.00032026613478919547, 'samples': 4156416, 'steps': 21647, 'loss/train': 1.7416625618934631} 01/29/2022 16:43:35 - INFO - codeparrot_training - Step 21648: {'lr': 0.0003202504317663128, 'samples': 4156608, 'steps': 21648, 'loss/train': 2.0301631093025208} 01/29/2022 16:43:39 - INFO - codeparrot_training - Step 21649: {'lr': 0.0003202347284424995, 'samples': 4156800, 'steps': 21649, 'loss/train': 1.2099173069000244} 01/29/2022 16:43:43 - INFO - codeparrot_training - Step 21650: {'lr': 0.00032021902481782304, 'samples': 4156992, 'steps': 21650, 'loss/train': 1.0580499172210693} 01/29/2022 16:43:47 - INFO - codeparrot_training - Step 21651: {'lr': 0.0003202033208923505, 'samples': 4157184, 'steps': 21651, 'loss/train': 1.5488204956054688} 01/29/2022 16:43:52 - INFO - codeparrot_training - Step 21652: {'lr': 0.0003201876166661493, 'samples': 4157376, 'steps': 21652, 'loss/train': 1.8973039984703064} 01/29/2022 16:43:57 - INFO - codeparrot_training - Step 21653: {'lr': 0.00032017191213928653, 'samples': 4157568, 'steps': 21653, 'loss/train': 1.1920217871665955} 01/29/2022 16:44:01 - INFO - codeparrot_training - Step 21654: {'lr': 0.0003201562073118297, 'samples': 4157760, 'steps': 21654, 'loss/train': 1.7116631269454956} 01/29/2022 16:44:06 - INFO - codeparrot_training - Step 21655: {'lr': 0.00032014050218384584, 'samples': 4157952, 'steps': 21655, 'loss/train': 1.709224820137024} 01/29/2022 16:44:10 - INFO - codeparrot_training - Step 21656: {'lr': 0.0003201247967554024, 'samples': 4158144, 'steps': 21656, 'loss/train': 1.5117892026901245} 01/29/2022 16:44:14 - INFO - codeparrot_training - Step 21657: {'lr': 0.0003201090910265666, 'samples': 4158336, 'steps': 21657, 'loss/train': 1.3655120730400085} 01/29/2022 16:44:21 - INFO - codeparrot_training - Step 21658: {'lr': 0.0003200933849974056, 'samples': 4158528, 'steps': 21658, 'loss/train': 1.6471683382987976} 01/29/2022 16:44:25 - INFO - codeparrot_training - Step 21659: {'lr': 0.000320077678667987, 'samples': 4158720, 'steps': 21659, 'loss/train': 1.3560667634010315} 01/29/2022 16:44:30 - INFO - codeparrot_training - Step 21660: {'lr': 0.00032006197203837775, 'samples': 4158912, 'steps': 21660, 'loss/train': 0.9726782441139221} 01/29/2022 16:44:34 - INFO - codeparrot_training - Step 21661: {'lr': 0.00032004626510864526, 'samples': 4159104, 'steps': 21661, 'loss/train': 1.4129663407802582} 01/29/2022 16:44:39 - INFO - codeparrot_training - Step 21662: {'lr': 0.00032003055787885684, 'samples': 4159296, 'steps': 21662, 'loss/train': 1.4805044531822205} 01/29/2022 16:44:44 - INFO - codeparrot_training - Step 21663: {'lr': 0.00032001485034907975, 'samples': 4159488, 'steps': 21663, 'loss/train': 1.674323558807373} 01/29/2022 16:44:48 - INFO - codeparrot_training - Step 21664: {'lr': 0.00031999914251938124, 'samples': 4159680, 'steps': 21664, 'loss/train': 1.720266580581665} 01/29/2022 16:44:52 - INFO - codeparrot_training - Step 21665: {'lr': 0.00031998343438982866, 'samples': 4159872, 'steps': 21665, 'loss/train': 1.0107362866401672} 01/29/2022 16:44:56 - INFO - codeparrot_training - Step 21666: {'lr': 0.0003199677259604893, 'samples': 4160064, 'steps': 21666, 'loss/train': 0.5399242043495178} 01/29/2022 16:45:02 - INFO - codeparrot_training - Step 21667: {'lr': 0.0003199520172314304, 'samples': 4160256, 'steps': 21667, 'loss/train': 1.8692690134048462} 01/29/2022 16:45:06 - INFO - codeparrot_training - Step 21668: {'lr': 0.00031993630820271925, 'samples': 4160448, 'steps': 21668, 'loss/train': 0.7354365438222885} 01/29/2022 16:45:10 - INFO - codeparrot_training - Step 21669: {'lr': 0.00031992059887442316, 'samples': 4160640, 'steps': 21669, 'loss/train': 0.7053664326667786} 01/29/2022 16:45:15 - INFO - codeparrot_training - Step 21670: {'lr': 0.00031990488924660943, 'samples': 4160832, 'steps': 21670, 'loss/train': 2.361015737056732} 01/29/2022 16:45:19 - INFO - codeparrot_training - Step 21671: {'lr': 0.00031988917931934536, 'samples': 4161024, 'steps': 21671, 'loss/train': 1.8439360857009888} 01/29/2022 16:45:24 - INFO - codeparrot_training - Step 21672: {'lr': 0.0003198734690926982, 'samples': 4161216, 'steps': 21672, 'loss/train': 1.7243177890777588} 01/29/2022 16:45:29 - INFO - codeparrot_training - Step 21673: {'lr': 0.00031985775856673536, 'samples': 4161408, 'steps': 21673, 'loss/train': 0.9780372977256775} 01/29/2022 16:45:33 - INFO - codeparrot_training - Step 21674: {'lr': 0.000319842047741524, 'samples': 4161600, 'steps': 21674, 'loss/train': 1.2857859134674072} 01/29/2022 16:45:37 - INFO - codeparrot_training - Step 21675: {'lr': 0.0003198263366171315, 'samples': 4161792, 'steps': 21675, 'loss/train': 0.3741970360279083} 01/29/2022 16:45:41 - INFO - codeparrot_training - Step 21676: {'lr': 0.00031981062519362513, 'samples': 4161984, 'steps': 21676, 'loss/train': 1.3427040874958038} 01/29/2022 16:45:48 - INFO - codeparrot_training - Step 21677: {'lr': 0.00031979491347107226, 'samples': 4162176, 'steps': 21677, 'loss/train': 1.1967697441577911} 01/29/2022 16:45:53 - INFO - codeparrot_training - Step 21678: {'lr': 0.0003197792014495402, 'samples': 4162368, 'steps': 21678, 'loss/train': 1.4727714657783508} 01/29/2022 16:45:57 - INFO - codeparrot_training - Step 21679: {'lr': 0.0003197634891290961, 'samples': 4162560, 'steps': 21679, 'loss/train': 1.9783902168273926} 01/29/2022 16:46:01 - INFO - codeparrot_training - Step 21680: {'lr': 0.00031974777650980735, 'samples': 4162752, 'steps': 21680, 'loss/train': 1.0317197442054749} 01/29/2022 16:46:05 - INFO - codeparrot_training - Step 21681: {'lr': 0.0003197320635917413, 'samples': 4162944, 'steps': 21681, 'loss/train': 1.8950217962265015} 01/29/2022 16:46:11 - INFO - codeparrot_training - Step 21682: {'lr': 0.0003197163503749652, 'samples': 4163136, 'steps': 21682, 'loss/train': 1.6023699045181274} 01/29/2022 16:46:15 - INFO - codeparrot_training - Step 21683: {'lr': 0.00031970063685954645, 'samples': 4163328, 'steps': 21683, 'loss/train': 1.3805947601795197} 01/29/2022 16:46:19 - INFO - codeparrot_training - Step 21684: {'lr': 0.0003196849230455523, 'samples': 4163520, 'steps': 21684, 'loss/train': 0.8036296963691711} 01/29/2022 16:46:24 - INFO - codeparrot_training - Step 21685: {'lr': 0.0003196692089330501, 'samples': 4163712, 'steps': 21685, 'loss/train': 1.5011937618255615} 01/29/2022 16:46:28 - INFO - codeparrot_training - Step 21686: {'lr': 0.000319653494522107, 'samples': 4163904, 'steps': 21686, 'loss/train': 1.360840380191803} 01/29/2022 16:46:33 - INFO - codeparrot_training - Step 21687: {'lr': 0.00031963777981279057, 'samples': 4164096, 'steps': 21687, 'loss/train': 2.038798213005066} 01/29/2022 16:46:37 - INFO - codeparrot_training - Step 21688: {'lr': 0.00031962206480516794, 'samples': 4164288, 'steps': 21688, 'loss/train': 0.09384938701987267} 01/29/2022 16:46:42 - INFO - codeparrot_training - Step 21689: {'lr': 0.00031960634949930656, 'samples': 4164480, 'steps': 21689, 'loss/train': 1.8104642629623413} 01/29/2022 16:46:46 - INFO - codeparrot_training - Step 21690: {'lr': 0.0003195906338952736, 'samples': 4164672, 'steps': 21690, 'loss/train': 1.9120075106620789} 01/29/2022 16:46:50 - INFO - codeparrot_training - Step 21691: {'lr': 0.00031957491799313646, 'samples': 4164864, 'steps': 21691, 'loss/train': 0.9831165075302124} 01/29/2022 16:46:57 - INFO - codeparrot_training - Step 21692: {'lr': 0.0003195592017929625, 'samples': 4165056, 'steps': 21692, 'loss/train': 2.047052800655365} 01/29/2022 16:47:02 - INFO - codeparrot_training - Step 21693: {'lr': 0.000319543485294819, 'samples': 4165248, 'steps': 21693, 'loss/train': 0.948269784450531} 01/29/2022 16:47:06 - INFO - codeparrot_training - Step 21694: {'lr': 0.0003195277684987733, 'samples': 4165440, 'steps': 21694, 'loss/train': 1.3810468018054962} 01/29/2022 16:47:10 - INFO - codeparrot_training - Step 21695: {'lr': 0.0003195120514048927, 'samples': 4165632, 'steps': 21695, 'loss/train': 1.4220552742481232} 01/29/2022 16:47:14 - INFO - codeparrot_training - Step 21696: {'lr': 0.00031949633401324464, 'samples': 4165824, 'steps': 21696, 'loss/train': 1.116628646850586} 01/29/2022 16:47:20 - INFO - codeparrot_training - Step 21697: {'lr': 0.00031948061632389624, 'samples': 4166016, 'steps': 21697, 'loss/train': 1.7008062601089478} 01/29/2022 16:47:24 - INFO - codeparrot_training - Step 21698: {'lr': 0.00031946489833691494, 'samples': 4166208, 'steps': 21698, 'loss/train': 1.2103594243526459} 01/29/2022 16:47:28 - INFO - codeparrot_training - Step 21699: {'lr': 0.0003194491800523681, 'samples': 4166400, 'steps': 21699, 'loss/train': 2.2195944786071777} 01/29/2022 16:47:32 - INFO - codeparrot_training - Step 21700: {'lr': 0.0003194334614703231, 'samples': 4166592, 'steps': 21700, 'loss/train': 1.844855546951294} 01/29/2022 16:47:37 - INFO - codeparrot_training - Step 21701: {'lr': 0.0003194177425908471, 'samples': 4166784, 'steps': 21701, 'loss/train': 1.391778141260147} 01/29/2022 16:47:42 - INFO - codeparrot_training - Step 21702: {'lr': 0.0003194020234140076, 'samples': 4166976, 'steps': 21702, 'loss/train': 0.7689953148365021} 01/29/2022 16:47:46 - INFO - codeparrot_training - Step 21703: {'lr': 0.00031938630393987176, 'samples': 4167168, 'steps': 21703, 'loss/train': 4.247654914855957} 01/29/2022 16:47:50 - INFO - codeparrot_training - Step 21704: {'lr': 0.00031937058416850716, 'samples': 4167360, 'steps': 21704, 'loss/train': 0.2708232253789902} 01/29/2022 16:47:55 - INFO - codeparrot_training - Step 21705: {'lr': 0.00031935486409998096, 'samples': 4167552, 'steps': 21705, 'loss/train': 1.8466315269470215} 01/29/2022 16:47:59 - INFO - codeparrot_training - Step 21706: {'lr': 0.0003193391437343605, 'samples': 4167744, 'steps': 21706, 'loss/train': 1.6700330972671509} 01/29/2022 16:48:06 - INFO - codeparrot_training - Step 21707: {'lr': 0.0003193234230717132, 'samples': 4167936, 'steps': 21707, 'loss/train': 1.7224161028862} 01/29/2022 16:48:11 - INFO - codeparrot_training - Step 21708: {'lr': 0.00031930770211210637, 'samples': 4168128, 'steps': 21708, 'loss/train': 2.170560300350189} 01/29/2022 16:48:15 - INFO - codeparrot_training - Step 21709: {'lr': 0.0003192919808556073, 'samples': 4168320, 'steps': 21709, 'loss/train': 1.1928541660308838} 01/29/2022 16:48:19 - INFO - codeparrot_training - Step 21710: {'lr': 0.00031927625930228343, 'samples': 4168512, 'steps': 21710, 'loss/train': 1.9204714894294739} 01/29/2022 16:48:24 - INFO - codeparrot_training - Step 21711: {'lr': 0.00031926053745220213, 'samples': 4168704, 'steps': 21711, 'loss/train': 0.4274754077196121} 01/29/2022 16:48:29 - INFO - codeparrot_training - Step 21712: {'lr': 0.0003192448153054306, 'samples': 4168896, 'steps': 21712, 'loss/train': 1.5927847623825073} 01/29/2022 16:48:33 - INFO - codeparrot_training - Step 21713: {'lr': 0.0003192290928620363, 'samples': 4169088, 'steps': 21713, 'loss/train': 1.3653227984905243} 01/29/2022 16:48:37 - INFO - codeparrot_training - Step 21714: {'lr': 0.0003192133701220865, 'samples': 4169280, 'steps': 21714, 'loss/train': 1.792011559009552} 01/29/2022 16:48:42 - INFO - codeparrot_training - Step 21715: {'lr': 0.0003191976470856487, 'samples': 4169472, 'steps': 21715, 'loss/train': 1.7093900442123413} 01/29/2022 16:48:46 - INFO - codeparrot_training - Step 21716: {'lr': 0.00031918192375279006, 'samples': 4169664, 'steps': 21716, 'loss/train': 1.247991532087326} 01/29/2022 16:48:52 - INFO - codeparrot_training - Step 21717: {'lr': 0.00031916620012357804, 'samples': 4169856, 'steps': 21717, 'loss/train': 1.9751880168914795} 01/29/2022 16:48:56 - INFO - codeparrot_training - Step 21718: {'lr': 0.00031915047619808, 'samples': 4170048, 'steps': 21718, 'loss/train': 1.023457556962967} 01/29/2022 16:49:01 - INFO - codeparrot_training - Step 21719: {'lr': 0.0003191347519763633, 'samples': 4170240, 'steps': 21719, 'loss/train': 1.9302175641059875} 01/29/2022 16:49:05 - INFO - codeparrot_training - Step 21720: {'lr': 0.00031911902745849525, 'samples': 4170432, 'steps': 21720, 'loss/train': 1.5428465008735657} 01/29/2022 16:49:09 - INFO - codeparrot_training - Step 21721: {'lr': 0.0003191033026445432, 'samples': 4170624, 'steps': 21721, 'loss/train': 0.9775018393993378} 01/29/2022 16:49:15 - INFO - codeparrot_training - Step 21722: {'lr': 0.00031908757753457465, 'samples': 4170816, 'steps': 21722, 'loss/train': 1.1753352284431458} 01/29/2022 16:49:19 - INFO - codeparrot_training - Step 21723: {'lr': 0.00031907185212865673, 'samples': 4171008, 'steps': 21723, 'loss/train': 1.109279304742813} 01/29/2022 16:49:23 - INFO - codeparrot_training - Step 21724: {'lr': 0.000319056126426857, 'samples': 4171200, 'steps': 21724, 'loss/train': 1.728000283241272} 01/29/2022 16:49:27 - INFO - codeparrot_training - Step 21725: {'lr': 0.0003190404004292427, 'samples': 4171392, 'steps': 21725, 'loss/train': 1.2982609570026398} 01/29/2022 16:49:32 - INFO - codeparrot_training - Step 21726: {'lr': 0.00031902467413588134, 'samples': 4171584, 'steps': 21726, 'loss/train': 2.6672126054763794} 01/29/2022 16:49:37 - INFO - codeparrot_training - Step 21727: {'lr': 0.00031900894754684006, 'samples': 4171776, 'steps': 21727, 'loss/train': 2.1456631422042847} 01/29/2022 16:49:41 - INFO - codeparrot_training - Step 21728: {'lr': 0.0003189932206621865, 'samples': 4171968, 'steps': 21728, 'loss/train': 0.8494352996349335} 01/29/2022 16:49:46 - INFO - codeparrot_training - Step 21729: {'lr': 0.00031897749348198777, 'samples': 4172160, 'steps': 21729, 'loss/train': 2.1679930686950684} 01/29/2022 16:49:50 - INFO - codeparrot_training - Step 21730: {'lr': 0.0003189617660063114, 'samples': 4172352, 'steps': 21730, 'loss/train': 1.4844428300857544} 01/29/2022 16:49:55 - INFO - codeparrot_training - Step 21731: {'lr': 0.0003189460382352248, 'samples': 4172544, 'steps': 21731, 'loss/train': 1.7935829758644104} 01/29/2022 16:50:00 - INFO - codeparrot_training - Step 21732: {'lr': 0.00031893031016879515, 'samples': 4172736, 'steps': 21732, 'loss/train': 1.5535878539085388} 01/29/2022 16:50:04 - INFO - codeparrot_training - Step 21733: {'lr': 0.00031891458180709003, 'samples': 4172928, 'steps': 21733, 'loss/train': 2.039648652076721} 01/29/2022 16:50:08 - INFO - codeparrot_training - Step 21734: {'lr': 0.0003188988531501766, 'samples': 4173120, 'steps': 21734, 'loss/train': 1.9077422618865967} 01/29/2022 16:50:12 - INFO - codeparrot_training - Step 21735: {'lr': 0.00031888312419812255, 'samples': 4173312, 'steps': 21735, 'loss/train': 1.9897461533546448} 01/29/2022 16:50:19 - INFO - codeparrot_training - Step 21736: {'lr': 0.00031886739495099494, 'samples': 4173504, 'steps': 21736, 'loss/train': 1.5613859295845032} 01/29/2022 16:50:24 - INFO - codeparrot_training - Step 21737: {'lr': 0.00031885166540886135, 'samples': 4173696, 'steps': 21737, 'loss/train': 1.4356939494609833} 01/29/2022 16:50:28 - INFO - codeparrot_training - Step 21738: {'lr': 0.000318835935571789, 'samples': 4173888, 'steps': 21738, 'loss/train': 2.007459819316864} 01/29/2022 16:50:32 - INFO - codeparrot_training - Step 21739: {'lr': 0.0003188202054398454, 'samples': 4174080, 'steps': 21739, 'loss/train': 1.3183461427688599} 01/29/2022 16:50:36 - INFO - codeparrot_training - Step 21740: {'lr': 0.00031880447501309787, 'samples': 4174272, 'steps': 21740, 'loss/train': 1.282684475183487} 01/29/2022 16:50:42 - INFO - codeparrot_training - Step 21741: {'lr': 0.0003187887442916139, 'samples': 4174464, 'steps': 21741, 'loss/train': 1.4699395895004272} 01/29/2022 16:50:46 - INFO - codeparrot_training - Step 21742: {'lr': 0.0003187730132754607, 'samples': 4174656, 'steps': 21742, 'loss/train': 0.10173298045992851} 01/29/2022 16:50:50 - INFO - codeparrot_training - Step 21743: {'lr': 0.0003187572819647058, 'samples': 4174848, 'steps': 21743, 'loss/train': 1.8550543785095215} 01/29/2022 16:50:54 - INFO - codeparrot_training - Step 21744: {'lr': 0.00031874155035941656, 'samples': 4175040, 'steps': 21744, 'loss/train': 1.816688597202301} 01/29/2022 16:50:59 - INFO - codeparrot_training - Step 21745: {'lr': 0.00031872581845966024, 'samples': 4175232, 'steps': 21745, 'loss/train': 1.2646970450878143} 01/29/2022 16:51:04 - INFO - codeparrot_training - Step 21746: {'lr': 0.0003187100862655044, 'samples': 4175424, 'steps': 21746, 'loss/train': 1.5526111721992493} 01/29/2022 16:51:08 - INFO - codeparrot_training - Step 21747: {'lr': 0.00031869435377701637, 'samples': 4175616, 'steps': 21747, 'loss/train': 2.46973192691803} 01/29/2022 16:51:12 - INFO - codeparrot_training - Step 21748: {'lr': 0.0003186786209942636, 'samples': 4175808, 'steps': 21748, 'loss/train': 2.1331032514572144} 01/29/2022 16:51:17 - INFO - codeparrot_training - Step 21749: {'lr': 0.00031866288791731334, 'samples': 4176000, 'steps': 21749, 'loss/train': 0.6094626635313034} 01/29/2022 16:51:21 - INFO - codeparrot_training - Step 21750: {'lr': 0.0003186471545462331, 'samples': 4176192, 'steps': 21750, 'loss/train': 2.11534720659256} 01/29/2022 16:51:28 - INFO - codeparrot_training - Step 21751: {'lr': 0.0003186314208810902, 'samples': 4176384, 'steps': 21751, 'loss/train': 1.9271062016487122} 01/29/2022 16:51:32 - INFO - codeparrot_training - Step 21752: {'lr': 0.0003186156869219522, 'samples': 4176576, 'steps': 21752, 'loss/train': 1.2975108325481415} 01/29/2022 16:51:36 - INFO - codeparrot_training - Step 21753: {'lr': 0.0003185999526688863, 'samples': 4176768, 'steps': 21753, 'loss/train': 2.776682138442993} 01/29/2022 16:51:41 - INFO - codeparrot_training - Step 21754: {'lr': 0.00031858421812196, 'samples': 4176960, 'steps': 21754, 'loss/train': 1.9503276944160461} 01/29/2022 16:51:45 - INFO - codeparrot_training - Step 21755: {'lr': 0.0003185684832812407, 'samples': 4177152, 'steps': 21755, 'loss/train': 1.1598930358886719} 01/29/2022 16:51:50 - INFO - codeparrot_training - Step 21756: {'lr': 0.00031855274814679576, 'samples': 4177344, 'steps': 21756, 'loss/train': 1.6222196817398071} 01/29/2022 16:51:54 - INFO - codeparrot_training - Step 21757: {'lr': 0.0003185370127186926, 'samples': 4177536, 'steps': 21757, 'loss/train': 2.2502835988998413} 01/29/2022 16:51:59 - INFO - codeparrot_training - Step 21758: {'lr': 0.00031852127699699874, 'samples': 4177728, 'steps': 21758, 'loss/train': 1.356017678976059} 01/29/2022 16:52:03 - INFO - codeparrot_training - Step 21759: {'lr': 0.0003185055409817814, 'samples': 4177920, 'steps': 21759, 'loss/train': 1.9852674007415771} 01/29/2022 16:52:07 - INFO - codeparrot_training - Step 21760: {'lr': 0.0003184898046731082, 'samples': 4178112, 'steps': 21760, 'loss/train': 1.6927916407585144} 01/29/2022 16:52:14 - INFO - codeparrot_training - Step 21761: {'lr': 0.0003184740680710462, 'samples': 4178304, 'steps': 21761, 'loss/train': 0.8874354958534241} 01/29/2022 16:52:19 - INFO - codeparrot_training - Step 21762: {'lr': 0.00031845833117566326, 'samples': 4178496, 'steps': 21762, 'loss/train': 2.1319981813430786} 01/29/2022 16:52:23 - INFO - codeparrot_training - Step 21763: {'lr': 0.0003184425939870264, 'samples': 4178688, 'steps': 21763, 'loss/train': 1.655391275882721} 01/29/2022 16:52:27 - INFO - codeparrot_training - Step 21764: {'lr': 0.0003184268565052033, 'samples': 4178880, 'steps': 21764, 'loss/train': 1.6270617842674255} 01/29/2022 16:52:31 - INFO - codeparrot_training - Step 21765: {'lr': 0.00031841111873026124, 'samples': 4179072, 'steps': 21765, 'loss/train': 1.486684262752533} 01/29/2022 16:52:37 - INFO - codeparrot_training - Step 21766: {'lr': 0.0003183953806622677, 'samples': 4179264, 'steps': 21766, 'loss/train': 1.0770657062530518} 01/29/2022 16:52:41 - INFO - codeparrot_training - Step 21767: {'lr': 0.00031837964230129004, 'samples': 4179456, 'steps': 21767, 'loss/train': 1.7214832305908203} 01/29/2022 16:52:45 - INFO - codeparrot_training - Step 21768: {'lr': 0.0003183639036473957, 'samples': 4179648, 'steps': 21768, 'loss/train': 0.7153350412845612} 01/29/2022 16:52:49 - INFO - codeparrot_training - Step 21769: {'lr': 0.0003183481647006521, 'samples': 4179840, 'steps': 21769, 'loss/train': 0.8063077926635742} 01/29/2022 16:52:54 - INFO - codeparrot_training - Step 21770: {'lr': 0.0003183324254611267, 'samples': 4180032, 'steps': 21770, 'loss/train': 0.8508336246013641} 01/29/2022 16:52:59 - INFO - codeparrot_training - Step 21771: {'lr': 0.00031831668592888684, 'samples': 4180224, 'steps': 21771, 'loss/train': 0.7759665548801422} 01/29/2022 16:53:03 - INFO - codeparrot_training - Step 21772: {'lr': 0.0003183009461040001, 'samples': 4180416, 'steps': 21772, 'loss/train': 1.1578088700771332} 01/29/2022 16:53:07 - INFO - codeparrot_training - Step 21773: {'lr': 0.0003182852059865337, 'samples': 4180608, 'steps': 21773, 'loss/train': 1.6623017191886902} 01/29/2022 16:53:12 - INFO - codeparrot_training - Step 21774: {'lr': 0.0003182694655765551, 'samples': 4180800, 'steps': 21774, 'loss/train': 1.3997341990470886} 01/29/2022 16:53:16 - INFO - codeparrot_training - Step 21775: {'lr': 0.00031825372487413186, 'samples': 4180992, 'steps': 21775, 'loss/train': 1.4729155004024506} 01/29/2022 16:53:23 - INFO - codeparrot_training - Step 21776: {'lr': 0.00031823798387933133, 'samples': 4181184, 'steps': 21776, 'loss/train': 1.5302125811576843} 01/29/2022 16:53:27 - INFO - codeparrot_training - Step 21777: {'lr': 0.00031822224259222095, 'samples': 4181376, 'steps': 21777, 'loss/train': 1.365927368402481} 01/29/2022 16:53:32 - INFO - codeparrot_training - Step 21778: {'lr': 0.0003182065010128682, 'samples': 4181568, 'steps': 21778, 'loss/train': 0.7887280583381653} 01/29/2022 16:53:36 - INFO - codeparrot_training - Step 21779: {'lr': 0.0003181907591413403, 'samples': 4181760, 'steps': 21779, 'loss/train': 1.5906663537025452} 01/29/2022 16:53:40 - INFO - codeparrot_training - Step 21780: {'lr': 0.00031817501697770496, 'samples': 4181952, 'steps': 21780, 'loss/train': 1.395660638809204} 01/29/2022 16:53:46 - INFO - codeparrot_training - Step 21781: {'lr': 0.00031815927452202955, 'samples': 4182144, 'steps': 21781, 'loss/train': 2.2700196504592896} 01/29/2022 16:53:50 - INFO - codeparrot_training - Step 21782: {'lr': 0.0003181435317743813, 'samples': 4182336, 'steps': 21782, 'loss/train': 0.03640385717153549} 01/29/2022 16:53:54 - INFO - codeparrot_training - Step 21783: {'lr': 0.00031812778873482796, 'samples': 4182528, 'steps': 21783, 'loss/train': 3.6400333642959595} 01/29/2022 16:53:58 - INFO - codeparrot_training - Step 21784: {'lr': 0.00031811204540343666, 'samples': 4182720, 'steps': 21784, 'loss/train': 0.711480125784874} 01/29/2022 16:54:03 - INFO - codeparrot_training - Step 21785: {'lr': 0.00031809630178027506, 'samples': 4182912, 'steps': 21785, 'loss/train': 0.9233027100563049} 01/29/2022 16:54:08 - INFO - codeparrot_training - Step 21786: {'lr': 0.0003180805578654105, 'samples': 4183104, 'steps': 21786, 'loss/train': 1.0405218601226807} 01/29/2022 16:54:12 - INFO - codeparrot_training - Step 21787: {'lr': 0.0003180648136589105, 'samples': 4183296, 'steps': 21787, 'loss/train': 1.0808599591255188} 01/29/2022 16:54:16 - INFO - codeparrot_training - Step 21788: {'lr': 0.00031804906916084235, 'samples': 4183488, 'steps': 21788, 'loss/train': 1.394388735294342} 01/29/2022 16:54:21 - INFO - codeparrot_training - Step 21789: {'lr': 0.0003180333243712737, 'samples': 4183680, 'steps': 21789, 'loss/train': 1.7585182785987854} 01/29/2022 16:54:25 - INFO - codeparrot_training - Step 21790: {'lr': 0.00031801757929027187, 'samples': 4183872, 'steps': 21790, 'loss/train': 1.146304339170456} 01/29/2022 16:54:31 - INFO - codeparrot_training - Step 21791: {'lr': 0.0003180018339179043, 'samples': 4184064, 'steps': 21791, 'loss/train': 1.6995513439178467} 01/29/2022 16:54:35 - INFO - codeparrot_training - Step 21792: {'lr': 0.00031798608825423847, 'samples': 4184256, 'steps': 21792, 'loss/train': 2.247085690498352} 01/29/2022 16:54:39 - INFO - codeparrot_training - Step 21793: {'lr': 0.0003179703422993418, 'samples': 4184448, 'steps': 21793, 'loss/train': 0.6395008563995361} 01/29/2022 16:54:43 - INFO - codeparrot_training - Step 21794: {'lr': 0.00031795459605328183, 'samples': 4184640, 'steps': 21794, 'loss/train': 2.0004347562789917} 01/29/2022 16:54:48 - INFO - codeparrot_training - Step 21795: {'lr': 0.0003179388495161259, 'samples': 4184832, 'steps': 21795, 'loss/train': 2.418584883213043} 01/29/2022 16:54:54 - INFO - codeparrot_training - Step 21796: {'lr': 0.00031792310268794155, 'samples': 4185024, 'steps': 21796, 'loss/train': 1.3734656274318695} 01/29/2022 16:54:59 - INFO - codeparrot_training - Step 21797: {'lr': 0.0003179073555687961, 'samples': 4185216, 'steps': 21797, 'loss/train': 0.5031610876321793} 01/29/2022 16:55:03 - INFO - codeparrot_training - Step 21798: {'lr': 0.00031789160815875724, 'samples': 4185408, 'steps': 21798, 'loss/train': 1.6209300756454468} 01/29/2022 16:55:07 - INFO - codeparrot_training - Step 21799: {'lr': 0.0003178758604578922, 'samples': 4185600, 'steps': 21799, 'loss/train': 0.15209781378507614} 01/29/2022 16:55:11 - INFO - codeparrot_training - Step 21800: {'lr': 0.00031786011246626855, 'samples': 4185792, 'steps': 21800, 'loss/train': 0.8598386943340302} 01/29/2022 16:55:17 - INFO - codeparrot_training - Step 21801: {'lr': 0.00031784436418395373, 'samples': 4185984, 'steps': 21801, 'loss/train': 1.5067782998085022} 01/29/2022 16:55:21 - INFO - codeparrot_training - Step 21802: {'lr': 0.0003178286156110152, 'samples': 4186176, 'steps': 21802, 'loss/train': 1.7469454407691956} 01/29/2022 16:55:26 - INFO - codeparrot_training - Step 21803: {'lr': 0.00031781286674752043, 'samples': 4186368, 'steps': 21803, 'loss/train': 2.3535014390945435} 01/29/2022 16:55:30 - INFO - codeparrot_training - Step 21804: {'lr': 0.00031779711759353683, 'samples': 4186560, 'steps': 21804, 'loss/train': 1.483799010515213} 01/29/2022 16:55:35 - INFO - codeparrot_training - Step 21805: {'lr': 0.00031778136814913195, 'samples': 4186752, 'steps': 21805, 'loss/train': 1.945353627204895} 01/29/2022 16:55:39 - INFO - codeparrot_training - Step 21806: {'lr': 0.0003177656184143732, 'samples': 4186944, 'steps': 21806, 'loss/train': 0.7294582575559616} 01/29/2022 16:55:44 - INFO - codeparrot_training - Step 21807: {'lr': 0.0003177498683893281, 'samples': 4187136, 'steps': 21807, 'loss/train': 1.76069837808609} 01/29/2022 16:55:48 - INFO - codeparrot_training - Step 21808: {'lr': 0.000317734118074064, 'samples': 4187328, 'steps': 21808, 'loss/train': 1.7307066321372986} 01/29/2022 16:55:52 - INFO - codeparrot_training - Step 21809: {'lr': 0.00031771836746864854, 'samples': 4187520, 'steps': 21809, 'loss/train': 1.6784087419509888} 01/29/2022 16:55:59 - INFO - codeparrot_training - Step 21810: {'lr': 0.000317702616573149, 'samples': 4187712, 'steps': 21810, 'loss/train': 1.3553941547870636} 01/29/2022 16:56:04 - INFO - codeparrot_training - Step 21811: {'lr': 0.000317686865387633, 'samples': 4187904, 'steps': 21811, 'loss/train': 1.9346283674240112} 01/29/2022 16:56:08 - INFO - codeparrot_training - Step 21812: {'lr': 0.0003176711139121679, 'samples': 4188096, 'steps': 21812, 'loss/train': 1.0816795527935028} 01/29/2022 16:56:12 - INFO - codeparrot_training - Step 21813: {'lr': 0.00031765536214682134, 'samples': 4188288, 'steps': 21813, 'loss/train': 1.6483667492866516} 01/29/2022 16:56:16 - INFO - codeparrot_training - Step 21814: {'lr': 0.00031763961009166055, 'samples': 4188480, 'steps': 21814, 'loss/train': 1.2279678583145142} 01/29/2022 16:56:21 - INFO - codeparrot_training - Step 21815: {'lr': 0.00031762385774675324, 'samples': 4188672, 'steps': 21815, 'loss/train': 1.7961080074310303} 01/29/2022 16:56:26 - INFO - codeparrot_training - Step 21816: {'lr': 0.0003176081051121668, 'samples': 4188864, 'steps': 21816, 'loss/train': 1.1663611829280853} 01/29/2022 16:56:30 - INFO - codeparrot_training - Step 21817: {'lr': 0.0003175923521879687, 'samples': 4189056, 'steps': 21817, 'loss/train': 1.7693644165992737} 01/29/2022 16:56:34 - INFO - codeparrot_training - Step 21818: {'lr': 0.0003175765989742264, 'samples': 4189248, 'steps': 21818, 'loss/train': 1.8004506826400757} 01/29/2022 16:56:39 - INFO - codeparrot_training - Step 21819: {'lr': 0.0003175608454710074, 'samples': 4189440, 'steps': 21819, 'loss/train': 1.7926608324050903} 01/29/2022 16:56:43 - INFO - codeparrot_training - Step 21820: {'lr': 0.00031754509167837927, 'samples': 4189632, 'steps': 21820, 'loss/train': 1.058098465204239} 01/29/2022 16:56:50 - INFO - codeparrot_training - Step 21821: {'lr': 0.00031752933759640937, 'samples': 4189824, 'steps': 21821, 'loss/train': 0.9235936403274536} 01/29/2022 16:56:54 - INFO - codeparrot_training - Step 21822: {'lr': 0.0003175135832251652, 'samples': 4190016, 'steps': 21822, 'loss/train': 1.7833823561668396} 01/29/2022 16:56:58 - INFO - codeparrot_training - Step 21823: {'lr': 0.00031749782856471426, 'samples': 4190208, 'steps': 21823, 'loss/train': 1.7505347728729248} 01/29/2022 16:57:03 - INFO - codeparrot_training - Step 21824: {'lr': 0.00031748207361512415, 'samples': 4190400, 'steps': 21824, 'loss/train': 1.067306399345398} 01/29/2022 16:57:08 - INFO - codeparrot_training - Step 21825: {'lr': 0.00031746631837646216, 'samples': 4190592, 'steps': 21825, 'loss/train': 1.848725974559784} 01/29/2022 16:57:12 - INFO - codeparrot_training - Step 21826: {'lr': 0.000317450562848796, 'samples': 4190784, 'steps': 21826, 'loss/train': 0.9604478180408478} 01/29/2022 16:57:16 - INFO - codeparrot_training - Step 21827: {'lr': 0.00031743480703219293, 'samples': 4190976, 'steps': 21827, 'loss/train': 2.0893993377685547} 01/29/2022 16:57:21 - INFO - codeparrot_training - Step 21828: {'lr': 0.00031741905092672057, 'samples': 4191168, 'steps': 21828, 'loss/train': 1.7101664543151855} 01/29/2022 16:57:25 - INFO - codeparrot_training - Step 21829: {'lr': 0.0003174032945324465, 'samples': 4191360, 'steps': 21829, 'loss/train': 2.1920260190963745} 01/29/2022 16:57:30 - INFO - codeparrot_training - Step 21830: {'lr': 0.00031738753784943803, 'samples': 4191552, 'steps': 21830, 'loss/train': 1.4534990787506104} 01/29/2022 16:57:34 - INFO - codeparrot_training - Step 21831: {'lr': 0.0003173717808777628, 'samples': 4191744, 'steps': 21831, 'loss/train': 0.6142755299806595} 01/29/2022 16:57:39 - INFO - codeparrot_training - Step 21832: {'lr': 0.00031735602361748815, 'samples': 4191936, 'steps': 21832, 'loss/train': 1.5211522579193115} 01/29/2022 16:57:43 - INFO - codeparrot_training - Step 21833: {'lr': 0.00031734026606868184, 'samples': 4192128, 'steps': 21833, 'loss/train': 2.106226623058319} 01/29/2022 16:57:47 - INFO - codeparrot_training - Step 21834: {'lr': 0.0003173245082314111, 'samples': 4192320, 'steps': 21834, 'loss/train': 1.3320466876029968} 01/29/2022 16:57:54 - INFO - codeparrot_training - Step 21835: {'lr': 0.0003173087501057436, 'samples': 4192512, 'steps': 21835, 'loss/train': 1.0445381104946136} 01/29/2022 16:57:59 - INFO - codeparrot_training - Step 21836: {'lr': 0.00031729299169174673, 'samples': 4192704, 'steps': 21836, 'loss/train': 1.4093973934650421} 01/29/2022 16:58:03 - INFO - codeparrot_training - Step 21837: {'lr': 0.0003172772329894882, 'samples': 4192896, 'steps': 21837, 'loss/train': 1.4528074264526367} 01/29/2022 16:58:07 - INFO - codeparrot_training - Step 21838: {'lr': 0.0003172614739990352, 'samples': 4193088, 'steps': 21838, 'loss/train': 1.761193871498108} 01/29/2022 16:58:11 - INFO - codeparrot_training - Step 21839: {'lr': 0.0003172457147204554, 'samples': 4193280, 'steps': 21839, 'loss/train': 1.1980274319648743} 01/29/2022 16:58:17 - INFO - codeparrot_training - Step 21840: {'lr': 0.0003172299551538164, 'samples': 4193472, 'steps': 21840, 'loss/train': 1.9360937476158142} 01/29/2022 16:58:21 - INFO - codeparrot_training - Step 21841: {'lr': 0.0003172141952991856, 'samples': 4193664, 'steps': 21841, 'loss/train': 1.2081020772457123} 01/29/2022 16:58:25 - INFO - codeparrot_training - Step 21842: {'lr': 0.00031719843515663055, 'samples': 4193856, 'steps': 21842, 'loss/train': 1.6817071437835693} 01/29/2022 16:58:30 - INFO - codeparrot_training - Step 21843: {'lr': 0.0003171826747262187, 'samples': 4194048, 'steps': 21843, 'loss/train': 0.7064977437257767} 01/29/2022 16:58:34 - INFO - codeparrot_training - Step 21844: {'lr': 0.0003171669140080177, 'samples': 4194240, 'steps': 21844, 'loss/train': 1.393408477306366} 01/29/2022 16:58:41 - INFO - codeparrot_training - Step 21845: {'lr': 0.00031715115300209477, 'samples': 4194432, 'steps': 21845, 'loss/train': 0.9350126087665558} 01/29/2022 16:58:45 - INFO - codeparrot_training - Step 21846: {'lr': 0.0003171353917085178, 'samples': 4194624, 'steps': 21846, 'loss/train': 0.8134288787841797} 01/29/2022 16:58:49 - INFO - codeparrot_training - Step 21847: {'lr': 0.00031711963012735414, 'samples': 4194816, 'steps': 21847, 'loss/train': 0.8366931080818176} 01/29/2022 16:58:54 - INFO - codeparrot_training - Step 21848: {'lr': 0.0003171038682586712, 'samples': 4195008, 'steps': 21848, 'loss/train': 0.6152466684579849} 01/29/2022 16:58:58 - INFO - codeparrot_training - Step 21849: {'lr': 0.0003170881061025366, 'samples': 4195200, 'steps': 21849, 'loss/train': 1.4330256879329681} 01/29/2022 16:59:03 - INFO - codeparrot_training - Step 21850: {'lr': 0.00031707234365901786, 'samples': 4195392, 'steps': 21850, 'loss/train': 1.4909862577915192} 01/29/2022 16:59:08 - INFO - codeparrot_training - Step 21851: {'lr': 0.0003170565809281826, 'samples': 4195584, 'steps': 21851, 'loss/train': 1.4750606417655945} 01/29/2022 16:59:12 - INFO - codeparrot_training - Step 21852: {'lr': 0.0003170408179100981, 'samples': 4195776, 'steps': 21852, 'loss/train': 1.6731877326965332} 01/29/2022 16:59:16 - INFO - codeparrot_training - Step 21853: {'lr': 0.0003170250546048321, 'samples': 4195968, 'steps': 21853, 'loss/train': 1.1722542643547058} 01/29/2022 16:59:20 - INFO - codeparrot_training - Step 21854: {'lr': 0.000317009291012452, 'samples': 4196160, 'steps': 21854, 'loss/train': 0.49439477920532227} 01/29/2022 16:59:26 - INFO - codeparrot_training - Step 21855: {'lr': 0.00031699352713302544, 'samples': 4196352, 'steps': 21855, 'loss/train': 1.355690360069275} 01/29/2022 16:59:30 - INFO - codeparrot_training - Step 21856: {'lr': 0.00031697776296661987, 'samples': 4196544, 'steps': 21856, 'loss/train': 1.7802395224571228} 01/29/2022 16:59:34 - INFO - codeparrot_training - Step 21857: {'lr': 0.0003169619985133028, 'samples': 4196736, 'steps': 21857, 'loss/train': 1.5989415049552917} 01/29/2022 16:59:39 - INFO - codeparrot_training - Step 21858: {'lr': 0.0003169462337731418, 'samples': 4196928, 'steps': 21858, 'loss/train': 1.7052034735679626} 01/29/2022 16:59:43 - INFO - codeparrot_training - Step 21859: {'lr': 0.0003169304687462044, 'samples': 4197120, 'steps': 21859, 'loss/train': 1.3602233827114105} 01/29/2022 16:59:48 - INFO - codeparrot_training - Step 21860: {'lr': 0.00031691470343255814, 'samples': 4197312, 'steps': 21860, 'loss/train': 1.2049561142921448} 01/29/2022 16:59:52 - INFO - codeparrot_training - Step 21861: {'lr': 0.00031689893783227053, 'samples': 4197504, 'steps': 21861, 'loss/train': 1.8848522901535034} 01/29/2022 16:59:57 - INFO - codeparrot_training - Step 21862: {'lr': 0.00031688317194540904, 'samples': 4197696, 'steps': 21862, 'loss/train': 1.7517718076705933} 01/29/2022 17:00:01 - INFO - codeparrot_training - Step 21863: {'lr': 0.0003168674057720413, 'samples': 4197888, 'steps': 21863, 'loss/train': 1.7948040962219238} 01/29/2022 17:00:05 - INFO - codeparrot_training - Step 21864: {'lr': 0.0003168516393122349, 'samples': 4198080, 'steps': 21864, 'loss/train': 1.9245446920394897} 01/29/2022 17:00:09 - INFO - codeparrot_training - Step 21865: {'lr': 0.0003168358725660573, 'samples': 4198272, 'steps': 21865, 'loss/train': 1.2946361303329468} 01/29/2022 17:00:17 - INFO - codeparrot_training - Step 21866: {'lr': 0.000316820105533576, 'samples': 4198464, 'steps': 21866, 'loss/train': 1.6310498714447021} 01/29/2022 17:00:21 - INFO - codeparrot_training - Step 21867: {'lr': 0.0003168043382148586, 'samples': 4198656, 'steps': 21867, 'loss/train': 1.6063522696495056} 01/29/2022 17:00:25 - INFO - codeparrot_training - Step 21868: {'lr': 0.0003167885706099726, 'samples': 4198848, 'steps': 21868, 'loss/train': 1.4796923696994781} 01/29/2022 17:00:30 - INFO - codeparrot_training - Step 21869: {'lr': 0.0003167728027189856, 'samples': 4199040, 'steps': 21869, 'loss/train': 1.9999296069145203} 01/29/2022 17:00:35 - INFO - codeparrot_training - Step 21870: {'lr': 0.00031675703454196513, 'samples': 4199232, 'steps': 21870, 'loss/train': 1.1867943406105042} 01/29/2022 17:00:39 - INFO - codeparrot_training - Step 21871: {'lr': 0.00031674126607897867, 'samples': 4199424, 'steps': 21871, 'loss/train': 1.9789668917655945} 01/29/2022 17:00:43 - INFO - codeparrot_training - Step 21872: {'lr': 0.00031672549733009395, 'samples': 4199616, 'steps': 21872, 'loss/train': 1.0743849277496338} 01/29/2022 17:00:48 - INFO - codeparrot_training - Step 21873: {'lr': 0.00031670972829537825, 'samples': 4199808, 'steps': 21873, 'loss/train': 1.0870343148708344} 01/29/2022 17:00:52 - INFO - codeparrot_training - Step 21874: {'lr': 0.0003166939589748993, 'samples': 4200000, 'steps': 21874, 'loss/train': 0.8027084469795227} 01/29/2022 17:00:57 - INFO - codeparrot_training - Step 21875: {'lr': 0.0003166781893687246, 'samples': 4200192, 'steps': 21875, 'loss/train': 1.48776376247406} 01/29/2022 17:01:01 - INFO - codeparrot_training - Step 21876: {'lr': 0.00031666241947692173, 'samples': 4200384, 'steps': 21876, 'loss/train': 1.1350001692771912} 01/29/2022 17:01:06 - INFO - codeparrot_training - Step 21877: {'lr': 0.0003166466492995582, 'samples': 4200576, 'steps': 21877, 'loss/train': 1.5760105848312378} 01/29/2022 17:01:10 - INFO - codeparrot_training - Step 21878: {'lr': 0.0003166308788367016, 'samples': 4200768, 'steps': 21878, 'loss/train': 1.2880882322788239} 01/29/2022 17:01:14 - INFO - codeparrot_training - Step 21879: {'lr': 0.00031661510808841947, 'samples': 4200960, 'steps': 21879, 'loss/train': 1.5061152577400208} 01/29/2022 17:01:21 - INFO - codeparrot_training - Step 21880: {'lr': 0.0003165993370547794, 'samples': 4201152, 'steps': 21880, 'loss/train': 2.3875350952148438} 01/29/2022 17:01:25 - INFO - codeparrot_training - Step 21881: {'lr': 0.0003165835657358489, 'samples': 4201344, 'steps': 21881, 'loss/train': 1.837882161140442} 01/29/2022 17:01:30 - INFO - codeparrot_training - Step 21882: {'lr': 0.00031656779413169543, 'samples': 4201536, 'steps': 21882, 'loss/train': 1.338644653558731} 01/29/2022 17:01:34 - INFO - codeparrot_training - Step 21883: {'lr': 0.00031655202224238686, 'samples': 4201728, 'steps': 21883, 'loss/train': 1.350147932767868} 01/29/2022 17:01:38 - INFO - codeparrot_training - Step 21884: {'lr': 0.0003165362500679905, 'samples': 4201920, 'steps': 21884, 'loss/train': 2.19058620929718} 01/29/2022 17:01:43 - INFO - codeparrot_training - Step 21885: {'lr': 0.00031652047760857393, 'samples': 4202112, 'steps': 21885, 'loss/train': 0.8803646564483643} 01/29/2022 17:01:48 - INFO - codeparrot_training - Step 21886: {'lr': 0.0003165047048642047, 'samples': 4202304, 'steps': 21886, 'loss/train': 1.9570280313491821} 01/29/2022 17:01:52 - INFO - codeparrot_training - Step 21887: {'lr': 0.00031648893183495053, 'samples': 4202496, 'steps': 21887, 'loss/train': 1.4002558887004852} 01/29/2022 17:01:56 - INFO - codeparrot_training - Step 21888: {'lr': 0.0003164731585208789, 'samples': 4202688, 'steps': 21888, 'loss/train': 1.4314650893211365} 01/29/2022 17:02:01 - INFO - codeparrot_training - Step 21889: {'lr': 0.00031645738492205736, 'samples': 4202880, 'steps': 21889, 'loss/train': 2.1127002239227295} 01/29/2022 17:02:07 - INFO - codeparrot_training - Step 21890: {'lr': 0.0003164416110385534, 'samples': 4203072, 'steps': 21890, 'loss/train': 1.4634187817573547} 01/29/2022 17:02:12 - INFO - codeparrot_training - Step 21891: {'lr': 0.0003164258368704347, 'samples': 4203264, 'steps': 21891, 'loss/train': 1.8652063608169556} 01/29/2022 17:02:16 - INFO - codeparrot_training - Step 21892: {'lr': 0.00031641006241776886, 'samples': 4203456, 'steps': 21892, 'loss/train': 1.6418184041976929} 01/29/2022 17:02:20 - INFO - codeparrot_training - Step 21893: {'lr': 0.0003163942876806234, 'samples': 4203648, 'steps': 21893, 'loss/train': 2.0900720357894897} 01/29/2022 17:02:24 - INFO - codeparrot_training - Step 21894: {'lr': 0.00031637851265906594, 'samples': 4203840, 'steps': 21894, 'loss/train': 1.8542656302452087} 01/29/2022 17:02:30 - INFO - codeparrot_training - Step 21895: {'lr': 0.0003163627373531639, 'samples': 4204032, 'steps': 21895, 'loss/train': 1.8499088287353516} 01/29/2022 17:02:34 - INFO - codeparrot_training - Step 21896: {'lr': 0.000316346961762985, 'samples': 4204224, 'steps': 21896, 'loss/train': 0.5030748546123505} 01/29/2022 17:02:38 - INFO - codeparrot_training - Step 21897: {'lr': 0.00031633118588859677, 'samples': 4204416, 'steps': 21897, 'loss/train': 0.8020403981208801} 01/29/2022 17:02:43 - INFO - codeparrot_training - Step 21898: {'lr': 0.00031631540973006683, 'samples': 4204608, 'steps': 21898, 'loss/train': 1.4467918574810028} 01/29/2022 17:02:47 - INFO - codeparrot_training - Step 21899: {'lr': 0.0003162996332874627, 'samples': 4204800, 'steps': 21899, 'loss/train': 2.06804358959198} 01/29/2022 17:02:55 - INFO - codeparrot_training - Step 21900: {'lr': 0.000316283856560852, 'samples': 4204992, 'steps': 21900, 'loss/train': 1.6934577226638794} 01/29/2022 17:02:59 - INFO - codeparrot_training - Step 21901: {'lr': 0.00031626807955030236, 'samples': 4205184, 'steps': 21901, 'loss/train': 1.247570425271988} 01/29/2022 17:03:03 - INFO - codeparrot_training - Step 21902: {'lr': 0.00031625230225588123, 'samples': 4205376, 'steps': 21902, 'loss/train': 0.9752469062805176} 01/29/2022 17:03:07 - INFO - codeparrot_training - Step 21903: {'lr': 0.0003162365246776564, 'samples': 4205568, 'steps': 21903, 'loss/train': 1.7275843620300293} 01/29/2022 17:03:12 - INFO - codeparrot_training - Step 21904: {'lr': 0.0003162207468156952, 'samples': 4205760, 'steps': 21904, 'loss/train': 1.2703028619289398} 01/29/2022 17:03:17 - INFO - codeparrot_training - Step 21905: {'lr': 0.00031620496867006543, 'samples': 4205952, 'steps': 21905, 'loss/train': 1.5578291416168213} 01/29/2022 17:03:21 - INFO - codeparrot_training - Step 21906: {'lr': 0.0003161891902408345, 'samples': 4206144, 'steps': 21906, 'loss/train': 3.07001531124115} 01/29/2022 17:03:25 - INFO - codeparrot_training - Step 21907: {'lr': 0.00031617341152807024, 'samples': 4206336, 'steps': 21907, 'loss/train': 1.817562460899353} 01/29/2022 17:03:30 - INFO - codeparrot_training - Step 21908: {'lr': 0.00031615763253183996, 'samples': 4206528, 'steps': 21908, 'loss/train': 1.4192020297050476} 01/29/2022 17:03:34 - INFO - codeparrot_training - Step 21909: {'lr': 0.00031614185325221143, 'samples': 4206720, 'steps': 21909, 'loss/train': 1.6752923727035522} 01/29/2022 17:03:39 - INFO - codeparrot_training - Step 21910: {'lr': 0.0003161260736892523, 'samples': 4206912, 'steps': 21910, 'loss/train': 1.583778977394104} 01/29/2022 17:03:44 - INFO - codeparrot_training - Step 21911: {'lr': 0.00031611029384302997, 'samples': 4207104, 'steps': 21911, 'loss/train': 2.00538432598114} 01/29/2022 17:03:48 - INFO - codeparrot_training - Step 21912: {'lr': 0.0003160945137136121, 'samples': 4207296, 'steps': 21912, 'loss/train': 1.0898158550262451} 01/29/2022 17:03:52 - INFO - codeparrot_training - Step 21913: {'lr': 0.0003160787333010664, 'samples': 4207488, 'steps': 21913, 'loss/train': 1.4969978034496307} 01/29/2022 17:03:56 - INFO - codeparrot_training - Step 21914: {'lr': 0.00031606295260546037, 'samples': 4207680, 'steps': 21914, 'loss/train': 1.7088069319725037} 01/29/2022 17:04:05 - INFO - codeparrot_training - Step 21915: {'lr': 0.00031604717162686156, 'samples': 4207872, 'steps': 21915, 'loss/train': 1.7773311138153076} 01/29/2022 17:04:09 - INFO - codeparrot_training - Step 21916: {'lr': 0.00031603139036533775, 'samples': 4208064, 'steps': 21916, 'loss/train': 1.7946009635925293} 01/29/2022 17:04:14 - INFO - codeparrot_training - Step 21917: {'lr': 0.0003160156088209564, 'samples': 4208256, 'steps': 21917, 'loss/train': 1.4791532456874847} 01/29/2022 17:04:18 - INFO - codeparrot_training - Step 21918: {'lr': 0.0003159998269937851, 'samples': 4208448, 'steps': 21918, 'loss/train': 1.7265422344207764} 01/29/2022 17:04:22 - INFO - codeparrot_training - Step 21919: {'lr': 0.0003159840448838915, 'samples': 4208640, 'steps': 21919, 'loss/train': 1.7834054231643677} 01/29/2022 17:04:28 - INFO - codeparrot_training - Step 21920: {'lr': 0.00031596826249134324, 'samples': 4208832, 'steps': 21920, 'loss/train': 1.6646021604537964} 01/29/2022 17:04:32 - INFO - codeparrot_training - Step 21921: {'lr': 0.0003159524798162079, 'samples': 4209024, 'steps': 21921, 'loss/train': 0.9384802579879761} 01/29/2022 17:04:36 - INFO - codeparrot_training - Step 21922: {'lr': 0.000315936696858553, 'samples': 4209216, 'steps': 21922, 'loss/train': 1.871411383152008} 01/29/2022 17:04:40 - INFO - codeparrot_training - Step 21923: {'lr': 0.00031592091361844633, 'samples': 4209408, 'steps': 21923, 'loss/train': 1.8290274739265442} 01/29/2022 17:04:45 - INFO - codeparrot_training - Step 21924: {'lr': 0.0003159051300959553, 'samples': 4209600, 'steps': 21924, 'loss/train': 1.7425875663757324} 01/29/2022 17:04:49 - INFO - codeparrot_training - Step 21925: {'lr': 0.0003158893462911477, 'samples': 4209792, 'steps': 21925, 'loss/train': 1.4397013485431671} 01/29/2022 17:04:56 - INFO - codeparrot_training - Step 21926: {'lr': 0.00031587356220409105, 'samples': 4209984, 'steps': 21926, 'loss/train': 1.2317276895046234} 01/29/2022 17:05:01 - INFO - codeparrot_training - Step 21927: {'lr': 0.0003158577778348529, 'samples': 4210176, 'steps': 21927, 'loss/train': 1.4103920459747314} 01/29/2022 17:05:05 - INFO - codeparrot_training - Step 21928: {'lr': 0.000315841993183501, 'samples': 4210368, 'steps': 21928, 'loss/train': 1.81086927652359} 01/29/2022 17:05:09 - INFO - codeparrot_training - Step 21929: {'lr': 0.000315826208250103, 'samples': 4210560, 'steps': 21929, 'loss/train': 1.1471849083900452} 01/29/2022 17:05:13 - INFO - codeparrot_training - Step 21930: {'lr': 0.0003158104230347263, 'samples': 4210752, 'steps': 21930, 'loss/train': 1.2368464171886444} 01/29/2022 17:05:19 - INFO - codeparrot_training - Step 21931: {'lr': 0.0003157946375374387, 'samples': 4210944, 'steps': 21931, 'loss/train': 1.9135052561759949} 01/29/2022 17:05:23 - INFO - codeparrot_training - Step 21932: {'lr': 0.0003157788517583077, 'samples': 4211136, 'steps': 21932, 'loss/train': 1.7393762469291687} 01/29/2022 17:05:28 - INFO - codeparrot_training - Step 21933: {'lr': 0.00031576306569740107, 'samples': 4211328, 'steps': 21933, 'loss/train': 1.1099904477596283} 01/29/2022 17:05:32 - INFO - codeparrot_training - Step 21934: {'lr': 0.00031574727935478633, 'samples': 4211520, 'steps': 21934, 'loss/train': 1.5629514455795288} 01/29/2022 17:05:36 - INFO - codeparrot_training - Step 21935: {'lr': 0.0003157314927305311, 'samples': 4211712, 'steps': 21935, 'loss/train': 1.9370695352554321} 01/29/2022 17:05:42 - INFO - codeparrot_training - Step 21936: {'lr': 0.00031571570582470304, 'samples': 4211904, 'steps': 21936, 'loss/train': 1.2328112125396729} 01/29/2022 17:05:46 - INFO - codeparrot_training - Step 21937: {'lr': 0.00031569991863736975, 'samples': 4212096, 'steps': 21937, 'loss/train': 1.6405768990516663} 01/29/2022 17:05:50 - INFO - codeparrot_training - Step 21938: {'lr': 0.00031568413116859887, 'samples': 4212288, 'steps': 21938, 'loss/train': 2.520641505718231} 01/29/2022 17:05:55 - INFO - codeparrot_training - Step 21939: {'lr': 0.000315668343418458, 'samples': 4212480, 'steps': 21939, 'loss/train': 2.027207672595978} 01/29/2022 17:05:59 - INFO - codeparrot_training - Step 21940: {'lr': 0.0003156525553870149, 'samples': 4212672, 'steps': 21940, 'loss/train': 0.5249409377574921} 01/29/2022 17:06:07 - INFO - codeparrot_training - Step 21941: {'lr': 0.000315636767074337, 'samples': 4212864, 'steps': 21941, 'loss/train': 1.4296153485774994} 01/29/2022 17:06:11 - INFO - codeparrot_training - Step 21942: {'lr': 0.00031562097848049207, 'samples': 4213056, 'steps': 21942, 'loss/train': 0.8040631413459778} 01/29/2022 17:06:15 - INFO - codeparrot_training - Step 21943: {'lr': 0.0003156051896055477, 'samples': 4213248, 'steps': 21943, 'loss/train': 2.2117486596107483} 01/29/2022 17:06:20 - INFO - codeparrot_training - Step 21944: {'lr': 0.0003155894004495716, 'samples': 4213440, 'steps': 21944, 'loss/train': 1.010810226202011} 01/29/2022 17:06:24 - INFO - codeparrot_training - Step 21945: {'lr': 0.0003155736110126312, 'samples': 4213632, 'steps': 21945, 'loss/train': 1.2275317311286926} 01/29/2022 17:06:29 - INFO - codeparrot_training - Step 21946: {'lr': 0.0003155578212947944, 'samples': 4213824, 'steps': 21946, 'loss/train': 1.8536049127578735} 01/29/2022 17:06:33 - INFO - codeparrot_training - Step 21947: {'lr': 0.0003155420312961286, 'samples': 4214016, 'steps': 21947, 'loss/train': 1.6266506910324097} 01/29/2022 17:06:38 - INFO - codeparrot_training - Step 21948: {'lr': 0.00031552624101670155, 'samples': 4214208, 'steps': 21948, 'loss/train': 1.7645124793052673} 01/29/2022 17:06:42 - INFO - codeparrot_training - Step 21949: {'lr': 0.000315510450456581, 'samples': 4214400, 'steps': 21949, 'loss/train': 0.9222072064876556} 01/29/2022 17:06:46 - INFO - codeparrot_training - Step 21950: {'lr': 0.0003154946596158343, 'samples': 4214592, 'steps': 21950, 'loss/train': 1.434403896331787} 01/29/2022 17:06:53 - INFO - codeparrot_training - Step 21951: {'lr': 0.0003154788684945295, 'samples': 4214784, 'steps': 21951, 'loss/train': 2.2499483227729797} 01/29/2022 17:06:58 - INFO - codeparrot_training - Step 21952: {'lr': 0.00031546307709273393, 'samples': 4214976, 'steps': 21952, 'loss/train': 1.5642866492271423} 01/29/2022 17:07:02 - INFO - codeparrot_training - Step 21953: {'lr': 0.0003154472854105153, 'samples': 4215168, 'steps': 21953, 'loss/train': 1.4844369292259216} 01/29/2022 17:07:06 - INFO - codeparrot_training - Step 21954: {'lr': 0.0003154314934479413, 'samples': 4215360, 'steps': 21954, 'loss/train': 0.9961740374565125} 01/29/2022 17:07:10 - INFO - codeparrot_training - Step 21955: {'lr': 0.00031541570120507956, 'samples': 4215552, 'steps': 21955, 'loss/train': 1.8764230012893677} 01/29/2022 17:07:16 - INFO - codeparrot_training - Step 21956: {'lr': 0.0003153999086819977, 'samples': 4215744, 'steps': 21956, 'loss/train': 1.2114048600196838} 01/29/2022 17:07:20 - INFO - codeparrot_training - Step 21957: {'lr': 0.00031538411587876344, 'samples': 4215936, 'steps': 21957, 'loss/train': 1.8651553988456726} 01/29/2022 17:07:24 - INFO - codeparrot_training - Step 21958: {'lr': 0.0003153683227954443, 'samples': 4216128, 'steps': 21958, 'loss/train': 1.451774775981903} 01/29/2022 17:07:28 - INFO - codeparrot_training - Step 21959: {'lr': 0.00031535252943210815, 'samples': 4216320, 'steps': 21959, 'loss/train': 5.734632611274719} 01/29/2022 17:07:33 - INFO - codeparrot_training - Step 21960: {'lr': 0.0003153367357888224, 'samples': 4216512, 'steps': 21960, 'loss/train': 2.0704899430274963} 01/29/2022 17:07:38 - INFO - codeparrot_training - Step 21961: {'lr': 0.00031532094186565487, 'samples': 4216704, 'steps': 21961, 'loss/train': 1.510785162448883} 01/29/2022 17:07:42 - INFO - codeparrot_training - Step 21962: {'lr': 0.0003153051476626732, 'samples': 4216896, 'steps': 21962, 'loss/train': 1.2958983778953552} 01/29/2022 17:07:47 - INFO - codeparrot_training - Step 21963: {'lr': 0.00031528935317994493, 'samples': 4217088, 'steps': 21963, 'loss/train': 1.6604715585708618} 01/29/2022 17:07:51 - INFO - codeparrot_training - Step 21964: {'lr': 0.00031527355841753793, 'samples': 4217280, 'steps': 21964, 'loss/train': 2.1937551498413086} 01/29/2022 17:07:55 - INFO - codeparrot_training - Step 21965: {'lr': 0.0003152577633755196, 'samples': 4217472, 'steps': 21965, 'loss/train': 1.8607547879219055} 01/29/2022 17:08:01 - INFO - codeparrot_training - Step 21966: {'lr': 0.0003152419680539578, 'samples': 4217664, 'steps': 21966, 'loss/train': 1.8116006255149841} 01/29/2022 17:08:05 - INFO - codeparrot_training - Step 21967: {'lr': 0.0003152261724529201, 'samples': 4217856, 'steps': 21967, 'loss/train': 1.5426276326179504} 01/29/2022 17:08:09 - INFO - codeparrot_training - Step 21968: {'lr': 0.0003152103765724743, 'samples': 4218048, 'steps': 21968, 'loss/train': 2.907401204109192} 01/29/2022 17:08:13 - INFO - codeparrot_training - Step 21969: {'lr': 0.0003151945804126878, 'samples': 4218240, 'steps': 21969, 'loss/train': 1.4348367154598236} 01/29/2022 17:08:18 - INFO - codeparrot_training - Step 21970: {'lr': 0.00031517878397362847, 'samples': 4218432, 'steps': 21970, 'loss/train': 1.490154504776001} 01/29/2022 17:08:25 - INFO - codeparrot_training - Step 21971: {'lr': 0.00031516298725536396, 'samples': 4218624, 'steps': 21971, 'loss/train': 2.0278578400611877} 01/29/2022 17:08:29 - INFO - codeparrot_training - Step 21972: {'lr': 0.00031514719025796183, 'samples': 4218816, 'steps': 21972, 'loss/train': 1.9266638159751892} 01/29/2022 17:08:34 - INFO - codeparrot_training - Step 21973: {'lr': 0.0003151313929814899, 'samples': 4219008, 'steps': 21973, 'loss/train': 1.2713515162467957} 01/29/2022 17:08:38 - INFO - codeparrot_training - Step 21974: {'lr': 0.0003151155954260157, 'samples': 4219200, 'steps': 21974, 'loss/train': 2.821620762348175} 01/29/2022 17:08:42 - INFO - codeparrot_training - Step 21975: {'lr': 0.00031509979759160707, 'samples': 4219392, 'steps': 21975, 'loss/train': 0.6480447202920914} 01/29/2022 17:08:47 - INFO - codeparrot_training - Step 21976: {'lr': 0.00031508399947833155, 'samples': 4219584, 'steps': 21976, 'loss/train': 1.624371349811554} 01/29/2022 17:08:52 - INFO - codeparrot_training - Step 21977: {'lr': 0.0003150682010862568, 'samples': 4219776, 'steps': 21977, 'loss/train': 2.2474345564842224} 01/29/2022 17:08:56 - INFO - codeparrot_training - Step 21978: {'lr': 0.00031505240241545054, 'samples': 4219968, 'steps': 21978, 'loss/train': 1.4283588230609894} 01/29/2022 17:09:00 - INFO - codeparrot_training - Step 21979: {'lr': 0.0003150366034659804, 'samples': 4220160, 'steps': 21979, 'loss/train': 1.228676587343216} 01/29/2022 17:09:04 - INFO - codeparrot_training - Step 21980: {'lr': 0.00031502080423791417, 'samples': 4220352, 'steps': 21980, 'loss/train': 2.111623227596283} 01/29/2022 17:09:10 - INFO - codeparrot_training - Step 21981: {'lr': 0.00031500500473131943, 'samples': 4220544, 'steps': 21981, 'loss/train': 1.6641186475753784} 01/29/2022 17:09:14 - INFO - codeparrot_training - Step 21982: {'lr': 0.00031498920494626396, 'samples': 4220736, 'steps': 21982, 'loss/train': 1.262068748474121} 01/29/2022 17:09:18 - INFO - codeparrot_training - Step 21983: {'lr': 0.00031497340488281534, 'samples': 4220928, 'steps': 21983, 'loss/train': 0.2394810989499092} 01/29/2022 17:09:23 - INFO - codeparrot_training - Step 21984: {'lr': 0.0003149576045410412, 'samples': 4221120, 'steps': 21984, 'loss/train': 1.8789442777633667} 01/29/2022 17:09:27 - INFO - codeparrot_training - Step 21985: {'lr': 0.00031494180392100936, 'samples': 4221312, 'steps': 21985, 'loss/train': 2.058067560195923} 01/29/2022 17:09:34 - INFO - codeparrot_training - Step 21986: {'lr': 0.00031492600302278746, 'samples': 4221504, 'steps': 21986, 'loss/train': 2.0062177777290344} 01/29/2022 17:09:38 - INFO - codeparrot_training - Step 21987: {'lr': 0.00031491020184644316, 'samples': 4221696, 'steps': 21987, 'loss/train': 1.7955408096313477} 01/29/2022 17:09:42 - INFO - codeparrot_training - Step 21988: {'lr': 0.0003148944003920442, 'samples': 4221888, 'steps': 21988, 'loss/train': 1.6403419375419617} 01/29/2022 17:09:47 - INFO - codeparrot_training - Step 21989: {'lr': 0.00031487859865965824, 'samples': 4222080, 'steps': 21989, 'loss/train': 0.7921958863735199} 01/29/2022 17:09:51 - INFO - codeparrot_training - Step 21990: {'lr': 0.000314862796649353, 'samples': 4222272, 'steps': 21990, 'loss/train': 1.575096309185028} 01/29/2022 17:09:56 - INFO - codeparrot_training - Step 21991: {'lr': 0.000314846994361196, 'samples': 4222464, 'steps': 21991, 'loss/train': 1.5445078611373901} 01/29/2022 17:10:00 - INFO - codeparrot_training - Step 21992: {'lr': 0.0003148311917952552, 'samples': 4222656, 'steps': 21992, 'loss/train': 1.301576167345047} 01/29/2022 17:10:05 - INFO - codeparrot_training - Step 21993: {'lr': 0.00031481538895159806, 'samples': 4222848, 'steps': 21993, 'loss/train': 2.365061581134796} 01/29/2022 17:10:09 - INFO - codeparrot_training - Step 21994: {'lr': 0.00031479958583029247, 'samples': 4223040, 'steps': 21994, 'loss/train': 1.181378960609436} 01/29/2022 17:10:13 - INFO - codeparrot_training - Step 21995: {'lr': 0.0003147837824314059, 'samples': 4223232, 'steps': 21995, 'loss/train': 1.2679975032806396} 01/29/2022 17:10:20 - INFO - codeparrot_training - Step 21996: {'lr': 0.00031476797875500627, 'samples': 4223424, 'steps': 21996, 'loss/train': 1.8905598521232605} 01/29/2022 17:10:24 - INFO - codeparrot_training - Step 21997: {'lr': 0.00031475217480116124, 'samples': 4223616, 'steps': 21997, 'loss/train': 1.374927431344986} 01/29/2022 17:10:29 - INFO - codeparrot_training - Step 21998: {'lr': 0.00031473637056993837, 'samples': 4223808, 'steps': 21998, 'loss/train': 1.3815385401248932} 01/29/2022 17:10:33 - INFO - codeparrot_training - Step 21999: {'lr': 0.0003147205660614055, 'samples': 4224000, 'steps': 21999, 'loss/train': 1.090144783258438} 01/29/2022 17:10:33 - INFO - codeparrot_training - Evaluating and saving model checkpoint 01/29/2022 17:11:06 - WARNING - huggingface_hub.repository - Several commits (11) will be pushed upstream. 01/29/2022 17:11:06 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 01/29/2022 17:12:10 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/neo-code-py 5707af2..7f31785 expert-dust-2 -> expert-dust-2 01/29/2022 17:12:15 - INFO - codeparrot_training - Step 22000: {'lr': 0.00031470476127563017, 'samples': 4224192, 'steps': 22000, 'loss/train': 0.7294651418924332} 01/29/2022 17:12:21 - INFO - codeparrot_training - Step 22001: {'lr': 0.00031468895621268036, 'samples': 4224384, 'steps': 22001, 'loss/train': 2.3257076740264893} 01/29/2022 17:12:25 - INFO - codeparrot_training - Step 22002: {'lr': 0.00031467315087262344, 'samples': 4224576, 'steps': 22002, 'loss/train': 2.015672028064728} 01/29/2022 17:12:29 - INFO - codeparrot_training - Step 22003: {'lr': 0.0003146573452555274, 'samples': 4224768, 'steps': 22003, 'loss/train': 1.4181785881519318} 01/29/2022 17:12:34 - INFO - codeparrot_training - Step 22004: {'lr': 0.0003146415393614597, 'samples': 4224960, 'steps': 22004, 'loss/train': 3.6781225204467773} 01/29/2022 17:12:38 - INFO - codeparrot_training - Step 22005: {'lr': 0.00031462573319048827, 'samples': 4225152, 'steps': 22005, 'loss/train': 1.7055628895759583} 01/29/2022 17:12:43 - INFO - codeparrot_training - Step 22006: {'lr': 0.0003146099267426806, 'samples': 4225344, 'steps': 22006, 'loss/train': 1.408981740474701} 01/29/2022 17:12:47 - INFO - codeparrot_training - Step 22007: {'lr': 0.00031459412001810474, 'samples': 4225536, 'steps': 22007, 'loss/train': 1.6812592148780823} 01/29/2022 17:12:52 - INFO - codeparrot_training - Step 22008: {'lr': 0.000314578313016828, 'samples': 4225728, 'steps': 22008, 'loss/train': 1.4872789978981018} 01/29/2022 17:12:56 - INFO - codeparrot_training - Step 22009: {'lr': 0.00031456250573891825, 'samples': 4225920, 'steps': 22009, 'loss/train': 1.6449995040893555} 01/29/2022 17:13:00 - INFO - codeparrot_training - Step 22010: {'lr': 0.0003145466981844434, 'samples': 4226112, 'steps': 22010, 'loss/train': 1.891547977924347} 01/29/2022 17:13:07 - INFO - codeparrot_training - Step 22011: {'lr': 0.00031453089035347084, 'samples': 4226304, 'steps': 22011, 'loss/train': 2.0247682929039} 01/29/2022 17:13:12 - INFO - codeparrot_training - Step 22012: {'lr': 0.0003145150822460685, 'samples': 4226496, 'steps': 22012, 'loss/train': 1.6697564721107483} 01/29/2022 17:13:16 - INFO - codeparrot_training - Step 22013: {'lr': 0.00031449927386230397, 'samples': 4226688, 'steps': 22013, 'loss/train': 1.3253786265850067} 01/29/2022 17:13:20 - INFO - codeparrot_training - Step 22014: {'lr': 0.0003144834652022451, 'samples': 4226880, 'steps': 22014, 'loss/train': 1.430372267961502} 01/29/2022 17:13:24 - INFO - codeparrot_training - Step 22015: {'lr': 0.0003144676562659595, 'samples': 4227072, 'steps': 22015, 'loss/train': 2.2930973768234253} 01/29/2022 17:13:30 - INFO - codeparrot_training - Step 22016: {'lr': 0.000314451847053515, 'samples': 4227264, 'steps': 22016, 'loss/train': 1.6549941301345825} 01/29/2022 17:13:34 - INFO - codeparrot_training - Step 22017: {'lr': 0.0003144360375649792, 'samples': 4227456, 'steps': 22017, 'loss/train': 0.5592041909694672} 01/29/2022 17:13:38 - INFO - codeparrot_training - Step 22018: {'lr': 0.0003144202278004199, 'samples': 4227648, 'steps': 22018, 'loss/train': 1.3562363684177399} 01/29/2022 17:13:42 - INFO - codeparrot_training - Step 22019: {'lr': 0.0003144044177599047, 'samples': 4227840, 'steps': 22019, 'loss/train': 0.9630638659000397} 01/29/2022 17:13:47 - INFO - codeparrot_training - Step 22020: {'lr': 0.00031438860744350156, 'samples': 4228032, 'steps': 22020, 'loss/train': 2.045553982257843} 01/29/2022 17:13:52 - INFO - codeparrot_training - Step 22021: {'lr': 0.000314372796851278, 'samples': 4228224, 'steps': 22021, 'loss/train': 1.2689038217067719} 01/29/2022 17:13:56 - INFO - codeparrot_training - Step 22022: {'lr': 0.00031435698598330185, 'samples': 4228416, 'steps': 22022, 'loss/train': 2.3658615946769714} 01/29/2022 17:14:01 - INFO - codeparrot_training - Step 22023: {'lr': 0.0003143411748396408, 'samples': 4228608, 'steps': 22023, 'loss/train': 1.384789913892746} 01/29/2022 17:14:05 - INFO - codeparrot_training - Step 22024: {'lr': 0.00031432536342036255, 'samples': 4228800, 'steps': 22024, 'loss/train': 1.129883050918579} 01/29/2022 17:14:09 - INFO - codeparrot_training - Step 22025: {'lr': 0.00031430955172553497, 'samples': 4228992, 'steps': 22025, 'loss/train': 1.4963155388832092} 01/29/2022 17:14:14 - INFO - codeparrot_training - Step 22026: {'lr': 0.00031429373975522555, 'samples': 4229184, 'steps': 22026, 'loss/train': 1.7065992951393127} 01/29/2022 17:14:18 - INFO - codeparrot_training - Step 22027: {'lr': 0.00031427792750950227, 'samples': 4229376, 'steps': 22027, 'loss/train': 1.350190132856369} 01/29/2022 17:14:23 - INFO - codeparrot_training - Step 22028: {'lr': 0.0003142621149884327, 'samples': 4229568, 'steps': 22028, 'loss/train': 1.9066177010536194} 01/29/2022 17:14:27 - INFO - codeparrot_training - Step 22029: {'lr': 0.00031424630219208474, 'samples': 4229760, 'steps': 22029, 'loss/train': 1.6766796112060547} 01/29/2022 17:14:31 - INFO - codeparrot_training - Step 22030: {'lr': 0.00031423048912052585, 'samples': 4229952, 'steps': 22030, 'loss/train': 1.5325748920440674} 01/29/2022 17:14:38 - INFO - codeparrot_training - Step 22031: {'lr': 0.0003142146757738241, 'samples': 4230144, 'steps': 22031, 'loss/train': 1.5229666829109192} 01/29/2022 17:14:42 - INFO - codeparrot_training - Step 22032: {'lr': 0.000314198862152047, 'samples': 4230336, 'steps': 22032, 'loss/train': 1.5144955515861511} 01/29/2022 17:14:47 - INFO - codeparrot_training - Step 22033: {'lr': 0.00031418304825526236, 'samples': 4230528, 'steps': 22033, 'loss/train': 1.776663064956665} 01/29/2022 17:14:51 - INFO - codeparrot_training - Step 22034: {'lr': 0.0003141672340835379, 'samples': 4230720, 'steps': 22034, 'loss/train': 1.6651784777641296} 01/29/2022 17:14:55 - INFO - codeparrot_training - Step 22035: {'lr': 0.0003141514196369414, 'samples': 4230912, 'steps': 22035, 'loss/train': 1.576603353023529} 01/29/2022 17:15:00 - INFO - codeparrot_training - Step 22036: {'lr': 0.0003141356049155406, 'samples': 4231104, 'steps': 22036, 'loss/train': 1.7657893896102905} 01/29/2022 17:15:05 - INFO - codeparrot_training - Step 22037: {'lr': 0.00031411978991940324, 'samples': 4231296, 'steps': 22037, 'loss/train': 1.5722812414169312} 01/29/2022 17:15:09 - INFO - codeparrot_training - Step 22038: {'lr': 0.000314103974648597, 'samples': 4231488, 'steps': 22038, 'loss/train': 1.6582667827606201} 01/29/2022 17:15:13 - INFO - codeparrot_training - Step 22039: {'lr': 0.0003140881591031898, 'samples': 4231680, 'steps': 22039, 'loss/train': 1.839832842350006} 01/29/2022 17:15:17 - INFO - codeparrot_training - Step 22040: {'lr': 0.0003140723432832492, 'samples': 4231872, 'steps': 22040, 'loss/train': 1.0712084770202637} 01/29/2022 17:15:25 - INFO - codeparrot_training - Step 22041: {'lr': 0.00031405652718884304, 'samples': 4232064, 'steps': 22041, 'loss/train': 1.984972894191742} 01/29/2022 17:15:29 - INFO - codeparrot_training - Step 22042: {'lr': 0.00031404071082003903, 'samples': 4232256, 'steps': 22042, 'loss/train': 2.6697086691856384} 01/29/2022 17:15:33 - INFO - codeparrot_training - Step 22043: {'lr': 0.000314024894176905, 'samples': 4232448, 'steps': 22043, 'loss/train': 0.5624170750379562} 01/29/2022 17:15:38 - INFO - codeparrot_training - Step 22044: {'lr': 0.00031400907725950865, 'samples': 4232640, 'steps': 22044, 'loss/train': 1.0981973111629486} 01/29/2022 17:15:43 - INFO - codeparrot_training - Step 22045: {'lr': 0.00031399326006791765, 'samples': 4232832, 'steps': 22045, 'loss/train': 2.2021772861480713} 01/29/2022 17:15:47 - INFO - codeparrot_training - Step 22046: {'lr': 0.00031397744260219996, 'samples': 4233024, 'steps': 22046, 'loss/train': 0.7804614007472992} 01/29/2022 17:15:52 - INFO - codeparrot_training - Step 22047: {'lr': 0.00031396162486242317, 'samples': 4233216, 'steps': 22047, 'loss/train': 2.0868544578552246} 01/29/2022 17:15:56 - INFO - codeparrot_training - Step 22048: {'lr': 0.0003139458068486551, 'samples': 4233408, 'steps': 22048, 'loss/train': 1.5211027264595032} 01/29/2022 17:16:00 - INFO - codeparrot_training - Step 22049: {'lr': 0.0003139299885609635, 'samples': 4233600, 'steps': 22049, 'loss/train': 1.6194189190864563} 01/29/2022 17:16:04 - INFO - codeparrot_training - Step 22050: {'lr': 0.00031391416999941606, 'samples': 4233792, 'steps': 22050, 'loss/train': 2.1286940574645996} 01/29/2022 17:16:10 - INFO - codeparrot_training - Step 22051: {'lr': 0.00031389835116408073, 'samples': 4233984, 'steps': 22051, 'loss/train': 1.686644732952118} 01/29/2022 17:16:14 - INFO - codeparrot_training - Step 22052: {'lr': 0.000313882532055025, 'samples': 4234176, 'steps': 22052, 'loss/train': 1.4268637597560883} 01/29/2022 17:16:18 - INFO - codeparrot_training - Step 22053: {'lr': 0.000313866712672317, 'samples': 4234368, 'steps': 22053, 'loss/train': 1.9818872809410095} 01/29/2022 17:16:23 - INFO - codeparrot_training - Step 22054: {'lr': 0.0003138508930160241, 'samples': 4234560, 'steps': 22054, 'loss/train': 1.6147372126579285} 01/29/2022 17:16:27 - INFO - codeparrot_training - Step 22055: {'lr': 0.0003138350730862144, 'samples': 4234752, 'steps': 22055, 'loss/train': 0.8677935898303986} 01/29/2022 17:16:34 - INFO - codeparrot_training - Step 22056: {'lr': 0.00031381925288295536, 'samples': 4234944, 'steps': 22056, 'loss/train': 1.8375222086906433} 01/29/2022 17:16:38 - INFO - codeparrot_training - Step 22057: {'lr': 0.000313803432406315, 'samples': 4235136, 'steps': 22057, 'loss/train': 1.5995088815689087} 01/29/2022 17:16:43 - INFO - codeparrot_training - Step 22058: {'lr': 0.000313787611656361, 'samples': 4235328, 'steps': 22058, 'loss/train': 1.803549349308014} 01/29/2022 17:16:47 - INFO - codeparrot_training - Step 22059: {'lr': 0.00031377179063316106, 'samples': 4235520, 'steps': 22059, 'loss/train': 1.7444196939468384} 01/29/2022 17:16:51 - INFO - codeparrot_training - Step 22060: {'lr': 0.0003137559693367831, 'samples': 4235712, 'steps': 22060, 'loss/train': 0.8833642601966858} 01/29/2022 17:16:57 - INFO - codeparrot_training - Step 22061: {'lr': 0.0003137401477672947, 'samples': 4235904, 'steps': 22061, 'loss/train': 1.0041523575782776} 01/29/2022 17:17:01 - INFO - codeparrot_training - Step 22062: {'lr': 0.00031372432592476383, 'samples': 4236096, 'steps': 22062, 'loss/train': 1.14071524143219} 01/29/2022 17:17:05 - INFO - codeparrot_training - Step 22063: {'lr': 0.0003137085038092582, 'samples': 4236288, 'steps': 22063, 'loss/train': 1.8395987749099731} 01/29/2022 17:17:10 - INFO - codeparrot_training - Step 22064: {'lr': 0.00031369268142084555, 'samples': 4236480, 'steps': 22064, 'loss/train': 1.5687727332115173} 01/29/2022 17:17:14 - INFO - codeparrot_training - Step 22065: {'lr': 0.00031367685875959375, 'samples': 4236672, 'steps': 22065, 'loss/train': 1.7396534085273743} 01/29/2022 17:17:19 - INFO - codeparrot_training - Step 22066: {'lr': 0.0003136610358255704, 'samples': 4236864, 'steps': 22066, 'loss/train': 1.9377068281173706} 01/29/2022 17:17:24 - INFO - codeparrot_training - Step 22067: {'lr': 0.00031364521261884336, 'samples': 4237056, 'steps': 22067, 'loss/train': 1.7466498613357544} 01/29/2022 17:17:28 - INFO - codeparrot_training - Step 22068: {'lr': 0.00031362938913948046, 'samples': 4237248, 'steps': 22068, 'loss/train': 2.048322916030884} 01/29/2022 17:17:32 - INFO - codeparrot_training - Step 22069: {'lr': 0.0003136135653875495, 'samples': 4237440, 'steps': 22069, 'loss/train': 1.6285280585289001} 01/29/2022 17:17:36 - INFO - codeparrot_training - Step 22070: {'lr': 0.00031359774136311823, 'samples': 4237632, 'steps': 22070, 'loss/train': 1.958620011806488} 01/29/2022 17:17:43 - INFO - codeparrot_training - Step 22071: {'lr': 0.0003135819170662545, 'samples': 4237824, 'steps': 22071, 'loss/train': 2.1949211955070496} 01/29/2022 17:17:48 - INFO - codeparrot_training - Step 22072: {'lr': 0.00031356609249702587, 'samples': 4238016, 'steps': 22072, 'loss/train': 1.0873979330062866} 01/29/2022 17:17:52 - INFO - codeparrot_training - Step 22073: {'lr': 0.0003135502676555004, 'samples': 4238208, 'steps': 22073, 'loss/train': 1.5199854969978333} 01/29/2022 17:17:56 - INFO - codeparrot_training - Step 22074: {'lr': 0.0003135344425417457, 'samples': 4238400, 'steps': 22074, 'loss/train': 1.9332205653190613} 01/29/2022 17:18:00 - INFO - codeparrot_training - Step 22075: {'lr': 0.0003135186171558297, 'samples': 4238592, 'steps': 22075, 'loss/train': 2.001096189022064} 01/29/2022 17:18:06 - INFO - codeparrot_training - Step 22076: {'lr': 0.00031350279149782004, 'samples': 4238784, 'steps': 22076, 'loss/train': 1.598426878452301} 01/29/2022 17:18:10 - INFO - codeparrot_training - Step 22077: {'lr': 0.0003134869655677846, 'samples': 4238976, 'steps': 22077, 'loss/train': 2.054503619670868} 01/29/2022 17:18:14 - INFO - codeparrot_training - Step 22078: {'lr': 0.00031347113936579116, 'samples': 4239168, 'steps': 22078, 'loss/train': 1.9067557454109192} 01/29/2022 17:18:18 - INFO - codeparrot_training - Step 22079: {'lr': 0.00031345531289190756, 'samples': 4239360, 'steps': 22079, 'loss/train': 1.8355541825294495} 01/29/2022 17:18:23 - INFO - codeparrot_training - Step 22080: {'lr': 0.0003134394861462014, 'samples': 4239552, 'steps': 22080, 'loss/train': 1.8083798289299011} 01/29/2022 17:18:28 - INFO - codeparrot_training - Step 22081: {'lr': 0.0003134236591287407, 'samples': 4239744, 'steps': 22081, 'loss/train': 1.2462304830551147} 01/29/2022 17:18:32 - INFO - codeparrot_training - Step 22082: {'lr': 0.0003134078318395933, 'samples': 4239936, 'steps': 22082, 'loss/train': 2.2743841409683228} 01/29/2022 17:18:36 - INFO - codeparrot_training - Step 22083: {'lr': 0.00031339200427882676, 'samples': 4240128, 'steps': 22083, 'loss/train': 1.7213464379310608} 01/29/2022 17:18:41 - INFO - codeparrot_training - Step 22084: {'lr': 0.00031337617644650907, 'samples': 4240320, 'steps': 22084, 'loss/train': 2.180558681488037} 01/29/2022 17:18:45 - INFO - codeparrot_training - Step 22085: {'lr': 0.00031336034834270786, 'samples': 4240512, 'steps': 22085, 'loss/train': 1.6368547081947327} 01/29/2022 17:18:52 - INFO - codeparrot_training - Step 22086: {'lr': 0.00031334451996749117, 'samples': 4240704, 'steps': 22086, 'loss/train': 2.209353983402252} 01/29/2022 17:18:56 - INFO - codeparrot_training - Step 22087: {'lr': 0.00031332869132092654, 'samples': 4240896, 'steps': 22087, 'loss/train': 1.6674118638038635} 01/29/2022 17:19:00 - INFO - codeparrot_training - Step 22088: {'lr': 0.00031331286240308205, 'samples': 4241088, 'steps': 22088, 'loss/train': 2.7394291162490845} 01/29/2022 17:19:05 - INFO - codeparrot_training - Step 22089: {'lr': 0.00031329703321402526, 'samples': 4241280, 'steps': 22089, 'loss/train': 2.1045597195625305} 01/29/2022 17:19:09 - INFO - codeparrot_training - Step 22090: {'lr': 0.00031328120375382414, 'samples': 4241472, 'steps': 22090, 'loss/train': 2.014697313308716} 01/29/2022 17:19:14 - INFO - codeparrot_training - Step 22091: {'lr': 0.0003132653740225464, 'samples': 4241664, 'steps': 22091, 'loss/train': 1.3775839805603027} 01/29/2022 17:19:19 - INFO - codeparrot_training - Step 22092: {'lr': 0.0003132495440202599, 'samples': 4241856, 'steps': 22092, 'loss/train': 1.9096525311470032} 01/29/2022 17:19:23 - INFO - codeparrot_training - Step 22093: {'lr': 0.0003132337137470324, 'samples': 4242048, 'steps': 22093, 'loss/train': 1.6905410885810852} 01/29/2022 17:19:27 - INFO - codeparrot_training - Step 22094: {'lr': 0.00031321788320293176, 'samples': 4242240, 'steps': 22094, 'loss/train': 1.1693287789821625} 01/29/2022 17:19:32 - INFO - codeparrot_training - Step 22095: {'lr': 0.00031320205238802583, 'samples': 4242432, 'steps': 22095, 'loss/train': 1.082742154598236} 01/29/2022 17:19:37 - INFO - codeparrot_training - Step 22096: {'lr': 0.0003131862213023823, 'samples': 4242624, 'steps': 22096, 'loss/train': 2.118521511554718} 01/29/2022 17:19:41 - INFO - codeparrot_training - Step 22097: {'lr': 0.0003131703899460692, 'samples': 4242816, 'steps': 22097, 'loss/train': 2.0646414756774902} 01/29/2022 17:19:46 - INFO - codeparrot_training - Step 22098: {'lr': 0.0003131545583191541, 'samples': 4243008, 'steps': 22098, 'loss/train': 6.935418605804443} 01/29/2022 17:19:50 - INFO - codeparrot_training - Step 22099: {'lr': 0.00031313872642170493, 'samples': 4243200, 'steps': 22099, 'loss/train': 2.029514729976654} 01/29/2022 17:19:54 - INFO - codeparrot_training - Step 22100: {'lr': 0.0003131228942537895, 'samples': 4243392, 'steps': 22100, 'loss/train': 0.9553154110908508} 01/29/2022 17:19:59 - INFO - codeparrot_training - Step 22101: {'lr': 0.00031310706181547567, 'samples': 4243584, 'steps': 22101, 'loss/train': 1.1325762569904327} 01/29/2022 17:20:06 - INFO - codeparrot_training - Step 22102: {'lr': 0.0003130912291068312, 'samples': 4243776, 'steps': 22102, 'loss/train': 1.2766943871974945} 01/29/2022 17:20:10 - INFO - codeparrot_training - Step 22103: {'lr': 0.000313075396127924, 'samples': 4243968, 'steps': 22103, 'loss/train': 2.0860885977745056} 01/29/2022 17:20:14 - INFO - codeparrot_training - Step 22104: {'lr': 0.0003130595628788217, 'samples': 4244160, 'steps': 22104, 'loss/train': 1.6747584342956543} 01/29/2022 17:20:19 - INFO - codeparrot_training - Step 22105: {'lr': 0.00031304372935959235, 'samples': 4244352, 'steps': 22105, 'loss/train': 1.2529475390911102} 01/29/2022 17:20:23 - INFO - codeparrot_training - Step 22106: {'lr': 0.00031302789557030364, 'samples': 4244544, 'steps': 22106, 'loss/train': 0.4839460104703903} 01/29/2022 17:20:28 - INFO - codeparrot_training - Step 22107: {'lr': 0.00031301206151102353, 'samples': 4244736, 'steps': 22107, 'loss/train': 1.918887734413147} 01/29/2022 17:20:32 - INFO - codeparrot_training - Step 22108: {'lr': 0.00031299622718181964, 'samples': 4244928, 'steps': 22108, 'loss/train': 2.3766363859176636} 01/29/2022 17:20:37 - INFO - codeparrot_training - Step 22109: {'lr': 0.00031298039258276, 'samples': 4245120, 'steps': 22109, 'loss/train': 1.6896631121635437} 01/29/2022 17:20:41 - INFO - codeparrot_training - Step 22110: {'lr': 0.0003129645577139123, 'samples': 4245312, 'steps': 22110, 'loss/train': 2.2402930855751038} 01/29/2022 17:20:45 - INFO - codeparrot_training - Step 22111: {'lr': 0.00031294872257534446, 'samples': 4245504, 'steps': 22111, 'loss/train': 0.9991984069347382} 01/29/2022 17:20:50 - INFO - codeparrot_training - Step 22112: {'lr': 0.0003129328871671243, 'samples': 4245696, 'steps': 22112, 'loss/train': 1.2491111755371094} 01/29/2022 17:20:55 - INFO - codeparrot_training - Step 22113: {'lr': 0.0003129170514893196, 'samples': 4245888, 'steps': 22113, 'loss/train': 1.5176453590393066} 01/29/2022 17:20:59 - INFO - codeparrot_training - Step 22114: {'lr': 0.0003129012155419983, 'samples': 4246080, 'steps': 22114, 'loss/train': 1.8644778728485107} 01/29/2022 17:21:03 - INFO - codeparrot_training - Step 22115: {'lr': 0.00031288537932522807, 'samples': 4246272, 'steps': 22115, 'loss/train': 1.7288707494735718} 01/29/2022 17:21:10 - INFO - codeparrot_training - Step 22116: {'lr': 0.0003128695428390769, 'samples': 4246464, 'steps': 22116, 'loss/train': 1.2865969240665436} 01/29/2022 17:21:15 - INFO - codeparrot_training - Step 22117: {'lr': 0.0003128537060836125, 'samples': 4246656, 'steps': 22117, 'loss/train': 1.9253059029579163} 01/29/2022 17:21:19 - INFO - codeparrot_training - Step 22118: {'lr': 0.0003128378690589028, 'samples': 4246848, 'steps': 22118, 'loss/train': 1.8014668822288513} 01/29/2022 17:21:23 - INFO - codeparrot_training - Step 22119: {'lr': 0.00031282203176501573, 'samples': 4247040, 'steps': 22119, 'loss/train': 1.5278865694999695} 01/29/2022 17:21:27 - INFO - codeparrot_training - Step 22120: {'lr': 0.0003128061942020189, 'samples': 4247232, 'steps': 22120, 'loss/train': 1.342042475938797} 01/29/2022 17:21:33 - INFO - codeparrot_training - Step 22121: {'lr': 0.00031279035636998037, 'samples': 4247424, 'steps': 22121, 'loss/train': 1.453108012676239} 01/29/2022 17:21:37 - INFO - codeparrot_training - Step 22122: {'lr': 0.0003127745182689678, 'samples': 4247616, 'steps': 22122, 'loss/train': 1.8244303464889526} 01/29/2022 17:21:41 - INFO - codeparrot_training - Step 22123: {'lr': 0.00031275867989904923, 'samples': 4247808, 'steps': 22123, 'loss/train': 1.6131058931350708} 01/29/2022 17:21:45 - INFO - codeparrot_training - Step 22124: {'lr': 0.0003127428412602923, 'samples': 4248000, 'steps': 22124, 'loss/train': 1.2411067485809326} 01/29/2022 17:21:50 - INFO - codeparrot_training - Step 22125: {'lr': 0.00031272700235276507, 'samples': 4248192, 'steps': 22125, 'loss/train': 2.2602809071540833} 01/29/2022 17:21:57 - INFO - codeparrot_training - Step 22126: {'lr': 0.00031271116317653513, 'samples': 4248384, 'steps': 22126, 'loss/train': 1.3730115294456482} 01/29/2022 17:22:01 - INFO - codeparrot_training - Step 22127: {'lr': 0.00031269532373167063, 'samples': 4248576, 'steps': 22127, 'loss/train': 1.6034016609191895} 01/29/2022 17:22:05 - INFO - codeparrot_training - Step 22128: {'lr': 0.0003126794840182392, 'samples': 4248768, 'steps': 22128, 'loss/train': 2.8368515968322754} 01/29/2022 17:22:09 - INFO - codeparrot_training - Step 22129: {'lr': 0.00031266364403630874, 'samples': 4248960, 'steps': 22129, 'loss/train': 1.6768561005592346} 01/29/2022 17:22:14 - INFO - codeparrot_training - Step 22130: {'lr': 0.0003126478037859472, 'samples': 4249152, 'steps': 22130, 'loss/train': 1.7565520405769348} 01/29/2022 17:22:19 - INFO - codeparrot_training - Step 22131: {'lr': 0.0003126319632672223, 'samples': 4249344, 'steps': 22131, 'loss/train': 1.9176501631736755} 01/29/2022 17:22:23 - INFO - codeparrot_training - Step 22132: {'lr': 0.00031261612248020197, 'samples': 4249536, 'steps': 22132, 'loss/train': 1.6004834175109863} 01/29/2022 17:22:28 - INFO - codeparrot_training - Step 22133: {'lr': 0.00031260028142495404, 'samples': 4249728, 'steps': 22133, 'loss/train': 1.6227040886878967} 01/29/2022 17:22:32 - INFO - codeparrot_training - Step 22134: {'lr': 0.0003125844401015465, 'samples': 4249920, 'steps': 22134, 'loss/train': 1.4090770483016968} 01/29/2022 17:22:36 - INFO - codeparrot_training - Step 22135: {'lr': 0.0003125685985100469, 'samples': 4250112, 'steps': 22135, 'loss/train': 2.5610931515693665} 01/29/2022 17:22:42 - INFO - codeparrot_training - Step 22136: {'lr': 0.0003125527566505234, 'samples': 4250304, 'steps': 22136, 'loss/train': 3.517988920211792} 01/29/2022 17:22:46 - INFO - codeparrot_training - Step 22137: {'lr': 0.0003125369145230438, 'samples': 4250496, 'steps': 22137, 'loss/train': 2.118595004081726} 01/29/2022 17:22:50 - INFO - codeparrot_training - Step 22138: {'lr': 0.0003125210721276758, 'samples': 4250688, 'steps': 22138, 'loss/train': 2.0677923560142517} 01/29/2022 17:22:54 - INFO - codeparrot_training - Step 22139: {'lr': 0.0003125052294644874, 'samples': 4250880, 'steps': 22139, 'loss/train': 1.4147033393383026} 01/29/2022 17:22:59 - INFO - codeparrot_training - Step 22140: {'lr': 0.0003124893865335466, 'samples': 4251072, 'steps': 22140, 'loss/train': 2.0103389024734497} 01/29/2022 17:23:06 - INFO - codeparrot_training - Step 22141: {'lr': 0.00031247354333492096, 'samples': 4251264, 'steps': 22141, 'loss/train': 1.762890636920929} 01/29/2022 17:23:10 - INFO - codeparrot_training - Step 22142: {'lr': 0.00031245769986867845, 'samples': 4251456, 'steps': 22142, 'loss/train': 1.938239336013794} 01/29/2022 17:23:14 - INFO - codeparrot_training - Step 22143: {'lr': 0.0003124418561348871, 'samples': 4251648, 'steps': 22143, 'loss/train': 1.3896063566207886} 01/29/2022 17:23:18 - INFO - codeparrot_training - Step 22144: {'lr': 0.0003124260121336146, 'samples': 4251840, 'steps': 22144, 'loss/train': 1.7648577690124512} 01/29/2022 17:23:23 - INFO - codeparrot_training - Step 22145: {'lr': 0.000312410167864929, 'samples': 4252032, 'steps': 22145, 'loss/train': 0.7795414924621582} 01/29/2022 17:23:28 - INFO - codeparrot_training - Step 22146: {'lr': 0.00031239432332889796, 'samples': 4252224, 'steps': 22146, 'loss/train': 1.6305744051933289} 01/29/2022 17:23:32 - INFO - codeparrot_training - Step 22147: {'lr': 0.00031237847852558947, 'samples': 4252416, 'steps': 22147, 'loss/train': 1.5376426577568054} 01/29/2022 17:23:36 - INFO - codeparrot_training - Step 22148: {'lr': 0.00031236263345507133, 'samples': 4252608, 'steps': 22148, 'loss/train': 2.0374646186828613} 01/29/2022 17:23:41 - INFO - codeparrot_training - Step 22149: {'lr': 0.0003123467881174116, 'samples': 4252800, 'steps': 22149, 'loss/train': 1.948259711265564} 01/29/2022 17:23:45 - INFO - codeparrot_training - Step 22150: {'lr': 0.0003123309425126779, 'samples': 4252992, 'steps': 22150, 'loss/train': 1.897462785243988} 01/29/2022 17:23:50 - INFO - codeparrot_training - Step 22151: {'lr': 0.00031231509664093833, 'samples': 4253184, 'steps': 22151, 'loss/train': 1.369369626045227} 01/29/2022 17:23:55 - INFO - codeparrot_training - Step 22152: {'lr': 0.00031229925050226063, 'samples': 4253376, 'steps': 22152, 'loss/train': 1.6196731925010681} 01/29/2022 17:23:59 - INFO - codeparrot_training - Step 22153: {'lr': 0.0003122834040967127, 'samples': 4253568, 'steps': 22153, 'loss/train': 1.067340463399887} 01/29/2022 17:24:03 - INFO - codeparrot_training - Step 22154: {'lr': 0.00031226755742436255, 'samples': 4253760, 'steps': 22154, 'loss/train': 1.2313039898872375} 01/29/2022 17:24:07 - INFO - codeparrot_training - Step 22155: {'lr': 0.0003122517104852778, 'samples': 4253952, 'steps': 22155, 'loss/train': 2.15651535987854} 01/29/2022 17:24:13 - INFO - codeparrot_training - Step 22156: {'lr': 0.0003122358632795266, 'samples': 4254144, 'steps': 22156, 'loss/train': 1.2799761593341827} 01/29/2022 17:24:17 - INFO - codeparrot_training - Step 22157: {'lr': 0.00031222001580717663, 'samples': 4254336, 'steps': 22157, 'loss/train': 0.9487791359424591} 01/29/2022 17:24:22 - INFO - codeparrot_training - Step 22158: {'lr': 0.000312204168068296, 'samples': 4254528, 'steps': 22158, 'loss/train': 2.2366417050361633} 01/29/2022 17:24:26 - INFO - codeparrot_training - Step 22159: {'lr': 0.00031218832006295235, 'samples': 4254720, 'steps': 22159, 'loss/train': 0.8271512389183044} 01/29/2022 17:24:30 - INFO - codeparrot_training - Step 22160: {'lr': 0.0003121724717912138, 'samples': 4254912, 'steps': 22160, 'loss/train': 1.8140308856964111} 01/29/2022 17:24:34 - INFO - codeparrot_training - Step 22161: {'lr': 0.000312156623253148, 'samples': 4255104, 'steps': 22161, 'loss/train': 1.1995005011558533} 01/29/2022 17:24:41 - INFO - codeparrot_training - Step 22162: {'lr': 0.00031214077444882297, 'samples': 4255296, 'steps': 22162, 'loss/train': 2.1305333375930786} 01/29/2022 17:24:46 - INFO - codeparrot_training - Step 22163: {'lr': 0.0003121249253783067, 'samples': 4255488, 'steps': 22163, 'loss/train': 1.5764726400375366} 01/29/2022 17:24:50 - INFO - codeparrot_training - Step 22164: {'lr': 0.00031210907604166686, 'samples': 4255680, 'steps': 22164, 'loss/train': 2.166677176952362} 01/29/2022 17:24:54 - INFO - codeparrot_training - Step 22165: {'lr': 0.0003120932264389715, 'samples': 4255872, 'steps': 22165, 'loss/train': 2.3122923374176025} 01/29/2022 17:24:58 - INFO - codeparrot_training - Step 22166: {'lr': 0.0003120773765702885, 'samples': 4256064, 'steps': 22166, 'loss/train': 2.731784999370575} 01/29/2022 17:25:04 - INFO - codeparrot_training - Step 22167: {'lr': 0.00031206152643568577, 'samples': 4256256, 'steps': 22167, 'loss/train': 1.627392053604126} 01/29/2022 17:25:08 - INFO - codeparrot_training - Step 22168: {'lr': 0.00031204567603523105, 'samples': 4256448, 'steps': 22168, 'loss/train': 1.4719775319099426} 01/29/2022 17:25:12 - INFO - codeparrot_training - Step 22169: {'lr': 0.00031202982536899246, 'samples': 4256640, 'steps': 22169, 'loss/train': 3.0744305849075317} 01/29/2022 17:25:17 - INFO - codeparrot_training - Step 22170: {'lr': 0.0003120139744370377, 'samples': 4256832, 'steps': 22170, 'loss/train': 1.5315842628479004} 01/29/2022 17:25:21 - INFO - codeparrot_training - Step 22171: {'lr': 0.0003119981232394349, 'samples': 4257024, 'steps': 22171, 'loss/train': 2.0551568269729614} 01/29/2022 17:25:28 - INFO - codeparrot_training - Step 22172: {'lr': 0.0003119822717762517, 'samples': 4257216, 'steps': 22172, 'loss/train': 1.7782668471336365} 01/29/2022 17:25:32 - INFO - codeparrot_training - Step 22173: {'lr': 0.0003119664200475562, 'samples': 4257408, 'steps': 22173, 'loss/train': 0.7942833602428436} 01/29/2022 17:25:37 - INFO - codeparrot_training - Step 22174: {'lr': 0.0003119505680534162, 'samples': 4257600, 'steps': 22174, 'loss/train': 2.2435750365257263} 01/29/2022 17:25:41 - INFO - codeparrot_training - Step 22175: {'lr': 0.00031193471579389967, 'samples': 4257792, 'steps': 22175, 'loss/train': 1.3651418387889862} 01/29/2022 17:25:45 - INFO - codeparrot_training - Step 22176: {'lr': 0.00031191886326907445, 'samples': 4257984, 'steps': 22176, 'loss/train': 1.7808401584625244} 01/29/2022 17:25:51 - INFO - codeparrot_training - Step 22177: {'lr': 0.0003119030104790085, 'samples': 4258176, 'steps': 22177, 'loss/train': 2.182638645172119} 01/29/2022 17:25:55 - INFO - codeparrot_training - Step 22178: {'lr': 0.00031188715742376966, 'samples': 4258368, 'steps': 22178, 'loss/train': 1.304250329732895} 01/29/2022 17:25:59 - INFO - codeparrot_training - Step 22179: {'lr': 0.0003118713041034259, 'samples': 4258560, 'steps': 22179, 'loss/train': 1.2365865111351013} 01/29/2022 17:26:03 - INFO - codeparrot_training - Step 22180: {'lr': 0.0003118554505180452, 'samples': 4258752, 'steps': 22180, 'loss/train': 0.9106057584285736} 01/29/2022 17:26:08 - INFO - codeparrot_training - Step 22181: {'lr': 0.0003118395966676953, 'samples': 4258944, 'steps': 22181, 'loss/train': 1.1767686009407043} 01/29/2022 17:26:12 - INFO - codeparrot_training - Step 22182: {'lr': 0.00031182374255244426, 'samples': 4259136, 'steps': 22182, 'loss/train': 2.4831928610801697} 01/29/2022 17:26:17 - INFO - codeparrot_training - Step 22183: {'lr': 0.0003118078881723599, 'samples': 4259328, 'steps': 22183, 'loss/train': 2.1979567408561707} 01/29/2022 17:26:21 - INFO - codeparrot_training - Step 22184: {'lr': 0.0003117920335275102, 'samples': 4259520, 'steps': 22184, 'loss/train': 1.8036679029464722} 01/29/2022 17:26:26 - INFO - codeparrot_training - Step 22185: {'lr': 0.000311776178617963, 'samples': 4259712, 'steps': 22185, 'loss/train': 1.111944168806076} 01/29/2022 17:26:30 - INFO - codeparrot_training - Step 22186: {'lr': 0.0003117603234437864, 'samples': 4259904, 'steps': 22186, 'loss/train': 1.6141878962516785} 01/29/2022 17:26:34 - INFO - codeparrot_training - Step 22187: {'lr': 0.0003117444680050481, 'samples': 4260096, 'steps': 22187, 'loss/train': 2.6029114723205566} 01/29/2022 17:26:41 - INFO - codeparrot_training - Step 22188: {'lr': 0.00031172861230181607, 'samples': 4260288, 'steps': 22188, 'loss/train': 1.4934545159339905} 01/29/2022 17:26:46 - INFO - codeparrot_training - Step 22189: {'lr': 0.0003117127563341583, 'samples': 4260480, 'steps': 22189, 'loss/train': 0.6954122185707092} 01/29/2022 17:26:50 - INFO - codeparrot_training - Step 22190: {'lr': 0.0003116969001021427, 'samples': 4260672, 'steps': 22190, 'loss/train': 1.2872409224510193} 01/29/2022 17:26:54 - INFO - codeparrot_training - Step 22191: {'lr': 0.00031168104360583716, 'samples': 4260864, 'steps': 22191, 'loss/train': 1.9373497366905212} 01/29/2022 17:26:59 - INFO - codeparrot_training - Step 22192: {'lr': 0.00031166518684530966, 'samples': 4261056, 'steps': 22192, 'loss/train': 1.4638495445251465} 01/29/2022 17:27:03 - INFO - codeparrot_training - Step 22193: {'lr': 0.000311649329820628, 'samples': 4261248, 'steps': 22193, 'loss/train': 1.9079418182373047} 01/29/2022 17:27:07 - INFO - codeparrot_training - Step 22194: {'lr': 0.0003116334725318602, 'samples': 4261440, 'steps': 22194, 'loss/train': 1.4627099633216858} 01/29/2022 17:27:13 - INFO - codeparrot_training - Step 22195: {'lr': 0.00031161761497907416, 'samples': 4261632, 'steps': 22195, 'loss/train': 2.284150242805481} 01/29/2022 17:27:17 - INFO - codeparrot_training - Step 22196: {'lr': 0.00031160175716233793, 'samples': 4261824, 'steps': 22196, 'loss/train': 1.8470045328140259} 01/29/2022 17:27:21 - INFO - codeparrot_training - Step 22197: {'lr': 0.0003115858990817192, 'samples': 4262016, 'steps': 22197, 'loss/train': 1.015902578830719} 01/29/2022 17:27:25 - INFO - codeparrot_training - Step 22198: {'lr': 0.0003115700407372861, 'samples': 4262208, 'steps': 22198, 'loss/train': 1.3035476803779602} 01/29/2022 17:27:29 - INFO - codeparrot_training - Step 22199: {'lr': 0.00031155418212910647, 'samples': 4262400, 'steps': 22199, 'loss/train': 1.7406651377677917} 01/29/2022 17:27:37 - INFO - codeparrot_training - Step 22200: {'lr': 0.00031153832325724826, 'samples': 4262592, 'steps': 22200, 'loss/train': 1.3886753618717194} 01/29/2022 17:27:41 - INFO - codeparrot_training - Step 22201: {'lr': 0.0003115224641217795, 'samples': 4262784, 'steps': 22201, 'loss/train': 1.6047552824020386} 01/29/2022 17:27:45 - INFO - codeparrot_training - Step 22202: {'lr': 0.0003115066047227679, 'samples': 4262976, 'steps': 22202, 'loss/train': 2.1328791975975037} 01/29/2022 17:27:49 - INFO - codeparrot_training - Step 22203: {'lr': 0.00031149074506028163, 'samples': 4263168, 'steps': 22203, 'loss/train': 2.022697627544403} 01/29/2022 17:27:54 - INFO - codeparrot_training - Step 22204: {'lr': 0.00031147488513438853, 'samples': 4263360, 'steps': 22204, 'loss/train': 1.7988225817680359} 01/29/2022 17:27:59 - INFO - codeparrot_training - Step 22205: {'lr': 0.00031145902494515655, 'samples': 4263552, 'steps': 22205, 'loss/train': 2.376934289932251} 01/29/2022 17:28:03 - INFO - codeparrot_training - Step 22206: {'lr': 0.0003114431644926536, 'samples': 4263744, 'steps': 22206, 'loss/train': 1.2991673648357391} 01/29/2022 17:28:07 - INFO - codeparrot_training - Step 22207: {'lr': 0.00031142730377694763, 'samples': 4263936, 'steps': 22207, 'loss/train': 1.0995968878269196} 01/29/2022 17:28:12 - INFO - codeparrot_training - Step 22208: {'lr': 0.0003114114427981066, 'samples': 4264128, 'steps': 22208, 'loss/train': 1.23307666182518} 01/29/2022 17:28:16 - INFO - codeparrot_training - Step 22209: {'lr': 0.00031139558155619844, 'samples': 4264320, 'steps': 22209, 'loss/train': 2.1707069277763367} 01/29/2022 17:28:21 - INFO - codeparrot_training - Step 22210: {'lr': 0.0003113797200512912, 'samples': 4264512, 'steps': 22210, 'loss/train': 2.2611259818077087} 01/29/2022 17:28:25 - INFO - codeparrot_training - Step 22211: {'lr': 0.0003113638582834526, 'samples': 4264704, 'steps': 22211, 'loss/train': 2.2152344584465027} 01/29/2022 17:28:30 - INFO - codeparrot_training - Step 22212: {'lr': 0.00031134799625275077, 'samples': 4264896, 'steps': 22212, 'loss/train': 1.2821714580059052} 01/29/2022 17:28:34 - INFO - codeparrot_training - Step 22213: {'lr': 0.0003113321339592536, 'samples': 4265088, 'steps': 22213, 'loss/train': 2.1260620951652527} 01/29/2022 17:28:38 - INFO - codeparrot_training - Step 22214: {'lr': 0.000311316271403029, 'samples': 4265280, 'steps': 22214, 'loss/train': 1.4978612065315247} 01/29/2022 17:28:44 - INFO - codeparrot_training - Step 22215: {'lr': 0.00031130040858414506, 'samples': 4265472, 'steps': 22215, 'loss/train': 2.284808099269867} 01/29/2022 17:28:48 - INFO - codeparrot_training - Step 22216: {'lr': 0.00031128454550266956, 'samples': 4265664, 'steps': 22216, 'loss/train': 0.8572927415370941} 01/29/2022 17:28:52 - INFO - codeparrot_training - Step 22217: {'lr': 0.0003112686821586706, 'samples': 4265856, 'steps': 22217, 'loss/train': 1.430612862110138} 01/29/2022 17:28:56 - INFO - codeparrot_training - Step 22218: {'lr': 0.000311252818552216, 'samples': 4266048, 'steps': 22218, 'loss/train': 1.5421258807182312} 01/29/2022 17:29:01 - INFO - codeparrot_training - Step 22219: {'lr': 0.00031123695468337375, 'samples': 4266240, 'steps': 22219, 'loss/train': 1.5326276421546936} 01/29/2022 17:29:08 - INFO - codeparrot_training - Step 22220: {'lr': 0.00031122109055221187, 'samples': 4266432, 'steps': 22220, 'loss/train': 1.7968772649765015} 01/29/2022 17:29:12 - INFO - codeparrot_training - Step 22221: {'lr': 0.00031120522615879834, 'samples': 4266624, 'steps': 22221, 'loss/train': 1.044331133365631} 01/29/2022 17:29:17 - INFO - codeparrot_training - Step 22222: {'lr': 0.00031118936150320093, 'samples': 4266816, 'steps': 22222, 'loss/train': 1.5359503626823425} 01/29/2022 17:29:21 - INFO - codeparrot_training - Step 22223: {'lr': 0.00031117349658548783, 'samples': 4267008, 'steps': 22223, 'loss/train': 1.8596057295799255} 01/29/2022 17:29:25 - INFO - codeparrot_training - Step 22224: {'lr': 0.00031115763140572686, 'samples': 4267200, 'steps': 22224, 'loss/train': 1.3208650946617126} 01/29/2022 17:29:30 - INFO - codeparrot_training - Step 22225: {'lr': 0.000311141765963986, 'samples': 4267392, 'steps': 22225, 'loss/train': 1.2995184659957886} 01/29/2022 17:29:35 - INFO - codeparrot_training - Step 22226: {'lr': 0.00031112590026033323, 'samples': 4267584, 'steps': 22226, 'loss/train': 1.6912626028060913} 01/29/2022 17:29:39 - INFO - codeparrot_training - Step 22227: {'lr': 0.00031111003429483647, 'samples': 4267776, 'steps': 22227, 'loss/train': 1.9819514751434326} 01/29/2022 17:29:43 - INFO - codeparrot_training - Step 22228: {'lr': 0.00031109416806756387, 'samples': 4267968, 'steps': 22228, 'loss/train': 1.4104354083538055} 01/29/2022 17:29:47 - INFO - codeparrot_training - Step 22229: {'lr': 0.0003110783015785831, 'samples': 4268160, 'steps': 22229, 'loss/train': 2.4265642762184143} 01/29/2022 17:29:53 - INFO - codeparrot_training - Step 22230: {'lr': 0.00031106243482796234, 'samples': 4268352, 'steps': 22230, 'loss/train': 1.1850082576274872} 01/29/2022 17:29:57 - INFO - codeparrot_training - Step 22231: {'lr': 0.0003110465678157695, 'samples': 4268544, 'steps': 22231, 'loss/train': 1.9728258848190308} 01/29/2022 17:30:01 - INFO - codeparrot_training - Step 22232: {'lr': 0.0003110307005420726, 'samples': 4268736, 'steps': 22232, 'loss/train': 1.3842046558856964} 01/29/2022 17:30:05 - INFO - codeparrot_training - Step 22233: {'lr': 0.00031101483300693944, 'samples': 4268928, 'steps': 22233, 'loss/train': 1.889630377292633} 01/29/2022 17:30:10 - INFO - codeparrot_training - Step 22234: {'lr': 0.00031099896521043826, 'samples': 4269120, 'steps': 22234, 'loss/train': 1.580434262752533} 01/29/2022 17:30:17 - INFO - codeparrot_training - Step 22235: {'lr': 0.0003109830971526367, 'samples': 4269312, 'steps': 22235, 'loss/train': 2.0423081517219543} 01/29/2022 17:30:21 - INFO - codeparrot_training - Step 22236: {'lr': 0.0003109672288336031, 'samples': 4269504, 'steps': 22236, 'loss/train': 0.8512862026691437} 01/29/2022 17:30:25 - INFO - codeparrot_training - Step 22237: {'lr': 0.00031095136025340514, 'samples': 4269696, 'steps': 22237, 'loss/train': 1.839228093624115} 01/29/2022 17:30:29 - INFO - codeparrot_training - Step 22238: {'lr': 0.00031093549141211096, 'samples': 4269888, 'steps': 22238, 'loss/train': 0.4697941839694977} 01/29/2022 17:30:34 - INFO - codeparrot_training - Step 22239: {'lr': 0.00031091962230978844, 'samples': 4270080, 'steps': 22239, 'loss/train': 1.7374649047851562} 01/29/2022 17:30:39 - INFO - codeparrot_training - Step 22240: {'lr': 0.0003109037529465056, 'samples': 4270272, 'steps': 22240, 'loss/train': 0.8268062174320221} 01/29/2022 17:30:43 - INFO - codeparrot_training - Step 22241: {'lr': 0.0003108878833223305, 'samples': 4270464, 'steps': 22241, 'loss/train': 1.6526865363121033} 01/29/2022 17:30:48 - INFO - codeparrot_training - Step 22242: {'lr': 0.00031087201343733096, 'samples': 4270656, 'steps': 22242, 'loss/train': 1.9559047222137451} 01/29/2022 17:30:52 - INFO - codeparrot_training - Step 22243: {'lr': 0.00031085614329157515, 'samples': 4270848, 'steps': 22243, 'loss/train': 1.6648249626159668} 01/29/2022 17:30:56 - INFO - codeparrot_training - Step 22244: {'lr': 0.00031084027288513083, 'samples': 4271040, 'steps': 22244, 'loss/train': 1.9838032722473145} 01/29/2022 17:31:03 - INFO - codeparrot_training - Step 22245: {'lr': 0.0003108244022180661, 'samples': 4271232, 'steps': 22245, 'loss/train': 1.7340908646583557} 01/29/2022 17:31:07 - INFO - codeparrot_training - Step 22246: {'lr': 0.000310808531290449, 'samples': 4271424, 'steps': 22246, 'loss/train': 2.201514780521393} 01/29/2022 17:31:12 - INFO - codeparrot_training - Step 22247: {'lr': 0.00031079266010234746, 'samples': 4271616, 'steps': 22247, 'loss/train': 0.9347780048847198} 01/29/2022 17:31:16 - INFO - codeparrot_training - Step 22248: {'lr': 0.00031077678865382944, 'samples': 4271808, 'steps': 22248, 'loss/train': 1.3156985342502594} 01/29/2022 17:31:20 - INFO - codeparrot_training - Step 22249: {'lr': 0.000310760916944963, 'samples': 4272000, 'steps': 22249, 'loss/train': 2.321380913257599} 01/29/2022 17:31:25 - INFO - codeparrot_training - Step 22250: {'lr': 0.000310745044975816, 'samples': 4272192, 'steps': 22250, 'loss/train': 0.6835193485021591} 01/29/2022 17:31:30 - INFO - codeparrot_training - Step 22251: {'lr': 0.00031072917274645656, 'samples': 4272384, 'steps': 22251, 'loss/train': 1.3493591845035553} 01/29/2022 17:31:34 - INFO - codeparrot_training - Step 22252: {'lr': 0.00031071330025695266, 'samples': 4272576, 'steps': 22252, 'loss/train': 1.6082607507705688} 01/29/2022 17:31:38 - INFO - codeparrot_training - Step 22253: {'lr': 0.0003106974275073722, 'samples': 4272768, 'steps': 22253, 'loss/train': 1.1315776705741882} 01/29/2022 17:31:42 - INFO - codeparrot_training - Step 22254: {'lr': 0.0003106815544977833, 'samples': 4272960, 'steps': 22254, 'loss/train': 0.9941892921924591} 01/29/2022 17:31:48 - INFO - codeparrot_training - Step 22255: {'lr': 0.00031066568122825383, 'samples': 4273152, 'steps': 22255, 'loss/train': 0.7491371333599091} 01/29/2022 17:31:52 - INFO - codeparrot_training - Step 22256: {'lr': 0.0003106498076988519, 'samples': 4273344, 'steps': 22256, 'loss/train': 1.4598349034786224} 01/29/2022 17:31:56 - INFO - codeparrot_training - Step 22257: {'lr': 0.0003106339339096454, 'samples': 4273536, 'steps': 22257, 'loss/train': 1.4117978811264038} 01/29/2022 17:32:00 - INFO - codeparrot_training - Step 22258: {'lr': 0.0003106180598607024, 'samples': 4273728, 'steps': 22258, 'loss/train': 1.5366851091384888} 01/29/2022 17:32:05 - INFO - codeparrot_training - Step 22259: {'lr': 0.00031060218555209094, 'samples': 4273920, 'steps': 22259, 'loss/train': 1.5645892024040222} 01/29/2022 17:32:10 - INFO - codeparrot_training - Step 22260: {'lr': 0.0003105863109838789, 'samples': 4274112, 'steps': 22260, 'loss/train': 1.5764747858047485} 01/29/2022 17:32:14 - INFO - codeparrot_training - Step 22261: {'lr': 0.0003105704361561343, 'samples': 4274304, 'steps': 22261, 'loss/train': 0.8120325207710266} 01/29/2022 17:32:19 - INFO - codeparrot_training - Step 22262: {'lr': 0.00031055456106892526, 'samples': 4274496, 'steps': 22262, 'loss/train': 1.0418902337551117} 01/29/2022 17:32:23 - INFO - codeparrot_training - Step 22263: {'lr': 0.0003105386857223197, 'samples': 4274688, 'steps': 22263, 'loss/train': 1.3311922252178192} 01/29/2022 17:32:27 - INFO - codeparrot_training - Step 22264: {'lr': 0.0003105228101163856, 'samples': 4274880, 'steps': 22264, 'loss/train': 0.7109927237033844} 01/29/2022 17:32:34 - INFO - codeparrot_training - Step 22265: {'lr': 0.000310506934251191, 'samples': 4275072, 'steps': 22265, 'loss/train': 0.811455488204956} 01/29/2022 17:32:38 - INFO - codeparrot_training - Step 22266: {'lr': 0.0003104910581268039, 'samples': 4275264, 'steps': 22266, 'loss/train': 1.5158439874649048} 01/29/2022 17:32:43 - INFO - codeparrot_training - Step 22267: {'lr': 0.00031047518174329234, 'samples': 4275456, 'steps': 22267, 'loss/train': 1.293841928243637} 01/29/2022 17:32:47 - INFO - codeparrot_training - Step 22268: {'lr': 0.00031045930510072427, 'samples': 4275648, 'steps': 22268, 'loss/train': 0.9207302033901215} 01/29/2022 17:32:51 - INFO - codeparrot_training - Step 22269: {'lr': 0.00031044342819916784, 'samples': 4275840, 'steps': 22269, 'loss/train': 0.7353072613477707} 01/29/2022 17:32:57 - INFO - codeparrot_training - Step 22270: {'lr': 0.0003104275510386908, 'samples': 4276032, 'steps': 22270, 'loss/train': 1.902368187904358} 01/29/2022 17:33:01 - INFO - codeparrot_training - Step 22271: {'lr': 0.0003104116736193615, 'samples': 4276224, 'steps': 22271, 'loss/train': 1.2144276201725006} 01/29/2022 17:33:05 - INFO - codeparrot_training - Step 22272: {'lr': 0.00031039579594124763, 'samples': 4276416, 'steps': 22272, 'loss/train': 1.589021623134613} 01/29/2022 17:33:09 - INFO - codeparrot_training - Step 22273: {'lr': 0.0003103799180044174, 'samples': 4276608, 'steps': 22273, 'loss/train': 0.9128741025924683} 01/29/2022 17:33:14 - INFO - codeparrot_training - Step 22274: {'lr': 0.00031036403980893874, 'samples': 4276800, 'steps': 22274, 'loss/train': 1.0794734358787537} 01/29/2022 17:33:19 - INFO - codeparrot_training - Step 22275: {'lr': 0.0003103481613548797, 'samples': 4276992, 'steps': 22275, 'loss/train': 2.076161205768585} 01/29/2022 17:33:23 - INFO - codeparrot_training - Step 22276: {'lr': 0.00031033228264230834, 'samples': 4277184, 'steps': 22276, 'loss/train': 1.8049179911613464} 01/29/2022 17:33:27 - INFO - codeparrot_training - Step 22277: {'lr': 0.0003103164036712926, 'samples': 4277376, 'steps': 22277, 'loss/train': 1.7470189332962036} 01/29/2022 17:33:32 - INFO - codeparrot_training - Step 22278: {'lr': 0.0003103005244419006, 'samples': 4277568, 'steps': 22278, 'loss/train': 2.076910972595215} 01/29/2022 17:33:36 - INFO - codeparrot_training - Step 22279: {'lr': 0.00031028464495420026, 'samples': 4277760, 'steps': 22279, 'loss/train': 2.146156132221222} 01/29/2022 17:33:43 - INFO - codeparrot_training - Step 22280: {'lr': 0.0003102687652082597, 'samples': 4277952, 'steps': 22280, 'loss/train': 0.6180006712675095} 01/29/2022 17:33:47 - INFO - codeparrot_training - Step 22281: {'lr': 0.00031025288520414686, 'samples': 4278144, 'steps': 22281, 'loss/train': 1.481559008359909} 01/29/2022 17:33:52 - INFO - codeparrot_training - Step 22282: {'lr': 0.0003102370049419297, 'samples': 4278336, 'steps': 22282, 'loss/train': 1.5847722887992859} 01/29/2022 17:33:56 - INFO - codeparrot_training - Step 22283: {'lr': 0.0003102211244216764, 'samples': 4278528, 'steps': 22283, 'loss/train': 1.6487449407577515} 01/29/2022 17:34:00 - INFO - codeparrot_training - Step 22284: {'lr': 0.000310205243643455, 'samples': 4278720, 'steps': 22284, 'loss/train': 1.7624748945236206} 01/29/2022 17:34:05 - INFO - codeparrot_training - Step 22285: {'lr': 0.00031018936260733337, 'samples': 4278912, 'steps': 22285, 'loss/train': 1.6961397528648376} 01/29/2022 17:34:10 - INFO - codeparrot_training - Step 22286: {'lr': 0.00031017348131337963, 'samples': 4279104, 'steps': 22286, 'loss/train': 1.9290170073509216} 01/29/2022 17:34:14 - INFO - codeparrot_training - Step 22287: {'lr': 0.00031015759976166186, 'samples': 4279296, 'steps': 22287, 'loss/train': 1.5726308226585388} 01/29/2022 17:34:18 - INFO - codeparrot_training - Step 22288: {'lr': 0.00031014171795224794, 'samples': 4279488, 'steps': 22288, 'loss/train': 1.4176044166088104} 01/29/2022 17:34:22 - INFO - codeparrot_training - Step 22289: {'lr': 0.00031012583588520607, 'samples': 4279680, 'steps': 22289, 'loss/train': 1.9355388879776} 01/29/2022 17:34:29 - INFO - codeparrot_training - Step 22290: {'lr': 0.00031010995356060416, 'samples': 4279872, 'steps': 22290, 'loss/train': 1.3103908896446228} 01/29/2022 17:34:34 - INFO - codeparrot_training - Step 22291: {'lr': 0.00031009407097851036, 'samples': 4280064, 'steps': 22291, 'loss/train': 1.4351138770580292} 01/29/2022 17:34:38 - INFO - codeparrot_training - Step 22292: {'lr': 0.0003100781881389926, 'samples': 4280256, 'steps': 22292, 'loss/train': 0.7290780544281006} 01/29/2022 17:34:42 - INFO - codeparrot_training - Step 22293: {'lr': 0.00031006230504211895, 'samples': 4280448, 'steps': 22293, 'loss/train': 2.0411185026168823} 01/29/2022 17:34:46 - INFO - codeparrot_training - Step 22294: {'lr': 0.0003100464216879574, 'samples': 4280640, 'steps': 22294, 'loss/train': 1.146393746137619} 01/29/2022 17:34:52 - INFO - codeparrot_training - Step 22295: {'lr': 0.0003100305380765762, 'samples': 4280832, 'steps': 22295, 'loss/train': 1.8149932622909546} 01/29/2022 17:34:56 - INFO - codeparrot_training - Step 22296: {'lr': 0.00031001465420804316, 'samples': 4281024, 'steps': 22296, 'loss/train': 2.023344576358795} 01/29/2022 17:35:00 - INFO - codeparrot_training - Step 22297: {'lr': 0.0003099987700824264, 'samples': 4281216, 'steps': 22297, 'loss/train': 1.3968597650527954} 01/29/2022 17:35:04 - INFO - codeparrot_training - Step 22298: {'lr': 0.00030998288569979393, 'samples': 4281408, 'steps': 22298, 'loss/train': 2.2361006140708923} 01/29/2022 17:35:09 - INFO - codeparrot_training - Step 22299: {'lr': 0.0003099670010602138, 'samples': 4281600, 'steps': 22299, 'loss/train': 1.6231005191802979} 01/29/2022 17:35:14 - INFO - codeparrot_training - Step 22300: {'lr': 0.00030995111616375417, 'samples': 4281792, 'steps': 22300, 'loss/train': 1.4484961330890656} 01/29/2022 17:35:18 - INFO - codeparrot_training - Step 22301: {'lr': 0.00030993523101048294, 'samples': 4281984, 'steps': 22301, 'loss/train': 0.8413499593734741} 01/29/2022 17:35:22 - INFO - codeparrot_training - Step 22302: {'lr': 0.0003099193456004682, 'samples': 4282176, 'steps': 22302, 'loss/train': 0.660102054476738} 01/29/2022 17:35:27 - INFO - codeparrot_training - Step 22303: {'lr': 0.00030990345993377807, 'samples': 4282368, 'steps': 22303, 'loss/train': 1.9921563863754272} 01/29/2022 17:35:31 - INFO - codeparrot_training - Step 22304: {'lr': 0.0003098875740104805, 'samples': 4282560, 'steps': 22304, 'loss/train': 2.132050037384033} 01/29/2022 17:35:38 - INFO - codeparrot_training - Step 22305: {'lr': 0.00030987168783064355, 'samples': 4282752, 'steps': 22305, 'loss/train': 1.6431142687797546} 01/29/2022 17:35:42 - INFO - codeparrot_training - Step 22306: {'lr': 0.0003098558013943353, 'samples': 4282944, 'steps': 22306, 'loss/train': 2.222215175628662} 01/29/2022 17:35:46 - INFO - codeparrot_training - Step 22307: {'lr': 0.00030983991470162386, 'samples': 4283136, 'steps': 22307, 'loss/train': 1.9035723209381104} 01/29/2022 17:35:51 - INFO - codeparrot_training - Step 22308: {'lr': 0.00030982402775257725, 'samples': 4283328, 'steps': 22308, 'loss/train': 2.1157860159873962} 01/29/2022 17:35:55 - INFO - codeparrot_training - Step 22309: {'lr': 0.0003098081405472634, 'samples': 4283520, 'steps': 22309, 'loss/train': 2.3537063598632812} 01/29/2022 17:36:00 - INFO - codeparrot_training - Step 22310: {'lr': 0.0003097922530857505, 'samples': 4283712, 'steps': 22310, 'loss/train': 1.9063045978546143} 01/29/2022 17:36:05 - INFO - codeparrot_training - Step 22311: {'lr': 0.0003097763653681066, 'samples': 4283904, 'steps': 22311, 'loss/train': 1.2291037738323212} 01/29/2022 17:36:09 - INFO - codeparrot_training - Step 22312: {'lr': 0.00030976047739439974, 'samples': 4284096, 'steps': 22312, 'loss/train': 1.926752507686615} 01/29/2022 17:36:13 - INFO - codeparrot_training - Step 22313: {'lr': 0.000309744589164698, 'samples': 4284288, 'steps': 22313, 'loss/train': 2.6930195689201355} 01/29/2022 17:36:17 - INFO - codeparrot_training - Step 22314: {'lr': 0.00030972870067906934, 'samples': 4284480, 'steps': 22314, 'loss/train': 2.834292948246002} 01/29/2022 17:36:23 - INFO - codeparrot_training - Step 22315: {'lr': 0.00030971281193758197, 'samples': 4284672, 'steps': 22315, 'loss/train': 2.2319204807281494} 01/29/2022 17:36:27 - INFO - codeparrot_training - Step 22316: {'lr': 0.00030969692294030376, 'samples': 4284864, 'steps': 22316, 'loss/train': 2.0661458373069763} 01/29/2022 17:36:31 - INFO - codeparrot_training - Step 22317: {'lr': 0.000309681033687303, 'samples': 4285056, 'steps': 22317, 'loss/train': 1.3257966935634613} 01/29/2022 17:36:35 - INFO - codeparrot_training - Step 22318: {'lr': 0.0003096651441786476, 'samples': 4285248, 'steps': 22318, 'loss/train': 2.265101909637451} 01/29/2022 17:36:40 - INFO - codeparrot_training - Step 22319: {'lr': 0.0003096492544144056, 'samples': 4285440, 'steps': 22319, 'loss/train': 1.5896835923194885} 01/29/2022 17:36:45 - INFO - codeparrot_training - Step 22320: {'lr': 0.00030963336439464523, 'samples': 4285632, 'steps': 22320, 'loss/train': 1.6637864112854004} 01/29/2022 17:36:49 - INFO - codeparrot_training - Step 22321: {'lr': 0.0003096174741194344, 'samples': 4285824, 'steps': 22321, 'loss/train': 1.4753691852092743} 01/29/2022 17:36:53 - INFO - codeparrot_training - Step 22322: {'lr': 0.00030960158358884127, 'samples': 4286016, 'steps': 22322, 'loss/train': 1.7418606877326965} 01/29/2022 17:36:58 - INFO - codeparrot_training - Step 22323: {'lr': 0.0003095856928029339, 'samples': 4286208, 'steps': 22323, 'loss/train': 1.9747254252433777} 01/29/2022 17:37:02 - INFO - codeparrot_training - Step 22324: {'lr': 0.00030956980176178033, 'samples': 4286400, 'steps': 22324, 'loss/train': 1.3094801008701324} 01/29/2022 17:37:10 - INFO - codeparrot_training - Step 22325: {'lr': 0.00030955391046544865, 'samples': 4286592, 'steps': 22325, 'loss/train': 1.7868545651435852} 01/29/2022 17:37:14 - INFO - codeparrot_training - Step 22326: {'lr': 0.0003095380189140069, 'samples': 4286784, 'steps': 22326, 'loss/train': 0.8536050617694855} 01/29/2022 17:37:19 - INFO - codeparrot_training - Step 22327: {'lr': 0.00030952212710752325, 'samples': 4286976, 'steps': 22327, 'loss/train': 2.478178918361664} 01/29/2022 17:37:23 - INFO - codeparrot_training - Step 22328: {'lr': 0.00030950623504606565, 'samples': 4287168, 'steps': 22328, 'loss/train': 2.2365174293518066} 01/29/2022 17:37:27 - INFO - codeparrot_training - Step 22329: {'lr': 0.0003094903427297023, 'samples': 4287360, 'steps': 22329, 'loss/train': 1.9033849239349365} 01/29/2022 17:37:33 - INFO - codeparrot_training - Step 22330: {'lr': 0.00030947445015850114, 'samples': 4287552, 'steps': 22330, 'loss/train': 2.090012848377228} 01/29/2022 17:37:37 - INFO - codeparrot_training - Step 22331: {'lr': 0.0003094585573325303, 'samples': 4287744, 'steps': 22331, 'loss/train': 1.7306132912635803} 01/29/2022 17:37:41 - INFO - codeparrot_training - Step 22332: {'lr': 0.00030944266425185794, 'samples': 4287936, 'steps': 22332, 'loss/train': 1.3800647556781769} 01/29/2022 17:37:45 - INFO - codeparrot_training - Step 22333: {'lr': 0.000309426770916552, 'samples': 4288128, 'steps': 22333, 'loss/train': 1.7704979181289673} 01/29/2022 17:37:50 - INFO - codeparrot_training - Step 22334: {'lr': 0.0003094108773266808, 'samples': 4288320, 'steps': 22334, 'loss/train': 1.5354090929031372} 01/29/2022 17:37:54 - INFO - codeparrot_training - Step 22335: {'lr': 0.00030939498348231214, 'samples': 4288512, 'steps': 22335, 'loss/train': 1.1425690054893494} 01/29/2022 17:37:59 - INFO - codeparrot_training - Step 22336: {'lr': 0.00030937908938351424, 'samples': 4288704, 'steps': 22336, 'loss/train': 1.8589419722557068} 01/29/2022 17:38:03 - INFO - codeparrot_training - Step 22337: {'lr': 0.0003093631950303552, 'samples': 4288896, 'steps': 22337, 'loss/train': 0.7466406673192978} 01/29/2022 17:38:08 - INFO - codeparrot_training - Step 22338: {'lr': 0.00030934730042290304, 'samples': 4289088, 'steps': 22338, 'loss/train': 1.763006865978241} 01/29/2022 17:38:12 - INFO - codeparrot_training - Step 22339: {'lr': 0.00030933140556122597, 'samples': 4289280, 'steps': 22339, 'loss/train': 1.777713418006897} 01/29/2022 17:38:16 - INFO - codeparrot_training - Step 22340: {'lr': 0.00030931551044539196, 'samples': 4289472, 'steps': 22340, 'loss/train': 2.299822211265564} 01/29/2022 17:38:23 - INFO - codeparrot_training - Step 22341: {'lr': 0.00030929961507546915, 'samples': 4289664, 'steps': 22341, 'loss/train': 1.2725092470645905} 01/29/2022 17:38:27 - INFO - codeparrot_training - Step 22342: {'lr': 0.0003092837194515256, 'samples': 4289856, 'steps': 22342, 'loss/train': 1.8722877502441406} 01/29/2022 17:38:32 - INFO - codeparrot_training - Step 22343: {'lr': 0.00030926782357362944, 'samples': 4290048, 'steps': 22343, 'loss/train': 2.0432660579681396} 01/29/2022 17:38:36 - INFO - codeparrot_training - Step 22344: {'lr': 0.0003092519274418487, 'samples': 4290240, 'steps': 22344, 'loss/train': 1.704979419708252} 01/29/2022 17:38:41 - INFO - codeparrot_training - Step 22345: {'lr': 0.0003092360310562516, 'samples': 4290432, 'steps': 22345, 'loss/train': 1.213603287935257} 01/29/2022 17:38:45 - INFO - codeparrot_training - Step 22346: {'lr': 0.00030922013441690607, 'samples': 4290624, 'steps': 22346, 'loss/train': 1.7939379215240479} 01/29/2022 17:38:50 - INFO - codeparrot_training - Step 22347: {'lr': 0.0003092042375238803, 'samples': 4290816, 'steps': 22347, 'loss/train': 1.5799673795700073} 01/29/2022 17:38:54 - INFO - codeparrot_training - Step 22348: {'lr': 0.00030918834037724235, 'samples': 4291008, 'steps': 22348, 'loss/train': 2.3476885557174683} 01/29/2022 17:38:58 - INFO - codeparrot_training - Step 22349: {'lr': 0.0003091724429770604, 'samples': 4291200, 'steps': 22349, 'loss/train': 1.6843151450157166} 01/29/2022 17:39:05 - INFO - codeparrot_training - Step 22350: {'lr': 0.0003091565453234025, 'samples': 4291392, 'steps': 22350, 'loss/train': 1.0479763448238373} 01/29/2022 17:39:10 - INFO - codeparrot_training - Step 22351: {'lr': 0.0003091406474163367, 'samples': 4291584, 'steps': 22351, 'loss/train': 1.6226261258125305} 01/29/2022 17:39:14 - INFO - codeparrot_training - Step 22352: {'lr': 0.0003091247492559312, 'samples': 4291776, 'steps': 22352, 'loss/train': 0.8822035789489746} 01/29/2022 17:39:18 - INFO - codeparrot_training - Step 22353: {'lr': 0.00030910885084225396, 'samples': 4291968, 'steps': 22353, 'loss/train': 1.8700103759765625} 01/29/2022 17:39:22 - INFO - codeparrot_training - Step 22354: {'lr': 0.0003090929521753733, 'samples': 4292160, 'steps': 22354, 'loss/train': 2.0320900082588196} 01/29/2022 17:39:27 - INFO - codeparrot_training - Step 22355: {'lr': 0.00030907705325535704, 'samples': 4292352, 'steps': 22355, 'loss/train': 2.1835582852363586} 01/29/2022 17:39:32 - INFO - codeparrot_training - Step 22356: {'lr': 0.0003090611540822736, 'samples': 4292544, 'steps': 22356, 'loss/train': 1.2920957207679749} 01/29/2022 17:39:36 - INFO - codeparrot_training - Step 22357: {'lr': 0.0003090452546561908, 'samples': 4292736, 'steps': 22357, 'loss/train': 1.5857702493667603} 01/29/2022 17:39:41 - INFO - codeparrot_training - Step 22358: {'lr': 0.000309029354977177, 'samples': 4292928, 'steps': 22358, 'loss/train': 0.9066223204135895} 01/29/2022 17:39:45 - INFO - codeparrot_training - Step 22359: {'lr': 0.00030901345504530007, 'samples': 4293120, 'steps': 22359, 'loss/train': 1.1899998486042023} 01/29/2022 17:39:49 - INFO - codeparrot_training - Step 22360: {'lr': 0.0003089975548606283, 'samples': 4293312, 'steps': 22360, 'loss/train': 0.5644245743751526} 01/29/2022 17:39:55 - INFO - codeparrot_training - Step 22361: {'lr': 0.0003089816544232298, 'samples': 4293504, 'steps': 22361, 'loss/train': 1.7287184000015259} 01/29/2022 17:39:59 - INFO - codeparrot_training - Step 22362: {'lr': 0.00030896575373317247, 'samples': 4293696, 'steps': 22362, 'loss/train': 2.126150965690613} 01/29/2022 17:40:03 - INFO - codeparrot_training - Step 22363: {'lr': 0.0003089498527905247, 'samples': 4293888, 'steps': 22363, 'loss/train': 1.7723152041435242} 01/29/2022 17:40:07 - INFO - codeparrot_training - Step 22364: {'lr': 0.00030893395159535444, 'samples': 4294080, 'steps': 22364, 'loss/train': 0.7271728813648224} 01/29/2022 17:40:12 - INFO - codeparrot_training - Step 22365: {'lr': 0.00030891805014772987, 'samples': 4294272, 'steps': 22365, 'loss/train': 1.3441012501716614} 01/29/2022 17:40:16 - INFO - codeparrot_training - Step 22366: {'lr': 0.000308902148447719, 'samples': 4294464, 'steps': 22366, 'loss/train': 2.6977989077568054} 01/29/2022 17:40:23 - INFO - codeparrot_training - Step 22367: {'lr': 0.00030888624649539015, 'samples': 4294656, 'steps': 22367, 'loss/train': 1.4721782505512238} 01/29/2022 17:40:27 - INFO - codeparrot_training - Step 22368: {'lr': 0.0003088703442908112, 'samples': 4294848, 'steps': 22368, 'loss/train': 0.4726148396730423} 01/29/2022 17:40:31 - INFO - codeparrot_training - Step 22369: {'lr': 0.0003088544418340505, 'samples': 4295040, 'steps': 22369, 'loss/train': 1.6314157247543335} 01/29/2022 17:40:36 - INFO - codeparrot_training - Step 22370: {'lr': 0.000308838539125176, 'samples': 4295232, 'steps': 22370, 'loss/train': 1.34530770778656} 01/29/2022 17:40:40 - INFO - codeparrot_training - Step 22371: {'lr': 0.00030882263616425587, 'samples': 4295424, 'steps': 22371, 'loss/train': 1.3191091418266296} 01/29/2022 17:40:45 - INFO - codeparrot_training - Step 22372: {'lr': 0.0003088067329513583, 'samples': 4295616, 'steps': 22372, 'loss/train': 1.6502763032913208} 01/29/2022 17:40:49 - INFO - codeparrot_training - Step 22373: {'lr': 0.0003087908294865513, 'samples': 4295808, 'steps': 22373, 'loss/train': 1.7070606350898743} 01/29/2022 17:40:54 - INFO - codeparrot_training - Step 22374: {'lr': 0.00030877492576990306, 'samples': 4296000, 'steps': 22374, 'loss/train': 1.7726672887802124} 01/29/2022 17:40:58 - INFO - codeparrot_training - Step 22375: {'lr': 0.0003087590218014817, 'samples': 4296192, 'steps': 22375, 'loss/train': 1.4497633874416351} 01/29/2022 17:41:02 - INFO - codeparrot_training - Step 22376: {'lr': 0.00030874311758135535, 'samples': 4296384, 'steps': 22376, 'loss/train': 2.1837775111198425} 01/29/2022 17:41:09 - INFO - codeparrot_training - Step 22377: {'lr': 0.00030872721310959216, 'samples': 4296576, 'steps': 22377, 'loss/train': 2.0151114463806152} 01/29/2022 17:41:14 - INFO - codeparrot_training - Step 22378: {'lr': 0.00030871130838626025, 'samples': 4296768, 'steps': 22378, 'loss/train': 2.0032755732536316} 01/29/2022 17:41:18 - INFO - codeparrot_training - Step 22379: {'lr': 0.0003086954034114277, 'samples': 4296960, 'steps': 22379, 'loss/train': 1.018265962600708} 01/29/2022 17:41:22 - INFO - codeparrot_training - Step 22380: {'lr': 0.00030867949818516274, 'samples': 4297152, 'steps': 22380, 'loss/train': 1.2838378250598907} 01/29/2022 17:41:26 - INFO - codeparrot_training - Step 22381: {'lr': 0.00030866359270753337, 'samples': 4297344, 'steps': 22381, 'loss/train': 1.8699955344200134} 01/29/2022 17:41:32 - INFO - codeparrot_training - Step 22382: {'lr': 0.00030864768697860784, 'samples': 4297536, 'steps': 22382, 'loss/train': 1.6996850967407227} 01/29/2022 17:41:36 - INFO - codeparrot_training - Step 22383: {'lr': 0.0003086317809984542, 'samples': 4297728, 'steps': 22383, 'loss/train': 0.8695022463798523} 01/29/2022 17:41:40 - INFO - codeparrot_training - Step 22384: {'lr': 0.0003086158747671406, 'samples': 4297920, 'steps': 22384, 'loss/train': 1.8485605716705322} 01/29/2022 17:41:44 - INFO - codeparrot_training - Step 22385: {'lr': 0.0003085999682847353, 'samples': 4298112, 'steps': 22385, 'loss/train': 1.7691872119903564} 01/29/2022 17:41:49 - INFO - codeparrot_training - Step 22386: {'lr': 0.00030858406155130625, 'samples': 4298304, 'steps': 22386, 'loss/train': 1.9380258321762085} 01/29/2022 17:41:54 - INFO - codeparrot_training - Step 22387: {'lr': 0.00030856815456692177, 'samples': 4298496, 'steps': 22387, 'loss/train': 0.49799448251724243} 01/29/2022 17:41:58 - INFO - codeparrot_training - Step 22388: {'lr': 0.00030855224733164987, 'samples': 4298688, 'steps': 22388, 'loss/train': 1.6145399808883667} 01/29/2022 17:42:02 - INFO - codeparrot_training - Step 22389: {'lr': 0.00030853633984555875, 'samples': 4298880, 'steps': 22389, 'loss/train': 1.6426806449890137} 01/29/2022 17:42:07 - INFO - codeparrot_training - Step 22390: {'lr': 0.0003085204321087165, 'samples': 4299072, 'steps': 22390, 'loss/train': 1.0875738859176636} 01/29/2022 17:42:11 - INFO - codeparrot_training - Step 22391: {'lr': 0.00030850452412119135, 'samples': 4299264, 'steps': 22391, 'loss/train': 1.536019206047058} 01/29/2022 17:42:16 - INFO - codeparrot_training - Step 22392: {'lr': 0.00030848861588305136, 'samples': 4299456, 'steps': 22392, 'loss/train': 2.095602035522461} 01/29/2022 17:42:20 - INFO - codeparrot_training - Step 22393: {'lr': 0.0003084727073943648, 'samples': 4299648, 'steps': 22393, 'loss/train': 1.5165832042694092} 01/29/2022 17:42:25 - INFO - codeparrot_training - Step 22394: {'lr': 0.0003084567986551996, 'samples': 4299840, 'steps': 22394, 'loss/train': 1.2301273047924042} 01/29/2022 17:42:29 - INFO - codeparrot_training - Step 22395: {'lr': 0.0003084408896656241, 'samples': 4300032, 'steps': 22395, 'loss/train': 1.1612260043621063} 01/29/2022 17:42:33 - INFO - codeparrot_training - Step 22396: {'lr': 0.0003084249804257064, 'samples': 4300224, 'steps': 22396, 'loss/train': 1.5454999208450317} 01/29/2022 17:42:40 - INFO - codeparrot_training - Step 22397: {'lr': 0.00030840907093551456, 'samples': 4300416, 'steps': 22397, 'loss/train': 1.8851879239082336} 01/29/2022 17:42:45 - INFO - codeparrot_training - Step 22398: {'lr': 0.0003083931611951169, 'samples': 4300608, 'steps': 22398, 'loss/train': 2.304607629776001} 01/29/2022 17:42:49 - INFO - codeparrot_training - Step 22399: {'lr': 0.0003083772512045814, 'samples': 4300800, 'steps': 22399, 'loss/train': 0.30224888026714325} 01/29/2022 17:42:53 - INFO - codeparrot_training - Step 22400: {'lr': 0.0003083613409639764, 'samples': 4300992, 'steps': 22400, 'loss/train': 1.338991105556488} 01/29/2022 17:42:57 - INFO - codeparrot_training - Step 22401: {'lr': 0.0003083454304733698, 'samples': 4301184, 'steps': 22401, 'loss/train': 2.3591830730438232} 01/29/2022 17:43:03 - INFO - codeparrot_training - Step 22402: {'lr': 0.00030832951973283, 'samples': 4301376, 'steps': 22402, 'loss/train': 2.3476309776306152} 01/29/2022 17:43:07 - INFO - codeparrot_training - Step 22403: {'lr': 0.000308313608742425, 'samples': 4301568, 'steps': 22403, 'loss/train': 1.132750242948532} 01/29/2022 17:43:11 - INFO - codeparrot_training - Step 22404: {'lr': 0.00030829769750222315, 'samples': 4301760, 'steps': 22404, 'loss/train': 1.8457310199737549} 01/29/2022 17:43:16 - INFO - codeparrot_training - Step 22405: {'lr': 0.00030828178601229235, 'samples': 4301952, 'steps': 22405, 'loss/train': 1.5083738565444946} 01/29/2022 17:43:21 - INFO - codeparrot_training - Step 22406: {'lr': 0.00030826587427270095, 'samples': 4302144, 'steps': 22406, 'loss/train': 2.1059165596961975} 01/29/2022 17:43:25 - INFO - codeparrot_training - Step 22407: {'lr': 0.000308249962283517, 'samples': 4302336, 'steps': 22407, 'loss/train': 1.6618793606758118} 01/29/2022 17:43:29 - INFO - codeparrot_training - Step 22408: {'lr': 0.0003082340500448087, 'samples': 4302528, 'steps': 22408, 'loss/train': 0.7046957910060883} 01/29/2022 17:43:34 - INFO - codeparrot_training - Step 22409: {'lr': 0.0003082181375566443, 'samples': 4302720, 'steps': 22409, 'loss/train': 2.4371745586395264} 01/29/2022 17:43:38 - INFO - codeparrot_training - Step 22410: {'lr': 0.0003082022248190918, 'samples': 4302912, 'steps': 22410, 'loss/train': 1.1420212090015411} 01/29/2022 17:43:45 - INFO - codeparrot_training - Step 22411: {'lr': 0.00030818631183221945, 'samples': 4303104, 'steps': 22411, 'loss/train': 1.1297741532325745} 01/29/2022 17:43:49 - INFO - codeparrot_training - Step 22412: {'lr': 0.0003081703985960955, 'samples': 4303296, 'steps': 22412, 'loss/train': 1.4712757766246796} 01/29/2022 17:43:54 - INFO - codeparrot_training - Step 22413: {'lr': 0.00030815448511078796, 'samples': 4303488, 'steps': 22413, 'loss/train': 1.00458624958992} 01/29/2022 17:43:58 - INFO - codeparrot_training - Step 22414: {'lr': 0.0003081385713763651, 'samples': 4303680, 'steps': 22414, 'loss/train': 1.9919090867042542} 01/29/2022 17:44:02 - INFO - codeparrot_training - Step 22415: {'lr': 0.00030812265739289503, 'samples': 4303872, 'steps': 22415, 'loss/train': 1.8439516425132751} 01/29/2022 17:44:07 - INFO - codeparrot_training - Step 22416: {'lr': 0.000308106743160446, 'samples': 4304064, 'steps': 22416, 'loss/train': 2.065950036048889} 01/29/2022 17:44:12 - INFO - codeparrot_training - Step 22417: {'lr': 0.00030809082867908614, 'samples': 4304256, 'steps': 22417, 'loss/train': 1.742898166179657} 01/29/2022 17:44:16 - INFO - codeparrot_training - Step 22418: {'lr': 0.0003080749139488836, 'samples': 4304448, 'steps': 22418, 'loss/train': 0.36529695242643356} 01/29/2022 17:44:20 - INFO - codeparrot_training - Step 22419: {'lr': 0.0003080589989699066, 'samples': 4304640, 'steps': 22419, 'loss/train': 0.9063854813575745} 01/29/2022 17:44:24 - INFO - codeparrot_training - Step 22420: {'lr': 0.00030804308374222315, 'samples': 4304832, 'steps': 22420, 'loss/train': 1.8355937004089355} 01/29/2022 17:44:31 - INFO - codeparrot_training - Step 22421: {'lr': 0.00030802716826590164, 'samples': 4305024, 'steps': 22421, 'loss/train': 1.2540223002433777} 01/29/2022 17:44:36 - INFO - codeparrot_training - Step 22422: {'lr': 0.0003080112525410102, 'samples': 4305216, 'steps': 22422, 'loss/train': 1.1382534205913544} 01/29/2022 17:44:40 - INFO - codeparrot_training - Step 22423: {'lr': 0.0003079953365676169, 'samples': 4305408, 'steps': 22423, 'loss/train': 2.1711888313293457} 01/29/2022 17:44:44 - INFO - codeparrot_training - Step 22424: {'lr': 0.00030797942034579013, 'samples': 4305600, 'steps': 22424, 'loss/train': 1.7582064270973206} 01/29/2022 17:44:48 - INFO - codeparrot_training - Step 22425: {'lr': 0.0003079635038755978, 'samples': 4305792, 'steps': 22425, 'loss/train': 1.579576849937439} 01/29/2022 17:44:53 - INFO - codeparrot_training - Step 22426: {'lr': 0.0003079475871571083, 'samples': 4305984, 'steps': 22426, 'loss/train': 1.3632395267486572} 01/29/2022 17:44:58 - INFO - codeparrot_training - Step 22427: {'lr': 0.0003079316701903897, 'samples': 4306176, 'steps': 22427, 'loss/train': 0.3846439719200134} 01/29/2022 17:45:02 - INFO - codeparrot_training - Step 22428: {'lr': 0.0003079157529755102, 'samples': 4306368, 'steps': 22428, 'loss/train': 1.976223886013031} 01/29/2022 17:45:06 - INFO - codeparrot_training - Step 22429: {'lr': 0.0003078998355125381, 'samples': 4306560, 'steps': 22429, 'loss/train': 1.2478366792201996} 01/29/2022 17:45:10 - INFO - codeparrot_training - Step 22430: {'lr': 0.0003078839178015414, 'samples': 4306752, 'steps': 22430, 'loss/train': 0.7927024662494659} 01/29/2022 17:45:16 - INFO - codeparrot_training - Step 22431: {'lr': 0.0003078679998425884, 'samples': 4306944, 'steps': 22431, 'loss/train': 0.7134803831577301} 01/29/2022 17:45:20 - INFO - codeparrot_training - Step 22432: {'lr': 0.0003078520816357472, 'samples': 4307136, 'steps': 22432, 'loss/train': 0.6326070874929428} 01/29/2022 17:45:25 - INFO - codeparrot_training - Step 22433: {'lr': 0.00030783616318108613, 'samples': 4307328, 'steps': 22433, 'loss/train': 0.40483640134334564} 01/29/2022 17:45:29 - INFO - codeparrot_training - Step 22434: {'lr': 0.0003078202444786733, 'samples': 4307520, 'steps': 22434, 'loss/train': 1.287265419960022} 01/29/2022 17:45:33 - INFO - codeparrot_training - Step 22435: {'lr': 0.0003078043255285769, 'samples': 4307712, 'steps': 22435, 'loss/train': 2.067869782447815} 01/29/2022 17:45:40 - INFO - codeparrot_training - Step 22436: {'lr': 0.00030778840633086514, 'samples': 4307904, 'steps': 22436, 'loss/train': 1.4772403836250305} 01/29/2022 17:45:44 - INFO - codeparrot_training - Step 22437: {'lr': 0.00030777248688560615, 'samples': 4308096, 'steps': 22437, 'loss/train': 1.7748751044273376} 01/29/2022 17:45:49 - INFO - codeparrot_training - Step 22438: {'lr': 0.0003077565671928682, 'samples': 4308288, 'steps': 22438, 'loss/train': 2.161752998828888} 01/29/2022 17:45:53 - INFO - codeparrot_training - Step 22439: {'lr': 0.00030774064725271944, 'samples': 4308480, 'steps': 22439, 'loss/train': 1.4943267703056335} 01/29/2022 17:45:57 - INFO - codeparrot_training - Step 22440: {'lr': 0.00030772472706522806, 'samples': 4308672, 'steps': 22440, 'loss/train': 1.1109226047992706} 01/29/2022 17:46:02 - INFO - codeparrot_training - Step 22441: {'lr': 0.00030770880663046236, 'samples': 4308864, 'steps': 22441, 'loss/train': 0.37880176305770874} 01/29/2022 17:46:07 - INFO - codeparrot_training - Step 22442: {'lr': 0.00030769288594849044, 'samples': 4309056, 'steps': 22442, 'loss/train': 1.882781982421875} 01/29/2022 17:46:11 - INFO - codeparrot_training - Step 22443: {'lr': 0.0003076769650193805, 'samples': 4309248, 'steps': 22443, 'loss/train': 1.352865993976593} 01/29/2022 17:46:15 - INFO - codeparrot_training - Step 22444: {'lr': 0.0003076610438432007, 'samples': 4309440, 'steps': 22444, 'loss/train': 2.614687979221344} 01/29/2022 17:46:19 - INFO - codeparrot_training - Step 22445: {'lr': 0.00030764512242001927, 'samples': 4309632, 'steps': 22445, 'loss/train': 1.2921480238437653} 01/29/2022 17:46:25 - INFO - codeparrot_training - Step 22446: {'lr': 0.00030762920074990457, 'samples': 4309824, 'steps': 22446, 'loss/train': 1.4249204099178314} 01/29/2022 17:46:29 - INFO - codeparrot_training - Step 22447: {'lr': 0.00030761327883292456, 'samples': 4310016, 'steps': 22447, 'loss/train': 2.6037368774414062} 01/29/2022 17:46:33 - INFO - codeparrot_training - Step 22448: {'lr': 0.00030759735666914767, 'samples': 4310208, 'steps': 22448, 'loss/train': 1.8982315063476562} 01/29/2022 17:46:37 - INFO - codeparrot_training - Step 22449: {'lr': 0.00030758143425864187, 'samples': 4310400, 'steps': 22449, 'loss/train': 1.5323554873466492} 01/29/2022 17:46:42 - INFO - codeparrot_training - Step 22450: {'lr': 0.00030756551160147563, 'samples': 4310592, 'steps': 22450, 'loss/train': 1.5064709186553955} 01/29/2022 17:46:47 - INFO - codeparrot_training - Step 22451: {'lr': 0.0003075495886977169, 'samples': 4310784, 'steps': 22451, 'loss/train': 1.736323893070221} 01/29/2022 17:46:51 - INFO - codeparrot_training - Step 22452: {'lr': 0.0003075336655474341, 'samples': 4310976, 'steps': 22452, 'loss/train': 1.7470561265945435} 01/29/2022 17:46:55 - INFO - codeparrot_training - Step 22453: {'lr': 0.0003075177421506952, 'samples': 4311168, 'steps': 22453, 'loss/train': 2.2157617807388306} 01/29/2022 17:47:00 - INFO - codeparrot_training - Step 22454: {'lr': 0.0003075018185075687, 'samples': 4311360, 'steps': 22454, 'loss/train': 1.2430177330970764} 01/29/2022 17:47:04 - INFO - codeparrot_training - Step 22455: {'lr': 0.0003074858946181226, 'samples': 4311552, 'steps': 22455, 'loss/train': 0.9690233767032623} 01/29/2022 17:47:11 - INFO - codeparrot_training - Step 22456: {'lr': 0.0003074699704824252, 'samples': 4311744, 'steps': 22456, 'loss/train': 1.3582372069358826} 01/29/2022 17:47:15 - INFO - codeparrot_training - Step 22457: {'lr': 0.0003074540461005447, 'samples': 4311936, 'steps': 22457, 'loss/train': 1.849898099899292} 01/29/2022 17:47:19 - INFO - codeparrot_training - Step 22458: {'lr': 0.00030743812147254935, 'samples': 4312128, 'steps': 22458, 'loss/train': 1.0980616807937622} 01/29/2022 17:47:24 - INFO - codeparrot_training - Step 22459: {'lr': 0.0003074221965985073, 'samples': 4312320, 'steps': 22459, 'loss/train': 0.9441007375717163} 01/29/2022 17:47:28 - INFO - codeparrot_training - Step 22460: {'lr': 0.0003074062714784867, 'samples': 4312512, 'steps': 22460, 'loss/train': 2.294106602668762} 01/29/2022 17:47:33 - INFO - codeparrot_training - Step 22461: {'lr': 0.000307390346112556, 'samples': 4312704, 'steps': 22461, 'loss/train': 1.6134881973266602} 01/29/2022 17:47:37 - INFO - codeparrot_training - Step 22462: {'lr': 0.0003073744205007832, 'samples': 4312896, 'steps': 22462, 'loss/train': 1.411789745092392} 01/29/2022 17:47:42 - INFO - codeparrot_training - Step 22463: {'lr': 0.0003073584946432366, 'samples': 4313088, 'steps': 22463, 'loss/train': 0.2625993937253952} 01/29/2022 17:47:46 - INFO - codeparrot_training - Step 22464: {'lr': 0.00030734256853998446, 'samples': 4313280, 'steps': 22464, 'loss/train': 1.9448096752166748} 01/29/2022 17:47:53 - INFO - codeparrot_training - Step 22465: {'lr': 0.00030732664219109497, 'samples': 4313472, 'steps': 22465, 'loss/train': 0.5085471868515015} 01/29/2022 17:47:57 - INFO - codeparrot_training - Step 22466: {'lr': 0.00030731071559663624, 'samples': 4313664, 'steps': 22466, 'loss/train': 1.8177709579467773} 01/29/2022 17:48:01 - INFO - codeparrot_training - Step 22467: {'lr': 0.0003072947887566768, 'samples': 4313856, 'steps': 22467, 'loss/train': 1.1331604421138763} 01/29/2022 17:48:06 - INFO - codeparrot_training - Step 22468: {'lr': 0.0003072788616712845, 'samples': 4314048, 'steps': 22468, 'loss/train': 2.07930064201355} 01/29/2022 17:48:10 - INFO - codeparrot_training - Step 22469: {'lr': 0.0003072629343405278, 'samples': 4314240, 'steps': 22469, 'loss/train': 1.8916436433792114} 01/29/2022 17:48:15 - INFO - codeparrot_training - Step 22470: {'lr': 0.00030724700676447485, 'samples': 4314432, 'steps': 22470, 'loss/train': 1.174730658531189} 01/29/2022 17:48:19 - INFO - codeparrot_training - Step 22471: {'lr': 0.00030723107894319393, 'samples': 4314624, 'steps': 22471, 'loss/train': 1.998643934726715} 01/29/2022 17:48:24 - INFO - codeparrot_training - Step 22472: {'lr': 0.00030721515087675326, 'samples': 4314816, 'steps': 22472, 'loss/train': 1.112708330154419} 01/29/2022 17:48:28 - INFO - codeparrot_training - Step 22473: {'lr': 0.00030719922256522105, 'samples': 4315008, 'steps': 22473, 'loss/train': 1.307160347700119} 01/29/2022 17:48:32 - INFO - codeparrot_training - Step 22474: {'lr': 0.0003071832940086655, 'samples': 4315200, 'steps': 22474, 'loss/train': 0.8839629292488098} 01/29/2022 17:48:38 - INFO - codeparrot_training - Step 22475: {'lr': 0.0003071673652071549, 'samples': 4315392, 'steps': 22475, 'loss/train': 1.9242262244224548} 01/29/2022 17:48:42 - INFO - codeparrot_training - Step 22476: {'lr': 0.0003071514361607575, 'samples': 4315584, 'steps': 22476, 'loss/train': 1.864431917667389} 01/29/2022 17:48:46 - INFO - codeparrot_training - Step 22477: {'lr': 0.0003071355068695414, 'samples': 4315776, 'steps': 22477, 'loss/train': 0.9099798202514648} 01/29/2022 17:48:50 - INFO - codeparrot_training - Step 22478: {'lr': 0.000307119577333575, 'samples': 4315968, 'steps': 22478, 'loss/train': 1.9412607550621033} 01/29/2022 17:48:55 - INFO - codeparrot_training - Step 22479: {'lr': 0.0003071036475529264, 'samples': 4316160, 'steps': 22479, 'loss/train': 1.875309705734253} 01/29/2022 17:49:02 - INFO - codeparrot_training - Step 22480: {'lr': 0.00030708771752766396, 'samples': 4316352, 'steps': 22480, 'loss/train': 1.2843638956546783} 01/29/2022 17:49:06 - INFO - codeparrot_training - Step 22481: {'lr': 0.0003070717872578558, 'samples': 4316544, 'steps': 22481, 'loss/train': 1.0168274939060211} 01/29/2022 17:49:10 - INFO - codeparrot_training - Step 22482: {'lr': 0.0003070558567435703, 'samples': 4316736, 'steps': 22482, 'loss/train': 1.393455058336258} 01/29/2022 17:49:14 - INFO - codeparrot_training - Step 22483: {'lr': 0.00030703992598487564, 'samples': 4316928, 'steps': 22483, 'loss/train': 1.6296265125274658} 01/29/2022 17:49:19 - INFO - codeparrot_training - Step 22484: {'lr': 0.00030702399498184005, 'samples': 4317120, 'steps': 22484, 'loss/train': 1.5735390186309814} 01/29/2022 17:49:25 - INFO - codeparrot_training - Step 22485: {'lr': 0.0003070080637345317, 'samples': 4317312, 'steps': 22485, 'loss/train': 1.9801311492919922} 01/29/2022 17:49:29 - INFO - codeparrot_training - Step 22486: {'lr': 0.00030699213224301896, 'samples': 4317504, 'steps': 22486, 'loss/train': 1.7641382217407227} 01/29/2022 17:49:33 - INFO - codeparrot_training - Step 22487: {'lr': 0.00030697620050737, 'samples': 4317696, 'steps': 22487, 'loss/train': 2.240931451320648} 01/29/2022 17:49:37 - INFO - codeparrot_training - Step 22488: {'lr': 0.00030696026852765313, 'samples': 4317888, 'steps': 22488, 'loss/train': 1.8874803185462952} 01/29/2022 17:49:41 - INFO - codeparrot_training - Step 22489: {'lr': 0.00030694433630393654, 'samples': 4318080, 'steps': 22489, 'loss/train': 0.9857945144176483} 01/29/2022 17:49:46 - INFO - codeparrot_training - Step 22490: {'lr': 0.00030692840383628845, 'samples': 4318272, 'steps': 22490, 'loss/train': 0.9573308229446411} 01/29/2022 17:49:51 - INFO - codeparrot_training - Step 22491: {'lr': 0.0003069124711247772, 'samples': 4318464, 'steps': 22491, 'loss/train': 0.8752251863479614} 01/29/2022 17:49:55 - INFO - codeparrot_training - Step 22492: {'lr': 0.000306896538169471, 'samples': 4318656, 'steps': 22492, 'loss/train': 2.9846282601356506} 01/29/2022 17:49:59 - INFO - codeparrot_training - Step 22493: {'lr': 0.000306880604970438, 'samples': 4318848, 'steps': 22493, 'loss/train': 1.91996169090271} 01/29/2022 17:50:04 - INFO - codeparrot_training - Step 22494: {'lr': 0.00030686467152774667, 'samples': 4319040, 'steps': 22494, 'loss/train': 1.4999444782733917} 01/29/2022 17:50:08 - INFO - codeparrot_training - Step 22495: {'lr': 0.0003068487378414651, 'samples': 4319232, 'steps': 22495, 'loss/train': 1.6021631956100464} 01/29/2022 17:50:15 - INFO - codeparrot_training - Step 22496: {'lr': 0.0003068328039116616, 'samples': 4319424, 'steps': 22496, 'loss/train': 1.8943807482719421} 01/29/2022 17:50:19 - INFO - codeparrot_training - Step 22497: {'lr': 0.0003068168697384044, 'samples': 4319616, 'steps': 22497, 'loss/train': 2.3626447319984436} 01/29/2022 17:50:24 - INFO - codeparrot_training - Step 22498: {'lr': 0.0003068009353217618, 'samples': 4319808, 'steps': 22498, 'loss/train': 1.657164216041565} 01/29/2022 17:50:28 - INFO - codeparrot_training - Step 22499: {'lr': 0.00030678500066180206, 'samples': 4320000, 'steps': 22499, 'loss/train': 1.6912046670913696} 01/29/2022 17:50:32 - INFO - codeparrot_training - Step 22500: {'lr': 0.0003067690657585933, 'samples': 4320192, 'steps': 22500, 'loss/train': 1.3313153386116028} 01/29/2022 17:50:37 - INFO - codeparrot_training - Step 22501: {'lr': 0.000306753130612204, 'samples': 4320384, 'steps': 22501, 'loss/train': 1.508746325969696} 01/29/2022 17:50:42 - INFO - codeparrot_training - Step 22502: {'lr': 0.0003067371952227022, 'samples': 4320576, 'steps': 22502, 'loss/train': 1.8611767888069153} 01/29/2022 17:50:46 - INFO - codeparrot_training - Step 22503: {'lr': 0.0003067212595901564, 'samples': 4320768, 'steps': 22503, 'loss/train': 1.7141515016555786} 01/29/2022 17:50:50 - INFO - codeparrot_training - Step 22504: {'lr': 0.00030670532371463463, 'samples': 4320960, 'steps': 22504, 'loss/train': 0.9104413390159607} 01/29/2022 17:50:54 - INFO - codeparrot_training - Step 22505: {'lr': 0.0003066893875962053, 'samples': 4321152, 'steps': 22505, 'loss/train': 2.8251532316207886} 01/29/2022 17:51:00 - INFO - codeparrot_training - Step 22506: {'lr': 0.0003066734512349366, 'samples': 4321344, 'steps': 22506, 'loss/train': 1.5102195739746094} 01/29/2022 17:51:04 - INFO - codeparrot_training - Step 22507: {'lr': 0.00030665751463089687, 'samples': 4321536, 'steps': 22507, 'loss/train': 1.6490201354026794} 01/29/2022 17:51:08 - INFO - codeparrot_training - Step 22508: {'lr': 0.0003066415777841543, 'samples': 4321728, 'steps': 22508, 'loss/train': 1.6245834231376648} 01/29/2022 17:51:12 - INFO - codeparrot_training - Step 22509: {'lr': 0.0003066256406947773, 'samples': 4321920, 'steps': 22509, 'loss/train': 1.3397253155708313} 01/29/2022 17:51:17 - INFO - codeparrot_training - Step 22510: {'lr': 0.0003066097033628339, 'samples': 4322112, 'steps': 22510, 'loss/train': 2.0956549644470215} 01/29/2022 17:51:22 - INFO - codeparrot_training - Step 22511: {'lr': 0.0003065937657883926, 'samples': 4322304, 'steps': 22511, 'loss/train': 1.8026633262634277} 01/29/2022 17:51:26 - INFO - codeparrot_training - Step 22512: {'lr': 0.0003065778279715215, 'samples': 4322496, 'steps': 22512, 'loss/train': 1.776664674282074} 01/29/2022 17:51:31 - INFO - codeparrot_training - Step 22513: {'lr': 0.000306561889912289, 'samples': 4322688, 'steps': 22513, 'loss/train': 1.276796668767929} 01/29/2022 17:51:35 - INFO - codeparrot_training - Step 22514: {'lr': 0.0003065459516107633, 'samples': 4322880, 'steps': 22514, 'loss/train': 1.8475749492645264} 01/29/2022 17:51:39 - INFO - codeparrot_training - Step 22515: {'lr': 0.0003065300130670127, 'samples': 4323072, 'steps': 22515, 'loss/train': 0.7460356950759888} 01/29/2022 17:51:46 - INFO - codeparrot_training - Step 22516: {'lr': 0.0003065140742811055, 'samples': 4323264, 'steps': 22516, 'loss/train': 1.7624480724334717} 01/29/2022 17:51:51 - INFO - codeparrot_training - Step 22517: {'lr': 0.0003064981352531099, 'samples': 4323456, 'steps': 22517, 'loss/train': 1.368649274110794} 01/29/2022 17:51:55 - INFO - codeparrot_training - Step 22518: {'lr': 0.0003064821959830943, 'samples': 4323648, 'steps': 22518, 'loss/train': 0.38379281759262085} 01/29/2022 17:51:59 - INFO - codeparrot_training - Step 22519: {'lr': 0.00030646625647112686, 'samples': 4323840, 'steps': 22519, 'loss/train': 0.8543847799301147} 01/29/2022 17:52:03 - INFO - codeparrot_training - Step 22520: {'lr': 0.000306450316717276, 'samples': 4324032, 'steps': 22520, 'loss/train': 1.516962468624115} 01/29/2022 17:52:09 - INFO - codeparrot_training - Step 22521: {'lr': 0.0003064343767216098, 'samples': 4324224, 'steps': 22521, 'loss/train': 1.367642194032669} 01/29/2022 17:52:13 - INFO - codeparrot_training - Step 22522: {'lr': 0.00030641843648419664, 'samples': 4324416, 'steps': 22522, 'loss/train': 1.8038250803947449} 01/29/2022 17:52:17 - INFO - codeparrot_training - Step 22523: {'lr': 0.0003064024960051048, 'samples': 4324608, 'steps': 22523, 'loss/train': 1.4777025282382965} 01/29/2022 17:52:22 - INFO - codeparrot_training - Step 22524: {'lr': 0.00030638655528440273, 'samples': 4324800, 'steps': 22524, 'loss/train': 1.5362305641174316} 01/29/2022 17:52:26 - INFO - codeparrot_training - Step 22525: {'lr': 0.0003063706143221584, 'samples': 4324992, 'steps': 22525, 'loss/train': 0.6055576354265213} 01/29/2022 17:52:33 - INFO - codeparrot_training - Step 22526: {'lr': 0.00030635467311844033, 'samples': 4325184, 'steps': 22526, 'loss/train': 1.963007390499115} 01/29/2022 17:52:37 - INFO - codeparrot_training - Step 22527: {'lr': 0.00030633873167331674, 'samples': 4325376, 'steps': 22527, 'loss/train': 1.479538768529892} 01/29/2022 17:52:41 - INFO - codeparrot_training - Step 22528: {'lr': 0.0003063227899868559, 'samples': 4325568, 'steps': 22528, 'loss/train': 1.7092180252075195} 01/29/2022 17:52:46 - INFO - codeparrot_training - Step 22529: {'lr': 0.00030630684805912613, 'samples': 4325760, 'steps': 22529, 'loss/train': 1.3496848046779633} 01/29/2022 17:52:50 - INFO - codeparrot_training - Step 22530: {'lr': 0.00030629090589019567, 'samples': 4325952, 'steps': 22530, 'loss/train': 2.225062608718872} 01/29/2022 17:52:55 - INFO - codeparrot_training - Step 22531: {'lr': 0.00030627496348013285, 'samples': 4326144, 'steps': 22531, 'loss/train': 2.9048418402671814} 01/29/2022 17:53:00 - INFO - codeparrot_training - Step 22532: {'lr': 0.0003062590208290059, 'samples': 4326336, 'steps': 22532, 'loss/train': 1.9783269166946411} 01/29/2022 17:53:04 - INFO - codeparrot_training - Step 22533: {'lr': 0.00030624307793688334, 'samples': 4326528, 'steps': 22533, 'loss/train': 1.9878484010696411} 01/29/2022 17:53:08 - INFO - codeparrot_training - Step 22534: {'lr': 0.00030622713480383314, 'samples': 4326720, 'steps': 22534, 'loss/train': 1.2239999771118164} 01/29/2022 17:53:12 - INFO - codeparrot_training - Step 22535: {'lr': 0.0003062111914299238, 'samples': 4326912, 'steps': 22535, 'loss/train': 1.7796590924263} 01/29/2022 17:53:18 - INFO - codeparrot_training - Step 22536: {'lr': 0.0003061952478152236, 'samples': 4327104, 'steps': 22536, 'loss/train': 1.6204179525375366} 01/29/2022 17:53:22 - INFO - codeparrot_training - Step 22537: {'lr': 0.0003061793039598007, 'samples': 4327296, 'steps': 22537, 'loss/train': 2.0070927143096924} 01/29/2022 17:53:26 - INFO - codeparrot_training - Step 22538: {'lr': 0.0003061633598637236, 'samples': 4327488, 'steps': 22538, 'loss/train': 0.029809992760419846} 01/29/2022 17:53:30 - INFO - codeparrot_training - Step 22539: {'lr': 0.00030614741552706045, 'samples': 4327680, 'steps': 22539, 'loss/train': 1.4179799258708954} 01/29/2022 17:53:35 - INFO - codeparrot_training - Step 22540: {'lr': 0.0003061314709498796, 'samples': 4327872, 'steps': 22540, 'loss/train': 1.3924202620983124} 01/29/2022 17:53:41 - INFO - codeparrot_training - Step 22541: {'lr': 0.00030611552613224935, 'samples': 4328064, 'steps': 22541, 'loss/train': 1.947918713092804} 01/29/2022 17:53:45 - INFO - codeparrot_training - Step 22542: {'lr': 0.00030609958107423804, 'samples': 4328256, 'steps': 22542, 'loss/train': 1.474412888288498} 01/29/2022 17:53:49 - INFO - codeparrot_training - Step 22543: {'lr': 0.0003060836357759139, 'samples': 4328448, 'steps': 22543, 'loss/train': 1.2721078097820282} 01/29/2022 17:53:53 - INFO - codeparrot_training - Step 22544: {'lr': 0.00030606769023734534, 'samples': 4328640, 'steps': 22544, 'loss/train': 1.742057740688324} 01/29/2022 17:53:58 - INFO - codeparrot_training - Step 22545: {'lr': 0.0003060517444586005, 'samples': 4328832, 'steps': 22545, 'loss/train': 1.5813894867897034} 01/29/2022 17:54:05 - INFO - codeparrot_training - Step 22546: {'lr': 0.0003060357984397479, 'samples': 4329024, 'steps': 22546, 'loss/train': 1.7311390042304993} 01/29/2022 17:54:09 - INFO - codeparrot_training - Step 22547: {'lr': 0.00030601985218085565, 'samples': 4329216, 'steps': 22547, 'loss/train': 1.296610951423645} 01/29/2022 17:54:14 - INFO - codeparrot_training - Step 22548: {'lr': 0.00030600390568199213, 'samples': 4329408, 'steps': 22548, 'loss/train': 0.9012918770313263} 01/29/2022 17:54:18 - INFO - codeparrot_training - Step 22549: {'lr': 0.00030598795894322574, 'samples': 4329600, 'steps': 22549, 'loss/train': 1.5588784217834473} 01/29/2022 17:54:22 - INFO - codeparrot_training - Step 22550: {'lr': 0.00030597201196462466, 'samples': 4329792, 'steps': 22550, 'loss/train': 1.9171850681304932} 01/29/2022 17:54:27 - INFO - codeparrot_training - Step 22551: {'lr': 0.00030595606474625726, 'samples': 4329984, 'steps': 22551, 'loss/train': 2.574224531650543} 01/29/2022 17:54:32 - INFO - codeparrot_training - Step 22552: {'lr': 0.00030594011728819184, 'samples': 4330176, 'steps': 22552, 'loss/train': 2.074785053730011} 01/29/2022 17:54:36 - INFO - codeparrot_training - Step 22553: {'lr': 0.00030592416959049666, 'samples': 4330368, 'steps': 22553, 'loss/train': 1.969417691230774} 01/29/2022 17:54:40 - INFO - codeparrot_training - Step 22554: {'lr': 0.00030590822165324017, 'samples': 4330560, 'steps': 22554, 'loss/train': 1.4495153725147247} 01/29/2022 17:54:44 - INFO - codeparrot_training - Step 22555: {'lr': 0.00030589227347649063, 'samples': 4330752, 'steps': 22555, 'loss/train': 2.2211485505104065} 01/29/2022 17:54:51 - INFO - codeparrot_training - Step 22556: {'lr': 0.00030587632506031624, 'samples': 4330944, 'steps': 22556, 'loss/train': 2.513738751411438} 01/29/2022 17:54:55 - INFO - codeparrot_training - Step 22557: {'lr': 0.0003058603764047855, 'samples': 4331136, 'steps': 22557, 'loss/train': 2.0188516974449158} 01/29/2022 17:55:00 - INFO - codeparrot_training - Step 22558: {'lr': 0.00030584442750996666, 'samples': 4331328, 'steps': 22558, 'loss/train': 2.053162693977356} 01/29/2022 17:55:04 - INFO - codeparrot_training - Step 22559: {'lr': 0.000305828478375928, 'samples': 4331520, 'steps': 22559, 'loss/train': 1.8646305799484253} 01/29/2022 17:55:08 - INFO - codeparrot_training - Step 22560: {'lr': 0.00030581252900273786, 'samples': 4331712, 'steps': 22560, 'loss/train': 1.633348524570465} 01/29/2022 17:55:14 - INFO - codeparrot_training - Step 22561: {'lr': 0.0003057965793904646, 'samples': 4331904, 'steps': 22561, 'loss/train': 1.4207072854042053} 01/29/2022 17:55:19 - INFO - codeparrot_training - Step 22562: {'lr': 0.00030578062953917645, 'samples': 4332096, 'steps': 22562, 'loss/train': 1.6239697337150574} 01/29/2022 17:55:23 - INFO - codeparrot_training - Step 22563: {'lr': 0.00030576467944894186, 'samples': 4332288, 'steps': 22563, 'loss/train': 1.4147182703018188} 01/29/2022 17:55:27 - INFO - codeparrot_training - Step 22564: {'lr': 0.000305748729119829, 'samples': 4332480, 'steps': 22564, 'loss/train': 1.7530642747879028} 01/29/2022 17:55:31 - INFO - codeparrot_training - Step 22565: {'lr': 0.00030573277855190634, 'samples': 4332672, 'steps': 22565, 'loss/train': 0.8965004682540894} 01/29/2022 17:55:36 - INFO - codeparrot_training - Step 22566: {'lr': 0.0003057168277452422, 'samples': 4332864, 'steps': 22566, 'loss/train': 1.691552460193634} 01/29/2022 17:55:41 - INFO - codeparrot_training - Step 22567: {'lr': 0.0003057008766999048, 'samples': 4333056, 'steps': 22567, 'loss/train': 2.2041388750076294} 01/29/2022 17:55:45 - INFO - codeparrot_training - Step 22568: {'lr': 0.0003056849254159625, 'samples': 4333248, 'steps': 22568, 'loss/train': 1.3536492884159088} 01/29/2022 17:55:50 - INFO - codeparrot_training - Step 22569: {'lr': 0.00030566897389348375, 'samples': 4333440, 'steps': 22569, 'loss/train': 1.8993836045265198} 01/29/2022 17:55:54 - INFO - codeparrot_training - Step 22570: {'lr': 0.00030565302213253677, 'samples': 4333632, 'steps': 22570, 'loss/train': 1.4931594729423523} 01/29/2022 17:56:01 - INFO - codeparrot_training - Step 22571: {'lr': 0.0003056370701331899, 'samples': 4333824, 'steps': 22571, 'loss/train': 2.6585533618927} 01/29/2022 17:56:05 - INFO - codeparrot_training - Step 22572: {'lr': 0.0003056211178955115, 'samples': 4334016, 'steps': 22572, 'loss/train': 0.9932528436183929} 01/29/2022 17:56:09 - INFO - codeparrot_training - Step 22573: {'lr': 0.00030560516541956983, 'samples': 4334208, 'steps': 22573, 'loss/train': 0.7322928607463837} 01/29/2022 17:56:14 - INFO - codeparrot_training - Step 22574: {'lr': 0.0003055892127054334, 'samples': 4334400, 'steps': 22574, 'loss/train': 1.8491452932357788} 01/29/2022 17:56:18 - INFO - codeparrot_training - Step 22575: {'lr': 0.00030557325975317037, 'samples': 4334592, 'steps': 22575, 'loss/train': 1.4021621346473694} 01/29/2022 17:56:23 - INFO - codeparrot_training - Step 22576: {'lr': 0.00030555730656284914, 'samples': 4334784, 'steps': 22576, 'loss/train': 1.2667794227600098} 01/29/2022 17:56:28 - INFO - codeparrot_training - Step 22577: {'lr': 0.000305541353134538, 'samples': 4334976, 'steps': 22577, 'loss/train': 1.5846553444862366} 01/29/2022 17:56:32 - INFO - codeparrot_training - Step 22578: {'lr': 0.0003055253994683054, 'samples': 4335168, 'steps': 22578, 'loss/train': 1.8019521832466125} 01/29/2022 17:56:36 - INFO - codeparrot_training - Step 22579: {'lr': 0.0003055094455642196, 'samples': 4335360, 'steps': 22579, 'loss/train': 1.7652212977409363} 01/29/2022 17:56:40 - INFO - codeparrot_training - Step 22580: {'lr': 0.000305493491422349, 'samples': 4335552, 'steps': 22580, 'loss/train': 0.8503713011741638} 01/29/2022 17:56:46 - INFO - codeparrot_training - Step 22581: {'lr': 0.0003054775370427619, 'samples': 4335744, 'steps': 22581, 'loss/train': 1.9921667575836182} 01/29/2022 17:56:50 - INFO - codeparrot_training - Step 22582: {'lr': 0.00030546158242552657, 'samples': 4335936, 'steps': 22582, 'loss/train': 1.3752292692661285} 01/29/2022 17:56:54 - INFO - codeparrot_training - Step 22583: {'lr': 0.00030544562757071154, 'samples': 4336128, 'steps': 22583, 'loss/train': 1.6832209825515747} 01/29/2022 17:56:58 - INFO - codeparrot_training - Step 22584: {'lr': 0.000305429672478385, 'samples': 4336320, 'steps': 22584, 'loss/train': 2.3017698526382446} 01/29/2022 17:57:02 - INFO - codeparrot_training - Step 22585: {'lr': 0.0003054137171486153, 'samples': 4336512, 'steps': 22585, 'loss/train': 1.1132575571537018} 01/29/2022 17:57:10 - INFO - codeparrot_training - Step 22586: {'lr': 0.0003053977615814709, 'samples': 4336704, 'steps': 22586, 'loss/train': 1.9218301177024841} 01/29/2022 17:57:14 - INFO - codeparrot_training - Step 22587: {'lr': 0.00030538180577702005, 'samples': 4336896, 'steps': 22587, 'loss/train': 2.005050837993622} 01/29/2022 17:57:18 - INFO - codeparrot_training - Step 22588: {'lr': 0.00030536584973533113, 'samples': 4337088, 'steps': 22588, 'loss/train': 1.9984524250030518} 01/29/2022 17:57:22 - INFO - codeparrot_training - Step 22589: {'lr': 0.0003053498934564725, 'samples': 4337280, 'steps': 22589, 'loss/train': 1.2944839596748352} 01/29/2022 17:57:26 - INFO - codeparrot_training - Step 22590: {'lr': 0.00030533393694051256, 'samples': 4337472, 'steps': 22590, 'loss/train': 1.8231110572814941} 01/29/2022 17:57:32 - INFO - codeparrot_training - Step 22591: {'lr': 0.0003053179801875195, 'samples': 4337664, 'steps': 22591, 'loss/train': 0.9996085166931152} 01/29/2022 17:57:36 - INFO - codeparrot_training - Step 22592: {'lr': 0.00030530202319756184, 'samples': 4337856, 'steps': 22592, 'loss/train': 1.9952545166015625} 01/29/2022 17:57:40 - INFO - codeparrot_training - Step 22593: {'lr': 0.0003052860659707078, 'samples': 4338048, 'steps': 22593, 'loss/train': 1.950823724269867} 01/29/2022 17:57:44 - INFO - codeparrot_training - Step 22594: {'lr': 0.0003052701085070259, 'samples': 4338240, 'steps': 22594, 'loss/train': 1.3987770080566406} 01/29/2022 17:57:49 - INFO - codeparrot_training - Step 22595: {'lr': 0.00030525415080658437, 'samples': 4338432, 'steps': 22595, 'loss/train': 2.1927865147590637} 01/29/2022 17:57:54 - INFO - codeparrot_training - Step 22596: {'lr': 0.0003052381928694516, 'samples': 4338624, 'steps': 22596, 'loss/train': 1.451328992843628} 01/29/2022 17:57:58 - INFO - codeparrot_training - Step 22597: {'lr': 0.00030522223469569594, 'samples': 4338816, 'steps': 22597, 'loss/train': 1.8796323537826538} 01/29/2022 17:58:02 - INFO - codeparrot_training - Step 22598: {'lr': 0.00030520627628538577, 'samples': 4339008, 'steps': 22598, 'loss/train': 2.1469831466674805} 01/29/2022 17:58:07 - INFO - codeparrot_training - Step 22599: {'lr': 0.0003051903176385894, 'samples': 4339200, 'steps': 22599, 'loss/train': 1.4557562470436096} 01/29/2022 17:58:11 - INFO - codeparrot_training - Step 22600: {'lr': 0.00030517435875537536, 'samples': 4339392, 'steps': 22600, 'loss/train': 1.2621381282806396} 01/29/2022 17:58:18 - INFO - codeparrot_training - Step 22601: {'lr': 0.0003051583996358118, 'samples': 4339584, 'steps': 22601, 'loss/train': 1.9712743163108826} 01/29/2022 17:58:22 - INFO - codeparrot_training - Step 22602: {'lr': 0.00030514244027996705, 'samples': 4339776, 'steps': 22602, 'loss/train': 1.5330138802528381} 01/29/2022 17:58:26 - INFO - codeparrot_training - Step 22603: {'lr': 0.00030512648068790985, 'samples': 4339968, 'steps': 22603, 'loss/train': 1.8916125297546387} 01/29/2022 17:58:30 - INFO - codeparrot_training - Step 22604: {'lr': 0.0003051105208597081, 'samples': 4340160, 'steps': 22604, 'loss/train': 1.6457176208496094} 01/29/2022 17:58:35 - INFO - codeparrot_training - Step 22605: {'lr': 0.00030509456079543044, 'samples': 4340352, 'steps': 22605, 'loss/train': 1.422050803899765} 01/29/2022 17:58:40 - INFO - codeparrot_training - Step 22606: {'lr': 0.0003050786004951452, 'samples': 4340544, 'steps': 22606, 'loss/train': 0.3679649233818054} 01/29/2022 17:58:44 - INFO - codeparrot_training - Step 22607: {'lr': 0.00030506263995892075, 'samples': 4340736, 'steps': 22607, 'loss/train': 2.257416844367981} 01/29/2022 17:58:48 - INFO - codeparrot_training - Step 22608: {'lr': 0.0003050466791868254, 'samples': 4340928, 'steps': 22608, 'loss/train': 1.317636787891388} 01/29/2022 17:58:53 - INFO - codeparrot_training - Step 22609: {'lr': 0.0003050307181789276, 'samples': 4341120, 'steps': 22609, 'loss/train': 2.2111347913742065} 01/29/2022 17:58:57 - INFO - codeparrot_training - Step 22610: {'lr': 0.0003050147569352956, 'samples': 4341312, 'steps': 22610, 'loss/train': 2.0741084218025208} 01/29/2022 17:59:02 - INFO - codeparrot_training - Step 22611: {'lr': 0.0003049987954559979, 'samples': 4341504, 'steps': 22611, 'loss/train': 1.4484303295612335} 01/29/2022 17:59:06 - INFO - codeparrot_training - Step 22612: {'lr': 0.0003049828337411028, 'samples': 4341696, 'steps': 22612, 'loss/train': 2.2598671317100525} 01/29/2022 17:59:10 - INFO - codeparrot_training - Step 22613: {'lr': 0.00030496687179067865, 'samples': 4341888, 'steps': 22613, 'loss/train': 1.9846315383911133} 01/29/2022 17:59:15 - INFO - codeparrot_training - Step 22614: {'lr': 0.000304950909604794, 'samples': 4342080, 'steps': 22614, 'loss/train': 1.234490990638733} 01/29/2022 17:59:19 - INFO - codeparrot_training - Step 22615: {'lr': 0.000304934947183517, 'samples': 4342272, 'steps': 22615, 'loss/train': 1.2129456102848053} 01/29/2022 17:59:26 - INFO - codeparrot_training - Step 22616: {'lr': 0.00030491898452691626, 'samples': 4342464, 'steps': 22616, 'loss/train': 1.4212449789047241} 01/29/2022 17:59:30 - INFO - codeparrot_training - Step 22617: {'lr': 0.0003049030216350599, 'samples': 4342656, 'steps': 22617, 'loss/train': 1.9088489413261414} 01/29/2022 17:59:34 - INFO - codeparrot_training - Step 22618: {'lr': 0.00030488705850801646, 'samples': 4342848, 'steps': 22618, 'loss/train': 2.5563583374023438} 01/29/2022 17:59:39 - INFO - codeparrot_training - Step 22619: {'lr': 0.00030487109514585426, 'samples': 4343040, 'steps': 22619, 'loss/train': 1.2687016725540161} 01/29/2022 17:59:43 - INFO - codeparrot_training - Step 22620: {'lr': 0.0003048551315486418, 'samples': 4343232, 'steps': 22620, 'loss/train': 1.4085635840892792} 01/29/2022 17:59:48 - INFO - codeparrot_training - Step 22621: {'lr': 0.00030483916771644734, 'samples': 4343424, 'steps': 22621, 'loss/train': 1.7604339122772217} 01/29/2022 17:59:53 - INFO - codeparrot_training - Step 22622: {'lr': 0.0003048232036493392, 'samples': 4343616, 'steps': 22622, 'loss/train': 1.391908586025238} 01/29/2022 17:59:57 - INFO - codeparrot_training - Step 22623: {'lr': 0.00030480723934738597, 'samples': 4343808, 'steps': 22623, 'loss/train': 2.3497663736343384} 01/29/2022 18:00:01 - INFO - codeparrot_training - Step 22624: {'lr': 0.00030479127481065595, 'samples': 4344000, 'steps': 22624, 'loss/train': 0.32307160645723343} 01/29/2022 18:00:05 - INFO - codeparrot_training - Step 22625: {'lr': 0.0003047753100392174, 'samples': 4344192, 'steps': 22625, 'loss/train': 0.5259219110012054} 01/29/2022 18:00:10 - INFO - codeparrot_training - Step 22626: {'lr': 0.00030475934503313893, 'samples': 4344384, 'steps': 22626, 'loss/train': 2.3010308146476746} 01/29/2022 18:00:15 - INFO - codeparrot_training - Step 22627: {'lr': 0.0003047433797924888, 'samples': 4344576, 'steps': 22627, 'loss/train': 1.9353511333465576} 01/29/2022 18:00:19 - INFO - codeparrot_training - Step 22628: {'lr': 0.00030472741431733535, 'samples': 4344768, 'steps': 22628, 'loss/train': 2.5447248816490173} 01/29/2022 18:00:23 - INFO - codeparrot_training - Step 22629: {'lr': 0.0003047114486077471, 'samples': 4344960, 'steps': 22629, 'loss/train': 2.856400966644287} 01/29/2022 18:00:27 - INFO - codeparrot_training - Step 22630: {'lr': 0.0003046954826637923, 'samples': 4345152, 'steps': 22630, 'loss/train': 1.3064830005168915} 01/29/2022 18:00:35 - INFO - codeparrot_training - Step 22631: {'lr': 0.0003046795164855395, 'samples': 4345344, 'steps': 22631, 'loss/train': 0.995220959186554} 01/29/2022 18:00:39 - INFO - codeparrot_training - Step 22632: {'lr': 0.000304663550073057, 'samples': 4345536, 'steps': 22632, 'loss/train': 1.7524884939193726} 01/29/2022 18:00:43 - INFO - codeparrot_training - Step 22633: {'lr': 0.00030464758342641315, 'samples': 4345728, 'steps': 22633, 'loss/train': 1.2462055385112762} 01/29/2022 18:00:47 - INFO - codeparrot_training - Step 22634: {'lr': 0.00030463161654567647, 'samples': 4345920, 'steps': 22634, 'loss/train': 1.8524560332298279} 01/29/2022 18:00:52 - INFO - codeparrot_training - Step 22635: {'lr': 0.00030461564943091524, 'samples': 4346112, 'steps': 22635, 'loss/train': 3.009069085121155} 01/29/2022 18:00:57 - INFO - codeparrot_training - Step 22636: {'lr': 0.0003045996820821979, 'samples': 4346304, 'steps': 22636, 'loss/train': 1.2293091416358948} 01/29/2022 18:01:01 - INFO - codeparrot_training - Step 22637: {'lr': 0.00030458371449959293, 'samples': 4346496, 'steps': 22637, 'loss/train': 0.9787772297859192} 01/29/2022 18:01:05 - INFO - codeparrot_training - Step 22638: {'lr': 0.00030456774668316864, 'samples': 4346688, 'steps': 22638, 'loss/train': 2.170802056789398} 01/29/2022 18:01:09 - INFO - codeparrot_training - Step 22639: {'lr': 0.0003045517786329934, 'samples': 4346880, 'steps': 22639, 'loss/train': 1.1713236272335052} 01/29/2022 18:01:14 - INFO - codeparrot_training - Step 22640: {'lr': 0.0003045358103491357, 'samples': 4347072, 'steps': 22640, 'loss/train': 1.7208958268165588} 01/29/2022 18:01:21 - INFO - codeparrot_training - Step 22641: {'lr': 0.00030451984183166384, 'samples': 4347264, 'steps': 22641, 'loss/train': 1.602948009967804} 01/29/2022 18:01:25 - INFO - codeparrot_training - Step 22642: {'lr': 0.0003045038730806464, 'samples': 4347456, 'steps': 22642, 'loss/train': 2.3609703183174133} 01/29/2022 18:01:29 - INFO - codeparrot_training - Step 22643: {'lr': 0.00030448790409615155, 'samples': 4347648, 'steps': 22643, 'loss/train': 1.9849843382835388} 01/29/2022 18:01:33 - INFO - codeparrot_training - Step 22644: {'lr': 0.00030447193487824796, 'samples': 4347840, 'steps': 22644, 'loss/train': 1.5695273280143738} 01/29/2022 18:01:37 - INFO - codeparrot_training - Step 22645: {'lr': 0.00030445596542700383, 'samples': 4348032, 'steps': 22645, 'loss/train': 1.5953527092933655} 01/29/2022 18:01:43 - INFO - codeparrot_training - Step 22646: {'lr': 0.00030443999574248764, 'samples': 4348224, 'steps': 22646, 'loss/train': 1.7006077766418457} 01/29/2022 18:01:47 - INFO - codeparrot_training - Step 22647: {'lr': 0.00030442402582476775, 'samples': 4348416, 'steps': 22647, 'loss/train': 1.3256417512893677} 01/29/2022 18:01:52 - INFO - codeparrot_training - Step 22648: {'lr': 0.00030440805567391274, 'samples': 4348608, 'steps': 22648, 'loss/train': 1.8449517488479614} 01/29/2022 18:01:56 - INFO - codeparrot_training - Step 22649: {'lr': 0.00030439208528999074, 'samples': 4348800, 'steps': 22649, 'loss/train': 2.6410857439041138} 01/29/2022 18:02:00 - INFO - codeparrot_training - Step 22650: {'lr': 0.0003043761146730704, 'samples': 4348992, 'steps': 22650, 'loss/train': 1.1950990855693817} 01/29/2022 18:02:04 - INFO - codeparrot_training - Step 22651: {'lr': 0.00030436014382322004, 'samples': 4349184, 'steps': 22651, 'loss/train': 2.3110309839248657} 01/29/2022 18:02:09 - INFO - codeparrot_training - Step 22652: {'lr': 0.00030434417274050805, 'samples': 4349376, 'steps': 22652, 'loss/train': 1.8466633558273315} 01/29/2022 18:02:14 - INFO - codeparrot_training - Step 22653: {'lr': 0.00030432820142500296, 'samples': 4349568, 'steps': 22653, 'loss/train': 1.9505801796913147} 01/29/2022 18:02:18 - INFO - codeparrot_training - Step 22654: {'lr': 0.00030431222987677305, 'samples': 4349760, 'steps': 22654, 'loss/train': 1.389021098613739} 01/29/2022 18:02:22 - INFO - codeparrot_training - Step 22655: {'lr': 0.0003042962580958868, 'samples': 4349952, 'steps': 22655, 'loss/train': 1.6259851455688477} 01/29/2022 18:02:27 - INFO - codeparrot_training - Step 22656: {'lr': 0.00030428028608241257, 'samples': 4350144, 'steps': 22656, 'loss/train': 1.5154681205749512} 01/29/2022 18:02:32 - INFO - codeparrot_training - Step 22657: {'lr': 0.00030426431383641893, 'samples': 4350336, 'steps': 22657, 'loss/train': 1.2798936367034912} 01/29/2022 18:02:36 - INFO - codeparrot_training - Step 22658: {'lr': 0.00030424834135797413, 'samples': 4350528, 'steps': 22658, 'loss/train': 1.3691999316215515} 01/29/2022 18:02:40 - INFO - codeparrot_training - Step 22659: {'lr': 0.00030423236864714676, 'samples': 4350720, 'steps': 22659, 'loss/train': 1.5329596996307373} 01/29/2022 18:02:44 - INFO - codeparrot_training - Step 22660: {'lr': 0.00030421639570400505, 'samples': 4350912, 'steps': 22660, 'loss/train': 1.892379641532898} 01/29/2022 18:02:51 - INFO - codeparrot_training - Step 22661: {'lr': 0.0003042004225286176, 'samples': 4351104, 'steps': 22661, 'loss/train': 2.069654703140259} 01/29/2022 18:02:56 - INFO - codeparrot_training - Step 22662: {'lr': 0.00030418444912105256, 'samples': 4351296, 'steps': 22662, 'loss/train': 2.0513160824775696} 01/29/2022 18:03:00 - INFO - codeparrot_training - Step 22663: {'lr': 0.0003041684754813787, 'samples': 4351488, 'steps': 22663, 'loss/train': 1.7221872210502625} 01/29/2022 18:03:04 - INFO - codeparrot_training - Step 22664: {'lr': 0.0003041525016096643, 'samples': 4351680, 'steps': 22664, 'loss/train': 1.647538661956787} 01/29/2022 18:03:08 - INFO - codeparrot_training - Step 22665: {'lr': 0.0003041365275059777, 'samples': 4351872, 'steps': 22665, 'loss/train': 2.309602975845337} 01/29/2022 18:03:12 - INFO - codeparrot_training - Step 22666: {'lr': 0.0003041205531703875, 'samples': 4352064, 'steps': 22666, 'loss/train': 1.45424285531044} 01/29/2022 18:03:18 - INFO - codeparrot_training - Step 22667: {'lr': 0.0003041045786029619, 'samples': 4352256, 'steps': 22667, 'loss/train': 2.542497217655182} 01/29/2022 18:03:22 - INFO - codeparrot_training - Step 22668: {'lr': 0.0003040886038037696, 'samples': 4352448, 'steps': 22668, 'loss/train': 1.9844672083854675} 01/29/2022 18:03:26 - INFO - codeparrot_training - Step 22669: {'lr': 0.00030407262877287883, 'samples': 4352640, 'steps': 22669, 'loss/train': 1.4924677312374115} 01/29/2022 18:03:31 - INFO - codeparrot_training - Step 22670: {'lr': 0.00030405665351035816, 'samples': 4352832, 'steps': 22670, 'loss/train': 1.7715550661087036} 01/29/2022 18:03:35 - INFO - codeparrot_training - Step 22671: {'lr': 0.0003040406780162759, 'samples': 4353024, 'steps': 22671, 'loss/train': 2.115138530731201} 01/29/2022 18:03:40 - INFO - codeparrot_training - Step 22672: {'lr': 0.00030402470229070054, 'samples': 4353216, 'steps': 22672, 'loss/train': 1.7915831208229065} 01/29/2022 18:03:45 - INFO - codeparrot_training - Step 22673: {'lr': 0.00030400872633370047, 'samples': 4353408, 'steps': 22673, 'loss/train': 1.029433786869049} 01/29/2022 18:03:49 - INFO - codeparrot_training - Step 22674: {'lr': 0.0003039927501453442, 'samples': 4353600, 'steps': 22674, 'loss/train': 1.8877118825912476} 01/29/2022 18:03:53 - INFO - codeparrot_training - Step 22675: {'lr': 0.0003039767737257002, 'samples': 4353792, 'steps': 22675, 'loss/train': 0.8239308893680573} 01/29/2022 18:03:57 - INFO - codeparrot_training - Step 22676: {'lr': 0.0003039607970748368, 'samples': 4353984, 'steps': 22676, 'loss/train': 2.4025379419326782} 01/29/2022 18:04:04 - INFO - codeparrot_training - Step 22677: {'lr': 0.00030394482019282246, 'samples': 4354176, 'steps': 22677, 'loss/train': 1.7011958956718445} 01/29/2022 18:04:09 - INFO - codeparrot_training - Step 22678: {'lr': 0.0003039288430797256, 'samples': 4354368, 'steps': 22678, 'loss/train': 1.6839097738265991} 01/29/2022 18:04:13 - INFO - codeparrot_training - Step 22679: {'lr': 0.0003039128657356147, 'samples': 4354560, 'steps': 22679, 'loss/train': 1.4638222754001617} 01/29/2022 18:04:17 - INFO - codeparrot_training - Step 22680: {'lr': 0.0003038968881605583, 'samples': 4354752, 'steps': 22680, 'loss/train': 1.6560405492782593} 01/29/2022 18:04:21 - INFO - codeparrot_training - Step 22681: {'lr': 0.00030388091035462466, 'samples': 4354944, 'steps': 22681, 'loss/train': 0.8733882308006287} 01/29/2022 18:04:27 - INFO - codeparrot_training - Step 22682: {'lr': 0.00030386493231788234, 'samples': 4355136, 'steps': 22682, 'loss/train': 2.0178637504577637} 01/29/2022 18:04:31 - INFO - codeparrot_training - Step 22683: {'lr': 0.0003038489540503997, 'samples': 4355328, 'steps': 22683, 'loss/train': 1.6941694021224976} 01/29/2022 18:04:35 - INFO - codeparrot_training - Step 22684: {'lr': 0.0003038329755522453, 'samples': 4355520, 'steps': 22684, 'loss/train': 0.8521456718444824} 01/29/2022 18:04:39 - INFO - codeparrot_training - Step 22685: {'lr': 0.0003038169968234875, 'samples': 4355712, 'steps': 22685, 'loss/train': 2.6067540049552917} 01/29/2022 18:04:43 - INFO - codeparrot_training - Step 22686: {'lr': 0.0003038010178641948, 'samples': 4355904, 'steps': 22686, 'loss/train': 1.5479949116706848} 01/29/2022 18:04:51 - INFO - codeparrot_training - Step 22687: {'lr': 0.00030378503867443555, 'samples': 4356096, 'steps': 22687, 'loss/train': 2.055567741394043} 01/29/2022 18:04:55 - INFO - codeparrot_training - Step 22688: {'lr': 0.0003037690592542784, 'samples': 4356288, 'steps': 22688, 'loss/train': 1.2559296190738678} 01/29/2022 18:04:59 - INFO - codeparrot_training - Step 22689: {'lr': 0.0003037530796037916, 'samples': 4356480, 'steps': 22689, 'loss/train': 1.4546558260917664} 01/29/2022 18:05:03 - INFO - codeparrot_training - Step 22690: {'lr': 0.0003037370997230436, 'samples': 4356672, 'steps': 22690, 'loss/train': 4.189022541046143} 01/29/2022 18:05:07 - INFO - codeparrot_training - Step 22691: {'lr': 0.000303721119612103, 'samples': 4356864, 'steps': 22691, 'loss/train': 2.484554171562195} 01/29/2022 18:05:13 - INFO - codeparrot_training - Step 22692: {'lr': 0.00030370513927103826, 'samples': 4357056, 'steps': 22692, 'loss/train': 0.9178165197372437} 01/29/2022 18:05:17 - INFO - codeparrot_training - Step 22693: {'lr': 0.0003036891586999176, 'samples': 4357248, 'steps': 22693, 'loss/train': 1.9032052159309387} 01/29/2022 18:05:21 - INFO - codeparrot_training - Step 22694: {'lr': 0.00030367317789880985, 'samples': 4357440, 'steps': 22694, 'loss/train': 1.4665686786174774} 01/29/2022 18:05:26 - INFO - codeparrot_training - Step 22695: {'lr': 0.000303657196867783, 'samples': 4357632, 'steps': 22695, 'loss/train': 1.399177998304367} 01/29/2022 18:05:30 - INFO - codeparrot_training - Step 22696: {'lr': 0.0003036412156069059, 'samples': 4357824, 'steps': 22696, 'loss/train': 0.7604847550392151} 01/29/2022 18:05:36 - INFO - codeparrot_training - Step 22697: {'lr': 0.00030362523411624686, 'samples': 4358016, 'steps': 22697, 'loss/train': 0.9895264506340027} 01/29/2022 18:05:40 - INFO - codeparrot_training - Step 22698: {'lr': 0.0003036092523958743, 'samples': 4358208, 'steps': 22698, 'loss/train': 2.0966153740882874} 01/29/2022 18:05:44 - INFO - codeparrot_training - Step 22699: {'lr': 0.00030359327044585685, 'samples': 4358400, 'steps': 22699, 'loss/train': 1.526306927204132} 01/29/2022 18:05:48 - INFO - codeparrot_training - Step 22700: {'lr': 0.00030357728826626266, 'samples': 4358592, 'steps': 22700, 'loss/train': 1.9600018858909607} 01/29/2022 18:05:55 - INFO - codeparrot_training - Step 22701: {'lr': 0.0003035613058571605, 'samples': 4358784, 'steps': 22701, 'loss/train': 1.3661752939224243} 01/29/2022 18:05:59 - INFO - codeparrot_training - Step 22702: {'lr': 0.00030354532321861865, 'samples': 4358976, 'steps': 22702, 'loss/train': 0.3968551754951477} 01/29/2022 18:06:03 - INFO - codeparrot_training - Step 22703: {'lr': 0.0003035293403507057, 'samples': 4359168, 'steps': 22703, 'loss/train': 1.915755271911621} 01/29/2022 18:06:07 - INFO - codeparrot_training - Step 22704: {'lr': 0.00030351335725349, 'samples': 4359360, 'steps': 22704, 'loss/train': 1.3465035259723663} 01/29/2022 18:06:12 - INFO - codeparrot_training - Step 22705: {'lr': 0.0003034973739270401, 'samples': 4359552, 'steps': 22705, 'loss/train': 2.231407642364502} 01/29/2022 18:06:17 - INFO - codeparrot_training - Step 22706: {'lr': 0.0003034813903714244, 'samples': 4359744, 'steps': 22706, 'loss/train': 1.1535758972167969} 01/29/2022 18:06:21 - INFO - codeparrot_training - Step 22707: {'lr': 0.0003034654065867115, 'samples': 4359936, 'steps': 22707, 'loss/train': 0.4374435245990753} 01/29/2022 18:06:25 - INFO - codeparrot_training - Step 22708: {'lr': 0.0003034494225729697, 'samples': 4360128, 'steps': 22708, 'loss/train': 1.6140959858894348} 01/29/2022 18:06:29 - INFO - codeparrot_training - Step 22709: {'lr': 0.0003034334383302676, 'samples': 4360320, 'steps': 22709, 'loss/train': 1.8658166527748108} 01/29/2022 18:06:34 - INFO - codeparrot_training - Step 22710: {'lr': 0.0003034174538586735, 'samples': 4360512, 'steps': 22710, 'loss/train': 1.9266820549964905} 01/29/2022 18:06:39 - INFO - codeparrot_training - Step 22711: {'lr': 0.00030340146915825605, 'samples': 4360704, 'steps': 22711, 'loss/train': 1.9317008256912231} 01/29/2022 18:06:43 - INFO - codeparrot_training - Step 22712: {'lr': 0.00030338548422908373, 'samples': 4360896, 'steps': 22712, 'loss/train': 1.4752547442913055} 01/29/2022 18:06:47 - INFO - codeparrot_training - Step 22713: {'lr': 0.00030336949907122483, 'samples': 4361088, 'steps': 22713, 'loss/train': 1.3171117901802063} 01/29/2022 18:06:52 - INFO - codeparrot_training - Step 22714: {'lr': 0.0003033535136847481, 'samples': 4361280, 'steps': 22714, 'loss/train': 0.7280515730381012} 01/29/2022 18:06:56 - INFO - codeparrot_training - Step 22715: {'lr': 0.0003033375280697218, 'samples': 4361472, 'steps': 22715, 'loss/train': 1.7159165740013123} 01/29/2022 18:07:01 - INFO - codeparrot_training - Step 22716: {'lr': 0.0003033215422262144, 'samples': 4361664, 'steps': 22716, 'loss/train': 1.1653330028057098} 01/29/2022 18:07:05 - INFO - codeparrot_training - Step 22717: {'lr': 0.0003033055561542945, 'samples': 4361856, 'steps': 22717, 'loss/train': 0.9497595727443695} 01/29/2022 18:07:09 - INFO - codeparrot_training - Step 22718: {'lr': 0.00030328956985403043, 'samples': 4362048, 'steps': 22718, 'loss/train': 1.3979547321796417} 01/29/2022 18:07:14 - INFO - codeparrot_training - Step 22719: {'lr': 0.0003032735833254909, 'samples': 4362240, 'steps': 22719, 'loss/train': 1.9712376594543457} 01/29/2022 18:07:18 - INFO - codeparrot_training - Step 22720: {'lr': 0.0003032575965687442, 'samples': 4362432, 'steps': 22720, 'loss/train': 2.0488414764404297} 01/29/2022 18:07:25 - INFO - codeparrot_training - Step 22721: {'lr': 0.0003032416095838588, 'samples': 4362624, 'steps': 22721, 'loss/train': 1.5909236669540405} 01/29/2022 18:07:30 - INFO - codeparrot_training - Step 22722: {'lr': 0.00030322562237090336, 'samples': 4362816, 'steps': 22722, 'loss/train': 1.1047427952289581} 01/29/2022 18:07:34 - INFO - codeparrot_training - Step 22723: {'lr': 0.00030320963492994616, 'samples': 4363008, 'steps': 22723, 'loss/train': 1.271057903766632} 01/29/2022 18:07:38 - INFO - codeparrot_training - Step 22724: {'lr': 0.00030319364726105584, 'samples': 4363200, 'steps': 22724, 'loss/train': 1.7241911888122559} 01/29/2022 18:07:42 - INFO - codeparrot_training - Step 22725: {'lr': 0.00030317765936430086, 'samples': 4363392, 'steps': 22725, 'loss/train': 1.9042607545852661} 01/29/2022 18:07:47 - INFO - codeparrot_training - Step 22726: {'lr': 0.0003031616712397496, 'samples': 4363584, 'steps': 22726, 'loss/train': 0.8329481184482574} 01/29/2022 18:07:52 - INFO - codeparrot_training - Step 22727: {'lr': 0.00030314568288747064, 'samples': 4363776, 'steps': 22727, 'loss/train': 2.1147741079330444} 01/29/2022 18:07:56 - INFO - codeparrot_training - Step 22728: {'lr': 0.00030312969430753244, 'samples': 4363968, 'steps': 22728, 'loss/train': 1.0707303285598755} 01/29/2022 18:08:00 - INFO - codeparrot_training - Step 22729: {'lr': 0.00030311370550000356, 'samples': 4364160, 'steps': 22729, 'loss/train': 1.3436425924301147} 01/29/2022 18:08:04 - INFO - codeparrot_training - Step 22730: {'lr': 0.0003030977164649523, 'samples': 4364352, 'steps': 22730, 'loss/train': 2.180356979370117} 01/29/2022 18:08:10 - INFO - codeparrot_training - Step 22731: {'lr': 0.0003030817272024474, 'samples': 4364544, 'steps': 22731, 'loss/train': 0.8277227282524109} 01/29/2022 18:08:14 - INFO - codeparrot_training - Step 22732: {'lr': 0.0003030657377125572, 'samples': 4364736, 'steps': 22732, 'loss/train': 1.5730479955673218} 01/29/2022 18:08:18 - INFO - codeparrot_training - Step 22733: {'lr': 0.0003030497479953503, 'samples': 4364928, 'steps': 22733, 'loss/train': 1.123435914516449} 01/29/2022 18:08:22 - INFO - codeparrot_training - Step 22734: {'lr': 0.00030303375805089503, 'samples': 4365120, 'steps': 22734, 'loss/train': 1.8944726586341858} 01/29/2022 18:08:27 - INFO - codeparrot_training - Step 22735: {'lr': 0.00030301776787925995, 'samples': 4365312, 'steps': 22735, 'loss/train': 2.150702476501465} 01/29/2022 18:08:34 - INFO - codeparrot_training - Step 22736: {'lr': 0.00030300177748051373, 'samples': 4365504, 'steps': 22736, 'loss/train': 1.6872096061706543} 01/29/2022 18:08:38 - INFO - codeparrot_training - Step 22737: {'lr': 0.0003029857868547246, 'samples': 4365696, 'steps': 22737, 'loss/train': 2.187905788421631} 01/29/2022 18:08:42 - INFO - codeparrot_training - Step 22738: {'lr': 0.0003029697960019612, 'samples': 4365888, 'steps': 22738, 'loss/train': 2.0301255583763123} 01/29/2022 18:08:46 - INFO - codeparrot_training - Step 22739: {'lr': 0.0003029538049222921, 'samples': 4366080, 'steps': 22739, 'loss/train': 1.8891257643699646} 01/29/2022 18:08:51 - INFO - codeparrot_training - Step 22740: {'lr': 0.00030293781361578567, 'samples': 4366272, 'steps': 22740, 'loss/train': 1.9368630051612854} 01/29/2022 18:08:56 - INFO - codeparrot_training - Step 22741: {'lr': 0.0003029218220825104, 'samples': 4366464, 'steps': 22741, 'loss/train': 2.4061295986175537} 01/29/2022 18:09:00 - INFO - codeparrot_training - Step 22742: {'lr': 0.00030290583032253486, 'samples': 4366656, 'steps': 22742, 'loss/train': 2.780810058116913} 01/29/2022 18:09:04 - INFO - codeparrot_training - Step 22743: {'lr': 0.00030288983833592757, 'samples': 4366848, 'steps': 22743, 'loss/train': 2.0040074586868286} 01/29/2022 18:09:09 - INFO - codeparrot_training - Step 22744: {'lr': 0.00030287384612275704, 'samples': 4367040, 'steps': 22744, 'loss/train': 1.382344901561737} 01/29/2022 18:09:13 - INFO - codeparrot_training - Step 22745: {'lr': 0.00030285785368309174, 'samples': 4367232, 'steps': 22745, 'loss/train': 1.4574239552021027} 01/29/2022 18:09:20 - INFO - codeparrot_training - Step 22746: {'lr': 0.0003028418610170001, 'samples': 4367424, 'steps': 22746, 'loss/train': 1.677512526512146} 01/29/2022 18:09:24 - INFO - codeparrot_training - Step 22747: {'lr': 0.00030282586812455076, 'samples': 4367616, 'steps': 22747, 'loss/train': 1.8202741742134094} 01/29/2022 18:09:28 - INFO - codeparrot_training - Step 22748: {'lr': 0.00030280987500581213, 'samples': 4367808, 'steps': 22748, 'loss/train': 2.5303276777267456} 01/29/2022 18:09:33 - INFO - codeparrot_training - Step 22749: {'lr': 0.00030279388166085287, 'samples': 4368000, 'steps': 22749, 'loss/train': 1.4555716216564178} 01/29/2022 18:09:37 - INFO - codeparrot_training - Step 22750: {'lr': 0.0003027778880897413, 'samples': 4368192, 'steps': 22750, 'loss/train': 1.3192435204982758} 01/29/2022 18:09:42 - INFO - codeparrot_training - Step 22751: {'lr': 0.000302761894292546, 'samples': 4368384, 'steps': 22751, 'loss/train': 0.8401957154273987} 01/29/2022 18:09:46 - INFO - codeparrot_training - Step 22752: {'lr': 0.0003027459002693356, 'samples': 4368576, 'steps': 22752, 'loss/train': 1.7822692394256592} 01/29/2022 18:09:51 - INFO - codeparrot_training - Step 22753: {'lr': 0.00030272990602017843, 'samples': 4368768, 'steps': 22753, 'loss/train': 1.6730920672416687} 01/29/2022 18:09:55 - INFO - codeparrot_training - Step 22754: {'lr': 0.0003027139115451431, 'samples': 4368960, 'steps': 22754, 'loss/train': 1.4833998084068298} 01/29/2022 18:09:59 - INFO - codeparrot_training - Step 22755: {'lr': 0.0003026979168442982, 'samples': 4369152, 'steps': 22755, 'loss/train': 1.79094797372818} 01/29/2022 18:10:04 - INFO - codeparrot_training - Step 22756: {'lr': 0.000302681921917712, 'samples': 4369344, 'steps': 22756, 'loss/train': 1.1190220713615417} 01/29/2022 18:10:08 - INFO - codeparrot_training - Step 22757: {'lr': 0.00030266592676545326, 'samples': 4369536, 'steps': 22757, 'loss/train': 2.3139299154281616} 01/29/2022 18:10:13 - INFO - codeparrot_training - Step 22758: {'lr': 0.0003026499313875903, 'samples': 4369728, 'steps': 22758, 'loss/train': 1.6533679962158203} 01/29/2022 18:10:17 - INFO - codeparrot_training - Step 22759: {'lr': 0.00030263393578419196, 'samples': 4369920, 'steps': 22759, 'loss/train': 1.7058168053627014} 01/29/2022 18:10:21 - INFO - codeparrot_training - Step 22760: {'lr': 0.0003026179399553264, 'samples': 4370112, 'steps': 22760, 'loss/train': 1.7461187839508057} 01/29/2022 18:10:28 - INFO - codeparrot_training - Step 22761: {'lr': 0.0003026019439010624, 'samples': 4370304, 'steps': 22761, 'loss/train': 0.9192829728126526} 01/29/2022 18:10:32 - INFO - codeparrot_training - Step 22762: {'lr': 0.0003025859476214683, 'samples': 4370496, 'steps': 22762, 'loss/train': 1.6669732332229614} 01/29/2022 18:10:37 - INFO - codeparrot_training - Step 22763: {'lr': 0.00030256995111661275, 'samples': 4370688, 'steps': 22763, 'loss/train': 1.9118657112121582} 01/29/2022 18:10:41 - INFO - codeparrot_training - Step 22764: {'lr': 0.0003025539543865642, 'samples': 4370880, 'steps': 22764, 'loss/train': 0.9710506796836853} 01/29/2022 18:10:45 - INFO - codeparrot_training - Step 22765: {'lr': 0.00030253795743139113, 'samples': 4371072, 'steps': 22765, 'loss/train': 1.6971184015274048} 01/29/2022 18:10:50 - INFO - codeparrot_training - Step 22766: {'lr': 0.0003025219602511622, 'samples': 4371264, 'steps': 22766, 'loss/train': 1.8292352557182312} 01/29/2022 18:10:55 - INFO - codeparrot_training - Step 22767: {'lr': 0.00030250596284594583, 'samples': 4371456, 'steps': 22767, 'loss/train': 1.5609031319618225} 01/29/2022 18:10:59 - INFO - codeparrot_training - Step 22768: {'lr': 0.0003024899652158107, 'samples': 4371648, 'steps': 22768, 'loss/train': 1.2597769796848297} 01/29/2022 18:11:03 - INFO - codeparrot_training - Step 22769: {'lr': 0.00030247396736082506, 'samples': 4371840, 'steps': 22769, 'loss/train': 1.7390357851982117} 01/29/2022 18:11:07 - INFO - codeparrot_training - Step 22770: {'lr': 0.0003024579692810577, 'samples': 4372032, 'steps': 22770, 'loss/train': 1.6391721367835999} 01/29/2022 18:11:13 - INFO - codeparrot_training - Step 22771: {'lr': 0.00030244197097657705, 'samples': 4372224, 'steps': 22771, 'loss/train': 2.0379050374031067} 01/29/2022 18:11:17 - INFO - codeparrot_training - Step 22772: {'lr': 0.0003024259724474516, 'samples': 4372416, 'steps': 22772, 'loss/train': 1.6789941787719727} 01/29/2022 18:11:21 - INFO - codeparrot_training - Step 22773: {'lr': 0.00030240997369375, 'samples': 4372608, 'steps': 22773, 'loss/train': 1.9578511118888855} 01/29/2022 18:11:25 - INFO - codeparrot_training - Step 22774: {'lr': 0.0003023939747155406, 'samples': 4372800, 'steps': 22774, 'loss/train': 2.55124694108963} 01/29/2022 18:11:29 - INFO - codeparrot_training - Step 22775: {'lr': 0.00030237797551289225, 'samples': 4372992, 'steps': 22775, 'loss/train': 2.1401272416114807} 01/29/2022 18:11:35 - INFO - codeparrot_training - Step 22776: {'lr': 0.0003023619760858731, 'samples': 4373184, 'steps': 22776, 'loss/train': 1.2365278601646423} 01/29/2022 18:11:39 - INFO - codeparrot_training - Step 22777: {'lr': 0.000302345976434552, 'samples': 4373376, 'steps': 22777, 'loss/train': 1.6188815832138062} 01/29/2022 18:11:43 - INFO - codeparrot_training - Step 22778: {'lr': 0.0003023299765589973, 'samples': 4373568, 'steps': 22778, 'loss/train': 1.409152328968048} 01/29/2022 18:11:47 - INFO - codeparrot_training - Step 22779: {'lr': 0.0003023139764592776, 'samples': 4373760, 'steps': 22779, 'loss/train': 1.8776968717575073} 01/29/2022 18:11:51 - INFO - codeparrot_training - Step 22780: {'lr': 0.0003022979761354614, 'samples': 4373952, 'steps': 22780, 'loss/train': 1.6018590331077576} 01/29/2022 18:11:59 - INFO - codeparrot_training - Step 22781: {'lr': 0.00030228197558761737, 'samples': 4374144, 'steps': 22781, 'loss/train': 1.7030276656150818} 01/29/2022 18:12:03 - INFO - codeparrot_training - Step 22782: {'lr': 0.00030226597481581387, 'samples': 4374336, 'steps': 22782, 'loss/train': 1.8905922174453735} 01/29/2022 18:12:07 - INFO - codeparrot_training - Step 22783: {'lr': 0.0003022499738201195, 'samples': 4374528, 'steps': 22783, 'loss/train': 1.4095516204833984} 01/29/2022 18:12:11 - INFO - codeparrot_training - Step 22784: {'lr': 0.0003022339726006029, 'samples': 4374720, 'steps': 22784, 'loss/train': 1.3552350103855133} 01/29/2022 18:12:15 - INFO - codeparrot_training - Step 22785: {'lr': 0.0003022179711573326, 'samples': 4374912, 'steps': 22785, 'loss/train': 1.4603311121463776} 01/29/2022 18:12:21 - INFO - codeparrot_training - Step 22786: {'lr': 0.000302201969490377, 'samples': 4375104, 'steps': 22786, 'loss/train': 1.8948265314102173} 01/29/2022 18:12:25 - INFO - codeparrot_training - Step 22787: {'lr': 0.00030218596759980476, 'samples': 4375296, 'steps': 22787, 'loss/train': 1.8659250140190125} 01/29/2022 18:12:29 - INFO - codeparrot_training - Step 22788: {'lr': 0.00030216996548568443, 'samples': 4375488, 'steps': 22788, 'loss/train': 1.5810002088546753} 01/29/2022 18:12:34 - INFO - codeparrot_training - Step 22789: {'lr': 0.0003021539631480845, 'samples': 4375680, 'steps': 22789, 'loss/train': 1.231415569782257} 01/29/2022 18:12:41 - INFO - codeparrot_training - Step 22790: {'lr': 0.0003021379605870736, 'samples': 4375872, 'steps': 22790, 'loss/train': 1.9083879590034485} 01/29/2022 18:12:45 - INFO - codeparrot_training - Step 22791: {'lr': 0.00030212195780272025, 'samples': 4376064, 'steps': 22791, 'loss/train': 0.31856435537338257} 01/29/2022 18:12:49 - INFO - codeparrot_training - Step 22792: {'lr': 0.000302105954795093, 'samples': 4376256, 'steps': 22792, 'loss/train': 0.5004651099443436} 01/29/2022 18:12:53 - INFO - codeparrot_training - Step 22793: {'lr': 0.00030208995156426024, 'samples': 4376448, 'steps': 22793, 'loss/train': 1.8137975335121155} 01/29/2022 18:12:58 - INFO - codeparrot_training - Step 22794: {'lr': 0.0003020739481102908, 'samples': 4376640, 'steps': 22794, 'loss/train': 1.3704683482646942} 01/29/2022 18:13:03 - INFO - codeparrot_training - Step 22795: {'lr': 0.00030205794443325296, 'samples': 4376832, 'steps': 22795, 'loss/train': 1.3576910197734833} 01/29/2022 18:13:07 - INFO - codeparrot_training - Step 22796: {'lr': 0.00030204194053321556, 'samples': 4377024, 'steps': 22796, 'loss/train': 1.5459160208702087} 01/29/2022 18:13:11 - INFO - codeparrot_training - Step 22797: {'lr': 0.00030202593641024696, 'samples': 4377216, 'steps': 22797, 'loss/train': 1.5351401567459106} 01/29/2022 18:13:15 - INFO - codeparrot_training - Step 22798: {'lr': 0.0003020099320644157, 'samples': 4377408, 'steps': 22798, 'loss/train': 4.13608717918396} 01/29/2022 18:13:20 - INFO - codeparrot_training - Step 22799: {'lr': 0.00030199392749579053, 'samples': 4377600, 'steps': 22799, 'loss/train': 1.297102689743042} 01/29/2022 18:13:25 - INFO - codeparrot_training - Step 22800: {'lr': 0.0003019779227044398, 'samples': 4377792, 'steps': 22800, 'loss/train': 1.0996491014957428} 01/29/2022 18:13:29 - INFO - codeparrot_training - Step 22801: {'lr': 0.0003019619176904322, 'samples': 4377984, 'steps': 22801, 'loss/train': 1.4743590652942657} 01/29/2022 18:13:33 - INFO - codeparrot_training - Step 22802: {'lr': 0.00030194591245383625, 'samples': 4378176, 'steps': 22802, 'loss/train': 2.5174614787101746} 01/29/2022 18:13:37 - INFO - codeparrot_training - Step 22803: {'lr': 0.00030192990699472053, 'samples': 4378368, 'steps': 22803, 'loss/train': 1.0262600183486938} 01/29/2022 18:13:42 - INFO - codeparrot_training - Step 22804: {'lr': 0.00030191390131315357, 'samples': 4378560, 'steps': 22804, 'loss/train': 1.8829777836799622} 01/29/2022 18:13:49 - INFO - codeparrot_training - Step 22805: {'lr': 0.0003018978954092039, 'samples': 4378752, 'steps': 22805, 'loss/train': 1.7279403805732727} 01/29/2022 18:13:53 - INFO - codeparrot_training - Step 22806: {'lr': 0.0003018818892829401, 'samples': 4378944, 'steps': 22806, 'loss/train': 1.3937139809131622} 01/29/2022 18:13:57 - INFO - codeparrot_training - Step 22807: {'lr': 0.00030186588293443077, 'samples': 4379136, 'steps': 22807, 'loss/train': 1.8231943845748901} 01/29/2022 18:14:01 - INFO - codeparrot_training - Step 22808: {'lr': 0.0003018498763637445, 'samples': 4379328, 'steps': 22808, 'loss/train': 1.8310550451278687} 01/29/2022 18:14:06 - INFO - codeparrot_training - Step 22809: {'lr': 0.0003018338695709498, 'samples': 4379520, 'steps': 22809, 'loss/train': 0.640752375125885} 01/29/2022 18:14:11 - INFO - codeparrot_training - Step 22810: {'lr': 0.0003018178625561153, 'samples': 4379712, 'steps': 22810, 'loss/train': 1.6814045906066895} 01/29/2022 18:14:15 - INFO - codeparrot_training - Step 22811: {'lr': 0.0003018018553193095, 'samples': 4379904, 'steps': 22811, 'loss/train': 1.8911942839622498} 01/29/2022 18:14:19 - INFO - codeparrot_training - Step 22812: {'lr': 0.00030178584786060106, 'samples': 4380096, 'steps': 22812, 'loss/train': 1.2076064050197601} 01/29/2022 18:14:24 - INFO - codeparrot_training - Step 22813: {'lr': 0.00030176984018005836, 'samples': 4380288, 'steps': 22813, 'loss/train': 2.244420289993286} 01/29/2022 18:14:28 - INFO - codeparrot_training - Step 22814: {'lr': 0.0003017538322777502, 'samples': 4380480, 'steps': 22814, 'loss/train': 1.3622834086418152} 01/29/2022 18:14:35 - INFO - codeparrot_training - Step 22815: {'lr': 0.00030173782415374503, 'samples': 4380672, 'steps': 22815, 'loss/train': 1.0713537633419037} 01/29/2022 18:14:39 - INFO - codeparrot_training - Step 22816: {'lr': 0.00030172181580811146, 'samples': 4380864, 'steps': 22816, 'loss/train': 1.9777650833129883} 01/29/2022 18:14:43 - INFO - codeparrot_training - Step 22817: {'lr': 0.0003017058072409181, 'samples': 4381056, 'steps': 22817, 'loss/train': 1.952803373336792} 01/29/2022 18:14:48 - INFO - codeparrot_training - Step 22818: {'lr': 0.0003016897984522334, 'samples': 4381248, 'steps': 22818, 'loss/train': 1.1806217730045319} 01/29/2022 18:14:52 - INFO - codeparrot_training - Step 22819: {'lr': 0.00030167378944212606, 'samples': 4381440, 'steps': 22819, 'loss/train': 1.0166111290454865} 01/29/2022 18:14:57 - INFO - codeparrot_training - Step 22820: {'lr': 0.0003016577802106645, 'samples': 4381632, 'steps': 22820, 'loss/train': 2.1523951292037964} 01/29/2022 18:15:01 - INFO - codeparrot_training - Step 22821: {'lr': 0.0003016417707579176, 'samples': 4381824, 'steps': 22821, 'loss/train': 1.309501200914383} 01/29/2022 18:15:05 - INFO - codeparrot_training - Step 22822: {'lr': 0.00030162576108395364, 'samples': 4382016, 'steps': 22822, 'loss/train': 2.0729949474334717} 01/29/2022 18:15:10 - INFO - codeparrot_training - Step 22823: {'lr': 0.0003016097511888414, 'samples': 4382208, 'steps': 22823, 'loss/train': 2.202359139919281} 01/29/2022 18:15:14 - INFO - codeparrot_training - Step 22824: {'lr': 0.0003015937410726493, 'samples': 4382400, 'steps': 22824, 'loss/train': 1.82895165681839} 01/29/2022 18:15:19 - INFO - codeparrot_training - Step 22825: {'lr': 0.000301577730735446, 'samples': 4382592, 'steps': 22825, 'loss/train': 1.6153437495231628} 01/29/2022 18:15:23 - INFO - codeparrot_training - Step 22826: {'lr': 0.00030156172017730006, 'samples': 4382784, 'steps': 22826, 'loss/train': 1.3816828429698944} 01/29/2022 18:15:27 - INFO - codeparrot_training - Step 22827: {'lr': 0.00030154570939828015, 'samples': 4382976, 'steps': 22827, 'loss/train': 0.12029841542243958} 01/29/2022 18:15:32 - INFO - codeparrot_training - Step 22828: {'lr': 0.0003015296983984547, 'samples': 4383168, 'steps': 22828, 'loss/train': 1.936715841293335} 01/29/2022 18:15:36 - INFO - codeparrot_training - Step 22829: {'lr': 0.00030151368717789244, 'samples': 4383360, 'steps': 22829, 'loss/train': 1.509410798549652} 01/29/2022 18:15:41 - INFO - codeparrot_training - Step 22830: {'lr': 0.0003014976757366619, 'samples': 4383552, 'steps': 22830, 'loss/train': 1.2906233668327332} 01/29/2022 18:15:46 - INFO - codeparrot_training - Step 22831: {'lr': 0.0003014816640748316, 'samples': 4383744, 'steps': 22831, 'loss/train': 0.45537570118904114} 01/29/2022 18:15:50 - INFO - codeparrot_training - Step 22832: {'lr': 0.00030146565219247033, 'samples': 4383936, 'steps': 22832, 'loss/train': 2.0401943922042847} 01/29/2022 18:15:54 - INFO - codeparrot_training - Step 22833: {'lr': 0.0003014496400896465, 'samples': 4384128, 'steps': 22833, 'loss/train': 2.1697239875793457} 01/29/2022 18:15:58 - INFO - codeparrot_training - Step 22834: {'lr': 0.0003014336277664287, 'samples': 4384320, 'steps': 22834, 'loss/train': 1.8260303735733032} 01/29/2022 18:16:05 - INFO - codeparrot_training - Step 22835: {'lr': 0.0003014176152228856, 'samples': 4384512, 'steps': 22835, 'loss/train': 0.5460586398839951} 01/29/2022 18:16:09 - INFO - codeparrot_training - Step 22836: {'lr': 0.00030140160245908584, 'samples': 4384704, 'steps': 22836, 'loss/train': 1.803138256072998} 01/29/2022 18:16:13 - INFO - codeparrot_training - Step 22837: {'lr': 0.0003013855894750978, 'samples': 4384896, 'steps': 22837, 'loss/train': 1.4124256074428558} 01/29/2022 18:16:17 - INFO - codeparrot_training - Step 22838: {'lr': 0.0003013695762709903, 'samples': 4385088, 'steps': 22838, 'loss/train': 1.141856074333191} 01/29/2022 18:16:22 - INFO - codeparrot_training - Step 22839: {'lr': 0.0003013535628468318, 'samples': 4385280, 'steps': 22839, 'loss/train': 1.673801064491272} 01/29/2022 18:16:27 - INFO - codeparrot_training - Step 22840: {'lr': 0.000301337549202691, 'samples': 4385472, 'steps': 22840, 'loss/train': 1.7270865440368652} 01/29/2022 18:16:32 - INFO - codeparrot_training - Step 22841: {'lr': 0.0003013215353386364, 'samples': 4385664, 'steps': 22841, 'loss/train': 1.0781209766864777} 01/29/2022 18:16:36 - INFO - codeparrot_training - Step 22842: {'lr': 0.00030130552125473667, 'samples': 4385856, 'steps': 22842, 'loss/train': 1.826519250869751} 01/29/2022 18:16:40 - INFO - codeparrot_training - Step 22843: {'lr': 0.00030128950695106034, 'samples': 4386048, 'steps': 22843, 'loss/train': 1.1155222356319427} 01/29/2022 18:16:44 - INFO - codeparrot_training - Step 22844: {'lr': 0.00030127349242767607, 'samples': 4386240, 'steps': 22844, 'loss/train': 1.4012559950351715} 01/29/2022 18:16:49 - INFO - codeparrot_training - Step 22845: {'lr': 0.0003012574776846524, 'samples': 4386432, 'steps': 22845, 'loss/train': 1.1858426928520203} 01/29/2022 18:16:54 - INFO - codeparrot_training - Step 22846: {'lr': 0.00030124146272205804, 'samples': 4386624, 'steps': 22846, 'loss/train': 1.4142619371414185} 01/29/2022 18:16:58 - INFO - codeparrot_training - Step 22847: {'lr': 0.00030122544753996143, 'samples': 4386816, 'steps': 22847, 'loss/train': 1.6412867903709412} 01/29/2022 18:17:02 - INFO - codeparrot_training - Step 22848: {'lr': 0.00030120943213843136, 'samples': 4387008, 'steps': 22848, 'loss/train': 1.6413654685020447} 01/29/2022 18:17:06 - INFO - codeparrot_training - Step 22849: {'lr': 0.0003011934165175363, 'samples': 4387200, 'steps': 22849, 'loss/train': 0.9747190475463867} 01/29/2022 18:17:13 - INFO - codeparrot_training - Step 22850: {'lr': 0.00030117740067734495, 'samples': 4387392, 'steps': 22850, 'loss/train': 0.5448332726955414} 01/29/2022 18:17:17 - INFO - codeparrot_training - Step 22851: {'lr': 0.0003011613846179258, 'samples': 4387584, 'steps': 22851, 'loss/train': 1.674925446510315} 01/29/2022 18:17:22 - INFO - codeparrot_training - Step 22852: {'lr': 0.0003011453683393476, 'samples': 4387776, 'steps': 22852, 'loss/train': 1.2593805491924286} 01/29/2022 18:17:26 - INFO - codeparrot_training - Step 22853: {'lr': 0.0003011293518416788, 'samples': 4387968, 'steps': 22853, 'loss/train': 1.5952216386795044} 01/29/2022 18:17:30 - INFO - codeparrot_training - Step 22854: {'lr': 0.00030111333512498813, 'samples': 4388160, 'steps': 22854, 'loss/train': 1.69060617685318} 01/29/2022 18:17:35 - INFO - codeparrot_training - Step 22855: {'lr': 0.00030109731818934413, 'samples': 4388352, 'steps': 22855, 'loss/train': 2.325329840183258} 01/29/2022 18:17:39 - INFO - codeparrot_training - Step 22856: {'lr': 0.00030108130103481554, 'samples': 4388544, 'steps': 22856, 'loss/train': 0.9819927513599396} 01/29/2022 18:17:44 - INFO - codeparrot_training - Step 22857: {'lr': 0.0003010652836614707, 'samples': 4388736, 'steps': 22857, 'loss/train': 1.9060855507850647} 01/29/2022 18:17:48 - INFO - codeparrot_training - Step 22858: {'lr': 0.00030104926606937856, 'samples': 4388928, 'steps': 22858, 'loss/train': 2.0269762873649597} 01/29/2022 18:17:52 - INFO - codeparrot_training - Step 22859: {'lr': 0.0003010332482586075, 'samples': 4389120, 'steps': 22859, 'loss/train': 0.5463093817234039} 01/29/2022 18:17:57 - INFO - codeparrot_training - Step 22860: {'lr': 0.0003010172302292263, 'samples': 4389312, 'steps': 22860, 'loss/train': 1.5653629302978516} 01/29/2022 18:18:01 - INFO - codeparrot_training - Step 22861: {'lr': 0.00030100121198130335, 'samples': 4389504, 'steps': 22861, 'loss/train': 2.037302255630493} 01/29/2022 18:18:06 - INFO - codeparrot_training - Step 22862: {'lr': 0.0003009851935149075, 'samples': 4389696, 'steps': 22862, 'loss/train': 0.9454594552516937} 01/29/2022 18:18:10 - INFO - codeparrot_training - Step 22863: {'lr': 0.0003009691748301072, 'samples': 4389888, 'steps': 22863, 'loss/train': 1.4355264902114868} 01/29/2022 18:18:14 - INFO - codeparrot_training - Step 22864: {'lr': 0.0003009531559269713, 'samples': 4390080, 'steps': 22864, 'loss/train': 1.7166073322296143} 01/29/2022 18:18:21 - INFO - codeparrot_training - Step 22865: {'lr': 0.00030093713680556805, 'samples': 4390272, 'steps': 22865, 'loss/train': 0.43563611805438995} 01/29/2022 18:18:26 - INFO - codeparrot_training - Step 22866: {'lr': 0.0003009211174659664, 'samples': 4390464, 'steps': 22866, 'loss/train': 1.112350881099701} 01/29/2022 18:18:30 - INFO - codeparrot_training - Step 22867: {'lr': 0.00030090509790823476, 'samples': 4390656, 'steps': 22867, 'loss/train': 1.817123293876648} 01/29/2022 18:18:34 - INFO - codeparrot_training - Step 22868: {'lr': 0.0003008890781324419, 'samples': 4390848, 'steps': 22868, 'loss/train': 1.3850375711917877} 01/29/2022 18:18:38 - INFO - codeparrot_training - Step 22869: {'lr': 0.0003008730581386564, 'samples': 4391040, 'steps': 22869, 'loss/train': 1.5528718829154968} 01/29/2022 18:18:44 - INFO - codeparrot_training - Step 22870: {'lr': 0.00030085703792694687, 'samples': 4391232, 'steps': 22870, 'loss/train': 2.2865429520606995} 01/29/2022 18:18:48 - INFO - codeparrot_training - Step 22871: {'lr': 0.00030084101749738195, 'samples': 4391424, 'steps': 22871, 'loss/train': 1.5292319655418396} 01/29/2022 18:18:52 - INFO - codeparrot_training - Step 22872: {'lr': 0.00030082499685003025, 'samples': 4391616, 'steps': 22872, 'loss/train': 1.5616831183433533} 01/29/2022 18:18:56 - INFO - codeparrot_training - Step 22873: {'lr': 0.0003008089759849604, 'samples': 4391808, 'steps': 22873, 'loss/train': 0.3576483204960823} 01/29/2022 18:19:00 - INFO - codeparrot_training - Step 22874: {'lr': 0.000300792954902241, 'samples': 4392000, 'steps': 22874, 'loss/train': 1.2270965874195099} 01/29/2022 18:19:08 - INFO - codeparrot_training - Step 22875: {'lr': 0.00030077693360194076, 'samples': 4392192, 'steps': 22875, 'loss/train': 1.4320040345191956} 01/29/2022 18:19:12 - INFO - codeparrot_training - Step 22876: {'lr': 0.0003007609120841282, 'samples': 4392384, 'steps': 22876, 'loss/train': 1.3605059087276459} 01/29/2022 18:19:16 - INFO - codeparrot_training - Step 22877: {'lr': 0.0003007448903488721, 'samples': 4392576, 'steps': 22877, 'loss/train': 1.3770816028118134} 01/29/2022 18:19:21 - INFO - codeparrot_training - Step 22878: {'lr': 0.00030072886839624093, 'samples': 4392768, 'steps': 22878, 'loss/train': 1.3724685609340668} 01/29/2022 18:19:25 - INFO - codeparrot_training - Step 22879: {'lr': 0.0003007128462263034, 'samples': 4392960, 'steps': 22879, 'loss/train': 1.7986995577812195} 01/29/2022 18:19:30 - INFO - codeparrot_training - Step 22880: {'lr': 0.0003006968238391281, 'samples': 4393152, 'steps': 22880, 'loss/train': 1.3728920817375183} 01/29/2022 18:19:34 - INFO - codeparrot_training - Step 22881: {'lr': 0.00030068080123478376, 'samples': 4393344, 'steps': 22881, 'loss/train': 0.7580476999282837} 01/29/2022 18:19:38 - INFO - codeparrot_training - Step 22882: {'lr': 0.000300664778413339, 'samples': 4393536, 'steps': 22882, 'loss/train': 1.8079531788825989} 01/29/2022 18:19:43 - INFO - codeparrot_training - Step 22883: {'lr': 0.00030064875537486236, 'samples': 4393728, 'steps': 22883, 'loss/train': 2.0214276909828186} 01/29/2022 18:19:47 - INFO - codeparrot_training - Step 22884: {'lr': 0.00030063273211942254, 'samples': 4393920, 'steps': 22884, 'loss/train': 1.3957232236862183} 01/29/2022 18:19:52 - INFO - codeparrot_training - Step 22885: {'lr': 0.0003006167086470882, 'samples': 4394112, 'steps': 22885, 'loss/train': 2.091041922569275} 01/29/2022 18:19:56 - INFO - codeparrot_training - Step 22886: {'lr': 0.00030060068495792793, 'samples': 4394304, 'steps': 22886, 'loss/train': 1.520008385181427} 01/29/2022 18:20:01 - INFO - codeparrot_training - Step 22887: {'lr': 0.0003005846610520104, 'samples': 4394496, 'steps': 22887, 'loss/train': 1.8482022285461426} 01/29/2022 18:20:05 - INFO - codeparrot_training - Step 22888: {'lr': 0.00030056863692940426, 'samples': 4394688, 'steps': 22888, 'loss/train': 1.130009651184082} 01/29/2022 18:20:09 - INFO - codeparrot_training - Step 22889: {'lr': 0.00030055261259017807, 'samples': 4394880, 'steps': 22889, 'loss/train': 1.6096422672271729} 01/29/2022 18:20:14 - INFO - codeparrot_training - Step 22890: {'lr': 0.00030053658803440064, 'samples': 4395072, 'steps': 22890, 'loss/train': 1.81514310836792} 01/29/2022 18:20:18 - INFO - codeparrot_training - Step 22891: {'lr': 0.00030052056326214046, 'samples': 4395264, 'steps': 22891, 'loss/train': 1.5317895412445068} 01/29/2022 18:20:23 - INFO - codeparrot_training - Step 22892: {'lr': 0.00030050453827346627, 'samples': 4395456, 'steps': 22892, 'loss/train': 1.9805434942245483} 01/29/2022 18:20:27 - INFO - codeparrot_training - Step 22893: {'lr': 0.0003004885130684467, 'samples': 4395648, 'steps': 22893, 'loss/train': 1.3283092081546783} 01/29/2022 18:20:31 - INFO - codeparrot_training - Step 22894: {'lr': 0.00030047248764715023, 'samples': 4395840, 'steps': 22894, 'loss/train': 1.4930495917797089} 01/29/2022 18:20:37 - INFO - codeparrot_training - Step 22895: {'lr': 0.0003004564620096457, 'samples': 4396032, 'steps': 22895, 'loss/train': 2.0945271849632263} 01/29/2022 18:20:42 - INFO - codeparrot_training - Step 22896: {'lr': 0.00030044043615600174, 'samples': 4396224, 'steps': 22896, 'loss/train': 1.868647813796997} 01/29/2022 18:20:46 - INFO - codeparrot_training - Step 22897: {'lr': 0.000300424410086287, 'samples': 4396416, 'steps': 22897, 'loss/train': 0.6828644424676895} 01/29/2022 18:20:50 - INFO - codeparrot_training - Step 22898: {'lr': 0.00030040838380057005, 'samples': 4396608, 'steps': 22898, 'loss/train': 1.6245787739753723} 01/29/2022 18:20:54 - INFO - codeparrot_training - Step 22899: {'lr': 0.00030039235729891964, 'samples': 4396800, 'steps': 22899, 'loss/train': 0.88378044962883} 01/29/2022 18:21:00 - INFO - codeparrot_training - Step 22900: {'lr': 0.0003003763305814043, 'samples': 4396992, 'steps': 22900, 'loss/train': 1.8239076733589172} 01/29/2022 18:21:04 - INFO - codeparrot_training - Step 22901: {'lr': 0.00030036030364809284, 'samples': 4397184, 'steps': 22901, 'loss/train': 1.773227334022522} 01/29/2022 18:21:08 - INFO - codeparrot_training - Step 22902: {'lr': 0.00030034427649905377, 'samples': 4397376, 'steps': 22902, 'loss/train': 1.754398226737976} 01/29/2022 18:21:12 - INFO - codeparrot_training - Step 22903: {'lr': 0.0003003282491343559, 'samples': 4397568, 'steps': 22903, 'loss/train': 2.2365134954452515} 01/29/2022 18:21:20 - INFO - codeparrot_training - Step 22904: {'lr': 0.00030031222155406763, 'samples': 4397760, 'steps': 22904, 'loss/train': 0.7728082537651062} 01/29/2022 18:21:24 - INFO - codeparrot_training - Step 22905: {'lr': 0.00030029619375825784, 'samples': 4397952, 'steps': 22905, 'loss/train': 1.6769290566444397} 01/29/2022 18:21:28 - INFO - codeparrot_training - Step 22906: {'lr': 0.00030028016574699517, 'samples': 4398144, 'steps': 22906, 'loss/train': 1.7948432564735413} 01/29/2022 18:21:33 - INFO - codeparrot_training - Step 22907: {'lr': 0.0003002641375203482, 'samples': 4398336, 'steps': 22907, 'loss/train': 1.1459676325321198} 01/29/2022 18:21:37 - INFO - codeparrot_training - Step 22908: {'lr': 0.0003002481090783856, 'samples': 4398528, 'steps': 22908, 'loss/train': 1.7443251013755798} 01/29/2022 18:21:41 - INFO - codeparrot_training - Step 22909: {'lr': 0.0003002320804211761, 'samples': 4398720, 'steps': 22909, 'loss/train': 0.48318345844745636} 01/29/2022 18:21:46 - INFO - codeparrot_training - Step 22910: {'lr': 0.00030021605154878836, 'samples': 4398912, 'steps': 22910, 'loss/train': 0.28127557039260864} 01/29/2022 18:21:51 - INFO - codeparrot_training - Step 22911: {'lr': 0.0003002000224612909, 'samples': 4399104, 'steps': 22911, 'loss/train': 1.9011738896369934} 01/29/2022 18:21:55 - INFO - codeparrot_training - Step 22912: {'lr': 0.0003001839931587526, 'samples': 4399296, 'steps': 22912, 'loss/train': 1.5495071411132812} 01/29/2022 18:21:59 - INFO - codeparrot_training - Step 22913: {'lr': 0.0003001679636412419, 'samples': 4399488, 'steps': 22913, 'loss/train': 1.8258352875709534} 01/29/2022 18:22:04 - INFO - codeparrot_training - Step 22914: {'lr': 0.0003001519339088277, 'samples': 4399680, 'steps': 22914, 'loss/train': 1.5052485466003418} 01/29/2022 18:22:09 - INFO - codeparrot_training - Step 22915: {'lr': 0.00030013590396157843, 'samples': 4399872, 'steps': 22915, 'loss/train': 1.5976653099060059} 01/29/2022 18:22:13 - INFO - codeparrot_training - Step 22916: {'lr': 0.0003001198737995628, 'samples': 4400064, 'steps': 22916, 'loss/train': 2.8728012442588806} 01/29/2022 18:22:17 - INFO - codeparrot_training - Step 22917: {'lr': 0.0003001038434228497, 'samples': 4400256, 'steps': 22917, 'loss/train': 2.209022283554077} 01/29/2022 18:22:21 - INFO - codeparrot_training - Step 22918: {'lr': 0.00030008781283150755, 'samples': 4400448, 'steps': 22918, 'loss/train': 1.693701982498169} 01/29/2022 18:22:29 - INFO - codeparrot_training - Step 22919: {'lr': 0.0003000717820256052, 'samples': 4400640, 'steps': 22919, 'loss/train': 1.3219149112701416} 01/29/2022 18:22:33 - INFO - codeparrot_training - Step 22920: {'lr': 0.00030005575100521117, 'samples': 4400832, 'steps': 22920, 'loss/train': 1.7044764161109924} 01/29/2022 18:22:38 - INFO - codeparrot_training - Step 22921: {'lr': 0.0003000397197703942, 'samples': 4401024, 'steps': 22921, 'loss/train': 2.485175371170044} 01/29/2022 18:22:42 - INFO - codeparrot_training - Step 22922: {'lr': 0.00030002368832122295, 'samples': 4401216, 'steps': 22922, 'loss/train': 1.2345252335071564} 01/29/2022 18:22:46 - INFO - codeparrot_training - Step 22923: {'lr': 0.00030000765665776617, 'samples': 4401408, 'steps': 22923, 'loss/train': 2.157676041126251} 01/29/2022 18:22:51 - INFO - codeparrot_training - Step 22924: {'lr': 0.0002999916247800924, 'samples': 4401600, 'steps': 22924, 'loss/train': 1.7933852076530457} 01/29/2022 18:22:55 - INFO - codeparrot_training - Step 22925: {'lr': 0.00029997559268827044, 'samples': 4401792, 'steps': 22925, 'loss/train': 1.7508324980735779} 01/29/2022 18:23:00 - INFO - codeparrot_training - Step 22926: {'lr': 0.0002999595603823689, 'samples': 4401984, 'steps': 22926, 'loss/train': 1.3589553236961365} 01/29/2022 18:23:04 - INFO - codeparrot_training - Step 22927: {'lr': 0.00029994352786245643, 'samples': 4402176, 'steps': 22927, 'loss/train': 0.9397151470184326} 01/29/2022 18:23:08 - INFO - codeparrot_training - Step 22928: {'lr': 0.0002999274951286017, 'samples': 4402368, 'steps': 22928, 'loss/train': 1.820344090461731} 01/29/2022 18:23:13 - INFO - codeparrot_training - Step 22929: {'lr': 0.0002999114621808735, 'samples': 4402560, 'steps': 22929, 'loss/train': 1.874934732913971} 01/29/2022 18:23:17 - INFO - codeparrot_training - Step 22930: {'lr': 0.0002998954290193405, 'samples': 4402752, 'steps': 22930, 'loss/train': 0.8018859028816223} 01/29/2022 18:23:22 - INFO - codeparrot_training - Step 22931: {'lr': 0.00029987939564407124, 'samples': 4402944, 'steps': 22931, 'loss/train': 1.581015408039093} 01/29/2022 18:23:26 - INFO - codeparrot_training - Step 22932: {'lr': 0.00029986336205513456, 'samples': 4403136, 'steps': 22932, 'loss/train': 0.8631349503993988} 01/29/2022 18:23:30 - INFO - codeparrot_training - Step 22933: {'lr': 0.00029984732825259904, 'samples': 4403328, 'steps': 22933, 'loss/train': 1.6551636457443237} 01/29/2022 18:23:37 - INFO - codeparrot_training - Step 22934: {'lr': 0.00029983129423653333, 'samples': 4403520, 'steps': 22934, 'loss/train': 1.7456670999526978} 01/29/2022 18:23:41 - INFO - codeparrot_training - Step 22935: {'lr': 0.00029981526000700626, 'samples': 4403712, 'steps': 22935, 'loss/train': 0.8519167900085449} 01/29/2022 18:23:46 - INFO - codeparrot_training - Step 22936: {'lr': 0.0002997992255640864, 'samples': 4403904, 'steps': 22936, 'loss/train': 1.9465434551239014} 01/29/2022 18:23:50 - INFO - codeparrot_training - Step 22937: {'lr': 0.0002997831909078425, 'samples': 4404096, 'steps': 22937, 'loss/train': 1.8705162405967712} 01/29/2022 18:23:54 - INFO - codeparrot_training - Step 22938: {'lr': 0.00029976715603834315, 'samples': 4404288, 'steps': 22938, 'loss/train': 1.7924251556396484} 01/29/2022 18:24:00 - INFO - codeparrot_training - Step 22939: {'lr': 0.00029975112095565723, 'samples': 4404480, 'steps': 22939, 'loss/train': 0.8445884585380554} 01/29/2022 18:24:04 - INFO - codeparrot_training - Step 22940: {'lr': 0.00029973508565985316, 'samples': 4404672, 'steps': 22940, 'loss/train': 2.1401864290237427} 01/29/2022 18:24:08 - INFO - codeparrot_training - Step 22941: {'lr': 0.0002997190501509999, 'samples': 4404864, 'steps': 22941, 'loss/train': 2.229790985584259} 01/29/2022 18:24:12 - INFO - codeparrot_training - Step 22942: {'lr': 0.00029970301442916594, 'samples': 4405056, 'steps': 22942, 'loss/train': 1.0542238354682922} 01/29/2022 18:24:17 - INFO - codeparrot_training - Step 22943: {'lr': 0.00029968697849442006, 'samples': 4405248, 'steps': 22943, 'loss/train': 1.2307583391666412} 01/29/2022 18:24:22 - INFO - codeparrot_training - Step 22944: {'lr': 0.0002996709423468309, 'samples': 4405440, 'steps': 22944, 'loss/train': 1.6441705226898193} 01/29/2022 18:24:26 - INFO - codeparrot_training - Step 22945: {'lr': 0.00029965490598646727, 'samples': 4405632, 'steps': 22945, 'loss/train': 2.2551414370536804} 01/29/2022 18:24:30 - INFO - codeparrot_training - Step 22946: {'lr': 0.00029963886941339774, 'samples': 4405824, 'steps': 22946, 'loss/train': 1.3485913574695587} 01/29/2022 18:24:35 - INFO - codeparrot_training - Step 22947: {'lr': 0.0002996228326276911, 'samples': 4406016, 'steps': 22947, 'loss/train': 1.8700127005577087} 01/29/2022 18:24:39 - INFO - codeparrot_training - Step 22948: {'lr': 0.0002996067956294159, 'samples': 4406208, 'steps': 22948, 'loss/train': 0.9711973965167999} 01/29/2022 18:24:44 - INFO - codeparrot_training - Step 22949: {'lr': 0.0002995907584186411, 'samples': 4406400, 'steps': 22949, 'loss/train': 3.2543052434921265} 01/29/2022 18:24:48 - INFO - codeparrot_training - Step 22950: {'lr': 0.00029957472099543516, 'samples': 4406592, 'steps': 22950, 'loss/train': 1.555031955242157} 01/29/2022 18:24:52 - INFO - codeparrot_training - Step 22951: {'lr': 0.00029955868335986686, 'samples': 4406784, 'steps': 22951, 'loss/train': 1.9286640286445618} 01/29/2022 18:24:57 - INFO - codeparrot_training - Step 22952: {'lr': 0.0002995426455120049, 'samples': 4406976, 'steps': 22952, 'loss/train': 1.962409257888794} 01/29/2022 18:25:01 - INFO - codeparrot_training - Step 22953: {'lr': 0.0002995266074519179, 'samples': 4407168, 'steps': 22953, 'loss/train': 1.0626579523086548} 01/29/2022 18:25:08 - INFO - codeparrot_training - Step 22954: {'lr': 0.00029951056917967476, 'samples': 4407360, 'steps': 22954, 'loss/train': 1.4268065392971039} 01/29/2022 18:25:12 - INFO - codeparrot_training - Step 22955: {'lr': 0.000299494530695344, 'samples': 4407552, 'steps': 22955, 'loss/train': 0.8782104849815369} 01/29/2022 18:25:16 - INFO - codeparrot_training - Step 22956: {'lr': 0.0002994784919989944, 'samples': 4407744, 'steps': 22956, 'loss/train': 0.9012706875801086} 01/29/2022 18:25:20 - INFO - codeparrot_training - Step 22957: {'lr': 0.00029946245309069464, 'samples': 4407936, 'steps': 22957, 'loss/train': 2.026245653629303} 01/29/2022 18:25:25 - INFO - codeparrot_training - Step 22958: {'lr': 0.0002994464139705135, 'samples': 4408128, 'steps': 22958, 'loss/train': 2.137509047985077} 01/29/2022 18:25:30 - INFO - codeparrot_training - Step 22959: {'lr': 0.00029943037463851953, 'samples': 4408320, 'steps': 22959, 'loss/train': 2.0421212911605835} 01/29/2022 18:25:34 - INFO - codeparrot_training - Step 22960: {'lr': 0.00029941433509478153, 'samples': 4408512, 'steps': 22960, 'loss/train': 0.18693409860134125} 01/29/2022 18:25:38 - INFO - codeparrot_training - Step 22961: {'lr': 0.00029939829533936823, 'samples': 4408704, 'steps': 22961, 'loss/train': 2.1646254658699036} 01/29/2022 18:25:42 - INFO - codeparrot_training - Step 22962: {'lr': 0.00029938225537234843, 'samples': 4408896, 'steps': 22962, 'loss/train': 1.7223967909812927} 01/29/2022 18:25:47 - INFO - codeparrot_training - Step 22963: {'lr': 0.0002993662151937906, 'samples': 4409088, 'steps': 22963, 'loss/train': 1.6265230178833008} 01/29/2022 18:25:54 - INFO - codeparrot_training - Step 22964: {'lr': 0.00029935017480376357, 'samples': 4409280, 'steps': 22964, 'loss/train': 2.0113776326179504} 01/29/2022 18:25:58 - INFO - codeparrot_training - Step 22965: {'lr': 0.00029933413420233615, 'samples': 4409472, 'steps': 22965, 'loss/train': 1.1648696064949036} 01/29/2022 18:26:02 - INFO - codeparrot_training - Step 22966: {'lr': 0.0002993180933895769, 'samples': 4409664, 'steps': 22966, 'loss/train': 1.8579869270324707} 01/29/2022 18:26:06 - INFO - codeparrot_training - Step 22967: {'lr': 0.00029930205236555464, 'samples': 4409856, 'steps': 22967, 'loss/train': 2.1379626989364624} 01/29/2022 18:26:11 - INFO - codeparrot_training - Step 22968: {'lr': 0.000299286011130338, 'samples': 4410048, 'steps': 22968, 'loss/train': 1.8259559869766235} 01/29/2022 18:26:16 - INFO - codeparrot_training - Step 22969: {'lr': 0.00029926996968399576, 'samples': 4410240, 'steps': 22969, 'loss/train': 1.6146785616874695} 01/29/2022 18:26:20 - INFO - codeparrot_training - Step 22970: {'lr': 0.0002992539280265966, 'samples': 4410432, 'steps': 22970, 'loss/train': 1.68508780002594} 01/29/2022 18:26:24 - INFO - codeparrot_training - Step 22971: {'lr': 0.00029923788615820936, 'samples': 4410624, 'steps': 22971, 'loss/train': 1.6271243691444397} 01/29/2022 18:26:29 - INFO - codeparrot_training - Step 22972: {'lr': 0.0002992218440789025, 'samples': 4410816, 'steps': 22972, 'loss/train': 1.5803046226501465} 01/29/2022 18:26:33 - INFO - codeparrot_training - Step 22973: {'lr': 0.00029920580178874497, 'samples': 4411008, 'steps': 22973, 'loss/train': 2.113477349281311} 01/29/2022 18:26:39 - INFO - codeparrot_training - Step 22974: {'lr': 0.00029918975928780537, 'samples': 4411200, 'steps': 22974, 'loss/train': 0.9925240874290466} 01/29/2022 18:26:44 - INFO - codeparrot_training - Step 22975: {'lr': 0.0002991737165761525, 'samples': 4411392, 'steps': 22975, 'loss/train': 2.055527687072754} 01/29/2022 18:26:48 - INFO - codeparrot_training - Step 22976: {'lr': 0.000299157673653855, 'samples': 4411584, 'steps': 22976, 'loss/train': 0.9820162653923035} 01/29/2022 18:26:52 - INFO - codeparrot_training - Step 22977: {'lr': 0.0002991416305209817, 'samples': 4411776, 'steps': 22977, 'loss/train': 1.787769377231598} 01/29/2022 18:26:56 - INFO - codeparrot_training - Step 22978: {'lr': 0.0002991255871776012, 'samples': 4411968, 'steps': 22978, 'loss/train': 1.492217481136322} 01/29/2022 18:27:02 - INFO - codeparrot_training - Step 22979: {'lr': 0.0002991095436237823, 'samples': 4412160, 'steps': 22979, 'loss/train': 0.47994978725910187} 01/29/2022 18:27:06 - INFO - codeparrot_training - Step 22980: {'lr': 0.00029909349985959377, 'samples': 4412352, 'steps': 22980, 'loss/train': 3.39680278301239} 01/29/2022 18:27:10 - INFO - codeparrot_training - Step 22981: {'lr': 0.00029907745588510416, 'samples': 4412544, 'steps': 22981, 'loss/train': 0.5173554718494415} 01/29/2022 18:27:15 - INFO - codeparrot_training - Step 22982: {'lr': 0.00029906141170038243, 'samples': 4412736, 'steps': 22982, 'loss/train': 1.4966984689235687} 01/29/2022 18:27:19 - INFO - codeparrot_training - Step 22983: {'lr': 0.00029904536730549706, 'samples': 4412928, 'steps': 22983, 'loss/train': 0.878593236207962} 01/29/2022 18:27:25 - INFO - codeparrot_training - Step 22984: {'lr': 0.00029902932270051705, 'samples': 4413120, 'steps': 22984, 'loss/train': 2.097237467765808} 01/29/2022 18:27:29 - INFO - codeparrot_training - Step 22985: {'lr': 0.00029901327788551087, 'samples': 4413312, 'steps': 22985, 'loss/train': 1.1548473536968231} 01/29/2022 18:27:33 - INFO - codeparrot_training - Step 22986: {'lr': 0.0002989972328605475, 'samples': 4413504, 'steps': 22986, 'loss/train': 1.7361928224563599} 01/29/2022 18:27:37 - INFO - codeparrot_training - Step 22987: {'lr': 0.0002989811876256954, 'samples': 4413696, 'steps': 22987, 'loss/train': 0.9563450217247009} 01/29/2022 18:27:42 - INFO - codeparrot_training - Step 22988: {'lr': 0.0002989651421810235, 'samples': 4413888, 'steps': 22988, 'loss/train': 2.6399272084236145} 01/29/2022 18:27:48 - INFO - codeparrot_training - Step 22989: {'lr': 0.00029894909652660053, 'samples': 4414080, 'steps': 22989, 'loss/train': 1.6022130846977234} 01/29/2022 18:27:53 - INFO - codeparrot_training - Step 22990: {'lr': 0.0002989330506624951, 'samples': 4414272, 'steps': 22990, 'loss/train': 1.8688042759895325} 01/29/2022 18:27:57 - INFO - codeparrot_training - Step 22991: {'lr': 0.0002989170045887761, 'samples': 4414464, 'steps': 22991, 'loss/train': 2.324509084224701} 01/29/2022 18:28:01 - INFO - codeparrot_training - Step 22992: {'lr': 0.00029890095830551204, 'samples': 4414656, 'steps': 22992, 'loss/train': 1.6287261843681335} 01/29/2022 18:28:05 - INFO - codeparrot_training - Step 22993: {'lr': 0.00029888491181277195, 'samples': 4414848, 'steps': 22993, 'loss/train': 1.263989120721817} 01/29/2022 18:28:10 - INFO - codeparrot_training - Step 22994: {'lr': 0.00029886886511062434, 'samples': 4415040, 'steps': 22994, 'loss/train': 1.062888264656067} 01/29/2022 18:28:15 - INFO - codeparrot_training - Step 22995: {'lr': 0.0002988528181991381, 'samples': 4415232, 'steps': 22995, 'loss/train': 1.923064112663269} 01/29/2022 18:28:19 - INFO - codeparrot_training - Step 22996: {'lr': 0.00029883677107838183, 'samples': 4415424, 'steps': 22996, 'loss/train': 2.177878439426422} 01/29/2022 18:28:23 - INFO - codeparrot_training - Step 22997: {'lr': 0.0002988207237484244, 'samples': 4415616, 'steps': 22997, 'loss/train': 1.8306997418403625} 01/29/2022 18:28:27 - INFO - codeparrot_training - Step 22998: {'lr': 0.0002988046762093344, 'samples': 4415808, 'steps': 22998, 'loss/train': 1.8007375001907349} 01/29/2022 18:28:33 - INFO - codeparrot_training - Step 22999: {'lr': 0.00029878862846118075, 'samples': 4416000, 'steps': 22999, 'loss/train': 1.5074341893196106} 01/29/2022 18:28:37 - INFO - codeparrot_training - Step 23000: {'lr': 0.0002987725805040321, 'samples': 4416192, 'steps': 23000, 'loss/train': 0.9491569697856903} 01/29/2022 18:28:41 - INFO - codeparrot_training - Step 23001: {'lr': 0.00029875653233795715, 'samples': 4416384, 'steps': 23001, 'loss/train': 5.5430556535720825} 01/29/2022 18:28:46 - INFO - codeparrot_training - Step 23002: {'lr': 0.0002987404839630248, 'samples': 4416576, 'steps': 23002, 'loss/train': 2.128088593482971} 01/29/2022 18:28:50 - INFO - codeparrot_training - Step 23003: {'lr': 0.00029872443537930357, 'samples': 4416768, 'steps': 23003, 'loss/train': 1.970357358455658} 01/29/2022 18:28:54 - INFO - codeparrot_training - Step 23004: {'lr': 0.0002987083865868624, 'samples': 4416960, 'steps': 23004, 'loss/train': 1.778695821762085} 01/29/2022 18:29:00 - INFO - codeparrot_training - Step 23005: {'lr': 0.0002986923375857699, 'samples': 4417152, 'steps': 23005, 'loss/train': 1.370459407567978} 01/29/2022 18:29:04 - INFO - codeparrot_training - Step 23006: {'lr': 0.00029867628837609503, 'samples': 4417344, 'steps': 23006, 'loss/train': 1.4542391002178192} 01/29/2022 18:29:08 - INFO - codeparrot_training - Step 23007: {'lr': 0.0002986602389579062, 'samples': 4417536, 'steps': 23007, 'loss/train': 1.631394624710083} 01/29/2022 18:29:12 - INFO - codeparrot_training - Step 23008: {'lr': 0.0002986441893312726, 'samples': 4417728, 'steps': 23008, 'loss/train': 2.1123728156089783} 01/29/2022 18:29:19 - INFO - codeparrot_training - Step 23009: {'lr': 0.0002986281394962626, 'samples': 4417920, 'steps': 23009, 'loss/train': 0.7910929620265961} 01/29/2022 18:29:24 - INFO - codeparrot_training - Step 23010: {'lr': 0.00029861208945294507, 'samples': 4418112, 'steps': 23010, 'loss/train': 1.1339275538921356} 01/29/2022 18:29:28 - INFO - codeparrot_training - Step 23011: {'lr': 0.00029859603920138876, 'samples': 4418304, 'steps': 23011, 'loss/train': 1.200260192155838} 01/29/2022 18:29:32 - INFO - codeparrot_training - Step 23012: {'lr': 0.00029857998874166253, 'samples': 4418496, 'steps': 23012, 'loss/train': 1.6534844040870667} 01/29/2022 18:29:36 - INFO - codeparrot_training - Step 23013: {'lr': 0.00029856393807383504, 'samples': 4418688, 'steps': 23013, 'loss/train': 0.534333810210228} 01/29/2022 18:29:42 - INFO - codeparrot_training - Step 23014: {'lr': 0.000298547887197975, 'samples': 4418880, 'steps': 23014, 'loss/train': 1.8801698684692383} 01/29/2022 18:29:46 - INFO - codeparrot_training - Step 23015: {'lr': 0.0002985318361141513, 'samples': 4419072, 'steps': 23015, 'loss/train': 0.6082141846418381} 01/29/2022 18:29:50 - INFO - codeparrot_training - Step 23016: {'lr': 0.0002985157848224326, 'samples': 4419264, 'steps': 23016, 'loss/train': 3.625856637954712} 01/29/2022 18:29:54 - INFO - codeparrot_training - Step 23017: {'lr': 0.00029849973332288763, 'samples': 4419456, 'steps': 23017, 'loss/train': 1.5938211679458618} 01/29/2022 18:29:59 - INFO - codeparrot_training - Step 23018: {'lr': 0.00029848368161558526, 'samples': 4419648, 'steps': 23018, 'loss/train': 1.8360373377799988} 01/29/2022 18:30:04 - INFO - codeparrot_training - Step 23019: {'lr': 0.0002984676297005942, 'samples': 4419840, 'steps': 23019, 'loss/train': 1.194874495267868} 01/29/2022 18:30:08 - INFO - codeparrot_training - Step 23020: {'lr': 0.0002984515775779832, 'samples': 4420032, 'steps': 23020, 'loss/train': 2.4280391335487366} 01/29/2022 18:30:12 - INFO - codeparrot_training - Step 23021: {'lr': 0.00029843552524782104, 'samples': 4420224, 'steps': 23021, 'loss/train': 1.140060156583786} 01/29/2022 18:30:17 - INFO - codeparrot_training - Step 23022: {'lr': 0.00029841947271017647, 'samples': 4420416, 'steps': 23022, 'loss/train': 1.230110228061676} 01/29/2022 18:30:21 - INFO - codeparrot_training - Step 23023: {'lr': 0.0002984034199651182, 'samples': 4420608, 'steps': 23023, 'loss/train': 1.8766189813613892} 01/29/2022 18:30:29 - INFO - codeparrot_training - Step 23024: {'lr': 0.00029838736701271514, 'samples': 4420800, 'steps': 23024, 'loss/train': 2.3643380999565125} 01/29/2022 18:30:33 - INFO - codeparrot_training - Step 23025: {'lr': 0.00029837131385303587, 'samples': 4420992, 'steps': 23025, 'loss/train': 1.4539784789085388} 01/29/2022 18:30:37 - INFO - codeparrot_training - Step 23026: {'lr': 0.0002983552604861493, 'samples': 4421184, 'steps': 23026, 'loss/train': 1.1441902220249176} 01/29/2022 18:30:41 - INFO - codeparrot_training - Step 23027: {'lr': 0.0002983392069121241, 'samples': 4421376, 'steps': 23027, 'loss/train': 1.4763380885124207} 01/29/2022 18:30:45 - INFO - codeparrot_training - Step 23028: {'lr': 0.00029832315313102915, 'samples': 4421568, 'steps': 23028, 'loss/train': 0.8269141316413879} 01/29/2022 18:30:50 - INFO - codeparrot_training - Step 23029: {'lr': 0.00029830709914293306, 'samples': 4421760, 'steps': 23029, 'loss/train': 2.123270273208618} 01/29/2022 18:30:55 - INFO - codeparrot_training - Step 23030: {'lr': 0.00029829104494790483, 'samples': 4421952, 'steps': 23030, 'loss/train': 1.2172421514987946} 01/29/2022 18:30:59 - INFO - codeparrot_training - Step 23031: {'lr': 0.00029827499054601306, 'samples': 4422144, 'steps': 23031, 'loss/train': 1.655597984790802} 01/29/2022 18:31:03 - INFO - codeparrot_training - Step 23032: {'lr': 0.0002982589359373265, 'samples': 4422336, 'steps': 23032, 'loss/train': 1.7984713912010193} 01/29/2022 18:31:07 - INFO - codeparrot_training - Step 23033: {'lr': 0.000298242881121914, 'samples': 4422528, 'steps': 23033, 'loss/train': 1.8164291381835938} 01/29/2022 18:31:12 - INFO - codeparrot_training - Step 23034: {'lr': 0.00029822682609984436, 'samples': 4422720, 'steps': 23034, 'loss/train': 1.6172073483467102} 01/29/2022 18:31:19 - INFO - codeparrot_training - Step 23035: {'lr': 0.00029821077087118625, 'samples': 4422912, 'steps': 23035, 'loss/train': 2.2695103883743286} 01/29/2022 18:31:23 - INFO - codeparrot_training - Step 23036: {'lr': 0.00029819471543600856, 'samples': 4423104, 'steps': 23036, 'loss/train': 2.0083000659942627} 01/29/2022 18:31:28 - INFO - codeparrot_training - Step 23037: {'lr': 0.00029817865979437996, 'samples': 4423296, 'steps': 23037, 'loss/train': 1.5501455068588257} 01/29/2022 18:31:32 - INFO - codeparrot_training - Step 23038: {'lr': 0.0002981626039463693, 'samples': 4423488, 'steps': 23038, 'loss/train': 1.0201160609722137} 01/29/2022 18:31:36 - INFO - codeparrot_training - Step 23039: {'lr': 0.0002981465478920454, 'samples': 4423680, 'steps': 23039, 'loss/train': 2.1868144869804382} 01/29/2022 18:31:41 - INFO - codeparrot_training - Step 23040: {'lr': 0.00029813049163147687, 'samples': 4423872, 'steps': 23040, 'loss/train': 0.36695674806833267} 01/29/2022 18:31:46 - INFO - codeparrot_training - Step 23041: {'lr': 0.0002981144351647327, 'samples': 4424064, 'steps': 23041, 'loss/train': 1.1891387701034546} 01/29/2022 18:31:50 - INFO - codeparrot_training - Step 23042: {'lr': 0.0002980983784918815, 'samples': 4424256, 'steps': 23042, 'loss/train': 2.0426174998283386} 01/29/2022 18:31:54 - INFO - codeparrot_training - Step 23043: {'lr': 0.0002980823216129921, 'samples': 4424448, 'steps': 23043, 'loss/train': 1.8515739440917969} 01/29/2022 18:31:58 - INFO - codeparrot_training - Step 23044: {'lr': 0.00029806626452813333, 'samples': 4424640, 'steps': 23044, 'loss/train': 1.966461718082428} 01/29/2022 18:32:04 - INFO - codeparrot_training - Step 23045: {'lr': 0.000298050207237374, 'samples': 4424832, 'steps': 23045, 'loss/train': 1.403796672821045} 01/29/2022 18:32:08 - INFO - codeparrot_training - Step 23046: {'lr': 0.0002980341497407828, 'samples': 4425024, 'steps': 23046, 'loss/train': 1.0729482471942902} 01/29/2022 18:32:12 - INFO - codeparrot_training - Step 23047: {'lr': 0.0002980180920384286, 'samples': 4425216, 'steps': 23047, 'loss/train': 2.042766809463501} 01/29/2022 18:32:17 - INFO - codeparrot_training - Step 23048: {'lr': 0.00029800203413038, 'samples': 4425408, 'steps': 23048, 'loss/train': 1.2288187444210052} 01/29/2022 18:32:21 - INFO - codeparrot_training - Step 23049: {'lr': 0.000297985976016706, 'samples': 4425600, 'steps': 23049, 'loss/train': 1.3425982296466827} 01/29/2022 18:32:26 - INFO - codeparrot_training - Step 23050: {'lr': 0.0002979699176974754, 'samples': 4425792, 'steps': 23050, 'loss/train': 1.9713072180747986} 01/29/2022 18:32:30 - INFO - codeparrot_training - Step 23051: {'lr': 0.0002979538591727568, 'samples': 4425984, 'steps': 23051, 'loss/train': 1.8561809062957764} 01/29/2022 18:32:35 - INFO - codeparrot_training - Step 23052: {'lr': 0.00029793780044261916, 'samples': 4426176, 'steps': 23052, 'loss/train': 1.2474626004695892} 01/29/2022 18:32:39 - INFO - codeparrot_training - Step 23053: {'lr': 0.0002979217415071311, 'samples': 4426368, 'steps': 23053, 'loss/train': 1.851986825466156} 01/29/2022 18:32:43 - INFO - codeparrot_training - Step 23054: {'lr': 0.00029790568236636166, 'samples': 4426560, 'steps': 23054, 'loss/train': 1.8090705871582031} 01/29/2022 18:32:50 - INFO - codeparrot_training - Step 23055: {'lr': 0.0002978896230203794, 'samples': 4426752, 'steps': 23055, 'loss/train': 1.6334649324417114} 01/29/2022 18:32:54 - INFO - codeparrot_training - Step 23056: {'lr': 0.0002978735634692533, 'samples': 4426944, 'steps': 23056, 'loss/train': 1.376938372850418} 01/29/2022 18:32:58 - INFO - codeparrot_training - Step 23057: {'lr': 0.0002978575037130519, 'samples': 4427136, 'steps': 23057, 'loss/train': 2.3455894589424133} 01/29/2022 18:33:03 - INFO - codeparrot_training - Step 23058: {'lr': 0.0002978414437518443, 'samples': 4427328, 'steps': 23058, 'loss/train': 1.396323412656784} 01/29/2022 18:33:07 - INFO - codeparrot_training - Step 23059: {'lr': 0.00029782538358569905, 'samples': 4427520, 'steps': 23059, 'loss/train': 1.0653463304042816} 01/29/2022 18:33:12 - INFO - codeparrot_training - Step 23060: {'lr': 0.0002978093232146851, 'samples': 4427712, 'steps': 23060, 'loss/train': 0.8534327745437622} 01/29/2022 18:33:16 - INFO - codeparrot_training - Step 23061: {'lr': 0.00029779326263887113, 'samples': 4427904, 'steps': 23061, 'loss/train': 2.0489278435707092} 01/29/2022 18:33:21 - INFO - codeparrot_training - Step 23062: {'lr': 0.00029777720185832605, 'samples': 4428096, 'steps': 23062, 'loss/train': 2.8205305337905884} 01/29/2022 18:33:25 - INFO - codeparrot_training - Step 23063: {'lr': 0.0002977611408731186, 'samples': 4428288, 'steps': 23063, 'loss/train': 2.5672622323036194} 01/29/2022 18:33:29 - INFO - codeparrot_training - Step 23064: {'lr': 0.0002977450796833176, 'samples': 4428480, 'steps': 23064, 'loss/train': 0.8126356601715088} 01/29/2022 18:33:35 - INFO - codeparrot_training - Step 23065: {'lr': 0.0002977290182889918, 'samples': 4428672, 'steps': 23065, 'loss/train': 1.7380022406578064} 01/29/2022 18:33:39 - INFO - codeparrot_training - Step 23066: {'lr': 0.00029771295669021, 'samples': 4428864, 'steps': 23066, 'loss/train': 1.7631163001060486} 01/29/2022 18:33:44 - INFO - codeparrot_training - Step 23067: {'lr': 0.00029769689488704117, 'samples': 4429056, 'steps': 23067, 'loss/train': 1.384735643863678} 01/29/2022 18:33:48 - INFO - codeparrot_training - Step 23068: {'lr': 0.00029768083287955394, 'samples': 4429248, 'steps': 23068, 'loss/train': 1.9085624814033508} 01/29/2022 18:33:52 - INFO - codeparrot_training - Step 23069: {'lr': 0.00029766477066781716, 'samples': 4429440, 'steps': 23069, 'loss/train': 1.8446475863456726} 01/29/2022 18:33:56 - INFO - codeparrot_training - Step 23070: {'lr': 0.0002976487082518996, 'samples': 4429632, 'steps': 23070, 'loss/train': 1.3955824077129364} 01/29/2022 18:34:03 - INFO - codeparrot_training - Step 23071: {'lr': 0.00029763264563187016, 'samples': 4429824, 'steps': 23071, 'loss/train': 0.9198881685733795} 01/29/2022 18:34:08 - INFO - codeparrot_training - Step 23072: {'lr': 0.0002976165828077975, 'samples': 4430016, 'steps': 23072, 'loss/train': 1.3852568864822388} 01/29/2022 18:34:12 - INFO - codeparrot_training - Step 23073: {'lr': 0.00029760051977975057, 'samples': 4430208, 'steps': 23073, 'loss/train': 1.3958100378513336} 01/29/2022 18:34:16 - INFO - codeparrot_training - Step 23074: {'lr': 0.00029758445654779814, 'samples': 4430400, 'steps': 23074, 'loss/train': 2.0058860778808594} 01/29/2022 18:34:20 - INFO - codeparrot_training - Step 23075: {'lr': 0.000297568393112009, 'samples': 4430592, 'steps': 23075, 'loss/train': 1.4790295958518982} 01/29/2022 18:34:25 - INFO - codeparrot_training - Step 23076: {'lr': 0.00029755232947245195, 'samples': 4430784, 'steps': 23076, 'loss/train': 1.613834023475647} 01/29/2022 18:34:30 - INFO - codeparrot_training - Step 23077: {'lr': 0.0002975362656291958, 'samples': 4430976, 'steps': 23077, 'loss/train': 0.9335027933120728} 01/29/2022 18:34:34 - INFO - codeparrot_training - Step 23078: {'lr': 0.0002975202015823095, 'samples': 4431168, 'steps': 23078, 'loss/train': 1.6172756552696228} 01/29/2022 18:34:38 - INFO - codeparrot_training - Step 23079: {'lr': 0.0002975041373318617, 'samples': 4431360, 'steps': 23079, 'loss/train': 2.1997944116592407} 01/29/2022 18:34:42 - INFO - codeparrot_training - Step 23080: {'lr': 0.0002974880728779212, 'samples': 4431552, 'steps': 23080, 'loss/train': 1.6990166902542114} 01/29/2022 18:34:49 - INFO - codeparrot_training - Step 23081: {'lr': 0.00029747200822055684, 'samples': 4431744, 'steps': 23081, 'loss/train': 1.6540609002113342} 01/29/2022 18:34:54 - INFO - codeparrot_training - Step 23082: {'lr': 0.0002974559433598376, 'samples': 4431936, 'steps': 23082, 'loss/train': 0.5006837099790573} 01/29/2022 18:34:58 - INFO - codeparrot_training - Step 23083: {'lr': 0.0002974398782958321, 'samples': 4432128, 'steps': 23083, 'loss/train': 1.4754895269870758} 01/29/2022 18:35:02 - INFO - codeparrot_training - Step 23084: {'lr': 0.00029742381302860923, 'samples': 4432320, 'steps': 23084, 'loss/train': 1.0189997255802155} 01/29/2022 18:35:06 - INFO - codeparrot_training - Step 23085: {'lr': 0.00029740774755823777, 'samples': 4432512, 'steps': 23085, 'loss/train': 1.7615439891815186} 01/29/2022 18:35:10 - INFO - codeparrot_training - Step 23086: {'lr': 0.0002973916818847866, 'samples': 4432704, 'steps': 23086, 'loss/train': 1.0611320436000824} 01/29/2022 18:35:16 - INFO - codeparrot_training - Step 23087: {'lr': 0.00029737561600832454, 'samples': 4432896, 'steps': 23087, 'loss/train': 1.8515342473983765} 01/29/2022 18:35:20 - INFO - codeparrot_training - Step 23088: {'lr': 0.00029735954992892035, 'samples': 4433088, 'steps': 23088, 'loss/train': 1.1151388585567474} 01/29/2022 18:35:24 - INFO - codeparrot_training - Step 23089: {'lr': 0.00029734348364664285, 'samples': 4433280, 'steps': 23089, 'loss/train': 0.4310491383075714} 01/29/2022 18:35:28 - INFO - codeparrot_training - Step 23090: {'lr': 0.000297327417161561, 'samples': 4433472, 'steps': 23090, 'loss/train': 2.0430171489715576} 01/29/2022 18:35:33 - INFO - codeparrot_training - Step 23091: {'lr': 0.0002973113504737435, 'samples': 4433664, 'steps': 23091, 'loss/train': 1.6617136001586914} 01/29/2022 18:35:38 - INFO - codeparrot_training - Step 23092: {'lr': 0.00029729528358325914, 'samples': 4433856, 'steps': 23092, 'loss/train': 3.135055661201477} 01/29/2022 18:35:42 - INFO - codeparrot_training - Step 23093: {'lr': 0.00029727921649017687, 'samples': 4434048, 'steps': 23093, 'loss/train': 1.7480605244636536} 01/29/2022 18:35:46 - INFO - codeparrot_training - Step 23094: {'lr': 0.0002972631491945653, 'samples': 4434240, 'steps': 23094, 'loss/train': 0.6088943481445312} 01/29/2022 18:35:51 - INFO - codeparrot_training - Step 23095: {'lr': 0.00029724708169649364, 'samples': 4434432, 'steps': 23095, 'loss/train': 1.4190166890621185} 01/29/2022 18:35:55 - INFO - codeparrot_training - Step 23096: {'lr': 0.0002972310139960303, 'samples': 4434624, 'steps': 23096, 'loss/train': 1.4016354382038116} 01/29/2022 18:36:02 - INFO - codeparrot_training - Step 23097: {'lr': 0.00029721494609324435, 'samples': 4434816, 'steps': 23097, 'loss/train': 2.0085148215293884} 01/29/2022 18:36:06 - INFO - codeparrot_training - Step 23098: {'lr': 0.00029719887798820453, 'samples': 4435008, 'steps': 23098, 'loss/train': 0.6903316229581833} 01/29/2022 18:36:10 - INFO - codeparrot_training - Step 23099: {'lr': 0.00029718280968097976, 'samples': 4435200, 'steps': 23099, 'loss/train': 1.5519681572914124} 01/29/2022 18:36:15 - INFO - codeparrot_training - Step 23100: {'lr': 0.00029716674117163884, 'samples': 4435392, 'steps': 23100, 'loss/train': 1.7665636539459229} 01/29/2022 18:36:19 - INFO - codeparrot_training - Step 23101: {'lr': 0.0002971506724602505, 'samples': 4435584, 'steps': 23101, 'loss/train': 2.001471519470215} 01/29/2022 18:36:24 - INFO - codeparrot_training - Step 23102: {'lr': 0.0002971346035468837, 'samples': 4435776, 'steps': 23102, 'loss/train': 2.2460432052612305} 01/29/2022 18:36:28 - INFO - codeparrot_training - Step 23103: {'lr': 0.0002971185344316072, 'samples': 4435968, 'steps': 23103, 'loss/train': 1.6324945092201233} 01/29/2022 18:36:32 - INFO - codeparrot_training - Step 23104: {'lr': 0.0002971024651144899, 'samples': 4436160, 'steps': 23104, 'loss/train': 1.6673092246055603} 01/29/2022 18:36:37 - INFO - codeparrot_training - Step 23105: {'lr': 0.0002970863955956005, 'samples': 4436352, 'steps': 23105, 'loss/train': 1.6129724979400635} 01/29/2022 18:36:41 - INFO - codeparrot_training - Step 23106: {'lr': 0.00029707032587500805, 'samples': 4436544, 'steps': 23106, 'loss/train': 1.5977883338928223} 01/29/2022 18:36:47 - INFO - codeparrot_training - Step 23107: {'lr': 0.00029705425595278126, 'samples': 4436736, 'steps': 23107, 'loss/train': 1.5122979283332825} 01/29/2022 18:36:51 - INFO - codeparrot_training - Step 23108: {'lr': 0.0002970381858289889, 'samples': 4436928, 'steps': 23108, 'loss/train': 1.386611670255661} 01/29/2022 18:36:55 - INFO - codeparrot_training - Step 23109: {'lr': 0.0002970221155037, 'samples': 4437120, 'steps': 23109, 'loss/train': 2.368249475955963} 01/29/2022 18:36:59 - INFO - codeparrot_training - Step 23110: {'lr': 0.0002970060449769832, 'samples': 4437312, 'steps': 23110, 'loss/train': 1.9223610162734985} 01/29/2022 18:37:03 - INFO - codeparrot_training - Step 23111: {'lr': 0.00029698997424890746, 'samples': 4437504, 'steps': 23111, 'loss/train': 1.6111754179000854} 01/29/2022 18:37:08 - INFO - codeparrot_training - Step 23112: {'lr': 0.0002969739033195415, 'samples': 4437696, 'steps': 23112, 'loss/train': 1.9728158712387085} 01/29/2022 18:37:14 - INFO - codeparrot_training - Step 23113: {'lr': 0.0002969578321889544, 'samples': 4437888, 'steps': 23113, 'loss/train': 1.6741247177124023} 01/29/2022 18:37:18 - INFO - codeparrot_training - Step 23114: {'lr': 0.00029694176085721474, 'samples': 4438080, 'steps': 23114, 'loss/train': 1.405007243156433} 01/29/2022 18:37:22 - INFO - codeparrot_training - Step 23115: {'lr': 0.0002969256893243916, 'samples': 4438272, 'steps': 23115, 'loss/train': 1.5891265869140625} 01/29/2022 18:37:27 - INFO - codeparrot_training - Step 23116: {'lr': 0.0002969096175905536, 'samples': 4438464, 'steps': 23116, 'loss/train': 2.036029100418091} 01/29/2022 18:37:31 - INFO - codeparrot_training - Step 23117: {'lr': 0.00029689354565576976, 'samples': 4438656, 'steps': 23117, 'loss/train': 2.907109558582306} 01/29/2022 18:37:36 - INFO - codeparrot_training - Step 23118: {'lr': 0.0002968774735201088, 'samples': 4438848, 'steps': 23118, 'loss/train': 1.739692747592926} 01/29/2022 18:37:40 - INFO - codeparrot_training - Step 23119: {'lr': 0.0002968614011836397, 'samples': 4439040, 'steps': 23119, 'loss/train': 0.09935282170772552} 01/29/2022 18:37:45 - INFO - codeparrot_training - Step 23120: {'lr': 0.0002968453286464312, 'samples': 4439232, 'steps': 23120, 'loss/train': 1.437962919473648} 01/29/2022 18:37:49 - INFO - codeparrot_training - Step 23121: {'lr': 0.00029682925590855213, 'samples': 4439424, 'steps': 23121, 'loss/train': 1.4750226438045502} 01/29/2022 18:37:53 - INFO - codeparrot_training - Step 23122: {'lr': 0.0002968131829700715, 'samples': 4439616, 'steps': 23122, 'loss/train': 0.8548922538757324} 01/29/2022 18:37:58 - INFO - codeparrot_training - Step 23123: {'lr': 0.000296797109831058, 'samples': 4439808, 'steps': 23123, 'loss/train': 1.9634762406349182} 01/29/2022 18:38:03 - INFO - codeparrot_training - Step 23124: {'lr': 0.00029678103649158057, 'samples': 4440000, 'steps': 23124, 'loss/train': 2.0205230712890625} 01/29/2022 18:38:07 - INFO - codeparrot_training - Step 23125: {'lr': 0.00029676496295170804, 'samples': 4440192, 'steps': 23125, 'loss/train': 1.1049490571022034} 01/29/2022 18:38:11 - INFO - codeparrot_training - Step 23126: {'lr': 0.0002967488892115092, 'samples': 4440384, 'steps': 23126, 'loss/train': 0.7669596076011658} 01/29/2022 18:38:15 - INFO - codeparrot_training - Step 23127: {'lr': 0.000296732815271053, 'samples': 4440576, 'steps': 23127, 'loss/train': 1.0560089349746704} 01/29/2022 18:38:23 - INFO - codeparrot_training - Step 23128: {'lr': 0.00029671674113040833, 'samples': 4440768, 'steps': 23128, 'loss/train': 1.059216320514679} 01/29/2022 18:38:27 - INFO - codeparrot_training - Step 23129: {'lr': 0.00029670066678964385, 'samples': 4440960, 'steps': 23129, 'loss/train': 2.0115979313850403} 01/29/2022 18:38:31 - INFO - codeparrot_training - Step 23130: {'lr': 0.0002966845922488286, 'samples': 4441152, 'steps': 23130, 'loss/train': 1.3042857348918915} 01/29/2022 18:38:35 - INFO - codeparrot_training - Step 23131: {'lr': 0.00029666851750803137, 'samples': 4441344, 'steps': 23131, 'loss/train': 1.5222651958465576} 01/29/2022 18:38:39 - INFO - codeparrot_training - Step 23132: {'lr': 0.00029665244256732107, 'samples': 4441536, 'steps': 23132, 'loss/train': 1.9952389597892761} 01/29/2022 18:38:45 - INFO - codeparrot_training - Step 23133: {'lr': 0.0002966363674267665, 'samples': 4441728, 'steps': 23133, 'loss/train': 1.5430419445037842} 01/29/2022 18:38:49 - INFO - codeparrot_training - Step 23134: {'lr': 0.00029662029208643646, 'samples': 4441920, 'steps': 23134, 'loss/train': 6.626663446426392} 01/29/2022 18:38:53 - INFO - codeparrot_training - Step 23135: {'lr': 0.0002966042165464, 'samples': 4442112, 'steps': 23135, 'loss/train': 1.9223134517669678} 01/29/2022 18:38:57 - INFO - codeparrot_training - Step 23136: {'lr': 0.0002965881408067258, 'samples': 4442304, 'steps': 23136, 'loss/train': 1.1105023920536041} 01/29/2022 18:39:03 - INFO - codeparrot_training - Step 23137: {'lr': 0.0002965720648674829, 'samples': 4442496, 'steps': 23137, 'loss/train': 1.785583734512329} 01/29/2022 18:39:07 - INFO - codeparrot_training - Step 23138: {'lr': 0.00029655598872873994, 'samples': 4442688, 'steps': 23138, 'loss/train': 1.8721386194229126} 01/29/2022 18:39:11 - INFO - codeparrot_training - Step 23139: {'lr': 0.000296539912390566, 'samples': 4442880, 'steps': 23139, 'loss/train': 1.4665687680244446} 01/29/2022 18:39:15 - INFO - codeparrot_training - Step 23140: {'lr': 0.0002965238358530298, 'samples': 4443072, 'steps': 23140, 'loss/train': 1.5840913653373718} 01/29/2022 18:39:20 - INFO - codeparrot_training - Step 23141: {'lr': 0.0002965077591162003, 'samples': 4443264, 'steps': 23141, 'loss/train': 3.4325594902038574} 01/29/2022 18:39:24 - INFO - codeparrot_training - Step 23142: {'lr': 0.00029649168218014627, 'samples': 4443456, 'steps': 23142, 'loss/train': 1.5362436175346375} 01/29/2022 18:39:31 - INFO - codeparrot_training - Step 23143: {'lr': 0.0002964756050449367, 'samples': 4443648, 'steps': 23143, 'loss/train': 1.7773287892341614} 01/29/2022 18:39:35 - INFO - codeparrot_training - Step 23144: {'lr': 0.0002964595277106403, 'samples': 4443840, 'steps': 23144, 'loss/train': 1.7885060906410217} 01/29/2022 18:39:40 - INFO - codeparrot_training - Step 23145: {'lr': 0.0002964434501773262, 'samples': 4444032, 'steps': 23145, 'loss/train': 1.2465850710868835} 01/29/2022 18:39:44 - INFO - codeparrot_training - Step 23146: {'lr': 0.00029642737244506295, 'samples': 4444224, 'steps': 23146, 'loss/train': 1.2899384200572968} 01/29/2022 18:39:48 - INFO - codeparrot_training - Step 23147: {'lr': 0.0002964112945139196, 'samples': 4444416, 'steps': 23147, 'loss/train': 1.8332045674324036} 01/29/2022 18:39:53 - INFO - codeparrot_training - Step 23148: {'lr': 0.0002963952163839651, 'samples': 4444608, 'steps': 23148, 'loss/train': 1.7137954831123352} 01/29/2022 18:39:58 - INFO - codeparrot_training - Step 23149: {'lr': 0.00029637913805526816, 'samples': 4444800, 'steps': 23149, 'loss/train': 2.582074463367462} 01/29/2022 18:40:02 - INFO - codeparrot_training - Step 23150: {'lr': 0.0002963630595278977, 'samples': 4444992, 'steps': 23150, 'loss/train': 1.6220165491104126} 01/29/2022 18:40:06 - INFO - codeparrot_training - Step 23151: {'lr': 0.0002963469808019227, 'samples': 4445184, 'steps': 23151, 'loss/train': 1.0220195353031158} 01/29/2022 18:40:10 - INFO - codeparrot_training - Step 23152: {'lr': 0.00029633090187741187, 'samples': 4445376, 'steps': 23152, 'loss/train': 1.4293614327907562} 01/29/2022 18:40:16 - INFO - codeparrot_training - Step 23153: {'lr': 0.00029631482275443424, 'samples': 4445568, 'steps': 23153, 'loss/train': 1.9287464618682861} 01/29/2022 18:40:20 - INFO - codeparrot_training - Step 23154: {'lr': 0.0002962987434330585, 'samples': 4445760, 'steps': 23154, 'loss/train': 1.8952546119689941} 01/29/2022 18:40:24 - INFO - codeparrot_training - Step 23155: {'lr': 0.0002962826639133538, 'samples': 4445952, 'steps': 23155, 'loss/train': 1.8633751273155212} 01/29/2022 18:40:28 - INFO - codeparrot_training - Step 23156: {'lr': 0.00029626658419538873, 'samples': 4446144, 'steps': 23156, 'loss/train': 2.079039216041565} 01/29/2022 18:40:32 - INFO - codeparrot_training - Step 23157: {'lr': 0.0002962505042792324, 'samples': 4446336, 'steps': 23157, 'loss/train': 3.2094790935516357} 01/29/2022 18:40:39 - INFO - codeparrot_training - Step 23158: {'lr': 0.0002962344241649535, 'samples': 4446528, 'steps': 23158, 'loss/train': 2.5170240998268127} 01/29/2022 18:40:44 - INFO - codeparrot_training - Step 23159: {'lr': 0.000296218343852621, 'samples': 4446720, 'steps': 23159, 'loss/train': 2.5689695477485657} 01/29/2022 18:40:48 - INFO - codeparrot_training - Step 23160: {'lr': 0.0002962022633423039, 'samples': 4446912, 'steps': 23160, 'loss/train': 1.6599812507629395} 01/29/2022 18:40:52 - INFO - codeparrot_training - Step 23161: {'lr': 0.00029618618263407094, 'samples': 4447104, 'steps': 23161, 'loss/train': 1.6597378849983215} 01/29/2022 18:40:56 - INFO - codeparrot_training - Step 23162: {'lr': 0.00029617010172799095, 'samples': 4447296, 'steps': 23162, 'loss/train': 0.7564560770988464} 01/29/2022 18:41:01 - INFO - codeparrot_training - Step 23163: {'lr': 0.00029615402062413307, 'samples': 4447488, 'steps': 23163, 'loss/train': 1.0986165404319763} 01/29/2022 18:41:06 - INFO - codeparrot_training - Step 23164: {'lr': 0.00029613793932256583, 'samples': 4447680, 'steps': 23164, 'loss/train': 0.9762145578861237} 01/29/2022 18:41:10 - INFO - codeparrot_training - Step 23165: {'lr': 0.00029612185782335845, 'samples': 4447872, 'steps': 23165, 'loss/train': 1.694276511669159} 01/29/2022 18:41:14 - INFO - codeparrot_training - Step 23166: {'lr': 0.00029610577612657963, 'samples': 4448064, 'steps': 23166, 'loss/train': 2.2981384992599487} 01/29/2022 18:41:18 - INFO - codeparrot_training - Step 23167: {'lr': 0.0002960896942322983, 'samples': 4448256, 'steps': 23167, 'loss/train': 2.0583161115646362} 01/29/2022 18:41:25 - INFO - codeparrot_training - Step 23168: {'lr': 0.0002960736121405834, 'samples': 4448448, 'steps': 23168, 'loss/train': 1.8880535960197449} 01/29/2022 18:41:30 - INFO - codeparrot_training - Step 23169: {'lr': 0.00029605752985150367, 'samples': 4448640, 'steps': 23169, 'loss/train': 0.3782931715250015} 01/29/2022 18:41:34 - INFO - codeparrot_training - Step 23170: {'lr': 0.00029604144736512816, 'samples': 4448832, 'steps': 23170, 'loss/train': 2.4504014253616333} 01/29/2022 18:41:38 - INFO - codeparrot_training - Step 23171: {'lr': 0.00029602536468152575, 'samples': 4449024, 'steps': 23171, 'loss/train': 1.1112326681613922} 01/29/2022 18:41:42 - INFO - codeparrot_training - Step 23172: {'lr': 0.00029600928180076525, 'samples': 4449216, 'steps': 23172, 'loss/train': 1.1164332926273346} 01/29/2022 18:41:48 - INFO - codeparrot_training - Step 23173: {'lr': 0.0002959931987229156, 'samples': 4449408, 'steps': 23173, 'loss/train': 1.7164258360862732} 01/29/2022 18:41:52 - INFO - codeparrot_training - Step 23174: {'lr': 0.0002959771154480457, 'samples': 4449600, 'steps': 23174, 'loss/train': 2.0142742395401} 01/29/2022 18:41:56 - INFO - codeparrot_training - Step 23175: {'lr': 0.0002959610319762244, 'samples': 4449792, 'steps': 23175, 'loss/train': 1.7904317378997803} 01/29/2022 18:42:00 - INFO - codeparrot_training - Step 23176: {'lr': 0.0002959449483075207, 'samples': 4449984, 'steps': 23176, 'loss/train': 1.7117793560028076} 01/29/2022 18:42:05 - INFO - codeparrot_training - Step 23177: {'lr': 0.0002959288644420034, 'samples': 4450176, 'steps': 23177, 'loss/train': 1.7281857132911682} 01/29/2022 18:42:10 - INFO - codeparrot_training - Step 23178: {'lr': 0.00029591278037974147, 'samples': 4450368, 'steps': 23178, 'loss/train': 1.9885963797569275} 01/29/2022 18:42:14 - INFO - codeparrot_training - Step 23179: {'lr': 0.0002958966961208037, 'samples': 4450560, 'steps': 23179, 'loss/train': 1.896978735923767} 01/29/2022 18:42:18 - INFO - codeparrot_training - Step 23180: {'lr': 0.0002958806116652591, 'samples': 4450752, 'steps': 23180, 'loss/train': 2.2111555337905884} 01/29/2022 18:42:22 - INFO - codeparrot_training - Step 23181: {'lr': 0.0002958645270131765, 'samples': 4450944, 'steps': 23181, 'loss/train': 1.2742020785808563} 01/29/2022 18:42:27 - INFO - codeparrot_training - Step 23182: {'lr': 0.0002958484421646248, 'samples': 4451136, 'steps': 23182, 'loss/train': 0.9614979028701782} 01/29/2022 18:42:32 - INFO - codeparrot_training - Step 23183: {'lr': 0.000295832357119673, 'samples': 4451328, 'steps': 23183, 'loss/train': 1.1897641718387604} 01/29/2022 18:42:36 - INFO - codeparrot_training - Step 23184: {'lr': 0.00029581627187838993, 'samples': 4451520, 'steps': 23184, 'loss/train': 1.7033300399780273} 01/29/2022 18:42:40 - INFO - codeparrot_training - Step 23185: {'lr': 0.0002958001864408445, 'samples': 4451712, 'steps': 23185, 'loss/train': 2.6749479174613953} 01/29/2022 18:42:44 - INFO - codeparrot_training - Step 23186: {'lr': 0.0002957841008071056, 'samples': 4451904, 'steps': 23186, 'loss/train': 1.611595094203949} 01/29/2022 18:42:49 - INFO - codeparrot_training - Step 23187: {'lr': 0.0002957680149772422, 'samples': 4452096, 'steps': 23187, 'loss/train': 0.3499368578195572} 01/29/2022 18:42:56 - INFO - codeparrot_training - Step 23188: {'lr': 0.000295751928951323, 'samples': 4452288, 'steps': 23188, 'loss/train': 1.6294294595718384} 01/29/2022 18:43:00 - INFO - codeparrot_training - Step 23189: {'lr': 0.0002957358427294172, 'samples': 4452480, 'steps': 23189, 'loss/train': 1.1983872950077057} 01/29/2022 18:43:04 - INFO - codeparrot_training - Step 23190: {'lr': 0.00029571975631159355, 'samples': 4452672, 'steps': 23190, 'loss/train': 0.7314898073673248} 01/29/2022 18:43:08 - INFO - codeparrot_training - Step 23191: {'lr': 0.000295703669697921, 'samples': 4452864, 'steps': 23191, 'loss/train': 0.9694561958312988} 01/29/2022 18:43:12 - INFO - codeparrot_training - Step 23192: {'lr': 0.0002956875828884684, 'samples': 4453056, 'steps': 23192, 'loss/train': 0.7855068147182465} 01/29/2022 18:43:18 - INFO - codeparrot_training - Step 23193: {'lr': 0.0002956714958833047, 'samples': 4453248, 'steps': 23193, 'loss/train': 1.4939375817775726} 01/29/2022 18:43:22 - INFO - codeparrot_training - Step 23194: {'lr': 0.00029565540868249884, 'samples': 4453440, 'steps': 23194, 'loss/train': 1.352573812007904} 01/29/2022 18:43:26 - INFO - codeparrot_training - Step 23195: {'lr': 0.0002956393212861197, 'samples': 4453632, 'steps': 23195, 'loss/train': 2.3461192846298218} 01/29/2022 18:43:30 - INFO - codeparrot_training - Step 23196: {'lr': 0.0002956232336942362, 'samples': 4453824, 'steps': 23196, 'loss/train': 1.3016488552093506} 01/29/2022 18:43:36 - INFO - codeparrot_training - Step 23197: {'lr': 0.0002956071459069173, 'samples': 4454016, 'steps': 23197, 'loss/train': 1.7917349338531494} 01/29/2022 18:43:40 - INFO - codeparrot_training - Step 23198: {'lr': 0.0002955910579242318, 'samples': 4454208, 'steps': 23198, 'loss/train': 2.090689480304718} 01/29/2022 18:43:44 - INFO - codeparrot_training - Step 23199: {'lr': 0.0002955749697462487, 'samples': 4454400, 'steps': 23199, 'loss/train': 1.257991075515747} 01/29/2022 18:43:48 - INFO - codeparrot_training - Step 23200: {'lr': 0.0002955588813730369, 'samples': 4454592, 'steps': 23200, 'loss/train': 2.010055661201477} 01/29/2022 18:43:52 - INFO - codeparrot_training - Step 23201: {'lr': 0.0002955427928046653, 'samples': 4454784, 'steps': 23201, 'loss/train': 1.7041840553283691} 01/29/2022 18:43:57 - INFO - codeparrot_training - Step 23202: {'lr': 0.0002955267040412029, 'samples': 4454976, 'steps': 23202, 'loss/train': 1.703305184841156} 01/29/2022 18:44:02 - INFO - codeparrot_training - Step 23203: {'lr': 0.0002955106150827185, 'samples': 4455168, 'steps': 23203, 'loss/train': 2.1972561478614807} 01/29/2022 18:44:07 - INFO - codeparrot_training - Step 23204: {'lr': 0.0002954945259292811, 'samples': 4455360, 'steps': 23204, 'loss/train': 0.8391433954238892} 01/29/2022 18:44:11 - INFO - codeparrot_training - Step 23205: {'lr': 0.0002954784365809596, 'samples': 4455552, 'steps': 23205, 'loss/train': 2.0140337347984314} 01/29/2022 18:44:15 - INFO - codeparrot_training - Step 23206: {'lr': 0.0002954623470378229, 'samples': 4455744, 'steps': 23206, 'loss/train': 0.3081681579351425} 01/29/2022 18:44:19 - INFO - codeparrot_training - Step 23207: {'lr': 0.00029544625729993997, 'samples': 4455936, 'steps': 23207, 'loss/train': 1.7021459341049194} 01/29/2022 18:44:27 - INFO - codeparrot_training - Step 23208: {'lr': 0.0002954301673673797, 'samples': 4456128, 'steps': 23208, 'loss/train': 1.9546167254447937} 01/29/2022 18:44:31 - INFO - codeparrot_training - Step 23209: {'lr': 0.00029541407724021095, 'samples': 4456320, 'steps': 23209, 'loss/train': 1.6733815670013428} 01/29/2022 18:44:36 - INFO - codeparrot_training - Step 23210: {'lr': 0.00029539798691850274, 'samples': 4456512, 'steps': 23210, 'loss/train': 0.9961519539356232} 01/29/2022 18:44:40 - INFO - codeparrot_training - Step 23211: {'lr': 0.00029538189640232406, 'samples': 4456704, 'steps': 23211, 'loss/train': 1.0113953053951263} 01/29/2022 18:44:44 - INFO - codeparrot_training - Step 23212: {'lr': 0.00029536580569174364, 'samples': 4456896, 'steps': 23212, 'loss/train': 1.078325629234314} 01/29/2022 18:44:49 - INFO - codeparrot_training - Step 23213: {'lr': 0.0002953497147868306, 'samples': 4457088, 'steps': 23213, 'loss/train': 1.4259792566299438} 01/29/2022 18:44:53 - INFO - codeparrot_training - Step 23214: {'lr': 0.0002953336236876537, 'samples': 4457280, 'steps': 23214, 'loss/train': 2.5550538897514343} 01/29/2022 18:44:58 - INFO - codeparrot_training - Step 23215: {'lr': 0.000295317532394282, 'samples': 4457472, 'steps': 23215, 'loss/train': 0.9792680740356445} 01/29/2022 18:45:02 - INFO - codeparrot_training - Step 23216: {'lr': 0.0002953014409067844, 'samples': 4457664, 'steps': 23216, 'loss/train': 1.2682044804096222} 01/29/2022 18:45:06 - INFO - codeparrot_training - Step 23217: {'lr': 0.00029528534922522974, 'samples': 4457856, 'steps': 23217, 'loss/train': 1.8119970560073853} 01/29/2022 18:45:14 - INFO - codeparrot_training - Step 23218: {'lr': 0.00029526925734968705, 'samples': 4458048, 'steps': 23218, 'loss/train': 1.2626051902770996} 01/29/2022 18:45:18 - INFO - codeparrot_training - Step 23219: {'lr': 0.0002952531652802252, 'samples': 4458240, 'steps': 23219, 'loss/train': 1.4585267901420593} 01/29/2022 18:45:22 - INFO - codeparrot_training - Step 23220: {'lr': 0.00029523707301691327, 'samples': 4458432, 'steps': 23220, 'loss/train': 1.9848253726959229} 01/29/2022 18:45:26 - INFO - codeparrot_training - Step 23221: {'lr': 0.00029522098055982, 'samples': 4458624, 'steps': 23221, 'loss/train': 2.4584553837776184} 01/29/2022 18:45:30 - INFO - codeparrot_training - Step 23222: {'lr': 0.00029520488790901446, 'samples': 4458816, 'steps': 23222, 'loss/train': 1.193576842546463} 01/29/2022 18:45:36 - INFO - codeparrot_training - Step 23223: {'lr': 0.00029518879506456556, 'samples': 4459008, 'steps': 23223, 'loss/train': 1.2331117987632751} 01/29/2022 18:45:40 - INFO - codeparrot_training - Step 23224: {'lr': 0.0002951727020265421, 'samples': 4459200, 'steps': 23224, 'loss/train': 1.643989384174347} 01/29/2022 18:45:44 - INFO - codeparrot_training - Step 23225: {'lr': 0.0002951566087950132, 'samples': 4459392, 'steps': 23225, 'loss/train': 1.7927902936935425} 01/29/2022 18:45:49 - INFO - codeparrot_training - Step 23226: {'lr': 0.0002951405153700477, 'samples': 4459584, 'steps': 23226, 'loss/train': 1.6314463019371033} 01/29/2022 18:45:53 - INFO - codeparrot_training - Step 23227: {'lr': 0.0002951244217517146, 'samples': 4459776, 'steps': 23227, 'loss/train': 1.9097882509231567} 01/29/2022 18:45:58 - INFO - codeparrot_training - Step 23228: {'lr': 0.0002951083279400828, 'samples': 4459968, 'steps': 23228, 'loss/train': 1.0329337120056152} 01/29/2022 18:46:02 - INFO - codeparrot_training - Step 23229: {'lr': 0.00029509223393522124, 'samples': 4460160, 'steps': 23229, 'loss/train': 0.8805965781211853} 01/29/2022 18:46:07 - INFO - codeparrot_training - Step 23230: {'lr': 0.00029507613973719883, 'samples': 4460352, 'steps': 23230, 'loss/train': 1.7975038290023804} 01/29/2022 18:46:11 - INFO - codeparrot_training - Step 23231: {'lr': 0.0002950600453460846, 'samples': 4460544, 'steps': 23231, 'loss/train': 0.7211261093616486} 01/29/2022 18:46:15 - INFO - codeparrot_training - Step 23232: {'lr': 0.0002950439507619474, 'samples': 4460736, 'steps': 23232, 'loss/train': 0.39314796030521393} 01/29/2022 18:46:22 - INFO - codeparrot_training - Step 23233: {'lr': 0.00029502785598485624, 'samples': 4460928, 'steps': 23233, 'loss/train': 0.741805762052536} 01/29/2022 18:46:26 - INFO - codeparrot_training - Step 23234: {'lr': 0.00029501176101488004, 'samples': 4461120, 'steps': 23234, 'loss/train': 1.6169832944869995} 01/29/2022 18:46:31 - INFO - codeparrot_training - Step 23235: {'lr': 0.00029499566585208776, 'samples': 4461312, 'steps': 23235, 'loss/train': 1.8583529591560364} 01/29/2022 18:46:35 - INFO - codeparrot_training - Step 23236: {'lr': 0.00029497957049654834, 'samples': 4461504, 'steps': 23236, 'loss/train': 1.118738830089569} 01/29/2022 18:46:39 - INFO - codeparrot_training - Step 23237: {'lr': 0.0002949634749483307, 'samples': 4461696, 'steps': 23237, 'loss/train': 0.2517586424946785} 01/29/2022 18:46:44 - INFO - codeparrot_training - Step 23238: {'lr': 0.0002949473792075039, 'samples': 4461888, 'steps': 23238, 'loss/train': 1.4251549243927002} 01/29/2022 18:46:48 - INFO - codeparrot_training - Step 23239: {'lr': 0.00029493128327413664, 'samples': 4462080, 'steps': 23239, 'loss/train': 1.3065955638885498} 01/29/2022 18:46:53 - INFO - codeparrot_training - Step 23240: {'lr': 0.0002949151871482982, 'samples': 4462272, 'steps': 23240, 'loss/train': 1.6747446656227112} 01/29/2022 18:46:57 - INFO - codeparrot_training - Step 23241: {'lr': 0.00029489909083005723, 'samples': 4462464, 'steps': 23241, 'loss/train': 1.3836702704429626} 01/29/2022 18:47:01 - INFO - codeparrot_training - Step 23242: {'lr': 0.0002948829943194829, 'samples': 4462656, 'steps': 23242, 'loss/train': 1.32831010222435} 01/29/2022 18:47:09 - INFO - codeparrot_training - Step 23243: {'lr': 0.00029486689761664396, 'samples': 4462848, 'steps': 23243, 'loss/train': 2.1221668124198914} 01/29/2022 18:47:13 - INFO - codeparrot_training - Step 23244: {'lr': 0.0002948508007216096, 'samples': 4463040, 'steps': 23244, 'loss/train': 1.1438261568546295} 01/29/2022 18:47:17 - INFO - codeparrot_training - Step 23245: {'lr': 0.00029483470363444856, 'samples': 4463232, 'steps': 23245, 'loss/train': 1.6109785437583923} 01/29/2022 18:47:21 - INFO - codeparrot_training - Step 23246: {'lr': 0.00029481860635522994, 'samples': 4463424, 'steps': 23246, 'loss/train': 1.984331488609314} 01/29/2022 18:47:26 - INFO - codeparrot_training - Step 23247: {'lr': 0.0002948025088840226, 'samples': 4463616, 'steps': 23247, 'loss/train': 1.5337965488433838} 01/29/2022 18:47:31 - INFO - codeparrot_training - Step 23248: {'lr': 0.0002947864112208956, 'samples': 4463808, 'steps': 23248, 'loss/train': 1.7255542874336243} 01/29/2022 18:47:35 - INFO - codeparrot_training - Step 23249: {'lr': 0.0002947703133659178, 'samples': 4464000, 'steps': 23249, 'loss/train': 2.4917845129966736} 01/29/2022 18:47:39 - INFO - codeparrot_training - Step 23250: {'lr': 0.00029475421531915827, 'samples': 4464192, 'steps': 23250, 'loss/train': 1.0447227358818054} 01/29/2022 18:47:43 - INFO - codeparrot_training - Step 23251: {'lr': 0.00029473811708068576, 'samples': 4464384, 'steps': 23251, 'loss/train': 0.9773289263248444} 01/29/2022 18:47:48 - INFO - codeparrot_training - Step 23252: {'lr': 0.0002947220186505694, 'samples': 4464576, 'steps': 23252, 'loss/train': 1.086462289094925} 01/29/2022 18:47:53 - INFO - codeparrot_training - Step 23253: {'lr': 0.00029470592002887815, 'samples': 4464768, 'steps': 23253, 'loss/train': 1.6926743388175964} 01/29/2022 18:47:57 - INFO - codeparrot_training - Step 23254: {'lr': 0.00029468982121568096, 'samples': 4464960, 'steps': 23254, 'loss/train': 1.1080255508422852} 01/29/2022 18:48:01 - INFO - codeparrot_training - Step 23255: {'lr': 0.0002946737222110467, 'samples': 4465152, 'steps': 23255, 'loss/train': 1.4396816790103912} 01/29/2022 18:48:06 - INFO - codeparrot_training - Step 23256: {'lr': 0.0002946576230150444, 'samples': 4465344, 'steps': 23256, 'loss/train': 1.1534215807914734} 01/29/2022 18:48:10 - INFO - codeparrot_training - Step 23257: {'lr': 0.00029464152362774305, 'samples': 4465536, 'steps': 23257, 'loss/train': 1.4394406378269196} 01/29/2022 18:48:15 - INFO - codeparrot_training - Step 23258: {'lr': 0.00029462542404921156, 'samples': 4465728, 'steps': 23258, 'loss/train': 1.5305541157722473} 01/29/2022 18:48:19 - INFO - codeparrot_training - Step 23259: {'lr': 0.00029460932427951897, 'samples': 4465920, 'steps': 23259, 'loss/train': 2.0810272693634033} 01/29/2022 18:48:23 - INFO - codeparrot_training - Step 23260: {'lr': 0.00029459322431873416, 'samples': 4466112, 'steps': 23260, 'loss/train': 1.840428650379181} 01/29/2022 18:48:28 - INFO - codeparrot_training - Step 23261: {'lr': 0.00029457712416692617, 'samples': 4466304, 'steps': 23261, 'loss/train': 1.1708111464977264} 01/29/2022 18:48:32 - INFO - codeparrot_training - Step 23262: {'lr': 0.0002945610238241639, 'samples': 4466496, 'steps': 23262, 'loss/train': 1.6881346106529236} 01/29/2022 18:48:39 - INFO - codeparrot_training - Step 23263: {'lr': 0.0002945449232905164, 'samples': 4466688, 'steps': 23263, 'loss/train': 2.1594282388687134} 01/29/2022 18:48:43 - INFO - codeparrot_training - Step 23264: {'lr': 0.0002945288225660525, 'samples': 4466880, 'steps': 23264, 'loss/train': 1.6376920938491821} 01/29/2022 18:48:48 - INFO - codeparrot_training - Step 23265: {'lr': 0.00029451272165084144, 'samples': 4467072, 'steps': 23265, 'loss/train': 1.9564931988716125} 01/29/2022 18:48:52 - INFO - codeparrot_training - Step 23266: {'lr': 0.00029449662054495184, 'samples': 4467264, 'steps': 23266, 'loss/train': 1.4839029908180237} 01/29/2022 18:48:56 - INFO - codeparrot_training - Step 23267: {'lr': 0.0002944805192484529, 'samples': 4467456, 'steps': 23267, 'loss/train': 1.2274959683418274} 01/29/2022 18:49:02 - INFO - codeparrot_training - Step 23268: {'lr': 0.0002944644177614136, 'samples': 4467648, 'steps': 23268, 'loss/train': 1.8114550709724426} 01/29/2022 18:49:06 - INFO - codeparrot_training - Step 23269: {'lr': 0.00029444831608390276, 'samples': 4467840, 'steps': 23269, 'loss/train': 1.6136836409568787} 01/29/2022 18:49:10 - INFO - codeparrot_training - Step 23270: {'lr': 0.0002944322142159895, 'samples': 4468032, 'steps': 23270, 'loss/train': 1.7151551842689514} 01/29/2022 18:49:14 - INFO - codeparrot_training - Step 23271: {'lr': 0.0002944161121577427, 'samples': 4468224, 'steps': 23271, 'loss/train': 1.9233795404434204} 01/29/2022 18:49:18 - INFO - codeparrot_training - Step 23272: {'lr': 0.0002944000099092314, 'samples': 4468416, 'steps': 23272, 'loss/train': 1.7427728176116943} 01/29/2022 18:49:24 - INFO - codeparrot_training - Step 23273: {'lr': 0.0002943839074705246, 'samples': 4468608, 'steps': 23273, 'loss/train': 1.8949243426322937} 01/29/2022 18:49:28 - INFO - codeparrot_training - Step 23274: {'lr': 0.00029436780484169125, 'samples': 4468800, 'steps': 23274, 'loss/train': 2.0702097415924072} 01/29/2022 18:49:32 - INFO - codeparrot_training - Step 23275: {'lr': 0.0002943517020228003, 'samples': 4468992, 'steps': 23275, 'loss/train': 0.8338253796100616} 01/29/2022 18:49:36 - INFO - codeparrot_training - Step 23276: {'lr': 0.00029433559901392067, 'samples': 4469184, 'steps': 23276, 'loss/train': 1.8925509452819824} 01/29/2022 18:49:40 - INFO - codeparrot_training - Step 23277: {'lr': 0.0002943194958151214, 'samples': 4469376, 'steps': 23277, 'loss/train': 0.9840590357780457} 01/29/2022 18:49:48 - INFO - codeparrot_training - Step 23278: {'lr': 0.00029430339242647157, 'samples': 4469568, 'steps': 23278, 'loss/train': 1.6704506278038025} 01/29/2022 18:49:53 - INFO - codeparrot_training - Step 23279: {'lr': 0.00029428728884804, 'samples': 4469760, 'steps': 23279, 'loss/train': 0.9872241318225861} 01/29/2022 18:49:57 - INFO - codeparrot_training - Step 23280: {'lr': 0.0002942711850798959, 'samples': 4469952, 'steps': 23280, 'loss/train': 1.8851932883262634} 01/29/2022 18:50:01 - INFO - codeparrot_training - Step 23281: {'lr': 0.00029425508112210794, 'samples': 4470144, 'steps': 23281, 'loss/train': 1.5206989645957947} 01/29/2022 18:50:05 - INFO - codeparrot_training - Step 23282: {'lr': 0.00029423897697474535, 'samples': 4470336, 'steps': 23282, 'loss/train': 0.7305037379264832} 01/29/2022 18:50:10 - INFO - codeparrot_training - Step 23283: {'lr': 0.00029422287263787696, 'samples': 4470528, 'steps': 23283, 'loss/train': 1.5482491850852966} 01/29/2022 18:50:15 - INFO - codeparrot_training - Step 23284: {'lr': 0.0002942067681115719, 'samples': 4470720, 'steps': 23284, 'loss/train': 2.070922315120697} 01/29/2022 18:50:19 - INFO - codeparrot_training - Step 23285: {'lr': 0.0002941906633958989, 'samples': 4470912, 'steps': 23285, 'loss/train': 0.38009458780288696} 01/29/2022 18:50:23 - INFO - codeparrot_training - Step 23286: {'lr': 0.0002941745584909274, 'samples': 4471104, 'steps': 23286, 'loss/train': 1.2880283296108246} 01/29/2022 18:50:27 - INFO - codeparrot_training - Step 23287: {'lr': 0.00029415845339672596, 'samples': 4471296, 'steps': 23287, 'loss/train': 0.8239124715328217} 01/29/2022 18:50:33 - INFO - codeparrot_training - Step 23288: {'lr': 0.00029414234811336377, 'samples': 4471488, 'steps': 23288, 'loss/train': 1.5126171112060547} 01/29/2022 18:50:37 - INFO - codeparrot_training - Step 23289: {'lr': 0.00029412624264090983, 'samples': 4471680, 'steps': 23289, 'loss/train': 0.28543784469366074} 01/29/2022 18:50:41 - INFO - codeparrot_training - Step 23290: {'lr': 0.00029411013697943294, 'samples': 4471872, 'steps': 23290, 'loss/train': 2.0454028844833374} 01/29/2022 18:50:45 - INFO - codeparrot_training - Step 23291: {'lr': 0.00029409403112900227, 'samples': 4472064, 'steps': 23291, 'loss/train': 0.943126380443573} 01/29/2022 18:50:49 - INFO - codeparrot_training - Step 23292: {'lr': 0.00029407792508968683, 'samples': 4472256, 'steps': 23292, 'loss/train': 1.8001209497451782} 01/29/2022 18:50:57 - INFO - codeparrot_training - Step 23293: {'lr': 0.0002940618188615555, 'samples': 4472448, 'steps': 23293, 'loss/train': 2.1383376717567444} 01/29/2022 18:51:01 - INFO - codeparrot_training - Step 23294: {'lr': 0.0002940457124446774, 'samples': 4472640, 'steps': 23294, 'loss/train': 1.557030200958252} 01/29/2022 18:51:05 - INFO - codeparrot_training - Step 23295: {'lr': 0.0002940296058391214, 'samples': 4472832, 'steps': 23295, 'loss/train': 1.4101864993572235} 01/29/2022 18:51:09 - INFO - codeparrot_training - Step 23296: {'lr': 0.00029401349904495654, 'samples': 4473024, 'steps': 23296, 'loss/train': 1.62547767162323} 01/29/2022 18:51:15 - INFO - codeparrot_training - Step 23297: {'lr': 0.00029399739206225186, 'samples': 4473216, 'steps': 23297, 'loss/train': 1.8235584497451782} 01/29/2022 18:51:19 - INFO - codeparrot_training - Step 23298: {'lr': 0.00029398128489107635, 'samples': 4473408, 'steps': 23298, 'loss/train': 1.682007372379303} 01/29/2022 18:51:23 - INFO - codeparrot_training - Step 23299: {'lr': 0.0002939651775314989, 'samples': 4473600, 'steps': 23299, 'loss/train': 1.9396677017211914} 01/29/2022 18:51:27 - INFO - codeparrot_training - Step 23300: {'lr': 0.0002939490699835887, 'samples': 4473792, 'steps': 23300, 'loss/train': 1.7669304013252258} 01/29/2022 18:51:32 - INFO - codeparrot_training - Step 23301: {'lr': 0.0002939329622474145, 'samples': 4473984, 'steps': 23301, 'loss/train': 2.681039571762085} 01/29/2022 18:51:36 - INFO - codeparrot_training - Step 23302: {'lr': 0.00029391685432304554, 'samples': 4474176, 'steps': 23302, 'loss/train': 1.3685748875141144} 01/29/2022 18:51:43 - INFO - codeparrot_training - Step 23303: {'lr': 0.0002939007462105507, 'samples': 4474368, 'steps': 23303, 'loss/train': 1.1285144984722137} 01/29/2022 18:51:48 - INFO - codeparrot_training - Step 23304: {'lr': 0.000293884637909999, 'samples': 4474560, 'steps': 23304, 'loss/train': 0.9999127686023712} 01/29/2022 18:51:52 - INFO - codeparrot_training - Step 23305: {'lr': 0.00029386852942145944, 'samples': 4474752, 'steps': 23305, 'loss/train': 0.5849349349737167} 01/29/2022 18:51:56 - INFO - codeparrot_training - Step 23306: {'lr': 0.000293852420745001, 'samples': 4474944, 'steps': 23306, 'loss/train': 1.226283073425293} 01/29/2022 18:52:00 - INFO - codeparrot_training - Step 23307: {'lr': 0.0002938363118806928, 'samples': 4475136, 'steps': 23307, 'loss/train': 1.9939870834350586} 01/29/2022 18:52:06 - INFO - codeparrot_training - Step 23308: {'lr': 0.00029382020282860366, 'samples': 4475328, 'steps': 23308, 'loss/train': 1.4510125815868378} 01/29/2022 18:52:10 - INFO - codeparrot_training - Step 23309: {'lr': 0.00029380409358880277, 'samples': 4475520, 'steps': 23309, 'loss/train': 1.4119104444980621} 01/29/2022 18:52:14 - INFO - codeparrot_training - Step 23310: {'lr': 0.000293787984161359, 'samples': 4475712, 'steps': 23310, 'loss/train': 1.4977566003799438} 01/29/2022 18:52:18 - INFO - codeparrot_training - Step 23311: {'lr': 0.0002937718745463414, 'samples': 4475904, 'steps': 23311, 'loss/train': 2.011587381362915} 01/29/2022 18:52:24 - INFO - codeparrot_training - Step 23312: {'lr': 0.00029375576474381903, 'samples': 4476096, 'steps': 23312, 'loss/train': 1.610706388950348} 01/29/2022 18:52:28 - INFO - codeparrot_training - Step 23313: {'lr': 0.0002937396547538609, 'samples': 4476288, 'steps': 23313, 'loss/train': 1.1237922012805939} 01/29/2022 18:52:32 - INFO - codeparrot_training - Step 23314: {'lr': 0.00029372354457653585, 'samples': 4476480, 'steps': 23314, 'loss/train': 7.362727403640747} 01/29/2022 18:52:36 - INFO - codeparrot_training - Step 23315: {'lr': 0.00029370743421191305, 'samples': 4476672, 'steps': 23315, 'loss/train': 2.5959951281547546} 01/29/2022 18:52:40 - INFO - codeparrot_training - Step 23316: {'lr': 0.0002936913236600616, 'samples': 4476864, 'steps': 23316, 'loss/train': 1.9263877272605896} 01/29/2022 18:52:45 - INFO - codeparrot_training - Step 23317: {'lr': 0.0002936752129210503, 'samples': 4477056, 'steps': 23317, 'loss/train': 1.5166170001029968} 01/29/2022 18:52:52 - INFO - codeparrot_training - Step 23318: {'lr': 0.00029365910199494823, 'samples': 4477248, 'steps': 23318, 'loss/train': 3.065377950668335} 01/29/2022 18:52:57 - INFO - codeparrot_training - Step 23319: {'lr': 0.00029364299088182445, 'samples': 4477440, 'steps': 23319, 'loss/train': 3.064932703971863} 01/29/2022 18:53:01 - INFO - codeparrot_training - Step 23320: {'lr': 0.000293626879581748, 'samples': 4477632, 'steps': 23320, 'loss/train': 1.877602458000183} 01/29/2022 18:53:05 - INFO - codeparrot_training - Step 23321: {'lr': 0.00029361076809478784, 'samples': 4477824, 'steps': 23321, 'loss/train': 1.4004883468151093} 01/29/2022 18:53:09 - INFO - codeparrot_training - Step 23322: {'lr': 0.00029359465642101297, 'samples': 4478016, 'steps': 23322, 'loss/train': 1.3742044866085052} 01/29/2022 18:53:15 - INFO - codeparrot_training - Step 23323: {'lr': 0.0002935785445604924, 'samples': 4478208, 'steps': 23323, 'loss/train': 1.5735953450202942} 01/29/2022 18:53:19 - INFO - codeparrot_training - Step 23324: {'lr': 0.0002935624325132953, 'samples': 4478400, 'steps': 23324, 'loss/train': 1.1457508206367493} 01/29/2022 18:53:23 - INFO - codeparrot_training - Step 23325: {'lr': 0.0002935463202794905, 'samples': 4478592, 'steps': 23325, 'loss/train': 1.3250703513622284} 01/29/2022 18:53:27 - INFO - codeparrot_training - Step 23326: {'lr': 0.00029353020785914707, 'samples': 4478784, 'steps': 23326, 'loss/train': 1.8481310606002808} 01/29/2022 18:53:31 - INFO - codeparrot_training - Step 23327: {'lr': 0.00029351409525233416, 'samples': 4478976, 'steps': 23327, 'loss/train': 1.072963535785675} 01/29/2022 18:53:37 - INFO - codeparrot_training - Step 23328: {'lr': 0.0002934979824591205, 'samples': 4479168, 'steps': 23328, 'loss/train': 1.936540961265564} 01/29/2022 18:53:41 - INFO - codeparrot_training - Step 23329: {'lr': 0.0002934818694795755, 'samples': 4479360, 'steps': 23329, 'loss/train': 1.6431094408035278} 01/29/2022 18:53:45 - INFO - codeparrot_training - Step 23330: {'lr': 0.0002934657563137678, 'samples': 4479552, 'steps': 23330, 'loss/train': 1.199665904045105} 01/29/2022 18:53:49 - INFO - codeparrot_training - Step 23331: {'lr': 0.00029344964296176666, 'samples': 4479744, 'steps': 23331, 'loss/train': 1.3808862268924713} 01/29/2022 18:53:53 - INFO - codeparrot_training - Step 23332: {'lr': 0.00029343352942364106, 'samples': 4479936, 'steps': 23332, 'loss/train': 2.2962589859962463} 01/29/2022 18:53:59 - INFO - codeparrot_training - Step 23333: {'lr': 0.00029341741569946007, 'samples': 4480128, 'steps': 23333, 'loss/train': 2.3964579105377197} 01/29/2022 18:54:03 - INFO - codeparrot_training - Step 23334: {'lr': 0.00029340130178929255, 'samples': 4480320, 'steps': 23334, 'loss/train': 1.3287532031536102} 01/29/2022 18:54:07 - INFO - codeparrot_training - Step 23335: {'lr': 0.0002933851876932077, 'samples': 4480512, 'steps': 23335, 'loss/train': 0.911943107843399} 01/29/2022 18:54:12 - INFO - codeparrot_training - Step 23336: {'lr': 0.00029336907341127443, 'samples': 4480704, 'steps': 23336, 'loss/train': 2.4040390849113464} 01/29/2022 18:54:16 - INFO - codeparrot_training - Step 23337: {'lr': 0.0002933529589435619, 'samples': 4480896, 'steps': 23337, 'loss/train': 1.0912826657295227} 01/29/2022 18:54:20 - INFO - codeparrot_training - Step 23338: {'lr': 0.000293336844290139, 'samples': 4481088, 'steps': 23338, 'loss/train': 0.9360928237438202} 01/29/2022 18:54:27 - INFO - codeparrot_training - Step 23339: {'lr': 0.0002933207294510748, 'samples': 4481280, 'steps': 23339, 'loss/train': 4.495097279548645} 01/29/2022 18:54:32 - INFO - codeparrot_training - Step 23340: {'lr': 0.0002933046144264384, 'samples': 4481472, 'steps': 23340, 'loss/train': 2.0485569834709167} 01/29/2022 18:54:36 - INFO - codeparrot_training - Step 23341: {'lr': 0.0002932884992162987, 'samples': 4481664, 'steps': 23341, 'loss/train': 1.872409164905548} 01/29/2022 18:54:40 - INFO - codeparrot_training - Step 23342: {'lr': 0.00029327238382072495, 'samples': 4481856, 'steps': 23342, 'loss/train': 1.8960413932800293} 01/29/2022 18:54:44 - INFO - codeparrot_training - Step 23343: {'lr': 0.000293256268239786, 'samples': 4482048, 'steps': 23343, 'loss/train': 1.773486614227295} 01/29/2022 18:54:48 - INFO - codeparrot_training - Step 23344: {'lr': 0.00029324015247355093, 'samples': 4482240, 'steps': 23344, 'loss/train': 1.8087866306304932} 01/29/2022 18:54:54 - INFO - codeparrot_training - Step 23345: {'lr': 0.0002932240365220887, 'samples': 4482432, 'steps': 23345, 'loss/train': 1.599815011024475} 01/29/2022 18:54:58 - INFO - codeparrot_training - Step 23346: {'lr': 0.00029320792038546855, 'samples': 4482624, 'steps': 23346, 'loss/train': 1.7377837300300598} 01/29/2022 18:55:02 - INFO - codeparrot_training - Step 23347: {'lr': 0.0002931918040637593, 'samples': 4482816, 'steps': 23347, 'loss/train': 1.5408373475074768} 01/29/2022 18:55:07 - INFO - codeparrot_training - Step 23348: {'lr': 0.0002931756875570301, 'samples': 4483008, 'steps': 23348, 'loss/train': 0.7731020450592041} 01/29/2022 18:55:15 - INFO - codeparrot_training - Step 23349: {'lr': 0.0002931595708653499, 'samples': 4483200, 'steps': 23349, 'loss/train': 2.102031111717224} 01/29/2022 18:55:19 - INFO - codeparrot_training - Step 23350: {'lr': 0.00029314345398878796, 'samples': 4483392, 'steps': 23350, 'loss/train': 2.515945851802826} 01/29/2022 18:55:23 - INFO - codeparrot_training - Step 23351: {'lr': 0.00029312733692741307, 'samples': 4483584, 'steps': 23351, 'loss/train': 1.2912079095840454} 01/29/2022 18:55:27 - INFO - codeparrot_training - Step 23352: {'lr': 0.00029311121968129435, 'samples': 4483776, 'steps': 23352, 'loss/train': 2.1166725754737854} 01/29/2022 18:55:31 - INFO - codeparrot_training - Step 23353: {'lr': 0.00029309510225050087, 'samples': 4483968, 'steps': 23353, 'loss/train': 1.8947714567184448} 01/29/2022 18:55:36 - INFO - codeparrot_training - Step 23354: {'lr': 0.00029307898463510164, 'samples': 4484160, 'steps': 23354, 'loss/train': 1.6648356914520264} 01/29/2022 18:55:40 - INFO - codeparrot_training - Step 23355: {'lr': 0.00029306286683516573, 'samples': 4484352, 'steps': 23355, 'loss/train': 1.5453450679779053} 01/29/2022 18:55:45 - INFO - codeparrot_training - Step 23356: {'lr': 0.00029304674885076215, 'samples': 4484544, 'steps': 23356, 'loss/train': 1.0176503956317902} 01/29/2022 18:55:49 - INFO - codeparrot_training - Step 23357: {'lr': 0.00029303063068196006, 'samples': 4484736, 'steps': 23357, 'loss/train': 1.9858978986740112} 01/29/2022 18:55:53 - INFO - codeparrot_training - Step 23358: {'lr': 0.0002930145123288283, 'samples': 4484928, 'steps': 23358, 'loss/train': 1.7870965003967285} 01/29/2022 18:55:57 - INFO - codeparrot_training - Step 23359: {'lr': 0.00029299839379143613, 'samples': 4485120, 'steps': 23359, 'loss/train': 2.196174681186676} 01/29/2022 18:56:02 - INFO - codeparrot_training - Step 23360: {'lr': 0.0002929822750698524, 'samples': 4485312, 'steps': 23360, 'loss/train': 1.3759199380874634} 01/29/2022 18:56:07 - INFO - codeparrot_training - Step 23361: {'lr': 0.0002929661561641463, 'samples': 4485504, 'steps': 23361, 'loss/train': 2.1584295630455017} 01/29/2022 18:56:11 - INFO - codeparrot_training - Step 23362: {'lr': 0.0002929500370743868, 'samples': 4485696, 'steps': 23362, 'loss/train': 1.3625726401805878} 01/29/2022 18:56:15 - INFO - codeparrot_training - Step 23363: {'lr': 0.000292933917800643, 'samples': 4485888, 'steps': 23363, 'loss/train': 1.2797605991363525} 01/29/2022 18:56:19 - INFO - codeparrot_training - Step 23364: {'lr': 0.0002929177983429839, 'samples': 4486080, 'steps': 23364, 'loss/train': 2.8687732815742493} 01/29/2022 18:56:26 - INFO - codeparrot_training - Step 23365: {'lr': 0.0002929016787014786, 'samples': 4486272, 'steps': 23365, 'loss/train': 2.208058476448059} 01/29/2022 18:56:30 - INFO - codeparrot_training - Step 23366: {'lr': 0.0002928855588761962, 'samples': 4486464, 'steps': 23366, 'loss/train': 2.0985467433929443} 01/29/2022 18:56:35 - INFO - codeparrot_training - Step 23367: {'lr': 0.0002928694388672056, 'samples': 4486656, 'steps': 23367, 'loss/train': 1.923707127571106} 01/29/2022 18:56:39 - INFO - codeparrot_training - Step 23368: {'lr': 0.00029285331867457597, 'samples': 4486848, 'steps': 23368, 'loss/train': 2.032481074333191} 01/29/2022 18:56:43 - INFO - codeparrot_training - Step 23369: {'lr': 0.00029283719829837636, 'samples': 4487040, 'steps': 23369, 'loss/train': 1.8166193962097168} 01/29/2022 18:56:49 - INFO - codeparrot_training - Step 23370: {'lr': 0.0002928210777386757, 'samples': 4487232, 'steps': 23370, 'loss/train': 0.4718962758779526} 01/29/2022 18:56:53 - INFO - codeparrot_training - Step 23371: {'lr': 0.0002928049569955432, 'samples': 4487424, 'steps': 23371, 'loss/train': 1.201136827468872} 01/29/2022 18:56:57 - INFO - codeparrot_training - Step 23372: {'lr': 0.0002927888360690478, 'samples': 4487616, 'steps': 23372, 'loss/train': 1.5234006643295288} 01/29/2022 18:57:01 - INFO - codeparrot_training - Step 23373: {'lr': 0.00029277271495925874, 'samples': 4487808, 'steps': 23373, 'loss/train': 1.2596802413463593} 01/29/2022 18:57:06 - INFO - codeparrot_training - Step 23374: {'lr': 0.00029275659366624493, 'samples': 4488000, 'steps': 23374, 'loss/train': 0.7350783348083496} 01/29/2022 18:57:11 - INFO - codeparrot_training - Step 23375: {'lr': 0.00029274047219007534, 'samples': 4488192, 'steps': 23375, 'loss/train': 0.9206145107746124} 01/29/2022 18:57:15 - INFO - codeparrot_training - Step 23376: {'lr': 0.00029272435053081917, 'samples': 4488384, 'steps': 23376, 'loss/train': 1.7407103776931763} 01/29/2022 18:57:20 - INFO - codeparrot_training - Step 23377: {'lr': 0.0002927082286885455, 'samples': 4488576, 'steps': 23377, 'loss/train': 1.094814956188202} 01/29/2022 18:57:24 - INFO - codeparrot_training - Step 23378: {'lr': 0.0002926921066633233, 'samples': 4488768, 'steps': 23378, 'loss/train': 1.7803650498390198} 01/29/2022 18:57:28 - INFO - codeparrot_training - Step 23379: {'lr': 0.0002926759844552217, 'samples': 4488960, 'steps': 23379, 'loss/train': 2.378867447376251} 01/29/2022 18:57:32 - INFO - codeparrot_training - Step 23380: {'lr': 0.0002926598620643097, 'samples': 4489152, 'steps': 23380, 'loss/train': 0.060898203402757645} 01/29/2022 18:57:39 - INFO - codeparrot_training - Step 23381: {'lr': 0.0002926437394906564, 'samples': 4489344, 'steps': 23381, 'loss/train': 1.8095946907997131} 01/29/2022 18:57:43 - INFO - codeparrot_training - Step 23382: {'lr': 0.00029262761673433087, 'samples': 4489536, 'steps': 23382, 'loss/train': 0.8861842453479767} 01/29/2022 18:57:48 - INFO - codeparrot_training - Step 23383: {'lr': 0.0002926114937954022, 'samples': 4489728, 'steps': 23383, 'loss/train': 1.2276331186294556} 01/29/2022 18:57:52 - INFO - codeparrot_training - Step 23384: {'lr': 0.0002925953706739394, 'samples': 4489920, 'steps': 23384, 'loss/train': 1.8839187026023865} 01/29/2022 18:57:56 - INFO - codeparrot_training - Step 23385: {'lr': 0.0002925792473700116, 'samples': 4490112, 'steps': 23385, 'loss/train': 1.9476351141929626} 01/29/2022 18:58:01 - INFO - codeparrot_training - Step 23386: {'lr': 0.00029256312388368773, 'samples': 4490304, 'steps': 23386, 'loss/train': 1.7674071192741394} 01/29/2022 18:58:05 - INFO - codeparrot_training - Step 23387: {'lr': 0.000292547000215037, 'samples': 4490496, 'steps': 23387, 'loss/train': 1.41392183303833} 01/29/2022 18:58:10 - INFO - codeparrot_training - Step 23388: {'lr': 0.0002925308763641284, 'samples': 4490688, 'steps': 23388, 'loss/train': 2.2168931365013123} 01/29/2022 18:58:14 - INFO - codeparrot_training - Step 23389: {'lr': 0.00029251475233103104, 'samples': 4490880, 'steps': 23389, 'loss/train': 1.4965595304965973} 01/29/2022 18:58:18 - INFO - codeparrot_training - Step 23390: {'lr': 0.000292498628115814, 'samples': 4491072, 'steps': 23390, 'loss/train': 1.8176636695861816} 01/29/2022 18:58:23 - INFO - codeparrot_training - Step 23391: {'lr': 0.00029248250371854636, 'samples': 4491264, 'steps': 23391, 'loss/train': 2.2741209268569946} 01/29/2022 18:58:28 - INFO - codeparrot_training - Step 23392: {'lr': 0.0002924663791392971, 'samples': 4491456, 'steps': 23392, 'loss/train': 1.4217088222503662} 01/29/2022 18:58:32 - INFO - codeparrot_training - Step 23393: {'lr': 0.0002924502543781354, 'samples': 4491648, 'steps': 23393, 'loss/train': 2.0711830258369446} 01/29/2022 18:58:36 - INFO - codeparrot_training - Step 23394: {'lr': 0.0002924341294351303, 'samples': 4491840, 'steps': 23394, 'loss/train': 1.3741532564163208} 01/29/2022 18:58:40 - INFO - codeparrot_training - Step 23395: {'lr': 0.0002924180043103508, 'samples': 4492032, 'steps': 23395, 'loss/train': 1.5941179990768433} 01/29/2022 18:58:47 - INFO - codeparrot_training - Step 23396: {'lr': 0.0002924018790038662, 'samples': 4492224, 'steps': 23396, 'loss/train': 5.665289282798767} 01/29/2022 18:58:51 - INFO - codeparrot_training - Step 23397: {'lr': 0.0002923857535157452, 'samples': 4492416, 'steps': 23397, 'loss/train': 1.8933770656585693} 01/29/2022 18:58:56 - INFO - codeparrot_training - Step 23398: {'lr': 0.00029236962784605727, 'samples': 4492608, 'steps': 23398, 'loss/train': 1.6870286464691162} 01/29/2022 18:59:00 - INFO - codeparrot_training - Step 23399: {'lr': 0.00029235350199487115, 'samples': 4492800, 'steps': 23399, 'loss/train': 1.9260896444320679} 01/29/2022 18:59:04 - INFO - codeparrot_training - Step 23400: {'lr': 0.0002923373759622561, 'samples': 4492992, 'steps': 23400, 'loss/train': 1.4178017377853394} 01/29/2022 18:59:09 - INFO - codeparrot_training - Step 23401: {'lr': 0.00029232124974828126, 'samples': 4493184, 'steps': 23401, 'loss/train': 2.1675020456314087} 01/29/2022 18:59:14 - INFO - codeparrot_training - Step 23402: {'lr': 0.0002923051233530156, 'samples': 4493376, 'steps': 23402, 'loss/train': 1.398179680109024} 01/29/2022 18:59:18 - INFO - codeparrot_training - Step 23403: {'lr': 0.00029228899677652826, 'samples': 4493568, 'steps': 23403, 'loss/train': 1.1035617291927338} 01/29/2022 18:59:22 - INFO - codeparrot_training - Step 23404: {'lr': 0.0002922728700188882, 'samples': 4493760, 'steps': 23404, 'loss/train': 0.8353758752346039} 01/29/2022 18:59:26 - INFO - codeparrot_training - Step 23405: {'lr': 0.00029225674308016464, 'samples': 4493952, 'steps': 23405, 'loss/train': 0.5655635297298431} 01/29/2022 18:59:31 - INFO - codeparrot_training - Step 23406: {'lr': 0.00029224061596042663, 'samples': 4494144, 'steps': 23406, 'loss/train': 1.6903476119041443} 01/29/2022 18:59:35 - INFO - codeparrot_training - Step 23407: {'lr': 0.0002922244886597432, 'samples': 4494336, 'steps': 23407, 'loss/train': 1.7021645307540894} 01/29/2022 18:59:40 - INFO - codeparrot_training - Step 23408: {'lr': 0.00029220836117818346, 'samples': 4494528, 'steps': 23408, 'loss/train': 0.03564136661589146} 01/29/2022 18:59:44 - INFO - codeparrot_training - Step 23409: {'lr': 0.00029219223351581653, 'samples': 4494720, 'steps': 23409, 'loss/train': 2.229494333267212} 01/29/2022 18:59:48 - INFO - codeparrot_training - Step 23410: {'lr': 0.00029217610567271147, 'samples': 4494912, 'steps': 23410, 'loss/train': 2.0820257663726807} 01/29/2022 18:59:55 - INFO - codeparrot_training - Step 23411: {'lr': 0.00029215997764893734, 'samples': 4495104, 'steps': 23411, 'loss/train': 2.163722813129425} 01/29/2022 18:59:59 - INFO - codeparrot_training - Step 23412: {'lr': 0.00029214384944456327, 'samples': 4495296, 'steps': 23412, 'loss/train': 1.4853903651237488} 01/29/2022 19:00:04 - INFO - codeparrot_training - Step 23413: {'lr': 0.00029212772105965837, 'samples': 4495488, 'steps': 23413, 'loss/train': 1.4210830628871918} 01/29/2022 19:00:08 - INFO - codeparrot_training - Step 23414: {'lr': 0.0002921115924942916, 'samples': 4495680, 'steps': 23414, 'loss/train': 1.2766062319278717} 01/29/2022 19:00:12 - INFO - codeparrot_training - Step 23415: {'lr': 0.0002920954637485322, 'samples': 4495872, 'steps': 23415, 'loss/train': 1.5876399278640747} 01/29/2022 19:00:17 - INFO - codeparrot_training - Step 23416: {'lr': 0.0002920793348224493, 'samples': 4496064, 'steps': 23416, 'loss/train': 1.630038321018219} 01/29/2022 19:00:22 - INFO - codeparrot_training - Step 23417: {'lr': 0.0002920632057161117, 'samples': 4496256, 'steps': 23417, 'loss/train': 2.072730839252472} 01/29/2022 19:00:26 - INFO - codeparrot_training - Step 23418: {'lr': 0.00029204707642958876, 'samples': 4496448, 'steps': 23418, 'loss/train': 1.281270056962967} 01/29/2022 19:00:30 - INFO - codeparrot_training - Step 23419: {'lr': 0.0002920309469629495, 'samples': 4496640, 'steps': 23419, 'loss/train': 1.8569276332855225} 01/29/2022 19:00:34 - INFO - codeparrot_training - Step 23420: {'lr': 0.000292014817316263, 'samples': 4496832, 'steps': 23420, 'loss/train': 0.9491633176803589} 01/29/2022 19:00:41 - INFO - codeparrot_training - Step 23421: {'lr': 0.00029199868748959836, 'samples': 4497024, 'steps': 23421, 'loss/train': 1.0134774148464203} 01/29/2022 19:00:46 - INFO - codeparrot_training - Step 23422: {'lr': 0.00029198255748302473, 'samples': 4497216, 'steps': 23422, 'loss/train': 2.256779193878174} 01/29/2022 19:00:50 - INFO - codeparrot_training - Step 23423: {'lr': 0.00029196642729661113, 'samples': 4497408, 'steps': 23423, 'loss/train': 2.050343155860901} 01/29/2022 19:00:54 - INFO - codeparrot_training - Step 23424: {'lr': 0.0002919502969304266, 'samples': 4497600, 'steps': 23424, 'loss/train': 1.6850246787071228} 01/29/2022 19:00:58 - INFO - codeparrot_training - Step 23425: {'lr': 0.0002919341663845404, 'samples': 4497792, 'steps': 23425, 'loss/train': 1.2777585983276367} 01/29/2022 19:01:03 - INFO - codeparrot_training - Step 23426: {'lr': 0.00029191803565902153, 'samples': 4497984, 'steps': 23426, 'loss/train': 1.154289722442627} 01/29/2022 19:01:08 - INFO - codeparrot_training - Step 23427: {'lr': 0.00029190190475393915, 'samples': 4498176, 'steps': 23427, 'loss/train': 1.3202232420444489} 01/29/2022 19:01:12 - INFO - codeparrot_training - Step 23428: {'lr': 0.0002918857736693622, 'samples': 4498368, 'steps': 23428, 'loss/train': 1.4825566112995148} 01/29/2022 19:01:16 - INFO - codeparrot_training - Step 23429: {'lr': 0.00029186964240536, 'samples': 4498560, 'steps': 23429, 'loss/train': 2.413850426673889} 01/29/2022 19:01:20 - INFO - codeparrot_training - Step 23430: {'lr': 0.0002918535109620015, 'samples': 4498752, 'steps': 23430, 'loss/train': 0.3662059083580971} 01/29/2022 19:01:25 - INFO - codeparrot_training - Step 23431: {'lr': 0.0002918373793393559, 'samples': 4498944, 'steps': 23431, 'loss/train': 1.5947660207748413} 01/29/2022 19:01:30 - INFO - codeparrot_training - Step 23432: {'lr': 0.0002918212475374922, 'samples': 4499136, 'steps': 23432, 'loss/train': 1.6040446758270264} 01/29/2022 19:01:34 - INFO - codeparrot_training - Step 23433: {'lr': 0.0002918051155564796, 'samples': 4499328, 'steps': 23433, 'loss/train': 1.9761255383491516} 01/29/2022 19:01:38 - INFO - codeparrot_training - Step 23434: {'lr': 0.0002917889833963871, 'samples': 4499520, 'steps': 23434, 'loss/train': 2.336623191833496} 01/29/2022 19:01:42 - INFO - codeparrot_training - Step 23435: {'lr': 0.00029177285105728393, 'samples': 4499712, 'steps': 23435, 'loss/train': 1.6058651804924011} 01/29/2022 19:01:49 - INFO - codeparrot_training - Step 23436: {'lr': 0.00029175671853923907, 'samples': 4499904, 'steps': 23436, 'loss/train': 1.0747953057289124} 01/29/2022 19:01:53 - INFO - codeparrot_training - Step 23437: {'lr': 0.00029174058584232167, 'samples': 4500096, 'steps': 23437, 'loss/train': 1.886991262435913} 01/29/2022 19:01:58 - INFO - codeparrot_training - Step 23438: {'lr': 0.00029172445296660096, 'samples': 4500288, 'steps': 23438, 'loss/train': 1.4761460423469543} 01/29/2022 19:02:02 - INFO - codeparrot_training - Step 23439: {'lr': 0.0002917083199121459, 'samples': 4500480, 'steps': 23439, 'loss/train': 1.9325183629989624} 01/29/2022 19:02:06 - INFO - codeparrot_training - Step 23440: {'lr': 0.00029169218667902556, 'samples': 4500672, 'steps': 23440, 'loss/train': 1.462662398815155} 01/29/2022 19:02:11 - INFO - codeparrot_training - Step 23441: {'lr': 0.0002916760532673092, 'samples': 4500864, 'steps': 23441, 'loss/train': 1.06521937251091} 01/29/2022 19:02:16 - INFO - codeparrot_training - Step 23442: {'lr': 0.0002916599196770659, 'samples': 4501056, 'steps': 23442, 'loss/train': 1.634423553943634} 01/29/2022 19:02:20 - INFO - codeparrot_training - Step 23443: {'lr': 0.00029164378590836466, 'samples': 4501248, 'steps': 23443, 'loss/train': 2.208073318004608} 01/29/2022 19:02:24 - INFO - codeparrot_training - Step 23444: {'lr': 0.0002916276519612747, 'samples': 4501440, 'steps': 23444, 'loss/train': 1.4235206544399261} 01/29/2022 19:02:28 - INFO - codeparrot_training - Step 23445: {'lr': 0.00029161151783586506, 'samples': 4501632, 'steps': 23445, 'loss/train': 1.533090054988861} 01/29/2022 19:02:33 - INFO - codeparrot_training - Step 23446: {'lr': 0.0002915953835322049, 'samples': 4501824, 'steps': 23446, 'loss/train': 1.6342588663101196} 01/29/2022 19:02:37 - INFO - codeparrot_training - Step 23447: {'lr': 0.0002915792490503634, 'samples': 4502016, 'steps': 23447, 'loss/train': 0.733200877904892} 01/29/2022 19:02:42 - INFO - codeparrot_training - Step 23448: {'lr': 0.0002915631143904095, 'samples': 4502208, 'steps': 23448, 'loss/train': 1.6744832396507263} 01/29/2022 19:02:46 - INFO - codeparrot_training - Step 23449: {'lr': 0.0002915469795524125, 'samples': 4502400, 'steps': 23449, 'loss/train': 1.7502082586288452} 01/29/2022 19:02:50 - INFO - codeparrot_training - Step 23450: {'lr': 0.00029153084453644135, 'samples': 4502592, 'steps': 23450, 'loss/train': 1.2573341131210327} 01/29/2022 19:02:57 - INFO - codeparrot_training - Step 23451: {'lr': 0.0002915147093425653, 'samples': 4502784, 'steps': 23451, 'loss/train': 0.8749573230743408} 01/29/2022 19:03:01 - INFO - codeparrot_training - Step 23452: {'lr': 0.0002914985739708534, 'samples': 4502976, 'steps': 23452, 'loss/train': 1.524575114250183} 01/29/2022 19:03:06 - INFO - codeparrot_training - Step 23453: {'lr': 0.00029148243842137486, 'samples': 4503168, 'steps': 23453, 'loss/train': 1.3035745024681091} 01/29/2022 19:03:10 - INFO - codeparrot_training - Step 23454: {'lr': 0.0002914663026941986, 'samples': 4503360, 'steps': 23454, 'loss/train': 1.414121925830841} 01/29/2022 19:03:14 - INFO - codeparrot_training - Step 23455: {'lr': 0.00029145016678939394, 'samples': 4503552, 'steps': 23455, 'loss/train': 1.3035902380943298} 01/29/2022 19:03:19 - INFO - codeparrot_training - Step 23456: {'lr': 0.00029143403070702994, 'samples': 4503744, 'steps': 23456, 'loss/train': 1.3982045352458954} 01/29/2022 19:03:24 - INFO - codeparrot_training - Step 23457: {'lr': 0.0002914178944471757, 'samples': 4503936, 'steps': 23457, 'loss/train': 1.4357421398162842} 01/29/2022 19:03:28 - INFO - codeparrot_training - Step 23458: {'lr': 0.00029140175800990034, 'samples': 4504128, 'steps': 23458, 'loss/train': 1.547856867313385} 01/29/2022 19:03:32 - INFO - codeparrot_training - Step 23459: {'lr': 0.000291385621395273, 'samples': 4504320, 'steps': 23459, 'loss/train': 2.009180009365082} 01/29/2022 19:03:36 - INFO - codeparrot_training - Step 23460: {'lr': 0.0002913694846033628, 'samples': 4504512, 'steps': 23460, 'loss/train': 1.4154377281665802} 01/29/2022 19:03:42 - INFO - codeparrot_training - Step 23461: {'lr': 0.00029135334763423883, 'samples': 4504704, 'steps': 23461, 'loss/train': 1.318810522556305} 01/29/2022 19:03:46 - INFO - codeparrot_training - Step 23462: {'lr': 0.0002913372104879703, 'samples': 4504896, 'steps': 23462, 'loss/train': 2.385393261909485} 01/29/2022 19:03:50 - INFO - codeparrot_training - Step 23463: {'lr': 0.00029132107316462625, 'samples': 4505088, 'steps': 23463, 'loss/train': 1.3901743590831757} 01/29/2022 19:03:54 - INFO - codeparrot_training - Step 23464: {'lr': 0.00029130493566427586, 'samples': 4505280, 'steps': 23464, 'loss/train': 1.726375937461853} 01/29/2022 19:03:59 - INFO - codeparrot_training - Step 23465: {'lr': 0.0002912887979869882, 'samples': 4505472, 'steps': 23465, 'loss/train': 1.1280149817466736} 01/29/2022 19:04:06 - INFO - codeparrot_training - Step 23466: {'lr': 0.00029127266013283254, 'samples': 4505664, 'steps': 23466, 'loss/train': 1.5110414028167725} 01/29/2022 19:04:10 - INFO - codeparrot_training - Step 23467: {'lr': 0.0002912565221018778, 'samples': 4505856, 'steps': 23467, 'loss/train': 1.598167598247528} 01/29/2022 19:04:14 - INFO - codeparrot_training - Step 23468: {'lr': 0.00029124038389419325, 'samples': 4506048, 'steps': 23468, 'loss/train': 1.4244639873504639} 01/29/2022 19:04:19 - INFO - codeparrot_training - Step 23469: {'lr': 0.00029122424550984805, 'samples': 4506240, 'steps': 23469, 'loss/train': 2.148797035217285} 01/29/2022 19:04:23 - INFO - codeparrot_training - Step 23470: {'lr': 0.00029120810694891126, 'samples': 4506432, 'steps': 23470, 'loss/train': 1.1480562686920166} 01/29/2022 19:04:28 - INFO - codeparrot_training - Step 23471: {'lr': 0.00029119196821145193, 'samples': 4506624, 'steps': 23471, 'loss/train': 1.9497599601745605} 01/29/2022 19:04:32 - INFO - codeparrot_training - Step 23472: {'lr': 0.00029117582929753935, 'samples': 4506816, 'steps': 23472, 'loss/train': 1.742314338684082} 01/29/2022 19:04:36 - INFO - codeparrot_training - Step 23473: {'lr': 0.00029115969020724263, 'samples': 4507008, 'steps': 23473, 'loss/train': 0.9851269125938416} 01/29/2022 19:04:40 - INFO - codeparrot_training - Step 23474: {'lr': 0.0002911435509406308, 'samples': 4507200, 'steps': 23474, 'loss/train': 1.542772114276886} 01/29/2022 19:04:45 - INFO - codeparrot_training - Step 23475: {'lr': 0.0002911274114977731, 'samples': 4507392, 'steps': 23475, 'loss/train': 2.123226821422577} 01/29/2022 19:04:50 - INFO - codeparrot_training - Step 23476: {'lr': 0.0002911112718787386, 'samples': 4507584, 'steps': 23476, 'loss/train': 1.1383949518203735} 01/29/2022 19:04:54 - INFO - codeparrot_training - Step 23477: {'lr': 0.0002910951320835965, 'samples': 4507776, 'steps': 23477, 'loss/train': 1.2849175035953522} 01/29/2022 19:04:58 - INFO - codeparrot_training - Step 23478: {'lr': 0.0002910789921124159, 'samples': 4507968, 'steps': 23478, 'loss/train': 1.966187596321106} 01/29/2022 19:05:03 - INFO - codeparrot_training - Step 23479: {'lr': 0.00029106285196526594, 'samples': 4508160, 'steps': 23479, 'loss/train': 0.7822984457015991} 01/29/2022 19:05:07 - INFO - codeparrot_training - Step 23480: {'lr': 0.00029104671164221574, 'samples': 4508352, 'steps': 23480, 'loss/train': 1.988376259803772} 01/29/2022 19:05:14 - INFO - codeparrot_training - Step 23481: {'lr': 0.0002910305711433345, 'samples': 4508544, 'steps': 23481, 'loss/train': 1.0579802691936493} 01/29/2022 19:05:18 - INFO - codeparrot_training - Step 23482: {'lr': 0.00029101443046869127, 'samples': 4508736, 'steps': 23482, 'loss/train': 1.953819751739502} 01/29/2022 19:05:22 - INFO - codeparrot_training - Step 23483: {'lr': 0.0002909982896183553, 'samples': 4508928, 'steps': 23483, 'loss/train': 0.9124419987201691} 01/29/2022 19:05:27 - INFO - codeparrot_training - Step 23484: {'lr': 0.0002909821485923956, 'samples': 4509120, 'steps': 23484, 'loss/train': 2.545562446117401} 01/29/2022 19:05:31 - INFO - codeparrot_training - Step 23485: {'lr': 0.0002909660073908814, 'samples': 4509312, 'steps': 23485, 'loss/train': 2.3818620443344116} 01/29/2022 19:05:36 - INFO - codeparrot_training - Step 23486: {'lr': 0.00029094986601388195, 'samples': 4509504, 'steps': 23486, 'loss/train': 2.2068700790405273} 01/29/2022 19:05:40 - INFO - codeparrot_training - Step 23487: {'lr': 0.00029093372446146613, 'samples': 4509696, 'steps': 23487, 'loss/train': 1.413188874721527} 01/29/2022 19:05:44 - INFO - codeparrot_training - Step 23488: {'lr': 0.0002909175827337033, 'samples': 4509888, 'steps': 23488, 'loss/train': 1.9557716846466064} 01/29/2022 19:05:49 - INFO - codeparrot_training - Step 23489: {'lr': 0.00029090144083066254, 'samples': 4510080, 'steps': 23489, 'loss/train': 1.7378736734390259} 01/29/2022 19:05:53 - INFO - codeparrot_training - Step 23490: {'lr': 0.00029088529875241306, 'samples': 4510272, 'steps': 23490, 'loss/train': 1.7797958850860596} 01/29/2022 19:05:58 - INFO - codeparrot_training - Step 23491: {'lr': 0.00029086915649902383, 'samples': 4510464, 'steps': 23491, 'loss/train': 2.420977234840393} 01/29/2022 19:06:02 - INFO - codeparrot_training - Step 23492: {'lr': 0.0002908530140705642, 'samples': 4510656, 'steps': 23492, 'loss/train': 1.1180883944034576} 01/29/2022 19:06:06 - INFO - codeparrot_training - Step 23493: {'lr': 0.0002908368714671031, 'samples': 4510848, 'steps': 23493, 'loss/train': 2.1636418104171753} 01/29/2022 19:06:11 - INFO - codeparrot_training - Step 23494: {'lr': 0.00029082072868870997, 'samples': 4511040, 'steps': 23494, 'loss/train': 1.6476813554763794} 01/29/2022 19:06:15 - INFO - codeparrot_training - Step 23495: {'lr': 0.00029080458573545366, 'samples': 4511232, 'steps': 23495, 'loss/train': 0.6330626159906387} 01/29/2022 19:06:22 - INFO - codeparrot_training - Step 23496: {'lr': 0.0002907884426074036, 'samples': 4511424, 'steps': 23496, 'loss/train': 0.10907256603240967} 01/29/2022 19:06:26 - INFO - codeparrot_training - Step 23497: {'lr': 0.0002907722993046287, 'samples': 4511616, 'steps': 23497, 'loss/train': 1.4759130477905273} 01/29/2022 19:06:31 - INFO - codeparrot_training - Step 23498: {'lr': 0.0002907561558271983, 'samples': 4511808, 'steps': 23498, 'loss/train': 1.6287830471992493} 01/29/2022 19:06:35 - INFO - codeparrot_training - Step 23499: {'lr': 0.0002907400121751814, 'samples': 4512000, 'steps': 23499, 'loss/train': 1.3182591497898102} 01/29/2022 19:06:39 - INFO - codeparrot_training - Step 23500: {'lr': 0.0002907238683486472, 'samples': 4512192, 'steps': 23500, 'loss/train': 1.9748808145523071} 01/29/2022 19:06:44 - INFO - codeparrot_training - Step 23501: {'lr': 0.0002907077243476649, 'samples': 4512384, 'steps': 23501, 'loss/train': 1.4046264588832855} 01/29/2022 19:06:48 - INFO - codeparrot_training - Step 23502: {'lr': 0.0002906915801723037, 'samples': 4512576, 'steps': 23502, 'loss/train': 1.0986999571323395} 01/29/2022 19:06:53 - INFO - codeparrot_training - Step 23503: {'lr': 0.00029067543582263265, 'samples': 4512768, 'steps': 23503, 'loss/train': 2.2202517986297607} 01/29/2022 19:06:57 - INFO - codeparrot_training - Step 23504: {'lr': 0.00029065929129872095, 'samples': 4512960, 'steps': 23504, 'loss/train': 1.6832519173622131} 01/29/2022 19:07:01 - INFO - codeparrot_training - Step 23505: {'lr': 0.0002906431466006378, 'samples': 4513152, 'steps': 23505, 'loss/train': 1.800916850566864} 01/29/2022 19:07:06 - INFO - codeparrot_training - Step 23506: {'lr': 0.0002906270017284522, 'samples': 4513344, 'steps': 23506, 'loss/train': 2.2701990008354187} 01/29/2022 19:07:10 - INFO - codeparrot_training - Step 23507: {'lr': 0.0002906108566822335, 'samples': 4513536, 'steps': 23507, 'loss/train': 1.4367680847644806} 01/29/2022 19:07:15 - INFO - codeparrot_training - Step 23508: {'lr': 0.0002905947114620508, 'samples': 4513728, 'steps': 23508, 'loss/train': 1.2189191579818726} 01/29/2022 19:07:19 - INFO - codeparrot_training - Step 23509: {'lr': 0.0002905785660679732, 'samples': 4513920, 'steps': 23509, 'loss/train': 0.5008187144994736} 01/29/2022 19:07:23 - INFO - codeparrot_training - Step 23510: {'lr': 0.00029056242050006995, 'samples': 4514112, 'steps': 23510, 'loss/train': 0.8042642176151276} 01/29/2022 19:07:28 - INFO - codeparrot_training - Step 23511: {'lr': 0.0002905462747584101, 'samples': 4514304, 'steps': 23511, 'loss/train': 2.079659879207611} 01/29/2022 19:07:32 - INFO - codeparrot_training - Step 23512: {'lr': 0.00029053012884306295, 'samples': 4514496, 'steps': 23512, 'loss/train': 1.538853943347931} 01/29/2022 19:07:37 - INFO - codeparrot_training - Step 23513: {'lr': 0.00029051398275409756, 'samples': 4514688, 'steps': 23513, 'loss/train': 1.7433398365974426} 01/29/2022 19:07:41 - INFO - codeparrot_training - Step 23514: {'lr': 0.0002904978364915832, 'samples': 4514880, 'steps': 23514, 'loss/train': 2.0995720624923706} 01/29/2022 19:07:45 - INFO - codeparrot_training - Step 23515: {'lr': 0.0002904816900555889, 'samples': 4515072, 'steps': 23515, 'loss/train': 2.2261358499526978} 01/29/2022 19:07:55 - INFO - codeparrot_training - Step 23516: {'lr': 0.000290465543446184, 'samples': 4515264, 'steps': 23516, 'loss/train': 1.4080660343170166} 01/29/2022 19:08:00 - INFO - codeparrot_training - Step 23517: {'lr': 0.0002904493966634374, 'samples': 4515456, 'steps': 23517, 'loss/train': 2.0122350454330444} 01/29/2022 19:08:04 - INFO - codeparrot_training - Step 23518: {'lr': 0.0002904332497074186, 'samples': 4515648, 'steps': 23518, 'loss/train': 1.3460581004619598} 01/29/2022 19:08:08 - INFO - codeparrot_training - Step 23519: {'lr': 0.00029041710257819643, 'samples': 4515840, 'steps': 23519, 'loss/train': 2.007681369781494} 01/29/2022 19:08:12 - INFO - codeparrot_training - Step 23520: {'lr': 0.0002904009552758403, 'samples': 4516032, 'steps': 23520, 'loss/train': 1.9106823205947876} 01/29/2022 19:08:18 - INFO - codeparrot_training - Step 23521: {'lr': 0.0002903848078004193, 'samples': 4516224, 'steps': 23521, 'loss/train': 1.7514288425445557} 01/29/2022 19:08:22 - INFO - codeparrot_training - Step 23522: {'lr': 0.00029036866015200256, 'samples': 4516416, 'steps': 23522, 'loss/train': 1.829567313194275} 01/29/2022 19:08:26 - INFO - codeparrot_training - Step 23523: {'lr': 0.0002903525123306594, 'samples': 4516608, 'steps': 23523, 'loss/train': 1.5408071279525757} 01/29/2022 19:08:30 - INFO - codeparrot_training - Step 23524: {'lr': 0.0002903363643364588, 'samples': 4516800, 'steps': 23524, 'loss/train': 1.0977406203746796} 01/29/2022 19:08:34 - INFO - codeparrot_training - Step 23525: {'lr': 0.0002903202161694701, 'samples': 4516992, 'steps': 23525, 'loss/train': 1.4390417039394379} 01/29/2022 19:08:42 - INFO - codeparrot_training - Step 23526: {'lr': 0.0002903040678297624, 'samples': 4517184, 'steps': 23526, 'loss/train': 1.7109475135803223} 01/29/2022 19:08:46 - INFO - codeparrot_training - Step 23527: {'lr': 0.00029028791931740483, 'samples': 4517376, 'steps': 23527, 'loss/train': 1.2296702563762665} 01/29/2022 19:08:50 - INFO - codeparrot_training - Step 23528: {'lr': 0.0002902717706324666, 'samples': 4517568, 'steps': 23528, 'loss/train': 2.0463956594467163} 01/29/2022 19:08:54 - INFO - codeparrot_training - Step 23529: {'lr': 0.00029025562177501695, 'samples': 4517760, 'steps': 23529, 'loss/train': 1.7227948307991028} 01/29/2022 19:08:58 - INFO - codeparrot_training - Step 23530: {'lr': 0.0002902394727451249, 'samples': 4517952, 'steps': 23530, 'loss/train': 1.3101400136947632} 01/29/2022 19:09:04 - INFO - codeparrot_training - Step 23531: {'lr': 0.00029022332354285986, 'samples': 4518144, 'steps': 23531, 'loss/train': 1.4097506403923035} 01/29/2022 19:09:08 - INFO - codeparrot_training - Step 23532: {'lr': 0.00029020717416829075, 'samples': 4518336, 'steps': 23532, 'loss/train': 1.9474639892578125} 01/29/2022 19:09:12 - INFO - codeparrot_training - Step 23533: {'lr': 0.0002901910246214869, 'samples': 4518528, 'steps': 23533, 'loss/train': 0.5948719829320908} 01/29/2022 19:09:17 - INFO - codeparrot_training - Step 23534: {'lr': 0.0002901748749025175, 'samples': 4518720, 'steps': 23534, 'loss/train': 1.0492455661296844} 01/29/2022 19:09:21 - INFO - codeparrot_training - Step 23535: {'lr': 0.00029015872501145164, 'samples': 4518912, 'steps': 23535, 'loss/train': 2.3044983744621277} 01/29/2022 19:09:26 - INFO - codeparrot_training - Step 23536: {'lr': 0.00029014257494835863, 'samples': 4519104, 'steps': 23536, 'loss/train': 1.0451931059360504} 01/29/2022 19:09:30 - INFO - codeparrot_training - Step 23537: {'lr': 0.0002901264247133075, 'samples': 4519296, 'steps': 23537, 'loss/train': 1.7027868032455444} 01/29/2022 19:09:34 - INFO - codeparrot_training - Step 23538: {'lr': 0.00029011027430636755, 'samples': 4519488, 'steps': 23538, 'loss/train': 2.0271987318992615} 01/29/2022 19:09:39 - INFO - codeparrot_training - Step 23539: {'lr': 0.00029009412372760793, 'samples': 4519680, 'steps': 23539, 'loss/train': 0.6688603609800339} 01/29/2022 19:09:43 - INFO - codeparrot_training - Step 23540: {'lr': 0.00029007797297709784, 'samples': 4519872, 'steps': 23540, 'loss/train': 2.1581037640571594} 01/29/2022 19:09:50 - INFO - codeparrot_training - Step 23541: {'lr': 0.00029006182205490634, 'samples': 4520064, 'steps': 23541, 'loss/train': 1.0117990672588348} 01/29/2022 19:09:54 - INFO - codeparrot_training - Step 23542: {'lr': 0.0002900456709611028, 'samples': 4520256, 'steps': 23542, 'loss/train': 1.702073335647583} 01/29/2022 19:09:58 - INFO - codeparrot_training - Step 23543: {'lr': 0.0002900295196957563, 'samples': 4520448, 'steps': 23543, 'loss/train': 2.1143898367881775} 01/29/2022 19:10:02 - INFO - codeparrot_training - Step 23544: {'lr': 0.00029001336825893603, 'samples': 4520640, 'steps': 23544, 'loss/train': 1.314928561449051} 01/29/2022 19:10:07 - INFO - codeparrot_training - Step 23545: {'lr': 0.0002899972166507112, 'samples': 4520832, 'steps': 23545, 'loss/train': 1.609722912311554} 01/29/2022 19:10:12 - INFO - codeparrot_training - Step 23546: {'lr': 0.000289981064871151, 'samples': 4521024, 'steps': 23546, 'loss/train': 1.8978588581085205} 01/29/2022 19:10:16 - INFO - codeparrot_training - Step 23547: {'lr': 0.00028996491292032465, 'samples': 4521216, 'steps': 23547, 'loss/train': 0.7476308941841125} 01/29/2022 19:10:20 - INFO - codeparrot_training - Step 23548: {'lr': 0.00028994876079830125, 'samples': 4521408, 'steps': 23548, 'loss/train': 1.5288351774215698} 01/29/2022 19:10:24 - INFO - codeparrot_training - Step 23549: {'lr': 0.00028993260850515015, 'samples': 4521600, 'steps': 23549, 'loss/train': 0.5534289926290512} 01/29/2022 19:10:29 - INFO - codeparrot_training - Step 23550: {'lr': 0.0002899164560409403, 'samples': 4521792, 'steps': 23550, 'loss/train': 1.5859671235084534} 01/29/2022 19:10:36 - INFO - codeparrot_training - Step 23551: {'lr': 0.00028990030340574107, 'samples': 4521984, 'steps': 23551, 'loss/train': 2.0358683466911316} 01/29/2022 19:10:40 - INFO - codeparrot_training - Step 23552: {'lr': 0.0002898841505996216, 'samples': 4522176, 'steps': 23552, 'loss/train': 2.123853027820587} 01/29/2022 19:10:44 - INFO - codeparrot_training - Step 23553: {'lr': 0.0002898679976226511, 'samples': 4522368, 'steps': 23553, 'loss/train': 2.005721390247345} 01/29/2022 19:10:49 - INFO - codeparrot_training - Step 23554: {'lr': 0.00028985184447489874, 'samples': 4522560, 'steps': 23554, 'loss/train': 2.650689125061035} 01/29/2022 19:10:53 - INFO - codeparrot_training - Step 23555: {'lr': 0.00028983569115643385, 'samples': 4522752, 'steps': 23555, 'loss/train': 1.0001804530620575} 01/29/2022 19:10:58 - INFO - codeparrot_training - Step 23556: {'lr': 0.0002898195376673254, 'samples': 4522944, 'steps': 23556, 'loss/train': 1.5973536372184753} 01/29/2022 19:11:02 - INFO - codeparrot_training - Step 23557: {'lr': 0.0002898033840076427, 'samples': 4523136, 'steps': 23557, 'loss/train': 1.8827658891677856} 01/29/2022 19:11:06 - INFO - codeparrot_training - Step 23558: {'lr': 0.00028978723017745496, 'samples': 4523328, 'steps': 23558, 'loss/train': 1.9671898484230042} 01/29/2022 19:11:10 - INFO - codeparrot_training - Step 23559: {'lr': 0.0002897710761768313, 'samples': 4523520, 'steps': 23559, 'loss/train': 0.7853535711765289} 01/29/2022 19:11:15 - INFO - codeparrot_training - Step 23560: {'lr': 0.00028975492200584106, 'samples': 4523712, 'steps': 23560, 'loss/train': 2.0321706533432007} 01/29/2022 19:11:20 - INFO - codeparrot_training - Step 23561: {'lr': 0.00028973876766455334, 'samples': 4523904, 'steps': 23561, 'loss/train': 2.468182861804962} 01/29/2022 19:11:24 - INFO - codeparrot_training - Step 23562: {'lr': 0.00028972261315303736, 'samples': 4524096, 'steps': 23562, 'loss/train': 1.599661946296692} 01/29/2022 19:11:28 - INFO - codeparrot_training - Step 23563: {'lr': 0.0002897064584713623, 'samples': 4524288, 'steps': 23563, 'loss/train': 1.8730634450912476} 01/29/2022 19:11:32 - INFO - codeparrot_training - Step 23564: {'lr': 0.0002896903036195974, 'samples': 4524480, 'steps': 23564, 'loss/train': 0.20208900421857834} 01/29/2022 19:11:37 - INFO - codeparrot_training - Step 23565: {'lr': 0.0002896741485978118, 'samples': 4524672, 'steps': 23565, 'loss/train': 1.4883933663368225} 01/29/2022 19:11:44 - INFO - codeparrot_training - Step 23566: {'lr': 0.0002896579934060748, 'samples': 4524864, 'steps': 23566, 'loss/train': 1.0558072328567505} 01/29/2022 19:11:48 - INFO - codeparrot_training - Step 23567: {'lr': 0.00028964183804445554, 'samples': 4525056, 'steps': 23567, 'loss/train': 1.8807506561279297} 01/29/2022 19:11:52 - INFO - codeparrot_training - Step 23568: {'lr': 0.00028962568251302326, 'samples': 4525248, 'steps': 23568, 'loss/train': 2.109527885913849} 01/29/2022 19:11:56 - INFO - codeparrot_training - Step 23569: {'lr': 0.00028960952681184705, 'samples': 4525440, 'steps': 23569, 'loss/train': 2.0506719946861267} 01/29/2022 19:12:01 - INFO - codeparrot_training - Step 23570: {'lr': 0.00028959337094099617, 'samples': 4525632, 'steps': 23570, 'loss/train': 2.146498739719391} 01/29/2022 19:12:06 - INFO - codeparrot_training - Step 23571: {'lr': 0.00028957721490054, 'samples': 4525824, 'steps': 23571, 'loss/train': 2.682684302330017} 01/29/2022 19:12:10 - INFO - codeparrot_training - Step 23572: {'lr': 0.00028956105869054757, 'samples': 4526016, 'steps': 23572, 'loss/train': 1.5999755859375} 01/29/2022 19:12:14 - INFO - codeparrot_training - Step 23573: {'lr': 0.0002895449023110881, 'samples': 4526208, 'steps': 23573, 'loss/train': 1.8513761758804321} 01/29/2022 19:12:19 - INFO - codeparrot_training - Step 23574: {'lr': 0.00028952874576223083, 'samples': 4526400, 'steps': 23574, 'loss/train': 1.1775789856910706} 01/29/2022 19:12:23 - INFO - codeparrot_training - Step 23575: {'lr': 0.000289512589044045, 'samples': 4526592, 'steps': 23575, 'loss/train': 1.6110325455665588} 01/29/2022 19:12:28 - INFO - codeparrot_training - Step 23576: {'lr': 0.0002894964321565997, 'samples': 4526784, 'steps': 23576, 'loss/train': 2.0046944618225098} 01/29/2022 19:12:32 - INFO - codeparrot_training - Step 23577: {'lr': 0.00028948027509996435, 'samples': 4526976, 'steps': 23577, 'loss/train': 1.1311782896518707} 01/29/2022 19:12:36 - INFO - codeparrot_training - Step 23578: {'lr': 0.0002894641178742079, 'samples': 4527168, 'steps': 23578, 'loss/train': 2.17206734418869} 01/29/2022 19:12:41 - INFO - codeparrot_training - Step 23579: {'lr': 0.0002894479604793998, 'samples': 4527360, 'steps': 23579, 'loss/train': 1.757177710533142} 01/29/2022 19:12:45 - INFO - codeparrot_training - Step 23580: {'lr': 0.0002894318029156091, 'samples': 4527552, 'steps': 23580, 'loss/train': 2.186424493789673} 01/29/2022 19:12:50 - INFO - codeparrot_training - Step 23581: {'lr': 0.0002894156451829051, 'samples': 4527744, 'steps': 23581, 'loss/train': 1.0172752439975739} 01/29/2022 19:12:54 - INFO - codeparrot_training - Step 23582: {'lr': 0.000289399487281357, 'samples': 4527936, 'steps': 23582, 'loss/train': 1.7343057990074158} 01/29/2022 19:12:59 - INFO - codeparrot_training - Step 23583: {'lr': 0.000289383329211034, 'samples': 4528128, 'steps': 23583, 'loss/train': 1.3295190632343292} 01/29/2022 19:13:03 - INFO - codeparrot_training - Step 23584: {'lr': 0.0002893671709720054, 'samples': 4528320, 'steps': 23584, 'loss/train': 2.192379355430603} 01/29/2022 19:13:10 - INFO - codeparrot_training - Step 23585: {'lr': 0.0002893510125643403, 'samples': 4528512, 'steps': 23585, 'loss/train': 1.437978744506836} 01/29/2022 19:13:14 - INFO - codeparrot_training - Step 23586: {'lr': 0.0002893348539881079, 'samples': 4528704, 'steps': 23586, 'loss/train': 1.7104053497314453} 01/29/2022 19:13:18 - INFO - codeparrot_training - Step 23587: {'lr': 0.0002893186952433775, 'samples': 4528896, 'steps': 23587, 'loss/train': 1.288418859243393} 01/29/2022 19:13:23 - INFO - codeparrot_training - Step 23588: {'lr': 0.00028930253633021826, 'samples': 4529088, 'steps': 23588, 'loss/train': 1.702209234237671} 01/29/2022 19:13:27 - INFO - codeparrot_training - Step 23589: {'lr': 0.0002892863772486995, 'samples': 4529280, 'steps': 23589, 'loss/train': 1.1587603390216827} 01/29/2022 19:13:32 - INFO - codeparrot_training - Step 23590: {'lr': 0.0002892702179988903, 'samples': 4529472, 'steps': 23590, 'loss/train': 1.911069631576538} 01/29/2022 19:13:36 - INFO - codeparrot_training - Step 23591: {'lr': 0.00028925405858086007, 'samples': 4529664, 'steps': 23591, 'loss/train': 1.5337275266647339} 01/29/2022 19:13:40 - INFO - codeparrot_training - Step 23592: {'lr': 0.0002892378989946779, 'samples': 4529856, 'steps': 23592, 'loss/train': 2.29597145318985} 01/29/2022 19:13:45 - INFO - codeparrot_training - Step 23593: {'lr': 0.00028922173924041296, 'samples': 4530048, 'steps': 23593, 'loss/train': 1.511126697063446} 01/29/2022 19:13:49 - INFO - codeparrot_training - Step 23594: {'lr': 0.00028920557931813454, 'samples': 4530240, 'steps': 23594, 'loss/train': 2.927673876285553} 01/29/2022 19:13:56 - INFO - codeparrot_training - Step 23595: {'lr': 0.00028918941922791185, 'samples': 4530432, 'steps': 23595, 'loss/train': 1.7779861092567444} 01/29/2022 19:14:00 - INFO - codeparrot_training - Step 23596: {'lr': 0.00028917325896981417, 'samples': 4530624, 'steps': 23596, 'loss/train': 1.590164601802826} 01/29/2022 19:14:05 - INFO - codeparrot_training - Step 23597: {'lr': 0.0002891570985439108, 'samples': 4530816, 'steps': 23597, 'loss/train': 2.3961090445518494} 01/29/2022 19:14:09 - INFO - codeparrot_training - Step 23598: {'lr': 0.00028914093795027063, 'samples': 4531008, 'steps': 23598, 'loss/train': 1.22804456949234} 01/29/2022 19:14:13 - INFO - codeparrot_training - Step 23599: {'lr': 0.0002891247771889633, 'samples': 4531200, 'steps': 23599, 'loss/train': 0.6672954261302948} 01/29/2022 19:14:17 - INFO - codeparrot_training - Step 23600: {'lr': 0.00028910861626005774, 'samples': 4531392, 'steps': 23600, 'loss/train': 2.1550764441490173} 01/29/2022 19:14:23 - INFO - codeparrot_training - Step 23601: {'lr': 0.0002890924551636234, 'samples': 4531584, 'steps': 23601, 'loss/train': 1.7952173352241516} 01/29/2022 19:14:27 - INFO - codeparrot_training - Step 23602: {'lr': 0.00028907629389972924, 'samples': 4531776, 'steps': 23602, 'loss/train': 2.0719361901283264} 01/29/2022 19:14:31 - INFO - codeparrot_training - Step 23603: {'lr': 0.00028906013246844474, 'samples': 4531968, 'steps': 23603, 'loss/train': 1.9146162271499634} 01/29/2022 19:14:35 - INFO - codeparrot_training - Step 23604: {'lr': 0.000289043970869839, 'samples': 4532160, 'steps': 23604, 'loss/train': 1.4733530580997467} 01/29/2022 19:14:40 - INFO - codeparrot_training - Step 23605: {'lr': 0.0002890278091039813, 'samples': 4532352, 'steps': 23605, 'loss/train': 1.8041006326675415} 01/29/2022 19:14:45 - INFO - codeparrot_training - Step 23606: {'lr': 0.00028901164717094085, 'samples': 4532544, 'steps': 23606, 'loss/train': 1.6420056223869324} 01/29/2022 19:14:49 - INFO - codeparrot_training - Step 23607: {'lr': 0.0002889954850707869, 'samples': 4532736, 'steps': 23607, 'loss/train': 1.7639518976211548} 01/29/2022 19:14:53 - INFO - codeparrot_training - Step 23608: {'lr': 0.0002889793228035887, 'samples': 4532928, 'steps': 23608, 'loss/train': 1.0020563006401062} 01/29/2022 19:14:57 - INFO - codeparrot_training - Step 23609: {'lr': 0.0002889631603694154, 'samples': 4533120, 'steps': 23609, 'loss/train': 1.9632509350776672} 01/29/2022 19:15:02 - INFO - codeparrot_training - Step 23610: {'lr': 0.00028894699776833637, 'samples': 4533312, 'steps': 23610, 'loss/train': 2.0334352254867554} 01/29/2022 19:15:09 - INFO - codeparrot_training - Step 23611: {'lr': 0.0002889308350004207, 'samples': 4533504, 'steps': 23611, 'loss/train': 2.1508902311325073} 01/29/2022 19:15:13 - INFO - codeparrot_training - Step 23612: {'lr': 0.00028891467206573773, 'samples': 4533696, 'steps': 23612, 'loss/train': 2.1062373518943787} 01/29/2022 19:15:17 - INFO - codeparrot_training - Step 23613: {'lr': 0.00028889850896435666, 'samples': 4533888, 'steps': 23613, 'loss/train': 1.3582224547863007} 01/29/2022 19:15:21 - INFO - codeparrot_training - Step 23614: {'lr': 0.00028888234569634673, 'samples': 4534080, 'steps': 23614, 'loss/train': 1.3187231719493866} 01/29/2022 19:15:26 - INFO - codeparrot_training - Step 23615: {'lr': 0.00028886618226177716, 'samples': 4534272, 'steps': 23615, 'loss/train': 1.0428127348423004} 01/29/2022 19:15:31 - INFO - codeparrot_training - Step 23616: {'lr': 0.00028885001866071723, 'samples': 4534464, 'steps': 23616, 'loss/train': 1.4518929719924927} 01/29/2022 19:15:35 - INFO - codeparrot_training - Step 23617: {'lr': 0.0002888338548932361, 'samples': 4534656, 'steps': 23617, 'loss/train': 1.3859311044216156} 01/29/2022 19:15:39 - INFO - codeparrot_training - Step 23618: {'lr': 0.00028881769095940316, 'samples': 4534848, 'steps': 23618, 'loss/train': 1.0455459952354431} 01/29/2022 19:15:43 - INFO - codeparrot_training - Step 23619: {'lr': 0.0002888015268592875, 'samples': 4535040, 'steps': 23619, 'loss/train': 1.89580500125885} 01/29/2022 19:15:48 - INFO - codeparrot_training - Step 23620: {'lr': 0.0002887853625929584, 'samples': 4535232, 'steps': 23620, 'loss/train': 2.0361133217811584} 01/29/2022 19:15:53 - INFO - codeparrot_training - Step 23621: {'lr': 0.00028876919816048517, 'samples': 4535424, 'steps': 23621, 'loss/train': 1.0005371868610382} 01/29/2022 19:15:57 - INFO - codeparrot_training - Step 23622: {'lr': 0.00028875303356193697, 'samples': 4535616, 'steps': 23622, 'loss/train': 1.7976120114326477} 01/29/2022 19:16:01 - INFO - codeparrot_training - Step 23623: {'lr': 0.0002887368687973831, 'samples': 4535808, 'steps': 23623, 'loss/train': 1.565843939781189} 01/29/2022 19:16:06 - INFO - codeparrot_training - Step 23624: {'lr': 0.00028872070386689274, 'samples': 4536000, 'steps': 23624, 'loss/train': 1.8201919198036194} 01/29/2022 19:16:10 - INFO - codeparrot_training - Step 23625: {'lr': 0.00028870453877053527, 'samples': 4536192, 'steps': 23625, 'loss/train': 2.30889755487442} 01/29/2022 19:16:17 - INFO - codeparrot_training - Step 23626: {'lr': 0.00028868837350837977, 'samples': 4536384, 'steps': 23626, 'loss/train': 0.7231777310371399} 01/29/2022 19:16:21 - INFO - codeparrot_training - Step 23627: {'lr': 0.0002886722080804956, 'samples': 4536576, 'steps': 23627, 'loss/train': 0.1940544918179512} 01/29/2022 19:16:25 - INFO - codeparrot_training - Step 23628: {'lr': 0.000288656042486952, 'samples': 4536768, 'steps': 23628, 'loss/train': 2.192134737968445} 01/29/2022 19:16:29 - INFO - codeparrot_training - Step 23629: {'lr': 0.0002886398767278181, 'samples': 4536960, 'steps': 23629, 'loss/train': 1.3044951260089874} 01/29/2022 19:16:34 - INFO - codeparrot_training - Step 23630: {'lr': 0.0002886237108031633, 'samples': 4537152, 'steps': 23630, 'loss/train': 1.27476105093956} 01/29/2022 19:16:39 - INFO - codeparrot_training - Step 23631: {'lr': 0.0002886075447130568, 'samples': 4537344, 'steps': 23631, 'loss/train': 1.9324843883514404} 01/29/2022 19:16:43 - INFO - codeparrot_training - Step 23632: {'lr': 0.00028859137845756784, 'samples': 4537536, 'steps': 23632, 'loss/train': 1.5678538084030151} 01/29/2022 19:16:47 - INFO - codeparrot_training - Step 23633: {'lr': 0.00028857521203676565, 'samples': 4537728, 'steps': 23633, 'loss/train': 1.8970937132835388} 01/29/2022 19:16:51 - INFO - codeparrot_training - Step 23634: {'lr': 0.0002885590454507195, 'samples': 4537920, 'steps': 23634, 'loss/train': 2.4781769514083862} 01/29/2022 19:16:56 - INFO - codeparrot_training - Step 23635: {'lr': 0.00028854287869949856, 'samples': 4538112, 'steps': 23635, 'loss/train': 1.6633277535438538} 01/29/2022 19:17:01 - INFO - codeparrot_training - Step 23636: {'lr': 0.00028852671178317233, 'samples': 4538304, 'steps': 23636, 'loss/train': 1.2453313171863556} 01/29/2022 19:17:05 - INFO - codeparrot_training - Step 23637: {'lr': 0.00028851054470180977, 'samples': 4538496, 'steps': 23637, 'loss/train': 1.7195470333099365} 01/29/2022 19:17:09 - INFO - codeparrot_training - Step 23638: {'lr': 0.0002884943774554803, 'samples': 4538688, 'steps': 23638, 'loss/train': 0.6917873471975327} 01/29/2022 19:17:14 - INFO - codeparrot_training - Step 23639: {'lr': 0.0002884782100442531, 'samples': 4538880, 'steps': 23639, 'loss/train': 1.4804065525531769} 01/29/2022 19:17:18 - INFO - codeparrot_training - Step 23640: {'lr': 0.0002884620424681976, 'samples': 4539072, 'steps': 23640, 'loss/train': 1.6097263097763062} 01/29/2022 19:17:25 - INFO - codeparrot_training - Step 23641: {'lr': 0.0002884458747273828, 'samples': 4539264, 'steps': 23641, 'loss/train': 1.4032634496688843} 01/29/2022 19:17:29 - INFO - codeparrot_training - Step 23642: {'lr': 0.0002884297068218781, 'samples': 4539456, 'steps': 23642, 'loss/train': 1.9821245670318604} 01/29/2022 19:17:33 - INFO - codeparrot_training - Step 23643: {'lr': 0.00028841353875175274, 'samples': 4539648, 'steps': 23643, 'loss/train': 0.9634590446949005} 01/29/2022 19:17:38 - INFO - codeparrot_training - Step 23644: {'lr': 0.000288397370517076, 'samples': 4539840, 'steps': 23644, 'loss/train': 1.4657481908798218} 01/29/2022 19:17:42 - INFO - codeparrot_training - Step 23645: {'lr': 0.0002883812021179171, 'samples': 4540032, 'steps': 23645, 'loss/train': 1.4241676926612854} 01/29/2022 19:17:47 - INFO - codeparrot_training - Step 23646: {'lr': 0.0002883650335543453, 'samples': 4540224, 'steps': 23646, 'loss/train': 1.4703748226165771} 01/29/2022 19:17:51 - INFO - codeparrot_training - Step 23647: {'lr': 0.00028834886482643, 'samples': 4540416, 'steps': 23647, 'loss/train': 1.4033105671405792} 01/29/2022 19:17:55 - INFO - codeparrot_training - Step 23648: {'lr': 0.00028833269593424017, 'samples': 4540608, 'steps': 23648, 'loss/train': 1.2777427732944489} 01/29/2022 19:18:00 - INFO - codeparrot_training - Step 23649: {'lr': 0.0002883165268778454, 'samples': 4540800, 'steps': 23649, 'loss/train': 1.685349941253662} 01/29/2022 19:18:04 - INFO - codeparrot_training - Step 23650: {'lr': 0.00028830035765731464, 'samples': 4540992, 'steps': 23650, 'loss/train': 2.9324432015419006} 01/29/2022 19:18:09 - INFO - codeparrot_training - Step 23651: {'lr': 0.0002882841882727174, 'samples': 4541184, 'steps': 23651, 'loss/train': 1.2459416091442108} 01/29/2022 19:18:13 - INFO - codeparrot_training - Step 23652: {'lr': 0.00028826801872412284, 'samples': 4541376, 'steps': 23652, 'loss/train': 2.5213379859924316} 01/29/2022 19:18:17 - INFO - codeparrot_training - Step 23653: {'lr': 0.00028825184901160023, 'samples': 4541568, 'steps': 23653, 'loss/train': 2.100342035293579} 01/29/2022 19:18:22 - INFO - codeparrot_training - Step 23654: {'lr': 0.00028823567913521884, 'samples': 4541760, 'steps': 23654, 'loss/train': 2.1708794832229614} 01/29/2022 19:18:26 - INFO - codeparrot_training - Step 23655: {'lr': 0.0002882195090950479, 'samples': 4541952, 'steps': 23655, 'loss/train': 1.84884774684906} 01/29/2022 19:18:33 - INFO - codeparrot_training - Step 23656: {'lr': 0.00028820333889115684, 'samples': 4542144, 'steps': 23656, 'loss/train': 0.8114182949066162} 01/29/2022 19:18:38 - INFO - codeparrot_training - Step 23657: {'lr': 0.0002881871685236147, 'samples': 4542336, 'steps': 23657, 'loss/train': 0.4368104785680771} 01/29/2022 19:18:42 - INFO - codeparrot_training - Step 23658: {'lr': 0.0002881709979924909, 'samples': 4542528, 'steps': 23658, 'loss/train': 1.8903606534004211} 01/29/2022 19:18:46 - INFO - codeparrot_training - Step 23659: {'lr': 0.00028815482729785467, 'samples': 4542720, 'steps': 23659, 'loss/train': 0.898266077041626} 01/29/2022 19:18:50 - INFO - codeparrot_training - Step 23660: {'lr': 0.00028813865643977527, 'samples': 4542912, 'steps': 23660, 'loss/train': 1.8084413409233093} 01/29/2022 19:18:55 - INFO - codeparrot_training - Step 23661: {'lr': 0.000288122485418322, 'samples': 4543104, 'steps': 23661, 'loss/train': 1.5018451809883118} 01/29/2022 19:19:00 - INFO - codeparrot_training - Step 23662: {'lr': 0.0002881063142335641, 'samples': 4543296, 'steps': 23662, 'loss/train': 2.0607866048812866} 01/29/2022 19:19:04 - INFO - codeparrot_training - Step 23663: {'lr': 0.0002880901428855708, 'samples': 4543488, 'steps': 23663, 'loss/train': 0.8942987322807312} 01/29/2022 19:19:08 - INFO - codeparrot_training - Step 23664: {'lr': 0.0002880739713744114, 'samples': 4543680, 'steps': 23664, 'loss/train': 1.0721558332443237} 01/29/2022 19:19:12 - INFO - codeparrot_training - Step 23665: {'lr': 0.00028805779970015525, 'samples': 4543872, 'steps': 23665, 'loss/train': 0.6317817270755768} 01/29/2022 19:19:17 - INFO - codeparrot_training - Step 23666: {'lr': 0.00028804162786287156, 'samples': 4544064, 'steps': 23666, 'loss/train': 1.6330751180648804} 01/29/2022 19:19:22 - INFO - codeparrot_training - Step 23667: {'lr': 0.0002880254558626297, 'samples': 4544256, 'steps': 23667, 'loss/train': 1.716603398323059} 01/29/2022 19:19:26 - INFO - codeparrot_training - Step 23668: {'lr': 0.0002880092836994987, 'samples': 4544448, 'steps': 23668, 'loss/train': 0.9604514837265015} 01/29/2022 19:19:30 - INFO - codeparrot_training - Step 23669: {'lr': 0.0002879931113735482, 'samples': 4544640, 'steps': 23669, 'loss/train': 1.6209686994552612} 01/29/2022 19:19:34 - INFO - codeparrot_training - Step 23670: {'lr': 0.0002879769388848471, 'samples': 4544832, 'steps': 23670, 'loss/train': 2.439521849155426} 01/29/2022 19:19:41 - INFO - codeparrot_training - Step 23671: {'lr': 0.0002879607662334649, 'samples': 4545024, 'steps': 23671, 'loss/train': 1.1243546605110168} 01/29/2022 19:19:45 - INFO - codeparrot_training - Step 23672: {'lr': 0.0002879445934194709, 'samples': 4545216, 'steps': 23672, 'loss/train': 1.020241767168045} 01/29/2022 19:19:50 - INFO - codeparrot_training - Step 23673: {'lr': 0.0002879284204429342, 'samples': 4545408, 'steps': 23673, 'loss/train': 0.5058597028255463} 01/29/2022 19:19:54 - INFO - codeparrot_training - Step 23674: {'lr': 0.0002879122473039243, 'samples': 4545600, 'steps': 23674, 'loss/train': 2.0158562064170837} 01/29/2022 19:19:58 - INFO - codeparrot_training - Step 23675: {'lr': 0.0002878960740025103, 'samples': 4545792, 'steps': 23675, 'loss/train': 1.4889556467533112} 01/29/2022 19:20:04 - INFO - codeparrot_training - Step 23676: {'lr': 0.00028787990053876156, 'samples': 4545984, 'steps': 23676, 'loss/train': 1.0995689928531647} 01/29/2022 19:20:08 - INFO - codeparrot_training - Step 23677: {'lr': 0.00028786372691274735, 'samples': 4546176, 'steps': 23677, 'loss/train': 0.9373913705348969} 01/29/2022 19:20:12 - INFO - codeparrot_training - Step 23678: {'lr': 0.00028784755312453704, 'samples': 4546368, 'steps': 23678, 'loss/train': 1.635275959968567} 01/29/2022 19:20:16 - INFO - codeparrot_training - Step 23679: {'lr': 0.0002878313791741997, 'samples': 4546560, 'steps': 23679, 'loss/train': 1.503688931465149} 01/29/2022 19:20:20 - INFO - codeparrot_training - Step 23680: {'lr': 0.00028781520506180485, 'samples': 4546752, 'steps': 23680, 'loss/train': 1.2362979054450989} 01/29/2022 19:20:26 - INFO - codeparrot_training - Step 23681: {'lr': 0.0002877990307874216, 'samples': 4546944, 'steps': 23681, 'loss/train': 2.723007917404175} 01/29/2022 19:20:30 - INFO - codeparrot_training - Step 23682: {'lr': 0.00028778285635111935, 'samples': 4547136, 'steps': 23682, 'loss/train': 1.6047493815422058} 01/29/2022 19:20:34 - INFO - codeparrot_training - Step 23683: {'lr': 0.0002877666817529673, 'samples': 4547328, 'steps': 23683, 'loss/train': 1.4539070427417755} 01/29/2022 19:20:38 - INFO - codeparrot_training - Step 23684: {'lr': 0.0002877505069930348, 'samples': 4547520, 'steps': 23684, 'loss/train': 1.8380354046821594} 01/29/2022 19:20:43 - INFO - codeparrot_training - Step 23685: {'lr': 0.0002877343320713911, 'samples': 4547712, 'steps': 23685, 'loss/train': 2.0078691244125366} 01/29/2022 19:20:50 - INFO - codeparrot_training - Step 23686: {'lr': 0.0002877181569881055, 'samples': 4547904, 'steps': 23686, 'loss/train': 1.735472023487091} 01/29/2022 19:20:54 - INFO - codeparrot_training - Step 23687: {'lr': 0.00028770198174324737, 'samples': 4548096, 'steps': 23687, 'loss/train': 1.4380398094654083} 01/29/2022 19:20:58 - INFO - codeparrot_training - Step 23688: {'lr': 0.00028768580633688586, 'samples': 4548288, 'steps': 23688, 'loss/train': 1.53056001663208} 01/29/2022 19:21:02 - INFO - codeparrot_training - Step 23689: {'lr': 0.00028766963076909033, 'samples': 4548480, 'steps': 23689, 'loss/train': 1.2517696022987366} 01/29/2022 19:21:07 - INFO - codeparrot_training - Step 23690: {'lr': 0.00028765345503993, 'samples': 4548672, 'steps': 23690, 'loss/train': 1.582670509815216} 01/29/2022 19:21:12 - INFO - codeparrot_training - Step 23691: {'lr': 0.0002876372791494743, 'samples': 4548864, 'steps': 23691, 'loss/train': 1.9783621430397034} 01/29/2022 19:21:16 - INFO - codeparrot_training - Step 23692: {'lr': 0.00028762110309779246, 'samples': 4549056, 'steps': 23692, 'loss/train': 0.7423478364944458} 01/29/2022 19:21:20 - INFO - codeparrot_training - Step 23693: {'lr': 0.0002876049268849537, 'samples': 4549248, 'steps': 23693, 'loss/train': 1.3109329640865326} 01/29/2022 19:21:24 - INFO - codeparrot_training - Step 23694: {'lr': 0.0002875887505110274, 'samples': 4549440, 'steps': 23694, 'loss/train': 1.7189496159553528} 01/29/2022 19:21:29 - INFO - codeparrot_training - Step 23695: {'lr': 0.00028757257397608285, 'samples': 4549632, 'steps': 23695, 'loss/train': 0.1487370729446411} 01/29/2022 19:21:34 - INFO - codeparrot_training - Step 23696: {'lr': 0.0002875563972801893, 'samples': 4549824, 'steps': 23696, 'loss/train': 0.7647911310195923} 01/29/2022 19:21:38 - INFO - codeparrot_training - Step 23697: {'lr': 0.0002875402204234161, 'samples': 4550016, 'steps': 23697, 'loss/train': 0.8909223675727844} 01/29/2022 19:21:42 - INFO - codeparrot_training - Step 23698: {'lr': 0.0002875240434058324, 'samples': 4550208, 'steps': 23698, 'loss/train': 1.7898362874984741} 01/29/2022 19:21:46 - INFO - codeparrot_training - Step 23699: {'lr': 0.0002875078662275078, 'samples': 4550400, 'steps': 23699, 'loss/train': 2.1902164220809937} 01/29/2022 19:21:51 - INFO - codeparrot_training - Step 23700: {'lr': 0.00028749168888851125, 'samples': 4550592, 'steps': 23700, 'loss/train': 1.5495579242706299} 01/29/2022 19:21:58 - INFO - codeparrot_training - Step 23701: {'lr': 0.0002874755113889123, 'samples': 4550784, 'steps': 23701, 'loss/train': 1.3093827366828918} 01/29/2022 19:22:02 - INFO - codeparrot_training - Step 23702: {'lr': 0.0002874593337287801, 'samples': 4550976, 'steps': 23702, 'loss/train': 1.3923381865024567} 01/29/2022 19:22:06 - INFO - codeparrot_training - Step 23703: {'lr': 0.00028744315590818406, 'samples': 4551168, 'steps': 23703, 'loss/train': 0.7172324806451797} 01/29/2022 19:22:10 - INFO - codeparrot_training - Step 23704: {'lr': 0.0002874269779271934, 'samples': 4551360, 'steps': 23704, 'loss/train': 2.0920695662498474} 01/29/2022 19:22:15 - INFO - codeparrot_training - Step 23705: {'lr': 0.0002874107997858775, 'samples': 4551552, 'steps': 23705, 'loss/train': 1.9279126524925232} 01/29/2022 19:22:20 - INFO - codeparrot_training - Step 23706: {'lr': 0.00028739462148430554, 'samples': 4551744, 'steps': 23706, 'loss/train': 0.4689645320177078} 01/29/2022 19:22:24 - INFO - codeparrot_training - Step 23707: {'lr': 0.0002873784430225469, 'samples': 4551936, 'steps': 23707, 'loss/train': 1.6635469794273376} 01/29/2022 19:22:28 - INFO - codeparrot_training - Step 23708: {'lr': 0.000287362264400671, 'samples': 4552128, 'steps': 23708, 'loss/train': 1.9938942790031433} 01/29/2022 19:22:32 - INFO - codeparrot_training - Step 23709: {'lr': 0.00028734608561874686, 'samples': 4552320, 'steps': 23709, 'loss/train': 1.080033391714096} 01/29/2022 19:22:37 - INFO - codeparrot_training - Step 23710: {'lr': 0.0002873299066768441, 'samples': 4552512, 'steps': 23710, 'loss/train': 2.1379358768463135} 01/29/2022 19:22:42 - INFO - codeparrot_training - Step 23711: {'lr': 0.0002873137275750317, 'samples': 4552704, 'steps': 23711, 'loss/train': 1.4008140563964844} 01/29/2022 19:22:46 - INFO - codeparrot_training - Step 23712: {'lr': 0.0002872975483133793, 'samples': 4552896, 'steps': 23712, 'loss/train': 2.440418243408203} 01/29/2022 19:22:50 - INFO - codeparrot_training - Step 23713: {'lr': 0.00028728136889195595, 'samples': 4553088, 'steps': 23713, 'loss/train': 1.624002456665039} 01/29/2022 19:22:54 - INFO - codeparrot_training - Step 23714: {'lr': 0.0002872651893108311, 'samples': 4553280, 'steps': 23714, 'loss/train': 2.178945779800415} 01/29/2022 19:22:59 - INFO - codeparrot_training - Step 23715: {'lr': 0.000287249009570074, 'samples': 4553472, 'steps': 23715, 'loss/train': 1.3063002526760101} 01/29/2022 19:23:06 - INFO - codeparrot_training - Step 23716: {'lr': 0.000287232829669754, 'samples': 4553664, 'steps': 23716, 'loss/train': 1.0008048713207245} 01/29/2022 19:23:10 - INFO - codeparrot_training - Step 23717: {'lr': 0.0002872166496099403, 'samples': 4553856, 'steps': 23717, 'loss/train': 1.2476704716682434} 01/29/2022 19:23:14 - INFO - codeparrot_training - Step 23718: {'lr': 0.0002872004693907024, 'samples': 4554048, 'steps': 23718, 'loss/train': 1.477052092552185} 01/29/2022 19:23:18 - INFO - codeparrot_training - Step 23719: {'lr': 0.0002871842890121094, 'samples': 4554240, 'steps': 23719, 'loss/train': 4.027720928192139} 01/29/2022 19:23:23 - INFO - codeparrot_training - Step 23720: {'lr': 0.0002871681084742308, 'samples': 4554432, 'steps': 23720, 'loss/train': 1.648006796836853} 01/29/2022 19:23:28 - INFO - codeparrot_training - Step 23721: {'lr': 0.0002871519277771358, 'samples': 4554624, 'steps': 23721, 'loss/train': 1.3107689023017883} 01/29/2022 19:23:32 - INFO - codeparrot_training - Step 23722: {'lr': 0.0002871357469208937, 'samples': 4554816, 'steps': 23722, 'loss/train': 2.1247002482414246} 01/29/2022 19:23:36 - INFO - codeparrot_training - Step 23723: {'lr': 0.0002871195659055739, 'samples': 4555008, 'steps': 23723, 'loss/train': 1.9375355243682861} 01/29/2022 19:23:40 - INFO - codeparrot_training - Step 23724: {'lr': 0.0002871033847312456, 'samples': 4555200, 'steps': 23724, 'loss/train': 1.8329717516899109} 01/29/2022 19:23:45 - INFO - codeparrot_training - Step 23725: {'lr': 0.0002870872033979782, 'samples': 4555392, 'steps': 23725, 'loss/train': 0.8448777794837952} 01/29/2022 19:23:52 - INFO - codeparrot_training - Step 23726: {'lr': 0.000287071021905841, 'samples': 4555584, 'steps': 23726, 'loss/train': 1.02071812748909} 01/29/2022 19:23:56 - INFO - codeparrot_training - Step 23727: {'lr': 0.00028705484025490333, 'samples': 4555776, 'steps': 23727, 'loss/train': 1.5726622939109802} 01/29/2022 19:24:00 - INFO - codeparrot_training - Step 23728: {'lr': 0.0002870386584452345, 'samples': 4555968, 'steps': 23728, 'loss/train': 1.6203492879867554} 01/29/2022 19:24:05 - INFO - codeparrot_training - Step 23729: {'lr': 0.00028702247647690383, 'samples': 4556160, 'steps': 23729, 'loss/train': 1.7342959642410278} 01/29/2022 19:24:09 - INFO - codeparrot_training - Step 23730: {'lr': 0.0002870062943499806, 'samples': 4556352, 'steps': 23730, 'loss/train': 1.0411593317985535} 01/29/2022 19:24:14 - INFO - codeparrot_training - Step 23731: {'lr': 0.0002869901120645341, 'samples': 4556544, 'steps': 23731, 'loss/train': 0.5245552808046341} 01/29/2022 19:24:18 - INFO - codeparrot_training - Step 23732: {'lr': 0.0002869739296206338, 'samples': 4556736, 'steps': 23732, 'loss/train': 2.3226277828216553} 01/29/2022 19:24:22 - INFO - codeparrot_training - Step 23733: {'lr': 0.0002869577470183489, 'samples': 4556928, 'steps': 23733, 'loss/train': 1.2922874987125397} 01/29/2022 19:24:27 - INFO - codeparrot_training - Step 23734: {'lr': 0.00028694156425774874, 'samples': 4557120, 'steps': 23734, 'loss/train': 2.087547719478607} 01/29/2022 19:24:31 - INFO - codeparrot_training - Step 23735: {'lr': 0.00028692538133890267, 'samples': 4557312, 'steps': 23735, 'loss/train': 0.8362129926681519} 01/29/2022 19:24:36 - INFO - codeparrot_training - Step 23736: {'lr': 0.00028690919826188, 'samples': 4557504, 'steps': 23736, 'loss/train': 2.1496256589889526} 01/29/2022 19:24:40 - INFO - codeparrot_training - Step 23737: {'lr': 0.00028689301502674995, 'samples': 4557696, 'steps': 23737, 'loss/train': 1.5346455574035645} 01/29/2022 19:24:44 - INFO - codeparrot_training - Step 23738: {'lr': 0.00028687683163358197, 'samples': 4557888, 'steps': 23738, 'loss/train': 0.8892926573753357} 01/29/2022 19:24:49 - INFO - codeparrot_training - Step 23739: {'lr': 0.00028686064808244546, 'samples': 4558080, 'steps': 23739, 'loss/train': 2.0266106128692627} 01/29/2022 19:24:53 - INFO - codeparrot_training - Step 23740: {'lr': 0.0002868444643734095, 'samples': 4558272, 'steps': 23740, 'loss/train': 1.892072081565857} 01/29/2022 19:24:58 - INFO - codeparrot_training - Step 23741: {'lr': 0.00028682828050654365, 'samples': 4558464, 'steps': 23741, 'loss/train': 2.1624661087989807} 01/29/2022 19:25:02 - INFO - codeparrot_training - Step 23742: {'lr': 0.0002868120964819171, 'samples': 4558656, 'steps': 23742, 'loss/train': 1.6489394903182983} 01/29/2022 19:25:07 - INFO - codeparrot_training - Step 23743: {'lr': 0.0002867959122995992, 'samples': 4558848, 'steps': 23743, 'loss/train': 0.7138748914003372} 01/29/2022 19:25:11 - INFO - codeparrot_training - Step 23744: {'lr': 0.0002867797279596593, 'samples': 4559040, 'steps': 23744, 'loss/train': 0.12176494672894478} 01/29/2022 19:25:15 - INFO - codeparrot_training - Step 23745: {'lr': 0.0002867635434621668, 'samples': 4559232, 'steps': 23745, 'loss/train': 1.2917332649230957} 01/29/2022 19:25:22 - INFO - codeparrot_training - Step 23746: {'lr': 0.0002867473588071909, 'samples': 4559424, 'steps': 23746, 'loss/train': 1.7220810055732727} 01/29/2022 19:25:26 - INFO - codeparrot_training - Step 23747: {'lr': 0.00028673117399480096, 'samples': 4559616, 'steps': 23747, 'loss/train': 1.638944149017334} 01/29/2022 19:25:30 - INFO - codeparrot_training - Step 23748: {'lr': 0.00028671498902506636, 'samples': 4559808, 'steps': 23748, 'loss/train': 1.7566072940826416} 01/29/2022 19:25:35 - INFO - codeparrot_training - Step 23749: {'lr': 0.00028669880389805647, 'samples': 4560000, 'steps': 23749, 'loss/train': 1.0826734006404877} 01/29/2022 19:25:39 - INFO - codeparrot_training - Step 23750: {'lr': 0.00028668261861384045, 'samples': 4560192, 'steps': 23750, 'loss/train': 1.6565784215927124} 01/29/2022 19:25:44 - INFO - codeparrot_training - Step 23751: {'lr': 0.00028666643317248777, 'samples': 4560384, 'steps': 23751, 'loss/train': 2.1252753138542175} 01/29/2022 19:25:48 - INFO - codeparrot_training - Step 23752: {'lr': 0.00028665024757406775, 'samples': 4560576, 'steps': 23752, 'loss/train': 0.653294026851654} 01/29/2022 19:25:52 - INFO - codeparrot_training - Step 23753: {'lr': 0.0002866340618186497, 'samples': 4560768, 'steps': 23753, 'loss/train': 2.214840888977051} 01/29/2022 19:25:57 - INFO - codeparrot_training - Step 23754: {'lr': 0.00028661787590630297, 'samples': 4560960, 'steps': 23754, 'loss/train': 1.1759944260120392} 01/29/2022 19:26:01 - INFO - codeparrot_training - Step 23755: {'lr': 0.00028660168983709683, 'samples': 4561152, 'steps': 23755, 'loss/train': 1.716211974620819} 01/29/2022 19:26:06 - INFO - codeparrot_training - Step 23756: {'lr': 0.00028658550361110075, 'samples': 4561344, 'steps': 23756, 'loss/train': 1.3853089213371277} 01/29/2022 19:26:10 - INFO - codeparrot_training - Step 23757: {'lr': 0.000286569317228384, 'samples': 4561536, 'steps': 23757, 'loss/train': 1.2177854776382446} 01/29/2022 19:26:14 - INFO - codeparrot_training - Step 23758: {'lr': 0.00028655313068901586, 'samples': 4561728, 'steps': 23758, 'loss/train': 1.4455585777759552} 01/29/2022 19:26:19 - INFO - codeparrot_training - Step 23759: {'lr': 0.0002865369439930657, 'samples': 4561920, 'steps': 23759, 'loss/train': 1.3968193531036377} 01/29/2022 19:26:23 - INFO - codeparrot_training - Step 23760: {'lr': 0.00028652075714060294, 'samples': 4562112, 'steps': 23760, 'loss/train': 1.2114752233028412} 01/29/2022 19:26:30 - INFO - codeparrot_training - Step 23761: {'lr': 0.0002865045701316968, 'samples': 4562304, 'steps': 23761, 'loss/train': 1.4791530668735504} 01/29/2022 19:26:34 - INFO - codeparrot_training - Step 23762: {'lr': 0.00028648838296641666, 'samples': 4562496, 'steps': 23762, 'loss/train': 1.3298044502735138} 01/29/2022 19:26:38 - INFO - codeparrot_training - Step 23763: {'lr': 0.00028647219564483195, 'samples': 4562688, 'steps': 23763, 'loss/train': 1.910481870174408} 01/29/2022 19:26:42 - INFO - codeparrot_training - Step 23764: {'lr': 0.00028645600816701186, 'samples': 4562880, 'steps': 23764, 'loss/train': 1.7682952880859375} 01/29/2022 19:26:47 - INFO - codeparrot_training - Step 23765: {'lr': 0.00028643982053302584, 'samples': 4563072, 'steps': 23765, 'loss/train': 1.5082772970199585} 01/29/2022 19:26:52 - INFO - codeparrot_training - Step 23766: {'lr': 0.00028642363274294317, 'samples': 4563264, 'steps': 23766, 'loss/train': 2.0256861448287964} 01/29/2022 19:26:56 - INFO - codeparrot_training - Step 23767: {'lr': 0.0002864074447968333, 'samples': 4563456, 'steps': 23767, 'loss/train': 1.7174050211906433} 01/29/2022 19:27:00 - INFO - codeparrot_training - Step 23768: {'lr': 0.0002863912566947654, 'samples': 4563648, 'steps': 23768, 'loss/train': 0.7418636083602905} 01/29/2022 19:27:04 - INFO - codeparrot_training - Step 23769: {'lr': 0.000286375068436809, 'samples': 4563840, 'steps': 23769, 'loss/train': 1.9094980359077454} 01/29/2022 19:27:09 - INFO - codeparrot_training - Step 23770: {'lr': 0.00028635888002303324, 'samples': 4564032, 'steps': 23770, 'loss/train': 1.7938088178634644} 01/29/2022 19:27:16 - INFO - codeparrot_training - Step 23771: {'lr': 0.00028634269145350765, 'samples': 4564224, 'steps': 23771, 'loss/train': 1.3138563930988312} 01/29/2022 19:27:20 - INFO - codeparrot_training - Step 23772: {'lr': 0.00028632650272830153, 'samples': 4564416, 'steps': 23772, 'loss/train': 1.6065797209739685} 01/29/2022 19:27:24 - INFO - codeparrot_training - Step 23773: {'lr': 0.00028631031384748426, 'samples': 4564608, 'steps': 23773, 'loss/train': 1.501475214958191} 01/29/2022 19:27:28 - INFO - codeparrot_training - Step 23774: {'lr': 0.000286294124811125, 'samples': 4564800, 'steps': 23774, 'loss/train': 1.3008224666118622} 01/29/2022 19:27:32 - INFO - codeparrot_training - Step 23775: {'lr': 0.0002862779356192933, 'samples': 4564992, 'steps': 23775, 'loss/train': 1.6909782886505127} 01/29/2022 19:27:38 - INFO - codeparrot_training - Step 23776: {'lr': 0.0002862617462720584, 'samples': 4565184, 'steps': 23776, 'loss/train': 0.7725518345832825} 01/29/2022 19:27:42 - INFO - codeparrot_training - Step 23777: {'lr': 0.00028624555676948975, 'samples': 4565376, 'steps': 23777, 'loss/train': 1.2059840261936188} 01/29/2022 19:27:46 - INFO - codeparrot_training - Step 23778: {'lr': 0.00028622936711165665, 'samples': 4565568, 'steps': 23778, 'loss/train': 4.271234035491943} 01/29/2022 19:27:51 - INFO - codeparrot_training - Step 23779: {'lr': 0.00028621317729862837, 'samples': 4565760, 'steps': 23779, 'loss/train': 1.623085856437683} 01/29/2022 19:27:56 - INFO - codeparrot_training - Step 23780: {'lr': 0.00028619698733047444, 'samples': 4565952, 'steps': 23780, 'loss/train': 1.645267367362976} 01/29/2022 19:28:00 - INFO - codeparrot_training - Step 23781: {'lr': 0.0002861807972072641, 'samples': 4566144, 'steps': 23781, 'loss/train': 1.791879415512085} 01/29/2022 19:28:04 - INFO - codeparrot_training - Step 23782: {'lr': 0.0002861646069290667, 'samples': 4566336, 'steps': 23782, 'loss/train': 1.2174807786941528} 01/29/2022 19:28:08 - INFO - codeparrot_training - Step 23783: {'lr': 0.0002861484164959515, 'samples': 4566528, 'steps': 23783, 'loss/train': 1.5905728340148926} 01/29/2022 19:28:13 - INFO - codeparrot_training - Step 23784: {'lr': 0.0002861322259079881, 'samples': 4566720, 'steps': 23784, 'loss/train': 1.6962445378303528} 01/29/2022 19:28:20 - INFO - codeparrot_training - Step 23785: {'lr': 0.00028611603516524566, 'samples': 4566912, 'steps': 23785, 'loss/train': 2.1118746399879456} 01/29/2022 19:28:24 - INFO - codeparrot_training - Step 23786: {'lr': 0.00028609984426779364, 'samples': 4567104, 'steps': 23786, 'loss/train': 1.920285165309906} 01/29/2022 19:28:28 - INFO - codeparrot_training - Step 23787: {'lr': 0.0002860836532157012, 'samples': 4567296, 'steps': 23787, 'loss/train': 1.371019721031189} 01/29/2022 19:28:32 - INFO - codeparrot_training - Step 23788: {'lr': 0.000286067462009038, 'samples': 4567488, 'steps': 23788, 'loss/train': 1.7535801529884338} 01/29/2022 19:28:36 - INFO - codeparrot_training - Step 23789: {'lr': 0.00028605127064787315, 'samples': 4567680, 'steps': 23789, 'loss/train': 1.6762797832489014} 01/29/2022 19:28:42 - INFO - codeparrot_training - Step 23790: {'lr': 0.00028603507913227617, 'samples': 4567872, 'steps': 23790, 'loss/train': 1.8376827836036682} 01/29/2022 19:28:46 - INFO - codeparrot_training - Step 23791: {'lr': 0.00028601888746231633, 'samples': 4568064, 'steps': 23791, 'loss/train': 7.076035737991333} 01/29/2022 19:28:50 - INFO - codeparrot_training - Step 23792: {'lr': 0.00028600269563806304, 'samples': 4568256, 'steps': 23792, 'loss/train': 2.0563488006591797} 01/29/2022 19:28:54 - INFO - codeparrot_training - Step 23793: {'lr': 0.0002859865036595856, 'samples': 4568448, 'steps': 23793, 'loss/train': 1.8380098342895508} 01/29/2022 19:28:59 - INFO - codeparrot_training - Step 23794: {'lr': 0.0002859703115269534, 'samples': 4568640, 'steps': 23794, 'loss/train': 1.7151578664779663} 01/29/2022 19:29:04 - INFO - codeparrot_training - Step 23795: {'lr': 0.0002859541192402359, 'samples': 4568832, 'steps': 23795, 'loss/train': 1.7719531059265137} 01/29/2022 19:29:08 - INFO - codeparrot_training - Step 23796: {'lr': 0.00028593792679950227, 'samples': 4569024, 'steps': 23796, 'loss/train': 1.539851188659668} 01/29/2022 19:29:12 - INFO - codeparrot_training - Step 23797: {'lr': 0.00028592173420482206, 'samples': 4569216, 'steps': 23797, 'loss/train': 1.4888718724250793} 01/29/2022 19:29:17 - INFO - codeparrot_training - Step 23798: {'lr': 0.0002859055414562644, 'samples': 4569408, 'steps': 23798, 'loss/train': 1.5513405203819275} 01/29/2022 19:29:21 - INFO - codeparrot_training - Step 23799: {'lr': 0.00028588934855389885, 'samples': 4569600, 'steps': 23799, 'loss/train': 1.4778040945529938} 01/29/2022 19:29:27 - INFO - codeparrot_training - Step 23800: {'lr': 0.0002858731554977948, 'samples': 4569792, 'steps': 23800, 'loss/train': 1.8277117609977722} 01/29/2022 19:29:31 - INFO - codeparrot_training - Step 23801: {'lr': 0.00028585696228802153, 'samples': 4569984, 'steps': 23801, 'loss/train': 2.074763596057892} 01/29/2022 19:29:35 - INFO - codeparrot_training - Step 23802: {'lr': 0.0002858407689246484, 'samples': 4570176, 'steps': 23802, 'loss/train': 1.2895670235157013} 01/29/2022 19:29:39 - INFO - codeparrot_training - Step 23803: {'lr': 0.0002858245754077448, 'samples': 4570368, 'steps': 23803, 'loss/train': 0.1970665007829666} 01/29/2022 19:29:44 - INFO - codeparrot_training - Step 23804: {'lr': 0.0002858083817373801, 'samples': 4570560, 'steps': 23804, 'loss/train': 0.7052941471338272} 01/29/2022 19:29:51 - INFO - codeparrot_training - Step 23805: {'lr': 0.00028579218791362367, 'samples': 4570752, 'steps': 23805, 'loss/train': 1.0911725163459778} 01/29/2022 19:29:55 - INFO - codeparrot_training - Step 23806: {'lr': 0.0002857759939365449, 'samples': 4570944, 'steps': 23806, 'loss/train': 1.2738105654716492} 01/29/2022 19:29:59 - INFO - codeparrot_training - Step 23807: {'lr': 0.000285759799806213, 'samples': 4571136, 'steps': 23807, 'loss/train': 2.2049102783203125} 01/29/2022 19:30:03 - INFO - codeparrot_training - Step 23808: {'lr': 0.00028574360552269755, 'samples': 4571328, 'steps': 23808, 'loss/train': 1.3672296702861786} 01/29/2022 19:30:07 - INFO - codeparrot_training - Step 23809: {'lr': 0.0002857274110860679, 'samples': 4571520, 'steps': 23809, 'loss/train': 1.8350075483322144} 01/29/2022 19:30:13 - INFO - codeparrot_training - Step 23810: {'lr': 0.00028571121649639337, 'samples': 4571712, 'steps': 23810, 'loss/train': 1.603756070137024} 01/29/2022 19:30:17 - INFO - codeparrot_training - Step 23811: {'lr': 0.0002856950217537432, 'samples': 4571904, 'steps': 23811, 'loss/train': 0.9147336781024933} 01/29/2022 19:30:22 - INFO - codeparrot_training - Step 23812: {'lr': 0.000285678826858187, 'samples': 4572096, 'steps': 23812, 'loss/train': 1.9245082139968872} 01/29/2022 19:30:26 - INFO - codeparrot_training - Step 23813: {'lr': 0.000285662631809794, 'samples': 4572288, 'steps': 23813, 'loss/train': 2.3410520553588867} 01/29/2022 19:30:30 - INFO - codeparrot_training - Step 23814: {'lr': 0.0002856464366086336, 'samples': 4572480, 'steps': 23814, 'loss/train': 1.513264775276184} 01/29/2022 19:30:37 - INFO - codeparrot_training - Step 23815: {'lr': 0.0002856302412547752, 'samples': 4572672, 'steps': 23815, 'loss/train': 1.2241946160793304} 01/29/2022 19:30:41 - INFO - codeparrot_training - Step 23816: {'lr': 0.0002856140457482882, 'samples': 4572864, 'steps': 23816, 'loss/train': 1.3713993430137634} 01/29/2022 19:30:45 - INFO - codeparrot_training - Step 23817: {'lr': 0.0002855978500892419, 'samples': 4573056, 'steps': 23817, 'loss/train': 0.9049513041973114} 01/29/2022 19:30:49 - INFO - codeparrot_training - Step 23818: {'lr': 0.00028558165427770567, 'samples': 4573248, 'steps': 23818, 'loss/train': 2.162637948989868} 01/29/2022 19:30:54 - INFO - codeparrot_training - Step 23819: {'lr': 0.00028556545831374903, 'samples': 4573440, 'steps': 23819, 'loss/train': 0.9633722305297852} 01/29/2022 19:30:59 - INFO - codeparrot_training - Step 23820: {'lr': 0.0002855492621974411, 'samples': 4573632, 'steps': 23820, 'loss/train': 2.0418970584869385} 01/29/2022 19:31:03 - INFO - codeparrot_training - Step 23821: {'lr': 0.0002855330659288516, 'samples': 4573824, 'steps': 23821, 'loss/train': 0.2941695749759674} 01/29/2022 19:31:07 - INFO - codeparrot_training - Step 23822: {'lr': 0.00028551686950804964, 'samples': 4574016, 'steps': 23822, 'loss/train': 1.7458866834640503} 01/29/2022 19:31:11 - INFO - codeparrot_training - Step 23823: {'lr': 0.0002855006729351046, 'samples': 4574208, 'steps': 23823, 'loss/train': 1.0646314322948456} 01/29/2022 19:31:16 - INFO - codeparrot_training - Step 23824: {'lr': 0.0002854844762100861, 'samples': 4574400, 'steps': 23824, 'loss/train': 1.6914478540420532} 01/29/2022 19:31:21 - INFO - codeparrot_training - Step 23825: {'lr': 0.0002854682793330633, 'samples': 4574592, 'steps': 23825, 'loss/train': 1.7167373299598694} 01/29/2022 19:31:25 - INFO - codeparrot_training - Step 23826: {'lr': 0.0002854520823041057, 'samples': 4574784, 'steps': 23826, 'loss/train': 1.381077915430069} 01/29/2022 19:31:29 - INFO - codeparrot_training - Step 23827: {'lr': 0.0002854358851232826, 'samples': 4574976, 'steps': 23827, 'loss/train': 1.4398386776447296} 01/29/2022 19:31:34 - INFO - codeparrot_training - Step 23828: {'lr': 0.0002854196877906635, 'samples': 4575168, 'steps': 23828, 'loss/train': 1.9371782541275024} 01/29/2022 19:31:38 - INFO - codeparrot_training - Step 23829: {'lr': 0.00028540349030631756, 'samples': 4575360, 'steps': 23829, 'loss/train': 1.711251139640808} 01/29/2022 19:31:45 - INFO - codeparrot_training - Step 23830: {'lr': 0.0002853872926703144, 'samples': 4575552, 'steps': 23830, 'loss/train': 0.7924161851406097} 01/29/2022 19:31:49 - INFO - codeparrot_training - Step 23831: {'lr': 0.0002853710948827233, 'samples': 4575744, 'steps': 23831, 'loss/train': 1.3826949298381805} 01/29/2022 19:31:54 - INFO - codeparrot_training - Step 23832: {'lr': 0.00028535489694361365, 'samples': 4575936, 'steps': 23832, 'loss/train': 1.1010820269584656} 01/29/2022 19:31:58 - INFO - codeparrot_training - Step 23833: {'lr': 0.00028533869885305485, 'samples': 4576128, 'steps': 23833, 'loss/train': 1.9540046453475952} 01/29/2022 19:32:02 - INFO - codeparrot_training - Step 23834: {'lr': 0.0002853225006111163, 'samples': 4576320, 'steps': 23834, 'loss/train': 0.8004572689533234} 01/29/2022 19:32:06 - INFO - codeparrot_training - Step 23835: {'lr': 0.00028530630221786736, 'samples': 4576512, 'steps': 23835, 'loss/train': 6.822926759719849} 01/29/2022 19:32:12 - INFO - codeparrot_training - Step 23836: {'lr': 0.00028529010367337745, 'samples': 4576704, 'steps': 23836, 'loss/train': 2.3045502305030823} 01/29/2022 19:32:16 - INFO - codeparrot_training - Step 23837: {'lr': 0.0002852739049777159, 'samples': 4576896, 'steps': 23837, 'loss/train': 1.4196816086769104} 01/29/2022 19:32:20 - INFO - codeparrot_training - Step 23838: {'lr': 0.00028525770613095213, 'samples': 4577088, 'steps': 23838, 'loss/train': 1.6034406423568726} 01/29/2022 19:32:24 - INFO - codeparrot_training - Step 23839: {'lr': 0.00028524150713315566, 'samples': 4577280, 'steps': 23839, 'loss/train': 1.4373348355293274} 01/29/2022 19:32:28 - INFO - codeparrot_training - Step 23840: {'lr': 0.00028522530798439564, 'samples': 4577472, 'steps': 23840, 'loss/train': 1.5730960965156555} 01/29/2022 19:32:36 - INFO - codeparrot_training - Step 23841: {'lr': 0.0002852091086847417, 'samples': 4577664, 'steps': 23841, 'loss/train': 2.0838144421577454} 01/29/2022 19:32:40 - INFO - codeparrot_training - Step 23842: {'lr': 0.000285192909234263, 'samples': 4577856, 'steps': 23842, 'loss/train': 1.7983176112174988} 01/29/2022 19:32:45 - INFO - codeparrot_training - Step 23843: {'lr': 0.0002851767096330291, 'samples': 4578048, 'steps': 23843, 'loss/train': 1.8224149346351624} 01/29/2022 19:32:49 - INFO - codeparrot_training - Step 23844: {'lr': 0.00028516050988110935, 'samples': 4578240, 'steps': 23844, 'loss/train': 1.6363706588745117} 01/29/2022 19:32:53 - INFO - codeparrot_training - Step 23845: {'lr': 0.00028514430997857317, 'samples': 4578432, 'steps': 23845, 'loss/train': 1.4135805666446686} 01/29/2022 19:32:57 - INFO - codeparrot_training - Step 23846: {'lr': 0.0002851281099254899, 'samples': 4578624, 'steps': 23846, 'loss/train': 2.20774644613266} 01/29/2022 19:33:02 - INFO - codeparrot_training - Step 23847: {'lr': 0.000285111909721929, 'samples': 4578816, 'steps': 23847, 'loss/train': 1.1829372346401215} 01/29/2022 19:33:07 - INFO - codeparrot_training - Step 23848: {'lr': 0.0002850957093679597, 'samples': 4579008, 'steps': 23848, 'loss/train': 1.3758874833583832} 01/29/2022 19:33:11 - INFO - codeparrot_training - Step 23849: {'lr': 0.0002850795088636516, 'samples': 4579200, 'steps': 23849, 'loss/train': 1.052115261554718} 01/29/2022 19:33:15 - INFO - codeparrot_training - Step 23850: {'lr': 0.0002850633082090741, 'samples': 4579392, 'steps': 23850, 'loss/train': 1.8243750929832458} 01/29/2022 19:33:19 - INFO - codeparrot_training - Step 23851: {'lr': 0.00028504710740429647, 'samples': 4579584, 'steps': 23851, 'loss/train': 1.0301754176616669} 01/29/2022 19:33:25 - INFO - codeparrot_training - Step 23852: {'lr': 0.0002850309064493882, 'samples': 4579776, 'steps': 23852, 'loss/train': 0.9863284528255463} 01/29/2022 19:33:29 - INFO - codeparrot_training - Step 23853: {'lr': 0.00028501470534441855, 'samples': 4579968, 'steps': 23853, 'loss/train': 0.96084925532341} 01/29/2022 19:33:33 - INFO - codeparrot_training - Step 23854: {'lr': 0.00028499850408945704, 'samples': 4580160, 'steps': 23854, 'loss/train': 1.420459359884262} 01/29/2022 19:33:37 - INFO - codeparrot_training - Step 23855: {'lr': 0.0002849823026845731, 'samples': 4580352, 'steps': 23855, 'loss/train': 2.1820918321609497} 01/29/2022 19:33:42 - INFO - codeparrot_training - Step 23856: {'lr': 0.00028496610112983605, 'samples': 4580544, 'steps': 23856, 'loss/train': 1.1218160390853882} 01/29/2022 19:33:47 - INFO - codeparrot_training - Step 23857: {'lr': 0.00028494989942531537, 'samples': 4580736, 'steps': 23857, 'loss/train': 1.9560842514038086} 01/29/2022 19:33:51 - INFO - codeparrot_training - Step 23858: {'lr': 0.0002849336975710804, 'samples': 4580928, 'steps': 23858, 'loss/train': 1.6119248270988464} 01/29/2022 19:33:55 - INFO - codeparrot_training - Step 23859: {'lr': 0.0002849174955672005, 'samples': 4581120, 'steps': 23859, 'loss/train': 1.2113977074623108} 01/29/2022 19:34:00 - INFO - codeparrot_training - Step 23860: {'lr': 0.0002849012934137452, 'samples': 4581312, 'steps': 23860, 'loss/train': 1.7344120144844055} 01/29/2022 19:34:04 - INFO - codeparrot_training - Step 23861: {'lr': 0.0002848850911107838, 'samples': 4581504, 'steps': 23861, 'loss/train': 1.6292213201522827} 01/29/2022 19:34:11 - INFO - codeparrot_training - Step 23862: {'lr': 0.0002848688886583858, 'samples': 4581696, 'steps': 23862, 'loss/train': 1.101608008146286} 01/29/2022 19:34:16 - INFO - codeparrot_training - Step 23863: {'lr': 0.0002848526860566205, 'samples': 4581888, 'steps': 23863, 'loss/train': 0.5139185786247253} 01/29/2022 19:34:20 - INFO - codeparrot_training - Step 23864: {'lr': 0.00028483648330555737, 'samples': 4582080, 'steps': 23864, 'loss/train': 1.9260830283164978} 01/29/2022 19:34:24 - INFO - codeparrot_training - Step 23865: {'lr': 0.0002848202804052659, 'samples': 4582272, 'steps': 23865, 'loss/train': 1.8986132740974426} 01/29/2022 19:34:28 - INFO - codeparrot_training - Step 23866: {'lr': 0.00028480407735581527, 'samples': 4582464, 'steps': 23866, 'loss/train': 1.4476490914821625} 01/29/2022 19:34:34 - INFO - codeparrot_training - Step 23867: {'lr': 0.00028478787415727515, 'samples': 4582656, 'steps': 23867, 'loss/train': 1.9308829307556152} 01/29/2022 19:34:38 - INFO - codeparrot_training - Step 23868: {'lr': 0.00028477167080971465, 'samples': 4582848, 'steps': 23868, 'loss/train': 2.2217541933059692} 01/29/2022 19:34:42 - INFO - codeparrot_training - Step 23869: {'lr': 0.0002847554673132035, 'samples': 4583040, 'steps': 23869, 'loss/train': 2.3728623390197754} 01/29/2022 19:34:47 - INFO - codeparrot_training - Step 23870: {'lr': 0.00028473926366781095, 'samples': 4583232, 'steps': 23870, 'loss/train': 1.3863560557365417} 01/29/2022 19:34:51 - INFO - codeparrot_training - Step 23871: {'lr': 0.00028472305987360636, 'samples': 4583424, 'steps': 23871, 'loss/train': 2.0632824897766113} 01/29/2022 19:34:55 - INFO - codeparrot_training - Step 23872: {'lr': 0.00028470685593065927, 'samples': 4583616, 'steps': 23872, 'loss/train': 1.2877518832683563} 01/29/2022 19:35:00 - INFO - codeparrot_training - Step 23873: {'lr': 0.00028469065183903895, 'samples': 4583808, 'steps': 23873, 'loss/train': 1.3399710059165955} 01/29/2022 19:35:05 - INFO - codeparrot_training - Step 23874: {'lr': 0.000284674447598815, 'samples': 4584000, 'steps': 23874, 'loss/train': 0.10148592665791512} 01/29/2022 19:35:09 - INFO - codeparrot_training - Step 23875: {'lr': 0.00028465824321005667, 'samples': 4584192, 'steps': 23875, 'loss/train': 1.0485747456550598} 01/29/2022 19:35:13 - INFO - codeparrot_training - Step 23876: {'lr': 0.0002846420386728334, 'samples': 4584384, 'steps': 23876, 'loss/train': 1.7209099531173706} 01/29/2022 19:35:17 - INFO - codeparrot_training - Step 23877: {'lr': 0.0002846258339872147, 'samples': 4584576, 'steps': 23877, 'loss/train': 1.7283049821853638} 01/29/2022 19:35:24 - INFO - codeparrot_training - Step 23878: {'lr': 0.0002846096291532699, 'samples': 4584768, 'steps': 23878, 'loss/train': 1.7433993816375732} 01/29/2022 19:35:29 - INFO - codeparrot_training - Step 23879: {'lr': 0.0002845934241710684, 'samples': 4584960, 'steps': 23879, 'loss/train': 2.1004786491394043} 01/29/2022 19:35:33 - INFO - codeparrot_training - Step 23880: {'lr': 0.0002845772190406798, 'samples': 4585152, 'steps': 23880, 'loss/train': 6.667693376541138} 01/29/2022 19:35:37 - INFO - codeparrot_training - Step 23881: {'lr': 0.0002845610137621732, 'samples': 4585344, 'steps': 23881, 'loss/train': 0.3637164607644081} 01/29/2022 19:35:41 - INFO - codeparrot_training - Step 23882: {'lr': 0.0002845448083356183, 'samples': 4585536, 'steps': 23882, 'loss/train': 1.4367566406726837} 01/29/2022 19:35:47 - INFO - codeparrot_training - Step 23883: {'lr': 0.00028452860276108436, 'samples': 4585728, 'steps': 23883, 'loss/train': 0.30817843973636627} 01/29/2022 19:35:51 - INFO - codeparrot_training - Step 23884: {'lr': 0.0002845123970386408, 'samples': 4585920, 'steps': 23884, 'loss/train': 0.657759815454483} 01/29/2022 19:35:55 - INFO - codeparrot_training - Step 23885: {'lr': 0.00028449619116835715, 'samples': 4586112, 'steps': 23885, 'loss/train': 2.3292782306671143} 01/29/2022 19:35:59 - INFO - codeparrot_training - Step 23886: {'lr': 0.0002844799851503028, 'samples': 4586304, 'steps': 23886, 'loss/train': 1.1195893585681915} 01/29/2022 19:36:04 - INFO - codeparrot_training - Step 23887: {'lr': 0.0002844637789845471, 'samples': 4586496, 'steps': 23887, 'loss/train': 2.125841438770294} 01/29/2022 19:36:09 - INFO - codeparrot_training - Step 23888: {'lr': 0.0002844475726711595, 'samples': 4586688, 'steps': 23888, 'loss/train': 2.2391082644462585} 01/29/2022 19:36:13 - INFO - codeparrot_training - Step 23889: {'lr': 0.0002844313662102095, 'samples': 4586880, 'steps': 23889, 'loss/train': 1.4712892770767212} 01/29/2022 19:36:17 - INFO - codeparrot_training - Step 23890: {'lr': 0.0002844151596017665, 'samples': 4587072, 'steps': 23890, 'loss/train': 2.506456196308136} 01/29/2022 19:36:21 - INFO - codeparrot_training - Step 23891: {'lr': 0.0002843989528458997, 'samples': 4587264, 'steps': 23891, 'loss/train': 1.4225609600543976} 01/29/2022 19:36:26 - INFO - codeparrot_training - Step 23892: {'lr': 0.0002843827459426789, 'samples': 4587456, 'steps': 23892, 'loss/train': 0.7289436310529709} 01/29/2022 19:36:33 - INFO - codeparrot_training - Step 23893: {'lr': 0.00028436653889217316, 'samples': 4587648, 'steps': 23893, 'loss/train': 1.9818010926246643} 01/29/2022 19:36:37 - INFO - codeparrot_training - Step 23894: {'lr': 0.00028435033169445223, 'samples': 4587840, 'steps': 23894, 'loss/train': 1.9826531410217285} 01/29/2022 19:36:41 - INFO - codeparrot_training - Step 23895: {'lr': 0.0002843341243495853, 'samples': 4588032, 'steps': 23895, 'loss/train': 1.2097082734107971} 01/29/2022 19:36:46 - INFO - codeparrot_training - Step 23896: {'lr': 0.0002843179168576419, 'samples': 4588224, 'steps': 23896, 'loss/train': 1.5351024270057678} 01/29/2022 19:36:50 - INFO - codeparrot_training - Step 23897: {'lr': 0.00028430170921869147, 'samples': 4588416, 'steps': 23897, 'loss/train': 1.1531972587108612} 01/29/2022 19:36:55 - INFO - codeparrot_training - Step 23898: {'lr': 0.0002842855014328034, 'samples': 4588608, 'steps': 23898, 'loss/train': 1.8952388763427734} 01/29/2022 19:37:00 - INFO - codeparrot_training - Step 23899: {'lr': 0.0002842692935000471, 'samples': 4588800, 'steps': 23899, 'loss/train': 1.4552857875823975} 01/29/2022 19:37:04 - INFO - codeparrot_training - Step 23900: {'lr': 0.00028425308542049207, 'samples': 4588992, 'steps': 23900, 'loss/train': 0.5217408835887909} 01/29/2022 19:37:08 - INFO - codeparrot_training - Step 23901: {'lr': 0.0002842368771942077, 'samples': 4589184, 'steps': 23901, 'loss/train': 2.226067364215851} 01/29/2022 19:37:12 - INFO - codeparrot_training - Step 23902: {'lr': 0.00028422066882126336, 'samples': 4589376, 'steps': 23902, 'loss/train': 3.2934683561325073} 01/29/2022 19:37:19 - INFO - codeparrot_training - Step 23903: {'lr': 0.0002842044603017285, 'samples': 4589568, 'steps': 23903, 'loss/train': 1.824639916419983} 01/29/2022 19:37:24 - INFO - codeparrot_training - Step 23904: {'lr': 0.00028418825163567275, 'samples': 4589760, 'steps': 23904, 'loss/train': 1.7718345522880554} 01/29/2022 19:37:28 - INFO - codeparrot_training - Step 23905: {'lr': 0.0002841720428231653, 'samples': 4589952, 'steps': 23905, 'loss/train': 1.5791105031967163} 01/29/2022 19:37:32 - INFO - codeparrot_training - Step 23906: {'lr': 0.00028415583386427566, 'samples': 4590144, 'steps': 23906, 'loss/train': 1.253352016210556} 01/29/2022 19:37:36 - INFO - codeparrot_training - Step 23907: {'lr': 0.0002841396247590733, 'samples': 4590336, 'steps': 23907, 'loss/train': 1.9852674007415771} 01/29/2022 19:37:41 - INFO - codeparrot_training - Step 23908: {'lr': 0.00028412341550762755, 'samples': 4590528, 'steps': 23908, 'loss/train': 2.7583696246147156} 01/29/2022 19:37:46 - INFO - codeparrot_training - Step 23909: {'lr': 0.00028410720611000804, 'samples': 4590720, 'steps': 23909, 'loss/train': 2.0297282338142395} 01/29/2022 19:37:50 - INFO - codeparrot_training - Step 23910: {'lr': 0.000284090996566284, 'samples': 4590912, 'steps': 23910, 'loss/train': 1.7018296122550964} 01/29/2022 19:37:54 - INFO - codeparrot_training - Step 23911: {'lr': 0.00028407478687652503, 'samples': 4591104, 'steps': 23911, 'loss/train': 1.6421799659729004} 01/29/2022 19:37:58 - INFO - codeparrot_training - Step 23912: {'lr': 0.0002840585770408004, 'samples': 4591296, 'steps': 23912, 'loss/train': 4.3295817375183105} 01/29/2022 19:38:03 - INFO - codeparrot_training - Step 23913: {'lr': 0.00028404236705917974, 'samples': 4591488, 'steps': 23913, 'loss/train': 1.6491249203681946} 01/29/2022 19:38:08 - INFO - codeparrot_training - Step 23914: {'lr': 0.00028402615693173236, 'samples': 4591680, 'steps': 23914, 'loss/train': 1.2246925234794617} 01/29/2022 19:38:12 - INFO - codeparrot_training - Step 23915: {'lr': 0.00028400994665852777, 'samples': 4591872, 'steps': 23915, 'loss/train': 1.3568209111690521} 01/29/2022 19:38:16 - INFO - codeparrot_training - Step 23916: {'lr': 0.00028399373623963525, 'samples': 4592064, 'steps': 23916, 'loss/train': 1.4262498915195465} 01/29/2022 19:38:20 - INFO - codeparrot_training - Step 23917: {'lr': 0.0002839775256751244, 'samples': 4592256, 'steps': 23917, 'loss/train': 2.5017465949058533} 01/29/2022 19:38:28 - INFO - codeparrot_training - Step 23918: {'lr': 0.00028396131496506466, 'samples': 4592448, 'steps': 23918, 'loss/train': 7.339595317840576} 01/29/2022 19:38:32 - INFO - codeparrot_training - Step 23919: {'lr': 0.00028394510410952544, 'samples': 4592640, 'steps': 23919, 'loss/train': 1.2172236442565918} 01/29/2022 19:38:36 - INFO - codeparrot_training - Step 23920: {'lr': 0.0002839288931085761, 'samples': 4592832, 'steps': 23920, 'loss/train': 1.8071399331092834} 01/29/2022 19:38:40 - INFO - codeparrot_training - Step 23921: {'lr': 0.0002839126819622862, 'samples': 4593024, 'steps': 23921, 'loss/train': 2.278142273426056} 01/29/2022 19:38:44 - INFO - codeparrot_training - Step 23922: {'lr': 0.00028389647067072517, 'samples': 4593216, 'steps': 23922, 'loss/train': 2.8620150089263916} 01/29/2022 19:38:50 - INFO - codeparrot_training - Step 23923: {'lr': 0.00028388025923396234, 'samples': 4593408, 'steps': 23923, 'loss/train': 1.8065795302391052} 01/29/2022 19:38:54 - INFO - codeparrot_training - Step 23924: {'lr': 0.0002838640476520673, 'samples': 4593600, 'steps': 23924, 'loss/train': 1.7671501636505127} 01/29/2022 19:38:58 - INFO - codeparrot_training - Step 23925: {'lr': 0.00028384783592510945, 'samples': 4593792, 'steps': 23925, 'loss/train': 7.2844648361206055} 01/29/2022 19:39:03 - INFO - codeparrot_training - Step 23926: {'lr': 0.00028383162405315823, 'samples': 4593984, 'steps': 23926, 'loss/train': 1.9071187376976013} 01/29/2022 19:39:07 - INFO - codeparrot_training - Step 23927: {'lr': 0.00028381541203628295, 'samples': 4594176, 'steps': 23927, 'loss/train': 1.6116403341293335} 01/29/2022 19:39:12 - INFO - codeparrot_training - Step 23928: {'lr': 0.0002837991998745533, 'samples': 4594368, 'steps': 23928, 'loss/train': 1.4861798286437988} 01/29/2022 19:39:16 - INFO - codeparrot_training - Step 23929: {'lr': 0.0002837829875680386, 'samples': 4594560, 'steps': 23929, 'loss/train': 1.7700303196907043} 01/29/2022 19:39:21 - INFO - codeparrot_training - Step 23930: {'lr': 0.00028376677511680827, 'samples': 4594752, 'steps': 23930, 'loss/train': 1.864320158958435} 01/29/2022 19:39:25 - INFO - codeparrot_training - Step 23931: {'lr': 0.0002837505625209318, 'samples': 4594944, 'steps': 23931, 'loss/train': 2.405535399913788} 01/29/2022 19:39:29 - INFO - codeparrot_training - Step 23932: {'lr': 0.0002837343497804787, 'samples': 4595136, 'steps': 23932, 'loss/train': 1.9334238767623901} 01/29/2022 19:39:34 - INFO - codeparrot_training - Step 23933: {'lr': 0.0002837181368955183, 'samples': 4595328, 'steps': 23933, 'loss/train': 1.53024423122406} 01/29/2022 19:39:38 - INFO - codeparrot_training - Step 23934: {'lr': 0.0002837019238661201, 'samples': 4595520, 'steps': 23934, 'loss/train': 1.4726594388484955} 01/29/2022 19:39:43 - INFO - codeparrot_training - Step 23935: {'lr': 0.00028368571069235354, 'samples': 4595712, 'steps': 23935, 'loss/train': 1.70952308177948} 01/29/2022 19:39:47 - INFO - codeparrot_training - Step 23936: {'lr': 0.00028366949737428814, 'samples': 4595904, 'steps': 23936, 'loss/train': 2.256426215171814} 01/29/2022 19:39:51 - INFO - codeparrot_training - Step 23937: {'lr': 0.00028365328391199334, 'samples': 4596096, 'steps': 23937, 'loss/train': 1.5048179626464844} 01/29/2022 19:39:58 - INFO - codeparrot_training - Step 23938: {'lr': 0.0002836370703055385, 'samples': 4596288, 'steps': 23938, 'loss/train': 0.08158080093562603} 01/29/2022 19:40:02 - INFO - codeparrot_training - Step 23939: {'lr': 0.0002836208565549932, 'samples': 4596480, 'steps': 23939, 'loss/train': 1.6059486865997314} 01/29/2022 19:40:07 - INFO - codeparrot_training - Step 23940: {'lr': 0.00028360464266042674, 'samples': 4596672, 'steps': 23940, 'loss/train': 1.704684555530548} 01/29/2022 19:40:11 - INFO - codeparrot_training - Step 23941: {'lr': 0.00028358842862190873, 'samples': 4596864, 'steps': 23941, 'loss/train': 1.7739797830581665} 01/29/2022 19:40:15 - INFO - codeparrot_training - Step 23942: {'lr': 0.00028357221443950847, 'samples': 4597056, 'steps': 23942, 'loss/train': 1.3783408105373383} 01/29/2022 19:40:19 - INFO - codeparrot_training - Step 23943: {'lr': 0.00028355600011329557, 'samples': 4597248, 'steps': 23943, 'loss/train': 1.5278881788253784} 01/29/2022 19:40:24 - INFO - codeparrot_training - Step 23944: {'lr': 0.00028353978564333936, 'samples': 4597440, 'steps': 23944, 'loss/train': 1.6490326523780823} 01/29/2022 19:40:29 - INFO - codeparrot_training - Step 23945: {'lr': 0.0002835235710297094, 'samples': 4597632, 'steps': 23945, 'loss/train': 3.061728358268738} 01/29/2022 19:40:33 - INFO - codeparrot_training - Step 23946: {'lr': 0.0002835073562724751, 'samples': 4597824, 'steps': 23946, 'loss/train': 1.5444656610488892} 01/29/2022 19:40:37 - INFO - codeparrot_training - Step 23947: {'lr': 0.00028349114137170593, 'samples': 4598016, 'steps': 23947, 'loss/train': 0.9231298863887787} 01/29/2022 19:40:41 - INFO - codeparrot_training - Step 23948: {'lr': 0.0002834749263274714, 'samples': 4598208, 'steps': 23948, 'loss/train': 1.6378228068351746} 01/29/2022 19:40:48 - INFO - codeparrot_training - Step 23949: {'lr': 0.00028345871113984086, 'samples': 4598400, 'steps': 23949, 'loss/train': 1.4845108687877655} 01/29/2022 19:40:52 - INFO - codeparrot_training - Step 23950: {'lr': 0.0002834424958088838, 'samples': 4598592, 'steps': 23950, 'loss/train': 1.4202479124069214} 01/29/2022 19:40:57 - INFO - codeparrot_training - Step 23951: {'lr': 0.00028342628033466974, 'samples': 4598784, 'steps': 23951, 'loss/train': 1.2285007238388062} 01/29/2022 19:41:01 - INFO - codeparrot_training - Step 23952: {'lr': 0.00028341006471726816, 'samples': 4598976, 'steps': 23952, 'loss/train': 2.1908154487609863} 01/29/2022 19:41:05 - INFO - codeparrot_training - Step 23953: {'lr': 0.0002833938489567484, 'samples': 4599168, 'steps': 23953, 'loss/train': 1.482117623090744} 01/29/2022 19:41:11 - INFO - codeparrot_training - Step 23954: {'lr': 0.00028337763305318, 'samples': 4599360, 'steps': 23954, 'loss/train': 0.8394040167331696} 01/29/2022 19:41:15 - INFO - codeparrot_training - Step 23955: {'lr': 0.00028336141700663244, 'samples': 4599552, 'steps': 23955, 'loss/train': 1.9457117915153503} 01/29/2022 19:41:19 - INFO - codeparrot_training - Step 23956: {'lr': 0.00028334520081717507, 'samples': 4599744, 'steps': 23956, 'loss/train': 1.0480142533779144} 01/29/2022 19:41:24 - INFO - codeparrot_training - Step 23957: {'lr': 0.0002833289844848776, 'samples': 4599936, 'steps': 23957, 'loss/train': 0.804270476102829} 01/29/2022 19:41:28 - INFO - codeparrot_training - Step 23958: {'lr': 0.0002833127680098092, 'samples': 4600128, 'steps': 23958, 'loss/train': 1.8483906984329224} 01/29/2022 19:41:32 - INFO - codeparrot_training - Step 23959: {'lr': 0.0002832965513920396, 'samples': 4600320, 'steps': 23959, 'loss/train': 1.3476849496364594} 01/29/2022 19:41:37 - INFO - codeparrot_training - Step 23960: {'lr': 0.0002832803346316381, 'samples': 4600512, 'steps': 23960, 'loss/train': 1.3335632979869843} 01/29/2022 19:41:41 - INFO - codeparrot_training - Step 23961: {'lr': 0.0002832641177286742, 'samples': 4600704, 'steps': 23961, 'loss/train': 1.489419847726822} 01/29/2022 19:41:46 - INFO - codeparrot_training - Step 23962: {'lr': 0.0002832479006832174, 'samples': 4600896, 'steps': 23962, 'loss/train': 2.156388223171234} 01/29/2022 19:41:50 - INFO - codeparrot_training - Step 23963: {'lr': 0.0002832316834953372, 'samples': 4601088, 'steps': 23963, 'loss/train': 1.30601567029953} 01/29/2022 19:41:54 - INFO - codeparrot_training - Step 23964: {'lr': 0.0002832154661651029, 'samples': 4601280, 'steps': 23964, 'loss/train': 1.8446361422538757} 01/29/2022 19:42:01 - INFO - codeparrot_training - Step 23965: {'lr': 0.00028319924869258425, 'samples': 4601472, 'steps': 23965, 'loss/train': 1.301607370376587} 01/29/2022 19:42:05 - INFO - codeparrot_training - Step 23966: {'lr': 0.0002831830310778504, 'samples': 4601664, 'steps': 23966, 'loss/train': 1.3221142888069153} 01/29/2022 19:42:10 - INFO - codeparrot_training - Step 23967: {'lr': 0.0002831668133209711, 'samples': 4601856, 'steps': 23967, 'loss/train': 1.6231216192245483} 01/29/2022 19:42:14 - INFO - codeparrot_training - Step 23968: {'lr': 0.0002831505954220156, 'samples': 4602048, 'steps': 23968, 'loss/train': 1.7397499680519104} 01/29/2022 19:42:18 - INFO - codeparrot_training - Step 23969: {'lr': 0.00028313437738105353, 'samples': 4602240, 'steps': 23969, 'loss/train': 1.442326694726944} 01/29/2022 19:42:23 - INFO - codeparrot_training - Step 23970: {'lr': 0.0002831181591981543, 'samples': 4602432, 'steps': 23970, 'loss/train': 2.5649591088294983} 01/29/2022 19:42:27 - INFO - codeparrot_training - Step 23971: {'lr': 0.0002831019408733874, 'samples': 4602624, 'steps': 23971, 'loss/train': 1.7808603644371033} 01/29/2022 19:42:32 - INFO - codeparrot_training - Step 23972: {'lr': 0.00028308572240682233, 'samples': 4602816, 'steps': 23972, 'loss/train': 2.0594301223754883} 01/29/2022 19:42:36 - INFO - codeparrot_training - Step 23973: {'lr': 0.00028306950379852844, 'samples': 4603008, 'steps': 23973, 'loss/train': 1.7291666865348816} 01/29/2022 19:42:40 - INFO - codeparrot_training - Step 23974: {'lr': 0.0002830532850485754, 'samples': 4603200, 'steps': 23974, 'loss/train': 1.3687503933906555} 01/29/2022 19:42:46 - INFO - codeparrot_training - Step 23975: {'lr': 0.0002830370661570325, 'samples': 4603392, 'steps': 23975, 'loss/train': 1.598259687423706} 01/29/2022 19:42:50 - INFO - codeparrot_training - Step 23976: {'lr': 0.00028302084712396937, 'samples': 4603584, 'steps': 23976, 'loss/train': 1.7897295355796814} 01/29/2022 19:42:54 - INFO - codeparrot_training - Step 23977: {'lr': 0.00028300462794945535, 'samples': 4603776, 'steps': 23977, 'loss/train': 1.8059313297271729} 01/29/2022 19:42:58 - INFO - codeparrot_training - Step 23978: {'lr': 0.00028298840863356006, 'samples': 4603968, 'steps': 23978, 'loss/train': 1.3636030554771423} 01/29/2022 19:43:03 - INFO - codeparrot_training - Step 23979: {'lr': 0.0002829721891763529, 'samples': 4604160, 'steps': 23979, 'loss/train': 1.709848701953888} 01/29/2022 19:43:10 - INFO - codeparrot_training - Step 23980: {'lr': 0.00028295596957790325, 'samples': 4604352, 'steps': 23980, 'loss/train': 0.4057074040174484} 01/29/2022 19:43:14 - INFO - codeparrot_training - Step 23981: {'lr': 0.0002829397498382808, 'samples': 4604544, 'steps': 23981, 'loss/train': 1.948394000530243} 01/29/2022 19:43:18 - INFO - codeparrot_training - Step 23982: {'lr': 0.00028292352995755487, 'samples': 4604736, 'steps': 23982, 'loss/train': 1.4339393377304077} 01/29/2022 19:43:22 - INFO - codeparrot_training - Step 23983: {'lr': 0.000282907309935795, 'samples': 4604928, 'steps': 23983, 'loss/train': 2.09488445520401} 01/29/2022 19:43:27 - INFO - codeparrot_training - Step 23984: {'lr': 0.00028289108977307066, 'samples': 4605120, 'steps': 23984, 'loss/train': 1.6891711950302124} 01/29/2022 19:43:32 - INFO - codeparrot_training - Step 23985: {'lr': 0.00028287486946945137, 'samples': 4605312, 'steps': 23985, 'loss/train': 1.6502711176872253} 01/29/2022 19:43:36 - INFO - codeparrot_training - Step 23986: {'lr': 0.0002828586490250065, 'samples': 4605504, 'steps': 23986, 'loss/train': 1.5294495820999146} 01/29/2022 19:43:40 - INFO - codeparrot_training - Step 23987: {'lr': 0.00028284242843980566, 'samples': 4605696, 'steps': 23987, 'loss/train': 1.2680242359638214} 01/29/2022 19:43:44 - INFO - codeparrot_training - Step 23988: {'lr': 0.00028282620771391824, 'samples': 4605888, 'steps': 23988, 'loss/train': 2.1197129487991333} 01/29/2022 19:43:49 - INFO - codeparrot_training - Step 23989: {'lr': 0.00028280998684741387, 'samples': 4606080, 'steps': 23989, 'loss/train': 1.4383395910263062} 01/29/2022 19:43:54 - INFO - codeparrot_training - Step 23990: {'lr': 0.00028279376584036187, 'samples': 4606272, 'steps': 23990, 'loss/train': 1.322454571723938} 01/29/2022 19:43:58 - INFO - codeparrot_training - Step 23991: {'lr': 0.0002827775446928318, 'samples': 4606464, 'steps': 23991, 'loss/train': 0.16137124970555305} 01/29/2022 19:44:02 - INFO - codeparrot_training - Step 23992: {'lr': 0.00028276132340489306, 'samples': 4606656, 'steps': 23992, 'loss/train': 1.6955174803733826} 01/29/2022 19:44:06 - INFO - codeparrot_training - Step 23993: {'lr': 0.0002827451019766153, 'samples': 4606848, 'steps': 23993, 'loss/train': 3.67300021648407} 01/29/2022 19:44:11 - INFO - codeparrot_training - Step 23994: {'lr': 0.00028272888040806795, 'samples': 4607040, 'steps': 23994, 'loss/train': 1.2293110191822052} 01/29/2022 19:44:18 - INFO - codeparrot_training - Step 23995: {'lr': 0.0002827126586993204, 'samples': 4607232, 'steps': 23995, 'loss/train': 2.323569595813751} 01/29/2022 19:44:22 - INFO - codeparrot_training - Step 23996: {'lr': 0.0002826964368504422, 'samples': 4607424, 'steps': 23996, 'loss/train': 1.1858670115470886} 01/29/2022 19:44:26 - INFO - codeparrot_training - Step 23997: {'lr': 0.0002826802148615029, 'samples': 4607616, 'steps': 23997, 'loss/train': 2.8211867809295654} 01/29/2022 19:44:31 - INFO - codeparrot_training - Step 23998: {'lr': 0.00028266399273257193, 'samples': 4607808, 'steps': 23998, 'loss/train': 2.0148983001708984} 01/29/2022 19:44:35 - INFO - codeparrot_training - Step 23999: {'lr': 0.0002826477704637188, 'samples': 4608000, 'steps': 23999, 'loss/train': 2.0956292152404785} 01/29/2022 19:44:35 - INFO - codeparrot_training - Evaluating and saving model checkpoint