{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.990948709353001, "eval_steps": 500, "global_step": 930, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0053637277908146165, "grad_norm": 6.338560672252144, "learning_rate": 8.60215053763441e-07, "loss": 1.0751, "step": 1 }, { "epoch": 0.010727455581629233, "grad_norm": 6.34459310995922, "learning_rate": 1.720430107526882e-06, "loss": 1.0724, "step": 2 }, { "epoch": 0.01609118337244385, "grad_norm": 6.24204412097311, "learning_rate": 2.580645161290323e-06, "loss": 1.0656, "step": 3 }, { "epoch": 0.021454911163258466, "grad_norm": 5.836679159124573, "learning_rate": 3.440860215053764e-06, "loss": 1.0643, "step": 4 }, { "epoch": 0.02681863895407308, "grad_norm": 4.516135892252434, "learning_rate": 4.3010752688172045e-06, "loss": 1.0254, "step": 5 }, { "epoch": 0.0321823667448877, "grad_norm": 2.6133078802773384, "learning_rate": 5.161290322580646e-06, "loss": 0.9857, "step": 6 }, { "epoch": 0.03754609453570231, "grad_norm": 2.356431435775483, "learning_rate": 6.021505376344087e-06, "loss": 0.9684, "step": 7 }, { "epoch": 0.04290982232651693, "grad_norm": 3.8129269226867466, "learning_rate": 6.881720430107528e-06, "loss": 0.9487, "step": 8 }, { "epoch": 0.048273550117331546, "grad_norm": 3.837805004980831, "learning_rate": 7.741935483870968e-06, "loss": 0.958, "step": 9 }, { "epoch": 0.05363727790814616, "grad_norm": 3.226513127256882, "learning_rate": 8.602150537634409e-06, "loss": 0.9352, "step": 10 }, { "epoch": 0.05900100569896078, "grad_norm": 3.1817577843556006, "learning_rate": 9.46236559139785e-06, "loss": 0.8991, "step": 11 }, { "epoch": 0.0643647334897754, "grad_norm": 2.734625287133208, "learning_rate": 1.0322580645161291e-05, "loss": 0.8769, "step": 12 }, { "epoch": 0.06972846128059002, "grad_norm": 1.9702603165063668, "learning_rate": 1.118279569892473e-05, "loss": 0.8565, "step": 13 }, { "epoch": 0.07509218907140462, "grad_norm": 1.4326323487408281, "learning_rate": 1.2043010752688173e-05, "loss": 0.8304, "step": 14 }, { "epoch": 0.08045591686221924, "grad_norm": 1.5771781448342905, "learning_rate": 1.2903225806451613e-05, "loss": 0.8223, "step": 15 }, { "epoch": 0.08581964465303386, "grad_norm": 1.2675059911437987, "learning_rate": 1.3763440860215056e-05, "loss": 0.8036, "step": 16 }, { "epoch": 0.09118337244384847, "grad_norm": 0.998013780742343, "learning_rate": 1.4623655913978497e-05, "loss": 0.7903, "step": 17 }, { "epoch": 0.09654710023466309, "grad_norm": 1.0123573361429459, "learning_rate": 1.5483870967741936e-05, "loss": 0.7909, "step": 18 }, { "epoch": 0.10191082802547771, "grad_norm": 0.867381543747644, "learning_rate": 1.6344086021505377e-05, "loss": 0.7807, "step": 19 }, { "epoch": 0.10727455581629232, "grad_norm": 0.8142913834018638, "learning_rate": 1.7204301075268818e-05, "loss": 0.7675, "step": 20 }, { "epoch": 0.11263828360710694, "grad_norm": 0.848005958837766, "learning_rate": 1.806451612903226e-05, "loss": 0.7644, "step": 21 }, { "epoch": 0.11800201139792156, "grad_norm": 0.7118287539298433, "learning_rate": 1.89247311827957e-05, "loss": 0.7569, "step": 22 }, { "epoch": 0.12336573918873617, "grad_norm": 0.5855004670255447, "learning_rate": 1.978494623655914e-05, "loss": 0.7487, "step": 23 }, { "epoch": 0.1287294669795508, "grad_norm": 0.6728204797382786, "learning_rate": 2.0645161290322582e-05, "loss": 0.7416, "step": 24 }, { "epoch": 0.1340931947703654, "grad_norm": 0.5967666869348331, "learning_rate": 2.1505376344086024e-05, "loss": 0.7363, "step": 25 }, { "epoch": 0.13945692256118003, "grad_norm": 0.6160085099956011, "learning_rate": 2.236559139784946e-05, "loss": 0.734, "step": 26 }, { "epoch": 0.14482065035199462, "grad_norm": 0.6363894242487178, "learning_rate": 2.3225806451612906e-05, "loss": 0.7262, "step": 27 }, { "epoch": 0.15018437814280924, "grad_norm": 0.6626486374605112, "learning_rate": 2.4086021505376347e-05, "loss": 0.7188, "step": 28 }, { "epoch": 0.15554810593362386, "grad_norm": 0.7590807753214363, "learning_rate": 2.4946236559139788e-05, "loss": 0.7223, "step": 29 }, { "epoch": 0.16091183372443849, "grad_norm": 0.8043401688063679, "learning_rate": 2.5806451612903226e-05, "loss": 0.7165, "step": 30 }, { "epoch": 0.1662755615152531, "grad_norm": 0.8116558016551599, "learning_rate": 2.6666666666666667e-05, "loss": 0.7128, "step": 31 }, { "epoch": 0.17163928930606773, "grad_norm": 0.7050286092608573, "learning_rate": 2.752688172043011e-05, "loss": 0.709, "step": 32 }, { "epoch": 0.17700301709688232, "grad_norm": 0.605916542924061, "learning_rate": 2.8387096774193552e-05, "loss": 0.7035, "step": 33 }, { "epoch": 0.18236674488769694, "grad_norm": 0.48538643613649285, "learning_rate": 2.9247311827956993e-05, "loss": 0.7025, "step": 34 }, { "epoch": 0.18773047267851156, "grad_norm": 0.4038151178490983, "learning_rate": 3.010752688172043e-05, "loss": 0.7043, "step": 35 }, { "epoch": 0.19309420046932618, "grad_norm": 0.5220993966401706, "learning_rate": 3.096774193548387e-05, "loss": 0.6972, "step": 36 }, { "epoch": 0.1984579282601408, "grad_norm": 1.1471863644219147, "learning_rate": 3.182795698924731e-05, "loss": 0.6992, "step": 37 }, { "epoch": 0.20382165605095542, "grad_norm": 11.637743206594976, "learning_rate": 3.2688172043010754e-05, "loss": 0.7018, "step": 38 }, { "epoch": 0.20918538384177002, "grad_norm": 2.51589445225404, "learning_rate": 3.3548387096774195e-05, "loss": 0.7004, "step": 39 }, { "epoch": 0.21454911163258464, "grad_norm": 0.8841029393673814, "learning_rate": 3.4408602150537636e-05, "loss": 0.6959, "step": 40 }, { "epoch": 0.21991283942339926, "grad_norm": 2.0189356445033613, "learning_rate": 3.526881720430108e-05, "loss": 0.6962, "step": 41 }, { "epoch": 0.22527656721421388, "grad_norm": 0.8438418075337472, "learning_rate": 3.612903225806452e-05, "loss": 0.6948, "step": 42 }, { "epoch": 0.2306402950050285, "grad_norm": 2.280049552210539, "learning_rate": 3.698924731182796e-05, "loss": 0.691, "step": 43 }, { "epoch": 0.23600402279584312, "grad_norm": 1.5063081335495634, "learning_rate": 3.78494623655914e-05, "loss": 0.6911, "step": 44 }, { "epoch": 0.24136775058665771, "grad_norm": 1.8084395582302646, "learning_rate": 3.870967741935484e-05, "loss": 0.6906, "step": 45 }, { "epoch": 0.24673147837747234, "grad_norm": 1.6588924960337472, "learning_rate": 3.956989247311828e-05, "loss": 0.6929, "step": 46 }, { "epoch": 0.25209520616828696, "grad_norm": 1.014045775957985, "learning_rate": 4.0430107526881724e-05, "loss": 0.6908, "step": 47 }, { "epoch": 0.2574589339591016, "grad_norm": 1.9198362282460184, "learning_rate": 4.1290322580645165e-05, "loss": 0.687, "step": 48 }, { "epoch": 0.2628226617499162, "grad_norm": 1.744156574216378, "learning_rate": 4.2150537634408606e-05, "loss": 0.6804, "step": 49 }, { "epoch": 0.2681863895407308, "grad_norm": 1.1668744438759826, "learning_rate": 4.301075268817205e-05, "loss": 0.6784, "step": 50 }, { "epoch": 0.27355011733154544, "grad_norm": 2.3338212759533854, "learning_rate": 4.387096774193548e-05, "loss": 0.6808, "step": 51 }, { "epoch": 0.27891384512236006, "grad_norm": 1.7244150150466557, "learning_rate": 4.473118279569892e-05, "loss": 0.6838, "step": 52 }, { "epoch": 0.2842775729131747, "grad_norm": 2.419331402256656, "learning_rate": 4.559139784946237e-05, "loss": 0.68, "step": 53 }, { "epoch": 0.28964130070398925, "grad_norm": 4.265232664271197, "learning_rate": 4.645161290322581e-05, "loss": 0.6877, "step": 54 }, { "epoch": 0.29500502849480387, "grad_norm": 1.6428312936958347, "learning_rate": 4.731182795698925e-05, "loss": 0.6866, "step": 55 }, { "epoch": 0.3003687562856185, "grad_norm": 1.0611731524926487, "learning_rate": 4.8172043010752693e-05, "loss": 0.6722, "step": 56 }, { "epoch": 0.3057324840764331, "grad_norm": 1.7504655396715425, "learning_rate": 4.9032258064516135e-05, "loss": 0.6823, "step": 57 }, { "epoch": 0.31109621186724773, "grad_norm": 1.4266527434683252, "learning_rate": 4.9892473118279576e-05, "loss": 0.678, "step": 58 }, { "epoch": 0.31645993965806235, "grad_norm": 1.3715035848857164, "learning_rate": 5.075268817204302e-05, "loss": 0.6669, "step": 59 }, { "epoch": 0.32182366744887697, "grad_norm": 1.4341866180855634, "learning_rate": 5.161290322580645e-05, "loss": 0.6809, "step": 60 }, { "epoch": 0.3271873952396916, "grad_norm": 1.5209173251883614, "learning_rate": 5.247311827956989e-05, "loss": 0.6694, "step": 61 }, { "epoch": 0.3325511230305062, "grad_norm": 1.3054315394599638, "learning_rate": 5.333333333333333e-05, "loss": 0.6637, "step": 62 }, { "epoch": 0.33791485082132083, "grad_norm": 1.5264153975136474, "learning_rate": 5.4193548387096774e-05, "loss": 0.6726, "step": 63 }, { "epoch": 0.34327857861213545, "grad_norm": 1.2283408808572964, "learning_rate": 5.505376344086022e-05, "loss": 0.6746, "step": 64 }, { "epoch": 0.3486423064029501, "grad_norm": 1.4653304361520723, "learning_rate": 5.591397849462366e-05, "loss": 0.6734, "step": 65 }, { "epoch": 0.35400603419376464, "grad_norm": 1.1322311838365562, "learning_rate": 5.6774193548387104e-05, "loss": 0.6665, "step": 66 }, { "epoch": 0.35936976198457926, "grad_norm": 1.4758546005558055, "learning_rate": 5.7634408602150545e-05, "loss": 0.6696, "step": 67 }, { "epoch": 0.3647334897753939, "grad_norm": 1.1524101733290173, "learning_rate": 5.8494623655913986e-05, "loss": 0.6698, "step": 68 }, { "epoch": 0.3700972175662085, "grad_norm": 1.7469295396906894, "learning_rate": 5.935483870967743e-05, "loss": 0.6709, "step": 69 }, { "epoch": 0.3754609453570231, "grad_norm": 1.1846563978993125, "learning_rate": 6.021505376344086e-05, "loss": 0.6673, "step": 70 }, { "epoch": 0.38082467314783774, "grad_norm": 1.1459305938713673, "learning_rate": 6.10752688172043e-05, "loss": 0.6575, "step": 71 }, { "epoch": 0.38618840093865237, "grad_norm": 1.8209165087943502, "learning_rate": 6.193548387096774e-05, "loss": 0.6742, "step": 72 }, { "epoch": 0.391552128729467, "grad_norm": 1.111422678506876, "learning_rate": 6.279569892473119e-05, "loss": 0.669, "step": 73 }, { "epoch": 0.3969158565202816, "grad_norm": 1.8170583952074655, "learning_rate": 6.365591397849463e-05, "loss": 0.6677, "step": 74 }, { "epoch": 0.4022795843110962, "grad_norm": 1.1599786495072175, "learning_rate": 6.451612903225807e-05, "loss": 0.6666, "step": 75 }, { "epoch": 0.40764331210191085, "grad_norm": 1.4841350723375708, "learning_rate": 6.537634408602151e-05, "loss": 0.6707, "step": 76 }, { "epoch": 0.41300703989272547, "grad_norm": 1.0958715407789212, "learning_rate": 6.623655913978495e-05, "loss": 0.6593, "step": 77 }, { "epoch": 0.41837076768354003, "grad_norm": 1.4849874395428133, "learning_rate": 6.709677419354839e-05, "loss": 0.6658, "step": 78 }, { "epoch": 0.42373449547435466, "grad_norm": 1.3922476398945494, "learning_rate": 6.795698924731183e-05, "loss": 0.6627, "step": 79 }, { "epoch": 0.4290982232651693, "grad_norm": 1.349253537020458, "learning_rate": 6.881720430107527e-05, "loss": 0.6651, "step": 80 }, { "epoch": 0.4344619510559839, "grad_norm": 1.214908872484667, "learning_rate": 6.967741935483871e-05, "loss": 0.667, "step": 81 }, { "epoch": 0.4398256788467985, "grad_norm": 1.345794151853099, "learning_rate": 7.053763440860215e-05, "loss": 0.663, "step": 82 }, { "epoch": 0.44518940663761314, "grad_norm": 1.8118348713875871, "learning_rate": 7.13978494623656e-05, "loss": 0.6627, "step": 83 }, { "epoch": 0.45055313442842776, "grad_norm": 1.106435642838999, "learning_rate": 7.225806451612904e-05, "loss": 0.6583, "step": 84 }, { "epoch": 0.4559168622192424, "grad_norm": 1.4776781772073297, "learning_rate": 7.311827956989248e-05, "loss": 0.6647, "step": 85 }, { "epoch": 0.461280590010057, "grad_norm": 1.0848264815933542, "learning_rate": 7.397849462365592e-05, "loss": 0.6557, "step": 86 }, { "epoch": 0.4666443178008716, "grad_norm": 1.8994900378306065, "learning_rate": 7.483870967741936e-05, "loss": 0.6561, "step": 87 }, { "epoch": 0.47200804559168624, "grad_norm": 1.1008594371177927, "learning_rate": 7.56989247311828e-05, "loss": 0.6541, "step": 88 }, { "epoch": 0.47737177338250086, "grad_norm": 1.7741666055604592, "learning_rate": 7.655913978494624e-05, "loss": 0.6532, "step": 89 }, { "epoch": 0.48273550117331543, "grad_norm": 1.1479464223196045, "learning_rate": 7.741935483870968e-05, "loss": 0.6639, "step": 90 }, { "epoch": 0.48809922896413005, "grad_norm": 1.3988914997706596, "learning_rate": 7.827956989247312e-05, "loss": 0.6511, "step": 91 }, { "epoch": 0.49346295675494467, "grad_norm": 1.2477219744293453, "learning_rate": 7.913978494623657e-05, "loss": 0.6605, "step": 92 }, { "epoch": 0.4988266845457593, "grad_norm": 0.9097553948609914, "learning_rate": 8e-05, "loss": 0.65, "step": 93 }, { "epoch": 0.5041904123365739, "grad_norm": 1.5836301376704878, "learning_rate": 7.999971824066397e-05, "loss": 0.6542, "step": 94 }, { "epoch": 0.5095541401273885, "grad_norm": 1.4923647833076734, "learning_rate": 7.99988729666253e-05, "loss": 0.6531, "step": 95 }, { "epoch": 0.5149178679182032, "grad_norm": 0.9849105190117686, "learning_rate": 7.999746418979217e-05, "loss": 0.643, "step": 96 }, { "epoch": 0.5202815957090178, "grad_norm": 1.012040139532396, "learning_rate": 7.99954919300114e-05, "loss": 0.6485, "step": 97 }, { "epoch": 0.5256453234998324, "grad_norm": 1.3986606888790023, "learning_rate": 7.999295621506808e-05, "loss": 0.6549, "step": 98 }, { "epoch": 0.531009051290647, "grad_norm": 1.1230453443468684, "learning_rate": 7.998985708068532e-05, "loss": 0.6459, "step": 99 }, { "epoch": 0.5363727790814616, "grad_norm": 1.3637771512073018, "learning_rate": 7.998619457052362e-05, "loss": 0.6398, "step": 100 }, { "epoch": 0.5417365068722763, "grad_norm": 1.11658331412697, "learning_rate": 7.998196873618028e-05, "loss": 0.6486, "step": 101 }, { "epoch": 0.5471002346630909, "grad_norm": 1.2713507835255546, "learning_rate": 7.997717963718872e-05, "loss": 0.6429, "step": 102 }, { "epoch": 0.5524639624539055, "grad_norm": 1.2063346906341978, "learning_rate": 7.997182734101763e-05, "loss": 0.6453, "step": 103 }, { "epoch": 0.5578276902447201, "grad_norm": 1.5786954989331352, "learning_rate": 7.996591192306995e-05, "loss": 0.6481, "step": 104 }, { "epoch": 0.5631914180355347, "grad_norm": 2.037493924714199, "learning_rate": 7.99594334666819e-05, "loss": 0.6452, "step": 105 }, { "epoch": 0.5685551458263494, "grad_norm": 1.076036499954452, "learning_rate": 7.995239206312176e-05, "loss": 0.6366, "step": 106 }, { "epoch": 0.573918873617164, "grad_norm": 2.8730059630882607, "learning_rate": 7.994478781158861e-05, "loss": 0.6585, "step": 107 }, { "epoch": 0.5792826014079785, "grad_norm": 2.526449754135261, "learning_rate": 7.993662081921086e-05, "loss": 0.6701, "step": 108 }, { "epoch": 0.5846463291987931, "grad_norm": 1.4700621099416509, "learning_rate": 7.992789120104486e-05, "loss": 0.6466, "step": 109 }, { "epoch": 0.5900100569896077, "grad_norm": 1.3276483909541417, "learning_rate": 7.991859908007314e-05, "loss": 0.6535, "step": 110 }, { "epoch": 0.5953737847804224, "grad_norm": 0.9621323964779972, "learning_rate": 7.990874458720283e-05, "loss": 0.6381, "step": 111 }, { "epoch": 0.600737512571237, "grad_norm": 1.7905082277492617, "learning_rate": 7.989832786126369e-05, "loss": 0.6477, "step": 112 }, { "epoch": 0.6061012403620516, "grad_norm": 0.7687516563695098, "learning_rate": 7.98873490490062e-05, "loss": 0.6512, "step": 113 }, { "epoch": 0.6114649681528662, "grad_norm": 1.9114078648875095, "learning_rate": 7.987580830509949e-05, "loss": 0.6532, "step": 114 }, { "epoch": 0.6168286959436808, "grad_norm": 1.1288630311317658, "learning_rate": 7.986370579212921e-05, "loss": 0.6521, "step": 115 }, { "epoch": 0.6221924237344955, "grad_norm": 1.7375799757138244, "learning_rate": 7.985104168059514e-05, "loss": 0.6482, "step": 116 }, { "epoch": 0.6275561515253101, "grad_norm": 1.4087756071428208, "learning_rate": 7.983781614890886e-05, "loss": 0.6542, "step": 117 }, { "epoch": 0.6329198793161247, "grad_norm": 1.4278272442020752, "learning_rate": 7.982402938339123e-05, "loss": 0.6467, "step": 118 }, { "epoch": 0.6382836071069393, "grad_norm": 0.9613841064478965, "learning_rate": 7.980968157826976e-05, "loss": 0.6392, "step": 119 }, { "epoch": 0.6436473348977539, "grad_norm": 1.4516378583492695, "learning_rate": 7.97947729356758e-05, "loss": 0.6405, "step": 120 }, { "epoch": 0.6490110626885686, "grad_norm": 0.8053425444626354, "learning_rate": 7.977930366564188e-05, "loss": 0.6413, "step": 121 }, { "epoch": 0.6543747904793832, "grad_norm": 1.0531976112785613, "learning_rate": 7.976327398609851e-05, "loss": 0.6371, "step": 122 }, { "epoch": 0.6597385182701978, "grad_norm": 0.7167882513685747, "learning_rate": 7.974668412287133e-05, "loss": 0.6363, "step": 123 }, { "epoch": 0.6651022460610124, "grad_norm": 0.9378177990169265, "learning_rate": 7.972953430967773e-05, "loss": 0.6341, "step": 124 }, { "epoch": 0.670465973851827, "grad_norm": 0.8552565042049741, "learning_rate": 7.971182478812374e-05, "loss": 0.6324, "step": 125 }, { "epoch": 0.6758297016426417, "grad_norm": 0.7453958012562631, "learning_rate": 7.96935558077005e-05, "loss": 0.63, "step": 126 }, { "epoch": 0.6811934294334563, "grad_norm": 0.814810784183303, "learning_rate": 7.967472762578082e-05, "loss": 0.6363, "step": 127 }, { "epoch": 0.6865571572242709, "grad_norm": 1.048879669741169, "learning_rate": 7.965534050761548e-05, "loss": 0.6316, "step": 128 }, { "epoch": 0.6919208850150855, "grad_norm": 1.5279961244780853, "learning_rate": 7.963539472632956e-05, "loss": 0.6455, "step": 129 }, { "epoch": 0.6972846128059002, "grad_norm": 0.7981994579669985, "learning_rate": 7.961489056291858e-05, "loss": 0.633, "step": 130 }, { "epoch": 0.7026483405967148, "grad_norm": 1.0473944191586428, "learning_rate": 7.95938283062445e-05, "loss": 0.631, "step": 131 }, { "epoch": 0.7080120683875293, "grad_norm": 1.2737207552833516, "learning_rate": 7.957220825303168e-05, "loss": 0.6318, "step": 132 }, { "epoch": 0.7133757961783439, "grad_norm": 0.8894810359025763, "learning_rate": 7.955003070786275e-05, "loss": 0.6269, "step": 133 }, { "epoch": 0.7187395239691585, "grad_norm": 1.122051061656217, "learning_rate": 7.95272959831742e-05, "loss": 0.6337, "step": 134 }, { "epoch": 0.7241032517599731, "grad_norm": 0.7525451166224594, "learning_rate": 7.950400439925207e-05, "loss": 0.6335, "step": 135 }, { "epoch": 0.7294669795507878, "grad_norm": 0.7016690417729449, "learning_rate": 7.948015628422745e-05, "loss": 0.6286, "step": 136 }, { "epoch": 0.7348307073416024, "grad_norm": 0.6789098247390296, "learning_rate": 7.945575197407177e-05, "loss": 0.6329, "step": 137 }, { "epoch": 0.740194435132417, "grad_norm": 0.6404674159758782, "learning_rate": 7.943079181259215e-05, "loss": 0.6287, "step": 138 }, { "epoch": 0.7455581629232316, "grad_norm": 0.9093960160044203, "learning_rate": 7.940527615142653e-05, "loss": 0.6289, "step": 139 }, { "epoch": 0.7509218907140462, "grad_norm": 1.1758965813037474, "learning_rate": 7.937920535003866e-05, "loss": 0.6351, "step": 140 }, { "epoch": 0.7562856185048609, "grad_norm": 0.742370532155386, "learning_rate": 7.935257977571317e-05, "loss": 0.6261, "step": 141 }, { "epoch": 0.7616493462956755, "grad_norm": 0.542139953471737, "learning_rate": 7.932539980355023e-05, "loss": 0.6268, "step": 142 }, { "epoch": 0.7670130740864901, "grad_norm": 0.6206950110046128, "learning_rate": 7.92976658164604e-05, "loss": 0.6202, "step": 143 }, { "epoch": 0.7723768018773047, "grad_norm": 0.729311584778343, "learning_rate": 7.926937820515918e-05, "loss": 0.6212, "step": 144 }, { "epoch": 0.7777405296681194, "grad_norm": 0.654400750098185, "learning_rate": 7.924053736816148e-05, "loss": 0.6182, "step": 145 }, { "epoch": 0.783104257458934, "grad_norm": 0.6195531403475248, "learning_rate": 7.921114371177607e-05, "loss": 0.6245, "step": 146 }, { "epoch": 0.7884679852497486, "grad_norm": 0.7976556334248974, "learning_rate": 7.918119765009979e-05, "loss": 0.6249, "step": 147 }, { "epoch": 0.7938317130405632, "grad_norm": 0.8396887583061823, "learning_rate": 7.915069960501177e-05, "loss": 0.6248, "step": 148 }, { "epoch": 0.7991954408313778, "grad_norm": 0.8502364179472985, "learning_rate": 7.911965000616746e-05, "loss": 0.6194, "step": 149 }, { "epoch": 0.8045591686221925, "grad_norm": 0.9898063147614041, "learning_rate": 7.908804929099256e-05, "loss": 0.62, "step": 150 }, { "epoch": 0.8099228964130071, "grad_norm": 1.2455473972915074, "learning_rate": 7.905589790467694e-05, "loss": 0.6198, "step": 151 }, { "epoch": 0.8152866242038217, "grad_norm": 0.6806078960215027, "learning_rate": 7.902319630016822e-05, "loss": 0.6162, "step": 152 }, { "epoch": 0.8206503519946363, "grad_norm": 0.4476574519688243, "learning_rate": 7.898994493816553e-05, "loss": 0.6214, "step": 153 }, { "epoch": 0.8260140797854509, "grad_norm": 0.8696643453540888, "learning_rate": 7.895614428711296e-05, "loss": 0.6148, "step": 154 }, { "epoch": 0.8313778075762656, "grad_norm": 1.246010486407204, "learning_rate": 7.892179482319297e-05, "loss": 0.6231, "step": 155 }, { "epoch": 0.8367415353670801, "grad_norm": 0.5040497252873622, "learning_rate": 7.888689703031963e-05, "loss": 0.6192, "step": 156 }, { "epoch": 0.8421052631578947, "grad_norm": 0.5624852670118361, "learning_rate": 7.885145140013192e-05, "loss": 0.6173, "step": 157 }, { "epoch": 0.8474689909487093, "grad_norm": 0.9368458860548771, "learning_rate": 7.88154584319867e-05, "loss": 0.6137, "step": 158 }, { "epoch": 0.8528327187395239, "grad_norm": 0.8845383469044678, "learning_rate": 7.87789186329517e-05, "loss": 0.6182, "step": 159 }, { "epoch": 0.8581964465303386, "grad_norm": 0.5894895395765093, "learning_rate": 7.87418325177984e-05, "loss": 0.6091, "step": 160 }, { "epoch": 0.8635601743211532, "grad_norm": 0.5731381526470704, "learning_rate": 7.870420060899476e-05, "loss": 0.6187, "step": 161 }, { "epoch": 0.8689239021119678, "grad_norm": 0.7549017840945705, "learning_rate": 7.866602343669785e-05, "loss": 0.6203, "step": 162 }, { "epoch": 0.8742876299027824, "grad_norm": 0.962856257611003, "learning_rate": 7.862730153874642e-05, "loss": 0.6176, "step": 163 }, { "epoch": 0.879651357693597, "grad_norm": 1.3065614325608164, "learning_rate": 7.858803546065328e-05, "loss": 0.6247, "step": 164 }, { "epoch": 0.8850150854844117, "grad_norm": 0.7514931987981934, "learning_rate": 7.854822575559764e-05, "loss": 0.6167, "step": 165 }, { "epoch": 0.8903788132752263, "grad_norm": 0.6171596328035264, "learning_rate": 7.85078729844173e-05, "loss": 0.6191, "step": 166 }, { "epoch": 0.8957425410660409, "grad_norm": 0.7129708826504616, "learning_rate": 7.846697771560075e-05, "loss": 0.6176, "step": 167 }, { "epoch": 0.9011062688568555, "grad_norm": 1.0448469281639305, "learning_rate": 7.842554052527918e-05, "loss": 0.6186, "step": 168 }, { "epoch": 0.9064699966476701, "grad_norm": 0.8929655231467754, "learning_rate": 7.838356199721836e-05, "loss": 0.6123, "step": 169 }, { "epoch": 0.9118337244384848, "grad_norm": 0.6547241466981105, "learning_rate": 7.834104272281041e-05, "loss": 0.6157, "step": 170 }, { "epoch": 0.9171974522292994, "grad_norm": 0.820018666142171, "learning_rate": 7.829798330106544e-05, "loss": 0.6151, "step": 171 }, { "epoch": 0.922561180020114, "grad_norm": 0.8525569097143616, "learning_rate": 7.825438433860314e-05, "loss": 0.6227, "step": 172 }, { "epoch": 0.9279249078109286, "grad_norm": 0.7094520065383435, "learning_rate": 7.821024644964429e-05, "loss": 0.6123, "step": 173 }, { "epoch": 0.9332886356017432, "grad_norm": 0.5885877988974711, "learning_rate": 7.816557025600196e-05, "loss": 0.6042, "step": 174 }, { "epoch": 0.9386523633925579, "grad_norm": 0.610585149507006, "learning_rate": 7.81203563870729e-05, "loss": 0.6115, "step": 175 }, { "epoch": 0.9440160911833725, "grad_norm": 0.4909861004310254, "learning_rate": 7.807460547982861e-05, "loss": 0.6136, "step": 176 }, { "epoch": 0.9493798189741871, "grad_norm": 0.552485503833097, "learning_rate": 7.802831817880633e-05, "loss": 0.607, "step": 177 }, { "epoch": 0.9547435467650017, "grad_norm": 0.7187876946145139, "learning_rate": 7.798149513610003e-05, "loss": 0.6081, "step": 178 }, { "epoch": 0.9601072745558162, "grad_norm": 0.6564549638726072, "learning_rate": 7.793413701135119e-05, "loss": 0.6092, "step": 179 }, { "epoch": 0.9654710023466309, "grad_norm": 0.7342350726235318, "learning_rate": 7.788624447173948e-05, "loss": 0.6082, "step": 180 }, { "epoch": 0.9708347301374455, "grad_norm": 0.6410201096556201, "learning_rate": 7.783781819197341e-05, "loss": 0.6098, "step": 181 }, { "epoch": 0.9761984579282601, "grad_norm": 0.7068555927099052, "learning_rate": 7.778885885428082e-05, "loss": 0.6159, "step": 182 }, { "epoch": 0.9815621857190747, "grad_norm": 0.972774444436176, "learning_rate": 7.773936714839922e-05, "loss": 0.6077, "step": 183 }, { "epoch": 0.9869259135098893, "grad_norm": 0.982305769799067, "learning_rate": 7.768934377156613e-05, "loss": 0.6074, "step": 184 }, { "epoch": 0.992289641300704, "grad_norm": 0.792349928717434, "learning_rate": 7.76387894285092e-05, "loss": 0.6071, "step": 185 }, { "epoch": 0.9976533690915186, "grad_norm": 0.6150582773472084, "learning_rate": 7.758770483143634e-05, "loss": 0.6055, "step": 186 }, { "epoch": 1.003687562856185, "grad_norm": 1.0092690602500856, "learning_rate": 7.75360907000257e-05, "loss": 1.0004, "step": 187 }, { "epoch": 1.0090512906469997, "grad_norm": 1.4743324999618854, "learning_rate": 7.748394776141539e-05, "loss": 0.5946, "step": 188 }, { "epoch": 1.0144150184378142, "grad_norm": 0.6277281735974461, "learning_rate": 7.743127675019344e-05, "loss": 0.5838, "step": 189 }, { "epoch": 1.019778746228629, "grad_norm": 1.448463639411485, "learning_rate": 7.737807840838728e-05, "loss": 0.5989, "step": 190 }, { "epoch": 1.0251424740194435, "grad_norm": 0.853892200703028, "learning_rate": 7.732435348545341e-05, "loss": 0.5831, "step": 191 }, { "epoch": 1.0305062018102582, "grad_norm": 0.9983332294058066, "learning_rate": 7.727010273826674e-05, "loss": 0.5929, "step": 192 }, { "epoch": 1.0358699296010727, "grad_norm": 0.828704229898241, "learning_rate": 7.721532693111002e-05, "loss": 0.5829, "step": 193 }, { "epoch": 1.0412336573918874, "grad_norm": 0.6753173564927042, "learning_rate": 7.716002683566298e-05, "loss": 0.5876, "step": 194 }, { "epoch": 1.046597385182702, "grad_norm": 0.7363007051587145, "learning_rate": 7.710420323099151e-05, "loss": 0.5817, "step": 195 }, { "epoch": 1.0519611129735167, "grad_norm": 0.6189597764944422, "learning_rate": 7.704785690353674e-05, "loss": 0.5813, "step": 196 }, { "epoch": 1.0573248407643312, "grad_norm": 0.548504465316847, "learning_rate": 7.699098864710385e-05, "loss": 0.5833, "step": 197 }, { "epoch": 1.062688568555146, "grad_norm": 0.5803042076698839, "learning_rate": 7.693359926285095e-05, "loss": 0.5862, "step": 198 }, { "epoch": 1.0680522963459604, "grad_norm": 0.43453807837921604, "learning_rate": 7.687568955927776e-05, "loss": 0.5786, "step": 199 }, { "epoch": 1.0734160241367752, "grad_norm": 0.5542433696373542, "learning_rate": 7.681726035221428e-05, "loss": 0.5712, "step": 200 }, { "epoch": 1.0787797519275897, "grad_norm": 0.616342629969921, "learning_rate": 7.675831246480923e-05, "loss": 0.5765, "step": 201 }, { "epoch": 1.0841434797184042, "grad_norm": 0.5795433672277299, "learning_rate": 7.66988467275185e-05, "loss": 0.5833, "step": 202 }, { "epoch": 1.089507207509219, "grad_norm": 0.4501970786991801, "learning_rate": 7.663886397809341e-05, "loss": 0.5791, "step": 203 }, { "epoch": 1.0948709353000334, "grad_norm": 0.4353910474176125, "learning_rate": 7.657836506156896e-05, "loss": 0.5741, "step": 204 }, { "epoch": 1.1002346630908482, "grad_norm": 0.3477330860396474, "learning_rate": 7.651735083025187e-05, "loss": 0.5753, "step": 205 }, { "epoch": 1.1055983908816627, "grad_norm": 0.3477663464279302, "learning_rate": 7.64558221437086e-05, "loss": 0.5771, "step": 206 }, { "epoch": 1.1109621186724774, "grad_norm": 0.4507396818143408, "learning_rate": 7.639377986875323e-05, "loss": 0.5735, "step": 207 }, { "epoch": 1.116325846463292, "grad_norm": 0.3794218987883004, "learning_rate": 7.63312248794353e-05, "loss": 0.5776, "step": 208 }, { "epoch": 1.1216895742541066, "grad_norm": 0.3574145388894302, "learning_rate": 7.626815805702741e-05, "loss": 0.5776, "step": 209 }, { "epoch": 1.1270533020449212, "grad_norm": 0.473623038030073, "learning_rate": 7.620458029001286e-05, "loss": 0.5687, "step": 210 }, { "epoch": 1.1324170298357359, "grad_norm": 0.513881497788833, "learning_rate": 7.61404924740731e-05, "loss": 0.571, "step": 211 }, { "epoch": 1.1377807576265504, "grad_norm": 0.5631346609022523, "learning_rate": 7.60758955120752e-05, "loss": 0.5784, "step": 212 }, { "epoch": 1.1431444854173651, "grad_norm": 0.6816476567358593, "learning_rate": 7.601079031405899e-05, "loss": 0.575, "step": 213 }, { "epoch": 1.1485082132081796, "grad_norm": 0.852628392408615, "learning_rate": 7.594517779722432e-05, "loss": 0.5799, "step": 214 }, { "epoch": 1.1538719409989944, "grad_norm": 0.8589566472396788, "learning_rate": 7.587905888591818e-05, "loss": 0.5774, "step": 215 }, { "epoch": 1.1592356687898089, "grad_norm": 0.7183732341560192, "learning_rate": 7.581243451162157e-05, "loss": 0.5725, "step": 216 }, { "epoch": 1.1645993965806236, "grad_norm": 0.5764553522343966, "learning_rate": 7.57453056129365e-05, "loss": 0.571, "step": 217 }, { "epoch": 1.1699631243714381, "grad_norm": 0.4929901875121555, "learning_rate": 7.567767313557262e-05, "loss": 0.5757, "step": 218 }, { "epoch": 1.1753268521622529, "grad_norm": 0.4723018655052622, "learning_rate": 7.560953803233407e-05, "loss": 0.5707, "step": 219 }, { "epoch": 1.1806905799530674, "grad_norm": 0.4230581385616684, "learning_rate": 7.554090126310589e-05, "loss": 0.5741, "step": 220 }, { "epoch": 1.1860543077438819, "grad_norm": 0.38511465865648165, "learning_rate": 7.547176379484063e-05, "loss": 0.5703, "step": 221 }, { "epoch": 1.1914180355346966, "grad_norm": 0.3152141271515254, "learning_rate": 7.540212660154462e-05, "loss": 0.5746, "step": 222 }, { "epoch": 1.1967817633255113, "grad_norm": 0.31639051914639366, "learning_rate": 7.533199066426435e-05, "loss": 0.5746, "step": 223 }, { "epoch": 1.2021454911163258, "grad_norm": 0.377554126720378, "learning_rate": 7.526135697107258e-05, "loss": 0.5679, "step": 224 }, { "epoch": 1.2075092189071404, "grad_norm": 0.3964575276420449, "learning_rate": 7.519022651705441e-05, "loss": 0.5703, "step": 225 }, { "epoch": 1.212872946697955, "grad_norm": 0.3112698936775079, "learning_rate": 7.511860030429333e-05, "loss": 0.5679, "step": 226 }, { "epoch": 1.2182366744887696, "grad_norm": 0.32379714073410504, "learning_rate": 7.504647934185706e-05, "loss": 0.5693, "step": 227 }, { "epoch": 1.2236004022795843, "grad_norm": 0.4040411094185397, "learning_rate": 7.497386464578329e-05, "loss": 0.5731, "step": 228 }, { "epoch": 1.2289641300703988, "grad_norm": 0.4908973335455194, "learning_rate": 7.490075723906548e-05, "loss": 0.5698, "step": 229 }, { "epoch": 1.2343278578612136, "grad_norm": 0.5659873818172788, "learning_rate": 7.482715815163833e-05, "loss": 0.5677, "step": 230 }, { "epoch": 1.239691585652028, "grad_norm": 0.6149372973578214, "learning_rate": 7.475306842036336e-05, "loss": 0.5686, "step": 231 }, { "epoch": 1.2450553134428428, "grad_norm": 0.6411124815404463, "learning_rate": 7.467848908901422e-05, "loss": 0.5731, "step": 232 }, { "epoch": 1.2504190412336573, "grad_norm": 0.7525181257988096, "learning_rate": 7.460342120826207e-05, "loss": 0.5634, "step": 233 }, { "epoch": 1.255782769024472, "grad_norm": 0.8271566546622127, "learning_rate": 7.452786583566072e-05, "loss": 0.5741, "step": 234 }, { "epoch": 1.2611464968152866, "grad_norm": 0.886416613492408, "learning_rate": 7.445182403563176e-05, "loss": 0.5728, "step": 235 }, { "epoch": 1.2665102246061013, "grad_norm": 0.8924881545698343, "learning_rate": 7.437529687944951e-05, "loss": 0.5717, "step": 236 }, { "epoch": 1.2718739523969158, "grad_norm": 0.7745460347324375, "learning_rate": 7.429828544522604e-05, "loss": 0.5722, "step": 237 }, { "epoch": 1.2772376801877305, "grad_norm": 0.7338807909601145, "learning_rate": 7.422079081789587e-05, "loss": 0.5695, "step": 238 }, { "epoch": 1.282601407978545, "grad_norm": 0.7633048091889993, "learning_rate": 7.414281408920074e-05, "loss": 0.5671, "step": 239 }, { "epoch": 1.2879651357693598, "grad_norm": 0.7923736165710045, "learning_rate": 7.40643563576742e-05, "loss": 0.5678, "step": 240 }, { "epoch": 1.2933288635601743, "grad_norm": 0.7486989851097655, "learning_rate": 7.398541872862619e-05, "loss": 0.5667, "step": 241 }, { "epoch": 1.298692591350989, "grad_norm": 0.6208239838109744, "learning_rate": 7.390600231412737e-05, "loss": 0.5736, "step": 242 }, { "epoch": 1.3040563191418035, "grad_norm": 0.7404604089133735, "learning_rate": 7.382610823299359e-05, "loss": 0.5727, "step": 243 }, { "epoch": 1.309420046932618, "grad_norm": 0.8945885885911449, "learning_rate": 7.374573761077001e-05, "loss": 0.5711, "step": 244 }, { "epoch": 1.3147837747234328, "grad_norm": 0.7417028846645015, "learning_rate": 7.366489157971525e-05, "loss": 0.5632, "step": 245 }, { "epoch": 1.3201475025142475, "grad_norm": 0.40901103836642044, "learning_rate": 7.358357127878557e-05, "loss": 0.5643, "step": 246 }, { "epoch": 1.325511230305062, "grad_norm": 0.4165629890181651, "learning_rate": 7.35017778536186e-05, "loss": 0.5655, "step": 247 }, { "epoch": 1.3308749580958765, "grad_norm": 0.5243219329421154, "learning_rate": 7.341951245651747e-05, "loss": 0.5633, "step": 248 }, { "epoch": 1.3362386858866913, "grad_norm": 0.4504046228772055, "learning_rate": 7.333677624643431e-05, "loss": 0.566, "step": 249 }, { "epoch": 1.341602413677506, "grad_norm": 0.35827695239534885, "learning_rate": 7.325357038895413e-05, "loss": 0.5743, "step": 250 }, { "epoch": 1.3469661414683205, "grad_norm": 0.32152110100309084, "learning_rate": 7.316989605627825e-05, "loss": 0.5629, "step": 251 }, { "epoch": 1.352329869259135, "grad_norm": 0.30834448489072347, "learning_rate": 7.308575442720796e-05, "loss": 0.5657, "step": 252 }, { "epoch": 1.3576935970499497, "grad_norm": 0.27834088197947626, "learning_rate": 7.300114668712767e-05, "loss": 0.5644, "step": 253 }, { "epoch": 1.3630573248407643, "grad_norm": 0.2294870511600659, "learning_rate": 7.291607402798843e-05, "loss": 0.5663, "step": 254 }, { "epoch": 1.368421052631579, "grad_norm": 0.31049794029517497, "learning_rate": 7.283053764829106e-05, "loss": 0.5662, "step": 255 }, { "epoch": 1.3737847804223935, "grad_norm": 0.3317352484927729, "learning_rate": 7.274453875306922e-05, "loss": 0.5624, "step": 256 }, { "epoch": 1.3791485082132082, "grad_norm": 0.2991395872386746, "learning_rate": 7.265807855387251e-05, "loss": 0.5671, "step": 257 }, { "epoch": 1.3845122360040227, "grad_norm": 0.36719705545840825, "learning_rate": 7.257115826874931e-05, "loss": 0.5696, "step": 258 }, { "epoch": 1.3898759637948375, "grad_norm": 0.42253987097227313, "learning_rate": 7.248377912222974e-05, "loss": 0.5619, "step": 259 }, { "epoch": 1.395239691585652, "grad_norm": 0.45185019659051934, "learning_rate": 7.239594234530831e-05, "loss": 0.5635, "step": 260 }, { "epoch": 1.4006034193764667, "grad_norm": 0.49545413351898454, "learning_rate": 7.23076491754266e-05, "loss": 0.5634, "step": 261 }, { "epoch": 1.4059671471672812, "grad_norm": 0.5802415947227182, "learning_rate": 7.221890085645588e-05, "loss": 0.5686, "step": 262 }, { "epoch": 1.411330874958096, "grad_norm": 0.6878874855186285, "learning_rate": 7.212969863867953e-05, "loss": 0.5648, "step": 263 }, { "epoch": 1.4166946027489105, "grad_norm": 0.815358307587891, "learning_rate": 7.204004377877539e-05, "loss": 0.5676, "step": 264 }, { "epoch": 1.4220583305397252, "grad_norm": 0.7718060228087203, "learning_rate": 7.194993753979818e-05, "loss": 0.567, "step": 265 }, { "epoch": 1.4274220583305397, "grad_norm": 0.5699309161254399, "learning_rate": 7.185938119116161e-05, "loss": 0.5613, "step": 266 }, { "epoch": 1.4327857861213542, "grad_norm": 0.5107074162600784, "learning_rate": 7.176837600862049e-05, "loss": 0.5658, "step": 267 }, { "epoch": 1.438149513912169, "grad_norm": 0.6522642870784296, "learning_rate": 7.167692327425282e-05, "loss": 0.5717, "step": 268 }, { "epoch": 1.4435132417029837, "grad_norm": 0.7567797167312894, "learning_rate": 7.15850242764417e-05, "loss": 0.5678, "step": 269 }, { "epoch": 1.4488769694937982, "grad_norm": 0.5791144265821572, "learning_rate": 7.149268030985714e-05, "loss": 0.5689, "step": 270 }, { "epoch": 1.4542406972846127, "grad_norm": 0.3877687625030928, "learning_rate": 7.139989267543787e-05, "loss": 0.5674, "step": 271 }, { "epoch": 1.4596044250754274, "grad_norm": 0.5209778834900987, "learning_rate": 7.130666268037303e-05, "loss": 0.5599, "step": 272 }, { "epoch": 1.4649681528662422, "grad_norm": 0.467485068394913, "learning_rate": 7.121299163808368e-05, "loss": 0.5666, "step": 273 }, { "epoch": 1.4703318806570567, "grad_norm": 0.3765790034913379, "learning_rate": 7.111888086820435e-05, "loss": 0.5687, "step": 274 }, { "epoch": 1.4756956084478712, "grad_norm": 0.4702163124649037, "learning_rate": 7.102433169656445e-05, "loss": 0.5653, "step": 275 }, { "epoch": 1.481059336238686, "grad_norm": 0.4595483441798163, "learning_rate": 7.092934545516958e-05, "loss": 0.5668, "step": 276 }, { "epoch": 1.4864230640295006, "grad_norm": 0.3806817840099515, "learning_rate": 7.083392348218274e-05, "loss": 0.5575, "step": 277 }, { "epoch": 1.4917867918203152, "grad_norm": 0.2933836832087611, "learning_rate": 7.073806712190551e-05, "loss": 0.566, "step": 278 }, { "epoch": 1.4971505196111297, "grad_norm": 0.29502601995192856, "learning_rate": 7.064177772475912e-05, "loss": 0.5579, "step": 279 }, { "epoch": 1.5025142474019444, "grad_norm": 0.2446527849116459, "learning_rate": 7.054505664726542e-05, "loss": 0.5588, "step": 280 }, { "epoch": 1.5078779751927591, "grad_norm": 0.23256693103187187, "learning_rate": 7.044790525202772e-05, "loss": 0.5575, "step": 281 }, { "epoch": 1.5132417029835736, "grad_norm": 0.31883595054809466, "learning_rate": 7.035032490771165e-05, "loss": 0.5608, "step": 282 }, { "epoch": 1.5186054307743881, "grad_norm": 0.35371535591743564, "learning_rate": 7.025231698902585e-05, "loss": 0.5622, "step": 283 }, { "epoch": 1.5239691585652029, "grad_norm": 0.30500733849822, "learning_rate": 7.015388287670264e-05, "loss": 0.561, "step": 284 }, { "epoch": 1.5293328863560174, "grad_norm": 0.27539890689680196, "learning_rate": 7.005502395747854e-05, "loss": 0.5632, "step": 285 }, { "epoch": 1.534696614146832, "grad_norm": 0.3163401305287628, "learning_rate": 6.995574162407471e-05, "loss": 0.5647, "step": 286 }, { "epoch": 1.5400603419376466, "grad_norm": 0.3537933994086675, "learning_rate": 6.985603727517736e-05, "loss": 0.5559, "step": 287 }, { "epoch": 1.5454240697284614, "grad_norm": 0.40878100621394714, "learning_rate": 6.975591231541805e-05, "loss": 0.5602, "step": 288 }, { "epoch": 1.5507877975192759, "grad_norm": 0.44323300216184397, "learning_rate": 6.96553681553539e-05, "loss": 0.5628, "step": 289 }, { "epoch": 1.5561515253100904, "grad_norm": 0.4339681253670531, "learning_rate": 6.955440621144766e-05, "loss": 0.5622, "step": 290 }, { "epoch": 1.561515253100905, "grad_norm": 0.42574808327461705, "learning_rate": 6.945302790604789e-05, "loss": 0.5656, "step": 291 }, { "epoch": 1.5668789808917198, "grad_norm": 0.49984403935729443, "learning_rate": 6.935123466736878e-05, "loss": 0.5615, "step": 292 }, { "epoch": 1.5722427086825344, "grad_norm": 0.6728978442736605, "learning_rate": 6.92490279294701e-05, "loss": 0.5612, "step": 293 }, { "epoch": 1.5776064364733489, "grad_norm": 0.7429424392949961, "learning_rate": 6.914640913223695e-05, "loss": 0.5666, "step": 294 }, { "epoch": 1.5829701642641636, "grad_norm": 0.6836320262283112, "learning_rate": 6.904337972135958e-05, "loss": 0.5607, "step": 295 }, { "epoch": 1.5883338920549783, "grad_norm": 0.5867869403119099, "learning_rate": 6.893994114831287e-05, "loss": 0.5608, "step": 296 }, { "epoch": 1.5936976198457928, "grad_norm": 0.5953453881284777, "learning_rate": 6.883609487033605e-05, "loss": 0.56, "step": 297 }, { "epoch": 1.5990613476366073, "grad_norm": 0.6338055543705274, "learning_rate": 6.8731842350412e-05, "loss": 0.5641, "step": 298 }, { "epoch": 1.604425075427422, "grad_norm": 0.6648867991249984, "learning_rate": 6.862718505724677e-05, "loss": 0.566, "step": 299 }, { "epoch": 1.6097888032182368, "grad_norm": 0.7077186281640041, "learning_rate": 6.852212446524881e-05, "loss": 0.5603, "step": 300 }, { "epoch": 1.6151525310090513, "grad_norm": 0.7046651305076743, "learning_rate": 6.84166620545083e-05, "loss": 0.5619, "step": 301 }, { "epoch": 1.6205162587998658, "grad_norm": 0.5221466006404026, "learning_rate": 6.831079931077615e-05, "loss": 0.5561, "step": 302 }, { "epoch": 1.6258799865906806, "grad_norm": 0.37944143421384124, "learning_rate": 6.820453772544318e-05, "loss": 0.5631, "step": 303 }, { "epoch": 1.6312437143814953, "grad_norm": 0.5844863656673001, "learning_rate": 6.80978787955191e-05, "loss": 0.5545, "step": 304 }, { "epoch": 1.6366074421723098, "grad_norm": 0.5727271056764134, "learning_rate": 6.799082402361131e-05, "loss": 0.5602, "step": 305 }, { "epoch": 1.6419711699631243, "grad_norm": 0.3690953808626758, "learning_rate": 6.788337491790397e-05, "loss": 0.5613, "step": 306 }, { "epoch": 1.647334897753939, "grad_norm": 0.31408506346634396, "learning_rate": 6.777553299213646e-05, "loss": 0.5587, "step": 307 }, { "epoch": 1.6526986255447536, "grad_norm": 0.36584423383126785, "learning_rate": 6.766729976558226e-05, "loss": 0.5579, "step": 308 }, { "epoch": 1.658062353335568, "grad_norm": 0.34138280754486466, "learning_rate": 6.755867676302747e-05, "loss": 0.5597, "step": 309 }, { "epoch": 1.6634260811263828, "grad_norm": 0.2631015494780705, "learning_rate": 6.744966551474936e-05, "loss": 0.558, "step": 310 }, { "epoch": 1.6687898089171975, "grad_norm": 0.34093472748141784, "learning_rate": 6.734026755649474e-05, "loss": 0.5619, "step": 311 }, { "epoch": 1.674153536708012, "grad_norm": 0.43169917288860177, "learning_rate": 6.723048442945845e-05, "loss": 0.5617, "step": 312 }, { "epoch": 1.6795172644988265, "grad_norm": 0.4064955294037096, "learning_rate": 6.712031768026154e-05, "loss": 0.562, "step": 313 }, { "epoch": 1.6848809922896413, "grad_norm": 0.4126204222058468, "learning_rate": 6.70097688609295e-05, "loss": 0.5592, "step": 314 }, { "epoch": 1.690244720080456, "grad_norm": 0.4791759638439358, "learning_rate": 6.689883952887042e-05, "loss": 0.5612, "step": 315 }, { "epoch": 1.6956084478712705, "grad_norm": 0.5431515205787689, "learning_rate": 6.678753124685306e-05, "loss": 0.5653, "step": 316 }, { "epoch": 1.700972175662085, "grad_norm": 0.6320848556741568, "learning_rate": 6.667584558298481e-05, "loss": 0.5576, "step": 317 }, { "epoch": 1.7063359034528998, "grad_norm": 0.7004122845515078, "learning_rate": 6.656378411068958e-05, "loss": 0.5609, "step": 318 }, { "epoch": 1.7116996312437145, "grad_norm": 0.6089676867039382, "learning_rate": 6.645134840868566e-05, "loss": 0.5647, "step": 319 }, { "epoch": 1.717063359034529, "grad_norm": 0.4272876666051787, "learning_rate": 6.633854006096351e-05, "loss": 0.5579, "step": 320 }, { "epoch": 1.7224270868253435, "grad_norm": 0.4606786107127789, "learning_rate": 6.622536065676338e-05, "loss": 0.5598, "step": 321 }, { "epoch": 1.7277908146161582, "grad_norm": 0.5009778503199392, "learning_rate": 6.611181179055296e-05, "loss": 0.5581, "step": 322 }, { "epoch": 1.733154542406973, "grad_norm": 0.4252208216862443, "learning_rate": 6.599789506200491e-05, "loss": 0.5598, "step": 323 }, { "epoch": 1.7385182701977875, "grad_norm": 0.3441171701556595, "learning_rate": 6.588361207597432e-05, "loss": 0.5592, "step": 324 }, { "epoch": 1.743881997988602, "grad_norm": 0.3879024267741338, "learning_rate": 6.576896444247609e-05, "loss": 0.5579, "step": 325 }, { "epoch": 1.7492457257794167, "grad_norm": 0.30111558862635124, "learning_rate": 6.565395377666228e-05, "loss": 0.5594, "step": 326 }, { "epoch": 1.7546094535702315, "grad_norm": 0.300732536138267, "learning_rate": 6.553858169879934e-05, "loss": 0.56, "step": 327 }, { "epoch": 1.759973181361046, "grad_norm": 0.340331678535412, "learning_rate": 6.542284983424528e-05, "loss": 0.5558, "step": 328 }, { "epoch": 1.7653369091518605, "grad_norm": 0.3902331189981612, "learning_rate": 6.530675981342674e-05, "loss": 0.5586, "step": 329 }, { "epoch": 1.7707006369426752, "grad_norm": 0.43726753368459903, "learning_rate": 6.51903132718161e-05, "loss": 0.5525, "step": 330 }, { "epoch": 1.7760643647334897, "grad_norm": 0.39280624763313293, "learning_rate": 6.507351184990837e-05, "loss": 0.555, "step": 331 }, { "epoch": 1.7814280925243042, "grad_norm": 0.4556458085809101, "learning_rate": 6.49563571931981e-05, "loss": 0.5605, "step": 332 }, { "epoch": 1.786791820315119, "grad_norm": 0.47834893977036586, "learning_rate": 6.48388509521562e-05, "loss": 0.5619, "step": 333 }, { "epoch": 1.7921555481059337, "grad_norm": 0.37688804258392605, "learning_rate": 6.47209947822067e-05, "loss": 0.5591, "step": 334 }, { "epoch": 1.7975192758967482, "grad_norm": 0.31185260137172, "learning_rate": 6.46027903437034e-05, "loss": 0.5601, "step": 335 }, { "epoch": 1.8028830036875627, "grad_norm": 0.38273388647684164, "learning_rate": 6.448423930190653e-05, "loss": 0.5586, "step": 336 }, { "epoch": 1.8082467314783774, "grad_norm": 0.3659890752118683, "learning_rate": 6.43653433269592e-05, "loss": 0.5593, "step": 337 }, { "epoch": 1.8136104592691922, "grad_norm": 0.28697986611796433, "learning_rate": 6.424610409386396e-05, "loss": 0.5531, "step": 338 }, { "epoch": 1.8189741870600067, "grad_norm": 0.3578108779196417, "learning_rate": 6.41265232824592e-05, "loss": 0.5576, "step": 339 }, { "epoch": 1.8243379148508212, "grad_norm": 0.3595477545374786, "learning_rate": 6.40066025773954e-05, "loss": 0.5536, "step": 340 }, { "epoch": 1.829701642641636, "grad_norm": 0.24495176220429252, "learning_rate": 6.388634366811146e-05, "loss": 0.5561, "step": 341 }, { "epoch": 1.8350653704324507, "grad_norm": 0.34910710828586194, "learning_rate": 6.376574824881092e-05, "loss": 0.5554, "step": 342 }, { "epoch": 1.8404290982232652, "grad_norm": 0.4115446924161809, "learning_rate": 6.364481801843802e-05, "loss": 0.556, "step": 343 }, { "epoch": 1.8457928260140797, "grad_norm": 0.45492770981427666, "learning_rate": 6.352355468065386e-05, "loss": 0.5574, "step": 344 }, { "epoch": 1.8511565538048944, "grad_norm": 0.4502627142016844, "learning_rate": 6.34019599438123e-05, "loss": 0.5578, "step": 345 }, { "epoch": 1.8565202815957091, "grad_norm": 0.3764500404139869, "learning_rate": 6.328003552093597e-05, "loss": 0.5554, "step": 346 }, { "epoch": 1.8618840093865237, "grad_norm": 0.3785656825858741, "learning_rate": 6.315778312969208e-05, "loss": 0.5551, "step": 347 }, { "epoch": 1.8672477371773382, "grad_norm": 0.43002728271919993, "learning_rate": 6.303520449236827e-05, "loss": 0.5502, "step": 348 }, { "epoch": 1.872611464968153, "grad_norm": 0.3968933759349479, "learning_rate": 6.291230133584829e-05, "loss": 0.5563, "step": 349 }, { "epoch": 1.8779751927589676, "grad_norm": 0.3445454221288414, "learning_rate": 6.278907539158775e-05, "loss": 0.5593, "step": 350 }, { "epoch": 1.8833389205497821, "grad_norm": 0.26807986583588095, "learning_rate": 6.266552839558965e-05, "loss": 0.5495, "step": 351 }, { "epoch": 1.8887026483405966, "grad_norm": 0.25008221285974236, "learning_rate": 6.254166208837998e-05, "loss": 0.5517, "step": 352 }, { "epoch": 1.8940663761314114, "grad_norm": 0.24820258665637415, "learning_rate": 6.241747821498315e-05, "loss": 0.5522, "step": 353 }, { "epoch": 1.899430103922226, "grad_norm": 0.27580712951039277, "learning_rate": 6.229297852489746e-05, "loss": 0.5583, "step": 354 }, { "epoch": 1.9047938317130404, "grad_norm": 0.25679911776420106, "learning_rate": 6.21681647720704e-05, "loss": 0.5499, "step": 355 }, { "epoch": 1.9101575595038551, "grad_norm": 0.2197401652483747, "learning_rate": 6.204303871487399e-05, "loss": 0.5549, "step": 356 }, { "epoch": 1.9155212872946699, "grad_norm": 0.2598523527517371, "learning_rate": 6.191760211607995e-05, "loss": 0.5581, "step": 357 }, { "epoch": 1.9208850150854844, "grad_norm": 0.25865343855556444, "learning_rate": 6.179185674283493e-05, "loss": 0.5561, "step": 358 }, { "epoch": 1.9262487428762989, "grad_norm": 0.23265157023397873, "learning_rate": 6.166580436663557e-05, "loss": 0.5538, "step": 359 }, { "epoch": 1.9316124706671136, "grad_norm": 0.21694798938619186, "learning_rate": 6.153944676330357e-05, "loss": 0.5488, "step": 360 }, { "epoch": 1.9369761984579283, "grad_norm": 0.2266361665583917, "learning_rate": 6.141278571296064e-05, "loss": 0.5479, "step": 361 }, { "epoch": 1.9423399262487429, "grad_norm": 0.20869559940686064, "learning_rate": 6.128582300000345e-05, "loss": 0.5507, "step": 362 }, { "epoch": 1.9477036540395574, "grad_norm": 0.24712363397279544, "learning_rate": 6.11585604130785e-05, "loss": 0.5521, "step": 363 }, { "epoch": 1.953067381830372, "grad_norm": 0.2625049015072699, "learning_rate": 6.103099974505689e-05, "loss": 0.5538, "step": 364 }, { "epoch": 1.9584311096211868, "grad_norm": 0.26045303486127896, "learning_rate": 6.0903142793009055e-05, "loss": 0.5521, "step": 365 }, { "epoch": 1.9637948374120013, "grad_norm": 0.2729210974767953, "learning_rate": 6.0774991358179505e-05, "loss": 0.5545, "step": 366 }, { "epoch": 1.9691585652028158, "grad_norm": 0.2661858868407906, "learning_rate": 6.064654724596141e-05, "loss": 0.5565, "step": 367 }, { "epoch": 1.9745222929936306, "grad_norm": 0.3161992287956601, "learning_rate": 6.051781226587114e-05, "loss": 0.5566, "step": 368 }, { "epoch": 1.9798860207844453, "grad_norm": 0.3247068065937088, "learning_rate": 6.038878823152283e-05, "loss": 0.5485, "step": 369 }, { "epoch": 1.9852497485752598, "grad_norm": 0.31042344714670567, "learning_rate": 6.0259476960602795e-05, "loss": 0.5529, "step": 370 }, { "epoch": 1.9906134763660743, "grad_norm": 0.28663489099284756, "learning_rate": 6.012988027484392e-05, "loss": 0.553, "step": 371 }, { "epoch": 1.995977204156889, "grad_norm": 0.32048941108555556, "learning_rate": 6.000000000000001e-05, "loss": 0.5491, "step": 372 }, { "epoch": 2.0020113979215557, "grad_norm": 0.5619929857332273, "learning_rate": 5.9869837965820064e-05, "loss": 0.9077, "step": 373 }, { "epoch": 2.00737512571237, "grad_norm": 0.9276754960897458, "learning_rate": 5.973939600602251e-05, "loss": 0.5122, "step": 374 }, { "epoch": 2.0127388535031847, "grad_norm": 1.1885282416108998, "learning_rate": 5.960867595826934e-05, "loss": 0.5236, "step": 375 }, { "epoch": 2.0181025812939994, "grad_norm": 0.6750751948664354, "learning_rate": 5.9477679664140256e-05, "loss": 0.5117, "step": 376 }, { "epoch": 2.0234663090848137, "grad_norm": 0.7850363466297992, "learning_rate": 5.934640896910668e-05, "loss": 0.5088, "step": 377 }, { "epoch": 2.0288300368756285, "grad_norm": 0.9014636634663518, "learning_rate": 5.921486572250582e-05, "loss": 0.5097, "step": 378 }, { "epoch": 2.034193764666443, "grad_norm": 0.8754469407524593, "learning_rate": 5.908305177751457e-05, "loss": 0.5147, "step": 379 }, { "epoch": 2.039557492457258, "grad_norm": 0.4769303794801558, "learning_rate": 5.89509689911234e-05, "loss": 0.5087, "step": 380 }, { "epoch": 2.044921220248072, "grad_norm": 0.7004101867500354, "learning_rate": 5.881861922411023e-05, "loss": 0.5088, "step": 381 }, { "epoch": 2.050284948038887, "grad_norm": 0.518970595004128, "learning_rate": 5.8686004341014175e-05, "loss": 0.5106, "step": 382 }, { "epoch": 2.0556486758297017, "grad_norm": 0.562641653776111, "learning_rate": 5.855312621010932e-05, "loss": 0.5076, "step": 383 }, { "epoch": 2.0610124036205164, "grad_norm": 0.6089557991308883, "learning_rate": 5.841998670337834e-05, "loss": 0.5098, "step": 384 }, { "epoch": 2.0663761314113307, "grad_norm": 0.3745335212494372, "learning_rate": 5.828658769648621e-05, "loss": 0.5107, "step": 385 }, { "epoch": 2.0717398592021454, "grad_norm": 0.5038019010238222, "learning_rate": 5.815293106875369e-05, "loss": 0.5086, "step": 386 }, { "epoch": 2.07710358699296, "grad_norm": 0.4417411581025683, "learning_rate": 5.8019018703130924e-05, "loss": 0.5082, "step": 387 }, { "epoch": 2.082467314783775, "grad_norm": 0.4645390192932457, "learning_rate": 5.788485248617088e-05, "loss": 0.5088, "step": 388 }, { "epoch": 2.087831042574589, "grad_norm": 0.38534363222977375, "learning_rate": 5.775043430800274e-05, "loss": 0.5044, "step": 389 }, { "epoch": 2.093194770365404, "grad_norm": 0.3177873591397125, "learning_rate": 5.761576606230538e-05, "loss": 0.5037, "step": 390 }, { "epoch": 2.0985584981562186, "grad_norm": 0.31327769202934297, "learning_rate": 5.7480849646280536e-05, "loss": 0.5065, "step": 391 }, { "epoch": 2.1039222259470334, "grad_norm": 0.3096007175172773, "learning_rate": 5.7345686960626216e-05, "loss": 0.5089, "step": 392 }, { "epoch": 2.1092859537378477, "grad_norm": 0.3250396986727379, "learning_rate": 5.7210279909509846e-05, "loss": 0.5075, "step": 393 }, { "epoch": 2.1146496815286624, "grad_norm": 0.3335529585038257, "learning_rate": 5.707463040054147e-05, "loss": 0.5034, "step": 394 }, { "epoch": 2.120013409319477, "grad_norm": 0.27510903683348653, "learning_rate": 5.693874034474686e-05, "loss": 0.5058, "step": 395 }, { "epoch": 2.125377137110292, "grad_norm": 0.30416211894239636, "learning_rate": 5.6802611656540605e-05, "loss": 0.5052, "step": 396 }, { "epoch": 2.130740864901106, "grad_norm": 0.3309159515345732, "learning_rate": 5.666624625369915e-05, "loss": 0.5011, "step": 397 }, { "epoch": 2.136104592691921, "grad_norm": 0.2633474754073962, "learning_rate": 5.652964605733378e-05, "loss": 0.5031, "step": 398 }, { "epoch": 2.1414683204827356, "grad_norm": 0.23995436691844568, "learning_rate": 5.6392812991863505e-05, "loss": 0.5013, "step": 399 }, { "epoch": 2.1468320482735503, "grad_norm": 0.21205441071144687, "learning_rate": 5.6255748984988026e-05, "loss": 0.5026, "step": 400 }, { "epoch": 2.1521957760643646, "grad_norm": 0.1978028326665556, "learning_rate": 5.61184559676605e-05, "loss": 0.5029, "step": 401 }, { "epoch": 2.1575595038551794, "grad_norm": 0.23744700406570274, "learning_rate": 5.598093587406042e-05, "loss": 0.5036, "step": 402 }, { "epoch": 2.162923231645994, "grad_norm": 0.22982879915535342, "learning_rate": 5.584319064156628e-05, "loss": 0.5007, "step": 403 }, { "epoch": 2.1682869594368084, "grad_norm": 0.2332943467736946, "learning_rate": 5.570522221072835e-05, "loss": 0.5023, "step": 404 }, { "epoch": 2.173650687227623, "grad_norm": 0.21085441964760665, "learning_rate": 5.5567032525241315e-05, "loss": 0.5009, "step": 405 }, { "epoch": 2.179014415018438, "grad_norm": 0.22875835498415426, "learning_rate": 5.542862353191686e-05, "loss": 0.5004, "step": 406 }, { "epoch": 2.1843781428092526, "grad_norm": 0.2508518468627133, "learning_rate": 5.528999718065629e-05, "loss": 0.5027, "step": 407 }, { "epoch": 2.189741870600067, "grad_norm": 0.2133758915716313, "learning_rate": 5.515115542442305e-05, "loss": 0.5039, "step": 408 }, { "epoch": 2.1951055983908816, "grad_norm": 0.21013033352445007, "learning_rate": 5.501210021921518e-05, "loss": 0.509, "step": 409 }, { "epoch": 2.2004693261816963, "grad_norm": 0.24089413117157768, "learning_rate": 5.48728335240378e-05, "loss": 0.5034, "step": 410 }, { "epoch": 2.205833053972511, "grad_norm": 0.2673713830759583, "learning_rate": 5.47333573008755e-05, "loss": 0.5003, "step": 411 }, { "epoch": 2.2111967817633253, "grad_norm": 0.26775112246936006, "learning_rate": 5.459367351466466e-05, "loss": 0.5023, "step": 412 }, { "epoch": 2.21656050955414, "grad_norm": 0.29448398829721556, "learning_rate": 5.445378413326582e-05, "loss": 0.5026, "step": 413 }, { "epoch": 2.221924237344955, "grad_norm": 0.26959315515895454, "learning_rate": 5.4313691127435975e-05, "loss": 0.5037, "step": 414 }, { "epoch": 2.2272879651357695, "grad_norm": 0.21092593980509475, "learning_rate": 5.417339647080071e-05, "loss": 0.5018, "step": 415 }, { "epoch": 2.232651692926584, "grad_norm": 0.27287941846913594, "learning_rate": 5.40329021398265e-05, "loss": 0.504, "step": 416 }, { "epoch": 2.2380154207173986, "grad_norm": 0.2243145959214538, "learning_rate": 5.389221011379281e-05, "loss": 0.5016, "step": 417 }, { "epoch": 2.2433791485082133, "grad_norm": 0.23263330992814404, "learning_rate": 5.3751322374764254e-05, "loss": 0.496, "step": 418 }, { "epoch": 2.248742876299028, "grad_norm": 0.2512271712538734, "learning_rate": 5.361024090756259e-05, "loss": 0.5035, "step": 419 }, { "epoch": 2.2541066040898423, "grad_norm": 0.2551912835527355, "learning_rate": 5.346896769973886e-05, "loss": 0.5007, "step": 420 }, { "epoch": 2.259470331880657, "grad_norm": 0.22385870672760472, "learning_rate": 5.3327504741545326e-05, "loss": 0.5029, "step": 421 }, { "epoch": 2.2648340596714718, "grad_norm": 0.19743299945786424, "learning_rate": 5.318585402590745e-05, "loss": 0.5028, "step": 422 }, { "epoch": 2.270197787462286, "grad_norm": 0.20277764076614224, "learning_rate": 5.3044017548395804e-05, "loss": 0.5005, "step": 423 }, { "epoch": 2.275561515253101, "grad_norm": 0.19076747830149005, "learning_rate": 5.290199730719798e-05, "loss": 0.4979, "step": 424 }, { "epoch": 2.2809252430439155, "grad_norm": 0.20942827686948592, "learning_rate": 5.275979530309042e-05, "loss": 0.5038, "step": 425 }, { "epoch": 2.2862889708347303, "grad_norm": 0.21537452339336768, "learning_rate": 5.2617413539410236e-05, "loss": 0.5053, "step": 426 }, { "epoch": 2.291652698625545, "grad_norm": 0.25627491704457783, "learning_rate": 5.247485402202697e-05, "loss": 0.4995, "step": 427 }, { "epoch": 2.2970164264163593, "grad_norm": 0.2829522803374003, "learning_rate": 5.2332118759314394e-05, "loss": 0.4969, "step": 428 }, { "epoch": 2.302380154207174, "grad_norm": 0.22098956424737304, "learning_rate": 5.218920976212215e-05, "loss": 0.504, "step": 429 }, { "epoch": 2.3077438819979887, "grad_norm": 0.2239527428611037, "learning_rate": 5.204612904374745e-05, "loss": 0.5001, "step": 430 }, { "epoch": 2.313107609788803, "grad_norm": 0.20447198338227954, "learning_rate": 5.1902878619906694e-05, "loss": 0.5008, "step": 431 }, { "epoch": 2.3184713375796178, "grad_norm": 0.1994784390562372, "learning_rate": 5.1759460508707085e-05, "loss": 0.4954, "step": 432 }, { "epoch": 2.3238350653704325, "grad_norm": 0.20376249784991973, "learning_rate": 5.1615876730618226e-05, "loss": 0.5051, "step": 433 }, { "epoch": 2.3291987931612472, "grad_norm": 0.1777004640090449, "learning_rate": 5.1472129308443616e-05, "loss": 0.5059, "step": 434 }, { "epoch": 2.3345625209520615, "grad_norm": 0.17919423701924383, "learning_rate": 5.132822026729216e-05, "loss": 0.4997, "step": 435 }, { "epoch": 2.3399262487428762, "grad_norm": 0.1952145263768232, "learning_rate": 5.118415163454968e-05, "loss": 0.4964, "step": 436 }, { "epoch": 2.345289976533691, "grad_norm": 0.23136811372687383, "learning_rate": 5.1039925439850244e-05, "loss": 0.4853, "step": 437 }, { "epoch": 2.3506537043245057, "grad_norm": 0.2527914130838122, "learning_rate": 5.0895543715047737e-05, "loss": 0.5063, "step": 438 }, { "epoch": 2.35601743211532, "grad_norm": 0.23487193932679698, "learning_rate": 5.075100849418708e-05, "loss": 0.5067, "step": 439 }, { "epoch": 2.3613811599061347, "grad_norm": 0.17865390789281577, "learning_rate": 5.060632181347568e-05, "loss": 0.5051, "step": 440 }, { "epoch": 2.3667448876969495, "grad_norm": 0.19380856376412295, "learning_rate": 5.046148571125468e-05, "loss": 0.5012, "step": 441 }, { "epoch": 2.3721086154877637, "grad_norm": 0.2136191065502088, "learning_rate": 5.031650222797028e-05, "loss": 0.4994, "step": 442 }, { "epoch": 2.3774723432785785, "grad_norm": 0.26432526505519316, "learning_rate": 5.0171373406144985e-05, "loss": 0.4977, "step": 443 }, { "epoch": 2.382836071069393, "grad_norm": 0.27436611300421115, "learning_rate": 5.002610129034883e-05, "loss": 0.5019, "step": 444 }, { "epoch": 2.388199798860208, "grad_norm": 0.20385491901178016, "learning_rate": 4.9880687927170534e-05, "loss": 0.4977, "step": 445 }, { "epoch": 2.3935635266510227, "grad_norm": 0.16876671632134482, "learning_rate": 4.973513536518875e-05, "loss": 0.4975, "step": 446 }, { "epoch": 2.398927254441837, "grad_norm": 0.1925770688555282, "learning_rate": 4.958944565494314e-05, "loss": 0.5, "step": 447 }, { "epoch": 2.4042909822326517, "grad_norm": 0.20179672034929158, "learning_rate": 4.944362084890548e-05, "loss": 0.4984, "step": 448 }, { "epoch": 2.4096547100234664, "grad_norm": 0.21265428318185095, "learning_rate": 4.929766300145083e-05, "loss": 0.4974, "step": 449 }, { "epoch": 2.4150184378142807, "grad_norm": 0.25978527351861586, "learning_rate": 4.915157416882849e-05, "loss": 0.5055, "step": 450 }, { "epoch": 2.4203821656050954, "grad_norm": 0.21775715781130206, "learning_rate": 4.9005356409133044e-05, "loss": 0.5016, "step": 451 }, { "epoch": 2.42574589339591, "grad_norm": 0.19027697490800918, "learning_rate": 4.88590117822755e-05, "loss": 0.5022, "step": 452 }, { "epoch": 2.431109621186725, "grad_norm": 0.22937605026076494, "learning_rate": 4.871254234995406e-05, "loss": 0.5041, "step": 453 }, { "epoch": 2.436473348977539, "grad_norm": 0.24637066107726857, "learning_rate": 4.856595017562525e-05, "loss": 0.5024, "step": 454 }, { "epoch": 2.441837076768354, "grad_norm": 0.21654301339170334, "learning_rate": 4.8419237324474734e-05, "loss": 0.499, "step": 455 }, { "epoch": 2.4472008045591687, "grad_norm": 0.23644059339454185, "learning_rate": 4.82724058633883e-05, "loss": 0.5013, "step": 456 }, { "epoch": 2.4525645323499834, "grad_norm": 0.22992449590363817, "learning_rate": 4.812545786092269e-05, "loss": 0.507, "step": 457 }, { "epoch": 2.4579282601407977, "grad_norm": 0.23917593718094943, "learning_rate": 4.7978395387276475e-05, "loss": 0.5017, "step": 458 }, { "epoch": 2.4632919879316124, "grad_norm": 0.20812657582063035, "learning_rate": 4.783122051426093e-05, "loss": 0.5019, "step": 459 }, { "epoch": 2.468655715722427, "grad_norm": 0.16762169107118158, "learning_rate": 4.768393531527077e-05, "loss": 0.5096, "step": 460 }, { "epoch": 2.474019443513242, "grad_norm": 0.21282987152563532, "learning_rate": 4.753654186525497e-05, "loss": 0.4999, "step": 461 }, { "epoch": 2.479383171304056, "grad_norm": 0.21742448911503381, "learning_rate": 4.738904224068758e-05, "loss": 0.5036, "step": 462 }, { "epoch": 2.484746899094871, "grad_norm": 0.1925300663006384, "learning_rate": 4.724143851953841e-05, "loss": 0.4977, "step": 463 }, { "epoch": 2.4901106268856856, "grad_norm": 0.18087700742417484, "learning_rate": 4.7093732781243777e-05, "loss": 0.5079, "step": 464 }, { "epoch": 2.4954743546765004, "grad_norm": 0.15639702761113433, "learning_rate": 4.694592710667723e-05, "loss": 0.5072, "step": 465 }, { "epoch": 2.5008380824673146, "grad_norm": 0.19597235332511273, "learning_rate": 4.6798023578120184e-05, "loss": 0.5018, "step": 466 }, { "epoch": 2.5062018102581294, "grad_norm": 0.1684316769198244, "learning_rate": 4.6650024279232666e-05, "loss": 0.4929, "step": 467 }, { "epoch": 2.511565538048944, "grad_norm": 0.21055212869799955, "learning_rate": 4.650193129502386e-05, "loss": 0.5006, "step": 468 }, { "epoch": 2.5169292658397584, "grad_norm": 0.24733144692822004, "learning_rate": 4.635374671182283e-05, "loss": 0.4994, "step": 469 }, { "epoch": 2.522292993630573, "grad_norm": 0.24533597541006635, "learning_rate": 4.620547261724906e-05, "loss": 0.4958, "step": 470 }, { "epoch": 2.527656721421388, "grad_norm": 0.2066880499616437, "learning_rate": 4.605711110018307e-05, "loss": 0.5005, "step": 471 }, { "epoch": 2.5330204492122026, "grad_norm": 0.16991247454378994, "learning_rate": 4.590866425073698e-05, "loss": 0.4937, "step": 472 }, { "epoch": 2.5383841770030173, "grad_norm": 0.1921709883210605, "learning_rate": 4.576013416022511e-05, "loss": 0.4986, "step": 473 }, { "epoch": 2.5437479047938316, "grad_norm": 0.21100456607229467, "learning_rate": 4.5611522921134394e-05, "loss": 0.5096, "step": 474 }, { "epoch": 2.5491116325846463, "grad_norm": 0.21922187130563403, "learning_rate": 4.546283262709506e-05, "loss": 0.4961, "step": 475 }, { "epoch": 2.554475360375461, "grad_norm": 0.2138728259158541, "learning_rate": 4.531406537285103e-05, "loss": 0.4978, "step": 476 }, { "epoch": 2.5598390881662754, "grad_norm": 0.17501395717864893, "learning_rate": 4.516522325423046e-05, "loss": 0.4978, "step": 477 }, { "epoch": 2.56520281595709, "grad_norm": 0.2043784564828151, "learning_rate": 4.5016308368116155e-05, "loss": 0.5071, "step": 478 }, { "epoch": 2.570566543747905, "grad_norm": 0.23306913353410968, "learning_rate": 4.486732281241611e-05, "loss": 0.4983, "step": 479 }, { "epoch": 2.5759302715387196, "grad_norm": 0.2446931512816206, "learning_rate": 4.471826868603385e-05, "loss": 0.4967, "step": 480 }, { "epoch": 2.5812939993295343, "grad_norm": 0.19818051787599317, "learning_rate": 4.456914808883898e-05, "loss": 0.5007, "step": 481 }, { "epoch": 2.5866577271203486, "grad_norm": 0.16087879372692104, "learning_rate": 4.4419963121637526e-05, "loss": 0.4944, "step": 482 }, { "epoch": 2.5920214549111633, "grad_norm": 0.20092825461048774, "learning_rate": 4.427071588614236e-05, "loss": 0.5002, "step": 483 }, { "epoch": 2.597385182701978, "grad_norm": 0.2281785256323823, "learning_rate": 4.412140848494356e-05, "loss": 0.5031, "step": 484 }, { "epoch": 2.6027489104927923, "grad_norm": 0.20204680377867434, "learning_rate": 4.397204302147886e-05, "loss": 0.496, "step": 485 }, { "epoch": 2.608112638283607, "grad_norm": 0.1977652017995082, "learning_rate": 4.3822621600003934e-05, "loss": 0.4959, "step": 486 }, { "epoch": 2.613476366074422, "grad_norm": 0.19244577039841376, "learning_rate": 4.367314632556281e-05, "loss": 0.4949, "step": 487 }, { "epoch": 2.618840093865236, "grad_norm": 0.24680562797983854, "learning_rate": 4.3523619303958196e-05, "loss": 0.5001, "step": 488 }, { "epoch": 2.624203821656051, "grad_norm": 0.2608000397905511, "learning_rate": 4.3374042641721787e-05, "loss": 0.5006, "step": 489 }, { "epoch": 2.6295675494468655, "grad_norm": 0.16561810253679096, "learning_rate": 4.322441844608469e-05, "loss": 0.4972, "step": 490 }, { "epoch": 2.6349312772376803, "grad_norm": 0.22607914890990072, "learning_rate": 4.3074748824947546e-05, "loss": 0.5002, "step": 491 }, { "epoch": 2.640295005028495, "grad_norm": 0.2918667204791213, "learning_rate": 4.292503588685105e-05, "loss": 0.5, "step": 492 }, { "epoch": 2.6456587328193093, "grad_norm": 0.2047093208105323, "learning_rate": 4.277528174094607e-05, "loss": 0.4999, "step": 493 }, { "epoch": 2.651022460610124, "grad_norm": 0.1740999252433492, "learning_rate": 4.262548849696407e-05, "loss": 0.495, "step": 494 }, { "epoch": 2.6563861884009388, "grad_norm": 0.1585431462983745, "learning_rate": 4.247565826518728e-05, "loss": 0.4913, "step": 495 }, { "epoch": 2.661749916191753, "grad_norm": 0.18550331479957272, "learning_rate": 4.2325793156419035e-05, "loss": 0.4991, "step": 496 }, { "epoch": 2.667113643982568, "grad_norm": 0.18017249852635744, "learning_rate": 4.217589528195403e-05, "loss": 0.4952, "step": 497 }, { "epoch": 2.6724773717733825, "grad_norm": 0.2170267245439509, "learning_rate": 4.202596675354851e-05, "loss": 0.4979, "step": 498 }, { "epoch": 2.6778410995641972, "grad_norm": 0.20466183171637414, "learning_rate": 4.187600968339064e-05, "loss": 0.4977, "step": 499 }, { "epoch": 2.683204827355012, "grad_norm": 0.1859091442920982, "learning_rate": 4.1726026184070625e-05, "loss": 0.5037, "step": 500 }, { "epoch": 2.6885685551458263, "grad_norm": 0.16577581911976427, "learning_rate": 4.157601836855103e-05, "loss": 0.4996, "step": 501 }, { "epoch": 2.693932282936641, "grad_norm": 0.20982822011955427, "learning_rate": 4.142598835013698e-05, "loss": 0.5003, "step": 502 }, { "epoch": 2.6992960107274557, "grad_norm": 0.1760144931085274, "learning_rate": 4.12759382424464e-05, "loss": 0.5048, "step": 503 }, { "epoch": 2.70465973851827, "grad_norm": 0.16308891153060442, "learning_rate": 4.11258701593802e-05, "loss": 0.4995, "step": 504 }, { "epoch": 2.7100234663090847, "grad_norm": 0.18485290922951852, "learning_rate": 4.0975786215092596e-05, "loss": 0.4974, "step": 505 }, { "epoch": 2.7153871940998995, "grad_norm": 0.16453966274325355, "learning_rate": 4.0825688523961176e-05, "loss": 0.4961, "step": 506 }, { "epoch": 2.7207509218907138, "grad_norm": 0.17078151543088202, "learning_rate": 4.0675579200557246e-05, "loss": 0.4996, "step": 507 }, { "epoch": 2.7261146496815285, "grad_norm": 0.18413562304090325, "learning_rate": 4.052546035961596e-05, "loss": 0.5047, "step": 508 }, { "epoch": 2.7314783774723432, "grad_norm": 0.1807528229446923, "learning_rate": 4.0375334116006596e-05, "loss": 0.4999, "step": 509 }, { "epoch": 2.736842105263158, "grad_norm": 0.1708195041200041, "learning_rate": 4.0225202584702643e-05, "loss": 0.5039, "step": 510 }, { "epoch": 2.7422058330539727, "grad_norm": 0.1772112061283108, "learning_rate": 4.0075067880752165e-05, "loss": 0.5006, "step": 511 }, { "epoch": 2.747569560844787, "grad_norm": 0.16771636982644764, "learning_rate": 3.9924932119247855e-05, "loss": 0.5008, "step": 512 }, { "epoch": 2.7529332886356017, "grad_norm": 0.16916780256561517, "learning_rate": 3.977479741529738e-05, "loss": 0.4939, "step": 513 }, { "epoch": 2.7582970164264164, "grad_norm": 0.17680260996970482, "learning_rate": 3.962466588399342e-05, "loss": 0.4997, "step": 514 }, { "epoch": 2.7636607442172307, "grad_norm": 0.1780764938281788, "learning_rate": 3.947453964038404e-05, "loss": 0.4939, "step": 515 }, { "epoch": 2.7690244720080455, "grad_norm": 0.20020918095732, "learning_rate": 3.932442079944276e-05, "loss": 0.4932, "step": 516 }, { "epoch": 2.77438819979886, "grad_norm": 0.19304746634023665, "learning_rate": 3.9174311476038824e-05, "loss": 0.5026, "step": 517 }, { "epoch": 2.779751927589675, "grad_norm": 0.19887309023558672, "learning_rate": 3.902421378490742e-05, "loss": 0.4953, "step": 518 }, { "epoch": 2.7851156553804897, "grad_norm": 0.204450641639583, "learning_rate": 3.887412984061979e-05, "loss": 0.5001, "step": 519 }, { "epoch": 2.790479383171304, "grad_norm": 0.1651139970176456, "learning_rate": 3.872406175755362e-05, "loss": 0.4985, "step": 520 }, { "epoch": 2.7958431109621187, "grad_norm": 0.2563097667637881, "learning_rate": 3.857401164986303e-05, "loss": 0.4948, "step": 521 }, { "epoch": 2.8012068387529334, "grad_norm": 0.234054045546983, "learning_rate": 3.842398163144899e-05, "loss": 0.4947, "step": 522 }, { "epoch": 2.8065705665437477, "grad_norm": 0.15811375667315455, "learning_rate": 3.827397381592939e-05, "loss": 0.5048, "step": 523 }, { "epoch": 2.8119342943345624, "grad_norm": 0.1987863698237863, "learning_rate": 3.8123990316609364e-05, "loss": 0.501, "step": 524 }, { "epoch": 2.817298022125377, "grad_norm": 0.17172619805487405, "learning_rate": 3.7974033246451496e-05, "loss": 0.4991, "step": 525 }, { "epoch": 2.822661749916192, "grad_norm": 0.16007263014351186, "learning_rate": 3.782410471804599e-05, "loss": 0.4972, "step": 526 }, { "epoch": 2.8280254777070066, "grad_norm": 0.16534850292679054, "learning_rate": 3.767420684358097e-05, "loss": 0.4984, "step": 527 }, { "epoch": 2.833389205497821, "grad_norm": 0.14773257937726625, "learning_rate": 3.752434173481273e-05, "loss": 0.504, "step": 528 }, { "epoch": 2.8387529332886356, "grad_norm": 0.17349714217162007, "learning_rate": 3.737451150303595e-05, "loss": 0.4984, "step": 529 }, { "epoch": 2.8441166610794504, "grad_norm": 0.15861554785434998, "learning_rate": 3.722471825905394e-05, "loss": 0.5046, "step": 530 }, { "epoch": 2.8494803888702647, "grad_norm": 0.17356548128864052, "learning_rate": 3.707496411314896e-05, "loss": 0.4965, "step": 531 }, { "epoch": 2.8548441166610794, "grad_norm": 0.16137459004699495, "learning_rate": 3.692525117505246e-05, "loss": 0.4966, "step": 532 }, { "epoch": 2.860207844451894, "grad_norm": 0.1543679059675596, "learning_rate": 3.677558155391532e-05, "loss": 0.4969, "step": 533 }, { "epoch": 2.8655715722427084, "grad_norm": 0.17840356309672536, "learning_rate": 3.662595735827822e-05, "loss": 0.4928, "step": 534 }, { "epoch": 2.870935300033523, "grad_norm": 0.15915427976526053, "learning_rate": 3.647638069604182e-05, "loss": 0.491, "step": 535 }, { "epoch": 2.876299027824338, "grad_norm": 0.1960858603291955, "learning_rate": 3.632685367443721e-05, "loss": 0.4981, "step": 536 }, { "epoch": 2.8816627556151526, "grad_norm": 0.20296992977396197, "learning_rate": 3.617737839999608e-05, "loss": 0.4984, "step": 537 }, { "epoch": 2.8870264834059673, "grad_norm": 0.2694415740935397, "learning_rate": 3.602795697852116e-05, "loss": 0.4907, "step": 538 }, { "epoch": 2.8923902111967816, "grad_norm": 0.19651514395841568, "learning_rate": 3.587859151505645e-05, "loss": 0.4907, "step": 539 }, { "epoch": 2.8977539389875964, "grad_norm": 0.17781434805395327, "learning_rate": 3.572928411385765e-05, "loss": 0.5042, "step": 540 }, { "epoch": 2.903117666778411, "grad_norm": 0.20415382211791988, "learning_rate": 3.558003687836249e-05, "loss": 0.4973, "step": 541 }, { "epoch": 2.9084813945692254, "grad_norm": 0.25361323116982154, "learning_rate": 3.5430851911161025e-05, "loss": 0.4957, "step": 542 }, { "epoch": 2.91384512236004, "grad_norm": 0.20713936406930122, "learning_rate": 3.528173131396617e-05, "loss": 0.5036, "step": 543 }, { "epoch": 2.919208850150855, "grad_norm": 0.17026848748787526, "learning_rate": 3.513267718758391e-05, "loss": 0.4975, "step": 544 }, { "epoch": 2.9245725779416696, "grad_norm": 0.21622589822789534, "learning_rate": 3.498369163188385e-05, "loss": 0.4976, "step": 545 }, { "epoch": 2.9299363057324843, "grad_norm": 0.19646465258305665, "learning_rate": 3.483477674576955e-05, "loss": 0.4987, "step": 546 }, { "epoch": 2.9353000335232986, "grad_norm": 0.17789915328753722, "learning_rate": 3.468593462714896e-05, "loss": 0.4952, "step": 547 }, { "epoch": 2.9406637613141133, "grad_norm": 0.16632647338954235, "learning_rate": 3.453716737290495e-05, "loss": 0.498, "step": 548 }, { "epoch": 2.946027489104928, "grad_norm": 0.18269935667563644, "learning_rate": 3.438847707886561e-05, "loss": 0.4993, "step": 549 }, { "epoch": 2.9513912168957424, "grad_norm": 0.1574198748294102, "learning_rate": 3.4239865839774906e-05, "loss": 0.4938, "step": 550 }, { "epoch": 2.956754944686557, "grad_norm": 0.16957001146132789, "learning_rate": 3.409133574926302e-05, "loss": 0.4969, "step": 551 }, { "epoch": 2.962118672477372, "grad_norm": 0.15435240809940032, "learning_rate": 3.394288889981695e-05, "loss": 0.4946, "step": 552 }, { "epoch": 2.9674824002681865, "grad_norm": 0.154260292592301, "learning_rate": 3.379452738275095e-05, "loss": 0.4929, "step": 553 }, { "epoch": 2.9728461280590013, "grad_norm": 0.15155738642592756, "learning_rate": 3.364625328817717e-05, "loss": 0.4972, "step": 554 }, { "epoch": 2.9782098558498156, "grad_norm": 0.17467421089382126, "learning_rate": 3.349806870497615e-05, "loss": 0.5008, "step": 555 }, { "epoch": 2.9835735836406303, "grad_norm": 0.17056105214141762, "learning_rate": 3.334997572076734e-05, "loss": 0.4905, "step": 556 }, { "epoch": 2.988937311431445, "grad_norm": 0.1673670147899835, "learning_rate": 3.320197642187983e-05, "loss": 0.494, "step": 557 }, { "epoch": 2.9943010392222593, "grad_norm": 0.1850966895118181, "learning_rate": 3.305407289332279e-05, "loss": 0.4992, "step": 558 }, { "epoch": 3.000335232986926, "grad_norm": 0.2722101703808084, "learning_rate": 3.2906267218756244e-05, "loss": 0.836, "step": 559 }, { "epoch": 3.0056989607777407, "grad_norm": 0.3632821992997869, "learning_rate": 3.2758561480461606e-05, "loss": 0.4507, "step": 560 }, { "epoch": 3.011062688568555, "grad_norm": 0.3428188312032687, "learning_rate": 3.261095775931244e-05, "loss": 0.46, "step": 561 }, { "epoch": 3.0164264163593697, "grad_norm": 0.2992573935904219, "learning_rate": 3.2463458134745036e-05, "loss": 0.4511, "step": 562 }, { "epoch": 3.0217901441501844, "grad_norm": 0.31765288073417064, "learning_rate": 3.2316064684729246e-05, "loss": 0.4476, "step": 563 }, { "epoch": 3.027153871940999, "grad_norm": 0.26053192252342355, "learning_rate": 3.2168779485739086e-05, "loss": 0.4473, "step": 564 }, { "epoch": 3.0325175997318135, "grad_norm": 0.298616923861466, "learning_rate": 3.2021604612723525e-05, "loss": 0.4521, "step": 565 }, { "epoch": 3.037881327522628, "grad_norm": 0.28423459110220495, "learning_rate": 3.187454213907733e-05, "loss": 0.4507, "step": 566 }, { "epoch": 3.043245055313443, "grad_norm": 0.23009487403692486, "learning_rate": 3.172759413661172e-05, "loss": 0.448, "step": 567 }, { "epoch": 3.0486087831042576, "grad_norm": 0.28376655658935424, "learning_rate": 3.1580762675525286e-05, "loss": 0.4509, "step": 568 }, { "epoch": 3.053972510895072, "grad_norm": 0.22381560292455216, "learning_rate": 3.143404982437476e-05, "loss": 0.4521, "step": 569 }, { "epoch": 3.0593362386858867, "grad_norm": 0.24070713732730523, "learning_rate": 3.128745765004594e-05, "loss": 0.4439, "step": 570 }, { "epoch": 3.0646999664767014, "grad_norm": 0.2632500650736694, "learning_rate": 3.114098821772451e-05, "loss": 0.4489, "step": 571 }, { "epoch": 3.070063694267516, "grad_norm": 0.19380561177767147, "learning_rate": 3.099464359086695e-05, "loss": 0.4473, "step": 572 }, { "epoch": 3.0754274220583304, "grad_norm": 0.28710006884974615, "learning_rate": 3.084842583117153e-05, "loss": 0.4527, "step": 573 }, { "epoch": 3.080791149849145, "grad_norm": 0.18312560702847464, "learning_rate": 3.0702336998549175e-05, "loss": 0.451, "step": 574 }, { "epoch": 3.08615487763996, "grad_norm": 0.23079325997690583, "learning_rate": 3.055637915109453e-05, "loss": 0.4427, "step": 575 }, { "epoch": 3.0915186054307746, "grad_norm": 0.19698651608881707, "learning_rate": 3.0410554345056876e-05, "loss": 0.4534, "step": 576 }, { "epoch": 3.096882333221589, "grad_norm": 0.18838697788327152, "learning_rate": 3.026486463481125e-05, "loss": 0.451, "step": 577 }, { "epoch": 3.1022460610124036, "grad_norm": 0.17680792058130787, "learning_rate": 3.0119312072829476e-05, "loss": 0.4435, "step": 578 }, { "epoch": 3.1076097888032184, "grad_norm": 0.1601761416116848, "learning_rate": 2.997389870965118e-05, "loss": 0.4511, "step": 579 }, { "epoch": 3.1129735165940327, "grad_norm": 0.17773164119539953, "learning_rate": 2.982862659385502e-05, "loss": 0.4513, "step": 580 }, { "epoch": 3.1183372443848474, "grad_norm": 0.17193804443547908, "learning_rate": 2.968349777202973e-05, "loss": 0.4484, "step": 581 }, { "epoch": 3.123700972175662, "grad_norm": 0.16914799438808237, "learning_rate": 2.9538514288745336e-05, "loss": 0.4463, "step": 582 }, { "epoch": 3.129064699966477, "grad_norm": 0.16873259416647984, "learning_rate": 2.939367818652434e-05, "loss": 0.4491, "step": 583 }, { "epoch": 3.134428427757291, "grad_norm": 0.16119241415751986, "learning_rate": 2.9248991505812944e-05, "loss": 0.4473, "step": 584 }, { "epoch": 3.139792155548106, "grad_norm": 0.15930346566056242, "learning_rate": 2.9104456284952277e-05, "loss": 0.4464, "step": 585 }, { "epoch": 3.1451558833389206, "grad_norm": 0.15103719252838246, "learning_rate": 2.8960074560149752e-05, "loss": 0.4458, "step": 586 }, { "epoch": 3.1505196111297353, "grad_norm": 0.14739728413169334, "learning_rate": 2.8815848365450336e-05, "loss": 0.4461, "step": 587 }, { "epoch": 3.1558833389205496, "grad_norm": 0.1636187894417558, "learning_rate": 2.867177973270784e-05, "loss": 0.4444, "step": 588 }, { "epoch": 3.1612470667113644, "grad_norm": 0.15742478053043885, "learning_rate": 2.8527870691556404e-05, "loss": 0.4476, "step": 589 }, { "epoch": 3.166610794502179, "grad_norm": 0.611468578577517, "learning_rate": 2.8384123269381784e-05, "loss": 0.4607, "step": 590 }, { "epoch": 3.171974522292994, "grad_norm": 0.17061785943367402, "learning_rate": 2.8240539491292938e-05, "loss": 0.4518, "step": 591 }, { "epoch": 3.177338250083808, "grad_norm": 0.14588623033185896, "learning_rate": 2.8097121380093323e-05, "loss": 0.4404, "step": 592 }, { "epoch": 3.182701977874623, "grad_norm": 0.18047802738317803, "learning_rate": 2.7953870956252562e-05, "loss": 0.4472, "step": 593 }, { "epoch": 3.1880657056654376, "grad_norm": 0.15324942876302464, "learning_rate": 2.7810790237877857e-05, "loss": 0.4471, "step": 594 }, { "epoch": 3.1934294334562523, "grad_norm": 0.19424542098598574, "learning_rate": 2.7667881240685606e-05, "loss": 0.4441, "step": 595 }, { "epoch": 3.1987931612470666, "grad_norm": 0.16898760852013042, "learning_rate": 2.7525145977973045e-05, "loss": 0.4435, "step": 596 }, { "epoch": 3.2041568890378813, "grad_norm": 0.19404967889677557, "learning_rate": 2.738258646058978e-05, "loss": 0.4427, "step": 597 }, { "epoch": 3.209520616828696, "grad_norm": 0.16937723295023943, "learning_rate": 2.7240204696909603e-05, "loss": 0.4493, "step": 598 }, { "epoch": 3.2148843446195103, "grad_norm": 0.14987599858224868, "learning_rate": 2.7098002692802033e-05, "loss": 0.4482, "step": 599 }, { "epoch": 3.220248072410325, "grad_norm": 0.14816239745817442, "learning_rate": 2.6955982451604206e-05, "loss": 0.4512, "step": 600 }, { "epoch": 3.22561180020114, "grad_norm": 0.13895435016957602, "learning_rate": 2.6814145974092566e-05, "loss": 0.4479, "step": 601 }, { "epoch": 3.2309755279919545, "grad_norm": 0.13200544957430468, "learning_rate": 2.6672495258454678e-05, "loss": 0.4543, "step": 602 }, { "epoch": 3.236339255782769, "grad_norm": 0.1456637489059305, "learning_rate": 2.6531032300261153e-05, "loss": 0.4475, "step": 603 }, { "epoch": 3.2417029835735836, "grad_norm": 0.13818311152408982, "learning_rate": 2.6389759092437418e-05, "loss": 0.4446, "step": 604 }, { "epoch": 3.2470667113643983, "grad_norm": 0.1330695845217543, "learning_rate": 2.6248677625235763e-05, "loss": 0.4513, "step": 605 }, { "epoch": 3.252430439155213, "grad_norm": 0.12671554312337807, "learning_rate": 2.6107789886207195e-05, "loss": 0.443, "step": 606 }, { "epoch": 3.2577941669460273, "grad_norm": 0.148616489659857, "learning_rate": 2.5967097860173514e-05, "loss": 0.4494, "step": 607 }, { "epoch": 3.263157894736842, "grad_norm": 0.13556510213657127, "learning_rate": 2.58266035291993e-05, "loss": 0.4535, "step": 608 }, { "epoch": 3.2685216225276568, "grad_norm": 0.15241618420915304, "learning_rate": 2.5686308872564028e-05, "loss": 0.4496, "step": 609 }, { "epoch": 3.2738853503184715, "grad_norm": 0.144757212701804, "learning_rate": 2.5546215866734185e-05, "loss": 0.4465, "step": 610 }, { "epoch": 3.279249078109286, "grad_norm": 0.14396415920602024, "learning_rate": 2.540632648533536e-05, "loss": 0.448, "step": 611 }, { "epoch": 3.2846128059001005, "grad_norm": 0.15210625698722938, "learning_rate": 2.526664269912452e-05, "loss": 0.4453, "step": 612 }, { "epoch": 3.2899765336909153, "grad_norm": 0.1536053230061205, "learning_rate": 2.5127166475962205e-05, "loss": 0.4487, "step": 613 }, { "epoch": 3.29534026148173, "grad_norm": 0.15509713077990947, "learning_rate": 2.4987899780784836e-05, "loss": 0.4457, "step": 614 }, { "epoch": 3.3007039892725443, "grad_norm": 0.13085919728595163, "learning_rate": 2.484884457557696e-05, "loss": 0.4443, "step": 615 }, { "epoch": 3.306067717063359, "grad_norm": 0.16886670984395816, "learning_rate": 2.4710002819343712e-05, "loss": 0.4538, "step": 616 }, { "epoch": 3.3114314448541737, "grad_norm": 0.12562181708750828, "learning_rate": 2.457137646808315e-05, "loss": 0.4477, "step": 617 }, { "epoch": 3.316795172644988, "grad_norm": 0.1577346712555077, "learning_rate": 2.44329674747587e-05, "loss": 0.4542, "step": 618 }, { "epoch": 3.3221589004358028, "grad_norm": 0.13554447992255508, "learning_rate": 2.4294777789271663e-05, "loss": 0.4495, "step": 619 }, { "epoch": 3.3275226282266175, "grad_norm": 0.14611739712605668, "learning_rate": 2.4156809358433728e-05, "loss": 0.4472, "step": 620 }, { "epoch": 3.332886356017432, "grad_norm": 0.1400607484697624, "learning_rate": 2.4019064125939603e-05, "loss": 0.4521, "step": 621 }, { "epoch": 3.338250083808247, "grad_norm": 0.14259331457489, "learning_rate": 2.3881544032339506e-05, "loss": 0.448, "step": 622 }, { "epoch": 3.3436138115990612, "grad_norm": 0.13315687312240615, "learning_rate": 2.3744251015011987e-05, "loss": 0.4463, "step": 623 }, { "epoch": 3.348977539389876, "grad_norm": 0.12598643415214605, "learning_rate": 2.360718700813651e-05, "loss": 0.4447, "step": 624 }, { "epoch": 3.3543412671806907, "grad_norm": 0.14066053069067935, "learning_rate": 2.347035394266623e-05, "loss": 0.446, "step": 625 }, { "epoch": 3.359704994971505, "grad_norm": 0.12802549460203494, "learning_rate": 2.333375374630086e-05, "loss": 0.4419, "step": 626 }, { "epoch": 3.3650687227623197, "grad_norm": 0.12761654852469786, "learning_rate": 2.3197388343459405e-05, "loss": 0.4501, "step": 627 }, { "epoch": 3.3704324505531345, "grad_norm": 0.14182894960939563, "learning_rate": 2.3061259655253165e-05, "loss": 0.4463, "step": 628 }, { "epoch": 3.375796178343949, "grad_norm": 0.41133093836911544, "learning_rate": 2.2925369599458543e-05, "loss": 0.4551, "step": 629 }, { "epoch": 3.3811599061347635, "grad_norm": 0.13182899206147553, "learning_rate": 2.2789720090490167e-05, "loss": 0.4446, "step": 630 }, { "epoch": 3.386523633925578, "grad_norm": 0.1170971831574601, "learning_rate": 2.265431303937379e-05, "loss": 0.4417, "step": 631 }, { "epoch": 3.391887361716393, "grad_norm": 0.12294985745042769, "learning_rate": 2.2519150353719478e-05, "loss": 0.4494, "step": 632 }, { "epoch": 3.3972510895072077, "grad_norm": 0.13039963065660728, "learning_rate": 2.2384233937694626e-05, "loss": 0.447, "step": 633 }, { "epoch": 3.402614817298022, "grad_norm": 0.11831506164421687, "learning_rate": 2.2249565691997263e-05, "loss": 0.4499, "step": 634 }, { "epoch": 3.4079785450888367, "grad_norm": 0.12143409559233521, "learning_rate": 2.2115147513829145e-05, "loss": 0.4476, "step": 635 }, { "epoch": 3.4133422728796514, "grad_norm": 0.1291445962430486, "learning_rate": 2.1980981296869083e-05, "loss": 0.4486, "step": 636 }, { "epoch": 3.418706000670466, "grad_norm": 0.13989097404575776, "learning_rate": 2.184706893124633e-05, "loss": 0.4387, "step": 637 }, { "epoch": 3.4240697284612804, "grad_norm": 0.3715077464881586, "learning_rate": 2.1713412303513803e-05, "loss": 0.4472, "step": 638 }, { "epoch": 3.429433456252095, "grad_norm": 0.1251774557356555, "learning_rate": 2.1580013296621657e-05, "loss": 0.4539, "step": 639 }, { "epoch": 3.43479718404291, "grad_norm": 0.13346719081780015, "learning_rate": 2.144687378989069e-05, "loss": 0.4494, "step": 640 }, { "epoch": 3.4401609118337246, "grad_norm": 0.13105974532551945, "learning_rate": 2.1313995658985825e-05, "loss": 0.4457, "step": 641 }, { "epoch": 3.445524639624539, "grad_norm": 0.12874889844793144, "learning_rate": 2.118138077588978e-05, "loss": 0.4471, "step": 642 }, { "epoch": 3.4508883674153537, "grad_norm": 0.13187822242861885, "learning_rate": 2.1049031008876603e-05, "loss": 0.4486, "step": 643 }, { "epoch": 3.4562520952061684, "grad_norm": 0.11696538375561746, "learning_rate": 2.0916948222485446e-05, "loss": 0.4509, "step": 644 }, { "epoch": 3.4616158229969827, "grad_norm": 0.121989247700871, "learning_rate": 2.0785134277494202e-05, "loss": 0.4462, "step": 645 }, { "epoch": 3.4669795507877974, "grad_norm": 0.893427164640271, "learning_rate": 2.065359103089333e-05, "loss": 0.4619, "step": 646 }, { "epoch": 3.472343278578612, "grad_norm": 0.45677961197929307, "learning_rate": 2.0522320335859768e-05, "loss": 0.4513, "step": 647 }, { "epoch": 3.477707006369427, "grad_norm": 0.13666000173799434, "learning_rate": 2.0391324041730665e-05, "loss": 0.4538, "step": 648 }, { "epoch": 3.4830707341602416, "grad_norm": 0.2608095297691742, "learning_rate": 2.026060399397751e-05, "loss": 0.4489, "step": 649 }, { "epoch": 3.488434461951056, "grad_norm": 0.18107927299339568, "learning_rate": 2.013016203417994e-05, "loss": 0.4526, "step": 650 }, { "epoch": 3.4937981897418706, "grad_norm": 0.14754327484468957, "learning_rate": 2.0000000000000012e-05, "loss": 0.4467, "step": 651 }, { "epoch": 3.4991619175326854, "grad_norm": 0.13398053465109833, "learning_rate": 1.9870119725156094e-05, "loss": 0.4526, "step": 652 }, { "epoch": 3.5045256453234996, "grad_norm": 0.13952089135435625, "learning_rate": 1.9740523039397225e-05, "loss": 0.4407, "step": 653 }, { "epoch": 3.5098893731143144, "grad_norm": 0.14080074484432106, "learning_rate": 1.9611211768477173e-05, "loss": 0.4454, "step": 654 }, { "epoch": 3.515253100905129, "grad_norm": 0.1198388915000314, "learning_rate": 1.948218773412886e-05, "loss": 0.4492, "step": 655 }, { "epoch": 3.520616828695944, "grad_norm": 0.13993217838224647, "learning_rate": 1.935345275403859e-05, "loss": 0.44, "step": 656 }, { "epoch": 3.5259805564867586, "grad_norm": 0.12889045963571127, "learning_rate": 1.9225008641820498e-05, "loss": 0.4443, "step": 657 }, { "epoch": 3.531344284277573, "grad_norm": 0.12378896201897356, "learning_rate": 1.909685720699096e-05, "loss": 0.4456, "step": 658 }, { "epoch": 3.5367080120683876, "grad_norm": 0.1276674265766728, "learning_rate": 1.8969000254943125e-05, "loss": 0.4481, "step": 659 }, { "epoch": 3.5420717398592023, "grad_norm": 0.12390916606673247, "learning_rate": 1.8841439586921515e-05, "loss": 0.4451, "step": 660 }, { "epoch": 3.5474354676500166, "grad_norm": 0.12273677799318142, "learning_rate": 1.871417699999656e-05, "loss": 0.4502, "step": 661 }, { "epoch": 3.5527991954408313, "grad_norm": 0.11996313423286693, "learning_rate": 1.858721428703937e-05, "loss": 0.4561, "step": 662 }, { "epoch": 3.558162923231646, "grad_norm": 0.12358256215052507, "learning_rate": 1.8460553236696448e-05, "loss": 0.4475, "step": 663 }, { "epoch": 3.5635266510224604, "grad_norm": 1.460296103471973, "learning_rate": 1.8334195633364435e-05, "loss": 0.4528, "step": 664 }, { "epoch": 3.568890378813275, "grad_norm": 0.12331491175077743, "learning_rate": 1.8208143257165085e-05, "loss": 0.4509, "step": 665 }, { "epoch": 3.57425410660409, "grad_norm": 0.11750213968895096, "learning_rate": 1.808239788392006e-05, "loss": 0.4489, "step": 666 }, { "epoch": 3.5796178343949046, "grad_norm": 0.11714707650553612, "learning_rate": 1.795696128512603e-05, "loss": 0.4466, "step": 667 }, { "epoch": 3.5849815621857193, "grad_norm": 0.12196764778181841, "learning_rate": 1.78318352279296e-05, "loss": 0.4447, "step": 668 }, { "epoch": 3.5903452899765336, "grad_norm": 0.12562029298509103, "learning_rate": 1.7707021475102548e-05, "loss": 0.4533, "step": 669 }, { "epoch": 3.5957090177673483, "grad_norm": 0.14349002013949913, "learning_rate": 1.758252178501686e-05, "loss": 0.4436, "step": 670 }, { "epoch": 3.601072745558163, "grad_norm": 0.11537479526408528, "learning_rate": 1.745833791162003e-05, "loss": 0.4482, "step": 671 }, { "epoch": 3.6064364733489773, "grad_norm": 0.3056266641244182, "learning_rate": 1.7334471604410367e-05, "loss": 0.4419, "step": 672 }, { "epoch": 3.611800201139792, "grad_norm": 0.11213414375074307, "learning_rate": 1.7210924608412257e-05, "loss": 0.4517, "step": 673 }, { "epoch": 3.617163928930607, "grad_norm": 0.13465101181054254, "learning_rate": 1.7087698664151724e-05, "loss": 0.4386, "step": 674 }, { "epoch": 3.6225276567214215, "grad_norm": 0.1265640161713725, "learning_rate": 1.6964795507631745e-05, "loss": 0.4495, "step": 675 }, { "epoch": 3.6278913845122363, "grad_norm": 0.13249341487141286, "learning_rate": 1.684221687030793e-05, "loss": 0.4505, "step": 676 }, { "epoch": 3.6332551123030505, "grad_norm": 0.10926988434260812, "learning_rate": 1.671996447906403e-05, "loss": 0.4507, "step": 677 }, { "epoch": 3.6386188400938653, "grad_norm": 0.1160947830417163, "learning_rate": 1.659804005618769e-05, "loss": 0.4548, "step": 678 }, { "epoch": 3.64398256788468, "grad_norm": 0.14474625249969114, "learning_rate": 1.6476445319346143e-05, "loss": 0.4415, "step": 679 }, { "epoch": 3.6493462956754943, "grad_norm": 0.11464961989670115, "learning_rate": 1.6355181981561976e-05, "loss": 0.4493, "step": 680 }, { "epoch": 3.654710023466309, "grad_norm": 0.11514676338990276, "learning_rate": 1.62342517511891e-05, "loss": 0.4464, "step": 681 }, { "epoch": 3.6600737512571238, "grad_norm": 0.10388737077706772, "learning_rate": 1.6113656331888563e-05, "loss": 0.4431, "step": 682 }, { "epoch": 3.665437479047938, "grad_norm": 0.21254990175343727, "learning_rate": 1.599339742260463e-05, "loss": 0.4494, "step": 683 }, { "epoch": 3.6708012068387528, "grad_norm": 0.11018791870383089, "learning_rate": 1.5873476717540818e-05, "loss": 0.4453, "step": 684 }, { "epoch": 3.6761649346295675, "grad_norm": 0.11202984963659915, "learning_rate": 1.575389590613604e-05, "loss": 0.4421, "step": 685 }, { "epoch": 3.6815286624203822, "grad_norm": 0.5609794652204741, "learning_rate": 1.5634656673040824e-05, "loss": 0.4453, "step": 686 }, { "epoch": 3.686892390211197, "grad_norm": 0.12056021537980104, "learning_rate": 1.5515760698093485e-05, "loss": 0.4422, "step": 687 }, { "epoch": 3.6922561180020113, "grad_norm": 0.1143368342750797, "learning_rate": 1.539720965629661e-05, "loss": 0.4459, "step": 688 }, { "epoch": 3.697619845792826, "grad_norm": 0.12221336375445264, "learning_rate": 1.5279005217793307e-05, "loss": 0.4454, "step": 689 }, { "epoch": 3.7029835735836407, "grad_norm": 0.12874007735808712, "learning_rate": 1.5161149047843813e-05, "loss": 0.4501, "step": 690 }, { "epoch": 3.708347301374455, "grad_norm": 0.23200442640856928, "learning_rate": 1.504364280680191e-05, "loss": 0.4421, "step": 691 }, { "epoch": 3.7137110291652697, "grad_norm": 0.118738423956062, "learning_rate": 1.492648815009163e-05, "loss": 0.4473, "step": 692 }, { "epoch": 3.7190747569560845, "grad_norm": 0.11440430785036737, "learning_rate": 1.4809686728183903e-05, "loss": 0.4441, "step": 693 }, { "epoch": 3.724438484746899, "grad_norm": 0.11276169565585069, "learning_rate": 1.4693240186573267e-05, "loss": 0.4441, "step": 694 }, { "epoch": 3.729802212537714, "grad_norm": 0.10878975487213734, "learning_rate": 1.4577150165754739e-05, "loss": 0.4455, "step": 695 }, { "epoch": 3.7351659403285282, "grad_norm": 0.11093388920860042, "learning_rate": 1.4461418301200665e-05, "loss": 0.4422, "step": 696 }, { "epoch": 3.740529668119343, "grad_norm": 0.10930519937275324, "learning_rate": 1.4346046223337737e-05, "loss": 0.4458, "step": 697 }, { "epoch": 3.7458933959101577, "grad_norm": 0.11767456655079721, "learning_rate": 1.4231035557523925e-05, "loss": 0.4455, "step": 698 }, { "epoch": 3.751257123700972, "grad_norm": 0.11061581216489039, "learning_rate": 1.4116387924025703e-05, "loss": 0.4451, "step": 699 }, { "epoch": 3.7566208514917867, "grad_norm": 0.12532437758806012, "learning_rate": 1.4002104937995103e-05, "loss": 0.4435, "step": 700 }, { "epoch": 3.7619845792826014, "grad_norm": 0.1101987534346296, "learning_rate": 1.388818820944704e-05, "loss": 0.4445, "step": 701 }, { "epoch": 3.767348307073416, "grad_norm": 0.11241200124363177, "learning_rate": 1.377463934323663e-05, "loss": 0.4424, "step": 702 }, { "epoch": 3.772712034864231, "grad_norm": 0.11769736086113589, "learning_rate": 1.3661459939036493e-05, "loss": 0.4441, "step": 703 }, { "epoch": 3.778075762655045, "grad_norm": 0.10530984853730485, "learning_rate": 1.354865159131435e-05, "loss": 0.4454, "step": 704 }, { "epoch": 3.78343949044586, "grad_norm": 0.128491542213919, "learning_rate": 1.3436215889310433e-05, "loss": 0.4483, "step": 705 }, { "epoch": 3.7888032182366747, "grad_norm": 0.10663013232583238, "learning_rate": 1.3324154417015205e-05, "loss": 0.4477, "step": 706 }, { "epoch": 3.794166946027489, "grad_norm": 0.11468548568373477, "learning_rate": 1.3212468753146955e-05, "loss": 0.4399, "step": 707 }, { "epoch": 3.7995306738183037, "grad_norm": 0.11277620565805507, "learning_rate": 1.3101160471129588e-05, "loss": 0.4412, "step": 708 }, { "epoch": 3.8048944016091184, "grad_norm": 0.10429631660245424, "learning_rate": 1.2990231139070519e-05, "loss": 0.439, "step": 709 }, { "epoch": 3.8102581293999327, "grad_norm": 0.11117524802508165, "learning_rate": 1.2879682319738467e-05, "loss": 0.448, "step": 710 }, { "epoch": 3.8156218571907474, "grad_norm": 0.10565379203871825, "learning_rate": 1.2769515570541554e-05, "loss": 0.4506, "step": 711 }, { "epoch": 3.820985584981562, "grad_norm": 0.10995579742720774, "learning_rate": 1.2659732443505263e-05, "loss": 0.4378, "step": 712 }, { "epoch": 3.826349312772377, "grad_norm": 0.1094854319106533, "learning_rate": 1.2550334485250661e-05, "loss": 0.4388, "step": 713 }, { "epoch": 3.8317130405631916, "grad_norm": 0.10426452233943717, "learning_rate": 1.2441323236972536e-05, "loss": 0.449, "step": 714 }, { "epoch": 3.837076768354006, "grad_norm": 0.11501738794377854, "learning_rate": 1.2332700234417745e-05, "loss": 0.4502, "step": 715 }, { "epoch": 3.8424404961448206, "grad_norm": 0.10783773069849821, "learning_rate": 1.222446700786355e-05, "loss": 0.443, "step": 716 }, { "epoch": 3.8478042239356354, "grad_norm": 0.1158852559257513, "learning_rate": 1.2116625082096043e-05, "loss": 0.4439, "step": 717 }, { "epoch": 3.8531679517264497, "grad_norm": 0.10640912577878574, "learning_rate": 1.2009175976388683e-05, "loss": 0.4416, "step": 718 }, { "epoch": 3.8585316795172644, "grad_norm": 0.11306196171729263, "learning_rate": 1.1902121204480928e-05, "loss": 0.4475, "step": 719 }, { "epoch": 3.863895407308079, "grad_norm": 0.10374780127454973, "learning_rate": 1.1795462274556835e-05, "loss": 0.4413, "step": 720 }, { "epoch": 3.869259135098894, "grad_norm": 0.10797835729481477, "learning_rate": 1.1689200689223862e-05, "loss": 0.4467, "step": 721 }, { "epoch": 3.8746228628897086, "grad_norm": 0.10466854161183786, "learning_rate": 1.1583337945491717e-05, "loss": 0.4435, "step": 722 }, { "epoch": 3.879986590680523, "grad_norm": 0.1037575420149514, "learning_rate": 1.1477875534751192e-05, "loss": 0.437, "step": 723 }, { "epoch": 3.8853503184713376, "grad_norm": 0.11088366593365082, "learning_rate": 1.1372814942753246e-05, "loss": 0.4431, "step": 724 }, { "epoch": 3.8907140462621523, "grad_norm": 0.09889840234162205, "learning_rate": 1.1268157649588018e-05, "loss": 0.4406, "step": 725 }, { "epoch": 3.8960777740529666, "grad_norm": 0.10215017922131649, "learning_rate": 1.1163905129663956e-05, "loss": 0.4483, "step": 726 }, { "epoch": 3.9014415018437814, "grad_norm": 0.1077840732925271, "learning_rate": 1.1060058851687128e-05, "loss": 0.4417, "step": 727 }, { "epoch": 3.906805229634596, "grad_norm": 0.10317372106874026, "learning_rate": 1.0956620278640427e-05, "loss": 0.4446, "step": 728 }, { "epoch": 3.912168957425411, "grad_norm": 0.1328432948301595, "learning_rate": 1.0853590867763054e-05, "loss": 0.431, "step": 729 }, { "epoch": 3.9175326852162256, "grad_norm": 0.10066539948258632, "learning_rate": 1.0750972070529922e-05, "loss": 0.4474, "step": 730 }, { "epoch": 3.92289641300704, "grad_norm": 0.10308611398222804, "learning_rate": 1.064876533263122e-05, "loss": 0.4466, "step": 731 }, { "epoch": 3.9282601407978546, "grad_norm": 0.10575171336887328, "learning_rate": 1.0546972093952114e-05, "loss": 0.4448, "step": 732 }, { "epoch": 3.9336238685886693, "grad_norm": 0.1090751248770618, "learning_rate": 1.0445593788552344e-05, "loss": 0.4405, "step": 733 }, { "epoch": 3.9389875963794836, "grad_norm": 0.10355561012627491, "learning_rate": 1.0344631844646128e-05, "loss": 0.4432, "step": 734 }, { "epoch": 3.9443513241702983, "grad_norm": 0.10595362039676634, "learning_rate": 1.024408768458196e-05, "loss": 0.4379, "step": 735 }, { "epoch": 3.949715051961113, "grad_norm": 0.11198347644741112, "learning_rate": 1.0143962724822653e-05, "loss": 0.4491, "step": 736 }, { "epoch": 3.9550787797519273, "grad_norm": 0.11560971244461928, "learning_rate": 1.0044258375925295e-05, "loss": 0.4482, "step": 737 }, { "epoch": 3.960442507542742, "grad_norm": 0.10482045247799242, "learning_rate": 9.944976042521465e-06, "loss": 0.4427, "step": 738 }, { "epoch": 3.965806235333557, "grad_norm": 0.10640773573235275, "learning_rate": 9.846117123297353e-06, "loss": 0.4455, "step": 739 }, { "epoch": 3.9711699631243715, "grad_norm": 0.1038041198852965, "learning_rate": 9.747683010974147e-06, "loss": 0.4427, "step": 740 }, { "epoch": 3.9765336909151863, "grad_norm": 0.10117239899529366, "learning_rate": 9.649675092288366e-06, "loss": 0.4413, "step": 741 }, { "epoch": 3.9818974187060006, "grad_norm": 0.10620766516638773, "learning_rate": 9.552094747972297e-06, "loss": 0.4444, "step": 742 }, { "epoch": 3.9872611464968153, "grad_norm": 0.10086629667003526, "learning_rate": 9.454943352734598e-06, "loss": 0.4417, "step": 743 }, { "epoch": 3.99262487428763, "grad_norm": 0.09919272930882302, "learning_rate": 9.358222275240884e-06, "loss": 0.4487, "step": 744 }, { "epoch": 3.9979886020784443, "grad_norm": 0.11767811064838585, "learning_rate": 9.26193287809451e-06, "loss": 0.5044, "step": 745 }, { "epoch": 4.004022795843111, "grad_norm": 0.23793487869261376, "learning_rate": 9.166076517817281e-06, "loss": 0.661, "step": 746 }, { "epoch": 4.009386523633926, "grad_norm": 0.13621220417679167, "learning_rate": 9.07065454483043e-06, "loss": 0.4149, "step": 747 }, { "epoch": 4.01475025142474, "grad_norm": 0.17673227803196773, "learning_rate": 8.975668303435556e-06, "loss": 0.4189, "step": 748 }, { "epoch": 4.020113979215555, "grad_norm": 0.18439349920499457, "learning_rate": 8.881119131795652e-06, "loss": 0.4107, "step": 749 }, { "epoch": 4.025477707006369, "grad_norm": 0.12260746721261573, "learning_rate": 8.787008361916332e-06, "loss": 0.4125, "step": 750 }, { "epoch": 4.030841434797184, "grad_norm": 0.15019363340139308, "learning_rate": 8.693337319626978e-06, "loss": 0.4128, "step": 751 }, { "epoch": 4.036205162587999, "grad_norm": 0.1375901687390116, "learning_rate": 8.60010732456214e-06, "loss": 0.4055, "step": 752 }, { "epoch": 4.041568890378813, "grad_norm": 0.12724391705577723, "learning_rate": 8.507319690142871e-06, "loss": 0.4146, "step": 753 }, { "epoch": 4.0469326181696275, "grad_norm": 0.13895885677671274, "learning_rate": 8.414975723558317e-06, "loss": 0.4142, "step": 754 }, { "epoch": 4.052296345960443, "grad_norm": 0.14044158258051825, "learning_rate": 8.323076725747192e-06, "loss": 0.4106, "step": 755 }, { "epoch": 4.057660073751257, "grad_norm": 0.11701974422585514, "learning_rate": 8.23162399137952e-06, "loss": 0.4192, "step": 756 }, { "epoch": 4.063023801542072, "grad_norm": 0.12971366592995717, "learning_rate": 8.140618808838408e-06, "loss": 0.4031, "step": 757 }, { "epoch": 4.068387529332886, "grad_norm": 0.12697445696160525, "learning_rate": 8.050062460201827e-06, "loss": 0.4131, "step": 758 }, { "epoch": 4.073751257123701, "grad_norm": 0.11378592889355851, "learning_rate": 7.959956221224626e-06, "loss": 0.4111, "step": 759 }, { "epoch": 4.079114984914516, "grad_norm": 0.12117141039045348, "learning_rate": 7.870301361320485e-06, "loss": 0.4165, "step": 760 }, { "epoch": 4.08447871270533, "grad_norm": 0.1197961111627781, "learning_rate": 7.781099143544124e-06, "loss": 0.4129, "step": 761 }, { "epoch": 4.089842440496144, "grad_norm": 0.10807056855124261, "learning_rate": 7.692350824573402e-06, "loss": 0.4134, "step": 762 }, { "epoch": 4.09520616828696, "grad_norm": 0.11324627112259107, "learning_rate": 7.604057654691699e-06, "loss": 0.4079, "step": 763 }, { "epoch": 4.100569896077774, "grad_norm": 0.11724445449675143, "learning_rate": 7.516220877770273e-06, "loss": 0.411, "step": 764 }, { "epoch": 4.105933623868589, "grad_norm": 0.10958796773895332, "learning_rate": 7.428841731250695e-06, "loss": 0.415, "step": 765 }, { "epoch": 4.111297351659403, "grad_norm": 0.11066322738663384, "learning_rate": 7.341921446127509e-06, "loss": 0.4106, "step": 766 }, { "epoch": 4.116661079450218, "grad_norm": 0.11185922785392959, "learning_rate": 7.255461246930791e-06, "loss": 0.4101, "step": 767 }, { "epoch": 4.122024807241033, "grad_norm": 0.11501328533411367, "learning_rate": 7.169462351708958e-06, "loss": 0.4154, "step": 768 }, { "epoch": 4.127388535031847, "grad_norm": 0.10891964674735306, "learning_rate": 7.083925972011583e-06, "loss": 0.4109, "step": 769 }, { "epoch": 4.132752262822661, "grad_norm": 0.1084357061608093, "learning_rate": 6.998853312872347e-06, "loss": 0.4176, "step": 770 }, { "epoch": 4.138115990613477, "grad_norm": 0.11666376677405135, "learning_rate": 6.914245572792064e-06, "loss": 0.4063, "step": 771 }, { "epoch": 4.143479718404291, "grad_norm": 0.10587157920780786, "learning_rate": 6.830103943721749e-06, "loss": 0.4149, "step": 772 }, { "epoch": 4.148843446195105, "grad_norm": 1.5402775013214236, "learning_rate": 6.7464296110458925e-06, "loss": 0.4212, "step": 773 }, { "epoch": 4.15420717398592, "grad_norm": 0.1127675778731525, "learning_rate": 6.6632237535656995e-06, "loss": 0.4123, "step": 774 }, { "epoch": 4.159570901776735, "grad_norm": 0.10636517110929453, "learning_rate": 6.58048754348255e-06, "loss": 0.4031, "step": 775 }, { "epoch": 4.16493462956755, "grad_norm": 0.10408511766608668, "learning_rate": 6.4982221463813965e-06, "loss": 0.4092, "step": 776 }, { "epoch": 4.170298357358364, "grad_norm": 0.10980308429743034, "learning_rate": 6.41642872121444e-06, "loss": 0.4133, "step": 777 }, { "epoch": 4.175662085149178, "grad_norm": 0.1811518845819661, "learning_rate": 6.335108420284748e-06, "loss": 0.4114, "step": 778 }, { "epoch": 4.1810258129399935, "grad_norm": 0.10929364439171839, "learning_rate": 6.254262389230006e-06, "loss": 0.4175, "step": 779 }, { "epoch": 4.186389540730808, "grad_norm": 0.10394422643680504, "learning_rate": 6.1738917670064194e-06, "loss": 0.418, "step": 780 }, { "epoch": 4.191753268521622, "grad_norm": 0.1012577329095609, "learning_rate": 6.09399768587263e-06, "loss": 0.4118, "step": 781 }, { "epoch": 4.197116996312437, "grad_norm": 0.10687807539475037, "learning_rate": 6.014581271373829e-06, "loss": 0.4062, "step": 782 }, { "epoch": 4.202480724103252, "grad_norm": 0.10262903089768421, "learning_rate": 5.935643642325808e-06, "loss": 0.4067, "step": 783 }, { "epoch": 4.207844451894067, "grad_norm": 0.09927029282657747, "learning_rate": 5.857185910799277e-06, "loss": 0.4141, "step": 784 }, { "epoch": 4.213208179684881, "grad_norm": 0.10324140566962033, "learning_rate": 5.779209182104133e-06, "loss": 0.4087, "step": 785 }, { "epoch": 4.218571907475695, "grad_norm": 0.09857973369656504, "learning_rate": 5.701714554773956e-06, "loss": 0.4122, "step": 786 }, { "epoch": 4.2239356352665105, "grad_norm": 0.10704699849654356, "learning_rate": 5.624703120550492e-06, "loss": 0.4132, "step": 787 }, { "epoch": 4.229299363057325, "grad_norm": 0.09932400738261876, "learning_rate": 5.548175964368248e-06, "loss": 0.4136, "step": 788 }, { "epoch": 4.234663090848139, "grad_norm": 0.10182388734864921, "learning_rate": 5.4721341643392845e-06, "loss": 0.4165, "step": 789 }, { "epoch": 4.240026818638954, "grad_norm": 0.10458339193532643, "learning_rate": 5.39657879173793e-06, "loss": 0.4136, "step": 790 }, { "epoch": 4.2453905464297685, "grad_norm": 0.10492190568796772, "learning_rate": 5.3215109109857835e-06, "loss": 0.4156, "step": 791 }, { "epoch": 4.250754274220584, "grad_norm": 0.09833089016292879, "learning_rate": 5.246931579636654e-06, "loss": 0.4213, "step": 792 }, { "epoch": 4.256118002011398, "grad_norm": 0.09631929734391337, "learning_rate": 5.172841848361674e-06, "loss": 0.4126, "step": 793 }, { "epoch": 4.261481729802212, "grad_norm": 0.09916119673198169, "learning_rate": 5.099242760934533e-06, "loss": 0.4093, "step": 794 }, { "epoch": 4.2668454575930275, "grad_norm": 0.09868198578261592, "learning_rate": 5.026135354216717e-06, "loss": 0.4136, "step": 795 }, { "epoch": 4.272209185383842, "grad_norm": 0.09504747159177464, "learning_rate": 4.953520658142958e-06, "loss": 0.4156, "step": 796 }, { "epoch": 4.277572913174656, "grad_norm": 0.09411067382234167, "learning_rate": 4.881399695706677e-06, "loss": 0.4136, "step": 797 }, { "epoch": 4.282936640965471, "grad_norm": 0.10081365538952461, "learning_rate": 4.809773482945601e-06, "loss": 0.4122, "step": 798 }, { "epoch": 4.2883003687562855, "grad_norm": 1.2440830128881666, "learning_rate": 4.738643028927432e-06, "loss": 0.4265, "step": 799 }, { "epoch": 4.293664096547101, "grad_norm": 0.09904601559918061, "learning_rate": 4.668009335735648e-06, "loss": 0.4119, "step": 800 }, { "epoch": 4.299027824337915, "grad_norm": 0.10410871952221983, "learning_rate": 4.5978733984553835e-06, "loss": 0.4116, "step": 801 }, { "epoch": 4.304391552128729, "grad_norm": 0.11936741762617652, "learning_rate": 4.528236205159386e-06, "loss": 0.4163, "step": 802 }, { "epoch": 4.309755279919544, "grad_norm": 0.7752585101606566, "learning_rate": 4.459098736894114e-06, "loss": 0.4168, "step": 803 }, { "epoch": 4.315119007710359, "grad_norm": 0.10147283835455163, "learning_rate": 4.39046196766594e-06, "loss": 0.409, "step": 804 }, { "epoch": 4.320482735501173, "grad_norm": 0.10494531512367294, "learning_rate": 4.322326864427387e-06, "loss": 0.4108, "step": 805 }, { "epoch": 4.325846463291988, "grad_norm": 0.09812759496228979, "learning_rate": 4.254694387063514e-06, "loss": 0.4139, "step": 806 }, { "epoch": 4.3312101910828025, "grad_norm": 0.10039748389385268, "learning_rate": 4.187565488378434e-06, "loss": 0.4099, "step": 807 }, { "epoch": 4.336573918873617, "grad_norm": 0.1000059293045232, "learning_rate": 4.120941114081833e-06, "loss": 0.4112, "step": 808 }, { "epoch": 4.341937646664432, "grad_norm": 0.09526496629638147, "learning_rate": 4.0548222027756835e-06, "loss": 0.4115, "step": 809 }, { "epoch": 4.347301374455246, "grad_norm": 0.09813119886581101, "learning_rate": 3.989209685941027e-06, "loss": 0.412, "step": 810 }, { "epoch": 4.352665102246061, "grad_norm": 0.09732532571840974, "learning_rate": 3.924104487924805e-06, "loss": 0.4061, "step": 811 }, { "epoch": 4.358028830036876, "grad_norm": 0.09543190553673306, "learning_rate": 3.859507525926897e-06, "loss": 0.4141, "step": 812 }, { "epoch": 4.36339255782769, "grad_norm": 0.09873317668913352, "learning_rate": 3.795419709987149e-06, "loss": 0.4153, "step": 813 }, { "epoch": 4.368756285618505, "grad_norm": 0.23096877402734253, "learning_rate": 3.7318419429726025e-06, "loss": 0.4102, "step": 814 }, { "epoch": 4.374120013409319, "grad_norm": 0.10185639946736366, "learning_rate": 3.6687751205647117e-06, "loss": 0.4145, "step": 815 }, { "epoch": 4.379483741200134, "grad_norm": 0.38509111321061773, "learning_rate": 3.606220131246776e-06, "loss": 0.4162, "step": 816 }, { "epoch": 4.384847468990949, "grad_norm": 0.09806919042690798, "learning_rate": 3.5441778562914242e-06, "loss": 0.4141, "step": 817 }, { "epoch": 4.390211196781763, "grad_norm": 0.0969544886038649, "learning_rate": 3.482649169748147e-06, "loss": 0.4212, "step": 818 }, { "epoch": 4.395574924572578, "grad_norm": 0.11496521800565809, "learning_rate": 3.4216349384310533e-06, "loss": 0.4115, "step": 819 }, { "epoch": 4.400938652363393, "grad_norm": 0.09584588428602607, "learning_rate": 3.3611360219065925e-06, "loss": 0.4113, "step": 820 }, { "epoch": 4.406302380154207, "grad_norm": 0.0925654699857564, "learning_rate": 3.3011532724815142e-06, "loss": 0.4102, "step": 821 }, { "epoch": 4.411666107945022, "grad_norm": 0.09392896960421455, "learning_rate": 3.241687535190776e-06, "loss": 0.4105, "step": 822 }, { "epoch": 4.417029835735836, "grad_norm": 0.09291428552953933, "learning_rate": 3.1827396477857264e-06, "loss": 0.4135, "step": 823 }, { "epoch": 4.422393563526651, "grad_norm": 0.09323644383847197, "learning_rate": 3.124310440722247e-06, "loss": 0.4159, "step": 824 }, { "epoch": 4.427757291317466, "grad_norm": 0.23059082156453775, "learning_rate": 3.0664007371490558e-06, "loss": 0.4185, "step": 825 }, { "epoch": 4.43312101910828, "grad_norm": 0.09839755683407715, "learning_rate": 3.009011352896152e-06, "loss": 0.4144, "step": 826 }, { "epoch": 4.4384847468990944, "grad_norm": 0.09990156860630273, "learning_rate": 2.9521430964632602e-06, "loss": 0.4056, "step": 827 }, { "epoch": 4.44384847468991, "grad_norm": 0.093523667605744, "learning_rate": 2.8957967690084986e-06, "loss": 0.4165, "step": 828 }, { "epoch": 4.449212202480724, "grad_norm": 0.09192510131863983, "learning_rate": 2.839973164337044e-06, "loss": 0.4115, "step": 829 }, { "epoch": 4.454575930271539, "grad_norm": 0.09410786987122588, "learning_rate": 2.7846730688900003e-06, "loss": 0.4138, "step": 830 }, { "epoch": 4.459939658062353, "grad_norm": 0.09707065586772506, "learning_rate": 2.729897261733263e-06, "loss": 0.4109, "step": 831 }, { "epoch": 4.465303385853168, "grad_norm": 0.09262040362541676, "learning_rate": 2.675646514546597e-06, "loss": 0.4139, "step": 832 }, { "epoch": 4.470667113643983, "grad_norm": 0.09786358492563979, "learning_rate": 2.6219215916127283e-06, "loss": 0.4112, "step": 833 }, { "epoch": 4.476030841434797, "grad_norm": 0.09168116998441043, "learning_rate": 2.568723249806575e-06, "loss": 0.4106, "step": 834 }, { "epoch": 4.481394569225611, "grad_norm": 0.09262516025294427, "learning_rate": 2.516052238584625e-06, "loss": 0.4149, "step": 835 }, { "epoch": 4.486758297016427, "grad_norm": 0.09205026270318697, "learning_rate": 2.463909299974323e-06, "loss": 0.4062, "step": 836 }, { "epoch": 4.492122024807241, "grad_norm": 0.09122547354614034, "learning_rate": 2.4122951685636674e-06, "loss": 0.4177, "step": 837 }, { "epoch": 4.497485752598056, "grad_norm": 0.09211390919384611, "learning_rate": 2.3612105714908173e-06, "loss": 0.4046, "step": 838 }, { "epoch": 4.50284948038887, "grad_norm": 0.091478554430513, "learning_rate": 2.310656228433894e-06, "loss": 0.412, "step": 839 }, { "epoch": 4.508213208179685, "grad_norm": 0.09399408591730637, "learning_rate": 2.260632851600795e-06, "loss": 0.4149, "step": 840 }, { "epoch": 4.5135769359705, "grad_norm": 0.09195607008303311, "learning_rate": 2.211141145719191e-06, "loss": 0.413, "step": 841 }, { "epoch": 4.518940663761314, "grad_norm": 0.47959804465608913, "learning_rate": 2.1621818080265955e-06, "loss": 0.4182, "step": 842 }, { "epoch": 4.524304391552128, "grad_norm": 0.09051553997088063, "learning_rate": 2.1137555282605325e-06, "loss": 0.4114, "step": 843 }, { "epoch": 4.5296681193429436, "grad_norm": 0.08829292877572836, "learning_rate": 2.0658629886488234e-06, "loss": 0.4165, "step": 844 }, { "epoch": 4.535031847133758, "grad_norm": 0.09080343947058375, "learning_rate": 2.0185048638999706e-06, "loss": 0.4142, "step": 845 }, { "epoch": 4.540395574924572, "grad_norm": 0.09243369068431759, "learning_rate": 1.9716818211936674e-06, "loss": 0.4005, "step": 846 }, { "epoch": 4.545759302715387, "grad_norm": 0.09353671619124167, "learning_rate": 1.925394520171393e-06, "loss": 0.4143, "step": 847 }, { "epoch": 4.551123030506202, "grad_norm": 0.09072830507170011, "learning_rate": 1.8796436129270955e-06, "loss": 0.4116, "step": 848 }, { "epoch": 4.556486758297017, "grad_norm": 0.09201634812096428, "learning_rate": 1.8344297439980475e-06, "loss": 0.4093, "step": 849 }, { "epoch": 4.561850486087831, "grad_norm": 0.20856475534127397, "learning_rate": 1.7897535503557196e-06, "loss": 0.4161, "step": 850 }, { "epoch": 4.567214213878645, "grad_norm": 0.0894737416874815, "learning_rate": 1.7456156613968644e-06, "loss": 0.4128, "step": 851 }, { "epoch": 4.5725779416694605, "grad_norm": 0.09249497786193772, "learning_rate": 1.702016698934581e-06, "loss": 0.4104, "step": 852 }, { "epoch": 4.577941669460275, "grad_norm": 0.10099307480578555, "learning_rate": 1.65895727718961e-06, "loss": 0.4134, "step": 853 }, { "epoch": 4.58330539725109, "grad_norm": 0.0894656321450511, "learning_rate": 1.6164380027816485e-06, "loss": 0.4153, "step": 854 }, { "epoch": 4.588669125041904, "grad_norm": 0.08952970498041048, "learning_rate": 1.5744594747208308e-06, "loss": 0.4118, "step": 855 }, { "epoch": 4.594032852832719, "grad_norm": 0.08843654007502101, "learning_rate": 1.5330222843992658e-06, "loss": 0.4156, "step": 856 }, { "epoch": 4.599396580623534, "grad_norm": 0.18591112023947728, "learning_rate": 1.492127015582714e-06, "loss": 0.4154, "step": 857 }, { "epoch": 4.604760308414348, "grad_norm": 0.0944881909519439, "learning_rate": 1.4517742444023665e-06, "loss": 0.4129, "step": 858 }, { "epoch": 4.610124036205162, "grad_norm": 0.09301181960862621, "learning_rate": 1.4119645393467196e-06, "loss": 0.4193, "step": 859 }, { "epoch": 4.6154877639959775, "grad_norm": 0.08874447140825721, "learning_rate": 1.3726984612535854e-06, "loss": 0.4126, "step": 860 }, { "epoch": 4.620851491786792, "grad_norm": 0.09037539019783776, "learning_rate": 1.3339765633021551e-06, "loss": 0.4164, "step": 861 }, { "epoch": 4.626215219577606, "grad_norm": 0.09159608133309274, "learning_rate": 1.2957993910052503e-06, "loss": 0.4156, "step": 862 }, { "epoch": 4.631578947368421, "grad_norm": 0.0876847986538798, "learning_rate": 1.2581674822016087e-06, "loss": 0.4133, "step": 863 }, { "epoch": 4.6369426751592355, "grad_norm": 0.08889717326718397, "learning_rate": 1.221081367048309e-06, "loss": 0.4173, "step": 864 }, { "epoch": 4.64230640295005, "grad_norm": 0.08911910146230485, "learning_rate": 1.1845415680133089e-06, "loss": 0.4041, "step": 865 }, { "epoch": 4.647670130740865, "grad_norm": 0.0912001297197578, "learning_rate": 1.1485485998680822e-06, "loss": 0.4057, "step": 866 }, { "epoch": 4.653033858531679, "grad_norm": 0.08889958044076257, "learning_rate": 1.1131029696803774e-06, "loss": 0.4078, "step": 867 }, { "epoch": 4.6583975863224945, "grad_norm": 0.09028811157516371, "learning_rate": 1.0782051768070477e-06, "loss": 0.417, "step": 868 }, { "epoch": 4.663761314113309, "grad_norm": 0.09213399459832071, "learning_rate": 1.0438557128870408e-06, "loss": 0.4148, "step": 869 }, { "epoch": 4.669125041904123, "grad_norm": 0.08974457665914247, "learning_rate": 1.010055061834474e-06, "loss": 0.4075, "step": 870 }, { "epoch": 4.674488769694938, "grad_norm": 0.08738862355780086, "learning_rate": 9.768036998317875e-07, "loss": 0.4167, "step": 871 }, { "epoch": 4.6798524974857525, "grad_norm": 0.08744381941505991, "learning_rate": 9.441020953230696e-07, "loss": 0.4044, "step": 872 }, { "epoch": 4.685216225276568, "grad_norm": 0.08768023798398505, "learning_rate": 9.119507090074342e-07, "loss": 0.4094, "step": 873 }, { "epoch": 4.690579953067382, "grad_norm": 0.08529166544500341, "learning_rate": 8.803499938325477e-07, "loss": 0.4129, "step": 874 }, { "epoch": 4.695943680858196, "grad_norm": 0.08637519839116772, "learning_rate": 8.493003949882373e-07, "loss": 0.4133, "step": 875 }, { "epoch": 4.701307408649011, "grad_norm": 0.0887897705514168, "learning_rate": 8.188023499002206e-07, "loss": 0.4051, "step": 876 }, { "epoch": 4.706671136439826, "grad_norm": 0.09055462354291803, "learning_rate": 7.888562882239425e-07, "loss": 0.4161, "step": 877 }, { "epoch": 4.71203486423064, "grad_norm": 0.08675152093159026, "learning_rate": 7.594626318385256e-07, "loss": 0.4039, "step": 878 }, { "epoch": 4.717398592021455, "grad_norm": 0.08841042189110157, "learning_rate": 7.30621794840829e-07, "loss": 0.411, "step": 879 }, { "epoch": 4.7227623198122695, "grad_norm": 0.09048394753620571, "learning_rate": 7.023341835396036e-07, "loss": 0.4158, "step": 880 }, { "epoch": 4.728126047603084, "grad_norm": 0.0937542868591101, "learning_rate": 6.746001964497773e-07, "loss": 0.4108, "step": 881 }, { "epoch": 4.733489775393899, "grad_norm": 0.09183387737661795, "learning_rate": 6.474202242868411e-07, "loss": 0.4132, "step": 882 }, { "epoch": 4.738853503184713, "grad_norm": 0.08673828591623912, "learning_rate": 6.207946499613382e-07, "loss": 0.4138, "step": 883 }, { "epoch": 4.7442172309755275, "grad_norm": 0.1221653497024461, "learning_rate": 5.947238485734819e-07, "loss": 0.411, "step": 884 }, { "epoch": 4.749580958766343, "grad_norm": 0.08800343336015064, "learning_rate": 5.692081874078481e-07, "loss": 0.4044, "step": 885 }, { "epoch": 4.754944686557157, "grad_norm": 0.08826254435249375, "learning_rate": 5.442480259282335e-07, "loss": 0.414, "step": 886 }, { "epoch": 4.760308414347972, "grad_norm": 0.0886779509344544, "learning_rate": 5.198437157725567e-07, "loss": 0.4116, "step": 887 }, { "epoch": 4.765672142138786, "grad_norm": 0.09045462951093008, "learning_rate": 4.959956007479338e-07, "loss": 0.4035, "step": 888 }, { "epoch": 4.771035869929601, "grad_norm": 0.08680316882639165, "learning_rate": 4.7270401682581567e-07, "loss": 0.415, "step": 889 }, { "epoch": 4.776399597720416, "grad_norm": 0.0867270897509586, "learning_rate": 4.499692921372667e-07, "loss": 0.4143, "step": 890 }, { "epoch": 4.78176332551123, "grad_norm": 0.08467163312345347, "learning_rate": 4.277917469683246e-07, "loss": 0.4081, "step": 891 }, { "epoch": 4.787127053302045, "grad_norm": 0.14351442138219858, "learning_rate": 4.061716937555149e-07, "loss": 0.4149, "step": 892 }, { "epoch": 4.79249078109286, "grad_norm": 0.08694159920502774, "learning_rate": 3.851094370814323e-07, "loss": 0.4173, "step": 893 }, { "epoch": 4.797854508883674, "grad_norm": 0.08529615404396881, "learning_rate": 3.646052736704464e-07, "loss": 0.4062, "step": 894 }, { "epoch": 4.803218236674489, "grad_norm": 0.08879884361022528, "learning_rate": 3.4465949238453144e-07, "loss": 0.4197, "step": 895 }, { "epoch": 4.808581964465303, "grad_norm": 0.0862432359092941, "learning_rate": 3.252723742191899e-07, "loss": 0.4139, "step": 896 }, { "epoch": 4.813945692256118, "grad_norm": 0.08670212847739438, "learning_rate": 3.064441922995043e-07, "loss": 0.421, "step": 897 }, { "epoch": 4.819309420046933, "grad_norm": 0.0861260158634291, "learning_rate": 2.8817521187626926e-07, "loss": 0.4114, "step": 898 }, { "epoch": 4.824673147837747, "grad_norm": 0.13041000271914696, "learning_rate": 2.704656903222791e-07, "loss": 0.4089, "step": 899 }, { "epoch": 4.830036875628561, "grad_norm": 0.0853917982287929, "learning_rate": 2.533158771286903e-07, "loss": 0.4189, "step": 900 }, { "epoch": 4.835400603419377, "grad_norm": 0.08585480202942483, "learning_rate": 2.3672601390148707e-07, "loss": 0.4102, "step": 901 }, { "epoch": 4.840764331210191, "grad_norm": 0.08514537079223002, "learning_rate": 2.206963343581281e-07, "loss": 0.4186, "step": 902 }, { "epoch": 4.846128059001006, "grad_norm": 0.0854519613880308, "learning_rate": 2.0522706432419382e-07, "loss": 0.4166, "step": 903 }, { "epoch": 4.85149178679182, "grad_norm": 0.08595768719036412, "learning_rate": 1.903184217302556e-07, "loss": 0.4138, "step": 904 }, { "epoch": 4.856855514582635, "grad_norm": 0.08593142424333411, "learning_rate": 1.7597061660877157e-07, "loss": 0.4187, "step": 905 }, { "epoch": 4.86221924237345, "grad_norm": 0.08605519303696639, "learning_rate": 1.6218385109114665e-07, "loss": 0.4152, "step": 906 }, { "epoch": 4.867582970164264, "grad_norm": 0.08778080722891127, "learning_rate": 1.4895831940486827e-07, "loss": 0.408, "step": 907 }, { "epoch": 4.872946697955078, "grad_norm": 0.5236720260419594, "learning_rate": 1.3629420787079738e-07, "loss": 0.4086, "step": 908 }, { "epoch": 4.878310425745894, "grad_norm": 0.08691961311755944, "learning_rate": 1.2419169490051287e-07, "loss": 0.4162, "step": 909 }, { "epoch": 4.883674153536708, "grad_norm": 0.08604130165924209, "learning_rate": 1.1265095099381118e-07, "loss": 0.4161, "step": 910 }, { "epoch": 4.889037881327523, "grad_norm": 0.08562809530743228, "learning_rate": 1.0167213873631732e-07, "loss": 0.4068, "step": 911 }, { "epoch": 4.894401609118337, "grad_norm": 0.08788772763541514, "learning_rate": 9.125541279717098e-08, "loss": 0.4117, "step": 912 }, { "epoch": 4.899765336909152, "grad_norm": 0.08465883837484985, "learning_rate": 8.14009199268595e-08, "loss": 0.412, "step": 913 }, { "epoch": 4.905129064699967, "grad_norm": 0.08635396133541602, "learning_rate": 7.210879895515277e-08, "loss": 0.4114, "step": 914 }, { "epoch": 4.910492792490781, "grad_norm": 0.08583085092727974, "learning_rate": 6.337918078914041e-08, "loss": 0.4129, "step": 915 }, { "epoch": 4.915856520281595, "grad_norm": 0.08664926249547512, "learning_rate": 5.521218841139764e-08, "loss": 0.4154, "step": 916 }, { "epoch": 4.9212202480724105, "grad_norm": 0.08448787096930517, "learning_rate": 4.7607936878235614e-08, "loss": 0.406, "step": 917 }, { "epoch": 4.926583975863225, "grad_norm": 0.0850836523204834, "learning_rate": 4.0566533318102676e-08, "loss": 0.4122, "step": 918 }, { "epoch": 4.931947703654039, "grad_norm": 0.08844457157599679, "learning_rate": 3.408807693006111e-08, "loss": 0.4186, "step": 919 }, { "epoch": 4.937311431444854, "grad_norm": 0.08507338195494901, "learning_rate": 2.817265898237942e-08, "loss": 0.4044, "step": 920 }, { "epoch": 4.942675159235669, "grad_norm": 0.08409639688790976, "learning_rate": 2.2820362811279973e-08, "loss": 0.4084, "step": 921 }, { "epoch": 4.948038887026484, "grad_norm": 0.08816595457478153, "learning_rate": 1.8031263819726642e-08, "loss": 0.4145, "step": 922 }, { "epoch": 4.953402614817298, "grad_norm": 0.08680564134665435, "learning_rate": 1.3805429476385634e-08, "loss": 0.4068, "step": 923 }, { "epoch": 4.958766342608112, "grad_norm": 0.09052432530510993, "learning_rate": 1.0142919314679588e-08, "loss": 0.4153, "step": 924 }, { "epoch": 4.9641300703989275, "grad_norm": 0.09320878153934302, "learning_rate": 7.043784931921593e-09, "loss": 0.4103, "step": 925 }, { "epoch": 4.969493798189742, "grad_norm": 0.08708424563628227, "learning_rate": 4.508069988617969e-09, "loss": 0.4131, "step": 926 }, { "epoch": 4.974857525980557, "grad_norm": 0.08389684177916036, "learning_rate": 2.5358102078376635e-09, "loss": 0.4079, "step": 927 }, { "epoch": 4.980221253771371, "grad_norm": 0.08673227922167183, "learning_rate": 1.127033374705988e-09, "loss": 0.4102, "step": 928 }, { "epoch": 4.9855849815621855, "grad_norm": 0.08435533244348804, "learning_rate": 2.8175933603158403e-10, "loss": 0.412, "step": 929 }, { "epoch": 4.990948709353001, "grad_norm": 0.0856513074559934, "learning_rate": 0.0, "loss": 0.4095, "step": 930 }, { "epoch": 4.990948709353001, "step": 930, "total_flos": 2.393166916865753e+19, "train_loss": 0.5230806940986265, "train_runtime": 141229.2058, "train_samples_per_second": 3.379, "train_steps_per_second": 0.007 } ], "logging_steps": 1.0, "max_steps": 930, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.393166916865753e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }