{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.999242079733212, "eval_steps": 500, "global_step": 1030, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004850689707442777, "grad_norm": 6.339065858846749, "learning_rate": 7.766990291262136e-07, "loss": 1.0147, "step": 1 }, { "epoch": 0.009701379414885554, "grad_norm": 6.370901874406726, "learning_rate": 1.5533980582524272e-06, "loss": 1.0217, "step": 2 }, { "epoch": 0.01455206912232833, "grad_norm": 6.253076167725343, "learning_rate": 2.330097087378641e-06, "loss": 1.0059, "step": 3 }, { "epoch": 0.01940275882977111, "grad_norm": 5.8627331903052005, "learning_rate": 3.1067961165048544e-06, "loss": 0.9987, "step": 4 }, { "epoch": 0.024253448537213885, "grad_norm": 4.624724616975588, "learning_rate": 3.883495145631068e-06, "loss": 0.9654, "step": 5 }, { "epoch": 0.02910413824465666, "grad_norm": 2.701973440331149, "learning_rate": 4.660194174757282e-06, "loss": 0.9221, "step": 6 }, { "epoch": 0.03395482795209944, "grad_norm": 2.5821400220833683, "learning_rate": 5.436893203883496e-06, "loss": 0.9118, "step": 7 }, { "epoch": 0.03880551765954222, "grad_norm": 3.1473551148693146, "learning_rate": 6.213592233009709e-06, "loss": 0.8818, "step": 8 }, { "epoch": 0.04365620736698499, "grad_norm": 3.8230150062051638, "learning_rate": 6.990291262135923e-06, "loss": 0.8912, "step": 9 }, { "epoch": 0.04850689707442777, "grad_norm": 3.4669236063777715, "learning_rate": 7.766990291262136e-06, "loss": 0.8779, "step": 10 }, { "epoch": 0.053357586781870546, "grad_norm": 2.712055876575345, "learning_rate": 8.54368932038835e-06, "loss": 0.8328, "step": 11 }, { "epoch": 0.05820827648931332, "grad_norm": 2.574263966000136, "learning_rate": 9.320388349514565e-06, "loss": 0.8149, "step": 12 }, { "epoch": 0.0630589661967561, "grad_norm": 1.8016385817876701, "learning_rate": 1.0097087378640778e-05, "loss": 0.7941, "step": 13 }, { "epoch": 0.06790965590419888, "grad_norm": 1.16110696712433, "learning_rate": 1.0873786407766991e-05, "loss": 0.7751, "step": 14 }, { "epoch": 0.07276034561164166, "grad_norm": 1.5441894400920566, "learning_rate": 1.1650485436893204e-05, "loss": 0.7603, "step": 15 }, { "epoch": 0.07761103531908443, "grad_norm": 1.3570174190036193, "learning_rate": 1.2427184466019418e-05, "loss": 0.7456, "step": 16 }, { "epoch": 0.08246172502652721, "grad_norm": 1.0160879152766609, "learning_rate": 1.3203883495145633e-05, "loss": 0.74, "step": 17 }, { "epoch": 0.08731241473396999, "grad_norm": 1.195923679791525, "learning_rate": 1.3980582524271846e-05, "loss": 0.7223, "step": 18 }, { "epoch": 0.09216310444141276, "grad_norm": 1.0381307779091873, "learning_rate": 1.475728155339806e-05, "loss": 0.7149, "step": 19 }, { "epoch": 0.09701379414885554, "grad_norm": 0.9191697728302082, "learning_rate": 1.5533980582524273e-05, "loss": 0.7032, "step": 20 }, { "epoch": 0.10186448385629832, "grad_norm": 1.0389109685950821, "learning_rate": 1.6310679611650486e-05, "loss": 0.697, "step": 21 }, { "epoch": 0.10671517356374109, "grad_norm": 0.8528569833940303, "learning_rate": 1.70873786407767e-05, "loss": 0.6913, "step": 22 }, { "epoch": 0.11156586327118387, "grad_norm": 0.7397574673832126, "learning_rate": 1.7864077669902916e-05, "loss": 0.6844, "step": 23 }, { "epoch": 0.11641655297862664, "grad_norm": 0.6762376097915315, "learning_rate": 1.864077669902913e-05, "loss": 0.6807, "step": 24 }, { "epoch": 0.12126724268606942, "grad_norm": 0.6801312007046909, "learning_rate": 1.9417475728155343e-05, "loss": 0.6651, "step": 25 }, { "epoch": 0.1261179323935122, "grad_norm": 0.5373129321939298, "learning_rate": 2.0194174757281556e-05, "loss": 0.6661, "step": 26 }, { "epoch": 0.13096862210095497, "grad_norm": 0.6155691125010336, "learning_rate": 2.097087378640777e-05, "loss": 0.6655, "step": 27 }, { "epoch": 0.13581931180839776, "grad_norm": 0.5373412410981904, "learning_rate": 2.1747572815533982e-05, "loss": 0.6541, "step": 28 }, { "epoch": 0.14067000151584053, "grad_norm": 0.5722094683121568, "learning_rate": 2.2524271844660196e-05, "loss": 0.6534, "step": 29 }, { "epoch": 0.14552069122328332, "grad_norm": 0.6369873796903149, "learning_rate": 2.330097087378641e-05, "loss": 0.6536, "step": 30 }, { "epoch": 0.15037138093072608, "grad_norm": 0.5246684440675834, "learning_rate": 2.4077669902912622e-05, "loss": 0.6545, "step": 31 }, { "epoch": 0.15522207063816887, "grad_norm": 0.504247506683658, "learning_rate": 2.4854368932038836e-05, "loss": 0.6314, "step": 32 }, { "epoch": 0.16007276034561163, "grad_norm": 0.5566944063536889, "learning_rate": 2.5631067961165052e-05, "loss": 0.6373, "step": 33 }, { "epoch": 0.16492345005305442, "grad_norm": 0.8570989167580252, "learning_rate": 2.6407766990291266e-05, "loss": 0.63, "step": 34 }, { "epoch": 0.16977413976049718, "grad_norm": 1.8667995731915865, "learning_rate": 2.718446601941748e-05, "loss": 0.6456, "step": 35 }, { "epoch": 0.17462482946793997, "grad_norm": 0.5164194413667431, "learning_rate": 2.7961165048543692e-05, "loss": 0.6308, "step": 36 }, { "epoch": 0.17947551917538274, "grad_norm": 1.6098330943831782, "learning_rate": 2.8737864077669905e-05, "loss": 0.6398, "step": 37 }, { "epoch": 0.18432620888282553, "grad_norm": 1.1492129920694993, "learning_rate": 2.951456310679612e-05, "loss": 0.6276, "step": 38 }, { "epoch": 0.18917689859026832, "grad_norm": 0.7127614761088336, "learning_rate": 3.0291262135922332e-05, "loss": 0.6218, "step": 39 }, { "epoch": 0.19402758829771108, "grad_norm": 1.0312378150228299, "learning_rate": 3.1067961165048545e-05, "loss": 0.6253, "step": 40 }, { "epoch": 0.19887827800515387, "grad_norm": 1.7574387168351864, "learning_rate": 3.184466019417476e-05, "loss": 0.6281, "step": 41 }, { "epoch": 0.20372896771259663, "grad_norm": 0.9416599187328968, "learning_rate": 3.262135922330097e-05, "loss": 0.6235, "step": 42 }, { "epoch": 0.20857965742003942, "grad_norm": 1.8301787236679616, "learning_rate": 3.339805825242719e-05, "loss": 0.622, "step": 43 }, { "epoch": 0.21343034712748218, "grad_norm": 0.9292078210446757, "learning_rate": 3.41747572815534e-05, "loss": 0.6107, "step": 44 }, { "epoch": 0.21828103683492497, "grad_norm": 1.994214362456412, "learning_rate": 3.4951456310679615e-05, "loss": 0.6151, "step": 45 }, { "epoch": 0.22313172654236774, "grad_norm": 1.3619344540131681, "learning_rate": 3.572815533980583e-05, "loss": 0.6089, "step": 46 }, { "epoch": 0.22798241624981053, "grad_norm": 1.6323821476629805, "learning_rate": 3.650485436893204e-05, "loss": 0.6165, "step": 47 }, { "epoch": 0.2328331059572533, "grad_norm": 1.1474300438640261, "learning_rate": 3.728155339805826e-05, "loss": 0.6104, "step": 48 }, { "epoch": 0.23768379566469608, "grad_norm": 1.1936059623728144, "learning_rate": 3.805825242718447e-05, "loss": 0.6086, "step": 49 }, { "epoch": 0.24253448537213884, "grad_norm": 1.4126137333521573, "learning_rate": 3.8834951456310685e-05, "loss": 0.6123, "step": 50 }, { "epoch": 0.24738517507958163, "grad_norm": 0.7835607330331523, "learning_rate": 3.9611650485436895e-05, "loss": 0.5979, "step": 51 }, { "epoch": 0.2522358647870244, "grad_norm": 1.3575257469759314, "learning_rate": 4.038834951456311e-05, "loss": 0.6134, "step": 52 }, { "epoch": 0.25708655449446716, "grad_norm": 0.9580505107131282, "learning_rate": 4.116504854368932e-05, "loss": 0.5987, "step": 53 }, { "epoch": 0.26193724420190995, "grad_norm": 1.1647956860260527, "learning_rate": 4.194174757281554e-05, "loss": 0.6134, "step": 54 }, { "epoch": 0.26678793390935274, "grad_norm": 0.9133078407230598, "learning_rate": 4.271844660194175e-05, "loss": 0.5978, "step": 55 }, { "epoch": 0.2716386236167955, "grad_norm": 1.3505836368180404, "learning_rate": 4.3495145631067965e-05, "loss": 0.5944, "step": 56 }, { "epoch": 0.2764893133242383, "grad_norm": 1.1628512424723636, "learning_rate": 4.4271844660194175e-05, "loss": 0.6052, "step": 57 }, { "epoch": 0.28134000303168105, "grad_norm": 1.6421048813051027, "learning_rate": 4.504854368932039e-05, "loss": 0.6008, "step": 58 }, { "epoch": 0.28619069273912384, "grad_norm": 1.3019802198524983, "learning_rate": 4.58252427184466e-05, "loss": 0.594, "step": 59 }, { "epoch": 0.29104138244656663, "grad_norm": 1.2774402772060065, "learning_rate": 4.660194174757282e-05, "loss": 0.5892, "step": 60 }, { "epoch": 0.2958920721540094, "grad_norm": 1.4435670386305743, "learning_rate": 4.737864077669903e-05, "loss": 0.5931, "step": 61 }, { "epoch": 0.30074276186145216, "grad_norm": 1.0284857540916943, "learning_rate": 4.8155339805825245e-05, "loss": 0.5932, "step": 62 }, { "epoch": 0.30559345156889495, "grad_norm": 1.1698103896183938, "learning_rate": 4.8932038834951454e-05, "loss": 0.5948, "step": 63 }, { "epoch": 0.31044414127633774, "grad_norm": 1.434142725222452, "learning_rate": 4.970873786407767e-05, "loss": 0.5897, "step": 64 }, { "epoch": 0.3152948309837805, "grad_norm": 1.3482179068151203, "learning_rate": 5.0485436893203895e-05, "loss": 0.5824, "step": 65 }, { "epoch": 0.32014552069122326, "grad_norm": 1.3069227496554443, "learning_rate": 5.1262135922330105e-05, "loss": 0.5802, "step": 66 }, { "epoch": 0.32499621039866605, "grad_norm": 1.1366236056516827, "learning_rate": 5.203883495145632e-05, "loss": 0.5804, "step": 67 }, { "epoch": 0.32984690010610884, "grad_norm": 1.6243350603336242, "learning_rate": 5.281553398058253e-05, "loss": 0.5812, "step": 68 }, { "epoch": 0.33469758981355163, "grad_norm": 1.0822457196908746, "learning_rate": 5.359223300970875e-05, "loss": 0.5862, "step": 69 }, { "epoch": 0.33954827952099437, "grad_norm": 1.6360957248140573, "learning_rate": 5.436893203883496e-05, "loss": 0.5912, "step": 70 }, { "epoch": 0.34439896922843716, "grad_norm": 2.3477845490488813, "learning_rate": 5.5145631067961174e-05, "loss": 0.5906, "step": 71 }, { "epoch": 0.34924965893587995, "grad_norm": 1.105543560593242, "learning_rate": 5.5922330097087384e-05, "loss": 0.5824, "step": 72 }, { "epoch": 0.35410034864332274, "grad_norm": 4.137098681881185, "learning_rate": 5.66990291262136e-05, "loss": 0.6359, "step": 73 }, { "epoch": 0.3589510383507655, "grad_norm": 3.8847979837997033, "learning_rate": 5.747572815533981e-05, "loss": 0.6486, "step": 74 }, { "epoch": 0.36380172805820826, "grad_norm": 1.1747128429519862, "learning_rate": 5.825242718446603e-05, "loss": 0.595, "step": 75 }, { "epoch": 0.36865241776565105, "grad_norm": 3.008245900701061, "learning_rate": 5.902912621359224e-05, "loss": 0.6387, "step": 76 }, { "epoch": 0.37350310747309384, "grad_norm": 2.3713075132931554, "learning_rate": 5.9805825242718454e-05, "loss": 0.6344, "step": 77 }, { "epoch": 0.37835379718053663, "grad_norm": 1.6213053074921984, "learning_rate": 6.0582524271844664e-05, "loss": 0.6048, "step": 78 }, { "epoch": 0.38320448688797937, "grad_norm": 1.6024154837501339, "learning_rate": 6.135922330097087e-05, "loss": 0.6203, "step": 79 }, { "epoch": 0.38805517659542216, "grad_norm": 1.1825030677591377, "learning_rate": 6.213592233009709e-05, "loss": 0.6052, "step": 80 }, { "epoch": 0.39290586630286495, "grad_norm": 1.3964525731128163, "learning_rate": 6.291262135922331e-05, "loss": 0.6112, "step": 81 }, { "epoch": 0.39775655601030774, "grad_norm": 1.1205074621871551, "learning_rate": 6.368932038834952e-05, "loss": 0.5977, "step": 82 }, { "epoch": 0.4026072457177505, "grad_norm": 1.045620374565707, "learning_rate": 6.446601941747573e-05, "loss": 0.5914, "step": 83 }, { "epoch": 0.40745793542519326, "grad_norm": 1.3974586249408472, "learning_rate": 6.524271844660194e-05, "loss": 0.5918, "step": 84 }, { "epoch": 0.41230862513263605, "grad_norm": 1.0818483302602913, "learning_rate": 6.601941747572816e-05, "loss": 0.5948, "step": 85 }, { "epoch": 0.41715931484007884, "grad_norm": 0.9808456957793906, "learning_rate": 6.679611650485438e-05, "loss": 0.5839, "step": 86 }, { "epoch": 0.4220100045475216, "grad_norm": 1.2035779456517084, "learning_rate": 6.757281553398058e-05, "loss": 0.5833, "step": 87 }, { "epoch": 0.42686069425496437, "grad_norm": 1.6887623926979713, "learning_rate": 6.83495145631068e-05, "loss": 0.5818, "step": 88 }, { "epoch": 0.43171138396240716, "grad_norm": 0.8023218391013366, "learning_rate": 6.912621359223301e-05, "loss": 0.5863, "step": 89 }, { "epoch": 0.43656207366984995, "grad_norm": 1.574900348178855, "learning_rate": 6.990291262135923e-05, "loss": 0.5821, "step": 90 }, { "epoch": 0.4414127633772927, "grad_norm": 0.9288518542917786, "learning_rate": 7.067961165048545e-05, "loss": 0.5814, "step": 91 }, { "epoch": 0.4462634530847355, "grad_norm": 1.6871845487045471, "learning_rate": 7.145631067961166e-05, "loss": 0.5819, "step": 92 }, { "epoch": 0.45111414279217826, "grad_norm": 1.147490028185953, "learning_rate": 7.223300970873787e-05, "loss": 0.5752, "step": 93 }, { "epoch": 0.45596483249962105, "grad_norm": 1.0734179177901382, "learning_rate": 7.300970873786408e-05, "loss": 0.5786, "step": 94 }, { "epoch": 0.46081552220706384, "grad_norm": 1.2263367009960806, "learning_rate": 7.37864077669903e-05, "loss": 0.5789, "step": 95 }, { "epoch": 0.4656662119145066, "grad_norm": 1.4570032389620742, "learning_rate": 7.456310679611652e-05, "loss": 0.5745, "step": 96 }, { "epoch": 0.47051690162194937, "grad_norm": 1.3246870726440927, "learning_rate": 7.533980582524272e-05, "loss": 0.5775, "step": 97 }, { "epoch": 0.47536759132939216, "grad_norm": 0.9415857506868542, "learning_rate": 7.611650485436894e-05, "loss": 0.5699, "step": 98 }, { "epoch": 0.48021828103683495, "grad_norm": 1.2384384474151087, "learning_rate": 7.689320388349515e-05, "loss": 0.5733, "step": 99 }, { "epoch": 0.4850689707442777, "grad_norm": 1.5627749991572353, "learning_rate": 7.766990291262137e-05, "loss": 0.5735, "step": 100 }, { "epoch": 0.4899196604517205, "grad_norm": 1.0078484211944914, "learning_rate": 7.844660194174757e-05, "loss": 0.5733, "step": 101 }, { "epoch": 0.49477035015916326, "grad_norm": 1.6421211712488573, "learning_rate": 7.922330097087379e-05, "loss": 0.576, "step": 102 }, { "epoch": 0.49962103986660605, "grad_norm": 0.8416126904816602, "learning_rate": 8e-05, "loss": 0.5697, "step": 103 }, { "epoch": 0.5044717295740488, "grad_norm": 1.7540458195736903, "learning_rate": 7.999977029531286e-05, "loss": 0.5799, "step": 104 }, { "epoch": 0.5093224192814916, "grad_norm": 1.050447231844734, "learning_rate": 7.999908118388965e-05, "loss": 0.5756, "step": 105 }, { "epoch": 0.5141731089889343, "grad_norm": 1.394330877960354, "learning_rate": 7.999793267364497e-05, "loss": 0.5713, "step": 106 }, { "epoch": 0.5190237986963772, "grad_norm": 1.5873678777006228, "learning_rate": 7.999632477776974e-05, "loss": 0.5733, "step": 107 }, { "epoch": 0.5238744884038199, "grad_norm": 0.8115732123781836, "learning_rate": 7.9994257514731e-05, "loss": 0.5661, "step": 108 }, { "epoch": 0.5287251781112627, "grad_norm": 1.1857754936411384, "learning_rate": 7.999173090827177e-05, "loss": 0.5719, "step": 109 }, { "epoch": 0.5335758678187055, "grad_norm": 0.8066157800150973, "learning_rate": 7.998874498741072e-05, "loss": 0.5695, "step": 110 }, { "epoch": 0.5384265575261482, "grad_norm": 1.4739854945603235, "learning_rate": 7.998529978644183e-05, "loss": 0.5712, "step": 111 }, { "epoch": 0.543277247233591, "grad_norm": 0.778875512416499, "learning_rate": 7.998139534493407e-05, "loss": 0.5609, "step": 112 }, { "epoch": 0.5481279369410338, "grad_norm": 0.9919152170469479, "learning_rate": 7.997703170773084e-05, "loss": 0.5648, "step": 113 }, { "epoch": 0.5529786266484766, "grad_norm": 1.3093959851041357, "learning_rate": 7.997220892494955e-05, "loss": 0.5757, "step": 114 }, { "epoch": 0.5578293163559194, "grad_norm": 0.9066117120369992, "learning_rate": 7.996692705198097e-05, "loss": 0.566, "step": 115 }, { "epoch": 0.5626800060633621, "grad_norm": 1.2498381901490132, "learning_rate": 7.996118614948869e-05, "loss": 0.5757, "step": 116 }, { "epoch": 0.567530695770805, "grad_norm": 1.0385189418340641, "learning_rate": 7.995498628340827e-05, "loss": 0.5697, "step": 117 }, { "epoch": 0.5723813854782477, "grad_norm": 1.3065564110858372, "learning_rate": 7.994832752494667e-05, "loss": 0.5672, "step": 118 }, { "epoch": 0.5772320751856904, "grad_norm": 0.9524061899396458, "learning_rate": 7.994120995058127e-05, "loss": 0.5624, "step": 119 }, { "epoch": 0.5820827648931333, "grad_norm": 1.018979775317657, "learning_rate": 7.993363364205907e-05, "loss": 0.554, "step": 120 }, { "epoch": 0.586933454600576, "grad_norm": 0.9276898951496916, "learning_rate": 7.992559868639576e-05, "loss": 0.556, "step": 121 }, { "epoch": 0.5917841443080188, "grad_norm": 1.1930439188526805, "learning_rate": 7.99171051758747e-05, "loss": 0.5526, "step": 122 }, { "epoch": 0.5966348340154616, "grad_norm": 0.6826863389197481, "learning_rate": 7.990815320804583e-05, "loss": 0.5609, "step": 123 }, { "epoch": 0.6014855237229043, "grad_norm": 0.8061980858751947, "learning_rate": 7.98987428857246e-05, "loss": 0.5586, "step": 124 }, { "epoch": 0.6063362134303472, "grad_norm": 0.9239948149603757, "learning_rate": 7.988887431699079e-05, "loss": 0.5507, "step": 125 }, { "epoch": 0.6111869031377899, "grad_norm": 1.0465957889843347, "learning_rate": 7.987854761518719e-05, "loss": 0.5568, "step": 126 }, { "epoch": 0.6160375928452326, "grad_norm": 0.8308130357408615, "learning_rate": 7.986776289891842e-05, "loss": 0.5591, "step": 127 }, { "epoch": 0.6208882825526755, "grad_norm": 1.0246744766037437, "learning_rate": 7.985652029204946e-05, "loss": 0.5563, "step": 128 }, { "epoch": 0.6257389722601182, "grad_norm": 1.3238612418839921, "learning_rate": 7.984481992370429e-05, "loss": 0.5491, "step": 129 }, { "epoch": 0.630589661967561, "grad_norm": 0.6948258976249133, "learning_rate": 7.983266192826437e-05, "loss": 0.5418, "step": 130 }, { "epoch": 0.6354403516750038, "grad_norm": 0.7288790333090353, "learning_rate": 7.982004644536716e-05, "loss": 0.5441, "step": 131 }, { "epoch": 0.6402910413824465, "grad_norm": 1.0943204811023435, "learning_rate": 7.98069736199044e-05, "loss": 0.5493, "step": 132 }, { "epoch": 0.6451417310898894, "grad_norm": 0.8706827515570799, "learning_rate": 7.979344360202055e-05, "loss": 0.5465, "step": 133 }, { "epoch": 0.6499924207973321, "grad_norm": 0.9127380693761118, "learning_rate": 7.977945654711108e-05, "loss": 0.5475, "step": 134 }, { "epoch": 0.654843110504775, "grad_norm": 0.9510188119086359, "learning_rate": 7.976501261582056e-05, "loss": 0.543, "step": 135 }, { "epoch": 0.6596938002122177, "grad_norm": 0.9350726421156861, "learning_rate": 7.975011197404092e-05, "loss": 0.5525, "step": 136 }, { "epoch": 0.6645444899196604, "grad_norm": 1.272258773056705, "learning_rate": 7.973475479290956e-05, "loss": 0.5518, "step": 137 }, { "epoch": 0.6693951796271033, "grad_norm": 0.9765265325518906, "learning_rate": 7.971894124880727e-05, "loss": 0.5417, "step": 138 }, { "epoch": 0.674245869334546, "grad_norm": 1.1917501609756302, "learning_rate": 7.970267152335632e-05, "loss": 0.5464, "step": 139 }, { "epoch": 0.6790965590419887, "grad_norm": 0.6719257463868904, "learning_rate": 7.968594580341832e-05, "loss": 0.544, "step": 140 }, { "epoch": 0.6839472487494316, "grad_norm": 0.7137208028607956, "learning_rate": 7.966876428109209e-05, "loss": 0.5351, "step": 141 }, { "epoch": 0.6887979384568743, "grad_norm": 0.7449688014358767, "learning_rate": 7.965112715371144e-05, "loss": 0.5397, "step": 142 }, { "epoch": 0.6936486281643172, "grad_norm": 0.5377305872108858, "learning_rate": 7.96330346238429e-05, "loss": 0.5346, "step": 143 }, { "epoch": 0.6984993178717599, "grad_norm": 0.5856757070627496, "learning_rate": 7.961448689928341e-05, "loss": 0.5395, "step": 144 }, { "epoch": 0.7033500075792026, "grad_norm": 0.6000325566823206, "learning_rate": 7.959548419305796e-05, "loss": 0.5447, "step": 145 }, { "epoch": 0.7082006972866455, "grad_norm": 0.7819963257560868, "learning_rate": 7.957602672341707e-05, "loss": 0.5364, "step": 146 }, { "epoch": 0.7130513869940882, "grad_norm": 1.1223449508846108, "learning_rate": 7.955611471383433e-05, "loss": 0.5381, "step": 147 }, { "epoch": 0.717902076701531, "grad_norm": 1.1418981667975974, "learning_rate": 7.953574839300385e-05, "loss": 0.5381, "step": 148 }, { "epoch": 0.7227527664089738, "grad_norm": 0.758286759296052, "learning_rate": 7.95149279948376e-05, "loss": 0.5398, "step": 149 }, { "epoch": 0.7276034561164165, "grad_norm": 0.7637204957772546, "learning_rate": 7.949365375846271e-05, "loss": 0.5386, "step": 150 }, { "epoch": 0.7324541458238594, "grad_norm": 0.6982030938329856, "learning_rate": 7.94719259282188e-05, "loss": 0.5328, "step": 151 }, { "epoch": 0.7373048355313021, "grad_norm": 0.7115887055025976, "learning_rate": 7.944974475365506e-05, "loss": 0.5406, "step": 152 }, { "epoch": 0.7421555252387448, "grad_norm": 0.6914212445412167, "learning_rate": 7.94271104895275e-05, "loss": 0.5375, "step": 153 }, { "epoch": 0.7470062149461877, "grad_norm": 0.6376946136665823, "learning_rate": 7.940402339579596e-05, "loss": 0.5322, "step": 154 }, { "epoch": 0.7518569046536304, "grad_norm": 0.662468788270689, "learning_rate": 7.93804837376211e-05, "loss": 0.5312, "step": 155 }, { "epoch": 0.7567075943610733, "grad_norm": 0.8116591959883654, "learning_rate": 7.935649178536142e-05, "loss": 0.5362, "step": 156 }, { "epoch": 0.761558284068516, "grad_norm": 1.2251954995336705, "learning_rate": 7.93320478145701e-05, "loss": 0.5454, "step": 157 }, { "epoch": 0.7664089737759587, "grad_norm": 1.1753293382340935, "learning_rate": 7.93071521059919e-05, "loss": 0.5369, "step": 158 }, { "epoch": 0.7712596634834016, "grad_norm": 0.5797209510428332, "learning_rate": 7.928180494555983e-05, "loss": 0.5255, "step": 159 }, { "epoch": 0.7761103531908443, "grad_norm": 0.9260629876609666, "learning_rate": 7.925600662439201e-05, "loss": 0.535, "step": 160 }, { "epoch": 0.780961042898287, "grad_norm": 1.2828815170548864, "learning_rate": 7.922975743878817e-05, "loss": 0.5293, "step": 161 }, { "epoch": 0.7858117326057299, "grad_norm": 0.5171701751512903, "learning_rate": 7.92030576902264e-05, "loss": 0.5298, "step": 162 }, { "epoch": 0.7906624223131726, "grad_norm": 0.9531992162184508, "learning_rate": 7.917590768535952e-05, "loss": 0.5315, "step": 163 }, { "epoch": 0.7955131120206155, "grad_norm": 1.1106333536493764, "learning_rate": 7.914830773601173e-05, "loss": 0.5279, "step": 164 }, { "epoch": 0.8003638017280582, "grad_norm": 0.45675301699833715, "learning_rate": 7.912025815917489e-05, "loss": 0.5372, "step": 165 }, { "epoch": 0.805214491435501, "grad_norm": 0.7189143844280501, "learning_rate": 7.909175927700499e-05, "loss": 0.535, "step": 166 }, { "epoch": 0.8100651811429438, "grad_norm": 0.8686823215812186, "learning_rate": 7.906281141681839e-05, "loss": 0.5333, "step": 167 }, { "epoch": 0.8149158708503865, "grad_norm": 0.588440050652225, "learning_rate": 7.903341491108798e-05, "loss": 0.5289, "step": 168 }, { "epoch": 0.8197665605578294, "grad_norm": 0.563873615495661, "learning_rate": 7.900357009743958e-05, "loss": 0.5331, "step": 169 }, { "epoch": 0.8246172502652721, "grad_norm": 0.545022127050129, "learning_rate": 7.897327731864784e-05, "loss": 0.5266, "step": 170 }, { "epoch": 0.8294679399727148, "grad_norm": 0.5592040867673563, "learning_rate": 7.894253692263244e-05, "loss": 0.522, "step": 171 }, { "epoch": 0.8343186296801577, "grad_norm": 0.6324827822327501, "learning_rate": 7.891134926245402e-05, "loss": 0.5297, "step": 172 }, { "epoch": 0.8391693193876004, "grad_norm": 0.8244822385641454, "learning_rate": 7.887971469631016e-05, "loss": 0.5319, "step": 173 }, { "epoch": 0.8440200090950432, "grad_norm": 1.2087031591715138, "learning_rate": 7.884763358753129e-05, "loss": 0.5408, "step": 174 }, { "epoch": 0.848870698802486, "grad_norm": 0.7427926592130248, "learning_rate": 7.881510630457643e-05, "loss": 0.5326, "step": 175 }, { "epoch": 0.8537213885099287, "grad_norm": 0.5074413186000275, "learning_rate": 7.878213322102908e-05, "loss": 0.5281, "step": 176 }, { "epoch": 0.8585720782173716, "grad_norm": 0.7191422226961306, "learning_rate": 7.874871471559282e-05, "loss": 0.5269, "step": 177 }, { "epoch": 0.8634227679248143, "grad_norm": 0.7351841845171684, "learning_rate": 7.8714851172087e-05, "loss": 0.5326, "step": 178 }, { "epoch": 0.868273457632257, "grad_norm": 0.661668481852005, "learning_rate": 7.868054297944237e-05, "loss": 0.5312, "step": 179 }, { "epoch": 0.8731241473396999, "grad_norm": 0.6257843185205204, "learning_rate": 7.864579053169657e-05, "loss": 0.5265, "step": 180 }, { "epoch": 0.8779748370471426, "grad_norm": 0.6272508186215946, "learning_rate": 7.86105942279896e-05, "loss": 0.5242, "step": 181 }, { "epoch": 0.8828255267545854, "grad_norm": 0.5859524441087289, "learning_rate": 7.857495447255925e-05, "loss": 0.5117, "step": 182 }, { "epoch": 0.8876762164620282, "grad_norm": 0.5314342493506371, "learning_rate": 7.853887167473646e-05, "loss": 0.5275, "step": 183 }, { "epoch": 0.892526906169471, "grad_norm": 0.5894396606474899, "learning_rate": 7.850234624894064e-05, "loss": 0.5236, "step": 184 }, { "epoch": 0.8973775958769138, "grad_norm": 0.7789309413273731, "learning_rate": 7.846537861467485e-05, "loss": 0.5269, "step": 185 }, { "epoch": 0.9022282855843565, "grad_norm": 0.9749399305462054, "learning_rate": 7.842796919652104e-05, "loss": 0.5177, "step": 186 }, { "epoch": 0.9070789752917993, "grad_norm": 0.8944031183061603, "learning_rate": 7.839011842413514e-05, "loss": 0.5236, "step": 187 }, { "epoch": 0.9119296649992421, "grad_norm": 0.6072136307213187, "learning_rate": 7.835182673224212e-05, "loss": 0.5237, "step": 188 }, { "epoch": 0.9167803547066848, "grad_norm": 0.44815995922956803, "learning_rate": 7.831309456063107e-05, "loss": 0.5193, "step": 189 }, { "epoch": 0.9216310444141277, "grad_norm": 0.5362363058315522, "learning_rate": 7.827392235415005e-05, "loss": 0.5242, "step": 190 }, { "epoch": 0.9264817341215704, "grad_norm": 0.5675209533538658, "learning_rate": 7.823431056270103e-05, "loss": 0.5223, "step": 191 }, { "epoch": 0.9313324238290132, "grad_norm": 0.5087214285921436, "learning_rate": 7.81942596412347e-05, "loss": 0.522, "step": 192 }, { "epoch": 0.936183113536456, "grad_norm": 0.46687413642954234, "learning_rate": 7.815377004974532e-05, "loss": 0.509, "step": 193 }, { "epoch": 0.9410338032438987, "grad_norm": 0.43127006261715695, "learning_rate": 7.811284225326529e-05, "loss": 0.522, "step": 194 }, { "epoch": 0.9458844929513415, "grad_norm": 0.43502529088952246, "learning_rate": 7.807147672185996e-05, "loss": 0.5258, "step": 195 }, { "epoch": 0.9507351826587843, "grad_norm": 0.5247431038277521, "learning_rate": 7.802967393062219e-05, "loss": 0.524, "step": 196 }, { "epoch": 0.955585872366227, "grad_norm": 0.6873590048364063, "learning_rate": 7.798743435966676e-05, "loss": 0.5227, "step": 197 }, { "epoch": 0.9604365620736699, "grad_norm": 0.8978763261543932, "learning_rate": 7.794475849412512e-05, "loss": 0.5143, "step": 198 }, { "epoch": 0.9652872517811126, "grad_norm": 0.8491329782406589, "learning_rate": 7.790164682413954e-05, "loss": 0.5186, "step": 199 }, { "epoch": 0.9701379414885554, "grad_norm": 0.5004070361840594, "learning_rate": 7.785809984485765e-05, "loss": 0.5185, "step": 200 }, { "epoch": 0.9749886311959982, "grad_norm": 0.4620691718653308, "learning_rate": 7.781411805642675e-05, "loss": 0.5179, "step": 201 }, { "epoch": 0.979839320903441, "grad_norm": 0.6078145800613444, "learning_rate": 7.776970196398795e-05, "loss": 0.5185, "step": 202 }, { "epoch": 0.9846900106108837, "grad_norm": 0.6744449960922595, "learning_rate": 7.77248520776705e-05, "loss": 0.5223, "step": 203 }, { "epoch": 0.9895407003183265, "grad_norm": 0.769858159356013, "learning_rate": 7.767956891258585e-05, "loss": 0.514, "step": 204 }, { "epoch": 0.9943913900257693, "grad_norm": 0.9703823352556248, "learning_rate": 7.763385298882177e-05, "loss": 0.5227, "step": 205 }, { "epoch": 0.9992420797332121, "grad_norm": 1.0790500827883502, "learning_rate": 7.758770483143634e-05, "loss": 0.5161, "step": 206 }, { "epoch": 1.0048506897074427, "grad_norm": 0.979399858326228, "learning_rate": 7.754112497045198e-05, "loss": 0.5126, "step": 207 }, { "epoch": 1.0097013794148855, "grad_norm": 0.8061955074549858, "learning_rate": 7.749411394084931e-05, "loss": 0.5093, "step": 208 }, { "epoch": 1.0145520691223284, "grad_norm": 0.769561002371883, "learning_rate": 7.744667228256102e-05, "loss": 0.5129, "step": 209 }, { "epoch": 1.0194027588297712, "grad_norm": 0.7563794216226459, "learning_rate": 7.739880054046567e-05, "loss": 0.504, "step": 210 }, { "epoch": 1.024253448537214, "grad_norm": 0.4861095042439798, "learning_rate": 7.735049926438143e-05, "loss": 0.5008, "step": 211 }, { "epoch": 1.0291041382446566, "grad_norm": 0.521763707621908, "learning_rate": 7.730176900905978e-05, "loss": 0.5039, "step": 212 }, { "epoch": 1.0339548279520994, "grad_norm": 0.6847654097095645, "learning_rate": 7.725261033417914e-05, "loss": 0.4987, "step": 213 }, { "epoch": 1.0388055176595423, "grad_norm": 0.45617035587869154, "learning_rate": 7.720302380433838e-05, "loss": 0.5082, "step": 214 }, { "epoch": 1.043656207366985, "grad_norm": 0.5136841747695677, "learning_rate": 7.715300998905045e-05, "loss": 0.4903, "step": 215 }, { "epoch": 1.0485068970744278, "grad_norm": 0.5201029165395914, "learning_rate": 7.710256946273572e-05, "loss": 0.5061, "step": 216 }, { "epoch": 1.0533575867818705, "grad_norm": 0.5369451831716502, "learning_rate": 7.705170280471546e-05, "loss": 0.4923, "step": 217 }, { "epoch": 1.0582082764893133, "grad_norm": 0.5292650633659572, "learning_rate": 7.700041059920516e-05, "loss": 0.4958, "step": 218 }, { "epoch": 1.0630589661967562, "grad_norm": 0.38347803199441816, "learning_rate": 7.694869343530781e-05, "loss": 0.4949, "step": 219 }, { "epoch": 1.067909655904199, "grad_norm": 0.42205008813365347, "learning_rate": 7.689655190700719e-05, "loss": 0.4958, "step": 220 }, { "epoch": 1.0727603456116417, "grad_norm": 0.43326896920779806, "learning_rate": 7.684398661316092e-05, "loss": 0.5034, "step": 221 }, { "epoch": 1.0776110353190844, "grad_norm": 0.3610486041319715, "learning_rate": 7.679099815749377e-05, "loss": 0.508, "step": 222 }, { "epoch": 1.0824617250265272, "grad_norm": 0.4631910613487445, "learning_rate": 7.673758714859052e-05, "loss": 0.5012, "step": 223 }, { "epoch": 1.08731241473397, "grad_norm": 0.5933537887890846, "learning_rate": 7.668375419988918e-05, "loss": 0.5003, "step": 224 }, { "epoch": 1.0921631044414128, "grad_norm": 0.7536616463697116, "learning_rate": 7.662949992967375e-05, "loss": 0.4984, "step": 225 }, { "epoch": 1.0970137941488556, "grad_norm": 0.8908783317227219, "learning_rate": 7.657482496106725e-05, "loss": 0.499, "step": 226 }, { "epoch": 1.1018644838562983, "grad_norm": 0.8949365262495667, "learning_rate": 7.651972992202449e-05, "loss": 0.4964, "step": 227 }, { "epoch": 1.106715173563741, "grad_norm": 0.7666080530207662, "learning_rate": 7.646421544532492e-05, "loss": 0.501, "step": 228 }, { "epoch": 1.1115658632711838, "grad_norm": 0.5974548399149405, "learning_rate": 7.640828216856532e-05, "loss": 0.5019, "step": 229 }, { "epoch": 1.1164165529786267, "grad_norm": 0.5208576739553936, "learning_rate": 7.635193073415246e-05, "loss": 0.4954, "step": 230 }, { "epoch": 1.1212672426860695, "grad_norm": 0.4449218126901089, "learning_rate": 7.62951617892958e-05, "loss": 0.4966, "step": 231 }, { "epoch": 1.1261179323935122, "grad_norm": 0.3413546355163314, "learning_rate": 7.623797598599995e-05, "loss": 0.4869, "step": 232 }, { "epoch": 1.130968622100955, "grad_norm": 0.4021494906350838, "learning_rate": 7.618037398105728e-05, "loss": 0.4876, "step": 233 }, { "epoch": 1.1358193118083977, "grad_norm": 0.5401093810240136, "learning_rate": 7.612235643604031e-05, "loss": 0.495, "step": 234 }, { "epoch": 1.1406700015158404, "grad_norm": 0.6450843987922678, "learning_rate": 7.606392401729415e-05, "loss": 0.4953, "step": 235 }, { "epoch": 1.1455206912232834, "grad_norm": 0.6413460171528268, "learning_rate": 7.600507739592879e-05, "loss": 0.4972, "step": 236 }, { "epoch": 1.150371380930726, "grad_norm": 0.6276158556341717, "learning_rate": 7.594581724781152e-05, "loss": 0.4957, "step": 237 }, { "epoch": 1.1552220706381688, "grad_norm": 0.7553018024510589, "learning_rate": 7.588614425355898e-05, "loss": 0.4955, "step": 238 }, { "epoch": 1.1600727603456116, "grad_norm": 0.8968837770832118, "learning_rate": 7.582605909852951e-05, "loss": 0.4937, "step": 239 }, { "epoch": 1.1649234500530543, "grad_norm": 0.7618870978199445, "learning_rate": 7.576556247281522e-05, "loss": 0.4969, "step": 240 }, { "epoch": 1.1697741397604973, "grad_norm": 0.6531717186517569, "learning_rate": 7.570465507123401e-05, "loss": 0.5042, "step": 241 }, { "epoch": 1.17462482946794, "grad_norm": 0.591515239054212, "learning_rate": 7.564333759332167e-05, "loss": 0.4906, "step": 242 }, { "epoch": 1.1794755191753827, "grad_norm": 0.49797581923871925, "learning_rate": 7.558161074332379e-05, "loss": 0.4966, "step": 243 }, { "epoch": 1.1843262088828255, "grad_norm": 0.5865910810734263, "learning_rate": 7.551947523018774e-05, "loss": 0.4997, "step": 244 }, { "epoch": 1.1891768985902682, "grad_norm": 0.5003536808662635, "learning_rate": 7.54569317675544e-05, "loss": 0.4954, "step": 245 }, { "epoch": 1.1940275882977112, "grad_norm": 0.4328856052217075, "learning_rate": 7.539398107375015e-05, "loss": 0.4979, "step": 246 }, { "epoch": 1.198878278005154, "grad_norm": 0.5623661241326378, "learning_rate": 7.533062387177843e-05, "loss": 0.4982, "step": 247 }, { "epoch": 1.2037289677125966, "grad_norm": 0.542587238982675, "learning_rate": 7.526686088931156e-05, "loss": 0.4991, "step": 248 }, { "epoch": 1.2085796574200394, "grad_norm": 0.4949553127282243, "learning_rate": 7.520269285868235e-05, "loss": 0.4908, "step": 249 }, { "epoch": 1.213430347127482, "grad_norm": 0.39917468694971237, "learning_rate": 7.513812051687564e-05, "loss": 0.4917, "step": 250 }, { "epoch": 1.218281036834925, "grad_norm": 0.44831825611716425, "learning_rate": 7.507314460551993e-05, "loss": 0.4898, "step": 251 }, { "epoch": 1.2231317265423678, "grad_norm": 0.4694728642442923, "learning_rate": 7.500776587087878e-05, "loss": 0.4929, "step": 252 }, { "epoch": 1.2279824162498105, "grad_norm": 0.4548469944052975, "learning_rate": 7.494198506384229e-05, "loss": 0.4826, "step": 253 }, { "epoch": 1.2328331059572533, "grad_norm": 0.6666754202357293, "learning_rate": 7.487580293991844e-05, "loss": 0.5021, "step": 254 }, { "epoch": 1.237683795664696, "grad_norm": 0.6791995766586792, "learning_rate": 7.480922025922443e-05, "loss": 0.4974, "step": 255 }, { "epoch": 1.242534485372139, "grad_norm": 0.48795487898431833, "learning_rate": 7.474223778647796e-05, "loss": 0.4934, "step": 256 }, { "epoch": 1.2473851750795817, "grad_norm": 0.32007742081193336, "learning_rate": 7.467485629098842e-05, "loss": 0.4907, "step": 257 }, { "epoch": 1.2522358647870244, "grad_norm": 0.2948779105297471, "learning_rate": 7.460707654664807e-05, "loss": 0.4974, "step": 258 }, { "epoch": 1.2570865544944672, "grad_norm": 0.2954711132741566, "learning_rate": 7.453889933192316e-05, "loss": 0.4893, "step": 259 }, { "epoch": 1.26193724420191, "grad_norm": 0.3234156565448516, "learning_rate": 7.447032542984502e-05, "loss": 0.4882, "step": 260 }, { "epoch": 1.2667879339093528, "grad_norm": 0.37736907562809757, "learning_rate": 7.440135562800093e-05, "loss": 0.487, "step": 261 }, { "epoch": 1.2716386236167956, "grad_norm": 0.4380964102542513, "learning_rate": 7.433199071852526e-05, "loss": 0.4965, "step": 262 }, { "epoch": 1.2764893133242383, "grad_norm": 0.5377640268616285, "learning_rate": 7.426223149809023e-05, "loss": 0.4922, "step": 263 }, { "epoch": 1.281340003031681, "grad_norm": 0.5447192325063156, "learning_rate": 7.419207876789685e-05, "loss": 0.4844, "step": 264 }, { "epoch": 1.2861906927391238, "grad_norm": 0.5569387883590096, "learning_rate": 7.412153333366567e-05, "loss": 0.4887, "step": 265 }, { "epoch": 1.2910413824465667, "grad_norm": 0.6127559785319273, "learning_rate": 7.405059600562751e-05, "loss": 0.4974, "step": 266 }, { "epoch": 1.2958920721540095, "grad_norm": 0.5372861059475705, "learning_rate": 7.397926759851425e-05, "loss": 0.4946, "step": 267 }, { "epoch": 1.3007427618614522, "grad_norm": 0.3715441206816088, "learning_rate": 7.390754893154933e-05, "loss": 0.4914, "step": 268 }, { "epoch": 1.305593451568895, "grad_norm": 0.3468457624853246, "learning_rate": 7.383544082843846e-05, "loss": 0.4906, "step": 269 }, { "epoch": 1.3104441412763377, "grad_norm": 0.37917435116867076, "learning_rate": 7.376294411736009e-05, "loss": 0.4877, "step": 270 }, { "epoch": 1.3152948309837806, "grad_norm": 0.44133896758740837, "learning_rate": 7.369005963095596e-05, "loss": 0.4962, "step": 271 }, { "epoch": 1.3201455206912232, "grad_norm": 0.5453886771483799, "learning_rate": 7.361678820632145e-05, "loss": 0.4918, "step": 272 }, { "epoch": 1.324996210398666, "grad_norm": 0.6036321305780495, "learning_rate": 7.354313068499607e-05, "loss": 0.4892, "step": 273 }, { "epoch": 1.3298469001061088, "grad_norm": 0.620838391242342, "learning_rate": 7.346908791295369e-05, "loss": 0.495, "step": 274 }, { "epoch": 1.3346975898135516, "grad_norm": 0.5848284090531766, "learning_rate": 7.339466074059292e-05, "loss": 0.4862, "step": 275 }, { "epoch": 1.3395482795209943, "grad_norm": 0.5915616847361488, "learning_rate": 7.331985002272726e-05, "loss": 0.4986, "step": 276 }, { "epoch": 1.344398969228437, "grad_norm": 0.5871009859314463, "learning_rate": 7.324465661857534e-05, "loss": 0.4868, "step": 277 }, { "epoch": 1.34924965893588, "grad_norm": 0.4325881770640798, "learning_rate": 7.316908139175105e-05, "loss": 0.4886, "step": 278 }, { "epoch": 1.3541003486433227, "grad_norm": 0.342133338929163, "learning_rate": 7.309312521025356e-05, "loss": 0.4909, "step": 279 }, { "epoch": 1.3589510383507655, "grad_norm": 0.4415991996544984, "learning_rate": 7.301678894645742e-05, "loss": 0.4915, "step": 280 }, { "epoch": 1.3638017280582082, "grad_norm": 0.510599069072842, "learning_rate": 7.294007347710251e-05, "loss": 0.4935, "step": 281 }, { "epoch": 1.368652417765651, "grad_norm": 0.5220646005978747, "learning_rate": 7.286297968328397e-05, "loss": 0.4834, "step": 282 }, { "epoch": 1.373503107473094, "grad_norm": 0.4836929320122624, "learning_rate": 7.27855084504421e-05, "loss": 0.4953, "step": 283 }, { "epoch": 1.3783537971805366, "grad_norm": 0.4678184328008887, "learning_rate": 7.270766066835217e-05, "loss": 0.4872, "step": 284 }, { "epoch": 1.3832044868879794, "grad_norm": 0.4479819209137706, "learning_rate": 7.262943723111419e-05, "loss": 0.4916, "step": 285 }, { "epoch": 1.388055176595422, "grad_norm": 0.44661460126650243, "learning_rate": 7.255083903714266e-05, "loss": 0.4866, "step": 286 }, { "epoch": 1.3929058663028648, "grad_norm": 0.46700839086853646, "learning_rate": 7.247186698915625e-05, "loss": 0.4879, "step": 287 }, { "epoch": 1.3977565560103078, "grad_norm": 0.5111356334340853, "learning_rate": 7.239252199416749e-05, "loss": 0.4812, "step": 288 }, { "epoch": 1.4026072457177505, "grad_norm": 0.5198819509993116, "learning_rate": 7.23128049634722e-05, "loss": 0.4809, "step": 289 }, { "epoch": 1.4074579354251933, "grad_norm": 0.5071119674882796, "learning_rate": 7.223271681263916e-05, "loss": 0.4839, "step": 290 }, { "epoch": 1.412308625132636, "grad_norm": 0.4871736388487796, "learning_rate": 7.215225846149957e-05, "loss": 0.4899, "step": 291 }, { "epoch": 1.4171593148400787, "grad_norm": 0.4644477177967194, "learning_rate": 7.207143083413643e-05, "loss": 0.4865, "step": 292 }, { "epoch": 1.4220100045475217, "grad_norm": 0.5305010693129898, "learning_rate": 7.1990234858874e-05, "loss": 0.4876, "step": 293 }, { "epoch": 1.4268606942549644, "grad_norm": 0.5381812012245146, "learning_rate": 7.190867146826707e-05, "loss": 0.4936, "step": 294 }, { "epoch": 1.4317113839624072, "grad_norm": 0.42926452605809334, "learning_rate": 7.182674159909031e-05, "loss": 0.4845, "step": 295 }, { "epoch": 1.43656207366985, "grad_norm": 0.42959757228135126, "learning_rate": 7.174444619232745e-05, "loss": 0.4952, "step": 296 }, { "epoch": 1.4414127633772926, "grad_norm": 0.4994771109955089, "learning_rate": 7.166178619316056e-05, "loss": 0.4912, "step": 297 }, { "epoch": 1.4462634530847356, "grad_norm": 0.4676921077663137, "learning_rate": 7.157876255095906e-05, "loss": 0.4875, "step": 298 }, { "epoch": 1.4511141427921783, "grad_norm": 0.430209471088095, "learning_rate": 7.149537621926895e-05, "loss": 0.4862, "step": 299 }, { "epoch": 1.455964832499621, "grad_norm": 0.5267214852499816, "learning_rate": 7.14116281558018e-05, "loss": 0.4879, "step": 300 }, { "epoch": 1.4608155222070638, "grad_norm": 0.5169308603441447, "learning_rate": 7.132751932242376e-05, "loss": 0.4984, "step": 301 }, { "epoch": 1.4656662119145065, "grad_norm": 0.3435076728807633, "learning_rate": 7.124305068514444e-05, "loss": 0.487, "step": 302 }, { "epoch": 1.4705169016219495, "grad_norm": 0.2662958717194974, "learning_rate": 7.1158223214106e-05, "loss": 0.4878, "step": 303 }, { "epoch": 1.4753675913293922, "grad_norm": 0.31770648802942325, "learning_rate": 7.107303788357177e-05, "loss": 0.4819, "step": 304 }, { "epoch": 1.480218281036835, "grad_norm": 0.29599579651368213, "learning_rate": 7.098749567191527e-05, "loss": 0.4852, "step": 305 }, { "epoch": 1.4850689707442777, "grad_norm": 0.3093179833213525, "learning_rate": 7.090159756160886e-05, "loss": 0.4877, "step": 306 }, { "epoch": 1.4899196604517204, "grad_norm": 0.38971280696993216, "learning_rate": 7.081534453921242e-05, "loss": 0.4852, "step": 307 }, { "epoch": 1.4947703501591634, "grad_norm": 0.3825128078756503, "learning_rate": 7.072873759536217e-05, "loss": 0.4913, "step": 308 }, { "epoch": 1.499621039866606, "grad_norm": 0.2886624346515238, "learning_rate": 7.064177772475912e-05, "loss": 0.4798, "step": 309 }, { "epoch": 1.5044717295740488, "grad_norm": 0.3712678777863381, "learning_rate": 7.05544659261578e-05, "loss": 0.4867, "step": 310 }, { "epoch": 1.5093224192814916, "grad_norm": 0.42331161544955054, "learning_rate": 7.046680320235466e-05, "loss": 0.4871, "step": 311 }, { "epoch": 1.5141731089889343, "grad_norm": 0.4018226592697959, "learning_rate": 7.037879056017663e-05, "loss": 0.4842, "step": 312 }, { "epoch": 1.5190237986963773, "grad_norm": 0.3959692369182176, "learning_rate": 7.029042901046952e-05, "loss": 0.4802, "step": 313 }, { "epoch": 1.5238744884038198, "grad_norm": 0.4052319536586186, "learning_rate": 7.020171956808645e-05, "loss": 0.4859, "step": 314 }, { "epoch": 1.5287251781112627, "grad_norm": 0.40673400305291324, "learning_rate": 7.011266325187615e-05, "loss": 0.496, "step": 315 }, { "epoch": 1.5335758678187055, "grad_norm": 0.44258718402573904, "learning_rate": 7.002326108467129e-05, "loss": 0.4864, "step": 316 }, { "epoch": 1.5384265575261482, "grad_norm": 0.4022422201093314, "learning_rate": 6.993351409327672e-05, "loss": 0.4763, "step": 317 }, { "epoch": 1.5432772472335912, "grad_norm": 0.39107692009497685, "learning_rate": 6.984342330845764e-05, "loss": 0.4952, "step": 318 }, { "epoch": 1.5481279369410337, "grad_norm": 0.36401667233363455, "learning_rate": 6.975298976492785e-05, "loss": 0.4952, "step": 319 }, { "epoch": 1.5529786266484766, "grad_norm": 0.3964209144895704, "learning_rate": 6.966221450133779e-05, "loss": 0.4901, "step": 320 }, { "epoch": 1.5578293163559194, "grad_norm": 0.41002635948711413, "learning_rate": 6.957109856026261e-05, "loss": 0.4917, "step": 321 }, { "epoch": 1.562680006063362, "grad_norm": 0.33043042887382146, "learning_rate": 6.94796429881903e-05, "loss": 0.4771, "step": 322 }, { "epoch": 1.567530695770805, "grad_norm": 0.3258167342621945, "learning_rate": 6.938784883550948e-05, "loss": 0.4889, "step": 323 }, { "epoch": 1.5723813854782476, "grad_norm": 0.33864970777234993, "learning_rate": 6.929571715649755e-05, "loss": 0.4866, "step": 324 }, { "epoch": 1.5772320751856905, "grad_norm": 0.3799335436470155, "learning_rate": 6.920324900930842e-05, "loss": 0.4907, "step": 325 }, { "epoch": 1.5820827648931333, "grad_norm": 0.38649667346458383, "learning_rate": 6.911044545596042e-05, "loss": 0.4854, "step": 326 }, { "epoch": 1.586933454600576, "grad_norm": 0.3332457174533195, "learning_rate": 6.901730756232411e-05, "loss": 0.4895, "step": 327 }, { "epoch": 1.591784144308019, "grad_norm": 0.3688581464993942, "learning_rate": 6.892383639811005e-05, "loss": 0.4958, "step": 328 }, { "epoch": 1.5966348340154615, "grad_norm": 0.43127701691860393, "learning_rate": 6.883003303685644e-05, "loss": 0.4844, "step": 329 }, { "epoch": 1.6014855237229044, "grad_norm": 0.5098788849460419, "learning_rate": 6.87358985559169e-05, "loss": 0.489, "step": 330 }, { "epoch": 1.6063362134303472, "grad_norm": 0.531466303384909, "learning_rate": 6.864143403644797e-05, "loss": 0.4945, "step": 331 }, { "epoch": 1.61118690313779, "grad_norm": 0.4575057116450561, "learning_rate": 6.85466405633968e-05, "loss": 0.4855, "step": 332 }, { "epoch": 1.6160375928452326, "grad_norm": 0.42418027914564915, "learning_rate": 6.845151922548865e-05, "loss": 0.4783, "step": 333 }, { "epoch": 1.6208882825526754, "grad_norm": 0.38431703300530295, "learning_rate": 6.835607111521439e-05, "loss": 0.4796, "step": 334 }, { "epoch": 1.6257389722601183, "grad_norm": 0.325044588754403, "learning_rate": 6.826029732881793e-05, "loss": 0.4928, "step": 335 }, { "epoch": 1.630589661967561, "grad_norm": 0.2723847494193817, "learning_rate": 6.816419896628363e-05, "loss": 0.4851, "step": 336 }, { "epoch": 1.6354403516750038, "grad_norm": 0.25325170865464947, "learning_rate": 6.806777713132374e-05, "loss": 0.4826, "step": 337 }, { "epoch": 1.6402910413824465, "grad_norm": 0.2793387163645126, "learning_rate": 6.79710329313656e-05, "loss": 0.4873, "step": 338 }, { "epoch": 1.6451417310898893, "grad_norm": 0.38835730960557174, "learning_rate": 6.787396747753903e-05, "loss": 0.4744, "step": 339 }, { "epoch": 1.6499924207973322, "grad_norm": 0.5896470389457479, "learning_rate": 6.777658188466354e-05, "loss": 0.4765, "step": 340 }, { "epoch": 1.654843110504775, "grad_norm": 0.7717199781637745, "learning_rate": 6.767887727123544e-05, "loss": 0.4931, "step": 341 }, { "epoch": 1.6596938002122177, "grad_norm": 1.0491183054565791, "learning_rate": 6.758085475941516e-05, "loss": 0.4875, "step": 342 }, { "epoch": 1.6645444899196604, "grad_norm": 1.0963514331569248, "learning_rate": 6.748251547501418e-05, "loss": 0.4783, "step": 343 }, { "epoch": 1.6693951796271032, "grad_norm": 0.6008540810802777, "learning_rate": 6.738386054748226e-05, "loss": 0.4836, "step": 344 }, { "epoch": 1.674245869334546, "grad_norm": 0.4448291346545642, "learning_rate": 6.728489110989434e-05, "loss": 0.4883, "step": 345 }, { "epoch": 1.6790965590419886, "grad_norm": 0.4942617693141126, "learning_rate": 6.718560829893762e-05, "loss": 0.4799, "step": 346 }, { "epoch": 1.6839472487494316, "grad_norm": 0.4996733158915889, "learning_rate": 6.708601325489844e-05, "loss": 0.4872, "step": 347 }, { "epoch": 1.6887979384568743, "grad_norm": 0.3715020675799419, "learning_rate": 6.698610712164924e-05, "loss": 0.4864, "step": 348 }, { "epoch": 1.693648628164317, "grad_norm": 0.27977348496462506, "learning_rate": 6.688589104663536e-05, "loss": 0.4731, "step": 349 }, { "epoch": 1.69849931787176, "grad_norm": 0.3329757061053683, "learning_rate": 6.67853661808619e-05, "loss": 0.4771, "step": 350 }, { "epoch": 1.7033500075792025, "grad_norm": 0.35849519050184514, "learning_rate": 6.668453367888052e-05, "loss": 0.4867, "step": 351 }, { "epoch": 1.7082006972866455, "grad_norm": 0.3093064728603477, "learning_rate": 6.658339469877613e-05, "loss": 0.478, "step": 352 }, { "epoch": 1.7130513869940882, "grad_norm": 0.2983932071893944, "learning_rate": 6.64819504021536e-05, "loss": 0.4814, "step": 353 }, { "epoch": 1.717902076701531, "grad_norm": 0.32895472335219694, "learning_rate": 6.638020195412448e-05, "loss": 0.4771, "step": 354 }, { "epoch": 1.722752766408974, "grad_norm": 0.4277599821762847, "learning_rate": 6.627815052329354e-05, "loss": 0.4925, "step": 355 }, { "epoch": 1.7276034561164164, "grad_norm": 0.4690652488456737, "learning_rate": 6.617579728174535e-05, "loss": 0.4854, "step": 356 }, { "epoch": 1.7324541458238594, "grad_norm": 0.43487992599318925, "learning_rate": 6.60731434050309e-05, "loss": 0.4777, "step": 357 }, { "epoch": 1.737304835531302, "grad_norm": 0.4398837483674929, "learning_rate": 6.597019007215401e-05, "loss": 0.4783, "step": 358 }, { "epoch": 1.7421555252387448, "grad_norm": 0.43050412759021583, "learning_rate": 6.586693846555788e-05, "loss": 0.4743, "step": 359 }, { "epoch": 1.7470062149461878, "grad_norm": 0.40202174280112624, "learning_rate": 6.576338977111134e-05, "loss": 0.48, "step": 360 }, { "epoch": 1.7518569046536303, "grad_norm": 0.36259944082265505, "learning_rate": 6.565954517809543e-05, "loss": 0.4747, "step": 361 }, { "epoch": 1.7567075943610733, "grad_norm": 0.28300782787905476, "learning_rate": 6.555540587918968e-05, "loss": 0.4778, "step": 362 }, { "epoch": 1.761558284068516, "grad_norm": 0.27286805994349533, "learning_rate": 6.545097307045831e-05, "loss": 0.4795, "step": 363 }, { "epoch": 1.7664089737759587, "grad_norm": 0.31934491384929364, "learning_rate": 6.534624795133662e-05, "loss": 0.4851, "step": 364 }, { "epoch": 1.7712596634834017, "grad_norm": 0.35245692182661065, "learning_rate": 6.524123172461711e-05, "loss": 0.4794, "step": 365 }, { "epoch": 1.7761103531908442, "grad_norm": 0.3396944161317505, "learning_rate": 6.51359255964358e-05, "loss": 0.4774, "step": 366 }, { "epoch": 1.7809610428982872, "grad_norm": 0.2877754553279699, "learning_rate": 6.503033077625824e-05, "loss": 0.4746, "step": 367 }, { "epoch": 1.78581173260573, "grad_norm": 0.3326775146217252, "learning_rate": 6.492444847686566e-05, "loss": 0.4849, "step": 368 }, { "epoch": 1.7906624223131726, "grad_norm": 0.419702080624426, "learning_rate": 6.481827991434111e-05, "loss": 0.4814, "step": 369 }, { "epoch": 1.7955131120206156, "grad_norm": 0.43780909545325103, "learning_rate": 6.471182630805538e-05, "loss": 0.4813, "step": 370 }, { "epoch": 1.800363801728058, "grad_norm": 0.5080371772111751, "learning_rate": 6.460508888065314e-05, "loss": 0.4865, "step": 371 }, { "epoch": 1.805214491435501, "grad_norm": 0.5541771901573003, "learning_rate": 6.449806885803873e-05, "loss": 0.4752, "step": 372 }, { "epoch": 1.8100651811429438, "grad_norm": 0.5077930985424878, "learning_rate": 6.439076746936219e-05, "loss": 0.4776, "step": 373 }, { "epoch": 1.8149158708503865, "grad_norm": 0.3974055597915673, "learning_rate": 6.428318594700509e-05, "loss": 0.4833, "step": 374 }, { "epoch": 1.8197665605578295, "grad_norm": 0.27838793632255576, "learning_rate": 6.417532552656647e-05, "loss": 0.4808, "step": 375 }, { "epoch": 1.824617250265272, "grad_norm": 0.25531453372215107, "learning_rate": 6.406718744684851e-05, "loss": 0.475, "step": 376 }, { "epoch": 1.829467939972715, "grad_norm": 0.37619306197372576, "learning_rate": 6.395877294984241e-05, "loss": 0.4718, "step": 377 }, { "epoch": 1.8343186296801577, "grad_norm": 0.45314622671289495, "learning_rate": 6.385008328071406e-05, "loss": 0.4858, "step": 378 }, { "epoch": 1.8391693193876004, "grad_norm": 0.4091697946782105, "learning_rate": 6.374111968778982e-05, "loss": 0.4797, "step": 379 }, { "epoch": 1.8440200090950432, "grad_norm": 0.31229108750141443, "learning_rate": 6.363188342254206e-05, "loss": 0.4819, "step": 380 }, { "epoch": 1.848870698802486, "grad_norm": 0.29389799920114323, "learning_rate": 6.352237573957488e-05, "loss": 0.4777, "step": 381 }, { "epoch": 1.8537213885099288, "grad_norm": 0.31830405845533455, "learning_rate": 6.341259789660969e-05, "loss": 0.4812, "step": 382 }, { "epoch": 1.8585720782173716, "grad_norm": 0.2783419658252336, "learning_rate": 6.330255115447076e-05, "loss": 0.4722, "step": 383 }, { "epoch": 1.8634227679248143, "grad_norm": 0.252195902565345, "learning_rate": 6.319223677707069e-05, "loss": 0.4786, "step": 384 }, { "epoch": 1.868273457632257, "grad_norm": 0.24931429544692238, "learning_rate": 6.308165603139598e-05, "loss": 0.4766, "step": 385 }, { "epoch": 1.8731241473396998, "grad_norm": 0.343056469344131, "learning_rate": 6.29708101874924e-05, "loss": 0.4746, "step": 386 }, { "epoch": 1.8779748370471427, "grad_norm": 0.39163026769861153, "learning_rate": 6.285970051845045e-05, "loss": 0.4777, "step": 387 }, { "epoch": 1.8828255267545853, "grad_norm": 0.3925083138992103, "learning_rate": 6.274832830039071e-05, "loss": 0.4762, "step": 388 }, { "epoch": 1.8876762164620282, "grad_norm": 0.3908015146094751, "learning_rate": 6.26366948124492e-05, "loss": 0.4882, "step": 389 }, { "epoch": 1.892526906169471, "grad_norm": 0.38768255703534454, "learning_rate": 6.25248013367627e-05, "loss": 0.4746, "step": 390 }, { "epoch": 1.8973775958769137, "grad_norm": 0.34797592572865116, "learning_rate": 6.241264915845401e-05, "loss": 0.4863, "step": 391 }, { "epoch": 1.9022282855843566, "grad_norm": 0.2939167601152598, "learning_rate": 6.230023956561716e-05, "loss": 0.4803, "step": 392 }, { "epoch": 1.9070789752917991, "grad_norm": 0.2683598696324213, "learning_rate": 6.218757384930268e-05, "loss": 0.4769, "step": 393 }, { "epoch": 1.911929664999242, "grad_norm": 0.2963955328357841, "learning_rate": 6.207465330350273e-05, "loss": 0.4798, "step": 394 }, { "epoch": 1.9167803547066848, "grad_norm": 0.34936564317737695, "learning_rate": 6.196147922513623e-05, "loss": 0.479, "step": 395 }, { "epoch": 1.9216310444141276, "grad_norm": 0.3023947037005716, "learning_rate": 6.184805291403402e-05, "loss": 0.4776, "step": 396 }, { "epoch": 1.9264817341215705, "grad_norm": 0.23472105439779495, "learning_rate": 6.173437567292383e-05, "loss": 0.475, "step": 397 }, { "epoch": 1.931332423829013, "grad_norm": 0.30151702935680424, "learning_rate": 6.162044880741544e-05, "loss": 0.4719, "step": 398 }, { "epoch": 1.936183113536456, "grad_norm": 0.4380813843564537, "learning_rate": 6.150627362598557e-05, "loss": 0.4871, "step": 399 }, { "epoch": 1.9410338032438987, "grad_norm": 0.5034940901090467, "learning_rate": 6.139185143996298e-05, "loss": 0.4806, "step": 400 }, { "epoch": 1.9458844929513415, "grad_norm": 0.44831770290541656, "learning_rate": 6.127718356351326e-05, "loss": 0.478, "step": 401 }, { "epoch": 1.9507351826587844, "grad_norm": 0.3881020850233725, "learning_rate": 6.116227131362385e-05, "loss": 0.4714, "step": 402 }, { "epoch": 1.955585872366227, "grad_norm": 0.3382903208345561, "learning_rate": 6.104711601008888e-05, "loss": 0.4779, "step": 403 }, { "epoch": 1.96043656207367, "grad_norm": 0.2830781275191087, "learning_rate": 6.0931718975493985e-05, "loss": 0.4846, "step": 404 }, { "epoch": 1.9652872517811126, "grad_norm": 0.27767803820547865, "learning_rate": 6.081608153520117e-05, "loss": 0.4691, "step": 405 }, { "epoch": 1.9701379414885554, "grad_norm": 0.35527234014372044, "learning_rate": 6.0700205017333525e-05, "loss": 0.4787, "step": 406 }, { "epoch": 1.9749886311959983, "grad_norm": 0.34300192919407774, "learning_rate": 6.058409075276002e-05, "loss": 0.4689, "step": 407 }, { "epoch": 1.9798393209034408, "grad_norm": 0.2964726842271146, "learning_rate": 6.046774007508019e-05, "loss": 0.475, "step": 408 }, { "epoch": 1.9846900106108838, "grad_norm": 0.2502113324820329, "learning_rate": 6.035115432060883e-05, "loss": 0.4747, "step": 409 }, { "epoch": 1.9895407003183265, "grad_norm": 0.23471156048166733, "learning_rate": 6.0234334828360655e-05, "loss": 0.4786, "step": 410 }, { "epoch": 1.9943913900257693, "grad_norm": 0.27137410563019304, "learning_rate": 6.011728294003494e-05, "loss": 0.4802, "step": 411 }, { "epoch": 1.9992420797332122, "grad_norm": 0.2960970008273601, "learning_rate": 6.000000000000001e-05, "loss": 0.4768, "step": 412 }, { "epoch": 2.004850689707443, "grad_norm": 0.33644070901215145, "learning_rate": 5.988248735527793e-05, "loss": 0.4473, "step": 413 }, { "epoch": 2.0097013794148855, "grad_norm": 0.39546749871672404, "learning_rate": 5.9764746355528994e-05, "loss": 0.4501, "step": 414 }, { "epoch": 2.0145520691223284, "grad_norm": 0.47967552460987467, "learning_rate": 5.964677835303615e-05, "loss": 0.4483, "step": 415 }, { "epoch": 2.019402758829771, "grad_norm": 0.43081443925289625, "learning_rate": 5.952858470268955e-05, "loss": 0.4468, "step": 416 }, { "epoch": 2.024253448537214, "grad_norm": 0.42692273173821377, "learning_rate": 5.941016676197098e-05, "loss": 0.4499, "step": 417 }, { "epoch": 2.029104138244657, "grad_norm": 0.5203095933335755, "learning_rate": 5.929152589093825e-05, "loss": 0.4498, "step": 418 }, { "epoch": 2.0339548279520994, "grad_norm": 0.5390774215298352, "learning_rate": 5.9172663452209554e-05, "loss": 0.449, "step": 419 }, { "epoch": 2.0388055176595423, "grad_norm": 0.5618164866931807, "learning_rate": 5.9053580810947845e-05, "loss": 0.4555, "step": 420 }, { "epoch": 2.043656207366985, "grad_norm": 0.6473211411897097, "learning_rate": 5.89342793348452e-05, "loss": 0.4537, "step": 421 }, { "epoch": 2.048506897074428, "grad_norm": 0.7641174380933736, "learning_rate": 5.881476039410699e-05, "loss": 0.4579, "step": 422 }, { "epoch": 2.0533575867818707, "grad_norm": 0.6589276143277975, "learning_rate": 5.869502536143629e-05, "loss": 0.4478, "step": 423 }, { "epoch": 2.0582082764893133, "grad_norm": 0.4055360078631115, "learning_rate": 5.857507561201802e-05, "loss": 0.4501, "step": 424 }, { "epoch": 2.063058966196756, "grad_norm": 0.3576976481492298, "learning_rate": 5.845491252350312e-05, "loss": 0.4479, "step": 425 }, { "epoch": 2.0679096559041987, "grad_norm": 0.444390387775824, "learning_rate": 5.833453747599286e-05, "loss": 0.4466, "step": 426 }, { "epoch": 2.0727603456116417, "grad_norm": 0.38706062123939283, "learning_rate": 5.821395185202285e-05, "loss": 0.449, "step": 427 }, { "epoch": 2.0776110353190846, "grad_norm": 0.3243650769978332, "learning_rate": 5.809315703654726e-05, "loss": 0.4581, "step": 428 }, { "epoch": 2.082461725026527, "grad_norm": 0.35862489527477903, "learning_rate": 5.797215441692284e-05, "loss": 0.4534, "step": 429 }, { "epoch": 2.08731241473397, "grad_norm": 0.3353344776943914, "learning_rate": 5.785094538289304e-05, "loss": 0.4537, "step": 430 }, { "epoch": 2.0921631044414126, "grad_norm": 0.3620686259692887, "learning_rate": 5.772953132657202e-05, "loss": 0.4553, "step": 431 }, { "epoch": 2.0970137941488556, "grad_norm": 0.307403523851727, "learning_rate": 5.7607913642428666e-05, "loss": 0.4424, "step": 432 }, { "epoch": 2.101864483856298, "grad_norm": 0.28710201149568576, "learning_rate": 5.7486093727270606e-05, "loss": 0.4462, "step": 433 }, { "epoch": 2.106715173563741, "grad_norm": 0.31451691469735704, "learning_rate": 5.736407298022809e-05, "loss": 0.4434, "step": 434 }, { "epoch": 2.111565863271184, "grad_norm": 0.3105010146819863, "learning_rate": 5.7241852802738e-05, "loss": 0.4533, "step": 435 }, { "epoch": 2.1164165529786265, "grad_norm": 0.2878408032383936, "learning_rate": 5.711943459852772e-05, "loss": 0.4427, "step": 436 }, { "epoch": 2.1212672426860695, "grad_norm": 0.3026608573456508, "learning_rate": 5.699681977359902e-05, "loss": 0.4385, "step": 437 }, { "epoch": 2.1261179323935124, "grad_norm": 0.2857208984253648, "learning_rate": 5.6874009736211896e-05, "loss": 0.4465, "step": 438 }, { "epoch": 2.130968622100955, "grad_norm": 0.28786201316205207, "learning_rate": 5.675100589686839e-05, "loss": 0.4472, "step": 439 }, { "epoch": 2.135819311808398, "grad_norm": 0.3397455432854385, "learning_rate": 5.662780966829646e-05, "loss": 0.4486, "step": 440 }, { "epoch": 2.1406700015158404, "grad_norm": 0.3060713829784068, "learning_rate": 5.650442246543364e-05, "loss": 0.4525, "step": 441 }, { "epoch": 2.1455206912232834, "grad_norm": 0.2554715156912105, "learning_rate": 5.638084570541088e-05, "loss": 0.4451, "step": 442 }, { "epoch": 2.150371380930726, "grad_norm": 0.31763468046624377, "learning_rate": 5.625708080753621e-05, "loss": 0.455, "step": 443 }, { "epoch": 2.155222070638169, "grad_norm": 0.3288505103534938, "learning_rate": 5.6133129193278525e-05, "loss": 0.4453, "step": 444 }, { "epoch": 2.160072760345612, "grad_norm": 0.27302062189682574, "learning_rate": 5.600899228625112e-05, "loss": 0.4523, "step": 445 }, { "epoch": 2.1649234500530543, "grad_norm": 0.2367917713116079, "learning_rate": 5.588467151219549e-05, "loss": 0.4481, "step": 446 }, { "epoch": 2.1697741397604973, "grad_norm": 0.27892881990044693, "learning_rate": 5.5760168298964874e-05, "loss": 0.4397, "step": 447 }, { "epoch": 2.17462482946794, "grad_norm": 0.3659363982261399, "learning_rate": 5.563548407650782e-05, "loss": 0.4464, "step": 448 }, { "epoch": 2.1794755191753827, "grad_norm": 0.3139974928184246, "learning_rate": 5.551062027685187e-05, "loss": 0.4487, "step": 449 }, { "epoch": 2.1843262088828257, "grad_norm": 0.25894377780084493, "learning_rate": 5.5385578334087006e-05, "loss": 0.4481, "step": 450 }, { "epoch": 2.189176898590268, "grad_norm": 0.17951964246489394, "learning_rate": 5.526035968434927e-05, "loss": 0.4469, "step": 451 }, { "epoch": 2.194027588297711, "grad_norm": 0.21554515294483917, "learning_rate": 5.513496576580418e-05, "loss": 0.4573, "step": 452 }, { "epoch": 2.1988782780051537, "grad_norm": 0.24799983544152385, "learning_rate": 5.5009398018630276e-05, "loss": 0.4498, "step": 453 }, { "epoch": 2.2037289677125966, "grad_norm": 0.186973375547793, "learning_rate": 5.4883657885002575e-05, "loss": 0.4449, "step": 454 }, { "epoch": 2.2085796574200396, "grad_norm": 0.1888483688795653, "learning_rate": 5.475774680907597e-05, "loss": 0.443, "step": 455 }, { "epoch": 2.213430347127482, "grad_norm": 0.18914695692387654, "learning_rate": 5.463166623696868e-05, "loss": 0.4434, "step": 456 }, { "epoch": 2.218281036834925, "grad_norm": 0.19070268199404652, "learning_rate": 5.450541761674562e-05, "loss": 0.4445, "step": 457 }, { "epoch": 2.2231317265423676, "grad_norm": 0.2259939851217927, "learning_rate": 5.437900239840179e-05, "loss": 0.4465, "step": 458 }, { "epoch": 2.2279824162498105, "grad_norm": 0.18023421500987896, "learning_rate": 5.42524220338456e-05, "loss": 0.4453, "step": 459 }, { "epoch": 2.2328331059572535, "grad_norm": 0.16598518940751159, "learning_rate": 5.412567797688219e-05, "loss": 0.4498, "step": 460 }, { "epoch": 2.237683795664696, "grad_norm": 0.16004882092407235, "learning_rate": 5.3998771683196754e-05, "loss": 0.4527, "step": 461 }, { "epoch": 2.242534485372139, "grad_norm": 0.17890202722776521, "learning_rate": 5.3871704610337836e-05, "loss": 0.4444, "step": 462 }, { "epoch": 2.2473851750795815, "grad_norm": 0.182744670257566, "learning_rate": 5.374447821770053e-05, "loss": 0.4431, "step": 463 }, { "epoch": 2.2522358647870244, "grad_norm": 0.19342699163866056, "learning_rate": 5.361709396650977e-05, "loss": 0.4404, "step": 464 }, { "epoch": 2.2570865544944674, "grad_norm": 0.20442779398031627, "learning_rate": 5.3489553319803566e-05, "loss": 0.4496, "step": 465 }, { "epoch": 2.26193724420191, "grad_norm": 0.23359410284964036, "learning_rate": 5.336185774241609e-05, "loss": 0.4469, "step": 466 }, { "epoch": 2.266787933909353, "grad_norm": 0.23295417695606166, "learning_rate": 5.3234008700961e-05, "loss": 0.4505, "step": 467 }, { "epoch": 2.2716386236167954, "grad_norm": 0.20207094824496044, "learning_rate": 5.3106007663814505e-05, "loss": 0.4406, "step": 468 }, { "epoch": 2.2764893133242383, "grad_norm": 0.1850252157104855, "learning_rate": 5.2977856101098484e-05, "loss": 0.4525, "step": 469 }, { "epoch": 2.281340003031681, "grad_norm": 0.1821206965545461, "learning_rate": 5.284955548466371e-05, "loss": 0.4592, "step": 470 }, { "epoch": 2.286190692739124, "grad_norm": 0.1913435003815255, "learning_rate": 5.272110728807279e-05, "loss": 0.4459, "step": 471 }, { "epoch": 2.2910413824465667, "grad_norm": 0.17908151714339782, "learning_rate": 5.25925129865834e-05, "loss": 0.4523, "step": 472 }, { "epoch": 2.2958920721540093, "grad_norm": 0.17796456682985312, "learning_rate": 5.246377405713121e-05, "loss": 0.4426, "step": 473 }, { "epoch": 2.300742761861452, "grad_norm": 0.16950491734508644, "learning_rate": 5.2334891978313006e-05, "loss": 0.4426, "step": 474 }, { "epoch": 2.305593451568895, "grad_norm": 0.18036359667208995, "learning_rate": 5.220586823036966e-05, "loss": 0.4458, "step": 475 }, { "epoch": 2.3104441412763377, "grad_norm": 0.1873060682555774, "learning_rate": 5.207670429516915e-05, "loss": 0.4433, "step": 476 }, { "epoch": 2.3152948309837806, "grad_norm": 0.1986452939709168, "learning_rate": 5.1947401656189546e-05, "loss": 0.4593, "step": 477 }, { "epoch": 2.320145520691223, "grad_norm": 0.20494449288937291, "learning_rate": 5.181796179850197e-05, "loss": 0.4424, "step": 478 }, { "epoch": 2.324996210398666, "grad_norm": 0.17952910191793728, "learning_rate": 5.168838620875352e-05, "loss": 0.4503, "step": 479 }, { "epoch": 2.3298469001061086, "grad_norm": 0.1909701609269039, "learning_rate": 5.155867637515019e-05, "loss": 0.4506, "step": 480 }, { "epoch": 2.3346975898135516, "grad_norm": 0.1838298898954926, "learning_rate": 5.142883378743984e-05, "loss": 0.4513, "step": 481 }, { "epoch": 2.3395482795209945, "grad_norm": 0.1818267641683358, "learning_rate": 5.129885993689502e-05, "loss": 0.4488, "step": 482 }, { "epoch": 2.344398969228437, "grad_norm": 0.21197375880432345, "learning_rate": 5.116875631629585e-05, "loss": 0.4456, "step": 483 }, { "epoch": 2.34924965893588, "grad_norm": 0.21240893965447508, "learning_rate": 5.10385244199129e-05, "loss": 0.4386, "step": 484 }, { "epoch": 2.354100348643323, "grad_norm": 0.19244612255162405, "learning_rate": 5.0908165743490047e-05, "loss": 0.4482, "step": 485 }, { "epoch": 2.3589510383507655, "grad_norm": 0.22440529731925618, "learning_rate": 5.0777681784227224e-05, "loss": 0.4496, "step": 486 }, { "epoch": 2.3638017280582084, "grad_norm": 0.249440062974833, "learning_rate": 5.064707404076327e-05, "loss": 0.4502, "step": 487 }, { "epoch": 2.368652417765651, "grad_norm": 0.2374206142112278, "learning_rate": 5.051634401315875e-05, "loss": 0.448, "step": 488 }, { "epoch": 2.373503107473094, "grad_norm": 0.21044332969367502, "learning_rate": 5.0385493202878656e-05, "loss": 0.4416, "step": 489 }, { "epoch": 2.3783537971805364, "grad_norm": 0.15343545111269605, "learning_rate": 5.025452311277522e-05, "loss": 0.4413, "step": 490 }, { "epoch": 2.3832044868879794, "grad_norm": 0.17472771019103053, "learning_rate": 5.01234352470706e-05, "loss": 0.4472, "step": 491 }, { "epoch": 2.3880551765954223, "grad_norm": 0.2225509747823868, "learning_rate": 4.999223111133968e-05, "loss": 0.4405, "step": 492 }, { "epoch": 2.392905866302865, "grad_norm": 0.27110633436791925, "learning_rate": 4.986091221249269e-05, "loss": 0.44, "step": 493 }, { "epoch": 2.397756556010308, "grad_norm": 0.24255464597168586, "learning_rate": 4.972948005875796e-05, "loss": 0.4432, "step": 494 }, { "epoch": 2.4026072457177503, "grad_norm": 0.2496648842091371, "learning_rate": 4.959793615966459e-05, "loss": 0.4401, "step": 495 }, { "epoch": 2.4074579354251933, "grad_norm": 0.24806426439634907, "learning_rate": 4.946628202602508e-05, "loss": 0.4526, "step": 496 }, { "epoch": 2.412308625132636, "grad_norm": 0.21808090914084832, "learning_rate": 4.933451916991802e-05, "loss": 0.4474, "step": 497 }, { "epoch": 2.4171593148400787, "grad_norm": 0.19833835766366836, "learning_rate": 4.920264910467066e-05, "loss": 0.4485, "step": 498 }, { "epoch": 2.4220100045475217, "grad_norm": 0.1904056029579938, "learning_rate": 4.9070673344841645e-05, "loss": 0.4471, "step": 499 }, { "epoch": 2.426860694254964, "grad_norm": 0.17821880940044135, "learning_rate": 4.893859340620348e-05, "loss": 0.4518, "step": 500 }, { "epoch": 2.431711383962407, "grad_norm": 0.16242846601925154, "learning_rate": 4.880641080572522e-05, "loss": 0.4426, "step": 501 }, { "epoch": 2.43656207366985, "grad_norm": 0.16230843192633562, "learning_rate": 4.8674127061555025e-05, "loss": 0.4492, "step": 502 }, { "epoch": 2.4414127633772926, "grad_norm": 0.18692985874064466, "learning_rate": 4.8541743693002676e-05, "loss": 0.4576, "step": 503 }, { "epoch": 2.4462634530847356, "grad_norm": 0.17489245993778632, "learning_rate": 4.8409262220522196e-05, "loss": 0.4476, "step": 504 }, { "epoch": 2.451114142792178, "grad_norm": 0.14061959670906948, "learning_rate": 4.8276684165694336e-05, "loss": 0.4479, "step": 505 }, { "epoch": 2.455964832499621, "grad_norm": 0.17289206898304424, "learning_rate": 4.814401105120914e-05, "loss": 0.4479, "step": 506 }, { "epoch": 2.460815522207064, "grad_norm": 0.21063084112901795, "learning_rate": 4.8011244400848414e-05, "loss": 0.4466, "step": 507 }, { "epoch": 2.4656662119145065, "grad_norm": 0.2134287283629687, "learning_rate": 4.787838573946825e-05, "loss": 0.4503, "step": 508 }, { "epoch": 2.4705169016219495, "grad_norm": 0.19387557882251144, "learning_rate": 4.774543659298152e-05, "loss": 0.4419, "step": 509 }, { "epoch": 2.475367591329392, "grad_norm": 0.1690053079886072, "learning_rate": 4.761239848834031e-05, "loss": 0.4443, "step": 510 }, { "epoch": 2.480218281036835, "grad_norm": 0.17406180900609755, "learning_rate": 4.747927295351845e-05, "loss": 0.4474, "step": 511 }, { "epoch": 2.485068970744278, "grad_norm": 0.2024050850623432, "learning_rate": 4.734606151749389e-05, "loss": 0.4473, "step": 512 }, { "epoch": 2.4899196604517204, "grad_norm": 0.22659357369802574, "learning_rate": 4.7212765710231204e-05, "loss": 0.4481, "step": 513 }, { "epoch": 2.4947703501591634, "grad_norm": 0.23595309939097722, "learning_rate": 4.707938706266397e-05, "loss": 0.4484, "step": 514 }, { "epoch": 2.499621039866606, "grad_norm": 0.19918584163751257, "learning_rate": 4.694592710667723e-05, "loss": 0.444, "step": 515 }, { "epoch": 2.504471729574049, "grad_norm": 0.18418670752131802, "learning_rate": 4.681238737508983e-05, "loss": 0.4424, "step": 516 }, { "epoch": 2.5093224192814914, "grad_norm": 0.1926237495649244, "learning_rate": 4.6678769401636894e-05, "loss": 0.4444, "step": 517 }, { "epoch": 2.5141731089889343, "grad_norm": 0.20706125086296728, "learning_rate": 4.6545074720952166e-05, "loss": 0.456, "step": 518 }, { "epoch": 2.5190237986963773, "grad_norm": 0.17699664563372686, "learning_rate": 4.641130486855038e-05, "loss": 0.4396, "step": 519 }, { "epoch": 2.52387448840382, "grad_norm": 0.18317752602670304, "learning_rate": 4.627746138080966e-05, "loss": 0.4432, "step": 520 }, { "epoch": 2.5287251781112627, "grad_norm": 0.2190424482227647, "learning_rate": 4.614354579495379e-05, "loss": 0.4448, "step": 521 }, { "epoch": 2.5335758678187057, "grad_norm": 0.20135719119048615, "learning_rate": 4.6009559649034695e-05, "loss": 0.4432, "step": 522 }, { "epoch": 2.538426557526148, "grad_norm": 0.20782420010728125, "learning_rate": 4.587550448191465e-05, "loss": 0.4474, "step": 523 }, { "epoch": 2.543277247233591, "grad_norm": 0.18668296726800496, "learning_rate": 4.5741381833248655e-05, "loss": 0.455, "step": 524 }, { "epoch": 2.5481279369410337, "grad_norm": 0.17935132627421838, "learning_rate": 4.560719324346677e-05, "loss": 0.4457, "step": 525 }, { "epoch": 2.5529786266484766, "grad_norm": 0.16835981784522308, "learning_rate": 4.547294025375641e-05, "loss": 0.4478, "step": 526 }, { "epoch": 2.557829316355919, "grad_norm": 0.18084487512355504, "learning_rate": 4.533862440604461e-05, "loss": 0.447, "step": 527 }, { "epoch": 2.562680006063362, "grad_norm": 0.17384784743298828, "learning_rate": 4.520424724298036e-05, "loss": 0.4408, "step": 528 }, { "epoch": 2.567530695770805, "grad_norm": 0.20150460275113774, "learning_rate": 4.5069810307916874e-05, "loss": 0.4441, "step": 529 }, { "epoch": 2.5723813854782476, "grad_norm": 0.2240004858996321, "learning_rate": 4.493531514489385e-05, "loss": 0.4425, "step": 530 }, { "epoch": 2.5772320751856905, "grad_norm": 0.2286831099325836, "learning_rate": 4.480076329861977e-05, "loss": 0.4433, "step": 531 }, { "epoch": 2.5820827648931335, "grad_norm": 0.2048648110357608, "learning_rate": 4.46661563144541e-05, "loss": 0.4487, "step": 532 }, { "epoch": 2.586933454600576, "grad_norm": 0.2121488428415987, "learning_rate": 4.453149573838962e-05, "loss": 0.4445, "step": 533 }, { "epoch": 2.591784144308019, "grad_norm": 0.18541474138380978, "learning_rate": 4.43967831170346e-05, "loss": 0.4494, "step": 534 }, { "epoch": 2.5966348340154615, "grad_norm": 0.17452762694525445, "learning_rate": 4.426201999759505e-05, "loss": 0.4484, "step": 535 }, { "epoch": 2.6014855237229044, "grad_norm": 0.21944224068377363, "learning_rate": 4.4127207927857e-05, "loss": 0.4419, "step": 536 }, { "epoch": 2.606336213430347, "grad_norm": 0.17656272332454842, "learning_rate": 4.3992348456168666e-05, "loss": 0.4568, "step": 537 }, { "epoch": 2.61118690313779, "grad_norm": 0.1892562653364182, "learning_rate": 4.385744313142267e-05, "loss": 0.4427, "step": 538 }, { "epoch": 2.616037592845233, "grad_norm": 0.21611454670373548, "learning_rate": 4.372249350303828e-05, "loss": 0.4418, "step": 539 }, { "epoch": 2.6208882825526754, "grad_norm": 0.17168747953224547, "learning_rate": 4.358750112094363e-05, "loss": 0.4544, "step": 540 }, { "epoch": 2.6257389722601183, "grad_norm": 0.17941819138400728, "learning_rate": 4.3452467535557846e-05, "loss": 0.4372, "step": 541 }, { "epoch": 2.6305896619675613, "grad_norm": 0.2025265834742146, "learning_rate": 4.3317394297773304e-05, "loss": 0.4517, "step": 542 }, { "epoch": 2.635440351675004, "grad_norm": 0.20441246530938206, "learning_rate": 4.3182282958937816e-05, "loss": 0.4333, "step": 543 }, { "epoch": 2.6402910413824463, "grad_norm": 0.2334105452950634, "learning_rate": 4.304713507083673e-05, "loss": 0.4481, "step": 544 }, { "epoch": 2.6451417310898893, "grad_norm": 0.26291969340773214, "learning_rate": 4.291195218567523e-05, "loss": 0.4466, "step": 545 }, { "epoch": 2.649992420797332, "grad_norm": 0.1863631298156993, "learning_rate": 4.277673585606046e-05, "loss": 0.4405, "step": 546 }, { "epoch": 2.6548431105047747, "grad_norm": 0.23226855973797117, "learning_rate": 4.264148763498364e-05, "loss": 0.4566, "step": 547 }, { "epoch": 2.6596938002122177, "grad_norm": 0.30482274820740174, "learning_rate": 4.250620907580226e-05, "loss": 0.4407, "step": 548 }, { "epoch": 2.6645444899196606, "grad_norm": 0.23781311620065457, "learning_rate": 4.237090173222231e-05, "loss": 0.4493, "step": 549 }, { "epoch": 2.669395179627103, "grad_norm": 0.1808214801234254, "learning_rate": 4.223556715828033e-05, "loss": 0.4511, "step": 550 }, { "epoch": 2.674245869334546, "grad_norm": 0.26315804734468673, "learning_rate": 4.2100206908325603e-05, "loss": 0.447, "step": 551 }, { "epoch": 2.6790965590419886, "grad_norm": 0.25781234163394623, "learning_rate": 4.196482253700235e-05, "loss": 0.4415, "step": 552 }, { "epoch": 2.6839472487494316, "grad_norm": 0.17133762584152984, "learning_rate": 4.182941559923179e-05, "loss": 0.4457, "step": 553 }, { "epoch": 2.688797938456874, "grad_norm": 0.2266803612041648, "learning_rate": 4.169398765019433e-05, "loss": 0.4422, "step": 554 }, { "epoch": 2.693648628164317, "grad_norm": 0.23286738752123257, "learning_rate": 4.15585402453117e-05, "loss": 0.4429, "step": 555 }, { "epoch": 2.69849931787176, "grad_norm": 0.20226496811604636, "learning_rate": 4.14230749402291e-05, "loss": 0.4421, "step": 556 }, { "epoch": 2.7033500075792025, "grad_norm": 0.21746634743317236, "learning_rate": 4.128759329079732e-05, "loss": 0.4318, "step": 557 }, { "epoch": 2.7082006972866455, "grad_norm": 0.24285493960537577, "learning_rate": 4.115209685305482e-05, "loss": 0.4374, "step": 558 }, { "epoch": 2.7130513869940884, "grad_norm": 0.20035101285126697, "learning_rate": 4.101658718320998e-05, "loss": 0.4429, "step": 559 }, { "epoch": 2.717902076701531, "grad_norm": 0.1733102653989901, "learning_rate": 4.088106583762309e-05, "loss": 0.4456, "step": 560 }, { "epoch": 2.722752766408974, "grad_norm": 0.25116764609287723, "learning_rate": 4.074553437278857e-05, "loss": 0.4494, "step": 561 }, { "epoch": 2.7276034561164164, "grad_norm": 0.19896329775589092, "learning_rate": 4.060999434531704e-05, "loss": 0.4449, "step": 562 }, { "epoch": 2.7324541458238594, "grad_norm": 0.16517536428811208, "learning_rate": 4.047444731191751e-05, "loss": 0.4426, "step": 563 }, { "epoch": 2.737304835531302, "grad_norm": 0.1656807626865065, "learning_rate": 4.033889482937943e-05, "loss": 0.4445, "step": 564 }, { "epoch": 2.742155525238745, "grad_norm": 0.15103159619749504, "learning_rate": 4.020333845455478e-05, "loss": 0.4565, "step": 565 }, { "epoch": 2.747006214946188, "grad_norm": 0.16996820086522443, "learning_rate": 4.0067779744340345e-05, "loss": 0.4459, "step": 566 }, { "epoch": 2.7518569046536303, "grad_norm": 0.1495970266083701, "learning_rate": 3.993222025565966e-05, "loss": 0.4447, "step": 567 }, { "epoch": 2.7567075943610733, "grad_norm": 0.15458974892236554, "learning_rate": 3.979666154544522e-05, "loss": 0.4452, "step": 568 }, { "epoch": 2.7615582840685162, "grad_norm": 0.1741093401099396, "learning_rate": 3.96611051706206e-05, "loss": 0.4421, "step": 569 }, { "epoch": 2.7664089737759587, "grad_norm": 0.1819530197226333, "learning_rate": 3.9525552688082494e-05, "loss": 0.4509, "step": 570 }, { "epoch": 2.7712596634834017, "grad_norm": 0.14996389947080183, "learning_rate": 3.939000565468297e-05, "loss": 0.4442, "step": 571 }, { "epoch": 2.776110353190844, "grad_norm": 0.19015205226216172, "learning_rate": 3.9254465627211444e-05, "loss": 0.4458, "step": 572 }, { "epoch": 2.780961042898287, "grad_norm": 0.2147271939320094, "learning_rate": 3.911893416237693e-05, "loss": 0.4423, "step": 573 }, { "epoch": 2.7858117326057297, "grad_norm": 0.17368493707493848, "learning_rate": 3.8983412816790045e-05, "loss": 0.4415, "step": 574 }, { "epoch": 2.7906624223131726, "grad_norm": 0.18366782229193682, "learning_rate": 3.8847903146945186e-05, "loss": 0.4419, "step": 575 }, { "epoch": 2.7955131120206156, "grad_norm": 0.1770373720928735, "learning_rate": 3.871240670920269e-05, "loss": 0.4477, "step": 576 }, { "epoch": 2.800363801728058, "grad_norm": 0.15508137985676013, "learning_rate": 3.85769250597709e-05, "loss": 0.4458, "step": 577 }, { "epoch": 2.805214491435501, "grad_norm": 0.15518418619016236, "learning_rate": 3.844145975468832e-05, "loss": 0.4403, "step": 578 }, { "epoch": 2.810065181142944, "grad_norm": 0.13016021632650948, "learning_rate": 3.830601234980569e-05, "loss": 0.4509, "step": 579 }, { "epoch": 2.8149158708503865, "grad_norm": 0.17125198843453068, "learning_rate": 3.8170584400768224e-05, "loss": 0.4492, "step": 580 }, { "epoch": 2.8197665605578295, "grad_norm": 0.18182599605587274, "learning_rate": 3.8035177462997664e-05, "loss": 0.4475, "step": 581 }, { "epoch": 2.824617250265272, "grad_norm": 0.16612208849559923, "learning_rate": 3.7899793091674396e-05, "loss": 0.4419, "step": 582 }, { "epoch": 2.829467939972715, "grad_norm": 0.15504382692612345, "learning_rate": 3.776443284171969e-05, "loss": 0.4421, "step": 583 }, { "epoch": 2.8343186296801575, "grad_norm": 0.14158493240403466, "learning_rate": 3.7629098267777706e-05, "loss": 0.4399, "step": 584 }, { "epoch": 2.8391693193876004, "grad_norm": 0.14521790840725082, "learning_rate": 3.7493790924197746e-05, "loss": 0.4328, "step": 585 }, { "epoch": 2.8440200090950434, "grad_norm": 0.14223117063886642, "learning_rate": 3.735851236501637e-05, "loss": 0.4403, "step": 586 }, { "epoch": 2.848870698802486, "grad_norm": 0.13716153813556975, "learning_rate": 3.722326414393954e-05, "loss": 0.4375, "step": 587 }, { "epoch": 2.853721388509929, "grad_norm": 0.13481835773066367, "learning_rate": 3.708804781432478e-05, "loss": 0.4465, "step": 588 }, { "epoch": 2.858572078217372, "grad_norm": 0.15098861155914894, "learning_rate": 3.6952864929163286e-05, "loss": 0.4478, "step": 589 }, { "epoch": 2.8634227679248143, "grad_norm": 0.14428029096945075, "learning_rate": 3.6817717041062204e-05, "loss": 0.4433, "step": 590 }, { "epoch": 2.868273457632257, "grad_norm": 0.14941685817838865, "learning_rate": 3.66826057022267e-05, "loss": 0.4426, "step": 591 }, { "epoch": 2.8731241473397, "grad_norm": 0.15814915640950794, "learning_rate": 3.654753246444217e-05, "loss": 0.437, "step": 592 }, { "epoch": 2.8779748370471427, "grad_norm": 0.1666350361712344, "learning_rate": 3.641249887905638e-05, "loss": 0.4404, "step": 593 }, { "epoch": 2.8828255267545853, "grad_norm": 0.14761581189993908, "learning_rate": 3.627750649696173e-05, "loss": 0.4418, "step": 594 }, { "epoch": 2.887676216462028, "grad_norm": 0.16433140815176697, "learning_rate": 3.614255686857734e-05, "loss": 0.4482, "step": 595 }, { "epoch": 2.892526906169471, "grad_norm": 0.14898117787478918, "learning_rate": 3.600765154383134e-05, "loss": 0.4407, "step": 596 }, { "epoch": 2.8973775958769137, "grad_norm": 0.1576311183401583, "learning_rate": 3.587279207214301e-05, "loss": 0.4502, "step": 597 }, { "epoch": 2.9022282855843566, "grad_norm": 0.12713739329535242, "learning_rate": 3.5737980002404965e-05, "loss": 0.4504, "step": 598 }, { "epoch": 2.907078975291799, "grad_norm": 0.1407162797770253, "learning_rate": 3.5603216882965415e-05, "loss": 0.4444, "step": 599 }, { "epoch": 2.911929664999242, "grad_norm": 0.15096239269549241, "learning_rate": 3.5468504261610387e-05, "loss": 0.4497, "step": 600 }, { "epoch": 2.9167803547066846, "grad_norm": 0.1341187523707445, "learning_rate": 3.5333843685545914e-05, "loss": 0.4449, "step": 601 }, { "epoch": 2.9216310444141276, "grad_norm": 0.1619335943958044, "learning_rate": 3.519923670138025e-05, "loss": 0.4434, "step": 602 }, { "epoch": 2.9264817341215705, "grad_norm": 0.16211928956809254, "learning_rate": 3.506468485510616e-05, "loss": 0.4394, "step": 603 }, { "epoch": 2.931332423829013, "grad_norm": 0.13565985896740923, "learning_rate": 3.493018969208314e-05, "loss": 0.4513, "step": 604 }, { "epoch": 2.936183113536456, "grad_norm": 0.14551792578562125, "learning_rate": 3.479575275701965e-05, "loss": 0.4425, "step": 605 }, { "epoch": 2.941033803243899, "grad_norm": 0.17680002225930672, "learning_rate": 3.4661375593955405e-05, "loss": 0.4384, "step": 606 }, { "epoch": 2.9458844929513415, "grad_norm": 0.16447736465809054, "learning_rate": 3.45270597462436e-05, "loss": 0.4451, "step": 607 }, { "epoch": 2.9507351826587844, "grad_norm": 0.15299942108774958, "learning_rate": 3.4392806756533233e-05, "loss": 0.4383, "step": 608 }, { "epoch": 2.955585872366227, "grad_norm": 0.1536623395643003, "learning_rate": 3.425861816675135e-05, "loss": 0.4453, "step": 609 }, { "epoch": 2.96043656207367, "grad_norm": 0.15727509499916784, "learning_rate": 3.4124495518085366e-05, "loss": 0.436, "step": 610 }, { "epoch": 2.9652872517811124, "grad_norm": 0.17906600408772821, "learning_rate": 3.399044035096532e-05, "loss": 0.4467, "step": 611 }, { "epoch": 2.9701379414885554, "grad_norm": 0.16243556143131102, "learning_rate": 3.3856454205046223e-05, "loss": 0.4364, "step": 612 }, { "epoch": 2.9749886311959983, "grad_norm": 0.1608114743453715, "learning_rate": 3.372253861919036e-05, "loss": 0.4517, "step": 613 }, { "epoch": 2.979839320903441, "grad_norm": 0.16249676695986184, "learning_rate": 3.3588695131449626e-05, "loss": 0.4464, "step": 614 }, { "epoch": 2.984690010610884, "grad_norm": 0.14682205641070967, "learning_rate": 3.3454925279047854e-05, "loss": 0.4446, "step": 615 }, { "epoch": 2.9895407003183267, "grad_norm": 0.158750060737996, "learning_rate": 3.3321230598363126e-05, "loss": 0.4449, "step": 616 }, { "epoch": 2.9943913900257693, "grad_norm": 0.15103569523913019, "learning_rate": 3.3187612624910185e-05, "loss": 0.4457, "step": 617 }, { "epoch": 2.999242079733212, "grad_norm": 0.13279281722008607, "learning_rate": 3.305407289332279e-05, "loss": 0.4524, "step": 618 }, { "epoch": 3.004850689707443, "grad_norm": 0.21786343264375474, "learning_rate": 3.2920612937336035e-05, "loss": 0.4186, "step": 619 }, { "epoch": 3.0097013794148855, "grad_norm": 0.1940657522375173, "learning_rate": 3.2787234289768816e-05, "loss": 0.4198, "step": 620 }, { "epoch": 3.0145520691223284, "grad_norm": 0.21773031084989108, "learning_rate": 3.2653938482506125e-05, "loss": 0.4257, "step": 621 }, { "epoch": 3.019402758829771, "grad_norm": 0.20326547030411232, "learning_rate": 3.252072704648157e-05, "loss": 0.4165, "step": 622 }, { "epoch": 3.024253448537214, "grad_norm": 0.1952882443077655, "learning_rate": 3.2387601511659695e-05, "loss": 0.4099, "step": 623 }, { "epoch": 3.029104138244657, "grad_norm": 0.1918042514566916, "learning_rate": 3.22545634070185e-05, "loss": 0.4152, "step": 624 }, { "epoch": 3.0339548279520994, "grad_norm": 0.19705116304049744, "learning_rate": 3.212161426053177e-05, "loss": 0.4128, "step": 625 }, { "epoch": 3.0388055176595423, "grad_norm": 0.19649678060258355, "learning_rate": 3.19887555991516e-05, "loss": 0.4129, "step": 626 }, { "epoch": 3.043656207366985, "grad_norm": 0.21708556979200094, "learning_rate": 3.1855988948790866e-05, "loss": 0.419, "step": 627 }, { "epoch": 3.048506897074428, "grad_norm": 0.1975715363358727, "learning_rate": 3.172331583430567e-05, "loss": 0.4179, "step": 628 }, { "epoch": 3.0533575867818707, "grad_norm": 0.23335868145268443, "learning_rate": 3.1590737779477825e-05, "loss": 0.4187, "step": 629 }, { "epoch": 3.0582082764893133, "grad_norm": 0.16312780113553543, "learning_rate": 3.145825630699734e-05, "loss": 0.4145, "step": 630 }, { "epoch": 3.063058966196756, "grad_norm": 0.1995031684584169, "learning_rate": 3.1325872938444995e-05, "loss": 0.4269, "step": 631 }, { "epoch": 3.0679096559041987, "grad_norm": 0.16321617015370155, "learning_rate": 3.119358919427478e-05, "loss": 0.4189, "step": 632 }, { "epoch": 3.0727603456116417, "grad_norm": 0.18410001775256316, "learning_rate": 3.106140659379652e-05, "loss": 0.4218, "step": 633 }, { "epoch": 3.0776110353190846, "grad_norm": 0.15016856412332835, "learning_rate": 3.092932665515837e-05, "loss": 0.4142, "step": 634 }, { "epoch": 3.082461725026527, "grad_norm": 0.1502499895895468, "learning_rate": 3.079735089532935e-05, "loss": 0.4092, "step": 635 }, { "epoch": 3.08731241473397, "grad_norm": 0.1564011362374121, "learning_rate": 3.0665480830082e-05, "loss": 0.4176, "step": 636 }, { "epoch": 3.0921631044414126, "grad_norm": 0.15490177035198396, "learning_rate": 3.0533717973974924e-05, "loss": 0.4171, "step": 637 }, { "epoch": 3.0970137941488556, "grad_norm": 0.16842243538179394, "learning_rate": 3.040206384033542e-05, "loss": 0.4142, "step": 638 }, { "epoch": 3.101864483856298, "grad_norm": 0.15647723657119908, "learning_rate": 3.0270519941242052e-05, "loss": 0.4045, "step": 639 }, { "epoch": 3.106715173563741, "grad_norm": 0.14187950473293476, "learning_rate": 3.0139087787507323e-05, "loss": 0.4162, "step": 640 }, { "epoch": 3.111565863271184, "grad_norm": 0.1486994798612613, "learning_rate": 3.0007768888660337e-05, "loss": 0.4162, "step": 641 }, { "epoch": 3.1164165529786265, "grad_norm": 0.14045982542928215, "learning_rate": 2.9876564752929406e-05, "loss": 0.423, "step": 642 }, { "epoch": 3.1212672426860695, "grad_norm": 0.14911944405394412, "learning_rate": 2.9745476887224806e-05, "loss": 0.4186, "step": 643 }, { "epoch": 3.1261179323935124, "grad_norm": 0.130768356118423, "learning_rate": 2.961450679712135e-05, "loss": 0.4149, "step": 644 }, { "epoch": 3.130968622100955, "grad_norm": 0.15344876789210227, "learning_rate": 2.9483655986841265e-05, "loss": 0.4185, "step": 645 }, { "epoch": 3.135819311808398, "grad_norm": 0.16373681347456412, "learning_rate": 2.9352925959236732e-05, "loss": 0.4199, "step": 646 }, { "epoch": 3.1406700015158404, "grad_norm": 0.13681917154319687, "learning_rate": 2.92223182157728e-05, "loss": 0.4187, "step": 647 }, { "epoch": 3.1455206912232834, "grad_norm": 0.16376916195733163, "learning_rate": 2.909183425650996e-05, "loss": 0.4144, "step": 648 }, { "epoch": 3.150371380930726, "grad_norm": 0.15231760580025508, "learning_rate": 2.8961475580087108e-05, "loss": 0.4065, "step": 649 }, { "epoch": 3.155222070638169, "grad_norm": 0.1482250629623317, "learning_rate": 2.8831243683704162e-05, "loss": 0.4167, "step": 650 }, { "epoch": 3.160072760345612, "grad_norm": 0.1279306947970543, "learning_rate": 2.8701140063104996e-05, "loss": 0.4163, "step": 651 }, { "epoch": 3.1649234500530543, "grad_norm": 0.14839352063841615, "learning_rate": 2.857116621256018e-05, "loss": 0.4066, "step": 652 }, { "epoch": 3.1697741397604973, "grad_norm": 0.14182099741632095, "learning_rate": 2.8441323624849827e-05, "loss": 0.4073, "step": 653 }, { "epoch": 3.17462482946794, "grad_norm": 0.14062989686711488, "learning_rate": 2.83116137912465e-05, "loss": 0.4148, "step": 654 }, { "epoch": 3.1794755191753827, "grad_norm": 0.13350262763245205, "learning_rate": 2.8182038201498038e-05, "loss": 0.4185, "step": 655 }, { "epoch": 3.1843262088828257, "grad_norm": 0.14000623741594295, "learning_rate": 2.8052598343810474e-05, "loss": 0.4084, "step": 656 }, { "epoch": 3.189176898590268, "grad_norm": 0.13922618624994615, "learning_rate": 2.7923295704830868e-05, "loss": 0.4209, "step": 657 }, { "epoch": 3.194027588297711, "grad_norm": 0.13424360868636123, "learning_rate": 2.7794131769630355e-05, "loss": 0.4203, "step": 658 }, { "epoch": 3.1988782780051537, "grad_norm": 0.1410971557670894, "learning_rate": 2.7665108021687007e-05, "loss": 0.4229, "step": 659 }, { "epoch": 3.2037289677125966, "grad_norm": 0.15216131334925778, "learning_rate": 2.753622594286879e-05, "loss": 0.4145, "step": 660 }, { "epoch": 3.2085796574200396, "grad_norm": 0.1261567069468345, "learning_rate": 2.7407487013416615e-05, "loss": 0.4083, "step": 661 }, { "epoch": 3.213430347127482, "grad_norm": 0.15853568579534694, "learning_rate": 2.727889271192722e-05, "loss": 0.4187, "step": 662 }, { "epoch": 3.218281036834925, "grad_norm": 0.1151093763460037, "learning_rate": 2.715044451533631e-05, "loss": 0.4164, "step": 663 }, { "epoch": 3.2231317265423676, "grad_norm": 0.14296777545925654, "learning_rate": 2.702214389890152e-05, "loss": 0.413, "step": 664 }, { "epoch": 3.2279824162498105, "grad_norm": 0.12844538251966497, "learning_rate": 2.6893992336185512e-05, "loss": 0.4035, "step": 665 }, { "epoch": 3.2328331059572535, "grad_norm": 0.13036276585432338, "learning_rate": 2.6765991299039025e-05, "loss": 0.4145, "step": 666 }, { "epoch": 3.237683795664696, "grad_norm": 0.12243507743746514, "learning_rate": 2.663814225758393e-05, "loss": 0.4117, "step": 667 }, { "epoch": 3.242534485372139, "grad_norm": 0.14245972122975337, "learning_rate": 2.6510446680196448e-05, "loss": 0.4195, "step": 668 }, { "epoch": 3.2473851750795815, "grad_norm": 0.12799039988824565, "learning_rate": 2.638290603349023e-05, "loss": 0.4203, "step": 669 }, { "epoch": 3.2522358647870244, "grad_norm": 0.12961756855915293, "learning_rate": 2.625552178229949e-05, "loss": 0.4159, "step": 670 }, { "epoch": 3.2570865544944674, "grad_norm": 0.1296944618260485, "learning_rate": 2.612829538966218e-05, "loss": 0.4111, "step": 671 }, { "epoch": 3.26193724420191, "grad_norm": 0.14508116679367689, "learning_rate": 2.6001228316803256e-05, "loss": 0.4196, "step": 672 }, { "epoch": 3.266787933909353, "grad_norm": 0.1216509600449271, "learning_rate": 2.5874322023117824e-05, "loss": 0.4162, "step": 673 }, { "epoch": 3.2716386236167954, "grad_norm": 0.1407698696044201, "learning_rate": 2.5747577966154404e-05, "loss": 0.4165, "step": 674 }, { "epoch": 3.2764893133242383, "grad_norm": 0.1142728140236527, "learning_rate": 2.5620997601598215e-05, "loss": 0.4076, "step": 675 }, { "epoch": 3.281340003031681, "grad_norm": 0.14212475123175447, "learning_rate": 2.5494582383254388e-05, "loss": 0.4174, "step": 676 }, { "epoch": 3.286190692739124, "grad_norm": 0.11798282133422631, "learning_rate": 2.5368333763031324e-05, "loss": 0.4131, "step": 677 }, { "epoch": 3.2910413824465667, "grad_norm": 0.14618443390836333, "learning_rate": 2.5242253190924034e-05, "loss": 0.4092, "step": 678 }, { "epoch": 3.2958920721540093, "grad_norm": 0.12815235524330332, "learning_rate": 2.5116342114997442e-05, "loss": 0.409, "step": 679 }, { "epoch": 3.300742761861452, "grad_norm": 0.1346079173937725, "learning_rate": 2.4990601981369737e-05, "loss": 0.4201, "step": 680 }, { "epoch": 3.305593451568895, "grad_norm": 0.12020401678661803, "learning_rate": 2.4865034234195834e-05, "loss": 0.4107, "step": 681 }, { "epoch": 3.3104441412763377, "grad_norm": 0.1342860022847603, "learning_rate": 2.4739640315650747e-05, "loss": 0.4145, "step": 682 }, { "epoch": 3.3152948309837806, "grad_norm": 0.11394559661374248, "learning_rate": 2.4614421665912997e-05, "loss": 0.4213, "step": 683 }, { "epoch": 3.320145520691223, "grad_norm": 0.126805017438777, "learning_rate": 2.4489379723148147e-05, "loss": 0.4129, "step": 684 }, { "epoch": 3.324996210398666, "grad_norm": 0.11817811298525939, "learning_rate": 2.4364515923492187e-05, "loss": 0.4193, "step": 685 }, { "epoch": 3.3298469001061086, "grad_norm": 0.1342332761111468, "learning_rate": 2.4239831701035143e-05, "loss": 0.418, "step": 686 }, { "epoch": 3.3346975898135516, "grad_norm": 0.12474527460295737, "learning_rate": 2.411532848780451e-05, "loss": 0.4166, "step": 687 }, { "epoch": 3.3395482795209945, "grad_norm": 0.1329059048345405, "learning_rate": 2.399100771374888e-05, "loss": 0.4138, "step": 688 }, { "epoch": 3.344398969228437, "grad_norm": 0.11628441384991241, "learning_rate": 2.3866870806721495e-05, "loss": 0.4111, "step": 689 }, { "epoch": 3.34924965893588, "grad_norm": 0.1453864432398833, "learning_rate": 2.37429191924638e-05, "loss": 0.42, "step": 690 }, { "epoch": 3.354100348643323, "grad_norm": 0.11264715608612798, "learning_rate": 2.361915429458913e-05, "loss": 0.417, "step": 691 }, { "epoch": 3.3589510383507655, "grad_norm": 0.14248256005839127, "learning_rate": 2.349557753456637e-05, "loss": 0.4168, "step": 692 }, { "epoch": 3.3638017280582084, "grad_norm": 0.12398727462550883, "learning_rate": 2.3372190331703556e-05, "loss": 0.4189, "step": 693 }, { "epoch": 3.368652417765651, "grad_norm": 0.11869463547338541, "learning_rate": 2.324899410313161e-05, "loss": 0.4125, "step": 694 }, { "epoch": 3.373503107473094, "grad_norm": 0.12377753865053726, "learning_rate": 2.3125990263788118e-05, "loss": 0.4186, "step": 695 }, { "epoch": 3.3783537971805364, "grad_norm": 0.13141561076446406, "learning_rate": 2.3003180226400986e-05, "loss": 0.4123, "step": 696 }, { "epoch": 3.3832044868879794, "grad_norm": 0.11816327910326484, "learning_rate": 2.288056540147229e-05, "loss": 0.4129, "step": 697 }, { "epoch": 3.3880551765954223, "grad_norm": 0.12260669143866527, "learning_rate": 2.275814719726201e-05, "loss": 0.4133, "step": 698 }, { "epoch": 3.392905866302865, "grad_norm": 0.133084483131333, "learning_rate": 2.263592701977193e-05, "loss": 0.4219, "step": 699 }, { "epoch": 3.397756556010308, "grad_norm": 0.11848073628628028, "learning_rate": 2.2513906272729397e-05, "loss": 0.4143, "step": 700 }, { "epoch": 3.4026072457177503, "grad_norm": 0.12858950370510128, "learning_rate": 2.239208635757133e-05, "loss": 0.4166, "step": 701 }, { "epoch": 3.4074579354251933, "grad_norm": 0.11722692816596028, "learning_rate": 2.2270468673428004e-05, "loss": 0.4259, "step": 702 }, { "epoch": 3.412308625132636, "grad_norm": 0.11830608786302087, "learning_rate": 2.2149054617106974e-05, "loss": 0.407, "step": 703 }, { "epoch": 3.4171593148400787, "grad_norm": 0.12285165179780408, "learning_rate": 2.2027845583077175e-05, "loss": 0.4231, "step": 704 }, { "epoch": 3.4220100045475217, "grad_norm": 0.12884220459818324, "learning_rate": 2.1906842963452757e-05, "loss": 0.4069, "step": 705 }, { "epoch": 3.426860694254964, "grad_norm": 0.10700415399515635, "learning_rate": 2.178604814797715e-05, "loss": 0.4149, "step": 706 }, { "epoch": 3.431711383962407, "grad_norm": 0.11853515304365536, "learning_rate": 2.1665462524007162e-05, "loss": 0.4125, "step": 707 }, { "epoch": 3.43656207366985, "grad_norm": 0.10772813217204756, "learning_rate": 2.1545087476496903e-05, "loss": 0.4216, "step": 708 }, { "epoch": 3.4414127633772926, "grad_norm": 0.12946073793938545, "learning_rate": 2.1424924387981996e-05, "loss": 0.4227, "step": 709 }, { "epoch": 3.4462634530847356, "grad_norm": 0.11256888467579416, "learning_rate": 2.1304974638563715e-05, "loss": 0.4116, "step": 710 }, { "epoch": 3.451114142792178, "grad_norm": 0.12334394272902899, "learning_rate": 2.1185239605893013e-05, "loss": 0.4217, "step": 711 }, { "epoch": 3.455964832499621, "grad_norm": 0.11467056354627854, "learning_rate": 2.106572066515482e-05, "loss": 0.4193, "step": 712 }, { "epoch": 3.460815522207064, "grad_norm": 0.11248104529388973, "learning_rate": 2.0946419189052162e-05, "loss": 0.4147, "step": 713 }, { "epoch": 3.4656662119145065, "grad_norm": 0.11839620237634582, "learning_rate": 2.0827336547790452e-05, "loss": 0.4214, "step": 714 }, { "epoch": 3.4705169016219495, "grad_norm": 0.11350680266976185, "learning_rate": 2.0708474109061752e-05, "loss": 0.416, "step": 715 }, { "epoch": 3.475367591329392, "grad_norm": 0.11310394993568851, "learning_rate": 2.0589833238029032e-05, "loss": 0.4001, "step": 716 }, { "epoch": 3.480218281036835, "grad_norm": 0.1142115014102716, "learning_rate": 2.0471415297310455e-05, "loss": 0.4158, "step": 717 }, { "epoch": 3.485068970744278, "grad_norm": 0.12114855144347077, "learning_rate": 2.0353221646963864e-05, "loss": 0.41, "step": 718 }, { "epoch": 3.4899196604517204, "grad_norm": 0.11155618948058961, "learning_rate": 2.0235253644471012e-05, "loss": 0.4226, "step": 719 }, { "epoch": 3.4947703501591634, "grad_norm": 0.12346483161916652, "learning_rate": 2.011751264472206e-05, "loss": 0.4163, "step": 720 }, { "epoch": 3.499621039866606, "grad_norm": 0.1314504891806893, "learning_rate": 2.0000000000000012e-05, "loss": 0.4185, "step": 721 }, { "epoch": 3.504471729574049, "grad_norm": 0.11587490363769114, "learning_rate": 1.9882717059965086e-05, "loss": 0.4243, "step": 722 }, { "epoch": 3.5093224192814914, "grad_norm": 0.147492515392888, "learning_rate": 1.9765665171639345e-05, "loss": 0.4139, "step": 723 }, { "epoch": 3.5141731089889343, "grad_norm": 0.13161876983358525, "learning_rate": 1.964884567939118e-05, "loss": 0.4023, "step": 724 }, { "epoch": 3.5190237986963773, "grad_norm": 0.1288012773595963, "learning_rate": 1.9532259924919823e-05, "loss": 0.416, "step": 725 }, { "epoch": 3.52387448840382, "grad_norm": 0.10979902623938809, "learning_rate": 1.9415909247239996e-05, "loss": 0.4198, "step": 726 }, { "epoch": 3.5287251781112627, "grad_norm": 0.12273321372911124, "learning_rate": 1.9299794982666485e-05, "loss": 0.4084, "step": 727 }, { "epoch": 3.5335758678187057, "grad_norm": 0.11485189001172974, "learning_rate": 1.9183918464798837e-05, "loss": 0.408, "step": 728 }, { "epoch": 3.538426557526148, "grad_norm": 0.1142263107632135, "learning_rate": 1.906828102450601e-05, "loss": 0.4131, "step": 729 }, { "epoch": 3.543277247233591, "grad_norm": 0.11235720950088848, "learning_rate": 1.895288398991114e-05, "loss": 0.4102, "step": 730 }, { "epoch": 3.5481279369410337, "grad_norm": 0.11476446551397697, "learning_rate": 1.8837728686376158e-05, "loss": 0.4176, "step": 731 }, { "epoch": 3.5529786266484766, "grad_norm": 0.12252689326129268, "learning_rate": 1.8722816436486754e-05, "loss": 0.4238, "step": 732 }, { "epoch": 3.557829316355919, "grad_norm": 0.10742367387279236, "learning_rate": 1.8608148560037036e-05, "loss": 0.4248, "step": 733 }, { "epoch": 3.562680006063362, "grad_norm": 0.11202485484827082, "learning_rate": 1.8493726374014442e-05, "loss": 0.4047, "step": 734 }, { "epoch": 3.567530695770805, "grad_norm": 0.10619927645163178, "learning_rate": 1.8379551192584588e-05, "loss": 0.4105, "step": 735 }, { "epoch": 3.5723813854782476, "grad_norm": 0.1104031551739802, "learning_rate": 1.826562432707619e-05, "loss": 0.4158, "step": 736 }, { "epoch": 3.5772320751856905, "grad_norm": 0.10920848831138874, "learning_rate": 1.8151947085965994e-05, "loss": 0.4157, "step": 737 }, { "epoch": 3.5820827648931335, "grad_norm": 0.10703865180962287, "learning_rate": 1.803852077486377e-05, "loss": 0.4144, "step": 738 }, { "epoch": 3.586933454600576, "grad_norm": 0.1043548569854331, "learning_rate": 1.7925346696497295e-05, "loss": 0.4082, "step": 739 }, { "epoch": 3.591784144308019, "grad_norm": 0.11138455070453787, "learning_rate": 1.781242615069733e-05, "loss": 0.4137, "step": 740 }, { "epoch": 3.5966348340154615, "grad_norm": 0.11820322046978973, "learning_rate": 1.7699760434382853e-05, "loss": 0.4108, "step": 741 }, { "epoch": 3.6014855237229044, "grad_norm": 0.11278678223744326, "learning_rate": 1.758735084154601e-05, "loss": 0.4189, "step": 742 }, { "epoch": 3.606336213430347, "grad_norm": 0.11323272041806605, "learning_rate": 1.7475198663237297e-05, "loss": 0.4123, "step": 743 }, { "epoch": 3.61118690313779, "grad_norm": 0.10904989882061365, "learning_rate": 1.736330518755082e-05, "loss": 0.4158, "step": 744 }, { "epoch": 3.616037592845233, "grad_norm": 0.11229772775053595, "learning_rate": 1.7251671699609313e-05, "loss": 0.4182, "step": 745 }, { "epoch": 3.6208882825526754, "grad_norm": 0.10951334908001022, "learning_rate": 1.7140299481549557e-05, "loss": 0.4213, "step": 746 }, { "epoch": 3.6257389722601183, "grad_norm": 0.11434493342191, "learning_rate": 1.7029189812507603e-05, "loss": 0.4224, "step": 747 }, { "epoch": 3.6305896619675613, "grad_norm": 0.10344385368037658, "learning_rate": 1.6918343968604027e-05, "loss": 0.4106, "step": 748 }, { "epoch": 3.635440351675004, "grad_norm": 0.11168549168378746, "learning_rate": 1.6807763222929315e-05, "loss": 0.408, "step": 749 }, { "epoch": 3.6402910413824463, "grad_norm": 0.11594040802990377, "learning_rate": 1.669744884552926e-05, "loss": 0.4169, "step": 750 }, { "epoch": 3.6451417310898893, "grad_norm": 0.11452998735643415, "learning_rate": 1.6587402103390314e-05, "loss": 0.4162, "step": 751 }, { "epoch": 3.649992420797332, "grad_norm": 0.11033578052460767, "learning_rate": 1.6477624260425137e-05, "loss": 0.4179, "step": 752 }, { "epoch": 3.6548431105047747, "grad_norm": 0.1198668006175803, "learning_rate": 1.6368116577457973e-05, "loss": 0.4124, "step": 753 }, { "epoch": 3.6596938002122177, "grad_norm": 0.10884921001547737, "learning_rate": 1.6258880312210195e-05, "loss": 0.4152, "step": 754 }, { "epoch": 3.6645444899196606, "grad_norm": 0.12009675101549108, "learning_rate": 1.6149916719285942e-05, "loss": 0.4147, "step": 755 }, { "epoch": 3.669395179627103, "grad_norm": 0.11742205741912104, "learning_rate": 1.6041227050157607e-05, "loss": 0.4096, "step": 756 }, { "epoch": 3.674245869334546, "grad_norm": 0.1123251375365744, "learning_rate": 1.5932812553151506e-05, "loss": 0.4128, "step": 757 }, { "epoch": 3.6790965590419886, "grad_norm": 0.12009728599563303, "learning_rate": 1.582467447343355e-05, "loss": 0.4179, "step": 758 }, { "epoch": 3.6839472487494316, "grad_norm": 0.10679356522304362, "learning_rate": 1.5716814052994928e-05, "loss": 0.4153, "step": 759 }, { "epoch": 3.688797938456874, "grad_norm": 0.11287526417521035, "learning_rate": 1.5609232530637827e-05, "loss": 0.4065, "step": 760 }, { "epoch": 3.693648628164317, "grad_norm": 0.10604174975295146, "learning_rate": 1.5501931141961278e-05, "loss": 0.4135, "step": 761 }, { "epoch": 3.69849931787176, "grad_norm": 0.10895447717123838, "learning_rate": 1.539491111934686e-05, "loss": 0.4102, "step": 762 }, { "epoch": 3.7033500075792025, "grad_norm": 0.11522106117016057, "learning_rate": 1.5288173691944613e-05, "loss": 0.4193, "step": 763 }, { "epoch": 3.7082006972866455, "grad_norm": 0.11668799730823959, "learning_rate": 1.5181720085658906e-05, "loss": 0.4131, "step": 764 }, { "epoch": 3.7130513869940884, "grad_norm": 0.1179757070897769, "learning_rate": 1.5075551523134358e-05, "loss": 0.4107, "step": 765 }, { "epoch": 3.717902076701531, "grad_norm": 0.11352197320512793, "learning_rate": 1.4969669223741771e-05, "loss": 0.4093, "step": 766 }, { "epoch": 3.722752766408974, "grad_norm": 0.11945491424386492, "learning_rate": 1.4864074403564216e-05, "loss": 0.4142, "step": 767 }, { "epoch": 3.7276034561164164, "grad_norm": 0.10892285815783607, "learning_rate": 1.4758768275382887e-05, "loss": 0.4205, "step": 768 }, { "epoch": 3.7324541458238594, "grad_norm": 0.12765157069597566, "learning_rate": 1.4653752048663394e-05, "loss": 0.412, "step": 769 }, { "epoch": 3.737304835531302, "grad_norm": 0.10956587187939422, "learning_rate": 1.4549026929541693e-05, "loss": 0.4148, "step": 770 }, { "epoch": 3.742155525238745, "grad_norm": 0.12138544117757244, "learning_rate": 1.4444594120810326e-05, "loss": 0.4115, "step": 771 }, { "epoch": 3.747006214946188, "grad_norm": 0.10697996518645103, "learning_rate": 1.4340454821904573e-05, "loss": 0.4194, "step": 772 }, { "epoch": 3.7518569046536303, "grad_norm": 0.10998535306961314, "learning_rate": 1.4236610228888683e-05, "loss": 0.4143, "step": 773 }, { "epoch": 3.7567075943610733, "grad_norm": 0.10998459629638649, "learning_rate": 1.4133061534442133e-05, "loss": 0.4121, "step": 774 }, { "epoch": 3.7615582840685162, "grad_norm": 0.1056241184630642, "learning_rate": 1.4029809927845981e-05, "loss": 0.4146, "step": 775 }, { "epoch": 3.7664089737759587, "grad_norm": 0.11617390368347923, "learning_rate": 1.3926856594969115e-05, "loss": 0.4125, "step": 776 }, { "epoch": 3.7712596634834017, "grad_norm": 0.10105741678809248, "learning_rate": 1.3824202718254655e-05, "loss": 0.4081, "step": 777 }, { "epoch": 3.776110353190844, "grad_norm": 0.12303653861990478, "learning_rate": 1.3721849476706477e-05, "loss": 0.413, "step": 778 }, { "epoch": 3.780961042898287, "grad_norm": 0.10128855312982177, "learning_rate": 1.3619798045875529e-05, "loss": 0.4117, "step": 779 }, { "epoch": 3.7858117326057297, "grad_norm": 0.1118697000457396, "learning_rate": 1.3518049597846412e-05, "loss": 0.4097, "step": 780 }, { "epoch": 3.7906624223131726, "grad_norm": 0.10689143155371568, "learning_rate": 1.3416605301223893e-05, "loss": 0.4146, "step": 781 }, { "epoch": 3.7955131120206156, "grad_norm": 0.09942573069367382, "learning_rate": 1.3315466321119486e-05, "loss": 0.415, "step": 782 }, { "epoch": 3.800363801728058, "grad_norm": 0.112223020708749, "learning_rate": 1.3214633819138105e-05, "loss": 0.4187, "step": 783 }, { "epoch": 3.805214491435501, "grad_norm": 0.09509376603334437, "learning_rate": 1.3114108953364655e-05, "loss": 0.4083, "step": 784 }, { "epoch": 3.810065181142944, "grad_norm": 0.1039694903593983, "learning_rate": 1.3013892878350771e-05, "loss": 0.415, "step": 785 }, { "epoch": 3.8149158708503865, "grad_norm": 0.09799457517015436, "learning_rate": 1.2913986745101567e-05, "loss": 0.4082, "step": 786 }, { "epoch": 3.8197665605578295, "grad_norm": 0.0994744466895985, "learning_rate": 1.2814391701062392e-05, "loss": 0.416, "step": 787 }, { "epoch": 3.824617250265272, "grad_norm": 0.09937882084294014, "learning_rate": 1.2715108890105663e-05, "loss": 0.4118, "step": 788 }, { "epoch": 3.829467939972715, "grad_norm": 0.09366936781008306, "learning_rate": 1.2616139452517748e-05, "loss": 0.4202, "step": 789 }, { "epoch": 3.8343186296801575, "grad_norm": 0.10103247526201467, "learning_rate": 1.2517484524985836e-05, "loss": 0.414, "step": 790 }, { "epoch": 3.8391693193876004, "grad_norm": 0.10154281976161658, "learning_rate": 1.2419145240584856e-05, "loss": 0.4169, "step": 791 }, { "epoch": 3.8440200090950434, "grad_norm": 0.09917419975099584, "learning_rate": 1.2321122728764566e-05, "loss": 0.4121, "step": 792 }, { "epoch": 3.848870698802486, "grad_norm": 0.10067017561953691, "learning_rate": 1.222341811533648e-05, "loss": 0.4177, "step": 793 }, { "epoch": 3.853721388509929, "grad_norm": 0.10318548830281854, "learning_rate": 1.2126032522460975e-05, "loss": 0.4211, "step": 794 }, { "epoch": 3.858572078217372, "grad_norm": 0.09616344349182201, "learning_rate": 1.2028967068634417e-05, "loss": 0.4204, "step": 795 }, { "epoch": 3.8634227679248143, "grad_norm": 0.10489574786705688, "learning_rate": 1.193222286867628e-05, "loss": 0.4119, "step": 796 }, { "epoch": 3.868273457632257, "grad_norm": 0.1008569871750799, "learning_rate": 1.1835801033716372e-05, "loss": 0.4086, "step": 797 }, { "epoch": 3.8731241473397, "grad_norm": 0.09122317803788128, "learning_rate": 1.1739702671182083e-05, "loss": 0.4214, "step": 798 }, { "epoch": 3.8779748370471427, "grad_norm": 0.0991904412555539, "learning_rate": 1.1643928884785618e-05, "loss": 0.4095, "step": 799 }, { "epoch": 3.8828255267545853, "grad_norm": 0.0959491624589887, "learning_rate": 1.1548480774511353e-05, "loss": 0.4218, "step": 800 }, { "epoch": 3.887676216462028, "grad_norm": 0.09541302821779325, "learning_rate": 1.1453359436603213e-05, "loss": 0.4218, "step": 801 }, { "epoch": 3.892526906169471, "grad_norm": 0.09054297930670004, "learning_rate": 1.1358565963552039e-05, "loss": 0.421, "step": 802 }, { "epoch": 3.8973775958769137, "grad_norm": 0.09175434223796135, "learning_rate": 1.126410144408312e-05, "loss": 0.4088, "step": 803 }, { "epoch": 3.9022282855843566, "grad_norm": 0.0933968969265613, "learning_rate": 1.1169966963143568e-05, "loss": 0.4105, "step": 804 }, { "epoch": 3.907078975291799, "grad_norm": 0.09827723734859882, "learning_rate": 1.1076163601889953e-05, "loss": 0.4114, "step": 805 }, { "epoch": 3.911929664999242, "grad_norm": 0.09185288445531155, "learning_rate": 1.098269243767589e-05, "loss": 0.4099, "step": 806 }, { "epoch": 3.9167803547066846, "grad_norm": 0.10344811490009215, "learning_rate": 1.0889554544039593e-05, "loss": 0.417, "step": 807 }, { "epoch": 3.9216310444141276, "grad_norm": 0.09584223413813131, "learning_rate": 1.0796750990691596e-05, "loss": 0.4092, "step": 808 }, { "epoch": 3.9264817341215705, "grad_norm": 0.10225202685968936, "learning_rate": 1.0704282843502459e-05, "loss": 0.4156, "step": 809 }, { "epoch": 3.931332423829013, "grad_norm": 0.10087206295646214, "learning_rate": 1.0612151164490525e-05, "loss": 0.4209, "step": 810 }, { "epoch": 3.936183113536456, "grad_norm": 0.09495839512641314, "learning_rate": 1.0520357011809707e-05, "loss": 0.4193, "step": 811 }, { "epoch": 3.941033803243899, "grad_norm": 0.10115287872844174, "learning_rate": 1.0428901439737387e-05, "loss": 0.415, "step": 812 }, { "epoch": 3.9458844929513415, "grad_norm": 0.09538434619439141, "learning_rate": 1.0337785498662223e-05, "loss": 0.4152, "step": 813 }, { "epoch": 3.9507351826587844, "grad_norm": 0.0955931932252973, "learning_rate": 1.024701023507216e-05, "loss": 0.4153, "step": 814 }, { "epoch": 3.955585872366227, "grad_norm": 0.10045702643945939, "learning_rate": 1.015657669154237e-05, "loss": 0.4156, "step": 815 }, { "epoch": 3.96043656207367, "grad_norm": 0.09335473216661304, "learning_rate": 1.00664859067233e-05, "loss": 0.4109, "step": 816 }, { "epoch": 3.9652872517811124, "grad_norm": 0.09645722262367523, "learning_rate": 9.976738915328719e-06, "loss": 0.4107, "step": 817 }, { "epoch": 3.9701379414885554, "grad_norm": 0.09446841411423582, "learning_rate": 9.887336748123864e-06, "loss": 0.4178, "step": 818 }, { "epoch": 3.9749886311959983, "grad_norm": 0.0903235693705833, "learning_rate": 9.798280431913558e-06, "loss": 0.4202, "step": 819 }, { "epoch": 3.979839320903441, "grad_norm": 0.0948091272844266, "learning_rate": 9.709570989530493e-06, "loss": 0.4123, "step": 820 }, { "epoch": 3.984690010610884, "grad_norm": 0.10103242493534337, "learning_rate": 9.621209439823388e-06, "loss": 0.4132, "step": 821 }, { "epoch": 3.9895407003183267, "grad_norm": 0.09199808648298305, "learning_rate": 9.533196797645354e-06, "loss": 0.4101, "step": 822 }, { "epoch": 3.9943913900257693, "grad_norm": 0.09866422487619428, "learning_rate": 9.44553407384221e-06, "loss": 0.412, "step": 823 }, { "epoch": 3.999242079733212, "grad_norm": 0.09950215009808663, "learning_rate": 9.358222275240884e-06, "loss": 0.4113, "step": 824 }, { "epoch": 4.0048506897074425, "grad_norm": 0.16227668213643984, "learning_rate": 9.271262404637835e-06, "loss": 0.4032, "step": 825 }, { "epoch": 4.009701379414886, "grad_norm": 0.11430363283166092, "learning_rate": 9.184655460787591e-06, "loss": 0.3988, "step": 826 }, { "epoch": 4.014552069122328, "grad_norm": 0.11751760908320912, "learning_rate": 9.098402438391161e-06, "loss": 0.3943, "step": 827 }, { "epoch": 4.019402758829771, "grad_norm": 0.13225951402360853, "learning_rate": 9.012504328084724e-06, "loss": 0.4024, "step": 828 }, { "epoch": 4.0242534485372135, "grad_norm": 0.1269725744791709, "learning_rate": 8.926962116428228e-06, "loss": 0.4, "step": 829 }, { "epoch": 4.029104138244657, "grad_norm": 0.12436207204015083, "learning_rate": 8.841776785894014e-06, "loss": 0.3994, "step": 830 }, { "epoch": 4.033954827952099, "grad_norm": 0.12882961013897004, "learning_rate": 8.756949314855565e-06, "loss": 0.3977, "step": 831 }, { "epoch": 4.038805517659542, "grad_norm": 0.11685196489455994, "learning_rate": 8.672480677576267e-06, "loss": 0.3906, "step": 832 }, { "epoch": 4.043656207366985, "grad_norm": 0.106568471380297, "learning_rate": 8.58837184419821e-06, "loss": 0.388, "step": 833 }, { "epoch": 4.048506897074428, "grad_norm": 0.1134401036269532, "learning_rate": 8.504623780731056e-06, "loss": 0.3918, "step": 834 }, { "epoch": 4.05335758678187, "grad_norm": 0.12093411095557142, "learning_rate": 8.421237449040962e-06, "loss": 0.397, "step": 835 }, { "epoch": 4.058208276489314, "grad_norm": 0.11674197768223235, "learning_rate": 8.338213806839453e-06, "loss": 0.393, "step": 836 }, { "epoch": 4.063058966196756, "grad_norm": 0.10082248897884766, "learning_rate": 8.255553807672547e-06, "loss": 0.3897, "step": 837 }, { "epoch": 4.067909655904199, "grad_norm": 0.11069126863357161, "learning_rate": 8.1732584009097e-06, "loss": 0.3882, "step": 838 }, { "epoch": 4.072760345611641, "grad_norm": 0.11738423041846735, "learning_rate": 8.091328531732925e-06, "loss": 0.3959, "step": 839 }, { "epoch": 4.077611035319085, "grad_norm": 0.10387395387726432, "learning_rate": 8.009765141126014e-06, "loss": 0.3891, "step": 840 }, { "epoch": 4.082461725026527, "grad_norm": 0.10003982048853301, "learning_rate": 7.928569165863584e-06, "loss": 0.3909, "step": 841 }, { "epoch": 4.08731241473397, "grad_norm": 0.10549262492820612, "learning_rate": 7.847741538500439e-06, "loss": 0.3875, "step": 842 }, { "epoch": 4.092163104441413, "grad_norm": 0.10986857014715488, "learning_rate": 7.767283187360846e-06, "loss": 0.3929, "step": 843 }, { "epoch": 4.097013794148856, "grad_norm": 0.09674503227632227, "learning_rate": 7.687195036527813e-06, "loss": 0.3928, "step": 844 }, { "epoch": 4.101864483856298, "grad_norm": 0.10305617355184506, "learning_rate": 7.60747800583252e-06, "loss": 0.4002, "step": 845 }, { "epoch": 4.1067151735637415, "grad_norm": 0.09835145173574618, "learning_rate": 7.52813301084375e-06, "loss": 0.3961, "step": 846 }, { "epoch": 4.111565863271184, "grad_norm": 0.09426867245890704, "learning_rate": 7.449160962857358e-06, "loss": 0.3946, "step": 847 }, { "epoch": 4.1164165529786265, "grad_norm": 0.0993123791236732, "learning_rate": 7.370562768885823e-06, "loss": 0.3937, "step": 848 }, { "epoch": 4.121267242686069, "grad_norm": 0.09692548506896993, "learning_rate": 7.292339331647848e-06, "loss": 0.3957, "step": 849 }, { "epoch": 4.126117932393512, "grad_norm": 0.0977639150148905, "learning_rate": 7.214491549557898e-06, "loss": 0.3969, "step": 850 }, { "epoch": 4.130968622100955, "grad_norm": 0.10100358672328608, "learning_rate": 7.1370203167160326e-06, "loss": 0.3937, "step": 851 }, { "epoch": 4.1358193118083975, "grad_norm": 0.0961508995875077, "learning_rate": 7.0599265228975e-06, "loss": 0.3965, "step": 852 }, { "epoch": 4.140670001515841, "grad_norm": 0.09450992963252156, "learning_rate": 6.983211053542591e-06, "loss": 0.4008, "step": 853 }, { "epoch": 4.145520691223283, "grad_norm": 0.10411768031595499, "learning_rate": 6.9068747897464535e-06, "loss": 0.4032, "step": 854 }, { "epoch": 4.150371380930726, "grad_norm": 0.09981896937364662, "learning_rate": 6.830918608248964e-06, "loss": 0.4002, "step": 855 }, { "epoch": 4.155222070638169, "grad_norm": 0.08754517232013051, "learning_rate": 6.755343381424659e-06, "loss": 0.3976, "step": 856 }, { "epoch": 4.160072760345612, "grad_norm": 0.09520671649907132, "learning_rate": 6.68014997727275e-06, "loss": 0.3921, "step": 857 }, { "epoch": 4.164923450053054, "grad_norm": 0.09719121632502135, "learning_rate": 6.605339259407104e-06, "loss": 0.3852, "step": 858 }, { "epoch": 4.169774139760497, "grad_norm": 0.08870928687089744, "learning_rate": 6.530912087046317e-06, "loss": 0.395, "step": 859 }, { "epoch": 4.17462482946794, "grad_norm": 0.08685694037296439, "learning_rate": 6.456869315003946e-06, "loss": 0.3941, "step": 860 }, { "epoch": 4.179475519175383, "grad_norm": 0.08904997313439429, "learning_rate": 6.3832117936785564e-06, "loss": 0.3997, "step": 861 }, { "epoch": 4.184326208882825, "grad_norm": 0.0910482544729166, "learning_rate": 6.309940369044047e-06, "loss": 0.3945, "step": 862 }, { "epoch": 4.189176898590269, "grad_norm": 0.0887134263309384, "learning_rate": 6.23705588263992e-06, "loss": 0.3881, "step": 863 }, { "epoch": 4.194027588297711, "grad_norm": 0.08977043998387933, "learning_rate": 6.164559171561553e-06, "loss": 0.3957, "step": 864 }, { "epoch": 4.198878278005154, "grad_norm": 0.09216637353138708, "learning_rate": 6.092451068450671e-06, "loss": 0.3969, "step": 865 }, { "epoch": 4.203728967712596, "grad_norm": 0.08464526503746563, "learning_rate": 6.020732401485751e-06, "loss": 0.3896, "step": 866 }, { "epoch": 4.20857965742004, "grad_norm": 0.08983668958241023, "learning_rate": 5.9494039943724845e-06, "loss": 0.4025, "step": 867 }, { "epoch": 4.213430347127482, "grad_norm": 0.08753368121579265, "learning_rate": 5.878466666334341e-06, "loss": 0.3954, "step": 868 }, { "epoch": 4.218281036834925, "grad_norm": 0.08687937990016806, "learning_rate": 5.80792123210316e-06, "loss": 0.391, "step": 869 }, { "epoch": 4.223131726542368, "grad_norm": 0.08701962255264487, "learning_rate": 5.737768501909773e-06, "loss": 0.3965, "step": 870 }, { "epoch": 4.2279824162498105, "grad_norm": 0.08811499746627446, "learning_rate": 5.668009281474751e-06, "loss": 0.3846, "step": 871 }, { "epoch": 4.232833105957253, "grad_norm": 0.08561633996061273, "learning_rate": 5.598644371999085e-06, "loss": 0.3919, "step": 872 }, { "epoch": 4.237683795664696, "grad_norm": 0.08321419945310787, "learning_rate": 5.5296745701549906e-06, "loss": 0.394, "step": 873 }, { "epoch": 4.242534485372139, "grad_norm": 0.08724643182123827, "learning_rate": 5.4611006680768305e-06, "loss": 0.3994, "step": 874 }, { "epoch": 4.2473851750795815, "grad_norm": 0.08361301460174729, "learning_rate": 5.3929234533519345e-06, "loss": 0.388, "step": 875 }, { "epoch": 4.252235864787025, "grad_norm": 0.0862576560444418, "learning_rate": 5.325143709011587e-06, "loss": 0.3942, "step": 876 }, { "epoch": 4.257086554494467, "grad_norm": 0.082606577377262, "learning_rate": 5.257762213522055e-06, "loss": 0.3907, "step": 877 }, { "epoch": 4.26193724420191, "grad_norm": 0.08601520095290771, "learning_rate": 5.19077974077558e-06, "loss": 0.3927, "step": 878 }, { "epoch": 4.266787933909352, "grad_norm": 0.08144847498379731, "learning_rate": 5.124197060081564e-06, "loss": 0.3981, "step": 879 }, { "epoch": 4.271638623616796, "grad_norm": 0.08369648223405692, "learning_rate": 5.058014936157714e-06, "loss": 0.3987, "step": 880 }, { "epoch": 4.276489313324238, "grad_norm": 0.08487450387282452, "learning_rate": 4.992234129121225e-06, "loss": 0.3993, "step": 881 }, { "epoch": 4.281340003031681, "grad_norm": 0.0835928002976722, "learning_rate": 4.926855394480079e-06, "loss": 0.394, "step": 882 }, { "epoch": 4.286190692739124, "grad_norm": 0.0877663910781713, "learning_rate": 4.861879483124372e-06, "loss": 0.3888, "step": 883 }, { "epoch": 4.291041382446567, "grad_norm": 0.08201902566316537, "learning_rate": 4.797307141317666e-06, "loss": 0.3978, "step": 884 }, { "epoch": 4.295892072154009, "grad_norm": 0.08252391090570044, "learning_rate": 4.7331391106884364e-06, "loss": 0.3949, "step": 885 }, { "epoch": 4.300742761861452, "grad_norm": 0.08121494056766716, "learning_rate": 4.6693761282215766e-06, "loss": 0.3922, "step": 886 }, { "epoch": 4.305593451568895, "grad_norm": 0.08414265476669228, "learning_rate": 4.606018926249851e-06, "loss": 0.389, "step": 887 }, { "epoch": 4.310444141276338, "grad_norm": 0.08528892500629966, "learning_rate": 4.543068232445596e-06, "loss": 0.3956, "step": 888 }, { "epoch": 4.31529483098378, "grad_norm": 0.08280199210655267, "learning_rate": 4.480524769812276e-06, "loss": 0.3938, "step": 889 }, { "epoch": 4.320145520691224, "grad_norm": 0.08884076809121294, "learning_rate": 4.418389256676206e-06, "loss": 0.3947, "step": 890 }, { "epoch": 4.324996210398666, "grad_norm": 0.08902762031211295, "learning_rate": 4.35666240667834e-06, "loss": 0.3907, "step": 891 }, { "epoch": 4.329846900106109, "grad_norm": 0.08464410835355753, "learning_rate": 4.295344928765999e-06, "loss": 0.3939, "step": 892 }, { "epoch": 4.334697589813552, "grad_norm": 0.08347421805827179, "learning_rate": 4.234437527184785e-06, "loss": 0.3985, "step": 893 }, { "epoch": 4.3395482795209945, "grad_norm": 0.08770395898608876, "learning_rate": 4.173940901470488e-06, "loss": 0.395, "step": 894 }, { "epoch": 4.344398969228437, "grad_norm": 0.08139356257791996, "learning_rate": 4.11385574644104e-06, "loss": 0.39, "step": 895 }, { "epoch": 4.34924965893588, "grad_norm": 0.0818047074224124, "learning_rate": 4.054182752188501e-06, "loss": 0.3948, "step": 896 }, { "epoch": 4.354100348643323, "grad_norm": 0.08370905382784635, "learning_rate": 3.994922604071217e-06, "loss": 0.3941, "step": 897 }, { "epoch": 4.3589510383507655, "grad_norm": 0.08543140680473625, "learning_rate": 3.936075982705871e-06, "loss": 0.3982, "step": 898 }, { "epoch": 4.363801728058208, "grad_norm": 0.08330546263387466, "learning_rate": 3.877643563959694e-06, "loss": 0.3986, "step": 899 }, { "epoch": 4.368652417765651, "grad_norm": 0.08181359063571672, "learning_rate": 3.819626018942732e-06, "loss": 0.3962, "step": 900 }, { "epoch": 4.373503107473094, "grad_norm": 0.08458157848919085, "learning_rate": 3.762024014000054e-06, "loss": 0.3974, "step": 901 }, { "epoch": 4.378353797180536, "grad_norm": 0.0797245693813596, "learning_rate": 3.7048382107042113e-06, "loss": 0.3849, "step": 902 }, { "epoch": 4.38320448688798, "grad_norm": 0.08552325047075819, "learning_rate": 3.6480692658475446e-06, "loss": 0.3908, "step": 903 }, { "epoch": 4.388055176595422, "grad_norm": 0.08191614537028945, "learning_rate": 3.5917178314346955e-06, "loss": 0.398, "step": 904 }, { "epoch": 4.392905866302865, "grad_norm": 0.08324472646577967, "learning_rate": 3.535784554675088e-06, "loss": 0.3941, "step": 905 }, { "epoch": 4.397756556010307, "grad_norm": 0.08386476178169076, "learning_rate": 3.480270077975525e-06, "loss": 0.395, "step": 906 }, { "epoch": 4.402607245717751, "grad_norm": 0.08759720428686872, "learning_rate": 3.42517503893276e-06, "loss": 0.3879, "step": 907 }, { "epoch": 4.407457935425193, "grad_norm": 0.0834695533465509, "learning_rate": 3.370500070326257e-06, "loss": 0.3832, "step": 908 }, { "epoch": 4.412308625132636, "grad_norm": 0.080547419306159, "learning_rate": 3.3162458001108332e-06, "loss": 0.3858, "step": 909 }, { "epoch": 4.417159314840079, "grad_norm": 0.08406690651169581, "learning_rate": 3.2624128514094778e-06, "loss": 0.3923, "step": 910 }, { "epoch": 4.422010004547522, "grad_norm": 0.08417275645743631, "learning_rate": 3.20900184250625e-06, "loss": 0.3933, "step": 911 }, { "epoch": 4.426860694254964, "grad_norm": 0.08104238775596306, "learning_rate": 3.1560133868390895e-06, "loss": 0.4023, "step": 912 }, { "epoch": 4.431711383962407, "grad_norm": 0.08376361425082632, "learning_rate": 3.1034480929928333e-06, "loss": 0.399, "step": 913 }, { "epoch": 4.43656207366985, "grad_norm": 0.08058350379685782, "learning_rate": 3.0513065646921957e-06, "loss": 0.3946, "step": 914 }, { "epoch": 4.441412763377293, "grad_norm": 0.08013760556674378, "learning_rate": 2.999589400794851e-06, "loss": 0.392, "step": 915 }, { "epoch": 4.446263453084735, "grad_norm": 0.08082216615145961, "learning_rate": 2.948297195284546e-06, "loss": 0.3916, "step": 916 }, { "epoch": 4.4511141427921785, "grad_norm": 0.09007717527915819, "learning_rate": 2.897430537264283e-06, "loss": 0.3947, "step": 917 }, { "epoch": 4.455964832499621, "grad_norm": 0.0866390993026153, "learning_rate": 2.8469900109495553e-06, "loss": 0.3942, "step": 918 }, { "epoch": 4.460815522207064, "grad_norm": 0.08034663815912857, "learning_rate": 2.79697619566162e-06, "loss": 0.3888, "step": 919 }, { "epoch": 4.465666211914507, "grad_norm": 0.08423465438674188, "learning_rate": 2.7473896658208743e-06, "loss": 0.391, "step": 920 }, { "epoch": 4.4705169016219495, "grad_norm": 0.08769412223500794, "learning_rate": 2.6982309909402293e-06, "loss": 0.3936, "step": 921 }, { "epoch": 4.475367591329392, "grad_norm": 0.0794092004599977, "learning_rate": 2.649500735618582e-06, "loss": 0.3993, "step": 922 }, { "epoch": 4.480218281036835, "grad_norm": 0.08321318361659834, "learning_rate": 2.6011994595343516e-06, "loss": 0.3965, "step": 923 }, { "epoch": 4.485068970744278, "grad_norm": 0.08400144661884966, "learning_rate": 2.5533277174389916e-06, "loss": 0.3927, "step": 924 }, { "epoch": 4.48991966045172, "grad_norm": 0.08039773750467258, "learning_rate": 2.5058860591506973e-06, "loss": 0.3927, "step": 925 }, { "epoch": 4.494770350159163, "grad_norm": 0.08132175209524008, "learning_rate": 2.4588750295480246e-06, "loss": 0.3888, "step": 926 }, { "epoch": 4.499621039866606, "grad_norm": 0.07920562392368859, "learning_rate": 2.4122951685636674e-06, "loss": 0.3896, "step": 927 }, { "epoch": 4.504471729574049, "grad_norm": 0.07833571502263627, "learning_rate": 2.366147011178246e-06, "loss": 0.398, "step": 928 }, { "epoch": 4.509322419281491, "grad_norm": 0.08232920047142565, "learning_rate": 2.320431087414159e-06, "loss": 0.3838, "step": 929 }, { "epoch": 4.514173108988935, "grad_norm": 0.08136833742803433, "learning_rate": 2.275147922329506e-06, "loss": 0.3935, "step": 930 }, { "epoch": 4.519023798696377, "grad_norm": 0.08383396678636225, "learning_rate": 2.230298036012055e-06, "loss": 0.3913, "step": 931 }, { "epoch": 4.52387448840382, "grad_norm": 0.08000391439211, "learning_rate": 2.1858819435732583e-06, "loss": 0.395, "step": 932 }, { "epoch": 4.528725178111262, "grad_norm": 0.080147749223563, "learning_rate": 2.141900155142351e-06, "loss": 0.3951, "step": 933 }, { "epoch": 4.533575867818706, "grad_norm": 0.07978655829531454, "learning_rate": 2.0983531758604726e-06, "loss": 0.4011, "step": 934 }, { "epoch": 4.538426557526148, "grad_norm": 0.08386421814730027, "learning_rate": 2.055241505874892e-06, "loss": 0.3999, "step": 935 }, { "epoch": 4.543277247233591, "grad_norm": 0.08174904946083562, "learning_rate": 2.0125656403332396e-06, "loss": 0.3968, "step": 936 }, { "epoch": 4.548127936941034, "grad_norm": 0.07927025601942647, "learning_rate": 1.970326069377828e-06, "loss": 0.4001, "step": 937 }, { "epoch": 4.552978626648477, "grad_norm": 0.0783638905343862, "learning_rate": 1.928523278140033e-06, "loss": 0.3943, "step": 938 }, { "epoch": 4.557829316355919, "grad_norm": 0.08012821453349714, "learning_rate": 1.887157746734718e-06, "loss": 0.3972, "step": 939 }, { "epoch": 4.562680006063362, "grad_norm": 0.08022389886437123, "learning_rate": 1.846229950254692e-06, "loss": 0.3888, "step": 940 }, { "epoch": 4.567530695770805, "grad_norm": 0.07938938767749412, "learning_rate": 1.8057403587652977e-06, "loss": 0.3892, "step": 941 }, { "epoch": 4.572381385478248, "grad_norm": 0.07979264917062728, "learning_rate": 1.7656894372989785e-06, "loss": 0.3991, "step": 942 }, { "epoch": 4.57723207518569, "grad_norm": 0.07900416574598519, "learning_rate": 1.726077645849955e-06, "loss": 0.4033, "step": 943 }, { "epoch": 4.5820827648931335, "grad_norm": 0.0816244017270053, "learning_rate": 1.6869054393689265e-06, "loss": 0.394, "step": 944 }, { "epoch": 4.586933454600576, "grad_norm": 0.07678354405433441, "learning_rate": 1.6481732677578798e-06, "loss": 0.4026, "step": 945 }, { "epoch": 4.5917841443080185, "grad_norm": 0.07913436605637802, "learning_rate": 1.60988157586488e-06, "loss": 0.378, "step": 946 }, { "epoch": 4.596634834015462, "grad_norm": 0.0793297266554538, "learning_rate": 1.5720308034789721e-06, "loss": 0.391, "step": 947 }, { "epoch": 4.601485523722904, "grad_norm": 0.07767980761029898, "learning_rate": 1.5346213853251546e-06, "loss": 0.3978, "step": 948 }, { "epoch": 4.606336213430347, "grad_norm": 0.08494507410525762, "learning_rate": 1.4976537510593646e-06, "loss": 0.3995, "step": 949 }, { "epoch": 4.61118690313779, "grad_norm": 0.08237577571806884, "learning_rate": 1.4611283252635412e-06, "loss": 0.4038, "step": 950 }, { "epoch": 4.616037592845233, "grad_norm": 0.07590194142960284, "learning_rate": 1.425045527440756e-06, "loss": 0.3956, "step": 951 }, { "epoch": 4.620888282552675, "grad_norm": 0.07954982455169565, "learning_rate": 1.3894057720104104e-06, "loss": 0.399, "step": 952 }, { "epoch": 4.625738972260118, "grad_norm": 0.07731849776688487, "learning_rate": 1.354209468303429e-06, "loss": 0.3828, "step": 953 }, { "epoch": 4.630589661967561, "grad_norm": 0.07970299041030604, "learning_rate": 1.3194570205576284e-06, "loss": 0.3954, "step": 954 }, { "epoch": 4.635440351675004, "grad_norm": 0.07874124868840192, "learning_rate": 1.2851488279130053e-06, "loss": 0.3876, "step": 955 }, { "epoch": 4.640291041382446, "grad_norm": 0.07691612719760402, "learning_rate": 1.2512852844071933e-06, "loss": 0.3949, "step": 956 }, { "epoch": 4.64514173108989, "grad_norm": 0.08126016720695953, "learning_rate": 1.2178667789709287e-06, "loss": 0.3919, "step": 957 }, { "epoch": 4.649992420797332, "grad_norm": 0.08002643782323322, "learning_rate": 1.1848936954235702e-06, "loss": 0.395, "step": 958 }, { "epoch": 4.654843110504775, "grad_norm": 0.08026194265591635, "learning_rate": 1.1523664124687284e-06, "loss": 0.3997, "step": 959 }, { "epoch": 4.659693800212217, "grad_norm": 0.07660352848559206, "learning_rate": 1.1202853036898476e-06, "loss": 0.3974, "step": 960 }, { "epoch": 4.664544489919661, "grad_norm": 0.08113236962163348, "learning_rate": 1.0886507375459908e-06, "loss": 0.3981, "step": 961 }, { "epoch": 4.669395179627103, "grad_norm": 0.07701309756204706, "learning_rate": 1.0574630773675687e-06, "loss": 0.3839, "step": 962 }, { "epoch": 4.674245869334546, "grad_norm": 0.07891978078549244, "learning_rate": 1.0267226813521635e-06, "loss": 0.3877, "step": 963 }, { "epoch": 4.679096559041989, "grad_norm": 0.07852321868608765, "learning_rate": 9.964299025604274e-07, "loss": 0.3921, "step": 964 }, { "epoch": 4.683947248749432, "grad_norm": 0.07818107376466682, "learning_rate": 9.66585088912022e-07, "loss": 0.3967, "step": 965 }, { "epoch": 4.688797938456874, "grad_norm": 0.076600770628096, "learning_rate": 9.371885831816319e-07, "loss": 0.395, "step": 966 }, { "epoch": 4.6936486281643175, "grad_norm": 0.07627250485843899, "learning_rate": 9.082407229950018e-07, "loss": 0.3976, "step": 967 }, { "epoch": 4.69849931787176, "grad_norm": 0.07905168555654667, "learning_rate": 8.797418408251101e-07, "loss": 0.3918, "step": 968 }, { "epoch": 4.7033500075792025, "grad_norm": 0.07625038396328788, "learning_rate": 8.516922639882819e-07, "loss": 0.3897, "step": 969 }, { "epoch": 4.708200697286646, "grad_norm": 0.08055085830124646, "learning_rate": 8.2409231464049e-07, "loss": 0.3901, "step": 970 }, { "epoch": 4.713051386994088, "grad_norm": 0.07599467665769233, "learning_rate": 7.969423097736162e-07, "loss": 0.3931, "step": 971 }, { "epoch": 4.717902076701531, "grad_norm": 0.08049982038815076, "learning_rate": 7.702425612118269e-07, "loss": 0.3962, "step": 972 }, { "epoch": 4.7227527664089735, "grad_norm": 0.07817439712988589, "learning_rate": 7.439933756079942e-07, "loss": 0.3913, "step": 973 }, { "epoch": 4.727603456116417, "grad_norm": 0.08001209525722262, "learning_rate": 7.181950544401695e-07, "loss": 0.3935, "step": 974 }, { "epoch": 4.732454145823859, "grad_norm": 0.07544329449164154, "learning_rate": 6.928478940081107e-07, "loss": 0.3984, "step": 975 }, { "epoch": 4.737304835531302, "grad_norm": 0.07792056514995602, "learning_rate": 6.679521854299032e-07, "loss": 0.3914, "step": 976 }, { "epoch": 4.742155525238745, "grad_norm": 0.07777159770613211, "learning_rate": 6.435082146385885e-07, "loss": 0.3901, "step": 977 }, { "epoch": 4.747006214946188, "grad_norm": 0.07684966968716175, "learning_rate": 6.195162623789052e-07, "loss": 0.393, "step": 978 }, { "epoch": 4.75185690465363, "grad_norm": 0.07663626226452194, "learning_rate": 5.959766042040426e-07, "loss": 0.3879, "step": 979 }, { "epoch": 4.756707594361073, "grad_norm": 0.07653692258850123, "learning_rate": 5.728895104724963e-07, "loss": 0.3959, "step": 980 }, { "epoch": 4.761558284068516, "grad_norm": 0.07731590848782262, "learning_rate": 5.502552463449418e-07, "loss": 0.3984, "step": 981 }, { "epoch": 4.766408973775959, "grad_norm": 0.07715742201671594, "learning_rate": 5.280740717812149e-07, "loss": 0.3977, "step": 982 }, { "epoch": 4.771259663483401, "grad_norm": 0.07733908081214975, "learning_rate": 5.063462415372967e-07, "loss": 0.3948, "step": 983 }, { "epoch": 4.776110353190845, "grad_norm": 0.07797117300936077, "learning_rate": 4.850720051624124e-07, "loss": 0.3914, "step": 984 }, { "epoch": 4.780961042898287, "grad_norm": 0.07721734718928677, "learning_rate": 4.642516069961556e-07, "loss": 0.392, "step": 985 }, { "epoch": 4.78581173260573, "grad_norm": 0.07648551851063208, "learning_rate": 4.438852861656751e-07, "loss": 0.3951, "step": 986 }, { "epoch": 4.790662422313172, "grad_norm": 0.07521119037010907, "learning_rate": 4.2397327658294076e-07, "loss": 0.3899, "step": 987 }, { "epoch": 4.795513112020616, "grad_norm": 0.07655879860272995, "learning_rate": 4.045158069420474e-07, "loss": 0.3963, "step": 988 }, { "epoch": 4.800363801728058, "grad_norm": 0.07622760847546149, "learning_rate": 3.8551310071659023e-07, "loss": 0.3975, "step": 989 }, { "epoch": 4.805214491435501, "grad_norm": 0.0750428400819878, "learning_rate": 3.6696537615711124e-07, "loss": 0.3968, "step": 990 }, { "epoch": 4.810065181142944, "grad_norm": 0.07852243689578567, "learning_rate": 3.4887284628857266e-07, "loss": 0.3932, "step": 991 }, { "epoch": 4.8149158708503865, "grad_norm": 0.0771348195411035, "learning_rate": 3.3123571890791405e-07, "loss": 0.3887, "step": 992 }, { "epoch": 4.819766560557829, "grad_norm": 0.07614331045752255, "learning_rate": 3.1405419658168125e-07, "loss": 0.394, "step": 993 }, { "epoch": 4.824617250265272, "grad_norm": 0.07563944264490313, "learning_rate": 2.973284766436857e-07, "loss": 0.3917, "step": 994 }, { "epoch": 4.829467939972715, "grad_norm": 0.07503164664083585, "learning_rate": 2.810587511927354e-07, "loss": 0.3901, "step": 995 }, { "epoch": 4.8343186296801575, "grad_norm": 0.07663595734772072, "learning_rate": 2.652452070904499e-07, "loss": 0.3923, "step": 996 }, { "epoch": 4.839169319387601, "grad_norm": 0.07606961706547127, "learning_rate": 2.498880259590797e-07, "loss": 0.3944, "step": 997 }, { "epoch": 4.844020009095043, "grad_norm": 0.0761182148197196, "learning_rate": 2.3498738417945034e-07, "loss": 0.3975, "step": 998 }, { "epoch": 4.848870698802486, "grad_norm": 0.08039104985634041, "learning_rate": 2.205434528889283e-07, "loss": 0.3971, "step": 999 }, { "epoch": 4.853721388509928, "grad_norm": 0.07775606427920397, "learning_rate": 2.0655639797944937e-07, "loss": 0.3903, "step": 1000 }, { "epoch": 4.858572078217372, "grad_norm": 0.07610940999590161, "learning_rate": 1.9302638009561782e-07, "loss": 0.396, "step": 1001 }, { "epoch": 4.863422767924814, "grad_norm": 0.07569450912275462, "learning_rate": 1.7995355463285457e-07, "loss": 0.3965, "step": 1002 }, { "epoch": 4.868273457632257, "grad_norm": 0.07643720956773195, "learning_rate": 1.6733807173562988e-07, "loss": 0.3913, "step": 1003 }, { "epoch": 4.8731241473397, "grad_norm": 0.07571887859535044, "learning_rate": 1.5518007629571342e-07, "loss": 0.3995, "step": 1004 }, { "epoch": 4.877974837047143, "grad_norm": 0.07657090064859913, "learning_rate": 1.4347970795054456e-07, "loss": 0.3967, "step": 1005 }, { "epoch": 4.882825526754585, "grad_norm": 0.07475277823538463, "learning_rate": 1.3223710108158483e-07, "loss": 0.3965, "step": 1006 }, { "epoch": 4.887676216462028, "grad_norm": 0.07465351831333361, "learning_rate": 1.214523848128124e-07, "loss": 0.3971, "step": 1007 }, { "epoch": 4.892526906169471, "grad_norm": 0.07637388169909817, "learning_rate": 1.111256830092211e-07, "loss": 0.3996, "step": 1008 }, { "epoch": 4.897377595876914, "grad_norm": 0.07639665552172381, "learning_rate": 1.0125711427540374e-07, "loss": 0.3949, "step": 1009 }, { "epoch": 4.902228285584356, "grad_norm": 0.07373092012101537, "learning_rate": 9.184679195417989e-08, "loss": 0.3889, "step": 1010 }, { "epoch": 4.9070789752918, "grad_norm": 0.07583698163385665, "learning_rate": 8.289482412531246e-08, "loss": 0.3984, "step": 1011 }, { "epoch": 4.911929664999242, "grad_norm": 0.07647787343235872, "learning_rate": 7.440131360424652e-08, "loss": 0.3887, "step": 1012 }, { "epoch": 4.916780354706685, "grad_norm": 0.07751142464425709, "learning_rate": 6.636635794094126e-08, "loss": 0.3908, "step": 1013 }, { "epoch": 4.921631044414128, "grad_norm": 0.07884353160777417, "learning_rate": 5.879004941874655e-08, "loss": 0.3951, "step": 1014 }, { "epoch": 4.9264817341215705, "grad_norm": 0.07661585400143268, "learning_rate": 5.16724750533415e-08, "loss": 0.398, "step": 1015 }, { "epoch": 4.931332423829013, "grad_norm": 0.07506936516416932, "learning_rate": 4.5013716591730815e-08, "loss": 0.3944, "step": 1016 }, { "epoch": 4.9361831135364564, "grad_norm": 0.07625120541429038, "learning_rate": 3.881385051132114e-08, "loss": 0.3997, "step": 1017 }, { "epoch": 4.941033803243899, "grad_norm": 0.0754548388472603, "learning_rate": 3.307294801902838e-08, "loss": 0.3968, "step": 1018 }, { "epoch": 4.9458844929513415, "grad_norm": 0.07563871272528049, "learning_rate": 2.7791075050460636e-08, "loss": 0.3977, "step": 1019 }, { "epoch": 4.950735182658784, "grad_norm": 0.0760975342061979, "learning_rate": 2.2968292269167637e-08, "loss": 0.3978, "step": 1020 }, { "epoch": 4.955585872366227, "grad_norm": 0.07484229156933196, "learning_rate": 1.8604655065939116e-08, "loss": 0.3888, "step": 1021 }, { "epoch": 4.96043656207367, "grad_norm": 0.07489129192172529, "learning_rate": 1.470021355816975e-08, "loss": 0.3952, "step": 1022 }, { "epoch": 4.965287251781112, "grad_norm": 0.07561234415729748, "learning_rate": 1.1255012589286297e-08, "loss": 0.3988, "step": 1023 }, { "epoch": 4.970137941488556, "grad_norm": 0.07589712959511802, "learning_rate": 8.269091728232426e-09, "loss": 0.3858, "step": 1024 }, { "epoch": 4.974988631195998, "grad_norm": 0.07684966203710471, "learning_rate": 5.742485269006892e-09, "loss": 0.3976, "step": 1025 }, { "epoch": 4.979839320903441, "grad_norm": 0.07450707523151534, "learning_rate": 3.6752222302727238e-09, "loss": 0.3922, "step": 1026 }, { "epoch": 4.984690010610883, "grad_norm": 0.0754634711281639, "learning_rate": 2.06732635503748e-09, "loss": 0.3894, "step": 1027 }, { "epoch": 4.989540700318327, "grad_norm": 0.07460842756020492, "learning_rate": 9.188161103557136e-10, "loss": 0.3893, "step": 1028 }, { "epoch": 4.994391390025769, "grad_norm": 0.07411367814575705, "learning_rate": 2.2970468714245132e-10, "loss": 0.3934, "step": 1029 }, { "epoch": 4.999242079733212, "grad_norm": 0.07533391242414708, "learning_rate": 0.0, "loss": 0.3877, "step": 1030 }, { "epoch": 4.999242079733212, "step": 1030, "total_flos": 2.739131934768418e+19, "train_loss": 0.07882811409755817, "train_runtime": 48155.7887, "train_samples_per_second": 10.958, "train_steps_per_second": 0.021 } ], "logging_steps": 1, "max_steps": 1030, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.739131934768418e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }