diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7252 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.999242079733212, + "eval_steps": 500, + "global_step": 1030, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.004850689707442777, + "grad_norm": 6.339065858846749, + "learning_rate": 7.766990291262136e-07, + "loss": 1.0147, + "step": 1 + }, + { + "epoch": 0.009701379414885554, + "grad_norm": 6.370901874406726, + "learning_rate": 1.5533980582524272e-06, + "loss": 1.0217, + "step": 2 + }, + { + "epoch": 0.01455206912232833, + "grad_norm": 6.253076167725343, + "learning_rate": 2.330097087378641e-06, + "loss": 1.0059, + "step": 3 + }, + { + "epoch": 0.01940275882977111, + "grad_norm": 5.8627331903052005, + "learning_rate": 3.1067961165048544e-06, + "loss": 0.9987, + "step": 4 + }, + { + "epoch": 0.024253448537213885, + "grad_norm": 4.624724616975588, + "learning_rate": 3.883495145631068e-06, + "loss": 0.9654, + "step": 5 + }, + { + "epoch": 0.02910413824465666, + "grad_norm": 2.701973440331149, + "learning_rate": 4.660194174757282e-06, + "loss": 0.9221, + "step": 6 + }, + { + "epoch": 0.03395482795209944, + "grad_norm": 2.5821400220833683, + "learning_rate": 5.436893203883496e-06, + "loss": 0.9118, + "step": 7 + }, + { + "epoch": 0.03880551765954222, + "grad_norm": 3.1473551148693146, + "learning_rate": 6.213592233009709e-06, + "loss": 0.8818, + "step": 8 + }, + { + "epoch": 0.04365620736698499, + "grad_norm": 3.8230150062051638, + "learning_rate": 6.990291262135923e-06, + "loss": 0.8912, + "step": 9 + }, + { + "epoch": 0.04850689707442777, + "grad_norm": 3.4669236063777715, + "learning_rate": 7.766990291262136e-06, + "loss": 0.8779, + "step": 10 + }, + { + "epoch": 0.053357586781870546, + "grad_norm": 2.712055876575345, + "learning_rate": 8.54368932038835e-06, + "loss": 0.8328, + "step": 11 + }, + { + "epoch": 0.05820827648931332, + "grad_norm": 2.574263966000136, + "learning_rate": 9.320388349514565e-06, + "loss": 0.8149, + "step": 12 + }, + { + "epoch": 0.0630589661967561, + "grad_norm": 1.8016385817876701, + "learning_rate": 1.0097087378640778e-05, + "loss": 0.7941, + "step": 13 + }, + { + "epoch": 0.06790965590419888, + "grad_norm": 1.16110696712433, + "learning_rate": 1.0873786407766991e-05, + "loss": 0.7751, + "step": 14 + }, + { + "epoch": 0.07276034561164166, + "grad_norm": 1.5441894400920566, + "learning_rate": 1.1650485436893204e-05, + "loss": 0.7603, + "step": 15 + }, + { + "epoch": 0.07761103531908443, + "grad_norm": 1.3570174190036193, + "learning_rate": 1.2427184466019418e-05, + "loss": 0.7456, + "step": 16 + }, + { + "epoch": 0.08246172502652721, + "grad_norm": 1.0160879152766609, + "learning_rate": 1.3203883495145633e-05, + "loss": 0.74, + "step": 17 + }, + { + "epoch": 0.08731241473396999, + "grad_norm": 1.195923679791525, + "learning_rate": 1.3980582524271846e-05, + "loss": 0.7223, + "step": 18 + }, + { + "epoch": 0.09216310444141276, + "grad_norm": 1.0381307779091873, + "learning_rate": 1.475728155339806e-05, + "loss": 0.7149, + "step": 19 + }, + { + "epoch": 0.09701379414885554, + "grad_norm": 0.9191697728302082, + "learning_rate": 1.5533980582524273e-05, + "loss": 0.7032, + "step": 20 + }, + { + "epoch": 0.10186448385629832, + "grad_norm": 1.0389109685950821, + "learning_rate": 1.6310679611650486e-05, + "loss": 0.697, + "step": 21 + }, + { + "epoch": 0.10671517356374109, + "grad_norm": 0.8528569833940303, + "learning_rate": 1.70873786407767e-05, + "loss": 0.6913, + "step": 22 + }, + { + "epoch": 0.11156586327118387, + "grad_norm": 0.7397574673832126, + "learning_rate": 1.7864077669902916e-05, + "loss": 0.6844, + "step": 23 + }, + { + "epoch": 0.11641655297862664, + "grad_norm": 0.6762376097915315, + "learning_rate": 1.864077669902913e-05, + "loss": 0.6807, + "step": 24 + }, + { + "epoch": 0.12126724268606942, + "grad_norm": 0.6801312007046909, + "learning_rate": 1.9417475728155343e-05, + "loss": 0.6651, + "step": 25 + }, + { + "epoch": 0.1261179323935122, + "grad_norm": 0.5373129321939298, + "learning_rate": 2.0194174757281556e-05, + "loss": 0.6661, + "step": 26 + }, + { + "epoch": 0.13096862210095497, + "grad_norm": 0.6155691125010336, + "learning_rate": 2.097087378640777e-05, + "loss": 0.6655, + "step": 27 + }, + { + "epoch": 0.13581931180839776, + "grad_norm": 0.5373412410981904, + "learning_rate": 2.1747572815533982e-05, + "loss": 0.6541, + "step": 28 + }, + { + "epoch": 0.14067000151584053, + "grad_norm": 0.5722094683121568, + "learning_rate": 2.2524271844660196e-05, + "loss": 0.6534, + "step": 29 + }, + { + "epoch": 0.14552069122328332, + "grad_norm": 0.6369873796903149, + "learning_rate": 2.330097087378641e-05, + "loss": 0.6536, + "step": 30 + }, + { + "epoch": 0.15037138093072608, + "grad_norm": 0.5246684440675834, + "learning_rate": 2.4077669902912622e-05, + "loss": 0.6545, + "step": 31 + }, + { + "epoch": 0.15522207063816887, + "grad_norm": 0.504247506683658, + "learning_rate": 2.4854368932038836e-05, + "loss": 0.6314, + "step": 32 + }, + { + "epoch": 0.16007276034561163, + "grad_norm": 0.5566944063536889, + "learning_rate": 2.5631067961165052e-05, + "loss": 0.6373, + "step": 33 + }, + { + "epoch": 0.16492345005305442, + "grad_norm": 0.8570989167580252, + "learning_rate": 2.6407766990291266e-05, + "loss": 0.63, + "step": 34 + }, + { + "epoch": 0.16977413976049718, + "grad_norm": 1.8667995731915865, + "learning_rate": 2.718446601941748e-05, + "loss": 0.6456, + "step": 35 + }, + { + "epoch": 0.17462482946793997, + "grad_norm": 0.5164194413667431, + "learning_rate": 2.7961165048543692e-05, + "loss": 0.6308, + "step": 36 + }, + { + "epoch": 0.17947551917538274, + "grad_norm": 1.6098330943831782, + "learning_rate": 2.8737864077669905e-05, + "loss": 0.6398, + "step": 37 + }, + { + "epoch": 0.18432620888282553, + "grad_norm": 1.1492129920694993, + "learning_rate": 2.951456310679612e-05, + "loss": 0.6276, + "step": 38 + }, + { + "epoch": 0.18917689859026832, + "grad_norm": 0.7127614761088336, + "learning_rate": 3.0291262135922332e-05, + "loss": 0.6218, + "step": 39 + }, + { + "epoch": 0.19402758829771108, + "grad_norm": 1.0312378150228299, + "learning_rate": 3.1067961165048545e-05, + "loss": 0.6253, + "step": 40 + }, + { + "epoch": 0.19887827800515387, + "grad_norm": 1.7574387168351864, + "learning_rate": 3.184466019417476e-05, + "loss": 0.6281, + "step": 41 + }, + { + "epoch": 0.20372896771259663, + "grad_norm": 0.9416599187328968, + "learning_rate": 3.262135922330097e-05, + "loss": 0.6235, + "step": 42 + }, + { + "epoch": 0.20857965742003942, + "grad_norm": 1.8301787236679616, + "learning_rate": 3.339805825242719e-05, + "loss": 0.622, + "step": 43 + }, + { + "epoch": 0.21343034712748218, + "grad_norm": 0.9292078210446757, + "learning_rate": 3.41747572815534e-05, + "loss": 0.6107, + "step": 44 + }, + { + "epoch": 0.21828103683492497, + "grad_norm": 1.994214362456412, + "learning_rate": 3.4951456310679615e-05, + "loss": 0.6151, + "step": 45 + }, + { + "epoch": 0.22313172654236774, + "grad_norm": 1.3619344540131681, + "learning_rate": 3.572815533980583e-05, + "loss": 0.6089, + "step": 46 + }, + { + "epoch": 0.22798241624981053, + "grad_norm": 1.6323821476629805, + "learning_rate": 3.650485436893204e-05, + "loss": 0.6165, + "step": 47 + }, + { + "epoch": 0.2328331059572533, + "grad_norm": 1.1474300438640261, + "learning_rate": 3.728155339805826e-05, + "loss": 0.6104, + "step": 48 + }, + { + "epoch": 0.23768379566469608, + "grad_norm": 1.1936059623728144, + "learning_rate": 3.805825242718447e-05, + "loss": 0.6086, + "step": 49 + }, + { + "epoch": 0.24253448537213884, + "grad_norm": 1.4126137333521573, + "learning_rate": 3.8834951456310685e-05, + "loss": 0.6123, + "step": 50 + }, + { + "epoch": 0.24738517507958163, + "grad_norm": 0.7835607330331523, + "learning_rate": 3.9611650485436895e-05, + "loss": 0.5979, + "step": 51 + }, + { + "epoch": 0.2522358647870244, + "grad_norm": 1.3575257469759314, + "learning_rate": 4.038834951456311e-05, + "loss": 0.6134, + "step": 52 + }, + { + "epoch": 0.25708655449446716, + "grad_norm": 0.9580505107131282, + "learning_rate": 4.116504854368932e-05, + "loss": 0.5987, + "step": 53 + }, + { + "epoch": 0.26193724420190995, + "grad_norm": 1.1647956860260527, + "learning_rate": 4.194174757281554e-05, + "loss": 0.6134, + "step": 54 + }, + { + "epoch": 0.26678793390935274, + "grad_norm": 0.9133078407230598, + "learning_rate": 4.271844660194175e-05, + "loss": 0.5978, + "step": 55 + }, + { + "epoch": 0.2716386236167955, + "grad_norm": 1.3505836368180404, + "learning_rate": 4.3495145631067965e-05, + "loss": 0.5944, + "step": 56 + }, + { + "epoch": 0.2764893133242383, + "grad_norm": 1.1628512424723636, + "learning_rate": 4.4271844660194175e-05, + "loss": 0.6052, + "step": 57 + }, + { + "epoch": 0.28134000303168105, + "grad_norm": 1.6421048813051027, + "learning_rate": 4.504854368932039e-05, + "loss": 0.6008, + "step": 58 + }, + { + "epoch": 0.28619069273912384, + "grad_norm": 1.3019802198524983, + "learning_rate": 4.58252427184466e-05, + "loss": 0.594, + "step": 59 + }, + { + "epoch": 0.29104138244656663, + "grad_norm": 1.2774402772060065, + "learning_rate": 4.660194174757282e-05, + "loss": 0.5892, + "step": 60 + }, + { + "epoch": 0.2958920721540094, + "grad_norm": 1.4435670386305743, + "learning_rate": 4.737864077669903e-05, + "loss": 0.5931, + "step": 61 + }, + { + "epoch": 0.30074276186145216, + "grad_norm": 1.0284857540916943, + "learning_rate": 4.8155339805825245e-05, + "loss": 0.5932, + "step": 62 + }, + { + "epoch": 0.30559345156889495, + "grad_norm": 1.1698103896183938, + "learning_rate": 4.8932038834951454e-05, + "loss": 0.5948, + "step": 63 + }, + { + "epoch": 0.31044414127633774, + "grad_norm": 1.434142725222452, + "learning_rate": 4.970873786407767e-05, + "loss": 0.5897, + "step": 64 + }, + { + "epoch": 0.3152948309837805, + "grad_norm": 1.3482179068151203, + "learning_rate": 5.0485436893203895e-05, + "loss": 0.5824, + "step": 65 + }, + { + "epoch": 0.32014552069122326, + "grad_norm": 1.3069227496554443, + "learning_rate": 5.1262135922330105e-05, + "loss": 0.5802, + "step": 66 + }, + { + "epoch": 0.32499621039866605, + "grad_norm": 1.1366236056516827, + "learning_rate": 5.203883495145632e-05, + "loss": 0.5804, + "step": 67 + }, + { + "epoch": 0.32984690010610884, + "grad_norm": 1.6243350603336242, + "learning_rate": 5.281553398058253e-05, + "loss": 0.5812, + "step": 68 + }, + { + "epoch": 0.33469758981355163, + "grad_norm": 1.0822457196908746, + "learning_rate": 5.359223300970875e-05, + "loss": 0.5862, + "step": 69 + }, + { + "epoch": 0.33954827952099437, + "grad_norm": 1.6360957248140573, + "learning_rate": 5.436893203883496e-05, + "loss": 0.5912, + "step": 70 + }, + { + "epoch": 0.34439896922843716, + "grad_norm": 2.3477845490488813, + "learning_rate": 5.5145631067961174e-05, + "loss": 0.5906, + "step": 71 + }, + { + "epoch": 0.34924965893587995, + "grad_norm": 1.105543560593242, + "learning_rate": 5.5922330097087384e-05, + "loss": 0.5824, + "step": 72 + }, + { + "epoch": 0.35410034864332274, + "grad_norm": 4.137098681881185, + "learning_rate": 5.66990291262136e-05, + "loss": 0.6359, + "step": 73 + }, + { + "epoch": 0.3589510383507655, + "grad_norm": 3.8847979837997033, + "learning_rate": 5.747572815533981e-05, + "loss": 0.6486, + "step": 74 + }, + { + "epoch": 0.36380172805820826, + "grad_norm": 1.1747128429519862, + "learning_rate": 5.825242718446603e-05, + "loss": 0.595, + "step": 75 + }, + { + "epoch": 0.36865241776565105, + "grad_norm": 3.008245900701061, + "learning_rate": 5.902912621359224e-05, + "loss": 0.6387, + "step": 76 + }, + { + "epoch": 0.37350310747309384, + "grad_norm": 2.3713075132931554, + "learning_rate": 5.9805825242718454e-05, + "loss": 0.6344, + "step": 77 + }, + { + "epoch": 0.37835379718053663, + "grad_norm": 1.6213053074921984, + "learning_rate": 6.0582524271844664e-05, + "loss": 0.6048, + "step": 78 + }, + { + "epoch": 0.38320448688797937, + "grad_norm": 1.6024154837501339, + "learning_rate": 6.135922330097087e-05, + "loss": 0.6203, + "step": 79 + }, + { + "epoch": 0.38805517659542216, + "grad_norm": 1.1825030677591377, + "learning_rate": 6.213592233009709e-05, + "loss": 0.6052, + "step": 80 + }, + { + "epoch": 0.39290586630286495, + "grad_norm": 1.3964525731128163, + "learning_rate": 6.291262135922331e-05, + "loss": 0.6112, + "step": 81 + }, + { + "epoch": 0.39775655601030774, + "grad_norm": 1.1205074621871551, + "learning_rate": 6.368932038834952e-05, + "loss": 0.5977, + "step": 82 + }, + { + "epoch": 0.4026072457177505, + "grad_norm": 1.045620374565707, + "learning_rate": 6.446601941747573e-05, + "loss": 0.5914, + "step": 83 + }, + { + "epoch": 0.40745793542519326, + "grad_norm": 1.3974586249408472, + "learning_rate": 6.524271844660194e-05, + "loss": 0.5918, + "step": 84 + }, + { + "epoch": 0.41230862513263605, + "grad_norm": 1.0818483302602913, + "learning_rate": 6.601941747572816e-05, + "loss": 0.5948, + "step": 85 + }, + { + "epoch": 0.41715931484007884, + "grad_norm": 0.9808456957793906, + "learning_rate": 6.679611650485438e-05, + "loss": 0.5839, + "step": 86 + }, + { + "epoch": 0.4220100045475216, + "grad_norm": 1.2035779456517084, + "learning_rate": 6.757281553398058e-05, + "loss": 0.5833, + "step": 87 + }, + { + "epoch": 0.42686069425496437, + "grad_norm": 1.6887623926979713, + "learning_rate": 6.83495145631068e-05, + "loss": 0.5818, + "step": 88 + }, + { + "epoch": 0.43171138396240716, + "grad_norm": 0.8023218391013366, + "learning_rate": 6.912621359223301e-05, + "loss": 0.5863, + "step": 89 + }, + { + "epoch": 0.43656207366984995, + "grad_norm": 1.574900348178855, + "learning_rate": 6.990291262135923e-05, + "loss": 0.5821, + "step": 90 + }, + { + "epoch": 0.4414127633772927, + "grad_norm": 0.9288518542917786, + "learning_rate": 7.067961165048545e-05, + "loss": 0.5814, + "step": 91 + }, + { + "epoch": 0.4462634530847355, + "grad_norm": 1.6871845487045471, + "learning_rate": 7.145631067961166e-05, + "loss": 0.5819, + "step": 92 + }, + { + "epoch": 0.45111414279217826, + "grad_norm": 1.147490028185953, + "learning_rate": 7.223300970873787e-05, + "loss": 0.5752, + "step": 93 + }, + { + "epoch": 0.45596483249962105, + "grad_norm": 1.0734179177901382, + "learning_rate": 7.300970873786408e-05, + "loss": 0.5786, + "step": 94 + }, + { + "epoch": 0.46081552220706384, + "grad_norm": 1.2263367009960806, + "learning_rate": 7.37864077669903e-05, + "loss": 0.5789, + "step": 95 + }, + { + "epoch": 0.4656662119145066, + "grad_norm": 1.4570032389620742, + "learning_rate": 7.456310679611652e-05, + "loss": 0.5745, + "step": 96 + }, + { + "epoch": 0.47051690162194937, + "grad_norm": 1.3246870726440927, + "learning_rate": 7.533980582524272e-05, + "loss": 0.5775, + "step": 97 + }, + { + "epoch": 0.47536759132939216, + "grad_norm": 0.9415857506868542, + "learning_rate": 7.611650485436894e-05, + "loss": 0.5699, + "step": 98 + }, + { + "epoch": 0.48021828103683495, + "grad_norm": 1.2384384474151087, + "learning_rate": 7.689320388349515e-05, + "loss": 0.5733, + "step": 99 + }, + { + "epoch": 0.4850689707442777, + "grad_norm": 1.5627749991572353, + "learning_rate": 7.766990291262137e-05, + "loss": 0.5735, + "step": 100 + }, + { + "epoch": 0.4899196604517205, + "grad_norm": 1.0078484211944914, + "learning_rate": 7.844660194174757e-05, + "loss": 0.5733, + "step": 101 + }, + { + "epoch": 0.49477035015916326, + "grad_norm": 1.6421211712488573, + "learning_rate": 7.922330097087379e-05, + "loss": 0.576, + "step": 102 + }, + { + "epoch": 0.49962103986660605, + "grad_norm": 0.8416126904816602, + "learning_rate": 8e-05, + "loss": 0.5697, + "step": 103 + }, + { + "epoch": 0.5044717295740488, + "grad_norm": 1.7540458195736903, + "learning_rate": 7.999977029531286e-05, + "loss": 0.5799, + "step": 104 + }, + { + "epoch": 0.5093224192814916, + "grad_norm": 1.050447231844734, + "learning_rate": 7.999908118388965e-05, + "loss": 0.5756, + "step": 105 + }, + { + "epoch": 0.5141731089889343, + "grad_norm": 1.394330877960354, + "learning_rate": 7.999793267364497e-05, + "loss": 0.5713, + "step": 106 + }, + { + "epoch": 0.5190237986963772, + "grad_norm": 1.5873678777006228, + "learning_rate": 7.999632477776974e-05, + "loss": 0.5733, + "step": 107 + }, + { + "epoch": 0.5238744884038199, + "grad_norm": 0.8115732123781836, + "learning_rate": 7.9994257514731e-05, + "loss": 0.5661, + "step": 108 + }, + { + "epoch": 0.5287251781112627, + "grad_norm": 1.1857754936411384, + "learning_rate": 7.999173090827177e-05, + "loss": 0.5719, + "step": 109 + }, + { + "epoch": 0.5335758678187055, + "grad_norm": 0.8066157800150973, + "learning_rate": 7.998874498741072e-05, + "loss": 0.5695, + "step": 110 + }, + { + "epoch": 0.5384265575261482, + "grad_norm": 1.4739854945603235, + "learning_rate": 7.998529978644183e-05, + "loss": 0.5712, + "step": 111 + }, + { + "epoch": 0.543277247233591, + "grad_norm": 0.778875512416499, + "learning_rate": 7.998139534493407e-05, + "loss": 0.5609, + "step": 112 + }, + { + "epoch": 0.5481279369410338, + "grad_norm": 0.9919152170469479, + "learning_rate": 7.997703170773084e-05, + "loss": 0.5648, + "step": 113 + }, + { + "epoch": 0.5529786266484766, + "grad_norm": 1.3093959851041357, + "learning_rate": 7.997220892494955e-05, + "loss": 0.5757, + "step": 114 + }, + { + "epoch": 0.5578293163559194, + "grad_norm": 0.9066117120369992, + "learning_rate": 7.996692705198097e-05, + "loss": 0.566, + "step": 115 + }, + { + "epoch": 0.5626800060633621, + "grad_norm": 1.2498381901490132, + "learning_rate": 7.996118614948869e-05, + "loss": 0.5757, + "step": 116 + }, + { + "epoch": 0.567530695770805, + "grad_norm": 1.0385189418340641, + "learning_rate": 7.995498628340827e-05, + "loss": 0.5697, + "step": 117 + }, + { + "epoch": 0.5723813854782477, + "grad_norm": 1.3065564110858372, + "learning_rate": 7.994832752494667e-05, + "loss": 0.5672, + "step": 118 + }, + { + "epoch": 0.5772320751856904, + "grad_norm": 0.9524061899396458, + "learning_rate": 7.994120995058127e-05, + "loss": 0.5624, + "step": 119 + }, + { + "epoch": 0.5820827648931333, + "grad_norm": 1.018979775317657, + "learning_rate": 7.993363364205907e-05, + "loss": 0.554, + "step": 120 + }, + { + "epoch": 0.586933454600576, + "grad_norm": 0.9276898951496916, + "learning_rate": 7.992559868639576e-05, + "loss": 0.556, + "step": 121 + }, + { + "epoch": 0.5917841443080188, + "grad_norm": 1.1930439188526805, + "learning_rate": 7.99171051758747e-05, + "loss": 0.5526, + "step": 122 + }, + { + "epoch": 0.5966348340154616, + "grad_norm": 0.6826863389197481, + "learning_rate": 7.990815320804583e-05, + "loss": 0.5609, + "step": 123 + }, + { + "epoch": 0.6014855237229043, + "grad_norm": 0.8061980858751947, + "learning_rate": 7.98987428857246e-05, + "loss": 0.5586, + "step": 124 + }, + { + "epoch": 0.6063362134303472, + "grad_norm": 0.9239948149603757, + "learning_rate": 7.988887431699079e-05, + "loss": 0.5507, + "step": 125 + }, + { + "epoch": 0.6111869031377899, + "grad_norm": 1.0465957889843347, + "learning_rate": 7.987854761518719e-05, + "loss": 0.5568, + "step": 126 + }, + { + "epoch": 0.6160375928452326, + "grad_norm": 0.8308130357408615, + "learning_rate": 7.986776289891842e-05, + "loss": 0.5591, + "step": 127 + }, + { + "epoch": 0.6208882825526755, + "grad_norm": 1.0246744766037437, + "learning_rate": 7.985652029204946e-05, + "loss": 0.5563, + "step": 128 + }, + { + "epoch": 0.6257389722601182, + "grad_norm": 1.3238612418839921, + "learning_rate": 7.984481992370429e-05, + "loss": 0.5491, + "step": 129 + }, + { + "epoch": 0.630589661967561, + "grad_norm": 0.6948258976249133, + "learning_rate": 7.983266192826437e-05, + "loss": 0.5418, + "step": 130 + }, + { + "epoch": 0.6354403516750038, + "grad_norm": 0.7288790333090353, + "learning_rate": 7.982004644536716e-05, + "loss": 0.5441, + "step": 131 + }, + { + "epoch": 0.6402910413824465, + "grad_norm": 1.0943204811023435, + "learning_rate": 7.98069736199044e-05, + "loss": 0.5493, + "step": 132 + }, + { + "epoch": 0.6451417310898894, + "grad_norm": 0.8706827515570799, + "learning_rate": 7.979344360202055e-05, + "loss": 0.5465, + "step": 133 + }, + { + "epoch": 0.6499924207973321, + "grad_norm": 0.9127380693761118, + "learning_rate": 7.977945654711108e-05, + "loss": 0.5475, + "step": 134 + }, + { + "epoch": 0.654843110504775, + "grad_norm": 0.9510188119086359, + "learning_rate": 7.976501261582056e-05, + "loss": 0.543, + "step": 135 + }, + { + "epoch": 0.6596938002122177, + "grad_norm": 0.9350726421156861, + "learning_rate": 7.975011197404092e-05, + "loss": 0.5525, + "step": 136 + }, + { + "epoch": 0.6645444899196604, + "grad_norm": 1.272258773056705, + "learning_rate": 7.973475479290956e-05, + "loss": 0.5518, + "step": 137 + }, + { + "epoch": 0.6693951796271033, + "grad_norm": 0.9765265325518906, + "learning_rate": 7.971894124880727e-05, + "loss": 0.5417, + "step": 138 + }, + { + "epoch": 0.674245869334546, + "grad_norm": 1.1917501609756302, + "learning_rate": 7.970267152335632e-05, + "loss": 0.5464, + "step": 139 + }, + { + "epoch": 0.6790965590419887, + "grad_norm": 0.6719257463868904, + "learning_rate": 7.968594580341832e-05, + "loss": 0.544, + "step": 140 + }, + { + "epoch": 0.6839472487494316, + "grad_norm": 0.7137208028607956, + "learning_rate": 7.966876428109209e-05, + "loss": 0.5351, + "step": 141 + }, + { + "epoch": 0.6887979384568743, + "grad_norm": 0.7449688014358767, + "learning_rate": 7.965112715371144e-05, + "loss": 0.5397, + "step": 142 + }, + { + "epoch": 0.6936486281643172, + "grad_norm": 0.5377305872108858, + "learning_rate": 7.96330346238429e-05, + "loss": 0.5346, + "step": 143 + }, + { + "epoch": 0.6984993178717599, + "grad_norm": 0.5856757070627496, + "learning_rate": 7.961448689928341e-05, + "loss": 0.5395, + "step": 144 + }, + { + "epoch": 0.7033500075792026, + "grad_norm": 0.6000325566823206, + "learning_rate": 7.959548419305796e-05, + "loss": 0.5447, + "step": 145 + }, + { + "epoch": 0.7082006972866455, + "grad_norm": 0.7819963257560868, + "learning_rate": 7.957602672341707e-05, + "loss": 0.5364, + "step": 146 + }, + { + "epoch": 0.7130513869940882, + "grad_norm": 1.1223449508846108, + "learning_rate": 7.955611471383433e-05, + "loss": 0.5381, + "step": 147 + }, + { + "epoch": 0.717902076701531, + "grad_norm": 1.1418981667975974, + "learning_rate": 7.953574839300385e-05, + "loss": 0.5381, + "step": 148 + }, + { + "epoch": 0.7227527664089738, + "grad_norm": 0.758286759296052, + "learning_rate": 7.95149279948376e-05, + "loss": 0.5398, + "step": 149 + }, + { + "epoch": 0.7276034561164165, + "grad_norm": 0.7637204957772546, + "learning_rate": 7.949365375846271e-05, + "loss": 0.5386, + "step": 150 + }, + { + "epoch": 0.7324541458238594, + "grad_norm": 0.6982030938329856, + "learning_rate": 7.94719259282188e-05, + "loss": 0.5328, + "step": 151 + }, + { + "epoch": 0.7373048355313021, + "grad_norm": 0.7115887055025976, + "learning_rate": 7.944974475365506e-05, + "loss": 0.5406, + "step": 152 + }, + { + "epoch": 0.7421555252387448, + "grad_norm": 0.6914212445412167, + "learning_rate": 7.94271104895275e-05, + "loss": 0.5375, + "step": 153 + }, + { + "epoch": 0.7470062149461877, + "grad_norm": 0.6376946136665823, + "learning_rate": 7.940402339579596e-05, + "loss": 0.5322, + "step": 154 + }, + { + "epoch": 0.7518569046536304, + "grad_norm": 0.662468788270689, + "learning_rate": 7.93804837376211e-05, + "loss": 0.5312, + "step": 155 + }, + { + "epoch": 0.7567075943610733, + "grad_norm": 0.8116591959883654, + "learning_rate": 7.935649178536142e-05, + "loss": 0.5362, + "step": 156 + }, + { + "epoch": 0.761558284068516, + "grad_norm": 1.2251954995336705, + "learning_rate": 7.93320478145701e-05, + "loss": 0.5454, + "step": 157 + }, + { + "epoch": 0.7664089737759587, + "grad_norm": 1.1753293382340935, + "learning_rate": 7.93071521059919e-05, + "loss": 0.5369, + "step": 158 + }, + { + "epoch": 0.7712596634834016, + "grad_norm": 0.5797209510428332, + "learning_rate": 7.928180494555983e-05, + "loss": 0.5255, + "step": 159 + }, + { + "epoch": 0.7761103531908443, + "grad_norm": 0.9260629876609666, + "learning_rate": 7.925600662439201e-05, + "loss": 0.535, + "step": 160 + }, + { + "epoch": 0.780961042898287, + "grad_norm": 1.2828815170548864, + "learning_rate": 7.922975743878817e-05, + "loss": 0.5293, + "step": 161 + }, + { + "epoch": 0.7858117326057299, + "grad_norm": 0.5171701751512903, + "learning_rate": 7.92030576902264e-05, + "loss": 0.5298, + "step": 162 + }, + { + "epoch": 0.7906624223131726, + "grad_norm": 0.9531992162184508, + "learning_rate": 7.917590768535952e-05, + "loss": 0.5315, + "step": 163 + }, + { + "epoch": 0.7955131120206155, + "grad_norm": 1.1106333536493764, + "learning_rate": 7.914830773601173e-05, + "loss": 0.5279, + "step": 164 + }, + { + "epoch": 0.8003638017280582, + "grad_norm": 0.45675301699833715, + "learning_rate": 7.912025815917489e-05, + "loss": 0.5372, + "step": 165 + }, + { + "epoch": 0.805214491435501, + "grad_norm": 0.7189143844280501, + "learning_rate": 7.909175927700499e-05, + "loss": 0.535, + "step": 166 + }, + { + "epoch": 0.8100651811429438, + "grad_norm": 0.8686823215812186, + "learning_rate": 7.906281141681839e-05, + "loss": 0.5333, + "step": 167 + }, + { + "epoch": 0.8149158708503865, + "grad_norm": 0.588440050652225, + "learning_rate": 7.903341491108798e-05, + "loss": 0.5289, + "step": 168 + }, + { + "epoch": 0.8197665605578294, + "grad_norm": 0.563873615495661, + "learning_rate": 7.900357009743958e-05, + "loss": 0.5331, + "step": 169 + }, + { + "epoch": 0.8246172502652721, + "grad_norm": 0.545022127050129, + "learning_rate": 7.897327731864784e-05, + "loss": 0.5266, + "step": 170 + }, + { + "epoch": 0.8294679399727148, + "grad_norm": 0.5592040867673563, + "learning_rate": 7.894253692263244e-05, + "loss": 0.522, + "step": 171 + }, + { + "epoch": 0.8343186296801577, + "grad_norm": 0.6324827822327501, + "learning_rate": 7.891134926245402e-05, + "loss": 0.5297, + "step": 172 + }, + { + "epoch": 0.8391693193876004, + "grad_norm": 0.8244822385641454, + "learning_rate": 7.887971469631016e-05, + "loss": 0.5319, + "step": 173 + }, + { + "epoch": 0.8440200090950432, + "grad_norm": 1.2087031591715138, + "learning_rate": 7.884763358753129e-05, + "loss": 0.5408, + "step": 174 + }, + { + "epoch": 0.848870698802486, + "grad_norm": 0.7427926592130248, + "learning_rate": 7.881510630457643e-05, + "loss": 0.5326, + "step": 175 + }, + { + "epoch": 0.8537213885099287, + "grad_norm": 0.5074413186000275, + "learning_rate": 7.878213322102908e-05, + "loss": 0.5281, + "step": 176 + }, + { + "epoch": 0.8585720782173716, + "grad_norm": 0.7191422226961306, + "learning_rate": 7.874871471559282e-05, + "loss": 0.5269, + "step": 177 + }, + { + "epoch": 0.8634227679248143, + "grad_norm": 0.7351841845171684, + "learning_rate": 7.8714851172087e-05, + "loss": 0.5326, + "step": 178 + }, + { + "epoch": 0.868273457632257, + "grad_norm": 0.661668481852005, + "learning_rate": 7.868054297944237e-05, + "loss": 0.5312, + "step": 179 + }, + { + "epoch": 0.8731241473396999, + "grad_norm": 0.6257843185205204, + "learning_rate": 7.864579053169657e-05, + "loss": 0.5265, + "step": 180 + }, + { + "epoch": 0.8779748370471426, + "grad_norm": 0.6272508186215946, + "learning_rate": 7.86105942279896e-05, + "loss": 0.5242, + "step": 181 + }, + { + "epoch": 0.8828255267545854, + "grad_norm": 0.5859524441087289, + "learning_rate": 7.857495447255925e-05, + "loss": 0.5117, + "step": 182 + }, + { + "epoch": 0.8876762164620282, + "grad_norm": 0.5314342493506371, + "learning_rate": 7.853887167473646e-05, + "loss": 0.5275, + "step": 183 + }, + { + "epoch": 0.892526906169471, + "grad_norm": 0.5894396606474899, + "learning_rate": 7.850234624894064e-05, + "loss": 0.5236, + "step": 184 + }, + { + "epoch": 0.8973775958769138, + "grad_norm": 0.7789309413273731, + "learning_rate": 7.846537861467485e-05, + "loss": 0.5269, + "step": 185 + }, + { + "epoch": 0.9022282855843565, + "grad_norm": 0.9749399305462054, + "learning_rate": 7.842796919652104e-05, + "loss": 0.5177, + "step": 186 + }, + { + "epoch": 0.9070789752917993, + "grad_norm": 0.8944031183061603, + "learning_rate": 7.839011842413514e-05, + "loss": 0.5236, + "step": 187 + }, + { + "epoch": 0.9119296649992421, + "grad_norm": 0.6072136307213187, + "learning_rate": 7.835182673224212e-05, + "loss": 0.5237, + "step": 188 + }, + { + "epoch": 0.9167803547066848, + "grad_norm": 0.44815995922956803, + "learning_rate": 7.831309456063107e-05, + "loss": 0.5193, + "step": 189 + }, + { + "epoch": 0.9216310444141277, + "grad_norm": 0.5362363058315522, + "learning_rate": 7.827392235415005e-05, + "loss": 0.5242, + "step": 190 + }, + { + "epoch": 0.9264817341215704, + "grad_norm": 0.5675209533538658, + "learning_rate": 7.823431056270103e-05, + "loss": 0.5223, + "step": 191 + }, + { + "epoch": 0.9313324238290132, + "grad_norm": 0.5087214285921436, + "learning_rate": 7.81942596412347e-05, + "loss": 0.522, + "step": 192 + }, + { + "epoch": 0.936183113536456, + "grad_norm": 0.46687413642954234, + "learning_rate": 7.815377004974532e-05, + "loss": 0.509, + "step": 193 + }, + { + "epoch": 0.9410338032438987, + "grad_norm": 0.43127006261715695, + "learning_rate": 7.811284225326529e-05, + "loss": 0.522, + "step": 194 + }, + { + "epoch": 0.9458844929513415, + "grad_norm": 0.43502529088952246, + "learning_rate": 7.807147672185996e-05, + "loss": 0.5258, + "step": 195 + }, + { + "epoch": 0.9507351826587843, + "grad_norm": 0.5247431038277521, + "learning_rate": 7.802967393062219e-05, + "loss": 0.524, + "step": 196 + }, + { + "epoch": 0.955585872366227, + "grad_norm": 0.6873590048364063, + "learning_rate": 7.798743435966676e-05, + "loss": 0.5227, + "step": 197 + }, + { + "epoch": 0.9604365620736699, + "grad_norm": 0.8978763261543932, + "learning_rate": 7.794475849412512e-05, + "loss": 0.5143, + "step": 198 + }, + { + "epoch": 0.9652872517811126, + "grad_norm": 0.8491329782406589, + "learning_rate": 7.790164682413954e-05, + "loss": 0.5186, + "step": 199 + }, + { + "epoch": 0.9701379414885554, + "grad_norm": 0.5004070361840594, + "learning_rate": 7.785809984485765e-05, + "loss": 0.5185, + "step": 200 + }, + { + "epoch": 0.9749886311959982, + "grad_norm": 0.4620691718653308, + "learning_rate": 7.781411805642675e-05, + "loss": 0.5179, + "step": 201 + }, + { + "epoch": 0.979839320903441, + "grad_norm": 0.6078145800613444, + "learning_rate": 7.776970196398795e-05, + "loss": 0.5185, + "step": 202 + }, + { + "epoch": 0.9846900106108837, + "grad_norm": 0.6744449960922595, + "learning_rate": 7.77248520776705e-05, + "loss": 0.5223, + "step": 203 + }, + { + "epoch": 0.9895407003183265, + "grad_norm": 0.769858159356013, + "learning_rate": 7.767956891258585e-05, + "loss": 0.514, + "step": 204 + }, + { + "epoch": 0.9943913900257693, + "grad_norm": 0.9703823352556248, + "learning_rate": 7.763385298882177e-05, + "loss": 0.5227, + "step": 205 + }, + { + "epoch": 0.9992420797332121, + "grad_norm": 1.0790500827883502, + "learning_rate": 7.758770483143634e-05, + "loss": 0.5161, + "step": 206 + }, + { + "epoch": 1.0048506897074427, + "grad_norm": 0.979399858326228, + "learning_rate": 7.754112497045198e-05, + "loss": 0.5126, + "step": 207 + }, + { + "epoch": 1.0097013794148855, + "grad_norm": 0.8061955074549858, + "learning_rate": 7.749411394084931e-05, + "loss": 0.5093, + "step": 208 + }, + { + "epoch": 1.0145520691223284, + "grad_norm": 0.769561002371883, + "learning_rate": 7.744667228256102e-05, + "loss": 0.5129, + "step": 209 + }, + { + "epoch": 1.0194027588297712, + "grad_norm": 0.7563794216226459, + "learning_rate": 7.739880054046567e-05, + "loss": 0.504, + "step": 210 + }, + { + "epoch": 1.024253448537214, + "grad_norm": 0.4861095042439798, + "learning_rate": 7.735049926438143e-05, + "loss": 0.5008, + "step": 211 + }, + { + "epoch": 1.0291041382446566, + "grad_norm": 0.521763707621908, + "learning_rate": 7.730176900905978e-05, + "loss": 0.5039, + "step": 212 + }, + { + "epoch": 1.0339548279520994, + "grad_norm": 0.6847654097095645, + "learning_rate": 7.725261033417914e-05, + "loss": 0.4987, + "step": 213 + }, + { + "epoch": 1.0388055176595423, + "grad_norm": 0.45617035587869154, + "learning_rate": 7.720302380433838e-05, + "loss": 0.5082, + "step": 214 + }, + { + "epoch": 1.043656207366985, + "grad_norm": 0.5136841747695677, + "learning_rate": 7.715300998905045e-05, + "loss": 0.4903, + "step": 215 + }, + { + "epoch": 1.0485068970744278, + "grad_norm": 0.5201029165395914, + "learning_rate": 7.710256946273572e-05, + "loss": 0.5061, + "step": 216 + }, + { + "epoch": 1.0533575867818705, + "grad_norm": 0.5369451831716502, + "learning_rate": 7.705170280471546e-05, + "loss": 0.4923, + "step": 217 + }, + { + "epoch": 1.0582082764893133, + "grad_norm": 0.5292650633659572, + "learning_rate": 7.700041059920516e-05, + "loss": 0.4958, + "step": 218 + }, + { + "epoch": 1.0630589661967562, + "grad_norm": 0.38347803199441816, + "learning_rate": 7.694869343530781e-05, + "loss": 0.4949, + "step": 219 + }, + { + "epoch": 1.067909655904199, + "grad_norm": 0.42205008813365347, + "learning_rate": 7.689655190700719e-05, + "loss": 0.4958, + "step": 220 + }, + { + "epoch": 1.0727603456116417, + "grad_norm": 0.43326896920779806, + "learning_rate": 7.684398661316092e-05, + "loss": 0.5034, + "step": 221 + }, + { + "epoch": 1.0776110353190844, + "grad_norm": 0.3610486041319715, + "learning_rate": 7.679099815749377e-05, + "loss": 0.508, + "step": 222 + }, + { + "epoch": 1.0824617250265272, + "grad_norm": 0.4631910613487445, + "learning_rate": 7.673758714859052e-05, + "loss": 0.5012, + "step": 223 + }, + { + "epoch": 1.08731241473397, + "grad_norm": 0.5933537887890846, + "learning_rate": 7.668375419988918e-05, + "loss": 0.5003, + "step": 224 + }, + { + "epoch": 1.0921631044414128, + "grad_norm": 0.7536616463697116, + "learning_rate": 7.662949992967375e-05, + "loss": 0.4984, + "step": 225 + }, + { + "epoch": 1.0970137941488556, + "grad_norm": 0.8908783317227219, + "learning_rate": 7.657482496106725e-05, + "loss": 0.499, + "step": 226 + }, + { + "epoch": 1.1018644838562983, + "grad_norm": 0.8949365262495667, + "learning_rate": 7.651972992202449e-05, + "loss": 0.4964, + "step": 227 + }, + { + "epoch": 1.106715173563741, + "grad_norm": 0.7666080530207662, + "learning_rate": 7.646421544532492e-05, + "loss": 0.501, + "step": 228 + }, + { + "epoch": 1.1115658632711838, + "grad_norm": 0.5974548399149405, + "learning_rate": 7.640828216856532e-05, + "loss": 0.5019, + "step": 229 + }, + { + "epoch": 1.1164165529786267, + "grad_norm": 0.5208576739553936, + "learning_rate": 7.635193073415246e-05, + "loss": 0.4954, + "step": 230 + }, + { + "epoch": 1.1212672426860695, + "grad_norm": 0.4449218126901089, + "learning_rate": 7.62951617892958e-05, + "loss": 0.4966, + "step": 231 + }, + { + "epoch": 1.1261179323935122, + "grad_norm": 0.3413546355163314, + "learning_rate": 7.623797598599995e-05, + "loss": 0.4869, + "step": 232 + }, + { + "epoch": 1.130968622100955, + "grad_norm": 0.4021494906350838, + "learning_rate": 7.618037398105728e-05, + "loss": 0.4876, + "step": 233 + }, + { + "epoch": 1.1358193118083977, + "grad_norm": 0.5401093810240136, + "learning_rate": 7.612235643604031e-05, + "loss": 0.495, + "step": 234 + }, + { + "epoch": 1.1406700015158404, + "grad_norm": 0.6450843987922678, + "learning_rate": 7.606392401729415e-05, + "loss": 0.4953, + "step": 235 + }, + { + "epoch": 1.1455206912232834, + "grad_norm": 0.6413460171528268, + "learning_rate": 7.600507739592879e-05, + "loss": 0.4972, + "step": 236 + }, + { + "epoch": 1.150371380930726, + "grad_norm": 0.6276158556341717, + "learning_rate": 7.594581724781152e-05, + "loss": 0.4957, + "step": 237 + }, + { + "epoch": 1.1552220706381688, + "grad_norm": 0.7553018024510589, + "learning_rate": 7.588614425355898e-05, + "loss": 0.4955, + "step": 238 + }, + { + "epoch": 1.1600727603456116, + "grad_norm": 0.8968837770832118, + "learning_rate": 7.582605909852951e-05, + "loss": 0.4937, + "step": 239 + }, + { + "epoch": 1.1649234500530543, + "grad_norm": 0.7618870978199445, + "learning_rate": 7.576556247281522e-05, + "loss": 0.4969, + "step": 240 + }, + { + "epoch": 1.1697741397604973, + "grad_norm": 0.6531717186517569, + "learning_rate": 7.570465507123401e-05, + "loss": 0.5042, + "step": 241 + }, + { + "epoch": 1.17462482946794, + "grad_norm": 0.591515239054212, + "learning_rate": 7.564333759332167e-05, + "loss": 0.4906, + "step": 242 + }, + { + "epoch": 1.1794755191753827, + "grad_norm": 0.49797581923871925, + "learning_rate": 7.558161074332379e-05, + "loss": 0.4966, + "step": 243 + }, + { + "epoch": 1.1843262088828255, + "grad_norm": 0.5865910810734263, + "learning_rate": 7.551947523018774e-05, + "loss": 0.4997, + "step": 244 + }, + { + "epoch": 1.1891768985902682, + "grad_norm": 0.5003536808662635, + "learning_rate": 7.54569317675544e-05, + "loss": 0.4954, + "step": 245 + }, + { + "epoch": 1.1940275882977112, + "grad_norm": 0.4328856052217075, + "learning_rate": 7.539398107375015e-05, + "loss": 0.4979, + "step": 246 + }, + { + "epoch": 1.198878278005154, + "grad_norm": 0.5623661241326378, + "learning_rate": 7.533062387177843e-05, + "loss": 0.4982, + "step": 247 + }, + { + "epoch": 1.2037289677125966, + "grad_norm": 0.542587238982675, + "learning_rate": 7.526686088931156e-05, + "loss": 0.4991, + "step": 248 + }, + { + "epoch": 1.2085796574200394, + "grad_norm": 0.4949553127282243, + "learning_rate": 7.520269285868235e-05, + "loss": 0.4908, + "step": 249 + }, + { + "epoch": 1.213430347127482, + "grad_norm": 0.39917468694971237, + "learning_rate": 7.513812051687564e-05, + "loss": 0.4917, + "step": 250 + }, + { + "epoch": 1.218281036834925, + "grad_norm": 0.44831825611716425, + "learning_rate": 7.507314460551993e-05, + "loss": 0.4898, + "step": 251 + }, + { + "epoch": 1.2231317265423678, + "grad_norm": 0.4694728642442923, + "learning_rate": 7.500776587087878e-05, + "loss": 0.4929, + "step": 252 + }, + { + "epoch": 1.2279824162498105, + "grad_norm": 0.4548469944052975, + "learning_rate": 7.494198506384229e-05, + "loss": 0.4826, + "step": 253 + }, + { + "epoch": 1.2328331059572533, + "grad_norm": 0.6666754202357293, + "learning_rate": 7.487580293991844e-05, + "loss": 0.5021, + "step": 254 + }, + { + "epoch": 1.237683795664696, + "grad_norm": 0.6791995766586792, + "learning_rate": 7.480922025922443e-05, + "loss": 0.4974, + "step": 255 + }, + { + "epoch": 1.242534485372139, + "grad_norm": 0.48795487898431833, + "learning_rate": 7.474223778647796e-05, + "loss": 0.4934, + "step": 256 + }, + { + "epoch": 1.2473851750795817, + "grad_norm": 0.32007742081193336, + "learning_rate": 7.467485629098842e-05, + "loss": 0.4907, + "step": 257 + }, + { + "epoch": 1.2522358647870244, + "grad_norm": 0.2948779105297471, + "learning_rate": 7.460707654664807e-05, + "loss": 0.4974, + "step": 258 + }, + { + "epoch": 1.2570865544944672, + "grad_norm": 0.2954711132741566, + "learning_rate": 7.453889933192316e-05, + "loss": 0.4893, + "step": 259 + }, + { + "epoch": 1.26193724420191, + "grad_norm": 0.3234156565448516, + "learning_rate": 7.447032542984502e-05, + "loss": 0.4882, + "step": 260 + }, + { + "epoch": 1.2667879339093528, + "grad_norm": 0.37736907562809757, + "learning_rate": 7.440135562800093e-05, + "loss": 0.487, + "step": 261 + }, + { + "epoch": 1.2716386236167956, + "grad_norm": 0.4380964102542513, + "learning_rate": 7.433199071852526e-05, + "loss": 0.4965, + "step": 262 + }, + { + "epoch": 1.2764893133242383, + "grad_norm": 0.5377640268616285, + "learning_rate": 7.426223149809023e-05, + "loss": 0.4922, + "step": 263 + }, + { + "epoch": 1.281340003031681, + "grad_norm": 0.5447192325063156, + "learning_rate": 7.419207876789685e-05, + "loss": 0.4844, + "step": 264 + }, + { + "epoch": 1.2861906927391238, + "grad_norm": 0.5569387883590096, + "learning_rate": 7.412153333366567e-05, + "loss": 0.4887, + "step": 265 + }, + { + "epoch": 1.2910413824465667, + "grad_norm": 0.6127559785319273, + "learning_rate": 7.405059600562751e-05, + "loss": 0.4974, + "step": 266 + }, + { + "epoch": 1.2958920721540095, + "grad_norm": 0.5372861059475705, + "learning_rate": 7.397926759851425e-05, + "loss": 0.4946, + "step": 267 + }, + { + "epoch": 1.3007427618614522, + "grad_norm": 0.3715441206816088, + "learning_rate": 7.390754893154933e-05, + "loss": 0.4914, + "step": 268 + }, + { + "epoch": 1.305593451568895, + "grad_norm": 0.3468457624853246, + "learning_rate": 7.383544082843846e-05, + "loss": 0.4906, + "step": 269 + }, + { + "epoch": 1.3104441412763377, + "grad_norm": 0.37917435116867076, + "learning_rate": 7.376294411736009e-05, + "loss": 0.4877, + "step": 270 + }, + { + "epoch": 1.3152948309837806, + "grad_norm": 0.44133896758740837, + "learning_rate": 7.369005963095596e-05, + "loss": 0.4962, + "step": 271 + }, + { + "epoch": 1.3201455206912232, + "grad_norm": 0.5453886771483799, + "learning_rate": 7.361678820632145e-05, + "loss": 0.4918, + "step": 272 + }, + { + "epoch": 1.324996210398666, + "grad_norm": 0.6036321305780495, + "learning_rate": 7.354313068499607e-05, + "loss": 0.4892, + "step": 273 + }, + { + "epoch": 1.3298469001061088, + "grad_norm": 0.620838391242342, + "learning_rate": 7.346908791295369e-05, + "loss": 0.495, + "step": 274 + }, + { + "epoch": 1.3346975898135516, + "grad_norm": 0.5848284090531766, + "learning_rate": 7.339466074059292e-05, + "loss": 0.4862, + "step": 275 + }, + { + "epoch": 1.3395482795209943, + "grad_norm": 0.5915616847361488, + "learning_rate": 7.331985002272726e-05, + "loss": 0.4986, + "step": 276 + }, + { + "epoch": 1.344398969228437, + "grad_norm": 0.5871009859314463, + "learning_rate": 7.324465661857534e-05, + "loss": 0.4868, + "step": 277 + }, + { + "epoch": 1.34924965893588, + "grad_norm": 0.4325881770640798, + "learning_rate": 7.316908139175105e-05, + "loss": 0.4886, + "step": 278 + }, + { + "epoch": 1.3541003486433227, + "grad_norm": 0.342133338929163, + "learning_rate": 7.309312521025356e-05, + "loss": 0.4909, + "step": 279 + }, + { + "epoch": 1.3589510383507655, + "grad_norm": 0.4415991996544984, + "learning_rate": 7.301678894645742e-05, + "loss": 0.4915, + "step": 280 + }, + { + "epoch": 1.3638017280582082, + "grad_norm": 0.510599069072842, + "learning_rate": 7.294007347710251e-05, + "loss": 0.4935, + "step": 281 + }, + { + "epoch": 1.368652417765651, + "grad_norm": 0.5220646005978747, + "learning_rate": 7.286297968328397e-05, + "loss": 0.4834, + "step": 282 + }, + { + "epoch": 1.373503107473094, + "grad_norm": 0.4836929320122624, + "learning_rate": 7.27855084504421e-05, + "loss": 0.4953, + "step": 283 + }, + { + "epoch": 1.3783537971805366, + "grad_norm": 0.4678184328008887, + "learning_rate": 7.270766066835217e-05, + "loss": 0.4872, + "step": 284 + }, + { + "epoch": 1.3832044868879794, + "grad_norm": 0.4479819209137706, + "learning_rate": 7.262943723111419e-05, + "loss": 0.4916, + "step": 285 + }, + { + "epoch": 1.388055176595422, + "grad_norm": 0.44661460126650243, + "learning_rate": 7.255083903714266e-05, + "loss": 0.4866, + "step": 286 + }, + { + "epoch": 1.3929058663028648, + "grad_norm": 0.46700839086853646, + "learning_rate": 7.247186698915625e-05, + "loss": 0.4879, + "step": 287 + }, + { + "epoch": 1.3977565560103078, + "grad_norm": 0.5111356334340853, + "learning_rate": 7.239252199416749e-05, + "loss": 0.4812, + "step": 288 + }, + { + "epoch": 1.4026072457177505, + "grad_norm": 0.5198819509993116, + "learning_rate": 7.23128049634722e-05, + "loss": 0.4809, + "step": 289 + }, + { + "epoch": 1.4074579354251933, + "grad_norm": 0.5071119674882796, + "learning_rate": 7.223271681263916e-05, + "loss": 0.4839, + "step": 290 + }, + { + "epoch": 1.412308625132636, + "grad_norm": 0.4871736388487796, + "learning_rate": 7.215225846149957e-05, + "loss": 0.4899, + "step": 291 + }, + { + "epoch": 1.4171593148400787, + "grad_norm": 0.4644477177967194, + "learning_rate": 7.207143083413643e-05, + "loss": 0.4865, + "step": 292 + }, + { + "epoch": 1.4220100045475217, + "grad_norm": 0.5305010693129898, + "learning_rate": 7.1990234858874e-05, + "loss": 0.4876, + "step": 293 + }, + { + "epoch": 1.4268606942549644, + "grad_norm": 0.5381812012245146, + "learning_rate": 7.190867146826707e-05, + "loss": 0.4936, + "step": 294 + }, + { + "epoch": 1.4317113839624072, + "grad_norm": 0.42926452605809334, + "learning_rate": 7.182674159909031e-05, + "loss": 0.4845, + "step": 295 + }, + { + "epoch": 1.43656207366985, + "grad_norm": 0.42959757228135126, + "learning_rate": 7.174444619232745e-05, + "loss": 0.4952, + "step": 296 + }, + { + "epoch": 1.4414127633772926, + "grad_norm": 0.4994771109955089, + "learning_rate": 7.166178619316056e-05, + "loss": 0.4912, + "step": 297 + }, + { + "epoch": 1.4462634530847356, + "grad_norm": 0.4676921077663137, + "learning_rate": 7.157876255095906e-05, + "loss": 0.4875, + "step": 298 + }, + { + "epoch": 1.4511141427921783, + "grad_norm": 0.430209471088095, + "learning_rate": 7.149537621926895e-05, + "loss": 0.4862, + "step": 299 + }, + { + "epoch": 1.455964832499621, + "grad_norm": 0.5267214852499816, + "learning_rate": 7.14116281558018e-05, + "loss": 0.4879, + "step": 300 + }, + { + "epoch": 1.4608155222070638, + "grad_norm": 0.5169308603441447, + "learning_rate": 7.132751932242376e-05, + "loss": 0.4984, + "step": 301 + }, + { + "epoch": 1.4656662119145065, + "grad_norm": 0.3435076728807633, + "learning_rate": 7.124305068514444e-05, + "loss": 0.487, + "step": 302 + }, + { + "epoch": 1.4705169016219495, + "grad_norm": 0.2662958717194974, + "learning_rate": 7.1158223214106e-05, + "loss": 0.4878, + "step": 303 + }, + { + "epoch": 1.4753675913293922, + "grad_norm": 0.31770648802942325, + "learning_rate": 7.107303788357177e-05, + "loss": 0.4819, + "step": 304 + }, + { + "epoch": 1.480218281036835, + "grad_norm": 0.29599579651368213, + "learning_rate": 7.098749567191527e-05, + "loss": 0.4852, + "step": 305 + }, + { + "epoch": 1.4850689707442777, + "grad_norm": 0.3093179833213525, + "learning_rate": 7.090159756160886e-05, + "loss": 0.4877, + "step": 306 + }, + { + "epoch": 1.4899196604517204, + "grad_norm": 0.38971280696993216, + "learning_rate": 7.081534453921242e-05, + "loss": 0.4852, + "step": 307 + }, + { + "epoch": 1.4947703501591634, + "grad_norm": 0.3825128078756503, + "learning_rate": 7.072873759536217e-05, + "loss": 0.4913, + "step": 308 + }, + { + "epoch": 1.499621039866606, + "grad_norm": 0.2886624346515238, + "learning_rate": 7.064177772475912e-05, + "loss": 0.4798, + "step": 309 + }, + { + "epoch": 1.5044717295740488, + "grad_norm": 0.3712678777863381, + "learning_rate": 7.05544659261578e-05, + "loss": 0.4867, + "step": 310 + }, + { + "epoch": 1.5093224192814916, + "grad_norm": 0.42331161544955054, + "learning_rate": 7.046680320235466e-05, + "loss": 0.4871, + "step": 311 + }, + { + "epoch": 1.5141731089889343, + "grad_norm": 0.4018226592697959, + "learning_rate": 7.037879056017663e-05, + "loss": 0.4842, + "step": 312 + }, + { + "epoch": 1.5190237986963773, + "grad_norm": 0.3959692369182176, + "learning_rate": 7.029042901046952e-05, + "loss": 0.4802, + "step": 313 + }, + { + "epoch": 1.5238744884038198, + "grad_norm": 0.4052319536586186, + "learning_rate": 7.020171956808645e-05, + "loss": 0.4859, + "step": 314 + }, + { + "epoch": 1.5287251781112627, + "grad_norm": 0.40673400305291324, + "learning_rate": 7.011266325187615e-05, + "loss": 0.496, + "step": 315 + }, + { + "epoch": 1.5335758678187055, + "grad_norm": 0.44258718402573904, + "learning_rate": 7.002326108467129e-05, + "loss": 0.4864, + "step": 316 + }, + { + "epoch": 1.5384265575261482, + "grad_norm": 0.4022422201093314, + "learning_rate": 6.993351409327672e-05, + "loss": 0.4763, + "step": 317 + }, + { + "epoch": 1.5432772472335912, + "grad_norm": 0.39107692009497685, + "learning_rate": 6.984342330845764e-05, + "loss": 0.4952, + "step": 318 + }, + { + "epoch": 1.5481279369410337, + "grad_norm": 0.36401667233363455, + "learning_rate": 6.975298976492785e-05, + "loss": 0.4952, + "step": 319 + }, + { + "epoch": 1.5529786266484766, + "grad_norm": 0.3964209144895704, + "learning_rate": 6.966221450133779e-05, + "loss": 0.4901, + "step": 320 + }, + { + "epoch": 1.5578293163559194, + "grad_norm": 0.41002635948711413, + "learning_rate": 6.957109856026261e-05, + "loss": 0.4917, + "step": 321 + }, + { + "epoch": 1.562680006063362, + "grad_norm": 0.33043042887382146, + "learning_rate": 6.94796429881903e-05, + "loss": 0.4771, + "step": 322 + }, + { + "epoch": 1.567530695770805, + "grad_norm": 0.3258167342621945, + "learning_rate": 6.938784883550948e-05, + "loss": 0.4889, + "step": 323 + }, + { + "epoch": 1.5723813854782476, + "grad_norm": 0.33864970777234993, + "learning_rate": 6.929571715649755e-05, + "loss": 0.4866, + "step": 324 + }, + { + "epoch": 1.5772320751856905, + "grad_norm": 0.3799335436470155, + "learning_rate": 6.920324900930842e-05, + "loss": 0.4907, + "step": 325 + }, + { + "epoch": 1.5820827648931333, + "grad_norm": 0.38649667346458383, + "learning_rate": 6.911044545596042e-05, + "loss": 0.4854, + "step": 326 + }, + { + "epoch": 1.586933454600576, + "grad_norm": 0.3332457174533195, + "learning_rate": 6.901730756232411e-05, + "loss": 0.4895, + "step": 327 + }, + { + "epoch": 1.591784144308019, + "grad_norm": 0.3688581464993942, + "learning_rate": 6.892383639811005e-05, + "loss": 0.4958, + "step": 328 + }, + { + "epoch": 1.5966348340154615, + "grad_norm": 0.43127701691860393, + "learning_rate": 6.883003303685644e-05, + "loss": 0.4844, + "step": 329 + }, + { + "epoch": 1.6014855237229044, + "grad_norm": 0.5098788849460419, + "learning_rate": 6.87358985559169e-05, + "loss": 0.489, + "step": 330 + }, + { + "epoch": 1.6063362134303472, + "grad_norm": 0.531466303384909, + "learning_rate": 6.864143403644797e-05, + "loss": 0.4945, + "step": 331 + }, + { + "epoch": 1.61118690313779, + "grad_norm": 0.4575057116450561, + "learning_rate": 6.85466405633968e-05, + "loss": 0.4855, + "step": 332 + }, + { + "epoch": 1.6160375928452326, + "grad_norm": 0.42418027914564915, + "learning_rate": 6.845151922548865e-05, + "loss": 0.4783, + "step": 333 + }, + { + "epoch": 1.6208882825526754, + "grad_norm": 0.38431703300530295, + "learning_rate": 6.835607111521439e-05, + "loss": 0.4796, + "step": 334 + }, + { + "epoch": 1.6257389722601183, + "grad_norm": 0.325044588754403, + "learning_rate": 6.826029732881793e-05, + "loss": 0.4928, + "step": 335 + }, + { + "epoch": 1.630589661967561, + "grad_norm": 0.2723847494193817, + "learning_rate": 6.816419896628363e-05, + "loss": 0.4851, + "step": 336 + }, + { + "epoch": 1.6354403516750038, + "grad_norm": 0.25325170865464947, + "learning_rate": 6.806777713132374e-05, + "loss": 0.4826, + "step": 337 + }, + { + "epoch": 1.6402910413824465, + "grad_norm": 0.2793387163645126, + "learning_rate": 6.79710329313656e-05, + "loss": 0.4873, + "step": 338 + }, + { + "epoch": 1.6451417310898893, + "grad_norm": 0.38835730960557174, + "learning_rate": 6.787396747753903e-05, + "loss": 0.4744, + "step": 339 + }, + { + "epoch": 1.6499924207973322, + "grad_norm": 0.5896470389457479, + "learning_rate": 6.777658188466354e-05, + "loss": 0.4765, + "step": 340 + }, + { + "epoch": 1.654843110504775, + "grad_norm": 0.7717199781637745, + "learning_rate": 6.767887727123544e-05, + "loss": 0.4931, + "step": 341 + }, + { + "epoch": 1.6596938002122177, + "grad_norm": 1.0491183054565791, + "learning_rate": 6.758085475941516e-05, + "loss": 0.4875, + "step": 342 + }, + { + "epoch": 1.6645444899196604, + "grad_norm": 1.0963514331569248, + "learning_rate": 6.748251547501418e-05, + "loss": 0.4783, + "step": 343 + }, + { + "epoch": 1.6693951796271032, + "grad_norm": 0.6008540810802777, + "learning_rate": 6.738386054748226e-05, + "loss": 0.4836, + "step": 344 + }, + { + "epoch": 1.674245869334546, + "grad_norm": 0.4448291346545642, + "learning_rate": 6.728489110989434e-05, + "loss": 0.4883, + "step": 345 + }, + { + "epoch": 1.6790965590419886, + "grad_norm": 0.4942617693141126, + "learning_rate": 6.718560829893762e-05, + "loss": 0.4799, + "step": 346 + }, + { + "epoch": 1.6839472487494316, + "grad_norm": 0.4996733158915889, + "learning_rate": 6.708601325489844e-05, + "loss": 0.4872, + "step": 347 + }, + { + "epoch": 1.6887979384568743, + "grad_norm": 0.3715020675799419, + "learning_rate": 6.698610712164924e-05, + "loss": 0.4864, + "step": 348 + }, + { + "epoch": 1.693648628164317, + "grad_norm": 0.27977348496462506, + "learning_rate": 6.688589104663536e-05, + "loss": 0.4731, + "step": 349 + }, + { + "epoch": 1.69849931787176, + "grad_norm": 0.3329757061053683, + "learning_rate": 6.67853661808619e-05, + "loss": 0.4771, + "step": 350 + }, + { + "epoch": 1.7033500075792025, + "grad_norm": 0.35849519050184514, + "learning_rate": 6.668453367888052e-05, + "loss": 0.4867, + "step": 351 + }, + { + "epoch": 1.7082006972866455, + "grad_norm": 0.3093064728603477, + "learning_rate": 6.658339469877613e-05, + "loss": 0.478, + "step": 352 + }, + { + "epoch": 1.7130513869940882, + "grad_norm": 0.2983932071893944, + "learning_rate": 6.64819504021536e-05, + "loss": 0.4814, + "step": 353 + }, + { + "epoch": 1.717902076701531, + "grad_norm": 0.32895472335219694, + "learning_rate": 6.638020195412448e-05, + "loss": 0.4771, + "step": 354 + }, + { + "epoch": 1.722752766408974, + "grad_norm": 0.4277599821762847, + "learning_rate": 6.627815052329354e-05, + "loss": 0.4925, + "step": 355 + }, + { + "epoch": 1.7276034561164164, + "grad_norm": 0.4690652488456737, + "learning_rate": 6.617579728174535e-05, + "loss": 0.4854, + "step": 356 + }, + { + "epoch": 1.7324541458238594, + "grad_norm": 0.43487992599318925, + "learning_rate": 6.60731434050309e-05, + "loss": 0.4777, + "step": 357 + }, + { + "epoch": 1.737304835531302, + "grad_norm": 0.4398837483674929, + "learning_rate": 6.597019007215401e-05, + "loss": 0.4783, + "step": 358 + }, + { + "epoch": 1.7421555252387448, + "grad_norm": 0.43050412759021583, + "learning_rate": 6.586693846555788e-05, + "loss": 0.4743, + "step": 359 + }, + { + "epoch": 1.7470062149461878, + "grad_norm": 0.40202174280112624, + "learning_rate": 6.576338977111134e-05, + "loss": 0.48, + "step": 360 + }, + { + "epoch": 1.7518569046536303, + "grad_norm": 0.36259944082265505, + "learning_rate": 6.565954517809543e-05, + "loss": 0.4747, + "step": 361 + }, + { + "epoch": 1.7567075943610733, + "grad_norm": 0.28300782787905476, + "learning_rate": 6.555540587918968e-05, + "loss": 0.4778, + "step": 362 + }, + { + "epoch": 1.761558284068516, + "grad_norm": 0.27286805994349533, + "learning_rate": 6.545097307045831e-05, + "loss": 0.4795, + "step": 363 + }, + { + "epoch": 1.7664089737759587, + "grad_norm": 0.31934491384929364, + "learning_rate": 6.534624795133662e-05, + "loss": 0.4851, + "step": 364 + }, + { + "epoch": 1.7712596634834017, + "grad_norm": 0.35245692182661065, + "learning_rate": 6.524123172461711e-05, + "loss": 0.4794, + "step": 365 + }, + { + "epoch": 1.7761103531908442, + "grad_norm": 0.3396944161317505, + "learning_rate": 6.51359255964358e-05, + "loss": 0.4774, + "step": 366 + }, + { + "epoch": 1.7809610428982872, + "grad_norm": 0.2877754553279699, + "learning_rate": 6.503033077625824e-05, + "loss": 0.4746, + "step": 367 + }, + { + "epoch": 1.78581173260573, + "grad_norm": 0.3326775146217252, + "learning_rate": 6.492444847686566e-05, + "loss": 0.4849, + "step": 368 + }, + { + "epoch": 1.7906624223131726, + "grad_norm": 0.419702080624426, + "learning_rate": 6.481827991434111e-05, + "loss": 0.4814, + "step": 369 + }, + { + "epoch": 1.7955131120206156, + "grad_norm": 0.43780909545325103, + "learning_rate": 6.471182630805538e-05, + "loss": 0.4813, + "step": 370 + }, + { + "epoch": 1.800363801728058, + "grad_norm": 0.5080371772111751, + "learning_rate": 6.460508888065314e-05, + "loss": 0.4865, + "step": 371 + }, + { + "epoch": 1.805214491435501, + "grad_norm": 0.5541771901573003, + "learning_rate": 6.449806885803873e-05, + "loss": 0.4752, + "step": 372 + }, + { + "epoch": 1.8100651811429438, + "grad_norm": 0.5077930985424878, + "learning_rate": 6.439076746936219e-05, + "loss": 0.4776, + "step": 373 + }, + { + "epoch": 1.8149158708503865, + "grad_norm": 0.3974055597915673, + "learning_rate": 6.428318594700509e-05, + "loss": 0.4833, + "step": 374 + }, + { + "epoch": 1.8197665605578295, + "grad_norm": 0.27838793632255576, + "learning_rate": 6.417532552656647e-05, + "loss": 0.4808, + "step": 375 + }, + { + "epoch": 1.824617250265272, + "grad_norm": 0.25531453372215107, + "learning_rate": 6.406718744684851e-05, + "loss": 0.475, + "step": 376 + }, + { + "epoch": 1.829467939972715, + "grad_norm": 0.37619306197372576, + "learning_rate": 6.395877294984241e-05, + "loss": 0.4718, + "step": 377 + }, + { + "epoch": 1.8343186296801577, + "grad_norm": 0.45314622671289495, + "learning_rate": 6.385008328071406e-05, + "loss": 0.4858, + "step": 378 + }, + { + "epoch": 1.8391693193876004, + "grad_norm": 0.4091697946782105, + "learning_rate": 6.374111968778982e-05, + "loss": 0.4797, + "step": 379 + }, + { + "epoch": 1.8440200090950432, + "grad_norm": 0.31229108750141443, + "learning_rate": 6.363188342254206e-05, + "loss": 0.4819, + "step": 380 + }, + { + "epoch": 1.848870698802486, + "grad_norm": 0.29389799920114323, + "learning_rate": 6.352237573957488e-05, + "loss": 0.4777, + "step": 381 + }, + { + "epoch": 1.8537213885099288, + "grad_norm": 0.31830405845533455, + "learning_rate": 6.341259789660969e-05, + "loss": 0.4812, + "step": 382 + }, + { + "epoch": 1.8585720782173716, + "grad_norm": 0.2783419658252336, + "learning_rate": 6.330255115447076e-05, + "loss": 0.4722, + "step": 383 + }, + { + "epoch": 1.8634227679248143, + "grad_norm": 0.252195902565345, + "learning_rate": 6.319223677707069e-05, + "loss": 0.4786, + "step": 384 + }, + { + "epoch": 1.868273457632257, + "grad_norm": 0.24931429544692238, + "learning_rate": 6.308165603139598e-05, + "loss": 0.4766, + "step": 385 + }, + { + "epoch": 1.8731241473396998, + "grad_norm": 0.343056469344131, + "learning_rate": 6.29708101874924e-05, + "loss": 0.4746, + "step": 386 + }, + { + "epoch": 1.8779748370471427, + "grad_norm": 0.39163026769861153, + "learning_rate": 6.285970051845045e-05, + "loss": 0.4777, + "step": 387 + }, + { + "epoch": 1.8828255267545853, + "grad_norm": 0.3925083138992103, + "learning_rate": 6.274832830039071e-05, + "loss": 0.4762, + "step": 388 + }, + { + "epoch": 1.8876762164620282, + "grad_norm": 0.3908015146094751, + "learning_rate": 6.26366948124492e-05, + "loss": 0.4882, + "step": 389 + }, + { + "epoch": 1.892526906169471, + "grad_norm": 0.38768255703534454, + "learning_rate": 6.25248013367627e-05, + "loss": 0.4746, + "step": 390 + }, + { + "epoch": 1.8973775958769137, + "grad_norm": 0.34797592572865116, + "learning_rate": 6.241264915845401e-05, + "loss": 0.4863, + "step": 391 + }, + { + "epoch": 1.9022282855843566, + "grad_norm": 0.2939167601152598, + "learning_rate": 6.230023956561716e-05, + "loss": 0.4803, + "step": 392 + }, + { + "epoch": 1.9070789752917991, + "grad_norm": 0.2683598696324213, + "learning_rate": 6.218757384930268e-05, + "loss": 0.4769, + "step": 393 + }, + { + "epoch": 1.911929664999242, + "grad_norm": 0.2963955328357841, + "learning_rate": 6.207465330350273e-05, + "loss": 0.4798, + "step": 394 + }, + { + "epoch": 1.9167803547066848, + "grad_norm": 0.34936564317737695, + "learning_rate": 6.196147922513623e-05, + "loss": 0.479, + "step": 395 + }, + { + "epoch": 1.9216310444141276, + "grad_norm": 0.3023947037005716, + "learning_rate": 6.184805291403402e-05, + "loss": 0.4776, + "step": 396 + }, + { + "epoch": 1.9264817341215705, + "grad_norm": 0.23472105439779495, + "learning_rate": 6.173437567292383e-05, + "loss": 0.475, + "step": 397 + }, + { + "epoch": 1.931332423829013, + "grad_norm": 0.30151702935680424, + "learning_rate": 6.162044880741544e-05, + "loss": 0.4719, + "step": 398 + }, + { + "epoch": 1.936183113536456, + "grad_norm": 0.4380813843564537, + "learning_rate": 6.150627362598557e-05, + "loss": 0.4871, + "step": 399 + }, + { + "epoch": 1.9410338032438987, + "grad_norm": 0.5034940901090467, + "learning_rate": 6.139185143996298e-05, + "loss": 0.4806, + "step": 400 + }, + { + "epoch": 1.9458844929513415, + "grad_norm": 0.44831770290541656, + "learning_rate": 6.127718356351326e-05, + "loss": 0.478, + "step": 401 + }, + { + "epoch": 1.9507351826587844, + "grad_norm": 0.3881020850233725, + "learning_rate": 6.116227131362385e-05, + "loss": 0.4714, + "step": 402 + }, + { + "epoch": 1.955585872366227, + "grad_norm": 0.3382903208345561, + "learning_rate": 6.104711601008888e-05, + "loss": 0.4779, + "step": 403 + }, + { + "epoch": 1.96043656207367, + "grad_norm": 0.2830781275191087, + "learning_rate": 6.0931718975493985e-05, + "loss": 0.4846, + "step": 404 + }, + { + "epoch": 1.9652872517811126, + "grad_norm": 0.27767803820547865, + "learning_rate": 6.081608153520117e-05, + "loss": 0.4691, + "step": 405 + }, + { + "epoch": 1.9701379414885554, + "grad_norm": 0.35527234014372044, + "learning_rate": 6.0700205017333525e-05, + "loss": 0.4787, + "step": 406 + }, + { + "epoch": 1.9749886311959983, + "grad_norm": 0.34300192919407774, + "learning_rate": 6.058409075276002e-05, + "loss": 0.4689, + "step": 407 + }, + { + "epoch": 1.9798393209034408, + "grad_norm": 0.2964726842271146, + "learning_rate": 6.046774007508019e-05, + "loss": 0.475, + "step": 408 + }, + { + "epoch": 1.9846900106108838, + "grad_norm": 0.2502113324820329, + "learning_rate": 6.035115432060883e-05, + "loss": 0.4747, + "step": 409 + }, + { + "epoch": 1.9895407003183265, + "grad_norm": 0.23471156048166733, + "learning_rate": 6.0234334828360655e-05, + "loss": 0.4786, + "step": 410 + }, + { + "epoch": 1.9943913900257693, + "grad_norm": 0.27137410563019304, + "learning_rate": 6.011728294003494e-05, + "loss": 0.4802, + "step": 411 + }, + { + "epoch": 1.9992420797332122, + "grad_norm": 0.2960970008273601, + "learning_rate": 6.000000000000001e-05, + "loss": 0.4768, + "step": 412 + }, + { + "epoch": 2.004850689707443, + "grad_norm": 0.33644070901215145, + "learning_rate": 5.988248735527793e-05, + "loss": 0.4473, + "step": 413 + }, + { + "epoch": 2.0097013794148855, + "grad_norm": 0.39546749871672404, + "learning_rate": 5.9764746355528994e-05, + "loss": 0.4501, + "step": 414 + }, + { + "epoch": 2.0145520691223284, + "grad_norm": 0.47967552460987467, + "learning_rate": 5.964677835303615e-05, + "loss": 0.4483, + "step": 415 + }, + { + "epoch": 2.019402758829771, + "grad_norm": 0.43081443925289625, + "learning_rate": 5.952858470268955e-05, + "loss": 0.4468, + "step": 416 + }, + { + "epoch": 2.024253448537214, + "grad_norm": 0.42692273173821377, + "learning_rate": 5.941016676197098e-05, + "loss": 0.4499, + "step": 417 + }, + { + "epoch": 2.029104138244657, + "grad_norm": 0.5203095933335755, + "learning_rate": 5.929152589093825e-05, + "loss": 0.4498, + "step": 418 + }, + { + "epoch": 2.0339548279520994, + "grad_norm": 0.5390774215298352, + "learning_rate": 5.9172663452209554e-05, + "loss": 0.449, + "step": 419 + }, + { + "epoch": 2.0388055176595423, + "grad_norm": 0.5618164866931807, + "learning_rate": 5.9053580810947845e-05, + "loss": 0.4555, + "step": 420 + }, + { + "epoch": 2.043656207366985, + "grad_norm": 0.6473211411897097, + "learning_rate": 5.89342793348452e-05, + "loss": 0.4537, + "step": 421 + }, + { + "epoch": 2.048506897074428, + "grad_norm": 0.7641174380933736, + "learning_rate": 5.881476039410699e-05, + "loss": 0.4579, + "step": 422 + }, + { + "epoch": 2.0533575867818707, + "grad_norm": 0.6589276143277975, + "learning_rate": 5.869502536143629e-05, + "loss": 0.4478, + "step": 423 + }, + { + "epoch": 2.0582082764893133, + "grad_norm": 0.4055360078631115, + "learning_rate": 5.857507561201802e-05, + "loss": 0.4501, + "step": 424 + }, + { + "epoch": 2.063058966196756, + "grad_norm": 0.3576976481492298, + "learning_rate": 5.845491252350312e-05, + "loss": 0.4479, + "step": 425 + }, + { + "epoch": 2.0679096559041987, + "grad_norm": 0.444390387775824, + "learning_rate": 5.833453747599286e-05, + "loss": 0.4466, + "step": 426 + }, + { + "epoch": 2.0727603456116417, + "grad_norm": 0.38706062123939283, + "learning_rate": 5.821395185202285e-05, + "loss": 0.449, + "step": 427 + }, + { + "epoch": 2.0776110353190846, + "grad_norm": 0.3243650769978332, + "learning_rate": 5.809315703654726e-05, + "loss": 0.4581, + "step": 428 + }, + { + "epoch": 2.082461725026527, + "grad_norm": 0.35862489527477903, + "learning_rate": 5.797215441692284e-05, + "loss": 0.4534, + "step": 429 + }, + { + "epoch": 2.08731241473397, + "grad_norm": 0.3353344776943914, + "learning_rate": 5.785094538289304e-05, + "loss": 0.4537, + "step": 430 + }, + { + "epoch": 2.0921631044414126, + "grad_norm": 0.3620686259692887, + "learning_rate": 5.772953132657202e-05, + "loss": 0.4553, + "step": 431 + }, + { + "epoch": 2.0970137941488556, + "grad_norm": 0.307403523851727, + "learning_rate": 5.7607913642428666e-05, + "loss": 0.4424, + "step": 432 + }, + { + "epoch": 2.101864483856298, + "grad_norm": 0.28710201149568576, + "learning_rate": 5.7486093727270606e-05, + "loss": 0.4462, + "step": 433 + }, + { + "epoch": 2.106715173563741, + "grad_norm": 0.31451691469735704, + "learning_rate": 5.736407298022809e-05, + "loss": 0.4434, + "step": 434 + }, + { + "epoch": 2.111565863271184, + "grad_norm": 0.3105010146819863, + "learning_rate": 5.7241852802738e-05, + "loss": 0.4533, + "step": 435 + }, + { + "epoch": 2.1164165529786265, + "grad_norm": 0.2878408032383936, + "learning_rate": 5.711943459852772e-05, + "loss": 0.4427, + "step": 436 + }, + { + "epoch": 2.1212672426860695, + "grad_norm": 0.3026608573456508, + "learning_rate": 5.699681977359902e-05, + "loss": 0.4385, + "step": 437 + }, + { + "epoch": 2.1261179323935124, + "grad_norm": 0.2857208984253648, + "learning_rate": 5.6874009736211896e-05, + "loss": 0.4465, + "step": 438 + }, + { + "epoch": 2.130968622100955, + "grad_norm": 0.28786201316205207, + "learning_rate": 5.675100589686839e-05, + "loss": 0.4472, + "step": 439 + }, + { + "epoch": 2.135819311808398, + "grad_norm": 0.3397455432854385, + "learning_rate": 5.662780966829646e-05, + "loss": 0.4486, + "step": 440 + }, + { + "epoch": 2.1406700015158404, + "grad_norm": 0.3060713829784068, + "learning_rate": 5.650442246543364e-05, + "loss": 0.4525, + "step": 441 + }, + { + "epoch": 2.1455206912232834, + "grad_norm": 0.2554715156912105, + "learning_rate": 5.638084570541088e-05, + "loss": 0.4451, + "step": 442 + }, + { + "epoch": 2.150371380930726, + "grad_norm": 0.31763468046624377, + "learning_rate": 5.625708080753621e-05, + "loss": 0.455, + "step": 443 + }, + { + "epoch": 2.155222070638169, + "grad_norm": 0.3288505103534938, + "learning_rate": 5.6133129193278525e-05, + "loss": 0.4453, + "step": 444 + }, + { + "epoch": 2.160072760345612, + "grad_norm": 0.27302062189682574, + "learning_rate": 5.600899228625112e-05, + "loss": 0.4523, + "step": 445 + }, + { + "epoch": 2.1649234500530543, + "grad_norm": 0.2367917713116079, + "learning_rate": 5.588467151219549e-05, + "loss": 0.4481, + "step": 446 + }, + { + "epoch": 2.1697741397604973, + "grad_norm": 0.27892881990044693, + "learning_rate": 5.5760168298964874e-05, + "loss": 0.4397, + "step": 447 + }, + { + "epoch": 2.17462482946794, + "grad_norm": 0.3659363982261399, + "learning_rate": 5.563548407650782e-05, + "loss": 0.4464, + "step": 448 + }, + { + "epoch": 2.1794755191753827, + "grad_norm": 0.3139974928184246, + "learning_rate": 5.551062027685187e-05, + "loss": 0.4487, + "step": 449 + }, + { + "epoch": 2.1843262088828257, + "grad_norm": 0.25894377780084493, + "learning_rate": 5.5385578334087006e-05, + "loss": 0.4481, + "step": 450 + }, + { + "epoch": 2.189176898590268, + "grad_norm": 0.17951964246489394, + "learning_rate": 5.526035968434927e-05, + "loss": 0.4469, + "step": 451 + }, + { + "epoch": 2.194027588297711, + "grad_norm": 0.21554515294483917, + "learning_rate": 5.513496576580418e-05, + "loss": 0.4573, + "step": 452 + }, + { + "epoch": 2.1988782780051537, + "grad_norm": 0.24799983544152385, + "learning_rate": 5.5009398018630276e-05, + "loss": 0.4498, + "step": 453 + }, + { + "epoch": 2.2037289677125966, + "grad_norm": 0.186973375547793, + "learning_rate": 5.4883657885002575e-05, + "loss": 0.4449, + "step": 454 + }, + { + "epoch": 2.2085796574200396, + "grad_norm": 0.1888483688795653, + "learning_rate": 5.475774680907597e-05, + "loss": 0.443, + "step": 455 + }, + { + "epoch": 2.213430347127482, + "grad_norm": 0.18914695692387654, + "learning_rate": 5.463166623696868e-05, + "loss": 0.4434, + "step": 456 + }, + { + "epoch": 2.218281036834925, + "grad_norm": 0.19070268199404652, + "learning_rate": 5.450541761674562e-05, + "loss": 0.4445, + "step": 457 + }, + { + "epoch": 2.2231317265423676, + "grad_norm": 0.2259939851217927, + "learning_rate": 5.437900239840179e-05, + "loss": 0.4465, + "step": 458 + }, + { + "epoch": 2.2279824162498105, + "grad_norm": 0.18023421500987896, + "learning_rate": 5.42524220338456e-05, + "loss": 0.4453, + "step": 459 + }, + { + "epoch": 2.2328331059572535, + "grad_norm": 0.16598518940751159, + "learning_rate": 5.412567797688219e-05, + "loss": 0.4498, + "step": 460 + }, + { + "epoch": 2.237683795664696, + "grad_norm": 0.16004882092407235, + "learning_rate": 5.3998771683196754e-05, + "loss": 0.4527, + "step": 461 + }, + { + "epoch": 2.242534485372139, + "grad_norm": 0.17890202722776521, + "learning_rate": 5.3871704610337836e-05, + "loss": 0.4444, + "step": 462 + }, + { + "epoch": 2.2473851750795815, + "grad_norm": 0.182744670257566, + "learning_rate": 5.374447821770053e-05, + "loss": 0.4431, + "step": 463 + }, + { + "epoch": 2.2522358647870244, + "grad_norm": 0.19342699163866056, + "learning_rate": 5.361709396650977e-05, + "loss": 0.4404, + "step": 464 + }, + { + "epoch": 2.2570865544944674, + "grad_norm": 0.20442779398031627, + "learning_rate": 5.3489553319803566e-05, + "loss": 0.4496, + "step": 465 + }, + { + "epoch": 2.26193724420191, + "grad_norm": 0.23359410284964036, + "learning_rate": 5.336185774241609e-05, + "loss": 0.4469, + "step": 466 + }, + { + "epoch": 2.266787933909353, + "grad_norm": 0.23295417695606166, + "learning_rate": 5.3234008700961e-05, + "loss": 0.4505, + "step": 467 + }, + { + "epoch": 2.2716386236167954, + "grad_norm": 0.20207094824496044, + "learning_rate": 5.3106007663814505e-05, + "loss": 0.4406, + "step": 468 + }, + { + "epoch": 2.2764893133242383, + "grad_norm": 0.1850252157104855, + "learning_rate": 5.2977856101098484e-05, + "loss": 0.4525, + "step": 469 + }, + { + "epoch": 2.281340003031681, + "grad_norm": 0.1821206965545461, + "learning_rate": 5.284955548466371e-05, + "loss": 0.4592, + "step": 470 + }, + { + "epoch": 2.286190692739124, + "grad_norm": 0.1913435003815255, + "learning_rate": 5.272110728807279e-05, + "loss": 0.4459, + "step": 471 + }, + { + "epoch": 2.2910413824465667, + "grad_norm": 0.17908151714339782, + "learning_rate": 5.25925129865834e-05, + "loss": 0.4523, + "step": 472 + }, + { + "epoch": 2.2958920721540093, + "grad_norm": 0.17796456682985312, + "learning_rate": 5.246377405713121e-05, + "loss": 0.4426, + "step": 473 + }, + { + "epoch": 2.300742761861452, + "grad_norm": 0.16950491734508644, + "learning_rate": 5.2334891978313006e-05, + "loss": 0.4426, + "step": 474 + }, + { + "epoch": 2.305593451568895, + "grad_norm": 0.18036359667208995, + "learning_rate": 5.220586823036966e-05, + "loss": 0.4458, + "step": 475 + }, + { + "epoch": 2.3104441412763377, + "grad_norm": 0.1873060682555774, + "learning_rate": 5.207670429516915e-05, + "loss": 0.4433, + "step": 476 + }, + { + "epoch": 2.3152948309837806, + "grad_norm": 0.1986452939709168, + "learning_rate": 5.1947401656189546e-05, + "loss": 0.4593, + "step": 477 + }, + { + "epoch": 2.320145520691223, + "grad_norm": 0.20494449288937291, + "learning_rate": 5.181796179850197e-05, + "loss": 0.4424, + "step": 478 + }, + { + "epoch": 2.324996210398666, + "grad_norm": 0.17952910191793728, + "learning_rate": 5.168838620875352e-05, + "loss": 0.4503, + "step": 479 + }, + { + "epoch": 2.3298469001061086, + "grad_norm": 0.1909701609269039, + "learning_rate": 5.155867637515019e-05, + "loss": 0.4506, + "step": 480 + }, + { + "epoch": 2.3346975898135516, + "grad_norm": 0.1838298898954926, + "learning_rate": 5.142883378743984e-05, + "loss": 0.4513, + "step": 481 + }, + { + "epoch": 2.3395482795209945, + "grad_norm": 0.1818267641683358, + "learning_rate": 5.129885993689502e-05, + "loss": 0.4488, + "step": 482 + }, + { + "epoch": 2.344398969228437, + "grad_norm": 0.21197375880432345, + "learning_rate": 5.116875631629585e-05, + "loss": 0.4456, + "step": 483 + }, + { + "epoch": 2.34924965893588, + "grad_norm": 0.21240893965447508, + "learning_rate": 5.10385244199129e-05, + "loss": 0.4386, + "step": 484 + }, + { + "epoch": 2.354100348643323, + "grad_norm": 0.19244612255162405, + "learning_rate": 5.0908165743490047e-05, + "loss": 0.4482, + "step": 485 + }, + { + "epoch": 2.3589510383507655, + "grad_norm": 0.22440529731925618, + "learning_rate": 5.0777681784227224e-05, + "loss": 0.4496, + "step": 486 + }, + { + "epoch": 2.3638017280582084, + "grad_norm": 0.249440062974833, + "learning_rate": 5.064707404076327e-05, + "loss": 0.4502, + "step": 487 + }, + { + "epoch": 2.368652417765651, + "grad_norm": 0.2374206142112278, + "learning_rate": 5.051634401315875e-05, + "loss": 0.448, + "step": 488 + }, + { + "epoch": 2.373503107473094, + "grad_norm": 0.21044332969367502, + "learning_rate": 5.0385493202878656e-05, + "loss": 0.4416, + "step": 489 + }, + { + "epoch": 2.3783537971805364, + "grad_norm": 0.15343545111269605, + "learning_rate": 5.025452311277522e-05, + "loss": 0.4413, + "step": 490 + }, + { + "epoch": 2.3832044868879794, + "grad_norm": 0.17472771019103053, + "learning_rate": 5.01234352470706e-05, + "loss": 0.4472, + "step": 491 + }, + { + "epoch": 2.3880551765954223, + "grad_norm": 0.2225509747823868, + "learning_rate": 4.999223111133968e-05, + "loss": 0.4405, + "step": 492 + }, + { + "epoch": 2.392905866302865, + "grad_norm": 0.27110633436791925, + "learning_rate": 4.986091221249269e-05, + "loss": 0.44, + "step": 493 + }, + { + "epoch": 2.397756556010308, + "grad_norm": 0.24255464597168586, + "learning_rate": 4.972948005875796e-05, + "loss": 0.4432, + "step": 494 + }, + { + "epoch": 2.4026072457177503, + "grad_norm": 0.2496648842091371, + "learning_rate": 4.959793615966459e-05, + "loss": 0.4401, + "step": 495 + }, + { + "epoch": 2.4074579354251933, + "grad_norm": 0.24806426439634907, + "learning_rate": 4.946628202602508e-05, + "loss": 0.4526, + "step": 496 + }, + { + "epoch": 2.412308625132636, + "grad_norm": 0.21808090914084832, + "learning_rate": 4.933451916991802e-05, + "loss": 0.4474, + "step": 497 + }, + { + "epoch": 2.4171593148400787, + "grad_norm": 0.19833835766366836, + "learning_rate": 4.920264910467066e-05, + "loss": 0.4485, + "step": 498 + }, + { + "epoch": 2.4220100045475217, + "grad_norm": 0.1904056029579938, + "learning_rate": 4.9070673344841645e-05, + "loss": 0.4471, + "step": 499 + }, + { + "epoch": 2.426860694254964, + "grad_norm": 0.17821880940044135, + "learning_rate": 4.893859340620348e-05, + "loss": 0.4518, + "step": 500 + }, + { + "epoch": 2.431711383962407, + "grad_norm": 0.16242846601925154, + "learning_rate": 4.880641080572522e-05, + "loss": 0.4426, + "step": 501 + }, + { + "epoch": 2.43656207366985, + "grad_norm": 0.16230843192633562, + "learning_rate": 4.8674127061555025e-05, + "loss": 0.4492, + "step": 502 + }, + { + "epoch": 2.4414127633772926, + "grad_norm": 0.18692985874064466, + "learning_rate": 4.8541743693002676e-05, + "loss": 0.4576, + "step": 503 + }, + { + "epoch": 2.4462634530847356, + "grad_norm": 0.17489245993778632, + "learning_rate": 4.8409262220522196e-05, + "loss": 0.4476, + "step": 504 + }, + { + "epoch": 2.451114142792178, + "grad_norm": 0.14061959670906948, + "learning_rate": 4.8276684165694336e-05, + "loss": 0.4479, + "step": 505 + }, + { + "epoch": 2.455964832499621, + "grad_norm": 0.17289206898304424, + "learning_rate": 4.814401105120914e-05, + "loss": 0.4479, + "step": 506 + }, + { + "epoch": 2.460815522207064, + "grad_norm": 0.21063084112901795, + "learning_rate": 4.8011244400848414e-05, + "loss": 0.4466, + "step": 507 + }, + { + "epoch": 2.4656662119145065, + "grad_norm": 0.2134287283629687, + "learning_rate": 4.787838573946825e-05, + "loss": 0.4503, + "step": 508 + }, + { + "epoch": 2.4705169016219495, + "grad_norm": 0.19387557882251144, + "learning_rate": 4.774543659298152e-05, + "loss": 0.4419, + "step": 509 + }, + { + "epoch": 2.475367591329392, + "grad_norm": 0.1690053079886072, + "learning_rate": 4.761239848834031e-05, + "loss": 0.4443, + "step": 510 + }, + { + "epoch": 2.480218281036835, + "grad_norm": 0.17406180900609755, + "learning_rate": 4.747927295351845e-05, + "loss": 0.4474, + "step": 511 + }, + { + "epoch": 2.485068970744278, + "grad_norm": 0.2024050850623432, + "learning_rate": 4.734606151749389e-05, + "loss": 0.4473, + "step": 512 + }, + { + "epoch": 2.4899196604517204, + "grad_norm": 0.22659357369802574, + "learning_rate": 4.7212765710231204e-05, + "loss": 0.4481, + "step": 513 + }, + { + "epoch": 2.4947703501591634, + "grad_norm": 0.23595309939097722, + "learning_rate": 4.707938706266397e-05, + "loss": 0.4484, + "step": 514 + }, + { + "epoch": 2.499621039866606, + "grad_norm": 0.19918584163751257, + "learning_rate": 4.694592710667723e-05, + "loss": 0.444, + "step": 515 + }, + { + "epoch": 2.504471729574049, + "grad_norm": 0.18418670752131802, + "learning_rate": 4.681238737508983e-05, + "loss": 0.4424, + "step": 516 + }, + { + "epoch": 2.5093224192814914, + "grad_norm": 0.1926237495649244, + "learning_rate": 4.6678769401636894e-05, + "loss": 0.4444, + "step": 517 + }, + { + "epoch": 2.5141731089889343, + "grad_norm": 0.20706125086296728, + "learning_rate": 4.6545074720952166e-05, + "loss": 0.456, + "step": 518 + }, + { + "epoch": 2.5190237986963773, + "grad_norm": 0.17699664563372686, + "learning_rate": 4.641130486855038e-05, + "loss": 0.4396, + "step": 519 + }, + { + "epoch": 2.52387448840382, + "grad_norm": 0.18317752602670304, + "learning_rate": 4.627746138080966e-05, + "loss": 0.4432, + "step": 520 + }, + { + "epoch": 2.5287251781112627, + "grad_norm": 0.2190424482227647, + "learning_rate": 4.614354579495379e-05, + "loss": 0.4448, + "step": 521 + }, + { + "epoch": 2.5335758678187057, + "grad_norm": 0.20135719119048615, + "learning_rate": 4.6009559649034695e-05, + "loss": 0.4432, + "step": 522 + }, + { + "epoch": 2.538426557526148, + "grad_norm": 0.20782420010728125, + "learning_rate": 4.587550448191465e-05, + "loss": 0.4474, + "step": 523 + }, + { + "epoch": 2.543277247233591, + "grad_norm": 0.18668296726800496, + "learning_rate": 4.5741381833248655e-05, + "loss": 0.455, + "step": 524 + }, + { + "epoch": 2.5481279369410337, + "grad_norm": 0.17935132627421838, + "learning_rate": 4.560719324346677e-05, + "loss": 0.4457, + "step": 525 + }, + { + "epoch": 2.5529786266484766, + "grad_norm": 0.16835981784522308, + "learning_rate": 4.547294025375641e-05, + "loss": 0.4478, + "step": 526 + }, + { + "epoch": 2.557829316355919, + "grad_norm": 0.18084487512355504, + "learning_rate": 4.533862440604461e-05, + "loss": 0.447, + "step": 527 + }, + { + "epoch": 2.562680006063362, + "grad_norm": 0.17384784743298828, + "learning_rate": 4.520424724298036e-05, + "loss": 0.4408, + "step": 528 + }, + { + "epoch": 2.567530695770805, + "grad_norm": 0.20150460275113774, + "learning_rate": 4.5069810307916874e-05, + "loss": 0.4441, + "step": 529 + }, + { + "epoch": 2.5723813854782476, + "grad_norm": 0.2240004858996321, + "learning_rate": 4.493531514489385e-05, + "loss": 0.4425, + "step": 530 + }, + { + "epoch": 2.5772320751856905, + "grad_norm": 0.2286831099325836, + "learning_rate": 4.480076329861977e-05, + "loss": 0.4433, + "step": 531 + }, + { + "epoch": 2.5820827648931335, + "grad_norm": 0.2048648110357608, + "learning_rate": 4.46661563144541e-05, + "loss": 0.4487, + "step": 532 + }, + { + "epoch": 2.586933454600576, + "grad_norm": 0.2121488428415987, + "learning_rate": 4.453149573838962e-05, + "loss": 0.4445, + "step": 533 + }, + { + "epoch": 2.591784144308019, + "grad_norm": 0.18541474138380978, + "learning_rate": 4.43967831170346e-05, + "loss": 0.4494, + "step": 534 + }, + { + "epoch": 2.5966348340154615, + "grad_norm": 0.17452762694525445, + "learning_rate": 4.426201999759505e-05, + "loss": 0.4484, + "step": 535 + }, + { + "epoch": 2.6014855237229044, + "grad_norm": 0.21944224068377363, + "learning_rate": 4.4127207927857e-05, + "loss": 0.4419, + "step": 536 + }, + { + "epoch": 2.606336213430347, + "grad_norm": 0.17656272332454842, + "learning_rate": 4.3992348456168666e-05, + "loss": 0.4568, + "step": 537 + }, + { + "epoch": 2.61118690313779, + "grad_norm": 0.1892562653364182, + "learning_rate": 4.385744313142267e-05, + "loss": 0.4427, + "step": 538 + }, + { + "epoch": 2.616037592845233, + "grad_norm": 0.21611454670373548, + "learning_rate": 4.372249350303828e-05, + "loss": 0.4418, + "step": 539 + }, + { + "epoch": 2.6208882825526754, + "grad_norm": 0.17168747953224547, + "learning_rate": 4.358750112094363e-05, + "loss": 0.4544, + "step": 540 + }, + { + "epoch": 2.6257389722601183, + "grad_norm": 0.17941819138400728, + "learning_rate": 4.3452467535557846e-05, + "loss": 0.4372, + "step": 541 + }, + { + "epoch": 2.6305896619675613, + "grad_norm": 0.2025265834742146, + "learning_rate": 4.3317394297773304e-05, + "loss": 0.4517, + "step": 542 + }, + { + "epoch": 2.635440351675004, + "grad_norm": 0.20441246530938206, + "learning_rate": 4.3182282958937816e-05, + "loss": 0.4333, + "step": 543 + }, + { + "epoch": 2.6402910413824463, + "grad_norm": 0.2334105452950634, + "learning_rate": 4.304713507083673e-05, + "loss": 0.4481, + "step": 544 + }, + { + "epoch": 2.6451417310898893, + "grad_norm": 0.26291969340773214, + "learning_rate": 4.291195218567523e-05, + "loss": 0.4466, + "step": 545 + }, + { + "epoch": 2.649992420797332, + "grad_norm": 0.1863631298156993, + "learning_rate": 4.277673585606046e-05, + "loss": 0.4405, + "step": 546 + }, + { + "epoch": 2.6548431105047747, + "grad_norm": 0.23226855973797117, + "learning_rate": 4.264148763498364e-05, + "loss": 0.4566, + "step": 547 + }, + { + "epoch": 2.6596938002122177, + "grad_norm": 0.30482274820740174, + "learning_rate": 4.250620907580226e-05, + "loss": 0.4407, + "step": 548 + }, + { + "epoch": 2.6645444899196606, + "grad_norm": 0.23781311620065457, + "learning_rate": 4.237090173222231e-05, + "loss": 0.4493, + "step": 549 + }, + { + "epoch": 2.669395179627103, + "grad_norm": 0.1808214801234254, + "learning_rate": 4.223556715828033e-05, + "loss": 0.4511, + "step": 550 + }, + { + "epoch": 2.674245869334546, + "grad_norm": 0.26315804734468673, + "learning_rate": 4.2100206908325603e-05, + "loss": 0.447, + "step": 551 + }, + { + "epoch": 2.6790965590419886, + "grad_norm": 0.25781234163394623, + "learning_rate": 4.196482253700235e-05, + "loss": 0.4415, + "step": 552 + }, + { + "epoch": 2.6839472487494316, + "grad_norm": 0.17133762584152984, + "learning_rate": 4.182941559923179e-05, + "loss": 0.4457, + "step": 553 + }, + { + "epoch": 2.688797938456874, + "grad_norm": 0.2266803612041648, + "learning_rate": 4.169398765019433e-05, + "loss": 0.4422, + "step": 554 + }, + { + "epoch": 2.693648628164317, + "grad_norm": 0.23286738752123257, + "learning_rate": 4.15585402453117e-05, + "loss": 0.4429, + "step": 555 + }, + { + "epoch": 2.69849931787176, + "grad_norm": 0.20226496811604636, + "learning_rate": 4.14230749402291e-05, + "loss": 0.4421, + "step": 556 + }, + { + "epoch": 2.7033500075792025, + "grad_norm": 0.21746634743317236, + "learning_rate": 4.128759329079732e-05, + "loss": 0.4318, + "step": 557 + }, + { + "epoch": 2.7082006972866455, + "grad_norm": 0.24285493960537577, + "learning_rate": 4.115209685305482e-05, + "loss": 0.4374, + "step": 558 + }, + { + "epoch": 2.7130513869940884, + "grad_norm": 0.20035101285126697, + "learning_rate": 4.101658718320998e-05, + "loss": 0.4429, + "step": 559 + }, + { + "epoch": 2.717902076701531, + "grad_norm": 0.1733102653989901, + "learning_rate": 4.088106583762309e-05, + "loss": 0.4456, + "step": 560 + }, + { + "epoch": 2.722752766408974, + "grad_norm": 0.25116764609287723, + "learning_rate": 4.074553437278857e-05, + "loss": 0.4494, + "step": 561 + }, + { + "epoch": 2.7276034561164164, + "grad_norm": 0.19896329775589092, + "learning_rate": 4.060999434531704e-05, + "loss": 0.4449, + "step": 562 + }, + { + "epoch": 2.7324541458238594, + "grad_norm": 0.16517536428811208, + "learning_rate": 4.047444731191751e-05, + "loss": 0.4426, + "step": 563 + }, + { + "epoch": 2.737304835531302, + "grad_norm": 0.1656807626865065, + "learning_rate": 4.033889482937943e-05, + "loss": 0.4445, + "step": 564 + }, + { + "epoch": 2.742155525238745, + "grad_norm": 0.15103159619749504, + "learning_rate": 4.020333845455478e-05, + "loss": 0.4565, + "step": 565 + }, + { + "epoch": 2.747006214946188, + "grad_norm": 0.16996820086522443, + "learning_rate": 4.0067779744340345e-05, + "loss": 0.4459, + "step": 566 + }, + { + "epoch": 2.7518569046536303, + "grad_norm": 0.1495970266083701, + "learning_rate": 3.993222025565966e-05, + "loss": 0.4447, + "step": 567 + }, + { + "epoch": 2.7567075943610733, + "grad_norm": 0.15458974892236554, + "learning_rate": 3.979666154544522e-05, + "loss": 0.4452, + "step": 568 + }, + { + "epoch": 2.7615582840685162, + "grad_norm": 0.1741093401099396, + "learning_rate": 3.96611051706206e-05, + "loss": 0.4421, + "step": 569 + }, + { + "epoch": 2.7664089737759587, + "grad_norm": 0.1819530197226333, + "learning_rate": 3.9525552688082494e-05, + "loss": 0.4509, + "step": 570 + }, + { + "epoch": 2.7712596634834017, + "grad_norm": 0.14996389947080183, + "learning_rate": 3.939000565468297e-05, + "loss": 0.4442, + "step": 571 + }, + { + "epoch": 2.776110353190844, + "grad_norm": 0.19015205226216172, + "learning_rate": 3.9254465627211444e-05, + "loss": 0.4458, + "step": 572 + }, + { + "epoch": 2.780961042898287, + "grad_norm": 0.2147271939320094, + "learning_rate": 3.911893416237693e-05, + "loss": 0.4423, + "step": 573 + }, + { + "epoch": 2.7858117326057297, + "grad_norm": 0.17368493707493848, + "learning_rate": 3.8983412816790045e-05, + "loss": 0.4415, + "step": 574 + }, + { + "epoch": 2.7906624223131726, + "grad_norm": 0.18366782229193682, + "learning_rate": 3.8847903146945186e-05, + "loss": 0.4419, + "step": 575 + }, + { + "epoch": 2.7955131120206156, + "grad_norm": 0.1770373720928735, + "learning_rate": 3.871240670920269e-05, + "loss": 0.4477, + "step": 576 + }, + { + "epoch": 2.800363801728058, + "grad_norm": 0.15508137985676013, + "learning_rate": 3.85769250597709e-05, + "loss": 0.4458, + "step": 577 + }, + { + "epoch": 2.805214491435501, + "grad_norm": 0.15518418619016236, + "learning_rate": 3.844145975468832e-05, + "loss": 0.4403, + "step": 578 + }, + { + "epoch": 2.810065181142944, + "grad_norm": 0.13016021632650948, + "learning_rate": 3.830601234980569e-05, + "loss": 0.4509, + "step": 579 + }, + { + "epoch": 2.8149158708503865, + "grad_norm": 0.17125198843453068, + "learning_rate": 3.8170584400768224e-05, + "loss": 0.4492, + "step": 580 + }, + { + "epoch": 2.8197665605578295, + "grad_norm": 0.18182599605587274, + "learning_rate": 3.8035177462997664e-05, + "loss": 0.4475, + "step": 581 + }, + { + "epoch": 2.824617250265272, + "grad_norm": 0.16612208849559923, + "learning_rate": 3.7899793091674396e-05, + "loss": 0.4419, + "step": 582 + }, + { + "epoch": 2.829467939972715, + "grad_norm": 0.15504382692612345, + "learning_rate": 3.776443284171969e-05, + "loss": 0.4421, + "step": 583 + }, + { + "epoch": 2.8343186296801575, + "grad_norm": 0.14158493240403466, + "learning_rate": 3.7629098267777706e-05, + "loss": 0.4399, + "step": 584 + }, + { + "epoch": 2.8391693193876004, + "grad_norm": 0.14521790840725082, + "learning_rate": 3.7493790924197746e-05, + "loss": 0.4328, + "step": 585 + }, + { + "epoch": 2.8440200090950434, + "grad_norm": 0.14223117063886642, + "learning_rate": 3.735851236501637e-05, + "loss": 0.4403, + "step": 586 + }, + { + "epoch": 2.848870698802486, + "grad_norm": 0.13716153813556975, + "learning_rate": 3.722326414393954e-05, + "loss": 0.4375, + "step": 587 + }, + { + "epoch": 2.853721388509929, + "grad_norm": 0.13481835773066367, + "learning_rate": 3.708804781432478e-05, + "loss": 0.4465, + "step": 588 + }, + { + "epoch": 2.858572078217372, + "grad_norm": 0.15098861155914894, + "learning_rate": 3.6952864929163286e-05, + "loss": 0.4478, + "step": 589 + }, + { + "epoch": 2.8634227679248143, + "grad_norm": 0.14428029096945075, + "learning_rate": 3.6817717041062204e-05, + "loss": 0.4433, + "step": 590 + }, + { + "epoch": 2.868273457632257, + "grad_norm": 0.14941685817838865, + "learning_rate": 3.66826057022267e-05, + "loss": 0.4426, + "step": 591 + }, + { + "epoch": 2.8731241473397, + "grad_norm": 0.15814915640950794, + "learning_rate": 3.654753246444217e-05, + "loss": 0.437, + "step": 592 + }, + { + "epoch": 2.8779748370471427, + "grad_norm": 0.1666350361712344, + "learning_rate": 3.641249887905638e-05, + "loss": 0.4404, + "step": 593 + }, + { + "epoch": 2.8828255267545853, + "grad_norm": 0.14761581189993908, + "learning_rate": 3.627750649696173e-05, + "loss": 0.4418, + "step": 594 + }, + { + "epoch": 2.887676216462028, + "grad_norm": 0.16433140815176697, + "learning_rate": 3.614255686857734e-05, + "loss": 0.4482, + "step": 595 + }, + { + "epoch": 2.892526906169471, + "grad_norm": 0.14898117787478918, + "learning_rate": 3.600765154383134e-05, + "loss": 0.4407, + "step": 596 + }, + { + "epoch": 2.8973775958769137, + "grad_norm": 0.1576311183401583, + "learning_rate": 3.587279207214301e-05, + "loss": 0.4502, + "step": 597 + }, + { + "epoch": 2.9022282855843566, + "grad_norm": 0.12713739329535242, + "learning_rate": 3.5737980002404965e-05, + "loss": 0.4504, + "step": 598 + }, + { + "epoch": 2.907078975291799, + "grad_norm": 0.1407162797770253, + "learning_rate": 3.5603216882965415e-05, + "loss": 0.4444, + "step": 599 + }, + { + "epoch": 2.911929664999242, + "grad_norm": 0.15096239269549241, + "learning_rate": 3.5468504261610387e-05, + "loss": 0.4497, + "step": 600 + }, + { + "epoch": 2.9167803547066846, + "grad_norm": 0.1341187523707445, + "learning_rate": 3.5333843685545914e-05, + "loss": 0.4449, + "step": 601 + }, + { + "epoch": 2.9216310444141276, + "grad_norm": 0.1619335943958044, + "learning_rate": 3.519923670138025e-05, + "loss": 0.4434, + "step": 602 + }, + { + "epoch": 2.9264817341215705, + "grad_norm": 0.16211928956809254, + "learning_rate": 3.506468485510616e-05, + "loss": 0.4394, + "step": 603 + }, + { + "epoch": 2.931332423829013, + "grad_norm": 0.13565985896740923, + "learning_rate": 3.493018969208314e-05, + "loss": 0.4513, + "step": 604 + }, + { + "epoch": 2.936183113536456, + "grad_norm": 0.14551792578562125, + "learning_rate": 3.479575275701965e-05, + "loss": 0.4425, + "step": 605 + }, + { + "epoch": 2.941033803243899, + "grad_norm": 0.17680002225930672, + "learning_rate": 3.4661375593955405e-05, + "loss": 0.4384, + "step": 606 + }, + { + "epoch": 2.9458844929513415, + "grad_norm": 0.16447736465809054, + "learning_rate": 3.45270597462436e-05, + "loss": 0.4451, + "step": 607 + }, + { + "epoch": 2.9507351826587844, + "grad_norm": 0.15299942108774958, + "learning_rate": 3.4392806756533233e-05, + "loss": 0.4383, + "step": 608 + }, + { + "epoch": 2.955585872366227, + "grad_norm": 0.1536623395643003, + "learning_rate": 3.425861816675135e-05, + "loss": 0.4453, + "step": 609 + }, + { + "epoch": 2.96043656207367, + "grad_norm": 0.15727509499916784, + "learning_rate": 3.4124495518085366e-05, + "loss": 0.436, + "step": 610 + }, + { + "epoch": 2.9652872517811124, + "grad_norm": 0.17906600408772821, + "learning_rate": 3.399044035096532e-05, + "loss": 0.4467, + "step": 611 + }, + { + "epoch": 2.9701379414885554, + "grad_norm": 0.16243556143131102, + "learning_rate": 3.3856454205046223e-05, + "loss": 0.4364, + "step": 612 + }, + { + "epoch": 2.9749886311959983, + "grad_norm": 0.1608114743453715, + "learning_rate": 3.372253861919036e-05, + "loss": 0.4517, + "step": 613 + }, + { + "epoch": 2.979839320903441, + "grad_norm": 0.16249676695986184, + "learning_rate": 3.3588695131449626e-05, + "loss": 0.4464, + "step": 614 + }, + { + "epoch": 2.984690010610884, + "grad_norm": 0.14682205641070967, + "learning_rate": 3.3454925279047854e-05, + "loss": 0.4446, + "step": 615 + }, + { + "epoch": 2.9895407003183267, + "grad_norm": 0.158750060737996, + "learning_rate": 3.3321230598363126e-05, + "loss": 0.4449, + "step": 616 + }, + { + "epoch": 2.9943913900257693, + "grad_norm": 0.15103569523913019, + "learning_rate": 3.3187612624910185e-05, + "loss": 0.4457, + "step": 617 + }, + { + "epoch": 2.999242079733212, + "grad_norm": 0.13279281722008607, + "learning_rate": 3.305407289332279e-05, + "loss": 0.4524, + "step": 618 + }, + { + "epoch": 3.004850689707443, + "grad_norm": 0.21786343264375474, + "learning_rate": 3.2920612937336035e-05, + "loss": 0.4186, + "step": 619 + }, + { + "epoch": 3.0097013794148855, + "grad_norm": 0.1940657522375173, + "learning_rate": 3.2787234289768816e-05, + "loss": 0.4198, + "step": 620 + }, + { + "epoch": 3.0145520691223284, + "grad_norm": 0.21773031084989108, + "learning_rate": 3.2653938482506125e-05, + "loss": 0.4257, + "step": 621 + }, + { + "epoch": 3.019402758829771, + "grad_norm": 0.20326547030411232, + "learning_rate": 3.252072704648157e-05, + "loss": 0.4165, + "step": 622 + }, + { + "epoch": 3.024253448537214, + "grad_norm": 0.1952882443077655, + "learning_rate": 3.2387601511659695e-05, + "loss": 0.4099, + "step": 623 + }, + { + "epoch": 3.029104138244657, + "grad_norm": 0.1918042514566916, + "learning_rate": 3.22545634070185e-05, + "loss": 0.4152, + "step": 624 + }, + { + "epoch": 3.0339548279520994, + "grad_norm": 0.19705116304049744, + "learning_rate": 3.212161426053177e-05, + "loss": 0.4128, + "step": 625 + }, + { + "epoch": 3.0388055176595423, + "grad_norm": 0.19649678060258355, + "learning_rate": 3.19887555991516e-05, + "loss": 0.4129, + "step": 626 + }, + { + "epoch": 3.043656207366985, + "grad_norm": 0.21708556979200094, + "learning_rate": 3.1855988948790866e-05, + "loss": 0.419, + "step": 627 + }, + { + "epoch": 3.048506897074428, + "grad_norm": 0.1975715363358727, + "learning_rate": 3.172331583430567e-05, + "loss": 0.4179, + "step": 628 + }, + { + "epoch": 3.0533575867818707, + "grad_norm": 0.23335868145268443, + "learning_rate": 3.1590737779477825e-05, + "loss": 0.4187, + "step": 629 + }, + { + "epoch": 3.0582082764893133, + "grad_norm": 0.16312780113553543, + "learning_rate": 3.145825630699734e-05, + "loss": 0.4145, + "step": 630 + }, + { + "epoch": 3.063058966196756, + "grad_norm": 0.1995031684584169, + "learning_rate": 3.1325872938444995e-05, + "loss": 0.4269, + "step": 631 + }, + { + "epoch": 3.0679096559041987, + "grad_norm": 0.16321617015370155, + "learning_rate": 3.119358919427478e-05, + "loss": 0.4189, + "step": 632 + }, + { + "epoch": 3.0727603456116417, + "grad_norm": 0.18410001775256316, + "learning_rate": 3.106140659379652e-05, + "loss": 0.4218, + "step": 633 + }, + { + "epoch": 3.0776110353190846, + "grad_norm": 0.15016856412332835, + "learning_rate": 3.092932665515837e-05, + "loss": 0.4142, + "step": 634 + }, + { + "epoch": 3.082461725026527, + "grad_norm": 0.1502499895895468, + "learning_rate": 3.079735089532935e-05, + "loss": 0.4092, + "step": 635 + }, + { + "epoch": 3.08731241473397, + "grad_norm": 0.1564011362374121, + "learning_rate": 3.0665480830082e-05, + "loss": 0.4176, + "step": 636 + }, + { + "epoch": 3.0921631044414126, + "grad_norm": 0.15490177035198396, + "learning_rate": 3.0533717973974924e-05, + "loss": 0.4171, + "step": 637 + }, + { + "epoch": 3.0970137941488556, + "grad_norm": 0.16842243538179394, + "learning_rate": 3.040206384033542e-05, + "loss": 0.4142, + "step": 638 + }, + { + "epoch": 3.101864483856298, + "grad_norm": 0.15647723657119908, + "learning_rate": 3.0270519941242052e-05, + "loss": 0.4045, + "step": 639 + }, + { + "epoch": 3.106715173563741, + "grad_norm": 0.14187950473293476, + "learning_rate": 3.0139087787507323e-05, + "loss": 0.4162, + "step": 640 + }, + { + "epoch": 3.111565863271184, + "grad_norm": 0.1486994798612613, + "learning_rate": 3.0007768888660337e-05, + "loss": 0.4162, + "step": 641 + }, + { + "epoch": 3.1164165529786265, + "grad_norm": 0.14045982542928215, + "learning_rate": 2.9876564752929406e-05, + "loss": 0.423, + "step": 642 + }, + { + "epoch": 3.1212672426860695, + "grad_norm": 0.14911944405394412, + "learning_rate": 2.9745476887224806e-05, + "loss": 0.4186, + "step": 643 + }, + { + "epoch": 3.1261179323935124, + "grad_norm": 0.130768356118423, + "learning_rate": 2.961450679712135e-05, + "loss": 0.4149, + "step": 644 + }, + { + "epoch": 3.130968622100955, + "grad_norm": 0.15344876789210227, + "learning_rate": 2.9483655986841265e-05, + "loss": 0.4185, + "step": 645 + }, + { + "epoch": 3.135819311808398, + "grad_norm": 0.16373681347456412, + "learning_rate": 2.9352925959236732e-05, + "loss": 0.4199, + "step": 646 + }, + { + "epoch": 3.1406700015158404, + "grad_norm": 0.13681917154319687, + "learning_rate": 2.92223182157728e-05, + "loss": 0.4187, + "step": 647 + }, + { + "epoch": 3.1455206912232834, + "grad_norm": 0.16376916195733163, + "learning_rate": 2.909183425650996e-05, + "loss": 0.4144, + "step": 648 + }, + { + "epoch": 3.150371380930726, + "grad_norm": 0.15231760580025508, + "learning_rate": 2.8961475580087108e-05, + "loss": 0.4065, + "step": 649 + }, + { + "epoch": 3.155222070638169, + "grad_norm": 0.1482250629623317, + "learning_rate": 2.8831243683704162e-05, + "loss": 0.4167, + "step": 650 + }, + { + "epoch": 3.160072760345612, + "grad_norm": 0.1279306947970543, + "learning_rate": 2.8701140063104996e-05, + "loss": 0.4163, + "step": 651 + }, + { + "epoch": 3.1649234500530543, + "grad_norm": 0.14839352063841615, + "learning_rate": 2.857116621256018e-05, + "loss": 0.4066, + "step": 652 + }, + { + "epoch": 3.1697741397604973, + "grad_norm": 0.14182099741632095, + "learning_rate": 2.8441323624849827e-05, + "loss": 0.4073, + "step": 653 + }, + { + "epoch": 3.17462482946794, + "grad_norm": 0.14062989686711488, + "learning_rate": 2.83116137912465e-05, + "loss": 0.4148, + "step": 654 + }, + { + "epoch": 3.1794755191753827, + "grad_norm": 0.13350262763245205, + "learning_rate": 2.8182038201498038e-05, + "loss": 0.4185, + "step": 655 + }, + { + "epoch": 3.1843262088828257, + "grad_norm": 0.14000623741594295, + "learning_rate": 2.8052598343810474e-05, + "loss": 0.4084, + "step": 656 + }, + { + "epoch": 3.189176898590268, + "grad_norm": 0.13922618624994615, + "learning_rate": 2.7923295704830868e-05, + "loss": 0.4209, + "step": 657 + }, + { + "epoch": 3.194027588297711, + "grad_norm": 0.13424360868636123, + "learning_rate": 2.7794131769630355e-05, + "loss": 0.4203, + "step": 658 + }, + { + "epoch": 3.1988782780051537, + "grad_norm": 0.1410971557670894, + "learning_rate": 2.7665108021687007e-05, + "loss": 0.4229, + "step": 659 + }, + { + "epoch": 3.2037289677125966, + "grad_norm": 0.15216131334925778, + "learning_rate": 2.753622594286879e-05, + "loss": 0.4145, + "step": 660 + }, + { + "epoch": 3.2085796574200396, + "grad_norm": 0.1261567069468345, + "learning_rate": 2.7407487013416615e-05, + "loss": 0.4083, + "step": 661 + }, + { + "epoch": 3.213430347127482, + "grad_norm": 0.15853568579534694, + "learning_rate": 2.727889271192722e-05, + "loss": 0.4187, + "step": 662 + }, + { + "epoch": 3.218281036834925, + "grad_norm": 0.1151093763460037, + "learning_rate": 2.715044451533631e-05, + "loss": 0.4164, + "step": 663 + }, + { + "epoch": 3.2231317265423676, + "grad_norm": 0.14296777545925654, + "learning_rate": 2.702214389890152e-05, + "loss": 0.413, + "step": 664 + }, + { + "epoch": 3.2279824162498105, + "grad_norm": 0.12844538251966497, + "learning_rate": 2.6893992336185512e-05, + "loss": 0.4035, + "step": 665 + }, + { + "epoch": 3.2328331059572535, + "grad_norm": 0.13036276585432338, + "learning_rate": 2.6765991299039025e-05, + "loss": 0.4145, + "step": 666 + }, + { + "epoch": 3.237683795664696, + "grad_norm": 0.12243507743746514, + "learning_rate": 2.663814225758393e-05, + "loss": 0.4117, + "step": 667 + }, + { + "epoch": 3.242534485372139, + "grad_norm": 0.14245972122975337, + "learning_rate": 2.6510446680196448e-05, + "loss": 0.4195, + "step": 668 + }, + { + "epoch": 3.2473851750795815, + "grad_norm": 0.12799039988824565, + "learning_rate": 2.638290603349023e-05, + "loss": 0.4203, + "step": 669 + }, + { + "epoch": 3.2522358647870244, + "grad_norm": 0.12961756855915293, + "learning_rate": 2.625552178229949e-05, + "loss": 0.4159, + "step": 670 + }, + { + "epoch": 3.2570865544944674, + "grad_norm": 0.1296944618260485, + "learning_rate": 2.612829538966218e-05, + "loss": 0.4111, + "step": 671 + }, + { + "epoch": 3.26193724420191, + "grad_norm": 0.14508116679367689, + "learning_rate": 2.6001228316803256e-05, + "loss": 0.4196, + "step": 672 + }, + { + "epoch": 3.266787933909353, + "grad_norm": 0.1216509600449271, + "learning_rate": 2.5874322023117824e-05, + "loss": 0.4162, + "step": 673 + }, + { + "epoch": 3.2716386236167954, + "grad_norm": 0.1407698696044201, + "learning_rate": 2.5747577966154404e-05, + "loss": 0.4165, + "step": 674 + }, + { + "epoch": 3.2764893133242383, + "grad_norm": 0.1142728140236527, + "learning_rate": 2.5620997601598215e-05, + "loss": 0.4076, + "step": 675 + }, + { + "epoch": 3.281340003031681, + "grad_norm": 0.14212475123175447, + "learning_rate": 2.5494582383254388e-05, + "loss": 0.4174, + "step": 676 + }, + { + "epoch": 3.286190692739124, + "grad_norm": 0.11798282133422631, + "learning_rate": 2.5368333763031324e-05, + "loss": 0.4131, + "step": 677 + }, + { + "epoch": 3.2910413824465667, + "grad_norm": 0.14618443390836333, + "learning_rate": 2.5242253190924034e-05, + "loss": 0.4092, + "step": 678 + }, + { + "epoch": 3.2958920721540093, + "grad_norm": 0.12815235524330332, + "learning_rate": 2.5116342114997442e-05, + "loss": 0.409, + "step": 679 + }, + { + "epoch": 3.300742761861452, + "grad_norm": 0.1346079173937725, + "learning_rate": 2.4990601981369737e-05, + "loss": 0.4201, + "step": 680 + }, + { + "epoch": 3.305593451568895, + "grad_norm": 0.12020401678661803, + "learning_rate": 2.4865034234195834e-05, + "loss": 0.4107, + "step": 681 + }, + { + "epoch": 3.3104441412763377, + "grad_norm": 0.1342860022847603, + "learning_rate": 2.4739640315650747e-05, + "loss": 0.4145, + "step": 682 + }, + { + "epoch": 3.3152948309837806, + "grad_norm": 0.11394559661374248, + "learning_rate": 2.4614421665912997e-05, + "loss": 0.4213, + "step": 683 + }, + { + "epoch": 3.320145520691223, + "grad_norm": 0.126805017438777, + "learning_rate": 2.4489379723148147e-05, + "loss": 0.4129, + "step": 684 + }, + { + "epoch": 3.324996210398666, + "grad_norm": 0.11817811298525939, + "learning_rate": 2.4364515923492187e-05, + "loss": 0.4193, + "step": 685 + }, + { + "epoch": 3.3298469001061086, + "grad_norm": 0.1342332761111468, + "learning_rate": 2.4239831701035143e-05, + "loss": 0.418, + "step": 686 + }, + { + "epoch": 3.3346975898135516, + "grad_norm": 0.12474527460295737, + "learning_rate": 2.411532848780451e-05, + "loss": 0.4166, + "step": 687 + }, + { + "epoch": 3.3395482795209945, + "grad_norm": 0.1329059048345405, + "learning_rate": 2.399100771374888e-05, + "loss": 0.4138, + "step": 688 + }, + { + "epoch": 3.344398969228437, + "grad_norm": 0.11628441384991241, + "learning_rate": 2.3866870806721495e-05, + "loss": 0.4111, + "step": 689 + }, + { + "epoch": 3.34924965893588, + "grad_norm": 0.1453864432398833, + "learning_rate": 2.37429191924638e-05, + "loss": 0.42, + "step": 690 + }, + { + "epoch": 3.354100348643323, + "grad_norm": 0.11264715608612798, + "learning_rate": 2.361915429458913e-05, + "loss": 0.417, + "step": 691 + }, + { + "epoch": 3.3589510383507655, + "grad_norm": 0.14248256005839127, + "learning_rate": 2.349557753456637e-05, + "loss": 0.4168, + "step": 692 + }, + { + "epoch": 3.3638017280582084, + "grad_norm": 0.12398727462550883, + "learning_rate": 2.3372190331703556e-05, + "loss": 0.4189, + "step": 693 + }, + { + "epoch": 3.368652417765651, + "grad_norm": 0.11869463547338541, + "learning_rate": 2.324899410313161e-05, + "loss": 0.4125, + "step": 694 + }, + { + "epoch": 3.373503107473094, + "grad_norm": 0.12377753865053726, + "learning_rate": 2.3125990263788118e-05, + "loss": 0.4186, + "step": 695 + }, + { + "epoch": 3.3783537971805364, + "grad_norm": 0.13141561076446406, + "learning_rate": 2.3003180226400986e-05, + "loss": 0.4123, + "step": 696 + }, + { + "epoch": 3.3832044868879794, + "grad_norm": 0.11816327910326484, + "learning_rate": 2.288056540147229e-05, + "loss": 0.4129, + "step": 697 + }, + { + "epoch": 3.3880551765954223, + "grad_norm": 0.12260669143866527, + "learning_rate": 2.275814719726201e-05, + "loss": 0.4133, + "step": 698 + }, + { + "epoch": 3.392905866302865, + "grad_norm": 0.133084483131333, + "learning_rate": 2.263592701977193e-05, + "loss": 0.4219, + "step": 699 + }, + { + "epoch": 3.397756556010308, + "grad_norm": 0.11848073628628028, + "learning_rate": 2.2513906272729397e-05, + "loss": 0.4143, + "step": 700 + }, + { + "epoch": 3.4026072457177503, + "grad_norm": 0.12858950370510128, + "learning_rate": 2.239208635757133e-05, + "loss": 0.4166, + "step": 701 + }, + { + "epoch": 3.4074579354251933, + "grad_norm": 0.11722692816596028, + "learning_rate": 2.2270468673428004e-05, + "loss": 0.4259, + "step": 702 + }, + { + "epoch": 3.412308625132636, + "grad_norm": 0.11830608786302087, + "learning_rate": 2.2149054617106974e-05, + "loss": 0.407, + "step": 703 + }, + { + "epoch": 3.4171593148400787, + "grad_norm": 0.12285165179780408, + "learning_rate": 2.2027845583077175e-05, + "loss": 0.4231, + "step": 704 + }, + { + "epoch": 3.4220100045475217, + "grad_norm": 0.12884220459818324, + "learning_rate": 2.1906842963452757e-05, + "loss": 0.4069, + "step": 705 + }, + { + "epoch": 3.426860694254964, + "grad_norm": 0.10700415399515635, + "learning_rate": 2.178604814797715e-05, + "loss": 0.4149, + "step": 706 + }, + { + "epoch": 3.431711383962407, + "grad_norm": 0.11853515304365536, + "learning_rate": 2.1665462524007162e-05, + "loss": 0.4125, + "step": 707 + }, + { + "epoch": 3.43656207366985, + "grad_norm": 0.10772813217204756, + "learning_rate": 2.1545087476496903e-05, + "loss": 0.4216, + "step": 708 + }, + { + "epoch": 3.4414127633772926, + "grad_norm": 0.12946073793938545, + "learning_rate": 2.1424924387981996e-05, + "loss": 0.4227, + "step": 709 + }, + { + "epoch": 3.4462634530847356, + "grad_norm": 0.11256888467579416, + "learning_rate": 2.1304974638563715e-05, + "loss": 0.4116, + "step": 710 + }, + { + "epoch": 3.451114142792178, + "grad_norm": 0.12334394272902899, + "learning_rate": 2.1185239605893013e-05, + "loss": 0.4217, + "step": 711 + }, + { + "epoch": 3.455964832499621, + "grad_norm": 0.11467056354627854, + "learning_rate": 2.106572066515482e-05, + "loss": 0.4193, + "step": 712 + }, + { + "epoch": 3.460815522207064, + "grad_norm": 0.11248104529388973, + "learning_rate": 2.0946419189052162e-05, + "loss": 0.4147, + "step": 713 + }, + { + "epoch": 3.4656662119145065, + "grad_norm": 0.11839620237634582, + "learning_rate": 2.0827336547790452e-05, + "loss": 0.4214, + "step": 714 + }, + { + "epoch": 3.4705169016219495, + "grad_norm": 0.11350680266976185, + "learning_rate": 2.0708474109061752e-05, + "loss": 0.416, + "step": 715 + }, + { + "epoch": 3.475367591329392, + "grad_norm": 0.11310394993568851, + "learning_rate": 2.0589833238029032e-05, + "loss": 0.4001, + "step": 716 + }, + { + "epoch": 3.480218281036835, + "grad_norm": 0.1142115014102716, + "learning_rate": 2.0471415297310455e-05, + "loss": 0.4158, + "step": 717 + }, + { + "epoch": 3.485068970744278, + "grad_norm": 0.12114855144347077, + "learning_rate": 2.0353221646963864e-05, + "loss": 0.41, + "step": 718 + }, + { + "epoch": 3.4899196604517204, + "grad_norm": 0.11155618948058961, + "learning_rate": 2.0235253644471012e-05, + "loss": 0.4226, + "step": 719 + }, + { + "epoch": 3.4947703501591634, + "grad_norm": 0.12346483161916652, + "learning_rate": 2.011751264472206e-05, + "loss": 0.4163, + "step": 720 + }, + { + "epoch": 3.499621039866606, + "grad_norm": 0.1314504891806893, + "learning_rate": 2.0000000000000012e-05, + "loss": 0.4185, + "step": 721 + }, + { + "epoch": 3.504471729574049, + "grad_norm": 0.11587490363769114, + "learning_rate": 1.9882717059965086e-05, + "loss": 0.4243, + "step": 722 + }, + { + "epoch": 3.5093224192814914, + "grad_norm": 0.147492515392888, + "learning_rate": 1.9765665171639345e-05, + "loss": 0.4139, + "step": 723 + }, + { + "epoch": 3.5141731089889343, + "grad_norm": 0.13161876983358525, + "learning_rate": 1.964884567939118e-05, + "loss": 0.4023, + "step": 724 + }, + { + "epoch": 3.5190237986963773, + "grad_norm": 0.1288012773595963, + "learning_rate": 1.9532259924919823e-05, + "loss": 0.416, + "step": 725 + }, + { + "epoch": 3.52387448840382, + "grad_norm": 0.10979902623938809, + "learning_rate": 1.9415909247239996e-05, + "loss": 0.4198, + "step": 726 + }, + { + "epoch": 3.5287251781112627, + "grad_norm": 0.12273321372911124, + "learning_rate": 1.9299794982666485e-05, + "loss": 0.4084, + "step": 727 + }, + { + "epoch": 3.5335758678187057, + "grad_norm": 0.11485189001172974, + "learning_rate": 1.9183918464798837e-05, + "loss": 0.408, + "step": 728 + }, + { + "epoch": 3.538426557526148, + "grad_norm": 0.1142263107632135, + "learning_rate": 1.906828102450601e-05, + "loss": 0.4131, + "step": 729 + }, + { + "epoch": 3.543277247233591, + "grad_norm": 0.11235720950088848, + "learning_rate": 1.895288398991114e-05, + "loss": 0.4102, + "step": 730 + }, + { + "epoch": 3.5481279369410337, + "grad_norm": 0.11476446551397697, + "learning_rate": 1.8837728686376158e-05, + "loss": 0.4176, + "step": 731 + }, + { + "epoch": 3.5529786266484766, + "grad_norm": 0.12252689326129268, + "learning_rate": 1.8722816436486754e-05, + "loss": 0.4238, + "step": 732 + }, + { + "epoch": 3.557829316355919, + "grad_norm": 0.10742367387279236, + "learning_rate": 1.8608148560037036e-05, + "loss": 0.4248, + "step": 733 + }, + { + "epoch": 3.562680006063362, + "grad_norm": 0.11202485484827082, + "learning_rate": 1.8493726374014442e-05, + "loss": 0.4047, + "step": 734 + }, + { + "epoch": 3.567530695770805, + "grad_norm": 0.10619927645163178, + "learning_rate": 1.8379551192584588e-05, + "loss": 0.4105, + "step": 735 + }, + { + "epoch": 3.5723813854782476, + "grad_norm": 0.1104031551739802, + "learning_rate": 1.826562432707619e-05, + "loss": 0.4158, + "step": 736 + }, + { + "epoch": 3.5772320751856905, + "grad_norm": 0.10920848831138874, + "learning_rate": 1.8151947085965994e-05, + "loss": 0.4157, + "step": 737 + }, + { + "epoch": 3.5820827648931335, + "grad_norm": 0.10703865180962287, + "learning_rate": 1.803852077486377e-05, + "loss": 0.4144, + "step": 738 + }, + { + "epoch": 3.586933454600576, + "grad_norm": 0.1043548569854331, + "learning_rate": 1.7925346696497295e-05, + "loss": 0.4082, + "step": 739 + }, + { + "epoch": 3.591784144308019, + "grad_norm": 0.11138455070453787, + "learning_rate": 1.781242615069733e-05, + "loss": 0.4137, + "step": 740 + }, + { + "epoch": 3.5966348340154615, + "grad_norm": 0.11820322046978973, + "learning_rate": 1.7699760434382853e-05, + "loss": 0.4108, + "step": 741 + }, + { + "epoch": 3.6014855237229044, + "grad_norm": 0.11278678223744326, + "learning_rate": 1.758735084154601e-05, + "loss": 0.4189, + "step": 742 + }, + { + "epoch": 3.606336213430347, + "grad_norm": 0.11323272041806605, + "learning_rate": 1.7475198663237297e-05, + "loss": 0.4123, + "step": 743 + }, + { + "epoch": 3.61118690313779, + "grad_norm": 0.10904989882061365, + "learning_rate": 1.736330518755082e-05, + "loss": 0.4158, + "step": 744 + }, + { + "epoch": 3.616037592845233, + "grad_norm": 0.11229772775053595, + "learning_rate": 1.7251671699609313e-05, + "loss": 0.4182, + "step": 745 + }, + { + "epoch": 3.6208882825526754, + "grad_norm": 0.10951334908001022, + "learning_rate": 1.7140299481549557e-05, + "loss": 0.4213, + "step": 746 + }, + { + "epoch": 3.6257389722601183, + "grad_norm": 0.11434493342191, + "learning_rate": 1.7029189812507603e-05, + "loss": 0.4224, + "step": 747 + }, + { + "epoch": 3.6305896619675613, + "grad_norm": 0.10344385368037658, + "learning_rate": 1.6918343968604027e-05, + "loss": 0.4106, + "step": 748 + }, + { + "epoch": 3.635440351675004, + "grad_norm": 0.11168549168378746, + "learning_rate": 1.6807763222929315e-05, + "loss": 0.408, + "step": 749 + }, + { + "epoch": 3.6402910413824463, + "grad_norm": 0.11594040802990377, + "learning_rate": 1.669744884552926e-05, + "loss": 0.4169, + "step": 750 + }, + { + "epoch": 3.6451417310898893, + "grad_norm": 0.11452998735643415, + "learning_rate": 1.6587402103390314e-05, + "loss": 0.4162, + "step": 751 + }, + { + "epoch": 3.649992420797332, + "grad_norm": 0.11033578052460767, + "learning_rate": 1.6477624260425137e-05, + "loss": 0.4179, + "step": 752 + }, + { + "epoch": 3.6548431105047747, + "grad_norm": 0.1198668006175803, + "learning_rate": 1.6368116577457973e-05, + "loss": 0.4124, + "step": 753 + }, + { + "epoch": 3.6596938002122177, + "grad_norm": 0.10884921001547737, + "learning_rate": 1.6258880312210195e-05, + "loss": 0.4152, + "step": 754 + }, + { + "epoch": 3.6645444899196606, + "grad_norm": 0.12009675101549108, + "learning_rate": 1.6149916719285942e-05, + "loss": 0.4147, + "step": 755 + }, + { + "epoch": 3.669395179627103, + "grad_norm": 0.11742205741912104, + "learning_rate": 1.6041227050157607e-05, + "loss": 0.4096, + "step": 756 + }, + { + "epoch": 3.674245869334546, + "grad_norm": 0.1123251375365744, + "learning_rate": 1.5932812553151506e-05, + "loss": 0.4128, + "step": 757 + }, + { + "epoch": 3.6790965590419886, + "grad_norm": 0.12009728599563303, + "learning_rate": 1.582467447343355e-05, + "loss": 0.4179, + "step": 758 + }, + { + "epoch": 3.6839472487494316, + "grad_norm": 0.10679356522304362, + "learning_rate": 1.5716814052994928e-05, + "loss": 0.4153, + "step": 759 + }, + { + "epoch": 3.688797938456874, + "grad_norm": 0.11287526417521035, + "learning_rate": 1.5609232530637827e-05, + "loss": 0.4065, + "step": 760 + }, + { + "epoch": 3.693648628164317, + "grad_norm": 0.10604174975295146, + "learning_rate": 1.5501931141961278e-05, + "loss": 0.4135, + "step": 761 + }, + { + "epoch": 3.69849931787176, + "grad_norm": 0.10895447717123838, + "learning_rate": 1.539491111934686e-05, + "loss": 0.4102, + "step": 762 + }, + { + "epoch": 3.7033500075792025, + "grad_norm": 0.11522106117016057, + "learning_rate": 1.5288173691944613e-05, + "loss": 0.4193, + "step": 763 + }, + { + "epoch": 3.7082006972866455, + "grad_norm": 0.11668799730823959, + "learning_rate": 1.5181720085658906e-05, + "loss": 0.4131, + "step": 764 + }, + { + "epoch": 3.7130513869940884, + "grad_norm": 0.1179757070897769, + "learning_rate": 1.5075551523134358e-05, + "loss": 0.4107, + "step": 765 + }, + { + "epoch": 3.717902076701531, + "grad_norm": 0.11352197320512793, + "learning_rate": 1.4969669223741771e-05, + "loss": 0.4093, + "step": 766 + }, + { + "epoch": 3.722752766408974, + "grad_norm": 0.11945491424386492, + "learning_rate": 1.4864074403564216e-05, + "loss": 0.4142, + "step": 767 + }, + { + "epoch": 3.7276034561164164, + "grad_norm": 0.10892285815783607, + "learning_rate": 1.4758768275382887e-05, + "loss": 0.4205, + "step": 768 + }, + { + "epoch": 3.7324541458238594, + "grad_norm": 0.12765157069597566, + "learning_rate": 1.4653752048663394e-05, + "loss": 0.412, + "step": 769 + }, + { + "epoch": 3.737304835531302, + "grad_norm": 0.10956587187939422, + "learning_rate": 1.4549026929541693e-05, + "loss": 0.4148, + "step": 770 + }, + { + "epoch": 3.742155525238745, + "grad_norm": 0.12138544117757244, + "learning_rate": 1.4444594120810326e-05, + "loss": 0.4115, + "step": 771 + }, + { + "epoch": 3.747006214946188, + "grad_norm": 0.10697996518645103, + "learning_rate": 1.4340454821904573e-05, + "loss": 0.4194, + "step": 772 + }, + { + "epoch": 3.7518569046536303, + "grad_norm": 0.10998535306961314, + "learning_rate": 1.4236610228888683e-05, + "loss": 0.4143, + "step": 773 + }, + { + "epoch": 3.7567075943610733, + "grad_norm": 0.10998459629638649, + "learning_rate": 1.4133061534442133e-05, + "loss": 0.4121, + "step": 774 + }, + { + "epoch": 3.7615582840685162, + "grad_norm": 0.1056241184630642, + "learning_rate": 1.4029809927845981e-05, + "loss": 0.4146, + "step": 775 + }, + { + "epoch": 3.7664089737759587, + "grad_norm": 0.11617390368347923, + "learning_rate": 1.3926856594969115e-05, + "loss": 0.4125, + "step": 776 + }, + { + "epoch": 3.7712596634834017, + "grad_norm": 0.10105741678809248, + "learning_rate": 1.3824202718254655e-05, + "loss": 0.4081, + "step": 777 + }, + { + "epoch": 3.776110353190844, + "grad_norm": 0.12303653861990478, + "learning_rate": 1.3721849476706477e-05, + "loss": 0.413, + "step": 778 + }, + { + "epoch": 3.780961042898287, + "grad_norm": 0.10128855312982177, + "learning_rate": 1.3619798045875529e-05, + "loss": 0.4117, + "step": 779 + }, + { + "epoch": 3.7858117326057297, + "grad_norm": 0.1118697000457396, + "learning_rate": 1.3518049597846412e-05, + "loss": 0.4097, + "step": 780 + }, + { + "epoch": 3.7906624223131726, + "grad_norm": 0.10689143155371568, + "learning_rate": 1.3416605301223893e-05, + "loss": 0.4146, + "step": 781 + }, + { + "epoch": 3.7955131120206156, + "grad_norm": 0.09942573069367382, + "learning_rate": 1.3315466321119486e-05, + "loss": 0.415, + "step": 782 + }, + { + "epoch": 3.800363801728058, + "grad_norm": 0.112223020708749, + "learning_rate": 1.3214633819138105e-05, + "loss": 0.4187, + "step": 783 + }, + { + "epoch": 3.805214491435501, + "grad_norm": 0.09509376603334437, + "learning_rate": 1.3114108953364655e-05, + "loss": 0.4083, + "step": 784 + }, + { + "epoch": 3.810065181142944, + "grad_norm": 0.1039694903593983, + "learning_rate": 1.3013892878350771e-05, + "loss": 0.415, + "step": 785 + }, + { + "epoch": 3.8149158708503865, + "grad_norm": 0.09799457517015436, + "learning_rate": 1.2913986745101567e-05, + "loss": 0.4082, + "step": 786 + }, + { + "epoch": 3.8197665605578295, + "grad_norm": 0.0994744466895985, + "learning_rate": 1.2814391701062392e-05, + "loss": 0.416, + "step": 787 + }, + { + "epoch": 3.824617250265272, + "grad_norm": 0.09937882084294014, + "learning_rate": 1.2715108890105663e-05, + "loss": 0.4118, + "step": 788 + }, + { + "epoch": 3.829467939972715, + "grad_norm": 0.09366936781008306, + "learning_rate": 1.2616139452517748e-05, + "loss": 0.4202, + "step": 789 + }, + { + "epoch": 3.8343186296801575, + "grad_norm": 0.10103247526201467, + "learning_rate": 1.2517484524985836e-05, + "loss": 0.414, + "step": 790 + }, + { + "epoch": 3.8391693193876004, + "grad_norm": 0.10154281976161658, + "learning_rate": 1.2419145240584856e-05, + "loss": 0.4169, + "step": 791 + }, + { + "epoch": 3.8440200090950434, + "grad_norm": 0.09917419975099584, + "learning_rate": 1.2321122728764566e-05, + "loss": 0.4121, + "step": 792 + }, + { + "epoch": 3.848870698802486, + "grad_norm": 0.10067017561953691, + "learning_rate": 1.222341811533648e-05, + "loss": 0.4177, + "step": 793 + }, + { + "epoch": 3.853721388509929, + "grad_norm": 0.10318548830281854, + "learning_rate": 1.2126032522460975e-05, + "loss": 0.4211, + "step": 794 + }, + { + "epoch": 3.858572078217372, + "grad_norm": 0.09616344349182201, + "learning_rate": 1.2028967068634417e-05, + "loss": 0.4204, + "step": 795 + }, + { + "epoch": 3.8634227679248143, + "grad_norm": 0.10489574786705688, + "learning_rate": 1.193222286867628e-05, + "loss": 0.4119, + "step": 796 + }, + { + "epoch": 3.868273457632257, + "grad_norm": 0.1008569871750799, + "learning_rate": 1.1835801033716372e-05, + "loss": 0.4086, + "step": 797 + }, + { + "epoch": 3.8731241473397, + "grad_norm": 0.09122317803788128, + "learning_rate": 1.1739702671182083e-05, + "loss": 0.4214, + "step": 798 + }, + { + "epoch": 3.8779748370471427, + "grad_norm": 0.0991904412555539, + "learning_rate": 1.1643928884785618e-05, + "loss": 0.4095, + "step": 799 + }, + { + "epoch": 3.8828255267545853, + "grad_norm": 0.0959491624589887, + "learning_rate": 1.1548480774511353e-05, + "loss": 0.4218, + "step": 800 + }, + { + "epoch": 3.887676216462028, + "grad_norm": 0.09541302821779325, + "learning_rate": 1.1453359436603213e-05, + "loss": 0.4218, + "step": 801 + }, + { + "epoch": 3.892526906169471, + "grad_norm": 0.09054297930670004, + "learning_rate": 1.1358565963552039e-05, + "loss": 0.421, + "step": 802 + }, + { + "epoch": 3.8973775958769137, + "grad_norm": 0.09175434223796135, + "learning_rate": 1.126410144408312e-05, + "loss": 0.4088, + "step": 803 + }, + { + "epoch": 3.9022282855843566, + "grad_norm": 0.0933968969265613, + "learning_rate": 1.1169966963143568e-05, + "loss": 0.4105, + "step": 804 + }, + { + "epoch": 3.907078975291799, + "grad_norm": 0.09827723734859882, + "learning_rate": 1.1076163601889953e-05, + "loss": 0.4114, + "step": 805 + }, + { + "epoch": 3.911929664999242, + "grad_norm": 0.09185288445531155, + "learning_rate": 1.098269243767589e-05, + "loss": 0.4099, + "step": 806 + }, + { + "epoch": 3.9167803547066846, + "grad_norm": 0.10344811490009215, + "learning_rate": 1.0889554544039593e-05, + "loss": 0.417, + "step": 807 + }, + { + "epoch": 3.9216310444141276, + "grad_norm": 0.09584223413813131, + "learning_rate": 1.0796750990691596e-05, + "loss": 0.4092, + "step": 808 + }, + { + "epoch": 3.9264817341215705, + "grad_norm": 0.10225202685968936, + "learning_rate": 1.0704282843502459e-05, + "loss": 0.4156, + "step": 809 + }, + { + "epoch": 3.931332423829013, + "grad_norm": 0.10087206295646214, + "learning_rate": 1.0612151164490525e-05, + "loss": 0.4209, + "step": 810 + }, + { + "epoch": 3.936183113536456, + "grad_norm": 0.09495839512641314, + "learning_rate": 1.0520357011809707e-05, + "loss": 0.4193, + "step": 811 + }, + { + "epoch": 3.941033803243899, + "grad_norm": 0.10115287872844174, + "learning_rate": 1.0428901439737387e-05, + "loss": 0.415, + "step": 812 + }, + { + "epoch": 3.9458844929513415, + "grad_norm": 0.09538434619439141, + "learning_rate": 1.0337785498662223e-05, + "loss": 0.4152, + "step": 813 + }, + { + "epoch": 3.9507351826587844, + "grad_norm": 0.0955931932252973, + "learning_rate": 1.024701023507216e-05, + "loss": 0.4153, + "step": 814 + }, + { + "epoch": 3.955585872366227, + "grad_norm": 0.10045702643945939, + "learning_rate": 1.015657669154237e-05, + "loss": 0.4156, + "step": 815 + }, + { + "epoch": 3.96043656207367, + "grad_norm": 0.09335473216661304, + "learning_rate": 1.00664859067233e-05, + "loss": 0.4109, + "step": 816 + }, + { + "epoch": 3.9652872517811124, + "grad_norm": 0.09645722262367523, + "learning_rate": 9.976738915328719e-06, + "loss": 0.4107, + "step": 817 + }, + { + "epoch": 3.9701379414885554, + "grad_norm": 0.09446841411423582, + "learning_rate": 9.887336748123864e-06, + "loss": 0.4178, + "step": 818 + }, + { + "epoch": 3.9749886311959983, + "grad_norm": 0.0903235693705833, + "learning_rate": 9.798280431913558e-06, + "loss": 0.4202, + "step": 819 + }, + { + "epoch": 3.979839320903441, + "grad_norm": 0.0948091272844266, + "learning_rate": 9.709570989530493e-06, + "loss": 0.4123, + "step": 820 + }, + { + "epoch": 3.984690010610884, + "grad_norm": 0.10103242493534337, + "learning_rate": 9.621209439823388e-06, + "loss": 0.4132, + "step": 821 + }, + { + "epoch": 3.9895407003183267, + "grad_norm": 0.09199808648298305, + "learning_rate": 9.533196797645354e-06, + "loss": 0.4101, + "step": 822 + }, + { + "epoch": 3.9943913900257693, + "grad_norm": 0.09866422487619428, + "learning_rate": 9.44553407384221e-06, + "loss": 0.412, + "step": 823 + }, + { + "epoch": 3.999242079733212, + "grad_norm": 0.09950215009808663, + "learning_rate": 9.358222275240884e-06, + "loss": 0.4113, + "step": 824 + }, + { + "epoch": 4.0048506897074425, + "grad_norm": 0.16227668213643984, + "learning_rate": 9.271262404637835e-06, + "loss": 0.4032, + "step": 825 + }, + { + "epoch": 4.009701379414886, + "grad_norm": 0.11430363283166092, + "learning_rate": 9.184655460787591e-06, + "loss": 0.3988, + "step": 826 + }, + { + "epoch": 4.014552069122328, + "grad_norm": 0.11751760908320912, + "learning_rate": 9.098402438391161e-06, + "loss": 0.3943, + "step": 827 + }, + { + "epoch": 4.019402758829771, + "grad_norm": 0.13225951402360853, + "learning_rate": 9.012504328084724e-06, + "loss": 0.4024, + "step": 828 + }, + { + "epoch": 4.0242534485372135, + "grad_norm": 0.1269725744791709, + "learning_rate": 8.926962116428228e-06, + "loss": 0.4, + "step": 829 + }, + { + "epoch": 4.029104138244657, + "grad_norm": 0.12436207204015083, + "learning_rate": 8.841776785894014e-06, + "loss": 0.3994, + "step": 830 + }, + { + "epoch": 4.033954827952099, + "grad_norm": 0.12882961013897004, + "learning_rate": 8.756949314855565e-06, + "loss": 0.3977, + "step": 831 + }, + { + "epoch": 4.038805517659542, + "grad_norm": 0.11685196489455994, + "learning_rate": 8.672480677576267e-06, + "loss": 0.3906, + "step": 832 + }, + { + "epoch": 4.043656207366985, + "grad_norm": 0.106568471380297, + "learning_rate": 8.58837184419821e-06, + "loss": 0.388, + "step": 833 + }, + { + "epoch": 4.048506897074428, + "grad_norm": 0.1134401036269532, + "learning_rate": 8.504623780731056e-06, + "loss": 0.3918, + "step": 834 + }, + { + "epoch": 4.05335758678187, + "grad_norm": 0.12093411095557142, + "learning_rate": 8.421237449040962e-06, + "loss": 0.397, + "step": 835 + }, + { + "epoch": 4.058208276489314, + "grad_norm": 0.11674197768223235, + "learning_rate": 8.338213806839453e-06, + "loss": 0.393, + "step": 836 + }, + { + "epoch": 4.063058966196756, + "grad_norm": 0.10082248897884766, + "learning_rate": 8.255553807672547e-06, + "loss": 0.3897, + "step": 837 + }, + { + "epoch": 4.067909655904199, + "grad_norm": 0.11069126863357161, + "learning_rate": 8.1732584009097e-06, + "loss": 0.3882, + "step": 838 + }, + { + "epoch": 4.072760345611641, + "grad_norm": 0.11738423041846735, + "learning_rate": 8.091328531732925e-06, + "loss": 0.3959, + "step": 839 + }, + { + "epoch": 4.077611035319085, + "grad_norm": 0.10387395387726432, + "learning_rate": 8.009765141126014e-06, + "loss": 0.3891, + "step": 840 + }, + { + "epoch": 4.082461725026527, + "grad_norm": 0.10003982048853301, + "learning_rate": 7.928569165863584e-06, + "loss": 0.3909, + "step": 841 + }, + { + "epoch": 4.08731241473397, + "grad_norm": 0.10549262492820612, + "learning_rate": 7.847741538500439e-06, + "loss": 0.3875, + "step": 842 + }, + { + "epoch": 4.092163104441413, + "grad_norm": 0.10986857014715488, + "learning_rate": 7.767283187360846e-06, + "loss": 0.3929, + "step": 843 + }, + { + "epoch": 4.097013794148856, + "grad_norm": 0.09674503227632227, + "learning_rate": 7.687195036527813e-06, + "loss": 0.3928, + "step": 844 + }, + { + "epoch": 4.101864483856298, + "grad_norm": 0.10305617355184506, + "learning_rate": 7.60747800583252e-06, + "loss": 0.4002, + "step": 845 + }, + { + "epoch": 4.1067151735637415, + "grad_norm": 0.09835145173574618, + "learning_rate": 7.52813301084375e-06, + "loss": 0.3961, + "step": 846 + }, + { + "epoch": 4.111565863271184, + "grad_norm": 0.09426867245890704, + "learning_rate": 7.449160962857358e-06, + "loss": 0.3946, + "step": 847 + }, + { + "epoch": 4.1164165529786265, + "grad_norm": 0.0993123791236732, + "learning_rate": 7.370562768885823e-06, + "loss": 0.3937, + "step": 848 + }, + { + "epoch": 4.121267242686069, + "grad_norm": 0.09692548506896993, + "learning_rate": 7.292339331647848e-06, + "loss": 0.3957, + "step": 849 + }, + { + "epoch": 4.126117932393512, + "grad_norm": 0.0977639150148905, + "learning_rate": 7.214491549557898e-06, + "loss": 0.3969, + "step": 850 + }, + { + "epoch": 4.130968622100955, + "grad_norm": 0.10100358672328608, + "learning_rate": 7.1370203167160326e-06, + "loss": 0.3937, + "step": 851 + }, + { + "epoch": 4.1358193118083975, + "grad_norm": 0.0961508995875077, + "learning_rate": 7.0599265228975e-06, + "loss": 0.3965, + "step": 852 + }, + { + "epoch": 4.140670001515841, + "grad_norm": 0.09450992963252156, + "learning_rate": 6.983211053542591e-06, + "loss": 0.4008, + "step": 853 + }, + { + "epoch": 4.145520691223283, + "grad_norm": 0.10411768031595499, + "learning_rate": 6.9068747897464535e-06, + "loss": 0.4032, + "step": 854 + }, + { + "epoch": 4.150371380930726, + "grad_norm": 0.09981896937364662, + "learning_rate": 6.830918608248964e-06, + "loss": 0.4002, + "step": 855 + }, + { + "epoch": 4.155222070638169, + "grad_norm": 0.08754517232013051, + "learning_rate": 6.755343381424659e-06, + "loss": 0.3976, + "step": 856 + }, + { + "epoch": 4.160072760345612, + "grad_norm": 0.09520671649907132, + "learning_rate": 6.68014997727275e-06, + "loss": 0.3921, + "step": 857 + }, + { + "epoch": 4.164923450053054, + "grad_norm": 0.09719121632502135, + "learning_rate": 6.605339259407104e-06, + "loss": 0.3852, + "step": 858 + }, + { + "epoch": 4.169774139760497, + "grad_norm": 0.08870928687089744, + "learning_rate": 6.530912087046317e-06, + "loss": 0.395, + "step": 859 + }, + { + "epoch": 4.17462482946794, + "grad_norm": 0.08685694037296439, + "learning_rate": 6.456869315003946e-06, + "loss": 0.3941, + "step": 860 + }, + { + "epoch": 4.179475519175383, + "grad_norm": 0.08904997313439429, + "learning_rate": 6.3832117936785564e-06, + "loss": 0.3997, + "step": 861 + }, + { + "epoch": 4.184326208882825, + "grad_norm": 0.0910482544729166, + "learning_rate": 6.309940369044047e-06, + "loss": 0.3945, + "step": 862 + }, + { + "epoch": 4.189176898590269, + "grad_norm": 0.0887134263309384, + "learning_rate": 6.23705588263992e-06, + "loss": 0.3881, + "step": 863 + }, + { + "epoch": 4.194027588297711, + "grad_norm": 0.08977043998387933, + "learning_rate": 6.164559171561553e-06, + "loss": 0.3957, + "step": 864 + }, + { + "epoch": 4.198878278005154, + "grad_norm": 0.09216637353138708, + "learning_rate": 6.092451068450671e-06, + "loss": 0.3969, + "step": 865 + }, + { + "epoch": 4.203728967712596, + "grad_norm": 0.08464526503746563, + "learning_rate": 6.020732401485751e-06, + "loss": 0.3896, + "step": 866 + }, + { + "epoch": 4.20857965742004, + "grad_norm": 0.08983668958241023, + "learning_rate": 5.9494039943724845e-06, + "loss": 0.4025, + "step": 867 + }, + { + "epoch": 4.213430347127482, + "grad_norm": 0.08753368121579265, + "learning_rate": 5.878466666334341e-06, + "loss": 0.3954, + "step": 868 + }, + { + "epoch": 4.218281036834925, + "grad_norm": 0.08687937990016806, + "learning_rate": 5.80792123210316e-06, + "loss": 0.391, + "step": 869 + }, + { + "epoch": 4.223131726542368, + "grad_norm": 0.08701962255264487, + "learning_rate": 5.737768501909773e-06, + "loss": 0.3965, + "step": 870 + }, + { + "epoch": 4.2279824162498105, + "grad_norm": 0.08811499746627446, + "learning_rate": 5.668009281474751e-06, + "loss": 0.3846, + "step": 871 + }, + { + "epoch": 4.232833105957253, + "grad_norm": 0.08561633996061273, + "learning_rate": 5.598644371999085e-06, + "loss": 0.3919, + "step": 872 + }, + { + "epoch": 4.237683795664696, + "grad_norm": 0.08321419945310787, + "learning_rate": 5.5296745701549906e-06, + "loss": 0.394, + "step": 873 + }, + { + "epoch": 4.242534485372139, + "grad_norm": 0.08724643182123827, + "learning_rate": 5.4611006680768305e-06, + "loss": 0.3994, + "step": 874 + }, + { + "epoch": 4.2473851750795815, + "grad_norm": 0.08361301460174729, + "learning_rate": 5.3929234533519345e-06, + "loss": 0.388, + "step": 875 + }, + { + "epoch": 4.252235864787025, + "grad_norm": 0.0862576560444418, + "learning_rate": 5.325143709011587e-06, + "loss": 0.3942, + "step": 876 + }, + { + "epoch": 4.257086554494467, + "grad_norm": 0.082606577377262, + "learning_rate": 5.257762213522055e-06, + "loss": 0.3907, + "step": 877 + }, + { + "epoch": 4.26193724420191, + "grad_norm": 0.08601520095290771, + "learning_rate": 5.19077974077558e-06, + "loss": 0.3927, + "step": 878 + }, + { + "epoch": 4.266787933909352, + "grad_norm": 0.08144847498379731, + "learning_rate": 5.124197060081564e-06, + "loss": 0.3981, + "step": 879 + }, + { + "epoch": 4.271638623616796, + "grad_norm": 0.08369648223405692, + "learning_rate": 5.058014936157714e-06, + "loss": 0.3987, + "step": 880 + }, + { + "epoch": 4.276489313324238, + "grad_norm": 0.08487450387282452, + "learning_rate": 4.992234129121225e-06, + "loss": 0.3993, + "step": 881 + }, + { + "epoch": 4.281340003031681, + "grad_norm": 0.0835928002976722, + "learning_rate": 4.926855394480079e-06, + "loss": 0.394, + "step": 882 + }, + { + "epoch": 4.286190692739124, + "grad_norm": 0.0877663910781713, + "learning_rate": 4.861879483124372e-06, + "loss": 0.3888, + "step": 883 + }, + { + "epoch": 4.291041382446567, + "grad_norm": 0.08201902566316537, + "learning_rate": 4.797307141317666e-06, + "loss": 0.3978, + "step": 884 + }, + { + "epoch": 4.295892072154009, + "grad_norm": 0.08252391090570044, + "learning_rate": 4.7331391106884364e-06, + "loss": 0.3949, + "step": 885 + }, + { + "epoch": 4.300742761861452, + "grad_norm": 0.08121494056766716, + "learning_rate": 4.6693761282215766e-06, + "loss": 0.3922, + "step": 886 + }, + { + "epoch": 4.305593451568895, + "grad_norm": 0.08414265476669228, + "learning_rate": 4.606018926249851e-06, + "loss": 0.389, + "step": 887 + }, + { + "epoch": 4.310444141276338, + "grad_norm": 0.08528892500629966, + "learning_rate": 4.543068232445596e-06, + "loss": 0.3956, + "step": 888 + }, + { + "epoch": 4.31529483098378, + "grad_norm": 0.08280199210655267, + "learning_rate": 4.480524769812276e-06, + "loss": 0.3938, + "step": 889 + }, + { + "epoch": 4.320145520691224, + "grad_norm": 0.08884076809121294, + "learning_rate": 4.418389256676206e-06, + "loss": 0.3947, + "step": 890 + }, + { + "epoch": 4.324996210398666, + "grad_norm": 0.08902762031211295, + "learning_rate": 4.35666240667834e-06, + "loss": 0.3907, + "step": 891 + }, + { + "epoch": 4.329846900106109, + "grad_norm": 0.08464410835355753, + "learning_rate": 4.295344928765999e-06, + "loss": 0.3939, + "step": 892 + }, + { + "epoch": 4.334697589813552, + "grad_norm": 0.08347421805827179, + "learning_rate": 4.234437527184785e-06, + "loss": 0.3985, + "step": 893 + }, + { + "epoch": 4.3395482795209945, + "grad_norm": 0.08770395898608876, + "learning_rate": 4.173940901470488e-06, + "loss": 0.395, + "step": 894 + }, + { + "epoch": 4.344398969228437, + "grad_norm": 0.08139356257791996, + "learning_rate": 4.11385574644104e-06, + "loss": 0.39, + "step": 895 + }, + { + "epoch": 4.34924965893588, + "grad_norm": 0.0818047074224124, + "learning_rate": 4.054182752188501e-06, + "loss": 0.3948, + "step": 896 + }, + { + "epoch": 4.354100348643323, + "grad_norm": 0.08370905382784635, + "learning_rate": 3.994922604071217e-06, + "loss": 0.3941, + "step": 897 + }, + { + "epoch": 4.3589510383507655, + "grad_norm": 0.08543140680473625, + "learning_rate": 3.936075982705871e-06, + "loss": 0.3982, + "step": 898 + }, + { + "epoch": 4.363801728058208, + "grad_norm": 0.08330546263387466, + "learning_rate": 3.877643563959694e-06, + "loss": 0.3986, + "step": 899 + }, + { + "epoch": 4.368652417765651, + "grad_norm": 0.08181359063571672, + "learning_rate": 3.819626018942732e-06, + "loss": 0.3962, + "step": 900 + }, + { + "epoch": 4.373503107473094, + "grad_norm": 0.08458157848919085, + "learning_rate": 3.762024014000054e-06, + "loss": 0.3974, + "step": 901 + }, + { + "epoch": 4.378353797180536, + "grad_norm": 0.0797245693813596, + "learning_rate": 3.7048382107042113e-06, + "loss": 0.3849, + "step": 902 + }, + { + "epoch": 4.38320448688798, + "grad_norm": 0.08552325047075819, + "learning_rate": 3.6480692658475446e-06, + "loss": 0.3908, + "step": 903 + }, + { + "epoch": 4.388055176595422, + "grad_norm": 0.08191614537028945, + "learning_rate": 3.5917178314346955e-06, + "loss": 0.398, + "step": 904 + }, + { + "epoch": 4.392905866302865, + "grad_norm": 0.08324472646577967, + "learning_rate": 3.535784554675088e-06, + "loss": 0.3941, + "step": 905 + }, + { + "epoch": 4.397756556010307, + "grad_norm": 0.08386476178169076, + "learning_rate": 3.480270077975525e-06, + "loss": 0.395, + "step": 906 + }, + { + "epoch": 4.402607245717751, + "grad_norm": 0.08759720428686872, + "learning_rate": 3.42517503893276e-06, + "loss": 0.3879, + "step": 907 + }, + { + "epoch": 4.407457935425193, + "grad_norm": 0.0834695533465509, + "learning_rate": 3.370500070326257e-06, + "loss": 0.3832, + "step": 908 + }, + { + "epoch": 4.412308625132636, + "grad_norm": 0.080547419306159, + "learning_rate": 3.3162458001108332e-06, + "loss": 0.3858, + "step": 909 + }, + { + "epoch": 4.417159314840079, + "grad_norm": 0.08406690651169581, + "learning_rate": 3.2624128514094778e-06, + "loss": 0.3923, + "step": 910 + }, + { + "epoch": 4.422010004547522, + "grad_norm": 0.08417275645743631, + "learning_rate": 3.20900184250625e-06, + "loss": 0.3933, + "step": 911 + }, + { + "epoch": 4.426860694254964, + "grad_norm": 0.08104238775596306, + "learning_rate": 3.1560133868390895e-06, + "loss": 0.4023, + "step": 912 + }, + { + "epoch": 4.431711383962407, + "grad_norm": 0.08376361425082632, + "learning_rate": 3.1034480929928333e-06, + "loss": 0.399, + "step": 913 + }, + { + "epoch": 4.43656207366985, + "grad_norm": 0.08058350379685782, + "learning_rate": 3.0513065646921957e-06, + "loss": 0.3946, + "step": 914 + }, + { + "epoch": 4.441412763377293, + "grad_norm": 0.08013760556674378, + "learning_rate": 2.999589400794851e-06, + "loss": 0.392, + "step": 915 + }, + { + "epoch": 4.446263453084735, + "grad_norm": 0.08082216615145961, + "learning_rate": 2.948297195284546e-06, + "loss": 0.3916, + "step": 916 + }, + { + "epoch": 4.4511141427921785, + "grad_norm": 0.09007717527915819, + "learning_rate": 2.897430537264283e-06, + "loss": 0.3947, + "step": 917 + }, + { + "epoch": 4.455964832499621, + "grad_norm": 0.0866390993026153, + "learning_rate": 2.8469900109495553e-06, + "loss": 0.3942, + "step": 918 + }, + { + "epoch": 4.460815522207064, + "grad_norm": 0.08034663815912857, + "learning_rate": 2.79697619566162e-06, + "loss": 0.3888, + "step": 919 + }, + { + "epoch": 4.465666211914507, + "grad_norm": 0.08423465438674188, + "learning_rate": 2.7473896658208743e-06, + "loss": 0.391, + "step": 920 + }, + { + "epoch": 4.4705169016219495, + "grad_norm": 0.08769412223500794, + "learning_rate": 2.6982309909402293e-06, + "loss": 0.3936, + "step": 921 + }, + { + "epoch": 4.475367591329392, + "grad_norm": 0.0794092004599977, + "learning_rate": 2.649500735618582e-06, + "loss": 0.3993, + "step": 922 + }, + { + "epoch": 4.480218281036835, + "grad_norm": 0.08321318361659834, + "learning_rate": 2.6011994595343516e-06, + "loss": 0.3965, + "step": 923 + }, + { + "epoch": 4.485068970744278, + "grad_norm": 0.08400144661884966, + "learning_rate": 2.5533277174389916e-06, + "loss": 0.3927, + "step": 924 + }, + { + "epoch": 4.48991966045172, + "grad_norm": 0.08039773750467258, + "learning_rate": 2.5058860591506973e-06, + "loss": 0.3927, + "step": 925 + }, + { + "epoch": 4.494770350159163, + "grad_norm": 0.08132175209524008, + "learning_rate": 2.4588750295480246e-06, + "loss": 0.3888, + "step": 926 + }, + { + "epoch": 4.499621039866606, + "grad_norm": 0.07920562392368859, + "learning_rate": 2.4122951685636674e-06, + "loss": 0.3896, + "step": 927 + }, + { + "epoch": 4.504471729574049, + "grad_norm": 0.07833571502263627, + "learning_rate": 2.366147011178246e-06, + "loss": 0.398, + "step": 928 + }, + { + "epoch": 4.509322419281491, + "grad_norm": 0.08232920047142565, + "learning_rate": 2.320431087414159e-06, + "loss": 0.3838, + "step": 929 + }, + { + "epoch": 4.514173108988935, + "grad_norm": 0.08136833742803433, + "learning_rate": 2.275147922329506e-06, + "loss": 0.3935, + "step": 930 + }, + { + "epoch": 4.519023798696377, + "grad_norm": 0.08383396678636225, + "learning_rate": 2.230298036012055e-06, + "loss": 0.3913, + "step": 931 + }, + { + "epoch": 4.52387448840382, + "grad_norm": 0.08000391439211, + "learning_rate": 2.1858819435732583e-06, + "loss": 0.395, + "step": 932 + }, + { + "epoch": 4.528725178111262, + "grad_norm": 0.080147749223563, + "learning_rate": 2.141900155142351e-06, + "loss": 0.3951, + "step": 933 + }, + { + "epoch": 4.533575867818706, + "grad_norm": 0.07978655829531454, + "learning_rate": 2.0983531758604726e-06, + "loss": 0.4011, + "step": 934 + }, + { + "epoch": 4.538426557526148, + "grad_norm": 0.08386421814730027, + "learning_rate": 2.055241505874892e-06, + "loss": 0.3999, + "step": 935 + }, + { + "epoch": 4.543277247233591, + "grad_norm": 0.08174904946083562, + "learning_rate": 2.0125656403332396e-06, + "loss": 0.3968, + "step": 936 + }, + { + "epoch": 4.548127936941034, + "grad_norm": 0.07927025601942647, + "learning_rate": 1.970326069377828e-06, + "loss": 0.4001, + "step": 937 + }, + { + "epoch": 4.552978626648477, + "grad_norm": 0.0783638905343862, + "learning_rate": 1.928523278140033e-06, + "loss": 0.3943, + "step": 938 + }, + { + "epoch": 4.557829316355919, + "grad_norm": 0.08012821453349714, + "learning_rate": 1.887157746734718e-06, + "loss": 0.3972, + "step": 939 + }, + { + "epoch": 4.562680006063362, + "grad_norm": 0.08022389886437123, + "learning_rate": 1.846229950254692e-06, + "loss": 0.3888, + "step": 940 + }, + { + "epoch": 4.567530695770805, + "grad_norm": 0.07938938767749412, + "learning_rate": 1.8057403587652977e-06, + "loss": 0.3892, + "step": 941 + }, + { + "epoch": 4.572381385478248, + "grad_norm": 0.07979264917062728, + "learning_rate": 1.7656894372989785e-06, + "loss": 0.3991, + "step": 942 + }, + { + "epoch": 4.57723207518569, + "grad_norm": 0.07900416574598519, + "learning_rate": 1.726077645849955e-06, + "loss": 0.4033, + "step": 943 + }, + { + "epoch": 4.5820827648931335, + "grad_norm": 0.0816244017270053, + "learning_rate": 1.6869054393689265e-06, + "loss": 0.394, + "step": 944 + }, + { + "epoch": 4.586933454600576, + "grad_norm": 0.07678354405433441, + "learning_rate": 1.6481732677578798e-06, + "loss": 0.4026, + "step": 945 + }, + { + "epoch": 4.5917841443080185, + "grad_norm": 0.07913436605637802, + "learning_rate": 1.60988157586488e-06, + "loss": 0.378, + "step": 946 + }, + { + "epoch": 4.596634834015462, + "grad_norm": 0.0793297266554538, + "learning_rate": 1.5720308034789721e-06, + "loss": 0.391, + "step": 947 + }, + { + "epoch": 4.601485523722904, + "grad_norm": 0.07767980761029898, + "learning_rate": 1.5346213853251546e-06, + "loss": 0.3978, + "step": 948 + }, + { + "epoch": 4.606336213430347, + "grad_norm": 0.08494507410525762, + "learning_rate": 1.4976537510593646e-06, + "loss": 0.3995, + "step": 949 + }, + { + "epoch": 4.61118690313779, + "grad_norm": 0.08237577571806884, + "learning_rate": 1.4611283252635412e-06, + "loss": 0.4038, + "step": 950 + }, + { + "epoch": 4.616037592845233, + "grad_norm": 0.07590194142960284, + "learning_rate": 1.425045527440756e-06, + "loss": 0.3956, + "step": 951 + }, + { + "epoch": 4.620888282552675, + "grad_norm": 0.07954982455169565, + "learning_rate": 1.3894057720104104e-06, + "loss": 0.399, + "step": 952 + }, + { + "epoch": 4.625738972260118, + "grad_norm": 0.07731849776688487, + "learning_rate": 1.354209468303429e-06, + "loss": 0.3828, + "step": 953 + }, + { + "epoch": 4.630589661967561, + "grad_norm": 0.07970299041030604, + "learning_rate": 1.3194570205576284e-06, + "loss": 0.3954, + "step": 954 + }, + { + "epoch": 4.635440351675004, + "grad_norm": 0.07874124868840192, + "learning_rate": 1.2851488279130053e-06, + "loss": 0.3876, + "step": 955 + }, + { + "epoch": 4.640291041382446, + "grad_norm": 0.07691612719760402, + "learning_rate": 1.2512852844071933e-06, + "loss": 0.3949, + "step": 956 + }, + { + "epoch": 4.64514173108989, + "grad_norm": 0.08126016720695953, + "learning_rate": 1.2178667789709287e-06, + "loss": 0.3919, + "step": 957 + }, + { + "epoch": 4.649992420797332, + "grad_norm": 0.08002643782323322, + "learning_rate": 1.1848936954235702e-06, + "loss": 0.395, + "step": 958 + }, + { + "epoch": 4.654843110504775, + "grad_norm": 0.08026194265591635, + "learning_rate": 1.1523664124687284e-06, + "loss": 0.3997, + "step": 959 + }, + { + "epoch": 4.659693800212217, + "grad_norm": 0.07660352848559206, + "learning_rate": 1.1202853036898476e-06, + "loss": 0.3974, + "step": 960 + }, + { + "epoch": 4.664544489919661, + "grad_norm": 0.08113236962163348, + "learning_rate": 1.0886507375459908e-06, + "loss": 0.3981, + "step": 961 + }, + { + "epoch": 4.669395179627103, + "grad_norm": 0.07701309756204706, + "learning_rate": 1.0574630773675687e-06, + "loss": 0.3839, + "step": 962 + }, + { + "epoch": 4.674245869334546, + "grad_norm": 0.07891978078549244, + "learning_rate": 1.0267226813521635e-06, + "loss": 0.3877, + "step": 963 + }, + { + "epoch": 4.679096559041989, + "grad_norm": 0.07852321868608765, + "learning_rate": 9.964299025604274e-07, + "loss": 0.3921, + "step": 964 + }, + { + "epoch": 4.683947248749432, + "grad_norm": 0.07818107376466682, + "learning_rate": 9.66585088912022e-07, + "loss": 0.3967, + "step": 965 + }, + { + "epoch": 4.688797938456874, + "grad_norm": 0.076600770628096, + "learning_rate": 9.371885831816319e-07, + "loss": 0.395, + "step": 966 + }, + { + "epoch": 4.6936486281643175, + "grad_norm": 0.07627250485843899, + "learning_rate": 9.082407229950018e-07, + "loss": 0.3976, + "step": 967 + }, + { + "epoch": 4.69849931787176, + "grad_norm": 0.07905168555654667, + "learning_rate": 8.797418408251101e-07, + "loss": 0.3918, + "step": 968 + }, + { + "epoch": 4.7033500075792025, + "grad_norm": 0.07625038396328788, + "learning_rate": 8.516922639882819e-07, + "loss": 0.3897, + "step": 969 + }, + { + "epoch": 4.708200697286646, + "grad_norm": 0.08055085830124646, + "learning_rate": 8.2409231464049e-07, + "loss": 0.3901, + "step": 970 + }, + { + "epoch": 4.713051386994088, + "grad_norm": 0.07599467665769233, + "learning_rate": 7.969423097736162e-07, + "loss": 0.3931, + "step": 971 + }, + { + "epoch": 4.717902076701531, + "grad_norm": 0.08049982038815076, + "learning_rate": 7.702425612118269e-07, + "loss": 0.3962, + "step": 972 + }, + { + "epoch": 4.7227527664089735, + "grad_norm": 0.07817439712988589, + "learning_rate": 7.439933756079942e-07, + "loss": 0.3913, + "step": 973 + }, + { + "epoch": 4.727603456116417, + "grad_norm": 0.08001209525722262, + "learning_rate": 7.181950544401695e-07, + "loss": 0.3935, + "step": 974 + }, + { + "epoch": 4.732454145823859, + "grad_norm": 0.07544329449164154, + "learning_rate": 6.928478940081107e-07, + "loss": 0.3984, + "step": 975 + }, + { + "epoch": 4.737304835531302, + "grad_norm": 0.07792056514995602, + "learning_rate": 6.679521854299032e-07, + "loss": 0.3914, + "step": 976 + }, + { + "epoch": 4.742155525238745, + "grad_norm": 0.07777159770613211, + "learning_rate": 6.435082146385885e-07, + "loss": 0.3901, + "step": 977 + }, + { + "epoch": 4.747006214946188, + "grad_norm": 0.07684966968716175, + "learning_rate": 6.195162623789052e-07, + "loss": 0.393, + "step": 978 + }, + { + "epoch": 4.75185690465363, + "grad_norm": 0.07663626226452194, + "learning_rate": 5.959766042040426e-07, + "loss": 0.3879, + "step": 979 + }, + { + "epoch": 4.756707594361073, + "grad_norm": 0.07653692258850123, + "learning_rate": 5.728895104724963e-07, + "loss": 0.3959, + "step": 980 + }, + { + "epoch": 4.761558284068516, + "grad_norm": 0.07731590848782262, + "learning_rate": 5.502552463449418e-07, + "loss": 0.3984, + "step": 981 + }, + { + "epoch": 4.766408973775959, + "grad_norm": 0.07715742201671594, + "learning_rate": 5.280740717812149e-07, + "loss": 0.3977, + "step": 982 + }, + { + "epoch": 4.771259663483401, + "grad_norm": 0.07733908081214975, + "learning_rate": 5.063462415372967e-07, + "loss": 0.3948, + "step": 983 + }, + { + "epoch": 4.776110353190845, + "grad_norm": 0.07797117300936077, + "learning_rate": 4.850720051624124e-07, + "loss": 0.3914, + "step": 984 + }, + { + "epoch": 4.780961042898287, + "grad_norm": 0.07721734718928677, + "learning_rate": 4.642516069961556e-07, + "loss": 0.392, + "step": 985 + }, + { + "epoch": 4.78581173260573, + "grad_norm": 0.07648551851063208, + "learning_rate": 4.438852861656751e-07, + "loss": 0.3951, + "step": 986 + }, + { + "epoch": 4.790662422313172, + "grad_norm": 0.07521119037010907, + "learning_rate": 4.2397327658294076e-07, + "loss": 0.3899, + "step": 987 + }, + { + "epoch": 4.795513112020616, + "grad_norm": 0.07655879860272995, + "learning_rate": 4.045158069420474e-07, + "loss": 0.3963, + "step": 988 + }, + { + "epoch": 4.800363801728058, + "grad_norm": 0.07622760847546149, + "learning_rate": 3.8551310071659023e-07, + "loss": 0.3975, + "step": 989 + }, + { + "epoch": 4.805214491435501, + "grad_norm": 0.0750428400819878, + "learning_rate": 3.6696537615711124e-07, + "loss": 0.3968, + "step": 990 + }, + { + "epoch": 4.810065181142944, + "grad_norm": 0.07852243689578567, + "learning_rate": 3.4887284628857266e-07, + "loss": 0.3932, + "step": 991 + }, + { + "epoch": 4.8149158708503865, + "grad_norm": 0.0771348195411035, + "learning_rate": 3.3123571890791405e-07, + "loss": 0.3887, + "step": 992 + }, + { + "epoch": 4.819766560557829, + "grad_norm": 0.07614331045752255, + "learning_rate": 3.1405419658168125e-07, + "loss": 0.394, + "step": 993 + }, + { + "epoch": 4.824617250265272, + "grad_norm": 0.07563944264490313, + "learning_rate": 2.973284766436857e-07, + "loss": 0.3917, + "step": 994 + }, + { + "epoch": 4.829467939972715, + "grad_norm": 0.07503164664083585, + "learning_rate": 2.810587511927354e-07, + "loss": 0.3901, + "step": 995 + }, + { + "epoch": 4.8343186296801575, + "grad_norm": 0.07663595734772072, + "learning_rate": 2.652452070904499e-07, + "loss": 0.3923, + "step": 996 + }, + { + "epoch": 4.839169319387601, + "grad_norm": 0.07606961706547127, + "learning_rate": 2.498880259590797e-07, + "loss": 0.3944, + "step": 997 + }, + { + "epoch": 4.844020009095043, + "grad_norm": 0.0761182148197196, + "learning_rate": 2.3498738417945034e-07, + "loss": 0.3975, + "step": 998 + }, + { + "epoch": 4.848870698802486, + "grad_norm": 0.08039104985634041, + "learning_rate": 2.205434528889283e-07, + "loss": 0.3971, + "step": 999 + }, + { + "epoch": 4.853721388509928, + "grad_norm": 0.07775606427920397, + "learning_rate": 2.0655639797944937e-07, + "loss": 0.3903, + "step": 1000 + }, + { + "epoch": 4.858572078217372, + "grad_norm": 0.07610940999590161, + "learning_rate": 1.9302638009561782e-07, + "loss": 0.396, + "step": 1001 + }, + { + "epoch": 4.863422767924814, + "grad_norm": 0.07569450912275462, + "learning_rate": 1.7995355463285457e-07, + "loss": 0.3965, + "step": 1002 + }, + { + "epoch": 4.868273457632257, + "grad_norm": 0.07643720956773195, + "learning_rate": 1.6733807173562988e-07, + "loss": 0.3913, + "step": 1003 + }, + { + "epoch": 4.8731241473397, + "grad_norm": 0.07571887859535044, + "learning_rate": 1.5518007629571342e-07, + "loss": 0.3995, + "step": 1004 + }, + { + "epoch": 4.877974837047143, + "grad_norm": 0.07657090064859913, + "learning_rate": 1.4347970795054456e-07, + "loss": 0.3967, + "step": 1005 + }, + { + "epoch": 4.882825526754585, + "grad_norm": 0.07475277823538463, + "learning_rate": 1.3223710108158483e-07, + "loss": 0.3965, + "step": 1006 + }, + { + "epoch": 4.887676216462028, + "grad_norm": 0.07465351831333361, + "learning_rate": 1.214523848128124e-07, + "loss": 0.3971, + "step": 1007 + }, + { + "epoch": 4.892526906169471, + "grad_norm": 0.07637388169909817, + "learning_rate": 1.111256830092211e-07, + "loss": 0.3996, + "step": 1008 + }, + { + "epoch": 4.897377595876914, + "grad_norm": 0.07639665552172381, + "learning_rate": 1.0125711427540374e-07, + "loss": 0.3949, + "step": 1009 + }, + { + "epoch": 4.902228285584356, + "grad_norm": 0.07373092012101537, + "learning_rate": 9.184679195417989e-08, + "loss": 0.3889, + "step": 1010 + }, + { + "epoch": 4.9070789752918, + "grad_norm": 0.07583698163385665, + "learning_rate": 8.289482412531246e-08, + "loss": 0.3984, + "step": 1011 + }, + { + "epoch": 4.911929664999242, + "grad_norm": 0.07647787343235872, + "learning_rate": 7.440131360424652e-08, + "loss": 0.3887, + "step": 1012 + }, + { + "epoch": 4.916780354706685, + "grad_norm": 0.07751142464425709, + "learning_rate": 6.636635794094126e-08, + "loss": 0.3908, + "step": 1013 + }, + { + "epoch": 4.921631044414128, + "grad_norm": 0.07884353160777417, + "learning_rate": 5.879004941874655e-08, + "loss": 0.3951, + "step": 1014 + }, + { + "epoch": 4.9264817341215705, + "grad_norm": 0.07661585400143268, + "learning_rate": 5.16724750533415e-08, + "loss": 0.398, + "step": 1015 + }, + { + "epoch": 4.931332423829013, + "grad_norm": 0.07506936516416932, + "learning_rate": 4.5013716591730815e-08, + "loss": 0.3944, + "step": 1016 + }, + { + "epoch": 4.9361831135364564, + "grad_norm": 0.07625120541429038, + "learning_rate": 3.881385051132114e-08, + "loss": 0.3997, + "step": 1017 + }, + { + "epoch": 4.941033803243899, + "grad_norm": 0.0754548388472603, + "learning_rate": 3.307294801902838e-08, + "loss": 0.3968, + "step": 1018 + }, + { + "epoch": 4.9458844929513415, + "grad_norm": 0.07563871272528049, + "learning_rate": 2.7791075050460636e-08, + "loss": 0.3977, + "step": 1019 + }, + { + "epoch": 4.950735182658784, + "grad_norm": 0.0760975342061979, + "learning_rate": 2.2968292269167637e-08, + "loss": 0.3978, + "step": 1020 + }, + { + "epoch": 4.955585872366227, + "grad_norm": 0.07484229156933196, + "learning_rate": 1.8604655065939116e-08, + "loss": 0.3888, + "step": 1021 + }, + { + "epoch": 4.96043656207367, + "grad_norm": 0.07489129192172529, + "learning_rate": 1.470021355816975e-08, + "loss": 0.3952, + "step": 1022 + }, + { + "epoch": 4.965287251781112, + "grad_norm": 0.07561234415729748, + "learning_rate": 1.1255012589286297e-08, + "loss": 0.3988, + "step": 1023 + }, + { + "epoch": 4.970137941488556, + "grad_norm": 0.07589712959511802, + "learning_rate": 8.269091728232426e-09, + "loss": 0.3858, + "step": 1024 + }, + { + "epoch": 4.974988631195998, + "grad_norm": 0.07684966203710471, + "learning_rate": 5.742485269006892e-09, + "loss": 0.3976, + "step": 1025 + }, + { + "epoch": 4.979839320903441, + "grad_norm": 0.07450707523151534, + "learning_rate": 3.6752222302727238e-09, + "loss": 0.3922, + "step": 1026 + }, + { + "epoch": 4.984690010610883, + "grad_norm": 0.0754634711281639, + "learning_rate": 2.06732635503748e-09, + "loss": 0.3894, + "step": 1027 + }, + { + "epoch": 4.989540700318327, + "grad_norm": 0.07460842756020492, + "learning_rate": 9.188161103557136e-10, + "loss": 0.3893, + "step": 1028 + }, + { + "epoch": 4.994391390025769, + "grad_norm": 0.07411367814575705, + "learning_rate": 2.2970468714245132e-10, + "loss": 0.3934, + "step": 1029 + }, + { + "epoch": 4.999242079733212, + "grad_norm": 0.07533391242414708, + "learning_rate": 0.0, + "loss": 0.3877, + "step": 1030 + }, + { + "epoch": 4.999242079733212, + "step": 1030, + "total_flos": 2.739131934768418e+19, + "train_loss": 0.07882811409755817, + "train_runtime": 48155.7887, + "train_samples_per_second": 10.958, + "train_steps_per_second": 0.021 + } + ], + "logging_steps": 1, + "max_steps": 1030, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.739131934768418e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}