{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1098, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00546448087431694, "grad_norm": 0.697250247001648, "learning_rate": 5.454545454545455e-07, "loss": 1.9263319969177246, "step": 2 }, { "epoch": 0.01092896174863388, "grad_norm": 0.721147894859314, "learning_rate": 1.6363636363636363e-06, "loss": 1.9639064073562622, "step": 4 }, { "epoch": 0.01639344262295082, "grad_norm": 0.7265484929084778, "learning_rate": 2.7272727272727272e-06, "loss": 1.8129602670669556, "step": 6 }, { "epoch": 0.02185792349726776, "grad_norm": 0.5450526475906372, "learning_rate": 3.818181818181818e-06, "loss": 1.9235541820526123, "step": 8 }, { "epoch": 0.0273224043715847, "grad_norm": 0.8771107196807861, "learning_rate": 4.90909090909091e-06, "loss": 2.251471519470215, "step": 10 }, { "epoch": 0.03278688524590164, "grad_norm": 10.971705436706543, "learning_rate": 6e-06, "loss": 3.0256073474884033, "step": 12 }, { "epoch": 0.03825136612021858, "grad_norm": 0.5691296458244324, "learning_rate": 7.090909090909091e-06, "loss": 1.903341293334961, "step": 14 }, { "epoch": 0.04371584699453552, "grad_norm": 0.632869303226471, "learning_rate": 8.181818181818181e-06, "loss": 1.824144721031189, "step": 16 }, { "epoch": 0.04918032786885246, "grad_norm": 2.687931776046753, "learning_rate": 9.272727272727273e-06, "loss": 2.273005962371826, "step": 18 }, { "epoch": 0.0546448087431694, "grad_norm": 1.9365599155426025, "learning_rate": 1.0363636363636364e-05, "loss": 1.9546592235565186, "step": 20 }, { "epoch": 0.060109289617486336, "grad_norm": 5.1370134353637695, "learning_rate": 1.1454545454545455e-05, "loss": 2.148141860961914, "step": 22 }, { "epoch": 0.06557377049180328, "grad_norm": 0.595227837562561, "learning_rate": 1.2545454545454545e-05, "loss": 1.7927855253219604, "step": 24 }, { "epoch": 0.07103825136612021, "grad_norm": 0.4974764585494995, "learning_rate": 1.3636363636363637e-05, "loss": 1.8154855966567993, "step": 26 }, { "epoch": 0.07650273224043716, "grad_norm": 0.4138644337654114, "learning_rate": 1.4727272727272728e-05, "loss": 1.7593921422958374, "step": 28 }, { "epoch": 0.08196721311475409, "grad_norm": 1.1251859664916992, "learning_rate": 1.5818181818181818e-05, "loss": 1.6035438776016235, "step": 30 }, { "epoch": 0.08743169398907104, "grad_norm": 3.0598487854003906, "learning_rate": 1.6909090909090907e-05, "loss": 1.9531118869781494, "step": 32 }, { "epoch": 0.09289617486338798, "grad_norm": 0.38799506425857544, "learning_rate": 1.8e-05, "loss": 1.632307529449463, "step": 34 }, { "epoch": 0.09836065573770492, "grad_norm": 1.2041815519332886, "learning_rate": 1.909090909090909e-05, "loss": 2.1407182216644287, "step": 36 }, { "epoch": 0.10382513661202186, "grad_norm": 1.2261812686920166, "learning_rate": 2.0181818181818183e-05, "loss": 1.7205183506011963, "step": 38 }, { "epoch": 0.1092896174863388, "grad_norm": 0.778466522693634, "learning_rate": 2.1272727272727273e-05, "loss": 1.4309196472167969, "step": 40 }, { "epoch": 0.11475409836065574, "grad_norm": 0.4726494550704956, "learning_rate": 2.2363636363636366e-05, "loss": 1.3979218006134033, "step": 42 }, { "epoch": 0.12021857923497267, "grad_norm": 0.46920815110206604, "learning_rate": 2.3454545454545456e-05, "loss": 1.5350127220153809, "step": 44 }, { "epoch": 0.12568306010928962, "grad_norm": 0.3451603353023529, "learning_rate": 2.454545454545455e-05, "loss": 1.2750825881958008, "step": 46 }, { "epoch": 0.13114754098360656, "grad_norm": 0.4018639326095581, "learning_rate": 2.5636363636363635e-05, "loss": 1.4977831840515137, "step": 48 }, { "epoch": 0.1366120218579235, "grad_norm": 0.32938382029533386, "learning_rate": 2.6727272727272728e-05, "loss": 1.455066204071045, "step": 50 }, { "epoch": 0.14207650273224043, "grad_norm": 0.3992089331150055, "learning_rate": 2.7818181818181818e-05, "loss": 1.50904381275177, "step": 52 }, { "epoch": 0.14754098360655737, "grad_norm": 0.48714780807495117, "learning_rate": 2.890909090909091e-05, "loss": 1.6593310832977295, "step": 54 }, { "epoch": 0.15300546448087432, "grad_norm": 0.5381816029548645, "learning_rate": 3e-05, "loss": 1.4189538955688477, "step": 56 }, { "epoch": 0.15846994535519127, "grad_norm": 2.2116336822509766, "learning_rate": 2.9997491688899256e-05, "loss": 1.0185623168945312, "step": 58 }, { "epoch": 0.16393442622950818, "grad_norm": 0.775851309299469, "learning_rate": 2.998996768768956e-05, "loss": 0.9615996479988098, "step": 60 }, { "epoch": 0.16939890710382513, "grad_norm": 0.5596693158149719, "learning_rate": 2.9977430792302124e-05, "loss": 1.397382378578186, "step": 62 }, { "epoch": 0.17486338797814208, "grad_norm": 0.37642741203308105, "learning_rate": 2.9959885661467903e-05, "loss": 1.1756441593170166, "step": 64 }, { "epoch": 0.18032786885245902, "grad_norm": 0.5881237983703613, "learning_rate": 2.993733881498636e-05, "loss": 0.8743929862976074, "step": 66 }, { "epoch": 0.18579234972677597, "grad_norm": 0.5072574615478516, "learning_rate": 2.9909798631302736e-05, "loss": 1.398439645767212, "step": 68 }, { "epoch": 0.1912568306010929, "grad_norm": 0.3287373185157776, "learning_rate": 2.987727534439457e-05, "loss": 1.2875562906265259, "step": 70 }, { "epoch": 0.19672131147540983, "grad_norm": 5.050961017608643, "learning_rate": 2.983978103996877e-05, "loss": 0.9359242916107178, "step": 72 }, { "epoch": 0.20218579234972678, "grad_norm": 0.4209338128566742, "learning_rate": 2.9797329650970525e-05, "loss": 1.1358091831207275, "step": 74 }, { "epoch": 0.20765027322404372, "grad_norm": 0.4743102490901947, "learning_rate": 2.974993695240579e-05, "loss": 1.252506971359253, "step": 76 }, { "epoch": 0.21311475409836064, "grad_norm": 0.7218373417854309, "learning_rate": 2.9697620555479297e-05, "loss": 1.3395963907241821, "step": 78 }, { "epoch": 0.2185792349726776, "grad_norm": 0.35181427001953125, "learning_rate": 2.9640399901050182e-05, "loss": 1.3289461135864258, "step": 80 }, { "epoch": 0.22404371584699453, "grad_norm": 0.6300466656684875, "learning_rate": 2.9578296252407734e-05, "loss": 1.3236968517303467, "step": 82 }, { "epoch": 0.22950819672131148, "grad_norm": 0.3324219286441803, "learning_rate": 2.9511332687369917e-05, "loss": 1.2778677940368652, "step": 84 }, { "epoch": 0.23497267759562843, "grad_norm": 0.3525979816913605, "learning_rate": 2.9439534089707624e-05, "loss": 1.2054697275161743, "step": 86 }, { "epoch": 0.24043715846994534, "grad_norm": 0.568080484867096, "learning_rate": 2.9362927139897832e-05, "loss": 0.8521115779876709, "step": 88 }, { "epoch": 0.2459016393442623, "grad_norm": 0.9504212141036987, "learning_rate": 2.9281540305209068e-05, "loss": 0.9611366987228394, "step": 90 }, { "epoch": 0.25136612021857924, "grad_norm": 0.5896546840667725, "learning_rate": 2.919540382912294e-05, "loss": 1.2202290296554565, "step": 92 }, { "epoch": 0.2568306010928962, "grad_norm": 0.4182462990283966, "learning_rate": 2.9104549720095634e-05, "loss": 0.7185679078102112, "step": 94 }, { "epoch": 0.26229508196721313, "grad_norm": 0.6521818041801453, "learning_rate": 2.9009011739663467e-05, "loss": 1.2726106643676758, "step": 96 }, { "epoch": 0.2677595628415301, "grad_norm": 0.5340232253074646, "learning_rate": 2.8908825389897094e-05, "loss": 1.2408103942871094, "step": 98 }, { "epoch": 0.273224043715847, "grad_norm": 0.3869662284851074, "learning_rate": 2.8804027900208843e-05, "loss": 1.1218898296356201, "step": 100 }, { "epoch": 0.2786885245901639, "grad_norm": 0.3763534724712372, "learning_rate": 2.8694658213518226e-05, "loss": 0.8273072242736816, "step": 102 }, { "epoch": 0.28415300546448086, "grad_norm": 0.36140915751457214, "learning_rate": 2.8580756971780686e-05, "loss": 1.3101215362548828, "step": 104 }, { "epoch": 0.2896174863387978, "grad_norm": 0.45537811517715454, "learning_rate": 2.846236650088497e-05, "loss": 1.2583913803100586, "step": 106 }, { "epoch": 0.29508196721311475, "grad_norm": 0.512202799320221, "learning_rate": 2.833953079492476e-05, "loss": 1.346620798110962, "step": 108 }, { "epoch": 0.3005464480874317, "grad_norm": 0.2695588767528534, "learning_rate": 2.82122954998504e-05, "loss": 1.265444278717041, "step": 110 }, { "epoch": 0.30601092896174864, "grad_norm": 0.3954983949661255, "learning_rate": 2.808070789650679e-05, "loss": 1.2438230514526367, "step": 112 }, { "epoch": 0.3114754098360656, "grad_norm": 0.6031639575958252, "learning_rate": 2.7944816883063727e-05, "loss": 1.0083987712860107, "step": 114 }, { "epoch": 0.31693989071038253, "grad_norm": 0.3518785834312439, "learning_rate": 2.7804672956845295e-05, "loss": 1.22913658618927, "step": 116 }, { "epoch": 0.3224043715846995, "grad_norm": 0.6181989908218384, "learning_rate": 2.766032819556495e-05, "loss": 1.2250205278396606, "step": 118 }, { "epoch": 0.32786885245901637, "grad_norm": 4.4156494140625, "learning_rate": 2.7511836237973366e-05, "loss": 0.9285832047462463, "step": 120 }, { "epoch": 0.3333333333333333, "grad_norm": 0.6863148212432861, "learning_rate": 2.735925226392618e-05, "loss": 1.5127151012420654, "step": 122 }, { "epoch": 0.33879781420765026, "grad_norm": 0.2905527651309967, "learning_rate": 2.7202632973879086e-05, "loss": 1.2349207401275635, "step": 124 }, { "epoch": 0.3442622950819672, "grad_norm": 0.2775297462940216, "learning_rate": 2.7042036567817838e-05, "loss": 1.1945700645446777, "step": 126 }, { "epoch": 0.34972677595628415, "grad_norm": 0.3885118067264557, "learning_rate": 2.6877522723631036e-05, "loss": 0.8623289465904236, "step": 128 }, { "epoch": 0.3551912568306011, "grad_norm": 0.48008739948272705, "learning_rate": 2.6709152574933727e-05, "loss": 1.019203543663025, "step": 130 }, { "epoch": 0.36065573770491804, "grad_norm": 1.9439845085144043, "learning_rate": 2.6536988688350067e-05, "loss": 0.9837391376495361, "step": 132 }, { "epoch": 0.366120218579235, "grad_norm": 0.30251508951187134, "learning_rate": 2.6361095040263437e-05, "loss": 1.2110865116119385, "step": 134 }, { "epoch": 0.37158469945355194, "grad_norm": 0.3148176074028015, "learning_rate": 2.618153699304274e-05, "loss": 1.2147680521011353, "step": 136 }, { "epoch": 0.3770491803278688, "grad_norm": 0.42469698190689087, "learning_rate": 2.599838127075361e-05, "loss": 1.2221580743789673, "step": 138 }, { "epoch": 0.3825136612021858, "grad_norm": 0.33319714665412903, "learning_rate": 2.5811695934363666e-05, "loss": 0.7449063062667847, "step": 140 }, { "epoch": 0.3879781420765027, "grad_norm": 0.4343816339969635, "learning_rate": 2.5621550356450914e-05, "loss": 0.8672135472297668, "step": 142 }, { "epoch": 0.39344262295081966, "grad_norm": 0.44463595747947693, "learning_rate": 2.5428015195424825e-05, "loss": 0.8414742946624756, "step": 144 }, { "epoch": 0.3989071038251366, "grad_norm": 0.7496324181556702, "learning_rate": 2.5231162369269498e-05, "loss": 0.7604200839996338, "step": 146 }, { "epoch": 0.40437158469945356, "grad_norm": 0.36488234996795654, "learning_rate": 2.503106502881889e-05, "loss": 1.2763407230377197, "step": 148 }, { "epoch": 0.4098360655737705, "grad_norm": 0.34441977739334106, "learning_rate": 2.4827797530573762e-05, "loss": 1.190218210220337, "step": 150 }, { "epoch": 0.41530054644808745, "grad_norm": 0.4284515082836151, "learning_rate": 2.4621435409070757e-05, "loss": 1.2155894041061401, "step": 152 }, { "epoch": 0.4207650273224044, "grad_norm": 0.6492432355880737, "learning_rate": 2.4412055348813602e-05, "loss": 1.1843308210372925, "step": 154 }, { "epoch": 0.4262295081967213, "grad_norm": 0.34537646174430847, "learning_rate": 2.4199735155777017e-05, "loss": 1.2178831100463867, "step": 156 }, { "epoch": 0.43169398907103823, "grad_norm": 1.2686742544174194, "learning_rate": 2.3984553728493914e-05, "loss": 1.0317703485488892, "step": 158 }, { "epoch": 0.4371584699453552, "grad_norm": 0.5937894582748413, "learning_rate": 2.3766591028736547e-05, "loss": 1.2135744094848633, "step": 160 }, { "epoch": 0.4426229508196721, "grad_norm": 2.5808207988739014, "learning_rate": 2.3545928051802588e-05, "loss": 0.8519853353500366, "step": 162 }, { "epoch": 0.44808743169398907, "grad_norm": 0.3145885765552521, "learning_rate": 2.332264679641717e-05, "loss": 1.2207863330841064, "step": 164 }, { "epoch": 0.453551912568306, "grad_norm": 1.2698609828948975, "learning_rate": 2.3096830234261996e-05, "loss": 0.843329131603241, "step": 166 }, { "epoch": 0.45901639344262296, "grad_norm": 0.2692093849182129, "learning_rate": 2.2868562279142912e-05, "loss": 1.206931233406067, "step": 168 }, { "epoch": 0.4644808743169399, "grad_norm": 0.31359997391700745, "learning_rate": 2.2637927755807458e-05, "loss": 1.532818078994751, "step": 170 }, { "epoch": 0.46994535519125685, "grad_norm": 0.8711612820625305, "learning_rate": 2.2405012368423786e-05, "loss": 0.9171653389930725, "step": 172 }, { "epoch": 0.47540983606557374, "grad_norm": 0.41555553674697876, "learning_rate": 2.2169902668732893e-05, "loss": 1.2091866731643677, "step": 174 }, { "epoch": 0.4808743169398907, "grad_norm": 0.3255589008331299, "learning_rate": 2.193268602388583e-05, "loss": 1.1792771816253662, "step": 176 }, { "epoch": 0.48633879781420764, "grad_norm": 2.388908624649048, "learning_rate": 2.1693450583977953e-05, "loss": 1.2991008758544922, "step": 178 }, { "epoch": 0.4918032786885246, "grad_norm": 0.8777170777320862, "learning_rate": 2.1452285249292147e-05, "loss": 0.9539348483085632, "step": 180 }, { "epoch": 0.4972677595628415, "grad_norm": 0.5306017398834229, "learning_rate": 2.12092796372634e-05, "loss": 1.217163324356079, "step": 182 }, { "epoch": 0.5027322404371585, "grad_norm": 0.3937690556049347, "learning_rate": 2.096452404917679e-05, "loss": 1.1728050708770752, "step": 184 }, { "epoch": 0.5081967213114754, "grad_norm": 0.4189966022968292, "learning_rate": 2.0718109436611348e-05, "loss": 1.1901968717575073, "step": 186 }, { "epoch": 0.5136612021857924, "grad_norm": 0.8861513733863831, "learning_rate": 2.0470127367642345e-05, "loss": 0.4577289819717407, "step": 188 }, { "epoch": 0.5191256830601093, "grad_norm": 0.8669730424880981, "learning_rate": 2.022066999281444e-05, "loss": 1.170414924621582, "step": 190 }, { "epoch": 0.5245901639344263, "grad_norm": 0.43784815073013306, "learning_rate": 1.9969830010898358e-05, "loss": 1.2003910541534424, "step": 192 }, { "epoch": 0.5300546448087432, "grad_norm": 0.2408360242843628, "learning_rate": 1.9717700634443903e-05, "loss": 1.196523904800415, "step": 194 }, { "epoch": 0.5355191256830601, "grad_norm": 0.275288462638855, "learning_rate": 1.9464375555142e-05, "loss": 0.9004436135292053, "step": 196 }, { "epoch": 0.5409836065573771, "grad_norm": 0.30962035059928894, "learning_rate": 1.9209948909008734e-05, "loss": 1.002712368965149, "step": 198 }, { "epoch": 0.546448087431694, "grad_norm": 0.37051403522491455, "learning_rate": 1.8954515241404218e-05, "loss": 1.0420843362808228, "step": 200 }, { "epoch": 0.5519125683060109, "grad_norm": 0.37618082761764526, "learning_rate": 1.8698169471899414e-05, "loss": 1.0357520580291748, "step": 202 }, { "epoch": 0.5573770491803278, "grad_norm": 0.3667483627796173, "learning_rate": 1.8441006859003842e-05, "loss": 1.15482759475708, "step": 204 }, { "epoch": 0.5628415300546448, "grad_norm": 0.35124123096466064, "learning_rate": 1.818312296476737e-05, "loss": 1.1959636211395264, "step": 206 }, { "epoch": 0.5683060109289617, "grad_norm": 0.35074105858802795, "learning_rate": 1.792461361926921e-05, "loss": 1.087142825126648, "step": 208 }, { "epoch": 0.5737704918032787, "grad_norm": 0.40441736578941345, "learning_rate": 1.766557488500727e-05, "loss": 1.23688805103302, "step": 210 }, { "epoch": 0.5792349726775956, "grad_norm": 0.40216121077537537, "learning_rate": 1.7406103021201212e-05, "loss": 1.175810694694519, "step": 212 }, { "epoch": 0.5846994535519126, "grad_norm": 0.356263130903244, "learning_rate": 1.7146294448022335e-05, "loss": 1.2632404565811157, "step": 214 }, { "epoch": 0.5901639344262295, "grad_norm": 0.379120409488678, "learning_rate": 1.688624571076371e-05, "loss": 1.1749930381774902, "step": 216 }, { "epoch": 0.5956284153005464, "grad_norm": 0.4688524901866913, "learning_rate": 1.6626053443963762e-05, "loss": 1.2324018478393555, "step": 218 }, { "epoch": 0.6010928961748634, "grad_norm": 0.3071102797985077, "learning_rate": 1.636581433549674e-05, "loss": 1.0937501192092896, "step": 220 }, { "epoch": 0.6065573770491803, "grad_norm": 0.4043320417404175, "learning_rate": 1.610562509064332e-05, "loss": 0.7812168598175049, "step": 222 }, { "epoch": 0.6120218579234973, "grad_norm": 0.37222233414649963, "learning_rate": 1.5845582396154786e-05, "loss": 1.1552724838256836, "step": 224 }, { "epoch": 0.6174863387978142, "grad_norm": 0.32638704776763916, "learning_rate": 1.5585782884324064e-05, "loss": 1.1811174154281616, "step": 226 }, { "epoch": 0.6229508196721312, "grad_norm": 1.021145224571228, "learning_rate": 1.5326323097077015e-05, "loss": 0.7545455694198608, "step": 228 }, { "epoch": 0.6284153005464481, "grad_norm": 1.044705867767334, "learning_rate": 1.5067299450097261e-05, "loss": 1.1588557958602905, "step": 230 }, { "epoch": 0.6338797814207651, "grad_norm": 0.417707622051239, "learning_rate": 1.4808808196998006e-05, "loss": 1.1800490617752075, "step": 232 }, { "epoch": 0.639344262295082, "grad_norm": 0.4615285396575928, "learning_rate": 1.4550945393554004e-05, "loss": 0.9214609861373901, "step": 234 }, { "epoch": 0.644808743169399, "grad_norm": 0.9104841947555542, "learning_rate": 1.4293806862007085e-05, "loss": 1.4733829498291016, "step": 236 }, { "epoch": 0.6502732240437158, "grad_norm": 0.5479267239570618, "learning_rate": 1.4037488155458448e-05, "loss": 1.122392177581787, "step": 238 }, { "epoch": 0.6557377049180327, "grad_norm": 1.7201780080795288, "learning_rate": 1.3782084522360981e-05, "loss": 1.0747902393341064, "step": 240 }, { "epoch": 0.6612021857923497, "grad_norm": 0.2870098948478699, "learning_rate": 1.3527690871124762e-05, "loss": 0.7035287022590637, "step": 242 }, { "epoch": 0.6666666666666666, "grad_norm": 0.618334174156189, "learning_rate": 1.3274401734848958e-05, "loss": 1.2590125799179077, "step": 244 }, { "epoch": 0.6721311475409836, "grad_norm": 0.9182770848274231, "learning_rate": 1.3022311236193156e-05, "loss": 1.1282920837402344, "step": 246 }, { "epoch": 0.6775956284153005, "grad_norm": 0.3048880398273468, "learning_rate": 1.2771513052401236e-05, "loss": 1.1468207836151123, "step": 248 }, { "epoch": 0.6830601092896175, "grad_norm": 0.46364668011665344, "learning_rate": 1.2522100380490744e-05, "loss": 0.41445863246917725, "step": 250 }, { "epoch": 0.6885245901639344, "grad_norm": 0.3374931216239929, "learning_rate": 1.2274165902620732e-05, "loss": 1.205442190170288, "step": 252 }, { "epoch": 0.6939890710382514, "grad_norm": 0.30988192558288574, "learning_rate": 1.2027801751650918e-05, "loss": 1.1606954336166382, "step": 254 }, { "epoch": 0.6994535519125683, "grad_norm": 0.6835072040557861, "learning_rate": 1.1783099476904972e-05, "loss": 0.9375866651535034, "step": 256 }, { "epoch": 0.7049180327868853, "grad_norm": 0.5492672324180603, "learning_rate": 1.1540150010150599e-05, "loss": 1.1717281341552734, "step": 258 }, { "epoch": 0.7103825136612022, "grad_norm": 1.321418285369873, "learning_rate": 1.1299043631809205e-05, "loss": 1.0042099952697754, "step": 260 }, { "epoch": 0.7158469945355191, "grad_norm": 0.5885441303253174, "learning_rate": 1.1059869937407486e-05, "loss": 1.1690725088119507, "step": 262 }, { "epoch": 0.7213114754098361, "grad_norm": 0.4145934581756592, "learning_rate": 1.082271780428362e-05, "loss": 1.1503442525863647, "step": 264 }, { "epoch": 0.726775956284153, "grad_norm": 0.3795332610607147, "learning_rate": 1.0587675358560278e-05, "loss": 0.7821759581565857, "step": 266 }, { "epoch": 0.73224043715847, "grad_norm": 0.34867045283317566, "learning_rate": 1.0354829942396837e-05, "loss": 1.1533390283584595, "step": 268 }, { "epoch": 0.7377049180327869, "grad_norm": 0.3432830572128296, "learning_rate": 1.012426808153287e-05, "loss": 1.1842344999313354, "step": 270 }, { "epoch": 0.7431693989071039, "grad_norm": 0.37027257680892944, "learning_rate": 9.896075453135039e-06, "loss": 1.2416757345199585, "step": 272 }, { "epoch": 0.7486338797814208, "grad_norm": 0.8126760721206665, "learning_rate": 9.67033685395934e-06, "loss": 1.2841261625289917, "step": 274 }, { "epoch": 0.7540983606557377, "grad_norm": 0.7494352459907532, "learning_rate": 9.447136168840466e-06, "loss": 1.2585190534591675, "step": 276 }, { "epoch": 0.7595628415300546, "grad_norm": 0.6677407026290894, "learning_rate": 9.226556339520069e-06, "loss": 0.826260507106781, "step": 278 }, { "epoch": 0.7650273224043715, "grad_norm": 0.294973224401474, "learning_rate": 9.008679333825478e-06, "loss": 1.1650235652923584, "step": 280 }, { "epoch": 0.7704918032786885, "grad_norm": 0.4066277742385864, "learning_rate": 8.793586115210326e-06, "loss": 1.0734792947769165, "step": 282 }, { "epoch": 0.7759562841530054, "grad_norm": 0.24723787605762482, "learning_rate": 8.581356612668382e-06, "loss": 1.1951000690460205, "step": 284 }, { "epoch": 0.7814207650273224, "grad_norm": 0.7921513319015503, "learning_rate": 8.372069691031804e-06, "loss": 1.210968017578125, "step": 286 }, { "epoch": 0.7868852459016393, "grad_norm": 0.9979972243309021, "learning_rate": 8.165803121664869e-06, "loss": 0.7500867247581482, "step": 288 }, { "epoch": 0.7923497267759563, "grad_norm": 0.3506968915462494, "learning_rate": 7.962633553563965e-06, "loss": 0.8001227974891663, "step": 290 }, { "epoch": 0.7978142076502732, "grad_norm": 0.33423033356666565, "learning_rate": 7.762636484874723e-06, "loss": 0.9556211233139038, "step": 292 }, { "epoch": 0.8032786885245902, "grad_norm": 0.3771698474884033, "learning_rate": 7.565886234836767e-06, "loss": 1.159753441810608, "step": 294 }, { "epoch": 0.8087431693989071, "grad_norm": 0.36711210012435913, "learning_rate": 7.3724559161665876e-06, "loss": 1.1728172302246094, "step": 296 }, { "epoch": 0.8142076502732241, "grad_norm": 0.3197503685951233, "learning_rate": 7.182417407888703e-06, "loss": 1.1757210493087769, "step": 298 }, { "epoch": 0.819672131147541, "grad_norm": 0.3136984705924988, "learning_rate": 6.995841328625321e-06, "loss": 0.8655977845191956, "step": 300 }, { "epoch": 0.825136612021858, "grad_norm": 0.393072247505188, "learning_rate": 6.812797010354325e-06, "loss": 1.1789922714233398, "step": 302 }, { "epoch": 0.8306010928961749, "grad_norm": 3.821265935897827, "learning_rate": 6.63335247264542e-06, "loss": 1.08378005027771, "step": 304 }, { "epoch": 0.8360655737704918, "grad_norm": 0.33668431639671326, "learning_rate": 6.457574397383919e-06, "loss": 1.1709719896316528, "step": 306 }, { "epoch": 0.8415300546448088, "grad_norm": 0.3578186631202698, "learning_rate": 6.285528103991665e-06, "loss": 1.193668246269226, "step": 308 }, { "epoch": 0.8469945355191257, "grad_norm": 0.3615029752254486, "learning_rate": 6.117277525154225e-06, "loss": 1.15645432472229, "step": 310 }, { "epoch": 0.8524590163934426, "grad_norm": 0.38210004568099976, "learning_rate": 5.952885183063397e-06, "loss": 1.1638695001602173, "step": 312 }, { "epoch": 0.8579234972677595, "grad_norm": 0.35445594787597656, "learning_rate": 5.792412166183841e-06, "loss": 1.1909141540527344, "step": 314 }, { "epoch": 0.8633879781420765, "grad_norm": 0.5096743702888489, "learning_rate": 5.635918106552546e-06, "loss": 1.201849102973938, "step": 316 }, { "epoch": 0.8688524590163934, "grad_norm": 1.2665657997131348, "learning_rate": 5.483461157619428e-06, "loss": 0.7697218656539917, "step": 318 }, { "epoch": 0.8743169398907104, "grad_norm": 0.31564080715179443, "learning_rate": 5.335097972637441e-06, "loss": 1.154910683631897, "step": 320 }, { "epoch": 0.8797814207650273, "grad_norm": 0.41027265787124634, "learning_rate": 5.1908836836101135e-06, "loss": 0.810014545917511, "step": 322 }, { "epoch": 0.8852459016393442, "grad_norm": 0.5105057954788208, "learning_rate": 5.050871880804414e-06, "loss": 1.4533857107162476, "step": 324 }, { "epoch": 0.8907103825136612, "grad_norm": 0.35378292202949524, "learning_rate": 4.915114592836521e-06, "loss": 1.1535991430282593, "step": 326 }, { "epoch": 0.8961748633879781, "grad_norm": 1.3628326654434204, "learning_rate": 4.783662267337909e-06, "loss": 1.1089421510696411, "step": 328 }, { "epoch": 0.9016393442622951, "grad_norm": 0.2328273057937622, "learning_rate": 4.656563752208907e-06, "loss": 1.2124766111373901, "step": 330 }, { "epoch": 0.907103825136612, "grad_norm": 0.31986290216445923, "learning_rate": 4.533866277466767e-06, "loss": 1.1944669485092163, "step": 332 }, { "epoch": 0.912568306010929, "grad_norm": 1.5621281862258911, "learning_rate": 4.415615437694876e-06, "loss": 1.3944669961929321, "step": 334 }, { "epoch": 0.9180327868852459, "grad_norm": 2.3418469429016113, "learning_rate": 4.3018551750997694e-06, "loss": 1.1178382635116577, "step": 336 }, { "epoch": 0.9234972677595629, "grad_norm": 0.40905457735061646, "learning_rate": 4.192627763182111e-06, "loss": 1.1970925331115723, "step": 338 }, { "epoch": 0.9289617486338798, "grad_norm": 1.2464805841445923, "learning_rate": 4.087973791027797e-06, "loss": 0.7709401249885559, "step": 340 }, { "epoch": 0.9344262295081968, "grad_norm": 0.713545560836792, "learning_rate": 3.987932148224993e-06, "loss": 1.1292576789855957, "step": 342 }, { "epoch": 0.9398907103825137, "grad_norm": 1.2035443782806396, "learning_rate": 3.8925400104126834e-06, "loss": 0.7497198581695557, "step": 344 }, { "epoch": 0.9453551912568307, "grad_norm": 0.3672787845134735, "learning_rate": 3.8018328254661618e-06, "loss": 1.1717313528060913, "step": 346 }, { "epoch": 0.9508196721311475, "grad_norm": 0.602628767490387, "learning_rate": 3.715844300324527e-06, "loss": 0.9975014328956604, "step": 348 }, { "epoch": 0.9562841530054644, "grad_norm": 0.4423633813858032, "learning_rate": 3.6346063884651327e-06, "loss": 1.1410292387008667, "step": 350 }, { "epoch": 0.9617486338797814, "grad_norm": 0.7007730007171631, "learning_rate": 3.558149278029624e-06, "loss": 1.3128447532653809, "step": 352 }, { "epoch": 0.9672131147540983, "grad_norm": 0.439485639333725, "learning_rate": 3.4865013806059817e-06, "loss": 0.7662681341171265, "step": 354 }, { "epoch": 0.9726775956284153, "grad_norm": 0.4370954930782318, "learning_rate": 3.419689320670712e-06, "loss": 0.7777884602546692, "step": 356 }, { "epoch": 0.9781420765027322, "grad_norm": 1.6398391723632812, "learning_rate": 3.35773792569517e-06, "loss": 1.1536012887954712, "step": 358 }, { "epoch": 0.9836065573770492, "grad_norm": 0.42289936542510986, "learning_rate": 3.300670216919602e-06, "loss": 0.8657046556472778, "step": 360 }, { "epoch": 0.9890710382513661, "grad_norm": 0.49119001626968384, "learning_rate": 3.2485074007984468e-06, "loss": 1.1398602724075317, "step": 362 }, { "epoch": 0.994535519125683, "grad_norm": 0.897072434425354, "learning_rate": 3.2012688611199566e-06, "loss": 1.2387840747833252, "step": 364 }, { "epoch": 1.0, "grad_norm": 0.38847413659095764, "learning_rate": 3.158972151803165e-06, "loss": 1.236254334449768, "step": 366 }, { "epoch": 1.005464480874317, "grad_norm": 0.31760314106941223, "learning_rate": 3.1216329903748095e-06, "loss": 0.761336088180542, "step": 368 }, { "epoch": 1.010928961748634, "grad_norm": 0.3967256247997284, "learning_rate": 3.089265252128686e-06, "loss": 1.2508467435836792, "step": 370 }, { "epoch": 1.0163934426229508, "grad_norm": 0.4457378387451172, "learning_rate": 3.061880964969555e-06, "loss": 1.246903419494629, "step": 372 }, { "epoch": 1.0218579234972678, "grad_norm": 0.3824722170829773, "learning_rate": 3.039490304943562e-06, "loss": 1.1185325384140015, "step": 374 }, { "epoch": 1.0273224043715847, "grad_norm": 0.3070431053638458, "learning_rate": 3.022101592456795e-06, "loss": 1.0686273574829102, "step": 376 }, { "epoch": 1.0327868852459017, "grad_norm": 0.28226983547210693, "learning_rate": 3.0097212891834095e-06, "loss": 1.0337367057800293, "step": 378 }, { "epoch": 1.0382513661202186, "grad_norm": 0.4869016110897064, "learning_rate": 3.0023539956644634e-06, "loss": 1.0311717987060547, "step": 380 }, { "epoch": 1.0437158469945356, "grad_norm": 0.8694993257522583, "learning_rate": 3.0000024495983428e-06, "loss": 0.7782408595085144, "step": 382 }, { "epoch": 1.0491803278688525, "grad_norm": 0.5946288108825684, "learning_rate": 3.002667524823434e-06, "loss": 1.0987261533737183, "step": 384 }, { "epoch": 1.0546448087431695, "grad_norm": 1.0870792865753174, "learning_rate": 3.010348230993402e-06, "loss": 1.0993095636367798, "step": 386 }, { "epoch": 1.0601092896174864, "grad_norm": 0.5876205563545227, "learning_rate": 3.0230417139451987e-06, "loss": 1.4858818054199219, "step": 388 }, { "epoch": 1.0655737704918034, "grad_norm": 0.8428175449371338, "learning_rate": 3.0407432567596883e-06, "loss": 0.9062157273292542, "step": 390 }, { "epoch": 1.0710382513661203, "grad_norm": 1.2778754234313965, "learning_rate": 3.0634462815144474e-06, "loss": 0.592805802822113, "step": 392 }, { "epoch": 1.0765027322404372, "grad_norm": 0.4570184648036957, "learning_rate": 3.0911423517281404e-06, "loss": 1.1192593574523926, "step": 394 }, { "epoch": 1.0819672131147542, "grad_norm": 0.42796361446380615, "learning_rate": 3.1238211754955294e-06, "loss": 1.0443531274795532, "step": 396 }, { "epoch": 1.0874316939890711, "grad_norm": 0.43680837750434875, "learning_rate": 3.161470609311961e-06, "loss": 1.0404291152954102, "step": 398 }, { "epoch": 1.092896174863388, "grad_norm": 0.40812206268310547, "learning_rate": 3.2040766625859115e-06, "loss": 1.1858930587768555, "step": 400 }, { "epoch": 1.098360655737705, "grad_norm": 0.5379334688186646, "learning_rate": 3.2516235028379157e-06, "loss": 1.0395185947418213, "step": 402 }, { "epoch": 1.1038251366120218, "grad_norm": 0.40824946761131287, "learning_rate": 3.304093461583944e-06, "loss": 1.026267170906067, "step": 404 }, { "epoch": 1.1092896174863387, "grad_norm": 0.603408694267273, "learning_rate": 3.3614670409010353e-06, "loss": 1.076138973236084, "step": 406 }, { "epoch": 1.1147540983606556, "grad_norm": 0.41980448365211487, "learning_rate": 3.4237229206727602e-06, "loss": 1.3695933818817139, "step": 408 }, { "epoch": 1.1202185792349726, "grad_norm": 0.869969367980957, "learning_rate": 3.490837966511817e-06, "loss": 1.047141671180725, "step": 410 }, { "epoch": 1.1256830601092895, "grad_norm": 0.38264161348342896, "learning_rate": 3.5627872383567937e-06, "loss": 1.0203361511230469, "step": 412 }, { "epoch": 1.1311475409836065, "grad_norm": 0.30198222398757935, "learning_rate": 3.6395439997399494e-06, "loss": 0.5207374095916748, "step": 414 }, { "epoch": 1.1366120218579234, "grad_norm": 1.9066449403762817, "learning_rate": 3.721079727722522e-06, "loss": 1.2205498218536377, "step": 416 }, { "epoch": 1.1420765027322404, "grad_norm": 0.28766530752182007, "learning_rate": 3.8073641234939055e-06, "loss": 1.1669960021972656, "step": 418 }, { "epoch": 1.1475409836065573, "grad_norm": 0.4494319558143616, "learning_rate": 3.898365123630732e-06, "loss": 1.0348408222198486, "step": 420 }, { "epoch": 1.1530054644808743, "grad_norm": 0.7580645680427551, "learning_rate": 3.994048912011692e-06, "loss": 1.0489870309829712, "step": 422 }, { "epoch": 1.1584699453551912, "grad_norm": 0.30486011505126953, "learning_rate": 4.094379932383666e-06, "loss": 1.0182819366455078, "step": 424 }, { "epoch": 1.1639344262295082, "grad_norm": 0.5283482074737549, "learning_rate": 4.199320901574489e-06, "loss": 0.7681307792663574, "step": 426 }, { "epoch": 1.169398907103825, "grad_norm": 0.6967251896858215, "learning_rate": 4.3088328233474185e-06, "loss": 1.1920503377914429, "step": 428 }, { "epoch": 1.174863387978142, "grad_norm": 0.30848848819732666, "learning_rate": 4.422875002892234e-06, "loss": 1.1227927207946777, "step": 430 }, { "epoch": 1.180327868852459, "grad_norm": 0.4338774085044861, "learning_rate": 4.54140506194747e-06, "loss": 1.0621066093444824, "step": 432 }, { "epoch": 1.185792349726776, "grad_norm": 0.5365703701972961, "learning_rate": 4.664378954548241e-06, "loss": 1.0539458990097046, "step": 434 }, { "epoch": 1.1912568306010929, "grad_norm": 0.9198355674743652, "learning_rate": 4.791750983393832e-06, "loss": 0.6073042750358582, "step": 436 }, { "epoch": 1.1967213114754098, "grad_norm": 0.2908204197883606, "learning_rate": 4.9234738168288466e-06, "loss": 0.4756653904914856, "step": 438 }, { "epoch": 1.2021857923497268, "grad_norm": 0.43431609869003296, "learning_rate": 5.059498506431758e-06, "loss": 0.9815635085105896, "step": 440 }, { "epoch": 1.2076502732240437, "grad_norm": 0.8783625364303589, "learning_rate": 5.199774505204206e-06, "loss": 1.324447512626648, "step": 442 }, { "epoch": 1.2131147540983607, "grad_norm": 0.8234128355979919, "learning_rate": 5.344249686354357e-06, "loss": 1.26932692527771, "step": 444 }, { "epoch": 1.2185792349726776, "grad_norm": 0.35708698630332947, "learning_rate": 5.492870362667299e-06, "loss": 0.5943037271499634, "step": 446 }, { "epoch": 1.2240437158469946, "grad_norm": 0.3496783375740051, "learning_rate": 5.645581306455302e-06, "loss": 0.8465991616249084, "step": 448 }, { "epoch": 1.2295081967213115, "grad_norm": 0.34306076169013977, "learning_rate": 5.802325770080506e-06, "loss": 1.0558314323425293, "step": 450 }, { "epoch": 1.2349726775956285, "grad_norm": 0.6199325919151306, "learning_rate": 5.96304550704246e-06, "loss": 0.7405421137809753, "step": 452 }, { "epoch": 1.2404371584699454, "grad_norm": 0.5972355604171753, "learning_rate": 6.127680793622588e-06, "loss": 1.023563027381897, "step": 454 }, { "epoch": 1.2459016393442623, "grad_norm": 0.36354729533195496, "learning_rate": 6.296170451077657e-06, "loss": 1.149239420890808, "step": 456 }, { "epoch": 1.2513661202185793, "grad_norm": 0.5815706849098206, "learning_rate": 6.468451868373856e-06, "loss": 0.5993931889533997, "step": 458 }, { "epoch": 1.2568306010928962, "grad_norm": 0.884687066078186, "learning_rate": 6.6444610254532e-06, "loss": 1.1250090599060059, "step": 460 }, { "epoch": 1.2622950819672132, "grad_norm": 0.5644488334655762, "learning_rate": 6.824132517023449e-06, "loss": 1.117283821105957, "step": 462 }, { "epoch": 1.2677595628415301, "grad_norm": 1.1574392318725586, "learning_rate": 7.007399576862872e-06, "loss": 0.8585411906242371, "step": 464 }, { "epoch": 1.273224043715847, "grad_norm": 0.31241562962532043, "learning_rate": 7.1941941026306275e-06, "loss": 1.159734845161438, "step": 466 }, { "epoch": 1.278688524590164, "grad_norm": 3.079608678817749, "learning_rate": 7.3844466811737555e-06, "loss": 0.43219655752182007, "step": 468 }, { "epoch": 1.2841530054644807, "grad_norm": 1.1515812873840332, "learning_rate": 7.578086614321175e-06, "loss": 0.32784971594810486, "step": 470 }, { "epoch": 1.289617486338798, "grad_norm": 0.8713538646697998, "learning_rate": 7.775041945155295e-06, "loss": 0.972320556640625, "step": 472 }, { "epoch": 1.2950819672131146, "grad_norm": 1.0352020263671875, "learning_rate": 7.975239484751258e-06, "loss": 0.8251127004623413, "step": 474 }, { "epoch": 1.3005464480874318, "grad_norm": 0.41140851378440857, "learning_rate": 8.178604839374125e-06, "loss": 1.0359759330749512, "step": 476 }, { "epoch": 1.3060109289617485, "grad_norm": 0.30868738889694214, "learning_rate": 8.385062438123673e-06, "loss": 1.13455331325531, "step": 478 }, { "epoch": 1.3114754098360657, "grad_norm": 0.603076159954071, "learning_rate": 8.594535561016661e-06, "loss": 1.1488956212997437, "step": 480 }, { "epoch": 1.3169398907103824, "grad_norm": 0.32255861163139343, "learning_rate": 8.806946367496155e-06, "loss": 1.1327383518218994, "step": 482 }, { "epoch": 1.3224043715846996, "grad_norm": 0.37436342239379883, "learning_rate": 9.02221592535712e-06, "loss": 1.0864595174789429, "step": 484 }, { "epoch": 1.3278688524590163, "grad_norm": 0.4781150221824646, "learning_rate": 9.240264240077859e-06, "loss": 0.2114480584859848, "step": 486 }, { "epoch": 1.3333333333333333, "grad_norm": 0.7249221205711365, "learning_rate": 9.461010284546016e-06, "loss": 1.14420747756958, "step": 488 }, { "epoch": 1.3387978142076502, "grad_norm": 0.31920719146728516, "learning_rate": 9.684372029168438e-06, "loss": 0.9852396249771118, "step": 490 }, { "epoch": 1.3442622950819672, "grad_norm": 0.31214451789855957, "learning_rate": 9.91026647235348e-06, "loss": 1.109579086303711, "step": 492 }, { "epoch": 1.349726775956284, "grad_norm": 1.5647644996643066, "learning_rate": 1.0138609671354586e-05, "loss": 0.6934401988983154, "step": 494 }, { "epoch": 1.355191256830601, "grad_norm": 0.3006077706813812, "learning_rate": 1.0369316773463458e-05, "loss": 1.0647120475769043, "step": 496 }, { "epoch": 1.360655737704918, "grad_norm": 0.3098089098930359, "learning_rate": 1.0602302047541566e-05, "loss": 0.8615930676460266, "step": 498 }, { "epoch": 1.366120218579235, "grad_norm": 0.30825164914131165, "learning_rate": 1.083747891587788e-05, "loss": 0.9996722340583801, "step": 500 }, { "epoch": 1.3715846994535519, "grad_norm": 0.31744763255119324, "learning_rate": 1.1074759986361392e-05, "loss": 1.1010159254074097, "step": 502 }, { "epoch": 1.3770491803278688, "grad_norm": 0.660419225692749, "learning_rate": 1.1314057084956073e-05, "loss": 0.9897991418838501, "step": 504 }, { "epoch": 1.3825136612021858, "grad_norm": 0.4726075530052185, "learning_rate": 1.1555281288466553e-05, "loss": 0.7116650342941284, "step": 506 }, { "epoch": 1.3879781420765027, "grad_norm": 0.4236489236354828, "learning_rate": 1.1798342957582084e-05, "loss": 0.7277122139930725, "step": 508 }, { "epoch": 1.3934426229508197, "grad_norm": 0.3505037724971771, "learning_rate": 1.2043151770186725e-05, "loss": 1.0869934558868408, "step": 510 }, { "epoch": 1.3989071038251366, "grad_norm": 0.4062495529651642, "learning_rate": 1.2289616754923078e-05, "loss": 1.0307718515396118, "step": 512 }, { "epoch": 1.4043715846994536, "grad_norm": 0.9875432252883911, "learning_rate": 1.253764632499752e-05, "loss": 0.6626445055007935, "step": 514 }, { "epoch": 1.4098360655737705, "grad_norm": 0.6780223846435547, "learning_rate": 1.2787148312213901e-05, "loss": 1.4220600128173828, "step": 516 }, { "epoch": 1.4153005464480874, "grad_norm": 0.32189661264419556, "learning_rate": 1.3038030001223439e-05, "loss": 0.7831554412841797, "step": 518 }, { "epoch": 1.4207650273224044, "grad_norm": 0.38265976309776306, "learning_rate": 1.3290198163977933e-05, "loss": 0.7503346800804138, "step": 520 }, { "epoch": 1.4262295081967213, "grad_norm": 0.7597147226333618, "learning_rate": 1.3543559094373372e-05, "loss": 0.7089603543281555, "step": 522 }, { "epoch": 1.4316939890710383, "grad_norm": 0.4683310091495514, "learning_rate": 1.3798018643071386e-05, "loss": 0.7786340117454529, "step": 524 }, { "epoch": 1.4371584699453552, "grad_norm": 0.4911711812019348, "learning_rate": 1.4053482252485178e-05, "loss": 0.9898138642311096, "step": 526 }, { "epoch": 1.4426229508196722, "grad_norm": 0.29452064633369446, "learning_rate": 1.4309854991917388e-05, "loss": 1.0076603889465332, "step": 528 }, { "epoch": 1.4480874316939891, "grad_norm": 0.31045207381248474, "learning_rate": 1.4567041592836413e-05, "loss": 0.6953315734863281, "step": 530 }, { "epoch": 1.453551912568306, "grad_norm": 0.39866623282432556, "learning_rate": 1.48249464842784e-05, "loss": 1.1244227886199951, "step": 532 }, { "epoch": 1.459016393442623, "grad_norm": 0.4015820026397705, "learning_rate": 1.508347382836153e-05, "loss": 1.071346402168274, "step": 534 }, { "epoch": 1.46448087431694, "grad_norm": 0.4267948269844055, "learning_rate": 1.534252755589961e-05, "loss": 1.0431689023971558, "step": 536 }, { "epoch": 1.469945355191257, "grad_norm": 0.24437487125396729, "learning_rate": 1.5602011402101432e-05, "loss": 0.6640901565551758, "step": 538 }, { "epoch": 1.4754098360655736, "grad_norm": 0.2949664890766144, "learning_rate": 1.5861828942343037e-05, "loss": 0.8600192070007324, "step": 540 }, { "epoch": 1.4808743169398908, "grad_norm": 3.951502561569214, "learning_rate": 1.612188362799917e-05, "loss": 1.1732059717178345, "step": 542 }, { "epoch": 1.4863387978142075, "grad_norm": 0.34538793563842773, "learning_rate": 1.6382078822320964e-05, "loss": 1.0121636390686035, "step": 544 }, { "epoch": 1.4918032786885247, "grad_norm": 0.8434890508651733, "learning_rate": 1.6642317836346324e-05, "loss": 0.7092042565345764, "step": 546 }, { "epoch": 1.4972677595628414, "grad_norm": 0.849798321723938, "learning_rate": 1.6902503964829644e-05, "loss": 0.8493767380714417, "step": 548 }, { "epoch": 1.5027322404371586, "grad_norm": 0.2791331708431244, "learning_rate": 1.7162540522177685e-05, "loss": 1.0140348672866821, "step": 550 }, { "epoch": 1.5081967213114753, "grad_norm": 0.6075169444084167, "learning_rate": 1.7422330878378113e-05, "loss": 0.8161216974258423, "step": 552 }, { "epoch": 1.5136612021857925, "grad_norm": 0.361250638961792, "learning_rate": 1.7681778494907298e-05, "loss": 1.0384575128555298, "step": 554 }, { "epoch": 1.5191256830601092, "grad_norm": 0.3741307854652405, "learning_rate": 1.794078696060429e-05, "loss": 0.7678899168968201, "step": 556 }, { "epoch": 1.5245901639344264, "grad_norm": 0.3114408552646637, "learning_rate": 1.819926002749727e-05, "loss": 1.02855384349823, "step": 558 }, { "epoch": 1.530054644808743, "grad_norm": 0.18622533977031708, "learning_rate": 1.84571016465695e-05, "loss": 0.2526892423629761, "step": 560 }, { "epoch": 1.5355191256830603, "grad_norm": 0.23625467717647552, "learning_rate": 1.8714216003451295e-05, "loss": 1.0640628337860107, "step": 562 }, { "epoch": 1.540983606557377, "grad_norm": 0.724408745765686, "learning_rate": 1.8970507554024827e-05, "loss": 0.7283908724784851, "step": 564 }, { "epoch": 1.5464480874316942, "grad_norm": 0.3644772171974182, "learning_rate": 1.922588105992838e-05, "loss": 1.1399354934692383, "step": 566 }, { "epoch": 1.5519125683060109, "grad_norm": 0.9722273349761963, "learning_rate": 1.9480241623947206e-05, "loss": 1.1557573080062866, "step": 568 }, { "epoch": 1.5573770491803278, "grad_norm": 0.4744189381599426, "learning_rate": 1.9733494725277413e-05, "loss": 1.133260726928711, "step": 570 }, { "epoch": 1.5628415300546448, "grad_norm": 2.321540594100952, "learning_rate": 1.998554625465005e-05, "loss": 0.6360920071601868, "step": 572 }, { "epoch": 1.5683060109289617, "grad_norm": 0.6282745599746704, "learning_rate": 2.0236302549302293e-05, "loss": 0.9884318709373474, "step": 574 }, { "epoch": 1.5737704918032787, "grad_norm": 0.25726932287216187, "learning_rate": 2.0485670427782644e-05, "loss": 1.072218656539917, "step": 576 }, { "epoch": 1.5792349726775956, "grad_norm": 0.3337240517139435, "learning_rate": 2.073355722457739e-05, "loss": 1.0474885702133179, "step": 578 }, { "epoch": 1.5846994535519126, "grad_norm": 0.3079088628292084, "learning_rate": 2.0979870824545165e-05, "loss": 1.0121105909347534, "step": 580 }, { "epoch": 1.5901639344262295, "grad_norm": 0.33754849433898926, "learning_rate": 2.1224519697147145e-05, "loss": 1.059688925743103, "step": 582 }, { "epoch": 1.5956284153005464, "grad_norm": 0.8730252385139465, "learning_rate": 2.1467412930459936e-05, "loss": 0.5801796317100525, "step": 584 }, { "epoch": 1.6010928961748634, "grad_norm": 1.4568897485733032, "learning_rate": 2.1708460264958595e-05, "loss": 1.0416557788848877, "step": 586 }, { "epoch": 1.6065573770491803, "grad_norm": 0.33535847067832947, "learning_rate": 2.194757212705718e-05, "loss": 1.0528165102005005, "step": 588 }, { "epoch": 1.6120218579234973, "grad_norm": 0.39254218339920044, "learning_rate": 2.2184659662394522e-05, "loss": 1.2890545129776, "step": 590 }, { "epoch": 1.6174863387978142, "grad_norm": 0.39055129885673523, "learning_rate": 2.24196347688526e-05, "loss": 1.0770164728164673, "step": 592 }, { "epoch": 1.6229508196721312, "grad_norm": 0.3632502853870392, "learning_rate": 2.265241012929541e-05, "loss": 1.0873066186904907, "step": 594 }, { "epoch": 1.6284153005464481, "grad_norm": 0.43228334188461304, "learning_rate": 2.28828992440162e-05, "loss": 1.009541630744934, "step": 596 }, { "epoch": 1.633879781420765, "grad_norm": 0.34224361181259155, "learning_rate": 2.3111016462880873e-05, "loss": 1.3436585664749146, "step": 598 }, { "epoch": 1.639344262295082, "grad_norm": 1.0117156505584717, "learning_rate": 2.333667701715578e-05, "loss": 0.9953033328056335, "step": 600 }, { "epoch": 1.644808743169399, "grad_norm": 0.4946202337741852, "learning_rate": 2.3559797051007815e-05, "loss": 1.3335758447647095, "step": 602 }, { "epoch": 1.650273224043716, "grad_norm": 0.30347269773483276, "learning_rate": 2.3780293652665477e-05, "loss": 0.9623563289642334, "step": 604 }, { "epoch": 1.6557377049180326, "grad_norm": 0.32685670256614685, "learning_rate": 2.399808488522895e-05, "loss": 1.137532353401184, "step": 606 }, { "epoch": 1.6612021857923498, "grad_norm": 0.3293339014053345, "learning_rate": 2.4213089817118078e-05, "loss": 1.055052399635315, "step": 608 }, { "epoch": 1.6666666666666665, "grad_norm": 0.46230047941207886, "learning_rate": 2.4425228552146573e-05, "loss": 1.1524425745010376, "step": 610 }, { "epoch": 1.6721311475409837, "grad_norm": 0.6461423635482788, "learning_rate": 2.4634422259211614e-05, "loss": 0.6941080689430237, "step": 612 }, { "epoch": 1.6775956284153004, "grad_norm": 1.0476462841033936, "learning_rate": 2.4840593201587626e-05, "loss": 0.8486734628677368, "step": 614 }, { "epoch": 1.6830601092896176, "grad_norm": 2.0571165084838867, "learning_rate": 2.5043664765813377e-05, "loss": 0.7663911581039429, "step": 616 }, { "epoch": 1.6885245901639343, "grad_norm": 0.4737246334552765, "learning_rate": 2.524356149016163e-05, "loss": 1.0844976902008057, "step": 618 }, { "epoch": 1.6939890710382515, "grad_norm": 0.3629366457462311, "learning_rate": 2.544020909268085e-05, "loss": 1.0529444217681885, "step": 620 }, { "epoch": 1.6994535519125682, "grad_norm": 0.3002009689807892, "learning_rate": 2.5633534498798598e-05, "loss": 1.004475474357605, "step": 622 }, { "epoch": 1.7049180327868854, "grad_norm": 6.076183319091797, "learning_rate": 2.5823465868475985e-05, "loss": 0.6815961599349976, "step": 624 }, { "epoch": 1.710382513661202, "grad_norm": 0.3336254954338074, "learning_rate": 2.60099326229037e-05, "loss": 1.1199313402175903, "step": 626 }, { "epoch": 1.7158469945355193, "grad_norm": 0.39660659432411194, "learning_rate": 2.619286547072914e-05, "loss": 1.038909673690796, "step": 628 }, { "epoch": 1.721311475409836, "grad_norm": 3.282057523727417, "learning_rate": 2.6372196433805214e-05, "loss": 0.709031879901886, "step": 630 }, { "epoch": 1.7267759562841531, "grad_norm": 0.4394005835056305, "learning_rate": 2.654785887245112e-05, "loss": 0.7367001175880432, "step": 632 }, { "epoch": 1.7322404371584699, "grad_norm": 0.32146531343460083, "learning_rate": 2.671978751021577e-05, "loss": 1.2364336252212524, "step": 634 }, { "epoch": 1.737704918032787, "grad_norm": 0.4605696499347687, "learning_rate": 2.6887918458134622e-05, "loss": 0.9947465658187866, "step": 636 }, { "epoch": 1.7431693989071038, "grad_norm": 0.3082851469516754, "learning_rate": 2.705218923847093e-05, "loss": 1.0885943174362183, "step": 638 }, { "epoch": 1.748633879781421, "grad_norm": 0.31364598870277405, "learning_rate": 2.7212538807932576e-05, "loss": 1.113791823387146, "step": 640 }, { "epoch": 1.7540983606557377, "grad_norm": 0.375470906496048, "learning_rate": 2.7368907580355843e-05, "loss": 0.7379390001296997, "step": 642 }, { "epoch": 1.7595628415300546, "grad_norm": 0.7673870921134949, "learning_rate": 2.7521237448847734e-05, "loss": 0.3575439751148224, "step": 644 }, { "epoch": 1.7650273224043715, "grad_norm": 0.4010154604911804, "learning_rate": 2.766947180737861e-05, "loss": 1.1416033506393433, "step": 646 }, { "epoch": 1.7704918032786885, "grad_norm": 0.27221986651420593, "learning_rate": 2.781355557181706e-05, "loss": 0.8822421431541443, "step": 648 }, { "epoch": 1.7759562841530054, "grad_norm": 0.496616005897522, "learning_rate": 2.7953435200399262e-05, "loss": 0.7749378085136414, "step": 650 }, { "epoch": 1.7814207650273224, "grad_norm": 0.7093772888183594, "learning_rate": 2.8089058713625194e-05, "loss": 0.7831916809082031, "step": 652 }, { "epoch": 1.7868852459016393, "grad_norm": 0.351391464471817, "learning_rate": 2.8220375713574307e-05, "loss": 1.0371301174163818, "step": 654 }, { "epoch": 1.7923497267759563, "grad_norm": 0.2960720658302307, "learning_rate": 2.8347337402633456e-05, "loss": 1.0718927383422852, "step": 656 }, { "epoch": 1.7978142076502732, "grad_norm": 0.6162258386611938, "learning_rate": 2.846989660163019e-05, "loss": 0.9962674379348755, "step": 658 }, { "epoch": 1.8032786885245902, "grad_norm": 0.4377414286136627, "learning_rate": 2.858800776736461e-05, "loss": 1.1022226810455322, "step": 660 }, { "epoch": 1.8087431693989071, "grad_norm": 0.2909747064113617, "learning_rate": 2.87016270095333e-05, "loss": 0.8855769634246826, "step": 662 }, { "epoch": 1.814207650273224, "grad_norm": 1.3076122999191284, "learning_rate": 2.8810712107039e-05, "loss": 0.6500188708305359, "step": 664 }, { "epoch": 1.819672131147541, "grad_norm": 0.303838849067688, "learning_rate": 2.8915222523680082e-05, "loss": 1.117383360862732, "step": 666 }, { "epoch": 1.825136612021858, "grad_norm": 0.28003278374671936, "learning_rate": 2.9015119423213857e-05, "loss": 1.067117691040039, "step": 668 }, { "epoch": 1.830601092896175, "grad_norm": 0.5866938829421997, "learning_rate": 2.9110365683788173e-05, "loss": 1.0645062923431396, "step": 670 }, { "epoch": 1.8360655737704918, "grad_norm": 0.4505460262298584, "learning_rate": 2.9200925911735956e-05, "loss": 1.0651450157165527, "step": 672 }, { "epoch": 1.8415300546448088, "grad_norm": 1.2405760288238525, "learning_rate": 2.9286766454727563e-05, "loss": 0.7814494967460632, "step": 674 }, { "epoch": 1.8469945355191257, "grad_norm": 0.34474292397499084, "learning_rate": 2.9367855414276073e-05, "loss": 1.151155710220337, "step": 676 }, { "epoch": 1.8524590163934427, "grad_norm": 0.3130297064781189, "learning_rate": 2.9444162657590747e-05, "loss": 0.960390567779541, "step": 678 }, { "epoch": 1.8579234972677594, "grad_norm": 0.37524649500846863, "learning_rate": 2.951565982877447e-05, "loss": 1.011980414390564, "step": 680 }, { "epoch": 1.8633879781420766, "grad_norm": 0.2684651017189026, "learning_rate": 2.9582320359360864e-05, "loss": 1.1165417432785034, "step": 682 }, { "epoch": 1.8688524590163933, "grad_norm": 0.6099388003349304, "learning_rate": 2.9644119478187126e-05, "loss": 0.9011784195899963, "step": 684 }, { "epoch": 1.8743169398907105, "grad_norm": 0.7181943655014038, "learning_rate": 2.9701034220599074e-05, "loss": 0.7465716004371643, "step": 686 }, { "epoch": 1.8797814207650272, "grad_norm": 0.31953608989715576, "learning_rate": 2.975304343698483e-05, "loss": 1.0670995712280273, "step": 688 }, { "epoch": 1.8852459016393444, "grad_norm": 0.3640364110469818, "learning_rate": 2.980012780063404e-05, "loss": 0.8072540760040283, "step": 690 }, { "epoch": 1.890710382513661, "grad_norm": 0.4549046754837036, "learning_rate": 2.9842269814919755e-05, "loss": 1.0721323490142822, "step": 692 }, { "epoch": 1.8961748633879782, "grad_norm": 0.9052984714508057, "learning_rate": 2.9879453819800156e-05, "loss": 0.8395594358444214, "step": 694 }, { "epoch": 1.901639344262295, "grad_norm": 0.32689180970191956, "learning_rate": 2.991166599763788e-05, "loss": 1.0846257209777832, "step": 696 }, { "epoch": 1.9071038251366121, "grad_norm": 0.43174540996551514, "learning_rate": 2.993889437833466e-05, "loss": 1.4478161334991455, "step": 698 }, { "epoch": 1.9125683060109289, "grad_norm": 0.30801597237586975, "learning_rate": 2.9961128843779457e-05, "loss": 1.0946073532104492, "step": 700 }, { "epoch": 1.918032786885246, "grad_norm": 0.29532310366630554, "learning_rate": 2.9978361131608348e-05, "loss": 1.0470151901245117, "step": 702 }, { "epoch": 1.9234972677595628, "grad_norm": 0.6861543655395508, "learning_rate": 2.999058483827483e-05, "loss": 1.1541780233383179, "step": 704 }, { "epoch": 1.92896174863388, "grad_norm": 0.373347669839859, "learning_rate": 2.9997795421429404e-05, "loss": 1.1399716138839722, "step": 706 }, { "epoch": 1.9344262295081966, "grad_norm": 0.31508007645606995, "learning_rate": 2.9999990201607516e-05, "loss": 1.1029908657073975, "step": 708 }, { "epoch": 1.9398907103825138, "grad_norm": 1.0937162637710571, "learning_rate": 2.999716836322524e-05, "loss": 1.2128394842147827, "step": 710 }, { "epoch": 1.9453551912568305, "grad_norm": 0.24902676045894623, "learning_rate": 2.9989330954882366e-05, "loss": 1.050254464149475, "step": 712 }, { "epoch": 1.9508196721311475, "grad_norm": 0.8138014078140259, "learning_rate": 2.9976480888972708e-05, "loss": 0.9359830021858215, "step": 714 }, { "epoch": 1.9562841530054644, "grad_norm": 0.4428923726081848, "learning_rate": 2.9958622940601907e-05, "loss": 1.1571717262268066, "step": 716 }, { "epoch": 1.9617486338797814, "grad_norm": 0.4176506996154785, "learning_rate": 2.9935763745812935e-05, "loss": 0.7543359398841858, "step": 718 }, { "epoch": 1.9672131147540983, "grad_norm": 0.32763925194740295, "learning_rate": 2.990791179912017e-05, "loss": 0.7445685863494873, "step": 720 }, { "epoch": 1.9726775956284153, "grad_norm": 0.3950430154800415, "learning_rate": 2.9875077450352817e-05, "loss": 1.0080370903015137, "step": 722 }, { "epoch": 1.9781420765027322, "grad_norm": 0.41049596667289734, "learning_rate": 2.9837272900808863e-05, "loss": 1.0963528156280518, "step": 724 }, { "epoch": 1.9836065573770492, "grad_norm": 0.9898220300674438, "learning_rate": 2.9794512198721092e-05, "loss": 0.7285370230674744, "step": 726 }, { "epoch": 1.989071038251366, "grad_norm": 0.3645866811275482, "learning_rate": 2.9746811234036736e-05, "loss": 0.8588154315948486, "step": 728 }, { "epoch": 1.994535519125683, "grad_norm": 0.23531635105609894, "learning_rate": 2.9694187732512702e-05, "loss": 1.1264981031417847, "step": 730 }, { "epoch": 2.0, "grad_norm": 0.6136480569839478, "learning_rate": 2.96366612491287e-05, "loss": 1.1605597734451294, "step": 732 }, { "epoch": 2.0054644808743167, "grad_norm": 0.3638727068901062, "learning_rate": 2.9574253160820573e-05, "loss": 0.5738942623138428, "step": 734 }, { "epoch": 2.010928961748634, "grad_norm": 0.38032037019729614, "learning_rate": 2.9506986658536562e-05, "loss": 0.8583514094352722, "step": 736 }, { "epoch": 2.0163934426229506, "grad_norm": 0.3164927363395691, "learning_rate": 2.9434886738619537e-05, "loss": 0.914330005645752, "step": 738 }, { "epoch": 2.021857923497268, "grad_norm": 0.2615019977092743, "learning_rate": 2.9357980193518312e-05, "loss": 0.9653261303901672, "step": 740 }, { "epoch": 2.0273224043715845, "grad_norm": 0.35523325204849243, "learning_rate": 2.927629560183153e-05, "loss": 0.9499996304512024, "step": 742 }, { "epoch": 2.0327868852459017, "grad_norm": 0.9985663294792175, "learning_rate": 2.91898633176878e-05, "loss": 1.293304681777954, "step": 744 }, { "epoch": 2.0382513661202184, "grad_norm": 0.4231744706630707, "learning_rate": 2.909871545946603e-05, "loss": 1.3876979351043701, "step": 746 }, { "epoch": 2.0437158469945356, "grad_norm": 0.36017516255378723, "learning_rate": 2.9002885897860252e-05, "loss": 1.1023402214050293, "step": 748 }, { "epoch": 2.0491803278688523, "grad_norm": 0.36883434653282166, "learning_rate": 2.8902410243293152e-05, "loss": 0.982643187046051, "step": 750 }, { "epoch": 2.0546448087431695, "grad_norm": 1.1026638746261597, "learning_rate": 2.8797325832683208e-05, "loss": 0.9915405511856079, "step": 752 }, { "epoch": 2.060109289617486, "grad_norm": 0.7558255195617676, "learning_rate": 2.868767171557021e-05, "loss": 0.5983301997184753, "step": 754 }, { "epoch": 2.0655737704918034, "grad_norm": 0.2940157353878021, "learning_rate": 2.8573488639604418e-05, "loss": 0.18914969265460968, "step": 756 }, { "epoch": 2.07103825136612, "grad_norm": 0.40500950813293457, "learning_rate": 2.845481903540464e-05, "loss": 1.0764596462249756, "step": 758 }, { "epoch": 2.0765027322404372, "grad_norm": 0.2900254428386688, "learning_rate": 2.8331707000790954e-05, "loss": 0.6835562586784363, "step": 760 }, { "epoch": 2.081967213114754, "grad_norm": 0.3581569492816925, "learning_rate": 2.820419828439788e-05, "loss": 1.024125576019287, "step": 762 }, { "epoch": 2.087431693989071, "grad_norm": 0.6350530982017517, "learning_rate": 2.8072340268674133e-05, "loss": 0.8588637709617615, "step": 764 }, { "epoch": 2.092896174863388, "grad_norm": 0.29905757308006287, "learning_rate": 2.793618195227521e-05, "loss": 1.1255227327346802, "step": 766 }, { "epoch": 2.098360655737705, "grad_norm": 0.6376034021377563, "learning_rate": 2.779577393185539e-05, "loss": 0.96219402551651, "step": 768 }, { "epoch": 2.1038251366120218, "grad_norm": 0.4236537516117096, "learning_rate": 2.765116838326597e-05, "loss": 1.024062991142273, "step": 770 }, { "epoch": 2.109289617486339, "grad_norm": 0.41378986835479736, "learning_rate": 2.750241904216663e-05, "loss": 0.9890310168266296, "step": 772 }, { "epoch": 2.1147540983606556, "grad_norm": 0.2886035442352295, "learning_rate": 2.7349581184057144e-05, "loss": 0.8476873636245728, "step": 774 }, { "epoch": 2.120218579234973, "grad_norm": 0.9165263772010803, "learning_rate": 2.719271160373693e-05, "loss": 0.6358259320259094, "step": 776 }, { "epoch": 2.1256830601092895, "grad_norm": 0.30565494298934937, "learning_rate": 2.703186859420002e-05, "loss": 0.9424027800559998, "step": 778 }, { "epoch": 2.1311475409836067, "grad_norm": 1.7092857360839844, "learning_rate": 2.6867111924973283e-05, "loss": 0.5238833427429199, "step": 780 }, { "epoch": 2.1366120218579234, "grad_norm": 0.6912135481834412, "learning_rate": 2.6698502819905935e-05, "loss": 0.6626419425010681, "step": 782 }, { "epoch": 2.1420765027322406, "grad_norm": 1.169651746749878, "learning_rate": 2.652610393441872e-05, "loss": 0.84641432762146, "step": 784 }, { "epoch": 2.1475409836065573, "grad_norm": 0.7925103902816772, "learning_rate": 2.6349979332220992e-05, "loss": 0.6377699971199036, "step": 786 }, { "epoch": 2.1530054644808745, "grad_norm": 14.70950984954834, "learning_rate": 2.6170194461504586e-05, "loss": 0.6009026169776917, "step": 788 }, { "epoch": 2.158469945355191, "grad_norm": 0.3301484286785126, "learning_rate": 2.5986816130623133e-05, "loss": 1.100122094154358, "step": 790 }, { "epoch": 2.1639344262295084, "grad_norm": 0.37357431650161743, "learning_rate": 2.579991248326594e-05, "loss": 0.920696496963501, "step": 792 }, { "epoch": 2.169398907103825, "grad_norm": 0.392766535282135, "learning_rate": 2.560955297313575e-05, "loss": 0.8580578565597534, "step": 794 }, { "epoch": 2.1748633879781423, "grad_norm": 0.4077216386795044, "learning_rate": 2.5415808338139595e-05, "loss": 0.6196969747543335, "step": 796 }, { "epoch": 2.180327868852459, "grad_norm": 0.40729326009750366, "learning_rate": 2.5218750574102465e-05, "loss": 0.6308495402336121, "step": 798 }, { "epoch": 2.185792349726776, "grad_norm": 0.2669661343097687, "learning_rate": 2.5018452908013522e-05, "loss": 0.9696865081787109, "step": 800 }, { "epoch": 2.191256830601093, "grad_norm": 0.49087321758270264, "learning_rate": 2.48149897708149e-05, "loss": 1.0563379526138306, "step": 802 }, { "epoch": 2.19672131147541, "grad_norm": 0.3579554855823517, "learning_rate": 2.4608436769743e-05, "loss": 1.1717143058776855, "step": 804 }, { "epoch": 2.202185792349727, "grad_norm": 0.2751805782318115, "learning_rate": 2.4398870660232684e-05, "loss": 0.939961850643158, "step": 806 }, { "epoch": 2.2076502732240435, "grad_norm": 1.9989415407180786, "learning_rate": 2.418636931739491e-05, "loss": 0.8161137700080872, "step": 808 }, { "epoch": 2.2131147540983607, "grad_norm": 0.5127248764038086, "learning_rate": 2.3971011707078125e-05, "loss": 1.0133951902389526, "step": 810 }, { "epoch": 2.2185792349726774, "grad_norm": 1.0294671058654785, "learning_rate": 2.3752877856524532e-05, "loss": 0.9799056649208069, "step": 812 }, { "epoch": 2.2240437158469946, "grad_norm": 0.8723068833351135, "learning_rate": 2.353204882463168e-05, "loss": 0.5277108550071716, "step": 814 }, { "epoch": 2.2295081967213113, "grad_norm": 0.42140084505081177, "learning_rate": 2.330860667183101e-05, "loss": 0.7577545642852783, "step": 816 }, { "epoch": 2.2349726775956285, "grad_norm": 1.2116085290908813, "learning_rate": 2.308263442959396e-05, "loss": 0.7222480773925781, "step": 818 }, { "epoch": 2.240437158469945, "grad_norm": 0.4509221911430359, "learning_rate": 2.2854216069577376e-05, "loss": 1.012153148651123, "step": 820 }, { "epoch": 2.2459016393442623, "grad_norm": 0.5047132968902588, "learning_rate": 2.2623436472419476e-05, "loss": 1.0092438459396362, "step": 822 }, { "epoch": 2.251366120218579, "grad_norm": 0.4681483507156372, "learning_rate": 2.2390381396198102e-05, "loss": 0.8176283836364746, "step": 824 }, { "epoch": 2.2568306010928962, "grad_norm": 0.9395934343338013, "learning_rate": 2.2155137444562842e-05, "loss": 0.8658889532089233, "step": 826 }, { "epoch": 2.262295081967213, "grad_norm": 0.8730800747871399, "learning_rate": 2.191779203455302e-05, "loss": 0.43443816900253296, "step": 828 }, { "epoch": 2.26775956284153, "grad_norm": 0.3437061607837677, "learning_rate": 2.1678433364113297e-05, "loss": 0.7622541189193726, "step": 830 }, { "epoch": 2.273224043715847, "grad_norm": 0.4882737696170807, "learning_rate": 2.1437150379319245e-05, "loss": 0.7550147771835327, "step": 832 }, { "epoch": 2.278688524590164, "grad_norm": 0.5824230313301086, "learning_rate": 2.1194032741324823e-05, "loss": 0.7233867645263672, "step": 834 }, { "epoch": 2.2841530054644807, "grad_norm": 0.455255925655365, "learning_rate": 2.0949170793044142e-05, "loss": 1.0538800954818726, "step": 836 }, { "epoch": 2.289617486338798, "grad_norm": 0.33487194776535034, "learning_rate": 2.070265552557985e-05, "loss": 0.9521304368972778, "step": 838 }, { "epoch": 2.2950819672131146, "grad_norm": 0.5424416661262512, "learning_rate": 2.0454578544410758e-05, "loss": 0.5176016092300415, "step": 840 }, { "epoch": 2.300546448087432, "grad_norm": 0.44442370533943176, "learning_rate": 2.0205032035351043e-05, "loss": 0.6318649649620056, "step": 842 }, { "epoch": 2.3060109289617485, "grad_norm": 0.17473050951957703, "learning_rate": 1.9954108730293875e-05, "loss": 0.5930679440498352, "step": 844 }, { "epoch": 2.3114754098360657, "grad_norm": 0.30008724331855774, "learning_rate": 1.9701901872752047e-05, "loss": 1.1380505561828613, "step": 846 }, { "epoch": 2.3169398907103824, "grad_norm": 0.3241089880466461, "learning_rate": 1.9448505183208607e-05, "loss": 1.0619306564331055, "step": 848 }, { "epoch": 2.3224043715846996, "grad_norm": 0.40371695160865784, "learning_rate": 1.919401282429013e-05, "loss": 0.9061838388442993, "step": 850 }, { "epoch": 2.3278688524590163, "grad_norm": 0.2833711504936218, "learning_rate": 1.893851936577567e-05, "loss": 0.9225311279296875, "step": 852 }, { "epoch": 2.3333333333333335, "grad_norm": 0.3210262060165405, "learning_rate": 1.868211974945461e-05, "loss": 0.7068516612052917, "step": 854 }, { "epoch": 2.33879781420765, "grad_norm": 0.2600228190422058, "learning_rate": 1.842490925384604e-05, "loss": 0.6802095174789429, "step": 856 }, { "epoch": 2.3442622950819674, "grad_norm": 0.3579116463661194, "learning_rate": 1.816698345879313e-05, "loss": 0.8223679065704346, "step": 858 }, { "epoch": 2.349726775956284, "grad_norm": 0.2626917064189911, "learning_rate": 1.790843820994548e-05, "loss": 0.6397656202316284, "step": 860 }, { "epoch": 2.3551912568306013, "grad_norm": 0.8502326011657715, "learning_rate": 1.7649369583142763e-05, "loss": 0.9318816065788269, "step": 862 }, { "epoch": 2.360655737704918, "grad_norm": 0.34465688467025757, "learning_rate": 1.738987384871274e-05, "loss": 0.8776544332504272, "step": 864 }, { "epoch": 2.366120218579235, "grad_norm": 0.35162287950515747, "learning_rate": 1.7130047435697118e-05, "loss": 0.6873989105224609, "step": 866 }, { "epoch": 2.371584699453552, "grad_norm": 2.411452054977417, "learning_rate": 1.6869986896018226e-05, "loss": 1.1587547063827515, "step": 868 }, { "epoch": 2.3770491803278686, "grad_norm": 0.40320128202438354, "learning_rate": 1.66097888686003e-05, "loss": 0.7416959404945374, "step": 870 }, { "epoch": 2.3825136612021858, "grad_norm": 0.7894490361213684, "learning_rate": 1.6349550043458252e-05, "loss": 0.5259535312652588, "step": 872 }, { "epoch": 2.387978142076503, "grad_norm": 0.5915789604187012, "learning_rate": 1.608936712576749e-05, "loss": 1.008756160736084, "step": 874 }, { "epoch": 2.3934426229508197, "grad_norm": 0.28783875703811646, "learning_rate": 1.582933679992809e-05, "loss": 0.7991337180137634, "step": 876 }, { "epoch": 2.3989071038251364, "grad_norm": 2.9894583225250244, "learning_rate": 1.556955569363678e-05, "loss": 0.49375149607658386, "step": 878 }, { "epoch": 2.4043715846994536, "grad_norm": 0.4313114285469055, "learning_rate": 1.531012034197988e-05, "loss": 1.1030757427215576, "step": 880 }, { "epoch": 2.4098360655737707, "grad_norm": 0.4268873631954193, "learning_rate": 1.5051127151560745e-05, "loss": 0.7422546148300171, "step": 882 }, { "epoch": 2.4153005464480874, "grad_norm": 0.4147166609764099, "learning_rate": 1.4792672364674816e-05, "loss": 0.7103762030601501, "step": 884 }, { "epoch": 2.420765027322404, "grad_norm": 0.7336945533752441, "learning_rate": 1.4534852023545968e-05, "loss": 0.8475983142852783, "step": 886 }, { "epoch": 2.4262295081967213, "grad_norm": 0.6543511748313904, "learning_rate": 1.4277761934636963e-05, "loss": 0.968014121055603, "step": 888 }, { "epoch": 2.431693989071038, "grad_norm": 0.49377578496932983, "learning_rate": 1.4021497633047664e-05, "loss": 0.6535788774490356, "step": 890 }, { "epoch": 2.4371584699453552, "grad_norm": 0.8991974592208862, "learning_rate": 1.3766154347013933e-05, "loss": 1.167250156402588, "step": 892 }, { "epoch": 2.442622950819672, "grad_norm": 0.5998942255973816, "learning_rate": 1.3511826962520809e-05, "loss": 0.7989946603775024, "step": 894 }, { "epoch": 2.448087431693989, "grad_norm": 0.25032103061676025, "learning_rate": 1.3258609988042627e-05, "loss": 0.4925026595592499, "step": 896 }, { "epoch": 2.453551912568306, "grad_norm": 0.3441256582736969, "learning_rate": 1.300659751942353e-05, "loss": 0.22913937270641327, "step": 898 }, { "epoch": 2.459016393442623, "grad_norm": 0.4925740659236908, "learning_rate": 1.2755883204911305e-05, "loss": 0.9754618406295776, "step": 900 }, { "epoch": 2.4644808743169397, "grad_norm": 0.3265496492385864, "learning_rate": 1.2506560210357541e-05, "loss": 0.47504350543022156, "step": 902 }, { "epoch": 2.469945355191257, "grad_norm": 0.30063438415527344, "learning_rate": 1.225872118459706e-05, "loss": 0.994674026966095, "step": 904 }, { "epoch": 2.4754098360655736, "grad_norm": 3.1096532344818115, "learning_rate": 1.2012458225019375e-05, "loss": 0.41736599802970886, "step": 906 }, { "epoch": 2.480874316939891, "grad_norm": 2.740457773208618, "learning_rate": 1.176786284334528e-05, "loss": 0.6344496011734009, "step": 908 }, { "epoch": 2.4863387978142075, "grad_norm": 0.3145305812358856, "learning_rate": 1.1525025931620855e-05, "loss": 0.9095264077186584, "step": 910 }, { "epoch": 2.4918032786885247, "grad_norm": 1.253567099571228, "learning_rate": 1.1284037728441877e-05, "loss": 0.5218656659126282, "step": 912 }, { "epoch": 2.4972677595628414, "grad_norm": 0.4414713382720947, "learning_rate": 1.1044987785420924e-05, "loss": 0.9621478319168091, "step": 914 }, { "epoch": 2.5027322404371586, "grad_norm": 0.2708357572555542, "learning_rate": 1.0807964933909975e-05, "loss": 1.0340259075164795, "step": 916 }, { "epoch": 2.5081967213114753, "grad_norm": 0.5265946388244629, "learning_rate": 1.0573057251990443e-05, "loss": 0.944985568523407, "step": 918 }, { "epoch": 2.5136612021857925, "grad_norm": 0.2796134948730469, "learning_rate": 1.0340352031743256e-05, "loss": 0.7973819375038147, "step": 920 }, { "epoch": 2.519125683060109, "grad_norm": 0.297129362821579, "learning_rate": 1.010993574681095e-05, "loss": 0.8349511027336121, "step": 922 }, { "epoch": 2.5245901639344264, "grad_norm": 0.8182933330535889, "learning_rate": 9.881894020263938e-06, "loss": 0.3968673050403595, "step": 924 }, { "epoch": 2.530054644808743, "grad_norm": 0.2900335192680359, "learning_rate": 9.656311592782831e-06, "loss": 0.9522480368614197, "step": 926 }, { "epoch": 2.5355191256830603, "grad_norm": 0.7127178311347961, "learning_rate": 9.433272291168689e-06, "loss": 0.9828237295150757, "step": 928 }, { "epoch": 2.540983606557377, "grad_norm": 0.6727446913719177, "learning_rate": 9.212858997192744e-06, "loss": 0.7021880149841309, "step": 930 }, { "epoch": 2.546448087431694, "grad_norm": 0.575061559677124, "learning_rate": 8.995153616797544e-06, "loss": 0.53633713722229, "step": 932 }, { "epoch": 2.551912568306011, "grad_norm": 0.6114161014556885, "learning_rate": 8.78023704966047e-06, "loss": 0.8347498178482056, "step": 934 }, { "epoch": 2.557377049180328, "grad_norm": 0.3482673168182373, "learning_rate": 8.568189159131336e-06, "loss": 0.8326533436775208, "step": 936 }, { "epoch": 2.5628415300546448, "grad_norm": 0.4061831533908844, "learning_rate": 8.359088742554941e-06, "loss": 0.9324700236320496, "step": 938 }, { "epoch": 2.5683060109289615, "grad_norm": 0.2786436676979065, "learning_rate": 8.15301350198999e-06, "loss": 0.8789687752723694, "step": 940 }, { "epoch": 2.5737704918032787, "grad_norm": 0.29912883043289185, "learning_rate": 7.950040015334789e-06, "loss": 0.8741180896759033, "step": 942 }, { "epoch": 2.579234972677596, "grad_norm": 0.23834018409252167, "learning_rate": 7.750243707870748e-06, "loss": 1.01088285446167, "step": 944 }, { "epoch": 2.5846994535519126, "grad_norm": 0.7520560026168823, "learning_rate": 7.553698824234314e-06, "loss": 0.9131101965904236, "step": 946 }, { "epoch": 2.5901639344262293, "grad_norm": 0.3973008096218109, "learning_rate": 7.360478400827475e-06, "loss": 0.6671414971351624, "step": 948 }, { "epoch": 2.5956284153005464, "grad_norm": 1.165856957435608, "learning_rate": 7.170654238677331e-06, "loss": 0.8358378410339355, "step": 950 }, { "epoch": 2.6010928961748636, "grad_norm": 0.6425595879554749, "learning_rate": 6.984296876754711e-06, "loss": 0.876006007194519, "step": 952 }, { "epoch": 2.6065573770491803, "grad_norm": 0.2834312319755554, "learning_rate": 6.801475565761783e-06, "loss": 0.8852624297142029, "step": 954 }, { "epoch": 2.612021857923497, "grad_norm": 0.26827678084373474, "learning_rate": 6.622258242398371e-06, "loss": 0.49164018034935, "step": 956 }, { "epoch": 2.6174863387978142, "grad_norm": 0.3386532962322235, "learning_rate": 6.4467115041165855e-06, "loss": 0.7272903919219971, "step": 958 }, { "epoch": 2.6229508196721314, "grad_norm": 0.2805633544921875, "learning_rate": 6.2749005843730336e-06, "loss": 0.7272099852561951, "step": 960 }, { "epoch": 2.628415300546448, "grad_norm": 0.27282318472862244, "learning_rate": 6.106889328388064e-06, "loss": 1.0292612314224243, "step": 962 }, { "epoch": 2.633879781420765, "grad_norm": 0.302123486995697, "learning_rate": 5.942740169420701e-06, "loss": 0.9444547891616821, "step": 964 }, { "epoch": 2.639344262295082, "grad_norm": 0.37525907158851624, "learning_rate": 5.7825141055683895e-06, "loss": 0.7231637239456177, "step": 966 }, { "epoch": 2.644808743169399, "grad_norm": 0.26892736554145813, "learning_rate": 5.62627067709992e-06, "loss": 0.9579343795776367, "step": 968 }, { "epoch": 2.650273224043716, "grad_norm": 0.3834379017353058, "learning_rate": 5.474067944330285e-06, "loss": 1.0427404642105103, "step": 970 }, { "epoch": 2.6557377049180326, "grad_norm": 0.576501727104187, "learning_rate": 5.325962466045282e-06, "loss": 1.0664690732955933, "step": 972 }, { "epoch": 2.66120218579235, "grad_norm": 0.32818320393562317, "learning_rate": 5.18200927848421e-06, "loss": 0.9675801992416382, "step": 974 }, { "epoch": 2.6666666666666665, "grad_norm": 0.4426153004169464, "learning_rate": 5.042261874888308e-06, "loss": 0.8803547024726868, "step": 976 }, { "epoch": 2.6721311475409837, "grad_norm": 0.6687554717063904, "learning_rate": 4.906772185622572e-06, "loss": 0.4762001633644104, "step": 978 }, { "epoch": 2.6775956284153004, "grad_norm": 0.25917065143585205, "learning_rate": 4.775590558878368e-06, "loss": 0.8584491014480591, "step": 980 }, { "epoch": 2.6830601092896176, "grad_norm": 0.8609603643417358, "learning_rate": 4.648765741963903e-06, "loss": 1.0697994232177734, "step": 982 }, { "epoch": 2.6885245901639343, "grad_norm": 0.26177000999450684, "learning_rate": 4.526344863189724e-06, "loss": 1.0349247455596924, "step": 984 }, { "epoch": 2.6939890710382515, "grad_norm": 0.6148212552070618, "learning_rate": 4.408373414355714e-06, "loss": 1.1050891876220703, "step": 986 }, { "epoch": 2.699453551912568, "grad_norm": 0.3300875425338745, "learning_rate": 4.29489523384628e-06, "loss": 0.9843365550041199, "step": 988 }, { "epoch": 2.7049180327868854, "grad_norm": 0.3000809848308563, "learning_rate": 4.185952490339899e-06, "loss": 0.8345898985862732, "step": 990 }, { "epoch": 2.710382513661202, "grad_norm": 0.45524969696998596, "learning_rate": 4.081585667139231e-06, "loss": 0.8222418427467346, "step": 992 }, { "epoch": 2.7158469945355193, "grad_norm": 0.5290429592132568, "learning_rate": 3.981833547127413e-06, "loss": 0.8905819058418274, "step": 994 }, { "epoch": 2.721311475409836, "grad_norm": 0.9160144925117493, "learning_rate": 3.886733198356298e-06, "loss": 0.8706901669502258, "step": 996 }, { "epoch": 2.726775956284153, "grad_norm": 0.24170571565628052, "learning_rate": 3.7963199602718717e-06, "loss": 0.829027533531189, "step": 998 }, { "epoch": 2.73224043715847, "grad_norm": 0.28936469554901123, "learning_rate": 3.7106274305821034e-06, "loss": 0.9818518161773682, "step": 1000 }, { "epoch": 2.737704918032787, "grad_norm": 1.1057279109954834, "learning_rate": 3.6296874527719515e-06, "loss": 0.8876405954360962, "step": 1002 }, { "epoch": 2.7431693989071038, "grad_norm": 0.22955042123794556, "learning_rate": 3.553530104270281e-06, "loss": 0.5577114820480347, "step": 1004 }, { "epoch": 2.748633879781421, "grad_norm": 0.7059992551803589, "learning_rate": 3.4821836852730384e-06, "loss": 0.8629549741744995, "step": 1006 }, { "epoch": 2.7540983606557377, "grad_norm": 0.41986677050590515, "learning_rate": 3.41567470822686e-06, "loss": 1.2205630540847778, "step": 1008 }, { "epoch": 2.7595628415300544, "grad_norm": 0.3675086796283722, "learning_rate": 3.354027887976989e-06, "loss": 0.6176258325576782, "step": 1010 }, { "epoch": 2.7650273224043715, "grad_norm": 0.3218194842338562, "learning_rate": 3.297266132583221e-06, "loss": 0.9429301023483276, "step": 1012 }, { "epoch": 2.7704918032786887, "grad_norm": 6.591115951538086, "learning_rate": 3.245410534807195e-06, "loss": 0.8208173513412476, "step": 1014 }, { "epoch": 2.7759562841530054, "grad_norm": 0.31964442133903503, "learning_rate": 3.1984803642743314e-06, "loss": 0.6169829368591309, "step": 1016 }, { "epoch": 2.781420765027322, "grad_norm": 1.38068425655365, "learning_rate": 3.1564930603131777e-06, "loss": 0.4389287531375885, "step": 1018 }, { "epoch": 2.7868852459016393, "grad_norm": 0.2846753001213074, "learning_rate": 3.1194642254749395e-06, "loss": 0.9181113243103027, "step": 1020 }, { "epoch": 2.7923497267759565, "grad_norm": 0.31778064370155334, "learning_rate": 3.0874076197355317e-06, "loss": 0.9030160903930664, "step": 1022 }, { "epoch": 2.797814207650273, "grad_norm": 0.442088782787323, "learning_rate": 3.0603351553823717e-06, "loss": 0.807093620300293, "step": 1024 }, { "epoch": 2.80327868852459, "grad_norm": 0.3773954510688782, "learning_rate": 3.038256892587734e-06, "loss": 0.8687778115272522, "step": 1026 }, { "epoch": 2.808743169398907, "grad_norm": 0.3855181336402893, "learning_rate": 3.0211810356703803e-06, "loss": 1.0619217157363892, "step": 1028 }, { "epoch": 2.8142076502732243, "grad_norm": 0.2924480736255646, "learning_rate": 3.0091139300468266e-06, "loss": 0.7161194086074829, "step": 1030 }, { "epoch": 2.819672131147541, "grad_norm": 0.2923072278499603, "learning_rate": 3.0020600598733656e-06, "loss": 0.725647509098053, "step": 1032 }, { "epoch": 2.8251366120218577, "grad_norm": 0.36642518639564514, "learning_rate": 3.000022046379753e-06, "loss": 1.2020186185836792, "step": 1034 }, { "epoch": 2.830601092896175, "grad_norm": 0.7068674564361572, "learning_rate": 3.0030006468951557e-06, "loss": 1.0905146598815918, "step": 1036 }, { "epoch": 2.836065573770492, "grad_norm": 0.3369429409503937, "learning_rate": 3.0109947545667246e-06, "loss": 0.25182783603668213, "step": 1038 }, { "epoch": 2.841530054644809, "grad_norm": 0.3530783951282501, "learning_rate": 3.024001398770901e-06, "loss": 1.0270302295684814, "step": 1040 }, { "epoch": 2.8469945355191255, "grad_norm": 0.32281166315078735, "learning_rate": 3.042015746217308e-06, "loss": 0.9714375138282776, "step": 1042 }, { "epoch": 2.8524590163934427, "grad_norm": 0.5103802680969238, "learning_rate": 3.0650311027448116e-06, "loss": 0.9001659154891968, "step": 1044 }, { "epoch": 2.8579234972677594, "grad_norm": 0.39303264021873474, "learning_rate": 3.0930389158090754e-06, "loss": 1.1310542821884155, "step": 1046 }, { "epoch": 2.8633879781420766, "grad_norm": 0.30728888511657715, "learning_rate": 3.1260287776607025e-06, "loss": 0.9291237592697144, "step": 1048 }, { "epoch": 2.8688524590163933, "grad_norm": 0.9881218075752258, "learning_rate": 3.163988429212773e-06, "loss": 0.9141870141029358, "step": 1050 }, { "epoch": 2.8743169398907105, "grad_norm": 0.21143440902233124, "learning_rate": 3.206903764596349e-06, "loss": 0.49286749958992004, "step": 1052 }, { "epoch": 2.879781420765027, "grad_norm": 0.32137101888656616, "learning_rate": 3.254758836402225e-06, "loss": 1.0972161293029785, "step": 1054 }, { "epoch": 2.8852459016393444, "grad_norm": 0.33947762846946716, "learning_rate": 3.3075358616070144e-06, "loss": 0.9067559242248535, "step": 1056 }, { "epoch": 2.890710382513661, "grad_norm": 0.32341116666793823, "learning_rate": 3.365215228181358e-06, "loss": 0.7935602068901062, "step": 1058 }, { "epoch": 2.8961748633879782, "grad_norm": 0.3355255424976349, "learning_rate": 3.4277755023777795e-06, "loss": 0.9974086880683899, "step": 1060 }, { "epoch": 2.901639344262295, "grad_norm": 0.27966246008872986, "learning_rate": 3.495193436695504e-06, "loss": 0.6481755375862122, "step": 1062 }, { "epoch": 2.907103825136612, "grad_norm": 1.3332988023757935, "learning_rate": 3.567443978519267e-06, "loss": 0.4348956346511841, "step": 1064 }, { "epoch": 2.912568306010929, "grad_norm": 0.3482104539871216, "learning_rate": 3.6445002794288992e-06, "loss": 0.7822635173797607, "step": 1066 }, { "epoch": 2.918032786885246, "grad_norm": 0.3401924669742584, "learning_rate": 3.7263337051762718e-06, "loss": 0.6737417578697205, "step": 1068 }, { "epoch": 2.9234972677595628, "grad_norm": 0.2622358798980713, "learning_rate": 3.8129138463257943e-06, "loss": 0.818148136138916, "step": 1070 }, { "epoch": 2.92896174863388, "grad_norm": 0.5667299628257751, "learning_rate": 3.904208529554625e-06, "loss": 0.9563145637512207, "step": 1072 }, { "epoch": 2.9344262295081966, "grad_norm": 1.3791451454162598, "learning_rate": 4.000183829608332e-06, "loss": 1.3551911115646362, "step": 1074 }, { "epoch": 2.939890710382514, "grad_norm": 0.43662652373313904, "learning_rate": 4.100804081907595e-06, "loss": 0.38690492510795593, "step": 1076 }, { "epoch": 2.9453551912568305, "grad_norm": 0.31867608428001404, "learning_rate": 4.206031895801176e-06, "loss": 0.9536612033843994, "step": 1078 }, { "epoch": 2.9508196721311473, "grad_norm": 0.5429739356040955, "learning_rate": 4.315828168460367e-06, "loss": 0.5728800296783447, "step": 1080 }, { "epoch": 2.9562841530054644, "grad_norm": 0.21107900142669678, "learning_rate": 4.430152099409704e-06, "loss": 0.1516314148902893, "step": 1082 }, { "epoch": 2.9617486338797816, "grad_norm": 0.22951926290988922, "learning_rate": 4.548961205688424e-06, "loss": 1.0313093662261963, "step": 1084 }, { "epoch": 2.9672131147540983, "grad_norm": 0.28624290227890015, "learning_rate": 4.672211337637246e-06, "loss": 0.69117271900177, "step": 1086 }, { "epoch": 2.972677595628415, "grad_norm": 0.5194671154022217, "learning_rate": 4.7998566953044445e-06, "loss": 0.8698742389678955, "step": 1088 }, { "epoch": 2.978142076502732, "grad_norm": 0.413565993309021, "learning_rate": 4.931849845465193e-06, "loss": 1.0937373638153076, "step": 1090 }, { "epoch": 2.9836065573770494, "grad_norm": 0.9167707562446594, "learning_rate": 5.06814173924782e-06, "loss": 1.0198386907577515, "step": 1092 }, { "epoch": 2.989071038251366, "grad_norm": 0.9425981044769287, "learning_rate": 5.208681730360458e-06, "loss": 0.5489972233772278, "step": 1094 }, { "epoch": 2.994535519125683, "grad_norm": 0.38925519585609436, "learning_rate": 5.3534175939112694e-06, "loss": 0.9535996913909912, "step": 1096 }, { "epoch": 3.0, "grad_norm": 0.33070504665374756, "learning_rate": 5.50229554581536e-06, "loss": 0.64783775806427, "step": 1098 }, { "epoch": 3.0, "step": 1098, "total_flos": 4.929269661905715e+18, "train_loss": 1.00607624289008, "train_runtime": 12927.7338, "train_samples_per_second": 5.096, "train_steps_per_second": 0.085 } ], "logging_steps": 2, "max_steps": 1098, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 99999, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4.929269661905715e+18, "train_batch_size": 3, "trial_name": null, "trial_params": null }