diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -2,749 +2,3380 @@ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, - "epoch": 2.0, + "epoch": 1.0, "eval_steps": 500, - "global_step": 88, + "global_step": 418, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { - "epoch": 0.022727272727272728, - "grad_norm": 1.4205019378037633, + "epoch": 0.0023923444976076554, + "grad_norm": 35.77700764414347, "learning_rate": 0.0, - "loss": 1.389, - "num_tokens": 131406.0, + "loss": 2.5866, + "num_tokens": 223542.0, "step": 1 }, { - "epoch": 0.045454545454545456, - "grad_norm": 1.3288118522808277, - "learning_rate": 3.3333333333333333e-06, - "loss": 1.4104, - "num_tokens": 269542.0, + "epoch": 0.004784688995215311, + "grad_norm": 34.45799391693798, + "learning_rate": 7.692307692307694e-07, + "loss": 2.4086, + "num_tokens": 463760.0, "step": 2 }, { - "epoch": 0.06818181818181818, - "grad_norm": 1.2602277316575312, - "learning_rate": 6.666666666666667e-06, - "loss": 1.3523, - "num_tokens": 415096.0, + "epoch": 0.007177033492822967, + "grad_norm": 37.09122577441204, + "learning_rate": 1.5384615384615387e-06, + "loss": 2.5261, + "num_tokens": 645652.0, "step": 3 }, { - "epoch": 0.09090909090909091, - "grad_norm": 0.9984816204796995, - "learning_rate": 1e-05, - "loss": 1.3315, - "num_tokens": 572593.0, + "epoch": 0.009569377990430622, + "grad_norm": 33.51928857090818, + "learning_rate": 2.307692307692308e-06, + "loss": 2.4839, + "num_tokens": 857399.0, "step": 4 }, { - "epoch": 0.11363636363636363, - "grad_norm": 0.6736617831588735, - "learning_rate": 9.996926770643605e-06, - "loss": 1.2977, - "num_tokens": 684308.0, + "epoch": 0.011961722488038277, + "grad_norm": 28.56245226187573, + "learning_rate": 3.0769230769230774e-06, + "loss": 2.492, + "num_tokens": 1083627.0, "step": 5 }, { - "epoch": 0.13636363636363635, - "grad_norm": 0.6541575722186684, - "learning_rate": 9.987711280236047e-06, - "loss": 1.4843, - "num_tokens": 874045.0, + "epoch": 0.014354066985645933, + "grad_norm": 23.350183104787792, + "learning_rate": 3.846153846153847e-06, + "loss": 2.3562, + "num_tokens": 1301423.0, "step": 6 }, { - "epoch": 0.1590909090909091, - "grad_norm": 0.9561934769276602, - "learning_rate": 9.972366116028733e-06, - "loss": 1.1514, - "num_tokens": 993747.0, + "epoch": 0.01674641148325359, + "grad_norm": 18.9808075909167, + "learning_rate": 4.615384615384616e-06, + "loss": 2.3909, + "num_tokens": 1492075.0, "step": 7 }, { - "epoch": 0.18181818181818182, - "grad_norm": 1.102114185447154, - "learning_rate": 9.950912237670158e-06, - "loss": 1.1769, - "num_tokens": 1095667.0, + "epoch": 0.019138755980861243, + "grad_norm": 10.708395743692822, + "learning_rate": 5.384615384615385e-06, + "loss": 2.2473, + "num_tokens": 1679562.0, "step": 8 }, { - "epoch": 0.20454545454545456, - "grad_norm": 1.104249818712161, - "learning_rate": 9.92337894857756e-06, - "loss": 1.1061, - "num_tokens": 1200992.0, + "epoch": 0.0215311004784689, + "grad_norm": 9.224705958016894, + "learning_rate": 6.153846153846155e-06, + "loss": 2.1681, + "num_tokens": 1888359.0, "step": 9 }, { - "epoch": 0.22727272727272727, - "grad_norm": 0.7355560266110254, - "learning_rate": 9.889803855911966e-06, - "loss": 1.1307, - "num_tokens": 1339956.0, + "epoch": 0.023923444976076555, + "grad_norm": 8.986278630054308, + "learning_rate": 6.923076923076923e-06, + "loss": 2.1438, + "num_tokens": 2058538.0, "step": 10 }, { - "epoch": 0.25, - "grad_norm": 0.8453382987560777, - "learning_rate": 9.850232819211344e-06, - "loss": 1.1827, - "num_tokens": 1421304.0, + "epoch": 0.02631578947368421, + "grad_norm": 4.193291761632025, + "learning_rate": 7.692307692307694e-06, + "loss": 2.0833, + "num_tokens": 2214696.0, "step": 11 }, { - "epoch": 0.2727272727272727, - "grad_norm": 0.6968595067429213, - "learning_rate": 9.804719887751984e-06, - "loss": 1.4597, - "num_tokens": 1568316.0, + "epoch": 0.028708133971291867, + "grad_norm": 3.5075199043533334, + "learning_rate": 8.461538461538462e-06, + "loss": 1.9944, + "num_tokens": 2401723.0, "step": 12 }, { - "epoch": 0.29545454545454547, - "grad_norm": 0.6529986853572142, - "learning_rate": 9.753327226723689e-06, - "loss": 1.3166, - "num_tokens": 1726457.0, + "epoch": 0.03110047846889952, + "grad_norm": 3.2504195652374874, + "learning_rate": 9.230769230769232e-06, + "loss": 2.0099, + "num_tokens": 2592218.0, "step": 13 }, { - "epoch": 0.3181818181818182, - "grad_norm": 0.6943250958530904, - "learning_rate": 9.696125032319601e-06, - "loss": 1.073, - "num_tokens": 1817768.0, + "epoch": 0.03349282296650718, + "grad_norm": 2.3589570152818795, + "learning_rate": 1e-05, + "loss": 1.9879, + "num_tokens": 2820770.0, "step": 14 }, { - "epoch": 0.3409090909090909, - "grad_norm": 0.6375339001131711, - "learning_rate": 9.633191435856654e-06, - "loss": 1.2662, - "num_tokens": 1957728.0, + "epoch": 0.03588516746411483, + "grad_norm": 3.6838613043114705, + "learning_rate": 9.999864615158956e-06, + "loss": 1.816, + "num_tokens": 3018187.0, "step": 15 }, { - "epoch": 0.36363636363636365, - "grad_norm": 0.6524047505159085, - "learning_rate": 9.564612397057581e-06, - "loss": 1.0518, - "num_tokens": 2068376.0, + "epoch": 0.03827751196172249, + "grad_norm": 2.804031211195472, + "learning_rate": 9.999458468782065e-06, + "loss": 1.9285, + "num_tokens": 3262110.0, "step": 16 }, { - "epoch": 0.38636363636363635, - "grad_norm": 0.6665766656788799, - "learning_rate": 9.490481586640279e-06, - "loss": 1.1479, - "num_tokens": 2162793.0, + "epoch": 0.04066985645933014, + "grad_norm": 2.7340483571073664, + "learning_rate": 9.998781585307577e-06, + "loss": 1.7399, + "num_tokens": 3461819.0, "step": 17 }, { - "epoch": 0.4090909090909091, - "grad_norm": 0.5776477859558063, - "learning_rate": 9.410900258374878e-06, - "loss": 1.3632, - "num_tokens": 2320690.0, + "epoch": 0.0430622009569378, + "grad_norm": 2.3573648641331952, + "learning_rate": 9.997834005464281e-06, + "loss": 1.8248, + "num_tokens": 3647749.0, "step": 18 }, { - "epoch": 0.4318181818181818, - "grad_norm": 0.5757090498766585, - "learning_rate": 9.325977110783264e-06, - "loss": 1.1998, - "num_tokens": 2453535.0, + "epoch": 0.045454545454545456, + "grad_norm": 2.7936113880349964, + "learning_rate": 9.996615786269036e-06, + "loss": 1.514, + "num_tokens": 3816664.0, "step": 19 }, { - "epoch": 0.45454545454545453, - "grad_norm": 0.5705398339175791, - "learning_rate": 9.235828138669978e-06, - "loss": 0.9882, - "num_tokens": 2584746.0, + "epoch": 0.04784688995215311, + "grad_norm": 1.5815077522510241, + "learning_rate": 9.995127001023362e-06, + "loss": 1.8473, + "num_tokens": 3981662.0, "step": 20 }, { - "epoch": 0.4772727272727273, - "grad_norm": 0.5482729927103823, - "learning_rate": 9.140576474687263e-06, - "loss": 1.3854, - "num_tokens": 2740354.0, + "epoch": 0.050239234449760764, + "grad_norm": 1.262398317459698, + "learning_rate": 9.993367739309013e-06, + "loss": 1.7998, + "num_tokens": 4166705.0, "step": 21 }, { - "epoch": 0.5, - "grad_norm": 0.645678039623043, - "learning_rate": 9.040352221150676e-06, - "loss": 1.1846, - "num_tokens": 2846842.0, + "epoch": 0.05263157894736842, + "grad_norm": 1.2085839356904715, + "learning_rate": 9.991338106982598e-06, + "loss": 1.7167, + "num_tokens": 4395460.0, "step": 22 }, { - "epoch": 0.5227272727272727, - "grad_norm": 0.5564650710505396, - "learning_rate": 8.935292272334964e-06, - "loss": 1.3385, - "num_tokens": 2977135.0, + "epoch": 0.05502392344497608, + "grad_norm": 0.9331954135736191, + "learning_rate": 9.98903822616921e-06, + "loss": 1.7225, + "num_tokens": 4660660.0, "step": 23 }, { - "epoch": 0.5454545454545454, - "grad_norm": 0.6234917245637664, - "learning_rate": 8.825540127492966e-06, - "loss": 1.1546, - "num_tokens": 3079059.0, + "epoch": 0.05741626794258373, + "grad_norm": 1.0079376114307204, + "learning_rate": 9.986468235255065e-06, + "loss": 1.7793, + "num_tokens": 4889032.0, "step": 24 }, { - "epoch": 0.5681818181818182, - "grad_norm": 0.6205509724863102, - "learning_rate": 8.711245694852887e-06, - "loss": 1.167, - "num_tokens": 3189124.0, + "epoch": 0.05980861244019139, + "grad_norm": 1.0174373092140978, + "learning_rate": 9.983628288879193e-06, + "loss": 1.6658, + "num_tokens": 5068345.0, "step": 25 }, { - "epoch": 0.5909090909090909, - "grad_norm": 0.5684731381431202, - "learning_rate": 8.592565086861683e-06, - "loss": 1.0909, - "num_tokens": 3327274.0, + "epoch": 0.06220095693779904, + "grad_norm": 0.9895227235288739, + "learning_rate": 9.98051855792412e-06, + "loss": 1.7126, + "num_tokens": 5284697.0, "step": 26 }, { - "epoch": 0.6136363636363636, - "grad_norm": 0.578100316939241, - "learning_rate": 8.469660406954254e-06, - "loss": 1.2355, - "num_tokens": 3473832.0, + "epoch": 0.0645933014354067, + "grad_norm": 0.864412154650663, + "learning_rate": 9.977139229505596e-06, + "loss": 1.658, + "num_tokens": 5521778.0, "step": 27 }, { - "epoch": 0.6363636363636364, - "grad_norm": 0.5736738369445964, - "learning_rate": 8.34269952813963e-06, - "loss": 1.3216, - "num_tokens": 3598469.0, + "epoch": 0.06698564593301436, + "grad_norm": 1.0342443056226127, + "learning_rate": 9.973490506961326e-06, + "loss": 1.5243, + "num_tokens": 5707434.0, "step": 28 }, { - "epoch": 0.6590909090909091, - "grad_norm": 0.49343606801406503, - "learning_rate": 8.211855863706654e-06, - "loss": 1.0988, - "num_tokens": 3745532.0, + "epoch": 0.06937799043062201, + "grad_norm": 0.8663151664394599, + "learning_rate": 9.969572609838745e-06, + "loss": 1.6671, + "num_tokens": 5937867.0, "step": 29 }, { - "epoch": 0.6818181818181818, - "grad_norm": 0.5794369000390701, - "learning_rate": 8.077308130362274e-06, - "loss": 1.2464, - "num_tokens": 3867813.0, + "epoch": 0.07177033492822966, + "grad_norm": 0.9781488889720827, + "learning_rate": 9.965385773881795e-06, + "loss": 1.6322, + "num_tokens": 6162073.0, "step": 30 }, { - "epoch": 0.7045454545454546, - "grad_norm": 0.5118110065353235, - "learning_rate": 7.939240104126024e-06, - "loss": 1.2701, - "num_tokens": 4030577.0, + "epoch": 0.07416267942583732, + "grad_norm": 0.9101239670857304, + "learning_rate": 9.960930251016752e-06, + "loss": 1.716, + "num_tokens": 6375008.0, "step": 31 }, { - "epoch": 0.7272727272727273, - "grad_norm": 0.5236737442701642, - "learning_rate": 7.797840369314081e-06, - "loss": 1.2471, - "num_tokens": 4182476.0, + "epoch": 0.07655502392344497, + "grad_norm": 0.9203816431810994, + "learning_rate": 9.956206309337067e-06, + "loss": 1.6335, + "num_tokens": 6568479.0, "step": 32 }, { - "epoch": 0.75, - "grad_norm": 0.5843504492331676, - "learning_rate": 7.65330206095579e-06, - "loss": 1.206, - "num_tokens": 4312690.0, + "epoch": 0.07894736842105263, + "grad_norm": 0.9924366034325527, + "learning_rate": 9.951214233087223e-06, + "loss": 1.6849, + "num_tokens": 6786875.0, "step": 33 }, { - "epoch": 0.7727272727272727, - "grad_norm": 0.5557510946066606, - "learning_rate": 7.5058226009944235e-06, - "loss": 1.0838, - "num_tokens": 4419944.0, + "epoch": 0.08133971291866028, + "grad_norm": 1.0591590540309344, + "learning_rate": 9.945954322645643e-06, + "loss": 1.5556, + "num_tokens": 6983383.0, "step": 34 }, { - "epoch": 0.7954545454545454, - "grad_norm": 0.5669433873537745, - "learning_rate": 7.355603428632566e-06, - "loss": 1.3686, - "num_tokens": 4551241.0, + "epoch": 0.08373205741626795, + "grad_norm": 0.859091081777175, + "learning_rate": 9.940426894506608e-06, + "loss": 1.6373, + "num_tokens": 7188567.0, "step": 35 }, { - "epoch": 0.8181818181818182, - "grad_norm": 0.5339287228804999, - "learning_rate": 7.2028497251903975e-06, - "loss": 0.9544, - "num_tokens": 4684645.0, + "epoch": 0.0861244019138756, + "grad_norm": 0.7819453253946674, + "learning_rate": 9.934632281261221e-06, + "loss": 1.5762, + "num_tokens": 7385552.0, "step": 36 }, { - "epoch": 0.8409090909090909, - "grad_norm": 0.549267331072764, - "learning_rate": 7.047770133852677e-06, - "loss": 1.1863, - "num_tokens": 4818559.0, + "epoch": 0.08851674641148326, + "grad_norm": 0.773342660575658, + "learning_rate": 9.928570831577396e-06, + "loss": 1.5561, + "num_tokens": 7583628.0, "step": 37 }, { - "epoch": 0.8636363636363636, - "grad_norm": 0.5012800140835663, - "learning_rate": 6.890576474687264e-06, - "loss": 1.1084, - "num_tokens": 4966509.0, + "epoch": 0.09090909090909091, + "grad_norm": 0.6689075869463446, + "learning_rate": 9.922242910178862e-06, + "loss": 1.6899, + "num_tokens": 7809190.0, "step": 38 }, { - "epoch": 0.8863636363636364, - "grad_norm": 0.5506570441743898, - "learning_rate": 6.731483455324374e-06, - "loss": 1.06, - "num_tokens": 5086656.0, + "epoch": 0.09330143540669857, + "grad_norm": 0.7505436942893252, + "learning_rate": 9.915648897823232e-06, + "loss": 1.554, + "num_tokens": 7997624.0, "step": 39 }, { - "epoch": 0.9090909090909091, - "grad_norm": 0.6125453057170752, - "learning_rate": 6.570708377691784e-06, - "loss": 1.324, - "num_tokens": 5221452.0, + "epoch": 0.09569377990430622, + "grad_norm": 0.7015919399979235, + "learning_rate": 9.908789191279093e-06, + "loss": 1.5389, + "num_tokens": 8183494.0, "step": 40 }, { - "epoch": 0.9318181818181818, - "grad_norm": 0.6117528065309762, - "learning_rate": 6.408470841206547e-06, - "loss": 1.0854, - "num_tokens": 5354389.0, + "epoch": 0.09808612440191387, + "grad_norm": 0.6466036781332534, + "learning_rate": 9.901664203302126e-06, + "loss": 1.6405, + "num_tokens": 8408569.0, "step": 41 }, { - "epoch": 0.9545454545454546, - "grad_norm": 0.5331557205029857, - "learning_rate": 6.244992442828586e-06, - "loss": 1.15, - "num_tokens": 5481387.0, + "epoch": 0.10047846889952153, + "grad_norm": 0.6548488840567466, + "learning_rate": 9.89427436261027e-06, + "loss": 1.5763, + "num_tokens": 8584336.0, "step": 42 }, { - "epoch": 0.9772727272727273, - "grad_norm": 0.5633865160193015, - "learning_rate": 6.080496474385916e-06, - "loss": 1.3434, - "num_tokens": 5610810.0, + "epoch": 0.10287081339712918, + "grad_norm": 0.699125932858908, + "learning_rate": 9.886620113857926e-06, + "loss": 1.5114, + "num_tokens": 8787695.0, "step": 43 }, { - "epoch": 1.0, - "grad_norm": 0.6410174981360411, - "learning_rate": 5.915207617584859e-06, - "loss": 1.2938, - "num_tokens": 5762262.0, - "step": 44 - }, - { - "epoch": 1.0, - "eval_loss": 0.7926232814788818, - "eval_num_tokens": 5762262.0, - "eval_runtime": 9.7087, - "eval_samples_per_second": 15.965, - "eval_steps_per_second": 4.017, + "epoch": 0.10526315789473684, + "grad_norm": 0.723472446208709, + "learning_rate": 9.878701917609208e-06, + "loss": 1.5305, + "num_tokens": 9005521.0, "step": 44 }, { - "epoch": 1.0227272727272727, - "grad_norm": 0.601059843631901, - "learning_rate": 5.749351637121866e-06, - "loss": 1.1418, - "num_tokens": 5907796.0, + "epoch": 0.1076555023923445, + "grad_norm": 0.7384196667665146, + "learning_rate": 9.870520250310223e-06, + "loss": 1.6305, + "num_tokens": 9221565.0, "step": 45 }, { - "epoch": 1.0454545454545454, - "grad_norm": 0.6205636328365246, - "learning_rate": 5.5831550723160854e-06, - "loss": 1.1986, - "num_tokens": 6036486.0, + "epoch": 0.11004784688995216, + "grad_norm": 0.629523841227617, + "learning_rate": 9.862075604260402e-06, + "loss": 1.6484, + "num_tokens": 9444154.0, "step": 46 }, { - "epoch": 1.0681818181818181, - "grad_norm": 0.5344322416393745, - "learning_rate": 5.416844927683916e-06, - "loss": 1.2355, - "num_tokens": 6204626.0, + "epoch": 0.11244019138755981, + "grad_norm": 0.6425557703652314, + "learning_rate": 9.853368487582888e-06, + "loss": 1.4761, + "num_tokens": 9628428.0, "step": 47 }, { - "epoch": 1.0909090909090908, - "grad_norm": 0.5607674809394576, - "learning_rate": 5.250648362878136e-06, - "loss": 1.0154, - "num_tokens": 6339028.0, + "epoch": 0.11483253588516747, + "grad_norm": 0.6988538672174611, + "learning_rate": 9.84439942419395e-06, + "loss": 1.6356, + "num_tokens": 9862779.0, "step": 48 }, { - "epoch": 1.1136363636363635, - "grad_norm": 0.5161307847392129, - "learning_rate": 5.084792382415142e-06, - "loss": 1.0456, - "num_tokens": 6483570.0, + "epoch": 0.11722488038277512, + "grad_norm": 0.6304374061279027, + "learning_rate": 9.835168953771463e-06, + "loss": 1.5105, + "num_tokens": 10111617.0, "step": 49 }, { - "epoch": 1.1363636363636362, - "grad_norm": 0.5680285537905805, - "learning_rate": 4.919503525614086e-06, - "loss": 1.1061, - "num_tokens": 6598215.0, + "epoch": 0.11961722488038277, + "grad_norm": 0.6166276961179814, + "learning_rate": 9.825677631722436e-06, + "loss": 1.6297, + "num_tokens": 10349939.0, "step": 50 }, { - "epoch": 1.1590909090909092, - "grad_norm": 0.5309756253818366, - "learning_rate": 4.755007557171415e-06, - "loss": 1.0321, - "num_tokens": 6741934.0, + "epoch": 0.12200956937799043, + "grad_norm": 0.7275282546024043, + "learning_rate": 9.815926029149593e-06, + "loss": 1.5068, + "num_tokens": 10534354.0, "step": 51 }, { - "epoch": 1.1818181818181819, - "grad_norm": 0.5535644437015048, - "learning_rate": 4.591529158793455e-06, - "loss": 1.0263, - "num_tokens": 6877014.0, + "epoch": 0.12440191387559808, + "grad_norm": 0.6327061243856574, + "learning_rate": 9.805914732817007e-06, + "loss": 1.5524, + "num_tokens": 10742091.0, "step": 52 }, { - "epoch": 1.2045454545454546, - "grad_norm": 0.6111696305729324, - "learning_rate": 4.429291622308217e-06, - "loss": 1.0381, - "num_tokens": 6996149.0, + "epoch": 0.12679425837320574, + "grad_norm": 0.7084417481062125, + "learning_rate": 9.795644345114796e-06, + "loss": 1.6218, + "num_tokens": 10961264.0, "step": 53 }, { - "epoch": 1.2272727272727273, - "grad_norm": 0.5273029166368002, - "learning_rate": 4.268516544675628e-06, - "loss": 1.2142, - "num_tokens": 7151424.0, + "epoch": 0.1291866028708134, + "grad_norm": 0.7292180354816592, + "learning_rate": 9.78511548402287e-06, + "loss": 1.3672, + "num_tokens": 11119849.0, "step": 54 }, { - "epoch": 1.25, - "grad_norm": 0.5406640820645169, - "learning_rate": 4.109423525312738e-06, - "loss": 1.186, - "num_tokens": 7293753.0, + "epoch": 0.13157894736842105, + "grad_norm": 0.6555927818623691, + "learning_rate": 9.77432878307376e-06, + "loss": 1.5114, + "num_tokens": 11325079.0, "step": 55 }, { - "epoch": 1.2727272727272727, - "grad_norm": 0.6579245501276235, - "learning_rate": 3.952229866147323e-06, - "loss": 0.7429, - "num_tokens": 7365428.0, + "epoch": 0.1339712918660287, + "grad_norm": 0.7613236601730216, + "learning_rate": 9.763284891314481e-06, + "loss": 1.2232, + "num_tokens": 11509772.0, "step": 56 }, { - "epoch": 1.2954545454545454, - "grad_norm": 0.617679571541095, - "learning_rate": 3.7971502748096043e-06, - "loss": 1.106, - "num_tokens": 7495875.0, + "epoch": 0.13636363636363635, + "grad_norm": 0.6430242992230627, + "learning_rate": 9.751984473267498e-06, + "loss": 1.5539, + "num_tokens": 11686001.0, "step": 57 }, { - "epoch": 1.3181818181818181, - "grad_norm": 0.5340097075891522, - "learning_rate": 3.6443965713674354e-06, - "loss": 1.0039, - "num_tokens": 7641432.0, + "epoch": 0.13875598086124402, + "grad_norm": 0.8426300537945989, + "learning_rate": 9.740428208890716e-06, + "loss": 1.3227, + "num_tokens": 11861406.0, "step": 58 }, { - "epoch": 1.3409090909090908, - "grad_norm": 0.5029638027608234, - "learning_rate": 3.4941773990055784e-06, - "loss": 1.0805, - "num_tokens": 7810665.0, + "epoch": 0.14114832535885166, + "grad_norm": 0.7044359465699243, + "learning_rate": 9.728616793536588e-06, + "loss": 1.5587, + "num_tokens": 12083010.0, "step": 59 }, { - "epoch": 1.3636363636363638, - "grad_norm": 0.5711313285447551, - "learning_rate": 3.3466979390442112e-06, - "loss": 0.8987, - "num_tokens": 7928413.0, + "epoch": 0.14354066985645933, + "grad_norm": 0.6499151456833749, + "learning_rate": 9.716550937910268e-06, + "loss": 1.4789, + "num_tokens": 12331238.0, "step": 60 }, { - "epoch": 1.3863636363636362, - "grad_norm": 0.6311451158664684, - "learning_rate": 3.20215963068592e-06, - "loss": 1.0411, - "num_tokens": 8045941.0, + "epoch": 0.145933014354067, + "grad_norm": 0.8468125913528113, + "learning_rate": 9.70423136802684e-06, + "loss": 1.397, + "num_tokens": 12510110.0, "step": 61 }, { - "epoch": 1.4090909090909092, - "grad_norm": 0.5723539780836542, - "learning_rate": 3.0607598958739782e-06, - "loss": 1.159, - "num_tokens": 8161378.0, + "epoch": 0.14832535885167464, + "grad_norm": 0.6549777551235627, + "learning_rate": 9.691658825167641e-06, + "loss": 1.5487, + "num_tokens": 12703234.0, "step": 62 }, { - "epoch": 1.4318181818181819, - "grad_norm": 0.6491007813040308, - "learning_rate": 2.922691869637727e-06, - "loss": 0.9775, - "num_tokens": 8249508.0, + "epoch": 0.1507177033492823, + "grad_norm": 1.861908798533634, + "learning_rate": 9.67883406583566e-06, + "loss": 1.5182, + "num_tokens": 12941740.0, "step": 63 }, { - "epoch": 1.4545454545454546, - "grad_norm": 0.5594437573758038, - "learning_rate": 2.788144136293347e-06, - "loss": 1.047, - "num_tokens": 8401172.0, + "epoch": 0.15311004784688995, + "grad_norm": 0.9710524431103752, + "learning_rate": 9.665757861710008e-06, + "loss": 0.9673, + "num_tokens": 13066657.0, "step": 64 }, { - "epoch": 1.4772727272727273, - "grad_norm": 0.7606841416397666, - "learning_rate": 2.6573004718603725e-06, - "loss": 1.1966, - "num_tokens": 8558052.0, + "epoch": 0.15550239234449761, + "grad_norm": 0.7709985947911734, + "learning_rate": 9.652430999599491e-06, + "loss": 1.2468, + "num_tokens": 13220140.0, "step": 65 }, { - "epoch": 1.5, - "grad_norm": 0.5482072444715032, - "learning_rate": 2.5303395930457496e-06, - "loss": 1.0915, - "num_tokens": 8697623.0, + "epoch": 0.15789473684210525, + "grad_norm": 1.148184964185836, + "learning_rate": 9.638854281395271e-06, + "loss": 1.3733, + "num_tokens": 13397562.0, "step": 66 }, { - "epoch": 1.5227272727272727, - "grad_norm": 0.5685205804070853, - "learning_rate": 2.4074349131383183e-06, - "loss": 0.8498, - "num_tokens": 8803080.0, + "epoch": 0.16028708133971292, + "grad_norm": 1.8909672590403999, + "learning_rate": 9.625028524022606e-06, + "loss": 1.5125, + "num_tokens": 13591137.0, "step": 67 }, { - "epoch": 1.5454545454545454, - "grad_norm": 0.5956789138114672, - "learning_rate": 2.2887543051471154e-06, - "loss": 0.9069, - "num_tokens": 8897749.0, + "epoch": 0.16267942583732056, + "grad_norm": 0.6070509391624586, + "learning_rate": 9.610954559391704e-06, + "loss": 1.576, + "num_tokens": 13812222.0, "step": 68 }, { - "epoch": 1.5681818181818183, - "grad_norm": 0.5822527883902986, - "learning_rate": 2.174459872507035e-06, - "loss": 0.9693, - "num_tokens": 9015995.0, + "epoch": 0.16507177033492823, + "grad_norm": 0.9685185344694189, + "learning_rate": 9.596633234347661e-06, + "loss": 1.3289, + "num_tokens": 14034014.0, "step": 69 }, { - "epoch": 1.5909090909090908, - "grad_norm": 0.5638254695880359, - "learning_rate": 2.064707727665037e-06, - "loss": 0.912, - "num_tokens": 9138851.0, + "epoch": 0.1674641148325359, + "grad_norm": 1.3816831844533364, + "learning_rate": 9.582065410619503e-06, + "loss": 1.6031, + "num_tokens": 14248962.0, "step": 70 }, { - "epoch": 1.6136363636363638, - "grad_norm": 0.50620677933753, - "learning_rate": 1.9596477788493256e-06, - "loss": 1.062, - "num_tokens": 9289386.0, + "epoch": 0.16985645933014354, + "grad_norm": 1.0562478172616532, + "learning_rate": 9.567251964768343e-06, + "loss": 1.1815, + "num_tokens": 14461235.0, "step": 71 }, { - "epoch": 1.6363636363636362, - "grad_norm": 0.5330266858174569, - "learning_rate": 1.8594235253127373e-06, - "loss": 1.0187, - "num_tokens": 9435061.0, + "epoch": 0.1722488038277512, + "grad_norm": 0.8780808898792398, + "learning_rate": 9.55219378813463e-06, + "loss": 1.4514, + "num_tokens": 14655207.0, "step": 72 }, { - "epoch": 1.6590909090909092, - "grad_norm": 0.6907150564057343, - "learning_rate": 1.764171861330023e-06, - "loss": 0.7628, - "num_tokens": 9510636.0, + "epoch": 0.17464114832535885, + "grad_norm": 1.1376167558495347, + "learning_rate": 9.53689178678452e-06, + "loss": 1.2722, + "num_tokens": 14832145.0, "step": 73 }, { - "epoch": 1.6818181818181817, - "grad_norm": 0.6110441029426633, - "learning_rate": 1.674022889216737e-06, - "loss": 1.0034, - "num_tokens": 9621814.0, + "epoch": 0.17703349282296652, + "grad_norm": 0.954111645763459, + "learning_rate": 9.521346881455356e-06, + "loss": 1.3542, + "num_tokens": 15010659.0, "step": 74 }, { - "epoch": 1.7045454545454546, - "grad_norm": 0.5037611057455024, - "learning_rate": 1.5890997416251226e-06, - "loss": 1.1213, - "num_tokens": 9782217.0, + "epoch": 0.17942583732057416, + "grad_norm": 1.1453698076350756, + "learning_rate": 9.505560007500263e-06, + "loss": 1.3355, + "num_tokens": 15213072.0, "step": 75 }, { - "epoch": 1.7272727272727273, - "grad_norm": 0.5861864173490295, - "learning_rate": 1.5095184133597219e-06, - "loss": 1.0996, - "num_tokens": 9910702.0, + "epoch": 0.18181818181818182, + "grad_norm": 0.7130576585117359, + "learning_rate": 9.489532114831876e-06, + "loss": 1.4303, + "num_tokens": 15403547.0, "step": 76 }, { - "epoch": 1.75, - "grad_norm": 0.5675991498361835, - "learning_rate": 1.4353876029424205e-06, - "loss": 1.1476, - "num_tokens": 10050003.0, + "epoch": 0.18421052631578946, + "grad_norm": 1.1940739116738854, + "learning_rate": 9.473264167865172e-06, + "loss": 1.4188, + "num_tokens": 15608337.0, "step": 77 }, { - "epoch": 1.7727272727272727, - "grad_norm": 0.5483178225982944, - "learning_rate": 1.3668085641433465e-06, - "loss": 1.1357, - "num_tokens": 10215027.0, + "epoch": 0.18660287081339713, + "grad_norm": 0.8288648375057176, + "learning_rate": 9.456757145459445e-06, + "loss": 1.5327, + "num_tokens": 15844122.0, "step": 78 }, { - "epoch": 1.7954545454545454, - "grad_norm": 0.5790711829756928, - "learning_rate": 1.3038749676803994e-06, - "loss": 1.0038, - "num_tokens": 10372778.0, + "epoch": 0.18899521531100477, + "grad_norm": 0.7407524557729414, + "learning_rate": 9.44001204085941e-06, + "loss": 1.2769, + "num_tokens": 16007831.0, "step": 79 }, { - "epoch": 1.8181818181818183, - "grad_norm": 0.7101041823981646, - "learning_rate": 1.2466727732763126e-06, - "loss": 0.943, - "num_tokens": 10479665.0, + "epoch": 0.19138755980861244, + "grad_norm": 0.6820580256711418, + "learning_rate": 9.423029861635431e-06, + "loss": 1.4512, + "num_tokens": 16192803.0, "step": 80 }, { - "epoch": 1.8409090909090908, - "grad_norm": 0.5732396202107574, - "learning_rate": 1.1952801122480168e-06, - "loss": 1.1247, - "num_tokens": 10601118.0, + "epoch": 0.1937799043062201, + "grad_norm": 0.7648355853081366, + "learning_rate": 9.405811629622904e-06, + "loss": 1.5147, + "num_tokens": 16372150.0, "step": 81 }, { - "epoch": 1.8636363636363638, - "grad_norm": 0.5473267173615615, - "learning_rate": 1.1497671807886569e-06, - "loss": 0.9684, - "num_tokens": 10720107.0, + "epoch": 0.19617224880382775, + "grad_norm": 0.9430997963743641, + "learning_rate": 9.388358380860763e-06, + "loss": 1.439, + "num_tokens": 16544020.0, "step": 82 }, { - "epoch": 1.8863636363636362, - "grad_norm": 0.5873037606094315, - "learning_rate": 1.1101961440880355e-06, - "loss": 0.8847, - "num_tokens": 10834398.0, + "epoch": 0.19856459330143542, + "grad_norm": 0.6625036619665103, + "learning_rate": 9.370671165529146e-06, + "loss": 1.362, + "num_tokens": 16724075.0, "step": 83 }, { - "epoch": 1.9090909090909092, - "grad_norm": 0.6086177345996866, - "learning_rate": 1.076621051422442e-06, - "loss": 1.0159, - "num_tokens": 10969053.0, + "epoch": 0.20095693779904306, + "grad_norm": 0.6779552456686411, + "learning_rate": 9.3527510478862e-06, + "loss": 1.296, + "num_tokens": 16892945.0, "step": 84 }, { - "epoch": 1.9318181818181817, - "grad_norm": 0.5510020714453475, - "learning_rate": 1.0490877623298432e-06, - "loss": 0.9697, - "num_tokens": 11087048.0, + "epoch": 0.20334928229665072, + "grad_norm": 0.8756233237648234, + "learning_rate": 9.334599106204051e-06, + "loss": 1.4463, + "num_tokens": 17052936.0, "step": 85 }, { - "epoch": 1.9545454545454546, - "grad_norm": 0.615419171592176, - "learning_rate": 1.0276338839712689e-06, - "loss": 1.0063, - "num_tokens": 11189987.0, + "epoch": 0.20574162679425836, + "grad_norm": 0.7900920573876685, + "learning_rate": 9.316216432703918e-06, + "loss": 1.1998, + "num_tokens": 17229219.0, "step": 86 }, { - "epoch": 1.9772727272727273, - "grad_norm": 0.4935034026694359, - "learning_rate": 1.012288719763954e-06, - "loss": 1.1786, - "num_tokens": 11343043.0, + "epoch": 0.20813397129186603, + "grad_norm": 0.6469526433608052, + "learning_rate": 9.29760413349039e-06, + "loss": 1.4746, + "num_tokens": 17439405.0, "step": 87 }, { - "epoch": 2.0, - "grad_norm": 0.4878283914253533, - "learning_rate": 1.0030732293563971e-06, - "loss": 1.1914, - "num_tokens": 11521575.0, + "epoch": 0.21052631578947367, + "grad_norm": 0.616638088055178, + "learning_rate": 9.278763328484875e-06, + "loss": 1.6045, + "num_tokens": 17642943.0, "step": 88 }, { - "epoch": 2.0, - "eval_loss": 0.7495966553688049, - "eval_num_tokens": 11521575.0, - "eval_runtime": 9.1602, - "eval_samples_per_second": 16.921, - "eval_steps_per_second": 4.258, - "step": 88 + "epoch": 0.21291866028708134, + "grad_norm": 0.950240919701955, + "learning_rate": 9.259695151358215e-06, + "loss": 1.4296, + "num_tokens": 17893610.0, + "step": 89 + }, + { + "epoch": 0.215311004784689, + "grad_norm": 0.8787225924649048, + "learning_rate": 9.240400749462467e-06, + "loss": 1.3394, + "num_tokens": 18041227.0, + "step": 90 + }, + { + "epoch": 0.21770334928229665, + "grad_norm": 0.6650651036759095, + "learning_rate": 9.220881283761868e-06, + "loss": 1.4976, + "num_tokens": 18231908.0, + "step": 91 + }, + { + "epoch": 0.22009569377990432, + "grad_norm": 1.0488690262480544, + "learning_rate": 9.20113792876298e-06, + "loss": 1.544, + "num_tokens": 18449826.0, + "step": 92 + }, + { + "epoch": 0.22248803827751196, + "grad_norm": 0.8128101002861978, + "learning_rate": 9.181171872444015e-06, + "loss": 1.6064, + "num_tokens": 18678175.0, + "step": 93 + }, + { + "epoch": 0.22488038277511962, + "grad_norm": 0.8352400142358486, + "learning_rate": 9.160984316183354e-06, + "loss": 1.3891, + "num_tokens": 18856695.0, + "step": 94 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 0.6330668052004139, + "learning_rate": 9.140576474687263e-06, + "loss": 1.478, + "num_tokens": 19058871.0, + "step": 95 + }, + { + "epoch": 0.22966507177033493, + "grad_norm": 0.6984617187165848, + "learning_rate": 9.1199495759168e-06, + "loss": 1.3507, + "num_tokens": 19251656.0, + "step": 96 + }, + { + "epoch": 0.23205741626794257, + "grad_norm": 0.86370290023878, + "learning_rate": 9.099104861013922e-06, + "loss": 1.3834, + "num_tokens": 19432481.0, + "step": 97 + }, + { + "epoch": 0.23444976076555024, + "grad_norm": 0.7464764871732519, + "learning_rate": 9.078043584226816e-06, + "loss": 1.2249, + "num_tokens": 19625994.0, + "step": 98 + }, + { + "epoch": 0.23684210526315788, + "grad_norm": 0.7046998605163254, + "learning_rate": 9.056767012834417e-06, + "loss": 1.3523, + "num_tokens": 19879656.0, + "step": 99 + }, + { + "epoch": 0.23923444976076555, + "grad_norm": 0.6797355825616871, + "learning_rate": 9.035276427070166e-06, + "loss": 1.4905, + "num_tokens": 20094011.0, + "step": 100 + }, + { + "epoch": 0.24162679425837322, + "grad_norm": 0.643387767261834, + "learning_rate": 9.013573120044968e-06, + "loss": 1.4817, + "num_tokens": 20324769.0, + "step": 101 + }, + { + "epoch": 0.24401913875598086, + "grad_norm": 0.6345521357259186, + "learning_rate": 8.991658397669384e-06, + "loss": 1.406, + "num_tokens": 20523715.0, + "step": 102 + }, + { + "epoch": 0.24641148325358853, + "grad_norm": 0.5851359669759257, + "learning_rate": 8.96953357857507e-06, + "loss": 1.5156, + "num_tokens": 20744615.0, + "step": 103 + }, + { + "epoch": 0.24880382775119617, + "grad_norm": 0.7543388243657427, + "learning_rate": 8.947199994035402e-06, + "loss": 1.4965, + "num_tokens": 20954429.0, + "step": 104 + }, + { + "epoch": 0.2511961722488038, + "grad_norm": 0.6708121582654688, + "learning_rate": 8.924658987885403e-06, + "loss": 1.4257, + "num_tokens": 21152905.0, + "step": 105 + }, + { + "epoch": 0.2535885167464115, + "grad_norm": 0.8402032350426654, + "learning_rate": 8.901911916440867e-06, + "loss": 1.5563, + "num_tokens": 21368251.0, + "step": 106 + }, + { + "epoch": 0.25598086124401914, + "grad_norm": 0.7320851769974586, + "learning_rate": 8.878960148416747e-06, + "loss": 1.3888, + "num_tokens": 21508325.0, + "step": 107 + }, + { + "epoch": 0.2583732057416268, + "grad_norm": 0.7657651705836107, + "learning_rate": 8.855805064844808e-06, + "loss": 1.2907, + "num_tokens": 21657394.0, + "step": 108 + }, + { + "epoch": 0.2607655502392344, + "grad_norm": 0.7516028881611917, + "learning_rate": 8.832448058990522e-06, + "loss": 1.3294, + "num_tokens": 21898147.0, + "step": 109 + }, + { + "epoch": 0.2631578947368421, + "grad_norm": 0.7412973068093045, + "learning_rate": 8.80889053626923e-06, + "loss": 1.3493, + "num_tokens": 22114888.0, + "step": 110 + }, + { + "epoch": 0.26555023923444976, + "grad_norm": 0.715207637159085, + "learning_rate": 8.785133914161586e-06, + "loss": 1.3189, + "num_tokens": 22309919.0, + "step": 111 + }, + { + "epoch": 0.2679425837320574, + "grad_norm": 0.8490500056983592, + "learning_rate": 8.761179622128264e-06, + "loss": 1.2958, + "num_tokens": 22463886.0, + "step": 112 + }, + { + "epoch": 0.2703349282296651, + "grad_norm": 0.6593276317049697, + "learning_rate": 8.737029101523931e-06, + "loss": 1.3876, + "num_tokens": 22695323.0, + "step": 113 + }, + { + "epoch": 0.2727272727272727, + "grad_norm": 0.809173159712039, + "learning_rate": 8.712683805510547e-06, + "loss": 1.2752, + "num_tokens": 22923680.0, + "step": 114 + }, + { + "epoch": 0.2751196172248804, + "grad_norm": 0.7247603972983714, + "learning_rate": 8.6881451989699e-06, + "loss": 1.1074, + "num_tokens": 23075695.0, + "step": 115 + }, + { + "epoch": 0.27751196172248804, + "grad_norm": 0.6591979377593177, + "learning_rate": 8.66341475841548e-06, + "loss": 1.4945, + "num_tokens": 23320851.0, + "step": 116 + }, + { + "epoch": 0.2799043062200957, + "grad_norm": 1.1647325657638634, + "learning_rate": 8.638493971903621e-06, + "loss": 1.2714, + "num_tokens": 23524534.0, + "step": 117 + }, + { + "epoch": 0.2822966507177033, + "grad_norm": 0.7586726684402458, + "learning_rate": 8.613384338943982e-06, + "loss": 1.3522, + "num_tokens": 23773345.0, + "step": 118 + }, + { + "epoch": 0.284688995215311, + "grad_norm": 0.6733449030482199, + "learning_rate": 8.588087370409303e-06, + "loss": 1.3563, + "num_tokens": 23961835.0, + "step": 119 + }, + { + "epoch": 0.28708133971291866, + "grad_norm": 1.1832042021851583, + "learning_rate": 8.562604588444498e-06, + "loss": 1.368, + "num_tokens": 24186825.0, + "step": 120 + }, + { + "epoch": 0.2894736842105263, + "grad_norm": 1.1589708241685623, + "learning_rate": 8.536937526375075e-06, + "loss": 1.4034, + "num_tokens": 24422495.0, + "step": 121 + }, + { + "epoch": 0.291866028708134, + "grad_norm": 0.7701824945643486, + "learning_rate": 8.511087728614863e-06, + "loss": 1.4039, + "num_tokens": 24638181.0, + "step": 122 + }, + { + "epoch": 0.2942583732057416, + "grad_norm": 0.6500381479650347, + "learning_rate": 8.485056750573088e-06, + "loss": 1.4504, + "num_tokens": 24875758.0, + "step": 123 + }, + { + "epoch": 0.2966507177033493, + "grad_norm": 0.6319890651859407, + "learning_rate": 8.458846158560787e-06, + "loss": 1.4881, + "num_tokens": 25089829.0, + "step": 124 + }, + { + "epoch": 0.29904306220095694, + "grad_norm": 1.388998615476455, + "learning_rate": 8.43245752969655e-06, + "loss": 1.1747, + "num_tokens": 25311276.0, + "step": 125 + }, + { + "epoch": 0.3014354066985646, + "grad_norm": 1.0789606093954458, + "learning_rate": 8.40589245181163e-06, + "loss": 1.1788, + "num_tokens": 25489153.0, + "step": 126 + }, + { + "epoch": 0.3038277511961722, + "grad_norm": 0.7049145346829824, + "learning_rate": 8.379152523354407e-06, + "loss": 1.3323, + "num_tokens": 25669945.0, + "step": 127 + }, + { + "epoch": 0.3062200956937799, + "grad_norm": 0.6756413517215413, + "learning_rate": 8.352239353294196e-06, + "loss": 1.4037, + "num_tokens": 25909865.0, + "step": 128 + }, + { + "epoch": 0.30861244019138756, + "grad_norm": 0.8708376581094076, + "learning_rate": 8.325154561024445e-06, + "loss": 1.3679, + "num_tokens": 26127238.0, + "step": 129 + }, + { + "epoch": 0.31100478468899523, + "grad_norm": 1.3947730736376298, + "learning_rate": 8.29789977626528e-06, + "loss": 1.368, + "num_tokens": 26410830.0, + "step": 130 + }, + { + "epoch": 0.3133971291866029, + "grad_norm": 0.6066250801672678, + "learning_rate": 8.270476638965463e-06, + "loss": 1.4124, + "num_tokens": 26580685.0, + "step": 131 + }, + { + "epoch": 0.3157894736842105, + "grad_norm": 0.6459251127241467, + "learning_rate": 8.242886799203696e-06, + "loss": 1.4815, + "num_tokens": 26810854.0, + "step": 132 + }, + { + "epoch": 0.3181818181818182, + "grad_norm": 0.8523013271806112, + "learning_rate": 8.215131917089342e-06, + "loss": 1.3218, + "num_tokens": 27020107.0, + "step": 133 + }, + { + "epoch": 0.32057416267942584, + "grad_norm": 0.80764930998251, + "learning_rate": 8.187213662662539e-06, + "loss": 1.4524, + "num_tokens": 27225900.0, + "step": 134 + }, + { + "epoch": 0.3229665071770335, + "grad_norm": 0.8176074504442626, + "learning_rate": 8.159133715793701e-06, + "loss": 1.1947, + "num_tokens": 27437322.0, + "step": 135 + }, + { + "epoch": 0.3253588516746411, + "grad_norm": 0.6612632769182687, + "learning_rate": 8.13089376608245e-06, + "loss": 1.4414, + "num_tokens": 27620222.0, + "step": 136 + }, + { + "epoch": 0.3277511961722488, + "grad_norm": 0.7997382570532242, + "learning_rate": 8.102495512755939e-06, + "loss": 1.1692, + "num_tokens": 27779969.0, + "step": 137 + }, + { + "epoch": 0.33014354066985646, + "grad_norm": 0.7611481255340015, + "learning_rate": 8.073940664566623e-06, + "loss": 1.2419, + "num_tokens": 27975654.0, + "step": 138 + }, + { + "epoch": 0.33253588516746413, + "grad_norm": 1.2608591606765718, + "learning_rate": 8.045230939689425e-06, + "loss": 1.389, + "num_tokens": 28167517.0, + "step": 139 + }, + { + "epoch": 0.3349282296650718, + "grad_norm": 0.5773709072619498, + "learning_rate": 8.016368065618361e-06, + "loss": 1.4468, + "num_tokens": 28484552.0, + "step": 140 + }, + { + "epoch": 0.3373205741626794, + "grad_norm": 0.725628062383284, + "learning_rate": 7.987353779062598e-06, + "loss": 1.1387, + "num_tokens": 28658376.0, + "step": 141 + }, + { + "epoch": 0.3397129186602871, + "grad_norm": 1.0360991727200277, + "learning_rate": 7.958189825841942e-06, + "loss": 1.0863, + "num_tokens": 28857873.0, + "step": 142 + }, + { + "epoch": 0.34210526315789475, + "grad_norm": 0.8786402024406298, + "learning_rate": 7.928877960781808e-06, + "loss": 1.2436, + "num_tokens": 29065902.0, + "step": 143 + }, + { + "epoch": 0.3444976076555024, + "grad_norm": 0.6828988509231846, + "learning_rate": 7.899419947607611e-06, + "loss": 1.4069, + "num_tokens": 29258786.0, + "step": 144 + }, + { + "epoch": 0.34688995215311, + "grad_norm": 0.7208210642568142, + "learning_rate": 7.869817558838654e-06, + "loss": 1.0879, + "num_tokens": 29468217.0, + "step": 145 + }, + { + "epoch": 0.3492822966507177, + "grad_norm": 0.7753656834569056, + "learning_rate": 7.840072575681468e-06, + "loss": 1.254, + "num_tokens": 29637430.0, + "step": 146 + }, + { + "epoch": 0.35167464114832536, + "grad_norm": 0.9152188366691545, + "learning_rate": 7.810186787922645e-06, + "loss": 1.3457, + "num_tokens": 29878720.0, + "step": 147 + }, + { + "epoch": 0.35406698564593303, + "grad_norm": 0.6836864741610811, + "learning_rate": 7.78016199382112e-06, + "loss": 1.3482, + "num_tokens": 30112527.0, + "step": 148 + }, + { + "epoch": 0.35645933014354064, + "grad_norm": 0.7649029535599635, + "learning_rate": 7.75e-06, + "loss": 1.2517, + "num_tokens": 30303711.0, + "step": 149 + }, + { + "epoch": 0.3588516746411483, + "grad_norm": 0.6409818000165072, + "learning_rate": 7.719702621337834e-06, + "loss": 1.4567, + "num_tokens": 30495350.0, + "step": 150 + }, + { + "epoch": 0.361244019138756, + "grad_norm": 0.6118573341165198, + "learning_rate": 7.68927168085942e-06, + "loss": 1.5922, + "num_tokens": 30731067.0, + "step": 151 + }, + { + "epoch": 0.36363636363636365, + "grad_norm": 0.651410536310813, + "learning_rate": 7.658709009626109e-06, + "loss": 1.4001, + "num_tokens": 30991669.0, + "step": 152 + }, + { + "epoch": 0.3660287081339713, + "grad_norm": 0.6691292954708529, + "learning_rate": 7.628016446625626e-06, + "loss": 1.3455, + "num_tokens": 31201627.0, + "step": 153 + }, + { + "epoch": 0.3684210526315789, + "grad_norm": 0.6334035930223118, + "learning_rate": 7.597195838661426e-06, + "loss": 1.5338, + "num_tokens": 31406112.0, + "step": 154 + }, + { + "epoch": 0.3708133971291866, + "grad_norm": 0.6127199012442186, + "learning_rate": 7.566249040241553e-06, + "loss": 1.4887, + "num_tokens": 31609455.0, + "step": 155 + }, + { + "epoch": 0.37320574162679426, + "grad_norm": 0.7297799820848448, + "learning_rate": 7.53517791346707e-06, + "loss": 1.4004, + "num_tokens": 31780471.0, + "step": 156 + }, + { + "epoch": 0.37559808612440193, + "grad_norm": 0.6946608625160634, + "learning_rate": 7.503984327920003e-06, + "loss": 1.4223, + "num_tokens": 32000025.0, + "step": 157 + }, + { + "epoch": 0.37799043062200954, + "grad_norm": 0.8430323980042668, + "learning_rate": 7.472670160550849e-06, + "loss": 1.4304, + "num_tokens": 32231426.0, + "step": 158 + }, + { + "epoch": 0.3803827751196172, + "grad_norm": 0.5357203180428419, + "learning_rate": 7.441237295565642e-06, + "loss": 1.3439, + "num_tokens": 32506275.0, + "step": 159 }, { - "epoch": 2.0, - "step": 88, - "total_flos": 33029095800832.0, - "train_loss": 1.1323178586634723, - "train_runtime": 824.9985, - "train_samples_per_second": 3.36, - "train_steps_per_second": 0.107 + "epoch": 0.3827751196172249, + "grad_norm": 0.7610722225382753, + "learning_rate": 7.409687624312569e-06, + "loss": 1.5259, + "num_tokens": 32731571.0, + "step": 160 + }, + { + "epoch": 0.38516746411483255, + "grad_norm": 0.8234610158166737, + "learning_rate": 7.378023045168181e-06, + "loss": 1.4657, + "num_tokens": 32929424.0, + "step": 161 + }, + { + "epoch": 0.3875598086124402, + "grad_norm": 0.6460594254871024, + "learning_rate": 7.346245463423148e-06, + "loss": 1.1413, + "num_tokens": 33091736.0, + "step": 162 + }, + { + "epoch": 0.38995215311004783, + "grad_norm": 0.6455217606881992, + "learning_rate": 7.314356791167626e-06, + "loss": 1.4832, + "num_tokens": 33325006.0, + "step": 163 + }, + { + "epoch": 0.3923444976076555, + "grad_norm": 0.6417174012861143, + "learning_rate": 7.282358947176207e-06, + "loss": 1.4427, + "num_tokens": 33527584.0, + "step": 164 + }, + { + "epoch": 0.39473684210526316, + "grad_norm": 0.7003025374700267, + "learning_rate": 7.250253856792452e-06, + "loss": 1.4317, + "num_tokens": 33749355.0, + "step": 165 + }, + { + "epoch": 0.39712918660287083, + "grad_norm": 0.694765245729581, + "learning_rate": 7.218043451813058e-06, + "loss": 1.5191, + "num_tokens": 34042210.0, + "step": 166 + }, + { + "epoch": 0.39952153110047844, + "grad_norm": 0.6862863313107851, + "learning_rate": 7.185729670371605e-06, + "loss": 1.2854, + "num_tokens": 34231648.0, + "step": 167 + }, + { + "epoch": 0.4019138755980861, + "grad_norm": 0.627780658507271, + "learning_rate": 7.153314456821942e-06, + "loss": 1.429, + "num_tokens": 34451588.0, + "step": 168 + }, + { + "epoch": 0.4043062200956938, + "grad_norm": 0.824208162193406, + "learning_rate": 7.120799761621198e-06, + "loss": 1.1818, + "num_tokens": 34618135.0, + "step": 169 + }, + { + "epoch": 0.40669856459330145, + "grad_norm": 0.9393638903104259, + "learning_rate": 7.08818754121241e-06, + "loss": 1.3077, + "num_tokens": 34832232.0, + "step": 170 + }, + { + "epoch": 0.4090909090909091, + "grad_norm": 0.6416833884125813, + "learning_rate": 7.0554797579068155e-06, + "loss": 1.3405, + "num_tokens": 35024526.0, + "step": 171 + }, + { + "epoch": 0.41148325358851673, + "grad_norm": 0.5995702891023682, + "learning_rate": 7.022678379765766e-06, + "loss": 1.4452, + "num_tokens": 35268266.0, + "step": 172 + }, + { + "epoch": 0.4138755980861244, + "grad_norm": 0.636425551662653, + "learning_rate": 6.989785380482313e-06, + "loss": 1.3153, + "num_tokens": 35477419.0, + "step": 173 + }, + { + "epoch": 0.41626794258373206, + "grad_norm": 0.6814014450370663, + "learning_rate": 6.956802739262446e-06, + "loss": 1.3158, + "num_tokens": 35674184.0, + "step": 174 + }, + { + "epoch": 0.41866028708133973, + "grad_norm": 0.6594946487026292, + "learning_rate": 6.923732440706005e-06, + "loss": 1.1427, + "num_tokens": 35840117.0, + "step": 175 + }, + { + "epoch": 0.42105263157894735, + "grad_norm": 0.7126174750307512, + "learning_rate": 6.890576474687264e-06, + "loss": 1.4535, + "num_tokens": 36065504.0, + "step": 176 + }, + { + "epoch": 0.423444976076555, + "grad_norm": 0.6199132849036029, + "learning_rate": 6.857336836235195e-06, + "loss": 1.332, + "num_tokens": 36231288.0, + "step": 177 + }, + { + "epoch": 0.4258373205741627, + "grad_norm": 0.6209510353715602, + "learning_rate": 6.824015525413428e-06, + "loss": 1.3095, + "num_tokens": 36423759.0, + "step": 178 + }, + { + "epoch": 0.42822966507177035, + "grad_norm": 0.7108089495676037, + "learning_rate": 6.790614547199908e-06, + "loss": 1.3967, + "num_tokens": 36654811.0, + "step": 179 + }, + { + "epoch": 0.430622009569378, + "grad_norm": 0.8599346708468241, + "learning_rate": 6.7571359113662405e-06, + "loss": 1.1825, + "num_tokens": 36806684.0, + "step": 180 + }, + { + "epoch": 0.43301435406698563, + "grad_norm": 0.7746452027624019, + "learning_rate": 6.723581632356783e-06, + "loss": 1.2524, + "num_tokens": 36974284.0, + "step": 181 + }, + { + "epoch": 0.4354066985645933, + "grad_norm": 0.6020692340908421, + "learning_rate": 6.689953729167411e-06, + "loss": 1.2342, + "num_tokens": 37200621.0, + "step": 182 + }, + { + "epoch": 0.43779904306220097, + "grad_norm": 0.797055893773105, + "learning_rate": 6.65625422522405e-06, + "loss": 1.0268, + "num_tokens": 37342950.0, + "step": 183 + }, + { + "epoch": 0.44019138755980863, + "grad_norm": 0.7123312279475232, + "learning_rate": 6.622485148260916e-06, + "loss": 1.3021, + "num_tokens": 37515788.0, + "step": 184 + }, + { + "epoch": 0.44258373205741625, + "grad_norm": 0.7937734591260545, + "learning_rate": 6.588648530198505e-06, + "loss": 1.4699, + "num_tokens": 37746273.0, + "step": 185 + }, + { + "epoch": 0.4449760765550239, + "grad_norm": 0.8697751290589487, + "learning_rate": 6.554746407021332e-06, + "loss": 1.171, + "num_tokens": 37900069.0, + "step": 186 + }, + { + "epoch": 0.4473684210526316, + "grad_norm": 0.6523686305235465, + "learning_rate": 6.520780818655421e-06, + "loss": 1.3579, + "num_tokens": 38100846.0, + "step": 187 + }, + { + "epoch": 0.44976076555023925, + "grad_norm": 0.76156891130206, + "learning_rate": 6.486753808845565e-06, + "loss": 1.2426, + "num_tokens": 38269916.0, + "step": 188 + }, + { + "epoch": 0.45215311004784686, + "grad_norm": 0.7778813369773113, + "learning_rate": 6.45266742503235e-06, + "loss": 1.417, + "num_tokens": 38507557.0, + "step": 189 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 0.6252852335577198, + "learning_rate": 6.418523718228952e-06, + "loss": 1.485, + "num_tokens": 38773090.0, + "step": 190 + }, + { + "epoch": 0.4569377990430622, + "grad_norm": 0.6334892736043869, + "learning_rate": 6.3843247428977365e-06, + "loss": 1.16, + "num_tokens": 38974351.0, + "step": 191 + }, + { + "epoch": 0.45933014354066987, + "grad_norm": 0.6580511686936932, + "learning_rate": 6.350072556826632e-06, + "loss": 1.3935, + "num_tokens": 39123913.0, + "step": 192 + }, + { + "epoch": 0.46172248803827753, + "grad_norm": 0.9741231496059587, + "learning_rate": 6.315769221005313e-06, + "loss": 1.3265, + "num_tokens": 39375811.0, + "step": 193 + }, + { + "epoch": 0.46411483253588515, + "grad_norm": 0.6742978137081723, + "learning_rate": 6.281416799501188e-06, + "loss": 1.3934, + "num_tokens": 39601804.0, + "step": 194 + }, + { + "epoch": 0.4665071770334928, + "grad_norm": 0.7390771455576962, + "learning_rate": 6.247017359335199e-06, + "loss": 1.171, + "num_tokens": 39757895.0, + "step": 195 + }, + { + "epoch": 0.4688995215311005, + "grad_norm": 0.6747486626145579, + "learning_rate": 6.2125729703574534e-06, + "loss": 1.3421, + "num_tokens": 39951954.0, + "step": 196 + }, + { + "epoch": 0.47129186602870815, + "grad_norm": 0.6361657014007149, + "learning_rate": 6.178085705122675e-06, + "loss": 1.3744, + "num_tokens": 40144065.0, + "step": 197 + }, + { + "epoch": 0.47368421052631576, + "grad_norm": 0.6813668201444038, + "learning_rate": 6.143557638765494e-06, + "loss": 1.2369, + "num_tokens": 40362129.0, + "step": 198 + }, + { + "epoch": 0.47607655502392343, + "grad_norm": 0.5383944924160711, + "learning_rate": 6.108990848875591e-06, + "loss": 1.4575, + "num_tokens": 40573664.0, + "step": 199 + }, + { + "epoch": 0.4784688995215311, + "grad_norm": 0.5793342283366394, + "learning_rate": 6.074387415372677e-06, + "loss": 1.2934, + "num_tokens": 40796949.0, + "step": 200 + }, + { + "epoch": 0.48086124401913877, + "grad_norm": 0.5637720873789647, + "learning_rate": 6.039749420381349e-06, + "loss": 1.2067, + "num_tokens": 41001222.0, + "step": 201 + }, + { + "epoch": 0.48325358851674644, + "grad_norm": 0.6798868345680159, + "learning_rate": 6.005078948105808e-06, + "loss": 1.1563, + "num_tokens": 41187424.0, + "step": 202 + }, + { + "epoch": 0.48564593301435405, + "grad_norm": 0.6423876571963225, + "learning_rate": 5.970378084704441e-06, + "loss": 1.2355, + "num_tokens": 41372314.0, + "step": 203 + }, + { + "epoch": 0.4880382775119617, + "grad_norm": 0.5968488672427387, + "learning_rate": 5.935648918164308e-06, + "loss": 1.4124, + "num_tokens": 41599872.0, + "step": 204 + }, + { + "epoch": 0.4904306220095694, + "grad_norm": 0.7153999682703835, + "learning_rate": 5.90089353817549e-06, + "loss": 1.2553, + "num_tokens": 41781334.0, + "step": 205 + }, + { + "epoch": 0.49282296650717705, + "grad_norm": 0.7701093892397903, + "learning_rate": 5.866114036005363e-06, + "loss": 0.9825, + "num_tokens": 41932610.0, + "step": 206 + }, + { + "epoch": 0.49521531100478466, + "grad_norm": 0.6134522844262603, + "learning_rate": 5.831312504372762e-06, + "loss": 1.4134, + "num_tokens": 42124179.0, + "step": 207 + }, + { + "epoch": 0.49760765550239233, + "grad_norm": 0.6289100662436853, + "learning_rate": 5.796491037322054e-06, + "loss": 1.1654, + "num_tokens": 42345381.0, + "step": 208 + }, + { + "epoch": 0.5, + "grad_norm": 0.5666466799765261, + "learning_rate": 5.761651730097142e-06, + "loss": 1.2846, + "num_tokens": 42538163.0, + "step": 209 + }, + { + "epoch": 0.5023923444976076, + "grad_norm": 0.7188005672009754, + "learning_rate": 5.726796679015392e-06, + "loss": 1.3365, + "num_tokens": 42694348.0, + "step": 210 + }, + { + "epoch": 0.5047846889952153, + "grad_norm": 0.5851351897769017, + "learning_rate": 5.691927981341488e-06, + "loss": 1.2529, + "num_tokens": 42927824.0, + "step": 211 + }, + { + "epoch": 0.507177033492823, + "grad_norm": 0.6489926998886038, + "learning_rate": 5.657047735161256e-06, + "loss": 1.2083, + "num_tokens": 43119784.0, + "step": 212 + }, + { + "epoch": 0.5095693779904307, + "grad_norm": 0.6806360393034492, + "learning_rate": 5.622158039255394e-06, + "loss": 1.2419, + "num_tokens": 43338927.0, + "step": 213 + }, + { + "epoch": 0.5119617224880383, + "grad_norm": 0.6900003239395834, + "learning_rate": 5.58726099297321e-06, + "loss": 1.3539, + "num_tokens": 43489368.0, + "step": 214 + }, + { + "epoch": 0.5143540669856459, + "grad_norm": 0.7208226618388429, + "learning_rate": 5.552358696106288e-06, + "loss": 1.0978, + "num_tokens": 43636289.0, + "step": 215 + }, + { + "epoch": 0.5167464114832536, + "grad_norm": 0.8421973876061544, + "learning_rate": 5.517453248762142e-06, + "loss": 1.2866, + "num_tokens": 43862748.0, + "step": 216 + }, + { + "epoch": 0.5191387559808612, + "grad_norm": 0.6260067887308286, + "learning_rate": 5.482546751237859e-06, + "loss": 1.3206, + "num_tokens": 44060052.0, + "step": 217 + }, + { + "epoch": 0.5215311004784688, + "grad_norm": 0.5760944390891047, + "learning_rate": 5.447641303893715e-06, + "loss": 1.439, + "num_tokens": 44297069.0, + "step": 218 + }, + { + "epoch": 0.5239234449760766, + "grad_norm": 0.61734789956525, + "learning_rate": 5.412739007026791e-06, + "loss": 1.4205, + "num_tokens": 44490168.0, + "step": 219 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6257750740469292, + "learning_rate": 5.377841960744607e-06, + "loss": 1.3976, + "num_tokens": 44704452.0, + "step": 220 + }, + { + "epoch": 0.5287081339712919, + "grad_norm": 0.5344707538543756, + "learning_rate": 5.342952264838748e-06, + "loss": 1.2854, + "num_tokens": 44963417.0, + "step": 221 + }, + { + "epoch": 0.5311004784688995, + "grad_norm": 0.7495273348676206, + "learning_rate": 5.308072018658512e-06, + "loss": 1.041, + "num_tokens": 45085562.0, + "step": 222 + }, + { + "epoch": 0.5334928229665071, + "grad_norm": 0.6174774052991956, + "learning_rate": 5.273203320984611e-06, + "loss": 1.3262, + "num_tokens": 45278857.0, + "step": 223 + }, + { + "epoch": 0.5358851674641149, + "grad_norm": 0.5347084709996872, + "learning_rate": 5.23834826990286e-06, + "loss": 1.2666, + "num_tokens": 45496428.0, + "step": 224 + }, + { + "epoch": 0.5382775119617225, + "grad_norm": 0.537469270834151, + "learning_rate": 5.203508962677947e-06, + "loss": 1.1741, + "num_tokens": 45712553.0, + "step": 225 + }, + { + "epoch": 0.5406698564593302, + "grad_norm": 0.6378408098836623, + "learning_rate": 5.168687495627239e-06, + "loss": 1.1603, + "num_tokens": 45901658.0, + "step": 226 + }, + { + "epoch": 0.5430622009569378, + "grad_norm": 0.6300812904077526, + "learning_rate": 5.1338859639946396e-06, + "loss": 1.3905, + "num_tokens": 46088856.0, + "step": 227 + }, + { + "epoch": 0.5454545454545454, + "grad_norm": 0.634911596104642, + "learning_rate": 5.099106461824513e-06, + "loss": 1.2321, + "num_tokens": 46291266.0, + "step": 228 + }, + { + "epoch": 0.5478468899521531, + "grad_norm": 0.6926590384685983, + "learning_rate": 5.064351081835695e-06, + "loss": 1.1985, + "num_tokens": 46454514.0, + "step": 229 + }, + { + "epoch": 0.5502392344497608, + "grad_norm": 0.646212813197843, + "learning_rate": 5.02962191529556e-06, + "loss": 1.162, + "num_tokens": 46613219.0, + "step": 230 + }, + { + "epoch": 0.5526315789473685, + "grad_norm": 0.6542386722720754, + "learning_rate": 4.9949210518941945e-06, + "loss": 1.4429, + "num_tokens": 46844393.0, + "step": 231 + }, + { + "epoch": 0.5550239234449761, + "grad_norm": 0.5729641542767329, + "learning_rate": 4.960250579618652e-06, + "loss": 1.3042, + "num_tokens": 47083488.0, + "step": 232 + }, + { + "epoch": 0.5574162679425837, + "grad_norm": 0.6558395223982498, + "learning_rate": 4.925612584627325e-06, + "loss": 1.4322, + "num_tokens": 47296155.0, + "step": 233 + }, + { + "epoch": 0.5598086124401914, + "grad_norm": 0.5293743728173484, + "learning_rate": 4.8910091511244115e-06, + "loss": 1.4863, + "num_tokens": 47513493.0, + "step": 234 + }, + { + "epoch": 0.562200956937799, + "grad_norm": 0.5943251941494672, + "learning_rate": 4.856442361234507e-06, + "loss": 1.4734, + "num_tokens": 47716889.0, + "step": 235 + }, + { + "epoch": 0.5645933014354066, + "grad_norm": 0.6271955944541076, + "learning_rate": 4.821914294877327e-06, + "loss": 1.4372, + "num_tokens": 47925440.0, + "step": 236 + }, + { + "epoch": 0.5669856459330144, + "grad_norm": 0.7789065855387718, + "learning_rate": 4.787427029642549e-06, + "loss": 1.0957, + "num_tokens": 48073033.0, + "step": 237 + }, + { + "epoch": 0.569377990430622, + "grad_norm": 0.6495651224848629, + "learning_rate": 4.752982640664804e-06, + "loss": 1.3268, + "num_tokens": 48248595.0, + "step": 238 + }, + { + "epoch": 0.5717703349282297, + "grad_norm": 0.6160766106630973, + "learning_rate": 4.718583200498814e-06, + "loss": 1.4054, + "num_tokens": 48424413.0, + "step": 239 + }, + { + "epoch": 0.5741626794258373, + "grad_norm": 0.623669847382596, + "learning_rate": 4.684230778994688e-06, + "loss": 1.3817, + "num_tokens": 48613957.0, + "step": 240 + }, + { + "epoch": 0.5765550239234449, + "grad_norm": 0.6760867003584786, + "learning_rate": 4.64992744317337e-06, + "loss": 1.4087, + "num_tokens": 48801450.0, + "step": 241 + }, + { + "epoch": 0.5789473684210527, + "grad_norm": 0.6726827041038008, + "learning_rate": 4.615675257102265e-06, + "loss": 1.3587, + "num_tokens": 48992762.0, + "step": 242 + }, + { + "epoch": 0.5813397129186603, + "grad_norm": 0.7586345648423406, + "learning_rate": 4.58147628177105e-06, + "loss": 1.405, + "num_tokens": 49206463.0, + "step": 243 + }, + { + "epoch": 0.583732057416268, + "grad_norm": 0.5748903626756076, + "learning_rate": 4.547332574967653e-06, + "loss": 1.2833, + "num_tokens": 49439116.0, + "step": 244 + }, + { + "epoch": 0.5861244019138756, + "grad_norm": 0.6459030659564946, + "learning_rate": 4.513246191154434e-06, + "loss": 1.3808, + "num_tokens": 49611111.0, + "step": 245 + }, + { + "epoch": 0.5885167464114832, + "grad_norm": 0.5897283591830758, + "learning_rate": 4.479219181344579e-06, + "loss": 1.414, + "num_tokens": 49808207.0, + "step": 246 + }, + { + "epoch": 0.5909090909090909, + "grad_norm": 0.5803885935290669, + "learning_rate": 4.44525359297867e-06, + "loss": 1.412, + "num_tokens": 50005343.0, + "step": 247 + }, + { + "epoch": 0.5933014354066986, + "grad_norm": 0.587903351707854, + "learning_rate": 4.4113514698014955e-06, + "loss": 1.3887, + "num_tokens": 50238374.0, + "step": 248 + }, + { + "epoch": 0.5956937799043063, + "grad_norm": 0.6203595406125716, + "learning_rate": 4.377514851739085e-06, + "loss": 1.2548, + "num_tokens": 50470788.0, + "step": 249 + }, + { + "epoch": 0.5980861244019139, + "grad_norm": 0.5500030062641514, + "learning_rate": 4.3437457747759515e-06, + "loss": 1.2764, + "num_tokens": 50686198.0, + "step": 250 + }, + { + "epoch": 0.6004784688995215, + "grad_norm": 0.7362861850087726, + "learning_rate": 4.310046270832592e-06, + "loss": 1.2957, + "num_tokens": 50846591.0, + "step": 251 + }, + { + "epoch": 0.6028708133971292, + "grad_norm": 0.69063454686759, + "learning_rate": 4.276418367643218e-06, + "loss": 1.2816, + "num_tokens": 51043154.0, + "step": 252 + }, + { + "epoch": 0.6052631578947368, + "grad_norm": 0.8279823527287432, + "learning_rate": 4.242864088633762e-06, + "loss": 1.2845, + "num_tokens": 51276675.0, + "step": 253 + }, + { + "epoch": 0.6076555023923444, + "grad_norm": 0.5263711391152379, + "learning_rate": 4.2093854528000955e-06, + "loss": 1.2058, + "num_tokens": 51505587.0, + "step": 254 + }, + { + "epoch": 0.6100478468899522, + "grad_norm": 0.6716320775000435, + "learning_rate": 4.175984474586572e-06, + "loss": 1.1494, + "num_tokens": 51659012.0, + "step": 255 + }, + { + "epoch": 0.6124401913875598, + "grad_norm": 0.584427424031427, + "learning_rate": 4.142663163764806e-06, + "loss": 1.4945, + "num_tokens": 51883796.0, + "step": 256 + }, + { + "epoch": 0.6148325358851675, + "grad_norm": 0.5336467093570565, + "learning_rate": 4.109423525312738e-06, + "loss": 1.4456, + "num_tokens": 52116569.0, + "step": 257 + }, + { + "epoch": 0.6172248803827751, + "grad_norm": 0.7082487916003727, + "learning_rate": 4.076267559293996e-06, + "loss": 1.2521, + "num_tokens": 52291317.0, + "step": 258 + }, + { + "epoch": 0.6196172248803827, + "grad_norm": 0.670162970807308, + "learning_rate": 4.043197260737556e-06, + "loss": 1.2754, + "num_tokens": 52459579.0, + "step": 259 + }, + { + "epoch": 0.6220095693779905, + "grad_norm": 0.6217645411272703, + "learning_rate": 4.0102146195176895e-06, + "loss": 0.9781, + "num_tokens": 52655741.0, + "step": 260 + }, + { + "epoch": 0.6244019138755981, + "grad_norm": 0.6680531388656962, + "learning_rate": 3.977321620234236e-06, + "loss": 1.2224, + "num_tokens": 52855219.0, + "step": 261 + }, + { + "epoch": 0.6267942583732058, + "grad_norm": 0.6570258266897226, + "learning_rate": 3.944520242093186e-06, + "loss": 1.0444, + "num_tokens": 53022058.0, + "step": 262 + }, + { + "epoch": 0.6291866028708134, + "grad_norm": 0.569909382172942, + "learning_rate": 3.911812458787592e-06, + "loss": 1.2736, + "num_tokens": 53260177.0, + "step": 263 + }, + { + "epoch": 0.631578947368421, + "grad_norm": 0.68299256588729, + "learning_rate": 3.8792002383788044e-06, + "loss": 1.1374, + "num_tokens": 53406591.0, + "step": 264 + }, + { + "epoch": 0.6339712918660287, + "grad_norm": 0.6110632336195695, + "learning_rate": 3.846685543178058e-06, + "loss": 1.3974, + "num_tokens": 53588048.0, + "step": 265 + }, + { + "epoch": 0.6363636363636364, + "grad_norm": 0.66529312464438, + "learning_rate": 3.8142703296283954e-06, + "loss": 1.1257, + "num_tokens": 53773418.0, + "step": 266 + }, + { + "epoch": 0.638755980861244, + "grad_norm": 0.7331354863770742, + "learning_rate": 3.7819565481869426e-06, + "loss": 1.135, + "num_tokens": 53949544.0, + "step": 267 + }, + { + "epoch": 0.6411483253588517, + "grad_norm": 0.691168541736748, + "learning_rate": 3.7497461432075477e-06, + "loss": 1.2486, + "num_tokens": 54137216.0, + "step": 268 + }, + { + "epoch": 0.6435406698564593, + "grad_norm": 0.640130099494791, + "learning_rate": 3.717641052823795e-06, + "loss": 1.0805, + "num_tokens": 54340688.0, + "step": 269 + }, + { + "epoch": 0.645933014354067, + "grad_norm": 0.5598721395410986, + "learning_rate": 3.6856432088323746e-06, + "loss": 1.3979, + "num_tokens": 54537725.0, + "step": 270 + }, + { + "epoch": 0.6483253588516746, + "grad_norm": 0.6272293974276336, + "learning_rate": 3.6537545365768543e-06, + "loss": 1.3884, + "num_tokens": 54744185.0, + "step": 271 + }, + { + "epoch": 0.6507177033492823, + "grad_norm": 0.6812681210406815, + "learning_rate": 3.6219769548318205e-06, + "loss": 1.2305, + "num_tokens": 54944311.0, + "step": 272 + }, + { + "epoch": 0.65311004784689, + "grad_norm": 0.5952820749936442, + "learning_rate": 3.5903123756874315e-06, + "loss": 1.3444, + "num_tokens": 55150654.0, + "step": 273 + }, + { + "epoch": 0.6555023923444976, + "grad_norm": 0.7839097202629515, + "learning_rate": 3.558762704434361e-06, + "loss": 1.1151, + "num_tokens": 55310137.0, + "step": 274 + }, + { + "epoch": 0.6578947368421053, + "grad_norm": 0.5761324938652522, + "learning_rate": 3.527329839449152e-06, + "loss": 1.5347, + "num_tokens": 55545292.0, + "step": 275 + }, + { + "epoch": 0.6602870813397129, + "grad_norm": 0.6610740055714768, + "learning_rate": 3.496015672079998e-06, + "loss": 1.1698, + "num_tokens": 55736137.0, + "step": 276 + }, + { + "epoch": 0.6626794258373205, + "grad_norm": 0.5351233279220332, + "learning_rate": 3.4648220865329312e-06, + "loss": 1.4086, + "num_tokens": 55973020.0, + "step": 277 + }, + { + "epoch": 0.6650717703349283, + "grad_norm": 0.5934609597098215, + "learning_rate": 3.4337509597584466e-06, + "loss": 1.2519, + "num_tokens": 56193548.0, + "step": 278 + }, + { + "epoch": 0.6674641148325359, + "grad_norm": 0.6315221359405954, + "learning_rate": 3.402804161338577e-06, + "loss": 1.4212, + "num_tokens": 56399279.0, + "step": 279 + }, + { + "epoch": 0.6698564593301436, + "grad_norm": 0.5278279782038036, + "learning_rate": 3.371983553374375e-06, + "loss": 1.476, + "num_tokens": 56627906.0, + "step": 280 + }, + { + "epoch": 0.6722488038277512, + "grad_norm": 0.576586533979081, + "learning_rate": 3.3412909903738937e-06, + "loss": 1.2902, + "num_tokens": 56809932.0, + "step": 281 + }, + { + "epoch": 0.6746411483253588, + "grad_norm": 0.6467781327397376, + "learning_rate": 3.310728319140581e-06, + "loss": 1.2593, + "num_tokens": 56984901.0, + "step": 282 + }, + { + "epoch": 0.6770334928229665, + "grad_norm": 0.5025331960324733, + "learning_rate": 3.2802973786621665e-06, + "loss": 1.3005, + "num_tokens": 57203056.0, + "step": 283 + }, + { + "epoch": 0.6794258373205742, + "grad_norm": 0.6049658416582278, + "learning_rate": 3.2500000000000015e-06, + "loss": 1.2841, + "num_tokens": 57416534.0, + "step": 284 + }, + { + "epoch": 0.6818181818181818, + "grad_norm": 0.6356567810366318, + "learning_rate": 3.2198380061788803e-06, + "loss": 1.35, + "num_tokens": 57634005.0, + "step": 285 + }, + { + "epoch": 0.6842105263157895, + "grad_norm": 0.6259305842227357, + "learning_rate": 3.1898132120773566e-06, + "loss": 1.1986, + "num_tokens": 57820926.0, + "step": 286 + }, + { + "epoch": 0.6866028708133971, + "grad_norm": 0.5597883896460356, + "learning_rate": 3.1599274243185314e-06, + "loss": 1.4769, + "num_tokens": 58055489.0, + "step": 287 + }, + { + "epoch": 0.6889952153110048, + "grad_norm": 0.6331651732983823, + "learning_rate": 3.1301824411613473e-06, + "loss": 1.2765, + "num_tokens": 58245472.0, + "step": 288 + }, + { + "epoch": 0.6913875598086124, + "grad_norm": 0.6737504201758567, + "learning_rate": 3.1005800523923906e-06, + "loss": 1.1567, + "num_tokens": 58473327.0, + "step": 289 + }, + { + "epoch": 0.69377990430622, + "grad_norm": 0.8142514854229884, + "learning_rate": 3.071122039218194e-06, + "loss": 1.3968, + "num_tokens": 58649683.0, + "step": 290 + }, + { + "epoch": 0.6961722488038278, + "grad_norm": 0.5368448130491544, + "learning_rate": 3.0418101741580586e-06, + "loss": 1.2969, + "num_tokens": 58868455.0, + "step": 291 + }, + { + "epoch": 0.6985645933014354, + "grad_norm": 0.6697873240820909, + "learning_rate": 3.012646220937403e-06, + "loss": 1.3331, + "num_tokens": 59087101.0, + "step": 292 + }, + { + "epoch": 0.7009569377990431, + "grad_norm": 0.6546487553486491, + "learning_rate": 2.98363193438164e-06, + "loss": 1.3721, + "num_tokens": 59275075.0, + "step": 293 + }, + { + "epoch": 0.7033492822966507, + "grad_norm": 0.5913784460775687, + "learning_rate": 2.9547690603105774e-06, + "loss": 1.3077, + "num_tokens": 59490836.0, + "step": 294 + }, + { + "epoch": 0.7057416267942583, + "grad_norm": 1.0971211488496566, + "learning_rate": 2.926059335433378e-06, + "loss": 1.3629, + "num_tokens": 59688459.0, + "step": 295 + }, + { + "epoch": 0.7081339712918661, + "grad_norm": 0.9520603620495983, + "learning_rate": 2.897504487244061e-06, + "loss": 1.1589, + "num_tokens": 59854480.0, + "step": 296 + }, + { + "epoch": 0.7105263157894737, + "grad_norm": 0.6060298756711499, + "learning_rate": 2.8691062339175512e-06, + "loss": 1.3453, + "num_tokens": 60038638.0, + "step": 297 + }, + { + "epoch": 0.7129186602870813, + "grad_norm": 0.6435411063885081, + "learning_rate": 2.8408662842063002e-06, + "loss": 1.278, + "num_tokens": 60203017.0, + "step": 298 + }, + { + "epoch": 0.715311004784689, + "grad_norm": 0.6129870408682873, + "learning_rate": 2.8127863373374637e-06, + "loss": 1.3159, + "num_tokens": 60465940.0, + "step": 299 + }, + { + "epoch": 0.7177033492822966, + "grad_norm": 0.7588206471875415, + "learning_rate": 2.7848680829106602e-06, + "loss": 1.3377, + "num_tokens": 60632204.0, + "step": 300 + }, + { + "epoch": 0.7200956937799043, + "grad_norm": 0.6755921460278633, + "learning_rate": 2.7571132007963074e-06, + "loss": 1.3766, + "num_tokens": 60862058.0, + "step": 301 + }, + { + "epoch": 0.722488038277512, + "grad_norm": 0.7825671027002948, + "learning_rate": 2.7295233610345384e-06, + "loss": 1.4489, + "num_tokens": 61079432.0, + "step": 302 + }, + { + "epoch": 0.7248803827751196, + "grad_norm": 0.7227060900229081, + "learning_rate": 2.7021002237347206e-06, + "loss": 1.2062, + "num_tokens": 61255175.0, + "step": 303 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 0.5029929904055921, + "learning_rate": 2.6748454389755576e-06, + "loss": 1.4672, + "num_tokens": 61519012.0, + "step": 304 + }, + { + "epoch": 0.7296650717703349, + "grad_norm": 0.5685050474018025, + "learning_rate": 2.647760646705804e-06, + "loss": 1.2661, + "num_tokens": 61737384.0, + "step": 305 + }, + { + "epoch": 0.7320574162679426, + "grad_norm": 0.5737444678346549, + "learning_rate": 2.620847476645594e-06, + "loss": 1.3634, + "num_tokens": 62014122.0, + "step": 306 + }, + { + "epoch": 0.7344497607655502, + "grad_norm": 0.5367398412947724, + "learning_rate": 2.5941075481883705e-06, + "loss": 1.2674, + "num_tokens": 62246863.0, + "step": 307 + }, + { + "epoch": 0.7368421052631579, + "grad_norm": 0.6243016009006609, + "learning_rate": 2.567542470303452e-06, + "loss": 1.168, + "num_tokens": 62409808.0, + "step": 308 + }, + { + "epoch": 0.7392344497607656, + "grad_norm": 0.5971922768594424, + "learning_rate": 2.5411538414392146e-06, + "loss": 1.2759, + "num_tokens": 62601276.0, + "step": 309 + }, + { + "epoch": 0.7416267942583732, + "grad_norm": 0.6815558823471888, + "learning_rate": 2.5149432494269134e-06, + "loss": 1.2409, + "num_tokens": 62811825.0, + "step": 310 + }, + { + "epoch": 0.7440191387559809, + "grad_norm": 0.5840128801515716, + "learning_rate": 2.4889122713851397e-06, + "loss": 1.3169, + "num_tokens": 63033086.0, + "step": 311 + }, + { + "epoch": 0.7464114832535885, + "grad_norm": 0.6273197788202933, + "learning_rate": 2.463062473624927e-06, + "loss": 1.3796, + "num_tokens": 63231740.0, + "step": 312 + }, + { + "epoch": 0.7488038277511961, + "grad_norm": 0.5265456617487613, + "learning_rate": 2.437395411555504e-06, + "loss": 1.5587, + "num_tokens": 63484868.0, + "step": 313 + }, + { + "epoch": 0.7511961722488039, + "grad_norm": 0.63416554890546, + "learning_rate": 2.4119126295906997e-06, + "loss": 1.337, + "num_tokens": 63670016.0, + "step": 314 + }, + { + "epoch": 0.7535885167464115, + "grad_norm": 0.5715322993600752, + "learning_rate": 2.3866156610560186e-06, + "loss": 1.3791, + "num_tokens": 63908153.0, + "step": 315 + }, + { + "epoch": 0.7559808612440191, + "grad_norm": 0.5352522962690041, + "learning_rate": 2.3615060280963797e-06, + "loss": 1.3869, + "num_tokens": 64124869.0, + "step": 316 + }, + { + "epoch": 0.7583732057416268, + "grad_norm": 0.6645309100106329, + "learning_rate": 2.3365852415845225e-06, + "loss": 1.1748, + "num_tokens": 64294032.0, + "step": 317 + }, + { + "epoch": 0.7607655502392344, + "grad_norm": 0.638253408315402, + "learning_rate": 2.3118548010301015e-06, + "loss": 1.3481, + "num_tokens": 64476190.0, + "step": 318 + }, + { + "epoch": 0.7631578947368421, + "grad_norm": 0.6354512288625345, + "learning_rate": 2.2873161944894552e-06, + "loss": 1.2763, + "num_tokens": 64648365.0, + "step": 319 + }, + { + "epoch": 0.7655502392344498, + "grad_norm": 0.6556279261993273, + "learning_rate": 2.262970898476071e-06, + "loss": 1.1877, + "num_tokens": 64835000.0, + "step": 320 + }, + { + "epoch": 0.7679425837320574, + "grad_norm": 0.5810049291016075, + "learning_rate": 2.2388203778717407e-06, + "loss": 1.2494, + "num_tokens": 65026464.0, + "step": 321 + }, + { + "epoch": 0.7703349282296651, + "grad_norm": 0.6418564147429942, + "learning_rate": 2.2148660858384147e-06, + "loss": 1.2868, + "num_tokens": 65251252.0, + "step": 322 + }, + { + "epoch": 0.7727272727272727, + "grad_norm": 0.48981392426231535, + "learning_rate": 2.1911094637307715e-06, + "loss": 1.3262, + "num_tokens": 65517715.0, + "step": 323 + }, + { + "epoch": 0.7751196172248804, + "grad_norm": 0.570398102904161, + "learning_rate": 2.1675519410094803e-06, + "loss": 1.2075, + "num_tokens": 65701842.0, + "step": 324 + }, + { + "epoch": 0.777511961722488, + "grad_norm": 0.718887759011956, + "learning_rate": 2.144194935155192e-06, + "loss": 1.0361, + "num_tokens": 65890598.0, + "step": 325 + }, + { + "epoch": 0.7799043062200957, + "grad_norm": 0.5958578032562041, + "learning_rate": 2.121039851583254e-06, + "loss": 1.1784, + "num_tokens": 66089193.0, + "step": 326 + }, + { + "epoch": 0.7822966507177034, + "grad_norm": 0.6321165266280917, + "learning_rate": 2.098088083559135e-06, + "loss": 1.4724, + "num_tokens": 66314970.0, + "step": 327 + }, + { + "epoch": 0.784688995215311, + "grad_norm": 0.7307933906135672, + "learning_rate": 2.0753410121145984e-06, + "loss": 1.3178, + "num_tokens": 66480313.0, + "step": 328 + }, + { + "epoch": 0.7870813397129187, + "grad_norm": 0.5974020419425694, + "learning_rate": 2.0528000059646e-06, + "loss": 1.319, + "num_tokens": 66682404.0, + "step": 329 + }, + { + "epoch": 0.7894736842105263, + "grad_norm": 0.6678260669267934, + "learning_rate": 2.0304664214249326e-06, + "loss": 1.1726, + "num_tokens": 66883951.0, + "step": 330 + }, + { + "epoch": 0.7918660287081339, + "grad_norm": 0.6478957893104581, + "learning_rate": 2.0083416023306163e-06, + "loss": 1.3654, + "num_tokens": 67061851.0, + "step": 331 + }, + { + "epoch": 0.7942583732057417, + "grad_norm": 0.6407419233083412, + "learning_rate": 1.986426879955034e-06, + "loss": 1.1735, + "num_tokens": 67244214.0, + "step": 332 + }, + { + "epoch": 0.7966507177033493, + "grad_norm": 0.5771873678756463, + "learning_rate": 1.9647235729298346e-06, + "loss": 1.2519, + "num_tokens": 67423338.0, + "step": 333 + }, + { + "epoch": 0.7990430622009569, + "grad_norm": 0.5247220569941472, + "learning_rate": 1.9432329871655837e-06, + "loss": 1.3961, + "num_tokens": 67651988.0, + "step": 334 + }, + { + "epoch": 0.8014354066985646, + "grad_norm": 0.5562304496267458, + "learning_rate": 1.9219564157731848e-06, + "loss": 1.3009, + "num_tokens": 67899296.0, + "step": 335 + }, + { + "epoch": 0.8038277511961722, + "grad_norm": 0.6411488634701462, + "learning_rate": 1.9008951389860785e-06, + "loss": 1.1231, + "num_tokens": 68079644.0, + "step": 336 + }, + { + "epoch": 0.80622009569378, + "grad_norm": 0.5322759860676641, + "learning_rate": 1.8800504240832012e-06, + "loss": 1.3515, + "num_tokens": 68293723.0, + "step": 337 + }, + { + "epoch": 0.8086124401913876, + "grad_norm": 0.5746223661246294, + "learning_rate": 1.8594235253127373e-06, + "loss": 1.2591, + "num_tokens": 68529723.0, + "step": 338 + }, + { + "epoch": 0.8110047846889952, + "grad_norm": 0.7085560876362006, + "learning_rate": 1.8390156838166464e-06, + "loss": 1.1149, + "num_tokens": 68705404.0, + "step": 339 + }, + { + "epoch": 0.8133971291866029, + "grad_norm": 0.5845527495613347, + "learning_rate": 1.8188281275559866e-06, + "loss": 1.2183, + "num_tokens": 68954531.0, + "step": 340 + }, + { + "epoch": 0.8157894736842105, + "grad_norm": 0.5202134256783918, + "learning_rate": 1.7988620712370197e-06, + "loss": 1.4849, + "num_tokens": 69236634.0, + "step": 341 + }, + { + "epoch": 0.8181818181818182, + "grad_norm": 0.5645505834969143, + "learning_rate": 1.7791187162381325e-06, + "loss": 1.4002, + "num_tokens": 69451514.0, + "step": 342 + }, + { + "epoch": 0.8205741626794258, + "grad_norm": 0.5607875212664545, + "learning_rate": 1.759599250537534e-06, + "loss": 1.3259, + "num_tokens": 69682913.0, + "step": 343 + }, + { + "epoch": 0.8229665071770335, + "grad_norm": 0.559929672578073, + "learning_rate": 1.740304848641787e-06, + "loss": 1.2551, + "num_tokens": 69901240.0, + "step": 344 + }, + { + "epoch": 0.8253588516746412, + "grad_norm": 0.6095925983926321, + "learning_rate": 1.7212366715151263e-06, + "loss": 1.287, + "num_tokens": 70116810.0, + "step": 345 + }, + { + "epoch": 0.8277511961722488, + "grad_norm": 0.6240251911568065, + "learning_rate": 1.702395866509612e-06, + "loss": 1.1325, + "num_tokens": 70301449.0, + "step": 346 + }, + { + "epoch": 0.8301435406698564, + "grad_norm": 0.5784795095657023, + "learning_rate": 1.6837835672960834e-06, + "loss": 1.2059, + "num_tokens": 70492602.0, + "step": 347 + }, + { + "epoch": 0.8325358851674641, + "grad_norm": 0.6504184788969882, + "learning_rate": 1.6654008937959498e-06, + "loss": 0.9334, + "num_tokens": 70656041.0, + "step": 348 + }, + { + "epoch": 0.8349282296650717, + "grad_norm": 0.5702583567689546, + "learning_rate": 1.6472489521138016e-06, + "loss": 1.2118, + "num_tokens": 70905608.0, + "step": 349 + }, + { + "epoch": 0.8373205741626795, + "grad_norm": 0.6274124275648116, + "learning_rate": 1.629328834470857e-06, + "loss": 1.1815, + "num_tokens": 71101191.0, + "step": 350 + }, + { + "epoch": 0.8397129186602871, + "grad_norm": 0.6033829272863582, + "learning_rate": 1.611641619139238e-06, + "loss": 1.2807, + "num_tokens": 71291926.0, + "step": 351 + }, + { + "epoch": 0.8421052631578947, + "grad_norm": 0.6273406564056667, + "learning_rate": 1.5941883703770968e-06, + "loss": 1.3962, + "num_tokens": 71533733.0, + "step": 352 + }, + { + "epoch": 0.8444976076555024, + "grad_norm": 0.5312463864078979, + "learning_rate": 1.57697013836457e-06, + "loss": 1.3811, + "num_tokens": 71747066.0, + "step": 353 + }, + { + "epoch": 0.84688995215311, + "grad_norm": 0.7082921095962508, + "learning_rate": 1.5599879591405917e-06, + "loss": 1.1701, + "num_tokens": 71900238.0, + "step": 354 + }, + { + "epoch": 0.8492822966507177, + "grad_norm": 0.6134430059504471, + "learning_rate": 1.5432428545405554e-06, + "loss": 1.2446, + "num_tokens": 72085575.0, + "step": 355 + }, + { + "epoch": 0.8516746411483254, + "grad_norm": 0.6204954269243784, + "learning_rate": 1.526735832134829e-06, + "loss": 1.1297, + "num_tokens": 72322159.0, + "step": 356 + }, + { + "epoch": 0.854066985645933, + "grad_norm": 0.565525186855862, + "learning_rate": 1.5104678851681253e-06, + "loss": 1.4408, + "num_tokens": 72528384.0, + "step": 357 + }, + { + "epoch": 0.8564593301435407, + "grad_norm": 0.5682317963462433, + "learning_rate": 1.4944399924997372e-06, + "loss": 1.3499, + "num_tokens": 72730257.0, + "step": 358 + }, + { + "epoch": 0.8588516746411483, + "grad_norm": 0.5129599811585936, + "learning_rate": 1.4786531185446455e-06, + "loss": 1.2999, + "num_tokens": 72969626.0, + "step": 359 + }, + { + "epoch": 0.861244019138756, + "grad_norm": 0.6444665971353251, + "learning_rate": 1.4631082132154806e-06, + "loss": 1.2785, + "num_tokens": 73149240.0, + "step": 360 + }, + { + "epoch": 0.8636363636363636, + "grad_norm": 0.5547734172552142, + "learning_rate": 1.4478062118653703e-06, + "loss": 1.4096, + "num_tokens": 73363813.0, + "step": 361 + }, + { + "epoch": 0.8660287081339713, + "grad_norm": 0.6371851858022577, + "learning_rate": 1.4327480352316581e-06, + "loss": 1.2693, + "num_tokens": 73551589.0, + "step": 362 + }, + { + "epoch": 0.868421052631579, + "grad_norm": 0.5546864610023559, + "learning_rate": 1.417934589380498e-06, + "loss": 1.403, + "num_tokens": 73790910.0, + "step": 363 + }, + { + "epoch": 0.8708133971291866, + "grad_norm": 0.6219072093620266, + "learning_rate": 1.4033667656523405e-06, + "loss": 1.2085, + "num_tokens": 73985624.0, + "step": 364 + }, + { + "epoch": 0.8732057416267942, + "grad_norm": 0.6235130422407937, + "learning_rate": 1.389045440608296e-06, + "loss": 1.0986, + "num_tokens": 74148795.0, + "step": 365 + }, + { + "epoch": 0.8755980861244019, + "grad_norm": 0.6006216529271874, + "learning_rate": 1.374971475977394e-06, + "loss": 1.3025, + "num_tokens": 74332841.0, + "step": 366 + }, + { + "epoch": 0.8779904306220095, + "grad_norm": 0.6404509379440522, + "learning_rate": 1.361145718604731e-06, + "loss": 1.2766, + "num_tokens": 74529785.0, + "step": 367 + }, + { + "epoch": 0.8803827751196173, + "grad_norm": 0.686225898948988, + "learning_rate": 1.3475690004005098e-06, + "loss": 1.2452, + "num_tokens": 74703039.0, + "step": 368 + }, + { + "epoch": 0.8827751196172249, + "grad_norm": 0.571283865051148, + "learning_rate": 1.3342421382899936e-06, + "loss": 1.2626, + "num_tokens": 74883297.0, + "step": 369 + }, + { + "epoch": 0.8851674641148325, + "grad_norm": 0.7382176987149066, + "learning_rate": 1.3211659341643412e-06, + "loss": 1.233, + "num_tokens": 75066196.0, + "step": 370 + }, + { + "epoch": 0.8875598086124402, + "grad_norm": 0.6611644800778096, + "learning_rate": 1.308341174832359e-06, + "loss": 1.1876, + "num_tokens": 75232348.0, + "step": 371 + }, + { + "epoch": 0.8899521531100478, + "grad_norm": 0.5130766984472136, + "learning_rate": 1.2957686319731623e-06, + "loss": 1.2034, + "num_tokens": 75458529.0, + "step": 372 + }, + { + "epoch": 0.8923444976076556, + "grad_norm": 0.5684819034489474, + "learning_rate": 1.2834490620897342e-06, + "loss": 1.1584, + "num_tokens": 75661699.0, + "step": 373 + }, + { + "epoch": 0.8947368421052632, + "grad_norm": 0.5422081420273679, + "learning_rate": 1.2713832064634127e-06, + "loss": 1.3424, + "num_tokens": 75916455.0, + "step": 374 + }, + { + "epoch": 0.8971291866028708, + "grad_norm": 0.617788912813985, + "learning_rate": 1.259571791109285e-06, + "loss": 1.3227, + "num_tokens": 76100165.0, + "step": 375 + }, + { + "epoch": 0.8995215311004785, + "grad_norm": 0.5559640103832523, + "learning_rate": 1.2480155267325039e-06, + "loss": 1.1962, + "num_tokens": 76326481.0, + "step": 376 + }, + { + "epoch": 0.9019138755980861, + "grad_norm": 0.5930770090555496, + "learning_rate": 1.2367151086855187e-06, + "loss": 1.2261, + "num_tokens": 76482451.0, + "step": 377 + }, + { + "epoch": 0.9043062200956937, + "grad_norm": 0.6461043468488586, + "learning_rate": 1.2256712169262415e-06, + "loss": 1.3309, + "num_tokens": 76692581.0, + "step": 378 + }, + { + "epoch": 0.9066985645933014, + "grad_norm": 0.6295579626427625, + "learning_rate": 1.2148845159771311e-06, + "loss": 1.0403, + "num_tokens": 76869394.0, + "step": 379 + }, + { + "epoch": 0.9090909090909091, + "grad_norm": 0.6317114832123317, + "learning_rate": 1.2043556548852065e-06, + "loss": 1.3877, + "num_tokens": 77057755.0, + "step": 380 + }, + { + "epoch": 0.9114832535885168, + "grad_norm": 0.5646978093938921, + "learning_rate": 1.1940852671829938e-06, + "loss": 1.1932, + "num_tokens": 77279900.0, + "step": 381 + }, + { + "epoch": 0.9138755980861244, + "grad_norm": 0.554856696245118, + "learning_rate": 1.184073970850408e-06, + "loss": 1.3931, + "num_tokens": 77474484.0, + "step": 382 + }, + { + "epoch": 0.916267942583732, + "grad_norm": 0.5125502030227311, + "learning_rate": 1.174322368277565e-06, + "loss": 1.5298, + "num_tokens": 77739918.0, + "step": 383 + }, + { + "epoch": 0.9186602870813397, + "grad_norm": 0.6245967453038902, + "learning_rate": 1.1648310462285386e-06, + "loss": 1.2966, + "num_tokens": 77876514.0, + "step": 384 + }, + { + "epoch": 0.9210526315789473, + "grad_norm": 0.510131355499661, + "learning_rate": 1.1556005758060517e-06, + "loss": 1.3328, + "num_tokens": 78102264.0, + "step": 385 + }, + { + "epoch": 0.9234449760765551, + "grad_norm": 0.5313647688128483, + "learning_rate": 1.146631512417113e-06, + "loss": 1.4715, + "num_tokens": 78336063.0, + "step": 386 + }, + { + "epoch": 0.9258373205741627, + "grad_norm": 0.5521683071808053, + "learning_rate": 1.1379243957395987e-06, + "loss": 1.231, + "num_tokens": 78529771.0, + "step": 387 + }, + { + "epoch": 0.9282296650717703, + "grad_norm": 0.7086439907733622, + "learning_rate": 1.1294797496897786e-06, + "loss": 1.1614, + "num_tokens": 78679587.0, + "step": 388 + }, + { + "epoch": 0.930622009569378, + "grad_norm": 0.5802516939308205, + "learning_rate": 1.121298082390793e-06, + "loss": 1.3184, + "num_tokens": 78878694.0, + "step": 389 + }, + { + "epoch": 0.9330143540669856, + "grad_norm": 0.5736760000180704, + "learning_rate": 1.113379886142075e-06, + "loss": 1.3361, + "num_tokens": 79073779.0, + "step": 390 + }, + { + "epoch": 0.9354066985645934, + "grad_norm": 0.5432525121846279, + "learning_rate": 1.105725637389732e-06, + "loss": 1.2006, + "num_tokens": 79270679.0, + "step": 391 + }, + { + "epoch": 0.937799043062201, + "grad_norm": 0.6631735593901804, + "learning_rate": 1.0983357966978747e-06, + "loss": 1.4223, + "num_tokens": 79451866.0, + "step": 392 + }, + { + "epoch": 0.9401913875598086, + "grad_norm": 0.5833059272480773, + "learning_rate": 1.0912108087209075e-06, + "loss": 1.3483, + "num_tokens": 79629121.0, + "step": 393 + }, + { + "epoch": 0.9425837320574163, + "grad_norm": 0.521975187666564, + "learning_rate": 1.084351102176769e-06, + "loss": 1.3392, + "num_tokens": 79864771.0, + "step": 394 + }, + { + "epoch": 0.9449760765550239, + "grad_norm": 0.5983490163395889, + "learning_rate": 1.0777570898211406e-06, + "loss": 1.138, + "num_tokens": 80090837.0, + "step": 395 + }, + { + "epoch": 0.9473684210526315, + "grad_norm": 0.6791397944349069, + "learning_rate": 1.0714291684226054e-06, + "loss": 1.1482, + "num_tokens": 80247684.0, + "step": 396 + }, + { + "epoch": 0.9497607655502392, + "grad_norm": 0.7135484583851197, + "learning_rate": 1.0653677187387787e-06, + "loss": 1.178, + "num_tokens": 80412689.0, + "step": 397 + }, + { + "epoch": 0.9521531100478469, + "grad_norm": 0.5467847425622099, + "learning_rate": 1.0595731054933937e-06, + "loss": 1.256, + "num_tokens": 80646002.0, + "step": 398 + }, + { + "epoch": 0.9545454545454546, + "grad_norm": 0.649264408997545, + "learning_rate": 1.0540456773543596e-06, + "loss": 1.2587, + "num_tokens": 80846542.0, + "step": 399 + }, + { + "epoch": 0.9569377990430622, + "grad_norm": 0.6408697258124432, + "learning_rate": 1.0487857669127782e-06, + "loss": 1.2991, + "num_tokens": 81039183.0, + "step": 400 + }, + { + "epoch": 0.9593301435406698, + "grad_norm": 0.6125041198323196, + "learning_rate": 1.0437936906629336e-06, + "loss": 1.2721, + "num_tokens": 81228745.0, + "step": 401 + }, + { + "epoch": 0.9617224880382775, + "grad_norm": 0.6078405250476319, + "learning_rate": 1.039069748983248e-06, + "loss": 1.3821, + "num_tokens": 81425019.0, + "step": 402 + }, + { + "epoch": 0.9641148325358851, + "grad_norm": 0.5595011118119477, + "learning_rate": 1.0346142261182064e-06, + "loss": 1.0514, + "num_tokens": 81634634.0, + "step": 403 + }, + { + "epoch": 0.9665071770334929, + "grad_norm": 0.4972617449675274, + "learning_rate": 1.0304273901612566e-06, + "loss": 1.341, + "num_tokens": 81852833.0, + "step": 404 + }, + { + "epoch": 0.9688995215311005, + "grad_norm": 0.5059294656620951, + "learning_rate": 1.0265094930386741e-06, + "loss": 1.309, + "num_tokens": 82114913.0, + "step": 405 + }, + { + "epoch": 0.9712918660287081, + "grad_norm": 0.5962308760747792, + "learning_rate": 1.0228607704944048e-06, + "loss": 1.3641, + "num_tokens": 82299779.0, + "step": 406 + }, + { + "epoch": 0.9736842105263158, + "grad_norm": 0.6131816008639356, + "learning_rate": 1.0194814420758806e-06, + "loss": 1.1839, + "num_tokens": 82469216.0, + "step": 407 + }, + { + "epoch": 0.9760765550239234, + "grad_norm": 0.6115969172998058, + "learning_rate": 1.0163717111208086e-06, + "loss": 1.3046, + "num_tokens": 82640081.0, + "step": 408 + }, + { + "epoch": 0.9784688995215312, + "grad_norm": 0.5612129186921679, + "learning_rate": 1.0135317647449362e-06, + "loss": 1.4059, + "num_tokens": 82843981.0, + "step": 409 + }, + { + "epoch": 0.9808612440191388, + "grad_norm": 0.587067444004318, + "learning_rate": 1.0109617738307914e-06, + "loss": 1.2261, + "num_tokens": 83020444.0, + "step": 410 + }, + { + "epoch": 0.9832535885167464, + "grad_norm": 0.5594996533209011, + "learning_rate": 1.0086618930174011e-06, + "loss": 1.2042, + "num_tokens": 83232598.0, + "step": 411 + }, + { + "epoch": 0.9856459330143541, + "grad_norm": 0.7071205597794052, + "learning_rate": 1.006632260690988e-06, + "loss": 1.2169, + "num_tokens": 83374569.0, + "step": 412 + }, + { + "epoch": 0.9880382775119617, + "grad_norm": 0.6748664203646131, + "learning_rate": 1.0048729989766396e-06, + "loss": 1.1071, + "num_tokens": 83586085.0, + "step": 413 + }, + { + "epoch": 0.9904306220095693, + "grad_norm": 0.5910387645740945, + "learning_rate": 1.0033842137309649e-06, + "loss": 1.4125, + "num_tokens": 83786689.0, + "step": 414 + }, + { + "epoch": 0.992822966507177, + "grad_norm": 0.6362289532425913, + "learning_rate": 1.0021659945357202e-06, + "loss": 1.2213, + "num_tokens": 83967541.0, + "step": 415 + }, + { + "epoch": 0.9952153110047847, + "grad_norm": 0.5429466958407293, + "learning_rate": 1.0012184146924225e-06, + "loss": 1.3142, + "num_tokens": 84191970.0, + "step": 416 + }, + { + "epoch": 0.9976076555023924, + "grad_norm": 0.5637271505857214, + "learning_rate": 1.0005415312179367e-06, + "loss": 1.3643, + "num_tokens": 84430751.0, + "step": 417 + }, + { + "epoch": 1.0, + "grad_norm": 0.5655593189952651, + "learning_rate": 1.0001353848410461e-06, + "loss": 1.2425, + "num_tokens": 84668558.0, + "step": 418 + }, + { + "epoch": 1.0, + "eval_loss": 0.7911388278007507, + "eval_num_tokens": 84668558.0, + "eval_runtime": 47.4555, + "eval_samples_per_second": 62.627, + "eval_steps_per_second": 7.839, + "step": 418 + }, + { + "epoch": 1.0, + "step": 418, + "total_flos": 179772395159552.0, + "train_loss": 1.3804568826581873, + "train_runtime": 1345.8555, + "train_samples_per_second": 19.872, + "train_steps_per_second": 0.311 } ], "logging_steps": 1, - "max_steps": 88, + "max_steps": 418, "num_input_tokens_seen": 0, - "num_train_epochs": 2, + "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { @@ -758,7 +3389,7 @@ "attributes": {} } }, - "total_flos": 33029095800832.0, + "total_flos": 179772395159552.0, "train_batch_size": 8, "trial_name": null, "trial_params": null