| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 418, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0023923444976076554, |
| "grad_norm": 35.77700764414347, |
| "learning_rate": 0.0, |
| "loss": 2.5866, |
| "num_tokens": 223542.0, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.004784688995215311, |
| "grad_norm": 34.45799391693798, |
| "learning_rate": 7.692307692307694e-07, |
| "loss": 2.4086, |
| "num_tokens": 463760.0, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.007177033492822967, |
| "grad_norm": 37.09122577441204, |
| "learning_rate": 1.5384615384615387e-06, |
| "loss": 2.5261, |
| "num_tokens": 645652.0, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.009569377990430622, |
| "grad_norm": 33.51928857090818, |
| "learning_rate": 2.307692307692308e-06, |
| "loss": 2.4839, |
| "num_tokens": 857399.0, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.011961722488038277, |
| "grad_norm": 28.56245226187573, |
| "learning_rate": 3.0769230769230774e-06, |
| "loss": 2.492, |
| "num_tokens": 1083627.0, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.014354066985645933, |
| "grad_norm": 23.350183104787792, |
| "learning_rate": 3.846153846153847e-06, |
| "loss": 2.3562, |
| "num_tokens": 1301423.0, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.01674641148325359, |
| "grad_norm": 18.9808075909167, |
| "learning_rate": 4.615384615384616e-06, |
| "loss": 2.3909, |
| "num_tokens": 1492075.0, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.019138755980861243, |
| "grad_norm": 10.708395743692822, |
| "learning_rate": 5.384615384615385e-06, |
| "loss": 2.2473, |
| "num_tokens": 1679562.0, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.0215311004784689, |
| "grad_norm": 9.224705958016894, |
| "learning_rate": 6.153846153846155e-06, |
| "loss": 2.1681, |
| "num_tokens": 1888359.0, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.023923444976076555, |
| "grad_norm": 8.986278630054308, |
| "learning_rate": 6.923076923076923e-06, |
| "loss": 2.1438, |
| "num_tokens": 2058538.0, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02631578947368421, |
| "grad_norm": 4.193291761632025, |
| "learning_rate": 7.692307692307694e-06, |
| "loss": 2.0833, |
| "num_tokens": 2214696.0, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.028708133971291867, |
| "grad_norm": 3.5075199043533334, |
| "learning_rate": 8.461538461538462e-06, |
| "loss": 1.9944, |
| "num_tokens": 2401723.0, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.03110047846889952, |
| "grad_norm": 3.2504195652374874, |
| "learning_rate": 9.230769230769232e-06, |
| "loss": 2.0099, |
| "num_tokens": 2592218.0, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.03349282296650718, |
| "grad_norm": 2.3589570152818795, |
| "learning_rate": 1e-05, |
| "loss": 1.9879, |
| "num_tokens": 2820770.0, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.03588516746411483, |
| "grad_norm": 3.6838613043114705, |
| "learning_rate": 9.999864615158956e-06, |
| "loss": 1.816, |
| "num_tokens": 3018187.0, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.03827751196172249, |
| "grad_norm": 2.804031211195472, |
| "learning_rate": 9.999458468782065e-06, |
| "loss": 1.9285, |
| "num_tokens": 3262110.0, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.04066985645933014, |
| "grad_norm": 2.7340483571073664, |
| "learning_rate": 9.998781585307577e-06, |
| "loss": 1.7399, |
| "num_tokens": 3461819.0, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.0430622009569378, |
| "grad_norm": 2.3573648641331952, |
| "learning_rate": 9.997834005464281e-06, |
| "loss": 1.8248, |
| "num_tokens": 3647749.0, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.045454545454545456, |
| "grad_norm": 2.7936113880349964, |
| "learning_rate": 9.996615786269036e-06, |
| "loss": 1.514, |
| "num_tokens": 3816664.0, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.04784688995215311, |
| "grad_norm": 1.5815077522510241, |
| "learning_rate": 9.995127001023362e-06, |
| "loss": 1.8473, |
| "num_tokens": 3981662.0, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.050239234449760764, |
| "grad_norm": 1.262398317459698, |
| "learning_rate": 9.993367739309013e-06, |
| "loss": 1.7998, |
| "num_tokens": 4166705.0, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.05263157894736842, |
| "grad_norm": 1.2085839356904715, |
| "learning_rate": 9.991338106982598e-06, |
| "loss": 1.7167, |
| "num_tokens": 4395460.0, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.05502392344497608, |
| "grad_norm": 0.9331954135736191, |
| "learning_rate": 9.98903822616921e-06, |
| "loss": 1.7225, |
| "num_tokens": 4660660.0, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.05741626794258373, |
| "grad_norm": 1.0079376114307204, |
| "learning_rate": 9.986468235255065e-06, |
| "loss": 1.7793, |
| "num_tokens": 4889032.0, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.05980861244019139, |
| "grad_norm": 1.0174373092140978, |
| "learning_rate": 9.983628288879193e-06, |
| "loss": 1.6658, |
| "num_tokens": 5068345.0, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.06220095693779904, |
| "grad_norm": 0.9895227235288739, |
| "learning_rate": 9.98051855792412e-06, |
| "loss": 1.7126, |
| "num_tokens": 5284697.0, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.0645933014354067, |
| "grad_norm": 0.864412154650663, |
| "learning_rate": 9.977139229505596e-06, |
| "loss": 1.658, |
| "num_tokens": 5521778.0, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.06698564593301436, |
| "grad_norm": 1.0342443056226127, |
| "learning_rate": 9.973490506961326e-06, |
| "loss": 1.5243, |
| "num_tokens": 5707434.0, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.06937799043062201, |
| "grad_norm": 0.8663151664394599, |
| "learning_rate": 9.969572609838745e-06, |
| "loss": 1.6671, |
| "num_tokens": 5937867.0, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.07177033492822966, |
| "grad_norm": 0.9781488889720827, |
| "learning_rate": 9.965385773881795e-06, |
| "loss": 1.6322, |
| "num_tokens": 6162073.0, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.07416267942583732, |
| "grad_norm": 0.9101239670857304, |
| "learning_rate": 9.960930251016752e-06, |
| "loss": 1.716, |
| "num_tokens": 6375008.0, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.07655502392344497, |
| "grad_norm": 0.9203816431810994, |
| "learning_rate": 9.956206309337067e-06, |
| "loss": 1.6335, |
| "num_tokens": 6568479.0, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.07894736842105263, |
| "grad_norm": 0.9924366034325527, |
| "learning_rate": 9.951214233087223e-06, |
| "loss": 1.6849, |
| "num_tokens": 6786875.0, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.08133971291866028, |
| "grad_norm": 1.0591590540309344, |
| "learning_rate": 9.945954322645643e-06, |
| "loss": 1.5556, |
| "num_tokens": 6983383.0, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.08373205741626795, |
| "grad_norm": 0.859091081777175, |
| "learning_rate": 9.940426894506608e-06, |
| "loss": 1.6373, |
| "num_tokens": 7188567.0, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.0861244019138756, |
| "grad_norm": 0.7819453253946674, |
| "learning_rate": 9.934632281261221e-06, |
| "loss": 1.5762, |
| "num_tokens": 7385552.0, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.08851674641148326, |
| "grad_norm": 0.773342660575658, |
| "learning_rate": 9.928570831577396e-06, |
| "loss": 1.5561, |
| "num_tokens": 7583628.0, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.09090909090909091, |
| "grad_norm": 0.6689075869463446, |
| "learning_rate": 9.922242910178862e-06, |
| "loss": 1.6899, |
| "num_tokens": 7809190.0, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.09330143540669857, |
| "grad_norm": 0.7505436942893252, |
| "learning_rate": 9.915648897823232e-06, |
| "loss": 1.554, |
| "num_tokens": 7997624.0, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.09569377990430622, |
| "grad_norm": 0.7015919399979235, |
| "learning_rate": 9.908789191279093e-06, |
| "loss": 1.5389, |
| "num_tokens": 8183494.0, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.09808612440191387, |
| "grad_norm": 0.6466036781332534, |
| "learning_rate": 9.901664203302126e-06, |
| "loss": 1.6405, |
| "num_tokens": 8408569.0, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.10047846889952153, |
| "grad_norm": 0.6548488840567466, |
| "learning_rate": 9.89427436261027e-06, |
| "loss": 1.5763, |
| "num_tokens": 8584336.0, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.10287081339712918, |
| "grad_norm": 0.699125932858908, |
| "learning_rate": 9.886620113857926e-06, |
| "loss": 1.5114, |
| "num_tokens": 8787695.0, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.10526315789473684, |
| "grad_norm": 0.723472446208709, |
| "learning_rate": 9.878701917609208e-06, |
| "loss": 1.5305, |
| "num_tokens": 9005521.0, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.1076555023923445, |
| "grad_norm": 0.7384196667665146, |
| "learning_rate": 9.870520250310223e-06, |
| "loss": 1.6305, |
| "num_tokens": 9221565.0, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.11004784688995216, |
| "grad_norm": 0.629523841227617, |
| "learning_rate": 9.862075604260402e-06, |
| "loss": 1.6484, |
| "num_tokens": 9444154.0, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.11244019138755981, |
| "grad_norm": 0.6425557703652314, |
| "learning_rate": 9.853368487582888e-06, |
| "loss": 1.4761, |
| "num_tokens": 9628428.0, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.11483253588516747, |
| "grad_norm": 0.6988538672174611, |
| "learning_rate": 9.84439942419395e-06, |
| "loss": 1.6356, |
| "num_tokens": 9862779.0, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.11722488038277512, |
| "grad_norm": 0.6304374061279027, |
| "learning_rate": 9.835168953771463e-06, |
| "loss": 1.5105, |
| "num_tokens": 10111617.0, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.11961722488038277, |
| "grad_norm": 0.6166276961179814, |
| "learning_rate": 9.825677631722436e-06, |
| "loss": 1.6297, |
| "num_tokens": 10349939.0, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.12200956937799043, |
| "grad_norm": 0.7275282546024043, |
| "learning_rate": 9.815926029149593e-06, |
| "loss": 1.5068, |
| "num_tokens": 10534354.0, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.12440191387559808, |
| "grad_norm": 0.6327061243856574, |
| "learning_rate": 9.805914732817007e-06, |
| "loss": 1.5524, |
| "num_tokens": 10742091.0, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.12679425837320574, |
| "grad_norm": 0.7084417481062125, |
| "learning_rate": 9.795644345114796e-06, |
| "loss": 1.6218, |
| "num_tokens": 10961264.0, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.1291866028708134, |
| "grad_norm": 0.7292180354816592, |
| "learning_rate": 9.78511548402287e-06, |
| "loss": 1.3672, |
| "num_tokens": 11119849.0, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.13157894736842105, |
| "grad_norm": 0.6555927818623691, |
| "learning_rate": 9.77432878307376e-06, |
| "loss": 1.5114, |
| "num_tokens": 11325079.0, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.1339712918660287, |
| "grad_norm": 0.7613236601730216, |
| "learning_rate": 9.763284891314481e-06, |
| "loss": 1.2232, |
| "num_tokens": 11509772.0, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.13636363636363635, |
| "grad_norm": 0.6430242992230627, |
| "learning_rate": 9.751984473267498e-06, |
| "loss": 1.5539, |
| "num_tokens": 11686001.0, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.13875598086124402, |
| "grad_norm": 0.8426300537945989, |
| "learning_rate": 9.740428208890716e-06, |
| "loss": 1.3227, |
| "num_tokens": 11861406.0, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.14114832535885166, |
| "grad_norm": 0.7044359465699243, |
| "learning_rate": 9.728616793536588e-06, |
| "loss": 1.5587, |
| "num_tokens": 12083010.0, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.14354066985645933, |
| "grad_norm": 0.6499151456833749, |
| "learning_rate": 9.716550937910268e-06, |
| "loss": 1.4789, |
| "num_tokens": 12331238.0, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.145933014354067, |
| "grad_norm": 0.8468125913528113, |
| "learning_rate": 9.70423136802684e-06, |
| "loss": 1.397, |
| "num_tokens": 12510110.0, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.14832535885167464, |
| "grad_norm": 0.6549777551235627, |
| "learning_rate": 9.691658825167641e-06, |
| "loss": 1.5487, |
| "num_tokens": 12703234.0, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.1507177033492823, |
| "grad_norm": 1.861908798533634, |
| "learning_rate": 9.67883406583566e-06, |
| "loss": 1.5182, |
| "num_tokens": 12941740.0, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.15311004784688995, |
| "grad_norm": 0.9710524431103752, |
| "learning_rate": 9.665757861710008e-06, |
| "loss": 0.9673, |
| "num_tokens": 13066657.0, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.15550239234449761, |
| "grad_norm": 0.7709985947911734, |
| "learning_rate": 9.652430999599491e-06, |
| "loss": 1.2468, |
| "num_tokens": 13220140.0, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.15789473684210525, |
| "grad_norm": 1.148184964185836, |
| "learning_rate": 9.638854281395271e-06, |
| "loss": 1.3733, |
| "num_tokens": 13397562.0, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.16028708133971292, |
| "grad_norm": 1.8909672590403999, |
| "learning_rate": 9.625028524022606e-06, |
| "loss": 1.5125, |
| "num_tokens": 13591137.0, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.16267942583732056, |
| "grad_norm": 0.6070509391624586, |
| "learning_rate": 9.610954559391704e-06, |
| "loss": 1.576, |
| "num_tokens": 13812222.0, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.16507177033492823, |
| "grad_norm": 0.9685185344694189, |
| "learning_rate": 9.596633234347661e-06, |
| "loss": 1.3289, |
| "num_tokens": 14034014.0, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.1674641148325359, |
| "grad_norm": 1.3816831844533364, |
| "learning_rate": 9.582065410619503e-06, |
| "loss": 1.6031, |
| "num_tokens": 14248962.0, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.16985645933014354, |
| "grad_norm": 1.0562478172616532, |
| "learning_rate": 9.567251964768343e-06, |
| "loss": 1.1815, |
| "num_tokens": 14461235.0, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.1722488038277512, |
| "grad_norm": 0.8780808898792398, |
| "learning_rate": 9.55219378813463e-06, |
| "loss": 1.4514, |
| "num_tokens": 14655207.0, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.17464114832535885, |
| "grad_norm": 1.1376167558495347, |
| "learning_rate": 9.53689178678452e-06, |
| "loss": 1.2722, |
| "num_tokens": 14832145.0, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.17703349282296652, |
| "grad_norm": 0.954111645763459, |
| "learning_rate": 9.521346881455356e-06, |
| "loss": 1.3542, |
| "num_tokens": 15010659.0, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.17942583732057416, |
| "grad_norm": 1.1453698076350756, |
| "learning_rate": 9.505560007500263e-06, |
| "loss": 1.3355, |
| "num_tokens": 15213072.0, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.18181818181818182, |
| "grad_norm": 0.7130576585117359, |
| "learning_rate": 9.489532114831876e-06, |
| "loss": 1.4303, |
| "num_tokens": 15403547.0, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.18421052631578946, |
| "grad_norm": 1.1940739116738854, |
| "learning_rate": 9.473264167865172e-06, |
| "loss": 1.4188, |
| "num_tokens": 15608337.0, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.18660287081339713, |
| "grad_norm": 0.8288648375057176, |
| "learning_rate": 9.456757145459445e-06, |
| "loss": 1.5327, |
| "num_tokens": 15844122.0, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.18899521531100477, |
| "grad_norm": 0.7407524557729414, |
| "learning_rate": 9.44001204085941e-06, |
| "loss": 1.2769, |
| "num_tokens": 16007831.0, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.19138755980861244, |
| "grad_norm": 0.6820580256711418, |
| "learning_rate": 9.423029861635431e-06, |
| "loss": 1.4512, |
| "num_tokens": 16192803.0, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.1937799043062201, |
| "grad_norm": 0.7648355853081366, |
| "learning_rate": 9.405811629622904e-06, |
| "loss": 1.5147, |
| "num_tokens": 16372150.0, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.19617224880382775, |
| "grad_norm": 0.9430997963743641, |
| "learning_rate": 9.388358380860763e-06, |
| "loss": 1.439, |
| "num_tokens": 16544020.0, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.19856459330143542, |
| "grad_norm": 0.6625036619665103, |
| "learning_rate": 9.370671165529146e-06, |
| "loss": 1.362, |
| "num_tokens": 16724075.0, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.20095693779904306, |
| "grad_norm": 0.6779552456686411, |
| "learning_rate": 9.3527510478862e-06, |
| "loss": 1.296, |
| "num_tokens": 16892945.0, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.20334928229665072, |
| "grad_norm": 0.8756233237648234, |
| "learning_rate": 9.334599106204051e-06, |
| "loss": 1.4463, |
| "num_tokens": 17052936.0, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.20574162679425836, |
| "grad_norm": 0.7900920573876685, |
| "learning_rate": 9.316216432703918e-06, |
| "loss": 1.1998, |
| "num_tokens": 17229219.0, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.20813397129186603, |
| "grad_norm": 0.6469526433608052, |
| "learning_rate": 9.29760413349039e-06, |
| "loss": 1.4746, |
| "num_tokens": 17439405.0, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.21052631578947367, |
| "grad_norm": 0.616638088055178, |
| "learning_rate": 9.278763328484875e-06, |
| "loss": 1.6045, |
| "num_tokens": 17642943.0, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.21291866028708134, |
| "grad_norm": 0.950240919701955, |
| "learning_rate": 9.259695151358215e-06, |
| "loss": 1.4296, |
| "num_tokens": 17893610.0, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.215311004784689, |
| "grad_norm": 0.8787225924649048, |
| "learning_rate": 9.240400749462467e-06, |
| "loss": 1.3394, |
| "num_tokens": 18041227.0, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.21770334928229665, |
| "grad_norm": 0.6650651036759095, |
| "learning_rate": 9.220881283761868e-06, |
| "loss": 1.4976, |
| "num_tokens": 18231908.0, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.22009569377990432, |
| "grad_norm": 1.0488690262480544, |
| "learning_rate": 9.20113792876298e-06, |
| "loss": 1.544, |
| "num_tokens": 18449826.0, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.22248803827751196, |
| "grad_norm": 0.8128101002861978, |
| "learning_rate": 9.181171872444015e-06, |
| "loss": 1.6064, |
| "num_tokens": 18678175.0, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.22488038277511962, |
| "grad_norm": 0.8352400142358486, |
| "learning_rate": 9.160984316183354e-06, |
| "loss": 1.3891, |
| "num_tokens": 18856695.0, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.22727272727272727, |
| "grad_norm": 0.6330668052004139, |
| "learning_rate": 9.140576474687263e-06, |
| "loss": 1.478, |
| "num_tokens": 19058871.0, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.22966507177033493, |
| "grad_norm": 0.6984617187165848, |
| "learning_rate": 9.1199495759168e-06, |
| "loss": 1.3507, |
| "num_tokens": 19251656.0, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.23205741626794257, |
| "grad_norm": 0.86370290023878, |
| "learning_rate": 9.099104861013922e-06, |
| "loss": 1.3834, |
| "num_tokens": 19432481.0, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.23444976076555024, |
| "grad_norm": 0.7464764871732519, |
| "learning_rate": 9.078043584226816e-06, |
| "loss": 1.2249, |
| "num_tokens": 19625994.0, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.23684210526315788, |
| "grad_norm": 0.7046998605163254, |
| "learning_rate": 9.056767012834417e-06, |
| "loss": 1.3523, |
| "num_tokens": 19879656.0, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.23923444976076555, |
| "grad_norm": 0.6797355825616871, |
| "learning_rate": 9.035276427070166e-06, |
| "loss": 1.4905, |
| "num_tokens": 20094011.0, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.24162679425837322, |
| "grad_norm": 0.643387767261834, |
| "learning_rate": 9.013573120044968e-06, |
| "loss": 1.4817, |
| "num_tokens": 20324769.0, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.24401913875598086, |
| "grad_norm": 0.6345521357259186, |
| "learning_rate": 8.991658397669384e-06, |
| "loss": 1.406, |
| "num_tokens": 20523715.0, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.24641148325358853, |
| "grad_norm": 0.5851359669759257, |
| "learning_rate": 8.96953357857507e-06, |
| "loss": 1.5156, |
| "num_tokens": 20744615.0, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.24880382775119617, |
| "grad_norm": 0.7543388243657427, |
| "learning_rate": 8.947199994035402e-06, |
| "loss": 1.4965, |
| "num_tokens": 20954429.0, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.2511961722488038, |
| "grad_norm": 0.6708121582654688, |
| "learning_rate": 8.924658987885403e-06, |
| "loss": 1.4257, |
| "num_tokens": 21152905.0, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.2535885167464115, |
| "grad_norm": 0.8402032350426654, |
| "learning_rate": 8.901911916440867e-06, |
| "loss": 1.5563, |
| "num_tokens": 21368251.0, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.25598086124401914, |
| "grad_norm": 0.7320851769974586, |
| "learning_rate": 8.878960148416747e-06, |
| "loss": 1.3888, |
| "num_tokens": 21508325.0, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.2583732057416268, |
| "grad_norm": 0.7657651705836107, |
| "learning_rate": 8.855805064844808e-06, |
| "loss": 1.2907, |
| "num_tokens": 21657394.0, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.2607655502392344, |
| "grad_norm": 0.7516028881611917, |
| "learning_rate": 8.832448058990522e-06, |
| "loss": 1.3294, |
| "num_tokens": 21898147.0, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.2631578947368421, |
| "grad_norm": 0.7412973068093045, |
| "learning_rate": 8.80889053626923e-06, |
| "loss": 1.3493, |
| "num_tokens": 22114888.0, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.26555023923444976, |
| "grad_norm": 0.715207637159085, |
| "learning_rate": 8.785133914161586e-06, |
| "loss": 1.3189, |
| "num_tokens": 22309919.0, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.2679425837320574, |
| "grad_norm": 0.8490500056983592, |
| "learning_rate": 8.761179622128264e-06, |
| "loss": 1.2958, |
| "num_tokens": 22463886.0, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.2703349282296651, |
| "grad_norm": 0.6593276317049697, |
| "learning_rate": 8.737029101523931e-06, |
| "loss": 1.3876, |
| "num_tokens": 22695323.0, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.2727272727272727, |
| "grad_norm": 0.809173159712039, |
| "learning_rate": 8.712683805510547e-06, |
| "loss": 1.2752, |
| "num_tokens": 22923680.0, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.2751196172248804, |
| "grad_norm": 0.7247603972983714, |
| "learning_rate": 8.6881451989699e-06, |
| "loss": 1.1074, |
| "num_tokens": 23075695.0, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.27751196172248804, |
| "grad_norm": 0.6591979377593177, |
| "learning_rate": 8.66341475841548e-06, |
| "loss": 1.4945, |
| "num_tokens": 23320851.0, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.2799043062200957, |
| "grad_norm": 1.1647325657638634, |
| "learning_rate": 8.638493971903621e-06, |
| "loss": 1.2714, |
| "num_tokens": 23524534.0, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.2822966507177033, |
| "grad_norm": 0.7586726684402458, |
| "learning_rate": 8.613384338943982e-06, |
| "loss": 1.3522, |
| "num_tokens": 23773345.0, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.284688995215311, |
| "grad_norm": 0.6733449030482199, |
| "learning_rate": 8.588087370409303e-06, |
| "loss": 1.3563, |
| "num_tokens": 23961835.0, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.28708133971291866, |
| "grad_norm": 1.1832042021851583, |
| "learning_rate": 8.562604588444498e-06, |
| "loss": 1.368, |
| "num_tokens": 24186825.0, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.2894736842105263, |
| "grad_norm": 1.1589708241685623, |
| "learning_rate": 8.536937526375075e-06, |
| "loss": 1.4034, |
| "num_tokens": 24422495.0, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.291866028708134, |
| "grad_norm": 0.7701824945643486, |
| "learning_rate": 8.511087728614863e-06, |
| "loss": 1.4039, |
| "num_tokens": 24638181.0, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.2942583732057416, |
| "grad_norm": 0.6500381479650347, |
| "learning_rate": 8.485056750573088e-06, |
| "loss": 1.4504, |
| "num_tokens": 24875758.0, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.2966507177033493, |
| "grad_norm": 0.6319890651859407, |
| "learning_rate": 8.458846158560787e-06, |
| "loss": 1.4881, |
| "num_tokens": 25089829.0, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.29904306220095694, |
| "grad_norm": 1.388998615476455, |
| "learning_rate": 8.43245752969655e-06, |
| "loss": 1.1747, |
| "num_tokens": 25311276.0, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.3014354066985646, |
| "grad_norm": 1.0789606093954458, |
| "learning_rate": 8.40589245181163e-06, |
| "loss": 1.1788, |
| "num_tokens": 25489153.0, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.3038277511961722, |
| "grad_norm": 0.7049145346829824, |
| "learning_rate": 8.379152523354407e-06, |
| "loss": 1.3323, |
| "num_tokens": 25669945.0, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.3062200956937799, |
| "grad_norm": 0.6756413517215413, |
| "learning_rate": 8.352239353294196e-06, |
| "loss": 1.4037, |
| "num_tokens": 25909865.0, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.30861244019138756, |
| "grad_norm": 0.8708376581094076, |
| "learning_rate": 8.325154561024445e-06, |
| "loss": 1.3679, |
| "num_tokens": 26127238.0, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.31100478468899523, |
| "grad_norm": 1.3947730736376298, |
| "learning_rate": 8.29789977626528e-06, |
| "loss": 1.368, |
| "num_tokens": 26410830.0, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.3133971291866029, |
| "grad_norm": 0.6066250801672678, |
| "learning_rate": 8.270476638965463e-06, |
| "loss": 1.4124, |
| "num_tokens": 26580685.0, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.3157894736842105, |
| "grad_norm": 0.6459251127241467, |
| "learning_rate": 8.242886799203696e-06, |
| "loss": 1.4815, |
| "num_tokens": 26810854.0, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.3181818181818182, |
| "grad_norm": 0.8523013271806112, |
| "learning_rate": 8.215131917089342e-06, |
| "loss": 1.3218, |
| "num_tokens": 27020107.0, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.32057416267942584, |
| "grad_norm": 0.80764930998251, |
| "learning_rate": 8.187213662662539e-06, |
| "loss": 1.4524, |
| "num_tokens": 27225900.0, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.3229665071770335, |
| "grad_norm": 0.8176074504442626, |
| "learning_rate": 8.159133715793701e-06, |
| "loss": 1.1947, |
| "num_tokens": 27437322.0, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.3253588516746411, |
| "grad_norm": 0.6612632769182687, |
| "learning_rate": 8.13089376608245e-06, |
| "loss": 1.4414, |
| "num_tokens": 27620222.0, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.3277511961722488, |
| "grad_norm": 0.7997382570532242, |
| "learning_rate": 8.102495512755939e-06, |
| "loss": 1.1692, |
| "num_tokens": 27779969.0, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.33014354066985646, |
| "grad_norm": 0.7611481255340015, |
| "learning_rate": 8.073940664566623e-06, |
| "loss": 1.2419, |
| "num_tokens": 27975654.0, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.33253588516746413, |
| "grad_norm": 1.2608591606765718, |
| "learning_rate": 8.045230939689425e-06, |
| "loss": 1.389, |
| "num_tokens": 28167517.0, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.3349282296650718, |
| "grad_norm": 0.5773709072619498, |
| "learning_rate": 8.016368065618361e-06, |
| "loss": 1.4468, |
| "num_tokens": 28484552.0, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.3373205741626794, |
| "grad_norm": 0.725628062383284, |
| "learning_rate": 7.987353779062598e-06, |
| "loss": 1.1387, |
| "num_tokens": 28658376.0, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.3397129186602871, |
| "grad_norm": 1.0360991727200277, |
| "learning_rate": 7.958189825841942e-06, |
| "loss": 1.0863, |
| "num_tokens": 28857873.0, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.34210526315789475, |
| "grad_norm": 0.8786402024406298, |
| "learning_rate": 7.928877960781808e-06, |
| "loss": 1.2436, |
| "num_tokens": 29065902.0, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.3444976076555024, |
| "grad_norm": 0.6828988509231846, |
| "learning_rate": 7.899419947607611e-06, |
| "loss": 1.4069, |
| "num_tokens": 29258786.0, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.34688995215311, |
| "grad_norm": 0.7208210642568142, |
| "learning_rate": 7.869817558838654e-06, |
| "loss": 1.0879, |
| "num_tokens": 29468217.0, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.3492822966507177, |
| "grad_norm": 0.7753656834569056, |
| "learning_rate": 7.840072575681468e-06, |
| "loss": 1.254, |
| "num_tokens": 29637430.0, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.35167464114832536, |
| "grad_norm": 0.9152188366691545, |
| "learning_rate": 7.810186787922645e-06, |
| "loss": 1.3457, |
| "num_tokens": 29878720.0, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.35406698564593303, |
| "grad_norm": 0.6836864741610811, |
| "learning_rate": 7.78016199382112e-06, |
| "loss": 1.3482, |
| "num_tokens": 30112527.0, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.35645933014354064, |
| "grad_norm": 0.7649029535599635, |
| "learning_rate": 7.75e-06, |
| "loss": 1.2517, |
| "num_tokens": 30303711.0, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.3588516746411483, |
| "grad_norm": 0.6409818000165072, |
| "learning_rate": 7.719702621337834e-06, |
| "loss": 1.4567, |
| "num_tokens": 30495350.0, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.361244019138756, |
| "grad_norm": 0.6118573341165198, |
| "learning_rate": 7.68927168085942e-06, |
| "loss": 1.5922, |
| "num_tokens": 30731067.0, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.36363636363636365, |
| "grad_norm": 0.651410536310813, |
| "learning_rate": 7.658709009626109e-06, |
| "loss": 1.4001, |
| "num_tokens": 30991669.0, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.3660287081339713, |
| "grad_norm": 0.6691292954708529, |
| "learning_rate": 7.628016446625626e-06, |
| "loss": 1.3455, |
| "num_tokens": 31201627.0, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.3684210526315789, |
| "grad_norm": 0.6334035930223118, |
| "learning_rate": 7.597195838661426e-06, |
| "loss": 1.5338, |
| "num_tokens": 31406112.0, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.3708133971291866, |
| "grad_norm": 0.6127199012442186, |
| "learning_rate": 7.566249040241553e-06, |
| "loss": 1.4887, |
| "num_tokens": 31609455.0, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.37320574162679426, |
| "grad_norm": 0.7297799820848448, |
| "learning_rate": 7.53517791346707e-06, |
| "loss": 1.4004, |
| "num_tokens": 31780471.0, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.37559808612440193, |
| "grad_norm": 0.6946608625160634, |
| "learning_rate": 7.503984327920003e-06, |
| "loss": 1.4223, |
| "num_tokens": 32000025.0, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.37799043062200954, |
| "grad_norm": 0.8430323980042668, |
| "learning_rate": 7.472670160550849e-06, |
| "loss": 1.4304, |
| "num_tokens": 32231426.0, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.3803827751196172, |
| "grad_norm": 0.5357203180428419, |
| "learning_rate": 7.441237295565642e-06, |
| "loss": 1.3439, |
| "num_tokens": 32506275.0, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.3827751196172249, |
| "grad_norm": 0.7610722225382753, |
| "learning_rate": 7.409687624312569e-06, |
| "loss": 1.5259, |
| "num_tokens": 32731571.0, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.38516746411483255, |
| "grad_norm": 0.8234610158166737, |
| "learning_rate": 7.378023045168181e-06, |
| "loss": 1.4657, |
| "num_tokens": 32929424.0, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.3875598086124402, |
| "grad_norm": 0.6460594254871024, |
| "learning_rate": 7.346245463423148e-06, |
| "loss": 1.1413, |
| "num_tokens": 33091736.0, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.38995215311004783, |
| "grad_norm": 0.6455217606881992, |
| "learning_rate": 7.314356791167626e-06, |
| "loss": 1.4832, |
| "num_tokens": 33325006.0, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.3923444976076555, |
| "grad_norm": 0.6417174012861143, |
| "learning_rate": 7.282358947176207e-06, |
| "loss": 1.4427, |
| "num_tokens": 33527584.0, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.39473684210526316, |
| "grad_norm": 0.7003025374700267, |
| "learning_rate": 7.250253856792452e-06, |
| "loss": 1.4317, |
| "num_tokens": 33749355.0, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.39712918660287083, |
| "grad_norm": 0.694765245729581, |
| "learning_rate": 7.218043451813058e-06, |
| "loss": 1.5191, |
| "num_tokens": 34042210.0, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.39952153110047844, |
| "grad_norm": 0.6862863313107851, |
| "learning_rate": 7.185729670371605e-06, |
| "loss": 1.2854, |
| "num_tokens": 34231648.0, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.4019138755980861, |
| "grad_norm": 0.627780658507271, |
| "learning_rate": 7.153314456821942e-06, |
| "loss": 1.429, |
| "num_tokens": 34451588.0, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.4043062200956938, |
| "grad_norm": 0.824208162193406, |
| "learning_rate": 7.120799761621198e-06, |
| "loss": 1.1818, |
| "num_tokens": 34618135.0, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.40669856459330145, |
| "grad_norm": 0.9393638903104259, |
| "learning_rate": 7.08818754121241e-06, |
| "loss": 1.3077, |
| "num_tokens": 34832232.0, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.4090909090909091, |
| "grad_norm": 0.6416833884125813, |
| "learning_rate": 7.0554797579068155e-06, |
| "loss": 1.3405, |
| "num_tokens": 35024526.0, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.41148325358851673, |
| "grad_norm": 0.5995702891023682, |
| "learning_rate": 7.022678379765766e-06, |
| "loss": 1.4452, |
| "num_tokens": 35268266.0, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.4138755980861244, |
| "grad_norm": 0.636425551662653, |
| "learning_rate": 6.989785380482313e-06, |
| "loss": 1.3153, |
| "num_tokens": 35477419.0, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.41626794258373206, |
| "grad_norm": 0.6814014450370663, |
| "learning_rate": 6.956802739262446e-06, |
| "loss": 1.3158, |
| "num_tokens": 35674184.0, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.41866028708133973, |
| "grad_norm": 0.6594946487026292, |
| "learning_rate": 6.923732440706005e-06, |
| "loss": 1.1427, |
| "num_tokens": 35840117.0, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.42105263157894735, |
| "grad_norm": 0.7126174750307512, |
| "learning_rate": 6.890576474687264e-06, |
| "loss": 1.4535, |
| "num_tokens": 36065504.0, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.423444976076555, |
| "grad_norm": 0.6199132849036029, |
| "learning_rate": 6.857336836235195e-06, |
| "loss": 1.332, |
| "num_tokens": 36231288.0, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.4258373205741627, |
| "grad_norm": 0.6209510353715602, |
| "learning_rate": 6.824015525413428e-06, |
| "loss": 1.3095, |
| "num_tokens": 36423759.0, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.42822966507177035, |
| "grad_norm": 0.7108089495676037, |
| "learning_rate": 6.790614547199908e-06, |
| "loss": 1.3967, |
| "num_tokens": 36654811.0, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.430622009569378, |
| "grad_norm": 0.8599346708468241, |
| "learning_rate": 6.7571359113662405e-06, |
| "loss": 1.1825, |
| "num_tokens": 36806684.0, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.43301435406698563, |
| "grad_norm": 0.7746452027624019, |
| "learning_rate": 6.723581632356783e-06, |
| "loss": 1.2524, |
| "num_tokens": 36974284.0, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.4354066985645933, |
| "grad_norm": 0.6020692340908421, |
| "learning_rate": 6.689953729167411e-06, |
| "loss": 1.2342, |
| "num_tokens": 37200621.0, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.43779904306220097, |
| "grad_norm": 0.797055893773105, |
| "learning_rate": 6.65625422522405e-06, |
| "loss": 1.0268, |
| "num_tokens": 37342950.0, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.44019138755980863, |
| "grad_norm": 0.7123312279475232, |
| "learning_rate": 6.622485148260916e-06, |
| "loss": 1.3021, |
| "num_tokens": 37515788.0, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.44258373205741625, |
| "grad_norm": 0.7937734591260545, |
| "learning_rate": 6.588648530198505e-06, |
| "loss": 1.4699, |
| "num_tokens": 37746273.0, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.4449760765550239, |
| "grad_norm": 0.8697751290589487, |
| "learning_rate": 6.554746407021332e-06, |
| "loss": 1.171, |
| "num_tokens": 37900069.0, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.4473684210526316, |
| "grad_norm": 0.6523686305235465, |
| "learning_rate": 6.520780818655421e-06, |
| "loss": 1.3579, |
| "num_tokens": 38100846.0, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.44976076555023925, |
| "grad_norm": 0.76156891130206, |
| "learning_rate": 6.486753808845565e-06, |
| "loss": 1.2426, |
| "num_tokens": 38269916.0, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.45215311004784686, |
| "grad_norm": 0.7778813369773113, |
| "learning_rate": 6.45266742503235e-06, |
| "loss": 1.417, |
| "num_tokens": 38507557.0, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.45454545454545453, |
| "grad_norm": 0.6252852335577198, |
| "learning_rate": 6.418523718228952e-06, |
| "loss": 1.485, |
| "num_tokens": 38773090.0, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.4569377990430622, |
| "grad_norm": 0.6334892736043869, |
| "learning_rate": 6.3843247428977365e-06, |
| "loss": 1.16, |
| "num_tokens": 38974351.0, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.45933014354066987, |
| "grad_norm": 0.6580511686936932, |
| "learning_rate": 6.350072556826632e-06, |
| "loss": 1.3935, |
| "num_tokens": 39123913.0, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.46172248803827753, |
| "grad_norm": 0.9741231496059587, |
| "learning_rate": 6.315769221005313e-06, |
| "loss": 1.3265, |
| "num_tokens": 39375811.0, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.46411483253588515, |
| "grad_norm": 0.6742978137081723, |
| "learning_rate": 6.281416799501188e-06, |
| "loss": 1.3934, |
| "num_tokens": 39601804.0, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.4665071770334928, |
| "grad_norm": 0.7390771455576962, |
| "learning_rate": 6.247017359335199e-06, |
| "loss": 1.171, |
| "num_tokens": 39757895.0, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.4688995215311005, |
| "grad_norm": 0.6747486626145579, |
| "learning_rate": 6.2125729703574534e-06, |
| "loss": 1.3421, |
| "num_tokens": 39951954.0, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.47129186602870815, |
| "grad_norm": 0.6361657014007149, |
| "learning_rate": 6.178085705122675e-06, |
| "loss": 1.3744, |
| "num_tokens": 40144065.0, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.47368421052631576, |
| "grad_norm": 0.6813668201444038, |
| "learning_rate": 6.143557638765494e-06, |
| "loss": 1.2369, |
| "num_tokens": 40362129.0, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.47607655502392343, |
| "grad_norm": 0.5383944924160711, |
| "learning_rate": 6.108990848875591e-06, |
| "loss": 1.4575, |
| "num_tokens": 40573664.0, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.4784688995215311, |
| "grad_norm": 0.5793342283366394, |
| "learning_rate": 6.074387415372677e-06, |
| "loss": 1.2934, |
| "num_tokens": 40796949.0, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.48086124401913877, |
| "grad_norm": 0.5637720873789647, |
| "learning_rate": 6.039749420381349e-06, |
| "loss": 1.2067, |
| "num_tokens": 41001222.0, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.48325358851674644, |
| "grad_norm": 0.6798868345680159, |
| "learning_rate": 6.005078948105808e-06, |
| "loss": 1.1563, |
| "num_tokens": 41187424.0, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.48564593301435405, |
| "grad_norm": 0.6423876571963225, |
| "learning_rate": 5.970378084704441e-06, |
| "loss": 1.2355, |
| "num_tokens": 41372314.0, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.4880382775119617, |
| "grad_norm": 0.5968488672427387, |
| "learning_rate": 5.935648918164308e-06, |
| "loss": 1.4124, |
| "num_tokens": 41599872.0, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.4904306220095694, |
| "grad_norm": 0.7153999682703835, |
| "learning_rate": 5.90089353817549e-06, |
| "loss": 1.2553, |
| "num_tokens": 41781334.0, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.49282296650717705, |
| "grad_norm": 0.7701093892397903, |
| "learning_rate": 5.866114036005363e-06, |
| "loss": 0.9825, |
| "num_tokens": 41932610.0, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.49521531100478466, |
| "grad_norm": 0.6134522844262603, |
| "learning_rate": 5.831312504372762e-06, |
| "loss": 1.4134, |
| "num_tokens": 42124179.0, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.49760765550239233, |
| "grad_norm": 0.6289100662436853, |
| "learning_rate": 5.796491037322054e-06, |
| "loss": 1.1654, |
| "num_tokens": 42345381.0, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5666466799765261, |
| "learning_rate": 5.761651730097142e-06, |
| "loss": 1.2846, |
| "num_tokens": 42538163.0, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.5023923444976076, |
| "grad_norm": 0.7188005672009754, |
| "learning_rate": 5.726796679015392e-06, |
| "loss": 1.3365, |
| "num_tokens": 42694348.0, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.5047846889952153, |
| "grad_norm": 0.5851351897769017, |
| "learning_rate": 5.691927981341488e-06, |
| "loss": 1.2529, |
| "num_tokens": 42927824.0, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.507177033492823, |
| "grad_norm": 0.6489926998886038, |
| "learning_rate": 5.657047735161256e-06, |
| "loss": 1.2083, |
| "num_tokens": 43119784.0, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.5095693779904307, |
| "grad_norm": 0.6806360393034492, |
| "learning_rate": 5.622158039255394e-06, |
| "loss": 1.2419, |
| "num_tokens": 43338927.0, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.5119617224880383, |
| "grad_norm": 0.6900003239395834, |
| "learning_rate": 5.58726099297321e-06, |
| "loss": 1.3539, |
| "num_tokens": 43489368.0, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.5143540669856459, |
| "grad_norm": 0.7208226618388429, |
| "learning_rate": 5.552358696106288e-06, |
| "loss": 1.0978, |
| "num_tokens": 43636289.0, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.5167464114832536, |
| "grad_norm": 0.8421973876061544, |
| "learning_rate": 5.517453248762142e-06, |
| "loss": 1.2866, |
| "num_tokens": 43862748.0, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.5191387559808612, |
| "grad_norm": 0.6260067887308286, |
| "learning_rate": 5.482546751237859e-06, |
| "loss": 1.3206, |
| "num_tokens": 44060052.0, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.5215311004784688, |
| "grad_norm": 0.5760944390891047, |
| "learning_rate": 5.447641303893715e-06, |
| "loss": 1.439, |
| "num_tokens": 44297069.0, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.5239234449760766, |
| "grad_norm": 0.61734789956525, |
| "learning_rate": 5.412739007026791e-06, |
| "loss": 1.4205, |
| "num_tokens": 44490168.0, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.5263157894736842, |
| "grad_norm": 0.6257750740469292, |
| "learning_rate": 5.377841960744607e-06, |
| "loss": 1.3976, |
| "num_tokens": 44704452.0, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.5287081339712919, |
| "grad_norm": 0.5344707538543756, |
| "learning_rate": 5.342952264838748e-06, |
| "loss": 1.2854, |
| "num_tokens": 44963417.0, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.5311004784688995, |
| "grad_norm": 0.7495273348676206, |
| "learning_rate": 5.308072018658512e-06, |
| "loss": 1.041, |
| "num_tokens": 45085562.0, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.5334928229665071, |
| "grad_norm": 0.6174774052991956, |
| "learning_rate": 5.273203320984611e-06, |
| "loss": 1.3262, |
| "num_tokens": 45278857.0, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.5358851674641149, |
| "grad_norm": 0.5347084709996872, |
| "learning_rate": 5.23834826990286e-06, |
| "loss": 1.2666, |
| "num_tokens": 45496428.0, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.5382775119617225, |
| "grad_norm": 0.537469270834151, |
| "learning_rate": 5.203508962677947e-06, |
| "loss": 1.1741, |
| "num_tokens": 45712553.0, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.5406698564593302, |
| "grad_norm": 0.6378408098836623, |
| "learning_rate": 5.168687495627239e-06, |
| "loss": 1.1603, |
| "num_tokens": 45901658.0, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.5430622009569378, |
| "grad_norm": 0.6300812904077526, |
| "learning_rate": 5.1338859639946396e-06, |
| "loss": 1.3905, |
| "num_tokens": 46088856.0, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.5454545454545454, |
| "grad_norm": 0.634911596104642, |
| "learning_rate": 5.099106461824513e-06, |
| "loss": 1.2321, |
| "num_tokens": 46291266.0, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.5478468899521531, |
| "grad_norm": 0.6926590384685983, |
| "learning_rate": 5.064351081835695e-06, |
| "loss": 1.1985, |
| "num_tokens": 46454514.0, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.5502392344497608, |
| "grad_norm": 0.646212813197843, |
| "learning_rate": 5.02962191529556e-06, |
| "loss": 1.162, |
| "num_tokens": 46613219.0, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.5526315789473685, |
| "grad_norm": 0.6542386722720754, |
| "learning_rate": 4.9949210518941945e-06, |
| "loss": 1.4429, |
| "num_tokens": 46844393.0, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.5550239234449761, |
| "grad_norm": 0.5729641542767329, |
| "learning_rate": 4.960250579618652e-06, |
| "loss": 1.3042, |
| "num_tokens": 47083488.0, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.5574162679425837, |
| "grad_norm": 0.6558395223982498, |
| "learning_rate": 4.925612584627325e-06, |
| "loss": 1.4322, |
| "num_tokens": 47296155.0, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.5598086124401914, |
| "grad_norm": 0.5293743728173484, |
| "learning_rate": 4.8910091511244115e-06, |
| "loss": 1.4863, |
| "num_tokens": 47513493.0, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.562200956937799, |
| "grad_norm": 0.5943251941494672, |
| "learning_rate": 4.856442361234507e-06, |
| "loss": 1.4734, |
| "num_tokens": 47716889.0, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.5645933014354066, |
| "grad_norm": 0.6271955944541076, |
| "learning_rate": 4.821914294877327e-06, |
| "loss": 1.4372, |
| "num_tokens": 47925440.0, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.5669856459330144, |
| "grad_norm": 0.7789065855387718, |
| "learning_rate": 4.787427029642549e-06, |
| "loss": 1.0957, |
| "num_tokens": 48073033.0, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.569377990430622, |
| "grad_norm": 0.6495651224848629, |
| "learning_rate": 4.752982640664804e-06, |
| "loss": 1.3268, |
| "num_tokens": 48248595.0, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.5717703349282297, |
| "grad_norm": 0.6160766106630973, |
| "learning_rate": 4.718583200498814e-06, |
| "loss": 1.4054, |
| "num_tokens": 48424413.0, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.5741626794258373, |
| "grad_norm": 0.623669847382596, |
| "learning_rate": 4.684230778994688e-06, |
| "loss": 1.3817, |
| "num_tokens": 48613957.0, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.5765550239234449, |
| "grad_norm": 0.6760867003584786, |
| "learning_rate": 4.64992744317337e-06, |
| "loss": 1.4087, |
| "num_tokens": 48801450.0, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.5789473684210527, |
| "grad_norm": 0.6726827041038008, |
| "learning_rate": 4.615675257102265e-06, |
| "loss": 1.3587, |
| "num_tokens": 48992762.0, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.5813397129186603, |
| "grad_norm": 0.7586345648423406, |
| "learning_rate": 4.58147628177105e-06, |
| "loss": 1.405, |
| "num_tokens": 49206463.0, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.583732057416268, |
| "grad_norm": 0.5748903626756076, |
| "learning_rate": 4.547332574967653e-06, |
| "loss": 1.2833, |
| "num_tokens": 49439116.0, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.5861244019138756, |
| "grad_norm": 0.6459030659564946, |
| "learning_rate": 4.513246191154434e-06, |
| "loss": 1.3808, |
| "num_tokens": 49611111.0, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.5885167464114832, |
| "grad_norm": 0.5897283591830758, |
| "learning_rate": 4.479219181344579e-06, |
| "loss": 1.414, |
| "num_tokens": 49808207.0, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.5909090909090909, |
| "grad_norm": 0.5803885935290669, |
| "learning_rate": 4.44525359297867e-06, |
| "loss": 1.412, |
| "num_tokens": 50005343.0, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.5933014354066986, |
| "grad_norm": 0.587903351707854, |
| "learning_rate": 4.4113514698014955e-06, |
| "loss": 1.3887, |
| "num_tokens": 50238374.0, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.5956937799043063, |
| "grad_norm": 0.6203595406125716, |
| "learning_rate": 4.377514851739085e-06, |
| "loss": 1.2548, |
| "num_tokens": 50470788.0, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.5980861244019139, |
| "grad_norm": 0.5500030062641514, |
| "learning_rate": 4.3437457747759515e-06, |
| "loss": 1.2764, |
| "num_tokens": 50686198.0, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.6004784688995215, |
| "grad_norm": 0.7362861850087726, |
| "learning_rate": 4.310046270832592e-06, |
| "loss": 1.2957, |
| "num_tokens": 50846591.0, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.6028708133971292, |
| "grad_norm": 0.69063454686759, |
| "learning_rate": 4.276418367643218e-06, |
| "loss": 1.2816, |
| "num_tokens": 51043154.0, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.6052631578947368, |
| "grad_norm": 0.8279823527287432, |
| "learning_rate": 4.242864088633762e-06, |
| "loss": 1.2845, |
| "num_tokens": 51276675.0, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.6076555023923444, |
| "grad_norm": 0.5263711391152379, |
| "learning_rate": 4.2093854528000955e-06, |
| "loss": 1.2058, |
| "num_tokens": 51505587.0, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.6100478468899522, |
| "grad_norm": 0.6716320775000435, |
| "learning_rate": 4.175984474586572e-06, |
| "loss": 1.1494, |
| "num_tokens": 51659012.0, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.6124401913875598, |
| "grad_norm": 0.584427424031427, |
| "learning_rate": 4.142663163764806e-06, |
| "loss": 1.4945, |
| "num_tokens": 51883796.0, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.6148325358851675, |
| "grad_norm": 0.5336467093570565, |
| "learning_rate": 4.109423525312738e-06, |
| "loss": 1.4456, |
| "num_tokens": 52116569.0, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.6172248803827751, |
| "grad_norm": 0.7082487916003727, |
| "learning_rate": 4.076267559293996e-06, |
| "loss": 1.2521, |
| "num_tokens": 52291317.0, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.6196172248803827, |
| "grad_norm": 0.670162970807308, |
| "learning_rate": 4.043197260737556e-06, |
| "loss": 1.2754, |
| "num_tokens": 52459579.0, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.6220095693779905, |
| "grad_norm": 0.6217645411272703, |
| "learning_rate": 4.0102146195176895e-06, |
| "loss": 0.9781, |
| "num_tokens": 52655741.0, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.6244019138755981, |
| "grad_norm": 0.6680531388656962, |
| "learning_rate": 3.977321620234236e-06, |
| "loss": 1.2224, |
| "num_tokens": 52855219.0, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.6267942583732058, |
| "grad_norm": 0.6570258266897226, |
| "learning_rate": 3.944520242093186e-06, |
| "loss": 1.0444, |
| "num_tokens": 53022058.0, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.6291866028708134, |
| "grad_norm": 0.569909382172942, |
| "learning_rate": 3.911812458787592e-06, |
| "loss": 1.2736, |
| "num_tokens": 53260177.0, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.631578947368421, |
| "grad_norm": 0.68299256588729, |
| "learning_rate": 3.8792002383788044e-06, |
| "loss": 1.1374, |
| "num_tokens": 53406591.0, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.6339712918660287, |
| "grad_norm": 0.6110632336195695, |
| "learning_rate": 3.846685543178058e-06, |
| "loss": 1.3974, |
| "num_tokens": 53588048.0, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.6363636363636364, |
| "grad_norm": 0.66529312464438, |
| "learning_rate": 3.8142703296283954e-06, |
| "loss": 1.1257, |
| "num_tokens": 53773418.0, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.638755980861244, |
| "grad_norm": 0.7331354863770742, |
| "learning_rate": 3.7819565481869426e-06, |
| "loss": 1.135, |
| "num_tokens": 53949544.0, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.6411483253588517, |
| "grad_norm": 0.691168541736748, |
| "learning_rate": 3.7497461432075477e-06, |
| "loss": 1.2486, |
| "num_tokens": 54137216.0, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.6435406698564593, |
| "grad_norm": 0.640130099494791, |
| "learning_rate": 3.717641052823795e-06, |
| "loss": 1.0805, |
| "num_tokens": 54340688.0, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.645933014354067, |
| "grad_norm": 0.5598721395410986, |
| "learning_rate": 3.6856432088323746e-06, |
| "loss": 1.3979, |
| "num_tokens": 54537725.0, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.6483253588516746, |
| "grad_norm": 0.6272293974276336, |
| "learning_rate": 3.6537545365768543e-06, |
| "loss": 1.3884, |
| "num_tokens": 54744185.0, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.6507177033492823, |
| "grad_norm": 0.6812681210406815, |
| "learning_rate": 3.6219769548318205e-06, |
| "loss": 1.2305, |
| "num_tokens": 54944311.0, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.65311004784689, |
| "grad_norm": 0.5952820749936442, |
| "learning_rate": 3.5903123756874315e-06, |
| "loss": 1.3444, |
| "num_tokens": 55150654.0, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.6555023923444976, |
| "grad_norm": 0.7839097202629515, |
| "learning_rate": 3.558762704434361e-06, |
| "loss": 1.1151, |
| "num_tokens": 55310137.0, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.6578947368421053, |
| "grad_norm": 0.5761324938652522, |
| "learning_rate": 3.527329839449152e-06, |
| "loss": 1.5347, |
| "num_tokens": 55545292.0, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.6602870813397129, |
| "grad_norm": 0.6610740055714768, |
| "learning_rate": 3.496015672079998e-06, |
| "loss": 1.1698, |
| "num_tokens": 55736137.0, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.6626794258373205, |
| "grad_norm": 0.5351233279220332, |
| "learning_rate": 3.4648220865329312e-06, |
| "loss": 1.4086, |
| "num_tokens": 55973020.0, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.6650717703349283, |
| "grad_norm": 0.5934609597098215, |
| "learning_rate": 3.4337509597584466e-06, |
| "loss": 1.2519, |
| "num_tokens": 56193548.0, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.6674641148325359, |
| "grad_norm": 0.6315221359405954, |
| "learning_rate": 3.402804161338577e-06, |
| "loss": 1.4212, |
| "num_tokens": 56399279.0, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.6698564593301436, |
| "grad_norm": 0.5278279782038036, |
| "learning_rate": 3.371983553374375e-06, |
| "loss": 1.476, |
| "num_tokens": 56627906.0, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.6722488038277512, |
| "grad_norm": 0.576586533979081, |
| "learning_rate": 3.3412909903738937e-06, |
| "loss": 1.2902, |
| "num_tokens": 56809932.0, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.6746411483253588, |
| "grad_norm": 0.6467781327397376, |
| "learning_rate": 3.310728319140581e-06, |
| "loss": 1.2593, |
| "num_tokens": 56984901.0, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.6770334928229665, |
| "grad_norm": 0.5025331960324733, |
| "learning_rate": 3.2802973786621665e-06, |
| "loss": 1.3005, |
| "num_tokens": 57203056.0, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.6794258373205742, |
| "grad_norm": 0.6049658416582278, |
| "learning_rate": 3.2500000000000015e-06, |
| "loss": 1.2841, |
| "num_tokens": 57416534.0, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.6818181818181818, |
| "grad_norm": 0.6356567810366318, |
| "learning_rate": 3.2198380061788803e-06, |
| "loss": 1.35, |
| "num_tokens": 57634005.0, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.6842105263157895, |
| "grad_norm": 0.6259305842227357, |
| "learning_rate": 3.1898132120773566e-06, |
| "loss": 1.1986, |
| "num_tokens": 57820926.0, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.6866028708133971, |
| "grad_norm": 0.5597883896460356, |
| "learning_rate": 3.1599274243185314e-06, |
| "loss": 1.4769, |
| "num_tokens": 58055489.0, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.6889952153110048, |
| "grad_norm": 0.6331651732983823, |
| "learning_rate": 3.1301824411613473e-06, |
| "loss": 1.2765, |
| "num_tokens": 58245472.0, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.6913875598086124, |
| "grad_norm": 0.6737504201758567, |
| "learning_rate": 3.1005800523923906e-06, |
| "loss": 1.1567, |
| "num_tokens": 58473327.0, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.69377990430622, |
| "grad_norm": 0.8142514854229884, |
| "learning_rate": 3.071122039218194e-06, |
| "loss": 1.3968, |
| "num_tokens": 58649683.0, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.6961722488038278, |
| "grad_norm": 0.5368448130491544, |
| "learning_rate": 3.0418101741580586e-06, |
| "loss": 1.2969, |
| "num_tokens": 58868455.0, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.6985645933014354, |
| "grad_norm": 0.6697873240820909, |
| "learning_rate": 3.012646220937403e-06, |
| "loss": 1.3331, |
| "num_tokens": 59087101.0, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.7009569377990431, |
| "grad_norm": 0.6546487553486491, |
| "learning_rate": 2.98363193438164e-06, |
| "loss": 1.3721, |
| "num_tokens": 59275075.0, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.7033492822966507, |
| "grad_norm": 0.5913784460775687, |
| "learning_rate": 2.9547690603105774e-06, |
| "loss": 1.3077, |
| "num_tokens": 59490836.0, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.7057416267942583, |
| "grad_norm": 1.0971211488496566, |
| "learning_rate": 2.926059335433378e-06, |
| "loss": 1.3629, |
| "num_tokens": 59688459.0, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.7081339712918661, |
| "grad_norm": 0.9520603620495983, |
| "learning_rate": 2.897504487244061e-06, |
| "loss": 1.1589, |
| "num_tokens": 59854480.0, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.7105263157894737, |
| "grad_norm": 0.6060298756711499, |
| "learning_rate": 2.8691062339175512e-06, |
| "loss": 1.3453, |
| "num_tokens": 60038638.0, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.7129186602870813, |
| "grad_norm": 0.6435411063885081, |
| "learning_rate": 2.8408662842063002e-06, |
| "loss": 1.278, |
| "num_tokens": 60203017.0, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.715311004784689, |
| "grad_norm": 0.6129870408682873, |
| "learning_rate": 2.8127863373374637e-06, |
| "loss": 1.3159, |
| "num_tokens": 60465940.0, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.7177033492822966, |
| "grad_norm": 0.7588206471875415, |
| "learning_rate": 2.7848680829106602e-06, |
| "loss": 1.3377, |
| "num_tokens": 60632204.0, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.7200956937799043, |
| "grad_norm": 0.6755921460278633, |
| "learning_rate": 2.7571132007963074e-06, |
| "loss": 1.3766, |
| "num_tokens": 60862058.0, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.722488038277512, |
| "grad_norm": 0.7825671027002948, |
| "learning_rate": 2.7295233610345384e-06, |
| "loss": 1.4489, |
| "num_tokens": 61079432.0, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.7248803827751196, |
| "grad_norm": 0.7227060900229081, |
| "learning_rate": 2.7021002237347206e-06, |
| "loss": 1.2062, |
| "num_tokens": 61255175.0, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.7272727272727273, |
| "grad_norm": 0.5029929904055921, |
| "learning_rate": 2.6748454389755576e-06, |
| "loss": 1.4672, |
| "num_tokens": 61519012.0, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.7296650717703349, |
| "grad_norm": 0.5685050474018025, |
| "learning_rate": 2.647760646705804e-06, |
| "loss": 1.2661, |
| "num_tokens": 61737384.0, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.7320574162679426, |
| "grad_norm": 0.5737444678346549, |
| "learning_rate": 2.620847476645594e-06, |
| "loss": 1.3634, |
| "num_tokens": 62014122.0, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.7344497607655502, |
| "grad_norm": 0.5367398412947724, |
| "learning_rate": 2.5941075481883705e-06, |
| "loss": 1.2674, |
| "num_tokens": 62246863.0, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.7368421052631579, |
| "grad_norm": 0.6243016009006609, |
| "learning_rate": 2.567542470303452e-06, |
| "loss": 1.168, |
| "num_tokens": 62409808.0, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.7392344497607656, |
| "grad_norm": 0.5971922768594424, |
| "learning_rate": 2.5411538414392146e-06, |
| "loss": 1.2759, |
| "num_tokens": 62601276.0, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.7416267942583732, |
| "grad_norm": 0.6815558823471888, |
| "learning_rate": 2.5149432494269134e-06, |
| "loss": 1.2409, |
| "num_tokens": 62811825.0, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.7440191387559809, |
| "grad_norm": 0.5840128801515716, |
| "learning_rate": 2.4889122713851397e-06, |
| "loss": 1.3169, |
| "num_tokens": 63033086.0, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.7464114832535885, |
| "grad_norm": 0.6273197788202933, |
| "learning_rate": 2.463062473624927e-06, |
| "loss": 1.3796, |
| "num_tokens": 63231740.0, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.7488038277511961, |
| "grad_norm": 0.5265456617487613, |
| "learning_rate": 2.437395411555504e-06, |
| "loss": 1.5587, |
| "num_tokens": 63484868.0, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.7511961722488039, |
| "grad_norm": 0.63416554890546, |
| "learning_rate": 2.4119126295906997e-06, |
| "loss": 1.337, |
| "num_tokens": 63670016.0, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.7535885167464115, |
| "grad_norm": 0.5715322993600752, |
| "learning_rate": 2.3866156610560186e-06, |
| "loss": 1.3791, |
| "num_tokens": 63908153.0, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.7559808612440191, |
| "grad_norm": 0.5352522962690041, |
| "learning_rate": 2.3615060280963797e-06, |
| "loss": 1.3869, |
| "num_tokens": 64124869.0, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.7583732057416268, |
| "grad_norm": 0.6645309100106329, |
| "learning_rate": 2.3365852415845225e-06, |
| "loss": 1.1748, |
| "num_tokens": 64294032.0, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.7607655502392344, |
| "grad_norm": 0.638253408315402, |
| "learning_rate": 2.3118548010301015e-06, |
| "loss": 1.3481, |
| "num_tokens": 64476190.0, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.7631578947368421, |
| "grad_norm": 0.6354512288625345, |
| "learning_rate": 2.2873161944894552e-06, |
| "loss": 1.2763, |
| "num_tokens": 64648365.0, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.7655502392344498, |
| "grad_norm": 0.6556279261993273, |
| "learning_rate": 2.262970898476071e-06, |
| "loss": 1.1877, |
| "num_tokens": 64835000.0, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.7679425837320574, |
| "grad_norm": 0.5810049291016075, |
| "learning_rate": 2.2388203778717407e-06, |
| "loss": 1.2494, |
| "num_tokens": 65026464.0, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.7703349282296651, |
| "grad_norm": 0.6418564147429942, |
| "learning_rate": 2.2148660858384147e-06, |
| "loss": 1.2868, |
| "num_tokens": 65251252.0, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.7727272727272727, |
| "grad_norm": 0.48981392426231535, |
| "learning_rate": 2.1911094637307715e-06, |
| "loss": 1.3262, |
| "num_tokens": 65517715.0, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.7751196172248804, |
| "grad_norm": 0.570398102904161, |
| "learning_rate": 2.1675519410094803e-06, |
| "loss": 1.2075, |
| "num_tokens": 65701842.0, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.777511961722488, |
| "grad_norm": 0.718887759011956, |
| "learning_rate": 2.144194935155192e-06, |
| "loss": 1.0361, |
| "num_tokens": 65890598.0, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.7799043062200957, |
| "grad_norm": 0.5958578032562041, |
| "learning_rate": 2.121039851583254e-06, |
| "loss": 1.1784, |
| "num_tokens": 66089193.0, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.7822966507177034, |
| "grad_norm": 0.6321165266280917, |
| "learning_rate": 2.098088083559135e-06, |
| "loss": 1.4724, |
| "num_tokens": 66314970.0, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.784688995215311, |
| "grad_norm": 0.7307933906135672, |
| "learning_rate": 2.0753410121145984e-06, |
| "loss": 1.3178, |
| "num_tokens": 66480313.0, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.7870813397129187, |
| "grad_norm": 0.5974020419425694, |
| "learning_rate": 2.0528000059646e-06, |
| "loss": 1.319, |
| "num_tokens": 66682404.0, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.7894736842105263, |
| "grad_norm": 0.6678260669267934, |
| "learning_rate": 2.0304664214249326e-06, |
| "loss": 1.1726, |
| "num_tokens": 66883951.0, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.7918660287081339, |
| "grad_norm": 0.6478957893104581, |
| "learning_rate": 2.0083416023306163e-06, |
| "loss": 1.3654, |
| "num_tokens": 67061851.0, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.7942583732057417, |
| "grad_norm": 0.6407419233083412, |
| "learning_rate": 1.986426879955034e-06, |
| "loss": 1.1735, |
| "num_tokens": 67244214.0, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.7966507177033493, |
| "grad_norm": 0.5771873678756463, |
| "learning_rate": 1.9647235729298346e-06, |
| "loss": 1.2519, |
| "num_tokens": 67423338.0, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.7990430622009569, |
| "grad_norm": 0.5247220569941472, |
| "learning_rate": 1.9432329871655837e-06, |
| "loss": 1.3961, |
| "num_tokens": 67651988.0, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.8014354066985646, |
| "grad_norm": 0.5562304496267458, |
| "learning_rate": 1.9219564157731848e-06, |
| "loss": 1.3009, |
| "num_tokens": 67899296.0, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.8038277511961722, |
| "grad_norm": 0.6411488634701462, |
| "learning_rate": 1.9008951389860785e-06, |
| "loss": 1.1231, |
| "num_tokens": 68079644.0, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.80622009569378, |
| "grad_norm": 0.5322759860676641, |
| "learning_rate": 1.8800504240832012e-06, |
| "loss": 1.3515, |
| "num_tokens": 68293723.0, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.8086124401913876, |
| "grad_norm": 0.5746223661246294, |
| "learning_rate": 1.8594235253127373e-06, |
| "loss": 1.2591, |
| "num_tokens": 68529723.0, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.8110047846889952, |
| "grad_norm": 0.7085560876362006, |
| "learning_rate": 1.8390156838166464e-06, |
| "loss": 1.1149, |
| "num_tokens": 68705404.0, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.8133971291866029, |
| "grad_norm": 0.5845527495613347, |
| "learning_rate": 1.8188281275559866e-06, |
| "loss": 1.2183, |
| "num_tokens": 68954531.0, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.8157894736842105, |
| "grad_norm": 0.5202134256783918, |
| "learning_rate": 1.7988620712370197e-06, |
| "loss": 1.4849, |
| "num_tokens": 69236634.0, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.8181818181818182, |
| "grad_norm": 0.5645505834969143, |
| "learning_rate": 1.7791187162381325e-06, |
| "loss": 1.4002, |
| "num_tokens": 69451514.0, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.8205741626794258, |
| "grad_norm": 0.5607875212664545, |
| "learning_rate": 1.759599250537534e-06, |
| "loss": 1.3259, |
| "num_tokens": 69682913.0, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.8229665071770335, |
| "grad_norm": 0.559929672578073, |
| "learning_rate": 1.740304848641787e-06, |
| "loss": 1.2551, |
| "num_tokens": 69901240.0, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.8253588516746412, |
| "grad_norm": 0.6095925983926321, |
| "learning_rate": 1.7212366715151263e-06, |
| "loss": 1.287, |
| "num_tokens": 70116810.0, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.8277511961722488, |
| "grad_norm": 0.6240251911568065, |
| "learning_rate": 1.702395866509612e-06, |
| "loss": 1.1325, |
| "num_tokens": 70301449.0, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.8301435406698564, |
| "grad_norm": 0.5784795095657023, |
| "learning_rate": 1.6837835672960834e-06, |
| "loss": 1.2059, |
| "num_tokens": 70492602.0, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.8325358851674641, |
| "grad_norm": 0.6504184788969882, |
| "learning_rate": 1.6654008937959498e-06, |
| "loss": 0.9334, |
| "num_tokens": 70656041.0, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.8349282296650717, |
| "grad_norm": 0.5702583567689546, |
| "learning_rate": 1.6472489521138016e-06, |
| "loss": 1.2118, |
| "num_tokens": 70905608.0, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.8373205741626795, |
| "grad_norm": 0.6274124275648116, |
| "learning_rate": 1.629328834470857e-06, |
| "loss": 1.1815, |
| "num_tokens": 71101191.0, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.8397129186602871, |
| "grad_norm": 0.6033829272863582, |
| "learning_rate": 1.611641619139238e-06, |
| "loss": 1.2807, |
| "num_tokens": 71291926.0, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.8421052631578947, |
| "grad_norm": 0.6273406564056667, |
| "learning_rate": 1.5941883703770968e-06, |
| "loss": 1.3962, |
| "num_tokens": 71533733.0, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.8444976076555024, |
| "grad_norm": 0.5312463864078979, |
| "learning_rate": 1.57697013836457e-06, |
| "loss": 1.3811, |
| "num_tokens": 71747066.0, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.84688995215311, |
| "grad_norm": 0.7082921095962508, |
| "learning_rate": 1.5599879591405917e-06, |
| "loss": 1.1701, |
| "num_tokens": 71900238.0, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.8492822966507177, |
| "grad_norm": 0.6134430059504471, |
| "learning_rate": 1.5432428545405554e-06, |
| "loss": 1.2446, |
| "num_tokens": 72085575.0, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.8516746411483254, |
| "grad_norm": 0.6204954269243784, |
| "learning_rate": 1.526735832134829e-06, |
| "loss": 1.1297, |
| "num_tokens": 72322159.0, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.854066985645933, |
| "grad_norm": 0.565525186855862, |
| "learning_rate": 1.5104678851681253e-06, |
| "loss": 1.4408, |
| "num_tokens": 72528384.0, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.8564593301435407, |
| "grad_norm": 0.5682317963462433, |
| "learning_rate": 1.4944399924997372e-06, |
| "loss": 1.3499, |
| "num_tokens": 72730257.0, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.8588516746411483, |
| "grad_norm": 0.5129599811585936, |
| "learning_rate": 1.4786531185446455e-06, |
| "loss": 1.2999, |
| "num_tokens": 72969626.0, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.861244019138756, |
| "grad_norm": 0.6444665971353251, |
| "learning_rate": 1.4631082132154806e-06, |
| "loss": 1.2785, |
| "num_tokens": 73149240.0, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.8636363636363636, |
| "grad_norm": 0.5547734172552142, |
| "learning_rate": 1.4478062118653703e-06, |
| "loss": 1.4096, |
| "num_tokens": 73363813.0, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.8660287081339713, |
| "grad_norm": 0.6371851858022577, |
| "learning_rate": 1.4327480352316581e-06, |
| "loss": 1.2693, |
| "num_tokens": 73551589.0, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.868421052631579, |
| "grad_norm": 0.5546864610023559, |
| "learning_rate": 1.417934589380498e-06, |
| "loss": 1.403, |
| "num_tokens": 73790910.0, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.8708133971291866, |
| "grad_norm": 0.6219072093620266, |
| "learning_rate": 1.4033667656523405e-06, |
| "loss": 1.2085, |
| "num_tokens": 73985624.0, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.8732057416267942, |
| "grad_norm": 0.6235130422407937, |
| "learning_rate": 1.389045440608296e-06, |
| "loss": 1.0986, |
| "num_tokens": 74148795.0, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.8755980861244019, |
| "grad_norm": 0.6006216529271874, |
| "learning_rate": 1.374971475977394e-06, |
| "loss": 1.3025, |
| "num_tokens": 74332841.0, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.8779904306220095, |
| "grad_norm": 0.6404509379440522, |
| "learning_rate": 1.361145718604731e-06, |
| "loss": 1.2766, |
| "num_tokens": 74529785.0, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.8803827751196173, |
| "grad_norm": 0.686225898948988, |
| "learning_rate": 1.3475690004005098e-06, |
| "loss": 1.2452, |
| "num_tokens": 74703039.0, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.8827751196172249, |
| "grad_norm": 0.571283865051148, |
| "learning_rate": 1.3342421382899936e-06, |
| "loss": 1.2626, |
| "num_tokens": 74883297.0, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.8851674641148325, |
| "grad_norm": 0.7382176987149066, |
| "learning_rate": 1.3211659341643412e-06, |
| "loss": 1.233, |
| "num_tokens": 75066196.0, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.8875598086124402, |
| "grad_norm": 0.6611644800778096, |
| "learning_rate": 1.308341174832359e-06, |
| "loss": 1.1876, |
| "num_tokens": 75232348.0, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.8899521531100478, |
| "grad_norm": 0.5130766984472136, |
| "learning_rate": 1.2957686319731623e-06, |
| "loss": 1.2034, |
| "num_tokens": 75458529.0, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.8923444976076556, |
| "grad_norm": 0.5684819034489474, |
| "learning_rate": 1.2834490620897342e-06, |
| "loss": 1.1584, |
| "num_tokens": 75661699.0, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.8947368421052632, |
| "grad_norm": 0.5422081420273679, |
| "learning_rate": 1.2713832064634127e-06, |
| "loss": 1.3424, |
| "num_tokens": 75916455.0, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.8971291866028708, |
| "grad_norm": 0.617788912813985, |
| "learning_rate": 1.259571791109285e-06, |
| "loss": 1.3227, |
| "num_tokens": 76100165.0, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.8995215311004785, |
| "grad_norm": 0.5559640103832523, |
| "learning_rate": 1.2480155267325039e-06, |
| "loss": 1.1962, |
| "num_tokens": 76326481.0, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.9019138755980861, |
| "grad_norm": 0.5930770090555496, |
| "learning_rate": 1.2367151086855187e-06, |
| "loss": 1.2261, |
| "num_tokens": 76482451.0, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.9043062200956937, |
| "grad_norm": 0.6461043468488586, |
| "learning_rate": 1.2256712169262415e-06, |
| "loss": 1.3309, |
| "num_tokens": 76692581.0, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.9066985645933014, |
| "grad_norm": 0.6295579626427625, |
| "learning_rate": 1.2148845159771311e-06, |
| "loss": 1.0403, |
| "num_tokens": 76869394.0, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.9090909090909091, |
| "grad_norm": 0.6317114832123317, |
| "learning_rate": 1.2043556548852065e-06, |
| "loss": 1.3877, |
| "num_tokens": 77057755.0, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.9114832535885168, |
| "grad_norm": 0.5646978093938921, |
| "learning_rate": 1.1940852671829938e-06, |
| "loss": 1.1932, |
| "num_tokens": 77279900.0, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.9138755980861244, |
| "grad_norm": 0.554856696245118, |
| "learning_rate": 1.184073970850408e-06, |
| "loss": 1.3931, |
| "num_tokens": 77474484.0, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.916267942583732, |
| "grad_norm": 0.5125502030227311, |
| "learning_rate": 1.174322368277565e-06, |
| "loss": 1.5298, |
| "num_tokens": 77739918.0, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.9186602870813397, |
| "grad_norm": 0.6245967453038902, |
| "learning_rate": 1.1648310462285386e-06, |
| "loss": 1.2966, |
| "num_tokens": 77876514.0, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.9210526315789473, |
| "grad_norm": 0.510131355499661, |
| "learning_rate": 1.1556005758060517e-06, |
| "loss": 1.3328, |
| "num_tokens": 78102264.0, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.9234449760765551, |
| "grad_norm": 0.5313647688128483, |
| "learning_rate": 1.146631512417113e-06, |
| "loss": 1.4715, |
| "num_tokens": 78336063.0, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.9258373205741627, |
| "grad_norm": 0.5521683071808053, |
| "learning_rate": 1.1379243957395987e-06, |
| "loss": 1.231, |
| "num_tokens": 78529771.0, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.9282296650717703, |
| "grad_norm": 0.7086439907733622, |
| "learning_rate": 1.1294797496897786e-06, |
| "loss": 1.1614, |
| "num_tokens": 78679587.0, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.930622009569378, |
| "grad_norm": 0.5802516939308205, |
| "learning_rate": 1.121298082390793e-06, |
| "loss": 1.3184, |
| "num_tokens": 78878694.0, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.9330143540669856, |
| "grad_norm": 0.5736760000180704, |
| "learning_rate": 1.113379886142075e-06, |
| "loss": 1.3361, |
| "num_tokens": 79073779.0, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.9354066985645934, |
| "grad_norm": 0.5432525121846279, |
| "learning_rate": 1.105725637389732e-06, |
| "loss": 1.2006, |
| "num_tokens": 79270679.0, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.937799043062201, |
| "grad_norm": 0.6631735593901804, |
| "learning_rate": 1.0983357966978747e-06, |
| "loss": 1.4223, |
| "num_tokens": 79451866.0, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.9401913875598086, |
| "grad_norm": 0.5833059272480773, |
| "learning_rate": 1.0912108087209075e-06, |
| "loss": 1.3483, |
| "num_tokens": 79629121.0, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.9425837320574163, |
| "grad_norm": 0.521975187666564, |
| "learning_rate": 1.084351102176769e-06, |
| "loss": 1.3392, |
| "num_tokens": 79864771.0, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.9449760765550239, |
| "grad_norm": 0.5983490163395889, |
| "learning_rate": 1.0777570898211406e-06, |
| "loss": 1.138, |
| "num_tokens": 80090837.0, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.9473684210526315, |
| "grad_norm": 0.6791397944349069, |
| "learning_rate": 1.0714291684226054e-06, |
| "loss": 1.1482, |
| "num_tokens": 80247684.0, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.9497607655502392, |
| "grad_norm": 0.7135484583851197, |
| "learning_rate": 1.0653677187387787e-06, |
| "loss": 1.178, |
| "num_tokens": 80412689.0, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.9521531100478469, |
| "grad_norm": 0.5467847425622099, |
| "learning_rate": 1.0595731054933937e-06, |
| "loss": 1.256, |
| "num_tokens": 80646002.0, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.9545454545454546, |
| "grad_norm": 0.649264408997545, |
| "learning_rate": 1.0540456773543596e-06, |
| "loss": 1.2587, |
| "num_tokens": 80846542.0, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.9569377990430622, |
| "grad_norm": 0.6408697258124432, |
| "learning_rate": 1.0487857669127782e-06, |
| "loss": 1.2991, |
| "num_tokens": 81039183.0, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.9593301435406698, |
| "grad_norm": 0.6125041198323196, |
| "learning_rate": 1.0437936906629336e-06, |
| "loss": 1.2721, |
| "num_tokens": 81228745.0, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.9617224880382775, |
| "grad_norm": 0.6078405250476319, |
| "learning_rate": 1.039069748983248e-06, |
| "loss": 1.3821, |
| "num_tokens": 81425019.0, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.9641148325358851, |
| "grad_norm": 0.5595011118119477, |
| "learning_rate": 1.0346142261182064e-06, |
| "loss": 1.0514, |
| "num_tokens": 81634634.0, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.9665071770334929, |
| "grad_norm": 0.4972617449675274, |
| "learning_rate": 1.0304273901612566e-06, |
| "loss": 1.341, |
| "num_tokens": 81852833.0, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.9688995215311005, |
| "grad_norm": 0.5059294656620951, |
| "learning_rate": 1.0265094930386741e-06, |
| "loss": 1.309, |
| "num_tokens": 82114913.0, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.9712918660287081, |
| "grad_norm": 0.5962308760747792, |
| "learning_rate": 1.0228607704944048e-06, |
| "loss": 1.3641, |
| "num_tokens": 82299779.0, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.9736842105263158, |
| "grad_norm": 0.6131816008639356, |
| "learning_rate": 1.0194814420758806e-06, |
| "loss": 1.1839, |
| "num_tokens": 82469216.0, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.9760765550239234, |
| "grad_norm": 0.6115969172998058, |
| "learning_rate": 1.0163717111208086e-06, |
| "loss": 1.3046, |
| "num_tokens": 82640081.0, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.9784688995215312, |
| "grad_norm": 0.5612129186921679, |
| "learning_rate": 1.0135317647449362e-06, |
| "loss": 1.4059, |
| "num_tokens": 82843981.0, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.9808612440191388, |
| "grad_norm": 0.587067444004318, |
| "learning_rate": 1.0109617738307914e-06, |
| "loss": 1.2261, |
| "num_tokens": 83020444.0, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.9832535885167464, |
| "grad_norm": 0.5594996533209011, |
| "learning_rate": 1.0086618930174011e-06, |
| "loss": 1.2042, |
| "num_tokens": 83232598.0, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.9856459330143541, |
| "grad_norm": 0.7071205597794052, |
| "learning_rate": 1.006632260690988e-06, |
| "loss": 1.2169, |
| "num_tokens": 83374569.0, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.9880382775119617, |
| "grad_norm": 0.6748664203646131, |
| "learning_rate": 1.0048729989766396e-06, |
| "loss": 1.1071, |
| "num_tokens": 83586085.0, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.9904306220095693, |
| "grad_norm": 0.5910387645740945, |
| "learning_rate": 1.0033842137309649e-06, |
| "loss": 1.4125, |
| "num_tokens": 83786689.0, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.992822966507177, |
| "grad_norm": 0.6362289532425913, |
| "learning_rate": 1.0021659945357202e-06, |
| "loss": 1.2213, |
| "num_tokens": 83967541.0, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.9952153110047847, |
| "grad_norm": 0.5429466958407293, |
| "learning_rate": 1.0012184146924225e-06, |
| "loss": 1.3142, |
| "num_tokens": 84191970.0, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.9976076555023924, |
| "grad_norm": 0.5637271505857214, |
| "learning_rate": 1.0005415312179367e-06, |
| "loss": 1.3643, |
| "num_tokens": 84430751.0, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5655593189952651, |
| "learning_rate": 1.0001353848410461e-06, |
| "loss": 1.2425, |
| "num_tokens": 84668558.0, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 0.7911388278007507, |
| "eval_num_tokens": 84668558.0, |
| "eval_runtime": 47.4555, |
| "eval_samples_per_second": 62.627, |
| "eval_steps_per_second": 7.839, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 418, |
| "total_flos": 179772395159552.0, |
| "train_loss": 1.3804568826581873, |
| "train_runtime": 1345.8555, |
| "train_samples_per_second": 19.872, |
| "train_steps_per_second": 0.311 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 418, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 179772395159552.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|