{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.996517107150174, "eval_steps": 500, "global_step": 760, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006556033599672198, "grad_norm": 6.308972265166019, "learning_rate": 1.0526315789473685e-06, "loss": 1.085, "step": 1 }, { "epoch": 0.013112067199344397, "grad_norm": 6.3669176951390085, "learning_rate": 2.105263157894737e-06, "loss": 1.0943, "step": 2 }, { "epoch": 0.019668100799016593, "grad_norm": 6.248172461223065, "learning_rate": 3.157894736842105e-06, "loss": 1.0945, "step": 3 }, { "epoch": 0.026224134398688793, "grad_norm": 5.720783796325484, "learning_rate": 4.210526315789474e-06, "loss": 1.0822, "step": 4 }, { "epoch": 0.03278016799836099, "grad_norm": 4.166580116763421, "learning_rate": 5.263157894736842e-06, "loss": 1.0376, "step": 5 }, { "epoch": 0.03933620159803319, "grad_norm": 2.5258613072829443, "learning_rate": 6.31578947368421e-06, "loss": 1.0088, "step": 6 }, { "epoch": 0.04589223519770539, "grad_norm": 3.241153982591279, "learning_rate": 7.368421052631579e-06, "loss": 0.9787, "step": 7 }, { "epoch": 0.05244826879737759, "grad_norm": 4.215997338140802, "learning_rate": 8.421052631578948e-06, "loss": 1.0021, "step": 8 }, { "epoch": 0.05900430239704978, "grad_norm": 3.7917272082574036, "learning_rate": 9.473684210526315e-06, "loss": 0.9883, "step": 9 }, { "epoch": 0.06556033599672198, "grad_norm": 3.1091596358257414, "learning_rate": 1.0526315789473684e-05, "loss": 0.9323, "step": 10 }, { "epoch": 0.07211636959639418, "grad_norm": 2.7854345292697746, "learning_rate": 1.1578947368421053e-05, "loss": 0.9234, "step": 11 }, { "epoch": 0.07867240319606637, "grad_norm": 1.8650709570608603, "learning_rate": 1.263157894736842e-05, "loss": 0.9023, "step": 12 }, { "epoch": 0.08522843679573858, "grad_norm": 1.574226758545731, "learning_rate": 1.3684210526315791e-05, "loss": 0.8693, "step": 13 }, { "epoch": 0.09178447039541078, "grad_norm": 1.4177768637857613, "learning_rate": 1.4736842105263159e-05, "loss": 0.8558, "step": 14 }, { "epoch": 0.09834050399508297, "grad_norm": 1.2125668795375901, "learning_rate": 1.578947368421053e-05, "loss": 0.8394, "step": 15 }, { "epoch": 0.10489653759475517, "grad_norm": 1.1398741474728522, "learning_rate": 1.6842105263157896e-05, "loss": 0.8373, "step": 16 }, { "epoch": 0.11145257119442738, "grad_norm": 1.0380532605315191, "learning_rate": 1.7894736842105264e-05, "loss": 0.8138, "step": 17 }, { "epoch": 0.11800860479409957, "grad_norm": 1.1944105144048436, "learning_rate": 1.894736842105263e-05, "loss": 0.8178, "step": 18 }, { "epoch": 0.12456463839377177, "grad_norm": 0.9889022037403388, "learning_rate": 2e-05, "loss": 0.803, "step": 19 }, { "epoch": 0.13112067199344396, "grad_norm": 0.7531044929728199, "learning_rate": 2.105263157894737e-05, "loss": 0.806, "step": 20 }, { "epoch": 0.13767670559311618, "grad_norm": 0.9451120842150798, "learning_rate": 2.210526315789474e-05, "loss": 0.7904, "step": 21 }, { "epoch": 0.14423273919278837, "grad_norm": 0.8762905130696029, "learning_rate": 2.3157894736842107e-05, "loss": 0.8001, "step": 22 }, { "epoch": 0.15078877279246056, "grad_norm": 0.6699562168478991, "learning_rate": 2.4210526315789474e-05, "loss": 0.7822, "step": 23 }, { "epoch": 0.15734480639213275, "grad_norm": 0.6276115227175743, "learning_rate": 2.526315789473684e-05, "loss": 0.7773, "step": 24 }, { "epoch": 0.16390083999180496, "grad_norm": 0.6290807862288751, "learning_rate": 2.6315789473684215e-05, "loss": 0.7783, "step": 25 }, { "epoch": 0.17045687359147715, "grad_norm": 0.5742564332284298, "learning_rate": 2.7368421052631583e-05, "loss": 0.7674, "step": 26 }, { "epoch": 0.17701290719114934, "grad_norm": 0.7412922803237626, "learning_rate": 2.842105263157895e-05, "loss": 0.7698, "step": 27 }, { "epoch": 0.18356894079082156, "grad_norm": 1.0547770655616728, "learning_rate": 2.9473684210526317e-05, "loss": 0.7655, "step": 28 }, { "epoch": 0.19012497439049375, "grad_norm": 1.5527104367024405, "learning_rate": 3.052631578947369e-05, "loss": 0.7707, "step": 29 }, { "epoch": 0.19668100799016594, "grad_norm": 0.7294937893627019, "learning_rate": 3.157894736842106e-05, "loss": 0.7578, "step": 30 }, { "epoch": 0.20323704158983816, "grad_norm": 0.9536771340964642, "learning_rate": 3.2631578947368426e-05, "loss": 0.748, "step": 31 }, { "epoch": 0.20979307518951035, "grad_norm": 1.6534518524100028, "learning_rate": 3.368421052631579e-05, "loss": 0.7588, "step": 32 }, { "epoch": 0.21634910878918254, "grad_norm": 0.6082911982013689, "learning_rate": 3.473684210526316e-05, "loss": 0.7424, "step": 33 }, { "epoch": 0.22290514238885475, "grad_norm": 1.4728558033024115, "learning_rate": 3.578947368421053e-05, "loss": 0.7423, "step": 34 }, { "epoch": 0.22946117598852694, "grad_norm": 0.8950252912362415, "learning_rate": 3.6842105263157895e-05, "loss": 0.7396, "step": 35 }, { "epoch": 0.23601720958819913, "grad_norm": 1.2202323688159324, "learning_rate": 3.789473684210526e-05, "loss": 0.7394, "step": 36 }, { "epoch": 0.24257324318787135, "grad_norm": 1.0375163505188818, "learning_rate": 3.8947368421052636e-05, "loss": 0.735, "step": 37 }, { "epoch": 0.24912927678754354, "grad_norm": 1.1817870027992188, "learning_rate": 4e-05, "loss": 0.7316, "step": 38 }, { "epoch": 0.25568531038721576, "grad_norm": 1.7472819001700346, "learning_rate": 4.105263157894738e-05, "loss": 0.7371, "step": 39 }, { "epoch": 0.2622413439868879, "grad_norm": 0.6997273968133122, "learning_rate": 4.210526315789474e-05, "loss": 0.7303, "step": 40 }, { "epoch": 0.26879737758656014, "grad_norm": 2.321753393397282, "learning_rate": 4.315789473684211e-05, "loss": 0.7458, "step": 41 }, { "epoch": 0.27535341118623236, "grad_norm": 1.366635958874213, "learning_rate": 4.421052631578948e-05, "loss": 0.7176, "step": 42 }, { "epoch": 0.2819094447859045, "grad_norm": 2.6212092693642552, "learning_rate": 4.5263157894736846e-05, "loss": 0.7315, "step": 43 }, { "epoch": 0.28846547838557673, "grad_norm": 2.5130277449596496, "learning_rate": 4.6315789473684214e-05, "loss": 0.753, "step": 44 }, { "epoch": 0.29502151198524895, "grad_norm": 1.6285728910727462, "learning_rate": 4.736842105263158e-05, "loss": 0.7297, "step": 45 }, { "epoch": 0.3015775455849211, "grad_norm": 2.0964313396830456, "learning_rate": 4.842105263157895e-05, "loss": 0.7265, "step": 46 }, { "epoch": 0.30813357918459333, "grad_norm": 2.0887773848236697, "learning_rate": 4.947368421052632e-05, "loss": 0.7239, "step": 47 }, { "epoch": 0.3146896127842655, "grad_norm": 1.1072655635857458, "learning_rate": 5.052631578947368e-05, "loss": 0.7193, "step": 48 }, { "epoch": 0.3212456463839377, "grad_norm": 1.4488623000707561, "learning_rate": 5.157894736842106e-05, "loss": 0.7203, "step": 49 }, { "epoch": 0.32780167998360993, "grad_norm": 1.2322683716693836, "learning_rate": 5.263157894736843e-05, "loss": 0.7238, "step": 50 }, { "epoch": 0.3343577135832821, "grad_norm": 1.38345500389746, "learning_rate": 5.368421052631579e-05, "loss": 0.7142, "step": 51 }, { "epoch": 0.3409137471829543, "grad_norm": 0.9094126438399451, "learning_rate": 5.4736842105263165e-05, "loss": 0.72, "step": 52 }, { "epoch": 0.3474697807826265, "grad_norm": 1.2022041950473747, "learning_rate": 5.5789473684210526e-05, "loss": 0.7075, "step": 53 }, { "epoch": 0.3540258143822987, "grad_norm": 1.6189179778780058, "learning_rate": 5.68421052631579e-05, "loss": 0.7105, "step": 54 }, { "epoch": 0.3605818479819709, "grad_norm": 1.2006446280190224, "learning_rate": 5.789473684210527e-05, "loss": 0.7151, "step": 55 }, { "epoch": 0.3671378815816431, "grad_norm": 1.6321920047861564, "learning_rate": 5.8947368421052634e-05, "loss": 0.7136, "step": 56 }, { "epoch": 0.3736939151813153, "grad_norm": 1.5287305591592921, "learning_rate": 6.000000000000001e-05, "loss": 0.7107, "step": 57 }, { "epoch": 0.3802499487809875, "grad_norm": 1.3257983915749183, "learning_rate": 6.105263157894738e-05, "loss": 0.7183, "step": 58 }, { "epoch": 0.3868059823806597, "grad_norm": 0.8173893638911733, "learning_rate": 6.210526315789474e-05, "loss": 0.7087, "step": 59 }, { "epoch": 0.3933620159803319, "grad_norm": 1.5027100216095572, "learning_rate": 6.315789473684212e-05, "loss": 0.7099, "step": 60 }, { "epoch": 0.3999180495800041, "grad_norm": 0.9118886385672123, "learning_rate": 6.421052631578948e-05, "loss": 0.7056, "step": 61 }, { "epoch": 0.4064740831796763, "grad_norm": 8.129014604966502, "learning_rate": 6.526315789473685e-05, "loss": 0.7359, "step": 62 }, { "epoch": 0.4130301167793485, "grad_norm": 3.8240665538108343, "learning_rate": 6.631578947368421e-05, "loss": 0.7738, "step": 63 }, { "epoch": 0.4195861503790207, "grad_norm": 3.710700934507178, "learning_rate": 6.736842105263159e-05, "loss": 0.7309, "step": 64 }, { "epoch": 0.4261421839786929, "grad_norm": 1.7575755137455045, "learning_rate": 6.842105263157895e-05, "loss": 0.7373, "step": 65 }, { "epoch": 0.4326982175783651, "grad_norm": 1.29443521888165, "learning_rate": 6.947368421052632e-05, "loss": 0.7149, "step": 66 }, { "epoch": 0.4392542511780373, "grad_norm": 1.9730328727947009, "learning_rate": 7.052631578947368e-05, "loss": 0.7309, "step": 67 }, { "epoch": 0.4458102847777095, "grad_norm": 1.5434748912530512, "learning_rate": 7.157894736842105e-05, "loss": 0.7307, "step": 68 }, { "epoch": 0.45236631837738167, "grad_norm": 1.3098144472083486, "learning_rate": 7.263157894736843e-05, "loss": 0.7226, "step": 69 }, { "epoch": 0.4589223519770539, "grad_norm": 1.4763764747327572, "learning_rate": 7.368421052631579e-05, "loss": 0.7122, "step": 70 }, { "epoch": 0.4654783855767261, "grad_norm": 1.782399664190018, "learning_rate": 7.473684210526316e-05, "loss": 0.7105, "step": 71 }, { "epoch": 0.47203441917639827, "grad_norm": 0.8645965374178922, "learning_rate": 7.578947368421052e-05, "loss": 0.7189, "step": 72 }, { "epoch": 0.4785904527760705, "grad_norm": 1.1783141500345207, "learning_rate": 7.68421052631579e-05, "loss": 0.7155, "step": 73 }, { "epoch": 0.4851464863757427, "grad_norm": 1.631405575969008, "learning_rate": 7.789473684210527e-05, "loss": 0.7163, "step": 74 }, { "epoch": 0.49170251997541486, "grad_norm": 1.799584170211315, "learning_rate": 7.894736842105263e-05, "loss": 0.724, "step": 75 }, { "epoch": 0.4982585535750871, "grad_norm": 131.28302876683892, "learning_rate": 8e-05, "loss": 0.9347, "step": 76 }, { "epoch": 0.5048145871747592, "grad_norm": 2.879295120261663, "learning_rate": 7.999957809295807e-05, "loss": 0.7474, "step": 77 }, { "epoch": 0.5113706207744315, "grad_norm": 0.9410659862896121, "learning_rate": 7.99983123807325e-05, "loss": 0.7124, "step": 78 }, { "epoch": 0.5179266543741037, "grad_norm": 2.4498983574935487, "learning_rate": 7.999620289002397e-05, "loss": 0.7294, "step": 79 }, { "epoch": 0.5244826879737758, "grad_norm": 1.3312297085268832, "learning_rate": 7.999324966533291e-05, "loss": 0.7155, "step": 80 }, { "epoch": 0.5310387215734481, "grad_norm": 1.9333142275612973, "learning_rate": 7.998945276895866e-05, "loss": 0.7249, "step": 81 }, { "epoch": 0.5375947551731203, "grad_norm": 1.5284340416931426, "learning_rate": 7.998481228099806e-05, "loss": 0.7199, "step": 82 }, { "epoch": 0.5441507887727924, "grad_norm": 1.5374661844560404, "learning_rate": 7.997932829934386e-05, "loss": 0.7174, "step": 83 }, { "epoch": 0.5507068223724647, "grad_norm": 1.099517635025565, "learning_rate": 7.997300093968255e-05, "loss": 0.7106, "step": 84 }, { "epoch": 0.5572628559721369, "grad_norm": 1.4642952447149862, "learning_rate": 7.996583033549204e-05, "loss": 0.7087, "step": 85 }, { "epoch": 0.563818889571809, "grad_norm": 1.4883363499473765, "learning_rate": 7.995781663803876e-05, "loss": 0.724, "step": 86 }, { "epoch": 0.5703749231714813, "grad_norm": 0.8588229153701384, "learning_rate": 7.994896001637443e-05, "loss": 0.7052, "step": 87 }, { "epoch": 0.5769309567711535, "grad_norm": 1.1329548393299773, "learning_rate": 7.993926065733265e-05, "loss": 0.7092, "step": 88 }, { "epoch": 0.5834869903708256, "grad_norm": 1.9381285339000986, "learning_rate": 7.99287187655248e-05, "loss": 0.7075, "step": 89 }, { "epoch": 0.5900430239704979, "grad_norm": 1.0298360769847832, "learning_rate": 7.991733456333579e-05, "loss": 0.7088, "step": 90 }, { "epoch": 0.5965990575701701, "grad_norm": 1.5913971706858665, "learning_rate": 7.990510829091938e-05, "loss": 0.7044, "step": 91 }, { "epoch": 0.6031550911698422, "grad_norm": 1.0752295835923937, "learning_rate": 7.98920402061931e-05, "loss": 0.698, "step": 92 }, { "epoch": 0.6097111247695144, "grad_norm": 1.3411827568812442, "learning_rate": 7.987813058483278e-05, "loss": 0.6897, "step": 93 }, { "epoch": 0.6162671583691867, "grad_norm": 1.0994663107424685, "learning_rate": 7.98633797202668e-05, "loss": 0.7009, "step": 94 }, { "epoch": 0.6228231919688588, "grad_norm": 1.5572209303699358, "learning_rate": 7.984778792366983e-05, "loss": 0.7001, "step": 95 }, { "epoch": 0.629379225568531, "grad_norm": 1.2622939681539915, "learning_rate": 7.98313555239563e-05, "loss": 0.7027, "step": 96 }, { "epoch": 0.6359352591682033, "grad_norm": 1.5287178198020728, "learning_rate": 7.98140828677735e-05, "loss": 0.6977, "step": 97 }, { "epoch": 0.6424912927678754, "grad_norm": 1.1853944350439, "learning_rate": 7.979597031949415e-05, "loss": 0.6943, "step": 98 }, { "epoch": 0.6490473263675476, "grad_norm": 1.1219402341610256, "learning_rate": 7.977701826120888e-05, "loss": 0.6984, "step": 99 }, { "epoch": 0.6556033599672199, "grad_norm": 1.8026221843535832, "learning_rate": 7.975722709271799e-05, "loss": 0.6955, "step": 100 }, { "epoch": 0.662159393566892, "grad_norm": 1.0073031848270164, "learning_rate": 7.973659723152317e-05, "loss": 0.6942, "step": 101 }, { "epoch": 0.6687154271665642, "grad_norm": 1.4709635304665, "learning_rate": 7.97151291128186e-05, "loss": 0.7024, "step": 102 }, { "epoch": 0.6752714607662365, "grad_norm": 1.5710635409318867, "learning_rate": 7.96928231894818e-05, "loss": 0.6917, "step": 103 }, { "epoch": 0.6818274943659086, "grad_norm": 0.8520728596618132, "learning_rate": 7.96696799320641e-05, "loss": 0.6821, "step": 104 }, { "epoch": 0.6883835279655808, "grad_norm": 1.3459049216152963, "learning_rate": 7.964569982878063e-05, "loss": 0.6916, "step": 105 }, { "epoch": 0.694939561565253, "grad_norm": 0.9826433336489447, "learning_rate": 7.962088338550013e-05, "loss": 0.6894, "step": 106 }, { "epoch": 0.7014955951649252, "grad_norm": 1.2896938337982984, "learning_rate": 7.959523112573422e-05, "loss": 0.6933, "step": 107 }, { "epoch": 0.7080516287645974, "grad_norm": 0.9867952374397386, "learning_rate": 7.956874359062632e-05, "loss": 0.697, "step": 108 }, { "epoch": 0.7146076623642696, "grad_norm": 1.2639647577953248, "learning_rate": 7.954142133894033e-05, "loss": 0.6894, "step": 109 }, { "epoch": 0.7211636959639418, "grad_norm": 1.357158506896498, "learning_rate": 7.951326494704878e-05, "loss": 0.691, "step": 110 }, { "epoch": 0.727719729563614, "grad_norm": 0.8936367138602801, "learning_rate": 7.948427500892065e-05, "loss": 0.6887, "step": 111 }, { "epoch": 0.7342757631632862, "grad_norm": 1.0046258611386971, "learning_rate": 7.94544521361089e-05, "loss": 0.6866, "step": 112 }, { "epoch": 0.7408317967629584, "grad_norm": 1.0742250060740215, "learning_rate": 7.942379695773753e-05, "loss": 0.6888, "step": 113 }, { "epoch": 0.7473878303626306, "grad_norm": 0.7284144457213086, "learning_rate": 7.939231012048833e-05, "loss": 0.6811, "step": 114 }, { "epoch": 0.7539438639623028, "grad_norm": 0.7052427321018496, "learning_rate": 7.93599922885872e-05, "loss": 0.6763, "step": 115 }, { "epoch": 0.760499897561975, "grad_norm": 0.798707901395817, "learning_rate": 7.932684414379021e-05, "loss": 0.692, "step": 116 }, { "epoch": 0.7670559311616472, "grad_norm": 1.5879889624779187, "learning_rate": 7.929286638536913e-05, "loss": 0.6908, "step": 117 }, { "epoch": 0.7736119647613194, "grad_norm": 0.8333519158322553, "learning_rate": 7.925805973009672e-05, "loss": 0.6734, "step": 118 }, { "epoch": 0.7801679983609916, "grad_norm": 1.0218779201091506, "learning_rate": 7.922242491223167e-05, "loss": 0.684, "step": 119 }, { "epoch": 0.7867240319606638, "grad_norm": 1.2082599250889219, "learning_rate": 7.918596268350296e-05, "loss": 0.6765, "step": 120 }, { "epoch": 0.793280065560336, "grad_norm": 1.179921751372486, "learning_rate": 7.914867381309418e-05, "loss": 0.6868, "step": 121 }, { "epoch": 0.7998360991600082, "grad_norm": 0.9486398114671882, "learning_rate": 7.911055908762718e-05, "loss": 0.6749, "step": 122 }, { "epoch": 0.8063921327596804, "grad_norm": 0.8000482601006661, "learning_rate": 7.90716193111455e-05, "loss": 0.6747, "step": 123 }, { "epoch": 0.8129481663593526, "grad_norm": 0.9517101245313929, "learning_rate": 7.903185530509743e-05, "loss": 0.6794, "step": 124 }, { "epoch": 0.8195041999590248, "grad_norm": 0.9986534633574224, "learning_rate": 7.899126790831869e-05, "loss": 0.6774, "step": 125 }, { "epoch": 0.826060233558697, "grad_norm": 1.1257859217497668, "learning_rate": 7.894985797701472e-05, "loss": 0.6793, "step": 126 }, { "epoch": 0.8326162671583692, "grad_norm": 1.0292173321571256, "learning_rate": 7.890762638474256e-05, "loss": 0.6826, "step": 127 }, { "epoch": 0.8391723007580414, "grad_norm": 0.7557871567857994, "learning_rate": 7.886457402239256e-05, "loss": 0.6792, "step": 128 }, { "epoch": 0.8457283343577136, "grad_norm": 1.10326991797423, "learning_rate": 7.882070179816944e-05, "loss": 0.6786, "step": 129 }, { "epoch": 0.8522843679573858, "grad_norm": 0.6776535159011668, "learning_rate": 7.877601063757323e-05, "loss": 0.6769, "step": 130 }, { "epoch": 0.858840401557058, "grad_norm": 0.8010200105501427, "learning_rate": 7.873050148337967e-05, "loss": 0.6748, "step": 131 }, { "epoch": 0.8653964351567301, "grad_norm": 0.9126529329815608, "learning_rate": 7.868417529562043e-05, "loss": 0.6632, "step": 132 }, { "epoch": 0.8719524687564024, "grad_norm": 1.145222649075524, "learning_rate": 7.863703305156273e-05, "loss": 0.6756, "step": 133 }, { "epoch": 0.8785085023560746, "grad_norm": 1.082669097422374, "learning_rate": 7.858907574568882e-05, "loss": 0.6765, "step": 134 }, { "epoch": 0.8850645359557467, "grad_norm": 1.0310987601522028, "learning_rate": 7.854030438967494e-05, "loss": 0.6738, "step": 135 }, { "epoch": 0.891620569555419, "grad_norm": 1.0160706638367318, "learning_rate": 7.849072001237001e-05, "loss": 0.6778, "step": 136 }, { "epoch": 0.8981766031550912, "grad_norm": 0.9374366244899177, "learning_rate": 7.844032365977396e-05, "loss": 0.6736, "step": 137 }, { "epoch": 0.9047326367547633, "grad_norm": 0.7008738978429749, "learning_rate": 7.838911639501557e-05, "loss": 0.6781, "step": 138 }, { "epoch": 0.9112886703544356, "grad_norm": 0.4532137724870558, "learning_rate": 7.833709929833012e-05, "loss": 0.6686, "step": 139 }, { "epoch": 0.9178447039541078, "grad_norm": 0.5139137753219752, "learning_rate": 7.828427346703657e-05, "loss": 0.6672, "step": 140 }, { "epoch": 0.9244007375537799, "grad_norm": 0.47106346244337904, "learning_rate": 7.823064001551445e-05, "loss": 0.6621, "step": 141 }, { "epoch": 0.9309567711534522, "grad_norm": 0.3844200151143341, "learning_rate": 7.81762000751803e-05, "loss": 0.6669, "step": 142 }, { "epoch": 0.9375128047531244, "grad_norm": 0.4844251889310036, "learning_rate": 7.812095479446383e-05, "loss": 0.6606, "step": 143 }, { "epoch": 0.9440688383527965, "grad_norm": 0.5614156709455322, "learning_rate": 7.806490533878368e-05, "loss": 0.6627, "step": 144 }, { "epoch": 0.9506248719524688, "grad_norm": 0.6295196907491479, "learning_rate": 7.800805289052286e-05, "loss": 0.6652, "step": 145 }, { "epoch": 0.957180905552141, "grad_norm": 0.7172605017331375, "learning_rate": 7.795039864900378e-05, "loss": 0.6632, "step": 146 }, { "epoch": 0.9637369391518131, "grad_norm": 0.9441139561703984, "learning_rate": 7.789194383046295e-05, "loss": 0.6681, "step": 147 }, { "epoch": 0.9702929727514854, "grad_norm": 1.3151350146737344, "learning_rate": 7.783268966802539e-05, "loss": 0.6703, "step": 148 }, { "epoch": 0.9768490063511576, "grad_norm": 0.5115168470062825, "learning_rate": 7.777263741167849e-05, "loss": 0.6647, "step": 149 }, { "epoch": 0.9834050399508297, "grad_norm": 1.0163356112728301, "learning_rate": 7.771178832824573e-05, "loss": 0.666, "step": 150 }, { "epoch": 0.9899610735505019, "grad_norm": 1.1732972717592203, "learning_rate": 7.765014370135999e-05, "loss": 0.6592, "step": 151 }, { "epoch": 0.9965171071501742, "grad_norm": 0.7851604401937097, "learning_rate": 7.758770483143634e-05, "loss": 0.6592, "step": 152 }, { "epoch": 1.0055316533497234, "grad_norm": 1.6420611019668498, "learning_rate": 7.752447303564475e-05, "loss": 1.1949, "step": 153 }, { "epoch": 1.0120876869493955, "grad_norm": 0.9808444668187579, "learning_rate": 7.74604496478822e-05, "loss": 0.6428, "step": 154 }, { "epoch": 1.0186437205490677, "grad_norm": 1.1426869091876994, "learning_rate": 7.73956360187446e-05, "loss": 0.6346, "step": 155 }, { "epoch": 1.02519975414874, "grad_norm": 1.147145143696543, "learning_rate": 7.733003351549829e-05, "loss": 0.6388, "step": 156 }, { "epoch": 1.0317557877484123, "grad_norm": 0.7817012939106086, "learning_rate": 7.726364352205117e-05, "loss": 0.6408, "step": 157 }, { "epoch": 1.0383118213480844, "grad_norm": 0.694581928412185, "learning_rate": 7.719646743892352e-05, "loss": 0.6284, "step": 158 }, { "epoch": 1.0448678549477566, "grad_norm": 0.6070626955744831, "learning_rate": 7.712850668321846e-05, "loss": 0.6405, "step": 159 }, { "epoch": 1.0514238885474287, "grad_norm": 0.6514623348225942, "learning_rate": 7.705976268859207e-05, "loss": 0.6395, "step": 160 }, { "epoch": 1.057979922147101, "grad_norm": 0.5619592930813068, "learning_rate": 7.699023690522315e-05, "loss": 0.629, "step": 161 }, { "epoch": 1.0645359557467733, "grad_norm": 0.5769129425178826, "learning_rate": 7.691993079978252e-05, "loss": 0.6381, "step": 162 }, { "epoch": 1.0710919893464454, "grad_norm": 0.8198780194642444, "learning_rate": 7.684884585540227e-05, "loss": 0.6363, "step": 163 }, { "epoch": 1.0776480229461176, "grad_norm": 0.7916477488592911, "learning_rate": 7.677698357164431e-05, "loss": 0.635, "step": 164 }, { "epoch": 1.0842040565457898, "grad_norm": 0.5931606825910369, "learning_rate": 7.670434546446886e-05, "loss": 0.6333, "step": 165 }, { "epoch": 1.090760090145462, "grad_norm": 0.5877139627697113, "learning_rate": 7.663093306620231e-05, "loss": 0.629, "step": 166 }, { "epoch": 1.097316123745134, "grad_norm": 0.852370483441162, "learning_rate": 7.655674792550507e-05, "loss": 0.6395, "step": 167 }, { "epoch": 1.1038721573448065, "grad_norm": 1.00327271618352, "learning_rate": 7.648179160733883e-05, "loss": 0.6478, "step": 168 }, { "epoch": 1.1104281909444786, "grad_norm": 1.0959509941262375, "learning_rate": 7.640606569293347e-05, "loss": 0.6415, "step": 169 }, { "epoch": 1.1169842245441508, "grad_norm": 0.7671153051839551, "learning_rate": 7.632957177975387e-05, "loss": 0.6401, "step": 170 }, { "epoch": 1.123540258143823, "grad_norm": 0.7544679819781906, "learning_rate": 7.625231148146601e-05, "loss": 0.6365, "step": 171 }, { "epoch": 1.1300962917434951, "grad_norm": 0.7948382138503196, "learning_rate": 7.61742864279031e-05, "loss": 0.6363, "step": 172 }, { "epoch": 1.1366523253431673, "grad_norm": 0.7047533773597212, "learning_rate": 7.609549826503115e-05, "loss": 0.6421, "step": 173 }, { "epoch": 1.1432083589428397, "grad_norm": 0.7374204351284568, "learning_rate": 7.601594865491414e-05, "loss": 0.6407, "step": 174 }, { "epoch": 1.1497643925425118, "grad_norm": 0.6911928835033512, "learning_rate": 7.593563927567916e-05, "loss": 0.625, "step": 175 }, { "epoch": 1.156320426142184, "grad_norm": 0.4803947112250499, "learning_rate": 7.585457182148081e-05, "loss": 0.628, "step": 176 }, { "epoch": 1.1628764597418562, "grad_norm": 0.3950061404414093, "learning_rate": 7.577274800246558e-05, "loss": 0.6357, "step": 177 }, { "epoch": 1.1694324933415283, "grad_norm": 0.4280389547638801, "learning_rate": 7.569016954473577e-05, "loss": 0.6434, "step": 178 }, { "epoch": 1.1759885269412005, "grad_norm": 0.5103954479058104, "learning_rate": 7.560683819031298e-05, "loss": 0.6325, "step": 179 }, { "epoch": 1.1825445605408729, "grad_norm": 0.48405163068225376, "learning_rate": 7.552275569710152e-05, "loss": 0.621, "step": 180 }, { "epoch": 1.189100594140545, "grad_norm": 0.49505546120752403, "learning_rate": 7.543792383885113e-05, "loss": 0.6335, "step": 181 }, { "epoch": 1.1956566277402172, "grad_norm": 0.5549014381017361, "learning_rate": 7.535234440511979e-05, "loss": 0.6401, "step": 182 }, { "epoch": 1.2022126613398894, "grad_norm": 0.5679718866243005, "learning_rate": 7.526601920123574e-05, "loss": 0.6339, "step": 183 }, { "epoch": 1.2087686949395615, "grad_norm": 0.5837881333403254, "learning_rate": 7.517895004825956e-05, "loss": 0.6201, "step": 184 }, { "epoch": 1.2153247285392337, "grad_norm": 0.5275232093842687, "learning_rate": 7.509113878294572e-05, "loss": 0.6313, "step": 185 }, { "epoch": 1.2218807621389058, "grad_norm": 0.4187731695243854, "learning_rate": 7.500258725770375e-05, "loss": 0.6297, "step": 186 }, { "epoch": 1.2284367957385782, "grad_norm": 0.3496822319709094, "learning_rate": 7.491329734055926e-05, "loss": 0.6341, "step": 187 }, { "epoch": 1.2349928293382504, "grad_norm": 0.2892298691529975, "learning_rate": 7.48232709151145e-05, "loss": 0.6289, "step": 188 }, { "epoch": 1.2415488629379225, "grad_norm": 0.3970392684873568, "learning_rate": 7.473250988050861e-05, "loss": 0.6327, "step": 189 }, { "epoch": 1.2481048965375947, "grad_norm": 0.5366046878238813, "learning_rate": 7.464101615137756e-05, "loss": 0.6234, "step": 190 }, { "epoch": 1.2546609301372669, "grad_norm": 0.6365895325361283, "learning_rate": 7.454879165781379e-05, "loss": 0.6322, "step": 191 }, { "epoch": 1.2612169637369393, "grad_norm": 0.6390242865748291, "learning_rate": 7.445583834532546e-05, "loss": 0.6252, "step": 192 }, { "epoch": 1.2677729973366114, "grad_norm": 0.7130289120451936, "learning_rate": 7.436215817479541e-05, "loss": 0.6369, "step": 193 }, { "epoch": 1.2743290309362836, "grad_norm": 0.8001337986119181, "learning_rate": 7.426775312243986e-05, "loss": 0.6341, "step": 194 }, { "epoch": 1.2808850645359557, "grad_norm": 0.9421645194421999, "learning_rate": 7.41726251797666e-05, "loss": 0.629, "step": 195 }, { "epoch": 1.287441098135628, "grad_norm": 1.0651209082990578, "learning_rate": 7.407677635353308e-05, "loss": 0.6274, "step": 196 }, { "epoch": 1.2939971317353, "grad_norm": 0.8478545298742448, "learning_rate": 7.398020866570407e-05, "loss": 0.6373, "step": 197 }, { "epoch": 1.3005531653349722, "grad_norm": 0.5323333743534195, "learning_rate": 7.388292415340888e-05, "loss": 0.631, "step": 198 }, { "epoch": 1.3071091989346446, "grad_norm": 0.37990592076356283, "learning_rate": 7.37849248688986e-05, "loss": 0.631, "step": 199 }, { "epoch": 1.3136652325343168, "grad_norm": 0.4732271711472085, "learning_rate": 7.368621287950264e-05, "loss": 0.6328, "step": 200 }, { "epoch": 1.320221266133989, "grad_norm": 0.5259491499827896, "learning_rate": 7.358679026758515e-05, "loss": 0.6283, "step": 201 }, { "epoch": 1.326777299733661, "grad_norm": 0.5000279931190099, "learning_rate": 7.348665913050115e-05, "loss": 0.6208, "step": 202 }, { "epoch": 1.3333333333333333, "grad_norm": 0.3912490196632365, "learning_rate": 7.338582158055224e-05, "loss": 0.6251, "step": 203 }, { "epoch": 1.3398893669330056, "grad_norm": 0.38905604815288286, "learning_rate": 7.328427974494201e-05, "loss": 0.6179, "step": 204 }, { "epoch": 1.3464454005326778, "grad_norm": 0.4906878706390755, "learning_rate": 7.318203576573126e-05, "loss": 0.6271, "step": 205 }, { "epoch": 1.35300143413235, "grad_norm": 0.4146373419913017, "learning_rate": 7.307909179979274e-05, "loss": 0.6256, "step": 206 }, { "epoch": 1.3595574677320221, "grad_norm": 0.29246282870192875, "learning_rate": 7.297545001876563e-05, "loss": 0.6219, "step": 207 }, { "epoch": 1.3661135013316943, "grad_norm": 0.27459922699544653, "learning_rate": 7.28711126090098e-05, "loss": 0.6346, "step": 208 }, { "epoch": 1.3726695349313665, "grad_norm": 0.3594793278431249, "learning_rate": 7.276608177155968e-05, "loss": 0.6234, "step": 209 }, { "epoch": 1.3792255685310386, "grad_norm": 0.46199653506152194, "learning_rate": 7.266035972207773e-05, "loss": 0.6328, "step": 210 }, { "epoch": 1.385781602130711, "grad_norm": 0.5084391547285126, "learning_rate": 7.25539486908078e-05, "loss": 0.6287, "step": 211 }, { "epoch": 1.3923376357303832, "grad_norm": 0.4970090588981773, "learning_rate": 7.24468509225281e-05, "loss": 0.6338, "step": 212 }, { "epoch": 1.3988936693300553, "grad_norm": 0.4297079110960391, "learning_rate": 7.233906867650373e-05, "loss": 0.6246, "step": 213 }, { "epoch": 1.4054497029297275, "grad_norm": 0.4277415770961674, "learning_rate": 7.223060422643914e-05, "loss": 0.6235, "step": 214 }, { "epoch": 1.4120057365293996, "grad_norm": 0.5211718034258063, "learning_rate": 7.212145986043007e-05, "loss": 0.626, "step": 215 }, { "epoch": 1.418561770129072, "grad_norm": 0.7113602181910502, "learning_rate": 7.201163788091536e-05, "loss": 0.626, "step": 216 }, { "epoch": 1.4251178037287442, "grad_norm": 0.8702986528721038, "learning_rate": 7.190114060462837e-05, "loss": 0.6285, "step": 217 }, { "epoch": 1.4316738373284164, "grad_norm": 0.88781618603766, "learning_rate": 7.178997036254799e-05, "loss": 0.625, "step": 218 }, { "epoch": 1.4382298709280885, "grad_norm": 0.8543987595137604, "learning_rate": 7.167812949984966e-05, "loss": 0.6369, "step": 219 }, { "epoch": 1.4447859045277607, "grad_norm": 0.6591694598278867, "learning_rate": 7.156562037585576e-05, "loss": 0.6309, "step": 220 }, { "epoch": 1.4513419381274328, "grad_norm": 0.435460183285974, "learning_rate": 7.145244536398584e-05, "loss": 0.6337, "step": 221 }, { "epoch": 1.457897971727105, "grad_norm": 0.5394064377655016, "learning_rate": 7.133860685170665e-05, "loss": 0.6272, "step": 222 }, { "epoch": 1.4644540053267772, "grad_norm": 0.6560512889140389, "learning_rate": 7.12241072404817e-05, "loss": 0.6317, "step": 223 }, { "epoch": 1.4710100389264495, "grad_norm": 0.5108872994150431, "learning_rate": 7.110894894572056e-05, "loss": 0.6266, "step": 224 }, { "epoch": 1.4775660725261217, "grad_norm": 0.4508746314491801, "learning_rate": 7.099313439672806e-05, "loss": 0.6222, "step": 225 }, { "epoch": 1.4841221061257939, "grad_norm": 0.5932762714408137, "learning_rate": 7.087666603665284e-05, "loss": 0.6174, "step": 226 }, { "epoch": 1.490678139725466, "grad_norm": 0.5703671444864514, "learning_rate": 7.0759546322436e-05, "loss": 0.6286, "step": 227 }, { "epoch": 1.4972341733251384, "grad_norm": 0.4219759927229514, "learning_rate": 7.064177772475912e-05, "loss": 0.621, "step": 228 }, { "epoch": 1.5037902069248106, "grad_norm": 0.3899488374409842, "learning_rate": 7.052336272799227e-05, "loss": 0.6368, "step": 229 }, { "epoch": 1.5103462405244827, "grad_norm": 0.3352814743100072, "learning_rate": 7.040430383014146e-05, "loss": 0.6273, "step": 230 }, { "epoch": 1.516902274124155, "grad_norm": 0.49330992250119887, "learning_rate": 7.02846035427961e-05, "loss": 0.6268, "step": 231 }, { "epoch": 1.523458307723827, "grad_norm": 0.7275390777504656, "learning_rate": 7.016426439107586e-05, "loss": 0.6198, "step": 232 }, { "epoch": 1.5300143413234992, "grad_norm": 0.5944986948743242, "learning_rate": 7.004328891357753e-05, "loss": 0.6321, "step": 233 }, { "epoch": 1.5365703749231714, "grad_norm": 0.40136071416301794, "learning_rate": 6.992167966232143e-05, "loss": 0.6205, "step": 234 }, { "epoch": 1.5431264085228436, "grad_norm": 0.42683320761735183, "learning_rate": 6.979943920269749e-05, "loss": 0.6282, "step": 235 }, { "epoch": 1.5496824421225157, "grad_norm": 0.37233786072575653, "learning_rate": 6.967657011341126e-05, "loss": 0.6216, "step": 236 }, { "epoch": 1.556238475722188, "grad_norm": 0.367274159532209, "learning_rate": 6.955307498642948e-05, "loss": 0.6224, "step": 237 }, { "epoch": 1.5627945093218603, "grad_norm": 0.3660447730807658, "learning_rate": 6.942895642692527e-05, "loss": 0.6202, "step": 238 }, { "epoch": 1.5693505429215324, "grad_norm": 0.38882305238122433, "learning_rate": 6.930421705322339e-05, "loss": 0.6195, "step": 239 }, { "epoch": 1.5759065765212048, "grad_norm": 0.5684648344288802, "learning_rate": 6.917885949674483e-05, "loss": 0.6228, "step": 240 }, { "epoch": 1.582462610120877, "grad_norm": 0.7431328174808526, "learning_rate": 6.905288640195141e-05, "loss": 0.626, "step": 241 }, { "epoch": 1.5890186437205491, "grad_norm": 0.7410518651663662, "learning_rate": 6.892630042628988e-05, "loss": 0.6248, "step": 242 }, { "epoch": 1.5955746773202213, "grad_norm": 0.6219837555078648, "learning_rate": 6.879910424013599e-05, "loss": 0.6295, "step": 243 }, { "epoch": 1.6021307109198935, "grad_norm": 0.5396699034196659, "learning_rate": 6.867130052673806e-05, "loss": 0.6231, "step": 244 }, { "epoch": 1.6086867445195656, "grad_norm": 0.4531223244262461, "learning_rate": 6.854289198216042e-05, "loss": 0.6346, "step": 245 }, { "epoch": 1.6152427781192378, "grad_norm": 0.3242170520225258, "learning_rate": 6.841388131522656e-05, "loss": 0.6292, "step": 246 }, { "epoch": 1.62179881171891, "grad_norm": 0.34710026048248854, "learning_rate": 6.828427124746191e-05, "loss": 0.6153, "step": 247 }, { "epoch": 1.628354845318582, "grad_norm": 0.44267957645174644, "learning_rate": 6.815406451303647e-05, "loss": 0.6205, "step": 248 }, { "epoch": 1.6349108789182545, "grad_norm": 0.5545408679102592, "learning_rate": 6.802326385870715e-05, "loss": 0.621, "step": 249 }, { "epoch": 1.6414669125179266, "grad_norm": 0.5443247431153629, "learning_rate": 6.789187204375981e-05, "loss": 0.622, "step": 250 }, { "epoch": 1.6480229461175988, "grad_norm": 0.4301197066284, "learning_rate": 6.775989183995108e-05, "loss": 0.6135, "step": 251 }, { "epoch": 1.6545789797172712, "grad_norm": 0.396698152583572, "learning_rate": 6.762732603144978e-05, "loss": 0.6216, "step": 252 }, { "epoch": 1.6611350133169434, "grad_norm": 0.3515261284616528, "learning_rate": 6.749417741477836e-05, "loss": 0.6187, "step": 253 }, { "epoch": 1.6676910469166155, "grad_norm": 0.35858581397311556, "learning_rate": 6.736044879875373e-05, "loss": 0.6185, "step": 254 }, { "epoch": 1.6742470805162877, "grad_norm": 0.34392278311565416, "learning_rate": 6.722614300442815e-05, "loss": 0.6154, "step": 255 }, { "epoch": 1.6808031141159598, "grad_norm": 0.36707892916862706, "learning_rate": 6.709126286502965e-05, "loss": 0.6179, "step": 256 }, { "epoch": 1.687359147715632, "grad_norm": 0.441219103876971, "learning_rate": 6.695581122590225e-05, "loss": 0.6227, "step": 257 }, { "epoch": 1.6939151813153042, "grad_norm": 0.3994835806954666, "learning_rate": 6.681979094444596e-05, "loss": 0.6192, "step": 258 }, { "epoch": 1.7004712149149763, "grad_norm": 0.433716188444932, "learning_rate": 6.668320489005654e-05, "loss": 0.622, "step": 259 }, { "epoch": 1.7070272485146485, "grad_norm": 0.47283003533467305, "learning_rate": 6.654605594406486e-05, "loss": 0.6253, "step": 260 }, { "epoch": 1.7135832821143209, "grad_norm": 0.49013239605830206, "learning_rate": 6.640834699967626e-05, "loss": 0.6214, "step": 261 }, { "epoch": 1.720139315713993, "grad_norm": 0.49324672918814405, "learning_rate": 6.627008096190938e-05, "loss": 0.6182, "step": 262 }, { "epoch": 1.7266953493136652, "grad_norm": 0.5169572714827575, "learning_rate": 6.6131260747535e-05, "loss": 0.6259, "step": 263 }, { "epoch": 1.7332513829133376, "grad_norm": 0.5873107078500235, "learning_rate": 6.59918892850144e-05, "loss": 0.6235, "step": 264 }, { "epoch": 1.7398074165130097, "grad_norm": 0.5020487886809576, "learning_rate": 6.585196951443763e-05, "loss": 0.6221, "step": 265 }, { "epoch": 1.746363450112682, "grad_norm": 0.329791614699069, "learning_rate": 6.571150438746157e-05, "loss": 0.6265, "step": 266 }, { "epoch": 1.752919483712354, "grad_norm": 0.2861139606661766, "learning_rate": 6.557049686724751e-05, "loss": 0.624, "step": 267 }, { "epoch": 1.7594755173120262, "grad_norm": 0.46750896287166216, "learning_rate": 6.542894992839873e-05, "loss": 0.6241, "step": 268 }, { "epoch": 1.7660315509116984, "grad_norm": 0.5684934196993104, "learning_rate": 6.528686655689774e-05, "loss": 0.6152, "step": 269 }, { "epoch": 1.7725875845113706, "grad_norm": 0.48490900499130857, "learning_rate": 6.514424975004329e-05, "loss": 0.6111, "step": 270 }, { "epoch": 1.7791436181110427, "grad_norm": 0.5506105973703118, "learning_rate": 6.500110251638715e-05, "loss": 0.6152, "step": 271 }, { "epoch": 1.7856996517107149, "grad_norm": 0.4539294761333573, "learning_rate": 6.48574278756706e-05, "loss": 0.6273, "step": 272 }, { "epoch": 1.7922556853103873, "grad_norm": 0.34705648209401146, "learning_rate": 6.471322885876077e-05, "loss": 0.6204, "step": 273 }, { "epoch": 1.7988117189100594, "grad_norm": 0.2427469168923526, "learning_rate": 6.456850850758673e-05, "loss": 0.6214, "step": 274 }, { "epoch": 1.8053677525097316, "grad_norm": 0.3004384124304078, "learning_rate": 6.44232698750752e-05, "loss": 0.6135, "step": 275 }, { "epoch": 1.811923786109404, "grad_norm": 0.29782974812920376, "learning_rate": 6.427751602508628e-05, "loss": 0.6104, "step": 276 }, { "epoch": 1.8184798197090761, "grad_norm": 0.29267694689244445, "learning_rate": 6.413125003234876e-05, "loss": 0.624, "step": 277 }, { "epoch": 1.8250358533087483, "grad_norm": 0.3462332305273243, "learning_rate": 6.398447498239527e-05, "loss": 0.612, "step": 278 }, { "epoch": 1.8315918869084205, "grad_norm": 0.3949978565357491, "learning_rate": 6.383719397149715e-05, "loss": 0.622, "step": 279 }, { "epoch": 1.8381479205080926, "grad_norm": 0.40536973102729335, "learning_rate": 6.368941010659921e-05, "loss": 0.6118, "step": 280 }, { "epoch": 1.8447039541077648, "grad_norm": 0.3621578404050215, "learning_rate": 6.354112650525407e-05, "loss": 0.6149, "step": 281 }, { "epoch": 1.851259987707437, "grad_norm": 0.3445107311397209, "learning_rate": 6.339234629555655e-05, "loss": 0.6196, "step": 282 }, { "epoch": 1.857816021307109, "grad_norm": 0.30605392165712425, "learning_rate": 6.324307261607754e-05, "loss": 0.6239, "step": 283 }, { "epoch": 1.8643720549067813, "grad_norm": 0.3080701040643038, "learning_rate": 6.309330861579786e-05, "loss": 0.629, "step": 284 }, { "epoch": 1.8709280885064536, "grad_norm": 0.3349937492345958, "learning_rate": 6.294305745404185e-05, "loss": 0.6224, "step": 285 }, { "epoch": 1.8774841221061258, "grad_norm": 0.3396848816828842, "learning_rate": 6.279232230041065e-05, "loss": 0.6182, "step": 286 }, { "epoch": 1.884040155705798, "grad_norm": 0.3410093342010557, "learning_rate": 6.26411063347154e-05, "loss": 0.6188, "step": 287 }, { "epoch": 1.8905961893054704, "grad_norm": 0.314163527323835, "learning_rate": 6.248941274691017e-05, "loss": 0.6169, "step": 288 }, { "epoch": 1.8971522229051425, "grad_norm": 0.3976946162007582, "learning_rate": 6.233724473702457e-05, "loss": 0.6195, "step": 289 }, { "epoch": 1.9037082565048147, "grad_norm": 0.5601353024746372, "learning_rate": 6.218460551509636e-05, "loss": 0.6206, "step": 290 }, { "epoch": 1.9102642901044868, "grad_norm": 0.6132983016524085, "learning_rate": 6.203149830110367e-05, "loss": 0.6138, "step": 291 }, { "epoch": 1.916820323704159, "grad_norm": 0.5021407478552538, "learning_rate": 6.18779263248971e-05, "loss": 0.6219, "step": 292 }, { "epoch": 1.9233763573038312, "grad_norm": 0.3890448472238098, "learning_rate": 6.172389282613151e-05, "loss": 0.6187, "step": 293 }, { "epoch": 1.9299323909035033, "grad_norm": 0.3361740572901459, "learning_rate": 6.156940105419785e-05, "loss": 0.6218, "step": 294 }, { "epoch": 1.9364884245031755, "grad_norm": 0.2907876046977178, "learning_rate": 6.141445426815443e-05, "loss": 0.6166, "step": 295 }, { "epoch": 1.9430444581028476, "grad_norm": 0.2632987500846716, "learning_rate": 6.125905573665824e-05, "loss": 0.6232, "step": 296 }, { "epoch": 1.9496004917025198, "grad_norm": 0.32120831457165155, "learning_rate": 6.110320873789604e-05, "loss": 0.6242, "step": 297 }, { "epoch": 1.9561565253021922, "grad_norm": 0.32170026271883145, "learning_rate": 6.094691655951512e-05, "loss": 0.6094, "step": 298 }, { "epoch": 1.9627125589018644, "grad_norm": 0.27126547953075625, "learning_rate": 6.079018249855402e-05, "loss": 0.622, "step": 299 }, { "epoch": 1.9692685925015365, "grad_norm": 0.25823866111956445, "learning_rate": 6.063300986137297e-05, "loss": 0.6195, "step": 300 }, { "epoch": 1.975824626101209, "grad_norm": 0.2777978831118673, "learning_rate": 6.047540196358405e-05, "loss": 0.6188, "step": 301 }, { "epoch": 1.982380659700881, "grad_norm": 0.3097245776135163, "learning_rate": 6.0317362129981375e-05, "loss": 0.6161, "step": 302 }, { "epoch": 1.9889366933005532, "grad_norm": 0.3108082557836966, "learning_rate": 6.015889369447088e-05, "loss": 0.6166, "step": 303 }, { "epoch": 1.9954927269002254, "grad_norm": 0.2659644339645489, "learning_rate": 6.000000000000001e-05, "loss": 0.6143, "step": 304 }, { "epoch": 2.006556033599672, "grad_norm": 0.3001511379848704, "learning_rate": 5.9840684398487186e-05, "loss": 0.5867, "step": 305 }, { "epoch": 2.0131120671993443, "grad_norm": 0.3404835526439102, "learning_rate": 5.968095025075114e-05, "loss": 0.5745, "step": 306 }, { "epoch": 2.0196681007990165, "grad_norm": 0.5738437895797627, "learning_rate": 5.952080092643993e-05, "loss": 0.5828, "step": 307 }, { "epoch": 2.0262241343986886, "grad_norm": 0.7841436327218428, "learning_rate": 5.936023980395997e-05, "loss": 0.583, "step": 308 }, { "epoch": 2.032780167998361, "grad_norm": 0.7382363346356913, "learning_rate": 5.919927027040463e-05, "loss": 0.5795, "step": 309 }, { "epoch": 2.039336201598033, "grad_norm": 0.6545848867535631, "learning_rate": 5.903789572148295e-05, "loss": 0.5835, "step": 310 }, { "epoch": 2.0458922351977056, "grad_norm": 0.516762214742333, "learning_rate": 5.887611956144782e-05, "loss": 0.5787, "step": 311 }, { "epoch": 2.0524482687973777, "grad_norm": 0.5327353778668117, "learning_rate": 5.871394520302432e-05, "loss": 0.5742, "step": 312 }, { "epoch": 2.05900430239705, "grad_norm": 0.5080742962254841, "learning_rate": 5.8551376067337626e-05, "loss": 0.5737, "step": 313 }, { "epoch": 2.065560335996722, "grad_norm": 0.5649012551531187, "learning_rate": 5.838841558384091e-05, "loss": 0.5764, "step": 314 }, { "epoch": 2.0721163695963942, "grad_norm": 0.5820350754778492, "learning_rate": 5.8225067190242925e-05, "loss": 0.5716, "step": 315 }, { "epoch": 2.0786724031960664, "grad_norm": 0.46076842055118317, "learning_rate": 5.806133433243558e-05, "loss": 0.5753, "step": 316 }, { "epoch": 2.0852284367957385, "grad_norm": 0.47933858405340773, "learning_rate": 5.789722046442114e-05, "loss": 0.575, "step": 317 }, { "epoch": 2.0917844703954107, "grad_norm": 0.5168151561479127, "learning_rate": 5.7732729048239444e-05, "loss": 0.5749, "step": 318 }, { "epoch": 2.098340503995083, "grad_norm": 0.40840899989073715, "learning_rate": 5.756786355389482e-05, "loss": 0.5802, "step": 319 }, { "epoch": 2.104896537594755, "grad_norm": 0.44997063706673934, "learning_rate": 5.740262745928293e-05, "loss": 0.5717, "step": 320 }, { "epoch": 2.111452571194427, "grad_norm": 0.5369629713906064, "learning_rate": 5.723702425011738e-05, "loss": 0.5751, "step": 321 }, { "epoch": 2.1180086047940994, "grad_norm": 0.4475818760429054, "learning_rate": 5.707105741985615e-05, "loss": 0.5765, "step": 322 }, { "epoch": 2.124564638393772, "grad_norm": 0.30932774384801376, "learning_rate": 5.6904730469627985e-05, "loss": 0.5785, "step": 323 }, { "epoch": 2.131120671993444, "grad_norm": 0.33040348186144713, "learning_rate": 5.673804690815845e-05, "loss": 0.5768, "step": 324 }, { "epoch": 2.1376767055931163, "grad_norm": 0.3576099667078179, "learning_rate": 5.6571010251695954e-05, "loss": 0.5815, "step": 325 }, { "epoch": 2.1442327391927885, "grad_norm": 0.4237003547581017, "learning_rate": 5.6403624023937614e-05, "loss": 0.5747, "step": 326 }, { "epoch": 2.1507887727924606, "grad_norm": 0.4700100373051473, "learning_rate": 5.62358917559548e-05, "loss": 0.5751, "step": 327 }, { "epoch": 2.1573448063921328, "grad_norm": 0.39617955284785017, "learning_rate": 5.606781698611879e-05, "loss": 0.5798, "step": 328 }, { "epoch": 2.163900839991805, "grad_norm": 0.28548306735776896, "learning_rate": 5.5899403260026006e-05, "loss": 0.5724, "step": 329 }, { "epoch": 2.170456873591477, "grad_norm": 0.3315286342160672, "learning_rate": 5.573065413042333e-05, "loss": 0.5721, "step": 330 }, { "epoch": 2.1770129071911493, "grad_norm": 0.3343441243180511, "learning_rate": 5.556157315713305e-05, "loss": 0.5783, "step": 331 }, { "epoch": 2.1835689407908214, "grad_norm": 0.29129418213987884, "learning_rate": 5.5392163906977835e-05, "loss": 0.577, "step": 332 }, { "epoch": 2.1901249743904936, "grad_norm": 0.25286013518278194, "learning_rate": 5.522242995370545e-05, "loss": 0.5698, "step": 333 }, { "epoch": 2.1966810079901657, "grad_norm": 0.5001313540973934, "learning_rate": 5.505237487791343e-05, "loss": 0.5915, "step": 334 }, { "epoch": 2.2032370415898384, "grad_norm": 0.2322461169124981, "learning_rate": 5.488200226697345e-05, "loss": 0.5707, "step": 335 }, { "epoch": 2.2097930751895105, "grad_norm": 0.24729941417632514, "learning_rate": 5.471131571495574e-05, "loss": 0.5688, "step": 336 }, { "epoch": 2.2163491087891827, "grad_norm": 0.25095902176236934, "learning_rate": 5.454031882255319e-05, "loss": 0.5804, "step": 337 }, { "epoch": 2.222905142388855, "grad_norm": 0.26393993849887953, "learning_rate": 5.4369015197005506e-05, "loss": 0.5741, "step": 338 }, { "epoch": 2.229461175988527, "grad_norm": 0.32581537912637687, "learning_rate": 5.419740845202292e-05, "loss": 0.5841, "step": 339 }, { "epoch": 2.236017209588199, "grad_norm": 0.2928430349319568, "learning_rate": 5.4025502207710184e-05, "loss": 0.5763, "step": 340 }, { "epoch": 2.2425732431878713, "grad_norm": 0.28344192622339104, "learning_rate": 5.385330009049003e-05, "loss": 0.5748, "step": 341 }, { "epoch": 2.2491292767875435, "grad_norm": 0.2108350778995995, "learning_rate": 5.368080573302676e-05, "loss": 0.5677, "step": 342 }, { "epoch": 2.2556853103872156, "grad_norm": 0.2654969873255356, "learning_rate": 5.3508022774149574e-05, "loss": 0.5759, "step": 343 }, { "epoch": 2.262241343986888, "grad_norm": 0.26893594961047856, "learning_rate": 5.333495485877583e-05, "loss": 0.5713, "step": 344 }, { "epoch": 2.26879737758656, "grad_norm": 0.2359706290249212, "learning_rate": 5.3161605637834135e-05, "loss": 0.5826, "step": 345 }, { "epoch": 2.2753534111862326, "grad_norm": 0.2001826448668912, "learning_rate": 5.298797876818735e-05, "loss": 0.5828, "step": 346 }, { "epoch": 2.2819094447859047, "grad_norm": 0.18687800048623612, "learning_rate": 5.2814077912555415e-05, "loss": 0.5674, "step": 347 }, { "epoch": 2.288465478385577, "grad_norm": 0.18869821797720956, "learning_rate": 5.263990673943811e-05, "loss": 0.5795, "step": 348 }, { "epoch": 2.295021511985249, "grad_norm": 0.18002946746844742, "learning_rate": 5.246546892303766e-05, "loss": 0.5766, "step": 349 }, { "epoch": 2.3015775455849212, "grad_norm": 0.2532930596607861, "learning_rate": 5.229076814318122e-05, "loss": 0.5742, "step": 350 }, { "epoch": 2.3081335791845934, "grad_norm": 0.3187321065178083, "learning_rate": 5.211580808524325e-05, "loss": 0.5739, "step": 351 }, { "epoch": 2.3146896127842655, "grad_norm": 0.3465322974483502, "learning_rate": 5.194059244006779e-05, "loss": 0.5699, "step": 352 }, { "epoch": 2.3212456463839377, "grad_norm": 0.48722607810944585, "learning_rate": 5.176512490389055e-05, "loss": 0.5756, "step": 353 }, { "epoch": 2.32780167998361, "grad_norm": 1.1074199903854496, "learning_rate": 5.158940917826099e-05, "loss": 0.5832, "step": 354 }, { "epoch": 2.334357713583282, "grad_norm": 0.3988094901700725, "learning_rate": 5.141344896996422e-05, "loss": 0.5871, "step": 355 }, { "epoch": 2.340913747182954, "grad_norm": 0.3381422893134184, "learning_rate": 5.123724799094279e-05, "loss": 0.5908, "step": 356 }, { "epoch": 2.3474697807826264, "grad_norm": 0.30952489856697063, "learning_rate": 5.106080995821836e-05, "loss": 0.5763, "step": 357 }, { "epoch": 2.3540258143822985, "grad_norm": 0.3302426550471252, "learning_rate": 5.088413859381341e-05, "loss": 0.5796, "step": 358 }, { "epoch": 2.3605818479819707, "grad_norm": 0.2513297320397786, "learning_rate": 5.070723762467254e-05, "loss": 0.5749, "step": 359 }, { "epoch": 2.3671378815816433, "grad_norm": 0.26942873256331556, "learning_rate": 5.053011078258397e-05, "loss": 0.5782, "step": 360 }, { "epoch": 2.3736939151813155, "grad_norm": 0.26811935982451485, "learning_rate": 5.0352761804100835e-05, "loss": 0.5893, "step": 361 }, { "epoch": 2.3802499487809876, "grad_norm": 0.24961381636090246, "learning_rate": 5.017519443046226e-05, "loss": 0.5752, "step": 362 }, { "epoch": 2.3868059823806598, "grad_norm": 0.2789479751757785, "learning_rate": 4.999741240751451e-05, "loss": 0.5819, "step": 363 }, { "epoch": 2.393362015980332, "grad_norm": 0.23968714486879794, "learning_rate": 4.981941948563197e-05, "loss": 0.5864, "step": 364 }, { "epoch": 2.399918049580004, "grad_norm": 0.3892339244230853, "learning_rate": 4.9641219419637985e-05, "loss": 0.589, "step": 365 }, { "epoch": 2.4064740831796763, "grad_norm": 0.19471837418765098, "learning_rate": 4.94628159687257e-05, "loss": 0.5841, "step": 366 }, { "epoch": 2.4130301167793484, "grad_norm": 0.19750205892208292, "learning_rate": 4.928421289637871e-05, "loss": 0.5735, "step": 367 }, { "epoch": 2.4195861503790206, "grad_norm": 0.2849687278419682, "learning_rate": 4.9105413970291747e-05, "loss": 0.5806, "step": 368 }, { "epoch": 2.4261421839786927, "grad_norm": 0.4039957657566193, "learning_rate": 4.892642296229107e-05, "loss": 0.5802, "step": 369 }, { "epoch": 2.432698217578365, "grad_norm": 0.339476402567376, "learning_rate": 4.874724364825504e-05, "loss": 0.583, "step": 370 }, { "epoch": 2.4392542511780375, "grad_norm": 0.2352273125693621, "learning_rate": 4.856787980803437e-05, "loss": 0.5822, "step": 371 }, { "epoch": 2.4458102847777097, "grad_norm": 0.20048904331921943, "learning_rate": 4.8388335225372416e-05, "loss": 0.5865, "step": 372 }, { "epoch": 2.452366318377382, "grad_norm": 0.20167869688700724, "learning_rate": 4.820861368782537e-05, "loss": 0.5753, "step": 373 }, { "epoch": 2.458922351977054, "grad_norm": 0.30771705944153094, "learning_rate": 4.802871898668237e-05, "loss": 0.5888, "step": 374 }, { "epoch": 2.465478385576726, "grad_norm": 0.38232619534443235, "learning_rate": 4.7848654916885446e-05, "loss": 0.5729, "step": 375 }, { "epoch": 2.4720344191763983, "grad_norm": 0.2297372798599205, "learning_rate": 4.7668425276949546e-05, "loss": 0.5806, "step": 376 }, { "epoch": 2.4785904527760705, "grad_norm": 0.22126295840938795, "learning_rate": 4.74880338688824e-05, "loss": 0.5761, "step": 377 }, { "epoch": 2.4851464863757426, "grad_norm": 0.2882635600513661, "learning_rate": 4.730748449810429e-05, "loss": 0.577, "step": 378 }, { "epoch": 2.491702519975415, "grad_norm": 0.306926273041758, "learning_rate": 4.712678097336773e-05, "loss": 0.5822, "step": 379 }, { "epoch": 2.498258553575087, "grad_norm": 0.22726474030427726, "learning_rate": 4.694592710667723e-05, "loss": 0.5795, "step": 380 }, { "epoch": 2.504814587174759, "grad_norm": 0.25322867792835857, "learning_rate": 4.6764926713208756e-05, "loss": 0.5762, "step": 381 }, { "epoch": 2.5113706207744313, "grad_norm": 0.22298267407895275, "learning_rate": 4.658378361122936e-05, "loss": 0.5737, "step": 382 }, { "epoch": 2.5179266543741035, "grad_norm": 0.20258291898015557, "learning_rate": 4.640250162201656e-05, "loss": 0.5787, "step": 383 }, { "epoch": 2.5244826879737756, "grad_norm": 0.22752302168713118, "learning_rate": 4.622108456977773e-05, "loss": 0.58, "step": 384 }, { "epoch": 2.5310387215734482, "grad_norm": 0.7757782276758474, "learning_rate": 4.6039536281569476e-05, "loss": 0.5823, "step": 385 }, { "epoch": 2.5375947551731204, "grad_norm": 0.2825695671491106, "learning_rate": 4.585786058721687e-05, "loss": 0.5617, "step": 386 }, { "epoch": 2.5441507887727925, "grad_norm": 0.18948706553634664, "learning_rate": 4.567606131923263e-05, "loss": 0.5747, "step": 387 }, { "epoch": 2.5507068223724647, "grad_norm": 0.21721693917710894, "learning_rate": 4.549414231273633e-05, "loss": 0.5829, "step": 388 }, { "epoch": 2.557262855972137, "grad_norm": 0.2383552245325273, "learning_rate": 4.531210740537347e-05, "loss": 0.5845, "step": 389 }, { "epoch": 2.563818889571809, "grad_norm": 0.30583160784152247, "learning_rate": 4.512996043723453e-05, "loss": 0.5865, "step": 390 }, { "epoch": 2.570374923171481, "grad_norm": 0.2812991313238803, "learning_rate": 4.494770525077392e-05, "loss": 0.5886, "step": 391 }, { "epoch": 2.5769309567711534, "grad_norm": 0.2573334610716088, "learning_rate": 4.476534569072895e-05, "loss": 0.5812, "step": 392 }, { "epoch": 2.5834869903708255, "grad_norm": 0.18065438415446586, "learning_rate": 4.458288560403878e-05, "loss": 0.5745, "step": 393 }, { "epoch": 2.590043023970498, "grad_norm": 0.2944192743209924, "learning_rate": 4.440032883976318e-05, "loss": 0.5771, "step": 394 }, { "epoch": 2.5965990575701703, "grad_norm": 0.23346564334137798, "learning_rate": 4.421767924900136e-05, "loss": 0.5773, "step": 395 }, { "epoch": 2.6031550911698424, "grad_norm": 0.2274103844648281, "learning_rate": 4.403494068481074e-05, "loss": 0.5792, "step": 396 }, { "epoch": 2.6097111247695146, "grad_norm": 0.18276656304603, "learning_rate": 4.385211700212567e-05, "loss": 0.5858, "step": 397 }, { "epoch": 2.6162671583691868, "grad_norm": 0.18285901538118296, "learning_rate": 4.3669212057676145e-05, "loss": 0.5799, "step": 398 }, { "epoch": 2.622823191968859, "grad_norm": 0.18097190512177241, "learning_rate": 4.348622970990634e-05, "loss": 0.5739, "step": 399 }, { "epoch": 2.629379225568531, "grad_norm": 0.1930953632606967, "learning_rate": 4.33031738188933e-05, "loss": 0.5753, "step": 400 }, { "epoch": 2.6359352591682033, "grad_norm": 0.17356931537284068, "learning_rate": 4.312004824626551e-05, "loss": 0.5777, "step": 401 }, { "epoch": 2.6424912927678754, "grad_norm": 0.1734706404199846, "learning_rate": 4.293685685512142e-05, "loss": 0.5752, "step": 402 }, { "epoch": 2.6490473263675476, "grad_norm": 0.19379812038615699, "learning_rate": 4.275360350994791e-05, "loss": 0.5795, "step": 403 }, { "epoch": 2.6556033599672197, "grad_norm": 0.1660813082194996, "learning_rate": 4.257029207653881e-05, "loss": 0.5826, "step": 404 }, { "epoch": 2.662159393566892, "grad_norm": 0.19954108364376924, "learning_rate": 4.238692642191336e-05, "loss": 0.5825, "step": 405 }, { "epoch": 2.668715427166564, "grad_norm": 0.19322396699850317, "learning_rate": 4.220351041423462e-05, "loss": 0.5841, "step": 406 }, { "epoch": 2.6752714607662362, "grad_norm": 0.17735120440062302, "learning_rate": 4.202004792272785e-05, "loss": 0.5798, "step": 407 }, { "epoch": 2.6818274943659084, "grad_norm": 0.17855572539839387, "learning_rate": 4.183654281759888e-05, "loss": 0.5785, "step": 408 }, { "epoch": 2.6883835279655806, "grad_norm": 0.1998138629729567, "learning_rate": 4.165299896995253e-05, "loss": 0.5782, "step": 409 }, { "epoch": 2.694939561565253, "grad_norm": 0.19733364770879444, "learning_rate": 4.1469420251710905e-05, "loss": 0.5738, "step": 410 }, { "epoch": 2.7014955951649253, "grad_norm": 0.17455960566932077, "learning_rate": 4.128581053553169e-05, "loss": 0.5804, "step": 411 }, { "epoch": 2.7080516287645975, "grad_norm": 0.2277904402914904, "learning_rate": 4.110217369472649e-05, "loss": 0.5823, "step": 412 }, { "epoch": 2.7146076623642696, "grad_norm": 0.17727829284042818, "learning_rate": 4.091851360317912e-05, "loss": 0.5806, "step": 413 }, { "epoch": 2.721163695963942, "grad_norm": 0.20033803661764574, "learning_rate": 4.07348341352639e-05, "loss": 0.5737, "step": 414 }, { "epoch": 2.727719729563614, "grad_norm": 0.28585140789050134, "learning_rate": 4.055113916576386e-05, "loss": 0.5922, "step": 415 }, { "epoch": 2.734275763163286, "grad_norm": 0.1770226397602663, "learning_rate": 4.0367432569789065e-05, "loss": 0.5708, "step": 416 }, { "epoch": 2.7408317967629583, "grad_norm": 0.19810748724107435, "learning_rate": 4.0183718222694823e-05, "loss": 0.5822, "step": 417 }, { "epoch": 2.7473878303626305, "grad_norm": 0.16916185476319684, "learning_rate": 4e-05, "loss": 0.5781, "step": 418 }, { "epoch": 2.753943863962303, "grad_norm": 0.1832943105404772, "learning_rate": 3.9816281777305176e-05, "loss": 0.5812, "step": 419 }, { "epoch": 2.7604998975619752, "grad_norm": 0.1502272563296327, "learning_rate": 3.963256743021095e-05, "loss": 0.5716, "step": 420 }, { "epoch": 2.7670559311616474, "grad_norm": 0.21339511828926633, "learning_rate": 3.944886083423615e-05, "loss": 0.5801, "step": 421 }, { "epoch": 2.7736119647613195, "grad_norm": 0.177457878421534, "learning_rate": 3.92651658647361e-05, "loss": 0.5747, "step": 422 }, { "epoch": 2.7801679983609917, "grad_norm": 0.1839395596929418, "learning_rate": 3.908148639682089e-05, "loss": 0.5843, "step": 423 }, { "epoch": 2.786724031960664, "grad_norm": 0.19753047943291085, "learning_rate": 3.889782630527353e-05, "loss": 0.574, "step": 424 }, { "epoch": 2.793280065560336, "grad_norm": 0.14640408330908058, "learning_rate": 3.8714189464468334e-05, "loss": 0.5786, "step": 425 }, { "epoch": 2.799836099160008, "grad_norm": 0.19103262838391485, "learning_rate": 3.853057974828911e-05, "loss": 0.5798, "step": 426 }, { "epoch": 2.8063921327596804, "grad_norm": 0.1673182640044429, "learning_rate": 3.834700103004747e-05, "loss": 0.5729, "step": 427 }, { "epoch": 2.8129481663593525, "grad_norm": 0.20105657869010776, "learning_rate": 3.816345718240113e-05, "loss": 0.5744, "step": 428 }, { "epoch": 2.8195041999590247, "grad_norm": 0.13814925979553985, "learning_rate": 3.797995207727217e-05, "loss": 0.5694, "step": 429 }, { "epoch": 2.826060233558697, "grad_norm": 0.16253212459987165, "learning_rate": 3.779648958576538e-05, "loss": 0.5732, "step": 430 }, { "epoch": 2.832616267158369, "grad_norm": 0.14306544183315967, "learning_rate": 3.7613073578086644e-05, "loss": 0.5737, "step": 431 }, { "epoch": 2.839172300758041, "grad_norm": 0.16148301199590984, "learning_rate": 3.74297079234612e-05, "loss": 0.5696, "step": 432 }, { "epoch": 2.8457283343577133, "grad_norm": 0.1591738101867343, "learning_rate": 3.7246396490052117e-05, "loss": 0.5752, "step": 433 }, { "epoch": 2.852284367957386, "grad_norm": 0.14229281054940401, "learning_rate": 3.706314314487859e-05, "loss": 0.5778, "step": 434 }, { "epoch": 2.858840401557058, "grad_norm": 0.18401979626969228, "learning_rate": 3.687995175373449e-05, "loss": 0.5785, "step": 435 }, { "epoch": 2.8653964351567303, "grad_norm": 0.14532042872787476, "learning_rate": 3.669682618110671e-05, "loss": 0.5717, "step": 436 }, { "epoch": 2.8719524687564024, "grad_norm": 0.14280350348324158, "learning_rate": 3.6513770290093674e-05, "loss": 0.5737, "step": 437 }, { "epoch": 2.8785085023560746, "grad_norm": 0.17828968160150296, "learning_rate": 3.6330787942323855e-05, "loss": 0.5739, "step": 438 }, { "epoch": 2.8850645359557467, "grad_norm": 0.1680340039645173, "learning_rate": 3.614788299787434e-05, "loss": 0.5719, "step": 439 }, { "epoch": 2.891620569555419, "grad_norm": 0.16773501838978377, "learning_rate": 3.5965059315189274e-05, "loss": 0.5861, "step": 440 }, { "epoch": 2.898176603155091, "grad_norm": 0.1466119984148444, "learning_rate": 3.578232075099866e-05, "loss": 0.5719, "step": 441 }, { "epoch": 2.9047326367547632, "grad_norm": 0.18860006392370388, "learning_rate": 3.559967116023683e-05, "loss": 0.573, "step": 442 }, { "epoch": 2.911288670354436, "grad_norm": 0.17180023728534166, "learning_rate": 3.541711439596122e-05, "loss": 0.5715, "step": 443 }, { "epoch": 2.917844703954108, "grad_norm": 0.16863095150334853, "learning_rate": 3.523465430927106e-05, "loss": 0.587, "step": 444 }, { "epoch": 2.92440073755378, "grad_norm": 0.18865259295148487, "learning_rate": 3.5052294749226094e-05, "loss": 0.5705, "step": 445 }, { "epoch": 2.9309567711534523, "grad_norm": 0.3900404738786457, "learning_rate": 3.4870039562765475e-05, "loss": 0.5808, "step": 446 }, { "epoch": 2.9375128047531245, "grad_norm": 0.1577690647524753, "learning_rate": 3.4687892594626536e-05, "loss": 0.5767, "step": 447 }, { "epoch": 2.9440688383527966, "grad_norm": 0.17930301136018978, "learning_rate": 3.4505857687263675e-05, "loss": 0.5794, "step": 448 }, { "epoch": 2.950624871952469, "grad_norm": 0.21442963322771297, "learning_rate": 3.432393868076739e-05, "loss": 0.5813, "step": 449 }, { "epoch": 2.957180905552141, "grad_norm": 0.5263365227065998, "learning_rate": 3.414213941278314e-05, "loss": 0.5712, "step": 450 }, { "epoch": 2.963736939151813, "grad_norm": 0.24486881857008144, "learning_rate": 3.396046371843052e-05, "loss": 0.5862, "step": 451 }, { "epoch": 2.9702929727514853, "grad_norm": 0.14980108475089587, "learning_rate": 3.377891543022229e-05, "loss": 0.5801, "step": 452 }, { "epoch": 2.9768490063511575, "grad_norm": 0.20536856095424458, "learning_rate": 3.3597498377983444e-05, "loss": 0.5802, "step": 453 }, { "epoch": 2.9834050399508296, "grad_norm": 0.18782303046647886, "learning_rate": 3.341621638877064e-05, "loss": 0.5784, "step": 454 }, { "epoch": 2.989961073550502, "grad_norm": 0.14515446375387317, "learning_rate": 3.3235073286791264e-05, "loss": 0.5754, "step": 455 }, { "epoch": 2.996517107150174, "grad_norm": 0.16161434398106808, "learning_rate": 3.305407289332279e-05, "loss": 0.5756, "step": 456 }, { "epoch": 3.0055316533497236, "grad_norm": 0.3533323244538683, "learning_rate": 3.287321902663229e-05, "loss": 1.0491, "step": 457 }, { "epoch": 3.0120876869493958, "grad_norm": 0.3045514737546258, "learning_rate": 3.269251550189573e-05, "loss": 0.5446, "step": 458 }, { "epoch": 3.018643720549068, "grad_norm": 0.25472049022532234, "learning_rate": 3.251196613111761e-05, "loss": 0.5407, "step": 459 }, { "epoch": 3.02519975414874, "grad_norm": 0.25510858573604617, "learning_rate": 3.2331574723050474e-05, "loss": 0.5411, "step": 460 }, { "epoch": 3.0317557877484123, "grad_norm": 0.38483951060765287, "learning_rate": 3.2151345083114574e-05, "loss": 0.5396, "step": 461 }, { "epoch": 3.0383118213480844, "grad_norm": 0.2558618267692145, "learning_rate": 3.197128101331764e-05, "loss": 0.5349, "step": 462 }, { "epoch": 3.0448678549477566, "grad_norm": 0.23801295747592094, "learning_rate": 3.179138631217463e-05, "loss": 0.5439, "step": 463 }, { "epoch": 3.0514238885474287, "grad_norm": 0.2746658371274729, "learning_rate": 3.161166477462759e-05, "loss": 0.5463, "step": 464 }, { "epoch": 3.057979922147101, "grad_norm": 0.23307565195751595, "learning_rate": 3.1432120191965647e-05, "loss": 0.538, "step": 465 }, { "epoch": 3.064535955746773, "grad_norm": 0.23849808366960823, "learning_rate": 3.125275635174497e-05, "loss": 0.5431, "step": 466 }, { "epoch": 3.0710919893464452, "grad_norm": 0.22477045443908192, "learning_rate": 3.1073577037708935e-05, "loss": 0.5422, "step": 467 }, { "epoch": 3.0776480229461174, "grad_norm": 0.2060382177048335, "learning_rate": 3.089458602970828e-05, "loss": 0.5446, "step": 468 }, { "epoch": 3.08420405654579, "grad_norm": 0.22200054079896087, "learning_rate": 3.0715787103621294e-05, "loss": 0.5479, "step": 469 }, { "epoch": 3.090760090145462, "grad_norm": 0.17039781849738478, "learning_rate": 3.0537184031274306e-05, "loss": 0.5305, "step": 470 }, { "epoch": 3.0973161237451343, "grad_norm": 0.16941181980633563, "learning_rate": 3.0358780580362025e-05, "loss": 0.5261, "step": 471 }, { "epoch": 3.1038721573448065, "grad_norm": 0.17660138294537184, "learning_rate": 3.0180580514368037e-05, "loss": 0.5312, "step": 472 }, { "epoch": 3.1104281909444786, "grad_norm": 0.18073996525501446, "learning_rate": 3.0002587592485497e-05, "loss": 0.5371, "step": 473 }, { "epoch": 3.116984224544151, "grad_norm": 0.3988440151798082, "learning_rate": 2.9824805569537747e-05, "loss": 0.5513, "step": 474 }, { "epoch": 3.123540258143823, "grad_norm": 0.20423237269747027, "learning_rate": 2.9647238195899168e-05, "loss": 0.5375, "step": 475 }, { "epoch": 3.130096291743495, "grad_norm": 0.17213193598614746, "learning_rate": 2.9469889217416045e-05, "loss": 0.5437, "step": 476 }, { "epoch": 3.1366523253431673, "grad_norm": 0.21207611299854595, "learning_rate": 2.9292762375327483e-05, "loss": 0.5339, "step": 477 }, { "epoch": 3.1432083589428395, "grad_norm": 0.15310274061749113, "learning_rate": 2.9115861406186593e-05, "loss": 0.5303, "step": 478 }, { "epoch": 3.1497643925425116, "grad_norm": 0.1877335953144904, "learning_rate": 2.8939190041781647e-05, "loss": 0.5319, "step": 479 }, { "epoch": 3.1563204261421838, "grad_norm": 0.1458292108542802, "learning_rate": 2.8762752009057232e-05, "loss": 0.54, "step": 480 }, { "epoch": 3.1628764597418564, "grad_norm": 0.17442833234197713, "learning_rate": 2.85865510300358e-05, "loss": 0.5377, "step": 481 }, { "epoch": 3.1694324933415285, "grad_norm": 0.15457635774978962, "learning_rate": 2.841059082173902e-05, "loss": 0.5389, "step": 482 }, { "epoch": 3.1759885269412007, "grad_norm": 0.16823476302413515, "learning_rate": 2.823487509610946e-05, "loss": 0.5435, "step": 483 }, { "epoch": 3.182544560540873, "grad_norm": 0.14604985181804828, "learning_rate": 2.805940755993223e-05, "loss": 0.5377, "step": 484 }, { "epoch": 3.189100594140545, "grad_norm": 0.14560052298572645, "learning_rate": 2.7884191914756757e-05, "loss": 0.5409, "step": 485 }, { "epoch": 3.195656627740217, "grad_norm": 0.1472508058071573, "learning_rate": 2.770923185681878e-05, "loss": 0.5455, "step": 486 }, { "epoch": 3.2022126613398894, "grad_norm": 0.13624234976069768, "learning_rate": 2.7534531076962356e-05, "loss": 0.5433, "step": 487 }, { "epoch": 3.2087686949395615, "grad_norm": 0.15264871662264248, "learning_rate": 2.7360093260561904e-05, "loss": 0.5372, "step": 488 }, { "epoch": 3.2153247285392337, "grad_norm": 0.1462659967641793, "learning_rate": 2.7185922087444602e-05, "loss": 0.538, "step": 489 }, { "epoch": 3.221880762138906, "grad_norm": 0.16303592219574492, "learning_rate": 2.7012021231812666e-05, "loss": 0.542, "step": 490 }, { "epoch": 3.228436795738578, "grad_norm": 0.14173407966037618, "learning_rate": 2.6838394362165875e-05, "loss": 0.5387, "step": 491 }, { "epoch": 3.23499282933825, "grad_norm": 0.15432095686939976, "learning_rate": 2.6665045141224193e-05, "loss": 0.5377, "step": 492 }, { "epoch": 3.2415488629379228, "grad_norm": 0.1482561893469378, "learning_rate": 2.6491977225850446e-05, "loss": 0.5371, "step": 493 }, { "epoch": 3.248104896537595, "grad_norm": 0.16624560752182815, "learning_rate": 2.6319194266973256e-05, "loss": 0.5359, "step": 494 }, { "epoch": 3.254660930137267, "grad_norm": 0.15016468812997047, "learning_rate": 2.6146699909509984e-05, "loss": 0.5411, "step": 495 }, { "epoch": 3.2612169637369393, "grad_norm": 0.2450944209511445, "learning_rate": 2.597449779228983e-05, "loss": 0.5429, "step": 496 }, { "epoch": 3.2677729973366114, "grad_norm": 0.2412941457078939, "learning_rate": 2.580259154797709e-05, "loss": 0.5343, "step": 497 }, { "epoch": 3.2743290309362836, "grad_norm": 0.14493021044588159, "learning_rate": 2.563098480299451e-05, "loss": 0.5409, "step": 498 }, { "epoch": 3.2808850645359557, "grad_norm": 0.16960807003233563, "learning_rate": 2.5459681177446803e-05, "loss": 0.5389, "step": 499 }, { "epoch": 3.287441098135628, "grad_norm": 0.14869909968414688, "learning_rate": 2.5288684285044283e-05, "loss": 0.5353, "step": 500 }, { "epoch": 3.2939971317353, "grad_norm": 0.14777225005600433, "learning_rate": 2.5117997733026566e-05, "loss": 0.5393, "step": 501 }, { "epoch": 3.3005531653349722, "grad_norm": 0.15139629935458157, "learning_rate": 2.4947625122086585e-05, "loss": 0.5435, "step": 502 }, { "epoch": 3.3071091989346444, "grad_norm": 0.1392849654786053, "learning_rate": 2.477757004629456e-05, "loss": 0.5375, "step": 503 }, { "epoch": 3.3136652325343166, "grad_norm": 0.1367733939357107, "learning_rate": 2.460783609302218e-05, "loss": 0.5426, "step": 504 }, { "epoch": 3.3202212661339887, "grad_norm": 0.12699654005182315, "learning_rate": 2.4438426842866966e-05, "loss": 0.5353, "step": 505 }, { "epoch": 3.3267772997336613, "grad_norm": 0.1390969751788505, "learning_rate": 2.4269345869576676e-05, "loss": 0.539, "step": 506 }, { "epoch": 3.3333333333333335, "grad_norm": 0.12248688460845146, "learning_rate": 2.4100596739973993e-05, "loss": 0.5392, "step": 507 }, { "epoch": 3.3398893669330056, "grad_norm": 0.13805750150926843, "learning_rate": 2.393218301388123e-05, "loss": 0.5401, "step": 508 }, { "epoch": 3.346445400532678, "grad_norm": 0.12180259855826225, "learning_rate": 2.3764108244045212e-05, "loss": 0.5465, "step": 509 }, { "epoch": 3.35300143413235, "grad_norm": 0.12245696463081471, "learning_rate": 2.35963759760624e-05, "loss": 0.5311, "step": 510 }, { "epoch": 3.359557467732022, "grad_norm": 0.12896584095171493, "learning_rate": 2.342898974830405e-05, "loss": 0.5391, "step": 511 }, { "epoch": 3.3661135013316943, "grad_norm": 0.12281667944621683, "learning_rate": 2.3261953091841553e-05, "loss": 0.5335, "step": 512 }, { "epoch": 3.3726695349313665, "grad_norm": 0.11829751242108535, "learning_rate": 2.3095269530372032e-05, "loss": 0.5447, "step": 513 }, { "epoch": 3.3792255685310386, "grad_norm": 0.13317495284187636, "learning_rate": 2.2928942580143855e-05, "loss": 0.5438, "step": 514 }, { "epoch": 3.3857816021307108, "grad_norm": 0.12369993983516651, "learning_rate": 2.276297574988263e-05, "loss": 0.5433, "step": 515 }, { "epoch": 3.392337635730383, "grad_norm": 0.1246149968311202, "learning_rate": 2.2597372540717083e-05, "loss": 0.5412, "step": 516 }, { "epoch": 3.3988936693300555, "grad_norm": 0.12919720585139893, "learning_rate": 2.2432136446105192e-05, "loss": 0.5442, "step": 517 }, { "epoch": 3.4054497029297277, "grad_norm": 0.12089802144087428, "learning_rate": 2.226727095176057e-05, "loss": 0.539, "step": 518 }, { "epoch": 3.4120057365294, "grad_norm": 0.13804336756907107, "learning_rate": 2.210277953557888e-05, "loss": 0.5462, "step": 519 }, { "epoch": 3.418561770129072, "grad_norm": 0.12982437086238163, "learning_rate": 2.1938665667564435e-05, "loss": 0.5344, "step": 520 }, { "epoch": 3.425117803728744, "grad_norm": 0.12345004982954612, "learning_rate": 2.177493280975708e-05, "loss": 0.5465, "step": 521 }, { "epoch": 3.4316738373284164, "grad_norm": 0.14573537922186017, "learning_rate": 2.1611584416159106e-05, "loss": 0.5323, "step": 522 }, { "epoch": 3.4382298709280885, "grad_norm": 0.1097589192258888, "learning_rate": 2.1448623932662377e-05, "loss": 0.5437, "step": 523 }, { "epoch": 3.4447859045277607, "grad_norm": 0.1317797633987106, "learning_rate": 2.1286054796975696e-05, "loss": 0.5377, "step": 524 }, { "epoch": 3.451341938127433, "grad_norm": 0.13321471541724095, "learning_rate": 2.1123880438552187e-05, "loss": 0.5518, "step": 525 }, { "epoch": 3.457897971727105, "grad_norm": 0.1352447131480959, "learning_rate": 2.096210427851706e-05, "loss": 0.532, "step": 526 }, { "epoch": 3.464454005326777, "grad_norm": 0.13508733978142928, "learning_rate": 2.0800729729595385e-05, "loss": 0.5314, "step": 527 }, { "epoch": 3.4710100389264493, "grad_norm": 0.1380525297606027, "learning_rate": 2.063976019604006e-05, "loss": 0.5438, "step": 528 }, { "epoch": 3.4775660725261215, "grad_norm": 0.13529984461100034, "learning_rate": 2.0479199073560084e-05, "loss": 0.5451, "step": 529 }, { "epoch": 3.4841221061257936, "grad_norm": 0.13383345341162217, "learning_rate": 2.0319049749248876e-05, "loss": 0.5335, "step": 530 }, { "epoch": 3.4906781397254663, "grad_norm": 0.12966276679499278, "learning_rate": 2.0159315601512817e-05, "loss": 0.5404, "step": 531 }, { "epoch": 3.4972341733251384, "grad_norm": 0.12661938479816495, "learning_rate": 2.0000000000000012e-05, "loss": 0.5361, "step": 532 }, { "epoch": 3.5037902069248106, "grad_norm": 0.25950697427544217, "learning_rate": 1.9841106305529133e-05, "loss": 0.5475, "step": 533 }, { "epoch": 3.5103462405244827, "grad_norm": 0.12388736409274281, "learning_rate": 1.9682637870018638e-05, "loss": 0.5433, "step": 534 }, { "epoch": 3.516902274124155, "grad_norm": 0.13692640136699888, "learning_rate": 1.9524598036415973e-05, "loss": 0.5379, "step": 535 }, { "epoch": 3.523458307723827, "grad_norm": 0.22971792027818427, "learning_rate": 1.9366990138627054e-05, "loss": 0.5354, "step": 536 }, { "epoch": 3.5300143413234992, "grad_norm": 0.13477440715558586, "learning_rate": 1.9209817501445978e-05, "loss": 0.5408, "step": 537 }, { "epoch": 3.5365703749231714, "grad_norm": 0.12000818819921334, "learning_rate": 1.9053083440484887e-05, "loss": 0.539, "step": 538 }, { "epoch": 3.5431264085228436, "grad_norm": 0.18018036064763043, "learning_rate": 1.889679126210397e-05, "loss": 0.54, "step": 539 }, { "epoch": 3.5496824421225157, "grad_norm": 0.13187429164193115, "learning_rate": 1.8740944263341773e-05, "loss": 0.5336, "step": 540 }, { "epoch": 3.5562384757221883, "grad_norm": 0.1115344100782877, "learning_rate": 1.8585545731845584e-05, "loss": 0.5457, "step": 541 }, { "epoch": 3.5627945093218605, "grad_norm": 0.4821099792222361, "learning_rate": 1.8430598945802156e-05, "loss": 0.5429, "step": 542 }, { "epoch": 3.5693505429215326, "grad_norm": 0.13726412619312758, "learning_rate": 1.8276107173868503e-05, "loss": 0.551, "step": 543 }, { "epoch": 3.575906576521205, "grad_norm": 0.11727031408692318, "learning_rate": 1.8122073675102935e-05, "loss": 0.5315, "step": 544 }, { "epoch": 3.582462610120877, "grad_norm": 0.12442269217723008, "learning_rate": 1.7968501698896346e-05, "loss": 0.5338, "step": 545 }, { "epoch": 3.589018643720549, "grad_norm": 0.11894782465239562, "learning_rate": 1.781539448490365e-05, "loss": 0.5365, "step": 546 }, { "epoch": 3.5955746773202213, "grad_norm": 0.11002626026600501, "learning_rate": 1.7662755262975432e-05, "loss": 0.5404, "step": 547 }, { "epoch": 3.6021307109198935, "grad_norm": 0.13251223380465485, "learning_rate": 1.7510587253089842e-05, "loss": 0.543, "step": 548 }, { "epoch": 3.6086867445195656, "grad_norm": 0.11425745596834401, "learning_rate": 1.7358893665284595e-05, "loss": 0.5345, "step": 549 }, { "epoch": 3.6152427781192378, "grad_norm": 0.11826825954302657, "learning_rate": 1.7207677699589355e-05, "loss": 0.5463, "step": 550 }, { "epoch": 3.62179881171891, "grad_norm": 0.10903975348972718, "learning_rate": 1.7056942545958167e-05, "loss": 0.5396, "step": 551 }, { "epoch": 3.628354845318582, "grad_norm": 0.11693940603496059, "learning_rate": 1.690669138420215e-05, "loss": 0.5305, "step": 552 }, { "epoch": 3.6349108789182543, "grad_norm": 0.1142560549805311, "learning_rate": 1.6756927383922473e-05, "loss": 0.5391, "step": 553 }, { "epoch": 3.6414669125179264, "grad_norm": 0.10839349884507322, "learning_rate": 1.6607653704443457e-05, "loss": 0.542, "step": 554 }, { "epoch": 3.6480229461175986, "grad_norm": 0.11225360540051456, "learning_rate": 1.6458873494745926e-05, "loss": 0.5388, "step": 555 }, { "epoch": 3.654578979717271, "grad_norm": 0.11188297317863019, "learning_rate": 1.6310589893400804e-05, "loss": 0.5381, "step": 556 }, { "epoch": 3.6611350133169434, "grad_norm": 0.7105498519897432, "learning_rate": 1.6162806028502852e-05, "loss": 0.5411, "step": 557 }, { "epoch": 3.6676910469166155, "grad_norm": 0.11475316554461859, "learning_rate": 1.601552501760473e-05, "loss": 0.5416, "step": 558 }, { "epoch": 3.6742470805162877, "grad_norm": 0.12194606250202103, "learning_rate": 1.5868749967651252e-05, "loss": 0.5396, "step": 559 }, { "epoch": 3.68080311411596, "grad_norm": 0.12661866626370497, "learning_rate": 1.5722483974913737e-05, "loss": 0.531, "step": 560 }, { "epoch": 3.687359147715632, "grad_norm": 0.11256332358317793, "learning_rate": 1.5576730124924822e-05, "loss": 0.5418, "step": 561 }, { "epoch": 3.693915181315304, "grad_norm": 0.11015606275564016, "learning_rate": 1.5431491492413288e-05, "loss": 0.5411, "step": 562 }, { "epoch": 3.7004712149149763, "grad_norm": 0.12074411265164606, "learning_rate": 1.528677114123923e-05, "loss": 0.53, "step": 563 }, { "epoch": 3.7070272485146485, "grad_norm": 0.10893671793883805, "learning_rate": 1.5142572124329418e-05, "loss": 0.5397, "step": 564 }, { "epoch": 3.713583282114321, "grad_norm": 0.11746065306540288, "learning_rate": 1.4998897483612865e-05, "loss": 0.5423, "step": 565 }, { "epoch": 3.7201393157139933, "grad_norm": 0.1126528745871064, "learning_rate": 1.4855750249956718e-05, "loss": 0.5426, "step": 566 }, { "epoch": 3.7266953493136654, "grad_norm": 0.3165608272055915, "learning_rate": 1.4713133443102283e-05, "loss": 0.5428, "step": 567 }, { "epoch": 3.7332513829133376, "grad_norm": 0.11800782137431694, "learning_rate": 1.457105007160129e-05, "loss": 0.5443, "step": 568 }, { "epoch": 3.7398074165130097, "grad_norm": 0.11122501505766147, "learning_rate": 1.44295031327525e-05, "loss": 0.5397, "step": 569 }, { "epoch": 3.746363450112682, "grad_norm": 0.12798481597310407, "learning_rate": 1.4288495612538427e-05, "loss": 0.5457, "step": 570 }, { "epoch": 3.752919483712354, "grad_norm": 0.10441587675286511, "learning_rate": 1.4148030485562362e-05, "loss": 0.5371, "step": 571 }, { "epoch": 3.7594755173120262, "grad_norm": 0.12743357362001942, "learning_rate": 1.4008110714985623e-05, "loss": 0.5392, "step": 572 }, { "epoch": 3.7660315509116984, "grad_norm": 0.1114224145991442, "learning_rate": 1.3868739252465017e-05, "loss": 0.5368, "step": 573 }, { "epoch": 3.7725875845113706, "grad_norm": 0.10668882720776016, "learning_rate": 1.3729919038090627e-05, "loss": 0.5386, "step": 574 }, { "epoch": 3.7791436181110427, "grad_norm": 0.11554119750738572, "learning_rate": 1.3591653000323764e-05, "loss": 0.5336, "step": 575 }, { "epoch": 3.785699651710715, "grad_norm": 0.11456186328555522, "learning_rate": 1.3453944055935151e-05, "loss": 0.5447, "step": 576 }, { "epoch": 3.792255685310387, "grad_norm": 0.12119615336966778, "learning_rate": 1.3316795109943476e-05, "loss": 0.5333, "step": 577 }, { "epoch": 3.798811718910059, "grad_norm": 0.1097823442621975, "learning_rate": 1.3180209055554043e-05, "loss": 0.5353, "step": 578 }, { "epoch": 3.8053677525097314, "grad_norm": 0.12113771390183198, "learning_rate": 1.3044188774097757e-05, "loss": 0.5406, "step": 579 }, { "epoch": 3.811923786109404, "grad_norm": 0.1076103282360743, "learning_rate": 1.2908737134970367e-05, "loss": 0.5428, "step": 580 }, { "epoch": 3.818479819709076, "grad_norm": 0.10778481828407349, "learning_rate": 1.2773856995571858e-05, "loss": 0.5346, "step": 581 }, { "epoch": 3.8250358533087483, "grad_norm": 0.10605992762588787, "learning_rate": 1.2639551201246278e-05, "loss": 0.5394, "step": 582 }, { "epoch": 3.8315918869084205, "grad_norm": 0.10492552888703595, "learning_rate": 1.2505822585221665e-05, "loss": 0.5402, "step": 583 }, { "epoch": 3.8381479205080926, "grad_norm": 0.10835031468844004, "learning_rate": 1.2372673968550229e-05, "loss": 0.5449, "step": 584 }, { "epoch": 3.8447039541077648, "grad_norm": 0.10610757084756944, "learning_rate": 1.2240108160048934e-05, "loss": 0.5366, "step": 585 }, { "epoch": 3.851259987707437, "grad_norm": 0.10287232050693294, "learning_rate": 1.2108127956240186e-05, "loss": 0.5367, "step": 586 }, { "epoch": 3.857816021307109, "grad_norm": 0.10929614703844563, "learning_rate": 1.1976736141292853e-05, "loss": 0.5335, "step": 587 }, { "epoch": 3.8643720549067813, "grad_norm": 0.1204731707453792, "learning_rate": 1.1845935486963546e-05, "loss": 0.5404, "step": 588 }, { "epoch": 3.870928088506454, "grad_norm": 0.10037277898866681, "learning_rate": 1.1715728752538103e-05, "loss": 0.5386, "step": 589 }, { "epoch": 3.877484122106126, "grad_norm": 0.10755482937489974, "learning_rate": 1.158611868477344e-05, "loss": 0.5396, "step": 590 }, { "epoch": 3.884040155705798, "grad_norm": 0.1125655323891997, "learning_rate": 1.1457108017839587e-05, "loss": 0.5522, "step": 591 }, { "epoch": 3.8905961893054704, "grad_norm": 0.09908512763395612, "learning_rate": 1.1328699473261957e-05, "loss": 0.5389, "step": 592 }, { "epoch": 3.8971522229051425, "grad_norm": 0.10011512841325966, "learning_rate": 1.1200895759864027e-05, "loss": 0.5402, "step": 593 }, { "epoch": 3.9037082565048147, "grad_norm": 0.11190228493508152, "learning_rate": 1.107369957371013e-05, "loss": 0.5402, "step": 594 }, { "epoch": 3.910264290104487, "grad_norm": 0.0997294109524961, "learning_rate": 1.09471135980486e-05, "loss": 0.5356, "step": 595 }, { "epoch": 3.916820323704159, "grad_norm": 0.10926000465284759, "learning_rate": 1.0821140503255174e-05, "loss": 0.5459, "step": 596 }, { "epoch": 3.923376357303831, "grad_norm": 0.10485038006440338, "learning_rate": 1.0695782946776619e-05, "loss": 0.5407, "step": 597 }, { "epoch": 3.9299323909035033, "grad_norm": 0.10291841117205842, "learning_rate": 1.0571043573074737e-05, "loss": 0.5373, "step": 598 }, { "epoch": 3.9364884245031755, "grad_norm": 0.10153144309813497, "learning_rate": 1.0446925013570545e-05, "loss": 0.5408, "step": 599 }, { "epoch": 3.9430444581028476, "grad_norm": 0.10605125481156527, "learning_rate": 1.0323429886588743e-05, "loss": 0.5411, "step": 600 }, { "epoch": 3.94960049170252, "grad_norm": 0.10215451404049354, "learning_rate": 1.020056079730252e-05, "loss": 0.5428, "step": 601 }, { "epoch": 3.956156525302192, "grad_norm": 0.10818454985255951, "learning_rate": 1.0078320337678584e-05, "loss": 0.5396, "step": 602 }, { "epoch": 3.962712558901864, "grad_norm": 0.10188342720575627, "learning_rate": 9.956711086422471e-06, "loss": 0.5486, "step": 603 }, { "epoch": 3.9692685925015363, "grad_norm": 0.10148315246310981, "learning_rate": 9.835735608924155e-06, "loss": 0.5341, "step": 604 }, { "epoch": 3.975824626101209, "grad_norm": 0.09241997326882184, "learning_rate": 9.715396457203918e-06, "loss": 0.5389, "step": 605 }, { "epoch": 3.982380659700881, "grad_norm": 0.102047541160635, "learning_rate": 9.595696169858542e-06, "loss": 0.5343, "step": 606 }, { "epoch": 3.9889366933005532, "grad_norm": 0.16329932937909056, "learning_rate": 9.476637272007748e-06, "loss": 0.5414, "step": 607 }, { "epoch": 3.9954927269002254, "grad_norm": 0.10040695165913859, "learning_rate": 9.358222275240884e-06, "loss": 0.5365, "step": 608 }, { "epoch": 4.006556033599672, "grad_norm": 0.16186764492729486, "learning_rate": 9.24045367756401e-06, "loss": 0.5171, "step": 609 }, { "epoch": 4.013112067199344, "grad_norm": 0.1322033843585489, "learning_rate": 9.123333963347166e-06, "loss": 0.5209, "step": 610 }, { "epoch": 4.0196681007990165, "grad_norm": 0.12245478327966862, "learning_rate": 9.006865603271952e-06, "loss": 0.5112, "step": 611 }, { "epoch": 4.026224134398689, "grad_norm": 0.117065406177187, "learning_rate": 8.89105105427945e-06, "loss": 0.5089, "step": 612 }, { "epoch": 4.032780167998361, "grad_norm": 0.1358673718015058, "learning_rate": 8.775892759518321e-06, "loss": 0.5245, "step": 613 }, { "epoch": 4.039336201598033, "grad_norm": 0.1436375868810231, "learning_rate": 8.661393148293355e-06, "loss": 0.523, "step": 614 }, { "epoch": 4.045892235197705, "grad_norm": 0.11924450152050672, "learning_rate": 8.547554636014177e-06, "loss": 0.506, "step": 615 }, { "epoch": 4.052448268797377, "grad_norm": 0.11988650232489285, "learning_rate": 8.434379624144261e-06, "loss": 0.515, "step": 616 }, { "epoch": 4.0590043023970495, "grad_norm": 0.13738230782732197, "learning_rate": 8.321870500150347e-06, "loss": 0.511, "step": 617 }, { "epoch": 4.065560335996722, "grad_norm": 0.1371441628587204, "learning_rate": 8.210029637452016e-06, "loss": 0.5139, "step": 618 }, { "epoch": 4.072116369596394, "grad_norm": 0.3628915461879963, "learning_rate": 8.098859395371641e-06, "loss": 0.5165, "step": 619 }, { "epoch": 4.078672403196066, "grad_norm": 0.11429063516858927, "learning_rate": 7.988362119084642e-06, "loss": 0.5136, "step": 620 }, { "epoch": 4.085228436795739, "grad_norm": 0.13045931327904303, "learning_rate": 7.87854013956994e-06, "loss": 0.5097, "step": 621 }, { "epoch": 4.091784470395411, "grad_norm": 0.11629419650039642, "learning_rate": 7.769395773560874e-06, "loss": 0.5157, "step": 622 }, { "epoch": 4.098340503995083, "grad_norm": 0.11362363257154101, "learning_rate": 7.660931323496283e-06, "loss": 0.5188, "step": 623 }, { "epoch": 4.1048965375947555, "grad_norm": 0.10992983384699095, "learning_rate": 7.553149077471915e-06, "loss": 0.5124, "step": 624 }, { "epoch": 4.111452571194428, "grad_norm": 0.10426021171852758, "learning_rate": 7.446051309192204e-06, "loss": 0.5142, "step": 625 }, { "epoch": 4.1180086047941, "grad_norm": 0.10863479923168164, "learning_rate": 7.3396402779222845e-06, "loss": 0.5127, "step": 626 }, { "epoch": 4.124564638393772, "grad_norm": 0.10635132221465905, "learning_rate": 7.233918228440324e-06, "loss": 0.5162, "step": 627 }, { "epoch": 4.131120671993444, "grad_norm": 0.09912102469646186, "learning_rate": 7.128887390990198e-06, "loss": 0.5224, "step": 628 }, { "epoch": 4.137676705593116, "grad_norm": 0.10290328673729038, "learning_rate": 7.024549981234377e-06, "loss": 0.5217, "step": 629 }, { "epoch": 4.1442327391927885, "grad_norm": 0.1038941435526061, "learning_rate": 6.9209082002072725e-06, "loss": 0.5133, "step": 630 }, { "epoch": 4.150788772792461, "grad_norm": 0.09603936143418888, "learning_rate": 6.817964234268748e-06, "loss": 0.5176, "step": 631 }, { "epoch": 4.157344806392133, "grad_norm": 0.11813529879188972, "learning_rate": 6.715720255058e-06, "loss": 0.5152, "step": 632 }, { "epoch": 4.163900839991805, "grad_norm": 0.09790007822885208, "learning_rate": 6.614178419447781e-06, "loss": 0.5158, "step": 633 }, { "epoch": 4.170456873591477, "grad_norm": 0.10792344691656378, "learning_rate": 6.513340869498859e-06, "loss": 0.5122, "step": 634 }, { "epoch": 4.177012907191149, "grad_norm": 0.24107917853631503, "learning_rate": 6.4132097324148556e-06, "loss": 0.5161, "step": 635 }, { "epoch": 4.183568940790821, "grad_norm": 0.09669004690900093, "learning_rate": 6.313787120497376e-06, "loss": 0.507, "step": 636 }, { "epoch": 4.190124974390494, "grad_norm": 0.10731913975620888, "learning_rate": 6.215075131101405e-06, "loss": 0.5087, "step": 637 }, { "epoch": 4.196681007990166, "grad_norm": 0.0997103006367049, "learning_rate": 6.117075846591123e-06, "loss": 0.5201, "step": 638 }, { "epoch": 4.203237041589838, "grad_norm": 0.09955705599544924, "learning_rate": 6.019791334295955e-06, "loss": 0.5076, "step": 639 }, { "epoch": 4.20979307518951, "grad_norm": 0.09540357195970695, "learning_rate": 5.923223646466923e-06, "loss": 0.5181, "step": 640 }, { "epoch": 4.216349108789182, "grad_norm": 0.0952939694361504, "learning_rate": 5.827374820233407e-06, "loss": 0.5195, "step": 641 }, { "epoch": 4.222905142388854, "grad_norm": 0.09942874484427339, "learning_rate": 5.732246877560146e-06, "loss": 0.5171, "step": 642 }, { "epoch": 4.2294611759885266, "grad_norm": 0.09564484595006637, "learning_rate": 5.637841825204588e-06, "loss": 0.5131, "step": 643 }, { "epoch": 4.236017209588199, "grad_norm": 0.1056325087529548, "learning_rate": 5.5441616546745646e-06, "loss": 0.5095, "step": 644 }, { "epoch": 4.242573243187872, "grad_norm": 0.09442156372590707, "learning_rate": 5.451208342186229e-06, "loss": 0.5139, "step": 645 }, { "epoch": 4.249129276787544, "grad_norm": 0.08963880451358089, "learning_rate": 5.358983848622452e-06, "loss": 0.5192, "step": 646 }, { "epoch": 4.255685310387216, "grad_norm": 0.0917773531851403, "learning_rate": 5.26749011949141e-06, "loss": 0.5112, "step": 647 }, { "epoch": 4.262241343986888, "grad_norm": 0.09650192405895454, "learning_rate": 5.176729084885508e-06, "loss": 0.5156, "step": 648 }, { "epoch": 4.26879737758656, "grad_norm": 0.09638202151344397, "learning_rate": 5.086702659440743e-06, "loss": 0.5132, "step": 649 }, { "epoch": 4.275353411186233, "grad_norm": 0.09515733205388084, "learning_rate": 4.99741274229625e-06, "loss": 0.5107, "step": 650 }, { "epoch": 4.281909444785905, "grad_norm": 0.08694276833101254, "learning_rate": 4.908861217054281e-06, "loss": 0.511, "step": 651 }, { "epoch": 4.288465478385577, "grad_norm": 0.09365787216266765, "learning_rate": 4.821049951740442e-06, "loss": 0.5196, "step": 652 }, { "epoch": 4.295021511985249, "grad_norm": 0.09875335551778978, "learning_rate": 4.733980798764273e-06, "loss": 0.5139, "step": 653 }, { "epoch": 4.301577545584921, "grad_norm": 0.0984165654822695, "learning_rate": 4.647655594880225e-06, "loss": 0.5194, "step": 654 }, { "epoch": 4.308133579184593, "grad_norm": 0.08981515555040814, "learning_rate": 4.562076161148881e-06, "loss": 0.5159, "step": 655 }, { "epoch": 4.3146896127842655, "grad_norm": 0.08699993576971486, "learning_rate": 4.4772443028985004e-06, "loss": 0.5107, "step": 656 }, { "epoch": 4.321245646383938, "grad_norm": 0.09531375447276737, "learning_rate": 4.393161809687021e-06, "loss": 0.5211, "step": 657 }, { "epoch": 4.32780167998361, "grad_norm": 0.16067767992274057, "learning_rate": 4.3098304552642385e-06, "loss": 0.5163, "step": 658 }, { "epoch": 4.334357713583282, "grad_norm": 0.09069517004519909, "learning_rate": 4.227251997534416e-06, "loss": 0.5107, "step": 659 }, { "epoch": 4.340913747182954, "grad_norm": 0.08688482436943581, "learning_rate": 4.1454281785191995e-06, "loss": 0.5118, "step": 660 }, { "epoch": 4.347469780782626, "grad_norm": 0.09109521686027529, "learning_rate": 4.064360724320846e-06, "loss": 0.5114, "step": 661 }, { "epoch": 4.3540258143822985, "grad_norm": 0.0885789818679224, "learning_rate": 3.984051345085855e-06, "loss": 0.5126, "step": 662 }, { "epoch": 4.360581847981971, "grad_norm": 0.09379559333937212, "learning_rate": 3.90450173496887e-06, "loss": 0.5207, "step": 663 }, { "epoch": 4.367137881581643, "grad_norm": 0.08937679173054215, "learning_rate": 3.825713572096903e-06, "loss": 0.5175, "step": 664 }, { "epoch": 4.373693915181315, "grad_norm": 0.08924834510849128, "learning_rate": 3.747688518534003e-06, "loss": 0.5186, "step": 665 }, { "epoch": 4.380249948780987, "grad_norm": 0.09109376121051856, "learning_rate": 3.6704282202461515e-06, "loss": 0.5161, "step": 666 }, { "epoch": 4.386805982380659, "grad_norm": 0.10119588536048828, "learning_rate": 3.5939343070665243e-06, "loss": 0.5032, "step": 667 }, { "epoch": 4.3933620159803315, "grad_norm": 0.09644206288351137, "learning_rate": 3.518208392661184e-06, "loss": 0.5188, "step": 668 }, { "epoch": 4.399918049580004, "grad_norm": 0.08778658376932175, "learning_rate": 3.4432520744949317e-06, "loss": 0.5133, "step": 669 }, { "epoch": 4.406474083179677, "grad_norm": 0.0941601011641409, "learning_rate": 3.3690669337977e-06, "loss": 0.5061, "step": 670 }, { "epoch": 4.413030116779349, "grad_norm": 0.09791658312405495, "learning_rate": 3.295654535531161e-06, "loss": 0.5193, "step": 671 }, { "epoch": 4.419586150379021, "grad_norm": 0.09354050139551012, "learning_rate": 3.2230164283556918e-06, "loss": 0.5172, "step": 672 }, { "epoch": 4.426142183978693, "grad_norm": 0.08893737381738466, "learning_rate": 3.151154144597741e-06, "loss": 0.5111, "step": 673 }, { "epoch": 4.432698217578365, "grad_norm": 0.08445332878405215, "learning_rate": 3.080069200217497e-06, "loss": 0.5097, "step": 674 }, { "epoch": 4.4392542511780375, "grad_norm": 0.08801285298434917, "learning_rate": 3.0097630947768695e-06, "loss": 0.5135, "step": 675 }, { "epoch": 4.44581028477771, "grad_norm": 0.09402581699959618, "learning_rate": 2.9402373114079295e-06, "loss": 0.5086, "step": 676 }, { "epoch": 4.452366318377382, "grad_norm": 0.0905312100821144, "learning_rate": 2.871493316781546e-06, "loss": 0.515, "step": 677 }, { "epoch": 4.458922351977054, "grad_norm": 0.09633255689013458, "learning_rate": 2.803532561076492e-06, "loss": 0.5123, "step": 678 }, { "epoch": 4.465478385576726, "grad_norm": 0.08803341472121784, "learning_rate": 2.7363564779488448e-06, "loss": 0.5135, "step": 679 }, { "epoch": 4.472034419176398, "grad_norm": 0.08797758072438489, "learning_rate": 2.669966484501716e-06, "loss": 0.5199, "step": 680 }, { "epoch": 4.4785904527760705, "grad_norm": 0.09112213281433264, "learning_rate": 2.6043639812554043e-06, "loss": 0.5205, "step": 681 }, { "epoch": 4.485146486375743, "grad_norm": 0.09544750033107532, "learning_rate": 2.5395503521178143e-06, "loss": 0.5253, "step": 682 }, { "epoch": 4.491702519975415, "grad_norm": 0.09376536821854904, "learning_rate": 2.4755269643552594e-06, "loss": 0.5164, "step": 683 }, { "epoch": 4.498258553575087, "grad_norm": 0.09454989320113857, "learning_rate": 2.4122951685636674e-06, "loss": 0.5187, "step": 684 }, { "epoch": 4.504814587174759, "grad_norm": 0.08421901449177807, "learning_rate": 2.3498562986400187e-06, "loss": 0.5154, "step": 685 }, { "epoch": 4.511370620774431, "grad_norm": 0.08753626044860159, "learning_rate": 2.2882116717542634e-06, "loss": 0.5176, "step": 686 }, { "epoch": 4.5179266543741035, "grad_norm": 0.0907417635028892, "learning_rate": 2.22736258832152e-06, "loss": 0.5244, "step": 687 }, { "epoch": 4.524482687973776, "grad_norm": 0.08866763487206057, "learning_rate": 2.1673103319746146e-06, "loss": 0.5188, "step": 688 }, { "epoch": 4.531038721573448, "grad_norm": 0.08406194293605605, "learning_rate": 2.1080561695370425e-06, "loss": 0.5117, "step": 689 }, { "epoch": 4.53759475517312, "grad_norm": 0.08781306185528281, "learning_rate": 2.049601350996233e-06, "loss": 0.5105, "step": 690 }, { "epoch": 4.544150788772792, "grad_norm": 0.0862714841925732, "learning_rate": 1.9919471094771523e-06, "loss": 0.513, "step": 691 }, { "epoch": 4.550706822372465, "grad_norm": 0.08316320143150852, "learning_rate": 1.93509466121633e-06, "loss": 0.5143, "step": 692 }, { "epoch": 4.557262855972137, "grad_norm": 0.08459303029799788, "learning_rate": 1.8790452055361764e-06, "loss": 0.5117, "step": 693 }, { "epoch": 4.5638188895718095, "grad_norm": 0.08646962847262735, "learning_rate": 1.8237999248197002e-06, "loss": 0.5122, "step": 694 }, { "epoch": 4.570374923171482, "grad_norm": 0.18896054046080968, "learning_rate": 1.7693599844855568e-06, "loss": 0.5187, "step": 695 }, { "epoch": 4.576930956771154, "grad_norm": 0.08583158851506041, "learning_rate": 1.7157265329634354e-06, "loss": 0.5206, "step": 696 }, { "epoch": 4.583486990370826, "grad_norm": 0.0823819160891982, "learning_rate": 1.6629007016698918e-06, "loss": 0.5119, "step": 697 }, { "epoch": 4.590043023970498, "grad_norm": 0.0856122092163256, "learning_rate": 1.6108836049844434e-06, "loss": 0.5192, "step": 698 }, { "epoch": 4.59659905757017, "grad_norm": 0.08366071868253996, "learning_rate": 1.5596763402260462e-06, "loss": 0.5169, "step": 699 }, { "epoch": 4.6031550911698424, "grad_norm": 0.08756847365451516, "learning_rate": 1.5092799876299835e-06, "loss": 0.5142, "step": 700 }, { "epoch": 4.609711124769515, "grad_norm": 0.08472854571793725, "learning_rate": 1.459695610325067e-06, "loss": 0.5191, "step": 701 }, { "epoch": 4.616267158369187, "grad_norm": 0.08619093020176126, "learning_rate": 1.4109242543111834e-06, "loss": 0.5166, "step": 702 }, { "epoch": 4.622823191968859, "grad_norm": 0.08301363626424127, "learning_rate": 1.3629669484372722e-06, "loss": 0.517, "step": 703 }, { "epoch": 4.629379225568531, "grad_norm": 0.08230801802672034, "learning_rate": 1.3158247043795735e-06, "loss": 0.5033, "step": 704 }, { "epoch": 4.635935259168203, "grad_norm": 0.08247077823518248, "learning_rate": 1.2694985166203311e-06, "loss": 0.523, "step": 705 }, { "epoch": 4.642491292767875, "grad_norm": 0.08077290324213217, "learning_rate": 1.2239893624267852e-06, "loss": 0.5218, "step": 706 }, { "epoch": 4.649047326367548, "grad_norm": 0.08354463750730508, "learning_rate": 1.1792982018305677e-06, "loss": 0.5143, "step": 707 }, { "epoch": 4.65560335996722, "grad_norm": 0.08393485609493388, "learning_rate": 1.1354259776074472e-06, "loss": 0.5151, "step": 708 }, { "epoch": 4.662159393566892, "grad_norm": 0.0832402976978181, "learning_rate": 1.0923736152574428e-06, "loss": 0.5233, "step": 709 }, { "epoch": 4.668715427166564, "grad_norm": 0.0835585475483202, "learning_rate": 1.050142022985292e-06, "loss": 0.5093, "step": 710 }, { "epoch": 4.675271460766236, "grad_norm": 0.08176206607867041, "learning_rate": 1.0087320916813127e-06, "loss": 0.5225, "step": 711 }, { "epoch": 4.681827494365908, "grad_norm": 0.08752232597844366, "learning_rate": 9.681446949025752e-07, "loss": 0.5156, "step": 712 }, { "epoch": 4.6883835279655806, "grad_norm": 0.08108785189147111, "learning_rate": 9.283806888545111e-07, "loss": 0.5138, "step": 713 }, { "epoch": 4.694939561565253, "grad_norm": 0.08223618174467849, "learning_rate": 8.89440912372832e-07, "loss": 0.5172, "step": 714 }, { "epoch": 4.701495595164925, "grad_norm": 0.08351263590200621, "learning_rate": 8.513261869058209e-07, "loss": 0.5282, "step": 715 }, { "epoch": 4.708051628764597, "grad_norm": 0.08286053605490928, "learning_rate": 8.140373164970428e-07, "loss": 0.5244, "step": 716 }, { "epoch": 4.714607662364269, "grad_norm": 0.08518895529322469, "learning_rate": 7.775750877683452e-07, "loss": 0.5171, "step": 717 }, { "epoch": 4.721163695963941, "grad_norm": 0.0813459247213106, "learning_rate": 7.419402699032852e-07, "loss": 0.5186, "step": 718 }, { "epoch": 4.7277197295636135, "grad_norm": 0.0803887295459458, "learning_rate": 7.071336146308883e-07, "loss": 0.5059, "step": 719 }, { "epoch": 4.734275763163287, "grad_norm": 0.08172953576855356, "learning_rate": 6.731558562097995e-07, "loss": 0.5224, "step": 720 }, { "epoch": 4.740831796762959, "grad_norm": 0.08549688805268194, "learning_rate": 6.400077114128023e-07, "loss": 0.5165, "step": 721 }, { "epoch": 4.747387830362631, "grad_norm": 0.08286626015188583, "learning_rate": 6.076898795116792e-07, "loss": 0.51, "step": 722 }, { "epoch": 4.753943863962303, "grad_norm": 0.0823372680890407, "learning_rate": 5.762030422624732e-07, "loss": 0.5212, "step": 723 }, { "epoch": 4.760499897561975, "grad_norm": 0.08161043456784334, "learning_rate": 5.455478638911071e-07, "loss": 0.5116, "step": 724 }, { "epoch": 4.767055931161647, "grad_norm": 0.08090694406998772, "learning_rate": 5.15724991079356e-07, "loss": 0.5223, "step": 725 }, { "epoch": 4.7736119647613195, "grad_norm": 0.08130626566721383, "learning_rate": 4.867350529512261e-07, "loss": 0.5104, "step": 726 }, { "epoch": 4.780167998360992, "grad_norm": 0.08487388342418309, "learning_rate": 4.5857866105966763e-07, "loss": 0.5129, "step": 727 }, { "epoch": 4.786724031960664, "grad_norm": 0.0816069016289385, "learning_rate": 4.3125640937368373e-07, "loss": 0.5123, "step": 728 }, { "epoch": 4.793280065560336, "grad_norm": 0.08049465996969124, "learning_rate": 4.047688742657885e-07, "loss": 0.5063, "step": 729 }, { "epoch": 4.799836099160008, "grad_norm": 0.0830575657951303, "learning_rate": 3.791166144998704e-07, "loss": 0.5121, "step": 730 }, { "epoch": 4.80639213275968, "grad_norm": 0.08387321061054255, "learning_rate": 3.54300171219375e-07, "loss": 0.5146, "step": 731 }, { "epoch": 4.8129481663593525, "grad_norm": 0.08224120804212026, "learning_rate": 3.3032006793590977e-07, "loss": 0.5206, "step": 732 }, { "epoch": 4.819504199959025, "grad_norm": 0.08072502199031102, "learning_rate": 3.0717681051819935e-07, "loss": 0.522, "step": 733 }, { "epoch": 4.826060233558697, "grad_norm": 0.07860877555923813, "learning_rate": 2.848708871814054e-07, "loss": 0.5188, "step": 734 }, { "epoch": 4.832616267158369, "grad_norm": 0.08001404199090854, "learning_rate": 2.634027684768414e-07, "loss": 0.5124, "step": 735 }, { "epoch": 4.839172300758041, "grad_norm": 0.08310618061839851, "learning_rate": 2.4277290728202063e-07, "loss": 0.5212, "step": 736 }, { "epoch": 4.845728334357713, "grad_norm": 0.0797549571951669, "learning_rate": 2.2298173879113481e-07, "loss": 0.5168, "step": 737 }, { "epoch": 4.8522843679573855, "grad_norm": 0.0824361700062477, "learning_rate": 2.040296805058528e-07, "loss": 0.5158, "step": 738 }, { "epoch": 4.858840401557058, "grad_norm": 0.08076093089733337, "learning_rate": 1.859171322265141e-07, "loss": 0.5253, "step": 739 }, { "epoch": 4.86539643515673, "grad_norm": 0.08043719703067892, "learning_rate": 1.6864447604370004e-07, "loss": 0.5244, "step": 740 }, { "epoch": 4.871952468756403, "grad_norm": 0.08277960997697664, "learning_rate": 1.522120763301782e-07, "loss": 0.5085, "step": 741 }, { "epoch": 4.878508502356075, "grad_norm": 0.08180092674268333, "learning_rate": 1.3662027973320614e-07, "loss": 0.5202, "step": 742 }, { "epoch": 4.885064535955747, "grad_norm": 0.08183966123917269, "learning_rate": 1.2186941516722173e-07, "loss": 0.5231, "step": 743 }, { "epoch": 4.891620569555419, "grad_norm": 0.07963446358425141, "learning_rate": 1.0795979380690657e-07, "loss": 0.5136, "step": 744 }, { "epoch": 4.8981766031550915, "grad_norm": 0.07852420651092652, "learning_rate": 9.489170908062228e-08, "loss": 0.5204, "step": 745 }, { "epoch": 4.904732636754764, "grad_norm": 0.12877804510272148, "learning_rate": 8.266543666421544e-08, "loss": 0.5185, "step": 746 }, { "epoch": 4.911288670354436, "grad_norm": 0.0869160748356661, "learning_rate": 7.128123447520452e-08, "loss": 0.5234, "step": 747 }, { "epoch": 4.917844703954108, "grad_norm": 0.08086821677688971, "learning_rate": 6.073934266735303e-08, "loss": 0.5129, "step": 748 }, { "epoch": 4.92440073755378, "grad_norm": 0.07880134225937953, "learning_rate": 5.10399836255715e-08, "loss": 0.5064, "step": 749 }, { "epoch": 4.930956771153452, "grad_norm": 0.07797418195960645, "learning_rate": 4.218336196125439e-08, "loss": 0.5106, "step": 750 }, { "epoch": 4.9375128047531245, "grad_norm": 0.08006201262285005, "learning_rate": 3.416966450795922e-08, "loss": 0.5099, "step": 751 }, { "epoch": 4.944068838352797, "grad_norm": 0.08059149745693742, "learning_rate": 2.699906031745414e-08, "loss": 0.5176, "step": 752 }, { "epoch": 4.950624871952469, "grad_norm": 0.08251393526468398, "learning_rate": 2.067170065615187e-08, "loss": 0.5235, "step": 753 }, { "epoch": 4.957180905552141, "grad_norm": 0.08234699138692923, "learning_rate": 1.5187719001943378e-08, "loss": 0.5122, "step": 754 }, { "epoch": 4.963736939151813, "grad_norm": 0.08191620156821956, "learning_rate": 1.0547231041346806e-08, "loss": 0.5237, "step": 755 }, { "epoch": 4.970292972751485, "grad_norm": 0.08084676979927179, "learning_rate": 6.750334667091629e-09, "loss": 0.5178, "step": 756 }, { "epoch": 4.9768490063511575, "grad_norm": 0.08117469896416148, "learning_rate": 3.797109976035884e-09, "loss": 0.5192, "step": 757 }, { "epoch": 4.98340503995083, "grad_norm": 0.0821326486604627, "learning_rate": 1.6876192675052695e-09, "loss": 0.519, "step": 758 }, { "epoch": 4.989961073550502, "grad_norm": 0.08465924625593949, "learning_rate": 4.219070419475557e-10, "loss": 0.5078, "step": 759 }, { "epoch": 4.996517107150174, "grad_norm": 0.08042205253823026, "learning_rate": 0.0, "loss": 0.5193, "step": 760 }, { "epoch": 4.996517107150174, "step": 760, "total_flos": 2.0225694010775175e+19, "train_loss": 0.10313788385767686, "train_runtime": 35636.9099, "train_samples_per_second": 10.956, "train_steps_per_second": 0.021 } ], "logging_steps": 1, "max_steps": 760, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.0225694010775175e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }