{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 4954, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002018927444794953, "grad_norm": 2.3389563305266345, "learning_rate": 1.8145161290322583e-07, "loss": 0.7881, "step": 10 }, { "epoch": 0.004037854889589906, "grad_norm": 2.024167058240825, "learning_rate": 3.830645161290323e-07, "loss": 0.8121, "step": 20 }, { "epoch": 0.006056782334384858, "grad_norm": 1.4351269819866417, "learning_rate": 5.846774193548388e-07, "loss": 0.7797, "step": 30 }, { "epoch": 0.008075709779179811, "grad_norm": 0.9310252821124614, "learning_rate": 7.862903225806453e-07, "loss": 0.7318, "step": 40 }, { "epoch": 0.010094637223974764, "grad_norm": 0.9557517206875442, "learning_rate": 9.879032258064516e-07, "loss": 0.7018, "step": 50 }, { "epoch": 0.012113564668769715, "grad_norm": 0.5829217778568054, "learning_rate": 1.1895161290322582e-06, "loss": 0.6708, "step": 60 }, { "epoch": 0.014132492113564668, "grad_norm": 0.5112881851342524, "learning_rate": 1.3911290322580644e-06, "loss": 0.6417, "step": 70 }, { "epoch": 0.016151419558359623, "grad_norm": 0.5140187941233466, "learning_rate": 1.5927419354838712e-06, "loss": 0.6227, "step": 80 }, { "epoch": 0.018170347003154574, "grad_norm": 0.4872446144280347, "learning_rate": 1.7943548387096777e-06, "loss": 0.615, "step": 90 }, { "epoch": 0.02018927444794953, "grad_norm": 0.5191969395080858, "learning_rate": 1.9959677419354837e-06, "loss": 0.6105, "step": 100 }, { "epoch": 0.02220820189274448, "grad_norm": 0.451802382492337, "learning_rate": 2.1975806451612907e-06, "loss": 0.5911, "step": 110 }, { "epoch": 0.02422712933753943, "grad_norm": 0.42163928465445194, "learning_rate": 2.399193548387097e-06, "loss": 0.5921, "step": 120 }, { "epoch": 0.026246056782334385, "grad_norm": 0.4552027864136731, "learning_rate": 2.6008064516129032e-06, "loss": 0.5818, "step": 130 }, { "epoch": 0.028264984227129336, "grad_norm": 0.44985963500335385, "learning_rate": 2.8024193548387097e-06, "loss": 0.5876, "step": 140 }, { "epoch": 0.03028391167192429, "grad_norm": 0.5111630527751778, "learning_rate": 3.0040322580645167e-06, "loss": 0.5728, "step": 150 }, { "epoch": 0.032302839116719245, "grad_norm": 0.44240607593167414, "learning_rate": 3.2056451612903228e-06, "loss": 0.5643, "step": 160 }, { "epoch": 0.03432176656151419, "grad_norm": 0.4865981695157576, "learning_rate": 3.4072580645161293e-06, "loss": 0.5617, "step": 170 }, { "epoch": 0.03634069400630915, "grad_norm": 0.6085288533687223, "learning_rate": 3.6088709677419358e-06, "loss": 0.5501, "step": 180 }, { "epoch": 0.0383596214511041, "grad_norm": 0.47289717909858964, "learning_rate": 3.8104838709677423e-06, "loss": 0.559, "step": 190 }, { "epoch": 0.04037854889589906, "grad_norm": 0.5389683685989423, "learning_rate": 4.012096774193548e-06, "loss": 0.5601, "step": 200 }, { "epoch": 0.042397476340694004, "grad_norm": 0.44905068010794785, "learning_rate": 4.213709677419355e-06, "loss": 0.546, "step": 210 }, { "epoch": 0.04441640378548896, "grad_norm": 0.4738265544351665, "learning_rate": 4.415322580645161e-06, "loss": 0.5527, "step": 220 }, { "epoch": 0.04643533123028391, "grad_norm": 0.48116551901846, "learning_rate": 4.616935483870968e-06, "loss": 0.5428, "step": 230 }, { "epoch": 0.04845425867507886, "grad_norm": 0.4514773717137442, "learning_rate": 4.818548387096774e-06, "loss": 0.5338, "step": 240 }, { "epoch": 0.050473186119873815, "grad_norm": 0.5092040454559039, "learning_rate": 5.020161290322581e-06, "loss": 0.5546, "step": 250 }, { "epoch": 0.05249211356466877, "grad_norm": 0.45858657596184155, "learning_rate": 5.221774193548388e-06, "loss": 0.5408, "step": 260 }, { "epoch": 0.054511041009463725, "grad_norm": 0.47766211942842146, "learning_rate": 5.423387096774194e-06, "loss": 0.5427, "step": 270 }, { "epoch": 0.05652996845425867, "grad_norm": 0.5108229913280311, "learning_rate": 5.625e-06, "loss": 0.5357, "step": 280 }, { "epoch": 0.05854889589905363, "grad_norm": 0.5232937676969737, "learning_rate": 5.8266129032258064e-06, "loss": 0.5316, "step": 290 }, { "epoch": 0.06056782334384858, "grad_norm": 0.5241114539289579, "learning_rate": 6.028225806451613e-06, "loss": 0.5444, "step": 300 }, { "epoch": 0.06258675078864354, "grad_norm": 0.9566252867340465, "learning_rate": 6.2298387096774194e-06, "loss": 0.5256, "step": 310 }, { "epoch": 0.06460567823343849, "grad_norm": 0.4700069264487592, "learning_rate": 6.431451612903226e-06, "loss": 0.535, "step": 320 }, { "epoch": 0.06662460567823345, "grad_norm": 0.46343674018411934, "learning_rate": 6.633064516129033e-06, "loss": 0.5281, "step": 330 }, { "epoch": 0.06864353312302839, "grad_norm": 0.46610206235853135, "learning_rate": 6.834677419354839e-06, "loss": 0.5324, "step": 340 }, { "epoch": 0.07066246056782334, "grad_norm": 0.5217091961421014, "learning_rate": 7.0362903225806454e-06, "loss": 0.5321, "step": 350 }, { "epoch": 0.0726813880126183, "grad_norm": 0.4532872443875366, "learning_rate": 7.2379032258064515e-06, "loss": 0.5235, "step": 360 }, { "epoch": 0.07470031545741325, "grad_norm": 0.5481191485070994, "learning_rate": 7.4395161290322585e-06, "loss": 0.5203, "step": 370 }, { "epoch": 0.0767192429022082, "grad_norm": 0.4817884817896886, "learning_rate": 7.641129032258065e-06, "loss": 0.5289, "step": 380 }, { "epoch": 0.07873817034700316, "grad_norm": 0.463062633118143, "learning_rate": 7.842741935483872e-06, "loss": 0.5199, "step": 390 }, { "epoch": 0.08075709779179811, "grad_norm": 0.5267129255916911, "learning_rate": 8.044354838709678e-06, "loss": 0.5239, "step": 400 }, { "epoch": 0.08277602523659305, "grad_norm": 0.4703183395271576, "learning_rate": 8.245967741935484e-06, "loss": 0.5217, "step": 410 }, { "epoch": 0.08479495268138801, "grad_norm": 0.4559407675951848, "learning_rate": 8.447580645161291e-06, "loss": 0.5187, "step": 420 }, { "epoch": 0.08681388012618296, "grad_norm": 0.4638358181971901, "learning_rate": 8.649193548387097e-06, "loss": 0.5194, "step": 430 }, { "epoch": 0.08883280757097792, "grad_norm": 0.47768712570505284, "learning_rate": 8.850806451612905e-06, "loss": 0.513, "step": 440 }, { "epoch": 0.09085173501577287, "grad_norm": 0.5140723294282747, "learning_rate": 9.05241935483871e-06, "loss": 0.5161, "step": 450 }, { "epoch": 0.09287066246056783, "grad_norm": 0.4719590441343577, "learning_rate": 9.254032258064517e-06, "loss": 0.5033, "step": 460 }, { "epoch": 0.09488958990536278, "grad_norm": 0.5031363320062582, "learning_rate": 9.455645161290323e-06, "loss": 0.5145, "step": 470 }, { "epoch": 0.09690851735015772, "grad_norm": 0.5027908604602157, "learning_rate": 9.65725806451613e-06, "loss": 0.5178, "step": 480 }, { "epoch": 0.09892744479495268, "grad_norm": 0.5046628095943158, "learning_rate": 9.858870967741936e-06, "loss": 0.5069, "step": 490 }, { "epoch": 0.10094637223974763, "grad_norm": 0.5522780839548007, "learning_rate": 9.99998882617222e-06, "loss": 0.5169, "step": 500 }, { "epoch": 0.10296529968454259, "grad_norm": 0.48260445699138166, "learning_rate": 9.999790181734308e-06, "loss": 0.5067, "step": 510 }, { "epoch": 0.10498422712933754, "grad_norm": 0.4875096921972213, "learning_rate": 9.999343241367322e-06, "loss": 0.5152, "step": 520 }, { "epoch": 0.1070031545741325, "grad_norm": 0.5080882476390746, "learning_rate": 9.998648027266884e-06, "loss": 0.5131, "step": 530 }, { "epoch": 0.10902208201892745, "grad_norm": 0.5107719615388011, "learning_rate": 9.997704573958209e-06, "loss": 0.5053, "step": 540 }, { "epoch": 0.1110410094637224, "grad_norm": 0.5361602322829081, "learning_rate": 9.996512928294389e-06, "loss": 0.5083, "step": 550 }, { "epoch": 0.11305993690851734, "grad_norm": 0.4851941991561818, "learning_rate": 9.995073149454057e-06, "loss": 0.4997, "step": 560 }, { "epoch": 0.1150788643533123, "grad_norm": 0.49258962377904003, "learning_rate": 9.99338530893846e-06, "loss": 0.516, "step": 570 }, { "epoch": 0.11709779179810725, "grad_norm": 0.4835880405802275, "learning_rate": 9.991449490567901e-06, "loss": 0.5073, "step": 580 }, { "epoch": 0.11911671924290221, "grad_norm": 0.507245673357914, "learning_rate": 9.989265790477581e-06, "loss": 0.5014, "step": 590 }, { "epoch": 0.12113564668769716, "grad_norm": 0.5146575464998254, "learning_rate": 9.986834317112817e-06, "loss": 0.5177, "step": 600 }, { "epoch": 0.12315457413249212, "grad_norm": 0.4662328996304102, "learning_rate": 9.984155191223663e-06, "loss": 0.5054, "step": 610 }, { "epoch": 0.12517350157728707, "grad_norm": 0.4571893765964666, "learning_rate": 9.981228545858913e-06, "loss": 0.5086, "step": 620 }, { "epoch": 0.12719242902208203, "grad_norm": 0.4697827268486345, "learning_rate": 9.978054526359493e-06, "loss": 0.5029, "step": 630 }, { "epoch": 0.12921135646687698, "grad_norm": 0.5031062337107917, "learning_rate": 9.97463329035124e-06, "loss": 0.5021, "step": 640 }, { "epoch": 0.13123028391167194, "grad_norm": 0.571029993720401, "learning_rate": 9.970965007737081e-06, "loss": 0.5077, "step": 650 }, { "epoch": 0.1332492113564669, "grad_norm": 0.45753840087765046, "learning_rate": 9.967049860688587e-06, "loss": 0.4961, "step": 660 }, { "epoch": 0.13526813880126182, "grad_norm": 0.5189371422618692, "learning_rate": 9.962888043636931e-06, "loss": 0.5036, "step": 670 }, { "epoch": 0.13728706624605677, "grad_norm": 0.47066047930695143, "learning_rate": 9.958479763263234e-06, "loss": 0.4972, "step": 680 }, { "epoch": 0.13930599369085173, "grad_norm": 0.49380951082560154, "learning_rate": 9.953825238488296e-06, "loss": 0.4955, "step": 690 }, { "epoch": 0.14132492113564668, "grad_norm": 0.49487805704511495, "learning_rate": 9.948924700461727e-06, "loss": 0.5013, "step": 700 }, { "epoch": 0.14334384858044164, "grad_norm": 0.4768179474157311, "learning_rate": 9.94377839255047e-06, "loss": 0.4945, "step": 710 }, { "epoch": 0.1453627760252366, "grad_norm": 0.5086022999656893, "learning_rate": 9.938386570326707e-06, "loss": 0.5033, "step": 720 }, { "epoch": 0.14738170347003154, "grad_norm": 0.4947936431118982, "learning_rate": 9.93274950155518e-06, "loss": 0.5013, "step": 730 }, { "epoch": 0.1494006309148265, "grad_norm": 0.49370518434763616, "learning_rate": 9.926867466179883e-06, "loss": 0.4923, "step": 740 }, { "epoch": 0.15141955835962145, "grad_norm": 0.49665569470371385, "learning_rate": 9.920740756310156e-06, "loss": 0.5038, "step": 750 }, { "epoch": 0.1534384858044164, "grad_norm": 0.46813454785234254, "learning_rate": 9.914369676206195e-06, "loss": 0.4875, "step": 760 }, { "epoch": 0.15545741324921136, "grad_norm": 0.49814972796278695, "learning_rate": 9.907754542263926e-06, "loss": 0.4939, "step": 770 }, { "epoch": 0.15747634069400632, "grad_norm": 0.5218798292518447, "learning_rate": 9.900895682999301e-06, "loss": 0.501, "step": 780 }, { "epoch": 0.15949526813880127, "grad_norm": 0.5913103279447773, "learning_rate": 9.893793439031982e-06, "loss": 0.4927, "step": 790 }, { "epoch": 0.16151419558359623, "grad_norm": 0.5125446596453367, "learning_rate": 9.88644816306842e-06, "loss": 0.4971, "step": 800 }, { "epoch": 0.16353312302839118, "grad_norm": 0.4483762668624327, "learning_rate": 9.878860219884347e-06, "loss": 0.4981, "step": 810 }, { "epoch": 0.1655520504731861, "grad_norm": 0.4787860866647759, "learning_rate": 9.871029986306658e-06, "loss": 0.4964, "step": 820 }, { "epoch": 0.16757097791798106, "grad_norm": 0.5030394247546482, "learning_rate": 9.862957851194694e-06, "loss": 0.4891, "step": 830 }, { "epoch": 0.16958990536277602, "grad_norm": 0.5239985323894606, "learning_rate": 9.854644215420933e-06, "loss": 0.501, "step": 840 }, { "epoch": 0.17160883280757097, "grad_norm": 0.4934736601735223, "learning_rate": 9.846089491851089e-06, "loss": 0.5012, "step": 850 }, { "epoch": 0.17362776025236593, "grad_norm": 0.5649851363661309, "learning_rate": 9.837294105323592e-06, "loss": 0.501, "step": 860 }, { "epoch": 0.17564668769716088, "grad_norm": 0.5032183289408247, "learning_rate": 9.828258492628506e-06, "loss": 0.5019, "step": 870 }, { "epoch": 0.17766561514195583, "grad_norm": 0.5408394773542722, "learning_rate": 9.818983102485832e-06, "loss": 0.4914, "step": 880 }, { "epoch": 0.1796845425867508, "grad_norm": 0.47080262308715565, "learning_rate": 9.809468395523225e-06, "loss": 0.4874, "step": 890 }, { "epoch": 0.18170347003154574, "grad_norm": 0.5707749159791521, "learning_rate": 9.799714844253109e-06, "loss": 0.4995, "step": 900 }, { "epoch": 0.1837223974763407, "grad_norm": 0.48122327046673424, "learning_rate": 9.789722933049227e-06, "loss": 0.4892, "step": 910 }, { "epoch": 0.18574132492113565, "grad_norm": 0.45082444705324104, "learning_rate": 9.779493158122578e-06, "loss": 0.4909, "step": 920 }, { "epoch": 0.1877602523659306, "grad_norm": 0.5068314998947344, "learning_rate": 9.769026027496773e-06, "loss": 0.4881, "step": 930 }, { "epoch": 0.18977917981072556, "grad_norm": 0.44089908003969003, "learning_rate": 9.758322060982813e-06, "loss": 0.4923, "step": 940 }, { "epoch": 0.19179810725552052, "grad_norm": 0.4786006068183123, "learning_rate": 9.747381790153262e-06, "loss": 0.4912, "step": 950 }, { "epoch": 0.19381703470031544, "grad_norm": 0.4531421936277408, "learning_rate": 9.736205758315864e-06, "loss": 0.4744, "step": 960 }, { "epoch": 0.1958359621451104, "grad_norm": 0.4492650045467347, "learning_rate": 9.724794520486551e-06, "loss": 0.485, "step": 970 }, { "epoch": 0.19785488958990535, "grad_norm": 0.4616235423638888, "learning_rate": 9.713148643361884e-06, "loss": 0.488, "step": 980 }, { "epoch": 0.1998738170347003, "grad_norm": 0.5078516721505932, "learning_rate": 9.701268705290907e-06, "loss": 0.4886, "step": 990 }, { "epoch": 0.20189274447949526, "grad_norm": 0.5292871585531458, "learning_rate": 9.689155296246431e-06, "loss": 0.4897, "step": 1000 }, { "epoch": 0.20391167192429022, "grad_norm": 0.4614297555636878, "learning_rate": 9.67680901779573e-06, "loss": 0.4809, "step": 1010 }, { "epoch": 0.20593059936908517, "grad_norm": 0.4853770956696164, "learning_rate": 9.664230483070669e-06, "loss": 0.4848, "step": 1020 }, { "epoch": 0.20794952681388013, "grad_norm": 0.45954875393494693, "learning_rate": 9.651420316737254e-06, "loss": 0.4817, "step": 1030 }, { "epoch": 0.20996845425867508, "grad_norm": 0.4539790244117171, "learning_rate": 9.638379154964615e-06, "loss": 0.4927, "step": 1040 }, { "epoch": 0.21198738170347003, "grad_norm": 0.4999117191134038, "learning_rate": 9.625107645393398e-06, "loss": 0.4906, "step": 1050 }, { "epoch": 0.214006309148265, "grad_norm": 0.5315614747354604, "learning_rate": 9.611606447103625e-06, "loss": 0.4701, "step": 1060 }, { "epoch": 0.21602523659305994, "grad_norm": 0.4465543761029164, "learning_rate": 9.597876230581946e-06, "loss": 0.4844, "step": 1070 }, { "epoch": 0.2180441640378549, "grad_norm": 0.5240713468337838, "learning_rate": 9.583917677688348e-06, "loss": 0.4802, "step": 1080 }, { "epoch": 0.22006309148264985, "grad_norm": 0.4797396408083518, "learning_rate": 9.56973148162229e-06, "loss": 0.4822, "step": 1090 }, { "epoch": 0.2220820189274448, "grad_norm": 0.5124904557581254, "learning_rate": 9.555318346888287e-06, "loss": 0.4887, "step": 1100 }, { "epoch": 0.22410094637223973, "grad_norm": 0.448720333545146, "learning_rate": 9.540678989260907e-06, "loss": 0.4804, "step": 1110 }, { "epoch": 0.2261198738170347, "grad_norm": 0.4479157257263002, "learning_rate": 9.525814135749238e-06, "loss": 0.4776, "step": 1120 }, { "epoch": 0.22813880126182964, "grad_norm": 0.4630993156519356, "learning_rate": 9.510724524560783e-06, "loss": 0.4753, "step": 1130 }, { "epoch": 0.2301577287066246, "grad_norm": 0.46574160333254905, "learning_rate": 9.495410905064795e-06, "loss": 0.4802, "step": 1140 }, { "epoch": 0.23217665615141955, "grad_norm": 0.532351729710861, "learning_rate": 9.479874037755061e-06, "loss": 0.4826, "step": 1150 }, { "epoch": 0.2341955835962145, "grad_norm": 0.4629220502691385, "learning_rate": 9.46411469421214e-06, "loss": 0.4802, "step": 1160 }, { "epoch": 0.23621451104100946, "grad_norm": 0.459458980682449, "learning_rate": 9.448133657065045e-06, "loss": 0.4863, "step": 1170 }, { "epoch": 0.23823343848580442, "grad_norm": 0.4652002127480942, "learning_rate": 9.431931719952372e-06, "loss": 0.4785, "step": 1180 }, { "epoch": 0.24025236593059937, "grad_norm": 0.466540922054408, "learning_rate": 9.415509687482892e-06, "loss": 0.4719, "step": 1190 }, { "epoch": 0.24227129337539433, "grad_norm": 0.49341259273484933, "learning_rate": 9.398868375195587e-06, "loss": 0.4926, "step": 1200 }, { "epoch": 0.24429022082018928, "grad_norm": 0.4728385366149605, "learning_rate": 9.382008609519159e-06, "loss": 0.4741, "step": 1210 }, { "epoch": 0.24630914826498423, "grad_norm": 0.5027959432042377, "learning_rate": 9.364931227730978e-06, "loss": 0.4826, "step": 1220 }, { "epoch": 0.2483280757097792, "grad_norm": 0.4883346568815589, "learning_rate": 9.34763707791551e-06, "loss": 0.4837, "step": 1230 }, { "epoch": 0.25034700315457414, "grad_norm": 0.4689919758806841, "learning_rate": 9.330127018922195e-06, "loss": 0.4834, "step": 1240 }, { "epoch": 0.25236593059936907, "grad_norm": 0.5054719137784303, "learning_rate": 9.312401920322793e-06, "loss": 0.4858, "step": 1250 }, { "epoch": 0.25438485804416405, "grad_norm": 0.4829012063372003, "learning_rate": 9.294462662368211e-06, "loss": 0.4757, "step": 1260 }, { "epoch": 0.256403785488959, "grad_norm": 0.4627206332424085, "learning_rate": 9.276310135944777e-06, "loss": 0.4783, "step": 1270 }, { "epoch": 0.25842271293375396, "grad_norm": 0.4497460790448332, "learning_rate": 9.257945242530002e-06, "loss": 0.4785, "step": 1280 }, { "epoch": 0.2604416403785489, "grad_norm": 0.44985051561838785, "learning_rate": 9.239368894147812e-06, "loss": 0.4786, "step": 1290 }, { "epoch": 0.26246056782334387, "grad_norm": 0.48240095905903635, "learning_rate": 9.220582013323258e-06, "loss": 0.4807, "step": 1300 }, { "epoch": 0.2644794952681388, "grad_norm": 0.48132105969556493, "learning_rate": 9.201585533036696e-06, "loss": 0.4706, "step": 1310 }, { "epoch": 0.2664984227129338, "grad_norm": 0.46675857398369297, "learning_rate": 9.18238039667746e-06, "loss": 0.4879, "step": 1320 }, { "epoch": 0.2685173501577287, "grad_norm": 0.4665634722250575, "learning_rate": 9.162967557997003e-06, "loss": 0.4799, "step": 1330 }, { "epoch": 0.27053627760252363, "grad_norm": 0.4432315628923224, "learning_rate": 9.14334798106155e-06, "loss": 0.4757, "step": 1340 }, { "epoch": 0.2725552050473186, "grad_norm": 0.5000831797296084, "learning_rate": 9.123522640204198e-06, "loss": 0.4805, "step": 1350 }, { "epoch": 0.27457413249211354, "grad_norm": 0.4441933490286196, "learning_rate": 9.103492519976555e-06, "loss": 0.4769, "step": 1360 }, { "epoch": 0.2765930599369085, "grad_norm": 0.4805005293661477, "learning_rate": 9.083258615099819e-06, "loss": 0.4777, "step": 1370 }, { "epoch": 0.27861198738170345, "grad_norm": 0.4445460559749252, "learning_rate": 9.062821930415404e-06, "loss": 0.4728, "step": 1380 }, { "epoch": 0.28063091482649843, "grad_norm": 0.4333540273890526, "learning_rate": 9.042183480835018e-06, "loss": 0.4705, "step": 1390 }, { "epoch": 0.28264984227129336, "grad_norm": 0.48259999792627817, "learning_rate": 9.021344291290275e-06, "loss": 0.4782, "step": 1400 }, { "epoch": 0.28466876971608834, "grad_norm": 0.439829603884435, "learning_rate": 9.000305396681788e-06, "loss": 0.4705, "step": 1410 }, { "epoch": 0.28668769716088327, "grad_norm": 0.4776753948778827, "learning_rate": 8.97906784182778e-06, "loss": 0.4694, "step": 1420 }, { "epoch": 0.28870662460567825, "grad_norm": 0.49821375314109334, "learning_rate": 8.95763268141219e-06, "loss": 0.4698, "step": 1430 }, { "epoch": 0.2907255520504732, "grad_norm": 0.4626475218122892, "learning_rate": 8.936000979932304e-06, "loss": 0.4766, "step": 1440 }, { "epoch": 0.29274447949526816, "grad_norm": 0.4797061046529278, "learning_rate": 8.914173811645884e-06, "loss": 0.4773, "step": 1450 }, { "epoch": 0.2947634069400631, "grad_norm": 0.5067089149884695, "learning_rate": 8.892152260517824e-06, "loss": 0.4652, "step": 1460 }, { "epoch": 0.29678233438485807, "grad_norm": 0.5216910139465747, "learning_rate": 8.869937420166312e-06, "loss": 0.475, "step": 1470 }, { "epoch": 0.298801261829653, "grad_norm": 0.4655884354714463, "learning_rate": 8.847530393808534e-06, "loss": 0.4811, "step": 1480 }, { "epoch": 0.3008201892744479, "grad_norm": 0.47533704121654896, "learning_rate": 8.824932294205868e-06, "loss": 0.4723, "step": 1490 }, { "epoch": 0.3028391167192429, "grad_norm": 0.479585079565894, "learning_rate": 8.802144243608633e-06, "loss": 0.4728, "step": 1500 }, { "epoch": 0.30485804416403783, "grad_norm": 0.4338897481002412, "learning_rate": 8.779167373700359e-06, "loss": 0.4687, "step": 1510 }, { "epoch": 0.3068769716088328, "grad_norm": 0.51284609503485, "learning_rate": 8.756002825541583e-06, "loss": 0.4795, "step": 1520 }, { "epoch": 0.30889589905362774, "grad_norm": 0.5019515223957067, "learning_rate": 8.73265174951318e-06, "loss": 0.4782, "step": 1530 }, { "epoch": 0.3109148264984227, "grad_norm": 0.5764514697725889, "learning_rate": 8.709115305259233e-06, "loss": 0.4722, "step": 1540 }, { "epoch": 0.31293375394321765, "grad_norm": 0.48602580093856246, "learning_rate": 8.685394661629453e-06, "loss": 0.4764, "step": 1550 }, { "epoch": 0.31495268138801263, "grad_norm": 0.43106495742778833, "learning_rate": 8.661490996621125e-06, "loss": 0.4749, "step": 1560 }, { "epoch": 0.31697160883280756, "grad_norm": 0.4879400234119546, "learning_rate": 8.637405497320599e-06, "loss": 0.4685, "step": 1570 }, { "epoch": 0.31899053627760254, "grad_norm": 0.48144632266967335, "learning_rate": 8.61313935984436e-06, "loss": 0.473, "step": 1580 }, { "epoch": 0.32100946372239747, "grad_norm": 0.4824945838927638, "learning_rate": 8.588693789279605e-06, "loss": 0.467, "step": 1590 }, { "epoch": 0.32302839116719245, "grad_norm": 0.48854234491790616, "learning_rate": 8.56406999962441e-06, "loss": 0.4715, "step": 1600 }, { "epoch": 0.3250473186119874, "grad_norm": 0.43356661154496373, "learning_rate": 8.539269213727434e-06, "loss": 0.4643, "step": 1610 }, { "epoch": 0.32706624605678236, "grad_norm": 0.45038524364573934, "learning_rate": 8.514292663227196e-06, "loss": 0.4753, "step": 1620 }, { "epoch": 0.3290851735015773, "grad_norm": 0.5390076730569227, "learning_rate": 8.489141588490914e-06, "loss": 0.4692, "step": 1630 }, { "epoch": 0.3311041009463722, "grad_norm": 0.5166665541948204, "learning_rate": 8.46381723855289e-06, "loss": 0.4626, "step": 1640 }, { "epoch": 0.3331230283911672, "grad_norm": 0.4973742178832253, "learning_rate": 8.438320871052501e-06, "loss": 0.4722, "step": 1650 }, { "epoch": 0.3351419558359621, "grad_norm": 0.4625137579015689, "learning_rate": 8.412653752171732e-06, "loss": 0.4591, "step": 1660 }, { "epoch": 0.3371608832807571, "grad_norm": 0.48424933592035946, "learning_rate": 8.386817156572298e-06, "loss": 0.48, "step": 1670 }, { "epoch": 0.33917981072555203, "grad_norm": 0.4532582849486311, "learning_rate": 8.360812367332342e-06, "loss": 0.469, "step": 1680 }, { "epoch": 0.341198738170347, "grad_norm": 0.4456128521040015, "learning_rate": 8.33464067588272e-06, "loss": 0.4701, "step": 1690 }, { "epoch": 0.34321766561514194, "grad_norm": 0.47975070454468693, "learning_rate": 8.30830338194286e-06, "loss": 0.4742, "step": 1700 }, { "epoch": 0.3452365930599369, "grad_norm": 0.44758135515111036, "learning_rate": 8.281801793456223e-06, "loss": 0.4563, "step": 1710 }, { "epoch": 0.34725552050473185, "grad_norm": 0.539530429343652, "learning_rate": 8.255137226525346e-06, "loss": 0.4678, "step": 1720 }, { "epoch": 0.34927444794952683, "grad_norm": 0.44634229380095736, "learning_rate": 8.22831100534648e-06, "loss": 0.4688, "step": 1730 }, { "epoch": 0.35129337539432176, "grad_norm": 0.41960408937669463, "learning_rate": 8.201324462143827e-06, "loss": 0.4542, "step": 1740 }, { "epoch": 0.35331230283911674, "grad_norm": 0.4791531544406239, "learning_rate": 8.174178937103397e-06, "loss": 0.4743, "step": 1750 }, { "epoch": 0.35533123028391167, "grad_norm": 0.4352818312435405, "learning_rate": 8.14687577830643e-06, "loss": 0.4604, "step": 1760 }, { "epoch": 0.35735015772870665, "grad_norm": 0.4672874879206997, "learning_rate": 8.11941634166246e-06, "loss": 0.4712, "step": 1770 }, { "epoch": 0.3593690851735016, "grad_norm": 0.4637056308756207, "learning_rate": 8.09180199084198e-06, "loss": 0.4668, "step": 1780 }, { "epoch": 0.3613880126182965, "grad_norm": 0.44264459219210456, "learning_rate": 8.064034097208712e-06, "loss": 0.4596, "step": 1790 }, { "epoch": 0.3634069400630915, "grad_norm": 0.49462230789807315, "learning_rate": 8.036114039751516e-06, "loss": 0.4667, "step": 1800 }, { "epoch": 0.3654258675078864, "grad_norm": 0.4634309345247295, "learning_rate": 8.0080432050159e-06, "loss": 0.4604, "step": 1810 }, { "epoch": 0.3674447949526814, "grad_norm": 0.4999816402185204, "learning_rate": 7.979822987035157e-06, "loss": 0.4644, "step": 1820 }, { "epoch": 0.3694637223974763, "grad_norm": 0.47642327436037435, "learning_rate": 7.951454787261148e-06, "loss": 0.4677, "step": 1830 }, { "epoch": 0.3714826498422713, "grad_norm": 0.4953986824691425, "learning_rate": 7.922940014494693e-06, "loss": 0.4593, "step": 1840 }, { "epoch": 0.37350157728706623, "grad_norm": 0.4817430868786167, "learning_rate": 7.894280084815618e-06, "loss": 0.4689, "step": 1850 }, { "epoch": 0.3755205047318612, "grad_norm": 0.4356746961237441, "learning_rate": 7.86547642151242e-06, "loss": 0.4611, "step": 1860 }, { "epoch": 0.37753943217665614, "grad_norm": 0.4905696785040258, "learning_rate": 7.836530455011594e-06, "loss": 0.4647, "step": 1870 }, { "epoch": 0.3795583596214511, "grad_norm": 0.4730705035791906, "learning_rate": 7.807443622806592e-06, "loss": 0.4643, "step": 1880 }, { "epoch": 0.38157728706624605, "grad_norm": 0.4626721574097504, "learning_rate": 7.778217369386434e-06, "loss": 0.4621, "step": 1890 }, { "epoch": 0.38359621451104103, "grad_norm": 0.5095748907639542, "learning_rate": 7.748853146163978e-06, "loss": 0.4702, "step": 1900 }, { "epoch": 0.38561514195583596, "grad_norm": 0.45666908318697785, "learning_rate": 7.719352411403836e-06, "loss": 0.4584, "step": 1910 }, { "epoch": 0.3876340694006309, "grad_norm": 0.5831960484166556, "learning_rate": 7.68971663014995e-06, "loss": 0.4581, "step": 1920 }, { "epoch": 0.38965299684542587, "grad_norm": 0.43083136772614855, "learning_rate": 7.659947274152857e-06, "loss": 0.4595, "step": 1930 }, { "epoch": 0.3916719242902208, "grad_norm": 0.44527597409724967, "learning_rate": 7.630045821796567e-06, "loss": 0.4603, "step": 1940 }, { "epoch": 0.3936908517350158, "grad_norm": 0.5085813030698595, "learning_rate": 7.60001375802518e-06, "loss": 0.4726, "step": 1950 }, { "epoch": 0.3957097791798107, "grad_norm": 0.5516780680340057, "learning_rate": 7.5698525742691145e-06, "loss": 0.4553, "step": 1960 }, { "epoch": 0.3977287066246057, "grad_norm": 0.4473799858576117, "learning_rate": 7.5395637683710534e-06, "loss": 0.4652, "step": 1970 }, { "epoch": 0.3997476340694006, "grad_norm": 0.4594068744960154, "learning_rate": 7.509148844511556e-06, "loss": 0.4569, "step": 1980 }, { "epoch": 0.4017665615141956, "grad_norm": 0.4195303708920141, "learning_rate": 7.478609313134367e-06, "loss": 0.4608, "step": 1990 }, { "epoch": 0.4037854889589905, "grad_norm": 0.5224611450690858, "learning_rate": 7.447946690871387e-06, "loss": 0.4618, "step": 2000 }, { "epoch": 0.4058044164037855, "grad_norm": 0.5037868351130359, "learning_rate": 7.4171625004673765e-06, "loss": 0.4549, "step": 2010 }, { "epoch": 0.40782334384858043, "grad_norm": 0.6617292048565804, "learning_rate": 7.386258270704318e-06, "loss": 0.4648, "step": 2020 }, { "epoch": 0.4098422712933754, "grad_norm": 0.46778481767707863, "learning_rate": 7.355235536325507e-06, "loss": 0.4618, "step": 2030 }, { "epoch": 0.41186119873817034, "grad_norm": 0.4467025867029599, "learning_rate": 7.32409583795932e-06, "loss": 0.4556, "step": 2040 }, { "epoch": 0.4138801261829653, "grad_norm": 0.4864826536631601, "learning_rate": 7.292840722042721e-06, "loss": 0.4649, "step": 2050 }, { "epoch": 0.41589905362776025, "grad_norm": 0.44674303304698915, "learning_rate": 7.261471740744453e-06, "loss": 0.4546, "step": 2060 }, { "epoch": 0.4179179810725552, "grad_norm": 0.4638264920705765, "learning_rate": 7.229990451887961e-06, "loss": 0.459, "step": 2070 }, { "epoch": 0.41993690851735016, "grad_norm": 0.44062835097290626, "learning_rate": 7.1983984188740215e-06, "loss": 0.4627, "step": 2080 }, { "epoch": 0.4219558359621451, "grad_norm": 0.4596889143174662, "learning_rate": 7.1666972106031105e-06, "loss": 0.4584, "step": 2090 }, { "epoch": 0.42397476340694007, "grad_norm": 0.4866244769293159, "learning_rate": 7.134888401397484e-06, "loss": 0.4629, "step": 2100 }, { "epoch": 0.425993690851735, "grad_norm": 0.5201676761862537, "learning_rate": 7.102973570922997e-06, "loss": 0.4549, "step": 2110 }, { "epoch": 0.42801261829653, "grad_norm": 0.4790871394995968, "learning_rate": 7.070954304110656e-06, "loss": 0.4632, "step": 2120 }, { "epoch": 0.4300315457413249, "grad_norm": 0.4659786396553266, "learning_rate": 7.03883219107791e-06, "loss": 0.4581, "step": 2130 }, { "epoch": 0.4320504731861199, "grad_norm": 0.4104121258537053, "learning_rate": 7.0066088270496775e-06, "loss": 0.4596, "step": 2140 }, { "epoch": 0.4340694006309148, "grad_norm": 0.46988504036618683, "learning_rate": 6.9742858122791355e-06, "loss": 0.4648, "step": 2150 }, { "epoch": 0.4360883280757098, "grad_norm": 0.6098242782979109, "learning_rate": 6.941864751968239e-06, "loss": 0.4548, "step": 2160 }, { "epoch": 0.4381072555205047, "grad_norm": 0.7557069029520974, "learning_rate": 6.90934725618801e-06, "loss": 0.461, "step": 2170 }, { "epoch": 0.4401261829652997, "grad_norm": 0.46196984209426056, "learning_rate": 6.876734939798584e-06, "loss": 0.4631, "step": 2180 }, { "epoch": 0.44214511041009463, "grad_norm": 0.45367052105694244, "learning_rate": 6.844029422368998e-06, "loss": 0.4567, "step": 2190 }, { "epoch": 0.4441640378548896, "grad_norm": 0.4923825492028847, "learning_rate": 6.811232328096778e-06, "loss": 0.4619, "step": 2200 }, { "epoch": 0.44618296529968454, "grad_norm": 0.4412882669817311, "learning_rate": 6.7783452857272725e-06, "loss": 0.4472, "step": 2210 }, { "epoch": 0.44820189274447947, "grad_norm": 0.4804805684578094, "learning_rate": 6.745369928472764e-06, "loss": 0.4634, "step": 2220 }, { "epoch": 0.45022082018927445, "grad_norm": 0.46765384977250785, "learning_rate": 6.712307893931364e-06, "loss": 0.4595, "step": 2230 }, { "epoch": 0.4522397476340694, "grad_norm": 1.0068029764875739, "learning_rate": 6.67916082400569e-06, "loss": 0.457, "step": 2240 }, { "epoch": 0.45425867507886436, "grad_norm": 0.4828558935559972, "learning_rate": 6.645930364821327e-06, "loss": 0.465, "step": 2250 }, { "epoch": 0.4562776025236593, "grad_norm": 0.4571716607551568, "learning_rate": 6.612618166645068e-06, "loss": 0.4521, "step": 2260 }, { "epoch": 0.45829652996845427, "grad_norm": 0.49891810921169605, "learning_rate": 6.579225883802981e-06, "loss": 0.4621, "step": 2270 }, { "epoch": 0.4603154574132492, "grad_norm": 0.6812817990255348, "learning_rate": 6.5457551745982275e-06, "loss": 0.4553, "step": 2280 }, { "epoch": 0.4623343848580442, "grad_norm": 0.466296009966539, "learning_rate": 6.512207701228727e-06, "loss": 0.4585, "step": 2290 }, { "epoch": 0.4643533123028391, "grad_norm": 0.48144166164862184, "learning_rate": 6.478585129704608e-06, "loss": 0.4579, "step": 2300 }, { "epoch": 0.4663722397476341, "grad_norm": 0.4575454070862058, "learning_rate": 6.444889129765465e-06, "loss": 0.453, "step": 2310 }, { "epoch": 0.468391167192429, "grad_norm": 0.5049212463992705, "learning_rate": 6.4111213747974385e-06, "loss": 0.4606, "step": 2320 }, { "epoch": 0.470410094637224, "grad_norm": 0.44196403700173553, "learning_rate": 6.377283541750123e-06, "loss": 0.4528, "step": 2330 }, { "epoch": 0.4724290220820189, "grad_norm": 0.44557725863878417, "learning_rate": 6.343377311053271e-06, "loss": 0.4614, "step": 2340 }, { "epoch": 0.4744479495268139, "grad_norm": 0.485138333868983, "learning_rate": 6.309404366533355e-06, "loss": 0.4662, "step": 2350 }, { "epoch": 0.47646687697160883, "grad_norm": 0.44703181568848005, "learning_rate": 6.275366395329934e-06, "loss": 0.4484, "step": 2360 }, { "epoch": 0.47848580441640376, "grad_norm": 0.46684239360571517, "learning_rate": 6.241265087811883e-06, "loss": 0.461, "step": 2370 }, { "epoch": 0.48050473186119874, "grad_norm": 0.4308686375763305, "learning_rate": 6.207102137493433e-06, "loss": 0.4572, "step": 2380 }, { "epoch": 0.48252365930599367, "grad_norm": 0.4349332573681559, "learning_rate": 6.172879240950075e-06, "loss": 0.4586, "step": 2390 }, { "epoch": 0.48454258675078865, "grad_norm": 0.4946893193107334, "learning_rate": 6.138598097734308e-06, "loss": 0.4613, "step": 2400 }, { "epoch": 0.4865615141955836, "grad_norm": 0.4278600145969151, "learning_rate": 6.104260410291237e-06, "loss": 0.4524, "step": 2410 }, { "epoch": 0.48858044164037856, "grad_norm": 0.438261930357501, "learning_rate": 6.069867883874017e-06, "loss": 0.4546, "step": 2420 }, { "epoch": 0.4905993690851735, "grad_norm": 0.4834328335496326, "learning_rate": 6.035422226459183e-06, "loss": 0.4601, "step": 2430 }, { "epoch": 0.49261829652996847, "grad_norm": 0.4560202224259225, "learning_rate": 6.000925148661822e-06, "loss": 0.4552, "step": 2440 }, { "epoch": 0.4946372239747634, "grad_norm": 0.46242134458299394, "learning_rate": 5.966378363650618e-06, "loss": 0.4534, "step": 2450 }, { "epoch": 0.4966561514195584, "grad_norm": 0.509798379798209, "learning_rate": 5.931783587062785e-06, "loss": 0.4478, "step": 2460 }, { "epoch": 0.4986750788643533, "grad_norm": 0.4916270732208669, "learning_rate": 5.897142536918854e-06, "loss": 0.4606, "step": 2470 }, { "epoch": 0.5006940063091483, "grad_norm": 0.4493237090311729, "learning_rate": 5.8624569335373625e-06, "loss": 0.4512, "step": 2480 }, { "epoch": 0.5027129337539432, "grad_norm": 0.4570244860395198, "learning_rate": 5.827728499449417e-06, "loss": 0.4503, "step": 2490 }, { "epoch": 0.5047318611987381, "grad_norm": 0.5044698787582333, "learning_rate": 5.792958959313155e-06, "loss": 0.4586, "step": 2500 }, { "epoch": 0.5067507886435331, "grad_norm": 0.4567238497826546, "learning_rate": 5.758150039828084e-06, "loss": 0.442, "step": 2510 }, { "epoch": 0.5087697160883281, "grad_norm": 0.49456514738525703, "learning_rate": 5.723303469649352e-06, "loss": 0.4524, "step": 2520 }, { "epoch": 0.510788643533123, "grad_norm": 0.46129063291343286, "learning_rate": 5.688420979301876e-06, "loss": 0.4533, "step": 2530 }, { "epoch": 0.512807570977918, "grad_norm": 0.46780830492612324, "learning_rate": 5.6535043010944226e-06, "loss": 0.4537, "step": 2540 }, { "epoch": 0.5148264984227129, "grad_norm": 0.49663726459668367, "learning_rate": 5.618555169033567e-06, "loss": 0.4624, "step": 2550 }, { "epoch": 0.5168454258675079, "grad_norm": 0.4323698197626643, "learning_rate": 5.583575318737588e-06, "loss": 0.4394, "step": 2560 }, { "epoch": 0.5188643533123029, "grad_norm": 0.433967581020834, "learning_rate": 5.548566487350264e-06, "loss": 0.4482, "step": 2570 }, { "epoch": 0.5208832807570978, "grad_norm": 0.5093937708469708, "learning_rate": 5.513530413454618e-06, "loss": 0.4488, "step": 2580 }, { "epoch": 0.5229022082018927, "grad_norm": 0.4479313579903872, "learning_rate": 5.478468836986568e-06, "loss": 0.4547, "step": 2590 }, { "epoch": 0.5249211356466877, "grad_norm": 0.4784953987086218, "learning_rate": 5.443383499148519e-06, "loss": 0.4536, "step": 2600 }, { "epoch": 0.5269400630914827, "grad_norm": 0.4311672354723456, "learning_rate": 5.408276142322904e-06, "loss": 0.4489, "step": 2610 }, { "epoch": 0.5289589905362776, "grad_norm": 0.4320377418884208, "learning_rate": 5.373148509985636e-06, "loss": 0.457, "step": 2620 }, { "epoch": 0.5309779179810725, "grad_norm": 0.4475608921551904, "learning_rate": 5.3380023466195455e-06, "loss": 0.4491, "step": 2630 }, { "epoch": 0.5329968454258676, "grad_norm": 0.4150312768857391, "learning_rate": 5.302839397627732e-06, "loss": 0.4509, "step": 2640 }, { "epoch": 0.5350157728706625, "grad_norm": 0.4842663915075593, "learning_rate": 5.267661409246893e-06, "loss": 0.4497, "step": 2650 }, { "epoch": 0.5370347003154574, "grad_norm": 0.48020608359904937, "learning_rate": 5.232470128460597e-06, "loss": 0.4455, "step": 2660 }, { "epoch": 0.5390536277602523, "grad_norm": 0.460861374110654, "learning_rate": 5.197267302912535e-06, "loss": 0.4633, "step": 2670 }, { "epoch": 0.5410725552050473, "grad_norm": 0.4861656655295904, "learning_rate": 5.162054680819724e-06, "loss": 0.455, "step": 2680 }, { "epoch": 0.5430914826498423, "grad_norm": 0.4280331150720691, "learning_rate": 5.12683401088569e-06, "loss": 0.4527, "step": 2690 }, { "epoch": 0.5451104100946372, "grad_norm": 0.4803934844859794, "learning_rate": 5.091607042213621e-06, "loss": 0.4507, "step": 2700 }, { "epoch": 0.5471293375394322, "grad_norm": 0.44679887041219724, "learning_rate": 5.056375524219516e-06, "loss": 0.4467, "step": 2710 }, { "epoch": 0.5491482649842271, "grad_norm": 0.4562693718725788, "learning_rate": 5.021141206545293e-06, "loss": 0.4481, "step": 2720 }, { "epoch": 0.5511671924290221, "grad_norm": 0.4568391909726876, "learning_rate": 4.9859058389719095e-06, "loss": 0.4467, "step": 2730 }, { "epoch": 0.553186119873817, "grad_norm": 0.4402113900197009, "learning_rate": 4.950671171332458e-06, "loss": 0.4438, "step": 2740 }, { "epoch": 0.555205047318612, "grad_norm": 0.4694901686129437, "learning_rate": 4.915438953425277e-06, "loss": 0.455, "step": 2750 }, { "epoch": 0.5572239747634069, "grad_norm": 0.5092377388071306, "learning_rate": 4.8802109349270415e-06, "loss": 0.4431, "step": 2760 }, { "epoch": 0.5592429022082019, "grad_norm": 0.5334478668234227, "learning_rate": 4.844988865305883e-06, "loss": 0.4457, "step": 2770 }, { "epoch": 0.5612618296529969, "grad_norm": 0.450190134489615, "learning_rate": 4.809774493734507e-06, "loss": 0.4468, "step": 2780 }, { "epoch": 0.5632807570977918, "grad_norm": 0.6136460189487958, "learning_rate": 4.774569569003314e-06, "loss": 0.4531, "step": 2790 }, { "epoch": 0.5652996845425867, "grad_norm": 0.47669716943494894, "learning_rate": 4.739375839433575e-06, "loss": 0.4559, "step": 2800 }, { "epoch": 0.5673186119873816, "grad_norm": 0.47396268716032475, "learning_rate": 4.704195052790584e-06, "loss": 0.4463, "step": 2810 }, { "epoch": 0.5693375394321767, "grad_norm": 0.5115114860936024, "learning_rate": 4.66902895619688e-06, "loss": 0.4494, "step": 2820 }, { "epoch": 0.5713564668769716, "grad_norm": 0.4561358589231192, "learning_rate": 4.633879296045474e-06, "loss": 0.4477, "step": 2830 }, { "epoch": 0.5733753943217665, "grad_norm": 0.4547946639578122, "learning_rate": 4.59874781791312e-06, "loss": 0.4502, "step": 2840 }, { "epoch": 0.5753943217665615, "grad_norm": 0.5036899795994354, "learning_rate": 4.56363626647363e-06, "loss": 0.4553, "step": 2850 }, { "epoch": 0.5774132492113565, "grad_norm": 0.46515739613875146, "learning_rate": 4.528546385411239e-06, "loss": 0.4373, "step": 2860 }, { "epoch": 0.5794321766561514, "grad_norm": 0.4834989187992637, "learning_rate": 4.493479917333989e-06, "loss": 0.453, "step": 2870 }, { "epoch": 0.5814511041009464, "grad_norm": 0.45922181466381384, "learning_rate": 4.4584386036872155e-06, "loss": 0.4527, "step": 2880 }, { "epoch": 0.5834700315457413, "grad_norm": 0.46803871925708684, "learning_rate": 4.42342418466705e-06, "loss": 0.4426, "step": 2890 }, { "epoch": 0.5854889589905363, "grad_norm": 0.47319916436776355, "learning_rate": 4.388438399134e-06, "loss": 0.4547, "step": 2900 }, { "epoch": 0.5875078864353313, "grad_norm": 0.4598569796278221, "learning_rate": 4.3534829845265996e-06, "loss": 0.4441, "step": 2910 }, { "epoch": 0.5895268138801262, "grad_norm": 0.47100191118178447, "learning_rate": 4.318559676775126e-06, "loss": 0.4472, "step": 2920 }, { "epoch": 0.5915457413249211, "grad_norm": 0.495923691841689, "learning_rate": 4.283670210215384e-06, "loss": 0.4516, "step": 2930 }, { "epoch": 0.5935646687697161, "grad_norm": 0.4391764043906036, "learning_rate": 4.248816317502585e-06, "loss": 0.4468, "step": 2940 }, { "epoch": 0.5955835962145111, "grad_norm": 0.47900304182889214, "learning_rate": 4.213999729525304e-06, "loss": 0.452, "step": 2950 }, { "epoch": 0.597602523659306, "grad_norm": 0.4894376668991494, "learning_rate": 4.179222175319502e-06, "loss": 0.4418, "step": 2960 }, { "epoch": 0.5996214511041009, "grad_norm": 0.4698266233405637, "learning_rate": 4.1444853819826845e-06, "loss": 0.4503, "step": 2970 }, { "epoch": 0.6016403785488958, "grad_norm": 0.4601994675920038, "learning_rate": 4.1097910745881175e-06, "loss": 0.4421, "step": 2980 }, { "epoch": 0.6036593059936909, "grad_norm": 0.6167906848346414, "learning_rate": 4.0751409760991585e-06, "loss": 0.4469, "step": 2990 }, { "epoch": 0.6056782334384858, "grad_norm": 0.5051790895352023, "learning_rate": 4.040536807283696e-06, "loss": 0.4502, "step": 3000 }, { "epoch": 0.6076971608832807, "grad_norm": 0.45322450050363433, "learning_rate": 4.005980286628698e-06, "loss": 0.4505, "step": 3010 }, { "epoch": 0.6097160883280757, "grad_norm": 0.4910869337898713, "learning_rate": 3.9714731302548525e-06, "loss": 0.4486, "step": 3020 }, { "epoch": 0.6117350157728707, "grad_norm": 0.447631013542978, "learning_rate": 3.937017051831367e-06, "loss": 0.4478, "step": 3030 }, { "epoch": 0.6137539432176656, "grad_norm": 0.7018069369756337, "learning_rate": 3.9026137624908466e-06, "loss": 0.4492, "step": 3040 }, { "epoch": 0.6157728706624606, "grad_norm": 0.4762076148803905, "learning_rate": 3.868264970744324e-06, "loss": 0.4513, "step": 3050 }, { "epoch": 0.6177917981072555, "grad_norm": 0.5020509623933802, "learning_rate": 3.833972382396412e-06, "loss": 0.4482, "step": 3060 }, { "epoch": 0.6198107255520505, "grad_norm": 0.44347011131653674, "learning_rate": 3.799737700460594e-06, "loss": 0.4441, "step": 3070 }, { "epoch": 0.6218296529968455, "grad_norm": 0.45353690385677753, "learning_rate": 3.7655626250746446e-06, "loss": 0.4443, "step": 3080 }, { "epoch": 0.6238485804416404, "grad_norm": 0.44404552147051596, "learning_rate": 3.7314488534162007e-06, "loss": 0.4509, "step": 3090 }, { "epoch": 0.6258675078864353, "grad_norm": 0.4759693579595236, "learning_rate": 3.6973980796184843e-06, "loss": 0.4573, "step": 3100 }, { "epoch": 0.6278864353312302, "grad_norm": 0.45248182104327295, "learning_rate": 3.6634119946861556e-06, "loss": 0.4442, "step": 3110 }, { "epoch": 0.6299053627760253, "grad_norm": 0.44605725901650006, "learning_rate": 3.629492286411349e-06, "loss": 0.4515, "step": 3120 }, { "epoch": 0.6319242902208202, "grad_norm": 0.4770842161429305, "learning_rate": 3.5956406392898484e-06, "loss": 0.4481, "step": 3130 }, { "epoch": 0.6339432176656151, "grad_norm": 0.4611511904942963, "learning_rate": 3.561858734437432e-06, "loss": 0.4553, "step": 3140 }, { "epoch": 0.63596214511041, "grad_norm": 0.4886490480648605, "learning_rate": 3.528148249506391e-06, "loss": 0.4541, "step": 3150 }, { "epoch": 0.6379810725552051, "grad_norm": 0.5063011038652176, "learning_rate": 3.494510858602217e-06, "loss": 0.448, "step": 3160 }, { "epoch": 0.64, "grad_norm": 0.5297665998286369, "learning_rate": 3.4609482322004463e-06, "loss": 0.45, "step": 3170 }, { "epoch": 0.6420189274447949, "grad_norm": 0.789602090268308, "learning_rate": 3.427462037063729e-06, "loss": 0.4499, "step": 3180 }, { "epoch": 0.6440378548895899, "grad_norm": 0.4595511607863831, "learning_rate": 3.394053936159035e-06, "loss": 0.4451, "step": 3190 }, { "epoch": 0.6460567823343849, "grad_norm": 0.5053416927018038, "learning_rate": 3.3607255885750756e-06, "loss": 0.4488, "step": 3200 }, { "epoch": 0.6480757097791798, "grad_norm": 0.47050871567377006, "learning_rate": 3.3274786494399102e-06, "loss": 0.4461, "step": 3210 }, { "epoch": 0.6500946372239748, "grad_norm": 0.46433780403203007, "learning_rate": 3.2943147698387562e-06, "loss": 0.4497, "step": 3220 }, { "epoch": 0.6521135646687697, "grad_norm": 0.45409078162209787, "learning_rate": 3.261235596731982e-06, "loss": 0.4457, "step": 3230 }, { "epoch": 0.6541324921135647, "grad_norm": 0.4218245687222197, "learning_rate": 3.228242772873328e-06, "loss": 0.4437, "step": 3240 }, { "epoch": 0.6561514195583596, "grad_norm": 0.47127385494827945, "learning_rate": 3.195337936728323e-06, "loss": 0.4488, "step": 3250 }, { "epoch": 0.6581703470031546, "grad_norm": 0.43568607166498186, "learning_rate": 3.1625227223929105e-06, "loss": 0.4415, "step": 3260 }, { "epoch": 0.6601892744479495, "grad_norm": 0.4200623699882891, "learning_rate": 3.1297987595123024e-06, "loss": 0.4423, "step": 3270 }, { "epoch": 0.6622082018927444, "grad_norm": 0.45902924831849934, "learning_rate": 3.0971676732000494e-06, "loss": 0.4488, "step": 3280 }, { "epoch": 0.6642271293375395, "grad_norm": 0.48610535709106045, "learning_rate": 3.0646310839573302e-06, "loss": 0.441, "step": 3290 }, { "epoch": 0.6662460567823344, "grad_norm": 0.5237594196219926, "learning_rate": 3.03219060759248e-06, "loss": 0.454, "step": 3300 }, { "epoch": 0.6682649842271293, "grad_norm": 0.4228282553579445, "learning_rate": 2.9998478551407524e-06, "loss": 0.4378, "step": 3310 }, { "epoch": 0.6702839116719242, "grad_norm": 0.46926328208349977, "learning_rate": 2.9676044327842963e-06, "loss": 0.4475, "step": 3320 }, { "epoch": 0.6723028391167193, "grad_norm": 0.47029450291352376, "learning_rate": 2.935461941772412e-06, "loss": 0.4418, "step": 3330 }, { "epoch": 0.6743217665615142, "grad_norm": 0.461936788764074, "learning_rate": 2.9034219783420183e-06, "loss": 0.438, "step": 3340 }, { "epoch": 0.6763406940063091, "grad_norm": 0.4879384251763818, "learning_rate": 2.871486133638378e-06, "loss": 0.4498, "step": 3350 }, { "epoch": 0.6783596214511041, "grad_norm": 0.47390435742643133, "learning_rate": 2.8396559936360934e-06, "loss": 0.4444, "step": 3360 }, { "epoch": 0.6803785488958991, "grad_norm": 0.43781719559966065, "learning_rate": 2.807933139060337e-06, "loss": 0.4481, "step": 3370 }, { "epoch": 0.682397476340694, "grad_norm": 0.452166890113836, "learning_rate": 2.776319145308345e-06, "loss": 0.4397, "step": 3380 }, { "epoch": 0.684416403785489, "grad_norm": 0.4480696892989974, "learning_rate": 2.7448155823711897e-06, "loss": 0.4415, "step": 3390 }, { "epoch": 0.6864353312302839, "grad_norm": 0.482631983733039, "learning_rate": 2.7134240147558124e-06, "loss": 0.4494, "step": 3400 }, { "epoch": 0.6884542586750788, "grad_norm": 0.4215351521880855, "learning_rate": 2.6821460014073187e-06, "loss": 0.4359, "step": 3410 }, { "epoch": 0.6904731861198738, "grad_norm": 0.44170545621316426, "learning_rate": 2.6509830956315684e-06, "loss": 0.4516, "step": 3420 }, { "epoch": 0.6924921135646688, "grad_norm": 0.43846533524679027, "learning_rate": 2.619936845018031e-06, "loss": 0.4357, "step": 3430 }, { "epoch": 0.6945110410094637, "grad_norm": 0.5008934216933979, "learning_rate": 2.5890087913629358e-06, "loss": 0.4467, "step": 3440 }, { "epoch": 0.6965299684542586, "grad_norm": 0.5250213600953301, "learning_rate": 2.558200470592701e-06, "loss": 0.449, "step": 3450 }, { "epoch": 0.6985488958990537, "grad_norm": 0.5362972001626155, "learning_rate": 2.527513412687654e-06, "loss": 0.436, "step": 3460 }, { "epoch": 0.7005678233438486, "grad_norm": 0.4447368579357635, "learning_rate": 2.4969491416060542e-06, "loss": 0.4393, "step": 3470 }, { "epoch": 0.7025867507886435, "grad_norm": 0.42761346324330285, "learning_rate": 2.4665091752084175e-06, "loss": 0.4423, "step": 3480 }, { "epoch": 0.7046056782334384, "grad_norm": 0.45180644732668995, "learning_rate": 2.4361950251821253e-06, "loss": 0.4351, "step": 3490 }, { "epoch": 0.7066246056782335, "grad_norm": 0.4655597774869552, "learning_rate": 2.406008196966359e-06, "loss": 0.445, "step": 3500 }, { "epoch": 0.7086435331230284, "grad_norm": 0.44607134328593223, "learning_rate": 2.375950189677342e-06, "loss": 0.4322, "step": 3510 }, { "epoch": 0.7106624605678233, "grad_norm": 0.45886523824853853, "learning_rate": 2.346022496033879e-06, "loss": 0.439, "step": 3520 }, { "epoch": 0.7126813880126183, "grad_norm": 0.439206301308698, "learning_rate": 2.316226602283243e-06, "loss": 0.445, "step": 3530 }, { "epoch": 0.7147003154574133, "grad_norm": 0.42444173885080955, "learning_rate": 2.2865639881273494e-06, "loss": 0.4402, "step": 3540 }, { "epoch": 0.7167192429022082, "grad_norm": 0.48281990057391616, "learning_rate": 2.25703612664928e-06, "loss": 0.445, "step": 3550 }, { "epoch": 0.7187381703470032, "grad_norm": 0.46127829547820787, "learning_rate": 2.227644484240135e-06, "loss": 0.4356, "step": 3560 }, { "epoch": 0.7207570977917981, "grad_norm": 0.4610423355631118, "learning_rate": 2.198390520526194e-06, "loss": 0.4438, "step": 3570 }, { "epoch": 0.722776025236593, "grad_norm": 0.4711885745663891, "learning_rate": 2.1692756882964404e-06, "loss": 0.4464, "step": 3580 }, { "epoch": 0.724794952681388, "grad_norm": 0.5420237747030388, "learning_rate": 2.140301433430414e-06, "loss": 0.4436, "step": 3590 }, { "epoch": 0.726813880126183, "grad_norm": 0.47745493287406926, "learning_rate": 2.1114691948264038e-06, "loss": 0.4433, "step": 3600 }, { "epoch": 0.7288328075709779, "grad_norm": 0.43683824200923127, "learning_rate": 2.082780404329988e-06, "loss": 0.4429, "step": 3610 }, { "epoch": 0.7308517350157728, "grad_norm": 0.42141576682539833, "learning_rate": 2.0542364866629292e-06, "loss": 0.4402, "step": 3620 }, { "epoch": 0.7328706624605679, "grad_norm": 0.4304526228372951, "learning_rate": 2.025838859352428e-06, "loss": 0.4462, "step": 3630 }, { "epoch": 0.7348895899053628, "grad_norm": 0.4960972537231228, "learning_rate": 1.997588932660713e-06, "loss": 0.4298, "step": 3640 }, { "epoch": 0.7369085173501577, "grad_norm": 0.5055917828329974, "learning_rate": 1.9694881095150125e-06, "loss": 0.4429, "step": 3650 }, { "epoch": 0.7389274447949526, "grad_norm": 0.4756073151541865, "learning_rate": 1.9415377854378903e-06, "loss": 0.4423, "step": 3660 }, { "epoch": 0.7409463722397477, "grad_norm": 0.4293238953341605, "learning_rate": 1.913739348477927e-06, "loss": 0.444, "step": 3670 }, { "epoch": 0.7429652996845426, "grad_norm": 0.4476184732471472, "learning_rate": 1.8860941791408038e-06, "loss": 0.4432, "step": 3680 }, { "epoch": 0.7449842271293375, "grad_norm": 0.4453341449071723, "learning_rate": 1.8586036503207316e-06, "loss": 0.4351, "step": 3690 }, { "epoch": 0.7470031545741325, "grad_norm": 0.4741404781356218, "learning_rate": 1.8312691272322768e-06, "loss": 0.4445, "step": 3700 }, { "epoch": 0.7490220820189274, "grad_norm": 0.4314595245056323, "learning_rate": 1.804091967342569e-06, "loss": 0.4392, "step": 3710 }, { "epoch": 0.7510410094637224, "grad_norm": 0.4960471983973586, "learning_rate": 1.777073520303874e-06, "loss": 0.4406, "step": 3720 }, { "epoch": 0.7530599369085174, "grad_norm": 0.47724657720097824, "learning_rate": 1.7502151278865798e-06, "loss": 0.4391, "step": 3730 }, { "epoch": 0.7550788643533123, "grad_norm": 0.45566327509907684, "learning_rate": 1.7235181239125587e-06, "loss": 0.4411, "step": 3740 }, { "epoch": 0.7570977917981072, "grad_norm": 0.47859031492685, "learning_rate": 1.6969838341889288e-06, "loss": 0.4412, "step": 3750 }, { "epoch": 0.7591167192429022, "grad_norm": 0.4239970850626806, "learning_rate": 1.67061357644221e-06, "loss": 0.4327, "step": 3760 }, { "epoch": 0.7611356466876972, "grad_norm": 0.41439508221708543, "learning_rate": 1.6444086602528835e-06, "loss": 0.4403, "step": 3770 }, { "epoch": 0.7631545741324921, "grad_norm": 0.4724942386548484, "learning_rate": 1.6183703869903666e-06, "loss": 0.4399, "step": 3780 }, { "epoch": 0.765173501577287, "grad_norm": 0.5068083200686946, "learning_rate": 1.59250004974837e-06, "loss": 0.4484, "step": 3790 }, { "epoch": 0.7671924290220821, "grad_norm": 0.47623913375530136, "learning_rate": 1.5667989332806894e-06, "loss": 0.442, "step": 3800 }, { "epoch": 0.769211356466877, "grad_norm": 0.5448383844625657, "learning_rate": 1.5412683139374068e-06, "loss": 0.4339, "step": 3810 }, { "epoch": 0.7712302839116719, "grad_norm": 0.4819612793724808, "learning_rate": 1.5159094596014934e-06, "loss": 0.4478, "step": 3820 }, { "epoch": 0.7732492113564668, "grad_norm": 0.4456007417706316, "learning_rate": 1.4907236296258598e-06, "loss": 0.4423, "step": 3830 }, { "epoch": 0.7752681388012618, "grad_norm": 0.43979298393219013, "learning_rate": 1.4657120747708014e-06, "loss": 0.4417, "step": 3840 }, { "epoch": 0.7772870662460568, "grad_norm": 0.5066920888924865, "learning_rate": 1.4408760371418895e-06, "loss": 0.4397, "step": 3850 }, { "epoch": 0.7793059936908517, "grad_norm": 0.47046124228822955, "learning_rate": 1.4162167501282925e-06, "loss": 0.4416, "step": 3860 }, { "epoch": 0.7813249211356467, "grad_norm": 0.4558898604297212, "learning_rate": 1.391735438341514e-06, "loss": 0.437, "step": 3870 }, { "epoch": 0.7833438485804416, "grad_norm": 0.4635342780752439, "learning_rate": 1.3674333175545818e-06, "loss": 0.4425, "step": 3880 }, { "epoch": 0.7853627760252366, "grad_norm": 0.450134125379297, "learning_rate": 1.343311594641674e-06, "loss": 0.4371, "step": 3890 }, { "epoch": 0.7873817034700316, "grad_norm": 0.48501901287598154, "learning_rate": 1.3193714675181813e-06, "loss": 0.4428, "step": 3900 }, { "epoch": 0.7894006309148265, "grad_norm": 0.41955405495588977, "learning_rate": 1.2956141250812144e-06, "loss": 0.4307, "step": 3910 }, { "epoch": 0.7914195583596214, "grad_norm": 0.41659445986128424, "learning_rate": 1.272040747150563e-06, "loss": 0.447, "step": 3920 }, { "epoch": 0.7934384858044164, "grad_norm": 0.44707673152149113, "learning_rate": 1.2486525044101123e-06, "loss": 0.4333, "step": 3930 }, { "epoch": 0.7954574132492114, "grad_norm": 0.4698766268506151, "learning_rate": 1.2254505583496922e-06, "loss": 0.4403, "step": 3940 }, { "epoch": 0.7974763406940063, "grad_norm": 0.47512471796296707, "learning_rate": 1.2024360612074055e-06, "loss": 0.4454, "step": 3950 }, { "epoch": 0.7994952681388012, "grad_norm": 0.4585910400457452, "learning_rate": 1.179610155912405e-06, "loss": 0.436, "step": 3960 }, { "epoch": 0.8015141955835963, "grad_norm": 0.4557389746705491, "learning_rate": 1.1569739760281295e-06, "loss": 0.4393, "step": 3970 }, { "epoch": 0.8035331230283912, "grad_norm": 0.4553106102818403, "learning_rate": 1.1345286456960186e-06, "loss": 0.4439, "step": 3980 }, { "epoch": 0.8055520504731861, "grad_norm": 0.4443887135897382, "learning_rate": 1.1122752795796748e-06, "loss": 0.4413, "step": 3990 }, { "epoch": 0.807570977917981, "grad_norm": 0.47272003456758904, "learning_rate": 1.0902149828095154e-06, "loss": 0.4422, "step": 4000 }, { "epoch": 0.809589905362776, "grad_norm": 0.43703108025040205, "learning_rate": 1.0683488509278928e-06, "loss": 0.4386, "step": 4010 }, { "epoch": 0.811608832807571, "grad_norm": 0.4441202945224273, "learning_rate": 1.0466779698346797e-06, "loss": 0.4363, "step": 4020 }, { "epoch": 0.8136277602523659, "grad_norm": 0.42622028909468485, "learning_rate": 1.025203415733348e-06, "loss": 0.4365, "step": 4030 }, { "epoch": 0.8156466876971609, "grad_norm": 0.440821020950604, "learning_rate": 1.0039262550775235e-06, "loss": 0.4345, "step": 4040 }, { "epoch": 0.8176656151419558, "grad_norm": 0.4670601139166515, "learning_rate": 9.828475445180212e-07, "loss": 0.4404, "step": 4050 }, { "epoch": 0.8196845425867508, "grad_norm": 0.4363041969561276, "learning_rate": 9.61968330850374e-07, "loss": 0.4369, "step": 4060 }, { "epoch": 0.8217034700315458, "grad_norm": 0.4866069797752605, "learning_rate": 9.412896509628439e-07, "loss": 0.4434, "step": 4070 }, { "epoch": 0.8237223974763407, "grad_norm": 0.4360948594973755, "learning_rate": 9.208125317849304e-07, "loss": 0.4395, "step": 4080 }, { "epoch": 0.8257413249211356, "grad_norm": 0.4490621547393745, "learning_rate": 9.005379902363753e-07, "loss": 0.4405, "step": 4090 }, { "epoch": 0.8277602523659306, "grad_norm": 0.4809678412101839, "learning_rate": 8.804670331766557e-07, "loss": 0.4449, "step": 4100 }, { "epoch": 0.8297791798107256, "grad_norm": 0.4626290445825763, "learning_rate": 8.606006573549836e-07, "loss": 0.4359, "step": 4110 }, { "epoch": 0.8317981072555205, "grad_norm": 0.5187638972410576, "learning_rate": 8.409398493608089e-07, "loss": 0.4413, "step": 4120 }, { "epoch": 0.8338170347003154, "grad_norm": 0.46587792642977804, "learning_rate": 8.214855855748249e-07, "loss": 0.4409, "step": 4130 }, { "epoch": 0.8358359621451104, "grad_norm": 0.4706901825283824, "learning_rate": 8.022388321204688e-07, "loss": 0.4354, "step": 4140 }, { "epoch": 0.8378548895899054, "grad_norm": 0.562791948402191, "learning_rate": 7.832005448159585e-07, "loss": 0.4445, "step": 4150 }, { "epoch": 0.8398738170347003, "grad_norm": 0.4543738469403552, "learning_rate": 7.643716691268171e-07, "loss": 0.4374, "step": 4160 }, { "epoch": 0.8418927444794952, "grad_norm": 1.6174532184269235, "learning_rate": 7.457531401189188e-07, "loss": 0.4408, "step": 4170 }, { "epoch": 0.8439116719242902, "grad_norm": 0.46922997370194347, "learning_rate": 7.273458824120555e-07, "loss": 0.4376, "step": 4180 }, { "epoch": 0.8459305993690852, "grad_norm": 0.44210687899012985, "learning_rate": 7.091508101340189e-07, "loss": 0.4407, "step": 4190 }, { "epoch": 0.8479495268138801, "grad_norm": 0.488809217492704, "learning_rate": 6.91168826875201e-07, "loss": 0.4489, "step": 4200 }, { "epoch": 0.8499684542586751, "grad_norm": 0.4335482219101724, "learning_rate": 6.734008256437253e-07, "loss": 0.435, "step": 4210 }, { "epoch": 0.85198738170347, "grad_norm": 0.4549494452467777, "learning_rate": 6.558476888210935e-07, "loss": 0.4375, "step": 4220 }, { "epoch": 0.854006309148265, "grad_norm": 0.44896647048001603, "learning_rate": 6.38510288118368e-07, "loss": 0.4409, "step": 4230 }, { "epoch": 0.85602523659306, "grad_norm": 0.43685823117883865, "learning_rate": 6.213894845328844e-07, "loss": 0.4418, "step": 4240 }, { "epoch": 0.8580441640378549, "grad_norm": 0.47792243578520655, "learning_rate": 6.044861283054876e-07, "loss": 0.4452, "step": 4250 }, { "epoch": 0.8600630914826498, "grad_norm": 0.43885235740502504, "learning_rate": 5.878010588783101e-07, "loss": 0.4423, "step": 4260 }, { "epoch": 0.8620820189274448, "grad_norm": 0.5121668333108534, "learning_rate": 5.713351048530874e-07, "loss": 0.4429, "step": 4270 }, { "epoch": 0.8641009463722398, "grad_norm": 0.44403508398564756, "learning_rate": 5.550890839500067e-07, "loss": 0.4511, "step": 4280 }, { "epoch": 0.8661198738170347, "grad_norm": 0.42769598938785636, "learning_rate": 5.390638029670908e-07, "loss": 0.4322, "step": 4290 }, { "epoch": 0.8681388012618296, "grad_norm": 0.4949663561795469, "learning_rate": 5.232600577401436e-07, "loss": 0.4457, "step": 4300 }, { "epoch": 0.8701577287066246, "grad_norm": 0.4489314407451146, "learning_rate": 5.076786331032224e-07, "loss": 0.4327, "step": 4310 }, { "epoch": 0.8721766561514196, "grad_norm": 0.48025463446625855, "learning_rate": 4.923203028496576e-07, "loss": 0.4447, "step": 4320 }, { "epoch": 0.8741955835962145, "grad_norm": 0.5743773868003562, "learning_rate": 4.771858296936311e-07, "loss": 0.4402, "step": 4330 }, { "epoch": 0.8762145110410094, "grad_norm": 0.4329672101913256, "learning_rate": 4.622759652322989e-07, "loss": 0.4455, "step": 4340 }, { "epoch": 0.8782334384858044, "grad_norm": 0.49048963795844497, "learning_rate": 4.475914499084605e-07, "loss": 0.4467, "step": 4350 }, { "epoch": 0.8802523659305994, "grad_norm": 0.4829611101804599, "learning_rate": 4.331330129737954e-07, "loss": 0.4287, "step": 4360 }, { "epoch": 0.8822712933753943, "grad_norm": 0.44140910883095635, "learning_rate": 4.189013724526392e-07, "loss": 0.4446, "step": 4370 }, { "epoch": 0.8842902208201893, "grad_norm": 0.45272062627566334, "learning_rate": 4.048972351063313e-07, "loss": 0.4354, "step": 4380 }, { "epoch": 0.8863091482649842, "grad_norm": 0.7049959167129851, "learning_rate": 3.9112129639811535e-07, "loss": 0.4346, "step": 4390 }, { "epoch": 0.8883280757097792, "grad_norm": 0.4811238411002169, "learning_rate": 3.7757424045860025e-07, "loss": 0.4372, "step": 4400 }, { "epoch": 0.8903470031545742, "grad_norm": 0.4540483251684103, "learning_rate": 3.642567400517838e-07, "loss": 0.4314, "step": 4410 }, { "epoch": 0.8923659305993691, "grad_norm": 0.47603601308214405, "learning_rate": 3.511694565416485e-07, "loss": 0.4382, "step": 4420 }, { "epoch": 0.894384858044164, "grad_norm": 0.4286142580967842, "learning_rate": 3.383130398593121e-07, "loss": 0.4399, "step": 4430 }, { "epoch": 0.8964037854889589, "grad_norm": 0.4337017987281163, "learning_rate": 3.2568812847074913e-07, "loss": 0.4437, "step": 4440 }, { "epoch": 0.898422712933754, "grad_norm": 0.48386529314250487, "learning_rate": 3.1329534934509186e-07, "loss": 0.4415, "step": 4450 }, { "epoch": 0.9004416403785489, "grad_norm": 0.46990087367597777, "learning_rate": 3.011353179234888e-07, "loss": 0.4317, "step": 4460 }, { "epoch": 0.9024605678233438, "grad_norm": 0.42506034975479673, "learning_rate": 2.8920863808853947e-07, "loss": 0.444, "step": 4470 }, { "epoch": 0.9044794952681388, "grad_norm": 0.6216117310744336, "learning_rate": 2.775159021343082e-07, "loss": 0.4367, "step": 4480 }, { "epoch": 0.9064984227129338, "grad_norm": 0.4583995655944248, "learning_rate": 2.660576907369111e-07, "loss": 0.4412, "step": 4490 }, { "epoch": 0.9085173501577287, "grad_norm": 0.4889160958013681, "learning_rate": 2.5483457292567517e-07, "loss": 0.4403, "step": 4500 }, { "epoch": 0.9105362776025236, "grad_norm": 0.4267838945833425, "learning_rate": 2.4384710605488195e-07, "loss": 0.4402, "step": 4510 }, { "epoch": 0.9125552050473186, "grad_norm": 0.4986776395611631, "learning_rate": 2.330958357760882e-07, "loss": 0.4399, "step": 4520 }, { "epoch": 0.9145741324921136, "grad_norm": 0.459784020601065, "learning_rate": 2.225812960110263e-07, "loss": 0.4316, "step": 4530 }, { "epoch": 0.9165930599369085, "grad_norm": 0.44426443349474987, "learning_rate": 2.123040089250944e-07, "loss": 0.4402, "step": 4540 }, { "epoch": 0.9186119873817035, "grad_norm": 0.49235316049109906, "learning_rate": 2.0226448490141715e-07, "loss": 0.4481, "step": 4550 }, { "epoch": 0.9206309148264984, "grad_norm": 0.43058790479448, "learning_rate": 1.9246322251550653e-07, "loss": 0.4317, "step": 4560 }, { "epoch": 0.9226498422712934, "grad_norm": 0.4966838385984267, "learning_rate": 1.829007085104978e-07, "loss": 0.4427, "step": 4570 }, { "epoch": 0.9246687697160884, "grad_norm": 0.4666275183152453, "learning_rate": 1.7357741777297987e-07, "loss": 0.4485, "step": 4580 }, { "epoch": 0.9266876971608833, "grad_norm": 0.4412830191687912, "learning_rate": 1.644938133094076e-07, "loss": 0.4312, "step": 4590 }, { "epoch": 0.9287066246056782, "grad_norm": 0.4886548137136921, "learning_rate": 1.5565034622311349e-07, "loss": 0.4461, "step": 4600 }, { "epoch": 0.9307255520504731, "grad_norm": 0.4652701957357695, "learning_rate": 1.4704745569190327e-07, "loss": 0.4375, "step": 4610 }, { "epoch": 0.9327444794952682, "grad_norm": 0.5298674695548953, "learning_rate": 1.3868556894624353e-07, "loss": 0.4399, "step": 4620 }, { "epoch": 0.9347634069400631, "grad_norm": 0.4562802443840814, "learning_rate": 1.30565101248048e-07, "loss": 0.4385, "step": 4630 }, { "epoch": 0.936782334384858, "grad_norm": 0.4578530629768795, "learning_rate": 1.226864558700541e-07, "loss": 0.4416, "step": 4640 }, { "epoch": 0.938801261829653, "grad_norm": 0.4951660779551572, "learning_rate": 1.1505002407579512e-07, "loss": 0.4424, "step": 4650 }, { "epoch": 0.940820189274448, "grad_norm": 0.44757841438265406, "learning_rate": 1.0765618510017118e-07, "loss": 0.4368, "step": 4660 }, { "epoch": 0.9428391167192429, "grad_norm": 1.5035499500846865, "learning_rate": 1.0050530613061327e-07, "loss": 0.4353, "step": 4670 }, { "epoch": 0.9448580441640378, "grad_norm": 0.4938389329390479, "learning_rate": 9.359774228885177e-08, "loss": 0.4391, "step": 4680 }, { "epoch": 0.9468769716088328, "grad_norm": 0.4711850612472287, "learning_rate": 8.693383661327882e-08, "loss": 0.4388, "step": 4690 }, { "epoch": 0.9488958990536278, "grad_norm": 0.48652374519922575, "learning_rate": 8.051392004191139e-08, "loss": 0.4486, "step": 4700 }, { "epoch": 0.9509148264984227, "grad_norm": 0.4473704324045684, "learning_rate": 7.433831139595727e-08, "loss": 0.4385, "step": 4710 }, { "epoch": 0.9529337539432177, "grad_norm": 0.509944701555344, "learning_rate": 6.840731736398542e-08, "loss": 0.4362, "step": 4720 }, { "epoch": 0.9549526813880126, "grad_norm": 0.4468123798191573, "learning_rate": 6.272123248668993e-08, "loss": 0.4356, "step": 4730 }, { "epoch": 0.9569716088328075, "grad_norm": 0.4432856047100861, "learning_rate": 5.728033914226494e-08, "loss": 0.4406, "step": 4740 }, { "epoch": 0.9589905362776026, "grad_norm": 0.8312641118207356, "learning_rate": 5.2084907532383156e-08, "loss": 0.4457, "step": 4750 }, { "epoch": 0.9610094637223975, "grad_norm": 0.4308428180533281, "learning_rate": 4.7135195668776e-08, "loss": 0.4337, "step": 4760 }, { "epoch": 0.9630283911671924, "grad_norm": 0.47706748470749133, "learning_rate": 4.243144936041943e-08, "loss": 0.4413, "step": 4770 }, { "epoch": 0.9650473186119873, "grad_norm": 0.46426542825118344, "learning_rate": 3.797390220132757e-08, "loss": 0.4325, "step": 4780 }, { "epoch": 0.9670662460567824, "grad_norm": 0.4324982289321672, "learning_rate": 3.3762775558952576e-08, "loss": 0.4377, "step": 4790 }, { "epoch": 0.9690851735015773, "grad_norm": 0.47719223462129745, "learning_rate": 2.9798278563190618e-08, "loss": 0.4439, "step": 4800 }, { "epoch": 0.9711041009463722, "grad_norm": 0.4540530080947163, "learning_rate": 2.6080608095997438e-08, "loss": 0.4377, "step": 4810 }, { "epoch": 0.9731230283911672, "grad_norm": 0.4413909184803219, "learning_rate": 2.2609948781607827e-08, "loss": 0.4424, "step": 4820 }, { "epoch": 0.9751419558359622, "grad_norm": 0.4712177319951425, "learning_rate": 1.9386472977370728e-08, "loss": 0.4412, "step": 4830 }, { "epoch": 0.9771608832807571, "grad_norm": 0.443140708468747, "learning_rate": 1.6410340765187772e-08, "loss": 0.4382, "step": 4840 }, { "epoch": 0.979179810725552, "grad_norm": 0.4782858905945508, "learning_rate": 1.3681699943564053e-08, "loss": 0.4387, "step": 4850 }, { "epoch": 0.981198738170347, "grad_norm": 0.6206662222549734, "learning_rate": 1.1200686020266805e-08, "loss": 0.4431, "step": 4860 }, { "epoch": 0.983217665615142, "grad_norm": 2.8460570799996985, "learning_rate": 8.967422205598541e-09, "loss": 0.4393, "step": 4870 }, { "epoch": 0.9852365930599369, "grad_norm": 0.4435087118229933, "learning_rate": 6.982019406275853e-09, "loss": 0.436, "step": 4880 }, { "epoch": 0.9872555205047319, "grad_norm": 0.4300508595922509, "learning_rate": 5.244576219922693e-09, "loss": 0.4316, "step": 4890 }, { "epoch": 0.9892744479495268, "grad_norm": 0.49712300151356087, "learning_rate": 3.755178930173742e-09, "loss": 0.4463, "step": 4900 }, { "epoch": 0.9912933753943217, "grad_norm": 0.4798533947795922, "learning_rate": 2.513901502389504e-09, "loss": 0.4327, "step": 4910 }, { "epoch": 0.9933123028391168, "grad_norm": 0.4936501161499046, "learning_rate": 1.520805579983131e-09, "loss": 0.4403, "step": 4920 }, { "epoch": 0.9953312302839117, "grad_norm": 0.4638813307919199, "learning_rate": 7.759404813600935e-10, "loss": 0.4385, "step": 4930 }, { "epoch": 0.9973501577287066, "grad_norm": 0.43789683220104403, "learning_rate": 2.7934319746625306e-10, "loss": 0.4387, "step": 4940 }, { "epoch": 0.9993690851735015, "grad_norm": 0.4980004114526748, "learning_rate": 3.103838995321962e-11, "loss": 0.4391, "step": 4950 }, { "epoch": 1.0, "step": 4954, "total_flos": 1.1092292323919266e+19, "train_loss": 0.46906891030845593, "train_runtime": 81648.0087, "train_samples_per_second": 7.765, "train_steps_per_second": 0.061 } ], "logging_steps": 10, "max_steps": 4954, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1092292323919266e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }