polyglot-tagger-v2.2 / trainer_state.json
DerivedFunction's picture
End of training
f3f95a2
{
"best_global_step": 60000,
"best_metric": 0.9576732752290649,
"best_model_checkpoint": "./polyglot-tagger-v2.2/checkpoint-60000",
"epoch": 2.0,
"eval_steps": 2500,
"global_step": 62154,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003217865589754316,
"grad_norm": 12.599648475646973,
"learning_rate": 4.9920359108022015e-05,
"loss": 8.340531616210937,
"step": 100
},
{
"epoch": 0.006435731179508632,
"grad_norm": 16.90213966369629,
"learning_rate": 4.9839913762589704e-05,
"loss": 2.2957839965820312,
"step": 200
},
{
"epoch": 0.009653596769262947,
"grad_norm": 4.964681625366211,
"learning_rate": 4.975946841715739e-05,
"loss": 1.2708330535888672,
"step": 300
},
{
"epoch": 0.012871462359017264,
"grad_norm": 9.05567455291748,
"learning_rate": 4.9679023071725076e-05,
"loss": 0.8362879180908203,
"step": 400
},
{
"epoch": 0.01608932794877158,
"grad_norm": 6.991761207580566,
"learning_rate": 4.959857772629276e-05,
"loss": 0.7179710388183593,
"step": 500
},
{
"epoch": 0.019307193538525894,
"grad_norm": 6.725792407989502,
"learning_rate": 4.951813238086044e-05,
"loss": 0.6352015686035156,
"step": 600
},
{
"epoch": 0.022525059128280213,
"grad_norm": 8.771743774414062,
"learning_rate": 4.943768703542813e-05,
"loss": 0.5740464782714844,
"step": 700
},
{
"epoch": 0.025742924718034528,
"grad_norm": 9.227128028869629,
"learning_rate": 4.935724168999582e-05,
"loss": 0.5493228149414062,
"step": 800
},
{
"epoch": 0.028960790307788843,
"grad_norm": 11.525238990783691,
"learning_rate": 4.927679634456351e-05,
"loss": 0.5246843719482421,
"step": 900
},
{
"epoch": 0.03217865589754316,
"grad_norm": 5.464803695678711,
"learning_rate": 4.919635099913119e-05,
"loss": 0.5209083938598633,
"step": 1000
},
{
"epoch": 0.03539652148729747,
"grad_norm": 16.38031005859375,
"learning_rate": 4.911590565369888e-05,
"loss": 0.4483478546142578,
"step": 1100
},
{
"epoch": 0.03861438707705179,
"grad_norm": 8.507850646972656,
"learning_rate": 4.903546030826657e-05,
"loss": 0.4467829513549805,
"step": 1200
},
{
"epoch": 0.04183225266680611,
"grad_norm": 10.285760879516602,
"learning_rate": 4.895501496283425e-05,
"loss": 0.44222518920898435,
"step": 1300
},
{
"epoch": 0.045050118256560426,
"grad_norm": 8.84359359741211,
"learning_rate": 4.887456961740194e-05,
"loss": 0.3752718734741211,
"step": 1400
},
{
"epoch": 0.04826798384631474,
"grad_norm": 5.227468490600586,
"learning_rate": 4.879412427196963e-05,
"loss": 0.4211923217773437,
"step": 1500
},
{
"epoch": 0.051485849436069056,
"grad_norm": 5.467648983001709,
"learning_rate": 4.871367892653732e-05,
"loss": 0.3770075988769531,
"step": 1600
},
{
"epoch": 0.05470371502582337,
"grad_norm": 7.041491508483887,
"learning_rate": 4.8633233581105e-05,
"loss": 0.39339431762695315,
"step": 1700
},
{
"epoch": 0.057921580615577686,
"grad_norm": 4.243207931518555,
"learning_rate": 4.855278823567269e-05,
"loss": 0.38666561126708987,
"step": 1800
},
{
"epoch": 0.061139446205332,
"grad_norm": 10.47419548034668,
"learning_rate": 4.847234289024038e-05,
"loss": 0.37535915374755857,
"step": 1900
},
{
"epoch": 0.06435731179508632,
"grad_norm": 7.903866291046143,
"learning_rate": 4.839189754480806e-05,
"loss": 0.3448226165771484,
"step": 2000
},
{
"epoch": 0.06757517738484063,
"grad_norm": 4.801281929016113,
"learning_rate": 4.831145219937574e-05,
"loss": 0.3735826873779297,
"step": 2100
},
{
"epoch": 0.07079304297459495,
"grad_norm": 6.289346218109131,
"learning_rate": 4.823100685394343e-05,
"loss": 0.3530204391479492,
"step": 2200
},
{
"epoch": 0.07401090856434926,
"grad_norm": 6.725964546203613,
"learning_rate": 4.815056150851112e-05,
"loss": 0.37652462005615234,
"step": 2300
},
{
"epoch": 0.07722877415410358,
"grad_norm": 14.810294151306152,
"learning_rate": 4.80701161630788e-05,
"loss": 0.3361575698852539,
"step": 2400
},
{
"epoch": 0.08044663974385789,
"grad_norm": 10.04340934753418,
"learning_rate": 4.798967081764649e-05,
"loss": 0.33932579040527344,
"step": 2500
},
{
"epoch": 0.08044663974385789,
"eval_accuracy": 0.9722890395064231,
"eval_f1": 0.8388115225531707,
"eval_loss": 0.10779489576816559,
"eval_precision": 0.7998833070774257,
"eval_recall": 0.8817226431355398,
"eval_runtime": 47.993,
"eval_samples_per_second": 520.91,
"eval_steps_per_second": 28.942,
"step": 2500
},
{
"epoch": 0.08366450533361222,
"grad_norm": 5.665945529937744,
"learning_rate": 4.790922547221418e-05,
"loss": 0.3147930717468262,
"step": 2600
},
{
"epoch": 0.08688237092336654,
"grad_norm": 5.679959297180176,
"learning_rate": 4.782878012678187e-05,
"loss": 0.363752555847168,
"step": 2700
},
{
"epoch": 0.09010023651312085,
"grad_norm": 5.174692153930664,
"learning_rate": 4.774833478134955e-05,
"loss": 0.3410203552246094,
"step": 2800
},
{
"epoch": 0.09331810210287517,
"grad_norm": 5.585685729980469,
"learning_rate": 4.766788943591724e-05,
"loss": 0.3211274337768555,
"step": 2900
},
{
"epoch": 0.09653596769262948,
"grad_norm": 5.2783684730529785,
"learning_rate": 4.758744409048493e-05,
"loss": 0.3486956787109375,
"step": 3000
},
{
"epoch": 0.0997538332823838,
"grad_norm": 6.57411527633667,
"learning_rate": 4.750699874505261e-05,
"loss": 0.33664627075195314,
"step": 3100
},
{
"epoch": 0.10297169887213811,
"grad_norm": 8.044240951538086,
"learning_rate": 4.74265533996203e-05,
"loss": 0.31327726364135744,
"step": 3200
},
{
"epoch": 0.10618956446189243,
"grad_norm": 6.461747169494629,
"learning_rate": 4.734610805418799e-05,
"loss": 0.32588130950927735,
"step": 3300
},
{
"epoch": 0.10940743005164674,
"grad_norm": 2.9853570461273193,
"learning_rate": 4.726566270875568e-05,
"loss": 0.28595542907714844,
"step": 3400
},
{
"epoch": 0.11262529564140106,
"grad_norm": 3.4314420223236084,
"learning_rate": 4.718521736332336e-05,
"loss": 0.32015865325927734,
"step": 3500
},
{
"epoch": 0.11584316123115537,
"grad_norm": 5.1868205070495605,
"learning_rate": 4.7104772017891044e-05,
"loss": 0.29926179885864257,
"step": 3600
},
{
"epoch": 0.11906102682090969,
"grad_norm": 5.119034767150879,
"learning_rate": 4.702432667245873e-05,
"loss": 0.30344295501708984,
"step": 3700
},
{
"epoch": 0.122278892410664,
"grad_norm": 10.863602638244629,
"learning_rate": 4.694388132702642e-05,
"loss": 0.282437686920166,
"step": 3800
},
{
"epoch": 0.12549675800041832,
"grad_norm": 3.252387046813965,
"learning_rate": 4.6863435981594104e-05,
"loss": 0.2832884407043457,
"step": 3900
},
{
"epoch": 0.12871462359017263,
"grad_norm": 12.616849899291992,
"learning_rate": 4.678299063616179e-05,
"loss": 0.299547119140625,
"step": 4000
},
{
"epoch": 0.13193248917992695,
"grad_norm": 7.093480110168457,
"learning_rate": 4.670254529072948e-05,
"loss": 0.26917409896850586,
"step": 4100
},
{
"epoch": 0.13515035476968126,
"grad_norm": 8.318635940551758,
"learning_rate": 4.6622099945297164e-05,
"loss": 0.28246700286865234,
"step": 4200
},
{
"epoch": 0.13836822035943558,
"grad_norm": 6.450869560241699,
"learning_rate": 4.654165459986485e-05,
"loss": 0.2775975036621094,
"step": 4300
},
{
"epoch": 0.1415860859491899,
"grad_norm": 7.253055572509766,
"learning_rate": 4.646120925443254e-05,
"loss": 0.2734929466247559,
"step": 4400
},
{
"epoch": 0.1448039515389442,
"grad_norm": 2.5073742866516113,
"learning_rate": 4.638076390900023e-05,
"loss": 0.2605240631103516,
"step": 4500
},
{
"epoch": 0.14802181712869852,
"grad_norm": 4.528222560882568,
"learning_rate": 4.6300318563567914e-05,
"loss": 0.2719747543334961,
"step": 4600
},
{
"epoch": 0.15123968271845284,
"grad_norm": 6.474442481994629,
"learning_rate": 4.62198732181356e-05,
"loss": 0.27226806640625,
"step": 4700
},
{
"epoch": 0.15445754830820715,
"grad_norm": 4.291607856750488,
"learning_rate": 4.613942787270329e-05,
"loss": 0.29810441970825197,
"step": 4800
},
{
"epoch": 0.15767541389796147,
"grad_norm": 2.773998737335205,
"learning_rate": 4.6058982527270974e-05,
"loss": 0.2813092041015625,
"step": 4900
},
{
"epoch": 0.16089327948771578,
"grad_norm": 8.771388053894043,
"learning_rate": 4.597853718183866e-05,
"loss": 0.3041937828063965,
"step": 5000
},
{
"epoch": 0.16089327948771578,
"eval_accuracy": 0.9766050453886422,
"eval_f1": 0.8726156897994607,
"eval_loss": 0.09095935523509979,
"eval_precision": 0.8421638308564865,
"eval_recall": 0.9053523880577816,
"eval_runtime": 38.7855,
"eval_samples_per_second": 644.572,
"eval_steps_per_second": 35.812,
"step": 5000
},
{
"epoch": 0.16411114507747013,
"grad_norm": 6.708637237548828,
"learning_rate": 4.5898091836406345e-05,
"loss": 0.25013633728027346,
"step": 5100
},
{
"epoch": 0.16732901066722444,
"grad_norm": 4.836703777313232,
"learning_rate": 4.5817646490974034e-05,
"loss": 0.27049646377563474,
"step": 5200
},
{
"epoch": 0.17054687625697876,
"grad_norm": 2.853039026260376,
"learning_rate": 4.5737201145541717e-05,
"loss": 0.2682499885559082,
"step": 5300
},
{
"epoch": 0.17376474184673307,
"grad_norm": 2.6769561767578125,
"learning_rate": 4.5656755800109406e-05,
"loss": 0.25310728073120115,
"step": 5400
},
{
"epoch": 0.1769826074364874,
"grad_norm": 3.740751028060913,
"learning_rate": 4.5576310454677095e-05,
"loss": 0.2738076400756836,
"step": 5500
},
{
"epoch": 0.1802004730262417,
"grad_norm": 5.003463268280029,
"learning_rate": 4.549586510924478e-05,
"loss": 0.23465911865234376,
"step": 5600
},
{
"epoch": 0.18341833861599602,
"grad_norm": 4.04716682434082,
"learning_rate": 4.5415419763812466e-05,
"loss": 0.2690970230102539,
"step": 5700
},
{
"epoch": 0.18663620420575033,
"grad_norm": 4.878298282623291,
"learning_rate": 4.5334974418380155e-05,
"loss": 0.254027156829834,
"step": 5800
},
{
"epoch": 0.18985406979550465,
"grad_norm": 6.428683757781982,
"learning_rate": 4.5254529072947844e-05,
"loss": 0.2576713562011719,
"step": 5900
},
{
"epoch": 0.19307193538525896,
"grad_norm": 4.079443454742432,
"learning_rate": 4.5174083727515526e-05,
"loss": 0.27297264099121094,
"step": 6000
},
{
"epoch": 0.19628980097501328,
"grad_norm": 1.6320624351501465,
"learning_rate": 4.5093638382083215e-05,
"loss": 0.24345703125,
"step": 6100
},
{
"epoch": 0.1995076665647676,
"grad_norm": 4.895444869995117,
"learning_rate": 4.5013193036650904e-05,
"loss": 0.2668925857543945,
"step": 6200
},
{
"epoch": 0.2027255321545219,
"grad_norm": 15.62102222442627,
"learning_rate": 4.493274769121859e-05,
"loss": 0.25465171813964843,
"step": 6300
},
{
"epoch": 0.20594339774427622,
"grad_norm": 3.3204216957092285,
"learning_rate": 4.4852302345786276e-05,
"loss": 0.2394422721862793,
"step": 6400
},
{
"epoch": 0.20916126333403054,
"grad_norm": 2.538597822189331,
"learning_rate": 4.4771857000353965e-05,
"loss": 0.2686458206176758,
"step": 6500
},
{
"epoch": 0.21237912892378485,
"grad_norm": 3.806776523590088,
"learning_rate": 4.469141165492165e-05,
"loss": 0.2445933151245117,
"step": 6600
},
{
"epoch": 0.21559699451353917,
"grad_norm": 5.519633769989014,
"learning_rate": 4.4610966309489336e-05,
"loss": 0.2316486930847168,
"step": 6700
},
{
"epoch": 0.21881486010329348,
"grad_norm": 2.570289134979248,
"learning_rate": 4.453052096405702e-05,
"loss": 0.25618520736694333,
"step": 6800
},
{
"epoch": 0.2220327256930478,
"grad_norm": 1.9880187511444092,
"learning_rate": 4.445007561862471e-05,
"loss": 0.22628854751586913,
"step": 6900
},
{
"epoch": 0.22525059128280212,
"grad_norm": 5.40452766418457,
"learning_rate": 4.4369630273192396e-05,
"loss": 0.23788238525390626,
"step": 7000
},
{
"epoch": 0.22846845687255643,
"grad_norm": 3.7318766117095947,
"learning_rate": 4.428918492776008e-05,
"loss": 0.23460798263549804,
"step": 7100
},
{
"epoch": 0.23168632246231075,
"grad_norm": 3.2017786502838135,
"learning_rate": 4.420873958232777e-05,
"loss": 0.2477778434753418,
"step": 7200
},
{
"epoch": 0.23490418805206506,
"grad_norm": 5.915706157684326,
"learning_rate": 4.412829423689546e-05,
"loss": 0.2488957405090332,
"step": 7300
},
{
"epoch": 0.23812205364181938,
"grad_norm": 2.817690849304199,
"learning_rate": 4.404784889146314e-05,
"loss": 0.24542184829711913,
"step": 7400
},
{
"epoch": 0.2413399192315737,
"grad_norm": 4.39451789855957,
"learning_rate": 4.396740354603083e-05,
"loss": 0.2323160743713379,
"step": 7500
},
{
"epoch": 0.2413399192315737,
"eval_accuracy": 0.9765879819893235,
"eval_f1": 0.8910072212088652,
"eval_loss": 0.08586511015892029,
"eval_precision": 0.8661088374239462,
"eval_recall": 0.9173795037367669,
"eval_runtime": 38.8792,
"eval_samples_per_second": 643.017,
"eval_steps_per_second": 35.726,
"step": 7500
},
{
"epoch": 0.244557784821328,
"grad_norm": 6.589978218078613,
"learning_rate": 4.388695820059852e-05,
"loss": 0.25242706298828127,
"step": 7600
},
{
"epoch": 0.24777565041108232,
"grad_norm": 8.314870834350586,
"learning_rate": 4.3806512855166206e-05,
"loss": 0.23692258834838867,
"step": 7700
},
{
"epoch": 0.25099351600083664,
"grad_norm": 1.829958200454712,
"learning_rate": 4.372606750973389e-05,
"loss": 0.2302927017211914,
"step": 7800
},
{
"epoch": 0.254211381590591,
"grad_norm": 2.879436492919922,
"learning_rate": 4.364562216430158e-05,
"loss": 0.22530841827392578,
"step": 7900
},
{
"epoch": 0.25742924718034527,
"grad_norm": 4.005130767822266,
"learning_rate": 4.3565176818869266e-05,
"loss": 0.22251163482666014,
"step": 8000
},
{
"epoch": 0.2606471127700996,
"grad_norm": 6.088663578033447,
"learning_rate": 4.348473147343695e-05,
"loss": 0.2417246437072754,
"step": 8100
},
{
"epoch": 0.2638649783598539,
"grad_norm": 2.821535587310791,
"learning_rate": 4.340428612800463e-05,
"loss": 0.2411360549926758,
"step": 8200
},
{
"epoch": 0.26708284394960824,
"grad_norm": 3.160569190979004,
"learning_rate": 4.332384078257232e-05,
"loss": 0.23668193817138672,
"step": 8300
},
{
"epoch": 0.2703007095393625,
"grad_norm": 1.9559069871902466,
"learning_rate": 4.324339543714001e-05,
"loss": 0.20959726333618164,
"step": 8400
},
{
"epoch": 0.27351857512911687,
"grad_norm": 3.570617914199829,
"learning_rate": 4.316295009170769e-05,
"loss": 0.23034614562988281,
"step": 8500
},
{
"epoch": 0.27673644071887116,
"grad_norm": 15.547330856323242,
"learning_rate": 4.308250474627538e-05,
"loss": 0.22180200576782227,
"step": 8600
},
{
"epoch": 0.2799543063086255,
"grad_norm": 1.9708352088928223,
"learning_rate": 4.300205940084307e-05,
"loss": 0.22687753677368164,
"step": 8700
},
{
"epoch": 0.2831721718983798,
"grad_norm": 2.9480574131011963,
"learning_rate": 4.292161405541076e-05,
"loss": 0.22114805221557618,
"step": 8800
},
{
"epoch": 0.28639003748813413,
"grad_norm": 1.7439308166503906,
"learning_rate": 4.284116870997844e-05,
"loss": 0.20100093841552735,
"step": 8900
},
{
"epoch": 0.2896079030778884,
"grad_norm": 8.816022872924805,
"learning_rate": 4.276072336454613e-05,
"loss": 0.21243839263916015,
"step": 9000
},
{
"epoch": 0.29282576866764276,
"grad_norm": 2.90824556350708,
"learning_rate": 4.268027801911382e-05,
"loss": 0.22030885696411132,
"step": 9100
},
{
"epoch": 0.29604363425739705,
"grad_norm": 1.6333738565444946,
"learning_rate": 4.25998326736815e-05,
"loss": 0.20564973831176758,
"step": 9200
},
{
"epoch": 0.2992614998471514,
"grad_norm": 2.048818349838257,
"learning_rate": 4.251938732824919e-05,
"loss": 0.22784378051757812,
"step": 9300
},
{
"epoch": 0.3024793654369057,
"grad_norm": 2.5882511138916016,
"learning_rate": 4.243894198281688e-05,
"loss": 0.2064011001586914,
"step": 9400
},
{
"epoch": 0.30569723102666,
"grad_norm": 4.423720359802246,
"learning_rate": 4.235849663738457e-05,
"loss": 0.22651321411132813,
"step": 9500
},
{
"epoch": 0.3089150966164143,
"grad_norm": 6.1782732009887695,
"learning_rate": 4.227805129195225e-05,
"loss": 0.2322876739501953,
"step": 9600
},
{
"epoch": 0.31213296220616865,
"grad_norm": 6.551365852355957,
"learning_rate": 4.219760594651993e-05,
"loss": 0.19593393325805664,
"step": 9700
},
{
"epoch": 0.31535082779592294,
"grad_norm": 3.024858236312866,
"learning_rate": 4.211716060108762e-05,
"loss": 0.21738645553588867,
"step": 9800
},
{
"epoch": 0.3185686933856773,
"grad_norm": 3.413978338241577,
"learning_rate": 4.203671525565531e-05,
"loss": 0.23522287368774414,
"step": 9900
},
{
"epoch": 0.32178655897543157,
"grad_norm": 5.521674156188965,
"learning_rate": 4.195626991022299e-05,
"loss": 0.2253404998779297,
"step": 10000
},
{
"epoch": 0.32178655897543157,
"eval_accuracy": 0.9818083480354339,
"eval_f1": 0.9029431672892276,
"eval_loss": 0.07069914042949677,
"eval_precision": 0.8804620179673655,
"eval_recall": 0.9266024362948766,
"eval_runtime": 38.4445,
"eval_samples_per_second": 650.288,
"eval_steps_per_second": 36.13,
"step": 10000
},
{
"epoch": 0.3250044245651859,
"grad_norm": 3.399749755859375,
"learning_rate": 4.187582456479068e-05,
"loss": 0.20138219833374024,
"step": 10100
},
{
"epoch": 0.32822229015494025,
"grad_norm": 6.1640753746032715,
"learning_rate": 4.179537921935837e-05,
"loss": 0.2074588966369629,
"step": 10200
},
{
"epoch": 0.33144015574469454,
"grad_norm": 2.652939796447754,
"learning_rate": 4.171493387392605e-05,
"loss": 0.20934364318847656,
"step": 10300
},
{
"epoch": 0.3346580213344489,
"grad_norm": 2.3221635818481445,
"learning_rate": 4.163448852849374e-05,
"loss": 0.20557579040527343,
"step": 10400
},
{
"epoch": 0.3378758869242032,
"grad_norm": 4.995593547821045,
"learning_rate": 4.155404318306143e-05,
"loss": 0.21284431457519531,
"step": 10500
},
{
"epoch": 0.3410937525139575,
"grad_norm": 5.589059829711914,
"learning_rate": 4.147359783762912e-05,
"loss": 0.22939311981201171,
"step": 10600
},
{
"epoch": 0.3443116181037118,
"grad_norm": 3.8397393226623535,
"learning_rate": 4.13931524921968e-05,
"loss": 0.20622701644897462,
"step": 10700
},
{
"epoch": 0.34752948369346615,
"grad_norm": 6.400503158569336,
"learning_rate": 4.131270714676449e-05,
"loss": 0.2185076332092285,
"step": 10800
},
{
"epoch": 0.35074734928322043,
"grad_norm": 8.725358009338379,
"learning_rate": 4.123226180133218e-05,
"loss": 0.22046119689941407,
"step": 10900
},
{
"epoch": 0.3539652148729748,
"grad_norm": 2.3542211055755615,
"learning_rate": 4.115181645589986e-05,
"loss": 0.21081886291503907,
"step": 11000
},
{
"epoch": 0.35718308046272906,
"grad_norm": 2.4933526515960693,
"learning_rate": 4.107137111046755e-05,
"loss": 0.21268159866333008,
"step": 11100
},
{
"epoch": 0.3604009460524834,
"grad_norm": 2.3393359184265137,
"learning_rate": 4.0990925765035234e-05,
"loss": 0.20191749572753906,
"step": 11200
},
{
"epoch": 0.3636188116422377,
"grad_norm": 7.511013507843018,
"learning_rate": 4.091048041960292e-05,
"loss": 0.22198728561401368,
"step": 11300
},
{
"epoch": 0.36683667723199204,
"grad_norm": 2.7252819538116455,
"learning_rate": 4.0830035074170606e-05,
"loss": 0.18651321411132812,
"step": 11400
},
{
"epoch": 0.3700545428217463,
"grad_norm": 1.7606490850448608,
"learning_rate": 4.0749589728738295e-05,
"loss": 0.20632423400878908,
"step": 11500
},
{
"epoch": 0.37327240841150067,
"grad_norm": 1.7523146867752075,
"learning_rate": 4.0669144383305984e-05,
"loss": 0.19867513656616212,
"step": 11600
},
{
"epoch": 0.37649027400125495,
"grad_norm": 1.9296563863754272,
"learning_rate": 4.058869903787367e-05,
"loss": 0.21059492111206055,
"step": 11700
},
{
"epoch": 0.3797081395910093,
"grad_norm": 2.119272232055664,
"learning_rate": 4.0508253692441355e-05,
"loss": 0.20072717666625978,
"step": 11800
},
{
"epoch": 0.3829260051807636,
"grad_norm": 1.8676605224609375,
"learning_rate": 4.0427808347009044e-05,
"loss": 0.2199223518371582,
"step": 11900
},
{
"epoch": 0.3861438707705179,
"grad_norm": 7.326332092285156,
"learning_rate": 4.034736300157673e-05,
"loss": 0.22979413986206054,
"step": 12000
},
{
"epoch": 0.3893617363602722,
"grad_norm": 6.403603553771973,
"learning_rate": 4.0266917656144415e-05,
"loss": 0.20364051818847656,
"step": 12100
},
{
"epoch": 0.39257960195002656,
"grad_norm": 7.924391269683838,
"learning_rate": 4.0186472310712104e-05,
"loss": 0.21068944931030273,
"step": 12200
},
{
"epoch": 0.39579746753978085,
"grad_norm": 1.8581446409225464,
"learning_rate": 4.010602696527979e-05,
"loss": 0.1954026412963867,
"step": 12300
},
{
"epoch": 0.3990153331295352,
"grad_norm": 4.413888931274414,
"learning_rate": 4.002558161984748e-05,
"loss": 0.20379119873046875,
"step": 12400
},
{
"epoch": 0.4022331987192895,
"grad_norm": 7.339420795440674,
"learning_rate": 3.9945136274415165e-05,
"loss": 0.21170831680297852,
"step": 12500
},
{
"epoch": 0.4022331987192895,
"eval_accuracy": 0.9805167004142786,
"eval_f1": 0.9110038183596832,
"eval_loss": 0.07153428345918655,
"eval_precision": 0.8942803330689929,
"eval_recall": 0.9283646981644178,
"eval_runtime": 38.6581,
"eval_samples_per_second": 646.695,
"eval_steps_per_second": 35.93,
"step": 12500
},
{
"epoch": 0.4054510643090438,
"grad_norm": 6.818169116973877,
"learning_rate": 3.9864690928982854e-05,
"loss": 0.1984395217895508,
"step": 12600
},
{
"epoch": 0.4086689298987981,
"grad_norm": 7.034976959228516,
"learning_rate": 3.9784245583550536e-05,
"loss": 0.20739023208618165,
"step": 12700
},
{
"epoch": 0.41188679548855245,
"grad_norm": 1.944700002670288,
"learning_rate": 3.9703800238118225e-05,
"loss": 0.2069017219543457,
"step": 12800
},
{
"epoch": 0.41510466107830674,
"grad_norm": 4.255757808685303,
"learning_rate": 3.962335489268591e-05,
"loss": 0.20037817001342773,
"step": 12900
},
{
"epoch": 0.4183225266680611,
"grad_norm": 3.274329900741577,
"learning_rate": 3.9542909547253596e-05,
"loss": 0.20747562408447265,
"step": 13000
},
{
"epoch": 0.42154039225781537,
"grad_norm": 8.684778213500977,
"learning_rate": 3.9462464201821285e-05,
"loss": 0.2101104164123535,
"step": 13100
},
{
"epoch": 0.4247582578475697,
"grad_norm": 2.2195351123809814,
"learning_rate": 3.938201885638897e-05,
"loss": 0.2074994659423828,
"step": 13200
},
{
"epoch": 0.427976123437324,
"grad_norm": 5.522794723510742,
"learning_rate": 3.9301573510956657e-05,
"loss": 0.20569501876831053,
"step": 13300
},
{
"epoch": 0.43119398902707834,
"grad_norm": 8.248220443725586,
"learning_rate": 3.9221128165524346e-05,
"loss": 0.20211294174194336,
"step": 13400
},
{
"epoch": 0.4344118546168327,
"grad_norm": 1.3259437084197998,
"learning_rate": 3.9140682820092035e-05,
"loss": 0.19619279861450195,
"step": 13500
},
{
"epoch": 0.43762972020658697,
"grad_norm": 3.264453172683716,
"learning_rate": 3.906023747465972e-05,
"loss": 0.20112485885620118,
"step": 13600
},
{
"epoch": 0.4408475857963413,
"grad_norm": 6.523902893066406,
"learning_rate": 3.8979792129227406e-05,
"loss": 0.17402765274047852,
"step": 13700
},
{
"epoch": 0.4440654513860956,
"grad_norm": 6.627138137817383,
"learning_rate": 3.8899346783795095e-05,
"loss": 0.18946147918701173,
"step": 13800
},
{
"epoch": 0.44728331697584994,
"grad_norm": 1.7494491338729858,
"learning_rate": 3.881890143836278e-05,
"loss": 0.20346969604492188,
"step": 13900
},
{
"epoch": 0.45050118256560423,
"grad_norm": 6.50437068939209,
"learning_rate": 3.8738456092930466e-05,
"loss": 0.19907255172729493,
"step": 14000
},
{
"epoch": 0.4537190481553586,
"grad_norm": 1.948766827583313,
"learning_rate": 3.8658010747498155e-05,
"loss": 0.19304462432861327,
"step": 14100
},
{
"epoch": 0.45693691374511286,
"grad_norm": 2.1899731159210205,
"learning_rate": 3.857756540206584e-05,
"loss": 0.19804704666137696,
"step": 14200
},
{
"epoch": 0.4601547793348672,
"grad_norm": 2.7552740573883057,
"learning_rate": 3.8497120056633527e-05,
"loss": 0.17073585510253905,
"step": 14300
},
{
"epoch": 0.4633726449246215,
"grad_norm": 3.2083916664123535,
"learning_rate": 3.841667471120121e-05,
"loss": 0.19360818862915039,
"step": 14400
},
{
"epoch": 0.46659051051437583,
"grad_norm": 5.908588409423828,
"learning_rate": 3.83362293657689e-05,
"loss": 0.17787866592407225,
"step": 14500
},
{
"epoch": 0.4698083761041301,
"grad_norm": 2.6561052799224854,
"learning_rate": 3.825578402033658e-05,
"loss": 0.20204214096069337,
"step": 14600
},
{
"epoch": 0.47302624169388446,
"grad_norm": 2.6253390312194824,
"learning_rate": 3.817533867490427e-05,
"loss": 0.18690492630004882,
"step": 14700
},
{
"epoch": 0.47624410728363875,
"grad_norm": 1.7459347248077393,
"learning_rate": 3.809489332947196e-05,
"loss": 0.17405952453613283,
"step": 14800
},
{
"epoch": 0.4794619728733931,
"grad_norm": 4.657139301300049,
"learning_rate": 3.801444798403965e-05,
"loss": 0.1884486961364746,
"step": 14900
},
{
"epoch": 0.4826798384631474,
"grad_norm": 3.4389936923980713,
"learning_rate": 3.793400263860733e-05,
"loss": 0.1895262336730957,
"step": 15000
},
{
"epoch": 0.4826798384631474,
"eval_accuracy": 0.9843394189343752,
"eval_f1": 0.9178045953844602,
"eval_loss": 0.05817751958966255,
"eval_precision": 0.8991940957952288,
"eval_recall": 0.9372017339627738,
"eval_runtime": 38.9423,
"eval_samples_per_second": 641.975,
"eval_steps_per_second": 35.668,
"step": 15000
},
{
"epoch": 0.4858977040529017,
"grad_norm": 2.704439640045166,
"learning_rate": 3.785355729317502e-05,
"loss": 0.1874333381652832,
"step": 15100
},
{
"epoch": 0.489115569642656,
"grad_norm": 3.225116729736328,
"learning_rate": 3.777311194774271e-05,
"loss": 0.20594432830810547,
"step": 15200
},
{
"epoch": 0.49233343523241035,
"grad_norm": 6.360146999359131,
"learning_rate": 3.76926666023104e-05,
"loss": 0.18481210708618165,
"step": 15300
},
{
"epoch": 0.49555130082216464,
"grad_norm": 6.072609901428223,
"learning_rate": 3.761222125687808e-05,
"loss": 0.19338289260864258,
"step": 15400
},
{
"epoch": 0.498769166411919,
"grad_norm": 6.828498363494873,
"learning_rate": 3.753177591144577e-05,
"loss": 0.17112209320068358,
"step": 15500
},
{
"epoch": 0.5019870320016733,
"grad_norm": 1.9960297346115112,
"learning_rate": 3.745133056601346e-05,
"loss": 0.177659912109375,
"step": 15600
},
{
"epoch": 0.5052048975914276,
"grad_norm": 2.489583969116211,
"learning_rate": 3.737088522058114e-05,
"loss": 0.1999117088317871,
"step": 15700
},
{
"epoch": 0.508422763181182,
"grad_norm": 1.4370602369308472,
"learning_rate": 3.729043987514882e-05,
"loss": 0.18860567092895508,
"step": 15800
},
{
"epoch": 0.5116406287709362,
"grad_norm": 1.5181185007095337,
"learning_rate": 3.720999452971651e-05,
"loss": 0.17629049301147462,
"step": 15900
},
{
"epoch": 0.5148584943606905,
"grad_norm": 6.286506652832031,
"learning_rate": 3.71295491842842e-05,
"loss": 0.18211414337158202,
"step": 16000
},
{
"epoch": 0.5180763599504449,
"grad_norm": 3.1108760833740234,
"learning_rate": 3.704910383885188e-05,
"loss": 0.18931018829345703,
"step": 16100
},
{
"epoch": 0.5212942255401992,
"grad_norm": 0.5399872064590454,
"learning_rate": 3.696865849341957e-05,
"loss": 0.18857126235961913,
"step": 16200
},
{
"epoch": 0.5245120911299535,
"grad_norm": 1.3765555620193481,
"learning_rate": 3.688821314798726e-05,
"loss": 0.18414220809936524,
"step": 16300
},
{
"epoch": 0.5277299567197078,
"grad_norm": 3.104323148727417,
"learning_rate": 3.680776780255494e-05,
"loss": 0.18788623809814453,
"step": 16400
},
{
"epoch": 0.5309478223094621,
"grad_norm": 1.609277367591858,
"learning_rate": 3.672732245712263e-05,
"loss": 0.17455484390258788,
"step": 16500
},
{
"epoch": 0.5341656878992165,
"grad_norm": 2.4873690605163574,
"learning_rate": 3.664687711169032e-05,
"loss": 0.17273141860961913,
"step": 16600
},
{
"epoch": 0.5373835534889707,
"grad_norm": 3.004922866821289,
"learning_rate": 3.656643176625801e-05,
"loss": 0.17702035903930663,
"step": 16700
},
{
"epoch": 0.540601419078725,
"grad_norm": 4.731108665466309,
"learning_rate": 3.648598642082569e-05,
"loss": 0.1878594970703125,
"step": 16800
},
{
"epoch": 0.5438192846684794,
"grad_norm": 3.8425683975219727,
"learning_rate": 3.640554107539338e-05,
"loss": 0.19668659210205078,
"step": 16900
},
{
"epoch": 0.5470371502582337,
"grad_norm": 1.5426945686340332,
"learning_rate": 3.632509572996107e-05,
"loss": 0.16921035766601564,
"step": 17000
},
{
"epoch": 0.5502550158479881,
"grad_norm": 1.886399507522583,
"learning_rate": 3.624465038452876e-05,
"loss": 0.15211493492126466,
"step": 17100
},
{
"epoch": 0.5534728814377423,
"grad_norm": 5.827595233917236,
"learning_rate": 3.616420503909644e-05,
"loss": 0.1691839599609375,
"step": 17200
},
{
"epoch": 0.5566907470274967,
"grad_norm": 5.728984832763672,
"learning_rate": 3.608375969366412e-05,
"loss": 0.1721731948852539,
"step": 17300
},
{
"epoch": 0.559908612617251,
"grad_norm": 2.2809104919433594,
"learning_rate": 3.600331434823181e-05,
"loss": 0.1917552185058594,
"step": 17400
},
{
"epoch": 0.5631264782070053,
"grad_norm": 2.936466693878174,
"learning_rate": 3.5922869002799494e-05,
"loss": 0.18653520584106445,
"step": 17500
},
{
"epoch": 0.5631264782070053,
"eval_accuracy": 0.9851181304305561,
"eval_f1": 0.9213907594344987,
"eval_loss": 0.055659033358097076,
"eval_precision": 0.9052643338588151,
"eval_recall": 0.9381021597355321,
"eval_runtime": 38.4689,
"eval_samples_per_second": 649.876,
"eval_steps_per_second": 36.107,
"step": 17500
},
{
"epoch": 0.5663443437967596,
"grad_norm": 5.648510456085205,
"learning_rate": 3.5842423657367184e-05,
"loss": 0.17743080139160156,
"step": 17600
},
{
"epoch": 0.5695622093865139,
"grad_norm": 2.0115761756896973,
"learning_rate": 3.576197831193487e-05,
"loss": 0.18602855682373046,
"step": 17700
},
{
"epoch": 0.5727800749762683,
"grad_norm": 5.880187511444092,
"learning_rate": 3.568153296650256e-05,
"loss": 0.18167713165283203,
"step": 17800
},
{
"epoch": 0.5759979405660226,
"grad_norm": 4.690539360046387,
"learning_rate": 3.5601087621070244e-05,
"loss": 0.17695968627929687,
"step": 17900
},
{
"epoch": 0.5792158061557768,
"grad_norm": 0.8806695342063904,
"learning_rate": 3.552064227563793e-05,
"loss": 0.19892288208007813,
"step": 18000
},
{
"epoch": 0.5824336717455312,
"grad_norm": 1.943228006362915,
"learning_rate": 3.544019693020562e-05,
"loss": 0.17242523193359374,
"step": 18100
},
{
"epoch": 0.5856515373352855,
"grad_norm": 3.5915210247039795,
"learning_rate": 3.5359751584773304e-05,
"loss": 0.181878662109375,
"step": 18200
},
{
"epoch": 0.5888694029250399,
"grad_norm": 2.8968303203582764,
"learning_rate": 3.527930623934099e-05,
"loss": 0.18603469848632812,
"step": 18300
},
{
"epoch": 0.5920872685147941,
"grad_norm": 5.2887678146362305,
"learning_rate": 3.519886089390868e-05,
"loss": 0.16267045974731445,
"step": 18400
},
{
"epoch": 0.5953051341045484,
"grad_norm": 1.6592695713043213,
"learning_rate": 3.511841554847637e-05,
"loss": 0.17949144363403322,
"step": 18500
},
{
"epoch": 0.5985229996943028,
"grad_norm": 0.979318380355835,
"learning_rate": 3.5037970203044054e-05,
"loss": 0.1779278564453125,
"step": 18600
},
{
"epoch": 0.6017408652840571,
"grad_norm": 1.6524354219436646,
"learning_rate": 3.495752485761174e-05,
"loss": 0.1590627098083496,
"step": 18700
},
{
"epoch": 0.6049587308738114,
"grad_norm": 3.5658767223358154,
"learning_rate": 3.4877079512179425e-05,
"loss": 0.16901378631591796,
"step": 18800
},
{
"epoch": 0.6081765964635657,
"grad_norm": 3.9737935066223145,
"learning_rate": 3.4796634166747114e-05,
"loss": 0.17097728729248046,
"step": 18900
},
{
"epoch": 0.61139446205332,
"grad_norm": 3.3529911041259766,
"learning_rate": 3.4716188821314796e-05,
"loss": 0.17232301712036133,
"step": 19000
},
{
"epoch": 0.6146123276430744,
"grad_norm": 3.2977654933929443,
"learning_rate": 3.4635743475882485e-05,
"loss": 0.16828372955322266,
"step": 19100
},
{
"epoch": 0.6178301932328286,
"grad_norm": 3.297731876373291,
"learning_rate": 3.4555298130450174e-05,
"loss": 0.17515941619873046,
"step": 19200
},
{
"epoch": 0.621048058822583,
"grad_norm": 7.258795261383057,
"learning_rate": 3.4474852785017856e-05,
"loss": 0.15848381996154784,
"step": 19300
},
{
"epoch": 0.6242659244123373,
"grad_norm": 3.2193796634674072,
"learning_rate": 3.4394407439585546e-05,
"loss": 0.1693815231323242,
"step": 19400
},
{
"epoch": 0.6274837900020916,
"grad_norm": 1.8956769704818726,
"learning_rate": 3.4313962094153235e-05,
"loss": 0.1632681083679199,
"step": 19500
},
{
"epoch": 0.6307016555918459,
"grad_norm": 1.5155613422393799,
"learning_rate": 3.4233516748720924e-05,
"loss": 0.1833029556274414,
"step": 19600
},
{
"epoch": 0.6339195211816002,
"grad_norm": 4.2836079597473145,
"learning_rate": 3.4153071403288606e-05,
"loss": 0.17777244567871095,
"step": 19700
},
{
"epoch": 0.6371373867713546,
"grad_norm": 5.861156463623047,
"learning_rate": 3.4072626057856295e-05,
"loss": 0.17992183685302734,
"step": 19800
},
{
"epoch": 0.6403552523611089,
"grad_norm": 3.4068503379821777,
"learning_rate": 3.3992180712423984e-05,
"loss": 0.17448163986206056,
"step": 19900
},
{
"epoch": 0.6435731179508631,
"grad_norm": 2.3339030742645264,
"learning_rate": 3.3911735366991666e-05,
"loss": 0.16662296295166015,
"step": 20000
},
{
"epoch": 0.6435731179508631,
"eval_accuracy": 0.985229301062481,
"eval_f1": 0.92316609857673,
"eval_loss": 0.056044481694698334,
"eval_precision": 0.9047148608263973,
"eval_recall": 0.9423856137687964,
"eval_runtime": 38.8504,
"eval_samples_per_second": 643.494,
"eval_steps_per_second": 35.753,
"step": 20000
},
{
"epoch": 0.6467909835406175,
"grad_norm": 3.3607404232025146,
"learning_rate": 3.3831290021559355e-05,
"loss": 0.17825605392456054,
"step": 20100
},
{
"epoch": 0.6500088491303718,
"grad_norm": 2.690410852432251,
"learning_rate": 3.3750844676127044e-05,
"loss": 0.1796116065979004,
"step": 20200
},
{
"epoch": 0.6532267147201262,
"grad_norm": 5.667492866516113,
"learning_rate": 3.3670399330694727e-05,
"loss": 0.15698086738586425,
"step": 20300
},
{
"epoch": 0.6564445803098805,
"grad_norm": 2.9585955142974854,
"learning_rate": 3.3589953985262416e-05,
"loss": 0.14914603233337403,
"step": 20400
},
{
"epoch": 0.6596624458996347,
"grad_norm": 3.539393901824951,
"learning_rate": 3.35095086398301e-05,
"loss": 0.15797216415405274,
"step": 20500
},
{
"epoch": 0.6628803114893891,
"grad_norm": 3.268399238586426,
"learning_rate": 3.342906329439779e-05,
"loss": 0.16447280883789062,
"step": 20600
},
{
"epoch": 0.6660981770791434,
"grad_norm": 1.261047601699829,
"learning_rate": 3.3348617948965476e-05,
"loss": 0.18879241943359376,
"step": 20700
},
{
"epoch": 0.6693160426688978,
"grad_norm": 2.1420369148254395,
"learning_rate": 3.326817260353316e-05,
"loss": 0.15382347106933594,
"step": 20800
},
{
"epoch": 0.672533908258652,
"grad_norm": 2.8946328163146973,
"learning_rate": 3.318772725810085e-05,
"loss": 0.15514832496643066,
"step": 20900
},
{
"epoch": 0.6757517738484063,
"grad_norm": 12.205283164978027,
"learning_rate": 3.3107281912668536e-05,
"loss": 0.18069747924804688,
"step": 21000
},
{
"epoch": 0.6789696394381607,
"grad_norm": 1.8224153518676758,
"learning_rate": 3.302683656723622e-05,
"loss": 0.17442039489746095,
"step": 21100
},
{
"epoch": 0.682187505027915,
"grad_norm": 3.392157554626465,
"learning_rate": 3.294639122180391e-05,
"loss": 0.16294483184814454,
"step": 21200
},
{
"epoch": 0.6854053706176693,
"grad_norm": 1.9432575702667236,
"learning_rate": 3.2865945876371597e-05,
"loss": 0.17553016662597656,
"step": 21300
},
{
"epoch": 0.6886232362074236,
"grad_norm": 3.2393903732299805,
"learning_rate": 3.2785500530939286e-05,
"loss": 0.17193687438964844,
"step": 21400
},
{
"epoch": 0.691841101797178,
"grad_norm": 3.1432416439056396,
"learning_rate": 3.270505518550697e-05,
"loss": 0.17834131240844728,
"step": 21500
},
{
"epoch": 0.6950589673869323,
"grad_norm": 1.5839548110961914,
"learning_rate": 3.262460984007466e-05,
"loss": 0.16497987747192383,
"step": 21600
},
{
"epoch": 0.6982768329766865,
"grad_norm": 3.1208765506744385,
"learning_rate": 3.2544164494642346e-05,
"loss": 0.18163761138916015,
"step": 21700
},
{
"epoch": 0.7014946985664409,
"grad_norm": 8.323920249938965,
"learning_rate": 3.246371914921003e-05,
"loss": 0.1704619598388672,
"step": 21800
},
{
"epoch": 0.7047125641561952,
"grad_norm": 0.9279898405075073,
"learning_rate": 3.238327380377772e-05,
"loss": 0.1771290397644043,
"step": 21900
},
{
"epoch": 0.7079304297459496,
"grad_norm": 3.2795073986053467,
"learning_rate": 3.23028284583454e-05,
"loss": 0.1524122428894043,
"step": 22000
},
{
"epoch": 0.7111482953357038,
"grad_norm": 5.583573818206787,
"learning_rate": 3.222238311291309e-05,
"loss": 0.16980670928955077,
"step": 22100
},
{
"epoch": 0.7143661609254581,
"grad_norm": 4.969387531280518,
"learning_rate": 3.214193776748077e-05,
"loss": 0.1560416603088379,
"step": 22200
},
{
"epoch": 0.7175840265152125,
"grad_norm": 3.3920233249664307,
"learning_rate": 3.206149242204846e-05,
"loss": 0.160349063873291,
"step": 22300
},
{
"epoch": 0.7208018921049668,
"grad_norm": 2.8945395946502686,
"learning_rate": 3.198104707661615e-05,
"loss": 0.18149927139282226,
"step": 22400
},
{
"epoch": 0.724019757694721,
"grad_norm": 1.1273301839828491,
"learning_rate": 3.190060173118384e-05,
"loss": 0.16230302810668945,
"step": 22500
},
{
"epoch": 0.724019757694721,
"eval_accuracy": 0.9838156242825616,
"eval_f1": 0.9249598360552048,
"eval_loss": 0.05883582681417465,
"eval_precision": 0.9098816590137007,
"eval_recall": 0.9405461725473045,
"eval_runtime": 38.7455,
"eval_samples_per_second": 645.236,
"eval_steps_per_second": 35.849,
"step": 22500
},
{
"epoch": 0.7272376232844754,
"grad_norm": 2.699639081954956,
"learning_rate": 3.182015638575152e-05,
"loss": 0.17641155242919923,
"step": 22600
},
{
"epoch": 0.7304554888742297,
"grad_norm": 4.614871978759766,
"learning_rate": 3.173971104031921e-05,
"loss": 0.15983717918395995,
"step": 22700
},
{
"epoch": 0.7336733544639841,
"grad_norm": 2.631415843963623,
"learning_rate": 3.16592656948869e-05,
"loss": 0.15378172874450682,
"step": 22800
},
{
"epoch": 0.7368912200537383,
"grad_norm": 1.4634898900985718,
"learning_rate": 3.157882034945458e-05,
"loss": 0.16743200302124023,
"step": 22900
},
{
"epoch": 0.7401090856434926,
"grad_norm": 3.4641151428222656,
"learning_rate": 3.149837500402227e-05,
"loss": 0.15191415786743165,
"step": 23000
},
{
"epoch": 0.743326951233247,
"grad_norm": 1.500380039215088,
"learning_rate": 3.141792965858996e-05,
"loss": 0.15310038566589357,
"step": 23100
},
{
"epoch": 0.7465448168230013,
"grad_norm": 2.184008836746216,
"learning_rate": 3.133748431315765e-05,
"loss": 0.16115007400512696,
"step": 23200
},
{
"epoch": 0.7497626824127556,
"grad_norm": 1.1047471761703491,
"learning_rate": 3.125703896772533e-05,
"loss": 0.153313045501709,
"step": 23300
},
{
"epoch": 0.7529805480025099,
"grad_norm": 3.3007121086120605,
"learning_rate": 3.117659362229301e-05,
"loss": 0.16872770309448243,
"step": 23400
},
{
"epoch": 0.7561984135922643,
"grad_norm": 6.547897815704346,
"learning_rate": 3.10961482768607e-05,
"loss": 0.18551017761230468,
"step": 23500
},
{
"epoch": 0.7594162791820186,
"grad_norm": 2.700106382369995,
"learning_rate": 3.1015702931428383e-05,
"loss": 0.15708372116088867,
"step": 23600
},
{
"epoch": 0.7626341447717729,
"grad_norm": 1.265189290046692,
"learning_rate": 3.093525758599607e-05,
"loss": 0.17149799346923827,
"step": 23700
},
{
"epoch": 0.7658520103615272,
"grad_norm": 6.2682204246521,
"learning_rate": 3.085481224056376e-05,
"loss": 0.15998153686523436,
"step": 23800
},
{
"epoch": 0.7690698759512815,
"grad_norm": 3.071823835372925,
"learning_rate": 3.077436689513145e-05,
"loss": 0.18916090011596678,
"step": 23900
},
{
"epoch": 0.7722877415410359,
"grad_norm": 2.476935625076294,
"learning_rate": 3.069392154969913e-05,
"loss": 0.14498735427856446,
"step": 24000
},
{
"epoch": 0.7755056071307902,
"grad_norm": 4.1790032386779785,
"learning_rate": 3.061347620426682e-05,
"loss": 0.1612476921081543,
"step": 24100
},
{
"epoch": 0.7787234727205444,
"grad_norm": 1.0075314044952393,
"learning_rate": 3.053303085883451e-05,
"loss": 0.14984708786010742,
"step": 24200
},
{
"epoch": 0.7819413383102988,
"grad_norm": 9.008003234863281,
"learning_rate": 3.04525855134022e-05,
"loss": 0.17614728927612305,
"step": 24300
},
{
"epoch": 0.7851592039000531,
"grad_norm": 0.790449321269989,
"learning_rate": 3.0372140167969882e-05,
"loss": 0.1434452724456787,
"step": 24400
},
{
"epoch": 0.7883770694898075,
"grad_norm": 2.6443631649017334,
"learning_rate": 3.029169482253757e-05,
"loss": 0.1625998878479004,
"step": 24500
},
{
"epoch": 0.7915949350795617,
"grad_norm": 1.1716026067733765,
"learning_rate": 3.0211249477105257e-05,
"loss": 0.1629466438293457,
"step": 24600
},
{
"epoch": 0.794812800669316,
"grad_norm": 1.9636585712432861,
"learning_rate": 3.013080413167294e-05,
"loss": 0.1526513195037842,
"step": 24700
},
{
"epoch": 0.7980306662590704,
"grad_norm": 6.442037582397461,
"learning_rate": 3.0050358786240628e-05,
"loss": 0.16380990982055665,
"step": 24800
},
{
"epoch": 0.8012485318488247,
"grad_norm": 2.0020787715911865,
"learning_rate": 2.9969913440808317e-05,
"loss": 0.16642404556274415,
"step": 24900
},
{
"epoch": 0.804466397438579,
"grad_norm": 3.425807476043701,
"learning_rate": 2.9889468095376006e-05,
"loss": 0.15638751029968262,
"step": 25000
},
{
"epoch": 0.804466397438579,
"eval_accuracy": 0.9871683237123339,
"eval_f1": 0.9358503491978916,
"eval_loss": 0.047558050602674484,
"eval_precision": 0.9231722273671564,
"eval_recall": 0.948881542557981,
"eval_runtime": 39.0981,
"eval_samples_per_second": 639.417,
"eval_steps_per_second": 35.526,
"step": 25000
},
{
"epoch": 0.8076842630283333,
"grad_norm": 1.7773306369781494,
"learning_rate": 2.980902274994369e-05,
"loss": 0.16818260192871093,
"step": 25100
},
{
"epoch": 0.8109021286180876,
"grad_norm": 2.457798957824707,
"learning_rate": 2.9728577404511378e-05,
"loss": 0.1736893844604492,
"step": 25200
},
{
"epoch": 0.814119994207842,
"grad_norm": 1.6925740242004395,
"learning_rate": 2.9648132059079063e-05,
"loss": 0.17085908889770507,
"step": 25300
},
{
"epoch": 0.8173378597975962,
"grad_norm": 2.4007983207702637,
"learning_rate": 2.956768671364675e-05,
"loss": 0.16143144607543947,
"step": 25400
},
{
"epoch": 0.8205557253873506,
"grad_norm": 2.328049898147583,
"learning_rate": 2.9487241368214435e-05,
"loss": 0.16922952651977538,
"step": 25500
},
{
"epoch": 0.8237735909771049,
"grad_norm": 2.7495510578155518,
"learning_rate": 2.9406796022782124e-05,
"loss": 0.16985824584960937,
"step": 25600
},
{
"epoch": 0.8269914565668592,
"grad_norm": 18.3980712890625,
"learning_rate": 2.9326350677349813e-05,
"loss": 0.15559816360473633,
"step": 25700
},
{
"epoch": 0.8302093221566135,
"grad_norm": 2.3473148345947266,
"learning_rate": 2.9245905331917495e-05,
"loss": 0.17103708267211915,
"step": 25800
},
{
"epoch": 0.8334271877463678,
"grad_norm": 2.7980849742889404,
"learning_rate": 2.9165459986485184e-05,
"loss": 0.1549097156524658,
"step": 25900
},
{
"epoch": 0.8366450533361222,
"grad_norm": 0.4983047842979431,
"learning_rate": 2.908501464105287e-05,
"loss": 0.15225051879882812,
"step": 26000
},
{
"epoch": 0.8398629189258765,
"grad_norm": 1.1898616552352905,
"learning_rate": 2.900456929562056e-05,
"loss": 0.14232300758361816,
"step": 26100
},
{
"epoch": 0.8430807845156307,
"grad_norm": 1.2203755378723145,
"learning_rate": 2.892412395018824e-05,
"loss": 0.16007034301757814,
"step": 26200
},
{
"epoch": 0.8462986501053851,
"grad_norm": 3.160578489303589,
"learning_rate": 2.884367860475593e-05,
"loss": 0.16384252548217773,
"step": 26300
},
{
"epoch": 0.8495165156951394,
"grad_norm": 5.011970043182373,
"learning_rate": 2.876323325932362e-05,
"loss": 0.15432297706604003,
"step": 26400
},
{
"epoch": 0.8527343812848938,
"grad_norm": 0.8514712452888489,
"learning_rate": 2.86827879138913e-05,
"loss": 0.15506745338439942,
"step": 26500
},
{
"epoch": 0.855952246874648,
"grad_norm": 9.957659721374512,
"learning_rate": 2.860234256845899e-05,
"loss": 0.15732930183410646,
"step": 26600
},
{
"epoch": 0.8591701124644023,
"grad_norm": 8.271772384643555,
"learning_rate": 2.852189722302668e-05,
"loss": 0.1552633285522461,
"step": 26700
},
{
"epoch": 0.8623879780541567,
"grad_norm": 2.02590274810791,
"learning_rate": 2.8441451877594365e-05,
"loss": 0.15745894432067872,
"step": 26800
},
{
"epoch": 0.865605843643911,
"grad_norm": 4.529170989990234,
"learning_rate": 2.836100653216205e-05,
"loss": 0.1467604446411133,
"step": 26900
},
{
"epoch": 0.8688237092336654,
"grad_norm": 5.628500938415527,
"learning_rate": 2.8280561186729736e-05,
"loss": 0.1447477436065674,
"step": 27000
},
{
"epoch": 0.8720415748234196,
"grad_norm": 12.87569522857666,
"learning_rate": 2.8200115841297425e-05,
"loss": 0.14656261444091798,
"step": 27100
},
{
"epoch": 0.8752594404131739,
"grad_norm": 3.9200944900512695,
"learning_rate": 2.8119670495865107e-05,
"loss": 0.16447486877441406,
"step": 27200
},
{
"epoch": 0.8784773060029283,
"grad_norm": 1.2977664470672607,
"learning_rate": 2.8039225150432797e-05,
"loss": 0.15925183296203613,
"step": 27300
},
{
"epoch": 0.8816951715926826,
"grad_norm": 2.786237955093384,
"learning_rate": 2.7958779805000486e-05,
"loss": 0.15786813735961913,
"step": 27400
},
{
"epoch": 0.8849130371824369,
"grad_norm": 4.701409339904785,
"learning_rate": 2.787833445956817e-05,
"loss": 0.14735154151916505,
"step": 27500
},
{
"epoch": 0.8849130371824369,
"eval_accuracy": 0.985545749558937,
"eval_f1": 0.9339101573386709,
"eval_loss": 0.052501607686281204,
"eval_precision": 0.9199600673862857,
"eval_recall": 0.9482898341930256,
"eval_runtime": 39.0393,
"eval_samples_per_second": 640.38,
"eval_steps_per_second": 35.58,
"step": 27500
},
{
"epoch": 0.8881309027721912,
"grad_norm": 1.7882884740829468,
"learning_rate": 2.7797889114135857e-05,
"loss": 0.14846997261047362,
"step": 27600
},
{
"epoch": 0.8913487683619455,
"grad_norm": 1.8413563966751099,
"learning_rate": 2.7717443768703542e-05,
"loss": 0.15191203117370605,
"step": 27700
},
{
"epoch": 0.8945666339516999,
"grad_norm": 15.72754192352295,
"learning_rate": 2.763699842327123e-05,
"loss": 0.1671205520629883,
"step": 27800
},
{
"epoch": 0.8977844995414541,
"grad_norm": 7.281972885131836,
"learning_rate": 2.755655307783892e-05,
"loss": 0.15349125862121582,
"step": 27900
},
{
"epoch": 0.9010023651312085,
"grad_norm": 2.539133071899414,
"learning_rate": 2.7476107732406603e-05,
"loss": 0.1401151466369629,
"step": 28000
},
{
"epoch": 0.9042202307209628,
"grad_norm": 1.8258554935455322,
"learning_rate": 2.7395662386974292e-05,
"loss": 0.14998969078063965,
"step": 28100
},
{
"epoch": 0.9074380963107171,
"grad_norm": 6.016611099243164,
"learning_rate": 2.731521704154198e-05,
"loss": 0.14337385177612305,
"step": 28200
},
{
"epoch": 0.9106559619004714,
"grad_norm": 3.221945285797119,
"learning_rate": 2.7234771696109663e-05,
"loss": 0.15551527023315428,
"step": 28300
},
{
"epoch": 0.9138738274902257,
"grad_norm": 1.4967395067214966,
"learning_rate": 2.715432635067735e-05,
"loss": 0.13511601448059082,
"step": 28400
},
{
"epoch": 0.9170916930799801,
"grad_norm": 2.9666309356689453,
"learning_rate": 2.7073881005245038e-05,
"loss": 0.1585793113708496,
"step": 28500
},
{
"epoch": 0.9203095586697344,
"grad_norm": 0.9828789234161377,
"learning_rate": 2.6993435659812727e-05,
"loss": 0.15394386291503906,
"step": 28600
},
{
"epoch": 0.9235274242594886,
"grad_norm": 0.7447288036346436,
"learning_rate": 2.691299031438041e-05,
"loss": 0.1461707878112793,
"step": 28700
},
{
"epoch": 0.926745289849243,
"grad_norm": 1.8215960264205933,
"learning_rate": 2.6832544968948098e-05,
"loss": 0.14361499786376952,
"step": 28800
},
{
"epoch": 0.9299631554389973,
"grad_norm": 5.320079326629639,
"learning_rate": 2.6752099623515787e-05,
"loss": 0.15090553283691407,
"step": 28900
},
{
"epoch": 0.9331810210287517,
"grad_norm": 1.9044791460037231,
"learning_rate": 2.667165427808347e-05,
"loss": 0.16826240539550782,
"step": 29000
},
{
"epoch": 0.9363988866185059,
"grad_norm": 2.378955602645874,
"learning_rate": 2.659120893265116e-05,
"loss": 0.14869707107543945,
"step": 29100
},
{
"epoch": 0.9396167522082602,
"grad_norm": 3.2531590461730957,
"learning_rate": 2.6510763587218844e-05,
"loss": 0.15361614227294923,
"step": 29200
},
{
"epoch": 0.9428346177980146,
"grad_norm": 3.1207923889160156,
"learning_rate": 2.6430318241786533e-05,
"loss": 0.14827921867370605,
"step": 29300
},
{
"epoch": 0.9460524833877689,
"grad_norm": 8.111202239990234,
"learning_rate": 2.6349872896354215e-05,
"loss": 0.14765314102172852,
"step": 29400
},
{
"epoch": 0.9492703489775232,
"grad_norm": 2.2104053497314453,
"learning_rate": 2.6269427550921905e-05,
"loss": 0.17542333602905275,
"step": 29500
},
{
"epoch": 0.9524882145672775,
"grad_norm": 0.9287620782852173,
"learning_rate": 2.6188982205489594e-05,
"loss": 0.14969840049743652,
"step": 29600
},
{
"epoch": 0.9557060801570318,
"grad_norm": 2.604642868041992,
"learning_rate": 2.6108536860057283e-05,
"loss": 0.16406068801879883,
"step": 29700
},
{
"epoch": 0.9589239457467862,
"grad_norm": 7.544447422027588,
"learning_rate": 2.6028091514624965e-05,
"loss": 0.1531221103668213,
"step": 29800
},
{
"epoch": 0.9621418113365404,
"grad_norm": 3.618058204650879,
"learning_rate": 2.594764616919265e-05,
"loss": 0.16164751052856446,
"step": 29900
},
{
"epoch": 0.9653596769262948,
"grad_norm": 2.8580358028411865,
"learning_rate": 2.586720082376034e-05,
"loss": 0.15803168296813966,
"step": 30000
},
{
"epoch": 0.9653596769262948,
"eval_accuracy": 0.9877764012153277,
"eval_f1": 0.9371598218703612,
"eval_loss": 0.04550632834434509,
"eval_precision": 0.9245153134660876,
"eval_recall": 0.9501550018651677,
"eval_runtime": 39.0526,
"eval_samples_per_second": 640.163,
"eval_steps_per_second": 35.567,
"step": 30000
},
{
"epoch": 0.9685775425160491,
"grad_norm": 1.4846822023391724,
"learning_rate": 2.5786755478328022e-05,
"loss": 0.138107852935791,
"step": 30100
},
{
"epoch": 0.9717954081058034,
"grad_norm": 2.2517590522766113,
"learning_rate": 2.570631013289571e-05,
"loss": 0.13565559387207032,
"step": 30200
},
{
"epoch": 0.9750132736955578,
"grad_norm": 3.6406161785125732,
"learning_rate": 2.56258647874634e-05,
"loss": 0.1619549560546875,
"step": 30300
},
{
"epoch": 0.978231139285312,
"grad_norm": 2.2304012775421143,
"learning_rate": 2.554541944203109e-05,
"loss": 0.13655080795288085,
"step": 30400
},
{
"epoch": 0.9814490048750664,
"grad_norm": 4.99883508682251,
"learning_rate": 2.546497409659877e-05,
"loss": 0.16668659210205078,
"step": 30500
},
{
"epoch": 0.9846668704648207,
"grad_norm": 2.1127796173095703,
"learning_rate": 2.538452875116646e-05,
"loss": 0.14745793342590333,
"step": 30600
},
{
"epoch": 0.987884736054575,
"grad_norm": 0.5811536908149719,
"learning_rate": 2.5304083405734146e-05,
"loss": 0.1281309223175049,
"step": 30700
},
{
"epoch": 0.9911026016443293,
"grad_norm": 3.0965957641601562,
"learning_rate": 2.5223638060301828e-05,
"loss": 0.14012389183044432,
"step": 30800
},
{
"epoch": 0.9943204672340836,
"grad_norm": 3.952867031097412,
"learning_rate": 2.5143192714869517e-05,
"loss": 0.18041374206542968,
"step": 30900
},
{
"epoch": 0.997538332823838,
"grad_norm": 2.5378260612487793,
"learning_rate": 2.5062747369437206e-05,
"loss": 0.14559303283691405,
"step": 31000
},
{
"epoch": 1.0007401090856436,
"grad_norm": 2.2179856300354004,
"learning_rate": 2.4982302024004892e-05,
"loss": 0.13755855560302735,
"step": 31100
},
{
"epoch": 1.0039579746753977,
"grad_norm": 1.619863510131836,
"learning_rate": 2.490185667857258e-05,
"loss": 0.11792629241943359,
"step": 31200
},
{
"epoch": 1.007175840265152,
"grad_norm": 1.596127986907959,
"learning_rate": 2.4821411333140267e-05,
"loss": 0.1253152847290039,
"step": 31300
},
{
"epoch": 1.0103937058549064,
"grad_norm": 5.185761451721191,
"learning_rate": 2.4740965987707952e-05,
"loss": 0.11788433074951171,
"step": 31400
},
{
"epoch": 1.0136115714446607,
"grad_norm": 4.540038585662842,
"learning_rate": 2.4660520642275638e-05,
"loss": 0.10257969856262207,
"step": 31500
},
{
"epoch": 1.016829437034415,
"grad_norm": 0.530720055103302,
"learning_rate": 2.4580075296843323e-05,
"loss": 0.11232701301574707,
"step": 31600
},
{
"epoch": 1.0200473026241694,
"grad_norm": 1.8916836977005005,
"learning_rate": 2.4499629951411013e-05,
"loss": 0.11638439178466797,
"step": 31700
},
{
"epoch": 1.0232651682139238,
"grad_norm": 1.7947441339492798,
"learning_rate": 2.4419184605978698e-05,
"loss": 0.1160141658782959,
"step": 31800
},
{
"epoch": 1.0264830338036781,
"grad_norm": 0.8208370804786682,
"learning_rate": 2.4338739260546387e-05,
"loss": 0.09769786834716797,
"step": 31900
},
{
"epoch": 1.0297008993934322,
"grad_norm": 2.4557855129241943,
"learning_rate": 2.4258293915114073e-05,
"loss": 0.1321820545196533,
"step": 32000
},
{
"epoch": 1.0329187649831866,
"grad_norm": 1.885315179824829,
"learning_rate": 2.4177848569681762e-05,
"loss": 0.10417740821838378,
"step": 32100
},
{
"epoch": 1.036136630572941,
"grad_norm": 0.7314161658287048,
"learning_rate": 2.4097403224249444e-05,
"loss": 0.11431515693664551,
"step": 32200
},
{
"epoch": 1.0393544961626953,
"grad_norm": 1.2714358568191528,
"learning_rate": 2.4016957878817133e-05,
"loss": 0.121004638671875,
"step": 32300
},
{
"epoch": 1.0425723617524496,
"grad_norm": 3.3896045684814453,
"learning_rate": 2.393651253338482e-05,
"loss": 0.11515800476074219,
"step": 32400
},
{
"epoch": 1.045790227342204,
"grad_norm": 1.288989782333374,
"learning_rate": 2.3856067187952504e-05,
"loss": 0.12445229530334473,
"step": 32500
},
{
"epoch": 1.045790227342204,
"eval_accuracy": 0.9881399033280868,
"eval_f1": 0.9421632544130543,
"eval_loss": 0.04471902176737785,
"eval_precision": 0.9309708286765537,
"eval_recall": 0.9536280727029496,
"eval_runtime": 38.6457,
"eval_samples_per_second": 646.903,
"eval_steps_per_second": 35.942,
"step": 32500
},
{
"epoch": 1.0490080929319583,
"grad_norm": 1.3189036846160889,
"learning_rate": 2.3775621842520194e-05,
"loss": 0.11217801094055176,
"step": 32600
},
{
"epoch": 1.0522259585217126,
"grad_norm": 3.2094037532806396,
"learning_rate": 2.369517649708788e-05,
"loss": 0.12271183013916015,
"step": 32700
},
{
"epoch": 1.055443824111467,
"grad_norm": 9.113394737243652,
"learning_rate": 2.3614731151655568e-05,
"loss": 0.11827351570129395,
"step": 32800
},
{
"epoch": 1.058661689701221,
"grad_norm": 1.4682255983352661,
"learning_rate": 2.3534285806223254e-05,
"loss": 0.12056323051452637,
"step": 32900
},
{
"epoch": 1.0618795552909754,
"grad_norm": 1.7021067142486572,
"learning_rate": 2.345384046079094e-05,
"loss": 0.10995281219482422,
"step": 33000
},
{
"epoch": 1.0650974208807298,
"grad_norm": 0.7944279313087463,
"learning_rate": 2.3373395115358625e-05,
"loss": 0.10448121070861817,
"step": 33100
},
{
"epoch": 1.0683152864704841,
"grad_norm": 1.9725573062896729,
"learning_rate": 2.3292949769926314e-05,
"loss": 0.12772043228149413,
"step": 33200
},
{
"epoch": 1.0715331520602385,
"grad_norm": 4.521146297454834,
"learning_rate": 2.3212504424494e-05,
"loss": 0.11528082847595215,
"step": 33300
},
{
"epoch": 1.0747510176499928,
"grad_norm": 3.5572879314422607,
"learning_rate": 2.3132059079061685e-05,
"loss": 0.1119957447052002,
"step": 33400
},
{
"epoch": 1.0779688832397472,
"grad_norm": 4.957686424255371,
"learning_rate": 2.3051613733629375e-05,
"loss": 0.1071746826171875,
"step": 33500
},
{
"epoch": 1.0811867488295013,
"grad_norm": 1.005800724029541,
"learning_rate": 2.297116838819706e-05,
"loss": 0.11798532485961914,
"step": 33600
},
{
"epoch": 1.0844046144192556,
"grad_norm": 1.6847596168518066,
"learning_rate": 2.2890723042764746e-05,
"loss": 0.12464913368225097,
"step": 33700
},
{
"epoch": 1.08762248000901,
"grad_norm": 2.2746946811676025,
"learning_rate": 2.281027769733243e-05,
"loss": 0.11306841850280762,
"step": 33800
},
{
"epoch": 1.0908403455987643,
"grad_norm": 7.0843706130981445,
"learning_rate": 2.272983235190012e-05,
"loss": 0.10698895454406739,
"step": 33900
},
{
"epoch": 1.0940582111885186,
"grad_norm": 1.731086015701294,
"learning_rate": 2.2649387006467806e-05,
"loss": 0.11211064338684082,
"step": 34000
},
{
"epoch": 1.097276076778273,
"grad_norm": 3.32621431350708,
"learning_rate": 2.2568941661035495e-05,
"loss": 0.10381333351135254,
"step": 34100
},
{
"epoch": 1.1004939423680273,
"grad_norm": 1.3450391292572021,
"learning_rate": 2.248849631560318e-05,
"loss": 0.12325675010681153,
"step": 34200
},
{
"epoch": 1.1037118079577817,
"grad_norm": 6.684245586395264,
"learning_rate": 2.2408050970170866e-05,
"loss": 0.12590248107910157,
"step": 34300
},
{
"epoch": 1.106929673547536,
"grad_norm": 1.1298221349716187,
"learning_rate": 2.2327605624738556e-05,
"loss": 0.11126781463623046,
"step": 34400
},
{
"epoch": 1.1101475391372901,
"grad_norm": 1.185738205909729,
"learning_rate": 2.2247160279306238e-05,
"loss": 0.11205435752868652,
"step": 34500
},
{
"epoch": 1.1133654047270445,
"grad_norm": 5.897251605987549,
"learning_rate": 2.2166714933873927e-05,
"loss": 0.10614056587219238,
"step": 34600
},
{
"epoch": 1.1165832703167988,
"grad_norm": 0.8509872555732727,
"learning_rate": 2.2086269588441612e-05,
"loss": 0.11909424781799316,
"step": 34700
},
{
"epoch": 1.1198011359065532,
"grad_norm": 1.7664456367492676,
"learning_rate": 2.20058242430093e-05,
"loss": 0.10704885482788086,
"step": 34800
},
{
"epoch": 1.1230190014963075,
"grad_norm": 4.05808687210083,
"learning_rate": 2.1925378897576987e-05,
"loss": 0.1148582649230957,
"step": 34900
},
{
"epoch": 1.1262368670860619,
"grad_norm": 0.689573347568512,
"learning_rate": 2.1844933552144676e-05,
"loss": 0.11986784934997559,
"step": 35000
},
{
"epoch": 1.1262368670860619,
"eval_accuracy": 0.988106293602156,
"eval_f1": 0.9430045230472126,
"eval_loss": 0.04454037919640541,
"eval_precision": 0.9315594603074993,
"eval_recall": 0.9547343100809097,
"eval_runtime": 38.4818,
"eval_samples_per_second": 649.658,
"eval_steps_per_second": 36.095,
"step": 35000
},
{
"epoch": 1.1294547326758162,
"grad_norm": 1.2704740762710571,
"learning_rate": 2.1764488206712362e-05,
"loss": 0.11099970817565918,
"step": 35100
},
{
"epoch": 1.1326725982655705,
"grad_norm": 5.144682884216309,
"learning_rate": 2.1684042861280047e-05,
"loss": 0.13487217903137208,
"step": 35200
},
{
"epoch": 1.1358904638553247,
"grad_norm": 0.7087175846099854,
"learning_rate": 2.1603597515847733e-05,
"loss": 0.12489970207214356,
"step": 35300
},
{
"epoch": 1.139108329445079,
"grad_norm": 1.559787631034851,
"learning_rate": 2.152315217041542e-05,
"loss": 0.10530885696411132,
"step": 35400
},
{
"epoch": 1.1423261950348333,
"grad_norm": 1.2589160203933716,
"learning_rate": 2.1442706824983108e-05,
"loss": 0.12130918502807617,
"step": 35500
},
{
"epoch": 1.1455440606245877,
"grad_norm": 1.1291875839233398,
"learning_rate": 2.1362261479550793e-05,
"loss": 0.11352409362792969,
"step": 35600
},
{
"epoch": 1.148761926214342,
"grad_norm": 5.766851425170898,
"learning_rate": 2.1281816134118483e-05,
"loss": 0.12505414009094237,
"step": 35700
},
{
"epoch": 1.1519797918040964,
"grad_norm": 4.1735334396362305,
"learning_rate": 2.1201370788686168e-05,
"loss": 0.10489568710327149,
"step": 35800
},
{
"epoch": 1.1551976573938507,
"grad_norm": 0.7965012192726135,
"learning_rate": 2.1120925443253857e-05,
"loss": 0.10613764762878418,
"step": 35900
},
{
"epoch": 1.158415522983605,
"grad_norm": 11.224800109863281,
"learning_rate": 2.104048009782154e-05,
"loss": 0.10370283126831055,
"step": 36000
},
{
"epoch": 1.1616333885733594,
"grad_norm": 2.968989372253418,
"learning_rate": 2.0960034752389225e-05,
"loss": 0.1149828052520752,
"step": 36100
},
{
"epoch": 1.1648512541631135,
"grad_norm": 2.068894863128662,
"learning_rate": 2.0879589406956914e-05,
"loss": 0.10543420791625976,
"step": 36200
},
{
"epoch": 1.1680691197528679,
"grad_norm": 1.2261930704116821,
"learning_rate": 2.07991440615246e-05,
"loss": 0.11141412734985351,
"step": 36300
},
{
"epoch": 1.1712869853426222,
"grad_norm": 3.737278699874878,
"learning_rate": 2.071869871609229e-05,
"loss": 0.11346253395080566,
"step": 36400
},
{
"epoch": 1.1745048509323766,
"grad_norm": 2.6356663703918457,
"learning_rate": 2.0638253370659974e-05,
"loss": 0.12470024108886718,
"step": 36500
},
{
"epoch": 1.177722716522131,
"grad_norm": 1.931261658668518,
"learning_rate": 2.0557808025227664e-05,
"loss": 0.12472087860107423,
"step": 36600
},
{
"epoch": 1.1809405821118852,
"grad_norm": 3.1169793605804443,
"learning_rate": 2.047736267979535e-05,
"loss": 0.10721843719482421,
"step": 36700
},
{
"epoch": 1.1841584477016396,
"grad_norm": 3.3439762592315674,
"learning_rate": 2.0396917334363035e-05,
"loss": 0.115928316116333,
"step": 36800
},
{
"epoch": 1.1873763132913937,
"grad_norm": 5.898672103881836,
"learning_rate": 2.031647198893072e-05,
"loss": 0.1268135643005371,
"step": 36900
},
{
"epoch": 1.190594178881148,
"grad_norm": 1.7046561241149902,
"learning_rate": 2.0236026643498406e-05,
"loss": 0.11421506881713867,
"step": 37000
},
{
"epoch": 1.1938120444709024,
"grad_norm": 1.223960041999817,
"learning_rate": 2.0155581298066095e-05,
"loss": 0.11065722465515136,
"step": 37100
},
{
"epoch": 1.1970299100606567,
"grad_norm": 5.873408794403076,
"learning_rate": 2.007513595263378e-05,
"loss": 0.12281797409057617,
"step": 37200
},
{
"epoch": 1.200247775650411,
"grad_norm": 1.8486683368682861,
"learning_rate": 1.999469060720147e-05,
"loss": 0.1131003189086914,
"step": 37300
},
{
"epoch": 1.2034656412401654,
"grad_norm": 3.4886019229888916,
"learning_rate": 1.9914245261769155e-05,
"loss": 0.11905000686645507,
"step": 37400
},
{
"epoch": 1.2066835068299198,
"grad_norm": 0.7239132523536682,
"learning_rate": 1.983379991633684e-05,
"loss": 0.10931044578552246,
"step": 37500
},
{
"epoch": 1.2066835068299198,
"eval_accuracy": 0.9886145760727707,
"eval_f1": 0.9460371209613769,
"eval_loss": 0.043151482939720154,
"eval_precision": 0.9363463404185619,
"eval_recall": 0.9559305900361457,
"eval_runtime": 38.5909,
"eval_samples_per_second": 647.822,
"eval_steps_per_second": 35.993,
"step": 37500
},
{
"epoch": 1.209901372419674,
"grad_norm": 2.1548306941986084,
"learning_rate": 1.9753354570904527e-05,
"loss": 0.11063767433166503,
"step": 37600
},
{
"epoch": 1.2131192380094284,
"grad_norm": 1.2177164554595947,
"learning_rate": 1.9672909225472216e-05,
"loss": 0.11377819061279297,
"step": 37700
},
{
"epoch": 1.2163371035991826,
"grad_norm": 0.8512117266654968,
"learning_rate": 1.95924638800399e-05,
"loss": 0.1323404121398926,
"step": 37800
},
{
"epoch": 1.219554969188937,
"grad_norm": 1.077217936515808,
"learning_rate": 1.9512018534607587e-05,
"loss": 0.0988365364074707,
"step": 37900
},
{
"epoch": 1.2227728347786913,
"grad_norm": 2.1567108631134033,
"learning_rate": 1.9431573189175276e-05,
"loss": 0.12120766639709472,
"step": 38000
},
{
"epoch": 1.2259907003684456,
"grad_norm": 3.18619966506958,
"learning_rate": 1.9351127843742962e-05,
"loss": 0.11067086219787597,
"step": 38100
},
{
"epoch": 1.2292085659582,
"grad_norm": 1.0777678489685059,
"learning_rate": 1.927068249831065e-05,
"loss": 0.11158087730407715,
"step": 38200
},
{
"epoch": 1.2324264315479543,
"grad_norm": 2.5809645652770996,
"learning_rate": 1.9190237152878333e-05,
"loss": 0.11114447593688964,
"step": 38300
},
{
"epoch": 1.2356442971377086,
"grad_norm": 1.8847070932388306,
"learning_rate": 1.9109791807446022e-05,
"loss": 0.11262718200683594,
"step": 38400
},
{
"epoch": 1.238862162727463,
"grad_norm": 7.437971591949463,
"learning_rate": 1.9029346462013708e-05,
"loss": 0.11268477439880371,
"step": 38500
},
{
"epoch": 1.242080028317217,
"grad_norm": 3.669605255126953,
"learning_rate": 1.8948901116581397e-05,
"loss": 0.11750943183898926,
"step": 38600
},
{
"epoch": 1.2452978939069714,
"grad_norm": 2.447467803955078,
"learning_rate": 1.8868455771149082e-05,
"loss": 0.10881879806518555,
"step": 38700
},
{
"epoch": 1.2485157594967258,
"grad_norm": 4.701931953430176,
"learning_rate": 1.8788010425716768e-05,
"loss": 0.12167869567871094,
"step": 38800
},
{
"epoch": 1.2517336250864801,
"grad_norm": 1.6426016092300415,
"learning_rate": 1.8707565080284457e-05,
"loss": 0.12139597892761231,
"step": 38900
},
{
"epoch": 1.2549514906762345,
"grad_norm": 4.622367858886719,
"learning_rate": 1.8627119734852143e-05,
"loss": 0.10588560104370118,
"step": 39000
},
{
"epoch": 1.2581693562659888,
"grad_norm": 1.7007027864456177,
"learning_rate": 1.854667438941983e-05,
"loss": 0.11511992454528809,
"step": 39100
},
{
"epoch": 1.2613872218557431,
"grad_norm": 2.8813023567199707,
"learning_rate": 1.8466229043987514e-05,
"loss": 0.09704429626464844,
"step": 39200
},
{
"epoch": 1.2646050874454975,
"grad_norm": 8.06043815612793,
"learning_rate": 1.8385783698555203e-05,
"loss": 0.11457429885864258,
"step": 39300
},
{
"epoch": 1.2678229530352518,
"grad_norm": 8.746806144714355,
"learning_rate": 1.830533835312289e-05,
"loss": 0.12408617973327636,
"step": 39400
},
{
"epoch": 1.271040818625006,
"grad_norm": 5.757315158843994,
"learning_rate": 1.8224893007690578e-05,
"loss": 0.10381578445434571,
"step": 39500
},
{
"epoch": 1.2742586842147603,
"grad_norm": 0.5180655121803284,
"learning_rate": 1.8144447662258263e-05,
"loss": 0.10882855415344238,
"step": 39600
},
{
"epoch": 1.2774765498045146,
"grad_norm": 2.3184099197387695,
"learning_rate": 1.806400231682595e-05,
"loss": 0.10384510040283203,
"step": 39700
},
{
"epoch": 1.280694415394269,
"grad_norm": 3.118455648422241,
"learning_rate": 1.7983556971393635e-05,
"loss": 0.1031002426147461,
"step": 39800
},
{
"epoch": 1.2839122809840233,
"grad_norm": 10.587071418762207,
"learning_rate": 1.790311162596132e-05,
"loss": 0.11156161308288574,
"step": 39900
},
{
"epoch": 1.2871301465737777,
"grad_norm": 2.5432746410369873,
"learning_rate": 1.782266628052901e-05,
"loss": 0.11637428283691406,
"step": 40000
},
{
"epoch": 1.2871301465737777,
"eval_accuracy": 0.989512214291476,
"eval_f1": 0.9486766456772123,
"eval_loss": 0.040956247597932816,
"eval_precision": 0.9392224099241068,
"eval_recall": 0.9583231499466176,
"eval_runtime": 38.6339,
"eval_samples_per_second": 647.1,
"eval_steps_per_second": 35.953,
"step": 40000
},
{
"epoch": 1.290348012163532,
"grad_norm": 1.766068458557129,
"learning_rate": 1.7742220935096695e-05,
"loss": 0.11921710968017578,
"step": 40100
},
{
"epoch": 1.2935658777532861,
"grad_norm": 1.7064458131790161,
"learning_rate": 1.7661775589664384e-05,
"loss": 0.12581809997558593,
"step": 40200
},
{
"epoch": 1.2967837433430405,
"grad_norm": 1.2497196197509766,
"learning_rate": 1.758133024423207e-05,
"loss": 0.10997524261474609,
"step": 40300
},
{
"epoch": 1.3000016089327948,
"grad_norm": 1.997997760772705,
"learning_rate": 1.750088489879976e-05,
"loss": 0.11178022384643554,
"step": 40400
},
{
"epoch": 1.3032194745225492,
"grad_norm": 1.8823689222335815,
"learning_rate": 1.7420439553367444e-05,
"loss": 0.09935328483581543,
"step": 40500
},
{
"epoch": 1.3064373401123035,
"grad_norm": 4.478331089019775,
"learning_rate": 1.733999420793513e-05,
"loss": 0.11027010917663574,
"step": 40600
},
{
"epoch": 1.3096552057020578,
"grad_norm": 3.932783365249634,
"learning_rate": 1.7259548862502816e-05,
"loss": 0.11079183578491211,
"step": 40700
},
{
"epoch": 1.3128730712918122,
"grad_norm": 2.1589443683624268,
"learning_rate": 1.71791035170705e-05,
"loss": 0.10328941345214844,
"step": 40800
},
{
"epoch": 1.3160909368815665,
"grad_norm": 2.9713399410247803,
"learning_rate": 1.709865817163819e-05,
"loss": 0.1030450439453125,
"step": 40900
},
{
"epoch": 1.3193088024713209,
"grad_norm": 4.732212066650391,
"learning_rate": 1.7018212826205876e-05,
"loss": 0.114616117477417,
"step": 41000
},
{
"epoch": 1.3225266680610752,
"grad_norm": 1.4628472328186035,
"learning_rate": 1.6937767480773565e-05,
"loss": 0.12563533782958985,
"step": 41100
},
{
"epoch": 1.3257445336508293,
"grad_norm": 3.7251741886138916,
"learning_rate": 1.685732213534125e-05,
"loss": 0.11473335266113281,
"step": 41200
},
{
"epoch": 1.3289623992405837,
"grad_norm": 4.524189472198486,
"learning_rate": 1.6776876789908936e-05,
"loss": 0.12875600814819335,
"step": 41300
},
{
"epoch": 1.332180264830338,
"grad_norm": 28.197296142578125,
"learning_rate": 1.6696431444476622e-05,
"loss": 0.11420975685119629,
"step": 41400
},
{
"epoch": 1.3353981304200924,
"grad_norm": 1.819801926612854,
"learning_rate": 1.6615986099044308e-05,
"loss": 0.09828336715698242,
"step": 41500
},
{
"epoch": 1.3386159960098467,
"grad_norm": 1.7372171878814697,
"learning_rate": 1.6535540753611997e-05,
"loss": 0.1114089298248291,
"step": 41600
},
{
"epoch": 1.341833861599601,
"grad_norm": 4.730764865875244,
"learning_rate": 1.6455095408179682e-05,
"loss": 0.11046221733093262,
"step": 41700
},
{
"epoch": 1.3450517271893552,
"grad_norm": 0.6920406222343445,
"learning_rate": 1.637465006274737e-05,
"loss": 0.1076749610900879,
"step": 41800
},
{
"epoch": 1.3482695927791095,
"grad_norm": 1.7678096294403076,
"learning_rate": 1.6294204717315057e-05,
"loss": 0.10196740150451661,
"step": 41900
},
{
"epoch": 1.3514874583688639,
"grad_norm": 7.219323635101318,
"learning_rate": 1.6213759371882746e-05,
"loss": 0.10450594902038574,
"step": 42000
},
{
"epoch": 1.3547053239586182,
"grad_norm": 3.9280190467834473,
"learning_rate": 1.613331402645043e-05,
"loss": 0.11356738090515137,
"step": 42100
},
{
"epoch": 1.3579231895483725,
"grad_norm": 0.7362053394317627,
"learning_rate": 1.6052868681018117e-05,
"loss": 0.11030459403991699,
"step": 42200
},
{
"epoch": 1.3611410551381269,
"grad_norm": 2.8355274200439453,
"learning_rate": 1.5972423335585803e-05,
"loss": 0.10215305328369141,
"step": 42300
},
{
"epoch": 1.3643589207278812,
"grad_norm": 2.266855001449585,
"learning_rate": 1.589197799015349e-05,
"loss": 0.10216259002685547,
"step": 42400
},
{
"epoch": 1.3675767863176356,
"grad_norm": 1.7423440217971802,
"learning_rate": 1.5811532644721178e-05,
"loss": 0.11463526725769042,
"step": 42500
},
{
"epoch": 1.3675767863176356,
"eval_accuracy": 0.9885556297842152,
"eval_f1": 0.9436256775454764,
"eval_loss": 0.043291687965393066,
"eval_precision": 0.9313770939571248,
"eval_recall": 0.9562007177679731,
"eval_runtime": 39.0115,
"eval_samples_per_second": 640.836,
"eval_steps_per_second": 35.605,
"step": 42500
},
{
"epoch": 1.37079465190739,
"grad_norm": 2.6859209537506104,
"learning_rate": 1.5731087299288863e-05,
"loss": 0.10509669303894043,
"step": 42600
},
{
"epoch": 1.3740125174971443,
"grad_norm": 1.952075481414795,
"learning_rate": 1.5650641953856552e-05,
"loss": 0.10759903907775879,
"step": 42700
},
{
"epoch": 1.3772303830868984,
"grad_norm": 2.712860107421875,
"learning_rate": 1.5570196608424238e-05,
"loss": 0.1039089298248291,
"step": 42800
},
{
"epoch": 1.3804482486766527,
"grad_norm": 0.966740608215332,
"learning_rate": 1.5489751262991924e-05,
"loss": 0.11702409744262696,
"step": 42900
},
{
"epoch": 1.383666114266407,
"grad_norm": 0.4163697063922882,
"learning_rate": 1.540930591755961e-05,
"loss": 0.10730401039123535,
"step": 43000
},
{
"epoch": 1.3868839798561614,
"grad_norm": 1.878928780555725,
"learning_rate": 1.53288605721273e-05,
"loss": 0.10289834022521972,
"step": 43100
},
{
"epoch": 1.3901018454459158,
"grad_norm": 0.8312145471572876,
"learning_rate": 1.5248415226694984e-05,
"loss": 0.0965440845489502,
"step": 43200
},
{
"epoch": 1.39331971103567,
"grad_norm": 2.221240282058716,
"learning_rate": 1.516796988126267e-05,
"loss": 0.11579631805419922,
"step": 43300
},
{
"epoch": 1.3965375766254244,
"grad_norm": 1.7558112144470215,
"learning_rate": 1.5087524535830357e-05,
"loss": 0.09858304977416993,
"step": 43400
},
{
"epoch": 1.3997554422151786,
"grad_norm": 2.686286449432373,
"learning_rate": 1.5007079190398043e-05,
"loss": 0.1057161808013916,
"step": 43500
},
{
"epoch": 1.402973307804933,
"grad_norm": 3.1591944694519043,
"learning_rate": 1.4926633844965732e-05,
"loss": 0.08528660774230958,
"step": 43600
},
{
"epoch": 1.4061911733946872,
"grad_norm": 5.5276994705200195,
"learning_rate": 1.4846188499533417e-05,
"loss": 0.11572813987731934,
"step": 43700
},
{
"epoch": 1.4094090389844416,
"grad_norm": 1.2573060989379883,
"learning_rate": 1.4765743154101105e-05,
"loss": 0.11284799575805664,
"step": 43800
},
{
"epoch": 1.412626904574196,
"grad_norm": 1.1216791868209839,
"learning_rate": 1.468529780866879e-05,
"loss": 0.11373154640197754,
"step": 43900
},
{
"epoch": 1.4158447701639503,
"grad_norm": 1.1175236701965332,
"learning_rate": 1.460485246323648e-05,
"loss": 0.11358193397521972,
"step": 44000
},
{
"epoch": 1.4190626357537046,
"grad_norm": 1.5515943765640259,
"learning_rate": 1.4524407117804165e-05,
"loss": 0.0964575481414795,
"step": 44100
},
{
"epoch": 1.422280501343459,
"grad_norm": 1.3432780504226685,
"learning_rate": 1.444396177237185e-05,
"loss": 0.12204754829406739,
"step": 44200
},
{
"epoch": 1.4254983669332133,
"grad_norm": 5.370590686798096,
"learning_rate": 1.4363516426939538e-05,
"loss": 0.10917285919189453,
"step": 44300
},
{
"epoch": 1.4287162325229676,
"grad_norm": 3.957749366760254,
"learning_rate": 1.4283071081507224e-05,
"loss": 0.1064769458770752,
"step": 44400
},
{
"epoch": 1.4319340981127218,
"grad_norm": 9.636327743530273,
"learning_rate": 1.4202625736074913e-05,
"loss": 0.11022495269775391,
"step": 44500
},
{
"epoch": 1.435151963702476,
"grad_norm": 2.991060495376587,
"learning_rate": 1.4122180390642597e-05,
"loss": 0.10742597579956055,
"step": 44600
},
{
"epoch": 1.4383698292922305,
"grad_norm": 2.648939847946167,
"learning_rate": 1.4041735045210286e-05,
"loss": 0.10721823692321777,
"step": 44700
},
{
"epoch": 1.4415876948819848,
"grad_norm": 0.6414406895637512,
"learning_rate": 1.3961289699777971e-05,
"loss": 0.09946588516235351,
"step": 44800
},
{
"epoch": 1.4448055604717391,
"grad_norm": 0.6889192461967468,
"learning_rate": 1.3880844354345659e-05,
"loss": 0.10531152725219727,
"step": 44900
},
{
"epoch": 1.4480234260614935,
"grad_norm": 9.125837326049805,
"learning_rate": 1.3800399008913344e-05,
"loss": 0.09133255004882812,
"step": 45000
},
{
"epoch": 1.4480234260614935,
"eval_accuracy": 0.9893700192971534,
"eval_f1": 0.9493190302072849,
"eval_loss": 0.043362099677324295,
"eval_precision": 0.9403339180474755,
"eval_recall": 0.9584775086505191,
"eval_runtime": 39.1132,
"eval_samples_per_second": 639.171,
"eval_steps_per_second": 35.512,
"step": 45000
},
{
"epoch": 1.4512412916512476,
"grad_norm": 2.54496431350708,
"learning_rate": 1.371995366348103e-05,
"loss": 0.11927374839782715,
"step": 45100
},
{
"epoch": 1.454459157241002,
"grad_norm": 2.2553350925445557,
"learning_rate": 1.3639508318048719e-05,
"loss": 0.10068492889404297,
"step": 45200
},
{
"epoch": 1.4576770228307563,
"grad_norm": 0.6272743940353394,
"learning_rate": 1.3559062972616405e-05,
"loss": 0.09143063545227051,
"step": 45300
},
{
"epoch": 1.4608948884205106,
"grad_norm": 2.9756674766540527,
"learning_rate": 1.3478617627184092e-05,
"loss": 0.11779884338378906,
"step": 45400
},
{
"epoch": 1.464112754010265,
"grad_norm": 1.4076873064041138,
"learning_rate": 1.3398172281751778e-05,
"loss": 0.09415918350219726,
"step": 45500
},
{
"epoch": 1.4673306196000193,
"grad_norm": 2.004279613494873,
"learning_rate": 1.3317726936319467e-05,
"loss": 0.10082883834838867,
"step": 45600
},
{
"epoch": 1.4705484851897737,
"grad_norm": 12.558138847351074,
"learning_rate": 1.3237281590887152e-05,
"loss": 0.10111323356628418,
"step": 45700
},
{
"epoch": 1.473766350779528,
"grad_norm": 3.433985948562622,
"learning_rate": 1.315683624545484e-05,
"loss": 0.10751867294311523,
"step": 45800
},
{
"epoch": 1.4769842163692823,
"grad_norm": 1.3402138948440552,
"learning_rate": 1.3076390900022525e-05,
"loss": 0.09543782234191894,
"step": 45900
},
{
"epoch": 1.4802020819590367,
"grad_norm": 2.1528687477111816,
"learning_rate": 1.2995945554590211e-05,
"loss": 0.10040821075439453,
"step": 46000
},
{
"epoch": 1.4834199475487908,
"grad_norm": 2.317426919937134,
"learning_rate": 1.2915500209157898e-05,
"loss": 0.10733115196228027,
"step": 46100
},
{
"epoch": 1.4866378131385451,
"grad_norm": 2.752028703689575,
"learning_rate": 1.2835054863725584e-05,
"loss": 0.09164957046508788,
"step": 46200
},
{
"epoch": 1.4898556787282995,
"grad_norm": 1.7984188795089722,
"learning_rate": 1.2754609518293273e-05,
"loss": 0.10080486297607422,
"step": 46300
},
{
"epoch": 1.4930735443180538,
"grad_norm": 2.4552342891693115,
"learning_rate": 1.2674164172860959e-05,
"loss": 0.10242198944091797,
"step": 46400
},
{
"epoch": 1.4962914099078082,
"grad_norm": 2.6899032592773438,
"learning_rate": 1.2593718827428646e-05,
"loss": 0.10129733085632324,
"step": 46500
},
{
"epoch": 1.4995092754975625,
"grad_norm": 1.909561038017273,
"learning_rate": 1.2513273481996332e-05,
"loss": 0.1006275749206543,
"step": 46600
},
{
"epoch": 1.5027271410873166,
"grad_norm": 19.42228889465332,
"learning_rate": 1.2432828136564019e-05,
"loss": 0.10171239852905273,
"step": 46700
},
{
"epoch": 1.505945006677071,
"grad_norm": 1.729035496711731,
"learning_rate": 1.2352382791131706e-05,
"loss": 0.11470888137817382,
"step": 46800
},
{
"epoch": 1.5091628722668253,
"grad_norm": 10.81437873840332,
"learning_rate": 1.2271937445699392e-05,
"loss": 0.09871227264404298,
"step": 46900
},
{
"epoch": 1.5123807378565797,
"grad_norm": 3.496492624282837,
"learning_rate": 1.219149210026708e-05,
"loss": 0.10564603805541992,
"step": 47000
},
{
"epoch": 1.515598603446334,
"grad_norm": 0.7296909093856812,
"learning_rate": 1.2111046754834767e-05,
"loss": 0.09953035354614258,
"step": 47100
},
{
"epoch": 1.5188164690360884,
"grad_norm": 0.6311995983123779,
"learning_rate": 1.2030601409402452e-05,
"loss": 0.10630006790161133,
"step": 47200
},
{
"epoch": 1.5220343346258427,
"grad_norm": 2.436532497406006,
"learning_rate": 1.1950156063970138e-05,
"loss": 0.09258381843566894,
"step": 47300
},
{
"epoch": 1.525252200215597,
"grad_norm": 0.557080864906311,
"learning_rate": 1.1869710718537825e-05,
"loss": 0.1228843879699707,
"step": 47400
},
{
"epoch": 1.5284700658053514,
"grad_norm": 7.565468788146973,
"learning_rate": 1.1789265373105513e-05,
"loss": 0.1068018627166748,
"step": 47500
},
{
"epoch": 1.5284700658053514,
"eval_accuracy": 0.989717492156007,
"eval_f1": 0.9515031847133757,
"eval_loss": 0.03974379226565361,
"eval_precision": 0.942391400345702,
"eval_recall": 0.9607928892090403,
"eval_runtime": 38.7298,
"eval_samples_per_second": 645.498,
"eval_steps_per_second": 35.864,
"step": 47500
},
{
"epoch": 1.5316879313951057,
"grad_norm": 2.948582410812378,
"learning_rate": 1.17088200276732e-05,
"loss": 0.10712558746337891,
"step": 47600
},
{
"epoch": 1.53490579698486,
"grad_norm": 6.355142593383789,
"learning_rate": 1.1628374682240886e-05,
"loss": 0.10939033508300781,
"step": 47700
},
{
"epoch": 1.5381236625746144,
"grad_norm": 0.6847637295722961,
"learning_rate": 1.1547929336808573e-05,
"loss": 0.09593602180480958,
"step": 47800
},
{
"epoch": 1.5413415281643685,
"grad_norm": 2.1787171363830566,
"learning_rate": 1.146748399137626e-05,
"loss": 0.09451790809631348,
"step": 47900
},
{
"epoch": 1.5445593937541229,
"grad_norm": 4.488222122192383,
"learning_rate": 1.1387038645943946e-05,
"loss": 0.1111738109588623,
"step": 48000
},
{
"epoch": 1.5477772593438772,
"grad_norm": 5.803431034088135,
"learning_rate": 1.1306593300511632e-05,
"loss": 0.09694368362426758,
"step": 48100
},
{
"epoch": 1.5509951249336316,
"grad_norm": 1.8297266960144043,
"learning_rate": 1.1226147955079319e-05,
"loss": 0.105281982421875,
"step": 48200
},
{
"epoch": 1.5542129905233857,
"grad_norm": 3.511244058609009,
"learning_rate": 1.1145702609647006e-05,
"loss": 0.10829096794128418,
"step": 48300
},
{
"epoch": 1.55743085611314,
"grad_norm": 7.947854995727539,
"learning_rate": 1.1065257264214692e-05,
"loss": 0.11103793144226075,
"step": 48400
},
{
"epoch": 1.5606487217028944,
"grad_norm": 3.7324254512786865,
"learning_rate": 1.098481191878238e-05,
"loss": 0.09467618942260742,
"step": 48500
},
{
"epoch": 1.5638665872926487,
"grad_norm": 8.261491775512695,
"learning_rate": 1.0904366573350067e-05,
"loss": 0.10849467277526856,
"step": 48600
},
{
"epoch": 1.567084452882403,
"grad_norm": 1.8083908557891846,
"learning_rate": 1.0823921227917754e-05,
"loss": 0.09797250747680664,
"step": 48700
},
{
"epoch": 1.5703023184721574,
"grad_norm": 0.7597082257270813,
"learning_rate": 1.074347588248544e-05,
"loss": 0.08994776725769044,
"step": 48800
},
{
"epoch": 1.5735201840619117,
"grad_norm": 1.0194493532180786,
"learning_rate": 1.0663030537053127e-05,
"loss": 0.1160097885131836,
"step": 48900
},
{
"epoch": 1.576738049651666,
"grad_norm": 3.0627150535583496,
"learning_rate": 1.0582585191620813e-05,
"loss": 0.08039675712585449,
"step": 49000
},
{
"epoch": 1.5799559152414204,
"grad_norm": 11.51146411895752,
"learning_rate": 1.05021398461885e-05,
"loss": 0.10077597618103028,
"step": 49100
},
{
"epoch": 1.5831737808311748,
"grad_norm": 3.410426616668701,
"learning_rate": 1.0421694500756186e-05,
"loss": 0.12183525085449219,
"step": 49200
},
{
"epoch": 1.5863916464209291,
"grad_norm": 4.240739822387695,
"learning_rate": 1.0341249155323873e-05,
"loss": 0.10310389518737793,
"step": 49300
},
{
"epoch": 1.5896095120106835,
"grad_norm": 4.760958194732666,
"learning_rate": 1.026080380989156e-05,
"loss": 0.10938605308532715,
"step": 49400
},
{
"epoch": 1.5928273776004376,
"grad_norm": 1.9922808408737183,
"learning_rate": 1.0180358464459248e-05,
"loss": 0.08123787879943847,
"step": 49500
},
{
"epoch": 1.596045243190192,
"grad_norm": 1.350643277168274,
"learning_rate": 1.0099913119026933e-05,
"loss": 0.12260754585266113,
"step": 49600
},
{
"epoch": 1.5992631087799463,
"grad_norm": 1.4271605014801025,
"learning_rate": 1.001946777359462e-05,
"loss": 0.09629259109497071,
"step": 49700
},
{
"epoch": 1.6024809743697006,
"grad_norm": 0.5193639397621155,
"learning_rate": 9.939022428162308e-06,
"loss": 0.10984975814819337,
"step": 49800
},
{
"epoch": 1.605698839959455,
"grad_norm": 9.31329345703125,
"learning_rate": 9.858577082729994e-06,
"loss": 0.09676812171936035,
"step": 49900
},
{
"epoch": 1.608916705549209,
"grad_norm": 1.29440176486969,
"learning_rate": 9.77813173729768e-06,
"loss": 0.09296704292297363,
"step": 50000
},
{
"epoch": 1.608916705549209,
"eval_accuracy": 0.9898457261872506,
"eval_f1": 0.9511017704750988,
"eval_loss": 0.038785140961408615,
"eval_precision": 0.9418761588819233,
"eval_recall": 0.9605098982518877,
"eval_runtime": 39.1145,
"eval_samples_per_second": 639.149,
"eval_steps_per_second": 35.511,
"step": 50000
},
{
"epoch": 1.6121345711389634,
"grad_norm": 2.5150434970855713,
"learning_rate": 9.697686391865367e-06,
"loss": 0.0893746280670166,
"step": 50100
},
{
"epoch": 1.6153524367287178,
"grad_norm": 3.5938878059387207,
"learning_rate": 9.617241046433054e-06,
"loss": 0.10516523361206055,
"step": 50200
},
{
"epoch": 1.618570302318472,
"grad_norm": 3.323432445526123,
"learning_rate": 9.53679570100074e-06,
"loss": 0.10093666076660156,
"step": 50300
},
{
"epoch": 1.6217881679082264,
"grad_norm": 3.201542854309082,
"learning_rate": 9.456350355568427e-06,
"loss": 0.09994193077087403,
"step": 50400
},
{
"epoch": 1.6250060334979808,
"grad_norm": 1.2016533613204956,
"learning_rate": 9.375905010136114e-06,
"loss": 0.09140503883361817,
"step": 50500
},
{
"epoch": 1.6282238990877351,
"grad_norm": 2.176725149154663,
"learning_rate": 9.295459664703802e-06,
"loss": 0.10795197486877442,
"step": 50600
},
{
"epoch": 1.6314417646774895,
"grad_norm": 2.084689140319824,
"learning_rate": 9.215014319271487e-06,
"loss": 0.10725319862365723,
"step": 50700
},
{
"epoch": 1.6346596302672438,
"grad_norm": 3.9410250186920166,
"learning_rate": 9.134568973839173e-06,
"loss": 0.1040007209777832,
"step": 50800
},
{
"epoch": 1.6378774958569982,
"grad_norm": 7.110863208770752,
"learning_rate": 9.05412362840686e-06,
"loss": 0.09510645866394044,
"step": 50900
},
{
"epoch": 1.6410953614467525,
"grad_norm": 2.1342461109161377,
"learning_rate": 8.973678282974548e-06,
"loss": 0.10462862014770508,
"step": 51000
},
{
"epoch": 1.6443132270365068,
"grad_norm": 2.3921563625335693,
"learning_rate": 8.893232937542233e-06,
"loss": 0.09732514381408691,
"step": 51100
},
{
"epoch": 1.647531092626261,
"grad_norm": 2.4982810020446777,
"learning_rate": 8.81278759210992e-06,
"loss": 0.08822668075561524,
"step": 51200
},
{
"epoch": 1.6507489582160153,
"grad_norm": 1.7426912784576416,
"learning_rate": 8.732342246677608e-06,
"loss": 0.08898674011230469,
"step": 51300
},
{
"epoch": 1.6539668238057696,
"grad_norm": 9.542119026184082,
"learning_rate": 8.651896901245295e-06,
"loss": 0.12028385162353515,
"step": 51400
},
{
"epoch": 1.657184689395524,
"grad_norm": 0.7342370748519897,
"learning_rate": 8.571451555812981e-06,
"loss": 0.10786977767944336,
"step": 51500
},
{
"epoch": 1.6604025549852781,
"grad_norm": 0.8785775303840637,
"learning_rate": 8.491006210380668e-06,
"loss": 0.09443745613098145,
"step": 51600
},
{
"epoch": 1.6636204205750325,
"grad_norm": 9.54134464263916,
"learning_rate": 8.410560864948354e-06,
"loss": 0.11094076156616212,
"step": 51700
},
{
"epoch": 1.6668382861647868,
"grad_norm": 23.422744750976562,
"learning_rate": 8.330115519516041e-06,
"loss": 0.10699220657348633,
"step": 51800
},
{
"epoch": 1.6700561517545411,
"grad_norm": 1.0714454650878906,
"learning_rate": 8.249670174083727e-06,
"loss": 0.10455569267272949,
"step": 51900
},
{
"epoch": 1.6732740173442955,
"grad_norm": 2.6620521545410156,
"learning_rate": 8.169224828651414e-06,
"loss": 0.09597659111022949,
"step": 52000
},
{
"epoch": 1.6764918829340498,
"grad_norm": 1.480083703994751,
"learning_rate": 8.088779483219102e-06,
"loss": 0.10422684669494629,
"step": 52100
},
{
"epoch": 1.6797097485238042,
"grad_norm": 2.0191445350646973,
"learning_rate": 8.008334137786787e-06,
"loss": 0.10290263175964355,
"step": 52200
},
{
"epoch": 1.6829276141135585,
"grad_norm": 1.1020231246948242,
"learning_rate": 7.927888792354475e-06,
"loss": 0.10773059844970703,
"step": 52300
},
{
"epoch": 1.6861454797033129,
"grad_norm": 2.2319722175598145,
"learning_rate": 7.847443446922162e-06,
"loss": 0.10715092658996582,
"step": 52400
},
{
"epoch": 1.6893633452930672,
"grad_norm": 3.737014055252075,
"learning_rate": 7.76699810148985e-06,
"loss": 0.08955263137817383,
"step": 52500
},
{
"epoch": 1.6893633452930672,
"eval_accuracy": 0.9901202917944698,
"eval_f1": 0.9537408288021826,
"eval_loss": 0.03707578778266907,
"eval_precision": 0.945334613829357,
"eval_recall": 0.962297886572079,
"eval_runtime": 38.6863,
"eval_samples_per_second": 646.224,
"eval_steps_per_second": 35.904,
"step": 52500
},
{
"epoch": 1.6925812108828215,
"grad_norm": 1.981602668762207,
"learning_rate": 7.686552756057535e-06,
"loss": 0.09512082099914551,
"step": 52600
},
{
"epoch": 1.6957990764725759,
"grad_norm": 0.9314746856689453,
"learning_rate": 7.606107410625221e-06,
"loss": 0.09555187225341796,
"step": 52700
},
{
"epoch": 1.69901694206233,
"grad_norm": 1.3338748216629028,
"learning_rate": 7.525662065192908e-06,
"loss": 0.08804367065429687,
"step": 52800
},
{
"epoch": 1.7022348076520843,
"grad_norm": 2.233142375946045,
"learning_rate": 7.445216719760595e-06,
"loss": 0.099904146194458,
"step": 52900
},
{
"epoch": 1.7054526732418387,
"grad_norm": 2.9169979095458984,
"learning_rate": 7.364771374328282e-06,
"loss": 0.09450019836425781,
"step": 53000
},
{
"epoch": 1.708670538831593,
"grad_norm": 1.4995619058609009,
"learning_rate": 7.284326028895968e-06,
"loss": 0.10291241645812989,
"step": 53100
},
{
"epoch": 1.7118884044213472,
"grad_norm": 0.7775556445121765,
"learning_rate": 7.203880683463656e-06,
"loss": 0.1126316738128662,
"step": 53200
},
{
"epoch": 1.7151062700111015,
"grad_norm": 4.055330753326416,
"learning_rate": 7.123435338031342e-06,
"loss": 0.10327481269836426,
"step": 53300
},
{
"epoch": 1.7183241356008558,
"grad_norm": 1.895665168762207,
"learning_rate": 7.04298999259903e-06,
"loss": 0.09612360954284668,
"step": 53400
},
{
"epoch": 1.7215420011906102,
"grad_norm": 2.202033758163452,
"learning_rate": 6.962544647166714e-06,
"loss": 0.09727460861206055,
"step": 53500
},
{
"epoch": 1.7247598667803645,
"grad_norm": 1.6725600957870483,
"learning_rate": 6.882099301734402e-06,
"loss": 0.1051039981842041,
"step": 53600
},
{
"epoch": 1.7279777323701189,
"grad_norm": 1.6951491832733154,
"learning_rate": 6.801653956302088e-06,
"loss": 0.09583574295043945,
"step": 53700
},
{
"epoch": 1.7311955979598732,
"grad_norm": 2.0571811199188232,
"learning_rate": 6.721208610869776e-06,
"loss": 0.09057200431823731,
"step": 53800
},
{
"epoch": 1.7344134635496276,
"grad_norm": 1.9436057806015015,
"learning_rate": 6.640763265437462e-06,
"loss": 0.10000760078430176,
"step": 53900
},
{
"epoch": 1.737631329139382,
"grad_norm": 0.3372837007045746,
"learning_rate": 6.560317920005149e-06,
"loss": 0.10574793815612793,
"step": 54000
},
{
"epoch": 1.7408491947291362,
"grad_norm": 2.41835880279541,
"learning_rate": 6.479872574572836e-06,
"loss": 0.10087454795837403,
"step": 54100
},
{
"epoch": 1.7440670603188906,
"grad_norm": 0.6734520792961121,
"learning_rate": 6.399427229140522e-06,
"loss": 0.08962540626525879,
"step": 54200
},
{
"epoch": 1.747284925908645,
"grad_norm": 2.5488386154174805,
"learning_rate": 6.31898188370821e-06,
"loss": 0.1008315372467041,
"step": 54300
},
{
"epoch": 1.7505027914983993,
"grad_norm": 3.4320902824401855,
"learning_rate": 6.238536538275895e-06,
"loss": 0.08948537826538086,
"step": 54400
},
{
"epoch": 1.7537206570881534,
"grad_norm": 3.819451093673706,
"learning_rate": 6.158091192843583e-06,
"loss": 0.08715433120727539,
"step": 54500
},
{
"epoch": 1.7569385226779077,
"grad_norm": 1.9689542055130005,
"learning_rate": 6.077645847411268e-06,
"loss": 0.0976317596435547,
"step": 54600
},
{
"epoch": 1.760156388267662,
"grad_norm": 2.7664942741394043,
"learning_rate": 5.997200501978956e-06,
"loss": 0.11424756050109863,
"step": 54700
},
{
"epoch": 1.7633742538574164,
"grad_norm": 1.188758134841919,
"learning_rate": 5.916755156546642e-06,
"loss": 0.09281901359558105,
"step": 54800
},
{
"epoch": 1.7665921194471705,
"grad_norm": 1.442600965499878,
"learning_rate": 5.83630981111433e-06,
"loss": 0.09055022239685058,
"step": 54900
},
{
"epoch": 1.7698099850369249,
"grad_norm": 2.042584180831909,
"learning_rate": 5.755864465682016e-06,
"loss": 0.08615251541137696,
"step": 55000
},
{
"epoch": 1.7698099850369249,
"eval_accuracy": 0.9903457354945594,
"eval_f1": 0.9534046525260119,
"eval_loss": 0.036240287125110626,
"eval_precision": 0.9445875312476327,
"eval_recall": 0.9623879291493549,
"eval_runtime": 38.9533,
"eval_samples_per_second": 641.795,
"eval_steps_per_second": 35.658,
"step": 55000
},
{
"epoch": 1.7730278506266792,
"grad_norm": 1.0843491554260254,
"learning_rate": 5.675419120249703e-06,
"loss": 0.08899230003356934,
"step": 55100
},
{
"epoch": 1.7762457162164336,
"grad_norm": 12.144911766052246,
"learning_rate": 5.594973774817389e-06,
"loss": 0.08881766319274903,
"step": 55200
},
{
"epoch": 1.779463581806188,
"grad_norm": 2.8250832557678223,
"learning_rate": 5.514528429385076e-06,
"loss": 0.0975367259979248,
"step": 55300
},
{
"epoch": 1.7826814473959423,
"grad_norm": 1.2422140836715698,
"learning_rate": 5.434083083952763e-06,
"loss": 0.10660443305969239,
"step": 55400
},
{
"epoch": 1.7858993129856966,
"grad_norm": 4.103642463684082,
"learning_rate": 5.353637738520449e-06,
"loss": 0.10672403335571289,
"step": 55500
},
{
"epoch": 1.789117178575451,
"grad_norm": 2.0695173740386963,
"learning_rate": 5.273192393088136e-06,
"loss": 0.08749826431274414,
"step": 55600
},
{
"epoch": 1.7923350441652053,
"grad_norm": 2.9659390449523926,
"learning_rate": 5.192747047655823e-06,
"loss": 0.10261861801147461,
"step": 55700
},
{
"epoch": 1.7955529097549596,
"grad_norm": 3.0467967987060547,
"learning_rate": 5.11230170222351e-06,
"loss": 0.1026226806640625,
"step": 55800
},
{
"epoch": 1.798770775344714,
"grad_norm": 2.826698064804077,
"learning_rate": 5.031856356791196e-06,
"loss": 0.08784229278564454,
"step": 55900
},
{
"epoch": 1.8019886409344683,
"grad_norm": 4.862614631652832,
"learning_rate": 4.951411011358883e-06,
"loss": 0.09360240936279297,
"step": 56000
},
{
"epoch": 1.8052065065242224,
"grad_norm": 1.5207270383834839,
"learning_rate": 4.87096566592657e-06,
"loss": 0.10859518051147461,
"step": 56100
},
{
"epoch": 1.8084243721139768,
"grad_norm": 2.692488670349121,
"learning_rate": 4.790520320494257e-06,
"loss": 0.0937327766418457,
"step": 56200
},
{
"epoch": 1.8116422377037311,
"grad_norm": 2.3452610969543457,
"learning_rate": 4.710074975061943e-06,
"loss": 0.0986541748046875,
"step": 56300
},
{
"epoch": 1.8148601032934855,
"grad_norm": 3.4485702514648438,
"learning_rate": 4.6296296296296296e-06,
"loss": 0.11264933586120605,
"step": 56400
},
{
"epoch": 1.8180779688832396,
"grad_norm": 0.7109314203262329,
"learning_rate": 4.549184284197316e-06,
"loss": 0.09030750274658203,
"step": 56500
},
{
"epoch": 1.821295834472994,
"grad_norm": 11.383523941040039,
"learning_rate": 4.468738938765003e-06,
"loss": 0.09089506149291993,
"step": 56600
},
{
"epoch": 1.8245137000627483,
"grad_norm": 2.5693790912628174,
"learning_rate": 4.38829359333269e-06,
"loss": 0.09092802047729492,
"step": 56700
},
{
"epoch": 1.8277315656525026,
"grad_norm": 4.628587245941162,
"learning_rate": 4.307848247900376e-06,
"loss": 0.10240273475646973,
"step": 56800
},
{
"epoch": 1.830949431242257,
"grad_norm": 2.5633656978607178,
"learning_rate": 4.227402902468063e-06,
"loss": 0.10268410682678222,
"step": 56900
},
{
"epoch": 1.8341672968320113,
"grad_norm": 1.9654608964920044,
"learning_rate": 4.14695755703575e-06,
"loss": 0.08537924766540528,
"step": 57000
},
{
"epoch": 1.8373851624217656,
"grad_norm": 5.80186653137207,
"learning_rate": 4.066512211603437e-06,
"loss": 0.09202526092529296,
"step": 57100
},
{
"epoch": 1.84060302801152,
"grad_norm": 1.1093907356262207,
"learning_rate": 3.986066866171124e-06,
"loss": 0.08391679763793945,
"step": 57200
},
{
"epoch": 1.8438208936012743,
"grad_norm": 2.4156336784362793,
"learning_rate": 3.90562152073881e-06,
"loss": 0.08964170455932617,
"step": 57300
},
{
"epoch": 1.8470387591910287,
"grad_norm": 0.8483083844184875,
"learning_rate": 3.825176175306497e-06,
"loss": 0.08196611404418945,
"step": 57400
},
{
"epoch": 1.850256624780783,
"grad_norm": 2.47489070892334,
"learning_rate": 3.7447308298741836e-06,
"loss": 0.09712170600891114,
"step": 57500
},
{
"epoch": 1.850256624780783,
"eval_accuracy": 0.990784730222486,
"eval_f1": 0.9556170744775851,
"eval_loss": 0.03545914590358734,
"eval_precision": 0.9478179448570814,
"eval_recall": 0.9635456194286155,
"eval_runtime": 39.1151,
"eval_samples_per_second": 639.14,
"eval_steps_per_second": 35.511,
"step": 57500
},
{
"epoch": 1.8534744903705374,
"grad_norm": 2.640423536300659,
"learning_rate": 3.6642854844418705e-06,
"loss": 0.09879606246948242,
"step": 57600
},
{
"epoch": 1.8566923559602917,
"grad_norm": 1.9036402702331543,
"learning_rate": 3.5838401390095566e-06,
"loss": 0.08787569999694825,
"step": 57700
},
{
"epoch": 1.8599102215500458,
"grad_norm": 2.5272982120513916,
"learning_rate": 3.5033947935772435e-06,
"loss": 0.0921150016784668,
"step": 57800
},
{
"epoch": 1.8631280871398002,
"grad_norm": 2.6478683948516846,
"learning_rate": 3.4229494481449304e-06,
"loss": 0.10192151069641113,
"step": 57900
},
{
"epoch": 1.8663459527295545,
"grad_norm": 1.6722396612167358,
"learning_rate": 3.3425041027126173e-06,
"loss": 0.08221748352050781,
"step": 58000
},
{
"epoch": 1.8695638183193088,
"grad_norm": 0.9039199948310852,
"learning_rate": 3.2620587572803042e-06,
"loss": 0.09699295043945312,
"step": 58100
},
{
"epoch": 1.872781683909063,
"grad_norm": 4.108741283416748,
"learning_rate": 3.1816134118479903e-06,
"loss": 0.10586893081665039,
"step": 58200
},
{
"epoch": 1.8759995494988173,
"grad_norm": 0.8022117018699646,
"learning_rate": 3.1011680664156772e-06,
"loss": 0.10414183616638184,
"step": 58300
},
{
"epoch": 1.8792174150885717,
"grad_norm": 0.757673978805542,
"learning_rate": 3.020722720983364e-06,
"loss": 0.09568047523498535,
"step": 58400
},
{
"epoch": 1.882435280678326,
"grad_norm": 2.535696506500244,
"learning_rate": 2.9402773755510506e-06,
"loss": 0.10529354095458984,
"step": 58500
},
{
"epoch": 1.8856531462680803,
"grad_norm": 3.3938286304473877,
"learning_rate": 2.8598320301187376e-06,
"loss": 0.09156073570251465,
"step": 58600
},
{
"epoch": 1.8888710118578347,
"grad_norm": 1.427871823310852,
"learning_rate": 2.779386684686424e-06,
"loss": 0.08934795379638671,
"step": 58700
},
{
"epoch": 1.892088877447589,
"grad_norm": 1.7206834554672241,
"learning_rate": 2.698941339254111e-06,
"loss": 0.10297789573669433,
"step": 58800
},
{
"epoch": 1.8953067430373434,
"grad_norm": 2.4081387519836426,
"learning_rate": 2.618495993821798e-06,
"loss": 0.10169391632080078,
"step": 58900
},
{
"epoch": 1.8985246086270977,
"grad_norm": 4.349306583404541,
"learning_rate": 2.5380506483894844e-06,
"loss": 0.09130579948425294,
"step": 59000
},
{
"epoch": 1.901742474216852,
"grad_norm": 1.038880467414856,
"learning_rate": 2.4576053029571713e-06,
"loss": 0.09172413825988769,
"step": 59100
},
{
"epoch": 1.9049603398066064,
"grad_norm": 1.437605619430542,
"learning_rate": 2.377159957524858e-06,
"loss": 0.10029353141784668,
"step": 59200
},
{
"epoch": 1.9081782053963607,
"grad_norm": 1.8541232347488403,
"learning_rate": 2.2967146120925443e-06,
"loss": 0.09666367530822755,
"step": 59300
},
{
"epoch": 1.9113960709861149,
"grad_norm": 4.376369953155518,
"learning_rate": 2.2162692666602312e-06,
"loss": 0.08466120719909669,
"step": 59400
},
{
"epoch": 1.9146139365758692,
"grad_norm": 1.9385297298431396,
"learning_rate": 2.1358239212279177e-06,
"loss": 0.08599072456359863,
"step": 59500
},
{
"epoch": 1.9178318021656235,
"grad_norm": 1.689815640449524,
"learning_rate": 2.0553785757956042e-06,
"loss": 0.09492490768432617,
"step": 59600
},
{
"epoch": 1.9210496677553779,
"grad_norm": 1.3797245025634766,
"learning_rate": 1.974933230363291e-06,
"loss": 0.09027796745300293,
"step": 59700
},
{
"epoch": 1.924267533345132,
"grad_norm": 0.4921301007270813,
"learning_rate": 1.894487884930978e-06,
"loss": 0.0897593879699707,
"step": 59800
},
{
"epoch": 1.9274853989348864,
"grad_norm": 1.489128828048706,
"learning_rate": 1.8140425394986648e-06,
"loss": 0.09694389343261718,
"step": 59900
},
{
"epoch": 1.9307032645246407,
"grad_norm": 0.7353283166885376,
"learning_rate": 1.7335971940663515e-06,
"loss": 0.0947257137298584,
"step": 60000
},
{
"epoch": 1.9307032645246407,
"eval_accuracy": 0.9908240277481897,
"eval_f1": 0.9576732752290649,
"eval_loss": 0.03446565568447113,
"eval_precision": 0.9507949387630905,
"eval_recall": 0.9646518568065757,
"eval_runtime": 38.8968,
"eval_samples_per_second": 642.726,
"eval_steps_per_second": 35.71,
"step": 60000
},
{
"epoch": 1.933921130114395,
"grad_norm": 1.3698302507400513,
"learning_rate": 1.653151848634038e-06,
"loss": 0.0832412052154541,
"step": 60100
},
{
"epoch": 1.9371389957041494,
"grad_norm": 4.799550533294678,
"learning_rate": 1.5727065032017249e-06,
"loss": 0.11331718444824218,
"step": 60200
},
{
"epoch": 1.9403568612939037,
"grad_norm": 1.4409183263778687,
"learning_rate": 1.4922611577694116e-06,
"loss": 0.09095165252685547,
"step": 60300
},
{
"epoch": 1.943574726883658,
"grad_norm": 0.7844113111495972,
"learning_rate": 1.4118158123370983e-06,
"loss": 0.07738580703735351,
"step": 60400
},
{
"epoch": 1.9467925924734124,
"grad_norm": 1.4404165744781494,
"learning_rate": 1.331370466904785e-06,
"loss": 0.08979734420776367,
"step": 60500
},
{
"epoch": 1.9500104580631668,
"grad_norm": 12.599478721618652,
"learning_rate": 1.2509251214724717e-06,
"loss": 0.10170120239257813,
"step": 60600
},
{
"epoch": 1.953228323652921,
"grad_norm": 16.47142791748047,
"learning_rate": 1.1704797760401584e-06,
"loss": 0.08839907646179199,
"step": 60700
},
{
"epoch": 1.9564461892426754,
"grad_norm": 1.4317628145217896,
"learning_rate": 1.0900344306078451e-06,
"loss": 0.10484655380249024,
"step": 60800
},
{
"epoch": 1.9596640548324298,
"grad_norm": 1.4019297361373901,
"learning_rate": 1.0095890851755318e-06,
"loss": 0.0975730323791504,
"step": 60900
},
{
"epoch": 1.9628819204221841,
"grad_norm": 1.428163766860962,
"learning_rate": 9.291437397432184e-07,
"loss": 0.09934920310974121,
"step": 61000
},
{
"epoch": 1.9660997860119382,
"grad_norm": 0.9141352772712708,
"learning_rate": 8.486983943109052e-07,
"loss": 0.09621753692626953,
"step": 61100
},
{
"epoch": 1.9693176516016926,
"grad_norm": 4.859970569610596,
"learning_rate": 7.682530488785919e-07,
"loss": 0.0880228328704834,
"step": 61200
},
{
"epoch": 1.972535517191447,
"grad_norm": 14.404598236083984,
"learning_rate": 6.878077034462786e-07,
"loss": 0.10370702743530273,
"step": 61300
},
{
"epoch": 1.9757533827812013,
"grad_norm": 1.890513300895691,
"learning_rate": 6.073623580139654e-07,
"loss": 0.08836669921875,
"step": 61400
},
{
"epoch": 1.9789712483709554,
"grad_norm": 3.365849494934082,
"learning_rate": 5.269170125816521e-07,
"loss": 0.0887114429473877,
"step": 61500
},
{
"epoch": 1.9821891139607097,
"grad_norm": 2.7111990451812744,
"learning_rate": 4.4647166714933875e-07,
"loss": 0.10604698181152344,
"step": 61600
},
{
"epoch": 1.985406979550464,
"grad_norm": 1.373115062713623,
"learning_rate": 3.6602632171702546e-07,
"loss": 0.08321118354797363,
"step": 61700
},
{
"epoch": 1.9886248451402184,
"grad_norm": 1.2105165719985962,
"learning_rate": 2.855809762847122e-07,
"loss": 0.09142358779907227,
"step": 61800
},
{
"epoch": 1.9918427107299728,
"grad_norm": 1.9538366794586182,
"learning_rate": 2.0513563085239887e-07,
"loss": 0.07895866394042969,
"step": 61900
},
{
"epoch": 1.995060576319727,
"grad_norm": 0.8726534843444824,
"learning_rate": 1.246902854200856e-07,
"loss": 0.10273164749145508,
"step": 62000
},
{
"epoch": 1.9982784419094815,
"grad_norm": 3.1368095874786377,
"learning_rate": 4.424493998777231e-08,
"loss": 0.08670869827270508,
"step": 62100
},
{
"epoch": 2.0,
"step": 62154,
"total_flos": 1.5480486185003853e+18,
"train_loss": 0.1780550116395906,
"train_runtime": 17614.9903,
"train_samples_per_second": 508.09,
"train_steps_per_second": 3.528
}
],
"logging_steps": 100,
"max_steps": 62154,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 2500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.5480486185003853e+18,
"train_batch_size": 18,
"trial_name": null,
"trial_params": null
}