d1_code_all_large / trainer_state.json
neginr's picture
End of training
b2c8055 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.998875140607424,
"eval_steps": 500,
"global_step": 555,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008998875140607425,
"grad_norm": 6.405796769298009,
"learning_rate": 1.4285714285714286e-06,
"loss": 1.0356,
"step": 1
},
{
"epoch": 0.01799775028121485,
"grad_norm": 6.457071617226858,
"learning_rate": 2.8571428571428573e-06,
"loss": 1.0356,
"step": 2
},
{
"epoch": 0.02699662542182227,
"grad_norm": 6.306094210313564,
"learning_rate": 4.2857142857142855e-06,
"loss": 1.0346,
"step": 3
},
{
"epoch": 0.0359955005624297,
"grad_norm": 4.578449546704467,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.9945,
"step": 4
},
{
"epoch": 0.04499437570303712,
"grad_norm": 2.7068662098456318,
"learning_rate": 7.1428571428571436e-06,
"loss": 0.9512,
"step": 5
},
{
"epoch": 0.05399325084364454,
"grad_norm": 2.4781853702406553,
"learning_rate": 8.571428571428571e-06,
"loss": 0.9632,
"step": 6
},
{
"epoch": 0.06299212598425197,
"grad_norm": 4.778781193622344,
"learning_rate": 1e-05,
"loss": 0.9478,
"step": 7
},
{
"epoch": 0.0719910011248594,
"grad_norm": 5.10213288638439,
"learning_rate": 1.1428571428571429e-05,
"loss": 0.9316,
"step": 8
},
{
"epoch": 0.08098987626546682,
"grad_norm": 5.371042207893317,
"learning_rate": 1.2857142857142859e-05,
"loss": 0.9182,
"step": 9
},
{
"epoch": 0.08998875140607424,
"grad_norm": 4.422548992141668,
"learning_rate": 1.4285714285714287e-05,
"loss": 0.9052,
"step": 10
},
{
"epoch": 0.09898762654668167,
"grad_norm": 2.908057739308309,
"learning_rate": 1.5714285714285715e-05,
"loss": 0.8738,
"step": 11
},
{
"epoch": 0.10798650168728909,
"grad_norm": 2.081438174306188,
"learning_rate": 1.7142857142857142e-05,
"loss": 0.8448,
"step": 12
},
{
"epoch": 0.11698537682789652,
"grad_norm": 1.8625844126627613,
"learning_rate": 1.8571428571428575e-05,
"loss": 0.8214,
"step": 13
},
{
"epoch": 0.12598425196850394,
"grad_norm": 1.4150826691986598,
"learning_rate": 2e-05,
"loss": 0.8118,
"step": 14
},
{
"epoch": 0.13498312710911137,
"grad_norm": 1.3470926001768453,
"learning_rate": 2.1428571428571428e-05,
"loss": 0.7967,
"step": 15
},
{
"epoch": 0.1439820022497188,
"grad_norm": 1.318138672473746,
"learning_rate": 2.2857142857142858e-05,
"loss": 0.7817,
"step": 16
},
{
"epoch": 0.1529808773903262,
"grad_norm": 1.132045894481426,
"learning_rate": 2.4285714285714285e-05,
"loss": 0.7702,
"step": 17
},
{
"epoch": 0.16197975253093364,
"grad_norm": 1.1216235493173696,
"learning_rate": 2.5714285714285718e-05,
"loss": 0.7625,
"step": 18
},
{
"epoch": 0.17097862767154107,
"grad_norm": 1.2949162303101338,
"learning_rate": 2.7142857142857148e-05,
"loss": 0.7552,
"step": 19
},
{
"epoch": 0.17997750281214847,
"grad_norm": 1.0018485261634034,
"learning_rate": 2.8571428571428574e-05,
"loss": 0.748,
"step": 20
},
{
"epoch": 0.1889763779527559,
"grad_norm": 0.9642142358632533,
"learning_rate": 3.0000000000000004e-05,
"loss": 0.7319,
"step": 21
},
{
"epoch": 0.19797525309336333,
"grad_norm": 1.3162346092752937,
"learning_rate": 3.142857142857143e-05,
"loss": 0.7376,
"step": 22
},
{
"epoch": 0.20697412823397077,
"grad_norm": 1.2462908815710223,
"learning_rate": 3.285714285714286e-05,
"loss": 0.7225,
"step": 23
},
{
"epoch": 0.21597300337457817,
"grad_norm": 1.1288729992725366,
"learning_rate": 3.4285714285714284e-05,
"loss": 0.7232,
"step": 24
},
{
"epoch": 0.2249718785151856,
"grad_norm": 1.2886081610849456,
"learning_rate": 3.571428571428572e-05,
"loss": 0.7235,
"step": 25
},
{
"epoch": 0.23397075365579303,
"grad_norm": 1.377579395059948,
"learning_rate": 3.714285714285715e-05,
"loss": 0.7136,
"step": 26
},
{
"epoch": 0.24296962879640044,
"grad_norm": 1.0739406693404967,
"learning_rate": 3.857142857142858e-05,
"loss": 0.7138,
"step": 27
},
{
"epoch": 0.25196850393700787,
"grad_norm": 1.2160719105328928,
"learning_rate": 4e-05,
"loss": 0.7106,
"step": 28
},
{
"epoch": 0.2609673790776153,
"grad_norm": 1.3411926888307302,
"learning_rate": 4.1428571428571437e-05,
"loss": 0.7056,
"step": 29
},
{
"epoch": 0.26996625421822273,
"grad_norm": 0.8344443792471369,
"learning_rate": 4.2857142857142856e-05,
"loss": 0.6906,
"step": 30
},
{
"epoch": 0.27896512935883017,
"grad_norm": 1.6378121927259697,
"learning_rate": 4.428571428571429e-05,
"loss": 0.7022,
"step": 31
},
{
"epoch": 0.2879640044994376,
"grad_norm": 1.312672534168789,
"learning_rate": 4.5714285714285716e-05,
"loss": 0.6925,
"step": 32
},
{
"epoch": 0.296962879640045,
"grad_norm": 1.1997985746020943,
"learning_rate": 4.714285714285715e-05,
"loss": 0.6894,
"step": 33
},
{
"epoch": 0.3059617547806524,
"grad_norm": 1.3374680408068491,
"learning_rate": 4.857142857142857e-05,
"loss": 0.6924,
"step": 34
},
{
"epoch": 0.31496062992125984,
"grad_norm": 0.9866632538970782,
"learning_rate": 5e-05,
"loss": 0.6856,
"step": 35
},
{
"epoch": 0.32395950506186727,
"grad_norm": 1.7617767807118945,
"learning_rate": 5.1428571428571436e-05,
"loss": 0.6882,
"step": 36
},
{
"epoch": 0.3329583802024747,
"grad_norm": 1.2162422657164105,
"learning_rate": 5.285714285714286e-05,
"loss": 0.6827,
"step": 37
},
{
"epoch": 0.34195725534308213,
"grad_norm": 1.4381817884560075,
"learning_rate": 5.4285714285714295e-05,
"loss": 0.6804,
"step": 38
},
{
"epoch": 0.35095613048368957,
"grad_norm": 1.1994316767887803,
"learning_rate": 5.5714285714285715e-05,
"loss": 0.675,
"step": 39
},
{
"epoch": 0.35995500562429694,
"grad_norm": 1.3859110222701834,
"learning_rate": 5.714285714285715e-05,
"loss": 0.6714,
"step": 40
},
{
"epoch": 0.3689538807649044,
"grad_norm": 1.0456569888416478,
"learning_rate": 5.8571428571428575e-05,
"loss": 0.666,
"step": 41
},
{
"epoch": 0.3779527559055118,
"grad_norm": 1.550645312960814,
"learning_rate": 6.000000000000001e-05,
"loss": 0.6716,
"step": 42
},
{
"epoch": 0.38695163104611924,
"grad_norm": 1.7520944490642492,
"learning_rate": 6.142857142857143e-05,
"loss": 0.6668,
"step": 43
},
{
"epoch": 0.39595050618672667,
"grad_norm": 1.220392434494151,
"learning_rate": 6.285714285714286e-05,
"loss": 0.6706,
"step": 44
},
{
"epoch": 0.4049493813273341,
"grad_norm": 1.5640678258841063,
"learning_rate": 6.428571428571429e-05,
"loss": 0.6666,
"step": 45
},
{
"epoch": 0.41394825646794153,
"grad_norm": 1.0339470330368512,
"learning_rate": 6.571428571428571e-05,
"loss": 0.6611,
"step": 46
},
{
"epoch": 0.4229471316085489,
"grad_norm": 1.5630693263603568,
"learning_rate": 6.714285714285715e-05,
"loss": 0.6654,
"step": 47
},
{
"epoch": 0.43194600674915634,
"grad_norm": 1.4317212538586248,
"learning_rate": 6.857142857142857e-05,
"loss": 0.666,
"step": 48
},
{
"epoch": 0.4409448818897638,
"grad_norm": 1.2652376704046324,
"learning_rate": 7.000000000000001e-05,
"loss": 0.6525,
"step": 49
},
{
"epoch": 0.4499437570303712,
"grad_norm": 1.9389193690579092,
"learning_rate": 7.142857142857143e-05,
"loss": 0.6706,
"step": 50
},
{
"epoch": 0.45894263217097864,
"grad_norm": 1.2095199143374322,
"learning_rate": 7.285714285714286e-05,
"loss": 0.6578,
"step": 51
},
{
"epoch": 0.46794150731158607,
"grad_norm": 2.1530205341425943,
"learning_rate": 7.42857142857143e-05,
"loss": 0.6659,
"step": 52
},
{
"epoch": 0.4769403824521935,
"grad_norm": 1.3498447393418538,
"learning_rate": 7.571428571428571e-05,
"loss": 0.6555,
"step": 53
},
{
"epoch": 0.4859392575928009,
"grad_norm": 1.7344034955388106,
"learning_rate": 7.714285714285715e-05,
"loss": 0.6655,
"step": 54
},
{
"epoch": 0.4949381327334083,
"grad_norm": 2.165550267732371,
"learning_rate": 7.857142857142858e-05,
"loss": 0.6593,
"step": 55
},
{
"epoch": 0.5039370078740157,
"grad_norm": 1.6288021163787398,
"learning_rate": 8e-05,
"loss": 0.6541,
"step": 56
},
{
"epoch": 0.5129358830146231,
"grad_norm": 1.6669527952524703,
"learning_rate": 7.999920726649282e-05,
"loss": 0.6421,
"step": 57
},
{
"epoch": 0.5219347581552306,
"grad_norm": 1.2703954706550948,
"learning_rate": 7.999682909739257e-05,
"loss": 0.6416,
"step": 58
},
{
"epoch": 0.530933633295838,
"grad_norm": 1.6224554752837215,
"learning_rate": 7.999286558696199e-05,
"loss": 0.6427,
"step": 59
},
{
"epoch": 0.5399325084364455,
"grad_norm": 1.3278739806750215,
"learning_rate": 7.998731689230145e-05,
"loss": 0.6535,
"step": 60
},
{
"epoch": 0.5489313835770528,
"grad_norm": 1.9995409784489284,
"learning_rate": 7.998018323334275e-05,
"loss": 0.6496,
"step": 61
},
{
"epoch": 0.5579302587176603,
"grad_norm": 1.5356642072108388,
"learning_rate": 7.997146489284042e-05,
"loss": 0.6437,
"step": 62
},
{
"epoch": 0.5669291338582677,
"grad_norm": 1.256225547258712,
"learning_rate": 7.996116221636049e-05,
"loss": 0.6478,
"step": 63
},
{
"epoch": 0.5759280089988752,
"grad_norm": 1.7177343714374778,
"learning_rate": 7.994927561226682e-05,
"loss": 0.641,
"step": 64
},
{
"epoch": 0.5849268841394826,
"grad_norm": 1.8288469591109016,
"learning_rate": 7.993580555170486e-05,
"loss": 0.6483,
"step": 65
},
{
"epoch": 0.59392575928009,
"grad_norm": 0.9174977020863746,
"learning_rate": 7.992075256858302e-05,
"loss": 0.639,
"step": 66
},
{
"epoch": 0.6029246344206974,
"grad_norm": 2.3068233658832846,
"learning_rate": 7.990411725955153e-05,
"loss": 0.6469,
"step": 67
},
{
"epoch": 0.6119235095613048,
"grad_norm": 1.331580694798539,
"learning_rate": 7.98859002839787e-05,
"loss": 0.6403,
"step": 68
},
{
"epoch": 0.6209223847019123,
"grad_norm": 1.9556860563217953,
"learning_rate": 7.986610236392491e-05,
"loss": 0.6436,
"step": 69
},
{
"epoch": 0.6299212598425197,
"grad_norm": 1.4030777265618954,
"learning_rate": 7.984472428411388e-05,
"loss": 0.6541,
"step": 70
},
{
"epoch": 0.6389201349831272,
"grad_norm": 1.1120271583327745,
"learning_rate": 7.98217668919016e-05,
"loss": 0.6408,
"step": 71
},
{
"epoch": 0.6479190101237345,
"grad_norm": 1.7515091640927072,
"learning_rate": 7.97972310972428e-05,
"loss": 0.6462,
"step": 72
},
{
"epoch": 0.6569178852643419,
"grad_norm": 1.2154666389278914,
"learning_rate": 7.977111787265479e-05,
"loss": 0.6348,
"step": 73
},
{
"epoch": 0.6659167604049494,
"grad_norm": 1.1948666759848607,
"learning_rate": 7.9743428253179e-05,
"loss": 0.6398,
"step": 74
},
{
"epoch": 0.6749156355455568,
"grad_norm": 1.4465768500981744,
"learning_rate": 7.971416333633984e-05,
"loss": 0.633,
"step": 75
},
{
"epoch": 0.6839145106861643,
"grad_norm": 0.7079358166882018,
"learning_rate": 7.968332428210136e-05,
"loss": 0.6344,
"step": 76
},
{
"epoch": 0.6929133858267716,
"grad_norm": 1.2354926269354385,
"learning_rate": 7.965091231282114e-05,
"loss": 0.6316,
"step": 77
},
{
"epoch": 0.7019122609673791,
"grad_norm": 0.8980787882545499,
"learning_rate": 7.961692871320186e-05,
"loss": 0.6384,
"step": 78
},
{
"epoch": 0.7109111361079865,
"grad_norm": 1.327053450288205,
"learning_rate": 7.958137483024044e-05,
"loss": 0.6348,
"step": 79
},
{
"epoch": 0.7199100112485939,
"grad_norm": 1.0872221047887007,
"learning_rate": 7.95442520731746e-05,
"loss": 0.6319,
"step": 80
},
{
"epoch": 0.7289088863892014,
"grad_norm": 1.0406772088732106,
"learning_rate": 7.9505561913427e-05,
"loss": 0.6255,
"step": 81
},
{
"epoch": 0.7379077615298087,
"grad_norm": 1.3043796341854046,
"learning_rate": 7.946530588454695e-05,
"loss": 0.6324,
"step": 82
},
{
"epoch": 0.7469066366704162,
"grad_norm": 1.0289090852613956,
"learning_rate": 7.942348558214958e-05,
"loss": 0.6326,
"step": 83
},
{
"epoch": 0.7559055118110236,
"grad_norm": 1.309994146626142,
"learning_rate": 7.938010266385268e-05,
"loss": 0.6362,
"step": 84
},
{
"epoch": 0.7649043869516311,
"grad_norm": 0.9949679280573976,
"learning_rate": 7.933515884921086e-05,
"loss": 0.6274,
"step": 85
},
{
"epoch": 0.7739032620922385,
"grad_norm": 1.3529279590081953,
"learning_rate": 7.928865591964751e-05,
"loss": 0.6356,
"step": 86
},
{
"epoch": 0.7829021372328459,
"grad_norm": 1.5394293742781517,
"learning_rate": 7.924059571838419e-05,
"loss": 0.6239,
"step": 87
},
{
"epoch": 0.7919010123734533,
"grad_norm": 0.8230315586506134,
"learning_rate": 7.919098015036746e-05,
"loss": 0.6254,
"step": 88
},
{
"epoch": 0.8008998875140607,
"grad_norm": 2.5412889042558646,
"learning_rate": 7.91398111821935e-05,
"loss": 0.6386,
"step": 89
},
{
"epoch": 0.8098987626546682,
"grad_norm": 1.6534609255338006,
"learning_rate": 7.908709084203006e-05,
"loss": 0.6433,
"step": 90
},
{
"epoch": 0.8188976377952756,
"grad_norm": 2.115550172588045,
"learning_rate": 7.903282121953619e-05,
"loss": 0.6274,
"step": 91
},
{
"epoch": 0.8278965129358831,
"grad_norm": 2.21745507570489,
"learning_rate": 7.897700446577928e-05,
"loss": 0.6395,
"step": 92
},
{
"epoch": 0.8368953880764904,
"grad_norm": 1.1348223149712242,
"learning_rate": 7.891964279314988e-05,
"loss": 0.6304,
"step": 93
},
{
"epoch": 0.8458942632170978,
"grad_norm": 1.1093790619140609,
"learning_rate": 7.886073847527397e-05,
"loss": 0.634,
"step": 94
},
{
"epoch": 0.8548931383577053,
"grad_norm": 1.0247299216217371,
"learning_rate": 7.88002938469229e-05,
"loss": 0.6187,
"step": 95
},
{
"epoch": 0.8638920134983127,
"grad_norm": 0.9690359150703725,
"learning_rate": 7.873831130392077e-05,
"loss": 0.6224,
"step": 96
},
{
"epoch": 0.8728908886389202,
"grad_norm": 0.8479503109980445,
"learning_rate": 7.867479330304951e-05,
"loss": 0.6129,
"step": 97
},
{
"epoch": 0.8818897637795275,
"grad_norm": 0.7677468105242354,
"learning_rate": 7.860974236195151e-05,
"loss": 0.6133,
"step": 98
},
{
"epoch": 0.890888638920135,
"grad_norm": 1.0066107483900497,
"learning_rate": 7.85431610590298e-05,
"loss": 0.6212,
"step": 99
},
{
"epoch": 0.8998875140607424,
"grad_norm": 0.8385258369213318,
"learning_rate": 7.847505203334588e-05,
"loss": 0.6162,
"step": 100
},
{
"epoch": 0.9088863892013498,
"grad_norm": 0.8097143309117704,
"learning_rate": 7.840541798451506e-05,
"loss": 0.6105,
"step": 101
},
{
"epoch": 0.9178852643419573,
"grad_norm": 0.6217223076576118,
"learning_rate": 7.833426167259955e-05,
"loss": 0.6091,
"step": 102
},
{
"epoch": 0.9268841394825647,
"grad_norm": 0.8171222793279067,
"learning_rate": 7.826158591799898e-05,
"loss": 0.6124,
"step": 103
},
{
"epoch": 0.9358830146231721,
"grad_norm": 0.825619425206217,
"learning_rate": 7.818739360133863e-05,
"loss": 0.6093,
"step": 104
},
{
"epoch": 0.9448818897637795,
"grad_norm": 0.9916352114638577,
"learning_rate": 7.811168766335531e-05,
"loss": 0.6152,
"step": 105
},
{
"epoch": 0.953880764904387,
"grad_norm": 1.0337582772307539,
"learning_rate": 7.803447110478067e-05,
"loss": 0.6183,
"step": 106
},
{
"epoch": 0.9628796400449944,
"grad_norm": 0.9165715880682846,
"learning_rate": 7.795574698622237e-05,
"loss": 0.6176,
"step": 107
},
{
"epoch": 0.9718785151856018,
"grad_norm": 0.6530649555546069,
"learning_rate": 7.787551842804276e-05,
"loss": 0.6099,
"step": 108
},
{
"epoch": 0.9808773903262092,
"grad_norm": 0.7668074713592107,
"learning_rate": 7.779378861023516e-05,
"loss": 0.6063,
"step": 109
},
{
"epoch": 0.9898762654668166,
"grad_norm": 0.715515594924498,
"learning_rate": 7.77105607722978e-05,
"loss": 0.6088,
"step": 110
},
{
"epoch": 0.9988751406074241,
"grad_norm": 0.7194843087763119,
"learning_rate": 7.762583821310548e-05,
"loss": 0.6027,
"step": 111
},
{
"epoch": 1.0078740157480315,
"grad_norm": 1.3373684763348388,
"learning_rate": 7.753962429077881e-05,
"loss": 1.109,
"step": 112
},
{
"epoch": 1.0168728908886389,
"grad_norm": 1.2665705644041705,
"learning_rate": 7.7451922422551e-05,
"loss": 0.5896,
"step": 113
},
{
"epoch": 1.0258717660292462,
"grad_norm": 0.7146916740896777,
"learning_rate": 7.736273608463253e-05,
"loss": 0.5776,
"step": 114
},
{
"epoch": 1.0348706411698538,
"grad_norm": 0.8865840936771311,
"learning_rate": 7.727206881207334e-05,
"loss": 0.5907,
"step": 115
},
{
"epoch": 1.0438695163104612,
"grad_norm": 0.7277897994412549,
"learning_rate": 7.717992419862268e-05,
"loss": 0.5873,
"step": 116
},
{
"epoch": 1.0528683914510686,
"grad_norm": 0.6536923705270321,
"learning_rate": 7.708630589658667e-05,
"loss": 0.5835,
"step": 117
},
{
"epoch": 1.061867266591676,
"grad_norm": 0.7162373408409907,
"learning_rate": 7.699121761668355e-05,
"loss": 0.5794,
"step": 118
},
{
"epoch": 1.0708661417322836,
"grad_norm": 1.0254372509038927,
"learning_rate": 7.689466312789661e-05,
"loss": 0.582,
"step": 119
},
{
"epoch": 1.079865016872891,
"grad_norm": 1.1137703756989226,
"learning_rate": 7.679664625732478e-05,
"loss": 0.5907,
"step": 120
},
{
"epoch": 1.0888638920134983,
"grad_norm": 0.5744577610457379,
"learning_rate": 7.669717089003094e-05,
"loss": 0.5742,
"step": 121
},
{
"epoch": 1.0978627671541057,
"grad_norm": 0.6451796147865981,
"learning_rate": 7.659624096888792e-05,
"loss": 0.5831,
"step": 122
},
{
"epoch": 1.106861642294713,
"grad_norm": 0.6276129538814031,
"learning_rate": 7.649386049442223e-05,
"loss": 0.5827,
"step": 123
},
{
"epoch": 1.1158605174353207,
"grad_norm": 0.7718645523573185,
"learning_rate": 7.639003352465551e-05,
"loss": 0.5745,
"step": 124
},
{
"epoch": 1.124859392575928,
"grad_norm": 0.9206192827571739,
"learning_rate": 7.628476417494368e-05,
"loss": 0.5839,
"step": 125
},
{
"epoch": 1.1338582677165354,
"grad_norm": 0.698663737317266,
"learning_rate": 7.617805661781374e-05,
"loss": 0.5754,
"step": 126
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.5240414441327959,
"learning_rate": 7.60699150827985e-05,
"loss": 0.5772,
"step": 127
},
{
"epoch": 1.1518560179977504,
"grad_norm": 0.5843162509749283,
"learning_rate": 7.596034385626888e-05,
"loss": 0.5727,
"step": 128
},
{
"epoch": 1.1608548931383578,
"grad_norm": 0.7391156549586344,
"learning_rate": 7.584934728126403e-05,
"loss": 0.5736,
"step": 129
},
{
"epoch": 1.1698537682789651,
"grad_norm": 0.6964429639639632,
"learning_rate": 7.573692975731914e-05,
"loss": 0.5677,
"step": 130
},
{
"epoch": 1.1788526434195725,
"grad_norm": 0.5118585657523064,
"learning_rate": 7.562309574029112e-05,
"loss": 0.5796,
"step": 131
},
{
"epoch": 1.18785151856018,
"grad_norm": 0.3947628953294446,
"learning_rate": 7.550784974218195e-05,
"loss": 0.5669,
"step": 132
},
{
"epoch": 1.1968503937007875,
"grad_norm": 0.37395524397542534,
"learning_rate": 7.539119633095983e-05,
"loss": 0.5641,
"step": 133
},
{
"epoch": 1.2058492688413949,
"grad_norm": 0.37361515124611633,
"learning_rate": 7.527314013037815e-05,
"loss": 0.5763,
"step": 134
},
{
"epoch": 1.2148481439820022,
"grad_norm": 0.4839280159646048,
"learning_rate": 7.515368581979224e-05,
"loss": 0.5703,
"step": 135
},
{
"epoch": 1.2238470191226096,
"grad_norm": 0.6174733585746707,
"learning_rate": 7.503283813397379e-05,
"loss": 0.5703,
"step": 136
},
{
"epoch": 1.232845894263217,
"grad_norm": 0.7319038998136339,
"learning_rate": 7.491060186292331e-05,
"loss": 0.5715,
"step": 137
},
{
"epoch": 1.2418447694038246,
"grad_norm": 0.7864687770694577,
"learning_rate": 7.478698185168019e-05,
"loss": 0.5714,
"step": 138
},
{
"epoch": 1.250843644544432,
"grad_norm": 0.7693307388456105,
"learning_rate": 7.466198300013066e-05,
"loss": 0.5761,
"step": 139
},
{
"epoch": 1.2598425196850394,
"grad_norm": 0.7329449732496208,
"learning_rate": 7.453561026281366e-05,
"loss": 0.5697,
"step": 140
},
{
"epoch": 1.2688413948256467,
"grad_norm": 0.7219015663079241,
"learning_rate": 7.440786864872433e-05,
"loss": 0.5725,
"step": 141
},
{
"epoch": 1.277840269966254,
"grad_norm": 0.7161519080047792,
"learning_rate": 7.427876322111558e-05,
"loss": 0.5737,
"step": 142
},
{
"epoch": 1.2868391451068617,
"grad_norm": 0.7985855229347674,
"learning_rate": 7.414829909729727e-05,
"loss": 0.5724,
"step": 143
},
{
"epoch": 1.295838020247469,
"grad_norm": 0.8867813272578328,
"learning_rate": 7.40164814484336e-05,
"loss": 0.5725,
"step": 144
},
{
"epoch": 1.3048368953880765,
"grad_norm": 0.8113361068419839,
"learning_rate": 7.388331549933787e-05,
"loss": 0.5743,
"step": 145
},
{
"epoch": 1.3138357705286838,
"grad_norm": 0.5476904687412559,
"learning_rate": 7.37488065282656e-05,
"loss": 0.5658,
"step": 146
},
{
"epoch": 1.3228346456692912,
"grad_norm": 0.45189065402105816,
"learning_rate": 7.361295986670522e-05,
"loss": 0.5708,
"step": 147
},
{
"epoch": 1.3318335208098988,
"grad_norm": 0.751469386998812,
"learning_rate": 7.347578089916672e-05,
"loss": 0.5629,
"step": 148
},
{
"epoch": 1.3408323959505062,
"grad_norm": 0.7162670873895776,
"learning_rate": 7.333727506296831e-05,
"loss": 0.5711,
"step": 149
},
{
"epoch": 1.3498312710911136,
"grad_norm": 0.5796332830646129,
"learning_rate": 7.319744784802087e-05,
"loss": 0.5701,
"step": 150
},
{
"epoch": 1.3588301462317212,
"grad_norm": 0.5806837398044375,
"learning_rate": 7.305630479661033e-05,
"loss": 0.568,
"step": 151
},
{
"epoch": 1.3678290213723285,
"grad_norm": 0.36656011961132856,
"learning_rate": 7.291385150317796e-05,
"loss": 0.5704,
"step": 152
},
{
"epoch": 1.376827896512936,
"grad_norm": 0.4854553631469011,
"learning_rate": 7.277009361409874e-05,
"loss": 0.5626,
"step": 153
},
{
"epoch": 1.3858267716535433,
"grad_norm": 0.7060268694492198,
"learning_rate": 7.262503682745744e-05,
"loss": 0.5682,
"step": 154
},
{
"epoch": 1.3948256467941507,
"grad_norm": 0.6290468975159552,
"learning_rate": 7.247868689282283e-05,
"loss": 0.5652,
"step": 155
},
{
"epoch": 1.4038245219347583,
"grad_norm": 0.5124920327952127,
"learning_rate": 7.233104961101974e-05,
"loss": 0.5702,
"step": 156
},
{
"epoch": 1.4128233970753656,
"grad_norm": 0.435666421319146,
"learning_rate": 7.21821308338992e-05,
"loss": 0.5656,
"step": 157
},
{
"epoch": 1.421822272215973,
"grad_norm": 0.31789150940846095,
"learning_rate": 7.203193646410642e-05,
"loss": 0.5625,
"step": 158
},
{
"epoch": 1.4308211473565804,
"grad_norm": 0.44522630815692826,
"learning_rate": 7.188047245484686e-05,
"loss": 0.5622,
"step": 159
},
{
"epoch": 1.4398200224971878,
"grad_norm": 0.5340918126651806,
"learning_rate": 7.172774480965033e-05,
"loss": 0.5663,
"step": 160
},
{
"epoch": 1.4488188976377954,
"grad_norm": 0.5126590833097094,
"learning_rate": 7.157375958213288e-05,
"loss": 0.5604,
"step": 161
},
{
"epoch": 1.4578177727784027,
"grad_norm": 0.5574211540917228,
"learning_rate": 7.141852287575701e-05,
"loss": 0.5644,
"step": 162
},
{
"epoch": 1.4668166479190101,
"grad_norm": 0.49894825274633764,
"learning_rate": 7.126204084358963e-05,
"loss": 0.5543,
"step": 163
},
{
"epoch": 1.4758155230596175,
"grad_norm": 0.383714620666216,
"learning_rate": 7.110431968805825e-05,
"loss": 0.5667,
"step": 164
},
{
"epoch": 1.4848143982002249,
"grad_norm": 0.3679127273669197,
"learning_rate": 7.094536566070514e-05,
"loss": 0.5649,
"step": 165
},
{
"epoch": 1.4938132733408325,
"grad_norm": 0.41052750978866387,
"learning_rate": 7.078518506193945e-05,
"loss": 0.5597,
"step": 166
},
{
"epoch": 1.5028121484814398,
"grad_norm": 0.4161268377444414,
"learning_rate": 7.062378424078758e-05,
"loss": 0.5683,
"step": 167
},
{
"epoch": 1.5118110236220472,
"grad_norm": 0.4143309696590428,
"learning_rate": 7.046116959464149e-05,
"loss": 0.5627,
"step": 168
},
{
"epoch": 1.5208098987626548,
"grad_norm": 0.47622613149080295,
"learning_rate": 7.02973475690051e-05,
"loss": 0.5619,
"step": 169
},
{
"epoch": 1.529808773903262,
"grad_norm": 0.5903824719429629,
"learning_rate": 7.013232465723888e-05,
"loss": 0.5685,
"step": 170
},
{
"epoch": 1.5388076490438696,
"grad_norm": 0.6410524866967787,
"learning_rate": 6.996610740030237e-05,
"loss": 0.5649,
"step": 171
},
{
"epoch": 1.547806524184477,
"grad_norm": 0.5120196634127667,
"learning_rate": 6.979870238649506e-05,
"loss": 0.5641,
"step": 172
},
{
"epoch": 1.5568053993250843,
"grad_norm": 0.46968763374354905,
"learning_rate": 6.963011625119514e-05,
"loss": 0.5607,
"step": 173
},
{
"epoch": 1.565804274465692,
"grad_norm": 0.7346438690156033,
"learning_rate": 6.94603556765965e-05,
"loss": 0.5632,
"step": 174
},
{
"epoch": 1.574803149606299,
"grad_norm": 0.9133676726822815,
"learning_rate": 6.928942739144394e-05,
"loss": 0.5561,
"step": 175
},
{
"epoch": 1.5838020247469067,
"grad_norm": 0.9699546226599123,
"learning_rate": 6.911733817076638e-05,
"loss": 0.5684,
"step": 176
},
{
"epoch": 1.592800899887514,
"grad_norm": 0.8546409930461957,
"learning_rate": 6.894409483560845e-05,
"loss": 0.5636,
"step": 177
},
{
"epoch": 1.6017997750281214,
"grad_norm": 0.5767270121981363,
"learning_rate": 6.876970425275993e-05,
"loss": 0.5636,
"step": 178
},
{
"epoch": 1.610798650168729,
"grad_norm": 0.4603868453512875,
"learning_rate": 6.859417333448376e-05,
"loss": 0.5657,
"step": 179
},
{
"epoch": 1.6197975253093362,
"grad_norm": 0.4657088093589826,
"learning_rate": 6.841750903824196e-05,
"loss": 0.553,
"step": 180
},
{
"epoch": 1.6287964004499438,
"grad_norm": 0.4245830377495125,
"learning_rate": 6.823971836641988e-05,
"loss": 0.553,
"step": 181
},
{
"epoch": 1.6377952755905512,
"grad_norm": 0.38313124117842345,
"learning_rate": 6.806080836604868e-05,
"loss": 0.5606,
"step": 182
},
{
"epoch": 1.6467941507311585,
"grad_norm": 0.3951074598265005,
"learning_rate": 6.788078612852596e-05,
"loss": 0.5611,
"step": 183
},
{
"epoch": 1.6557930258717661,
"grad_norm": 0.3324519467665917,
"learning_rate": 6.769965878933468e-05,
"loss": 0.5589,
"step": 184
},
{
"epoch": 1.6647919010123733,
"grad_norm": 0.30525105111986356,
"learning_rate": 6.751743352776041e-05,
"loss": 0.5513,
"step": 185
},
{
"epoch": 1.6737907761529809,
"grad_norm": 0.38317669421888395,
"learning_rate": 6.733411756660668e-05,
"loss": 0.561,
"step": 186
},
{
"epoch": 1.6827896512935883,
"grad_norm": 0.36452922899831897,
"learning_rate": 6.714971817190872e-05,
"loss": 0.5565,
"step": 187
},
{
"epoch": 1.6917885264341956,
"grad_norm": 0.45451136606739156,
"learning_rate": 6.696424265264549e-05,
"loss": 0.5554,
"step": 188
},
{
"epoch": 1.7007874015748032,
"grad_norm": 0.5850188409579592,
"learning_rate": 6.677769836044991e-05,
"loss": 0.5603,
"step": 189
},
{
"epoch": 1.7097862767154106,
"grad_norm": 0.6094742834131825,
"learning_rate": 6.659009268931756e-05,
"loss": 0.5574,
"step": 190
},
{
"epoch": 1.718785151856018,
"grad_norm": 0.6284162963180803,
"learning_rate": 6.64014330753135e-05,
"loss": 0.5507,
"step": 191
},
{
"epoch": 1.7277840269966256,
"grad_norm": 0.6446183014432966,
"learning_rate": 6.621172699627761e-05,
"loss": 0.5551,
"step": 192
},
{
"epoch": 1.7367829021372327,
"grad_norm": 0.6604713611773748,
"learning_rate": 6.602098197152817e-05,
"loss": 0.555,
"step": 193
},
{
"epoch": 1.7457817772778403,
"grad_norm": 0.7414347856141247,
"learning_rate": 6.582920556156378e-05,
"loss": 0.5631,
"step": 194
},
{
"epoch": 1.7547806524184477,
"grad_norm": 0.7773047456674285,
"learning_rate": 6.563640536776375e-05,
"loss": 0.56,
"step": 195
},
{
"epoch": 1.763779527559055,
"grad_norm": 0.6417760152496987,
"learning_rate": 6.544258903208679e-05,
"loss": 0.5518,
"step": 196
},
{
"epoch": 1.7727784026996627,
"grad_norm": 0.3607864424657309,
"learning_rate": 6.524776423676806e-05,
"loss": 0.55,
"step": 197
},
{
"epoch": 1.7817772778402698,
"grad_norm": 0.4683312923869542,
"learning_rate": 6.505193870401472e-05,
"loss": 0.5579,
"step": 198
},
{
"epoch": 1.7907761529808774,
"grad_norm": 0.6344422826239488,
"learning_rate": 6.485512019569986e-05,
"loss": 0.5628,
"step": 199
},
{
"epoch": 1.7997750281214848,
"grad_norm": 0.48250024105262346,
"learning_rate": 6.465731651305475e-05,
"loss": 0.5561,
"step": 200
},
{
"epoch": 1.8087739032620922,
"grad_norm": 0.4011620009633319,
"learning_rate": 6.445853549635982e-05,
"loss": 0.5589,
"step": 201
},
{
"epoch": 1.8177727784026998,
"grad_norm": 0.46120746083720787,
"learning_rate": 6.425878502463363e-05,
"loss": 0.5464,
"step": 202
},
{
"epoch": 1.826771653543307,
"grad_norm": 0.3366919397807797,
"learning_rate": 6.405807301532082e-05,
"loss": 0.5583,
"step": 203
},
{
"epoch": 1.8357705286839145,
"grad_norm": 0.3775528944092533,
"learning_rate": 6.38564074239781e-05,
"loss": 0.5515,
"step": 204
},
{
"epoch": 1.844769403824522,
"grad_norm": 0.4646001384485239,
"learning_rate": 6.365379624395911e-05,
"loss": 0.5545,
"step": 205
},
{
"epoch": 1.8537682789651293,
"grad_norm": 0.3173359890548618,
"learning_rate": 6.345024750609735e-05,
"loss": 0.5511,
"step": 206
},
{
"epoch": 1.862767154105737,
"grad_norm": 0.30771535160757224,
"learning_rate": 6.324576927838811e-05,
"loss": 0.5569,
"step": 207
},
{
"epoch": 1.871766029246344,
"grad_norm": 0.39138826318539643,
"learning_rate": 6.30403696656685e-05,
"loss": 0.5507,
"step": 208
},
{
"epoch": 1.8807649043869517,
"grad_norm": 0.39402910206893743,
"learning_rate": 6.28340568092963e-05,
"loss": 0.551,
"step": 209
},
{
"epoch": 1.889763779527559,
"grad_norm": 0.42067586027851417,
"learning_rate": 6.26268388868272e-05,
"loss": 0.5552,
"step": 210
},
{
"epoch": 1.8987626546681664,
"grad_norm": 0.39427161933244775,
"learning_rate": 6.241872411169075e-05,
"loss": 0.5515,
"step": 211
},
{
"epoch": 1.907761529808774,
"grad_norm": 0.2940525255889463,
"learning_rate": 6.220972073286469e-05,
"loss": 0.5452,
"step": 212
},
{
"epoch": 1.9167604049493814,
"grad_norm": 0.34048373966582285,
"learning_rate": 6.199983703454813e-05,
"loss": 0.5509,
"step": 213
},
{
"epoch": 1.9257592800899888,
"grad_norm": 0.299832377853422,
"learning_rate": 6.178908133583306e-05,
"loss": 0.5456,
"step": 214
},
{
"epoch": 1.9347581552305961,
"grad_norm": 0.2392599130105496,
"learning_rate": 6.157746199037473e-05,
"loss": 0.5525,
"step": 215
},
{
"epoch": 1.9437570303712035,
"grad_norm": 0.2884892083080534,
"learning_rate": 6.136498738606038e-05,
"loss": 0.5494,
"step": 216
},
{
"epoch": 1.952755905511811,
"grad_norm": 0.25915586035077326,
"learning_rate": 6.115166594467696e-05,
"loss": 0.5578,
"step": 217
},
{
"epoch": 1.9617547806524185,
"grad_norm": 0.2404861202851113,
"learning_rate": 6.093750612157719e-05,
"loss": 0.5501,
"step": 218
},
{
"epoch": 1.9707536557930259,
"grad_norm": 0.27079067912862353,
"learning_rate": 6.0722516405344436e-05,
"loss": 0.5544,
"step": 219
},
{
"epoch": 1.9797525309336335,
"grad_norm": 0.27569696549284556,
"learning_rate": 6.050670531745629e-05,
"loss": 0.5436,
"step": 220
},
{
"epoch": 1.9887514060742406,
"grad_norm": 0.39262044690390413,
"learning_rate": 6.0290081411946785e-05,
"loss": 0.5589,
"step": 221
},
{
"epoch": 1.9977502812148482,
"grad_norm": 0.3759638408126938,
"learning_rate": 6.007265327506734e-05,
"loss": 0.5489,
"step": 222
},
{
"epoch": 2.0067491563554554,
"grad_norm": 0.6721068777680985,
"learning_rate": 5.985442952494643e-05,
"loss": 0.9973,
"step": 223
},
{
"epoch": 2.015748031496063,
"grad_norm": 1.2532035078989137,
"learning_rate": 5.9635418811248e-05,
"loss": 0.5251,
"step": 224
},
{
"epoch": 2.0247469066366706,
"grad_norm": 1.3965146746879655,
"learning_rate": 5.941562981482859e-05,
"loss": 0.5222,
"step": 225
},
{
"epoch": 2.0337457817772777,
"grad_norm": 0.47181130829523693,
"learning_rate": 5.9195071247393325e-05,
"loss": 0.5175,
"step": 226
},
{
"epoch": 2.0427446569178853,
"grad_norm": 1.2945631540598115,
"learning_rate": 5.897375185115052e-05,
"loss": 0.5219,
"step": 227
},
{
"epoch": 2.0517435320584925,
"grad_norm": 0.753799335003198,
"learning_rate": 5.8751680398465244e-05,
"loss": 0.5248,
"step": 228
},
{
"epoch": 2.0607424071991,
"grad_norm": 0.7482718095479951,
"learning_rate": 5.8528865691511564e-05,
"loss": 0.531,
"step": 229
},
{
"epoch": 2.0697412823397077,
"grad_norm": 0.9072745588806806,
"learning_rate": 5.83053165619237e-05,
"loss": 0.5185,
"step": 230
},
{
"epoch": 2.078740157480315,
"grad_norm": 0.7594832390100079,
"learning_rate": 5.808104187044592e-05,
"loss": 0.5238,
"step": 231
},
{
"epoch": 2.0877390326209224,
"grad_norm": 0.7018923770640809,
"learning_rate": 5.785605050658134e-05,
"loss": 0.5116,
"step": 232
},
{
"epoch": 2.09673790776153,
"grad_norm": 0.6538828939619331,
"learning_rate": 5.7630351388239654e-05,
"loss": 0.5188,
"step": 233
},
{
"epoch": 2.105736782902137,
"grad_norm": 0.5009696448905507,
"learning_rate": 5.7403953461383515e-05,
"loss": 0.5189,
"step": 234
},
{
"epoch": 2.1147356580427448,
"grad_norm": 0.5418468893978772,
"learning_rate": 5.717686569967406e-05,
"loss": 0.513,
"step": 235
},
{
"epoch": 2.123734533183352,
"grad_norm": 0.5552312277008503,
"learning_rate": 5.694909710411517e-05,
"loss": 0.5136,
"step": 236
},
{
"epoch": 2.1327334083239595,
"grad_norm": 0.4618715663984918,
"learning_rate": 5.672065670269674e-05,
"loss": 0.5226,
"step": 237
},
{
"epoch": 2.141732283464567,
"grad_norm": 0.4634338024463565,
"learning_rate": 5.649155355003677e-05,
"loss": 0.517,
"step": 238
},
{
"epoch": 2.1507311586051743,
"grad_norm": 0.35434206430397164,
"learning_rate": 5.6261796727022575e-05,
"loss": 0.5166,
"step": 239
},
{
"epoch": 2.159730033745782,
"grad_norm": 0.46283864431852784,
"learning_rate": 5.603139534045075e-05,
"loss": 0.5112,
"step": 240
},
{
"epoch": 2.168728908886389,
"grad_norm": 0.42575715836926703,
"learning_rate": 5.5800358522666254e-05,
"loss": 0.5078,
"step": 241
},
{
"epoch": 2.1777277840269966,
"grad_norm": 0.33272119313147885,
"learning_rate": 5.556869543120043e-05,
"loss": 0.5194,
"step": 242
},
{
"epoch": 2.1867266591676042,
"grad_norm": 0.3307000934681126,
"learning_rate": 5.533641524840805e-05,
"loss": 0.5177,
"step": 243
},
{
"epoch": 2.1957255343082114,
"grad_norm": 0.3028678329115125,
"learning_rate": 5.5103527181103266e-05,
"loss": 0.5169,
"step": 244
},
{
"epoch": 2.204724409448819,
"grad_norm": 0.3208549750095541,
"learning_rate": 5.4870040460194834e-05,
"loss": 0.5166,
"step": 245
},
{
"epoch": 2.213723284589426,
"grad_norm": 0.24593852068559713,
"learning_rate": 5.463596434032011e-05,
"loss": 0.5162,
"step": 246
},
{
"epoch": 2.2227221597300337,
"grad_norm": 0.26893184750686033,
"learning_rate": 5.440130809947824e-05,
"loss": 0.5211,
"step": 247
},
{
"epoch": 2.2317210348706413,
"grad_norm": 0.2453740819753266,
"learning_rate": 5.4166081038662484e-05,
"loss": 0.5114,
"step": 248
},
{
"epoch": 2.2407199100112485,
"grad_norm": 0.22939493074869363,
"learning_rate": 5.3930292481491494e-05,
"loss": 0.5101,
"step": 249
},
{
"epoch": 2.249718785151856,
"grad_norm": 0.25733357689003183,
"learning_rate": 5.3693951773839736e-05,
"loss": 0.5111,
"step": 250
},
{
"epoch": 2.2587176602924632,
"grad_norm": 0.2252436543398025,
"learning_rate": 5.345706828346715e-05,
"loss": 0.516,
"step": 251
},
{
"epoch": 2.267716535433071,
"grad_norm": 0.28121431385301127,
"learning_rate": 5.3219651399647716e-05,
"loss": 0.5168,
"step": 252
},
{
"epoch": 2.2767154105736784,
"grad_norm": 0.2450809438398232,
"learning_rate": 5.298171053279739e-05,
"loss": 0.5089,
"step": 253
},
{
"epoch": 2.2857142857142856,
"grad_norm": 0.22424110667284097,
"learning_rate": 5.274325511410106e-05,
"loss": 0.5165,
"step": 254
},
{
"epoch": 2.294713160854893,
"grad_norm": 0.2360828175959895,
"learning_rate": 5.250429459513876e-05,
"loss": 0.5148,
"step": 255
},
{
"epoch": 2.303712035995501,
"grad_norm": 0.2188304429809361,
"learning_rate": 5.226483844751099e-05,
"loss": 0.5098,
"step": 256
},
{
"epoch": 2.312710911136108,
"grad_norm": 0.2334512539619716,
"learning_rate": 5.202489616246333e-05,
"loss": 0.5158,
"step": 257
},
{
"epoch": 2.3217097862767155,
"grad_norm": 0.19345732584763242,
"learning_rate": 5.178447725051026e-05,
"loss": 0.5137,
"step": 258
},
{
"epoch": 2.3307086614173227,
"grad_norm": 0.22726718301152563,
"learning_rate": 5.154359124105814e-05,
"loss": 0.5149,
"step": 259
},
{
"epoch": 2.3397075365579303,
"grad_norm": 0.2149304217989204,
"learning_rate": 5.1302247682027494e-05,
"loss": 0.5122,
"step": 260
},
{
"epoch": 2.3487064116985374,
"grad_norm": 0.2574574491521757,
"learning_rate": 5.106045613947466e-05,
"loss": 0.5106,
"step": 261
},
{
"epoch": 2.357705286839145,
"grad_norm": 0.2244044383603563,
"learning_rate": 5.08182261972125e-05,
"loss": 0.5098,
"step": 262
},
{
"epoch": 2.3667041619797526,
"grad_norm": 0.1722686877882969,
"learning_rate": 5.057556745643058e-05,
"loss": 0.5152,
"step": 263
},
{
"epoch": 2.37570303712036,
"grad_norm": 0.20786422486929515,
"learning_rate": 5.033248953531466e-05,
"loss": 0.5096,
"step": 264
},
{
"epoch": 2.3847019122609674,
"grad_norm": 0.20518803463398436,
"learning_rate": 5.0089002068665376e-05,
"loss": 0.5108,
"step": 265
},
{
"epoch": 2.393700787401575,
"grad_norm": 0.19783852278781122,
"learning_rate": 4.9845114707516395e-05,
"loss": 0.5169,
"step": 266
},
{
"epoch": 2.402699662542182,
"grad_norm": 0.15894601748576243,
"learning_rate": 4.9600837118751874e-05,
"loss": 0.5077,
"step": 267
},
{
"epoch": 2.4116985376827897,
"grad_norm": 0.19443345762052888,
"learning_rate": 4.93561789847233e-05,
"loss": 0.5211,
"step": 268
},
{
"epoch": 2.420697412823397,
"grad_norm": 0.19638270288569096,
"learning_rate": 4.91111500028657e-05,
"loss": 0.5098,
"step": 269
},
{
"epoch": 2.4296962879640045,
"grad_norm": 0.17816319122872262,
"learning_rate": 4.886575988531329e-05,
"loss": 0.5175,
"step": 270
},
{
"epoch": 2.438695163104612,
"grad_norm": 0.17651511328007938,
"learning_rate": 4.862001835851449e-05,
"loss": 0.5136,
"step": 271
},
{
"epoch": 2.4476940382452193,
"grad_norm": 0.1840295184610878,
"learning_rate": 4.837393516284642e-05,
"loss": 0.5099,
"step": 272
},
{
"epoch": 2.456692913385827,
"grad_norm": 0.1914170149726011,
"learning_rate": 4.8127520052228815e-05,
"loss": 0.5165,
"step": 273
},
{
"epoch": 2.465691788526434,
"grad_norm": 0.184680647588443,
"learning_rate": 4.788078279373744e-05,
"loss": 0.5084,
"step": 274
},
{
"epoch": 2.4746906636670416,
"grad_norm": 0.22287775595578857,
"learning_rate": 4.763373316721687e-05,
"loss": 0.5134,
"step": 275
},
{
"epoch": 2.483689538807649,
"grad_norm": 0.25180987790543646,
"learning_rate": 4.7386380964892984e-05,
"loss": 0.5137,
"step": 276
},
{
"epoch": 2.4926884139482564,
"grad_norm": 0.2645098272392606,
"learning_rate": 4.713873599098471e-05,
"loss": 0.5128,
"step": 277
},
{
"epoch": 2.501687289088864,
"grad_norm": 0.16671664940075334,
"learning_rate": 4.689080806131547e-05,
"loss": 0.5114,
"step": 278
},
{
"epoch": 2.5106861642294716,
"grad_norm": 0.22458637707427415,
"learning_rate": 4.664260700292416e-05,
"loss": 0.5095,
"step": 279
},
{
"epoch": 2.5196850393700787,
"grad_norm": 0.20330358237599677,
"learning_rate": 4.639414265367554e-05,
"loss": 0.509,
"step": 280
},
{
"epoch": 2.5286839145106863,
"grad_norm": 0.19076267160273108,
"learning_rate": 4.614542486187033e-05,
"loss": 0.5135,
"step": 281
},
{
"epoch": 2.5376827896512935,
"grad_norm": 0.19390672620858349,
"learning_rate": 4.589646348585494e-05,
"loss": 0.5129,
"step": 282
},
{
"epoch": 2.546681664791901,
"grad_norm": 0.17431590192861887,
"learning_rate": 4.564726839363059e-05,
"loss": 0.5119,
"step": 283
},
{
"epoch": 2.555680539932508,
"grad_norm": 0.19796573953993535,
"learning_rate": 4.539784946246225e-05,
"loss": 0.5226,
"step": 284
},
{
"epoch": 2.564679415073116,
"grad_norm": 0.21806700674738302,
"learning_rate": 4.5148216578487134e-05,
"loss": 0.5205,
"step": 285
},
{
"epoch": 2.5736782902137234,
"grad_norm": 0.21026084289689406,
"learning_rate": 4.4898379636322815e-05,
"loss": 0.5078,
"step": 286
},
{
"epoch": 2.5826771653543306,
"grad_norm": 0.18917285015389398,
"learning_rate": 4.4648348538675064e-05,
"loss": 0.5146,
"step": 287
},
{
"epoch": 2.591676040494938,
"grad_norm": 0.17876905148977648,
"learning_rate": 4.4398133195945326e-05,
"loss": 0.5124,
"step": 288
},
{
"epoch": 2.6006749156355458,
"grad_norm": 0.19991148697525993,
"learning_rate": 4.414774352583791e-05,
"loss": 0.5141,
"step": 289
},
{
"epoch": 2.609673790776153,
"grad_norm": 0.1718371307146163,
"learning_rate": 4.3897189452966895e-05,
"loss": 0.5126,
"step": 290
},
{
"epoch": 2.6186726659167605,
"grad_norm": 0.1685835159809121,
"learning_rate": 4.364648090846271e-05,
"loss": 0.5133,
"step": 291
},
{
"epoch": 2.6276715410573677,
"grad_norm": 0.14950354028807825,
"learning_rate": 4.339562782957857e-05,
"loss": 0.5083,
"step": 292
},
{
"epoch": 2.6366704161979753,
"grad_norm": 0.17804715853771555,
"learning_rate": 4.314464015929649e-05,
"loss": 0.5194,
"step": 293
},
{
"epoch": 2.6456692913385824,
"grad_norm": 0.17707266681302694,
"learning_rate": 4.28935278459333e-05,
"loss": 0.5128,
"step": 294
},
{
"epoch": 2.65466816647919,
"grad_norm": 0.15273475864497393,
"learning_rate": 4.264230084274624e-05,
"loss": 0.5148,
"step": 295
},
{
"epoch": 2.6636670416197976,
"grad_norm": 0.19292123693836521,
"learning_rate": 4.239096910753846e-05,
"loss": 0.5097,
"step": 296
},
{
"epoch": 2.6726659167604048,
"grad_norm": 0.19491285573674288,
"learning_rate": 4.213954260226438e-05,
"loss": 0.5108,
"step": 297
},
{
"epoch": 2.6816647919010124,
"grad_norm": 0.16377738874278228,
"learning_rate": 4.188803129263476e-05,
"loss": 0.5084,
"step": 298
},
{
"epoch": 2.69066366704162,
"grad_norm": 0.17707808785911283,
"learning_rate": 4.163644514772172e-05,
"loss": 0.5078,
"step": 299
},
{
"epoch": 2.699662542182227,
"grad_norm": 0.16129691634220186,
"learning_rate": 4.1384794139563614e-05,
"loss": 0.5177,
"step": 300
},
{
"epoch": 2.7086614173228347,
"grad_norm": 0.19221250489172526,
"learning_rate": 4.113308824276977e-05,
"loss": 0.5149,
"step": 301
},
{
"epoch": 2.7176602924634423,
"grad_norm": 0.17286893566543513,
"learning_rate": 4.0881337434125086e-05,
"loss": 0.5088,
"step": 302
},
{
"epoch": 2.7266591676040495,
"grad_norm": 0.1755459663317774,
"learning_rate": 4.0629551692194634e-05,
"loss": 0.509,
"step": 303
},
{
"epoch": 2.735658042744657,
"grad_norm": 0.1791201340425473,
"learning_rate": 4.037774099692815e-05,
"loss": 0.5143,
"step": 304
},
{
"epoch": 2.7446569178852642,
"grad_norm": 0.18143794043338418,
"learning_rate": 4.0125915329264396e-05,
"loss": 0.5123,
"step": 305
},
{
"epoch": 2.753655793025872,
"grad_norm": 0.15111984385935012,
"learning_rate": 3.9874084670735624e-05,
"loss": 0.5132,
"step": 306
},
{
"epoch": 2.762654668166479,
"grad_norm": 0.17406268964374014,
"learning_rate": 3.962225900307187e-05,
"loss": 0.5088,
"step": 307
},
{
"epoch": 2.7716535433070866,
"grad_norm": 0.1845218510232061,
"learning_rate": 3.937044830780537e-05,
"loss": 0.5115,
"step": 308
},
{
"epoch": 2.780652418447694,
"grad_norm": 0.14408452277861603,
"learning_rate": 3.9118662565874934e-05,
"loss": 0.5088,
"step": 309
},
{
"epoch": 2.7896512935883013,
"grad_norm": 0.2121818690921875,
"learning_rate": 3.886691175723025e-05,
"loss": 0.515,
"step": 310
},
{
"epoch": 2.798650168728909,
"grad_norm": 0.17415478680558893,
"learning_rate": 3.8615205860436406e-05,
"loss": 0.5102,
"step": 311
},
{
"epoch": 2.8076490438695165,
"grad_norm": 0.17710673827275958,
"learning_rate": 3.83635548522783e-05,
"loss": 0.5142,
"step": 312
},
{
"epoch": 2.8166479190101237,
"grad_norm": 0.2131327323874231,
"learning_rate": 3.811196870736526e-05,
"loss": 0.5143,
"step": 313
},
{
"epoch": 2.8256467941507313,
"grad_norm": 0.17135728004223627,
"learning_rate": 3.786045739773564e-05,
"loss": 0.5066,
"step": 314
},
{
"epoch": 2.8346456692913384,
"grad_norm": 0.1787641234702102,
"learning_rate": 3.7609030892461554e-05,
"loss": 0.5077,
"step": 315
},
{
"epoch": 2.843644544431946,
"grad_norm": 0.1737180877569213,
"learning_rate": 3.735769915725378e-05,
"loss": 0.5104,
"step": 316
},
{
"epoch": 2.852643419572553,
"grad_norm": 0.1759592145643783,
"learning_rate": 3.710647215406672e-05,
"loss": 0.5076,
"step": 317
},
{
"epoch": 2.861642294713161,
"grad_norm": 0.1674590421194775,
"learning_rate": 3.6855359840703525e-05,
"loss": 0.5117,
"step": 318
},
{
"epoch": 2.8706411698537684,
"grad_norm": 0.19195727087026487,
"learning_rate": 3.660437217042145e-05,
"loss": 0.515,
"step": 319
},
{
"epoch": 2.8796400449943755,
"grad_norm": 0.2079526174400504,
"learning_rate": 3.63535190915373e-05,
"loss": 0.5114,
"step": 320
},
{
"epoch": 2.888638920134983,
"grad_norm": 0.16261938739039505,
"learning_rate": 3.610281054703311e-05,
"loss": 0.512,
"step": 321
},
{
"epoch": 2.8976377952755907,
"grad_norm": 0.17301946358427484,
"learning_rate": 3.58522564741621e-05,
"loss": 0.513,
"step": 322
},
{
"epoch": 2.906636670416198,
"grad_norm": 0.15809619671547212,
"learning_rate": 3.560186680405469e-05,
"loss": 0.5092,
"step": 323
},
{
"epoch": 2.9156355455568055,
"grad_norm": 0.18005047534402024,
"learning_rate": 3.535165146132494e-05,
"loss": 0.5143,
"step": 324
},
{
"epoch": 2.924634420697413,
"grad_norm": 0.17055434106531733,
"learning_rate": 3.51016203636772e-05,
"loss": 0.5044,
"step": 325
},
{
"epoch": 2.9336332958380202,
"grad_norm": 0.2091344061636353,
"learning_rate": 3.485178342151287e-05,
"loss": 0.5151,
"step": 326
},
{
"epoch": 2.942632170978628,
"grad_norm": 0.17776622234467257,
"learning_rate": 3.460215053753776e-05,
"loss": 0.506,
"step": 327
},
{
"epoch": 2.951631046119235,
"grad_norm": 0.2020120609501547,
"learning_rate": 3.435273160636942e-05,
"loss": 0.5106,
"step": 328
},
{
"epoch": 2.9606299212598426,
"grad_norm": 0.17580656405878176,
"learning_rate": 3.410353651414507e-05,
"loss": 0.509,
"step": 329
},
{
"epoch": 2.9696287964004497,
"grad_norm": 0.1874074410148767,
"learning_rate": 3.385457513812968e-05,
"loss": 0.5119,
"step": 330
},
{
"epoch": 2.9786276715410573,
"grad_norm": 0.18596738197007068,
"learning_rate": 3.360585734632448e-05,
"loss": 0.5105,
"step": 331
},
{
"epoch": 2.987626546681665,
"grad_norm": 0.18012608783890774,
"learning_rate": 3.3357392997075854e-05,
"loss": 0.51,
"step": 332
},
{
"epoch": 2.996625421822272,
"grad_norm": 0.17695247800719235,
"learning_rate": 3.3109191938684535e-05,
"loss": 0.5096,
"step": 333
},
{
"epoch": 3.0056242969628797,
"grad_norm": 0.4041448073125774,
"learning_rate": 3.28612640090153e-05,
"loss": 0.9217,
"step": 334
},
{
"epoch": 3.014623172103487,
"grad_norm": 0.40335779907517166,
"learning_rate": 3.261361903510703e-05,
"loss": 0.4749,
"step": 335
},
{
"epoch": 3.0236220472440944,
"grad_norm": 0.23571838314313354,
"learning_rate": 3.2366266832783145e-05,
"loss": 0.4795,
"step": 336
},
{
"epoch": 3.032620922384702,
"grad_norm": 0.34962469353680325,
"learning_rate": 3.211921720626258e-05,
"loss": 0.4772,
"step": 337
},
{
"epoch": 3.041619797525309,
"grad_norm": 0.3250568724209024,
"learning_rate": 3.187247994777119e-05,
"loss": 0.4756,
"step": 338
},
{
"epoch": 3.050618672665917,
"grad_norm": 0.27683092123819425,
"learning_rate": 3.1626064837153596e-05,
"loss": 0.478,
"step": 339
},
{
"epoch": 3.0596175478065244,
"grad_norm": 0.29128324637433173,
"learning_rate": 3.1379981641485524e-05,
"loss": 0.4767,
"step": 340
},
{
"epoch": 3.0686164229471316,
"grad_norm": 0.2783998825962506,
"learning_rate": 3.113424011468672e-05,
"loss": 0.4758,
"step": 341
},
{
"epoch": 3.077615298087739,
"grad_norm": 0.25380987160914664,
"learning_rate": 3.0888849997134316e-05,
"loss": 0.4663,
"step": 342
},
{
"epoch": 3.0866141732283463,
"grad_norm": 0.23893069831482153,
"learning_rate": 3.064382101527671e-05,
"loss": 0.4788,
"step": 343
},
{
"epoch": 3.095613048368954,
"grad_norm": 0.2546443581638444,
"learning_rate": 3.039916288124814e-05,
"loss": 0.4794,
"step": 344
},
{
"epoch": 3.1046119235095615,
"grad_norm": 0.2193178093849395,
"learning_rate": 3.015488529248362e-05,
"loss": 0.479,
"step": 345
},
{
"epoch": 3.1136107986501687,
"grad_norm": 0.23966451017854798,
"learning_rate": 2.9910997931334637e-05,
"loss": 0.4754,
"step": 346
},
{
"epoch": 3.1226096737907763,
"grad_norm": 0.2258415034497983,
"learning_rate": 2.9667510464685345e-05,
"loss": 0.4825,
"step": 347
},
{
"epoch": 3.1316085489313834,
"grad_norm": 0.23731192604698906,
"learning_rate": 2.9424432543569428e-05,
"loss": 0.4723,
"step": 348
},
{
"epoch": 3.140607424071991,
"grad_norm": 0.17268744214169082,
"learning_rate": 2.918177380278752e-05,
"loss": 0.4763,
"step": 349
},
{
"epoch": 3.1496062992125986,
"grad_norm": 0.23327367459895917,
"learning_rate": 2.893954386052535e-05,
"loss": 0.4786,
"step": 350
},
{
"epoch": 3.1586051743532058,
"grad_norm": 0.1956862643734178,
"learning_rate": 2.8697752317972513e-05,
"loss": 0.4765,
"step": 351
},
{
"epoch": 3.1676040494938134,
"grad_norm": 0.18627973371325526,
"learning_rate": 2.845640875894188e-05,
"loss": 0.4745,
"step": 352
},
{
"epoch": 3.1766029246344205,
"grad_norm": 0.2105328504046573,
"learning_rate": 2.8215522749489742e-05,
"loss": 0.4804,
"step": 353
},
{
"epoch": 3.185601799775028,
"grad_norm": 0.16272715621667633,
"learning_rate": 2.7975103837536672e-05,
"loss": 0.472,
"step": 354
},
{
"epoch": 3.1946006749156357,
"grad_norm": 0.21097547189014748,
"learning_rate": 2.7735161552489022e-05,
"loss": 0.4749,
"step": 355
},
{
"epoch": 3.203599550056243,
"grad_norm": 0.17765983026821935,
"learning_rate": 2.749570540486125e-05,
"loss": 0.4774,
"step": 356
},
{
"epoch": 3.2125984251968505,
"grad_norm": 0.18185508796151498,
"learning_rate": 2.7256744885898942e-05,
"loss": 0.4769,
"step": 357
},
{
"epoch": 3.2215973003374576,
"grad_norm": 0.17545115382309487,
"learning_rate": 2.7018289467202623e-05,
"loss": 0.4707,
"step": 358
},
{
"epoch": 3.230596175478065,
"grad_norm": 0.16118530505673961,
"learning_rate": 2.6780348600352284e-05,
"loss": 0.4751,
"step": 359
},
{
"epoch": 3.239595050618673,
"grad_norm": 0.17928864347147178,
"learning_rate": 2.6542931716532856e-05,
"loss": 0.4807,
"step": 360
},
{
"epoch": 3.24859392575928,
"grad_norm": 0.16478061360464155,
"learning_rate": 2.630604822616027e-05,
"loss": 0.4736,
"step": 361
},
{
"epoch": 3.2575928008998876,
"grad_norm": 0.1478559482793346,
"learning_rate": 2.6069707518508523e-05,
"loss": 0.4781,
"step": 362
},
{
"epoch": 3.2665916760404947,
"grad_norm": 0.16844157556458098,
"learning_rate": 2.583391896133753e-05,
"loss": 0.4765,
"step": 363
},
{
"epoch": 3.2755905511811023,
"grad_norm": 0.1337175278664685,
"learning_rate": 2.5598691900521778e-05,
"loss": 0.4741,
"step": 364
},
{
"epoch": 3.28458942632171,
"grad_norm": 0.13894808378571294,
"learning_rate": 2.5364035659679914e-05,
"loss": 0.4762,
"step": 365
},
{
"epoch": 3.293588301462317,
"grad_norm": 0.14038211286961325,
"learning_rate": 2.512995953980518e-05,
"loss": 0.4785,
"step": 366
},
{
"epoch": 3.3025871766029247,
"grad_norm": 0.13577552755343492,
"learning_rate": 2.4896472818896743e-05,
"loss": 0.4798,
"step": 367
},
{
"epoch": 3.3115860517435323,
"grad_norm": 0.14414082887532978,
"learning_rate": 2.4663584751591977e-05,
"loss": 0.4784,
"step": 368
},
{
"epoch": 3.3205849268841394,
"grad_norm": 0.1499138881871864,
"learning_rate": 2.443130456879958e-05,
"loss": 0.475,
"step": 369
},
{
"epoch": 3.329583802024747,
"grad_norm": 0.15378837222105818,
"learning_rate": 2.4199641477333766e-05,
"loss": 0.4695,
"step": 370
},
{
"epoch": 3.338582677165354,
"grad_norm": 0.15500949142269607,
"learning_rate": 2.3968604659549266e-05,
"loss": 0.4808,
"step": 371
},
{
"epoch": 3.3475815523059618,
"grad_norm": 0.13510134534596135,
"learning_rate": 2.3738203272977446e-05,
"loss": 0.4762,
"step": 372
},
{
"epoch": 3.3565804274465694,
"grad_norm": 0.1580364610829008,
"learning_rate": 2.350844644996325e-05,
"loss": 0.4756,
"step": 373
},
{
"epoch": 3.3655793025871765,
"grad_norm": 0.1447596119360514,
"learning_rate": 2.3279343297303293e-05,
"loss": 0.476,
"step": 374
},
{
"epoch": 3.374578177727784,
"grad_norm": 0.13783465555399632,
"learning_rate": 2.305090289588485e-05,
"loss": 0.478,
"step": 375
},
{
"epoch": 3.3835770528683913,
"grad_norm": 0.1422413809528974,
"learning_rate": 2.2823134300325948e-05,
"loss": 0.4717,
"step": 376
},
{
"epoch": 3.392575928008999,
"grad_norm": 0.12982390766734286,
"learning_rate": 2.25960465386165e-05,
"loss": 0.4787,
"step": 377
},
{
"epoch": 3.4015748031496065,
"grad_norm": 0.1445554728575935,
"learning_rate": 2.2369648611760352e-05,
"loss": 0.4728,
"step": 378
},
{
"epoch": 3.4105736782902136,
"grad_norm": 0.14091047744757867,
"learning_rate": 2.2143949493418654e-05,
"loss": 0.4778,
"step": 379
},
{
"epoch": 3.4195725534308212,
"grad_norm": 0.15001963385136857,
"learning_rate": 2.1918958129554106e-05,
"loss": 0.4764,
"step": 380
},
{
"epoch": 3.4285714285714284,
"grad_norm": 0.13768741985271635,
"learning_rate": 2.1694683438076317e-05,
"loss": 0.471,
"step": 381
},
{
"epoch": 3.437570303712036,
"grad_norm": 0.14948373435356904,
"learning_rate": 2.147113430848844e-05,
"loss": 0.4697,
"step": 382
},
{
"epoch": 3.4465691788526436,
"grad_norm": 0.13361989178999917,
"learning_rate": 2.1248319601534772e-05,
"loss": 0.4707,
"step": 383
},
{
"epoch": 3.4555680539932507,
"grad_norm": 0.13151795053856452,
"learning_rate": 2.102624814884949e-05,
"loss": 0.4747,
"step": 384
},
{
"epoch": 3.4645669291338583,
"grad_norm": 0.14805117724302444,
"learning_rate": 2.080492875260668e-05,
"loss": 0.4765,
"step": 385
},
{
"epoch": 3.4735658042744655,
"grad_norm": 0.15118784969552582,
"learning_rate": 2.0584370185171418e-05,
"loss": 0.4727,
"step": 386
},
{
"epoch": 3.482564679415073,
"grad_norm": 0.14047665231367965,
"learning_rate": 2.0364581188752012e-05,
"loss": 0.4819,
"step": 387
},
{
"epoch": 3.4915635545556807,
"grad_norm": 0.15664210880968063,
"learning_rate": 2.014557047505357e-05,
"loss": 0.466,
"step": 388
},
{
"epoch": 3.500562429696288,
"grad_norm": 0.13241547005518312,
"learning_rate": 1.992734672493267e-05,
"loss": 0.4772,
"step": 389
},
{
"epoch": 3.5095613048368954,
"grad_norm": 0.15495785626518518,
"learning_rate": 1.970991858805322e-05,
"loss": 0.4768,
"step": 390
},
{
"epoch": 3.518560179977503,
"grad_norm": 0.1451652460628155,
"learning_rate": 1.9493294682543715e-05,
"loss": 0.4752,
"step": 391
},
{
"epoch": 3.52755905511811,
"grad_norm": 0.15292105555983718,
"learning_rate": 1.927748359465558e-05,
"loss": 0.4768,
"step": 392
},
{
"epoch": 3.536557930258718,
"grad_norm": 0.1457418788935837,
"learning_rate": 1.9062493878422823e-05,
"loss": 0.4768,
"step": 393
},
{
"epoch": 3.545556805399325,
"grad_norm": 0.13975287470844083,
"learning_rate": 1.884833405532304e-05,
"loss": 0.4757,
"step": 394
},
{
"epoch": 3.5545556805399325,
"grad_norm": 0.14640008673656155,
"learning_rate": 1.863501261393963e-05,
"loss": 0.4813,
"step": 395
},
{
"epoch": 3.5635545556805397,
"grad_norm": 0.12836792930570248,
"learning_rate": 1.8422538009625285e-05,
"loss": 0.4738,
"step": 396
},
{
"epoch": 3.5725534308211473,
"grad_norm": 0.13526217288902667,
"learning_rate": 1.8210918664166945e-05,
"loss": 0.4762,
"step": 397
},
{
"epoch": 3.581552305961755,
"grad_norm": 0.1420752062696319,
"learning_rate": 1.8000162965451884e-05,
"loss": 0.4803,
"step": 398
},
{
"epoch": 3.590551181102362,
"grad_norm": 0.1279945537635078,
"learning_rate": 1.7790279267135317e-05,
"loss": 0.4829,
"step": 399
},
{
"epoch": 3.5995500562429696,
"grad_norm": 0.13720312932073742,
"learning_rate": 1.758127588830928e-05,
"loss": 0.4756,
"step": 400
},
{
"epoch": 3.6085489313835772,
"grad_norm": 0.11571164848869406,
"learning_rate": 1.737316111317281e-05,
"loss": 0.4788,
"step": 401
},
{
"epoch": 3.6175478065241844,
"grad_norm": 0.1267184023160666,
"learning_rate": 1.716594319070371e-05,
"loss": 0.4672,
"step": 402
},
{
"epoch": 3.626546681664792,
"grad_norm": 0.12093756371303144,
"learning_rate": 1.695963033433151e-05,
"loss": 0.4767,
"step": 403
},
{
"epoch": 3.6355455568053996,
"grad_norm": 0.1281599915828755,
"learning_rate": 1.6754230721611896e-05,
"loss": 0.4762,
"step": 404
},
{
"epoch": 3.6445444319460067,
"grad_norm": 0.12299228584682197,
"learning_rate": 1.654975249390265e-05,
"loss": 0.477,
"step": 405
},
{
"epoch": 3.653543307086614,
"grad_norm": 0.13855705326136183,
"learning_rate": 1.634620375604091e-05,
"loss": 0.4791,
"step": 406
},
{
"epoch": 3.6625421822272215,
"grad_norm": 0.12445550210040791,
"learning_rate": 1.6143592576021897e-05,
"loss": 0.4746,
"step": 407
},
{
"epoch": 3.671541057367829,
"grad_norm": 0.1328319119905009,
"learning_rate": 1.594192698467919e-05,
"loss": 0.4753,
"step": 408
},
{
"epoch": 3.6805399325084363,
"grad_norm": 0.12084524067527391,
"learning_rate": 1.574121497536638e-05,
"loss": 0.4778,
"step": 409
},
{
"epoch": 3.689538807649044,
"grad_norm": 0.12320850922380423,
"learning_rate": 1.5541464503640195e-05,
"loss": 0.4718,
"step": 410
},
{
"epoch": 3.6985376827896514,
"grad_norm": 0.11554496711321419,
"learning_rate": 1.534268348694524e-05,
"loss": 0.4734,
"step": 411
},
{
"epoch": 3.7075365579302586,
"grad_norm": 0.11815716191127555,
"learning_rate": 1.5144879804300163e-05,
"loss": 0.4701,
"step": 412
},
{
"epoch": 3.716535433070866,
"grad_norm": 0.12388808731535488,
"learning_rate": 1.4948061295985286e-05,
"loss": 0.4786,
"step": 413
},
{
"epoch": 3.725534308211474,
"grad_norm": 0.11077052559901988,
"learning_rate": 1.4752235763231944e-05,
"loss": 0.4716,
"step": 414
},
{
"epoch": 3.734533183352081,
"grad_norm": 0.11829062763031385,
"learning_rate": 1.4557410967913219e-05,
"loss": 0.4819,
"step": 415
},
{
"epoch": 3.7435320584926886,
"grad_norm": 0.12116279675489595,
"learning_rate": 1.4363594632236249e-05,
"loss": 0.4746,
"step": 416
},
{
"epoch": 3.7525309336332957,
"grad_norm": 0.11924444626526026,
"learning_rate": 1.4170794438436236e-05,
"loss": 0.4826,
"step": 417
},
{
"epoch": 3.7615298087739033,
"grad_norm": 0.1117894145977932,
"learning_rate": 1.3979018028471858e-05,
"loss": 0.4719,
"step": 418
},
{
"epoch": 3.7705286839145105,
"grad_norm": 0.11651706087134892,
"learning_rate": 1.3788273003722404e-05,
"loss": 0.4789,
"step": 419
},
{
"epoch": 3.779527559055118,
"grad_norm": 0.11055442276894334,
"learning_rate": 1.3598566924686511e-05,
"loss": 0.4796,
"step": 420
},
{
"epoch": 3.7885264341957257,
"grad_norm": 0.10345352334778583,
"learning_rate": 1.3409907310682462e-05,
"loss": 0.4709,
"step": 421
},
{
"epoch": 3.797525309336333,
"grad_norm": 0.10232180540617608,
"learning_rate": 1.3222301639550099e-05,
"loss": 0.4766,
"step": 422
},
{
"epoch": 3.8065241844769404,
"grad_norm": 0.10715279829216998,
"learning_rate": 1.3035757347354526e-05,
"loss": 0.4776,
"step": 423
},
{
"epoch": 3.815523059617548,
"grad_norm": 0.1047852036739672,
"learning_rate": 1.2850281828091298e-05,
"loss": 0.4854,
"step": 424
},
{
"epoch": 3.824521934758155,
"grad_norm": 0.10690185941320834,
"learning_rate": 1.2665882433393338e-05,
"loss": 0.4729,
"step": 425
},
{
"epoch": 3.8335208098987628,
"grad_norm": 0.11484140024382984,
"learning_rate": 1.24825664722396e-05,
"loss": 0.475,
"step": 426
},
{
"epoch": 3.84251968503937,
"grad_norm": 0.11568039124788146,
"learning_rate": 1.2300341210665336e-05,
"loss": 0.4758,
"step": 427
},
{
"epoch": 3.8515185601799775,
"grad_norm": 0.10120504659325706,
"learning_rate": 1.211921387147406e-05,
"loss": 0.4732,
"step": 428
},
{
"epoch": 3.8605174353205847,
"grad_norm": 0.1194598344922678,
"learning_rate": 1.1939191633951328e-05,
"loss": 0.4768,
"step": 429
},
{
"epoch": 3.8695163104611923,
"grad_norm": 0.1090408000804906,
"learning_rate": 1.1760281633580136e-05,
"loss": 0.4741,
"step": 430
},
{
"epoch": 3.8785151856018,
"grad_norm": 0.10733618962308203,
"learning_rate": 1.1582490961758057e-05,
"loss": 0.479,
"step": 431
},
{
"epoch": 3.887514060742407,
"grad_norm": 0.11031251576967621,
"learning_rate": 1.1405826665516253e-05,
"loss": 0.4768,
"step": 432
},
{
"epoch": 3.8965129358830146,
"grad_norm": 0.11348215054291673,
"learning_rate": 1.1230295747240092e-05,
"loss": 0.4714,
"step": 433
},
{
"epoch": 3.905511811023622,
"grad_norm": 0.10993620934878778,
"learning_rate": 1.1055905164391567e-05,
"loss": 0.4785,
"step": 434
},
{
"epoch": 3.9145106861642294,
"grad_norm": 0.10346360621624287,
"learning_rate": 1.0882661829233619e-05,
"loss": 0.4789,
"step": 435
},
{
"epoch": 3.923509561304837,
"grad_norm": 0.10965908377773818,
"learning_rate": 1.071057260855608e-05,
"loss": 0.4759,
"step": 436
},
{
"epoch": 3.9325084364454446,
"grad_norm": 0.10838534658950197,
"learning_rate": 1.0539644323403514e-05,
"loss": 0.4732,
"step": 437
},
{
"epoch": 3.9415073115860517,
"grad_norm": 0.10164569710162663,
"learning_rate": 1.0369883748804868e-05,
"loss": 0.4777,
"step": 438
},
{
"epoch": 3.9505061867266593,
"grad_norm": 0.11037796996055305,
"learning_rate": 1.0201297613504946e-05,
"loss": 0.4765,
"step": 439
},
{
"epoch": 3.9595050618672665,
"grad_norm": 0.10427806979539742,
"learning_rate": 1.0033892599697638e-05,
"loss": 0.4724,
"step": 440
},
{
"epoch": 3.968503937007874,
"grad_norm": 0.10553954203749775,
"learning_rate": 9.86767534276114e-06,
"loss": 0.4827,
"step": 441
},
{
"epoch": 3.9775028121484812,
"grad_norm": 0.10616804220182864,
"learning_rate": 9.702652430994917e-06,
"loss": 0.4752,
"step": 442
},
{
"epoch": 3.986501687289089,
"grad_norm": 0.10240539280603339,
"learning_rate": 9.538830405358523e-06,
"loss": 0.4754,
"step": 443
},
{
"epoch": 3.9955005624296964,
"grad_norm": 0.10491697645612037,
"learning_rate": 9.376215759212423e-06,
"loss": 0.477,
"step": 444
},
{
"epoch": 4.008998875140607,
"grad_norm": 0.17334062662555766,
"learning_rate": 9.214814938060561e-06,
"loss": 0.4526,
"step": 445
},
{
"epoch": 4.017997750281215,
"grad_norm": 0.13435271730849074,
"learning_rate": 9.054634339294867e-06,
"loss": 0.4497,
"step": 446
},
{
"epoch": 4.026996625421822,
"grad_norm": 0.11872819062021366,
"learning_rate": 8.895680311941745e-06,
"loss": 0.4455,
"step": 447
},
{
"epoch": 4.0359955005624295,
"grad_norm": 0.1319008455279487,
"learning_rate": 8.737959156410385e-06,
"loss": 0.4527,
"step": 448
},
{
"epoch": 4.0449943757030375,
"grad_norm": 0.14206834817022773,
"learning_rate": 8.581477124243002e-06,
"loss": 0.451,
"step": 449
},
{
"epoch": 4.053993250843645,
"grad_norm": 0.14857898517649198,
"learning_rate": 8.426240417867121e-06,
"loss": 0.4511,
"step": 450
},
{
"epoch": 4.062992125984252,
"grad_norm": 0.12827796758216584,
"learning_rate": 8.272255190349678e-06,
"loss": 0.4517,
"step": 451
},
{
"epoch": 4.071991001124859,
"grad_norm": 0.12515412380900934,
"learning_rate": 8.119527545153137e-06,
"loss": 0.4514,
"step": 452
},
{
"epoch": 4.080989876265467,
"grad_norm": 0.13907555587381168,
"learning_rate": 7.968063535893588e-06,
"loss": 0.4465,
"step": 453
},
{
"epoch": 4.089988751406074,
"grad_norm": 0.1388856297974169,
"learning_rate": 7.817869166100812e-06,
"loss": 0.4485,
"step": 454
},
{
"epoch": 4.098987626546681,
"grad_norm": 0.12576078763204848,
"learning_rate": 7.668950388980261e-06,
"loss": 0.4506,
"step": 455
},
{
"epoch": 4.107986501687289,
"grad_norm": 0.12043647761005791,
"learning_rate": 7.521313107177182e-06,
"loss": 0.4507,
"step": 456
},
{
"epoch": 4.116985376827897,
"grad_norm": 0.12541943358764632,
"learning_rate": 7.374963172542564e-06,
"loss": 0.4524,
"step": 457
},
{
"epoch": 4.125984251968504,
"grad_norm": 0.12465374418622191,
"learning_rate": 7.229906385901264e-06,
"loss": 0.4463,
"step": 458
},
{
"epoch": 4.134983127109112,
"grad_norm": 0.1172582061392934,
"learning_rate": 7.086148496822054e-06,
"loss": 0.4541,
"step": 459
},
{
"epoch": 4.143982002249719,
"grad_norm": 0.11434456228260406,
"learning_rate": 6.943695203389689e-06,
"loss": 0.4522,
"step": 460
},
{
"epoch": 4.152980877390326,
"grad_norm": 0.126721520706509,
"learning_rate": 6.802552151979132e-06,
"loss": 0.4496,
"step": 461
},
{
"epoch": 4.161979752530933,
"grad_norm": 0.11357636546475693,
"learning_rate": 6.662724937031697e-06,
"loss": 0.4481,
"step": 462
},
{
"epoch": 4.170978627671541,
"grad_norm": 0.10812363979662522,
"learning_rate": 6.524219100833291e-06,
"loss": 0.45,
"step": 463
},
{
"epoch": 4.179977502812148,
"grad_norm": 0.10736860028431898,
"learning_rate": 6.387040133294786e-06,
"loss": 0.4556,
"step": 464
},
{
"epoch": 4.188976377952756,
"grad_norm": 0.11484073744710532,
"learning_rate": 6.2511934717343955e-06,
"loss": 0.4585,
"step": 465
},
{
"epoch": 4.197975253093364,
"grad_norm": 0.1116733301395841,
"learning_rate": 6.116684500662127e-06,
"loss": 0.4509,
"step": 466
},
{
"epoch": 4.206974128233971,
"grad_norm": 0.10184483968685969,
"learning_rate": 5.983518551566403e-06,
"loss": 0.4465,
"step": 467
},
{
"epoch": 4.215973003374578,
"grad_norm": 0.10648831010048188,
"learning_rate": 5.8517009027027285e-06,
"loss": 0.4529,
"step": 468
},
{
"epoch": 4.224971878515186,
"grad_norm": 0.11078629147369119,
"learning_rate": 5.72123677888444e-06,
"loss": 0.4504,
"step": 469
},
{
"epoch": 4.233970753655793,
"grad_norm": 0.10270056261316057,
"learning_rate": 5.592131351275671e-06,
"loss": 0.4488,
"step": 470
},
{
"epoch": 4.2429696287964,
"grad_norm": 0.0996620878043424,
"learning_rate": 5.464389737186348e-06,
"loss": 0.4479,
"step": 471
},
{
"epoch": 4.251968503937007,
"grad_norm": 0.10061707844415065,
"learning_rate": 5.338016999869351e-06,
"loss": 0.4483,
"step": 472
},
{
"epoch": 4.2609673790776155,
"grad_norm": 0.10888494466945672,
"learning_rate": 5.213018148319835e-06,
"loss": 0.4517,
"step": 473
},
{
"epoch": 4.269966254218223,
"grad_norm": 0.10212521131683365,
"learning_rate": 5.089398137076704e-06,
"loss": 0.4506,
"step": 474
},
{
"epoch": 4.27896512935883,
"grad_norm": 0.10500341792664687,
"learning_rate": 4.967161866026229e-06,
"loss": 0.451,
"step": 475
},
{
"epoch": 4.287964004499438,
"grad_norm": 0.10466244170953495,
"learning_rate": 4.846314180207774e-06,
"loss": 0.4473,
"step": 476
},
{
"epoch": 4.296962879640045,
"grad_norm": 0.09722711168374878,
"learning_rate": 4.726859869621847e-06,
"loss": 0.4481,
"step": 477
},
{
"epoch": 4.305961754780652,
"grad_norm": 0.10204450180802838,
"learning_rate": 4.608803669040187e-06,
"loss": 0.4454,
"step": 478
},
{
"epoch": 4.31496062992126,
"grad_norm": 0.0995565612941236,
"learning_rate": 4.492150257818066e-06,
"loss": 0.4498,
"step": 479
},
{
"epoch": 4.323959505061867,
"grad_norm": 0.10534766800687224,
"learning_rate": 4.376904259708892e-06,
"loss": 0.4528,
"step": 480
},
{
"epoch": 4.3329583802024745,
"grad_norm": 0.09462904906757555,
"learning_rate": 4.263070242680866e-06,
"loss": 0.447,
"step": 481
},
{
"epoch": 4.3419572553430825,
"grad_norm": 0.09711700163085293,
"learning_rate": 4.1506527187359765e-06,
"loss": 0.4552,
"step": 482
},
{
"epoch": 4.35095613048369,
"grad_norm": 0.09726591017318084,
"learning_rate": 4.039656143731128e-06,
"loss": 0.449,
"step": 483
},
{
"epoch": 4.359955005624297,
"grad_norm": 0.10398933972434006,
"learning_rate": 3.930084917201508e-06,
"loss": 0.4541,
"step": 484
},
{
"epoch": 4.368953880764904,
"grad_norm": 0.09827974212844832,
"learning_rate": 3.821943382186275e-06,
"loss": 0.4472,
"step": 485
},
{
"epoch": 4.377952755905512,
"grad_norm": 0.09502494293413481,
"learning_rate": 3.715235825056338e-06,
"loss": 0.4541,
"step": 486
},
{
"epoch": 4.386951631046119,
"grad_norm": 0.09221995391540017,
"learning_rate": 3.609966475344493e-06,
"loss": 0.4507,
"step": 487
},
{
"epoch": 4.395950506186726,
"grad_norm": 0.09866352346696798,
"learning_rate": 3.506139505577779e-06,
"loss": 0.4538,
"step": 488
},
{
"epoch": 4.404949381327334,
"grad_norm": 0.09418594379926497,
"learning_rate": 3.4037590311121015e-06,
"loss": 0.4547,
"step": 489
},
{
"epoch": 4.4139482564679415,
"grad_norm": 0.09399762424115403,
"learning_rate": 3.302829109969072e-06,
"loss": 0.4516,
"step": 490
},
{
"epoch": 4.422947131608549,
"grad_norm": 0.09056395808047757,
"learning_rate": 3.2033537426752236e-06,
"loss": 0.4556,
"step": 491
},
{
"epoch": 4.431946006749157,
"grad_norm": 0.09212248672508358,
"learning_rate": 3.1053368721033974e-06,
"loss": 0.4566,
"step": 492
},
{
"epoch": 4.440944881889764,
"grad_norm": 0.0899827038953071,
"learning_rate": 3.0087823833164596e-06,
"loss": 0.4455,
"step": 493
},
{
"epoch": 4.449943757030371,
"grad_norm": 0.09311934934388971,
"learning_rate": 2.9136941034133424e-06,
"loss": 0.4469,
"step": 494
},
{
"epoch": 4.458942632170979,
"grad_norm": 0.0942657183382691,
"learning_rate": 2.8200758013773313e-06,
"loss": 0.4501,
"step": 495
},
{
"epoch": 4.467941507311586,
"grad_norm": 0.09095763881082566,
"learning_rate": 2.7279311879266645e-06,
"loss": 0.4522,
"step": 496
},
{
"epoch": 4.476940382452193,
"grad_norm": 0.09128227123247035,
"learning_rate": 2.637263915367476e-06,
"loss": 0.4539,
"step": 497
},
{
"epoch": 4.4859392575928005,
"grad_norm": 0.0896975324322128,
"learning_rate": 2.5480775774490195e-06,
"loss": 0.4505,
"step": 498
},
{
"epoch": 4.494938132733409,
"grad_norm": 0.09373464127242032,
"learning_rate": 2.4603757092212057e-06,
"loss": 0.4506,
"step": 499
},
{
"epoch": 4.503937007874016,
"grad_norm": 0.09045392446753084,
"learning_rate": 2.374161786894513e-06,
"loss": 0.4535,
"step": 500
},
{
"epoch": 4.512935883014623,
"grad_norm": 0.09322805283679514,
"learning_rate": 2.2894392277022125e-06,
"loss": 0.4511,
"step": 501
},
{
"epoch": 4.521934758155231,
"grad_norm": 0.09493164625872488,
"learning_rate": 2.206211389764854e-06,
"loss": 0.4531,
"step": 502
},
{
"epoch": 4.530933633295838,
"grad_norm": 0.09141196271877726,
"learning_rate": 2.124481571957242e-06,
"loss": 0.4537,
"step": 503
},
{
"epoch": 4.539932508436445,
"grad_norm": 0.0875486494434062,
"learning_rate": 2.0442530137776374e-06,
"loss": 0.4431,
"step": 504
},
{
"epoch": 4.548931383577052,
"grad_norm": 0.09177205535576874,
"learning_rate": 1.9655288952193442e-06,
"loss": 0.4567,
"step": 505
},
{
"epoch": 4.55793025871766,
"grad_norm": 0.08907941923419668,
"learning_rate": 1.8883123366446955e-06,
"loss": 0.4507,
"step": 506
},
{
"epoch": 4.566929133858268,
"grad_norm": 0.09083887942548946,
"learning_rate": 1.8126063986613652e-06,
"loss": 0.45,
"step": 507
},
{
"epoch": 4.575928008998876,
"grad_norm": 0.08937321221433324,
"learning_rate": 1.7384140820010253e-06,
"loss": 0.4489,
"step": 508
},
{
"epoch": 4.584926884139483,
"grad_norm": 0.08801698232366557,
"learning_rate": 1.6657383274004545e-06,
"loss": 0.4534,
"step": 509
},
{
"epoch": 4.59392575928009,
"grad_norm": 0.08808939037763655,
"learning_rate": 1.5945820154849512e-06,
"loss": 0.4469,
"step": 510
},
{
"epoch": 4.602924634420697,
"grad_norm": 0.0903921066592697,
"learning_rate": 1.524947966654131e-06,
"loss": 0.4535,
"step": 511
},
{
"epoch": 4.611923509561305,
"grad_norm": 0.0899536812400529,
"learning_rate": 1.4568389409702e-06,
"loss": 0.4535,
"step": 512
},
{
"epoch": 4.620922384701912,
"grad_norm": 0.08721249506811954,
"learning_rate": 1.390257638048489e-06,
"loss": 0.4552,
"step": 513
},
{
"epoch": 4.6299212598425195,
"grad_norm": 0.08874618721675122,
"learning_rate": 1.3252066969504874e-06,
"loss": 0.4502,
"step": 514
},
{
"epoch": 4.6389201349831275,
"grad_norm": 0.09036518689753495,
"learning_rate": 1.261688696079233e-06,
"loss": 0.451,
"step": 515
},
{
"epoch": 4.647919010123735,
"grad_norm": 0.08929772055357187,
"learning_rate": 1.1997061530771004e-06,
"loss": 0.4513,
"step": 516
},
{
"epoch": 4.656917885264342,
"grad_norm": 0.08770289069775795,
"learning_rate": 1.1392615247260275e-06,
"loss": 0.4509,
"step": 517
},
{
"epoch": 4.665916760404949,
"grad_norm": 0.0861779113751004,
"learning_rate": 1.080357206850131e-06,
"loss": 0.4497,
"step": 518
},
{
"epoch": 4.674915635545557,
"grad_norm": 0.09018663786133461,
"learning_rate": 1.0229955342207254e-06,
"loss": 0.4556,
"step": 519
},
{
"epoch": 4.683914510686164,
"grad_norm": 0.08621786512649365,
"learning_rate": 9.67178780463809e-07,
"loss": 0.4537,
"step": 520
},
{
"epoch": 4.692913385826771,
"grad_norm": 0.08431669753032205,
"learning_rate": 9.129091579699412e-07,
"loss": 0.4534,
"step": 521
},
{
"epoch": 4.701912260967379,
"grad_norm": 0.08666217491427604,
"learning_rate": 8.601888178065177e-07,
"loss": 0.4501,
"step": 522
},
{
"epoch": 4.7109111361079865,
"grad_norm": 0.08654873368758471,
"learning_rate": 8.090198496325485e-07,
"loss": 0.4506,
"step": 523
},
{
"epoch": 4.719910011248594,
"grad_norm": 0.08652439676904433,
"learning_rate": 7.594042816158187e-07,
"loss": 0.4529,
"step": 524
},
{
"epoch": 4.728908886389202,
"grad_norm": 0.08831640949919531,
"learning_rate": 7.113440803524896e-07,
"loss": 0.4531,
"step": 525
},
{
"epoch": 4.737907761529809,
"grad_norm": 0.0866028677777177,
"learning_rate": 6.648411507891528e-07,
"loss": 0.4522,
"step": 526
},
{
"epoch": 4.746906636670416,
"grad_norm": 0.08731496032068242,
"learning_rate": 6.198973361473349e-07,
"loss": 0.4495,
"step": 527
},
{
"epoch": 4.755905511811024,
"grad_norm": 0.08758993950549104,
"learning_rate": 5.765144178504222e-07,
"loss": 0.4526,
"step": 528
},
{
"epoch": 4.764904386951631,
"grad_norm": 0.08831218479369252,
"learning_rate": 5.34694115453065e-07,
"loss": 0.4532,
"step": 529
},
{
"epoch": 4.773903262092238,
"grad_norm": 0.0860726407842809,
"learning_rate": 4.944380865730125e-07,
"loss": 0.4529,
"step": 530
},
{
"epoch": 4.7829021372328455,
"grad_norm": 0.0840695127911042,
"learning_rate": 4.5574792682541167e-07,
"loss": 0.4502,
"step": 531
},
{
"epoch": 4.791901012373454,
"grad_norm": 0.0827762260768475,
"learning_rate": 4.186251697595678e-07,
"loss": 0.4471,
"step": 532
},
{
"epoch": 4.800899887514061,
"grad_norm": 0.08825259092439537,
"learning_rate": 3.83071286798149e-07,
"loss": 0.4434,
"step": 533
},
{
"epoch": 4.809898762654668,
"grad_norm": 0.08855814892569033,
"learning_rate": 3.4908768717887286e-07,
"loss": 0.45,
"step": 534
},
{
"epoch": 4.818897637795276,
"grad_norm": 0.08776896957164552,
"learning_rate": 3.1667571789864015e-07,
"loss": 0.4539,
"step": 535
},
{
"epoch": 4.827896512935883,
"grad_norm": 0.0848390478817017,
"learning_rate": 2.858366636601639e-07,
"loss": 0.451,
"step": 536
},
{
"epoch": 4.83689538807649,
"grad_norm": 0.0832820856667789,
"learning_rate": 2.5657174682101936e-07,
"loss": 0.4501,
"step": 537
},
{
"epoch": 4.845894263217097,
"grad_norm": 0.08506972205806178,
"learning_rate": 2.288821273452113e-07,
"loss": 0.4495,
"step": 538
},
{
"epoch": 4.854893138357705,
"grad_norm": 0.08489253043277004,
"learning_rate": 2.027689027572066e-07,
"loss": 0.4527,
"step": 539
},
{
"epoch": 4.863892013498313,
"grad_norm": 0.08426467385333178,
"learning_rate": 1.7823310809840456e-07,
"loss": 0.4464,
"step": 540
},
{
"epoch": 4.872890888638921,
"grad_norm": 0.08767870830184761,
"learning_rate": 1.55275715886134e-07,
"loss": 0.4519,
"step": 541
},
{
"epoch": 4.881889763779528,
"grad_norm": 0.08723387825162499,
"learning_rate": 1.3389763607509765e-07,
"loss": 0.455,
"step": 542
},
{
"epoch": 4.890888638920135,
"grad_norm": 0.08523501703956923,
"learning_rate": 1.1409971602130754e-07,
"loss": 0.4502,
"step": 543
},
{
"epoch": 4.899887514060742,
"grad_norm": 0.08495566625522424,
"learning_rate": 9.588274044848523e-08,
"loss": 0.4541,
"step": 544
},
{
"epoch": 4.90888638920135,
"grad_norm": 0.08415284728722193,
"learning_rate": 7.924743141698888e-08,
"loss": 0.4534,
"step": 545
},
{
"epoch": 4.917885264341957,
"grad_norm": 0.0899650650008716,
"learning_rate": 6.419444829515175e-08,
"loss": 0.4514,
"step": 546
},
{
"epoch": 4.926884139482564,
"grad_norm": 0.08441613482621164,
"learning_rate": 5.072438773318755e-08,
"loss": 0.4483,
"step": 547
},
{
"epoch": 4.9358830146231725,
"grad_norm": 0.08443189586686374,
"learning_rate": 3.8837783639507076e-08,
"loss": 0.453,
"step": 548
},
{
"epoch": 4.94488188976378,
"grad_norm": 0.08693922903585893,
"learning_rate": 2.8535107159584076e-08,
"loss": 0.4519,
"step": 549
},
{
"epoch": 4.953880764904387,
"grad_norm": 0.0869995258137961,
"learning_rate": 1.9816766657254626e-08,
"loss": 0.4498,
"step": 550
},
{
"epoch": 4.962879640044994,
"grad_norm": 0.08843911191962896,
"learning_rate": 1.268310769855674e-08,
"loss": 0.4495,
"step": 551
},
{
"epoch": 4.971878515185602,
"grad_norm": 0.08550795511559632,
"learning_rate": 7.134413038012433e-09,
"loss": 0.4487,
"step": 552
},
{
"epoch": 4.980877390326209,
"grad_norm": 0.08541041297147574,
"learning_rate": 3.170902607432247e-09,
"loss": 0.4502,
"step": 553
},
{
"epoch": 4.989876265466816,
"grad_norm": 0.08600009481079209,
"learning_rate": 7.92733507188892e-10,
"loss": 0.4553,
"step": 554
},
{
"epoch": 4.998875140607424,
"grad_norm": 0.0856629080592209,
"learning_rate": 0.0,
"loss": 0.4591,
"step": 555
},
{
"epoch": 4.998875140607424,
"step": 555,
"total_flos": 2.8638155373837025e+18,
"train_loss": 0.09022162760700192,
"train_runtime": 11331.6206,
"train_samples_per_second": 25.101,
"train_steps_per_second": 0.049
}
],
"logging_steps": 1,
"max_steps": 555,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.8638155373837025e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}