nemo_nano_code_100k / trainer_state.json
ryanmarten's picture
Upload model
c5ced6a verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.956190476190476,
"eval_steps": 500,
"global_step": 325,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.015238095238095238,
"grad_norm": 6.41825703019793,
"learning_rate": 2.4242424242424244e-06,
"loss": 1.0212,
"step": 1
},
{
"epoch": 0.030476190476190476,
"grad_norm": 6.32460329298527,
"learning_rate": 4.848484848484849e-06,
"loss": 1.0078,
"step": 2
},
{
"epoch": 0.045714285714285714,
"grad_norm": 5.823493836887707,
"learning_rate": 7.272727272727273e-06,
"loss": 0.9926,
"step": 3
},
{
"epoch": 0.06095238095238095,
"grad_norm": 4.215092249626398,
"learning_rate": 9.696969696969698e-06,
"loss": 0.9511,
"step": 4
},
{
"epoch": 0.0761904761904762,
"grad_norm": 2.755821783551069,
"learning_rate": 1.2121212121212122e-05,
"loss": 0.924,
"step": 5
},
{
"epoch": 0.09142857142857143,
"grad_norm": 5.79854862437273,
"learning_rate": 1.4545454545454546e-05,
"loss": 0.9436,
"step": 6
},
{
"epoch": 0.10666666666666667,
"grad_norm": 6.970153409745764,
"learning_rate": 1.6969696969696972e-05,
"loss": 0.9199,
"step": 7
},
{
"epoch": 0.1219047619047619,
"grad_norm": 6.398045150850651,
"learning_rate": 1.9393939393939395e-05,
"loss": 0.9368,
"step": 8
},
{
"epoch": 0.13714285714285715,
"grad_norm": 3.7323118823746717,
"learning_rate": 2.1818181818181818e-05,
"loss": 0.8624,
"step": 9
},
{
"epoch": 0.1523809523809524,
"grad_norm": 2.4649345087328935,
"learning_rate": 2.4242424242424244e-05,
"loss": 0.8177,
"step": 10
},
{
"epoch": 0.1676190476190476,
"grad_norm": 2.624350847053939,
"learning_rate": 2.6666666666666667e-05,
"loss": 0.7966,
"step": 11
},
{
"epoch": 0.18285714285714286,
"grad_norm": 1.495708592085656,
"learning_rate": 2.9090909090909093e-05,
"loss": 0.7813,
"step": 12
},
{
"epoch": 0.1980952380952381,
"grad_norm": 1.4871227985381805,
"learning_rate": 3.151515151515152e-05,
"loss": 0.7453,
"step": 13
},
{
"epoch": 0.21333333333333335,
"grad_norm": 2.060287535099846,
"learning_rate": 3.3939393939393945e-05,
"loss": 0.7371,
"step": 14
},
{
"epoch": 0.22857142857142856,
"grad_norm": 1.4326857447271233,
"learning_rate": 3.6363636363636364e-05,
"loss": 0.7234,
"step": 15
},
{
"epoch": 0.2438095238095238,
"grad_norm": 2.1165784341208105,
"learning_rate": 3.878787878787879e-05,
"loss": 0.7143,
"step": 16
},
{
"epoch": 0.259047619047619,
"grad_norm": 1.7376117944761271,
"learning_rate": 4.1212121212121216e-05,
"loss": 0.7091,
"step": 17
},
{
"epoch": 0.2742857142857143,
"grad_norm": 1.9235494481574669,
"learning_rate": 4.3636363636363636e-05,
"loss": 0.7033,
"step": 18
},
{
"epoch": 0.2895238095238095,
"grad_norm": 1.2885542410242774,
"learning_rate": 4.606060606060607e-05,
"loss": 0.693,
"step": 19
},
{
"epoch": 0.3047619047619048,
"grad_norm": 2.0682542365556746,
"learning_rate": 4.848484848484849e-05,
"loss": 0.6782,
"step": 20
},
{
"epoch": 0.32,
"grad_norm": 1.1294865223043438,
"learning_rate": 5.0909090909090914e-05,
"loss": 0.6734,
"step": 21
},
{
"epoch": 0.3352380952380952,
"grad_norm": 1.6892295840685134,
"learning_rate": 5.333333333333333e-05,
"loss": 0.6701,
"step": 22
},
{
"epoch": 0.3504761904761905,
"grad_norm": 1.7015145087043497,
"learning_rate": 5.5757575757575766e-05,
"loss": 0.6747,
"step": 23
},
{
"epoch": 0.3657142857142857,
"grad_norm": 1.105513245543281,
"learning_rate": 5.8181818181818185e-05,
"loss": 0.6569,
"step": 24
},
{
"epoch": 0.38095238095238093,
"grad_norm": 1.9301412200235886,
"learning_rate": 6.060606060606061e-05,
"loss": 0.6672,
"step": 25
},
{
"epoch": 0.3961904761904762,
"grad_norm": 1.5372986311163992,
"learning_rate": 6.303030303030304e-05,
"loss": 0.6575,
"step": 26
},
{
"epoch": 0.4114285714285714,
"grad_norm": 1.841770844497413,
"learning_rate": 6.545454545454546e-05,
"loss": 0.6439,
"step": 27
},
{
"epoch": 0.4266666666666667,
"grad_norm": 1.285994127925713,
"learning_rate": 6.787878787878789e-05,
"loss": 0.6421,
"step": 28
},
{
"epoch": 0.4419047619047619,
"grad_norm": 1.306641316576376,
"learning_rate": 7.03030303030303e-05,
"loss": 0.6446,
"step": 29
},
{
"epoch": 0.45714285714285713,
"grad_norm": 2.060289337588267,
"learning_rate": 7.272727272727273e-05,
"loss": 0.6418,
"step": 30
},
{
"epoch": 0.4723809523809524,
"grad_norm": 1.5995121846928928,
"learning_rate": 7.515151515151517e-05,
"loss": 0.6318,
"step": 31
},
{
"epoch": 0.4876190476190476,
"grad_norm": 1.5361156868763157,
"learning_rate": 7.757575757575758e-05,
"loss": 0.6356,
"step": 32
},
{
"epoch": 0.5028571428571429,
"grad_norm": 1.9790958633956652,
"learning_rate": 8e-05,
"loss": 0.6315,
"step": 33
},
{
"epoch": 0.518095238095238,
"grad_norm": 1.0983212679170282,
"learning_rate": 7.999768495280586e-05,
"loss": 0.6212,
"step": 34
},
{
"epoch": 0.5333333333333333,
"grad_norm": 2.356052428059575,
"learning_rate": 7.999074007919565e-05,
"loss": 0.6366,
"step": 35
},
{
"epoch": 0.5485714285714286,
"grad_norm": 1.9118106657858278,
"learning_rate": 7.997916618305483e-05,
"loss": 0.6337,
"step": 36
},
{
"epoch": 0.5638095238095238,
"grad_norm": 2.086875421088417,
"learning_rate": 7.996296460408921e-05,
"loss": 0.6243,
"step": 37
},
{
"epoch": 0.579047619047619,
"grad_norm": 1.6815412319889804,
"learning_rate": 7.994213721766979e-05,
"loss": 0.6222,
"step": 38
},
{
"epoch": 0.5942857142857143,
"grad_norm": 1.9619509595749545,
"learning_rate": 7.99166864346157e-05,
"loss": 0.6185,
"step": 39
},
{
"epoch": 0.6095238095238096,
"grad_norm": 1.1850043703613138,
"learning_rate": 7.988661520091513e-05,
"loss": 0.6222,
"step": 40
},
{
"epoch": 0.6247619047619047,
"grad_norm": 1.5413225249585591,
"learning_rate": 7.985192699738432e-05,
"loss": 0.6131,
"step": 41
},
{
"epoch": 0.64,
"grad_norm": 2.2994933443512697,
"learning_rate": 7.981262583926472e-05,
"loss": 0.6326,
"step": 42
},
{
"epoch": 0.6552380952380953,
"grad_norm": 0.8610367129145644,
"learning_rate": 7.976871627575808e-05,
"loss": 0.6057,
"step": 43
},
{
"epoch": 0.6704761904761904,
"grad_norm": 2.630926210907478,
"learning_rate": 7.972020338950004e-05,
"loss": 0.6366,
"step": 44
},
{
"epoch": 0.6857142857142857,
"grad_norm": 1.542557855004854,
"learning_rate": 7.96670927959716e-05,
"loss": 0.624,
"step": 45
},
{
"epoch": 0.700952380952381,
"grad_norm": 1.9643148826366752,
"learning_rate": 7.960939064284934e-05,
"loss": 0.619,
"step": 46
},
{
"epoch": 0.7161904761904762,
"grad_norm": 1.248523269079334,
"learning_rate": 7.954710360929362e-05,
"loss": 0.6108,
"step": 47
},
{
"epoch": 0.7314285714285714,
"grad_norm": 1.225316205009982,
"learning_rate": 7.948023890517557e-05,
"loss": 0.6126,
"step": 48
},
{
"epoch": 0.7466666666666667,
"grad_norm": 1.4200551325493893,
"learning_rate": 7.940880427024243e-05,
"loss": 0.6093,
"step": 49
},
{
"epoch": 0.7619047619047619,
"grad_norm": 1.3484644010507392,
"learning_rate": 7.933280797322181e-05,
"loss": 0.5995,
"step": 50
},
{
"epoch": 0.7771428571428571,
"grad_norm": 1.317639647361387,
"learning_rate": 7.925225881086437e-05,
"loss": 0.5967,
"step": 51
},
{
"epoch": 0.7923809523809524,
"grad_norm": 1.5340582175990225,
"learning_rate": 7.916716610692578e-05,
"loss": 0.5981,
"step": 52
},
{
"epoch": 0.8076190476190476,
"grad_norm": 1.0386828999827622,
"learning_rate": 7.907753971108728e-05,
"loss": 0.5896,
"step": 53
},
{
"epoch": 0.8228571428571428,
"grad_norm": 1.5332648656890953,
"learning_rate": 7.898338999781567e-05,
"loss": 0.589,
"step": 54
},
{
"epoch": 0.8380952380952381,
"grad_norm": 1.1339766871931323,
"learning_rate": 7.888472786516246e-05,
"loss": 0.5937,
"step": 55
},
{
"epoch": 0.8533333333333334,
"grad_norm": 1.5164034405411189,
"learning_rate": 7.87815647335023e-05,
"loss": 0.5883,
"step": 56
},
{
"epoch": 0.8685714285714285,
"grad_norm": 0.9700436403678562,
"learning_rate": 7.86739125442111e-05,
"loss": 0.5931,
"step": 57
},
{
"epoch": 0.8838095238095238,
"grad_norm": 1.3463613799484717,
"learning_rate": 7.856178375828384e-05,
"loss": 0.589,
"step": 58
},
{
"epoch": 0.8990476190476191,
"grad_norm": 0.8752856535881799,
"learning_rate": 7.844519135489204e-05,
"loss": 0.5754,
"step": 59
},
{
"epoch": 0.9142857142857143,
"grad_norm": 1.1854440512880053,
"learning_rate": 7.832414882988153e-05,
"loss": 0.5894,
"step": 60
},
{
"epoch": 0.9295238095238095,
"grad_norm": 1.7051861411911757,
"learning_rate": 7.819867019421023e-05,
"loss": 0.5885,
"step": 61
},
{
"epoch": 0.9447619047619048,
"grad_norm": 0.6380913241526577,
"learning_rate": 7.806876997232625e-05,
"loss": 0.5834,
"step": 62
},
{
"epoch": 0.96,
"grad_norm": 1.3953430391233206,
"learning_rate": 7.793446320048687e-05,
"loss": 0.5783,
"step": 63
},
{
"epoch": 0.9752380952380952,
"grad_norm": 1.0079170376946163,
"learning_rate": 7.779576542501781e-05,
"loss": 0.5839,
"step": 64
},
{
"epoch": 0.9904761904761905,
"grad_norm": 1.1878161560623453,
"learning_rate": 7.765269270051389e-05,
"loss": 0.575,
"step": 65
},
{
"epoch": 1.0066666666666666,
"grad_norm": 1.495861192796852,
"learning_rate": 7.750526158798056e-05,
"loss": 0.8303,
"step": 66
},
{
"epoch": 1.0219047619047619,
"grad_norm": 1.3683956168224771,
"learning_rate": 7.735348915291701e-05,
"loss": 0.571,
"step": 67
},
{
"epoch": 1.0371428571428571,
"grad_norm": 0.8831683489698775,
"learning_rate": 7.719739296334073e-05,
"loss": 0.5548,
"step": 68
},
{
"epoch": 1.0523809523809524,
"grad_norm": 0.9884633873935289,
"learning_rate": 7.7036991087754e-05,
"loss": 0.5619,
"step": 69
},
{
"epoch": 1.0676190476190477,
"grad_norm": 1.658491152165491,
"learning_rate": 7.687230209305242e-05,
"loss": 0.5876,
"step": 70
},
{
"epoch": 1.0828571428571427,
"grad_norm": 0.8352075307968011,
"learning_rate": 7.670334504237575e-05,
"loss": 0.5685,
"step": 71
},
{
"epoch": 1.098095238095238,
"grad_norm": 1.429719144977111,
"learning_rate": 7.65301394929013e-05,
"loss": 0.5764,
"step": 72
},
{
"epoch": 1.1133333333333333,
"grad_norm": 0.9027318597099093,
"learning_rate": 7.635270549358012e-05,
"loss": 0.5602,
"step": 73
},
{
"epoch": 1.1285714285714286,
"grad_norm": 1.0536650805805048,
"learning_rate": 7.617106358281634e-05,
"loss": 0.5765,
"step": 74
},
{
"epoch": 1.1438095238095238,
"grad_norm": 0.8336607908560256,
"learning_rate": 7.598523478608974e-05,
"loss": 0.5583,
"step": 75
},
{
"epoch": 1.159047619047619,
"grad_norm": 0.7166113488943966,
"learning_rate": 7.579524061352206e-05,
"loss": 0.551,
"step": 76
},
{
"epoch": 1.1742857142857144,
"grad_norm": 0.8105369080449557,
"learning_rate": 7.560110305738708e-05,
"loss": 0.5659,
"step": 77
},
{
"epoch": 1.1895238095238094,
"grad_norm": 0.6283210931425559,
"learning_rate": 7.540284458956503e-05,
"loss": 0.5379,
"step": 78
},
{
"epoch": 1.2047619047619047,
"grad_norm": 0.8126105033449552,
"learning_rate": 7.520048815894144e-05,
"loss": 0.5476,
"step": 79
},
{
"epoch": 1.22,
"grad_norm": 0.5252506521520861,
"learning_rate": 7.49940571887506e-05,
"loss": 0.5551,
"step": 80
},
{
"epoch": 1.2352380952380952,
"grad_norm": 0.6609919242663752,
"learning_rate": 7.478357557386444e-05,
"loss": 0.5407,
"step": 81
},
{
"epoch": 1.2504761904761905,
"grad_norm": 0.576238102776681,
"learning_rate": 7.456906767802658e-05,
"loss": 0.5465,
"step": 82
},
{
"epoch": 1.2657142857142858,
"grad_norm": 0.5233988001346253,
"learning_rate": 7.435055833103213e-05,
"loss": 0.55,
"step": 83
},
{
"epoch": 1.2809523809523808,
"grad_norm": 0.44023553757218015,
"learning_rate": 7.412807282585362e-05,
"loss": 0.5412,
"step": 84
},
{
"epoch": 1.2961904761904761,
"grad_norm": 0.4831178702636027,
"learning_rate": 7.390163691571325e-05,
"loss": 0.5398,
"step": 85
},
{
"epoch": 1.3114285714285714,
"grad_norm": 0.454634628403172,
"learning_rate": 7.367127681110199e-05,
"loss": 0.534,
"step": 86
},
{
"epoch": 1.3266666666666667,
"grad_norm": 0.44419681534445155,
"learning_rate": 7.343701917674547e-05,
"loss": 0.5421,
"step": 87
},
{
"epoch": 1.341904761904762,
"grad_norm": 0.4888479476988804,
"learning_rate": 7.319889112851771e-05,
"loss": 0.5371,
"step": 88
},
{
"epoch": 1.3571428571428572,
"grad_norm": 0.7143397739870562,
"learning_rate": 7.295692023030217e-05,
"loss": 0.5377,
"step": 89
},
{
"epoch": 1.3723809523809525,
"grad_norm": 0.9465106459896011,
"learning_rate": 7.271113449080131e-05,
"loss": 0.5479,
"step": 90
},
{
"epoch": 1.3876190476190478,
"grad_norm": 1.2678262246780991,
"learning_rate": 7.246156236029445e-05,
"loss": 0.5452,
"step": 91
},
{
"epoch": 1.4028571428571428,
"grad_norm": 0.6281248994163343,
"learning_rate": 7.220823272734461e-05,
"loss": 0.5401,
"step": 92
},
{
"epoch": 1.418095238095238,
"grad_norm": 0.34247752257094766,
"learning_rate": 7.19511749154546e-05,
"loss": 0.5394,
"step": 93
},
{
"epoch": 1.4333333333333333,
"grad_norm": 0.7014077226631056,
"learning_rate": 7.169041867967272e-05,
"loss": 0.5407,
"step": 94
},
{
"epoch": 1.4485714285714286,
"grad_norm": 0.9221215435654039,
"learning_rate": 7.142599420314857e-05,
"loss": 0.5464,
"step": 95
},
{
"epoch": 1.4638095238095237,
"grad_norm": 0.8724875653280793,
"learning_rate": 7.115793209363929e-05,
"loss": 0.5379,
"step": 96
},
{
"epoch": 1.479047619047619,
"grad_norm": 0.734302784896464,
"learning_rate": 7.088626337996657e-05,
"loss": 0.5248,
"step": 97
},
{
"epoch": 1.4942857142857142,
"grad_norm": 0.6804548302488894,
"learning_rate": 7.06110195084251e-05,
"loss": 0.5403,
"step": 98
},
{
"epoch": 1.5095238095238095,
"grad_norm": 0.8180102079360684,
"learning_rate": 7.03322323391425e-05,
"loss": 0.5356,
"step": 99
},
{
"epoch": 1.5247619047619048,
"grad_norm": 0.7089574204038593,
"learning_rate": 7.004993414239147e-05,
"loss": 0.5367,
"step": 100
},
{
"epoch": 1.54,
"grad_norm": 0.4534978500691297,
"learning_rate": 6.976415759485443e-05,
"loss": 0.5301,
"step": 101
},
{
"epoch": 1.5552380952380953,
"grad_norm": 0.5791842384826449,
"learning_rate": 6.947493577584111e-05,
"loss": 0.5338,
"step": 102
},
{
"epoch": 1.5704761904761906,
"grad_norm": 0.6373667696236239,
"learning_rate": 6.918230216345951e-05,
"loss": 0.5255,
"step": 103
},
{
"epoch": 1.5857142857142859,
"grad_norm": 0.5813750117468534,
"learning_rate": 6.888629063074082e-05,
"loss": 0.5209,
"step": 104
},
{
"epoch": 1.600952380952381,
"grad_norm": 0.5731639034091425,
"learning_rate": 6.858693544171845e-05,
"loss": 0.5215,
"step": 105
},
{
"epoch": 1.6161904761904762,
"grad_norm": 0.5351450945596543,
"learning_rate": 6.828427124746191e-05,
"loss": 0.5301,
"step": 106
},
{
"epoch": 1.6314285714285715,
"grad_norm": 0.6016816279659157,
"learning_rate": 6.797833308206588e-05,
"loss": 0.529,
"step": 107
},
{
"epoch": 1.6466666666666665,
"grad_norm": 0.7659521464895882,
"learning_rate": 6.766915635859497e-05,
"loss": 0.533,
"step": 108
},
{
"epoch": 1.6619047619047618,
"grad_norm": 0.9630524913355004,
"learning_rate": 6.735677686498443e-05,
"loss": 0.5297,
"step": 109
},
{
"epoch": 1.677142857142857,
"grad_norm": 1.1492200252998914,
"learning_rate": 6.70412307598978e-05,
"loss": 0.5264,
"step": 110
},
{
"epoch": 1.6923809523809523,
"grad_norm": 0.5857314920473611,
"learning_rate": 6.672255456854135e-05,
"loss": 0.5241,
"step": 111
},
{
"epoch": 1.7076190476190476,
"grad_norm": 0.48173626706101424,
"learning_rate": 6.640078517843619e-05,
"loss": 0.5346,
"step": 112
},
{
"epoch": 1.7228571428571429,
"grad_norm": 0.7730414410041352,
"learning_rate": 6.60759598351485e-05,
"loss": 0.5229,
"step": 113
},
{
"epoch": 1.7380952380952381,
"grad_norm": 0.6363250099986629,
"learning_rate": 6.57481161379783e-05,
"loss": 0.5243,
"step": 114
},
{
"epoch": 1.7533333333333334,
"grad_norm": 0.45624023704141803,
"learning_rate": 6.54172920356071e-05,
"loss": 0.5247,
"step": 115
},
{
"epoch": 1.7685714285714287,
"grad_norm": 0.4715396085670591,
"learning_rate": 6.508352582170543e-05,
"loss": 0.5144,
"step": 116
},
{
"epoch": 1.783809523809524,
"grad_norm": 0.5945001964397519,
"learning_rate": 6.474685613050014e-05,
"loss": 0.5279,
"step": 117
},
{
"epoch": 1.799047619047619,
"grad_norm": 0.5588842197173406,
"learning_rate": 6.440732193230241e-05,
"loss": 0.522,
"step": 118
},
{
"epoch": 1.8142857142857143,
"grad_norm": 0.36397509002448036,
"learning_rate": 6.40649625289969e-05,
"loss": 0.5206,
"step": 119
},
{
"epoch": 1.8295238095238096,
"grad_norm": 0.5225246500798869,
"learning_rate": 6.371981754949241e-05,
"loss": 0.5233,
"step": 120
},
{
"epoch": 1.8447619047619046,
"grad_norm": 0.6419719236551851,
"learning_rate": 6.337192694513474e-05,
"loss": 0.5265,
"step": 121
},
{
"epoch": 1.8599999999999999,
"grad_norm": 0.41672473350375433,
"learning_rate": 6.30213309850823e-05,
"loss": 0.5165,
"step": 122
},
{
"epoch": 1.8752380952380951,
"grad_norm": 0.29619919377664744,
"learning_rate": 6.266807025164472e-05,
"loss": 0.5158,
"step": 123
},
{
"epoch": 1.8904761904761904,
"grad_norm": 0.41731867813128104,
"learning_rate": 6.231218563558551e-05,
"loss": 0.518,
"step": 124
},
{
"epoch": 1.9057142857142857,
"grad_norm": 0.3648760435140768,
"learning_rate": 6.195371833138878e-05,
"loss": 0.5165,
"step": 125
},
{
"epoch": 1.920952380952381,
"grad_norm": 0.3531467699123105,
"learning_rate": 6.159270983249084e-05,
"loss": 0.5247,
"step": 126
},
{
"epoch": 1.9361904761904762,
"grad_norm": 0.45588003744063627,
"learning_rate": 6.122920192647734e-05,
"loss": 0.5197,
"step": 127
},
{
"epoch": 1.9514285714285715,
"grad_norm": 0.5510619573466758,
"learning_rate": 6.086323669024616e-05,
"loss": 0.5256,
"step": 128
},
{
"epoch": 1.9666666666666668,
"grad_norm": 0.5646853258922581,
"learning_rate": 6.049485648513696e-05,
"loss": 0.527,
"step": 129
},
{
"epoch": 1.981904761904762,
"grad_norm": 0.4479646401937566,
"learning_rate": 6.012410395202774e-05,
"loss": 0.5234,
"step": 130
},
{
"epoch": 1.997142857142857,
"grad_norm": 0.575986791359234,
"learning_rate": 5.9751022006399095e-05,
"loss": 0.7009,
"step": 131
},
{
"epoch": 2.013333333333333,
"grad_norm": 0.814219365805337,
"learning_rate": 5.9375653833366585e-05,
"loss": 0.5404,
"step": 132
},
{
"epoch": 2.0285714285714285,
"grad_norm": 0.8146199046124171,
"learning_rate": 5.8998042882682024e-05,
"loss": 0.4877,
"step": 133
},
{
"epoch": 2.0438095238095237,
"grad_norm": 0.7629490780519704,
"learning_rate": 5.861823286370396e-05,
"loss": 0.4849,
"step": 134
},
{
"epoch": 2.059047619047619,
"grad_norm": 0.6570978207497692,
"learning_rate": 5.823626774033836e-05,
"loss": 0.4887,
"step": 135
},
{
"epoch": 2.0742857142857143,
"grad_norm": 0.5305731216336181,
"learning_rate": 5.7852191725949564e-05,
"loss": 0.4881,
"step": 136
},
{
"epoch": 2.0895238095238096,
"grad_norm": 0.512493977688762,
"learning_rate": 5.746604927824257e-05,
"loss": 0.4844,
"step": 137
},
{
"epoch": 2.104761904761905,
"grad_norm": 0.4828622337949408,
"learning_rate": 5.707788509411685e-05,
"loss": 0.4887,
"step": 138
},
{
"epoch": 2.12,
"grad_norm": 0.4884027438000281,
"learning_rate": 5.668774410449268e-05,
"loss": 0.4908,
"step": 139
},
{
"epoch": 2.1352380952380954,
"grad_norm": 0.42776751878992564,
"learning_rate": 5.629567146911023e-05,
"loss": 0.4872,
"step": 140
},
{
"epoch": 2.1504761904761907,
"grad_norm": 0.36115257273207996,
"learning_rate": 5.590171257130225e-05,
"loss": 0.4861,
"step": 141
},
{
"epoch": 2.1657142857142855,
"grad_norm": 0.38436733504559367,
"learning_rate": 5.550591301274073e-05,
"loss": 0.4844,
"step": 142
},
{
"epoch": 2.1809523809523808,
"grad_norm": 0.3435583779900254,
"learning_rate": 5.5108318608158595e-05,
"loss": 0.4836,
"step": 143
},
{
"epoch": 2.196190476190476,
"grad_norm": 0.2857883944660014,
"learning_rate": 5.4708975380046356e-05,
"loss": 0.4796,
"step": 144
},
{
"epoch": 2.2114285714285713,
"grad_norm": 0.31512140613431383,
"learning_rate": 5.430792955332503e-05,
"loss": 0.4863,
"step": 145
},
{
"epoch": 2.2266666666666666,
"grad_norm": 0.31769228977835245,
"learning_rate": 5.3905227549995366e-05,
"loss": 0.4809,
"step": 146
},
{
"epoch": 2.241904761904762,
"grad_norm": 0.27924100773876076,
"learning_rate": 5.350091598376454e-05,
"loss": 0.4805,
"step": 147
},
{
"epoch": 2.257142857142857,
"grad_norm": 0.2662485345450648,
"learning_rate": 5.3095041654650384e-05,
"loss": 0.4831,
"step": 148
},
{
"epoch": 2.2723809523809524,
"grad_norm": 0.23461248952656025,
"learning_rate": 5.268765154356425e-05,
"loss": 0.4816,
"step": 149
},
{
"epoch": 2.2876190476190477,
"grad_norm": 0.240243576104443,
"learning_rate": 5.227879280687281e-05,
"loss": 0.4788,
"step": 150
},
{
"epoch": 2.302857142857143,
"grad_norm": 0.24392104367270967,
"learning_rate": 5.186851277093961e-05,
"loss": 0.4851,
"step": 151
},
{
"epoch": 2.318095238095238,
"grad_norm": 0.19828908319240954,
"learning_rate": 5.1456858926646974e-05,
"loss": 0.4817,
"step": 152
},
{
"epoch": 2.3333333333333335,
"grad_norm": 0.2354905746190516,
"learning_rate": 5.104387892389877e-05,
"loss": 0.4816,
"step": 153
},
{
"epoch": 2.3485714285714288,
"grad_norm": 0.24408900516772156,
"learning_rate": 5.062962056610478e-05,
"loss": 0.4833,
"step": 154
},
{
"epoch": 2.3638095238095236,
"grad_norm": 0.22588988217182931,
"learning_rate": 5.021413180464748e-05,
"loss": 0.4884,
"step": 155
},
{
"epoch": 2.379047619047619,
"grad_norm": 0.20985654308992252,
"learning_rate": 4.979746073333145e-05,
"loss": 0.4777,
"step": 156
},
{
"epoch": 2.394285714285714,
"grad_norm": 0.24110384355800776,
"learning_rate": 4.9379655582816396e-05,
"loss": 0.4867,
"step": 157
},
{
"epoch": 2.4095238095238094,
"grad_norm": 0.20401637758198887,
"learning_rate": 4.896076471503439e-05,
"loss": 0.4766,
"step": 158
},
{
"epoch": 2.4247619047619047,
"grad_norm": 0.19932449853927872,
"learning_rate": 4.854083661759185e-05,
"loss": 0.4884,
"step": 159
},
{
"epoch": 2.44,
"grad_norm": 0.24533376047455172,
"learning_rate": 4.8119919898156934e-05,
"loss": 0.4834,
"step": 160
},
{
"epoch": 2.455238095238095,
"grad_norm": 0.19336383545060395,
"learning_rate": 4.76980632788332e-05,
"loss": 0.4779,
"step": 161
},
{
"epoch": 2.4704761904761905,
"grad_norm": 0.19824370981981676,
"learning_rate": 4.7275315590519765e-05,
"loss": 0.4823,
"step": 162
},
{
"epoch": 2.4857142857142858,
"grad_norm": 0.20431330071012052,
"learning_rate": 4.6851725767259106e-05,
"loss": 0.4781,
"step": 163
},
{
"epoch": 2.500952380952381,
"grad_norm": 0.2082524151757895,
"learning_rate": 4.642734284057283e-05,
"loss": 0.4755,
"step": 164
},
{
"epoch": 2.5161904761904763,
"grad_norm": 0.2129960852067495,
"learning_rate": 4.600221593378611e-05,
"loss": 0.4796,
"step": 165
},
{
"epoch": 2.5314285714285716,
"grad_norm": 0.17147652502249622,
"learning_rate": 4.5576394256341596e-05,
"loss": 0.4784,
"step": 166
},
{
"epoch": 2.546666666666667,
"grad_norm": 0.2328806782705143,
"learning_rate": 4.514992709810324e-05,
"loss": 0.4839,
"step": 167
},
{
"epoch": 2.5619047619047617,
"grad_norm": 0.2628457128916824,
"learning_rate": 4.472286382365096e-05,
"loss": 0.4899,
"step": 168
},
{
"epoch": 2.5771428571428574,
"grad_norm": 0.20530497111629914,
"learning_rate": 4.429525386656653e-05,
"loss": 0.4865,
"step": 169
},
{
"epoch": 2.5923809523809522,
"grad_norm": 0.18615860925149122,
"learning_rate": 4.38671467237115e-05,
"loss": 0.4841,
"step": 170
},
{
"epoch": 2.6076190476190475,
"grad_norm": 0.181524499480342,
"learning_rate": 4.343859194949787e-05,
"loss": 0.4792,
"step": 171
},
{
"epoch": 2.6228571428571428,
"grad_norm": 0.2025029996461714,
"learning_rate": 4.3009639150152e-05,
"loss": 0.4788,
"step": 172
},
{
"epoch": 2.638095238095238,
"grad_norm": 0.19560730560439193,
"learning_rate": 4.2580337977972664e-05,
"loss": 0.4827,
"step": 173
},
{
"epoch": 2.6533333333333333,
"grad_norm": 0.16741577940585486,
"learning_rate": 4.215073812558352e-05,
"loss": 0.4885,
"step": 174
},
{
"epoch": 2.6685714285714286,
"grad_norm": 0.1739859347419123,
"learning_rate": 4.1720889320181224e-05,
"loss": 0.4772,
"step": 175
},
{
"epoch": 2.683809523809524,
"grad_norm": 0.19324214245630159,
"learning_rate": 4.1290841317779325e-05,
"loss": 0.4884,
"step": 176
},
{
"epoch": 2.699047619047619,
"grad_norm": 0.162133413448313,
"learning_rate": 4.0860643897448894e-05,
"loss": 0.4789,
"step": 177
},
{
"epoch": 2.7142857142857144,
"grad_norm": 0.1910560430197172,
"learning_rate": 4.043034685555647e-05,
"loss": 0.4807,
"step": 178
},
{
"epoch": 2.7295238095238097,
"grad_norm": 0.14823981314032625,
"learning_rate": 4e-05,
"loss": 0.481,
"step": 179
},
{
"epoch": 2.744761904761905,
"grad_norm": 0.1784212235327859,
"learning_rate": 3.9569653144443546e-05,
"loss": 0.481,
"step": 180
},
{
"epoch": 2.76,
"grad_norm": 0.200733513756011,
"learning_rate": 3.913935610255112e-05,
"loss": 0.4763,
"step": 181
},
{
"epoch": 2.7752380952380955,
"grad_norm": 0.14938330915090647,
"learning_rate": 3.870915868222068e-05,
"loss": 0.4819,
"step": 182
},
{
"epoch": 2.7904761904761903,
"grad_norm": 0.1739261226284681,
"learning_rate": 3.8279110679818775e-05,
"loss": 0.4738,
"step": 183
},
{
"epoch": 2.8057142857142856,
"grad_norm": 0.19838890478019414,
"learning_rate": 3.78492618744165e-05,
"loss": 0.4783,
"step": 184
},
{
"epoch": 2.820952380952381,
"grad_norm": 0.1528772994448381,
"learning_rate": 3.7419662022027356e-05,
"loss": 0.4769,
"step": 185
},
{
"epoch": 2.836190476190476,
"grad_norm": 0.1919042097767604,
"learning_rate": 3.6990360849848005e-05,
"loss": 0.4757,
"step": 186
},
{
"epoch": 2.8514285714285714,
"grad_norm": 0.17566953990872838,
"learning_rate": 3.656140805050215e-05,
"loss": 0.4786,
"step": 187
},
{
"epoch": 2.8666666666666667,
"grad_norm": 0.1637970789000741,
"learning_rate": 3.6132853276288506e-05,
"loss": 0.4739,
"step": 188
},
{
"epoch": 2.881904761904762,
"grad_norm": 0.18549958217752188,
"learning_rate": 3.570474613343348e-05,
"loss": 0.4768,
"step": 189
},
{
"epoch": 2.8971428571428572,
"grad_norm": 0.18002674770058677,
"learning_rate": 3.527713617634904e-05,
"loss": 0.4733,
"step": 190
},
{
"epoch": 2.9123809523809525,
"grad_norm": 0.15898767756027013,
"learning_rate": 3.4850072901896766e-05,
"loss": 0.4802,
"step": 191
},
{
"epoch": 2.9276190476190473,
"grad_norm": 0.17974934358276728,
"learning_rate": 3.442360574365843e-05,
"loss": 0.4834,
"step": 192
},
{
"epoch": 2.942857142857143,
"grad_norm": 0.14203142815133346,
"learning_rate": 3.39977840662139e-05,
"loss": 0.4747,
"step": 193
},
{
"epoch": 2.958095238095238,
"grad_norm": 0.17952115377014885,
"learning_rate": 3.3572657159427186e-05,
"loss": 0.4808,
"step": 194
},
{
"epoch": 2.9733333333333336,
"grad_norm": 0.17353334331590287,
"learning_rate": 3.31482742327409e-05,
"loss": 0.4751,
"step": 195
},
{
"epoch": 2.9885714285714284,
"grad_norm": 0.1528037773950146,
"learning_rate": 3.272468440948025e-05,
"loss": 0.4734,
"step": 196
},
{
"epoch": 3.0047619047619047,
"grad_norm": 0.20480536463226645,
"learning_rate": 3.230193672116681e-05,
"loss": 0.6928,
"step": 197
},
{
"epoch": 3.02,
"grad_norm": 0.19151123516717333,
"learning_rate": 3.188008010184306e-05,
"loss": 0.4452,
"step": 198
},
{
"epoch": 3.0352380952380953,
"grad_norm": 0.2104984649163227,
"learning_rate": 3.145916338240816e-05,
"loss": 0.4546,
"step": 199
},
{
"epoch": 3.0504761904761906,
"grad_norm": 0.19741714449829997,
"learning_rate": 3.103923528496562e-05,
"loss": 0.4452,
"step": 200
},
{
"epoch": 3.065714285714286,
"grad_norm": 0.19623857285807467,
"learning_rate": 3.062034441718362e-05,
"loss": 0.4483,
"step": 201
},
{
"epoch": 3.080952380952381,
"grad_norm": 0.25035024771569697,
"learning_rate": 3.0202539266668568e-05,
"loss": 0.4506,
"step": 202
},
{
"epoch": 3.0961904761904764,
"grad_norm": 0.16968910324281286,
"learning_rate": 2.9785868195352525e-05,
"loss": 0.4449,
"step": 203
},
{
"epoch": 3.111428571428571,
"grad_norm": 0.23581488026041375,
"learning_rate": 2.9370379433895228e-05,
"loss": 0.4512,
"step": 204
},
{
"epoch": 3.1266666666666665,
"grad_norm": 0.15672745746884265,
"learning_rate": 2.8956121076101244e-05,
"loss": 0.449,
"step": 205
},
{
"epoch": 3.1419047619047618,
"grad_norm": 0.20702186752340365,
"learning_rate": 2.8543141073353026e-05,
"loss": 0.4439,
"step": 206
},
{
"epoch": 3.157142857142857,
"grad_norm": 0.16133838526688454,
"learning_rate": 2.8131487229060387e-05,
"loss": 0.4427,
"step": 207
},
{
"epoch": 3.1723809523809523,
"grad_norm": 0.1854062242780086,
"learning_rate": 2.7721207193127212e-05,
"loss": 0.4472,
"step": 208
},
{
"epoch": 3.1876190476190476,
"grad_norm": 0.17182738319155882,
"learning_rate": 2.7312348456435766e-05,
"loss": 0.4453,
"step": 209
},
{
"epoch": 3.202857142857143,
"grad_norm": 0.16697225397908244,
"learning_rate": 2.6904958345349633e-05,
"loss": 0.4561,
"step": 210
},
{
"epoch": 3.218095238095238,
"grad_norm": 0.16821522604659328,
"learning_rate": 2.6499084016235475e-05,
"loss": 0.444,
"step": 211
},
{
"epoch": 3.2333333333333334,
"grad_norm": 0.14853827941880265,
"learning_rate": 2.609477245000464e-05,
"loss": 0.44,
"step": 212
},
{
"epoch": 3.2485714285714287,
"grad_norm": 0.1502037122279702,
"learning_rate": 2.569207044667498e-05,
"loss": 0.4494,
"step": 213
},
{
"epoch": 3.263809523809524,
"grad_norm": 0.16216155417715333,
"learning_rate": 2.529102461995364e-05,
"loss": 0.4471,
"step": 214
},
{
"epoch": 3.279047619047619,
"grad_norm": 0.1586744564208733,
"learning_rate": 2.48916813918414e-05,
"loss": 0.4486,
"step": 215
},
{
"epoch": 3.2942857142857145,
"grad_norm": 0.14092622353201845,
"learning_rate": 2.449408698725928e-05,
"loss": 0.445,
"step": 216
},
{
"epoch": 3.3095238095238093,
"grad_norm": 0.15562276697944483,
"learning_rate": 2.409828742869777e-05,
"loss": 0.4415,
"step": 217
},
{
"epoch": 3.3247619047619046,
"grad_norm": 0.14631955274946826,
"learning_rate": 2.370432853088978e-05,
"loss": 0.454,
"step": 218
},
{
"epoch": 3.34,
"grad_norm": 0.14585582091880941,
"learning_rate": 2.3312255895507336e-05,
"loss": 0.4473,
"step": 219
},
{
"epoch": 3.355238095238095,
"grad_norm": 0.1458201114996886,
"learning_rate": 2.2922114905883167e-05,
"loss": 0.4444,
"step": 220
},
{
"epoch": 3.3704761904761904,
"grad_norm": 0.1402608135797642,
"learning_rate": 2.2533950721757447e-05,
"loss": 0.4465,
"step": 221
},
{
"epoch": 3.3857142857142857,
"grad_norm": 0.14672569191797688,
"learning_rate": 2.2147808274050436e-05,
"loss": 0.4484,
"step": 222
},
{
"epoch": 3.400952380952381,
"grad_norm": 0.15377591576947733,
"learning_rate": 2.1763732259661657e-05,
"loss": 0.4479,
"step": 223
},
{
"epoch": 3.416190476190476,
"grad_norm": 0.14739415052119614,
"learning_rate": 2.138176713629605e-05,
"loss": 0.4475,
"step": 224
},
{
"epoch": 3.4314285714285715,
"grad_norm": 0.14607883746358746,
"learning_rate": 2.1001957117318002e-05,
"loss": 0.4421,
"step": 225
},
{
"epoch": 3.4466666666666668,
"grad_norm": 0.1465432249850391,
"learning_rate": 2.0624346166633425e-05,
"loss": 0.4396,
"step": 226
},
{
"epoch": 3.461904761904762,
"grad_norm": 0.12349201931738121,
"learning_rate": 2.0248977993600912e-05,
"loss": 0.4491,
"step": 227
},
{
"epoch": 3.4771428571428573,
"grad_norm": 0.14981067132432477,
"learning_rate": 1.987589604797227e-05,
"loss": 0.4448,
"step": 228
},
{
"epoch": 3.4923809523809526,
"grad_norm": 0.1137656845152472,
"learning_rate": 1.9505143514863056e-05,
"loss": 0.4489,
"step": 229
},
{
"epoch": 3.5076190476190474,
"grad_norm": 0.1398849247762614,
"learning_rate": 1.913676330975385e-05,
"loss": 0.4546,
"step": 230
},
{
"epoch": 3.522857142857143,
"grad_norm": 0.12820525613752842,
"learning_rate": 1.8770798073522657e-05,
"loss": 0.447,
"step": 231
},
{
"epoch": 3.538095238095238,
"grad_norm": 0.13773852598915348,
"learning_rate": 1.8407290167509163e-05,
"loss": 0.4543,
"step": 232
},
{
"epoch": 3.5533333333333332,
"grad_norm": 0.12262425169332791,
"learning_rate": 1.8046281668611242e-05,
"loss": 0.444,
"step": 233
},
{
"epoch": 3.5685714285714285,
"grad_norm": 0.11207677036541486,
"learning_rate": 1.7687814364414508e-05,
"loss": 0.4451,
"step": 234
},
{
"epoch": 3.583809523809524,
"grad_norm": 0.11456531045490899,
"learning_rate": 1.73319297483553e-05,
"loss": 0.4465,
"step": 235
},
{
"epoch": 3.599047619047619,
"grad_norm": 0.12011483524657235,
"learning_rate": 1.6978669014917723e-05,
"loss": 0.4406,
"step": 236
},
{
"epoch": 3.6142857142857143,
"grad_norm": 0.11200693227331737,
"learning_rate": 1.6628073054865262e-05,
"loss": 0.4539,
"step": 237
},
{
"epoch": 3.6295238095238096,
"grad_norm": 0.11945653235133909,
"learning_rate": 1.6280182450507593e-05,
"loss": 0.4505,
"step": 238
},
{
"epoch": 3.644761904761905,
"grad_norm": 0.11328522908106976,
"learning_rate": 1.593503747100309e-05,
"loss": 0.44,
"step": 239
},
{
"epoch": 3.66,
"grad_norm": 0.11584353620347927,
"learning_rate": 1.5592678067697594e-05,
"loss": 0.4401,
"step": 240
},
{
"epoch": 3.675238095238095,
"grad_norm": 0.11338268232248719,
"learning_rate": 1.5253143869499875e-05,
"loss": 0.4526,
"step": 241
},
{
"epoch": 3.6904761904761907,
"grad_norm": 0.11825521687064344,
"learning_rate": 1.4916474178294578e-05,
"loss": 0.4472,
"step": 242
},
{
"epoch": 3.7057142857142855,
"grad_norm": 0.10393717835800384,
"learning_rate": 1.4582707964392909e-05,
"loss": 0.4503,
"step": 243
},
{
"epoch": 3.720952380952381,
"grad_norm": 0.11344323679399201,
"learning_rate": 1.4251883862021715e-05,
"loss": 0.4507,
"step": 244
},
{
"epoch": 3.736190476190476,
"grad_norm": 0.11795276330820621,
"learning_rate": 1.3924040164851497e-05,
"loss": 0.4509,
"step": 245
},
{
"epoch": 3.7514285714285713,
"grad_norm": 0.10446068928724489,
"learning_rate": 1.359921482156382e-05,
"loss": 0.453,
"step": 246
},
{
"epoch": 3.7666666666666666,
"grad_norm": 0.11963166659223025,
"learning_rate": 1.3277445431458653e-05,
"loss": 0.4423,
"step": 247
},
{
"epoch": 3.781904761904762,
"grad_norm": 0.11133235152901447,
"learning_rate": 1.2958769240102206e-05,
"loss": 0.4465,
"step": 248
},
{
"epoch": 3.797142857142857,
"grad_norm": 0.10488189725095455,
"learning_rate": 1.264322313501559e-05,
"loss": 0.4495,
"step": 249
},
{
"epoch": 3.8123809523809524,
"grad_norm": 0.09788048647298295,
"learning_rate": 1.2330843641405057e-05,
"loss": 0.4451,
"step": 250
},
{
"epoch": 3.8276190476190477,
"grad_norm": 0.1068932171231341,
"learning_rate": 1.2021666917934125e-05,
"loss": 0.4473,
"step": 251
},
{
"epoch": 3.842857142857143,
"grad_norm": 0.09965606097681808,
"learning_rate": 1.1715728752538103e-05,
"loss": 0.4443,
"step": 252
},
{
"epoch": 3.8580952380952382,
"grad_norm": 0.10053631876045145,
"learning_rate": 1.1413064558281564e-05,
"loss": 0.4468,
"step": 253
},
{
"epoch": 3.873333333333333,
"grad_norm": 0.09801079877642455,
"learning_rate": 1.1113709369259187e-05,
"loss": 0.4426,
"step": 254
},
{
"epoch": 3.888571428571429,
"grad_norm": 0.09483370655189176,
"learning_rate": 1.0817697836540493e-05,
"loss": 0.4493,
"step": 255
},
{
"epoch": 3.9038095238095236,
"grad_norm": 0.09358289511323944,
"learning_rate": 1.052506422415891e-05,
"loss": 0.4415,
"step": 256
},
{
"epoch": 3.919047619047619,
"grad_norm": 0.09937735008586661,
"learning_rate": 1.0235842405145587e-05,
"loss": 0.447,
"step": 257
},
{
"epoch": 3.934285714285714,
"grad_norm": 0.09713725112426527,
"learning_rate": 9.950065857608537e-06,
"loss": 0.4449,
"step": 258
},
{
"epoch": 3.9495238095238094,
"grad_norm": 0.10061824714900855,
"learning_rate": 9.667767660857508e-06,
"loss": 0.4378,
"step": 259
},
{
"epoch": 3.9647619047619047,
"grad_norm": 0.09975526240303212,
"learning_rate": 9.388980491574901e-06,
"loss": 0.4485,
"step": 260
},
{
"epoch": 3.98,
"grad_norm": 0.09583315993538538,
"learning_rate": 9.11373662003343e-06,
"loss": 0.4518,
"step": 261
},
{
"epoch": 3.9952380952380953,
"grad_norm": 0.12153372041343533,
"learning_rate": 8.842067906360716e-06,
"loss": 0.559,
"step": 262
},
{
"epoch": 4.011428571428572,
"grad_norm": 0.14086095004056465,
"learning_rate": 8.574005796851419e-06,
"loss": 0.5283,
"step": 263
},
{
"epoch": 4.026666666666666,
"grad_norm": 0.11788004209905123,
"learning_rate": 8.309581320327282e-06,
"loss": 0.4248,
"step": 264
},
{
"epoch": 4.041904761904762,
"grad_norm": 0.11207857563036874,
"learning_rate": 8.048825084545413e-06,
"loss": 0.4208,
"step": 265
},
{
"epoch": 4.057142857142857,
"grad_norm": 0.11130142780853505,
"learning_rate": 7.791767272655403e-06,
"loss": 0.4282,
"step": 266
},
{
"epoch": 4.072380952380953,
"grad_norm": 0.11636579346899405,
"learning_rate": 7.5384376397055644e-06,
"loss": 0.4275,
"step": 267
},
{
"epoch": 4.0876190476190475,
"grad_norm": 0.12911191155727517,
"learning_rate": 7.2888655091986995e-06,
"loss": 0.4296,
"step": 268
},
{
"epoch": 4.102857142857143,
"grad_norm": 0.1207757027022162,
"learning_rate": 7.043079769697833e-06,
"loss": 0.4301,
"step": 269
},
{
"epoch": 4.118095238095238,
"grad_norm": 0.10942499693335449,
"learning_rate": 6.801108871482287e-06,
"loss": 0.432,
"step": 270
},
{
"epoch": 4.133333333333334,
"grad_norm": 0.11446276362313004,
"learning_rate": 6.562980823254519e-06,
"loss": 0.4313,
"step": 271
},
{
"epoch": 4.148571428571429,
"grad_norm": 0.11355557887705738,
"learning_rate": 6.328723188898029e-06,
"loss": 0.4279,
"step": 272
},
{
"epoch": 4.163809523809523,
"grad_norm": 0.1062761145201427,
"learning_rate": 6.098363084286765e-06,
"loss": 0.4259,
"step": 273
},
{
"epoch": 4.179047619047619,
"grad_norm": 0.10273617911128365,
"learning_rate": 5.8719271741464056e-06,
"loss": 0.4225,
"step": 274
},
{
"epoch": 4.194285714285714,
"grad_norm": 0.10086062111767927,
"learning_rate": 5.6494416689678855e-06,
"loss": 0.4189,
"step": 275
},
{
"epoch": 4.20952380952381,
"grad_norm": 0.09559781846200713,
"learning_rate": 5.430932321973425e-06,
"loss": 0.4158,
"step": 276
},
{
"epoch": 4.2247619047619045,
"grad_norm": 0.09465874774217785,
"learning_rate": 5.21642442613556e-06,
"loss": 0.4299,
"step": 277
},
{
"epoch": 4.24,
"grad_norm": 0.09527880228230132,
"learning_rate": 5.005942811249412e-06,
"loss": 0.4243,
"step": 278
},
{
"epoch": 4.255238095238095,
"grad_norm": 0.09428002792877421,
"learning_rate": 4.799511841058575e-06,
"loss": 0.4262,
"step": 279
},
{
"epoch": 4.270476190476191,
"grad_norm": 0.09763553814742547,
"learning_rate": 4.597155410434972e-06,
"loss": 0.4247,
"step": 280
},
{
"epoch": 4.285714285714286,
"grad_norm": 0.09287137908820702,
"learning_rate": 4.398896942612934e-06,
"loss": 0.4238,
"step": 281
},
{
"epoch": 4.300952380952381,
"grad_norm": 0.09888369046987704,
"learning_rate": 4.204759386477961e-06,
"loss": 0.4391,
"step": 282
},
{
"epoch": 4.316190476190476,
"grad_norm": 0.09609824332962343,
"learning_rate": 4.01476521391027e-06,
"loss": 0.4352,
"step": 283
},
{
"epoch": 4.331428571428571,
"grad_norm": 0.09233308942677047,
"learning_rate": 3.828936417183671e-06,
"loss": 0.43,
"step": 284
},
{
"epoch": 4.346666666666667,
"grad_norm": 0.0890288964909307,
"learning_rate": 3.6472945064198827e-06,
"loss": 0.427,
"step": 285
},
{
"epoch": 4.3619047619047615,
"grad_norm": 0.08989519773450554,
"learning_rate": 3.4698605070987077e-06,
"loss": 0.4257,
"step": 286
},
{
"epoch": 4.377142857142857,
"grad_norm": 0.09526649707374434,
"learning_rate": 3.2966549576242477e-06,
"loss": 0.4357,
"step": 287
},
{
"epoch": 4.392380952380952,
"grad_norm": 0.09131870660723178,
"learning_rate": 3.127697906947584e-06,
"loss": 0.4272,
"step": 288
},
{
"epoch": 4.407619047619048,
"grad_norm": 0.08945756240639792,
"learning_rate": 2.963008912246008e-06,
"loss": 0.4347,
"step": 289
},
{
"epoch": 4.422857142857143,
"grad_norm": 0.08483664610068463,
"learning_rate": 2.802607036659284e-06,
"loss": 0.4294,
"step": 290
},
{
"epoch": 4.438095238095238,
"grad_norm": 0.08486549550090346,
"learning_rate": 2.646510847082997e-06,
"loss": 0.4303,
"step": 291
},
{
"epoch": 4.453333333333333,
"grad_norm": 0.084355078151828,
"learning_rate": 2.494738412019442e-06,
"loss": 0.423,
"step": 292
},
{
"epoch": 4.468571428571429,
"grad_norm": 0.08518364314817348,
"learning_rate": 2.3473072994861167e-06,
"loss": 0.427,
"step": 293
},
{
"epoch": 4.483809523809524,
"grad_norm": 0.08287931866258626,
"learning_rate": 2.2042345749821913e-06,
"loss": 0.4216,
"step": 294
},
{
"epoch": 4.499047619047619,
"grad_norm": 0.08289838691379334,
"learning_rate": 2.065536799513135e-06,
"loss": 0.4288,
"step": 295
},
{
"epoch": 4.514285714285714,
"grad_norm": 0.0850326818939402,
"learning_rate": 1.9312300276737516e-06,
"loss": 0.4235,
"step": 296
},
{
"epoch": 4.529523809523809,
"grad_norm": 0.08392478938826013,
"learning_rate": 1.8013298057897932e-06,
"loss": 0.4293,
"step": 297
},
{
"epoch": 4.544761904761905,
"grad_norm": 0.0892781985318678,
"learning_rate": 1.6758511701184809e-06,
"loss": 0.4238,
"step": 298
},
{
"epoch": 4.5600000000000005,
"grad_norm": 0.08180406525124756,
"learning_rate": 1.5548086451079747e-06,
"loss": 0.4384,
"step": 299
},
{
"epoch": 4.575238095238095,
"grad_norm": 0.08011550631601072,
"learning_rate": 1.4382162417161748e-06,
"loss": 0.4206,
"step": 300
},
{
"epoch": 4.59047619047619,
"grad_norm": 0.08066581660175723,
"learning_rate": 1.3260874557889002e-06,
"loss": 0.4287,
"step": 301
},
{
"epoch": 4.605714285714286,
"grad_norm": 0.0806215224586864,
"learning_rate": 1.2184352664977105e-06,
"loss": 0.4175,
"step": 302
},
{
"epoch": 4.620952380952381,
"grad_norm": 0.07969848747023248,
"learning_rate": 1.115272134837544e-06,
"loss": 0.4312,
"step": 303
},
{
"epoch": 4.636190476190476,
"grad_norm": 0.0797095157185979,
"learning_rate": 1.0166100021843283e-06,
"loss": 0.4346,
"step": 304
},
{
"epoch": 4.651428571428571,
"grad_norm": 0.0792961159548818,
"learning_rate": 9.224602889127321e-07,
"loss": 0.4286,
"step": 305
},
{
"epoch": 4.666666666666667,
"grad_norm": 0.08285323866652071,
"learning_rate": 8.328338930742296e-07,
"loss": 0.4267,
"step": 306
},
{
"epoch": 4.681904761904762,
"grad_norm": 0.07823875843722773,
"learning_rate": 7.477411891356268e-07,
"loss": 0.4276,
"step": 307
},
{
"epoch": 4.6971428571428575,
"grad_norm": 0.07763022247982207,
"learning_rate": 6.671920267782029e-07,
"loss": 0.4282,
"step": 308
},
{
"epoch": 4.712380952380952,
"grad_norm": 0.07871343041787499,
"learning_rate": 5.911957297575743e-07,
"loss": 0.4323,
"step": 309
},
{
"epoch": 4.727619047619047,
"grad_norm": 0.07681136995775628,
"learning_rate": 5.197610948244469e-07,
"loss": 0.4256,
"step": 310
},
{
"epoch": 4.742857142857143,
"grad_norm": 0.07665250706053243,
"learning_rate": 4.5289639070638103e-07,
"loss": 0.4171,
"step": 311
},
{
"epoch": 4.758095238095238,
"grad_norm": 0.07835687105047397,
"learning_rate": 3.906093571506597e-07,
"loss": 0.428,
"step": 312
},
{
"epoch": 4.773333333333333,
"grad_norm": 0.07949715755664716,
"learning_rate": 3.329072040284009e-07,
"loss": 0.4208,
"step": 313
},
{
"epoch": 4.788571428571428,
"grad_norm": 0.0773084553588578,
"learning_rate": 2.797966104999805e-07,
"loss": 0.4332,
"step": 314
},
{
"epoch": 4.803809523809524,
"grad_norm": 0.07795013678746698,
"learning_rate": 2.3128372424192635e-07,
"loss": 0.4377,
"step": 315
},
{
"epoch": 4.819047619047619,
"grad_norm": 0.07666311315306534,
"learning_rate": 1.8737416073529635e-07,
"loss": 0.4272,
"step": 316
},
{
"epoch": 4.8342857142857145,
"grad_norm": 0.07693103942282913,
"learning_rate": 1.4807300261568735e-07,
"loss": 0.4287,
"step": 317
},
{
"epoch": 4.849523809523809,
"grad_norm": 0.08092920067828374,
"learning_rate": 1.1338479908488798e-07,
"loss": 0.4337,
"step": 318
},
{
"epoch": 4.864761904761905,
"grad_norm": 0.07921080037622698,
"learning_rate": 8.331356538431313e-08,
"loss": 0.4262,
"step": 319
},
{
"epoch": 4.88,
"grad_norm": 0.08216408996820164,
"learning_rate": 5.7862782330215804e-08,
"loss": 0.4317,
"step": 320
},
{
"epoch": 4.895238095238096,
"grad_norm": 0.07591222440722502,
"learning_rate": 3.703539591080052e-08,
"loss": 0.4232,
"step": 321
},
{
"epoch": 4.91047619047619,
"grad_norm": 0.08045135907636206,
"learning_rate": 2.0833816945184936e-08,
"loss": 0.4297,
"step": 322
},
{
"epoch": 4.925714285714285,
"grad_norm": 0.07684321088822434,
"learning_rate": 9.25992080436533e-09,
"loss": 0.4276,
"step": 323
},
{
"epoch": 4.940952380952381,
"grad_norm": 0.07680665420588856,
"learning_rate": 2.315047194136888e-09,
"loss": 0.4271,
"step": 324
},
{
"epoch": 4.956190476190476,
"grad_norm": 0.07663586105720761,
"learning_rate": 0.0,
"loss": 0.4248,
"step": 325
},
{
"epoch": 4.956190476190476,
"step": 325,
"total_flos": 8.641809758367515e+18,
"train_loss": 0.5190348960803105,
"train_runtime": 75842.3225,
"train_samples_per_second": 2.214,
"train_steps_per_second": 0.004
}
],
"logging_steps": 1,
"max_steps": 325,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.641809758367515e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}