{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1290, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.023255813953488372, "grad_norm": 2.1102249816437895, "learning_rate": 2.0930232558139536e-06, "loss": 0.4502, "step": 10 }, { "epoch": 0.046511627906976744, "grad_norm": 0.6087936521314388, "learning_rate": 4.418604651162791e-06, "loss": 0.3222, "step": 20 }, { "epoch": 0.06976744186046512, "grad_norm": 0.5276269917240805, "learning_rate": 6.744186046511628e-06, "loss": 0.2413, "step": 30 }, { "epoch": 0.09302325581395349, "grad_norm": 0.33927814937972933, "learning_rate": 9.069767441860465e-06, "loss": 0.2011, "step": 40 }, { "epoch": 0.11627906976744186, "grad_norm": 0.41749972892158266, "learning_rate": 1.1395348837209304e-05, "loss": 0.1824, "step": 50 }, { "epoch": 0.13953488372093023, "grad_norm": 0.3467307268335015, "learning_rate": 1.3720930232558139e-05, "loss": 0.1777, "step": 60 }, { "epoch": 0.16279069767441862, "grad_norm": 0.2909147698009027, "learning_rate": 1.6046511627906977e-05, "loss": 0.1641, "step": 70 }, { "epoch": 0.18604651162790697, "grad_norm": 0.6246257892448243, "learning_rate": 1.8372093023255815e-05, "loss": 0.1515, "step": 80 }, { "epoch": 0.20930232558139536, "grad_norm": 0.35908006141784865, "learning_rate": 2.069767441860465e-05, "loss": 0.1491, "step": 90 }, { "epoch": 0.23255813953488372, "grad_norm": 0.27603388668684564, "learning_rate": 2.302325581395349e-05, "loss": 0.1481, "step": 100 }, { "epoch": 0.2558139534883721, "grad_norm": 0.2527784608738212, "learning_rate": 2.5348837209302324e-05, "loss": 0.1397, "step": 110 }, { "epoch": 0.27906976744186046, "grad_norm": 0.29932144986900905, "learning_rate": 2.7674418604651162e-05, "loss": 0.1404, "step": 120 }, { "epoch": 0.3023255813953488, "grad_norm": 0.3393202145121213, "learning_rate": 3e-05, "loss": 0.142, "step": 130 }, { "epoch": 0.32558139534883723, "grad_norm": 0.24298726842558038, "learning_rate": 2.999450876449861e-05, "loss": 0.1353, "step": 140 }, { "epoch": 0.3488372093023256, "grad_norm": 0.17700673786533003, "learning_rate": 2.9978039078483434e-05, "loss": 0.1316, "step": 150 }, { "epoch": 0.37209302325581395, "grad_norm": 0.23624869484478003, "learning_rate": 2.9950603000477722e-05, "loss": 0.1296, "step": 160 }, { "epoch": 0.3953488372093023, "grad_norm": 0.2450597575472511, "learning_rate": 2.991222061821024e-05, "loss": 0.1322, "step": 170 }, { "epoch": 0.4186046511627907, "grad_norm": 0.2629976539114997, "learning_rate": 2.9862920033907654e-05, "loss": 0.1272, "step": 180 }, { "epoch": 0.4418604651162791, "grad_norm": 0.158909112774518, "learning_rate": 2.980273734371914e-05, "loss": 0.1265, "step": 190 }, { "epoch": 0.46511627906976744, "grad_norm": 0.16963032221578891, "learning_rate": 2.9731716611288024e-05, "loss": 0.1275, "step": 200 }, { "epoch": 0.4883720930232558, "grad_norm": 0.207022304132148, "learning_rate": 2.9649909835489935e-05, "loss": 0.1262, "step": 210 }, { "epoch": 0.5116279069767442, "grad_norm": 0.18726792852194557, "learning_rate": 2.955737691236108e-05, "loss": 0.1198, "step": 220 }, { "epoch": 0.5348837209302325, "grad_norm": 0.1701049540713391, "learning_rate": 2.945418559124446e-05, "loss": 0.1209, "step": 230 }, { "epoch": 0.5581395348837209, "grad_norm": 0.19958209659130646, "learning_rate": 2.9340411425186207e-05, "loss": 0.122, "step": 240 }, { "epoch": 0.5813953488372093, "grad_norm": 0.22288049794980888, "learning_rate": 2.921613771561829e-05, "loss": 0.1155, "step": 250 }, { "epoch": 0.6046511627906976, "grad_norm": 0.17056830281402427, "learning_rate": 2.908145545136816e-05, "loss": 0.1123, "step": 260 }, { "epoch": 0.627906976744186, "grad_norm": 0.1726845936656448, "learning_rate": 2.8936463242039924e-05, "loss": 0.1139, "step": 270 }, { "epoch": 0.6511627906976745, "grad_norm": 0.17997669987525944, "learning_rate": 2.8781267245815898e-05, "loss": 0.1085, "step": 280 }, { "epoch": 0.6744186046511628, "grad_norm": 0.19695001871413667, "learning_rate": 2.8615981091731296e-05, "loss": 0.113, "step": 290 }, { "epoch": 0.6976744186046512, "grad_norm": 0.19008651260372486, "learning_rate": 2.8440725796479075e-05, "loss": 0.1172, "step": 300 }, { "epoch": 0.7209302325581395, "grad_norm": 0.18432725372928307, "learning_rate": 2.825562967580579e-05, "loss": 0.113, "step": 310 }, { "epoch": 0.7441860465116279, "grad_norm": 0.14972827867736163, "learning_rate": 2.8060828250563294e-05, "loss": 0.1113, "step": 320 }, { "epoch": 0.7674418604651163, "grad_norm": 0.1708006253644624, "learning_rate": 2.7856464147485202e-05, "loss": 0.1118, "step": 330 }, { "epoch": 0.7906976744186046, "grad_norm": 0.2545153275078911, "learning_rate": 2.764268699476058e-05, "loss": 0.1126, "step": 340 }, { "epoch": 0.813953488372093, "grad_norm": 0.13582442076029375, "learning_rate": 2.7419653312481482e-05, "loss": 0.1084, "step": 350 }, { "epoch": 0.8372093023255814, "grad_norm": 0.1885021789487428, "learning_rate": 2.7187526398044463e-05, "loss": 0.1177, "step": 360 }, { "epoch": 0.8604651162790697, "grad_norm": 0.18897101004109837, "learning_rate": 2.6946476206589972e-05, "loss": 0.1044, "step": 370 }, { "epoch": 0.8837209302325582, "grad_norm": 0.12704249962452174, "learning_rate": 2.6696679226567202e-05, "loss": 0.1099, "step": 380 }, { "epoch": 0.9069767441860465, "grad_norm": 0.19659069921161837, "learning_rate": 2.6438318350515467e-05, "loss": 0.1084, "step": 390 }, { "epoch": 0.9302325581395349, "grad_norm": 0.17360269942962656, "learning_rate": 2.617158274115673e-05, "loss": 0.1086, "step": 400 }, { "epoch": 0.9534883720930233, "grad_norm": 0.15978057253464667, "learning_rate": 2.5896667692897334e-05, "loss": 0.098, "step": 410 }, { "epoch": 0.9767441860465116, "grad_norm": 0.12225147895714812, "learning_rate": 2.5613774488840333e-05, "loss": 0.1083, "step": 420 }, { "epoch": 1.0, "grad_norm": 0.11136831453413676, "learning_rate": 2.532311025341309e-05, "loss": 0.1064, "step": 430 }, { "epoch": 1.0232558139534884, "grad_norm": 0.11500268275827884, "learning_rate": 2.502488780071807e-05, "loss": 0.1053, "step": 440 }, { "epoch": 1.0465116279069768, "grad_norm": 0.13683341482554096, "learning_rate": 2.4719325478717893e-05, "loss": 0.103, "step": 450 }, { "epoch": 1.069767441860465, "grad_norm": 0.1391528532587655, "learning_rate": 2.440664700936861e-05, "loss": 0.102, "step": 460 }, { "epoch": 1.0930232558139534, "grad_norm": 0.1591189829381333, "learning_rate": 2.408708132481842e-05, "loss": 0.1017, "step": 470 }, { "epoch": 1.1162790697674418, "grad_norm": 0.10083202954080314, "learning_rate": 2.376086239979158e-05, "loss": 0.1047, "step": 480 }, { "epoch": 1.1395348837209303, "grad_norm": 0.2036430998599195, "learning_rate": 2.3428229080280407e-05, "loss": 0.1018, "step": 490 }, { "epoch": 1.1627906976744187, "grad_norm": 0.15460019789887855, "learning_rate": 2.3089424908670642e-05, "loss": 0.1006, "step": 500 }, { "epoch": 1.1860465116279069, "grad_norm": 0.1171523300082374, "learning_rate": 2.2744697945428307e-05, "loss": 0.1007, "step": 510 }, { "epoch": 1.2093023255813953, "grad_norm": 0.19738915842381824, "learning_rate": 2.2394300587478566e-05, "loss": 0.0972, "step": 520 }, { "epoch": 1.2325581395348837, "grad_norm": 0.16778096884891686, "learning_rate": 2.2038489383409652e-05, "loss": 0.1027, "step": 530 }, { "epoch": 1.255813953488372, "grad_norm": 0.13067156491026763, "learning_rate": 2.167752484563696e-05, "loss": 0.0997, "step": 540 }, { "epoch": 1.2790697674418605, "grad_norm": 0.16129671511714355, "learning_rate": 2.13116712596651e-05, "loss": 0.1025, "step": 550 }, { "epoch": 1.302325581395349, "grad_norm": 0.15785271361590572, "learning_rate": 2.0941196490587352e-05, "loss": 0.1002, "step": 560 }, { "epoch": 1.3255813953488373, "grad_norm": 0.13841764309679408, "learning_rate": 2.05663717869643e-05, "loss": 0.0994, "step": 570 }, { "epoch": 1.3488372093023255, "grad_norm": 0.17509131465232344, "learning_rate": 2.0187471582225173e-05, "loss": 0.0993, "step": 580 }, { "epoch": 1.372093023255814, "grad_norm": 0.1221451212745489, "learning_rate": 1.9804773293737416e-05, "loss": 0.0998, "step": 590 }, { "epoch": 1.3953488372093024, "grad_norm": 0.14688306712971377, "learning_rate": 1.9418557119691434e-05, "loss": 0.0988, "step": 600 }, { "epoch": 1.4186046511627908, "grad_norm": 0.10658674461240546, "learning_rate": 1.902910583394938e-05, "loss": 0.0994, "step": 610 }, { "epoch": 1.441860465116279, "grad_norm": 0.11935538384825148, "learning_rate": 1.8636704579008096e-05, "loss": 0.0985, "step": 620 }, { "epoch": 1.4651162790697674, "grad_norm": 0.12721218184322167, "learning_rate": 1.824164065722783e-05, "loss": 0.0994, "step": 630 }, { "epoch": 1.4883720930232558, "grad_norm": 0.09137892917036859, "learning_rate": 1.7844203320479614e-05, "loss": 0.1047, "step": 640 }, { "epoch": 1.5116279069767442, "grad_norm": 0.23254378397439393, "learning_rate": 1.7444683558365182e-05, "loss": 0.0959, "step": 650 }, { "epoch": 1.5348837209302326, "grad_norm": 0.12936296282790546, "learning_rate": 1.7043373885164703e-05, "loss": 0.0968, "step": 660 }, { "epoch": 1.558139534883721, "grad_norm": 0.15526942233091734, "learning_rate": 1.664056812566812e-05, "loss": 0.1004, "step": 670 }, { "epoch": 1.5813953488372094, "grad_norm": 0.12963665517042525, "learning_rate": 1.623656120004698e-05, "loss": 0.0921, "step": 680 }, { "epoch": 1.6046511627906976, "grad_norm": 0.11191785293454326, "learning_rate": 1.5831648907924337e-05, "loss": 0.0969, "step": 690 }, { "epoch": 1.627906976744186, "grad_norm": 0.13773389105774028, "learning_rate": 1.5426127711800636e-05, "loss": 0.0994, "step": 700 }, { "epoch": 1.6511627906976745, "grad_norm": 0.10576569528021193, "learning_rate": 1.5020294519994381e-05, "loss": 0.0957, "step": 710 }, { "epoch": 1.6744186046511627, "grad_norm": 0.1437216064703046, "learning_rate": 1.4614446469256305e-05, "loss": 0.0945, "step": 720 }, { "epoch": 1.697674418604651, "grad_norm": 0.10803257779082216, "learning_rate": 1.4208880707216323e-05, "loss": 0.095, "step": 730 }, { "epoch": 1.7209302325581395, "grad_norm": 0.12123283799800175, "learning_rate": 1.3803894174822518e-05, "loss": 0.0944, "step": 740 }, { "epoch": 1.744186046511628, "grad_norm": 0.100414996108917, "learning_rate": 1.3399783388931468e-05, "loss": 0.0954, "step": 750 }, { "epoch": 1.7674418604651163, "grad_norm": 0.1369011425168964, "learning_rate": 1.2996844225209033e-05, "loss": 0.0984, "step": 760 }, { "epoch": 1.7906976744186047, "grad_norm": 0.1359084312900855, "learning_rate": 1.2595371701500639e-05, "loss": 0.0962, "step": 770 }, { "epoch": 1.8139534883720931, "grad_norm": 0.13125777727884955, "learning_rate": 1.219565976182963e-05, "loss": 0.0972, "step": 780 }, { "epoch": 1.8372093023255816, "grad_norm": 0.13341223023541443, "learning_rate": 1.1798001061181799e-05, "loss": 0.0952, "step": 790 }, { "epoch": 1.8604651162790697, "grad_norm": 0.08587984083521312, "learning_rate": 1.1402686751233723e-05, "loss": 0.0969, "step": 800 }, { "epoch": 1.8837209302325582, "grad_norm": 0.11630108959701958, "learning_rate": 1.101000626718182e-05, "loss": 0.0915, "step": 810 }, { "epoch": 1.9069767441860463, "grad_norm": 0.0951067714002941, "learning_rate": 1.0620247115828044e-05, "loss": 0.0966, "step": 820 }, { "epoch": 1.9302325581395348, "grad_norm": 0.10789836511240755, "learning_rate": 1.0233694665077584e-05, "loss": 0.0946, "step": 830 }, { "epoch": 1.9534883720930232, "grad_norm": 0.10824454265797709, "learning_rate": 9.850631935002531e-06, "loss": 0.0956, "step": 840 }, { "epoch": 1.9767441860465116, "grad_norm": 0.10283011375161531, "learning_rate": 9.471339390624574e-06, "loss": 0.0949, "step": 850 }, { "epoch": 2.0, "grad_norm": 0.15809185521039087, "learning_rate": 9.0960947365684e-06, "loss": 0.0934, "step": 860 }, { "epoch": 2.0232558139534884, "grad_norm": 0.15384352891044772, "learning_rate": 8.725172713736136e-06, "loss": 0.0889, "step": 870 }, { "epoch": 2.046511627906977, "grad_norm": 0.11467341711361749, "learning_rate": 8.358844898151791e-06, "loss": 0.0882, "step": 880 }, { "epoch": 2.0697674418604652, "grad_norm": 0.10965841387662369, "learning_rate": 7.997379502122849e-06, "loss": 0.0944, "step": 890 }, { "epoch": 2.0930232558139537, "grad_norm": 0.1024892604780862, "learning_rate": 7.641041177864661e-06, "loss": 0.0892, "step": 900 }, { "epoch": 2.116279069767442, "grad_norm": 0.14766897087136052, "learning_rate": 7.290090823731452e-06, "loss": 0.0904, "step": 910 }, { "epoch": 2.13953488372093, "grad_norm": 0.11551061066180639, "learning_rate": 6.944785393195742e-06, "loss": 0.0933, "step": 920 }, { "epoch": 2.1627906976744184, "grad_norm": 0.1307266761701476, "learning_rate": 6.605377706716049e-06, "loss": 0.0866, "step": 930 }, { "epoch": 2.186046511627907, "grad_norm": 0.10681432279549767, "learning_rate": 6.27211626663071e-06, "loss": 0.0869, "step": 940 }, { "epoch": 2.2093023255813953, "grad_norm": 0.1471123103304021, "learning_rate": 5.945245075213187e-06, "loss": 0.0858, "step": 950 }, { "epoch": 2.2325581395348837, "grad_norm": 0.1518492318767705, "learning_rate": 5.625003456022247e-06, "loss": 0.0854, "step": 960 }, { "epoch": 2.255813953488372, "grad_norm": 0.12598219156122908, "learning_rate": 5.311625878677658e-06, "loss": 0.0853, "step": 970 }, { "epoch": 2.2790697674418605, "grad_norm": 0.15991189489983731, "learning_rate": 5.005341787189832e-06, "loss": 0.0914, "step": 980 }, { "epoch": 2.302325581395349, "grad_norm": 0.10711098788498662, "learning_rate": 4.706375431968998e-06, "loss": 0.0833, "step": 990 }, { "epoch": 2.3255813953488373, "grad_norm": 0.15238996613139555, "learning_rate": 4.414945705636949e-06, "loss": 0.088, "step": 1000 }, { "epoch": 2.3488372093023258, "grad_norm": 0.13302573278619106, "learning_rate": 4.131265982761614e-06, "loss": 0.0872, "step": 1010 }, { "epoch": 2.3720930232558137, "grad_norm": 0.09653650821534612, "learning_rate": 3.855543963631685e-06, "loss": 0.0895, "step": 1020 }, { "epoch": 2.395348837209302, "grad_norm": 0.12648429239194942, "learning_rate": 3.587981522185829e-06, "loss": 0.0854, "step": 1030 }, { "epoch": 2.4186046511627906, "grad_norm": 0.1545950486769342, "learning_rate": 3.328774558207692e-06, "loss": 0.0909, "step": 1040 }, { "epoch": 2.441860465116279, "grad_norm": 0.10549590353114048, "learning_rate": 3.0781128538949714e-06, "loss": 0.0878, "step": 1050 }, { "epoch": 2.4651162790697674, "grad_norm": 0.18177777322706803, "learning_rate": 2.8361799349076143e-06, "loss": 0.0852, "step": 1060 }, { "epoch": 2.488372093023256, "grad_norm": 0.11868713958151146, "learning_rate": 2.6031529359967833e-06, "loss": 0.0891, "step": 1070 }, { "epoch": 2.511627906976744, "grad_norm": 0.1284488259520864, "learning_rate": 2.3792024713130284e-06, "loss": 0.089, "step": 1080 }, { "epoch": 2.5348837209302326, "grad_norm": 0.14486344360735834, "learning_rate": 2.164492509488657e-06, "loss": 0.0803, "step": 1090 }, { "epoch": 2.558139534883721, "grad_norm": 0.12768774247632075, "learning_rate": 1.9591802535856433e-06, "loss": 0.0852, "step": 1100 }, { "epoch": 2.5813953488372094, "grad_norm": 0.12486343748182972, "learning_rate": 1.763416025997126e-06, "loss": 0.0854, "step": 1110 }, { "epoch": 2.604651162790698, "grad_norm": 0.1325469478337174, "learning_rate": 1.5773431583866227e-06, "loss": 0.09, "step": 1120 }, { "epoch": 2.6279069767441863, "grad_norm": 0.1483837067276731, "learning_rate": 1.4010978867456664e-06, "loss": 0.0864, "step": 1130 }, { "epoch": 2.6511627906976747, "grad_norm": 0.10867631647084727, "learning_rate": 1.2348092516466032e-06, "loss": 0.0859, "step": 1140 }, { "epoch": 2.6744186046511627, "grad_norm": 0.10578223359112342, "learning_rate": 1.0785990037636335e-06, "loss": 0.0852, "step": 1150 }, { "epoch": 2.697674418604651, "grad_norm": 0.11026052474123807, "learning_rate": 9.325815147312739e-07, "loss": 0.0858, "step": 1160 }, { "epoch": 2.7209302325581395, "grad_norm": 0.1433636551446035, "learning_rate": 7.968636934054741e-07, "loss": 0.0833, "step": 1170 }, { "epoch": 2.744186046511628, "grad_norm": 0.12573674370411025, "learning_rate": 6.71544907588712e-07, "loss": 0.0783, "step": 1180 }, { "epoch": 2.7674418604651163, "grad_norm": 0.13603777866417854, "learning_rate": 5.567169112764109e-07, "loss": 0.0856, "step": 1190 }, { "epoch": 2.7906976744186047, "grad_norm": 0.12748468973009236, "learning_rate": 4.524637774778984e-07, "loss": 0.0832, "step": 1200 }, { "epoch": 2.813953488372093, "grad_norm": 0.10740073013695721, "learning_rate": 3.588618366610941e-07, "loss": 0.0859, "step": 1210 }, { "epoch": 2.8372093023255816, "grad_norm": 0.16114245090981888, "learning_rate": 2.7597962086605255e-07, "loss": 0.0842, "step": 1220 }, { "epoch": 2.8604651162790695, "grad_norm": 0.1597114067458733, "learning_rate": 2.038778135282171e-07, "loss": 0.0891, "step": 1230 }, { "epoch": 2.883720930232558, "grad_norm": 0.15931507855411944, "learning_rate": 1.4260920504814366e-07, "loss": 0.086, "step": 1240 }, { "epoch": 2.9069767441860463, "grad_norm": 0.15611356894003056, "learning_rate": 9.221865414023201e-08, "loss": 0.0808, "step": 1250 }, { "epoch": 2.9302325581395348, "grad_norm": 0.12919745509778163, "learning_rate": 5.2743054988758085e-08, "loss": 0.0813, "step": 1260 }, { "epoch": 2.953488372093023, "grad_norm": 0.11402034227218683, "learning_rate": 2.4211310235258687e-08, "loss": 0.0787, "step": 1270 }, { "epoch": 2.9767441860465116, "grad_norm": 0.14204729077366124, "learning_rate": 6.6443098170271276e-09, "loss": 0.0846, "step": 1280 }, { "epoch": 3.0, "grad_norm": 0.1646401717020459, "learning_rate": 5.4915672239586045e-11, "loss": 0.088, "step": 1290 } ], "logging_steps": 10, "max_steps": 1290, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 173923848683520.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }