| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9999958772917105, | |
| "eval_steps": 1000, | |
| "global_step": 121279, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.00024811801548585953, | |
| "loss": 8.2906, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.00029403430324938403, | |
| "loss": 5.1339, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.00029980426002857634, | |
| "loss": 4.2501, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.00029955648791285023, | |
| "loss": 3.94, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.0002993087157971242, | |
| "loss": 3.7765, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.0002990609436813981, | |
| "loss": 3.6756, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.000298813171565672, | |
| "loss": 3.6106, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029856539944994587, | |
| "loss": 3.5704, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029831762733421977, | |
| "loss": 3.5304, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029806985521849366, | |
| "loss": 3.5018, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_accuracy": 0.3767958311915216, | |
| "eval_loss": 3.509509325027466, | |
| "eval_runtime": 36.7301, | |
| "eval_samples_per_second": 309.447, | |
| "eval_steps_per_second": 2.586, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.0002978220831027676, | |
| "loss": 3.4716, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029757678870819877, | |
| "loss": 3.4476, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029732901659247266, | |
| "loss": 3.4368, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029708124447674656, | |
| "loss": 3.4075, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029683347236102045, | |
| "loss": 3.3945, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029658570024529435, | |
| "loss": 3.3788, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.00029633792812956824, | |
| "loss": 3.3738, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.0002960901560138422, | |
| "loss": 3.3585, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002958423838981161, | |
| "loss": 3.3479, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.00029559461178239, | |
| "loss": 3.3375, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "eval_accuracy": 0.391758262312935, | |
| "eval_loss": 3.3602521419525146, | |
| "eval_runtime": 36.8476, | |
| "eval_samples_per_second": 308.46, | |
| "eval_steps_per_second": 2.578, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002953468396666639, | |
| "loss": 3.3351, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.00029509906755093777, | |
| "loss": 3.3283, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002948512954352117, | |
| "loss": 3.3185, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002946035233194856, | |
| "loss": 3.309, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002943557512037595, | |
| "loss": 3.3015, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002941079790880334, | |
| "loss": 3.3015, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002938602069723073, | |
| "loss": 3.2913, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002936124348565812, | |
| "loss": 3.2871, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.00029336466274085515, | |
| "loss": 3.2761, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.0002931193683462863, | |
| "loss": 3.2694, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "eval_accuracy": 0.39822664917193284, | |
| "eval_loss": 3.2995240688323975, | |
| "eval_runtime": 36.3126, | |
| "eval_samples_per_second": 313.004, | |
| "eval_steps_per_second": 2.616, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002928715962305602, | |
| "loss": 3.2757, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002926238241148341, | |
| "loss": 3.2597, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.000292376051999108, | |
| "loss": 3.2603, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002921282798833819, | |
| "loss": 3.2499, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002918805077676558, | |
| "loss": 3.2522, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.00029163273565192973, | |
| "loss": 3.2511, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002913849635362036, | |
| "loss": 3.2399, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002911371914204775, | |
| "loss": 3.2459, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002908894193047514, | |
| "loss": 3.2364, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002906416471890253, | |
| "loss": 3.236, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "eval_accuracy": 0.4022114581269458, | |
| "eval_loss": 3.262552261352539, | |
| "eval_runtime": 36.6585, | |
| "eval_samples_per_second": 310.051, | |
| "eval_steps_per_second": 2.591, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0002903938750732992, | |
| "loss": 3.2373, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.00029014610295757315, | |
| "loss": 3.2318, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028989833084184705, | |
| "loss": 3.2394, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028965055872612095, | |
| "loss": 3.2198, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028940278661039484, | |
| "loss": 3.2214, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.000289157492215826, | |
| "loss": 3.2195, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0002889097201000999, | |
| "loss": 3.22, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0002886619479843738, | |
| "loss": 3.2135, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028841417586864774, | |
| "loss": 3.2163, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028816640375292163, | |
| "loss": 3.2131, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_accuracy": 0.40479479480866404, | |
| "eval_loss": 3.2350122928619385, | |
| "eval_runtime": 36.6665, | |
| "eval_samples_per_second": 309.983, | |
| "eval_steps_per_second": 2.591, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0002879186316371955, | |
| "loss": 3.2066, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0002876708595214694, | |
| "loss": 3.2067, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0002874230874057433, | |
| "loss": 3.2011, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.00028717531529001727, | |
| "loss": 3.1956, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028692754317429116, | |
| "loss": 3.1997, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028667977105856506, | |
| "loss": 3.2032, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028643199894283895, | |
| "loss": 3.1976, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028618422682711285, | |
| "loss": 3.1987, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028593645471138674, | |
| "loss": 3.1932, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.0002856886825956607, | |
| "loss": 3.1934, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "eval_accuracy": 0.4072977702527638, | |
| "eval_loss": 3.21467661857605, | |
| "eval_runtime": 35.9987, | |
| "eval_samples_per_second": 315.734, | |
| "eval_steps_per_second": 2.639, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028544091047993453, | |
| "loss": 3.1823, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028519313836420843, | |
| "loss": 3.1893, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.0002849453662484824, | |
| "loss": 3.1883, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.0002846975941327563, | |
| "loss": 3.1824, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028444982201703017, | |
| "loss": 3.1806, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00028420204990130406, | |
| "loss": 3.178, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028395427778557796, | |
| "loss": 3.1819, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028370898339100917, | |
| "loss": 3.1905, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028346121127528306, | |
| "loss": 3.1775, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028321343915955696, | |
| "loss": 3.177, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "eval_accuracy": 0.40887866977212045, | |
| "eval_loss": 3.2006027698516846, | |
| "eval_runtime": 36.3445, | |
| "eval_samples_per_second": 312.73, | |
| "eval_steps_per_second": 2.614, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028296566704383085, | |
| "loss": 3.1713, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002827178949281048, | |
| "loss": 3.1711, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002824701228123787, | |
| "loss": 3.1705, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00028222235069665254, | |
| "loss": 3.1755, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002819745785809265, | |
| "loss": 3.1673, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002817268064652004, | |
| "loss": 3.1703, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002814790343494743, | |
| "loss": 3.1608, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0002812312622337482, | |
| "loss": 3.1637, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00028098349011802207, | |
| "loss": 3.1726, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00028073571800229597, | |
| "loss": 3.1653, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "eval_accuracy": 0.4097224627671145, | |
| "eval_loss": 3.1890077590942383, | |
| "eval_runtime": 36.3132, | |
| "eval_samples_per_second": 312.999, | |
| "eval_steps_per_second": 2.616, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002804879458865699, | |
| "loss": 3.1585, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002802401737708438, | |
| "loss": 3.1608, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002799924016551177, | |
| "loss": 3.1606, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002797446295393916, | |
| "loss": 3.1605, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002794968574236655, | |
| "loss": 3.1588, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00027924908530793945, | |
| "loss": 3.1691, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00027900131319221334, | |
| "loss": 3.1556, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00027875354107648724, | |
| "loss": 3.1505, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002785082466819184, | |
| "loss": 3.162, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0002782604745661923, | |
| "loss": 3.1548, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "eval_accuracy": 0.4109264957928121, | |
| "eval_loss": 3.1779375076293945, | |
| "eval_runtime": 36.2677, | |
| "eval_samples_per_second": 313.392, | |
| "eval_steps_per_second": 2.619, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002780127024504662, | |
| "loss": 3.1461, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002777649303347401, | |
| "loss": 3.1508, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00027751715821901403, | |
| "loss": 3.1521, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002772718638244452, | |
| "loss": 3.1486, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002770240917087191, | |
| "loss": 3.1457, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.000276776319592993, | |
| "loss": 3.1577, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00027652854747726687, | |
| "loss": 3.1528, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002762807753615408, | |
| "loss": 3.1404, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002760330032458147, | |
| "loss": 3.145, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00027578523113008855, | |
| "loss": 3.152, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_accuracy": 0.4119192341851163, | |
| "eval_loss": 3.169926404953003, | |
| "eval_runtime": 36.3986, | |
| "eval_samples_per_second": 312.265, | |
| "eval_steps_per_second": 2.61, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002755374590143625, | |
| "loss": 3.1468, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002752896868986364, | |
| "loss": 3.1396, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0002750419147829103, | |
| "loss": 3.1375, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002747941426671842, | |
| "loss": 3.1463, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002745463705514581, | |
| "loss": 3.1433, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00027429859843573203, | |
| "loss": 3.145, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00027405082632000593, | |
| "loss": 3.1393, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002738055319254371, | |
| "loss": 3.1389, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.000273557759809711, | |
| "loss": 3.1366, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00027330998769398493, | |
| "loss": 3.1416, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "eval_accuracy": 0.41301242412444356, | |
| "eval_loss": 3.1622183322906494, | |
| "eval_runtime": 36.132, | |
| "eval_samples_per_second": 314.569, | |
| "eval_steps_per_second": 2.629, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002730622155782588, | |
| "loss": 3.1402, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002728144434625327, | |
| "loss": 3.1331, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002725666713468066, | |
| "loss": 3.1401, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002723188992310805, | |
| "loss": 3.1359, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0002720711271153544, | |
| "loss": 3.1345, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0002718233549996283, | |
| "loss": 3.1287, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0002715755828839022, | |
| "loss": 3.1352, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0002713278107681761, | |
| "loss": 3.1246, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00027108003865245004, | |
| "loss": 3.1382, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00027083226653672394, | |
| "loss": 3.1387, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "eval_accuracy": 0.4138569098887274, | |
| "eval_loss": 3.156071662902832, | |
| "eval_runtime": 37.0052, | |
| "eval_samples_per_second": 307.146, | |
| "eval_steps_per_second": 2.567, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00027058449442099783, | |
| "loss": 3.1329, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.000270339200026429, | |
| "loss": 3.1254, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00027009142791070294, | |
| "loss": 3.1315, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00026984365579497683, | |
| "loss": 3.1257, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00026959588367925073, | |
| "loss": 3.1171, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0002693481115635246, | |
| "loss": 3.128, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0002691003394477985, | |
| "loss": 3.1234, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0002688525673320724, | |
| "loss": 3.1354, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0002686047952163463, | |
| "loss": 3.1286, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0002683570231006202, | |
| "loss": 3.1192, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "eval_accuracy": 0.4142968183877596, | |
| "eval_loss": 3.149528741836548, | |
| "eval_runtime": 36.1498, | |
| "eval_samples_per_second": 314.414, | |
| "eval_steps_per_second": 2.628, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026810925098489415, | |
| "loss": 3.1155, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026786147886916805, | |
| "loss": 3.1268, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026761370675344194, | |
| "loss": 3.1216, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026736593463771584, | |
| "loss": 3.1262, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.000267120640243147, | |
| "loss": 3.1215, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026687286812742094, | |
| "loss": 3.1259, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026662509601169484, | |
| "loss": 3.1206, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026637732389596874, | |
| "loss": 3.1161, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00026612955178024263, | |
| "loss": 3.1127, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.0002658817796645165, | |
| "loss": 3.1221, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_accuracy": 0.4149895876775741, | |
| "eval_loss": 3.1430952548980713, | |
| "eval_runtime": 36.2908, | |
| "eval_samples_per_second": 313.192, | |
| "eval_steps_per_second": 2.618, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.0002656340075487905, | |
| "loss": 3.1211, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.0002653862354330643, | |
| "loss": 3.1159, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.0002651384633173382, | |
| "loss": 3.1211, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026489069120161216, | |
| "loss": 3.1168, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026464291908588606, | |
| "loss": 3.1166, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026439514697015995, | |
| "loss": 3.1168, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.0002641498525755911, | |
| "loss": 3.1165, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026390208045986506, | |
| "loss": 3.1115, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026365430834413895, | |
| "loss": 3.1124, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026340653622841285, | |
| "loss": 3.1136, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_accuracy": 0.4154509720245905, | |
| "eval_loss": 3.139134168624878, | |
| "eval_runtime": 36.9033, | |
| "eval_samples_per_second": 307.994, | |
| "eval_steps_per_second": 2.574, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00026315876411268674, | |
| "loss": 3.1096, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00026291099199696064, | |
| "loss": 3.1066, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00026266321988123453, | |
| "loss": 3.1172, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002624154477655085, | |
| "loss": 3.1209, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002621676756497823, | |
| "loss": 3.1057, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002619199035340562, | |
| "loss": 3.1105, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00026167213141833017, | |
| "loss": 3.1094, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002614268370237613, | |
| "loss": 3.1113, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002611790649080352, | |
| "loss": 3.1052, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002609312927923091, | |
| "loss": 3.106, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "eval_accuracy": 0.41599687422496434, | |
| "eval_loss": 3.1348280906677246, | |
| "eval_runtime": 36.2431, | |
| "eval_samples_per_second": 313.605, | |
| "eval_steps_per_second": 2.621, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00026068599839774027, | |
| "loss": 3.1111, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00026043822628201417, | |
| "loss": 3.1061, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0002601904541662881, | |
| "loss": 3.1147, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.000259942682050562, | |
| "loss": 3.1053, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0002596949099348359, | |
| "loss": 3.1002, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0002594471378191098, | |
| "loss": 3.113, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0002591993657033837, | |
| "loss": 3.1075, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025895159358765765, | |
| "loss": 3.109, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025870382147193154, | |
| "loss": 3.0998, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025845604935620544, | |
| "loss": 3.1023, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "eval_accuracy": 0.416494975344341, | |
| "eval_loss": 3.1311593055725098, | |
| "eval_runtime": 36.4415, | |
| "eval_samples_per_second": 311.897, | |
| "eval_steps_per_second": 2.607, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025820827724047933, | |
| "loss": 3.1055, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0002579605051247532, | |
| "loss": 3.1019, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0002577127330090271, | |
| "loss": 3.1029, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025746496089330107, | |
| "loss": 3.1028, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00025721718877757497, | |
| "loss": 3.1048, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025696941666184886, | |
| "loss": 3.1033, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025672164454612276, | |
| "loss": 3.109, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025647387243039665, | |
| "loss": 3.0972, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.0002562261003146706, | |
| "loss": 3.1006, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.0002559783281989445, | |
| "loss": 3.1062, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "eval_accuracy": 0.4171718109404897, | |
| "eval_loss": 3.1264827251434326, | |
| "eval_runtime": 37.3996, | |
| "eval_samples_per_second": 303.907, | |
| "eval_steps_per_second": 2.54, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025573055608321834, | |
| "loss": 3.1048, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.0002554827839674923, | |
| "loss": 3.0967, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025523748957292344, | |
| "loss": 3.0937, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025498971745719734, | |
| "loss": 3.1036, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00025474194534147123, | |
| "loss": 3.0981, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.0002544941732257452, | |
| "loss": 3.1013, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.0002542464011100191, | |
| "loss": 3.0977, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.000253998628994293, | |
| "loss": 3.0961, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025375085687856687, | |
| "loss": 3.0993, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025350308476284076, | |
| "loss": 3.1007, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_accuracy": 0.4176775325220543, | |
| "eval_loss": 3.1230428218841553, | |
| "eval_runtime": 36.1092, | |
| "eval_samples_per_second": 314.767, | |
| "eval_steps_per_second": 2.631, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025325531264711466, | |
| "loss": 3.0957, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0002530075405313886, | |
| "loss": 3.099, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0002527597684156625, | |
| "loss": 3.1044, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025251199629993635, | |
| "loss": 3.0985, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0002522642241842103, | |
| "loss": 3.1026, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025201892978964145, | |
| "loss": 3.0926, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025177115767391535, | |
| "loss": 3.0932, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00025152338555818924, | |
| "loss": 3.0912, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0002512756134424632, | |
| "loss": 3.0914, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0002510278413267371, | |
| "loss": 3.0979, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_accuracy": 0.41779807437848204, | |
| "eval_loss": 3.1200578212738037, | |
| "eval_runtime": 36.2694, | |
| "eval_samples_per_second": 313.377, | |
| "eval_steps_per_second": 2.619, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.000250780069211011, | |
| "loss": 3.088, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0002505322970952849, | |
| "loss": 3.0936, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00025028452497955877, | |
| "loss": 3.0931, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0002500392305849899, | |
| "loss": 3.09, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0002497914584692638, | |
| "loss": 3.0961, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024954368635353777, | |
| "loss": 3.0979, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024929591423781167, | |
| "loss": 3.0899, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024904814212208556, | |
| "loss": 3.0919, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024880037000635946, | |
| "loss": 3.0944, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024855259789063335, | |
| "loss": 3.0897, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "eval_accuracy": 0.4178625019224348, | |
| "eval_loss": 3.1168224811553955, | |
| "eval_runtime": 36.6661, | |
| "eval_samples_per_second": 309.987, | |
| "eval_steps_per_second": 2.591, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00024830730349606456, | |
| "loss": 3.091, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0002480595313803384, | |
| "loss": 3.0925, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024781175926461235, | |
| "loss": 3.0861, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024756398714888625, | |
| "loss": 3.0845, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024731621503316014, | |
| "loss": 3.0944, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024706844291743404, | |
| "loss": 3.083, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024682067080170793, | |
| "loss": 3.0803, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024657289868598183, | |
| "loss": 3.0899, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.0002463251265702558, | |
| "loss": 3.0872, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.0002460773544545297, | |
| "loss": 3.0863, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "eval_accuracy": 0.4188947281642584, | |
| "eval_loss": 3.1127541065216064, | |
| "eval_runtime": 36.4488, | |
| "eval_samples_per_second": 311.835, | |
| "eval_steps_per_second": 2.606, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024582958233880357, | |
| "loss": 3.0979, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024558181022307746, | |
| "loss": 3.0893, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00024533403810735136, | |
| "loss": 3.0898, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.0002450862659916253, | |
| "loss": 3.0875, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002448384938758992, | |
| "loss": 3.0822, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002445907217601731, | |
| "loss": 3.08, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.000244342949644447, | |
| "loss": 3.0835, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002440951775287209, | |
| "loss": 3.0913, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00024384740541299479, | |
| "loss": 3.0822, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002435996332972687, | |
| "loss": 3.0898, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "eval_accuracy": 0.4191060227976518, | |
| "eval_loss": 3.1097447872161865, | |
| "eval_runtime": 36.3549, | |
| "eval_samples_per_second": 312.64, | |
| "eval_steps_per_second": 2.613, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002433518611815426, | |
| "loss": 3.076, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002431040890658165, | |
| "loss": 3.0892, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00024285631695009042, | |
| "loss": 3.0842, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00024260854483436432, | |
| "loss": 3.0857, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0002423607727186382, | |
| "loss": 3.0845, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00024211300060291213, | |
| "loss": 3.0933, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024186522848718603, | |
| "loss": 3.0855, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024161745637145995, | |
| "loss": 3.0843, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024136968425573385, | |
| "loss": 3.097, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024112191214000774, | |
| "loss": 3.0825, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_accuracy": 0.41910394448978233, | |
| "eval_loss": 3.107358694076538, | |
| "eval_runtime": 36.6977, | |
| "eval_samples_per_second": 309.72, | |
| "eval_steps_per_second": 2.589, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024087414002428166, | |
| "loss": 3.0926, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024062636790855556, | |
| "loss": 3.079, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00024037859579282945, | |
| "loss": 3.0831, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.0002401333013982606, | |
| "loss": 3.0865, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00023988552928253453, | |
| "loss": 3.0748, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00023963775716680843, | |
| "loss": 3.073, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00023938998505108232, | |
| "loss": 3.0822, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00023914221293535625, | |
| "loss": 3.0817, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023889444081963014, | |
| "loss": 3.0803, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023864666870390404, | |
| "loss": 3.0808, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_accuracy": 0.41998445425713654, | |
| "eval_loss": 3.103720188140869, | |
| "eval_runtime": 36.4784, | |
| "eval_samples_per_second": 311.582, | |
| "eval_steps_per_second": 2.604, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023839889658817796, | |
| "loss": 3.0778, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023815112447245185, | |
| "loss": 3.0755, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023790335235672575, | |
| "loss": 3.0817, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023765558024099967, | |
| "loss": 3.0759, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023740780812527357, | |
| "loss": 3.0813, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023716003600954746, | |
| "loss": 3.08, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023691226389382138, | |
| "loss": 3.0746, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023666449177809528, | |
| "loss": 3.0786, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023641919738352644, | |
| "loss": 3.0857, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00023617142526780033, | |
| "loss": 3.0774, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_accuracy": 0.4197142742341089, | |
| "eval_loss": 3.1032252311706543, | |
| "eval_runtime": 36.4807, | |
| "eval_samples_per_second": 311.562, | |
| "eval_steps_per_second": 2.604, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023592365315207425, | |
| "loss": 3.0776, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023567588103634815, | |
| "loss": 3.0806, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023542810892062204, | |
| "loss": 3.0768, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023518033680489597, | |
| "loss": 3.0733, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023493256468916986, | |
| "loss": 3.0822, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023468479257344378, | |
| "loss": 3.0774, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023443702045771768, | |
| "loss": 3.0763, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00023418924834199157, | |
| "loss": 3.0773, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0002339414762262655, | |
| "loss": 3.0774, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0002336937041105394, | |
| "loss": 3.0652, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "eval_accuracy": 0.42021445366135496, | |
| "eval_loss": 3.098003387451172, | |
| "eval_runtime": 37.7336, | |
| "eval_samples_per_second": 301.217, | |
| "eval_steps_per_second": 2.518, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0002334459319948133, | |
| "loss": 3.0767, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0002331981598790872, | |
| "loss": 3.0743, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023295038776336108, | |
| "loss": 3.0693, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023270509336879226, | |
| "loss": 3.0829, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023245732125306616, | |
| "loss": 3.069, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.0002322120268584973, | |
| "loss": 3.0764, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.0002319642547427712, | |
| "loss": 3.0722, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023171648262704513, | |
| "loss": 3.0703, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023146871051131902, | |
| "loss": 3.0752, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023122093839559292, | |
| "loss": 3.0693, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "eval_accuracy": 0.4207465004759325, | |
| "eval_loss": 3.096764087677002, | |
| "eval_runtime": 36.5764, | |
| "eval_samples_per_second": 310.747, | |
| "eval_steps_per_second": 2.597, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023097316627986684, | |
| "loss": 3.0629, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023072539416414074, | |
| "loss": 3.0749, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023047762204841466, | |
| "loss": 3.0658, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00023022984993268855, | |
| "loss": 3.0806, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00022998207781696245, | |
| "loss": 3.0677, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022973430570123637, | |
| "loss": 3.0779, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022948653358551027, | |
| "loss": 3.0714, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022923876146978416, | |
| "loss": 3.0707, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022899098935405808, | |
| "loss": 3.0757, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022874321723833198, | |
| "loss": 3.0665, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_accuracy": 0.4209293915684435, | |
| "eval_loss": 3.0943939685821533, | |
| "eval_runtime": 36.661, | |
| "eval_samples_per_second": 310.03, | |
| "eval_steps_per_second": 2.591, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022849544512260588, | |
| "loss": 3.0677, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0002282476730068798, | |
| "loss": 3.0662, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0002279999008911537, | |
| "loss": 3.0649, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0002277521287754276, | |
| "loss": 3.067, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0002275043566597015, | |
| "loss": 3.0686, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0002272565845439754, | |
| "loss": 3.0726, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00022700881242824933, | |
| "loss": 3.0639, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022676104031252322, | |
| "loss": 3.0679, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.0002265132681967971, | |
| "loss": 3.0778, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022626549608107104, | |
| "loss": 3.0657, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "eval_accuracy": 0.4210423129626833, | |
| "eval_loss": 3.09199595451355, | |
| "eval_runtime": 36.5589, | |
| "eval_samples_per_second": 310.896, | |
| "eval_steps_per_second": 2.599, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.0002260177239653449, | |
| "loss": 3.0664, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.0002257699518496188, | |
| "loss": 3.0657, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022552217973389273, | |
| "loss": 3.0674, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022527440761816662, | |
| "loss": 3.0611, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022502663550244052, | |
| "loss": 3.0694, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022477886338671444, | |
| "loss": 3.0563, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022453109127098833, | |
| "loss": 3.0611, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022428331915526223, | |
| "loss": 3.0647, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00022403554703953615, | |
| "loss": 3.0645, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022378777492381005, | |
| "loss": 3.0608, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_accuracy": 0.4213305049872461, | |
| "eval_loss": 3.0911319255828857, | |
| "eval_runtime": 36.982, | |
| "eval_samples_per_second": 307.339, | |
| "eval_steps_per_second": 2.569, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022354248052924123, | |
| "loss": 3.0683, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0002232947084135151, | |
| "loss": 3.0719, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022304693629778905, | |
| "loss": 3.0614, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0002228016419032202, | |
| "loss": 3.0655, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0002225538697874941, | |
| "loss": 3.0682, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.000222306097671768, | |
| "loss": 3.0588, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022205832555604192, | |
| "loss": 3.0709, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0002218105534403158, | |
| "loss": 3.0687, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0002215627813245897, | |
| "loss": 3.0679, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022131500920886363, | |
| "loss": 3.0647, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_accuracy": 0.42134713145020164, | |
| "eval_loss": 3.089580774307251, | |
| "eval_runtime": 36.6102, | |
| "eval_samples_per_second": 310.46, | |
| "eval_steps_per_second": 2.595, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00022106723709313752, | |
| "loss": 3.054, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00022081946497741142, | |
| "loss": 3.0674, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00022057169286168534, | |
| "loss": 3.0678, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00022032392074595924, | |
| "loss": 3.06, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0002200786263513904, | |
| "loss": 3.0647, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0002198308542356643, | |
| "loss": 3.0619, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0002195830821199382, | |
| "loss": 3.0655, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0002193353100042121, | |
| "loss": 3.066, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.000219087537888486, | |
| "loss": 3.0615, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00021883976577275992, | |
| "loss": 3.0604, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "eval_accuracy": 0.4216789679400228, | |
| "eval_loss": 3.0860743522644043, | |
| "eval_runtime": 36.6297, | |
| "eval_samples_per_second": 310.295, | |
| "eval_steps_per_second": 2.594, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00021859199365703382, | |
| "loss": 3.0556, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00021834422154130771, | |
| "loss": 3.0664, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00021809644942558164, | |
| "loss": 3.0585, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021784867730985553, | |
| "loss": 3.0628, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021760090519412945, | |
| "loss": 3.0653, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021735313307840335, | |
| "loss": 3.0616, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021710536096267724, | |
| "loss": 3.0661, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021685758884695117, | |
| "loss": 3.0638, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021660981673122506, | |
| "loss": 3.0547, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021636204461549893, | |
| "loss": 3.0577, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "eval_accuracy": 0.422070382588768, | |
| "eval_loss": 3.084482431411743, | |
| "eval_runtime": 36.4437, | |
| "eval_samples_per_second": 311.879, | |
| "eval_steps_per_second": 2.607, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021611427249977288, | |
| "loss": 3.0627, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021586650038404675, | |
| "loss": 3.0604, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021561872826832064, | |
| "loss": 3.061, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021537095615259457, | |
| "loss": 3.0572, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00021512566175802575, | |
| "loss": 3.057, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021487788964229964, | |
| "loss": 3.0622, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021463011752657354, | |
| "loss": 3.0586, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021438234541084746, | |
| "loss": 3.0567, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021413457329512136, | |
| "loss": 3.0617, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021388680117939525, | |
| "loss": 3.0606, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "eval_accuracy": 0.4220225815077708, | |
| "eval_loss": 3.081350088119507, | |
| "eval_runtime": 39.7683, | |
| "eval_samples_per_second": 285.806, | |
| "eval_steps_per_second": 2.389, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021363902906366917, | |
| "loss": 3.0641, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021339125694794307, | |
| "loss": 3.0528, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021314348483221694, | |
| "loss": 3.0602, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.0002128957127164909, | |
| "loss": 3.0572, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021265041832192204, | |
| "loss": 3.0549, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021240512392735317, | |
| "loss": 3.0561, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00021215735181162712, | |
| "loss": 3.0499, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.000211909579695901, | |
| "loss": 3.0517, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0002116618075801749, | |
| "loss": 3.0606, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0002114140354644488, | |
| "loss": 3.0515, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "eval_accuracy": 0.4227111941818464, | |
| "eval_loss": 3.080108642578125, | |
| "eval_runtime": 36.5107, | |
| "eval_samples_per_second": 311.306, | |
| "eval_steps_per_second": 2.602, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0002111662633487227, | |
| "loss": 3.0534, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00021091849123299662, | |
| "loss": 3.0623, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00021067071911727052, | |
| "loss": 3.0529, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00021042294700154441, | |
| "loss": 3.0513, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00021017517488581834, | |
| "loss": 3.0526, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00020992740277009223, | |
| "loss": 3.0523, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00020967963065436613, | |
| "loss": 3.052, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0002094343362597973, | |
| "loss": 3.0554, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00020918656414407123, | |
| "loss": 3.0506, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0002089387920283451, | |
| "loss": 3.0527, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_accuracy": 0.42249435739413443, | |
| "eval_loss": 3.0771751403808594, | |
| "eval_runtime": 36.5791, | |
| "eval_samples_per_second": 310.724, | |
| "eval_steps_per_second": 2.597, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.000208691019912619, | |
| "loss": 3.0438, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020844324779689292, | |
| "loss": 3.0513, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020819547568116681, | |
| "loss": 3.0513, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0002079477035654407, | |
| "loss": 3.0597, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020769993144971463, | |
| "loss": 3.0448, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020745215933398853, | |
| "loss": 3.049, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020720438721826242, | |
| "loss": 3.0482, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020695661510253634, | |
| "loss": 3.0529, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020670884298681024, | |
| "loss": 3.0511, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020646354859224142, | |
| "loss": 3.0507, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_accuracy": 0.4227541458778149, | |
| "eval_loss": 3.075801134109497, | |
| "eval_runtime": 36.6131, | |
| "eval_samples_per_second": 310.436, | |
| "eval_steps_per_second": 2.595, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0002062157764765153, | |
| "loss": 3.0569, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00020596800436078924, | |
| "loss": 3.0587, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.0002057202322450631, | |
| "loss": 3.0528, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.000205472460129337, | |
| "loss": 3.0474, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020522468801361093, | |
| "loss": 3.0521, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020497691589788482, | |
| "loss": 3.0479, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020472914378215872, | |
| "loss": 3.0475, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020448137166643264, | |
| "loss": 3.0441, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020423359955070653, | |
| "loss": 3.0514, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020398582743498046, | |
| "loss": 3.0433, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_accuracy": 0.4233707105457498, | |
| "eval_loss": 3.0738978385925293, | |
| "eval_runtime": 36.3934, | |
| "eval_samples_per_second": 312.31, | |
| "eval_steps_per_second": 2.61, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020373805531925435, | |
| "loss": 3.0446, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020349028320352825, | |
| "loss": 3.055, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020324251108780217, | |
| "loss": 3.0532, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00020299473897207606, | |
| "loss": 3.0457, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020274944457750725, | |
| "loss": 3.0481, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020250167246178112, | |
| "loss": 3.0524, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020225390034605506, | |
| "loss": 3.0513, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020200612823032893, | |
| "loss": 3.0518, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020175835611460283, | |
| "loss": 3.0441, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020151058399887675, | |
| "loss": 3.0546, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "eval_accuracy": 0.42341574054958775, | |
| "eval_loss": 3.0717380046844482, | |
| "eval_runtime": 36.8684, | |
| "eval_samples_per_second": 308.286, | |
| "eval_steps_per_second": 2.577, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020126281188315065, | |
| "loss": 3.0486, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020101503976742454, | |
| "loss": 3.0451, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020076726765169846, | |
| "loss": 3.0468, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020051949553597236, | |
| "loss": 3.0454, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020027172342024625, | |
| "loss": 3.0439, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00020002642902567744, | |
| "loss": 3.0444, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019977865690995136, | |
| "loss": 3.0425, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019953088479422525, | |
| "loss": 3.0405, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019928311267849912, | |
| "loss": 3.0408, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019903534056277307, | |
| "loss": 3.0484, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "eval_accuracy": 0.42364643272309593, | |
| "eval_loss": 3.0696725845336914, | |
| "eval_runtime": 36.5687, | |
| "eval_samples_per_second": 310.813, | |
| "eval_steps_per_second": 2.598, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019878756844704694, | |
| "loss": 3.0496, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019853979633132084, | |
| "loss": 3.0508, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019829202421559476, | |
| "loss": 3.0486, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019804425209986865, | |
| "loss": 3.0481, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019779647998414255, | |
| "loss": 3.032, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019754870786841647, | |
| "loss": 3.0444, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00019730093575269037, | |
| "loss": 3.0446, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.0001970531636369643, | |
| "loss": 3.0434, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019680539152123818, | |
| "loss": 3.0394, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019656009712666937, | |
| "loss": 3.0441, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "eval_accuracy": 0.4235577582539997, | |
| "eval_loss": 3.0694241523742676, | |
| "eval_runtime": 36.751, | |
| "eval_samples_per_second": 309.271, | |
| "eval_steps_per_second": 2.585, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019631232501094326, | |
| "loss": 3.0452, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019606455289521713, | |
| "loss": 3.0439, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019581678077949108, | |
| "loss": 3.0425, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019556900866376495, | |
| "loss": 3.0391, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019532123654803884, | |
| "loss": 3.0443, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019507346443231277, | |
| "loss": 3.0485, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019482817003774395, | |
| "loss": 3.046, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019458039792201784, | |
| "loss": 3.0427, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019433262580629174, | |
| "loss": 3.0363, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00019408485369056566, | |
| "loss": 3.0292, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "eval_accuracy": 0.4242269733879605, | |
| "eval_loss": 3.0662310123443604, | |
| "eval_runtime": 36.6029, | |
| "eval_samples_per_second": 310.522, | |
| "eval_steps_per_second": 2.595, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019383708157483956, | |
| "loss": 3.0414, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019358930945911345, | |
| "loss": 3.0474, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019334153734338737, | |
| "loss": 3.0427, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019309376522766127, | |
| "loss": 3.0452, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001928459931119352, | |
| "loss": 3.0393, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001925982209962091, | |
| "loss": 3.0402, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019235044888048295, | |
| "loss": 3.0395, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019210515448591414, | |
| "loss": 3.0315, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019185738237018803, | |
| "loss": 3.0371, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019160961025446196, | |
| "loss": 3.0384, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_accuracy": 0.4244334186363252, | |
| "eval_loss": 3.064300775527954, | |
| "eval_runtime": 37.8137, | |
| "eval_samples_per_second": 300.579, | |
| "eval_steps_per_second": 2.512, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019136183813873585, | |
| "loss": 3.0465, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00019111406602300975, | |
| "loss": 3.0364, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00019086629390728367, | |
| "loss": 3.0378, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00019061852179155756, | |
| "loss": 3.0403, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00019037074967583149, | |
| "loss": 3.0372, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00019012297756010538, | |
| "loss": 3.0368, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018987520544437928, | |
| "loss": 3.0392, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.0001896274333286532, | |
| "loss": 3.0387, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.0001893796612129271, | |
| "loss": 3.0345, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018913188909720096, | |
| "loss": 3.0367, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "eval_accuracy": 0.42403784737184114, | |
| "eval_loss": 3.062938928604126, | |
| "eval_runtime": 39.8903, | |
| "eval_samples_per_second": 284.931, | |
| "eval_steps_per_second": 2.382, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018888659470263214, | |
| "loss": 3.0391, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018863882258690607, | |
| "loss": 3.0395, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018839105047117996, | |
| "loss": 3.0362, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00018814327835545386, | |
| "loss": 3.038, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018789550623972778, | |
| "loss": 3.0367, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018764773412400168, | |
| "loss": 3.0363, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018739996200827557, | |
| "loss": 3.0405, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.0001871521898925495, | |
| "loss": 3.0352, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.0001869044177768234, | |
| "loss": 3.0348, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018665664566109728, | |
| "loss": 3.0337, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "eval_accuracy": 0.4246010688044603, | |
| "eval_loss": 3.0621790885925293, | |
| "eval_runtime": 36.366, | |
| "eval_samples_per_second": 312.545, | |
| "eval_steps_per_second": 2.612, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.0001864088735453712, | |
| "loss": 3.0339, | |
| "step": 46100 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018616357915080236, | |
| "loss": 3.0373, | |
| "step": 46200 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018591580703507626, | |
| "loss": 3.0311, | |
| "step": 46300 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018566803491935015, | |
| "loss": 3.0392, | |
| "step": 46400 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018542026280362407, | |
| "loss": 3.0461, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00018517249068789797, | |
| "loss": 3.0335, | |
| "step": 46600 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018492471857217186, | |
| "loss": 3.0345, | |
| "step": 46700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001846769464564458, | |
| "loss": 3.0327, | |
| "step": 46800 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018442917434071968, | |
| "loss": 3.0338, | |
| "step": 46900 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018418140222499358, | |
| "loss": 3.0385, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "eval_accuracy": 0.4245407978762465, | |
| "eval_loss": 3.0599210262298584, | |
| "eval_runtime": 38.501, | |
| "eval_samples_per_second": 295.213, | |
| "eval_steps_per_second": 2.467, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001839336301092675, | |
| "loss": 3.0416, | |
| "step": 47100 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001836858579935414, | |
| "loss": 3.0355, | |
| "step": 47200 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018343808587781532, | |
| "loss": 3.0313, | |
| "step": 47300 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001831903137620892, | |
| "loss": 3.0264, | |
| "step": 47400 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018294254164636308, | |
| "loss": 3.0319, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018269476953063703, | |
| "loss": 3.0357, | |
| "step": 47600 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001824469974149109, | |
| "loss": 3.035, | |
| "step": 47700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.0001821992252991848, | |
| "loss": 3.0333, | |
| "step": 47800 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00018195145318345872, | |
| "loss": 3.0326, | |
| "step": 47900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001817036810677326, | |
| "loss": 3.0319, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_accuracy": 0.4249543811422657, | |
| "eval_loss": 3.057422399520874, | |
| "eval_runtime": 36.2469, | |
| "eval_samples_per_second": 313.572, | |
| "eval_steps_per_second": 2.621, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001814559089520065, | |
| "loss": 3.0296, | |
| "step": 48100 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00018120813683628043, | |
| "loss": 3.0391, | |
| "step": 48200 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00018096036472055432, | |
| "loss": 3.0309, | |
| "step": 48300 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00018071259260482822, | |
| "loss": 3.0335, | |
| "step": 48400 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00018046482048910214, | |
| "loss": 3.0328, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00018021704837337604, | |
| "loss": 3.0353, | |
| "step": 48600 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00017996927625764996, | |
| "loss": 3.0348, | |
| "step": 48700 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00017972150414192385, | |
| "loss": 3.0222, | |
| "step": 48800 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00017947373202619775, | |
| "loss": 3.0283, | |
| "step": 48900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00017922595991047167, | |
| "loss": 3.0255, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_accuracy": 0.4249474534493676, | |
| "eval_loss": 3.05733585357666, | |
| "eval_runtime": 36.7665, | |
| "eval_samples_per_second": 309.14, | |
| "eval_steps_per_second": 2.584, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00017897818779474557, | |
| "loss": 3.028, | |
| "step": 49100 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017873289340017672, | |
| "loss": 3.0368, | |
| "step": 49200 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017848512128445062, | |
| "loss": 3.0396, | |
| "step": 49300 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017823734916872454, | |
| "loss": 3.0283, | |
| "step": 49400 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017798957705299844, | |
| "loss": 3.0278, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017774180493727233, | |
| "loss": 3.0374, | |
| "step": 49600 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017749403282154625, | |
| "loss": 3.0317, | |
| "step": 49700 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017724626070582015, | |
| "loss": 3.0318, | |
| "step": 49800 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017699848859009404, | |
| "loss": 3.0324, | |
| "step": 49900 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017675071647436797, | |
| "loss": 3.021, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "eval_accuracy": 0.4252966091714341, | |
| "eval_loss": 3.055666923522949, | |
| "eval_runtime": 36.7556, | |
| "eval_samples_per_second": 309.232, | |
| "eval_steps_per_second": 2.585, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017650294435864186, | |
| "loss": 3.0338, | |
| "step": 50100 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00017625517224291576, | |
| "loss": 3.0348, | |
| "step": 50200 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.0001760098778483469, | |
| "loss": 3.0247, | |
| "step": 50300 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017576210573262086, | |
| "loss": 3.0318, | |
| "step": 50400 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017551433361689473, | |
| "loss": 3.023, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017526656150116863, | |
| "loss": 3.0271, | |
| "step": 50600 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017501878938544255, | |
| "loss": 3.0305, | |
| "step": 50700 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017477101726971644, | |
| "loss": 3.0332, | |
| "step": 50800 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017452324515399034, | |
| "loss": 3.0292, | |
| "step": 50900 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017427547303826426, | |
| "loss": 3.0305, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "eval_accuracy": 0.42537212102402383, | |
| "eval_loss": 3.052976608276367, | |
| "eval_runtime": 36.1546, | |
| "eval_samples_per_second": 314.372, | |
| "eval_steps_per_second": 2.628, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017402770092253816, | |
| "loss": 3.0373, | |
| "step": 51100 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017377992880681205, | |
| "loss": 3.0281, | |
| "step": 51200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017353215669108597, | |
| "loss": 3.0308, | |
| "step": 51300 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017328438457535987, | |
| "loss": 3.0293, | |
| "step": 51400 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00017303661245963376, | |
| "loss": 3.0236, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.0001727888403439077, | |
| "loss": 3.0272, | |
| "step": 51600 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017254354594933887, | |
| "loss": 3.0267, | |
| "step": 51700 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017229577383361274, | |
| "loss": 3.0305, | |
| "step": 51800 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017204800171788663, | |
| "loss": 3.0243, | |
| "step": 51900 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017180022960216056, | |
| "loss": 3.0248, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "eval_accuracy": 0.42570326474455517, | |
| "eval_loss": 3.0528042316436768, | |
| "eval_runtime": 36.159, | |
| "eval_samples_per_second": 314.334, | |
| "eval_steps_per_second": 2.627, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017155245748643445, | |
| "loss": 3.031, | |
| "step": 52100 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017130468537070835, | |
| "loss": 3.0278, | |
| "step": 52200 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017105691325498227, | |
| "loss": 3.0219, | |
| "step": 52300 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017080914113925616, | |
| "loss": 3.0313, | |
| "step": 52400 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017056136902353009, | |
| "loss": 3.0308, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017031359690780398, | |
| "loss": 3.0214, | |
| "step": 52600 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00017006582479207788, | |
| "loss": 3.0325, | |
| "step": 52700 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001698180526763518, | |
| "loss": 3.0235, | |
| "step": 52800 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001695702805606257, | |
| "loss": 3.0314, | |
| "step": 52900 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016932498616605688, | |
| "loss": 3.0269, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "eval_accuracy": 0.42607943846892443, | |
| "eval_loss": 3.049508571624756, | |
| "eval_runtime": 38.1346, | |
| "eval_samples_per_second": 298.05, | |
| "eval_steps_per_second": 2.491, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016907721405033074, | |
| "loss": 3.0216, | |
| "step": 53100 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001688294419346047, | |
| "loss": 3.023, | |
| "step": 53200 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016858166981887856, | |
| "loss": 3.0213, | |
| "step": 53300 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016833389770315246, | |
| "loss": 3.0316, | |
| "step": 53400 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016808612558742638, | |
| "loss": 3.0297, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016783835347170028, | |
| "loss": 3.0295, | |
| "step": 53600 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00016759058135597417, | |
| "loss": 3.0267, | |
| "step": 53700 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001673428092402481, | |
| "loss": 3.0158, | |
| "step": 53800 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.000167095037124522, | |
| "loss": 3.0209, | |
| "step": 53900 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016684726500879588, | |
| "loss": 3.0136, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "eval_accuracy": 0.425942962918831, | |
| "eval_loss": 3.048839569091797, | |
| "eval_runtime": 38.3696, | |
| "eval_samples_per_second": 296.224, | |
| "eval_steps_per_second": 2.476, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001665994928930698, | |
| "loss": 3.0268, | |
| "step": 54100 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001663517207773437, | |
| "loss": 3.018, | |
| "step": 54200 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001661039486616176, | |
| "loss": 3.026, | |
| "step": 54300 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016585617654589152, | |
| "loss": 3.0194, | |
| "step": 54400 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016560840443016541, | |
| "loss": 3.0255, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016536063231443934, | |
| "loss": 3.0169, | |
| "step": 54600 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016511286019871323, | |
| "loss": 3.0285, | |
| "step": 54700 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016486508808298713, | |
| "loss": 3.0255, | |
| "step": 54800 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016461731596726105, | |
| "loss": 3.0233, | |
| "step": 54900 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016437202157269218, | |
| "loss": 3.0156, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "eval_accuracy": 0.42623946817487157, | |
| "eval_loss": 3.0467984676361084, | |
| "eval_runtime": 36.8565, | |
| "eval_samples_per_second": 308.385, | |
| "eval_steps_per_second": 2.578, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00016412672717812336, | |
| "loss": 3.0207, | |
| "step": 55100 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016387895506239728, | |
| "loss": 3.025, | |
| "step": 55200 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016363118294667118, | |
| "loss": 3.019, | |
| "step": 55300 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016338341083094507, | |
| "loss": 3.013, | |
| "step": 55400 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.000163135638715219, | |
| "loss": 3.0253, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001628878665994929, | |
| "loss": 3.0123, | |
| "step": 55600 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016264009448376676, | |
| "loss": 3.0203, | |
| "step": 55700 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016239232236804068, | |
| "loss": 3.0326, | |
| "step": 55800 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016214455025231458, | |
| "loss": 3.0228, | |
| "step": 55900 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00016189677813658847, | |
| "loss": 3.022, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "eval_accuracy": 0.42677844268234727, | |
| "eval_loss": 3.0453531742095947, | |
| "eval_runtime": 36.3795, | |
| "eval_samples_per_second": 312.429, | |
| "eval_steps_per_second": 2.611, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001616490060208624, | |
| "loss": 3.0227, | |
| "step": 56100 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001614012339051363, | |
| "loss": 3.0261, | |
| "step": 56200 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001611534617894102, | |
| "loss": 3.0218, | |
| "step": 56300 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001609056896736841, | |
| "loss": 3.0216, | |
| "step": 56400 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.000160657917557958, | |
| "loss": 3.0183, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00016041014544223192, | |
| "loss": 3.0174, | |
| "step": 56600 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00016016237332650582, | |
| "loss": 3.0241, | |
| "step": 56700 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015991460121077972, | |
| "loss": 3.0181, | |
| "step": 56800 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015966682909505364, | |
| "loss": 3.0194, | |
| "step": 56900 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015941905697932753, | |
| "loss": 3.0193, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "eval_accuracy": 0.42686365330499443, | |
| "eval_loss": 3.044196367263794, | |
| "eval_runtime": 36.1843, | |
| "eval_samples_per_second": 314.114, | |
| "eval_steps_per_second": 2.625, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015917128486360143, | |
| "loss": 3.0154, | |
| "step": 57100 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015892351274787535, | |
| "loss": 3.0283, | |
| "step": 57200 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001586782183533065, | |
| "loss": 3.0184, | |
| "step": 57300 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001584304462375804, | |
| "loss": 3.0122, | |
| "step": 57400 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001581826741218543, | |
| "loss": 3.0183, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015793490200612822, | |
| "loss": 3.0212, | |
| "step": 57600 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00015768712989040211, | |
| "loss": 3.0084, | |
| "step": 57700 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.000157439357774676, | |
| "loss": 3.0144, | |
| "step": 57800 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001571940633801072, | |
| "loss": 3.0292, | |
| "step": 57900 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00015694629126438111, | |
| "loss": 3.0222, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_accuracy": 0.42703615285815827, | |
| "eval_loss": 3.0416929721832275, | |
| "eval_runtime": 36.2277, | |
| "eval_samples_per_second": 313.737, | |
| "eval_steps_per_second": 2.622, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.000156698519148655, | |
| "loss": 3.0173, | |
| "step": 58100 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00015645074703292888, | |
| "loss": 3.0236, | |
| "step": 58200 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00015620297491720283, | |
| "loss": 3.0151, | |
| "step": 58300 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001559552028014767, | |
| "loss": 3.0253, | |
| "step": 58400 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001557074306857506, | |
| "loss": 3.0175, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001554596585700245, | |
| "loss": 3.0128, | |
| "step": 58600 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001552118864542984, | |
| "loss": 3.0129, | |
| "step": 58700 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001549641143385723, | |
| "loss": 3.0188, | |
| "step": 58800 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015471634222284623, | |
| "loss": 3.0085, | |
| "step": 58900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015446857010712012, | |
| "loss": 3.0111, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "eval_accuracy": 0.42760214536793667, | |
| "eval_loss": 3.039332389831543, | |
| "eval_runtime": 36.2138, | |
| "eval_samples_per_second": 313.859, | |
| "eval_steps_per_second": 2.623, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015422079799139402, | |
| "loss": 2.9987, | |
| "step": 59100 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001539755035968252, | |
| "loss": 3.0151, | |
| "step": 59200 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015372773148109912, | |
| "loss": 3.011, | |
| "step": 59300 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015347995936537302, | |
| "loss": 3.0117, | |
| "step": 59400 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015323218724964689, | |
| "loss": 3.0211, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015298441513392083, | |
| "loss": 3.0137, | |
| "step": 59600 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001527366430181947, | |
| "loss": 3.0184, | |
| "step": 59700 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001524888709024686, | |
| "loss": 3.009, | |
| "step": 59800 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015224109878674252, | |
| "loss": 3.0189, | |
| "step": 59900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00015199332667101642, | |
| "loss": 3.0148, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "eval_accuracy": 0.4273499773464442, | |
| "eval_loss": 3.0384342670440674, | |
| "eval_runtime": 37.3986, | |
| "eval_samples_per_second": 303.915, | |
| "eval_steps_per_second": 2.54, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015174555455529034, | |
| "loss": 3.0113, | |
| "step": 60100 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015149778243956423, | |
| "loss": 3.0134, | |
| "step": 60200 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015125001032383813, | |
| "loss": 3.0105, | |
| "step": 60300 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.0001510047159292693, | |
| "loss": 3.0107, | |
| "step": 60400 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.0001507569438135432, | |
| "loss": 3.0143, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015050917169781713, | |
| "loss": 3.0049, | |
| "step": 60600 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015026139958209102, | |
| "loss": 3.0171, | |
| "step": 60700 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00015001362746636495, | |
| "loss": 3.0145, | |
| "step": 60800 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00014976585535063884, | |
| "loss": 3.0095, | |
| "step": 60900 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.0001495180832349127, | |
| "loss": 3.0077, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "eval_accuracy": 0.4275841333664015, | |
| "eval_loss": 3.0363619327545166, | |
| "eval_runtime": 36.6829, | |
| "eval_samples_per_second": 309.845, | |
| "eval_steps_per_second": 2.59, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00014927031111918663, | |
| "loss": 3.0168, | |
| "step": 61100 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00014902253900346053, | |
| "loss": 3.0083, | |
| "step": 61200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014877476688773445, | |
| "loss": 3.0024, | |
| "step": 61300 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014852699477200835, | |
| "loss": 3.018, | |
| "step": 61400 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014827922265628224, | |
| "loss": 3.0216, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014803145054055616, | |
| "loss": 3.0105, | |
| "step": 61600 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014778615614598732, | |
| "loss": 3.0053, | |
| "step": 61700 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014753838403026121, | |
| "loss": 3.0133, | |
| "step": 61800 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014729061191453514, | |
| "loss": 3.0113, | |
| "step": 61900 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014704283979880903, | |
| "loss": 3.0167, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_accuracy": 0.42764232598674595, | |
| "eval_loss": 3.0357508659362793, | |
| "eval_runtime": 36.3375, | |
| "eval_samples_per_second": 312.79, | |
| "eval_steps_per_second": 2.614, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014679506768308293, | |
| "loss": 3.0026, | |
| "step": 62100 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014654729556735685, | |
| "loss": 3.0072, | |
| "step": 62200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014629952345163074, | |
| "loss": 3.0118, | |
| "step": 62300 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00014605175133590464, | |
| "loss": 3.0098, | |
| "step": 62400 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014580397922017853, | |
| "loss": 3.0079, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014555620710445246, | |
| "loss": 3.008, | |
| "step": 62600 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014530843498872635, | |
| "loss": 3.0051, | |
| "step": 62700 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014506066287300025, | |
| "loss": 3.0086, | |
| "step": 62800 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014481289075727417, | |
| "loss": 3.0124, | |
| "step": 62900 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014456511864154807, | |
| "loss": 3.0049, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "eval_accuracy": 0.42800118047886987, | |
| "eval_loss": 3.0342743396759033, | |
| "eval_runtime": 36.3466, | |
| "eval_samples_per_second": 312.711, | |
| "eval_steps_per_second": 2.614, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014431982424697922, | |
| "loss": 3.017, | |
| "step": 63100 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014407205213125314, | |
| "loss": 3.0057, | |
| "step": 63200 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014382428001552704, | |
| "loss": 3.0143, | |
| "step": 63300 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.0001435789856209582, | |
| "loss": 3.0131, | |
| "step": 63400 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00014333121350523212, | |
| "loss": 3.0048, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.000143083441389506, | |
| "loss": 3.0061, | |
| "step": 63600 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014283566927377993, | |
| "loss": 3.0144, | |
| "step": 63700 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.0001425878971580538, | |
| "loss": 3.0082, | |
| "step": 63800 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014234012504232772, | |
| "loss": 3.0089, | |
| "step": 63900 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014209235292660162, | |
| "loss": 3.016, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "eval_accuracy": 0.4281286500281957, | |
| "eval_loss": 3.032212257385254, | |
| "eval_runtime": 36.9386, | |
| "eval_samples_per_second": 307.699, | |
| "eval_steps_per_second": 2.572, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014184458081087552, | |
| "loss": 3.0087, | |
| "step": 64100 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014159680869514944, | |
| "loss": 3.0083, | |
| "step": 64200 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014134903657942333, | |
| "loss": 3.0078, | |
| "step": 64300 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014110126446369726, | |
| "loss": 3.0104, | |
| "step": 64400 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014085349234797115, | |
| "loss": 3.0069, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014060572023224505, | |
| "loss": 2.9968, | |
| "step": 64600 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014035794811651897, | |
| "loss": 3.0041, | |
| "step": 64700 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00014011017600079286, | |
| "loss": 3.0097, | |
| "step": 64800 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013986240388506676, | |
| "loss": 3.0105, | |
| "step": 64900 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013961463176934068, | |
| "loss": 3.0103, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "eval_accuracy": 0.4285138297533326, | |
| "eval_loss": 3.0296883583068848, | |
| "eval_runtime": 36.5876, | |
| "eval_samples_per_second": 310.652, | |
| "eval_steps_per_second": 2.597, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013936685965361458, | |
| "loss": 2.9976, | |
| "step": 65100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013911908753788847, | |
| "loss": 3.0013, | |
| "step": 65200 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013887131542216237, | |
| "loss": 3.0069, | |
| "step": 65300 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.0001386235433064363, | |
| "loss": 3.0034, | |
| "step": 65400 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013837577119071018, | |
| "loss": 3.0107, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013812799907498408, | |
| "loss": 3.0011, | |
| "step": 65600 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.000137880226959258, | |
| "loss": 3.0037, | |
| "step": 65700 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.0001376324548435319, | |
| "loss": 2.9999, | |
| "step": 65800 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.0001373846827278058, | |
| "loss": 3.0036, | |
| "step": 65900 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00013713691061207971, | |
| "loss": 3.0066, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "eval_accuracy": 0.42835587835525485, | |
| "eval_loss": 3.0290277004241943, | |
| "eval_runtime": 36.1991, | |
| "eval_samples_per_second": 313.985, | |
| "eval_steps_per_second": 2.624, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001368891384963536, | |
| "loss": 3.0091, | |
| "step": 66100 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013664136638062753, | |
| "loss": 3.0024, | |
| "step": 66200 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001363935942649014, | |
| "loss": 2.9975, | |
| "step": 66300 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013614829987033258, | |
| "loss": 3.0005, | |
| "step": 66400 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013590052775460648, | |
| "loss": 3.0091, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013565275563888037, | |
| "loss": 3.0056, | |
| "step": 66600 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001354049835231543, | |
| "loss": 3.0058, | |
| "step": 66700 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001351572114074282, | |
| "loss": 3.0075, | |
| "step": 66800 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001349094392917021, | |
| "loss": 3.0081, | |
| "step": 66900 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.000134661667175976, | |
| "loss": 2.9958, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "eval_accuracy": 0.42845632990227794, | |
| "eval_loss": 3.0280661582946777, | |
| "eval_runtime": 36.2004, | |
| "eval_samples_per_second": 313.975, | |
| "eval_steps_per_second": 2.624, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001344138950602499, | |
| "loss": 2.9959, | |
| "step": 67100 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013416612294452383, | |
| "loss": 3.0026, | |
| "step": 67200 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00013391835082879772, | |
| "loss": 2.9972, | |
| "step": 67300 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013367057871307162, | |
| "loss": 3.0, | |
| "step": 67400 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013342280659734554, | |
| "loss": 3.0026, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001331750344816194, | |
| "loss": 3.0058, | |
| "step": 67600 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013292726236589333, | |
| "loss": 3.0031, | |
| "step": 67700 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013268196797132449, | |
| "loss": 2.9992, | |
| "step": 67800 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001324341958555984, | |
| "loss": 3.0035, | |
| "step": 67900 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001321864237398723, | |
| "loss": 3.0062, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "eval_accuracy": 0.428811027778663, | |
| "eval_loss": 3.0265986919403076, | |
| "eval_runtime": 36.5821, | |
| "eval_samples_per_second": 310.698, | |
| "eval_steps_per_second": 2.597, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001319386516241462, | |
| "loss": 2.9966, | |
| "step": 68100 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013169087950842012, | |
| "loss": 3.0086, | |
| "step": 68200 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013144310739269402, | |
| "loss": 3.0005, | |
| "step": 68300 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001311953352769679, | |
| "loss": 2.9978, | |
| "step": 68400 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.00013094756316124183, | |
| "loss": 3.0039, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00013069979104551573, | |
| "loss": 3.0034, | |
| "step": 68600 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00013045201892978962, | |
| "loss": 2.997, | |
| "step": 68700 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00013020424681406355, | |
| "loss": 2.9994, | |
| "step": 68800 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012995647469833744, | |
| "loss": 2.9964, | |
| "step": 68900 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012970870258261134, | |
| "loss": 2.9985, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "eval_accuracy": 0.42893919009727866, | |
| "eval_loss": 3.0245213508605957, | |
| "eval_runtime": 36.3523, | |
| "eval_samples_per_second": 312.662, | |
| "eval_steps_per_second": 2.613, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012946093046688523, | |
| "loss": 2.9977, | |
| "step": 69100 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012921315835115915, | |
| "loss": 3.0076, | |
| "step": 69200 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012896538623543305, | |
| "loss": 3.0036, | |
| "step": 69300 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012871761411970695, | |
| "loss": 2.9946, | |
| "step": 69400 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012846984200398087, | |
| "loss": 3.0045, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012822206988825476, | |
| "loss": 3.0005, | |
| "step": 69600 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.00012797429777252866, | |
| "loss": 3.0016, | |
| "step": 69700 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012772900337795984, | |
| "loss": 2.996, | |
| "step": 69800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.000127483708983391, | |
| "loss": 3.0003, | |
| "step": 69900 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.0001272359368676649, | |
| "loss": 3.0031, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "eval_accuracy": 0.42916849673220725, | |
| "eval_loss": 3.0224156379699707, | |
| "eval_runtime": 36.4457, | |
| "eval_samples_per_second": 311.861, | |
| "eval_steps_per_second": 2.607, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012698816475193881, | |
| "loss": 3.001, | |
| "step": 70100 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.0001267403926362127, | |
| "loss": 3.0014, | |
| "step": 70200 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012649262052048663, | |
| "loss": 3.0013, | |
| "step": 70300 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.0001262448484047605, | |
| "loss": 2.9942, | |
| "step": 70400 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012599707628903442, | |
| "loss": 3.0038, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012574930417330832, | |
| "loss": 2.9929, | |
| "step": 70600 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.0001255015320575822, | |
| "loss": 2.9973, | |
| "step": 70700 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012525375994185614, | |
| "loss": 2.9922, | |
| "step": 70800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.00012500598782613003, | |
| "loss": 2.9953, | |
| "step": 70900 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012476069343156121, | |
| "loss": 2.9894, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "eval_accuracy": 0.4295162669156941, | |
| "eval_loss": 3.0213873386383057, | |
| "eval_runtime": 36.4185, | |
| "eval_samples_per_second": 312.094, | |
| "eval_steps_per_second": 2.609, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.0001245129213158351, | |
| "loss": 2.9974, | |
| "step": 71100 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.000124265149200109, | |
| "loss": 3.0013, | |
| "step": 71200 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012401737708438293, | |
| "loss": 3.0012, | |
| "step": 71300 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012376960496865682, | |
| "loss": 2.9985, | |
| "step": 71400 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012352183285293072, | |
| "loss": 3.0029, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012327406073720464, | |
| "loss": 2.9952, | |
| "step": 71600 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012302628862147853, | |
| "loss": 3.0009, | |
| "step": 71700 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012277851650575243, | |
| "loss": 2.993, | |
| "step": 71800 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012253074439002633, | |
| "loss": 2.994, | |
| "step": 71900 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012228297227430025, | |
| "loss": 2.9929, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "eval_accuracy": 0.4295869293832552, | |
| "eval_loss": 3.0192549228668213, | |
| "eval_runtime": 36.4669, | |
| "eval_samples_per_second": 311.68, | |
| "eval_steps_per_second": 2.605, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00012203520015857414, | |
| "loss": 2.996, | |
| "step": 72100 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012178742804284805, | |
| "loss": 3.002, | |
| "step": 72200 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012153965592712196, | |
| "loss": 3.0022, | |
| "step": 72300 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012129188381139587, | |
| "loss": 2.9964, | |
| "step": 72400 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012104411169566975, | |
| "loss": 2.9914, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012079633957994366, | |
| "loss": 2.9934, | |
| "step": 72600 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012054856746421757, | |
| "loss": 2.9932, | |
| "step": 72700 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012030079534849146, | |
| "loss": 2.9928, | |
| "step": 72800 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00012005550095392263, | |
| "loss": 2.9955, | |
| "step": 72900 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00011980772883819654, | |
| "loss": 2.9904, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "eval_accuracy": 0.4296423509264404, | |
| "eval_loss": 3.0176117420196533, | |
| "eval_runtime": 36.4739, | |
| "eval_samples_per_second": 311.62, | |
| "eval_steps_per_second": 2.605, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00011955995672247044, | |
| "loss": 3.0003, | |
| "step": 73100 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00011931218460674435, | |
| "loss": 2.9891, | |
| "step": 73200 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.00011906441249101825, | |
| "loss": 2.9927, | |
| "step": 73300 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011881664037529216, | |
| "loss": 2.9979, | |
| "step": 73400 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011856886825956606, | |
| "loss": 2.9961, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011832109614383997, | |
| "loss": 2.9888, | |
| "step": 73600 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011807332402811386, | |
| "loss": 2.9908, | |
| "step": 73700 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011782555191238777, | |
| "loss": 2.9999, | |
| "step": 73800 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011757777979666167, | |
| "loss": 2.997, | |
| "step": 73900 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011733000768093558, | |
| "loss": 2.9989, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "eval_accuracy": 0.43006701850109663, | |
| "eval_loss": 3.0170629024505615, | |
| "eval_runtime": 36.6925, | |
| "eval_samples_per_second": 309.764, | |
| "eval_steps_per_second": 2.589, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011708223556520948, | |
| "loss": 2.9978, | |
| "step": 74100 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011683446344948338, | |
| "loss": 2.9962, | |
| "step": 74200 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.00011658669133375729, | |
| "loss": 3.0012, | |
| "step": 74300 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.0001163389192180312, | |
| "loss": 2.9862, | |
| "step": 74400 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 0.0001160911471023051, | |
| "loss": 2.9931, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.000115843374986579, | |
| "loss": 2.9931, | |
| "step": 74600 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011559808059201017, | |
| "loss": 2.9819, | |
| "step": 74700 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011535030847628408, | |
| "loss": 2.9885, | |
| "step": 74800 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011510253636055796, | |
| "loss": 2.9905, | |
| "step": 74900 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011485476424483187, | |
| "loss": 2.9959, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "eval_accuracy": 0.4301397592765272, | |
| "eval_loss": 3.015258550643921, | |
| "eval_runtime": 36.8121, | |
| "eval_samples_per_second": 308.757, | |
| "eval_steps_per_second": 2.581, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011460699212910578, | |
| "loss": 2.9742, | |
| "step": 75100 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011435922001337967, | |
| "loss": 2.9913, | |
| "step": 75200 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011411144789765358, | |
| "loss": 2.9913, | |
| "step": 75300 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00011386367578192749, | |
| "loss": 2.9888, | |
| "step": 75400 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.0001136159036662014, | |
| "loss": 2.9901, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.0001133681315504753, | |
| "loss": 2.9909, | |
| "step": 75600 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.0001131203594347492, | |
| "loss": 2.9901, | |
| "step": 75700 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011287506504018037, | |
| "loss": 2.984, | |
| "step": 75800 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011262729292445427, | |
| "loss": 2.9969, | |
| "step": 75900 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011237952080872818, | |
| "loss": 2.9847, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "eval_accuracy": 0.43058243885271863, | |
| "eval_loss": 3.0142199993133545, | |
| "eval_runtime": 36.2994, | |
| "eval_samples_per_second": 313.118, | |
| "eval_steps_per_second": 2.617, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011213174869300209, | |
| "loss": 2.994, | |
| "step": 76100 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.000111883976577276, | |
| "loss": 2.9892, | |
| "step": 76200 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011163620446154988, | |
| "loss": 2.9926, | |
| "step": 76300 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011138843234582379, | |
| "loss": 2.989, | |
| "step": 76400 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.0001111406602300977, | |
| "loss": 2.992, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011089288811437159, | |
| "loss": 2.9784, | |
| "step": 76600 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.0001106451159986455, | |
| "loss": 2.9865, | |
| "step": 76700 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011039734388291941, | |
| "loss": 2.9886, | |
| "step": 76800 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00011014957176719332, | |
| "loss": 2.9855, | |
| "step": 76900 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.00010990179965146721, | |
| "loss": 2.9892, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "eval_accuracy": 0.4308013539483, | |
| "eval_loss": 3.0127484798431396, | |
| "eval_runtime": 37.0779, | |
| "eval_samples_per_second": 306.544, | |
| "eval_steps_per_second": 2.562, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010965402753574112, | |
| "loss": 2.986, | |
| "step": 77100 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010940625542001503, | |
| "loss": 2.9875, | |
| "step": 77200 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010916096102544618, | |
| "loss": 2.9868, | |
| "step": 77300 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.0001089131889097201, | |
| "loss": 2.9904, | |
| "step": 77400 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.000108665416793994, | |
| "loss": 2.9861, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010841764467826791, | |
| "loss": 2.9866, | |
| "step": 77600 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010816987256254179, | |
| "loss": 2.9794, | |
| "step": 77700 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.0001079221004468157, | |
| "loss": 2.9848, | |
| "step": 77800 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010767432833108961, | |
| "loss": 2.982, | |
| "step": 77900 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.0001074265562153635, | |
| "loss": 2.9924, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "eval_accuracy": 0.4309537631920592, | |
| "eval_loss": 3.010981321334839, | |
| "eval_runtime": 36.5579, | |
| "eval_samples_per_second": 310.904, | |
| "eval_steps_per_second": 2.599, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010717878409963741, | |
| "loss": 2.9795, | |
| "step": 78100 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00010693101198391132, | |
| "loss": 2.9867, | |
| "step": 78200 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010668323986818523, | |
| "loss": 2.9761, | |
| "step": 78300 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010643546775245913, | |
| "loss": 2.9887, | |
| "step": 78400 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010618769563673304, | |
| "loss": 2.9941, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010593992352100695, | |
| "loss": 2.9888, | |
| "step": 78600 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010569215140528084, | |
| "loss": 2.9859, | |
| "step": 78700 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010544437928955474, | |
| "loss": 2.9828, | |
| "step": 78800 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010519660717382864, | |
| "loss": 2.9819, | |
| "step": 78900 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.0001049513127792598, | |
| "loss": 2.991, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "eval_accuracy": 0.4311969252127841, | |
| "eval_loss": 3.009610652923584, | |
| "eval_runtime": 36.4211, | |
| "eval_samples_per_second": 312.071, | |
| "eval_steps_per_second": 2.608, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010470354066353371, | |
| "loss": 2.9902, | |
| "step": 79100 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010445576854780762, | |
| "loss": 2.9855, | |
| "step": 79200 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010420799643208153, | |
| "loss": 2.9842, | |
| "step": 79300 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 0.00010396022431635542, | |
| "loss": 2.9881, | |
| "step": 79400 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010371245220062933, | |
| "loss": 2.9834, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010346468008490324, | |
| "loss": 2.9834, | |
| "step": 79600 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010321690796917713, | |
| "loss": 2.9728, | |
| "step": 79700 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010296913585345104, | |
| "loss": 2.988, | |
| "step": 79800 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010272136373772495, | |
| "loss": 2.988, | |
| "step": 79900 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010247606934315612, | |
| "loss": 2.9824, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "eval_accuracy": 0.4311297265916721, | |
| "eval_loss": 3.0079753398895264, | |
| "eval_runtime": 37.4243, | |
| "eval_samples_per_second": 303.707, | |
| "eval_steps_per_second": 2.538, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010222829722743002, | |
| "loss": 2.9804, | |
| "step": 80100 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010198052511170393, | |
| "loss": 2.9889, | |
| "step": 80200 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010173275299597783, | |
| "loss": 2.9813, | |
| "step": 80300 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010148498088025172, | |
| "loss": 2.9886, | |
| "step": 80400 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010123720876452562, | |
| "loss": 2.9851, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.00010098943664879953, | |
| "loss": 2.982, | |
| "step": 80600 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 0.00010074166453307344, | |
| "loss": 2.9822, | |
| "step": 80700 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 0.00010049389241734734, | |
| "loss": 2.9797, | |
| "step": 80800 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 0.00010024612030162125, | |
| "loss": 2.9913, | |
| "step": 80900 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.999834818589516e-05, | |
| "loss": 2.9879, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "eval_accuracy": 0.43145879200433396, | |
| "eval_loss": 3.0059893131256104, | |
| "eval_runtime": 37.641, | |
| "eval_samples_per_second": 301.958, | |
| "eval_steps_per_second": 2.524, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.975057607016905e-05, | |
| "loss": 2.9876, | |
| "step": 81100 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.950280395444296e-05, | |
| "loss": 2.9798, | |
| "step": 81200 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.925503183871687e-05, | |
| "loss": 2.9783, | |
| "step": 81300 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.900973744414804e-05, | |
| "loss": 2.9856, | |
| "step": 81400 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.876196532842193e-05, | |
| "loss": 2.99, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.851419321269584e-05, | |
| "loss": 2.9815, | |
| "step": 81600 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.826642109696975e-05, | |
| "loss": 2.9816, | |
| "step": 81700 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.801864898124363e-05, | |
| "loss": 2.9838, | |
| "step": 81800 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.777087686551754e-05, | |
| "loss": 2.9795, | |
| "step": 81900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.752310474979145e-05, | |
| "loss": 2.9764, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_accuracy": 0.4320621940557624, | |
| "eval_loss": 3.004152774810791, | |
| "eval_runtime": 37.3223, | |
| "eval_samples_per_second": 304.537, | |
| "eval_steps_per_second": 2.545, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.727533263406536e-05, | |
| "loss": 2.9796, | |
| "step": 82100 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.702756051833925e-05, | |
| "loss": 2.9796, | |
| "step": 82200 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.677978840261316e-05, | |
| "loss": 2.9842, | |
| "step": 82300 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.653201628688707e-05, | |
| "loss": 2.9867, | |
| "step": 82400 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.628424417116097e-05, | |
| "loss": 2.9787, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.603647205543488e-05, | |
| "loss": 2.9812, | |
| "step": 82600 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.578869993970878e-05, | |
| "loss": 2.9854, | |
| "step": 82700 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.554092782398269e-05, | |
| "loss": 2.9859, | |
| "step": 82800 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.529315570825657e-05, | |
| "loss": 2.9781, | |
| "step": 82900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.504538359253048e-05, | |
| "loss": 2.9827, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_accuracy": 0.43152391231757653, | |
| "eval_loss": 3.0029940605163574, | |
| "eval_runtime": 37.7162, | |
| "eval_samples_per_second": 301.356, | |
| "eval_steps_per_second": 2.519, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.479761147680439e-05, | |
| "loss": 2.9819, | |
| "step": 83100 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.454983936107829e-05, | |
| "loss": 2.9774, | |
| "step": 83200 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.430454496650946e-05, | |
| "loss": 2.9792, | |
| "step": 83300 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.405677285078337e-05, | |
| "loss": 2.9814, | |
| "step": 83400 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.380900073505726e-05, | |
| "loss": 2.9847, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.356370634048843e-05, | |
| "loss": 2.9777, | |
| "step": 83600 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.331593422476234e-05, | |
| "loss": 2.9804, | |
| "step": 83700 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.306816210903625e-05, | |
| "loss": 2.9823, | |
| "step": 83800 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.282038999331014e-05, | |
| "loss": 2.9715, | |
| "step": 83900 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.257261787758405e-05, | |
| "loss": 2.9769, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "eval_accuracy": 0.4324085787006696, | |
| "eval_loss": 3.0011510848999023, | |
| "eval_runtime": 36.6978, | |
| "eval_samples_per_second": 309.719, | |
| "eval_steps_per_second": 2.589, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.232484576185796e-05, | |
| "loss": 2.9795, | |
| "step": 84100 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.207707364613186e-05, | |
| "loss": 2.9795, | |
| "step": 84200 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.182930153040575e-05, | |
| "loss": 2.9859, | |
| "step": 84300 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.158152941467966e-05, | |
| "loss": 2.9744, | |
| "step": 84400 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.133375729895357e-05, | |
| "loss": 2.9785, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.108598518322746e-05, | |
| "loss": 2.9776, | |
| "step": 84600 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.083821306750137e-05, | |
| "loss": 2.9782, | |
| "step": 84700 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.059044095177528e-05, | |
| "loss": 2.9744, | |
| "step": 84800 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.034266883604918e-05, | |
| "loss": 2.9852, | |
| "step": 84900 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.009489672032309e-05, | |
| "loss": 2.9788, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "eval_accuracy": 0.43223330807034654, | |
| "eval_loss": 3.0002310276031494, | |
| "eval_runtime": 36.492, | |
| "eval_samples_per_second": 311.466, | |
| "eval_steps_per_second": 2.603, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.9847124604597e-05, | |
| "loss": 2.982, | |
| "step": 85100 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.95993524888709e-05, | |
| "loss": 2.974, | |
| "step": 85200 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.93515803731448e-05, | |
| "loss": 2.9794, | |
| "step": 85300 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.910380825741871e-05, | |
| "loss": 2.973, | |
| "step": 85400 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.885603614169262e-05, | |
| "loss": 2.9741, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.86082640259665e-05, | |
| "loss": 2.9796, | |
| "step": 85600 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.83604919102404e-05, | |
| "loss": 2.9687, | |
| "step": 85700 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.811271979451432e-05, | |
| "loss": 2.9755, | |
| "step": 85800 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.786494767878822e-05, | |
| "loss": 2.9721, | |
| "step": 85900 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.761717556306212e-05, | |
| "loss": 2.9734, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "eval_accuracy": 0.43254782532792235, | |
| "eval_loss": 2.998685121536255, | |
| "eval_runtime": 36.4931, | |
| "eval_samples_per_second": 311.456, | |
| "eval_steps_per_second": 2.603, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.736940344733603e-05, | |
| "loss": 2.9778, | |
| "step": 86100 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.712163133160994e-05, | |
| "loss": 2.9772, | |
| "step": 86200 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.687385921588383e-05, | |
| "loss": 2.9683, | |
| "step": 86300 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.662608710015774e-05, | |
| "loss": 2.9746, | |
| "step": 86400 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.637831498443165e-05, | |
| "loss": 2.979, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.613054286870556e-05, | |
| "loss": 2.9792, | |
| "step": 86600 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.588524847413671e-05, | |
| "loss": 2.967, | |
| "step": 86700 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.563747635841062e-05, | |
| "loss": 2.9794, | |
| "step": 86800 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.538970424268453e-05, | |
| "loss": 2.9738, | |
| "step": 86900 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.514193212695841e-05, | |
| "loss": 2.9769, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "eval_accuracy": 0.4327799030400102, | |
| "eval_loss": 2.997548818588257, | |
| "eval_runtime": 39.3891, | |
| "eval_samples_per_second": 288.557, | |
| "eval_steps_per_second": 2.412, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.489416001123232e-05, | |
| "loss": 2.9792, | |
| "step": 87100 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.464638789550623e-05, | |
| "loss": 2.9722, | |
| "step": 87200 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.439861577978014e-05, | |
| "loss": 2.9805, | |
| "step": 87300 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.415084366405404e-05, | |
| "loss": 2.9765, | |
| "step": 87400 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.390307154832794e-05, | |
| "loss": 2.9794, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.365529943260185e-05, | |
| "loss": 2.9697, | |
| "step": 87600 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.341000503803301e-05, | |
| "loss": 2.9733, | |
| "step": 87700 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.316223292230692e-05, | |
| "loss": 2.9746, | |
| "step": 87800 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.291446080658083e-05, | |
| "loss": 2.9655, | |
| "step": 87900 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.266668869085474e-05, | |
| "loss": 2.9676, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "eval_accuracy": 0.432616409487614, | |
| "eval_loss": 2.9959168434143066, | |
| "eval_runtime": 36.4885, | |
| "eval_samples_per_second": 311.495, | |
| "eval_steps_per_second": 2.604, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.241891657512863e-05, | |
| "loss": 2.9681, | |
| "step": 88100 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.21736221805598e-05, | |
| "loss": 2.9675, | |
| "step": 88200 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.192585006483371e-05, | |
| "loss": 2.9789, | |
| "step": 88300 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.167807794910759e-05, | |
| "loss": 2.9689, | |
| "step": 88400 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.14303058333815e-05, | |
| "loss": 2.9683, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.118253371765541e-05, | |
| "loss": 2.9729, | |
| "step": 88600 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.09347616019293e-05, | |
| "loss": 2.9676, | |
| "step": 88700 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.068698948620321e-05, | |
| "loss": 2.9708, | |
| "step": 88800 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.043921737047712e-05, | |
| "loss": 2.9669, | |
| "step": 88900 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.019144525475103e-05, | |
| "loss": 2.9677, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "eval_accuracy": 0.43304869752445824, | |
| "eval_loss": 2.9942848682403564, | |
| "eval_runtime": 36.9235, | |
| "eval_samples_per_second": 307.826, | |
| "eval_steps_per_second": 2.573, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 7.994367313902492e-05, | |
| "loss": 2.9678, | |
| "step": 89100 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.969590102329883e-05, | |
| "loss": 2.9667, | |
| "step": 89200 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.945060662873e-05, | |
| "loss": 2.9667, | |
| "step": 89300 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.92028345130039e-05, | |
| "loss": 2.9772, | |
| "step": 89400 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.895506239727781e-05, | |
| "loss": 2.9686, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.870729028155172e-05, | |
| "loss": 2.9651, | |
| "step": 89600 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.845951816582562e-05, | |
| "loss": 2.9738, | |
| "step": 89700 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.82117460500995e-05, | |
| "loss": 2.9674, | |
| "step": 89800 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.796397393437342e-05, | |
| "loss": 2.9731, | |
| "step": 89900 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.771620181864732e-05, | |
| "loss": 2.9739, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "eval_accuracy": 0.4330424626008499, | |
| "eval_loss": 2.993264675140381, | |
| "eval_runtime": 36.5912, | |
| "eval_samples_per_second": 310.621, | |
| "eval_steps_per_second": 2.596, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.746842970292122e-05, | |
| "loss": 2.966, | |
| "step": 90100 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.722065758719513e-05, | |
| "loss": 2.9695, | |
| "step": 90200 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.697288547146904e-05, | |
| "loss": 2.9719, | |
| "step": 90300 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.672511335574295e-05, | |
| "loss": 2.9691, | |
| "step": 90400 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.647734124001684e-05, | |
| "loss": 2.9707, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.622956912429075e-05, | |
| "loss": 2.9685, | |
| "step": 90600 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.598179700856466e-05, | |
| "loss": 2.9652, | |
| "step": 90700 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.573402489283854e-05, | |
| "loss": 2.9687, | |
| "step": 90800 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.548625277711245e-05, | |
| "loss": 2.9723, | |
| "step": 90900 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.523848066138636e-05, | |
| "loss": 2.9691, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "eval_accuracy": 0.4334020098622636, | |
| "eval_loss": 2.9914395809173584, | |
| "eval_runtime": 37.2256, | |
| "eval_samples_per_second": 305.327, | |
| "eval_steps_per_second": 2.552, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.499070854566027e-05, | |
| "loss": 2.9663, | |
| "step": 91100 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.474293642993418e-05, | |
| "loss": 2.9696, | |
| "step": 91200 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.449764203536533e-05, | |
| "loss": 2.969, | |
| "step": 91300 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.42523476407965e-05, | |
| "loss": 2.9687, | |
| "step": 91400 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 7.400705324622766e-05, | |
| "loss": 2.9709, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.375928113050156e-05, | |
| "loss": 2.9647, | |
| "step": 91600 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.351150901477547e-05, | |
| "loss": 2.9715, | |
| "step": 91700 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.326373689904938e-05, | |
| "loss": 2.9707, | |
| "step": 91800 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.301596478332328e-05, | |
| "loss": 2.9665, | |
| "step": 91900 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.276819266759719e-05, | |
| "loss": 2.969, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "eval_accuracy": 0.4335689672611089, | |
| "eval_loss": 2.990133285522461, | |
| "eval_runtime": 36.2216, | |
| "eval_samples_per_second": 313.79, | |
| "eval_steps_per_second": 2.623, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.252042055187108e-05, | |
| "loss": 2.9663, | |
| "step": 92100 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.227264843614499e-05, | |
| "loss": 2.9625, | |
| "step": 92200 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.20248763204189e-05, | |
| "loss": 2.9639, | |
| "step": 92300 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.17771042046928e-05, | |
| "loss": 2.9647, | |
| "step": 92400 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.15293320889667e-05, | |
| "loss": 2.9621, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.12815599732406e-05, | |
| "loss": 2.9648, | |
| "step": 92600 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 7.103378785751451e-05, | |
| "loss": 2.9671, | |
| "step": 92700 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 7.078601574178842e-05, | |
| "loss": 2.9601, | |
| "step": 92800 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 7.053824362606231e-05, | |
| "loss": 2.9639, | |
| "step": 92900 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 7.029047151033622e-05, | |
| "loss": 2.9602, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "eval_accuracy": 0.43370405727262273, | |
| "eval_loss": 2.9889016151428223, | |
| "eval_runtime": 37.7589, | |
| "eval_samples_per_second": 301.015, | |
| "eval_steps_per_second": 2.516, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 7.004269939461012e-05, | |
| "loss": 2.9696, | |
| "step": 93100 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.979492727888402e-05, | |
| "loss": 2.966, | |
| "step": 93200 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.954715516315793e-05, | |
| "loss": 2.9673, | |
| "step": 93300 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.929938304743184e-05, | |
| "loss": 2.9631, | |
| "step": 93400 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.905161093170574e-05, | |
| "loss": 2.9601, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.880383881597965e-05, | |
| "loss": 2.959, | |
| "step": 93600 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.855854442141082e-05, | |
| "loss": 2.9669, | |
| "step": 93700 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.831077230568471e-05, | |
| "loss": 2.9667, | |
| "step": 93800 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 6.806300018995862e-05, | |
| "loss": 2.9612, | |
| "step": 93900 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.781522807423251e-05, | |
| "loss": 2.965, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "eval_accuracy": 0.43392020129104486, | |
| "eval_loss": 2.987159490585327, | |
| "eval_runtime": 37.9274, | |
| "eval_samples_per_second": 299.678, | |
| "eval_steps_per_second": 2.505, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.756745595850642e-05, | |
| "loss": 2.9569, | |
| "step": 94100 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.731968384278033e-05, | |
| "loss": 2.9578, | |
| "step": 94200 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.707191172705423e-05, | |
| "loss": 2.9567, | |
| "step": 94300 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.682413961132814e-05, | |
| "loss": 2.9798, | |
| "step": 94400 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.657636749560203e-05, | |
| "loss": 2.9641, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.632859537987594e-05, | |
| "loss": 2.959, | |
| "step": 94600 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.608082326414985e-05, | |
| "loss": 2.9565, | |
| "step": 94700 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.583305114842376e-05, | |
| "loss": 2.9648, | |
| "step": 94800 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.558775675385491e-05, | |
| "loss": 2.9584, | |
| "step": 94900 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.533998463812882e-05, | |
| "loss": 2.9627, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "eval_accuracy": 0.4341079417685846, | |
| "eval_loss": 2.985278606414795, | |
| "eval_runtime": 36.3639, | |
| "eval_samples_per_second": 312.562, | |
| "eval_steps_per_second": 2.612, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.509221252240273e-05, | |
| "loss": 2.9647, | |
| "step": 95100 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 6.484444040667663e-05, | |
| "loss": 2.9584, | |
| "step": 95200 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.459666829095054e-05, | |
| "loss": 2.9659, | |
| "step": 95300 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.434889617522443e-05, | |
| "loss": 2.9635, | |
| "step": 95400 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.410112405949834e-05, | |
| "loss": 2.9615, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.385335194377225e-05, | |
| "loss": 2.9669, | |
| "step": 95600 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.360557982804614e-05, | |
| "loss": 2.9563, | |
| "step": 95700 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.335780771232005e-05, | |
| "loss": 2.9603, | |
| "step": 95800 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.311003559659395e-05, | |
| "loss": 2.9576, | |
| "step": 95900 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.286226348086786e-05, | |
| "loss": 2.9542, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "eval_accuracy": 0.4340469780710809, | |
| "eval_loss": 2.984398603439331, | |
| "eval_runtime": 36.0861, | |
| "eval_samples_per_second": 314.969, | |
| "eval_steps_per_second": 2.633, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.261449136514177e-05, | |
| "loss": 2.9567, | |
| "step": 96100 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.236671924941566e-05, | |
| "loss": 2.9652, | |
| "step": 96200 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.211894713368957e-05, | |
| "loss": 2.9593, | |
| "step": 96300 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 6.187117501796346e-05, | |
| "loss": 2.9532, | |
| "step": 96400 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.162340290223737e-05, | |
| "loss": 2.9562, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.137563078651128e-05, | |
| "loss": 2.9667, | |
| "step": 96600 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.112785867078519e-05, | |
| "loss": 2.9647, | |
| "step": 96700 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.0880086555059086e-05, | |
| "loss": 2.9587, | |
| "step": 96800 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.0632314439332995e-05, | |
| "loss": 2.961, | |
| "step": 96900 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.03845423236069e-05, | |
| "loss": 2.9552, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_accuracy": 0.43437950733019187, | |
| "eval_loss": 2.9822094440460205, | |
| "eval_runtime": 36.4432, | |
| "eval_samples_per_second": 311.883, | |
| "eval_steps_per_second": 2.607, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 6.013924792903806e-05, | |
| "loss": 2.9531, | |
| "step": 97100 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.989147581331197e-05, | |
| "loss": 2.9497, | |
| "step": 97200 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.964618141874313e-05, | |
| "loss": 2.9569, | |
| "step": 97300 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.939840930301703e-05, | |
| "loss": 2.9532, | |
| "step": 97400 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.915063718729094e-05, | |
| "loss": 2.9589, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.890286507156484e-05, | |
| "loss": 2.9626, | |
| "step": 97600 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.8655092955838746e-05, | |
| "loss": 2.9566, | |
| "step": 97700 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.840732084011265e-05, | |
| "loss": 2.9493, | |
| "step": 97800 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.8159548724386557e-05, | |
| "loss": 2.9595, | |
| "step": 97900 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.791177660866046e-05, | |
| "loss": 2.9576, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "eval_accuracy": 0.43472242812865003, | |
| "eval_loss": 2.98115611076355, | |
| "eval_runtime": 36.3414, | |
| "eval_samples_per_second": 312.757, | |
| "eval_steps_per_second": 2.614, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.7664004492934354e-05, | |
| "loss": 2.9558, | |
| "step": 98100 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.741623237720826e-05, | |
| "loss": 2.9584, | |
| "step": 98200 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.7168460261482165e-05, | |
| "loss": 2.9556, | |
| "step": 98300 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.6920688145756073e-05, | |
| "loss": 2.9576, | |
| "step": 98400 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.6672916030029976e-05, | |
| "loss": 2.9546, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.6425143914303884e-05, | |
| "loss": 2.9607, | |
| "step": 98600 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.6177371798577786e-05, | |
| "loss": 2.9464, | |
| "step": 98700 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.593207740400895e-05, | |
| "loss": 2.9614, | |
| "step": 98800 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.568430528828286e-05, | |
| "loss": 2.9592, | |
| "step": 98900 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.543653317255675e-05, | |
| "loss": 2.9579, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "eval_accuracy": 0.4348104098284565, | |
| "eval_loss": 2.9802134037017822, | |
| "eval_runtime": 37.635, | |
| "eval_samples_per_second": 302.006, | |
| "eval_steps_per_second": 2.524, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.518876105683066e-05, | |
| "loss": 2.9605, | |
| "step": 99100 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.4940988941104564e-05, | |
| "loss": 2.9602, | |
| "step": 99200 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.4693216825378466e-05, | |
| "loss": 2.9587, | |
| "step": 99300 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.4445444709652375e-05, | |
| "loss": 2.9472, | |
| "step": 99400 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.419767259392627e-05, | |
| "loss": 2.9589, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.394990047820018e-05, | |
| "loss": 2.9673, | |
| "step": 99600 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.370212836247408e-05, | |
| "loss": 2.9532, | |
| "step": 99700 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.345435624674799e-05, | |
| "loss": 2.9508, | |
| "step": 99800 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.320658413102189e-05, | |
| "loss": 2.9546, | |
| "step": 99900 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 5.2961289736453054e-05, | |
| "loss": 2.9508, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "eval_accuracy": 0.4348886927582055, | |
| "eval_loss": 2.9783637523651123, | |
| "eval_runtime": 36.4236, | |
| "eval_samples_per_second": 312.05, | |
| "eval_steps_per_second": 2.608, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.271599534188422e-05, | |
| "loss": 2.9495, | |
| "step": 100100 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.2468223226158125e-05, | |
| "loss": 2.9514, | |
| "step": 100200 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.222045111043202e-05, | |
| "loss": 2.9546, | |
| "step": 100300 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.197267899470593e-05, | |
| "loss": 2.9544, | |
| "step": 100400 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.172490687897983e-05, | |
| "loss": 2.9568, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.147713476325374e-05, | |
| "loss": 2.9591, | |
| "step": 100600 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.122936264752764e-05, | |
| "loss": 2.9574, | |
| "step": 100700 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.098159053180155e-05, | |
| "loss": 2.9591, | |
| "step": 100800 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.073381841607545e-05, | |
| "loss": 2.9438, | |
| "step": 100900 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.048604630034935e-05, | |
| "loss": 2.9551, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "eval_accuracy": 0.43532513741078865, | |
| "eval_loss": 2.977102041244507, | |
| "eval_runtime": 36.649, | |
| "eval_samples_per_second": 310.131, | |
| "eval_steps_per_second": 2.592, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.023827418462326e-05, | |
| "loss": 2.9481, | |
| "step": 101100 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.999050206889716e-05, | |
| "loss": 2.9504, | |
| "step": 101200 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.974272995317107e-05, | |
| "loss": 2.9566, | |
| "step": 101300 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.949495783744497e-05, | |
| "loss": 2.9571, | |
| "step": 101400 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.924718572171888e-05, | |
| "loss": 2.9469, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.8999413605992774e-05, | |
| "loss": 2.9507, | |
| "step": 101600 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.875164149026668e-05, | |
| "loss": 2.9506, | |
| "step": 101700 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.8503869374540585e-05, | |
| "loss": 2.944, | |
| "step": 101800 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.825609725881449e-05, | |
| "loss": 2.9513, | |
| "step": 101900 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.8008325143088396e-05, | |
| "loss": 2.9535, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "eval_accuracy": 0.4356812208257533, | |
| "eval_loss": 2.975886106491089, | |
| "eval_runtime": 37.6554, | |
| "eval_samples_per_second": 301.843, | |
| "eval_steps_per_second": 2.523, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.77605530273623e-05, | |
| "loss": 2.9564, | |
| "step": 102100 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.751525863279347e-05, | |
| "loss": 2.9465, | |
| "step": 102200 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.726748651706736e-05, | |
| "loss": 2.9458, | |
| "step": 102300 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.7019714401341264e-05, | |
| "loss": 2.9552, | |
| "step": 102400 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.677194228561517e-05, | |
| "loss": 2.9505, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.6524170169889075e-05, | |
| "loss": 2.9547, | |
| "step": 102600 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.6276398054162984e-05, | |
| "loss": 2.9493, | |
| "step": 102700 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.6028625938436886e-05, | |
| "loss": 2.9543, | |
| "step": 102800 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.5780853822710795e-05, | |
| "loss": 2.9447, | |
| "step": 102900 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.553308170698469e-05, | |
| "loss": 2.9479, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "eval_accuracy": 0.43567152205569587, | |
| "eval_loss": 2.9743244647979736, | |
| "eval_runtime": 36.3875, | |
| "eval_samples_per_second": 312.36, | |
| "eval_steps_per_second": 2.611, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.528530959125859e-05, | |
| "loss": 2.9493, | |
| "step": 103100 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.50375374755325e-05, | |
| "loss": 2.952, | |
| "step": 103200 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.47897653598064e-05, | |
| "loss": 2.954, | |
| "step": 103300 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.454199324408031e-05, | |
| "loss": 2.9462, | |
| "step": 103400 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.429422112835421e-05, | |
| "loss": 2.9468, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.4046449012628116e-05, | |
| "loss": 2.9514, | |
| "step": 103600 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.379867689690202e-05, | |
| "loss": 2.9504, | |
| "step": 103700 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.355090478117592e-05, | |
| "loss": 2.9568, | |
| "step": 103800 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.330313266544983e-05, | |
| "loss": 2.9525, | |
| "step": 103900 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.305536054972373e-05, | |
| "loss": 2.9542, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "eval_accuracy": 0.43588905161269764, | |
| "eval_loss": 2.973242998123169, | |
| "eval_runtime": 36.37, | |
| "eval_samples_per_second": 312.51, | |
| "eval_steps_per_second": 2.612, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.280758843399764e-05, | |
| "loss": 2.9467, | |
| "step": 104100 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.25622940394288e-05, | |
| "loss": 2.9516, | |
| "step": 104200 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.23145219237027e-05, | |
| "loss": 2.9421, | |
| "step": 104300 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.2066749807976606e-05, | |
| "loss": 2.9451, | |
| "step": 104400 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.182145541340777e-05, | |
| "loss": 2.9481, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.157368329768168e-05, | |
| "loss": 2.9514, | |
| "step": 104600 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.132591118195558e-05, | |
| "loss": 2.9486, | |
| "step": 104700 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.107813906622948e-05, | |
| "loss": 2.9429, | |
| "step": 104800 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.083036695050339e-05, | |
| "loss": 2.948, | |
| "step": 104900 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.0582594834777285e-05, | |
| "loss": 2.9481, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "eval_accuracy": 0.4360442319336161, | |
| "eval_loss": 2.971482992172241, | |
| "eval_runtime": 36.0911, | |
| "eval_samples_per_second": 314.925, | |
| "eval_steps_per_second": 2.632, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.0334822719051194e-05, | |
| "loss": 2.9432, | |
| "step": 105100 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.0087050603325096e-05, | |
| "loss": 2.9543, | |
| "step": 105200 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.9839278487599005e-05, | |
| "loss": 2.9504, | |
| "step": 105300 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.959150637187291e-05, | |
| "loss": 2.9536, | |
| "step": 105400 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.9343734256146816e-05, | |
| "loss": 2.945, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.909596214042071e-05, | |
| "loss": 2.9532, | |
| "step": 105600 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.884819002469461e-05, | |
| "loss": 2.9493, | |
| "step": 105700 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.860041790896852e-05, | |
| "loss": 2.948, | |
| "step": 105800 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.8352645793242424e-05, | |
| "loss": 2.9508, | |
| "step": 105900 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.810735139867359e-05, | |
| "loss": 2.941, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "eval_accuracy": 0.4362319724111558, | |
| "eval_loss": 2.969744920730591, | |
| "eval_runtime": 36.2819, | |
| "eval_samples_per_second": 313.269, | |
| "eval_steps_per_second": 2.618, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.7859579282947495e-05, | |
| "loss": 2.944, | |
| "step": 106100 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.76118071672214e-05, | |
| "loss": 2.9433, | |
| "step": 106200 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.7364035051495306e-05, | |
| "loss": 2.9442, | |
| "step": 106300 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.711626293576921e-05, | |
| "loss": 2.9525, | |
| "step": 106400 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.686849082004311e-05, | |
| "loss": 2.9423, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.662071870431701e-05, | |
| "loss": 2.9505, | |
| "step": 106600 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.6372946588590914e-05, | |
| "loss": 2.948, | |
| "step": 106700 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.612517447286482e-05, | |
| "loss": 2.9446, | |
| "step": 106800 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.5877402357138725e-05, | |
| "loss": 2.9429, | |
| "step": 106900 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.562963024141263e-05, | |
| "loss": 2.9435, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "eval_accuracy": 0.4365194716664288, | |
| "eval_loss": 2.9684131145477295, | |
| "eval_runtime": 36.3894, | |
| "eval_samples_per_second": 312.343, | |
| "eval_steps_per_second": 2.611, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.5381858125686536e-05, | |
| "loss": 2.9433, | |
| "step": 107100 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.513408600996043e-05, | |
| "loss": 2.9447, | |
| "step": 107200 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.488631389423434e-05, | |
| "loss": 2.9476, | |
| "step": 107300 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.463854177850824e-05, | |
| "loss": 2.9508, | |
| "step": 107400 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.4390769662782144e-05, | |
| "loss": 2.9423, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.414299754705605e-05, | |
| "loss": 2.9491, | |
| "step": 107600 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.3897703152487215e-05, | |
| "loss": 2.943, | |
| "step": 107700 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.364993103676112e-05, | |
| "loss": 2.9444, | |
| "step": 107800 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.3402158921035026e-05, | |
| "loss": 2.9475, | |
| "step": 107900 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.315438680530893e-05, | |
| "loss": 2.9403, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "eval_accuracy": 0.4367591698407046, | |
| "eval_loss": 2.9674224853515625, | |
| "eval_runtime": 36.6777, | |
| "eval_samples_per_second": 309.888, | |
| "eval_steps_per_second": 2.59, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.290661468958283e-05, | |
| "loss": 2.9395, | |
| "step": 108100 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.265884257385674e-05, | |
| "loss": 2.9413, | |
| "step": 108200 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.241107045813064e-05, | |
| "loss": 2.9495, | |
| "step": 108300 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.216329834240454e-05, | |
| "loss": 2.9461, | |
| "step": 108400 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.1915526226678445e-05, | |
| "loss": 2.939, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.166775411095235e-05, | |
| "loss": 2.9436, | |
| "step": 108600 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.1419981995226256e-05, | |
| "loss": 2.9414, | |
| "step": 108700 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.117220987950016e-05, | |
| "loss": 2.9388, | |
| "step": 108800 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.092691548493133e-05, | |
| "loss": 2.9425, | |
| "step": 108900 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.067914336920522e-05, | |
| "loss": 2.9453, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "eval_accuracy": 0.43670790491325834, | |
| "eval_loss": 2.9661126136779785, | |
| "eval_runtime": 37.1173, | |
| "eval_samples_per_second": 306.218, | |
| "eval_steps_per_second": 2.559, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.0433848974636392e-05, | |
| "loss": 2.9453, | |
| "step": 109100 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.0186076858910297e-05, | |
| "loss": 2.9398, | |
| "step": 109200 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 2.99383047431842e-05, | |
| "loss": 2.9452, | |
| "step": 109300 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 2.9690532627458105e-05, | |
| "loss": 2.9423, | |
| "step": 109400 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 2.9442760511732007e-05, | |
| "loss": 2.945, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 2.919498839600591e-05, | |
| "loss": 2.941, | |
| "step": 109600 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 2.8947216280279814e-05, | |
| "loss": 2.9404, | |
| "step": 109700 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.869944416455372e-05, | |
| "loss": 2.9441, | |
| "step": 109800 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.845167204882762e-05, | |
| "loss": 2.9427, | |
| "step": 109900 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.8203899933101527e-05, | |
| "loss": 2.9396, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "eval_accuracy": 0.43717206033743405, | |
| "eval_loss": 2.964357852935791, | |
| "eval_runtime": 36.5503, | |
| "eval_samples_per_second": 310.969, | |
| "eval_steps_per_second": 2.599, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.7956127817375432e-05, | |
| "loss": 2.9405, | |
| "step": 110100 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.7708355701649334e-05, | |
| "loss": 2.9487, | |
| "step": 110200 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.7460583585923236e-05, | |
| "loss": 2.9415, | |
| "step": 110300 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.7212811470197142e-05, | |
| "loss": 2.9365, | |
| "step": 110400 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.6965039354471044e-05, | |
| "loss": 2.9468, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.671726723874495e-05, | |
| "loss": 2.9311, | |
| "step": 110600 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.6469495123018855e-05, | |
| "loss": 2.9402, | |
| "step": 110700 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.6221723007292757e-05, | |
| "loss": 2.9361, | |
| "step": 110800 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 2.5973950891566662e-05, | |
| "loss": 2.9328, | |
| "step": 110900 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.5726178775840564e-05, | |
| "loss": 2.9375, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "eval_accuracy": 0.43724757219002386, | |
| "eval_loss": 2.9633212089538574, | |
| "eval_runtime": 36.4613, | |
| "eval_samples_per_second": 311.728, | |
| "eval_steps_per_second": 2.606, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.5478406660114466e-05, | |
| "loss": 2.9379, | |
| "step": 111100 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.523063454438837e-05, | |
| "loss": 2.933, | |
| "step": 111200 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.4982862428662277e-05, | |
| "loss": 2.9385, | |
| "step": 111300 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.473509031293618e-05, | |
| "loss": 2.9362, | |
| "step": 111400 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.4487318197210084e-05, | |
| "loss": 2.9342, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.423954608148399e-05, | |
| "loss": 2.9392, | |
| "step": 111600 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.3991773965757892e-05, | |
| "loss": 2.9382, | |
| "step": 111700 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.3746479571189054e-05, | |
| "loss": 2.9376, | |
| "step": 111800 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.349870745546296e-05, | |
| "loss": 2.9352, | |
| "step": 111900 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.3250935339736865e-05, | |
| "loss": 2.9284, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "eval_accuracy": 0.43742907774395523, | |
| "eval_loss": 2.96207857131958, | |
| "eval_runtime": 36.2575, | |
| "eval_samples_per_second": 313.48, | |
| "eval_steps_per_second": 2.62, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.300316322401077e-05, | |
| "loss": 2.9384, | |
| "step": 112100 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.2755391108284673e-05, | |
| "loss": 2.933, | |
| "step": 112200 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.2507618992558575e-05, | |
| "loss": 2.935, | |
| "step": 112300 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.2259846876832477e-05, | |
| "loss": 2.938, | |
| "step": 112400 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.2012074761106382e-05, | |
| "loss": 2.9392, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.1764302645380288e-05, | |
| "loss": 2.9409, | |
| "step": 112600 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.151653052965419e-05, | |
| "loss": 2.9445, | |
| "step": 112700 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.1268758413928095e-05, | |
| "loss": 2.9378, | |
| "step": 112800 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.1020986298202e-05, | |
| "loss": 2.9357, | |
| "step": 112900 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.0773214182475906e-05, | |
| "loss": 2.9418, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "eval_accuracy": 0.4375648605247589, | |
| "eval_loss": 2.9606027603149414, | |
| "eval_runtime": 38.0645, | |
| "eval_samples_per_second": 298.598, | |
| "eval_steps_per_second": 2.496, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.0525442066749805e-05, | |
| "loss": 2.9426, | |
| "step": 113100 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.028014767218097e-05, | |
| "loss": 2.9436, | |
| "step": 113200 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.0032375556454876e-05, | |
| "loss": 2.9355, | |
| "step": 113300 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.978460344072878e-05, | |
| "loss": 2.9296, | |
| "step": 113400 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.9536831325002683e-05, | |
| "loss": 2.9288, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.9289059209276585e-05, | |
| "loss": 2.9385, | |
| "step": 113600 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.904128709355049e-05, | |
| "loss": 2.9348, | |
| "step": 113700 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.8793514977824393e-05, | |
| "loss": 2.9296, | |
| "step": 113800 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.8545742862098298e-05, | |
| "loss": 2.9354, | |
| "step": 113900 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.8297970746372204e-05, | |
| "loss": 2.934, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "eval_accuracy": 0.4376666976103616, | |
| "eval_loss": 2.9594342708587646, | |
| "eval_runtime": 36.6834, | |
| "eval_samples_per_second": 309.841, | |
| "eval_steps_per_second": 2.59, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.8050198630646106e-05, | |
| "loss": 2.9292, | |
| "step": 114100 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.7802426514920008e-05, | |
| "loss": 2.9371, | |
| "step": 114200 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.7557132120351174e-05, | |
| "loss": 2.9319, | |
| "step": 114300 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.730936000462508e-05, | |
| "loss": 2.9402, | |
| "step": 114400 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.706158788889898e-05, | |
| "loss": 2.9303, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.6813815773172883e-05, | |
| "loss": 2.9395, | |
| "step": 114600 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.656604365744679e-05, | |
| "loss": 2.9359, | |
| "step": 114700 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.6318271541720694e-05, | |
| "loss": 2.9392, | |
| "step": 114800 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.60704994259946e-05, | |
| "loss": 2.9353, | |
| "step": 114900 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.58227273102685e-05, | |
| "loss": 2.9374, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "eval_accuracy": 0.4379791365600679, | |
| "eval_loss": 2.958286762237549, | |
| "eval_runtime": 36.2601, | |
| "eval_samples_per_second": 313.458, | |
| "eval_steps_per_second": 2.62, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5574955194542403e-05, | |
| "loss": 2.9304, | |
| "step": 115100 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.532718307881631e-05, | |
| "loss": 2.9345, | |
| "step": 115200 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5079410963090212e-05, | |
| "loss": 2.9344, | |
| "step": 115300 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4834116568521378e-05, | |
| "loss": 2.9346, | |
| "step": 115400 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.458634445279528e-05, | |
| "loss": 2.936, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4338572337069184e-05, | |
| "loss": 2.9414, | |
| "step": 115600 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.409080022134309e-05, | |
| "loss": 2.938, | |
| "step": 115700 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.3843028105616993e-05, | |
| "loss": 2.9345, | |
| "step": 115800 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.3595255989890895e-05, | |
| "loss": 2.9356, | |
| "step": 115900 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.33474838741648e-05, | |
| "loss": 2.9302, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_accuracy": 0.43818142519269376, | |
| "eval_loss": 2.9568593502044678, | |
| "eval_runtime": 36.5356, | |
| "eval_samples_per_second": 311.094, | |
| "eval_steps_per_second": 2.6, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.3102189479595965e-05, | |
| "loss": 2.9352, | |
| "step": 116100 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.2854417363869869e-05, | |
| "loss": 2.9384, | |
| "step": 116200 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.2606645248143774e-05, | |
| "loss": 2.9256, | |
| "step": 116300 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.2358873132417676e-05, | |
| "loss": 2.9308, | |
| "step": 116400 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.211110101669158e-05, | |
| "loss": 2.9365, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1863328900965485e-05, | |
| "loss": 2.9291, | |
| "step": 116600 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1615556785239389e-05, | |
| "loss": 2.9293, | |
| "step": 116700 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1367784669513291e-05, | |
| "loss": 2.9411, | |
| "step": 116800 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1120012553787196e-05, | |
| "loss": 2.9315, | |
| "step": 116900 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.08722404380611e-05, | |
| "loss": 2.9273, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_accuracy": 0.4382222985807928, | |
| "eval_loss": 2.9559996128082275, | |
| "eval_runtime": 36.528, | |
| "eval_samples_per_second": 311.158, | |
| "eval_steps_per_second": 2.601, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.0624468322335004e-05, | |
| "loss": 2.9355, | |
| "step": 117100 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.0376696206608908e-05, | |
| "loss": 2.9227, | |
| "step": 117200 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.0128924090882811e-05, | |
| "loss": 2.9323, | |
| "step": 117300 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.881151975156715e-06, | |
| "loss": 2.9335, | |
| "step": 117400 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.63337985943062e-06, | |
| "loss": 2.9313, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.385607743704522e-06, | |
| "loss": 2.932, | |
| "step": 117600 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.137835627978426e-06, | |
| "loss": 2.934, | |
| "step": 117700 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.89006351225233e-06, | |
| "loss": 2.9361, | |
| "step": 117800 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.642291396526234e-06, | |
| "loss": 2.9318, | |
| "step": 117900 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.394519280800137e-06, | |
| "loss": 2.9338, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "eval_accuracy": 0.4383601596694659, | |
| "eval_loss": 2.9548416137695312, | |
| "eval_runtime": 36.5258, | |
| "eval_samples_per_second": 311.177, | |
| "eval_steps_per_second": 2.601, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.146747165074043e-06, | |
| "loss": 2.9222, | |
| "step": 118100 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 7.898975049347945e-06, | |
| "loss": 2.9346, | |
| "step": 118200 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 7.65120293362185e-06, | |
| "loss": 2.9239, | |
| "step": 118300 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 7.403430817895753e-06, | |
| "loss": 2.9302, | |
| "step": 118400 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 7.155658702169658e-06, | |
| "loss": 2.9323, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 6.9078865864435605e-06, | |
| "loss": 2.9315, | |
| "step": 118600 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 6.660114470717465e-06, | |
| "loss": 2.9308, | |
| "step": 118700 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 6.412342354991369e-06, | |
| "loss": 2.9289, | |
| "step": 118800 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 6.1645702392652725e-06, | |
| "loss": 2.9301, | |
| "step": 118900 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 5.916798123539176e-06, | |
| "loss": 2.9304, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "eval_accuracy": 0.43854651460842603, | |
| "eval_loss": 2.953854560852051, | |
| "eval_runtime": 37.1399, | |
| "eval_samples_per_second": 306.032, | |
| "eval_steps_per_second": 2.558, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 5.669026007813081e-06, | |
| "loss": 2.9354, | |
| "step": 119100 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 5.423731613244245e-06, | |
| "loss": 2.9288, | |
| "step": 119200 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 5.175959497518149e-06, | |
| "loss": 2.9323, | |
| "step": 119300 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.928187381792053e-06, | |
| "loss": 2.9298, | |
| "step": 119400 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.680415266065956e-06, | |
| "loss": 2.9226, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.435120871497121e-06, | |
| "loss": 2.9276, | |
| "step": 119600 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.187348755771026e-06, | |
| "loss": 2.926, | |
| "step": 119700 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.9395766400449295e-06, | |
| "loss": 2.9256, | |
| "step": 119800 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.6918045243188332e-06, | |
| "loss": 2.9231, | |
| "step": 119900 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.4440324085927365e-06, | |
| "loss": 2.9361, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "eval_accuracy": 0.43850217737387787, | |
| "eval_loss": 2.9531476497650146, | |
| "eval_runtime": 36.7198, | |
| "eval_samples_per_second": 309.534, | |
| "eval_steps_per_second": 2.587, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.1962602928666402e-06, | |
| "loss": 2.9236, | |
| "step": 120100 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.948488177140544e-06, | |
| "loss": 2.9354, | |
| "step": 120200 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.700716061414448e-06, | |
| "loss": 2.9264, | |
| "step": 120300 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.452943945688352e-06, | |
| "loss": 2.925, | |
| "step": 120400 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.205171829962256e-06, | |
| "loss": 2.932, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.9573997142361597e-06, | |
| "loss": 2.9291, | |
| "step": 120600 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.7096275985100634e-06, | |
| "loss": 2.9221, | |
| "step": 120700 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.4643332039412285e-06, | |
| "loss": 2.9315, | |
| "step": 120800 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.2165610882151322e-06, | |
| "loss": 2.9287, | |
| "step": 120900 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.68788972489036e-07, | |
| "loss": 2.927, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.43872455631590834, | |
| "eval_loss": 2.9526402950286865, | |
| "eval_runtime": 36.2748, | |
| "eval_samples_per_second": 313.331, | |
| "eval_steps_per_second": 2.619, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 7.210168567629398e-07, | |
| "loss": 2.9372, | |
| "step": 121100 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.732447410368437e-07, | |
| "loss": 2.9285, | |
| "step": 121200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 121279, | |
| "total_flos": 1.477015264316925e+20, | |
| "train_loss": 3.0377933233989878, | |
| "train_runtime": 170019.0203, | |
| "train_samples_per_second": 171.199, | |
| "train_steps_per_second": 0.713 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 121279, | |
| "num_train_epochs": 1, | |
| "save_steps": 12128, | |
| "total_flos": 1.477015264316925e+20, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |