| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.9921259842519685, |
| "eval_steps": 500, |
| "global_step": 570, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05249343832020997, |
| "grad_norm": 5.856240742467741, |
| "learning_rate": 5e-06, |
| "loss": 0.9788, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.10498687664041995, |
| "grad_norm": 1.504758475031017, |
| "learning_rate": 5e-06, |
| "loss": 0.8672, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.15748031496062992, |
| "grad_norm": 0.9699579752356402, |
| "learning_rate": 5e-06, |
| "loss": 0.8312, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.2099737532808399, |
| "grad_norm": 2.3401202479822167, |
| "learning_rate": 5e-06, |
| "loss": 0.8176, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.26246719160104987, |
| "grad_norm": 1.4328785905158876, |
| "learning_rate": 5e-06, |
| "loss": 0.8021, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.31496062992125984, |
| "grad_norm": 1.776027549071983, |
| "learning_rate": 5e-06, |
| "loss": 0.7924, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.3674540682414698, |
| "grad_norm": 1.6731596205975654, |
| "learning_rate": 5e-06, |
| "loss": 0.7862, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.4199475065616798, |
| "grad_norm": 0.7241122381033481, |
| "learning_rate": 5e-06, |
| "loss": 0.7745, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.47244094488188976, |
| "grad_norm": 0.667641025148353, |
| "learning_rate": 5e-06, |
| "loss": 0.7684, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.5249343832020997, |
| "grad_norm": 0.5504196367515327, |
| "learning_rate": 5e-06, |
| "loss": 0.7659, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.5774278215223098, |
| "grad_norm": 0.5097698406084038, |
| "learning_rate": 5e-06, |
| "loss": 0.7644, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.6299212598425197, |
| "grad_norm": 0.7982727448245113, |
| "learning_rate": 5e-06, |
| "loss": 0.7621, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.6824146981627297, |
| "grad_norm": 0.7024212096277765, |
| "learning_rate": 5e-06, |
| "loss": 0.7587, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.7349081364829396, |
| "grad_norm": 0.6368075359040738, |
| "learning_rate": 5e-06, |
| "loss": 0.7528, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.7874015748031497, |
| "grad_norm": 0.5954968543585505, |
| "learning_rate": 5e-06, |
| "loss": 0.7523, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.8398950131233596, |
| "grad_norm": 0.670163545471037, |
| "learning_rate": 5e-06, |
| "loss": 0.7534, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.8923884514435696, |
| "grad_norm": 0.5996022754065202, |
| "learning_rate": 5e-06, |
| "loss": 0.7517, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.9448818897637795, |
| "grad_norm": 0.8415836628541776, |
| "learning_rate": 5e-06, |
| "loss": 0.7459, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.9973753280839895, |
| "grad_norm": 0.690137189983478, |
| "learning_rate": 5e-06, |
| "loss": 0.7488, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.0498687664041995, |
| "grad_norm": 0.5870306499065036, |
| "learning_rate": 5e-06, |
| "loss": 0.7462, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.1023622047244095, |
| "grad_norm": 0.6711935804641513, |
| "learning_rate": 5e-06, |
| "loss": 0.7045, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.1548556430446195, |
| "grad_norm": 0.6793309667692513, |
| "learning_rate": 5e-06, |
| "loss": 0.7054, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.2073490813648293, |
| "grad_norm": 0.4945921653765713, |
| "learning_rate": 5e-06, |
| "loss": 0.7007, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.2598425196850394, |
| "grad_norm": 0.7994708362053626, |
| "learning_rate": 5e-06, |
| "loss": 0.7035, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.3123359580052494, |
| "grad_norm": 0.6645996514564108, |
| "learning_rate": 5e-06, |
| "loss": 0.7029, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.3648293963254594, |
| "grad_norm": 0.6524179166081678, |
| "learning_rate": 5e-06, |
| "loss": 0.7069, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.4173228346456692, |
| "grad_norm": 0.8192112613994083, |
| "learning_rate": 5e-06, |
| "loss": 0.7058, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.4698162729658792, |
| "grad_norm": 0.7359384256295017, |
| "learning_rate": 5e-06, |
| "loss": 0.7096, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.5223097112860893, |
| "grad_norm": 0.8479895809019217, |
| "learning_rate": 5e-06, |
| "loss": 0.7014, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.574803149606299, |
| "grad_norm": 0.5643654766970888, |
| "learning_rate": 5e-06, |
| "loss": 0.7027, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.627296587926509, |
| "grad_norm": 0.5431958308145564, |
| "learning_rate": 5e-06, |
| "loss": 0.7014, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.6797900262467191, |
| "grad_norm": 0.520988761182218, |
| "learning_rate": 5e-06, |
| "loss": 0.7004, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.7322834645669292, |
| "grad_norm": 0.5293919501601649, |
| "learning_rate": 5e-06, |
| "loss": 0.7014, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.7847769028871392, |
| "grad_norm": 0.5558746861224813, |
| "learning_rate": 5e-06, |
| "loss": 0.7051, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.8372703412073492, |
| "grad_norm": 0.5864088226331533, |
| "learning_rate": 5e-06, |
| "loss": 0.7028, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.889763779527559, |
| "grad_norm": 0.5203657185554663, |
| "learning_rate": 5e-06, |
| "loss": 0.7029, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.942257217847769, |
| "grad_norm": 0.5249283309342107, |
| "learning_rate": 5e-06, |
| "loss": 0.6985, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.9947506561679789, |
| "grad_norm": 0.5014731402840785, |
| "learning_rate": 5e-06, |
| "loss": 0.698, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.047244094488189, |
| "grad_norm": 0.6606881961316096, |
| "learning_rate": 5e-06, |
| "loss": 0.701, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.099737532808399, |
| "grad_norm": 0.7956310597917616, |
| "learning_rate": 5e-06, |
| "loss": 0.6552, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.152230971128609, |
| "grad_norm": 0.7409823582769325, |
| "learning_rate": 5e-06, |
| "loss": 0.6529, |
| "step": 410 |
| }, |
| { |
| "epoch": 2.204724409448819, |
| "grad_norm": 0.5301522848133626, |
| "learning_rate": 5e-06, |
| "loss": 0.6583, |
| "step": 420 |
| }, |
| { |
| "epoch": 2.257217847769029, |
| "grad_norm": 0.5707394351699061, |
| "learning_rate": 5e-06, |
| "loss": 0.6567, |
| "step": 430 |
| }, |
| { |
| "epoch": 2.309711286089239, |
| "grad_norm": 0.8305197146184463, |
| "learning_rate": 5e-06, |
| "loss": 0.6591, |
| "step": 440 |
| }, |
| { |
| "epoch": 2.362204724409449, |
| "grad_norm": 0.6345171704808862, |
| "learning_rate": 5e-06, |
| "loss": 0.6586, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.4146981627296586, |
| "grad_norm": 0.540143075416848, |
| "learning_rate": 5e-06, |
| "loss": 0.6577, |
| "step": 460 |
| }, |
| { |
| "epoch": 2.4671916010498687, |
| "grad_norm": 0.6047688082843984, |
| "learning_rate": 5e-06, |
| "loss": 0.6583, |
| "step": 470 |
| }, |
| { |
| "epoch": 2.5196850393700787, |
| "grad_norm": 0.5089956964635003, |
| "learning_rate": 5e-06, |
| "loss": 0.6597, |
| "step": 480 |
| }, |
| { |
| "epoch": 2.5721784776902887, |
| "grad_norm": 0.6189665519218218, |
| "learning_rate": 5e-06, |
| "loss": 0.6593, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.6246719160104988, |
| "grad_norm": 0.6080333484551841, |
| "learning_rate": 5e-06, |
| "loss": 0.6586, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.677165354330709, |
| "grad_norm": 0.6796282219494648, |
| "learning_rate": 5e-06, |
| "loss": 0.6613, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.729658792650919, |
| "grad_norm": 0.6734905288037869, |
| "learning_rate": 5e-06, |
| "loss": 0.6593, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.7821522309711284, |
| "grad_norm": 0.7251898875558835, |
| "learning_rate": 5e-06, |
| "loss": 0.6579, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.8346456692913384, |
| "grad_norm": 0.5903651504718952, |
| "learning_rate": 5e-06, |
| "loss": 0.6613, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.8871391076115485, |
| "grad_norm": 0.6998542288142676, |
| "learning_rate": 5e-06, |
| "loss": 0.6589, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.9396325459317585, |
| "grad_norm": 0.6240795370124482, |
| "learning_rate": 5e-06, |
| "loss": 0.6594, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.9921259842519685, |
| "grad_norm": 0.5469561259952066, |
| "learning_rate": 5e-06, |
| "loss": 0.6579, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.9921259842519685, |
| "step": 570, |
| "total_flos": 954352470589440.0, |
| "train_loss": 0.718004734474316, |
| "train_runtime": 8541.7768, |
| "train_samples_per_second": 34.252, |
| "train_steps_per_second": 0.067 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 570, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 954352470589440.0, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|