| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "global_step": 20373, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.07, |
| "learning_rate": 4.882079900826349e-05, |
| "loss": 1.5901, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 4.539443723136697e-05, |
| "loss": 1.4477, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 4.004414420573958e-05, |
| "loss": 1.3924, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 3.3274645598733375e-05, |
| "loss": 1.3816, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 2.5724549368023616e-05, |
| "loss": 1.3555, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.8106101990647016e-05, |
| "loss": 1.3471, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.1137997922835163e-05, |
| "loss": 1.3297, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 5.477580782771013e-06, |
| "loss": 1.3251, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.6588321308710408e-06, |
| "loss": 1.3444, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 4.1997742936272785e-08, |
| "loss": 1.3264, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 7.796034360648269e-07, |
| "loss": 1.3304, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 3.802066381068883e-06, |
| "loss": 1.3264, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 8.8242592737696e-06, |
| "loss": 1.3417, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5372408126985777e-05, |
| "loss": 1.3275, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 2.282878625097739e-05, |
| "loss": 1.2826, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 3.0489988167458534e-05, |
| "loss": 1.2881, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 3.76332861246046e-05, |
| "loss": 1.2777, |
| "step": 8500 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 4.3584809399588655e-05, |
| "loss": 1.278, |
| "step": 9000 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 4.778311462055443e-05, |
| "loss": 1.2832, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 4.983215013308785e-05, |
| "loss": 1.257, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 4.953861796051032e-05, |
| "loss": 1.2585, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 4.6930208777142614e-05, |
| "loss": 1.29, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 4.225298967865532e-05, |
| "loss": 1.2467, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 3.59481911770088e-05, |
| "loss": 1.2725, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 2.8610583243870337e-05, |
| "loss": 1.2567, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 2.09323670433783e-05, |
| "loss": 1.2283, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 1.363787538820378e-05, |
| "loss": 1.246, |
| "step": 13500 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 7.4152420218663585e-06, |
| "loss": 1.1957, |
| "step": 14000 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 2.8514857793098454e-06, |
| "loss": 1.1595, |
| "step": 14500 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.771335315155861e-07, |
| "loss": 1.2108, |
| "step": 15000 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.2560596844390935e-07, |
| "loss": 1.1781, |
| "step": 15500 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.4111976063067728e-06, |
| "loss": 1.2147, |
| "step": 16000 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 6.727728298952271e-06, |
| "loss": 1.189, |
| "step": 16500 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.2767993464490152e-05, |
| "loss": 1.2059, |
| "step": 17000 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.9962178169036043e-05, |
| "loss": 1.1812, |
| "step": 17500 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 2.7631611233523096e-05, |
| "loss": 1.1945, |
| "step": 18000 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 3.505278841189536e-05, |
| "loss": 1.2114, |
| "step": 18500 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 4.152562494506565e-05, |
| "loss": 1.1838, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 4.643949881229301e-05, |
| "loss": 1.2117, |
| "step": 19500 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 4.933085441857718e-05, |
| "loss": 1.1867, |
| "step": 20000 |
| }, |
| { |
| "epoch": 3.0, |
| "step": 20373, |
| "total_flos": 4.098837932462899e+16, |
| "train_loss": 1.2774244995746322, |
| "train_runtime": 7751.6429, |
| "train_samples_per_second": 42.048, |
| "train_steps_per_second": 2.628 |
| } |
| ], |
| "max_steps": 20373, |
| "num_train_epochs": 3, |
| "total_flos": 4.098837932462899e+16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|