| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.9970788704965923, |
| "eval_steps": 500, |
| "global_step": 1026, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0029211295034079843, |
| "grad_norm": 6.789194746397651, |
| "learning_rate": 9.70873786407767e-08, |
| "loss": 1.0797, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.005842259006815969, |
| "grad_norm": 6.623744256841628, |
| "learning_rate": 1.941747572815534e-07, |
| "loss": 1.0936, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.008763388510223954, |
| "grad_norm": 6.814087231706784, |
| "learning_rate": 2.9126213592233014e-07, |
| "loss": 1.0915, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.011684518013631937, |
| "grad_norm": 6.979864404527602, |
| "learning_rate": 3.883495145631068e-07, |
| "loss": 1.113, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.014605647517039922, |
| "grad_norm": 7.130506329535252, |
| "learning_rate": 4.854368932038835e-07, |
| "loss": 1.1201, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.017526777020447908, |
| "grad_norm": 6.457316225305797, |
| "learning_rate": 5.825242718446603e-07, |
| "loss": 1.0589, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.02044790652385589, |
| "grad_norm": 6.634803833935598, |
| "learning_rate": 6.79611650485437e-07, |
| "loss": 1.0948, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.023369036027263874, |
| "grad_norm": 6.51924054198785, |
| "learning_rate": 7.766990291262136e-07, |
| "loss": 1.1033, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.02629016553067186, |
| "grad_norm": 6.172807303107381, |
| "learning_rate": 8.737864077669904e-07, |
| "loss": 1.1025, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.029211295034079845, |
| "grad_norm": 6.123481829151969, |
| "learning_rate": 9.70873786407767e-07, |
| "loss": 1.0593, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.03213242453748783, |
| "grad_norm": 5.067066736988861, |
| "learning_rate": 1.0679611650485437e-06, |
| "loss": 1.0514, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.035053554040895815, |
| "grad_norm": 5.316963931526159, |
| "learning_rate": 1.1650485436893206e-06, |
| "loss": 1.0649, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.0379746835443038, |
| "grad_norm": 4.8023708760386326, |
| "learning_rate": 1.2621359223300972e-06, |
| "loss": 1.0595, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.04089581304771178, |
| "grad_norm": 4.307477266760391, |
| "learning_rate": 1.359223300970874e-06, |
| "loss": 1.0045, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.043816942551119765, |
| "grad_norm": 2.9737946154856254, |
| "learning_rate": 1.4563106796116506e-06, |
| "loss": 0.9864, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.04673807205452775, |
| "grad_norm": 2.969891758892917, |
| "learning_rate": 1.5533980582524272e-06, |
| "loss": 1.0038, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.04965920155793573, |
| "grad_norm": 2.6410254985919264, |
| "learning_rate": 1.650485436893204e-06, |
| "loss": 0.9688, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.05258033106134372, |
| "grad_norm": 2.6767839969522385, |
| "learning_rate": 1.7475728155339808e-06, |
| "loss": 0.9777, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.055501460564751706, |
| "grad_norm": 2.3845837305117867, |
| "learning_rate": 1.8446601941747574e-06, |
| "loss": 0.9803, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.05842259006815969, |
| "grad_norm": 2.485668257022799, |
| "learning_rate": 1.941747572815534e-06, |
| "loss": 0.9512, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.06134371957156767, |
| "grad_norm": 2.9203150216318057, |
| "learning_rate": 2.0388349514563107e-06, |
| "loss": 0.93, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.06426484907497566, |
| "grad_norm": 2.868722547204036, |
| "learning_rate": 2.1359223300970874e-06, |
| "loss": 0.9171, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.06718597857838364, |
| "grad_norm": 2.679648431038279, |
| "learning_rate": 2.2330097087378645e-06, |
| "loss": 0.9199, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.07010710808179163, |
| "grad_norm": 2.459073576199394, |
| "learning_rate": 2.330097087378641e-06, |
| "loss": 0.9157, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.0730282375851996, |
| "grad_norm": 2.014055908464458, |
| "learning_rate": 2.427184466019418e-06, |
| "loss": 0.9164, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.0759493670886076, |
| "grad_norm": 1.5955860458303692, |
| "learning_rate": 2.5242718446601945e-06, |
| "loss": 0.8909, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.07887049659201557, |
| "grad_norm": 1.3790379745407235, |
| "learning_rate": 2.621359223300971e-06, |
| "loss": 0.8631, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.08179162609542356, |
| "grad_norm": 1.6462939916147095, |
| "learning_rate": 2.718446601941748e-06, |
| "loss": 0.8384, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.08471275559883155, |
| "grad_norm": 1.7416274692927092, |
| "learning_rate": 2.8155339805825245e-06, |
| "loss": 0.8643, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.08763388510223953, |
| "grad_norm": 1.4689264445022938, |
| "learning_rate": 2.912621359223301e-06, |
| "loss": 0.8401, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.09055501460564752, |
| "grad_norm": 1.2970139226424346, |
| "learning_rate": 3.0097087378640778e-06, |
| "loss": 0.8423, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.0934761441090555, |
| "grad_norm": 1.1674121947058942, |
| "learning_rate": 3.1067961165048544e-06, |
| "loss": 0.837, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.09639727361246349, |
| "grad_norm": 1.1908632755730892, |
| "learning_rate": 3.2038834951456315e-06, |
| "loss": 0.8203, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.09931840311587146, |
| "grad_norm": 1.1042928737436872, |
| "learning_rate": 3.300970873786408e-06, |
| "loss": 0.799, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.10223953261927946, |
| "grad_norm": 1.1603936762022113, |
| "learning_rate": 3.398058252427185e-06, |
| "loss": 0.8303, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.10516066212268745, |
| "grad_norm": 1.0681312398436076, |
| "learning_rate": 3.4951456310679615e-06, |
| "loss": 0.8052, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.10808179162609542, |
| "grad_norm": 0.8925175456575719, |
| "learning_rate": 3.592233009708738e-06, |
| "loss": 0.8069, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.11100292112950341, |
| "grad_norm": 0.8822021824942768, |
| "learning_rate": 3.689320388349515e-06, |
| "loss": 0.7812, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.11392405063291139, |
| "grad_norm": 0.8935693976115221, |
| "learning_rate": 3.7864077669902915e-06, |
| "loss": 0.7787, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.11684518013631938, |
| "grad_norm": 0.9243397791705507, |
| "learning_rate": 3.883495145631068e-06, |
| "loss": 0.7936, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.11976630963972736, |
| "grad_norm": 0.8181188534202998, |
| "learning_rate": 3.980582524271845e-06, |
| "loss": 0.7752, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.12268743914313535, |
| "grad_norm": 0.723717052367605, |
| "learning_rate": 4.0776699029126215e-06, |
| "loss": 0.7544, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.12560856864654332, |
| "grad_norm": 0.7268642480625847, |
| "learning_rate": 4.1747572815533986e-06, |
| "loss": 0.7871, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.12852969814995133, |
| "grad_norm": 0.7154200711509474, |
| "learning_rate": 4.271844660194175e-06, |
| "loss": 0.7617, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.1314508276533593, |
| "grad_norm": 0.7425189888635864, |
| "learning_rate": 4.368932038834952e-06, |
| "loss": 0.7661, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.13437195715676728, |
| "grad_norm": 0.6964525874804132, |
| "learning_rate": 4.466019417475729e-06, |
| "loss": 0.7578, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.13729308666017526, |
| "grad_norm": 0.7077568254698756, |
| "learning_rate": 4.563106796116505e-06, |
| "loss": 0.7466, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.14021421616358326, |
| "grad_norm": 0.728109048064247, |
| "learning_rate": 4.660194174757282e-06, |
| "loss": 0.7536, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.14313534566699124, |
| "grad_norm": 0.7073551382013681, |
| "learning_rate": 4.7572815533980585e-06, |
| "loss": 0.7386, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.1460564751703992, |
| "grad_norm": 0.6810942703152736, |
| "learning_rate": 4.854368932038836e-06, |
| "loss": 0.7771, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.14897760467380722, |
| "grad_norm": 0.7280450266974076, |
| "learning_rate": 4.951456310679612e-06, |
| "loss": 0.739, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.1518987341772152, |
| "grad_norm": 0.7021483972343961, |
| "learning_rate": 5.048543689320389e-06, |
| "loss": 0.7437, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.15481986368062317, |
| "grad_norm": 0.7167957676962433, |
| "learning_rate": 5.145631067961165e-06, |
| "loss": 0.764, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.15774099318403115, |
| "grad_norm": 0.6140023964252928, |
| "learning_rate": 5.242718446601942e-06, |
| "loss": 0.7338, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.16066212268743915, |
| "grad_norm": 0.7304557287827438, |
| "learning_rate": 5.3398058252427185e-06, |
| "loss": 0.7815, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.16358325219084713, |
| "grad_norm": 0.6396185255120146, |
| "learning_rate": 5.436893203883496e-06, |
| "loss": 0.7349, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.1665043816942551, |
| "grad_norm": 0.6558732255969539, |
| "learning_rate": 5.533980582524272e-06, |
| "loss": 0.7212, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.1694255111976631, |
| "grad_norm": 0.5940933645304805, |
| "learning_rate": 5.631067961165049e-06, |
| "loss": 0.7577, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.17234664070107109, |
| "grad_norm": 0.6218585913117457, |
| "learning_rate": 5.728155339805825e-06, |
| "loss": 0.7431, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.17526777020447906, |
| "grad_norm": 0.6908136674947283, |
| "learning_rate": 5.825242718446602e-06, |
| "loss": 0.7368, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.17818889970788704, |
| "grad_norm": 0.6522752947974526, |
| "learning_rate": 5.9223300970873785e-06, |
| "loss": 0.7126, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.18111002921129504, |
| "grad_norm": 0.6780346336214896, |
| "learning_rate": 6.0194174757281556e-06, |
| "loss": 0.7367, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.18403115871470302, |
| "grad_norm": 0.7722253573433183, |
| "learning_rate": 6.116504854368932e-06, |
| "loss": 0.7228, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.186952288218111, |
| "grad_norm": 0.6399768773236866, |
| "learning_rate": 6.213592233009709e-06, |
| "loss": 0.7244, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.189873417721519, |
| "grad_norm": 0.6674128228125541, |
| "learning_rate": 6.310679611650487e-06, |
| "loss": 0.7376, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.19279454722492698, |
| "grad_norm": 0.6775138710637573, |
| "learning_rate": 6.407766990291263e-06, |
| "loss": 0.7286, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.19571567672833495, |
| "grad_norm": 0.7443299390748634, |
| "learning_rate": 6.50485436893204e-06, |
| "loss": 0.7123, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.19863680623174293, |
| "grad_norm": 0.6712217876186718, |
| "learning_rate": 6.601941747572816e-06, |
| "loss": 0.7118, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.20155793573515093, |
| "grad_norm": 0.6173482204736721, |
| "learning_rate": 6.6990291262135935e-06, |
| "loss": 0.7141, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.2044790652385589, |
| "grad_norm": 0.7586706797857404, |
| "learning_rate": 6.79611650485437e-06, |
| "loss": 0.7182, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.2074001947419669, |
| "grad_norm": 0.6711328351671544, |
| "learning_rate": 6.893203883495147e-06, |
| "loss": 0.7182, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.2103213242453749, |
| "grad_norm": 0.6423529972707454, |
| "learning_rate": 6.990291262135923e-06, |
| "loss": 0.7147, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.21324245374878287, |
| "grad_norm": 0.7175157009799245, |
| "learning_rate": 7.0873786407767e-06, |
| "loss": 0.7078, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.21616358325219084, |
| "grad_norm": 0.7156449836663106, |
| "learning_rate": 7.184466019417476e-06, |
| "loss": 0.7026, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.21908471275559882, |
| "grad_norm": 0.591120785534527, |
| "learning_rate": 7.2815533980582534e-06, |
| "loss": 0.6996, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.22200584225900682, |
| "grad_norm": 0.6568235675952798, |
| "learning_rate": 7.37864077669903e-06, |
| "loss": 0.7098, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.2249269717624148, |
| "grad_norm": 0.6969907394816692, |
| "learning_rate": 7.475728155339807e-06, |
| "loss": 0.7112, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.22784810126582278, |
| "grad_norm": 0.674624972312595, |
| "learning_rate": 7.572815533980583e-06, |
| "loss": 0.6905, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.23076923076923078, |
| "grad_norm": 0.6004655479528318, |
| "learning_rate": 7.66990291262136e-06, |
| "loss": 0.6987, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.23369036027263876, |
| "grad_norm": 0.672439140786889, |
| "learning_rate": 7.766990291262136e-06, |
| "loss": 0.7059, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.23661148977604674, |
| "grad_norm": 0.6379167803971234, |
| "learning_rate": 7.864077669902913e-06, |
| "loss": 0.6993, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.2395326192794547, |
| "grad_norm": 0.624979376741808, |
| "learning_rate": 7.96116504854369e-06, |
| "loss": 0.6961, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.24245374878286272, |
| "grad_norm": 0.5967840653189634, |
| "learning_rate": 8.058252427184466e-06, |
| "loss": 0.7115, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.2453748782862707, |
| "grad_norm": 0.6448351126797235, |
| "learning_rate": 8.155339805825243e-06, |
| "loss": 0.7058, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.24829600778967867, |
| "grad_norm": 0.5456246974361787, |
| "learning_rate": 8.25242718446602e-06, |
| "loss": 0.6971, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.25121713729308665, |
| "grad_norm": 0.5981523190285354, |
| "learning_rate": 8.349514563106797e-06, |
| "loss": 0.6644, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.25413826679649465, |
| "grad_norm": 0.6094281430676193, |
| "learning_rate": 8.446601941747573e-06, |
| "loss": 0.6822, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.25705939629990265, |
| "grad_norm": 0.645486530934357, |
| "learning_rate": 8.54368932038835e-06, |
| "loss": 0.6767, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.2599805258033106, |
| "grad_norm": 0.6225273389721123, |
| "learning_rate": 8.640776699029127e-06, |
| "loss": 0.7152, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.2629016553067186, |
| "grad_norm": 0.6432259026110182, |
| "learning_rate": 8.737864077669904e-06, |
| "loss": 0.7159, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.26582278481012656, |
| "grad_norm": 0.6162731782600254, |
| "learning_rate": 8.834951456310681e-06, |
| "loss": 0.7132, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.26874391431353456, |
| "grad_norm": 0.6542304058964258, |
| "learning_rate": 8.932038834951458e-06, |
| "loss": 0.6809, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.27166504381694256, |
| "grad_norm": 0.6079172030969984, |
| "learning_rate": 9.029126213592233e-06, |
| "loss": 0.6824, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.2745861733203505, |
| "grad_norm": 0.6519383743853913, |
| "learning_rate": 9.12621359223301e-06, |
| "loss": 0.6899, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.2775073028237585, |
| "grad_norm": 0.7282902638094394, |
| "learning_rate": 9.223300970873788e-06, |
| "loss": 0.6922, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.2804284323271665, |
| "grad_norm": 0.5989786182176935, |
| "learning_rate": 9.320388349514565e-06, |
| "loss": 0.6734, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.28334956183057447, |
| "grad_norm": 0.7160709546948157, |
| "learning_rate": 9.41747572815534e-06, |
| "loss": 0.6815, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.2862706913339825, |
| "grad_norm": 0.6813096412081009, |
| "learning_rate": 9.514563106796117e-06, |
| "loss": 0.6885, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.2891918208373905, |
| "grad_norm": 0.6540975722149734, |
| "learning_rate": 9.611650485436894e-06, |
| "loss": 0.69, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.2921129503407984, |
| "grad_norm": 0.8050571281257926, |
| "learning_rate": 9.708737864077671e-06, |
| "loss": 0.691, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.29503407984420643, |
| "grad_norm": 0.6433559989032654, |
| "learning_rate": 9.805825242718447e-06, |
| "loss": 0.6986, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.29795520934761444, |
| "grad_norm": 0.8412736504392738, |
| "learning_rate": 9.902912621359224e-06, |
| "loss": 0.6926, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.3008763388510224, |
| "grad_norm": 0.7443726086923518, |
| "learning_rate": 1e-05, |
| "loss": 0.6906, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.3037974683544304, |
| "grad_norm": 0.6597015006842325, |
| "learning_rate": 9.999971037507608e-06, |
| "loss": 0.675, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.30671859785783834, |
| "grad_norm": 0.5931435595661035, |
| "learning_rate": 9.99988415036596e-06, |
| "loss": 0.6802, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.30963972736124634, |
| "grad_norm": 0.7214217523040783, |
| "learning_rate": 9.99973933958164e-06, |
| "loss": 0.7041, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.31256085686465435, |
| "grad_norm": 0.7234513096207073, |
| "learning_rate": 9.999536606832288e-06, |
| "loss": 0.6872, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.3154819863680623, |
| "grad_norm": 0.7879752038918911, |
| "learning_rate": 9.999275954466555e-06, |
| "loss": 0.6873, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.3184031158714703, |
| "grad_norm": 0.6655845938433153, |
| "learning_rate": 9.998957385504103e-06, |
| "loss": 0.6976, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.3213242453748783, |
| "grad_norm": 0.8522730059744493, |
| "learning_rate": 9.99858090363555e-06, |
| "loss": 0.6719, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.32424537487828625, |
| "grad_norm": 0.6292219731062122, |
| "learning_rate": 9.998146513222436e-06, |
| "loss": 0.6993, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.32716650438169426, |
| "grad_norm": 0.7181176135878521, |
| "learning_rate": 9.997654219297176e-06, |
| "loss": 0.6901, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.33008763388510226, |
| "grad_norm": 0.6962928810640735, |
| "learning_rate": 9.997104027562991e-06, |
| "loss": 0.6951, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.3330087633885102, |
| "grad_norm": 0.6849230790401417, |
| "learning_rate": 9.996495944393853e-06, |
| "loss": 0.6828, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.3359298928919182, |
| "grad_norm": 0.7135171270031142, |
| "learning_rate": 9.995829976834402e-06, |
| "loss": 0.6737, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.3388510223953262, |
| "grad_norm": 0.6814570642632325, |
| "learning_rate": 9.995106132599869e-06, |
| "loss": 0.6875, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.34177215189873417, |
| "grad_norm": 0.8248995841087691, |
| "learning_rate": 9.99432442007599e-06, |
| "loss": 0.6873, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.34469328140214217, |
| "grad_norm": 0.6919031766051941, |
| "learning_rate": 9.993484848318899e-06, |
| "loss": 0.6835, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.3476144109055501, |
| "grad_norm": 0.7748697122331325, |
| "learning_rate": 9.992587427055036e-06, |
| "loss": 0.6933, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.3505355404089581, |
| "grad_norm": 0.6232356121853884, |
| "learning_rate": 9.99163216668102e-06, |
| "loss": 0.6634, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.35345666991236613, |
| "grad_norm": 0.8152682417581196, |
| "learning_rate": 9.990619078263543e-06, |
| "loss": 0.6833, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.3563777994157741, |
| "grad_norm": 0.6829212246748637, |
| "learning_rate": 9.989548173539229e-06, |
| "loss": 0.6904, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.3592989289191821, |
| "grad_norm": 0.7764702428398512, |
| "learning_rate": 9.988419464914505e-06, |
| "loss": 0.6911, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.3622200584225901, |
| "grad_norm": 0.792332480824063, |
| "learning_rate": 9.98723296546546e-06, |
| "loss": 0.6817, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.36514118792599803, |
| "grad_norm": 0.6594623726005864, |
| "learning_rate": 9.985988688937684e-06, |
| "loss": 0.6873, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.36806231742940604, |
| "grad_norm": 0.8269823477396988, |
| "learning_rate": 9.984686649746119e-06, |
| "loss": 0.693, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.37098344693281404, |
| "grad_norm": 0.6422126697095933, |
| "learning_rate": 9.983326862974882e-06, |
| "loss": 0.6576, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.373904576436222, |
| "grad_norm": 0.8200568586438982, |
| "learning_rate": 9.981909344377101e-06, |
| "loss": 0.6929, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.37682570593963, |
| "grad_norm": 0.7400126500706381, |
| "learning_rate": 9.980434110374725e-06, |
| "loss": 0.6557, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.379746835443038, |
| "grad_norm": 0.713854878991036, |
| "learning_rate": 9.978901178058333e-06, |
| "loss": 0.6942, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.38266796494644595, |
| "grad_norm": 0.7483541113835968, |
| "learning_rate": 9.977310565186945e-06, |
| "loss": 0.6781, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.38558909444985395, |
| "grad_norm": 0.7131870241688308, |
| "learning_rate": 9.975662290187802e-06, |
| "loss": 0.6564, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.3885102239532619, |
| "grad_norm": 0.7880618680074065, |
| "learning_rate": 9.973956372156166e-06, |
| "loss": 0.6752, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.3914313534566699, |
| "grad_norm": 0.5977156256261835, |
| "learning_rate": 9.972192830855095e-06, |
| "loss": 0.6763, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.3943524829600779, |
| "grad_norm": 0.76186235107972, |
| "learning_rate": 9.970371686715205e-06, |
| "loss": 0.7014, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.39727361246348586, |
| "grad_norm": 0.8438134698760479, |
| "learning_rate": 9.96849296083445e-06, |
| "loss": 0.6902, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.40019474196689386, |
| "grad_norm": 0.6120514167187466, |
| "learning_rate": 9.966556674977864e-06, |
| "loss": 0.663, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.40311587147030187, |
| "grad_norm": 0.7891889540366791, |
| "learning_rate": 9.964562851577307e-06, |
| "loss": 0.6458, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.4060370009737098, |
| "grad_norm": 0.8381022052177086, |
| "learning_rate": 9.962511513731219e-06, |
| "loss": 0.6728, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.4089581304771178, |
| "grad_norm": 0.7489332182994131, |
| "learning_rate": 9.960402685204347e-06, |
| "loss": 0.6585, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.4118792599805258, |
| "grad_norm": 0.804869829411989, |
| "learning_rate": 9.958236390427458e-06, |
| "loss": 0.6784, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.4148003894839338, |
| "grad_norm": 0.7474713341428297, |
| "learning_rate": 9.956012654497073e-06, |
| "loss": 0.6488, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.4177215189873418, |
| "grad_norm": 0.8115239389486957, |
| "learning_rate": 9.953731503175166e-06, |
| "loss": 0.6894, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.4206426484907498, |
| "grad_norm": 0.781168520393115, |
| "learning_rate": 9.951392962888868e-06, |
| "loss": 0.6534, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.42356377799415773, |
| "grad_norm": 0.7722795439014734, |
| "learning_rate": 9.948997060730161e-06, |
| "loss": 0.6504, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.42648490749756574, |
| "grad_norm": 0.8668435801796398, |
| "learning_rate": 9.946543824455563e-06, |
| "loss": 0.6507, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.4294060370009737, |
| "grad_norm": 0.8391877814865175, |
| "learning_rate": 9.94403328248581e-06, |
| "loss": 0.6702, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.4323271665043817, |
| "grad_norm": 0.6805575537389376, |
| "learning_rate": 9.941465463905522e-06, |
| "loss": 0.6744, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.4352482960077897, |
| "grad_norm": 0.7033559759136386, |
| "learning_rate": 9.938840398462872e-06, |
| "loss": 0.6732, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.43816942551119764, |
| "grad_norm": 0.8071774421297472, |
| "learning_rate": 9.936158116569231e-06, |
| "loss": 0.6704, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.44109055501460565, |
| "grad_norm": 0.7788309365778397, |
| "learning_rate": 9.933418649298831e-06, |
| "loss": 0.6697, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.44401168451801365, |
| "grad_norm": 0.633484330960411, |
| "learning_rate": 9.930622028388388e-06, |
| "loss": 0.6533, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.4469328140214216, |
| "grad_norm": 0.6677671353777382, |
| "learning_rate": 9.92776828623675e-06, |
| "loss": 0.6323, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.4498539435248296, |
| "grad_norm": 0.6580113405415786, |
| "learning_rate": 9.924857455904511e-06, |
| "loss": 0.6569, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.4527750730282376, |
| "grad_norm": 0.7153967279296722, |
| "learning_rate": 9.921889571113629e-06, |
| "loss": 0.6651, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.45569620253164556, |
| "grad_norm": 0.6507727168616313, |
| "learning_rate": 9.918864666247042e-06, |
| "loss": 0.6709, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.45861733203505356, |
| "grad_norm": 0.6614595067208825, |
| "learning_rate": 9.915782776348263e-06, |
| "loss": 0.6558, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.46153846153846156, |
| "grad_norm": 0.8137858393206525, |
| "learning_rate": 9.912643937120978e-06, |
| "loss": 0.6756, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.4644595910418695, |
| "grad_norm": 0.6306212983973334, |
| "learning_rate": 9.909448184928629e-06, |
| "loss": 0.6416, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.4673807205452775, |
| "grad_norm": 0.6836208681318199, |
| "learning_rate": 9.906195556793996e-06, |
| "loss": 0.6662, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.47030185004868547, |
| "grad_norm": 0.6141476817991299, |
| "learning_rate": 9.902886090398764e-06, |
| "loss": 0.6774, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.47322297955209347, |
| "grad_norm": 0.7052661650958192, |
| "learning_rate": 9.899519824083095e-06, |
| "loss": 0.6618, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.4761441090555015, |
| "grad_norm": 0.5986401947237227, |
| "learning_rate": 9.896096796845172e-06, |
| "loss": 0.6738, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.4790652385589094, |
| "grad_norm": 0.6916970139226505, |
| "learning_rate": 9.892617048340754e-06, |
| "loss": 0.6588, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.4819863680623174, |
| "grad_norm": 0.6805676670949773, |
| "learning_rate": 9.889080618882719e-06, |
| "loss": 0.6826, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.48490749756572543, |
| "grad_norm": 0.6658431348883423, |
| "learning_rate": 9.88548754944059e-06, |
| "loss": 0.6702, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.4878286270691334, |
| "grad_norm": 0.682259295410329, |
| "learning_rate": 9.881837881640064e-06, |
| "loss": 0.6735, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.4907497565725414, |
| "grad_norm": 0.5993509164653038, |
| "learning_rate": 9.878131657762535e-06, |
| "loss": 0.6545, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.4936708860759494, |
| "grad_norm": 0.7534948952289838, |
| "learning_rate": 9.874368920744594e-06, |
| "loss": 0.6812, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.49659201557935734, |
| "grad_norm": 0.707512873305316, |
| "learning_rate": 9.870549714177538e-06, |
| "loss": 0.6513, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.49951314508276534, |
| "grad_norm": 0.7200355139629895, |
| "learning_rate": 9.866674082306861e-06, |
| "loss": 0.6438, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.5024342745861733, |
| "grad_norm": 0.8156295486958224, |
| "learning_rate": 9.86274207003175e-06, |
| "loss": 0.6564, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.5053554040895814, |
| "grad_norm": 0.6117511012510686, |
| "learning_rate": 9.858753722904552e-06, |
| "loss": 0.6827, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.5082765335929893, |
| "grad_norm": 0.8219493047877929, |
| "learning_rate": 9.854709087130261e-06, |
| "loss": 0.6718, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.5111976630963972, |
| "grad_norm": 0.6662487126424073, |
| "learning_rate": 9.850608209565967e-06, |
| "loss": 0.6388, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.5141187925998053, |
| "grad_norm": 0.7125438710540766, |
| "learning_rate": 9.84645113772032e-06, |
| "loss": 0.6589, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.5170399221032133, |
| "grad_norm": 0.7487609228616714, |
| "learning_rate": 9.842237919752994e-06, |
| "loss": 0.6544, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.5199610516066212, |
| "grad_norm": 0.6468420309416394, |
| "learning_rate": 9.8379686044741e-06, |
| "loss": 0.6565, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.5228821811100293, |
| "grad_norm": 0.7546649500098744, |
| "learning_rate": 9.833643241343642e-06, |
| "loss": 0.6647, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.5258033106134372, |
| "grad_norm": 0.6258186572488958, |
| "learning_rate": 9.829261880470941e-06, |
| "loss": 0.6392, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.5287244401168452, |
| "grad_norm": 0.5997447293335689, |
| "learning_rate": 9.82482457261405e-06, |
| "loss": 0.6398, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.5316455696202531, |
| "grad_norm": 0.6452320212378018, |
| "learning_rate": 9.820331369179166e-06, |
| "loss": 0.6611, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.5345666991236612, |
| "grad_norm": 0.5580735581912285, |
| "learning_rate": 9.815782322220036e-06, |
| "loss": 0.6548, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.5374878286270691, |
| "grad_norm": 0.6058080142971995, |
| "learning_rate": 9.811177484437357e-06, |
| "loss": 0.6664, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.5404089581304771, |
| "grad_norm": 0.7487147947448509, |
| "learning_rate": 9.806516909178161e-06, |
| "loss": 0.665, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.5433300876338851, |
| "grad_norm": 0.5994532711002538, |
| "learning_rate": 9.801800650435194e-06, |
| "loss": 0.6345, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.5462512171372931, |
| "grad_norm": 0.6402638834025774, |
| "learning_rate": 9.797028762846305e-06, |
| "loss": 0.6689, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.549172346640701, |
| "grad_norm": 0.8110120277002857, |
| "learning_rate": 9.792201301693793e-06, |
| "loss": 0.6623, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.5520934761441091, |
| "grad_norm": 0.6022982451173915, |
| "learning_rate": 9.787318322903784e-06, |
| "loss": 0.642, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.555014605647517, |
| "grad_norm": 0.8359964918822578, |
| "learning_rate": 9.78237988304557e-06, |
| "loss": 0.6828, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.557935735150925, |
| "grad_norm": 0.5971974835914099, |
| "learning_rate": 9.77738603933096e-06, |
| "loss": 0.6637, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.560856864654333, |
| "grad_norm": 0.9007553762322157, |
| "learning_rate": 9.772336849613624e-06, |
| "loss": 0.6489, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.563777994157741, |
| "grad_norm": 0.6097333173160772, |
| "learning_rate": 9.767232372388406e-06, |
| "loss": 0.6195, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.5666991236611489, |
| "grad_norm": 0.7430323349181741, |
| "learning_rate": 9.762072666790658e-06, |
| "loss": 0.6602, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.569620253164557, |
| "grad_norm": 0.698590106062137, |
| "learning_rate": 9.756857792595555e-06, |
| "loss": 0.654, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.572541382667965, |
| "grad_norm": 0.6098458915248055, |
| "learning_rate": 9.751587810217398e-06, |
| "loss": 0.6571, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.5754625121713729, |
| "grad_norm": 0.6600018141821303, |
| "learning_rate": 9.746262780708919e-06, |
| "loss": 0.6572, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.578383641674781, |
| "grad_norm": 0.6601639946678165, |
| "learning_rate": 9.740882765760567e-06, |
| "loss": 0.6593, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.5813047711781889, |
| "grad_norm": 0.726967683938266, |
| "learning_rate": 9.735447827699798e-06, |
| "loss": 0.6573, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.5842259006815969, |
| "grad_norm": 0.6076134837821863, |
| "learning_rate": 9.729958029490353e-06, |
| "loss": 0.6495, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.5871470301850049, |
| "grad_norm": 0.693728259825805, |
| "learning_rate": 9.72441343473153e-06, |
| "loss": 0.6384, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.5900681596884129, |
| "grad_norm": 0.7278423168034551, |
| "learning_rate": 9.718814107657441e-06, |
| "loss": 0.6584, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.5929892891918208, |
| "grad_norm": 0.7100671054561837, |
| "learning_rate": 9.713160113136272e-06, |
| "loss": 0.6555, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.5959104186952289, |
| "grad_norm": 0.6955835624438068, |
| "learning_rate": 9.707451516669533e-06, |
| "loss": 0.6581, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.5988315481986368, |
| "grad_norm": 0.6862859891275203, |
| "learning_rate": 9.701688384391296e-06, |
| "loss": 0.6471, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.6017526777020448, |
| "grad_norm": 0.7918106833642026, |
| "learning_rate": 9.695870783067434e-06, |
| "loss": 0.6351, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.6046738072054528, |
| "grad_norm": 0.762255183423834, |
| "learning_rate": 9.689998780094839e-06, |
| "loss": 0.6464, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.6075949367088608, |
| "grad_norm": 0.5926349601655899, |
| "learning_rate": 9.684072443500645e-06, |
| "loss": 0.6342, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.6105160662122687, |
| "grad_norm": 0.863234457455766, |
| "learning_rate": 9.678091841941446e-06, |
| "loss": 0.653, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.6134371957156767, |
| "grad_norm": 0.7588656251837851, |
| "learning_rate": 9.672057044702492e-06, |
| "loss": 0.6379, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.6163583252190847, |
| "grad_norm": 0.6108635991637165, |
| "learning_rate": 9.665968121696892e-06, |
| "loss": 0.6605, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.6192794547224927, |
| "grad_norm": 0.8356230484629192, |
| "learning_rate": 9.659825143464798e-06, |
| "loss": 0.6458, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.6222005842259006, |
| "grad_norm": 0.6240730332192024, |
| "learning_rate": 9.653628181172596e-06, |
| "loss": 0.6506, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.6251217137293087, |
| "grad_norm": 0.6689297135107584, |
| "learning_rate": 9.647377306612075e-06, |
| "loss": 0.6299, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.6280428432327166, |
| "grad_norm": 0.7685374427252067, |
| "learning_rate": 9.641072592199599e-06, |
| "loss": 0.6634, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.6309639727361246, |
| "grad_norm": 0.6331465350705314, |
| "learning_rate": 9.634714110975263e-06, |
| "loss": 0.6705, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.6338851022395326, |
| "grad_norm": 0.7142109675799595, |
| "learning_rate": 9.628301936602053e-06, |
| "loss": 0.6539, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.6368062317429406, |
| "grad_norm": 0.8729837863809322, |
| "learning_rate": 9.62183614336499e-06, |
| "loss": 0.6596, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.6397273612463485, |
| "grad_norm": 0.5991176634004923, |
| "learning_rate": 9.61531680617027e-06, |
| "loss": 0.6656, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.6426484907497566, |
| "grad_norm": 0.7145894148388473, |
| "learning_rate": 9.608744000544392e-06, |
| "loss": 0.6643, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.6455696202531646, |
| "grad_norm": 0.7598912110336243, |
| "learning_rate": 9.602117802633293e-06, |
| "loss": 0.6291, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.6484907497565725, |
| "grad_norm": 0.6815208811474045, |
| "learning_rate": 9.595438289201453e-06, |
| "loss": 0.6472, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.6514118792599806, |
| "grad_norm": 0.5760568859368258, |
| "learning_rate": 9.588705537631014e-06, |
| "loss": 0.6563, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.6543330087633885, |
| "grad_norm": 0.7022056147869815, |
| "learning_rate": 9.581919625920886e-06, |
| "loss": 0.6524, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.6572541382667965, |
| "grad_norm": 0.67470404537858, |
| "learning_rate": 9.575080632685832e-06, |
| "loss": 0.6436, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.6601752677702045, |
| "grad_norm": 0.6156654623500659, |
| "learning_rate": 9.568188637155569e-06, |
| "loss": 0.6256, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.6630963972736125, |
| "grad_norm": 0.6870167927139845, |
| "learning_rate": 9.561243719173844e-06, |
| "loss": 0.628, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.6660175267770204, |
| "grad_norm": 0.6043043393160271, |
| "learning_rate": 9.554245959197511e-06, |
| "loss": 0.6631, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.6689386562804285, |
| "grad_norm": 0.6424377779531785, |
| "learning_rate": 9.5471954382956e-06, |
| "loss": 0.6455, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.6718597857838364, |
| "grad_norm": 0.6747246252989533, |
| "learning_rate": 9.54009223814837e-06, |
| "loss": 0.6482, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.6747809152872444, |
| "grad_norm": 0.6198749042634925, |
| "learning_rate": 9.532936441046376e-06, |
| "loss": 0.6679, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.6777020447906524, |
| "grad_norm": 0.6700871145900489, |
| "learning_rate": 9.525728129889505e-06, |
| "loss": 0.6704, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.6806231742940604, |
| "grad_norm": 0.6486221862584836, |
| "learning_rate": 9.51846738818602e-06, |
| "loss": 0.6533, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.6835443037974683, |
| "grad_norm": 0.6158430372621209, |
| "learning_rate": 9.511154300051591e-06, |
| "loss": 0.6391, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.6864654333008764, |
| "grad_norm": 0.6059458334544432, |
| "learning_rate": 9.503788950208324e-06, |
| "loss": 0.6326, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.6893865628042843, |
| "grad_norm": 0.6630441105155737, |
| "learning_rate": 9.49637142398377e-06, |
| "loss": 0.6419, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.6923076923076923, |
| "grad_norm": 0.7121010143311686, |
| "learning_rate": 9.48890180730995e-06, |
| "loss": 0.6366, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.6952288218111002, |
| "grad_norm": 0.6536311654933115, |
| "learning_rate": 9.481380186722354e-06, |
| "loss": 0.6475, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.6981499513145083, |
| "grad_norm": 0.5870586837084283, |
| "learning_rate": 9.473806649358929e-06, |
| "loss": 0.6664, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.7010710808179162, |
| "grad_norm": 0.7293444660664181, |
| "learning_rate": 9.466181282959083e-06, |
| "loss": 0.6294, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.7039922103213242, |
| "grad_norm": 0.6803091119725557, |
| "learning_rate": 9.458504175862665e-06, |
| "loss": 0.6543, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.7069133398247323, |
| "grad_norm": 0.513109460376802, |
| "learning_rate": 9.450775417008936e-06, |
| "loss": 0.6529, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.7098344693281402, |
| "grad_norm": 0.6591044352211995, |
| "learning_rate": 9.442995095935542e-06, |
| "loss": 0.6485, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.7127555988315482, |
| "grad_norm": 0.5639394652214005, |
| "learning_rate": 9.43516330277748e-06, |
| "loss": 0.6354, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.7156767283349562, |
| "grad_norm": 0.5382276491132706, |
| "learning_rate": 9.427280128266049e-06, |
| "loss": 0.6338, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.7185978578383642, |
| "grad_norm": 0.5783621915913141, |
| "learning_rate": 9.419345663727805e-06, |
| "loss": 0.6541, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.7215189873417721, |
| "grad_norm": 0.5457758477722148, |
| "learning_rate": 9.411360001083496e-06, |
| "loss": 0.6649, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.7244401168451802, |
| "grad_norm": 0.5701118395223765, |
| "learning_rate": 9.403323232846994e-06, |
| "loss": 0.6305, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.7273612463485881, |
| "grad_norm": 0.5986045250901076, |
| "learning_rate": 9.395235452124239e-06, |
| "loss": 0.6315, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.7302823758519961, |
| "grad_norm": 0.5915757822980239, |
| "learning_rate": 9.387096752612144e-06, |
| "loss": 0.6563, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.7332035053554041, |
| "grad_norm": 0.6447044009504002, |
| "learning_rate": 9.378907228597518e-06, |
| "loss": 0.6543, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.7361246348588121, |
| "grad_norm": 0.6146190085950654, |
| "learning_rate": 9.370666974955973e-06, |
| "loss": 0.6474, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.73904576436222, |
| "grad_norm": 0.6186340229955254, |
| "learning_rate": 9.362376087150822e-06, |
| "loss": 0.6498, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.7419668938656281, |
| "grad_norm": 0.6697188050730257, |
| "learning_rate": 9.354034661231976e-06, |
| "loss": 0.629, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.744888023369036, |
| "grad_norm": 0.7166126450253048, |
| "learning_rate": 9.345642793834825e-06, |
| "loss": 0.6476, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.747809152872444, |
| "grad_norm": 0.5909733136537622, |
| "learning_rate": 9.337200582179134e-06, |
| "loss": 0.6338, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.750730282375852, |
| "grad_norm": 0.6690338784895201, |
| "learning_rate": 9.328708124067893e-06, |
| "loss": 0.6425, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.75365141187926, |
| "grad_norm": 0.6170552061635086, |
| "learning_rate": 9.320165517886207e-06, |
| "loss": 0.649, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.7565725413826679, |
| "grad_norm": 0.6767600272762853, |
| "learning_rate": 9.31157286260014e-06, |
| "loss": 0.6496, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.759493670886076, |
| "grad_norm": 0.7264560653599718, |
| "learning_rate": 9.302930257755579e-06, |
| "loss": 0.6583, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.762414800389484, |
| "grad_norm": 0.5453999254984644, |
| "learning_rate": 9.294237803477076e-06, |
| "loss": 0.648, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.7653359298928919, |
| "grad_norm": 0.7523663818704205, |
| "learning_rate": 9.285495600466683e-06, |
| "loss": 0.6488, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.7682570593963, |
| "grad_norm": 0.7485014295621598, |
| "learning_rate": 9.276703750002801e-06, |
| "loss": 0.6255, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.7711781888997079, |
| "grad_norm": 0.686777506781002, |
| "learning_rate": 9.267862353938988e-06, |
| "loss": 0.6534, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.7740993184031159, |
| "grad_norm": 0.7655220545156425, |
| "learning_rate": 9.258971514702789e-06, |
| "loss": 0.6439, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.7770204479065238, |
| "grad_norm": 0.5995690372205543, |
| "learning_rate": 9.250031335294551e-06, |
| "loss": 0.6264, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.7799415774099319, |
| "grad_norm": 0.6306408667729854, |
| "learning_rate": 9.241041919286227e-06, |
| "loss": 0.633, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.7828627069133398, |
| "grad_norm": 0.7405970033463782, |
| "learning_rate": 9.232003370820171e-06, |
| "loss": 0.6355, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.7857838364167478, |
| "grad_norm": 0.5761877077710947, |
| "learning_rate": 9.222915794607942e-06, |
| "loss": 0.6431, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.7887049659201558, |
| "grad_norm": 0.6417671555725009, |
| "learning_rate": 9.213779295929082e-06, |
| "loss": 0.6302, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.7916260954235638, |
| "grad_norm": 0.5476966311891922, |
| "learning_rate": 9.204593980629898e-06, |
| "loss": 0.6307, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.7945472249269717, |
| "grad_norm": 0.6325575856222458, |
| "learning_rate": 9.195359955122244e-06, |
| "loss": 0.6316, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.7974683544303798, |
| "grad_norm": 0.6487642715668982, |
| "learning_rate": 9.186077326382275e-06, |
| "loss": 0.6324, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.8003894839337877, |
| "grad_norm": 0.5751549331819923, |
| "learning_rate": 9.176746201949216e-06, |
| "loss": 0.6585, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.8033106134371957, |
| "grad_norm": 0.6208214920916966, |
| "learning_rate": 9.167366689924116e-06, |
| "loss": 0.6517, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.8062317429406037, |
| "grad_norm": 0.6579852013112687, |
| "learning_rate": 9.157938898968594e-06, |
| "loss": 0.643, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.8091528724440117, |
| "grad_norm": 0.5985871969783593, |
| "learning_rate": 9.14846293830358e-06, |
| "loss": 0.6386, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.8120740019474196, |
| "grad_norm": 0.5776495482063276, |
| "learning_rate": 9.138938917708047e-06, |
| "loss": 0.6367, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.8149951314508277, |
| "grad_norm": 0.6461097088775256, |
| "learning_rate": 9.129366947517746e-06, |
| "loss": 0.6311, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.8179162609542356, |
| "grad_norm": 0.6054324701596054, |
| "learning_rate": 9.119747138623925e-06, |
| "loss": 0.6365, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.8208373904576436, |
| "grad_norm": 0.6046664159813615, |
| "learning_rate": 9.110079602472035e-06, |
| "loss": 0.6549, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.8237585199610516, |
| "grad_norm": 0.690747949343666, |
| "learning_rate": 9.100364451060457e-06, |
| "loss": 0.6477, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.8266796494644596, |
| "grad_norm": 0.6352884441285447, |
| "learning_rate": 9.090601796939192e-06, |
| "loss": 0.6315, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.8296007789678675, |
| "grad_norm": 0.7031617950372325, |
| "learning_rate": 9.080791753208553e-06, |
| "loss": 0.6304, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.8325219084712756, |
| "grad_norm": 0.6955969158961154, |
| "learning_rate": 9.070934433517872e-06, |
| "loss": 0.6371, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.8354430379746836, |
| "grad_norm": 0.8508450672946094, |
| "learning_rate": 9.061029952064165e-06, |
| "loss": 0.6392, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.8383641674780915, |
| "grad_norm": 0.6682143045272909, |
| "learning_rate": 9.05107842359082e-06, |
| "loss": 0.6354, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.8412852969814996, |
| "grad_norm": 0.730935179569577, |
| "learning_rate": 9.041079963386263e-06, |
| "loss": 0.6365, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.8442064264849075, |
| "grad_norm": 0.9284592753787645, |
| "learning_rate": 9.031034687282627e-06, |
| "loss": 0.6512, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.8471275559883155, |
| "grad_norm": 0.5711740420462373, |
| "learning_rate": 9.020942711654404e-06, |
| "loss": 0.6253, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.8500486854917235, |
| "grad_norm": 0.7521676041243963, |
| "learning_rate": 9.0108041534171e-06, |
| "loss": 0.6346, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.8529698149951315, |
| "grad_norm": 0.7268819094436544, |
| "learning_rate": 9.000619130025885e-06, |
| "loss": 0.6321, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.8558909444985394, |
| "grad_norm": 0.6255002973375909, |
| "learning_rate": 8.99038775947422e-06, |
| "loss": 0.6448, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.8588120740019474, |
| "grad_norm": 0.5654580978730834, |
| "learning_rate": 8.980110160292503e-06, |
| "loss": 0.6546, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.8617332035053554, |
| "grad_norm": 0.5744283394449378, |
| "learning_rate": 8.969786451546691e-06, |
| "loss": 0.6354, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.8646543330087634, |
| "grad_norm": 0.5607277367845812, |
| "learning_rate": 8.959416752836915e-06, |
| "loss": 0.6315, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.8675754625121713, |
| "grad_norm": 0.6883032069319117, |
| "learning_rate": 8.949001184296107e-06, |
| "loss": 0.6284, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.8704965920155794, |
| "grad_norm": 0.5793320184156081, |
| "learning_rate": 8.938539866588593e-06, |
| "loss": 0.6299, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.8734177215189873, |
| "grad_norm": 0.6122270930586687, |
| "learning_rate": 8.928032920908709e-06, |
| "loss": 0.6463, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.8763388510223953, |
| "grad_norm": 0.6718320925279334, |
| "learning_rate": 8.917480468979387e-06, |
| "loss": 0.6405, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.8792599805258033, |
| "grad_norm": 0.6132181340646334, |
| "learning_rate": 8.906882633050753e-06, |
| "loss": 0.6426, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.8821811100292113, |
| "grad_norm": 0.6206805066838101, |
| "learning_rate": 8.896239535898702e-06, |
| "loss": 0.65, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.8851022395326192, |
| "grad_norm": 0.7456156741229341, |
| "learning_rate": 8.885551300823483e-06, |
| "loss": 0.6395, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.8880233690360273, |
| "grad_norm": 0.6186279743363227, |
| "learning_rate": 8.874818051648267e-06, |
| "loss": 0.6236, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.8909444985394352, |
| "grad_norm": 0.7192434268898347, |
| "learning_rate": 8.864039912717713e-06, |
| "loss": 0.6427, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.8938656280428432, |
| "grad_norm": 0.6528244466221247, |
| "learning_rate": 8.853217008896526e-06, |
| "loss": 0.6478, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.8967867575462513, |
| "grad_norm": 0.6240468261028331, |
| "learning_rate": 8.842349465568018e-06, |
| "loss": 0.6354, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.8997078870496592, |
| "grad_norm": 0.6086442105860419, |
| "learning_rate": 8.831437408632639e-06, |
| "loss": 0.6175, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.9026290165530672, |
| "grad_norm": 0.6206978097636743, |
| "learning_rate": 8.820480964506542e-06, |
| "loss": 0.6329, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.9055501460564752, |
| "grad_norm": 0.67571865721595, |
| "learning_rate": 8.809480260120096e-06, |
| "loss": 0.6302, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.9084712755598832, |
| "grad_norm": 0.612775453081801, |
| "learning_rate": 8.798435422916425e-06, |
| "loss": 0.6248, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.9113924050632911, |
| "grad_norm": 0.5939423168299965, |
| "learning_rate": 8.787346580849939e-06, |
| "loss": 0.6274, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.9143135345666992, |
| "grad_norm": 0.6578489213437513, |
| "learning_rate": 8.776213862384838e-06, |
| "loss": 0.6367, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.9172346640701071, |
| "grad_norm": 0.7439730364901744, |
| "learning_rate": 8.76503739649363e-06, |
| "loss": 0.6584, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.9201557935735151, |
| "grad_norm": 0.5758626313580656, |
| "learning_rate": 8.753817312655642e-06, |
| "loss": 0.638, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.9230769230769231, |
| "grad_norm": 0.7372145224476075, |
| "learning_rate": 8.742553740855507e-06, |
| "loss": 0.6391, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.9259980525803311, |
| "grad_norm": 0.7004119821194315, |
| "learning_rate": 8.73124681158167e-06, |
| "loss": 0.6426, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.928919182083739, |
| "grad_norm": 0.5393578063337247, |
| "learning_rate": 8.719896655824878e-06, |
| "loss": 0.6326, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.9318403115871471, |
| "grad_norm": 0.5698772754745327, |
| "learning_rate": 8.708503405076646e-06, |
| "loss": 0.634, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.934761441090555, |
| "grad_norm": 0.588256486663727, |
| "learning_rate": 8.697067191327748e-06, |
| "loss": 0.6328, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.937682570593963, |
| "grad_norm": 0.6194561945270076, |
| "learning_rate": 8.685588147066688e-06, |
| "loss": 0.6303, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.9406037000973709, |
| "grad_norm": 0.5910869397544244, |
| "learning_rate": 8.67406640527816e-06, |
| "loss": 0.6494, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.943524829600779, |
| "grad_norm": 0.6055073823945398, |
| "learning_rate": 8.662502099441505e-06, |
| "loss": 0.6334, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.9464459591041869, |
| "grad_norm": 0.5417594595582104, |
| "learning_rate": 8.650895363529172e-06, |
| "loss": 0.6279, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.9493670886075949, |
| "grad_norm": 0.5878187442362004, |
| "learning_rate": 8.639246332005163e-06, |
| "loss": 0.639, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.952288218111003, |
| "grad_norm": 0.5233625621707794, |
| "learning_rate": 8.627555139823468e-06, |
| "loss": 0.614, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.9552093476144109, |
| "grad_norm": 0.5656952115933153, |
| "learning_rate": 8.615821922426517e-06, |
| "loss": 0.6214, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.9581304771178188, |
| "grad_norm": 0.5806200502649499, |
| "learning_rate": 8.604046815743598e-06, |
| "loss": 0.6424, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.9610516066212269, |
| "grad_norm": 0.5561534960958242, |
| "learning_rate": 8.592229956189283e-06, |
| "loss": 0.638, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.9639727361246349, |
| "grad_norm": 0.5340425065767049, |
| "learning_rate": 8.580371480661857e-06, |
| "loss": 0.6238, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.9668938656280428, |
| "grad_norm": 0.5501524304300768, |
| "learning_rate": 8.568471526541721e-06, |
| "loss": 0.6518, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.9698149951314509, |
| "grad_norm": 0.571968183721703, |
| "learning_rate": 8.556530231689809e-06, |
| "loss": 0.6588, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.9727361246348588, |
| "grad_norm": 0.5870934675467651, |
| "learning_rate": 8.544547734445983e-06, |
| "loss": 0.6441, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.9756572541382668, |
| "grad_norm": 0.5633142366316923, |
| "learning_rate": 8.532524173627438e-06, |
| "loss": 0.6558, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.9785783836416748, |
| "grad_norm": 0.6080503864783372, |
| "learning_rate": 8.520459688527091e-06, |
| "loss": 0.6239, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.9814995131450828, |
| "grad_norm": 0.6010534017830508, |
| "learning_rate": 8.508354418911966e-06, |
| "loss": 0.6392, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.9844206426484907, |
| "grad_norm": 0.6110941694269748, |
| "learning_rate": 8.496208505021572e-06, |
| "loss": 0.6334, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.9873417721518988, |
| "grad_norm": 0.5748899254330844, |
| "learning_rate": 8.484022087566284e-06, |
| "loss": 0.6213, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.9902629016553067, |
| "grad_norm": 0.5910067428214469, |
| "learning_rate": 8.471795307725713e-06, |
| "loss": 0.6313, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.9931840311587147, |
| "grad_norm": 0.6409845798522262, |
| "learning_rate": 8.459528307147066e-06, |
| "loss": 0.6223, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.9961051606621227, |
| "grad_norm": 0.5798933129205326, |
| "learning_rate": 8.447221227943507e-06, |
| "loss": 0.6375, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.9990262901655307, |
| "grad_norm": 0.5323356127721257, |
| "learning_rate": 8.434874212692513e-06, |
| "loss": 0.631, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.0019474196689386, |
| "grad_norm": 1.1893567012708939, |
| "learning_rate": 8.422487404434214e-06, |
| "loss": 1.0195, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.0048685491723466, |
| "grad_norm": 0.6455689118062985, |
| "learning_rate": 8.41006094666975e-06, |
| "loss": 0.5891, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.0077896786757545, |
| "grad_norm": 0.6583690984771934, |
| "learning_rate": 8.397594983359591e-06, |
| "loss": 0.5549, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.0107108081791627, |
| "grad_norm": 0.6011562108844719, |
| "learning_rate": 8.385089658921892e-06, |
| "loss": 0.4942, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.0136319376825706, |
| "grad_norm": 0.7194629629820957, |
| "learning_rate": 8.372545118230793e-06, |
| "loss": 0.5879, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.0165530671859786, |
| "grad_norm": 0.7106922347864785, |
| "learning_rate": 8.35996150661476e-06, |
| "loss": 0.6464, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.0194741966893865, |
| "grad_norm": 0.7004876649145284, |
| "learning_rate": 8.347338969854898e-06, |
| "loss": 0.5635, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.0223953261927945, |
| "grad_norm": 0.7176313477690597, |
| "learning_rate": 8.334677654183254e-06, |
| "loss": 0.6121, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.0253164556962024, |
| "grad_norm": 0.6767419736782746, |
| "learning_rate": 8.321977706281135e-06, |
| "loss": 0.5923, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.0282375851996106, |
| "grad_norm": 0.6606355795156919, |
| "learning_rate": 8.309239273277394e-06, |
| "loss": 0.5375, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.0311587147030186, |
| "grad_norm": 0.7777491012749531, |
| "learning_rate": 8.296462502746743e-06, |
| "loss": 0.5971, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.0340798442064265, |
| "grad_norm": 0.6239185790928177, |
| "learning_rate": 8.283647542708026e-06, |
| "loss": 0.6017, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.0370009737098345, |
| "grad_norm": 0.659621365452411, |
| "learning_rate": 8.27079454162252e-06, |
| "loss": 0.523, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.0399221032132424, |
| "grad_norm": 0.7314219237543246, |
| "learning_rate": 8.2579036483922e-06, |
| "loss": 0.5992, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.0428432327166504, |
| "grad_norm": 0.6317967581739343, |
| "learning_rate": 8.244975012358028e-06, |
| "loss": 0.58, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.0457643622200585, |
| "grad_norm": 0.709552964482946, |
| "learning_rate": 8.232008783298211e-06, |
| "loss": 0.5703, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.0486854917234665, |
| "grad_norm": 0.5856088171700015, |
| "learning_rate": 8.219005111426472e-06, |
| "loss": 0.5851, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.0516066212268744, |
| "grad_norm": 0.72335786152197, |
| "learning_rate": 8.205964147390313e-06, |
| "loss": 0.5762, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.0545277507302824, |
| "grad_norm": 0.6082315528170907, |
| "learning_rate": 8.19288604226926e-06, |
| "loss": 0.6045, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.0574488802336903, |
| "grad_norm": 0.6904542946382585, |
| "learning_rate": 8.179770947573124e-06, |
| "loss": 0.5649, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.0603700097370983, |
| "grad_norm": 0.5889683647522176, |
| "learning_rate": 8.166619015240236e-06, |
| "loss": 0.5681, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.0632911392405062, |
| "grad_norm": 0.5621253187560026, |
| "learning_rate": 8.15343039763569e-06, |
| "loss": 0.6013, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.0662122687439144, |
| "grad_norm": 0.5587583825073225, |
| "learning_rate": 8.140205247549583e-06, |
| "loss": 0.6026, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.0691333982473223, |
| "grad_norm": 0.5053935078299595, |
| "learning_rate": 8.126943718195239e-06, |
| "loss": 0.5693, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.0720545277507303, |
| "grad_norm": 0.5469581161571481, |
| "learning_rate": 8.113645963207432e-06, |
| "loss": 0.6007, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.0749756572541382, |
| "grad_norm": 0.5430360387939689, |
| "learning_rate": 8.100312136640618e-06, |
| "loss": 0.5754, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.0778967867575462, |
| "grad_norm": 0.529838189847217, |
| "learning_rate": 8.086942392967131e-06, |
| "loss": 0.5758, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.0808179162609541, |
| "grad_norm": 0.5537166527008645, |
| "learning_rate": 8.073536887075417e-06, |
| "loss": 0.5596, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.0837390457643623, |
| "grad_norm": 0.575683387457894, |
| "learning_rate": 8.060095774268217e-06, |
| "loss": 0.6149, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.0866601752677703, |
| "grad_norm": 0.5212823944156575, |
| "learning_rate": 8.046619210260785e-06, |
| "loss": 0.5376, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.0895813047711782, |
| "grad_norm": 0.6132577983715938, |
| "learning_rate": 8.03310735117907e-06, |
| "loss": 0.611, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.0925024342745862, |
| "grad_norm": 0.5741650450345563, |
| "learning_rate": 8.019560353557923e-06, |
| "loss": 0.5428, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.095423563777994, |
| "grad_norm": 0.6898994227426227, |
| "learning_rate": 8.005978374339264e-06, |
| "loss": 0.6122, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.098344693281402, |
| "grad_norm": 0.6195159909025971, |
| "learning_rate": 7.992361570870289e-06, |
| "loss": 0.6298, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.1012658227848102, |
| "grad_norm": 0.5592176863621418, |
| "learning_rate": 7.978710100901617e-06, |
| "loss": 0.527, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.1041869522882182, |
| "grad_norm": 0.6077274644298606, |
| "learning_rate": 7.965024122585491e-06, |
| "loss": 0.5733, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.1071080817916261, |
| "grad_norm": 0.5404386498094536, |
| "learning_rate": 7.951303794473926e-06, |
| "loss": 0.5786, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.110029211295034, |
| "grad_norm": 0.6285033579389189, |
| "learning_rate": 7.937549275516882e-06, |
| "loss": 0.5593, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.112950340798442, |
| "grad_norm": 0.6136636789874864, |
| "learning_rate": 7.92376072506042e-06, |
| "loss": 0.5887, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.11587147030185, |
| "grad_norm": 0.5563073653338333, |
| "learning_rate": 7.909938302844856e-06, |
| "loss": 0.637, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.1187925998052581, |
| "grad_norm": 0.4984935085580574, |
| "learning_rate": 7.896082169002903e-06, |
| "loss": 0.537, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.121713729308666, |
| "grad_norm": 0.5688731697204966, |
| "learning_rate": 7.882192484057837e-06, |
| "loss": 0.5977, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.124634858812074, |
| "grad_norm": 0.5409938929481367, |
| "learning_rate": 7.868269408921614e-06, |
| "loss": 0.5477, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.127555988315482, |
| "grad_norm": 0.6311630545441865, |
| "learning_rate": 7.854313104893014e-06, |
| "loss": 0.5595, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.13047711781889, |
| "grad_norm": 0.5517528259351719, |
| "learning_rate": 7.84032373365578e-06, |
| "loss": 0.5588, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.1333982473222979, |
| "grad_norm": 0.515690277716415, |
| "learning_rate": 7.826301457276733e-06, |
| "loss": 0.5767, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.136319376825706, |
| "grad_norm": 0.5872560884226455, |
| "learning_rate": 7.812246438203905e-06, |
| "loss": 0.5618, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.139240506329114, |
| "grad_norm": 0.5449550621943328, |
| "learning_rate": 7.798158839264645e-06, |
| "loss": 0.5373, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.142161635832522, |
| "grad_norm": 0.6192261787272578, |
| "learning_rate": 7.784038823663746e-06, |
| "loss": 0.6117, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.14508276533593, |
| "grad_norm": 0.5911797274948096, |
| "learning_rate": 7.769886554981549e-06, |
| "loss": 0.5832, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.1480038948393378, |
| "grad_norm": 0.5393860544929336, |
| "learning_rate": 7.755702197172036e-06, |
| "loss": 0.5511, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.1509250243427458, |
| "grad_norm": 0.5686675548722897, |
| "learning_rate": 7.741485914560958e-06, |
| "loss": 0.601, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.1538461538461537, |
| "grad_norm": 0.5965421717249094, |
| "learning_rate": 7.7272378718439e-06, |
| "loss": 0.6142, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.156767283349562, |
| "grad_norm": 0.5976378598496599, |
| "learning_rate": 7.712958234084395e-06, |
| "loss": 0.532, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.1596884128529699, |
| "grad_norm": 0.5866514447746561, |
| "learning_rate": 7.698647166712003e-06, |
| "loss": 0.6436, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.1626095423563778, |
| "grad_norm": 0.5577115650955226, |
| "learning_rate": 7.684304835520395e-06, |
| "loss": 0.5524, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.1655306718597858, |
| "grad_norm": 0.6320330036922427, |
| "learning_rate": 7.669931406665437e-06, |
| "loss": 0.631, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.1684518013631937, |
| "grad_norm": 0.5274101108563934, |
| "learning_rate": 7.655527046663254e-06, |
| "loss": 0.5369, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.1713729308666017, |
| "grad_norm": 0.5563468540404465, |
| "learning_rate": 7.641091922388316e-06, |
| "loss": 0.577, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.1742940603700098, |
| "grad_norm": 0.5043411412651241, |
| "learning_rate": 7.626626201071494e-06, |
| "loss": 0.5623, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.1772151898734178, |
| "grad_norm": 0.5658997847969963, |
| "learning_rate": 7.612130050298126e-06, |
| "loss": 0.5613, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.1801363193768257, |
| "grad_norm": 0.5617715925867288, |
| "learning_rate": 7.597603638006071e-06, |
| "loss": 0.5796, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.1830574488802337, |
| "grad_norm": 0.5730674124102592, |
| "learning_rate": 7.5830471324837765e-06, |
| "loss": 0.6102, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.1859785783836416, |
| "grad_norm": 0.5322280548113558, |
| "learning_rate": 7.56846070236831e-06, |
| "loss": 0.5392, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.1888997078870496, |
| "grad_norm": 0.5636446588596294, |
| "learning_rate": 7.55384451664342e-06, |
| "loss": 0.5805, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.1918208373904577, |
| "grad_norm": 0.5515192754192108, |
| "learning_rate": 7.539198744637577e-06, |
| "loss": 0.5647, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.1947419668938657, |
| "grad_norm": 0.649611846767762, |
| "learning_rate": 7.524523556022003e-06, |
| "loss": 0.5804, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.1976630963972736, |
| "grad_norm": 0.5400876783820088, |
| "learning_rate": 7.5098191208087144e-06, |
| "loss": 0.5295, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.2005842259006816, |
| "grad_norm": 0.727016701945529, |
| "learning_rate": 7.495085609348549e-06, |
| "loss": 0.6035, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.2035053554040895, |
| "grad_norm": 0.6474638029028308, |
| "learning_rate": 7.4803231923291905e-06, |
| "loss": 0.5905, |
| "step": 412 |
| }, |
| { |
| "epoch": 1.2064264849074975, |
| "grad_norm": 0.5322418094330414, |
| "learning_rate": 7.465532040773195e-06, |
| "loss": 0.5696, |
| "step": 413 |
| }, |
| { |
| "epoch": 1.2093476144109054, |
| "grad_norm": 0.6033615790880061, |
| "learning_rate": 7.45071232603601e-06, |
| "loss": 0.5495, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.2122687439143136, |
| "grad_norm": 0.5696185950396394, |
| "learning_rate": 7.4358642198039835e-06, |
| "loss": 0.5761, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.2151898734177216, |
| "grad_norm": 0.58236484440868, |
| "learning_rate": 7.420987894092383e-06, |
| "loss": 0.6225, |
| "step": 416 |
| }, |
| { |
| "epoch": 1.2181110029211295, |
| "grad_norm": 0.6023725223738031, |
| "learning_rate": 7.406083521243396e-06, |
| "loss": 0.5539, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.2210321324245375, |
| "grad_norm": 0.5964833664567649, |
| "learning_rate": 7.391151273924135e-06, |
| "loss": 0.5766, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.2239532619279454, |
| "grad_norm": 0.5579327351939204, |
| "learning_rate": 7.376191325124644e-06, |
| "loss": 0.6037, |
| "step": 419 |
| }, |
| { |
| "epoch": 1.2268743914313536, |
| "grad_norm": 0.662691054067833, |
| "learning_rate": 7.36120384815588e-06, |
| "loss": 0.5775, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.2297955209347615, |
| "grad_norm": 0.5457870053427097, |
| "learning_rate": 7.34618901664772e-06, |
| "loss": 0.6032, |
| "step": 421 |
| }, |
| { |
| "epoch": 1.2327166504381695, |
| "grad_norm": 0.528437147629629, |
| "learning_rate": 7.33114700454694e-06, |
| "loss": 0.5551, |
| "step": 422 |
| }, |
| { |
| "epoch": 1.2356377799415774, |
| "grad_norm": 0.6396983335282087, |
| "learning_rate": 7.316077986115206e-06, |
| "loss": 0.5546, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.2385589094449854, |
| "grad_norm": 0.6059608136412865, |
| "learning_rate": 7.300982135927051e-06, |
| "loss": 0.5889, |
| "step": 424 |
| }, |
| { |
| "epoch": 1.2414800389483933, |
| "grad_norm": 0.5395833042938699, |
| "learning_rate": 7.285859628867851e-06, |
| "loss": 0.5765, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.2444011684518013, |
| "grad_norm": 0.5938150264550871, |
| "learning_rate": 7.270710640131806e-06, |
| "loss": 0.6058, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.2473222979552094, |
| "grad_norm": 0.5435408340953944, |
| "learning_rate": 7.255535345219905e-06, |
| "loss": 0.5933, |
| "step": 427 |
| }, |
| { |
| "epoch": 1.2502434274586174, |
| "grad_norm": 0.5055763559518435, |
| "learning_rate": 7.240333919937893e-06, |
| "loss": 0.5989, |
| "step": 428 |
| }, |
| { |
| "epoch": 1.2531645569620253, |
| "grad_norm": 0.5862621574043387, |
| "learning_rate": 7.2251065403942355e-06, |
| "loss": 0.5888, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.2560856864654333, |
| "grad_norm": 0.5234746081250162, |
| "learning_rate": 7.209853382998077e-06, |
| "loss": 0.5537, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.2590068159688412, |
| "grad_norm": 0.5736429396541277, |
| "learning_rate": 7.1945746244572e-06, |
| "loss": 0.547, |
| "step": 431 |
| }, |
| { |
| "epoch": 1.2619279454722494, |
| "grad_norm": 0.5662950178895197, |
| "learning_rate": 7.179270441775976e-06, |
| "loss": 0.6502, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.2648490749756571, |
| "grad_norm": 0.6193942417064999, |
| "learning_rate": 7.163941012253317e-06, |
| "loss": 0.5182, |
| "step": 433 |
| }, |
| { |
| "epoch": 1.2677702044790653, |
| "grad_norm": 0.5954433533450075, |
| "learning_rate": 7.148586513480614e-06, |
| "loss": 0.5961, |
| "step": 434 |
| }, |
| { |
| "epoch": 1.2706913339824732, |
| "grad_norm": 0.5407062847357796, |
| "learning_rate": 7.133207123339689e-06, |
| "loss": 0.5475, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.2736124634858812, |
| "grad_norm": 0.6133121400132443, |
| "learning_rate": 7.117803020000733e-06, |
| "loss": 0.59, |
| "step": 436 |
| }, |
| { |
| "epoch": 1.2765335929892891, |
| "grad_norm": 0.5272070801107334, |
| "learning_rate": 7.102374381920233e-06, |
| "loss": 0.542, |
| "step": 437 |
| }, |
| { |
| "epoch": 1.279454722492697, |
| "grad_norm": 0.6014559091825684, |
| "learning_rate": 7.086921387838916e-06, |
| "loss": 0.6136, |
| "step": 438 |
| }, |
| { |
| "epoch": 1.2823758519961053, |
| "grad_norm": 0.4925204082847554, |
| "learning_rate": 7.071444216779669e-06, |
| "loss": 0.5707, |
| "step": 439 |
| }, |
| { |
| "epoch": 1.2852969814995132, |
| "grad_norm": 0.5297523991584766, |
| "learning_rate": 7.055943048045476e-06, |
| "loss": 0.5917, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.2882181110029212, |
| "grad_norm": 0.5742994575633306, |
| "learning_rate": 7.040418061217325e-06, |
| "loss": 0.6161, |
| "step": 441 |
| }, |
| { |
| "epoch": 1.2911392405063291, |
| "grad_norm": 0.5144128093922836, |
| "learning_rate": 7.024869436152144e-06, |
| "loss": 0.5481, |
| "step": 442 |
| }, |
| { |
| "epoch": 1.294060370009737, |
| "grad_norm": 0.5651127309438425, |
| "learning_rate": 7.009297352980706e-06, |
| "loss": 0.5789, |
| "step": 443 |
| }, |
| { |
| "epoch": 1.296981499513145, |
| "grad_norm": 0.46327530559493646, |
| "learning_rate": 6.99370199210555e-06, |
| "loss": 0.5435, |
| "step": 444 |
| }, |
| { |
| "epoch": 1.299902629016553, |
| "grad_norm": 0.5306445037454199, |
| "learning_rate": 6.978083534198878e-06, |
| "loss": 0.6135, |
| "step": 445 |
| }, |
| { |
| "epoch": 1.3028237585199611, |
| "grad_norm": 0.6551321587944974, |
| "learning_rate": 6.962442160200484e-06, |
| "loss": 0.6111, |
| "step": 446 |
| }, |
| { |
| "epoch": 1.305744888023369, |
| "grad_norm": 0.5887843505365383, |
| "learning_rate": 6.9467780513156335e-06, |
| "loss": 0.5428, |
| "step": 447 |
| }, |
| { |
| "epoch": 1.308666017526777, |
| "grad_norm": 0.5871163724329924, |
| "learning_rate": 6.931091389012983e-06, |
| "loss": 0.5595, |
| "step": 448 |
| }, |
| { |
| "epoch": 1.311587147030185, |
| "grad_norm": 0.6876137396335796, |
| "learning_rate": 6.915382355022465e-06, |
| "loss": 0.5776, |
| "step": 449 |
| }, |
| { |
| "epoch": 1.314508276533593, |
| "grad_norm": 0.5978349360901288, |
| "learning_rate": 6.899651131333194e-06, |
| "loss": 0.559, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.317429406037001, |
| "grad_norm": 0.5600538501243326, |
| "learning_rate": 6.8838979001913454e-06, |
| "loss": 0.5782, |
| "step": 451 |
| }, |
| { |
| "epoch": 1.3203505355404088, |
| "grad_norm": 0.6670604315541686, |
| "learning_rate": 6.868122844098057e-06, |
| "loss": 0.6669, |
| "step": 452 |
| }, |
| { |
| "epoch": 1.323271665043817, |
| "grad_norm": 0.6019460261148103, |
| "learning_rate": 6.852326145807302e-06, |
| "loss": 0.6006, |
| "step": 453 |
| }, |
| { |
| "epoch": 1.326192794547225, |
| "grad_norm": 0.4932938142869286, |
| "learning_rate": 6.836507988323785e-06, |
| "loss": 0.4971, |
| "step": 454 |
| }, |
| { |
| "epoch": 1.3291139240506329, |
| "grad_norm": 0.6389612562939274, |
| "learning_rate": 6.82066855490081e-06, |
| "loss": 0.5994, |
| "step": 455 |
| }, |
| { |
| "epoch": 1.3320350535540408, |
| "grad_norm": 0.569391901245458, |
| "learning_rate": 6.804808029038168e-06, |
| "loss": 0.5776, |
| "step": 456 |
| }, |
| { |
| "epoch": 1.3349561830574488, |
| "grad_norm": 0.5627619450291675, |
| "learning_rate": 6.788926594480001e-06, |
| "loss": 0.5894, |
| "step": 457 |
| }, |
| { |
| "epoch": 1.337877312560857, |
| "grad_norm": 0.5525138776130833, |
| "learning_rate": 6.773024435212678e-06, |
| "loss": 0.5507, |
| "step": 458 |
| }, |
| { |
| "epoch": 1.340798442064265, |
| "grad_norm": 0.6243616187017386, |
| "learning_rate": 6.75710173546267e-06, |
| "loss": 0.6052, |
| "step": 459 |
| }, |
| { |
| "epoch": 1.3437195715676729, |
| "grad_norm": 0.46296070627628405, |
| "learning_rate": 6.741158679694403e-06, |
| "loss": 0.5284, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.3466407010710808, |
| "grad_norm": 0.556033009188379, |
| "learning_rate": 6.7251954526081294e-06, |
| "loss": 0.6225, |
| "step": 461 |
| }, |
| { |
| "epoch": 1.3495618305744888, |
| "grad_norm": 0.5425579804205766, |
| "learning_rate": 6.709212239137785e-06, |
| "loss": 0.5919, |
| "step": 462 |
| }, |
| { |
| "epoch": 1.352482960077897, |
| "grad_norm": 0.5628452971401441, |
| "learning_rate": 6.693209224448853e-06, |
| "loss": 0.6162, |
| "step": 463 |
| }, |
| { |
| "epoch": 1.3554040895813046, |
| "grad_norm": 0.5283138648819659, |
| "learning_rate": 6.677186593936207e-06, |
| "loss": 0.5022, |
| "step": 464 |
| }, |
| { |
| "epoch": 1.3583252190847128, |
| "grad_norm": 0.5688480741205626, |
| "learning_rate": 6.661144533221974e-06, |
| "loss": 0.594, |
| "step": 465 |
| }, |
| { |
| "epoch": 1.3612463485881208, |
| "grad_norm": 0.6112831896983953, |
| "learning_rate": 6.645083228153377e-06, |
| "loss": 0.5803, |
| "step": 466 |
| }, |
| { |
| "epoch": 1.3641674780915287, |
| "grad_norm": 0.5385606514191121, |
| "learning_rate": 6.629002864800589e-06, |
| "loss": 0.5834, |
| "step": 467 |
| }, |
| { |
| "epoch": 1.3670886075949367, |
| "grad_norm": 0.5366278285884805, |
| "learning_rate": 6.612903629454568e-06, |
| "loss": 0.5343, |
| "step": 468 |
| }, |
| { |
| "epoch": 1.3700097370983446, |
| "grad_norm": 0.6713338764048196, |
| "learning_rate": 6.5967857086249065e-06, |
| "loss": 0.5573, |
| "step": 469 |
| }, |
| { |
| "epoch": 1.3729308666017528, |
| "grad_norm": 0.5538883029206723, |
| "learning_rate": 6.58064928903767e-06, |
| "loss": 0.6164, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.3758519961051607, |
| "grad_norm": 0.5005412280561692, |
| "learning_rate": 6.56449455763323e-06, |
| "loss": 0.5576, |
| "step": 471 |
| }, |
| { |
| "epoch": 1.3787731256085687, |
| "grad_norm": 0.5868063901075268, |
| "learning_rate": 6.548321701564099e-06, |
| "loss": 0.5933, |
| "step": 472 |
| }, |
| { |
| "epoch": 1.3816942551119766, |
| "grad_norm": 0.5499777044713934, |
| "learning_rate": 6.5321309081927665e-06, |
| "loss": 0.562, |
| "step": 473 |
| }, |
| { |
| "epoch": 1.3846153846153846, |
| "grad_norm": 0.5209457126764143, |
| "learning_rate": 6.515922365089524e-06, |
| "loss": 0.5793, |
| "step": 474 |
| }, |
| { |
| "epoch": 1.3875365141187925, |
| "grad_norm": 0.5045261657801648, |
| "learning_rate": 6.499696260030297e-06, |
| "loss": 0.594, |
| "step": 475 |
| }, |
| { |
| "epoch": 1.3904576436222005, |
| "grad_norm": 0.5243175331043878, |
| "learning_rate": 6.483452780994459e-06, |
| "loss": 0.534, |
| "step": 476 |
| }, |
| { |
| "epoch": 1.3933787731256086, |
| "grad_norm": 0.5169962119701911, |
| "learning_rate": 6.467192116162668e-06, |
| "loss": 0.5403, |
| "step": 477 |
| }, |
| { |
| "epoch": 1.3962999026290166, |
| "grad_norm": 0.5140189723701601, |
| "learning_rate": 6.450914453914674e-06, |
| "loss": 0.6058, |
| "step": 478 |
| }, |
| { |
| "epoch": 1.3992210321324245, |
| "grad_norm": 0.5330866177837568, |
| "learning_rate": 6.434619982827147e-06, |
| "loss": 0.5882, |
| "step": 479 |
| }, |
| { |
| "epoch": 1.4021421616358325, |
| "grad_norm": 0.5414814038608168, |
| "learning_rate": 6.418308891671484e-06, |
| "loss": 0.5954, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.4050632911392404, |
| "grad_norm": 0.5376659643418761, |
| "learning_rate": 6.401981369411626e-06, |
| "loss": 0.5633, |
| "step": 481 |
| }, |
| { |
| "epoch": 1.4079844206426486, |
| "grad_norm": 0.5438938648229659, |
| "learning_rate": 6.385637605201871e-06, |
| "loss": 0.5677, |
| "step": 482 |
| }, |
| { |
| "epoch": 1.4109055501460563, |
| "grad_norm": 0.533597042074437, |
| "learning_rate": 6.3692777883846746e-06, |
| "loss": 0.6217, |
| "step": 483 |
| }, |
| { |
| "epoch": 1.4138266796494645, |
| "grad_norm": 0.5530416412659575, |
| "learning_rate": 6.3529021084884655e-06, |
| "loss": 0.5135, |
| "step": 484 |
| }, |
| { |
| "epoch": 1.4167478091528725, |
| "grad_norm": 0.6426214488830264, |
| "learning_rate": 6.336510755225447e-06, |
| "loss": 0.6039, |
| "step": 485 |
| }, |
| { |
| "epoch": 1.4196689386562804, |
| "grad_norm": 0.5321542466408685, |
| "learning_rate": 6.320103918489395e-06, |
| "loss": 0.5615, |
| "step": 486 |
| }, |
| { |
| "epoch": 1.4225900681596884, |
| "grad_norm": 0.5807878117878842, |
| "learning_rate": 6.303681788353465e-06, |
| "loss": 0.6417, |
| "step": 487 |
| }, |
| { |
| "epoch": 1.4255111976630963, |
| "grad_norm": 0.6006151574662654, |
| "learning_rate": 6.287244555067984e-06, |
| "loss": 0.5794, |
| "step": 488 |
| }, |
| { |
| "epoch": 1.4284323271665045, |
| "grad_norm": 0.5126167984608686, |
| "learning_rate": 6.270792409058247e-06, |
| "loss": 0.5242, |
| "step": 489 |
| }, |
| { |
| "epoch": 1.4313534566699124, |
| "grad_norm": 0.509995152360523, |
| "learning_rate": 6.25432554092232e-06, |
| "loss": 0.5828, |
| "step": 490 |
| }, |
| { |
| "epoch": 1.4342745861733204, |
| "grad_norm": 0.5960222110178056, |
| "learning_rate": 6.237844141428817e-06, |
| "loss": 0.568, |
| "step": 491 |
| }, |
| { |
| "epoch": 1.4371957156767283, |
| "grad_norm": 0.5173153132658744, |
| "learning_rate": 6.221348401514703e-06, |
| "loss": 0.5796, |
| "step": 492 |
| }, |
| { |
| "epoch": 1.4401168451801363, |
| "grad_norm": 0.5671800858222442, |
| "learning_rate": 6.204838512283073e-06, |
| "loss": 0.5506, |
| "step": 493 |
| }, |
| { |
| "epoch": 1.4430379746835442, |
| "grad_norm": 0.6040430910549448, |
| "learning_rate": 6.188314665000944e-06, |
| "loss": 0.5444, |
| "step": 494 |
| }, |
| { |
| "epoch": 1.4459591041869522, |
| "grad_norm": 0.5391742690410167, |
| "learning_rate": 6.171777051097037e-06, |
| "loss": 0.5649, |
| "step": 495 |
| }, |
| { |
| "epoch": 1.4488802336903603, |
| "grad_norm": 0.597749185130725, |
| "learning_rate": 6.155225862159558e-06, |
| "loss": 0.6408, |
| "step": 496 |
| }, |
| { |
| "epoch": 1.4518013631937683, |
| "grad_norm": 0.5229846439398993, |
| "learning_rate": 6.138661289933981e-06, |
| "loss": 0.5494, |
| "step": 497 |
| }, |
| { |
| "epoch": 1.4547224926971762, |
| "grad_norm": 0.5920796120572965, |
| "learning_rate": 6.1220835263208256e-06, |
| "loss": 0.5992, |
| "step": 498 |
| }, |
| { |
| "epoch": 1.4576436222005842, |
| "grad_norm": 0.5775415962111137, |
| "learning_rate": 6.105492763373431e-06, |
| "loss": 0.5911, |
| "step": 499 |
| }, |
| { |
| "epoch": 1.4605647517039921, |
| "grad_norm": 0.5917654314551983, |
| "learning_rate": 6.088889193295738e-06, |
| "loss": 0.5643, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.4634858812074003, |
| "grad_norm": 0.5199838430286094, |
| "learning_rate": 6.072273008440052e-06, |
| "loss": 0.5472, |
| "step": 501 |
| }, |
| { |
| "epoch": 1.4664070107108083, |
| "grad_norm": 0.5522718232673819, |
| "learning_rate": 6.0556444013048265e-06, |
| "loss": 0.546, |
| "step": 502 |
| }, |
| { |
| "epoch": 1.4693281402142162, |
| "grad_norm": 0.525908351198303, |
| "learning_rate": 6.039003564532423e-06, |
| "loss": 0.6015, |
| "step": 503 |
| }, |
| { |
| "epoch": 1.4722492697176242, |
| "grad_norm": 0.5105926082337128, |
| "learning_rate": 6.0223506909068875e-06, |
| "loss": 0.5785, |
| "step": 504 |
| }, |
| { |
| "epoch": 1.475170399221032, |
| "grad_norm": 0.5832929477157227, |
| "learning_rate": 6.005685973351708e-06, |
| "loss": 0.6027, |
| "step": 505 |
| }, |
| { |
| "epoch": 1.47809152872444, |
| "grad_norm": 0.5447366515820563, |
| "learning_rate": 5.989009604927587e-06, |
| "loss": 0.5833, |
| "step": 506 |
| }, |
| { |
| "epoch": 1.481012658227848, |
| "grad_norm": 0.5245433114840506, |
| "learning_rate": 5.972321778830202e-06, |
| "loss": 0.4944, |
| "step": 507 |
| }, |
| { |
| "epoch": 1.4839337877312562, |
| "grad_norm": 0.5377622062276557, |
| "learning_rate": 5.9556226883879685e-06, |
| "loss": 0.5518, |
| "step": 508 |
| }, |
| { |
| "epoch": 1.4868549172346641, |
| "grad_norm": 0.6351508971275112, |
| "learning_rate": 5.938912527059798e-06, |
| "loss": 0.5982, |
| "step": 509 |
| }, |
| { |
| "epoch": 1.489776046738072, |
| "grad_norm": 0.5264532680548314, |
| "learning_rate": 5.922191488432857e-06, |
| "loss": 0.5472, |
| "step": 510 |
| }, |
| { |
| "epoch": 1.49269717624148, |
| "grad_norm": 0.5618733075790842, |
| "learning_rate": 5.90545976622033e-06, |
| "loss": 0.6387, |
| "step": 511 |
| }, |
| { |
| "epoch": 1.495618305744888, |
| "grad_norm": 0.5042404211846953, |
| "learning_rate": 5.888717554259165e-06, |
| "loss": 0.5188, |
| "step": 512 |
| }, |
| { |
| "epoch": 1.4985394352482961, |
| "grad_norm": 0.5996827758246558, |
| "learning_rate": 5.871965046507838e-06, |
| "loss": 0.5845, |
| "step": 513 |
| }, |
| { |
| "epoch": 1.5014605647517039, |
| "grad_norm": 0.513160385214818, |
| "learning_rate": 5.855202437044102e-06, |
| "loss": 0.5407, |
| "step": 514 |
| }, |
| { |
| "epoch": 1.504381694255112, |
| "grad_norm": 0.49930093522201097, |
| "learning_rate": 5.838429920062734e-06, |
| "loss": 0.5888, |
| "step": 515 |
| }, |
| { |
| "epoch": 1.50730282375852, |
| "grad_norm": 0.5820994183473087, |
| "learning_rate": 5.8216476898732935e-06, |
| "loss": 0.6026, |
| "step": 516 |
| }, |
| { |
| "epoch": 1.510223953261928, |
| "grad_norm": 0.5965529197255538, |
| "learning_rate": 5.804855940897866e-06, |
| "loss": 0.5772, |
| "step": 517 |
| }, |
| { |
| "epoch": 1.5131450827653359, |
| "grad_norm": 0.4891063083201174, |
| "learning_rate": 5.788054867668811e-06, |
| "loss": 0.5383, |
| "step": 518 |
| }, |
| { |
| "epoch": 1.5160662122687438, |
| "grad_norm": 0.549598064227309, |
| "learning_rate": 5.771244664826512e-06, |
| "loss": 0.5701, |
| "step": 519 |
| }, |
| { |
| "epoch": 1.518987341772152, |
| "grad_norm": 0.5201662235289578, |
| "learning_rate": 5.754425527117118e-06, |
| "loss": 0.5221, |
| "step": 520 |
| }, |
| { |
| "epoch": 1.5219084712755597, |
| "grad_norm": 0.5727695969794172, |
| "learning_rate": 5.737597649390288e-06, |
| "loss": 0.5849, |
| "step": 521 |
| }, |
| { |
| "epoch": 1.524829600778968, |
| "grad_norm": 0.5387267148344115, |
| "learning_rate": 5.720761226596935e-06, |
| "loss": 0.5899, |
| "step": 522 |
| }, |
| { |
| "epoch": 1.5277507302823758, |
| "grad_norm": 0.5196417961644373, |
| "learning_rate": 5.703916453786965e-06, |
| "loss": 0.6075, |
| "step": 523 |
| }, |
| { |
| "epoch": 1.5306718597857838, |
| "grad_norm": 0.5162526391071218, |
| "learning_rate": 5.6870635261070186e-06, |
| "loss": 0.5524, |
| "step": 524 |
| }, |
| { |
| "epoch": 1.533592989289192, |
| "grad_norm": 0.5712502081281199, |
| "learning_rate": 5.670202638798213e-06, |
| "loss": 0.5639, |
| "step": 525 |
| }, |
| { |
| "epoch": 1.5365141187925997, |
| "grad_norm": 0.569067876773126, |
| "learning_rate": 5.653333987193876e-06, |
| "loss": 0.5437, |
| "step": 526 |
| }, |
| { |
| "epoch": 1.5394352482960079, |
| "grad_norm": 0.5925750045532109, |
| "learning_rate": 5.636457766717286e-06, |
| "loss": 0.591, |
| "step": 527 |
| }, |
| { |
| "epoch": 1.5423563777994158, |
| "grad_norm": 0.5141531017914309, |
| "learning_rate": 5.619574172879405e-06, |
| "loss": 0.5471, |
| "step": 528 |
| }, |
| { |
| "epoch": 1.5452775073028238, |
| "grad_norm": 0.6597290642553237, |
| "learning_rate": 5.6026834012766155e-06, |
| "loss": 0.5901, |
| "step": 529 |
| }, |
| { |
| "epoch": 1.5481986368062317, |
| "grad_norm": 0.5884715259486929, |
| "learning_rate": 5.585785647588458e-06, |
| "loss": 0.5878, |
| "step": 530 |
| }, |
| { |
| "epoch": 1.5511197663096397, |
| "grad_norm": 0.5255721650946908, |
| "learning_rate": 5.568881107575353e-06, |
| "loss": 0.5447, |
| "step": 531 |
| }, |
| { |
| "epoch": 1.5540408958130478, |
| "grad_norm": 0.5091097959641512, |
| "learning_rate": 5.55196997707635e-06, |
| "loss": 0.6276, |
| "step": 532 |
| }, |
| { |
| "epoch": 1.5569620253164556, |
| "grad_norm": 0.5119232421542075, |
| "learning_rate": 5.53505245200684e-06, |
| "loss": 0.5658, |
| "step": 533 |
| }, |
| { |
| "epoch": 1.5598831548198637, |
| "grad_norm": 0.556141615427262, |
| "learning_rate": 5.518128728356303e-06, |
| "loss": 0.5951, |
| "step": 534 |
| }, |
| { |
| "epoch": 1.5628042843232717, |
| "grad_norm": 0.4969794250430925, |
| "learning_rate": 5.501199002186024e-06, |
| "loss": 0.5457, |
| "step": 535 |
| }, |
| { |
| "epoch": 1.5657254138266796, |
| "grad_norm": 0.5377430121732353, |
| "learning_rate": 5.48426346962683e-06, |
| "loss": 0.587, |
| "step": 536 |
| }, |
| { |
| "epoch": 1.5686465433300878, |
| "grad_norm": 0.4763376786579273, |
| "learning_rate": 5.467322326876813e-06, |
| "loss": 0.5352, |
| "step": 537 |
| }, |
| { |
| "epoch": 1.5715676728334955, |
| "grad_norm": 0.5630031793193719, |
| "learning_rate": 5.450375770199063e-06, |
| "loss": 0.652, |
| "step": 538 |
| }, |
| { |
| "epoch": 1.5744888023369037, |
| "grad_norm": 0.505233569384271, |
| "learning_rate": 5.433423995919383e-06, |
| "loss": 0.5633, |
| "step": 539 |
| }, |
| { |
| "epoch": 1.5774099318403116, |
| "grad_norm": 0.4918114442897718, |
| "learning_rate": 5.416467200424032e-06, |
| "loss": 0.5422, |
| "step": 540 |
| }, |
| { |
| "epoch": 1.5803310613437196, |
| "grad_norm": 0.5991334895146722, |
| "learning_rate": 5.399505580157428e-06, |
| "loss": 0.558, |
| "step": 541 |
| }, |
| { |
| "epoch": 1.5832521908471275, |
| "grad_norm": 0.5363474139996335, |
| "learning_rate": 5.382539331619896e-06, |
| "loss": 0.5699, |
| "step": 542 |
| }, |
| { |
| "epoch": 1.5861733203505355, |
| "grad_norm": 0.5807115654118472, |
| "learning_rate": 5.365568651365369e-06, |
| "loss": 0.5688, |
| "step": 543 |
| }, |
| { |
| "epoch": 1.5890944498539437, |
| "grad_norm": 0.5904366045058206, |
| "learning_rate": 5.34859373599913e-06, |
| "loss": 0.5969, |
| "step": 544 |
| }, |
| { |
| "epoch": 1.5920155793573514, |
| "grad_norm": 0.491932507277215, |
| "learning_rate": 5.33161478217552e-06, |
| "loss": 0.5542, |
| "step": 545 |
| }, |
| { |
| "epoch": 1.5949367088607596, |
| "grad_norm": 0.5011301426277018, |
| "learning_rate": 5.314631986595669e-06, |
| "loss": 0.569, |
| "step": 546 |
| }, |
| { |
| "epoch": 1.5978578383641675, |
| "grad_norm": 0.6106059864005979, |
| "learning_rate": 5.297645546005208e-06, |
| "loss": 0.6051, |
| "step": 547 |
| }, |
| { |
| "epoch": 1.6007789678675755, |
| "grad_norm": 0.5958177689979132, |
| "learning_rate": 5.280655657192003e-06, |
| "loss": 0.5266, |
| "step": 548 |
| }, |
| { |
| "epoch": 1.6037000973709834, |
| "grad_norm": 0.5395470595824067, |
| "learning_rate": 5.263662516983863e-06, |
| "loss": 0.5657, |
| "step": 549 |
| }, |
| { |
| "epoch": 1.6066212268743914, |
| "grad_norm": 0.6212970783782318, |
| "learning_rate": 5.246666322246267e-06, |
| "loss": 0.5619, |
| "step": 550 |
| }, |
| { |
| "epoch": 1.6095423563777995, |
| "grad_norm": 0.6613037032143712, |
| "learning_rate": 5.229667269880078e-06, |
| "loss": 0.6092, |
| "step": 551 |
| }, |
| { |
| "epoch": 1.6124634858812072, |
| "grad_norm": 0.48332526807644743, |
| "learning_rate": 5.212665556819264e-06, |
| "loss": 0.5401, |
| "step": 552 |
| }, |
| { |
| "epoch": 1.6153846153846154, |
| "grad_norm": 0.5325857980417065, |
| "learning_rate": 5.195661380028625e-06, |
| "loss": 0.6035, |
| "step": 553 |
| }, |
| { |
| "epoch": 1.6183057448880234, |
| "grad_norm": 0.5442785714221408, |
| "learning_rate": 5.178654936501492e-06, |
| "loss": 0.5494, |
| "step": 554 |
| }, |
| { |
| "epoch": 1.6212268743914313, |
| "grad_norm": 0.6437807568653632, |
| "learning_rate": 5.1616464232574635e-06, |
| "loss": 0.6254, |
| "step": 555 |
| }, |
| { |
| "epoch": 1.6241480038948395, |
| "grad_norm": 0.4492125406814553, |
| "learning_rate": 5.1446360373401125e-06, |
| "loss": 0.5301, |
| "step": 556 |
| }, |
| { |
| "epoch": 1.6270691333982472, |
| "grad_norm": 0.5622471364004906, |
| "learning_rate": 5.127623975814709e-06, |
| "loss": 0.5715, |
| "step": 557 |
| }, |
| { |
| "epoch": 1.6299902629016554, |
| "grad_norm": 0.6240851711756873, |
| "learning_rate": 5.110610435765935e-06, |
| "loss": 0.5876, |
| "step": 558 |
| }, |
| { |
| "epoch": 1.6329113924050633, |
| "grad_norm": 0.4713533254066173, |
| "learning_rate": 5.093595614295599e-06, |
| "loss": 0.4901, |
| "step": 559 |
| }, |
| { |
| "epoch": 1.6358325219084713, |
| "grad_norm": 0.5796358881475803, |
| "learning_rate": 5.076579708520355e-06, |
| "loss": 0.6325, |
| "step": 560 |
| }, |
| { |
| "epoch": 1.6387536514118792, |
| "grad_norm": 0.4866837205814519, |
| "learning_rate": 5.059562915569424e-06, |
| "loss": 0.5202, |
| "step": 561 |
| }, |
| { |
| "epoch": 1.6416747809152872, |
| "grad_norm": 0.5697352103252339, |
| "learning_rate": 5.0425454325822946e-06, |
| "loss": 0.6339, |
| "step": 562 |
| }, |
| { |
| "epoch": 1.6445959104186954, |
| "grad_norm": 0.5721771538016244, |
| "learning_rate": 5.0255274567064594e-06, |
| "loss": 0.5364, |
| "step": 563 |
| }, |
| { |
| "epoch": 1.647517039922103, |
| "grad_norm": 0.5605730886016151, |
| "learning_rate": 5.008509185095114e-06, |
| "loss": 0.5717, |
| "step": 564 |
| }, |
| { |
| "epoch": 1.6504381694255112, |
| "grad_norm": 0.472681806190808, |
| "learning_rate": 4.991490814904888e-06, |
| "loss": 0.555, |
| "step": 565 |
| }, |
| { |
| "epoch": 1.6533592989289192, |
| "grad_norm": 0.5791761573446532, |
| "learning_rate": 4.974472543293544e-06, |
| "loss": 0.6077, |
| "step": 566 |
| }, |
| { |
| "epoch": 1.6562804284323271, |
| "grad_norm": 0.5167447517838185, |
| "learning_rate": 4.957454567417708e-06, |
| "loss": 0.5805, |
| "step": 567 |
| }, |
| { |
| "epoch": 1.6592015579357353, |
| "grad_norm": 0.5561818381815233, |
| "learning_rate": 4.940437084430579e-06, |
| "loss": 0.594, |
| "step": 568 |
| }, |
| { |
| "epoch": 1.662122687439143, |
| "grad_norm": 0.44095968784824047, |
| "learning_rate": 4.923420291479646e-06, |
| "loss": 0.5235, |
| "step": 569 |
| }, |
| { |
| "epoch": 1.6650438169425512, |
| "grad_norm": 0.49918372558187907, |
| "learning_rate": 4.906404385704402e-06, |
| "loss": 0.6011, |
| "step": 570 |
| }, |
| { |
| "epoch": 1.667964946445959, |
| "grad_norm": 0.5067863457643954, |
| "learning_rate": 4.8893895642340665e-06, |
| "loss": 0.5415, |
| "step": 571 |
| }, |
| { |
| "epoch": 1.6708860759493671, |
| "grad_norm": 0.49737844087035643, |
| "learning_rate": 4.872376024185291e-06, |
| "loss": 0.583, |
| "step": 572 |
| }, |
| { |
| "epoch": 1.673807205452775, |
| "grad_norm": 0.4659962080708305, |
| "learning_rate": 4.855363962659889e-06, |
| "loss": 0.5627, |
| "step": 573 |
| }, |
| { |
| "epoch": 1.676728334956183, |
| "grad_norm": 0.4972582768404136, |
| "learning_rate": 4.838353576742538e-06, |
| "loss": 0.5911, |
| "step": 574 |
| }, |
| { |
| "epoch": 1.6796494644595912, |
| "grad_norm": 0.44592731268572705, |
| "learning_rate": 4.82134506349851e-06, |
| "loss": 0.5459, |
| "step": 575 |
| }, |
| { |
| "epoch": 1.682570593962999, |
| "grad_norm": 0.506185991053395, |
| "learning_rate": 4.804338619971377e-06, |
| "loss": 0.5306, |
| "step": 576 |
| }, |
| { |
| "epoch": 1.685491723466407, |
| "grad_norm": 0.5177923049058465, |
| "learning_rate": 4.787334443180735e-06, |
| "loss": 0.6209, |
| "step": 577 |
| }, |
| { |
| "epoch": 1.688412852969815, |
| "grad_norm": 0.4601309987230691, |
| "learning_rate": 4.7703327301199244e-06, |
| "loss": 0.5145, |
| "step": 578 |
| }, |
| { |
| "epoch": 1.691333982473223, |
| "grad_norm": 0.4643978865329562, |
| "learning_rate": 4.753333677753734e-06, |
| "loss": 0.5788, |
| "step": 579 |
| }, |
| { |
| "epoch": 1.694255111976631, |
| "grad_norm": 0.5283219640600681, |
| "learning_rate": 4.736337483016138e-06, |
| "loss": 0.6056, |
| "step": 580 |
| }, |
| { |
| "epoch": 1.6971762414800389, |
| "grad_norm": 0.5571423915108132, |
| "learning_rate": 4.719344342808e-06, |
| "loss": 0.5791, |
| "step": 581 |
| }, |
| { |
| "epoch": 1.700097370983447, |
| "grad_norm": 0.49563030247464135, |
| "learning_rate": 4.702354453994794e-06, |
| "loss": 0.5373, |
| "step": 582 |
| }, |
| { |
| "epoch": 1.7030185004868548, |
| "grad_norm": 0.49902926183967655, |
| "learning_rate": 4.6853680134043345e-06, |
| "loss": 0.5823, |
| "step": 583 |
| }, |
| { |
| "epoch": 1.705939629990263, |
| "grad_norm": 0.5116289187531037, |
| "learning_rate": 4.668385217824482e-06, |
| "loss": 0.5557, |
| "step": 584 |
| }, |
| { |
| "epoch": 1.7088607594936709, |
| "grad_norm": 0.5412976872055105, |
| "learning_rate": 4.651406264000871e-06, |
| "loss": 0.5535, |
| "step": 585 |
| }, |
| { |
| "epoch": 1.7117818889970788, |
| "grad_norm": 0.5297049362432723, |
| "learning_rate": 4.634431348634632e-06, |
| "loss": 0.5872, |
| "step": 586 |
| }, |
| { |
| "epoch": 1.714703018500487, |
| "grad_norm": 0.46815419228452076, |
| "learning_rate": 4.6174606683801055e-06, |
| "loss": 0.5612, |
| "step": 587 |
| }, |
| { |
| "epoch": 1.7176241480038947, |
| "grad_norm": 0.5047028920506614, |
| "learning_rate": 4.600494419842573e-06, |
| "loss": 0.5457, |
| "step": 588 |
| }, |
| { |
| "epoch": 1.720545277507303, |
| "grad_norm": 0.5021848097551853, |
| "learning_rate": 4.58353279957597e-06, |
| "loss": 0.5512, |
| "step": 589 |
| }, |
| { |
| "epoch": 1.7234664070107109, |
| "grad_norm": 0.5567736723329815, |
| "learning_rate": 4.5665760040806174e-06, |
| "loss": 0.577, |
| "step": 590 |
| }, |
| { |
| "epoch": 1.7263875365141188, |
| "grad_norm": 0.4982671338092423, |
| "learning_rate": 4.549624229800938e-06, |
| "loss": 0.548, |
| "step": 591 |
| }, |
| { |
| "epoch": 1.7293086660175268, |
| "grad_norm": 0.49329094074335034, |
| "learning_rate": 4.532677673123188e-06, |
| "loss": 0.5563, |
| "step": 592 |
| }, |
| { |
| "epoch": 1.7322297955209347, |
| "grad_norm": 0.4890277332079673, |
| "learning_rate": 4.5157365303731705e-06, |
| "loss": 0.5747, |
| "step": 593 |
| }, |
| { |
| "epoch": 1.7351509250243429, |
| "grad_norm": 0.5600537499346627, |
| "learning_rate": 4.498800997813976e-06, |
| "loss": 0.5816, |
| "step": 594 |
| }, |
| { |
| "epoch": 1.7380720545277506, |
| "grad_norm": 0.5086959673884651, |
| "learning_rate": 4.481871271643698e-06, |
| "loss": 0.5671, |
| "step": 595 |
| }, |
| { |
| "epoch": 1.7409931840311588, |
| "grad_norm": 0.5157431177024923, |
| "learning_rate": 4.464947547993162e-06, |
| "loss": 0.5731, |
| "step": 596 |
| }, |
| { |
| "epoch": 1.7439143135345667, |
| "grad_norm": 0.5137961841228513, |
| "learning_rate": 4.4480300229236525e-06, |
| "loss": 0.544, |
| "step": 597 |
| }, |
| { |
| "epoch": 1.7468354430379747, |
| "grad_norm": 0.5937636838238008, |
| "learning_rate": 4.431118892424649e-06, |
| "loss": 0.6082, |
| "step": 598 |
| }, |
| { |
| "epoch": 1.7497565725413826, |
| "grad_norm": 0.5529433304939138, |
| "learning_rate": 4.414214352411544e-06, |
| "loss": 0.5569, |
| "step": 599 |
| }, |
| { |
| "epoch": 1.7526777020447906, |
| "grad_norm": 0.5159596510951358, |
| "learning_rate": 4.397316598723385e-06, |
| "loss": 0.5455, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.7555988315481987, |
| "grad_norm": 0.5506758519748508, |
| "learning_rate": 4.3804258271205965e-06, |
| "loss": 0.5505, |
| "step": 601 |
| }, |
| { |
| "epoch": 1.7585199610516065, |
| "grad_norm": 0.5523849090934733, |
| "learning_rate": 4.363542233282715e-06, |
| "loss": 0.5355, |
| "step": 602 |
| }, |
| { |
| "epoch": 1.7614410905550146, |
| "grad_norm": 0.6508686455735765, |
| "learning_rate": 4.346666012806126e-06, |
| "loss": 0.602, |
| "step": 603 |
| }, |
| { |
| "epoch": 1.7643622200584226, |
| "grad_norm": 0.46084404235460724, |
| "learning_rate": 4.329797361201788e-06, |
| "loss": 0.5028, |
| "step": 604 |
| }, |
| { |
| "epoch": 1.7672833495618305, |
| "grad_norm": 0.5857372990733761, |
| "learning_rate": 4.312936473892984e-06, |
| "loss": 0.6468, |
| "step": 605 |
| }, |
| { |
| "epoch": 1.7702044790652387, |
| "grad_norm": 0.4935847914878853, |
| "learning_rate": 4.296083546213037e-06, |
| "loss": 0.5578, |
| "step": 606 |
| }, |
| { |
| "epoch": 1.7731256085686464, |
| "grad_norm": 0.5144278362398599, |
| "learning_rate": 4.279238773403066e-06, |
| "loss": 0.5337, |
| "step": 607 |
| }, |
| { |
| "epoch": 1.7760467380720546, |
| "grad_norm": 0.46428380216043086, |
| "learning_rate": 4.2624023506097116e-06, |
| "loss": 0.6219, |
| "step": 608 |
| }, |
| { |
| "epoch": 1.7789678675754625, |
| "grad_norm": 0.4492920362378749, |
| "learning_rate": 4.245574472882882e-06, |
| "loss": 0.5061, |
| "step": 609 |
| }, |
| { |
| "epoch": 1.7818889970788705, |
| "grad_norm": 0.48814416218826545, |
| "learning_rate": 4.228755335173488e-06, |
| "loss": 0.5045, |
| "step": 610 |
| }, |
| { |
| "epoch": 1.7848101265822784, |
| "grad_norm": 0.4935317746542341, |
| "learning_rate": 4.21194513233119e-06, |
| "loss": 0.6334, |
| "step": 611 |
| }, |
| { |
| "epoch": 1.7877312560856864, |
| "grad_norm": 0.5119284354029839, |
| "learning_rate": 4.1951440591021375e-06, |
| "loss": 0.5599, |
| "step": 612 |
| }, |
| { |
| "epoch": 1.7906523855890946, |
| "grad_norm": 0.4827371863106343, |
| "learning_rate": 4.17835231012671e-06, |
| "loss": 0.5692, |
| "step": 613 |
| }, |
| { |
| "epoch": 1.7935735150925023, |
| "grad_norm": 0.5212508350905807, |
| "learning_rate": 4.161570079937268e-06, |
| "loss": 0.5947, |
| "step": 614 |
| }, |
| { |
| "epoch": 1.7964946445959105, |
| "grad_norm": 0.5235515777898554, |
| "learning_rate": 4.1447975629559e-06, |
| "loss": 0.5808, |
| "step": 615 |
| }, |
| { |
| "epoch": 1.7994157740993184, |
| "grad_norm": 0.5336005328533747, |
| "learning_rate": 4.128034953492163e-06, |
| "loss": 0.5358, |
| "step": 616 |
| }, |
| { |
| "epoch": 1.8023369036027264, |
| "grad_norm": 0.5597776990519019, |
| "learning_rate": 4.1112824457408375e-06, |
| "loss": 0.5884, |
| "step": 617 |
| }, |
| { |
| "epoch": 1.8052580331061345, |
| "grad_norm": 0.5036193557727575, |
| "learning_rate": 4.094540233779672e-06, |
| "loss": 0.5688, |
| "step": 618 |
| }, |
| { |
| "epoch": 1.8081791626095423, |
| "grad_norm": 0.5338396120419266, |
| "learning_rate": 4.077808511567145e-06, |
| "loss": 0.5608, |
| "step": 619 |
| }, |
| { |
| "epoch": 1.8111002921129504, |
| "grad_norm": 0.4958715717085618, |
| "learning_rate": 4.061087472940204e-06, |
| "loss": 0.5543, |
| "step": 620 |
| }, |
| { |
| "epoch": 1.8140214216163584, |
| "grad_norm": 0.6029334087991165, |
| "learning_rate": 4.044377311612033e-06, |
| "loss": 0.5691, |
| "step": 621 |
| }, |
| { |
| "epoch": 1.8169425511197663, |
| "grad_norm": 0.47915777984754265, |
| "learning_rate": 4.0276782211698e-06, |
| "loss": 0.5391, |
| "step": 622 |
| }, |
| { |
| "epoch": 1.8198636806231743, |
| "grad_norm": 0.5235008297199194, |
| "learning_rate": 4.010990395072414e-06, |
| "loss": 0.5736, |
| "step": 623 |
| }, |
| { |
| "epoch": 1.8227848101265822, |
| "grad_norm": 0.5590747505915074, |
| "learning_rate": 3.9943140266482935e-06, |
| "loss": 0.6075, |
| "step": 624 |
| }, |
| { |
| "epoch": 1.8257059396299904, |
| "grad_norm": 0.47598042280541497, |
| "learning_rate": 3.977649309093113e-06, |
| "loss": 0.5344, |
| "step": 625 |
| }, |
| { |
| "epoch": 1.8286270691333981, |
| "grad_norm": 0.48621841809715133, |
| "learning_rate": 3.960996435467577e-06, |
| "loss": 0.6179, |
| "step": 626 |
| }, |
| { |
| "epoch": 1.8315481986368063, |
| "grad_norm": 0.5017298117872958, |
| "learning_rate": 3.944355598695174e-06, |
| "loss": 0.5443, |
| "step": 627 |
| }, |
| { |
| "epoch": 1.8344693281402142, |
| "grad_norm": 0.5380565465468942, |
| "learning_rate": 3.9277269915599505e-06, |
| "loss": 0.57, |
| "step": 628 |
| }, |
| { |
| "epoch": 1.8373904576436222, |
| "grad_norm": 0.5990209575217125, |
| "learning_rate": 3.911110806704265e-06, |
| "loss": 0.5784, |
| "step": 629 |
| }, |
| { |
| "epoch": 1.8403115871470301, |
| "grad_norm": 0.49509817541993284, |
| "learning_rate": 3.89450723662657e-06, |
| "loss": 0.5628, |
| "step": 630 |
| }, |
| { |
| "epoch": 1.843232716650438, |
| "grad_norm": 0.550605497861562, |
| "learning_rate": 3.877916473679176e-06, |
| "loss": 0.6147, |
| "step": 631 |
| }, |
| { |
| "epoch": 1.8461538461538463, |
| "grad_norm": 0.5524680706044535, |
| "learning_rate": 3.861338710066021e-06, |
| "loss": 0.5125, |
| "step": 632 |
| }, |
| { |
| "epoch": 1.849074975657254, |
| "grad_norm": 0.5494313577644265, |
| "learning_rate": 3.8447741378404436e-06, |
| "loss": 0.5562, |
| "step": 633 |
| }, |
| { |
| "epoch": 1.8519961051606622, |
| "grad_norm": 0.5058622045652387, |
| "learning_rate": 3.828222948902965e-06, |
| "loss": 0.5967, |
| "step": 634 |
| }, |
| { |
| "epoch": 1.85491723466407, |
| "grad_norm": 0.4488395706127018, |
| "learning_rate": 3.8116853349990574e-06, |
| "loss": 0.4735, |
| "step": 635 |
| }, |
| { |
| "epoch": 1.857838364167478, |
| "grad_norm": 0.5073104983823334, |
| "learning_rate": 3.7951614877169285e-06, |
| "loss": 0.6214, |
| "step": 636 |
| }, |
| { |
| "epoch": 1.8607594936708862, |
| "grad_norm": 0.5020623526859631, |
| "learning_rate": 3.7786515984852977e-06, |
| "loss": 0.5801, |
| "step": 637 |
| }, |
| { |
| "epoch": 1.863680623174294, |
| "grad_norm": 0.5066558405528695, |
| "learning_rate": 3.762155858571184e-06, |
| "loss": 0.5571, |
| "step": 638 |
| }, |
| { |
| "epoch": 1.8666017526777021, |
| "grad_norm": 0.5016476079074594, |
| "learning_rate": 3.7456744590776807e-06, |
| "loss": 0.554, |
| "step": 639 |
| }, |
| { |
| "epoch": 1.86952288218111, |
| "grad_norm": 0.5522667996432958, |
| "learning_rate": 3.729207590941753e-06, |
| "loss": 0.5684, |
| "step": 640 |
| }, |
| { |
| "epoch": 1.872444011684518, |
| "grad_norm": 0.5063071640192653, |
| "learning_rate": 3.712755444932018e-06, |
| "loss": 0.5536, |
| "step": 641 |
| }, |
| { |
| "epoch": 1.875365141187926, |
| "grad_norm": 0.47979895935138983, |
| "learning_rate": 3.6963182116465358e-06, |
| "loss": 0.5607, |
| "step": 642 |
| }, |
| { |
| "epoch": 1.878286270691334, |
| "grad_norm": 0.5466280214439997, |
| "learning_rate": 3.6798960815106065e-06, |
| "loss": 0.5912, |
| "step": 643 |
| }, |
| { |
| "epoch": 1.881207400194742, |
| "grad_norm": 0.50137359619968, |
| "learning_rate": 3.663489244774555e-06, |
| "loss": 0.6032, |
| "step": 644 |
| }, |
| { |
| "epoch": 1.8841285296981498, |
| "grad_norm": 0.5176702766961846, |
| "learning_rate": 3.647097891511536e-06, |
| "loss": 0.5441, |
| "step": 645 |
| }, |
| { |
| "epoch": 1.887049659201558, |
| "grad_norm": 0.45011045640480307, |
| "learning_rate": 3.630722211615328e-06, |
| "loss": 0.5064, |
| "step": 646 |
| }, |
| { |
| "epoch": 1.889970788704966, |
| "grad_norm": 0.5394614468332969, |
| "learning_rate": 3.614362394798131e-06, |
| "loss": 0.6368, |
| "step": 647 |
| }, |
| { |
| "epoch": 1.8928919182083739, |
| "grad_norm": 0.4874425981153431, |
| "learning_rate": 3.5980186305883746e-06, |
| "loss": 0.5701, |
| "step": 648 |
| }, |
| { |
| "epoch": 1.895813047711782, |
| "grad_norm": 0.507766689537816, |
| "learning_rate": 3.5816911083285165e-06, |
| "loss": 0.5798, |
| "step": 649 |
| }, |
| { |
| "epoch": 1.8987341772151898, |
| "grad_norm": 0.4863274459783003, |
| "learning_rate": 3.565380017172854e-06, |
| "loss": 0.5244, |
| "step": 650 |
| }, |
| { |
| "epoch": 1.901655306718598, |
| "grad_norm": 0.5805740988868203, |
| "learning_rate": 3.5490855460853275e-06, |
| "loss": 0.6026, |
| "step": 651 |
| }, |
| { |
| "epoch": 1.904576436222006, |
| "grad_norm": 0.5151920826178348, |
| "learning_rate": 3.5328078838373338e-06, |
| "loss": 0.5578, |
| "step": 652 |
| }, |
| { |
| "epoch": 1.9074975657254138, |
| "grad_norm": 0.4988300821200546, |
| "learning_rate": 3.516547219005542e-06, |
| "loss": 0.5639, |
| "step": 653 |
| }, |
| { |
| "epoch": 1.9104186952288218, |
| "grad_norm": 0.5020574341639038, |
| "learning_rate": 3.500303739969704e-06, |
| "loss": 0.5553, |
| "step": 654 |
| }, |
| { |
| "epoch": 1.9133398247322297, |
| "grad_norm": 0.47286299047516567, |
| "learning_rate": 3.4840776349104755e-06, |
| "loss": 0.5649, |
| "step": 655 |
| }, |
| { |
| "epoch": 1.916260954235638, |
| "grad_norm": 0.4681307130388586, |
| "learning_rate": 3.4678690918072335e-06, |
| "loss": 0.5682, |
| "step": 656 |
| }, |
| { |
| "epoch": 1.9191820837390456, |
| "grad_norm": 0.46768532251073486, |
| "learning_rate": 3.451678298435902e-06, |
| "loss": 0.5486, |
| "step": 657 |
| }, |
| { |
| "epoch": 1.9221032132424538, |
| "grad_norm": 0.4674675744224569, |
| "learning_rate": 3.4355054423667712e-06, |
| "loss": 0.5421, |
| "step": 658 |
| }, |
| { |
| "epoch": 1.9250243427458618, |
| "grad_norm": 0.5073113729998869, |
| "learning_rate": 3.4193507109623323e-06, |
| "loss": 0.5802, |
| "step": 659 |
| }, |
| { |
| "epoch": 1.9279454722492697, |
| "grad_norm": 0.47796072007470125, |
| "learning_rate": 3.4032142913750956e-06, |
| "loss": 0.5454, |
| "step": 660 |
| }, |
| { |
| "epoch": 1.9308666017526777, |
| "grad_norm": 0.44978612526614165, |
| "learning_rate": 3.3870963705454353e-06, |
| "loss": 0.6016, |
| "step": 661 |
| }, |
| { |
| "epoch": 1.9337877312560856, |
| "grad_norm": 0.49209754086259494, |
| "learning_rate": 3.370997135199413e-06, |
| "loss": 0.5898, |
| "step": 662 |
| }, |
| { |
| "epoch": 1.9367088607594938, |
| "grad_norm": 0.5261340174516683, |
| "learning_rate": 3.3549167718466245e-06, |
| "loss": 0.5438, |
| "step": 663 |
| }, |
| { |
| "epoch": 1.9396299902629015, |
| "grad_norm": 0.5215555650313894, |
| "learning_rate": 3.3388554667780272e-06, |
| "loss": 0.59, |
| "step": 664 |
| }, |
| { |
| "epoch": 1.9425511197663097, |
| "grad_norm": 0.4360310531256464, |
| "learning_rate": 3.322813406063794e-06, |
| "loss": 0.5361, |
| "step": 665 |
| }, |
| { |
| "epoch": 1.9454722492697176, |
| "grad_norm": 0.4338971494593997, |
| "learning_rate": 3.3067907755511473e-06, |
| "loss": 0.5217, |
| "step": 666 |
| }, |
| { |
| "epoch": 1.9483933787731256, |
| "grad_norm": 0.5993978224848855, |
| "learning_rate": 3.290787760862215e-06, |
| "loss": 0.6499, |
| "step": 667 |
| }, |
| { |
| "epoch": 1.9513145082765337, |
| "grad_norm": 0.4277968919618892, |
| "learning_rate": 3.274804547391872e-06, |
| "loss": 0.5059, |
| "step": 668 |
| }, |
| { |
| "epoch": 1.9542356377799415, |
| "grad_norm": 0.4833235638298786, |
| "learning_rate": 3.258841320305598e-06, |
| "loss": 0.6043, |
| "step": 669 |
| }, |
| { |
| "epoch": 1.9571567672833496, |
| "grad_norm": 0.4876241950113585, |
| "learning_rate": 3.242898264537331e-06, |
| "loss": 0.5982, |
| "step": 670 |
| }, |
| { |
| "epoch": 1.9600778967867576, |
| "grad_norm": 0.4287569825122473, |
| "learning_rate": 3.226975564787322e-06, |
| "loss": 0.5291, |
| "step": 671 |
| }, |
| { |
| "epoch": 1.9629990262901655, |
| "grad_norm": 0.559198431896861, |
| "learning_rate": 3.211073405520001e-06, |
| "loss": 0.5722, |
| "step": 672 |
| }, |
| { |
| "epoch": 1.9659201557935735, |
| "grad_norm": 0.5529747384120326, |
| "learning_rate": 3.195191970961833e-06, |
| "loss": 0.5766, |
| "step": 673 |
| }, |
| { |
| "epoch": 1.9688412852969814, |
| "grad_norm": 0.4620747878088523, |
| "learning_rate": 3.1793314450991895e-06, |
| "loss": 0.5911, |
| "step": 674 |
| }, |
| { |
| "epoch": 1.9717624148003896, |
| "grad_norm": 0.46006255679284874, |
| "learning_rate": 3.1634920116762175e-06, |
| "loss": 0.5622, |
| "step": 675 |
| }, |
| { |
| "epoch": 1.9746835443037973, |
| "grad_norm": 0.4259271639046577, |
| "learning_rate": 3.1476738541926993e-06, |
| "loss": 0.5287, |
| "step": 676 |
| }, |
| { |
| "epoch": 1.9776046738072055, |
| "grad_norm": 0.4729000058273843, |
| "learning_rate": 3.1318771559019455e-06, |
| "loss": 0.5863, |
| "step": 677 |
| }, |
| { |
| "epoch": 1.9805258033106135, |
| "grad_norm": 0.5100371866983662, |
| "learning_rate": 3.1161020998086566e-06, |
| "loss": 0.5687, |
| "step": 678 |
| }, |
| { |
| "epoch": 1.9834469328140214, |
| "grad_norm": 0.4379626484222088, |
| "learning_rate": 3.1003488686668076e-06, |
| "loss": 0.5155, |
| "step": 679 |
| }, |
| { |
| "epoch": 1.9863680623174296, |
| "grad_norm": 0.4844416917506815, |
| "learning_rate": 3.0846176449775363e-06, |
| "loss": 0.5782, |
| "step": 680 |
| }, |
| { |
| "epoch": 1.9892891918208373, |
| "grad_norm": 0.48815074844900985, |
| "learning_rate": 3.0689086109870188e-06, |
| "loss": 0.5709, |
| "step": 681 |
| }, |
| { |
| "epoch": 1.9922103213242455, |
| "grad_norm": 0.5064903715857387, |
| "learning_rate": 3.0532219486843686e-06, |
| "loss": 0.596, |
| "step": 682 |
| }, |
| { |
| "epoch": 1.9951314508276532, |
| "grad_norm": 0.492762116678178, |
| "learning_rate": 3.0375578397995178e-06, |
| "loss": 0.5766, |
| "step": 683 |
| }, |
| { |
| "epoch": 1.9980525803310614, |
| "grad_norm": 0.44675106308795903, |
| "learning_rate": 3.021916465801122e-06, |
| "loss": 0.5206, |
| "step": 684 |
| }, |
| { |
| "epoch": 2.0009737098344695, |
| "grad_norm": 0.9800800703433351, |
| "learning_rate": 3.0062980078944515e-06, |
| "loss": 0.8991, |
| "step": 685 |
| }, |
| { |
| "epoch": 2.0038948393378773, |
| "grad_norm": 0.4864022689636578, |
| "learning_rate": 2.990702647019294e-06, |
| "loss": 0.4938, |
| "step": 686 |
| }, |
| { |
| "epoch": 2.0068159688412854, |
| "grad_norm": 0.47889455472294706, |
| "learning_rate": 2.9751305638478555e-06, |
| "loss": 0.5913, |
| "step": 687 |
| }, |
| { |
| "epoch": 2.009737098344693, |
| "grad_norm": 0.41456980849801855, |
| "learning_rate": 2.9595819387826753e-06, |
| "loss": 0.4387, |
| "step": 688 |
| }, |
| { |
| "epoch": 2.0126582278481013, |
| "grad_norm": 0.5149845371997606, |
| "learning_rate": 2.9440569519545258e-06, |
| "loss": 0.5165, |
| "step": 689 |
| }, |
| { |
| "epoch": 2.015579357351509, |
| "grad_norm": 0.5752113401087547, |
| "learning_rate": 2.9285557832203328e-06, |
| "loss": 0.5373, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.0185004868549172, |
| "grad_norm": 0.5062618329379694, |
| "learning_rate": 2.9130786121610866e-06, |
| "loss": 0.4947, |
| "step": 691 |
| }, |
| { |
| "epoch": 2.0214216163583254, |
| "grad_norm": 0.49684882097652555, |
| "learning_rate": 2.897625618079769e-06, |
| "loss": 0.5208, |
| "step": 692 |
| }, |
| { |
| "epoch": 2.024342745861733, |
| "grad_norm": 0.53299035788322, |
| "learning_rate": 2.88219697999927e-06, |
| "loss": 0.5438, |
| "step": 693 |
| }, |
| { |
| "epoch": 2.0272638753651413, |
| "grad_norm": 0.5543675569999611, |
| "learning_rate": 2.8667928766603115e-06, |
| "loss": 0.5444, |
| "step": 694 |
| }, |
| { |
| "epoch": 2.030185004868549, |
| "grad_norm": 0.4572356814121919, |
| "learning_rate": 2.851413486519388e-06, |
| "loss": 0.4634, |
| "step": 695 |
| }, |
| { |
| "epoch": 2.033106134371957, |
| "grad_norm": 0.48340574114252344, |
| "learning_rate": 2.8360589877466848e-06, |
| "loss": 0.5391, |
| "step": 696 |
| }, |
| { |
| "epoch": 2.036027263875365, |
| "grad_norm": 0.5141120553439197, |
| "learning_rate": 2.8207295582240248e-06, |
| "loss": 0.4871, |
| "step": 697 |
| }, |
| { |
| "epoch": 2.038948393378773, |
| "grad_norm": 0.4913089013833701, |
| "learning_rate": 2.8054253755428017e-06, |
| "loss": 0.5574, |
| "step": 698 |
| }, |
| { |
| "epoch": 2.0418695228821813, |
| "grad_norm": 0.4601265937011902, |
| "learning_rate": 2.7901466170019242e-06, |
| "loss": 0.4977, |
| "step": 699 |
| }, |
| { |
| "epoch": 2.044790652385589, |
| "grad_norm": 0.5120604434660413, |
| "learning_rate": 2.774893459605766e-06, |
| "loss": 0.56, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.047711781888997, |
| "grad_norm": 0.47951033771571855, |
| "learning_rate": 2.7596660800621076e-06, |
| "loss": 0.5028, |
| "step": 701 |
| }, |
| { |
| "epoch": 2.050632911392405, |
| "grad_norm": 0.48501372243432067, |
| "learning_rate": 2.744464654780096e-06, |
| "loss": 0.5174, |
| "step": 702 |
| }, |
| { |
| "epoch": 2.053554040895813, |
| "grad_norm": 0.4741865276944199, |
| "learning_rate": 2.7292893598681934e-06, |
| "loss": 0.5104, |
| "step": 703 |
| }, |
| { |
| "epoch": 2.0564751703992212, |
| "grad_norm": 0.48778437702309935, |
| "learning_rate": 2.714140371132149e-06, |
| "loss": 0.5375, |
| "step": 704 |
| }, |
| { |
| "epoch": 2.059396299902629, |
| "grad_norm": 0.49018615472755583, |
| "learning_rate": 2.69901786407295e-06, |
| "loss": 0.5105, |
| "step": 705 |
| }, |
| { |
| "epoch": 2.062317429406037, |
| "grad_norm": 0.48782945068406763, |
| "learning_rate": 2.6839220138847966e-06, |
| "loss": 0.4899, |
| "step": 706 |
| }, |
| { |
| "epoch": 2.065238558909445, |
| "grad_norm": 0.4732567901538321, |
| "learning_rate": 2.6688529954530612e-06, |
| "loss": 0.5388, |
| "step": 707 |
| }, |
| { |
| "epoch": 2.068159688412853, |
| "grad_norm": 0.4935373768584108, |
| "learning_rate": 2.653810983352282e-06, |
| "loss": 0.5822, |
| "step": 708 |
| }, |
| { |
| "epoch": 2.0710808179162608, |
| "grad_norm": 0.44379266935434936, |
| "learning_rate": 2.6387961518441223e-06, |
| "loss": 0.4639, |
| "step": 709 |
| }, |
| { |
| "epoch": 2.074001947419669, |
| "grad_norm": 0.47555161755659076, |
| "learning_rate": 2.6238086748753587e-06, |
| "loss": 0.5153, |
| "step": 710 |
| }, |
| { |
| "epoch": 2.076923076923077, |
| "grad_norm": 0.4924498329014108, |
| "learning_rate": 2.6088487260758643e-06, |
| "loss": 0.5245, |
| "step": 711 |
| }, |
| { |
| "epoch": 2.079844206426485, |
| "grad_norm": 0.4492288702216835, |
| "learning_rate": 2.593916478756605e-06, |
| "loss": 0.4816, |
| "step": 712 |
| }, |
| { |
| "epoch": 2.082765335929893, |
| "grad_norm": 0.5008247473603998, |
| "learning_rate": 2.579012105907619e-06, |
| "loss": 0.5824, |
| "step": 713 |
| }, |
| { |
| "epoch": 2.0856864654333007, |
| "grad_norm": 0.479453244836006, |
| "learning_rate": 2.5641357801960186e-06, |
| "loss": 0.4738, |
| "step": 714 |
| }, |
| { |
| "epoch": 2.088607594936709, |
| "grad_norm": 0.49048895532662706, |
| "learning_rate": 2.5492876739639912e-06, |
| "loss": 0.5299, |
| "step": 715 |
| }, |
| { |
| "epoch": 2.091528724440117, |
| "grad_norm": 0.5369724976128885, |
| "learning_rate": 2.534467959226806e-06, |
| "loss": 0.5162, |
| "step": 716 |
| }, |
| { |
| "epoch": 2.094449853943525, |
| "grad_norm": 0.5137260447049424, |
| "learning_rate": 2.519676807670811e-06, |
| "loss": 0.5081, |
| "step": 717 |
| }, |
| { |
| "epoch": 2.097370983446933, |
| "grad_norm": 0.4930859741487903, |
| "learning_rate": 2.504914390651453e-06, |
| "loss": 0.5576, |
| "step": 718 |
| }, |
| { |
| "epoch": 2.1002921129503407, |
| "grad_norm": 0.4775125343363975, |
| "learning_rate": 2.4901808791912864e-06, |
| "loss": 0.4998, |
| "step": 719 |
| }, |
| { |
| "epoch": 2.103213242453749, |
| "grad_norm": 0.5214779418043487, |
| "learning_rate": 2.475476443977996e-06, |
| "loss": 0.5696, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.1061343719571566, |
| "grad_norm": 0.47977761871037206, |
| "learning_rate": 2.460801255362425e-06, |
| "loss": 0.4764, |
| "step": 721 |
| }, |
| { |
| "epoch": 2.1090555014605648, |
| "grad_norm": 0.48061705012654404, |
| "learning_rate": 2.446155483356582e-06, |
| "loss": 0.5357, |
| "step": 722 |
| }, |
| { |
| "epoch": 2.111976630963973, |
| "grad_norm": 0.45629547557990546, |
| "learning_rate": 2.4315392976316923e-06, |
| "loss": 0.4915, |
| "step": 723 |
| }, |
| { |
| "epoch": 2.1148977604673806, |
| "grad_norm": 0.4780522247938775, |
| "learning_rate": 2.4169528675162256e-06, |
| "loss": 0.4838, |
| "step": 724 |
| }, |
| { |
| "epoch": 2.117818889970789, |
| "grad_norm": 0.49801842180761746, |
| "learning_rate": 2.40239636199393e-06, |
| "loss": 0.537, |
| "step": 725 |
| }, |
| { |
| "epoch": 2.1207400194741965, |
| "grad_norm": 0.4846689678826097, |
| "learning_rate": 2.3878699497018763e-06, |
| "loss": 0.549, |
| "step": 726 |
| }, |
| { |
| "epoch": 2.1236611489776047, |
| "grad_norm": 0.5018761582798917, |
| "learning_rate": 2.373373798928507e-06, |
| "loss": 0.5042, |
| "step": 727 |
| }, |
| { |
| "epoch": 2.1265822784810124, |
| "grad_norm": 0.5072064832309269, |
| "learning_rate": 2.358908077611684e-06, |
| "loss": 0.5795, |
| "step": 728 |
| }, |
| { |
| "epoch": 2.1295034079844206, |
| "grad_norm": 0.46610002849697485, |
| "learning_rate": 2.344472953336747e-06, |
| "loss": 0.468, |
| "step": 729 |
| }, |
| { |
| "epoch": 2.132424537487829, |
| "grad_norm": 0.5002554866584938, |
| "learning_rate": 2.3300685933345656e-06, |
| "loss": 0.5022, |
| "step": 730 |
| }, |
| { |
| "epoch": 2.1353456669912365, |
| "grad_norm": 0.5324744280306557, |
| "learning_rate": 2.3156951644796065e-06, |
| "loss": 0.5877, |
| "step": 731 |
| }, |
| { |
| "epoch": 2.1382667964946447, |
| "grad_norm": 0.4804432916120911, |
| "learning_rate": 2.3013528332879976e-06, |
| "loss": 0.4814, |
| "step": 732 |
| }, |
| { |
| "epoch": 2.1411879259980524, |
| "grad_norm": 0.47091232830887136, |
| "learning_rate": 2.287041765915606e-06, |
| "loss": 0.4953, |
| "step": 733 |
| }, |
| { |
| "epoch": 2.1441090555014606, |
| "grad_norm": 0.4535727155626379, |
| "learning_rate": 2.272762128156101e-06, |
| "loss": 0.5269, |
| "step": 734 |
| }, |
| { |
| "epoch": 2.1470301850048688, |
| "grad_norm": 0.47112551281102016, |
| "learning_rate": 2.2585140854390432e-06, |
| "loss": 0.5244, |
| "step": 735 |
| }, |
| { |
| "epoch": 2.1499513145082765, |
| "grad_norm": 0.4545532988740129, |
| "learning_rate": 2.2442978028279634e-06, |
| "loss": 0.5341, |
| "step": 736 |
| }, |
| { |
| "epoch": 2.1528724440116846, |
| "grad_norm": 0.5024829916260531, |
| "learning_rate": 2.2301134450184535e-06, |
| "loss": 0.5243, |
| "step": 737 |
| }, |
| { |
| "epoch": 2.1557935735150924, |
| "grad_norm": 0.457679806572502, |
| "learning_rate": 2.215961176336255e-06, |
| "loss": 0.5157, |
| "step": 738 |
| }, |
| { |
| "epoch": 2.1587147030185005, |
| "grad_norm": 0.41799952326682965, |
| "learning_rate": 2.2018411607353572e-06, |
| "loss": 0.4653, |
| "step": 739 |
| }, |
| { |
| "epoch": 2.1616358325219083, |
| "grad_norm": 0.4787265660707973, |
| "learning_rate": 2.187753561796097e-06, |
| "loss": 0.5521, |
| "step": 740 |
| }, |
| { |
| "epoch": 2.1645569620253164, |
| "grad_norm": 0.48532252970802797, |
| "learning_rate": 2.1736985427232684e-06, |
| "loss": 0.5736, |
| "step": 741 |
| }, |
| { |
| "epoch": 2.1674780915287246, |
| "grad_norm": 0.46670045472968474, |
| "learning_rate": 2.159676266344222e-06, |
| "loss": 0.4856, |
| "step": 742 |
| }, |
| { |
| "epoch": 2.1703992210321323, |
| "grad_norm": 0.456907097163635, |
| "learning_rate": 2.1456868951069875e-06, |
| "loss": 0.513, |
| "step": 743 |
| }, |
| { |
| "epoch": 2.1733203505355405, |
| "grad_norm": 0.43480032818372993, |
| "learning_rate": 2.1317305910783886e-06, |
| "loss": 0.5183, |
| "step": 744 |
| }, |
| { |
| "epoch": 2.1762414800389482, |
| "grad_norm": 0.4646674951011356, |
| "learning_rate": 2.117807515942163e-06, |
| "loss": 0.5169, |
| "step": 745 |
| }, |
| { |
| "epoch": 2.1791626095423564, |
| "grad_norm": 0.5402952715425572, |
| "learning_rate": 2.1039178309970975e-06, |
| "loss": 0.5256, |
| "step": 746 |
| }, |
| { |
| "epoch": 2.1820837390457646, |
| "grad_norm": 0.44604300646980855, |
| "learning_rate": 2.090061697155147e-06, |
| "loss": 0.5264, |
| "step": 747 |
| }, |
| { |
| "epoch": 2.1850048685491723, |
| "grad_norm": 0.4855629840620991, |
| "learning_rate": 2.076239274939582e-06, |
| "loss": 0.497, |
| "step": 748 |
| }, |
| { |
| "epoch": 2.1879259980525805, |
| "grad_norm": 0.450371819506377, |
| "learning_rate": 2.062450724483118e-06, |
| "loss": 0.492, |
| "step": 749 |
| }, |
| { |
| "epoch": 2.190847127555988, |
| "grad_norm": 0.4675791607919664, |
| "learning_rate": 2.0486962055260744e-06, |
| "loss": 0.4833, |
| "step": 750 |
| }, |
| { |
| "epoch": 2.1937682570593964, |
| "grad_norm": 0.4931924675592549, |
| "learning_rate": 2.03497587741451e-06, |
| "loss": 0.5325, |
| "step": 751 |
| }, |
| { |
| "epoch": 2.196689386562804, |
| "grad_norm": 0.4728333199503108, |
| "learning_rate": 2.021289899098384e-06, |
| "loss": 0.5753, |
| "step": 752 |
| }, |
| { |
| "epoch": 2.1996105160662123, |
| "grad_norm": 0.44477975654172386, |
| "learning_rate": 2.0076384291297134e-06, |
| "loss": 0.5095, |
| "step": 753 |
| }, |
| { |
| "epoch": 2.2025316455696204, |
| "grad_norm": 0.42502487470067785, |
| "learning_rate": 1.994021625660737e-06, |
| "loss": 0.5004, |
| "step": 754 |
| }, |
| { |
| "epoch": 2.205452775073028, |
| "grad_norm": 0.4682554593875959, |
| "learning_rate": 1.9804396464420798e-06, |
| "loss": 0.5377, |
| "step": 755 |
| }, |
| { |
| "epoch": 2.2083739045764363, |
| "grad_norm": 0.478543398252641, |
| "learning_rate": 1.966892648820932e-06, |
| "loss": 0.5214, |
| "step": 756 |
| }, |
| { |
| "epoch": 2.211295034079844, |
| "grad_norm": 0.4638445072960299, |
| "learning_rate": 1.953380789739216e-06, |
| "loss": 0.522, |
| "step": 757 |
| }, |
| { |
| "epoch": 2.2142161635832522, |
| "grad_norm": 0.44455755591464263, |
| "learning_rate": 1.939904225731783e-06, |
| "loss": 0.5279, |
| "step": 758 |
| }, |
| { |
| "epoch": 2.21713729308666, |
| "grad_norm": 0.43381712587340415, |
| "learning_rate": 1.9264631129245836e-06, |
| "loss": 0.5082, |
| "step": 759 |
| }, |
| { |
| "epoch": 2.220058422590068, |
| "grad_norm": 0.46180168699161817, |
| "learning_rate": 1.9130576070328695e-06, |
| "loss": 0.5129, |
| "step": 760 |
| }, |
| { |
| "epoch": 2.2229795520934763, |
| "grad_norm": 0.4554045010755852, |
| "learning_rate": 1.8996878633593829e-06, |
| "loss": 0.499, |
| "step": 761 |
| }, |
| { |
| "epoch": 2.225900681596884, |
| "grad_norm": 0.4749182830100526, |
| "learning_rate": 1.8863540367925676e-06, |
| "loss": 0.5343, |
| "step": 762 |
| }, |
| { |
| "epoch": 2.228821811100292, |
| "grad_norm": 0.45836107720097835, |
| "learning_rate": 1.873056281804762e-06, |
| "loss": 0.4869, |
| "step": 763 |
| }, |
| { |
| "epoch": 2.2317429406037, |
| "grad_norm": 0.45195906937962993, |
| "learning_rate": 1.8597947524504178e-06, |
| "loss": 0.5201, |
| "step": 764 |
| }, |
| { |
| "epoch": 2.234664070107108, |
| "grad_norm": 0.45112191767699233, |
| "learning_rate": 1.8465696023643115e-06, |
| "loss": 0.5354, |
| "step": 765 |
| }, |
| { |
| "epoch": 2.2375851996105163, |
| "grad_norm": 0.49025509449815313, |
| "learning_rate": 1.8333809847597644e-06, |
| "loss": 0.558, |
| "step": 766 |
| }, |
| { |
| "epoch": 2.240506329113924, |
| "grad_norm": 0.4486301410086504, |
| "learning_rate": 1.8202290524268761e-06, |
| "loss": 0.4795, |
| "step": 767 |
| }, |
| { |
| "epoch": 2.243427458617332, |
| "grad_norm": 0.45776452388300076, |
| "learning_rate": 1.8071139577307416e-06, |
| "loss": 0.5228, |
| "step": 768 |
| }, |
| { |
| "epoch": 2.24634858812074, |
| "grad_norm": 0.4414886979982915, |
| "learning_rate": 1.7940358526096885e-06, |
| "loss": 0.4903, |
| "step": 769 |
| }, |
| { |
| "epoch": 2.249269717624148, |
| "grad_norm": 0.44672029042681155, |
| "learning_rate": 1.7809948885735295e-06, |
| "loss": 0.5906, |
| "step": 770 |
| }, |
| { |
| "epoch": 2.252190847127556, |
| "grad_norm": 0.40774965094554166, |
| "learning_rate": 1.7679912167017922e-06, |
| "loss": 0.4726, |
| "step": 771 |
| }, |
| { |
| "epoch": 2.255111976630964, |
| "grad_norm": 0.42904714451405873, |
| "learning_rate": 1.7550249876419746e-06, |
| "loss": 0.5263, |
| "step": 772 |
| }, |
| { |
| "epoch": 2.258033106134372, |
| "grad_norm": 0.46232426979933267, |
| "learning_rate": 1.7420963516078016e-06, |
| "loss": 0.5207, |
| "step": 773 |
| }, |
| { |
| "epoch": 2.26095423563778, |
| "grad_norm": 0.43964498189412426, |
| "learning_rate": 1.7292054583774809e-06, |
| "loss": 0.5634, |
| "step": 774 |
| }, |
| { |
| "epoch": 2.263875365141188, |
| "grad_norm": 0.4614885797674806, |
| "learning_rate": 1.7163524572919748e-06, |
| "loss": 0.5167, |
| "step": 775 |
| }, |
| { |
| "epoch": 2.2667964946445958, |
| "grad_norm": 0.4326167271312955, |
| "learning_rate": 1.7035374972532593e-06, |
| "loss": 0.4734, |
| "step": 776 |
| }, |
| { |
| "epoch": 2.269717624148004, |
| "grad_norm": 0.477108582339082, |
| "learning_rate": 1.6907607267226079e-06, |
| "loss": 0.5436, |
| "step": 777 |
| }, |
| { |
| "epoch": 2.272638753651412, |
| "grad_norm": 0.45201769752670545, |
| "learning_rate": 1.6780222937188662e-06, |
| "loss": 0.5207, |
| "step": 778 |
| }, |
| { |
| "epoch": 2.27555988315482, |
| "grad_norm": 0.48262208131523315, |
| "learning_rate": 1.665322345816746e-06, |
| "loss": 0.5304, |
| "step": 779 |
| }, |
| { |
| "epoch": 2.278481012658228, |
| "grad_norm": 0.45623936201495924, |
| "learning_rate": 1.6526610301451028e-06, |
| "loss": 0.4604, |
| "step": 780 |
| }, |
| { |
| "epoch": 2.2814021421616357, |
| "grad_norm": 0.45941621411519823, |
| "learning_rate": 1.6400384933852403e-06, |
| "loss": 0.508, |
| "step": 781 |
| }, |
| { |
| "epoch": 2.284323271665044, |
| "grad_norm": 0.4640162246141768, |
| "learning_rate": 1.6274548817692088e-06, |
| "loss": 0.5429, |
| "step": 782 |
| }, |
| { |
| "epoch": 2.2872444011684516, |
| "grad_norm": 0.4517120340093356, |
| "learning_rate": 1.6149103410781086e-06, |
| "loss": 0.5126, |
| "step": 783 |
| }, |
| { |
| "epoch": 2.29016553067186, |
| "grad_norm": 0.4528239429323758, |
| "learning_rate": 1.6024050166404097e-06, |
| "loss": 0.4995, |
| "step": 784 |
| }, |
| { |
| "epoch": 2.293086660175268, |
| "grad_norm": 0.4440949852742247, |
| "learning_rate": 1.5899390533302538e-06, |
| "loss": 0.5333, |
| "step": 785 |
| }, |
| { |
| "epoch": 2.2960077896786757, |
| "grad_norm": 0.3970036577732869, |
| "learning_rate": 1.5775125955657877e-06, |
| "loss": 0.5012, |
| "step": 786 |
| }, |
| { |
| "epoch": 2.298928919182084, |
| "grad_norm": 0.4484643953022932, |
| "learning_rate": 1.5651257873074898e-06, |
| "loss": 0.5413, |
| "step": 787 |
| }, |
| { |
| "epoch": 2.3018500486854916, |
| "grad_norm": 0.4679410901074801, |
| "learning_rate": 1.5527787720564946e-06, |
| "loss": 0.5229, |
| "step": 788 |
| }, |
| { |
| "epoch": 2.3047711781888998, |
| "grad_norm": 0.450714999626221, |
| "learning_rate": 1.5404716928529356e-06, |
| "loss": 0.5079, |
| "step": 789 |
| }, |
| { |
| "epoch": 2.3076923076923075, |
| "grad_norm": 0.46250776338412736, |
| "learning_rate": 1.5282046922742876e-06, |
| "loss": 0.5056, |
| "step": 790 |
| }, |
| { |
| "epoch": 2.3106134371957157, |
| "grad_norm": 0.47191060686599, |
| "learning_rate": 1.515977912433717e-06, |
| "loss": 0.5279, |
| "step": 791 |
| }, |
| { |
| "epoch": 2.313534566699124, |
| "grad_norm": 0.493016587785878, |
| "learning_rate": 1.50379149497843e-06, |
| "loss": 0.5483, |
| "step": 792 |
| }, |
| { |
| "epoch": 2.3164556962025316, |
| "grad_norm": 0.47810281591953624, |
| "learning_rate": 1.4916455810880358e-06, |
| "loss": 0.4898, |
| "step": 793 |
| }, |
| { |
| "epoch": 2.3193768257059397, |
| "grad_norm": 0.4478075131535074, |
| "learning_rate": 1.4795403114729095e-06, |
| "loss": 0.5101, |
| "step": 794 |
| }, |
| { |
| "epoch": 2.3222979552093475, |
| "grad_norm": 0.4875641000416008, |
| "learning_rate": 1.4674758263725614e-06, |
| "loss": 0.5854, |
| "step": 795 |
| }, |
| { |
| "epoch": 2.3252190847127556, |
| "grad_norm": 0.42525631573701944, |
| "learning_rate": 1.4554522655540176e-06, |
| "loss": 0.4989, |
| "step": 796 |
| }, |
| { |
| "epoch": 2.3281402142161634, |
| "grad_norm": 0.41706543335088775, |
| "learning_rate": 1.4434697683101928e-06, |
| "loss": 0.516, |
| "step": 797 |
| }, |
| { |
| "epoch": 2.3310613437195715, |
| "grad_norm": 0.4524630863860165, |
| "learning_rate": 1.4315284734582802e-06, |
| "loss": 0.4936, |
| "step": 798 |
| }, |
| { |
| "epoch": 2.3339824732229797, |
| "grad_norm": 0.5066460944480237, |
| "learning_rate": 1.4196285193381431e-06, |
| "loss": 0.5665, |
| "step": 799 |
| }, |
| { |
| "epoch": 2.3369036027263874, |
| "grad_norm": 0.4886439078673485, |
| "learning_rate": 1.4077700438107183e-06, |
| "loss": 0.5191, |
| "step": 800 |
| }, |
| { |
| "epoch": 2.3398247322297956, |
| "grad_norm": 0.42502615563238366, |
| "learning_rate": 1.3959531842564046e-06, |
| "loss": 0.4785, |
| "step": 801 |
| }, |
| { |
| "epoch": 2.3427458617332033, |
| "grad_norm": 0.45982217381683227, |
| "learning_rate": 1.3841780775734847e-06, |
| "loss": 0.5643, |
| "step": 802 |
| }, |
| { |
| "epoch": 2.3456669912366115, |
| "grad_norm": 0.43225051785834356, |
| "learning_rate": 1.3724448601765328e-06, |
| "loss": 0.5629, |
| "step": 803 |
| }, |
| { |
| "epoch": 2.3485881207400197, |
| "grad_norm": 0.44393943057319624, |
| "learning_rate": 1.3607536679948397e-06, |
| "loss": 0.456, |
| "step": 804 |
| }, |
| { |
| "epoch": 2.3515092502434274, |
| "grad_norm": 0.44935770423521926, |
| "learning_rate": 1.3491046364708294e-06, |
| "loss": 0.52, |
| "step": 805 |
| }, |
| { |
| "epoch": 2.3544303797468356, |
| "grad_norm": 0.4573183127076476, |
| "learning_rate": 1.337497900558497e-06, |
| "loss": 0.5231, |
| "step": 806 |
| }, |
| { |
| "epoch": 2.3573515092502433, |
| "grad_norm": 0.4471873705684597, |
| "learning_rate": 1.325933594721841e-06, |
| "loss": 0.5191, |
| "step": 807 |
| }, |
| { |
| "epoch": 2.3602726387536515, |
| "grad_norm": 0.46232878806411926, |
| "learning_rate": 1.3144118529333126e-06, |
| "loss": 0.4641, |
| "step": 808 |
| }, |
| { |
| "epoch": 2.3631937682570596, |
| "grad_norm": 0.46569289026754496, |
| "learning_rate": 1.3029328086722537e-06, |
| "loss": 0.5363, |
| "step": 809 |
| }, |
| { |
| "epoch": 2.3661148977604674, |
| "grad_norm": 0.4515090725602521, |
| "learning_rate": 1.2914965949233572e-06, |
| "loss": 0.5278, |
| "step": 810 |
| }, |
| { |
| "epoch": 2.3690360272638755, |
| "grad_norm": 0.45127137221258873, |
| "learning_rate": 1.2801033441751244e-06, |
| "loss": 0.4919, |
| "step": 811 |
| }, |
| { |
| "epoch": 2.3719571567672832, |
| "grad_norm": 0.45307816143888285, |
| "learning_rate": 1.268753188418329e-06, |
| "loss": 0.5063, |
| "step": 812 |
| }, |
| { |
| "epoch": 2.3748782862706914, |
| "grad_norm": 0.4681309102712795, |
| "learning_rate": 1.257446259144494e-06, |
| "loss": 0.5185, |
| "step": 813 |
| }, |
| { |
| "epoch": 2.377799415774099, |
| "grad_norm": 0.4665552455306928, |
| "learning_rate": 1.24618268734436e-06, |
| "loss": 0.5426, |
| "step": 814 |
| }, |
| { |
| "epoch": 2.3807205452775073, |
| "grad_norm": 0.45997968978470405, |
| "learning_rate": 1.2349626035063705e-06, |
| "loss": 0.4996, |
| "step": 815 |
| }, |
| { |
| "epoch": 2.3836416747809155, |
| "grad_norm": 0.470113793180687, |
| "learning_rate": 1.2237861376151632e-06, |
| "loss": 0.5392, |
| "step": 816 |
| }, |
| { |
| "epoch": 2.386562804284323, |
| "grad_norm": 0.47230105708445114, |
| "learning_rate": 1.2126534191500622e-06, |
| "loss": 0.5361, |
| "step": 817 |
| }, |
| { |
| "epoch": 2.3894839337877314, |
| "grad_norm": 0.4439367141535501, |
| "learning_rate": 1.2015645770835765e-06, |
| "loss": 0.4732, |
| "step": 818 |
| }, |
| { |
| "epoch": 2.392405063291139, |
| "grad_norm": 0.4656016270422327, |
| "learning_rate": 1.1905197398799074e-06, |
| "loss": 0.5227, |
| "step": 819 |
| }, |
| { |
| "epoch": 2.3953261927945473, |
| "grad_norm": 0.46417888892808024, |
| "learning_rate": 1.1795190354934587e-06, |
| "loss": 0.5606, |
| "step": 820 |
| }, |
| { |
| "epoch": 2.398247322297955, |
| "grad_norm": 0.41153069172218604, |
| "learning_rate": 1.168562591367361e-06, |
| "loss": 0.5035, |
| "step": 821 |
| }, |
| { |
| "epoch": 2.401168451801363, |
| "grad_norm": 0.4591933177826969, |
| "learning_rate": 1.1576505344319843e-06, |
| "loss": 0.5485, |
| "step": 822 |
| }, |
| { |
| "epoch": 2.4040895813047714, |
| "grad_norm": 0.48641885058814227, |
| "learning_rate": 1.146782991103475e-06, |
| "loss": 0.5122, |
| "step": 823 |
| }, |
| { |
| "epoch": 2.407010710808179, |
| "grad_norm": 0.452636414728882, |
| "learning_rate": 1.1359600872822879e-06, |
| "loss": 0.528, |
| "step": 824 |
| }, |
| { |
| "epoch": 2.4099318403115872, |
| "grad_norm": 0.4637194713579063, |
| "learning_rate": 1.1251819483517334e-06, |
| "loss": 0.5335, |
| "step": 825 |
| }, |
| { |
| "epoch": 2.412852969814995, |
| "grad_norm": 0.46499658066713967, |
| "learning_rate": 1.1144486991765175e-06, |
| "loss": 0.519, |
| "step": 826 |
| }, |
| { |
| "epoch": 2.415774099318403, |
| "grad_norm": 0.4519765342482912, |
| "learning_rate": 1.1037604641012995e-06, |
| "loss": 0.5488, |
| "step": 827 |
| }, |
| { |
| "epoch": 2.418695228821811, |
| "grad_norm": 0.4382247666420335, |
| "learning_rate": 1.0931173669492472e-06, |
| "loss": 0.4853, |
| "step": 828 |
| }, |
| { |
| "epoch": 2.421616358325219, |
| "grad_norm": 0.45306336606053793, |
| "learning_rate": 1.0825195310206132e-06, |
| "loss": 0.5224, |
| "step": 829 |
| }, |
| { |
| "epoch": 2.424537487828627, |
| "grad_norm": 0.4473053801916244, |
| "learning_rate": 1.0719670790912928e-06, |
| "loss": 0.4971, |
| "step": 830 |
| }, |
| { |
| "epoch": 2.427458617332035, |
| "grad_norm": 0.44465614500766465, |
| "learning_rate": 1.0614601334114099e-06, |
| "loss": 0.5404, |
| "step": 831 |
| }, |
| { |
| "epoch": 2.430379746835443, |
| "grad_norm": 0.4251639911624754, |
| "learning_rate": 1.0509988157038952e-06, |
| "loss": 0.5179, |
| "step": 832 |
| }, |
| { |
| "epoch": 2.433300876338851, |
| "grad_norm": 0.40203118130358384, |
| "learning_rate": 1.0405832471630862e-06, |
| "loss": 0.4465, |
| "step": 833 |
| }, |
| { |
| "epoch": 2.436222005842259, |
| "grad_norm": 0.49627881222358805, |
| "learning_rate": 1.030213548453311e-06, |
| "loss": 0.5782, |
| "step": 834 |
| }, |
| { |
| "epoch": 2.439143135345667, |
| "grad_norm": 0.4352033965269781, |
| "learning_rate": 1.019889839707498e-06, |
| "loss": 0.471, |
| "step": 835 |
| }, |
| { |
| "epoch": 2.442064264849075, |
| "grad_norm": 0.4112365791877552, |
| "learning_rate": 1.0096122405257802e-06, |
| "loss": 0.525, |
| "step": 836 |
| }, |
| { |
| "epoch": 2.444985394352483, |
| "grad_norm": 0.46113597408698487, |
| "learning_rate": 9.99380869974116e-07, |
| "loss": 0.55, |
| "step": 837 |
| }, |
| { |
| "epoch": 2.447906523855891, |
| "grad_norm": 0.42894572321057406, |
| "learning_rate": 9.891958465828999e-07, |
| "loss": 0.4883, |
| "step": 838 |
| }, |
| { |
| "epoch": 2.450827653359299, |
| "grad_norm": 0.4306576340142323, |
| "learning_rate": 9.790572883455974e-07, |
| "loss": 0.5127, |
| "step": 839 |
| }, |
| { |
| "epoch": 2.453748782862707, |
| "grad_norm": 0.47843377943065934, |
| "learning_rate": 9.689653127173743e-07, |
| "loss": 0.5376, |
| "step": 840 |
| }, |
| { |
| "epoch": 2.456669912366115, |
| "grad_norm": 0.48286119801655825, |
| "learning_rate": 9.589200366137375e-07, |
| "loss": 0.5108, |
| "step": 841 |
| }, |
| { |
| "epoch": 2.459591041869523, |
| "grad_norm": 0.4840439892081914, |
| "learning_rate": 9.489215764091813e-07, |
| "loss": 0.5436, |
| "step": 842 |
| }, |
| { |
| "epoch": 2.4625121713729308, |
| "grad_norm": 0.4360507831964012, |
| "learning_rate": 9.389700479358365e-07, |
| "loss": 0.4733, |
| "step": 843 |
| }, |
| { |
| "epoch": 2.465433300876339, |
| "grad_norm": 0.449450069287443, |
| "learning_rate": 9.290655664821296e-07, |
| "loss": 0.5599, |
| "step": 844 |
| }, |
| { |
| "epoch": 2.4683544303797467, |
| "grad_norm": 0.41280059188581175, |
| "learning_rate": 9.192082467914465e-07, |
| "loss": 0.4792, |
| "step": 845 |
| }, |
| { |
| "epoch": 2.471275559883155, |
| "grad_norm": 0.5093589634384059, |
| "learning_rate": 9.093982030608095e-07, |
| "loss": 0.5397, |
| "step": 846 |
| }, |
| { |
| "epoch": 2.474196689386563, |
| "grad_norm": 0.45448988250512634, |
| "learning_rate": 8.996355489395442e-07, |
| "loss": 0.4975, |
| "step": 847 |
| }, |
| { |
| "epoch": 2.4771178188899707, |
| "grad_norm": 0.44480146821328176, |
| "learning_rate": 8.899203975279674e-07, |
| "loss": 0.4914, |
| "step": 848 |
| }, |
| { |
| "epoch": 2.480038948393379, |
| "grad_norm": 0.4896804183179627, |
| "learning_rate": 8.802528613760775e-07, |
| "loss": 0.5599, |
| "step": 849 |
| }, |
| { |
| "epoch": 2.4829600778967866, |
| "grad_norm": 0.4529609749837371, |
| "learning_rate": 8.706330524822548e-07, |
| "loss": 0.4598, |
| "step": 850 |
| }, |
| { |
| "epoch": 2.485881207400195, |
| "grad_norm": 0.4141270566207645, |
| "learning_rate": 8.610610822919546e-07, |
| "loss": 0.5195, |
| "step": 851 |
| }, |
| { |
| "epoch": 2.4888023369036025, |
| "grad_norm": 0.4413472687231943, |
| "learning_rate": 8.515370616964219e-07, |
| "loss": 0.5181, |
| "step": 852 |
| }, |
| { |
| "epoch": 2.4917234664070107, |
| "grad_norm": 0.4120555785416349, |
| "learning_rate": 8.420611010314062e-07, |
| "loss": 0.5267, |
| "step": 853 |
| }, |
| { |
| "epoch": 2.494644595910419, |
| "grad_norm": 0.4661650695044543, |
| "learning_rate": 8.326333100758843e-07, |
| "loss": 0.5086, |
| "step": 854 |
| }, |
| { |
| "epoch": 2.4975657254138266, |
| "grad_norm": 0.5073361612635612, |
| "learning_rate": 8.232537980507848e-07, |
| "loss": 0.5536, |
| "step": 855 |
| }, |
| { |
| "epoch": 2.5004868549172348, |
| "grad_norm": 0.4961949890878462, |
| "learning_rate": 8.13922673617727e-07, |
| "loss": 0.5238, |
| "step": 856 |
| }, |
| { |
| "epoch": 2.5034079844206425, |
| "grad_norm": 0.4720376448013444, |
| "learning_rate": 8.046400448777575e-07, |
| "loss": 0.4823, |
| "step": 857 |
| }, |
| { |
| "epoch": 2.5063291139240507, |
| "grad_norm": 0.49809468171059407, |
| "learning_rate": 7.954060193701019e-07, |
| "loss": 0.5466, |
| "step": 858 |
| }, |
| { |
| "epoch": 2.5092502434274584, |
| "grad_norm": 0.4639329084639315, |
| "learning_rate": 7.862207040709191e-07, |
| "loss": 0.5104, |
| "step": 859 |
| }, |
| { |
| "epoch": 2.5121713729308666, |
| "grad_norm": 0.47993248882289974, |
| "learning_rate": 7.770842053920585e-07, |
| "loss": 0.5379, |
| "step": 860 |
| }, |
| { |
| "epoch": 2.5150925024342747, |
| "grad_norm": 0.42897728377766037, |
| "learning_rate": 7.679966291798297e-07, |
| "loss": 0.482, |
| "step": 861 |
| }, |
| { |
| "epoch": 2.5180136319376825, |
| "grad_norm": 0.45778561195288503, |
| "learning_rate": 7.589580807137742e-07, |
| "loss": 0.5032, |
| "step": 862 |
| }, |
| { |
| "epoch": 2.5209347614410906, |
| "grad_norm": 0.46270673489951974, |
| "learning_rate": 7.4996866470545e-07, |
| "loss": 0.5706, |
| "step": 863 |
| }, |
| { |
| "epoch": 2.523855890944499, |
| "grad_norm": 0.42072713048943156, |
| "learning_rate": 7.410284852972127e-07, |
| "loss": 0.5204, |
| "step": 864 |
| }, |
| { |
| "epoch": 2.5267770204479065, |
| "grad_norm": 0.40846266187572267, |
| "learning_rate": 7.321376460610136e-07, |
| "loss": 0.4916, |
| "step": 865 |
| }, |
| { |
| "epoch": 2.5296981499513143, |
| "grad_norm": 0.4443443498829621, |
| "learning_rate": 7.232962499972002e-07, |
| "loss": 0.5395, |
| "step": 866 |
| }, |
| { |
| "epoch": 2.5326192794547224, |
| "grad_norm": 0.459368960082765, |
| "learning_rate": 7.145043995333173e-07, |
| "loss": 0.4924, |
| "step": 867 |
| }, |
| { |
| "epoch": 2.5355404089581306, |
| "grad_norm": 0.47824210909228654, |
| "learning_rate": 7.057621965229267e-07, |
| "loss": 0.513, |
| "step": 868 |
| }, |
| { |
| "epoch": 2.5384615384615383, |
| "grad_norm": 0.438214605329431, |
| "learning_rate": 6.970697422444228e-07, |
| "loss": 0.5695, |
| "step": 869 |
| }, |
| { |
| "epoch": 2.5413826679649465, |
| "grad_norm": 0.45798584370168316, |
| "learning_rate": 6.884271373998608e-07, |
| "loss": 0.4925, |
| "step": 870 |
| }, |
| { |
| "epoch": 2.5443037974683547, |
| "grad_norm": 0.4811217026678713, |
| "learning_rate": 6.798344821137947e-07, |
| "loss": 0.5083, |
| "step": 871 |
| }, |
| { |
| "epoch": 2.5472249269717624, |
| "grad_norm": 0.4418809012550695, |
| "learning_rate": 6.71291875932108e-07, |
| "loss": 0.5352, |
| "step": 872 |
| }, |
| { |
| "epoch": 2.5501460564751706, |
| "grad_norm": 0.42576085849329387, |
| "learning_rate": 6.62799417820868e-07, |
| "loss": 0.5262, |
| "step": 873 |
| }, |
| { |
| "epoch": 2.5530671859785783, |
| "grad_norm": 0.4287609653415205, |
| "learning_rate": 6.543572061651738e-07, |
| "loss": 0.5399, |
| "step": 874 |
| }, |
| { |
| "epoch": 2.5559883154819865, |
| "grad_norm": 0.45232434071279953, |
| "learning_rate": 6.459653387680248e-07, |
| "loss": 0.4725, |
| "step": 875 |
| }, |
| { |
| "epoch": 2.558909444985394, |
| "grad_norm": 0.4554887075692202, |
| "learning_rate": 6.376239128491784e-07, |
| "loss": 0.5122, |
| "step": 876 |
| }, |
| { |
| "epoch": 2.5618305744888024, |
| "grad_norm": 0.46277880760014206, |
| "learning_rate": 6.293330250440277e-07, |
| "loss": 0.4941, |
| "step": 877 |
| }, |
| { |
| "epoch": 2.5647517039922105, |
| "grad_norm": 0.4962082268099505, |
| "learning_rate": 6.210927714024834e-07, |
| "loss": 0.5592, |
| "step": 878 |
| }, |
| { |
| "epoch": 2.5676728334956183, |
| "grad_norm": 0.44324064458458867, |
| "learning_rate": 6.129032473878582e-07, |
| "loss": 0.5282, |
| "step": 879 |
| }, |
| { |
| "epoch": 2.5705939629990264, |
| "grad_norm": 0.43320029335686094, |
| "learning_rate": 6.047645478757635e-07, |
| "loss": 0.4562, |
| "step": 880 |
| }, |
| { |
| "epoch": 2.573515092502434, |
| "grad_norm": 0.45163259818241996, |
| "learning_rate": 5.966767671530078e-07, |
| "loss": 0.5501, |
| "step": 881 |
| }, |
| { |
| "epoch": 2.5764362220058423, |
| "grad_norm": 0.4552021870162526, |
| "learning_rate": 5.88639998916506e-07, |
| "loss": 0.5359, |
| "step": 882 |
| }, |
| { |
| "epoch": 2.57935735150925, |
| "grad_norm": 0.4604883255796744, |
| "learning_rate": 5.806543362721945e-07, |
| "loss": 0.5274, |
| "step": 883 |
| }, |
| { |
| "epoch": 2.5822784810126582, |
| "grad_norm": 0.41454954591229876, |
| "learning_rate": 5.727198717339511e-07, |
| "loss": 0.4765, |
| "step": 884 |
| }, |
| { |
| "epoch": 2.5851996105160664, |
| "grad_norm": 0.4325932298835654, |
| "learning_rate": 5.648366972225222e-07, |
| "loss": 0.5282, |
| "step": 885 |
| }, |
| { |
| "epoch": 2.588120740019474, |
| "grad_norm": 0.433038404211758, |
| "learning_rate": 5.570049040644609e-07, |
| "loss": 0.5137, |
| "step": 886 |
| }, |
| { |
| "epoch": 2.5910418695228823, |
| "grad_norm": 0.45496347053471886, |
| "learning_rate": 5.492245829910664e-07, |
| "loss": 0.4973, |
| "step": 887 |
| }, |
| { |
| "epoch": 2.59396299902629, |
| "grad_norm": 0.44790019595001196, |
| "learning_rate": 5.414958241373358e-07, |
| "loss": 0.4881, |
| "step": 888 |
| }, |
| { |
| "epoch": 2.596884128529698, |
| "grad_norm": 0.4506105888324243, |
| "learning_rate": 5.33818717040917e-07, |
| "loss": 0.5446, |
| "step": 889 |
| }, |
| { |
| "epoch": 2.599805258033106, |
| "grad_norm": 0.45920237343152387, |
| "learning_rate": 5.261933506410722e-07, |
| "loss": 0.5544, |
| "step": 890 |
| }, |
| { |
| "epoch": 2.602726387536514, |
| "grad_norm": 0.433968288589714, |
| "learning_rate": 5.186198132776459e-07, |
| "loss": 0.4792, |
| "step": 891 |
| }, |
| { |
| "epoch": 2.6056475170399223, |
| "grad_norm": 0.45054587006233904, |
| "learning_rate": 5.110981926900488e-07, |
| "loss": 0.5271, |
| "step": 892 |
| }, |
| { |
| "epoch": 2.60856864654333, |
| "grad_norm": 0.4402130423120147, |
| "learning_rate": 5.036285760162307e-07, |
| "loss": 0.4772, |
| "step": 893 |
| }, |
| { |
| "epoch": 2.611489776046738, |
| "grad_norm": 0.4560401508104557, |
| "learning_rate": 4.962110497916789e-07, |
| "loss": 0.5078, |
| "step": 894 |
| }, |
| { |
| "epoch": 2.6144109055501463, |
| "grad_norm": 0.4574934740247989, |
| "learning_rate": 4.888456999484098e-07, |
| "loss": 0.5177, |
| "step": 895 |
| }, |
| { |
| "epoch": 2.617332035053554, |
| "grad_norm": 0.42797938213430475, |
| "learning_rate": 4.815326118139813e-07, |
| "loss": 0.5068, |
| "step": 896 |
| }, |
| { |
| "epoch": 2.620253164556962, |
| "grad_norm": 0.4508198643702282, |
| "learning_rate": 4.742718701104965e-07, |
| "loss": 0.4836, |
| "step": 897 |
| }, |
| { |
| "epoch": 2.62317429406037, |
| "grad_norm": 0.4936637557435792, |
| "learning_rate": 4.670635589536254e-07, |
| "loss": 0.5688, |
| "step": 898 |
| }, |
| { |
| "epoch": 2.626095423563778, |
| "grad_norm": 0.4446810251135343, |
| "learning_rate": 4.599077618516312e-07, |
| "loss": 0.5213, |
| "step": 899 |
| }, |
| { |
| "epoch": 2.629016553067186, |
| "grad_norm": 0.4393121918571539, |
| "learning_rate": 4.528045617044019e-07, |
| "loss": 0.515, |
| "step": 900 |
| }, |
| { |
| "epoch": 2.631937682570594, |
| "grad_norm": 0.4546453714565344, |
| "learning_rate": 4.457540408024896e-07, |
| "loss": 0.5119, |
| "step": 901 |
| }, |
| { |
| "epoch": 2.634858812074002, |
| "grad_norm": 0.43919005512991516, |
| "learning_rate": 4.387562808261575e-07, |
| "loss": 0.5273, |
| "step": 902 |
| }, |
| { |
| "epoch": 2.63777994157741, |
| "grad_norm": 0.4491615831317555, |
| "learning_rate": 4.3181136284443204e-07, |
| "loss": 0.4994, |
| "step": 903 |
| }, |
| { |
| "epoch": 2.6407010710808176, |
| "grad_norm": 0.48434931343594284, |
| "learning_rate": 4.249193673141694e-07, |
| "loss": 0.5186, |
| "step": 904 |
| }, |
| { |
| "epoch": 2.643622200584226, |
| "grad_norm": 0.41259506287384623, |
| "learning_rate": 4.180803740791156e-07, |
| "loss": 0.5319, |
| "step": 905 |
| }, |
| { |
| "epoch": 2.646543330087634, |
| "grad_norm": 0.4076433170920647, |
| "learning_rate": 4.112944623689863e-07, |
| "loss": 0.5178, |
| "step": 906 |
| }, |
| { |
| "epoch": 2.6494644595910417, |
| "grad_norm": 0.421345630456112, |
| "learning_rate": 4.0456171079854833e-07, |
| "loss": 0.4951, |
| "step": 907 |
| }, |
| { |
| "epoch": 2.65238558909445, |
| "grad_norm": 0.4435092242276153, |
| "learning_rate": 3.978821973667074e-07, |
| "loss": 0.4837, |
| "step": 908 |
| }, |
| { |
| "epoch": 2.655306718597858, |
| "grad_norm": 0.44521873058867134, |
| "learning_rate": 3.9125599945560866e-07, |
| "loss": 0.5526, |
| "step": 909 |
| }, |
| { |
| "epoch": 2.6582278481012658, |
| "grad_norm": 0.45505539256105115, |
| "learning_rate": 3.846831938297324e-07, |
| "loss": 0.5158, |
| "step": 910 |
| }, |
| { |
| "epoch": 2.661148977604674, |
| "grad_norm": 0.433074017412774, |
| "learning_rate": 3.7816385663501097e-07, |
| "loss": 0.4809, |
| "step": 911 |
| }, |
| { |
| "epoch": 2.6640701071080817, |
| "grad_norm": 0.4502224951078667, |
| "learning_rate": 3.716980633979489e-07, |
| "loss": 0.5363, |
| "step": 912 |
| }, |
| { |
| "epoch": 2.66699123661149, |
| "grad_norm": 0.43979723124115566, |
| "learning_rate": 3.6528588902473905e-07, |
| "loss": 0.508, |
| "step": 913 |
| }, |
| { |
| "epoch": 2.6699123661148976, |
| "grad_norm": 0.46901546979749387, |
| "learning_rate": 3.589274078004029e-07, |
| "loss": 0.5489, |
| "step": 914 |
| }, |
| { |
| "epoch": 2.6728334956183057, |
| "grad_norm": 0.4432062648822634, |
| "learning_rate": 3.5262269338792623e-07, |
| "loss": 0.526, |
| "step": 915 |
| }, |
| { |
| "epoch": 2.675754625121714, |
| "grad_norm": 0.43511333463780544, |
| "learning_rate": 3.463718188274046e-07, |
| "loss": 0.5332, |
| "step": 916 |
| }, |
| { |
| "epoch": 2.6786757546251216, |
| "grad_norm": 0.38843912794513447, |
| "learning_rate": 3.401748565352031e-07, |
| "loss": 0.4997, |
| "step": 917 |
| }, |
| { |
| "epoch": 2.68159688412853, |
| "grad_norm": 0.41986039406023057, |
| "learning_rate": 3.340318783031099e-07, |
| "loss": 0.4804, |
| "step": 918 |
| }, |
| { |
| "epoch": 2.6845180136319375, |
| "grad_norm": 0.44648799987073357, |
| "learning_rate": 3.279429552975094e-07, |
| "loss": 0.57, |
| "step": 919 |
| }, |
| { |
| "epoch": 2.6874391431353457, |
| "grad_norm": 0.38625226404630086, |
| "learning_rate": 3.219081580585548e-07, |
| "loss": 0.418, |
| "step": 920 |
| }, |
| { |
| "epoch": 2.6903602726387534, |
| "grad_norm": 0.46495901890128216, |
| "learning_rate": 3.159275564993558e-07, |
| "loss": 0.562, |
| "step": 921 |
| }, |
| { |
| "epoch": 2.6932814021421616, |
| "grad_norm": 0.4542095138457356, |
| "learning_rate": 3.100012199051627e-07, |
| "loss": 0.48, |
| "step": 922 |
| }, |
| { |
| "epoch": 2.6962025316455698, |
| "grad_norm": 0.461034837465003, |
| "learning_rate": 3.0412921693256657e-07, |
| "loss": 0.5802, |
| "step": 923 |
| }, |
| { |
| "epoch": 2.6991236611489775, |
| "grad_norm": 0.4124404226906978, |
| "learning_rate": 2.9831161560870346e-07, |
| "loss": 0.5124, |
| "step": 924 |
| }, |
| { |
| "epoch": 2.7020447906523857, |
| "grad_norm": 0.420405602577591, |
| "learning_rate": 2.9254848333046817e-07, |
| "loss": 0.4959, |
| "step": 925 |
| }, |
| { |
| "epoch": 2.704965920155794, |
| "grad_norm": 0.42637463336396936, |
| "learning_rate": 2.8683988686372956e-07, |
| "loss": 0.529, |
| "step": 926 |
| }, |
| { |
| "epoch": 2.7078870496592016, |
| "grad_norm": 0.4156040595949887, |
| "learning_rate": 2.811858923425609e-07, |
| "loss": 0.5184, |
| "step": 927 |
| }, |
| { |
| "epoch": 2.7108081791626093, |
| "grad_norm": 0.4312166608407746, |
| "learning_rate": 2.755865652684703e-07, |
| "loss": 0.5438, |
| "step": 928 |
| }, |
| { |
| "epoch": 2.7137293086660175, |
| "grad_norm": 0.4369289965460609, |
| "learning_rate": 2.7004197050964744e-07, |
| "loss": 0.4978, |
| "step": 929 |
| }, |
| { |
| "epoch": 2.7166504381694256, |
| "grad_norm": 0.426344227648129, |
| "learning_rate": 2.645521723002037e-07, |
| "loss": 0.5041, |
| "step": 930 |
| }, |
| { |
| "epoch": 2.7195715676728334, |
| "grad_norm": 0.4417752376579622, |
| "learning_rate": 2.591172342394349e-07, |
| "loss": 0.5587, |
| "step": 931 |
| }, |
| { |
| "epoch": 2.7224926971762415, |
| "grad_norm": 0.457814912149392, |
| "learning_rate": 2.537372192910825e-07, |
| "loss": 0.5175, |
| "step": 932 |
| }, |
| { |
| "epoch": 2.7254138266796497, |
| "grad_norm": 0.45549699443779185, |
| "learning_rate": 2.4841218978260194e-07, |
| "loss": 0.5033, |
| "step": 933 |
| }, |
| { |
| "epoch": 2.7283349561830574, |
| "grad_norm": 0.44509990954989737, |
| "learning_rate": 2.4314220740444595e-07, |
| "loss": 0.5192, |
| "step": 934 |
| }, |
| { |
| "epoch": 2.731256085686465, |
| "grad_norm": 0.4542595414389855, |
| "learning_rate": 2.3792733320934348e-07, |
| "loss": 0.5193, |
| "step": 935 |
| }, |
| { |
| "epoch": 2.7341772151898733, |
| "grad_norm": 0.46788823553431785, |
| "learning_rate": 2.3276762761159588e-07, |
| "loss": 0.5226, |
| "step": 936 |
| }, |
| { |
| "epoch": 2.7370983446932815, |
| "grad_norm": 0.46548939441743264, |
| "learning_rate": 2.276631503863763e-07, |
| "loss": 0.5234, |
| "step": 937 |
| }, |
| { |
| "epoch": 2.7400194741966892, |
| "grad_norm": 0.43071296003232673, |
| "learning_rate": 2.2261396066903916e-07, |
| "loss": 0.4968, |
| "step": 938 |
| }, |
| { |
| "epoch": 2.7429406037000974, |
| "grad_norm": 0.4422590771192652, |
| "learning_rate": 2.176201169544312e-07, |
| "loss": 0.5135, |
| "step": 939 |
| }, |
| { |
| "epoch": 2.7458617332035056, |
| "grad_norm": 0.43839294000442797, |
| "learning_rate": 2.12681677096217e-07, |
| "loss": 0.5459, |
| "step": 940 |
| }, |
| { |
| "epoch": 2.7487828627069133, |
| "grad_norm": 0.4296191280132258, |
| "learning_rate": 2.0779869830620692e-07, |
| "loss": 0.4538, |
| "step": 941 |
| }, |
| { |
| "epoch": 2.7517039922103215, |
| "grad_norm": 0.4119484440139385, |
| "learning_rate": 2.029712371536957e-07, |
| "loss": 0.4839, |
| "step": 942 |
| }, |
| { |
| "epoch": 2.754625121713729, |
| "grad_norm": 0.4512158763524109, |
| "learning_rate": 1.9819934956480603e-07, |
| "loss": 0.5445, |
| "step": 943 |
| }, |
| { |
| "epoch": 2.7575462512171374, |
| "grad_norm": 0.4212292286246555, |
| "learning_rate": 1.9348309082184102e-07, |
| "loss": 0.4934, |
| "step": 944 |
| }, |
| { |
| "epoch": 2.760467380720545, |
| "grad_norm": 0.44496919909800775, |
| "learning_rate": 1.888225155626433e-07, |
| "loss": 0.5456, |
| "step": 945 |
| }, |
| { |
| "epoch": 2.7633885102239533, |
| "grad_norm": 0.41713723572900646, |
| "learning_rate": 1.8421767777996425e-07, |
| "loss": 0.4462, |
| "step": 946 |
| }, |
| { |
| "epoch": 2.7663096397273614, |
| "grad_norm": 0.4457624171080001, |
| "learning_rate": 1.7966863082083507e-07, |
| "loss": 0.5427, |
| "step": 947 |
| }, |
| { |
| "epoch": 2.769230769230769, |
| "grad_norm": 0.4348007177346548, |
| "learning_rate": 1.7517542738595071e-07, |
| "loss": 0.5568, |
| "step": 948 |
| }, |
| { |
| "epoch": 2.7721518987341773, |
| "grad_norm": 0.41399848976523823, |
| "learning_rate": 1.7073811952905862e-07, |
| "loss": 0.5083, |
| "step": 949 |
| }, |
| { |
| "epoch": 2.775073028237585, |
| "grad_norm": 0.43232436981734923, |
| "learning_rate": 1.6635675865635859e-07, |
| "loss": 0.4505, |
| "step": 950 |
| }, |
| { |
| "epoch": 2.7779941577409932, |
| "grad_norm": 0.4750164143516587, |
| "learning_rate": 1.6203139552590175e-07, |
| "loss": 0.5831, |
| "step": 951 |
| }, |
| { |
| "epoch": 2.780915287244401, |
| "grad_norm": 0.39972505599387415, |
| "learning_rate": 1.5776208024700702e-07, |
| "loss": 0.4751, |
| "step": 952 |
| }, |
| { |
| "epoch": 2.783836416747809, |
| "grad_norm": 0.42708956986618357, |
| "learning_rate": 1.5354886227967934e-07, |
| "loss": 0.532, |
| "step": 953 |
| }, |
| { |
| "epoch": 2.7867575462512173, |
| "grad_norm": 0.4435213719711565, |
| "learning_rate": 1.4939179043403474e-07, |
| "loss": 0.5347, |
| "step": 954 |
| }, |
| { |
| "epoch": 2.789678675754625, |
| "grad_norm": 0.4416374818886015, |
| "learning_rate": 1.4529091286973994e-07, |
| "loss": 0.5459, |
| "step": 955 |
| }, |
| { |
| "epoch": 2.792599805258033, |
| "grad_norm": 0.45153869521194906, |
| "learning_rate": 1.4124627709544814e-07, |
| "loss": 0.4918, |
| "step": 956 |
| }, |
| { |
| "epoch": 2.7955209347614414, |
| "grad_norm": 0.4440135260222283, |
| "learning_rate": 1.3725792996825083e-07, |
| "loss": 0.5016, |
| "step": 957 |
| }, |
| { |
| "epoch": 2.798442064264849, |
| "grad_norm": 0.39484795107770926, |
| "learning_rate": 1.3332591769314006e-07, |
| "loss": 0.5043, |
| "step": 958 |
| }, |
| { |
| "epoch": 2.801363193768257, |
| "grad_norm": 0.444235962409132, |
| "learning_rate": 1.2945028582246445e-07, |
| "loss": 0.5637, |
| "step": 959 |
| }, |
| { |
| "epoch": 2.804284323271665, |
| "grad_norm": 0.4583481416669065, |
| "learning_rate": 1.2563107925540774e-07, |
| "loss": 0.5017, |
| "step": 960 |
| }, |
| { |
| "epoch": 2.807205452775073, |
| "grad_norm": 0.4215520268166574, |
| "learning_rate": 1.2186834223746612e-07, |
| "loss": 0.5136, |
| "step": 961 |
| }, |
| { |
| "epoch": 2.810126582278481, |
| "grad_norm": 0.43820018627048524, |
| "learning_rate": 1.1816211835993684e-07, |
| "loss": 0.4964, |
| "step": 962 |
| }, |
| { |
| "epoch": 2.813047711781889, |
| "grad_norm": 0.4420212484091613, |
| "learning_rate": 1.1451245055941251e-07, |
| "loss": 0.5112, |
| "step": 963 |
| }, |
| { |
| "epoch": 2.8159688412852972, |
| "grad_norm": 0.4365334218610012, |
| "learning_rate": 1.1091938111728374e-07, |
| "loss": 0.5273, |
| "step": 964 |
| }, |
| { |
| "epoch": 2.818889970788705, |
| "grad_norm": 0.4377608647741322, |
| "learning_rate": 1.0738295165924783e-07, |
| "loss": 0.5398, |
| "step": 965 |
| }, |
| { |
| "epoch": 2.8218111002921127, |
| "grad_norm": 0.45046127023142907, |
| "learning_rate": 1.0390320315482982e-07, |
| "loss": 0.5062, |
| "step": 966 |
| }, |
| { |
| "epoch": 2.824732229795521, |
| "grad_norm": 0.4019879301034501, |
| "learning_rate": 1.0048017591690607e-07, |
| "loss": 0.5114, |
| "step": 967 |
| }, |
| { |
| "epoch": 2.827653359298929, |
| "grad_norm": 0.4402917529148519, |
| "learning_rate": 9.711390960123634e-08, |
| "loss": 0.5359, |
| "step": 968 |
| }, |
| { |
| "epoch": 2.8305744888023368, |
| "grad_norm": 0.41904135734811376, |
| "learning_rate": 9.380444320600591e-08, |
| "loss": 0.4684, |
| "step": 969 |
| }, |
| { |
| "epoch": 2.833495618305745, |
| "grad_norm": 0.4659811321879393, |
| "learning_rate": 9.055181507137245e-08, |
| "loss": 0.5426, |
| "step": 970 |
| }, |
| { |
| "epoch": 2.836416747809153, |
| "grad_norm": 0.440634928654449, |
| "learning_rate": 8.73560628790232e-08, |
| "loss": 0.5258, |
| "step": 971 |
| }, |
| { |
| "epoch": 2.839337877312561, |
| "grad_norm": 0.44241019235537177, |
| "learning_rate": 8.421722365173856e-08, |
| "loss": 0.553, |
| "step": 972 |
| }, |
| { |
| "epoch": 2.842259006815969, |
| "grad_norm": 0.4085933163681831, |
| "learning_rate": 8.113533375295968e-08, |
| "loss": 0.4804, |
| "step": 973 |
| }, |
| { |
| "epoch": 2.8451801363193767, |
| "grad_norm": 0.42283609192144683, |
| "learning_rate": 7.81104288863721e-08, |
| "loss": 0.5268, |
| "step": 974 |
| }, |
| { |
| "epoch": 2.848101265822785, |
| "grad_norm": 0.4414452447566111, |
| "learning_rate": 7.514254409549005e-08, |
| "loss": 0.5059, |
| "step": 975 |
| }, |
| { |
| "epoch": 2.8510223953261926, |
| "grad_norm": 0.477876908869347, |
| "learning_rate": 7.223171376325e-08, |
| "loss": 0.5401, |
| "step": 976 |
| }, |
| { |
| "epoch": 2.853943524829601, |
| "grad_norm": 0.45667617596842697, |
| "learning_rate": 6.93779716116122e-08, |
| "loss": 0.5179, |
| "step": 977 |
| }, |
| { |
| "epoch": 2.856864654333009, |
| "grad_norm": 0.43661296400815486, |
| "learning_rate": 6.658135070116978e-08, |
| "loss": 0.479, |
| "step": 978 |
| }, |
| { |
| "epoch": 2.8597857838364167, |
| "grad_norm": 0.43447801900111893, |
| "learning_rate": 6.384188343076914e-08, |
| "loss": 0.4933, |
| "step": 979 |
| }, |
| { |
| "epoch": 2.862706913339825, |
| "grad_norm": 0.4567696696725672, |
| "learning_rate": 6.115960153712963e-08, |
| "loss": 0.5225, |
| "step": 980 |
| }, |
| { |
| "epoch": 2.8656280428432326, |
| "grad_norm": 0.4423237005020546, |
| "learning_rate": 5.8534536094478345e-08, |
| "loss": 0.5649, |
| "step": 981 |
| }, |
| { |
| "epoch": 2.8685491723466408, |
| "grad_norm": 0.41379887497949286, |
| "learning_rate": 5.596671751419147e-08, |
| "loss": 0.474, |
| "step": 982 |
| }, |
| { |
| "epoch": 2.8714703018500485, |
| "grad_norm": 0.4382842717392665, |
| "learning_rate": 5.3456175544437936e-08, |
| "loss": 0.529, |
| "step": 983 |
| }, |
| { |
| "epoch": 2.8743914313534566, |
| "grad_norm": 0.4463860151059245, |
| "learning_rate": 5.100293926984023e-08, |
| "loss": 0.4909, |
| "step": 984 |
| }, |
| { |
| "epoch": 2.877312560856865, |
| "grad_norm": 0.4057084453405752, |
| "learning_rate": 4.860703711113246e-08, |
| "loss": 0.5029, |
| "step": 985 |
| }, |
| { |
| "epoch": 2.8802336903602725, |
| "grad_norm": 0.42518495194251527, |
| "learning_rate": 4.626849682483392e-08, |
| "loss": 0.5376, |
| "step": 986 |
| }, |
| { |
| "epoch": 2.8831548198636807, |
| "grad_norm": 0.45413911287155506, |
| "learning_rate": 4.398734550292716e-08, |
| "loss": 0.5467, |
| "step": 987 |
| }, |
| { |
| "epoch": 2.8860759493670884, |
| "grad_norm": 0.4142515305933239, |
| "learning_rate": 4.176360957254266e-08, |
| "loss": 0.4568, |
| "step": 988 |
| }, |
| { |
| "epoch": 2.8889970788704966, |
| "grad_norm": 0.435060369756503, |
| "learning_rate": 3.959731479565465e-08, |
| "loss": 0.5319, |
| "step": 989 |
| }, |
| { |
| "epoch": 2.8919182083739043, |
| "grad_norm": 0.43929179209673974, |
| "learning_rate": 3.748848626878132e-08, |
| "loss": 0.5782, |
| "step": 990 |
| }, |
| { |
| "epoch": 2.8948393378773125, |
| "grad_norm": 0.4061515904911967, |
| "learning_rate": 3.543714842269508e-08, |
| "loss": 0.462, |
| "step": 991 |
| }, |
| { |
| "epoch": 2.8977604673807207, |
| "grad_norm": 0.46420928088916874, |
| "learning_rate": 3.344332502213887e-08, |
| "loss": 0.5218, |
| "step": 992 |
| }, |
| { |
| "epoch": 2.9006815968841284, |
| "grad_norm": 0.4346617171265348, |
| "learning_rate": 3.150703916554976e-08, |
| "loss": 0.4832, |
| "step": 993 |
| }, |
| { |
| "epoch": 2.9036027263875366, |
| "grad_norm": 0.45371971197536476, |
| "learning_rate": 2.962831328479465e-08, |
| "loss": 0.5087, |
| "step": 994 |
| }, |
| { |
| "epoch": 2.9065238558909448, |
| "grad_norm": 0.47344580171350403, |
| "learning_rate": 2.7807169144906108e-08, |
| "loss": 0.5639, |
| "step": 995 |
| }, |
| { |
| "epoch": 2.9094449853943525, |
| "grad_norm": 0.4343079692320858, |
| "learning_rate": 2.6043627843834185e-08, |
| "loss": 0.534, |
| "step": 996 |
| }, |
| { |
| "epoch": 2.91236611489776, |
| "grad_norm": 0.41472366849725406, |
| "learning_rate": 2.4337709812199405e-08, |
| "loss": 0.471, |
| "step": 997 |
| }, |
| { |
| "epoch": 2.9152872444011684, |
| "grad_norm": 0.41847164257771846, |
| "learning_rate": 2.2689434813056856e-08, |
| "loss": 0.5252, |
| "step": 998 |
| }, |
| { |
| "epoch": 2.9182083739045765, |
| "grad_norm": 0.4540400278681399, |
| "learning_rate": 2.1098821941667457e-08, |
| "loss": 0.5683, |
| "step": 999 |
| }, |
| { |
| "epoch": 2.9211295034079843, |
| "grad_norm": 0.43916840754064074, |
| "learning_rate": 1.9565889625275945e-08, |
| "loss": 0.4844, |
| "step": 1000 |
| }, |
| { |
| "epoch": 2.9240506329113924, |
| "grad_norm": 0.4345074027775515, |
| "learning_rate": 1.8090655622899356e-08, |
| "loss": 0.4881, |
| "step": 1001 |
| }, |
| { |
| "epoch": 2.9269717624148006, |
| "grad_norm": 0.4796907378778938, |
| "learning_rate": 1.6673137025118303e-08, |
| "loss": 0.5576, |
| "step": 1002 |
| }, |
| { |
| "epoch": 2.9298928919182083, |
| "grad_norm": 0.4751897325081864, |
| "learning_rate": 1.5313350253882143e-08, |
| "loss": 0.5248, |
| "step": 1003 |
| }, |
| { |
| "epoch": 2.9328140214216165, |
| "grad_norm": 0.4566958532328187, |
| "learning_rate": 1.4011311062316347e-08, |
| "loss": 0.5207, |
| "step": 1004 |
| }, |
| { |
| "epoch": 2.9357351509250242, |
| "grad_norm": 0.4143592301627743, |
| "learning_rate": 1.2767034534540978e-08, |
| "loss": 0.483, |
| "step": 1005 |
| }, |
| { |
| "epoch": 2.9386562804284324, |
| "grad_norm": 0.4151441658060871, |
| "learning_rate": 1.158053508549528e-08, |
| "loss": 0.4624, |
| "step": 1006 |
| }, |
| { |
| "epoch": 2.94157740993184, |
| "grad_norm": 0.44443161265224573, |
| "learning_rate": 1.0451826460772252e-08, |
| "loss": 0.5295, |
| "step": 1007 |
| }, |
| { |
| "epoch": 2.9444985394352483, |
| "grad_norm": 0.4244569182328679, |
| "learning_rate": 9.38092173645766e-09, |
| "loss": 0.5164, |
| "step": 1008 |
| }, |
| { |
| "epoch": 2.9474196689386565, |
| "grad_norm": 0.42619195275616983, |
| "learning_rate": 8.367833318980168e-09, |
| "loss": 0.5474, |
| "step": 1009 |
| }, |
| { |
| "epoch": 2.950340798442064, |
| "grad_norm": 0.44631991928366016, |
| "learning_rate": 7.412572944965335e-09, |
| "loss": 0.4825, |
| "step": 1010 |
| }, |
| { |
| "epoch": 2.9532619279454724, |
| "grad_norm": 0.4028816094942567, |
| "learning_rate": 6.515151681101284e-09, |
| "loss": 0.507, |
| "step": 1011 |
| }, |
| { |
| "epoch": 2.95618305744888, |
| "grad_norm": 0.4346618256973241, |
| "learning_rate": 5.675579924011576e-09, |
| "loss": 0.5173, |
| "step": 1012 |
| }, |
| { |
| "epoch": 2.9591041869522883, |
| "grad_norm": 0.4213345827232603, |
| "learning_rate": 4.89386740013198e-09, |
| "loss": 0.4984, |
| "step": 1013 |
| }, |
| { |
| "epoch": 2.962025316455696, |
| "grad_norm": 0.40444422520787165, |
| "learning_rate": 4.170023165600001e-09, |
| "loss": 0.5463, |
| "step": 1014 |
| }, |
| { |
| "epoch": 2.964946445959104, |
| "grad_norm": 0.4158643063044851, |
| "learning_rate": 3.5040556061483043e-09, |
| "loss": 0.4929, |
| "step": 1015 |
| }, |
| { |
| "epoch": 2.9678675754625123, |
| "grad_norm": 0.45968510533530904, |
| "learning_rate": 2.8959724370092313e-09, |
| "loss": 0.5618, |
| "step": 1016 |
| }, |
| { |
| "epoch": 2.97078870496592, |
| "grad_norm": 0.45874234552575655, |
| "learning_rate": 2.3457807028248734e-09, |
| "loss": 0.5245, |
| "step": 1017 |
| }, |
| { |
| "epoch": 2.9737098344693282, |
| "grad_norm": 0.4715342361612812, |
| "learning_rate": 1.8534867775649169e-09, |
| "loss": 0.5129, |
| "step": 1018 |
| }, |
| { |
| "epoch": 2.976630963972736, |
| "grad_norm": 0.4184866954459166, |
| "learning_rate": 1.4190963644522549e-09, |
| "loss": 0.4844, |
| "step": 1019 |
| }, |
| { |
| "epoch": 2.979552093476144, |
| "grad_norm": 0.41399082545193333, |
| "learning_rate": 1.0426144958985974e-09, |
| "loss": 0.5588, |
| "step": 1020 |
| }, |
| { |
| "epoch": 2.982473222979552, |
| "grad_norm": 0.445097303921185, |
| "learning_rate": 7.240455334456276e-10, |
| "loss": 0.4614, |
| "step": 1021 |
| }, |
| { |
| "epoch": 2.98539435248296, |
| "grad_norm": 0.4214572233508389, |
| "learning_rate": 4.633931677139325e-10, |
| "loss": 0.5183, |
| "step": 1022 |
| }, |
| { |
| "epoch": 2.988315481986368, |
| "grad_norm": 0.4246075700283406, |
| "learning_rate": 2.60660418360259e-10, |
| "loss": 0.5797, |
| "step": 1023 |
| }, |
| { |
| "epoch": 2.991236611489776, |
| "grad_norm": 0.4332145327952596, |
| "learning_rate": 1.158496340419868e-10, |
| "loss": 0.5132, |
| "step": 1024 |
| }, |
| { |
| "epoch": 2.994157740993184, |
| "grad_norm": 0.4486911892623868, |
| "learning_rate": 2.8962492393258546e-11, |
| "loss": 0.5159, |
| "step": 1025 |
| }, |
| { |
| "epoch": 2.9970788704965923, |
| "grad_norm": 0.4412334672273621, |
| "learning_rate": 0.0, |
| "loss": 0.5245, |
| "step": 1026 |
| }, |
| { |
| "epoch": 2.9970788704965923, |
| "step": 1026, |
| "total_flos": 883131557314560.0, |
| "train_loss": 0.5971221869800523, |
| "train_runtime": 15429.9437, |
| "train_samples_per_second": 6.384, |
| "train_steps_per_second": 0.066 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1026, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 883131557314560.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|