| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.979591836734694, | |
| "eval_steps": 500, | |
| "global_step": 610, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 312.0, | |
| "learning_rate": 3.278688524590164e-06, | |
| "loss": 33.9196, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 354.0, | |
| "learning_rate": 1.6393442622950818e-05, | |
| "loss": 40.6749, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 125.0, | |
| "learning_rate": 3.2786885245901635e-05, | |
| "loss": 27.6429, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 29.875, | |
| "learning_rate": 4.918032786885246e-05, | |
| "loss": 19.7282, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 10.1875, | |
| "learning_rate": 6.557377049180327e-05, | |
| "loss": 16.6712, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 5.53125, | |
| "learning_rate": 8.19672131147541e-05, | |
| "loss": 14.7944, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 7.09375, | |
| "learning_rate": 9.836065573770493e-05, | |
| "loss": 14.0597, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 11.875, | |
| "learning_rate": 0.00011475409836065574, | |
| "loss": 12.5767, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 55.75, | |
| "learning_rate": 0.00013114754098360654, | |
| "loss": 9.9687, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 9.5, | |
| "learning_rate": 0.00014754098360655738, | |
| "loss": 2.8926, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 2.78125, | |
| "learning_rate": 0.0001639344262295082, | |
| "loss": 1.5807, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.609375, | |
| "learning_rate": 0.00018032786885245904, | |
| "loss": 1.3009, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 0.00019672131147540985, | |
| "loss": 1.1905, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 1.4921875, | |
| "learning_rate": 0.00019997380453683512, | |
| "loss": 1.1472, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 3.5, | |
| "learning_rate": 0.00019986740898848306, | |
| "loss": 1.0454, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 1.84375, | |
| "learning_rate": 0.00019967926316561134, | |
| "loss": 0.9724, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 2.421875, | |
| "learning_rate": 0.00019940952108215526, | |
| "loss": 0.9521, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 6.9375, | |
| "learning_rate": 0.00019905840354577972, | |
| "loss": 0.9019, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 0.00019862619797712846, | |
| "loss": 0.8665, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 0.00019811325817454517, | |
| "loss": 0.8629, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 1.046875, | |
| "learning_rate": 0.00019752000402445825, | |
| "loss": 0.7888, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.1875, | |
| "learning_rate": 0.0001968469211576668, | |
| "loss": 0.8141, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 0.00019609456055180897, | |
| "loss": 0.8212, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 6.96875, | |
| "learning_rate": 0.00019526353808033825, | |
| "loss": 0.8073, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 0.00019435453400837684, | |
| "loss": 0.8192, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.2006253004074097, | |
| "eval_runtime": 1.1043, | |
| "eval_samples_per_second": 1.811, | |
| "eval_steps_per_second": 0.906, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 0.62890625, | |
| "learning_rate": 0.00019336829243585856, | |
| "loss": 0.7633, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "grad_norm": 1.3125, | |
| "learning_rate": 0.0001923056206884176, | |
| "loss": 0.7091, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "grad_norm": 0.578125, | |
| "learning_rate": 0.00019116738865652134, | |
| "loss": 0.6773, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "grad_norm": 1.2578125, | |
| "learning_rate": 0.00018995452808338823, | |
| "loss": 0.7058, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "grad_norm": 1.140625, | |
| "learning_rate": 0.00018866803180227402, | |
| "loss": 0.7189, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "grad_norm": 1.71875, | |
| "learning_rate": 0.00018730895292375018, | |
| "loss": 0.6698, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "grad_norm": 1.515625, | |
| "learning_rate": 0.00018587840397364007, | |
| "loss": 0.6941, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "grad_norm": 0.482421875, | |
| "learning_rate": 0.00018437755598231856, | |
| "loss": 0.6511, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "grad_norm": 1.3203125, | |
| "learning_rate": 0.00018280763752612052, | |
| "loss": 0.6672, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "grad_norm": 0.72265625, | |
| "learning_rate": 0.00018116993372164263, | |
| "loss": 0.698, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "grad_norm": 0.7890625, | |
| "learning_rate": 0.0001794657851737625, | |
| "loss": 0.6736, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "grad_norm": 0.5234375, | |
| "learning_rate": 0.00017769658687823525, | |
| "loss": 0.651, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 0.0001758637870797665, | |
| "loss": 0.6809, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "grad_norm": 1.1875, | |
| "learning_rate": 0.0001739688860864967, | |
| "loss": 0.6331, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "grad_norm": 0.8359375, | |
| "learning_rate": 0.00017201343504186643, | |
| "loss": 0.6562, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "grad_norm": 1.5625, | |
| "learning_rate": 0.00016999903465486914, | |
| "loss": 0.6416, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "grad_norm": 1.3359375, | |
| "learning_rate": 0.00016792733388972932, | |
| "loss": 0.6957, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "grad_norm": 0.93359375, | |
| "learning_rate": 0.0001658000286160805, | |
| "loss": 0.6857, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 0.00016361886022074612, | |
| "loss": 0.6589, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 0.62109375, | |
| "learning_rate": 0.0001613856141822612, | |
| "loss": 0.6635, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 0.00015910211860930065, | |
| "loss": 0.6252, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 0.78515625, | |
| "learning_rate": 0.0001567702427442113, | |
| "loss": 0.6187, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "grad_norm": 1.8984375, | |
| "learning_rate": 0.00015439189543287247, | |
| "loss": 0.6194, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "grad_norm": 1.328125, | |
| "learning_rate": 0.00015196902356213714, | |
| "loss": 0.6396, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.87109375, | |
| "learning_rate": 0.0001495036104661337, | |
| "loss": 0.6377, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.1303967237472534, | |
| "eval_runtime": 1.0903, | |
| "eval_samples_per_second": 1.834, | |
| "eval_steps_per_second": 0.917, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "grad_norm": 1.0234375, | |
| "learning_rate": 0.000146997674302732, | |
| "loss": 0.5446, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "grad_norm": 0.76171875, | |
| "learning_rate": 0.0001444532664015037, | |
| "loss": 0.5915, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "grad_norm": 1.28125, | |
| "learning_rate": 0.0001418724695845277, | |
| "loss": 0.5521, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "grad_norm": 1.3203125, | |
| "learning_rate": 0.0001392573964614172, | |
| "loss": 0.5279, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 0.470703125, | |
| "learning_rate": 0.00013661018769996227, | |
| "loss": 0.558, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "grad_norm": 0.82421875, | |
| "learning_rate": 0.00013393301027380476, | |
| "loss": 0.5422, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 0.00013122805568857948, | |
| "loss": 0.5401, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "grad_norm": 0.6171875, | |
| "learning_rate": 0.00012849753818797353, | |
| "loss": 0.5458, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 0.00012574369294117297, | |
| "loss": 0.5309, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "grad_norm": 0.490234375, | |
| "learning_rate": 0.0001229687742131796, | |
| "loss": 0.5276, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "grad_norm": 0.6796875, | |
| "learning_rate": 0.00012017505351949661, | |
| "loss": 0.5322, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "grad_norm": 0.53515625, | |
| "learning_rate": 0.00011736481776669306, | |
| "loss": 0.5242, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.00011454036738036899, | |
| "loss": 0.5104, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "grad_norm": 1.0234375, | |
| "learning_rate": 0.00011170401442205474, | |
| "loss": 0.6306, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "grad_norm": 0.5078125, | |
| "learning_rate": 0.00010885808069658451, | |
| "loss": 0.5246, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "grad_norm": 0.86328125, | |
| "learning_rate": 0.00010600489585149484, | |
| "loss": 0.5307, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 0.0001031467954700025, | |
| "loss": 0.523, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "grad_norm": 0.44921875, | |
| "learning_rate": 0.00010028611915912405, | |
| "loss": 0.5449, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "grad_norm": 0.443359375, | |
| "learning_rate": 9.742520863450115e-05, | |
| "loss": 0.5226, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "grad_norm": 0.466796875, | |
| "learning_rate": 9.456640580350018e-05, | |
| "loss": 0.5173, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "grad_norm": 0.76953125, | |
| "learning_rate": 9.171205084815444e-05, | |
| "loss": 0.5281, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "grad_norm": 0.6796875, | |
| "learning_rate": 8.886448030951912e-05, | |
| "loss": 0.543, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "grad_norm": 0.427734375, | |
| "learning_rate": 8.602602517500651e-05, | |
| "loss": 0.5459, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "grad_norm": 0.5, | |
| "learning_rate": 8.319900897026733e-05, | |
| "loss": 0.5334, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.1456480026245117, | |
| "eval_runtime": 1.0913, | |
| "eval_samples_per_second": 1.833, | |
| "eval_steps_per_second": 0.916, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "grad_norm": 0.494140625, | |
| "learning_rate": 8.038574585718032e-05, | |
| "loss": 0.5025, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "grad_norm": 0.69921875, | |
| "learning_rate": 7.758853873950676e-05, | |
| "loss": 0.4247, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 7.480967737776089e-05, | |
| "loss": 0.4551, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 7.205143651483906e-05, | |
| "loss": 0.4667, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "grad_norm": 0.59375, | |
| "learning_rate": 6.931607401394229e-05, | |
| "loss": 0.4581, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "grad_norm": 0.7421875, | |
| "learning_rate": 6.660582901031621e-05, | |
| "loss": 0.4441, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 6.392292007832168e-05, | |
| "loss": 0.4247, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "grad_norm": 0.66796875, | |
| "learning_rate": 6.126954341533599e-05, | |
| "loss": 0.4214, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "grad_norm": 0.48046875, | |
| "learning_rate": 5.864787104397194e-05, | |
| "loss": 0.4413, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "grad_norm": 0.65234375, | |
| "learning_rate": 5.6060049034085815e-05, | |
| "loss": 0.4648, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 5.3508195746030296e-05, | |
| "loss": 0.5039, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "grad_norm": 0.498046875, | |
| "learning_rate": 5.099440009658957e-05, | |
| "loss": 0.4493, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "grad_norm": 0.43359375, | |
| "learning_rate": 4.852071984901696e-05, | |
| "loss": 0.4343, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 4.6089179928574487e-05, | |
| "loss": 0.4354, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "grad_norm": 0.48046875, | |
| "learning_rate": 4.3701770764953124e-05, | |
| "loss": 0.4694, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "grad_norm": 0.431640625, | |
| "learning_rate": 4.136044666293044e-05, | |
| "loss": 0.4446, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "grad_norm": 0.447265625, | |
| "learning_rate": 3.9067124202600194e-05, | |
| "loss": 0.4398, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "grad_norm": 0.4609375, | |
| "learning_rate": 3.6823680670482485e-05, | |
| "loss": 0.4449, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "grad_norm": 0.4765625, | |
| "learning_rate": 3.463195252279939e-05, | |
| "loss": 0.4443, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "grad_norm": 0.423828125, | |
| "learning_rate": 3.24937338821738e-05, | |
| "loss": 0.4346, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 3.0410775068981613e-05, | |
| "loss": 0.4423, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "grad_norm": 0.828125, | |
| "learning_rate": 2.8384781168560693e-05, | |
| "loss": 0.4304, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "grad_norm": 0.490234375, | |
| "learning_rate": 2.6417410635448013e-05, | |
| "loss": 0.443, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "grad_norm": 0.46875, | |
| "learning_rate": 2.451027393578863e-05, | |
| "loss": 0.4392, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 0.5, | |
| "learning_rate": 2.2664932229027024e-05, | |
| "loss": 0.4454, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.1934518814086914, | |
| "eval_runtime": 1.0901, | |
| "eval_samples_per_second": 1.835, | |
| "eval_steps_per_second": 0.917, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "grad_norm": 0.45703125, | |
| "learning_rate": 2.088289608996071e-05, | |
| "loss": 0.3925, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "grad_norm": 0.4921875, | |
| "learning_rate": 1.9165624272201567e-05, | |
| "loss": 0.403, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "grad_norm": 0.5, | |
| "learning_rate": 1.7514522514057553e-05, | |
| "loss": 0.3824, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "grad_norm": 0.462890625, | |
| "learning_rate": 1.593094238781191e-05, | |
| "loss": 0.3818, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 1.4416180193342144e-05, | |
| "loss": 0.4017, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "grad_norm": 0.4609375, | |
| "learning_rate": 1.2971475896984475e-05, | |
| "loss": 0.4047, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "grad_norm": 0.466796875, | |
| "learning_rate": 1.1598012116512092e-05, | |
| "loss": 0.3735, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 1.0296913153058163e-05, | |
| "loss": 0.3954, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "grad_norm": 0.49609375, | |
| "learning_rate": 9.069244070776428e-06, | |
| "loss": 0.3955, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "grad_norm": 0.5, | |
| "learning_rate": 7.916009824992254e-06, | |
| "loss": 0.3824, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 6.838154439558253e-06, | |
| "loss": 0.3846, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 5.836560234087418e-06, | |
| "loss": 0.4087, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 4.912047101696848e-06, | |
| "loss": 0.3785, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 4.065371837853016e-06, | |
| "loss": 0.4013, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 3.2972275208679625e-06, | |
| "loss": 0.4005, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "grad_norm": 0.498046875, | |
| "learning_rate": 2.608242944553607e-06, | |
| "loss": 0.3842, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "grad_norm": 0.470703125, | |
| "learning_rate": 1.9989821034986032e-06, | |
| "loss": 0.3791, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "grad_norm": 0.494140625, | |
| "learning_rate": 1.4699437313891007e-06, | |
| "loss": 0.398, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "grad_norm": 0.482421875, | |
| "learning_rate": 1.02156089275125e-06, | |
| "loss": 0.3836, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "grad_norm": 0.55078125, | |
| "learning_rate": 6.542006284497304e-07, | |
| "loss": 0.3827, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 3.681636552324452e-07, | |
| "loss": 0.3846, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "grad_norm": 0.4609375, | |
| "learning_rate": 1.6368411956741991e-07, | |
| "loss": 0.4347, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 4.092940597322237e-08, | |
| "loss": 0.3899, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "grad_norm": 0.458984375, | |
| "learning_rate": 0.0, | |
| "loss": 0.408, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "eval_loss": 1.2246639728546143, | |
| "eval_runtime": 1.1185, | |
| "eval_samples_per_second": 1.788, | |
| "eval_steps_per_second": 0.894, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "step": 610, | |
| "total_flos": 4.678055725535068e+17, | |
| "train_loss": 1.826815680011374, | |
| "train_runtime": 4235.156, | |
| "train_samples_per_second": 1.156, | |
| "train_steps_per_second": 0.144 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 610, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 100, | |
| "total_flos": 4.678055725535068e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |