| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 1004, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.00099601593625498, |
| "grad_norm": 0.4463295638561249, |
| "learning_rate": 9.900990099009901e-08, |
| "loss": 2.6197, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.00199203187250996, |
| "grad_norm": 0.8444207906723022, |
| "learning_rate": 1.9801980198019803e-07, |
| "loss": 2.7588, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.00298804780876494, |
| "grad_norm": 0.6591606140136719, |
| "learning_rate": 2.9702970297029703e-07, |
| "loss": 2.9263, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.00398406374501992, |
| "grad_norm": 0.5656299591064453, |
| "learning_rate": 3.9603960396039606e-07, |
| "loss": 2.4296, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0049800796812749, |
| "grad_norm": 0.5051721930503845, |
| "learning_rate": 4.950495049504951e-07, |
| "loss": 2.3939, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.00597609561752988, |
| "grad_norm": 0.7762399911880493, |
| "learning_rate": 5.940594059405941e-07, |
| "loss": 2.6638, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.0069721115537848604, |
| "grad_norm": 0.5301679968833923, |
| "learning_rate": 6.930693069306931e-07, |
| "loss": 2.2199, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.00796812749003984, |
| "grad_norm": 0.6617525219917297, |
| "learning_rate": 7.920792079207921e-07, |
| "loss": 2.8019, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.008964143426294821, |
| "grad_norm": 0.7944237589836121, |
| "learning_rate": 8.910891089108911e-07, |
| "loss": 2.7146, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.0099601593625498, |
| "grad_norm": 0.6918312907218933, |
| "learning_rate": 9.900990099009902e-07, |
| "loss": 2.6128, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.010956175298804782, |
| "grad_norm": 0.550072193145752, |
| "learning_rate": 1.0891089108910893e-06, |
| "loss": 2.5252, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.01195219123505976, |
| "grad_norm": 0.4420550763607025, |
| "learning_rate": 1.1881188118811881e-06, |
| "loss": 2.6964, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.012948207171314742, |
| "grad_norm": 0.4683515429496765, |
| "learning_rate": 1.2871287128712872e-06, |
| "loss": 2.6433, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.013944223107569721, |
| "grad_norm": 0.5689812898635864, |
| "learning_rate": 1.3861386138613863e-06, |
| "loss": 2.3309, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.014940239043824702, |
| "grad_norm": 0.5711223483085632, |
| "learning_rate": 1.4851485148514852e-06, |
| "loss": 2.4396, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.01593625498007968, |
| "grad_norm": 0.4562544822692871, |
| "learning_rate": 1.5841584158415842e-06, |
| "loss": 2.234, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.01693227091633466, |
| "grad_norm": 0.33882570266723633, |
| "learning_rate": 1.6831683168316833e-06, |
| "loss": 2.5468, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.017928286852589643, |
| "grad_norm": 0.46446338295936584, |
| "learning_rate": 1.7821782178217822e-06, |
| "loss": 2.6143, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.018924302788844622, |
| "grad_norm": 0.625619649887085, |
| "learning_rate": 1.8811881188118813e-06, |
| "loss": 2.5565, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.0199203187250996, |
| "grad_norm": 0.5139931440353394, |
| "learning_rate": 1.9801980198019803e-06, |
| "loss": 2.8371, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.02091633466135458, |
| "grad_norm": 0.45826011896133423, |
| "learning_rate": 2.0792079207920794e-06, |
| "loss": 2.5883, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.021912350597609563, |
| "grad_norm": 0.5945838093757629, |
| "learning_rate": 2.1782178217821785e-06, |
| "loss": 2.6956, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.022908366533864542, |
| "grad_norm": 0.6705940961837769, |
| "learning_rate": 2.2772277227722776e-06, |
| "loss": 3.1889, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.02390438247011952, |
| "grad_norm": 0.489014208316803, |
| "learning_rate": 2.3762376237623762e-06, |
| "loss": 2.646, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.0249003984063745, |
| "grad_norm": 0.5540168285369873, |
| "learning_rate": 2.4752475247524753e-06, |
| "loss": 2.6909, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.025896414342629483, |
| "grad_norm": 0.34993091225624084, |
| "learning_rate": 2.5742574257425744e-06, |
| "loss": 2.5703, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.026892430278884463, |
| "grad_norm": 0.6204649209976196, |
| "learning_rate": 2.6732673267326735e-06, |
| "loss": 2.6973, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.027888446215139442, |
| "grad_norm": 0.6948006749153137, |
| "learning_rate": 2.7722772277227726e-06, |
| "loss": 2.8826, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.02888446215139442, |
| "grad_norm": 0.4865665137767792, |
| "learning_rate": 2.8712871287128712e-06, |
| "loss": 2.6566, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.029880478087649404, |
| "grad_norm": 0.7654755711555481, |
| "learning_rate": 2.9702970297029703e-06, |
| "loss": 2.9627, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.030876494023904383, |
| "grad_norm": 0.636715829372406, |
| "learning_rate": 3.0693069306930694e-06, |
| "loss": 2.3846, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.03187250996015936, |
| "grad_norm": 0.3698335289955139, |
| "learning_rate": 3.1683168316831685e-06, |
| "loss": 2.6724, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.03286852589641434, |
| "grad_norm": 0.7592146396636963, |
| "learning_rate": 3.2673267326732676e-06, |
| "loss": 2.9239, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.03386454183266932, |
| "grad_norm": 0.7194887399673462, |
| "learning_rate": 3.3663366336633666e-06, |
| "loss": 2.4727, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.0348605577689243, |
| "grad_norm": 0.4150688350200653, |
| "learning_rate": 3.4653465346534653e-06, |
| "loss": 2.6407, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.035856573705179286, |
| "grad_norm": 0.4450097382068634, |
| "learning_rate": 3.5643564356435644e-06, |
| "loss": 2.6709, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.036852589641434265, |
| "grad_norm": 0.4754781424999237, |
| "learning_rate": 3.6633663366336635e-06, |
| "loss": 2.6848, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.037848605577689244, |
| "grad_norm": 0.41208407282829285, |
| "learning_rate": 3.7623762376237625e-06, |
| "loss": 2.5176, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.03884462151394422, |
| "grad_norm": 0.41328731179237366, |
| "learning_rate": 3.861386138613862e-06, |
| "loss": 2.3275, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.0398406374501992, |
| "grad_norm": 0.5368106365203857, |
| "learning_rate": 3.960396039603961e-06, |
| "loss": 2.4877, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.04083665338645418, |
| "grad_norm": 0.37100547552108765, |
| "learning_rate": 4.05940594059406e-06, |
| "loss": 2.5933, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.04183266932270916, |
| "grad_norm": 0.4816776216030121, |
| "learning_rate": 4.158415841584159e-06, |
| "loss": 2.844, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.04282868525896414, |
| "grad_norm": 0.4209342896938324, |
| "learning_rate": 4.2574257425742575e-06, |
| "loss": 2.5422, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.043824701195219126, |
| "grad_norm": 0.6714078783988953, |
| "learning_rate": 4.356435643564357e-06, |
| "loss": 2.7081, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.044820717131474105, |
| "grad_norm": 0.26568883657455444, |
| "learning_rate": 4.455445544554456e-06, |
| "loss": 2.9217, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.045816733067729085, |
| "grad_norm": 0.37946802377700806, |
| "learning_rate": 4.554455445544555e-06, |
| "loss": 2.4118, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.046812749003984064, |
| "grad_norm": 0.6484291553497314, |
| "learning_rate": 4.653465346534654e-06, |
| "loss": 2.547, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.04780876494023904, |
| "grad_norm": 0.29198533296585083, |
| "learning_rate": 4.7524752475247525e-06, |
| "loss": 2.519, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.04880478087649402, |
| "grad_norm": 0.7583147883415222, |
| "learning_rate": 4.851485148514852e-06, |
| "loss": 2.9767, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.049800796812749, |
| "grad_norm": 0.5130609273910522, |
| "learning_rate": 4.950495049504951e-06, |
| "loss": 2.7004, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.05079681274900399, |
| "grad_norm": 0.31620916724205017, |
| "learning_rate": 5.04950495049505e-06, |
| "loss": 2.4462, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.05179282868525897, |
| "grad_norm": 0.8395189046859741, |
| "learning_rate": 5.148514851485149e-06, |
| "loss": 2.5588, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.052788844621513946, |
| "grad_norm": 0.387138307094574, |
| "learning_rate": 5.247524752475248e-06, |
| "loss": 2.1448, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.053784860557768925, |
| "grad_norm": 1.2158163785934448, |
| "learning_rate": 5.346534653465347e-06, |
| "loss": 2.3136, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.054780876494023904, |
| "grad_norm": 0.2387009710073471, |
| "learning_rate": 5.4455445544554465e-06, |
| "loss": 2.1985, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.055776892430278883, |
| "grad_norm": 0.3181290924549103, |
| "learning_rate": 5.544554455445545e-06, |
| "loss": 2.3419, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.05677290836653386, |
| "grad_norm": 0.37027299404144287, |
| "learning_rate": 5.643564356435644e-06, |
| "loss": 2.6245, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.05776892430278884, |
| "grad_norm": 1.2025309801101685, |
| "learning_rate": 5.7425742574257425e-06, |
| "loss": 3.9658, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.05876494023904383, |
| "grad_norm": 0.2420024573802948, |
| "learning_rate": 5.841584158415842e-06, |
| "loss": 2.6491, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.05976095617529881, |
| "grad_norm": 0.5019764304161072, |
| "learning_rate": 5.940594059405941e-06, |
| "loss": 2.5349, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.060756972111553786, |
| "grad_norm": 0.42150792479515076, |
| "learning_rate": 6.03960396039604e-06, |
| "loss": 3.0192, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.061752988047808766, |
| "grad_norm": 0.21461670100688934, |
| "learning_rate": 6.138613861386139e-06, |
| "loss": 2.4131, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.06274900398406374, |
| "grad_norm": 0.27715393900871277, |
| "learning_rate": 6.237623762376238e-06, |
| "loss": 2.0699, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.06374501992031872, |
| "grad_norm": 0.23175019025802612, |
| "learning_rate": 6.336633663366337e-06, |
| "loss": 2.454, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.0647410358565737, |
| "grad_norm": 1.4909917116165161, |
| "learning_rate": 6.4356435643564364e-06, |
| "loss": 2.394, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.06573705179282868, |
| "grad_norm": 0.25836002826690674, |
| "learning_rate": 6.534653465346535e-06, |
| "loss": 2.1991, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.06673306772908366, |
| "grad_norm": 0.24367666244506836, |
| "learning_rate": 6.633663366336635e-06, |
| "loss": 2.1038, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.06772908366533864, |
| "grad_norm": 0.2366018295288086, |
| "learning_rate": 6.732673267326733e-06, |
| "loss": 2.4102, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.06872509960159362, |
| "grad_norm": 0.2741665542125702, |
| "learning_rate": 6.831683168316833e-06, |
| "loss": 2.4163, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.0697211155378486, |
| "grad_norm": 1.1350017786026, |
| "learning_rate": 6.930693069306931e-06, |
| "loss": 3.1909, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.07071713147410359, |
| "grad_norm": 0.7466657161712646, |
| "learning_rate": 7.02970297029703e-06, |
| "loss": 3.0505, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.07171314741035857, |
| "grad_norm": 0.6016573309898376, |
| "learning_rate": 7.128712871287129e-06, |
| "loss": 2.523, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.07270916334661355, |
| "grad_norm": 0.29950985312461853, |
| "learning_rate": 7.227722772277228e-06, |
| "loss": 2.2338, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.07370517928286853, |
| "grad_norm": 0.3367365598678589, |
| "learning_rate": 7.326732673267327e-06, |
| "loss": 2.3668, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.07470119521912351, |
| "grad_norm": 0.23957543075084686, |
| "learning_rate": 7.425742574257426e-06, |
| "loss": 2.3929, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.07569721115537849, |
| "grad_norm": 0.2996574342250824, |
| "learning_rate": 7.524752475247525e-06, |
| "loss": 2.4929, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.07669322709163347, |
| "grad_norm": 1.1166422367095947, |
| "learning_rate": 7.6237623762376246e-06, |
| "loss": 2.2256, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.07768924302788845, |
| "grad_norm": 0.3733150362968445, |
| "learning_rate": 7.722772277227724e-06, |
| "loss": 2.4912, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.07868525896414343, |
| "grad_norm": 0.3624296486377716, |
| "learning_rate": 7.821782178217822e-06, |
| "loss": 2.7605, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.0796812749003984, |
| "grad_norm": 0.47846829891204834, |
| "learning_rate": 7.920792079207921e-06, |
| "loss": 2.7636, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.08067729083665338, |
| "grad_norm": 0.3782709836959839, |
| "learning_rate": 8.019801980198021e-06, |
| "loss": 2.4481, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.08167330677290836, |
| "grad_norm": 0.5004844665527344, |
| "learning_rate": 8.11881188118812e-06, |
| "loss": 2.4166, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.08266932270916334, |
| "grad_norm": 0.20830737054347992, |
| "learning_rate": 8.217821782178218e-06, |
| "loss": 2.4728, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.08366533864541832, |
| "grad_norm": 0.2479114830493927, |
| "learning_rate": 8.316831683168318e-06, |
| "loss": 2.2449, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.0846613545816733, |
| "grad_norm": 0.42911332845687866, |
| "learning_rate": 8.415841584158416e-06, |
| "loss": 2.2295, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.08565737051792828, |
| "grad_norm": 0.44220131635665894, |
| "learning_rate": 8.514851485148515e-06, |
| "loss": 2.4495, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.08665338645418327, |
| "grad_norm": 0.23947738111019135, |
| "learning_rate": 8.613861386138615e-06, |
| "loss": 2.1415, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.08764940239043825, |
| "grad_norm": 0.42801541090011597, |
| "learning_rate": 8.712871287128714e-06, |
| "loss": 2.3226, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.08864541832669323, |
| "grad_norm": 0.39098042249679565, |
| "learning_rate": 8.811881188118812e-06, |
| "loss": 2.3063, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.08964143426294821, |
| "grad_norm": 0.29923197627067566, |
| "learning_rate": 8.910891089108911e-06, |
| "loss": 2.6163, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.09063745019920319, |
| "grad_norm": 0.2684191167354584, |
| "learning_rate": 9.009900990099011e-06, |
| "loss": 2.6845, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.09163346613545817, |
| "grad_norm": 0.27097082138061523, |
| "learning_rate": 9.10891089108911e-06, |
| "loss": 2.598, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.09262948207171315, |
| "grad_norm": 0.2647894024848938, |
| "learning_rate": 9.20792079207921e-06, |
| "loss": 2.2659, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.09362549800796813, |
| "grad_norm": 0.38580745458602905, |
| "learning_rate": 9.306930693069308e-06, |
| "loss": 2.5818, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.0946215139442231, |
| "grad_norm": 0.23376502096652985, |
| "learning_rate": 9.405940594059405e-06, |
| "loss": 2.207, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.09561752988047809, |
| "grad_norm": 0.7030872702598572, |
| "learning_rate": 9.504950495049505e-06, |
| "loss": 2.1312, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.09661354581673307, |
| "grad_norm": 0.23579809069633484, |
| "learning_rate": 9.603960396039604e-06, |
| "loss": 2.0685, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.09760956175298804, |
| "grad_norm": 0.3901154100894928, |
| "learning_rate": 9.702970297029704e-06, |
| "loss": 2.662, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.09860557768924302, |
| "grad_norm": 0.2687411606311798, |
| "learning_rate": 9.801980198019802e-06, |
| "loss": 2.9062, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.099601593625498, |
| "grad_norm": 0.1916651427745819, |
| "learning_rate": 9.900990099009901e-06, |
| "loss": 2.1874, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.10059760956175298, |
| "grad_norm": 0.3915342688560486, |
| "learning_rate": 1e-05, |
| "loss": 2.4196, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.10159362549800798, |
| "grad_norm": 0.4256736636161804, |
| "learning_rate": 9.999969740355927e-06, |
| "loss": 2.3229, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.10258964143426295, |
| "grad_norm": 0.24454592168331146, |
| "learning_rate": 9.999878961789962e-06, |
| "loss": 2.4725, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.10358565737051793, |
| "grad_norm": 0.4549209475517273, |
| "learning_rate": 9.999727665400876e-06, |
| "loss": 2.6802, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.10458167330677291, |
| "grad_norm": 0.27968448400497437, |
| "learning_rate": 9.999515853019941e-06, |
| "loss": 2.154, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.10557768924302789, |
| "grad_norm": 0.30781856179237366, |
| "learning_rate": 9.999243527210904e-06, |
| "loss": 2.1358, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.10657370517928287, |
| "grad_norm": 0.22190262377262115, |
| "learning_rate": 9.998910691269957e-06, |
| "loss": 2.1951, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.10756972111553785, |
| "grad_norm": 0.36984801292419434, |
| "learning_rate": 9.998517349225698e-06, |
| "loss": 2.2283, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.10856573705179283, |
| "grad_norm": 0.24837668240070343, |
| "learning_rate": 9.998063505839084e-06, |
| "loss": 2.4556, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.10956175298804781, |
| "grad_norm": 0.17527616024017334, |
| "learning_rate": 9.99754916660337e-06, |
| "loss": 2.2414, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.11055776892430279, |
| "grad_norm": 0.2681446969509125, |
| "learning_rate": 9.996974337744047e-06, |
| "loss": 2.6504, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.11155378486055777, |
| "grad_norm": 0.41997164487838745, |
| "learning_rate": 9.99633902621876e-06, |
| "loss": 2.4704, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.11254980079681275, |
| "grad_norm": 0.43319636583328247, |
| "learning_rate": 9.995643239717228e-06, |
| "loss": 2.4391, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.11354581673306773, |
| "grad_norm": 0.3344462811946869, |
| "learning_rate": 9.994886986661155e-06, |
| "loss": 2.4113, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.1145418326693227, |
| "grad_norm": 0.2086816132068634, |
| "learning_rate": 9.994070276204115e-06, |
| "loss": 2.1469, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.11553784860557768, |
| "grad_norm": 0.35499969124794006, |
| "learning_rate": 9.993193118231463e-06, |
| "loss": 2.6212, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.11653386454183266, |
| "grad_norm": 0.4640713036060333, |
| "learning_rate": 9.992255523360187e-06, |
| "loss": 2.5926, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.11752988047808766, |
| "grad_norm": 0.3302168548107147, |
| "learning_rate": 9.991257502938805e-06, |
| "loss": 2.0769, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.11852589641434264, |
| "grad_norm": 0.4918990135192871, |
| "learning_rate": 9.990199069047216e-06, |
| "loss": 2.4818, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.11952191235059761, |
| "grad_norm": 0.25973260402679443, |
| "learning_rate": 9.989080234496548e-06, |
| "loss": 2.3236, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.1205179282868526, |
| "grad_norm": 0.26280921697616577, |
| "learning_rate": 9.98790101282902e-06, |
| "loss": 2.4276, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.12151394422310757, |
| "grad_norm": 0.21018964052200317, |
| "learning_rate": 9.986661418317759e-06, |
| "loss": 2.3006, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.12250996015936255, |
| "grad_norm": 0.30688363313674927, |
| "learning_rate": 9.985361465966644e-06, |
| "loss": 2.3402, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.12350597609561753, |
| "grad_norm": 0.27469494938850403, |
| "learning_rate": 9.984001171510112e-06, |
| "loss": 2.7987, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.12450199203187251, |
| "grad_norm": 0.2825527489185333, |
| "learning_rate": 9.982580551412972e-06, |
| "loss": 2.4642, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.1254980079681275, |
| "grad_norm": 0.6109227538108826, |
| "learning_rate": 9.98109962287021e-06, |
| "loss": 2.6041, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.12649402390438247, |
| "grad_norm": 0.3983345031738281, |
| "learning_rate": 9.979558403806773e-06, |
| "loss": 2.4599, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.12749003984063745, |
| "grad_norm": 0.3466341495513916, |
| "learning_rate": 9.977956912877356e-06, |
| "loss": 2.1902, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.12848605577689243, |
| "grad_norm": 0.3762282729148865, |
| "learning_rate": 9.97629516946618e-06, |
| "loss": 2.2643, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.1294820717131474, |
| "grad_norm": 0.523991048336029, |
| "learning_rate": 9.974573193686747e-06, |
| "loss": 2.196, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.13047808764940239, |
| "grad_norm": 0.23254041373729706, |
| "learning_rate": 9.97279100638161e-06, |
| "loss": 2.4402, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.13147410358565736, |
| "grad_norm": 0.3137255311012268, |
| "learning_rate": 9.970948629122108e-06, |
| "loss": 2.5905, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.13247011952191234, |
| "grad_norm": 0.21106691658496857, |
| "learning_rate": 9.969046084208116e-06, |
| "loss": 2.3683, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.13346613545816732, |
| "grad_norm": 0.4183836877346039, |
| "learning_rate": 9.967083394667763e-06, |
| "loss": 2.0614, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.1344621513944223, |
| "grad_norm": 0.4468408226966858, |
| "learning_rate": 9.965060584257165e-06, |
| "loss": 2.4639, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.13545816733067728, |
| "grad_norm": 0.22207094728946686, |
| "learning_rate": 9.962977677460132e-06, |
| "loss": 2.2261, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.13645418326693226, |
| "grad_norm": 0.2465856373310089, |
| "learning_rate": 9.960834699487873e-06, |
| "loss": 2.2444, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.13745019920318724, |
| "grad_norm": 0.3648821711540222, |
| "learning_rate": 9.958631676278686e-06, |
| "loss": 2.6132, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.13844621513944222, |
| "grad_norm": 0.21419532597064972, |
| "learning_rate": 9.956368634497648e-06, |
| "loss": 2.4863, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.1394422310756972, |
| "grad_norm": 0.5358874797821045, |
| "learning_rate": 9.9540456015363e-06, |
| "loss": 2.401, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.14043824701195218, |
| "grad_norm": 0.25208160281181335, |
| "learning_rate": 9.951662605512298e-06, |
| "loss": 2.5901, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.14143426294820718, |
| "grad_norm": 0.2659305930137634, |
| "learning_rate": 9.949219675269089e-06, |
| "loss": 2.3229, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.14243027888446216, |
| "grad_norm": 0.5667638778686523, |
| "learning_rate": 9.946716840375552e-06, |
| "loss": 2.4998, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.14342629482071714, |
| "grad_norm": 0.315893292427063, |
| "learning_rate": 9.944154131125643e-06, |
| "loss": 2.4017, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.14442231075697212, |
| "grad_norm": 0.43832895159721375, |
| "learning_rate": 9.941531578538032e-06, |
| "loss": 2.473, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.1454183266932271, |
| "grad_norm": 0.2750052511692047, |
| "learning_rate": 9.938849214355722e-06, |
| "loss": 2.3208, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.14641434262948208, |
| "grad_norm": 0.26100143790245056, |
| "learning_rate": 9.936107071045665e-06, |
| "loss": 2.2151, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.14741035856573706, |
| "grad_norm": 0.3080121576786041, |
| "learning_rate": 9.933305181798374e-06, |
| "loss": 2.0868, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.14840637450199204, |
| "grad_norm": 0.24721984565258026, |
| "learning_rate": 9.93044358052752e-06, |
| "loss": 2.3312, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.14940239043824702, |
| "grad_norm": 0.5916289687156677, |
| "learning_rate": 9.927522301869515e-06, |
| "loss": 2.6542, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.150398406374502, |
| "grad_norm": 0.22790588438510895, |
| "learning_rate": 9.924541381183099e-06, |
| "loss": 2.3121, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.15139442231075698, |
| "grad_norm": 0.3201110064983368, |
| "learning_rate": 9.921500854548916e-06, |
| "loss": 2.8776, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.15239043824701196, |
| "grad_norm": 0.3063032627105713, |
| "learning_rate": 9.918400758769063e-06, |
| "loss": 2.3604, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.15338645418326693, |
| "grad_norm": 0.41891732811927795, |
| "learning_rate": 9.915241131366657e-06, |
| "loss": 2.1036, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.15438247011952191, |
| "grad_norm": 1.0909143686294556, |
| "learning_rate": 9.912022010585385e-06, |
| "loss": 2.5247, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.1553784860557769, |
| "grad_norm": 0.26673582196235657, |
| "learning_rate": 9.90874343538902e-06, |
| "loss": 2.159, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.15637450199203187, |
| "grad_norm": 0.3614170551300049, |
| "learning_rate": 9.905405445460972e-06, |
| "loss": 2.4383, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.15737051792828685, |
| "grad_norm": 0.5078898668289185, |
| "learning_rate": 9.902008081203796e-06, |
| "loss": 2.2543, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.15836653386454183, |
| "grad_norm": 0.3936934769153595, |
| "learning_rate": 9.898551383738707e-06, |
| "loss": 2.538, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.1593625498007968, |
| "grad_norm": 0.6516975164413452, |
| "learning_rate": 9.895035394905073e-06, |
| "loss": 2.5301, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.1603585657370518, |
| "grad_norm": 0.26518794894218445, |
| "learning_rate": 9.89146015725993e-06, |
| "loss": 2.2118, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.16135458167330677, |
| "grad_norm": 0.41366127133369446, |
| "learning_rate": 9.887825714077439e-06, |
| "loss": 2.4799, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.16235059760956175, |
| "grad_norm": 0.5400006771087646, |
| "learning_rate": 9.884132109348386e-06, |
| "loss": 2.5019, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.16334661354581673, |
| "grad_norm": 0.36508408188819885, |
| "learning_rate": 9.880379387779637e-06, |
| "loss": 2.7938, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.1643426294820717, |
| "grad_norm": 0.4240388572216034, |
| "learning_rate": 9.876567594793597e-06, |
| "loss": 2.5002, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.16533864541832669, |
| "grad_norm": 0.277864933013916, |
| "learning_rate": 9.87269677652767e-06, |
| "loss": 2.5436, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.16633466135458166, |
| "grad_norm": 0.25840163230895996, |
| "learning_rate": 9.868766979833686e-06, |
| "loss": 2.2811, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.16733067729083664, |
| "grad_norm": 0.32198214530944824, |
| "learning_rate": 9.864778252277344e-06, |
| "loss": 2.3215, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.16832669322709162, |
| "grad_norm": 0.613046407699585, |
| "learning_rate": 9.86073064213764e-06, |
| "loss": 2.8133, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.1693227091633466, |
| "grad_norm": 0.38470038771629333, |
| "learning_rate": 9.856624198406262e-06, |
| "loss": 2.4133, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.17031872509960158, |
| "grad_norm": 0.3675747811794281, |
| "learning_rate": 9.852458970787027e-06, |
| "loss": 2.0617, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.17131474103585656, |
| "grad_norm": 0.26074767112731934, |
| "learning_rate": 9.848235009695255e-06, |
| "loss": 2.132, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.17231075697211157, |
| "grad_norm": 0.5974801778793335, |
| "learning_rate": 9.84395236625717e-06, |
| "loss": 2.3888, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.17330677290836655, |
| "grad_norm": 0.2652048170566559, |
| "learning_rate": 9.839611092309278e-06, |
| "loss": 2.4468, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.17430278884462153, |
| "grad_norm": 0.6336271166801453, |
| "learning_rate": 9.835211240397743e-06, |
| "loss": 2.3256, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.1752988047808765, |
| "grad_norm": 0.3853505849838257, |
| "learning_rate": 9.830752863777741e-06, |
| "loss": 2.5527, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.17629482071713148, |
| "grad_norm": 0.25374558568000793, |
| "learning_rate": 9.826236016412833e-06, |
| "loss": 2.5593, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.17729083665338646, |
| "grad_norm": 0.29101264476776123, |
| "learning_rate": 9.821660752974294e-06, |
| "loss": 2.6399, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.17828685258964144, |
| "grad_norm": 0.7464101910591125, |
| "learning_rate": 9.817027128840462e-06, |
| "loss": 2.1674, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.17928286852589642, |
| "grad_norm": 0.28557366132736206, |
| "learning_rate": 9.812335200096064e-06, |
| "loss": 2.7127, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.1802788844621514, |
| "grad_norm": 0.5655897259712219, |
| "learning_rate": 9.807585023531536e-06, |
| "loss": 2.0397, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.18127490039840638, |
| "grad_norm": 0.2831386625766754, |
| "learning_rate": 9.802776656642341e-06, |
| "loss": 2.1947, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.18227091633466136, |
| "grad_norm": 0.30917420983314514, |
| "learning_rate": 9.797910157628265e-06, |
| "loss": 2.3951, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.18326693227091634, |
| "grad_norm": 0.3886703550815582, |
| "learning_rate": 9.792985585392722e-06, |
| "loss": 2.4107, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.18426294820717132, |
| "grad_norm": 0.51981121301651, |
| "learning_rate": 9.78800299954203e-06, |
| "loss": 2.3655, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.1852589641434263, |
| "grad_norm": 0.31090375781059265, |
| "learning_rate": 9.782962460384701e-06, |
| "loss": 1.984, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.18625498007968128, |
| "grad_norm": 0.4561314880847931, |
| "learning_rate": 9.777864028930705e-06, |
| "loss": 3.0161, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.18725099601593626, |
| "grad_norm": 0.3265978693962097, |
| "learning_rate": 9.772707766890726e-06, |
| "loss": 2.6738, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.18824701195219123, |
| "grad_norm": 0.5627899765968323, |
| "learning_rate": 9.767493736675429e-06, |
| "loss": 2.4544, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.1892430278884462, |
| "grad_norm": 0.3551636338233948, |
| "learning_rate": 9.762222001394692e-06, |
| "loss": 2.4696, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.1902390438247012, |
| "grad_norm": 0.27445298433303833, |
| "learning_rate": 9.756892624856848e-06, |
| "loss": 2.5626, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.19123505976095617, |
| "grad_norm": 0.4346907436847687, |
| "learning_rate": 9.751505671567914e-06, |
| "loss": 2.6588, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.19223107569721115, |
| "grad_norm": 1.7177170515060425, |
| "learning_rate": 9.746061206730801e-06, |
| "loss": 3.3538, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.19322709163346613, |
| "grad_norm": 0.294007807970047, |
| "learning_rate": 9.740559296244543e-06, |
| "loss": 2.7963, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.1942231075697211, |
| "grad_norm": 0.3322044014930725, |
| "learning_rate": 9.735000006703475e-06, |
| "loss": 2.1763, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.1952191235059761, |
| "grad_norm": 0.2852723002433777, |
| "learning_rate": 9.72938340539645e-06, |
| "loss": 2.2182, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.19621513944223107, |
| "grad_norm": 0.2600834369659424, |
| "learning_rate": 9.723709560306009e-06, |
| "loss": 2.7632, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.19721115537848605, |
| "grad_norm": 0.27677562832832336, |
| "learning_rate": 9.717978540107566e-06, |
| "loss": 2.3831, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.19820717131474103, |
| "grad_norm": 0.4312080144882202, |
| "learning_rate": 9.712190414168573e-06, |
| "loss": 2.4096, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.199203187250996, |
| "grad_norm": 0.7516922950744629, |
| "learning_rate": 9.706345252547681e-06, |
| "loss": 3.0072, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.20019920318725098, |
| "grad_norm": 0.47257497906684875, |
| "learning_rate": 9.700443125993897e-06, |
| "loss": 2.4537, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.20119521912350596, |
| "grad_norm": 0.5163850784301758, |
| "learning_rate": 9.694484105945719e-06, |
| "loss": 2.4488, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.20219123505976094, |
| "grad_norm": 0.2632780373096466, |
| "learning_rate": 9.688468264530278e-06, |
| "loss": 2.5477, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.20318725099601595, |
| "grad_norm": 1.0932762622833252, |
| "learning_rate": 9.682395674562459e-06, |
| "loss": 2.8381, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.20418326693227093, |
| "grad_norm": 0.568217396736145, |
| "learning_rate": 9.676266409544031e-06, |
| "loss": 2.2398, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.2051792828685259, |
| "grad_norm": 0.5864899754524231, |
| "learning_rate": 9.670080543662742e-06, |
| "loss": 2.5067, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.2061752988047809, |
| "grad_norm": 0.38742220401763916, |
| "learning_rate": 9.663838151791431e-06, |
| "loss": 2.3831, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.20717131474103587, |
| "grad_norm": 0.441034197807312, |
| "learning_rate": 9.657539309487123e-06, |
| "loss": 2.3785, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.20816733067729085, |
| "grad_norm": 1.6438182592391968, |
| "learning_rate": 9.651184092990109e-06, |
| "loss": 3.6952, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.20916334661354583, |
| "grad_norm": 0.35267430543899536, |
| "learning_rate": 9.644772579223023e-06, |
| "loss": 2.4354, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.2101593625498008, |
| "grad_norm": 0.27790936827659607, |
| "learning_rate": 9.638304845789916e-06, |
| "loss": 2.3407, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.21115537848605578, |
| "grad_norm": 0.4317843019962311, |
| "learning_rate": 9.631780970975311e-06, |
| "loss": 2.2805, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.21215139442231076, |
| "grad_norm": 0.35801681876182556, |
| "learning_rate": 9.625201033743262e-06, |
| "loss": 2.3219, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.21314741035856574, |
| "grad_norm": 0.3666556477546692, |
| "learning_rate": 9.618565113736388e-06, |
| "loss": 2.1962, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.21414342629482072, |
| "grad_norm": 0.36347630620002747, |
| "learning_rate": 9.611873291274927e-06, |
| "loss": 2.1945, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.2151394422310757, |
| "grad_norm": 0.47142109274864197, |
| "learning_rate": 9.60512564735574e-06, |
| "loss": 2.1907, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.21613545816733068, |
| "grad_norm": 0.3300761580467224, |
| "learning_rate": 9.598322263651352e-06, |
| "loss": 2.0638, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.21713147410358566, |
| "grad_norm": 0.3918429911136627, |
| "learning_rate": 9.591463222508947e-06, |
| "loss": 2.4349, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.21812749003984064, |
| "grad_norm": 0.3837280869483948, |
| "learning_rate": 9.584548606949384e-06, |
| "loss": 2.2359, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.21912350597609562, |
| "grad_norm": 0.47225990891456604, |
| "learning_rate": 9.577578500666187e-06, |
| "loss": 2.4696, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.2201195219123506, |
| "grad_norm": 0.3448033630847931, |
| "learning_rate": 9.570552988024527e-06, |
| "loss": 2.3639, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.22111553784860558, |
| "grad_norm": 0.39937150478363037, |
| "learning_rate": 9.563472154060212e-06, |
| "loss": 2.3513, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.22211155378486055, |
| "grad_norm": 0.3486849367618561, |
| "learning_rate": 9.556336084478645e-06, |
| "loss": 2.3674, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.22310756972111553, |
| "grad_norm": 0.4388813376426697, |
| "learning_rate": 9.5491448656538e-06, |
| "loss": 2.4748, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.2241035856573705, |
| "grad_norm": 0.4307428300380707, |
| "learning_rate": 9.541898584627164e-06, |
| "loss": 2.1206, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.2250996015936255, |
| "grad_norm": 0.5265683531761169, |
| "learning_rate": 9.534597329106688e-06, |
| "loss": 2.589, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.22609561752988047, |
| "grad_norm": 0.5943540930747986, |
| "learning_rate": 9.527241187465735e-06, |
| "loss": 2.8641, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.22709163346613545, |
| "grad_norm": 0.3554113209247589, |
| "learning_rate": 9.519830248741991e-06, |
| "loss": 2.0978, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.22808764940239043, |
| "grad_norm": 0.43764352798461914, |
| "learning_rate": 9.512364602636405e-06, |
| "loss": 2.2777, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.2290836653386454, |
| "grad_norm": 0.27372264862060547, |
| "learning_rate": 9.504844339512096e-06, |
| "loss": 2.3366, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.2300796812749004, |
| "grad_norm": 0.5419708490371704, |
| "learning_rate": 9.497269550393257e-06, |
| "loss": 2.5115, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.23107569721115537, |
| "grad_norm": 0.3294195234775543, |
| "learning_rate": 9.489640326964058e-06, |
| "loss": 2.3812, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.23207171314741035, |
| "grad_norm": 0.3676604926586151, |
| "learning_rate": 9.481956761567531e-06, |
| "loss": 2.1645, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.23306772908366533, |
| "grad_norm": 0.24499647319316864, |
| "learning_rate": 9.47421894720446e-06, |
| "loss": 2.436, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.2340637450199203, |
| "grad_norm": 0.4562065601348877, |
| "learning_rate": 9.466426977532246e-06, |
| "loss": 2.4614, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.2350597609561753, |
| "grad_norm": 0.4152824580669403, |
| "learning_rate": 9.458580946863784e-06, |
| "loss": 2.6406, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.2360557768924303, |
| "grad_norm": 0.2812240421772003, |
| "learning_rate": 9.45068095016631e-06, |
| "loss": 2.4129, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.23705179282868527, |
| "grad_norm": 0.37695300579071045, |
| "learning_rate": 9.442727083060258e-06, |
| "loss": 2.8288, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.23804780876494025, |
| "grad_norm": 0.29094114899635315, |
| "learning_rate": 9.434719441818106e-06, |
| "loss": 2.2392, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.23904382470119523, |
| "grad_norm": 0.6004308462142944, |
| "learning_rate": 9.426658123363202e-06, |
| "loss": 2.6978, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.2400398406374502, |
| "grad_norm": 0.824376106262207, |
| "learning_rate": 9.418543225268598e-06, |
| "loss": 2.8384, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.2410358565737052, |
| "grad_norm": 0.37361350655555725, |
| "learning_rate": 9.410374845755862e-06, |
| "loss": 2.7737, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.24203187250996017, |
| "grad_norm": 0.3311799168586731, |
| "learning_rate": 9.402153083693898e-06, |
| "loss": 2.6569, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.24302788844621515, |
| "grad_norm": 0.31674501299858093, |
| "learning_rate": 9.393878038597748e-06, |
| "loss": 2.4791, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.24402390438247012, |
| "grad_norm": 0.5512855052947998, |
| "learning_rate": 9.385549810627374e-06, |
| "loss": 2.0412, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.2450199203187251, |
| "grad_norm": 0.48202139139175415, |
| "learning_rate": 9.377168500586465e-06, |
| "loss": 2.4472, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.24601593625498008, |
| "grad_norm": 0.29134997725486755, |
| "learning_rate": 9.3687342099212e-06, |
| "loss": 2.1911, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.24701195219123506, |
| "grad_norm": 0.27262917160987854, |
| "learning_rate": 9.36024704071904e-06, |
| "loss": 2.3633, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.24800796812749004, |
| "grad_norm": 0.2618001699447632, |
| "learning_rate": 9.351707095707465e-06, |
| "loss": 2.3556, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.24900398406374502, |
| "grad_norm": 0.4861814081668854, |
| "learning_rate": 9.343114478252758e-06, |
| "loss": 2.2809, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 0.3331791162490845, |
| "learning_rate": 9.334469292358736e-06, |
| "loss": 2.2747, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.250996015936255, |
| "grad_norm": 0.3599317669868469, |
| "learning_rate": 9.3257716426655e-06, |
| "loss": 2.3204, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.25199203187250996, |
| "grad_norm": 0.28026479482650757, |
| "learning_rate": 9.317021634448162e-06, |
| "loss": 2.5997, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.25298804780876494, |
| "grad_norm": 0.4968087077140808, |
| "learning_rate": 9.308219373615574e-06, |
| "loss": 2.4623, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.2539840637450199, |
| "grad_norm": 0.5899234414100647, |
| "learning_rate": 9.299364966709051e-06, |
| "loss": 2.4678, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.2549800796812749, |
| "grad_norm": 0.27741050720214844, |
| "learning_rate": 9.290458520901072e-06, |
| "loss": 2.4373, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.2559760956175299, |
| "grad_norm": 0.44141483306884766, |
| "learning_rate": 9.28150014399399e-06, |
| "loss": 2.3013, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.25697211155378485, |
| "grad_norm": 0.4108343720436096, |
| "learning_rate": 9.272489944418724e-06, |
| "loss": 2.6281, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.25796812749003983, |
| "grad_norm": 0.4309611916542053, |
| "learning_rate": 9.263428031233444e-06, |
| "loss": 2.6192, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.2589641434262948, |
| "grad_norm": 0.3191240727901459, |
| "learning_rate": 9.25431451412226e-06, |
| "loss": 2.3667, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.2599601593625498, |
| "grad_norm": 0.4311404824256897, |
| "learning_rate": 9.245149503393884e-06, |
| "loss": 2.5286, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.26095617529880477, |
| "grad_norm": 0.8753085136413574, |
| "learning_rate": 9.235933109980302e-06, |
| "loss": 2.6609, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.26195219123505975, |
| "grad_norm": 0.679023265838623, |
| "learning_rate": 9.226665445435428e-06, |
| "loss": 2.4715, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.26294820717131473, |
| "grad_norm": 0.4910929501056671, |
| "learning_rate": 9.217346621933753e-06, |
| "loss": 2.0939, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.2639442231075697, |
| "grad_norm": 1.3370636701583862, |
| "learning_rate": 9.207976752268992e-06, |
| "loss": 2.4367, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.2649402390438247, |
| "grad_norm": 0.4148995578289032, |
| "learning_rate": 9.19855594985271e-06, |
| "loss": 2.5403, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.26593625498007967, |
| "grad_norm": 0.5130553841590881, |
| "learning_rate": 9.189084328712961e-06, |
| "loss": 2.654, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.26693227091633465, |
| "grad_norm": 0.5000612139701843, |
| "learning_rate": 9.179562003492898e-06, |
| "loss": 2.2451, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.2679282868525896, |
| "grad_norm": 1.3651481866836548, |
| "learning_rate": 9.16998908944939e-06, |
| "loss": 2.3566, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.2689243027888446, |
| "grad_norm": 0.4688972532749176, |
| "learning_rate": 9.160365702451625e-06, |
| "loss": 2.4274, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.2699203187250996, |
| "grad_norm": 0.44729602336883545, |
| "learning_rate": 9.150691958979712e-06, |
| "loss": 2.3431, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.27091633466135456, |
| "grad_norm": 0.4126404821872711, |
| "learning_rate": 9.14096797612326e-06, |
| "loss": 2.4358, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.27191235059760954, |
| "grad_norm": 0.8661454319953918, |
| "learning_rate": 9.131193871579975e-06, |
| "loss": 2.6125, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.2729083665338645, |
| "grad_norm": 0.35947325825691223, |
| "learning_rate": 9.121369763654228e-06, |
| "loss": 2.2618, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.2739043824701195, |
| "grad_norm": 0.3399883210659027, |
| "learning_rate": 9.111495771255623e-06, |
| "loss": 2.1576, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.2749003984063745, |
| "grad_norm": 0.4308667778968811, |
| "learning_rate": 9.101572013897555e-06, |
| "loss": 2.0927, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.27589641434262946, |
| "grad_norm": 0.3370983302593231, |
| "learning_rate": 9.091598611695774e-06, |
| "loss": 2.3073, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.27689243027888444, |
| "grad_norm": 0.30387991666793823, |
| "learning_rate": 9.081575685366919e-06, |
| "loss": 2.5888, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.2778884462151394, |
| "grad_norm": 0.6190817952156067, |
| "learning_rate": 9.071503356227063e-06, |
| "loss": 2.5076, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.2788844621513944, |
| "grad_norm": 0.43932202458381653, |
| "learning_rate": 9.061381746190243e-06, |
| "loss": 2.3828, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.2798804780876494, |
| "grad_norm": 0.4108044505119324, |
| "learning_rate": 9.051210977766987e-06, |
| "loss": 2.4859, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.28087649402390436, |
| "grad_norm": 0.7741344571113586, |
| "learning_rate": 9.040991174062827e-06, |
| "loss": 2.332, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.2818725099601594, |
| "grad_norm": 0.42518022656440735, |
| "learning_rate": 9.030722458776815e-06, |
| "loss": 2.5427, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.28286852589641437, |
| "grad_norm": 0.3040229082107544, |
| "learning_rate": 9.020404956200016e-06, |
| "loss": 2.5173, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.28386454183266935, |
| "grad_norm": 0.4257875680923462, |
| "learning_rate": 9.010038791214012e-06, |
| "loss": 2.3872, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.2848605577689243, |
| "grad_norm": 0.6529532074928284, |
| "learning_rate": 8.999624089289389e-06, |
| "loss": 2.2936, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.2858565737051793, |
| "grad_norm": 0.2723180055618286, |
| "learning_rate": 8.989160976484218e-06, |
| "loss": 2.5736, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.2868525896414343, |
| "grad_norm": 0.3371571898460388, |
| "learning_rate": 8.978649579442525e-06, |
| "loss": 2.3213, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.28784860557768926, |
| "grad_norm": 0.4722624719142914, |
| "learning_rate": 8.968090025392763e-06, |
| "loss": 2.4039, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.28884462151394424, |
| "grad_norm": 0.6963698863983154, |
| "learning_rate": 8.957482442146271e-06, |
| "loss": 2.4849, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.2898406374501992, |
| "grad_norm": 0.41670724749565125, |
| "learning_rate": 8.946826958095726e-06, |
| "loss": 2.4297, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.2908366533864542, |
| "grad_norm": 0.44924449920654297, |
| "learning_rate": 8.936123702213593e-06, |
| "loss": 2.29, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.2918326693227092, |
| "grad_norm": 0.5405289530754089, |
| "learning_rate": 8.925372804050554e-06, |
| "loss": 2.4732, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.29282868525896416, |
| "grad_norm": 0.5333283543586731, |
| "learning_rate": 8.914574393733953e-06, |
| "loss": 2.3553, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.29382470119521914, |
| "grad_norm": 0.4173821806907654, |
| "learning_rate": 8.903728601966206e-06, |
| "loss": 2.4729, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.2948207171314741, |
| "grad_norm": 0.6668480634689331, |
| "learning_rate": 8.892835560023236e-06, |
| "loss": 2.7302, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.2958167330677291, |
| "grad_norm": 0.5601832270622253, |
| "learning_rate": 8.881895399752873e-06, |
| "loss": 2.8201, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.2968127490039841, |
| "grad_norm": 0.7715175151824951, |
| "learning_rate": 8.870908253573255e-06, |
| "loss": 2.439, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.29780876494023906, |
| "grad_norm": 0.6411163210868835, |
| "learning_rate": 8.85987425447124e-06, |
| "loss": 2.2098, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.29880478087649404, |
| "grad_norm": 1.8174595832824707, |
| "learning_rate": 8.848793536000779e-06, |
| "loss": 2.862, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.299800796812749, |
| "grad_norm": 0.4861983060836792, |
| "learning_rate": 8.837666232281312e-06, |
| "loss": 1.964, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.300796812749004, |
| "grad_norm": 0.629531979560852, |
| "learning_rate": 8.826492477996138e-06, |
| "loss": 2.4866, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.301792828685259, |
| "grad_norm": 0.33506232500076294, |
| "learning_rate": 8.81527240839079e-06, |
| "loss": 2.1813, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.30278884462151395, |
| "grad_norm": 0.5065098404884338, |
| "learning_rate": 8.80400615927139e-06, |
| "loss": 2.2313, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.30378486055776893, |
| "grad_norm": 0.31633898615837097, |
| "learning_rate": 8.792693867003017e-06, |
| "loss": 2.5764, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.3047808764940239, |
| "grad_norm": 0.6082801818847656, |
| "learning_rate": 8.781335668508044e-06, |
| "loss": 2.1408, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.3057768924302789, |
| "grad_norm": 0.3309324383735657, |
| "learning_rate": 8.76993170126449e-06, |
| "loss": 2.5198, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.30677290836653387, |
| "grad_norm": 0.38401421904563904, |
| "learning_rate": 8.758482103304348e-06, |
| "loss": 2.2784, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.30776892430278885, |
| "grad_norm": 0.4021207392215729, |
| "learning_rate": 8.746987013211924e-06, |
| "loss": 2.1789, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.30876494023904383, |
| "grad_norm": 1.5585757493972778, |
| "learning_rate": 8.735446570122151e-06, |
| "loss": 2.429, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.3097609561752988, |
| "grad_norm": 0.5734106302261353, |
| "learning_rate": 8.72386091371891e-06, |
| "loss": 2.3943, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.3107569721115538, |
| "grad_norm": 0.5181722044944763, |
| "learning_rate": 8.712230184233337e-06, |
| "loss": 2.4501, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.31175298804780877, |
| "grad_norm": 0.42989951372146606, |
| "learning_rate": 8.700554522442124e-06, |
| "loss": 2.3612, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.31274900398406374, |
| "grad_norm": 0.8372073769569397, |
| "learning_rate": 8.688834069665819e-06, |
| "loss": 2.6138, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.3137450199203187, |
| "grad_norm": 0.5447811484336853, |
| "learning_rate": 8.677068967767117e-06, |
| "loss": 2.4036, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.3147410358565737, |
| "grad_norm": 0.28892412781715393, |
| "learning_rate": 8.665259359149132e-06, |
| "loss": 2.4249, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.3157370517928287, |
| "grad_norm": 0.33981916308403015, |
| "learning_rate": 8.653405386753688e-06, |
| "loss": 2.8255, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.31673306772908366, |
| "grad_norm": 0.37266361713409424, |
| "learning_rate": 8.64150719405958e-06, |
| "loss": 2.4619, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.31772908366533864, |
| "grad_norm": 0.611991286277771, |
| "learning_rate": 8.629564925080838e-06, |
| "loss": 2.6266, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.3187250996015936, |
| "grad_norm": 0.7753398418426514, |
| "learning_rate": 8.617578724364984e-06, |
| "loss": 2.5113, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.3197211155378486, |
| "grad_norm": 0.7812793850898743, |
| "learning_rate": 8.605548736991284e-06, |
| "loss": 2.1736, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.3207171314741036, |
| "grad_norm": 0.5114774703979492, |
| "learning_rate": 8.593475108568995e-06, |
| "loss": 2.7169, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.32171314741035856, |
| "grad_norm": 0.3701231777667999, |
| "learning_rate": 8.581357985235595e-06, |
| "loss": 2.2407, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.32270916334661354, |
| "grad_norm": 1.135130763053894, |
| "learning_rate": 8.569197513655022e-06, |
| "loss": 2.0901, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.3237051792828685, |
| "grad_norm": 0.2848869562149048, |
| "learning_rate": 8.55699384101589e-06, |
| "loss": 2.592, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.3247011952191235, |
| "grad_norm": 0.5609592795372009, |
| "learning_rate": 8.544747115029717e-06, |
| "loss": 2.3673, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.3256972111553785, |
| "grad_norm": 0.3470471203327179, |
| "learning_rate": 8.53245748392913e-06, |
| "loss": 2.377, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.32669322709163345, |
| "grad_norm": 0.9178757667541504, |
| "learning_rate": 8.520125096466072e-06, |
| "loss": 2.7617, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.32768924302788843, |
| "grad_norm": 0.46402791142463684, |
| "learning_rate": 8.50775010191001e-06, |
| "loss": 2.2848, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.3286852589641434, |
| "grad_norm": 0.4459151327610016, |
| "learning_rate": 8.495332650046112e-06, |
| "loss": 2.1105, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.3296812749003984, |
| "grad_norm": 0.7026370763778687, |
| "learning_rate": 8.48287289117345e-06, |
| "loss": 2.6823, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.33067729083665337, |
| "grad_norm": 0.7429327964782715, |
| "learning_rate": 8.470370976103171e-06, |
| "loss": 2.2689, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.33167330677290835, |
| "grad_norm": 0.37948814034461975, |
| "learning_rate": 8.457827056156673e-06, |
| "loss": 2.628, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.33266932270916333, |
| "grad_norm": 0.3749179244041443, |
| "learning_rate": 8.44524128316378e-06, |
| "loss": 2.4552, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.3336653386454183, |
| "grad_norm": 0.421303391456604, |
| "learning_rate": 8.432613809460895e-06, |
| "loss": 2.1797, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.3346613545816733, |
| "grad_norm": 0.5645405054092407, |
| "learning_rate": 8.419944787889162e-06, |
| "loss": 2.4308, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.33565737051792827, |
| "grad_norm": 0.7171806693077087, |
| "learning_rate": 8.407234371792614e-06, |
| "loss": 2.3868, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.33665338645418325, |
| "grad_norm": 0.7937443852424622, |
| "learning_rate": 8.394482715016318e-06, |
| "loss": 2.433, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.3376494023904382, |
| "grad_norm": 0.5325895547866821, |
| "learning_rate": 8.381689971904514e-06, |
| "loss": 2.5116, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.3386454183266932, |
| "grad_norm": 0.3550787568092346, |
| "learning_rate": 8.368856297298742e-06, |
| "loss": 2.4187, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.3396414342629482, |
| "grad_norm": 0.5118217468261719, |
| "learning_rate": 8.355981846535972e-06, |
| "loss": 2.1325, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.34063745019920316, |
| "grad_norm": 0.35231295228004456, |
| "learning_rate": 8.343066775446724e-06, |
| "loss": 2.3751, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.34163346613545814, |
| "grad_norm": 0.6949347853660583, |
| "learning_rate": 8.330111240353178e-06, |
| "loss": 2.5615, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.3426294820717131, |
| "grad_norm": 0.5718231797218323, |
| "learning_rate": 8.317115398067289e-06, |
| "loss": 2.2858, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.3436254980079681, |
| "grad_norm": 0.6337103843688965, |
| "learning_rate": 8.30407940588888e-06, |
| "loss": 2.5088, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.34462151394422313, |
| "grad_norm": 0.4129788875579834, |
| "learning_rate": 8.29100342160374e-06, |
| "loss": 2.2748, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.3456175298804781, |
| "grad_norm": 0.3462570607662201, |
| "learning_rate": 8.27788760348173e-06, |
| "loss": 2.3205, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.3466135458167331, |
| "grad_norm": 0.4417884349822998, |
| "learning_rate": 8.26473211027484e-06, |
| "loss": 2.3901, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.34760956175298807, |
| "grad_norm": 0.45579978823661804, |
| "learning_rate": 8.251537101215287e-06, |
| "loss": 2.2336, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.34860557768924305, |
| "grad_norm": 0.6957226991653442, |
| "learning_rate": 8.238302736013587e-06, |
| "loss": 2.7518, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.34960159362549803, |
| "grad_norm": 0.696114718914032, |
| "learning_rate": 8.225029174856602e-06, |
| "loss": 2.0373, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.350597609561753, |
| "grad_norm": 0.40747031569480896, |
| "learning_rate": 8.211716578405635e-06, |
| "loss": 2.4626, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.351593625498008, |
| "grad_norm": 0.47290411591529846, |
| "learning_rate": 8.198365107794457e-06, |
| "loss": 2.5871, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.35258964143426297, |
| "grad_norm": 0.592217206954956, |
| "learning_rate": 8.184974924627365e-06, |
| "loss": 2.3886, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.35358565737051795, |
| "grad_norm": 0.8985310196876526, |
| "learning_rate": 8.171546190977231e-06, |
| "loss": 2.6021, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.3545816733067729, |
| "grad_norm": 0.7999231815338135, |
| "learning_rate": 8.158079069383535e-06, |
| "loss": 2.2757, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.3555776892430279, |
| "grad_norm": 0.6230331063270569, |
| "learning_rate": 8.1445737228504e-06, |
| "loss": 2.4343, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.3565737051792829, |
| "grad_norm": 0.43640607595443726, |
| "learning_rate": 8.131030314844617e-06, |
| "loss": 1.9181, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.35756972111553786, |
| "grad_norm": 0.5745819211006165, |
| "learning_rate": 8.117449009293668e-06, |
| "loss": 2.9202, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.35856573705179284, |
| "grad_norm": 0.46643760800361633, |
| "learning_rate": 8.103829970583742e-06, |
| "loss": 2.5197, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.3595617529880478, |
| "grad_norm": 0.5368001461029053, |
| "learning_rate": 8.090173363557748e-06, |
| "loss": 2.3562, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.3605577689243028, |
| "grad_norm": 0.5117561221122742, |
| "learning_rate": 8.076479353513308e-06, |
| "loss": 2.2398, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.3615537848605578, |
| "grad_norm": 1.2853957414627075, |
| "learning_rate": 8.06274810620077e-06, |
| "loss": 2.041, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.36254980079681276, |
| "grad_norm": 0.5592566132545471, |
| "learning_rate": 8.0489797878212e-06, |
| "loss": 2.614, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.36354581673306774, |
| "grad_norm": 0.3889990448951721, |
| "learning_rate": 8.035174565024362e-06, |
| "loss": 2.4095, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.3645418326693227, |
| "grad_norm": 0.39642640948295593, |
| "learning_rate": 8.021332604906709e-06, |
| "loss": 1.9971, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.3655378486055777, |
| "grad_norm": 0.5725635290145874, |
| "learning_rate": 8.007454075009352e-06, |
| "loss": 2.4322, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.3665338645418327, |
| "grad_norm": 0.538329541683197, |
| "learning_rate": 7.993539143316044e-06, |
| "loss": 2.3927, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.36752988047808766, |
| "grad_norm": 0.8502817153930664, |
| "learning_rate": 7.979587978251136e-06, |
| "loss": 2.6487, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.36852589641434264, |
| "grad_norm": 0.5096033811569214, |
| "learning_rate": 7.965600748677545e-06, |
| "loss": 2.4393, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.3695219123505976, |
| "grad_norm": 0.39716872572898865, |
| "learning_rate": 7.951577623894701e-06, |
| "loss": 2.9555, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.3705179282868526, |
| "grad_norm": 0.6120476126670837, |
| "learning_rate": 7.937518773636518e-06, |
| "loss": 2.1758, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.3715139442231076, |
| "grad_norm": 0.4519326984882355, |
| "learning_rate": 7.923424368069312e-06, |
| "loss": 2.4508, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.37250996015936255, |
| "grad_norm": 0.4252610504627228, |
| "learning_rate": 7.909294577789765e-06, |
| "loss": 2.134, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.37350597609561753, |
| "grad_norm": 0.5511481761932373, |
| "learning_rate": 7.895129573822844e-06, |
| "loss": 2.4735, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.3745019920318725, |
| "grad_norm": 1.1340324878692627, |
| "learning_rate": 7.880929527619742e-06, |
| "loss": 2.7602, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.3754980079681275, |
| "grad_norm": 0.45862501859664917, |
| "learning_rate": 7.866694611055796e-06, |
| "loss": 2.5242, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.37649402390438247, |
| "grad_norm": 0.48843666911125183, |
| "learning_rate": 7.852424996428412e-06, |
| "loss": 2.3878, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.37749003984063745, |
| "grad_norm": 1.07990562915802, |
| "learning_rate": 7.838120856454967e-06, |
| "loss": 2.2745, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.3784860557768924, |
| "grad_norm": 0.466766893863678, |
| "learning_rate": 7.823782364270743e-06, |
| "loss": 2.5844, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.3794820717131474, |
| "grad_norm": 0.6437628865242004, |
| "learning_rate": 7.809409693426803e-06, |
| "loss": 2.381, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.3804780876494024, |
| "grad_norm": 0.6378084421157837, |
| "learning_rate": 7.79500301788791e-06, |
| "loss": 2.4076, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.38147410358565736, |
| "grad_norm": 0.6559402346611023, |
| "learning_rate": 7.780562512030414e-06, |
| "loss": 2.2115, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.38247011952191234, |
| "grad_norm": 0.4882892668247223, |
| "learning_rate": 7.766088350640141e-06, |
| "loss": 2.179, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.3834661354581673, |
| "grad_norm": 0.37981244921684265, |
| "learning_rate": 7.75158070891028e-06, |
| "loss": 2.3806, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.3844621513944223, |
| "grad_norm": 0.6670547723770142, |
| "learning_rate": 7.737039762439263e-06, |
| "loss": 2.0258, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.3854581673306773, |
| "grad_norm": 0.5870895981788635, |
| "learning_rate": 7.722465687228634e-06, |
| "loss": 2.4609, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.38645418326693226, |
| "grad_norm": 0.5210617184638977, |
| "learning_rate": 7.707858659680924e-06, |
| "loss": 2.0152, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.38745019920318724, |
| "grad_norm": 0.8872121572494507, |
| "learning_rate": 7.693218856597515e-06, |
| "loss": 2.4107, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.3884462151394422, |
| "grad_norm": 0.408750057220459, |
| "learning_rate": 7.6785464551765e-06, |
| "loss": 2.4381, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.3894422310756972, |
| "grad_norm": 0.4789107143878937, |
| "learning_rate": 7.663841633010539e-06, |
| "loss": 2.3057, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.3904382470119522, |
| "grad_norm": 0.37599480152130127, |
| "learning_rate": 7.649104568084701e-06, |
| "loss": 2.4101, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.39143426294820716, |
| "grad_norm": 0.6352246403694153, |
| "learning_rate": 7.634335438774325e-06, |
| "loss": 2.2202, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.39243027888446214, |
| "grad_norm": 0.5684521198272705, |
| "learning_rate": 7.619534423842852e-06, |
| "loss": 2.7473, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.3934262948207171, |
| "grad_norm": 0.5998817682266235, |
| "learning_rate": 7.604701702439652e-06, |
| "loss": 2.109, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.3944223107569721, |
| "grad_norm": 0.69579017162323, |
| "learning_rate": 7.589837454097879e-06, |
| "loss": 2.6706, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.3954183266932271, |
| "grad_norm": 0.5617753267288208, |
| "learning_rate": 7.574941858732279e-06, |
| "loss": 2.4175, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.39641434262948205, |
| "grad_norm": 0.509094774723053, |
| "learning_rate": 7.560015096637015e-06, |
| "loss": 2.4367, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.39741035856573703, |
| "grad_norm": 0.5854381322860718, |
| "learning_rate": 7.54505734848349e-06, |
| "loss": 2.3366, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.398406374501992, |
| "grad_norm": 0.7329273819923401, |
| "learning_rate": 7.53006879531816e-06, |
| "loss": 2.2252, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.399402390438247, |
| "grad_norm": 0.7855085730552673, |
| "learning_rate": 7.515049618560337e-06, |
| "loss": 2.2244, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.40039840637450197, |
| "grad_norm": 0.608400821685791, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 2.2232, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.40139442231075695, |
| "grad_norm": 0.5910929441452026, |
| "learning_rate": 7.484920121795589e-06, |
| "loss": 2.4034, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.40239043824701193, |
| "grad_norm": 0.5254145264625549, |
| "learning_rate": 7.469810166471802e-06, |
| "loss": 2.3053, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.4033864541832669, |
| "grad_norm": 0.5197448134422302, |
| "learning_rate": 7.454670316917387e-06, |
| "loss": 2.2416, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.4043824701195219, |
| "grad_norm": 0.759675145149231, |
| "learning_rate": 7.4395007563829295e-06, |
| "loss": 2.4197, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.40537848605577687, |
| "grad_norm": 0.4646972417831421, |
| "learning_rate": 7.424301668478626e-06, |
| "loss": 2.5638, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.4063745019920319, |
| "grad_norm": 0.5564824938774109, |
| "learning_rate": 7.4090732371720774e-06, |
| "loss": 2.5038, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.4073705179282869, |
| "grad_norm": 0.8897591829299927, |
| "learning_rate": 7.393815646786047e-06, |
| "loss": 2.6751, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.40836653386454186, |
| "grad_norm": 0.8237727284431458, |
| "learning_rate": 7.378529081996233e-06, |
| "loss": 2.2649, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.40936254980079684, |
| "grad_norm": 0.4922022521495819, |
| "learning_rate": 7.363213727829045e-06, |
| "loss": 2.2894, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.4103585657370518, |
| "grad_norm": 0.48476412892341614, |
| "learning_rate": 7.347869769659346e-06, |
| "loss": 2.4317, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.4113545816733068, |
| "grad_norm": 0.7627730965614319, |
| "learning_rate": 7.332497393208221e-06, |
| "loss": 3.0171, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.4123505976095618, |
| "grad_norm": 1.4489892721176147, |
| "learning_rate": 7.317096784540728e-06, |
| "loss": 3.0745, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.41334661354581675, |
| "grad_norm": 0.44949018955230713, |
| "learning_rate": 7.301668130063639e-06, |
| "loss": 2.4086, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.41434262948207173, |
| "grad_norm": 0.9026828408241272, |
| "learning_rate": 7.286211616523193e-06, |
| "loss": 2.4219, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.4153386454183267, |
| "grad_norm": 0.5528742074966431, |
| "learning_rate": 7.2707274310028306e-06, |
| "loss": 2.069, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.4163346613545817, |
| "grad_norm": 0.8069695830345154, |
| "learning_rate": 7.255215760920925e-06, |
| "loss": 2.2901, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.41733067729083667, |
| "grad_norm": 0.9854758977890015, |
| "learning_rate": 7.239676794028526e-06, |
| "loss": 2.533, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.41832669322709165, |
| "grad_norm": 0.8304996490478516, |
| "learning_rate": 7.224110718407075e-06, |
| "loss": 2.4076, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.41932270916334663, |
| "grad_norm": 0.5824740529060364, |
| "learning_rate": 7.208517722466135e-06, |
| "loss": 2.2298, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.4203187250996016, |
| "grad_norm": 0.740998387336731, |
| "learning_rate": 7.192897994941111e-06, |
| "loss": 2.2991, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.4213147410358566, |
| "grad_norm": 0.9516714215278625, |
| "learning_rate": 7.177251724890957e-06, |
| "loss": 2.2348, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.42231075697211157, |
| "grad_norm": 1.330517053604126, |
| "learning_rate": 7.1615791016959024e-06, |
| "loss": 2.5969, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.42330677290836655, |
| "grad_norm": 0.8636577129364014, |
| "learning_rate": 7.145880315055145e-06, |
| "loss": 2.4203, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.4243027888446215, |
| "grad_norm": 0.40839532017707825, |
| "learning_rate": 7.1301555549845634e-06, |
| "loss": 2.428, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.4252988047808765, |
| "grad_norm": 0.6337350010871887, |
| "learning_rate": 7.114405011814415e-06, |
| "loss": 2.3018, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.4262948207171315, |
| "grad_norm": 0.8653415441513062, |
| "learning_rate": 7.098628876187031e-06, |
| "loss": 2.8947, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.42729083665338646, |
| "grad_norm": 0.7558097839355469, |
| "learning_rate": 7.082827339054513e-06, |
| "loss": 2.2061, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.42828685258964144, |
| "grad_norm": 0.8000844717025757, |
| "learning_rate": 7.067000591676416e-06, |
| "loss": 2.4076, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.4292828685258964, |
| "grad_norm": 0.7074631452560425, |
| "learning_rate": 7.051148825617435e-06, |
| "loss": 2.7278, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.4302788844621514, |
| "grad_norm": 1.1026921272277832, |
| "learning_rate": 7.035272232745093e-06, |
| "loss": 2.5602, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.4312749003984064, |
| "grad_norm": 0.8589175939559937, |
| "learning_rate": 7.019371005227407e-06, |
| "loss": 2.3789, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.43227091633466136, |
| "grad_norm": 0.8680172562599182, |
| "learning_rate": 7.003445335530572e-06, |
| "loss": 2.6989, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.43326693227091634, |
| "grad_norm": 0.7565051317214966, |
| "learning_rate": 6.987495416416627e-06, |
| "loss": 2.3688, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.4342629482071713, |
| "grad_norm": 1.050288438796997, |
| "learning_rate": 6.9715214409411204e-06, |
| "loss": 2.1948, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.4352589641434263, |
| "grad_norm": 0.5728120803833008, |
| "learning_rate": 6.95552360245078e-06, |
| "loss": 2.2716, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.4362549800796813, |
| "grad_norm": 1.0542654991149902, |
| "learning_rate": 6.939502094581164e-06, |
| "loss": 2.5845, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.43725099601593626, |
| "grad_norm": 0.7242105603218079, |
| "learning_rate": 6.923457111254322e-06, |
| "loss": 2.2048, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.43824701195219123, |
| "grad_norm": 1.0688732862472534, |
| "learning_rate": 6.9073888466764495e-06, |
| "loss": 2.1883, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.4392430278884462, |
| "grad_norm": 0.8276563286781311, |
| "learning_rate": 6.891297495335531e-06, |
| "loss": 2.2532, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.4402390438247012, |
| "grad_norm": 0.7350841760635376, |
| "learning_rate": 6.875183251998993e-06, |
| "loss": 2.2517, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.44123505976095617, |
| "grad_norm": 0.7074620127677917, |
| "learning_rate": 6.859046311711344e-06, |
| "loss": 2.1943, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.44223107569721115, |
| "grad_norm": 0.9273977279663086, |
| "learning_rate": 6.84288686979181e-06, |
| "loss": 2.4566, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.44322709163346613, |
| "grad_norm": 0.54817795753479, |
| "learning_rate": 6.8267051218319766e-06, |
| "loss": 2.1578, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.4442231075697211, |
| "grad_norm": 0.8614233136177063, |
| "learning_rate": 6.810501263693416e-06, |
| "loss": 2.3546, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.4452191235059761, |
| "grad_norm": 0.5919457077980042, |
| "learning_rate": 6.7942754915053225e-06, |
| "loss": 2.4907, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.44621513944223107, |
| "grad_norm": 0.7229816317558289, |
| "learning_rate": 6.77802800166213e-06, |
| "loss": 2.2884, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.44721115537848605, |
| "grad_norm": 1.1251389980316162, |
| "learning_rate": 6.761758990821143e-06, |
| "loss": 2.508, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.448207171314741, |
| "grad_norm": 0.7033310532569885, |
| "learning_rate": 6.745468655900156e-06, |
| "loss": 2.4315, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.449203187250996, |
| "grad_norm": 1.2203772068023682, |
| "learning_rate": 6.7291571940750575e-06, |
| "loss": 2.6156, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.450199203187251, |
| "grad_norm": 0.7150283455848694, |
| "learning_rate": 6.712824802777465e-06, |
| "loss": 2.2121, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.45119521912350596, |
| "grad_norm": 0.9422833323478699, |
| "learning_rate": 6.696471679692313e-06, |
| "loss": 2.2294, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.45219123505976094, |
| "grad_norm": 0.6846040487289429, |
| "learning_rate": 6.680098022755478e-06, |
| "loss": 2.5567, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.4531872509960159, |
| "grad_norm": 0.5450727343559265, |
| "learning_rate": 6.66370403015137e-06, |
| "loss": 2.4599, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.4541832669322709, |
| "grad_norm": 0.5914618968963623, |
| "learning_rate": 6.647289900310545e-06, |
| "loss": 2.5134, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.4551792828685259, |
| "grad_norm": 0.8993861079216003, |
| "learning_rate": 6.63085583190729e-06, |
| "loss": 2.5729, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.45617529880478086, |
| "grad_norm": 0.5783509016036987, |
| "learning_rate": 6.614402023857231e-06, |
| "loss": 2.5881, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.45717131474103584, |
| "grad_norm": 0.6340298652648926, |
| "learning_rate": 6.597928675314918e-06, |
| "loss": 2.4509, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.4581673306772908, |
| "grad_norm": 0.7813217043876648, |
| "learning_rate": 6.581435985671418e-06, |
| "loss": 2.242, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.4591633466135458, |
| "grad_norm": 0.8051680326461792, |
| "learning_rate": 6.564924154551895e-06, |
| "loss": 2.398, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.4601593625498008, |
| "grad_norm": 0.6447633504867554, |
| "learning_rate": 6.548393381813205e-06, |
| "loss": 2.9214, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.46115537848605576, |
| "grad_norm": 0.5684821605682373, |
| "learning_rate": 6.5318438675414665e-06, |
| "loss": 2.2545, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.46215139442231074, |
| "grad_norm": 0.6067225337028503, |
| "learning_rate": 6.515275812049644e-06, |
| "loss": 2.5541, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.4631474103585657, |
| "grad_norm": 0.5634474754333496, |
| "learning_rate": 6.498689415875121e-06, |
| "loss": 2.581, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.4641434262948207, |
| "grad_norm": 0.4764470160007477, |
| "learning_rate": 6.48208487977728e-06, |
| "loss": 2.2492, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.4651394422310757, |
| "grad_norm": 0.6636649370193481, |
| "learning_rate": 6.4654624047350575e-06, |
| "loss": 2.418, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.46613545816733065, |
| "grad_norm": 0.6568376421928406, |
| "learning_rate": 6.448822191944526e-06, |
| "loss": 2.4608, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.46713147410358563, |
| "grad_norm": 0.5867657661437988, |
| "learning_rate": 6.432164442816452e-06, |
| "loss": 2.7101, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.4681274900398406, |
| "grad_norm": 0.5845500826835632, |
| "learning_rate": 6.41548935897386e-06, |
| "loss": 2.1822, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.46912350597609564, |
| "grad_norm": 0.9894917011260986, |
| "learning_rate": 6.398797142249591e-06, |
| "loss": 2.4232, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.4701195219123506, |
| "grad_norm": 0.5987226366996765, |
| "learning_rate": 6.3820879946838585e-06, |
| "loss": 2.6506, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.4711155378486056, |
| "grad_norm": 1.552355408668518, |
| "learning_rate": 6.365362118521807e-06, |
| "loss": 3.0233, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.4721115537848606, |
| "grad_norm": 0.6667497158050537, |
| "learning_rate": 6.348619716211058e-06, |
| "loss": 2.6748, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.47310756972111556, |
| "grad_norm": 0.970600962638855, |
| "learning_rate": 6.33186099039927e-06, |
| "loss": 2.4542, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.47410358565737054, |
| "grad_norm": 0.6683152914047241, |
| "learning_rate": 6.31508614393167e-06, |
| "loss": 2.4034, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.4750996015936255, |
| "grad_norm": 0.9013263583183289, |
| "learning_rate": 6.2982953798486124e-06, |
| "loss": 2.4747, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.4760956175298805, |
| "grad_norm": 0.8827518820762634, |
| "learning_rate": 6.2814889013831174e-06, |
| "loss": 2.5649, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.4770916334661355, |
| "grad_norm": 0.8040870428085327, |
| "learning_rate": 6.264666911958404e-06, |
| "loss": 2.2855, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.47808764940239046, |
| "grad_norm": 0.9028819799423218, |
| "learning_rate": 6.247829615185441e-06, |
| "loss": 2.3607, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.47908366533864544, |
| "grad_norm": 0.8722829818725586, |
| "learning_rate": 6.230977214860468e-06, |
| "loss": 2.2346, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.4800796812749004, |
| "grad_norm": 0.6933993697166443, |
| "learning_rate": 6.214109914962542e-06, |
| "loss": 2.7604, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.4810756972111554, |
| "grad_norm": 0.6176011562347412, |
| "learning_rate": 6.1972279196510565e-06, |
| "loss": 2.8278, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.4820717131474104, |
| "grad_norm": 0.7204033732414246, |
| "learning_rate": 6.180331433263283e-06, |
| "loss": 2.4275, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.48306772908366535, |
| "grad_norm": 1.1777395009994507, |
| "learning_rate": 6.1634206603118844e-06, |
| "loss": 2.501, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.48406374501992033, |
| "grad_norm": 1.3905079364776611, |
| "learning_rate": 6.146495805482451e-06, |
| "loss": 2.9668, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.4850597609561753, |
| "grad_norm": 1.8294525146484375, |
| "learning_rate": 6.129557073631013e-06, |
| "loss": 2.0932, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.4860557768924303, |
| "grad_norm": 1.0543792247772217, |
| "learning_rate": 6.112604669781572e-06, |
| "loss": 2.1609, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.48705179282868527, |
| "grad_norm": 0.6805559992790222, |
| "learning_rate": 6.09563879912361e-06, |
| "loss": 2.5514, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.48804780876494025, |
| "grad_norm": 0.913993239402771, |
| "learning_rate": 6.07865966700961e-06, |
| "loss": 2.1365, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.48904382470119523, |
| "grad_norm": 1.127585768699646, |
| "learning_rate": 6.06166747895257e-06, |
| "loss": 2.5846, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.4900398406374502, |
| "grad_norm": 0.5136232376098633, |
| "learning_rate": 6.044662440623512e-06, |
| "loss": 2.3032, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.4910358565737052, |
| "grad_norm": 1.1162834167480469, |
| "learning_rate": 6.027644757849004e-06, |
| "loss": 2.4506, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.49203187250996017, |
| "grad_norm": 0.8343062996864319, |
| "learning_rate": 6.0106146366086514e-06, |
| "loss": 2.5259, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.49302788844621515, |
| "grad_norm": 1.2255266904830933, |
| "learning_rate": 5.99357228303262e-06, |
| "loss": 2.4964, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.4940239043824701, |
| "grad_norm": 0.8280947804450989, |
| "learning_rate": 5.976517903399128e-06, |
| "loss": 2.3576, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.4950199203187251, |
| "grad_norm": 0.5358011722564697, |
| "learning_rate": 5.959451704131962e-06, |
| "loss": 2.4103, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.4960159362549801, |
| "grad_norm": 0.7300974130630493, |
| "learning_rate": 5.9423738917979655e-06, |
| "loss": 2.4792, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.49701195219123506, |
| "grad_norm": 1.1368849277496338, |
| "learning_rate": 5.92528467310455e-06, |
| "loss": 2.7343, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.49800796812749004, |
| "grad_norm": 1.1836261749267578, |
| "learning_rate": 5.908184254897183e-06, |
| "loss": 2.263, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.499003984063745, |
| "grad_norm": 0.7548766136169434, |
| "learning_rate": 5.891072844156895e-06, |
| "loss": 2.5232, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 0.5676743984222412, |
| "learning_rate": 5.87395064799776e-06, |
| "loss": 2.532, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.500996015936255, |
| "grad_norm": 0.6243408918380737, |
| "learning_rate": 5.856817873664409e-06, |
| "loss": 2.6337, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.50199203187251, |
| "grad_norm": 0.6920816898345947, |
| "learning_rate": 5.839674728529499e-06, |
| "loss": 2.3761, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.5029880478087649, |
| "grad_norm": 0.773199737071991, |
| "learning_rate": 5.8225214200912195e-06, |
| "loss": 2.3334, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.5039840637450199, |
| "grad_norm": 0.8243865370750427, |
| "learning_rate": 5.8053581559707754e-06, |
| "loss": 2.2201, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.5049800796812749, |
| "grad_norm": 0.6483944654464722, |
| "learning_rate": 5.788185143909868e-06, |
| "loss": 2.2019, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.5059760956175299, |
| "grad_norm": 1.1562165021896362, |
| "learning_rate": 5.7710025917681954e-06, |
| "loss": 2.6738, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.5069721115537849, |
| "grad_norm": 0.7714682221412659, |
| "learning_rate": 5.753810707520918e-06, |
| "loss": 2.5662, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.5079681274900398, |
| "grad_norm": 1.8939898014068604, |
| "learning_rate": 5.736609699256158e-06, |
| "loss": 2.4841, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.5089641434262948, |
| "grad_norm": 1.0769683122634888, |
| "learning_rate": 5.719399775172468e-06, |
| "loss": 2.2339, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.5099601593625498, |
| "grad_norm": 0.7231705188751221, |
| "learning_rate": 5.702181143576323e-06, |
| "loss": 2.2754, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.5109561752988048, |
| "grad_norm": 1.0154212713241577, |
| "learning_rate": 5.684954012879583e-06, |
| "loss": 2.5029, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.5119521912350598, |
| "grad_norm": 1.130210041999817, |
| "learning_rate": 5.66771859159699e-06, |
| "loss": 2.5272, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.5129482071713147, |
| "grad_norm": 0.6206554770469666, |
| "learning_rate": 5.6504750883436275e-06, |
| "loss": 2.3235, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.5139442231075697, |
| "grad_norm": 1.0481131076812744, |
| "learning_rate": 5.633223711832403e-06, |
| "loss": 2.2866, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.5149402390438247, |
| "grad_norm": 0.7321045398712158, |
| "learning_rate": 5.615964670871524e-06, |
| "loss": 2.4402, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.5159362549800797, |
| "grad_norm": 0.7684382796287537, |
| "learning_rate": 5.5986981743619615e-06, |
| "loss": 2.3404, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.5169322709163346, |
| "grad_norm": 1.1246601343154907, |
| "learning_rate": 5.581424431294936e-06, |
| "loss": 2.2032, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.5179282868525896, |
| "grad_norm": 0.617168128490448, |
| "learning_rate": 5.56414365074937e-06, |
| "loss": 2.5379, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.5189243027888446, |
| "grad_norm": 0.7718055248260498, |
| "learning_rate": 5.546856041889374e-06, |
| "loss": 2.4803, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.5199203187250996, |
| "grad_norm": 1.5422130823135376, |
| "learning_rate": 5.5295618139617e-06, |
| "loss": 2.3014, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.5209163346613546, |
| "grad_norm": 0.5689607262611389, |
| "learning_rate": 5.512261176293226e-06, |
| "loss": 2.1094, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.5219123505976095, |
| "grad_norm": 0.6268129944801331, |
| "learning_rate": 5.494954338288404e-06, |
| "loss": 2.4562, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.5229083665338645, |
| "grad_norm": 1.1070692539215088, |
| "learning_rate": 5.477641509426739e-06, |
| "loss": 2.0365, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.5239043824701195, |
| "grad_norm": 1.2830649614334106, |
| "learning_rate": 5.460322899260245e-06, |
| "loss": 2.2455, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.5249003984063745, |
| "grad_norm": 0.6706056594848633, |
| "learning_rate": 5.442998717410916e-06, |
| "loss": 2.3095, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.5258964143426295, |
| "grad_norm": 0.6565206050872803, |
| "learning_rate": 5.425669173568179e-06, |
| "loss": 2.5873, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.5268924302788844, |
| "grad_norm": 0.8396487236022949, |
| "learning_rate": 5.408334477486369e-06, |
| "loss": 2.4121, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.5278884462151394, |
| "grad_norm": 0.8442867398262024, |
| "learning_rate": 5.390994838982178e-06, |
| "loss": 2.1112, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.5288844621513944, |
| "grad_norm": 0.6712931990623474, |
| "learning_rate": 5.373650467932122e-06, |
| "loss": 2.5735, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.5298804780876494, |
| "grad_norm": 0.8547630906105042, |
| "learning_rate": 5.356301574269998e-06, |
| "loss": 2.5899, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.5308764940239044, |
| "grad_norm": 1.1971904039382935, |
| "learning_rate": 5.338948367984347e-06, |
| "loss": 2.7327, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.5318725099601593, |
| "grad_norm": 0.8025546669960022, |
| "learning_rate": 5.321591059115906e-06, |
| "loss": 2.3269, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.5328685258964143, |
| "grad_norm": 0.9815241098403931, |
| "learning_rate": 5.30422985775507e-06, |
| "loss": 2.4555, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.5338645418326693, |
| "grad_norm": 0.9032710194587708, |
| "learning_rate": 5.286864974039349e-06, |
| "loss": 2.4246, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.5348605577689243, |
| "grad_norm": 1.508058786392212, |
| "learning_rate": 5.269496618150823e-06, |
| "loss": 2.1206, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.5358565737051793, |
| "grad_norm": 1.0455362796783447, |
| "learning_rate": 5.2521250003136005e-06, |
| "loss": 2.4406, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.5368525896414342, |
| "grad_norm": 0.8184682726860046, |
| "learning_rate": 5.234750330791268e-06, |
| "loss": 2.1303, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.5378486055776892, |
| "grad_norm": 0.8355326056480408, |
| "learning_rate": 5.217372819884353e-06, |
| "loss": 2.3468, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.5388446215139442, |
| "grad_norm": 1.0594871044158936, |
| "learning_rate": 5.199992677927775e-06, |
| "loss": 2.2558, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.5398406374501992, |
| "grad_norm": 1.3626909255981445, |
| "learning_rate": 5.182610115288296e-06, |
| "loss": 2.4741, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.5408366533864541, |
| "grad_norm": 0.6529922485351562, |
| "learning_rate": 5.165225342361978e-06, |
| "loss": 2.3131, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.5418326693227091, |
| "grad_norm": 0.8616756796836853, |
| "learning_rate": 5.147838569571642e-06, |
| "loss": 2.2786, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.5428286852589641, |
| "grad_norm": 1.9104212522506714, |
| "learning_rate": 5.1304500073643045e-06, |
| "loss": 2.2784, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.5438247011952191, |
| "grad_norm": 0.7414434552192688, |
| "learning_rate": 5.1130598662086514e-06, |
| "loss": 2.1551, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.5448207171314741, |
| "grad_norm": 0.889681339263916, |
| "learning_rate": 5.095668356592474e-06, |
| "loss": 1.9545, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.545816733067729, |
| "grad_norm": 0.8562681674957275, |
| "learning_rate": 5.078275689020129e-06, |
| "loss": 2.1665, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.546812749003984, |
| "grad_norm": 0.810234010219574, |
| "learning_rate": 5.060882074009988e-06, |
| "loss": 2.5324, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.547808764940239, |
| "grad_norm": 1.867493748664856, |
| "learning_rate": 5.043487722091891e-06, |
| "loss": 2.6972, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.548804780876494, |
| "grad_norm": 0.8987257480621338, |
| "learning_rate": 5.026092843804599e-06, |
| "loss": 2.3632, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.549800796812749, |
| "grad_norm": 0.9021519422531128, |
| "learning_rate": 5.0086976496932374e-06, |
| "loss": 2.2825, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.5507968127490039, |
| "grad_norm": 1.0589499473571777, |
| "learning_rate": 4.991302350306764e-06, |
| "loss": 2.6046, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.5517928286852589, |
| "grad_norm": 0.8575243949890137, |
| "learning_rate": 4.973907156195405e-06, |
| "loss": 2.6962, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.5527888446215139, |
| "grad_norm": 1.7027395963668823, |
| "learning_rate": 4.956512277908109e-06, |
| "loss": 2.4405, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.5537848605577689, |
| "grad_norm": 0.8842843770980835, |
| "learning_rate": 4.939117925990013e-06, |
| "loss": 2.4694, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.5547808764940239, |
| "grad_norm": 0.9840981364250183, |
| "learning_rate": 4.921724310979872e-06, |
| "loss": 2.5034, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.5557768924302788, |
| "grad_norm": 0.8198688626289368, |
| "learning_rate": 4.904331643407527e-06, |
| "loss": 2.5664, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.5567729083665338, |
| "grad_norm": 1.1393426656723022, |
| "learning_rate": 4.886940133791349e-06, |
| "loss": 2.4415, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.5577689243027888, |
| "grad_norm": 0.835932731628418, |
| "learning_rate": 4.869549992635697e-06, |
| "loss": 2.3719, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.5587649402390438, |
| "grad_norm": 0.952167272567749, |
| "learning_rate": 4.8521614304283615e-06, |
| "loss": 2.3092, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.5597609561752988, |
| "grad_norm": 0.8146136999130249, |
| "learning_rate": 4.834774657638023e-06, |
| "loss": 2.5503, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.5607569721115537, |
| "grad_norm": 0.7990023493766785, |
| "learning_rate": 4.817389884711706e-06, |
| "loss": 2.3319, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.5617529880478087, |
| "grad_norm": 0.5947994589805603, |
| "learning_rate": 4.800007322072226e-06, |
| "loss": 2.1591, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.5627490039840638, |
| "grad_norm": 0.8183361291885376, |
| "learning_rate": 4.7826271801156485e-06, |
| "loss": 2.431, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.5637450199203188, |
| "grad_norm": 0.896101713180542, |
| "learning_rate": 4.765249669208733e-06, |
| "loss": 2.3313, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.5647410358565738, |
| "grad_norm": 1.7690149545669556, |
| "learning_rate": 4.747874999686401e-06, |
| "loss": 2.1365, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.5657370517928287, |
| "grad_norm": 0.8712881803512573, |
| "learning_rate": 4.730503381849179e-06, |
| "loss": 2.4139, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.5667330677290837, |
| "grad_norm": 1.0946391820907593, |
| "learning_rate": 4.713135025960652e-06, |
| "loss": 2.1844, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.5677290836653387, |
| "grad_norm": 0.9877662658691406, |
| "learning_rate": 4.695770142244931e-06, |
| "loss": 2.4189, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.5687250996015937, |
| "grad_norm": 3.4181249141693115, |
| "learning_rate": 4.6784089408840955e-06, |
| "loss": 2.7926, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.5697211155378487, |
| "grad_norm": 0.7407424449920654, |
| "learning_rate": 4.661051632015655e-06, |
| "loss": 2.4036, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.5707171314741036, |
| "grad_norm": 0.8603093028068542, |
| "learning_rate": 4.643698425730004e-06, |
| "loss": 2.0895, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.5717131474103586, |
| "grad_norm": 1.5615819692611694, |
| "learning_rate": 4.626349532067879e-06, |
| "loss": 2.6668, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.5727091633466136, |
| "grad_norm": 0.9211438298225403, |
| "learning_rate": 4.609005161017824e-06, |
| "loss": 2.7534, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.5737051792828686, |
| "grad_norm": 0.6863355040550232, |
| "learning_rate": 4.591665522513633e-06, |
| "loss": 2.4823, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.5747011952191236, |
| "grad_norm": 1.1497968435287476, |
| "learning_rate": 4.574330826431822e-06, |
| "loss": 2.7213, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.5756972111553785, |
| "grad_norm": 0.6396723389625549, |
| "learning_rate": 4.557001282589086e-06, |
| "loss": 2.342, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.5766932270916335, |
| "grad_norm": 0.7893930673599243, |
| "learning_rate": 4.5396771007397565e-06, |
| "loss": 2.426, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.5776892430278885, |
| "grad_norm": 1.0288350582122803, |
| "learning_rate": 4.5223584905732635e-06, |
| "loss": 2.3415, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.5786852589641435, |
| "grad_norm": 1.215003252029419, |
| "learning_rate": 4.505045661711596e-06, |
| "loss": 2.2311, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.5796812749003984, |
| "grad_norm": 1.2418211698532104, |
| "learning_rate": 4.487738823706775e-06, |
| "loss": 2.6043, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.5806772908366534, |
| "grad_norm": 0.658145546913147, |
| "learning_rate": 4.470438186038301e-06, |
| "loss": 2.5859, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.5816733067729084, |
| "grad_norm": 0.9392750859260559, |
| "learning_rate": 4.4531439581106295e-06, |
| "loss": 2.4185, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.5826693227091634, |
| "grad_norm": 0.726355254650116, |
| "learning_rate": 4.43585634925063e-06, |
| "loss": 2.656, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.5836653386454184, |
| "grad_norm": 0.8283027410507202, |
| "learning_rate": 4.418575568705066e-06, |
| "loss": 2.6963, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.5846613545816733, |
| "grad_norm": 2.0478458404541016, |
| "learning_rate": 4.401301825638039e-06, |
| "loss": 2.8958, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.5856573705179283, |
| "grad_norm": 0.9227387309074402, |
| "learning_rate": 4.3840353291284776e-06, |
| "loss": 2.5498, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.5866533864541833, |
| "grad_norm": 1.2917126417160034, |
| "learning_rate": 4.366776288167598e-06, |
| "loss": 2.4996, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.5876494023904383, |
| "grad_norm": 0.9034551978111267, |
| "learning_rate": 4.349524911656373e-06, |
| "loss": 2.4281, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.5886454183266933, |
| "grad_norm": 1.010878324508667, |
| "learning_rate": 4.332281408403011e-06, |
| "loss": 2.2912, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.5896414342629482, |
| "grad_norm": 1.7873375415802002, |
| "learning_rate": 4.315045987120417e-06, |
| "loss": 2.5171, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.5906374501992032, |
| "grad_norm": 0.8005262613296509, |
| "learning_rate": 4.297818856423679e-06, |
| "loss": 2.33, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.5916334661354582, |
| "grad_norm": 0.6728765368461609, |
| "learning_rate": 4.280600224827533e-06, |
| "loss": 2.1524, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.5926294820717132, |
| "grad_norm": 0.8610662221908569, |
| "learning_rate": 4.2633903007438445e-06, |
| "loss": 2.7524, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.5936254980079682, |
| "grad_norm": 1.0232973098754883, |
| "learning_rate": 4.2461892924790825e-06, |
| "loss": 2.6512, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.5946215139442231, |
| "grad_norm": 1.0610368251800537, |
| "learning_rate": 4.228997408231806e-06, |
| "loss": 2.4996, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.5956175298804781, |
| "grad_norm": 1.2796133756637573, |
| "learning_rate": 4.2118148560901325e-06, |
| "loss": 2.2488, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.5966135458167331, |
| "grad_norm": 1.5423349142074585, |
| "learning_rate": 4.194641844029227e-06, |
| "loss": 2.6293, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.5976095617529881, |
| "grad_norm": 1.5228114128112793, |
| "learning_rate": 4.1774785799087805e-06, |
| "loss": 2.3751, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.598605577689243, |
| "grad_norm": 0.9803175330162048, |
| "learning_rate": 4.160325271470502e-06, |
| "loss": 2.5003, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.599601593625498, |
| "grad_norm": 1.0139139890670776, |
| "learning_rate": 4.143182126335594e-06, |
| "loss": 2.5435, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.600597609561753, |
| "grad_norm": 0.8577011227607727, |
| "learning_rate": 4.12604935200224e-06, |
| "loss": 2.6227, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.601593625498008, |
| "grad_norm": 1.015549659729004, |
| "learning_rate": 4.108927155843108e-06, |
| "loss": 2.3803, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.602589641434263, |
| "grad_norm": 1.5675932168960571, |
| "learning_rate": 4.091815745102818e-06, |
| "loss": 2.364, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.603585657370518, |
| "grad_norm": 0.7266266345977783, |
| "learning_rate": 4.074715326895453e-06, |
| "loss": 2.6198, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.6045816733067729, |
| "grad_norm": 0.617874026298523, |
| "learning_rate": 4.0576261082020345e-06, |
| "loss": 2.6864, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.6055776892430279, |
| "grad_norm": 1.3024420738220215, |
| "learning_rate": 4.040548295868039e-06, |
| "loss": 2.4742, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.6065737051792829, |
| "grad_norm": 0.7109612822532654, |
| "learning_rate": 4.023482096600873e-06, |
| "loss": 2.6456, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.6075697211155379, |
| "grad_norm": 1.2119102478027344, |
| "learning_rate": 4.006427716967382e-06, |
| "loss": 2.3745, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.6085657370517928, |
| "grad_norm": 1.3014880418777466, |
| "learning_rate": 3.9893853633913485e-06, |
| "loss": 2.6845, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.6095617529880478, |
| "grad_norm": 1.3628534078598022, |
| "learning_rate": 3.9723552421509975e-06, |
| "loss": 2.6973, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.6105577689243028, |
| "grad_norm": 0.9777284860610962, |
| "learning_rate": 3.955337559376489e-06, |
| "loss": 2.4989, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.6115537848605578, |
| "grad_norm": 0.6070024371147156, |
| "learning_rate": 3.938332521047434e-06, |
| "loss": 2.0082, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.6125498007968128, |
| "grad_norm": 0.6223677396774292, |
| "learning_rate": 3.921340332990392e-06, |
| "loss": 2.2016, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.6135458167330677, |
| "grad_norm": 1.2076197862625122, |
| "learning_rate": 3.904361200876391e-06, |
| "loss": 2.7328, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.6145418326693227, |
| "grad_norm": 0.7502063512802124, |
| "learning_rate": 3.887395330218429e-06, |
| "loss": 2.1634, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.6155378486055777, |
| "grad_norm": 1.090084195137024, |
| "learning_rate": 3.8704429263689865e-06, |
| "loss": 2.2409, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.6165338645418327, |
| "grad_norm": 1.7830555438995361, |
| "learning_rate": 3.853504194517551e-06, |
| "loss": 2.5541, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.6175298804780877, |
| "grad_norm": 1.0715655088424683, |
| "learning_rate": 3.836579339688116e-06, |
| "loss": 2.7304, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.6185258964143426, |
| "grad_norm": 0.7255896925926208, |
| "learning_rate": 3.819668566736719e-06, |
| "loss": 2.5671, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.6195219123505976, |
| "grad_norm": 1.475665807723999, |
| "learning_rate": 3.802772080348943e-06, |
| "loss": 2.2374, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.6205179282868526, |
| "grad_norm": 1.1244341135025024, |
| "learning_rate": 3.7858900850374596e-06, |
| "loss": 2.2705, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.6215139442231076, |
| "grad_norm": 1.270950436592102, |
| "learning_rate": 3.769022785139534e-06, |
| "loss": 2.427, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.6225099601593626, |
| "grad_norm": 0.9996942281723022, |
| "learning_rate": 3.752170384814562e-06, |
| "loss": 2.3181, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.6235059760956175, |
| "grad_norm": 0.9702761173248291, |
| "learning_rate": 3.7353330880415963e-06, |
| "loss": 2.4871, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.6245019920318725, |
| "grad_norm": 0.7174897193908691, |
| "learning_rate": 3.7185110986168842e-06, |
| "loss": 2.6481, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.6254980079681275, |
| "grad_norm": 1.0198302268981934, |
| "learning_rate": 3.701704620151389e-06, |
| "loss": 2.4368, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.6264940239043825, |
| "grad_norm": 0.6317278742790222, |
| "learning_rate": 3.6849138560683305e-06, |
| "loss": 2.2506, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.6274900398406374, |
| "grad_norm": 1.6083205938339233, |
| "learning_rate": 3.6681390096007315e-06, |
| "loss": 2.441, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.6284860557768924, |
| "grad_norm": 1.1788543462753296, |
| "learning_rate": 3.651380283788942e-06, |
| "loss": 2.0867, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.6294820717131474, |
| "grad_norm": 1.6041985750198364, |
| "learning_rate": 3.634637881478196e-06, |
| "loss": 2.7786, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.6304780876494024, |
| "grad_norm": 0.7498704195022583, |
| "learning_rate": 3.617912005316142e-06, |
| "loss": 2.5885, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.6314741035856574, |
| "grad_norm": 1.2260042428970337, |
| "learning_rate": 3.6012028577504106e-06, |
| "loss": 2.5491, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.6324701195219123, |
| "grad_norm": 0.766639232635498, |
| "learning_rate": 3.5845106410261417e-06, |
| "loss": 2.6436, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.6334661354581673, |
| "grad_norm": 0.8522284626960754, |
| "learning_rate": 3.56783555718355e-06, |
| "loss": 2.361, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.6344621513944223, |
| "grad_norm": 1.09912109375, |
| "learning_rate": 3.551177808055476e-06, |
| "loss": 2.5303, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.6354581673306773, |
| "grad_norm": 1.4560422897338867, |
| "learning_rate": 3.534537595264944e-06, |
| "loss": 2.6122, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.6364541832669323, |
| "grad_norm": 0.858035147190094, |
| "learning_rate": 3.5179151202227214e-06, |
| "loss": 2.3591, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.6374501992031872, |
| "grad_norm": 4.184999942779541, |
| "learning_rate": 3.5013105841248794e-06, |
| "loss": 2.3339, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.6384462151394422, |
| "grad_norm": 1.2636277675628662, |
| "learning_rate": 3.4847241879503574e-06, |
| "loss": 2.6084, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.6394422310756972, |
| "grad_norm": 1.3735069036483765, |
| "learning_rate": 3.4681561324585356e-06, |
| "loss": 2.4582, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.6404382470119522, |
| "grad_norm": 1.3198506832122803, |
| "learning_rate": 3.451606618186796e-06, |
| "loss": 2.2207, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.6414342629482072, |
| "grad_norm": 0.895077109336853, |
| "learning_rate": 3.435075845448105e-06, |
| "loss": 2.141, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.6424302788844621, |
| "grad_norm": 1.5022435188293457, |
| "learning_rate": 3.418564014328583e-06, |
| "loss": 2.5608, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.6434262948207171, |
| "grad_norm": 0.9838452935218811, |
| "learning_rate": 3.402071324685082e-06, |
| "loss": 2.372, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.6444223107569721, |
| "grad_norm": 0.6322600841522217, |
| "learning_rate": 3.3855979761427705e-06, |
| "loss": 2.5499, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.6454183266932271, |
| "grad_norm": 1.3608890771865845, |
| "learning_rate": 3.3691441680927105e-06, |
| "loss": 2.4689, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.646414342629482, |
| "grad_norm": 0.9520907998085022, |
| "learning_rate": 3.352710099689457e-06, |
| "loss": 2.5012, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.647410358565737, |
| "grad_norm": 3.0419979095458984, |
| "learning_rate": 3.3362959698486307e-06, |
| "loss": 2.2773, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.648406374501992, |
| "grad_norm": 1.0915313959121704, |
| "learning_rate": 3.3199019772445253e-06, |
| "loss": 2.3744, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.649402390438247, |
| "grad_norm": 1.543050765991211, |
| "learning_rate": 3.3035283203076877e-06, |
| "loss": 2.3499, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.650398406374502, |
| "grad_norm": 1.0574357509613037, |
| "learning_rate": 3.287175197222537e-06, |
| "loss": 2.516, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.651394422310757, |
| "grad_norm": 1.3127410411834717, |
| "learning_rate": 3.2708428059249437e-06, |
| "loss": 2.4012, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.6523904382470119, |
| "grad_norm": 0.9456487894058228, |
| "learning_rate": 3.254531344099847e-06, |
| "loss": 2.773, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.6533864541832669, |
| "grad_norm": 1.1509116888046265, |
| "learning_rate": 3.2382410091788567e-06, |
| "loss": 2.7622, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.6543824701195219, |
| "grad_norm": 1.0328110456466675, |
| "learning_rate": 3.221971998337872e-06, |
| "loss": 2.5343, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.6553784860557769, |
| "grad_norm": 1.723029613494873, |
| "learning_rate": 3.2057245084946796e-06, |
| "loss": 2.8968, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.6563745019920318, |
| "grad_norm": 1.13263738155365, |
| "learning_rate": 3.189498736306584e-06, |
| "loss": 2.4689, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.6573705179282868, |
| "grad_norm": 1.4411126375198364, |
| "learning_rate": 3.173294878168025e-06, |
| "loss": 2.0715, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.6583665338645418, |
| "grad_norm": 2.4463119506835938, |
| "learning_rate": 3.1571131302081916e-06, |
| "loss": 2.5004, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.6593625498007968, |
| "grad_norm": 1.063270926475525, |
| "learning_rate": 3.140953688288658e-06, |
| "loss": 2.4079, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.6603585657370518, |
| "grad_norm": 1.860757827758789, |
| "learning_rate": 3.1248167480010083e-06, |
| "loss": 2.6755, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.6613545816733067, |
| "grad_norm": 1.5435043573379517, |
| "learning_rate": 3.1087025046644704e-06, |
| "loss": 2.7499, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.6623505976095617, |
| "grad_norm": 0.7970728278160095, |
| "learning_rate": 3.0926111533235526e-06, |
| "loss": 2.1911, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.6633466135458167, |
| "grad_norm": 1.1135482788085938, |
| "learning_rate": 3.0765428887456794e-06, |
| "loss": 2.6387, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.6643426294820717, |
| "grad_norm": 1.2876728773117065, |
| "learning_rate": 3.0604979054188367e-06, |
| "loss": 2.6715, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.6653386454183267, |
| "grad_norm": 0.6579734683036804, |
| "learning_rate": 3.044476397549221e-06, |
| "loss": 2.1833, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.6663346613545816, |
| "grad_norm": 1.7546638250350952, |
| "learning_rate": 3.0284785590588804e-06, |
| "loss": 2.5761, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.6673306772908366, |
| "grad_norm": 1.1617887020111084, |
| "learning_rate": 3.012504583583374e-06, |
| "loss": 2.4205, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.6683266932270916, |
| "grad_norm": 1.4457294940948486, |
| "learning_rate": 2.9965546644694287e-06, |
| "loss": 2.178, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.6693227091633466, |
| "grad_norm": 0.9334515333175659, |
| "learning_rate": 2.9806289947725947e-06, |
| "loss": 2.5343, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.6703187250996016, |
| "grad_norm": 1.115212082862854, |
| "learning_rate": 2.9647277672549093e-06, |
| "loss": 2.1731, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.6713147410358565, |
| "grad_norm": 1.1038217544555664, |
| "learning_rate": 2.948851174382565e-06, |
| "loss": 2.3589, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.6723107569721115, |
| "grad_norm": 1.4897500276565552, |
| "learning_rate": 2.9329994083235857e-06, |
| "loss": 2.4302, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.6733067729083665, |
| "grad_norm": 1.7196754217147827, |
| "learning_rate": 2.9171726609454875e-06, |
| "loss": 2.5387, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.6743027888446215, |
| "grad_norm": 1.271872878074646, |
| "learning_rate": 2.9013711238129693e-06, |
| "loss": 2.1938, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.6752988047808764, |
| "grad_norm": 1.0383085012435913, |
| "learning_rate": 2.885594988185587e-06, |
| "loss": 2.5842, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.6762948207171314, |
| "grad_norm": 1.9233471155166626, |
| "learning_rate": 2.8698444450154395e-06, |
| "loss": 2.4421, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.6772908366533864, |
| "grad_norm": 0.9812890291213989, |
| "learning_rate": 2.8541196849448582e-06, |
| "loss": 2.1876, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.6782868525896414, |
| "grad_norm": 1.217011570930481, |
| "learning_rate": 2.8384208983040997e-06, |
| "loss": 2.5115, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.6792828685258964, |
| "grad_norm": 1.4093648195266724, |
| "learning_rate": 2.8227482751090445e-06, |
| "loss": 2.5296, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.6802788844621513, |
| "grad_norm": 1.0479772090911865, |
| "learning_rate": 2.8071020050588927e-06, |
| "loss": 2.3801, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.6812749003984063, |
| "grad_norm": 0.761779248714447, |
| "learning_rate": 2.7914822775338678e-06, |
| "loss": 2.397, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.6822709163346613, |
| "grad_norm": 0.7536188364028931, |
| "learning_rate": 2.775889281592927e-06, |
| "loss": 2.2802, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.6832669322709163, |
| "grad_norm": 1.1621276140213013, |
| "learning_rate": 2.760323205971476e-06, |
| "loss": 2.3802, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.6842629482071713, |
| "grad_norm": 1.2401965856552124, |
| "learning_rate": 2.744784239079077e-06, |
| "loss": 1.9567, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.6852589641434262, |
| "grad_norm": 0.9456545114517212, |
| "learning_rate": 2.7292725689971732e-06, |
| "loss": 2.547, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.6862549800796812, |
| "grad_norm": 1.3253943920135498, |
| "learning_rate": 2.7137883834768076e-06, |
| "loss": 2.2105, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.6872509960159362, |
| "grad_norm": 1.525397777557373, |
| "learning_rate": 2.6983318699363627e-06, |
| "loss": 2.3682, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.6882470119521913, |
| "grad_norm": 0.9517590403556824, |
| "learning_rate": 2.6829032154592745e-06, |
| "loss": 2.2159, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.6892430278884463, |
| "grad_norm": 0.8040021061897278, |
| "learning_rate": 2.6675026067917808e-06, |
| "loss": 2.3967, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.6902390438247012, |
| "grad_norm": 1.6833242177963257, |
| "learning_rate": 2.652130230340655e-06, |
| "loss": 2.9864, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.6912350597609562, |
| "grad_norm": 1.104771614074707, |
| "learning_rate": 2.636786272170956e-06, |
| "loss": 2.5124, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.6922310756972112, |
| "grad_norm": 1.1763907670974731, |
| "learning_rate": 2.621470918003768e-06, |
| "loss": 2.6426, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.6932270916334662, |
| "grad_norm": 2.304222345352173, |
| "learning_rate": 2.6061843532139563e-06, |
| "loss": 2.4522, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.6942231075697212, |
| "grad_norm": 0.8599796891212463, |
| "learning_rate": 2.5909267628279234e-06, |
| "loss": 2.6796, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.6952191235059761, |
| "grad_norm": 1.0061733722686768, |
| "learning_rate": 2.5756983315213748e-06, |
| "loss": 2.5076, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.6962151394422311, |
| "grad_norm": 1.392606258392334, |
| "learning_rate": 2.560499243617074e-06, |
| "loss": 2.5134, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.6972111553784861, |
| "grad_norm": 1.2116351127624512, |
| "learning_rate": 2.5453296830826135e-06, |
| "loss": 2.0634, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.6982071713147411, |
| "grad_norm": 0.7071558237075806, |
| "learning_rate": 2.5301898335281994e-06, |
| "loss": 2.1104, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.6992031872509961, |
| "grad_norm": 1.8307946920394897, |
| "learning_rate": 2.5150798782044123e-06, |
| "loss": 2.8147, |
| "step": 702 |
| }, |
| { |
| "epoch": 0.700199203187251, |
| "grad_norm": 0.9716182351112366, |
| "learning_rate": 2.5000000000000015e-06, |
| "loss": 2.4836, |
| "step": 703 |
| }, |
| { |
| "epoch": 0.701195219123506, |
| "grad_norm": 0.7655389308929443, |
| "learning_rate": 2.4849503814396624e-06, |
| "loss": 2.2803, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.702191235059761, |
| "grad_norm": 1.1354485750198364, |
| "learning_rate": 2.469931204681841e-06, |
| "loss": 2.5936, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.703187250996016, |
| "grad_norm": 0.9272159337997437, |
| "learning_rate": 2.4549426515165116e-06, |
| "loss": 2.6629, |
| "step": 706 |
| }, |
| { |
| "epoch": 0.704183266932271, |
| "grad_norm": 1.20318603515625, |
| "learning_rate": 2.439984903362988e-06, |
| "loss": 2.341, |
| "step": 707 |
| }, |
| { |
| "epoch": 0.7051792828685259, |
| "grad_norm": 1.0813405513763428, |
| "learning_rate": 2.425058141267722e-06, |
| "loss": 2.5484, |
| "step": 708 |
| }, |
| { |
| "epoch": 0.7061752988047809, |
| "grad_norm": 0.6365978121757507, |
| "learning_rate": 2.4101625459021212e-06, |
| "loss": 2.2276, |
| "step": 709 |
| }, |
| { |
| "epoch": 0.7071713147410359, |
| "grad_norm": 1.4600951671600342, |
| "learning_rate": 2.3952982975603494e-06, |
| "loss": 2.7489, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.7081673306772909, |
| "grad_norm": 1.0905722379684448, |
| "learning_rate": 2.3804655761571517e-06, |
| "loss": 2.7045, |
| "step": 711 |
| }, |
| { |
| "epoch": 0.7091633466135459, |
| "grad_norm": 1.2118492126464844, |
| "learning_rate": 2.3656645612256747e-06, |
| "loss": 2.2625, |
| "step": 712 |
| }, |
| { |
| "epoch": 0.7101593625498008, |
| "grad_norm": 2.2730562686920166, |
| "learning_rate": 2.3508954319153e-06, |
| "loss": 2.9233, |
| "step": 713 |
| }, |
| { |
| "epoch": 0.7111553784860558, |
| "grad_norm": 1.1532260179519653, |
| "learning_rate": 2.3361583669894634e-06, |
| "loss": 2.6882, |
| "step": 714 |
| }, |
| { |
| "epoch": 0.7121513944223108, |
| "grad_norm": 1.74001944065094, |
| "learning_rate": 2.321453544823499e-06, |
| "loss": 2.296, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.7131474103585658, |
| "grad_norm": 1.6281747817993164, |
| "learning_rate": 2.306781143402485e-06, |
| "loss": 2.5453, |
| "step": 716 |
| }, |
| { |
| "epoch": 0.7141434262948207, |
| "grad_norm": 1.2188794612884521, |
| "learning_rate": 2.2921413403190774e-06, |
| "loss": 2.3351, |
| "step": 717 |
| }, |
| { |
| "epoch": 0.7151394422310757, |
| "grad_norm": 1.180245280265808, |
| "learning_rate": 2.2775343127713685e-06, |
| "loss": 2.8909, |
| "step": 718 |
| }, |
| { |
| "epoch": 0.7161354581673307, |
| "grad_norm": 1.207853078842163, |
| "learning_rate": 2.2629602375607373e-06, |
| "loss": 2.2249, |
| "step": 719 |
| }, |
| { |
| "epoch": 0.7171314741035857, |
| "grad_norm": 1.1911535263061523, |
| "learning_rate": 2.24841929108972e-06, |
| "loss": 1.9728, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.7181274900398407, |
| "grad_norm": 2.913970470428467, |
| "learning_rate": 2.23391164935986e-06, |
| "loss": 3.3, |
| "step": 721 |
| }, |
| { |
| "epoch": 0.7191235059760956, |
| "grad_norm": 1.0136604309082031, |
| "learning_rate": 2.219437487969588e-06, |
| "loss": 2.3078, |
| "step": 722 |
| }, |
| { |
| "epoch": 0.7201195219123506, |
| "grad_norm": 1.5370888710021973, |
| "learning_rate": 2.20499698211209e-06, |
| "loss": 2.4226, |
| "step": 723 |
| }, |
| { |
| "epoch": 0.7211155378486056, |
| "grad_norm": 0.8609825968742371, |
| "learning_rate": 2.190590306573198e-06, |
| "loss": 2.5837, |
| "step": 724 |
| }, |
| { |
| "epoch": 0.7221115537848606, |
| "grad_norm": 0.7969903945922852, |
| "learning_rate": 2.1762176357292582e-06, |
| "loss": 2.5065, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.7231075697211156, |
| "grad_norm": 0.6895061135292053, |
| "learning_rate": 2.1618791435450334e-06, |
| "loss": 2.4443, |
| "step": 726 |
| }, |
| { |
| "epoch": 0.7241035856573705, |
| "grad_norm": 1.005803108215332, |
| "learning_rate": 2.1475750035715914e-06, |
| "loss": 2.8449, |
| "step": 727 |
| }, |
| { |
| "epoch": 0.7250996015936255, |
| "grad_norm": 1.464055061340332, |
| "learning_rate": 2.1333053889442033e-06, |
| "loss": 2.436, |
| "step": 728 |
| }, |
| { |
| "epoch": 0.7260956175298805, |
| "grad_norm": 0.7166134715080261, |
| "learning_rate": 2.1190704723802587e-06, |
| "loss": 2.6141, |
| "step": 729 |
| }, |
| { |
| "epoch": 0.7270916334661355, |
| "grad_norm": 1.2269198894500732, |
| "learning_rate": 2.104870426177157e-06, |
| "loss": 2.0039, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.7280876494023905, |
| "grad_norm": 1.233473539352417, |
| "learning_rate": 2.0907054222102367e-06, |
| "loss": 2.4503, |
| "step": 731 |
| }, |
| { |
| "epoch": 0.7290836653386454, |
| "grad_norm": 0.8751947283744812, |
| "learning_rate": 2.0765756319306897e-06, |
| "loss": 2.482, |
| "step": 732 |
| }, |
| { |
| "epoch": 0.7300796812749004, |
| "grad_norm": 1.954285979270935, |
| "learning_rate": 2.0624812263634847e-06, |
| "loss": 2.6237, |
| "step": 733 |
| }, |
| { |
| "epoch": 0.7310756972111554, |
| "grad_norm": 0.7244362235069275, |
| "learning_rate": 2.048422376105299e-06, |
| "loss": 2.3371, |
| "step": 734 |
| }, |
| { |
| "epoch": 0.7320717131474104, |
| "grad_norm": 0.7712534666061401, |
| "learning_rate": 2.034399251322458e-06, |
| "loss": 2.6775, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.7330677290836654, |
| "grad_norm": 1.0466793775558472, |
| "learning_rate": 2.020412021748866e-06, |
| "loss": 2.2773, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.7340637450199203, |
| "grad_norm": 1.417794942855835, |
| "learning_rate": 2.0064608566839584e-06, |
| "loss": 2.7359, |
| "step": 737 |
| }, |
| { |
| "epoch": 0.7350597609561753, |
| "grad_norm": 0.5706871747970581, |
| "learning_rate": 1.9925459249906488e-06, |
| "loss": 2.246, |
| "step": 738 |
| }, |
| { |
| "epoch": 0.7360557768924303, |
| "grad_norm": 1.112219214439392, |
| "learning_rate": 1.978667395093293e-06, |
| "loss": 2.5444, |
| "step": 739 |
| }, |
| { |
| "epoch": 0.7370517928286853, |
| "grad_norm": 1.5537924766540527, |
| "learning_rate": 1.964825434975639e-06, |
| "loss": 2.2497, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.7380478087649402, |
| "grad_norm": 0.7418034672737122, |
| "learning_rate": 1.9510202121788003e-06, |
| "loss": 2.4711, |
| "step": 741 |
| }, |
| { |
| "epoch": 0.7390438247011952, |
| "grad_norm": 1.0376439094543457, |
| "learning_rate": 1.9372518937992306e-06, |
| "loss": 2.1369, |
| "step": 742 |
| }, |
| { |
| "epoch": 0.7400398406374502, |
| "grad_norm": 1.6938295364379883, |
| "learning_rate": 1.923520646486695e-06, |
| "loss": 2.7013, |
| "step": 743 |
| }, |
| { |
| "epoch": 0.7410358565737052, |
| "grad_norm": 1.1227657794952393, |
| "learning_rate": 1.9098266364422554e-06, |
| "loss": 2.1956, |
| "step": 744 |
| }, |
| { |
| "epoch": 0.7420318725099602, |
| "grad_norm": 0.8521560430526733, |
| "learning_rate": 1.8961700294162578e-06, |
| "loss": 2.7621, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.7430278884462151, |
| "grad_norm": 1.3367222547531128, |
| "learning_rate": 1.8825509907063328e-06, |
| "loss": 2.3669, |
| "step": 746 |
| }, |
| { |
| "epoch": 0.7440239043824701, |
| "grad_norm": 1.0971968173980713, |
| "learning_rate": 1.8689696851553847e-06, |
| "loss": 2.2727, |
| "step": 747 |
| }, |
| { |
| "epoch": 0.7450199203187251, |
| "grad_norm": 0.7232230305671692, |
| "learning_rate": 1.8554262771496017e-06, |
| "loss": 2.4247, |
| "step": 748 |
| }, |
| { |
| "epoch": 0.7460159362549801, |
| "grad_norm": 0.779901921749115, |
| "learning_rate": 1.8419209306164653e-06, |
| "loss": 2.4956, |
| "step": 749 |
| }, |
| { |
| "epoch": 0.7470119521912351, |
| "grad_norm": 0.9150820970535278, |
| "learning_rate": 1.82845380902277e-06, |
| "loss": 2.6319, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.74800796812749, |
| "grad_norm": 1.0264116525650024, |
| "learning_rate": 1.8150250753726363e-06, |
| "loss": 2.537, |
| "step": 751 |
| }, |
| { |
| "epoch": 0.749003984063745, |
| "grad_norm": 1.3325294256210327, |
| "learning_rate": 1.8016348922055448e-06, |
| "loss": 2.5891, |
| "step": 752 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.9217858910560608, |
| "learning_rate": 1.7882834215943645e-06, |
| "loss": 2.3572, |
| "step": 753 |
| }, |
| { |
| "epoch": 0.750996015936255, |
| "grad_norm": 1.020738959312439, |
| "learning_rate": 1.7749708251433983e-06, |
| "loss": 2.4734, |
| "step": 754 |
| }, |
| { |
| "epoch": 0.75199203187251, |
| "grad_norm": 0.9455721378326416, |
| "learning_rate": 1.7616972639864166e-06, |
| "loss": 2.4533, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.7529880478087649, |
| "grad_norm": 1.7625263929367065, |
| "learning_rate": 1.7484628987847125e-06, |
| "loss": 2.5292, |
| "step": 756 |
| }, |
| { |
| "epoch": 0.7539840637450199, |
| "grad_norm": 1.2456424236297607, |
| "learning_rate": 1.7352678897251606e-06, |
| "loss": 2.5379, |
| "step": 757 |
| }, |
| { |
| "epoch": 0.7549800796812749, |
| "grad_norm": 1.9081121683120728, |
| "learning_rate": 1.7221123965182712e-06, |
| "loss": 2.413, |
| "step": 758 |
| }, |
| { |
| "epoch": 0.7559760956175299, |
| "grad_norm": 1.062225341796875, |
| "learning_rate": 1.7089965783962608e-06, |
| "loss": 2.458, |
| "step": 759 |
| }, |
| { |
| "epoch": 0.7569721115537849, |
| "grad_norm": 1.1116987466812134, |
| "learning_rate": 1.6959205941111228e-06, |
| "loss": 2.4556, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.7579681274900398, |
| "grad_norm": 1.6234967708587646, |
| "learning_rate": 1.6828846019327128e-06, |
| "loss": 2.5499, |
| "step": 761 |
| }, |
| { |
| "epoch": 0.7589641434262948, |
| "grad_norm": 0.741877555847168, |
| "learning_rate": 1.6698887596468232e-06, |
| "loss": 2.1629, |
| "step": 762 |
| }, |
| { |
| "epoch": 0.7599601593625498, |
| "grad_norm": 2.112726926803589, |
| "learning_rate": 1.6569332245532777e-06, |
| "loss": 2.85, |
| "step": 763 |
| }, |
| { |
| "epoch": 0.7609561752988048, |
| "grad_norm": 0.9073076844215393, |
| "learning_rate": 1.6440181534640277e-06, |
| "loss": 2.4195, |
| "step": 764 |
| }, |
| { |
| "epoch": 0.7619521912350598, |
| "grad_norm": 1.2635924816131592, |
| "learning_rate": 1.6311437027012582e-06, |
| "loss": 2.3853, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.7629482071713147, |
| "grad_norm": 1.6688510179519653, |
| "learning_rate": 1.618310028095486e-06, |
| "loss": 2.186, |
| "step": 766 |
| }, |
| { |
| "epoch": 0.7639442231075697, |
| "grad_norm": 1.076957106590271, |
| "learning_rate": 1.6055172849836826e-06, |
| "loss": 2.3887, |
| "step": 767 |
| }, |
| { |
| "epoch": 0.7649402390438247, |
| "grad_norm": 1.0081124305725098, |
| "learning_rate": 1.5927656282073861e-06, |
| "loss": 2.4315, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.7659362549800797, |
| "grad_norm": 1.2986465692520142, |
| "learning_rate": 1.5800552121108392e-06, |
| "loss": 2.6633, |
| "step": 769 |
| }, |
| { |
| "epoch": 0.7669322709163346, |
| "grad_norm": 0.7207338213920593, |
| "learning_rate": 1.567386190539107e-06, |
| "loss": 2.6924, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.7679282868525896, |
| "grad_norm": 0.6458574533462524, |
| "learning_rate": 1.5547587168362204e-06, |
| "loss": 2.7688, |
| "step": 771 |
| }, |
| { |
| "epoch": 0.7689243027888446, |
| "grad_norm": 1.0633124113082886, |
| "learning_rate": 1.5421729438433274e-06, |
| "loss": 2.1328, |
| "step": 772 |
| }, |
| { |
| "epoch": 0.7699203187250996, |
| "grad_norm": 2.143666982650757, |
| "learning_rate": 1.5296290238968303e-06, |
| "loss": 2.29, |
| "step": 773 |
| }, |
| { |
| "epoch": 0.7709163346613546, |
| "grad_norm": 0.5651401281356812, |
| "learning_rate": 1.517127108826551e-06, |
| "loss": 2.4732, |
| "step": 774 |
| }, |
| { |
| "epoch": 0.7719123505976095, |
| "grad_norm": 0.8489325642585754, |
| "learning_rate": 1.5046673499538893e-06, |
| "loss": 2.3174, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.7729083665338645, |
| "grad_norm": 1.1251336336135864, |
| "learning_rate": 1.4922498980899907e-06, |
| "loss": 2.2915, |
| "step": 776 |
| }, |
| { |
| "epoch": 0.7739043824701195, |
| "grad_norm": 0.7484387755393982, |
| "learning_rate": 1.4798749035339278e-06, |
| "loss": 2.3685, |
| "step": 777 |
| }, |
| { |
| "epoch": 0.7749003984063745, |
| "grad_norm": 1.1463130712509155, |
| "learning_rate": 1.4675425160708723e-06, |
| "loss": 2.468, |
| "step": 778 |
| }, |
| { |
| "epoch": 0.7758964143426295, |
| "grad_norm": 1.5645790100097656, |
| "learning_rate": 1.4552528849702852e-06, |
| "loss": 2.6442, |
| "step": 779 |
| }, |
| { |
| "epoch": 0.7768924302788844, |
| "grad_norm": 1.8811829090118408, |
| "learning_rate": 1.4430061589841122e-06, |
| "loss": 2.5609, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.7778884462151394, |
| "grad_norm": 0.8737534284591675, |
| "learning_rate": 1.4308024863449805e-06, |
| "loss": 2.6824, |
| "step": 781 |
| }, |
| { |
| "epoch": 0.7788844621513944, |
| "grad_norm": 1.1957892179489136, |
| "learning_rate": 1.4186420147644053e-06, |
| "loss": 2.3529, |
| "step": 782 |
| }, |
| { |
| "epoch": 0.7798804780876494, |
| "grad_norm": 1.2302711009979248, |
| "learning_rate": 1.4065248914310066e-06, |
| "loss": 2.513, |
| "step": 783 |
| }, |
| { |
| "epoch": 0.7808764940239044, |
| "grad_norm": 0.5240752100944519, |
| "learning_rate": 1.3944512630087182e-06, |
| "loss": 2.4043, |
| "step": 784 |
| }, |
| { |
| "epoch": 0.7818725099601593, |
| "grad_norm": 1.9195410013198853, |
| "learning_rate": 1.3824212756350196e-06, |
| "loss": 2.8095, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.7828685258964143, |
| "grad_norm": 0.9604887962341309, |
| "learning_rate": 1.3704350749191642e-06, |
| "loss": 2.3252, |
| "step": 786 |
| }, |
| { |
| "epoch": 0.7838645418326693, |
| "grad_norm": 0.9721193313598633, |
| "learning_rate": 1.3584928059404207e-06, |
| "loss": 2.4578, |
| "step": 787 |
| }, |
| { |
| "epoch": 0.7848605577689243, |
| "grad_norm": 1.9194726943969727, |
| "learning_rate": 1.3465946132463125e-06, |
| "loss": 2.623, |
| "step": 788 |
| }, |
| { |
| "epoch": 0.7858565737051793, |
| "grad_norm": 1.482784390449524, |
| "learning_rate": 1.3347406408508695e-06, |
| "loss": 2.7708, |
| "step": 789 |
| }, |
| { |
| "epoch": 0.7868525896414342, |
| "grad_norm": 0.7451381683349609, |
| "learning_rate": 1.3229310322328847e-06, |
| "loss": 2.4386, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.7878486055776892, |
| "grad_norm": 0.6679832339286804, |
| "learning_rate": 1.3111659303341824e-06, |
| "loss": 2.37, |
| "step": 791 |
| }, |
| { |
| "epoch": 0.7888446215139442, |
| "grad_norm": 0.8974138498306274, |
| "learning_rate": 1.2994454775578785e-06, |
| "loss": 2.2855, |
| "step": 792 |
| }, |
| { |
| "epoch": 0.7898406374501992, |
| "grad_norm": 1.3459084033966064, |
| "learning_rate": 1.2877698157666663e-06, |
| "loss": 2.7191, |
| "step": 793 |
| }, |
| { |
| "epoch": 0.7908366533864541, |
| "grad_norm": 1.0975403785705566, |
| "learning_rate": 1.2761390862810907e-06, |
| "loss": 2.2521, |
| "step": 794 |
| }, |
| { |
| "epoch": 0.7918326693227091, |
| "grad_norm": 0.9908530712127686, |
| "learning_rate": 1.2645534298778506e-06, |
| "loss": 2.603, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.7928286852589641, |
| "grad_norm": 0.7879658937454224, |
| "learning_rate": 1.253012986788078e-06, |
| "loss": 2.6744, |
| "step": 796 |
| }, |
| { |
| "epoch": 0.7938247011952191, |
| "grad_norm": 0.9611647129058838, |
| "learning_rate": 1.2415178966956531e-06, |
| "loss": 2.3191, |
| "step": 797 |
| }, |
| { |
| "epoch": 0.7948207171314741, |
| "grad_norm": 2.035386085510254, |
| "learning_rate": 1.2300682987355122e-06, |
| "loss": 2.46, |
| "step": 798 |
| }, |
| { |
| "epoch": 0.795816733067729, |
| "grad_norm": 1.4089851379394531, |
| "learning_rate": 1.2186643314919571e-06, |
| "loss": 2.164, |
| "step": 799 |
| }, |
| { |
| "epoch": 0.796812749003984, |
| "grad_norm": 0.944324254989624, |
| "learning_rate": 1.2073061329969843e-06, |
| "loss": 2.4692, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.797808764940239, |
| "grad_norm": 1.0982811450958252, |
| "learning_rate": 1.1959938407286099e-06, |
| "loss": 2.2721, |
| "step": 801 |
| }, |
| { |
| "epoch": 0.798804780876494, |
| "grad_norm": 0.8596687316894531, |
| "learning_rate": 1.1847275916092116e-06, |
| "loss": 2.3065, |
| "step": 802 |
| }, |
| { |
| "epoch": 0.799800796812749, |
| "grad_norm": 0.8045834898948669, |
| "learning_rate": 1.1735075220038634e-06, |
| "loss": 2.1781, |
| "step": 803 |
| }, |
| { |
| "epoch": 0.8007968127490039, |
| "grad_norm": 1.5618336200714111, |
| "learning_rate": 1.1623337677186902e-06, |
| "loss": 2.4166, |
| "step": 804 |
| }, |
| { |
| "epoch": 0.8017928286852589, |
| "grad_norm": 1.3380889892578125, |
| "learning_rate": 1.151206463999222e-06, |
| "loss": 2.5112, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.8027888446215139, |
| "grad_norm": 1.37197744846344, |
| "learning_rate": 1.1401257455287612e-06, |
| "loss": 2.4657, |
| "step": 806 |
| }, |
| { |
| "epoch": 0.8037848605577689, |
| "grad_norm": 1.3427671194076538, |
| "learning_rate": 1.1290917464267458e-06, |
| "loss": 2.5771, |
| "step": 807 |
| }, |
| { |
| "epoch": 0.8047808764940239, |
| "grad_norm": 0.5480353832244873, |
| "learning_rate": 1.1181046002471292e-06, |
| "loss": 2.2615, |
| "step": 808 |
| }, |
| { |
| "epoch": 0.8057768924302788, |
| "grad_norm": 1.03799569606781, |
| "learning_rate": 1.107164439976764e-06, |
| "loss": 2.1258, |
| "step": 809 |
| }, |
| { |
| "epoch": 0.8067729083665338, |
| "grad_norm": 0.874638020992279, |
| "learning_rate": 1.0962713980337947e-06, |
| "loss": 2.3982, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.8077689243027888, |
| "grad_norm": 0.7668205499649048, |
| "learning_rate": 1.085425606266049e-06, |
| "loss": 2.5981, |
| "step": 811 |
| }, |
| { |
| "epoch": 0.8087649402390438, |
| "grad_norm": 1.033339262008667, |
| "learning_rate": 1.0746271959494453e-06, |
| "loss": 2.425, |
| "step": 812 |
| }, |
| { |
| "epoch": 0.8097609561752988, |
| "grad_norm": 0.7686687707901001, |
| "learning_rate": 1.063876297786407e-06, |
| "loss": 2.287, |
| "step": 813 |
| }, |
| { |
| "epoch": 0.8107569721115537, |
| "grad_norm": 0.8868098855018616, |
| "learning_rate": 1.0531730419042736e-06, |
| "loss": 2.4047, |
| "step": 814 |
| }, |
| { |
| "epoch": 0.8117529880478087, |
| "grad_norm": 0.9515554308891296, |
| "learning_rate": 1.04251755785373e-06, |
| "loss": 2.2864, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.8127490039840638, |
| "grad_norm": 0.96048903465271, |
| "learning_rate": 1.0319099746072375e-06, |
| "loss": 2.305, |
| "step": 816 |
| }, |
| { |
| "epoch": 0.8137450199203188, |
| "grad_norm": 1.2892875671386719, |
| "learning_rate": 1.0213504205574758e-06, |
| "loss": 2.8046, |
| "step": 817 |
| }, |
| { |
| "epoch": 0.8147410358565738, |
| "grad_norm": 1.2894792556762695, |
| "learning_rate": 1.0108390235157828e-06, |
| "loss": 2.3662, |
| "step": 818 |
| }, |
| { |
| "epoch": 0.8157370517928287, |
| "grad_norm": 0.894437849521637, |
| "learning_rate": 1.0003759107106116e-06, |
| "loss": 2.3213, |
| "step": 819 |
| }, |
| { |
| "epoch": 0.8167330677290837, |
| "grad_norm": 0.8480390906333923, |
| "learning_rate": 9.899612087859883e-07, |
| "loss": 2.2743, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.8177290836653387, |
| "grad_norm": 0.6957425475120544, |
| "learning_rate": 9.795950437999852e-07, |
| "loss": 2.6014, |
| "step": 821 |
| }, |
| { |
| "epoch": 0.8187250996015937, |
| "grad_norm": 1.1951571702957153, |
| "learning_rate": 9.692775412231863e-07, |
| "loss": 2.5359, |
| "step": 822 |
| }, |
| { |
| "epoch": 0.8197211155378487, |
| "grad_norm": 1.5518149137496948, |
| "learning_rate": 9.590088259371738e-07, |
| "loss": 2.5717, |
| "step": 823 |
| }, |
| { |
| "epoch": 0.8207171314741036, |
| "grad_norm": 0.9850301146507263, |
| "learning_rate": 9.487890222330137e-07, |
| "loss": 2.3225, |
| "step": 824 |
| }, |
| { |
| "epoch": 0.8217131474103586, |
| "grad_norm": 1.0830625295639038, |
| "learning_rate": 9.386182538097582e-07, |
| "loss": 2.49, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.8227091633466136, |
| "grad_norm": 1.1903777122497559, |
| "learning_rate": 9.284966437729387e-07, |
| "loss": 2.3532, |
| "step": 826 |
| }, |
| { |
| "epoch": 0.8237051792828686, |
| "grad_norm": 1.2485320568084717, |
| "learning_rate": 9.184243146330829e-07, |
| "loss": 2.286, |
| "step": 827 |
| }, |
| { |
| "epoch": 0.8247011952191236, |
| "grad_norm": 1.5328834056854248, |
| "learning_rate": 9.084013883042276e-07, |
| "loss": 2.5148, |
| "step": 828 |
| }, |
| { |
| "epoch": 0.8256972111553785, |
| "grad_norm": 1.7866473197937012, |
| "learning_rate": 8.984279861024453e-07, |
| "loss": 2.7636, |
| "step": 829 |
| }, |
| { |
| "epoch": 0.8266932270916335, |
| "grad_norm": 0.8635814785957336, |
| "learning_rate": 8.885042287443785e-07, |
| "loss": 2.579, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.8276892430278885, |
| "grad_norm": 1.480765461921692, |
| "learning_rate": 8.786302363457733e-07, |
| "loss": 2.6228, |
| "step": 831 |
| }, |
| { |
| "epoch": 0.8286852589641435, |
| "grad_norm": 0.6803283095359802, |
| "learning_rate": 8.688061284200266e-07, |
| "loss": 2.4377, |
| "step": 832 |
| }, |
| { |
| "epoch": 0.8296812749003984, |
| "grad_norm": 1.2872114181518555, |
| "learning_rate": 8.590320238767425e-07, |
| "loss": 2.4269, |
| "step": 833 |
| }, |
| { |
| "epoch": 0.8306772908366534, |
| "grad_norm": 1.054971694946289, |
| "learning_rate": 8.493080410202914e-07, |
| "loss": 2.4407, |
| "step": 834 |
| }, |
| { |
| "epoch": 0.8316733067729084, |
| "grad_norm": 0.8980826735496521, |
| "learning_rate": 8.396342975483751e-07, |
| "loss": 2.352, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.8326693227091634, |
| "grad_norm": 0.7561918497085571, |
| "learning_rate": 8.30010910550611e-07, |
| "loss": 2.3562, |
| "step": 836 |
| }, |
| { |
| "epoch": 0.8336653386454184, |
| "grad_norm": 0.6130694150924683, |
| "learning_rate": 8.204379965071036e-07, |
| "loss": 2.3059, |
| "step": 837 |
| }, |
| { |
| "epoch": 0.8346613545816733, |
| "grad_norm": 1.1262505054473877, |
| "learning_rate": 8.109156712870397e-07, |
| "loss": 2.7065, |
| "step": 838 |
| }, |
| { |
| "epoch": 0.8356573705179283, |
| "grad_norm": 2.2383527755737305, |
| "learning_rate": 8.014440501472909e-07, |
| "loss": 2.6112, |
| "step": 839 |
| }, |
| { |
| "epoch": 0.8366533864541833, |
| "grad_norm": 1.0032474994659424, |
| "learning_rate": 7.920232477310102e-07, |
| "loss": 2.4155, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.8376494023904383, |
| "grad_norm": 1.5110204219818115, |
| "learning_rate": 7.826533780662481e-07, |
| "loss": 2.2312, |
| "step": 841 |
| }, |
| { |
| "epoch": 0.8386454183266933, |
| "grad_norm": 1.206811785697937, |
| "learning_rate": 7.733345545645726e-07, |
| "loss": 2.7882, |
| "step": 842 |
| }, |
| { |
| "epoch": 0.8396414342629482, |
| "grad_norm": 1.1954952478408813, |
| "learning_rate": 7.640668900196985e-07, |
| "loss": 2.5765, |
| "step": 843 |
| }, |
| { |
| "epoch": 0.8406374501992032, |
| "grad_norm": 0.8079789280891418, |
| "learning_rate": 7.54850496606117e-07, |
| "loss": 2.0988, |
| "step": 844 |
| }, |
| { |
| "epoch": 0.8416334661354582, |
| "grad_norm": 1.204300880432129, |
| "learning_rate": 7.456854858777418e-07, |
| "loss": 2.6726, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.8426294820717132, |
| "grad_norm": 2.1050732135772705, |
| "learning_rate": 7.365719687665568e-07, |
| "loss": 2.4657, |
| "step": 846 |
| }, |
| { |
| "epoch": 0.8436254980079682, |
| "grad_norm": 1.1028344631195068, |
| "learning_rate": 7.27510055581278e-07, |
| "loss": 2.7165, |
| "step": 847 |
| }, |
| { |
| "epoch": 0.8446215139442231, |
| "grad_norm": 0.9454997777938843, |
| "learning_rate": 7.184998560060114e-07, |
| "loss": 2.4538, |
| "step": 848 |
| }, |
| { |
| "epoch": 0.8456175298804781, |
| "grad_norm": 1.000157117843628, |
| "learning_rate": 7.095414790989292e-07, |
| "loss": 2.2186, |
| "step": 849 |
| }, |
| { |
| "epoch": 0.8466135458167331, |
| "grad_norm": 1.7517778873443604, |
| "learning_rate": 7.006350332909495e-07, |
| "loss": 2.4932, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.8476095617529881, |
| "grad_norm": 1.2720731496810913, |
| "learning_rate": 6.917806263844268e-07, |
| "loss": 2.287, |
| "step": 851 |
| }, |
| { |
| "epoch": 0.848605577689243, |
| "grad_norm": 1.1185457706451416, |
| "learning_rate": 6.829783655518402e-07, |
| "loss": 2.4009, |
| "step": 852 |
| }, |
| { |
| "epoch": 0.849601593625498, |
| "grad_norm": 1.228535771369934, |
| "learning_rate": 6.742283573345004e-07, |
| "loss": 2.4729, |
| "step": 853 |
| }, |
| { |
| "epoch": 0.850597609561753, |
| "grad_norm": 1.3845924139022827, |
| "learning_rate": 6.655307076412637e-07, |
| "loss": 2.5723, |
| "step": 854 |
| }, |
| { |
| "epoch": 0.851593625498008, |
| "grad_norm": 1.2278180122375488, |
| "learning_rate": 6.568855217472425e-07, |
| "loss": 2.4958, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.852589641434263, |
| "grad_norm": 0.7195264101028442, |
| "learning_rate": 6.482929042925363e-07, |
| "loss": 2.5879, |
| "step": 856 |
| }, |
| { |
| "epoch": 0.853585657370518, |
| "grad_norm": 1.1685850620269775, |
| "learning_rate": 6.397529592809615e-07, |
| "loss": 2.3411, |
| "step": 857 |
| }, |
| { |
| "epoch": 0.8545816733067729, |
| "grad_norm": 0.935212254524231, |
| "learning_rate": 6.312657900788e-07, |
| "loss": 2.3266, |
| "step": 858 |
| }, |
| { |
| "epoch": 0.8555776892430279, |
| "grad_norm": 1.0523936748504639, |
| "learning_rate": 6.228314994135376e-07, |
| "loss": 2.5915, |
| "step": 859 |
| }, |
| { |
| "epoch": 0.8565737051792829, |
| "grad_norm": 0.7740164399147034, |
| "learning_rate": 6.14450189372628e-07, |
| "loss": 2.7521, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.8575697211155379, |
| "grad_norm": 2.0991780757904053, |
| "learning_rate": 6.061219614022535e-07, |
| "loss": 2.2285, |
| "step": 861 |
| }, |
| { |
| "epoch": 0.8585657370517928, |
| "grad_norm": 0.990088164806366, |
| "learning_rate": 5.978469163061018e-07, |
| "loss": 2.3091, |
| "step": 862 |
| }, |
| { |
| "epoch": 0.8595617529880478, |
| "grad_norm": 1.3237099647521973, |
| "learning_rate": 5.896251542441395e-07, |
| "loss": 2.4856, |
| "step": 863 |
| }, |
| { |
| "epoch": 0.8605577689243028, |
| "grad_norm": 0.9690184593200684, |
| "learning_rate": 5.814567747314049e-07, |
| "loss": 2.3608, |
| "step": 864 |
| }, |
| { |
| "epoch": 0.8615537848605578, |
| "grad_norm": 2.3030495643615723, |
| "learning_rate": 5.733418766367988e-07, |
| "loss": 2.4188, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.8625498007968128, |
| "grad_norm": 1.1400495767593384, |
| "learning_rate": 5.652805581818943e-07, |
| "loss": 2.2339, |
| "step": 866 |
| }, |
| { |
| "epoch": 0.8635458167330677, |
| "grad_norm": 1.5023630857467651, |
| "learning_rate": 5.572729169397422e-07, |
| "loss": 2.1393, |
| "step": 867 |
| }, |
| { |
| "epoch": 0.8645418326693227, |
| "grad_norm": 0.9161491394042969, |
| "learning_rate": 5.493190498336903e-07, |
| "loss": 2.3602, |
| "step": 868 |
| }, |
| { |
| "epoch": 0.8655378486055777, |
| "grad_norm": 0.9129965901374817, |
| "learning_rate": 5.414190531362162e-07, |
| "loss": 2.3639, |
| "step": 869 |
| }, |
| { |
| "epoch": 0.8665338645418327, |
| "grad_norm": 1.9090954065322876, |
| "learning_rate": 5.335730224677538e-07, |
| "loss": 2.2505, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.8675298804780877, |
| "grad_norm": 0.8173048496246338, |
| "learning_rate": 5.25781052795541e-07, |
| "loss": 2.2072, |
| "step": 871 |
| }, |
| { |
| "epoch": 0.8685258964143426, |
| "grad_norm": 2.2648603916168213, |
| "learning_rate": 5.180432384324691e-07, |
| "loss": 2.458, |
| "step": 872 |
| }, |
| { |
| "epoch": 0.8695219123505976, |
| "grad_norm": 0.906757652759552, |
| "learning_rate": 5.103596730359428e-07, |
| "loss": 2.5185, |
| "step": 873 |
| }, |
| { |
| "epoch": 0.8705179282868526, |
| "grad_norm": 0.9674282073974609, |
| "learning_rate": 5.027304496067431e-07, |
| "loss": 2.3208, |
| "step": 874 |
| }, |
| { |
| "epoch": 0.8715139442231076, |
| "grad_norm": 1.4019418954849243, |
| "learning_rate": 4.951556604879049e-07, |
| "loss": 2.531, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.8725099601593626, |
| "grad_norm": 1.1170932054519653, |
| "learning_rate": 4.876353973635955e-07, |
| "loss": 2.2724, |
| "step": 876 |
| }, |
| { |
| "epoch": 0.8735059760956175, |
| "grad_norm": 0.8795150518417358, |
| "learning_rate": 4.8016975125801e-07, |
| "loss": 2.4447, |
| "step": 877 |
| }, |
| { |
| "epoch": 0.8745019920318725, |
| "grad_norm": 1.4134328365325928, |
| "learning_rate": 4.727588125342669e-07, |
| "loss": 2.4728, |
| "step": 878 |
| }, |
| { |
| "epoch": 0.8754980079681275, |
| "grad_norm": 0.8537651300430298, |
| "learning_rate": 4.6540267089331294e-07, |
| "loss": 1.8693, |
| "step": 879 |
| }, |
| { |
| "epoch": 0.8764940239043825, |
| "grad_norm": 0.973147988319397, |
| "learning_rate": 4.581014153728386e-07, |
| "loss": 2.6805, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.8774900398406374, |
| "grad_norm": 2.0054306983947754, |
| "learning_rate": 4.508551343462014e-07, |
| "loss": 2.8345, |
| "step": 881 |
| }, |
| { |
| "epoch": 0.8784860557768924, |
| "grad_norm": 1.3600786924362183, |
| "learning_rate": 4.4366391552135567e-07, |
| "loss": 2.4685, |
| "step": 882 |
| }, |
| { |
| "epoch": 0.8794820717131474, |
| "grad_norm": 1.042197823524475, |
| "learning_rate": 4.3652784593978927e-07, |
| "loss": 2.2609, |
| "step": 883 |
| }, |
| { |
| "epoch": 0.8804780876494024, |
| "grad_norm": 1.342214822769165, |
| "learning_rate": 4.29447011975474e-07, |
| "loss": 2.3443, |
| "step": 884 |
| }, |
| { |
| "epoch": 0.8814741035856574, |
| "grad_norm": 1.0559214353561401, |
| "learning_rate": 4.224214993338149e-07, |
| "loss": 2.5197, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.8824701195219123, |
| "grad_norm": 0.9082587361335754, |
| "learning_rate": 4.154513930506171e-07, |
| "loss": 2.4609, |
| "step": 886 |
| }, |
| { |
| "epoch": 0.8834661354581673, |
| "grad_norm": 1.555578589439392, |
| "learning_rate": 4.0853677749105426e-07, |
| "loss": 2.5883, |
| "step": 887 |
| }, |
| { |
| "epoch": 0.8844621513944223, |
| "grad_norm": 0.6700481176376343, |
| "learning_rate": 4.0167773634865017e-07, |
| "loss": 2.4754, |
| "step": 888 |
| }, |
| { |
| "epoch": 0.8854581673306773, |
| "grad_norm": 1.7728710174560547, |
| "learning_rate": 3.9487435264426056e-07, |
| "loss": 2.5669, |
| "step": 889 |
| }, |
| { |
| "epoch": 0.8864541832669323, |
| "grad_norm": 0.8938197493553162, |
| "learning_rate": 3.8812670872507454e-07, |
| "loss": 2.7976, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.8874501992031872, |
| "grad_norm": 1.6636402606964111, |
| "learning_rate": 3.8143488626361135e-07, |
| "loss": 2.4172, |
| "step": 891 |
| }, |
| { |
| "epoch": 0.8884462151394422, |
| "grad_norm": 1.0676062107086182, |
| "learning_rate": 3.747989662567403e-07, |
| "loss": 2.582, |
| "step": 892 |
| }, |
| { |
| "epoch": 0.8894422310756972, |
| "grad_norm": 1.280415654182434, |
| "learning_rate": 3.6821902902469066e-07, |
| "loss": 2.2295, |
| "step": 893 |
| }, |
| { |
| "epoch": 0.8904382470119522, |
| "grad_norm": 1.0814982652664185, |
| "learning_rate": 3.6169515421008494e-07, |
| "loss": 2.7001, |
| "step": 894 |
| }, |
| { |
| "epoch": 0.8914342629482072, |
| "grad_norm": 1.077160120010376, |
| "learning_rate": 3.5522742077697734e-07, |
| "loss": 2.49, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.8924302788844621, |
| "grad_norm": 1.2483303546905518, |
| "learning_rate": 3.4881590700989175e-07, |
| "loss": 2.3397, |
| "step": 896 |
| }, |
| { |
| "epoch": 0.8934262948207171, |
| "grad_norm": 0.6463543772697449, |
| "learning_rate": 3.4246069051287747e-07, |
| "loss": 2.3176, |
| "step": 897 |
| }, |
| { |
| "epoch": 0.8944223107569721, |
| "grad_norm": 0.8918944001197815, |
| "learning_rate": 3.3616184820856936e-07, |
| "loss": 2.5445, |
| "step": 898 |
| }, |
| { |
| "epoch": 0.8954183266932271, |
| "grad_norm": 0.963518500328064, |
| "learning_rate": 3.299194563372604e-07, |
| "loss": 2.3521, |
| "step": 899 |
| }, |
| { |
| "epoch": 0.896414342629482, |
| "grad_norm": 1.8490091562271118, |
| "learning_rate": 3.237335904559713e-07, |
| "loss": 1.8583, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.897410358565737, |
| "grad_norm": 1.2349917888641357, |
| "learning_rate": 3.176043254375422e-07, |
| "loss": 2.5093, |
| "step": 901 |
| }, |
| { |
| "epoch": 0.898406374501992, |
| "grad_norm": 1.3500816822052002, |
| "learning_rate": 3.1153173546972395e-07, |
| "loss": 2.3507, |
| "step": 902 |
| }, |
| { |
| "epoch": 0.899402390438247, |
| "grad_norm": 1.2740628719329834, |
| "learning_rate": 3.055158940542818e-07, |
| "loss": 2.4008, |
| "step": 903 |
| }, |
| { |
| "epoch": 0.900398406374502, |
| "grad_norm": 33.82315444946289, |
| "learning_rate": 2.9955687400610336e-07, |
| "loss": 2.3794, |
| "step": 904 |
| }, |
| { |
| "epoch": 0.901394422310757, |
| "grad_norm": 0.8968676328659058, |
| "learning_rate": 2.9365474745231935e-07, |
| "loss": 2.3304, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.9023904382470119, |
| "grad_norm": 1.474859356880188, |
| "learning_rate": 2.878095858314278e-07, |
| "loss": 2.2761, |
| "step": 906 |
| }, |
| { |
| "epoch": 0.9033864541832669, |
| "grad_norm": 1.0338733196258545, |
| "learning_rate": 2.820214598924348e-07, |
| "loss": 2.222, |
| "step": 907 |
| }, |
| { |
| "epoch": 0.9043824701195219, |
| "grad_norm": 1.2461026906967163, |
| "learning_rate": 2.7629043969399193e-07, |
| "loss": 2.129, |
| "step": 908 |
| }, |
| { |
| "epoch": 0.9053784860557769, |
| "grad_norm": 1.0711873769760132, |
| "learning_rate": 2.7061659460355047e-07, |
| "loss": 2.5671, |
| "step": 909 |
| }, |
| { |
| "epoch": 0.9063745019920318, |
| "grad_norm": 2.0772130489349365, |
| "learning_rate": 2.6499999329652525e-07, |
| "loss": 2.4514, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.9073705179282868, |
| "grad_norm": 0.8682186603546143, |
| "learning_rate": 2.594407037554586e-07, |
| "loss": 2.4655, |
| "step": 911 |
| }, |
| { |
| "epoch": 0.9083665338645418, |
| "grad_norm": 2.363909959793091, |
| "learning_rate": 2.539387932691995e-07, |
| "loss": 3.0087, |
| "step": 912 |
| }, |
| { |
| "epoch": 0.9093625498007968, |
| "grad_norm": 1.1324294805526733, |
| "learning_rate": 2.4849432843208786e-07, |
| "loss": 2.6718, |
| "step": 913 |
| }, |
| { |
| "epoch": 0.9103585657370518, |
| "grad_norm": 1.2291409969329834, |
| "learning_rate": 2.431073751431529e-07, |
| "loss": 2.4459, |
| "step": 914 |
| }, |
| { |
| "epoch": 0.9113545816733067, |
| "grad_norm": 1.1384942531585693, |
| "learning_rate": 2.377779986053097e-07, |
| "loss": 2.3891, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.9123505976095617, |
| "grad_norm": 0.9154942631721497, |
| "learning_rate": 2.3250626332457226e-07, |
| "loss": 2.3965, |
| "step": 916 |
| }, |
| { |
| "epoch": 0.9133466135458167, |
| "grad_norm": 3.3389575481414795, |
| "learning_rate": 2.2729223310927473e-07, |
| "loss": 2.4395, |
| "step": 917 |
| }, |
| { |
| "epoch": 0.9143426294820717, |
| "grad_norm": 1.6811953783035278, |
| "learning_rate": 2.2213597106929608e-07, |
| "loss": 2.6017, |
| "step": 918 |
| }, |
| { |
| "epoch": 0.9153386454183267, |
| "grad_norm": 0.8396251201629639, |
| "learning_rate": 2.1703753961529906e-07, |
| "loss": 2.7736, |
| "step": 919 |
| }, |
| { |
| "epoch": 0.9163346613545816, |
| "grad_norm": 1.4626351594924927, |
| "learning_rate": 2.1199700045797077e-07, |
| "loss": 2.2861, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.9173306772908366, |
| "grad_norm": 1.5617847442626953, |
| "learning_rate": 2.070144146072789e-07, |
| "loss": 2.6273, |
| "step": 921 |
| }, |
| { |
| "epoch": 0.9183266932270916, |
| "grad_norm": 1.4151337146759033, |
| "learning_rate": 2.0208984237173546e-07, |
| "loss": 2.1617, |
| "step": 922 |
| }, |
| { |
| "epoch": 0.9193227091633466, |
| "grad_norm": 0.9167352318763733, |
| "learning_rate": 1.9722334335766092e-07, |
| "loss": 2.3805, |
| "step": 923 |
| }, |
| { |
| "epoch": 0.9203187250996016, |
| "grad_norm": 0.840559720993042, |
| "learning_rate": 1.9241497646846463e-07, |
| "loss": 2.3247, |
| "step": 924 |
| }, |
| { |
| "epoch": 0.9213147410358565, |
| "grad_norm": 0.9646689891815186, |
| "learning_rate": 1.876647999039377e-07, |
| "loss": 2.5301, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.9223107569721115, |
| "grad_norm": 0.7615554928779602, |
| "learning_rate": 1.829728711595391e-07, |
| "loss": 2.3153, |
| "step": 926 |
| }, |
| { |
| "epoch": 0.9233067729083665, |
| "grad_norm": 1.138389229774475, |
| "learning_rate": 1.7833924702570725e-07, |
| "loss": 2.573, |
| "step": 927 |
| }, |
| { |
| "epoch": 0.9243027888446215, |
| "grad_norm": 1.0490334033966064, |
| "learning_rate": 1.7376398358716852e-07, |
| "loss": 2.6447, |
| "step": 928 |
| }, |
| { |
| "epoch": 0.9252988047808764, |
| "grad_norm": 0.9057884812355042, |
| "learning_rate": 1.6924713622225975e-07, |
| "loss": 2.6325, |
| "step": 929 |
| }, |
| { |
| "epoch": 0.9262948207171314, |
| "grad_norm": 2.5391931533813477, |
| "learning_rate": 1.6478875960225904e-07, |
| "loss": 2.7326, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.9272908366533864, |
| "grad_norm": 1.9867279529571533, |
| "learning_rate": 1.6038890769072223e-07, |
| "loss": 2.4295, |
| "step": 931 |
| }, |
| { |
| "epoch": 0.9282868525896414, |
| "grad_norm": 0.7426679730415344, |
| "learning_rate": 1.5604763374283073e-07, |
| "loss": 2.3705, |
| "step": 932 |
| }, |
| { |
| "epoch": 0.9292828685258964, |
| "grad_norm": 1.033311367034912, |
| "learning_rate": 1.5176499030474578e-07, |
| "loss": 2.2442, |
| "step": 933 |
| }, |
| { |
| "epoch": 0.9302788844621513, |
| "grad_norm": 1.394702672958374, |
| "learning_rate": 1.4754102921297363e-07, |
| "loss": 2.2417, |
| "step": 934 |
| }, |
| { |
| "epoch": 0.9312749003984063, |
| "grad_norm": 1.109711766242981, |
| "learning_rate": 1.4337580159373864e-07, |
| "loss": 2.1194, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.9322709163346613, |
| "grad_norm": 0.810978353023529, |
| "learning_rate": 1.3926935786236218e-07, |
| "loss": 2.251, |
| "step": 936 |
| }, |
| { |
| "epoch": 0.9332669322709163, |
| "grad_norm": 0.9075368642807007, |
| "learning_rate": 1.3522174772265585e-07, |
| "loss": 2.3295, |
| "step": 937 |
| }, |
| { |
| "epoch": 0.9342629482071713, |
| "grad_norm": 1.0174400806427002, |
| "learning_rate": 1.3123302016631477e-07, |
| "loss": 2.4231, |
| "step": 938 |
| }, |
| { |
| "epoch": 0.9352589641434262, |
| "grad_norm": 1.4848259687423706, |
| "learning_rate": 1.2730322347233037e-07, |
| "loss": 2.327, |
| "step": 939 |
| }, |
| { |
| "epoch": 0.9362549800796812, |
| "grad_norm": 0.7531813979148865, |
| "learning_rate": 1.2343240520640287e-07, |
| "loss": 2.5221, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.9372509960159362, |
| "grad_norm": 2.318554162979126, |
| "learning_rate": 1.196206122203647e-07, |
| "loss": 2.3946, |
| "step": 941 |
| }, |
| { |
| "epoch": 0.9382470119521913, |
| "grad_norm": 0.8973721861839294, |
| "learning_rate": 1.158678906516153e-07, |
| "loss": 2.1919, |
| "step": 942 |
| }, |
| { |
| "epoch": 0.9392430278884463, |
| "grad_norm": 1.4439376592636108, |
| "learning_rate": 1.1217428592256218e-07, |
| "loss": 2.3653, |
| "step": 943 |
| }, |
| { |
| "epoch": 0.9402390438247012, |
| "grad_norm": 1.6569935083389282, |
| "learning_rate": 1.0853984274007246e-07, |
| "loss": 2.6982, |
| "step": 944 |
| }, |
| { |
| "epoch": 0.9412350597609562, |
| "grad_norm": 1.5078299045562744, |
| "learning_rate": 1.0496460509492767e-07, |
| "loss": 2.4708, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.9422310756972112, |
| "grad_norm": 2.2158310413360596, |
| "learning_rate": 1.0144861626129599e-07, |
| "loss": 2.5999, |
| "step": 946 |
| }, |
| { |
| "epoch": 0.9432270916334662, |
| "grad_norm": 1.4059021472930908, |
| "learning_rate": 9.799191879620474e-08, |
| "loss": 2.4373, |
| "step": 947 |
| }, |
| { |
| "epoch": 0.9442231075697212, |
| "grad_norm": 2.9164271354675293, |
| "learning_rate": 9.459455453902866e-08, |
| "loss": 2.6697, |
| "step": 948 |
| }, |
| { |
| "epoch": 0.9452191235059761, |
| "grad_norm": 1.275817632675171, |
| "learning_rate": 9.125656461098142e-08, |
| "loss": 2.48, |
| "step": 949 |
| }, |
| { |
| "epoch": 0.9462151394422311, |
| "grad_norm": 0.821499764919281, |
| "learning_rate": 8.797798941461655e-08, |
| "loss": 2.4301, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.9472111553784861, |
| "grad_norm": 1.4214954376220703, |
| "learning_rate": 8.475886863334282e-08, |
| "loss": 2.2847, |
| "step": 951 |
| }, |
| { |
| "epoch": 0.9482071713147411, |
| "grad_norm": 0.9953071475028992, |
| "learning_rate": 8.15992412309391e-08, |
| "loss": 2.464, |
| "step": 952 |
| }, |
| { |
| "epoch": 0.9492031872509961, |
| "grad_norm": 1.4973928928375244, |
| "learning_rate": 7.84991454510864e-08, |
| "loss": 2.0293, |
| "step": 953 |
| }, |
| { |
| "epoch": 0.950199203187251, |
| "grad_norm": 1.326232671737671, |
| "learning_rate": 7.545861881690097e-08, |
| "loss": 2.4445, |
| "step": 954 |
| }, |
| { |
| "epoch": 0.951195219123506, |
| "grad_norm": 1.0779331922531128, |
| "learning_rate": 7.247769813048644e-08, |
| "loss": 2.7232, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.952191235059761, |
| "grad_norm": 1.0593082904815674, |
| "learning_rate": 6.955641947248127e-08, |
| "loss": 2.8634, |
| "step": 956 |
| }, |
| { |
| "epoch": 0.953187250996016, |
| "grad_norm": 0.8761929869651794, |
| "learning_rate": 6.669481820162638e-08, |
| "loss": 2.3328, |
| "step": 957 |
| }, |
| { |
| "epoch": 0.954183266932271, |
| "grad_norm": 0.9143054485321045, |
| "learning_rate": 6.389292895433608e-08, |
| "loss": 2.2261, |
| "step": 958 |
| }, |
| { |
| "epoch": 0.9551792828685259, |
| "grad_norm": 0.8204777240753174, |
| "learning_rate": 6.115078564427946e-08, |
| "loss": 2.4155, |
| "step": 959 |
| }, |
| { |
| "epoch": 0.9561752988047809, |
| "grad_norm": 0.7546234726905823, |
| "learning_rate": 5.8468421461968517e-08, |
| "loss": 2.267, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.9571713147410359, |
| "grad_norm": 1.0376356840133667, |
| "learning_rate": 5.584586887435739e-08, |
| "loss": 2.4496, |
| "step": 961 |
| }, |
| { |
| "epoch": 0.9581673306772909, |
| "grad_norm": 0.6573870182037354, |
| "learning_rate": 5.3283159624448745e-08, |
| "loss": 2.4904, |
| "step": 962 |
| }, |
| { |
| "epoch": 0.9591633466135459, |
| "grad_norm": 1.3613762855529785, |
| "learning_rate": 5.0780324730911877e-08, |
| "loss": 2.4824, |
| "step": 963 |
| }, |
| { |
| "epoch": 0.9601593625498008, |
| "grad_norm": 1.4304169416427612, |
| "learning_rate": 4.833739448770247e-08, |
| "loss": 2.9062, |
| "step": 964 |
| }, |
| { |
| "epoch": 0.9611553784860558, |
| "grad_norm": 0.9547715783119202, |
| "learning_rate": 4.5954398463700647e-08, |
| "loss": 2.4977, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.9621513944223108, |
| "grad_norm": 1.3909553289413452, |
| "learning_rate": 4.3631365502351805e-08, |
| "loss": 2.2116, |
| "step": 966 |
| }, |
| { |
| "epoch": 0.9631474103585658, |
| "grad_norm": 0.7387050986289978, |
| "learning_rate": 4.136832372131583e-08, |
| "loss": 2.5225, |
| "step": 967 |
| }, |
| { |
| "epoch": 0.9641434262948207, |
| "grad_norm": 1.2469770908355713, |
| "learning_rate": 3.916530051212841e-08, |
| "loss": 2.4759, |
| "step": 968 |
| }, |
| { |
| "epoch": 0.9651394422310757, |
| "grad_norm": 1.3780826330184937, |
| "learning_rate": 3.702232253986804e-08, |
| "loss": 2.0538, |
| "step": 969 |
| }, |
| { |
| "epoch": 0.9661354581673307, |
| "grad_norm": 0.9699292778968811, |
| "learning_rate": 3.4939415742835655e-08, |
| "loss": 2.3441, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.9671314741035857, |
| "grad_norm": 1.147615671157837, |
| "learning_rate": 3.2916605332238284e-08, |
| "loss": 2.4042, |
| "step": 971 |
| }, |
| { |
| "epoch": 0.9681274900398407, |
| "grad_norm": 1.1738359928131104, |
| "learning_rate": 3.095391579188589e-08, |
| "loss": 2.539, |
| "step": 972 |
| }, |
| { |
| "epoch": 0.9691235059760956, |
| "grad_norm": 0.640042781829834, |
| "learning_rate": 2.9051370877892226e-08, |
| "loss": 2.3044, |
| "step": 973 |
| }, |
| { |
| "epoch": 0.9701195219123506, |
| "grad_norm": 0.7774790525436401, |
| "learning_rate": 2.7208993618390578e-08, |
| "loss": 2.2616, |
| "step": 974 |
| }, |
| { |
| "epoch": 0.9711155378486056, |
| "grad_norm": 1.0389803647994995, |
| "learning_rate": 2.5426806313252895e-08, |
| "loss": 2.4425, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.9721115537848606, |
| "grad_norm": 1.3041914701461792, |
| "learning_rate": 2.370483053382111e-08, |
| "loss": 2.39, |
| "step": 976 |
| }, |
| { |
| "epoch": 0.9731075697211156, |
| "grad_norm": 0.6934490203857422, |
| "learning_rate": 2.2043087122644023e-08, |
| "loss": 2.2232, |
| "step": 977 |
| }, |
| { |
| "epoch": 0.9741035856573705, |
| "grad_norm": 0.6868986487388611, |
| "learning_rate": 2.0441596193227497e-08, |
| "loss": 2.4806, |
| "step": 978 |
| }, |
| { |
| "epoch": 0.9750996015936255, |
| "grad_norm": 1.1924256086349487, |
| "learning_rate": 1.8900377129790205e-08, |
| "loss": 2.4314, |
| "step": 979 |
| }, |
| { |
| "epoch": 0.9760956175298805, |
| "grad_norm": 0.7938891053199768, |
| "learning_rate": 1.741944858702771e-08, |
| "loss": 2.4715, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.9770916334661355, |
| "grad_norm": 0.9900745749473572, |
| "learning_rate": 1.5998828489888762e-08, |
| "loss": 2.0915, |
| "step": 981 |
| }, |
| { |
| "epoch": 0.9780876494023905, |
| "grad_norm": 3.0510518550872803, |
| "learning_rate": 1.4638534033356578e-08, |
| "loss": 3.3239, |
| "step": 982 |
| }, |
| { |
| "epoch": 0.9790836653386454, |
| "grad_norm": 1.1174182891845703, |
| "learning_rate": 1.333858168224178e-08, |
| "loss": 2.1631, |
| "step": 983 |
| }, |
| { |
| "epoch": 0.9800796812749004, |
| "grad_norm": 0.7828091979026794, |
| "learning_rate": 1.2098987170982013e-08, |
| "loss": 2.2998, |
| "step": 984 |
| }, |
| { |
| "epoch": 0.9810756972111554, |
| "grad_norm": 0.8114204406738281, |
| "learning_rate": 1.0919765503453195e-08, |
| "loss": 2.3996, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.9820717131474104, |
| "grad_norm": 0.625230073928833, |
| "learning_rate": 9.800930952786336e-09, |
| "loss": 2.6785, |
| "step": 986 |
| }, |
| { |
| "epoch": 0.9830677290836654, |
| "grad_norm": 1.6107351779937744, |
| "learning_rate": 8.742497061195455e-09, |
| "loss": 2.6999, |
| "step": 987 |
| }, |
| { |
| "epoch": 0.9840637450199203, |
| "grad_norm": 1.5219416618347168, |
| "learning_rate": 7.744476639813814e-09, |
| "loss": 2.3396, |
| "step": 988 |
| }, |
| { |
| "epoch": 0.9850597609561753, |
| "grad_norm": 1.0381386280059814, |
| "learning_rate": 6.806881768539053e-09, |
| "loss": 2.2097, |
| "step": 989 |
| }, |
| { |
| "epoch": 0.9860557768924303, |
| "grad_norm": 1.1309791803359985, |
| "learning_rate": 5.929723795884967e-09, |
| "loss": 2.4901, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.9870517928286853, |
| "grad_norm": 0.979224443435669, |
| "learning_rate": 5.113013338847173e-09, |
| "loss": 2.355, |
| "step": 991 |
| }, |
| { |
| "epoch": 0.9880478087649402, |
| "grad_norm": 0.9343250393867493, |
| "learning_rate": 4.356760282773209e-09, |
| "loss": 2.682, |
| "step": 992 |
| }, |
| { |
| "epoch": 0.9890438247011952, |
| "grad_norm": 4.211667060852051, |
| "learning_rate": 3.660973781242083e-09, |
| "loss": 3.9138, |
| "step": 993 |
| }, |
| { |
| "epoch": 0.9900398406374502, |
| "grad_norm": 0.9402066469192505, |
| "learning_rate": 3.0256622559543537e-09, |
| "loss": 2.1841, |
| "step": 994 |
| }, |
| { |
| "epoch": 0.9910358565737052, |
| "grad_norm": 1.136916995048523, |
| "learning_rate": 2.4508333966305473e-09, |
| "loss": 2.2469, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.9920318725099602, |
| "grad_norm": 1.080809473991394, |
| "learning_rate": 1.936494160916791e-09, |
| "loss": 2.4922, |
| "step": 996 |
| }, |
| { |
| "epoch": 0.9930278884462151, |
| "grad_norm": 0.9956486821174622, |
| "learning_rate": 1.4826507743032071e-09, |
| "loss": 2.5901, |
| "step": 997 |
| }, |
| { |
| "epoch": 0.9940239043824701, |
| "grad_norm": 2.1008529663085938, |
| "learning_rate": 1.089308730043981e-09, |
| "loss": 2.8828, |
| "step": 998 |
| }, |
| { |
| "epoch": 0.9950199203187251, |
| "grad_norm": 1.3167147636413574, |
| "learning_rate": 7.564727890968515e-10, |
| "loss": 2.3331, |
| "step": 999 |
| }, |
| { |
| "epoch": 0.9960159362549801, |
| "grad_norm": 1.1669059991836548, |
| "learning_rate": 4.841469800592746e-10, |
| "loss": 1.9942, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.9970119521912351, |
| "grad_norm": 0.9161001443862915, |
| "learning_rate": 2.723345991245685e-10, |
| "loss": 2.2932, |
| "step": 1001 |
| }, |
| { |
| "epoch": 0.99800796812749, |
| "grad_norm": 0.8516436815261841, |
| "learning_rate": 1.210382100397256e-10, |
| "loss": 2.2065, |
| "step": 1002 |
| }, |
| { |
| "epoch": 0.999003984063745, |
| "grad_norm": 0.7772925496101379, |
| "learning_rate": 3.0259644074326355e-11, |
| "loss": 2.4638, |
| "step": 1003 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 1.244535207748413, |
| "learning_rate": 0.0, |
| "loss": 2.3463, |
| "step": 1004 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1004, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.5282748355775386e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|