{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 3740, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0013377926421404682, "grad_norm": 3.9157129680105482, "learning_rate": 0.0, "loss": 1.7311, "step": 1 }, { "epoch": 0.0026755852842809363, "grad_norm": 4.049782645831512, "learning_rate": 2.6737967914438503e-08, "loss": 1.3927, "step": 2 }, { "epoch": 0.004013377926421404, "grad_norm": 4.212697489080718, "learning_rate": 5.3475935828877005e-08, "loss": 1.3993, "step": 3 }, { "epoch": 0.005351170568561873, "grad_norm": 4.036085706160469, "learning_rate": 8.021390374331552e-08, "loss": 1.5013, "step": 4 }, { "epoch": 0.006688963210702341, "grad_norm": 4.331588281326752, "learning_rate": 1.0695187165775401e-07, "loss": 1.4964, "step": 5 }, { "epoch": 0.008026755852842809, "grad_norm": 4.061367677570281, "learning_rate": 1.3368983957219251e-07, "loss": 1.7278, "step": 6 }, { "epoch": 0.009364548494983277, "grad_norm": 4.64908615833987, "learning_rate": 1.6042780748663104e-07, "loss": 1.7473, "step": 7 }, { "epoch": 0.010702341137123745, "grad_norm": 3.8970270028697254, "learning_rate": 1.8716577540106952e-07, "loss": 1.3707, "step": 8 }, { "epoch": 0.012040133779264214, "grad_norm": 4.759032293575562, "learning_rate": 2.1390374331550802e-07, "loss": 1.4481, "step": 9 }, { "epoch": 0.013377926421404682, "grad_norm": 4.469136095528119, "learning_rate": 2.4064171122994655e-07, "loss": 1.3708, "step": 10 }, { "epoch": 0.01471571906354515, "grad_norm": 4.464271320833816, "learning_rate": 2.6737967914438503e-07, "loss": 1.4898, "step": 11 }, { "epoch": 0.016053511705685617, "grad_norm": 4.506424032244469, "learning_rate": 2.9411764705882356e-07, "loss": 1.5292, "step": 12 }, { "epoch": 0.017391304347826087, "grad_norm": 4.578319076515605, "learning_rate": 3.208556149732621e-07, "loss": 1.5398, "step": 13 }, { "epoch": 0.018729096989966554, "grad_norm": 4.457240144831283, "learning_rate": 3.4759358288770056e-07, "loss": 1.3577, "step": 14 }, { "epoch": 0.020066889632107024, "grad_norm": 4.262045560611147, "learning_rate": 3.7433155080213904e-07, "loss": 1.6834, "step": 15 }, { "epoch": 0.02140468227424749, "grad_norm": 3.7072903332289524, "learning_rate": 4.0106951871657757e-07, "loss": 1.5117, "step": 16 }, { "epoch": 0.02274247491638796, "grad_norm": 4.205803896237137, "learning_rate": 4.2780748663101604e-07, "loss": 1.5635, "step": 17 }, { "epoch": 0.024080267558528427, "grad_norm": 4.493307347404639, "learning_rate": 4.5454545454545457e-07, "loss": 1.731, "step": 18 }, { "epoch": 0.025418060200668897, "grad_norm": 3.726931233704262, "learning_rate": 4.812834224598931e-07, "loss": 1.3936, "step": 19 }, { "epoch": 0.026755852842809364, "grad_norm": 3.5606565544100417, "learning_rate": 5.080213903743316e-07, "loss": 1.5265, "step": 20 }, { "epoch": 0.028093645484949834, "grad_norm": 3.6811814085799126, "learning_rate": 5.347593582887701e-07, "loss": 1.6604, "step": 21 }, { "epoch": 0.0294314381270903, "grad_norm": 3.7180870560331845, "learning_rate": 5.614973262032086e-07, "loss": 1.3211, "step": 22 }, { "epoch": 0.03076923076923077, "grad_norm": 3.66894156025768, "learning_rate": 5.882352941176471e-07, "loss": 1.6908, "step": 23 }, { "epoch": 0.032107023411371234, "grad_norm": 3.3513591211683087, "learning_rate": 6.149732620320856e-07, "loss": 1.4616, "step": 24 }, { "epoch": 0.033444816053511704, "grad_norm": 3.6023000599469612, "learning_rate": 6.417112299465242e-07, "loss": 1.4588, "step": 25 }, { "epoch": 0.034782608695652174, "grad_norm": 2.9456517569403617, "learning_rate": 6.684491978609627e-07, "loss": 1.5313, "step": 26 }, { "epoch": 0.036120401337792644, "grad_norm": 2.92900382905626, "learning_rate": 6.951871657754011e-07, "loss": 1.6627, "step": 27 }, { "epoch": 0.03745819397993311, "grad_norm": 2.4668458145583476, "learning_rate": 7.219251336898397e-07, "loss": 1.6153, "step": 28 }, { "epoch": 0.03879598662207358, "grad_norm": 2.6058143682819543, "learning_rate": 7.486631016042781e-07, "loss": 1.4459, "step": 29 }, { "epoch": 0.04013377926421405, "grad_norm": 2.3539752200838504, "learning_rate": 7.754010695187167e-07, "loss": 1.3755, "step": 30 }, { "epoch": 0.04147157190635452, "grad_norm": 2.485058003179185, "learning_rate": 8.021390374331551e-07, "loss": 1.5149, "step": 31 }, { "epoch": 0.04280936454849498, "grad_norm": 2.6259358209233916, "learning_rate": 8.288770053475937e-07, "loss": 1.5015, "step": 32 }, { "epoch": 0.04414715719063545, "grad_norm": 2.3377621182324746, "learning_rate": 8.556149732620321e-07, "loss": 1.6409, "step": 33 }, { "epoch": 0.04548494983277592, "grad_norm": 2.2223661961123824, "learning_rate": 8.823529411764707e-07, "loss": 1.5505, "step": 34 }, { "epoch": 0.046822742474916385, "grad_norm": 2.241977565470677, "learning_rate": 9.090909090909091e-07, "loss": 1.5814, "step": 35 }, { "epoch": 0.048160535117056855, "grad_norm": 2.052913240447271, "learning_rate": 9.358288770053477e-07, "loss": 1.6391, "step": 36 }, { "epoch": 0.049498327759197325, "grad_norm": 2.0246240201258447, "learning_rate": 9.625668449197862e-07, "loss": 1.4019, "step": 37 }, { "epoch": 0.050836120401337795, "grad_norm": 2.210266030781263, "learning_rate": 9.893048128342248e-07, "loss": 1.4404, "step": 38 }, { "epoch": 0.05217391304347826, "grad_norm": 2.2412850008570406, "learning_rate": 1.0160427807486633e-06, "loss": 1.3527, "step": 39 }, { "epoch": 0.05351170568561873, "grad_norm": 2.1552940528581397, "learning_rate": 1.0427807486631017e-06, "loss": 1.2864, "step": 40 }, { "epoch": 0.0548494983277592, "grad_norm": 2.3103612677395815, "learning_rate": 1.0695187165775401e-06, "loss": 1.4832, "step": 41 }, { "epoch": 0.05618729096989967, "grad_norm": 2.0488148875233025, "learning_rate": 1.0962566844919787e-06, "loss": 1.5442, "step": 42 }, { "epoch": 0.05752508361204013, "grad_norm": 2.176336296355304, "learning_rate": 1.1229946524064172e-06, "loss": 1.3233, "step": 43 }, { "epoch": 0.0588628762541806, "grad_norm": 2.2152172893629016, "learning_rate": 1.1497326203208558e-06, "loss": 1.4669, "step": 44 }, { "epoch": 0.06020066889632107, "grad_norm": 1.9483716964602458, "learning_rate": 1.1764705882352942e-06, "loss": 1.3597, "step": 45 }, { "epoch": 0.06153846153846154, "grad_norm": 1.6640274629340461, "learning_rate": 1.2032085561497326e-06, "loss": 1.2395, "step": 46 }, { "epoch": 0.06287625418060201, "grad_norm": 1.9140782138139933, "learning_rate": 1.2299465240641713e-06, "loss": 1.3121, "step": 47 }, { "epoch": 0.06421404682274247, "grad_norm": 1.5010929655101481, "learning_rate": 1.2566844919786097e-06, "loss": 1.2685, "step": 48 }, { "epoch": 0.06555183946488294, "grad_norm": 1.7243179864941898, "learning_rate": 1.2834224598930483e-06, "loss": 1.5761, "step": 49 }, { "epoch": 0.06688963210702341, "grad_norm": 1.5371591257249981, "learning_rate": 1.3101604278074868e-06, "loss": 1.4041, "step": 50 }, { "epoch": 0.06822742474916388, "grad_norm": 1.5186764141310858, "learning_rate": 1.3368983957219254e-06, "loss": 1.5414, "step": 51 }, { "epoch": 0.06956521739130435, "grad_norm": 1.437650352056891, "learning_rate": 1.3636363636363636e-06, "loss": 1.6225, "step": 52 }, { "epoch": 0.07090301003344482, "grad_norm": 1.315880970269414, "learning_rate": 1.3903743315508022e-06, "loss": 1.3719, "step": 53 }, { "epoch": 0.07224080267558529, "grad_norm": 1.5488345555258765, "learning_rate": 1.4171122994652409e-06, "loss": 1.4724, "step": 54 }, { "epoch": 0.07357859531772576, "grad_norm": 1.2793540534972754, "learning_rate": 1.4438502673796793e-06, "loss": 1.2898, "step": 55 }, { "epoch": 0.07491638795986622, "grad_norm": 1.292007497908228, "learning_rate": 1.4705882352941177e-06, "loss": 1.3846, "step": 56 }, { "epoch": 0.07625418060200669, "grad_norm": 1.4970549423924973, "learning_rate": 1.4973262032085562e-06, "loss": 1.4086, "step": 57 }, { "epoch": 0.07759197324414716, "grad_norm": 1.2614149719222902, "learning_rate": 1.5240641711229948e-06, "loss": 1.3551, "step": 58 }, { "epoch": 0.07892976588628763, "grad_norm": 1.272695339186833, "learning_rate": 1.5508021390374334e-06, "loss": 1.2577, "step": 59 }, { "epoch": 0.0802675585284281, "grad_norm": 1.4282839095950055, "learning_rate": 1.5775401069518716e-06, "loss": 1.3459, "step": 60 }, { "epoch": 0.08160535117056857, "grad_norm": 1.2291019123827382, "learning_rate": 1.6042780748663103e-06, "loss": 1.167, "step": 61 }, { "epoch": 0.08294314381270904, "grad_norm": 1.2579493104162365, "learning_rate": 1.631016042780749e-06, "loss": 1.4053, "step": 62 }, { "epoch": 0.08428093645484949, "grad_norm": 1.1407572115933866, "learning_rate": 1.6577540106951873e-06, "loss": 1.5485, "step": 63 }, { "epoch": 0.08561872909698996, "grad_norm": 1.4625393999954348, "learning_rate": 1.684491978609626e-06, "loss": 1.2999, "step": 64 }, { "epoch": 0.08695652173913043, "grad_norm": 1.40376775796651, "learning_rate": 1.7112299465240642e-06, "loss": 1.2589, "step": 65 }, { "epoch": 0.0882943143812709, "grad_norm": 1.119786008097653, "learning_rate": 1.7379679144385028e-06, "loss": 1.422, "step": 66 }, { "epoch": 0.08963210702341137, "grad_norm": 1.2223872358067744, "learning_rate": 1.7647058823529414e-06, "loss": 1.4599, "step": 67 }, { "epoch": 0.09096989966555184, "grad_norm": 1.1313190991099473, "learning_rate": 1.7914438502673799e-06, "loss": 1.1415, "step": 68 }, { "epoch": 0.09230769230769231, "grad_norm": 1.064362635947767, "learning_rate": 1.8181818181818183e-06, "loss": 1.2546, "step": 69 }, { "epoch": 0.09364548494983277, "grad_norm": 1.1577071430857597, "learning_rate": 1.8449197860962567e-06, "loss": 1.1437, "step": 70 }, { "epoch": 0.09498327759197324, "grad_norm": 1.2945508502036516, "learning_rate": 1.8716577540106954e-06, "loss": 1.091, "step": 71 }, { "epoch": 0.09632107023411371, "grad_norm": 1.238397592005461, "learning_rate": 1.898395721925134e-06, "loss": 1.5206, "step": 72 }, { "epoch": 0.09765886287625418, "grad_norm": 1.0394833819705573, "learning_rate": 1.9251336898395724e-06, "loss": 1.1004, "step": 73 }, { "epoch": 0.09899665551839465, "grad_norm": 1.2764847091521805, "learning_rate": 1.951871657754011e-06, "loss": 1.4334, "step": 74 }, { "epoch": 0.10033444816053512, "grad_norm": 0.9447536784425749, "learning_rate": 1.9786096256684497e-06, "loss": 1.0329, "step": 75 }, { "epoch": 0.10167224080267559, "grad_norm": 1.1485164309192772, "learning_rate": 2.0053475935828877e-06, "loss": 1.5485, "step": 76 }, { "epoch": 0.10301003344481606, "grad_norm": 1.5255866472190742, "learning_rate": 2.0320855614973265e-06, "loss": 1.3535, "step": 77 }, { "epoch": 0.10434782608695652, "grad_norm": 1.006960274565403, "learning_rate": 2.058823529411765e-06, "loss": 1.2198, "step": 78 }, { "epoch": 0.10568561872909699, "grad_norm": 1.1023594195384814, "learning_rate": 2.0855614973262034e-06, "loss": 1.5153, "step": 79 }, { "epoch": 0.10702341137123746, "grad_norm": 1.3392253084497152, "learning_rate": 2.112299465240642e-06, "loss": 1.324, "step": 80 }, { "epoch": 0.10836120401337793, "grad_norm": 1.3249628069750998, "learning_rate": 2.1390374331550802e-06, "loss": 1.4813, "step": 81 }, { "epoch": 0.1096989966555184, "grad_norm": 1.044388153738812, "learning_rate": 2.165775401069519e-06, "loss": 1.5487, "step": 82 }, { "epoch": 0.11103678929765887, "grad_norm": 1.1016043902322221, "learning_rate": 2.1925133689839575e-06, "loss": 1.4622, "step": 83 }, { "epoch": 0.11237458193979934, "grad_norm": 1.194937755180522, "learning_rate": 2.219251336898396e-06, "loss": 1.4046, "step": 84 }, { "epoch": 0.11371237458193979, "grad_norm": 0.9863221636698583, "learning_rate": 2.2459893048128343e-06, "loss": 1.2929, "step": 85 }, { "epoch": 0.11505016722408026, "grad_norm": 0.9981193826727828, "learning_rate": 2.2727272727272728e-06, "loss": 1.573, "step": 86 }, { "epoch": 0.11638795986622073, "grad_norm": 1.0956659394719481, "learning_rate": 2.2994652406417116e-06, "loss": 1.3076, "step": 87 }, { "epoch": 0.1177257525083612, "grad_norm": 1.3268684618382045, "learning_rate": 2.32620320855615e-06, "loss": 1.284, "step": 88 }, { "epoch": 0.11906354515050167, "grad_norm": 1.2824807316851756, "learning_rate": 2.3529411764705885e-06, "loss": 1.5232, "step": 89 }, { "epoch": 0.12040133779264214, "grad_norm": 1.090395382252868, "learning_rate": 2.379679144385027e-06, "loss": 1.5873, "step": 90 }, { "epoch": 0.12173913043478261, "grad_norm": 0.961570824490146, "learning_rate": 2.4064171122994653e-06, "loss": 1.4555, "step": 91 }, { "epoch": 0.12307692307692308, "grad_norm": 1.0844086182237676, "learning_rate": 2.433155080213904e-06, "loss": 1.5585, "step": 92 }, { "epoch": 0.12441471571906354, "grad_norm": 1.0387969165914852, "learning_rate": 2.4598930481283426e-06, "loss": 1.3507, "step": 93 }, { "epoch": 0.12575250836120402, "grad_norm": 1.193870363520448, "learning_rate": 2.486631016042781e-06, "loss": 1.4139, "step": 94 }, { "epoch": 0.12709030100334448, "grad_norm": 1.233428049637306, "learning_rate": 2.5133689839572194e-06, "loss": 1.324, "step": 95 }, { "epoch": 0.12842809364548494, "grad_norm": 0.9058660867208012, "learning_rate": 2.5401069518716583e-06, "loss": 1.3779, "step": 96 }, { "epoch": 0.12976588628762542, "grad_norm": 1.4494523843774527, "learning_rate": 2.5668449197860967e-06, "loss": 1.5868, "step": 97 }, { "epoch": 0.13110367892976588, "grad_norm": 0.9973494547531304, "learning_rate": 2.5935828877005347e-06, "loss": 1.3485, "step": 98 }, { "epoch": 0.13244147157190636, "grad_norm": 0.9253419174739222, "learning_rate": 2.6203208556149735e-06, "loss": 1.051, "step": 99 }, { "epoch": 0.13377926421404682, "grad_norm": 1.0932737193240503, "learning_rate": 2.647058823529412e-06, "loss": 1.5658, "step": 100 }, { "epoch": 0.1351170568561873, "grad_norm": 0.9684334785137063, "learning_rate": 2.673796791443851e-06, "loss": 1.4145, "step": 101 }, { "epoch": 0.13645484949832776, "grad_norm": 0.8760131116929545, "learning_rate": 2.7005347593582892e-06, "loss": 1.2181, "step": 102 }, { "epoch": 0.13779264214046824, "grad_norm": 1.3354854545054526, "learning_rate": 2.7272727272727272e-06, "loss": 1.4523, "step": 103 }, { "epoch": 0.1391304347826087, "grad_norm": 0.9862886301227201, "learning_rate": 2.754010695187166e-06, "loss": 1.4093, "step": 104 }, { "epoch": 0.14046822742474915, "grad_norm": 0.9528355705184072, "learning_rate": 2.7807486631016045e-06, "loss": 1.1985, "step": 105 }, { "epoch": 0.14180602006688964, "grad_norm": 1.092551622625183, "learning_rate": 2.807486631016043e-06, "loss": 1.587, "step": 106 }, { "epoch": 0.1431438127090301, "grad_norm": 1.0617594677821316, "learning_rate": 2.8342245989304818e-06, "loss": 1.3647, "step": 107 }, { "epoch": 0.14448160535117058, "grad_norm": 1.12250031002164, "learning_rate": 2.8609625668449198e-06, "loss": 1.3693, "step": 108 }, { "epoch": 0.14581939799331103, "grad_norm": 1.1221720220339526, "learning_rate": 2.8877005347593586e-06, "loss": 1.2924, "step": 109 }, { "epoch": 0.14715719063545152, "grad_norm": 1.4099168655036936, "learning_rate": 2.914438502673797e-06, "loss": 1.3234, "step": 110 }, { "epoch": 0.14849498327759197, "grad_norm": 1.0799202546626252, "learning_rate": 2.9411764705882355e-06, "loss": 1.5002, "step": 111 }, { "epoch": 0.14983277591973243, "grad_norm": 1.1499380143399163, "learning_rate": 2.9679144385026743e-06, "loss": 1.1992, "step": 112 }, { "epoch": 0.15117056856187291, "grad_norm": 1.1516456091248333, "learning_rate": 2.9946524064171123e-06, "loss": 1.5489, "step": 113 }, { "epoch": 0.15250836120401337, "grad_norm": 1.2012221809667467, "learning_rate": 3.0213903743315507e-06, "loss": 1.2675, "step": 114 }, { "epoch": 0.15384615384615385, "grad_norm": 1.2992030026312718, "learning_rate": 3.0481283422459896e-06, "loss": 1.3792, "step": 115 }, { "epoch": 0.1551839464882943, "grad_norm": 1.3735364854126975, "learning_rate": 3.074866310160428e-06, "loss": 1.3471, "step": 116 }, { "epoch": 0.1565217391304348, "grad_norm": 1.1565233341538352, "learning_rate": 3.101604278074867e-06, "loss": 1.1866, "step": 117 }, { "epoch": 0.15785953177257525, "grad_norm": 0.9715400196509453, "learning_rate": 3.128342245989305e-06, "loss": 1.4531, "step": 118 }, { "epoch": 0.1591973244147157, "grad_norm": 1.0088011823243286, "learning_rate": 3.1550802139037433e-06, "loss": 1.3062, "step": 119 }, { "epoch": 0.1605351170568562, "grad_norm": 0.961411904701495, "learning_rate": 3.181818181818182e-06, "loss": 1.3072, "step": 120 }, { "epoch": 0.16187290969899665, "grad_norm": 1.1048310041829936, "learning_rate": 3.2085561497326205e-06, "loss": 1.5988, "step": 121 }, { "epoch": 0.16321070234113713, "grad_norm": 1.0900999421209308, "learning_rate": 3.2352941176470594e-06, "loss": 1.3049, "step": 122 }, { "epoch": 0.1645484949832776, "grad_norm": 0.9950228472990127, "learning_rate": 3.262032085561498e-06, "loss": 1.2853, "step": 123 }, { "epoch": 0.16588628762541807, "grad_norm": 0.877217867645291, "learning_rate": 3.288770053475936e-06, "loss": 1.1708, "step": 124 }, { "epoch": 0.16722408026755853, "grad_norm": 0.9342888397691156, "learning_rate": 3.3155080213903747e-06, "loss": 1.2191, "step": 125 }, { "epoch": 0.16856187290969898, "grad_norm": 1.162551523878704, "learning_rate": 3.342245989304813e-06, "loss": 1.4792, "step": 126 }, { "epoch": 0.16989966555183947, "grad_norm": 0.9109682170591216, "learning_rate": 3.368983957219252e-06, "loss": 1.1461, "step": 127 }, { "epoch": 0.17123745819397992, "grad_norm": 1.017278635942957, "learning_rate": 3.3957219251336904e-06, "loss": 1.0772, "step": 128 }, { "epoch": 0.1725752508361204, "grad_norm": 1.0628320454739395, "learning_rate": 3.4224598930481284e-06, "loss": 1.3881, "step": 129 }, { "epoch": 0.17391304347826086, "grad_norm": 1.1255887796845363, "learning_rate": 3.449197860962567e-06, "loss": 1.3848, "step": 130 }, { "epoch": 0.17525083612040135, "grad_norm": 1.2168160706758417, "learning_rate": 3.4759358288770056e-06, "loss": 1.5587, "step": 131 }, { "epoch": 0.1765886287625418, "grad_norm": 0.8869532805176883, "learning_rate": 3.5026737967914445e-06, "loss": 1.2815, "step": 132 }, { "epoch": 0.17792642140468226, "grad_norm": 0.938322865914247, "learning_rate": 3.529411764705883e-06, "loss": 1.4471, "step": 133 }, { "epoch": 0.17926421404682275, "grad_norm": 1.0181887221316095, "learning_rate": 3.556149732620321e-06, "loss": 1.3823, "step": 134 }, { "epoch": 0.1806020066889632, "grad_norm": 0.9759359622705379, "learning_rate": 3.5828877005347597e-06, "loss": 1.3245, "step": 135 }, { "epoch": 0.18193979933110369, "grad_norm": 1.1109838742918499, "learning_rate": 3.609625668449198e-06, "loss": 1.2477, "step": 136 }, { "epoch": 0.18327759197324414, "grad_norm": 1.3230085751127434, "learning_rate": 3.6363636363636366e-06, "loss": 0.9357, "step": 137 }, { "epoch": 0.18461538461538463, "grad_norm": 1.4189480206501361, "learning_rate": 3.6631016042780754e-06, "loss": 1.3706, "step": 138 }, { "epoch": 0.18595317725752508, "grad_norm": 0.9586296078564682, "learning_rate": 3.6898395721925134e-06, "loss": 1.3951, "step": 139 }, { "epoch": 0.18729096989966554, "grad_norm": 1.3021663529006131, "learning_rate": 3.716577540106952e-06, "loss": 1.3043, "step": 140 }, { "epoch": 0.18862876254180602, "grad_norm": 0.8556504998042965, "learning_rate": 3.7433155080213907e-06, "loss": 0.9773, "step": 141 }, { "epoch": 0.18996655518394648, "grad_norm": 1.2510634763362827, "learning_rate": 3.770053475935829e-06, "loss": 1.336, "step": 142 }, { "epoch": 0.19130434782608696, "grad_norm": 1.4575730886997018, "learning_rate": 3.796791443850268e-06, "loss": 1.2618, "step": 143 }, { "epoch": 0.19264214046822742, "grad_norm": 1.0839747784663591, "learning_rate": 3.8235294117647055e-06, "loss": 1.1063, "step": 144 }, { "epoch": 0.1939799331103679, "grad_norm": 0.9607295068925171, "learning_rate": 3.850267379679145e-06, "loss": 1.3843, "step": 145 }, { "epoch": 0.19531772575250836, "grad_norm": 1.0962701443846363, "learning_rate": 3.877005347593583e-06, "loss": 1.1803, "step": 146 }, { "epoch": 0.19665551839464884, "grad_norm": 0.9332528904639927, "learning_rate": 3.903743315508022e-06, "loss": 1.2812, "step": 147 }, { "epoch": 0.1979933110367893, "grad_norm": 0.9358629966595745, "learning_rate": 3.93048128342246e-06, "loss": 0.9131, "step": 148 }, { "epoch": 0.19933110367892976, "grad_norm": 1.258263393278227, "learning_rate": 3.957219251336899e-06, "loss": 1.2332, "step": 149 }, { "epoch": 0.20066889632107024, "grad_norm": 1.4757554841146654, "learning_rate": 3.983957219251337e-06, "loss": 1.3761, "step": 150 }, { "epoch": 0.2020066889632107, "grad_norm": 0.9761300744571975, "learning_rate": 4.010695187165775e-06, "loss": 1.2609, "step": 151 }, { "epoch": 0.20334448160535118, "grad_norm": 1.1808835420947998, "learning_rate": 4.037433155080215e-06, "loss": 0.9771, "step": 152 }, { "epoch": 0.20468227424749164, "grad_norm": 1.247054213214014, "learning_rate": 4.064171122994653e-06, "loss": 1.4017, "step": 153 }, { "epoch": 0.20602006688963212, "grad_norm": 0.8864808101130358, "learning_rate": 4.0909090909090915e-06, "loss": 1.2883, "step": 154 }, { "epoch": 0.20735785953177258, "grad_norm": 0.8975249183288195, "learning_rate": 4.11764705882353e-06, "loss": 1.0806, "step": 155 }, { "epoch": 0.20869565217391303, "grad_norm": 0.9253917440547867, "learning_rate": 4.144385026737968e-06, "loss": 1.2307, "step": 156 }, { "epoch": 0.21003344481605352, "grad_norm": 1.013114760136161, "learning_rate": 4.171122994652407e-06, "loss": 1.2244, "step": 157 }, { "epoch": 0.21137123745819397, "grad_norm": 0.913050799297298, "learning_rate": 4.197860962566845e-06, "loss": 1.11, "step": 158 }, { "epoch": 0.21270903010033446, "grad_norm": 1.1076091914325155, "learning_rate": 4.224598930481284e-06, "loss": 1.0744, "step": 159 }, { "epoch": 0.2140468227424749, "grad_norm": 0.9808637517919976, "learning_rate": 4.251336898395722e-06, "loss": 1.0954, "step": 160 }, { "epoch": 0.2153846153846154, "grad_norm": 0.9115123718698217, "learning_rate": 4.2780748663101604e-06, "loss": 1.1236, "step": 161 }, { "epoch": 0.21672240802675585, "grad_norm": 0.9468677486631801, "learning_rate": 4.304812834224599e-06, "loss": 1.3057, "step": 162 }, { "epoch": 0.2180602006688963, "grad_norm": 0.97634947596904, "learning_rate": 4.331550802139038e-06, "loss": 1.3313, "step": 163 }, { "epoch": 0.2193979933110368, "grad_norm": 0.9962808592767697, "learning_rate": 4.3582887700534766e-06, "loss": 1.2327, "step": 164 }, { "epoch": 0.22073578595317725, "grad_norm": 1.1771627304892343, "learning_rate": 4.385026737967915e-06, "loss": 1.0875, "step": 165 }, { "epoch": 0.22207357859531773, "grad_norm": 0.851134643499975, "learning_rate": 4.411764705882353e-06, "loss": 1.2778, "step": 166 }, { "epoch": 0.2234113712374582, "grad_norm": 1.140107363229054, "learning_rate": 4.438502673796792e-06, "loss": 1.0467, "step": 167 }, { "epoch": 0.22474916387959867, "grad_norm": 1.001304154868117, "learning_rate": 4.46524064171123e-06, "loss": 1.0919, "step": 168 }, { "epoch": 0.22608695652173913, "grad_norm": 0.9665701906955589, "learning_rate": 4.491978609625669e-06, "loss": 1.2897, "step": 169 }, { "epoch": 0.22742474916387959, "grad_norm": 1.0298792367324974, "learning_rate": 4.518716577540107e-06, "loss": 1.2398, "step": 170 }, { "epoch": 0.22876254180602007, "grad_norm": 0.9695535035292172, "learning_rate": 4.5454545454545455e-06, "loss": 1.2557, "step": 171 }, { "epoch": 0.23010033444816053, "grad_norm": 0.9456014855854386, "learning_rate": 4.572192513368984e-06, "loss": 0.96, "step": 172 }, { "epoch": 0.231438127090301, "grad_norm": 1.0548525294780986, "learning_rate": 4.598930481283423e-06, "loss": 1.4763, "step": 173 }, { "epoch": 0.23277591973244147, "grad_norm": 1.027718357731009, "learning_rate": 4.625668449197862e-06, "loss": 1.1049, "step": 174 }, { "epoch": 0.23411371237458195, "grad_norm": 0.8406733721638855, "learning_rate": 4.6524064171123e-06, "loss": 1.0567, "step": 175 }, { "epoch": 0.2354515050167224, "grad_norm": 1.3481800370425767, "learning_rate": 4.6791443850267385e-06, "loss": 1.2175, "step": 176 }, { "epoch": 0.23678929765886286, "grad_norm": 0.9282177578411976, "learning_rate": 4.705882352941177e-06, "loss": 1.0576, "step": 177 }, { "epoch": 0.23812709030100335, "grad_norm": 1.0101569467540727, "learning_rate": 4.732620320855615e-06, "loss": 1.5222, "step": 178 }, { "epoch": 0.2394648829431438, "grad_norm": 0.9565148626998742, "learning_rate": 4.759358288770054e-06, "loss": 1.0428, "step": 179 }, { "epoch": 0.2408026755852843, "grad_norm": 0.9131481936151572, "learning_rate": 4.786096256684493e-06, "loss": 1.1425, "step": 180 }, { "epoch": 0.24214046822742474, "grad_norm": 1.0125438636949295, "learning_rate": 4.812834224598931e-06, "loss": 1.1828, "step": 181 }, { "epoch": 0.24347826086956523, "grad_norm": 1.067416973901263, "learning_rate": 4.839572192513369e-06, "loss": 1.2701, "step": 182 }, { "epoch": 0.24481605351170568, "grad_norm": 0.9254029396392351, "learning_rate": 4.866310160427808e-06, "loss": 1.3387, "step": 183 }, { "epoch": 0.24615384615384617, "grad_norm": 1.022491872782546, "learning_rate": 4.893048128342247e-06, "loss": 1.25, "step": 184 }, { "epoch": 0.24749163879598662, "grad_norm": 1.0323928219576903, "learning_rate": 4.919786096256685e-06, "loss": 1.2111, "step": 185 }, { "epoch": 0.24882943143812708, "grad_norm": 0.9247916356532084, "learning_rate": 4.9465240641711236e-06, "loss": 1.3932, "step": 186 }, { "epoch": 0.25016722408026754, "grad_norm": 0.991035450461571, "learning_rate": 4.973262032085562e-06, "loss": 1.2033, "step": 187 }, { "epoch": 0.25150501672240805, "grad_norm": 1.288375923614787, "learning_rate": 5e-06, "loss": 1.2444, "step": 188 }, { "epoch": 0.2528428093645485, "grad_norm": 0.9437283453049607, "learning_rate": 5.026737967914439e-06, "loss": 1.2995, "step": 189 }, { "epoch": 0.25418060200668896, "grad_norm": 1.1057778511280403, "learning_rate": 5.053475935828877e-06, "loss": 1.264, "step": 190 }, { "epoch": 0.2555183946488294, "grad_norm": 1.0840549410491302, "learning_rate": 5.0802139037433165e-06, "loss": 0.9565, "step": 191 }, { "epoch": 0.2568561872909699, "grad_norm": 1.151733723146191, "learning_rate": 5.106951871657755e-06, "loss": 1.3935, "step": 192 }, { "epoch": 0.2581939799331104, "grad_norm": 1.2809387527948872, "learning_rate": 5.133689839572193e-06, "loss": 1.1428, "step": 193 }, { "epoch": 0.25953177257525084, "grad_norm": 1.1251565942818054, "learning_rate": 5.160427807486631e-06, "loss": 1.2814, "step": 194 }, { "epoch": 0.2608695652173913, "grad_norm": 1.1828306218103561, "learning_rate": 5.187165775401069e-06, "loss": 1.2492, "step": 195 }, { "epoch": 0.26220735785953175, "grad_norm": 0.9469866389182032, "learning_rate": 5.213903743315508e-06, "loss": 1.0437, "step": 196 }, { "epoch": 0.26354515050167227, "grad_norm": 0.9821917811924334, "learning_rate": 5.240641711229947e-06, "loss": 1.2794, "step": 197 }, { "epoch": 0.2648829431438127, "grad_norm": 1.1912180641780492, "learning_rate": 5.2673796791443855e-06, "loss": 1.4338, "step": 198 }, { "epoch": 0.2662207357859532, "grad_norm": 1.1797213371291784, "learning_rate": 5.294117647058824e-06, "loss": 1.1933, "step": 199 }, { "epoch": 0.26755852842809363, "grad_norm": 1.0772086932482343, "learning_rate": 5.320855614973262e-06, "loss": 1.5565, "step": 200 }, { "epoch": 0.2688963210702341, "grad_norm": 1.1525992329334942, "learning_rate": 5.347593582887702e-06, "loss": 1.2621, "step": 201 }, { "epoch": 0.2702341137123746, "grad_norm": 0.9281049069037984, "learning_rate": 5.37433155080214e-06, "loss": 1.4235, "step": 202 }, { "epoch": 0.27157190635451506, "grad_norm": 1.0423036350461647, "learning_rate": 5.4010695187165785e-06, "loss": 1.1545, "step": 203 }, { "epoch": 0.2729096989966555, "grad_norm": 1.067311032397672, "learning_rate": 5.427807486631016e-06, "loss": 1.201, "step": 204 }, { "epoch": 0.27424749163879597, "grad_norm": 1.2025360778389023, "learning_rate": 5.4545454545454545e-06, "loss": 1.4862, "step": 205 }, { "epoch": 0.2755852842809365, "grad_norm": 0.9428425769190734, "learning_rate": 5.481283422459893e-06, "loss": 1.2426, "step": 206 }, { "epoch": 0.27692307692307694, "grad_norm": 0.9435116428036413, "learning_rate": 5.508021390374332e-06, "loss": 1.1978, "step": 207 }, { "epoch": 0.2782608695652174, "grad_norm": 0.9177177968669087, "learning_rate": 5.5347593582887706e-06, "loss": 1.0433, "step": 208 }, { "epoch": 0.27959866220735785, "grad_norm": 1.0342527468174432, "learning_rate": 5.561497326203209e-06, "loss": 1.1882, "step": 209 }, { "epoch": 0.2809364548494983, "grad_norm": 1.0581175046158087, "learning_rate": 5.588235294117647e-06, "loss": 1.0081, "step": 210 }, { "epoch": 0.2822742474916388, "grad_norm": 1.0272667138268619, "learning_rate": 5.614973262032086e-06, "loss": 1.1372, "step": 211 }, { "epoch": 0.2836120401337793, "grad_norm": 0.9383120184179475, "learning_rate": 5.641711229946525e-06, "loss": 1.2233, "step": 212 }, { "epoch": 0.28494983277591973, "grad_norm": 1.2634798916350611, "learning_rate": 5.6684491978609635e-06, "loss": 1.2067, "step": 213 }, { "epoch": 0.2862876254180602, "grad_norm": 1.0966792667095455, "learning_rate": 5.695187165775401e-06, "loss": 1.3432, "step": 214 }, { "epoch": 0.28762541806020064, "grad_norm": 1.2004379340305802, "learning_rate": 5.7219251336898395e-06, "loss": 1.2964, "step": 215 }, { "epoch": 0.28896321070234116, "grad_norm": 1.151402224840367, "learning_rate": 5.748663101604278e-06, "loss": 1.2529, "step": 216 }, { "epoch": 0.2903010033444816, "grad_norm": 1.10871313186554, "learning_rate": 5.775401069518717e-06, "loss": 1.2003, "step": 217 }, { "epoch": 0.29163879598662207, "grad_norm": 1.0357595774530752, "learning_rate": 5.802139037433156e-06, "loss": 1.1847, "step": 218 }, { "epoch": 0.2929765886287625, "grad_norm": 1.0466250054314141, "learning_rate": 5.828877005347594e-06, "loss": 1.1091, "step": 219 }, { "epoch": 0.29431438127090304, "grad_norm": 1.1243023261822038, "learning_rate": 5.8556149732620325e-06, "loss": 1.3212, "step": 220 }, { "epoch": 0.2956521739130435, "grad_norm": 1.1475795858266287, "learning_rate": 5.882352941176471e-06, "loss": 1.3045, "step": 221 }, { "epoch": 0.29698996655518395, "grad_norm": 1.263511137257789, "learning_rate": 5.90909090909091e-06, "loss": 1.3914, "step": 222 }, { "epoch": 0.2983277591973244, "grad_norm": 1.0167405489541992, "learning_rate": 5.935828877005349e-06, "loss": 1.2596, "step": 223 }, { "epoch": 0.29966555183946486, "grad_norm": 0.9311557619274806, "learning_rate": 5.962566844919787e-06, "loss": 1.2965, "step": 224 }, { "epoch": 0.3010033444816054, "grad_norm": 0.8930322917385488, "learning_rate": 5.989304812834225e-06, "loss": 1.0934, "step": 225 }, { "epoch": 0.30234113712374583, "grad_norm": 1.1958158643859074, "learning_rate": 6.016042780748663e-06, "loss": 1.1692, "step": 226 }, { "epoch": 0.3036789297658863, "grad_norm": 1.2169262028241012, "learning_rate": 6.0427807486631015e-06, "loss": 1.2712, "step": 227 }, { "epoch": 0.30501672240802674, "grad_norm": 1.1001497794209407, "learning_rate": 6.069518716577541e-06, "loss": 1.3441, "step": 228 }, { "epoch": 0.3063545150501672, "grad_norm": 1.0781590881874723, "learning_rate": 6.096256684491979e-06, "loss": 1.4163, "step": 229 }, { "epoch": 0.3076923076923077, "grad_norm": 0.9916687284766107, "learning_rate": 6.122994652406418e-06, "loss": 1.0175, "step": 230 }, { "epoch": 0.30903010033444817, "grad_norm": 1.1611848788623675, "learning_rate": 6.149732620320856e-06, "loss": 1.2523, "step": 231 }, { "epoch": 0.3103678929765886, "grad_norm": 1.0314341955961284, "learning_rate": 6.176470588235295e-06, "loss": 1.1485, "step": 232 }, { "epoch": 0.3117056856187291, "grad_norm": 1.275468917299554, "learning_rate": 6.203208556149734e-06, "loss": 1.243, "step": 233 }, { "epoch": 0.3130434782608696, "grad_norm": 1.1055465436348912, "learning_rate": 6.229946524064172e-06, "loss": 1.0791, "step": 234 }, { "epoch": 0.31438127090301005, "grad_norm": 1.2520919625732445, "learning_rate": 6.25668449197861e-06, "loss": 1.2821, "step": 235 }, { "epoch": 0.3157190635451505, "grad_norm": 0.9223924013455241, "learning_rate": 6.283422459893048e-06, "loss": 1.3735, "step": 236 }, { "epoch": 0.31705685618729096, "grad_norm": 1.2776476971932909, "learning_rate": 6.3101604278074865e-06, "loss": 1.2886, "step": 237 }, { "epoch": 0.3183946488294314, "grad_norm": 1.010311116228789, "learning_rate": 6.336898395721926e-06, "loss": 1.2268, "step": 238 }, { "epoch": 0.3197324414715719, "grad_norm": 1.2178767564598971, "learning_rate": 6.363636363636364e-06, "loss": 1.2774, "step": 239 }, { "epoch": 0.3210702341137124, "grad_norm": 0.9765032197760644, "learning_rate": 6.390374331550803e-06, "loss": 1.0641, "step": 240 }, { "epoch": 0.32240802675585284, "grad_norm": 1.1716785123013587, "learning_rate": 6.417112299465241e-06, "loss": 1.2325, "step": 241 }, { "epoch": 0.3237458193979933, "grad_norm": 1.0505371223409037, "learning_rate": 6.4438502673796795e-06, "loss": 1.3489, "step": 242 }, { "epoch": 0.3250836120401338, "grad_norm": 1.258047724389353, "learning_rate": 6.470588235294119e-06, "loss": 1.4232, "step": 243 }, { "epoch": 0.32642140468227426, "grad_norm": 1.5098420660490974, "learning_rate": 6.497326203208557e-06, "loss": 1.3314, "step": 244 }, { "epoch": 0.3277591973244147, "grad_norm": 0.903018915452519, "learning_rate": 6.524064171122996e-06, "loss": 1.1156, "step": 245 }, { "epoch": 0.3290969899665552, "grad_norm": 1.0956719712028697, "learning_rate": 6.550802139037433e-06, "loss": 1.3226, "step": 246 }, { "epoch": 0.33043478260869563, "grad_norm": 0.9045803074873201, "learning_rate": 6.577540106951872e-06, "loss": 1.2028, "step": 247 }, { "epoch": 0.33177257525083614, "grad_norm": 0.9752833364587686, "learning_rate": 6.60427807486631e-06, "loss": 1.2958, "step": 248 }, { "epoch": 0.3331103678929766, "grad_norm": 1.1777906079473246, "learning_rate": 6.631016042780749e-06, "loss": 1.0394, "step": 249 }, { "epoch": 0.33444816053511706, "grad_norm": 1.1975021767953604, "learning_rate": 6.657754010695188e-06, "loss": 1.2409, "step": 250 }, { "epoch": 0.3357859531772575, "grad_norm": 1.1343138081364368, "learning_rate": 6.684491978609626e-06, "loss": 1.2547, "step": 251 }, { "epoch": 0.33712374581939797, "grad_norm": 1.1919178444838614, "learning_rate": 6.711229946524065e-06, "loss": 1.4559, "step": 252 }, { "epoch": 0.3384615384615385, "grad_norm": 0.9205273785138814, "learning_rate": 6.737967914438504e-06, "loss": 1.1704, "step": 253 }, { "epoch": 0.33979933110367894, "grad_norm": 1.0901122003316708, "learning_rate": 6.764705882352942e-06, "loss": 1.3695, "step": 254 }, { "epoch": 0.3411371237458194, "grad_norm": 1.2229242593238725, "learning_rate": 6.791443850267381e-06, "loss": 1.2577, "step": 255 }, { "epoch": 0.34247491638795985, "grad_norm": 1.0556295997803857, "learning_rate": 6.818181818181818e-06, "loss": 1.0669, "step": 256 }, { "epoch": 0.34381270903010036, "grad_norm": 1.0316230028449263, "learning_rate": 6.844919786096257e-06, "loss": 1.3137, "step": 257 }, { "epoch": 0.3451505016722408, "grad_norm": 0.9849274396462254, "learning_rate": 6.871657754010695e-06, "loss": 0.8761, "step": 258 }, { "epoch": 0.3464882943143813, "grad_norm": 1.0005527918208867, "learning_rate": 6.898395721925134e-06, "loss": 1.2667, "step": 259 }, { "epoch": 0.34782608695652173, "grad_norm": 1.2148178262632785, "learning_rate": 6.925133689839573e-06, "loss": 1.243, "step": 260 }, { "epoch": 0.3491638795986622, "grad_norm": 0.9488822006146334, "learning_rate": 6.951871657754011e-06, "loss": 1.0177, "step": 261 }, { "epoch": 0.3505016722408027, "grad_norm": 1.2310213324751118, "learning_rate": 6.97860962566845e-06, "loss": 1.2505, "step": 262 }, { "epoch": 0.35183946488294315, "grad_norm": 0.9820397298365069, "learning_rate": 7.005347593582889e-06, "loss": 1.2833, "step": 263 }, { "epoch": 0.3531772575250836, "grad_norm": 1.2051778189594624, "learning_rate": 7.032085561497327e-06, "loss": 1.1921, "step": 264 }, { "epoch": 0.35451505016722407, "grad_norm": 1.2057369028811422, "learning_rate": 7.058823529411766e-06, "loss": 1.0696, "step": 265 }, { "epoch": 0.3558528428093645, "grad_norm": 0.9306541434674512, "learning_rate": 7.085561497326203e-06, "loss": 1.0584, "step": 266 }, { "epoch": 0.35719063545150503, "grad_norm": 1.1728251885966816, "learning_rate": 7.112299465240642e-06, "loss": 1.3446, "step": 267 }, { "epoch": 0.3585284280936455, "grad_norm": 1.0837590445291865, "learning_rate": 7.13903743315508e-06, "loss": 1.2475, "step": 268 }, { "epoch": 0.35986622073578595, "grad_norm": 1.1099792646719011, "learning_rate": 7.1657754010695195e-06, "loss": 1.1991, "step": 269 }, { "epoch": 0.3612040133779264, "grad_norm": 0.8961489941355324, "learning_rate": 7.192513368983958e-06, "loss": 1.1424, "step": 270 }, { "epoch": 0.3625418060200669, "grad_norm": 1.1135359986723232, "learning_rate": 7.219251336898396e-06, "loss": 1.1197, "step": 271 }, { "epoch": 0.36387959866220737, "grad_norm": 1.218465509799198, "learning_rate": 7.245989304812835e-06, "loss": 1.4389, "step": 272 }, { "epoch": 0.3652173913043478, "grad_norm": 1.0230676275950505, "learning_rate": 7.272727272727273e-06, "loss": 1.4047, "step": 273 }, { "epoch": 0.3665551839464883, "grad_norm": 1.1925929362397132, "learning_rate": 7.2994652406417124e-06, "loss": 1.2023, "step": 274 }, { "epoch": 0.36789297658862874, "grad_norm": 1.0063916488196771, "learning_rate": 7.326203208556151e-06, "loss": 1.4423, "step": 275 }, { "epoch": 0.36923076923076925, "grad_norm": 1.0752177073687528, "learning_rate": 7.352941176470589e-06, "loss": 1.3234, "step": 276 }, { "epoch": 0.3705685618729097, "grad_norm": 1.0806443170175464, "learning_rate": 7.379679144385027e-06, "loss": 1.4123, "step": 277 }, { "epoch": 0.37190635451505016, "grad_norm": 0.8902041922980004, "learning_rate": 7.406417112299465e-06, "loss": 1.1878, "step": 278 }, { "epoch": 0.3732441471571906, "grad_norm": 1.1687368660059196, "learning_rate": 7.433155080213904e-06, "loss": 1.3435, "step": 279 }, { "epoch": 0.3745819397993311, "grad_norm": 1.0573906161001407, "learning_rate": 7.459893048128343e-06, "loss": 1.1239, "step": 280 }, { "epoch": 0.3759197324414716, "grad_norm": 1.1716015635271027, "learning_rate": 7.486631016042781e-06, "loss": 1.3871, "step": 281 }, { "epoch": 0.37725752508361204, "grad_norm": 1.2419323095811095, "learning_rate": 7.51336898395722e-06, "loss": 1.2096, "step": 282 }, { "epoch": 0.3785953177257525, "grad_norm": 1.1264243143202848, "learning_rate": 7.540106951871658e-06, "loss": 1.5293, "step": 283 }, { "epoch": 0.37993311036789296, "grad_norm": 1.0493134440372986, "learning_rate": 7.5668449197860975e-06, "loss": 1.1268, "step": 284 }, { "epoch": 0.38127090301003347, "grad_norm": 1.0437987226550618, "learning_rate": 7.593582887700536e-06, "loss": 1.4346, "step": 285 }, { "epoch": 0.3826086956521739, "grad_norm": 1.2688372164013815, "learning_rate": 7.620320855614974e-06, "loss": 1.4024, "step": 286 }, { "epoch": 0.3839464882943144, "grad_norm": 0.9631358529884743, "learning_rate": 7.647058823529411e-06, "loss": 1.0309, "step": 287 }, { "epoch": 0.38528428093645484, "grad_norm": 1.072240242516224, "learning_rate": 7.67379679144385e-06, "loss": 1.331, "step": 288 }, { "epoch": 0.3866220735785953, "grad_norm": 0.973037495510108, "learning_rate": 7.70053475935829e-06, "loss": 1.0663, "step": 289 }, { "epoch": 0.3879598662207358, "grad_norm": 1.046273724116429, "learning_rate": 7.727272727272727e-06, "loss": 1.1228, "step": 290 }, { "epoch": 0.38929765886287626, "grad_norm": 1.124187058934406, "learning_rate": 7.754010695187166e-06, "loss": 1.2674, "step": 291 }, { "epoch": 0.3906354515050167, "grad_norm": 0.9071136019002644, "learning_rate": 7.780748663101606e-06, "loss": 1.1872, "step": 292 }, { "epoch": 0.3919732441471572, "grad_norm": 0.9094884502494917, "learning_rate": 7.807486631016043e-06, "loss": 1.1579, "step": 293 }, { "epoch": 0.3933110367892977, "grad_norm": 1.0501534594127808, "learning_rate": 7.834224598930483e-06, "loss": 1.3831, "step": 294 }, { "epoch": 0.39464882943143814, "grad_norm": 0.9666127672825461, "learning_rate": 7.86096256684492e-06, "loss": 1.2415, "step": 295 }, { "epoch": 0.3959866220735786, "grad_norm": 1.0494180828841164, "learning_rate": 7.88770053475936e-06, "loss": 1.2464, "step": 296 }, { "epoch": 0.39732441471571905, "grad_norm": 1.0362380028576592, "learning_rate": 7.914438502673799e-06, "loss": 1.235, "step": 297 }, { "epoch": 0.3986622073578595, "grad_norm": 1.0495614520584642, "learning_rate": 7.941176470588236e-06, "loss": 0.9618, "step": 298 }, { "epoch": 0.4, "grad_norm": 0.9377769957616789, "learning_rate": 7.967914438502674e-06, "loss": 1.2903, "step": 299 }, { "epoch": 0.4013377926421405, "grad_norm": 1.0558307255935884, "learning_rate": 7.994652406417113e-06, "loss": 1.1972, "step": 300 }, { "epoch": 0.40267558528428093, "grad_norm": 1.070607453768897, "learning_rate": 8.02139037433155e-06, "loss": 1.3614, "step": 301 }, { "epoch": 0.4040133779264214, "grad_norm": 1.294423445881669, "learning_rate": 8.04812834224599e-06, "loss": 1.2616, "step": 302 }, { "epoch": 0.40535117056856185, "grad_norm": 0.973443620226616, "learning_rate": 8.07486631016043e-06, "loss": 1.0042, "step": 303 }, { "epoch": 0.40668896321070236, "grad_norm": 1.3923968553510182, "learning_rate": 8.101604278074867e-06, "loss": 1.2321, "step": 304 }, { "epoch": 0.4080267558528428, "grad_norm": 1.0262494988539328, "learning_rate": 8.128342245989306e-06, "loss": 1.0736, "step": 305 }, { "epoch": 0.40936454849498327, "grad_norm": 1.0157893693860662, "learning_rate": 8.155080213903744e-06, "loss": 1.2096, "step": 306 }, { "epoch": 0.4107023411371237, "grad_norm": 1.1631142238692702, "learning_rate": 8.181818181818183e-06, "loss": 1.0864, "step": 307 }, { "epoch": 0.41204013377926424, "grad_norm": 1.088624307519204, "learning_rate": 8.20855614973262e-06, "loss": 1.1996, "step": 308 }, { "epoch": 0.4133779264214047, "grad_norm": 1.021140195463474, "learning_rate": 8.23529411764706e-06, "loss": 1.3784, "step": 309 }, { "epoch": 0.41471571906354515, "grad_norm": 1.2708654808503461, "learning_rate": 8.262032085561497e-06, "loss": 1.0753, "step": 310 }, { "epoch": 0.4160535117056856, "grad_norm": 1.1529335856678133, "learning_rate": 8.288770053475937e-06, "loss": 1.1959, "step": 311 }, { "epoch": 0.41739130434782606, "grad_norm": 1.2279635966744422, "learning_rate": 8.315508021390374e-06, "loss": 1.1053, "step": 312 }, { "epoch": 0.4187290969899666, "grad_norm": 1.047059905376027, "learning_rate": 8.342245989304813e-06, "loss": 1.4733, "step": 313 }, { "epoch": 0.42006688963210703, "grad_norm": 0.9690077639823466, "learning_rate": 8.368983957219253e-06, "loss": 1.4009, "step": 314 }, { "epoch": 0.4214046822742475, "grad_norm": 0.918506333560281, "learning_rate": 8.39572192513369e-06, "loss": 1.0039, "step": 315 }, { "epoch": 0.42274247491638794, "grad_norm": 1.0171176178314516, "learning_rate": 8.42245989304813e-06, "loss": 1.3321, "step": 316 }, { "epoch": 0.4240802675585284, "grad_norm": 1.0112434023096923, "learning_rate": 8.449197860962567e-06, "loss": 1.1803, "step": 317 }, { "epoch": 0.4254180602006689, "grad_norm": 1.100777568915651, "learning_rate": 8.475935828877005e-06, "loss": 1.0405, "step": 318 }, { "epoch": 0.42675585284280937, "grad_norm": 1.0195083206991282, "learning_rate": 8.502673796791444e-06, "loss": 1.0981, "step": 319 }, { "epoch": 0.4280936454849498, "grad_norm": 1.0260263159196596, "learning_rate": 8.529411764705883e-06, "loss": 1.1459, "step": 320 }, { "epoch": 0.4294314381270903, "grad_norm": 0.9403570207213072, "learning_rate": 8.556149732620321e-06, "loss": 1.3919, "step": 321 }, { "epoch": 0.4307692307692308, "grad_norm": 1.233576549911364, "learning_rate": 8.58288770053476e-06, "loss": 1.3778, "step": 322 }, { "epoch": 0.43210702341137125, "grad_norm": 1.2618879289631673, "learning_rate": 8.609625668449198e-06, "loss": 1.1033, "step": 323 }, { "epoch": 0.4334448160535117, "grad_norm": 1.1803531281291175, "learning_rate": 8.636363636363637e-06, "loss": 1.1826, "step": 324 }, { "epoch": 0.43478260869565216, "grad_norm": 1.2235921801481053, "learning_rate": 8.663101604278076e-06, "loss": 1.1241, "step": 325 }, { "epoch": 0.4361204013377926, "grad_norm": 1.1765156673713362, "learning_rate": 8.689839572192514e-06, "loss": 1.205, "step": 326 }, { "epoch": 0.43745819397993313, "grad_norm": 0.9779466937406754, "learning_rate": 8.716577540106953e-06, "loss": 0.9837, "step": 327 }, { "epoch": 0.4387959866220736, "grad_norm": 1.1489661319859135, "learning_rate": 8.743315508021392e-06, "loss": 1.2539, "step": 328 }, { "epoch": 0.44013377926421404, "grad_norm": 0.989982267221648, "learning_rate": 8.77005347593583e-06, "loss": 1.0887, "step": 329 }, { "epoch": 0.4414715719063545, "grad_norm": 1.2740789339145953, "learning_rate": 8.796791443850268e-06, "loss": 1.174, "step": 330 }, { "epoch": 0.442809364548495, "grad_norm": 1.1554514938611635, "learning_rate": 8.823529411764707e-06, "loss": 1.2302, "step": 331 }, { "epoch": 0.44414715719063547, "grad_norm": 0.9312582664485676, "learning_rate": 8.850267379679144e-06, "loss": 1.2639, "step": 332 }, { "epoch": 0.4454849498327759, "grad_norm": 1.0145007961451649, "learning_rate": 8.877005347593584e-06, "loss": 1.0097, "step": 333 }, { "epoch": 0.4468227424749164, "grad_norm": 1.1586382841024734, "learning_rate": 8.903743315508023e-06, "loss": 0.9566, "step": 334 }, { "epoch": 0.44816053511705684, "grad_norm": 0.938638190261139, "learning_rate": 8.93048128342246e-06, "loss": 1.095, "step": 335 }, { "epoch": 0.44949832775919735, "grad_norm": 1.0386423955640514, "learning_rate": 8.9572192513369e-06, "loss": 1.3404, "step": 336 }, { "epoch": 0.4508361204013378, "grad_norm": 1.097698327264974, "learning_rate": 8.983957219251337e-06, "loss": 1.3331, "step": 337 }, { "epoch": 0.45217391304347826, "grad_norm": 1.021650404910539, "learning_rate": 9.010695187165777e-06, "loss": 1.1318, "step": 338 }, { "epoch": 0.4535117056856187, "grad_norm": 1.0455874311848476, "learning_rate": 9.037433155080214e-06, "loss": 1.3927, "step": 339 }, { "epoch": 0.45484949832775917, "grad_norm": 2.3753676468197007, "learning_rate": 9.064171122994653e-06, "loss": 1.1758, "step": 340 }, { "epoch": 0.4561872909698997, "grad_norm": 1.0748483399752888, "learning_rate": 9.090909090909091e-06, "loss": 1.2107, "step": 341 }, { "epoch": 0.45752508361204014, "grad_norm": 1.21372421922117, "learning_rate": 9.11764705882353e-06, "loss": 1.4119, "step": 342 }, { "epoch": 0.4588628762541806, "grad_norm": 1.2177144823477217, "learning_rate": 9.144385026737968e-06, "loss": 1.0879, "step": 343 }, { "epoch": 0.46020066889632105, "grad_norm": 1.2080932428626807, "learning_rate": 9.171122994652407e-06, "loss": 1.0344, "step": 344 }, { "epoch": 0.46153846153846156, "grad_norm": 1.2123336491288588, "learning_rate": 9.197860962566846e-06, "loss": 1.3929, "step": 345 }, { "epoch": 0.462876254180602, "grad_norm": 1.0725596853002646, "learning_rate": 9.224598930481284e-06, "loss": 0.9641, "step": 346 }, { "epoch": 0.4642140468227425, "grad_norm": 1.0609276036948314, "learning_rate": 9.251336898395723e-06, "loss": 1.0328, "step": 347 }, { "epoch": 0.46555183946488293, "grad_norm": 1.0566485098453948, "learning_rate": 9.278074866310161e-06, "loss": 1.0151, "step": 348 }, { "epoch": 0.4668896321070234, "grad_norm": 0.9673711442746381, "learning_rate": 9.3048128342246e-06, "loss": 1.1382, "step": 349 }, { "epoch": 0.4682274247491639, "grad_norm": 1.2730086662726603, "learning_rate": 9.331550802139038e-06, "loss": 1.3532, "step": 350 }, { "epoch": 0.46956521739130436, "grad_norm": 1.0928686947198454, "learning_rate": 9.358288770053477e-06, "loss": 1.3905, "step": 351 }, { "epoch": 0.4709030100334448, "grad_norm": 1.0201185273201718, "learning_rate": 9.385026737967915e-06, "loss": 1.159, "step": 352 }, { "epoch": 0.47224080267558527, "grad_norm": 1.3043249811292532, "learning_rate": 9.411764705882354e-06, "loss": 1.3979, "step": 353 }, { "epoch": 0.4735785953177257, "grad_norm": 1.1477426700535256, "learning_rate": 9.438502673796791e-06, "loss": 1.3231, "step": 354 }, { "epoch": 0.47491638795986624, "grad_norm": 1.3118504190748892, "learning_rate": 9.46524064171123e-06, "loss": 1.1002, "step": 355 }, { "epoch": 0.4762541806020067, "grad_norm": 1.1162797335341523, "learning_rate": 9.49197860962567e-06, "loss": 1.3024, "step": 356 }, { "epoch": 0.47759197324414715, "grad_norm": 1.1416803725229607, "learning_rate": 9.518716577540108e-06, "loss": 1.3772, "step": 357 }, { "epoch": 0.4789297658862876, "grad_norm": 1.431097758637806, "learning_rate": 9.545454545454547e-06, "loss": 1.4081, "step": 358 }, { "epoch": 0.4802675585284281, "grad_norm": 1.143696539349071, "learning_rate": 9.572192513368986e-06, "loss": 1.3409, "step": 359 }, { "epoch": 0.4816053511705686, "grad_norm": 1.2295766000755, "learning_rate": 9.598930481283422e-06, "loss": 1.2881, "step": 360 }, { "epoch": 0.48294314381270903, "grad_norm": 1.1496995774805734, "learning_rate": 9.625668449197861e-06, "loss": 1.3206, "step": 361 }, { "epoch": 0.4842809364548495, "grad_norm": 0.9885989371794952, "learning_rate": 9.6524064171123e-06, "loss": 1.1715, "step": 362 }, { "epoch": 0.48561872909698994, "grad_norm": 1.1086942664965442, "learning_rate": 9.679144385026738e-06, "loss": 1.3225, "step": 363 }, { "epoch": 0.48695652173913045, "grad_norm": 1.1209484608360967, "learning_rate": 9.705882352941177e-06, "loss": 1.0414, "step": 364 }, { "epoch": 0.4882943143812709, "grad_norm": 1.185329308515873, "learning_rate": 9.732620320855617e-06, "loss": 1.1672, "step": 365 }, { "epoch": 0.48963210702341137, "grad_norm": 1.1258182080862529, "learning_rate": 9.759358288770054e-06, "loss": 1.2185, "step": 366 }, { "epoch": 0.4909698996655518, "grad_norm": 1.2419101831471426, "learning_rate": 9.786096256684493e-06, "loss": 1.3881, "step": 367 }, { "epoch": 0.49230769230769234, "grad_norm": 1.0401564466981865, "learning_rate": 9.812834224598931e-06, "loss": 1.0845, "step": 368 }, { "epoch": 0.4936454849498328, "grad_norm": 1.0229147505460416, "learning_rate": 9.83957219251337e-06, "loss": 1.2725, "step": 369 }, { "epoch": 0.49498327759197325, "grad_norm": 1.0366315076732977, "learning_rate": 9.866310160427808e-06, "loss": 1.1698, "step": 370 }, { "epoch": 0.4963210702341137, "grad_norm": 1.1754655103784883, "learning_rate": 9.893048128342247e-06, "loss": 1.2406, "step": 371 }, { "epoch": 0.49765886287625416, "grad_norm": 1.18334309200908, "learning_rate": 9.919786096256685e-06, "loss": 1.2422, "step": 372 }, { "epoch": 0.49899665551839467, "grad_norm": 1.264751646172798, "learning_rate": 9.946524064171124e-06, "loss": 1.2276, "step": 373 }, { "epoch": 0.5003344481605351, "grad_norm": 1.4098773700534457, "learning_rate": 9.973262032085562e-06, "loss": 1.3232, "step": 374 }, { "epoch": 0.5016722408026756, "grad_norm": 1.1543630385642156, "learning_rate": 1e-05, "loss": 1.4878, "step": 375 }, { "epoch": 0.5030100334448161, "grad_norm": 1.1980277120933867, "learning_rate": 9.999997822232566e-06, "loss": 1.1958, "step": 376 }, { "epoch": 0.5043478260869565, "grad_norm": 0.9586006072608856, "learning_rate": 9.99999128893216e-06, "loss": 1.3427, "step": 377 }, { "epoch": 0.505685618729097, "grad_norm": 0.938647666928365, "learning_rate": 9.999980400104472e-06, "loss": 1.0203, "step": 378 }, { "epoch": 0.5070234113712374, "grad_norm": 1.1453796916788488, "learning_rate": 9.99996515575899e-06, "loss": 1.0816, "step": 379 }, { "epoch": 0.5083612040133779, "grad_norm": 1.09561182998107, "learning_rate": 9.99994555590899e-06, "loss": 1.4898, "step": 380 }, { "epoch": 0.5096989966555184, "grad_norm": 1.2303332603803678, "learning_rate": 9.99992160057155e-06, "loss": 1.4717, "step": 381 }, { "epoch": 0.5110367892976588, "grad_norm": 0.9126481448749092, "learning_rate": 9.999893289767533e-06, "loss": 1.153, "step": 382 }, { "epoch": 0.5123745819397993, "grad_norm": 1.15785652105225, "learning_rate": 9.999860623521604e-06, "loss": 1.2227, "step": 383 }, { "epoch": 0.5137123745819397, "grad_norm": 1.1681511305030072, "learning_rate": 9.999823601862217e-06, "loss": 1.172, "step": 384 }, { "epoch": 0.5150501672240803, "grad_norm": 0.9866938253086767, "learning_rate": 9.999782224821624e-06, "loss": 1.2337, "step": 385 }, { "epoch": 0.5163879598662208, "grad_norm": 1.1443603460708736, "learning_rate": 9.999736492435867e-06, "loss": 1.2197, "step": 386 }, { "epoch": 0.5177257525083612, "grad_norm": 1.0249746373052577, "learning_rate": 9.999686404744782e-06, "loss": 1.4569, "step": 387 }, { "epoch": 0.5190635451505017, "grad_norm": 1.204668950832557, "learning_rate": 9.999631961792006e-06, "loss": 1.4209, "step": 388 }, { "epoch": 0.5204013377926422, "grad_norm": 1.0272346315343277, "learning_rate": 9.99957316362496e-06, "loss": 1.0917, "step": 389 }, { "epoch": 0.5217391304347826, "grad_norm": 1.0148145147502554, "learning_rate": 9.999510010294864e-06, "loss": 1.1398, "step": 390 }, { "epoch": 0.5230769230769231, "grad_norm": 1.0768396827284614, "learning_rate": 9.999442501856736e-06, "loss": 1.4531, "step": 391 }, { "epoch": 0.5244147157190635, "grad_norm": 0.9334969393305891, "learning_rate": 9.999370638369377e-06, "loss": 1.2758, "step": 392 }, { "epoch": 0.525752508361204, "grad_norm": 1.2382664799439471, "learning_rate": 9.999294419895389e-06, "loss": 1.0486, "step": 393 }, { "epoch": 0.5270903010033445, "grad_norm": 1.099758704833212, "learning_rate": 9.99921384650117e-06, "loss": 1.236, "step": 394 }, { "epoch": 0.5284280936454849, "grad_norm": 0.9561287836860403, "learning_rate": 9.999128918256904e-06, "loss": 1.3743, "step": 395 }, { "epoch": 0.5297658862876254, "grad_norm": 1.0022051694854943, "learning_rate": 9.999039635236576e-06, "loss": 1.2374, "step": 396 }, { "epoch": 0.5311036789297658, "grad_norm": 1.1226920683441717, "learning_rate": 9.998945997517957e-06, "loss": 1.1392, "step": 397 }, { "epoch": 0.5324414715719064, "grad_norm": 0.9518934789849339, "learning_rate": 9.99884800518262e-06, "loss": 1.1682, "step": 398 }, { "epoch": 0.5337792642140469, "grad_norm": 0.9397939073647785, "learning_rate": 9.998745658315924e-06, "loss": 1.1536, "step": 399 }, { "epoch": 0.5351170568561873, "grad_norm": 1.061090936389956, "learning_rate": 9.998638957007024e-06, "loss": 1.2079, "step": 400 }, { "epoch": 0.5364548494983278, "grad_norm": 0.8985287906069113, "learning_rate": 9.998527901348869e-06, "loss": 1.2256, "step": 401 }, { "epoch": 0.5377926421404682, "grad_norm": 1.097540015740794, "learning_rate": 9.998412491438201e-06, "loss": 1.3826, "step": 402 }, { "epoch": 0.5391304347826087, "grad_norm": 1.1171008287187654, "learning_rate": 9.998292727375554e-06, "loss": 1.3815, "step": 403 }, { "epoch": 0.5404682274247492, "grad_norm": 1.215258145581717, "learning_rate": 9.998168609265254e-06, "loss": 1.2969, "step": 404 }, { "epoch": 0.5418060200668896, "grad_norm": 1.0820830142669478, "learning_rate": 9.998040137215423e-06, "loss": 1.4355, "step": 405 }, { "epoch": 0.5431438127090301, "grad_norm": 1.1362065506028982, "learning_rate": 9.997907311337973e-06, "loss": 1.15, "step": 406 }, { "epoch": 0.5444816053511705, "grad_norm": 1.1988875655106281, "learning_rate": 9.99777013174861e-06, "loss": 1.2032, "step": 407 }, { "epoch": 0.545819397993311, "grad_norm": 0.9090946859409978, "learning_rate": 9.99762859856683e-06, "loss": 0.9534, "step": 408 }, { "epoch": 0.5471571906354515, "grad_norm": 0.9540045567794444, "learning_rate": 9.997482711915926e-06, "loss": 1.0648, "step": 409 }, { "epoch": 0.5484949832775919, "grad_norm": 0.9961931800197278, "learning_rate": 9.997332471922981e-06, "loss": 1.2449, "step": 410 }, { "epoch": 0.5498327759197325, "grad_norm": 1.0879975783486475, "learning_rate": 9.99717787871887e-06, "loss": 1.1854, "step": 411 }, { "epoch": 0.551170568561873, "grad_norm": 1.1098923006725112, "learning_rate": 9.997018932438256e-06, "loss": 1.0927, "step": 412 }, { "epoch": 0.5525083612040134, "grad_norm": 0.9503277588477231, "learning_rate": 9.996855633219605e-06, "loss": 1.3003, "step": 413 }, { "epoch": 0.5538461538461539, "grad_norm": 1.0320370439724704, "learning_rate": 9.996687981205163e-06, "loss": 1.3607, "step": 414 }, { "epoch": 0.5551839464882943, "grad_norm": 1.2891401008117247, "learning_rate": 9.996515976540974e-06, "loss": 1.2753, "step": 415 }, { "epoch": 0.5565217391304348, "grad_norm": 1.0500480019856635, "learning_rate": 9.996339619376876e-06, "loss": 1.1194, "step": 416 }, { "epoch": 0.5578595317725753, "grad_norm": 1.0588239866889981, "learning_rate": 9.99615890986649e-06, "loss": 1.1929, "step": 417 }, { "epoch": 0.5591973244147157, "grad_norm": 1.1664544138274497, "learning_rate": 9.995973848167234e-06, "loss": 1.1064, "step": 418 }, { "epoch": 0.5605351170568562, "grad_norm": 1.103239905683686, "learning_rate": 9.99578443444032e-06, "loss": 1.2503, "step": 419 }, { "epoch": 0.5618729096989966, "grad_norm": 1.1377839116168003, "learning_rate": 9.995590668850745e-06, "loss": 1.0718, "step": 420 }, { "epoch": 0.5632107023411371, "grad_norm": 1.1683164828275518, "learning_rate": 9.9953925515673e-06, "loss": 1.225, "step": 421 }, { "epoch": 0.5645484949832776, "grad_norm": 1.0369187452379705, "learning_rate": 9.995190082762566e-06, "loss": 1.3464, "step": 422 }, { "epoch": 0.565886287625418, "grad_norm": 1.3872966641153328, "learning_rate": 9.994983262612916e-06, "loss": 1.0848, "step": 423 }, { "epoch": 0.5672240802675586, "grad_norm": 1.0174072305889925, "learning_rate": 9.99477209129851e-06, "loss": 1.1697, "step": 424 }, { "epoch": 0.568561872909699, "grad_norm": 1.1931533814777413, "learning_rate": 9.994556569003305e-06, "loss": 1.054, "step": 425 }, { "epoch": 0.5698996655518395, "grad_norm": 0.9429015235669953, "learning_rate": 9.994336695915041e-06, "loss": 1.2349, "step": 426 }, { "epoch": 0.57123745819398, "grad_norm": 1.2970097784577572, "learning_rate": 9.99411247222525e-06, "loss": 1.3045, "step": 427 }, { "epoch": 0.5725752508361204, "grad_norm": 1.1074058909271967, "learning_rate": 9.993883898129259e-06, "loss": 1.3267, "step": 428 }, { "epoch": 0.5739130434782609, "grad_norm": 1.038581519046633, "learning_rate": 9.993650973826177e-06, "loss": 1.3265, "step": 429 }, { "epoch": 0.5752508361204013, "grad_norm": 1.0436226834674196, "learning_rate": 9.993413699518906e-06, "loss": 1.3916, "step": 430 }, { "epoch": 0.5765886287625418, "grad_norm": 1.0089184411644652, "learning_rate": 9.99317207541414e-06, "loss": 1.3844, "step": 431 }, { "epoch": 0.5779264214046823, "grad_norm": 0.9568702115459079, "learning_rate": 9.992926101722355e-06, "loss": 1.0924, "step": 432 }, { "epoch": 0.5792642140468227, "grad_norm": 1.6164535528686894, "learning_rate": 9.992675778657824e-06, "loss": 1.1339, "step": 433 }, { "epoch": 0.5806020066889632, "grad_norm": 0.9986790260467662, "learning_rate": 9.992421106438606e-06, "loss": 1.2536, "step": 434 }, { "epoch": 0.5819397993311036, "grad_norm": 1.0595031040162188, "learning_rate": 9.992162085286543e-06, "loss": 0.9984, "step": 435 }, { "epoch": 0.5832775919732441, "grad_norm": 0.9633024096333119, "learning_rate": 9.991898715427274e-06, "loss": 1.2491, "step": 436 }, { "epoch": 0.5846153846153846, "grad_norm": 1.1311404921219397, "learning_rate": 9.991630997090222e-06, "loss": 1.2774, "step": 437 }, { "epoch": 0.585953177257525, "grad_norm": 0.9634073117439108, "learning_rate": 9.991358930508599e-06, "loss": 0.9111, "step": 438 }, { "epoch": 0.5872909698996656, "grad_norm": 1.038090065126639, "learning_rate": 9.991082515919402e-06, "loss": 1.3188, "step": 439 }, { "epoch": 0.5886287625418061, "grad_norm": 1.0547398478525392, "learning_rate": 9.990801753563418e-06, "loss": 1.3794, "step": 440 }, { "epoch": 0.5899665551839465, "grad_norm": 1.2236179066958357, "learning_rate": 9.990516643685222e-06, "loss": 1.2108, "step": 441 }, { "epoch": 0.591304347826087, "grad_norm": 1.4002157378024767, "learning_rate": 9.990227186533174e-06, "loss": 1.0618, "step": 442 }, { "epoch": 0.5926421404682274, "grad_norm": 0.9128789950881795, "learning_rate": 9.989933382359423e-06, "loss": 1.1629, "step": 443 }, { "epoch": 0.5939799331103679, "grad_norm": 1.0545825979019847, "learning_rate": 9.989635231419903e-06, "loss": 1.1663, "step": 444 }, { "epoch": 0.5953177257525084, "grad_norm": 1.0499914209423973, "learning_rate": 9.989332733974337e-06, "loss": 1.1207, "step": 445 }, { "epoch": 0.5966555183946488, "grad_norm": 2.3009150833649716, "learning_rate": 9.989025890286233e-06, "loss": 1.395, "step": 446 }, { "epoch": 0.5979933110367893, "grad_norm": 0.905139911557014, "learning_rate": 9.988714700622882e-06, "loss": 1.0147, "step": 447 }, { "epoch": 0.5993311036789297, "grad_norm": 1.1688754898346634, "learning_rate": 9.988399165255365e-06, "loss": 1.4547, "step": 448 }, { "epoch": 0.6006688963210702, "grad_norm": 1.2089218212969008, "learning_rate": 9.988079284458547e-06, "loss": 1.1131, "step": 449 }, { "epoch": 0.6020066889632107, "grad_norm": 1.076240911301496, "learning_rate": 9.987755058511079e-06, "loss": 1.1393, "step": 450 }, { "epoch": 0.6033444816053511, "grad_norm": 1.01997639931134, "learning_rate": 9.987426487695396e-06, "loss": 1.274, "step": 451 }, { "epoch": 0.6046822742474917, "grad_norm": 1.4148860483757058, "learning_rate": 9.987093572297716e-06, "loss": 1.22, "step": 452 }, { "epoch": 0.6060200668896321, "grad_norm": 0.9812685731896932, "learning_rate": 9.986756312608048e-06, "loss": 1.0958, "step": 453 }, { "epoch": 0.6073578595317726, "grad_norm": 1.1281169625932292, "learning_rate": 9.98641470892018e-06, "loss": 1.2758, "step": 454 }, { "epoch": 0.6086956521739131, "grad_norm": 1.0756694067702124, "learning_rate": 9.986068761531681e-06, "loss": 1.1781, "step": 455 }, { "epoch": 0.6100334448160535, "grad_norm": 1.0695533443126792, "learning_rate": 9.985718470743916e-06, "loss": 1.2998, "step": 456 }, { "epoch": 0.611371237458194, "grad_norm": 0.9545126885392965, "learning_rate": 9.985363836862021e-06, "loss": 1.0279, "step": 457 }, { "epoch": 0.6127090301003344, "grad_norm": 1.7120115629029613, "learning_rate": 9.98500486019492e-06, "loss": 1.2777, "step": 458 }, { "epoch": 0.6140468227424749, "grad_norm": 1.159333327096624, "learning_rate": 9.98464154105532e-06, "loss": 1.35, "step": 459 }, { "epoch": 0.6153846153846154, "grad_norm": 1.106493446626155, "learning_rate": 9.984273879759713e-06, "loss": 1.2708, "step": 460 }, { "epoch": 0.6167224080267558, "grad_norm": 0.9318669469946661, "learning_rate": 9.983901876628369e-06, "loss": 1.1813, "step": 461 }, { "epoch": 0.6180602006688963, "grad_norm": 1.171952638758082, "learning_rate": 9.983525531985343e-06, "loss": 1.0678, "step": 462 }, { "epoch": 0.6193979933110368, "grad_norm": 1.2845656943331367, "learning_rate": 9.983144846158472e-06, "loss": 1.2528, "step": 463 }, { "epoch": 0.6207357859531772, "grad_norm": 0.9331050598817028, "learning_rate": 9.982759819479375e-06, "loss": 1.1001, "step": 464 }, { "epoch": 0.6220735785953178, "grad_norm": 1.2195170474761798, "learning_rate": 9.982370452283451e-06, "loss": 1.2283, "step": 465 }, { "epoch": 0.6234113712374582, "grad_norm": 1.295182876902021, "learning_rate": 9.981976744909878e-06, "loss": 1.2799, "step": 466 }, { "epoch": 0.6247491638795987, "grad_norm": 1.0458027864025168, "learning_rate": 9.98157869770162e-06, "loss": 1.0312, "step": 467 }, { "epoch": 0.6260869565217392, "grad_norm": 1.1932705962995398, "learning_rate": 9.981176311005419e-06, "loss": 1.1319, "step": 468 }, { "epoch": 0.6274247491638796, "grad_norm": 0.9974230584841888, "learning_rate": 9.980769585171795e-06, "loss": 1.338, "step": 469 }, { "epoch": 0.6287625418060201, "grad_norm": 1.3905980363773291, "learning_rate": 9.980358520555048e-06, "loss": 1.3953, "step": 470 }, { "epoch": 0.6301003344481605, "grad_norm": 1.0971085546880637, "learning_rate": 9.979943117513265e-06, "loss": 1.1905, "step": 471 }, { "epoch": 0.631438127090301, "grad_norm": 1.0183611498542111, "learning_rate": 9.9795233764083e-06, "loss": 1.1635, "step": 472 }, { "epoch": 0.6327759197324415, "grad_norm": 1.4074865823839777, "learning_rate": 9.979099297605798e-06, "loss": 1.3124, "step": 473 }, { "epoch": 0.6341137123745819, "grad_norm": 1.004930186866696, "learning_rate": 9.978670881475173e-06, "loss": 1.2622, "step": 474 }, { "epoch": 0.6354515050167224, "grad_norm": 1.3527422647523326, "learning_rate": 9.978238128389623e-06, "loss": 1.1636, "step": 475 }, { "epoch": 0.6367892976588628, "grad_norm": 1.1388599786784637, "learning_rate": 9.977801038726123e-06, "loss": 1.23, "step": 476 }, { "epoch": 0.6381270903010033, "grad_norm": 1.1443829496135467, "learning_rate": 9.977359612865424e-06, "loss": 1.2255, "step": 477 }, { "epoch": 0.6394648829431439, "grad_norm": 0.8827965135698503, "learning_rate": 9.976913851192053e-06, "loss": 1.2244, "step": 478 }, { "epoch": 0.6408026755852843, "grad_norm": 1.2135411782913195, "learning_rate": 9.976463754094321e-06, "loss": 1.7073, "step": 479 }, { "epoch": 0.6421404682274248, "grad_norm": 1.1251507521684694, "learning_rate": 9.976009321964306e-06, "loss": 1.1929, "step": 480 }, { "epoch": 0.6434782608695652, "grad_norm": 1.4349042520454818, "learning_rate": 9.97555055519787e-06, "loss": 1.2269, "step": 481 }, { "epoch": 0.6448160535117057, "grad_norm": 1.1025834179587153, "learning_rate": 9.975087454194645e-06, "loss": 1.0182, "step": 482 }, { "epoch": 0.6461538461538462, "grad_norm": 1.0874312261493098, "learning_rate": 9.974620019358046e-06, "loss": 1.2064, "step": 483 }, { "epoch": 0.6474916387959866, "grad_norm": 0.9253402013587224, "learning_rate": 9.974148251095253e-06, "loss": 1.0463, "step": 484 }, { "epoch": 0.6488294314381271, "grad_norm": 1.2218517389432153, "learning_rate": 9.973672149817232e-06, "loss": 1.1593, "step": 485 }, { "epoch": 0.6501672240802676, "grad_norm": 1.1703826526240575, "learning_rate": 9.973191715938715e-06, "loss": 1.4027, "step": 486 }, { "epoch": 0.651505016722408, "grad_norm": 0.9193252422711862, "learning_rate": 9.972706949878212e-06, "loss": 1.1294, "step": 487 }, { "epoch": 0.6528428093645485, "grad_norm": 0.9194967768727665, "learning_rate": 9.972217852058006e-06, "loss": 1.1333, "step": 488 }, { "epoch": 0.6541806020066889, "grad_norm": 1.0786482750233548, "learning_rate": 9.971724422904154e-06, "loss": 1.124, "step": 489 }, { "epoch": 0.6555183946488294, "grad_norm": 1.0896868164545908, "learning_rate": 9.971226662846485e-06, "loss": 1.1783, "step": 490 }, { "epoch": 0.65685618729097, "grad_norm": 1.3461753262155707, "learning_rate": 9.970724572318602e-06, "loss": 1.18, "step": 491 }, { "epoch": 0.6581939799331104, "grad_norm": 1.0743624744456561, "learning_rate": 9.97021815175788e-06, "loss": 1.1465, "step": 492 }, { "epoch": 0.6595317725752509, "grad_norm": 1.17753086528407, "learning_rate": 9.969707401605464e-06, "loss": 1.0365, "step": 493 }, { "epoch": 0.6608695652173913, "grad_norm": 1.2923483683085784, "learning_rate": 9.969192322306271e-06, "loss": 1.325, "step": 494 }, { "epoch": 0.6622073578595318, "grad_norm": 1.0877745442956304, "learning_rate": 9.968672914308995e-06, "loss": 1.0952, "step": 495 }, { "epoch": 0.6635451505016723, "grad_norm": 0.9739845067358635, "learning_rate": 9.96814917806609e-06, "loss": 1.1839, "step": 496 }, { "epoch": 0.6648829431438127, "grad_norm": 0.9095856606113901, "learning_rate": 9.96762111403379e-06, "loss": 1.1059, "step": 497 }, { "epoch": 0.6662207357859532, "grad_norm": 0.9166518068480027, "learning_rate": 9.967088722672094e-06, "loss": 0.9872, "step": 498 }, { "epoch": 0.6675585284280936, "grad_norm": 1.245720644628499, "learning_rate": 9.966552004444772e-06, "loss": 1.3664, "step": 499 }, { "epoch": 0.6688963210702341, "grad_norm": 1.0071283363596744, "learning_rate": 9.966010959819363e-06, "loss": 1.1375, "step": 500 }, { "epoch": 0.6702341137123746, "grad_norm": 0.9549688939619114, "learning_rate": 9.965465589267176e-06, "loss": 1.0361, "step": 501 }, { "epoch": 0.671571906354515, "grad_norm": 1.1541351859720144, "learning_rate": 9.964915893263285e-06, "loss": 1.2606, "step": 502 }, { "epoch": 0.6729096989966555, "grad_norm": 0.973023279992627, "learning_rate": 9.964361872286534e-06, "loss": 1.1136, "step": 503 }, { "epoch": 0.6742474916387959, "grad_norm": 1.3308402373882993, "learning_rate": 9.963803526819537e-06, "loss": 1.03, "step": 504 }, { "epoch": 0.6755852842809364, "grad_norm": 1.0583220304696679, "learning_rate": 9.963240857348671e-06, "loss": 1.2051, "step": 505 }, { "epoch": 0.676923076923077, "grad_norm": 0.9396068850633498, "learning_rate": 9.962673864364081e-06, "loss": 1.0718, "step": 506 }, { "epoch": 0.6782608695652174, "grad_norm": 1.2179840213299078, "learning_rate": 9.96210254835968e-06, "loss": 1.063, "step": 507 }, { "epoch": 0.6795986622073579, "grad_norm": 1.0099653739425316, "learning_rate": 9.961526909833143e-06, "loss": 1.1354, "step": 508 }, { "epoch": 0.6809364548494983, "grad_norm": 0.905278702833277, "learning_rate": 9.960946949285915e-06, "loss": 1.0928, "step": 509 }, { "epoch": 0.6822742474916388, "grad_norm": 0.9528069771931961, "learning_rate": 9.960362667223202e-06, "loss": 1.0005, "step": 510 }, { "epoch": 0.6836120401337793, "grad_norm": 1.1294424087512411, "learning_rate": 9.959774064153977e-06, "loss": 1.1921, "step": 511 }, { "epoch": 0.6849498327759197, "grad_norm": 1.160689931550795, "learning_rate": 9.959181140590977e-06, "loss": 1.2544, "step": 512 }, { "epoch": 0.6862876254180602, "grad_norm": 0.9848642472275341, "learning_rate": 9.9585838970507e-06, "loss": 1.3328, "step": 513 }, { "epoch": 0.6876254180602007, "grad_norm": 0.9849719187307819, "learning_rate": 9.95798233405341e-06, "loss": 1.2841, "step": 514 }, { "epoch": 0.6889632107023411, "grad_norm": 0.9573892051977116, "learning_rate": 9.957376452123133e-06, "loss": 1.1036, "step": 515 }, { "epoch": 0.6903010033444816, "grad_norm": 0.9279145379957967, "learning_rate": 9.956766251787657e-06, "loss": 1.213, "step": 516 }, { "epoch": 0.691638795986622, "grad_norm": 1.1031284525939864, "learning_rate": 9.956151733578533e-06, "loss": 1.1917, "step": 517 }, { "epoch": 0.6929765886287625, "grad_norm": 0.9038401611266098, "learning_rate": 9.955532898031069e-06, "loss": 1.3808, "step": 518 }, { "epoch": 0.6943143812709031, "grad_norm": 0.8976268004052095, "learning_rate": 9.954909745684339e-06, "loss": 1.1009, "step": 519 }, { "epoch": 0.6956521739130435, "grad_norm": 0.9238469054634825, "learning_rate": 9.954282277081173e-06, "loss": 1.128, "step": 520 }, { "epoch": 0.696989966555184, "grad_norm": 1.0206641638606757, "learning_rate": 9.953650492768167e-06, "loss": 1.0612, "step": 521 }, { "epoch": 0.6983277591973244, "grad_norm": 0.9720990656282066, "learning_rate": 9.95301439329567e-06, "loss": 1.0757, "step": 522 }, { "epoch": 0.6996655518394649, "grad_norm": 1.1062674817848017, "learning_rate": 9.952373979217795e-06, "loss": 1.1528, "step": 523 }, { "epoch": 0.7010033444816054, "grad_norm": 1.013413617289394, "learning_rate": 9.951729251092408e-06, "loss": 1.1977, "step": 524 }, { "epoch": 0.7023411371237458, "grad_norm": 1.1688594219152204, "learning_rate": 9.951080209481138e-06, "loss": 1.0873, "step": 525 }, { "epoch": 0.7036789297658863, "grad_norm": 1.0364459435051967, "learning_rate": 9.950426854949371e-06, "loss": 1.0942, "step": 526 }, { "epoch": 0.7050167224080267, "grad_norm": 1.0079121966124014, "learning_rate": 9.949769188066247e-06, "loss": 1.3527, "step": 527 }, { "epoch": 0.7063545150501672, "grad_norm": 1.0612448442357814, "learning_rate": 9.949107209404664e-06, "loss": 1.3056, "step": 528 }, { "epoch": 0.7076923076923077, "grad_norm": 1.0936272070430593, "learning_rate": 9.948440919541277e-06, "loss": 1.0413, "step": 529 }, { "epoch": 0.7090301003344481, "grad_norm": 1.3083555450938178, "learning_rate": 9.947770319056496e-06, "loss": 1.3328, "step": 530 }, { "epoch": 0.7103678929765886, "grad_norm": 0.9353830351662236, "learning_rate": 9.947095408534483e-06, "loss": 1.1827, "step": 531 }, { "epoch": 0.711705685618729, "grad_norm": 0.9758561860687325, "learning_rate": 9.946416188563163e-06, "loss": 1.4016, "step": 532 }, { "epoch": 0.7130434782608696, "grad_norm": 1.0110154481566989, "learning_rate": 9.945732659734204e-06, "loss": 1.2355, "step": 533 }, { "epoch": 0.7143812709030101, "grad_norm": 0.9369626573333347, "learning_rate": 9.945044822643033e-06, "loss": 1.2793, "step": 534 }, { "epoch": 0.7157190635451505, "grad_norm": 1.0129976601822213, "learning_rate": 9.944352677888833e-06, "loss": 0.9829, "step": 535 }, { "epoch": 0.717056856187291, "grad_norm": 0.977251414985897, "learning_rate": 9.943656226074534e-06, "loss": 1.1012, "step": 536 }, { "epoch": 0.7183946488294315, "grad_norm": 1.1327594724305379, "learning_rate": 9.94295546780682e-06, "loss": 1.3853, "step": 537 }, { "epoch": 0.7197324414715719, "grad_norm": 1.0396136634539315, "learning_rate": 9.942250403696126e-06, "loss": 1.1819, "step": 538 }, { "epoch": 0.7210702341137124, "grad_norm": 1.1381876909137765, "learning_rate": 9.94154103435664e-06, "loss": 1.2021, "step": 539 }, { "epoch": 0.7224080267558528, "grad_norm": 1.1741482275177033, "learning_rate": 9.940827360406297e-06, "loss": 1.1952, "step": 540 }, { "epoch": 0.7237458193979933, "grad_norm": 1.1385043856276729, "learning_rate": 9.940109382466785e-06, "loss": 1.0588, "step": 541 }, { "epoch": 0.7250836120401338, "grad_norm": 1.0240670255213438, "learning_rate": 9.939387101163538e-06, "loss": 1.1192, "step": 542 }, { "epoch": 0.7264214046822742, "grad_norm": 0.910428916550738, "learning_rate": 9.93866051712574e-06, "loss": 1.0892, "step": 543 }, { "epoch": 0.7277591973244147, "grad_norm": 1.316199835473011, "learning_rate": 9.937929630986324e-06, "loss": 1.2626, "step": 544 }, { "epoch": 0.7290969899665551, "grad_norm": 0.9360830882519046, "learning_rate": 9.937194443381972e-06, "loss": 1.2297, "step": 545 }, { "epoch": 0.7304347826086957, "grad_norm": 1.2043046934799437, "learning_rate": 9.936454954953108e-06, "loss": 1.3662, "step": 546 }, { "epoch": 0.7317725752508362, "grad_norm": 1.0497397264672619, "learning_rate": 9.935711166343909e-06, "loss": 1.221, "step": 547 }, { "epoch": 0.7331103678929766, "grad_norm": 1.0258925449943213, "learning_rate": 9.934963078202289e-06, "loss": 1.1471, "step": 548 }, { "epoch": 0.7344481605351171, "grad_norm": 1.1581927211301382, "learning_rate": 9.934210691179918e-06, "loss": 1.1441, "step": 549 }, { "epoch": 0.7357859531772575, "grad_norm": 1.0492056325155064, "learning_rate": 9.933454005932204e-06, "loss": 1.1393, "step": 550 }, { "epoch": 0.737123745819398, "grad_norm": 1.1590657214970634, "learning_rate": 9.932693023118299e-06, "loss": 1.6682, "step": 551 }, { "epoch": 0.7384615384615385, "grad_norm": 1.0668146799343072, "learning_rate": 9.931927743401102e-06, "loss": 1.237, "step": 552 }, { "epoch": 0.7397993311036789, "grad_norm": 1.025553964343556, "learning_rate": 9.931158167447254e-06, "loss": 1.1489, "step": 553 }, { "epoch": 0.7411371237458194, "grad_norm": 0.9066638630246214, "learning_rate": 9.930384295927137e-06, "loss": 1.2373, "step": 554 }, { "epoch": 0.7424749163879598, "grad_norm": 0.9798093772567751, "learning_rate": 9.929606129514875e-06, "loss": 1.1959, "step": 555 }, { "epoch": 0.7438127090301003, "grad_norm": 1.1267558890154552, "learning_rate": 9.928823668888337e-06, "loss": 1.2176, "step": 556 }, { "epoch": 0.7451505016722408, "grad_norm": 1.337925075176693, "learning_rate": 9.928036914729129e-06, "loss": 1.0327, "step": 557 }, { "epoch": 0.7464882943143812, "grad_norm": 1.058543866651578, "learning_rate": 9.927245867722596e-06, "loss": 1.0753, "step": 558 }, { "epoch": 0.7478260869565218, "grad_norm": 1.0803821544600893, "learning_rate": 9.926450528557828e-06, "loss": 1.4577, "step": 559 }, { "epoch": 0.7491638795986622, "grad_norm": 1.054946960758878, "learning_rate": 9.925650897927646e-06, "loss": 1.2532, "step": 560 }, { "epoch": 0.7505016722408027, "grad_norm": 1.209567679000771, "learning_rate": 9.924846976528618e-06, "loss": 1.3871, "step": 561 }, { "epoch": 0.7518394648829432, "grad_norm": 1.027552282071518, "learning_rate": 9.924038765061042e-06, "loss": 1.138, "step": 562 }, { "epoch": 0.7531772575250836, "grad_norm": 1.145330187044862, "learning_rate": 9.923226264228958e-06, "loss": 1.2387, "step": 563 }, { "epoch": 0.7545150501672241, "grad_norm": 1.0598835009480434, "learning_rate": 9.922409474740142e-06, "loss": 1.4123, "step": 564 }, { "epoch": 0.7558528428093646, "grad_norm": 0.9342441765267993, "learning_rate": 9.921588397306105e-06, "loss": 1.0851, "step": 565 }, { "epoch": 0.757190635451505, "grad_norm": 0.9669328837678545, "learning_rate": 9.920763032642094e-06, "loss": 1.2588, "step": 566 }, { "epoch": 0.7585284280936455, "grad_norm": 1.1047397626640783, "learning_rate": 9.919933381467088e-06, "loss": 1.1856, "step": 567 }, { "epoch": 0.7598662207357859, "grad_norm": 0.9957790754468706, "learning_rate": 9.919099444503804e-06, "loss": 1.0517, "step": 568 }, { "epoch": 0.7612040133779264, "grad_norm": 1.3832207904962894, "learning_rate": 9.918261222478687e-06, "loss": 1.3141, "step": 569 }, { "epoch": 0.7625418060200669, "grad_norm": 0.9843289341190617, "learning_rate": 9.91741871612192e-06, "loss": 1.1259, "step": 570 }, { "epoch": 0.7638795986622073, "grad_norm": 1.016016100218584, "learning_rate": 9.916571926167417e-06, "loss": 1.2582, "step": 571 }, { "epoch": 0.7652173913043478, "grad_norm": 0.9633748184540706, "learning_rate": 9.915720853352821e-06, "loss": 1.171, "step": 572 }, { "epoch": 0.7665551839464882, "grad_norm": 1.1908295026627416, "learning_rate": 9.91486549841951e-06, "loss": 1.3272, "step": 573 }, { "epoch": 0.7678929765886288, "grad_norm": 0.8505250202248061, "learning_rate": 9.914005862112587e-06, "loss": 0.8131, "step": 574 }, { "epoch": 0.7692307692307693, "grad_norm": 0.937455188289705, "learning_rate": 9.913141945180888e-06, "loss": 1.0711, "step": 575 }, { "epoch": 0.7705685618729097, "grad_norm": 0.8779788649318374, "learning_rate": 9.912273748376976e-06, "loss": 1.1633, "step": 576 }, { "epoch": 0.7719063545150502, "grad_norm": 0.9512182349139586, "learning_rate": 9.911401272457145e-06, "loss": 1.1487, "step": 577 }, { "epoch": 0.7732441471571906, "grad_norm": 1.0104093376463148, "learning_rate": 9.910524518181416e-06, "loss": 1.0962, "step": 578 }, { "epoch": 0.7745819397993311, "grad_norm": 1.0305324917602166, "learning_rate": 9.909643486313533e-06, "loss": 1.1404, "step": 579 }, { "epoch": 0.7759197324414716, "grad_norm": 1.0455408762919678, "learning_rate": 9.908758177620972e-06, "loss": 1.2228, "step": 580 }, { "epoch": 0.777257525083612, "grad_norm": 0.9742655471982674, "learning_rate": 9.907868592874927e-06, "loss": 1.1097, "step": 581 }, { "epoch": 0.7785953177257525, "grad_norm": 0.8537518807967347, "learning_rate": 9.906974732850327e-06, "loss": 0.896, "step": 582 }, { "epoch": 0.7799331103678929, "grad_norm": 0.9663495646909485, "learning_rate": 9.906076598325815e-06, "loss": 1.1363, "step": 583 }, { "epoch": 0.7812709030100334, "grad_norm": 1.2375045223425296, "learning_rate": 9.905174190083763e-06, "loss": 1.4156, "step": 584 }, { "epoch": 0.782608695652174, "grad_norm": 1.5764510792062418, "learning_rate": 9.904267508910269e-06, "loss": 1.1571, "step": 585 }, { "epoch": 0.7839464882943143, "grad_norm": 1.1261514674181803, "learning_rate": 9.903356555595143e-06, "loss": 1.1688, "step": 586 }, { "epoch": 0.7852842809364549, "grad_norm": 1.0487094186622077, "learning_rate": 9.90244133093193e-06, "loss": 1.2924, "step": 587 }, { "epoch": 0.7866220735785954, "grad_norm": 1.0462817809824114, "learning_rate": 9.901521835717882e-06, "loss": 1.294, "step": 588 }, { "epoch": 0.7879598662207358, "grad_norm": 1.069464433630998, "learning_rate": 9.900598070753981e-06, "loss": 1.1601, "step": 589 }, { "epoch": 0.7892976588628763, "grad_norm": 1.1657186278518388, "learning_rate": 9.899670036844926e-06, "loss": 1.4394, "step": 590 }, { "epoch": 0.7906354515050167, "grad_norm": 1.0037130022081504, "learning_rate": 9.898737734799134e-06, "loss": 1.3087, "step": 591 }, { "epoch": 0.7919732441471572, "grad_norm": 0.994189473022677, "learning_rate": 9.897801165428736e-06, "loss": 1.1171, "step": 592 }, { "epoch": 0.7933110367892977, "grad_norm": 1.4660051651651518, "learning_rate": 9.896860329549585e-06, "loss": 1.2769, "step": 593 }, { "epoch": 0.7946488294314381, "grad_norm": 1.1817434656343826, "learning_rate": 9.895915227981254e-06, "loss": 1.094, "step": 594 }, { "epoch": 0.7959866220735786, "grad_norm": 1.0160523912303874, "learning_rate": 9.894965861547023e-06, "loss": 1.235, "step": 595 }, { "epoch": 0.797324414715719, "grad_norm": 1.0364796157931995, "learning_rate": 9.894012231073895e-06, "loss": 1.0702, "step": 596 }, { "epoch": 0.7986622073578595, "grad_norm": 0.9437935708217852, "learning_rate": 9.89305433739258e-06, "loss": 1.0795, "step": 597 }, { "epoch": 0.8, "grad_norm": 1.01149985446928, "learning_rate": 9.892092181337512e-06, "loss": 1.372, "step": 598 }, { "epoch": 0.8013377926421404, "grad_norm": 1.2364038760514962, "learning_rate": 9.891125763746824e-06, "loss": 1.2504, "step": 599 }, { "epoch": 0.802675585284281, "grad_norm": 1.010372750107206, "learning_rate": 9.890155085462376e-06, "loss": 1.2659, "step": 600 }, { "epoch": 0.8040133779264214, "grad_norm": 1.0591661453830967, "learning_rate": 9.889180147329731e-06, "loss": 1.0995, "step": 601 }, { "epoch": 0.8053511705685619, "grad_norm": 1.1678318531107514, "learning_rate": 9.888200950198162e-06, "loss": 1.0227, "step": 602 }, { "epoch": 0.8066889632107024, "grad_norm": 1.0336540763985116, "learning_rate": 9.887217494920655e-06, "loss": 1.0588, "step": 603 }, { "epoch": 0.8080267558528428, "grad_norm": 1.2646684085301063, "learning_rate": 9.886229782353907e-06, "loss": 1.2838, "step": 604 }, { "epoch": 0.8093645484949833, "grad_norm": 0.9048045455343249, "learning_rate": 9.88523781335832e-06, "loss": 1.0947, "step": 605 }, { "epoch": 0.8107023411371237, "grad_norm": 0.9865529775375038, "learning_rate": 9.884241588798004e-06, "loss": 1.2895, "step": 606 }, { "epoch": 0.8120401337792642, "grad_norm": 1.437371451010746, "learning_rate": 9.88324110954078e-06, "loss": 1.3675, "step": 607 }, { "epoch": 0.8133779264214047, "grad_norm": 1.015356862165022, "learning_rate": 9.88223637645817e-06, "loss": 1.2799, "step": 608 }, { "epoch": 0.8147157190635451, "grad_norm": 0.9267283771545635, "learning_rate": 9.881227390425404e-06, "loss": 1.1732, "step": 609 }, { "epoch": 0.8160535117056856, "grad_norm": 0.9390354378108255, "learning_rate": 9.880214152321417e-06, "loss": 1.2491, "step": 610 }, { "epoch": 0.8173913043478261, "grad_norm": 0.7758397143455721, "learning_rate": 9.879196663028847e-06, "loss": 1.0027, "step": 611 }, { "epoch": 0.8187290969899665, "grad_norm": 1.0975501295489378, "learning_rate": 9.87817492343404e-06, "loss": 1.2996, "step": 612 }, { "epoch": 0.820066889632107, "grad_norm": 0.958568511270922, "learning_rate": 9.877148934427037e-06, "loss": 1.2764, "step": 613 }, { "epoch": 0.8214046822742475, "grad_norm": 1.2112640732751092, "learning_rate": 9.876118696901585e-06, "loss": 1.3764, "step": 614 }, { "epoch": 0.822742474916388, "grad_norm": 0.9310384373744037, "learning_rate": 9.875084211755127e-06, "loss": 1.2756, "step": 615 }, { "epoch": 0.8240802675585285, "grad_norm": 1.0442088425569003, "learning_rate": 9.874045479888819e-06, "loss": 1.2984, "step": 616 }, { "epoch": 0.8254180602006689, "grad_norm": 1.032430621931905, "learning_rate": 9.873002502207502e-06, "loss": 1.1027, "step": 617 }, { "epoch": 0.8267558528428094, "grad_norm": 0.9489166947271507, "learning_rate": 9.871955279619721e-06, "loss": 1.2451, "step": 618 }, { "epoch": 0.8280936454849498, "grad_norm": 0.9584460373716269, "learning_rate": 9.87090381303772e-06, "loss": 1.1171, "step": 619 }, { "epoch": 0.8294314381270903, "grad_norm": 1.2026452265020962, "learning_rate": 9.86984810337744e-06, "loss": 1.1376, "step": 620 }, { "epoch": 0.8307692307692308, "grad_norm": 1.0012410173222073, "learning_rate": 9.868788151558513e-06, "loss": 1.4247, "step": 621 }, { "epoch": 0.8321070234113712, "grad_norm": 0.9499756152780574, "learning_rate": 9.867723958504275e-06, "loss": 1.1078, "step": 622 }, { "epoch": 0.8334448160535117, "grad_norm": 1.1309609959834643, "learning_rate": 9.86665552514175e-06, "loss": 1.2136, "step": 623 }, { "epoch": 0.8347826086956521, "grad_norm": 1.1826392316102614, "learning_rate": 9.865582852401659e-06, "loss": 1.2328, "step": 624 }, { "epoch": 0.8361204013377926, "grad_norm": 0.9359524520677287, "learning_rate": 9.86450594121841e-06, "loss": 1.1151, "step": 625 }, { "epoch": 0.8374581939799332, "grad_norm": 1.63758491195004, "learning_rate": 9.863424792530114e-06, "loss": 1.2172, "step": 626 }, { "epoch": 0.8387959866220736, "grad_norm": 1.052721772102972, "learning_rate": 9.862339407278564e-06, "loss": 1.3793, "step": 627 }, { "epoch": 0.8401337792642141, "grad_norm": 1.093955855872764, "learning_rate": 9.861249786409248e-06, "loss": 1.1461, "step": 628 }, { "epoch": 0.8414715719063545, "grad_norm": 0.9840779033887354, "learning_rate": 9.860155930871341e-06, "loss": 1.2899, "step": 629 }, { "epoch": 0.842809364548495, "grad_norm": 0.9887964710505275, "learning_rate": 9.859057841617709e-06, "loss": 1.1405, "step": 630 }, { "epoch": 0.8441471571906355, "grad_norm": 0.9860121573619206, "learning_rate": 9.857955519604906e-06, "loss": 1.2388, "step": 631 }, { "epoch": 0.8454849498327759, "grad_norm": 1.0191404074782313, "learning_rate": 9.856848965793168e-06, "loss": 1.3195, "step": 632 }, { "epoch": 0.8468227424749164, "grad_norm": 0.9230659675490308, "learning_rate": 9.855738181146427e-06, "loss": 1.1839, "step": 633 }, { "epoch": 0.8481605351170568, "grad_norm": 1.337445510382536, "learning_rate": 9.854623166632296e-06, "loss": 1.3889, "step": 634 }, { "epoch": 0.8494983277591973, "grad_norm": 1.0997790599749402, "learning_rate": 9.853503923222066e-06, "loss": 1.2648, "step": 635 }, { "epoch": 0.8508361204013378, "grad_norm": 0.9874582813633553, "learning_rate": 9.852380451890723e-06, "loss": 1.0541, "step": 636 }, { "epoch": 0.8521739130434782, "grad_norm": 0.9610201823235474, "learning_rate": 9.851252753616928e-06, "loss": 1.1751, "step": 637 }, { "epoch": 0.8535117056856187, "grad_norm": 1.0487975170192587, "learning_rate": 9.850120829383027e-06, "loss": 1.1861, "step": 638 }, { "epoch": 0.8548494983277592, "grad_norm": 1.0535883038163176, "learning_rate": 9.848984680175049e-06, "loss": 1.1886, "step": 639 }, { "epoch": 0.8561872909698997, "grad_norm": 1.0852938871562892, "learning_rate": 9.847844306982698e-06, "loss": 1.0049, "step": 640 }, { "epoch": 0.8575250836120402, "grad_norm": 0.9846737582212536, "learning_rate": 9.846699710799365e-06, "loss": 1.0314, "step": 641 }, { "epoch": 0.8588628762541806, "grad_norm": 1.04837205148603, "learning_rate": 9.845550892622113e-06, "loss": 1.1988, "step": 642 }, { "epoch": 0.8602006688963211, "grad_norm": 1.195018397904588, "learning_rate": 9.844397853451687e-06, "loss": 0.9557, "step": 643 }, { "epoch": 0.8615384615384616, "grad_norm": 0.9957745872206203, "learning_rate": 9.843240594292507e-06, "loss": 0.9541, "step": 644 }, { "epoch": 0.862876254180602, "grad_norm": 2.623783959336471, "learning_rate": 9.84207911615267e-06, "loss": 1.4788, "step": 645 }, { "epoch": 0.8642140468227425, "grad_norm": 1.187987575629985, "learning_rate": 9.840913420043945e-06, "loss": 1.1591, "step": 646 }, { "epoch": 0.8655518394648829, "grad_norm": 1.0920584812669722, "learning_rate": 9.839743506981783e-06, "loss": 1.3126, "step": 647 }, { "epoch": 0.8668896321070234, "grad_norm": 1.1832742890589614, "learning_rate": 9.838569377985298e-06, "loss": 1.0795, "step": 648 }, { "epoch": 0.8682274247491639, "grad_norm": 1.0363042312199444, "learning_rate": 9.837391034077286e-06, "loss": 1.1162, "step": 649 }, { "epoch": 0.8695652173913043, "grad_norm": 0.9555267224848123, "learning_rate": 9.836208476284208e-06, "loss": 1.1598, "step": 650 }, { "epoch": 0.8709030100334448, "grad_norm": 0.8767782336961009, "learning_rate": 9.835021705636201e-06, "loss": 0.9192, "step": 651 }, { "epoch": 0.8722408026755852, "grad_norm": 0.9644007359125557, "learning_rate": 9.833830723167067e-06, "loss": 1.1107, "step": 652 }, { "epoch": 0.8735785953177257, "grad_norm": 1.1040343631316711, "learning_rate": 9.83263552991428e-06, "loss": 0.964, "step": 653 }, { "epoch": 0.8749163879598663, "grad_norm": 1.3609993858329352, "learning_rate": 9.83143612691898e-06, "loss": 1.0381, "step": 654 }, { "epoch": 0.8762541806020067, "grad_norm": 1.1524129373591085, "learning_rate": 9.830232515225976e-06, "loss": 1.0936, "step": 655 }, { "epoch": 0.8775919732441472, "grad_norm": 1.1083192431380482, "learning_rate": 9.829024695883746e-06, "loss": 1.283, "step": 656 }, { "epoch": 0.8789297658862876, "grad_norm": 0.931560053220981, "learning_rate": 9.827812669944423e-06, "loss": 1.2189, "step": 657 }, { "epoch": 0.8802675585284281, "grad_norm": 1.1907417713552158, "learning_rate": 9.826596438463818e-06, "loss": 1.4279, "step": 658 }, { "epoch": 0.8816053511705686, "grad_norm": 1.0736222305587315, "learning_rate": 9.825376002501393e-06, "loss": 0.981, "step": 659 }, { "epoch": 0.882943143812709, "grad_norm": 1.152322269410841, "learning_rate": 9.824151363120283e-06, "loss": 1.2287, "step": 660 }, { "epoch": 0.8842809364548495, "grad_norm": 1.2400212737028145, "learning_rate": 9.822922521387277e-06, "loss": 1.4516, "step": 661 }, { "epoch": 0.88561872909699, "grad_norm": 1.2140849639939946, "learning_rate": 9.821689478372827e-06, "loss": 1.1112, "step": 662 }, { "epoch": 0.8869565217391304, "grad_norm": 1.3174482906181668, "learning_rate": 9.82045223515105e-06, "loss": 1.2981, "step": 663 }, { "epoch": 0.8882943143812709, "grad_norm": 1.1359070296211902, "learning_rate": 9.819210792799711e-06, "loss": 1.2935, "step": 664 }, { "epoch": 0.8896321070234113, "grad_norm": 0.9988657352146671, "learning_rate": 9.817965152400244e-06, "loss": 1.081, "step": 665 }, { "epoch": 0.8909698996655518, "grad_norm": 1.009326511806942, "learning_rate": 9.816715315037733e-06, "loss": 1.0283, "step": 666 }, { "epoch": 0.8923076923076924, "grad_norm": 0.9748309506972497, "learning_rate": 9.815461281800917e-06, "loss": 1.3762, "step": 667 }, { "epoch": 0.8936454849498328, "grad_norm": 0.926074360078927, "learning_rate": 9.814203053782201e-06, "loss": 1.1209, "step": 668 }, { "epoch": 0.8949832775919733, "grad_norm": 0.9616386615509389, "learning_rate": 9.812940632077629e-06, "loss": 1.1824, "step": 669 }, { "epoch": 0.8963210702341137, "grad_norm": 1.0856196970841454, "learning_rate": 9.811674017786908e-06, "loss": 1.4442, "step": 670 }, { "epoch": 0.8976588628762542, "grad_norm": 1.2432021017671862, "learning_rate": 9.810403212013395e-06, "loss": 1.1755, "step": 671 }, { "epoch": 0.8989966555183947, "grad_norm": 0.9830775627395631, "learning_rate": 9.809128215864096e-06, "loss": 1.2645, "step": 672 }, { "epoch": 0.9003344481605351, "grad_norm": 1.051863253066761, "learning_rate": 9.807849030449671e-06, "loss": 1.2007, "step": 673 }, { "epoch": 0.9016722408026756, "grad_norm": 1.0549347118056278, "learning_rate": 9.806565656884426e-06, "loss": 1.1851, "step": 674 }, { "epoch": 0.903010033444816, "grad_norm": 1.1752115222157493, "learning_rate": 9.805278096286318e-06, "loss": 1.1542, "step": 675 }, { "epoch": 0.9043478260869565, "grad_norm": 1.1080216839406412, "learning_rate": 9.803986349776948e-06, "loss": 1.303, "step": 676 }, { "epoch": 0.905685618729097, "grad_norm": 0.8495055989690289, "learning_rate": 9.802690418481569e-06, "loss": 1.0052, "step": 677 }, { "epoch": 0.9070234113712374, "grad_norm": 1.1443128780878427, "learning_rate": 9.80139030352907e-06, "loss": 1.461, "step": 678 }, { "epoch": 0.9083612040133779, "grad_norm": 0.851958904855734, "learning_rate": 9.800086006051996e-06, "loss": 1.084, "step": 679 }, { "epoch": 0.9096989966555183, "grad_norm": 0.9726349208012318, "learning_rate": 9.798777527186527e-06, "loss": 1.223, "step": 680 }, { "epoch": 0.9110367892976589, "grad_norm": 1.6156052765378088, "learning_rate": 9.797464868072489e-06, "loss": 1.0188, "step": 681 }, { "epoch": 0.9123745819397994, "grad_norm": 1.011043739936154, "learning_rate": 9.796148029853345e-06, "loss": 0.9256, "step": 682 }, { "epoch": 0.9137123745819398, "grad_norm": 1.2650744438156902, "learning_rate": 9.794827013676206e-06, "loss": 1.3598, "step": 683 }, { "epoch": 0.9150501672240803, "grad_norm": 1.076243883701245, "learning_rate": 9.793501820691818e-06, "loss": 1.1814, "step": 684 }, { "epoch": 0.9163879598662207, "grad_norm": 0.9082656247988505, "learning_rate": 9.792172452054565e-06, "loss": 1.0986, "step": 685 }, { "epoch": 0.9177257525083612, "grad_norm": 0.9060472251756602, "learning_rate": 9.790838908922468e-06, "loss": 1.1307, "step": 686 }, { "epoch": 0.9190635451505017, "grad_norm": 0.9538689227947832, "learning_rate": 9.789501192457188e-06, "loss": 1.1956, "step": 687 }, { "epoch": 0.9204013377926421, "grad_norm": 0.8997324462454521, "learning_rate": 9.788159303824018e-06, "loss": 1.0314, "step": 688 }, { "epoch": 0.9217391304347826, "grad_norm": 0.9556582001296928, "learning_rate": 9.786813244191885e-06, "loss": 1.0832, "step": 689 }, { "epoch": 0.9230769230769231, "grad_norm": 0.9359237452425293, "learning_rate": 9.785463014733356e-06, "loss": 1.2016, "step": 690 }, { "epoch": 0.9244147157190635, "grad_norm": 0.8326924247918515, "learning_rate": 9.78410861662462e-06, "loss": 0.9061, "step": 691 }, { "epoch": 0.925752508361204, "grad_norm": 1.1690512254909176, "learning_rate": 9.782750051045504e-06, "loss": 0.8892, "step": 692 }, { "epoch": 0.9270903010033444, "grad_norm": 1.0046587384279455, "learning_rate": 9.781387319179465e-06, "loss": 1.1656, "step": 693 }, { "epoch": 0.928428093645485, "grad_norm": 1.2485150097353308, "learning_rate": 9.78002042221359e-06, "loss": 0.9339, "step": 694 }, { "epoch": 0.9297658862876255, "grad_norm": 0.950760809910313, "learning_rate": 9.778649361338588e-06, "loss": 1.3059, "step": 695 }, { "epoch": 0.9311036789297659, "grad_norm": 1.0258541445453178, "learning_rate": 9.777274137748802e-06, "loss": 1.2519, "step": 696 }, { "epoch": 0.9324414715719064, "grad_norm": 1.1693613657390223, "learning_rate": 9.775894752642199e-06, "loss": 1.2794, "step": 697 }, { "epoch": 0.9337792642140468, "grad_norm": 1.116376806692642, "learning_rate": 9.774511207220369e-06, "loss": 1.1119, "step": 698 }, { "epoch": 0.9351170568561873, "grad_norm": 1.00372876701593, "learning_rate": 9.773123502688532e-06, "loss": 1.12, "step": 699 }, { "epoch": 0.9364548494983278, "grad_norm": 1.18124217949937, "learning_rate": 9.771731640255525e-06, "loss": 0.9667, "step": 700 }, { "epoch": 0.9377926421404682, "grad_norm": 1.0244923545015718, "learning_rate": 9.770335621133808e-06, "loss": 1.0506, "step": 701 }, { "epoch": 0.9391304347826087, "grad_norm": 0.9583148071274727, "learning_rate": 9.768935446539464e-06, "loss": 1.3069, "step": 702 }, { "epoch": 0.9404682274247491, "grad_norm": 0.9545288450998501, "learning_rate": 9.767531117692196e-06, "loss": 1.3368, "step": 703 }, { "epoch": 0.9418060200668896, "grad_norm": 1.296229209934714, "learning_rate": 9.766122635815322e-06, "loss": 1.4115, "step": 704 }, { "epoch": 0.9431438127090301, "grad_norm": 1.1002739676081217, "learning_rate": 9.764710002135784e-06, "loss": 0.9887, "step": 705 }, { "epoch": 0.9444816053511705, "grad_norm": 1.0180674681651405, "learning_rate": 9.763293217884133e-06, "loss": 1.0949, "step": 706 }, { "epoch": 0.945819397993311, "grad_norm": 1.167401790239325, "learning_rate": 9.761872284294542e-06, "loss": 1.1099, "step": 707 }, { "epoch": 0.9471571906354515, "grad_norm": 1.2906725541994704, "learning_rate": 9.760447202604796e-06, "loss": 1.085, "step": 708 }, { "epoch": 0.948494983277592, "grad_norm": 0.970864770528708, "learning_rate": 9.759017974056292e-06, "loss": 1.223, "step": 709 }, { "epoch": 0.9498327759197325, "grad_norm": 1.1164613347044747, "learning_rate": 9.757584599894045e-06, "loss": 1.1954, "step": 710 }, { "epoch": 0.9511705685618729, "grad_norm": 1.0655843728717254, "learning_rate": 9.756147081366673e-06, "loss": 1.262, "step": 711 }, { "epoch": 0.9525083612040134, "grad_norm": 1.044031545026209, "learning_rate": 9.75470541972641e-06, "loss": 1.1687, "step": 712 }, { "epoch": 0.9538461538461539, "grad_norm": 1.1992017112356412, "learning_rate": 9.753259616229096e-06, "loss": 1.3466, "step": 713 }, { "epoch": 0.9551839464882943, "grad_norm": 1.032172654098663, "learning_rate": 9.751809672134184e-06, "loss": 1.5412, "step": 714 }, { "epoch": 0.9565217391304348, "grad_norm": 1.1027236373345535, "learning_rate": 9.750355588704728e-06, "loss": 1.0804, "step": 715 }, { "epoch": 0.9578595317725752, "grad_norm": 0.9769958741069888, "learning_rate": 9.748897367207391e-06, "loss": 1.2932, "step": 716 }, { "epoch": 0.9591973244147157, "grad_norm": 0.9415186409526449, "learning_rate": 9.747435008912438e-06, "loss": 1.0295, "step": 717 }, { "epoch": 0.9605351170568562, "grad_norm": 0.9473867298857055, "learning_rate": 9.745968515093741e-06, "loss": 1.0263, "step": 718 }, { "epoch": 0.9618729096989966, "grad_norm": 0.9424872115272992, "learning_rate": 9.744497887028774e-06, "loss": 1.201, "step": 719 }, { "epoch": 0.9632107023411371, "grad_norm": 1.0992987613328598, "learning_rate": 9.74302312599861e-06, "loss": 1.4418, "step": 720 }, { "epoch": 0.9645484949832775, "grad_norm": 0.9029941170471982, "learning_rate": 9.741544233287924e-06, "loss": 1.0173, "step": 721 }, { "epoch": 0.9658862876254181, "grad_norm": 0.8319462560564428, "learning_rate": 9.74006121018499e-06, "loss": 0.8575, "step": 722 }, { "epoch": 0.9672240802675586, "grad_norm": 0.9773509161903545, "learning_rate": 9.73857405798168e-06, "loss": 1.217, "step": 723 }, { "epoch": 0.968561872909699, "grad_norm": 1.0006838987780111, "learning_rate": 9.73708277797346e-06, "loss": 1.3616, "step": 724 }, { "epoch": 0.9698996655518395, "grad_norm": 0.9737005904910055, "learning_rate": 9.735587371459399e-06, "loss": 1.2078, "step": 725 }, { "epoch": 0.9712374581939799, "grad_norm": 1.0258262445589679, "learning_rate": 9.734087839742152e-06, "loss": 1.0354, "step": 726 }, { "epoch": 0.9725752508361204, "grad_norm": 1.0751183342511959, "learning_rate": 9.732584184127973e-06, "loss": 1.3379, "step": 727 }, { "epoch": 0.9739130434782609, "grad_norm": 0.9445482302291927, "learning_rate": 9.731076405926706e-06, "loss": 1.1237, "step": 728 }, { "epoch": 0.9752508361204013, "grad_norm": 0.9197841326726465, "learning_rate": 9.729564506451791e-06, "loss": 1.0121, "step": 729 }, { "epoch": 0.9765886287625418, "grad_norm": 1.0986332743474945, "learning_rate": 9.72804848702025e-06, "loss": 1.3125, "step": 730 }, { "epoch": 0.9779264214046822, "grad_norm": 1.41339410414861, "learning_rate": 9.7265283489527e-06, "loss": 1.1952, "step": 731 }, { "epoch": 0.9792642140468227, "grad_norm": 1.2804842658240465, "learning_rate": 9.725004093573343e-06, "loss": 1.3374, "step": 732 }, { "epoch": 0.9806020066889632, "grad_norm": 1.0297098118475143, "learning_rate": 9.72347572220997e-06, "loss": 1.5325, "step": 733 }, { "epoch": 0.9819397993311036, "grad_norm": 1.1114975939734417, "learning_rate": 9.721943236193952e-06, "loss": 1.254, "step": 734 }, { "epoch": 0.9832775919732442, "grad_norm": 1.09901986349397, "learning_rate": 9.720406636860252e-06, "loss": 1.2009, "step": 735 }, { "epoch": 0.9846153846153847, "grad_norm": 1.213670835739425, "learning_rate": 9.718865925547411e-06, "loss": 1.3843, "step": 736 }, { "epoch": 0.9859531772575251, "grad_norm": 1.1157562841880595, "learning_rate": 9.717321103597556e-06, "loss": 1.2804, "step": 737 }, { "epoch": 0.9872909698996656, "grad_norm": 1.0770936592338518, "learning_rate": 9.715772172356388e-06, "loss": 1.2369, "step": 738 }, { "epoch": 0.988628762541806, "grad_norm": 1.098161678153236, "learning_rate": 9.714219133173194e-06, "loss": 1.1639, "step": 739 }, { "epoch": 0.9899665551839465, "grad_norm": 1.0478582987733194, "learning_rate": 9.712661987400838e-06, "loss": 1.1853, "step": 740 }, { "epoch": 0.991304347826087, "grad_norm": 0.9730376626552903, "learning_rate": 9.711100736395758e-06, "loss": 1.233, "step": 741 }, { "epoch": 0.9926421404682274, "grad_norm": 1.013803692440576, "learning_rate": 9.709535381517973e-06, "loss": 1.4875, "step": 742 }, { "epoch": 0.9939799331103679, "grad_norm": 1.3620815657427228, "learning_rate": 9.707965924131074e-06, "loss": 1.2778, "step": 743 }, { "epoch": 0.9953177257525083, "grad_norm": 1.005090856341657, "learning_rate": 9.706392365602224e-06, "loss": 1.1868, "step": 744 }, { "epoch": 0.9966555183946488, "grad_norm": 1.010059168348425, "learning_rate": 9.704814707302166e-06, "loss": 1.204, "step": 745 }, { "epoch": 0.9979933110367893, "grad_norm": 0.9928158472267319, "learning_rate": 9.703232950605203e-06, "loss": 1.0633, "step": 746 }, { "epoch": 0.9993311036789297, "grad_norm": 1.0394222300326905, "learning_rate": 9.70164709688922e-06, "loss": 1.3204, "step": 747 }, { "epoch": 1.0, "grad_norm": 1.0394222300326905, "learning_rate": 9.70005714753566e-06, "loss": 1.1957, "step": 748 }, { "epoch": 1.0013377926421405, "grad_norm": 1.6408764682701042, "learning_rate": 9.698463103929542e-06, "loss": 1.1639, "step": 749 }, { "epoch": 1.002675585284281, "grad_norm": 0.9986871663193632, "learning_rate": 9.69686496745945e-06, "loss": 1.0556, "step": 750 }, { "epoch": 1.0040133779264213, "grad_norm": 1.0050611036793766, "learning_rate": 9.695262739517528e-06, "loss": 1.0091, "step": 751 }, { "epoch": 1.0053511705685618, "grad_norm": 0.7495931778855482, "learning_rate": 9.69365642149949e-06, "loss": 1.3198, "step": 752 }, { "epoch": 1.0066889632107023, "grad_norm": 1.0479566546770847, "learning_rate": 9.69204601480461e-06, "loss": 1.04, "step": 753 }, { "epoch": 1.0080267558528428, "grad_norm": 1.053132617077858, "learning_rate": 9.690431520835725e-06, "loss": 1.0764, "step": 754 }, { "epoch": 1.0093645484949834, "grad_norm": 0.9465668124282555, "learning_rate": 9.688812940999232e-06, "loss": 1.1983, "step": 755 }, { "epoch": 1.0107023411371236, "grad_norm": 0.9950541402919411, "learning_rate": 9.687190276705088e-06, "loss": 1.0921, "step": 756 }, { "epoch": 1.0120401337792642, "grad_norm": 1.1839286226927221, "learning_rate": 9.685563529366806e-06, "loss": 1.3702, "step": 757 }, { "epoch": 1.0133779264214047, "grad_norm": 0.9043987800002394, "learning_rate": 9.683932700401457e-06, "loss": 1.0289, "step": 758 }, { "epoch": 1.0147157190635452, "grad_norm": 1.1765735230079775, "learning_rate": 9.682297791229668e-06, "loss": 1.2154, "step": 759 }, { "epoch": 1.0160535117056857, "grad_norm": 1.0174099426597931, "learning_rate": 9.68065880327562e-06, "loss": 1.115, "step": 760 }, { "epoch": 1.017391304347826, "grad_norm": 1.087321549880045, "learning_rate": 9.679015737967046e-06, "loss": 1.1258, "step": 761 }, { "epoch": 1.0187290969899665, "grad_norm": 1.224619779654506, "learning_rate": 9.677368596735232e-06, "loss": 1.2102, "step": 762 }, { "epoch": 1.020066889632107, "grad_norm": 0.8711385280783899, "learning_rate": 9.675717381015014e-06, "loss": 0.8649, "step": 763 }, { "epoch": 1.0214046822742475, "grad_norm": 1.0222722003710831, "learning_rate": 9.674062092244779e-06, "loss": 1.1274, "step": 764 }, { "epoch": 1.022742474916388, "grad_norm": 1.0795141815297804, "learning_rate": 9.67240273186646e-06, "loss": 1.3098, "step": 765 }, { "epoch": 1.0240802675585283, "grad_norm": 1.0752063100846632, "learning_rate": 9.670739301325534e-06, "loss": 1.105, "step": 766 }, { "epoch": 1.0254180602006688, "grad_norm": 1.1540608745412908, "learning_rate": 9.669071802071032e-06, "loss": 0.9314, "step": 767 }, { "epoch": 1.0267558528428093, "grad_norm": 1.1516185266778138, "learning_rate": 9.66740023555552e-06, "loss": 1.0527, "step": 768 }, { "epoch": 1.0280936454849499, "grad_norm": 1.0550912880342065, "learning_rate": 9.665724603235115e-06, "loss": 0.8196, "step": 769 }, { "epoch": 1.0294314381270904, "grad_norm": 0.9063643704754969, "learning_rate": 9.66404490656947e-06, "loss": 1.2121, "step": 770 }, { "epoch": 1.0307692307692307, "grad_norm": 1.0499869357373308, "learning_rate": 9.66236114702178e-06, "loss": 1.1471, "step": 771 }, { "epoch": 1.0321070234113712, "grad_norm": 0.9647138316411162, "learning_rate": 9.66067332605878e-06, "loss": 1.1688, "step": 772 }, { "epoch": 1.0334448160535117, "grad_norm": 0.9660690337663761, "learning_rate": 9.658981445150744e-06, "loss": 1.2949, "step": 773 }, { "epoch": 1.0347826086956522, "grad_norm": 1.0444635514745335, "learning_rate": 9.65728550577148e-06, "loss": 1.1788, "step": 774 }, { "epoch": 1.0361204013377927, "grad_norm": 1.1475649966565253, "learning_rate": 9.655585509398334e-06, "loss": 1.0751, "step": 775 }, { "epoch": 1.037458193979933, "grad_norm": 1.0671960969200227, "learning_rate": 9.65388145751218e-06, "loss": 1.1433, "step": 776 }, { "epoch": 1.0387959866220735, "grad_norm": 1.0764237507217969, "learning_rate": 9.652173351597435e-06, "loss": 1.0925, "step": 777 }, { "epoch": 1.040133779264214, "grad_norm": 1.1186060645699878, "learning_rate": 9.650461193142042e-06, "loss": 1.1132, "step": 778 }, { "epoch": 1.0414715719063545, "grad_norm": 0.6544335363296623, "learning_rate": 9.648744983637471e-06, "loss": 1.0562, "step": 779 }, { "epoch": 1.042809364548495, "grad_norm": 1.3907488948756037, "learning_rate": 9.647024724578724e-06, "loss": 1.1951, "step": 780 }, { "epoch": 1.0441471571906356, "grad_norm": 0.8502982062457299, "learning_rate": 9.645300417464332e-06, "loss": 1.0556, "step": 781 }, { "epoch": 1.0454849498327758, "grad_norm": 1.1354147919657973, "learning_rate": 9.643572063796352e-06, "loss": 1.0825, "step": 782 }, { "epoch": 1.0468227424749164, "grad_norm": 1.010988528502207, "learning_rate": 9.641839665080363e-06, "loss": 1.1542, "step": 783 }, { "epoch": 1.0481605351170569, "grad_norm": 0.9839939249469988, "learning_rate": 9.640103222825472e-06, "loss": 1.1856, "step": 784 }, { "epoch": 1.0494983277591974, "grad_norm": 1.0256844282049855, "learning_rate": 9.638362738544302e-06, "loss": 1.2218, "step": 785 }, { "epoch": 1.050836120401338, "grad_norm": 0.7329901252105672, "learning_rate": 9.636618213753006e-06, "loss": 1.0696, "step": 786 }, { "epoch": 1.0521739130434782, "grad_norm": 1.0260663997677806, "learning_rate": 9.634869649971247e-06, "loss": 1.3372, "step": 787 }, { "epoch": 1.0535117056856187, "grad_norm": 1.2981813417048746, "learning_rate": 9.633117048722213e-06, "loss": 0.9215, "step": 788 }, { "epoch": 1.0548494983277592, "grad_norm": 0.9517028873297264, "learning_rate": 9.631360411532609e-06, "loss": 1.2883, "step": 789 }, { "epoch": 1.0561872909698997, "grad_norm": 0.9919806212230781, "learning_rate": 9.629599739932652e-06, "loss": 1.04, "step": 790 }, { "epoch": 1.0575250836120402, "grad_norm": 0.8700407578427113, "learning_rate": 9.627835035456074e-06, "loss": 1.192, "step": 791 }, { "epoch": 1.0588628762541805, "grad_norm": 0.9537573140231194, "learning_rate": 9.626066299640124e-06, "loss": 1.022, "step": 792 }, { "epoch": 1.060200668896321, "grad_norm": 1.0975691566801855, "learning_rate": 9.62429353402556e-06, "loss": 0.9994, "step": 793 }, { "epoch": 1.0615384615384615, "grad_norm": 0.9226170598096517, "learning_rate": 9.62251674015665e-06, "loss": 0.9967, "step": 794 }, { "epoch": 1.062876254180602, "grad_norm": 0.967626012882805, "learning_rate": 9.620735919581168e-06, "loss": 0.8248, "step": 795 }, { "epoch": 1.0642140468227426, "grad_norm": 0.9586446643087508, "learning_rate": 9.618951073850404e-06, "loss": 1.167, "step": 796 }, { "epoch": 1.0655518394648829, "grad_norm": 1.1841251169343072, "learning_rate": 9.617162204519147e-06, "loss": 0.9226, "step": 797 }, { "epoch": 1.0668896321070234, "grad_norm": 1.4467652858312365, "learning_rate": 9.615369313145695e-06, "loss": 1.1059, "step": 798 }, { "epoch": 1.0682274247491639, "grad_norm": 1.0408468446113, "learning_rate": 9.61357240129185e-06, "loss": 1.0336, "step": 799 }, { "epoch": 1.0695652173913044, "grad_norm": 1.082952331861326, "learning_rate": 9.611771470522908e-06, "loss": 1.2171, "step": 800 }, { "epoch": 1.070903010033445, "grad_norm": 0.940071124432128, "learning_rate": 9.609966522407678e-06, "loss": 1.1014, "step": 801 }, { "epoch": 1.0722408026755852, "grad_norm": 0.8898724931425187, "learning_rate": 9.60815755851846e-06, "loss": 1.07, "step": 802 }, { "epoch": 1.0735785953177257, "grad_norm": 1.0255812636112083, "learning_rate": 9.60634458043106e-06, "loss": 1.1243, "step": 803 }, { "epoch": 1.0749163879598662, "grad_norm": 1.084959639355751, "learning_rate": 9.60452758972477e-06, "loss": 1.1234, "step": 804 }, { "epoch": 1.0762541806020067, "grad_norm": 1.097076333845879, "learning_rate": 9.602706587982384e-06, "loss": 1.0338, "step": 805 }, { "epoch": 1.0775919732441472, "grad_norm": 1.0272909031251145, "learning_rate": 9.600881576790194e-06, "loss": 1.1076, "step": 806 }, { "epoch": 1.0789297658862875, "grad_norm": 0.914577049273855, "learning_rate": 9.599052557737973e-06, "loss": 1.2142, "step": 807 }, { "epoch": 1.080267558528428, "grad_norm": 1.1377812179466673, "learning_rate": 9.597219532418997e-06, "loss": 0.9761, "step": 808 }, { "epoch": 1.0816053511705686, "grad_norm": 0.7539738507619589, "learning_rate": 9.59538250243003e-06, "loss": 0.9145, "step": 809 }, { "epoch": 1.082943143812709, "grad_norm": 1.0840297035797992, "learning_rate": 9.593541469371313e-06, "loss": 1.1941, "step": 810 }, { "epoch": 1.0842809364548496, "grad_norm": 0.9145863729580779, "learning_rate": 9.591696434846589e-06, "loss": 1.2826, "step": 811 }, { "epoch": 1.0856187290969899, "grad_norm": 1.1587921612803063, "learning_rate": 9.589847400463079e-06, "loss": 1.0963, "step": 812 }, { "epoch": 1.0869565217391304, "grad_norm": 0.9692056094010871, "learning_rate": 9.58799436783149e-06, "loss": 1.0991, "step": 813 }, { "epoch": 1.0882943143812709, "grad_norm": 0.9552399945032205, "learning_rate": 9.586137338566012e-06, "loss": 1.2025, "step": 814 }, { "epoch": 1.0896321070234114, "grad_norm": 1.169027297900495, "learning_rate": 9.584276314284316e-06, "loss": 1.0992, "step": 815 }, { "epoch": 1.090969899665552, "grad_norm": 0.8881842659440236, "learning_rate": 9.58241129660755e-06, "loss": 1.2728, "step": 816 }, { "epoch": 1.0923076923076924, "grad_norm": 0.9263228036846691, "learning_rate": 9.580542287160348e-06, "loss": 0.9823, "step": 817 }, { "epoch": 1.0936454849498327, "grad_norm": 1.3012451866912513, "learning_rate": 9.578669287570817e-06, "loss": 1.2993, "step": 818 }, { "epoch": 1.0949832775919732, "grad_norm": 0.9889008568013318, "learning_rate": 9.576792299470537e-06, "loss": 1.0028, "step": 819 }, { "epoch": 1.0963210702341137, "grad_norm": 0.7102897326399057, "learning_rate": 9.574911324494569e-06, "loss": 1.0844, "step": 820 }, { "epoch": 1.0976588628762542, "grad_norm": 0.8614614807966973, "learning_rate": 9.573026364281441e-06, "loss": 1.3268, "step": 821 }, { "epoch": 1.0989966555183948, "grad_norm": 1.2474999817391657, "learning_rate": 9.571137420473154e-06, "loss": 1.1072, "step": 822 }, { "epoch": 1.100334448160535, "grad_norm": 1.1269807264319953, "learning_rate": 9.569244494715183e-06, "loss": 1.0698, "step": 823 }, { "epoch": 1.1016722408026756, "grad_norm": 1.2072597212966658, "learning_rate": 9.567347588656468e-06, "loss": 0.8886, "step": 824 }, { "epoch": 1.103010033444816, "grad_norm": 0.9104453841445563, "learning_rate": 9.565446703949417e-06, "loss": 0.9279, "step": 825 }, { "epoch": 1.1043478260869566, "grad_norm": 1.048446043918377, "learning_rate": 9.563541842249903e-06, "loss": 1.0862, "step": 826 }, { "epoch": 1.105685618729097, "grad_norm": 0.9507951672779461, "learning_rate": 9.561633005217264e-06, "loss": 1.1606, "step": 827 }, { "epoch": 1.1070234113712374, "grad_norm": 0.9166166854406071, "learning_rate": 9.559720194514303e-06, "loss": 1.0172, "step": 828 }, { "epoch": 1.108361204013378, "grad_norm": 1.1039122410416018, "learning_rate": 9.557803411807283e-06, "loss": 1.0806, "step": 829 }, { "epoch": 1.1096989966555184, "grad_norm": 0.9790719676293573, "learning_rate": 9.555882658765924e-06, "loss": 1.0611, "step": 830 }, { "epoch": 1.111036789297659, "grad_norm": 0.99834402134704, "learning_rate": 9.55395793706341e-06, "loss": 1.2855, "step": 831 }, { "epoch": 1.1123745819397994, "grad_norm": 0.9691363473449872, "learning_rate": 9.552029248376378e-06, "loss": 1.1683, "step": 832 }, { "epoch": 1.1137123745819397, "grad_norm": 1.1480601757353315, "learning_rate": 9.550096594384923e-06, "loss": 1.2027, "step": 833 }, { "epoch": 1.1150501672240802, "grad_norm": 0.9516731432875061, "learning_rate": 9.548159976772593e-06, "loss": 1.3705, "step": 834 }, { "epoch": 1.1163879598662207, "grad_norm": 1.0437933724391346, "learning_rate": 9.54621939722639e-06, "loss": 1.1246, "step": 835 }, { "epoch": 1.1177257525083613, "grad_norm": 1.1336025582335585, "learning_rate": 9.544274857436763e-06, "loss": 1.1342, "step": 836 }, { "epoch": 1.1190635451505018, "grad_norm": 0.9915791096625699, "learning_rate": 9.542326359097619e-06, "loss": 1.2518, "step": 837 }, { "epoch": 1.120401337792642, "grad_norm": 1.2136708294495961, "learning_rate": 9.540373903906306e-06, "loss": 1.3529, "step": 838 }, { "epoch": 1.1217391304347826, "grad_norm": 0.8643161003686142, "learning_rate": 9.538417493563621e-06, "loss": 0.9916, "step": 839 }, { "epoch": 1.123076923076923, "grad_norm": 0.978325207329163, "learning_rate": 9.536457129773808e-06, "loss": 1.025, "step": 840 }, { "epoch": 1.1244147157190636, "grad_norm": 0.7791506071336896, "learning_rate": 9.534492814244552e-06, "loss": 1.0539, "step": 841 }, { "epoch": 1.125752508361204, "grad_norm": 1.163056394911356, "learning_rate": 9.532524548686984e-06, "loss": 1.1995, "step": 842 }, { "epoch": 1.1270903010033444, "grad_norm": 0.9023114669088667, "learning_rate": 9.530552334815672e-06, "loss": 0.7732, "step": 843 }, { "epoch": 1.128428093645485, "grad_norm": 1.0995210679327945, "learning_rate": 9.528576174348625e-06, "loss": 1.0073, "step": 844 }, { "epoch": 1.1297658862876254, "grad_norm": 0.9068529315538739, "learning_rate": 9.526596069007292e-06, "loss": 1.0384, "step": 845 }, { "epoch": 1.131103678929766, "grad_norm": 1.2370174517917405, "learning_rate": 9.524612020516556e-06, "loss": 1.3982, "step": 846 }, { "epoch": 1.1324414715719064, "grad_norm": 0.8171521910268423, "learning_rate": 9.522624030604735e-06, "loss": 0.863, "step": 847 }, { "epoch": 1.1337792642140467, "grad_norm": 0.9337414557598569, "learning_rate": 9.520632101003579e-06, "loss": 1.3148, "step": 848 }, { "epoch": 1.1351170568561872, "grad_norm": 0.9774501464648273, "learning_rate": 9.518636233448276e-06, "loss": 0.9698, "step": 849 }, { "epoch": 1.1364548494983278, "grad_norm": 1.0513317190496965, "learning_rate": 9.516636429677437e-06, "loss": 1.2311, "step": 850 }, { "epoch": 1.1377926421404683, "grad_norm": 1.0778873427689237, "learning_rate": 9.514632691433108e-06, "loss": 1.1864, "step": 851 }, { "epoch": 1.1391304347826088, "grad_norm": 1.4358666405944671, "learning_rate": 9.512625020460754e-06, "loss": 0.9318, "step": 852 }, { "epoch": 1.140468227424749, "grad_norm": 1.0131334298302108, "learning_rate": 9.510613418509276e-06, "loss": 1.0401, "step": 853 }, { "epoch": 1.1418060200668896, "grad_norm": 1.0823393021415018, "learning_rate": 9.508597887330993e-06, "loss": 1.1483, "step": 854 }, { "epoch": 1.14314381270903, "grad_norm": 1.022287679390198, "learning_rate": 9.506578428681648e-06, "loss": 1.1261, "step": 855 }, { "epoch": 1.1444816053511706, "grad_norm": 1.1770471513587295, "learning_rate": 9.504555044320407e-06, "loss": 1.1826, "step": 856 }, { "epoch": 1.1458193979933111, "grad_norm": 0.8743356976118994, "learning_rate": 9.50252773600985e-06, "loss": 1.1252, "step": 857 }, { "epoch": 1.1471571906354514, "grad_norm": 1.069858265027939, "learning_rate": 9.500496505515986e-06, "loss": 1.0242, "step": 858 }, { "epoch": 1.148494983277592, "grad_norm": 1.1920802650002664, "learning_rate": 9.498461354608228e-06, "loss": 1.0065, "step": 859 }, { "epoch": 1.1498327759197324, "grad_norm": 0.9372165808054905, "learning_rate": 9.496422285059412e-06, "loss": 1.0566, "step": 860 }, { "epoch": 1.151170568561873, "grad_norm": 0.9547693817636563, "learning_rate": 9.494379298645788e-06, "loss": 1.1936, "step": 861 }, { "epoch": 1.1525083612040135, "grad_norm": 1.2869280455065049, "learning_rate": 9.492332397147013e-06, "loss": 1.0637, "step": 862 }, { "epoch": 1.1538461538461537, "grad_norm": 0.9108107684924931, "learning_rate": 9.490281582346159e-06, "loss": 1.1279, "step": 863 }, { "epoch": 1.1551839464882943, "grad_norm": 1.0067192949954282, "learning_rate": 9.488226856029704e-06, "loss": 1.2389, "step": 864 }, { "epoch": 1.1565217391304348, "grad_norm": 0.9137684549837446, "learning_rate": 9.486168219987534e-06, "loss": 1.1251, "step": 865 }, { "epoch": 1.1578595317725753, "grad_norm": 0.9640785296561794, "learning_rate": 9.484105676012943e-06, "loss": 1.0087, "step": 866 }, { "epoch": 1.1591973244147158, "grad_norm": 1.1002147600863734, "learning_rate": 9.482039225902623e-06, "loss": 1.118, "step": 867 }, { "epoch": 1.160535117056856, "grad_norm": 0.9471999192057262, "learning_rate": 9.47996887145668e-06, "loss": 1.2312, "step": 868 }, { "epoch": 1.1618729096989966, "grad_norm": 1.0555806920103628, "learning_rate": 9.47789461447861e-06, "loss": 0.8614, "step": 869 }, { "epoch": 1.163210702341137, "grad_norm": 1.1883870143819468, "learning_rate": 9.475816456775313e-06, "loss": 1.1225, "step": 870 }, { "epoch": 1.1645484949832776, "grad_norm": 1.3153143491632646, "learning_rate": 9.473734400157086e-06, "loss": 0.9837, "step": 871 }, { "epoch": 1.1658862876254181, "grad_norm": 1.2570529765077223, "learning_rate": 9.471648446437625e-06, "loss": 1.1283, "step": 872 }, { "epoch": 1.1672240802675584, "grad_norm": 0.9202958796753631, "learning_rate": 9.469558597434018e-06, "loss": 1.395, "step": 873 }, { "epoch": 1.168561872909699, "grad_norm": 1.1760038908230006, "learning_rate": 9.467464854966746e-06, "loss": 1.2269, "step": 874 }, { "epoch": 1.1698996655518394, "grad_norm": 0.9771821576508082, "learning_rate": 9.465367220859684e-06, "loss": 1.0409, "step": 875 }, { "epoch": 1.17123745819398, "grad_norm": 1.2256168863617884, "learning_rate": 9.463265696940095e-06, "loss": 1.0898, "step": 876 }, { "epoch": 1.1725752508361205, "grad_norm": 1.112078686605078, "learning_rate": 9.461160285038632e-06, "loss": 0.9407, "step": 877 }, { "epoch": 1.1739130434782608, "grad_norm": 0.7614154935082874, "learning_rate": 9.459050986989333e-06, "loss": 1.2148, "step": 878 }, { "epoch": 1.1752508361204013, "grad_norm": 1.2260621223637347, "learning_rate": 9.456937804629623e-06, "loss": 1.4171, "step": 879 }, { "epoch": 1.1765886287625418, "grad_norm": 1.4263586719827277, "learning_rate": 9.45482073980031e-06, "loss": 1.1501, "step": 880 }, { "epoch": 1.1779264214046823, "grad_norm": 1.2030799278981377, "learning_rate": 9.452699794345583e-06, "loss": 1.3214, "step": 881 }, { "epoch": 1.1792642140468228, "grad_norm": 1.1914019409264442, "learning_rate": 9.45057497011301e-06, "loss": 1.1279, "step": 882 }, { "epoch": 1.180602006688963, "grad_norm": 1.188392379501908, "learning_rate": 9.448446268953549e-06, "loss": 0.9096, "step": 883 }, { "epoch": 1.1819397993311036, "grad_norm": 0.9775377863086502, "learning_rate": 9.446313692721515e-06, "loss": 1.079, "step": 884 }, { "epoch": 1.1832775919732441, "grad_norm": 0.9052053919290663, "learning_rate": 9.444177243274619e-06, "loss": 0.9625, "step": 885 }, { "epoch": 1.1846153846153846, "grad_norm": 1.0148689152497823, "learning_rate": 9.442036922473932e-06, "loss": 1.1621, "step": 886 }, { "epoch": 1.1859531772575251, "grad_norm": 0.934308582945314, "learning_rate": 9.439892732183903e-06, "loss": 1.2337, "step": 887 }, { "epoch": 1.1872909698996654, "grad_norm": 1.2428064371104839, "learning_rate": 9.437744674272353e-06, "loss": 1.2238, "step": 888 }, { "epoch": 1.188628762541806, "grad_norm": 0.9566232554125114, "learning_rate": 9.435592750610469e-06, "loss": 1.2328, "step": 889 }, { "epoch": 1.1899665551839465, "grad_norm": 0.9981931465694369, "learning_rate": 9.433436963072807e-06, "loss": 1.144, "step": 890 }, { "epoch": 1.191304347826087, "grad_norm": 1.0851356948214035, "learning_rate": 9.43127731353729e-06, "loss": 1.0835, "step": 891 }, { "epoch": 1.1926421404682275, "grad_norm": 1.0034046261712184, "learning_rate": 9.429113803885199e-06, "loss": 1.2565, "step": 892 }, { "epoch": 1.193979933110368, "grad_norm": 1.0975275103812339, "learning_rate": 9.426946436001189e-06, "loss": 1.0973, "step": 893 }, { "epoch": 1.1953177257525083, "grad_norm": 0.9462916518787362, "learning_rate": 9.424775211773263e-06, "loss": 1.1611, "step": 894 }, { "epoch": 1.1966555183946488, "grad_norm": 1.0800117804154359, "learning_rate": 9.422600133092795e-06, "loss": 1.1473, "step": 895 }, { "epoch": 1.1979933110367893, "grad_norm": 0.98496366761679, "learning_rate": 9.420421201854507e-06, "loss": 1.0126, "step": 896 }, { "epoch": 1.1993311036789298, "grad_norm": 1.2741726827455453, "learning_rate": 9.418238419956484e-06, "loss": 1.2013, "step": 897 }, { "epoch": 1.2006688963210703, "grad_norm": 0.9795881156814314, "learning_rate": 9.41605178930016e-06, "loss": 1.3234, "step": 898 }, { "epoch": 1.2020066889632106, "grad_norm": 1.100340022861075, "learning_rate": 9.413861311790327e-06, "loss": 1.2747, "step": 899 }, { "epoch": 1.2033444816053511, "grad_norm": 1.2729615738164586, "learning_rate": 9.411666989335123e-06, "loss": 1.2689, "step": 900 }, { "epoch": 1.2046822742474916, "grad_norm": 1.1806118469027582, "learning_rate": 9.409468823846038e-06, "loss": 0.9771, "step": 901 }, { "epoch": 1.2060200668896321, "grad_norm": 1.5114012053119223, "learning_rate": 9.40726681723791e-06, "loss": 1.0193, "step": 902 }, { "epoch": 1.2073578595317727, "grad_norm": 1.3789247852985993, "learning_rate": 9.405060971428924e-06, "loss": 1.1007, "step": 903 }, { "epoch": 1.208695652173913, "grad_norm": 0.9861729035977734, "learning_rate": 9.402851288340604e-06, "loss": 0.9183, "step": 904 }, { "epoch": 1.2100334448160535, "grad_norm": 1.1842020860786115, "learning_rate": 9.400637769897823e-06, "loss": 1.1133, "step": 905 }, { "epoch": 1.211371237458194, "grad_norm": 0.9962171392109489, "learning_rate": 9.398420418028789e-06, "loss": 1.0996, "step": 906 }, { "epoch": 1.2127090301003345, "grad_norm": 0.7729906089361527, "learning_rate": 9.396199234665056e-06, "loss": 1.2429, "step": 907 }, { "epoch": 1.214046822742475, "grad_norm": 1.0947150273361879, "learning_rate": 9.393974221741514e-06, "loss": 1.2886, "step": 908 }, { "epoch": 1.2153846153846155, "grad_norm": 1.324610791333157, "learning_rate": 9.391745381196382e-06, "loss": 1.0299, "step": 909 }, { "epoch": 1.2167224080267558, "grad_norm": 1.1072631678617357, "learning_rate": 9.38951271497122e-06, "loss": 1.1941, "step": 910 }, { "epoch": 1.2180602006688963, "grad_norm": 0.8971622035054051, "learning_rate": 9.387276225010925e-06, "loss": 1.3145, "step": 911 }, { "epoch": 1.2193979933110368, "grad_norm": 1.062155382747387, "learning_rate": 9.38503591326371e-06, "loss": 1.1417, "step": 912 }, { "epoch": 1.2207357859531773, "grad_norm": 1.1357774558559108, "learning_rate": 9.382791781681133e-06, "loss": 1.3981, "step": 913 }, { "epoch": 1.2220735785953178, "grad_norm": 1.0293576335892933, "learning_rate": 9.380543832218069e-06, "loss": 1.3537, "step": 914 }, { "epoch": 1.2234113712374581, "grad_norm": 1.083515501364652, "learning_rate": 9.378292066832723e-06, "loss": 1.3363, "step": 915 }, { "epoch": 1.2247491638795986, "grad_norm": 1.0555688894762045, "learning_rate": 9.376036487486626e-06, "loss": 0.8515, "step": 916 }, { "epoch": 1.2260869565217392, "grad_norm": 0.8274493649395899, "learning_rate": 9.373777096144625e-06, "loss": 1.1986, "step": 917 }, { "epoch": 1.2274247491638797, "grad_norm": 1.0800070413821699, "learning_rate": 9.371513894774894e-06, "loss": 1.1482, "step": 918 }, { "epoch": 1.2287625418060202, "grad_norm": 0.9455830697507178, "learning_rate": 9.369246885348926e-06, "loss": 1.0273, "step": 919 }, { "epoch": 1.2301003344481605, "grad_norm": 0.9951278792702288, "learning_rate": 9.366976069841524e-06, "loss": 1.3085, "step": 920 }, { "epoch": 1.231438127090301, "grad_norm": 1.1206109218966005, "learning_rate": 9.364701450230813e-06, "loss": 1.135, "step": 921 }, { "epoch": 1.2327759197324415, "grad_norm": 0.8896594494502921, "learning_rate": 9.362423028498229e-06, "loss": 0.9632, "step": 922 }, { "epoch": 1.234113712374582, "grad_norm": 1.1309491622967742, "learning_rate": 9.360140806628523e-06, "loss": 1.2211, "step": 923 }, { "epoch": 1.2354515050167225, "grad_norm": 1.1871555619071246, "learning_rate": 9.357854786609754e-06, "loss": 0.9845, "step": 924 }, { "epoch": 1.2367892976588628, "grad_norm": 0.9941420477965409, "learning_rate": 9.355564970433288e-06, "loss": 1.1786, "step": 925 }, { "epoch": 1.2381270903010033, "grad_norm": 0.8475853034818221, "learning_rate": 9.353271360093802e-06, "loss": 1.2488, "step": 926 }, { "epoch": 1.2394648829431438, "grad_norm": 0.99569268072867, "learning_rate": 9.350973957589278e-06, "loss": 1.0709, "step": 927 }, { "epoch": 1.2408026755852843, "grad_norm": 1.1530626970543558, "learning_rate": 9.348672764920995e-06, "loss": 1.341, "step": 928 }, { "epoch": 1.2421404682274249, "grad_norm": 0.9284404251569781, "learning_rate": 9.346367784093538e-06, "loss": 1.1175, "step": 929 }, { "epoch": 1.2434782608695651, "grad_norm": 0.78798957487457, "learning_rate": 9.344059017114796e-06, "loss": 0.8881, "step": 930 }, { "epoch": 1.2448160535117057, "grad_norm": 1.0160672517131157, "learning_rate": 9.341746465995947e-06, "loss": 0.9314, "step": 931 }, { "epoch": 1.2461538461538462, "grad_norm": 0.7959803736548032, "learning_rate": 9.339430132751474e-06, "loss": 1.026, "step": 932 }, { "epoch": 1.2474916387959867, "grad_norm": 0.8280717508326519, "learning_rate": 9.33711001939915e-06, "loss": 1.0149, "step": 933 }, { "epoch": 1.2488294314381272, "grad_norm": 1.0212364700884051, "learning_rate": 9.33478612796004e-06, "loss": 1.4217, "step": 934 }, { "epoch": 1.2501672240802675, "grad_norm": 1.0104747608490596, "learning_rate": 9.332458460458507e-06, "loss": 1.0759, "step": 935 }, { "epoch": 1.251505016722408, "grad_norm": 0.8500703005600044, "learning_rate": 9.330127018922195e-06, "loss": 1.0337, "step": 936 }, { "epoch": 1.2528428093645485, "grad_norm": 1.001423667327852, "learning_rate": 9.327791805382038e-06, "loss": 1.2797, "step": 937 }, { "epoch": 1.254180602006689, "grad_norm": 1.1398310144451185, "learning_rate": 9.325452821872258e-06, "loss": 0.9084, "step": 938 }, { "epoch": 1.2555183946488295, "grad_norm": 0.9855284783108956, "learning_rate": 9.32311007043036e-06, "loss": 1.2154, "step": 939 }, { "epoch": 1.2568561872909698, "grad_norm": 1.1227518738432103, "learning_rate": 9.320763553097132e-06, "loss": 1.2445, "step": 940 }, { "epoch": 1.2581939799331103, "grad_norm": 0.9886666664581908, "learning_rate": 9.31841327191664e-06, "loss": 1.2513, "step": 941 }, { "epoch": 1.2595317725752508, "grad_norm": 1.171838470296137, "learning_rate": 9.316059228936231e-06, "loss": 1.4469, "step": 942 }, { "epoch": 1.2608695652173914, "grad_norm": 1.2414176464889668, "learning_rate": 9.31370142620653e-06, "loss": 1.0213, "step": 943 }, { "epoch": 1.2622073578595319, "grad_norm": 1.0838544508775279, "learning_rate": 9.311339865781432e-06, "loss": 1.1637, "step": 944 }, { "epoch": 1.2635451505016722, "grad_norm": 0.9976970711957279, "learning_rate": 9.30897454971811e-06, "loss": 1.2593, "step": 945 }, { "epoch": 1.2648829431438127, "grad_norm": 1.3075639192768365, "learning_rate": 9.30660548007701e-06, "loss": 0.8424, "step": 946 }, { "epoch": 1.2662207357859532, "grad_norm": 0.9878711682420908, "learning_rate": 9.30423265892184e-06, "loss": 1.4686, "step": 947 }, { "epoch": 1.2675585284280937, "grad_norm": 0.9216199429408262, "learning_rate": 9.301856088319584e-06, "loss": 1.0657, "step": 948 }, { "epoch": 1.2688963210702342, "grad_norm": 0.8153456669708955, "learning_rate": 9.299475770340492e-06, "loss": 1.0484, "step": 949 }, { "epoch": 1.2702341137123745, "grad_norm": 1.141826747650476, "learning_rate": 9.297091707058071e-06, "loss": 1.2122, "step": 950 }, { "epoch": 1.271571906354515, "grad_norm": 1.0548231614815127, "learning_rate": 9.294703900549096e-06, "loss": 1.1274, "step": 951 }, { "epoch": 1.2729096989966555, "grad_norm": 0.8025101765181377, "learning_rate": 9.292312352893603e-06, "loss": 1.0601, "step": 952 }, { "epoch": 1.274247491638796, "grad_norm": 0.9454762636482521, "learning_rate": 9.289917066174887e-06, "loss": 1.1113, "step": 953 }, { "epoch": 1.2755852842809365, "grad_norm": 1.1232794327507734, "learning_rate": 9.287518042479495e-06, "loss": 1.1421, "step": 954 }, { "epoch": 1.2769230769230768, "grad_norm": 1.0883587608599672, "learning_rate": 9.285115283897237e-06, "loss": 1.2585, "step": 955 }, { "epoch": 1.2782608695652173, "grad_norm": 1.3167119929302997, "learning_rate": 9.282708792521173e-06, "loss": 1.1979, "step": 956 }, { "epoch": 1.2795986622073579, "grad_norm": 0.9202724975299253, "learning_rate": 9.280298570447612e-06, "loss": 1.1667, "step": 957 }, { "epoch": 1.2809364548494984, "grad_norm": 1.2111843000278935, "learning_rate": 9.277884619776116e-06, "loss": 0.9902, "step": 958 }, { "epoch": 1.2822742474916389, "grad_norm": 1.0505780609535809, "learning_rate": 9.275466942609495e-06, "loss": 1.1328, "step": 959 }, { "epoch": 1.2836120401337792, "grad_norm": 1.2924022138145947, "learning_rate": 9.273045541053805e-06, "loss": 1.1672, "step": 960 }, { "epoch": 1.2849498327759197, "grad_norm": 1.0543003349712545, "learning_rate": 9.270620417218344e-06, "loss": 0.8698, "step": 961 }, { "epoch": 1.2862876254180602, "grad_norm": 1.0506811888006515, "learning_rate": 9.268191573215653e-06, "loss": 1.2288, "step": 962 }, { "epoch": 1.2876254180602007, "grad_norm": 1.1178668565864505, "learning_rate": 9.265759011161519e-06, "loss": 1.1215, "step": 963 }, { "epoch": 1.2889632107023412, "grad_norm": 0.9248306809417991, "learning_rate": 9.263322733174962e-06, "loss": 0.9293, "step": 964 }, { "epoch": 1.2903010033444815, "grad_norm": 0.8612045696007308, "learning_rate": 9.26088274137824e-06, "loss": 0.9889, "step": 965 }, { "epoch": 1.291638795986622, "grad_norm": 1.1578547457717252, "learning_rate": 9.258439037896846e-06, "loss": 1.4979, "step": 966 }, { "epoch": 1.2929765886287625, "grad_norm": 1.1968474363085408, "learning_rate": 9.25599162485951e-06, "loss": 1.148, "step": 967 }, { "epoch": 1.294314381270903, "grad_norm": 0.7123714720625102, "learning_rate": 9.25354050439819e-06, "loss": 1.2305, "step": 968 }, { "epoch": 1.2956521739130435, "grad_norm": 1.1577174629550213, "learning_rate": 9.251085678648072e-06, "loss": 1.218, "step": 969 }, { "epoch": 1.2969899665551838, "grad_norm": 1.128362959302019, "learning_rate": 9.248627149747573e-06, "loss": 1.2517, "step": 970 }, { "epoch": 1.2983277591973243, "grad_norm": 1.0278131271412625, "learning_rate": 9.246164919838334e-06, "loss": 1.0379, "step": 971 }, { "epoch": 1.2996655518394649, "grad_norm": 1.1327739322221286, "learning_rate": 9.243698991065222e-06, "loss": 1.1504, "step": 972 }, { "epoch": 1.3010033444816054, "grad_norm": 0.738034340931029, "learning_rate": 9.241229365576325e-06, "loss": 1.2362, "step": 973 }, { "epoch": 1.3023411371237459, "grad_norm": 1.167318770134274, "learning_rate": 9.238756045522949e-06, "loss": 1.177, "step": 974 }, { "epoch": 1.3036789297658862, "grad_norm": 1.232453095447205, "learning_rate": 9.236279033059622e-06, "loss": 1.1932, "step": 975 }, { "epoch": 1.3050167224080267, "grad_norm": 1.0677236459299584, "learning_rate": 9.233798330344085e-06, "loss": 0.8692, "step": 976 }, { "epoch": 1.3063545150501672, "grad_norm": 1.0025917308610022, "learning_rate": 9.231313939537298e-06, "loss": 1.0749, "step": 977 }, { "epoch": 1.3076923076923077, "grad_norm": 0.9969360913613065, "learning_rate": 9.22882586280343e-06, "loss": 1.1581, "step": 978 }, { "epoch": 1.3090301003344482, "grad_norm": 1.2187971421995694, "learning_rate": 9.226334102309862e-06, "loss": 1.4001, "step": 979 }, { "epoch": 1.3103678929765885, "grad_norm": 0.9412587501353228, "learning_rate": 9.223838660227183e-06, "loss": 0.9952, "step": 980 }, { "epoch": 1.311705685618729, "grad_norm": 0.8341998538992306, "learning_rate": 9.221339538729191e-06, "loss": 1.4597, "step": 981 }, { "epoch": 1.3130434782608695, "grad_norm": 1.2200306685942388, "learning_rate": 9.21883673999289e-06, "loss": 1.3731, "step": 982 }, { "epoch": 1.31438127090301, "grad_norm": 1.2357597756266794, "learning_rate": 9.21633026619848e-06, "loss": 1.0922, "step": 983 }, { "epoch": 1.3157190635451506, "grad_norm": 1.2588843578328757, "learning_rate": 9.213820119529372e-06, "loss": 1.0693, "step": 984 }, { "epoch": 1.3170568561872908, "grad_norm": 0.8330967419393992, "learning_rate": 9.211306302172174e-06, "loss": 1.0741, "step": 985 }, { "epoch": 1.3183946488294314, "grad_norm": 1.3707396001207675, "learning_rate": 9.208788816316684e-06, "loss": 0.8881, "step": 986 }, { "epoch": 1.3197324414715719, "grad_norm": 0.8906130602891312, "learning_rate": 9.206267664155906e-06, "loss": 1.2124, "step": 987 }, { "epoch": 1.3210702341137124, "grad_norm": 1.0554690787203567, "learning_rate": 9.203742847886033e-06, "loss": 1.2271, "step": 988 }, { "epoch": 1.322408026755853, "grad_norm": 1.0158699704854073, "learning_rate": 9.201214369706448e-06, "loss": 1.2224, "step": 989 }, { "epoch": 1.3237458193979932, "grad_norm": 1.065477003802122, "learning_rate": 9.198682231819727e-06, "loss": 1.1764, "step": 990 }, { "epoch": 1.325083612040134, "grad_norm": 1.0285856318625264, "learning_rate": 9.196146436431635e-06, "loss": 1.2031, "step": 991 }, { "epoch": 1.3264214046822742, "grad_norm": 1.1271754354182744, "learning_rate": 9.193606985751117e-06, "loss": 0.9404, "step": 992 }, { "epoch": 1.3277591973244147, "grad_norm": 0.7295778263745322, "learning_rate": 9.191063881990308e-06, "loss": 0.9814, "step": 993 }, { "epoch": 1.3290969899665552, "grad_norm": 1.3202132112096634, "learning_rate": 9.188517127364524e-06, "loss": 1.1647, "step": 994 }, { "epoch": 1.3304347826086955, "grad_norm": 1.277242134478278, "learning_rate": 9.185966724092261e-06, "loss": 1.1454, "step": 995 }, { "epoch": 1.3317725752508363, "grad_norm": 1.3113372615278618, "learning_rate": 9.183412674395193e-06, "loss": 1.135, "step": 996 }, { "epoch": 1.3331103678929765, "grad_norm": 1.0149629932558504, "learning_rate": 9.180854980498168e-06, "loss": 0.9981, "step": 997 }, { "epoch": 1.334448160535117, "grad_norm": 1.0996746517963596, "learning_rate": 9.178293644629214e-06, "loss": 1.0311, "step": 998 }, { "epoch": 1.3357859531772576, "grad_norm": 0.9792728068441171, "learning_rate": 9.17572866901953e-06, "loss": 1.2036, "step": 999 }, { "epoch": 1.3371237458193979, "grad_norm": 1.6252472948257513, "learning_rate": 9.173160055903478e-06, "loss": 0.9096, "step": 1000 }, { "epoch": 1.3384615384615386, "grad_norm": 1.291592072336438, "learning_rate": 9.1705878075186e-06, "loss": 1.0733, "step": 1001 }, { "epoch": 1.3397993311036789, "grad_norm": 0.9802004801324091, "learning_rate": 9.168011926105598e-06, "loss": 1.1831, "step": 1002 }, { "epoch": 1.3411371237458194, "grad_norm": 1.0157672585082693, "learning_rate": 9.165432413908341e-06, "loss": 1.2467, "step": 1003 }, { "epoch": 1.34247491638796, "grad_norm": 0.9307554895133545, "learning_rate": 9.162849273173857e-06, "loss": 1.216, "step": 1004 }, { "epoch": 1.3438127090301004, "grad_norm": 1.141154230353349, "learning_rate": 9.160262506152343e-06, "loss": 1.1639, "step": 1005 }, { "epoch": 1.345150501672241, "grad_norm": 1.2535913734626059, "learning_rate": 9.157672115097145e-06, "loss": 1.1186, "step": 1006 }, { "epoch": 1.3464882943143812, "grad_norm": 0.9377774978092356, "learning_rate": 9.155078102264773e-06, "loss": 1.1629, "step": 1007 }, { "epoch": 1.3478260869565217, "grad_norm": 0.9736944790695241, "learning_rate": 9.15248046991489e-06, "loss": 1.3593, "step": 1008 }, { "epoch": 1.3491638795986622, "grad_norm": 1.040635034987279, "learning_rate": 9.14987922031031e-06, "loss": 1.1777, "step": 1009 }, { "epoch": 1.3505016722408028, "grad_norm": 1.2111650068777156, "learning_rate": 9.147274355717002e-06, "loss": 1.179, "step": 1010 }, { "epoch": 1.3518394648829433, "grad_norm": 1.3677734428156727, "learning_rate": 9.14466587840408e-06, "loss": 1.0616, "step": 1011 }, { "epoch": 1.3531772575250836, "grad_norm": 1.540049973039558, "learning_rate": 9.142053790643806e-06, "loss": 1.2221, "step": 1012 }, { "epoch": 1.354515050167224, "grad_norm": 0.8696767205825195, "learning_rate": 9.13943809471159e-06, "loss": 1.0958, "step": 1013 }, { "epoch": 1.3558528428093646, "grad_norm": 0.9053386457775117, "learning_rate": 9.136818792885981e-06, "loss": 1.0145, "step": 1014 }, { "epoch": 1.357190635451505, "grad_norm": 1.0283800674910595, "learning_rate": 9.134195887448673e-06, "loss": 1.1682, "step": 1015 }, { "epoch": 1.3585284280936456, "grad_norm": 0.8612371090181168, "learning_rate": 9.131569380684497e-06, "loss": 0.8679, "step": 1016 }, { "epoch": 1.359866220735786, "grad_norm": 1.185420571630535, "learning_rate": 9.12893927488142e-06, "loss": 1.3117, "step": 1017 }, { "epoch": 1.3612040133779264, "grad_norm": 0.8232803666017838, "learning_rate": 9.126305572330547e-06, "loss": 1.2087, "step": 1018 }, { "epoch": 1.362541806020067, "grad_norm": 1.0573370670249163, "learning_rate": 9.123668275326113e-06, "loss": 0.8893, "step": 1019 }, { "epoch": 1.3638795986622074, "grad_norm": 0.9751335178232687, "learning_rate": 9.121027386165487e-06, "loss": 1.0187, "step": 1020 }, { "epoch": 1.365217391304348, "grad_norm": 1.148517904517755, "learning_rate": 9.118382907149164e-06, "loss": 1.2998, "step": 1021 }, { "epoch": 1.3665551839464882, "grad_norm": 1.0836590932059602, "learning_rate": 9.115734840580772e-06, "loss": 0.9957, "step": 1022 }, { "epoch": 1.3678929765886287, "grad_norm": 0.8632598155524693, "learning_rate": 9.113083188767057e-06, "loss": 1.1309, "step": 1023 }, { "epoch": 1.3692307692307693, "grad_norm": 1.1359804273346945, "learning_rate": 9.110427954017891e-06, "loss": 1.3467, "step": 1024 }, { "epoch": 1.3705685618729098, "grad_norm": 1.1762041822126832, "learning_rate": 9.107769138646273e-06, "loss": 1.285, "step": 1025 }, { "epoch": 1.3719063545150503, "grad_norm": 1.153304628986612, "learning_rate": 9.105106744968308e-06, "loss": 1.005, "step": 1026 }, { "epoch": 1.3732441471571906, "grad_norm": 1.0767181684536593, "learning_rate": 9.10244077530323e-06, "loss": 1.4105, "step": 1027 }, { "epoch": 1.374581939799331, "grad_norm": 1.2754937092811478, "learning_rate": 9.099771231973382e-06, "loss": 0.9559, "step": 1028 }, { "epoch": 1.3759197324414716, "grad_norm": 1.1214968095160858, "learning_rate": 9.097098117304223e-06, "loss": 1.039, "step": 1029 }, { "epoch": 1.377257525083612, "grad_norm": 1.0148977366100682, "learning_rate": 9.094421433624322e-06, "loss": 0.8895, "step": 1030 }, { "epoch": 1.3785953177257526, "grad_norm": 1.0660147138950848, "learning_rate": 9.09174118326536e-06, "loss": 1.0823, "step": 1031 }, { "epoch": 1.379933110367893, "grad_norm": 0.9389813278680752, "learning_rate": 9.089057368562113e-06, "loss": 0.9956, "step": 1032 }, { "epoch": 1.3812709030100334, "grad_norm": 0.8758042792830794, "learning_rate": 9.086369991852478e-06, "loss": 1.1367, "step": 1033 }, { "epoch": 1.382608695652174, "grad_norm": 0.9912078623981783, "learning_rate": 9.083679055477446e-06, "loss": 1.2896, "step": 1034 }, { "epoch": 1.3839464882943144, "grad_norm": 0.8454896744277268, "learning_rate": 9.08098456178111e-06, "loss": 1.0223, "step": 1035 }, { "epoch": 1.385284280936455, "grad_norm": 0.9401126266709989, "learning_rate": 9.078286513110661e-06, "loss": 1.0534, "step": 1036 }, { "epoch": 1.3866220735785952, "grad_norm": 1.1476211616829115, "learning_rate": 9.07558491181639e-06, "loss": 1.1639, "step": 1037 }, { "epoch": 1.3879598662207357, "grad_norm": 0.9514447200495244, "learning_rate": 9.07287976025168e-06, "loss": 1.1034, "step": 1038 }, { "epoch": 1.3892976588628763, "grad_norm": 1.2803856099475737, "learning_rate": 9.070171060773007e-06, "loss": 1.0645, "step": 1039 }, { "epoch": 1.3906354515050168, "grad_norm": 1.2976298914382969, "learning_rate": 9.067458815739938e-06, "loss": 1.1181, "step": 1040 }, { "epoch": 1.3919732441471573, "grad_norm": 1.1866193220582426, "learning_rate": 9.064743027515127e-06, "loss": 1.2833, "step": 1041 }, { "epoch": 1.3933110367892976, "grad_norm": 1.134563836623694, "learning_rate": 9.062023698464322e-06, "loss": 1.1397, "step": 1042 }, { "epoch": 1.394648829431438, "grad_norm": 1.171680155446973, "learning_rate": 9.059300830956343e-06, "loss": 1.1776, "step": 1043 }, { "epoch": 1.3959866220735786, "grad_norm": 0.8451507032489148, "learning_rate": 9.056574427363102e-06, "loss": 1.1535, "step": 1044 }, { "epoch": 1.397324414715719, "grad_norm": 0.8937297945878082, "learning_rate": 9.053844490059589e-06, "loss": 0.96, "step": 1045 }, { "epoch": 1.3986622073578596, "grad_norm": 1.2631465813372595, "learning_rate": 9.051111021423868e-06, "loss": 1.1292, "step": 1046 }, { "epoch": 1.4, "grad_norm": 1.0918199005803273, "learning_rate": 9.048374023837086e-06, "loss": 1.0206, "step": 1047 }, { "epoch": 1.4013377926421404, "grad_norm": 0.8388489987826043, "learning_rate": 9.045633499683457e-06, "loss": 1.1443, "step": 1048 }, { "epoch": 1.402675585284281, "grad_norm": 1.10124109981722, "learning_rate": 9.042889451350274e-06, "loss": 1.4338, "step": 1049 }, { "epoch": 1.4040133779264214, "grad_norm": 0.9476273053317312, "learning_rate": 9.040141881227897e-06, "loss": 1.1255, "step": 1050 }, { "epoch": 1.405351170568562, "grad_norm": 1.0806049266926796, "learning_rate": 9.03739079170975e-06, "loss": 0.9859, "step": 1051 }, { "epoch": 1.4066889632107022, "grad_norm": 1.1568586469682658, "learning_rate": 9.034636185192329e-06, "loss": 1.0416, "step": 1052 }, { "epoch": 1.4080267558528428, "grad_norm": 1.01549526358625, "learning_rate": 9.03187806407519e-06, "loss": 0.879, "step": 1053 }, { "epoch": 1.4093645484949833, "grad_norm": 0.9094373972659422, "learning_rate": 9.029116430760952e-06, "loss": 1.4453, "step": 1054 }, { "epoch": 1.4107023411371238, "grad_norm": 1.1420934761976616, "learning_rate": 9.026351287655294e-06, "loss": 1.1192, "step": 1055 }, { "epoch": 1.4120401337792643, "grad_norm": 0.920945440375503, "learning_rate": 9.023582637166948e-06, "loss": 1.2273, "step": 1056 }, { "epoch": 1.4133779264214046, "grad_norm": 0.9710314101829718, "learning_rate": 9.020810481707709e-06, "loss": 1.2842, "step": 1057 }, { "epoch": 1.414715719063545, "grad_norm": 1.1689334993941947, "learning_rate": 9.01803482369242e-06, "loss": 1.2135, "step": 1058 }, { "epoch": 1.4160535117056856, "grad_norm": 0.9057096187441303, "learning_rate": 9.015255665538972e-06, "loss": 1.1475, "step": 1059 }, { "epoch": 1.4173913043478261, "grad_norm": 1.1536274840355867, "learning_rate": 9.012473009668314e-06, "loss": 1.2087, "step": 1060 }, { "epoch": 1.4187290969899666, "grad_norm": 1.1144400181879779, "learning_rate": 9.009686858504434e-06, "loss": 1.1501, "step": 1061 }, { "epoch": 1.420066889632107, "grad_norm": 0.8457246625706478, "learning_rate": 9.00689721447437e-06, "loss": 0.8883, "step": 1062 }, { "epoch": 1.4214046822742474, "grad_norm": 0.8189767001732856, "learning_rate": 9.004104080008198e-06, "loss": 1.137, "step": 1063 }, { "epoch": 1.422742474916388, "grad_norm": 1.1026145642008875, "learning_rate": 9.001307457539038e-06, "loss": 0.9439, "step": 1064 }, { "epoch": 1.4240802675585285, "grad_norm": 1.3112835389493291, "learning_rate": 8.998507349503048e-06, "loss": 0.8846, "step": 1065 }, { "epoch": 1.425418060200669, "grad_norm": 1.0585382097859057, "learning_rate": 8.99570375833942e-06, "loss": 1.0604, "step": 1066 }, { "epoch": 1.4267558528428093, "grad_norm": 1.2814187487186803, "learning_rate": 8.992896686490384e-06, "loss": 0.9554, "step": 1067 }, { "epoch": 1.4280936454849498, "grad_norm": 0.8602405144846962, "learning_rate": 8.990086136401199e-06, "loss": 1.2247, "step": 1068 }, { "epoch": 1.4294314381270903, "grad_norm": 1.055841571087125, "learning_rate": 8.987272110520154e-06, "loss": 1.1507, "step": 1069 }, { "epoch": 1.4307692307692308, "grad_norm": 1.0283278588629359, "learning_rate": 8.984454611298565e-06, "loss": 1.1116, "step": 1070 }, { "epoch": 1.4321070234113713, "grad_norm": 0.897806081572082, "learning_rate": 8.981633641190779e-06, "loss": 1.1058, "step": 1071 }, { "epoch": 1.4334448160535116, "grad_norm": 1.0579335899735345, "learning_rate": 8.978809202654161e-06, "loss": 1.1507, "step": 1072 }, { "epoch": 1.434782608695652, "grad_norm": 1.010539387751505, "learning_rate": 8.975981298149099e-06, "loss": 1.1664, "step": 1073 }, { "epoch": 1.4361204013377926, "grad_norm": 1.2801902837866137, "learning_rate": 8.973149930139e-06, "loss": 1.2103, "step": 1074 }, { "epoch": 1.4374581939799331, "grad_norm": 0.953749945828665, "learning_rate": 8.97031510109029e-06, "loss": 1.1638, "step": 1075 }, { "epoch": 1.4387959866220736, "grad_norm": 0.9690532148965925, "learning_rate": 8.967476813472407e-06, "loss": 1.28, "step": 1076 }, { "epoch": 1.440133779264214, "grad_norm": 0.8459889059432347, "learning_rate": 8.964635069757803e-06, "loss": 1.1778, "step": 1077 }, { "epoch": 1.4414715719063544, "grad_norm": 0.9260674596836813, "learning_rate": 8.96178987242194e-06, "loss": 1.0857, "step": 1078 }, { "epoch": 1.442809364548495, "grad_norm": 0.8696258882126179, "learning_rate": 8.958941223943292e-06, "loss": 1.0541, "step": 1079 }, { "epoch": 1.4441471571906355, "grad_norm": 0.9786966895141043, "learning_rate": 8.956089126803333e-06, "loss": 1.0725, "step": 1080 }, { "epoch": 1.445484949832776, "grad_norm": 0.962035271982183, "learning_rate": 8.953233583486548e-06, "loss": 1.3845, "step": 1081 }, { "epoch": 1.4468227424749163, "grad_norm": 0.9751864294242292, "learning_rate": 8.950374596480419e-06, "loss": 1.1981, "step": 1082 }, { "epoch": 1.4481605351170568, "grad_norm": 1.2514760634258262, "learning_rate": 8.94751216827543e-06, "loss": 1.2162, "step": 1083 }, { "epoch": 1.4494983277591973, "grad_norm": 1.1871668065860543, "learning_rate": 8.944646301365061e-06, "loss": 1.1798, "step": 1084 }, { "epoch": 1.4508361204013378, "grad_norm": 0.9593486555197452, "learning_rate": 8.94177699824579e-06, "loss": 1.1185, "step": 1085 }, { "epoch": 1.4521739130434783, "grad_norm": 0.9101268027923279, "learning_rate": 8.938904261417088e-06, "loss": 1.1435, "step": 1086 }, { "epoch": 1.4535117056856186, "grad_norm": 0.9851770345525475, "learning_rate": 8.936028093381414e-06, "loss": 1.087, "step": 1087 }, { "epoch": 1.4548494983277591, "grad_norm": 1.006026928575297, "learning_rate": 8.933148496644218e-06, "loss": 1.2445, "step": 1088 }, { "epoch": 1.4561872909698996, "grad_norm": 1.229364994982501, "learning_rate": 8.930265473713939e-06, "loss": 1.0448, "step": 1089 }, { "epoch": 1.4575250836120401, "grad_norm": 0.9194918776615906, "learning_rate": 8.927379027101994e-06, "loss": 1.1095, "step": 1090 }, { "epoch": 1.4588628762541807, "grad_norm": 1.1474046353688254, "learning_rate": 8.924489159322792e-06, "loss": 1.1941, "step": 1091 }, { "epoch": 1.460200668896321, "grad_norm": 1.1300485922192796, "learning_rate": 8.921595872893714e-06, "loss": 1.1129, "step": 1092 }, { "epoch": 1.4615384615384617, "grad_norm": 1.4743179277898695, "learning_rate": 8.918699170335123e-06, "loss": 0.9632, "step": 1093 }, { "epoch": 1.462876254180602, "grad_norm": 0.8967975224368405, "learning_rate": 8.915799054170357e-06, "loss": 0.9569, "step": 1094 }, { "epoch": 1.4642140468227425, "grad_norm": 1.032363367160883, "learning_rate": 8.912895526925726e-06, "loss": 1.098, "step": 1095 }, { "epoch": 1.465551839464883, "grad_norm": 1.21081468531522, "learning_rate": 8.909988591130514e-06, "loss": 1.1063, "step": 1096 }, { "epoch": 1.4668896321070233, "grad_norm": 0.9092890616332162, "learning_rate": 8.907078249316972e-06, "loss": 1.1477, "step": 1097 }, { "epoch": 1.468227424749164, "grad_norm": 1.2050541103545973, "learning_rate": 8.904164504020321e-06, "loss": 0.9697, "step": 1098 }, { "epoch": 1.4695652173913043, "grad_norm": 0.9723451857282444, "learning_rate": 8.901247357778742e-06, "loss": 1.3754, "step": 1099 }, { "epoch": 1.4709030100334448, "grad_norm": 1.1286827373042139, "learning_rate": 8.898326813133385e-06, "loss": 1.0211, "step": 1100 }, { "epoch": 1.4722408026755853, "grad_norm": 1.1429249301857436, "learning_rate": 8.895402872628352e-06, "loss": 1.3752, "step": 1101 }, { "epoch": 1.4735785953177256, "grad_norm": 0.8776389391399207, "learning_rate": 8.892475538810714e-06, "loss": 0.915, "step": 1102 }, { "epoch": 1.4749163879598663, "grad_norm": 1.4624829624969364, "learning_rate": 8.889544814230487e-06, "loss": 1.1164, "step": 1103 }, { "epoch": 1.4762541806020066, "grad_norm": 1.2710722411531215, "learning_rate": 8.886610701440648e-06, "loss": 1.1418, "step": 1104 }, { "epoch": 1.4775919732441471, "grad_norm": 1.003174632151788, "learning_rate": 8.883673202997121e-06, "loss": 1.2061, "step": 1105 }, { "epoch": 1.4789297658862877, "grad_norm": 1.2170006106575848, "learning_rate": 8.880732321458785e-06, "loss": 1.1745, "step": 1106 }, { "epoch": 1.4802675585284282, "grad_norm": 1.1814123885398127, "learning_rate": 8.87778805938746e-06, "loss": 0.9821, "step": 1107 }, { "epoch": 1.4816053511705687, "grad_norm": 1.1107657349622195, "learning_rate": 8.874840419347912e-06, "loss": 1.2665, "step": 1108 }, { "epoch": 1.482943143812709, "grad_norm": 1.2099832466044216, "learning_rate": 8.871889403907853e-06, "loss": 1.0819, "step": 1109 }, { "epoch": 1.4842809364548495, "grad_norm": 0.9696980761794896, "learning_rate": 8.868935015637932e-06, "loss": 1.2597, "step": 1110 }, { "epoch": 1.48561872909699, "grad_norm": 0.8534736898742808, "learning_rate": 8.865977257111738e-06, "loss": 1.1181, "step": 1111 }, { "epoch": 1.4869565217391305, "grad_norm": 1.0793790250225996, "learning_rate": 8.863016130905795e-06, "loss": 1.4813, "step": 1112 }, { "epoch": 1.488294314381271, "grad_norm": 0.8767633119767375, "learning_rate": 8.86005163959956e-06, "loss": 1.1534, "step": 1113 }, { "epoch": 1.4896321070234113, "grad_norm": 0.8945650435557001, "learning_rate": 8.857083785775423e-06, "loss": 0.8981, "step": 1114 }, { "epoch": 1.4909698996655518, "grad_norm": 1.3540206494765203, "learning_rate": 8.854112572018702e-06, "loss": 1.1694, "step": 1115 }, { "epoch": 1.4923076923076923, "grad_norm": 0.9122027070944845, "learning_rate": 8.851138000917641e-06, "loss": 1.2447, "step": 1116 }, { "epoch": 1.4936454849498328, "grad_norm": 1.1243946415651414, "learning_rate": 8.84816007506341e-06, "loss": 0.9898, "step": 1117 }, { "epoch": 1.4949832775919734, "grad_norm": 0.956083732392081, "learning_rate": 8.845178797050102e-06, "loss": 1.2784, "step": 1118 }, { "epoch": 1.4963210702341136, "grad_norm": 0.9069244858462663, "learning_rate": 8.842194169474727e-06, "loss": 1.1064, "step": 1119 }, { "epoch": 1.4976588628762542, "grad_norm": 0.8070632836901401, "learning_rate": 8.839206194937218e-06, "loss": 0.8523, "step": 1120 }, { "epoch": 1.4989966555183947, "grad_norm": 1.0773805623896053, "learning_rate": 8.836214876040416e-06, "loss": 1.2055, "step": 1121 }, { "epoch": 1.500334448160535, "grad_norm": 0.9459109465639987, "learning_rate": 8.833220215390085e-06, "loss": 0.9727, "step": 1122 }, { "epoch": 1.5016722408026757, "grad_norm": 1.033314698269303, "learning_rate": 8.83022221559489e-06, "loss": 1.1034, "step": 1123 }, { "epoch": 1.503010033444816, "grad_norm": 0.9350455991502298, "learning_rate": 8.827220879266414e-06, "loss": 1.0387, "step": 1124 }, { "epoch": 1.5043478260869565, "grad_norm": 1.1649009050993357, "learning_rate": 8.824216209019139e-06, "loss": 0.8804, "step": 1125 }, { "epoch": 1.505685618729097, "grad_norm": 0.813287029990514, "learning_rate": 8.821208207470454e-06, "loss": 1.2782, "step": 1126 }, { "epoch": 1.5070234113712373, "grad_norm": 0.8699091132179207, "learning_rate": 8.818196877240652e-06, "loss": 1.0785, "step": 1127 }, { "epoch": 1.508361204013378, "grad_norm": 1.0764192705507325, "learning_rate": 8.815182220952922e-06, "loss": 1.2375, "step": 1128 }, { "epoch": 1.5096989966555183, "grad_norm": 0.9520320462130694, "learning_rate": 8.812164241233354e-06, "loss": 0.9402, "step": 1129 }, { "epoch": 1.5110367892976588, "grad_norm": 1.0260389546095927, "learning_rate": 8.80914294071093e-06, "loss": 1.1972, "step": 1130 }, { "epoch": 1.5123745819397993, "grad_norm": 1.0003957419879672, "learning_rate": 8.806118322017525e-06, "loss": 1.1994, "step": 1131 }, { "epoch": 1.5137123745819396, "grad_norm": 1.356305526525121, "learning_rate": 8.803090387787909e-06, "loss": 1.0372, "step": 1132 }, { "epoch": 1.5150501672240804, "grad_norm": 0.7900155667802884, "learning_rate": 8.800059140659731e-06, "loss": 1.1371, "step": 1133 }, { "epoch": 1.5163879598662207, "grad_norm": 1.2933607036346915, "learning_rate": 8.797024583273536e-06, "loss": 1.0781, "step": 1134 }, { "epoch": 1.5177257525083612, "grad_norm": 1.0458002724865052, "learning_rate": 8.793986718272747e-06, "loss": 1.2656, "step": 1135 }, { "epoch": 1.5190635451505017, "grad_norm": 1.2089468714187932, "learning_rate": 8.790945548303669e-06, "loss": 1.0676, "step": 1136 }, { "epoch": 1.5204013377926422, "grad_norm": 0.7584589559032697, "learning_rate": 8.787901076015487e-06, "loss": 1.2696, "step": 1137 }, { "epoch": 1.5217391304347827, "grad_norm": 0.9098971952133185, "learning_rate": 8.784853304060262e-06, "loss": 0.9967, "step": 1138 }, { "epoch": 1.523076923076923, "grad_norm": 0.7831978936416883, "learning_rate": 8.781802235092927e-06, "loss": 1.0385, "step": 1139 }, { "epoch": 1.5244147157190635, "grad_norm": 1.1718692317545067, "learning_rate": 8.778747871771293e-06, "loss": 0.9845, "step": 1140 }, { "epoch": 1.525752508361204, "grad_norm": 0.8700806586295358, "learning_rate": 8.775690216756035e-06, "loss": 0.9047, "step": 1141 }, { "epoch": 1.5270903010033445, "grad_norm": 0.8731930556791816, "learning_rate": 8.772629272710698e-06, "loss": 0.9775, "step": 1142 }, { "epoch": 1.528428093645485, "grad_norm": 1.0331360510977525, "learning_rate": 8.769565042301692e-06, "loss": 0.9859, "step": 1143 }, { "epoch": 1.5297658862876253, "grad_norm": 1.4281375751484708, "learning_rate": 8.766497528198289e-06, "loss": 1.3005, "step": 1144 }, { "epoch": 1.5311036789297658, "grad_norm": 0.8534915814871332, "learning_rate": 8.763426733072624e-06, "loss": 1.3438, "step": 1145 }, { "epoch": 1.5324414715719064, "grad_norm": 1.3740638602212059, "learning_rate": 8.760352659599684e-06, "loss": 1.2747, "step": 1146 }, { "epoch": 1.5337792642140469, "grad_norm": 0.946974671249906, "learning_rate": 8.757275310457321e-06, "loss": 1.1641, "step": 1147 }, { "epoch": 1.5351170568561874, "grad_norm": 0.8460737010634026, "learning_rate": 8.754194688326229e-06, "loss": 1.1374, "step": 1148 }, { "epoch": 1.5364548494983277, "grad_norm": 1.1209454500111449, "learning_rate": 8.751110795889966e-06, "loss": 1.0966, "step": 1149 }, { "epoch": 1.5377926421404682, "grad_norm": 1.0351398121148045, "learning_rate": 8.748023635834927e-06, "loss": 1.2918, "step": 1150 }, { "epoch": 1.5391304347826087, "grad_norm": 1.1466329919159575, "learning_rate": 8.744933210850363e-06, "loss": 1.157, "step": 1151 }, { "epoch": 1.5404682274247492, "grad_norm": 1.1053018652880033, "learning_rate": 8.741839523628361e-06, "loss": 0.9831, "step": 1152 }, { "epoch": 1.5418060200668897, "grad_norm": 1.0507973430293625, "learning_rate": 8.738742576863855e-06, "loss": 1.1507, "step": 1153 }, { "epoch": 1.54314381270903, "grad_norm": 1.0546913844769603, "learning_rate": 8.735642373254617e-06, "loss": 0.9618, "step": 1154 }, { "epoch": 1.5444816053511705, "grad_norm": 1.0490704186815605, "learning_rate": 8.732538915501257e-06, "loss": 1.1712, "step": 1155 }, { "epoch": 1.545819397993311, "grad_norm": 1.0607148768593042, "learning_rate": 8.729432206307218e-06, "loss": 1.2971, "step": 1156 }, { "epoch": 1.5471571906354515, "grad_norm": 0.8246637143510166, "learning_rate": 8.726322248378775e-06, "loss": 1.2606, "step": 1157 }, { "epoch": 1.548494983277592, "grad_norm": 0.9409959068074953, "learning_rate": 8.723209044425034e-06, "loss": 0.959, "step": 1158 }, { "epoch": 1.5498327759197323, "grad_norm": 0.954098032469285, "learning_rate": 8.72009259715793e-06, "loss": 0.9239, "step": 1159 }, { "epoch": 1.551170568561873, "grad_norm": 0.8597422299333858, "learning_rate": 8.71697290929222e-06, "loss": 1.1544, "step": 1160 }, { "epoch": 1.5525083612040134, "grad_norm": 0.8694344319564208, "learning_rate": 8.71384998354549e-06, "loss": 1.3259, "step": 1161 }, { "epoch": 1.5538461538461539, "grad_norm": 0.9519830955997358, "learning_rate": 8.710723822638138e-06, "loss": 0.9604, "step": 1162 }, { "epoch": 1.5551839464882944, "grad_norm": 0.9582922980272383, "learning_rate": 8.707594429293387e-06, "loss": 1.1606, "step": 1163 }, { "epoch": 1.5565217391304347, "grad_norm": 1.316864570850179, "learning_rate": 8.704461806237272e-06, "loss": 1.2251, "step": 1164 }, { "epoch": 1.5578595317725754, "grad_norm": 0.8924699471263677, "learning_rate": 8.701325956198643e-06, "loss": 1.1456, "step": 1165 }, { "epoch": 1.5591973244147157, "grad_norm": 1.1332051614987897, "learning_rate": 8.69818688190916e-06, "loss": 1.1023, "step": 1166 }, { "epoch": 1.5605351170568562, "grad_norm": 1.0359154108166875, "learning_rate": 8.695044586103297e-06, "loss": 0.9923, "step": 1167 }, { "epoch": 1.5618729096989967, "grad_norm": 1.1514827811231336, "learning_rate": 8.691899071518323e-06, "loss": 1.0823, "step": 1168 }, { "epoch": 1.563210702341137, "grad_norm": 1.1817078740795017, "learning_rate": 8.688750340894324e-06, "loss": 1.3059, "step": 1169 }, { "epoch": 1.5645484949832777, "grad_norm": 0.8681057467397233, "learning_rate": 8.685598396974178e-06, "loss": 0.8943, "step": 1170 }, { "epoch": 1.565886287625418, "grad_norm": 0.8576458308650835, "learning_rate": 8.682443242503564e-06, "loss": 1.208, "step": 1171 }, { "epoch": 1.5672240802675586, "grad_norm": 1.1725762977989187, "learning_rate": 8.679284880230963e-06, "loss": 1.0411, "step": 1172 }, { "epoch": 1.568561872909699, "grad_norm": 1.2431135737574128, "learning_rate": 8.676123312907641e-06, "loss": 1.0882, "step": 1173 }, { "epoch": 1.5698996655518394, "grad_norm": 1.0270956981841943, "learning_rate": 8.672958543287666e-06, "loss": 1.0164, "step": 1174 }, { "epoch": 1.57123745819398, "grad_norm": 0.7819241195537215, "learning_rate": 8.66979057412789e-06, "loss": 1.0167, "step": 1175 }, { "epoch": 1.5725752508361204, "grad_norm": 1.1612840323737177, "learning_rate": 8.666619408187953e-06, "loss": 0.9558, "step": 1176 }, { "epoch": 1.5739130434782609, "grad_norm": 1.1336598900627357, "learning_rate": 8.663445048230278e-06, "loss": 1.1997, "step": 1177 }, { "epoch": 1.5752508361204014, "grad_norm": 0.9867194896402057, "learning_rate": 8.660267497020074e-06, "loss": 1.1716, "step": 1178 }, { "epoch": 1.5765886287625417, "grad_norm": 0.9804698503760182, "learning_rate": 8.657086757325328e-06, "loss": 1.1924, "step": 1179 }, { "epoch": 1.5779264214046824, "grad_norm": 1.0654139026958176, "learning_rate": 8.653902831916803e-06, "loss": 0.9838, "step": 1180 }, { "epoch": 1.5792642140468227, "grad_norm": 0.8690231482189016, "learning_rate": 8.650715723568039e-06, "loss": 0.9488, "step": 1181 }, { "epoch": 1.5806020066889632, "grad_norm": 0.7339262764427701, "learning_rate": 8.64752543505535e-06, "loss": 1.0587, "step": 1182 }, { "epoch": 1.5819397993311037, "grad_norm": 1.1254407318589923, "learning_rate": 8.644331969157815e-06, "loss": 1.1791, "step": 1183 }, { "epoch": 1.583277591973244, "grad_norm": 0.7812456344178278, "learning_rate": 8.641135328657288e-06, "loss": 1.0318, "step": 1184 }, { "epoch": 1.5846153846153848, "grad_norm": 1.1223025552418775, "learning_rate": 8.637935516338384e-06, "loss": 1.2596, "step": 1185 }, { "epoch": 1.585953177257525, "grad_norm": 0.9618692829939213, "learning_rate": 8.63473253498848e-06, "loss": 0.959, "step": 1186 }, { "epoch": 1.5872909698996656, "grad_norm": 1.0795073892720355, "learning_rate": 8.63152638739772e-06, "loss": 1.1304, "step": 1187 }, { "epoch": 1.588628762541806, "grad_norm": 1.2544236736883991, "learning_rate": 8.628317076358997e-06, "loss": 0.8854, "step": 1188 }, { "epoch": 1.5899665551839464, "grad_norm": 0.9797092510711526, "learning_rate": 8.625104604667965e-06, "loss": 1.0831, "step": 1189 }, { "epoch": 1.591304347826087, "grad_norm": 1.139775279315503, "learning_rate": 8.62188897512303e-06, "loss": 1.0727, "step": 1190 }, { "epoch": 1.5926421404682274, "grad_norm": 0.9861894422535459, "learning_rate": 8.61867019052535e-06, "loss": 1.0482, "step": 1191 }, { "epoch": 1.593979933110368, "grad_norm": 1.1206371682099745, "learning_rate": 8.615448253678834e-06, "loss": 1.1137, "step": 1192 }, { "epoch": 1.5953177257525084, "grad_norm": 1.0156871622099137, "learning_rate": 8.61222316739013e-06, "loss": 1.0722, "step": 1193 }, { "epoch": 1.5966555183946487, "grad_norm": 1.116664402678087, "learning_rate": 8.608994934468633e-06, "loss": 1.0547, "step": 1194 }, { "epoch": 1.5979933110367894, "grad_norm": 0.7407740236872212, "learning_rate": 8.60576355772648e-06, "loss": 1.1413, "step": 1195 }, { "epoch": 1.5993311036789297, "grad_norm": 0.8352871592976983, "learning_rate": 8.602529039978546e-06, "loss": 0.9358, "step": 1196 }, { "epoch": 1.6006688963210702, "grad_norm": 0.9954165511209426, "learning_rate": 8.599291384042442e-06, "loss": 1.0077, "step": 1197 }, { "epoch": 1.6020066889632107, "grad_norm": 1.072161100016342, "learning_rate": 8.596050592738514e-06, "loss": 1.0988, "step": 1198 }, { "epoch": 1.603344481605351, "grad_norm": 1.1145711656355572, "learning_rate": 8.592806668889835e-06, "loss": 0.9907, "step": 1199 }, { "epoch": 1.6046822742474918, "grad_norm": 1.073877327909215, "learning_rate": 8.58955961532221e-06, "loss": 1.0066, "step": 1200 }, { "epoch": 1.606020066889632, "grad_norm": 1.0562716103800187, "learning_rate": 8.586309434864173e-06, "loss": 1.1913, "step": 1201 }, { "epoch": 1.6073578595317726, "grad_norm": 0.7678773824789086, "learning_rate": 8.583056130346977e-06, "loss": 0.8551, "step": 1202 }, { "epoch": 1.608695652173913, "grad_norm": 1.2053346056943466, "learning_rate": 8.579799704604597e-06, "loss": 0.7789, "step": 1203 }, { "epoch": 1.6100334448160534, "grad_norm": 1.1557197271509942, "learning_rate": 8.57654016047373e-06, "loss": 1.1855, "step": 1204 }, { "epoch": 1.611371237458194, "grad_norm": 1.346048237718495, "learning_rate": 8.573277500793788e-06, "loss": 1.3565, "step": 1205 }, { "epoch": 1.6127090301003344, "grad_norm": 1.0100758101513907, "learning_rate": 8.570011728406895e-06, "loss": 1.2186, "step": 1206 }, { "epoch": 1.614046822742475, "grad_norm": 0.9638446465115132, "learning_rate": 8.56674284615789e-06, "loss": 1.0827, "step": 1207 }, { "epoch": 1.6153846153846154, "grad_norm": 0.9445345008263832, "learning_rate": 8.563470856894316e-06, "loss": 1.0354, "step": 1208 }, { "epoch": 1.6167224080267557, "grad_norm": 1.0126399746426198, "learning_rate": 8.560195763466428e-06, "loss": 0.9891, "step": 1209 }, { "epoch": 1.6180602006688964, "grad_norm": 0.9979224936562032, "learning_rate": 8.556917568727182e-06, "loss": 0.953, "step": 1210 }, { "epoch": 1.6193979933110367, "grad_norm": 1.4031645596963058, "learning_rate": 8.553636275532236e-06, "loss": 1.2496, "step": 1211 }, { "epoch": 1.6207357859531772, "grad_norm": 1.087371240863874, "learning_rate": 8.550351886739949e-06, "loss": 1.2095, "step": 1212 }, { "epoch": 1.6220735785953178, "grad_norm": 0.8451677722349469, "learning_rate": 8.547064405211376e-06, "loss": 1.155, "step": 1213 }, { "epoch": 1.623411371237458, "grad_norm": 1.2052785306092122, "learning_rate": 8.54377383381026e-06, "loss": 1.1309, "step": 1214 }, { "epoch": 1.6247491638795988, "grad_norm": 1.0369212796999792, "learning_rate": 8.540480175403045e-06, "loss": 1.0946, "step": 1215 }, { "epoch": 1.626086956521739, "grad_norm": 1.0923716258491323, "learning_rate": 8.53718343285886e-06, "loss": 0.9171, "step": 1216 }, { "epoch": 1.6274247491638796, "grad_norm": 0.9272265904567695, "learning_rate": 8.533883609049517e-06, "loss": 1.0643, "step": 1217 }, { "epoch": 1.62876254180602, "grad_norm": 1.0248005753248346, "learning_rate": 8.530580706849518e-06, "loss": 1.2792, "step": 1218 }, { "epoch": 1.6301003344481604, "grad_norm": 1.24597929366335, "learning_rate": 8.527274729136042e-06, "loss": 1.0031, "step": 1219 }, { "epoch": 1.6314381270903011, "grad_norm": 0.988631501728731, "learning_rate": 8.523965678788952e-06, "loss": 1.2688, "step": 1220 }, { "epoch": 1.6327759197324414, "grad_norm": 0.9281924502881461, "learning_rate": 8.520653558690785e-06, "loss": 1.2002, "step": 1221 }, { "epoch": 1.634113712374582, "grad_norm": 1.1978131472418236, "learning_rate": 8.51733837172675e-06, "loss": 1.1197, "step": 1222 }, { "epoch": 1.6354515050167224, "grad_norm": 0.702460307680178, "learning_rate": 8.51402012078473e-06, "loss": 1.04, "step": 1223 }, { "epoch": 1.6367892976588627, "grad_norm": 1.1260513569098463, "learning_rate": 8.510698808755275e-06, "loss": 1.2528, "step": 1224 }, { "epoch": 1.6381270903010035, "grad_norm": 0.850262455107949, "learning_rate": 8.507374438531606e-06, "loss": 0.9783, "step": 1225 }, { "epoch": 1.6394648829431437, "grad_norm": 0.8829586553620289, "learning_rate": 8.504047013009605e-06, "loss": 1.1231, "step": 1226 }, { "epoch": 1.6408026755852843, "grad_norm": 0.973146017016233, "learning_rate": 8.500716535087815e-06, "loss": 0.9897, "step": 1227 }, { "epoch": 1.6421404682274248, "grad_norm": 0.8523347772123873, "learning_rate": 8.497383007667435e-06, "loss": 1.272, "step": 1228 }, { "epoch": 1.643478260869565, "grad_norm": 1.083919315216665, "learning_rate": 8.494046433652327e-06, "loss": 1.107, "step": 1229 }, { "epoch": 1.6448160535117058, "grad_norm": 1.1104852019894857, "learning_rate": 8.490706815949006e-06, "loss": 1.2359, "step": 1230 }, { "epoch": 1.646153846153846, "grad_norm": 0.8011220320312396, "learning_rate": 8.487364157466633e-06, "loss": 1.0471, "step": 1231 }, { "epoch": 1.6474916387959866, "grad_norm": 1.054373302574096, "learning_rate": 8.484018461117023e-06, "loss": 1.2019, "step": 1232 }, { "epoch": 1.648829431438127, "grad_norm": 1.0531053551941703, "learning_rate": 8.480669729814635e-06, "loss": 1.179, "step": 1233 }, { "epoch": 1.6501672240802676, "grad_norm": 0.904788788934545, "learning_rate": 8.477317966476569e-06, "loss": 1.3204, "step": 1234 }, { "epoch": 1.6515050167224081, "grad_norm": 0.8694695821752231, "learning_rate": 8.473963174022574e-06, "loss": 1.2846, "step": 1235 }, { "epoch": 1.6528428093645484, "grad_norm": 1.0567208687486753, "learning_rate": 8.470605355375033e-06, "loss": 1.2265, "step": 1236 }, { "epoch": 1.654180602006689, "grad_norm": 0.819752282025206, "learning_rate": 8.467244513458961e-06, "loss": 1.117, "step": 1237 }, { "epoch": 1.6555183946488294, "grad_norm": 0.8766336661526526, "learning_rate": 8.463880651202014e-06, "loss": 1.1757, "step": 1238 }, { "epoch": 1.65685618729097, "grad_norm": 1.1844990076124804, "learning_rate": 8.460513771534475e-06, "loss": 1.1032, "step": 1239 }, { "epoch": 1.6581939799331105, "grad_norm": 0.9553892306756859, "learning_rate": 8.457143877389258e-06, "loss": 1.0345, "step": 1240 }, { "epoch": 1.6595317725752508, "grad_norm": 0.8033900690443401, "learning_rate": 8.453770971701899e-06, "loss": 1.0858, "step": 1241 }, { "epoch": 1.6608695652173913, "grad_norm": 0.884425297039905, "learning_rate": 8.450395057410561e-06, "loss": 1.1123, "step": 1242 }, { "epoch": 1.6622073578595318, "grad_norm": 1.3047257901751035, "learning_rate": 8.447016137456025e-06, "loss": 1.1191, "step": 1243 }, { "epoch": 1.6635451505016723, "grad_norm": 0.7991726831678995, "learning_rate": 8.443634214781693e-06, "loss": 1.1673, "step": 1244 }, { "epoch": 1.6648829431438128, "grad_norm": 1.0022954868293463, "learning_rate": 8.440249292333583e-06, "loss": 0.9888, "step": 1245 }, { "epoch": 1.666220735785953, "grad_norm": 0.6247344328233743, "learning_rate": 8.43686137306032e-06, "loss": 1.3064, "step": 1246 }, { "epoch": 1.6675585284280936, "grad_norm": 1.2012101657869967, "learning_rate": 8.43347045991315e-06, "loss": 1.1051, "step": 1247 }, { "epoch": 1.6688963210702341, "grad_norm": 1.0346450485954488, "learning_rate": 8.430076555845917e-06, "loss": 1.1058, "step": 1248 }, { "epoch": 1.6702341137123746, "grad_norm": 1.0637037784826295, "learning_rate": 8.426679663815073e-06, "loss": 1.151, "step": 1249 }, { "epoch": 1.6715719063545151, "grad_norm": 0.9180727194136292, "learning_rate": 8.42327978677968e-06, "loss": 1.0432, "step": 1250 }, { "epoch": 1.6729096989966554, "grad_norm": 0.9547444307143926, "learning_rate": 8.41987692770139e-06, "loss": 1.0897, "step": 1251 }, { "epoch": 1.674247491638796, "grad_norm": 0.9355505450196803, "learning_rate": 8.41647108954446e-06, "loss": 1.2009, "step": 1252 }, { "epoch": 1.6755852842809364, "grad_norm": 0.8471222262768042, "learning_rate": 8.413062275275737e-06, "loss": 1.0515, "step": 1253 }, { "epoch": 1.676923076923077, "grad_norm": 0.95485006119562, "learning_rate": 8.409650487864662e-06, "loss": 1.1211, "step": 1254 }, { "epoch": 1.6782608695652175, "grad_norm": 1.0809434992660047, "learning_rate": 8.40623573028327e-06, "loss": 1.1016, "step": 1255 }, { "epoch": 1.6795986622073578, "grad_norm": 1.0113337586718916, "learning_rate": 8.402818005506181e-06, "loss": 1.1307, "step": 1256 }, { "epoch": 1.6809364548494983, "grad_norm": 0.9345817740192036, "learning_rate": 8.399397316510596e-06, "loss": 1.0443, "step": 1257 }, { "epoch": 1.6822742474916388, "grad_norm": 1.1228934940977542, "learning_rate": 8.395973666276301e-06, "loss": 0.9326, "step": 1258 }, { "epoch": 1.6836120401337793, "grad_norm": 1.1280254365541247, "learning_rate": 8.392547057785662e-06, "loss": 0.7853, "step": 1259 }, { "epoch": 1.6849498327759198, "grad_norm": 0.9532106175776842, "learning_rate": 8.389117494023622e-06, "loss": 1.1768, "step": 1260 }, { "epoch": 1.68628762541806, "grad_norm": 1.2336978948722364, "learning_rate": 8.385684977977698e-06, "loss": 1.0436, "step": 1261 }, { "epoch": 1.6876254180602008, "grad_norm": 1.1672465833026704, "learning_rate": 8.382249512637978e-06, "loss": 1.1496, "step": 1262 }, { "epoch": 1.6889632107023411, "grad_norm": 1.12803962223924, "learning_rate": 8.378811100997122e-06, "loss": 1.2643, "step": 1263 }, { "epoch": 1.6903010033444816, "grad_norm": 0.9347004754768953, "learning_rate": 8.375369746050353e-06, "loss": 1.901, "step": 1264 }, { "epoch": 1.6916387959866221, "grad_norm": 12.485348359139412, "learning_rate": 8.371925450795458e-06, "loss": 0.9912, "step": 1265 }, { "epoch": 1.6929765886287624, "grad_norm": 1.011625623647366, "learning_rate": 8.368478218232787e-06, "loss": 0.9341, "step": 1266 }, { "epoch": 1.6943143812709032, "grad_norm": 0.958094933784197, "learning_rate": 8.365028051365249e-06, "loss": 1.0204, "step": 1267 }, { "epoch": 1.6956521739130435, "grad_norm": 1.2369784619817465, "learning_rate": 8.361574953198309e-06, "loss": 1.0851, "step": 1268 }, { "epoch": 1.696989966555184, "grad_norm": 1.0807854227766303, "learning_rate": 8.358118926739984e-06, "loss": 0.9979, "step": 1269 }, { "epoch": 1.6983277591973245, "grad_norm": 1.2628971121318586, "learning_rate": 8.354659975000843e-06, "loss": 1.0301, "step": 1270 }, { "epoch": 1.6996655518394648, "grad_norm": 1.0837980701675767, "learning_rate": 8.351198100994003e-06, "loss": 1.0063, "step": 1271 }, { "epoch": 1.7010033444816055, "grad_norm": 1.4234022852675794, "learning_rate": 8.347733307735127e-06, "loss": 1.0085, "step": 1272 }, { "epoch": 1.7023411371237458, "grad_norm": 1.3485663726389747, "learning_rate": 8.34426559824242e-06, "loss": 0.9618, "step": 1273 }, { "epoch": 1.7036789297658863, "grad_norm": 0.7229804116006995, "learning_rate": 8.340794975536627e-06, "loss": 0.9438, "step": 1274 }, { "epoch": 1.7050167224080268, "grad_norm": 0.9198042862053263, "learning_rate": 8.337321442641036e-06, "loss": 1.2629, "step": 1275 }, { "epoch": 1.706354515050167, "grad_norm": 1.0852725630312077, "learning_rate": 8.33384500258146e-06, "loss": 1.0342, "step": 1276 }, { "epoch": 1.7076923076923078, "grad_norm": 1.0435920750908407, "learning_rate": 8.330365658386252e-06, "loss": 1.0698, "step": 1277 }, { "epoch": 1.7090301003344481, "grad_norm": 1.0183688669189477, "learning_rate": 8.326883413086295e-06, "loss": 1.1028, "step": 1278 }, { "epoch": 1.7103678929765886, "grad_norm": 1.0744281953718309, "learning_rate": 8.323398269714994e-06, "loss": 1.1599, "step": 1279 }, { "epoch": 1.7117056856187292, "grad_norm": 1.1283493829538693, "learning_rate": 8.319910231308285e-06, "loss": 0.9858, "step": 1280 }, { "epoch": 1.7130434782608694, "grad_norm": 0.790983156134335, "learning_rate": 8.316419300904622e-06, "loss": 1.3884, "step": 1281 }, { "epoch": 1.7143812709030102, "grad_norm": 1.2645696158125823, "learning_rate": 8.312925481544976e-06, "loss": 1.0545, "step": 1282 }, { "epoch": 1.7157190635451505, "grad_norm": 0.9058635492852943, "learning_rate": 8.309428776272838e-06, "loss": 0.8911, "step": 1283 }, { "epoch": 1.717056856187291, "grad_norm": 0.9548665132692693, "learning_rate": 8.305929188134216e-06, "loss": 1.0141, "step": 1284 }, { "epoch": 1.7183946488294315, "grad_norm": 0.8591424822209708, "learning_rate": 8.302426720177624e-06, "loss": 1.0826, "step": 1285 }, { "epoch": 1.7197324414715718, "grad_norm": 1.3630508256148148, "learning_rate": 8.298921375454083e-06, "loss": 1.295, "step": 1286 }, { "epoch": 1.7210702341137125, "grad_norm": 0.8730993636321397, "learning_rate": 8.295413157017127e-06, "loss": 0.9973, "step": 1287 }, { "epoch": 1.7224080267558528, "grad_norm": 0.8725264741544888, "learning_rate": 8.291902067922791e-06, "loss": 0.9554, "step": 1288 }, { "epoch": 1.7237458193979933, "grad_norm": 1.1644959293943713, "learning_rate": 8.288388111229601e-06, "loss": 1.1459, "step": 1289 }, { "epoch": 1.7250836120401338, "grad_norm": 0.9838887495788354, "learning_rate": 8.284871289998599e-06, "loss": 1.1131, "step": 1290 }, { "epoch": 1.7264214046822741, "grad_norm": 1.1256157363316397, "learning_rate": 8.281351607293307e-06, "loss": 1.2088, "step": 1291 }, { "epoch": 1.7277591973244149, "grad_norm": 1.1993316613434746, "learning_rate": 8.277829066179746e-06, "loss": 1.1903, "step": 1292 }, { "epoch": 1.7290969899665551, "grad_norm": 1.1663839928937763, "learning_rate": 8.274303669726427e-06, "loss": 1.2299, "step": 1293 }, { "epoch": 1.7304347826086957, "grad_norm": 1.1125154968942774, "learning_rate": 8.270775421004345e-06, "loss": 1.0332, "step": 1294 }, { "epoch": 1.7317725752508362, "grad_norm": 1.1343874931035824, "learning_rate": 8.267244323086985e-06, "loss": 1.1467, "step": 1295 }, { "epoch": 1.7331103678929765, "grad_norm": 0.9988975180436451, "learning_rate": 8.263710379050311e-06, "loss": 1.3127, "step": 1296 }, { "epoch": 1.7344481605351172, "grad_norm": 1.2234473698320223, "learning_rate": 8.260173591972765e-06, "loss": 1.0297, "step": 1297 }, { "epoch": 1.7357859531772575, "grad_norm": 0.9795840419937375, "learning_rate": 8.256633964935268e-06, "loss": 1.2612, "step": 1298 }, { "epoch": 1.737123745819398, "grad_norm": 1.1705942388468749, "learning_rate": 8.25309150102121e-06, "loss": 1.1342, "step": 1299 }, { "epoch": 1.7384615384615385, "grad_norm": 0.9515154908877435, "learning_rate": 8.249546203316461e-06, "loss": 1.2835, "step": 1300 }, { "epoch": 1.7397993311036788, "grad_norm": 1.3584428595084501, "learning_rate": 8.245998074909354e-06, "loss": 1.2228, "step": 1301 }, { "epoch": 1.7411371237458195, "grad_norm": 0.8483262222536273, "learning_rate": 8.242447118890686e-06, "loss": 1.1637, "step": 1302 }, { "epoch": 1.7424749163879598, "grad_norm": 0.9459336538049941, "learning_rate": 8.23889333835372e-06, "loss": 0.9896, "step": 1303 }, { "epoch": 1.7438127090301003, "grad_norm": 0.9693844169042332, "learning_rate": 8.235336736394179e-06, "loss": 1.3315, "step": 1304 }, { "epoch": 1.7451505016722408, "grad_norm": 1.025229503388539, "learning_rate": 8.231777316110245e-06, "loss": 0.9206, "step": 1305 }, { "epoch": 1.7464882943143811, "grad_norm": 0.9702694598417486, "learning_rate": 8.228215080602554e-06, "loss": 1.0746, "step": 1306 }, { "epoch": 1.7478260869565219, "grad_norm": 1.069126590711673, "learning_rate": 8.22465003297419e-06, "loss": 1.0233, "step": 1307 }, { "epoch": 1.7491638795986622, "grad_norm": 0.8847732413145254, "learning_rate": 8.221082176330697e-06, "loss": 1.2228, "step": 1308 }, { "epoch": 1.7505016722408027, "grad_norm": 1.3003470332939664, "learning_rate": 8.217511513780056e-06, "loss": 1.254, "step": 1309 }, { "epoch": 1.7518394648829432, "grad_norm": 1.076094285836178, "learning_rate": 8.213938048432697e-06, "loss": 0.9872, "step": 1310 }, { "epoch": 1.7531772575250835, "grad_norm": 0.9551157153532198, "learning_rate": 8.210361783401491e-06, "loss": 1.1095, "step": 1311 }, { "epoch": 1.7545150501672242, "grad_norm": 0.808097456519594, "learning_rate": 8.206782721801747e-06, "loss": 1.3173, "step": 1312 }, { "epoch": 1.7558528428093645, "grad_norm": 1.0709922007681472, "learning_rate": 8.203200866751212e-06, "loss": 1.1375, "step": 1313 }, { "epoch": 1.757190635451505, "grad_norm": 0.9734908853322948, "learning_rate": 8.19961622137006e-06, "loss": 0.8683, "step": 1314 }, { "epoch": 1.7585284280936455, "grad_norm": 1.044194087314653, "learning_rate": 8.196028788780905e-06, "loss": 1.0535, "step": 1315 }, { "epoch": 1.7598662207357858, "grad_norm": 1.0560708061952861, "learning_rate": 8.192438572108786e-06, "loss": 1.0115, "step": 1316 }, { "epoch": 1.7612040133779265, "grad_norm": 0.8574251862075222, "learning_rate": 8.188845574481162e-06, "loss": 1.2823, "step": 1317 }, { "epoch": 1.7625418060200668, "grad_norm": 0.8728276566516617, "learning_rate": 8.185249799027919e-06, "loss": 0.962, "step": 1318 }, { "epoch": 1.7638795986622073, "grad_norm": 0.9197258182073648, "learning_rate": 8.181651248881364e-06, "loss": 1.0319, "step": 1319 }, { "epoch": 1.7652173913043478, "grad_norm": 1.1625762832353825, "learning_rate": 8.178049927176217e-06, "loss": 1.1819, "step": 1320 }, { "epoch": 1.7665551839464881, "grad_norm": 1.0381611713416592, "learning_rate": 8.174445837049614e-06, "loss": 1.2826, "step": 1321 }, { "epoch": 1.7678929765886289, "grad_norm": 1.0300290732346031, "learning_rate": 8.170838981641108e-06, "loss": 1.0224, "step": 1322 }, { "epoch": 1.7692307692307692, "grad_norm": 1.025202157088203, "learning_rate": 8.167229364092648e-06, "loss": 1.1876, "step": 1323 }, { "epoch": 1.7705685618729097, "grad_norm": 1.0852655913182763, "learning_rate": 8.163616987548605e-06, "loss": 1.3563, "step": 1324 }, { "epoch": 1.7719063545150502, "grad_norm": 1.3712649127644416, "learning_rate": 8.16000185515574e-06, "loss": 0.784, "step": 1325 }, { "epoch": 1.7732441471571905, "grad_norm": 1.1728348042483838, "learning_rate": 8.15638397006322e-06, "loss": 1.2695, "step": 1326 }, { "epoch": 1.7745819397993312, "grad_norm": 0.640307543289293, "learning_rate": 8.152763335422612e-06, "loss": 0.8601, "step": 1327 }, { "epoch": 1.7759197324414715, "grad_norm": 1.154487316794997, "learning_rate": 8.14913995438788e-06, "loss": 0.914, "step": 1328 }, { "epoch": 1.777257525083612, "grad_norm": 0.9859145771135289, "learning_rate": 8.145513830115367e-06, "loss": 1.0989, "step": 1329 }, { "epoch": 1.7785953177257525, "grad_norm": 1.1647135922366698, "learning_rate": 8.141884965763822e-06, "loss": 1.1759, "step": 1330 }, { "epoch": 1.7799331103678928, "grad_norm": 1.0342530846453222, "learning_rate": 8.138253364494374e-06, "loss": 0.9837, "step": 1331 }, { "epoch": 1.7812709030100335, "grad_norm": 0.837938910567284, "learning_rate": 8.134619029470535e-06, "loss": 1.005, "step": 1332 }, { "epoch": 1.7826086956521738, "grad_norm": 1.2533933644721498, "learning_rate": 8.130981963858195e-06, "loss": 1.1515, "step": 1333 }, { "epoch": 1.7839464882943143, "grad_norm": 1.128249881494306, "learning_rate": 8.127342170825635e-06, "loss": 1.0572, "step": 1334 }, { "epoch": 1.7852842809364549, "grad_norm": 1.0792848072350354, "learning_rate": 8.1236996535435e-06, "loss": 0.9562, "step": 1335 }, { "epoch": 1.7866220735785954, "grad_norm": 0.9017651295186162, "learning_rate": 8.120054415184811e-06, "loss": 1.1188, "step": 1336 }, { "epoch": 1.7879598662207359, "grad_norm": 1.0351541665435258, "learning_rate": 8.116406458924964e-06, "loss": 1.253, "step": 1337 }, { "epoch": 1.7892976588628762, "grad_norm": 0.9063543467723881, "learning_rate": 8.112755787941718e-06, "loss": 0.8942, "step": 1338 }, { "epoch": 1.7906354515050167, "grad_norm": 1.108837740597617, "learning_rate": 8.109102405415195e-06, "loss": 1.1704, "step": 1339 }, { "epoch": 1.7919732441471572, "grad_norm": 0.8740200627102102, "learning_rate": 8.105446314527885e-06, "loss": 1.0889, "step": 1340 }, { "epoch": 1.7933110367892977, "grad_norm": 1.077544667607713, "learning_rate": 8.101787518464634e-06, "loss": 1.0473, "step": 1341 }, { "epoch": 1.7946488294314382, "grad_norm": 0.9601629911662485, "learning_rate": 8.098126020412644e-06, "loss": 0.8402, "step": 1342 }, { "epoch": 1.7959866220735785, "grad_norm": 0.9124585121177934, "learning_rate": 8.094461823561473e-06, "loss": 1.0298, "step": 1343 }, { "epoch": 1.797324414715719, "grad_norm": 1.095611074989212, "learning_rate": 8.090794931103026e-06, "loss": 1.2043, "step": 1344 }, { "epoch": 1.7986622073578595, "grad_norm": 0.8571760254072552, "learning_rate": 8.087125346231562e-06, "loss": 0.9309, "step": 1345 }, { "epoch": 1.8, "grad_norm": 0.9486218660387196, "learning_rate": 8.083453072143678e-06, "loss": 1.2729, "step": 1346 }, { "epoch": 1.8013377926421406, "grad_norm": 0.7660307471692658, "learning_rate": 8.079778112038318e-06, "loss": 1.1521, "step": 1347 }, { "epoch": 1.8026755852842808, "grad_norm": 1.1184480446830192, "learning_rate": 8.07610046911677e-06, "loss": 1.0746, "step": 1348 }, { "epoch": 1.8040133779264214, "grad_norm": 0.898887204302731, "learning_rate": 8.072420146582649e-06, "loss": 0.9599, "step": 1349 }, { "epoch": 1.8053511705685619, "grad_norm": 1.0713434137629256, "learning_rate": 8.068737147641913e-06, "loss": 1.1387, "step": 1350 }, { "epoch": 1.8066889632107024, "grad_norm": 0.9002000319830386, "learning_rate": 8.065051475502847e-06, "loss": 0.9544, "step": 1351 }, { "epoch": 1.808026755852843, "grad_norm": 1.0506106934660453, "learning_rate": 8.061363133376065e-06, "loss": 0.937, "step": 1352 }, { "epoch": 1.8093645484949832, "grad_norm": 0.8946140922173809, "learning_rate": 8.057672124474508e-06, "loss": 0.9293, "step": 1353 }, { "epoch": 1.8107023411371237, "grad_norm": 1.0852819402480598, "learning_rate": 8.05397845201344e-06, "loss": 0.9373, "step": 1354 }, { "epoch": 1.8120401337792642, "grad_norm": 0.9951456673085664, "learning_rate": 8.050282119210443e-06, "loss": 1.1825, "step": 1355 }, { "epoch": 1.8133779264214047, "grad_norm": 1.145097780346333, "learning_rate": 8.046583129285422e-06, "loss": 1.1968, "step": 1356 }, { "epoch": 1.8147157190635452, "grad_norm": 1.0301782695725679, "learning_rate": 8.042881485460591e-06, "loss": 0.9833, "step": 1357 }, { "epoch": 1.8160535117056855, "grad_norm": 1.036066455456211, "learning_rate": 8.039177190960476e-06, "loss": 1.1485, "step": 1358 }, { "epoch": 1.8173913043478263, "grad_norm": 1.0087505037894362, "learning_rate": 8.035470249011916e-06, "loss": 0.8872, "step": 1359 }, { "epoch": 1.8187290969899665, "grad_norm": 0.843404797650954, "learning_rate": 8.031760662844053e-06, "loss": 0.957, "step": 1360 }, { "epoch": 1.820066889632107, "grad_norm": 0.9564551925613984, "learning_rate": 8.028048435688333e-06, "loss": 0.997, "step": 1361 }, { "epoch": 1.8214046822742476, "grad_norm": 1.157529863796313, "learning_rate": 8.024333570778507e-06, "loss": 1.1043, "step": 1362 }, { "epoch": 1.8227424749163879, "grad_norm": 1.2840567851736746, "learning_rate": 8.020616071350613e-06, "loss": 1.1875, "step": 1363 }, { "epoch": 1.8240802675585286, "grad_norm": 1.0243558525465912, "learning_rate": 8.016895940642994e-06, "loss": 1.1155, "step": 1364 }, { "epoch": 1.8254180602006689, "grad_norm": 1.0065336580216615, "learning_rate": 8.013173181896283e-06, "loss": 1.2216, "step": 1365 }, { "epoch": 1.8267558528428094, "grad_norm": 1.3363122339146638, "learning_rate": 8.0094477983534e-06, "loss": 0.9814, "step": 1366 }, { "epoch": 1.82809364548495, "grad_norm": 0.7779427282023568, "learning_rate": 8.005719793259552e-06, "loss": 1.1153, "step": 1367 }, { "epoch": 1.8294314381270902, "grad_norm": 1.0755501199431554, "learning_rate": 8.00198916986223e-06, "loss": 1.0431, "step": 1368 }, { "epoch": 1.830769230769231, "grad_norm": 0.9748043016050415, "learning_rate": 7.998255931411208e-06, "loss": 1.3053, "step": 1369 }, { "epoch": 1.8321070234113712, "grad_norm": 0.9119017627719722, "learning_rate": 7.994520081158534e-06, "loss": 1.3737, "step": 1370 }, { "epoch": 1.8334448160535117, "grad_norm": 1.0839599067026016, "learning_rate": 7.990781622358535e-06, "loss": 1.3495, "step": 1371 }, { "epoch": 1.8347826086956522, "grad_norm": 1.106571160893858, "learning_rate": 7.987040558267807e-06, "loss": 1.212, "step": 1372 }, { "epoch": 1.8361204013377925, "grad_norm": 0.8973960159008225, "learning_rate": 7.983296892145218e-06, "loss": 1.0797, "step": 1373 }, { "epoch": 1.8374581939799333, "grad_norm": 1.309958952369714, "learning_rate": 7.979550627251901e-06, "loss": 1.215, "step": 1374 }, { "epoch": 1.8387959866220736, "grad_norm": 0.9699885667783271, "learning_rate": 7.975801766851255e-06, "loss": 1.1415, "step": 1375 }, { "epoch": 1.840133779264214, "grad_norm": 1.1347701097920875, "learning_rate": 7.972050314208934e-06, "loss": 1.1734, "step": 1376 }, { "epoch": 1.8414715719063546, "grad_norm": 1.1141809401002065, "learning_rate": 7.968296272592862e-06, "loss": 1.1199, "step": 1377 }, { "epoch": 1.8428093645484949, "grad_norm": 0.994538507291041, "learning_rate": 7.964539645273204e-06, "loss": 1.2589, "step": 1378 }, { "epoch": 1.8441471571906356, "grad_norm": 1.0797115950007015, "learning_rate": 7.960780435522387e-06, "loss": 1.1857, "step": 1379 }, { "epoch": 1.845484949832776, "grad_norm": 1.071479932808848, "learning_rate": 7.957018646615085e-06, "loss": 0.9244, "step": 1380 }, { "epoch": 1.8468227424749164, "grad_norm": 1.2237347557367744, "learning_rate": 7.953254281828217e-06, "loss": 1.0616, "step": 1381 }, { "epoch": 1.848160535117057, "grad_norm": 1.0647950890199858, "learning_rate": 7.94948734444095e-06, "loss": 1.1283, "step": 1382 }, { "epoch": 1.8494983277591972, "grad_norm": 0.8956932162810435, "learning_rate": 7.945717837734688e-06, "loss": 1.1059, "step": 1383 }, { "epoch": 1.850836120401338, "grad_norm": 1.2183462072686708, "learning_rate": 7.941945764993074e-06, "loss": 1.0443, "step": 1384 }, { "epoch": 1.8521739130434782, "grad_norm": 1.3440026282322615, "learning_rate": 7.938171129501988e-06, "loss": 1.1114, "step": 1385 }, { "epoch": 1.8535117056856187, "grad_norm": 0.9545914604665438, "learning_rate": 7.934393934549542e-06, "loss": 1.0984, "step": 1386 }, { "epoch": 1.8548494983277592, "grad_norm": 0.8724043196916478, "learning_rate": 7.930614183426074e-06, "loss": 1.3964, "step": 1387 }, { "epoch": 1.8561872909698995, "grad_norm": 0.8980616747034307, "learning_rate": 7.926831879424154e-06, "loss": 1.1877, "step": 1388 }, { "epoch": 1.8575250836120403, "grad_norm": 1.3232092325368487, "learning_rate": 7.923047025838573e-06, "loss": 0.9576, "step": 1389 }, { "epoch": 1.8588628762541806, "grad_norm": 0.9329274655920894, "learning_rate": 7.919259625966342e-06, "loss": 0.9117, "step": 1390 }, { "epoch": 1.860200668896321, "grad_norm": 1.330835281841632, "learning_rate": 7.915469683106694e-06, "loss": 1.2209, "step": 1391 }, { "epoch": 1.8615384615384616, "grad_norm": 0.8340917406907226, "learning_rate": 7.91167720056107e-06, "loss": 1.1289, "step": 1392 }, { "epoch": 1.8628762541806019, "grad_norm": 0.989609153475318, "learning_rate": 7.907882181633134e-06, "loss": 1.2396, "step": 1393 }, { "epoch": 1.8642140468227426, "grad_norm": 1.2827534308303896, "learning_rate": 7.90408462962875e-06, "loss": 1.1345, "step": 1394 }, { "epoch": 1.865551839464883, "grad_norm": 0.9394931378300206, "learning_rate": 7.900284547855992e-06, "loss": 1.0589, "step": 1395 }, { "epoch": 1.8668896321070234, "grad_norm": 1.001945914780929, "learning_rate": 7.896481939625139e-06, "loss": 1.0639, "step": 1396 }, { "epoch": 1.868227424749164, "grad_norm": 0.9262520462960626, "learning_rate": 7.892676808248666e-06, "loss": 1.0904, "step": 1397 }, { "epoch": 1.8695652173913042, "grad_norm": 1.0160689799454656, "learning_rate": 7.888869157041257e-06, "loss": 1.3072, "step": 1398 }, { "epoch": 1.870903010033445, "grad_norm": 1.0512481754834342, "learning_rate": 7.885058989319776e-06, "loss": 1.0011, "step": 1399 }, { "epoch": 1.8722408026755852, "grad_norm": 0.7857036309870001, "learning_rate": 7.88124630840329e-06, "loss": 1.1994, "step": 1400 }, { "epoch": 1.8735785953177257, "grad_norm": 1.1898975324996341, "learning_rate": 7.87743111761305e-06, "loss": 1.1238, "step": 1401 }, { "epoch": 1.8749163879598663, "grad_norm": 1.2029977282844446, "learning_rate": 7.8736134202725e-06, "loss": 1.4371, "step": 1402 }, { "epoch": 1.8762541806020065, "grad_norm": 0.880927567870842, "learning_rate": 7.869793219707258e-06, "loss": 0.898, "step": 1403 }, { "epoch": 1.8775919732441473, "grad_norm": 1.2064042173452543, "learning_rate": 7.865970519245129e-06, "loss": 1.1744, "step": 1404 }, { "epoch": 1.8789297658862876, "grad_norm": 0.8969098232691118, "learning_rate": 7.862145322216092e-06, "loss": 1.2647, "step": 1405 }, { "epoch": 1.880267558528428, "grad_norm": 1.204617412056039, "learning_rate": 7.858317631952307e-06, "loss": 1.1179, "step": 1406 }, { "epoch": 1.8816053511705686, "grad_norm": 0.8426741969207092, "learning_rate": 7.8544874517881e-06, "loss": 1.1105, "step": 1407 }, { "epoch": 1.8829431438127089, "grad_norm": 0.8452371234073803, "learning_rate": 7.850654785059966e-06, "loss": 1.2007, "step": 1408 }, { "epoch": 1.8842809364548496, "grad_norm": 0.9416266665393999, "learning_rate": 7.846819635106569e-06, "loss": 1.1575, "step": 1409 }, { "epoch": 1.88561872909699, "grad_norm": 1.024385196220546, "learning_rate": 7.842982005268733e-06, "loss": 1.0665, "step": 1410 }, { "epoch": 1.8869565217391304, "grad_norm": 0.9009258579598443, "learning_rate": 7.83914189888945e-06, "loss": 1.1775, "step": 1411 }, { "epoch": 1.888294314381271, "grad_norm": 1.058805866553999, "learning_rate": 7.835299319313854e-06, "loss": 1.1797, "step": 1412 }, { "epoch": 1.8896321070234112, "grad_norm": 1.012187557263587, "learning_rate": 7.831454269889251e-06, "loss": 0.9086, "step": 1413 }, { "epoch": 1.890969899665552, "grad_norm": 0.7361897402567618, "learning_rate": 7.827606753965086e-06, "loss": 1.07, "step": 1414 }, { "epoch": 1.8923076923076922, "grad_norm": 1.0653457743777064, "learning_rate": 7.823756774892961e-06, "loss": 1.0483, "step": 1415 }, { "epoch": 1.8936454849498328, "grad_norm": 0.8702391644139729, "learning_rate": 7.819904336026615e-06, "loss": 1.2836, "step": 1416 }, { "epoch": 1.8949832775919733, "grad_norm": 0.9727001548574422, "learning_rate": 7.816049440721937e-06, "loss": 1.1671, "step": 1417 }, { "epoch": 1.8963210702341136, "grad_norm": 0.8600602509763807, "learning_rate": 7.812192092336951e-06, "loss": 0.9611, "step": 1418 }, { "epoch": 1.8976588628762543, "grad_norm": 1.0638193768382502, "learning_rate": 7.808332294231824e-06, "loss": 1.0931, "step": 1419 }, { "epoch": 1.8989966555183946, "grad_norm": 1.0216135636851302, "learning_rate": 7.80447004976885e-06, "loss": 1.1348, "step": 1420 }, { "epoch": 1.900334448160535, "grad_norm": 1.366158231672114, "learning_rate": 7.800605362312456e-06, "loss": 1.1803, "step": 1421 }, { "epoch": 1.9016722408026756, "grad_norm": 0.7621199733830623, "learning_rate": 7.796738235229203e-06, "loss": 1.2148, "step": 1422 }, { "epoch": 1.903010033444816, "grad_norm": 1.0618869860533038, "learning_rate": 7.792868671887768e-06, "loss": 1.3271, "step": 1423 }, { "epoch": 1.9043478260869566, "grad_norm": 0.8987644640942961, "learning_rate": 7.788996675658955e-06, "loss": 0.8803, "step": 1424 }, { "epoch": 1.905685618729097, "grad_norm": 1.0251777448075632, "learning_rate": 7.785122249915688e-06, "loss": 1.1117, "step": 1425 }, { "epoch": 1.9070234113712374, "grad_norm": 0.91117358038687, "learning_rate": 7.781245398033009e-06, "loss": 1.1119, "step": 1426 }, { "epoch": 1.908361204013378, "grad_norm": 0.9592705419655734, "learning_rate": 7.777366123388065e-06, "loss": 1.1767, "step": 1427 }, { "epoch": 1.9096989966555182, "grad_norm": 1.1657302309548256, "learning_rate": 7.773484429360122e-06, "loss": 0.8874, "step": 1428 }, { "epoch": 1.911036789297659, "grad_norm": 0.797089442615022, "learning_rate": 7.769600319330553e-06, "loss": 0.9462, "step": 1429 }, { "epoch": 1.9123745819397993, "grad_norm": 1.1042223250806713, "learning_rate": 7.765713796682829e-06, "loss": 1.2103, "step": 1430 }, { "epoch": 1.9137123745819398, "grad_norm": 0.9557854186997792, "learning_rate": 7.76182486480253e-06, "loss": 0.9703, "step": 1431 }, { "epoch": 1.9150501672240803, "grad_norm": 0.8481631059991172, "learning_rate": 7.75793352707733e-06, "loss": 1.3355, "step": 1432 }, { "epoch": 1.9163879598662206, "grad_norm": 1.0535012102181593, "learning_rate": 7.754039786897004e-06, "loss": 1.1375, "step": 1433 }, { "epoch": 1.9177257525083613, "grad_norm": 0.8579524780839797, "learning_rate": 7.750143647653409e-06, "loss": 1.0552, "step": 1434 }, { "epoch": 1.9190635451505016, "grad_norm": 0.795817548535621, "learning_rate": 7.746245112740507e-06, "loss": 1.2562, "step": 1435 }, { "epoch": 1.920401337792642, "grad_norm": 0.9273868241881819, "learning_rate": 7.742344185554335e-06, "loss": 0.9566, "step": 1436 }, { "epoch": 1.9217391304347826, "grad_norm": 1.1051274081497386, "learning_rate": 7.738440869493018e-06, "loss": 1.0123, "step": 1437 }, { "epoch": 1.9230769230769231, "grad_norm": 0.8008943490596314, "learning_rate": 7.734535167956761e-06, "loss": 1.2234, "step": 1438 }, { "epoch": 1.9244147157190636, "grad_norm": 1.2681498476082467, "learning_rate": 7.73062708434785e-06, "loss": 1.3369, "step": 1439 }, { "epoch": 1.925752508361204, "grad_norm": 0.8633150635486877, "learning_rate": 7.726716622070643e-06, "loss": 1.0542, "step": 1440 }, { "epoch": 1.9270903010033444, "grad_norm": 1.2226587116929093, "learning_rate": 7.722803784531572e-06, "loss": 1.0685, "step": 1441 }, { "epoch": 1.928428093645485, "grad_norm": 1.0019218630990392, "learning_rate": 7.718888575139134e-06, "loss": 1.11, "step": 1442 }, { "epoch": 1.9297658862876255, "grad_norm": 1.5517059578278822, "learning_rate": 7.714970997303898e-06, "loss": 0.8446, "step": 1443 }, { "epoch": 1.931103678929766, "grad_norm": 1.1728263113266852, "learning_rate": 7.711051054438491e-06, "loss": 1.2924, "step": 1444 }, { "epoch": 1.9324414715719063, "grad_norm": 1.16411623992991, "learning_rate": 7.707128749957606e-06, "loss": 0.8534, "step": 1445 }, { "epoch": 1.9337792642140468, "grad_norm": 0.9760210094866245, "learning_rate": 7.703204087277989e-06, "loss": 1.1113, "step": 1446 }, { "epoch": 1.9351170568561873, "grad_norm": 1.0813554838173045, "learning_rate": 7.699277069818439e-06, "loss": 1.0378, "step": 1447 }, { "epoch": 1.9364548494983278, "grad_norm": 0.7748899039720765, "learning_rate": 7.69534770099981e-06, "loss": 1.2517, "step": 1448 }, { "epoch": 1.9377926421404683, "grad_norm": 1.0202401273729143, "learning_rate": 7.691415984244998e-06, "loss": 1.3606, "step": 1449 }, { "epoch": 1.9391304347826086, "grad_norm": 0.8660984776106343, "learning_rate": 7.687481922978955e-06, "loss": 1.2127, "step": 1450 }, { "epoch": 1.9404682274247491, "grad_norm": 0.9569688483285013, "learning_rate": 7.683545520628667e-06, "loss": 0.817, "step": 1451 }, { "epoch": 1.9418060200668896, "grad_norm": 0.8861515738668474, "learning_rate": 7.679606780623162e-06, "loss": 1.0454, "step": 1452 }, { "epoch": 1.9431438127090301, "grad_norm": 1.2063822816593837, "learning_rate": 7.675665706393502e-06, "loss": 1.1881, "step": 1453 }, { "epoch": 1.9444816053511706, "grad_norm": 1.128577303304323, "learning_rate": 7.671722301372788e-06, "loss": 1.0146, "step": 1454 }, { "epoch": 1.945819397993311, "grad_norm": 0.9241336215093544, "learning_rate": 7.667776568996143e-06, "loss": 1.0314, "step": 1455 }, { "epoch": 1.9471571906354515, "grad_norm": 1.044447426514981, "learning_rate": 7.663828512700724e-06, "loss": 0.995, "step": 1456 }, { "epoch": 1.948494983277592, "grad_norm": 0.9659524547821133, "learning_rate": 7.65987813592571e-06, "loss": 1.0474, "step": 1457 }, { "epoch": 1.9498327759197325, "grad_norm": 0.933642475380743, "learning_rate": 7.655925442112303e-06, "loss": 1.2083, "step": 1458 }, { "epoch": 1.951170568561873, "grad_norm": 1.0465759035641653, "learning_rate": 7.651970434703724e-06, "loss": 1.2151, "step": 1459 }, { "epoch": 1.9525083612040133, "grad_norm": 0.9151229813768318, "learning_rate": 7.648013117145203e-06, "loss": 1.2635, "step": 1460 }, { "epoch": 1.953846153846154, "grad_norm": 1.1000356612672397, "learning_rate": 7.64405349288399e-06, "loss": 1.0403, "step": 1461 }, { "epoch": 1.9551839464882943, "grad_norm": 0.7893958987104329, "learning_rate": 7.640091565369339e-06, "loss": 1.2213, "step": 1462 }, { "epoch": 1.9565217391304348, "grad_norm": 1.0062489151585723, "learning_rate": 7.636127338052513e-06, "loss": 1.4027, "step": 1463 }, { "epoch": 1.9578595317725753, "grad_norm": 1.110702940757215, "learning_rate": 7.63216081438678e-06, "loss": 1.2155, "step": 1464 }, { "epoch": 1.9591973244147156, "grad_norm": 0.8797975080131499, "learning_rate": 7.628191997827405e-06, "loss": 1.0768, "step": 1465 }, { "epoch": 1.9605351170568563, "grad_norm": 0.8895256474909151, "learning_rate": 7.624220891831653e-06, "loss": 0.8581, "step": 1466 }, { "epoch": 1.9618729096989966, "grad_norm": 1.056455396167, "learning_rate": 7.62024749985878e-06, "loss": 1.0339, "step": 1467 }, { "epoch": 1.9632107023411371, "grad_norm": 1.0287021782043828, "learning_rate": 7.616271825370037e-06, "loss": 1.067, "step": 1468 }, { "epoch": 1.9645484949832777, "grad_norm": 1.0297478805529892, "learning_rate": 7.612293871828662e-06, "loss": 1.4511, "step": 1469 }, { "epoch": 1.965886287625418, "grad_norm": 0.7336666890059218, "learning_rate": 7.6083136426998786e-06, "loss": 0.943, "step": 1470 }, { "epoch": 1.9672240802675587, "grad_norm": 1.125372757291073, "learning_rate": 7.604331141450889e-06, "loss": 1.1556, "step": 1471 }, { "epoch": 1.968561872909699, "grad_norm": 1.2354019422728049, "learning_rate": 7.600346371550882e-06, "loss": 1.3348, "step": 1472 }, { "epoch": 1.9698996655518395, "grad_norm": 0.8780276725269451, "learning_rate": 7.596359336471015e-06, "loss": 0.9639, "step": 1473 }, { "epoch": 1.97123745819398, "grad_norm": 1.146416342605179, "learning_rate": 7.592370039684424e-06, "loss": 1.2375, "step": 1474 }, { "epoch": 1.9725752508361203, "grad_norm": 0.9125185686446718, "learning_rate": 7.588378484666214e-06, "loss": 1.0303, "step": 1475 }, { "epoch": 1.973913043478261, "grad_norm": 0.8334019474244095, "learning_rate": 7.584384674893454e-06, "loss": 1.0832, "step": 1476 }, { "epoch": 1.9752508361204013, "grad_norm": 1.1413497965551236, "learning_rate": 7.58038861384518e-06, "loss": 1.1445, "step": 1477 }, { "epoch": 1.9765886287625418, "grad_norm": 0.9141609655312544, "learning_rate": 7.576390305002389e-06, "loss": 1.3798, "step": 1478 }, { "epoch": 1.9779264214046823, "grad_norm": 0.9244003229491952, "learning_rate": 7.572389751848037e-06, "loss": 1.0318, "step": 1479 }, { "epoch": 1.9792642140468226, "grad_norm": 1.0404344250905808, "learning_rate": 7.568386957867033e-06, "loss": 1.1127, "step": 1480 }, { "epoch": 1.9806020066889634, "grad_norm": 0.9352623828531544, "learning_rate": 7.564381926546238e-06, "loss": 1.2063, "step": 1481 }, { "epoch": 1.9819397993311036, "grad_norm": 0.8459983926096373, "learning_rate": 7.560374661374463e-06, "loss": 1.1938, "step": 1482 }, { "epoch": 1.9832775919732442, "grad_norm": 1.145707223409049, "learning_rate": 7.556365165842466e-06, "loss": 1.0017, "step": 1483 }, { "epoch": 1.9846153846153847, "grad_norm": 0.884604064739179, "learning_rate": 7.552353443442944e-06, "loss": 0.8679, "step": 1484 }, { "epoch": 1.985953177257525, "grad_norm": 1.1232716024200489, "learning_rate": 7.548339497670538e-06, "loss": 0.7912, "step": 1485 }, { "epoch": 1.9872909698996657, "grad_norm": 1.0728879754137894, "learning_rate": 7.544323332021826e-06, "loss": 1.01, "step": 1486 }, { "epoch": 1.988628762541806, "grad_norm": 1.0072455505036682, "learning_rate": 7.540304949995314e-06, "loss": 1.0761, "step": 1487 }, { "epoch": 1.9899665551839465, "grad_norm": 0.8779619715391508, "learning_rate": 7.536284355091443e-06, "loss": 1.1961, "step": 1488 }, { "epoch": 1.991304347826087, "grad_norm": 1.0306166393879193, "learning_rate": 7.532261550812585e-06, "loss": 1.0611, "step": 1489 }, { "epoch": 1.9926421404682273, "grad_norm": 1.0182897735663456, "learning_rate": 7.528236540663031e-06, "loss": 1.0615, "step": 1490 }, { "epoch": 1.993979933110368, "grad_norm": 1.112623075010365, "learning_rate": 7.524209328148995e-06, "loss": 1.1867, "step": 1491 }, { "epoch": 1.9953177257525083, "grad_norm": 1.2642431045436857, "learning_rate": 7.520179916778608e-06, "loss": 1.3197, "step": 1492 }, { "epoch": 1.9966555183946488, "grad_norm": 0.9952009933520749, "learning_rate": 7.516148310061921e-06, "loss": 1.3218, "step": 1493 }, { "epoch": 1.9979933110367893, "grad_norm": 1.3238381401771386, "learning_rate": 7.512114511510893e-06, "loss": 1.0888, "step": 1494 }, { "epoch": 1.9993311036789296, "grad_norm": 0.9622769041531732, "learning_rate": 7.508078524639397e-06, "loss": 1.2493, "step": 1495 }, { "epoch": 2.0, "grad_norm": 2.1734249378876833, "learning_rate": 7.504040352963206e-06, "loss": 0.9008, "step": 1496 }, { "epoch": 2.0013377926421403, "grad_norm": 1.1582268467642853, "learning_rate": 7.500000000000001e-06, "loss": 0.9529, "step": 1497 }, { "epoch": 2.002675585284281, "grad_norm": 1.1626751672999238, "learning_rate": 7.495957469269361e-06, "loss": 1.1358, "step": 1498 }, { "epoch": 2.0040133779264213, "grad_norm": 1.0506950881146222, "learning_rate": 7.491912764292764e-06, "loss": 1.1929, "step": 1499 }, { "epoch": 2.005351170568562, "grad_norm": 0.9607017692058403, "learning_rate": 7.487865888593579e-06, "loss": 1.147, "step": 1500 }, { "epoch": 2.0066889632107023, "grad_norm": 0.8911123942925315, "learning_rate": 7.483816845697069e-06, "loss": 1.0209, "step": 1501 }, { "epoch": 2.0080267558528426, "grad_norm": 0.8764119861268106, "learning_rate": 7.479765639130384e-06, "loss": 1.0242, "step": 1502 }, { "epoch": 2.0093645484949834, "grad_norm": 0.9761671485774847, "learning_rate": 7.4757122724225575e-06, "loss": 1.0344, "step": 1503 }, { "epoch": 2.0107023411371236, "grad_norm": 1.0504777721889755, "learning_rate": 7.471656749104503e-06, "loss": 1.1148, "step": 1504 }, { "epoch": 2.0120401337792644, "grad_norm": 0.9271748599900005, "learning_rate": 7.467599072709019e-06, "loss": 1.0993, "step": 1505 }, { "epoch": 2.0133779264214047, "grad_norm": 1.0179807890036645, "learning_rate": 7.463539246770775e-06, "loss": 0.7663, "step": 1506 }, { "epoch": 2.014715719063545, "grad_norm": 1.099208282213893, "learning_rate": 7.459477274826312e-06, "loss": 1.021, "step": 1507 }, { "epoch": 2.0160535117056857, "grad_norm": 1.0874363655632202, "learning_rate": 7.4554131604140425e-06, "loss": 0.9462, "step": 1508 }, { "epoch": 2.017391304347826, "grad_norm": 0.9638962330716166, "learning_rate": 7.451346907074245e-06, "loss": 0.9788, "step": 1509 }, { "epoch": 2.0187290969899667, "grad_norm": 1.0124415029835236, "learning_rate": 7.447278518349062e-06, "loss": 0.9533, "step": 1510 }, { "epoch": 2.020066889632107, "grad_norm": 0.9584494581847505, "learning_rate": 7.443207997782495e-06, "loss": 1.0745, "step": 1511 }, { "epoch": 2.0214046822742473, "grad_norm": 1.1380659333835579, "learning_rate": 7.439135348920403e-06, "loss": 1.1317, "step": 1512 }, { "epoch": 2.022742474916388, "grad_norm": 1.1711449542620707, "learning_rate": 7.435060575310498e-06, "loss": 1.0659, "step": 1513 }, { "epoch": 2.0240802675585283, "grad_norm": 0.9452129077976054, "learning_rate": 7.430983680502344e-06, "loss": 1.0413, "step": 1514 }, { "epoch": 2.025418060200669, "grad_norm": 1.1288752130823188, "learning_rate": 7.426904668047352e-06, "loss": 1.2516, "step": 1515 }, { "epoch": 2.0267558528428093, "grad_norm": 1.068721793637266, "learning_rate": 7.4228235414987805e-06, "loss": 1.0441, "step": 1516 }, { "epoch": 2.0280936454849496, "grad_norm": 1.0085699604215819, "learning_rate": 7.418740304411725e-06, "loss": 0.933, "step": 1517 }, { "epoch": 2.0294314381270904, "grad_norm": 0.8834381269789208, "learning_rate": 7.4146549603431225e-06, "loss": 1.0373, "step": 1518 }, { "epoch": 2.0307692307692307, "grad_norm": 1.3949824707575578, "learning_rate": 7.4105675128517456e-06, "loss": 1.1588, "step": 1519 }, { "epoch": 2.0321070234113714, "grad_norm": 0.9328754311685377, "learning_rate": 7.4064779654981966e-06, "loss": 0.6705, "step": 1520 }, { "epoch": 2.0334448160535117, "grad_norm": 0.8718440609028515, "learning_rate": 7.40238632184491e-06, "loss": 1.2494, "step": 1521 }, { "epoch": 2.034782608695652, "grad_norm": 0.9626318699161587, "learning_rate": 7.398292585456144e-06, "loss": 1.1309, "step": 1522 }, { "epoch": 2.0361204013377927, "grad_norm": 0.8990652170929269, "learning_rate": 7.39419675989798e-06, "loss": 1.0722, "step": 1523 }, { "epoch": 2.037458193979933, "grad_norm": 1.1225887115614213, "learning_rate": 7.390098848738324e-06, "loss": 0.9333, "step": 1524 }, { "epoch": 2.0387959866220737, "grad_norm": 0.907691318085966, "learning_rate": 7.385998855546892e-06, "loss": 1.019, "step": 1525 }, { "epoch": 2.040133779264214, "grad_norm": 1.0206233270251608, "learning_rate": 7.381896783895217e-06, "loss": 0.9829, "step": 1526 }, { "epoch": 2.0414715719063543, "grad_norm": 0.8762307862648366, "learning_rate": 7.377792637356644e-06, "loss": 0.9782, "step": 1527 }, { "epoch": 2.042809364548495, "grad_norm": 0.9893952113825407, "learning_rate": 7.373686419506321e-06, "loss": 0.9199, "step": 1528 }, { "epoch": 2.0441471571906353, "grad_norm": 1.040955985201199, "learning_rate": 7.369578133921205e-06, "loss": 1.1304, "step": 1529 }, { "epoch": 2.045484949832776, "grad_norm": 0.8835502758121206, "learning_rate": 7.365467784180051e-06, "loss": 0.9639, "step": 1530 }, { "epoch": 2.0468227424749164, "grad_norm": 0.9911742092622862, "learning_rate": 7.361355373863415e-06, "loss": 1.0293, "step": 1531 }, { "epoch": 2.0481605351170566, "grad_norm": 0.9904967528872938, "learning_rate": 7.357240906553644e-06, "loss": 0.8114, "step": 1532 }, { "epoch": 2.0494983277591974, "grad_norm": 0.8007738210168789, "learning_rate": 7.35312438583488e-06, "loss": 1.0243, "step": 1533 }, { "epoch": 2.0508361204013377, "grad_norm": 0.9136436656011789, "learning_rate": 7.349005815293055e-06, "loss": 0.9717, "step": 1534 }, { "epoch": 2.0521739130434784, "grad_norm": 0.9031175479582793, "learning_rate": 7.344885198515881e-06, "loss": 1.1065, "step": 1535 }, { "epoch": 2.0535117056856187, "grad_norm": 0.7854891246648976, "learning_rate": 7.340762539092858e-06, "loss": 0.8025, "step": 1536 }, { "epoch": 2.054849498327759, "grad_norm": 0.9103590589091035, "learning_rate": 7.336637840615265e-06, "loss": 1.0525, "step": 1537 }, { "epoch": 2.0561872909698997, "grad_norm": 0.9982011367525587, "learning_rate": 7.332511106676151e-06, "loss": 0.9762, "step": 1538 }, { "epoch": 2.05752508361204, "grad_norm": 0.8919991524243384, "learning_rate": 7.3283823408703466e-06, "loss": 1.1202, "step": 1539 }, { "epoch": 2.0588628762541807, "grad_norm": 1.0980009263292232, "learning_rate": 7.324251546794449e-06, "loss": 0.8688, "step": 1540 }, { "epoch": 2.060200668896321, "grad_norm": 0.95701723970262, "learning_rate": 7.320118728046818e-06, "loss": 1.0164, "step": 1541 }, { "epoch": 2.0615384615384613, "grad_norm": 0.8742716992961577, "learning_rate": 7.315983888227583e-06, "loss": 0.9058, "step": 1542 }, { "epoch": 2.062876254180602, "grad_norm": 2.5002868924145623, "learning_rate": 7.3118470309386325e-06, "loss": 1.3302, "step": 1543 }, { "epoch": 2.0642140468227423, "grad_norm": 1.120183931066008, "learning_rate": 7.3077081597836105e-06, "loss": 1.0024, "step": 1544 }, { "epoch": 2.065551839464883, "grad_norm": 1.1024110866988368, "learning_rate": 7.303567278367918e-06, "loss": 1.03, "step": 1545 }, { "epoch": 2.0668896321070234, "grad_norm": 1.0518549165294815, "learning_rate": 7.299424390298704e-06, "loss": 1.0297, "step": 1546 }, { "epoch": 2.068227424749164, "grad_norm": 1.0086944969091147, "learning_rate": 7.295279499184867e-06, "loss": 1.0135, "step": 1547 }, { "epoch": 2.0695652173913044, "grad_norm": 0.9422139325071827, "learning_rate": 7.291132608637053e-06, "loss": 1.3351, "step": 1548 }, { "epoch": 2.0709030100334447, "grad_norm": 1.0942983982025076, "learning_rate": 7.2869837222676445e-06, "loss": 0.844, "step": 1549 }, { "epoch": 2.0722408026755854, "grad_norm": 0.8644939172286806, "learning_rate": 7.282832843690768e-06, "loss": 0.9546, "step": 1550 }, { "epoch": 2.0735785953177257, "grad_norm": 0.9022114318978104, "learning_rate": 7.278679976522279e-06, "loss": 1.0088, "step": 1551 }, { "epoch": 2.074916387959866, "grad_norm": 0.8742289512766747, "learning_rate": 7.274525124379773e-06, "loss": 1.1198, "step": 1552 }, { "epoch": 2.0762541806020067, "grad_norm": 0.9306082895373801, "learning_rate": 7.2703682908825675e-06, "loss": 1.0528, "step": 1553 }, { "epoch": 2.077591973244147, "grad_norm": 1.1607914723546446, "learning_rate": 7.266209479651712e-06, "loss": 0.8514, "step": 1554 }, { "epoch": 2.0789297658862878, "grad_norm": 0.9683210199017045, "learning_rate": 7.262048694309976e-06, "loss": 0.9411, "step": 1555 }, { "epoch": 2.080267558528428, "grad_norm": 0.986832829104616, "learning_rate": 7.257885938481845e-06, "loss": 1.1061, "step": 1556 }, { "epoch": 2.0816053511705688, "grad_norm": 1.0141602650504782, "learning_rate": 7.253721215793528e-06, "loss": 1.055, "step": 1557 }, { "epoch": 2.082943143812709, "grad_norm": 1.1134592581900762, "learning_rate": 7.249554529872941e-06, "loss": 1.1756, "step": 1558 }, { "epoch": 2.0842809364548494, "grad_norm": 0.9799036915908764, "learning_rate": 7.245385884349716e-06, "loss": 0.9738, "step": 1559 }, { "epoch": 2.08561872909699, "grad_norm": 1.1185118804521705, "learning_rate": 7.241215282855189e-06, "loss": 0.9275, "step": 1560 }, { "epoch": 2.0869565217391304, "grad_norm": 0.8726796085247968, "learning_rate": 7.2370427290224e-06, "loss": 0.9987, "step": 1561 }, { "epoch": 2.088294314381271, "grad_norm": 0.9311492622455193, "learning_rate": 7.232868226486087e-06, "loss": 0.8278, "step": 1562 }, { "epoch": 2.0896321070234114, "grad_norm": 0.8268299607291658, "learning_rate": 7.2286917788826926e-06, "loss": 0.9852, "step": 1563 }, { "epoch": 2.0909698996655517, "grad_norm": 0.9279091774022891, "learning_rate": 7.224513389850345e-06, "loss": 1.0667, "step": 1564 }, { "epoch": 2.0923076923076924, "grad_norm": 0.8972842123500341, "learning_rate": 7.2203330630288714e-06, "loss": 1.0934, "step": 1565 }, { "epoch": 2.0936454849498327, "grad_norm": 1.0550523422221478, "learning_rate": 7.216150802059782e-06, "loss": 1.0473, "step": 1566 }, { "epoch": 2.0949832775919734, "grad_norm": 1.0723661697302294, "learning_rate": 7.211966610586274e-06, "loss": 0.9941, "step": 1567 }, { "epoch": 2.0963210702341137, "grad_norm": 0.9188268381394111, "learning_rate": 7.2077804922532245e-06, "loss": 0.8575, "step": 1568 }, { "epoch": 2.097658862876254, "grad_norm": 0.926612578008097, "learning_rate": 7.203592450707193e-06, "loss": 1.0835, "step": 1569 }, { "epoch": 2.0989966555183948, "grad_norm": 0.9286533514031133, "learning_rate": 7.1994024895964095e-06, "loss": 0.8713, "step": 1570 }, { "epoch": 2.100334448160535, "grad_norm": 0.8106350743183318, "learning_rate": 7.195210612570781e-06, "loss": 1.1181, "step": 1571 }, { "epoch": 2.101672240802676, "grad_norm": 0.8859518343332231, "learning_rate": 7.1910168232818765e-06, "loss": 1.1115, "step": 1572 }, { "epoch": 2.103010033444816, "grad_norm": 0.8921358663157047, "learning_rate": 7.1868211253829375e-06, "loss": 0.9319, "step": 1573 }, { "epoch": 2.1043478260869564, "grad_norm": 1.2359796085541126, "learning_rate": 7.182623522528866e-06, "loss": 1.2686, "step": 1574 }, { "epoch": 2.105685618729097, "grad_norm": 0.9669266359037734, "learning_rate": 7.178424018376224e-06, "loss": 1.1997, "step": 1575 }, { "epoch": 2.1070234113712374, "grad_norm": 1.0963437547025836, "learning_rate": 7.174222616583228e-06, "loss": 0.8662, "step": 1576 }, { "epoch": 2.108361204013378, "grad_norm": 1.2662405302280832, "learning_rate": 7.170019320809747e-06, "loss": 0.8786, "step": 1577 }, { "epoch": 2.1096989966555184, "grad_norm": 1.0387076985750996, "learning_rate": 7.165814134717303e-06, "loss": 0.7584, "step": 1578 }, { "epoch": 2.1110367892976587, "grad_norm": 1.2454485824635395, "learning_rate": 7.161607061969061e-06, "loss": 0.9695, "step": 1579 }, { "epoch": 2.1123745819397994, "grad_norm": 1.1424653258797268, "learning_rate": 7.157398106229834e-06, "loss": 1.1259, "step": 1580 }, { "epoch": 2.1137123745819397, "grad_norm": 1.1098155968580288, "learning_rate": 7.153187271166071e-06, "loss": 0.9751, "step": 1581 }, { "epoch": 2.1150501672240805, "grad_norm": 1.0538075199899775, "learning_rate": 7.148974560445859e-06, "loss": 0.7937, "step": 1582 }, { "epoch": 2.1163879598662207, "grad_norm": 0.7689814169317805, "learning_rate": 7.144759977738921e-06, "loss": 0.8453, "step": 1583 }, { "epoch": 2.117725752508361, "grad_norm": 1.057767759352348, "learning_rate": 7.14054352671661e-06, "loss": 1.0723, "step": 1584 }, { "epoch": 2.1190635451505018, "grad_norm": 0.884020608922426, "learning_rate": 7.136325211051905e-06, "loss": 0.9379, "step": 1585 }, { "epoch": 2.120401337792642, "grad_norm": 1.0113816318518056, "learning_rate": 7.132105034419411e-06, "loss": 1.24, "step": 1586 }, { "epoch": 2.121739130434783, "grad_norm": 1.0912744146319473, "learning_rate": 7.127883000495353e-06, "loss": 1.0013, "step": 1587 }, { "epoch": 2.123076923076923, "grad_norm": 0.9945799776201907, "learning_rate": 7.123659112957571e-06, "loss": 0.9119, "step": 1588 }, { "epoch": 2.1244147157190634, "grad_norm": 0.9650105464826692, "learning_rate": 7.119433375485527e-06, "loss": 0.919, "step": 1589 }, { "epoch": 2.125752508361204, "grad_norm": 1.0984889176158492, "learning_rate": 7.1152057917602904e-06, "loss": 0.72, "step": 1590 }, { "epoch": 2.1270903010033444, "grad_norm": 0.8563439949633473, "learning_rate": 7.110976365464537e-06, "loss": 1.2571, "step": 1591 }, { "epoch": 2.128428093645485, "grad_norm": 0.9272621221395067, "learning_rate": 7.10674510028255e-06, "loss": 1.1691, "step": 1592 }, { "epoch": 2.1297658862876254, "grad_norm": 0.9817560346020463, "learning_rate": 7.102511999900213e-06, "loss": 1.0293, "step": 1593 }, { "epoch": 2.1311036789297657, "grad_norm": 1.0037563173395745, "learning_rate": 7.098277068005012e-06, "loss": 1.2248, "step": 1594 }, { "epoch": 2.1324414715719064, "grad_norm": 0.8918184341308957, "learning_rate": 7.094040308286023e-06, "loss": 1.29, "step": 1595 }, { "epoch": 2.1337792642140467, "grad_norm": 0.9935845434148353, "learning_rate": 7.089801724433918e-06, "loss": 0.9344, "step": 1596 }, { "epoch": 2.1351170568561875, "grad_norm": 0.8443424150716751, "learning_rate": 7.085561320140958e-06, "loss": 0.8446, "step": 1597 }, { "epoch": 2.1364548494983278, "grad_norm": 0.990710990487738, "learning_rate": 7.081319099100986e-06, "loss": 1.2576, "step": 1598 }, { "epoch": 2.137792642140468, "grad_norm": 0.9357968882403901, "learning_rate": 7.0770750650094335e-06, "loss": 1.1033, "step": 1599 }, { "epoch": 2.139130434782609, "grad_norm": 1.198335023423348, "learning_rate": 7.072829221563305e-06, "loss": 1.0207, "step": 1600 }, { "epoch": 2.140468227424749, "grad_norm": 0.9660427465710644, "learning_rate": 7.068581572461188e-06, "loss": 0.964, "step": 1601 }, { "epoch": 2.14180602006689, "grad_norm": 0.9788042428106781, "learning_rate": 7.064332121403237e-06, "loss": 1.1498, "step": 1602 }, { "epoch": 2.14314381270903, "grad_norm": 0.8804350134885616, "learning_rate": 7.060080872091178e-06, "loss": 1.1632, "step": 1603 }, { "epoch": 2.1444816053511704, "grad_norm": 0.9082035643442317, "learning_rate": 7.055827828228304e-06, "loss": 0.9953, "step": 1604 }, { "epoch": 2.145819397993311, "grad_norm": 0.8991671470504425, "learning_rate": 7.051572993519474e-06, "loss": 1.0027, "step": 1605 }, { "epoch": 2.1471571906354514, "grad_norm": 1.245321244401268, "learning_rate": 7.0473163716711004e-06, "loss": 1.0669, "step": 1606 }, { "epoch": 2.148494983277592, "grad_norm": 1.1257629456641853, "learning_rate": 7.043057966391158e-06, "loss": 0.9414, "step": 1607 }, { "epoch": 2.1498327759197324, "grad_norm": 0.9563130341058275, "learning_rate": 7.038797781389174e-06, "loss": 0.9271, "step": 1608 }, { "epoch": 2.1511705685618727, "grad_norm": 1.0107338646092578, "learning_rate": 7.034535820376225e-06, "loss": 0.9667, "step": 1609 }, { "epoch": 2.1525083612040135, "grad_norm": 1.0322613586658083, "learning_rate": 7.030272087064933e-06, "loss": 1.0269, "step": 1610 }, { "epoch": 2.1538461538461537, "grad_norm": 1.0151823156319928, "learning_rate": 7.026006585169467e-06, "loss": 0.8121, "step": 1611 }, { "epoch": 2.1551839464882945, "grad_norm": 1.039016196392125, "learning_rate": 7.021739318405537e-06, "loss": 1.0454, "step": 1612 }, { "epoch": 2.1565217391304348, "grad_norm": 1.0967238027546424, "learning_rate": 7.017470290490386e-06, "loss": 1.0978, "step": 1613 }, { "epoch": 2.157859531772575, "grad_norm": 1.0701759010496619, "learning_rate": 7.013199505142796e-06, "loss": 1.1658, "step": 1614 }, { "epoch": 2.159197324414716, "grad_norm": 1.0496336576645666, "learning_rate": 7.008926966083078e-06, "loss": 1.0683, "step": 1615 }, { "epoch": 2.160535117056856, "grad_norm": 1.4355032295972046, "learning_rate": 7.004652677033069e-06, "loss": 0.9017, "step": 1616 }, { "epoch": 2.161872909698997, "grad_norm": 1.0546785006735477, "learning_rate": 7.0003766417161335e-06, "loss": 1.0432, "step": 1617 }, { "epoch": 2.163210702341137, "grad_norm": 0.9267158594340015, "learning_rate": 6.996098863857155e-06, "loss": 1.1516, "step": 1618 }, { "epoch": 2.1645484949832774, "grad_norm": 1.0946972115962772, "learning_rate": 6.991819347182536e-06, "loss": 0.8605, "step": 1619 }, { "epoch": 2.165886287625418, "grad_norm": 1.0229044563408163, "learning_rate": 6.987538095420193e-06, "loss": 1.3349, "step": 1620 }, { "epoch": 2.1672240802675584, "grad_norm": 1.033397386443758, "learning_rate": 6.983255112299554e-06, "loss": 1.2895, "step": 1621 }, { "epoch": 2.168561872909699, "grad_norm": 0.8479210509790241, "learning_rate": 6.978970401551557e-06, "loss": 1.0595, "step": 1622 }, { "epoch": 2.1698996655518394, "grad_norm": 0.926790752177307, "learning_rate": 6.974683966908642e-06, "loss": 1.3309, "step": 1623 }, { "epoch": 2.1712374581939797, "grad_norm": 1.09234429692103, "learning_rate": 6.970395812104751e-06, "loss": 1.1525, "step": 1624 }, { "epoch": 2.1725752508361205, "grad_norm": 0.9125947488373529, "learning_rate": 6.966105940875328e-06, "loss": 0.9653, "step": 1625 }, { "epoch": 2.1739130434782608, "grad_norm": 0.9584785791408102, "learning_rate": 6.961814356957308e-06, "loss": 1.0525, "step": 1626 }, { "epoch": 2.1752508361204015, "grad_norm": 0.950857126604837, "learning_rate": 6.9575210640891215e-06, "loss": 0.9845, "step": 1627 }, { "epoch": 2.1765886287625418, "grad_norm": 1.1068741554761037, "learning_rate": 6.953226066010683e-06, "loss": 1.0577, "step": 1628 }, { "epoch": 2.177926421404682, "grad_norm": 1.1854612737245693, "learning_rate": 6.948929366463397e-06, "loss": 0.9095, "step": 1629 }, { "epoch": 2.179264214046823, "grad_norm": 1.6315940106445606, "learning_rate": 6.944630969190149e-06, "loss": 0.8848, "step": 1630 }, { "epoch": 2.180602006688963, "grad_norm": 1.1332355459827985, "learning_rate": 6.940330877935304e-06, "loss": 0.8604, "step": 1631 }, { "epoch": 2.181939799331104, "grad_norm": 1.1164455103424604, "learning_rate": 6.936029096444697e-06, "loss": 1.1826, "step": 1632 }, { "epoch": 2.183277591973244, "grad_norm": 1.0512510587966684, "learning_rate": 6.931725628465643e-06, "loss": 1.0968, "step": 1633 }, { "epoch": 2.184615384615385, "grad_norm": 1.085936343513785, "learning_rate": 6.927420477746923e-06, "loss": 1.1461, "step": 1634 }, { "epoch": 2.185953177257525, "grad_norm": 1.0484168875099238, "learning_rate": 6.923113648038784e-06, "loss": 0.9299, "step": 1635 }, { "epoch": 2.1872909698996654, "grad_norm": 0.9813639344300368, "learning_rate": 6.918805143092935e-06, "loss": 0.8385, "step": 1636 }, { "epoch": 2.188628762541806, "grad_norm": 1.1184294138886763, "learning_rate": 6.9144949666625434e-06, "loss": 1.0458, "step": 1637 }, { "epoch": 2.1899665551839465, "grad_norm": 1.1159501127084988, "learning_rate": 6.910183122502236e-06, "loss": 0.9269, "step": 1638 }, { "epoch": 2.1913043478260867, "grad_norm": 0.9531368382157029, "learning_rate": 6.9058696143680895e-06, "loss": 0.8517, "step": 1639 }, { "epoch": 2.1926421404682275, "grad_norm": 0.9944254695578043, "learning_rate": 6.9015544460176296e-06, "loss": 1.1225, "step": 1640 }, { "epoch": 2.1939799331103678, "grad_norm": 1.089789800890709, "learning_rate": 6.897237621209831e-06, "loss": 1.0794, "step": 1641 }, { "epoch": 2.1953177257525085, "grad_norm": 0.9850752955142484, "learning_rate": 6.89291914370511e-06, "loss": 1.0846, "step": 1642 }, { "epoch": 2.196655518394649, "grad_norm": 1.1622212945367352, "learning_rate": 6.888599017265321e-06, "loss": 0.9808, "step": 1643 }, { "epoch": 2.1979933110367895, "grad_norm": 0.8693624581546519, "learning_rate": 6.884277245653758e-06, "loss": 0.9668, "step": 1644 }, { "epoch": 2.19933110367893, "grad_norm": 1.1153644361137305, "learning_rate": 6.8799538326351455e-06, "loss": 0.9929, "step": 1645 }, { "epoch": 2.20066889632107, "grad_norm": 0.9623235871111888, "learning_rate": 6.87562878197564e-06, "loss": 1.0034, "step": 1646 }, { "epoch": 2.202006688963211, "grad_norm": 1.1321013145571686, "learning_rate": 6.87130209744282e-06, "loss": 0.9378, "step": 1647 }, { "epoch": 2.203344481605351, "grad_norm": 0.8985842137610206, "learning_rate": 6.866973782805694e-06, "loss": 1.2326, "step": 1648 }, { "epoch": 2.2046822742474914, "grad_norm": 0.8923917864876674, "learning_rate": 6.862643841834686e-06, "loss": 0.9736, "step": 1649 }, { "epoch": 2.206020066889632, "grad_norm": 1.0960131560902968, "learning_rate": 6.858312278301638e-06, "loss": 0.8121, "step": 1650 }, { "epoch": 2.2073578595317724, "grad_norm": 1.0949493952016784, "learning_rate": 6.8539790959798045e-06, "loss": 1.0529, "step": 1651 }, { "epoch": 2.208695652173913, "grad_norm": 0.9738197106288115, "learning_rate": 6.849644298643852e-06, "loss": 1.1681, "step": 1652 }, { "epoch": 2.2100334448160535, "grad_norm": 1.2247699342115956, "learning_rate": 6.845307890069851e-06, "loss": 1.1953, "step": 1653 }, { "epoch": 2.211371237458194, "grad_norm": 0.8718340814181453, "learning_rate": 6.840969874035278e-06, "loss": 1.0641, "step": 1654 }, { "epoch": 2.2127090301003345, "grad_norm": 1.1140233077061839, "learning_rate": 6.83663025431901e-06, "loss": 0.9801, "step": 1655 }, { "epoch": 2.2140468227424748, "grad_norm": 0.9307036230813608, "learning_rate": 6.832289034701318e-06, "loss": 1.2634, "step": 1656 }, { "epoch": 2.2153846153846155, "grad_norm": 0.9335690677874119, "learning_rate": 6.82794621896387e-06, "loss": 0.7687, "step": 1657 }, { "epoch": 2.216722408026756, "grad_norm": 0.8917926395935936, "learning_rate": 6.823601810889723e-06, "loss": 0.9246, "step": 1658 }, { "epoch": 2.218060200668896, "grad_norm": 0.9404152311132599, "learning_rate": 6.8192558142633215e-06, "loss": 1.0063, "step": 1659 }, { "epoch": 2.219397993311037, "grad_norm": 0.9717701087279421, "learning_rate": 6.814908232870493e-06, "loss": 1.1156, "step": 1660 }, { "epoch": 2.220735785953177, "grad_norm": 0.8929197984539806, "learning_rate": 6.810559070498446e-06, "loss": 1.0952, "step": 1661 }, { "epoch": 2.222073578595318, "grad_norm": 0.981423175892025, "learning_rate": 6.806208330935766e-06, "loss": 1.234, "step": 1662 }, { "epoch": 2.223411371237458, "grad_norm": 1.0682396766812956, "learning_rate": 6.801856017972412e-06, "loss": 0.9929, "step": 1663 }, { "epoch": 2.224749163879599, "grad_norm": 1.1482191073384618, "learning_rate": 6.797502135399716e-06, "loss": 1.0558, "step": 1664 }, { "epoch": 2.226086956521739, "grad_norm": 1.0109207400393598, "learning_rate": 6.7931466870103735e-06, "loss": 1.0375, "step": 1665 }, { "epoch": 2.2274247491638794, "grad_norm": 1.2009092908430017, "learning_rate": 6.788789676598449e-06, "loss": 0.9589, "step": 1666 }, { "epoch": 2.22876254180602, "grad_norm": 1.2743687221471387, "learning_rate": 6.78443110795936e-06, "loss": 1.173, "step": 1667 }, { "epoch": 2.2301003344481605, "grad_norm": 1.087781083336, "learning_rate": 6.78007098488989e-06, "loss": 0.9936, "step": 1668 }, { "epoch": 2.231438127090301, "grad_norm": 1.0884988943596365, "learning_rate": 6.77570931118817e-06, "loss": 1.0309, "step": 1669 }, { "epoch": 2.2327759197324415, "grad_norm": 1.1772464556983573, "learning_rate": 6.771346090653687e-06, "loss": 0.9968, "step": 1670 }, { "epoch": 2.234113712374582, "grad_norm": 0.9939696297827828, "learning_rate": 6.766981327087271e-06, "loss": 1.0763, "step": 1671 }, { "epoch": 2.2354515050167225, "grad_norm": 0.8770951447151907, "learning_rate": 6.762615024291098e-06, "loss": 0.9335, "step": 1672 }, { "epoch": 2.236789297658863, "grad_norm": 0.8161380646640386, "learning_rate": 6.758247186068684e-06, "loss": 1.1425, "step": 1673 }, { "epoch": 2.2381270903010035, "grad_norm": 1.1393387347592638, "learning_rate": 6.753877816224886e-06, "loss": 1.0891, "step": 1674 }, { "epoch": 2.239464882943144, "grad_norm": 0.9079858711513, "learning_rate": 6.749506918565891e-06, "loss": 0.9867, "step": 1675 }, { "epoch": 2.240802675585284, "grad_norm": 0.930209830624639, "learning_rate": 6.7451344968992184e-06, "loss": 1.1791, "step": 1676 }, { "epoch": 2.242140468227425, "grad_norm": 0.8883247836902459, "learning_rate": 6.740760555033715e-06, "loss": 0.9948, "step": 1677 }, { "epoch": 2.243478260869565, "grad_norm": 0.8964088738092246, "learning_rate": 6.736385096779552e-06, "loss": 1.1841, "step": 1678 }, { "epoch": 2.244816053511706, "grad_norm": 0.8049486051306619, "learning_rate": 6.732008125948223e-06, "loss": 0.8352, "step": 1679 }, { "epoch": 2.246153846153846, "grad_norm": 0.8920255198634045, "learning_rate": 6.727629646352536e-06, "loss": 1.1618, "step": 1680 }, { "epoch": 2.2474916387959865, "grad_norm": 0.8346549072581853, "learning_rate": 6.723249661806617e-06, "loss": 1.1592, "step": 1681 }, { "epoch": 2.248829431438127, "grad_norm": 0.9452326400550823, "learning_rate": 6.718868176125899e-06, "loss": 1.0405, "step": 1682 }, { "epoch": 2.2501672240802675, "grad_norm": 0.9070563072251671, "learning_rate": 6.714485193127126e-06, "loss": 1.0527, "step": 1683 }, { "epoch": 2.251505016722408, "grad_norm": 0.8995024391445783, "learning_rate": 6.710100716628345e-06, "loss": 0.8433, "step": 1684 }, { "epoch": 2.2528428093645485, "grad_norm": 0.8856513662655207, "learning_rate": 6.705714750448904e-06, "loss": 0.9342, "step": 1685 }, { "epoch": 2.254180602006689, "grad_norm": 0.9467722236751861, "learning_rate": 6.701327298409448e-06, "loss": 1.2602, "step": 1686 }, { "epoch": 2.2555183946488295, "grad_norm": 1.1180095078348111, "learning_rate": 6.6969383643319175e-06, "loss": 1.245, "step": 1687 }, { "epoch": 2.25685618729097, "grad_norm": 0.9409231207847799, "learning_rate": 6.692547952039543e-06, "loss": 1.105, "step": 1688 }, { "epoch": 2.2581939799331106, "grad_norm": 1.1599883338766073, "learning_rate": 6.688156065356845e-06, "loss": 0.9702, "step": 1689 }, { "epoch": 2.259531772575251, "grad_norm": 1.00690798679603, "learning_rate": 6.683762708109625e-06, "loss": 1.2723, "step": 1690 }, { "epoch": 2.260869565217391, "grad_norm": 0.8877604720515682, "learning_rate": 6.679367884124968e-06, "loss": 1.1652, "step": 1691 }, { "epoch": 2.262207357859532, "grad_norm": 0.9404568013917908, "learning_rate": 6.674971597231236e-06, "loss": 1.352, "step": 1692 }, { "epoch": 2.263545150501672, "grad_norm": 1.3002118900072408, "learning_rate": 6.670573851258063e-06, "loss": 0.8027, "step": 1693 }, { "epoch": 2.264882943143813, "grad_norm": 1.1778131678747588, "learning_rate": 6.66617465003636e-06, "loss": 1.0843, "step": 1694 }, { "epoch": 2.266220735785953, "grad_norm": 1.0939185403265157, "learning_rate": 6.6617739973982985e-06, "loss": 0.8548, "step": 1695 }, { "epoch": 2.2675585284280935, "grad_norm": 1.0980237585506671, "learning_rate": 6.6573718971773204e-06, "loss": 1.0047, "step": 1696 }, { "epoch": 2.268896321070234, "grad_norm": 1.0205374823634772, "learning_rate": 6.652968353208122e-06, "loss": 1.0511, "step": 1697 }, { "epoch": 2.2702341137123745, "grad_norm": 0.9943152871486743, "learning_rate": 6.648563369326666e-06, "loss": 1.1437, "step": 1698 }, { "epoch": 2.2715719063545152, "grad_norm": 0.986555128752096, "learning_rate": 6.644156949370162e-06, "loss": 0.956, "step": 1699 }, { "epoch": 2.2729096989966555, "grad_norm": 0.8887478432369301, "learning_rate": 6.639749097177073e-06, "loss": 1.0844, "step": 1700 }, { "epoch": 2.274247491638796, "grad_norm": 0.9741893739866891, "learning_rate": 6.635339816587109e-06, "loss": 1.049, "step": 1701 }, { "epoch": 2.2755852842809365, "grad_norm": 0.9369498053232155, "learning_rate": 6.630929111441227e-06, "loss": 0.9595, "step": 1702 }, { "epoch": 2.276923076923077, "grad_norm": 0.8313489462670471, "learning_rate": 6.626516985581621e-06, "loss": 1.1659, "step": 1703 }, { "epoch": 2.2782608695652176, "grad_norm": 0.9446120980514827, "learning_rate": 6.622103442851728e-06, "loss": 1.0167, "step": 1704 }, { "epoch": 2.279598662207358, "grad_norm": 1.0446212128414376, "learning_rate": 6.617688487096213e-06, "loss": 1.1366, "step": 1705 }, { "epoch": 2.280936454849498, "grad_norm": 0.8793003723136374, "learning_rate": 6.613272122160975e-06, "loss": 0.9764, "step": 1706 }, { "epoch": 2.282274247491639, "grad_norm": 0.9134144896063494, "learning_rate": 6.60885435189314e-06, "loss": 1.0232, "step": 1707 }, { "epoch": 2.283612040133779, "grad_norm": 1.0270309577748888, "learning_rate": 6.60443518014106e-06, "loss": 0.9097, "step": 1708 }, { "epoch": 2.28494983277592, "grad_norm": 1.381517509756264, "learning_rate": 6.600014610754306e-06, "loss": 0.7478, "step": 1709 }, { "epoch": 2.28628762541806, "grad_norm": 0.9595026437833496, "learning_rate": 6.595592647583666e-06, "loss": 1.0483, "step": 1710 }, { "epoch": 2.2876254180602005, "grad_norm": 0.8724954936872927, "learning_rate": 6.591169294481143e-06, "loss": 1.2282, "step": 1711 }, { "epoch": 2.288963210702341, "grad_norm": 0.9636645673465111, "learning_rate": 6.586744555299953e-06, "loss": 1.2578, "step": 1712 }, { "epoch": 2.2903010033444815, "grad_norm": 1.0155768073535967, "learning_rate": 6.582318433894513e-06, "loss": 1.3243, "step": 1713 }, { "epoch": 2.2916387959866222, "grad_norm": 0.9663889665924534, "learning_rate": 6.577890934120451e-06, "loss": 0.9565, "step": 1714 }, { "epoch": 2.2929765886287625, "grad_norm": 0.9914914783102737, "learning_rate": 6.573462059834593e-06, "loss": 1.1519, "step": 1715 }, { "epoch": 2.294314381270903, "grad_norm": 0.9550362616485057, "learning_rate": 6.569031814894962e-06, "loss": 1.2136, "step": 1716 }, { "epoch": 2.2956521739130435, "grad_norm": 1.1296225589936086, "learning_rate": 6.5646002031607726e-06, "loss": 0.8848, "step": 1717 }, { "epoch": 2.296989966555184, "grad_norm": 0.825946409335774, "learning_rate": 6.560167228492436e-06, "loss": 0.8905, "step": 1718 }, { "epoch": 2.2983277591973246, "grad_norm": 0.877387246456996, "learning_rate": 6.555732894751548e-06, "loss": 0.8906, "step": 1719 }, { "epoch": 2.299665551839465, "grad_norm": 0.9363721152314193, "learning_rate": 6.551297205800884e-06, "loss": 1.152, "step": 1720 }, { "epoch": 2.3010033444816056, "grad_norm": 1.1096076660957175, "learning_rate": 6.546860165504406e-06, "loss": 1.2146, "step": 1721 }, { "epoch": 2.302341137123746, "grad_norm": 0.8580943756373745, "learning_rate": 6.5424217777272506e-06, "loss": 0.8623, "step": 1722 }, { "epoch": 2.303678929765886, "grad_norm": 1.1105914132324963, "learning_rate": 6.537982046335727e-06, "loss": 1.0021, "step": 1723 }, { "epoch": 2.305016722408027, "grad_norm": 0.8952603940735279, "learning_rate": 6.533540975197319e-06, "loss": 0.8653, "step": 1724 }, { "epoch": 2.306354515050167, "grad_norm": 0.9733532297356293, "learning_rate": 6.529098568180672e-06, "loss": 1.1669, "step": 1725 }, { "epoch": 2.3076923076923075, "grad_norm": 1.0508833423024984, "learning_rate": 6.524654829155599e-06, "loss": 1.1663, "step": 1726 }, { "epoch": 2.309030100334448, "grad_norm": 1.2151431835983468, "learning_rate": 6.520209761993072e-06, "loss": 0.9077, "step": 1727 }, { "epoch": 2.3103678929765885, "grad_norm": 0.925383318526305, "learning_rate": 6.515763370565218e-06, "loss": 0.9406, "step": 1728 }, { "epoch": 2.3117056856187292, "grad_norm": 0.9119232418185506, "learning_rate": 6.511315658745323e-06, "loss": 1.0873, "step": 1729 }, { "epoch": 2.3130434782608695, "grad_norm": 1.232086296069661, "learning_rate": 6.506866630407817e-06, "loss": 1.2099, "step": 1730 }, { "epoch": 2.3143812709030103, "grad_norm": 1.172988317220349, "learning_rate": 6.502416289428282e-06, "loss": 1.0942, "step": 1731 }, { "epoch": 2.3157190635451506, "grad_norm": 0.8582508819075977, "learning_rate": 6.4979646396834375e-06, "loss": 1.1078, "step": 1732 }, { "epoch": 2.317056856187291, "grad_norm": 0.8209150045374413, "learning_rate": 6.4935116850511495e-06, "loss": 1.1266, "step": 1733 }, { "epoch": 2.3183946488294316, "grad_norm": 1.0993105605806752, "learning_rate": 6.489057429410418e-06, "loss": 1.0566, "step": 1734 }, { "epoch": 2.319732441471572, "grad_norm": 0.8909898824752885, "learning_rate": 6.484601876641375e-06, "loss": 1.135, "step": 1735 }, { "epoch": 2.321070234113712, "grad_norm": 0.9631005704568818, "learning_rate": 6.480145030625284e-06, "loss": 1.0337, "step": 1736 }, { "epoch": 2.322408026755853, "grad_norm": 1.2234854150390555, "learning_rate": 6.475686895244534e-06, "loss": 1.0968, "step": 1737 }, { "epoch": 2.323745819397993, "grad_norm": 1.0277084072502627, "learning_rate": 6.471227474382639e-06, "loss": 0.9851, "step": 1738 }, { "epoch": 2.325083612040134, "grad_norm": 1.0053029411723076, "learning_rate": 6.466766771924231e-06, "loss": 1.0284, "step": 1739 }, { "epoch": 2.326421404682274, "grad_norm": 1.2852260471865329, "learning_rate": 6.462304791755059e-06, "loss": 1.067, "step": 1740 }, { "epoch": 2.327759197324415, "grad_norm": 0.879962292102815, "learning_rate": 6.457841537761985e-06, "loss": 1.1757, "step": 1741 }, { "epoch": 2.3290969899665552, "grad_norm": 1.0428342236935424, "learning_rate": 6.453377013832981e-06, "loss": 0.8825, "step": 1742 }, { "epoch": 2.3304347826086955, "grad_norm": 0.9241082811296258, "learning_rate": 6.448911223857124e-06, "loss": 1.2167, "step": 1743 }, { "epoch": 2.3317725752508363, "grad_norm": 0.9176123079361356, "learning_rate": 6.444444171724595e-06, "loss": 1.0225, "step": 1744 }, { "epoch": 2.3331103678929765, "grad_norm": 1.0523884379910635, "learning_rate": 6.4399758613266775e-06, "loss": 1.0784, "step": 1745 }, { "epoch": 2.334448160535117, "grad_norm": 0.9319141982681433, "learning_rate": 6.435506296555742e-06, "loss": 0.9225, "step": 1746 }, { "epoch": 2.3357859531772576, "grad_norm": 0.8831747582975312, "learning_rate": 6.431035481305261e-06, "loss": 0.9665, "step": 1747 }, { "epoch": 2.337123745819398, "grad_norm": 1.2323574383748228, "learning_rate": 6.426563419469793e-06, "loss": 1.1224, "step": 1748 }, { "epoch": 2.3384615384615386, "grad_norm": 0.9305889476084916, "learning_rate": 6.422090114944982e-06, "loss": 1.0748, "step": 1749 }, { "epoch": 2.339799331103679, "grad_norm": 1.4285770170980012, "learning_rate": 6.417615571627555e-06, "loss": 1.0111, "step": 1750 }, { "epoch": 2.3411371237458196, "grad_norm": 1.2747758059944416, "learning_rate": 6.4131397934153175e-06, "loss": 1.1446, "step": 1751 }, { "epoch": 2.34247491638796, "grad_norm": 1.0309494035530535, "learning_rate": 6.408662784207149e-06, "loss": 1.1612, "step": 1752 }, { "epoch": 2.3438127090301, "grad_norm": 1.115000536809722, "learning_rate": 6.404184547903006e-06, "loss": 1.1119, "step": 1753 }, { "epoch": 2.345150501672241, "grad_norm": 0.8496108880067366, "learning_rate": 6.399705088403912e-06, "loss": 0.9741, "step": 1754 }, { "epoch": 2.346488294314381, "grad_norm": 0.9983096855639143, "learning_rate": 6.3952244096119535e-06, "loss": 1.0468, "step": 1755 }, { "epoch": 2.3478260869565215, "grad_norm": 0.8577291671075368, "learning_rate": 6.3907425154302815e-06, "loss": 1.0682, "step": 1756 }, { "epoch": 2.3491638795986622, "grad_norm": 0.9632537603881958, "learning_rate": 6.386259409763107e-06, "loss": 0.9658, "step": 1757 }, { "epoch": 2.3505016722408025, "grad_norm": 0.9196271090769537, "learning_rate": 6.381775096515692e-06, "loss": 0.9769, "step": 1758 }, { "epoch": 2.3518394648829433, "grad_norm": 0.8637469322102747, "learning_rate": 6.377289579594355e-06, "loss": 0.9658, "step": 1759 }, { "epoch": 2.3531772575250836, "grad_norm": 1.0191868222925913, "learning_rate": 6.372802862906459e-06, "loss": 1.2753, "step": 1760 }, { "epoch": 2.3545150501672243, "grad_norm": 1.0337923242846507, "learning_rate": 6.368314950360416e-06, "loss": 1.1736, "step": 1761 }, { "epoch": 2.3558528428093646, "grad_norm": 0.9402932235561753, "learning_rate": 6.3638258458656766e-06, "loss": 0.8375, "step": 1762 }, { "epoch": 2.357190635451505, "grad_norm": 0.9150755431684604, "learning_rate": 6.3593355533327314e-06, "loss": 1.1089, "step": 1763 }, { "epoch": 2.3585284280936456, "grad_norm": 0.9991567466958132, "learning_rate": 6.354844076673108e-06, "loss": 0.8381, "step": 1764 }, { "epoch": 2.359866220735786, "grad_norm": 0.9745455814216744, "learning_rate": 6.35035141979936e-06, "loss": 1.3054, "step": 1765 }, { "epoch": 2.361204013377926, "grad_norm": 1.1260054385031413, "learning_rate": 6.345857586625073e-06, "loss": 0.9744, "step": 1766 }, { "epoch": 2.362541806020067, "grad_norm": 0.9053381989995765, "learning_rate": 6.341362581064856e-06, "loss": 0.8853, "step": 1767 }, { "epoch": 2.363879598662207, "grad_norm": 0.9009027154294925, "learning_rate": 6.336866407034341e-06, "loss": 1.0145, "step": 1768 }, { "epoch": 2.365217391304348, "grad_norm": 0.896851363663586, "learning_rate": 6.332369068450175e-06, "loss": 1.0335, "step": 1769 }, { "epoch": 2.3665551839464882, "grad_norm": 0.8159188205575251, "learning_rate": 6.327870569230022e-06, "loss": 0.832, "step": 1770 }, { "epoch": 2.367892976588629, "grad_norm": 0.9863759273403112, "learning_rate": 6.323370913292557e-06, "loss": 0.9623, "step": 1771 }, { "epoch": 2.3692307692307693, "grad_norm": 0.8623896783547793, "learning_rate": 6.318870104557459e-06, "loss": 1.0823, "step": 1772 }, { "epoch": 2.3705685618729095, "grad_norm": 1.12218801496878, "learning_rate": 6.314368146945418e-06, "loss": 1.2916, "step": 1773 }, { "epoch": 2.3719063545150503, "grad_norm": 1.151261892738765, "learning_rate": 6.309865044378115e-06, "loss": 1.1095, "step": 1774 }, { "epoch": 2.3732441471571906, "grad_norm": 1.30056607420811, "learning_rate": 6.3053608007782385e-06, "loss": 1.0596, "step": 1775 }, { "epoch": 2.374581939799331, "grad_norm": 0.9345595499445907, "learning_rate": 6.300855420069465e-06, "loss": 0.9365, "step": 1776 }, { "epoch": 2.3759197324414716, "grad_norm": 0.9061814632303258, "learning_rate": 6.296348906176462e-06, "loss": 0.9026, "step": 1777 }, { "epoch": 2.377257525083612, "grad_norm": 1.527868429212941, "learning_rate": 6.2918412630248874e-06, "loss": 0.9518, "step": 1778 }, { "epoch": 2.3785953177257526, "grad_norm": 0.9741388057494006, "learning_rate": 6.28733249454138e-06, "loss": 1.084, "step": 1779 }, { "epoch": 2.379933110367893, "grad_norm": 1.0655724750908069, "learning_rate": 6.2828226046535575e-06, "loss": 1.3668, "step": 1780 }, { "epoch": 2.3812709030100336, "grad_norm": 0.9360204690734704, "learning_rate": 6.278311597290019e-06, "loss": 1.2028, "step": 1781 }, { "epoch": 2.382608695652174, "grad_norm": 0.886286393778401, "learning_rate": 6.273799476380332e-06, "loss": 1.2127, "step": 1782 }, { "epoch": 2.383946488294314, "grad_norm": 1.1406292346920468, "learning_rate": 6.269286245855039e-06, "loss": 0.8765, "step": 1783 }, { "epoch": 2.385284280936455, "grad_norm": 0.9803635088990008, "learning_rate": 6.264771909645646e-06, "loss": 0.8658, "step": 1784 }, { "epoch": 2.3866220735785952, "grad_norm": 0.858596389848438, "learning_rate": 6.260256471684622e-06, "loss": 1.2508, "step": 1785 }, { "epoch": 2.387959866220736, "grad_norm": 1.617353472162379, "learning_rate": 6.255739935905396e-06, "loss": 1.0221, "step": 1786 }, { "epoch": 2.3892976588628763, "grad_norm": 0.9262867405353749, "learning_rate": 6.2512223062423545e-06, "loss": 1.1425, "step": 1787 }, { "epoch": 2.3906354515050166, "grad_norm": 0.9789158942759759, "learning_rate": 6.246703586630838e-06, "loss": 1.0076, "step": 1788 }, { "epoch": 2.3919732441471573, "grad_norm": 0.9063303363007783, "learning_rate": 6.242183781007132e-06, "loss": 1.1954, "step": 1789 }, { "epoch": 2.3933110367892976, "grad_norm": 0.9916241900305185, "learning_rate": 6.237662893308471e-06, "loss": 1.1721, "step": 1790 }, { "epoch": 2.3946488294314383, "grad_norm": 1.2496673112557866, "learning_rate": 6.233140927473033e-06, "loss": 1.0905, "step": 1791 }, { "epoch": 2.3959866220735786, "grad_norm": 0.9080105398781408, "learning_rate": 6.228617887439931e-06, "loss": 1.2302, "step": 1792 }, { "epoch": 2.397324414715719, "grad_norm": 1.1218463084624217, "learning_rate": 6.224093777149222e-06, "loss": 1.2796, "step": 1793 }, { "epoch": 2.3986622073578596, "grad_norm": 0.986803492371461, "learning_rate": 6.219568600541886e-06, "loss": 1.0635, "step": 1794 }, { "epoch": 2.4, "grad_norm": 1.0357531843811614, "learning_rate": 6.2150423615598376e-06, "loss": 0.9532, "step": 1795 }, { "epoch": 2.4013377926421406, "grad_norm": 0.9359289129192679, "learning_rate": 6.210515064145915e-06, "loss": 1.0594, "step": 1796 }, { "epoch": 2.402675585284281, "grad_norm": 0.9297797400323677, "learning_rate": 6.205986712243876e-06, "loss": 1.0915, "step": 1797 }, { "epoch": 2.4040133779264212, "grad_norm": 1.2010579395213947, "learning_rate": 6.201457309798403e-06, "loss": 1.0539, "step": 1798 }, { "epoch": 2.405351170568562, "grad_norm": 0.9234978122300295, "learning_rate": 6.196926860755088e-06, "loss": 1.064, "step": 1799 }, { "epoch": 2.4066889632107022, "grad_norm": 0.8928244271533496, "learning_rate": 6.192395369060439e-06, "loss": 0.9472, "step": 1800 }, { "epoch": 2.408026755852843, "grad_norm": 0.8904500004481711, "learning_rate": 6.187862838661869e-06, "loss": 1.1723, "step": 1801 }, { "epoch": 2.4093645484949833, "grad_norm": 1.0151585065538982, "learning_rate": 6.183329273507693e-06, "loss": 1.0384, "step": 1802 }, { "epoch": 2.4107023411371236, "grad_norm": 0.9169390210145765, "learning_rate": 6.178794677547138e-06, "loss": 1.0065, "step": 1803 }, { "epoch": 2.4120401337792643, "grad_norm": 1.1431816319113761, "learning_rate": 6.174259054730316e-06, "loss": 1.2818, "step": 1804 }, { "epoch": 2.4133779264214046, "grad_norm": 1.0574249444630548, "learning_rate": 6.169722409008244e-06, "loss": 1.109, "step": 1805 }, { "epoch": 2.4147157190635453, "grad_norm": 1.0204656927160234, "learning_rate": 6.165184744332824e-06, "loss": 1.1323, "step": 1806 }, { "epoch": 2.4160535117056856, "grad_norm": 0.9297374226272005, "learning_rate": 6.160646064656845e-06, "loss": 1.2739, "step": 1807 }, { "epoch": 2.417391304347826, "grad_norm": 0.9611381561627882, "learning_rate": 6.156106373933988e-06, "loss": 0.9829, "step": 1808 }, { "epoch": 2.4187290969899666, "grad_norm": 0.8670510601466382, "learning_rate": 6.151565676118805e-06, "loss": 1.0201, "step": 1809 }, { "epoch": 2.420066889632107, "grad_norm": 1.0622989623826509, "learning_rate": 6.147023975166731e-06, "loss": 0.9948, "step": 1810 }, { "epoch": 2.4214046822742477, "grad_norm": 0.9562290000039385, "learning_rate": 6.142481275034072e-06, "loss": 1.0164, "step": 1811 }, { "epoch": 2.422742474916388, "grad_norm": 1.136732573748159, "learning_rate": 6.137937579678007e-06, "loss": 0.8463, "step": 1812 }, { "epoch": 2.4240802675585282, "grad_norm": 0.9006387465415493, "learning_rate": 6.133392893056583e-06, "loss": 1.0157, "step": 1813 }, { "epoch": 2.425418060200669, "grad_norm": 0.9469140120639599, "learning_rate": 6.128847219128703e-06, "loss": 1.0597, "step": 1814 }, { "epoch": 2.4267558528428093, "grad_norm": 0.9619954264466929, "learning_rate": 6.124300561854139e-06, "loss": 1.1195, "step": 1815 }, { "epoch": 2.42809364548495, "grad_norm": 1.3710636250299848, "learning_rate": 6.119752925193516e-06, "loss": 0.9452, "step": 1816 }, { "epoch": 2.4294314381270903, "grad_norm": 1.052477160748896, "learning_rate": 6.1152043131083095e-06, "loss": 0.7988, "step": 1817 }, { "epoch": 2.430769230769231, "grad_norm": 0.8503676588926308, "learning_rate": 6.1106547295608495e-06, "loss": 0.9992, "step": 1818 }, { "epoch": 2.4321070234113713, "grad_norm": 0.8486614382567056, "learning_rate": 6.106104178514309e-06, "loss": 0.955, "step": 1819 }, { "epoch": 2.4334448160535116, "grad_norm": 1.279160801052509, "learning_rate": 6.101552663932704e-06, "loss": 1.145, "step": 1820 }, { "epoch": 2.4347826086956523, "grad_norm": 1.2135526744763765, "learning_rate": 6.097000189780893e-06, "loss": 0.8207, "step": 1821 }, { "epoch": 2.4361204013377926, "grad_norm": 1.1170225855849358, "learning_rate": 6.092446760024564e-06, "loss": 0.8541, "step": 1822 }, { "epoch": 2.437458193979933, "grad_norm": 1.2913635839228608, "learning_rate": 6.087892378630245e-06, "loss": 1.0664, "step": 1823 }, { "epoch": 2.4387959866220736, "grad_norm": 1.045247194213941, "learning_rate": 6.0833370495652885e-06, "loss": 1.1295, "step": 1824 }, { "epoch": 2.440133779264214, "grad_norm": 0.8909762851004575, "learning_rate": 6.0787807767978736e-06, "loss": 0.9059, "step": 1825 }, { "epoch": 2.4414715719063547, "grad_norm": 1.0706454636609757, "learning_rate": 6.074223564296999e-06, "loss": 1.0788, "step": 1826 }, { "epoch": 2.442809364548495, "grad_norm": 0.792961373037822, "learning_rate": 6.0696654160324875e-06, "loss": 1.0257, "step": 1827 }, { "epoch": 2.4441471571906357, "grad_norm": 0.8967726443972275, "learning_rate": 6.065106335974972e-06, "loss": 1.1203, "step": 1828 }, { "epoch": 2.445484949832776, "grad_norm": 0.9910215813763193, "learning_rate": 6.0605463280958995e-06, "loss": 1.1554, "step": 1829 }, { "epoch": 2.4468227424749163, "grad_norm": 1.0095036770406762, "learning_rate": 6.055985396367526e-06, "loss": 1.0632, "step": 1830 }, { "epoch": 2.448160535117057, "grad_norm": 1.1323629554377628, "learning_rate": 6.051423544762909e-06, "loss": 0.9528, "step": 1831 }, { "epoch": 2.4494983277591973, "grad_norm": 0.829915662938606, "learning_rate": 6.046860777255907e-06, "loss": 0.9525, "step": 1832 }, { "epoch": 2.4508361204013376, "grad_norm": 1.0076659753370572, "learning_rate": 6.042297097821184e-06, "loss": 1.1764, "step": 1833 }, { "epoch": 2.4521739130434783, "grad_norm": 1.0654574249298172, "learning_rate": 6.0377325104341885e-06, "loss": 1.1507, "step": 1834 }, { "epoch": 2.4535117056856186, "grad_norm": 1.0866785118371032, "learning_rate": 6.033167019071168e-06, "loss": 0.9635, "step": 1835 }, { "epoch": 2.4548494983277593, "grad_norm": 1.0679064480554037, "learning_rate": 6.028600627709151e-06, "loss": 1.1319, "step": 1836 }, { "epoch": 2.4561872909698996, "grad_norm": 1.0470508136341101, "learning_rate": 6.024033340325954e-06, "loss": 1.0843, "step": 1837 }, { "epoch": 2.4575250836120404, "grad_norm": 0.9238506421156313, "learning_rate": 6.019465160900173e-06, "loss": 1.0109, "step": 1838 }, { "epoch": 2.4588628762541807, "grad_norm": 1.1584881278378014, "learning_rate": 6.014896093411181e-06, "loss": 1.0489, "step": 1839 }, { "epoch": 2.460200668896321, "grad_norm": 1.2287531004801386, "learning_rate": 6.010326141839125e-06, "loss": 1.0262, "step": 1840 }, { "epoch": 2.4615384615384617, "grad_norm": 1.0224224437613938, "learning_rate": 6.005755310164919e-06, "loss": 1.0786, "step": 1841 }, { "epoch": 2.462876254180602, "grad_norm": 1.1727952550900722, "learning_rate": 6.001183602370249e-06, "loss": 1.2537, "step": 1842 }, { "epoch": 2.4642140468227423, "grad_norm": 0.8421844454487292, "learning_rate": 5.996611022437562e-06, "loss": 1.013, "step": 1843 }, { "epoch": 2.465551839464883, "grad_norm": 1.0983948585384216, "learning_rate": 5.992037574350062e-06, "loss": 1.184, "step": 1844 }, { "epoch": 2.4668896321070233, "grad_norm": 1.028360451187395, "learning_rate": 5.987463262091715e-06, "loss": 1.1036, "step": 1845 }, { "epoch": 2.468227424749164, "grad_norm": 1.4828947615356682, "learning_rate": 5.982888089647232e-06, "loss": 1.2192, "step": 1846 }, { "epoch": 2.4695652173913043, "grad_norm": 0.8949037039743718, "learning_rate": 5.978312061002078e-06, "loss": 1.2335, "step": 1847 }, { "epoch": 2.470903010033445, "grad_norm": 1.1164804683013325, "learning_rate": 5.973735180142468e-06, "loss": 1.016, "step": 1848 }, { "epoch": 2.4722408026755853, "grad_norm": 1.108832068690222, "learning_rate": 5.9691574510553505e-06, "loss": 1.1468, "step": 1849 }, { "epoch": 2.4735785953177256, "grad_norm": 0.9485220506936197, "learning_rate": 5.9645788777284195e-06, "loss": 1.0618, "step": 1850 }, { "epoch": 2.4749163879598663, "grad_norm": 1.20026141650198, "learning_rate": 5.959999464150101e-06, "loss": 0.8262, "step": 1851 }, { "epoch": 2.4762541806020066, "grad_norm": 0.9685042930884532, "learning_rate": 5.9554192143095535e-06, "loss": 1.034, "step": 1852 }, { "epoch": 2.477591973244147, "grad_norm": 1.0069039615944846, "learning_rate": 5.950838132196667e-06, "loss": 0.8872, "step": 1853 }, { "epoch": 2.4789297658862877, "grad_norm": 1.2457715131088862, "learning_rate": 5.946256221802052e-06, "loss": 0.8776, "step": 1854 }, { "epoch": 2.480267558528428, "grad_norm": 1.3344581892161573, "learning_rate": 5.941673487117043e-06, "loss": 0.8913, "step": 1855 }, { "epoch": 2.4816053511705687, "grad_norm": 0.9730755143212169, "learning_rate": 5.937089932133693e-06, "loss": 0.9792, "step": 1856 }, { "epoch": 2.482943143812709, "grad_norm": 0.9685322309693766, "learning_rate": 5.932505560844766e-06, "loss": 1.0704, "step": 1857 }, { "epoch": 2.4842809364548497, "grad_norm": 0.8922808946672333, "learning_rate": 5.927920377243743e-06, "loss": 0.9256, "step": 1858 }, { "epoch": 2.48561872909699, "grad_norm": 0.9131196337561178, "learning_rate": 5.923334385324809e-06, "loss": 0.8877, "step": 1859 }, { "epoch": 2.4869565217391303, "grad_norm": 1.238616575908337, "learning_rate": 5.918747589082853e-06, "loss": 1.1629, "step": 1860 }, { "epoch": 2.488294314381271, "grad_norm": 1.2458023063237929, "learning_rate": 5.914159992513464e-06, "loss": 0.9581, "step": 1861 }, { "epoch": 2.4896321070234113, "grad_norm": 1.1373737175365306, "learning_rate": 5.90957159961293e-06, "loss": 1.0327, "step": 1862 }, { "epoch": 2.4909698996655516, "grad_norm": 1.3467860098547657, "learning_rate": 5.904982414378233e-06, "loss": 0.8151, "step": 1863 }, { "epoch": 2.4923076923076923, "grad_norm": 1.023230826140762, "learning_rate": 5.900392440807044e-06, "loss": 1.0694, "step": 1864 }, { "epoch": 2.4936454849498326, "grad_norm": 0.9341880367242312, "learning_rate": 5.895801682897721e-06, "loss": 1.0428, "step": 1865 }, { "epoch": 2.4949832775919734, "grad_norm": 1.259288590460806, "learning_rate": 5.891210144649303e-06, "loss": 0.9245, "step": 1866 }, { "epoch": 2.4963210702341136, "grad_norm": 0.9629860129252424, "learning_rate": 5.886617830061514e-06, "loss": 1.2541, "step": 1867 }, { "epoch": 2.4976588628762544, "grad_norm": 1.0997977303624278, "learning_rate": 5.88202474313475e-06, "loss": 1.0737, "step": 1868 }, { "epoch": 2.4989966555183947, "grad_norm": 1.0092090652822643, "learning_rate": 5.877430887870081e-06, "loss": 1.0514, "step": 1869 }, { "epoch": 2.500334448160535, "grad_norm": 1.0027493653038564, "learning_rate": 5.872836268269246e-06, "loss": 0.9853, "step": 1870 }, { "epoch": 2.5016722408026757, "grad_norm": 1.0135373843771214, "learning_rate": 5.8682408883346535e-06, "loss": 0.7677, "step": 1871 }, { "epoch": 2.503010033444816, "grad_norm": 0.9945858597171933, "learning_rate": 5.863644752069364e-06, "loss": 0.9137, "step": 1872 }, { "epoch": 2.5043478260869563, "grad_norm": 1.3018962781719858, "learning_rate": 5.859047863477112e-06, "loss": 1.0215, "step": 1873 }, { "epoch": 2.505685618729097, "grad_norm": 1.1481256579255743, "learning_rate": 5.854450226562274e-06, "loss": 1.0877, "step": 1874 }, { "epoch": 2.5070234113712373, "grad_norm": 0.8713706321370671, "learning_rate": 5.849851845329884e-06, "loss": 0.7728, "step": 1875 }, { "epoch": 2.508361204013378, "grad_norm": 0.9416946139045507, "learning_rate": 5.845252723785626e-06, "loss": 1.0968, "step": 1876 }, { "epoch": 2.5096989966555183, "grad_norm": 1.016166396982064, "learning_rate": 5.8406528659358234e-06, "loss": 1.2588, "step": 1877 }, { "epoch": 2.511036789297659, "grad_norm": 0.9441751193451301, "learning_rate": 5.836052275787448e-06, "loss": 1.0376, "step": 1878 }, { "epoch": 2.5123745819397993, "grad_norm": 0.8053995876721759, "learning_rate": 5.831450957348106e-06, "loss": 1.0426, "step": 1879 }, { "epoch": 2.5137123745819396, "grad_norm": 0.918627127156628, "learning_rate": 5.826848914626035e-06, "loss": 0.9145, "step": 1880 }, { "epoch": 2.5150501672240804, "grad_norm": 0.8396272705335169, "learning_rate": 5.822246151630109e-06, "loss": 0.7867, "step": 1881 }, { "epoch": 2.5163879598662207, "grad_norm": 1.1084988576493078, "learning_rate": 5.817642672369825e-06, "loss": 0.9727, "step": 1882 }, { "epoch": 2.517725752508361, "grad_norm": 0.8346283338974226, "learning_rate": 5.813038480855308e-06, "loss": 1.1383, "step": 1883 }, { "epoch": 2.5190635451505017, "grad_norm": 0.917736341725873, "learning_rate": 5.808433581097301e-06, "loss": 0.8762, "step": 1884 }, { "epoch": 2.5204013377926424, "grad_norm": 0.9725960027836682, "learning_rate": 5.803827977107163e-06, "loss": 1.364, "step": 1885 }, { "epoch": 2.5217391304347827, "grad_norm": 0.9959421210174695, "learning_rate": 5.799221672896868e-06, "loss": 0.8543, "step": 1886 }, { "epoch": 2.523076923076923, "grad_norm": 0.9280499556607265, "learning_rate": 5.794614672479e-06, "loss": 0.9744, "step": 1887 }, { "epoch": 2.5244147157190637, "grad_norm": 0.9805934851071515, "learning_rate": 5.79000697986675e-06, "loss": 0.9721, "step": 1888 }, { "epoch": 2.525752508361204, "grad_norm": 1.118527271730031, "learning_rate": 5.7853985990739115e-06, "loss": 1.0334, "step": 1889 }, { "epoch": 2.5270903010033443, "grad_norm": 0.8892792935318041, "learning_rate": 5.780789534114875e-06, "loss": 1.1976, "step": 1890 }, { "epoch": 2.528428093645485, "grad_norm": 0.9819059287760276, "learning_rate": 5.77617978900463e-06, "loss": 1.2351, "step": 1891 }, { "epoch": 2.5297658862876253, "grad_norm": 0.8721961433000648, "learning_rate": 5.771569367758757e-06, "loss": 1.0802, "step": 1892 }, { "epoch": 2.5311036789297656, "grad_norm": 0.9157346336429528, "learning_rate": 5.766958274393428e-06, "loss": 1.1282, "step": 1893 }, { "epoch": 2.5324414715719064, "grad_norm": 1.096885777038834, "learning_rate": 5.762346512925397e-06, "loss": 1.0004, "step": 1894 }, { "epoch": 2.533779264214047, "grad_norm": 1.2609239045489362, "learning_rate": 5.757734087372003e-06, "loss": 0.9585, "step": 1895 }, { "epoch": 2.5351170568561874, "grad_norm": 0.8718733117979934, "learning_rate": 5.753121001751161e-06, "loss": 0.9607, "step": 1896 }, { "epoch": 2.5364548494983277, "grad_norm": 1.0289900020716696, "learning_rate": 5.748507260081361e-06, "loss": 1.0082, "step": 1897 }, { "epoch": 2.5377926421404684, "grad_norm": 1.0435059104357125, "learning_rate": 5.743892866381668e-06, "loss": 1.2096, "step": 1898 }, { "epoch": 2.5391304347826087, "grad_norm": 0.821656534242571, "learning_rate": 5.739277824671711e-06, "loss": 1.0517, "step": 1899 }, { "epoch": 2.540468227424749, "grad_norm": 0.803194418905504, "learning_rate": 5.734662138971686e-06, "loss": 1.1375, "step": 1900 }, { "epoch": 2.5418060200668897, "grad_norm": 0.9537980732746415, "learning_rate": 5.730045813302347e-06, "loss": 0.9804, "step": 1901 }, { "epoch": 2.54314381270903, "grad_norm": 0.9121177309165219, "learning_rate": 5.725428851685011e-06, "loss": 1.0512, "step": 1902 }, { "epoch": 2.5444816053511703, "grad_norm": 1.0846188922251343, "learning_rate": 5.720811258141541e-06, "loss": 1.1392, "step": 1903 }, { "epoch": 2.545819397993311, "grad_norm": 0.8122548353151819, "learning_rate": 5.716193036694359e-06, "loss": 0.9435, "step": 1904 }, { "epoch": 2.5471571906354518, "grad_norm": 0.9336890753237073, "learning_rate": 5.711574191366427e-06, "loss": 1.0983, "step": 1905 }, { "epoch": 2.548494983277592, "grad_norm": 1.0144204084231636, "learning_rate": 5.706954726181255e-06, "loss": 0.9105, "step": 1906 }, { "epoch": 2.5498327759197323, "grad_norm": 0.8177336253439044, "learning_rate": 5.70233464516289e-06, "loss": 0.9492, "step": 1907 }, { "epoch": 2.551170568561873, "grad_norm": 0.8192751956211891, "learning_rate": 5.697713952335918e-06, "loss": 0.9396, "step": 1908 }, { "epoch": 2.5525083612040134, "grad_norm": 0.9330571364906027, "learning_rate": 5.693092651725457e-06, "loss": 1.1091, "step": 1909 }, { "epoch": 2.5538461538461537, "grad_norm": 1.1395113135597805, "learning_rate": 5.688470747357153e-06, "loss": 1.0843, "step": 1910 }, { "epoch": 2.5551839464882944, "grad_norm": 0.8809127865438304, "learning_rate": 5.683848243257181e-06, "loss": 1.2292, "step": 1911 }, { "epoch": 2.5565217391304347, "grad_norm": 1.0252197808102461, "learning_rate": 5.679225143452233e-06, "loss": 0.8343, "step": 1912 }, { "epoch": 2.5578595317725754, "grad_norm": 0.9907276223888806, "learning_rate": 5.674601451969527e-06, "loss": 1.0526, "step": 1913 }, { "epoch": 2.5591973244147157, "grad_norm": 0.8876660716986997, "learning_rate": 5.669977172836791e-06, "loss": 0.9109, "step": 1914 }, { "epoch": 2.5605351170568564, "grad_norm": 1.1527217571936268, "learning_rate": 5.66535231008227e-06, "loss": 0.9719, "step": 1915 }, { "epoch": 2.5618729096989967, "grad_norm": 1.0104429729223172, "learning_rate": 5.66072686773471e-06, "loss": 1.1673, "step": 1916 }, { "epoch": 2.563210702341137, "grad_norm": 1.0398661617822784, "learning_rate": 5.656100849823366e-06, "loss": 0.9115, "step": 1917 }, { "epoch": 2.5645484949832777, "grad_norm": 1.0046401732796846, "learning_rate": 5.651474260377998e-06, "loss": 1.2627, "step": 1918 }, { "epoch": 2.565886287625418, "grad_norm": 0.8396226668507198, "learning_rate": 5.646847103428859e-06, "loss": 1.0467, "step": 1919 }, { "epoch": 2.5672240802675583, "grad_norm": 0.8546190467952353, "learning_rate": 5.642219383006696e-06, "loss": 0.7591, "step": 1920 }, { "epoch": 2.568561872909699, "grad_norm": 1.0430153127569541, "learning_rate": 5.63759110314275e-06, "loss": 1.024, "step": 1921 }, { "epoch": 2.5698996655518394, "grad_norm": 0.9694595826620183, "learning_rate": 5.632962267868747e-06, "loss": 1.0599, "step": 1922 }, { "epoch": 2.57123745819398, "grad_norm": 0.935509811055879, "learning_rate": 5.628332881216899e-06, "loss": 1.0919, "step": 1923 }, { "epoch": 2.5725752508361204, "grad_norm": 0.9237883903767292, "learning_rate": 5.623702947219896e-06, "loss": 1.1007, "step": 1924 }, { "epoch": 2.573913043478261, "grad_norm": 1.1096491859711155, "learning_rate": 5.619072469910907e-06, "loss": 0.945, "step": 1925 }, { "epoch": 2.5752508361204014, "grad_norm": 0.9693313771385114, "learning_rate": 5.614441453323571e-06, "loss": 0.9275, "step": 1926 }, { "epoch": 2.5765886287625417, "grad_norm": 0.9425965149552739, "learning_rate": 5.609809901492e-06, "loss": 1.0277, "step": 1927 }, { "epoch": 2.5779264214046824, "grad_norm": 0.9375700547674981, "learning_rate": 5.605177818450772e-06, "loss": 1.237, "step": 1928 }, { "epoch": 2.5792642140468227, "grad_norm": 0.8865601017233428, "learning_rate": 5.600545208234927e-06, "loss": 1.017, "step": 1929 }, { "epoch": 2.580602006688963, "grad_norm": 0.9807087501510431, "learning_rate": 5.595912074879961e-06, "loss": 1.0991, "step": 1930 }, { "epoch": 2.5819397993311037, "grad_norm": 1.360982307133714, "learning_rate": 5.591278422421831e-06, "loss": 1.0345, "step": 1931 }, { "epoch": 2.583277591973244, "grad_norm": 0.9539630116058583, "learning_rate": 5.586644254896945e-06, "loss": 0.7301, "step": 1932 }, { "epoch": 2.5846153846153848, "grad_norm": 0.8796419390770085, "learning_rate": 5.5820095763421565e-06, "loss": 1.0674, "step": 1933 }, { "epoch": 2.585953177257525, "grad_norm": 1.1032640906871372, "learning_rate": 5.5773743907947674e-06, "loss": 1.0765, "step": 1934 }, { "epoch": 2.587290969899666, "grad_norm": 0.8977977262539288, "learning_rate": 5.57273870229252e-06, "loss": 1.1015, "step": 1935 }, { "epoch": 2.588628762541806, "grad_norm": 1.0350008353315647, "learning_rate": 5.568102514873595e-06, "loss": 0.8903, "step": 1936 }, { "epoch": 2.5899665551839464, "grad_norm": 0.7615704827822232, "learning_rate": 5.5634658325766066e-06, "loss": 0.9071, "step": 1937 }, { "epoch": 2.591304347826087, "grad_norm": 1.1085351183158516, "learning_rate": 5.558828659440603e-06, "loss": 0.9379, "step": 1938 }, { "epoch": 2.5926421404682274, "grad_norm": 0.8266047473381335, "learning_rate": 5.5541909995050554e-06, "loss": 1.0832, "step": 1939 }, { "epoch": 2.5939799331103677, "grad_norm": 0.8886832292511988, "learning_rate": 5.549552856809865e-06, "loss": 0.9141, "step": 1940 }, { "epoch": 2.5953177257525084, "grad_norm": 1.0537302600059542, "learning_rate": 5.544914235395347e-06, "loss": 0.6952, "step": 1941 }, { "epoch": 2.5966555183946487, "grad_norm": 1.3062143931434573, "learning_rate": 5.540275139302241e-06, "loss": 0.7625, "step": 1942 }, { "epoch": 2.5979933110367894, "grad_norm": 0.8069165812313676, "learning_rate": 5.53563557257169e-06, "loss": 0.9526, "step": 1943 }, { "epoch": 2.5993311036789297, "grad_norm": 1.0357916137292713, "learning_rate": 5.5309955392452585e-06, "loss": 1.1504, "step": 1944 }, { "epoch": 2.6006688963210705, "grad_norm": 1.1278422997826076, "learning_rate": 5.526355043364909e-06, "loss": 0.8668, "step": 1945 }, { "epoch": 2.6020066889632107, "grad_norm": 0.8423221487274121, "learning_rate": 5.521714088973012e-06, "loss": 0.9747, "step": 1946 }, { "epoch": 2.603344481605351, "grad_norm": 1.0657030228902173, "learning_rate": 5.517072680112332e-06, "loss": 1.0417, "step": 1947 }, { "epoch": 2.6046822742474918, "grad_norm": 0.9908897959063151, "learning_rate": 5.512430820826035e-06, "loss": 1.0121, "step": 1948 }, { "epoch": 2.606020066889632, "grad_norm": 0.8310847137031142, "learning_rate": 5.507788515157677e-06, "loss": 1.1773, "step": 1949 }, { "epoch": 2.6073578595317723, "grad_norm": 1.2545341021863374, "learning_rate": 5.503145767151201e-06, "loss": 1.0918, "step": 1950 }, { "epoch": 2.608695652173913, "grad_norm": 0.9734825616432529, "learning_rate": 5.498502580850938e-06, "loss": 1.0359, "step": 1951 }, { "epoch": 2.6100334448160534, "grad_norm": 1.2153552439003585, "learning_rate": 5.493858960301602e-06, "loss": 0.9598, "step": 1952 }, { "epoch": 2.611371237458194, "grad_norm": 0.8869246637199315, "learning_rate": 5.4892149095482815e-06, "loss": 1.0845, "step": 1953 }, { "epoch": 2.6127090301003344, "grad_norm": 1.0028747593477232, "learning_rate": 5.484570432636441e-06, "loss": 1.0593, "step": 1954 }, { "epoch": 2.614046822742475, "grad_norm": 1.0224966040391446, "learning_rate": 5.479925533611917e-06, "loss": 0.8783, "step": 1955 }, { "epoch": 2.6153846153846154, "grad_norm": 0.9414735279925366, "learning_rate": 5.475280216520913e-06, "loss": 0.9393, "step": 1956 }, { "epoch": 2.6167224080267557, "grad_norm": 0.8853842113573018, "learning_rate": 5.470634485409999e-06, "loss": 0.9031, "step": 1957 }, { "epoch": 2.6180602006688964, "grad_norm": 1.1185341834512839, "learning_rate": 5.465988344326103e-06, "loss": 0.8406, "step": 1958 }, { "epoch": 2.6193979933110367, "grad_norm": 1.2171683437368614, "learning_rate": 5.46134179731651e-06, "loss": 1.1127, "step": 1959 }, { "epoch": 2.620735785953177, "grad_norm": 1.3666922954333647, "learning_rate": 5.456694848428861e-06, "loss": 1.0306, "step": 1960 }, { "epoch": 2.6220735785953178, "grad_norm": 1.1901538460246868, "learning_rate": 5.452047501711144e-06, "loss": 1.1696, "step": 1961 }, { "epoch": 2.623411371237458, "grad_norm": 0.8468798606729907, "learning_rate": 5.4473997612116956e-06, "loss": 1.0776, "step": 1962 }, { "epoch": 2.624749163879599, "grad_norm": 0.991844564056409, "learning_rate": 5.442751630979195e-06, "loss": 1.2522, "step": 1963 }, { "epoch": 2.626086956521739, "grad_norm": 0.8025929585625525, "learning_rate": 5.438103115062662e-06, "loss": 1.0826, "step": 1964 }, { "epoch": 2.62742474916388, "grad_norm": 1.145602932349733, "learning_rate": 5.4334542175114495e-06, "loss": 0.8354, "step": 1965 }, { "epoch": 2.62876254180602, "grad_norm": 1.0859546483299118, "learning_rate": 5.428804942375243e-06, "loss": 1.083, "step": 1966 }, { "epoch": 2.6301003344481604, "grad_norm": 0.9527961941378715, "learning_rate": 5.424155293704063e-06, "loss": 0.9124, "step": 1967 }, { "epoch": 2.631438127090301, "grad_norm": 0.9588114544745916, "learning_rate": 5.419505275548249e-06, "loss": 1.1473, "step": 1968 }, { "epoch": 2.6327759197324414, "grad_norm": 0.8556844303742533, "learning_rate": 5.414854891958464e-06, "loss": 0.7245, "step": 1969 }, { "epoch": 2.6341137123745817, "grad_norm": 1.03902924453866, "learning_rate": 5.41020414698569e-06, "loss": 1.244, "step": 1970 }, { "epoch": 2.6354515050167224, "grad_norm": 1.0923647667446104, "learning_rate": 5.40555304468122e-06, "loss": 1.3032, "step": 1971 }, { "epoch": 2.6367892976588627, "grad_norm": 1.0713563540430706, "learning_rate": 5.400901589096667e-06, "loss": 1.1685, "step": 1972 }, { "epoch": 2.6381270903010035, "grad_norm": 1.0007460264091526, "learning_rate": 5.396249784283943e-06, "loss": 0.8341, "step": 1973 }, { "epoch": 2.6394648829431437, "grad_norm": 0.8554620690299624, "learning_rate": 5.391597634295269e-06, "loss": 0.9919, "step": 1974 }, { "epoch": 2.6408026755852845, "grad_norm": 0.9640354648199924, "learning_rate": 5.386945143183164e-06, "loss": 0.7844, "step": 1975 }, { "epoch": 2.6421404682274248, "grad_norm": 1.0022947434992557, "learning_rate": 5.382292315000448e-06, "loss": 0.9117, "step": 1976 }, { "epoch": 2.643478260869565, "grad_norm": 0.8046864629123232, "learning_rate": 5.377639153800229e-06, "loss": 0.8658, "step": 1977 }, { "epoch": 2.644816053511706, "grad_norm": 1.1324613384512894, "learning_rate": 5.37298566363591e-06, "loss": 1.1285, "step": 1978 }, { "epoch": 2.646153846153846, "grad_norm": 1.1887186657712017, "learning_rate": 5.368331848561178e-06, "loss": 0.885, "step": 1979 }, { "epoch": 2.6474916387959864, "grad_norm": 0.8601318384361885, "learning_rate": 5.363677712630004e-06, "loss": 1.2082, "step": 1980 }, { "epoch": 2.648829431438127, "grad_norm": 1.073808474878058, "learning_rate": 5.359023259896638e-06, "loss": 1.1095, "step": 1981 }, { "epoch": 2.650167224080268, "grad_norm": 1.3131209592506836, "learning_rate": 5.354368494415607e-06, "loss": 1.0159, "step": 1982 }, { "epoch": 2.651505016722408, "grad_norm": 0.8454478488997724, "learning_rate": 5.34971342024171e-06, "loss": 1.1938, "step": 1983 }, { "epoch": 2.6528428093645484, "grad_norm": 1.0350528132142418, "learning_rate": 5.345058041430013e-06, "loss": 1.0649, "step": 1984 }, { "epoch": 2.654180602006689, "grad_norm": 0.8098385968786911, "learning_rate": 5.3404023620358494e-06, "loss": 0.9742, "step": 1985 }, { "epoch": 2.6555183946488294, "grad_norm": 0.7456772008511848, "learning_rate": 5.335746386114814e-06, "loss": 1.0516, "step": 1986 }, { "epoch": 2.6568561872909697, "grad_norm": 1.1519708420071342, "learning_rate": 5.3310901177227615e-06, "loss": 1.0178, "step": 1987 }, { "epoch": 2.6581939799331105, "grad_norm": 0.8082887870326021, "learning_rate": 5.326433560915798e-06, "loss": 0.8056, "step": 1988 }, { "epoch": 2.6595317725752508, "grad_norm": 1.3934857816585744, "learning_rate": 5.321776719750283e-06, "loss": 0.8731, "step": 1989 }, { "epoch": 2.660869565217391, "grad_norm": 0.9044700612166542, "learning_rate": 5.317119598282823e-06, "loss": 0.8245, "step": 1990 }, { "epoch": 2.6622073578595318, "grad_norm": 1.072517564086636, "learning_rate": 5.31246220057027e-06, "loss": 0.9202, "step": 1991 }, { "epoch": 2.6635451505016725, "grad_norm": 0.9483529132066457, "learning_rate": 5.3078045306697154e-06, "loss": 1.0952, "step": 1992 }, { "epoch": 2.664882943143813, "grad_norm": 0.9457345258793196, "learning_rate": 5.303146592638487e-06, "loss": 1.0415, "step": 1993 }, { "epoch": 2.666220735785953, "grad_norm": 1.1170046959469802, "learning_rate": 5.298488390534148e-06, "loss": 1.2689, "step": 1994 }, { "epoch": 2.667558528428094, "grad_norm": 0.9431044925057017, "learning_rate": 5.29382992841449e-06, "loss": 1.0551, "step": 1995 }, { "epoch": 2.668896321070234, "grad_norm": 0.887750669776736, "learning_rate": 5.289171210337531e-06, "loss": 1.095, "step": 1996 }, { "epoch": 2.6702341137123744, "grad_norm": 1.0771814690001333, "learning_rate": 5.284512240361516e-06, "loss": 0.9789, "step": 1997 }, { "epoch": 2.671571906354515, "grad_norm": 0.8607876829939497, "learning_rate": 5.279853022544904e-06, "loss": 1.313, "step": 1998 }, { "epoch": 2.6729096989966554, "grad_norm": 0.9186701384376283, "learning_rate": 5.275193560946372e-06, "loss": 1.03, "step": 1999 }, { "epoch": 2.6742474916387957, "grad_norm": 1.0368433473911216, "learning_rate": 5.27053385962481e-06, "loss": 1.0505, "step": 2000 }, { "epoch": 2.6755852842809364, "grad_norm": 1.0300778044505035, "learning_rate": 5.265873922639315e-06, "loss": 0.8823, "step": 2001 }, { "epoch": 2.676923076923077, "grad_norm": 0.7211640921751338, "learning_rate": 5.261213754049193e-06, "loss": 0.9908, "step": 2002 }, { "epoch": 2.6782608695652175, "grad_norm": 1.4263444500899591, "learning_rate": 5.2565533579139484e-06, "loss": 0.9722, "step": 2003 }, { "epoch": 2.6795986622073578, "grad_norm": 0.8478994107194656, "learning_rate": 5.251892738293285e-06, "loss": 1.0755, "step": 2004 }, { "epoch": 2.6809364548494985, "grad_norm": 1.0147826970829485, "learning_rate": 5.247231899247099e-06, "loss": 0.9632, "step": 2005 }, { "epoch": 2.682274247491639, "grad_norm": 1.0071150044761523, "learning_rate": 5.242570844835484e-06, "loss": 0.8738, "step": 2006 }, { "epoch": 2.683612040133779, "grad_norm": 0.8891990683258876, "learning_rate": 5.237909579118713e-06, "loss": 1.0993, "step": 2007 }, { "epoch": 2.68494983277592, "grad_norm": 1.0306598018862323, "learning_rate": 5.233248106157248e-06, "loss": 0.9385, "step": 2008 }, { "epoch": 2.68628762541806, "grad_norm": 0.9491802943039358, "learning_rate": 5.228586430011732e-06, "loss": 0.8757, "step": 2009 }, { "epoch": 2.687625418060201, "grad_norm": 0.9717254429501214, "learning_rate": 5.223924554742982e-06, "loss": 1.3153, "step": 2010 }, { "epoch": 2.688963210702341, "grad_norm": 0.9114767010895825, "learning_rate": 5.21926248441199e-06, "loss": 1.1032, "step": 2011 }, { "epoch": 2.690301003344482, "grad_norm": 0.7625216819046183, "learning_rate": 5.21460022307992e-06, "loss": 1.0165, "step": 2012 }, { "epoch": 2.691638795986622, "grad_norm": 1.2565329097847, "learning_rate": 5.209937774808098e-06, "loss": 1.1106, "step": 2013 }, { "epoch": 2.6929765886287624, "grad_norm": 0.8035261009787338, "learning_rate": 5.205275143658018e-06, "loss": 0.9146, "step": 2014 }, { "epoch": 2.694314381270903, "grad_norm": 1.054103312504531, "learning_rate": 5.2006123336913275e-06, "loss": 1.2047, "step": 2015 }, { "epoch": 2.6956521739130435, "grad_norm": 1.2400904137241422, "learning_rate": 5.195949348969833e-06, "loss": 0.8325, "step": 2016 }, { "epoch": 2.6969899665551837, "grad_norm": 0.9839927184111734, "learning_rate": 5.191286193555496e-06, "loss": 1.0033, "step": 2017 }, { "epoch": 2.6983277591973245, "grad_norm": 1.3289391435315843, "learning_rate": 5.186622871510421e-06, "loss": 0.7817, "step": 2018 }, { "epoch": 2.6996655518394648, "grad_norm": 0.9539716587817372, "learning_rate": 5.181959386896862e-06, "loss": 0.8488, "step": 2019 }, { "epoch": 2.7010033444816055, "grad_norm": 1.0111233312514996, "learning_rate": 5.177295743777212e-06, "loss": 1.0555, "step": 2020 }, { "epoch": 2.702341137123746, "grad_norm": 1.1638567156170443, "learning_rate": 5.172631946214003e-06, "loss": 1.1047, "step": 2021 }, { "epoch": 2.7036789297658865, "grad_norm": 0.8808394284402741, "learning_rate": 5.167967998269902e-06, "loss": 1.034, "step": 2022 }, { "epoch": 2.705016722408027, "grad_norm": 0.9439351213624038, "learning_rate": 5.1633039040077046e-06, "loss": 1.13, "step": 2023 }, { "epoch": 2.706354515050167, "grad_norm": 0.856314358833399, "learning_rate": 5.15863966749034e-06, "loss": 0.8313, "step": 2024 }, { "epoch": 2.707692307692308, "grad_norm": 0.9665796903762968, "learning_rate": 5.153975292780852e-06, "loss": 0.9827, "step": 2025 }, { "epoch": 2.709030100334448, "grad_norm": 1.1637688547588272, "learning_rate": 5.149310783942414e-06, "loss": 0.9911, "step": 2026 }, { "epoch": 2.7103678929765884, "grad_norm": 1.1000349082517762, "learning_rate": 5.144646145038311e-06, "loss": 1.0339, "step": 2027 }, { "epoch": 2.711705685618729, "grad_norm": 1.0578869111890712, "learning_rate": 5.139981380131943e-06, "loss": 1.2419, "step": 2028 }, { "epoch": 2.7130434782608694, "grad_norm": 0.8896231918449652, "learning_rate": 5.135316493286818e-06, "loss": 1.2384, "step": 2029 }, { "epoch": 2.71438127090301, "grad_norm": 0.90794760780843, "learning_rate": 5.1306514885665524e-06, "loss": 0.9547, "step": 2030 }, { "epoch": 2.7157190635451505, "grad_norm": 0.994025454944422, "learning_rate": 5.125986370034862e-06, "loss": 1.0472, "step": 2031 }, { "epoch": 2.717056856187291, "grad_norm": 0.7999740182178026, "learning_rate": 5.121321141755568e-06, "loss": 1.0645, "step": 2032 }, { "epoch": 2.7183946488294315, "grad_norm": 0.8545134624385412, "learning_rate": 5.116655807792581e-06, "loss": 1.0706, "step": 2033 }, { "epoch": 2.719732441471572, "grad_norm": 0.8589577229013902, "learning_rate": 5.111990372209906e-06, "loss": 1.0047, "step": 2034 }, { "epoch": 2.7210702341137125, "grad_norm": 0.8718874221891096, "learning_rate": 5.107324839071638e-06, "loss": 0.9774, "step": 2035 }, { "epoch": 2.722408026755853, "grad_norm": 0.9124714954706943, "learning_rate": 5.102659212441953e-06, "loss": 0.9153, "step": 2036 }, { "epoch": 2.723745819397993, "grad_norm": 0.8639143511687885, "learning_rate": 5.097993496385112e-06, "loss": 1.0552, "step": 2037 }, { "epoch": 2.725083612040134, "grad_norm": 0.779626954239031, "learning_rate": 5.093327694965453e-06, "loss": 0.9163, "step": 2038 }, { "epoch": 2.726421404682274, "grad_norm": 0.8494359620285198, "learning_rate": 5.088661812247389e-06, "loss": 1.0679, "step": 2039 }, { "epoch": 2.727759197324415, "grad_norm": 0.9219851380772683, "learning_rate": 5.083995852295402e-06, "loss": 1.1049, "step": 2040 }, { "epoch": 2.729096989966555, "grad_norm": 0.7983448094744993, "learning_rate": 5.07932981917404e-06, "loss": 1.0789, "step": 2041 }, { "epoch": 2.730434782608696, "grad_norm": 0.9165521958084314, "learning_rate": 5.0746637169479205e-06, "loss": 1.011, "step": 2042 }, { "epoch": 2.731772575250836, "grad_norm": 0.8564544215456857, "learning_rate": 5.069997549681718e-06, "loss": 1.2687, "step": 2043 }, { "epoch": 2.7331103678929765, "grad_norm": 0.8482111619141011, "learning_rate": 5.06533132144016e-06, "loss": 0.9946, "step": 2044 }, { "epoch": 2.734448160535117, "grad_norm": 0.8210359903049502, "learning_rate": 5.060665036288034e-06, "loss": 1.0085, "step": 2045 }, { "epoch": 2.7357859531772575, "grad_norm": 0.8424335022713717, "learning_rate": 5.0559986982901695e-06, "loss": 1.0998, "step": 2046 }, { "epoch": 2.7371237458193978, "grad_norm": 0.95149059277568, "learning_rate": 5.05133231151145e-06, "loss": 0.8663, "step": 2047 }, { "epoch": 2.7384615384615385, "grad_norm": 1.402393251928025, "learning_rate": 5.046665880016795e-06, "loss": 0.8451, "step": 2048 }, { "epoch": 2.739799331103679, "grad_norm": 0.9711511129370838, "learning_rate": 5.041999407871168e-06, "loss": 1.1569, "step": 2049 }, { "epoch": 2.7411371237458195, "grad_norm": 0.842401871735485, "learning_rate": 5.037332899139563e-06, "loss": 1.1615, "step": 2050 }, { "epoch": 2.74247491638796, "grad_norm": 1.113542826355207, "learning_rate": 5.0326663578870095e-06, "loss": 1.1637, "step": 2051 }, { "epoch": 2.7438127090301005, "grad_norm": 0.9078916196536618, "learning_rate": 5.0279997881785635e-06, "loss": 1.0868, "step": 2052 }, { "epoch": 2.745150501672241, "grad_norm": 1.007772343768085, "learning_rate": 5.0233331940793074e-06, "loss": 1.0424, "step": 2053 }, { "epoch": 2.746488294314381, "grad_norm": 0.9526173078329278, "learning_rate": 5.018666579654342e-06, "loss": 0.9457, "step": 2054 }, { "epoch": 2.747826086956522, "grad_norm": 0.9812288113846908, "learning_rate": 5.01399994896879e-06, "loss": 0.8903, "step": 2055 }, { "epoch": 2.749163879598662, "grad_norm": 0.94904078957751, "learning_rate": 5.009333306087784e-06, "loss": 1.0795, "step": 2056 }, { "epoch": 2.7505016722408024, "grad_norm": 1.1140701825801318, "learning_rate": 5.00466665507647e-06, "loss": 0.9823, "step": 2057 }, { "epoch": 2.751839464882943, "grad_norm": 0.8619530476042616, "learning_rate": 5e-06, "loss": 1.1366, "step": 2058 }, { "epoch": 2.7531772575250835, "grad_norm": 1.3188920279216703, "learning_rate": 4.995333344923531e-06, "loss": 1.336, "step": 2059 }, { "epoch": 2.754515050167224, "grad_norm": 0.8711436415276719, "learning_rate": 4.990666693912218e-06, "loss": 0.828, "step": 2060 }, { "epoch": 2.7558528428093645, "grad_norm": 1.0441657209743407, "learning_rate": 4.986000051031212e-06, "loss": 1.1161, "step": 2061 }, { "epoch": 2.7571906354515052, "grad_norm": 0.9288777106819964, "learning_rate": 4.9813334203456595e-06, "loss": 1.1881, "step": 2062 }, { "epoch": 2.7585284280936455, "grad_norm": 1.0136345320086588, "learning_rate": 4.976666805920694e-06, "loss": 1.2515, "step": 2063 }, { "epoch": 2.759866220735786, "grad_norm": 0.9481211938364827, "learning_rate": 4.972000211821438e-06, "loss": 1.1366, "step": 2064 }, { "epoch": 2.7612040133779265, "grad_norm": 0.879139145467375, "learning_rate": 4.967333642112992e-06, "loss": 1.1813, "step": 2065 }, { "epoch": 2.762541806020067, "grad_norm": 0.9280017199591174, "learning_rate": 4.9626671008604385e-06, "loss": 1.2583, "step": 2066 }, { "epoch": 2.763879598662207, "grad_norm": 1.2231086153825528, "learning_rate": 4.958000592128834e-06, "loss": 0.8401, "step": 2067 }, { "epoch": 2.765217391304348, "grad_norm": 1.090772709554255, "learning_rate": 4.953334119983206e-06, "loss": 1.0997, "step": 2068 }, { "epoch": 2.766555183946488, "grad_norm": 1.1321380553032347, "learning_rate": 4.948667688488552e-06, "loss": 0.8018, "step": 2069 }, { "epoch": 2.767892976588629, "grad_norm": 0.9367979088444879, "learning_rate": 4.944001301709832e-06, "loss": 0.8533, "step": 2070 }, { "epoch": 2.769230769230769, "grad_norm": 0.9047276613930044, "learning_rate": 4.9393349637119695e-06, "loss": 1.2479, "step": 2071 }, { "epoch": 2.77056856187291, "grad_norm": 0.9614366932466819, "learning_rate": 4.934668678559842e-06, "loss": 1.0697, "step": 2072 }, { "epoch": 2.77190635451505, "grad_norm": 0.9468646456230952, "learning_rate": 4.930002450318282e-06, "loss": 1.1047, "step": 2073 }, { "epoch": 2.7732441471571905, "grad_norm": 0.8607255087142659, "learning_rate": 4.925336283052079e-06, "loss": 1.0691, "step": 2074 }, { "epoch": 2.774581939799331, "grad_norm": 0.9151967788988621, "learning_rate": 4.9206701808259605e-06, "loss": 0.9245, "step": 2075 }, { "epoch": 2.7759197324414715, "grad_norm": 0.84656793548214, "learning_rate": 4.9160041477046e-06, "loss": 1.0852, "step": 2076 }, { "epoch": 2.777257525083612, "grad_norm": 1.1020726153214433, "learning_rate": 4.911338187752612e-06, "loss": 0.9771, "step": 2077 }, { "epoch": 2.7785953177257525, "grad_norm": 0.8632477458247616, "learning_rate": 4.906672305034548e-06, "loss": 0.9636, "step": 2078 }, { "epoch": 2.779933110367893, "grad_norm": 1.3735055376725158, "learning_rate": 4.9020065036148885e-06, "loss": 1.1439, "step": 2079 }, { "epoch": 2.7812709030100335, "grad_norm": 0.8834094409743487, "learning_rate": 4.8973407875580485e-06, "loss": 1.0455, "step": 2080 }, { "epoch": 2.782608695652174, "grad_norm": 0.9271128872927928, "learning_rate": 4.892675160928364e-06, "loss": 0.9916, "step": 2081 }, { "epoch": 2.7839464882943146, "grad_norm": 1.029602105928199, "learning_rate": 4.888009627790095e-06, "loss": 1.0144, "step": 2082 }, { "epoch": 2.785284280936455, "grad_norm": 0.9928959222857744, "learning_rate": 4.8833441922074194e-06, "loss": 0.8691, "step": 2083 }, { "epoch": 2.786622073578595, "grad_norm": 1.2587962524136056, "learning_rate": 4.878678858244432e-06, "loss": 1.0029, "step": 2084 }, { "epoch": 2.787959866220736, "grad_norm": 0.8751423574935386, "learning_rate": 4.874013629965138e-06, "loss": 0.9717, "step": 2085 }, { "epoch": 2.789297658862876, "grad_norm": 1.0016249870520093, "learning_rate": 4.869348511433449e-06, "loss": 1.0406, "step": 2086 }, { "epoch": 2.7906354515050165, "grad_norm": 0.8970517592828624, "learning_rate": 4.864683506713183e-06, "loss": 1.4133, "step": 2087 }, { "epoch": 2.791973244147157, "grad_norm": 0.8794000987179231, "learning_rate": 4.860018619868058e-06, "loss": 1.0199, "step": 2088 }, { "epoch": 2.793311036789298, "grad_norm": 0.9448233496639538, "learning_rate": 4.85535385496169e-06, "loss": 0.9353, "step": 2089 }, { "epoch": 2.794648829431438, "grad_norm": 0.9203657628501833, "learning_rate": 4.850689216057587e-06, "loss": 1.3894, "step": 2090 }, { "epoch": 2.7959866220735785, "grad_norm": 1.3186869489382844, "learning_rate": 4.846024707219149e-06, "loss": 1.1177, "step": 2091 }, { "epoch": 2.7973244147157192, "grad_norm": 0.9583828317172524, "learning_rate": 4.841360332509663e-06, "loss": 0.7275, "step": 2092 }, { "epoch": 2.7986622073578595, "grad_norm": 0.8547943874695131, "learning_rate": 4.836696095992296e-06, "loss": 0.9166, "step": 2093 }, { "epoch": 2.8, "grad_norm": 0.8530533843131057, "learning_rate": 4.8320320017301e-06, "loss": 1.0914, "step": 2094 }, { "epoch": 2.8013377926421406, "grad_norm": 0.7705690798981317, "learning_rate": 4.827368053785999e-06, "loss": 1.1937, "step": 2095 }, { "epoch": 2.802675585284281, "grad_norm": 1.1283755219372047, "learning_rate": 4.82270425622279e-06, "loss": 1.1382, "step": 2096 }, { "epoch": 2.804013377926421, "grad_norm": 0.90243620806898, "learning_rate": 4.818040613103139e-06, "loss": 1.308, "step": 2097 }, { "epoch": 2.805351170568562, "grad_norm": 0.9571831075701341, "learning_rate": 4.81337712848958e-06, "loss": 1.1399, "step": 2098 }, { "epoch": 2.8066889632107026, "grad_norm": 0.9081458714839811, "learning_rate": 4.808713806444506e-06, "loss": 1.0961, "step": 2099 }, { "epoch": 2.808026755852843, "grad_norm": 0.8450633579421545, "learning_rate": 4.804050651030168e-06, "loss": 0.8236, "step": 2100 }, { "epoch": 2.809364548494983, "grad_norm": 0.918063331863433, "learning_rate": 4.799387666308675e-06, "loss": 1.116, "step": 2101 }, { "epoch": 2.810702341137124, "grad_norm": 0.9046017014814403, "learning_rate": 4.794724856341985e-06, "loss": 1.051, "step": 2102 }, { "epoch": 2.812040133779264, "grad_norm": 0.7955248913100311, "learning_rate": 4.790062225191902e-06, "loss": 0.9847, "step": 2103 }, { "epoch": 2.8133779264214045, "grad_norm": 0.9705171881104572, "learning_rate": 4.785399776920081e-06, "loss": 1.0941, "step": 2104 }, { "epoch": 2.8147157190635452, "grad_norm": 1.016441043016156, "learning_rate": 4.780737515588011e-06, "loss": 0.8593, "step": 2105 }, { "epoch": 2.8160535117056855, "grad_norm": 0.9366941890535689, "learning_rate": 4.77607544525702e-06, "loss": 1.1205, "step": 2106 }, { "epoch": 2.8173913043478263, "grad_norm": 0.8517777180861654, "learning_rate": 4.77141356998827e-06, "loss": 1.0215, "step": 2107 }, { "epoch": 2.8187290969899665, "grad_norm": 1.051517014028963, "learning_rate": 4.7667518938427534e-06, "loss": 1.1441, "step": 2108 }, { "epoch": 2.8200668896321073, "grad_norm": 0.8600411247017793, "learning_rate": 4.762090420881289e-06, "loss": 1.275, "step": 2109 }, { "epoch": 2.8214046822742476, "grad_norm": 0.820903976581373, "learning_rate": 4.757429155164518e-06, "loss": 1.1913, "step": 2110 }, { "epoch": 2.822742474916388, "grad_norm": 0.9811827799531575, "learning_rate": 4.752768100752902e-06, "loss": 0.9677, "step": 2111 }, { "epoch": 2.8240802675585286, "grad_norm": 0.9798691275083052, "learning_rate": 4.748107261706716e-06, "loss": 0.9274, "step": 2112 }, { "epoch": 2.825418060200669, "grad_norm": 0.9052063512793861, "learning_rate": 4.7434466420860515e-06, "loss": 0.9802, "step": 2113 }, { "epoch": 2.826755852842809, "grad_norm": 0.8908922370123022, "learning_rate": 4.7387862459508074e-06, "loss": 1.0184, "step": 2114 }, { "epoch": 2.82809364548495, "grad_norm": 0.8472792100337703, "learning_rate": 4.734126077360685e-06, "loss": 1.066, "step": 2115 }, { "epoch": 2.82943143812709, "grad_norm": 1.1626702406038687, "learning_rate": 4.729466140375192e-06, "loss": 1.0298, "step": 2116 }, { "epoch": 2.830769230769231, "grad_norm": 0.9810191609796189, "learning_rate": 4.724806439053629e-06, "loss": 0.9603, "step": 2117 }, { "epoch": 2.832107023411371, "grad_norm": 1.101845085093852, "learning_rate": 4.720146977455098e-06, "loss": 1.0763, "step": 2118 }, { "epoch": 2.833444816053512, "grad_norm": 0.9229847585586196, "learning_rate": 4.715487759638486e-06, "loss": 0.9569, "step": 2119 }, { "epoch": 2.8347826086956522, "grad_norm": 1.0515133712388038, "learning_rate": 4.7108287896624695e-06, "loss": 1.0453, "step": 2120 }, { "epoch": 2.8361204013377925, "grad_norm": 0.9514991535957462, "learning_rate": 4.706170071585513e-06, "loss": 1.4366, "step": 2121 }, { "epoch": 2.8374581939799333, "grad_norm": 0.9748406290981186, "learning_rate": 4.7015116094658544e-06, "loss": 0.8685, "step": 2122 }, { "epoch": 2.8387959866220736, "grad_norm": 1.0746843170230378, "learning_rate": 4.6968534073615145e-06, "loss": 1.2807, "step": 2123 }, { "epoch": 2.840133779264214, "grad_norm": 1.083474449007014, "learning_rate": 4.692195469330286e-06, "loss": 1.0639, "step": 2124 }, { "epoch": 2.8414715719063546, "grad_norm": 1.3741172858818798, "learning_rate": 4.687537799429731e-06, "loss": 0.7963, "step": 2125 }, { "epoch": 2.842809364548495, "grad_norm": 0.9515801039515189, "learning_rate": 4.682880401717178e-06, "loss": 1.0452, "step": 2126 }, { "epoch": 2.8441471571906356, "grad_norm": 0.8332306883123399, "learning_rate": 4.678223280249718e-06, "loss": 0.9232, "step": 2127 }, { "epoch": 2.845484949832776, "grad_norm": 1.100245404868723, "learning_rate": 4.673566439084204e-06, "loss": 0.8286, "step": 2128 }, { "epoch": 2.8468227424749166, "grad_norm": 0.952130717459923, "learning_rate": 4.66890988227724e-06, "loss": 0.9249, "step": 2129 }, { "epoch": 2.848160535117057, "grad_norm": 1.0160360060282778, "learning_rate": 4.664253613885187e-06, "loss": 0.9573, "step": 2130 }, { "epoch": 2.849498327759197, "grad_norm": 1.0939157444501788, "learning_rate": 4.659597637964153e-06, "loss": 0.9166, "step": 2131 }, { "epoch": 2.850836120401338, "grad_norm": 0.7950739517336851, "learning_rate": 4.65494195856999e-06, "loss": 1.1353, "step": 2132 }, { "epoch": 2.8521739130434782, "grad_norm": 0.9273680109337571, "learning_rate": 4.650286579758291e-06, "loss": 1.254, "step": 2133 }, { "epoch": 2.8535117056856185, "grad_norm": 0.8171584822222895, "learning_rate": 4.645631505584393e-06, "loss": 1.2102, "step": 2134 }, { "epoch": 2.8548494983277592, "grad_norm": 1.1511728130434329, "learning_rate": 4.640976740103363e-06, "loss": 0.8835, "step": 2135 }, { "epoch": 2.8561872909698995, "grad_norm": 0.8901774735007809, "learning_rate": 4.636322287369997e-06, "loss": 0.8334, "step": 2136 }, { "epoch": 2.8575250836120403, "grad_norm": 1.0531613832147984, "learning_rate": 4.6316681514388235e-06, "loss": 1.2434, "step": 2137 }, { "epoch": 2.8588628762541806, "grad_norm": 0.972524283083011, "learning_rate": 4.6270143363640914e-06, "loss": 1.0877, "step": 2138 }, { "epoch": 2.8602006688963213, "grad_norm": 0.8241305644226401, "learning_rate": 4.622360846199772e-06, "loss": 1.016, "step": 2139 }, { "epoch": 2.8615384615384616, "grad_norm": 0.8450730029637682, "learning_rate": 4.617707684999554e-06, "loss": 0.9429, "step": 2140 }, { "epoch": 2.862876254180602, "grad_norm": 1.2887534802466298, "learning_rate": 4.613054856816837e-06, "loss": 0.9059, "step": 2141 }, { "epoch": 2.8642140468227426, "grad_norm": 0.8327568829669055, "learning_rate": 4.608402365704734e-06, "loss": 1.0201, "step": 2142 }, { "epoch": 2.865551839464883, "grad_norm": 1.0050355383185585, "learning_rate": 4.603750215716057e-06, "loss": 0.8979, "step": 2143 }, { "epoch": 2.866889632107023, "grad_norm": 0.9278657515330697, "learning_rate": 4.599098410903334e-06, "loss": 1.1357, "step": 2144 }, { "epoch": 2.868227424749164, "grad_norm": 0.9604492687486095, "learning_rate": 4.594446955318781e-06, "loss": 0.8912, "step": 2145 }, { "epoch": 2.869565217391304, "grad_norm": 0.9926552287800351, "learning_rate": 4.589795853014313e-06, "loss": 0.9337, "step": 2146 }, { "epoch": 2.870903010033445, "grad_norm": 0.9244030423511244, "learning_rate": 4.585145108041538e-06, "loss": 0.8478, "step": 2147 }, { "epoch": 2.8722408026755852, "grad_norm": 0.979948889077756, "learning_rate": 4.580494724451752e-06, "loss": 0.9351, "step": 2148 }, { "epoch": 2.873578595317726, "grad_norm": 0.8301368562274422, "learning_rate": 4.575844706295938e-06, "loss": 0.8562, "step": 2149 }, { "epoch": 2.8749163879598663, "grad_norm": 1.036620945644359, "learning_rate": 4.5711950576247585e-06, "loss": 1.1383, "step": 2150 }, { "epoch": 2.8762541806020065, "grad_norm": 0.9800203047698068, "learning_rate": 4.566545782488554e-06, "loss": 0.913, "step": 2151 }, { "epoch": 2.8775919732441473, "grad_norm": 1.1296683679011041, "learning_rate": 4.5618968849373415e-06, "loss": 1.2587, "step": 2152 }, { "epoch": 2.8789297658862876, "grad_norm": 1.2192864838834991, "learning_rate": 4.557248369020806e-06, "loss": 0.8515, "step": 2153 }, { "epoch": 2.880267558528428, "grad_norm": 0.9933035899147166, "learning_rate": 4.552600238788306e-06, "loss": 1.0374, "step": 2154 }, { "epoch": 2.8816053511705686, "grad_norm": 1.0275504079700946, "learning_rate": 4.5479524982888575e-06, "loss": 1.1569, "step": 2155 }, { "epoch": 2.882943143812709, "grad_norm": 0.8827571353878665, "learning_rate": 4.543305151571141e-06, "loss": 1.2842, "step": 2156 }, { "epoch": 2.8842809364548496, "grad_norm": 0.9475834453259495, "learning_rate": 4.53865820268349e-06, "loss": 1.0111, "step": 2157 }, { "epoch": 2.88561872909699, "grad_norm": 0.8742707472033884, "learning_rate": 4.534011655673898e-06, "loss": 0.918, "step": 2158 }, { "epoch": 2.8869565217391306, "grad_norm": 0.8390966265838724, "learning_rate": 4.529365514590002e-06, "loss": 0.9084, "step": 2159 }, { "epoch": 2.888294314381271, "grad_norm": 0.8868946830269732, "learning_rate": 4.524719783479088e-06, "loss": 1.0529, "step": 2160 }, { "epoch": 2.8896321070234112, "grad_norm": 0.8243743576496338, "learning_rate": 4.5200744663880856e-06, "loss": 0.9598, "step": 2161 }, { "epoch": 2.890969899665552, "grad_norm": 0.8014690508279467, "learning_rate": 4.515429567363562e-06, "loss": 0.8806, "step": 2162 }, { "epoch": 2.8923076923076922, "grad_norm": 0.8386197157102914, "learning_rate": 4.510785090451719e-06, "loss": 1.0161, "step": 2163 }, { "epoch": 2.8936454849498325, "grad_norm": 0.9099733046217923, "learning_rate": 4.506141039698398e-06, "loss": 0.795, "step": 2164 }, { "epoch": 2.8949832775919733, "grad_norm": 1.363626415462802, "learning_rate": 4.501497419149062e-06, "loss": 0.9406, "step": 2165 }, { "epoch": 2.8963210702341136, "grad_norm": 1.0676726329286932, "learning_rate": 4.4968542328488e-06, "loss": 0.9637, "step": 2166 }, { "epoch": 2.8976588628762543, "grad_norm": 0.8106444029544972, "learning_rate": 4.492211484842324e-06, "loss": 1.0136, "step": 2167 }, { "epoch": 2.8989966555183946, "grad_norm": 1.0249876304054801, "learning_rate": 4.4875691791739655e-06, "loss": 1.0718, "step": 2168 }, { "epoch": 2.9003344481605353, "grad_norm": 0.9628782350920088, "learning_rate": 4.482927319887669e-06, "loss": 0.8177, "step": 2169 }, { "epoch": 2.9016722408026756, "grad_norm": 0.9092833319238383, "learning_rate": 4.478285911026989e-06, "loss": 1.0598, "step": 2170 }, { "epoch": 2.903010033444816, "grad_norm": 1.1075217084116349, "learning_rate": 4.4736449566350924e-06, "loss": 1.0914, "step": 2171 }, { "epoch": 2.9043478260869566, "grad_norm": 1.0236951113140738, "learning_rate": 4.469004460754743e-06, "loss": 1.132, "step": 2172 }, { "epoch": 2.905685618729097, "grad_norm": 1.2166517603457607, "learning_rate": 4.46436442742831e-06, "loss": 0.9665, "step": 2173 }, { "epoch": 2.907023411371237, "grad_norm": 0.9643471230529127, "learning_rate": 4.45972486069776e-06, "loss": 1.3379, "step": 2174 }, { "epoch": 2.908361204013378, "grad_norm": 0.7704049445166805, "learning_rate": 4.455085764604653e-06, "loss": 0.9521, "step": 2175 }, { "epoch": 2.9096989966555182, "grad_norm": 1.0532919139227808, "learning_rate": 4.450447143190136e-06, "loss": 0.9924, "step": 2176 }, { "epoch": 2.911036789297659, "grad_norm": 1.0764575173479025, "learning_rate": 4.445809000494945e-06, "loss": 1.0835, "step": 2177 }, { "epoch": 2.9123745819397993, "grad_norm": 0.8055815673558907, "learning_rate": 4.441171340559399e-06, "loss": 1.168, "step": 2178 }, { "epoch": 2.91371237458194, "grad_norm": 0.932307497783027, "learning_rate": 4.436534167423395e-06, "loss": 0.9742, "step": 2179 }, { "epoch": 2.9150501672240803, "grad_norm": 0.7988900591101925, "learning_rate": 4.431897485126408e-06, "loss": 1.0432, "step": 2180 }, { "epoch": 2.9163879598662206, "grad_norm": 0.8236135332389911, "learning_rate": 4.427261297707482e-06, "loss": 0.9436, "step": 2181 }, { "epoch": 2.9177257525083613, "grad_norm": 1.0062321178970686, "learning_rate": 4.422625609205235e-06, "loss": 1.1284, "step": 2182 }, { "epoch": 2.9190635451505016, "grad_norm": 0.7807286927633281, "learning_rate": 4.417990423657845e-06, "loss": 1.0405, "step": 2183 }, { "epoch": 2.920401337792642, "grad_norm": 0.8980047063480919, "learning_rate": 4.413355745103057e-06, "loss": 0.8469, "step": 2184 }, { "epoch": 2.9217391304347826, "grad_norm": 0.9598609560177991, "learning_rate": 4.40872157757817e-06, "loss": 1.0707, "step": 2185 }, { "epoch": 2.9230769230769234, "grad_norm": 0.8496924158425851, "learning_rate": 4.404087925120041e-06, "loss": 1.0733, "step": 2186 }, { "epoch": 2.9244147157190636, "grad_norm": 0.9155469569120517, "learning_rate": 4.399454791765076e-06, "loss": 1.0098, "step": 2187 }, { "epoch": 2.925752508361204, "grad_norm": 0.9094467673841312, "learning_rate": 4.3948221815492294e-06, "loss": 0.8724, "step": 2188 }, { "epoch": 2.9270903010033447, "grad_norm": 0.8734966841048291, "learning_rate": 4.390190098508001e-06, "loss": 1.0473, "step": 2189 }, { "epoch": 2.928428093645485, "grad_norm": 0.909829336813761, "learning_rate": 4.3855585466764305e-06, "loss": 1.1335, "step": 2190 }, { "epoch": 2.9297658862876252, "grad_norm": 0.9122556487599842, "learning_rate": 4.3809275300890956e-06, "loss": 0.9776, "step": 2191 }, { "epoch": 2.931103678929766, "grad_norm": 1.104600535553276, "learning_rate": 4.376297052780106e-06, "loss": 1.0062, "step": 2192 }, { "epoch": 2.9324414715719063, "grad_norm": 1.1785663008759506, "learning_rate": 4.371667118783101e-06, "loss": 1.1907, "step": 2193 }, { "epoch": 2.9337792642140466, "grad_norm": 0.9395255322856512, "learning_rate": 4.367037732131254e-06, "loss": 0.9411, "step": 2194 }, { "epoch": 2.9351170568561873, "grad_norm": 0.8571439499806015, "learning_rate": 4.362408896857251e-06, "loss": 1.0038, "step": 2195 }, { "epoch": 2.936454849498328, "grad_norm": 0.9811675412482117, "learning_rate": 4.357780616993305e-06, "loss": 0.881, "step": 2196 }, { "epoch": 2.9377926421404683, "grad_norm": 1.0381886678203711, "learning_rate": 4.353152896571143e-06, "loss": 0.9622, "step": 2197 }, { "epoch": 2.9391304347826086, "grad_norm": 1.0288840255238316, "learning_rate": 4.348525739622003e-06, "loss": 0.9272, "step": 2198 }, { "epoch": 2.9404682274247493, "grad_norm": 0.8224869456811106, "learning_rate": 4.343899150176635e-06, "loss": 1.1026, "step": 2199 }, { "epoch": 2.9418060200668896, "grad_norm": 1.4350644119410405, "learning_rate": 4.339273132265294e-06, "loss": 1.1128, "step": 2200 }, { "epoch": 2.94314381270903, "grad_norm": 1.1115311450189365, "learning_rate": 4.334647689917734e-06, "loss": 0.9621, "step": 2201 }, { "epoch": 2.9444816053511706, "grad_norm": 1.2165473137530287, "learning_rate": 4.3300228271632105e-06, "loss": 1.0025, "step": 2202 }, { "epoch": 2.945819397993311, "grad_norm": 1.0002056578014809, "learning_rate": 4.325398548030473e-06, "loss": 1.099, "step": 2203 }, { "epoch": 2.9471571906354512, "grad_norm": 0.9353721450618303, "learning_rate": 4.320774856547767e-06, "loss": 1.0095, "step": 2204 }, { "epoch": 2.948494983277592, "grad_norm": 0.8979358543958192, "learning_rate": 4.316151756742821e-06, "loss": 1.1979, "step": 2205 }, { "epoch": 2.9498327759197327, "grad_norm": 1.2354774303736997, "learning_rate": 4.311529252642848e-06, "loss": 1.0048, "step": 2206 }, { "epoch": 2.951170568561873, "grad_norm": 1.3347143904650174, "learning_rate": 4.306907348274545e-06, "loss": 1.1042, "step": 2207 }, { "epoch": 2.9525083612040133, "grad_norm": 0.8721137765396098, "learning_rate": 4.302286047664083e-06, "loss": 0.8563, "step": 2208 }, { "epoch": 2.953846153846154, "grad_norm": 1.150184698511428, "learning_rate": 4.2976653548371115e-06, "loss": 1.0858, "step": 2209 }, { "epoch": 2.9551839464882943, "grad_norm": 0.7835504185064788, "learning_rate": 4.293045273818748e-06, "loss": 1.1395, "step": 2210 }, { "epoch": 2.9565217391304346, "grad_norm": 0.8514797080755964, "learning_rate": 4.2884258086335755e-06, "loss": 1.1937, "step": 2211 }, { "epoch": 2.9578595317725753, "grad_norm": 0.9012583144094606, "learning_rate": 4.283806963305644e-06, "loss": 0.6423, "step": 2212 }, { "epoch": 2.9591973244147156, "grad_norm": 1.2197159409686593, "learning_rate": 4.27918874185846e-06, "loss": 0.7805, "step": 2213 }, { "epoch": 2.9605351170568563, "grad_norm": 1.2909665514334512, "learning_rate": 4.274571148314991e-06, "loss": 0.8117, "step": 2214 }, { "epoch": 2.9618729096989966, "grad_norm": 0.9072991768112973, "learning_rate": 4.269954186697654e-06, "loss": 0.9991, "step": 2215 }, { "epoch": 2.9632107023411374, "grad_norm": 1.1946047127936665, "learning_rate": 4.265337861028316e-06, "loss": 1.117, "step": 2216 }, { "epoch": 2.9645484949832777, "grad_norm": 1.0756752385471564, "learning_rate": 4.26072217532829e-06, "loss": 1.1586, "step": 2217 }, { "epoch": 2.965886287625418, "grad_norm": 0.9433068261195041, "learning_rate": 4.256107133618333e-06, "loss": 1.0544, "step": 2218 }, { "epoch": 2.9672240802675587, "grad_norm": 1.030276261126104, "learning_rate": 4.251492739918641e-06, "loss": 0.9733, "step": 2219 }, { "epoch": 2.968561872909699, "grad_norm": 1.0623858943349442, "learning_rate": 4.2468789982488415e-06, "loss": 0.9634, "step": 2220 }, { "epoch": 2.9698996655518393, "grad_norm": 1.1079755664425932, "learning_rate": 4.242265912628e-06, "loss": 0.9113, "step": 2221 }, { "epoch": 2.97123745819398, "grad_norm": 1.2099369931898494, "learning_rate": 4.2376534870746054e-06, "loss": 1.197, "step": 2222 }, { "epoch": 2.9725752508361203, "grad_norm": 0.9833103360032374, "learning_rate": 4.233041725606573e-06, "loss": 0.862, "step": 2223 }, { "epoch": 2.973913043478261, "grad_norm": 1.0991568571729, "learning_rate": 4.228430632241244e-06, "loss": 1.0698, "step": 2224 }, { "epoch": 2.9752508361204013, "grad_norm": 0.9933241001672425, "learning_rate": 4.223820210995372e-06, "loss": 1.2393, "step": 2225 }, { "epoch": 2.976588628762542, "grad_norm": 0.7622915531440776, "learning_rate": 4.219210465885127e-06, "loss": 0.9523, "step": 2226 }, { "epoch": 2.9779264214046823, "grad_norm": 1.0647285270096358, "learning_rate": 4.21460140092609e-06, "loss": 1.0805, "step": 2227 }, { "epoch": 2.9792642140468226, "grad_norm": 1.2262940685708386, "learning_rate": 4.209993020133251e-06, "loss": 0.8826, "step": 2228 }, { "epoch": 2.9806020066889634, "grad_norm": 0.8751827810434091, "learning_rate": 4.205385327521002e-06, "loss": 1.1968, "step": 2229 }, { "epoch": 2.9819397993311036, "grad_norm": 0.8637375502907045, "learning_rate": 4.200778327103134e-06, "loss": 1.0491, "step": 2230 }, { "epoch": 2.983277591973244, "grad_norm": 0.8965256036003005, "learning_rate": 4.19617202289284e-06, "loss": 1.069, "step": 2231 }, { "epoch": 2.9846153846153847, "grad_norm": 0.9069980664760637, "learning_rate": 4.191566418902701e-06, "loss": 1.1705, "step": 2232 }, { "epoch": 2.985953177257525, "grad_norm": 0.9873634903093494, "learning_rate": 4.1869615191446925e-06, "loss": 1.0531, "step": 2233 }, { "epoch": 2.9872909698996657, "grad_norm": 0.9861481045430089, "learning_rate": 4.182357327630175e-06, "loss": 1.1116, "step": 2234 }, { "epoch": 2.988628762541806, "grad_norm": 1.0880540303606516, "learning_rate": 4.177753848369892e-06, "loss": 1.4256, "step": 2235 }, { "epoch": 2.9899665551839467, "grad_norm": 0.9321097604546673, "learning_rate": 4.173151085373966e-06, "loss": 1.0086, "step": 2236 }, { "epoch": 2.991304347826087, "grad_norm": 0.9684702803006461, "learning_rate": 4.168549042651896e-06, "loss": 1.1107, "step": 2237 }, { "epoch": 2.9926421404682273, "grad_norm": 0.7632630180810329, "learning_rate": 4.163947724212553e-06, "loss": 0.947, "step": 2238 }, { "epoch": 2.993979933110368, "grad_norm": 0.875774217600485, "learning_rate": 4.159347134064177e-06, "loss": 0.9652, "step": 2239 }, { "epoch": 2.9953177257525083, "grad_norm": 0.9516971175410415, "learning_rate": 4.154747276214377e-06, "loss": 1.0533, "step": 2240 }, { "epoch": 2.9966555183946486, "grad_norm": 1.3751994748290512, "learning_rate": 4.1501481546701185e-06, "loss": 1.0887, "step": 2241 }, { "epoch": 2.9979933110367893, "grad_norm": 1.4135858231714669, "learning_rate": 4.145549773437728e-06, "loss": 0.8112, "step": 2242 }, { "epoch": 2.9993311036789296, "grad_norm": 0.8998574273660873, "learning_rate": 4.140952136522889e-06, "loss": 1.083, "step": 2243 }, { "epoch": 3.0, "grad_norm": 0.8998574273660873, "learning_rate": 4.136355247930636e-06, "loss": 0.9392, "step": 2244 }, { "epoch": 3.0013377926421403, "grad_norm": 1.9626957244337444, "learning_rate": 4.131759111665349e-06, "loss": 0.8255, "step": 2245 }, { "epoch": 3.002675585284281, "grad_norm": 1.1756967833955676, "learning_rate": 4.127163731730755e-06, "loss": 1.1136, "step": 2246 }, { "epoch": 3.0040133779264213, "grad_norm": 1.2543183265998001, "learning_rate": 4.12256911212992e-06, "loss": 1.1294, "step": 2247 }, { "epoch": 3.005351170568562, "grad_norm": 0.7986562370437158, "learning_rate": 4.117975256865252e-06, "loss": 1.0797, "step": 2248 }, { "epoch": 3.0066889632107023, "grad_norm": 0.926202438146485, "learning_rate": 4.113382169938488e-06, "loss": 1.1404, "step": 2249 }, { "epoch": 3.0080267558528426, "grad_norm": 0.9718186308100057, "learning_rate": 4.108789855350699e-06, "loss": 1.2245, "step": 2250 }, { "epoch": 3.0093645484949834, "grad_norm": 0.7160509558050419, "learning_rate": 4.104198317102283e-06, "loss": 0.8653, "step": 2251 }, { "epoch": 3.0107023411371236, "grad_norm": 0.9557358114026884, "learning_rate": 4.099607559192959e-06, "loss": 1.1455, "step": 2252 }, { "epoch": 3.0120401337792644, "grad_norm": 0.9815737629176171, "learning_rate": 4.095017585621767e-06, "loss": 1.0578, "step": 2253 }, { "epoch": 3.0133779264214047, "grad_norm": 0.9491402279193684, "learning_rate": 4.090428400387071e-06, "loss": 0.759, "step": 2254 }, { "epoch": 3.014715719063545, "grad_norm": 0.921367154810677, "learning_rate": 4.0858400074865364e-06, "loss": 0.9234, "step": 2255 }, { "epoch": 3.0160535117056857, "grad_norm": 0.9330736186117555, "learning_rate": 4.081252410917148e-06, "loss": 0.987, "step": 2256 }, { "epoch": 3.017391304347826, "grad_norm": 0.8852190445637175, "learning_rate": 4.076665614675191e-06, "loss": 1.0111, "step": 2257 }, { "epoch": 3.0187290969899667, "grad_norm": 1.058971089444735, "learning_rate": 4.0720796227562585e-06, "loss": 1.0488, "step": 2258 }, { "epoch": 3.020066889632107, "grad_norm": 0.9619239718634741, "learning_rate": 4.067494439155236e-06, "loss": 0.8568, "step": 2259 }, { "epoch": 3.0214046822742473, "grad_norm": 1.3241878220512173, "learning_rate": 4.0629100678663104e-06, "loss": 0.8395, "step": 2260 }, { "epoch": 3.022742474916388, "grad_norm": 1.0809631781518991, "learning_rate": 4.05832651288296e-06, "loss": 1.0348, "step": 2261 }, { "epoch": 3.0240802675585283, "grad_norm": 1.219236996786812, "learning_rate": 4.053743778197951e-06, "loss": 0.835, "step": 2262 }, { "epoch": 3.025418060200669, "grad_norm": 1.2319725178956005, "learning_rate": 4.049161867803334e-06, "loss": 1.0538, "step": 2263 }, { "epoch": 3.0267558528428093, "grad_norm": 0.8812225896314884, "learning_rate": 4.0445807856904465e-06, "loss": 1.1573, "step": 2264 }, { "epoch": 3.0280936454849496, "grad_norm": 0.7586488488391143, "learning_rate": 4.0400005358499e-06, "loss": 0.9569, "step": 2265 }, { "epoch": 3.0294314381270904, "grad_norm": 1.0899870355630707, "learning_rate": 4.035421122271581e-06, "loss": 1.0971, "step": 2266 }, { "epoch": 3.0307692307692307, "grad_norm": 0.886747549252554, "learning_rate": 4.03084254894465e-06, "loss": 0.7763, "step": 2267 }, { "epoch": 3.0321070234113714, "grad_norm": 0.917687078996204, "learning_rate": 4.026264819857533e-06, "loss": 1.1311, "step": 2268 }, { "epoch": 3.0334448160535117, "grad_norm": 0.8919187261539682, "learning_rate": 4.021687938997923e-06, "loss": 1.1886, "step": 2269 }, { "epoch": 3.034782608695652, "grad_norm": 0.8386279900516811, "learning_rate": 4.017111910352771e-06, "loss": 0.924, "step": 2270 }, { "epoch": 3.0361204013377927, "grad_norm": 0.8008637811561009, "learning_rate": 4.012536737908288e-06, "loss": 1.0951, "step": 2271 }, { "epoch": 3.037458193979933, "grad_norm": 0.9744412700392986, "learning_rate": 4.007962425649939e-06, "loss": 1.1505, "step": 2272 }, { "epoch": 3.0387959866220737, "grad_norm": 0.7495567750588298, "learning_rate": 4.003388977562439e-06, "loss": 1.2363, "step": 2273 }, { "epoch": 3.040133779264214, "grad_norm": 1.2660059969553998, "learning_rate": 3.998816397629752e-06, "loss": 0.6949, "step": 2274 }, { "epoch": 3.0414715719063543, "grad_norm": 0.9331328097111112, "learning_rate": 3.994244689835083e-06, "loss": 0.9564, "step": 2275 }, { "epoch": 3.042809364548495, "grad_norm": 1.0148552128399582, "learning_rate": 3.989673858160878e-06, "loss": 0.9058, "step": 2276 }, { "epoch": 3.0441471571906353, "grad_norm": 0.9493639744375073, "learning_rate": 3.985103906588821e-06, "loss": 0.9184, "step": 2277 }, { "epoch": 3.045484949832776, "grad_norm": 1.043961530917917, "learning_rate": 3.980534839099829e-06, "loss": 0.9228, "step": 2278 }, { "epoch": 3.0468227424749164, "grad_norm": 1.3286672646165805, "learning_rate": 3.975966659674048e-06, "loss": 0.7688, "step": 2279 }, { "epoch": 3.0481605351170566, "grad_norm": 1.1956319988477413, "learning_rate": 3.971399372290851e-06, "loss": 1.1017, "step": 2280 }, { "epoch": 3.0494983277591974, "grad_norm": 0.9189485297095035, "learning_rate": 3.966832980928834e-06, "loss": 1.1582, "step": 2281 }, { "epoch": 3.0508361204013377, "grad_norm": 0.9078217173774105, "learning_rate": 3.962267489565813e-06, "loss": 0.9933, "step": 2282 }, { "epoch": 3.0521739130434784, "grad_norm": 0.9008697763218828, "learning_rate": 3.957702902178816e-06, "loss": 0.9843, "step": 2283 }, { "epoch": 3.0535117056856187, "grad_norm": 0.8377739350345415, "learning_rate": 3.953139222744093e-06, "loss": 1.0992, "step": 2284 }, { "epoch": 3.054849498327759, "grad_norm": 0.8224648959870562, "learning_rate": 3.9485764552370934e-06, "loss": 0.8743, "step": 2285 }, { "epoch": 3.0561872909698997, "grad_norm": 0.836579052791847, "learning_rate": 3.944014603632476e-06, "loss": 0.9883, "step": 2286 }, { "epoch": 3.05752508361204, "grad_norm": 1.0627700280393044, "learning_rate": 3.939453671904101e-06, "loss": 1.0695, "step": 2287 }, { "epoch": 3.0588628762541807, "grad_norm": 0.8595517158001753, "learning_rate": 3.93489366402503e-06, "loss": 0.8943, "step": 2288 }, { "epoch": 3.060200668896321, "grad_norm": 0.81472846270611, "learning_rate": 3.930334583967514e-06, "loss": 1.1432, "step": 2289 }, { "epoch": 3.0615384615384613, "grad_norm": 0.8477628000731692, "learning_rate": 3.9257764357030025e-06, "loss": 0.9149, "step": 2290 }, { "epoch": 3.062876254180602, "grad_norm": 0.9355022050583489, "learning_rate": 3.92121922320213e-06, "loss": 1.0984, "step": 2291 }, { "epoch": 3.0642140468227423, "grad_norm": 1.0419191884447119, "learning_rate": 3.916662950434714e-06, "loss": 0.7541, "step": 2292 }, { "epoch": 3.065551839464883, "grad_norm": 1.1864323098324807, "learning_rate": 3.912107621369755e-06, "loss": 1.2359, "step": 2293 }, { "epoch": 3.0668896321070234, "grad_norm": 0.786428526058848, "learning_rate": 3.907553239975437e-06, "loss": 0.9253, "step": 2294 }, { "epoch": 3.068227424749164, "grad_norm": 1.041243210957351, "learning_rate": 3.902999810219109e-06, "loss": 0.9415, "step": 2295 }, { "epoch": 3.0695652173913044, "grad_norm": 1.226559467642848, "learning_rate": 3.898447336067297e-06, "loss": 1.1043, "step": 2296 }, { "epoch": 3.0709030100334447, "grad_norm": 0.8858315138988206, "learning_rate": 3.893895821485692e-06, "loss": 1.0851, "step": 2297 }, { "epoch": 3.0722408026755854, "grad_norm": 1.1135826136732963, "learning_rate": 3.889345270439152e-06, "loss": 0.964, "step": 2298 }, { "epoch": 3.0735785953177257, "grad_norm": 0.8909853148142841, "learning_rate": 3.884795686891692e-06, "loss": 1.0001, "step": 2299 }, { "epoch": 3.074916387959866, "grad_norm": 0.9417551917861404, "learning_rate": 3.8802470748064855e-06, "loss": 1.2083, "step": 2300 }, { "epoch": 3.0762541806020067, "grad_norm": 1.0961352332098144, "learning_rate": 3.875699438145862e-06, "loss": 1.1041, "step": 2301 }, { "epoch": 3.077591973244147, "grad_norm": 0.9267570677255196, "learning_rate": 3.871152780871298e-06, "loss": 0.9794, "step": 2302 }, { "epoch": 3.0789297658862878, "grad_norm": 0.9113684495395631, "learning_rate": 3.866607106943418e-06, "loss": 0.5799, "step": 2303 }, { "epoch": 3.080267558528428, "grad_norm": 0.9934877515160159, "learning_rate": 3.862062420321993e-06, "loss": 0.7645, "step": 2304 }, { "epoch": 3.0816053511705688, "grad_norm": 1.0920950610846198, "learning_rate": 3.857518724965929e-06, "loss": 1.0298, "step": 2305 }, { "epoch": 3.082943143812709, "grad_norm": 1.2343429432307742, "learning_rate": 3.852976024833271e-06, "loss": 1.0762, "step": 2306 }, { "epoch": 3.0842809364548494, "grad_norm": 1.0676651201950829, "learning_rate": 3.8484343238811976e-06, "loss": 0.9159, "step": 2307 }, { "epoch": 3.08561872909699, "grad_norm": 1.183980987670622, "learning_rate": 3.8438936260660145e-06, "loss": 0.997, "step": 2308 }, { "epoch": 3.0869565217391304, "grad_norm": 0.9328707611214607, "learning_rate": 3.839353935343156e-06, "loss": 1.0063, "step": 2309 }, { "epoch": 3.088294314381271, "grad_norm": 0.8139324543603579, "learning_rate": 3.834815255667179e-06, "loss": 1.0236, "step": 2310 }, { "epoch": 3.0896321070234114, "grad_norm": 0.8681240945979088, "learning_rate": 3.8302775909917585e-06, "loss": 0.948, "step": 2311 }, { "epoch": 3.0909698996655517, "grad_norm": 0.9125591813367786, "learning_rate": 3.8257409452696845e-06, "loss": 0.9993, "step": 2312 }, { "epoch": 3.0923076923076924, "grad_norm": 1.1960000165714102, "learning_rate": 3.821205322452863e-06, "loss": 1.2935, "step": 2313 }, { "epoch": 3.0936454849498327, "grad_norm": 0.8191887092775252, "learning_rate": 3.816670726492307e-06, "loss": 1.0419, "step": 2314 }, { "epoch": 3.0949832775919734, "grad_norm": 0.85564435159501, "learning_rate": 3.812137161338133e-06, "loss": 1.0225, "step": 2315 }, { "epoch": 3.0963210702341137, "grad_norm": 0.7109070805821186, "learning_rate": 3.8076046309395627e-06, "loss": 1.3909, "step": 2316 }, { "epoch": 3.097658862876254, "grad_norm": 1.039652548554782, "learning_rate": 3.803073139244913e-06, "loss": 0.9543, "step": 2317 }, { "epoch": 3.0989966555183948, "grad_norm": 0.8938548309852771, "learning_rate": 3.7985426902015987e-06, "loss": 1.178, "step": 2318 }, { "epoch": 3.100334448160535, "grad_norm": 0.9752107467840749, "learning_rate": 3.794013287756125e-06, "loss": 1.0953, "step": 2319 }, { "epoch": 3.101672240802676, "grad_norm": 0.7213094109059524, "learning_rate": 3.789484935854088e-06, "loss": 0.7675, "step": 2320 }, { "epoch": 3.103010033444816, "grad_norm": 1.0006399206885757, "learning_rate": 3.784957638440165e-06, "loss": 1.1556, "step": 2321 }, { "epoch": 3.1043478260869564, "grad_norm": 0.8104592280516204, "learning_rate": 3.7804313994581143e-06, "loss": 0.9379, "step": 2322 }, { "epoch": 3.105685618729097, "grad_norm": 1.0952182152746652, "learning_rate": 3.775906222850778e-06, "loss": 0.9969, "step": 2323 }, { "epoch": 3.1070234113712374, "grad_norm": 0.978126709534746, "learning_rate": 3.7713821125600687e-06, "loss": 0.8142, "step": 2324 }, { "epoch": 3.108361204013378, "grad_norm": 0.7863338553085566, "learning_rate": 3.766859072526969e-06, "loss": 0.858, "step": 2325 }, { "epoch": 3.1096989966555184, "grad_norm": 0.801045395499979, "learning_rate": 3.7623371066915305e-06, "loss": 1.0738, "step": 2326 }, { "epoch": 3.1110367892976587, "grad_norm": 0.9292266606328248, "learning_rate": 3.7578162189928696e-06, "loss": 0.91, "step": 2327 }, { "epoch": 3.1123745819397994, "grad_norm": 1.001962424845012, "learning_rate": 3.7532964133691634e-06, "loss": 1.0713, "step": 2328 }, { "epoch": 3.1137123745819397, "grad_norm": 0.791620330667513, "learning_rate": 3.748777693757646e-06, "loss": 1.005, "step": 2329 }, { "epoch": 3.1150501672240805, "grad_norm": 0.8592511864488701, "learning_rate": 3.7442600640946045e-06, "loss": 0.9271, "step": 2330 }, { "epoch": 3.1163879598662207, "grad_norm": 0.9219902050072623, "learning_rate": 3.7397435283153795e-06, "loss": 0.9682, "step": 2331 }, { "epoch": 3.117725752508361, "grad_norm": 0.8814888564003621, "learning_rate": 3.735228090354354e-06, "loss": 0.8586, "step": 2332 }, { "epoch": 3.1190635451505018, "grad_norm": 0.9496653400325253, "learning_rate": 3.730713754144961e-06, "loss": 0.9621, "step": 2333 }, { "epoch": 3.120401337792642, "grad_norm": 1.1330719739823185, "learning_rate": 3.726200523619668e-06, "loss": 0.9503, "step": 2334 }, { "epoch": 3.121739130434783, "grad_norm": 0.9850281771423463, "learning_rate": 3.721688402709982e-06, "loss": 0.7379, "step": 2335 }, { "epoch": 3.123076923076923, "grad_norm": 0.9918713269849789, "learning_rate": 3.7171773953464437e-06, "loss": 0.9769, "step": 2336 }, { "epoch": 3.1244147157190634, "grad_norm": 0.7673762185083779, "learning_rate": 3.712667505458622e-06, "loss": 1.0485, "step": 2337 }, { "epoch": 3.125752508361204, "grad_norm": 0.9872475435686526, "learning_rate": 3.708158736975114e-06, "loss": 0.9736, "step": 2338 }, { "epoch": 3.1270903010033444, "grad_norm": 0.9266682563897016, "learning_rate": 3.7036510938235394e-06, "loss": 0.9607, "step": 2339 }, { "epoch": 3.128428093645485, "grad_norm": 0.8568816452550514, "learning_rate": 3.6991445799305376e-06, "loss": 1.1888, "step": 2340 }, { "epoch": 3.1297658862876254, "grad_norm": 1.1361547478793266, "learning_rate": 3.694639199221764e-06, "loss": 0.8582, "step": 2341 }, { "epoch": 3.1311036789297657, "grad_norm": 0.9282501340797914, "learning_rate": 3.690134955621885e-06, "loss": 0.7953, "step": 2342 }, { "epoch": 3.1324414715719064, "grad_norm": 0.794427580796897, "learning_rate": 3.685631853054583e-06, "loss": 1.0535, "step": 2343 }, { "epoch": 3.1337792642140467, "grad_norm": 0.9532299545727602, "learning_rate": 3.68112989544254e-06, "loss": 1.0588, "step": 2344 }, { "epoch": 3.1351170568561875, "grad_norm": 1.1153003727904418, "learning_rate": 3.6766290867074444e-06, "loss": 0.7912, "step": 2345 }, { "epoch": 3.1364548494983278, "grad_norm": 1.00659174889467, "learning_rate": 3.6721294307699786e-06, "loss": 1.0933, "step": 2346 }, { "epoch": 3.137792642140468, "grad_norm": 0.9209087105388066, "learning_rate": 3.667630931549826e-06, "loss": 1.0872, "step": 2347 }, { "epoch": 3.139130434782609, "grad_norm": 1.0699006976136751, "learning_rate": 3.6631335929656608e-06, "loss": 0.8653, "step": 2348 }, { "epoch": 3.140468227424749, "grad_norm": 0.8574746511510144, "learning_rate": 3.658637418935146e-06, "loss": 0.952, "step": 2349 }, { "epoch": 3.14180602006689, "grad_norm": 0.8379533595931271, "learning_rate": 3.6541424133749293e-06, "loss": 0.8323, "step": 2350 }, { "epoch": 3.14314381270903, "grad_norm": 0.97699823473477, "learning_rate": 3.6496485802006433e-06, "loss": 1.1314, "step": 2351 }, { "epoch": 3.1444816053511704, "grad_norm": 0.9865025748403312, "learning_rate": 3.645155923326893e-06, "loss": 1.1173, "step": 2352 }, { "epoch": 3.145819397993311, "grad_norm": 0.8231086501128544, "learning_rate": 3.640664446667268e-06, "loss": 0.9596, "step": 2353 }, { "epoch": 3.1471571906354514, "grad_norm": 0.8445925701247757, "learning_rate": 3.6361741541343242e-06, "loss": 1.1145, "step": 2354 }, { "epoch": 3.148494983277592, "grad_norm": 0.8830826522150906, "learning_rate": 3.6316850496395863e-06, "loss": 1.2565, "step": 2355 }, { "epoch": 3.1498327759197324, "grad_norm": 1.0164799685465966, "learning_rate": 3.6271971370935432e-06, "loss": 0.8678, "step": 2356 }, { "epoch": 3.1511705685618727, "grad_norm": 1.1977478266600015, "learning_rate": 3.622710420405647e-06, "loss": 0.9285, "step": 2357 }, { "epoch": 3.1525083612040135, "grad_norm": 1.0110652189344689, "learning_rate": 3.61822490348431e-06, "loss": 0.9322, "step": 2358 }, { "epoch": 3.1538461538461537, "grad_norm": 0.9909772107913533, "learning_rate": 3.613740590236895e-06, "loss": 0.7669, "step": 2359 }, { "epoch": 3.1551839464882945, "grad_norm": 1.2369516710375952, "learning_rate": 3.6092574845697193e-06, "loss": 1.2352, "step": 2360 }, { "epoch": 3.1565217391304348, "grad_norm": 1.1021334237501235, "learning_rate": 3.6047755903880478e-06, "loss": 0.8673, "step": 2361 }, { "epoch": 3.157859531772575, "grad_norm": 0.9891565481547364, "learning_rate": 3.6002949115960884e-06, "loss": 1.0448, "step": 2362 }, { "epoch": 3.159197324414716, "grad_norm": 1.1701706251938715, "learning_rate": 3.595815452096994e-06, "loss": 1.001, "step": 2363 }, { "epoch": 3.160535117056856, "grad_norm": 0.8075202371665549, "learning_rate": 3.5913372157928515e-06, "loss": 1.0273, "step": 2364 }, { "epoch": 3.161872909698997, "grad_norm": 0.889543601013753, "learning_rate": 3.5868602065846846e-06, "loss": 0.8611, "step": 2365 }, { "epoch": 3.163210702341137, "grad_norm": 1.1932872005398552, "learning_rate": 3.5823844283724464e-06, "loss": 0.8867, "step": 2366 }, { "epoch": 3.1645484949832774, "grad_norm": 0.8847454043796521, "learning_rate": 3.577909885055019e-06, "loss": 0.9613, "step": 2367 }, { "epoch": 3.165886287625418, "grad_norm": 1.0528590797300794, "learning_rate": 3.573436580530208e-06, "loss": 0.9741, "step": 2368 }, { "epoch": 3.1672240802675584, "grad_norm": 0.8204545529590835, "learning_rate": 3.56896451869474e-06, "loss": 0.9181, "step": 2369 }, { "epoch": 3.168561872909699, "grad_norm": 0.9773930825420676, "learning_rate": 3.56449370344426e-06, "loss": 0.8454, "step": 2370 }, { "epoch": 3.1698996655518394, "grad_norm": 1.0087294106207274, "learning_rate": 3.560024138673326e-06, "loss": 1.2045, "step": 2371 }, { "epoch": 3.1712374581939797, "grad_norm": 0.8983022308775929, "learning_rate": 3.5555558282754045e-06, "loss": 1.0419, "step": 2372 }, { "epoch": 3.1725752508361205, "grad_norm": 1.201816573695687, "learning_rate": 3.5510887761428764e-06, "loss": 1.1034, "step": 2373 }, { "epoch": 3.1739130434782608, "grad_norm": 1.0716457169768858, "learning_rate": 3.546622986167021e-06, "loss": 0.9419, "step": 2374 }, { "epoch": 3.1752508361204015, "grad_norm": 0.8464069959889379, "learning_rate": 3.5421584622380167e-06, "loss": 1.1396, "step": 2375 }, { "epoch": 3.1765886287625418, "grad_norm": 0.8481491095556335, "learning_rate": 3.5376952082449425e-06, "loss": 1.0077, "step": 2376 }, { "epoch": 3.177926421404682, "grad_norm": 0.8368049269188386, "learning_rate": 3.5332332280757706e-06, "loss": 1.0746, "step": 2377 }, { "epoch": 3.179264214046823, "grad_norm": 0.9020742372196259, "learning_rate": 3.5287725256173627e-06, "loss": 1.1028, "step": 2378 }, { "epoch": 3.180602006688963, "grad_norm": 1.2499250273190887, "learning_rate": 3.524313104755468e-06, "loss": 0.9001, "step": 2379 }, { "epoch": 3.181939799331104, "grad_norm": 1.2434671045553947, "learning_rate": 3.5198549693747185e-06, "loss": 0.8936, "step": 2380 }, { "epoch": 3.183277591973244, "grad_norm": 0.8169537208052383, "learning_rate": 3.5153981233586277e-06, "loss": 0.9992, "step": 2381 }, { "epoch": 3.184615384615385, "grad_norm": 1.0102779396460306, "learning_rate": 3.510942570589583e-06, "loss": 0.8721, "step": 2382 }, { "epoch": 3.185953177257525, "grad_norm": 1.042890780675582, "learning_rate": 3.5064883149488505e-06, "loss": 0.8906, "step": 2383 }, { "epoch": 3.1872909698996654, "grad_norm": 1.2140328419799882, "learning_rate": 3.5020353603165634e-06, "loss": 0.9106, "step": 2384 }, { "epoch": 3.188628762541806, "grad_norm": 0.9133225495684494, "learning_rate": 3.4975837105717203e-06, "loss": 0.9598, "step": 2385 }, { "epoch": 3.1899665551839465, "grad_norm": 1.0335438442832354, "learning_rate": 3.4931333695921843e-06, "loss": 0.8615, "step": 2386 }, { "epoch": 3.1913043478260867, "grad_norm": 1.1097601697434951, "learning_rate": 3.488684341254679e-06, "loss": 0.6517, "step": 2387 }, { "epoch": 3.1926421404682275, "grad_norm": 1.105118016079598, "learning_rate": 3.484236629434783e-06, "loss": 0.9281, "step": 2388 }, { "epoch": 3.1939799331103678, "grad_norm": 0.9249947908550734, "learning_rate": 3.4797902380069305e-06, "loss": 0.8841, "step": 2389 }, { "epoch": 3.1953177257525085, "grad_norm": 1.0239887910945935, "learning_rate": 3.475345170844403e-06, "loss": 0.8969, "step": 2390 }, { "epoch": 3.196655518394649, "grad_norm": 0.9180205146000875, "learning_rate": 3.4709014318193298e-06, "loss": 1.1486, "step": 2391 }, { "epoch": 3.1979933110367895, "grad_norm": 1.0152671213305413, "learning_rate": 3.466459024802682e-06, "loss": 0.9357, "step": 2392 }, { "epoch": 3.19933110367893, "grad_norm": 0.8908087465740738, "learning_rate": 3.4620179536642727e-06, "loss": 1.0629, "step": 2393 }, { "epoch": 3.20066889632107, "grad_norm": 0.7876195037069597, "learning_rate": 3.4575782222727507e-06, "loss": 0.9634, "step": 2394 }, { "epoch": 3.202006688963211, "grad_norm": 0.9599090500406533, "learning_rate": 3.453139834495596e-06, "loss": 0.8506, "step": 2395 }, { "epoch": 3.203344481605351, "grad_norm": 1.0034915279884615, "learning_rate": 3.448702794199118e-06, "loss": 0.7569, "step": 2396 }, { "epoch": 3.2046822742474914, "grad_norm": 1.032914605290529, "learning_rate": 3.4442671052484545e-06, "loss": 0.9482, "step": 2397 }, { "epoch": 3.206020066889632, "grad_norm": 0.9320332284569552, "learning_rate": 3.439832771507565e-06, "loss": 0.8418, "step": 2398 }, { "epoch": 3.2073578595317724, "grad_norm": 1.1725004931340859, "learning_rate": 3.4353997968392295e-06, "loss": 1.2198, "step": 2399 }, { "epoch": 3.208695652173913, "grad_norm": 0.8996895325541554, "learning_rate": 3.4309681851050414e-06, "loss": 0.9908, "step": 2400 }, { "epoch": 3.2100334448160535, "grad_norm": 1.0303536775526907, "learning_rate": 3.4265379401654096e-06, "loss": 0.8098, "step": 2401 }, { "epoch": 3.211371237458194, "grad_norm": 1.1166885040215047, "learning_rate": 3.4221090658795484e-06, "loss": 1.0373, "step": 2402 }, { "epoch": 3.2127090301003345, "grad_norm": 0.9236814107096714, "learning_rate": 3.4176815661054884e-06, "loss": 0.903, "step": 2403 }, { "epoch": 3.2140468227424748, "grad_norm": 1.0327699155287526, "learning_rate": 3.4132554447000487e-06, "loss": 0.828, "step": 2404 }, { "epoch": 3.2153846153846155, "grad_norm": 0.9558709805386082, "learning_rate": 3.4088307055188574e-06, "loss": 1.0265, "step": 2405 }, { "epoch": 3.216722408026756, "grad_norm": 0.9316353217298369, "learning_rate": 3.4044073524163344e-06, "loss": 0.8285, "step": 2406 }, { "epoch": 3.218060200668896, "grad_norm": 1.2399484399057912, "learning_rate": 3.3999853892456945e-06, "loss": 1.1503, "step": 2407 }, { "epoch": 3.219397993311037, "grad_norm": 0.9381198630750133, "learning_rate": 3.3955648198589407e-06, "loss": 0.9076, "step": 2408 }, { "epoch": 3.220735785953177, "grad_norm": 0.8178663597223617, "learning_rate": 3.3911456481068613e-06, "loss": 0.9137, "step": 2409 }, { "epoch": 3.222073578595318, "grad_norm": 0.8323616295055155, "learning_rate": 3.386727877839027e-06, "loss": 0.971, "step": 2410 }, { "epoch": 3.223411371237458, "grad_norm": 0.9056302868259969, "learning_rate": 3.3823115129037897e-06, "loss": 0.9884, "step": 2411 }, { "epoch": 3.224749163879599, "grad_norm": 1.1498377805629294, "learning_rate": 3.3778965571482723e-06, "loss": 0.9259, "step": 2412 }, { "epoch": 3.226086956521739, "grad_norm": 1.8279483011327384, "learning_rate": 3.3734830144183783e-06, "loss": 0.806, "step": 2413 }, { "epoch": 3.2274247491638794, "grad_norm": 1.0565908307609497, "learning_rate": 3.369070888558774e-06, "loss": 0.7789, "step": 2414 }, { "epoch": 3.22876254180602, "grad_norm": 0.8533209295578117, "learning_rate": 3.3646601834128924e-06, "loss": 0.6469, "step": 2415 }, { "epoch": 3.2301003344481605, "grad_norm": 0.8017925642947263, "learning_rate": 3.360250902822929e-06, "loss": 1.0516, "step": 2416 }, { "epoch": 3.231438127090301, "grad_norm": 1.266017154263993, "learning_rate": 3.35584305062984e-06, "loss": 1.166, "step": 2417 }, { "epoch": 3.2327759197324415, "grad_norm": 1.1487158745157784, "learning_rate": 3.3514366306733348e-06, "loss": 1.0631, "step": 2418 }, { "epoch": 3.234113712374582, "grad_norm": 0.9012652112376609, "learning_rate": 3.3470316467918785e-06, "loss": 1.206, "step": 2419 }, { "epoch": 3.2354515050167225, "grad_norm": 6.1854486078230435, "learning_rate": 3.3426281028226817e-06, "loss": 1.1729, "step": 2420 }, { "epoch": 3.236789297658863, "grad_norm": 1.2097691115118259, "learning_rate": 3.3382260026017027e-06, "loss": 0.9613, "step": 2421 }, { "epoch": 3.2381270903010035, "grad_norm": 0.9677751174837788, "learning_rate": 3.3338253499636407e-06, "loss": 0.9866, "step": 2422 }, { "epoch": 3.239464882943144, "grad_norm": 0.8580660361854529, "learning_rate": 3.329426148741937e-06, "loss": 1.0396, "step": 2423 }, { "epoch": 3.240802675585284, "grad_norm": 1.7381441806074633, "learning_rate": 3.3250284027687652e-06, "loss": 0.9024, "step": 2424 }, { "epoch": 3.242140468227425, "grad_norm": 0.9687627000483536, "learning_rate": 3.320632115875033e-06, "loss": 0.9586, "step": 2425 }, { "epoch": 3.243478260869565, "grad_norm": 1.0242094140669546, "learning_rate": 3.3162372918903764e-06, "loss": 0.8951, "step": 2426 }, { "epoch": 3.244816053511706, "grad_norm": 1.1279603587339548, "learning_rate": 3.311843934643157e-06, "loss": 0.9969, "step": 2427 }, { "epoch": 3.246153846153846, "grad_norm": 0.9194954494014211, "learning_rate": 3.307452047960459e-06, "loss": 0.5939, "step": 2428 }, { "epoch": 3.2474916387959865, "grad_norm": 1.334544541653259, "learning_rate": 3.3030616356680854e-06, "loss": 0.9216, "step": 2429 }, { "epoch": 3.248829431438127, "grad_norm": 0.8803916678002883, "learning_rate": 3.298672701590555e-06, "loss": 1.1308, "step": 2430 }, { "epoch": 3.2501672240802675, "grad_norm": 0.9458052484671156, "learning_rate": 3.2942852495510992e-06, "loss": 0.913, "step": 2431 }, { "epoch": 3.251505016722408, "grad_norm": 0.9369076525944816, "learning_rate": 3.289899283371657e-06, "loss": 0.7865, "step": 2432 }, { "epoch": 3.2528428093645485, "grad_norm": 1.1817359695151943, "learning_rate": 3.2855148068728753e-06, "loss": 1.0302, "step": 2433 }, { "epoch": 3.254180602006689, "grad_norm": 1.1611655042831446, "learning_rate": 3.2811318238741026e-06, "loss": 0.9445, "step": 2434 }, { "epoch": 3.2555183946488295, "grad_norm": 1.2225039541550322, "learning_rate": 3.276750338193385e-06, "loss": 0.9116, "step": 2435 }, { "epoch": 3.25685618729097, "grad_norm": 1.009074077094541, "learning_rate": 3.272370353647465e-06, "loss": 0.9599, "step": 2436 }, { "epoch": 3.2581939799331106, "grad_norm": 0.8726304265576502, "learning_rate": 3.2679918740517785e-06, "loss": 1.0028, "step": 2437 }, { "epoch": 3.259531772575251, "grad_norm": 1.135634035679661, "learning_rate": 3.263614903220449e-06, "loss": 0.7423, "step": 2438 }, { "epoch": 3.260869565217391, "grad_norm": 0.9329620402524612, "learning_rate": 3.2592394449662867e-06, "loss": 1.1575, "step": 2439 }, { "epoch": 3.262207357859532, "grad_norm": 1.075104572911778, "learning_rate": 3.2548655031007837e-06, "loss": 1.004, "step": 2440 }, { "epoch": 3.263545150501672, "grad_norm": 0.7007476097414075, "learning_rate": 3.250493081434112e-06, "loss": 1.1221, "step": 2441 }, { "epoch": 3.264882943143813, "grad_norm": 0.9753483281854181, "learning_rate": 3.2461221837751146e-06, "loss": 0.6864, "step": 2442 }, { "epoch": 3.266220735785953, "grad_norm": 1.406136170157171, "learning_rate": 3.241752813931316e-06, "loss": 0.7899, "step": 2443 }, { "epoch": 3.2675585284280935, "grad_norm": 1.2059443515812682, "learning_rate": 3.237384975708904e-06, "loss": 1.1211, "step": 2444 }, { "epoch": 3.268896321070234, "grad_norm": 1.1132524077308281, "learning_rate": 3.233018672912731e-06, "loss": 0.833, "step": 2445 }, { "epoch": 3.2702341137123745, "grad_norm": 0.9704560739595096, "learning_rate": 3.228653909346314e-06, "loss": 1.1046, "step": 2446 }, { "epoch": 3.2715719063545152, "grad_norm": 0.8481700806675437, "learning_rate": 3.224290688811831e-06, "loss": 0.8544, "step": 2447 }, { "epoch": 3.2729096989966555, "grad_norm": 0.7270555326991734, "learning_rate": 3.2199290151101115e-06, "loss": 0.9505, "step": 2448 }, { "epoch": 3.274247491638796, "grad_norm": 0.8549112650591137, "learning_rate": 3.2155688920406415e-06, "loss": 0.925, "step": 2449 }, { "epoch": 3.2755852842809365, "grad_norm": 1.0331284051263003, "learning_rate": 3.2112103234015535e-06, "loss": 0.8699, "step": 2450 }, { "epoch": 3.276923076923077, "grad_norm": 1.0009264089487173, "learning_rate": 3.2068533129896273e-06, "loss": 1.0949, "step": 2451 }, { "epoch": 3.2782608695652176, "grad_norm": 1.0384995277433373, "learning_rate": 3.2024978646002848e-06, "loss": 0.7674, "step": 2452 }, { "epoch": 3.279598662207358, "grad_norm": 1.1441585028835612, "learning_rate": 3.1981439820275883e-06, "loss": 1.1118, "step": 2453 }, { "epoch": 3.280936454849498, "grad_norm": 1.033547291185661, "learning_rate": 3.1937916690642356e-06, "loss": 1.0041, "step": 2454 }, { "epoch": 3.282274247491639, "grad_norm": 0.8737082723731726, "learning_rate": 3.189440929501556e-06, "loss": 1.1036, "step": 2455 }, { "epoch": 3.283612040133779, "grad_norm": 1.2905772628923033, "learning_rate": 3.185091767129509e-06, "loss": 0.911, "step": 2456 }, { "epoch": 3.28494983277592, "grad_norm": 0.8163112680241311, "learning_rate": 3.1807441857366798e-06, "loss": 1.0118, "step": 2457 }, { "epoch": 3.28628762541806, "grad_norm": 0.7412500440701302, "learning_rate": 3.1763981891102785e-06, "loss": 0.9435, "step": 2458 }, { "epoch": 3.2876254180602005, "grad_norm": 1.0737102587338974, "learning_rate": 3.172053781036132e-06, "loss": 1.0366, "step": 2459 }, { "epoch": 3.288963210702341, "grad_norm": 1.0268044492919626, "learning_rate": 3.167710965298684e-06, "loss": 1.0558, "step": 2460 }, { "epoch": 3.2903010033444815, "grad_norm": 1.1987236897530682, "learning_rate": 3.1633697456809932e-06, "loss": 0.8332, "step": 2461 }, { "epoch": 3.2916387959866222, "grad_norm": 1.370760148400418, "learning_rate": 3.159030125964723e-06, "loss": 0.9726, "step": 2462 }, { "epoch": 3.2929765886287625, "grad_norm": 0.9613985236519932, "learning_rate": 3.1546921099301507e-06, "loss": 0.9209, "step": 2463 }, { "epoch": 3.294314381270903, "grad_norm": 0.8480628079044641, "learning_rate": 3.15035570135615e-06, "loss": 0.9295, "step": 2464 }, { "epoch": 3.2956521739130435, "grad_norm": 0.9893510606722586, "learning_rate": 3.1460209040201967e-06, "loss": 0.9513, "step": 2465 }, { "epoch": 3.296989966555184, "grad_norm": 1.0095565134892797, "learning_rate": 3.141687721698363e-06, "loss": 0.961, "step": 2466 }, { "epoch": 3.2983277591973246, "grad_norm": 0.9292828158737918, "learning_rate": 3.1373561581653152e-06, "loss": 0.8379, "step": 2467 }, { "epoch": 3.299665551839465, "grad_norm": 0.9052301340424151, "learning_rate": 3.1330262171943073e-06, "loss": 0.8895, "step": 2468 }, { "epoch": 3.3010033444816056, "grad_norm": 1.170582246748635, "learning_rate": 3.1286979025571817e-06, "loss": 1.1792, "step": 2469 }, { "epoch": 3.302341137123746, "grad_norm": 0.8312784595481998, "learning_rate": 3.1243712180243633e-06, "loss": 0.8264, "step": 2470 }, { "epoch": 3.303678929765886, "grad_norm": 0.9541031452211025, "learning_rate": 3.120046167364857e-06, "loss": 1.1396, "step": 2471 }, { "epoch": 3.305016722408027, "grad_norm": 0.9907301864346499, "learning_rate": 3.1157227543462428e-06, "loss": 0.9008, "step": 2472 }, { "epoch": 3.306354515050167, "grad_norm": 0.8880990744236433, "learning_rate": 3.11140098273468e-06, "loss": 0.9939, "step": 2473 }, { "epoch": 3.3076923076923075, "grad_norm": 1.040443294630524, "learning_rate": 3.107080856294892e-06, "loss": 0.6927, "step": 2474 }, { "epoch": 3.309030100334448, "grad_norm": 0.8997277159750869, "learning_rate": 3.1027623787901706e-06, "loss": 0.9264, "step": 2475 }, { "epoch": 3.3103678929765885, "grad_norm": 1.0535420119358403, "learning_rate": 3.098445553982372e-06, "loss": 1.1213, "step": 2476 }, { "epoch": 3.3117056856187292, "grad_norm": 0.7920158992881686, "learning_rate": 3.0941303856319126e-06, "loss": 0.7492, "step": 2477 }, { "epoch": 3.3130434782608695, "grad_norm": 0.8638423187209764, "learning_rate": 3.0898168774977654e-06, "loss": 0.9158, "step": 2478 }, { "epoch": 3.3143812709030103, "grad_norm": 1.0597874363825717, "learning_rate": 3.0855050333374574e-06, "loss": 0.7531, "step": 2479 }, { "epoch": 3.3157190635451506, "grad_norm": 0.8532559046248658, "learning_rate": 3.0811948569070666e-06, "loss": 0.9397, "step": 2480 }, { "epoch": 3.317056856187291, "grad_norm": 1.3245491822406341, "learning_rate": 3.076886351961217e-06, "loss": 1.0035, "step": 2481 }, { "epoch": 3.3183946488294316, "grad_norm": 0.8010061523862972, "learning_rate": 3.072579522253076e-06, "loss": 0.9585, "step": 2482 }, { "epoch": 3.319732441471572, "grad_norm": 1.4147426694309815, "learning_rate": 3.0682743715343565e-06, "loss": 0.8623, "step": 2483 }, { "epoch": 3.321070234113712, "grad_norm": 1.511236693678775, "learning_rate": 3.063970903555304e-06, "loss": 1.2242, "step": 2484 }, { "epoch": 3.322408026755853, "grad_norm": 0.9900518344930515, "learning_rate": 3.0596691220646978e-06, "loss": 1.0069, "step": 2485 }, { "epoch": 3.323745819397993, "grad_norm": 1.0203139816044606, "learning_rate": 3.0553690308098517e-06, "loss": 1.1393, "step": 2486 }, { "epoch": 3.325083612040134, "grad_norm": 1.0697057121414144, "learning_rate": 3.0510706335366034e-06, "loss": 0.8366, "step": 2487 }, { "epoch": 3.326421404682274, "grad_norm": 0.9769181035230453, "learning_rate": 3.046773933989319e-06, "loss": 1.0082, "step": 2488 }, { "epoch": 3.327759197324415, "grad_norm": 0.8750706616967903, "learning_rate": 3.042478935910881e-06, "loss": 0.8664, "step": 2489 }, { "epoch": 3.3290969899665552, "grad_norm": 1.0131301557645256, "learning_rate": 3.0381856430426935e-06, "loss": 1.1373, "step": 2490 }, { "epoch": 3.3304347826086955, "grad_norm": 0.7361299172746528, "learning_rate": 3.033894059124675e-06, "loss": 0.9108, "step": 2491 }, { "epoch": 3.3317725752508363, "grad_norm": 1.162724540192755, "learning_rate": 3.0296041878952497e-06, "loss": 1.064, "step": 2492 }, { "epoch": 3.3331103678929765, "grad_norm": 0.9625213462606929, "learning_rate": 3.02531603309136e-06, "loss": 1.0056, "step": 2493 }, { "epoch": 3.334448160535117, "grad_norm": 1.2434115950967894, "learning_rate": 3.0210295984484446e-06, "loss": 1.0654, "step": 2494 }, { "epoch": 3.3357859531772576, "grad_norm": 1.122603520628483, "learning_rate": 3.016744887700447e-06, "loss": 1.0339, "step": 2495 }, { "epoch": 3.337123745819398, "grad_norm": 1.1948714140055905, "learning_rate": 3.0124619045798087e-06, "loss": 0.8189, "step": 2496 }, { "epoch": 3.3384615384615386, "grad_norm": 1.196122355506986, "learning_rate": 3.0081806528174655e-06, "loss": 1.1724, "step": 2497 }, { "epoch": 3.339799331103679, "grad_norm": 1.186839033043405, "learning_rate": 3.0039011361428466e-06, "loss": 0.8375, "step": 2498 }, { "epoch": 3.3411371237458196, "grad_norm": 1.0128888329131995, "learning_rate": 2.9996233582838686e-06, "loss": 1.1077, "step": 2499 }, { "epoch": 3.34247491638796, "grad_norm": 0.7148409186161675, "learning_rate": 2.995347322966933e-06, "loss": 0.942, "step": 2500 }, { "epoch": 3.3438127090301, "grad_norm": 0.8846433912663603, "learning_rate": 2.9910730339169245e-06, "loss": 1.0979, "step": 2501 }, { "epoch": 3.345150501672241, "grad_norm": 1.0013639262761973, "learning_rate": 2.9868004948572044e-06, "loss": 0.9148, "step": 2502 }, { "epoch": 3.346488294314381, "grad_norm": 1.2091338165133347, "learning_rate": 2.982529709509615e-06, "loss": 0.9816, "step": 2503 }, { "epoch": 3.3478260869565215, "grad_norm": 0.7979188151590147, "learning_rate": 2.978260681594465e-06, "loss": 0.911, "step": 2504 }, { "epoch": 3.3491638795986622, "grad_norm": 1.0696725599048187, "learning_rate": 2.973993414830534e-06, "loss": 0.9803, "step": 2505 }, { "epoch": 3.3505016722408025, "grad_norm": 1.1583912070466795, "learning_rate": 2.9697279129350686e-06, "loss": 1.035, "step": 2506 }, { "epoch": 3.3518394648829433, "grad_norm": 1.0395177749554907, "learning_rate": 2.965464179623777e-06, "loss": 1.1188, "step": 2507 }, { "epoch": 3.3531772575250836, "grad_norm": 0.9052046284123316, "learning_rate": 2.9612022186108267e-06, "loss": 0.797, "step": 2508 }, { "epoch": 3.3545150501672243, "grad_norm": 1.0939463393390496, "learning_rate": 2.956942033608843e-06, "loss": 1.0211, "step": 2509 }, { "epoch": 3.3558528428093646, "grad_norm": 1.0172460375034797, "learning_rate": 2.952683628328901e-06, "loss": 0.7776, "step": 2510 }, { "epoch": 3.357190635451505, "grad_norm": 0.9109644076224833, "learning_rate": 2.948427006480528e-06, "loss": 0.9864, "step": 2511 }, { "epoch": 3.3585284280936456, "grad_norm": 1.2013938856694961, "learning_rate": 2.9441721717716966e-06, "loss": 0.6534, "step": 2512 }, { "epoch": 3.359866220735786, "grad_norm": 1.384503337175571, "learning_rate": 2.9399191279088236e-06, "loss": 1.0412, "step": 2513 }, { "epoch": 3.361204013377926, "grad_norm": 1.2390093480895572, "learning_rate": 2.9356678785967646e-06, "loss": 0.9292, "step": 2514 }, { "epoch": 3.362541806020067, "grad_norm": 1.008564878080212, "learning_rate": 2.9314184275388134e-06, "loss": 1.0757, "step": 2515 }, { "epoch": 3.363879598662207, "grad_norm": 1.027020428446574, "learning_rate": 2.9271707784366952e-06, "loss": 0.9587, "step": 2516 }, { "epoch": 3.365217391304348, "grad_norm": 1.009745916654776, "learning_rate": 2.9229249349905686e-06, "loss": 1.0138, "step": 2517 }, { "epoch": 3.3665551839464882, "grad_norm": 0.9840554661008071, "learning_rate": 2.918680900899017e-06, "loss": 0.9239, "step": 2518 }, { "epoch": 3.367892976588629, "grad_norm": 1.0086494124799652, "learning_rate": 2.914438679859046e-06, "loss": 0.5756, "step": 2519 }, { "epoch": 3.3692307692307693, "grad_norm": 0.9673177419512934, "learning_rate": 2.910198275566085e-06, "loss": 0.8701, "step": 2520 }, { "epoch": 3.3705685618729095, "grad_norm": 1.21529135723859, "learning_rate": 2.9059596917139804e-06, "loss": 0.9399, "step": 2521 }, { "epoch": 3.3719063545150503, "grad_norm": 1.0877902649121924, "learning_rate": 2.9017229319949897e-06, "loss": 0.7422, "step": 2522 }, { "epoch": 3.3732441471571906, "grad_norm": 1.0430075860166066, "learning_rate": 2.897488000099788e-06, "loss": 0.8929, "step": 2523 }, { "epoch": 3.374581939799331, "grad_norm": 0.8394478038341568, "learning_rate": 2.893254899717452e-06, "loss": 1.0496, "step": 2524 }, { "epoch": 3.3759197324414716, "grad_norm": 0.9475425881785573, "learning_rate": 2.8890236345354648e-06, "loss": 0.7295, "step": 2525 }, { "epoch": 3.377257525083612, "grad_norm": 0.8547896271412092, "learning_rate": 2.8847942082397112e-06, "loss": 1.06, "step": 2526 }, { "epoch": 3.3785953177257526, "grad_norm": 1.002792556814478, "learning_rate": 2.8805666245144735e-06, "loss": 0.7466, "step": 2527 }, { "epoch": 3.379933110367893, "grad_norm": 0.9072226095472068, "learning_rate": 2.8763408870424305e-06, "loss": 1.0124, "step": 2528 }, { "epoch": 3.3812709030100336, "grad_norm": 0.7950165110467008, "learning_rate": 2.8721169995046503e-06, "loss": 1.0298, "step": 2529 }, { "epoch": 3.382608695652174, "grad_norm": 1.0795696910112553, "learning_rate": 2.8678949655805915e-06, "loss": 1.0811, "step": 2530 }, { "epoch": 3.383946488294314, "grad_norm": 1.1462512195266532, "learning_rate": 2.863674788948097e-06, "loss": 0.8376, "step": 2531 }, { "epoch": 3.385284280936455, "grad_norm": 0.9660432762675125, "learning_rate": 2.85945647328339e-06, "loss": 1.142, "step": 2532 }, { "epoch": 3.3866220735785952, "grad_norm": 0.7974930819566493, "learning_rate": 2.8552400222610788e-06, "loss": 1.0876, "step": 2533 }, { "epoch": 3.387959866220736, "grad_norm": 0.8200071970160381, "learning_rate": 2.851025439554142e-06, "loss": 0.9373, "step": 2534 }, { "epoch": 3.3892976588628763, "grad_norm": 0.9235177849085524, "learning_rate": 2.846812728833931e-06, "loss": 0.9947, "step": 2535 }, { "epoch": 3.3906354515050166, "grad_norm": 1.0119940392099256, "learning_rate": 2.8426018937701678e-06, "loss": 0.9617, "step": 2536 }, { "epoch": 3.3919732441471573, "grad_norm": 1.0648266339946815, "learning_rate": 2.8383929380309406e-06, "loss": 0.9199, "step": 2537 }, { "epoch": 3.3933110367892976, "grad_norm": 0.8985717192511151, "learning_rate": 2.834185865282699e-06, "loss": 0.9008, "step": 2538 }, { "epoch": 3.3946488294314383, "grad_norm": 1.0843509939230773, "learning_rate": 2.829980679190254e-06, "loss": 0.9785, "step": 2539 }, { "epoch": 3.3959866220735786, "grad_norm": 1.039942868078978, "learning_rate": 2.8257773834167736e-06, "loss": 0.9884, "step": 2540 }, { "epoch": 3.397324414715719, "grad_norm": 0.9872186843038104, "learning_rate": 2.8215759816237748e-06, "loss": 1.0364, "step": 2541 }, { "epoch": 3.3986622073578596, "grad_norm": 0.9570300114523181, "learning_rate": 2.817376477471132e-06, "loss": 0.954, "step": 2542 }, { "epoch": 3.4, "grad_norm": 1.4551108976478087, "learning_rate": 2.8131788746170612e-06, "loss": 0.8223, "step": 2543 }, { "epoch": 3.4013377926421406, "grad_norm": 0.9708101817641233, "learning_rate": 2.808983176718125e-06, "loss": 0.9775, "step": 2544 }, { "epoch": 3.402675585284281, "grad_norm": 0.8630020057329093, "learning_rate": 2.804789387429222e-06, "loss": 0.7488, "step": 2545 }, { "epoch": 3.4040133779264212, "grad_norm": 0.881185849521178, "learning_rate": 2.800597510403592e-06, "loss": 1.0118, "step": 2546 }, { "epoch": 3.405351170568562, "grad_norm": 1.1696780026897742, "learning_rate": 2.796407549292809e-06, "loss": 1.0132, "step": 2547 }, { "epoch": 3.4066889632107022, "grad_norm": 1.1302299256117632, "learning_rate": 2.792219507746777e-06, "loss": 0.9397, "step": 2548 }, { "epoch": 3.408026755852843, "grad_norm": 0.953138130905063, "learning_rate": 2.788033389413729e-06, "loss": 1.1177, "step": 2549 }, { "epoch": 3.4093645484949833, "grad_norm": 1.1418709876412967, "learning_rate": 2.7838491979402205e-06, "loss": 1.0098, "step": 2550 }, { "epoch": 3.4107023411371236, "grad_norm": 1.2751517255259077, "learning_rate": 2.7796669369711294e-06, "loss": 1.0529, "step": 2551 }, { "epoch": 3.4120401337792643, "grad_norm": 0.7514601859608498, "learning_rate": 2.7754866101496558e-06, "loss": 1.0766, "step": 2552 }, { "epoch": 3.4133779264214046, "grad_norm": 0.9138762693618729, "learning_rate": 2.771308221117309e-06, "loss": 0.8973, "step": 2553 }, { "epoch": 3.4147157190635453, "grad_norm": 1.0109090140321113, "learning_rate": 2.7671317735139136e-06, "loss": 1.1036, "step": 2554 }, { "epoch": 3.4160535117056856, "grad_norm": 1.17167176838633, "learning_rate": 2.762957270977602e-06, "loss": 0.7949, "step": 2555 }, { "epoch": 3.417391304347826, "grad_norm": 1.650855987171061, "learning_rate": 2.758784717144812e-06, "loss": 0.8792, "step": 2556 }, { "epoch": 3.4187290969899666, "grad_norm": 0.8239226043342552, "learning_rate": 2.754614115650285e-06, "loss": 1.1659, "step": 2557 }, { "epoch": 3.420066889632107, "grad_norm": 0.778959014569038, "learning_rate": 2.7504454701270604e-06, "loss": 1.0357, "step": 2558 }, { "epoch": 3.4214046822742477, "grad_norm": 0.7040834220428136, "learning_rate": 2.7462787842064753e-06, "loss": 1.0481, "step": 2559 }, { "epoch": 3.422742474916388, "grad_norm": 1.2173422438314223, "learning_rate": 2.742114061518157e-06, "loss": 0.8994, "step": 2560 }, { "epoch": 3.4240802675585282, "grad_norm": 0.8178650076591513, "learning_rate": 2.7379513056900254e-06, "loss": 0.9766, "step": 2561 }, { "epoch": 3.425418060200669, "grad_norm": 0.8534651295115426, "learning_rate": 2.7337905203482884e-06, "loss": 1.208, "step": 2562 }, { "epoch": 3.4267558528428093, "grad_norm": 1.0272873032813963, "learning_rate": 2.7296317091174325e-06, "loss": 1.252, "step": 2563 }, { "epoch": 3.42809364548495, "grad_norm": 0.8472319080372678, "learning_rate": 2.725474875620228e-06, "loss": 0.939, "step": 2564 }, { "epoch": 3.4294314381270903, "grad_norm": 1.1346166980655121, "learning_rate": 2.7213200234777215e-06, "loss": 1.0015, "step": 2565 }, { "epoch": 3.430769230769231, "grad_norm": 1.0302672384833151, "learning_rate": 2.717167156309234e-06, "loss": 0.8438, "step": 2566 }, { "epoch": 3.4321070234113713, "grad_norm": 1.0718516911593854, "learning_rate": 2.7130162777323567e-06, "loss": 1.2175, "step": 2567 }, { "epoch": 3.4334448160535116, "grad_norm": 1.043592202410892, "learning_rate": 2.708867391362948e-06, "loss": 1.0544, "step": 2568 }, { "epoch": 3.4347826086956523, "grad_norm": 0.953944871141496, "learning_rate": 2.7047205008151332e-06, "loss": 0.9827, "step": 2569 }, { "epoch": 3.4361204013377926, "grad_norm": 1.0257756852818092, "learning_rate": 2.700575609701298e-06, "loss": 1.1416, "step": 2570 }, { "epoch": 3.437458193979933, "grad_norm": 0.8736915666094205, "learning_rate": 2.696432721632082e-06, "loss": 0.925, "step": 2571 }, { "epoch": 3.4387959866220736, "grad_norm": 0.7416142373333507, "learning_rate": 2.692291840216389e-06, "loss": 1.0213, "step": 2572 }, { "epoch": 3.440133779264214, "grad_norm": 1.0065679209341245, "learning_rate": 2.6881529690613687e-06, "loss": 0.8979, "step": 2573 }, { "epoch": 3.4414715719063547, "grad_norm": 0.9427641091650956, "learning_rate": 2.6840161117724184e-06, "loss": 0.9529, "step": 2574 }, { "epoch": 3.442809364548495, "grad_norm": 1.1159968965810934, "learning_rate": 2.6798812719531843e-06, "loss": 0.8927, "step": 2575 }, { "epoch": 3.4441471571906357, "grad_norm": 1.243904450098762, "learning_rate": 2.6757484532055537e-06, "loss": 0.7709, "step": 2576 }, { "epoch": 3.445484949832776, "grad_norm": 1.1184747591735278, "learning_rate": 2.671617659129655e-06, "loss": 0.9148, "step": 2577 }, { "epoch": 3.4468227424749163, "grad_norm": 1.1000990668883732, "learning_rate": 2.667488893323851e-06, "loss": 0.9404, "step": 2578 }, { "epoch": 3.448160535117057, "grad_norm": 1.065362098375987, "learning_rate": 2.6633621593847387e-06, "loss": 0.9887, "step": 2579 }, { "epoch": 3.4494983277591973, "grad_norm": 1.0781173320240445, "learning_rate": 2.6592374609071446e-06, "loss": 0.9998, "step": 2580 }, { "epoch": 3.4508361204013376, "grad_norm": 0.826523121767694, "learning_rate": 2.65511480148412e-06, "loss": 0.9777, "step": 2581 }, { "epoch": 3.4521739130434783, "grad_norm": 1.178450365550267, "learning_rate": 2.6509941847069466e-06, "loss": 0.9756, "step": 2582 }, { "epoch": 3.4535117056856186, "grad_norm": 0.9984543915985522, "learning_rate": 2.646875614165121e-06, "loss": 1.0619, "step": 2583 }, { "epoch": 3.4548494983277593, "grad_norm": 1.011689855104505, "learning_rate": 2.6427590934463576e-06, "loss": 0.8067, "step": 2584 }, { "epoch": 3.4561872909698996, "grad_norm": 1.2155967655630164, "learning_rate": 2.6386446261365874e-06, "loss": 0.886, "step": 2585 }, { "epoch": 3.4575250836120404, "grad_norm": 1.0141612588195372, "learning_rate": 2.6345322158199503e-06, "loss": 0.915, "step": 2586 }, { "epoch": 3.4588628762541807, "grad_norm": 0.8759902243600399, "learning_rate": 2.630421866078797e-06, "loss": 0.8263, "step": 2587 }, { "epoch": 3.460200668896321, "grad_norm": 0.946686767441225, "learning_rate": 2.626313580493681e-06, "loss": 0.9624, "step": 2588 }, { "epoch": 3.4615384615384617, "grad_norm": 1.0283610456803287, "learning_rate": 2.6222073626433587e-06, "loss": 0.9307, "step": 2589 }, { "epoch": 3.462876254180602, "grad_norm": 0.8360038734040618, "learning_rate": 2.618103216104785e-06, "loss": 0.7859, "step": 2590 }, { "epoch": 3.4642140468227423, "grad_norm": 0.99082727785375, "learning_rate": 2.6140011444531086e-06, "loss": 0.8035, "step": 2591 }, { "epoch": 3.465551839464883, "grad_norm": 1.1023645431701752, "learning_rate": 2.6099011512616767e-06, "loss": 1.0737, "step": 2592 }, { "epoch": 3.4668896321070233, "grad_norm": 1.1608595498436767, "learning_rate": 2.60580324010202e-06, "loss": 0.9986, "step": 2593 }, { "epoch": 3.468227424749164, "grad_norm": 0.8379070492517701, "learning_rate": 2.6017074145438583e-06, "loss": 0.9836, "step": 2594 }, { "epoch": 3.4695652173913043, "grad_norm": 1.0779680145811208, "learning_rate": 2.597613678155092e-06, "loss": 0.9969, "step": 2595 }, { "epoch": 3.470903010033445, "grad_norm": 0.6644419534827318, "learning_rate": 2.593522034501805e-06, "loss": 1.0829, "step": 2596 }, { "epoch": 3.4722408026755853, "grad_norm": 1.1742152479345769, "learning_rate": 2.5894324871482557e-06, "loss": 0.8863, "step": 2597 }, { "epoch": 3.4735785953177256, "grad_norm": 1.0977562781708208, "learning_rate": 2.585345039656878e-06, "loss": 0.8894, "step": 2598 }, { "epoch": 3.4749163879598663, "grad_norm": 1.225066512160864, "learning_rate": 2.5812596955882756e-06, "loss": 0.8941, "step": 2599 }, { "epoch": 3.4762541806020066, "grad_norm": 1.1294157490145151, "learning_rate": 2.5771764585012203e-06, "loss": 0.9131, "step": 2600 }, { "epoch": 3.477591973244147, "grad_norm": 1.4296318446282068, "learning_rate": 2.573095331952646e-06, "loss": 1.0907, "step": 2601 }, { "epoch": 3.4789297658862877, "grad_norm": 0.9255122437508664, "learning_rate": 2.5690163194976576e-06, "loss": 1.0411, "step": 2602 }, { "epoch": 3.480267558528428, "grad_norm": 0.9659902532974164, "learning_rate": 2.5649394246895044e-06, "loss": 0.7694, "step": 2603 }, { "epoch": 3.4816053511705687, "grad_norm": 1.0359611783281082, "learning_rate": 2.560864651079599e-06, "loss": 1.0204, "step": 2604 }, { "epoch": 3.482943143812709, "grad_norm": 0.879938395306546, "learning_rate": 2.556792002217507e-06, "loss": 0.9592, "step": 2605 }, { "epoch": 3.4842809364548497, "grad_norm": 0.9339679295307174, "learning_rate": 2.5527214816509398e-06, "loss": 1.1232, "step": 2606 }, { "epoch": 3.48561872909699, "grad_norm": 0.9096766167150621, "learning_rate": 2.5486530929257574e-06, "loss": 0.9043, "step": 2607 }, { "epoch": 3.4869565217391303, "grad_norm": 1.1476888691867142, "learning_rate": 2.544586839585961e-06, "loss": 0.9308, "step": 2608 }, { "epoch": 3.488294314381271, "grad_norm": 0.9260131318073177, "learning_rate": 2.540522725173692e-06, "loss": 0.8158, "step": 2609 }, { "epoch": 3.4896321070234113, "grad_norm": 1.3912721417737308, "learning_rate": 2.5364607532292283e-06, "loss": 0.8645, "step": 2610 }, { "epoch": 3.4909698996655516, "grad_norm": 0.9854773301476857, "learning_rate": 2.532400927290982e-06, "loss": 1.0356, "step": 2611 }, { "epoch": 3.4923076923076923, "grad_norm": 1.1010700683193784, "learning_rate": 2.5283432508954976e-06, "loss": 0.9757, "step": 2612 }, { "epoch": 3.4936454849498326, "grad_norm": 0.8799837406621035, "learning_rate": 2.5242877275774446e-06, "loss": 0.7457, "step": 2613 }, { "epoch": 3.4949832775919734, "grad_norm": 1.0006239333115725, "learning_rate": 2.520234360869617e-06, "loss": 0.8915, "step": 2614 }, { "epoch": 3.4963210702341136, "grad_norm": 0.8586049587440574, "learning_rate": 2.5161831543029314e-06, "loss": 0.9881, "step": 2615 }, { "epoch": 3.4976588628762544, "grad_norm": 0.7770249370697508, "learning_rate": 2.512134111406422e-06, "loss": 1.1822, "step": 2616 }, { "epoch": 3.4989966555183947, "grad_norm": 0.7302395422325426, "learning_rate": 2.508087235707237e-06, "loss": 1.0265, "step": 2617 }, { "epoch": 3.500334448160535, "grad_norm": 1.1078392218696895, "learning_rate": 2.5040425307306404e-06, "loss": 0.9143, "step": 2618 }, { "epoch": 3.5016722408026757, "grad_norm": 0.9864592223052762, "learning_rate": 2.5000000000000015e-06, "loss": 1.0889, "step": 2619 }, { "epoch": 3.503010033444816, "grad_norm": 1.2233928401175742, "learning_rate": 2.4959596470367965e-06, "loss": 1.0693, "step": 2620 }, { "epoch": 3.5043478260869563, "grad_norm": 0.8850624344337126, "learning_rate": 2.4919214753606043e-06, "loss": 0.9228, "step": 2621 }, { "epoch": 3.505685618729097, "grad_norm": 1.1164743796228342, "learning_rate": 2.4878854884891067e-06, "loss": 0.8725, "step": 2622 }, { "epoch": 3.5070234113712373, "grad_norm": 0.8438338923423279, "learning_rate": 2.4838516899380806e-06, "loss": 0.8596, "step": 2623 }, { "epoch": 3.508361204013378, "grad_norm": 0.9944502606973887, "learning_rate": 2.4798200832213933e-06, "loss": 1.1284, "step": 2624 }, { "epoch": 3.5096989966555183, "grad_norm": 1.2363316032161562, "learning_rate": 2.475790671851007e-06, "loss": 1.0441, "step": 2625 }, { "epoch": 3.511036789297659, "grad_norm": 1.0481420587847672, "learning_rate": 2.4717634593369704e-06, "loss": 1.0631, "step": 2626 }, { "epoch": 3.5123745819397993, "grad_norm": 1.09350132861457, "learning_rate": 2.4677384491874155e-06, "loss": 0.8755, "step": 2627 }, { "epoch": 3.5137123745819396, "grad_norm": 1.297621173505391, "learning_rate": 2.463715644908557e-06, "loss": 1.0238, "step": 2628 }, { "epoch": 3.5150501672240804, "grad_norm": 1.3189402817219391, "learning_rate": 2.459695050004688e-06, "loss": 0.6466, "step": 2629 }, { "epoch": 3.5163879598662207, "grad_norm": 0.77677396915294, "learning_rate": 2.4556766679781763e-06, "loss": 1.2467, "step": 2630 }, { "epoch": 3.517725752508361, "grad_norm": 1.1193999171772278, "learning_rate": 2.4516605023294626e-06, "loss": 1.2083, "step": 2631 }, { "epoch": 3.5190635451505017, "grad_norm": 0.9058608705170266, "learning_rate": 2.447646556557057e-06, "loss": 1.0238, "step": 2632 }, { "epoch": 3.5204013377926424, "grad_norm": 1.0068188529574027, "learning_rate": 2.443634834157536e-06, "loss": 0.9252, "step": 2633 }, { "epoch": 3.5217391304347827, "grad_norm": 0.8916670031477756, "learning_rate": 2.4396253386255386e-06, "loss": 0.9121, "step": 2634 }, { "epoch": 3.523076923076923, "grad_norm": 0.92910172156366, "learning_rate": 2.4356180734537643e-06, "loss": 1.0694, "step": 2635 }, { "epoch": 3.5244147157190637, "grad_norm": 0.9902860879843776, "learning_rate": 2.4316130421329696e-06, "loss": 0.8427, "step": 2636 }, { "epoch": 3.525752508361204, "grad_norm": 0.8316022321922317, "learning_rate": 2.4276102481519655e-06, "loss": 1.0361, "step": 2637 }, { "epoch": 3.5270903010033443, "grad_norm": 1.0406697545272485, "learning_rate": 2.4236096949976136e-06, "loss": 0.8721, "step": 2638 }, { "epoch": 3.528428093645485, "grad_norm": 1.126397922599676, "learning_rate": 2.4196113861548233e-06, "loss": 1.0608, "step": 2639 }, { "epoch": 3.5297658862876253, "grad_norm": 1.2575766920831981, "learning_rate": 2.41561532510655e-06, "loss": 1.1745, "step": 2640 }, { "epoch": 3.5311036789297656, "grad_norm": 1.0050877640608478, "learning_rate": 2.411621515333788e-06, "loss": 0.9256, "step": 2641 }, { "epoch": 3.5324414715719064, "grad_norm": 0.9017635590395471, "learning_rate": 2.407629960315577e-06, "loss": 1.0257, "step": 2642 }, { "epoch": 3.533779264214047, "grad_norm": 1.011533126996821, "learning_rate": 2.403640663528986e-06, "loss": 0.8704, "step": 2643 }, { "epoch": 3.5351170568561874, "grad_norm": 1.1523341607563513, "learning_rate": 2.3996536284491197e-06, "loss": 1.0452, "step": 2644 }, { "epoch": 3.5364548494983277, "grad_norm": 0.8755995859824671, "learning_rate": 2.3956688585491117e-06, "loss": 1.0619, "step": 2645 }, { "epoch": 3.5377926421404684, "grad_norm": 0.9637073046505461, "learning_rate": 2.391686357300123e-06, "loss": 1.0455, "step": 2646 }, { "epoch": 3.5391304347826087, "grad_norm": 1.1945645135850032, "learning_rate": 2.3877061281713393e-06, "loss": 1.0753, "step": 2647 }, { "epoch": 3.540468227424749, "grad_norm": 1.0221804433878559, "learning_rate": 2.383728174629964e-06, "loss": 0.8933, "step": 2648 }, { "epoch": 3.5418060200668897, "grad_norm": 0.8220189791662256, "learning_rate": 2.379752500141222e-06, "loss": 0.8549, "step": 2649 }, { "epoch": 3.54314381270903, "grad_norm": 0.7174582576901627, "learning_rate": 2.3757791081683497e-06, "loss": 0.9133, "step": 2650 }, { "epoch": 3.5444816053511703, "grad_norm": 1.0271028899362393, "learning_rate": 2.371808002172595e-06, "loss": 0.833, "step": 2651 }, { "epoch": 3.545819397993311, "grad_norm": 1.2433553118119778, "learning_rate": 2.3678391856132203e-06, "loss": 0.873, "step": 2652 }, { "epoch": 3.5471571906354518, "grad_norm": 0.9939584614134485, "learning_rate": 2.363872661947488e-06, "loss": 1.1617, "step": 2653 }, { "epoch": 3.548494983277592, "grad_norm": 0.8556866748204482, "learning_rate": 2.3599084346306626e-06, "loss": 0.9204, "step": 2654 }, { "epoch": 3.5498327759197323, "grad_norm": 0.8992088584594958, "learning_rate": 2.355946507116012e-06, "loss": 1.1148, "step": 2655 }, { "epoch": 3.551170568561873, "grad_norm": 0.8921679236547468, "learning_rate": 2.3519868828547974e-06, "loss": 1.0158, "step": 2656 }, { "epoch": 3.5525083612040134, "grad_norm": 0.9261187481138963, "learning_rate": 2.348029565296277e-06, "loss": 0.9972, "step": 2657 }, { "epoch": 3.5538461538461537, "grad_norm": 0.8921138355366143, "learning_rate": 2.344074557887696e-06, "loss": 0.9718, "step": 2658 }, { "epoch": 3.5551839464882944, "grad_norm": 1.0664475873261803, "learning_rate": 2.3401218640742894e-06, "loss": 1.0868, "step": 2659 }, { "epoch": 3.5565217391304347, "grad_norm": 0.79566938075384, "learning_rate": 2.336171487299277e-06, "loss": 1.0895, "step": 2660 }, { "epoch": 3.5578595317725754, "grad_norm": 1.0450414436779794, "learning_rate": 2.332223431003859e-06, "loss": 0.9189, "step": 2661 }, { "epoch": 3.5591973244147157, "grad_norm": 1.2580000310053763, "learning_rate": 2.3282776986272143e-06, "loss": 0.8283, "step": 2662 }, { "epoch": 3.5605351170568564, "grad_norm": 1.1659315315598233, "learning_rate": 2.324334293606499e-06, "loss": 0.9879, "step": 2663 }, { "epoch": 3.5618729096989967, "grad_norm": 1.141711013945576, "learning_rate": 2.3203932193768398e-06, "loss": 0.9568, "step": 2664 }, { "epoch": 3.563210702341137, "grad_norm": 1.2411918684238694, "learning_rate": 2.3164544793713345e-06, "loss": 0.8523, "step": 2665 }, { "epoch": 3.5645484949832777, "grad_norm": 1.1200405541989333, "learning_rate": 2.3125180770210464e-06, "loss": 0.8101, "step": 2666 }, { "epoch": 3.565886287625418, "grad_norm": 1.0621279003779422, "learning_rate": 2.3085840157550036e-06, "loss": 0.857, "step": 2667 }, { "epoch": 3.5672240802675583, "grad_norm": 1.049768856141284, "learning_rate": 2.3046522990001944e-06, "loss": 1.0257, "step": 2668 }, { "epoch": 3.568561872909699, "grad_norm": 0.8479354198151497, "learning_rate": 2.3007229301815643e-06, "loss": 1.0921, "step": 2669 }, { "epoch": 3.5698996655518394, "grad_norm": 1.0603210667183196, "learning_rate": 2.296795912722014e-06, "loss": 1.09, "step": 2670 }, { "epoch": 3.57123745819398, "grad_norm": 0.7928070444290581, "learning_rate": 2.2928712500423938e-06, "loss": 1.0306, "step": 2671 }, { "epoch": 3.5725752508361204, "grad_norm": 1.2713540326888335, "learning_rate": 2.288948945561509e-06, "loss": 0.8312, "step": 2672 }, { "epoch": 3.573913043478261, "grad_norm": 0.9764585432984232, "learning_rate": 2.2850290026961032e-06, "loss": 0.8621, "step": 2673 }, { "epoch": 3.5752508361204014, "grad_norm": 1.1312896745908467, "learning_rate": 2.2811114248608675e-06, "loss": 0.935, "step": 2674 }, { "epoch": 3.5765886287625417, "grad_norm": 1.0711223150616442, "learning_rate": 2.2771962154684303e-06, "loss": 1.0672, "step": 2675 }, { "epoch": 3.5779264214046824, "grad_norm": 1.31021035838723, "learning_rate": 2.2732833779293583e-06, "loss": 0.7289, "step": 2676 }, { "epoch": 3.5792642140468227, "grad_norm": 0.8238364849764371, "learning_rate": 2.2693729156521518e-06, "loss": 1.0225, "step": 2677 }, { "epoch": 3.580602006688963, "grad_norm": 0.9362049504179459, "learning_rate": 2.2654648320432403e-06, "loss": 0.862, "step": 2678 }, { "epoch": 3.5819397993311037, "grad_norm": 0.9892490592715557, "learning_rate": 2.2615591305069846e-06, "loss": 1.0119, "step": 2679 }, { "epoch": 3.583277591973244, "grad_norm": 1.016957823509135, "learning_rate": 2.2576558144456677e-06, "loss": 1.0203, "step": 2680 }, { "epoch": 3.5846153846153848, "grad_norm": 0.8619383931716925, "learning_rate": 2.2537548872594935e-06, "loss": 0.8703, "step": 2681 }, { "epoch": 3.585953177257525, "grad_norm": 0.8330995540952685, "learning_rate": 2.2498563523465905e-06, "loss": 1.013, "step": 2682 }, { "epoch": 3.587290969899666, "grad_norm": 1.1697769917787084, "learning_rate": 2.2459602131029977e-06, "loss": 1.0636, "step": 2683 }, { "epoch": 3.588628762541806, "grad_norm": 1.0829710984627965, "learning_rate": 2.24206647292267e-06, "loss": 1.0409, "step": 2684 }, { "epoch": 3.5899665551839464, "grad_norm": 0.9954903130878003, "learning_rate": 2.238175135197471e-06, "loss": 0.9446, "step": 2685 }, { "epoch": 3.591304347826087, "grad_norm": 0.9306907070399849, "learning_rate": 2.234286203317172e-06, "loss": 1.1009, "step": 2686 }, { "epoch": 3.5926421404682274, "grad_norm": 1.1719750475798396, "learning_rate": 2.230399680669449e-06, "loss": 1.1037, "step": 2687 }, { "epoch": 3.5939799331103677, "grad_norm": 0.8771655070674338, "learning_rate": 2.226515570639879e-06, "loss": 0.9771, "step": 2688 }, { "epoch": 3.5953177257525084, "grad_norm": 0.8397455460162695, "learning_rate": 2.2226338766119366e-06, "loss": 1.0886, "step": 2689 }, { "epoch": 3.5966555183946487, "grad_norm": 0.8906776080501645, "learning_rate": 2.2187546019669938e-06, "loss": 0.9514, "step": 2690 }, { "epoch": 3.5979933110367894, "grad_norm": 0.8762952317832369, "learning_rate": 2.2148777500843125e-06, "loss": 0.9919, "step": 2691 }, { "epoch": 3.5993311036789297, "grad_norm": 0.8349745574626892, "learning_rate": 2.2110033243410462e-06, "loss": 0.5883, "step": 2692 }, { "epoch": 3.6006688963210705, "grad_norm": 0.9753616555701272, "learning_rate": 2.207131328112234e-06, "loss": 0.8594, "step": 2693 }, { "epoch": 3.6020066889632107, "grad_norm": 0.8694884014909465, "learning_rate": 2.2032617647707995e-06, "loss": 1.0664, "step": 2694 }, { "epoch": 3.603344481605351, "grad_norm": 0.9214666277405478, "learning_rate": 2.1993946376875447e-06, "loss": 0.9905, "step": 2695 }, { "epoch": 3.6046822742474918, "grad_norm": 0.9816627908968569, "learning_rate": 2.1955299502311523e-06, "loss": 0.9237, "step": 2696 }, { "epoch": 3.606020066889632, "grad_norm": 0.975995121023001, "learning_rate": 2.1916677057681786e-06, "loss": 0.9517, "step": 2697 }, { "epoch": 3.6073578595317723, "grad_norm": 1.0197973979992887, "learning_rate": 2.1878079076630502e-06, "loss": 0.9517, "step": 2698 }, { "epoch": 3.608695652173913, "grad_norm": 0.7812489153661716, "learning_rate": 2.1839505592780658e-06, "loss": 1.1673, "step": 2699 }, { "epoch": 3.6100334448160534, "grad_norm": 1.1287593519839942, "learning_rate": 2.180095663973388e-06, "loss": 1.0585, "step": 2700 }, { "epoch": 3.611371237458194, "grad_norm": 0.9567091115246962, "learning_rate": 2.1762432251070404e-06, "loss": 0.9723, "step": 2701 }, { "epoch": 3.6127090301003344, "grad_norm": 1.0083695290342913, "learning_rate": 2.172393246034914e-06, "loss": 0.9764, "step": 2702 }, { "epoch": 3.614046822742475, "grad_norm": 0.9709168275602744, "learning_rate": 2.1685457301107506e-06, "loss": 0.9618, "step": 2703 }, { "epoch": 3.6153846153846154, "grad_norm": 0.9415629610230364, "learning_rate": 2.1647006806861472e-06, "loss": 0.9021, "step": 2704 }, { "epoch": 3.6167224080267557, "grad_norm": 1.2729046550058043, "learning_rate": 2.1608581011105533e-06, "loss": 1.0775, "step": 2705 }, { "epoch": 3.6180602006688964, "grad_norm": 0.9879644477950768, "learning_rate": 2.1570179947312674e-06, "loss": 0.9015, "step": 2706 }, { "epoch": 3.6193979933110367, "grad_norm": 1.033928230961644, "learning_rate": 2.1531803648934333e-06, "loss": 0.7971, "step": 2707 }, { "epoch": 3.620735785953177, "grad_norm": 1.1848487881187362, "learning_rate": 2.149345214940036e-06, "loss": 0.8105, "step": 2708 }, { "epoch": 3.6220735785953178, "grad_norm": 0.8462556082180163, "learning_rate": 2.145512548211902e-06, "loss": 0.6914, "step": 2709 }, { "epoch": 3.623411371237458, "grad_norm": 0.9479488341387159, "learning_rate": 2.1416823680476945e-06, "loss": 0.8787, "step": 2710 }, { "epoch": 3.624749163879599, "grad_norm": 0.9688098122044059, "learning_rate": 2.137854677783907e-06, "loss": 1.0255, "step": 2711 }, { "epoch": 3.626086956521739, "grad_norm": 1.0363814137903087, "learning_rate": 2.1340294807548716e-06, "loss": 1.2742, "step": 2712 }, { "epoch": 3.62742474916388, "grad_norm": 0.8640813320789164, "learning_rate": 2.130206780292743e-06, "loss": 0.9986, "step": 2713 }, { "epoch": 3.62876254180602, "grad_norm": 1.0744630845414274, "learning_rate": 2.1263865797275007e-06, "loss": 0.8004, "step": 2714 }, { "epoch": 3.6301003344481604, "grad_norm": 1.0396489143715608, "learning_rate": 2.1225688823869494e-06, "loss": 0.8865, "step": 2715 }, { "epoch": 3.631438127090301, "grad_norm": 1.0509752154322822, "learning_rate": 2.118753691596711e-06, "loss": 0.9771, "step": 2716 }, { "epoch": 3.6327759197324414, "grad_norm": 0.8257402506075157, "learning_rate": 2.1149410106802252e-06, "loss": 1.2837, "step": 2717 }, { "epoch": 3.6341137123745817, "grad_norm": 0.8390294705648464, "learning_rate": 2.1111308429587446e-06, "loss": 0.8512, "step": 2718 }, { "epoch": 3.6354515050167224, "grad_norm": 1.154006656367974, "learning_rate": 2.1073231917513336e-06, "loss": 1.0584, "step": 2719 }, { "epoch": 3.6367892976588627, "grad_norm": 1.1635985147545358, "learning_rate": 2.1035180603748635e-06, "loss": 0.9704, "step": 2720 }, { "epoch": 3.6381270903010035, "grad_norm": 0.9507753419666803, "learning_rate": 2.09971545214401e-06, "loss": 1.2823, "step": 2721 }, { "epoch": 3.6394648829431437, "grad_norm": 0.7346727517758653, "learning_rate": 2.095915370371252e-06, "loss": 1.0891, "step": 2722 }, { "epoch": 3.6408026755852845, "grad_norm": 0.7834810418499231, "learning_rate": 2.0921178183668676e-06, "loss": 1.1809, "step": 2723 }, { "epoch": 3.6421404682274248, "grad_norm": 0.8718231826990991, "learning_rate": 2.088322799438931e-06, "loss": 0.9463, "step": 2724 }, { "epoch": 3.643478260869565, "grad_norm": 1.0170734706128155, "learning_rate": 2.084530316893309e-06, "loss": 1.0289, "step": 2725 }, { "epoch": 3.644816053511706, "grad_norm": 0.9111186782793496, "learning_rate": 2.08074037403366e-06, "loss": 0.8134, "step": 2726 }, { "epoch": 3.646153846153846, "grad_norm": 0.8352530801557646, "learning_rate": 2.0769529741614297e-06, "loss": 0.9424, "step": 2727 }, { "epoch": 3.6474916387959864, "grad_norm": 0.9171887986878957, "learning_rate": 2.0731681205758485e-06, "loss": 1.0971, "step": 2728 }, { "epoch": 3.648829431438127, "grad_norm": 0.7551712107052351, "learning_rate": 2.069385816573928e-06, "loss": 1.2117, "step": 2729 }, { "epoch": 3.650167224080268, "grad_norm": 1.0150909194719748, "learning_rate": 2.065606065450461e-06, "loss": 1.0468, "step": 2730 }, { "epoch": 3.651505016722408, "grad_norm": 1.0603390256264829, "learning_rate": 2.061828870498012e-06, "loss": 1.2303, "step": 2731 }, { "epoch": 3.6528428093645484, "grad_norm": 0.8586569241097899, "learning_rate": 2.0580542350069266e-06, "loss": 0.928, "step": 2732 }, { "epoch": 3.654180602006689, "grad_norm": 0.7838236866421338, "learning_rate": 2.054282162265313e-06, "loss": 0.9782, "step": 2733 }, { "epoch": 3.6555183946488294, "grad_norm": 0.9867871940464066, "learning_rate": 2.050512655559051e-06, "loss": 1.1337, "step": 2734 }, { "epoch": 3.6568561872909697, "grad_norm": 0.9396151132683499, "learning_rate": 2.046745718171784e-06, "loss": 0.9541, "step": 2735 }, { "epoch": 3.6581939799331105, "grad_norm": 1.0922886527854037, "learning_rate": 2.0429813533849174e-06, "loss": 0.7772, "step": 2736 }, { "epoch": 3.6595317725752508, "grad_norm": 0.9954514103829812, "learning_rate": 2.0392195644776153e-06, "loss": 1.1712, "step": 2737 }, { "epoch": 3.660869565217391, "grad_norm": 0.8403645354001298, "learning_rate": 2.0354603547267985e-06, "loss": 0.8314, "step": 2738 }, { "epoch": 3.6622073578595318, "grad_norm": 1.0413643530651313, "learning_rate": 2.0317037274071412e-06, "loss": 0.7924, "step": 2739 }, { "epoch": 3.6635451505016725, "grad_norm": 0.9821149205100547, "learning_rate": 2.0279496857910667e-06, "loss": 1.2099, "step": 2740 }, { "epoch": 3.664882943143813, "grad_norm": 0.9337912406806408, "learning_rate": 2.0241982331487465e-06, "loss": 1.208, "step": 2741 }, { "epoch": 3.666220735785953, "grad_norm": 0.8795748230195375, "learning_rate": 2.0204493727480996e-06, "loss": 0.8879, "step": 2742 }, { "epoch": 3.667558528428094, "grad_norm": 0.7470946552382264, "learning_rate": 2.016703107854783e-06, "loss": 1.102, "step": 2743 }, { "epoch": 3.668896321070234, "grad_norm": 1.0197767640947069, "learning_rate": 2.0129594417321937e-06, "loss": 0.9312, "step": 2744 }, { "epoch": 3.6702341137123744, "grad_norm": 0.9227391967214871, "learning_rate": 2.009218377641466e-06, "loss": 0.9427, "step": 2745 }, { "epoch": 3.671571906354515, "grad_norm": 1.0172181913541982, "learning_rate": 2.0054799188414666e-06, "loss": 1.066, "step": 2746 }, { "epoch": 3.6729096989966554, "grad_norm": 0.9843556052009177, "learning_rate": 2.0017440685887934e-06, "loss": 0.8246, "step": 2747 }, { "epoch": 3.6742474916387957, "grad_norm": 0.847951550998236, "learning_rate": 1.998010830137771e-06, "loss": 0.7829, "step": 2748 }, { "epoch": 3.6755852842809364, "grad_norm": 0.7193707688516939, "learning_rate": 1.99428020674045e-06, "loss": 1.0638, "step": 2749 }, { "epoch": 3.676923076923077, "grad_norm": 1.017164402155477, "learning_rate": 1.9905522016466023e-06, "loss": 1.1112, "step": 2750 }, { "epoch": 3.6782608695652175, "grad_norm": 0.7521117045259352, "learning_rate": 1.9868268181037186e-06, "loss": 0.8123, "step": 2751 }, { "epoch": 3.6795986622073578, "grad_norm": 0.9428879856151394, "learning_rate": 1.9831040593570076e-06, "loss": 1.1641, "step": 2752 }, { "epoch": 3.6809364548494985, "grad_norm": 0.849768747256127, "learning_rate": 1.9793839286493894e-06, "loss": 1.1129, "step": 2753 }, { "epoch": 3.682274247491639, "grad_norm": 0.82348862772769, "learning_rate": 1.9756664292214962e-06, "loss": 0.9108, "step": 2754 }, { "epoch": 3.683612040133779, "grad_norm": 1.0066860545714527, "learning_rate": 1.971951564311668e-06, "loss": 0.8513, "step": 2755 }, { "epoch": 3.68494983277592, "grad_norm": 1.127323878945603, "learning_rate": 1.968239337155949e-06, "loss": 0.8183, "step": 2756 }, { "epoch": 3.68628762541806, "grad_norm": 0.9931731183951017, "learning_rate": 1.964529750988086e-06, "loss": 0.9309, "step": 2757 }, { "epoch": 3.687625418060201, "grad_norm": 1.2963944592558958, "learning_rate": 1.960822809039526e-06, "loss": 0.9418, "step": 2758 }, { "epoch": 3.688963210702341, "grad_norm": 0.9818863792599234, "learning_rate": 1.9571185145394117e-06, "loss": 0.9365, "step": 2759 }, { "epoch": 3.690301003344482, "grad_norm": 0.8579138294272908, "learning_rate": 1.95341687071458e-06, "loss": 1.1079, "step": 2760 }, { "epoch": 3.691638795986622, "grad_norm": 0.8967769054843223, "learning_rate": 1.949717880789557e-06, "loss": 1.0739, "step": 2761 }, { "epoch": 3.6929765886287624, "grad_norm": 0.9595863274539762, "learning_rate": 1.9460215479865613e-06, "loss": 1.0794, "step": 2762 }, { "epoch": 3.694314381270903, "grad_norm": 0.8322248321588366, "learning_rate": 1.9423278755254933e-06, "loss": 0.9069, "step": 2763 }, { "epoch": 3.6956521739130435, "grad_norm": 1.1503418883147012, "learning_rate": 1.9386368666239364e-06, "loss": 0.8121, "step": 2764 }, { "epoch": 3.6969899665551837, "grad_norm": 0.9642588742579251, "learning_rate": 1.9349485244971543e-06, "loss": 0.8779, "step": 2765 }, { "epoch": 3.6983277591973245, "grad_norm": 1.027757220444089, "learning_rate": 1.9312628523580882e-06, "loss": 1.084, "step": 2766 }, { "epoch": 3.6996655518394648, "grad_norm": 0.9490159785769965, "learning_rate": 1.927579853417352e-06, "loss": 1.1382, "step": 2767 }, { "epoch": 3.7010033444816055, "grad_norm": 1.0823449390940032, "learning_rate": 1.923899530883232e-06, "loss": 1.0185, "step": 2768 }, { "epoch": 3.702341137123746, "grad_norm": 1.107103017402264, "learning_rate": 1.9202218879616824e-06, "loss": 0.9893, "step": 2769 }, { "epoch": 3.7036789297658865, "grad_norm": 0.9266200795914082, "learning_rate": 1.9165469278563243e-06, "loss": 0.9016, "step": 2770 }, { "epoch": 3.705016722408027, "grad_norm": 0.968012347238075, "learning_rate": 1.912874653768439e-06, "loss": 0.9031, "step": 2771 }, { "epoch": 3.706354515050167, "grad_norm": 1.1902336949565613, "learning_rate": 1.9092050688969736e-06, "loss": 1.1172, "step": 2772 }, { "epoch": 3.707692307692308, "grad_norm": 1.164462984537177, "learning_rate": 1.9055381764385272e-06, "loss": 1.0184, "step": 2773 }, { "epoch": 3.709030100334448, "grad_norm": 1.3376869174008537, "learning_rate": 1.9018739795873558e-06, "loss": 0.8173, "step": 2774 }, { "epoch": 3.7103678929765884, "grad_norm": 0.90222295114351, "learning_rate": 1.8982124815353665e-06, "loss": 0.8171, "step": 2775 }, { "epoch": 3.711705685618729, "grad_norm": 1.078907597322207, "learning_rate": 1.8945536854721153e-06, "loss": 0.779, "step": 2776 }, { "epoch": 3.7130434782608694, "grad_norm": 0.8690351934844226, "learning_rate": 1.8908975945848063e-06, "loss": 1.0604, "step": 2777 }, { "epoch": 3.71438127090301, "grad_norm": 0.8929122792213389, "learning_rate": 1.8872442120582845e-06, "loss": 0.9779, "step": 2778 }, { "epoch": 3.7157190635451505, "grad_norm": 0.8855406833184847, "learning_rate": 1.8835935410750372e-06, "loss": 0.9411, "step": 2779 }, { "epoch": 3.717056856187291, "grad_norm": 0.8141250927341466, "learning_rate": 1.8799455848151898e-06, "loss": 1.1299, "step": 2780 }, { "epoch": 3.7183946488294315, "grad_norm": 1.1858304067454142, "learning_rate": 1.8763003464565022e-06, "loss": 1.0345, "step": 2781 }, { "epoch": 3.719732441471572, "grad_norm": 0.831191412791092, "learning_rate": 1.872657829174367e-06, "loss": 1.0153, "step": 2782 }, { "epoch": 3.7210702341137125, "grad_norm": 1.2807870826242909, "learning_rate": 1.8690180361418058e-06, "loss": 1.24, "step": 2783 }, { "epoch": 3.722408026755853, "grad_norm": 0.7620764941526961, "learning_rate": 1.865380970529469e-06, "loss": 0.8978, "step": 2784 }, { "epoch": 3.723745819397993, "grad_norm": 0.8889568927756514, "learning_rate": 1.8617466355056285e-06, "loss": 1.0678, "step": 2785 }, { "epoch": 3.725083612040134, "grad_norm": 0.7998274293000868, "learning_rate": 1.8581150342361792e-06, "loss": 0.9766, "step": 2786 }, { "epoch": 3.726421404682274, "grad_norm": 1.1522352807668508, "learning_rate": 1.854486169884635e-06, "loss": 0.9761, "step": 2787 }, { "epoch": 3.727759197324415, "grad_norm": 1.261450327827233, "learning_rate": 1.850860045612124e-06, "loss": 0.8387, "step": 2788 }, { "epoch": 3.729096989966555, "grad_norm": 1.1358440227317024, "learning_rate": 1.8472366645773892e-06, "loss": 0.8726, "step": 2789 }, { "epoch": 3.730434782608696, "grad_norm": 0.894732114238354, "learning_rate": 1.8436160299367806e-06, "loss": 0.8689, "step": 2790 }, { "epoch": 3.731772575250836, "grad_norm": 0.9166851369601461, "learning_rate": 1.8399981448442623e-06, "loss": 0.8891, "step": 2791 }, { "epoch": 3.7331103678929765, "grad_norm": 0.9415008398300495, "learning_rate": 1.8363830124513975e-06, "loss": 1.0049, "step": 2792 }, { "epoch": 3.734448160535117, "grad_norm": 1.1275616482091408, "learning_rate": 1.8327706359073526e-06, "loss": 1.036, "step": 2793 }, { "epoch": 3.7357859531772575, "grad_norm": 0.7936242609782902, "learning_rate": 1.8291610183588949e-06, "loss": 1.0889, "step": 2794 }, { "epoch": 3.7371237458193978, "grad_norm": 1.6287692846801745, "learning_rate": 1.8255541629503865e-06, "loss": 1.0093, "step": 2795 }, { "epoch": 3.7384615384615385, "grad_norm": 0.8914692219109877, "learning_rate": 1.8219500728237849e-06, "loss": 0.8853, "step": 2796 }, { "epoch": 3.739799331103679, "grad_norm": 0.8446419277654786, "learning_rate": 1.8183487511186381e-06, "loss": 0.9802, "step": 2797 }, { "epoch": 3.7411371237458195, "grad_norm": 0.8708686418780199, "learning_rate": 1.8147502009720825e-06, "loss": 1.3251, "step": 2798 }, { "epoch": 3.74247491638796, "grad_norm": 0.8218537089723053, "learning_rate": 1.8111544255188402e-06, "loss": 1.1006, "step": 2799 }, { "epoch": 3.7438127090301005, "grad_norm": 0.9457170946106264, "learning_rate": 1.807561427891214e-06, "loss": 0.8641, "step": 2800 }, { "epoch": 3.745150501672241, "grad_norm": 0.9074997680814795, "learning_rate": 1.8039712112190938e-06, "loss": 0.6878, "step": 2801 }, { "epoch": 3.746488294314381, "grad_norm": 0.9759087349030134, "learning_rate": 1.8003837786299399e-06, "loss": 0.8245, "step": 2802 }, { "epoch": 3.747826086956522, "grad_norm": 1.238365069410728, "learning_rate": 1.79679913324879e-06, "loss": 0.8937, "step": 2803 }, { "epoch": 3.749163879598662, "grad_norm": 0.7643381124221825, "learning_rate": 1.7932172781982532e-06, "loss": 1.0932, "step": 2804 }, { "epoch": 3.7505016722408024, "grad_norm": 0.9638870804674365, "learning_rate": 1.7896382165985094e-06, "loss": 1.0172, "step": 2805 }, { "epoch": 3.751839464882943, "grad_norm": 0.8568740103975314, "learning_rate": 1.7860619515673034e-06, "loss": 0.8797, "step": 2806 }, { "epoch": 3.7531772575250835, "grad_norm": 1.1660926384034307, "learning_rate": 1.7824884862199448e-06, "loss": 0.9015, "step": 2807 }, { "epoch": 3.754515050167224, "grad_norm": 0.9979284647784386, "learning_rate": 1.7789178236693045e-06, "loss": 1.1144, "step": 2808 }, { "epoch": 3.7558528428093645, "grad_norm": 1.2322185671188195, "learning_rate": 1.7753499670258106e-06, "loss": 0.9138, "step": 2809 }, { "epoch": 3.7571906354515052, "grad_norm": 1.0126173056129406, "learning_rate": 1.771784919397449e-06, "loss": 0.8645, "step": 2810 }, { "epoch": 3.7585284280936455, "grad_norm": 1.0131781013840013, "learning_rate": 1.768222683889757e-06, "loss": 1.1445, "step": 2811 }, { "epoch": 3.759866220735786, "grad_norm": 1.0675784345578605, "learning_rate": 1.764663263605823e-06, "loss": 1.131, "step": 2812 }, { "epoch": 3.7612040133779265, "grad_norm": 0.8445742848474783, "learning_rate": 1.7611066616462824e-06, "loss": 0.9629, "step": 2813 }, { "epoch": 3.762541806020067, "grad_norm": 0.9786862454788335, "learning_rate": 1.7575528811093168e-06, "loss": 0.9737, "step": 2814 }, { "epoch": 3.763879598662207, "grad_norm": 0.8726555170776208, "learning_rate": 1.7540019250906481e-06, "loss": 0.8647, "step": 2815 }, { "epoch": 3.765217391304348, "grad_norm": 0.957712510897573, "learning_rate": 1.75045379668354e-06, "loss": 0.9732, "step": 2816 }, { "epoch": 3.766555183946488, "grad_norm": 0.8255303120545551, "learning_rate": 1.746908498978791e-06, "loss": 1.0484, "step": 2817 }, { "epoch": 3.767892976588629, "grad_norm": 0.7580534035459822, "learning_rate": 1.7433660350647347e-06, "loss": 1.0995, "step": 2818 }, { "epoch": 3.769230769230769, "grad_norm": 1.0043779486810942, "learning_rate": 1.7398264080272371e-06, "loss": 1.0585, "step": 2819 }, { "epoch": 3.77056856187291, "grad_norm": 1.075911454201079, "learning_rate": 1.7362896209496894e-06, "loss": 0.75, "step": 2820 }, { "epoch": 3.77190635451505, "grad_norm": 1.3717590049766688, "learning_rate": 1.732755676913015e-06, "loss": 0.786, "step": 2821 }, { "epoch": 3.7732441471571905, "grad_norm": 1.070374482852983, "learning_rate": 1.7292245789956552e-06, "loss": 0.8778, "step": 2822 }, { "epoch": 3.774581939799331, "grad_norm": 1.1169369073705302, "learning_rate": 1.7256963302735752e-06, "loss": 0.8256, "step": 2823 }, { "epoch": 3.7759197324414715, "grad_norm": 0.9448582617380252, "learning_rate": 1.7221709338202558e-06, "loss": 0.8763, "step": 2824 }, { "epoch": 3.777257525083612, "grad_norm": 0.7508295110629416, "learning_rate": 1.718648392706695e-06, "loss": 0.8903, "step": 2825 }, { "epoch": 3.7785953177257525, "grad_norm": 0.9917715377357603, "learning_rate": 1.715128710001403e-06, "loss": 1.2075, "step": 2826 }, { "epoch": 3.779933110367893, "grad_norm": 0.8708181468292401, "learning_rate": 1.7116118887703997e-06, "loss": 1.1093, "step": 2827 }, { "epoch": 3.7812709030100335, "grad_norm": 0.9621009141174194, "learning_rate": 1.708097932077213e-06, "loss": 0.9573, "step": 2828 }, { "epoch": 3.782608695652174, "grad_norm": 0.9439546311801256, "learning_rate": 1.7045868429828745e-06, "loss": 0.8027, "step": 2829 }, { "epoch": 3.7839464882943146, "grad_norm": 1.2199318395225127, "learning_rate": 1.7010786245459166e-06, "loss": 1.0371, "step": 2830 }, { "epoch": 3.785284280936455, "grad_norm": 0.8003302448634537, "learning_rate": 1.697573279822377e-06, "loss": 0.8238, "step": 2831 }, { "epoch": 3.786622073578595, "grad_norm": 0.9350494580932758, "learning_rate": 1.6940708118657838e-06, "loss": 0.9926, "step": 2832 }, { "epoch": 3.787959866220736, "grad_norm": 0.9695767022650797, "learning_rate": 1.6905712237271616e-06, "loss": 0.8623, "step": 2833 }, { "epoch": 3.789297658862876, "grad_norm": 0.8733389757132203, "learning_rate": 1.6870745184550257e-06, "loss": 1.1263, "step": 2834 }, { "epoch": 3.7906354515050165, "grad_norm": 0.9200075983510804, "learning_rate": 1.6835806990953802e-06, "loss": 1.074, "step": 2835 }, { "epoch": 3.791973244147157, "grad_norm": 0.9080925331570424, "learning_rate": 1.680089768691716e-06, "loss": 0.9807, "step": 2836 }, { "epoch": 3.793311036789298, "grad_norm": 0.9352333234101757, "learning_rate": 1.6766017302850068e-06, "loss": 0.8771, "step": 2837 }, { "epoch": 3.794648829431438, "grad_norm": 0.8197893941293359, "learning_rate": 1.6731165869137073e-06, "loss": 1.0408, "step": 2838 }, { "epoch": 3.7959866220735785, "grad_norm": 0.860951322754363, "learning_rate": 1.6696343416137495e-06, "loss": 0.9773, "step": 2839 }, { "epoch": 3.7973244147157192, "grad_norm": 0.9490948921018401, "learning_rate": 1.6661549974185426e-06, "loss": 1.0102, "step": 2840 }, { "epoch": 3.7986622073578595, "grad_norm": 0.8602825863854174, "learning_rate": 1.6626785573589667e-06, "loss": 0.7096, "step": 2841 }, { "epoch": 3.8, "grad_norm": 0.9777394178037245, "learning_rate": 1.6592050244633733e-06, "loss": 1.2239, "step": 2842 }, { "epoch": 3.8013377926421406, "grad_norm": 0.7612557530764098, "learning_rate": 1.6557344017575817e-06, "loss": 0.9275, "step": 2843 }, { "epoch": 3.802675585284281, "grad_norm": 0.9561739607946889, "learning_rate": 1.6522666922648745e-06, "loss": 0.7662, "step": 2844 }, { "epoch": 3.804013377926421, "grad_norm": 1.2199806095219485, "learning_rate": 1.6488018990059985e-06, "loss": 1.0419, "step": 2845 }, { "epoch": 3.805351170568562, "grad_norm": 1.1404194694766714, "learning_rate": 1.6453400249991587e-06, "loss": 0.8886, "step": 2846 }, { "epoch": 3.8066889632107026, "grad_norm": 1.251263783272409, "learning_rate": 1.6418810732600177e-06, "loss": 0.7183, "step": 2847 }, { "epoch": 3.808026755852843, "grad_norm": 1.052113160494803, "learning_rate": 1.6384250468016932e-06, "loss": 0.9257, "step": 2848 }, { "epoch": 3.809364548494983, "grad_norm": 1.0047156335731757, "learning_rate": 1.6349719486347533e-06, "loss": 0.9949, "step": 2849 }, { "epoch": 3.810702341137124, "grad_norm": 0.9057465985285285, "learning_rate": 1.6315217817672142e-06, "loss": 0.8827, "step": 2850 }, { "epoch": 3.812040133779264, "grad_norm": 0.8828710446309934, "learning_rate": 1.6280745492045435e-06, "loss": 1.2262, "step": 2851 }, { "epoch": 3.8133779264214045, "grad_norm": 0.9758173564591484, "learning_rate": 1.6246302539496483e-06, "loss": 0.9632, "step": 2852 }, { "epoch": 3.8147157190635452, "grad_norm": 1.0219371355102544, "learning_rate": 1.6211888990028785e-06, "loss": 0.6997, "step": 2853 }, { "epoch": 3.8160535117056855, "grad_norm": 0.9184624021088534, "learning_rate": 1.617750487362022e-06, "loss": 0.9641, "step": 2854 }, { "epoch": 3.8173913043478263, "grad_norm": 0.7693382609556989, "learning_rate": 1.614315022022303e-06, "loss": 0.9865, "step": 2855 }, { "epoch": 3.8187290969899665, "grad_norm": 0.8697318759448883, "learning_rate": 1.6108825059763794e-06, "loss": 0.9817, "step": 2856 }, { "epoch": 3.8200668896321073, "grad_norm": 0.9447889671293613, "learning_rate": 1.6074529422143398e-06, "loss": 0.7263, "step": 2857 }, { "epoch": 3.8214046822742476, "grad_norm": 0.9930397481749877, "learning_rate": 1.6040263337237017e-06, "loss": 0.8853, "step": 2858 }, { "epoch": 3.822742474916388, "grad_norm": 0.8513963690251027, "learning_rate": 1.6006026834894068e-06, "loss": 1.1614, "step": 2859 }, { "epoch": 3.8240802675585286, "grad_norm": 0.8789990068196665, "learning_rate": 1.5971819944938194e-06, "loss": 1.0991, "step": 2860 }, { "epoch": 3.825418060200669, "grad_norm": 1.0186508792438667, "learning_rate": 1.5937642697167288e-06, "loss": 1.0806, "step": 2861 }, { "epoch": 3.826755852842809, "grad_norm": 0.9205860419195312, "learning_rate": 1.5903495121353373e-06, "loss": 1.0763, "step": 2862 }, { "epoch": 3.82809364548495, "grad_norm": 0.8238345592083953, "learning_rate": 1.5869377247242645e-06, "loss": 0.847, "step": 2863 }, { "epoch": 3.82943143812709, "grad_norm": 1.038699898261887, "learning_rate": 1.5835289104555417e-06, "loss": 0.9154, "step": 2864 }, { "epoch": 3.830769230769231, "grad_norm": 1.247783410669057, "learning_rate": 1.5801230722986104e-06, "loss": 0.9717, "step": 2865 }, { "epoch": 3.832107023411371, "grad_norm": 1.0096538101048007, "learning_rate": 1.5767202132203207e-06, "loss": 0.8924, "step": 2866 }, { "epoch": 3.833444816053512, "grad_norm": 0.9824461473420792, "learning_rate": 1.5733203361849265e-06, "loss": 0.914, "step": 2867 }, { "epoch": 3.8347826086956522, "grad_norm": 0.8186006852549992, "learning_rate": 1.5699234441540845e-06, "loss": 0.8806, "step": 2868 }, { "epoch": 3.8361204013377925, "grad_norm": 1.1397618816692596, "learning_rate": 1.5665295400868513e-06, "loss": 0.7942, "step": 2869 }, { "epoch": 3.8374581939799333, "grad_norm": 0.9398836586460928, "learning_rate": 1.5631386269396798e-06, "loss": 0.9332, "step": 2870 }, { "epoch": 3.8387959866220736, "grad_norm": 1.0112888950580399, "learning_rate": 1.5597507076664187e-06, "loss": 1.1208, "step": 2871 }, { "epoch": 3.840133779264214, "grad_norm": 0.8309386253528712, "learning_rate": 1.5563657852183072e-06, "loss": 0.9988, "step": 2872 }, { "epoch": 3.8414715719063546, "grad_norm": 0.8804665000908655, "learning_rate": 1.5529838625439763e-06, "loss": 1.007, "step": 2873 }, { "epoch": 3.842809364548495, "grad_norm": 0.9560081451478236, "learning_rate": 1.549604942589441e-06, "loss": 1.0596, "step": 2874 }, { "epoch": 3.8441471571906356, "grad_norm": 1.0355007968531724, "learning_rate": 1.546229028298103e-06, "loss": 0.9527, "step": 2875 }, { "epoch": 3.845484949832776, "grad_norm": 1.011220198524945, "learning_rate": 1.5428561226107442e-06, "loss": 0.9427, "step": 2876 }, { "epoch": 3.8468227424749166, "grad_norm": 1.1014773481586744, "learning_rate": 1.5394862284655266e-06, "loss": 1.0797, "step": 2877 }, { "epoch": 3.848160535117057, "grad_norm": 1.3397298517437326, "learning_rate": 1.5361193487979881e-06, "loss": 0.9488, "step": 2878 }, { "epoch": 3.849498327759197, "grad_norm": 0.9864328913633427, "learning_rate": 1.5327554865410415e-06, "loss": 1.0637, "step": 2879 }, { "epoch": 3.850836120401338, "grad_norm": 1.115935831698273, "learning_rate": 1.5293946446249686e-06, "loss": 0.7503, "step": 2880 }, { "epoch": 3.8521739130434782, "grad_norm": 0.947183002313095, "learning_rate": 1.526036825977426e-06, "loss": 0.9915, "step": 2881 }, { "epoch": 3.8535117056856185, "grad_norm": 0.7830464608162625, "learning_rate": 1.5226820335234316e-06, "loss": 0.6873, "step": 2882 }, { "epoch": 3.8548494983277592, "grad_norm": 1.4253625631081908, "learning_rate": 1.5193302701853674e-06, "loss": 0.7802, "step": 2883 }, { "epoch": 3.8561872909698995, "grad_norm": 1.2698886500112376, "learning_rate": 1.5159815388829784e-06, "loss": 0.7133, "step": 2884 }, { "epoch": 3.8575250836120403, "grad_norm": 1.0515389004135351, "learning_rate": 1.5126358425333677e-06, "loss": 0.9267, "step": 2885 }, { "epoch": 3.8588628762541806, "grad_norm": 0.6968456439331615, "learning_rate": 1.509293184050995e-06, "loss": 1.015, "step": 2886 }, { "epoch": 3.8602006688963213, "grad_norm": 0.9523465912569055, "learning_rate": 1.5059535663476731e-06, "loss": 0.8665, "step": 2887 }, { "epoch": 3.8615384615384616, "grad_norm": 1.1900485094050761, "learning_rate": 1.5026169923325668e-06, "loss": 1.0919, "step": 2888 }, { "epoch": 3.862876254180602, "grad_norm": 1.1892909048518783, "learning_rate": 1.499283464912188e-06, "loss": 0.8508, "step": 2889 }, { "epoch": 3.8642140468227426, "grad_norm": 0.8734625790121906, "learning_rate": 1.4959529869903948e-06, "loss": 1.1085, "step": 2890 }, { "epoch": 3.865551839464883, "grad_norm": 0.8824829650793081, "learning_rate": 1.4926255614683931e-06, "loss": 0.8793, "step": 2891 }, { "epoch": 3.866889632107023, "grad_norm": 1.3974244861424894, "learning_rate": 1.4893011912447248e-06, "loss": 0.8778, "step": 2892 }, { "epoch": 3.868227424749164, "grad_norm": 1.2917469722077715, "learning_rate": 1.4859798792152713e-06, "loss": 0.9199, "step": 2893 }, { "epoch": 3.869565217391304, "grad_norm": 0.7970921946060576, "learning_rate": 1.4826616282732509e-06, "loss": 1.1941, "step": 2894 }, { "epoch": 3.870903010033445, "grad_norm": 0.8940357774151968, "learning_rate": 1.4793464413092161e-06, "loss": 1.0928, "step": 2895 }, { "epoch": 3.8722408026755852, "grad_norm": 1.2169371023400217, "learning_rate": 1.4760343212110484e-06, "loss": 0.8983, "step": 2896 }, { "epoch": 3.873578595317726, "grad_norm": 0.7739852549459412, "learning_rate": 1.4727252708639589e-06, "loss": 1.0997, "step": 2897 }, { "epoch": 3.8749163879598663, "grad_norm": 0.9630023163339092, "learning_rate": 1.4694192931504842e-06, "loss": 1.0888, "step": 2898 }, { "epoch": 3.8762541806020065, "grad_norm": 1.2608082948389938, "learning_rate": 1.4661163909504855e-06, "loss": 0.8956, "step": 2899 }, { "epoch": 3.8775919732441473, "grad_norm": 1.254666830672735, "learning_rate": 1.4628165671411426e-06, "loss": 0.9282, "step": 2900 }, { "epoch": 3.8789297658862876, "grad_norm": 0.8164425109649939, "learning_rate": 1.459519824596956e-06, "loss": 0.9209, "step": 2901 }, { "epoch": 3.880267558528428, "grad_norm": 1.1343854808948166, "learning_rate": 1.4562261661897415e-06, "loss": 0.644, "step": 2902 }, { "epoch": 3.8816053511705686, "grad_norm": 1.0938850954498724, "learning_rate": 1.4529355947886265e-06, "loss": 0.9414, "step": 2903 }, { "epoch": 3.882943143812709, "grad_norm": 0.7026755594711415, "learning_rate": 1.4496481132600516e-06, "loss": 0.6901, "step": 2904 }, { "epoch": 3.8842809364548496, "grad_norm": 1.3366215806762543, "learning_rate": 1.4463637244677648e-06, "loss": 0.8843, "step": 2905 }, { "epoch": 3.88561872909699, "grad_norm": 1.0617941561521118, "learning_rate": 1.4430824312728197e-06, "loss": 0.9417, "step": 2906 }, { "epoch": 3.8869565217391306, "grad_norm": 0.8649352261964967, "learning_rate": 1.4398042365335745e-06, "loss": 1.1127, "step": 2907 }, { "epoch": 3.888294314381271, "grad_norm": 1.1072272931587088, "learning_rate": 1.4365291431056871e-06, "loss": 0.9462, "step": 2908 }, { "epoch": 3.8896321070234112, "grad_norm": 0.8564383544769428, "learning_rate": 1.4332571538421136e-06, "loss": 0.9552, "step": 2909 }, { "epoch": 3.890969899665552, "grad_norm": 0.746407888753958, "learning_rate": 1.4299882715931062e-06, "loss": 0.9952, "step": 2910 }, { "epoch": 3.8923076923076922, "grad_norm": 0.9338117507524845, "learning_rate": 1.4267224992062134e-06, "loss": 0.9612, "step": 2911 }, { "epoch": 3.8936454849498325, "grad_norm": 1.0828530795024522, "learning_rate": 1.4234598395262706e-06, "loss": 1.0822, "step": 2912 }, { "epoch": 3.8949832775919733, "grad_norm": 0.9949504447638382, "learning_rate": 1.4202002953954042e-06, "loss": 0.8974, "step": 2913 }, { "epoch": 3.8963210702341136, "grad_norm": 0.8706740335180897, "learning_rate": 1.4169438696530246e-06, "loss": 1.0745, "step": 2914 }, { "epoch": 3.8976588628762543, "grad_norm": 0.9595047728101556, "learning_rate": 1.4136905651358284e-06, "loss": 1.0799, "step": 2915 }, { "epoch": 3.8989966555183946, "grad_norm": 0.8269443331126609, "learning_rate": 1.410440384677791e-06, "loss": 0.8089, "step": 2916 }, { "epoch": 3.9003344481605353, "grad_norm": 0.9142806676904466, "learning_rate": 1.4071933311101675e-06, "loss": 0.9404, "step": 2917 }, { "epoch": 3.9016722408026756, "grad_norm": 1.0740253403883315, "learning_rate": 1.4039494072614884e-06, "loss": 0.9896, "step": 2918 }, { "epoch": 3.903010033444816, "grad_norm": 1.0236637299496854, "learning_rate": 1.4007086159575595e-06, "loss": 1.1502, "step": 2919 }, { "epoch": 3.9043478260869566, "grad_norm": 0.9097860466590479, "learning_rate": 1.3974709600214541e-06, "loss": 0.9835, "step": 2920 }, { "epoch": 3.905685618729097, "grad_norm": 1.2111461839483841, "learning_rate": 1.3942364422735205e-06, "loss": 0.9651, "step": 2921 }, { "epoch": 3.907023411371237, "grad_norm": 0.7842769044569, "learning_rate": 1.3910050655313679e-06, "loss": 0.9779, "step": 2922 }, { "epoch": 3.908361204013378, "grad_norm": 0.8670565820694772, "learning_rate": 1.3877768326098712e-06, "loss": 0.8038, "step": 2923 }, { "epoch": 3.9096989966555182, "grad_norm": 0.946256873290899, "learning_rate": 1.3845517463211667e-06, "loss": 0.9532, "step": 2924 }, { "epoch": 3.911036789297659, "grad_norm": 0.8878495604794058, "learning_rate": 1.3813298094746491e-06, "loss": 0.8859, "step": 2925 }, { "epoch": 3.9123745819397993, "grad_norm": 0.843341121899057, "learning_rate": 1.3781110248769709e-06, "loss": 0.7116, "step": 2926 }, { "epoch": 3.91371237458194, "grad_norm": 0.8584176858495275, "learning_rate": 1.374895395332037e-06, "loss": 0.8925, "step": 2927 }, { "epoch": 3.9150501672240803, "grad_norm": 0.939253357788684, "learning_rate": 1.371682923641005e-06, "loss": 1.0083, "step": 2928 }, { "epoch": 3.9163879598662206, "grad_norm": 1.0575732216175493, "learning_rate": 1.3684736126022812e-06, "loss": 1.1236, "step": 2929 }, { "epoch": 3.9177257525083613, "grad_norm": 1.0477679176774317, "learning_rate": 1.3652674650115193e-06, "loss": 0.9723, "step": 2930 }, { "epoch": 3.9190635451505016, "grad_norm": 1.1589029074484534, "learning_rate": 1.362064483661617e-06, "loss": 1.0272, "step": 2931 }, { "epoch": 3.920401337792642, "grad_norm": 0.9214620179959191, "learning_rate": 1.3588646713427128e-06, "loss": 1.256, "step": 2932 }, { "epoch": 3.9217391304347826, "grad_norm": 0.7911924098275, "learning_rate": 1.3556680308421865e-06, "loss": 1.167, "step": 2933 }, { "epoch": 3.9230769230769234, "grad_norm": 0.8243396203451439, "learning_rate": 1.352474564944653e-06, "loss": 0.9643, "step": 2934 }, { "epoch": 3.9244147157190636, "grad_norm": 0.7561625100524417, "learning_rate": 1.349284276431963e-06, "loss": 0.8563, "step": 2935 }, { "epoch": 3.925752508361204, "grad_norm": 0.9897121065807176, "learning_rate": 1.3460971680831996e-06, "loss": 0.7655, "step": 2936 }, { "epoch": 3.9270903010033447, "grad_norm": 1.2027642873290771, "learning_rate": 1.3429132426746743e-06, "loss": 1.0102, "step": 2937 }, { "epoch": 3.928428093645485, "grad_norm": 1.1557889928757372, "learning_rate": 1.339732502979928e-06, "loss": 0.8893, "step": 2938 }, { "epoch": 3.9297658862876252, "grad_norm": 0.9290886367429305, "learning_rate": 1.3365549517697234e-06, "loss": 1.0286, "step": 2939 }, { "epoch": 3.931103678929766, "grad_norm": 0.9465631284644397, "learning_rate": 1.3333805918120473e-06, "loss": 0.8154, "step": 2940 }, { "epoch": 3.9324414715719063, "grad_norm": 0.7592763792340298, "learning_rate": 1.33020942587211e-06, "loss": 1.0259, "step": 2941 }, { "epoch": 3.9337792642140466, "grad_norm": 1.0486604402541748, "learning_rate": 1.3270414567123342e-06, "loss": 0.7261, "step": 2942 }, { "epoch": 3.9351170568561873, "grad_norm": 1.0984565981098793, "learning_rate": 1.3238766870923592e-06, "loss": 0.9907, "step": 2943 }, { "epoch": 3.936454849498328, "grad_norm": 0.9060443071850033, "learning_rate": 1.3207151197690392e-06, "loss": 1.0279, "step": 2944 }, { "epoch": 3.9377926421404683, "grad_norm": 0.9253891817224952, "learning_rate": 1.3175567574964372e-06, "loss": 1.103, "step": 2945 }, { "epoch": 3.9391304347826086, "grad_norm": 0.6964454143768536, "learning_rate": 1.3144016030258244e-06, "loss": 1.0703, "step": 2946 }, { "epoch": 3.9404682274247493, "grad_norm": 0.9161897490159745, "learning_rate": 1.3112496591056778e-06, "loss": 1.1433, "step": 2947 }, { "epoch": 3.9418060200668896, "grad_norm": 0.8019087900014917, "learning_rate": 1.3081009284816776e-06, "loss": 0.9976, "step": 2948 }, { "epoch": 3.94314381270903, "grad_norm": 0.9149438925265041, "learning_rate": 1.3049554138967052e-06, "loss": 1.0089, "step": 2949 }, { "epoch": 3.9444816053511706, "grad_norm": 0.9831910873581642, "learning_rate": 1.301813118090839e-06, "loss": 1.0747, "step": 2950 }, { "epoch": 3.945819397993311, "grad_norm": 0.8728878941959286, "learning_rate": 1.2986740438013579e-06, "loss": 1.0207, "step": 2951 }, { "epoch": 3.9471571906354512, "grad_norm": 1.2002156242440467, "learning_rate": 1.2955381937627293e-06, "loss": 0.9955, "step": 2952 }, { "epoch": 3.948494983277592, "grad_norm": 0.8455620638876246, "learning_rate": 1.2924055707066141e-06, "loss": 0.8705, "step": 2953 }, { "epoch": 3.9498327759197327, "grad_norm": 0.9383058323988297, "learning_rate": 1.2892761773618628e-06, "loss": 0.9221, "step": 2954 }, { "epoch": 3.951170568561873, "grad_norm": 1.22905656280789, "learning_rate": 1.286150016454511e-06, "loss": 1.041, "step": 2955 }, { "epoch": 3.9525083612040133, "grad_norm": 0.951661269537117, "learning_rate": 1.2830270907077797e-06, "loss": 0.9148, "step": 2956 }, { "epoch": 3.953846153846154, "grad_norm": 1.18457472930354, "learning_rate": 1.279907402842071e-06, "loss": 1.0042, "step": 2957 }, { "epoch": 3.9551839464882943, "grad_norm": 0.9399740274127215, "learning_rate": 1.2767909555749676e-06, "loss": 0.8243, "step": 2958 }, { "epoch": 3.9565217391304346, "grad_norm": 0.8949601578241434, "learning_rate": 1.2736777516212267e-06, "loss": 1.093, "step": 2959 }, { "epoch": 3.9578595317725753, "grad_norm": 1.0265363637006253, "learning_rate": 1.2705677936927841e-06, "loss": 0.9797, "step": 2960 }, { "epoch": 3.9591973244147156, "grad_norm": 0.8684861094473476, "learning_rate": 1.267461084498744e-06, "loss": 1.0616, "step": 2961 }, { "epoch": 3.9605351170568563, "grad_norm": 1.0028014816770434, "learning_rate": 1.2643576267453832e-06, "loss": 1.1303, "step": 2962 }, { "epoch": 3.9618729096989966, "grad_norm": 0.9014937883310552, "learning_rate": 1.2612574231361463e-06, "loss": 1.129, "step": 2963 }, { "epoch": 3.9632107023411374, "grad_norm": 0.8447278310832329, "learning_rate": 1.2581604763716404e-06, "loss": 1.0309, "step": 2964 }, { "epoch": 3.9645484949832777, "grad_norm": 0.8086740623803703, "learning_rate": 1.2550667891496394e-06, "loss": 0.9844, "step": 2965 }, { "epoch": 3.965886287625418, "grad_norm": 1.1428375600563587, "learning_rate": 1.2519763641650739e-06, "loss": 0.9213, "step": 2966 }, { "epoch": 3.9672240802675587, "grad_norm": 0.9348549605318561, "learning_rate": 1.2488892041100364e-06, "loss": 1.0515, "step": 2967 }, { "epoch": 3.968561872909699, "grad_norm": 0.9443944939084737, "learning_rate": 1.2458053116737722e-06, "loss": 1.0087, "step": 2968 }, { "epoch": 3.9698996655518393, "grad_norm": 1.3675764377805784, "learning_rate": 1.2427246895426826e-06, "loss": 1.0065, "step": 2969 }, { "epoch": 3.97123745819398, "grad_norm": 1.1154429625714943, "learning_rate": 1.2396473404003162e-06, "loss": 0.8541, "step": 2970 }, { "epoch": 3.9725752508361203, "grad_norm": 0.7291221221235983, "learning_rate": 1.2365732669273778e-06, "loss": 0.945, "step": 2971 }, { "epoch": 3.973913043478261, "grad_norm": 0.9798622157548187, "learning_rate": 1.233502471801712e-06, "loss": 0.8811, "step": 2972 }, { "epoch": 3.9752508361204013, "grad_norm": 0.8091397523734227, "learning_rate": 1.2304349576983094e-06, "loss": 1.0031, "step": 2973 }, { "epoch": 3.976588628762542, "grad_norm": 1.0553868487578657, "learning_rate": 1.2273707272893038e-06, "loss": 1.1969, "step": 2974 }, { "epoch": 3.9779264214046823, "grad_norm": 0.9472962663566439, "learning_rate": 1.2243097832439672e-06, "loss": 0.8081, "step": 2975 }, { "epoch": 3.9792642140468226, "grad_norm": 0.7850369942234027, "learning_rate": 1.2212521282287093e-06, "loss": 1.1026, "step": 2976 }, { "epoch": 3.9806020066889634, "grad_norm": 0.8741532063380797, "learning_rate": 1.2181977649070749e-06, "loss": 1.1587, "step": 2977 }, { "epoch": 3.9819397993311036, "grad_norm": 0.9707440620020908, "learning_rate": 1.2151466959397406e-06, "loss": 1.0318, "step": 2978 }, { "epoch": 3.983277591973244, "grad_norm": 1.25431483033151, "learning_rate": 1.2120989239845149e-06, "loss": 0.9938, "step": 2979 }, { "epoch": 3.9846153846153847, "grad_norm": 1.0095426218589978, "learning_rate": 1.209054451696331e-06, "loss": 0.9969, "step": 2980 }, { "epoch": 3.985953177257525, "grad_norm": 1.3366443790845448, "learning_rate": 1.206013281727253e-06, "loss": 0.7042, "step": 2981 }, { "epoch": 3.9872909698996657, "grad_norm": 1.0064577380052662, "learning_rate": 1.202975416726464e-06, "loss": 1.0553, "step": 2982 }, { "epoch": 3.988628762541806, "grad_norm": 0.9733697828736418, "learning_rate": 1.1999408593402688e-06, "loss": 1.2231, "step": 2983 }, { "epoch": 3.9899665551839467, "grad_norm": 1.176234255026895, "learning_rate": 1.1969096122120927e-06, "loss": 1.1721, "step": 2984 }, { "epoch": 3.991304347826087, "grad_norm": 1.207765170631155, "learning_rate": 1.1938816779824753e-06, "loss": 1.0478, "step": 2985 }, { "epoch": 3.9926421404682273, "grad_norm": 1.0204744181726053, "learning_rate": 1.190857059289071e-06, "loss": 0.9718, "step": 2986 }, { "epoch": 3.993979933110368, "grad_norm": 1.0404125108229358, "learning_rate": 1.1878357587666468e-06, "loss": 0.9873, "step": 2987 }, { "epoch": 3.9953177257525083, "grad_norm": 0.9203869397088814, "learning_rate": 1.1848177790470784e-06, "loss": 1.1694, "step": 2988 }, { "epoch": 3.9966555183946486, "grad_norm": 1.2765278565456248, "learning_rate": 1.1818031227593491e-06, "loss": 0.8035, "step": 2989 }, { "epoch": 3.9979933110367893, "grad_norm": 1.1410236254791233, "learning_rate": 1.1787917925295467e-06, "loss": 0.9816, "step": 2990 }, { "epoch": 3.9993311036789296, "grad_norm": 0.7974138054696138, "learning_rate": 1.1757837909808628e-06, "loss": 0.6811, "step": 2991 }, { "epoch": 4.0, "grad_norm": 1.441206905114398, "learning_rate": 1.1727791207335876e-06, "loss": 1.2157, "step": 2992 }, { "epoch": 4.001337792642141, "grad_norm": 0.8280063351912375, "learning_rate": 1.1697777844051105e-06, "loss": 1.0405, "step": 2993 }, { "epoch": 4.002675585284281, "grad_norm": 1.021464006118735, "learning_rate": 1.1667797846099172e-06, "loss": 1.1537, "step": 2994 }, { "epoch": 4.004013377926421, "grad_norm": 1.250544768387613, "learning_rate": 1.163785123959585e-06, "loss": 0.799, "step": 2995 }, { "epoch": 4.005351170568562, "grad_norm": 0.8663404728133578, "learning_rate": 1.1607938050627849e-06, "loss": 0.6797, "step": 2996 }, { "epoch": 4.006688963210703, "grad_norm": 1.2452650451711589, "learning_rate": 1.157805830525275e-06, "loss": 1.0364, "step": 2997 }, { "epoch": 4.008026755852843, "grad_norm": 0.7929359153341754, "learning_rate": 1.1548212029499006e-06, "loss": 0.9627, "step": 2998 }, { "epoch": 4.009364548494983, "grad_norm": 0.8158722204906339, "learning_rate": 1.1518399249365924e-06, "loss": 0.8799, "step": 2999 }, { "epoch": 4.010702341137124, "grad_norm": 0.9871320743876424, "learning_rate": 1.1488619990823602e-06, "loss": 0.9388, "step": 3000 }, { "epoch": 4.012040133779264, "grad_norm": 0.8122481790696807, "learning_rate": 1.1458874279812992e-06, "loss": 1.1517, "step": 3001 }, { "epoch": 4.013377926421405, "grad_norm": 0.8799358618226677, "learning_rate": 1.1429162142245775e-06, "loss": 0.9716, "step": 3002 }, { "epoch": 4.014715719063545, "grad_norm": 0.7836927913576146, "learning_rate": 1.1399483604004403e-06, "loss": 0.9862, "step": 3003 }, { "epoch": 4.016053511705685, "grad_norm": 0.8810321412135649, "learning_rate": 1.1369838690942059e-06, "loss": 0.7127, "step": 3004 }, { "epoch": 4.017391304347826, "grad_norm": 0.7665908081956377, "learning_rate": 1.1340227428882627e-06, "loss": 0.9913, "step": 3005 }, { "epoch": 4.018729096989967, "grad_norm": 0.8749644054182748, "learning_rate": 1.1310649843620686e-06, "loss": 0.8887, "step": 3006 }, { "epoch": 4.0200668896321075, "grad_norm": 0.8478311915854941, "learning_rate": 1.1281105960921484e-06, "loss": 1.0023, "step": 3007 }, { "epoch": 4.021404682274247, "grad_norm": 1.076175242844535, "learning_rate": 1.1251595806520893e-06, "loss": 0.7934, "step": 3008 }, { "epoch": 4.022742474916388, "grad_norm": 1.031261770876538, "learning_rate": 1.1222119406125426e-06, "loss": 0.8056, "step": 3009 }, { "epoch": 4.024080267558529, "grad_norm": 0.8872452682612977, "learning_rate": 1.1192676785412154e-06, "loss": 0.9876, "step": 3010 }, { "epoch": 4.025418060200669, "grad_norm": 1.046031924722863, "learning_rate": 1.1163267970028786e-06, "loss": 0.7953, "step": 3011 }, { "epoch": 4.026755852842809, "grad_norm": 1.0272943382143616, "learning_rate": 1.1133892985593532e-06, "loss": 0.8765, "step": 3012 }, { "epoch": 4.02809364548495, "grad_norm": 1.1866047504935449, "learning_rate": 1.1104551857695133e-06, "loss": 0.7898, "step": 3013 }, { "epoch": 4.02943143812709, "grad_norm": 1.242421754110628, "learning_rate": 1.1075244611892872e-06, "loss": 0.7327, "step": 3014 }, { "epoch": 4.030769230769231, "grad_norm": 0.9686373724405863, "learning_rate": 1.1045971273716476e-06, "loss": 0.9544, "step": 3015 }, { "epoch": 4.032107023411371, "grad_norm": 0.8706934811615044, "learning_rate": 1.1016731868666169e-06, "loss": 0.8446, "step": 3016 }, { "epoch": 4.033444816053512, "grad_norm": 0.8944548695448159, "learning_rate": 1.0987526422212585e-06, "loss": 0.945, "step": 3017 }, { "epoch": 4.034782608695652, "grad_norm": 0.7871596058805923, "learning_rate": 1.0958354959796807e-06, "loss": 1.0302, "step": 3018 }, { "epoch": 4.036120401337793, "grad_norm": 0.8952270955107063, "learning_rate": 1.0929217506830292e-06, "loss": 1.3021, "step": 3019 }, { "epoch": 4.037458193979933, "grad_norm": 0.9188011782915074, "learning_rate": 1.0900114088694874e-06, "loss": 1.209, "step": 3020 }, { "epoch": 4.038795986622073, "grad_norm": 0.8044853645953933, "learning_rate": 1.0871044730742752e-06, "loss": 1.0376, "step": 3021 }, { "epoch": 4.040133779264214, "grad_norm": 1.0504682241721726, "learning_rate": 1.084200945829645e-06, "loss": 0.8985, "step": 3022 }, { "epoch": 4.041471571906355, "grad_norm": 0.7799441975322731, "learning_rate": 1.081300829664878e-06, "loss": 0.9506, "step": 3023 }, { "epoch": 4.042809364548495, "grad_norm": 1.040662267371888, "learning_rate": 1.0784041271062867e-06, "loss": 1.0795, "step": 3024 }, { "epoch": 4.044147157190635, "grad_norm": 1.2943742605274544, "learning_rate": 1.075510840677209e-06, "loss": 0.9327, "step": 3025 }, { "epoch": 4.045484949832776, "grad_norm": 1.3681990992814457, "learning_rate": 1.072620972898007e-06, "loss": 0.8332, "step": 3026 }, { "epoch": 4.046822742474917, "grad_norm": 0.9023874696344266, "learning_rate": 1.0697345262860638e-06, "loss": 1.0672, "step": 3027 }, { "epoch": 4.048160535117057, "grad_norm": 0.8790391060213004, "learning_rate": 1.0668515033557835e-06, "loss": 1.1021, "step": 3028 }, { "epoch": 4.049498327759197, "grad_norm": 0.7947940646374899, "learning_rate": 1.0639719066185867e-06, "loss": 0.8694, "step": 3029 }, { "epoch": 4.050836120401338, "grad_norm": 1.0765342053359679, "learning_rate": 1.061095738582913e-06, "loss": 1.383, "step": 3030 }, { "epoch": 4.052173913043478, "grad_norm": 0.8627709463180413, "learning_rate": 1.05822300175421e-06, "loss": 1.0304, "step": 3031 }, { "epoch": 4.053511705685619, "grad_norm": 0.9766195414609894, "learning_rate": 1.0553536986349393e-06, "loss": 1.0444, "step": 3032 }, { "epoch": 4.054849498327759, "grad_norm": 1.3960004314298025, "learning_rate": 1.0524878317245713e-06, "loss": 0.928, "step": 3033 }, { "epoch": 4.056187290969899, "grad_norm": 0.9027069373685171, "learning_rate": 1.0496254035195819e-06, "loss": 0.9591, "step": 3034 }, { "epoch": 4.05752508361204, "grad_norm": 0.8522403001247284, "learning_rate": 1.0467664165134534e-06, "loss": 0.8013, "step": 3035 }, { "epoch": 4.058862876254181, "grad_norm": 0.9841337246919105, "learning_rate": 1.043910873196668e-06, "loss": 0.8514, "step": 3036 }, { "epoch": 4.0602006688963215, "grad_norm": 0.8308246608632991, "learning_rate": 1.0410587760567104e-06, "loss": 1.0901, "step": 3037 }, { "epoch": 4.061538461538461, "grad_norm": 0.9945926648450533, "learning_rate": 1.0382101275780615e-06, "loss": 1.0146, "step": 3038 }, { "epoch": 4.062876254180602, "grad_norm": 0.8021949913771701, "learning_rate": 1.0353649302421982e-06, "loss": 1.09, "step": 3039 }, { "epoch": 4.064214046822743, "grad_norm": 0.8640538692952772, "learning_rate": 1.0325231865275936e-06, "loss": 0.8885, "step": 3040 }, { "epoch": 4.065551839464883, "grad_norm": 0.813073700118874, "learning_rate": 1.0296848989097103e-06, "loss": 1.087, "step": 3041 }, { "epoch": 4.066889632107023, "grad_norm": 0.9181471338018907, "learning_rate": 1.0268500698609996e-06, "loss": 0.7517, "step": 3042 }, { "epoch": 4.068227424749164, "grad_norm": 0.9063384282623147, "learning_rate": 1.0240187018509012e-06, "loss": 0.8741, "step": 3043 }, { "epoch": 4.069565217391304, "grad_norm": 1.0021965708360927, "learning_rate": 1.0211907973458391e-06, "loss": 0.6118, "step": 3044 }, { "epoch": 4.070903010033445, "grad_norm": 0.8603616228055635, "learning_rate": 1.0183663588092214e-06, "loss": 0.9874, "step": 3045 }, { "epoch": 4.072240802675585, "grad_norm": 0.9385409162116537, "learning_rate": 1.015545388701435e-06, "loss": 1.0934, "step": 3046 }, { "epoch": 4.073578595317726, "grad_norm": 0.891804368886903, "learning_rate": 1.012727889479848e-06, "loss": 1.0759, "step": 3047 }, { "epoch": 4.074916387959866, "grad_norm": 1.1499469422611253, "learning_rate": 1.0099138635988026e-06, "loss": 0.9738, "step": 3048 }, { "epoch": 4.076254180602007, "grad_norm": 0.8842361714606165, "learning_rate": 1.007103313509617e-06, "loss": 0.8134, "step": 3049 }, { "epoch": 4.0775919732441475, "grad_norm": 1.0170977399235832, "learning_rate": 1.0042962416605805e-06, "loss": 0.8521, "step": 3050 }, { "epoch": 4.078929765886287, "grad_norm": 0.8553708618629486, "learning_rate": 1.0014926504969535e-06, "loss": 1.0022, "step": 3051 }, { "epoch": 4.080267558528428, "grad_norm": 0.8141786816650026, "learning_rate": 9.986925424609633e-07, "loss": 1.1278, "step": 3052 }, { "epoch": 4.081605351170569, "grad_norm": 0.9594205287613319, "learning_rate": 9.95895919991804e-07, "loss": 0.9511, "step": 3053 }, { "epoch": 4.082943143812709, "grad_norm": 0.8011465054836964, "learning_rate": 9.93102785525632e-07, "loss": 1.0766, "step": 3054 }, { "epoch": 4.084280936454849, "grad_norm": 0.9384385426532851, "learning_rate": 9.903131414955674e-07, "loss": 0.9776, "step": 3055 }, { "epoch": 4.08561872909699, "grad_norm": 0.894558537671233, "learning_rate": 9.87526990331688e-07, "loss": 1.0396, "step": 3056 }, { "epoch": 4.086956521739131, "grad_norm": 0.956217280743456, "learning_rate": 9.847443344610296e-07, "loss": 1.0106, "step": 3057 }, { "epoch": 4.088294314381271, "grad_norm": 1.3412648339259645, "learning_rate": 9.819651763075833e-07, "loss": 0.7196, "step": 3058 }, { "epoch": 4.089632107023411, "grad_norm": 0.9947518131313292, "learning_rate": 9.791895182922911e-07, "loss": 0.9064, "step": 3059 }, { "epoch": 4.090969899665552, "grad_norm": 1.3824352903445005, "learning_rate": 9.764173628330514e-07, "loss": 0.7517, "step": 3060 }, { "epoch": 4.092307692307692, "grad_norm": 0.7973019178545656, "learning_rate": 9.73648712344707e-07, "loss": 0.7543, "step": 3061 }, { "epoch": 4.093645484949833, "grad_norm": 0.8646427936209167, "learning_rate": 9.708835692390483e-07, "loss": 1.0431, "step": 3062 }, { "epoch": 4.0949832775919734, "grad_norm": 0.9513512864214965, "learning_rate": 9.681219359248106e-07, "loss": 0.8463, "step": 3063 }, { "epoch": 4.096321070234113, "grad_norm": 1.2828211433191599, "learning_rate": 9.65363814807672e-07, "loss": 0.6538, "step": 3064 }, { "epoch": 4.097658862876254, "grad_norm": 1.1622693033673321, "learning_rate": 9.626092082902511e-07, "loss": 1.0957, "step": 3065 }, { "epoch": 4.098996655518395, "grad_norm": 0.9000928312390667, "learning_rate": 9.59858118772105e-07, "loss": 0.6632, "step": 3066 }, { "epoch": 4.1003344481605355, "grad_norm": 0.9938360042093286, "learning_rate": 9.571105486497268e-07, "loss": 1.1561, "step": 3067 }, { "epoch": 4.101672240802675, "grad_norm": 0.7723906516538165, "learning_rate": 9.543665003165442e-07, "loss": 0.8583, "step": 3068 }, { "epoch": 4.103010033444816, "grad_norm": 1.207359321858884, "learning_rate": 9.516259761629148e-07, "loss": 0.9469, "step": 3069 }, { "epoch": 4.104347826086957, "grad_norm": 1.136869290046632, "learning_rate": 9.488889785761324e-07, "loss": 0.9074, "step": 3070 }, { "epoch": 4.105685618729097, "grad_norm": 1.1523569489309455, "learning_rate": 9.461555099404119e-07, "loss": 0.7287, "step": 3071 }, { "epoch": 4.107023411371237, "grad_norm": 0.843532372414803, "learning_rate": 9.434255726368974e-07, "loss": 0.9696, "step": 3072 }, { "epoch": 4.108361204013378, "grad_norm": 1.1374881477381689, "learning_rate": 9.406991690436567e-07, "loss": 0.8005, "step": 3073 }, { "epoch": 4.109698996655518, "grad_norm": 1.0779663887379347, "learning_rate": 9.379763015356785e-07, "loss": 0.9444, "step": 3074 }, { "epoch": 4.111036789297659, "grad_norm": 0.8769857252650088, "learning_rate": 9.352569724848715e-07, "loss": 0.8381, "step": 3075 }, { "epoch": 4.112374581939799, "grad_norm": 1.1668425222894656, "learning_rate": 9.325411842600629e-07, "loss": 0.9653, "step": 3076 }, { "epoch": 4.11371237458194, "grad_norm": 1.2985269038965672, "learning_rate": 9.298289392269944e-07, "loss": 0.8724, "step": 3077 }, { "epoch": 4.11505016722408, "grad_norm": 0.9562891271073687, "learning_rate": 9.271202397483214e-07, "loss": 1.0257, "step": 3078 }, { "epoch": 4.116387959866221, "grad_norm": 1.1930106505165903, "learning_rate": 9.244150881836117e-07, "loss": 0.9237, "step": 3079 }, { "epoch": 4.1177257525083615, "grad_norm": 1.0045448406353186, "learning_rate": 9.217134868893401e-07, "loss": 0.7192, "step": 3080 }, { "epoch": 4.119063545150501, "grad_norm": 0.921817822307301, "learning_rate": 9.190154382188921e-07, "loss": 1.009, "step": 3081 }, { "epoch": 4.120401337792642, "grad_norm": 0.9147782035415966, "learning_rate": 9.163209445225557e-07, "loss": 1.1223, "step": 3082 }, { "epoch": 4.121739130434783, "grad_norm": 1.0133110211327268, "learning_rate": 9.13630008147523e-07, "loss": 0.9081, "step": 3083 }, { "epoch": 4.123076923076923, "grad_norm": 0.9392673767952657, "learning_rate": 9.109426314378878e-07, "loss": 0.7439, "step": 3084 }, { "epoch": 4.124414715719063, "grad_norm": 1.179978194854883, "learning_rate": 9.082588167346428e-07, "loss": 0.7247, "step": 3085 }, { "epoch": 4.125752508361204, "grad_norm": 1.2514088793078848, "learning_rate": 9.055785663756778e-07, "loss": 1.0025, "step": 3086 }, { "epoch": 4.127090301003345, "grad_norm": 0.793545613519867, "learning_rate": 9.029018826957775e-07, "loss": 0.8609, "step": 3087 }, { "epoch": 4.128428093645485, "grad_norm": 0.8398599947096222, "learning_rate": 9.002287680266192e-07, "loss": 0.7975, "step": 3088 }, { "epoch": 4.129765886287625, "grad_norm": 1.245952308836207, "learning_rate": 8.975592246967713e-07, "loss": 0.8091, "step": 3089 }, { "epoch": 4.131103678929766, "grad_norm": 0.8119053430887362, "learning_rate": 8.948932550316935e-07, "loss": 0.8401, "step": 3090 }, { "epoch": 4.132441471571906, "grad_norm": 0.9288364454699554, "learning_rate": 8.922308613537295e-07, "loss": 1.1727, "step": 3091 }, { "epoch": 4.133779264214047, "grad_norm": 0.8450586743404618, "learning_rate": 8.895720459821089e-07, "loss": 0.7835, "step": 3092 }, { "epoch": 4.1351170568561875, "grad_norm": 1.186721089611921, "learning_rate": 8.86916811232944e-07, "loss": 0.9538, "step": 3093 }, { "epoch": 4.136454849498328, "grad_norm": 0.998729855042994, "learning_rate": 8.842651594192292e-07, "loss": 0.938, "step": 3094 }, { "epoch": 4.137792642140468, "grad_norm": 1.0181840341509647, "learning_rate": 8.816170928508367e-07, "loss": 1.0502, "step": 3095 }, { "epoch": 4.139130434782609, "grad_norm": 0.8307292637779197, "learning_rate": 8.78972613834515e-07, "loss": 0.8279, "step": 3096 }, { "epoch": 4.1404682274247495, "grad_norm": 0.9792904768984686, "learning_rate": 8.763317246738889e-07, "loss": 0.8911, "step": 3097 }, { "epoch": 4.141806020066889, "grad_norm": 1.4039316383152245, "learning_rate": 8.736944276694548e-07, "loss": 0.9123, "step": 3098 }, { "epoch": 4.14314381270903, "grad_norm": 0.89467462211456, "learning_rate": 8.710607251185799e-07, "loss": 0.9509, "step": 3099 }, { "epoch": 4.144481605351171, "grad_norm": 0.9387856632572124, "learning_rate": 8.684306193155034e-07, "loss": 1.0477, "step": 3100 }, { "epoch": 4.145819397993311, "grad_norm": 0.7799350854895337, "learning_rate": 8.658041125513267e-07, "loss": 0.93, "step": 3101 }, { "epoch": 4.147157190635451, "grad_norm": 0.8638870864714358, "learning_rate": 8.631812071140189e-07, "loss": 0.8992, "step": 3102 }, { "epoch": 4.148494983277592, "grad_norm": 0.9199664103295864, "learning_rate": 8.605619052884106e-07, "loss": 0.7236, "step": 3103 }, { "epoch": 4.149832775919732, "grad_norm": 0.7928289577012629, "learning_rate": 8.579462093561947e-07, "loss": 0.7944, "step": 3104 }, { "epoch": 4.151170568561873, "grad_norm": 0.8407996865350826, "learning_rate": 8.553341215959215e-07, "loss": 0.7933, "step": 3105 }, { "epoch": 4.1525083612040135, "grad_norm": 0.8939677968656649, "learning_rate": 8.527256442829995e-07, "loss": 0.8467, "step": 3106 }, { "epoch": 4.153846153846154, "grad_norm": 1.0682388817265194, "learning_rate": 8.50120779689691e-07, "loss": 0.6048, "step": 3107 }, { "epoch": 4.155183946488294, "grad_norm": 1.0000683946914148, "learning_rate": 8.475195300851113e-07, "loss": 1.1365, "step": 3108 }, { "epoch": 4.156521739130435, "grad_norm": 0.9073343216574019, "learning_rate": 8.449218977352281e-07, "loss": 0.7883, "step": 3109 }, { "epoch": 4.1578595317725755, "grad_norm": 0.9012449386524988, "learning_rate": 8.423278849028565e-07, "loss": 0.7777, "step": 3110 }, { "epoch": 4.159197324414715, "grad_norm": 0.9670495860986333, "learning_rate": 8.397374938476594e-07, "loss": 0.752, "step": 3111 }, { "epoch": 4.160535117056856, "grad_norm": 0.9480848970798407, "learning_rate": 8.371507268261436e-07, "loss": 1.1032, "step": 3112 }, { "epoch": 4.161872909698997, "grad_norm": 1.0836372769718732, "learning_rate": 8.345675860916613e-07, "loss": 0.8259, "step": 3113 }, { "epoch": 4.1632107023411375, "grad_norm": 0.8280088845746553, "learning_rate": 8.31988073894403e-07, "loss": 1.1212, "step": 3114 }, { "epoch": 4.164548494983277, "grad_norm": 1.2468168407663562, "learning_rate": 8.294121924814014e-07, "loss": 0.8732, "step": 3115 }, { "epoch": 4.165886287625418, "grad_norm": 1.0931653984214362, "learning_rate": 8.26839944096523e-07, "loss": 0.8795, "step": 3116 }, { "epoch": 4.167224080267559, "grad_norm": 1.1957401369294458, "learning_rate": 8.242713309804729e-07, "loss": 0.7203, "step": 3117 }, { "epoch": 4.168561872909699, "grad_norm": 0.9986988984640831, "learning_rate": 8.217063553707865e-07, "loss": 0.7565, "step": 3118 }, { "epoch": 4.169899665551839, "grad_norm": 0.9684301827248636, "learning_rate": 8.191450195018313e-07, "loss": 0.89, "step": 3119 }, { "epoch": 4.17123745819398, "grad_norm": 0.8943697055135564, "learning_rate": 8.165873256048079e-07, "loss": 0.8865, "step": 3120 }, { "epoch": 4.17257525083612, "grad_norm": 0.7659067538302902, "learning_rate": 8.140332759077397e-07, "loss": 0.9287, "step": 3121 }, { "epoch": 4.173913043478261, "grad_norm": 1.0804697002469983, "learning_rate": 8.114828726354762e-07, "loss": 0.9765, "step": 3122 }, { "epoch": 4.1752508361204015, "grad_norm": 1.055301761316934, "learning_rate": 8.089361180096927e-07, "loss": 0.876, "step": 3123 }, { "epoch": 4.176588628762542, "grad_norm": 1.0360732365343885, "learning_rate": 8.063930142488846e-07, "loss": 0.9689, "step": 3124 }, { "epoch": 4.177926421404682, "grad_norm": 0.8310583371441614, "learning_rate": 8.03853563568367e-07, "loss": 0.9337, "step": 3125 }, { "epoch": 4.179264214046823, "grad_norm": 1.0234549929226564, "learning_rate": 8.013177681802736e-07, "loss": 0.9529, "step": 3126 }, { "epoch": 4.1806020066889635, "grad_norm": 1.3066692371590671, "learning_rate": 7.987856302935532e-07, "loss": 1.1059, "step": 3127 }, { "epoch": 4.181939799331103, "grad_norm": 0.8657148565377589, "learning_rate": 7.962571521139684e-07, "loss": 1.212, "step": 3128 }, { "epoch": 4.183277591973244, "grad_norm": 1.1220397811235636, "learning_rate": 7.937323358440935e-07, "loss": 0.9416, "step": 3129 }, { "epoch": 4.184615384615385, "grad_norm": 1.227938040802187, "learning_rate": 7.912111836833158e-07, "loss": 0.9961, "step": 3130 }, { "epoch": 4.185953177257525, "grad_norm": 1.068381561914134, "learning_rate": 7.886936978278276e-07, "loss": 0.9523, "step": 3131 }, { "epoch": 4.187290969899665, "grad_norm": 0.884543288019862, "learning_rate": 7.861798804706278e-07, "loss": 0.5968, "step": 3132 }, { "epoch": 4.188628762541806, "grad_norm": 1.179030680196975, "learning_rate": 7.836697338015203e-07, "loss": 0.7554, "step": 3133 }, { "epoch": 4.189966555183947, "grad_norm": 1.0575335881679198, "learning_rate": 7.811632600071117e-07, "loss": 0.9989, "step": 3134 }, { "epoch": 4.191304347826087, "grad_norm": 0.7756849667928976, "learning_rate": 7.786604612708093e-07, "loss": 0.5765, "step": 3135 }, { "epoch": 4.1926421404682275, "grad_norm": 0.9849881684872787, "learning_rate": 7.761613397728174e-07, "loss": 1.0517, "step": 3136 }, { "epoch": 4.193979933110368, "grad_norm": 0.9615421264794369, "learning_rate": 7.73665897690139e-07, "loss": 0.9135, "step": 3137 }, { "epoch": 4.195317725752508, "grad_norm": 1.023357245970346, "learning_rate": 7.711741371965703e-07, "loss": 0.8132, "step": 3138 }, { "epoch": 4.196655518394649, "grad_norm": 0.8705180458094645, "learning_rate": 7.686860604627022e-07, "loss": 0.8627, "step": 3139 }, { "epoch": 4.1979933110367895, "grad_norm": 0.9228030349638685, "learning_rate": 7.662016696559149e-07, "loss": 0.9614, "step": 3140 }, { "epoch": 4.199331103678929, "grad_norm": 0.9279241122025407, "learning_rate": 7.637209669403789e-07, "loss": 1.0397, "step": 3141 }, { "epoch": 4.20066889632107, "grad_norm": 0.9745939128733336, "learning_rate": 7.612439544770517e-07, "loss": 0.7624, "step": 3142 }, { "epoch": 4.202006688963211, "grad_norm": 1.26975486707695, "learning_rate": 7.587706344236762e-07, "loss": 0.7061, "step": 3143 }, { "epoch": 4.203344481605352, "grad_norm": 0.8606916470265408, "learning_rate": 7.563010089347789e-07, "loss": 0.8805, "step": 3144 }, { "epoch": 4.204682274247491, "grad_norm": 0.7834805187132585, "learning_rate": 7.538350801616673e-07, "loss": 1.0439, "step": 3145 }, { "epoch": 4.206020066889632, "grad_norm": 1.0137713065609797, "learning_rate": 7.513728502524286e-07, "loss": 1.0556, "step": 3146 }, { "epoch": 4.207357859531773, "grad_norm": 1.059398593893106, "learning_rate": 7.489143213519301e-07, "loss": 1.0305, "step": 3147 }, { "epoch": 4.208695652173913, "grad_norm": 0.8061273759118233, "learning_rate": 7.464594956018124e-07, "loss": 0.7012, "step": 3148 }, { "epoch": 4.2100334448160535, "grad_norm": 0.9541460655516458, "learning_rate": 7.440083751404902e-07, "loss": 1.0981, "step": 3149 }, { "epoch": 4.211371237458194, "grad_norm": 0.9762952725088351, "learning_rate": 7.415609621031539e-07, "loss": 0.8372, "step": 3150 }, { "epoch": 4.212709030100334, "grad_norm": 0.8536947174714196, "learning_rate": 7.39117258621761e-07, "loss": 0.864, "step": 3151 }, { "epoch": 4.214046822742475, "grad_norm": 1.0857205218407486, "learning_rate": 7.366772668250394e-07, "loss": 0.7624, "step": 3152 }, { "epoch": 4.2153846153846155, "grad_norm": 0.8565901836574558, "learning_rate": 7.342409888384816e-07, "loss": 1.2038, "step": 3153 }, { "epoch": 4.216722408026756, "grad_norm": 0.8284954287940969, "learning_rate": 7.318084267843473e-07, "loss": 1.2776, "step": 3154 }, { "epoch": 4.218060200668896, "grad_norm": 1.0205773002903182, "learning_rate": 7.29379582781658e-07, "loss": 0.8366, "step": 3155 }, { "epoch": 4.219397993311037, "grad_norm": 1.033429712095725, "learning_rate": 7.269544589461968e-07, "loss": 0.8127, "step": 3156 }, { "epoch": 4.2207357859531776, "grad_norm": 1.1359955354175142, "learning_rate": 7.245330573905058e-07, "loss": 0.9961, "step": 3157 }, { "epoch": 4.222073578595317, "grad_norm": 0.8384511857293151, "learning_rate": 7.221153802238845e-07, "loss": 1.0077, "step": 3158 }, { "epoch": 4.223411371237458, "grad_norm": 0.8719451486548933, "learning_rate": 7.197014295523879e-07, "loss": 1.0659, "step": 3159 }, { "epoch": 4.224749163879599, "grad_norm": 0.7971499004089717, "learning_rate": 7.172912074788274e-07, "loss": 0.7209, "step": 3160 }, { "epoch": 4.226086956521739, "grad_norm": 0.9508273145188315, "learning_rate": 7.148847161027622e-07, "loss": 1.2636, "step": 3161 }, { "epoch": 4.2274247491638794, "grad_norm": 1.0962583956624776, "learning_rate": 7.12481957520505e-07, "loss": 0.9316, "step": 3162 }, { "epoch": 4.22876254180602, "grad_norm": 1.089518073127911, "learning_rate": 7.100829338251147e-07, "loss": 1.2339, "step": 3163 }, { "epoch": 4.230100334448161, "grad_norm": 1.0289294364071757, "learning_rate": 7.076876471063976e-07, "loss": 0.7853, "step": 3164 }, { "epoch": 4.231438127090301, "grad_norm": 0.9807035981748832, "learning_rate": 7.052960994509056e-07, "loss": 1.0173, "step": 3165 }, { "epoch": 4.2327759197324415, "grad_norm": 1.1254775085779083, "learning_rate": 7.029082929419312e-07, "loss": 0.856, "step": 3166 }, { "epoch": 4.234113712374582, "grad_norm": 0.9983504422253662, "learning_rate": 7.005242296595099e-07, "loss": 1.0626, "step": 3167 }, { "epoch": 4.235451505016722, "grad_norm": 0.8181544238762191, "learning_rate": 6.981439116804161e-07, "loss": 0.9955, "step": 3168 }, { "epoch": 4.236789297658863, "grad_norm": 1.1951492817705467, "learning_rate": 6.957673410781617e-07, "loss": 1.0999, "step": 3169 }, { "epoch": 4.2381270903010035, "grad_norm": 0.8994872039829458, "learning_rate": 6.93394519922993e-07, "loss": 0.9645, "step": 3170 }, { "epoch": 4.239464882943143, "grad_norm": 1.0909223821493759, "learning_rate": 6.910254502818914e-07, "loss": 1.0498, "step": 3171 }, { "epoch": 4.240802675585284, "grad_norm": 0.9031490717787896, "learning_rate": 6.886601342185701e-07, "loss": 0.8792, "step": 3172 }, { "epoch": 4.242140468227425, "grad_norm": 0.9030748907305851, "learning_rate": 6.862985737934724e-07, "loss": 1.0646, "step": 3173 }, { "epoch": 4.243478260869566, "grad_norm": 0.960885910371266, "learning_rate": 6.839407710637696e-07, "loss": 0.7684, "step": 3174 }, { "epoch": 4.244816053511705, "grad_norm": 1.016816212406333, "learning_rate": 6.815867280833611e-07, "loss": 0.8752, "step": 3175 }, { "epoch": 4.246153846153846, "grad_norm": 1.0293604766882445, "learning_rate": 6.792364469028695e-07, "loss": 0.7042, "step": 3176 }, { "epoch": 4.247491638795987, "grad_norm": 0.8105875336615989, "learning_rate": 6.768899295696413e-07, "loss": 1.0089, "step": 3177 }, { "epoch": 4.248829431438127, "grad_norm": 1.0072837604999378, "learning_rate": 6.745471781277435e-07, "loss": 1.0579, "step": 3178 }, { "epoch": 4.2501672240802675, "grad_norm": 0.8109672041441471, "learning_rate": 6.722081946179631e-07, "loss": 0.9156, "step": 3179 }, { "epoch": 4.251505016722408, "grad_norm": 0.9461325114479914, "learning_rate": 6.698729810778065e-07, "loss": 0.9651, "step": 3180 }, { "epoch": 4.252842809364548, "grad_norm": 0.9954555201916447, "learning_rate": 6.675415395414942e-07, "loss": 0.7834, "step": 3181 }, { "epoch": 4.254180602006689, "grad_norm": 0.8395208174045494, "learning_rate": 6.652138720399598e-07, "loss": 0.9042, "step": 3182 }, { "epoch": 4.2555183946488295, "grad_norm": 0.8693346099680652, "learning_rate": 6.628899806008515e-07, "loss": 1.0094, "step": 3183 }, { "epoch": 4.25685618729097, "grad_norm": 1.5605875729390795, "learning_rate": 6.605698672485278e-07, "loss": 0.8906, "step": 3184 }, { "epoch": 4.25819397993311, "grad_norm": 1.22256034661141, "learning_rate": 6.582535340040547e-07, "loss": 0.8629, "step": 3185 }, { "epoch": 4.259531772575251, "grad_norm": 1.2895380798160643, "learning_rate": 6.55940982885207e-07, "loss": 0.9419, "step": 3186 }, { "epoch": 4.260869565217392, "grad_norm": 1.0254455321361278, "learning_rate": 6.536322159064634e-07, "loss": 0.748, "step": 3187 }, { "epoch": 4.262207357859531, "grad_norm": 1.1462912055481853, "learning_rate": 6.513272350790079e-07, "loss": 0.9622, "step": 3188 }, { "epoch": 4.263545150501672, "grad_norm": 1.1106600778285725, "learning_rate": 6.490260424107231e-07, "loss": 0.7451, "step": 3189 }, { "epoch": 4.264882943143813, "grad_norm": 0.9484948651157307, "learning_rate": 6.467286399061967e-07, "loss": 0.7425, "step": 3190 }, { "epoch": 4.266220735785954, "grad_norm": 1.1813886564496818, "learning_rate": 6.444350295667112e-07, "loss": 1.1083, "step": 3191 }, { "epoch": 4.2675585284280935, "grad_norm": 0.8125567367637633, "learning_rate": 6.421452133902467e-07, "loss": 0.7122, "step": 3192 }, { "epoch": 4.268896321070234, "grad_norm": 0.8360852396557095, "learning_rate": 6.398591933714771e-07, "loss": 0.9651, "step": 3193 }, { "epoch": 4.270234113712375, "grad_norm": 1.0009306510993563, "learning_rate": 6.375769715017716e-07, "loss": 0.9345, "step": 3194 }, { "epoch": 4.271571906354515, "grad_norm": 0.7378652064327724, "learning_rate": 6.352985497691883e-07, "loss": 0.7055, "step": 3195 }, { "epoch": 4.2729096989966555, "grad_norm": 0.9262939593854368, "learning_rate": 6.330239301584773e-07, "loss": 0.9308, "step": 3196 }, { "epoch": 4.274247491638796, "grad_norm": 1.0612625069417436, "learning_rate": 6.307531146510754e-07, "loss": 1.0129, "step": 3197 }, { "epoch": 4.275585284280936, "grad_norm": 0.9127126341209495, "learning_rate": 6.284861052251062e-07, "loss": 0.9167, "step": 3198 }, { "epoch": 4.276923076923077, "grad_norm": 1.0193368499896465, "learning_rate": 6.262229038553752e-07, "loss": 1.1987, "step": 3199 }, { "epoch": 4.278260869565218, "grad_norm": 0.8466978237069193, "learning_rate": 6.239635125133753e-07, "loss": 0.9475, "step": 3200 }, { "epoch": 4.279598662207357, "grad_norm": 0.9294648426066319, "learning_rate": 6.217079331672777e-07, "loss": 0.8343, "step": 3201 }, { "epoch": 4.280936454849498, "grad_norm": 0.8279140501206602, "learning_rate": 6.194561677819327e-07, "loss": 1.1389, "step": 3202 }, { "epoch": 4.282274247491639, "grad_norm": 0.8579548763986076, "learning_rate": 6.172082183188688e-07, "loss": 0.9986, "step": 3203 }, { "epoch": 4.28361204013378, "grad_norm": 1.1335065652188618, "learning_rate": 6.14964086736291e-07, "loss": 0.9076, "step": 3204 }, { "epoch": 4.2849498327759195, "grad_norm": 1.0351498103582868, "learning_rate": 6.12723774989078e-07, "loss": 1.128, "step": 3205 }, { "epoch": 4.28628762541806, "grad_norm": 0.8232257710534557, "learning_rate": 6.104872850287802e-07, "loss": 0.9787, "step": 3206 }, { "epoch": 4.287625418060201, "grad_norm": 1.4979228299040184, "learning_rate": 6.082546188036204e-07, "loss": 0.8411, "step": 3207 }, { "epoch": 4.288963210702341, "grad_norm": 0.8147359707569908, "learning_rate": 6.060257782584889e-07, "loss": 1.185, "step": 3208 }, { "epoch": 4.2903010033444815, "grad_norm": 1.0086257982608322, "learning_rate": 6.038007653349437e-07, "loss": 1.0129, "step": 3209 }, { "epoch": 4.291638795986622, "grad_norm": 0.9104896433966623, "learning_rate": 6.015795819712117e-07, "loss": 0.9135, "step": 3210 }, { "epoch": 4.292976588628763, "grad_norm": 0.9085717643455332, "learning_rate": 5.99362230102179e-07, "loss": 0.7951, "step": 3211 }, { "epoch": 4.294314381270903, "grad_norm": 0.8469118210688888, "learning_rate": 5.971487116593977e-07, "loss": 1.0017, "step": 3212 }, { "epoch": 4.2956521739130435, "grad_norm": 1.252083133515555, "learning_rate": 5.949390285710777e-07, "loss": 0.9104, "step": 3213 }, { "epoch": 4.296989966555184, "grad_norm": 1.3104102666220023, "learning_rate": 5.927331827620902e-07, "loss": 0.8892, "step": 3214 }, { "epoch": 4.298327759197324, "grad_norm": 0.9773600216538079, "learning_rate": 5.905311761539622e-07, "loss": 0.9189, "step": 3215 }, { "epoch": 4.299665551839465, "grad_norm": 1.0892191300107572, "learning_rate": 5.883330106648782e-07, "loss": 0.6306, "step": 3216 }, { "epoch": 4.301003344481606, "grad_norm": 0.8359028399946533, "learning_rate": 5.861386882096743e-07, "loss": 0.9844, "step": 3217 }, { "epoch": 4.302341137123745, "grad_norm": 0.8997116154133661, "learning_rate": 5.839482106998406e-07, "loss": 1.0311, "step": 3218 }, { "epoch": 4.303678929765886, "grad_norm": 0.9141345895588102, "learning_rate": 5.817615800435167e-07, "loss": 1.0072, "step": 3219 }, { "epoch": 4.305016722408027, "grad_norm": 0.9741489348280798, "learning_rate": 5.795787981454931e-07, "loss": 1.0079, "step": 3220 }, { "epoch": 4.306354515050167, "grad_norm": 0.9519199641690299, "learning_rate": 5.773998669072057e-07, "loss": 0.9164, "step": 3221 }, { "epoch": 4.3076923076923075, "grad_norm": 0.8324629048311498, "learning_rate": 5.752247882267365e-07, "loss": 1.02, "step": 3222 }, { "epoch": 4.309030100334448, "grad_norm": 1.0696773372292196, "learning_rate": 5.730535639988122e-07, "loss": 1.0489, "step": 3223 }, { "epoch": 4.310367892976589, "grad_norm": 0.9091144927751381, "learning_rate": 5.708861961148004e-07, "loss": 1.0724, "step": 3224 }, { "epoch": 4.311705685618729, "grad_norm": 1.11346260683661, "learning_rate": 5.687226864627115e-07, "loss": 0.951, "step": 3225 }, { "epoch": 4.3130434782608695, "grad_norm": 1.0276544740403917, "learning_rate": 5.665630369271935e-07, "loss": 1.0079, "step": 3226 }, { "epoch": 4.31438127090301, "grad_norm": 0.7450401819846576, "learning_rate": 5.644072493895325e-07, "loss": 0.8038, "step": 3227 }, { "epoch": 4.31571906354515, "grad_norm": 0.7772910865675335, "learning_rate": 5.622553257276487e-07, "loss": 1.0292, "step": 3228 }, { "epoch": 4.317056856187291, "grad_norm": 1.0203963630710733, "learning_rate": 5.60107267816098e-07, "loss": 0.9876, "step": 3229 }, { "epoch": 4.318394648829432, "grad_norm": 0.8902033472862917, "learning_rate": 5.579630775260697e-07, "loss": 0.8848, "step": 3230 }, { "epoch": 4.319732441471572, "grad_norm": 0.9731414276836905, "learning_rate": 5.558227567253832e-07, "loss": 0.7456, "step": 3231 }, { "epoch": 4.321070234113712, "grad_norm": 1.031986123024288, "learning_rate": 5.53686307278486e-07, "loss": 0.9622, "step": 3232 }, { "epoch": 4.322408026755853, "grad_norm": 0.8362944863510848, "learning_rate": 5.515537310464536e-07, "loss": 0.9896, "step": 3233 }, { "epoch": 4.323745819397994, "grad_norm": 0.8064186525921855, "learning_rate": 5.494250298869896e-07, "loss": 1.0672, "step": 3234 }, { "epoch": 4.3250836120401335, "grad_norm": 0.8728100471077637, "learning_rate": 5.473002056544191e-07, "loss": 1.1079, "step": 3235 }, { "epoch": 4.326421404682274, "grad_norm": 0.8884031387081069, "learning_rate": 5.45179260199692e-07, "loss": 1.1479, "step": 3236 }, { "epoch": 4.327759197324415, "grad_norm": 1.1016172554054504, "learning_rate": 5.430621953703785e-07, "loss": 0.6965, "step": 3237 }, { "epoch": 4.329096989966555, "grad_norm": 1.1640786934157223, "learning_rate": 5.409490130106682e-07, "loss": 0.9455, "step": 3238 }, { "epoch": 4.3304347826086955, "grad_norm": 0.9223253078170822, "learning_rate": 5.388397149613683e-07, "loss": 0.9567, "step": 3239 }, { "epoch": 4.331772575250836, "grad_norm": 1.0451609155251829, "learning_rate": 5.367343030599054e-07, "loss": 0.947, "step": 3240 }, { "epoch": 4.333110367892977, "grad_norm": 0.8689704139694082, "learning_rate": 5.346327791403167e-07, "loss": 0.9076, "step": 3241 }, { "epoch": 4.334448160535117, "grad_norm": 1.163500246662526, "learning_rate": 5.32535145033255e-07, "loss": 0.6732, "step": 3242 }, { "epoch": 4.335785953177258, "grad_norm": 0.7746597265280392, "learning_rate": 5.304414025659832e-07, "loss": 0.8366, "step": 3243 }, { "epoch": 4.337123745819398, "grad_norm": 0.9726660249515784, "learning_rate": 5.283515535623762e-07, "loss": 0.9477, "step": 3244 }, { "epoch": 4.338461538461538, "grad_norm": 0.8420737688962859, "learning_rate": 5.262655998429151e-07, "loss": 0.9446, "step": 3245 }, { "epoch": 4.339799331103679, "grad_norm": 0.8284651228281594, "learning_rate": 5.241835432246888e-07, "loss": 1.047, "step": 3246 }, { "epoch": 4.34113712374582, "grad_norm": 0.8358225106663048, "learning_rate": 5.221053855213914e-07, "loss": 0.6564, "step": 3247 }, { "epoch": 4.3424749163879595, "grad_norm": 0.9112261926028632, "learning_rate": 5.200311285433213e-07, "loss": 0.9758, "step": 3248 }, { "epoch": 4.3438127090301, "grad_norm": 0.8725749410344665, "learning_rate": 5.179607740973764e-07, "loss": 1.0344, "step": 3249 }, { "epoch": 4.345150501672241, "grad_norm": 0.8533908731879286, "learning_rate": 5.158943239870585e-07, "loss": 0.9131, "step": 3250 }, { "epoch": 4.346488294314382, "grad_norm": 0.9042211420854239, "learning_rate": 5.13831780012467e-07, "loss": 1.119, "step": 3251 }, { "epoch": 4.3478260869565215, "grad_norm": 1.3673752260824164, "learning_rate": 5.117731439702972e-07, "loss": 0.8972, "step": 3252 }, { "epoch": 4.349163879598662, "grad_norm": 1.1263078476253985, "learning_rate": 5.097184176538423e-07, "loss": 1.0521, "step": 3253 }, { "epoch": 4.350501672240803, "grad_norm": 1.0703875244278505, "learning_rate": 5.076676028529875e-07, "loss": 0.9646, "step": 3254 }, { "epoch": 4.351839464882943, "grad_norm": 0.9784736831592834, "learning_rate": 5.056207013542131e-07, "loss": 0.9684, "step": 3255 }, { "epoch": 4.3531772575250836, "grad_norm": 1.1583439130072843, "learning_rate": 5.035777149405891e-07, "loss": 0.9023, "step": 3256 }, { "epoch": 4.354515050167224, "grad_norm": 1.009419348310741, "learning_rate": 5.015386453917742e-07, "loss": 1.0526, "step": 3257 }, { "epoch": 4.355852842809364, "grad_norm": 0.9861440419908915, "learning_rate": 4.995034944840171e-07, "loss": 1.1898, "step": 3258 }, { "epoch": 4.357190635451505, "grad_norm": 1.0386986734027706, "learning_rate": 4.974722639901503e-07, "loss": 0.8989, "step": 3259 }, { "epoch": 4.358528428093646, "grad_norm": 0.875267390246978, "learning_rate": 4.954449556795948e-07, "loss": 0.8087, "step": 3260 }, { "epoch": 4.359866220735786, "grad_norm": 1.0343021525377463, "learning_rate": 4.934215713183527e-07, "loss": 0.9718, "step": 3261 }, { "epoch": 4.361204013377926, "grad_norm": 1.0608636028963798, "learning_rate": 4.914021126690083e-07, "loss": 0.8199, "step": 3262 }, { "epoch": 4.362541806020067, "grad_norm": 0.8545949875898324, "learning_rate": 4.89386581490725e-07, "loss": 1.278, "step": 3263 }, { "epoch": 4.363879598662208, "grad_norm": 1.2727763431405499, "learning_rate": 4.873749795392469e-07, "loss": 0.8553, "step": 3264 }, { "epoch": 4.3652173913043475, "grad_norm": 1.1126590486104577, "learning_rate": 4.853673085668947e-07, "loss": 0.9358, "step": 3265 }, { "epoch": 4.366555183946488, "grad_norm": 0.8568023811718782, "learning_rate": 4.833635703225637e-07, "loss": 1.0844, "step": 3266 }, { "epoch": 4.367892976588629, "grad_norm": 0.8736267762059607, "learning_rate": 4.813637665517251e-07, "loss": 0.8018, "step": 3267 }, { "epoch": 4.36923076923077, "grad_norm": 0.9732032211213986, "learning_rate": 4.793678989964207e-07, "loss": 1.166, "step": 3268 }, { "epoch": 4.3705685618729095, "grad_norm": 0.854628091919865, "learning_rate": 4.773759693952662e-07, "loss": 1.1204, "step": 3269 }, { "epoch": 4.37190635451505, "grad_norm": 1.1520331016229406, "learning_rate": 4.7538797948344485e-07, "loss": 0.9792, "step": 3270 }, { "epoch": 4.373244147157191, "grad_norm": 1.2558888734883746, "learning_rate": 4.7340393099270854e-07, "loss": 0.8754, "step": 3271 }, { "epoch": 4.374581939799331, "grad_norm": 1.5665556967193173, "learning_rate": 4.7142382565137535e-07, "loss": 0.6989, "step": 3272 }, { "epoch": 4.375919732441472, "grad_norm": 0.7672678626455475, "learning_rate": 4.6944766518432936e-07, "loss": 0.6436, "step": 3273 }, { "epoch": 4.377257525083612, "grad_norm": 0.813639636296344, "learning_rate": 4.6747545131301755e-07, "loss": 0.7883, "step": 3274 }, { "epoch": 4.378595317725752, "grad_norm": 1.5029149127316581, "learning_rate": 4.6550718575544883e-07, "loss": 0.7582, "step": 3275 }, { "epoch": 4.379933110367893, "grad_norm": 0.9302590433692418, "learning_rate": 4.635428702261929e-07, "loss": 1.1632, "step": 3276 }, { "epoch": 4.381270903010034, "grad_norm": 1.06120782079749, "learning_rate": 4.615825064363799e-07, "loss": 0.9033, "step": 3277 }, { "epoch": 4.3826086956521735, "grad_norm": 4.741833293135936, "learning_rate": 4.5962609609369436e-07, "loss": 0.9459, "step": 3278 }, { "epoch": 4.383946488294314, "grad_norm": 0.8330929488441083, "learning_rate": 4.576736409023813e-07, "loss": 0.9786, "step": 3279 }, { "epoch": 4.385284280936455, "grad_norm": 0.8713540602171218, "learning_rate": 4.5572514256323697e-07, "loss": 1.0041, "step": 3280 }, { "epoch": 4.386622073578596, "grad_norm": 1.050963892487914, "learning_rate": 4.537806027736114e-07, "loss": 0.9844, "step": 3281 }, { "epoch": 4.3879598662207355, "grad_norm": 0.7907398388275455, "learning_rate": 4.5184002322740784e-07, "loss": 0.9777, "step": 3282 }, { "epoch": 4.389297658862876, "grad_norm": 0.86861311211029, "learning_rate": 4.4990340561507805e-07, "loss": 0.7002, "step": 3283 }, { "epoch": 4.390635451505017, "grad_norm": 1.3080065767500808, "learning_rate": 4.479707516236231e-07, "loss": 0.9305, "step": 3284 }, { "epoch": 4.391973244147157, "grad_norm": 1.3375625119117656, "learning_rate": 4.460420629365919e-07, "loss": 0.6828, "step": 3285 }, { "epoch": 4.393311036789298, "grad_norm": 1.0303706315057242, "learning_rate": 4.441173412340777e-07, "loss": 0.8039, "step": 3286 }, { "epoch": 4.394648829431438, "grad_norm": 0.8192324604819926, "learning_rate": 4.4219658819271925e-07, "loss": 0.873, "step": 3287 }, { "epoch": 4.395986622073579, "grad_norm": 0.8962105925896519, "learning_rate": 4.402798054856977e-07, "loss": 0.9188, "step": 3288 }, { "epoch": 4.397324414715719, "grad_norm": 0.8320976146788925, "learning_rate": 4.383669947827368e-07, "loss": 1.0179, "step": 3289 }, { "epoch": 4.39866220735786, "grad_norm": 0.8337403534715965, "learning_rate": 4.364581577500987e-07, "loss": 1.0326, "step": 3290 }, { "epoch": 4.4, "grad_norm": 0.8610858620331471, "learning_rate": 4.3455329605058436e-07, "loss": 0.9706, "step": 3291 }, { "epoch": 4.40133779264214, "grad_norm": 0.7787262850588783, "learning_rate": 4.3265241134353265e-07, "loss": 0.9576, "step": 3292 }, { "epoch": 4.402675585284281, "grad_norm": 0.9667414419186946, "learning_rate": 4.307555052848178e-07, "loss": 1.0507, "step": 3293 }, { "epoch": 4.404013377926422, "grad_norm": 0.8220618070992547, "learning_rate": 4.288625795268464e-07, "loss": 0.8218, "step": 3294 }, { "epoch": 4.4053511705685615, "grad_norm": 0.8765415801110905, "learning_rate": 4.269736357185611e-07, "loss": 1.0276, "step": 3295 }, { "epoch": 4.406688963210702, "grad_norm": 0.9831104777805477, "learning_rate": 4.250886755054329e-07, "loss": 1.0433, "step": 3296 }, { "epoch": 4.408026755852843, "grad_norm": 1.2440680168975817, "learning_rate": 4.232077005294638e-07, "loss": 0.8346, "step": 3297 }, { "epoch": 4.409364548494983, "grad_norm": 0.8550198151761699, "learning_rate": 4.213307124291838e-07, "loss": 0.9206, "step": 3298 }, { "epoch": 4.410702341137124, "grad_norm": 1.0901968509003754, "learning_rate": 4.194577128396521e-07, "loss": 0.8911, "step": 3299 }, { "epoch": 4.412040133779264, "grad_norm": 0.9997655349192132, "learning_rate": 4.175887033924503e-07, "loss": 1.16, "step": 3300 }, { "epoch": 4.413377926421405, "grad_norm": 0.7850210608172012, "learning_rate": 4.15723685715686e-07, "loss": 0.941, "step": 3301 }, { "epoch": 4.414715719063545, "grad_norm": 1.1631241303044457, "learning_rate": 4.1386266143398855e-07, "loss": 0.8766, "step": 3302 }, { "epoch": 4.416053511705686, "grad_norm": 0.972574098932745, "learning_rate": 4.120056321685101e-07, "loss": 0.8334, "step": 3303 }, { "epoch": 4.417391304347826, "grad_norm": 0.8232627627078564, "learning_rate": 4.10152599536921e-07, "loss": 0.9811, "step": 3304 }, { "epoch": 4.418729096989966, "grad_norm": 1.0121669678050997, "learning_rate": 4.0830356515341173e-07, "loss": 0.8152, "step": 3305 }, { "epoch": 4.420066889632107, "grad_norm": 1.106590308729796, "learning_rate": 4.064585306286878e-07, "loss": 1.0658, "step": 3306 }, { "epoch": 4.421404682274248, "grad_norm": 1.0992254774469414, "learning_rate": 4.046174975699729e-07, "loss": 0.8662, "step": 3307 }, { "epoch": 4.422742474916388, "grad_norm": 0.9552494829860144, "learning_rate": 4.027804675810021e-07, "loss": 0.9564, "step": 3308 }, { "epoch": 4.424080267558528, "grad_norm": 0.7573452955700372, "learning_rate": 4.009474422620269e-07, "loss": 0.84, "step": 3309 }, { "epoch": 4.425418060200669, "grad_norm": 0.7615747918627476, "learning_rate": 3.9911842320980777e-07, "loss": 0.8557, "step": 3310 }, { "epoch": 4.42675585284281, "grad_norm": 1.0960660508833895, "learning_rate": 3.972934120176164e-07, "loss": 0.7745, "step": 3311 }, { "epoch": 4.4280936454849495, "grad_norm": 0.8418998724903212, "learning_rate": 3.9547241027523164e-07, "loss": 1.1389, "step": 3312 }, { "epoch": 4.42943143812709, "grad_norm": 0.8647254023997626, "learning_rate": 3.936554195689418e-07, "loss": 0.9105, "step": 3313 }, { "epoch": 4.430769230769231, "grad_norm": 0.8482863174342049, "learning_rate": 3.9184244148154025e-07, "loss": 0.9505, "step": 3314 }, { "epoch": 4.432107023411371, "grad_norm": 0.9909219214478295, "learning_rate": 3.900334775923237e-07, "loss": 0.7749, "step": 3315 }, { "epoch": 4.433444816053512, "grad_norm": 1.000388864482503, "learning_rate": 3.882285294770938e-07, "loss": 1.0027, "step": 3316 }, { "epoch": 4.434782608695652, "grad_norm": 0.892561614225708, "learning_rate": 3.864275987081539e-07, "loss": 1.0592, "step": 3317 }, { "epoch": 4.436120401337792, "grad_norm": 0.9648290473883153, "learning_rate": 3.846306868543054e-07, "loss": 1.1871, "step": 3318 }, { "epoch": 4.437458193979933, "grad_norm": 0.7648748148445198, "learning_rate": 3.828377954808538e-07, "loss": 0.8624, "step": 3319 }, { "epoch": 4.438795986622074, "grad_norm": 1.3709352240740371, "learning_rate": 3.8104892614959757e-07, "loss": 0.8036, "step": 3320 }, { "epoch": 4.440133779264214, "grad_norm": 1.3429411329362049, "learning_rate": 3.7926408041883355e-07, "loss": 0.9687, "step": 3321 }, { "epoch": 4.441471571906354, "grad_norm": 1.0129709725781808, "learning_rate": 3.774832598433531e-07, "loss": 1.1284, "step": 3322 }, { "epoch": 4.442809364548495, "grad_norm": 0.7886228216082827, "learning_rate": 3.7570646597444196e-07, "loss": 0.9327, "step": 3323 }, { "epoch": 4.444147157190636, "grad_norm": 0.777261970649713, "learning_rate": 3.7393370035987697e-07, "loss": 0.9187, "step": 3324 }, { "epoch": 4.4454849498327755, "grad_norm": 1.3746223665412896, "learning_rate": 3.721649645439268e-07, "loss": 1.0112, "step": 3325 }, { "epoch": 4.446822742474916, "grad_norm": 1.0729972234332097, "learning_rate": 3.704002600673501e-07, "loss": 0.9889, "step": 3326 }, { "epoch": 4.448160535117057, "grad_norm": 0.9242516938756785, "learning_rate": 3.6863958846739213e-07, "loss": 1.0484, "step": 3327 }, { "epoch": 4.449498327759198, "grad_norm": 1.0987231456219912, "learning_rate": 3.668829512777866e-07, "loss": 1.1536, "step": 3328 }, { "epoch": 4.450836120401338, "grad_norm": 0.859937961741623, "learning_rate": 3.651303500287534e-07, "loss": 1.0275, "step": 3329 }, { "epoch": 4.452173913043478, "grad_norm": 0.9713756337062875, "learning_rate": 3.6338178624699516e-07, "loss": 0.6971, "step": 3330 }, { "epoch": 4.453511705685619, "grad_norm": 0.9780617338967879, "learning_rate": 3.6163726145569787e-07, "loss": 0.8886, "step": 3331 }, { "epoch": 4.454849498327759, "grad_norm": 0.9474752024728547, "learning_rate": 3.5989677717452933e-07, "loss": 0.9238, "step": 3332 }, { "epoch": 4.4561872909699, "grad_norm": 0.8723529375145571, "learning_rate": 3.581603349196372e-07, "loss": 0.8441, "step": 3333 }, { "epoch": 4.45752508361204, "grad_norm": 0.8737954357639945, "learning_rate": 3.564279362036488e-07, "loss": 0.8399, "step": 3334 }, { "epoch": 4.45886287625418, "grad_norm": 0.8593399741094002, "learning_rate": 3.5469958253566807e-07, "loss": 0.9584, "step": 3335 }, { "epoch": 4.460200668896321, "grad_norm": 1.1165729812710947, "learning_rate": 3.5297527542127675e-07, "loss": 0.8613, "step": 3336 }, { "epoch": 4.461538461538462, "grad_norm": 0.9056926757698858, "learning_rate": 3.512550163625311e-07, "loss": 1.1393, "step": 3337 }, { "epoch": 4.462876254180602, "grad_norm": 1.0016425194293404, "learning_rate": 3.495388068579586e-07, "loss": 0.7519, "step": 3338 }, { "epoch": 4.464214046822742, "grad_norm": 0.844831296493241, "learning_rate": 3.4782664840256387e-07, "loss": 0.8048, "step": 3339 }, { "epoch": 4.465551839464883, "grad_norm": 1.0235713958557358, "learning_rate": 3.461185424878194e-07, "loss": 0.9116, "step": 3340 }, { "epoch": 4.466889632107024, "grad_norm": 0.8431519539283305, "learning_rate": 3.4441449060166776e-07, "loss": 0.8961, "step": 3341 }, { "epoch": 4.468227424749164, "grad_norm": 1.2097030231831, "learning_rate": 3.427144942285215e-07, "loss": 0.7847, "step": 3342 }, { "epoch": 4.469565217391304, "grad_norm": 0.8606694357108752, "learning_rate": 3.4101855484925727e-07, "loss": 1.0765, "step": 3343 }, { "epoch": 4.470903010033445, "grad_norm": 0.9741802787595952, "learning_rate": 3.39326673941221e-07, "loss": 1.1714, "step": 3344 }, { "epoch": 4.472240802675585, "grad_norm": 0.9977555158982664, "learning_rate": 3.3763885297822153e-07, "loss": 0.6756, "step": 3345 }, { "epoch": 4.473578595317726, "grad_norm": 0.8962144818106258, "learning_rate": 3.359550934305322e-07, "loss": 0.8524, "step": 3346 }, { "epoch": 4.474916387959866, "grad_norm": 0.9167784559409103, "learning_rate": 3.342753967648865e-07, "loss": 0.9038, "step": 3347 }, { "epoch": 4.476254180602007, "grad_norm": 0.9771914453565219, "learning_rate": 3.3259976444448005e-07, "loss": 0.9382, "step": 3348 }, { "epoch": 4.477591973244147, "grad_norm": 0.9626082751555878, "learning_rate": 3.3092819792896913e-07, "loss": 0.82, "step": 3349 }, { "epoch": 4.478929765886288, "grad_norm": 0.868225332657918, "learning_rate": 3.2926069867446673e-07, "loss": 0.9995, "step": 3350 }, { "epoch": 4.480267558528428, "grad_norm": 0.7770007022860822, "learning_rate": 3.275972681335421e-07, "loss": 1.1193, "step": 3351 }, { "epoch": 4.481605351170568, "grad_norm": 0.8495343478164067, "learning_rate": 3.259379077552216e-07, "loss": 0.921, "step": 3352 }, { "epoch": 4.482943143812709, "grad_norm": 0.9790907682695584, "learning_rate": 3.2428261898498625e-07, "loss": 0.828, "step": 3353 }, { "epoch": 4.48428093645485, "grad_norm": 0.8108325501769046, "learning_rate": 3.226314032647687e-07, "loss": 0.7342, "step": 3354 }, { "epoch": 4.4856187290969896, "grad_norm": 1.2052309949105424, "learning_rate": 3.209842620329545e-07, "loss": 0.8963, "step": 3355 }, { "epoch": 4.48695652173913, "grad_norm": 1.2859254440410668, "learning_rate": 3.1934119672438093e-07, "loss": 0.8384, "step": 3356 }, { "epoch": 4.488294314381271, "grad_norm": 1.0083432595850839, "learning_rate": 3.1770220877033243e-07, "loss": 0.7995, "step": 3357 }, { "epoch": 4.489632107023412, "grad_norm": 0.9086625957852971, "learning_rate": 3.160672995985431e-07, "loss": 0.9145, "step": 3358 }, { "epoch": 4.490969899665552, "grad_norm": 0.8651742460002565, "learning_rate": 3.1443647063319425e-07, "loss": 1.0403, "step": 3359 }, { "epoch": 4.492307692307692, "grad_norm": 0.9543861432393508, "learning_rate": 3.128097232949123e-07, "loss": 1.0136, "step": 3360 }, { "epoch": 4.493645484949833, "grad_norm": 0.8145244849125057, "learning_rate": 3.111870590007682e-07, "loss": 1.1008, "step": 3361 }, { "epoch": 4.494983277591973, "grad_norm": 0.7836142428802556, "learning_rate": 3.0956847916427556e-07, "loss": 0.9224, "step": 3362 }, { "epoch": 4.496321070234114, "grad_norm": 0.9090020544755741, "learning_rate": 3.0795398519539113e-07, "loss": 0.8807, "step": 3363 }, { "epoch": 4.497658862876254, "grad_norm": 0.8236789802407011, "learning_rate": 3.0634357850051144e-07, "loss": 1.0436, "step": 3364 }, { "epoch": 4.498996655518395, "grad_norm": 0.986811912511882, "learning_rate": 3.0473726048247386e-07, "loss": 1.0973, "step": 3365 }, { "epoch": 4.500334448160535, "grad_norm": 0.9021165112274132, "learning_rate": 3.031350325405519e-07, "loss": 1.0356, "step": 3366 }, { "epoch": 4.501672240802676, "grad_norm": 0.7823949005375891, "learning_rate": 3.015368960704584e-07, "loss": 1.2074, "step": 3367 }, { "epoch": 4.503010033444816, "grad_norm": 0.8151257770546859, "learning_rate": 2.9994285246433996e-07, "loss": 0.8429, "step": 3368 }, { "epoch": 4.504347826086956, "grad_norm": 0.9943892620064357, "learning_rate": 2.9835290311078123e-07, "loss": 0.7633, "step": 3369 }, { "epoch": 4.505685618729097, "grad_norm": 1.029872138632414, "learning_rate": 2.967670493947966e-07, "loss": 1.0595, "step": 3370 }, { "epoch": 4.507023411371238, "grad_norm": 1.2690240861657687, "learning_rate": 2.9518529269783604e-07, "loss": 0.9556, "step": 3371 }, { "epoch": 4.508361204013378, "grad_norm": 0.8983066938187856, "learning_rate": 2.936076343977762e-07, "loss": 1.0593, "step": 3372 }, { "epoch": 4.509698996655518, "grad_norm": 0.8210137341727911, "learning_rate": 2.9203407586892776e-07, "loss": 0.897, "step": 3373 }, { "epoch": 4.511036789297659, "grad_norm": 0.9047597338337872, "learning_rate": 2.9046461848202865e-07, "loss": 0.9204, "step": 3374 }, { "epoch": 4.512374581939799, "grad_norm": 0.7939199784402367, "learning_rate": 2.888992636042437e-07, "loss": 1.0671, "step": 3375 }, { "epoch": 4.51371237458194, "grad_norm": 1.0926170797953798, "learning_rate": 2.873380125991643e-07, "loss": 0.7891, "step": 3376 }, { "epoch": 4.51505016722408, "grad_norm": 0.9109996993183433, "learning_rate": 2.857808668268075e-07, "loss": 0.7829, "step": 3377 }, { "epoch": 4.516387959866221, "grad_norm": 1.2752968820242863, "learning_rate": 2.842278276436128e-07, "loss": 1.0446, "step": 3378 }, { "epoch": 4.517725752508361, "grad_norm": 0.918935620441367, "learning_rate": 2.8267889640244516e-07, "loss": 0.9501, "step": 3379 }, { "epoch": 4.519063545150502, "grad_norm": 1.068778677984516, "learning_rate": 2.811340744525887e-07, "loss": 0.8304, "step": 3380 }, { "epoch": 4.520401337792642, "grad_norm": 1.009319990645498, "learning_rate": 2.7959336313974847e-07, "loss": 1.0823, "step": 3381 }, { "epoch": 4.521739130434782, "grad_norm": 1.1824712045231998, "learning_rate": 2.7805676380604883e-07, "loss": 0.9777, "step": 3382 }, { "epoch": 4.523076923076923, "grad_norm": 0.8450417610363067, "learning_rate": 2.7652427779003234e-07, "loss": 1.2397, "step": 3383 }, { "epoch": 4.524414715719064, "grad_norm": 1.33125128782691, "learning_rate": 2.7499590642665773e-07, "loss": 0.8342, "step": 3384 }, { "epoch": 4.5257525083612045, "grad_norm": 0.755463994570284, "learning_rate": 2.734716510473007e-07, "loss": 1.1781, "step": 3385 }, { "epoch": 4.527090301003344, "grad_norm": 0.9480275593028992, "learning_rate": 2.7195151297975065e-07, "loss": 1.2138, "step": 3386 }, { "epoch": 4.528428093645485, "grad_norm": 0.7504287291529974, "learning_rate": 2.704354935482095e-07, "loss": 0.7706, "step": 3387 }, { "epoch": 4.529765886287626, "grad_norm": 1.1152246626071045, "learning_rate": 2.689235940732926e-07, "loss": 0.8348, "step": 3388 }, { "epoch": 4.531103678929766, "grad_norm": 1.1454769609254412, "learning_rate": 2.6741581587202747e-07, "loss": 0.9092, "step": 3389 }, { "epoch": 4.532441471571906, "grad_norm": 0.9922003682380308, "learning_rate": 2.6591216025784904e-07, "loss": 0.9889, "step": 3390 }, { "epoch": 4.533779264214047, "grad_norm": 1.2793334004588013, "learning_rate": 2.644126285406029e-07, "loss": 0.8984, "step": 3391 }, { "epoch": 4.535117056856187, "grad_norm": 0.9349535027840511, "learning_rate": 2.629172220265408e-07, "loss": 0.9026, "step": 3392 }, { "epoch": 4.536454849498328, "grad_norm": 0.8415636985216985, "learning_rate": 2.6142594201832183e-07, "loss": 1.1222, "step": 3393 }, { "epoch": 4.537792642140468, "grad_norm": 0.96951441512124, "learning_rate": 2.5993878981501133e-07, "loss": 0.7206, "step": 3394 }, { "epoch": 4.539130434782608, "grad_norm": 0.9485759000655818, "learning_rate": 2.584557667120768e-07, "loss": 1.0282, "step": 3395 }, { "epoch": 4.540468227424749, "grad_norm": 0.7633055179535401, "learning_rate": 2.5697687400139115e-07, "loss": 1.0351, "step": 3396 }, { "epoch": 4.54180602006689, "grad_norm": 0.7744568991969988, "learning_rate": 2.5550211297122705e-07, "loss": 0.8413, "step": 3397 }, { "epoch": 4.5431438127090304, "grad_norm": 1.0162979143837443, "learning_rate": 2.540314849062592e-07, "loss": 0.9634, "step": 3398 }, { "epoch": 4.54448160535117, "grad_norm": 0.9386789530449096, "learning_rate": 2.525649910875627e-07, "loss": 0.7397, "step": 3399 }, { "epoch": 4.545819397993311, "grad_norm": 0.8736392076860505, "learning_rate": 2.511026327926114e-07, "loss": 1.1635, "step": 3400 }, { "epoch": 4.547157190635452, "grad_norm": 0.8288214026719115, "learning_rate": 2.4964441129527337e-07, "loss": 1.2165, "step": 3401 }, { "epoch": 4.548494983277592, "grad_norm": 0.8729835718352784, "learning_rate": 2.481903278658171e-07, "loss": 1.0239, "step": 3402 }, { "epoch": 4.549832775919732, "grad_norm": 1.176229552435499, "learning_rate": 2.4674038377090423e-07, "loss": 0.5456, "step": 3403 }, { "epoch": 4.551170568561873, "grad_norm": 0.8194739350743141, "learning_rate": 2.452945802735918e-07, "loss": 0.9159, "step": 3404 }, { "epoch": 4.552508361204014, "grad_norm": 0.822990311330519, "learning_rate": 2.438529186333288e-07, "loss": 0.7428, "step": 3405 }, { "epoch": 4.553846153846154, "grad_norm": 0.8244275124250793, "learning_rate": 2.42415400105957e-07, "loss": 0.7869, "step": 3406 }, { "epoch": 4.555183946488294, "grad_norm": 0.7186492208260264, "learning_rate": 2.4098202594370844e-07, "loss": 0.8713, "step": 3407 }, { "epoch": 4.556521739130435, "grad_norm": 0.9903943324217569, "learning_rate": 2.3955279739520496e-07, "loss": 0.8449, "step": 3408 }, { "epoch": 4.557859531772575, "grad_norm": 1.2442405311262235, "learning_rate": 2.3812771570545846e-07, "loss": 0.8858, "step": 3409 }, { "epoch": 4.559197324414716, "grad_norm": 0.885118531720378, "learning_rate": 2.3670678211586805e-07, "loss": 1.3003, "step": 3410 }, { "epoch": 4.560535117056856, "grad_norm": 0.8275344781357615, "learning_rate": 2.3528999786421758e-07, "loss": 0.9603, "step": 3411 }, { "epoch": 4.561872909698996, "grad_norm": 0.9775949034203717, "learning_rate": 2.338773641846781e-07, "loss": 1.0803, "step": 3412 }, { "epoch": 4.563210702341137, "grad_norm": 1.1311245538211447, "learning_rate": 2.3246888230780474e-07, "loss": 1.0841, "step": 3413 }, { "epoch": 4.564548494983278, "grad_norm": 0.8919642679034281, "learning_rate": 2.3106455346053603e-07, "loss": 0.9662, "step": 3414 }, { "epoch": 4.565886287625418, "grad_norm": 1.028812902124353, "learning_rate": 2.2966437886619286e-07, "loss": 0.9011, "step": 3415 }, { "epoch": 4.567224080267558, "grad_norm": 1.0139941874084075, "learning_rate": 2.2826835974447626e-07, "loss": 0.9148, "step": 3416 }, { "epoch": 4.568561872909699, "grad_norm": 1.2457625856027297, "learning_rate": 2.2687649731146844e-07, "loss": 0.8878, "step": 3417 }, { "epoch": 4.56989966555184, "grad_norm": 0.8341484772722453, "learning_rate": 2.2548879277963065e-07, "loss": 1.1036, "step": 3418 }, { "epoch": 4.57123745819398, "grad_norm": 0.960786497905212, "learning_rate": 2.2410524735780205e-07, "loss": 1.07, "step": 3419 }, { "epoch": 4.57257525083612, "grad_norm": 0.9039392106982292, "learning_rate": 2.227258622511991e-07, "loss": 0.894, "step": 3420 }, { "epoch": 4.573913043478261, "grad_norm": 0.8680079712986077, "learning_rate": 2.2135063866141337e-07, "loss": 0.9793, "step": 3421 }, { "epoch": 4.575250836120401, "grad_norm": 1.0777084172896587, "learning_rate": 2.1997957778641166e-07, "loss": 0.9602, "step": 3422 }, { "epoch": 4.576588628762542, "grad_norm": 0.9856863738018464, "learning_rate": 2.1861268082053466e-07, "loss": 0.9474, "step": 3423 }, { "epoch": 4.577926421404682, "grad_norm": 0.892974148651928, "learning_rate": 2.1724994895449603e-07, "loss": 1.0154, "step": 3424 }, { "epoch": 4.579264214046823, "grad_norm": 0.8479979764026874, "learning_rate": 2.1589138337538062e-07, "loss": 1.0718, "step": 3425 }, { "epoch": 4.580602006688963, "grad_norm": 1.058560541713886, "learning_rate": 2.1453698526664513e-07, "loss": 0.9531, "step": 3426 }, { "epoch": 4.581939799331104, "grad_norm": 0.9414270970726396, "learning_rate": 2.1318675580811409e-07, "loss": 1.2466, "step": 3427 }, { "epoch": 4.5832775919732445, "grad_norm": 0.8085813722659113, "learning_rate": 2.1184069617598225e-07, "loss": 0.9407, "step": 3428 }, { "epoch": 4.584615384615384, "grad_norm": 0.9094800726805624, "learning_rate": 2.104988075428127e-07, "loss": 0.9931, "step": 3429 }, { "epoch": 4.585953177257525, "grad_norm": 0.916091602203396, "learning_rate": 2.0916109107753267e-07, "loss": 0.9076, "step": 3430 }, { "epoch": 4.587290969899666, "grad_norm": 0.8023699105729402, "learning_rate": 2.0782754794543668e-07, "loss": 1.0037, "step": 3431 }, { "epoch": 4.588628762541806, "grad_norm": 0.813699519431517, "learning_rate": 2.0649817930818326e-07, "loss": 0.8313, "step": 3432 }, { "epoch": 4.589966555183946, "grad_norm": 1.0023181870270244, "learning_rate": 2.0517298632379445e-07, "loss": 0.853, "step": 3433 }, { "epoch": 4.591304347826087, "grad_norm": 0.8000780882942163, "learning_rate": 2.038519701466557e-07, "loss": 0.7854, "step": 3434 }, { "epoch": 4.592642140468227, "grad_norm": 0.8500497213934377, "learning_rate": 2.0253513192751374e-07, "loss": 0.927, "step": 3435 }, { "epoch": 4.593979933110368, "grad_norm": 1.0473786242085332, "learning_rate": 2.012224728134743e-07, "loss": 0.9041, "step": 3436 }, { "epoch": 4.595317725752508, "grad_norm": 0.7359870013252958, "learning_rate": 1.999139939480049e-07, "loss": 0.9483, "step": 3437 }, { "epoch": 4.596655518394649, "grad_norm": 0.732331855492198, "learning_rate": 1.9860969647092998e-07, "loss": 0.7356, "step": 3438 }, { "epoch": 4.597993311036789, "grad_norm": 0.9918538430878612, "learning_rate": 1.9730958151843282e-07, "loss": 1.1462, "step": 3439 }, { "epoch": 4.59933110367893, "grad_norm": 0.7738572059149563, "learning_rate": 1.9601365022305196e-07, "loss": 0.8487, "step": 3440 }, { "epoch": 4.6006688963210705, "grad_norm": 1.1496869897363813, "learning_rate": 1.947219037136827e-07, "loss": 0.9046, "step": 3441 }, { "epoch": 4.602006688963211, "grad_norm": 1.3316165935009354, "learning_rate": 1.934343431155744e-07, "loss": 0.6587, "step": 3442 }, { "epoch": 4.603344481605351, "grad_norm": 0.8386969738627418, "learning_rate": 1.9215096955032986e-07, "loss": 1.251, "step": 3443 }, { "epoch": 4.604682274247492, "grad_norm": 0.8790227188202673, "learning_rate": 1.908717841359048e-07, "loss": 1.0033, "step": 3444 }, { "epoch": 4.6060200668896325, "grad_norm": 0.9300572800897733, "learning_rate": 1.8959678798660674e-07, "loss": 1.1125, "step": 3445 }, { "epoch": 4.607357859531772, "grad_norm": 0.8714576432223715, "learning_rate": 1.883259822130934e-07, "loss": 0.8029, "step": 3446 }, { "epoch": 4.608695652173913, "grad_norm": 0.9972855497995139, "learning_rate": 1.8705936792237255e-07, "loss": 0.7296, "step": 3447 }, { "epoch": 4.610033444816054, "grad_norm": 1.0118438262952532, "learning_rate": 1.8579694621780054e-07, "loss": 0.8967, "step": 3448 }, { "epoch": 4.611371237458194, "grad_norm": 1.063416878699577, "learning_rate": 1.845387181990821e-07, "loss": 0.9749, "step": 3449 }, { "epoch": 4.612709030100334, "grad_norm": 0.7926819231375314, "learning_rate": 1.8328468496226882e-07, "loss": 0.9976, "step": 3450 }, { "epoch": 4.614046822742475, "grad_norm": 1.3547956134122818, "learning_rate": 1.8203484759975743e-07, "loss": 0.887, "step": 3451 }, { "epoch": 4.615384615384615, "grad_norm": 1.194969770051571, "learning_rate": 1.807892072002898e-07, "loss": 1.0186, "step": 3452 }, { "epoch": 4.616722408026756, "grad_norm": 1.0436626546918009, "learning_rate": 1.7954776484895188e-07, "loss": 1.0946, "step": 3453 }, { "epoch": 4.618060200668896, "grad_norm": 0.9137471468769923, "learning_rate": 1.78310521627173e-07, "loss": 0.6565, "step": 3454 }, { "epoch": 4.619397993311037, "grad_norm": 1.0764811882504208, "learning_rate": 1.770774786127244e-07, "loss": 0.7672, "step": 3455 }, { "epoch": 4.620735785953177, "grad_norm": 0.9393079605693597, "learning_rate": 1.7584863687971852e-07, "loss": 0.9866, "step": 3456 }, { "epoch": 4.622073578595318, "grad_norm": 1.0792576987210787, "learning_rate": 1.7462399749860748e-07, "loss": 0.7724, "step": 3457 }, { "epoch": 4.6234113712374585, "grad_norm": 1.2170154165370628, "learning_rate": 1.7340356153618343e-07, "loss": 1.0697, "step": 3458 }, { "epoch": 4.624749163879598, "grad_norm": 0.8155243067731742, "learning_rate": 1.7218733005557707e-07, "loss": 0.6469, "step": 3459 }, { "epoch": 4.626086956521739, "grad_norm": 0.8329455604674131, "learning_rate": 1.7097530411625596e-07, "loss": 1.062, "step": 3460 }, { "epoch": 4.62742474916388, "grad_norm": 0.85286952747454, "learning_rate": 1.6976748477402384e-07, "loss": 0.8772, "step": 3461 }, { "epoch": 4.6287625418060205, "grad_norm": 1.0561042416335729, "learning_rate": 1.6856387308102073e-07, "loss": 0.8711, "step": 3462 }, { "epoch": 4.63010033444816, "grad_norm": 0.9376069795261381, "learning_rate": 1.6736447008572132e-07, "loss": 0.8314, "step": 3463 }, { "epoch": 4.631438127090301, "grad_norm": 0.8707062008634633, "learning_rate": 1.6616927683293427e-07, "loss": 1.1072, "step": 3464 }, { "epoch": 4.632775919732442, "grad_norm": 1.0699230760478347, "learning_rate": 1.6497829436380009e-07, "loss": 1.0059, "step": 3465 }, { "epoch": 4.634113712374582, "grad_norm": 1.0128511593254887, "learning_rate": 1.6379152371579277e-07, "loss": 0.6898, "step": 3466 }, { "epoch": 4.635451505016722, "grad_norm": 0.8528203964323592, "learning_rate": 1.6260896592271534e-07, "loss": 0.9796, "step": 3467 }, { "epoch": 4.636789297658863, "grad_norm": 0.9005975357915842, "learning_rate": 1.614306220147027e-07, "loss": 1.1045, "step": 3468 }, { "epoch": 4.638127090301003, "grad_norm": 0.9381267784620926, "learning_rate": 1.6025649301821877e-07, "loss": 0.9681, "step": 3469 }, { "epoch": 4.639464882943144, "grad_norm": 0.9255815356409202, "learning_rate": 1.5908657995605536e-07, "loss": 0.8658, "step": 3470 }, { "epoch": 4.6408026755852845, "grad_norm": 0.7424164143634934, "learning_rate": 1.5792088384733174e-07, "loss": 0.8946, "step": 3471 }, { "epoch": 4.642140468227424, "grad_norm": 0.7092175894640276, "learning_rate": 1.5675940570749393e-07, "loss": 0.9614, "step": 3472 }, { "epoch": 4.643478260869565, "grad_norm": 0.8180309998270601, "learning_rate": 1.5560214654831375e-07, "loss": 1.0602, "step": 3473 }, { "epoch": 4.644816053511706, "grad_norm": 0.8905274983641664, "learning_rate": 1.5444910737788755e-07, "loss": 1.1586, "step": 3474 }, { "epoch": 4.6461538461538465, "grad_norm": 0.9705753067606857, "learning_rate": 1.5330028920063634e-07, "loss": 0.9347, "step": 3475 }, { "epoch": 4.647491638795986, "grad_norm": 0.882761105856875, "learning_rate": 1.5215569301730293e-07, "loss": 0.7916, "step": 3476 }, { "epoch": 4.648829431438127, "grad_norm": 0.7921752152782053, "learning_rate": 1.510153198249531e-07, "loss": 1.0675, "step": 3477 }, { "epoch": 4.650167224080268, "grad_norm": 1.4253612607463002, "learning_rate": 1.4987917061697387e-07, "loss": 0.7924, "step": 3478 }, { "epoch": 4.651505016722408, "grad_norm": 0.9396030300473629, "learning_rate": 1.4874724638307303e-07, "loss": 0.8392, "step": 3479 }, { "epoch": 4.652842809364548, "grad_norm": 0.906516015529776, "learning_rate": 1.4761954810927791e-07, "loss": 0.8453, "step": 3480 }, { "epoch": 4.654180602006689, "grad_norm": 0.883690777408552, "learning_rate": 1.4649607677793388e-07, "loss": 1.0435, "step": 3481 }, { "epoch": 4.65551839464883, "grad_norm": 0.9369833474092168, "learning_rate": 1.4537683336770526e-07, "loss": 0.8118, "step": 3482 }, { "epoch": 4.65685618729097, "grad_norm": 0.9124124791655851, "learning_rate": 1.4426181885357215e-07, "loss": 0.68, "step": 3483 }, { "epoch": 4.6581939799331105, "grad_norm": 1.034303996952988, "learning_rate": 1.4315103420683152e-07, "loss": 0.7143, "step": 3484 }, { "epoch": 4.659531772575251, "grad_norm": 0.8521338955517384, "learning_rate": 1.42044480395096e-07, "loss": 0.8405, "step": 3485 }, { "epoch": 4.660869565217391, "grad_norm": 1.1361535951549429, "learning_rate": 1.4094215838229176e-07, "loss": 0.9735, "step": 3486 }, { "epoch": 4.662207357859532, "grad_norm": 0.7879898961262175, "learning_rate": 1.3984406912865954e-07, "loss": 0.8295, "step": 3487 }, { "epoch": 4.6635451505016725, "grad_norm": 1.240739509240088, "learning_rate": 1.3875021359075257e-07, "loss": 0.818, "step": 3488 }, { "epoch": 4.664882943143812, "grad_norm": 0.9167699725854205, "learning_rate": 1.376605927214364e-07, "loss": 0.928, "step": 3489 }, { "epoch": 4.666220735785953, "grad_norm": 0.8874265896364817, "learning_rate": 1.3657520746988674e-07, "loss": 0.8537, "step": 3490 }, { "epoch": 4.667558528428094, "grad_norm": 0.9395693303738865, "learning_rate": 1.354940587815906e-07, "loss": 0.995, "step": 3491 }, { "epoch": 4.668896321070234, "grad_norm": 1.1109943938618938, "learning_rate": 1.3441714759834358e-07, "loss": 0.898, "step": 3492 }, { "epoch": 4.670234113712374, "grad_norm": 0.9841937284084397, "learning_rate": 1.333444748582513e-07, "loss": 0.9239, "step": 3493 }, { "epoch": 4.671571906354515, "grad_norm": 1.05667020655409, "learning_rate": 1.3227604149572638e-07, "loss": 0.8398, "step": 3494 }, { "epoch": 4.672909698996656, "grad_norm": 1.1063447133698028, "learning_rate": 1.312118484414876e-07, "loss": 0.9151, "step": 3495 }, { "epoch": 4.674247491638796, "grad_norm": 0.856960136251213, "learning_rate": 1.3015189662256234e-07, "loss": 0.7284, "step": 3496 }, { "epoch": 4.6755852842809364, "grad_norm": 0.8611535016809787, "learning_rate": 1.2909618696228088e-07, "loss": 0.7693, "step": 3497 }, { "epoch": 4.676923076923077, "grad_norm": 0.9336538117477514, "learning_rate": 1.2804472038027983e-07, "loss": 0.7314, "step": 3498 }, { "epoch": 4.678260869565217, "grad_norm": 0.9348074496571795, "learning_rate": 1.2699749779249926e-07, "loss": 0.8685, "step": 3499 }, { "epoch": 4.679598662207358, "grad_norm": 1.0942363671399087, "learning_rate": 1.259545201111817e-07, "loss": 0.9607, "step": 3500 }, { "epoch": 4.6809364548494985, "grad_norm": 0.9987968161802744, "learning_rate": 1.2491578824487204e-07, "loss": 0.9277, "step": 3501 }, { "epoch": 4.682274247491639, "grad_norm": 0.8044871069588523, "learning_rate": 1.2388130309841762e-07, "loss": 1.0693, "step": 3502 }, { "epoch": 4.683612040133779, "grad_norm": 0.7968739693589157, "learning_rate": 1.2285106557296479e-07, "loss": 0.9864, "step": 3503 }, { "epoch": 4.68494983277592, "grad_norm": 0.8200679096372285, "learning_rate": 1.2182507656596177e-07, "loss": 0.9366, "step": 3504 }, { "epoch": 4.6862876254180605, "grad_norm": 1.013005090665651, "learning_rate": 1.2080333697115366e-07, "loss": 0.792, "step": 3505 }, { "epoch": 4.6876254180602, "grad_norm": 0.781560104940914, "learning_rate": 1.1978584767858513e-07, "loss": 0.9814, "step": 3506 }, { "epoch": 4.688963210702341, "grad_norm": 0.9521785794726048, "learning_rate": 1.1877260957459835e-07, "loss": 0.6636, "step": 3507 }, { "epoch": 4.690301003344482, "grad_norm": 1.0538240898191993, "learning_rate": 1.1776362354183224e-07, "loss": 0.8073, "step": 3508 }, { "epoch": 4.691638795986622, "grad_norm": 0.7906971392340251, "learning_rate": 1.1675889045922151e-07, "loss": 0.9244, "step": 3509 }, { "epoch": 4.692976588628762, "grad_norm": 1.2573967718073353, "learning_rate": 1.157584112019966e-07, "loss": 0.8913, "step": 3510 }, { "epoch": 4.694314381270903, "grad_norm": 0.8938693477263647, "learning_rate": 1.1476218664168093e-07, "loss": 1.0888, "step": 3511 }, { "epoch": 4.695652173913043, "grad_norm": 0.8893009410067202, "learning_rate": 1.1377021764609364e-07, "loss": 0.8227, "step": 3512 }, { "epoch": 4.696989966555184, "grad_norm": 0.8760175767320074, "learning_rate": 1.1278250507934518e-07, "loss": 0.8944, "step": 3513 }, { "epoch": 4.6983277591973245, "grad_norm": 0.9586059545132696, "learning_rate": 1.1179904980183897e-07, "loss": 0.9528, "step": 3514 }, { "epoch": 4.699665551839465, "grad_norm": 0.9935546336396704, "learning_rate": 1.1081985267027029e-07, "loss": 1.0307, "step": 3515 }, { "epoch": 4.701003344481605, "grad_norm": 0.7918841133686778, "learning_rate": 1.0984491453762402e-07, "loss": 0.8117, "step": 3516 }, { "epoch": 4.702341137123746, "grad_norm": 0.89185795196407, "learning_rate": 1.0887423625317584e-07, "loss": 0.9573, "step": 3517 }, { "epoch": 4.7036789297658865, "grad_norm": 1.1074040484098786, "learning_rate": 1.079078186624899e-07, "loss": 0.8649, "step": 3518 }, { "epoch": 4.705016722408026, "grad_norm": 1.3524395739718107, "learning_rate": 1.0694566260742001e-07, "loss": 1.2378, "step": 3519 }, { "epoch": 4.706354515050167, "grad_norm": 1.0240778333712888, "learning_rate": 1.0598776892610685e-07, "loss": 0.9552, "step": 3520 }, { "epoch": 4.707692307692308, "grad_norm": 0.8709071604273074, "learning_rate": 1.0503413845297739e-07, "loss": 1.0103, "step": 3521 }, { "epoch": 4.709030100334449, "grad_norm": 0.9364109904259138, "learning_rate": 1.0408477201874712e-07, "loss": 1.0887, "step": 3522 }, { "epoch": 4.710367892976588, "grad_norm": 0.9444390697926941, "learning_rate": 1.0313967045041507e-07, "loss": 0.9663, "step": 3523 }, { "epoch": 4.711705685618729, "grad_norm": 0.7445367192907572, "learning_rate": 1.02198834571266e-07, "loss": 0.9642, "step": 3524 }, { "epoch": 4.71304347826087, "grad_norm": 0.793889573272855, "learning_rate": 1.0126226520086823e-07, "loss": 0.9203, "step": 3525 }, { "epoch": 4.71438127090301, "grad_norm": 1.0945481456551998, "learning_rate": 1.0032996315507415e-07, "loss": 1.1239, "step": 3526 }, { "epoch": 4.7157190635451505, "grad_norm": 1.06696202328999, "learning_rate": 9.940192924601855e-08, "loss": 0.9876, "step": 3527 }, { "epoch": 4.717056856187291, "grad_norm": 0.9355102359214357, "learning_rate": 9.847816428211809e-08, "loss": 1.0959, "step": 3528 }, { "epoch": 4.718394648829431, "grad_norm": 1.050545506330983, "learning_rate": 9.755866906807188e-08, "loss": 1.136, "step": 3529 }, { "epoch": 4.719732441471572, "grad_norm": 1.0444908250633647, "learning_rate": 9.664344440485696e-08, "loss": 0.8649, "step": 3530 }, { "epoch": 4.7210702341137125, "grad_norm": 0.9180756237353728, "learning_rate": 9.573249108973281e-08, "loss": 1.0754, "step": 3531 }, { "epoch": 4.722408026755852, "grad_norm": 0.8832493444846164, "learning_rate": 9.482580991623747e-08, "loss": 0.8097, "step": 3532 }, { "epoch": 4.723745819397993, "grad_norm": 1.051670615288861, "learning_rate": 9.39234016741869e-08, "loss": 0.9238, "step": 3533 }, { "epoch": 4.725083612040134, "grad_norm": 0.9344708929469139, "learning_rate": 9.302526714967508e-08, "loss": 1.063, "step": 3534 }, { "epoch": 4.726421404682275, "grad_norm": 1.0967543051849848, "learning_rate": 9.21314071250734e-08, "loss": 0.5486, "step": 3535 }, { "epoch": 4.727759197324414, "grad_norm": 0.8387410903860719, "learning_rate": 9.124182237902957e-08, "loss": 0.8579, "step": 3536 }, { "epoch": 4.729096989966555, "grad_norm": 0.8071107168725659, "learning_rate": 9.035651368646647e-08, "loss": 1.0974, "step": 3537 }, { "epoch": 4.730434782608696, "grad_norm": 0.8548180870208943, "learning_rate": 8.94754818185839e-08, "loss": 1.2543, "step": 3538 }, { "epoch": 4.731772575250837, "grad_norm": 0.9968361102311571, "learning_rate": 8.859872754285403e-08, "loss": 0.9308, "step": 3539 }, { "epoch": 4.7331103678929765, "grad_norm": 0.8057672797620808, "learning_rate": 8.772625162302373e-08, "loss": 0.8316, "step": 3540 }, { "epoch": 4.734448160535117, "grad_norm": 0.9414750040902594, "learning_rate": 8.68580548191128e-08, "loss": 1.0771, "step": 3541 }, { "epoch": 4.735785953177258, "grad_norm": 0.9949583805295964, "learning_rate": 8.599413788741407e-08, "loss": 0.9103, "step": 3542 }, { "epoch": 4.737123745819398, "grad_norm": 0.8928261418897531, "learning_rate": 8.513450158049109e-08, "loss": 1.1696, "step": 3543 }, { "epoch": 4.7384615384615385, "grad_norm": 1.0329611737393305, "learning_rate": 8.427914664717929e-08, "loss": 0.9537, "step": 3544 }, { "epoch": 4.739799331103679, "grad_norm": 0.9467196777819108, "learning_rate": 8.342807383258378e-08, "loss": 0.9309, "step": 3545 }, { "epoch": 4.741137123745819, "grad_norm": 0.7888952639505493, "learning_rate": 8.258128387808095e-08, "loss": 0.726, "step": 3546 }, { "epoch": 4.74247491638796, "grad_norm": 0.8568013717915883, "learning_rate": 8.173877752131465e-08, "loss": 0.917, "step": 3547 }, { "epoch": 4.7438127090301005, "grad_norm": 1.0764363245159285, "learning_rate": 8.090055549619835e-08, "loss": 0.999, "step": 3548 }, { "epoch": 4.74515050167224, "grad_norm": 0.8684579561757797, "learning_rate": 8.006661853291298e-08, "loss": 0.9223, "step": 3549 }, { "epoch": 4.746488294314381, "grad_norm": 0.944020573073878, "learning_rate": 7.923696735790687e-08, "loss": 1.052, "step": 3550 }, { "epoch": 4.747826086956522, "grad_norm": 0.7805473724216769, "learning_rate": 7.84116026938947e-08, "loss": 0.9458, "step": 3551 }, { "epoch": 4.749163879598662, "grad_norm": 0.8789571397999069, "learning_rate": 7.7590525259858e-08, "loss": 1.0353, "step": 3552 }, { "epoch": 4.750501672240802, "grad_norm": 0.9807426158373517, "learning_rate": 7.677373577104296e-08, "loss": 0.8201, "step": 3553 }, { "epoch": 4.751839464882943, "grad_norm": 0.8752953985263725, "learning_rate": 7.59612349389599e-08, "loss": 1.0533, "step": 3554 }, { "epoch": 4.753177257525084, "grad_norm": 0.9154444853048325, "learning_rate": 7.515302347138486e-08, "loss": 1.2328, "step": 3555 }, { "epoch": 4.754515050167224, "grad_norm": 0.8891253073645375, "learning_rate": 7.434910207235579e-08, "loss": 1.0668, "step": 3556 }, { "epoch": 4.7558528428093645, "grad_norm": 1.0436039669892019, "learning_rate": 7.354947144217417e-08, "loss": 0.8405, "step": 3557 }, { "epoch": 4.757190635451505, "grad_norm": 1.066437836539782, "learning_rate": 7.275413227740446e-08, "loss": 0.9958, "step": 3558 }, { "epoch": 4.758528428093646, "grad_norm": 0.8919772176315582, "learning_rate": 7.196308527087192e-08, "loss": 0.9404, "step": 3559 }, { "epoch": 4.759866220735786, "grad_norm": 1.1669139366804988, "learning_rate": 7.117633111166311e-08, "loss": 0.7053, "step": 3560 }, { "epoch": 4.7612040133779265, "grad_norm": 0.9186520764934619, "learning_rate": 7.03938704851248e-08, "loss": 0.9764, "step": 3561 }, { "epoch": 4.762541806020067, "grad_norm": 1.0479869191174038, "learning_rate": 6.9615704072864e-08, "loss": 0.7335, "step": 3562 }, { "epoch": 4.763879598662207, "grad_norm": 0.7838492313814837, "learning_rate": 6.884183255274734e-08, "loss": 1.0323, "step": 3563 }, { "epoch": 4.765217391304348, "grad_norm": 0.9463683073989582, "learning_rate": 6.807225659889894e-08, "loss": 0.8675, "step": 3564 }, { "epoch": 4.766555183946489, "grad_norm": 0.91784123770435, "learning_rate": 6.730697688170251e-08, "loss": 1.0109, "step": 3565 }, { "epoch": 4.767892976588628, "grad_norm": 0.7959268413919922, "learning_rate": 6.654599406779816e-08, "loss": 0.9702, "step": 3566 }, { "epoch": 4.769230769230769, "grad_norm": 0.9621728087703091, "learning_rate": 6.578930882008283e-08, "loss": 1.0135, "step": 3567 }, { "epoch": 4.77056856187291, "grad_norm": 1.0673148272940787, "learning_rate": 6.503692179771148e-08, "loss": 0.8608, "step": 3568 }, { "epoch": 4.77190635451505, "grad_norm": 0.9742339333745336, "learning_rate": 6.428883365609261e-08, "loss": 0.9003, "step": 3569 }, { "epoch": 4.7732441471571905, "grad_norm": 0.907182885969716, "learning_rate": 6.354504504689219e-08, "loss": 1.1181, "step": 3570 }, { "epoch": 4.774581939799331, "grad_norm": 1.6966939646223784, "learning_rate": 6.280555661802857e-08, "loss": 0.6456, "step": 3571 }, { "epoch": 4.775919732441472, "grad_norm": 0.935786727999216, "learning_rate": 6.207036901367536e-08, "loss": 1.0958, "step": 3572 }, { "epoch": 4.777257525083612, "grad_norm": 0.8812588883039691, "learning_rate": 6.133948287426028e-08, "loss": 1.0218, "step": 3573 }, { "epoch": 4.7785953177257525, "grad_norm": 1.180266522446011, "learning_rate": 6.061289883646293e-08, "loss": 1.0187, "step": 3574 }, { "epoch": 4.779933110367893, "grad_norm": 0.8561016530602747, "learning_rate": 5.98906175332159e-08, "loss": 0.9424, "step": 3575 }, { "epoch": 4.781270903010033, "grad_norm": 1.4219579270899205, "learning_rate": 5.917263959370312e-08, "loss": 0.7247, "step": 3576 }, { "epoch": 4.782608695652174, "grad_norm": 1.053331358133385, "learning_rate": 5.84589656433604e-08, "loss": 0.847, "step": 3577 }, { "epoch": 4.783946488294315, "grad_norm": 1.2398214841891633, "learning_rate": 5.7749596303874335e-08, "loss": 1.0295, "step": 3578 }, { "epoch": 4.785284280936455, "grad_norm": 0.7766688325313668, "learning_rate": 5.704453219318118e-08, "loss": 1.1494, "step": 3579 }, { "epoch": 4.786622073578595, "grad_norm": 1.2477252703402582, "learning_rate": 5.634377392546741e-08, "loss": 1.0521, "step": 3580 }, { "epoch": 4.787959866220736, "grad_norm": 0.9417052878771464, "learning_rate": 5.564732211116808e-08, "loss": 0.9041, "step": 3581 }, { "epoch": 4.789297658862877, "grad_norm": 1.0231528188987, "learning_rate": 5.495517735696732e-08, "loss": 0.9294, "step": 3582 }, { "epoch": 4.7906354515050165, "grad_norm": 0.9013561104916261, "learning_rate": 5.426734026579783e-08, "loss": 1.0121, "step": 3583 }, { "epoch": 4.791973244147157, "grad_norm": 0.9531480720521034, "learning_rate": 5.358381143683866e-08, "loss": 1.1347, "step": 3584 }, { "epoch": 4.793311036789298, "grad_norm": 0.8775476612062648, "learning_rate": 5.2904591465516855e-08, "loss": 1.0287, "step": 3585 }, { "epoch": 4.794648829431438, "grad_norm": 1.0470709155650175, "learning_rate": 5.2229680943505225e-08, "loss": 1.0187, "step": 3586 }, { "epoch": 4.7959866220735785, "grad_norm": 1.3569422777147828, "learning_rate": 5.155908045872349e-08, "loss": 1.1084, "step": 3587 }, { "epoch": 4.797324414715719, "grad_norm": 0.7811419804972676, "learning_rate": 5.089279059533658e-08, "loss": 0.9117, "step": 3588 }, { "epoch": 4.798662207357859, "grad_norm": 0.7647692676704854, "learning_rate": 5.023081193375357e-08, "loss": 0.9462, "step": 3589 }, { "epoch": 4.8, "grad_norm": 0.8565466740700886, "learning_rate": 4.957314505062927e-08, "loss": 0.9504, "step": 3590 }, { "epoch": 4.801337792642141, "grad_norm": 1.0338214957469998, "learning_rate": 4.891979051886153e-08, "loss": 1.1831, "step": 3591 }, { "epoch": 4.802675585284281, "grad_norm": 0.8791329315237887, "learning_rate": 4.827074890759231e-08, "loss": 0.9887, "step": 3592 }, { "epoch": 4.804013377926421, "grad_norm": 0.7996129320678891, "learning_rate": 4.762602078220657e-08, "loss": 0.7683, "step": 3593 }, { "epoch": 4.805351170568562, "grad_norm": 0.7911330356298462, "learning_rate": 4.698560670433061e-08, "loss": 0.8692, "step": 3594 }, { "epoch": 4.806688963210703, "grad_norm": 1.0490094016556581, "learning_rate": 4.634950723183429e-08, "loss": 1.0683, "step": 3595 }, { "epoch": 4.8080267558528424, "grad_norm": 0.9601113089976723, "learning_rate": 4.57177229188277e-08, "loss": 1.0684, "step": 3596 }, { "epoch": 4.809364548494983, "grad_norm": 0.7994965164975164, "learning_rate": 4.509025431566283e-08, "loss": 1.0894, "step": 3597 }, { "epoch": 4.810702341137124, "grad_norm": 0.865130744239625, "learning_rate": 4.446710196893245e-08, "loss": 0.9783, "step": 3598 }, { "epoch": 4.812040133779265, "grad_norm": 0.9098944938472839, "learning_rate": 4.384826642146844e-08, "loss": 1.1447, "step": 3599 }, { "epoch": 4.8133779264214045, "grad_norm": 0.7425413775051963, "learning_rate": 4.323374821234294e-08, "loss": 0.8007, "step": 3600 }, { "epoch": 4.814715719063545, "grad_norm": 1.1187544469930526, "learning_rate": 4.262354787686718e-08, "loss": 0.9312, "step": 3601 }, { "epoch": 4.816053511705686, "grad_norm": 0.9954840386101953, "learning_rate": 4.20176659465904e-08, "loss": 0.9234, "step": 3602 }, { "epoch": 4.817391304347826, "grad_norm": 0.8676951358630718, "learning_rate": 4.141610294930043e-08, "loss": 0.8822, "step": 3603 }, { "epoch": 4.8187290969899665, "grad_norm": 1.0876699778634284, "learning_rate": 4.081885940902419e-08, "loss": 1.0462, "step": 3604 }, { "epoch": 4.820066889632107, "grad_norm": 0.9369444293621885, "learning_rate": 4.02259358460233e-08, "loss": 1.0199, "step": 3605 }, { "epoch": 4.821404682274247, "grad_norm": 0.7948413590791451, "learning_rate": 3.963733277679904e-08, "loss": 0.9882, "step": 3606 }, { "epoch": 4.822742474916388, "grad_norm": 1.0600960393954963, "learning_rate": 3.905305071408627e-08, "loss": 0.9888, "step": 3607 }, { "epoch": 4.824080267558529, "grad_norm": 1.0072246355668262, "learning_rate": 3.847309016685785e-08, "loss": 0.8878, "step": 3608 }, { "epoch": 4.825418060200668, "grad_norm": 0.9352539752479899, "learning_rate": 3.7897451640321326e-08, "loss": 1.009, "step": 3609 }, { "epoch": 4.826755852842809, "grad_norm": 0.9061499961724098, "learning_rate": 3.732613563591947e-08, "loss": 0.7794, "step": 3610 }, { "epoch": 4.82809364548495, "grad_norm": 0.9244562956890087, "learning_rate": 3.675914265132974e-08, "loss": 0.8731, "step": 3611 }, { "epoch": 4.829431438127091, "grad_norm": 0.7568553063056798, "learning_rate": 3.619647318046371e-08, "loss": 0.8424, "step": 3612 }, { "epoch": 4.8307692307692305, "grad_norm": 0.8603743199897458, "learning_rate": 3.563812771346598e-08, "loss": 1.0515, "step": 3613 }, { "epoch": 4.832107023411371, "grad_norm": 0.965900446056626, "learning_rate": 3.508410673671636e-08, "loss": 0.9168, "step": 3614 }, { "epoch": 4.833444816053512, "grad_norm": 1.230516683923745, "learning_rate": 3.4534410732825485e-08, "loss": 0.9609, "step": 3615 }, { "epoch": 4.834782608695652, "grad_norm": 0.9406234210944177, "learning_rate": 3.398904018063809e-08, "loss": 0.6856, "step": 3616 }, { "epoch": 4.8361204013377925, "grad_norm": 0.8891349334716699, "learning_rate": 3.344799555522915e-08, "loss": 0.8033, "step": 3617 }, { "epoch": 4.837458193979933, "grad_norm": 1.0414698717299962, "learning_rate": 3.291127732790722e-08, "loss": 1.0036, "step": 3618 }, { "epoch": 4.838795986622074, "grad_norm": 0.9255640985961983, "learning_rate": 3.2378885966211636e-08, "loss": 0.837, "step": 3619 }, { "epoch": 4.840133779264214, "grad_norm": 0.9669854730308083, "learning_rate": 3.185082193391143e-08, "loss": 0.6586, "step": 3620 }, { "epoch": 4.841471571906355, "grad_norm": 0.8270073314265747, "learning_rate": 3.1327085691006954e-08, "loss": 0.8435, "step": 3621 }, { "epoch": 4.842809364548495, "grad_norm": 0.9914761882401728, "learning_rate": 3.080767769372939e-08, "loss": 0.9796, "step": 3622 }, { "epoch": 4.844147157190635, "grad_norm": 1.0136887771510588, "learning_rate": 3.029259839453791e-08, "loss": 0.9792, "step": 3623 }, { "epoch": 4.845484949832776, "grad_norm": 0.8504278722658047, "learning_rate": 2.978184824212138e-08, "loss": 0.8335, "step": 3624 }, { "epoch": 4.846822742474917, "grad_norm": 0.904800904141204, "learning_rate": 2.927542768139835e-08, "loss": 0.9898, "step": 3625 }, { "epoch": 4.8481605351170565, "grad_norm": 0.930186516284019, "learning_rate": 2.877333715351538e-08, "loss": 1.0879, "step": 3626 }, { "epoch": 4.849498327759197, "grad_norm": 0.9486254628756674, "learning_rate": 2.8275577095846495e-08, "loss": 0.7933, "step": 3627 }, { "epoch": 4.850836120401338, "grad_norm": 1.1835820300152324, "learning_rate": 2.7782147941994298e-08, "loss": 0.7401, "step": 3628 }, { "epoch": 4.852173913043478, "grad_norm": 1.2050154421742225, "learning_rate": 2.7293050121788843e-08, "loss": 0.9765, "step": 3629 }, { "epoch": 4.8535117056856185, "grad_norm": 1.2216796980707199, "learning_rate": 2.6808284061285996e-08, "loss": 0.8846, "step": 3630 }, { "epoch": 4.854849498327759, "grad_norm": 1.258159756988173, "learning_rate": 2.6327850182769065e-08, "loss": 1.0481, "step": 3631 }, { "epoch": 4.8561872909699, "grad_norm": 1.3856756823770011, "learning_rate": 2.5851748904747176e-08, "loss": 0.6018, "step": 3632 }, { "epoch": 4.85752508361204, "grad_norm": 1.0342191401327228, "learning_rate": 2.5379980641955792e-08, "loss": 0.8789, "step": 3633 }, { "epoch": 4.858862876254181, "grad_norm": 0.9030460540240082, "learning_rate": 2.491254580535507e-08, "loss": 0.8706, "step": 3634 }, { "epoch": 4.860200668896321, "grad_norm": 1.212708611581969, "learning_rate": 2.4449444802130962e-08, "loss": 0.7945, "step": 3635 }, { "epoch": 4.861538461538462, "grad_norm": 0.8681543516797424, "learning_rate": 2.399067803569466e-08, "loss": 0.8044, "step": 3636 }, { "epoch": 4.862876254180602, "grad_norm": 1.02146924433177, "learning_rate": 2.3536245905679823e-08, "loss": 0.9055, "step": 3637 }, { "epoch": 4.864214046822743, "grad_norm": 0.9912335137875817, "learning_rate": 2.3086148807946463e-08, "loss": 0.8891, "step": 3638 }, { "epoch": 4.865551839464883, "grad_norm": 0.8744863410381111, "learning_rate": 2.264038713457706e-08, "loss": 0.9178, "step": 3639 }, { "epoch": 4.866889632107023, "grad_norm": 0.8418559450793044, "learning_rate": 2.219896127387766e-08, "loss": 0.7995, "step": 3640 }, { "epoch": 4.868227424749164, "grad_norm": 1.0079317221823951, "learning_rate": 2.1761871610376794e-08, "loss": 0.8259, "step": 3641 }, { "epoch": 4.869565217391305, "grad_norm": 0.9591213532635959, "learning_rate": 2.1329118524827662e-08, "loss": 1.1427, "step": 3642 }, { "epoch": 4.8709030100334445, "grad_norm": 1.2186727999817784, "learning_rate": 2.0900702394203165e-08, "loss": 0.8897, "step": 3643 }, { "epoch": 4.872240802675585, "grad_norm": 1.3480849858355848, "learning_rate": 2.047662359170033e-08, "loss": 0.6864, "step": 3644 }, { "epoch": 4.873578595317726, "grad_norm": 1.222736479610807, "learning_rate": 2.0056882486736982e-08, "loss": 0.6999, "step": 3645 }, { "epoch": 4.874916387959866, "grad_norm": 0.8196743325954746, "learning_rate": 1.9641479444952317e-08, "loss": 0.8386, "step": 3646 }, { "epoch": 4.8762541806020065, "grad_norm": 0.8933020866770496, "learning_rate": 1.9230414828206866e-08, "loss": 1.0148, "step": 3647 }, { "epoch": 4.877591973244147, "grad_norm": 0.9496488391374859, "learning_rate": 1.8823688994582533e-08, "loss": 1.0035, "step": 3648 }, { "epoch": 4.878929765886287, "grad_norm": 0.7850799898316354, "learning_rate": 1.842130229838035e-08, "loss": 0.9197, "step": 3649 }, { "epoch": 4.880267558528428, "grad_norm": 0.8029918978133967, "learning_rate": 1.802325509012215e-08, "loss": 0.8664, "step": 3650 }, { "epoch": 4.881605351170569, "grad_norm": 0.7617762836460311, "learning_rate": 1.762954771655001e-08, "loss": 1.037, "step": 3651 }, { "epoch": 4.882943143812709, "grad_norm": 0.900255218114835, "learning_rate": 1.724018052062515e-08, "loss": 1.0719, "step": 3652 }, { "epoch": 4.884280936454849, "grad_norm": 0.804588086832841, "learning_rate": 1.6855153841527915e-08, "loss": 1.0203, "step": 3653 }, { "epoch": 4.88561872909699, "grad_norm": 1.1526743779861206, "learning_rate": 1.647446801465724e-08, "loss": 0.9739, "step": 3654 }, { "epoch": 4.886956521739131, "grad_norm": 0.8944063854946138, "learning_rate": 1.60981233716323e-08, "loss": 1.1764, "step": 3655 }, { "epoch": 4.888294314381271, "grad_norm": 0.933564967290555, "learning_rate": 1.5726120240288632e-08, "loss": 1.1616, "step": 3656 }, { "epoch": 4.889632107023411, "grad_norm": 0.9141885286874358, "learning_rate": 1.5358458944680356e-08, "loss": 0.9085, "step": 3657 }, { "epoch": 4.890969899665552, "grad_norm": 0.8197206853874404, "learning_rate": 1.4995139805081272e-08, "loss": 0.8495, "step": 3658 }, { "epoch": 4.892307692307693, "grad_norm": 0.9081676651300283, "learning_rate": 1.4636163137980441e-08, "loss": 0.9825, "step": 3659 }, { "epoch": 4.8936454849498325, "grad_norm": 1.000950107553287, "learning_rate": 1.4281529256084393e-08, "loss": 1.0434, "step": 3660 }, { "epoch": 4.894983277591973, "grad_norm": 0.7885760494891673, "learning_rate": 1.393123846831823e-08, "loss": 1.1846, "step": 3661 }, { "epoch": 4.896321070234114, "grad_norm": 0.899330819260637, "learning_rate": 1.358529107982176e-08, "loss": 1.1127, "step": 3662 }, { "epoch": 4.897658862876254, "grad_norm": 0.9867787553132703, "learning_rate": 1.3243687391952809e-08, "loss": 1.1504, "step": 3663 }, { "epoch": 4.898996655518395, "grad_norm": 0.9279973156126679, "learning_rate": 1.2906427702284452e-08, "loss": 0.8552, "step": 3664 }, { "epoch": 4.900334448160535, "grad_norm": 1.1451109956838246, "learning_rate": 1.2573512304605574e-08, "loss": 0.909, "step": 3665 }, { "epoch": 4.901672240802675, "grad_norm": 0.9189256505096647, "learning_rate": 1.2244941488921968e-08, "loss": 0.9051, "step": 3666 }, { "epoch": 4.903010033444816, "grad_norm": 1.1904767087891894, "learning_rate": 1.1920715541453576e-08, "loss": 0.8246, "step": 3667 }, { "epoch": 4.904347826086957, "grad_norm": 0.7873257494834027, "learning_rate": 1.160083474463558e-08, "loss": 0.9535, "step": 3668 }, { "epoch": 4.905685618729097, "grad_norm": 0.932475179974749, "learning_rate": 1.1285299377118974e-08, "loss": 0.8256, "step": 3669 }, { "epoch": 4.907023411371237, "grad_norm": 1.213384851113868, "learning_rate": 1.0974109713768333e-08, "loss": 0.9768, "step": 3670 }, { "epoch": 4.908361204013378, "grad_norm": 1.0785020893006685, "learning_rate": 1.0667266025662925e-08, "loss": 0.9983, "step": 3671 }, { "epoch": 4.909698996655519, "grad_norm": 0.9936152354213452, "learning_rate": 1.0364768580097273e-08, "loss": 1.1733, "step": 3672 }, { "epoch": 4.9110367892976585, "grad_norm": 1.0017449698134087, "learning_rate": 1.006661764057837e-08, "loss": 0.8201, "step": 3673 }, { "epoch": 4.912374581939799, "grad_norm": 0.8125204964959144, "learning_rate": 9.772813466827347e-09, "loss": 0.9885, "step": 3674 }, { "epoch": 4.91371237458194, "grad_norm": 1.132470887013327, "learning_rate": 9.48335631477948e-09, "loss": 0.9424, "step": 3675 }, { "epoch": 4.915050167224081, "grad_norm": 1.051089309081772, "learning_rate": 9.19824643658307e-09, "loss": 0.9777, "step": 3676 }, { "epoch": 4.916387959866221, "grad_norm": 1.469020161541693, "learning_rate": 8.917484080599448e-09, "loss": 0.8002, "step": 3677 }, { "epoch": 4.917725752508361, "grad_norm": 1.0323600938413462, "learning_rate": 8.641069491401866e-09, "loss": 0.8918, "step": 3678 }, { "epoch": 4.919063545150502, "grad_norm": 0.811113309724705, "learning_rate": 8.369002909777713e-09, "loss": 0.9079, "step": 3679 }, { "epoch": 4.920401337792642, "grad_norm": 1.2220757777678652, "learning_rate": 8.101284572725743e-09, "loss": 0.9515, "step": 3680 }, { "epoch": 4.921739130434783, "grad_norm": 1.3475228234063932, "learning_rate": 7.837914713457184e-09, "loss": 0.9831, "step": 3681 }, { "epoch": 4.923076923076923, "grad_norm": 1.2403105361999662, "learning_rate": 7.57889356139574e-09, "loss": 0.9264, "step": 3682 }, { "epoch": 4.924414715719063, "grad_norm": 0.9509100882689079, "learning_rate": 7.324221342176474e-09, "loss": 0.8981, "step": 3683 }, { "epoch": 4.925752508361204, "grad_norm": 1.286993652275378, "learning_rate": 7.073898277645819e-09, "loss": 0.9094, "step": 3684 }, { "epoch": 4.927090301003345, "grad_norm": 0.9452357505768586, "learning_rate": 6.8279245858621226e-09, "loss": 0.8639, "step": 3685 }, { "epoch": 4.9284280936454845, "grad_norm": 1.0058204901416155, "learning_rate": 6.586300481095098e-09, "loss": 0.7919, "step": 3686 }, { "epoch": 4.929765886287625, "grad_norm": 0.8202949993275135, "learning_rate": 6.349026173824713e-09, "loss": 1.0955, "step": 3687 }, { "epoch": 4.931103678929766, "grad_norm": 1.0024099130217239, "learning_rate": 6.116101870742297e-09, "loss": 0.8047, "step": 3688 }, { "epoch": 4.932441471571907, "grad_norm": 1.1096266674902966, "learning_rate": 5.88752777474999e-09, "loss": 0.9566, "step": 3689 }, { "epoch": 4.933779264214047, "grad_norm": 0.851099122679725, "learning_rate": 5.6633040849601865e-09, "loss": 1.1837, "step": 3690 }, { "epoch": 4.935117056856187, "grad_norm": 0.8834423216417991, "learning_rate": 5.443430996695531e-09, "loss": 0.9961, "step": 3691 }, { "epoch": 4.936454849498328, "grad_norm": 0.8458841771557586, "learning_rate": 5.227908701490036e-09, "loss": 0.8603, "step": 3692 }, { "epoch": 4.937792642140468, "grad_norm": 0.9196597362865587, "learning_rate": 5.016737387085191e-09, "loss": 0.839, "step": 3693 }, { "epoch": 4.939130434782609, "grad_norm": 1.240988092825689, "learning_rate": 4.8099172374349576e-09, "loss": 0.7513, "step": 3694 }, { "epoch": 4.940468227424749, "grad_norm": 1.0952676076509933, "learning_rate": 4.607448432701333e-09, "loss": 0.9713, "step": 3695 }, { "epoch": 4.94180602006689, "grad_norm": 0.99885851176294, "learning_rate": 4.409331149256013e-09, "loss": 1.0735, "step": 3696 }, { "epoch": 4.94314381270903, "grad_norm": 0.9869538158899511, "learning_rate": 4.2155655596809455e-09, "loss": 1.0368, "step": 3697 }, { "epoch": 4.944481605351171, "grad_norm": 1.1643907349867157, "learning_rate": 4.026151832766112e-09, "loss": 1.0188, "step": 3698 }, { "epoch": 4.945819397993311, "grad_norm": 0.9199194137293993, "learning_rate": 3.841090133511749e-09, "loss": 0.7009, "step": 3699 }, { "epoch": 4.947157190635451, "grad_norm": 0.8943943976349935, "learning_rate": 3.66038062312557e-09, "loss": 1.0754, "step": 3700 }, { "epoch": 4.948494983277592, "grad_norm": 0.7759229244991375, "learning_rate": 3.484023459026098e-09, "loss": 0.9754, "step": 3701 }, { "epoch": 4.949832775919733, "grad_norm": 0.8417285867665204, "learning_rate": 3.3120187948382233e-09, "loss": 0.8639, "step": 3702 }, { "epoch": 4.9511705685618725, "grad_norm": 1.0335744059900291, "learning_rate": 3.144366780396535e-09, "loss": 0.8075, "step": 3703 }, { "epoch": 4.952508361204013, "grad_norm": 0.7616683518313792, "learning_rate": 2.981067561744211e-09, "loss": 1.107, "step": 3704 }, { "epoch": 4.953846153846154, "grad_norm": 0.7513828627273843, "learning_rate": 2.8221212811324616e-09, "loss": 1.0082, "step": 3705 }, { "epoch": 4.955183946488294, "grad_norm": 0.9074621794838299, "learning_rate": 2.6675280770199763e-09, "loss": 0.8957, "step": 3706 }, { "epoch": 4.956521739130435, "grad_norm": 0.8614664203520158, "learning_rate": 2.5172880840745873e-09, "loss": 0.9517, "step": 3707 }, { "epoch": 4.957859531772575, "grad_norm": 0.9319035483703629, "learning_rate": 2.371401433170495e-09, "loss": 1.4146, "step": 3708 }, { "epoch": 4.959197324414716, "grad_norm": 0.9861102410596019, "learning_rate": 2.229868251391598e-09, "loss": 1.2755, "step": 3709 }, { "epoch": 4.960535117056856, "grad_norm": 0.8844921642085508, "learning_rate": 2.0926886620281637e-09, "loss": 0.9124, "step": 3710 }, { "epoch": 4.961872909698997, "grad_norm": 1.0833177596315435, "learning_rate": 1.959862784577937e-09, "loss": 0.9497, "step": 3711 }, { "epoch": 4.963210702341137, "grad_norm": 0.8042507720478866, "learning_rate": 1.8313907347466963e-09, "loss": 0.9569, "step": 3712 }, { "epoch": 4.964548494983277, "grad_norm": 1.124252745776976, "learning_rate": 1.7072726244471427e-09, "loss": 0.8946, "step": 3713 }, { "epoch": 4.965886287625418, "grad_norm": 0.8373239913951783, "learning_rate": 1.5875085617994558e-09, "loss": 1.0133, "step": 3714 }, { "epoch": 4.967224080267559, "grad_norm": 0.8662005144977608, "learning_rate": 1.4720986511312928e-09, "loss": 1.0994, "step": 3715 }, { "epoch": 4.968561872909699, "grad_norm": 0.8475234074395852, "learning_rate": 1.36104299297668e-09, "loss": 0.9788, "step": 3716 }, { "epoch": 4.969899665551839, "grad_norm": 0.8666432594928133, "learning_rate": 1.2543416840771206e-09, "loss": 0.8235, "step": 3717 }, { "epoch": 4.97123745819398, "grad_norm": 0.9062567576958915, "learning_rate": 1.1519948173810414e-09, "loss": 1.0076, "step": 3718 }, { "epoch": 4.972575250836121, "grad_norm": 1.5092276234315132, "learning_rate": 1.054002482043237e-09, "loss": 0.8338, "step": 3719 }, { "epoch": 4.973913043478261, "grad_norm": 1.0480375428277175, "learning_rate": 9.603647634259806e-10, "loss": 0.8128, "step": 3720 }, { "epoch": 4.975250836120401, "grad_norm": 0.951244258756974, "learning_rate": 8.710817430968021e-10, "loss": 0.7106, "step": 3721 }, { "epoch": 4.976588628762542, "grad_norm": 0.8932821495333715, "learning_rate": 7.861534988312658e-10, "loss": 0.9915, "step": 3722 }, { "epoch": 4.977926421404682, "grad_norm": 0.8686400758868342, "learning_rate": 7.055801046113031e-10, "loss": 0.7885, "step": 3723 }, { "epoch": 4.979264214046823, "grad_norm": 0.7668892337681118, "learning_rate": 6.293616306246586e-10, "loss": 0.9195, "step": 3724 }, { "epoch": 4.980602006688963, "grad_norm": 1.0447053765573426, "learning_rate": 5.574981432659998e-10, "loss": 0.7025, "step": 3725 }, { "epoch": 4.981939799331103, "grad_norm": 0.88158886706116, "learning_rate": 4.899897051358071e-10, "loss": 0.7837, "step": 3726 }, { "epoch": 4.983277591973244, "grad_norm": 0.8210884734389188, "learning_rate": 4.2683637504092877e-10, "loss": 0.8882, "step": 3727 }, { "epoch": 4.984615384615385, "grad_norm": 0.8732828958593359, "learning_rate": 3.6803820799513613e-10, "loss": 0.8839, "step": 3728 }, { "epoch": 4.985953177257525, "grad_norm": 1.0575139455054343, "learning_rate": 3.1359525521801326e-10, "loss": 0.819, "step": 3729 }, { "epoch": 4.987290969899665, "grad_norm": 0.9288685799347429, "learning_rate": 2.6350756413440203e-10, "loss": 1.0515, "step": 3730 }, { "epoch": 4.988628762541806, "grad_norm": 1.0199255827105047, "learning_rate": 2.1777517837717755e-10, "loss": 0.9645, "step": 3731 }, { "epoch": 4.989966555183947, "grad_norm": 1.1246212873065353, "learning_rate": 1.7639813778336233e-10, "loss": 0.7914, "step": 3732 }, { "epoch": 4.9913043478260875, "grad_norm": 0.8868786719604985, "learning_rate": 1.3937647839690205e-10, "loss": 1.2467, "step": 3733 }, { "epoch": 4.992642140468227, "grad_norm": 0.954991383234249, "learning_rate": 1.0671023246755507e-10, "loss": 0.9009, "step": 3734 }, { "epoch": 4.993979933110368, "grad_norm": 0.9972286782354367, "learning_rate": 7.839942845144777e-11, "loss": 1.0393, "step": 3735 }, { "epoch": 4.995317725752509, "grad_norm": 0.8862485050387628, "learning_rate": 5.444409100996417e-11, "loss": 1.0069, "step": 3736 }, { "epoch": 4.996655518394649, "grad_norm": 0.9388449213773438, "learning_rate": 3.4844241011411375e-11, "loss": 0.9753, "step": 3737 }, { "epoch": 4.997993311036789, "grad_norm": 0.9055123919909552, "learning_rate": 1.959989552824393e-11, "loss": 1.0543, "step": 3738 }, { "epoch": 4.99933110367893, "grad_norm": 0.8843720333971772, "learning_rate": 8.711067840949661e-12, "loss": 0.9633, "step": 3739 }, { "epoch": 5.0, "grad_norm": 0.8843720333971772, "learning_rate": 2.1777674347189805e-12, "loss": 1.0245, "step": 3740 }, { "epoch": 5.0, "step": 3740, "total_flos": 428590214676480.0, "train_loss": 1.0611437029857687, "train_runtime": 76806.1738, "train_samples_per_second": 0.389, "train_steps_per_second": 0.049 } ], "logging_steps": 1, "max_steps": 3740, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 428590214676480.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }