{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 88688, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 95.03720582402384, "learning_rate": 7.515971439308531e-08, "loss": 3.8316, "step": 10 }, { "epoch": 0.0, "grad_norm": 22.60961881561774, "learning_rate": 1.5031942878617061e-07, "loss": 4.0681, "step": 20 }, { "epoch": 0.0, "grad_norm": 19.407385250646676, "learning_rate": 2.2547914317925593e-07, "loss": 3.7058, "step": 30 }, { "epoch": 0.0, "grad_norm": 10.58159167291745, "learning_rate": 3.0063885757234123e-07, "loss": 3.6046, "step": 40 }, { "epoch": 0.0, "grad_norm": 16.042680239079196, "learning_rate": 3.757985719654265e-07, "loss": 3.3742, "step": 50 }, { "epoch": 0.0, "grad_norm": 10.533367861877702, "learning_rate": 4.5095828635851187e-07, "loss": 3.3528, "step": 60 }, { "epoch": 0.0, "grad_norm": 5.360818531602447, "learning_rate": 5.261180007515972e-07, "loss": 3.2943, "step": 70 }, { "epoch": 0.0, "grad_norm": 4.6402167652093596, "learning_rate": 6.012777151446825e-07, "loss": 3.1295, "step": 80 }, { "epoch": 0.0, "grad_norm": 5.201681309312079, "learning_rate": 6.764374295377678e-07, "loss": 3.1071, "step": 90 }, { "epoch": 0.0, "grad_norm": 3.8128318113363147, "learning_rate": 7.51597143930853e-07, "loss": 3.0349, "step": 100 }, { "epoch": 0.0, "grad_norm": 4.984705496424015, "learning_rate": 8.267568583239384e-07, "loss": 3.1114, "step": 110 }, { "epoch": 0.0, "grad_norm": 3.80939172041421, "learning_rate": 9.019165727170237e-07, "loss": 3.2174, "step": 120 }, { "epoch": 0.0, "grad_norm": 4.458634517215646, "learning_rate": 9.77076287110109e-07, "loss": 3.1459, "step": 130 }, { "epoch": 0.0, "grad_norm": 4.549231638487384, "learning_rate": 1.0522360015031944e-06, "loss": 3.1007, "step": 140 }, { "epoch": 0.0, "grad_norm": 6.047796073463416, "learning_rate": 1.1273957158962797e-06, "loss": 3.1193, "step": 150 }, { "epoch": 0.0, "grad_norm": 5.065856045643684, "learning_rate": 1.202555430289365e-06, "loss": 2.9893, "step": 160 }, { "epoch": 0.0, "grad_norm": 3.72502351149555, "learning_rate": 1.2777151446824504e-06, "loss": 2.8768, "step": 170 }, { "epoch": 0.0, "grad_norm": 3.48869137886807, "learning_rate": 1.3528748590755356e-06, "loss": 2.801, "step": 180 }, { "epoch": 0.0, "grad_norm": 3.0933081794303208, "learning_rate": 1.428034573468621e-06, "loss": 2.8449, "step": 190 }, { "epoch": 0.0, "grad_norm": 4.082967960062426, "learning_rate": 1.503194287861706e-06, "loss": 2.8939, "step": 200 }, { "epoch": 0.0, "grad_norm": 3.9204399299948642, "learning_rate": 1.5783540022547915e-06, "loss": 2.7646, "step": 210 }, { "epoch": 0.0, "grad_norm": 3.1214589641852735, "learning_rate": 1.6535137166478768e-06, "loss": 2.8401, "step": 220 }, { "epoch": 0.01, "grad_norm": 2.966325117745589, "learning_rate": 1.7286734310409622e-06, "loss": 2.7687, "step": 230 }, { "epoch": 0.01, "grad_norm": 4.63562435922088, "learning_rate": 1.8038331454340475e-06, "loss": 2.8577, "step": 240 }, { "epoch": 0.01, "grad_norm": 4.255548675469941, "learning_rate": 1.878992859827133e-06, "loss": 2.8726, "step": 250 }, { "epoch": 0.01, "grad_norm": 2.7702690665795284, "learning_rate": 1.954152574220218e-06, "loss": 2.8574, "step": 260 }, { "epoch": 0.01, "grad_norm": 3.160295386049676, "learning_rate": 2.0293122886133034e-06, "loss": 2.7705, "step": 270 }, { "epoch": 0.01, "grad_norm": 3.30353964112357, "learning_rate": 2.104472003006389e-06, "loss": 2.846, "step": 280 }, { "epoch": 0.01, "grad_norm": 3.1877955493912156, "learning_rate": 2.179631717399474e-06, "loss": 2.8751, "step": 290 }, { "epoch": 0.01, "grad_norm": 2.6527725645743674, "learning_rate": 2.2547914317925593e-06, "loss": 2.8154, "step": 300 }, { "epoch": 0.01, "grad_norm": 3.35945641093628, "learning_rate": 2.329951146185645e-06, "loss": 2.7584, "step": 310 }, { "epoch": 0.01, "grad_norm": 3.320801767010909, "learning_rate": 2.40511086057873e-06, "loss": 2.8174, "step": 320 }, { "epoch": 0.01, "grad_norm": 3.4749206427808934, "learning_rate": 2.4802705749718153e-06, "loss": 2.8256, "step": 330 }, { "epoch": 0.01, "grad_norm": 2.905805421333625, "learning_rate": 2.5554302893649007e-06, "loss": 2.7489, "step": 340 }, { "epoch": 0.01, "grad_norm": 3.553762566552604, "learning_rate": 2.6305900037579858e-06, "loss": 2.8857, "step": 350 }, { "epoch": 0.01, "grad_norm": 4.371054747885646, "learning_rate": 2.7057497181510712e-06, "loss": 2.828, "step": 360 }, { "epoch": 0.01, "grad_norm": 3.5024781016028483, "learning_rate": 2.7809094325441567e-06, "loss": 2.7623, "step": 370 }, { "epoch": 0.01, "grad_norm": 3.0399017760796547, "learning_rate": 2.856069146937242e-06, "loss": 2.7234, "step": 380 }, { "epoch": 0.01, "grad_norm": 2.9491427443128315, "learning_rate": 2.9312288613303267e-06, "loss": 2.8057, "step": 390 }, { "epoch": 0.01, "grad_norm": 3.5768795637844533, "learning_rate": 3.006388575723412e-06, "loss": 2.6538, "step": 400 }, { "epoch": 0.01, "grad_norm": 3.9431994815447546, "learning_rate": 3.0815482901164976e-06, "loss": 2.7372, "step": 410 }, { "epoch": 0.01, "grad_norm": 4.7313819774667385, "learning_rate": 3.156708004509583e-06, "loss": 2.7818, "step": 420 }, { "epoch": 0.01, "grad_norm": 3.41677520638338, "learning_rate": 3.2318677189026685e-06, "loss": 2.727, "step": 430 }, { "epoch": 0.01, "grad_norm": 2.567948819321526, "learning_rate": 3.3070274332957536e-06, "loss": 2.5756, "step": 440 }, { "epoch": 0.01, "grad_norm": 2.9941576022264225, "learning_rate": 3.382187147688839e-06, "loss": 2.732, "step": 450 }, { "epoch": 0.01, "grad_norm": 3.163899042749549, "learning_rate": 3.4573468620819245e-06, "loss": 2.7437, "step": 460 }, { "epoch": 0.01, "grad_norm": 3.4120263084516536, "learning_rate": 3.53250657647501e-06, "loss": 2.7112, "step": 470 }, { "epoch": 0.01, "grad_norm": 2.406468539097057, "learning_rate": 3.607666290868095e-06, "loss": 2.7846, "step": 480 }, { "epoch": 0.01, "grad_norm": 3.393012764613268, "learning_rate": 3.6828260052611804e-06, "loss": 2.8518, "step": 490 }, { "epoch": 0.01, "grad_norm": 4.59665529924303, "learning_rate": 3.757985719654266e-06, "loss": 2.7689, "step": 500 }, { "epoch": 0.01, "grad_norm": 3.4196947095059946, "learning_rate": 3.833145434047351e-06, "loss": 2.7865, "step": 510 }, { "epoch": 0.01, "grad_norm": 3.4181899493508836, "learning_rate": 3.908305148440436e-06, "loss": 2.7394, "step": 520 }, { "epoch": 0.01, "grad_norm": 2.498252181943101, "learning_rate": 3.983464862833522e-06, "loss": 2.8291, "step": 530 }, { "epoch": 0.01, "grad_norm": 3.464288370714309, "learning_rate": 4.058624577226607e-06, "loss": 2.7362, "step": 540 }, { "epoch": 0.01, "grad_norm": 4.1812854544104905, "learning_rate": 4.133784291619693e-06, "loss": 2.7795, "step": 550 }, { "epoch": 0.01, "grad_norm": 3.05433638779607, "learning_rate": 4.208944006012778e-06, "loss": 2.8088, "step": 560 }, { "epoch": 0.01, "grad_norm": 4.338417918242724, "learning_rate": 4.284103720405863e-06, "loss": 2.8252, "step": 570 }, { "epoch": 0.01, "grad_norm": 2.8531148444337586, "learning_rate": 4.359263434798948e-06, "loss": 2.7267, "step": 580 }, { "epoch": 0.01, "grad_norm": 3.1104850832898676, "learning_rate": 4.434423149192034e-06, "loss": 2.8945, "step": 590 }, { "epoch": 0.01, "grad_norm": 2.7519579438140314, "learning_rate": 4.509582863585119e-06, "loss": 2.7379, "step": 600 }, { "epoch": 0.01, "grad_norm": 3.4519679210933503, "learning_rate": 4.584742577978204e-06, "loss": 2.7569, "step": 610 }, { "epoch": 0.01, "grad_norm": 3.5946084291207945, "learning_rate": 4.65990229237129e-06, "loss": 2.7538, "step": 620 }, { "epoch": 0.01, "grad_norm": 2.8368669270608637, "learning_rate": 4.735062006764375e-06, "loss": 2.8342, "step": 630 }, { "epoch": 0.01, "grad_norm": 2.753200126538317, "learning_rate": 4.81022172115746e-06, "loss": 2.7296, "step": 640 }, { "epoch": 0.01, "grad_norm": 3.2893787353247594, "learning_rate": 4.8853814355505455e-06, "loss": 2.8404, "step": 650 }, { "epoch": 0.01, "grad_norm": 2.710652643862423, "learning_rate": 4.9605411499436306e-06, "loss": 2.7239, "step": 660 }, { "epoch": 0.02, "grad_norm": 4.214358320782758, "learning_rate": 5.035700864336716e-06, "loss": 2.8073, "step": 670 }, { "epoch": 0.02, "grad_norm": 3.2079396507760585, "learning_rate": 5.1108605787298015e-06, "loss": 2.7874, "step": 680 }, { "epoch": 0.02, "grad_norm": 4.272772840898047, "learning_rate": 5.1860202931228865e-06, "loss": 2.7287, "step": 690 }, { "epoch": 0.02, "grad_norm": 3.019687348582402, "learning_rate": 5.2611800075159715e-06, "loss": 2.7535, "step": 700 }, { "epoch": 0.02, "grad_norm": 2.9094533088162757, "learning_rate": 5.336339721909057e-06, "loss": 2.7841, "step": 710 }, { "epoch": 0.02, "grad_norm": 3.0904960435059, "learning_rate": 5.4114994363021424e-06, "loss": 2.8449, "step": 720 }, { "epoch": 0.02, "grad_norm": 3.2516189896971808, "learning_rate": 5.486659150695228e-06, "loss": 2.7745, "step": 730 }, { "epoch": 0.02, "grad_norm": 5.555388098601699, "learning_rate": 5.561818865088313e-06, "loss": 2.8021, "step": 740 }, { "epoch": 0.02, "grad_norm": 4.369719762064073, "learning_rate": 5.636978579481398e-06, "loss": 2.85, "step": 750 }, { "epoch": 0.02, "grad_norm": 2.9250268949985956, "learning_rate": 5.712138293874484e-06, "loss": 2.6548, "step": 760 }, { "epoch": 0.02, "grad_norm": 3.09038608258868, "learning_rate": 5.787298008267569e-06, "loss": 2.8149, "step": 770 }, { "epoch": 0.02, "grad_norm": 4.043172775908715, "learning_rate": 5.8624577226606534e-06, "loss": 2.6717, "step": 780 }, { "epoch": 0.02, "grad_norm": 3.7427966986997747, "learning_rate": 5.93761743705374e-06, "loss": 2.8, "step": 790 }, { "epoch": 0.02, "grad_norm": 3.7477858857033226, "learning_rate": 6.012777151446824e-06, "loss": 2.8216, "step": 800 }, { "epoch": 0.02, "grad_norm": 4.640149169407717, "learning_rate": 6.087936865839911e-06, "loss": 2.6453, "step": 810 }, { "epoch": 0.02, "grad_norm": 3.7434024964001713, "learning_rate": 6.163096580232995e-06, "loss": 2.7491, "step": 820 }, { "epoch": 0.02, "grad_norm": 3.1393871397779884, "learning_rate": 6.23825629462608e-06, "loss": 2.8601, "step": 830 }, { "epoch": 0.02, "grad_norm": 2.9405762700147124, "learning_rate": 6.313416009019166e-06, "loss": 2.7405, "step": 840 }, { "epoch": 0.02, "grad_norm": 2.6912695896317853, "learning_rate": 6.388575723412251e-06, "loss": 2.7094, "step": 850 }, { "epoch": 0.02, "grad_norm": 3.144546161450778, "learning_rate": 6.463735437805337e-06, "loss": 2.7101, "step": 860 }, { "epoch": 0.02, "grad_norm": 2.9474457164554075, "learning_rate": 6.538895152198422e-06, "loss": 2.8321, "step": 870 }, { "epoch": 0.02, "grad_norm": 4.084868403400088, "learning_rate": 6.614054866591507e-06, "loss": 2.8372, "step": 880 }, { "epoch": 0.02, "grad_norm": 3.089184711173101, "learning_rate": 6.689214580984593e-06, "loss": 2.7976, "step": 890 }, { "epoch": 0.02, "grad_norm": 4.709064646381547, "learning_rate": 6.764374295377678e-06, "loss": 2.7403, "step": 900 }, { "epoch": 0.02, "grad_norm": 3.8330105859011008, "learning_rate": 6.839534009770763e-06, "loss": 2.7203, "step": 910 }, { "epoch": 0.02, "grad_norm": 3.5951412959126605, "learning_rate": 6.914693724163849e-06, "loss": 2.6352, "step": 920 }, { "epoch": 0.02, "grad_norm": 3.5034343630261757, "learning_rate": 6.989853438556934e-06, "loss": 2.8003, "step": 930 }, { "epoch": 0.02, "grad_norm": 2.6226996077233236, "learning_rate": 7.06501315295002e-06, "loss": 2.7928, "step": 940 }, { "epoch": 0.02, "grad_norm": 3.909329962102813, "learning_rate": 7.140172867343105e-06, "loss": 2.7292, "step": 950 }, { "epoch": 0.02, "grad_norm": 3.174083397866648, "learning_rate": 7.21533258173619e-06, "loss": 2.862, "step": 960 }, { "epoch": 0.02, "grad_norm": 2.7402531034342243, "learning_rate": 7.290492296129276e-06, "loss": 2.7015, "step": 970 }, { "epoch": 0.02, "grad_norm": 3.584531877145179, "learning_rate": 7.365652010522361e-06, "loss": 2.7996, "step": 980 }, { "epoch": 0.02, "grad_norm": 2.9821405363653377, "learning_rate": 7.440811724915446e-06, "loss": 2.7288, "step": 990 }, { "epoch": 0.02, "grad_norm": 2.442415229786894, "learning_rate": 7.515971439308532e-06, "loss": 2.7239, "step": 1000 }, { "epoch": 0.02, "grad_norm": 2.723818755445013, "learning_rate": 7.591131153701617e-06, "loss": 2.7496, "step": 1010 }, { "epoch": 0.02, "grad_norm": 4.626529726613246, "learning_rate": 7.666290868094702e-06, "loss": 2.8494, "step": 1020 }, { "epoch": 0.02, "grad_norm": 7.011143581734417, "learning_rate": 7.741450582487787e-06, "loss": 2.767, "step": 1030 }, { "epoch": 0.02, "grad_norm": 2.516276050570755, "learning_rate": 7.816610296880872e-06, "loss": 2.6905, "step": 1040 }, { "epoch": 0.02, "grad_norm": 3.795712718841949, "learning_rate": 7.891770011273959e-06, "loss": 2.7443, "step": 1050 }, { "epoch": 0.02, "grad_norm": 2.969146205584421, "learning_rate": 7.966929725667044e-06, "loss": 2.7071, "step": 1060 }, { "epoch": 0.02, "grad_norm": 2.6979085229959856, "learning_rate": 8.042089440060129e-06, "loss": 2.7099, "step": 1070 }, { "epoch": 0.02, "grad_norm": 2.5500650966370286, "learning_rate": 8.117249154453214e-06, "loss": 2.7826, "step": 1080 }, { "epoch": 0.02, "grad_norm": 3.6928270358347124, "learning_rate": 8.192408868846299e-06, "loss": 2.761, "step": 1090 }, { "epoch": 0.02, "grad_norm": 2.460337068142729, "learning_rate": 8.267568583239385e-06, "loss": 2.7428, "step": 1100 }, { "epoch": 0.03, "grad_norm": 2.269967349898418, "learning_rate": 8.34272829763247e-06, "loss": 2.7861, "step": 1110 }, { "epoch": 0.03, "grad_norm": 2.357446065583618, "learning_rate": 8.417888012025555e-06, "loss": 2.8114, "step": 1120 }, { "epoch": 0.03, "grad_norm": 3.5963275679000652, "learning_rate": 8.49304772641864e-06, "loss": 2.7154, "step": 1130 }, { "epoch": 0.03, "grad_norm": 2.7242340320764034, "learning_rate": 8.568207440811726e-06, "loss": 2.7532, "step": 1140 }, { "epoch": 0.03, "grad_norm": 3.041794123381994, "learning_rate": 8.64336715520481e-06, "loss": 2.7445, "step": 1150 }, { "epoch": 0.03, "grad_norm": 2.921482221030251, "learning_rate": 8.718526869597896e-06, "loss": 2.6676, "step": 1160 }, { "epoch": 0.03, "grad_norm": 2.975417501858349, "learning_rate": 8.79368658399098e-06, "loss": 2.7314, "step": 1170 }, { "epoch": 0.03, "grad_norm": 2.3054172662199472, "learning_rate": 8.868846298384067e-06, "loss": 2.786, "step": 1180 }, { "epoch": 0.03, "grad_norm": 3.444437720291339, "learning_rate": 8.944006012777152e-06, "loss": 2.6698, "step": 1190 }, { "epoch": 0.03, "grad_norm": 5.283509617632523, "learning_rate": 9.019165727170237e-06, "loss": 2.7284, "step": 1200 }, { "epoch": 0.03, "grad_norm": 3.1100864194737485, "learning_rate": 9.094325441563322e-06, "loss": 2.6557, "step": 1210 }, { "epoch": 0.03, "grad_norm": 3.9478795442142647, "learning_rate": 9.169485155956407e-06, "loss": 2.6876, "step": 1220 }, { "epoch": 0.03, "grad_norm": 2.56514947371657, "learning_rate": 9.244644870349494e-06, "loss": 2.8169, "step": 1230 }, { "epoch": 0.03, "grad_norm": 3.638768911531967, "learning_rate": 9.31980458474258e-06, "loss": 2.7331, "step": 1240 }, { "epoch": 0.03, "grad_norm": 2.2947718496858225, "learning_rate": 9.394964299135664e-06, "loss": 2.7059, "step": 1250 }, { "epoch": 0.03, "grad_norm": 2.1721616997730866, "learning_rate": 9.47012401352875e-06, "loss": 2.8024, "step": 1260 }, { "epoch": 0.03, "grad_norm": 2.270572852407896, "learning_rate": 9.545283727921834e-06, "loss": 2.7808, "step": 1270 }, { "epoch": 0.03, "grad_norm": 4.374484379789212, "learning_rate": 9.62044344231492e-06, "loss": 2.6319, "step": 1280 }, { "epoch": 0.03, "grad_norm": 3.133250439278576, "learning_rate": 9.695603156708006e-06, "loss": 2.8632, "step": 1290 }, { "epoch": 0.03, "grad_norm": 2.520906441946405, "learning_rate": 9.770762871101091e-06, "loss": 2.6632, "step": 1300 }, { "epoch": 0.03, "grad_norm": 6.834035872039535, "learning_rate": 9.845922585494176e-06, "loss": 2.7887, "step": 1310 }, { "epoch": 0.03, "grad_norm": 2.455979653734193, "learning_rate": 9.921082299887261e-06, "loss": 2.7575, "step": 1320 }, { "epoch": 0.03, "grad_norm": 2.713754872670945, "learning_rate": 9.996242014280346e-06, "loss": 2.6911, "step": 1330 }, { "epoch": 0.03, "grad_norm": 2.712683276952514, "learning_rate": 1.0071401728673431e-05, "loss": 2.6959, "step": 1340 }, { "epoch": 0.03, "grad_norm": 2.28766944280398, "learning_rate": 1.0146561443066518e-05, "loss": 2.7896, "step": 1350 }, { "epoch": 0.03, "grad_norm": 5.8474488227317725, "learning_rate": 1.0221721157459603e-05, "loss": 2.6944, "step": 1360 }, { "epoch": 0.03, "grad_norm": 5.81665380699611, "learning_rate": 1.0296880871852688e-05, "loss": 2.6487, "step": 1370 }, { "epoch": 0.03, "grad_norm": 5.276102214809851, "learning_rate": 1.0372040586245773e-05, "loss": 2.6217, "step": 1380 }, { "epoch": 0.03, "grad_norm": 2.9148455538005633, "learning_rate": 1.0447200300638858e-05, "loss": 2.6949, "step": 1390 }, { "epoch": 0.03, "grad_norm": 2.917583352691097, "learning_rate": 1.0522360015031943e-05, "loss": 2.8012, "step": 1400 }, { "epoch": 0.03, "grad_norm": 2.4082305528674937, "learning_rate": 1.059751972942503e-05, "loss": 2.5955, "step": 1410 }, { "epoch": 0.03, "grad_norm": 3.0966228814804375, "learning_rate": 1.0672679443818115e-05, "loss": 2.7541, "step": 1420 }, { "epoch": 0.03, "grad_norm": 2.6572256642773175, "learning_rate": 1.07478391582112e-05, "loss": 2.899, "step": 1430 }, { "epoch": 0.03, "grad_norm": 3.1249395865686753, "learning_rate": 1.0822998872604285e-05, "loss": 2.7242, "step": 1440 }, { "epoch": 0.03, "grad_norm": 3.331959379013374, "learning_rate": 1.089815858699737e-05, "loss": 2.6248, "step": 1450 }, { "epoch": 0.03, "grad_norm": 2.647555571108222, "learning_rate": 1.0973318301390457e-05, "loss": 2.7634, "step": 1460 }, { "epoch": 0.03, "grad_norm": 3.478255829297643, "learning_rate": 1.1048478015783542e-05, "loss": 2.8924, "step": 1470 }, { "epoch": 0.03, "grad_norm": 2.258603587154346, "learning_rate": 1.1123637730176627e-05, "loss": 2.7008, "step": 1480 }, { "epoch": 0.03, "grad_norm": 3.022898055095653, "learning_rate": 1.1198797444569712e-05, "loss": 2.8386, "step": 1490 }, { "epoch": 0.03, "grad_norm": 6.1970200658356145, "learning_rate": 1.1273957158962797e-05, "loss": 2.6887, "step": 1500 }, { "epoch": 0.03, "grad_norm": 2.6237202261144175, "learning_rate": 1.1349116873355883e-05, "loss": 2.8548, "step": 1510 }, { "epoch": 0.03, "grad_norm": 2.3773812175988205, "learning_rate": 1.1424276587748968e-05, "loss": 2.7636, "step": 1520 }, { "epoch": 0.03, "grad_norm": 2.803597405958123, "learning_rate": 1.1499436302142054e-05, "loss": 2.7214, "step": 1530 }, { "epoch": 0.03, "grad_norm": 2.359490308903166, "learning_rate": 1.1574596016535139e-05, "loss": 2.8205, "step": 1540 }, { "epoch": 0.03, "grad_norm": 2.9051046274935257, "learning_rate": 1.1649755730928224e-05, "loss": 2.6978, "step": 1550 }, { "epoch": 0.04, "grad_norm": 2.533407720316074, "learning_rate": 1.1724915445321307e-05, "loss": 2.6147, "step": 1560 }, { "epoch": 0.04, "grad_norm": 2.787991641903505, "learning_rate": 1.1800075159714395e-05, "loss": 2.766, "step": 1570 }, { "epoch": 0.04, "grad_norm": 3.305320068100113, "learning_rate": 1.187523487410748e-05, "loss": 2.752, "step": 1580 }, { "epoch": 0.04, "grad_norm": 3.9642367497683564, "learning_rate": 1.1950394588500565e-05, "loss": 2.805, "step": 1590 }, { "epoch": 0.04, "grad_norm": 3.035371546987386, "learning_rate": 1.2025554302893649e-05, "loss": 2.8673, "step": 1600 }, { "epoch": 0.04, "grad_norm": 3.173692570241095, "learning_rate": 1.2100714017286734e-05, "loss": 2.7707, "step": 1610 }, { "epoch": 0.04, "grad_norm": 2.4551298701408357, "learning_rate": 1.2175873731679822e-05, "loss": 2.7409, "step": 1620 }, { "epoch": 0.04, "grad_norm": 2.7020637286798026, "learning_rate": 1.2251033446072907e-05, "loss": 2.7564, "step": 1630 }, { "epoch": 0.04, "grad_norm": 2.926526344838212, "learning_rate": 1.232619316046599e-05, "loss": 2.7962, "step": 1640 }, { "epoch": 0.04, "grad_norm": 3.1249317099021714, "learning_rate": 1.2401352874859076e-05, "loss": 2.829, "step": 1650 }, { "epoch": 0.04, "grad_norm": 3.7654775666236895, "learning_rate": 1.247651258925216e-05, "loss": 2.6808, "step": 1660 }, { "epoch": 0.04, "grad_norm": 3.744408523035668, "learning_rate": 1.2551672303645247e-05, "loss": 2.821, "step": 1670 }, { "epoch": 0.04, "grad_norm": 3.5410208230850597, "learning_rate": 1.2626832018038332e-05, "loss": 2.6907, "step": 1680 }, { "epoch": 0.04, "grad_norm": 2.3006164816618915, "learning_rate": 1.2701991732431417e-05, "loss": 2.7101, "step": 1690 }, { "epoch": 0.04, "grad_norm": 2.450196293485681, "learning_rate": 1.2777151446824502e-05, "loss": 2.6883, "step": 1700 }, { "epoch": 0.04, "grad_norm": 4.20628362822278, "learning_rate": 1.2852311161217587e-05, "loss": 2.6779, "step": 1710 }, { "epoch": 0.04, "grad_norm": 3.6184921252186886, "learning_rate": 1.2927470875610674e-05, "loss": 2.7325, "step": 1720 }, { "epoch": 0.04, "grad_norm": 2.626948026691171, "learning_rate": 1.300263059000376e-05, "loss": 2.7781, "step": 1730 }, { "epoch": 0.04, "grad_norm": 2.520163212935597, "learning_rate": 1.3077790304396844e-05, "loss": 2.7743, "step": 1740 }, { "epoch": 0.04, "grad_norm": 2.6381176131023025, "learning_rate": 1.315295001878993e-05, "loss": 2.6825, "step": 1750 }, { "epoch": 0.04, "grad_norm": 2.6846863282370927, "learning_rate": 1.3228109733183014e-05, "loss": 2.7234, "step": 1760 }, { "epoch": 0.04, "grad_norm": 3.2129364886242775, "learning_rate": 1.33032694475761e-05, "loss": 2.6381, "step": 1770 }, { "epoch": 0.04, "grad_norm": 3.617057954822954, "learning_rate": 1.3378429161969186e-05, "loss": 2.7838, "step": 1780 }, { "epoch": 0.04, "grad_norm": 3.457240222167099, "learning_rate": 1.3453588876362271e-05, "loss": 2.7084, "step": 1790 }, { "epoch": 0.04, "grad_norm": 4.817992941779454, "learning_rate": 1.3528748590755356e-05, "loss": 2.652, "step": 1800 }, { "epoch": 0.04, "grad_norm": 2.421287333310825, "learning_rate": 1.3603908305148441e-05, "loss": 2.7085, "step": 1810 }, { "epoch": 0.04, "grad_norm": 2.4541912876726424, "learning_rate": 1.3679068019541526e-05, "loss": 2.8116, "step": 1820 }, { "epoch": 0.04, "grad_norm": 4.48325373086683, "learning_rate": 1.3754227733934613e-05, "loss": 2.7404, "step": 1830 }, { "epoch": 0.04, "grad_norm": 2.6844930953540707, "learning_rate": 1.3829387448327698e-05, "loss": 2.6829, "step": 1840 }, { "epoch": 0.04, "grad_norm": 4.194516442354704, "learning_rate": 1.3904547162720783e-05, "loss": 2.6746, "step": 1850 }, { "epoch": 0.04, "grad_norm": 3.4715344867067817, "learning_rate": 1.3979706877113868e-05, "loss": 2.7863, "step": 1860 }, { "epoch": 0.04, "grad_norm": 2.6799620884290567, "learning_rate": 1.4054866591506953e-05, "loss": 2.6668, "step": 1870 }, { "epoch": 0.04, "grad_norm": 2.734720252287986, "learning_rate": 1.413002630590004e-05, "loss": 2.7154, "step": 1880 }, { "epoch": 0.04, "grad_norm": 4.084508688732074, "learning_rate": 1.4205186020293125e-05, "loss": 2.6856, "step": 1890 }, { "epoch": 0.04, "grad_norm": 5.260904360558632, "learning_rate": 1.428034573468621e-05, "loss": 2.7287, "step": 1900 }, { "epoch": 0.04, "grad_norm": 4.316671325576979, "learning_rate": 1.4355505449079295e-05, "loss": 2.7799, "step": 1910 }, { "epoch": 0.04, "grad_norm": 5.7857740457730635, "learning_rate": 1.443066516347238e-05, "loss": 2.7402, "step": 1920 }, { "epoch": 0.04, "grad_norm": 3.4908168752107533, "learning_rate": 1.4505824877865465e-05, "loss": 2.8095, "step": 1930 }, { "epoch": 0.04, "grad_norm": 3.889405047130587, "learning_rate": 1.4580984592258552e-05, "loss": 2.7798, "step": 1940 }, { "epoch": 0.04, "grad_norm": 2.2054198160381504, "learning_rate": 1.4656144306651637e-05, "loss": 2.8111, "step": 1950 }, { "epoch": 0.04, "grad_norm": 2.568005716360079, "learning_rate": 1.4731304021044722e-05, "loss": 2.6407, "step": 1960 }, { "epoch": 0.04, "grad_norm": 2.204469163559196, "learning_rate": 1.4806463735437807e-05, "loss": 2.7224, "step": 1970 }, { "epoch": 0.04, "grad_norm": 2.5853001798200776, "learning_rate": 1.4881623449830892e-05, "loss": 2.6465, "step": 1980 }, { "epoch": 0.04, "grad_norm": 2.388092590944376, "learning_rate": 1.4956783164223978e-05, "loss": 2.6617, "step": 1990 }, { "epoch": 0.05, "grad_norm": 2.611731098250706, "learning_rate": 1.5031942878617063e-05, "loss": 2.7992, "step": 2000 }, { "epoch": 0.05, "grad_norm": 4.8331252447819635, "learning_rate": 1.5107102593010148e-05, "loss": 2.6326, "step": 2010 }, { "epoch": 0.05, "grad_norm": 4.669892917002436, "learning_rate": 1.5182262307403233e-05, "loss": 2.7049, "step": 2020 }, { "epoch": 0.05, "grad_norm": 3.5612291536867593, "learning_rate": 1.5257422021796317e-05, "loss": 2.7633, "step": 2030 }, { "epoch": 0.05, "grad_norm": 3.093273834786886, "learning_rate": 1.5332581736189404e-05, "loss": 2.7127, "step": 2040 }, { "epoch": 0.05, "grad_norm": 2.352512208624229, "learning_rate": 1.540774145058249e-05, "loss": 2.6959, "step": 2050 }, { "epoch": 0.05, "grad_norm": 2.1267620179222413, "learning_rate": 1.5482901164975574e-05, "loss": 2.8181, "step": 2060 }, { "epoch": 0.05, "grad_norm": 6.829009870497989, "learning_rate": 1.555806087936866e-05, "loss": 2.7797, "step": 2070 }, { "epoch": 0.05, "grad_norm": 5.100887356943923, "learning_rate": 1.5633220593761744e-05, "loss": 2.7301, "step": 2080 }, { "epoch": 0.05, "grad_norm": 2.241157268697907, "learning_rate": 1.570838030815483e-05, "loss": 2.7788, "step": 2090 }, { "epoch": 0.05, "grad_norm": 2.426295966851187, "learning_rate": 1.5783540022547917e-05, "loss": 2.8305, "step": 2100 }, { "epoch": 0.05, "grad_norm": 3.1563278651264213, "learning_rate": 1.5858699736941e-05, "loss": 2.7877, "step": 2110 }, { "epoch": 0.05, "grad_norm": 3.5003651705261962, "learning_rate": 1.5933859451334087e-05, "loss": 2.7581, "step": 2120 }, { "epoch": 0.05, "grad_norm": 2.590500112411194, "learning_rate": 1.600901916572717e-05, "loss": 2.7514, "step": 2130 }, { "epoch": 0.05, "grad_norm": 2.545516437393321, "learning_rate": 1.6084178880120257e-05, "loss": 2.7304, "step": 2140 }, { "epoch": 0.05, "grad_norm": 4.073624478846258, "learning_rate": 1.6159338594513344e-05, "loss": 2.7662, "step": 2150 }, { "epoch": 0.05, "grad_norm": 2.379803910065847, "learning_rate": 1.6234498308906427e-05, "loss": 2.7113, "step": 2160 }, { "epoch": 0.05, "grad_norm": 3.970383432519192, "learning_rate": 1.6309658023299514e-05, "loss": 2.6236, "step": 2170 }, { "epoch": 0.05, "grad_norm": 4.666379922199015, "learning_rate": 1.6384817737692597e-05, "loss": 2.7609, "step": 2180 }, { "epoch": 0.05, "grad_norm": 2.7423145418026684, "learning_rate": 1.6459977452085684e-05, "loss": 2.8015, "step": 2190 }, { "epoch": 0.05, "grad_norm": 2.3372559741977774, "learning_rate": 1.653513716647877e-05, "loss": 2.8699, "step": 2200 }, { "epoch": 0.05, "grad_norm": 7.9735079020620985, "learning_rate": 1.6610296880871854e-05, "loss": 2.757, "step": 2210 }, { "epoch": 0.05, "grad_norm": 2.731163398930488, "learning_rate": 1.668545659526494e-05, "loss": 2.7593, "step": 2220 }, { "epoch": 0.05, "grad_norm": 4.057229998176962, "learning_rate": 1.6760616309658024e-05, "loss": 2.782, "step": 2230 }, { "epoch": 0.05, "grad_norm": 2.6080144507446796, "learning_rate": 1.683577602405111e-05, "loss": 2.8282, "step": 2240 }, { "epoch": 0.05, "grad_norm": 2.2833099647163606, "learning_rate": 1.6910935738444198e-05, "loss": 2.748, "step": 2250 }, { "epoch": 0.05, "grad_norm": 3.676420386436387, "learning_rate": 1.698609545283728e-05, "loss": 2.845, "step": 2260 }, { "epoch": 0.05, "grad_norm": 2.6072097587266496, "learning_rate": 1.7061255167230368e-05, "loss": 2.8564, "step": 2270 }, { "epoch": 0.05, "grad_norm": 3.729264685868158, "learning_rate": 1.713641488162345e-05, "loss": 2.7286, "step": 2280 }, { "epoch": 0.05, "grad_norm": 2.9716692892807934, "learning_rate": 1.7211574596016534e-05, "loss": 2.6822, "step": 2290 }, { "epoch": 0.05, "grad_norm": 2.7901216503222632, "learning_rate": 1.728673431040962e-05, "loss": 2.8068, "step": 2300 }, { "epoch": 0.05, "grad_norm": 3.7137753747661204, "learning_rate": 1.7361894024802708e-05, "loss": 2.741, "step": 2310 }, { "epoch": 0.05, "grad_norm": 2.131209041102546, "learning_rate": 1.743705373919579e-05, "loss": 2.7939, "step": 2320 }, { "epoch": 0.05, "grad_norm": 2.3890405321147794, "learning_rate": 1.7512213453588878e-05, "loss": 2.7757, "step": 2330 }, { "epoch": 0.05, "grad_norm": 2.4432494112392833, "learning_rate": 1.758737316798196e-05, "loss": 2.8355, "step": 2340 }, { "epoch": 0.05, "grad_norm": 2.4664731744727137, "learning_rate": 1.7662532882375048e-05, "loss": 2.6958, "step": 2350 }, { "epoch": 0.05, "grad_norm": 2.6071164995653775, "learning_rate": 1.7737692596768135e-05, "loss": 2.8254, "step": 2360 }, { "epoch": 0.05, "grad_norm": 2.68189763162259, "learning_rate": 1.7812852311161218e-05, "loss": 2.7171, "step": 2370 }, { "epoch": 0.05, "grad_norm": 2.779282629231184, "learning_rate": 1.7888012025554305e-05, "loss": 2.7564, "step": 2380 }, { "epoch": 0.05, "grad_norm": 3.0476734543151958, "learning_rate": 1.7963171739947388e-05, "loss": 2.7255, "step": 2390 }, { "epoch": 0.05, "grad_norm": 2.640144511665021, "learning_rate": 1.8038331454340475e-05, "loss": 2.7694, "step": 2400 }, { "epoch": 0.05, "grad_norm": 2.2073430627236807, "learning_rate": 1.811349116873356e-05, "loss": 2.847, "step": 2410 }, { "epoch": 0.05, "grad_norm": 3.5324600497874084, "learning_rate": 1.8188650883126645e-05, "loss": 2.8425, "step": 2420 }, { "epoch": 0.05, "grad_norm": 2.524810381227783, "learning_rate": 1.826381059751973e-05, "loss": 2.8193, "step": 2430 }, { "epoch": 0.06, "grad_norm": 2.6894394000129838, "learning_rate": 1.8338970311912815e-05, "loss": 2.9274, "step": 2440 }, { "epoch": 0.06, "grad_norm": 5.344469336364714, "learning_rate": 1.84141300263059e-05, "loss": 2.637, "step": 2450 }, { "epoch": 0.06, "grad_norm": 2.8946528694898412, "learning_rate": 1.848928974069899e-05, "loss": 2.7645, "step": 2460 }, { "epoch": 0.06, "grad_norm": 5.0739730468430295, "learning_rate": 1.856444945509207e-05, "loss": 2.8731, "step": 2470 }, { "epoch": 0.06, "grad_norm": 4.258290301899105, "learning_rate": 1.863960916948516e-05, "loss": 2.7591, "step": 2480 }, { "epoch": 0.06, "grad_norm": 2.5170263078734405, "learning_rate": 1.8714768883878242e-05, "loss": 2.8494, "step": 2490 }, { "epoch": 0.06, "grad_norm": 3.127491689479058, "learning_rate": 1.878992859827133e-05, "loss": 2.6618, "step": 2500 }, { "epoch": 0.06, "grad_norm": 6.1532565880792385, "learning_rate": 1.8865088312664412e-05, "loss": 2.7972, "step": 2510 }, { "epoch": 0.06, "grad_norm": 4.526456013755865, "learning_rate": 1.89402480270575e-05, "loss": 2.8431, "step": 2520 }, { "epoch": 0.06, "grad_norm": 2.902329768596828, "learning_rate": 1.9015407741450585e-05, "loss": 2.7147, "step": 2530 }, { "epoch": 0.06, "grad_norm": 3.1165905582934315, "learning_rate": 1.909056745584367e-05, "loss": 2.769, "step": 2540 }, { "epoch": 0.06, "grad_norm": 4.125665501333099, "learning_rate": 1.9165727170236755e-05, "loss": 2.8083, "step": 2550 }, { "epoch": 0.06, "grad_norm": 2.1813597628611383, "learning_rate": 1.924088688462984e-05, "loss": 2.8354, "step": 2560 }, { "epoch": 0.06, "grad_norm": 3.1807718396442306, "learning_rate": 1.9316046599022925e-05, "loss": 2.6271, "step": 2570 }, { "epoch": 0.06, "grad_norm": 2.142136265106513, "learning_rate": 1.9391206313416012e-05, "loss": 2.7201, "step": 2580 }, { "epoch": 0.06, "grad_norm": 1.9521613733726517, "learning_rate": 1.9466366027809095e-05, "loss": 2.7863, "step": 2590 }, { "epoch": 0.06, "grad_norm": 2.783840238239663, "learning_rate": 1.9541525742202182e-05, "loss": 2.8285, "step": 2600 }, { "epoch": 0.06, "grad_norm": 4.9605175886968045, "learning_rate": 1.9616685456595265e-05, "loss": 2.8407, "step": 2610 }, { "epoch": 0.06, "grad_norm": 2.629376492254008, "learning_rate": 1.9691845170988352e-05, "loss": 2.8104, "step": 2620 }, { "epoch": 0.06, "grad_norm": 3.4850231233314215, "learning_rate": 1.976700488538144e-05, "loss": 2.6677, "step": 2630 }, { "epoch": 0.06, "grad_norm": 3.641122644981293, "learning_rate": 1.9842164599774522e-05, "loss": 2.8034, "step": 2640 }, { "epoch": 0.06, "grad_norm": 2.8105126365771693, "learning_rate": 1.991732431416761e-05, "loss": 2.8789, "step": 2650 }, { "epoch": 0.06, "grad_norm": 3.104613397587185, "learning_rate": 1.9992484028560692e-05, "loss": 2.6908, "step": 2660 }, { "epoch": 0.06, "grad_norm": 4.288765788049623, "learning_rate": 1.9999999459886285e-05, "loss": 2.7413, "step": 2670 }, { "epoch": 0.06, "grad_norm": 4.805359214827431, "learning_rate": 1.999999759282659e-05, "loss": 2.7447, "step": 2680 }, { "epoch": 0.06, "grad_norm": 2.9200282108911884, "learning_rate": 1.99999943921531e-05, "loss": 2.9021, "step": 2690 }, { "epoch": 0.06, "grad_norm": 2.3181394641341013, "learning_rate": 1.999998985786624e-05, "loss": 2.9005, "step": 2700 }, { "epoch": 0.06, "grad_norm": 3.4591495812183184, "learning_rate": 1.999998398996661e-05, "loss": 2.7513, "step": 2710 }, { "epoch": 0.06, "grad_norm": 3.4862128969019186, "learning_rate": 1.9999976788455004e-05, "loss": 2.8001, "step": 2720 }, { "epoch": 0.06, "grad_norm": 2.9126057091828264, "learning_rate": 1.9999968253332373e-05, "loss": 2.8018, "step": 2730 }, { "epoch": 0.06, "grad_norm": 5.1472181487226925, "learning_rate": 1.9999958384599857e-05, "loss": 2.7213, "step": 2740 }, { "epoch": 0.06, "grad_norm": 2.528601366763359, "learning_rate": 1.9999947182258773e-05, "loss": 2.7772, "step": 2750 }, { "epoch": 0.06, "grad_norm": 7.506345765857245, "learning_rate": 1.9999934646310616e-05, "loss": 2.7628, "step": 2760 }, { "epoch": 0.06, "grad_norm": 4.7954161374246285, "learning_rate": 1.999992077675706e-05, "loss": 2.7257, "step": 2770 }, { "epoch": 0.06, "grad_norm": 3.6233984389511815, "learning_rate": 1.9999905573599946e-05, "loss": 2.8035, "step": 2780 }, { "epoch": 0.06, "grad_norm": 2.3767621362386815, "learning_rate": 1.999988903684131e-05, "loss": 2.7578, "step": 2790 }, { "epoch": 0.06, "grad_norm": 2.399573705897267, "learning_rate": 1.9999871166483355e-05, "loss": 2.8445, "step": 2800 }, { "epoch": 0.06, "grad_norm": 2.5449198137630993, "learning_rate": 1.9999851962528462e-05, "loss": 2.7496, "step": 2810 }, { "epoch": 0.06, "grad_norm": 2.847899140938146, "learning_rate": 1.9999831424979197e-05, "loss": 2.875, "step": 2820 }, { "epoch": 0.06, "grad_norm": 2.842369425182328, "learning_rate": 1.9999809553838292e-05, "loss": 2.7623, "step": 2830 }, { "epoch": 0.06, "grad_norm": 1.828629543102104, "learning_rate": 1.9999786349108667e-05, "loss": 2.7117, "step": 2840 }, { "epoch": 0.06, "grad_norm": 1.9641981849721524, "learning_rate": 1.999976181079342e-05, "loss": 2.7977, "step": 2850 }, { "epoch": 0.06, "grad_norm": 3.0203335608467334, "learning_rate": 1.999973593889582e-05, "loss": 2.7908, "step": 2860 }, { "epoch": 0.06, "grad_norm": 2.8932678995909353, "learning_rate": 1.9999708733419316e-05, "loss": 2.8569, "step": 2870 }, { "epoch": 0.06, "grad_norm": 5.611483449345251, "learning_rate": 1.9999680194367535e-05, "loss": 2.8757, "step": 2880 }, { "epoch": 0.07, "grad_norm": 2.8127592130799774, "learning_rate": 1.999965032174429e-05, "loss": 2.773, "step": 2890 }, { "epoch": 0.07, "grad_norm": 4.580143721966985, "learning_rate": 1.9999619115553562e-05, "loss": 2.8644, "step": 2900 }, { "epoch": 0.07, "grad_norm": 2.7864967167472443, "learning_rate": 1.999958657579951e-05, "loss": 2.8086, "step": 2910 }, { "epoch": 0.07, "grad_norm": 4.395367613361058, "learning_rate": 1.9999552702486476e-05, "loss": 2.6765, "step": 2920 }, { "epoch": 0.07, "grad_norm": 2.943975365484581, "learning_rate": 1.9999517495618975e-05, "loss": 2.8768, "step": 2930 }, { "epoch": 0.07, "grad_norm": 2.245521730519683, "learning_rate": 1.9999480955201703e-05, "loss": 2.8719, "step": 2940 }, { "epoch": 0.07, "grad_norm": 2.141401355780759, "learning_rate": 1.9999443081239534e-05, "loss": 2.7603, "step": 2950 }, { "epoch": 0.07, "grad_norm": 4.075832900012646, "learning_rate": 1.999940387373752e-05, "loss": 2.7317, "step": 2960 }, { "epoch": 0.07, "grad_norm": 2.830657773480418, "learning_rate": 1.9999363332700887e-05, "loss": 2.8149, "step": 2970 }, { "epoch": 0.07, "grad_norm": 2.8410824362621803, "learning_rate": 1.9999321458135043e-05, "loss": 2.8591, "step": 2980 }, { "epoch": 0.07, "grad_norm": 2.105415489380484, "learning_rate": 1.9999278250045573e-05, "loss": 2.7353, "step": 2990 }, { "epoch": 0.07, "grad_norm": 3.6101983770287887, "learning_rate": 1.9999233708438242e-05, "loss": 2.9135, "step": 3000 }, { "epoch": 0.07, "grad_norm": 2.3367392151989654, "learning_rate": 1.9999187833318983e-05, "loss": 2.7239, "step": 3010 }, { "epoch": 0.07, "grad_norm": 2.3734151583788408, "learning_rate": 1.999914062469392e-05, "loss": 2.7851, "step": 3020 }, { "epoch": 0.07, "grad_norm": 2.470793925582182, "learning_rate": 1.9999092082569342e-05, "loss": 2.7657, "step": 3030 }, { "epoch": 0.07, "grad_norm": 7.86800738974802, "learning_rate": 1.9999042206951732e-05, "loss": 2.8695, "step": 3040 }, { "epoch": 0.07, "grad_norm": 3.2097185599393017, "learning_rate": 1.9998990997847735e-05, "loss": 2.8633, "step": 3050 }, { "epoch": 0.07, "grad_norm": 2.6802291911467635, "learning_rate": 1.9998938455264187e-05, "loss": 2.823, "step": 3060 }, { "epoch": 0.07, "grad_norm": 2.600479733149364, "learning_rate": 1.9998884579208084e-05, "loss": 2.8059, "step": 3070 }, { "epoch": 0.07, "grad_norm": 2.4581650153009296, "learning_rate": 1.999882936968662e-05, "loss": 2.8985, "step": 3080 }, { "epoch": 0.07, "grad_norm": 3.9633486635018245, "learning_rate": 1.9998772826707154e-05, "loss": 2.7108, "step": 3090 }, { "epoch": 0.07, "grad_norm": 2.3070067590894556, "learning_rate": 1.9998714950277228e-05, "loss": 2.8327, "step": 3100 }, { "epoch": 0.07, "grad_norm": 2.684839158448384, "learning_rate": 1.999865574040456e-05, "loss": 2.8547, "step": 3110 }, { "epoch": 0.07, "grad_norm": 2.299232154115436, "learning_rate": 1.9998595197097047e-05, "loss": 2.7893, "step": 3120 }, { "epoch": 0.07, "grad_norm": 2.5145589903884984, "learning_rate": 1.9998533320362764e-05, "loss": 2.7539, "step": 3130 }, { "epoch": 0.07, "grad_norm": 7.385454067106267, "learning_rate": 1.9998470110209964e-05, "loss": 2.7707, "step": 3140 }, { "epoch": 0.07, "grad_norm": 2.348121370061157, "learning_rate": 1.9998405566647066e-05, "loss": 2.8139, "step": 3150 }, { "epoch": 0.07, "grad_norm": 3.7378268451875925, "learning_rate": 1.9998339689682694e-05, "loss": 2.8468, "step": 3160 }, { "epoch": 0.07, "grad_norm": 2.6029048413998135, "learning_rate": 1.9998272479325623e-05, "loss": 2.8102, "step": 3170 }, { "epoch": 0.07, "grad_norm": 2.6788283818254395, "learning_rate": 1.9998203935584817e-05, "loss": 2.6873, "step": 3180 }, { "epoch": 0.07, "grad_norm": 3.0422830506819487, "learning_rate": 1.999813405846942e-05, "loss": 2.7436, "step": 3190 }, { "epoch": 0.07, "grad_norm": 3.516401019264344, "learning_rate": 1.999806284798875e-05, "loss": 2.8206, "step": 3200 }, { "epoch": 0.07, "grad_norm": 3.8132853584454773, "learning_rate": 1.9997990304152303e-05, "loss": 2.8768, "step": 3210 }, { "epoch": 0.07, "grad_norm": 3.1251527992717394, "learning_rate": 1.9997916426969757e-05, "loss": 2.7891, "step": 3220 }, { "epoch": 0.07, "grad_norm": 2.3048219341941367, "learning_rate": 1.999784121645096e-05, "loss": 2.853, "step": 3230 }, { "epoch": 0.07, "grad_norm": 2.7864087286569386, "learning_rate": 1.999776467260594e-05, "loss": 2.8975, "step": 3240 }, { "epoch": 0.07, "grad_norm": 5.453680898044389, "learning_rate": 1.9997686795444912e-05, "loss": 2.8872, "step": 3250 }, { "epoch": 0.07, "grad_norm": 2.8430296416335183, "learning_rate": 1.999760758497826e-05, "loss": 2.78, "step": 3260 }, { "epoch": 0.07, "grad_norm": 2.295578244550342, "learning_rate": 1.9997527041216542e-05, "loss": 2.9183, "step": 3270 }, { "epoch": 0.07, "grad_norm": 2.2661790084286944, "learning_rate": 1.9997445164170505e-05, "loss": 2.795, "step": 3280 }, { "epoch": 0.07, "grad_norm": 2.5115802068841364, "learning_rate": 1.999736195385107e-05, "loss": 2.7682, "step": 3290 }, { "epoch": 0.07, "grad_norm": 3.68544486484894, "learning_rate": 1.9997277410269325e-05, "loss": 2.7555, "step": 3300 }, { "epoch": 0.07, "grad_norm": 3.3994831046680316, "learning_rate": 1.9997191533436557e-05, "loss": 2.8599, "step": 3310 }, { "epoch": 0.07, "grad_norm": 2.644407302163393, "learning_rate": 1.9997104323364206e-05, "loss": 2.8076, "step": 3320 }, { "epoch": 0.08, "grad_norm": 2.265976750340429, "learning_rate": 1.9997015780063913e-05, "loss": 2.8112, "step": 3330 }, { "epoch": 0.08, "grad_norm": 2.6313337197075413, "learning_rate": 1.999692590354748e-05, "loss": 2.756, "step": 3340 }, { "epoch": 0.08, "grad_norm": 2.158640700606804, "learning_rate": 1.9996834693826894e-05, "loss": 2.7526, "step": 3350 }, { "epoch": 0.08, "grad_norm": 3.179790461612602, "learning_rate": 1.999674215091432e-05, "loss": 2.7423, "step": 3360 }, { "epoch": 0.08, "grad_norm": 4.349706822872941, "learning_rate": 1.9996648274822102e-05, "loss": 2.772, "step": 3370 }, { "epoch": 0.08, "grad_norm": 2.2271099941355357, "learning_rate": 1.9996553065562758e-05, "loss": 2.8362, "step": 3380 }, { "epoch": 0.08, "grad_norm": 3.772870953956124, "learning_rate": 1.9996456523148982e-05, "loss": 2.7726, "step": 3390 }, { "epoch": 0.08, "grad_norm": 2.499557308769303, "learning_rate": 1.999635864759365e-05, "loss": 2.7707, "step": 3400 }, { "epoch": 0.08, "grad_norm": 6.094012045062048, "learning_rate": 1.999625943890982e-05, "loss": 2.6944, "step": 3410 }, { "epoch": 0.08, "grad_norm": 2.4138189167281987, "learning_rate": 1.9996158897110716e-05, "loss": 2.8283, "step": 3420 }, { "epoch": 0.08, "grad_norm": 3.3282110261338116, "learning_rate": 1.9996057022209748e-05, "loss": 2.7832, "step": 3430 }, { "epoch": 0.08, "grad_norm": 2.870370716950427, "learning_rate": 1.9995953814220505e-05, "loss": 2.8192, "step": 3440 }, { "epoch": 0.08, "grad_norm": 4.526502056502257, "learning_rate": 1.9995849273156748e-05, "loss": 2.7363, "step": 3450 }, { "epoch": 0.08, "grad_norm": 4.853427121408125, "learning_rate": 1.999574339903242e-05, "loss": 2.6914, "step": 3460 }, { "epoch": 0.08, "grad_norm": 4.727059937766468, "learning_rate": 1.9995636191861644e-05, "loss": 2.8407, "step": 3470 }, { "epoch": 0.08, "grad_norm": 4.1624548453580115, "learning_rate": 1.9995527651658712e-05, "loss": 2.8077, "step": 3480 }, { "epoch": 0.08, "grad_norm": 3.828526953243561, "learning_rate": 1.9995417778438096e-05, "loss": 2.7901, "step": 3490 }, { "epoch": 0.08, "grad_norm": 5.472523502005989, "learning_rate": 1.999530657221446e-05, "loss": 2.7978, "step": 3500 }, { "epoch": 0.08, "grad_norm": 2.3307243801863993, "learning_rate": 1.9995194033002628e-05, "loss": 2.8725, "step": 3510 }, { "epoch": 0.08, "grad_norm": 2.611526035704928, "learning_rate": 1.9995080160817607e-05, "loss": 2.7103, "step": 3520 }, { "epoch": 0.08, "grad_norm": 5.016295013664585, "learning_rate": 1.9994964955674585e-05, "loss": 2.8064, "step": 3530 }, { "epoch": 0.08, "grad_norm": 2.7475073189617807, "learning_rate": 1.9994848417588922e-05, "loss": 2.7321, "step": 3540 }, { "epoch": 0.08, "grad_norm": 2.388982497651055, "learning_rate": 1.999473054657617e-05, "loss": 2.8319, "step": 3550 }, { "epoch": 0.08, "grad_norm": 2.1534968661847023, "learning_rate": 1.999461134265204e-05, "loss": 2.7642, "step": 3560 }, { "epoch": 0.08, "grad_norm": 2.256234967258948, "learning_rate": 1.999449080583243e-05, "loss": 2.6581, "step": 3570 }, { "epoch": 0.08, "grad_norm": 2.201174221456859, "learning_rate": 1.9994368936133416e-05, "loss": 2.8235, "step": 3580 }, { "epoch": 0.08, "grad_norm": 3.1741377877046317, "learning_rate": 1.999424573357125e-05, "loss": 2.8669, "step": 3590 }, { "epoch": 0.08, "grad_norm": 2.4211973889222307, "learning_rate": 1.9994121198162365e-05, "loss": 2.7387, "step": 3600 }, { "epoch": 0.08, "grad_norm": 2.922248601093943, "learning_rate": 1.9993995329923367e-05, "loss": 2.6548, "step": 3610 }, { "epoch": 0.08, "grad_norm": 2.3147563336791763, "learning_rate": 1.9993868128871043e-05, "loss": 2.7526, "step": 3620 }, { "epoch": 0.08, "grad_norm": 2.3590710798292185, "learning_rate": 1.9993739595022354e-05, "loss": 2.8058, "step": 3630 }, { "epoch": 0.08, "grad_norm": 2.22227164151924, "learning_rate": 1.999360972839445e-05, "loss": 2.8238, "step": 3640 }, { "epoch": 0.08, "grad_norm": 3.0644796731551223, "learning_rate": 1.9993478529004636e-05, "loss": 2.7644, "step": 3650 }, { "epoch": 0.08, "grad_norm": 2.19161859463938, "learning_rate": 1.9993345996870423e-05, "loss": 2.7665, "step": 3660 }, { "epoch": 0.08, "grad_norm": 2.4712388908080096, "learning_rate": 1.9993212132009475e-05, "loss": 2.8085, "step": 3670 }, { "epoch": 0.08, "grad_norm": 3.5396140895826034, "learning_rate": 1.999307693443965e-05, "loss": 2.7264, "step": 3680 }, { "epoch": 0.08, "grad_norm": 2.863085626066982, "learning_rate": 1.9992940404178975e-05, "loss": 2.8313, "step": 3690 }, { "epoch": 0.08, "grad_norm": 3.137184554472705, "learning_rate": 1.9992802541245662e-05, "loss": 2.8496, "step": 3700 }, { "epoch": 0.08, "grad_norm": 2.343967645953022, "learning_rate": 1.9992663345658098e-05, "loss": 2.732, "step": 3710 }, { "epoch": 0.08, "grad_norm": 3.3663363239021398, "learning_rate": 1.999252281743484e-05, "loss": 2.8152, "step": 3720 }, { "epoch": 0.08, "grad_norm": 2.6805740908354836, "learning_rate": 1.9992380956594627e-05, "loss": 2.7629, "step": 3730 }, { "epoch": 0.08, "grad_norm": 2.5102586157990774, "learning_rate": 1.9992237763156386e-05, "loss": 2.7303, "step": 3740 }, { "epoch": 0.08, "grad_norm": 2.4798099393259574, "learning_rate": 1.9992093237139215e-05, "loss": 2.8282, "step": 3750 }, { "epoch": 0.08, "grad_norm": 3.4776995009061644, "learning_rate": 1.9991947378562378e-05, "loss": 2.8168, "step": 3760 }, { "epoch": 0.09, "grad_norm": 3.2432945863644376, "learning_rate": 1.9991800187445332e-05, "loss": 2.7464, "step": 3770 }, { "epoch": 0.09, "grad_norm": 3.0314175455157244, "learning_rate": 1.999165166380771e-05, "loss": 2.7744, "step": 3780 }, { "epoch": 0.09, "grad_norm": 3.0318531419576704, "learning_rate": 1.9991501807669316e-05, "loss": 2.7786, "step": 3790 }, { "epoch": 0.09, "grad_norm": 2.345413703194256, "learning_rate": 1.9991350619050133e-05, "loss": 2.7839, "step": 3800 }, { "epoch": 0.09, "grad_norm": 2.2122701549137855, "learning_rate": 1.9991198097970325e-05, "loss": 2.836, "step": 3810 }, { "epoch": 0.09, "grad_norm": 5.51776410209793, "learning_rate": 1.999104424445023e-05, "loss": 2.7363, "step": 3820 }, { "epoch": 0.09, "grad_norm": 2.4628275131984423, "learning_rate": 1.9990889058510378e-05, "loss": 2.7602, "step": 3830 }, { "epoch": 0.09, "grad_norm": 3.129926738751418, "learning_rate": 1.9990732540171453e-05, "loss": 2.7629, "step": 3840 }, { "epoch": 0.09, "grad_norm": 2.954591082583205, "learning_rate": 1.999057468945433e-05, "loss": 2.8226, "step": 3850 }, { "epoch": 0.09, "grad_norm": 4.344125110460846, "learning_rate": 1.9990415506380062e-05, "loss": 2.7674, "step": 3860 }, { "epoch": 0.09, "grad_norm": 2.02130471067519, "learning_rate": 1.999025499096988e-05, "loss": 2.7617, "step": 3870 }, { "epoch": 0.09, "grad_norm": 2.9311597979498982, "learning_rate": 1.9990093143245185e-05, "loss": 2.8998, "step": 3880 }, { "epoch": 0.09, "grad_norm": 2.267112788369172, "learning_rate": 1.9989929963227567e-05, "loss": 2.8643, "step": 3890 }, { "epoch": 0.09, "grad_norm": 2.1060075368503433, "learning_rate": 1.9989765450938786e-05, "loss": 2.8062, "step": 3900 }, { "epoch": 0.09, "grad_norm": 2.591512660903145, "learning_rate": 1.998959960640078e-05, "loss": 2.7377, "step": 3910 }, { "epoch": 0.09, "grad_norm": 3.1078510812959323, "learning_rate": 1.998943242963567e-05, "loss": 2.8264, "step": 3920 }, { "epoch": 0.09, "grad_norm": 3.8219858979871453, "learning_rate": 1.9989263920665746e-05, "loss": 2.923, "step": 3930 }, { "epoch": 0.09, "grad_norm": 4.621751906064776, "learning_rate": 1.998909407951348e-05, "loss": 2.7452, "step": 3940 }, { "epoch": 0.09, "grad_norm": 3.7086506974501994, "learning_rate": 1.998892290620153e-05, "loss": 2.7816, "step": 3950 }, { "epoch": 0.09, "grad_norm": 2.4473165234263607, "learning_rate": 1.9988750400752718e-05, "loss": 2.6906, "step": 3960 }, { "epoch": 0.09, "grad_norm": 2.326538202228531, "learning_rate": 1.9988576563190046e-05, "loss": 2.7842, "step": 3970 }, { "epoch": 0.09, "grad_norm": 2.642820196142749, "learning_rate": 1.9988401393536707e-05, "loss": 2.7951, "step": 3980 }, { "epoch": 0.09, "grad_norm": 2.970629486360716, "learning_rate": 1.9988224891816055e-05, "loss": 2.7482, "step": 3990 }, { "epoch": 0.09, "grad_norm": 2.606652513118783, "learning_rate": 1.998804705805163e-05, "loss": 2.7056, "step": 4000 }, { "epoch": 0.09, "grad_norm": 2.3275390311226607, "learning_rate": 1.998786789226715e-05, "loss": 2.8103, "step": 4010 }, { "epoch": 0.09, "grad_norm": 3.1300631644041093, "learning_rate": 1.9987687394486506e-05, "loss": 2.807, "step": 4020 }, { "epoch": 0.09, "grad_norm": 2.4182285902008345, "learning_rate": 1.998750556473377e-05, "loss": 2.7364, "step": 4030 }, { "epoch": 0.09, "grad_norm": 4.103393472633442, "learning_rate": 1.9987322403033192e-05, "loss": 2.8541, "step": 4040 }, { "epoch": 0.09, "grad_norm": 2.343070418863726, "learning_rate": 1.99871379094092e-05, "loss": 2.7217, "step": 4050 }, { "epoch": 0.09, "grad_norm": 2.664782203919707, "learning_rate": 1.9986952083886395e-05, "loss": 2.9016, "step": 4060 }, { "epoch": 0.09, "grad_norm": 2.2658124582371886, "learning_rate": 1.9986764926489562e-05, "loss": 2.885, "step": 4070 }, { "epoch": 0.09, "grad_norm": 2.5318463708896957, "learning_rate": 1.998657643724366e-05, "loss": 2.8839, "step": 4080 }, { "epoch": 0.09, "grad_norm": 3.289113459089288, "learning_rate": 1.9986386616173823e-05, "loss": 2.8796, "step": 4090 }, { "epoch": 0.09, "grad_norm": 1.9611542477563804, "learning_rate": 1.998619546330537e-05, "loss": 2.8441, "step": 4100 }, { "epoch": 0.09, "grad_norm": 2.513400873695121, "learning_rate": 1.998600297866379e-05, "loss": 2.708, "step": 4110 }, { "epoch": 0.09, "grad_norm": 1.9155315853508317, "learning_rate": 1.9985809162274757e-05, "loss": 2.8666, "step": 4120 }, { "epoch": 0.09, "grad_norm": 2.486462256385746, "learning_rate": 1.9985614014164114e-05, "loss": 2.8315, "step": 4130 }, { "epoch": 0.09, "grad_norm": 5.219549535801043, "learning_rate": 1.998541753435789e-05, "loss": 2.7969, "step": 4140 }, { "epoch": 0.09, "grad_norm": 3.1587348739359262, "learning_rate": 1.9985219722882286e-05, "loss": 2.7957, "step": 4150 }, { "epoch": 0.09, "grad_norm": 2.2614701096066567, "learning_rate": 1.9985020579763682e-05, "loss": 2.686, "step": 4160 }, { "epoch": 0.09, "grad_norm": 3.271343092419369, "learning_rate": 1.998482010502864e-05, "loss": 2.8058, "step": 4170 }, { "epoch": 0.09, "grad_norm": 2.5701200413737357, "learning_rate": 1.998461829870389e-05, "loss": 2.7491, "step": 4180 }, { "epoch": 0.09, "grad_norm": 2.726394254067148, "learning_rate": 1.9984415160816346e-05, "loss": 2.791, "step": 4190 }, { "epoch": 0.09, "grad_norm": 2.0610826012832932, "learning_rate": 1.9984210691393105e-05, "loss": 2.9968, "step": 4200 }, { "epoch": 0.09, "grad_norm": 2.6247099594724026, "learning_rate": 1.9984004890461426e-05, "loss": 2.7999, "step": 4210 }, { "epoch": 0.1, "grad_norm": 8.160301278003612, "learning_rate": 1.9983797758048763e-05, "loss": 2.8512, "step": 4220 }, { "epoch": 0.1, "grad_norm": 2.214089162101829, "learning_rate": 1.9983589294182735e-05, "loss": 2.8667, "step": 4230 }, { "epoch": 0.1, "grad_norm": 3.068676579846673, "learning_rate": 1.9983379498891145e-05, "loss": 2.7433, "step": 4240 }, { "epoch": 0.1, "grad_norm": 4.614860007873676, "learning_rate": 1.998316837220197e-05, "loss": 2.6846, "step": 4250 }, { "epoch": 0.1, "grad_norm": 6.24732736512667, "learning_rate": 1.9982955914143366e-05, "loss": 2.744, "step": 4260 }, { "epoch": 0.1, "grad_norm": 2.3715840333016205, "learning_rate": 1.9982742124743666e-05, "loss": 2.7213, "step": 4270 }, { "epoch": 0.1, "grad_norm": 2.6568955147401048, "learning_rate": 1.9982527004031383e-05, "loss": 2.7786, "step": 4280 }, { "epoch": 0.1, "grad_norm": 2.2503174073248524, "learning_rate": 1.9982310552035208e-05, "loss": 2.7536, "step": 4290 }, { "epoch": 0.1, "grad_norm": 2.1809577727093146, "learning_rate": 1.9982092768784004e-05, "loss": 2.8158, "step": 4300 }, { "epoch": 0.1, "grad_norm": 2.4175779691322816, "learning_rate": 1.9981873654306814e-05, "loss": 2.7134, "step": 4310 }, { "epoch": 0.1, "grad_norm": 5.4203809110354575, "learning_rate": 1.9981653208632865e-05, "loss": 2.7311, "step": 4320 }, { "epoch": 0.1, "grad_norm": 2.3527729978888545, "learning_rate": 1.998143143179155e-05, "loss": 2.7609, "step": 4330 }, { "epoch": 0.1, "grad_norm": 2.702893185876208, "learning_rate": 1.9981208323812452e-05, "loss": 2.7458, "step": 4340 }, { "epoch": 0.1, "grad_norm": 2.1349670268690897, "learning_rate": 1.9980983884725315e-05, "loss": 2.6925, "step": 4350 }, { "epoch": 0.1, "grad_norm": 3.1992721494976304, "learning_rate": 1.9980758114560076e-05, "loss": 2.8286, "step": 4360 }, { "epoch": 0.1, "grad_norm": 2.062120080334543, "learning_rate": 1.9980531013346848e-05, "loss": 2.7298, "step": 4370 }, { "epoch": 0.1, "grad_norm": 1.99344591417991, "learning_rate": 1.998030258111591e-05, "loss": 2.7085, "step": 4380 }, { "epoch": 0.1, "grad_norm": 2.2353178409907213, "learning_rate": 1.9980072817897733e-05, "loss": 2.7181, "step": 4390 }, { "epoch": 0.1, "grad_norm": 3.0051619038680224, "learning_rate": 1.9979841723722955e-05, "loss": 2.8256, "step": 4400 }, { "epoch": 0.1, "grad_norm": 2.3443358473982028, "learning_rate": 1.9979609298622393e-05, "loss": 2.7479, "step": 4410 }, { "epoch": 0.1, "grad_norm": 2.530767716139881, "learning_rate": 1.997937554262705e-05, "loss": 2.6231, "step": 4420 }, { "epoch": 0.1, "grad_norm": 2.3961284816150537, "learning_rate": 1.9979140455768093e-05, "loss": 2.7826, "step": 4430 }, { "epoch": 0.1, "grad_norm": 1.9471764663507463, "learning_rate": 1.9978904038076872e-05, "loss": 2.8189, "step": 4440 }, { "epoch": 0.1, "grad_norm": 4.088988091849032, "learning_rate": 1.9978666289584928e-05, "loss": 2.707, "step": 4450 }, { "epoch": 0.1, "grad_norm": 2.544305413635871, "learning_rate": 1.9978427210323958e-05, "loss": 2.6677, "step": 4460 }, { "epoch": 0.1, "grad_norm": 3.627514435741834, "learning_rate": 1.9978186800325846e-05, "loss": 2.826, "step": 4470 }, { "epoch": 0.1, "grad_norm": 2.263030665226886, "learning_rate": 1.997794505962266e-05, "loss": 2.8683, "step": 4480 }, { "epoch": 0.1, "grad_norm": 2.2855402745371918, "learning_rate": 1.9977701988246627e-05, "loss": 2.7382, "step": 4490 }, { "epoch": 0.1, "grad_norm": 4.1331051801607295, "learning_rate": 1.9977457586230177e-05, "loss": 2.8904, "step": 4500 }, { "epoch": 0.1, "grad_norm": 3.8313481768532047, "learning_rate": 1.997721185360589e-05, "loss": 2.8074, "step": 4510 }, { "epoch": 0.1, "grad_norm": 2.627614870198135, "learning_rate": 1.9976964790406552e-05, "loss": 2.8925, "step": 4520 }, { "epoch": 0.1, "grad_norm": 2.4106278912094337, "learning_rate": 1.99767163966651e-05, "loss": 2.788, "step": 4530 }, { "epoch": 0.1, "grad_norm": 2.6111692640970325, "learning_rate": 1.9976466672414667e-05, "loss": 2.8106, "step": 4540 }, { "epoch": 0.1, "grad_norm": 2.1783777555415234, "learning_rate": 1.9976215617688553e-05, "loss": 2.8998, "step": 4550 }, { "epoch": 0.1, "grad_norm": 2.381482266374249, "learning_rate": 1.997596323252024e-05, "loss": 2.7829, "step": 4560 }, { "epoch": 0.1, "grad_norm": 2.1145197122407637, "learning_rate": 1.9975709516943388e-05, "loss": 2.7585, "step": 4570 }, { "epoch": 0.1, "grad_norm": 3.062678132666394, "learning_rate": 1.997545447099183e-05, "loss": 2.7107, "step": 4580 }, { "epoch": 0.1, "grad_norm": 2.871004777902555, "learning_rate": 1.9975198094699577e-05, "loss": 2.9109, "step": 4590 }, { "epoch": 0.1, "grad_norm": 2.9891506578012996, "learning_rate": 1.9974940388100827e-05, "loss": 2.6957, "step": 4600 }, { "epoch": 0.1, "grad_norm": 2.6012034224505425, "learning_rate": 1.997468135122994e-05, "loss": 2.7246, "step": 4610 }, { "epoch": 0.1, "grad_norm": 2.980657445206224, "learning_rate": 1.9974420984121473e-05, "loss": 2.823, "step": 4620 }, { "epoch": 0.1, "grad_norm": 2.365359130974841, "learning_rate": 1.9974159286810137e-05, "loss": 2.6795, "step": 4630 }, { "epoch": 0.1, "grad_norm": 2.2936721377250233, "learning_rate": 1.997389625933084e-05, "loss": 2.7512, "step": 4640 }, { "epoch": 0.1, "grad_norm": 2.9144035410387246, "learning_rate": 1.9973631901718654e-05, "loss": 2.7502, "step": 4650 }, { "epoch": 0.11, "grad_norm": 1.9694373114442558, "learning_rate": 1.9973366214008842e-05, "loss": 2.8195, "step": 4660 }, { "epoch": 0.11, "grad_norm": 2.0216045085625383, "learning_rate": 1.9973099196236826e-05, "loss": 2.7865, "step": 4670 }, { "epoch": 0.11, "grad_norm": 2.329080734449, "learning_rate": 1.9972830848438224e-05, "loss": 2.8209, "step": 4680 }, { "epoch": 0.11, "grad_norm": 2.5003268559023475, "learning_rate": 1.997256117064882e-05, "loss": 2.6985, "step": 4690 }, { "epoch": 0.11, "grad_norm": 2.135611176432461, "learning_rate": 1.997229016290458e-05, "loss": 2.8746, "step": 4700 }, { "epoch": 0.11, "grad_norm": 2.1627164119611324, "learning_rate": 1.9972017825241647e-05, "loss": 2.8414, "step": 4710 }, { "epoch": 0.11, "grad_norm": 2.4785514720496735, "learning_rate": 1.9971744157696333e-05, "loss": 2.8047, "step": 4720 }, { "epoch": 0.11, "grad_norm": 2.1461017089953467, "learning_rate": 1.9971469160305145e-05, "loss": 2.6634, "step": 4730 }, { "epoch": 0.11, "grad_norm": 4.207363570195497, "learning_rate": 1.9971192833104752e-05, "loss": 2.8201, "step": 4740 }, { "epoch": 0.11, "grad_norm": 2.641618557473545, "learning_rate": 1.9970915176132e-05, "loss": 2.8256, "step": 4750 }, { "epoch": 0.11, "grad_norm": 2.208252341714473, "learning_rate": 1.997063618942393e-05, "loss": 2.6899, "step": 4760 }, { "epoch": 0.11, "grad_norm": 2.2269865646308205, "learning_rate": 1.997035587301774e-05, "loss": 2.8002, "step": 4770 }, { "epoch": 0.11, "grad_norm": 2.6234271645320946, "learning_rate": 1.9970074226950812e-05, "loss": 2.7874, "step": 4780 }, { "epoch": 0.11, "grad_norm": 3.0705036565746187, "learning_rate": 1.9969791251260715e-05, "loss": 2.8146, "step": 4790 }, { "epoch": 0.11, "grad_norm": 2.1698583514542347, "learning_rate": 1.9969506945985175e-05, "loss": 2.8781, "step": 4800 }, { "epoch": 0.11, "grad_norm": 3.2878311537742664, "learning_rate": 1.996922131116212e-05, "loss": 2.9144, "step": 4810 }, { "epoch": 0.11, "grad_norm": 2.1361248490571456, "learning_rate": 1.9968934346829632e-05, "loss": 2.7352, "step": 4820 }, { "epoch": 0.11, "grad_norm": 2.6902848007832807, "learning_rate": 1.996864605302599e-05, "loss": 2.8302, "step": 4830 }, { "epoch": 0.11, "grad_norm": 2.453323597952177, "learning_rate": 1.996835642978963e-05, "loss": 2.8013, "step": 4840 }, { "epoch": 0.11, "grad_norm": 3.150985155391749, "learning_rate": 1.996806547715919e-05, "loss": 2.7441, "step": 4850 }, { "epoch": 0.11, "grad_norm": 2.49995521271767, "learning_rate": 1.9967773195173462e-05, "loss": 2.6781, "step": 4860 }, { "epoch": 0.11, "grad_norm": 3.0217981087670944, "learning_rate": 1.996747958387143e-05, "loss": 2.7349, "step": 4870 }, { "epoch": 0.11, "grad_norm": 2.051729842998712, "learning_rate": 1.9967184643292247e-05, "loss": 2.8983, "step": 4880 }, { "epoch": 0.11, "grad_norm": 2.219110466348218, "learning_rate": 1.9966888373475252e-05, "loss": 2.8073, "step": 4890 }, { "epoch": 0.11, "grad_norm": 3.49295998458521, "learning_rate": 1.996659077445995e-05, "loss": 2.7974, "step": 4900 }, { "epoch": 0.11, "grad_norm": 2.309137307595954, "learning_rate": 1.9966291846286028e-05, "loss": 2.86, "step": 4910 }, { "epoch": 0.11, "grad_norm": 3.314379976029997, "learning_rate": 1.996599158899336e-05, "loss": 2.807, "step": 4920 }, { "epoch": 0.11, "grad_norm": 2.443661297790473, "learning_rate": 1.9965690002621984e-05, "loss": 2.691, "step": 4930 }, { "epoch": 0.11, "grad_norm": 2.716691596137297, "learning_rate": 1.996538708721212e-05, "loss": 2.7329, "step": 4940 }, { "epoch": 0.11, "grad_norm": 2.5384696746398294, "learning_rate": 1.996508284280416e-05, "loss": 2.7428, "step": 4950 }, { "epoch": 0.11, "grad_norm": 2.783831789928233, "learning_rate": 1.996477726943869e-05, "loss": 2.8378, "step": 4960 }, { "epoch": 0.11, "grad_norm": 3.505701003937554, "learning_rate": 1.996447036715645e-05, "loss": 2.7307, "step": 4970 }, { "epoch": 0.11, "grad_norm": 2.6057758831639126, "learning_rate": 1.9964162135998374e-05, "loss": 2.8181, "step": 4980 }, { "epoch": 0.11, "grad_norm": 2.85902224170803, "learning_rate": 1.9963852576005575e-05, "loss": 2.7531, "step": 4990 }, { "epoch": 0.11, "grad_norm": 1.944353317802165, "learning_rate": 1.9963541687219323e-05, "loss": 2.8301, "step": 5000 }, { "epoch": 0.11, "grad_norm": 2.2758814299281838, "learning_rate": 1.996322946968109e-05, "loss": 2.8082, "step": 5010 }, { "epoch": 0.11, "grad_norm": 2.6302119520059706, "learning_rate": 1.9962915923432506e-05, "loss": 2.8235, "step": 5020 }, { "epoch": 0.11, "grad_norm": 2.7821872629571347, "learning_rate": 1.996260104851539e-05, "loss": 2.7058, "step": 5030 }, { "epoch": 0.11, "grad_norm": 2.6707604726319674, "learning_rate": 1.9962284844971733e-05, "loss": 2.8403, "step": 5040 }, { "epoch": 0.11, "grad_norm": 2.396619805724571, "learning_rate": 1.9961967312843704e-05, "loss": 2.8184, "step": 5050 }, { "epoch": 0.11, "grad_norm": 2.6287245270569173, "learning_rate": 1.996164845217365e-05, "loss": 2.7218, "step": 5060 }, { "epoch": 0.11, "grad_norm": 2.0380517821368884, "learning_rate": 1.9961328263004097e-05, "loss": 2.8331, "step": 5070 }, { "epoch": 0.11, "grad_norm": 2.1357475089681537, "learning_rate": 1.9961006745377742e-05, "loss": 2.7984, "step": 5080 }, { "epoch": 0.11, "grad_norm": 4.110547681971177, "learning_rate": 1.9960683899337466e-05, "loss": 2.6924, "step": 5090 }, { "epoch": 0.12, "grad_norm": 2.2309146965201077, "learning_rate": 1.996035972492632e-05, "loss": 2.7315, "step": 5100 }, { "epoch": 0.12, "grad_norm": 2.4109321879384047, "learning_rate": 1.9960034222187542e-05, "loss": 2.8681, "step": 5110 }, { "epoch": 0.12, "grad_norm": 2.3440413400830717, "learning_rate": 1.995970739116454e-05, "loss": 2.6563, "step": 5120 }, { "epoch": 0.12, "grad_norm": 2.7402513471031282, "learning_rate": 1.9959379231900894e-05, "loss": 2.7373, "step": 5130 }, { "epoch": 0.12, "grad_norm": 2.470622191220145, "learning_rate": 1.9959049744440378e-05, "loss": 2.6625, "step": 5140 }, { "epoch": 0.12, "grad_norm": 3.297959805879017, "learning_rate": 1.9958718928826926e-05, "loss": 2.8264, "step": 5150 }, { "epoch": 0.12, "grad_norm": 2.8579758244276317, "learning_rate": 1.9958386785104658e-05, "loss": 2.7606, "step": 5160 }, { "epoch": 0.12, "grad_norm": 2.4467732641171915, "learning_rate": 1.995805331331787e-05, "loss": 2.7733, "step": 5170 }, { "epoch": 0.12, "grad_norm": 2.616435507426269, "learning_rate": 1.9957718513511034e-05, "loss": 2.8355, "step": 5180 }, { "epoch": 0.12, "grad_norm": 1.961989225595134, "learning_rate": 1.99573823857288e-05, "loss": 2.8032, "step": 5190 }, { "epoch": 0.12, "grad_norm": 2.1963748874923934, "learning_rate": 1.995704493001599e-05, "loss": 2.7835, "step": 5200 }, { "epoch": 0.12, "grad_norm": 3.231042102970438, "learning_rate": 1.9956706146417612e-05, "loss": 2.6949, "step": 5210 }, { "epoch": 0.12, "grad_norm": 2.725070730852553, "learning_rate": 1.9956366034978847e-05, "loss": 2.7638, "step": 5220 }, { "epoch": 0.12, "grad_norm": 2.295579869772909, "learning_rate": 1.995602459574505e-05, "loss": 2.709, "step": 5230 }, { "epoch": 0.12, "grad_norm": 2.3136673235782106, "learning_rate": 1.9955681828761756e-05, "loss": 2.9157, "step": 5240 }, { "epoch": 0.12, "grad_norm": 2.0176731948130517, "learning_rate": 1.9955337734074683e-05, "loss": 2.7226, "step": 5250 }, { "epoch": 0.12, "grad_norm": 2.260587197181578, "learning_rate": 1.9954992311729713e-05, "loss": 2.8341, "step": 5260 }, { "epoch": 0.12, "grad_norm": 5.027497865243849, "learning_rate": 1.995464556177291e-05, "loss": 2.678, "step": 5270 }, { "epoch": 0.12, "grad_norm": 2.2703061959226076, "learning_rate": 1.9954297484250523e-05, "loss": 2.9094, "step": 5280 }, { "epoch": 0.12, "grad_norm": 2.6672898446249915, "learning_rate": 1.9953948079208975e-05, "loss": 2.8149, "step": 5290 }, { "epoch": 0.12, "grad_norm": 3.639007727980739, "learning_rate": 1.9953597346694854e-05, "loss": 2.7128, "step": 5300 }, { "epoch": 0.12, "grad_norm": 2.6751234481803103, "learning_rate": 1.9953245286754938e-05, "loss": 2.7705, "step": 5310 }, { "epoch": 0.12, "grad_norm": 2.532805593089366, "learning_rate": 1.995289189943618e-05, "loss": 2.8677, "step": 5320 }, { "epoch": 0.12, "grad_norm": 3.358987338864299, "learning_rate": 1.995253718478571e-05, "loss": 2.7586, "step": 5330 }, { "epoch": 0.12, "grad_norm": 3.402112107235774, "learning_rate": 1.9952181142850826e-05, "loss": 2.6912, "step": 5340 }, { "epoch": 0.12, "grad_norm": 3.143806013657559, "learning_rate": 1.9951823773679017e-05, "loss": 2.8119, "step": 5350 }, { "epoch": 0.12, "grad_norm": 2.534246618042724, "learning_rate": 1.9951465077317938e-05, "loss": 2.678, "step": 5360 }, { "epoch": 0.12, "grad_norm": 2.197574255075016, "learning_rate": 1.995110505381543e-05, "loss": 2.8334, "step": 5370 }, { "epoch": 0.12, "grad_norm": 2.055599555714795, "learning_rate": 1.99507437032195e-05, "loss": 2.764, "step": 5380 }, { "epoch": 0.12, "grad_norm": 2.4144719966934294, "learning_rate": 1.995038102557834e-05, "loss": 2.9053, "step": 5390 }, { "epoch": 0.12, "grad_norm": 1.9967983181374025, "learning_rate": 1.995001702094032e-05, "loss": 2.7855, "step": 5400 }, { "epoch": 0.12, "grad_norm": 3.3652493780071704, "learning_rate": 1.9949651689353985e-05, "loss": 2.9293, "step": 5410 }, { "epoch": 0.12, "grad_norm": 2.336991327185288, "learning_rate": 1.9949285030868054e-05, "loss": 2.9152, "step": 5420 }, { "epoch": 0.12, "grad_norm": 1.8303091327200283, "learning_rate": 1.9948917045531424e-05, "loss": 2.778, "step": 5430 }, { "epoch": 0.12, "grad_norm": 2.6409384932920283, "learning_rate": 1.994854773339317e-05, "loss": 2.8126, "step": 5440 }, { "epoch": 0.12, "grad_norm": 3.1361230420767265, "learning_rate": 1.9948177094502548e-05, "loss": 2.7575, "step": 5450 }, { "epoch": 0.12, "grad_norm": 2.6396159800419445, "learning_rate": 1.994780512890898e-05, "loss": 2.6984, "step": 5460 }, { "epoch": 0.12, "grad_norm": 3.045151623954712, "learning_rate": 1.994743183666208e-05, "loss": 2.6244, "step": 5470 }, { "epoch": 0.12, "grad_norm": 2.484620557130041, "learning_rate": 1.9947057217811625e-05, "loss": 2.7727, "step": 5480 }, { "epoch": 0.12, "grad_norm": 5.9769486758337855, "learning_rate": 1.9946681272407576e-05, "loss": 2.8151, "step": 5490 }, { "epoch": 0.12, "grad_norm": 2.0234067524922406, "learning_rate": 1.9946304000500068e-05, "loss": 2.8507, "step": 5500 }, { "epoch": 0.12, "grad_norm": 2.4771593162977217, "learning_rate": 1.9945925402139422e-05, "loss": 2.8573, "step": 5510 }, { "epoch": 0.12, "grad_norm": 2.0894413871761968, "learning_rate": 1.9945545477376118e-05, "loss": 2.7462, "step": 5520 }, { "epoch": 0.12, "grad_norm": 3.2947149345207762, "learning_rate": 1.994516422626083e-05, "loss": 2.6844, "step": 5530 }, { "epoch": 0.12, "grad_norm": 2.5187923759696655, "learning_rate": 1.99447816488444e-05, "loss": 2.7722, "step": 5540 }, { "epoch": 0.13, "grad_norm": 1.9830326942075478, "learning_rate": 1.994439774517785e-05, "loss": 2.8505, "step": 5550 }, { "epoch": 0.13, "grad_norm": 2.7417445543228762, "learning_rate": 1.9944012515312375e-05, "loss": 2.8577, "step": 5560 }, { "epoch": 0.13, "grad_norm": 2.0526732317481047, "learning_rate": 1.9943625959299356e-05, "loss": 2.764, "step": 5570 }, { "epoch": 0.13, "grad_norm": 2.085870150441627, "learning_rate": 1.9943238077190343e-05, "loss": 2.7256, "step": 5580 }, { "epoch": 0.13, "grad_norm": 2.7364667725484875, "learning_rate": 1.9942848869037057e-05, "loss": 2.7719, "step": 5590 }, { "epoch": 0.13, "grad_norm": 2.300649281095665, "learning_rate": 1.994245833489141e-05, "loss": 2.8006, "step": 5600 }, { "epoch": 0.13, "grad_norm": 3.267686194814767, "learning_rate": 1.994206647480548e-05, "loss": 2.8313, "step": 5610 }, { "epoch": 0.13, "grad_norm": 3.7300565501299108, "learning_rate": 1.9941673288831532e-05, "loss": 2.8215, "step": 5620 }, { "epoch": 0.13, "grad_norm": 3.497117333801037, "learning_rate": 1.9941278777022003e-05, "loss": 2.7387, "step": 5630 }, { "epoch": 0.13, "grad_norm": 2.0362572737488005, "learning_rate": 1.9940882939429495e-05, "loss": 2.8158, "step": 5640 }, { "epoch": 0.13, "grad_norm": 2.1751498918016554, "learning_rate": 1.9940485776106805e-05, "loss": 2.8001, "step": 5650 }, { "epoch": 0.13, "grad_norm": 2.4300257336774442, "learning_rate": 1.9940087287106903e-05, "loss": 2.7121, "step": 5660 }, { "epoch": 0.13, "grad_norm": 2.5295144364391278, "learning_rate": 1.993968747248292e-05, "loss": 2.8954, "step": 5670 }, { "epoch": 0.13, "grad_norm": 3.0410623049706964, "learning_rate": 1.993928633228819e-05, "loss": 2.7761, "step": 5680 }, { "epoch": 0.13, "grad_norm": 2.599192339813641, "learning_rate": 1.9938883866576196e-05, "loss": 2.8935, "step": 5690 }, { "epoch": 0.13, "grad_norm": 3.60999951618649, "learning_rate": 1.993848007540062e-05, "loss": 2.7967, "step": 5700 }, { "epoch": 0.13, "grad_norm": 3.230835600411114, "learning_rate": 1.9938074958815313e-05, "loss": 2.7996, "step": 5710 }, { "epoch": 0.13, "grad_norm": 1.9187204237470397, "learning_rate": 1.9937668516874297e-05, "loss": 2.7841, "step": 5720 }, { "epoch": 0.13, "grad_norm": 1.983001829173743, "learning_rate": 1.993726074963178e-05, "loss": 2.82, "step": 5730 }, { "epoch": 0.13, "grad_norm": 2.7577420901555616, "learning_rate": 1.9936851657142138e-05, "loss": 2.8855, "step": 5740 }, { "epoch": 0.13, "grad_norm": 2.0871717056445096, "learning_rate": 1.993644123945993e-05, "loss": 2.8434, "step": 5750 }, { "epoch": 0.13, "grad_norm": 2.6288893846164245, "learning_rate": 1.993602949663989e-05, "loss": 2.7573, "step": 5760 }, { "epoch": 0.13, "grad_norm": 2.5128346118969214, "learning_rate": 1.993561642873693e-05, "loss": 2.6796, "step": 5770 }, { "epoch": 0.13, "grad_norm": 4.530266249368864, "learning_rate": 1.9935202035806137e-05, "loss": 2.8444, "step": 5780 }, { "epoch": 0.13, "grad_norm": 3.6316393079058784, "learning_rate": 1.993478631790277e-05, "loss": 2.9395, "step": 5790 }, { "epoch": 0.13, "grad_norm": 3.0343613655902506, "learning_rate": 1.993436927508228e-05, "loss": 2.6837, "step": 5800 }, { "epoch": 0.13, "grad_norm": 2.309847990941736, "learning_rate": 1.9933950907400273e-05, "loss": 2.8185, "step": 5810 }, { "epoch": 0.13, "grad_norm": 2.445817456488047, "learning_rate": 1.9933531214912553e-05, "loss": 2.7116, "step": 5820 }, { "epoch": 0.13, "grad_norm": 3.3238997136147654, "learning_rate": 1.9933110197675086e-05, "loss": 2.777, "step": 5830 }, { "epoch": 0.13, "grad_norm": 3.308944494125035, "learning_rate": 1.9932687855744018e-05, "loss": 2.7674, "step": 5840 }, { "epoch": 0.13, "grad_norm": 3.5086198415656726, "learning_rate": 1.9932264189175675e-05, "loss": 2.8218, "step": 5850 }, { "epoch": 0.13, "grad_norm": 2.3447813550268664, "learning_rate": 1.9931839198026553e-05, "loss": 2.7758, "step": 5860 }, { "epoch": 0.13, "grad_norm": 2.2871696559232486, "learning_rate": 1.993141288235334e-05, "loss": 2.6995, "step": 5870 }, { "epoch": 0.13, "grad_norm": 4.138798686264896, "learning_rate": 1.993098524221288e-05, "loss": 2.8084, "step": 5880 }, { "epoch": 0.13, "grad_norm": 2.134880023901073, "learning_rate": 1.9930556277662213e-05, "loss": 2.8583, "step": 5890 }, { "epoch": 0.13, "grad_norm": 2.0829510026338354, "learning_rate": 1.9930125988758536e-05, "loss": 2.8409, "step": 5900 }, { "epoch": 0.13, "grad_norm": 2.3697520026046393, "learning_rate": 1.992969437555924e-05, "loss": 2.7151, "step": 5910 }, { "epoch": 0.13, "grad_norm": 4.787018084596695, "learning_rate": 1.9929261438121882e-05, "loss": 2.8581, "step": 5920 }, { "epoch": 0.13, "grad_norm": 1.8850261852621846, "learning_rate": 1.9928827176504206e-05, "loss": 2.8547, "step": 5930 }, { "epoch": 0.13, "grad_norm": 1.948714244577613, "learning_rate": 1.9928391590764114e-05, "loss": 2.8024, "step": 5940 }, { "epoch": 0.13, "grad_norm": 2.2943628176566966, "learning_rate": 1.9927954680959707e-05, "loss": 2.8641, "step": 5950 }, { "epoch": 0.13, "grad_norm": 2.5907126034647217, "learning_rate": 1.9927516447149244e-05, "loss": 2.7623, "step": 5960 }, { "epoch": 0.13, "grad_norm": 2.3080735196599726, "learning_rate": 1.9927076889391176e-05, "loss": 2.8614, "step": 5970 }, { "epoch": 0.13, "grad_norm": 2.842588817575145, "learning_rate": 1.9926636007744116e-05, "loss": 2.7578, "step": 5980 }, { "epoch": 0.14, "grad_norm": 3.1382479070914635, "learning_rate": 1.9926193802266866e-05, "loss": 2.7959, "step": 5990 }, { "epoch": 0.14, "grad_norm": 2.194188905915112, "learning_rate": 1.9925750273018395e-05, "loss": 2.7841, "step": 6000 }, { "epoch": 0.14, "grad_norm": 2.2088907078936786, "learning_rate": 1.9925305420057856e-05, "loss": 2.9106, "step": 6010 }, { "epoch": 0.14, "grad_norm": 2.8168331452076316, "learning_rate": 1.9924859243444575e-05, "loss": 2.8985, "step": 6020 }, { "epoch": 0.14, "grad_norm": 1.9968194468537916, "learning_rate": 1.992441174323805e-05, "loss": 2.8374, "step": 6030 }, { "epoch": 0.14, "grad_norm": 2.914526243879575, "learning_rate": 1.9923962919497968e-05, "loss": 2.9356, "step": 6040 }, { "epoch": 0.14, "grad_norm": 2.027020870339381, "learning_rate": 1.9923512772284176e-05, "loss": 2.9215, "step": 6050 }, { "epoch": 0.14, "grad_norm": 2.614924068384043, "learning_rate": 1.9923061301656716e-05, "loss": 2.8013, "step": 6060 }, { "epoch": 0.14, "grad_norm": 3.696841189011048, "learning_rate": 1.9922608507675787e-05, "loss": 2.6975, "step": 6070 }, { "epoch": 0.14, "grad_norm": 2.8913327499645964, "learning_rate": 1.992215439040178e-05, "loss": 2.8119, "step": 6080 }, { "epoch": 0.14, "grad_norm": 4.355580915481562, "learning_rate": 1.9921698949895257e-05, "loss": 2.762, "step": 6090 }, { "epoch": 0.14, "grad_norm": 2.1964320732347953, "learning_rate": 1.9921242186216954e-05, "loss": 2.7249, "step": 6100 }, { "epoch": 0.14, "grad_norm": 2.6833321536783594, "learning_rate": 1.9920784099427787e-05, "loss": 2.8406, "step": 6110 }, { "epoch": 0.14, "grad_norm": 2.307753865657894, "learning_rate": 1.9920324689588847e-05, "loss": 2.8127, "step": 6120 }, { "epoch": 0.14, "grad_norm": 3.980031368048901, "learning_rate": 1.9919863956761398e-05, "loss": 2.8417, "step": 6130 }, { "epoch": 0.14, "grad_norm": 2.6143319205482123, "learning_rate": 1.9919401901006893e-05, "loss": 2.7934, "step": 6140 }, { "epoch": 0.14, "grad_norm": 2.154658196540038, "learning_rate": 1.991893852238694e-05, "loss": 2.7489, "step": 6150 }, { "epoch": 0.14, "grad_norm": 2.6413975205517004, "learning_rate": 1.9918473820963345e-05, "loss": 2.813, "step": 6160 }, { "epoch": 0.14, "grad_norm": 2.8003463219385125, "learning_rate": 1.991800779679808e-05, "loss": 2.6713, "step": 6170 }, { "epoch": 0.14, "grad_norm": 2.206474229760586, "learning_rate": 1.9917540449953293e-05, "loss": 2.7125, "step": 6180 }, { "epoch": 0.14, "grad_norm": 2.186751962085766, "learning_rate": 1.9917071780491306e-05, "loss": 2.8428, "step": 6190 }, { "epoch": 0.14, "grad_norm": 3.67542809249096, "learning_rate": 1.991660178847463e-05, "loss": 2.6336, "step": 6200 }, { "epoch": 0.14, "grad_norm": 2.537464944481492, "learning_rate": 1.991613047396594e-05, "loss": 2.8186, "step": 6210 }, { "epoch": 0.14, "grad_norm": 3.0922286427280246, "learning_rate": 1.9915657837028088e-05, "loss": 2.9464, "step": 6220 }, { "epoch": 0.14, "grad_norm": 2.569960453495521, "learning_rate": 1.9915183877724107e-05, "loss": 2.7588, "step": 6230 }, { "epoch": 0.14, "grad_norm": 2.0545798173676735, "learning_rate": 1.991470859611721e-05, "loss": 2.7306, "step": 6240 }, { "epoch": 0.14, "grad_norm": 3.3613552870131573, "learning_rate": 1.9914231992270774e-05, "loss": 2.8251, "step": 6250 }, { "epoch": 0.14, "grad_norm": 3.2486249099193736, "learning_rate": 1.9913754066248364e-05, "loss": 2.702, "step": 6260 }, { "epoch": 0.14, "grad_norm": 2.2862650505898263, "learning_rate": 1.9913274818113715e-05, "loss": 2.6694, "step": 6270 }, { "epoch": 0.14, "grad_norm": 2.1277366480078572, "learning_rate": 1.9912794247930742e-05, "loss": 2.9849, "step": 6280 }, { "epoch": 0.14, "grad_norm": 6.066269511608426, "learning_rate": 1.9912312355763533e-05, "loss": 2.8148, "step": 6290 }, { "epoch": 0.14, "grad_norm": 2.7055502547081383, "learning_rate": 1.9911829141676354e-05, "loss": 2.7867, "step": 6300 }, { "epoch": 0.14, "grad_norm": 2.931343729599391, "learning_rate": 1.9911344605733648e-05, "loss": 2.6411, "step": 6310 }, { "epoch": 0.14, "grad_norm": 1.9353234211464536, "learning_rate": 1.991085874800003e-05, "loss": 2.8158, "step": 6320 }, { "epoch": 0.14, "grad_norm": 5.277986852859951, "learning_rate": 1.99103715685403e-05, "loss": 2.8455, "step": 6330 }, { "epoch": 0.14, "grad_norm": 4.171236470585959, "learning_rate": 1.9909883067419426e-05, "loss": 2.7542, "step": 6340 }, { "epoch": 0.14, "grad_norm": 2.847864909709101, "learning_rate": 1.9909393244702557e-05, "loss": 2.7513, "step": 6350 }, { "epoch": 0.14, "grad_norm": 3.970604886678481, "learning_rate": 1.9908902100455015e-05, "loss": 2.7181, "step": 6360 }, { "epoch": 0.14, "grad_norm": 2.4978799099361297, "learning_rate": 1.9908409634742295e-05, "loss": 2.771, "step": 6370 }, { "epoch": 0.14, "grad_norm": 2.087121695430665, "learning_rate": 1.9907915847630082e-05, "loss": 2.7593, "step": 6380 }, { "epoch": 0.14, "grad_norm": 1.9131271856532073, "learning_rate": 1.9907420739184223e-05, "loss": 2.8533, "step": 6390 }, { "epoch": 0.14, "grad_norm": 3.009071534547379, "learning_rate": 1.9906924309470748e-05, "loss": 2.8534, "step": 6400 }, { "epoch": 0.14, "grad_norm": 2.0845279860323505, "learning_rate": 1.9906426558555857e-05, "loss": 2.7376, "step": 6410 }, { "epoch": 0.14, "grad_norm": 2.528905266336736, "learning_rate": 1.990592748650594e-05, "loss": 2.6728, "step": 6420 }, { "epoch": 0.15, "grad_norm": 3.3895352585437393, "learning_rate": 1.9905427093387542e-05, "loss": 2.7762, "step": 6430 }, { "epoch": 0.15, "grad_norm": 2.291465389163573, "learning_rate": 1.990492537926741e-05, "loss": 2.8742, "step": 6440 }, { "epoch": 0.15, "grad_norm": 2.3041850373557233, "learning_rate": 1.9904422344212437e-05, "loss": 2.7935, "step": 6450 }, { "epoch": 0.15, "grad_norm": 1.9898718245511635, "learning_rate": 1.9903917988289722e-05, "loss": 2.8851, "step": 6460 }, { "epoch": 0.15, "grad_norm": 2.154115601031775, "learning_rate": 1.9903412311566524e-05, "loss": 2.8643, "step": 6470 }, { "epoch": 0.15, "grad_norm": 2.2151583077679935, "learning_rate": 1.9902905314110275e-05, "loss": 2.7173, "step": 6480 }, { "epoch": 0.15, "grad_norm": 4.655368106241364, "learning_rate": 1.9902396995988594e-05, "loss": 2.7703, "step": 6490 }, { "epoch": 0.15, "grad_norm": 2.736857239746388, "learning_rate": 1.9901887357269272e-05, "loss": 2.8684, "step": 6500 }, { "epoch": 0.15, "grad_norm": 2.583619252782247, "learning_rate": 1.990137639802027e-05, "loss": 2.7576, "step": 6510 }, { "epoch": 0.15, "grad_norm": 2.1606970833337305, "learning_rate": 1.9900864118309733e-05, "loss": 2.7992, "step": 6520 }, { "epoch": 0.15, "grad_norm": 2.9737036102567544, "learning_rate": 1.990035051820598e-05, "loss": 2.823, "step": 6530 }, { "epoch": 0.15, "grad_norm": 1.9314936830432459, "learning_rate": 1.9899835597777505e-05, "loss": 2.7594, "step": 6540 }, { "epoch": 0.15, "grad_norm": 2.3899101921915746, "learning_rate": 1.9899319357092976e-05, "loss": 2.6745, "step": 6550 }, { "epoch": 0.15, "grad_norm": 1.892373839349803, "learning_rate": 1.9898801796221246e-05, "loss": 2.8232, "step": 6560 }, { "epoch": 0.15, "grad_norm": 2.3194352856961533, "learning_rate": 1.989828291523133e-05, "loss": 2.7108, "step": 6570 }, { "epoch": 0.15, "grad_norm": 4.526078983333026, "learning_rate": 1.9897762714192432e-05, "loss": 2.7871, "step": 6580 }, { "epoch": 0.15, "grad_norm": 3.208024720358828, "learning_rate": 1.9897241193173923e-05, "loss": 2.7604, "step": 6590 }, { "epoch": 0.15, "grad_norm": 2.808133343334559, "learning_rate": 1.9896718352245358e-05, "loss": 2.7455, "step": 6600 }, { "epoch": 0.15, "grad_norm": 4.430475184537566, "learning_rate": 1.9896194191476464e-05, "loss": 2.7627, "step": 6610 }, { "epoch": 0.15, "grad_norm": 2.1512832368235464, "learning_rate": 1.9895668710937137e-05, "loss": 2.8264, "step": 6620 }, { "epoch": 0.15, "grad_norm": 2.1814637507468784, "learning_rate": 1.9895141910697462e-05, "loss": 2.8236, "step": 6630 }, { "epoch": 0.15, "grad_norm": 5.021165949899958, "learning_rate": 1.989461379082769e-05, "loss": 2.7978, "step": 6640 }, { "epoch": 0.15, "grad_norm": 3.530963431902333, "learning_rate": 1.9894084351398258e-05, "loss": 2.7761, "step": 6650 }, { "epoch": 0.15, "grad_norm": 1.9183801728714047, "learning_rate": 1.9893553592479763e-05, "loss": 2.7114, "step": 6660 }, { "epoch": 0.15, "grad_norm": 3.417630578193585, "learning_rate": 1.9893021514142997e-05, "loss": 2.7788, "step": 6670 }, { "epoch": 0.15, "grad_norm": 3.000621322413861, "learning_rate": 1.9892488116458916e-05, "loss": 2.7854, "step": 6680 }, { "epoch": 0.15, "grad_norm": 2.709303528381348, "learning_rate": 1.989195339949865e-05, "loss": 2.8182, "step": 6690 }, { "epoch": 0.15, "grad_norm": 2.046050594890124, "learning_rate": 1.989141736333352e-05, "loss": 2.8975, "step": 6700 }, { "epoch": 0.15, "grad_norm": 1.9468393257696541, "learning_rate": 1.9890880008034997e-05, "loss": 2.8564, "step": 6710 }, { "epoch": 0.15, "grad_norm": 1.84769820465814, "learning_rate": 1.989034133367476e-05, "loss": 2.8159, "step": 6720 }, { "epoch": 0.15, "grad_norm": 1.8971525915742204, "learning_rate": 1.9889801340324635e-05, "loss": 2.823, "step": 6730 }, { "epoch": 0.15, "grad_norm": 2.0069815919152103, "learning_rate": 1.988926002805664e-05, "loss": 2.85, "step": 6740 }, { "epoch": 0.15, "grad_norm": 2.575968904579521, "learning_rate": 1.988871739694297e-05, "loss": 2.7379, "step": 6750 }, { "epoch": 0.15, "grad_norm": 1.9072995698416606, "learning_rate": 1.988817344705599e-05, "loss": 2.7945, "step": 6760 }, { "epoch": 0.15, "grad_norm": 2.078534285467484, "learning_rate": 1.988762817846823e-05, "loss": 2.7819, "step": 6770 }, { "epoch": 0.15, "grad_norm": 3.3478658282054337, "learning_rate": 1.9887081591252426e-05, "loss": 2.8427, "step": 6780 }, { "epoch": 0.15, "grad_norm": 2.1856030020962645, "learning_rate": 1.9886533685481457e-05, "loss": 2.8425, "step": 6790 }, { "epoch": 0.15, "grad_norm": 1.996909008694423, "learning_rate": 1.9885984461228402e-05, "loss": 2.7475, "step": 6800 }, { "epoch": 0.15, "grad_norm": 2.2480336125387366, "learning_rate": 1.98854339185665e-05, "loss": 2.7617, "step": 6810 }, { "epoch": 0.15, "grad_norm": 1.901183468970178, "learning_rate": 1.9884882057569174e-05, "loss": 2.7441, "step": 6820 }, { "epoch": 0.15, "grad_norm": 2.0563053273749445, "learning_rate": 1.988432887831002e-05, "loss": 2.9282, "step": 6830 }, { "epoch": 0.15, "grad_norm": 3.2200931519193157, "learning_rate": 1.9883774380862818e-05, "loss": 2.7108, "step": 6840 }, { "epoch": 0.15, "grad_norm": 2.275950691472631, "learning_rate": 1.9883218565301507e-05, "loss": 2.6481, "step": 6850 }, { "epoch": 0.15, "grad_norm": 1.9366235142971633, "learning_rate": 1.9882661431700216e-05, "loss": 2.8212, "step": 6860 }, { "epoch": 0.15, "grad_norm": 2.1914953046969066, "learning_rate": 1.9882102980133244e-05, "loss": 2.7487, "step": 6870 }, { "epoch": 0.16, "grad_norm": 1.7796330042700534, "learning_rate": 1.9881543210675068e-05, "loss": 2.7838, "step": 6880 }, { "epoch": 0.16, "grad_norm": 3.753532181831686, "learning_rate": 1.988098212340034e-05, "loss": 2.7947, "step": 6890 }, { "epoch": 0.16, "grad_norm": 2.338194918509687, "learning_rate": 1.9880419718383885e-05, "loss": 2.7421, "step": 6900 }, { "epoch": 0.16, "grad_norm": 3.971465156438877, "learning_rate": 1.9879855995700707e-05, "loss": 2.7412, "step": 6910 }, { "epoch": 0.16, "grad_norm": 3.030013058822201, "learning_rate": 1.9879290955425982e-05, "loss": 2.6859, "step": 6920 }, { "epoch": 0.16, "grad_norm": 2.1664234405267857, "learning_rate": 1.9878724597635072e-05, "loss": 2.6961, "step": 6930 }, { "epoch": 0.16, "grad_norm": 2.2188440168641117, "learning_rate": 1.9878156922403503e-05, "loss": 2.8099, "step": 6940 }, { "epoch": 0.16, "grad_norm": 3.204733633076102, "learning_rate": 1.987758792980698e-05, "loss": 2.8147, "step": 6950 }, { "epoch": 0.16, "grad_norm": 1.785058482947381, "learning_rate": 1.9877017619921384e-05, "loss": 2.7723, "step": 6960 }, { "epoch": 0.16, "grad_norm": 1.937663579410184, "learning_rate": 1.9876445992822776e-05, "loss": 2.7491, "step": 6970 }, { "epoch": 0.16, "grad_norm": 1.9076914319633478, "learning_rate": 1.987587304858739e-05, "loss": 2.7934, "step": 6980 }, { "epoch": 0.16, "grad_norm": 2.3478324794579244, "learning_rate": 1.9875298787291626e-05, "loss": 2.7979, "step": 6990 }, { "epoch": 0.16, "grad_norm": 2.2369825281522986, "learning_rate": 1.9874723209012076e-05, "loss": 2.7094, "step": 7000 }, { "epoch": 0.16, "grad_norm": 3.351449243809098, "learning_rate": 1.9874146313825496e-05, "loss": 2.7508, "step": 7010 }, { "epoch": 0.16, "grad_norm": 2.075648733256519, "learning_rate": 1.987356810180883e-05, "loss": 2.8448, "step": 7020 }, { "epoch": 0.16, "grad_norm": 2.3397739260745163, "learning_rate": 1.987298857303918e-05, "loss": 2.7347, "step": 7030 }, { "epoch": 0.16, "grad_norm": 1.9528226287671457, "learning_rate": 1.9872407727593833e-05, "loss": 2.7783, "step": 7040 }, { "epoch": 0.16, "grad_norm": 2.251823105190714, "learning_rate": 1.9871825565550255e-05, "loss": 2.7287, "step": 7050 }, { "epoch": 0.16, "grad_norm": 3.721301825967627, "learning_rate": 1.987124208698608e-05, "loss": 2.7808, "step": 7060 }, { "epoch": 0.16, "grad_norm": 2.2951789286061035, "learning_rate": 1.9870657291979126e-05, "loss": 2.7598, "step": 7070 }, { "epoch": 0.16, "grad_norm": 2.2603515438952293, "learning_rate": 1.9870071180607383e-05, "loss": 2.7087, "step": 7080 }, { "epoch": 0.16, "grad_norm": 1.8456285012541112, "learning_rate": 1.986948375294901e-05, "loss": 2.859, "step": 7090 }, { "epoch": 0.16, "grad_norm": 2.2298489742755487, "learning_rate": 1.9868895009082353e-05, "loss": 2.6704, "step": 7100 }, { "epoch": 0.16, "grad_norm": 2.283042143307112, "learning_rate": 1.986830494908592e-05, "loss": 2.7838, "step": 7110 }, { "epoch": 0.16, "grad_norm": 2.8917727844111636, "learning_rate": 1.986771357303841e-05, "loss": 2.6893, "step": 7120 }, { "epoch": 0.16, "grad_norm": 6.2824340616744525, "learning_rate": 1.986712088101869e-05, "loss": 2.7521, "step": 7130 }, { "epoch": 0.16, "grad_norm": 2.232092751071759, "learning_rate": 1.9866526873105795e-05, "loss": 2.7984, "step": 7140 }, { "epoch": 0.16, "grad_norm": 1.741350801461087, "learning_rate": 1.9865931549378946e-05, "loss": 2.8474, "step": 7150 }, { "epoch": 0.16, "grad_norm": 2.84680140253457, "learning_rate": 1.986533490991754e-05, "loss": 2.8207, "step": 7160 }, { "epoch": 0.16, "grad_norm": 2.3287677413892616, "learning_rate": 1.9864736954801144e-05, "loss": 2.8227, "step": 7170 }, { "epoch": 0.16, "grad_norm": 3.0804631628214127, "learning_rate": 1.98641376841095e-05, "loss": 2.7904, "step": 7180 }, { "epoch": 0.16, "grad_norm": 3.2510710781599736, "learning_rate": 1.9863537097922526e-05, "loss": 2.8067, "step": 7190 }, { "epoch": 0.16, "grad_norm": 3.4392553625601514, "learning_rate": 1.9862935196320318e-05, "loss": 2.8173, "step": 7200 }, { "epoch": 0.16, "grad_norm": 3.4642146793581556, "learning_rate": 1.9862331979383152e-05, "loss": 2.7241, "step": 7210 }, { "epoch": 0.16, "grad_norm": 2.236583959390296, "learning_rate": 1.986172744719147e-05, "loss": 2.7655, "step": 7220 }, { "epoch": 0.16, "grad_norm": 1.8227933114552675, "learning_rate": 1.986112159982589e-05, "loss": 2.9201, "step": 7230 }, { "epoch": 0.16, "grad_norm": 2.6149103187399056, "learning_rate": 1.986051443736721e-05, "loss": 2.8146, "step": 7240 }, { "epoch": 0.16, "grad_norm": 2.6423487331811546, "learning_rate": 1.985990595989641e-05, "loss": 2.862, "step": 7250 }, { "epoch": 0.16, "grad_norm": 2.574745736284546, "learning_rate": 1.985929616749463e-05, "loss": 2.7864, "step": 7260 }, { "epoch": 0.16, "grad_norm": 3.834220624199108, "learning_rate": 1.985868506024319e-05, "loss": 2.7232, "step": 7270 }, { "epoch": 0.16, "grad_norm": 3.108137380798239, "learning_rate": 1.9858072638223594e-05, "loss": 2.7852, "step": 7280 }, { "epoch": 0.16, "grad_norm": 3.7651473357906586, "learning_rate": 1.985745890151751e-05, "loss": 2.6533, "step": 7290 }, { "epoch": 0.16, "grad_norm": 2.168775517977004, "learning_rate": 1.9856843850206794e-05, "loss": 2.7487, "step": 7300 }, { "epoch": 0.16, "grad_norm": 1.970704326745524, "learning_rate": 1.985622748437347e-05, "loss": 2.7631, "step": 7310 }, { "epoch": 0.17, "grad_norm": 2.2738331667292897, "learning_rate": 1.985560980409973e-05, "loss": 2.571, "step": 7320 }, { "epoch": 0.17, "grad_norm": 2.0715183714513072, "learning_rate": 1.9854990809467952e-05, "loss": 2.7761, "step": 7330 }, { "epoch": 0.17, "grad_norm": 2.319475049451401, "learning_rate": 1.9854370500560685e-05, "loss": 2.9174, "step": 7340 }, { "epoch": 0.17, "grad_norm": 2.1075739634464923, "learning_rate": 1.985374887746066e-05, "loss": 2.7851, "step": 7350 }, { "epoch": 0.17, "grad_norm": 1.9488996365166418, "learning_rate": 1.9853125940250773e-05, "loss": 2.8109, "step": 7360 }, { "epoch": 0.17, "grad_norm": 2.1438738272239166, "learning_rate": 1.9852501689014095e-05, "loss": 2.7199, "step": 7370 }, { "epoch": 0.17, "grad_norm": 2.9071067306089344, "learning_rate": 1.985187612383389e-05, "loss": 2.8538, "step": 7380 }, { "epoch": 0.17, "grad_norm": 2.785527804325284, "learning_rate": 1.985124924479357e-05, "loss": 2.7603, "step": 7390 }, { "epoch": 0.17, "grad_norm": 1.9189993899795572, "learning_rate": 1.9850621051976747e-05, "loss": 2.7666, "step": 7400 }, { "epoch": 0.17, "grad_norm": 1.9358928431146367, "learning_rate": 1.9849991545467192e-05, "loss": 2.7499, "step": 7410 }, { "epoch": 0.17, "grad_norm": 2.1806144879699443, "learning_rate": 1.9849360725348855e-05, "loss": 2.8159, "step": 7420 }, { "epoch": 0.17, "grad_norm": 2.216211845374149, "learning_rate": 1.9848728591705872e-05, "loss": 2.6935, "step": 7430 }, { "epoch": 0.17, "grad_norm": 2.125777875463922, "learning_rate": 1.9848095144622538e-05, "loss": 2.7538, "step": 7440 }, { "epoch": 0.17, "grad_norm": 2.5019923413138487, "learning_rate": 1.984746038418333e-05, "loss": 2.8138, "step": 7450 }, { "epoch": 0.17, "grad_norm": 2.126201613134078, "learning_rate": 1.9846824310472907e-05, "loss": 2.7392, "step": 7460 }, { "epoch": 0.17, "grad_norm": 2.184896681373899, "learning_rate": 1.984618692357609e-05, "loss": 2.7175, "step": 7470 }, { "epoch": 0.17, "grad_norm": 3.230793464315536, "learning_rate": 1.9845548223577883e-05, "loss": 2.7836, "step": 7480 }, { "epoch": 0.17, "grad_norm": 2.207500770377193, "learning_rate": 1.984490821056347e-05, "loss": 2.7012, "step": 7490 }, { "epoch": 0.17, "grad_norm": 1.780043421937432, "learning_rate": 1.9844266884618194e-05, "loss": 2.8991, "step": 7500 }, { "epoch": 0.17, "grad_norm": 2.1205628710625524, "learning_rate": 1.984362424582759e-05, "loss": 2.681, "step": 7510 }, { "epoch": 0.17, "grad_norm": 2.4726015333675893, "learning_rate": 1.9842980294277357e-05, "loss": 2.7951, "step": 7520 }, { "epoch": 0.17, "grad_norm": 2.2286371883593254, "learning_rate": 1.984233503005338e-05, "loss": 2.9121, "step": 7530 }, { "epoch": 0.17, "grad_norm": 2.559244077890361, "learning_rate": 1.9841688453241706e-05, "loss": 2.8912, "step": 7540 }, { "epoch": 0.17, "grad_norm": 3.1656621853979683, "learning_rate": 1.9841040563928568e-05, "loss": 2.781, "step": 7550 }, { "epoch": 0.17, "grad_norm": 1.991491583656647, "learning_rate": 1.9840391362200364e-05, "loss": 2.8035, "step": 7560 }, { "epoch": 0.17, "grad_norm": 3.2717696071061955, "learning_rate": 1.9839740848143677e-05, "loss": 2.8314, "step": 7570 }, { "epoch": 0.17, "grad_norm": 2.7167345007050536, "learning_rate": 1.983908902184526e-05, "loss": 2.7693, "step": 7580 }, { "epoch": 0.17, "grad_norm": 4.448619834242887, "learning_rate": 1.9838435883392035e-05, "loss": 2.6854, "step": 7590 }, { "epoch": 0.17, "grad_norm": 2.043938896498377, "learning_rate": 1.9837781432871115e-05, "loss": 2.872, "step": 7600 }, { "epoch": 0.17, "grad_norm": 1.8291297608233734, "learning_rate": 1.9837125670369774e-05, "loss": 2.7263, "step": 7610 }, { "epoch": 0.17, "grad_norm": 2.139421740980679, "learning_rate": 1.9836468595975463e-05, "loss": 2.9021, "step": 7620 }, { "epoch": 0.17, "grad_norm": 1.9293256622632395, "learning_rate": 1.983581020977582e-05, "loss": 2.7197, "step": 7630 }, { "epoch": 0.17, "grad_norm": 2.0308124938753465, "learning_rate": 1.9835150511858635e-05, "loss": 2.7921, "step": 7640 }, { "epoch": 0.17, "grad_norm": 2.421399390150751, "learning_rate": 1.9834489502311894e-05, "loss": 2.6348, "step": 7650 }, { "epoch": 0.17, "grad_norm": 2.0821866971415863, "learning_rate": 1.9833827181223753e-05, "loss": 2.8018, "step": 7660 }, { "epoch": 0.17, "grad_norm": 3.032304656402104, "learning_rate": 1.983316354868253e-05, "loss": 2.7378, "step": 7670 }, { "epoch": 0.17, "grad_norm": 3.1202042371910372, "learning_rate": 1.983249860477674e-05, "loss": 2.7304, "step": 7680 }, { "epoch": 0.17, "grad_norm": 2.374585805292361, "learning_rate": 1.983183234959505e-05, "loss": 2.752, "step": 7690 }, { "epoch": 0.17, "grad_norm": 3.19915233485909, "learning_rate": 1.9831164783226317e-05, "loss": 2.7657, "step": 7700 }, { "epoch": 0.17, "grad_norm": 3.089675612683068, "learning_rate": 1.9830495905759573e-05, "loss": 2.6808, "step": 7710 }, { "epoch": 0.17, "grad_norm": 2.4552726369220848, "learning_rate": 1.9829825717284016e-05, "loss": 2.7833, "step": 7720 }, { "epoch": 0.17, "grad_norm": 2.7824765749267657, "learning_rate": 1.9829154217889018e-05, "loss": 2.7867, "step": 7730 }, { "epoch": 0.17, "grad_norm": 2.0589028316828992, "learning_rate": 1.9828481407664143e-05, "loss": 2.8464, "step": 7740 }, { "epoch": 0.17, "grad_norm": 2.368108253737227, "learning_rate": 1.9827807286699114e-05, "loss": 2.8842, "step": 7750 }, { "epoch": 0.17, "grad_norm": 2.323521791013013, "learning_rate": 1.9827131855083823e-05, "loss": 2.8175, "step": 7760 }, { "epoch": 0.18, "grad_norm": 1.810377116073987, "learning_rate": 1.9826455112908358e-05, "loss": 2.8427, "step": 7770 }, { "epoch": 0.18, "grad_norm": 2.2249531615968063, "learning_rate": 1.982577706026297e-05, "loss": 2.7403, "step": 7780 }, { "epoch": 0.18, "grad_norm": 2.0713656413942534, "learning_rate": 1.9825097697238074e-05, "loss": 2.8481, "step": 7790 }, { "epoch": 0.18, "grad_norm": 2.9983581697877324, "learning_rate": 1.9824417023924282e-05, "loss": 2.7384, "step": 7800 }, { "epoch": 0.18, "grad_norm": 2.0466793578050337, "learning_rate": 1.9823735040412362e-05, "loss": 2.7904, "step": 7810 }, { "epoch": 0.18, "grad_norm": 3.0418134981977683, "learning_rate": 1.9823051746793275e-05, "loss": 2.7663, "step": 7820 }, { "epoch": 0.18, "grad_norm": 2.2760815605684668, "learning_rate": 1.9822367143158134e-05, "loss": 2.905, "step": 7830 }, { "epoch": 0.18, "grad_norm": 2.4520887555033073, "learning_rate": 1.9821681229598243e-05, "loss": 2.863, "step": 7840 }, { "epoch": 0.18, "grad_norm": 1.8694949039894355, "learning_rate": 1.9820994006205085e-05, "loss": 2.7823, "step": 7850 }, { "epoch": 0.18, "grad_norm": 2.4696264794981007, "learning_rate": 1.982030547307029e-05, "loss": 2.8058, "step": 7860 }, { "epoch": 0.18, "grad_norm": 2.275143087765355, "learning_rate": 1.9819615630285703e-05, "loss": 2.864, "step": 7870 }, { "epoch": 0.18, "grad_norm": 1.9474986546869546, "learning_rate": 1.9818924477943304e-05, "loss": 2.8877, "step": 7880 }, { "epoch": 0.18, "grad_norm": 1.9911399318881877, "learning_rate": 1.981823201613528e-05, "loss": 2.8716, "step": 7890 }, { "epoch": 0.18, "grad_norm": 1.9143243773713794, "learning_rate": 1.9817538244953975e-05, "loss": 2.8228, "step": 7900 }, { "epoch": 0.18, "grad_norm": 2.300045022941775, "learning_rate": 1.9816843164491905e-05, "loss": 2.7614, "step": 7910 }, { "epoch": 0.18, "grad_norm": 2.9774419037632627, "learning_rate": 1.981614677484177e-05, "loss": 2.7542, "step": 7920 }, { "epoch": 0.18, "grad_norm": 2.6569282741462334, "learning_rate": 1.9815449076096445e-05, "loss": 2.8052, "step": 7930 }, { "epoch": 0.18, "grad_norm": 2.4254499980427457, "learning_rate": 1.9814750068348974e-05, "loss": 2.8025, "step": 7940 }, { "epoch": 0.18, "grad_norm": 4.741113518107711, "learning_rate": 1.981404975169258e-05, "loss": 2.6942, "step": 7950 }, { "epoch": 0.18, "grad_norm": 2.239686598735432, "learning_rate": 1.9813348126220652e-05, "loss": 2.7109, "step": 7960 }, { "epoch": 0.18, "grad_norm": 2.772665022101185, "learning_rate": 1.9812645192026767e-05, "loss": 2.7214, "step": 7970 }, { "epoch": 0.18, "grad_norm": 2.584534432015499, "learning_rate": 1.9811940949204662e-05, "loss": 2.7626, "step": 7980 }, { "epoch": 0.18, "grad_norm": 3.184179854250384, "learning_rate": 1.9811235397848265e-05, "loss": 2.804, "step": 7990 }, { "epoch": 0.18, "grad_norm": 2.2391983172096297, "learning_rate": 1.9810528538051658e-05, "loss": 2.767, "step": 8000 }, { "epoch": 0.18, "grad_norm": 2.794434777706539, "learning_rate": 1.980982036990912e-05, "loss": 2.7833, "step": 8010 }, { "epoch": 0.18, "grad_norm": 2.2981522871368547, "learning_rate": 1.9809110893515084e-05, "loss": 2.7695, "step": 8020 }, { "epoch": 0.18, "grad_norm": 2.347337553058434, "learning_rate": 1.9808400108964173e-05, "loss": 2.7741, "step": 8030 }, { "epoch": 0.18, "grad_norm": 2.2258648255745856, "learning_rate": 1.980768801635118e-05, "loss": 2.7162, "step": 8040 }, { "epoch": 0.18, "grad_norm": 2.283012510156272, "learning_rate": 1.9806974615771063e-05, "loss": 2.8373, "step": 8050 }, { "epoch": 0.18, "grad_norm": 2.4076441544696876, "learning_rate": 1.9806259907318964e-05, "loss": 2.7608, "step": 8060 }, { "epoch": 0.18, "grad_norm": 5.646583504499038, "learning_rate": 1.98055438910902e-05, "loss": 2.6631, "step": 8070 }, { "epoch": 0.18, "grad_norm": 2.6296081801124846, "learning_rate": 1.9804826567180263e-05, "loss": 2.7595, "step": 8080 }, { "epoch": 0.18, "grad_norm": 3.0597461197940308, "learning_rate": 1.980410793568481e-05, "loss": 2.8465, "step": 8090 }, { "epoch": 0.18, "grad_norm": 2.060987519683762, "learning_rate": 1.980338799669968e-05, "loss": 2.8306, "step": 8100 }, { "epoch": 0.18, "grad_norm": 2.2603270430958258, "learning_rate": 1.980266675032089e-05, "loss": 2.817, "step": 8110 }, { "epoch": 0.18, "grad_norm": 2.898478578085197, "learning_rate": 1.980194419664462e-05, "loss": 2.6832, "step": 8120 }, { "epoch": 0.18, "grad_norm": 2.0064950999593414, "learning_rate": 1.9801220335767236e-05, "loss": 2.8735, "step": 8130 }, { "epoch": 0.18, "grad_norm": 3.2249675638117914, "learning_rate": 1.980049516778527e-05, "loss": 2.779, "step": 8140 }, { "epoch": 0.18, "grad_norm": 2.582710268445914, "learning_rate": 1.9799768692795433e-05, "loss": 2.7632, "step": 8150 }, { "epoch": 0.18, "grad_norm": 2.4862016261356445, "learning_rate": 1.9799040910894607e-05, "loss": 2.7895, "step": 8160 }, { "epoch": 0.18, "grad_norm": 2.8224321271313184, "learning_rate": 1.9798311822179852e-05, "loss": 2.7574, "step": 8170 }, { "epoch": 0.18, "grad_norm": 3.2633318701503042, "learning_rate": 1.97975814267484e-05, "loss": 2.7631, "step": 8180 }, { "epoch": 0.18, "grad_norm": 3.3231954098568126, "learning_rate": 1.9796849724697656e-05, "loss": 2.8053, "step": 8190 }, { "epoch": 0.18, "grad_norm": 2.0403101840863758, "learning_rate": 1.9796116716125203e-05, "loss": 2.7799, "step": 8200 }, { "epoch": 0.19, "grad_norm": 2.036352464153201, "learning_rate": 1.979538240112879e-05, "loss": 2.7478, "step": 8210 }, { "epoch": 0.19, "grad_norm": 2.463032563067108, "learning_rate": 1.9794646779806356e-05, "loss": 2.6908, "step": 8220 }, { "epoch": 0.19, "grad_norm": 2.11529359017937, "learning_rate": 1.9793909852255998e-05, "loss": 2.7408, "step": 8230 }, { "epoch": 0.19, "grad_norm": 2.0547673491273306, "learning_rate": 1.9793171618575994e-05, "loss": 2.736, "step": 8240 }, { "epoch": 0.19, "grad_norm": 1.942727805282583, "learning_rate": 1.97924320788648e-05, "loss": 2.6862, "step": 8250 }, { "epoch": 0.19, "grad_norm": 1.9917187136712629, "learning_rate": 1.9791691233221036e-05, "loss": 2.8366, "step": 8260 }, { "epoch": 0.19, "grad_norm": 2.509656753560939, "learning_rate": 1.9790949081743505e-05, "loss": 2.7784, "step": 8270 }, { "epoch": 0.19, "grad_norm": 1.9988093405863812, "learning_rate": 1.9790205624531184e-05, "loss": 2.7868, "step": 8280 }, { "epoch": 0.19, "grad_norm": 2.5197210626791953, "learning_rate": 1.978946086168322e-05, "loss": 2.7596, "step": 8290 }, { "epoch": 0.19, "grad_norm": 1.8250245887219183, "learning_rate": 1.9788714793298936e-05, "loss": 2.7694, "step": 8300 }, { "epoch": 0.19, "grad_norm": 2.1109947824519626, "learning_rate": 1.9787967419477825e-05, "loss": 2.8089, "step": 8310 }, { "epoch": 0.19, "grad_norm": 2.731210792592392, "learning_rate": 1.978721874031956e-05, "loss": 2.8564, "step": 8320 }, { "epoch": 0.19, "grad_norm": 1.8293837380367588, "learning_rate": 1.9786468755923986e-05, "loss": 2.7939, "step": 8330 }, { "epoch": 0.19, "grad_norm": 3.0545568265931333, "learning_rate": 1.9785717466391126e-05, "loss": 2.7823, "step": 8340 }, { "epoch": 0.19, "grad_norm": 2.3486528602558314, "learning_rate": 1.9784964871821166e-05, "loss": 2.7852, "step": 8350 }, { "epoch": 0.19, "grad_norm": 1.9113831009116968, "learning_rate": 1.978421097231448e-05, "loss": 2.7687, "step": 8360 }, { "epoch": 0.19, "grad_norm": 4.93944182157382, "learning_rate": 1.97834557679716e-05, "loss": 2.7277, "step": 8370 }, { "epoch": 0.19, "grad_norm": 2.63205363791064, "learning_rate": 1.978269925889325e-05, "loss": 2.715, "step": 8380 }, { "epoch": 0.19, "grad_norm": 2.2524962685768783, "learning_rate": 1.9781941445180317e-05, "loss": 2.6939, "step": 8390 }, { "epoch": 0.19, "grad_norm": 2.644390641764994, "learning_rate": 1.978118232693386e-05, "loss": 2.7386, "step": 8400 }, { "epoch": 0.19, "grad_norm": 2.013763659948002, "learning_rate": 1.9780421904255124e-05, "loss": 2.7665, "step": 8410 }, { "epoch": 0.19, "grad_norm": 1.8969083380617073, "learning_rate": 1.977966017724551e-05, "loss": 2.7969, "step": 8420 }, { "epoch": 0.19, "grad_norm": 1.867684488909579, "learning_rate": 1.977889714600661e-05, "loss": 2.8625, "step": 8430 }, { "epoch": 0.19, "grad_norm": 2.968667793330346, "learning_rate": 1.977813281064018e-05, "loss": 2.798, "step": 8440 }, { "epoch": 0.19, "grad_norm": 2.6689542995569258, "learning_rate": 1.9777367171248155e-05, "loss": 2.7208, "step": 8450 }, { "epoch": 0.19, "grad_norm": 4.5537700492642506, "learning_rate": 1.9776600227932642e-05, "loss": 2.811, "step": 8460 }, { "epoch": 0.19, "grad_norm": 2.325344379467167, "learning_rate": 1.9775831980795915e-05, "loss": 2.6506, "step": 8470 }, { "epoch": 0.19, "grad_norm": 2.050196433740164, "learning_rate": 1.977506242994044e-05, "loss": 2.6527, "step": 8480 }, { "epoch": 0.19, "grad_norm": 1.855092460467338, "learning_rate": 1.9774291575468835e-05, "loss": 2.7331, "step": 8490 }, { "epoch": 0.19, "grad_norm": 1.8883421475150903, "learning_rate": 1.9773519417483907e-05, "loss": 2.758, "step": 8500 }, { "epoch": 0.19, "grad_norm": 2.2325523414224153, "learning_rate": 1.9772745956088633e-05, "loss": 2.7346, "step": 8510 }, { "epoch": 0.19, "grad_norm": 3.0974332495235233, "learning_rate": 1.977197119138616e-05, "loss": 2.7167, "step": 8520 }, { "epoch": 0.19, "grad_norm": 2.866197523133204, "learning_rate": 1.977119512347981e-05, "loss": 2.7813, "step": 8530 }, { "epoch": 0.19, "grad_norm": 2.9629577306788986, "learning_rate": 1.977041775247309e-05, "loss": 2.723, "step": 8540 }, { "epoch": 0.19, "grad_norm": 1.9082422766593536, "learning_rate": 1.976963907846966e-05, "loss": 2.8531, "step": 8550 }, { "epoch": 0.19, "grad_norm": 2.6350920321996623, "learning_rate": 1.9768859101573373e-05, "loss": 2.7497, "step": 8560 }, { "epoch": 0.19, "grad_norm": 2.450827281153569, "learning_rate": 1.976807782188824e-05, "loss": 2.7888, "step": 8570 }, { "epoch": 0.19, "grad_norm": 1.8922345442019677, "learning_rate": 1.9767295239518464e-05, "loss": 2.6985, "step": 8580 }, { "epoch": 0.19, "grad_norm": 1.986988153985144, "learning_rate": 1.9766511354568403e-05, "loss": 2.8186, "step": 8590 }, { "epoch": 0.19, "grad_norm": 2.2515559344439793, "learning_rate": 1.97657261671426e-05, "loss": 2.7135, "step": 8600 }, { "epoch": 0.19, "grad_norm": 2.0733223019099154, "learning_rate": 1.9764939677345767e-05, "loss": 2.7489, "step": 8610 }, { "epoch": 0.19, "grad_norm": 1.9776648557037269, "learning_rate": 1.9764151885282796e-05, "loss": 2.7966, "step": 8620 }, { "epoch": 0.19, "grad_norm": 1.710016817143567, "learning_rate": 1.976336279105874e-05, "loss": 2.7381, "step": 8630 }, { "epoch": 0.19, "grad_norm": 2.076852045747656, "learning_rate": 1.9762572394778844e-05, "loss": 2.7044, "step": 8640 }, { "epoch": 0.2, "grad_norm": 1.9394523895294256, "learning_rate": 1.9761780696548504e-05, "loss": 2.6847, "step": 8650 }, { "epoch": 0.2, "grad_norm": 2.050550154211569, "learning_rate": 1.9760987696473316e-05, "loss": 2.8418, "step": 8660 }, { "epoch": 0.2, "grad_norm": 1.8017201067874828, "learning_rate": 1.9760193394659025e-05, "loss": 2.8307, "step": 8670 }, { "epoch": 0.2, "grad_norm": 1.8038224536255467, "learning_rate": 1.975939779121156e-05, "loss": 2.8754, "step": 8680 }, { "epoch": 0.2, "grad_norm": 3.450398857068499, "learning_rate": 1.9758600886237035e-05, "loss": 2.7619, "step": 8690 }, { "epoch": 0.2, "grad_norm": 1.8952055765739566, "learning_rate": 1.9757802679841713e-05, "loss": 2.7576, "step": 8700 }, { "epoch": 0.2, "grad_norm": 3.198153704094149, "learning_rate": 1.9757003172132052e-05, "loss": 2.8633, "step": 8710 }, { "epoch": 0.2, "grad_norm": 3.3818538450971287, "learning_rate": 1.975620236321467e-05, "loss": 2.759, "step": 8720 }, { "epoch": 0.2, "grad_norm": 2.6904081205231574, "learning_rate": 1.9755400253196373e-05, "loss": 2.7211, "step": 8730 }, { "epoch": 0.2, "grad_norm": 2.4098565460186183, "learning_rate": 1.9754596842184125e-05, "loss": 2.8198, "step": 8740 }, { "epoch": 0.2, "grad_norm": 2.5871875508096576, "learning_rate": 1.9753792130285067e-05, "loss": 2.7334, "step": 8750 }, { "epoch": 0.2, "grad_norm": 1.8970160347730445, "learning_rate": 1.975298611760652e-05, "loss": 2.7739, "step": 8760 }, { "epoch": 0.2, "grad_norm": 2.2821561224317737, "learning_rate": 1.9752178804255978e-05, "loss": 2.8074, "step": 8770 }, { "epoch": 0.2, "grad_norm": 2.067011023748488, "learning_rate": 1.9751370190341104e-05, "loss": 2.8068, "step": 8780 }, { "epoch": 0.2, "grad_norm": 4.2790198504383685, "learning_rate": 1.9750560275969732e-05, "loss": 2.8478, "step": 8790 }, { "epoch": 0.2, "grad_norm": 1.9925778532042455, "learning_rate": 1.9749749061249876e-05, "loss": 2.802, "step": 8800 }, { "epoch": 0.2, "grad_norm": 2.023560382554052, "learning_rate": 1.9748936546289724e-05, "loss": 2.9319, "step": 8810 }, { "epoch": 0.2, "grad_norm": 1.972030912990174, "learning_rate": 1.9748122731197626e-05, "loss": 2.8045, "step": 8820 }, { "epoch": 0.2, "grad_norm": 1.9944422284610348, "learning_rate": 1.9747307616082122e-05, "loss": 2.6793, "step": 8830 }, { "epoch": 0.2, "grad_norm": 2.0993019573538536, "learning_rate": 1.9746491201051913e-05, "loss": 2.8422, "step": 8840 }, { "epoch": 0.2, "grad_norm": 2.0461086506415413, "learning_rate": 1.9745673486215874e-05, "loss": 2.8194, "step": 8850 }, { "epoch": 0.2, "grad_norm": 2.422104815806678, "learning_rate": 1.9744854471683062e-05, "loss": 2.721, "step": 8860 }, { "epoch": 0.2, "grad_norm": 2.1085345036695156, "learning_rate": 1.9744034157562703e-05, "loss": 2.801, "step": 8870 }, { "epoch": 0.2, "grad_norm": 2.306126782680104, "learning_rate": 1.9743212543964186e-05, "loss": 2.6817, "step": 8880 }, { "epoch": 0.2, "grad_norm": 1.7899038223945587, "learning_rate": 1.9742389630997095e-05, "loss": 2.7638, "step": 8890 }, { "epoch": 0.2, "grad_norm": 2.7536215583752996, "learning_rate": 1.9741565418771163e-05, "loss": 2.7283, "step": 8900 }, { "epoch": 0.2, "grad_norm": 2.0970480488610868, "learning_rate": 1.974073990739632e-05, "loss": 2.761, "step": 8910 }, { "epoch": 0.2, "grad_norm": 2.19423343852118, "learning_rate": 1.973991309698265e-05, "loss": 2.7765, "step": 8920 }, { "epoch": 0.2, "grad_norm": 2.2530734008123914, "learning_rate": 1.9739084987640415e-05, "loss": 2.7113, "step": 8930 }, { "epoch": 0.2, "grad_norm": 2.5989207140876394, "learning_rate": 1.9738255579480058e-05, "loss": 2.6712, "step": 8940 }, { "epoch": 0.2, "grad_norm": 2.382407442419438, "learning_rate": 1.973742487261219e-05, "loss": 2.8699, "step": 8950 }, { "epoch": 0.2, "grad_norm": 1.9079481326492784, "learning_rate": 1.9736592867147593e-05, "loss": 2.7963, "step": 8960 }, { "epoch": 0.2, "grad_norm": 3.5410692073141505, "learning_rate": 1.9735759563197224e-05, "loss": 2.7617, "step": 8970 }, { "epoch": 0.2, "grad_norm": 1.9049090430220255, "learning_rate": 1.973492496087222e-05, "loss": 2.7462, "step": 8980 }, { "epoch": 0.2, "grad_norm": 2.4573644115768833, "learning_rate": 1.9734089060283875e-05, "loss": 2.5379, "step": 8990 }, { "epoch": 0.2, "grad_norm": 4.2652963707273965, "learning_rate": 1.973325186154367e-05, "loss": 2.7781, "step": 9000 }, { "epoch": 0.2, "grad_norm": 2.122250948818721, "learning_rate": 1.973241336476326e-05, "loss": 2.7602, "step": 9010 }, { "epoch": 0.2, "grad_norm": 2.6230933185065, "learning_rate": 1.973157357005446e-05, "loss": 2.7455, "step": 9020 }, { "epoch": 0.2, "grad_norm": 2.4185198124250524, "learning_rate": 1.973073247752927e-05, "loss": 2.7976, "step": 9030 }, { "epoch": 0.2, "grad_norm": 1.9707023808436606, "learning_rate": 1.972989008729986e-05, "loss": 2.8381, "step": 9040 }, { "epoch": 0.2, "grad_norm": 3.4999605949218795, "learning_rate": 1.9729046399478573e-05, "loss": 2.7183, "step": 9050 }, { "epoch": 0.2, "grad_norm": 1.8417796901789367, "learning_rate": 1.972820141417792e-05, "loss": 2.898, "step": 9060 }, { "epoch": 0.2, "grad_norm": 2.019550119087328, "learning_rate": 1.9727355131510593e-05, "loss": 2.8403, "step": 9070 }, { "epoch": 0.2, "grad_norm": 3.660361754843532, "learning_rate": 1.9726507551589452e-05, "loss": 2.6804, "step": 9080 }, { "epoch": 0.2, "grad_norm": 2.0161277541110953, "learning_rate": 1.9725658674527533e-05, "loss": 2.7804, "step": 9090 }, { "epoch": 0.21, "grad_norm": 2.1908956408155436, "learning_rate": 1.9724808500438043e-05, "loss": 2.7107, "step": 9100 }, { "epoch": 0.21, "grad_norm": 2.087884654000585, "learning_rate": 1.9723957029434366e-05, "loss": 2.8359, "step": 9110 }, { "epoch": 0.21, "grad_norm": 1.9675560116517132, "learning_rate": 1.9723104261630044e-05, "loss": 2.7274, "step": 9120 }, { "epoch": 0.21, "grad_norm": 2.1109691172450353, "learning_rate": 1.9722250197138817e-05, "loss": 2.6828, "step": 9130 }, { "epoch": 0.21, "grad_norm": 1.7438836470498078, "learning_rate": 1.9721394836074578e-05, "loss": 2.7123, "step": 9140 }, { "epoch": 0.21, "grad_norm": 2.3037296367480646, "learning_rate": 1.9720538178551397e-05, "loss": 2.6897, "step": 9150 }, { "epoch": 0.21, "grad_norm": 2.3009098830256707, "learning_rate": 1.9719680224683518e-05, "loss": 2.8224, "step": 9160 }, { "epoch": 0.21, "grad_norm": 2.0167930329831982, "learning_rate": 1.9718820974585365e-05, "loss": 2.7731, "step": 9170 }, { "epoch": 0.21, "grad_norm": 1.86222412323092, "learning_rate": 1.9717960428371528e-05, "loss": 2.739, "step": 9180 }, { "epoch": 0.21, "grad_norm": 2.1118497793418425, "learning_rate": 1.9717098586156767e-05, "loss": 2.6866, "step": 9190 }, { "epoch": 0.21, "grad_norm": 2.37416863757663, "learning_rate": 1.971623544805602e-05, "loss": 2.9245, "step": 9200 }, { "epoch": 0.21, "grad_norm": 2.380754798610577, "learning_rate": 1.97153710141844e-05, "loss": 2.6724, "step": 9210 }, { "epoch": 0.21, "grad_norm": 2.0536833713039977, "learning_rate": 1.9714505284657178e-05, "loss": 2.6208, "step": 9220 }, { "epoch": 0.21, "grad_norm": 1.7080454281294544, "learning_rate": 1.971363825958982e-05, "loss": 2.8374, "step": 9230 }, { "epoch": 0.21, "grad_norm": 2.149429751864897, "learning_rate": 1.971276993909795e-05, "loss": 2.8651, "step": 9240 }, { "epoch": 0.21, "grad_norm": 1.8460152595518649, "learning_rate": 1.9711900323297372e-05, "loss": 2.7399, "step": 9250 }, { "epoch": 0.21, "grad_norm": 3.4594992378675573, "learning_rate": 1.9711029412304052e-05, "loss": 2.7721, "step": 9260 }, { "epoch": 0.21, "grad_norm": 2.205306782861007, "learning_rate": 1.971015720623414e-05, "loss": 2.6537, "step": 9270 }, { "epoch": 0.21, "grad_norm": 2.3709425791697285, "learning_rate": 1.9709283705203956e-05, "loss": 2.8382, "step": 9280 }, { "epoch": 0.21, "grad_norm": 2.0480285571251433, "learning_rate": 1.970840890932999e-05, "loss": 2.7732, "step": 9290 }, { "epoch": 0.21, "grad_norm": 3.8804762899493515, "learning_rate": 1.9707532818728902e-05, "loss": 2.7877, "step": 9300 }, { "epoch": 0.21, "grad_norm": 4.000239763185581, "learning_rate": 1.970665543351754e-05, "loss": 2.7848, "step": 9310 }, { "epoch": 0.21, "grad_norm": 2.399790470307285, "learning_rate": 1.9705776753812897e-05, "loss": 2.7252, "step": 9320 }, { "epoch": 0.21, "grad_norm": 2.3691733301576807, "learning_rate": 1.970489677973217e-05, "loss": 2.735, "step": 9330 }, { "epoch": 0.21, "grad_norm": 1.9008710358701064, "learning_rate": 1.9704015511392704e-05, "loss": 2.761, "step": 9340 }, { "epoch": 0.21, "grad_norm": 2.0693623774352687, "learning_rate": 1.970313294891203e-05, "loss": 2.699, "step": 9350 }, { "epoch": 0.21, "grad_norm": 2.3941455627648414, "learning_rate": 1.970224909240785e-05, "loss": 2.8186, "step": 9360 }, { "epoch": 0.21, "grad_norm": 2.1638620407703644, "learning_rate": 1.970136394199803e-05, "loss": 2.8298, "step": 9370 }, { "epoch": 0.21, "grad_norm": 2.8948298931207908, "learning_rate": 1.970047749780062e-05, "loss": 2.635, "step": 9380 }, { "epoch": 0.21, "grad_norm": 2.5907495133699254, "learning_rate": 1.9699589759933834e-05, "loss": 2.7034, "step": 9390 }, { "epoch": 0.21, "grad_norm": 2.4386363055794584, "learning_rate": 1.9698700728516065e-05, "loss": 2.7474, "step": 9400 }, { "epoch": 0.21, "grad_norm": 2.123695793187867, "learning_rate": 1.9697810403665874e-05, "loss": 2.7697, "step": 9410 }, { "epoch": 0.21, "grad_norm": 1.8852597795201749, "learning_rate": 1.9696918785502e-05, "loss": 2.7118, "step": 9420 }, { "epoch": 0.21, "grad_norm": 2.0883430423875957, "learning_rate": 1.9696025874143342e-05, "loss": 2.6715, "step": 9430 }, { "epoch": 0.21, "grad_norm": 2.085285063506428, "learning_rate": 1.9695131669708987e-05, "loss": 2.8703, "step": 9440 }, { "epoch": 0.21, "grad_norm": 1.916543578176851, "learning_rate": 1.9694236172318188e-05, "loss": 2.7358, "step": 9450 }, { "epoch": 0.21, "grad_norm": 2.8377399084690103, "learning_rate": 1.9693339382090362e-05, "loss": 2.777, "step": 9460 }, { "epoch": 0.21, "grad_norm": 2.3036381304504485, "learning_rate": 1.969244129914511e-05, "loss": 2.7715, "step": 9470 }, { "epoch": 0.21, "grad_norm": 2.066241723622545, "learning_rate": 1.969154192360221e-05, "loss": 2.7758, "step": 9480 }, { "epoch": 0.21, "grad_norm": 2.050198641897065, "learning_rate": 1.969064125558159e-05, "loss": 2.6633, "step": 9490 }, { "epoch": 0.21, "grad_norm": 3.2157812330417452, "learning_rate": 1.9689739295203375e-05, "loss": 2.6379, "step": 9500 }, { "epoch": 0.21, "grad_norm": 1.9751022683261426, "learning_rate": 1.968883604258785e-05, "loss": 2.7149, "step": 9510 }, { "epoch": 0.21, "grad_norm": 1.945023771733285, "learning_rate": 1.9687931497855466e-05, "loss": 2.6804, "step": 9520 }, { "epoch": 0.21, "grad_norm": 2.1642606497732704, "learning_rate": 1.9687025661126862e-05, "loss": 2.6699, "step": 9530 }, { "epoch": 0.22, "grad_norm": 1.8465889614126567, "learning_rate": 1.968611853252284e-05, "loss": 2.7505, "step": 9540 }, { "epoch": 0.22, "grad_norm": 2.060699418298319, "learning_rate": 1.9685210112164377e-05, "loss": 2.6547, "step": 9550 }, { "epoch": 0.22, "grad_norm": 3.945121490009671, "learning_rate": 1.968430040017262e-05, "loss": 2.7153, "step": 9560 }, { "epoch": 0.22, "grad_norm": 1.8225377072198161, "learning_rate": 1.9683389396668892e-05, "loss": 2.7589, "step": 9570 }, { "epoch": 0.22, "grad_norm": 2.527509735367638, "learning_rate": 1.968247710177468e-05, "loss": 2.7308, "step": 9580 }, { "epoch": 0.22, "grad_norm": 1.9713354162271155, "learning_rate": 1.968156351561165e-05, "loss": 2.8204, "step": 9590 }, { "epoch": 0.22, "grad_norm": 1.8307056353977704, "learning_rate": 1.9680648638301644e-05, "loss": 2.596, "step": 9600 }, { "epoch": 0.22, "grad_norm": 2.26252532507046, "learning_rate": 1.9679732469966668e-05, "loss": 2.8957, "step": 9610 }, { "epoch": 0.22, "grad_norm": 1.9095834701132657, "learning_rate": 1.9678815010728905e-05, "loss": 2.8985, "step": 9620 }, { "epoch": 0.22, "grad_norm": 2.234000140338573, "learning_rate": 1.9677896260710706e-05, "loss": 2.7047, "step": 9630 }, { "epoch": 0.22, "grad_norm": 2.638291734321406, "learning_rate": 1.96769762200346e-05, "loss": 2.8061, "step": 9640 }, { "epoch": 0.22, "grad_norm": 1.8850076074144158, "learning_rate": 1.9676054888823283e-05, "loss": 2.7464, "step": 9650 }, { "epoch": 0.22, "grad_norm": 1.8581241342886992, "learning_rate": 1.9675132267199628e-05, "loss": 2.7133, "step": 9660 }, { "epoch": 0.22, "grad_norm": 3.8818122392900296, "learning_rate": 1.9674208355286673e-05, "loss": 2.7734, "step": 9670 }, { "epoch": 0.22, "grad_norm": 3.386602882981185, "learning_rate": 1.967328315320763e-05, "loss": 2.7671, "step": 9680 }, { "epoch": 0.22, "grad_norm": 2.341236876868426, "learning_rate": 1.9672356661085893e-05, "loss": 2.6711, "step": 9690 }, { "epoch": 0.22, "grad_norm": 3.4100376842530946, "learning_rate": 1.967142887904502e-05, "loss": 2.8461, "step": 9700 }, { "epoch": 0.22, "grad_norm": 1.9065740636501807, "learning_rate": 1.9670499807208732e-05, "loss": 2.6679, "step": 9710 }, { "epoch": 0.22, "grad_norm": 1.8771555920533032, "learning_rate": 1.966956944570094e-05, "loss": 2.7689, "step": 9720 }, { "epoch": 0.22, "grad_norm": 1.898426179558017, "learning_rate": 1.9668637794645712e-05, "loss": 2.877, "step": 9730 }, { "epoch": 0.22, "grad_norm": 1.98514936656849, "learning_rate": 1.9667704854167305e-05, "loss": 2.799, "step": 9740 }, { "epoch": 0.22, "grad_norm": 1.9761023881752422, "learning_rate": 1.966677062439012e-05, "loss": 2.8341, "step": 9750 }, { "epoch": 0.22, "grad_norm": 2.818980518478803, "learning_rate": 1.9665835105438766e-05, "loss": 2.8118, "step": 9760 }, { "epoch": 0.22, "grad_norm": 4.958467211957016, "learning_rate": 1.9664898297437994e-05, "loss": 2.7644, "step": 9770 }, { "epoch": 0.22, "grad_norm": 3.1589829206906557, "learning_rate": 1.9663960200512737e-05, "loss": 2.7687, "step": 9780 }, { "epoch": 0.22, "grad_norm": 2.0031136748524316, "learning_rate": 1.9663020814788106e-05, "loss": 2.6823, "step": 9790 }, { "epoch": 0.22, "grad_norm": 1.9432532759365029, "learning_rate": 1.9662080140389377e-05, "loss": 2.7663, "step": 9800 }, { "epoch": 0.22, "grad_norm": 1.894269659875283, "learning_rate": 1.9661138177442e-05, "loss": 2.8253, "step": 9810 }, { "epoch": 0.22, "grad_norm": 2.897703714778192, "learning_rate": 1.96601949260716e-05, "loss": 2.8472, "step": 9820 }, { "epoch": 0.22, "grad_norm": 1.7594723024963528, "learning_rate": 1.9659250386403962e-05, "loss": 2.8181, "step": 9830 }, { "epoch": 0.22, "grad_norm": 2.6379340593901195, "learning_rate": 1.9658304558565054e-05, "loss": 2.6916, "step": 9840 }, { "epoch": 0.22, "grad_norm": 1.8289865414652329, "learning_rate": 1.965735744268102e-05, "loss": 2.7323, "step": 9850 }, { "epoch": 0.22, "grad_norm": 1.8572050641502864, "learning_rate": 1.965640903887816e-05, "loss": 2.7421, "step": 9860 }, { "epoch": 0.22, "grad_norm": 2.3297231924760076, "learning_rate": 1.965545934728296e-05, "loss": 2.5862, "step": 9870 }, { "epoch": 0.22, "grad_norm": 2.01674931227531, "learning_rate": 1.9654508368022065e-05, "loss": 2.897, "step": 9880 }, { "epoch": 0.22, "grad_norm": 2.3360513628412503, "learning_rate": 1.9653556101222308e-05, "loss": 2.7545, "step": 9890 }, { "epoch": 0.22, "grad_norm": 3.9614438486053696, "learning_rate": 1.9652602547010683e-05, "loss": 2.8364, "step": 9900 }, { "epoch": 0.22, "grad_norm": 2.0182264634309504, "learning_rate": 1.965164770551435e-05, "loss": 2.7695, "step": 9910 }, { "epoch": 0.22, "grad_norm": 1.9310545832380765, "learning_rate": 1.9650691576860655e-05, "loss": 2.767, "step": 9920 }, { "epoch": 0.22, "grad_norm": 2.074777554794828, "learning_rate": 1.9649734161177108e-05, "loss": 2.756, "step": 9930 }, { "epoch": 0.22, "grad_norm": 1.7864673752491895, "learning_rate": 1.964877545859139e-05, "loss": 2.8251, "step": 9940 }, { "epoch": 0.22, "grad_norm": 1.7894201073701457, "learning_rate": 1.9647815469231353e-05, "loss": 2.7976, "step": 9950 }, { "epoch": 0.22, "grad_norm": 2.193305977941157, "learning_rate": 1.9646854193225025e-05, "loss": 2.8435, "step": 9960 }, { "epoch": 0.22, "grad_norm": 2.4485678344812585, "learning_rate": 1.9645891630700604e-05, "loss": 2.8565, "step": 9970 }, { "epoch": 0.23, "grad_norm": 4.3207767563538475, "learning_rate": 1.9644927781786456e-05, "loss": 2.6951, "step": 9980 }, { "epoch": 0.23, "grad_norm": 1.8656021995614318, "learning_rate": 1.964396264661112e-05, "loss": 2.7616, "step": 9990 }, { "epoch": 0.23, "grad_norm": 2.8820334190046495, "learning_rate": 1.9642996225303314e-05, "loss": 2.7582, "step": 10000 }, { "epoch": 0.23, "grad_norm": 2.3542875044737155, "learning_rate": 1.9642028517991916e-05, "loss": 2.8006, "step": 10010 }, { "epoch": 0.23, "grad_norm": 2.2049195609706573, "learning_rate": 1.9641059524805987e-05, "loss": 2.7703, "step": 10020 }, { "epoch": 0.23, "grad_norm": 2.7900153950440405, "learning_rate": 1.964008924587474e-05, "loss": 2.7177, "step": 10030 }, { "epoch": 0.23, "grad_norm": 2.2823744583974244, "learning_rate": 1.963911768132759e-05, "loss": 2.7639, "step": 10040 }, { "epoch": 0.23, "grad_norm": 1.9281083578829903, "learning_rate": 1.9638144831294094e-05, "loss": 2.7915, "step": 10050 }, { "epoch": 0.23, "grad_norm": 2.0652397742826265, "learning_rate": 1.9637170695903994e-05, "loss": 2.7421, "step": 10060 }, { "epoch": 0.23, "grad_norm": 2.208228472036979, "learning_rate": 1.9636195275287205e-05, "loss": 2.8291, "step": 10070 }, { "epoch": 0.23, "grad_norm": 2.0172331844051246, "learning_rate": 1.9635218569573814e-05, "loss": 2.7581, "step": 10080 }, { "epoch": 0.23, "grad_norm": 1.9060968915584477, "learning_rate": 1.9634240578894068e-05, "loss": 2.6234, "step": 10090 }, { "epoch": 0.23, "grad_norm": 1.9225392871253482, "learning_rate": 1.9633261303378396e-05, "loss": 2.8338, "step": 10100 }, { "epoch": 0.23, "grad_norm": 2.1491615393524137, "learning_rate": 1.96322807431574e-05, "loss": 2.81, "step": 10110 }, { "epoch": 0.23, "grad_norm": 2.0244713980427504, "learning_rate": 1.9631298898361847e-05, "loss": 2.6182, "step": 10120 }, { "epoch": 0.23, "grad_norm": 2.1726041743621836, "learning_rate": 1.963031576912267e-05, "loss": 2.7879, "step": 10130 }, { "epoch": 0.23, "grad_norm": 2.88562679640263, "learning_rate": 1.962933135557099e-05, "loss": 2.8523, "step": 10140 }, { "epoch": 0.23, "grad_norm": 1.8195082864311165, "learning_rate": 1.9628345657838085e-05, "loss": 2.8332, "step": 10150 }, { "epoch": 0.23, "grad_norm": 1.7570776820233134, "learning_rate": 1.962735867605541e-05, "loss": 2.6548, "step": 10160 }, { "epoch": 0.23, "grad_norm": 2.762546105230515, "learning_rate": 1.962637041035459e-05, "loss": 2.8289, "step": 10170 }, { "epoch": 0.23, "grad_norm": 1.9654061759704817, "learning_rate": 1.9625380860867423e-05, "loss": 2.8075, "step": 10180 }, { "epoch": 0.23, "grad_norm": 1.8591912510000292, "learning_rate": 1.9624390027725876e-05, "loss": 2.7378, "step": 10190 }, { "epoch": 0.23, "grad_norm": 2.001177695138059, "learning_rate": 1.9623397911062083e-05, "loss": 2.784, "step": 10200 }, { "epoch": 0.23, "grad_norm": 2.075943411254483, "learning_rate": 1.962240451100836e-05, "loss": 2.7826, "step": 10210 }, { "epoch": 0.23, "grad_norm": 2.458126259866497, "learning_rate": 1.962140982769719e-05, "loss": 2.7282, "step": 10220 }, { "epoch": 0.23, "grad_norm": 1.8826828675436176, "learning_rate": 1.9620413861261222e-05, "loss": 2.8245, "step": 10230 }, { "epoch": 0.23, "grad_norm": 1.9693431880050636, "learning_rate": 1.9619416611833275e-05, "loss": 2.8223, "step": 10240 }, { "epoch": 0.23, "grad_norm": 2.4088200328468745, "learning_rate": 1.9618418079546352e-05, "loss": 2.7842, "step": 10250 }, { "epoch": 0.23, "grad_norm": 1.834132608932584, "learning_rate": 1.9617418264533613e-05, "loss": 2.7656, "step": 10260 }, { "epoch": 0.23, "grad_norm": 2.0884433697393527, "learning_rate": 1.9616417166928395e-05, "loss": 2.7377, "step": 10270 }, { "epoch": 0.23, "grad_norm": 2.4014923478950307, "learning_rate": 1.961541478686421e-05, "loss": 2.805, "step": 10280 }, { "epoch": 0.23, "grad_norm": 3.059720009007472, "learning_rate": 1.9614411124474734e-05, "loss": 2.714, "step": 10290 }, { "epoch": 0.23, "grad_norm": 1.9860113050500283, "learning_rate": 1.9613406179893817e-05, "loss": 2.7656, "step": 10300 }, { "epoch": 0.23, "grad_norm": 3.7804085590736465, "learning_rate": 1.961239995325548e-05, "loss": 2.7835, "step": 10310 }, { "epoch": 0.23, "grad_norm": 2.1799722315861674, "learning_rate": 1.961139244469391e-05, "loss": 2.7151, "step": 10320 }, { "epoch": 0.23, "grad_norm": 2.022843765782694, "learning_rate": 1.961038365434348e-05, "loss": 2.7551, "step": 10330 }, { "epoch": 0.23, "grad_norm": 2.5816033712240314, "learning_rate": 1.960937358233872e-05, "loss": 2.7276, "step": 10340 }, { "epoch": 0.23, "grad_norm": 2.6277408109890636, "learning_rate": 1.9608362228814326e-05, "loss": 2.803, "step": 10350 }, { "epoch": 0.23, "grad_norm": 2.0532421294387793, "learning_rate": 1.9607349593905186e-05, "loss": 2.8041, "step": 10360 }, { "epoch": 0.23, "grad_norm": 1.9092105690907386, "learning_rate": 1.9606335677746336e-05, "loss": 2.612, "step": 10370 }, { "epoch": 0.23, "grad_norm": 2.255234665950484, "learning_rate": 1.9605320480473003e-05, "loss": 2.7828, "step": 10380 }, { "epoch": 0.23, "grad_norm": 2.244071043197074, "learning_rate": 1.9604304002220567e-05, "loss": 2.7879, "step": 10390 }, { "epoch": 0.23, "grad_norm": 2.5750312851080936, "learning_rate": 1.960328624312459e-05, "loss": 2.8042, "step": 10400 }, { "epoch": 0.23, "grad_norm": 2.4730883498517073, "learning_rate": 1.9602267203320803e-05, "loss": 2.854, "step": 10410 }, { "epoch": 0.23, "grad_norm": 2.001887834268218, "learning_rate": 1.9601246882945105e-05, "loss": 2.6377, "step": 10420 }, { "epoch": 0.24, "grad_norm": 2.5318634798760757, "learning_rate": 1.9600225282133567e-05, "loss": 2.8495, "step": 10430 }, { "epoch": 0.24, "grad_norm": 2.174523796420796, "learning_rate": 1.9599202401022432e-05, "loss": 2.8629, "step": 10440 }, { "epoch": 0.24, "grad_norm": 2.3008377097020447, "learning_rate": 1.9598178239748113e-05, "loss": 2.7618, "step": 10450 }, { "epoch": 0.24, "grad_norm": 2.322639582420326, "learning_rate": 1.9597152798447196e-05, "loss": 2.6794, "step": 10460 }, { "epoch": 0.24, "grad_norm": 2.5639652158769186, "learning_rate": 1.959612607725643e-05, "loss": 2.661, "step": 10470 }, { "epoch": 0.24, "grad_norm": 1.9155393897479793, "learning_rate": 1.959509807631274e-05, "loss": 2.8078, "step": 10480 }, { "epoch": 0.24, "grad_norm": 1.997496657399573, "learning_rate": 1.9594068795753232e-05, "loss": 2.6979, "step": 10490 }, { "epoch": 0.24, "grad_norm": 2.194310224084581, "learning_rate": 1.959303823571516e-05, "loss": 2.8118, "step": 10500 }, { "epoch": 0.24, "grad_norm": 2.6953080005217394, "learning_rate": 1.9592006396335967e-05, "loss": 2.7095, "step": 10510 }, { "epoch": 0.24, "grad_norm": 2.723378118479027, "learning_rate": 1.9590973277753256e-05, "loss": 2.7377, "step": 10520 }, { "epoch": 0.24, "grad_norm": 1.6432964945444346, "learning_rate": 1.9589938880104814e-05, "loss": 2.7312, "step": 10530 }, { "epoch": 0.24, "grad_norm": 2.663437102485018, "learning_rate": 1.958890320352858e-05, "loss": 2.8224, "step": 10540 }, { "epoch": 0.24, "grad_norm": 2.0419137197947146, "learning_rate": 1.9587866248162676e-05, "loss": 2.7595, "step": 10550 }, { "epoch": 0.24, "grad_norm": 1.820184898514923, "learning_rate": 1.9586828014145396e-05, "loss": 2.6633, "step": 10560 }, { "epoch": 0.24, "grad_norm": 2.7681727159206844, "learning_rate": 1.9585788501615197e-05, "loss": 2.6834, "step": 10570 }, { "epoch": 0.24, "grad_norm": 1.7573716200722844, "learning_rate": 1.958474771071071e-05, "loss": 2.8302, "step": 10580 }, { "epoch": 0.24, "grad_norm": 2.3115794959697453, "learning_rate": 1.9583705641570737e-05, "loss": 2.7215, "step": 10590 }, { "epoch": 0.24, "grad_norm": 3.246785437141259, "learning_rate": 1.958266229433425e-05, "loss": 2.8128, "step": 10600 }, { "epoch": 0.24, "grad_norm": 2.4066436990062585, "learning_rate": 1.9581617669140388e-05, "loss": 2.6747, "step": 10610 }, { "epoch": 0.24, "grad_norm": 1.9330215478083879, "learning_rate": 1.9580571766128467e-05, "loss": 2.7928, "step": 10620 }, { "epoch": 0.24, "grad_norm": 2.0280869624579116, "learning_rate": 1.9579524585437975e-05, "loss": 2.8065, "step": 10630 }, { "epoch": 0.24, "grad_norm": 1.8547668849377257, "learning_rate": 1.9578476127208557e-05, "loss": 2.7616, "step": 10640 }, { "epoch": 0.24, "grad_norm": 1.8198724656214709, "learning_rate": 1.957742639158004e-05, "loss": 2.7319, "step": 10650 }, { "epoch": 0.24, "grad_norm": 2.2073370547932356, "learning_rate": 1.9576375378692414e-05, "loss": 2.7487, "step": 10660 }, { "epoch": 0.24, "grad_norm": 1.9031721096114789, "learning_rate": 1.9575323088685852e-05, "loss": 2.8817, "step": 10670 }, { "epoch": 0.24, "grad_norm": 3.230919965478331, "learning_rate": 1.9574269521700686e-05, "loss": 2.7243, "step": 10680 }, { "epoch": 0.24, "grad_norm": 2.5060234210512387, "learning_rate": 1.9573214677877418e-05, "loss": 2.7193, "step": 10690 }, { "epoch": 0.24, "grad_norm": 2.597925496423927, "learning_rate": 1.9572158557356723e-05, "loss": 2.7885, "step": 10700 }, { "epoch": 0.24, "grad_norm": 2.7918568644286514, "learning_rate": 1.9571101160279453e-05, "loss": 2.6535, "step": 10710 }, { "epoch": 0.24, "grad_norm": 1.8198064842383384, "learning_rate": 1.957004248678662e-05, "loss": 2.7317, "step": 10720 }, { "epoch": 0.24, "grad_norm": 1.9640690205709614, "learning_rate": 1.9568982537019406e-05, "loss": 2.7342, "step": 10730 }, { "epoch": 0.24, "grad_norm": 4.0099536141573715, "learning_rate": 1.9567921311119176e-05, "loss": 2.7815, "step": 10740 }, { "epoch": 0.24, "grad_norm": 1.8948189391163555, "learning_rate": 1.956685880922745e-05, "loss": 2.7182, "step": 10750 }, { "epoch": 0.24, "grad_norm": 2.165485724365357, "learning_rate": 1.9565795031485927e-05, "loss": 2.6933, "step": 10760 }, { "epoch": 0.24, "grad_norm": 2.2052072297207324, "learning_rate": 1.956472997803647e-05, "loss": 2.7627, "step": 10770 }, { "epoch": 0.24, "grad_norm": 1.7792363941031735, "learning_rate": 1.9563663649021126e-05, "loss": 2.8788, "step": 10780 }, { "epoch": 0.24, "grad_norm": 1.5901296380963261, "learning_rate": 1.9562596044582096e-05, "loss": 2.7674, "step": 10790 }, { "epoch": 0.24, "grad_norm": 2.499347793214378, "learning_rate": 1.9561527164861752e-05, "loss": 2.7627, "step": 10800 }, { "epoch": 0.24, "grad_norm": 1.7054432933619155, "learning_rate": 1.9560457010002652e-05, "loss": 2.7166, "step": 10810 }, { "epoch": 0.24, "grad_norm": 2.544522224151244, "learning_rate": 1.9559385580147502e-05, "loss": 2.7617, "step": 10820 }, { "epoch": 0.24, "grad_norm": 1.9468702086270415, "learning_rate": 1.95583128754392e-05, "loss": 2.7863, "step": 10830 }, { "epoch": 0.24, "grad_norm": 2.091899215913959, "learning_rate": 1.9557238896020797e-05, "loss": 2.7048, "step": 10840 }, { "epoch": 0.24, "grad_norm": 1.8732233637575413, "learning_rate": 1.9556163642035524e-05, "loss": 2.8265, "step": 10850 }, { "epoch": 0.24, "grad_norm": 1.8589452725718163, "learning_rate": 1.9555087113626773e-05, "loss": 2.721, "step": 10860 }, { "epoch": 0.25, "grad_norm": 2.6991405092522114, "learning_rate": 1.9554009310938118e-05, "loss": 2.6718, "step": 10870 }, { "epoch": 0.25, "grad_norm": 2.4863861392968274, "learning_rate": 1.9552930234113292e-05, "loss": 2.7575, "step": 10880 }, { "epoch": 0.25, "grad_norm": 2.913569708769738, "learning_rate": 1.9551849883296202e-05, "loss": 2.7375, "step": 10890 }, { "epoch": 0.25, "grad_norm": 2.1909761111301296, "learning_rate": 1.9550768258630928e-05, "loss": 2.7639, "step": 10900 }, { "epoch": 0.25, "grad_norm": 2.125095543625551, "learning_rate": 1.9549685360261716e-05, "loss": 2.7543, "step": 10910 }, { "epoch": 0.25, "grad_norm": 2.485483850560963, "learning_rate": 1.9548601188332982e-05, "loss": 2.8132, "step": 10920 }, { "epoch": 0.25, "grad_norm": 1.9058523902307527, "learning_rate": 1.9547515742989312e-05, "loss": 2.7345, "step": 10930 }, { "epoch": 0.25, "grad_norm": 1.9862822606855741, "learning_rate": 1.9546429024375463e-05, "loss": 2.724, "step": 10940 }, { "epoch": 0.25, "grad_norm": 4.2044450643358955, "learning_rate": 1.9545341032636362e-05, "loss": 2.7921, "step": 10950 }, { "epoch": 0.25, "grad_norm": 2.308759459578628, "learning_rate": 1.9544251767917107e-05, "loss": 2.7437, "step": 10960 }, { "epoch": 0.25, "grad_norm": 1.9377579419428057, "learning_rate": 1.954316123036296e-05, "loss": 2.8313, "step": 10970 }, { "epoch": 0.25, "grad_norm": 1.8422281060237176, "learning_rate": 1.954206942011936e-05, "loss": 2.7316, "step": 10980 }, { "epoch": 0.25, "grad_norm": 2.1252795036130676, "learning_rate": 1.954097633733191e-05, "loss": 2.5678, "step": 10990 }, { "epoch": 0.25, "grad_norm": 1.8974864872201767, "learning_rate": 1.9539881982146385e-05, "loss": 2.7072, "step": 11000 }, { "epoch": 0.25, "grad_norm": 2.1799631703819413, "learning_rate": 1.953878635470873e-05, "loss": 2.8075, "step": 11010 }, { "epoch": 0.25, "grad_norm": 2.402902335881672, "learning_rate": 1.9537689455165064e-05, "loss": 2.7122, "step": 11020 }, { "epoch": 0.25, "grad_norm": 2.1516033324749895, "learning_rate": 1.9536591283661665e-05, "loss": 2.7412, "step": 11030 }, { "epoch": 0.25, "grad_norm": 2.2408431788588095, "learning_rate": 1.9535491840344987e-05, "loss": 2.7416, "step": 11040 }, { "epoch": 0.25, "grad_norm": 2.052994976617454, "learning_rate": 1.9534391125361658e-05, "loss": 2.787, "step": 11050 }, { "epoch": 0.25, "grad_norm": 1.9304134185751305, "learning_rate": 1.9533289138858466e-05, "loss": 2.8372, "step": 11060 }, { "epoch": 0.25, "grad_norm": 2.464313492638104, "learning_rate": 1.9532185880982375e-05, "loss": 2.7943, "step": 11070 }, { "epoch": 0.25, "grad_norm": 2.3952818074881383, "learning_rate": 1.953108135188052e-05, "loss": 2.7821, "step": 11080 }, { "epoch": 0.25, "grad_norm": 2.5532108361789723, "learning_rate": 1.9529975551700198e-05, "loss": 2.8, "step": 11090 }, { "epoch": 0.25, "grad_norm": 2.640430462095355, "learning_rate": 1.9528868480588884e-05, "loss": 2.799, "step": 11100 }, { "epoch": 0.25, "grad_norm": 2.1830289979438486, "learning_rate": 1.9527760138694215e-05, "loss": 2.794, "step": 11110 }, { "epoch": 0.25, "grad_norm": 2.6792273831501623, "learning_rate": 1.9526650526164003e-05, "loss": 2.7232, "step": 11120 }, { "epoch": 0.25, "grad_norm": 2.0457028713046905, "learning_rate": 1.952553964314623e-05, "loss": 2.7154, "step": 11130 }, { "epoch": 0.25, "grad_norm": 2.9365973951927304, "learning_rate": 1.952442748978904e-05, "loss": 2.8097, "step": 11140 }, { "epoch": 0.25, "grad_norm": 2.2177868078948797, "learning_rate": 1.9523314066240753e-05, "loss": 2.747, "step": 11150 }, { "epoch": 0.25, "grad_norm": 1.9723242882302456, "learning_rate": 1.9522199372649856e-05, "loss": 2.7686, "step": 11160 }, { "epoch": 0.25, "grad_norm": 2.1363995522472172, "learning_rate": 1.952108340916501e-05, "loss": 2.8277, "step": 11170 }, { "epoch": 0.25, "grad_norm": 2.5963137045438254, "learning_rate": 1.9519966175935038e-05, "loss": 2.6642, "step": 11180 }, { "epoch": 0.25, "grad_norm": 2.969214359096815, "learning_rate": 1.9518847673108937e-05, "loss": 2.7421, "step": 11190 }, { "epoch": 0.25, "grad_norm": 3.2130076495105673, "learning_rate": 1.9517727900835875e-05, "loss": 2.8335, "step": 11200 }, { "epoch": 0.25, "grad_norm": 1.8979769681647807, "learning_rate": 1.951660685926518e-05, "loss": 2.7333, "step": 11210 }, { "epoch": 0.25, "grad_norm": 1.7597478515491518, "learning_rate": 1.9515484548546356e-05, "loss": 2.8486, "step": 11220 }, { "epoch": 0.25, "grad_norm": 1.983130076669053, "learning_rate": 1.9514360968829083e-05, "loss": 2.8454, "step": 11230 }, { "epoch": 0.25, "grad_norm": 2.8137730837928694, "learning_rate": 1.9513236120263198e-05, "loss": 2.8703, "step": 11240 }, { "epoch": 0.25, "grad_norm": 2.061176140828701, "learning_rate": 1.9512110002998713e-05, "loss": 2.7658, "step": 11250 }, { "epoch": 0.25, "grad_norm": 2.48948011740651, "learning_rate": 1.951098261718581e-05, "loss": 2.8017, "step": 11260 }, { "epoch": 0.25, "grad_norm": 1.9826013282296289, "learning_rate": 1.9509853962974836e-05, "loss": 2.8142, "step": 11270 }, { "epoch": 0.25, "grad_norm": 2.22680537866824, "learning_rate": 1.9508724040516314e-05, "loss": 2.842, "step": 11280 }, { "epoch": 0.25, "grad_norm": 2.5634161182690374, "learning_rate": 1.9507592849960928e-05, "loss": 2.6549, "step": 11290 }, { "epoch": 0.25, "grad_norm": 1.970858358733636, "learning_rate": 1.9506460391459537e-05, "loss": 2.7383, "step": 11300 }, { "epoch": 0.26, "grad_norm": 2.0135047503389476, "learning_rate": 1.9505326665163168e-05, "loss": 2.7831, "step": 11310 }, { "epoch": 0.26, "grad_norm": 2.087541934286038, "learning_rate": 1.9504191671223016e-05, "loss": 2.7162, "step": 11320 }, { "epoch": 0.26, "grad_norm": 2.2117768562100086, "learning_rate": 1.950305540979044e-05, "loss": 2.7109, "step": 11330 }, { "epoch": 0.26, "grad_norm": 2.4761889078310015, "learning_rate": 1.9501917881016987e-05, "loss": 2.7964, "step": 11340 }, { "epoch": 0.26, "grad_norm": 1.7925592573528104, "learning_rate": 1.9500779085054343e-05, "loss": 2.7753, "step": 11350 }, { "epoch": 0.26, "grad_norm": 2.0830218815835386, "learning_rate": 1.9499639022054392e-05, "loss": 2.8281, "step": 11360 }, { "epoch": 0.26, "grad_norm": 2.3282666117243176, "learning_rate": 1.9498497692169166e-05, "loss": 2.7631, "step": 11370 }, { "epoch": 0.26, "grad_norm": 1.8327407088491532, "learning_rate": 1.949735509555088e-05, "loss": 2.6373, "step": 11380 }, { "epoch": 0.26, "grad_norm": 1.7548984812781492, "learning_rate": 1.949621123235191e-05, "loss": 2.7958, "step": 11390 }, { "epoch": 0.26, "grad_norm": 2.305293965497714, "learning_rate": 1.9495066102724804e-05, "loss": 2.6859, "step": 11400 }, { "epoch": 0.26, "grad_norm": 2.8868830796248623, "learning_rate": 1.9493919706822278e-05, "loss": 2.6788, "step": 11410 }, { "epoch": 0.26, "grad_norm": 2.5488382962914957, "learning_rate": 1.9492772044797215e-05, "loss": 2.7971, "step": 11420 }, { "epoch": 0.26, "grad_norm": 2.4882555146649548, "learning_rate": 1.949162311680267e-05, "loss": 2.7872, "step": 11430 }, { "epoch": 0.26, "grad_norm": 1.8558282014267242, "learning_rate": 1.949047292299187e-05, "loss": 2.7849, "step": 11440 }, { "epoch": 0.26, "grad_norm": 2.227750807854553, "learning_rate": 1.9489321463518198e-05, "loss": 2.7319, "step": 11450 }, { "epoch": 0.26, "grad_norm": 2.0233229438790357, "learning_rate": 1.948816873853522e-05, "loss": 2.6979, "step": 11460 }, { "epoch": 0.26, "grad_norm": 1.9235973101543722, "learning_rate": 1.9487014748196664e-05, "loss": 2.6038, "step": 11470 }, { "epoch": 0.26, "grad_norm": 1.8143831596331892, "learning_rate": 1.948585949265643e-05, "loss": 2.8149, "step": 11480 }, { "epoch": 0.26, "grad_norm": 2.080360521254768, "learning_rate": 1.948470297206858e-05, "loss": 2.7612, "step": 11490 }, { "epoch": 0.26, "grad_norm": 2.3508258202458125, "learning_rate": 1.948354518658735e-05, "loss": 2.8926, "step": 11500 }, { "epoch": 0.26, "grad_norm": 3.3677520975383968, "learning_rate": 1.9482386136367148e-05, "loss": 2.758, "step": 11510 }, { "epoch": 0.26, "grad_norm": 2.906583309024663, "learning_rate": 1.9481225821562542e-05, "loss": 2.8248, "step": 11520 }, { "epoch": 0.26, "grad_norm": 2.1505143520321552, "learning_rate": 1.9480064242328278e-05, "loss": 2.906, "step": 11530 }, { "epoch": 0.26, "grad_norm": 2.0964653251453047, "learning_rate": 1.947890139881926e-05, "loss": 2.8219, "step": 11540 }, { "epoch": 0.26, "grad_norm": 2.5602174103146513, "learning_rate": 1.9477737291190567e-05, "loss": 2.6752, "step": 11550 }, { "epoch": 0.26, "grad_norm": 2.617390769901438, "learning_rate": 1.947657191959745e-05, "loss": 2.7088, "step": 11560 }, { "epoch": 0.26, "grad_norm": 2.0373842840235326, "learning_rate": 1.9475405284195322e-05, "loss": 2.7172, "step": 11570 }, { "epoch": 0.26, "grad_norm": 1.6385924968439423, "learning_rate": 1.9474237385139774e-05, "loss": 2.8172, "step": 11580 }, { "epoch": 0.26, "grad_norm": 2.126642215552026, "learning_rate": 1.9473068222586548e-05, "loss": 2.748, "step": 11590 }, { "epoch": 0.26, "grad_norm": 1.9002127373452484, "learning_rate": 1.947189779669157e-05, "loss": 2.7424, "step": 11600 }, { "epoch": 0.26, "grad_norm": 2.5593131940921117, "learning_rate": 1.9470726107610932e-05, "loss": 2.7557, "step": 11610 }, { "epoch": 0.26, "grad_norm": 2.1633502296669267, "learning_rate": 1.946955315550089e-05, "loss": 2.7981, "step": 11620 }, { "epoch": 0.26, "grad_norm": 2.8836955921686434, "learning_rate": 1.9468378940517864e-05, "loss": 2.72, "step": 11630 }, { "epoch": 0.26, "grad_norm": 3.984233516797135, "learning_rate": 1.946720346281846e-05, "loss": 2.7823, "step": 11640 }, { "epoch": 0.26, "grad_norm": 2.0189522465166903, "learning_rate": 1.946602672255944e-05, "loss": 2.7257, "step": 11650 }, { "epoch": 0.26, "grad_norm": 2.219509366988483, "learning_rate": 1.9464848719897725e-05, "loss": 2.8121, "step": 11660 }, { "epoch": 0.26, "grad_norm": 1.8852002326319848, "learning_rate": 1.9463669454990426e-05, "loss": 2.7605, "step": 11670 }, { "epoch": 0.26, "grad_norm": 2.076213353550997, "learning_rate": 1.946248892799481e-05, "loss": 2.7924, "step": 11680 }, { "epoch": 0.26, "grad_norm": 2.077944120475819, "learning_rate": 1.9461307139068307e-05, "loss": 2.7481, "step": 11690 }, { "epoch": 0.26, "grad_norm": 1.8178991813321592, "learning_rate": 1.946012408836853e-05, "loss": 2.8099, "step": 11700 }, { "epoch": 0.26, "grad_norm": 2.505135976889072, "learning_rate": 1.9458939776053247e-05, "loss": 2.686, "step": 11710 }, { "epoch": 0.26, "grad_norm": 2.521043708848944, "learning_rate": 1.9457754202280402e-05, "loss": 2.8813, "step": 11720 }, { "epoch": 0.26, "grad_norm": 2.597366487741212, "learning_rate": 1.9456567367208108e-05, "loss": 2.7963, "step": 11730 }, { "epoch": 0.26, "grad_norm": 3.0268108663116498, "learning_rate": 1.9455379270994635e-05, "loss": 2.7774, "step": 11740 }, { "epoch": 0.26, "grad_norm": 2.7856857022692023, "learning_rate": 1.9454189913798436e-05, "loss": 2.8575, "step": 11750 }, { "epoch": 0.27, "grad_norm": 1.713124019473379, "learning_rate": 1.945299929577812e-05, "loss": 2.8842, "step": 11760 }, { "epoch": 0.27, "grad_norm": 2.8201673306578, "learning_rate": 1.9451807417092476e-05, "loss": 2.686, "step": 11770 }, { "epoch": 0.27, "grad_norm": 3.7862261833433903, "learning_rate": 1.945061427790045e-05, "loss": 2.6256, "step": 11780 }, { "epoch": 0.27, "grad_norm": 4.542408668699285, "learning_rate": 1.9449419878361158e-05, "loss": 2.815, "step": 11790 }, { "epoch": 0.27, "grad_norm": 1.9181190733338473, "learning_rate": 1.9448224218633895e-05, "loss": 2.754, "step": 11800 }, { "epoch": 0.27, "grad_norm": 1.9856679627707106, "learning_rate": 1.944702729887811e-05, "loss": 2.7302, "step": 11810 }, { "epoch": 0.27, "grad_norm": 2.4310886460206476, "learning_rate": 1.9445829119253425e-05, "loss": 2.7398, "step": 11820 }, { "epoch": 0.27, "grad_norm": 1.972844939489742, "learning_rate": 1.9444629679919635e-05, "loss": 2.708, "step": 11830 }, { "epoch": 0.27, "grad_norm": 2.091759730077857, "learning_rate": 1.9443428981036698e-05, "loss": 2.8354, "step": 11840 }, { "epoch": 0.27, "grad_norm": 1.861758903545625, "learning_rate": 1.9442227022764738e-05, "loss": 2.7141, "step": 11850 }, { "epoch": 0.27, "grad_norm": 2.179407083191411, "learning_rate": 1.9441023805264054e-05, "loss": 2.6794, "step": 11860 }, { "epoch": 0.27, "grad_norm": 1.8267713004816848, "learning_rate": 1.9439819328695108e-05, "loss": 2.7885, "step": 11870 }, { "epoch": 0.27, "grad_norm": 2.1270386117544176, "learning_rate": 1.9438613593218525e-05, "loss": 2.7714, "step": 11880 }, { "epoch": 0.27, "grad_norm": 2.9617768965293663, "learning_rate": 1.943740659899511e-05, "loss": 2.776, "step": 11890 }, { "epoch": 0.27, "grad_norm": 2.0001831616554986, "learning_rate": 1.9436198346185828e-05, "loss": 2.9347, "step": 11900 }, { "epoch": 0.27, "grad_norm": 2.1428472354859807, "learning_rate": 1.943498883495181e-05, "loss": 2.7649, "step": 11910 }, { "epoch": 0.27, "grad_norm": 2.0570023253205987, "learning_rate": 1.9433778065454366e-05, "loss": 2.74, "step": 11920 }, { "epoch": 0.27, "grad_norm": 1.9013973007019245, "learning_rate": 1.9432566037854957e-05, "loss": 2.666, "step": 11930 }, { "epoch": 0.27, "grad_norm": 1.976976028873789, "learning_rate": 1.9431352752315226e-05, "loss": 2.7012, "step": 11940 }, { "epoch": 0.27, "grad_norm": 1.8840713344722726, "learning_rate": 1.9430138208996977e-05, "loss": 2.7732, "step": 11950 }, { "epoch": 0.27, "grad_norm": 2.9796135868896694, "learning_rate": 1.9428922408062184e-05, "loss": 2.7324, "step": 11960 }, { "epoch": 0.27, "grad_norm": 2.0039207098341723, "learning_rate": 1.9427705349672986e-05, "loss": 2.8098, "step": 11970 }, { "epoch": 0.27, "grad_norm": 1.800758280786929, "learning_rate": 1.942648703399169e-05, "loss": 2.7831, "step": 11980 }, { "epoch": 0.27, "grad_norm": 1.9691168319288923, "learning_rate": 1.9425267461180778e-05, "loss": 2.8317, "step": 11990 }, { "epoch": 0.27, "grad_norm": 2.262699059455376, "learning_rate": 1.942404663140289e-05, "loss": 2.7235, "step": 12000 }, { "epoch": 0.27, "grad_norm": 2.129091397748688, "learning_rate": 1.942282454482084e-05, "loss": 2.6903, "step": 12010 }, { "epoch": 0.27, "grad_norm": 2.3461273164704846, "learning_rate": 1.9421601201597606e-05, "loss": 2.8101, "step": 12020 }, { "epoch": 0.27, "grad_norm": 2.1226002161051225, "learning_rate": 1.9420376601896333e-05, "loss": 2.7238, "step": 12030 }, { "epoch": 0.27, "grad_norm": 2.9644533530365473, "learning_rate": 1.9419150745880336e-05, "loss": 2.6971, "step": 12040 }, { "epoch": 0.27, "grad_norm": 3.034007806185856, "learning_rate": 1.94179236337131e-05, "loss": 2.7338, "step": 12050 }, { "epoch": 0.27, "grad_norm": 2.0175176393529304, "learning_rate": 1.941669526555827e-05, "loss": 2.7355, "step": 12060 }, { "epoch": 0.27, "grad_norm": 1.865898762301003, "learning_rate": 1.9415465641579666e-05, "loss": 2.7299, "step": 12070 }, { "epoch": 0.27, "grad_norm": 2.0675995564880556, "learning_rate": 1.9414234761941272e-05, "loss": 2.6049, "step": 12080 }, { "epoch": 0.27, "grad_norm": 2.462373644364351, "learning_rate": 1.9413002626807238e-05, "loss": 2.8101, "step": 12090 }, { "epoch": 0.27, "grad_norm": 2.922523773197924, "learning_rate": 1.9411769236341885e-05, "loss": 2.7396, "step": 12100 }, { "epoch": 0.27, "grad_norm": 2.5529836424008816, "learning_rate": 1.94105345907097e-05, "loss": 2.7305, "step": 12110 }, { "epoch": 0.27, "grad_norm": 2.318514561235473, "learning_rate": 1.940929869007534e-05, "loss": 2.7323, "step": 12120 }, { "epoch": 0.27, "grad_norm": 1.8240360409573875, "learning_rate": 1.9408061534603616e-05, "loss": 2.8358, "step": 12130 }, { "epoch": 0.27, "grad_norm": 2.229343760079257, "learning_rate": 1.9406823124459526e-05, "loss": 2.7624, "step": 12140 }, { "epoch": 0.27, "grad_norm": 2.263569044427036, "learning_rate": 1.9405583459808227e-05, "loss": 2.7009, "step": 12150 }, { "epoch": 0.27, "grad_norm": 1.848558462472167, "learning_rate": 1.9404342540815034e-05, "loss": 2.7164, "step": 12160 }, { "epoch": 0.27, "grad_norm": 1.8046914437175468, "learning_rate": 1.9403100367645446e-05, "loss": 2.7189, "step": 12170 }, { "epoch": 0.27, "grad_norm": 2.1141332133631106, "learning_rate": 1.9401856940465117e-05, "loss": 2.8355, "step": 12180 }, { "epoch": 0.27, "grad_norm": 2.1852039108130485, "learning_rate": 1.9400612259439874e-05, "loss": 2.7986, "step": 12190 }, { "epoch": 0.28, "grad_norm": 3.0316178507937486, "learning_rate": 1.9399366324735708e-05, "loss": 2.7306, "step": 12200 }, { "epoch": 0.28, "grad_norm": 1.940923557653609, "learning_rate": 1.9398119136518782e-05, "loss": 2.7465, "step": 12210 }, { "epoch": 0.28, "grad_norm": 2.299704362604376, "learning_rate": 1.9396870694955414e-05, "loss": 2.6549, "step": 12220 }, { "epoch": 0.28, "grad_norm": 2.369159047126358, "learning_rate": 1.9395621000212106e-05, "loss": 2.8279, "step": 12230 }, { "epoch": 0.28, "grad_norm": 2.0404620996861387, "learning_rate": 1.939437005245552e-05, "loss": 2.7194, "step": 12240 }, { "epoch": 0.28, "grad_norm": 1.9122144994122758, "learning_rate": 1.9393117851852475e-05, "loss": 2.8087, "step": 12250 }, { "epoch": 0.28, "grad_norm": 1.9847778663351214, "learning_rate": 1.939186439856998e-05, "loss": 2.6134, "step": 12260 }, { "epoch": 0.28, "grad_norm": 1.9959285896910846, "learning_rate": 1.9390609692775185e-05, "loss": 2.593, "step": 12270 }, { "epoch": 0.28, "grad_norm": 1.9351782941865794, "learning_rate": 1.938935373463542e-05, "loss": 2.7823, "step": 12280 }, { "epoch": 0.28, "grad_norm": 1.9687214185307729, "learning_rate": 1.9388096524318193e-05, "loss": 2.8403, "step": 12290 }, { "epoch": 0.28, "grad_norm": 1.9363722537606098, "learning_rate": 1.9386838061991156e-05, "loss": 2.6519, "step": 12300 }, { "epoch": 0.28, "grad_norm": 1.8321615373440234, "learning_rate": 1.9385578347822143e-05, "loss": 2.5853, "step": 12310 }, { "epoch": 0.28, "grad_norm": 2.0411027938101856, "learning_rate": 1.938431738197915e-05, "loss": 2.7206, "step": 12320 }, { "epoch": 0.28, "grad_norm": 2.1690523837968376, "learning_rate": 1.9383055164630345e-05, "loss": 2.7625, "step": 12330 }, { "epoch": 0.28, "grad_norm": 1.7057733547012242, "learning_rate": 1.9381791695944055e-05, "loss": 2.7144, "step": 12340 }, { "epoch": 0.28, "grad_norm": 2.2861976389611067, "learning_rate": 1.9380526976088777e-05, "loss": 2.7775, "step": 12350 }, { "epoch": 0.28, "grad_norm": 1.8046044967923731, "learning_rate": 1.9379261005233182e-05, "loss": 2.7323, "step": 12360 }, { "epoch": 0.28, "grad_norm": 2.103523871850075, "learning_rate": 1.9377993783546097e-05, "loss": 2.7414, "step": 12370 }, { "epoch": 0.28, "grad_norm": 3.347271060280652, "learning_rate": 1.937672531119652e-05, "loss": 2.6485, "step": 12380 }, { "epoch": 0.28, "grad_norm": 2.4136871019651536, "learning_rate": 1.937545558835362e-05, "loss": 2.7153, "step": 12390 }, { "epoch": 0.28, "grad_norm": 1.9735451523770793, "learning_rate": 1.9374184615186726e-05, "loss": 2.7186, "step": 12400 }, { "epoch": 0.28, "grad_norm": 2.4065605857202046, "learning_rate": 1.9372912391865334e-05, "loss": 2.7147, "step": 12410 }, { "epoch": 0.28, "grad_norm": 1.9924937074602278, "learning_rate": 1.937163891855912e-05, "loss": 2.7335, "step": 12420 }, { "epoch": 0.28, "grad_norm": 1.990044575349334, "learning_rate": 1.9370364195437902e-05, "loss": 2.7024, "step": 12430 }, { "epoch": 0.28, "grad_norm": 1.7413158998040172, "learning_rate": 1.936908822267169e-05, "loss": 2.8464, "step": 12440 }, { "epoch": 0.28, "grad_norm": 2.1134074301727077, "learning_rate": 1.9367811000430643e-05, "loss": 2.7581, "step": 12450 }, { "epoch": 0.28, "grad_norm": 2.3160613622435298, "learning_rate": 1.9366532528885095e-05, "loss": 2.9102, "step": 12460 }, { "epoch": 0.28, "grad_norm": 2.1826887369545735, "learning_rate": 1.936525280820555e-05, "loss": 2.6762, "step": 12470 }, { "epoch": 0.28, "grad_norm": 2.43798845826892, "learning_rate": 1.936397183856267e-05, "loss": 2.7181, "step": 12480 }, { "epoch": 0.28, "grad_norm": 1.7499635410671217, "learning_rate": 1.9362689620127278e-05, "loss": 2.8781, "step": 12490 }, { "epoch": 0.28, "grad_norm": 2.190915217671094, "learning_rate": 1.9361406153070384e-05, "loss": 2.7566, "step": 12500 }, { "epoch": 0.28, "grad_norm": 1.8662095160892973, "learning_rate": 1.9360121437563148e-05, "loss": 2.6134, "step": 12510 }, { "epoch": 0.28, "grad_norm": 2.1737681054463236, "learning_rate": 1.9358835473776906e-05, "loss": 2.6827, "step": 12520 }, { "epoch": 0.28, "grad_norm": 1.9185433051471423, "learning_rate": 1.935754826188315e-05, "loss": 2.7129, "step": 12530 }, { "epoch": 0.28, "grad_norm": 2.4221267446903156, "learning_rate": 1.9356259802053547e-05, "loss": 2.6278, "step": 12540 }, { "epoch": 0.28, "grad_norm": 1.967345399018631, "learning_rate": 1.935497009445993e-05, "loss": 2.7105, "step": 12550 }, { "epoch": 0.28, "grad_norm": 1.9919478318837378, "learning_rate": 1.9353679139274294e-05, "loss": 2.7369, "step": 12560 }, { "epoch": 0.28, "grad_norm": 2.3928925681701387, "learning_rate": 1.93523869366688e-05, "loss": 2.8807, "step": 12570 }, { "epoch": 0.28, "grad_norm": 2.0373969177044065, "learning_rate": 1.935109348681578e-05, "loss": 2.7023, "step": 12580 }, { "epoch": 0.28, "grad_norm": 1.8114674753026694, "learning_rate": 1.9349798789887734e-05, "loss": 2.6321, "step": 12590 }, { "epoch": 0.28, "grad_norm": 2.071164648675773, "learning_rate": 1.934850284605732e-05, "loss": 2.7116, "step": 12600 }, { "epoch": 0.28, "grad_norm": 2.0923305131591, "learning_rate": 1.934720565549737e-05, "loss": 2.735, "step": 12610 }, { "epoch": 0.28, "grad_norm": 2.1580063001839904, "learning_rate": 1.9345907218380872e-05, "loss": 2.7609, "step": 12620 }, { "epoch": 0.28, "grad_norm": 2.3746265326258436, "learning_rate": 1.9344607534880998e-05, "loss": 2.7955, "step": 12630 }, { "epoch": 0.29, "grad_norm": 2.668062170500229, "learning_rate": 1.934330660517107e-05, "loss": 2.7031, "step": 12640 }, { "epoch": 0.29, "grad_norm": 2.0398652548928062, "learning_rate": 1.9342004429424582e-05, "loss": 2.7292, "step": 12650 }, { "epoch": 0.29, "grad_norm": 2.08309566178143, "learning_rate": 1.9340701007815196e-05, "loss": 2.6601, "step": 12660 }, { "epoch": 0.29, "grad_norm": 2.3857116975193855, "learning_rate": 1.9339396340516737e-05, "loss": 2.6145, "step": 12670 }, { "epoch": 0.29, "grad_norm": 2.7465438298337506, "learning_rate": 1.9338090427703195e-05, "loss": 2.7486, "step": 12680 }, { "epoch": 0.29, "grad_norm": 2.886382211378277, "learning_rate": 1.9336783269548728e-05, "loss": 2.7113, "step": 12690 }, { "epoch": 0.29, "grad_norm": 2.125879838630425, "learning_rate": 1.9335474866227666e-05, "loss": 2.658, "step": 12700 }, { "epoch": 0.29, "grad_norm": 1.636660254813869, "learning_rate": 1.9334165217914496e-05, "loss": 2.701, "step": 12710 }, { "epoch": 0.29, "grad_norm": 2.3255009282258188, "learning_rate": 1.933285432478387e-05, "loss": 2.5935, "step": 12720 }, { "epoch": 0.29, "grad_norm": 2.015440272215456, "learning_rate": 1.9331542187010622e-05, "loss": 2.703, "step": 12730 }, { "epoch": 0.29, "grad_norm": 2.043027003782106, "learning_rate": 1.933022880476973e-05, "loss": 2.7249, "step": 12740 }, { "epoch": 0.29, "grad_norm": 3.8084655567367487, "learning_rate": 1.9328914178236354e-05, "loss": 2.6986, "step": 12750 }, { "epoch": 0.29, "grad_norm": 1.6473688503719097, "learning_rate": 1.9327598307585814e-05, "loss": 2.8412, "step": 12760 }, { "epoch": 0.29, "grad_norm": 2.0131453395481365, "learning_rate": 1.9326281192993594e-05, "loss": 2.7886, "step": 12770 }, { "epoch": 0.29, "grad_norm": 2.03508303785929, "learning_rate": 1.9324962834635344e-05, "loss": 2.7638, "step": 12780 }, { "epoch": 0.29, "grad_norm": 2.0890083896739706, "learning_rate": 1.932364323268689e-05, "loss": 2.6797, "step": 12790 }, { "epoch": 0.29, "grad_norm": 2.133139252915853, "learning_rate": 1.932232238732421e-05, "loss": 2.6848, "step": 12800 }, { "epoch": 0.29, "grad_norm": 3.1931019672982126, "learning_rate": 1.932100029872346e-05, "loss": 2.67, "step": 12810 }, { "epoch": 0.29, "grad_norm": 2.0142678333017057, "learning_rate": 1.9319676967060943e-05, "loss": 2.7177, "step": 12820 }, { "epoch": 0.29, "grad_norm": 2.0991889930280654, "learning_rate": 1.931835239251315e-05, "loss": 2.6841, "step": 12830 }, { "epoch": 0.29, "grad_norm": 2.5575662415970677, "learning_rate": 1.931702657525673e-05, "loss": 2.7741, "step": 12840 }, { "epoch": 0.29, "grad_norm": 1.8842605456469457, "learning_rate": 1.9315699515468493e-05, "loss": 2.6894, "step": 12850 }, { "epoch": 0.29, "grad_norm": 1.7672643828374355, "learning_rate": 1.931437121332541e-05, "loss": 2.7506, "step": 12860 }, { "epoch": 0.29, "grad_norm": 2.349274300414016, "learning_rate": 1.9313041669004638e-05, "loss": 2.6385, "step": 12870 }, { "epoch": 0.29, "grad_norm": 2.2092709788247586, "learning_rate": 1.9311710882683477e-05, "loss": 2.6386, "step": 12880 }, { "epoch": 0.29, "grad_norm": 2.3972627977688434, "learning_rate": 1.9310378854539415e-05, "loss": 2.7723, "step": 12890 }, { "epoch": 0.29, "grad_norm": 2.0872453137823523, "learning_rate": 1.9309045584750077e-05, "loss": 2.8102, "step": 12900 }, { "epoch": 0.29, "grad_norm": 1.895615496313984, "learning_rate": 1.930771107349328e-05, "loss": 2.6579, "step": 12910 }, { "epoch": 0.29, "grad_norm": 2.2915135292357656, "learning_rate": 1.9306375320946994e-05, "loss": 2.5297, "step": 12920 }, { "epoch": 0.29, "grad_norm": 2.3078121200732693, "learning_rate": 1.9305038327289356e-05, "loss": 2.7727, "step": 12930 }, { "epoch": 0.29, "grad_norm": 1.8430131773175416, "learning_rate": 1.930370009269867e-05, "loss": 2.6782, "step": 12940 }, { "epoch": 0.29, "grad_norm": 2.1672888119588545, "learning_rate": 1.9302360617353405e-05, "loss": 2.771, "step": 12950 }, { "epoch": 0.29, "grad_norm": 2.302723451241502, "learning_rate": 1.930101990143219e-05, "loss": 2.7594, "step": 12960 }, { "epoch": 0.29, "grad_norm": 2.0187094846963496, "learning_rate": 1.9299677945113838e-05, "loss": 2.6931, "step": 12970 }, { "epoch": 0.29, "grad_norm": 1.955745764751732, "learning_rate": 1.9298334748577304e-05, "loss": 2.6759, "step": 12980 }, { "epoch": 0.29, "grad_norm": 2.1986902795131193, "learning_rate": 1.929699031200172e-05, "loss": 2.8221, "step": 12990 }, { "epoch": 0.29, "grad_norm": 1.775288980827097, "learning_rate": 1.929564463556638e-05, "loss": 2.7258, "step": 13000 }, { "epoch": 0.29, "grad_norm": 2.4662565713796423, "learning_rate": 1.9294297719450755e-05, "loss": 2.7904, "step": 13010 }, { "epoch": 0.29, "grad_norm": 2.3488773508610463, "learning_rate": 1.9292949563834458e-05, "loss": 2.7612, "step": 13020 }, { "epoch": 0.29, "grad_norm": 2.3609994104236915, "learning_rate": 1.929160016889729e-05, "loss": 2.7562, "step": 13030 }, { "epoch": 0.29, "grad_norm": 1.7060481040991329, "learning_rate": 1.9290249534819206e-05, "loss": 2.7558, "step": 13040 }, { "epoch": 0.29, "grad_norm": 1.9849938054636227, "learning_rate": 1.9288897661780326e-05, "loss": 2.7228, "step": 13050 }, { "epoch": 0.29, "grad_norm": 2.177862256119405, "learning_rate": 1.9287544549960944e-05, "loss": 2.7046, "step": 13060 }, { "epoch": 0.29, "grad_norm": 1.6790837751503986, "learning_rate": 1.9286190199541505e-05, "loss": 2.624, "step": 13070 }, { "epoch": 0.29, "grad_norm": 2.2729878271423782, "learning_rate": 1.9284834610702635e-05, "loss": 2.7946, "step": 13080 }, { "epoch": 0.3, "grad_norm": 2.925681708681495, "learning_rate": 1.9283477783625107e-05, "loss": 2.7109, "step": 13090 }, { "epoch": 0.3, "grad_norm": 2.022054747749002, "learning_rate": 1.928211971848988e-05, "loss": 2.6471, "step": 13100 }, { "epoch": 0.3, "grad_norm": 2.145295728265416, "learning_rate": 1.9280760415478064e-05, "loss": 2.6624, "step": 13110 }, { "epoch": 0.3, "grad_norm": 1.615407209628871, "learning_rate": 1.9279399874770933e-05, "loss": 2.7688, "step": 13120 }, { "epoch": 0.3, "grad_norm": 2.041546802076249, "learning_rate": 1.927803809654994e-05, "loss": 2.7103, "step": 13130 }, { "epoch": 0.3, "grad_norm": 1.875695533886479, "learning_rate": 1.927667508099668e-05, "loss": 2.579, "step": 13140 }, { "epoch": 0.3, "grad_norm": 1.8315996871535996, "learning_rate": 1.927531082829294e-05, "loss": 2.762, "step": 13150 }, { "epoch": 0.3, "grad_norm": 2.334727485912962, "learning_rate": 1.9273945338620653e-05, "loss": 2.7599, "step": 13160 }, { "epoch": 0.3, "grad_norm": 2.2271673065330364, "learning_rate": 1.927257861216192e-05, "loss": 2.6785, "step": 13170 }, { "epoch": 0.3, "grad_norm": 2.014542488500857, "learning_rate": 1.9271210649099013e-05, "loss": 2.7781, "step": 13180 }, { "epoch": 0.3, "grad_norm": 1.7755966693563607, "learning_rate": 1.9269841449614367e-05, "loss": 2.7856, "step": 13190 }, { "epoch": 0.3, "grad_norm": 1.838859423531961, "learning_rate": 1.926847101389058e-05, "loss": 2.7141, "step": 13200 }, { "epoch": 0.3, "grad_norm": 2.1553260689616325, "learning_rate": 1.9267099342110408e-05, "loss": 2.7784, "step": 13210 }, { "epoch": 0.3, "grad_norm": 1.8329776992184292, "learning_rate": 1.9265726434456787e-05, "loss": 2.7755, "step": 13220 }, { "epoch": 0.3, "grad_norm": 1.8888350498632116, "learning_rate": 1.926435229111281e-05, "loss": 2.6588, "step": 13230 }, { "epoch": 0.3, "grad_norm": 1.9720053312816601, "learning_rate": 1.9262976912261727e-05, "loss": 2.5876, "step": 13240 }, { "epoch": 0.3, "grad_norm": 2.0287625726839376, "learning_rate": 1.9261600298086973e-05, "loss": 2.6806, "step": 13250 }, { "epoch": 0.3, "grad_norm": 2.0631474868893385, "learning_rate": 1.9260222448772122e-05, "loss": 2.7556, "step": 13260 }, { "epoch": 0.3, "grad_norm": 1.9444892624129309, "learning_rate": 1.9258843364500934e-05, "loss": 2.7328, "step": 13270 }, { "epoch": 0.3, "grad_norm": 2.431187666446992, "learning_rate": 1.9257463045457326e-05, "loss": 2.6455, "step": 13280 }, { "epoch": 0.3, "grad_norm": 2.091667204650452, "learning_rate": 1.9256081491825373e-05, "loss": 2.8045, "step": 13290 }, { "epoch": 0.3, "grad_norm": 2.3507928528956534, "learning_rate": 1.9254698703789323e-05, "loss": 2.7198, "step": 13300 }, { "epoch": 0.3, "grad_norm": 1.7899202013635092, "learning_rate": 1.925331468153359e-05, "loss": 2.7023, "step": 13310 }, { "epoch": 0.3, "grad_norm": 1.896159678409647, "learning_rate": 1.925192942524275e-05, "loss": 2.7303, "step": 13320 }, { "epoch": 0.3, "grad_norm": 1.8362578238085112, "learning_rate": 1.925054293510154e-05, "loss": 2.8212, "step": 13330 }, { "epoch": 0.3, "grad_norm": 2.254739240819426, "learning_rate": 1.9249155211294865e-05, "loss": 2.6299, "step": 13340 }, { "epoch": 0.3, "grad_norm": 1.9139360866714263, "learning_rate": 1.924776625400779e-05, "loss": 2.7266, "step": 13350 }, { "epoch": 0.3, "grad_norm": 2.538632819511414, "learning_rate": 1.9246376063425553e-05, "loss": 2.8252, "step": 13360 }, { "epoch": 0.3, "grad_norm": 1.811018238409982, "learning_rate": 1.924498463973355e-05, "loss": 2.7558, "step": 13370 }, { "epoch": 0.3, "grad_norm": 2.115792866396537, "learning_rate": 1.9243591983117344e-05, "loss": 2.723, "step": 13380 }, { "epoch": 0.3, "grad_norm": 1.8602319810787926, "learning_rate": 1.924219809376266e-05, "loss": 2.6499, "step": 13390 }, { "epoch": 0.3, "grad_norm": 2.4208334128801465, "learning_rate": 1.924080297185539e-05, "loss": 2.7112, "step": 13400 }, { "epoch": 0.3, "grad_norm": 2.0731395307887093, "learning_rate": 1.9239406617581592e-05, "loss": 2.6367, "step": 13410 }, { "epoch": 0.3, "grad_norm": 2.463500270297942, "learning_rate": 1.9238009031127484e-05, "loss": 2.8277, "step": 13420 }, { "epoch": 0.3, "grad_norm": 2.345817675541325, "learning_rate": 1.9236610212679445e-05, "loss": 2.88, "step": 13430 }, { "epoch": 0.3, "grad_norm": 1.8325197207247217, "learning_rate": 1.9235210162424032e-05, "loss": 2.6571, "step": 13440 }, { "epoch": 0.3, "grad_norm": 2.313651517866765, "learning_rate": 1.923380888054795e-05, "loss": 2.6854, "step": 13450 }, { "epoch": 0.3, "grad_norm": 1.918242212973636, "learning_rate": 1.9232406367238082e-05, "loss": 2.7967, "step": 13460 }, { "epoch": 0.3, "grad_norm": 2.1423417832027583, "learning_rate": 1.9231002622681467e-05, "loss": 2.7554, "step": 13470 }, { "epoch": 0.3, "grad_norm": 1.8192423551548145, "learning_rate": 1.9229597647065312e-05, "loss": 2.7741, "step": 13480 }, { "epoch": 0.3, "grad_norm": 1.7112417019716788, "learning_rate": 1.9228191440576983e-05, "loss": 2.7834, "step": 13490 }, { "epoch": 0.3, "grad_norm": 1.9723301007241958, "learning_rate": 1.9226784003404012e-05, "loss": 2.7262, "step": 13500 }, { "epoch": 0.3, "grad_norm": 2.319619891616245, "learning_rate": 1.9225375335734108e-05, "loss": 2.6794, "step": 13510 }, { "epoch": 0.3, "grad_norm": 1.7913891226893939, "learning_rate": 1.9223965437755116e-05, "loss": 2.6539, "step": 13520 }, { "epoch": 0.31, "grad_norm": 3.0175854044345636, "learning_rate": 1.9222554309655078e-05, "loss": 2.7692, "step": 13530 }, { "epoch": 0.31, "grad_norm": 2.4555109884598636, "learning_rate": 1.9221141951622174e-05, "loss": 2.7804, "step": 13540 }, { "epoch": 0.31, "grad_norm": 2.6207568118893794, "learning_rate": 1.921972836384476e-05, "loss": 2.68, "step": 13550 }, { "epoch": 0.31, "grad_norm": 2.045103684879059, "learning_rate": 1.9218313546511357e-05, "loss": 2.7388, "step": 13560 }, { "epoch": 0.31, "grad_norm": 1.7403925096951027, "learning_rate": 1.9216897499810643e-05, "loss": 2.7078, "step": 13570 }, { "epoch": 0.31, "grad_norm": 2.4380608453903334, "learning_rate": 1.921548022393147e-05, "loss": 2.8081, "step": 13580 }, { "epoch": 0.31, "grad_norm": 2.4870184465303944, "learning_rate": 1.921406171906284e-05, "loss": 2.6677, "step": 13590 }, { "epoch": 0.31, "grad_norm": 1.8735018212410834, "learning_rate": 1.9212641985393935e-05, "loss": 2.7597, "step": 13600 }, { "epoch": 0.31, "grad_norm": 1.7215033667884199, "learning_rate": 1.9211221023114083e-05, "loss": 2.7501, "step": 13610 }, { "epoch": 0.31, "grad_norm": 1.9463268285468949, "learning_rate": 1.9209798832412798e-05, "loss": 2.6434, "step": 13620 }, { "epoch": 0.31, "grad_norm": 2.259973367851505, "learning_rate": 1.9208375413479735e-05, "loss": 2.7015, "step": 13630 }, { "epoch": 0.31, "grad_norm": 2.288167127598039, "learning_rate": 1.9206950766504726e-05, "loss": 2.75, "step": 13640 }, { "epoch": 0.31, "grad_norm": 1.689365528567376, "learning_rate": 1.9205524891677764e-05, "loss": 2.7272, "step": 13650 }, { "epoch": 0.31, "grad_norm": 1.8208715720363986, "learning_rate": 1.920409778918901e-05, "loss": 2.8014, "step": 13660 }, { "epoch": 0.31, "grad_norm": 1.8039168232512348, "learning_rate": 1.920266945922878e-05, "loss": 2.6413, "step": 13670 }, { "epoch": 0.31, "grad_norm": 2.0852618380963923, "learning_rate": 1.9201239901987556e-05, "loss": 2.7307, "step": 13680 }, { "epoch": 0.31, "grad_norm": 2.1711664653284437, "learning_rate": 1.9199809117655987e-05, "loss": 2.7784, "step": 13690 }, { "epoch": 0.31, "grad_norm": 1.8301565469091168, "learning_rate": 1.9198377106424888e-05, "loss": 2.6542, "step": 13700 }, { "epoch": 0.31, "grad_norm": 2.1770282109627908, "learning_rate": 1.9196943868485232e-05, "loss": 2.8264, "step": 13710 }, { "epoch": 0.31, "grad_norm": 2.8545735979273315, "learning_rate": 1.9195509404028157e-05, "loss": 2.6387, "step": 13720 }, { "epoch": 0.31, "grad_norm": 1.9029151105534938, "learning_rate": 1.919407371324497e-05, "loss": 2.7438, "step": 13730 }, { "epoch": 0.31, "grad_norm": 1.9125179341562266, "learning_rate": 1.9192636796327127e-05, "loss": 2.7604, "step": 13740 }, { "epoch": 0.31, "grad_norm": 2.6785006404684175, "learning_rate": 1.919119865346627e-05, "loss": 2.6978, "step": 13750 }, { "epoch": 0.31, "grad_norm": 1.8475612606287184, "learning_rate": 1.9189759284854177e-05, "loss": 2.7699, "step": 13760 }, { "epoch": 0.31, "grad_norm": 1.5476527443884047, "learning_rate": 1.9188318690682815e-05, "loss": 2.6974, "step": 13770 }, { "epoch": 0.31, "grad_norm": 1.8454114966943085, "learning_rate": 1.91868768711443e-05, "loss": 2.6179, "step": 13780 }, { "epoch": 0.31, "grad_norm": 2.178009837944662, "learning_rate": 1.918543382643092e-05, "loss": 2.7551, "step": 13790 }, { "epoch": 0.31, "grad_norm": 1.9876121137246363, "learning_rate": 1.918398955673511e-05, "loss": 2.7311, "step": 13800 }, { "epoch": 0.31, "grad_norm": 1.8989092803765255, "learning_rate": 1.918254406224949e-05, "loss": 2.6557, "step": 13810 }, { "epoch": 0.31, "grad_norm": 1.7985220671820714, "learning_rate": 1.918109734316683e-05, "loss": 2.7383, "step": 13820 }, { "epoch": 0.31, "grad_norm": 2.6988600353849184, "learning_rate": 1.9179649399680068e-05, "loss": 2.766, "step": 13830 }, { "epoch": 0.31, "grad_norm": 2.944807583026322, "learning_rate": 1.91782002319823e-05, "loss": 2.7394, "step": 13840 }, { "epoch": 0.31, "grad_norm": 1.8261677008498345, "learning_rate": 1.9176749840266793e-05, "loss": 2.8066, "step": 13850 }, { "epoch": 0.31, "grad_norm": 1.8049084714045056, "learning_rate": 1.9175298224726972e-05, "loss": 2.7583, "step": 13860 }, { "epoch": 0.31, "grad_norm": 1.9588791171582667, "learning_rate": 1.9173845385556427e-05, "loss": 2.6545, "step": 13870 }, { "epoch": 0.31, "grad_norm": 2.3501802685579927, "learning_rate": 1.9172391322948907e-05, "loss": 2.7445, "step": 13880 }, { "epoch": 0.31, "grad_norm": 1.8250125474707504, "learning_rate": 1.9170936037098334e-05, "loss": 2.5986, "step": 13890 }, { "epoch": 0.31, "grad_norm": 1.9538924903284354, "learning_rate": 1.916947952819878e-05, "loss": 2.7672, "step": 13900 }, { "epoch": 0.31, "grad_norm": 1.963335461586374, "learning_rate": 1.9168021796444495e-05, "loss": 2.693, "step": 13910 }, { "epoch": 0.31, "grad_norm": 2.049825263330607, "learning_rate": 1.9166562842029876e-05, "loss": 2.7244, "step": 13920 }, { "epoch": 0.31, "grad_norm": 2.1728354939985937, "learning_rate": 1.91651026651495e-05, "loss": 2.746, "step": 13930 }, { "epoch": 0.31, "grad_norm": 1.8790417637659844, "learning_rate": 1.916364126599809e-05, "loss": 2.6842, "step": 13940 }, { "epoch": 0.31, "grad_norm": 1.7570835566420755, "learning_rate": 1.9162178644770543e-05, "loss": 2.8248, "step": 13950 }, { "epoch": 0.31, "grad_norm": 2.199669184112789, "learning_rate": 1.9160714801661918e-05, "loss": 2.7388, "step": 13960 }, { "epoch": 0.32, "grad_norm": 3.102285564181996, "learning_rate": 1.9159249736867434e-05, "loss": 2.8414, "step": 13970 }, { "epoch": 0.32, "grad_norm": 1.7942868869056496, "learning_rate": 1.9157783450582478e-05, "loss": 2.6739, "step": 13980 }, { "epoch": 0.32, "grad_norm": 2.099410058693645, "learning_rate": 1.915631594300259e-05, "loss": 2.8353, "step": 13990 }, { "epoch": 0.32, "grad_norm": 1.9197926536443306, "learning_rate": 1.9154847214323476e-05, "loss": 2.7106, "step": 14000 }, { "epoch": 0.32, "grad_norm": 1.8371010105139705, "learning_rate": 1.915337726474102e-05, "loss": 2.6921, "step": 14010 }, { "epoch": 0.32, "grad_norm": 2.6154382948682304, "learning_rate": 1.9151906094451246e-05, "loss": 2.7736, "step": 14020 }, { "epoch": 0.32, "grad_norm": 1.9810016205946828, "learning_rate": 1.9150433703650357e-05, "loss": 2.7488, "step": 14030 }, { "epoch": 0.32, "grad_norm": 1.7855161104980921, "learning_rate": 1.9148960092534707e-05, "loss": 2.6894, "step": 14040 }, { "epoch": 0.32, "grad_norm": 1.8714068444017973, "learning_rate": 1.9147485261300826e-05, "loss": 2.7908, "step": 14050 }, { "epoch": 0.32, "grad_norm": 2.626807978176742, "learning_rate": 1.9146009210145395e-05, "loss": 2.7453, "step": 14060 }, { "epoch": 0.32, "grad_norm": 2.404095084954914, "learning_rate": 1.914453193926526e-05, "loss": 2.697, "step": 14070 }, { "epoch": 0.32, "grad_norm": 1.6204205223031976, "learning_rate": 1.9143053448857444e-05, "loss": 2.6692, "step": 14080 }, { "epoch": 0.32, "grad_norm": 2.203609970245674, "learning_rate": 1.9141573739119106e-05, "loss": 2.7492, "step": 14090 }, { "epoch": 0.32, "grad_norm": 1.6993533706056514, "learning_rate": 1.914009281024759e-05, "loss": 2.683, "step": 14100 }, { "epoch": 0.32, "grad_norm": 1.7039969604553284, "learning_rate": 1.9138610662440393e-05, "loss": 2.6868, "step": 14110 }, { "epoch": 0.32, "grad_norm": 1.6210972124549605, "learning_rate": 1.9137127295895175e-05, "loss": 2.7302, "step": 14120 }, { "epoch": 0.32, "grad_norm": 2.0753347920711134, "learning_rate": 1.9135642710809763e-05, "loss": 2.7478, "step": 14130 }, { "epoch": 0.32, "grad_norm": 2.913971655868837, "learning_rate": 1.913415690738214e-05, "loss": 2.6594, "step": 14140 }, { "epoch": 0.32, "grad_norm": 1.8489010558228622, "learning_rate": 1.9132669885810457e-05, "loss": 2.8358, "step": 14150 }, { "epoch": 0.32, "grad_norm": 1.8723031098238345, "learning_rate": 1.9131181646293026e-05, "loss": 2.7489, "step": 14160 }, { "epoch": 0.32, "grad_norm": 2.2606253659540676, "learning_rate": 1.912969218902832e-05, "loss": 2.7218, "step": 14170 }, { "epoch": 0.32, "grad_norm": 1.9587522606869923, "learning_rate": 1.9128201514214973e-05, "loss": 2.8446, "step": 14180 }, { "epoch": 0.32, "grad_norm": 1.9243145593533404, "learning_rate": 1.9126709622051783e-05, "loss": 2.7435, "step": 14190 }, { "epoch": 0.32, "grad_norm": 1.6748277831294351, "learning_rate": 1.9125216512737715e-05, "loss": 2.7403, "step": 14200 }, { "epoch": 0.32, "grad_norm": 2.174195706316875, "learning_rate": 1.912372218647189e-05, "loss": 2.6829, "step": 14210 }, { "epoch": 0.32, "grad_norm": 1.7499636720269351, "learning_rate": 1.9122226643453592e-05, "loss": 2.6815, "step": 14220 }, { "epoch": 0.32, "grad_norm": 1.952701163957575, "learning_rate": 1.9120729883882273e-05, "loss": 2.6006, "step": 14230 }, { "epoch": 0.32, "grad_norm": 2.1822803944155513, "learning_rate": 1.911923190795754e-05, "loss": 2.6926, "step": 14240 }, { "epoch": 0.32, "grad_norm": 12.660991607400168, "learning_rate": 1.911773271587916e-05, "loss": 2.6677, "step": 14250 }, { "epoch": 0.32, "grad_norm": 6.743633414678927, "learning_rate": 1.911623230784708e-05, "loss": 2.8324, "step": 14260 }, { "epoch": 0.32, "grad_norm": 3.536383204117804, "learning_rate": 1.9114730684061386e-05, "loss": 2.8172, "step": 14270 }, { "epoch": 0.32, "grad_norm": 2.058619379971638, "learning_rate": 1.9113227844722342e-05, "loss": 2.7101, "step": 14280 }, { "epoch": 0.32, "grad_norm": 2.067654822650506, "learning_rate": 1.9111723790030362e-05, "loss": 2.8531, "step": 14290 }, { "epoch": 0.32, "grad_norm": 1.859797560169627, "learning_rate": 1.9110218520186038e-05, "loss": 2.7059, "step": 14300 }, { "epoch": 0.32, "grad_norm": 2.074763901600279, "learning_rate": 1.910871203539011e-05, "loss": 2.8097, "step": 14310 }, { "epoch": 0.32, "grad_norm": 2.0267958077003763, "learning_rate": 1.9107204335843485e-05, "loss": 2.7228, "step": 14320 }, { "epoch": 0.32, "grad_norm": 2.070876230375417, "learning_rate": 1.9105695421747234e-05, "loss": 2.7423, "step": 14330 }, { "epoch": 0.32, "grad_norm": 1.795958930141179, "learning_rate": 1.9104185293302585e-05, "loss": 2.6854, "step": 14340 }, { "epoch": 0.32, "grad_norm": 2.180473292769334, "learning_rate": 1.910267395071093e-05, "loss": 2.7917, "step": 14350 }, { "epoch": 0.32, "grad_norm": 2.625294251080841, "learning_rate": 1.9101161394173825e-05, "loss": 2.7167, "step": 14360 }, { "epoch": 0.32, "grad_norm": 2.3927447866314915, "learning_rate": 1.9099647623892994e-05, "loss": 2.7265, "step": 14370 }, { "epoch": 0.32, "grad_norm": 1.6570759900669052, "learning_rate": 1.9098132640070303e-05, "loss": 2.8047, "step": 14380 }, { "epoch": 0.32, "grad_norm": 2.0488173571158614, "learning_rate": 1.9096616442907802e-05, "loss": 2.6795, "step": 14390 }, { "epoch": 0.32, "grad_norm": 1.8910105054095059, "learning_rate": 1.909509903260769e-05, "loss": 2.8595, "step": 14400 }, { "epoch": 0.32, "grad_norm": 1.6745085350194289, "learning_rate": 1.909358040937233e-05, "loss": 2.6079, "step": 14410 }, { "epoch": 0.33, "grad_norm": 1.789731445375366, "learning_rate": 1.9092060573404246e-05, "loss": 2.8374, "step": 14420 }, { "epoch": 0.33, "grad_norm": 2.6843346416078617, "learning_rate": 1.909053952490613e-05, "loss": 2.7106, "step": 14430 }, { "epoch": 0.33, "grad_norm": 1.7545447678719175, "learning_rate": 1.908901726408083e-05, "loss": 2.7954, "step": 14440 }, { "epoch": 0.33, "grad_norm": 2.013109370072844, "learning_rate": 1.9087493791131355e-05, "loss": 2.7171, "step": 14450 }, { "epoch": 0.33, "grad_norm": 2.4522691599136666, "learning_rate": 1.9085969106260878e-05, "loss": 2.694, "step": 14460 }, { "epoch": 0.33, "grad_norm": 1.7572347143067781, "learning_rate": 1.9084443209672734e-05, "loss": 2.8361, "step": 14470 }, { "epoch": 0.33, "grad_norm": 2.4085082629520107, "learning_rate": 1.9082916101570418e-05, "loss": 2.79, "step": 14480 }, { "epoch": 0.33, "grad_norm": 2.0342771591601783, "learning_rate": 1.9081387782157587e-05, "loss": 2.7767, "step": 14490 }, { "epoch": 0.33, "grad_norm": 1.6752376556449549, "learning_rate": 1.9079858251638064e-05, "loss": 2.7709, "step": 14500 }, { "epoch": 0.33, "grad_norm": 2.0615915929650406, "learning_rate": 1.9078327510215822e-05, "loss": 2.771, "step": 14510 }, { "epoch": 0.33, "grad_norm": 2.055149460419816, "learning_rate": 1.9076795558095006e-05, "loss": 2.6547, "step": 14520 }, { "epoch": 0.33, "grad_norm": 1.6752102303970169, "learning_rate": 1.9075262395479927e-05, "loss": 2.7309, "step": 14530 }, { "epoch": 0.33, "grad_norm": 1.786695147073347, "learning_rate": 1.9073728022575034e-05, "loss": 2.7025, "step": 14540 }, { "epoch": 0.33, "grad_norm": 1.759957437887734, "learning_rate": 1.9072192439584968e-05, "loss": 2.8014, "step": 14550 }, { "epoch": 0.33, "grad_norm": 1.7659836342861794, "learning_rate": 1.9070655646714504e-05, "loss": 2.7426, "step": 14560 }, { "epoch": 0.33, "grad_norm": 1.778290965445482, "learning_rate": 1.9069117644168604e-05, "loss": 2.7591, "step": 14570 }, { "epoch": 0.33, "grad_norm": 1.9948009334264865, "learning_rate": 1.9067578432152368e-05, "loss": 2.7079, "step": 14580 }, { "epoch": 0.33, "grad_norm": 2.0518870314990236, "learning_rate": 1.9066038010871072e-05, "loss": 2.6522, "step": 14590 }, { "epoch": 0.33, "grad_norm": 1.877621804769306, "learning_rate": 1.9064496380530144e-05, "loss": 2.778, "step": 14600 }, { "epoch": 0.33, "grad_norm": 1.7876752723154974, "learning_rate": 1.9062953541335184e-05, "loss": 2.6615, "step": 14610 }, { "epoch": 0.33, "grad_norm": 1.8573592886449424, "learning_rate": 1.9061409493491947e-05, "loss": 2.6625, "step": 14620 }, { "epoch": 0.33, "grad_norm": 1.9088687977129006, "learning_rate": 1.9059864237206344e-05, "loss": 2.726, "step": 14630 }, { "epoch": 0.33, "grad_norm": 2.4580274407338623, "learning_rate": 1.9058317772684457e-05, "loss": 2.7071, "step": 14640 }, { "epoch": 0.33, "grad_norm": 1.9890558080405971, "learning_rate": 1.9056770100132523e-05, "loss": 2.6931, "step": 14650 }, { "epoch": 0.33, "grad_norm": 1.9444594912182664, "learning_rate": 1.9055221219756946e-05, "loss": 2.7235, "step": 14660 }, { "epoch": 0.33, "grad_norm": 1.7891782667068348, "learning_rate": 1.905367113176428e-05, "loss": 2.647, "step": 14670 }, { "epoch": 0.33, "grad_norm": 2.38972892496578, "learning_rate": 1.9052119836361246e-05, "loss": 2.7806, "step": 14680 }, { "epoch": 0.33, "grad_norm": 1.7295808823684948, "learning_rate": 1.905056733375474e-05, "loss": 2.6105, "step": 14690 }, { "epoch": 0.33, "grad_norm": 1.7405475872562481, "learning_rate": 1.904901362415179e-05, "loss": 2.7587, "step": 14700 }, { "epoch": 0.33, "grad_norm": 1.8926794358390744, "learning_rate": 1.904745870775961e-05, "loss": 2.5605, "step": 14710 }, { "epoch": 0.33, "grad_norm": 1.7938398277319305, "learning_rate": 1.9045902584785562e-05, "loss": 2.6769, "step": 14720 }, { "epoch": 0.33, "grad_norm": 2.0170416147858092, "learning_rate": 1.9044345255437178e-05, "loss": 2.7376, "step": 14730 }, { "epoch": 0.33, "grad_norm": 1.9856988457242795, "learning_rate": 1.904278671992214e-05, "loss": 2.7869, "step": 14740 }, { "epoch": 0.33, "grad_norm": 1.7163448014728104, "learning_rate": 1.90412269784483e-05, "loss": 2.6695, "step": 14750 }, { "epoch": 0.33, "grad_norm": 2.3077452231785536, "learning_rate": 1.9039666031223664e-05, "loss": 2.6336, "step": 14760 }, { "epoch": 0.33, "grad_norm": 1.8913469827679354, "learning_rate": 1.9038103878456404e-05, "loss": 2.7627, "step": 14770 }, { "epoch": 0.33, "grad_norm": 1.7187478388429789, "learning_rate": 1.9036540520354853e-05, "loss": 2.7493, "step": 14780 }, { "epoch": 0.33, "grad_norm": 1.7771052108565932, "learning_rate": 1.9034975957127496e-05, "loss": 2.5715, "step": 14790 }, { "epoch": 0.33, "grad_norm": 2.0576436537616423, "learning_rate": 1.9033410188983e-05, "loss": 2.6668, "step": 14800 }, { "epoch": 0.33, "grad_norm": 2.02217009977994, "learning_rate": 1.903184321613016e-05, "loss": 2.6761, "step": 14810 }, { "epoch": 0.33, "grad_norm": 1.953682294331569, "learning_rate": 1.9030275038777956e-05, "loss": 2.8173, "step": 14820 }, { "epoch": 0.33, "grad_norm": 1.8332850176028097, "learning_rate": 1.902870565713553e-05, "loss": 2.6944, "step": 14830 }, { "epoch": 0.33, "grad_norm": 1.76942531536127, "learning_rate": 1.9027135071412167e-05, "loss": 2.7248, "step": 14840 }, { "epoch": 0.33, "grad_norm": 1.916407250731294, "learning_rate": 1.9025563281817326e-05, "loss": 2.7519, "step": 14850 }, { "epoch": 0.34, "grad_norm": 1.6991921406001413, "learning_rate": 1.9023990288560624e-05, "loss": 2.7525, "step": 14860 }, { "epoch": 0.34, "grad_norm": 1.7840795991718559, "learning_rate": 1.9022416091851843e-05, "loss": 2.7433, "step": 14870 }, { "epoch": 0.34, "grad_norm": 1.7837347567127828, "learning_rate": 1.9020840691900908e-05, "loss": 2.6638, "step": 14880 }, { "epoch": 0.34, "grad_norm": 2.065233147468619, "learning_rate": 1.9019264088917925e-05, "loss": 2.7477, "step": 14890 }, { "epoch": 0.34, "grad_norm": 2.0127780030061273, "learning_rate": 1.901768628311315e-05, "loss": 2.7424, "step": 14900 }, { "epoch": 0.34, "grad_norm": 2.3880663155870208, "learning_rate": 1.9016107274697e-05, "loss": 2.6574, "step": 14910 }, { "epoch": 0.34, "grad_norm": 1.97464115177824, "learning_rate": 1.9014527063880056e-05, "loss": 2.6127, "step": 14920 }, { "epoch": 0.34, "grad_norm": 1.5990820806666817, "learning_rate": 1.9012945650873058e-05, "loss": 2.6564, "step": 14930 }, { "epoch": 0.34, "grad_norm": 1.9344749232581768, "learning_rate": 1.90113630358869e-05, "loss": 2.634, "step": 14940 }, { "epoch": 0.34, "grad_norm": 1.827143159759275, "learning_rate": 1.900977921913265e-05, "loss": 2.6322, "step": 14950 }, { "epoch": 0.34, "grad_norm": 2.080855181540868, "learning_rate": 1.9008194200821523e-05, "loss": 2.7018, "step": 14960 }, { "epoch": 0.34, "grad_norm": 2.016634110597166, "learning_rate": 1.9006607981164895e-05, "loss": 2.6537, "step": 14970 }, { "epoch": 0.34, "grad_norm": 1.992579255174244, "learning_rate": 1.9005020560374314e-05, "loss": 2.7427, "step": 14980 }, { "epoch": 0.34, "grad_norm": 1.8620438510806325, "learning_rate": 1.9003431938661478e-05, "loss": 2.7633, "step": 14990 }, { "epoch": 0.34, "grad_norm": 2.156787533658386, "learning_rate": 1.9001842116238246e-05, "loss": 2.6213, "step": 15000 }, { "epoch": 0.34, "grad_norm": 2.195516736726624, "learning_rate": 1.9000251093316645e-05, "loss": 2.7007, "step": 15010 }, { "epoch": 0.34, "grad_norm": 1.6091124967641321, "learning_rate": 1.899865887010885e-05, "loss": 2.752, "step": 15020 }, { "epoch": 0.34, "grad_norm": 2.077661532114242, "learning_rate": 1.8997065446827202e-05, "loss": 2.8461, "step": 15030 }, { "epoch": 0.34, "grad_norm": 1.9517667807567696, "learning_rate": 1.899547082368421e-05, "loss": 2.7726, "step": 15040 }, { "epoch": 0.34, "grad_norm": 2.3851078464657136, "learning_rate": 1.8993875000892525e-05, "loss": 2.6152, "step": 15050 }, { "epoch": 0.34, "grad_norm": 1.7571599302558176, "learning_rate": 1.899227797866497e-05, "loss": 2.7386, "step": 15060 }, { "epoch": 0.34, "grad_norm": 1.8798670475598374, "learning_rate": 1.8990679757214533e-05, "loss": 2.6772, "step": 15070 }, { "epoch": 0.34, "grad_norm": 1.9683155253670102, "learning_rate": 1.8989080336754346e-05, "loss": 2.7047, "step": 15080 }, { "epoch": 0.34, "grad_norm": 2.08723371329161, "learning_rate": 1.8987479717497716e-05, "loss": 2.8292, "step": 15090 }, { "epoch": 0.34, "grad_norm": 2.114011043401869, "learning_rate": 1.8985877899658106e-05, "loss": 2.7838, "step": 15100 }, { "epoch": 0.34, "grad_norm": 2.3400889561422042, "learning_rate": 1.898427488344913e-05, "loss": 2.6941, "step": 15110 }, { "epoch": 0.34, "grad_norm": 2.0106328600955337, "learning_rate": 1.8982670669084574e-05, "loss": 2.6191, "step": 15120 }, { "epoch": 0.34, "grad_norm": 2.0775219990056164, "learning_rate": 1.8981065256778375e-05, "loss": 2.7242, "step": 15130 }, { "epoch": 0.34, "grad_norm": 4.060176875218205, "learning_rate": 1.8979458646744633e-05, "loss": 2.6929, "step": 15140 }, { "epoch": 0.34, "grad_norm": 2.061988840413874, "learning_rate": 1.8977850839197608e-05, "loss": 2.8271, "step": 15150 }, { "epoch": 0.34, "grad_norm": 2.3151529582297483, "learning_rate": 1.8976241834351724e-05, "loss": 2.758, "step": 15160 }, { "epoch": 0.34, "grad_norm": 1.8762348988498374, "learning_rate": 1.8974631632421556e-05, "loss": 2.7021, "step": 15170 }, { "epoch": 0.34, "grad_norm": 2.2711323297689665, "learning_rate": 1.8973020233621842e-05, "loss": 2.7336, "step": 15180 }, { "epoch": 0.34, "grad_norm": 1.880849844845327, "learning_rate": 1.897140763816748e-05, "loss": 2.7887, "step": 15190 }, { "epoch": 0.34, "grad_norm": 1.9021380630650526, "learning_rate": 1.896979384627353e-05, "loss": 2.6995, "step": 15200 }, { "epoch": 0.34, "grad_norm": 2.08303800990317, "learning_rate": 1.8968178858155214e-05, "loss": 2.734, "step": 15210 }, { "epoch": 0.34, "grad_norm": 2.546798783166087, "learning_rate": 1.89665626740279e-05, "loss": 2.7202, "step": 15220 }, { "epoch": 0.34, "grad_norm": 1.875966356432404, "learning_rate": 1.896494529410713e-05, "loss": 2.7932, "step": 15230 }, { "epoch": 0.34, "grad_norm": 2.1721102133348436, "learning_rate": 1.89633267186086e-05, "loss": 2.7707, "step": 15240 }, { "epoch": 0.34, "grad_norm": 1.9870089210767372, "learning_rate": 1.896170694774816e-05, "loss": 2.7089, "step": 15250 }, { "epoch": 0.34, "grad_norm": 2.1662613401594473, "learning_rate": 1.896008598174184e-05, "loss": 2.7035, "step": 15260 }, { "epoch": 0.34, "grad_norm": 2.3814625240038647, "learning_rate": 1.8958463820805795e-05, "loss": 2.7314, "step": 15270 }, { "epoch": 0.34, "grad_norm": 1.997741103308146, "learning_rate": 1.8956840465156367e-05, "loss": 2.6965, "step": 15280 }, { "epoch": 0.34, "grad_norm": 2.3424047152662553, "learning_rate": 1.8955215915010054e-05, "loss": 2.7002, "step": 15290 }, { "epoch": 0.35, "grad_norm": 1.6441157831657491, "learning_rate": 1.89535901705835e-05, "loss": 2.7791, "step": 15300 }, { "epoch": 0.35, "grad_norm": 1.9955390730789118, "learning_rate": 1.8951963232093522e-05, "loss": 2.6091, "step": 15310 }, { "epoch": 0.35, "grad_norm": 1.62035010955052, "learning_rate": 1.8950335099757092e-05, "loss": 2.7776, "step": 15320 }, { "epoch": 0.35, "grad_norm": 1.7866800256750694, "learning_rate": 1.894870577379133e-05, "loss": 2.7069, "step": 15330 }, { "epoch": 0.35, "grad_norm": 2.0369750012632144, "learning_rate": 1.894707525441354e-05, "loss": 2.8241, "step": 15340 }, { "epoch": 0.35, "grad_norm": 2.076440652975239, "learning_rate": 1.894544354184116e-05, "loss": 2.5627, "step": 15350 }, { "epoch": 0.35, "grad_norm": 1.8731062950013813, "learning_rate": 1.89438106362918e-05, "loss": 2.6481, "step": 15360 }, { "epoch": 0.35, "grad_norm": 1.7155055802177646, "learning_rate": 1.8942176537983226e-05, "loss": 2.7353, "step": 15370 }, { "epoch": 0.35, "grad_norm": 1.8280656899209866, "learning_rate": 1.8940541247133363e-05, "loss": 2.7442, "step": 15380 }, { "epoch": 0.35, "grad_norm": 1.9909177319516134, "learning_rate": 1.8938904763960303e-05, "loss": 2.7361, "step": 15390 }, { "epoch": 0.35, "grad_norm": 2.2146971913247016, "learning_rate": 1.8937267088682278e-05, "loss": 2.6867, "step": 15400 }, { "epoch": 0.35, "grad_norm": 1.7880734668565934, "learning_rate": 1.89356282215177e-05, "loss": 2.5937, "step": 15410 }, { "epoch": 0.35, "grad_norm": 1.903919289996242, "learning_rate": 1.8933988162685125e-05, "loss": 2.7178, "step": 15420 }, { "epoch": 0.35, "grad_norm": 2.349329129863165, "learning_rate": 1.8932346912403278e-05, "loss": 2.7518, "step": 15430 }, { "epoch": 0.35, "grad_norm": 1.7492105538758564, "learning_rate": 1.8930704470891036e-05, "loss": 2.7456, "step": 15440 }, { "epoch": 0.35, "grad_norm": 2.1633174021258323, "learning_rate": 1.8929060838367436e-05, "loss": 2.8063, "step": 15450 }, { "epoch": 0.35, "grad_norm": 1.9177595566064534, "learning_rate": 1.892741601505168e-05, "loss": 2.6545, "step": 15460 }, { "epoch": 0.35, "grad_norm": 1.7117869997253559, "learning_rate": 1.8925770001163117e-05, "loss": 2.6741, "step": 15470 }, { "epoch": 0.35, "grad_norm": 1.7081705932818565, "learning_rate": 1.8924122796921266e-05, "loss": 2.6469, "step": 15480 }, { "epoch": 0.35, "grad_norm": 2.545728126122648, "learning_rate": 1.89224744025458e-05, "loss": 2.7241, "step": 15490 }, { "epoch": 0.35, "grad_norm": 1.942260667212414, "learning_rate": 1.8920824818256554e-05, "loss": 2.8086, "step": 15500 }, { "epoch": 0.35, "grad_norm": 2.3499303849149347, "learning_rate": 1.8919174044273513e-05, "loss": 2.6778, "step": 15510 }, { "epoch": 0.35, "grad_norm": 1.858348237685157, "learning_rate": 1.891752208081683e-05, "loss": 2.7613, "step": 15520 }, { "epoch": 0.35, "grad_norm": 2.0886310740713254, "learning_rate": 1.8915868928106813e-05, "loss": 2.6338, "step": 15530 }, { "epoch": 0.35, "grad_norm": 1.9166565680525904, "learning_rate": 1.891421458636393e-05, "loss": 2.7682, "step": 15540 }, { "epoch": 0.35, "grad_norm": 1.9257908370311196, "learning_rate": 1.8912559055808802e-05, "loss": 2.733, "step": 15550 }, { "epoch": 0.35, "grad_norm": 1.7769926397763545, "learning_rate": 1.8910902336662214e-05, "loss": 2.7178, "step": 15560 }, { "epoch": 0.35, "grad_norm": 1.6434487189587281, "learning_rate": 1.8909244429145115e-05, "loss": 2.7807, "step": 15570 }, { "epoch": 0.35, "grad_norm": 3.5432721212555727, "learning_rate": 1.89075853334786e-05, "loss": 2.6518, "step": 15580 }, { "epoch": 0.35, "grad_norm": 2.0806297435800665, "learning_rate": 1.8905925049883928e-05, "loss": 2.659, "step": 15590 }, { "epoch": 0.35, "grad_norm": 1.8800879672184405, "learning_rate": 1.8904263578582517e-05, "loss": 2.7482, "step": 15600 }, { "epoch": 0.35, "grad_norm": 1.7336680712865171, "learning_rate": 1.8902600919795948e-05, "loss": 2.818, "step": 15610 }, { "epoch": 0.35, "grad_norm": 2.081478867784675, "learning_rate": 1.890093707374595e-05, "loss": 2.6468, "step": 15620 }, { "epoch": 0.35, "grad_norm": 1.6351408418047753, "learning_rate": 1.8899272040654417e-05, "loss": 2.7231, "step": 15630 }, { "epoch": 0.35, "grad_norm": 1.9537938213699717, "learning_rate": 1.8897605820743403e-05, "loss": 2.62, "step": 15640 }, { "epoch": 0.35, "grad_norm": 2.1448020080771184, "learning_rate": 1.8895938414235114e-05, "loss": 2.7335, "step": 15650 }, { "epoch": 0.35, "grad_norm": 2.1409604720538034, "learning_rate": 1.889426982135192e-05, "loss": 2.7907, "step": 15660 }, { "epoch": 0.35, "grad_norm": 1.748474428663511, "learning_rate": 1.8892600042316342e-05, "loss": 2.7531, "step": 15670 }, { "epoch": 0.35, "grad_norm": 1.709721428469744, "learning_rate": 1.889092907735107e-05, "loss": 2.6389, "step": 15680 }, { "epoch": 0.35, "grad_norm": 1.885079155879625, "learning_rate": 1.8889256926678946e-05, "loss": 2.7881, "step": 15690 }, { "epoch": 0.35, "grad_norm": 1.8317428567629088, "learning_rate": 1.8887583590522967e-05, "loss": 2.8878, "step": 15700 }, { "epoch": 0.35, "grad_norm": 1.9629195290436938, "learning_rate": 1.888590906910629e-05, "loss": 2.8276, "step": 15710 }, { "epoch": 0.35, "grad_norm": 1.9451587971352402, "learning_rate": 1.8884233362652242e-05, "loss": 2.6976, "step": 15720 }, { "epoch": 0.35, "grad_norm": 2.033665060213413, "learning_rate": 1.8882556471384285e-05, "loss": 2.7417, "step": 15730 }, { "epoch": 0.35, "grad_norm": 1.6741888006397863, "learning_rate": 1.888087839552606e-05, "loss": 2.733, "step": 15740 }, { "epoch": 0.36, "grad_norm": 1.8642610726583944, "learning_rate": 1.887919913530135e-05, "loss": 2.6502, "step": 15750 }, { "epoch": 0.36, "grad_norm": 1.948075547120618, "learning_rate": 1.887751869093411e-05, "loss": 2.6266, "step": 15760 }, { "epoch": 0.36, "grad_norm": 1.661955102377463, "learning_rate": 1.8875837062648445e-05, "loss": 2.773, "step": 15770 }, { "epoch": 0.36, "grad_norm": 1.5900491493764388, "learning_rate": 1.8874154250668617e-05, "loss": 2.7749, "step": 15780 }, { "epoch": 0.36, "grad_norm": 1.621634237862706, "learning_rate": 1.887247025521905e-05, "loss": 2.6112, "step": 15790 }, { "epoch": 0.36, "grad_norm": 2.043860712532612, "learning_rate": 1.8870785076524326e-05, "loss": 2.7347, "step": 15800 }, { "epoch": 0.36, "grad_norm": 2.5627176274483054, "learning_rate": 1.886909871480918e-05, "loss": 2.7423, "step": 15810 }, { "epoch": 0.36, "grad_norm": 1.621725936493362, "learning_rate": 1.8867411170298507e-05, "loss": 2.8126, "step": 15820 }, { "epoch": 0.36, "grad_norm": 1.9453177475456636, "learning_rate": 1.886572244321736e-05, "loss": 2.8315, "step": 15830 }, { "epoch": 0.36, "grad_norm": 1.7874864476355794, "learning_rate": 1.8864032533790956e-05, "loss": 2.7164, "step": 15840 }, { "epoch": 0.36, "grad_norm": 2.4180390365699633, "learning_rate": 1.8862341442244653e-05, "loss": 2.7825, "step": 15850 }, { "epoch": 0.36, "grad_norm": 2.2445211224912964, "learning_rate": 1.886064916880399e-05, "loss": 2.5851, "step": 15860 }, { "epoch": 0.36, "grad_norm": 1.7842964935243117, "learning_rate": 1.885895571369464e-05, "loss": 2.7298, "step": 15870 }, { "epoch": 0.36, "grad_norm": 2.0796278083636857, "learning_rate": 1.885726107714245e-05, "loss": 2.7485, "step": 15880 }, { "epoch": 0.36, "grad_norm": 2.1670541740363456, "learning_rate": 1.8855565259373422e-05, "loss": 2.699, "step": 15890 }, { "epoch": 0.36, "grad_norm": 1.7903480324158951, "learning_rate": 1.8853868260613706e-05, "loss": 2.8689, "step": 15900 }, { "epoch": 0.36, "grad_norm": 1.844965766001667, "learning_rate": 1.885217008108962e-05, "loss": 2.7533, "step": 15910 }, { "epoch": 0.36, "grad_norm": 1.827565698790594, "learning_rate": 1.885047072102763e-05, "loss": 2.6327, "step": 15920 }, { "epoch": 0.36, "grad_norm": 1.9365250447576692, "learning_rate": 1.8848770180654377e-05, "loss": 2.7459, "step": 15930 }, { "epoch": 0.36, "grad_norm": 2.088050124555697, "learning_rate": 1.8847068460196633e-05, "loss": 2.6115, "step": 15940 }, { "epoch": 0.36, "grad_norm": 1.8182219037872722, "learning_rate": 1.8845365559881353e-05, "loss": 2.7278, "step": 15950 }, { "epoch": 0.36, "grad_norm": 2.6237950625274147, "learning_rate": 1.8843661479935635e-05, "loss": 2.753, "step": 15960 }, { "epoch": 0.36, "grad_norm": 2.0826647839954755, "learning_rate": 1.8841956220586736e-05, "loss": 2.6358, "step": 15970 }, { "epoch": 0.36, "grad_norm": 2.3418084201182623, "learning_rate": 1.884024978206207e-05, "loss": 2.5742, "step": 15980 }, { "epoch": 0.36, "grad_norm": 1.6162835187336495, "learning_rate": 1.8838542164589216e-05, "loss": 2.7326, "step": 15990 }, { "epoch": 0.36, "grad_norm": 1.928897565546441, "learning_rate": 1.8836833368395902e-05, "loss": 2.7144, "step": 16000 }, { "epoch": 0.36, "grad_norm": 2.0945694856518826, "learning_rate": 1.883512339371001e-05, "loss": 2.7419, "step": 16010 }, { "epoch": 0.36, "grad_norm": 2.06301890293257, "learning_rate": 1.883341224075959e-05, "loss": 2.7028, "step": 16020 }, { "epoch": 0.36, "grad_norm": 2.61600840783163, "learning_rate": 1.8831699909772846e-05, "loss": 2.7085, "step": 16030 }, { "epoch": 0.36, "grad_norm": 2.245221177470541, "learning_rate": 1.882998640097813e-05, "loss": 2.7176, "step": 16040 }, { "epoch": 0.36, "grad_norm": 1.7229079636846147, "learning_rate": 1.8828271714603962e-05, "loss": 2.7146, "step": 16050 }, { "epoch": 0.36, "grad_norm": 2.022252877866903, "learning_rate": 1.8826555850879016e-05, "loss": 2.6582, "step": 16060 }, { "epoch": 0.36, "grad_norm": 1.822006046971535, "learning_rate": 1.882483881003212e-05, "loss": 2.6948, "step": 16070 }, { "epoch": 0.36, "grad_norm": 1.7318099660613875, "learning_rate": 1.882312059229226e-05, "loss": 2.7261, "step": 16080 }, { "epoch": 0.36, "grad_norm": 2.050451889906987, "learning_rate": 1.8821401197888586e-05, "loss": 2.6878, "step": 16090 }, { "epoch": 0.36, "grad_norm": 1.5691729534794416, "learning_rate": 1.881968062705039e-05, "loss": 2.5184, "step": 16100 }, { "epoch": 0.36, "grad_norm": 2.548379059523761, "learning_rate": 1.8817958880007134e-05, "loss": 2.6027, "step": 16110 }, { "epoch": 0.36, "grad_norm": 1.7310816107325842, "learning_rate": 1.8816235956988436e-05, "loss": 2.7968, "step": 16120 }, { "epoch": 0.36, "grad_norm": 1.5787069935622515, "learning_rate": 1.8814511858224063e-05, "loss": 2.8389, "step": 16130 }, { "epoch": 0.36, "grad_norm": 1.774233574576138, "learning_rate": 1.8812786583943944e-05, "loss": 2.6432, "step": 16140 }, { "epoch": 0.36, "grad_norm": 2.8988014976459398, "learning_rate": 1.8811060134378163e-05, "loss": 2.7876, "step": 16150 }, { "epoch": 0.36, "grad_norm": 3.309396159047528, "learning_rate": 1.8809332509756964e-05, "loss": 2.7323, "step": 16160 }, { "epoch": 0.36, "grad_norm": 2.308553581558245, "learning_rate": 1.8807603710310744e-05, "loss": 2.7402, "step": 16170 }, { "epoch": 0.36, "grad_norm": 1.8296796147895265, "learning_rate": 1.8805873736270062e-05, "loss": 2.7775, "step": 16180 }, { "epoch": 0.37, "grad_norm": 1.67380094016608, "learning_rate": 1.8804142587865626e-05, "loss": 2.7694, "step": 16190 }, { "epoch": 0.37, "grad_norm": 1.5325807528524151, "learning_rate": 1.8802410265328302e-05, "loss": 2.7015, "step": 16200 }, { "epoch": 0.37, "grad_norm": 2.039136474776179, "learning_rate": 1.880067676888912e-05, "loss": 2.6815, "step": 16210 }, { "epoch": 0.37, "grad_norm": 1.8808919511870537, "learning_rate": 1.8798942098779256e-05, "loss": 2.7122, "step": 16220 }, { "epoch": 0.37, "grad_norm": 2.0488504065331017, "learning_rate": 1.8797206255230058e-05, "loss": 2.5031, "step": 16230 }, { "epoch": 0.37, "grad_norm": 2.0605768066013526, "learning_rate": 1.879546923847301e-05, "loss": 2.6757, "step": 16240 }, { "epoch": 0.37, "grad_norm": 2.0061850670574692, "learning_rate": 1.8793731048739767e-05, "loss": 2.6879, "step": 16250 }, { "epoch": 0.37, "grad_norm": 1.667835504567859, "learning_rate": 1.8791991686262138e-05, "loss": 2.5927, "step": 16260 }, { "epoch": 0.37, "grad_norm": 2.753272302107966, "learning_rate": 1.8790251151272083e-05, "loss": 2.6959, "step": 16270 }, { "epoch": 0.37, "grad_norm": 1.7563799136167741, "learning_rate": 1.8788509444001728e-05, "loss": 2.7566, "step": 16280 }, { "epoch": 0.37, "grad_norm": 1.832550194761647, "learning_rate": 1.8786766564683345e-05, "loss": 2.7675, "step": 16290 }, { "epoch": 0.37, "grad_norm": 2.3458309975760474, "learning_rate": 1.8785022513549368e-05, "loss": 2.6145, "step": 16300 }, { "epoch": 0.37, "grad_norm": 1.9760093188902317, "learning_rate": 1.8783277290832385e-05, "loss": 2.6431, "step": 16310 }, { "epoch": 0.37, "grad_norm": 2.4082361685293674, "learning_rate": 1.878153089676514e-05, "loss": 2.6837, "step": 16320 }, { "epoch": 0.37, "grad_norm": 1.9059001884939588, "learning_rate": 1.877978333158054e-05, "loss": 2.684, "step": 16330 }, { "epoch": 0.37, "grad_norm": 1.8526667560005332, "learning_rate": 1.877803459551164e-05, "loss": 2.7031, "step": 16340 }, { "epoch": 0.37, "grad_norm": 1.9762059826698868, "learning_rate": 1.8776284688791653e-05, "loss": 2.7836, "step": 16350 }, { "epoch": 0.37, "grad_norm": 1.9553481577534366, "learning_rate": 1.877453361165395e-05, "loss": 2.7553, "step": 16360 }, { "epoch": 0.37, "grad_norm": 1.8077383923558321, "learning_rate": 1.8772781364332058e-05, "loss": 2.6825, "step": 16370 }, { "epoch": 0.37, "grad_norm": 1.7708957098022446, "learning_rate": 1.8771027947059655e-05, "loss": 2.5825, "step": 16380 }, { "epoch": 0.37, "grad_norm": 1.5862025835156857, "learning_rate": 1.8769273360070585e-05, "loss": 2.6726, "step": 16390 }, { "epoch": 0.37, "grad_norm": 1.5171176154167356, "learning_rate": 1.876751760359884e-05, "loss": 2.669, "step": 16400 }, { "epoch": 0.37, "grad_norm": 1.8128284419211471, "learning_rate": 1.8765760677878567e-05, "loss": 2.7366, "step": 16410 }, { "epoch": 0.37, "grad_norm": 2.648736906461833, "learning_rate": 1.8764002583144075e-05, "loss": 2.6097, "step": 16420 }, { "epoch": 0.37, "grad_norm": 1.9043032442228724, "learning_rate": 1.8762243319629826e-05, "loss": 2.7485, "step": 16430 }, { "epoch": 0.37, "grad_norm": 1.6178331174530869, "learning_rate": 1.8760482887570438e-05, "loss": 2.7271, "step": 16440 }, { "epoch": 0.37, "grad_norm": 1.9129579444392335, "learning_rate": 1.8758721287200684e-05, "loss": 2.8202, "step": 16450 }, { "epoch": 0.37, "grad_norm": 2.4822901478051955, "learning_rate": 1.875695851875549e-05, "loss": 2.7343, "step": 16460 }, { "epoch": 0.37, "grad_norm": 1.777605320445451, "learning_rate": 1.8755194582469955e-05, "loss": 2.7903, "step": 16470 }, { "epoch": 0.37, "grad_norm": 2.246838671728247, "learning_rate": 1.8753429478579302e-05, "loss": 2.669, "step": 16480 }, { "epoch": 0.37, "grad_norm": 1.5662181189183526, "learning_rate": 1.8751663207318938e-05, "loss": 2.6924, "step": 16490 }, { "epoch": 0.37, "grad_norm": 1.9389926308289156, "learning_rate": 1.874989576892441e-05, "loss": 2.7341, "step": 16500 }, { "epoch": 0.37, "grad_norm": 2.1058463047711165, "learning_rate": 1.8748127163631436e-05, "loss": 2.7831, "step": 16510 }, { "epoch": 0.37, "grad_norm": 1.71398519461996, "learning_rate": 1.8746357391675868e-05, "loss": 2.7512, "step": 16520 }, { "epoch": 0.37, "grad_norm": 1.769503338633784, "learning_rate": 1.8744586453293732e-05, "loss": 2.6772, "step": 16530 }, { "epoch": 0.37, "grad_norm": 1.6723322027297285, "learning_rate": 1.8742814348721202e-05, "loss": 2.6966, "step": 16540 }, { "epoch": 0.37, "grad_norm": 1.9055406728759243, "learning_rate": 1.8741041078194607e-05, "loss": 2.6672, "step": 16550 }, { "epoch": 0.37, "grad_norm": 2.5491895075808575, "learning_rate": 1.873926664195043e-05, "loss": 2.6875, "step": 16560 }, { "epoch": 0.37, "grad_norm": 1.7583377671916856, "learning_rate": 1.873749104022532e-05, "loss": 2.7585, "step": 16570 }, { "epoch": 0.37, "grad_norm": 1.6191157942503631, "learning_rate": 1.8735714273256067e-05, "loss": 2.8172, "step": 16580 }, { "epoch": 0.37, "grad_norm": 1.8866218962084598, "learning_rate": 1.873393634127963e-05, "loss": 2.6635, "step": 16590 }, { "epoch": 0.37, "grad_norm": 1.6657802420348786, "learning_rate": 1.873215724453311e-05, "loss": 2.7739, "step": 16600 }, { "epoch": 0.37, "grad_norm": 2.073037765128659, "learning_rate": 1.873037698325377e-05, "loss": 2.7314, "step": 16610 }, { "epoch": 0.37, "grad_norm": 1.9770678969926292, "learning_rate": 1.8728595557679027e-05, "loss": 2.6499, "step": 16620 }, { "epoch": 0.38, "grad_norm": 1.82677916363696, "learning_rate": 1.8726812968046465e-05, "loss": 2.81, "step": 16630 }, { "epoch": 0.38, "grad_norm": 1.7902586178377144, "learning_rate": 1.87250292145938e-05, "loss": 2.6902, "step": 16640 }, { "epoch": 0.38, "grad_norm": 1.9749217607486587, "learning_rate": 1.872324429755892e-05, "loss": 2.8569, "step": 16650 }, { "epoch": 0.38, "grad_norm": 2.297091846453394, "learning_rate": 1.872145821717987e-05, "loss": 2.6339, "step": 16660 }, { "epoch": 0.38, "grad_norm": 1.8892153619083814, "learning_rate": 1.8719670973694837e-05, "loss": 2.8187, "step": 16670 }, { "epoch": 0.38, "grad_norm": 1.8400328510990949, "learning_rate": 1.871788256734217e-05, "loss": 2.6922, "step": 16680 }, { "epoch": 0.38, "grad_norm": 1.7452888857325544, "learning_rate": 1.871609299836038e-05, "loss": 2.7088, "step": 16690 }, { "epoch": 0.38, "grad_norm": 2.1733931806780564, "learning_rate": 1.8714302266988126e-05, "loss": 2.6236, "step": 16700 }, { "epoch": 0.38, "grad_norm": 1.8834477089901829, "learning_rate": 1.871251037346421e-05, "loss": 2.6819, "step": 16710 }, { "epoch": 0.38, "grad_norm": 2.308577228985731, "learning_rate": 1.8710717318027617e-05, "loss": 2.7695, "step": 16720 }, { "epoch": 0.38, "grad_norm": 1.8860339549134904, "learning_rate": 1.8708923100917463e-05, "loss": 2.6891, "step": 16730 }, { "epoch": 0.38, "grad_norm": 1.892668429022229, "learning_rate": 1.8707127722373028e-05, "loss": 2.6918, "step": 16740 }, { "epoch": 0.38, "grad_norm": 1.6983646046363476, "learning_rate": 1.870533118263375e-05, "loss": 2.9483, "step": 16750 }, { "epoch": 0.38, "grad_norm": 1.7530087023895837, "learning_rate": 1.8703533481939214e-05, "loss": 2.6507, "step": 16760 }, { "epoch": 0.38, "grad_norm": 2.0531817338800438, "learning_rate": 1.8701734620529165e-05, "loss": 2.665, "step": 16770 }, { "epoch": 0.38, "grad_norm": 1.7678701979567983, "learning_rate": 1.8699934598643503e-05, "loss": 2.653, "step": 16780 }, { "epoch": 0.38, "grad_norm": 2.6457762424343314, "learning_rate": 1.8698133416522282e-05, "loss": 2.7401, "step": 16790 }, { "epoch": 0.38, "grad_norm": 2.3169467021669345, "learning_rate": 1.8696331074405706e-05, "loss": 2.8302, "step": 16800 }, { "epoch": 0.38, "grad_norm": 2.048365082592648, "learning_rate": 1.869452757253414e-05, "loss": 2.7007, "step": 16810 }, { "epoch": 0.38, "grad_norm": 2.2525542361117257, "learning_rate": 1.8692722911148105e-05, "loss": 2.7387, "step": 16820 }, { "epoch": 0.38, "grad_norm": 2.088665080584738, "learning_rate": 1.8690917090488266e-05, "loss": 2.6666, "step": 16830 }, { "epoch": 0.38, "grad_norm": 2.6191022793563725, "learning_rate": 1.8689110110795458e-05, "loss": 2.6598, "step": 16840 }, { "epoch": 0.38, "grad_norm": 3.4829926014791903, "learning_rate": 1.8687301972310658e-05, "loss": 2.7178, "step": 16850 }, { "epoch": 0.38, "grad_norm": 1.8688794006898868, "learning_rate": 1.8685492675275002e-05, "loss": 2.7727, "step": 16860 }, { "epoch": 0.38, "grad_norm": 1.8585333670623854, "learning_rate": 1.8683682219929783e-05, "loss": 2.7555, "step": 16870 }, { "epoch": 0.38, "grad_norm": 1.768467849626366, "learning_rate": 1.868187060651644e-05, "loss": 2.6987, "step": 16880 }, { "epoch": 0.38, "grad_norm": 2.245064881271469, "learning_rate": 1.8680057835276577e-05, "loss": 2.63, "step": 16890 }, { "epoch": 0.38, "grad_norm": 2.1520455343066485, "learning_rate": 1.8678243906451947e-05, "loss": 2.825, "step": 16900 }, { "epoch": 0.38, "grad_norm": 1.897480083495051, "learning_rate": 1.867642882028446e-05, "loss": 2.7234, "step": 16910 }, { "epoch": 0.38, "grad_norm": 1.7351283273623557, "learning_rate": 1.8674612577016172e-05, "loss": 2.7455, "step": 16920 }, { "epoch": 0.38, "grad_norm": 2.3471458764285384, "learning_rate": 1.867279517688931e-05, "loss": 2.758, "step": 16930 }, { "epoch": 0.38, "grad_norm": 1.7762580730199002, "learning_rate": 1.8670976620146233e-05, "loss": 2.6987, "step": 16940 }, { "epoch": 0.38, "grad_norm": 1.8879921475634387, "learning_rate": 1.8669156907029474e-05, "loss": 2.7482, "step": 16950 }, { "epoch": 0.38, "grad_norm": 2.071228111545676, "learning_rate": 1.8667336037781713e-05, "loss": 2.6808, "step": 16960 }, { "epoch": 0.38, "grad_norm": 1.8132848625732272, "learning_rate": 1.866551401264578e-05, "loss": 2.7484, "step": 16970 }, { "epoch": 0.38, "grad_norm": 2.0149081719475976, "learning_rate": 1.8663690831864666e-05, "loss": 2.8102, "step": 16980 }, { "epoch": 0.38, "grad_norm": 2.0983801203632844, "learning_rate": 1.8661866495681508e-05, "loss": 2.7159, "step": 16990 }, { "epoch": 0.38, "grad_norm": 1.6428542255006828, "learning_rate": 1.866004100433961e-05, "loss": 2.6065, "step": 17000 }, { "epoch": 0.38, "grad_norm": 1.7256991321797999, "learning_rate": 1.8658214358082415e-05, "loss": 2.7309, "step": 17010 }, { "epoch": 0.38, "grad_norm": 2.052172673124891, "learning_rate": 1.865638655715353e-05, "loss": 2.7786, "step": 17020 }, { "epoch": 0.38, "grad_norm": 2.096862605734884, "learning_rate": 1.8654557601796714e-05, "loss": 2.6175, "step": 17030 }, { "epoch": 0.38, "grad_norm": 1.8714369734807896, "learning_rate": 1.8652727492255876e-05, "loss": 2.7538, "step": 17040 }, { "epoch": 0.38, "grad_norm": 2.169572720038037, "learning_rate": 1.8650896228775085e-05, "loss": 2.5536, "step": 17050 }, { "epoch": 0.38, "grad_norm": 2.1226610040026666, "learning_rate": 1.8649063811598563e-05, "loss": 2.689, "step": 17060 }, { "epoch": 0.38, "grad_norm": 1.9656501372414472, "learning_rate": 1.8647230240970674e-05, "loss": 2.725, "step": 17070 }, { "epoch": 0.39, "grad_norm": 2.0381950599786465, "learning_rate": 1.8645395517135955e-05, "loss": 2.6998, "step": 17080 }, { "epoch": 0.39, "grad_norm": 1.7948482813810338, "learning_rate": 1.8643559640339087e-05, "loss": 2.7242, "step": 17090 }, { "epoch": 0.39, "grad_norm": 1.92446583978594, "learning_rate": 1.86417226108249e-05, "loss": 2.596, "step": 17100 }, { "epoch": 0.39, "grad_norm": 2.0752477972823167, "learning_rate": 1.8639884428838385e-05, "loss": 2.7258, "step": 17110 }, { "epoch": 0.39, "grad_norm": 2.0517515983974612, "learning_rate": 1.8638045094624686e-05, "loss": 2.7261, "step": 17120 }, { "epoch": 0.39, "grad_norm": 2.710228138214641, "learning_rate": 1.86362046084291e-05, "loss": 2.6136, "step": 17130 }, { "epoch": 0.39, "grad_norm": 1.8167570921039298, "learning_rate": 1.863436297049707e-05, "loss": 2.6704, "step": 17140 }, { "epoch": 0.39, "grad_norm": 1.6927219065586898, "learning_rate": 1.8632520181074207e-05, "loss": 2.7756, "step": 17150 }, { "epoch": 0.39, "grad_norm": 2.260562666317957, "learning_rate": 1.8630676240406265e-05, "loss": 2.7104, "step": 17160 }, { "epoch": 0.39, "grad_norm": 1.8746433983282138, "learning_rate": 1.8628831148739157e-05, "loss": 2.6377, "step": 17170 }, { "epoch": 0.39, "grad_norm": 2.0236720706961604, "learning_rate": 1.8626984906318944e-05, "loss": 2.672, "step": 17180 }, { "epoch": 0.39, "grad_norm": 2.0806179025530707, "learning_rate": 1.8625137513391843e-05, "loss": 2.7694, "step": 17190 }, { "epoch": 0.39, "grad_norm": 1.753738692575318, "learning_rate": 1.862328897020423e-05, "loss": 2.7376, "step": 17200 }, { "epoch": 0.39, "grad_norm": 1.7440261640486519, "learning_rate": 1.862143927700262e-05, "loss": 2.6564, "step": 17210 }, { "epoch": 0.39, "grad_norm": 2.3606388635522033, "learning_rate": 1.8619588434033703e-05, "loss": 2.6802, "step": 17220 }, { "epoch": 0.39, "grad_norm": 1.7501323479299644, "learning_rate": 1.8617736441544297e-05, "loss": 2.7913, "step": 17230 }, { "epoch": 0.39, "grad_norm": 2.362439499377099, "learning_rate": 1.86158832997814e-05, "loss": 2.6398, "step": 17240 }, { "epoch": 0.39, "grad_norm": 1.9227606189945832, "learning_rate": 1.8614029008992137e-05, "loss": 2.7466, "step": 17250 }, { "epoch": 0.39, "grad_norm": 1.878758128432856, "learning_rate": 1.8612173569423805e-05, "loss": 2.7423, "step": 17260 }, { "epoch": 0.39, "grad_norm": 1.7800518976758128, "learning_rate": 1.861031698132385e-05, "loss": 2.7525, "step": 17270 }, { "epoch": 0.39, "grad_norm": 1.8855647360462622, "learning_rate": 1.8608459244939862e-05, "loss": 2.8094, "step": 17280 }, { "epoch": 0.39, "grad_norm": 2.307349617795376, "learning_rate": 1.86066003605196e-05, "loss": 2.7772, "step": 17290 }, { "epoch": 0.39, "grad_norm": 1.9235873102385452, "learning_rate": 1.860474032831096e-05, "loss": 2.6958, "step": 17300 }, { "epoch": 0.39, "grad_norm": 2.200007830447651, "learning_rate": 1.8602879148562007e-05, "loss": 2.7022, "step": 17310 }, { "epoch": 0.39, "grad_norm": 2.1075472839458103, "learning_rate": 1.860101682152094e-05, "loss": 2.6652, "step": 17320 }, { "epoch": 0.39, "grad_norm": 1.9348670596049902, "learning_rate": 1.859915334743613e-05, "loss": 2.6965, "step": 17330 }, { "epoch": 0.39, "grad_norm": 2.015849637480781, "learning_rate": 1.8597288726556086e-05, "loss": 2.6437, "step": 17340 }, { "epoch": 0.39, "grad_norm": 2.5256720174566736, "learning_rate": 1.8595422959129482e-05, "loss": 2.6812, "step": 17350 }, { "epoch": 0.39, "grad_norm": 1.8571064103668593, "learning_rate": 1.8593556045405137e-05, "loss": 2.6456, "step": 17360 }, { "epoch": 0.39, "grad_norm": 2.235799719531208, "learning_rate": 1.8591687985632028e-05, "loss": 2.7812, "step": 17370 }, { "epoch": 0.39, "grad_norm": 1.7777057824009026, "learning_rate": 1.8589818780059277e-05, "loss": 2.6843, "step": 17380 }, { "epoch": 0.39, "grad_norm": 1.5278883151920482, "learning_rate": 1.858794842893617e-05, "loss": 2.6909, "step": 17390 }, { "epoch": 0.39, "grad_norm": 1.8510103741936377, "learning_rate": 1.858607693251213e-05, "loss": 2.737, "step": 17400 }, { "epoch": 0.39, "grad_norm": 1.9150977639401374, "learning_rate": 1.8584204291036755e-05, "loss": 2.7276, "step": 17410 }, { "epoch": 0.39, "grad_norm": 1.9816972755153517, "learning_rate": 1.8582330504759773e-05, "loss": 2.6325, "step": 17420 }, { "epoch": 0.39, "grad_norm": 1.9276758046377565, "learning_rate": 1.858045557393108e-05, "loss": 2.7169, "step": 17430 }, { "epoch": 0.39, "grad_norm": 2.107155013936715, "learning_rate": 1.8578579498800716e-05, "loss": 2.6961, "step": 17440 }, { "epoch": 0.39, "grad_norm": 1.5494263495652187, "learning_rate": 1.8576702279618878e-05, "loss": 2.7358, "step": 17450 }, { "epoch": 0.39, "grad_norm": 1.648055645238316, "learning_rate": 1.857482391663592e-05, "loss": 2.8179, "step": 17460 }, { "epoch": 0.39, "grad_norm": 2.0716501156374956, "learning_rate": 1.8572944410102338e-05, "loss": 2.6198, "step": 17470 }, { "epoch": 0.39, "grad_norm": 2.0533307141059733, "learning_rate": 1.8571063760268785e-05, "loss": 2.8441, "step": 17480 }, { "epoch": 0.39, "grad_norm": 1.7359868634074005, "learning_rate": 1.8569181967386066e-05, "loss": 2.7642, "step": 17490 }, { "epoch": 0.39, "grad_norm": 1.595814662829922, "learning_rate": 1.8567299031705146e-05, "loss": 2.567, "step": 17500 }, { "epoch": 0.39, "grad_norm": 1.6727813200066504, "learning_rate": 1.8565414953477133e-05, "loss": 2.7518, "step": 17510 }, { "epoch": 0.4, "grad_norm": 2.2153450141644484, "learning_rate": 1.856352973295328e-05, "loss": 2.5619, "step": 17520 }, { "epoch": 0.4, "grad_norm": 2.062442290861878, "learning_rate": 1.8561643370385024e-05, "loss": 2.8378, "step": 17530 }, { "epoch": 0.4, "grad_norm": 1.9827062679062988, "learning_rate": 1.8559755866023915e-05, "loss": 2.6906, "step": 17540 }, { "epoch": 0.4, "grad_norm": 1.6268724576633746, "learning_rate": 1.855786722012168e-05, "loss": 2.7065, "step": 17550 }, { "epoch": 0.4, "grad_norm": 2.1625578207667573, "learning_rate": 1.855597743293019e-05, "loss": 2.7377, "step": 17560 }, { "epoch": 0.4, "grad_norm": 1.6793823567793784, "learning_rate": 1.8554086504701472e-05, "loss": 2.5679, "step": 17570 }, { "epoch": 0.4, "grad_norm": 1.8826077485036141, "learning_rate": 1.8552194435687702e-05, "loss": 2.5761, "step": 17580 }, { "epoch": 0.4, "grad_norm": 1.9150651389093991, "learning_rate": 1.8550301226141206e-05, "loss": 2.7165, "step": 17590 }, { "epoch": 0.4, "grad_norm": 2.5769135992325434, "learning_rate": 1.8548406876314466e-05, "loss": 2.7745, "step": 17600 }, { "epoch": 0.4, "grad_norm": 2.5651206211334165, "learning_rate": 1.854651138646012e-05, "loss": 2.6615, "step": 17610 }, { "epoch": 0.4, "grad_norm": 1.9881208473496126, "learning_rate": 1.854461475683095e-05, "loss": 2.7391, "step": 17620 }, { "epoch": 0.4, "grad_norm": 1.5952698918846338, "learning_rate": 1.8542716987679886e-05, "loss": 2.7055, "step": 17630 }, { "epoch": 0.4, "grad_norm": 1.9020896732638684, "learning_rate": 1.8540818079260034e-05, "loss": 2.5775, "step": 17640 }, { "epoch": 0.4, "grad_norm": 2.072343614233518, "learning_rate": 1.853891803182462e-05, "loss": 2.728, "step": 17650 }, { "epoch": 0.4, "grad_norm": 1.7383155186255268, "learning_rate": 1.853701684562704e-05, "loss": 2.7128, "step": 17660 }, { "epoch": 0.4, "grad_norm": 2.6565421190641683, "learning_rate": 1.8535114520920844e-05, "loss": 2.6481, "step": 17670 }, { "epoch": 0.4, "grad_norm": 1.8283543122486883, "learning_rate": 1.8533211057959724e-05, "loss": 2.6319, "step": 17680 }, { "epoch": 0.4, "grad_norm": 1.8026105043265528, "learning_rate": 1.8531306456997534e-05, "loss": 2.6346, "step": 17690 }, { "epoch": 0.4, "grad_norm": 1.605436639399799, "learning_rate": 1.852940071828827e-05, "loss": 2.6838, "step": 17700 }, { "epoch": 0.4, "grad_norm": 1.9984694657130544, "learning_rate": 1.852749384208608e-05, "loss": 2.6207, "step": 17710 }, { "epoch": 0.4, "grad_norm": 2.013675401054261, "learning_rate": 1.8525585828645275e-05, "loss": 2.6814, "step": 17720 }, { "epoch": 0.4, "grad_norm": 2.0731242136627275, "learning_rate": 1.852367667822031e-05, "loss": 2.7802, "step": 17730 }, { "epoch": 0.4, "grad_norm": 1.5932150816920383, "learning_rate": 1.8521766391065784e-05, "loss": 2.7344, "step": 17740 }, { "epoch": 0.4, "grad_norm": 2.1924319480061714, "learning_rate": 1.8519854967436468e-05, "loss": 2.6251, "step": 17750 }, { "epoch": 0.4, "grad_norm": 1.9381705120350785, "learning_rate": 1.851794240758726e-05, "loss": 2.5645, "step": 17760 }, { "epoch": 0.4, "grad_norm": 2.365504907309499, "learning_rate": 1.851602871177323e-05, "loss": 2.7888, "step": 17770 }, { "epoch": 0.4, "grad_norm": 1.8824486453334133, "learning_rate": 1.8514113880249587e-05, "loss": 2.6833, "step": 17780 }, { "epoch": 0.4, "grad_norm": 1.8046137171774996, "learning_rate": 1.8512197913271697e-05, "loss": 2.8551, "step": 17790 }, { "epoch": 0.4, "grad_norm": 1.924842800600316, "learning_rate": 1.8510280811095077e-05, "loss": 2.7192, "step": 17800 }, { "epoch": 0.4, "grad_norm": 1.8902964654272654, "learning_rate": 1.850836257397539e-05, "loss": 2.6754, "step": 17810 }, { "epoch": 0.4, "grad_norm": 1.6479477293666454, "learning_rate": 1.8506443202168462e-05, "loss": 2.6883, "step": 17820 }, { "epoch": 0.4, "grad_norm": 1.8054459189891803, "learning_rate": 1.850452269593026e-05, "loss": 2.7035, "step": 17830 }, { "epoch": 0.4, "grad_norm": 2.4318050277486476, "learning_rate": 1.85026010555169e-05, "loss": 2.7689, "step": 17840 }, { "epoch": 0.4, "grad_norm": 1.868081535382185, "learning_rate": 1.8500678281184666e-05, "loss": 2.6648, "step": 17850 }, { "epoch": 0.4, "grad_norm": 2.452408223386878, "learning_rate": 1.849875437318997e-05, "loss": 2.6176, "step": 17860 }, { "epoch": 0.4, "grad_norm": 1.7791467367287037, "learning_rate": 1.8496829331789393e-05, "loss": 2.7546, "step": 17870 }, { "epoch": 0.4, "grad_norm": 2.0284434714595876, "learning_rate": 1.8494903157239662e-05, "loss": 2.6087, "step": 17880 }, { "epoch": 0.4, "grad_norm": 1.8276844899162572, "learning_rate": 1.849297584979765e-05, "loss": 2.5862, "step": 17890 }, { "epoch": 0.4, "grad_norm": 2.4445707920385926, "learning_rate": 1.8491047409720396e-05, "loss": 2.6727, "step": 17900 }, { "epoch": 0.4, "grad_norm": 1.8761085506149389, "learning_rate": 1.8489117837265067e-05, "loss": 2.7571, "step": 17910 }, { "epoch": 0.4, "grad_norm": 2.04987547886408, "learning_rate": 1.8487187132688995e-05, "loss": 2.6211, "step": 17920 }, { "epoch": 0.4, "grad_norm": 1.790289204019114, "learning_rate": 1.8485255296249673e-05, "loss": 2.7637, "step": 17930 }, { "epoch": 0.4, "grad_norm": 1.4791920132927872, "learning_rate": 1.8483322328204718e-05, "loss": 2.6896, "step": 17940 }, { "epoch": 0.4, "grad_norm": 4.0007509114366595, "learning_rate": 1.8481388228811923e-05, "loss": 2.7047, "step": 17950 }, { "epoch": 0.41, "grad_norm": 1.9530252011332214, "learning_rate": 1.847945299832922e-05, "loss": 2.5892, "step": 17960 }, { "epoch": 0.41, "grad_norm": 2.2399268854522485, "learning_rate": 1.8477516637014697e-05, "loss": 2.7553, "step": 17970 }, { "epoch": 0.41, "grad_norm": 2.49421395381093, "learning_rate": 1.8475579145126582e-05, "loss": 2.6896, "step": 17980 }, { "epoch": 0.41, "grad_norm": 1.6991021198997347, "learning_rate": 1.847364052292327e-05, "loss": 2.6663, "step": 17990 }, { "epoch": 0.41, "grad_norm": 1.9276552307935177, "learning_rate": 1.8471700770663296e-05, "loss": 2.6075, "step": 18000 }, { "epoch": 0.41, "grad_norm": 2.203774251295314, "learning_rate": 1.8469759888605346e-05, "loss": 2.6302, "step": 18010 }, { "epoch": 0.41, "grad_norm": 2.2993655081636124, "learning_rate": 1.8467817877008257e-05, "loss": 2.7428, "step": 18020 }, { "epoch": 0.41, "grad_norm": 1.7505353325870732, "learning_rate": 1.8465874736131023e-05, "loss": 2.6649, "step": 18030 }, { "epoch": 0.41, "grad_norm": 1.579114603300732, "learning_rate": 1.8463930466232782e-05, "loss": 2.7058, "step": 18040 }, { "epoch": 0.41, "grad_norm": 1.8328375288919947, "learning_rate": 1.8461985067572825e-05, "loss": 2.657, "step": 18050 }, { "epoch": 0.41, "grad_norm": 1.9936206784335528, "learning_rate": 1.8460038540410595e-05, "loss": 2.5905, "step": 18060 }, { "epoch": 0.41, "grad_norm": 1.6273413415200928, "learning_rate": 1.8458090885005676e-05, "loss": 2.5997, "step": 18070 }, { "epoch": 0.41, "grad_norm": 2.6512038083406857, "learning_rate": 1.8456142101617824e-05, "loss": 2.7526, "step": 18080 }, { "epoch": 0.41, "grad_norm": 1.8736973846530318, "learning_rate": 1.8454192190506917e-05, "loss": 2.7533, "step": 18090 }, { "epoch": 0.41, "grad_norm": 1.661994986364647, "learning_rate": 1.8452241151933003e-05, "loss": 2.6527, "step": 18100 }, { "epoch": 0.41, "grad_norm": 1.853378026911607, "learning_rate": 1.8450288986156278e-05, "loss": 2.816, "step": 18110 }, { "epoch": 0.41, "grad_norm": 1.9606752435020736, "learning_rate": 1.8448335693437088e-05, "loss": 2.6416, "step": 18120 }, { "epoch": 0.41, "grad_norm": 1.7577439829733106, "learning_rate": 1.8446381274035916e-05, "loss": 2.6467, "step": 18130 }, { "epoch": 0.41, "grad_norm": 1.7442688493054879, "learning_rate": 1.8444425728213416e-05, "loss": 2.7337, "step": 18140 }, { "epoch": 0.41, "grad_norm": 2.0189880835649663, "learning_rate": 1.844246905623038e-05, "loss": 2.6789, "step": 18150 }, { "epoch": 0.41, "grad_norm": 2.2154091731152974, "learning_rate": 1.844051125834775e-05, "loss": 2.6434, "step": 18160 }, { "epoch": 0.41, "grad_norm": 1.8181838397885974, "learning_rate": 1.8438552334826624e-05, "loss": 2.752, "step": 18170 }, { "epoch": 0.41, "grad_norm": 1.8320127441070644, "learning_rate": 1.8436592285928247e-05, "loss": 2.6998, "step": 18180 }, { "epoch": 0.41, "grad_norm": 2.0473419750048665, "learning_rate": 1.843463111191401e-05, "loss": 2.804, "step": 18190 }, { "epoch": 0.41, "grad_norm": 1.8108026506770651, "learning_rate": 1.843266881304546e-05, "loss": 2.8184, "step": 18200 }, { "epoch": 0.41, "grad_norm": 2.270074128117075, "learning_rate": 1.8430705389584293e-05, "loss": 2.5738, "step": 18210 }, { "epoch": 0.41, "grad_norm": 1.7235488440438373, "learning_rate": 1.842874084179235e-05, "loss": 2.6129, "step": 18220 }, { "epoch": 0.41, "grad_norm": 1.9794521547402146, "learning_rate": 1.842677516993163e-05, "loss": 2.6992, "step": 18230 }, { "epoch": 0.41, "grad_norm": 1.8156810808500126, "learning_rate": 1.8424808374264275e-05, "loss": 2.6496, "step": 18240 }, { "epoch": 0.41, "grad_norm": 1.831748709491938, "learning_rate": 1.8422840455052586e-05, "loss": 2.621, "step": 18250 }, { "epoch": 0.41, "grad_norm": 2.538279165326366, "learning_rate": 1.8420871412559e-05, "loss": 2.7532, "step": 18260 }, { "epoch": 0.41, "grad_norm": 2.057852308898526, "learning_rate": 1.8418901247046114e-05, "loss": 2.8095, "step": 18270 }, { "epoch": 0.41, "grad_norm": 1.5923119731130493, "learning_rate": 1.8416929958776674e-05, "loss": 2.7333, "step": 18280 }, { "epoch": 0.41, "grad_norm": 2.3703154896382563, "learning_rate": 1.8414957548013567e-05, "loss": 2.6986, "step": 18290 }, { "epoch": 0.41, "grad_norm": 1.7085273899494189, "learning_rate": 1.8412984015019847e-05, "loss": 2.7485, "step": 18300 }, { "epoch": 0.41, "grad_norm": 1.9257928400270843, "learning_rate": 1.8411009360058698e-05, "loss": 2.679, "step": 18310 }, { "epoch": 0.41, "grad_norm": 2.760301812506775, "learning_rate": 1.8409033583393468e-05, "loss": 2.7606, "step": 18320 }, { "epoch": 0.41, "grad_norm": 2.3003080272677656, "learning_rate": 1.8407056685287647e-05, "loss": 2.6941, "step": 18330 }, { "epoch": 0.41, "grad_norm": 2.1143944087279785, "learning_rate": 1.8405078666004877e-05, "loss": 2.6555, "step": 18340 }, { "epoch": 0.41, "grad_norm": 1.830731713785575, "learning_rate": 1.840309952580895e-05, "loss": 2.6555, "step": 18350 }, { "epoch": 0.41, "grad_norm": 2.063991363162046, "learning_rate": 1.840111926496381e-05, "loss": 2.5789, "step": 18360 }, { "epoch": 0.41, "grad_norm": 1.7076320524593518, "learning_rate": 1.839913788373354e-05, "loss": 2.6219, "step": 18370 }, { "epoch": 0.41, "grad_norm": 2.268102441631159, "learning_rate": 1.8397155382382386e-05, "loss": 2.553, "step": 18380 }, { "epoch": 0.41, "grad_norm": 1.8317145339935448, "learning_rate": 1.8395171761174734e-05, "loss": 2.624, "step": 18390 }, { "epoch": 0.41, "grad_norm": 1.7255974654318669, "learning_rate": 1.8393187020375127e-05, "loss": 2.6986, "step": 18400 }, { "epoch": 0.42, "grad_norm": 1.7467048299700534, "learning_rate": 1.8391201160248243e-05, "loss": 2.706, "step": 18410 }, { "epoch": 0.42, "grad_norm": 2.0657629118922682, "learning_rate": 1.8389214181058932e-05, "loss": 2.633, "step": 18420 }, { "epoch": 0.42, "grad_norm": 1.8613890644059012, "learning_rate": 1.838722608307217e-05, "loss": 2.6388, "step": 18430 }, { "epoch": 0.42, "grad_norm": 1.9240332838634442, "learning_rate": 1.83852368665531e-05, "loss": 2.659, "step": 18440 }, { "epoch": 0.42, "grad_norm": 2.1884372649364296, "learning_rate": 1.8383246531767003e-05, "loss": 2.8088, "step": 18450 }, { "epoch": 0.42, "grad_norm": 1.9330593544916725, "learning_rate": 1.8381255078979314e-05, "loss": 2.7625, "step": 18460 }, { "epoch": 0.42, "grad_norm": 2.2172776027679384, "learning_rate": 1.8379262508455614e-05, "loss": 2.7363, "step": 18470 }, { "epoch": 0.42, "grad_norm": 2.0899161662163084, "learning_rate": 1.8377268820461632e-05, "loss": 2.6925, "step": 18480 }, { "epoch": 0.42, "grad_norm": 2.006515915314581, "learning_rate": 1.8375274015263258e-05, "loss": 2.7602, "step": 18490 }, { "epoch": 0.42, "grad_norm": 1.8538961464873462, "learning_rate": 1.8373278093126518e-05, "loss": 2.7408, "step": 18500 }, { "epoch": 0.42, "grad_norm": 1.798209462133884, "learning_rate": 1.837128105431759e-05, "loss": 2.8725, "step": 18510 }, { "epoch": 0.42, "grad_norm": 1.8395382731861438, "learning_rate": 1.8369282899102798e-05, "loss": 2.7553, "step": 18520 }, { "epoch": 0.42, "grad_norm": 1.6422884241610192, "learning_rate": 1.8367283627748626e-05, "loss": 2.7521, "step": 18530 }, { "epoch": 0.42, "grad_norm": 1.6276039535265607, "learning_rate": 1.8365283240521696e-05, "loss": 2.7987, "step": 18540 }, { "epoch": 0.42, "grad_norm": 1.9603118021981996, "learning_rate": 1.8363281737688785e-05, "loss": 2.6696, "step": 18550 }, { "epoch": 0.42, "grad_norm": 1.6295343516033722, "learning_rate": 1.8361279119516812e-05, "loss": 2.8408, "step": 18560 }, { "epoch": 0.42, "grad_norm": 1.7790130852781907, "learning_rate": 1.8359275386272852e-05, "loss": 2.8552, "step": 18570 }, { "epoch": 0.42, "grad_norm": 1.8848528719345834, "learning_rate": 1.8357270538224128e-05, "loss": 2.6924, "step": 18580 }, { "epoch": 0.42, "grad_norm": 1.7737316342536407, "learning_rate": 1.8355264575638002e-05, "loss": 2.5877, "step": 18590 }, { "epoch": 0.42, "grad_norm": 2.02333154866352, "learning_rate": 1.8353257498781997e-05, "loss": 2.7414, "step": 18600 }, { "epoch": 0.42, "grad_norm": 1.6288821974897496, "learning_rate": 1.8351249307923783e-05, "loss": 2.6769, "step": 18610 }, { "epoch": 0.42, "grad_norm": 2.403035251522631, "learning_rate": 1.8349240003331167e-05, "loss": 2.6422, "step": 18620 }, { "epoch": 0.42, "grad_norm": 2.86433780886025, "learning_rate": 1.8347229585272118e-05, "loss": 2.6855, "step": 18630 }, { "epoch": 0.42, "grad_norm": 2.152966838502077, "learning_rate": 1.8345218054014747e-05, "loss": 2.5894, "step": 18640 }, { "epoch": 0.42, "grad_norm": 1.7044680983826972, "learning_rate": 1.8343205409827314e-05, "loss": 2.5606, "step": 18650 }, { "epoch": 0.42, "grad_norm": 2.1105923930701835, "learning_rate": 1.834119165297823e-05, "loss": 2.6673, "step": 18660 }, { "epoch": 0.42, "grad_norm": 1.6865106200532478, "learning_rate": 1.833917678373605e-05, "loss": 2.8073, "step": 18670 }, { "epoch": 0.42, "grad_norm": 2.360003559535709, "learning_rate": 1.833716080236948e-05, "loss": 2.7695, "step": 18680 }, { "epoch": 0.42, "grad_norm": 2.0606625055751637, "learning_rate": 1.8335143709147382e-05, "loss": 2.7746, "step": 18690 }, { "epoch": 0.42, "grad_norm": 1.9764021828841125, "learning_rate": 1.8333125504338745e-05, "loss": 2.7528, "step": 18700 }, { "epoch": 0.42, "grad_norm": 1.7246478070175342, "learning_rate": 1.8331106188212728e-05, "loss": 2.7291, "step": 18710 }, { "epoch": 0.42, "grad_norm": 1.6633471133921318, "learning_rate": 1.8329085761038628e-05, "loss": 2.6447, "step": 18720 }, { "epoch": 0.42, "grad_norm": 1.7807134359448407, "learning_rate": 1.832706422308589e-05, "loss": 2.6169, "step": 18730 }, { "epoch": 0.42, "grad_norm": 1.9529097851248958, "learning_rate": 1.8325041574624113e-05, "loss": 2.7992, "step": 18740 }, { "epoch": 0.42, "grad_norm": 1.8709296468708374, "learning_rate": 1.832301781592304e-05, "loss": 2.7432, "step": 18750 }, { "epoch": 0.42, "grad_norm": 1.6341001179850692, "learning_rate": 1.832099294725256e-05, "loss": 2.697, "step": 18760 }, { "epoch": 0.42, "grad_norm": 1.6829017495165757, "learning_rate": 1.831896696888271e-05, "loss": 2.732, "step": 18770 }, { "epoch": 0.42, "grad_norm": 2.1769930916243787, "learning_rate": 1.8316939881083686e-05, "loss": 2.7438, "step": 18780 }, { "epoch": 0.42, "grad_norm": 1.7587914945437202, "learning_rate": 1.8314911684125814e-05, "loss": 2.6605, "step": 18790 }, { "epoch": 0.42, "grad_norm": 2.051570445096864, "learning_rate": 1.831288237827958e-05, "loss": 2.6662, "step": 18800 }, { "epoch": 0.42, "grad_norm": 2.0383269562853923, "learning_rate": 1.831085196381562e-05, "loss": 2.6763, "step": 18810 }, { "epoch": 0.42, "grad_norm": 1.8164964281530431, "learning_rate": 1.8308820441004707e-05, "loss": 2.6405, "step": 18820 }, { "epoch": 0.42, "grad_norm": 2.0181321904365785, "learning_rate": 1.830678781011777e-05, "loss": 2.7609, "step": 18830 }, { "epoch": 0.42, "grad_norm": 1.6547658772560048, "learning_rate": 1.830475407142588e-05, "loss": 2.7316, "step": 18840 }, { "epoch": 0.43, "grad_norm": 2.129755026223448, "learning_rate": 1.830271922520027e-05, "loss": 2.6921, "step": 18850 }, { "epoch": 0.43, "grad_norm": 1.815554330816424, "learning_rate": 1.8300683271712295e-05, "loss": 2.7727, "step": 18860 }, { "epoch": 0.43, "grad_norm": 1.8434099305961509, "learning_rate": 1.8298646211233487e-05, "loss": 2.7524, "step": 18870 }, { "epoch": 0.43, "grad_norm": 2.833398396305068, "learning_rate": 1.8296608044035497e-05, "loss": 2.7093, "step": 18880 }, { "epoch": 0.43, "grad_norm": 1.8195590193237003, "learning_rate": 1.829456877039015e-05, "loss": 2.7948, "step": 18890 }, { "epoch": 0.43, "grad_norm": 2.0041256897652917, "learning_rate": 1.82925283905694e-05, "loss": 2.6288, "step": 18900 }, { "epoch": 0.43, "grad_norm": 1.8138467825795077, "learning_rate": 1.8290486904845357e-05, "loss": 2.7736, "step": 18910 }, { "epoch": 0.43, "grad_norm": 2.2322524207777326, "learning_rate": 1.8288444313490278e-05, "loss": 2.7728, "step": 18920 }, { "epoch": 0.43, "grad_norm": 1.7644074435941082, "learning_rate": 1.828640061677656e-05, "loss": 2.7795, "step": 18930 }, { "epoch": 0.43, "grad_norm": 2.0723328333015822, "learning_rate": 1.828435581497676e-05, "loss": 2.661, "step": 18940 }, { "epoch": 0.43, "grad_norm": 1.9174333811079098, "learning_rate": 1.828230990836357e-05, "loss": 2.6612, "step": 18950 }, { "epoch": 0.43, "grad_norm": 1.8105168376568075, "learning_rate": 1.8280262897209842e-05, "loss": 2.7294, "step": 18960 }, { "epoch": 0.43, "grad_norm": 1.7909534368203073, "learning_rate": 1.8278214781788563e-05, "loss": 2.6801, "step": 18970 }, { "epoch": 0.43, "grad_norm": 2.0552008768781205, "learning_rate": 1.8276165562372872e-05, "loss": 2.7241, "step": 18980 }, { "epoch": 0.43, "grad_norm": 1.8070372467521163, "learning_rate": 1.8274115239236056e-05, "loss": 2.7395, "step": 18990 }, { "epoch": 0.43, "grad_norm": 1.695578796513189, "learning_rate": 1.8272063812651554e-05, "loss": 2.7843, "step": 19000 }, { "epoch": 0.43, "grad_norm": 1.9449163008504597, "learning_rate": 1.827001128289294e-05, "loss": 2.5708, "step": 19010 }, { "epoch": 0.43, "grad_norm": 1.8446640185337564, "learning_rate": 1.826795765023395e-05, "loss": 2.6584, "step": 19020 }, { "epoch": 0.43, "grad_norm": 1.9398514382519882, "learning_rate": 1.826590291494845e-05, "loss": 2.7827, "step": 19030 }, { "epoch": 0.43, "grad_norm": 1.6007993114903982, "learning_rate": 1.8263847077310477e-05, "loss": 2.744, "step": 19040 }, { "epoch": 0.43, "grad_norm": 1.8118359223897873, "learning_rate": 1.8261790137594185e-05, "loss": 2.8392, "step": 19050 }, { "epoch": 0.43, "grad_norm": 1.783256269494908, "learning_rate": 1.8259732096073894e-05, "loss": 2.7453, "step": 19060 }, { "epoch": 0.43, "grad_norm": 1.8483004483289598, "learning_rate": 1.8257672953024074e-05, "loss": 2.6756, "step": 19070 }, { "epoch": 0.43, "grad_norm": 2.1353874363332217, "learning_rate": 1.825561270871933e-05, "loss": 2.7149, "step": 19080 }, { "epoch": 0.43, "grad_norm": 2.398834092415735, "learning_rate": 1.8253551363434418e-05, "loss": 2.8006, "step": 19090 }, { "epoch": 0.43, "grad_norm": 1.824595639617488, "learning_rate": 1.8251488917444245e-05, "loss": 2.6661, "step": 19100 }, { "epoch": 0.43, "grad_norm": 1.7920024609745244, "learning_rate": 1.824942537102386e-05, "loss": 2.8304, "step": 19110 }, { "epoch": 0.43, "grad_norm": 2.6224936427889984, "learning_rate": 1.8247360724448464e-05, "loss": 2.6729, "step": 19120 }, { "epoch": 0.43, "grad_norm": 3.223128608145342, "learning_rate": 1.8245294977993392e-05, "loss": 2.5754, "step": 19130 }, { "epoch": 0.43, "grad_norm": 1.953496435424518, "learning_rate": 1.824322813193415e-05, "loss": 2.6585, "step": 19140 }, { "epoch": 0.43, "grad_norm": 1.9547292622162475, "learning_rate": 1.824116018654636e-05, "loss": 2.6878, "step": 19150 }, { "epoch": 0.43, "grad_norm": 2.272203472541914, "learning_rate": 1.8239091142105815e-05, "loss": 2.7162, "step": 19160 }, { "epoch": 0.43, "grad_norm": 2.40670243317455, "learning_rate": 1.8237020998888445e-05, "loss": 2.6546, "step": 19170 }, { "epoch": 0.43, "grad_norm": 2.1762786765683737, "learning_rate": 1.8234949757170325e-05, "loss": 2.7796, "step": 19180 }, { "epoch": 0.43, "grad_norm": 2.022731552691724, "learning_rate": 1.8232877417227677e-05, "loss": 2.6769, "step": 19190 }, { "epoch": 0.43, "grad_norm": 1.9046728897409695, "learning_rate": 1.8230803979336874e-05, "loss": 2.7381, "step": 19200 }, { "epoch": 0.43, "grad_norm": 1.5122032186276588, "learning_rate": 1.8228729443774432e-05, "loss": 2.7193, "step": 19210 }, { "epoch": 0.43, "grad_norm": 2.5420595954084018, "learning_rate": 1.8226653810817018e-05, "loss": 2.7213, "step": 19220 }, { "epoch": 0.43, "grad_norm": 1.8562329405983076, "learning_rate": 1.8224577080741436e-05, "loss": 2.7244, "step": 19230 }, { "epoch": 0.43, "grad_norm": 1.9344803652151805, "learning_rate": 1.8222499253824645e-05, "loss": 2.8057, "step": 19240 }, { "epoch": 0.43, "grad_norm": 2.0316298692104184, "learning_rate": 1.822042033034374e-05, "loss": 2.7464, "step": 19250 }, { "epoch": 0.43, "grad_norm": 2.135852153060799, "learning_rate": 1.8218340310575978e-05, "loss": 2.726, "step": 19260 }, { "epoch": 0.43, "grad_norm": 1.8544411342404665, "learning_rate": 1.8216259194798748e-05, "loss": 2.6576, "step": 19270 }, { "epoch": 0.43, "grad_norm": 1.8515500519875348, "learning_rate": 1.8214176983289595e-05, "loss": 2.5559, "step": 19280 }, { "epoch": 0.44, "grad_norm": 1.723898971352434, "learning_rate": 1.8212093676326202e-05, "loss": 2.7424, "step": 19290 }, { "epoch": 0.44, "grad_norm": 1.7046373737795153, "learning_rate": 1.82100092741864e-05, "loss": 2.6616, "step": 19300 }, { "epoch": 0.44, "grad_norm": 2.6037947462486017, "learning_rate": 1.820792377714817e-05, "loss": 2.6305, "step": 19310 }, { "epoch": 0.44, "grad_norm": 1.4854469112725612, "learning_rate": 1.820583718548964e-05, "loss": 2.6443, "step": 19320 }, { "epoch": 0.44, "grad_norm": 1.802461253737003, "learning_rate": 1.820374949948908e-05, "loss": 2.6121, "step": 19330 }, { "epoch": 0.44, "grad_norm": 2.038654848082819, "learning_rate": 1.8201660719424902e-05, "loss": 2.7375, "step": 19340 }, { "epoch": 0.44, "grad_norm": 1.7927639662096082, "learning_rate": 1.819957084557567e-05, "loss": 2.6832, "step": 19350 }, { "epoch": 0.44, "grad_norm": 1.723418290437915, "learning_rate": 1.8197479878220096e-05, "loss": 2.5812, "step": 19360 }, { "epoch": 0.44, "grad_norm": 1.644219360971015, "learning_rate": 1.8195387817637034e-05, "loss": 2.7165, "step": 19370 }, { "epoch": 0.44, "grad_norm": 2.3543870893961354, "learning_rate": 1.819329466410548e-05, "loss": 2.6111, "step": 19380 }, { "epoch": 0.44, "grad_norm": 2.235396256157343, "learning_rate": 1.8191200417904582e-05, "loss": 2.7432, "step": 19390 }, { "epoch": 0.44, "grad_norm": 1.9890553294848812, "learning_rate": 1.8189105079313633e-05, "loss": 2.729, "step": 19400 }, { "epoch": 0.44, "grad_norm": 1.7452980387544035, "learning_rate": 1.818700864861207e-05, "loss": 2.6602, "step": 19410 }, { "epoch": 0.44, "grad_norm": 2.177204783598027, "learning_rate": 1.8184911126079478e-05, "loss": 2.6615, "step": 19420 }, { "epoch": 0.44, "grad_norm": 1.663521660433675, "learning_rate": 1.8182812511995578e-05, "loss": 2.7105, "step": 19430 }, { "epoch": 0.44, "grad_norm": 2.377830800004827, "learning_rate": 1.818071280664025e-05, "loss": 2.6517, "step": 19440 }, { "epoch": 0.44, "grad_norm": 1.7383708654962153, "learning_rate": 1.8178612010293516e-05, "loss": 2.7173, "step": 19450 }, { "epoch": 0.44, "grad_norm": 1.8932070237790932, "learning_rate": 1.8176510123235537e-05, "loss": 2.715, "step": 19460 }, { "epoch": 0.44, "grad_norm": 1.702948817812986, "learning_rate": 1.817440714574662e-05, "loss": 2.6843, "step": 19470 }, { "epoch": 0.44, "grad_norm": 1.9431544099867146, "learning_rate": 1.817230307810723e-05, "loss": 2.6987, "step": 19480 }, { "epoch": 0.44, "grad_norm": 1.652640804315001, "learning_rate": 1.817019792059796e-05, "loss": 2.7703, "step": 19490 }, { "epoch": 0.44, "grad_norm": 2.058648615744542, "learning_rate": 1.8168091673499568e-05, "loss": 2.6709, "step": 19500 }, { "epoch": 0.44, "grad_norm": 1.802449333107703, "learning_rate": 1.8165984337092934e-05, "loss": 2.7508, "step": 19510 }, { "epoch": 0.44, "grad_norm": 1.5506602646412715, "learning_rate": 1.81638759116591e-05, "loss": 2.7495, "step": 19520 }, { "epoch": 0.44, "grad_norm": 2.021391296242878, "learning_rate": 1.816176639747925e-05, "loss": 2.8401, "step": 19530 }, { "epoch": 0.44, "grad_norm": 1.8306881859348472, "learning_rate": 1.815965579483471e-05, "loss": 2.7029, "step": 19540 }, { "epoch": 0.44, "grad_norm": 1.8047727112307121, "learning_rate": 1.8157544104006956e-05, "loss": 2.5905, "step": 19550 }, { "epoch": 0.44, "grad_norm": 1.9864268463530192, "learning_rate": 1.81554313252776e-05, "loss": 2.8149, "step": 19560 }, { "epoch": 0.44, "grad_norm": 1.8278166653506196, "learning_rate": 1.815331745892841e-05, "loss": 2.6127, "step": 19570 }, { "epoch": 0.44, "grad_norm": 1.60633190331812, "learning_rate": 1.8151202505241294e-05, "loss": 2.6244, "step": 19580 }, { "epoch": 0.44, "grad_norm": 1.8822227604017612, "learning_rate": 1.8149086464498303e-05, "loss": 2.7468, "step": 19590 }, { "epoch": 0.44, "grad_norm": 1.9358461323124134, "learning_rate": 1.8146969336981636e-05, "loss": 2.6519, "step": 19600 }, { "epoch": 0.44, "grad_norm": 1.832339167797039, "learning_rate": 1.8144851122973638e-05, "loss": 2.6759, "step": 19610 }, { "epoch": 0.44, "grad_norm": 2.0853239564787756, "learning_rate": 1.8142731822756796e-05, "loss": 2.8328, "step": 19620 }, { "epoch": 0.44, "grad_norm": 1.911377808405896, "learning_rate": 1.814061143661374e-05, "loss": 2.7123, "step": 19630 }, { "epoch": 0.44, "grad_norm": 1.9707102964619643, "learning_rate": 1.813848996482725e-05, "loss": 2.7159, "step": 19640 }, { "epoch": 0.44, "grad_norm": 2.403571838219083, "learning_rate": 1.8136367407680252e-05, "loss": 2.6703, "step": 19650 }, { "epoch": 0.44, "grad_norm": 2.3913549205013576, "learning_rate": 1.8134243765455805e-05, "loss": 2.6893, "step": 19660 }, { "epoch": 0.44, "grad_norm": 1.90978076937604, "learning_rate": 1.813211903843713e-05, "loss": 2.7964, "step": 19670 }, { "epoch": 0.44, "grad_norm": 2.0777614226331345, "learning_rate": 1.812999322690758e-05, "loss": 2.7756, "step": 19680 }, { "epoch": 0.44, "grad_norm": 2.0272775074028813, "learning_rate": 1.812786633115065e-05, "loss": 2.7019, "step": 19690 }, { "epoch": 0.44, "grad_norm": 1.7109863535037402, "learning_rate": 1.8125738351449995e-05, "loss": 2.7095, "step": 19700 }, { "epoch": 0.44, "grad_norm": 1.5618456574836173, "learning_rate": 1.8123609288089402e-05, "loss": 2.7943, "step": 19710 }, { "epoch": 0.44, "grad_norm": 2.1743195071005337, "learning_rate": 1.8121479141352806e-05, "loss": 2.6231, "step": 19720 }, { "epoch": 0.44, "grad_norm": 1.7225367207079585, "learning_rate": 1.811934791152428e-05, "loss": 2.6903, "step": 19730 }, { "epoch": 0.45, "grad_norm": 1.7980324875448834, "learning_rate": 1.811721559888806e-05, "loss": 2.7277, "step": 19740 }, { "epoch": 0.45, "grad_norm": 2.0166351055925844, "learning_rate": 1.8115082203728506e-05, "loss": 2.6083, "step": 19750 }, { "epoch": 0.45, "grad_norm": 1.6686095174452011, "learning_rate": 1.8112947726330135e-05, "loss": 2.6164, "step": 19760 }, { "epoch": 0.45, "grad_norm": 1.7187193156566756, "learning_rate": 1.81108121669776e-05, "loss": 2.7706, "step": 19770 }, { "epoch": 0.45, "grad_norm": 1.6842736981484863, "learning_rate": 1.81086755259557e-05, "loss": 2.7262, "step": 19780 }, { "epoch": 0.45, "grad_norm": 1.8798686338917419, "learning_rate": 1.8106537803549387e-05, "loss": 2.7115, "step": 19790 }, { "epoch": 0.45, "grad_norm": 1.7797897383192, "learning_rate": 1.810439900004375e-05, "loss": 2.6899, "step": 19800 }, { "epoch": 0.45, "grad_norm": 2.1631430144508585, "learning_rate": 1.810225911572402e-05, "loss": 2.7539, "step": 19810 }, { "epoch": 0.45, "grad_norm": 1.7446174406016368, "learning_rate": 1.8100118150875574e-05, "loss": 2.6624, "step": 19820 }, { "epoch": 0.45, "grad_norm": 1.5524125002370295, "learning_rate": 1.8097976105783935e-05, "loss": 2.6341, "step": 19830 }, { "epoch": 0.45, "grad_norm": 1.6996915964537325, "learning_rate": 1.809583298073477e-05, "loss": 2.7082, "step": 19840 }, { "epoch": 0.45, "grad_norm": 1.7464576932991027, "learning_rate": 1.8093688776013893e-05, "loss": 2.8004, "step": 19850 }, { "epoch": 0.45, "grad_norm": 1.7525501703725612, "learning_rate": 1.8091543491907254e-05, "loss": 2.6951, "step": 19860 }, { "epoch": 0.45, "grad_norm": 1.7120024358250319, "learning_rate": 1.8089397128700946e-05, "loss": 2.8379, "step": 19870 }, { "epoch": 0.45, "grad_norm": 1.6629631714667412, "learning_rate": 1.808724968668122e-05, "loss": 2.7277, "step": 19880 }, { "epoch": 0.45, "grad_norm": 1.8755053554392003, "learning_rate": 1.808510116613446e-05, "loss": 2.6254, "step": 19890 }, { "epoch": 0.45, "grad_norm": 1.929839069897064, "learning_rate": 1.8082951567347193e-05, "loss": 2.6728, "step": 19900 }, { "epoch": 0.45, "grad_norm": 2.1660693416778134, "learning_rate": 1.8080800890606092e-05, "loss": 2.6986, "step": 19910 }, { "epoch": 0.45, "grad_norm": 1.7538181940419704, "learning_rate": 1.8078649136197976e-05, "loss": 2.7633, "step": 19920 }, { "epoch": 0.45, "grad_norm": 1.5701834880630343, "learning_rate": 1.807649630440981e-05, "loss": 2.7295, "step": 19930 }, { "epoch": 0.45, "grad_norm": 1.7339199411324753, "learning_rate": 1.807434239552869e-05, "loss": 2.7139, "step": 19940 }, { "epoch": 0.45, "grad_norm": 1.8612595850726623, "learning_rate": 1.8072187409841873e-05, "loss": 2.8345, "step": 19950 }, { "epoch": 0.45, "grad_norm": 2.1856005223821575, "learning_rate": 1.807003134763674e-05, "loss": 2.787, "step": 19960 }, { "epoch": 0.45, "grad_norm": 3.1984452794506417, "learning_rate": 1.806787420920084e-05, "loss": 2.6333, "step": 19970 }, { "epoch": 0.45, "grad_norm": 2.281850241744834, "learning_rate": 1.8065715994821843e-05, "loss": 2.6736, "step": 19980 }, { "epoch": 0.45, "grad_norm": 2.2092531310554433, "learning_rate": 1.8063556704787574e-05, "loss": 2.692, "step": 19990 }, { "epoch": 0.45, "grad_norm": 1.7522866993012407, "learning_rate": 1.8061396339386e-05, "loss": 2.5918, "step": 20000 }, { "epoch": 0.45, "grad_norm": 1.7813065436112323, "learning_rate": 1.8059234898905226e-05, "loss": 2.8253, "step": 20010 }, { "epoch": 0.45, "grad_norm": 1.6385965040895405, "learning_rate": 1.805707238363351e-05, "loss": 2.671, "step": 20020 }, { "epoch": 0.45, "grad_norm": 2.0205671434034764, "learning_rate": 1.805490879385925e-05, "loss": 2.735, "step": 20030 }, { "epoch": 0.45, "grad_norm": 1.5969428335345448, "learning_rate": 1.8052744129870975e-05, "loss": 2.7573, "step": 20040 }, { "epoch": 0.45, "grad_norm": 2.325512554099412, "learning_rate": 1.8050578391957377e-05, "loss": 2.7049, "step": 20050 }, { "epoch": 0.45, "grad_norm": 2.409648376232075, "learning_rate": 1.804841158040728e-05, "loss": 2.6304, "step": 20060 }, { "epoch": 0.45, "grad_norm": 2.2195306537767276, "learning_rate": 1.804624369550965e-05, "loss": 2.728, "step": 20070 }, { "epoch": 0.45, "grad_norm": 2.0443486719817434, "learning_rate": 1.8044074737553605e-05, "loss": 2.7248, "step": 20080 }, { "epoch": 0.45, "grad_norm": 1.566617210930172, "learning_rate": 1.804190470682839e-05, "loss": 2.7263, "step": 20090 }, { "epoch": 0.45, "grad_norm": 2.000731851011744, "learning_rate": 1.8039733603623417e-05, "loss": 2.7315, "step": 20100 }, { "epoch": 0.45, "grad_norm": 1.9027513995218768, "learning_rate": 1.803756142822821e-05, "loss": 2.5853, "step": 20110 }, { "epoch": 0.45, "grad_norm": 1.6740741649375077, "learning_rate": 1.8035388180932476e-05, "loss": 2.835, "step": 20120 }, { "epoch": 0.45, "grad_norm": 1.661752286998715, "learning_rate": 1.8033213862026024e-05, "loss": 2.7151, "step": 20130 }, { "epoch": 0.45, "grad_norm": 2.100028019128485, "learning_rate": 1.803103847179883e-05, "loss": 2.646, "step": 20140 }, { "epoch": 0.45, "grad_norm": 2.1863399912556787, "learning_rate": 1.802886201054101e-05, "loss": 2.6603, "step": 20150 }, { "epoch": 0.45, "grad_norm": 1.7690660515636527, "learning_rate": 1.8026684478542815e-05, "loss": 2.8338, "step": 20160 }, { "epoch": 0.45, "grad_norm": 1.776395799459516, "learning_rate": 1.8024505876094648e-05, "loss": 2.6901, "step": 20170 }, { "epoch": 0.46, "grad_norm": 2.1217212164948918, "learning_rate": 1.8022326203487044e-05, "loss": 2.6977, "step": 20180 }, { "epoch": 0.46, "grad_norm": 1.8882759815709917, "learning_rate": 1.8020145461010696e-05, "loss": 2.6132, "step": 20190 }, { "epoch": 0.46, "grad_norm": 2.003545523590419, "learning_rate": 1.8017963648956425e-05, "loss": 2.6361, "step": 20200 }, { "epoch": 0.46, "grad_norm": 1.8492074595938905, "learning_rate": 1.8015780767615202e-05, "loss": 2.7281, "step": 20210 }, { "epoch": 0.46, "grad_norm": 1.9635523381468358, "learning_rate": 1.801359681727814e-05, "loss": 2.748, "step": 20220 }, { "epoch": 0.46, "grad_norm": 1.775108743720469, "learning_rate": 1.8011411798236492e-05, "loss": 2.6627, "step": 20230 }, { "epoch": 0.46, "grad_norm": 1.716505704930579, "learning_rate": 1.800922571078166e-05, "loss": 2.6177, "step": 20240 }, { "epoch": 0.46, "grad_norm": 1.9003012785879168, "learning_rate": 1.8007038555205177e-05, "loss": 2.6273, "step": 20250 }, { "epoch": 0.46, "grad_norm": 1.6313599274627792, "learning_rate": 1.8004850331798724e-05, "loss": 2.7095, "step": 20260 }, { "epoch": 0.46, "grad_norm": 1.7280708643072586, "learning_rate": 1.8002661040854133e-05, "loss": 2.6089, "step": 20270 }, { "epoch": 0.46, "grad_norm": 2.2650413549280217, "learning_rate": 1.800047068266337e-05, "loss": 2.7207, "step": 20280 }, { "epoch": 0.46, "grad_norm": 1.8987883823887393, "learning_rate": 1.799827925751854e-05, "loss": 2.7185, "step": 20290 }, { "epoch": 0.46, "grad_norm": 1.9213911582895526, "learning_rate": 1.7996086765711896e-05, "loss": 2.6663, "step": 20300 }, { "epoch": 0.46, "grad_norm": 2.013548590877338, "learning_rate": 1.7993893207535837e-05, "loss": 2.6541, "step": 20310 }, { "epoch": 0.46, "grad_norm": 2.0333592008727748, "learning_rate": 1.7991698583282887e-05, "loss": 2.6683, "step": 20320 }, { "epoch": 0.46, "grad_norm": 1.7370585551488933, "learning_rate": 1.7989502893245733e-05, "loss": 2.675, "step": 20330 }, { "epoch": 0.46, "grad_norm": 1.8227674303874322, "learning_rate": 1.79873061377172e-05, "loss": 2.7179, "step": 20340 }, { "epoch": 0.46, "grad_norm": 1.7571398179954418, "learning_rate": 1.7985108316990236e-05, "loss": 2.6925, "step": 20350 }, { "epoch": 0.46, "grad_norm": 1.6541320710746585, "learning_rate": 1.7982909431357957e-05, "loss": 2.6649, "step": 20360 }, { "epoch": 0.46, "grad_norm": 1.5190691624358759, "learning_rate": 1.7980709481113606e-05, "loss": 2.7694, "step": 20370 }, { "epoch": 0.46, "grad_norm": 2.3920172009197964, "learning_rate": 1.797850846655057e-05, "loss": 2.7181, "step": 20380 }, { "epoch": 0.46, "grad_norm": 2.0199410083822458, "learning_rate": 1.7976306387962387e-05, "loss": 2.7501, "step": 20390 }, { "epoch": 0.46, "grad_norm": 1.7435131458995234, "learning_rate": 1.797410324564272e-05, "loss": 2.7021, "step": 20400 }, { "epoch": 0.46, "grad_norm": 2.294872798477835, "learning_rate": 1.7971899039885383e-05, "loss": 2.6983, "step": 20410 }, { "epoch": 0.46, "grad_norm": 1.7712933377497355, "learning_rate": 1.7969693770984336e-05, "loss": 2.7653, "step": 20420 }, { "epoch": 0.46, "grad_norm": 2.2630328498316548, "learning_rate": 1.796748743923368e-05, "loss": 2.69, "step": 20430 }, { "epoch": 0.46, "grad_norm": 2.0713348498249604, "learning_rate": 1.7965280044927648e-05, "loss": 2.7008, "step": 20440 }, { "epoch": 0.46, "grad_norm": 1.8035034382353627, "learning_rate": 1.7963071588360627e-05, "loss": 2.6031, "step": 20450 }, { "epoch": 0.46, "grad_norm": 1.863242118345726, "learning_rate": 1.7960862069827135e-05, "loss": 2.7411, "step": 20460 }, { "epoch": 0.46, "grad_norm": 1.7669586794017347, "learning_rate": 1.7958651489621836e-05, "loss": 2.536, "step": 20470 }, { "epoch": 0.46, "grad_norm": 1.613307461738401, "learning_rate": 1.795643984803954e-05, "loss": 2.613, "step": 20480 }, { "epoch": 0.46, "grad_norm": 2.0358711424267413, "learning_rate": 1.79542271453752e-05, "loss": 2.597, "step": 20490 }, { "epoch": 0.46, "grad_norm": 1.8583654031023824, "learning_rate": 1.795201338192389e-05, "loss": 2.6172, "step": 20500 }, { "epoch": 0.46, "grad_norm": 1.903134436056768, "learning_rate": 1.794979855798085e-05, "loss": 2.7365, "step": 20510 }, { "epoch": 0.46, "grad_norm": 1.6939430444625219, "learning_rate": 1.794758267384145e-05, "loss": 2.7753, "step": 20520 }, { "epoch": 0.46, "grad_norm": 1.684147792627642, "learning_rate": 1.7945365729801206e-05, "loss": 2.713, "step": 20530 }, { "epoch": 0.46, "grad_norm": 1.7696492551071819, "learning_rate": 1.7943147726155775e-05, "loss": 2.583, "step": 20540 }, { "epoch": 0.46, "grad_norm": 1.7747246158595016, "learning_rate": 1.7940928663200946e-05, "loss": 2.7065, "step": 20550 }, { "epoch": 0.46, "grad_norm": 1.8372215987904648, "learning_rate": 1.7938708541232657e-05, "loss": 2.6177, "step": 20560 }, { "epoch": 0.46, "grad_norm": 1.8839166050880796, "learning_rate": 1.7936487360546992e-05, "loss": 2.7033, "step": 20570 }, { "epoch": 0.46, "grad_norm": 1.6917818618826557, "learning_rate": 1.7934265121440167e-05, "loss": 2.6588, "step": 20580 }, { "epoch": 0.46, "grad_norm": 2.0318881866645, "learning_rate": 1.7932041824208546e-05, "loss": 2.6615, "step": 20590 }, { "epoch": 0.46, "grad_norm": 1.7349356150423045, "learning_rate": 1.7929817469148628e-05, "loss": 2.6079, "step": 20600 }, { "epoch": 0.46, "grad_norm": 1.6118852845288898, "learning_rate": 1.7927592056557056e-05, "loss": 2.7243, "step": 20610 }, { "epoch": 0.47, "grad_norm": 1.6231916264006996, "learning_rate": 1.792536558673062e-05, "loss": 2.8169, "step": 20620 }, { "epoch": 0.47, "grad_norm": 1.772734106366193, "learning_rate": 1.7923138059966235e-05, "loss": 2.7034, "step": 20630 }, { "epoch": 0.47, "grad_norm": 2.3696088262456807, "learning_rate": 1.792090947656097e-05, "loss": 2.7752, "step": 20640 }, { "epoch": 0.47, "grad_norm": 1.7688392458850721, "learning_rate": 1.7918679836812042e-05, "loss": 2.7244, "step": 20650 }, { "epoch": 0.47, "grad_norm": 1.7655929437452775, "learning_rate": 1.7916449141016787e-05, "loss": 2.7246, "step": 20660 }, { "epoch": 0.47, "grad_norm": 2.078775733506154, "learning_rate": 1.79142173894727e-05, "loss": 2.6387, "step": 20670 }, { "epoch": 0.47, "grad_norm": 1.8595299916820154, "learning_rate": 1.791198458247741e-05, "loss": 2.7402, "step": 20680 }, { "epoch": 0.47, "grad_norm": 2.368393840738484, "learning_rate": 1.7909750720328683e-05, "loss": 2.7042, "step": 20690 }, { "epoch": 0.47, "grad_norm": 2.2076276197703955, "learning_rate": 1.7907515803324436e-05, "loss": 2.8076, "step": 20700 }, { "epoch": 0.47, "grad_norm": 1.5601238086032851, "learning_rate": 1.7905279831762717e-05, "loss": 2.7776, "step": 20710 }, { "epoch": 0.47, "grad_norm": 1.6048020885335252, "learning_rate": 1.7903042805941718e-05, "loss": 2.6798, "step": 20720 }, { "epoch": 0.47, "grad_norm": 1.9704163404941388, "learning_rate": 1.7900804726159774e-05, "loss": 2.749, "step": 20730 }, { "epoch": 0.47, "grad_norm": 2.831940562858933, "learning_rate": 1.7898565592715356e-05, "loss": 2.6901, "step": 20740 }, { "epoch": 0.47, "grad_norm": 1.9308976462252416, "learning_rate": 1.7896325405907084e-05, "loss": 2.5941, "step": 20750 }, { "epoch": 0.47, "grad_norm": 2.1313209321946966, "learning_rate": 1.7894084166033702e-05, "loss": 2.6441, "step": 20760 }, { "epoch": 0.47, "grad_norm": 2.0544933881928746, "learning_rate": 1.7891841873394113e-05, "loss": 2.7901, "step": 20770 }, { "epoch": 0.47, "grad_norm": 2.0848280216135513, "learning_rate": 1.788959852828735e-05, "loss": 2.6556, "step": 20780 }, { "epoch": 0.47, "grad_norm": 2.000325266927202, "learning_rate": 1.788735413101259e-05, "loss": 2.72, "step": 20790 }, { "epoch": 0.47, "grad_norm": 1.6566919262406814, "learning_rate": 1.788510868186915e-05, "loss": 2.6078, "step": 20800 }, { "epoch": 0.47, "grad_norm": 1.7948826753595777, "learning_rate": 1.7882862181156482e-05, "loss": 2.7584, "step": 20810 }, { "epoch": 0.47, "grad_norm": 1.8584805439348593, "learning_rate": 1.7880614629174184e-05, "loss": 2.6184, "step": 20820 }, { "epoch": 0.47, "grad_norm": 1.8623558131944793, "learning_rate": 1.7878366026221994e-05, "loss": 2.5703, "step": 20830 }, { "epoch": 0.47, "grad_norm": 1.8489156810826244, "learning_rate": 1.787611637259979e-05, "loss": 2.7271, "step": 20840 }, { "epoch": 0.47, "grad_norm": 2.4328315909966967, "learning_rate": 1.787386566860758e-05, "loss": 2.6918, "step": 20850 }, { "epoch": 0.47, "grad_norm": 2.4192801908945656, "learning_rate": 1.7871613914545535e-05, "loss": 2.7099, "step": 20860 }, { "epoch": 0.47, "grad_norm": 1.972440771468184, "learning_rate": 1.7869361110713944e-05, "loss": 2.6979, "step": 20870 }, { "epoch": 0.47, "grad_norm": 1.9926049713691338, "learning_rate": 1.7867107257413248e-05, "loss": 2.6882, "step": 20880 }, { "epoch": 0.47, "grad_norm": 1.9607441853151502, "learning_rate": 1.7864852354944018e-05, "loss": 2.6002, "step": 20890 }, { "epoch": 0.47, "grad_norm": 2.052469492867211, "learning_rate": 1.7862596403606974e-05, "loss": 2.6705, "step": 20900 }, { "epoch": 0.47, "grad_norm": 1.6302185197547414, "learning_rate": 1.7860339403702976e-05, "loss": 2.7643, "step": 20910 }, { "epoch": 0.47, "grad_norm": 1.790184072800793, "learning_rate": 1.7858081355533015e-05, "loss": 2.7338, "step": 20920 }, { "epoch": 0.47, "grad_norm": 1.8909401524397007, "learning_rate": 1.7855822259398232e-05, "loss": 2.5286, "step": 20930 }, { "epoch": 0.47, "grad_norm": 1.6430838410467867, "learning_rate": 1.7853562115599903e-05, "loss": 2.7336, "step": 20940 }, { "epoch": 0.47, "grad_norm": 1.9389970789008801, "learning_rate": 1.7851300924439442e-05, "loss": 2.7776, "step": 20950 }, { "epoch": 0.47, "grad_norm": 3.0557900158210733, "learning_rate": 1.7849038686218405e-05, "loss": 2.6236, "step": 20960 }, { "epoch": 0.47, "grad_norm": 2.2513838867738354, "learning_rate": 1.784677540123849e-05, "loss": 2.5771, "step": 20970 }, { "epoch": 0.47, "grad_norm": 1.5730125936822865, "learning_rate": 1.7844511069801527e-05, "loss": 2.79, "step": 20980 }, { "epoch": 0.47, "grad_norm": 1.7853659734055676, "learning_rate": 1.7842245692209493e-05, "loss": 2.7313, "step": 20990 }, { "epoch": 0.47, "grad_norm": 1.9831184828565456, "learning_rate": 1.7839979268764504e-05, "loss": 2.6885, "step": 21000 }, { "epoch": 0.47, "grad_norm": 1.695049687392938, "learning_rate": 1.7837711799768807e-05, "loss": 2.6867, "step": 21010 }, { "epoch": 0.47, "grad_norm": 1.9006968965251145, "learning_rate": 1.7835443285524804e-05, "loss": 2.6296, "step": 21020 }, { "epoch": 0.47, "grad_norm": 1.6815883726227028, "learning_rate": 1.783317372633502e-05, "loss": 2.613, "step": 21030 }, { "epoch": 0.47, "grad_norm": 1.899254957848313, "learning_rate": 1.783090312250213e-05, "loss": 2.5697, "step": 21040 }, { "epoch": 0.47, "grad_norm": 1.8035686557116188, "learning_rate": 1.7828631474328942e-05, "loss": 2.7907, "step": 21050 }, { "epoch": 0.47, "grad_norm": 1.7619263563687015, "learning_rate": 1.782635878211841e-05, "loss": 2.5518, "step": 21060 }, { "epoch": 0.48, "grad_norm": 1.7165904051190046, "learning_rate": 1.7824085046173626e-05, "loss": 2.7656, "step": 21070 }, { "epoch": 0.48, "grad_norm": 1.5838757710835076, "learning_rate": 1.7821810266797808e-05, "loss": 2.7497, "step": 21080 }, { "epoch": 0.48, "grad_norm": 1.486599125988114, "learning_rate": 1.781953444429433e-05, "loss": 2.6909, "step": 21090 }, { "epoch": 0.48, "grad_norm": 2.849088213570644, "learning_rate": 1.7817257578966703e-05, "loss": 2.6274, "step": 21100 }, { "epoch": 0.48, "grad_norm": 2.358578161388383, "learning_rate": 1.7814979671118567e-05, "loss": 2.6994, "step": 21110 }, { "epoch": 0.48, "grad_norm": 1.9172614319980361, "learning_rate": 1.781270072105371e-05, "loss": 2.6899, "step": 21120 }, { "epoch": 0.48, "grad_norm": 1.7032641661797572, "learning_rate": 1.7810420729076057e-05, "loss": 2.7957, "step": 21130 }, { "epoch": 0.48, "grad_norm": 1.8968407210576266, "learning_rate": 1.7808139695489666e-05, "loss": 2.6948, "step": 21140 }, { "epoch": 0.48, "grad_norm": 1.7904369942371126, "learning_rate": 1.7805857620598744e-05, "loss": 2.6754, "step": 21150 }, { "epoch": 0.48, "grad_norm": 1.6249723575019421, "learning_rate": 1.7803574504707627e-05, "loss": 2.6924, "step": 21160 }, { "epoch": 0.48, "grad_norm": 1.7257006615252959, "learning_rate": 1.7801290348120798e-05, "loss": 2.7192, "step": 21170 }, { "epoch": 0.48, "grad_norm": 1.7199041739067031, "learning_rate": 1.7799005151142876e-05, "loss": 2.6053, "step": 21180 }, { "epoch": 0.48, "grad_norm": 1.8304899645588284, "learning_rate": 1.779671891407861e-05, "loss": 2.8466, "step": 21190 }, { "epoch": 0.48, "grad_norm": 1.9174132739248464, "learning_rate": 1.779443163723291e-05, "loss": 2.5722, "step": 21200 }, { "epoch": 0.48, "grad_norm": 2.1158769218471782, "learning_rate": 1.77921433209108e-05, "loss": 2.7146, "step": 21210 }, { "epoch": 0.48, "grad_norm": 1.6584368499068125, "learning_rate": 1.7789853965417455e-05, "loss": 2.6696, "step": 21220 }, { "epoch": 0.48, "grad_norm": 1.7177619068785759, "learning_rate": 1.778756357105819e-05, "loss": 2.7177, "step": 21230 }, { "epoch": 0.48, "grad_norm": 2.0327652038474255, "learning_rate": 1.778527213813845e-05, "loss": 2.6538, "step": 21240 }, { "epoch": 0.48, "grad_norm": 2.1256241710743122, "learning_rate": 1.7782979666963827e-05, "loss": 2.6648, "step": 21250 }, { "epoch": 0.48, "grad_norm": 1.6542787154515048, "learning_rate": 1.778068615784005e-05, "loss": 2.5764, "step": 21260 }, { "epoch": 0.48, "grad_norm": 2.1290853821587414, "learning_rate": 1.777839161107298e-05, "loss": 2.728, "step": 21270 }, { "epoch": 0.48, "grad_norm": 1.7362453187801954, "learning_rate": 1.7776096026968628e-05, "loss": 2.6682, "step": 21280 }, { "epoch": 0.48, "grad_norm": 1.8475895204305828, "learning_rate": 1.777379940583313e-05, "loss": 2.6013, "step": 21290 }, { "epoch": 0.48, "grad_norm": 1.9107455752496423, "learning_rate": 1.7771501747972766e-05, "loss": 2.738, "step": 21300 }, { "epoch": 0.48, "grad_norm": 1.9815424501103942, "learning_rate": 1.7769203053693958e-05, "loss": 2.6683, "step": 21310 }, { "epoch": 0.48, "grad_norm": 1.5085694481317526, "learning_rate": 1.776690332330326e-05, "loss": 2.7588, "step": 21320 }, { "epoch": 0.48, "grad_norm": 1.8017028501456922, "learning_rate": 1.7764602557107374e-05, "loss": 2.7275, "step": 21330 }, { "epoch": 0.48, "grad_norm": 1.9566090553361613, "learning_rate": 1.7762300755413128e-05, "loss": 2.7254, "step": 21340 }, { "epoch": 0.48, "grad_norm": 1.9335783657096939, "learning_rate": 1.7759997918527493e-05, "loss": 2.7357, "step": 21350 }, { "epoch": 0.48, "grad_norm": 1.7575342796737632, "learning_rate": 1.775769404675758e-05, "loss": 2.7723, "step": 21360 }, { "epoch": 0.48, "grad_norm": 2.6068872862715846, "learning_rate": 1.775538914041064e-05, "loss": 2.7139, "step": 21370 }, { "epoch": 0.48, "grad_norm": 1.6570851856318813, "learning_rate": 1.7753083199794054e-05, "loss": 2.7716, "step": 21380 }, { "epoch": 0.48, "grad_norm": 1.7871059032803018, "learning_rate": 1.775077622521534e-05, "loss": 2.7991, "step": 21390 }, { "epoch": 0.48, "grad_norm": 1.693627790036093, "learning_rate": 1.7748468216982174e-05, "loss": 2.6558, "step": 21400 }, { "epoch": 0.48, "grad_norm": 1.9622049791855345, "learning_rate": 1.7746159175402347e-05, "loss": 2.6687, "step": 21410 }, { "epoch": 0.48, "grad_norm": 1.608182249260063, "learning_rate": 1.7743849100783794e-05, "loss": 2.6184, "step": 21420 }, { "epoch": 0.48, "grad_norm": 2.185413455363706, "learning_rate": 1.7741537993434593e-05, "loss": 2.719, "step": 21430 }, { "epoch": 0.48, "grad_norm": 2.1403888888341838, "learning_rate": 1.7739225853662958e-05, "loss": 2.8029, "step": 21440 }, { "epoch": 0.48, "grad_norm": 1.8119226014542327, "learning_rate": 1.7736912681777235e-05, "loss": 2.7413, "step": 21450 }, { "epoch": 0.48, "grad_norm": 1.6064978385762947, "learning_rate": 1.7734598478085913e-05, "loss": 2.7688, "step": 21460 }, { "epoch": 0.48, "grad_norm": 2.1934353381679887, "learning_rate": 1.7732283242897617e-05, "loss": 2.6084, "step": 21470 }, { "epoch": 0.48, "grad_norm": 1.7237906257809537, "learning_rate": 1.772996697652112e-05, "loss": 2.7348, "step": 21480 }, { "epoch": 0.48, "grad_norm": 2.287758585810658, "learning_rate": 1.7727649679265306e-05, "loss": 2.7184, "step": 21490 }, { "epoch": 0.48, "grad_norm": 2.079970671389805, "learning_rate": 1.772533135143922e-05, "loss": 2.7196, "step": 21500 }, { "epoch": 0.49, "grad_norm": 1.9160614476271782, "learning_rate": 1.772301199335204e-05, "loss": 2.6485, "step": 21510 }, { "epoch": 0.49, "grad_norm": 1.8382181143158958, "learning_rate": 1.772069160531308e-05, "loss": 2.6844, "step": 21520 }, { "epoch": 0.49, "grad_norm": 1.8734671679020645, "learning_rate": 1.771837018763179e-05, "loss": 2.547, "step": 21530 }, { "epoch": 0.49, "grad_norm": 1.682476559569447, "learning_rate": 1.771604774061775e-05, "loss": 2.7138, "step": 21540 }, { "epoch": 0.49, "grad_norm": 2.1211092695846276, "learning_rate": 1.771372426458069e-05, "loss": 2.57, "step": 21550 }, { "epoch": 0.49, "grad_norm": 1.7943146334619542, "learning_rate": 1.7711399759830477e-05, "loss": 2.5646, "step": 21560 }, { "epoch": 0.49, "grad_norm": 1.7247631472769778, "learning_rate": 1.77090742266771e-05, "loss": 2.7682, "step": 21570 }, { "epoch": 0.49, "grad_norm": 1.9720394892296593, "learning_rate": 1.7706747665430703e-05, "loss": 2.6559, "step": 21580 }, { "epoch": 0.49, "grad_norm": 1.8810381770728017, "learning_rate": 1.7704420076401556e-05, "loss": 2.6813, "step": 21590 }, { "epoch": 0.49, "grad_norm": 1.8982513434101904, "learning_rate": 1.7702091459900072e-05, "loss": 2.6742, "step": 21600 }, { "epoch": 0.49, "grad_norm": 1.8601802889666859, "learning_rate": 1.7699761816236797e-05, "loss": 2.6773, "step": 21610 }, { "epoch": 0.49, "grad_norm": 1.5591863545706428, "learning_rate": 1.7697431145722412e-05, "loss": 2.5778, "step": 21620 }, { "epoch": 0.49, "grad_norm": 2.14046422380736, "learning_rate": 1.7695099448667748e-05, "loss": 2.6069, "step": 21630 }, { "epoch": 0.49, "grad_norm": 1.9623299940334293, "learning_rate": 1.7692766725383754e-05, "loss": 2.7362, "step": 21640 }, { "epoch": 0.49, "grad_norm": 2.3609093167544937, "learning_rate": 1.769043297618153e-05, "loss": 2.6917, "step": 21650 }, { "epoch": 0.49, "grad_norm": 1.77188238233252, "learning_rate": 1.7688098201372307e-05, "loss": 2.6149, "step": 21660 }, { "epoch": 0.49, "grad_norm": 1.77731247764444, "learning_rate": 1.7685762401267458e-05, "loss": 2.7308, "step": 21670 }, { "epoch": 0.49, "grad_norm": 1.8650079069660146, "learning_rate": 1.768342557617848e-05, "loss": 2.6639, "step": 21680 }, { "epoch": 0.49, "grad_norm": 1.898389761166727, "learning_rate": 1.768108772641702e-05, "loss": 2.5904, "step": 21690 }, { "epoch": 0.49, "grad_norm": 1.6244697455549357, "learning_rate": 1.7678748852294857e-05, "loss": 2.727, "step": 21700 }, { "epoch": 0.49, "grad_norm": 2.445710423189469, "learning_rate": 1.7676408954123906e-05, "loss": 2.6694, "step": 21710 }, { "epoch": 0.49, "grad_norm": 2.022889235959734, "learning_rate": 1.767406803221622e-05, "loss": 2.7341, "step": 21720 }, { "epoch": 0.49, "grad_norm": 1.7697891887850006, "learning_rate": 1.7671726086883988e-05, "loss": 2.7173, "step": 21730 }, { "epoch": 0.49, "grad_norm": 2.053184923337992, "learning_rate": 1.7669383118439537e-05, "loss": 2.7056, "step": 21740 }, { "epoch": 0.49, "grad_norm": 1.9488693003471562, "learning_rate": 1.766703912719532e-05, "loss": 2.7623, "step": 21750 }, { "epoch": 0.49, "grad_norm": 1.6935226069763791, "learning_rate": 1.7664694113463943e-05, "loss": 2.7021, "step": 21760 }, { "epoch": 0.49, "grad_norm": 1.6952809236063404, "learning_rate": 1.766234807755814e-05, "loss": 2.7227, "step": 21770 }, { "epoch": 0.49, "grad_norm": 1.9933149482740242, "learning_rate": 1.766000101979078e-05, "loss": 2.7868, "step": 21780 }, { "epoch": 0.49, "grad_norm": 2.2766897422952734, "learning_rate": 1.7657652940474868e-05, "loss": 2.6576, "step": 21790 }, { "epoch": 0.49, "grad_norm": 2.211110919182422, "learning_rate": 1.765530383992355e-05, "loss": 2.7176, "step": 21800 }, { "epoch": 0.49, "grad_norm": 1.8139475380806356, "learning_rate": 1.76529537184501e-05, "loss": 2.619, "step": 21810 }, { "epoch": 0.49, "grad_norm": 2.0502608669771716, "learning_rate": 1.7650602576367942e-05, "loss": 2.7538, "step": 21820 }, { "epoch": 0.49, "grad_norm": 1.8604258359891908, "learning_rate": 1.7648250413990624e-05, "loss": 2.7405, "step": 21830 }, { "epoch": 0.49, "grad_norm": 2.3838671403378964, "learning_rate": 1.7645897231631834e-05, "loss": 2.7273, "step": 21840 }, { "epoch": 0.49, "grad_norm": 1.8651703564809954, "learning_rate": 1.764354302960539e-05, "loss": 2.5752, "step": 21850 }, { "epoch": 0.49, "grad_norm": 2.3415780368987957, "learning_rate": 1.7641187808225263e-05, "loss": 2.7851, "step": 21860 }, { "epoch": 0.49, "grad_norm": 2.026461565520274, "learning_rate": 1.7638831567805537e-05, "loss": 2.6505, "step": 21870 }, { "epoch": 0.49, "grad_norm": 1.6063540260924751, "learning_rate": 1.763647430866045e-05, "loss": 2.8333, "step": 21880 }, { "epoch": 0.49, "grad_norm": 1.702798847730682, "learning_rate": 1.763411603110437e-05, "loss": 2.824, "step": 21890 }, { "epoch": 0.49, "grad_norm": 1.658096093937909, "learning_rate": 1.7631756735451795e-05, "loss": 2.7898, "step": 21900 }, { "epoch": 0.49, "grad_norm": 1.722507722543026, "learning_rate": 1.762939642201737e-05, "loss": 2.6741, "step": 21910 }, { "epoch": 0.49, "grad_norm": 1.7673850737402446, "learning_rate": 1.7627035091115864e-05, "loss": 2.6755, "step": 21920 }, { "epoch": 0.49, "grad_norm": 2.0773313654837966, "learning_rate": 1.7624672743062195e-05, "loss": 2.7913, "step": 21930 }, { "epoch": 0.49, "grad_norm": 1.8609150066188884, "learning_rate": 1.7622309378171402e-05, "loss": 2.648, "step": 21940 }, { "epoch": 0.49, "grad_norm": 1.7935025003651752, "learning_rate": 1.761994499675867e-05, "loss": 2.7704, "step": 21950 }, { "epoch": 0.5, "grad_norm": 1.6526099863643342, "learning_rate": 1.7617579599139317e-05, "loss": 2.6738, "step": 21960 }, { "epoch": 0.5, "grad_norm": 2.140716664025171, "learning_rate": 1.761521318562879e-05, "loss": 2.7149, "step": 21970 }, { "epoch": 0.5, "grad_norm": 1.722931554824386, "learning_rate": 1.7612845756542688e-05, "loss": 2.7467, "step": 21980 }, { "epoch": 0.5, "grad_norm": 2.06932434497539, "learning_rate": 1.7610477312196724e-05, "loss": 2.6863, "step": 21990 }, { "epoch": 0.5, "grad_norm": 1.9046868023004078, "learning_rate": 1.7608107852906762e-05, "loss": 2.7489, "step": 22000 }, { "epoch": 0.5, "grad_norm": 2.1741096580341734, "learning_rate": 1.7605737378988793e-05, "loss": 2.7091, "step": 22010 }, { "epoch": 0.5, "grad_norm": 1.8573761196973266, "learning_rate": 1.7603365890758956e-05, "loss": 2.67, "step": 22020 }, { "epoch": 0.5, "grad_norm": 1.7850070498800878, "learning_rate": 1.7600993388533502e-05, "loss": 2.7056, "step": 22030 }, { "epoch": 0.5, "grad_norm": 1.6251844182625488, "learning_rate": 1.7598619872628843e-05, "loss": 2.5498, "step": 22040 }, { "epoch": 0.5, "grad_norm": 1.8337138946090572, "learning_rate": 1.759624534336151e-05, "loss": 2.588, "step": 22050 }, { "epoch": 0.5, "grad_norm": 1.5603913307031847, "learning_rate": 1.7593869801048175e-05, "loss": 2.773, "step": 22060 }, { "epoch": 0.5, "grad_norm": 1.7850991755932097, "learning_rate": 1.7591493246005642e-05, "loss": 2.6568, "step": 22070 }, { "epoch": 0.5, "grad_norm": 1.6255278404980067, "learning_rate": 1.758911567855085e-05, "loss": 2.6258, "step": 22080 }, { "epoch": 0.5, "grad_norm": 2.2075888638996766, "learning_rate": 1.7586737099000876e-05, "loss": 2.6786, "step": 22090 }, { "epoch": 0.5, "grad_norm": 1.5650533940272244, "learning_rate": 1.7584357507672937e-05, "loss": 2.7178, "step": 22100 }, { "epoch": 0.5, "grad_norm": 2.004323094017556, "learning_rate": 1.758197690488437e-05, "loss": 2.5828, "step": 22110 }, { "epoch": 0.5, "grad_norm": 1.7341902806619032, "learning_rate": 1.757959529095266e-05, "loss": 2.6129, "step": 22120 }, { "epoch": 0.5, "grad_norm": 1.5707287207405385, "learning_rate": 1.757721266619542e-05, "loss": 2.8562, "step": 22130 }, { "epoch": 0.5, "grad_norm": 1.6840067216483576, "learning_rate": 1.75748290309304e-05, "loss": 2.5989, "step": 22140 }, { "epoch": 0.5, "grad_norm": 1.7511350250947257, "learning_rate": 1.757244438547549e-05, "loss": 2.6175, "step": 22150 }, { "epoch": 0.5, "grad_norm": 1.61064317603792, "learning_rate": 1.7570058730148704e-05, "loss": 2.6241, "step": 22160 }, { "epoch": 0.5, "grad_norm": 1.9798957733683797, "learning_rate": 1.7567672065268196e-05, "loss": 2.7436, "step": 22170 }, { "epoch": 0.5, "grad_norm": 1.6434706853533827, "learning_rate": 1.7565284391152265e-05, "loss": 2.702, "step": 22180 }, { "epoch": 0.5, "grad_norm": 1.774243301321108, "learning_rate": 1.7562895708119317e-05, "loss": 2.6565, "step": 22190 }, { "epoch": 0.5, "grad_norm": 1.6903833754708126, "learning_rate": 1.756050601648793e-05, "loss": 2.7207, "step": 22200 }, { "epoch": 0.5, "grad_norm": 2.1971781150015537, "learning_rate": 1.755811531657678e-05, "loss": 2.4832, "step": 22210 }, { "epoch": 0.5, "grad_norm": 1.5934163076753947, "learning_rate": 1.7555723608704705e-05, "loss": 2.7679, "step": 22220 }, { "epoch": 0.5, "grad_norm": 1.6126202105884173, "learning_rate": 1.755333089319066e-05, "loss": 2.663, "step": 22230 }, { "epoch": 0.5, "grad_norm": 1.8575638795147769, "learning_rate": 1.7550937170353745e-05, "loss": 2.6084, "step": 22240 }, { "epoch": 0.5, "grad_norm": 2.077940499965233, "learning_rate": 1.7548542440513187e-05, "loss": 2.5857, "step": 22250 }, { "epoch": 0.5, "grad_norm": 1.7851476485250057, "learning_rate": 1.7546146703988355e-05, "loss": 2.7035, "step": 22260 }, { "epoch": 0.5, "grad_norm": 1.7603356675645438, "learning_rate": 1.7543749961098744e-05, "loss": 2.7823, "step": 22270 }, { "epoch": 0.5, "grad_norm": 1.7882597197628982, "learning_rate": 1.7541352212163986e-05, "loss": 2.7148, "step": 22280 }, { "epoch": 0.5, "grad_norm": 1.612239397572481, "learning_rate": 1.7538953457503854e-05, "loss": 2.6733, "step": 22290 }, { "epoch": 0.5, "grad_norm": 1.7844039214289313, "learning_rate": 1.7536553697438244e-05, "loss": 2.8684, "step": 22300 }, { "epoch": 0.5, "grad_norm": 1.7399811237994653, "learning_rate": 1.753415293228719e-05, "loss": 2.6981, "step": 22310 }, { "epoch": 0.5, "grad_norm": 1.8128751623615282, "learning_rate": 1.7531751162370866e-05, "loss": 2.6165, "step": 22320 }, { "epoch": 0.5, "grad_norm": 1.8199972889486353, "learning_rate": 1.7529348388009574e-05, "loss": 2.6638, "step": 22330 }, { "epoch": 0.5, "grad_norm": 2.890732978896469, "learning_rate": 1.7526944609523754e-05, "loss": 2.5752, "step": 22340 }, { "epoch": 0.5, "grad_norm": 1.82577965836627, "learning_rate": 1.752453982723397e-05, "loss": 2.6387, "step": 22350 }, { "epoch": 0.5, "grad_norm": 1.8485576850761132, "learning_rate": 1.7522134041460934e-05, "loss": 2.7801, "step": 22360 }, { "epoch": 0.5, "grad_norm": 1.6792619810019607, "learning_rate": 1.7519727252525483e-05, "loss": 2.7568, "step": 22370 }, { "epoch": 0.5, "grad_norm": 1.7762005602108835, "learning_rate": 1.751731946074859e-05, "loss": 2.6509, "step": 22380 }, { "epoch": 0.5, "grad_norm": 1.7911627987254766, "learning_rate": 1.751491066645136e-05, "loss": 2.6084, "step": 22390 }, { "epoch": 0.51, "grad_norm": 1.9646039264268562, "learning_rate": 1.7512500869955033e-05, "loss": 2.6521, "step": 22400 }, { "epoch": 0.51, "grad_norm": 1.6944471393982168, "learning_rate": 1.7510090071580985e-05, "loss": 2.6839, "step": 22410 }, { "epoch": 0.51, "grad_norm": 1.8956993797976518, "learning_rate": 1.7507678271650722e-05, "loss": 2.6932, "step": 22420 }, { "epoch": 0.51, "grad_norm": 1.9721754282037347, "learning_rate": 1.7505265470485885e-05, "loss": 2.7024, "step": 22430 }, { "epoch": 0.51, "grad_norm": 1.7188651246231688, "learning_rate": 1.750285166840825e-05, "loss": 2.6364, "step": 22440 }, { "epoch": 0.51, "grad_norm": 1.8880890501380927, "learning_rate": 1.7500436865739725e-05, "loss": 2.858, "step": 22450 }, { "epoch": 0.51, "grad_norm": 1.736522784350448, "learning_rate": 1.749802106280235e-05, "loss": 2.6285, "step": 22460 }, { "epoch": 0.51, "grad_norm": 2.3152048770140876, "learning_rate": 1.74956042599183e-05, "loss": 2.7763, "step": 22470 }, { "epoch": 0.51, "grad_norm": 1.748380995083513, "learning_rate": 1.7493186457409886e-05, "loss": 2.62, "step": 22480 }, { "epoch": 0.51, "grad_norm": 1.7656830373643808, "learning_rate": 1.7490767655599542e-05, "loss": 2.6912, "step": 22490 }, { "epoch": 0.51, "grad_norm": 1.8033876921825556, "learning_rate": 1.7488347854809853e-05, "loss": 2.6186, "step": 22500 }, { "epoch": 0.51, "grad_norm": 1.797413743081371, "learning_rate": 1.7485927055363522e-05, "loss": 2.6929, "step": 22510 }, { "epoch": 0.51, "grad_norm": 2.200375329740435, "learning_rate": 1.748350525758339e-05, "loss": 2.5945, "step": 22520 }, { "epoch": 0.51, "grad_norm": 2.319779251728127, "learning_rate": 1.7481082461792427e-05, "loss": 2.6911, "step": 22530 }, { "epoch": 0.51, "grad_norm": 2.2455392529626024, "learning_rate": 1.7478658668313753e-05, "loss": 2.7132, "step": 22540 }, { "epoch": 0.51, "grad_norm": 1.9942159460361506, "learning_rate": 1.7476233877470596e-05, "loss": 2.6415, "step": 22550 }, { "epoch": 0.51, "grad_norm": 1.732020853155954, "learning_rate": 1.7473808089586332e-05, "loss": 2.7012, "step": 22560 }, { "epoch": 0.51, "grad_norm": 1.8603024812239806, "learning_rate": 1.7471381304984473e-05, "loss": 2.5678, "step": 22570 }, { "epoch": 0.51, "grad_norm": 1.7961439772531358, "learning_rate": 1.7468953523988655e-05, "loss": 2.6376, "step": 22580 }, { "epoch": 0.51, "grad_norm": 1.9618710817264626, "learning_rate": 1.7466524746922654e-05, "loss": 2.6107, "step": 22590 }, { "epoch": 0.51, "grad_norm": 2.1462520181281675, "learning_rate": 1.7464094974110368e-05, "loss": 2.8106, "step": 22600 }, { "epoch": 0.51, "grad_norm": 1.9096984202359455, "learning_rate": 1.7461664205875838e-05, "loss": 2.6327, "step": 22610 }, { "epoch": 0.51, "grad_norm": 1.8974459886954076, "learning_rate": 1.7459232442543237e-05, "loss": 2.6246, "step": 22620 }, { "epoch": 0.51, "grad_norm": 1.8519368501839248, "learning_rate": 1.7456799684436864e-05, "loss": 2.6373, "step": 22630 }, { "epoch": 0.51, "grad_norm": 1.8394457596188227, "learning_rate": 1.745436593188116e-05, "loss": 2.7601, "step": 22640 }, { "epoch": 0.51, "grad_norm": 1.6279699565368195, "learning_rate": 1.7451931185200694e-05, "loss": 2.7239, "step": 22650 }, { "epoch": 0.51, "grad_norm": 1.8739420356506187, "learning_rate": 1.744949544472016e-05, "loss": 2.6575, "step": 22660 }, { "epoch": 0.51, "grad_norm": 1.766473691877, "learning_rate": 1.7447058710764397e-05, "loss": 2.6382, "step": 22670 }, { "epoch": 0.51, "grad_norm": 1.9527446200955068, "learning_rate": 1.7444620983658373e-05, "loss": 2.7007, "step": 22680 }, { "epoch": 0.51, "grad_norm": 1.6516689834975564, "learning_rate": 1.7442182263727182e-05, "loss": 2.7018, "step": 22690 }, { "epoch": 0.51, "grad_norm": 1.6772483973963856, "learning_rate": 1.743974255129606e-05, "loss": 2.718, "step": 22700 }, { "epoch": 0.51, "grad_norm": 1.652679703406694, "learning_rate": 1.7437301846690366e-05, "loss": 2.6938, "step": 22710 }, { "epoch": 0.51, "grad_norm": 1.6555948173876525, "learning_rate": 1.74348601502356e-05, "loss": 2.6651, "step": 22720 }, { "epoch": 0.51, "grad_norm": 1.9741906976032972, "learning_rate": 1.7432417462257385e-05, "loss": 2.6521, "step": 22730 }, { "epoch": 0.51, "grad_norm": 1.915833387009373, "learning_rate": 1.742997378308149e-05, "loss": 2.7281, "step": 22740 }, { "epoch": 0.51, "grad_norm": 1.7308379850789235, "learning_rate": 1.7427529113033794e-05, "loss": 2.6144, "step": 22750 }, { "epoch": 0.51, "grad_norm": 2.3846789698554227, "learning_rate": 1.7425083452440335e-05, "loss": 2.7437, "step": 22760 }, { "epoch": 0.51, "grad_norm": 1.6831291077067558, "learning_rate": 1.742263680162726e-05, "loss": 2.648, "step": 22770 }, { "epoch": 0.51, "grad_norm": 1.5571085551066184, "learning_rate": 1.7420189160920867e-05, "loss": 2.6783, "step": 22780 }, { "epoch": 0.51, "grad_norm": 1.6245139611550994, "learning_rate": 1.741774053064757e-05, "loss": 2.6441, "step": 22790 }, { "epoch": 0.51, "grad_norm": 1.8402523320570723, "learning_rate": 1.7415290911133922e-05, "loss": 2.6548, "step": 22800 }, { "epoch": 0.51, "grad_norm": 2.3164900747704777, "learning_rate": 1.741284030270661e-05, "loss": 2.665, "step": 22810 }, { "epoch": 0.51, "grad_norm": 1.8259931627021626, "learning_rate": 1.741038870569245e-05, "loss": 2.6124, "step": 22820 }, { "epoch": 0.51, "grad_norm": 1.6443031460408375, "learning_rate": 1.7407936120418393e-05, "loss": 2.6135, "step": 22830 }, { "epoch": 0.52, "grad_norm": 1.8557093981576702, "learning_rate": 1.7405482547211513e-05, "loss": 2.7389, "step": 22840 }, { "epoch": 0.52, "grad_norm": 1.6826013539590698, "learning_rate": 1.7403027986399025e-05, "loss": 2.5792, "step": 22850 }, { "epoch": 0.52, "grad_norm": 1.875148626218132, "learning_rate": 1.740057243830828e-05, "loss": 2.8134, "step": 22860 }, { "epoch": 0.52, "grad_norm": 2.159826958911109, "learning_rate": 1.739811590326674e-05, "loss": 2.7669, "step": 22870 }, { "epoch": 0.52, "grad_norm": 1.755732725225071, "learning_rate": 1.739565838160202e-05, "loss": 2.5496, "step": 22880 }, { "epoch": 0.52, "grad_norm": 1.6588125261739965, "learning_rate": 1.7393199873641856e-05, "loss": 2.7356, "step": 22890 }, { "epoch": 0.52, "grad_norm": 2.0670920035233165, "learning_rate": 1.7390740379714122e-05, "loss": 2.7296, "step": 22900 }, { "epoch": 0.52, "grad_norm": 1.7208586863052566, "learning_rate": 1.7388279900146816e-05, "loss": 2.6711, "step": 22910 }, { "epoch": 0.52, "grad_norm": 1.748218923577889, "learning_rate": 1.738581843526807e-05, "loss": 2.5418, "step": 22920 }, { "epoch": 0.52, "grad_norm": 1.849394069615745, "learning_rate": 1.738335598540615e-05, "loss": 2.6783, "step": 22930 }, { "epoch": 0.52, "grad_norm": 2.0099766560763617, "learning_rate": 1.7380892550889456e-05, "loss": 2.7059, "step": 22940 }, { "epoch": 0.52, "grad_norm": 1.654952780726336, "learning_rate": 1.737842813204651e-05, "loss": 2.7723, "step": 22950 }, { "epoch": 0.52, "grad_norm": 1.926542892076294, "learning_rate": 1.7375962729205972e-05, "loss": 2.6415, "step": 22960 }, { "epoch": 0.52, "grad_norm": 1.9448653332480474, "learning_rate": 1.737349634269663e-05, "loss": 2.5753, "step": 22970 }, { "epoch": 0.52, "grad_norm": 1.699230473612858, "learning_rate": 1.7371028972847406e-05, "loss": 2.7466, "step": 22980 }, { "epoch": 0.52, "grad_norm": 1.8338642791363597, "learning_rate": 1.736856061998735e-05, "loss": 2.663, "step": 22990 }, { "epoch": 0.52, "grad_norm": 1.9669955876425642, "learning_rate": 1.736609128444565e-05, "loss": 2.6603, "step": 23000 }, { "epoch": 0.52, "grad_norm": 1.7245059171105204, "learning_rate": 1.7363620966551616e-05, "loss": 2.6937, "step": 23010 }, { "epoch": 0.52, "grad_norm": 1.9087943647391723, "learning_rate": 1.7361149666634695e-05, "loss": 2.715, "step": 23020 }, { "epoch": 0.52, "grad_norm": 1.635282207151815, "learning_rate": 1.735867738502446e-05, "loss": 2.7125, "step": 23030 }, { "epoch": 0.52, "grad_norm": 1.6239897812714714, "learning_rate": 1.7356204122050623e-05, "loss": 2.683, "step": 23040 }, { "epoch": 0.52, "grad_norm": 1.6894745965552624, "learning_rate": 1.7353729878043015e-05, "loss": 2.6814, "step": 23050 }, { "epoch": 0.52, "grad_norm": 1.7464624392660766, "learning_rate": 1.735125465333161e-05, "loss": 2.5371, "step": 23060 }, { "epoch": 0.52, "grad_norm": 1.8319208449904527, "learning_rate": 1.7348778448246507e-05, "loss": 2.7489, "step": 23070 }, { "epoch": 0.52, "grad_norm": 1.6672727205026177, "learning_rate": 1.7346301263117937e-05, "loss": 2.7878, "step": 23080 }, { "epoch": 0.52, "grad_norm": 1.7038739138407533, "learning_rate": 1.7343823098276257e-05, "loss": 2.6465, "step": 23090 }, { "epoch": 0.52, "grad_norm": 1.7853445272962956, "learning_rate": 1.7341343954051962e-05, "loss": 2.5538, "step": 23100 }, { "epoch": 0.52, "grad_norm": 1.7410867785539594, "learning_rate": 1.733886383077567e-05, "loss": 2.786, "step": 23110 }, { "epoch": 0.52, "grad_norm": 2.042980218145061, "learning_rate": 1.7336382728778142e-05, "loss": 2.6221, "step": 23120 }, { "epoch": 0.52, "grad_norm": 1.6858137574511234, "learning_rate": 1.733390064839025e-05, "loss": 2.6672, "step": 23130 }, { "epoch": 0.52, "grad_norm": 1.8221026004377496, "learning_rate": 1.733141758994302e-05, "loss": 2.6976, "step": 23140 }, { "epoch": 0.52, "grad_norm": 1.8095182172705824, "learning_rate": 1.7328933553767582e-05, "loss": 2.7974, "step": 23150 }, { "epoch": 0.52, "grad_norm": 1.6236925948787133, "learning_rate": 1.7326448540195223e-05, "loss": 2.5702, "step": 23160 }, { "epoch": 0.52, "grad_norm": 2.0577221407532096, "learning_rate": 1.7323962549557346e-05, "loss": 2.6976, "step": 23170 }, { "epoch": 0.52, "grad_norm": 1.9125150919833493, "learning_rate": 1.732147558218548e-05, "loss": 2.6638, "step": 23180 }, { "epoch": 0.52, "grad_norm": 1.6953579235803413, "learning_rate": 1.7318987638411293e-05, "loss": 2.6714, "step": 23190 }, { "epoch": 0.52, "grad_norm": 1.820494831531721, "learning_rate": 1.7316498718566585e-05, "loss": 2.7157, "step": 23200 }, { "epoch": 0.52, "grad_norm": 1.9617907241872838, "learning_rate": 1.7314008822983275e-05, "loss": 2.5933, "step": 23210 }, { "epoch": 0.52, "grad_norm": 1.6255830778806082, "learning_rate": 1.7311517951993426e-05, "loss": 2.6718, "step": 23220 }, { "epoch": 0.52, "grad_norm": 1.7300370054182854, "learning_rate": 1.7309026105929216e-05, "loss": 2.7163, "step": 23230 }, { "epoch": 0.52, "grad_norm": 1.9018377274735354, "learning_rate": 1.7306533285122974e-05, "loss": 2.679, "step": 23240 }, { "epoch": 0.52, "grad_norm": 1.8116920244456332, "learning_rate": 1.730403948990713e-05, "loss": 2.7351, "step": 23250 }, { "epoch": 0.52, "grad_norm": 1.8706117681195027, "learning_rate": 1.7301544720614273e-05, "loss": 2.6958, "step": 23260 }, { "epoch": 0.52, "grad_norm": 2.003762294761966, "learning_rate": 1.72990489775771e-05, "loss": 2.7133, "step": 23270 }, { "epoch": 0.52, "grad_norm": 1.9606142298856644, "learning_rate": 1.7296552261128452e-05, "loss": 2.6677, "step": 23280 }, { "epoch": 0.53, "grad_norm": 1.8381423526869458, "learning_rate": 1.7294054571601294e-05, "loss": 2.705, "step": 23290 }, { "epoch": 0.53, "grad_norm": 1.8114157131657713, "learning_rate": 1.729155590932872e-05, "loss": 2.6714, "step": 23300 }, { "epoch": 0.53, "grad_norm": 1.745554247328976, "learning_rate": 1.7289056274643955e-05, "loss": 2.7043, "step": 23310 }, { "epoch": 0.53, "grad_norm": 1.878561358836879, "learning_rate": 1.7286555667880354e-05, "loss": 2.649, "step": 23320 }, { "epoch": 0.53, "grad_norm": 2.098673675586361, "learning_rate": 1.728405408937141e-05, "loss": 2.6385, "step": 23330 }, { "epoch": 0.53, "grad_norm": 1.6318278480190436, "learning_rate": 1.728155153945072e-05, "loss": 2.6265, "step": 23340 }, { "epoch": 0.53, "grad_norm": 1.971936089976185, "learning_rate": 1.727904801845204e-05, "loss": 2.7226, "step": 23350 }, { "epoch": 0.53, "grad_norm": 2.000349713448849, "learning_rate": 1.727654352670924e-05, "loss": 2.7717, "step": 23360 }, { "epoch": 0.53, "grad_norm": 2.2234878510712717, "learning_rate": 1.7274038064556323e-05, "loss": 2.6667, "step": 23370 }, { "epoch": 0.53, "grad_norm": 1.7872255179061463, "learning_rate": 1.727153163232742e-05, "loss": 2.555, "step": 23380 }, { "epoch": 0.53, "grad_norm": 1.5377532009631565, "learning_rate": 1.7269024230356792e-05, "loss": 2.7357, "step": 23390 }, { "epoch": 0.53, "grad_norm": 1.9626648795714479, "learning_rate": 1.7266515858978828e-05, "loss": 2.8283, "step": 23400 }, { "epoch": 0.53, "grad_norm": 1.7387933063050784, "learning_rate": 1.7264006518528057e-05, "loss": 2.6032, "step": 23410 }, { "epoch": 0.53, "grad_norm": 1.6989387336275685, "learning_rate": 1.7261496209339114e-05, "loss": 2.6378, "step": 23420 }, { "epoch": 0.53, "grad_norm": 1.9316390668608774, "learning_rate": 1.725898493174679e-05, "loss": 2.5958, "step": 23430 }, { "epoch": 0.53, "grad_norm": 1.7736303238976372, "learning_rate": 1.7256472686085984e-05, "loss": 2.6091, "step": 23440 }, { "epoch": 0.53, "grad_norm": 2.1089694767976157, "learning_rate": 1.7253959472691743e-05, "loss": 2.7223, "step": 23450 }, { "epoch": 0.53, "grad_norm": 1.685084646156437, "learning_rate": 1.7251445291899218e-05, "loss": 2.7464, "step": 23460 }, { "epoch": 0.53, "grad_norm": 1.5755201833607866, "learning_rate": 1.7248930144043715e-05, "loss": 2.6861, "step": 23470 }, { "epoch": 0.53, "grad_norm": 1.7564400433651053, "learning_rate": 1.7246414029460653e-05, "loss": 2.7483, "step": 23480 }, { "epoch": 0.53, "grad_norm": 1.9218672550948166, "learning_rate": 1.7243896948485585e-05, "loss": 2.5867, "step": 23490 }, { "epoch": 0.53, "grad_norm": 1.8055518590485553, "learning_rate": 1.7241378901454195e-05, "loss": 2.6386, "step": 23500 }, { "epoch": 0.53, "grad_norm": 1.7589802551605682, "learning_rate": 1.723885988870229e-05, "loss": 2.8155, "step": 23510 }, { "epoch": 0.53, "grad_norm": 1.7644750165307008, "learning_rate": 1.723633991056581e-05, "loss": 2.686, "step": 23520 }, { "epoch": 0.53, "grad_norm": 1.5893099582197177, "learning_rate": 1.7233818967380825e-05, "loss": 2.7139, "step": 23530 }, { "epoch": 0.53, "grad_norm": 1.7557998336848706, "learning_rate": 1.723129705948353e-05, "loss": 2.607, "step": 23540 }, { "epoch": 0.53, "grad_norm": 1.7601851964168107, "learning_rate": 1.722877418721025e-05, "loss": 2.7926, "step": 23550 }, { "epoch": 0.53, "grad_norm": 1.6598059083245138, "learning_rate": 1.7226250350897437e-05, "loss": 2.6744, "step": 23560 }, { "epoch": 0.53, "grad_norm": 1.5695384745638419, "learning_rate": 1.7223725550881675e-05, "loss": 2.6601, "step": 23570 }, { "epoch": 0.53, "grad_norm": 1.6137182889215584, "learning_rate": 1.7221199787499678e-05, "loss": 2.7269, "step": 23580 }, { "epoch": 0.53, "grad_norm": 1.9565162443454345, "learning_rate": 1.721867306108828e-05, "loss": 2.6104, "step": 23590 }, { "epoch": 0.53, "grad_norm": 2.1589006436697664, "learning_rate": 1.7216145371984455e-05, "loss": 2.7381, "step": 23600 }, { "epoch": 0.53, "grad_norm": 1.6887952236075374, "learning_rate": 1.721361672052529e-05, "loss": 2.6975, "step": 23610 }, { "epoch": 0.53, "grad_norm": 1.7241281809800402, "learning_rate": 1.7211087107048024e-05, "loss": 2.6145, "step": 23620 }, { "epoch": 0.53, "grad_norm": 1.6715477935159442, "learning_rate": 1.7208556531889995e-05, "loss": 2.8536, "step": 23630 }, { "epoch": 0.53, "grad_norm": 1.8184474438513947, "learning_rate": 1.720602499538869e-05, "loss": 2.5774, "step": 23640 }, { "epoch": 0.53, "grad_norm": 2.2092620481795517, "learning_rate": 1.720349249788172e-05, "loss": 2.6462, "step": 23650 }, { "epoch": 0.53, "grad_norm": 2.421772381270146, "learning_rate": 1.720095903970682e-05, "loss": 2.8714, "step": 23660 }, { "epoch": 0.53, "grad_norm": 1.6505177800166706, "learning_rate": 1.7198424621201855e-05, "loss": 2.756, "step": 23670 }, { "epoch": 0.53, "grad_norm": 2.0455937111885394, "learning_rate": 1.719588924270482e-05, "loss": 2.7337, "step": 23680 }, { "epoch": 0.53, "grad_norm": 1.7370265796722701, "learning_rate": 1.719335290455384e-05, "loss": 2.6297, "step": 23690 }, { "epoch": 0.53, "grad_norm": 1.7954802280382407, "learning_rate": 1.7190815607087157e-05, "loss": 2.6803, "step": 23700 }, { "epoch": 0.53, "grad_norm": 1.5284085679624695, "learning_rate": 1.7188277350643156e-05, "loss": 2.687, "step": 23710 }, { "epoch": 0.53, "grad_norm": 1.6795283833905204, "learning_rate": 1.7185738135560336e-05, "loss": 2.6626, "step": 23720 }, { "epoch": 0.54, "grad_norm": 1.630758557788928, "learning_rate": 1.7183197962177336e-05, "loss": 2.621, "step": 23730 }, { "epoch": 0.54, "grad_norm": 1.8392997383187673, "learning_rate": 1.7180656830832913e-05, "loss": 2.6459, "step": 23740 }, { "epoch": 0.54, "grad_norm": 2.223745897121185, "learning_rate": 1.717811474186596e-05, "loss": 2.6181, "step": 23750 }, { "epoch": 0.54, "grad_norm": 1.752185264729432, "learning_rate": 1.7175571695615487e-05, "loss": 2.7614, "step": 23760 }, { "epoch": 0.54, "grad_norm": 2.01502031572315, "learning_rate": 1.7173027692420646e-05, "loss": 2.7229, "step": 23770 }, { "epoch": 0.54, "grad_norm": 1.6354967946291774, "learning_rate": 1.7170482732620706e-05, "loss": 2.6712, "step": 23780 }, { "epoch": 0.54, "grad_norm": 1.6293997267027014, "learning_rate": 1.7167936816555063e-05, "loss": 2.5644, "step": 23790 }, { "epoch": 0.54, "grad_norm": 1.6466962268945424, "learning_rate": 1.7165389944563248e-05, "loss": 2.7109, "step": 23800 }, { "epoch": 0.54, "grad_norm": 1.5776009584284507, "learning_rate": 1.7162842116984912e-05, "loss": 2.7679, "step": 23810 }, { "epoch": 0.54, "grad_norm": 1.765792598435173, "learning_rate": 1.716029333415984e-05, "loss": 2.7134, "step": 23820 }, { "epoch": 0.54, "grad_norm": 1.7665595970573393, "learning_rate": 1.715774359642794e-05, "loss": 2.7486, "step": 23830 }, { "epoch": 0.54, "grad_norm": 1.5021561943453643, "learning_rate": 1.715519290412925e-05, "loss": 2.6447, "step": 23840 }, { "epoch": 0.54, "grad_norm": 1.9448656666901778, "learning_rate": 1.7152641257603934e-05, "loss": 2.7136, "step": 23850 }, { "epoch": 0.54, "grad_norm": 1.7632489054535454, "learning_rate": 1.715008865719228e-05, "loss": 2.7153, "step": 23860 }, { "epoch": 0.54, "grad_norm": 1.9283097185209004, "learning_rate": 1.7147535103234706e-05, "loss": 2.6381, "step": 23870 }, { "epoch": 0.54, "grad_norm": 1.8470500899660995, "learning_rate": 1.7144980596071764e-05, "loss": 2.621, "step": 23880 }, { "epoch": 0.54, "grad_norm": 1.6964225507604345, "learning_rate": 1.7142425136044122e-05, "loss": 2.7752, "step": 23890 }, { "epoch": 0.54, "grad_norm": 1.7092600550837045, "learning_rate": 1.713986872349258e-05, "loss": 2.7611, "step": 23900 }, { "epoch": 0.54, "grad_norm": 1.4836942654650842, "learning_rate": 1.7137311358758063e-05, "loss": 2.6681, "step": 23910 }, { "epoch": 0.54, "grad_norm": 1.7262997822995876, "learning_rate": 1.7134753042181626e-05, "loss": 2.6854, "step": 23920 }, { "epoch": 0.54, "grad_norm": 1.8981072120475828, "learning_rate": 1.7132193774104453e-05, "loss": 2.6959, "step": 23930 }, { "epoch": 0.54, "grad_norm": 1.6887630935579172, "learning_rate": 1.712963355486785e-05, "loss": 2.778, "step": 23940 }, { "epoch": 0.54, "grad_norm": 1.7525680162214863, "learning_rate": 1.712707238481325e-05, "loss": 2.627, "step": 23950 }, { "epoch": 0.54, "grad_norm": 1.7343946498893508, "learning_rate": 1.7124510264282214e-05, "loss": 2.6259, "step": 23960 }, { "epoch": 0.54, "grad_norm": 1.6963265741844247, "learning_rate": 1.7121947193616432e-05, "loss": 2.7518, "step": 23970 }, { "epoch": 0.54, "grad_norm": 1.6008465002738088, "learning_rate": 1.7119383173157716e-05, "loss": 2.8132, "step": 23980 }, { "epoch": 0.54, "grad_norm": 1.9253687858122133, "learning_rate": 1.711681820324801e-05, "loss": 2.6637, "step": 23990 }, { "epoch": 0.54, "grad_norm": 1.8893871401105027, "learning_rate": 1.711425228422938e-05, "loss": 2.6543, "step": 24000 }, { "epoch": 0.54, "grad_norm": 1.8481024974072209, "learning_rate": 1.7111685416444026e-05, "loss": 2.6499, "step": 24010 }, { "epoch": 0.54, "grad_norm": 1.8168686955583517, "learning_rate": 1.710911760023426e-05, "loss": 2.6774, "step": 24020 }, { "epoch": 0.54, "grad_norm": 1.6836272432801742, "learning_rate": 1.710654883594254e-05, "loss": 2.6615, "step": 24030 }, { "epoch": 0.54, "grad_norm": 2.2067455327806065, "learning_rate": 1.7103979123911427e-05, "loss": 2.6473, "step": 24040 }, { "epoch": 0.54, "grad_norm": 1.5906077623598474, "learning_rate": 1.7101408464483633e-05, "loss": 2.699, "step": 24050 }, { "epoch": 0.54, "grad_norm": 2.417927568717569, "learning_rate": 1.7098836858001978e-05, "loss": 2.6411, "step": 24060 }, { "epoch": 0.54, "grad_norm": 1.812992989658272, "learning_rate": 1.7096264304809416e-05, "loss": 2.6266, "step": 24070 }, { "epoch": 0.54, "grad_norm": 2.3919310447291715, "learning_rate": 1.7093690805249034e-05, "loss": 2.6997, "step": 24080 }, { "epoch": 0.54, "grad_norm": 2.088697720102489, "learning_rate": 1.7091116359664027e-05, "loss": 2.7746, "step": 24090 }, { "epoch": 0.54, "grad_norm": 1.7958322355685084, "learning_rate": 1.7088540968397732e-05, "loss": 2.5843, "step": 24100 }, { "epoch": 0.54, "grad_norm": 1.6126804649985549, "learning_rate": 1.708596463179361e-05, "loss": 2.6574, "step": 24110 }, { "epoch": 0.54, "grad_norm": 1.6435356146577174, "learning_rate": 1.7083387350195236e-05, "loss": 2.7012, "step": 24120 }, { "epoch": 0.54, "grad_norm": 1.6785919759023467, "learning_rate": 1.7080809123946327e-05, "loss": 2.6595, "step": 24130 }, { "epoch": 0.54, "grad_norm": 1.5753515680407244, "learning_rate": 1.7078229953390715e-05, "loss": 2.7, "step": 24140 }, { "epoch": 0.54, "grad_norm": 1.5696666449294887, "learning_rate": 1.7075649838872365e-05, "loss": 2.6893, "step": 24150 }, { "epoch": 0.54, "grad_norm": 1.8204765406749162, "learning_rate": 1.7073068780735363e-05, "loss": 2.6754, "step": 24160 }, { "epoch": 0.55, "grad_norm": 1.8246504856520998, "learning_rate": 1.7070486779323926e-05, "loss": 2.5805, "step": 24170 }, { "epoch": 0.55, "grad_norm": 1.8527858678625002, "learning_rate": 1.7067903834982384e-05, "loss": 2.622, "step": 24180 }, { "epoch": 0.55, "grad_norm": 1.6658158673942316, "learning_rate": 1.7065319948055212e-05, "loss": 2.516, "step": 24190 }, { "epoch": 0.55, "grad_norm": 1.8291598333842718, "learning_rate": 1.7062735118886994e-05, "loss": 2.6767, "step": 24200 }, { "epoch": 0.55, "grad_norm": 2.2069320128210754, "learning_rate": 1.7060149347822452e-05, "loss": 2.7926, "step": 24210 }, { "epoch": 0.55, "grad_norm": 2.070611019760841, "learning_rate": 1.7057562635206426e-05, "loss": 2.8142, "step": 24220 }, { "epoch": 0.55, "grad_norm": 1.5292033679940367, "learning_rate": 1.7054974981383883e-05, "loss": 2.8268, "step": 24230 }, { "epoch": 0.55, "grad_norm": 1.7701809897864111, "learning_rate": 1.7052386386699913e-05, "loss": 2.5838, "step": 24240 }, { "epoch": 0.55, "grad_norm": 2.142836424459652, "learning_rate": 1.7049796851499745e-05, "loss": 2.7113, "step": 24250 }, { "epoch": 0.55, "grad_norm": 1.8028206797889939, "learning_rate": 1.7047206376128712e-05, "loss": 2.7708, "step": 24260 }, { "epoch": 0.55, "grad_norm": 1.7520412684111564, "learning_rate": 1.7044614960932286e-05, "loss": 2.6658, "step": 24270 }, { "epoch": 0.55, "grad_norm": 1.6665462475737283, "learning_rate": 1.7042022606256067e-05, "loss": 2.6326, "step": 24280 }, { "epoch": 0.55, "grad_norm": 1.6079266366561717, "learning_rate": 1.703942931244577e-05, "loss": 2.6192, "step": 24290 }, { "epoch": 0.55, "grad_norm": 1.7402477845041175, "learning_rate": 1.703683507984724e-05, "loss": 2.6967, "step": 24300 }, { "epoch": 0.55, "grad_norm": 1.6220990759720502, "learning_rate": 1.703423990880645e-05, "loss": 2.6329, "step": 24310 }, { "epoch": 0.55, "grad_norm": 2.1028366203375835, "learning_rate": 1.7031643799669493e-05, "loss": 2.7069, "step": 24320 }, { "epoch": 0.55, "grad_norm": 1.7268514859325925, "learning_rate": 1.7029046752782593e-05, "loss": 2.5602, "step": 24330 }, { "epoch": 0.55, "grad_norm": 1.4812337435731233, "learning_rate": 1.70264487684921e-05, "loss": 2.6085, "step": 24340 }, { "epoch": 0.55, "grad_norm": 1.5256757643329573, "learning_rate": 1.7023849847144473e-05, "loss": 2.5978, "step": 24350 }, { "epoch": 0.55, "grad_norm": 1.716829701376031, "learning_rate": 1.702124998908632e-05, "loss": 2.6466, "step": 24360 }, { "epoch": 0.55, "grad_norm": 1.7894292017695932, "learning_rate": 1.701864919466435e-05, "loss": 2.5727, "step": 24370 }, { "epoch": 0.55, "grad_norm": 1.9828656492227887, "learning_rate": 1.7016047464225415e-05, "loss": 2.6716, "step": 24380 }, { "epoch": 0.55, "grad_norm": 1.8676352408646706, "learning_rate": 1.701344479811649e-05, "loss": 2.5216, "step": 24390 }, { "epoch": 0.55, "grad_norm": 1.6955670086266224, "learning_rate": 1.701084119668466e-05, "loss": 2.7593, "step": 24400 }, { "epoch": 0.55, "grad_norm": 1.6571809857908613, "learning_rate": 1.7008236660277154e-05, "loss": 2.7102, "step": 24410 }, { "epoch": 0.55, "grad_norm": 1.6059070087648264, "learning_rate": 1.7005631189241313e-05, "loss": 2.7104, "step": 24420 }, { "epoch": 0.55, "grad_norm": 1.7066827985919075, "learning_rate": 1.7003024783924604e-05, "loss": 2.7569, "step": 24430 }, { "epoch": 0.55, "grad_norm": 1.6453784477135205, "learning_rate": 1.7000417444674622e-05, "loss": 2.5794, "step": 24440 }, { "epoch": 0.55, "grad_norm": 1.80408091905499, "learning_rate": 1.699780917183909e-05, "loss": 2.6452, "step": 24450 }, { "epoch": 0.55, "grad_norm": 1.6054416219883152, "learning_rate": 1.6995199965765845e-05, "loss": 2.6652, "step": 24460 }, { "epoch": 0.55, "grad_norm": 1.8511971955490827, "learning_rate": 1.6992589826802857e-05, "loss": 2.6635, "step": 24470 }, { "epoch": 0.55, "grad_norm": 1.6465628435501094, "learning_rate": 1.698997875529822e-05, "loss": 2.6978, "step": 24480 }, { "epoch": 0.55, "grad_norm": 2.330184667005359, "learning_rate": 1.6987366751600144e-05, "loss": 2.6779, "step": 24490 }, { "epoch": 0.55, "grad_norm": 1.9517118417454744, "learning_rate": 1.6984753816056978e-05, "loss": 2.7488, "step": 24500 }, { "epoch": 0.55, "grad_norm": 1.5912646234063847, "learning_rate": 1.698213994901718e-05, "loss": 2.6995, "step": 24510 }, { "epoch": 0.55, "grad_norm": 1.92316660261019, "learning_rate": 1.697952515082934e-05, "loss": 2.7418, "step": 24520 }, { "epoch": 0.55, "grad_norm": 1.795339191674215, "learning_rate": 1.6976909421842172e-05, "loss": 2.5763, "step": 24530 }, { "epoch": 0.55, "grad_norm": 1.6703911800119293, "learning_rate": 1.6974292762404516e-05, "loss": 2.6525, "step": 24540 }, { "epoch": 0.55, "grad_norm": 2.1324899227938583, "learning_rate": 1.697167517286533e-05, "loss": 2.6964, "step": 24550 }, { "epoch": 0.55, "grad_norm": 2.5455395017759512, "learning_rate": 1.69690566535737e-05, "loss": 2.5776, "step": 24560 }, { "epoch": 0.55, "grad_norm": 2.3211438854945947, "learning_rate": 1.6966437204878837e-05, "loss": 2.6639, "step": 24570 }, { "epoch": 0.55, "grad_norm": 1.8192957893868973, "learning_rate": 1.6963816827130074e-05, "loss": 2.6454, "step": 24580 }, { "epoch": 0.55, "grad_norm": 1.968317873779225, "learning_rate": 1.6961195520676864e-05, "loss": 2.6792, "step": 24590 }, { "epoch": 0.55, "grad_norm": 1.5291317719974626, "learning_rate": 1.6958573285868795e-05, "loss": 2.7354, "step": 24600 }, { "epoch": 0.55, "grad_norm": 1.7264833872312548, "learning_rate": 1.6955950123055566e-05, "loss": 2.6509, "step": 24610 }, { "epoch": 0.56, "grad_norm": 2.048096711050953, "learning_rate": 1.695332603258701e-05, "loss": 2.7313, "step": 24620 }, { "epoch": 0.56, "grad_norm": 1.6994061132384353, "learning_rate": 1.6950701014813078e-05, "loss": 2.6669, "step": 24630 }, { "epoch": 0.56, "grad_norm": 1.729360095122331, "learning_rate": 1.6948075070083844e-05, "loss": 2.665, "step": 24640 }, { "epoch": 0.56, "grad_norm": 1.7959642203670934, "learning_rate": 1.694544819874951e-05, "loss": 2.7243, "step": 24650 }, { "epoch": 0.56, "grad_norm": 1.8253570816151938, "learning_rate": 1.6942820401160402e-05, "loss": 2.7018, "step": 24660 }, { "epoch": 0.56, "grad_norm": 2.242812112229444, "learning_rate": 1.694019167766696e-05, "loss": 2.7349, "step": 24670 }, { "epoch": 0.56, "grad_norm": 1.634194712593525, "learning_rate": 1.693756202861976e-05, "loss": 2.6297, "step": 24680 }, { "epoch": 0.56, "grad_norm": 1.7064997610241321, "learning_rate": 1.6934931454369494e-05, "loss": 2.6758, "step": 24690 }, { "epoch": 0.56, "grad_norm": 1.7294548322365946, "learning_rate": 1.6932299955266978e-05, "loss": 2.54, "step": 24700 }, { "epoch": 0.56, "grad_norm": 1.9805004772610066, "learning_rate": 1.692966753166315e-05, "loss": 2.597, "step": 24710 }, { "epoch": 0.56, "grad_norm": 1.8515564212731759, "learning_rate": 1.692703418390908e-05, "loss": 2.6945, "step": 24720 }, { "epoch": 0.56, "grad_norm": 1.6722217410224116, "learning_rate": 1.6924399912355953e-05, "loss": 2.6236, "step": 24730 }, { "epoch": 0.56, "grad_norm": 3.0800367294101663, "learning_rate": 1.6921764717355072e-05, "loss": 2.7337, "step": 24740 }, { "epoch": 0.56, "grad_norm": 2.4387659833984947, "learning_rate": 1.6919128599257882e-05, "loss": 2.6694, "step": 24750 }, { "epoch": 0.56, "grad_norm": 1.7949060721514327, "learning_rate": 1.6916491558415932e-05, "loss": 2.7956, "step": 24760 }, { "epoch": 0.56, "grad_norm": 1.8929528865018244, "learning_rate": 1.6913853595180906e-05, "loss": 2.6609, "step": 24770 }, { "epoch": 0.56, "grad_norm": 1.835302284694053, "learning_rate": 1.69112147099046e-05, "loss": 2.5786, "step": 24780 }, { "epoch": 0.56, "grad_norm": 1.6641681990561197, "learning_rate": 1.6908574902938943e-05, "loss": 2.6093, "step": 24790 }, { "epoch": 0.56, "grad_norm": 1.7116291373051344, "learning_rate": 1.6905934174635984e-05, "loss": 2.6945, "step": 24800 }, { "epoch": 0.56, "grad_norm": 1.7755998669484665, "learning_rate": 1.6903292525347895e-05, "loss": 2.6887, "step": 24810 }, { "epoch": 0.56, "grad_norm": 1.8567684859741904, "learning_rate": 1.690064995542697e-05, "loss": 2.5915, "step": 24820 }, { "epoch": 0.56, "grad_norm": 1.4831592632238773, "learning_rate": 1.689800646522562e-05, "loss": 2.6459, "step": 24830 }, { "epoch": 0.56, "grad_norm": 1.7281304477839263, "learning_rate": 1.689536205509639e-05, "loss": 2.6842, "step": 24840 }, { "epoch": 0.56, "grad_norm": 2.254394765215246, "learning_rate": 1.6892716725391947e-05, "loss": 2.6325, "step": 24850 }, { "epoch": 0.56, "grad_norm": 2.060237863542605, "learning_rate": 1.6890070476465066e-05, "loss": 2.8339, "step": 24860 }, { "epoch": 0.56, "grad_norm": 1.905188579143855, "learning_rate": 1.688742330866866e-05, "loss": 2.6992, "step": 24870 }, { "epoch": 0.56, "grad_norm": 1.6482560549811465, "learning_rate": 1.688477522235576e-05, "loss": 2.7389, "step": 24880 }, { "epoch": 0.56, "grad_norm": 1.7282545229390023, "learning_rate": 1.6882126217879517e-05, "loss": 2.6872, "step": 24890 }, { "epoch": 0.56, "grad_norm": 1.5893381608524861, "learning_rate": 1.6879476295593202e-05, "loss": 2.6077, "step": 24900 }, { "epoch": 0.56, "grad_norm": 1.6959402788386126, "learning_rate": 1.6876825455850217e-05, "loss": 2.6208, "step": 24910 }, { "epoch": 0.56, "grad_norm": 1.8875282195451295, "learning_rate": 1.687417369900408e-05, "loss": 2.7944, "step": 24920 }, { "epoch": 0.56, "grad_norm": 1.753085397766255, "learning_rate": 1.6871521025408437e-05, "loss": 2.6784, "step": 24930 }, { "epoch": 0.56, "grad_norm": 1.9206505271737404, "learning_rate": 1.6868867435417046e-05, "loss": 2.6952, "step": 24940 }, { "epoch": 0.56, "grad_norm": 2.156446139855591, "learning_rate": 1.6866212929383795e-05, "loss": 2.7989, "step": 24950 }, { "epoch": 0.56, "grad_norm": 1.6254559412345007, "learning_rate": 1.6863557507662696e-05, "loss": 2.6205, "step": 24960 }, { "epoch": 0.56, "grad_norm": 2.2375595327155318, "learning_rate": 1.686090117060788e-05, "loss": 2.7598, "step": 24970 }, { "epoch": 0.56, "grad_norm": 1.881830492378842, "learning_rate": 1.68582439185736e-05, "loss": 2.6656, "step": 24980 }, { "epoch": 0.56, "grad_norm": 1.960930873008514, "learning_rate": 1.685558575191422e-05, "loss": 2.7979, "step": 24990 }, { "epoch": 0.56, "grad_norm": 1.6518323569570124, "learning_rate": 1.685292667098425e-05, "loss": 2.6371, "step": 25000 }, { "epoch": 0.56, "grad_norm": 1.6963557749797311, "learning_rate": 1.6850266676138308e-05, "loss": 2.646, "step": 25010 }, { "epoch": 0.56, "grad_norm": 1.6138358419578263, "learning_rate": 1.684760576773113e-05, "loss": 2.6425, "step": 25020 }, { "epoch": 0.56, "grad_norm": 1.5933409740642543, "learning_rate": 1.684494394611757e-05, "loss": 2.5873, "step": 25030 }, { "epoch": 0.56, "grad_norm": 1.7079795152101078, "learning_rate": 1.6842281211652633e-05, "loss": 2.7399, "step": 25040 }, { "epoch": 0.56, "grad_norm": 1.6478930973574677, "learning_rate": 1.6839617564691408e-05, "loss": 2.7094, "step": 25050 }, { "epoch": 0.57, "grad_norm": 2.1170046158924363, "learning_rate": 1.6836953005589133e-05, "loss": 2.6979, "step": 25060 }, { "epoch": 0.57, "grad_norm": 2.0333942528912297, "learning_rate": 1.683428753470115e-05, "loss": 2.6209, "step": 25070 }, { "epoch": 0.57, "grad_norm": 1.8440140458069607, "learning_rate": 1.6831621152382932e-05, "loss": 2.5796, "step": 25080 }, { "epoch": 0.57, "grad_norm": 1.8965274120326088, "learning_rate": 1.6828953858990073e-05, "loss": 2.7112, "step": 25090 }, { "epoch": 0.57, "grad_norm": 2.033773502799209, "learning_rate": 1.682628565487829e-05, "loss": 2.7311, "step": 25100 }, { "epoch": 0.57, "grad_norm": 2.219310911319307, "learning_rate": 1.682361654040341e-05, "loss": 2.7193, "step": 25110 }, { "epoch": 0.57, "grad_norm": 1.722787443149689, "learning_rate": 1.6820946515921398e-05, "loss": 2.6948, "step": 25120 }, { "epoch": 0.57, "grad_norm": 1.9305905804625703, "learning_rate": 1.6818275581788323e-05, "loss": 2.7271, "step": 25130 }, { "epoch": 0.57, "grad_norm": 1.9024304752876613, "learning_rate": 1.68156037383604e-05, "loss": 2.7264, "step": 25140 }, { "epoch": 0.57, "grad_norm": 1.5378984116034877, "learning_rate": 1.6812930985993935e-05, "loss": 2.6945, "step": 25150 }, { "epoch": 0.57, "grad_norm": 1.5199462660785337, "learning_rate": 1.6810257325045373e-05, "loss": 2.6917, "step": 25160 }, { "epoch": 0.57, "grad_norm": 1.9696168770409617, "learning_rate": 1.6807582755871283e-05, "loss": 2.7607, "step": 25170 }, { "epoch": 0.57, "grad_norm": 1.9557903071172236, "learning_rate": 1.6804907278828343e-05, "loss": 2.6968, "step": 25180 }, { "epoch": 0.57, "grad_norm": 1.8166716537837875, "learning_rate": 1.6802230894273362e-05, "loss": 2.73, "step": 25190 }, { "epoch": 0.57, "grad_norm": 1.6764710737942743, "learning_rate": 1.679955360256327e-05, "loss": 2.6531, "step": 25200 }, { "epoch": 0.57, "grad_norm": 1.6817643999940184, "learning_rate": 1.6796875404055106e-05, "loss": 2.556, "step": 25210 }, { "epoch": 0.57, "grad_norm": 2.0202598579228583, "learning_rate": 1.6794196299106045e-05, "loss": 2.699, "step": 25220 }, { "epoch": 0.57, "grad_norm": 1.8707444853598643, "learning_rate": 1.679151628807337e-05, "loss": 2.751, "step": 25230 }, { "epoch": 0.57, "grad_norm": 1.6743304067349862, "learning_rate": 1.6788835371314497e-05, "loss": 2.552, "step": 25240 }, { "epoch": 0.57, "grad_norm": 1.5944910730564028, "learning_rate": 1.6786153549186957e-05, "loss": 2.6816, "step": 25250 }, { "epoch": 0.57, "grad_norm": 1.6404914417294314, "learning_rate": 1.6783470822048397e-05, "loss": 2.6029, "step": 25260 }, { "epoch": 0.57, "grad_norm": 1.523047228024887, "learning_rate": 1.6780787190256588e-05, "loss": 2.812, "step": 25270 }, { "epoch": 0.57, "grad_norm": 1.537493039303056, "learning_rate": 1.677810265416943e-05, "loss": 2.6731, "step": 25280 }, { "epoch": 0.57, "grad_norm": 1.7653299612337043, "learning_rate": 1.6775417214144934e-05, "loss": 2.8003, "step": 25290 }, { "epoch": 0.57, "grad_norm": 1.6324628006654, "learning_rate": 1.677273087054123e-05, "loss": 2.6155, "step": 25300 }, { "epoch": 0.57, "grad_norm": 1.7117872533163982, "learning_rate": 1.6770043623716578e-05, "loss": 2.6414, "step": 25310 }, { "epoch": 0.57, "grad_norm": 1.5778962248747042, "learning_rate": 1.6767355474029346e-05, "loss": 2.7455, "step": 25320 }, { "epoch": 0.57, "grad_norm": 1.6637577674531938, "learning_rate": 1.676466642183804e-05, "loss": 2.6971, "step": 25330 }, { "epoch": 0.57, "grad_norm": 1.784440164855448, "learning_rate": 1.6761976467501266e-05, "loss": 2.6039, "step": 25340 }, { "epoch": 0.57, "grad_norm": 1.8703783125889382, "learning_rate": 1.6759285611377763e-05, "loss": 2.5484, "step": 25350 }, { "epoch": 0.57, "grad_norm": 1.6805995249773285, "learning_rate": 1.6756593853826394e-05, "loss": 2.5736, "step": 25360 }, { "epoch": 0.57, "grad_norm": 1.5298272607925387, "learning_rate": 1.6753901195206124e-05, "loss": 2.8302, "step": 25370 }, { "epoch": 0.57, "grad_norm": 1.9074264533132148, "learning_rate": 1.6751207635876055e-05, "loss": 2.6955, "step": 25380 }, { "epoch": 0.57, "grad_norm": 1.6055324989499433, "learning_rate": 1.6748513176195406e-05, "loss": 2.6372, "step": 25390 }, { "epoch": 0.57, "grad_norm": 2.0451054891162332, "learning_rate": 1.6745817816523513e-05, "loss": 2.5098, "step": 25400 }, { "epoch": 0.57, "grad_norm": 2.002496683786847, "learning_rate": 1.6743121557219835e-05, "loss": 2.5848, "step": 25410 }, { "epoch": 0.57, "grad_norm": 1.8070552446987864, "learning_rate": 1.6740424398643943e-05, "loss": 2.8049, "step": 25420 }, { "epoch": 0.57, "grad_norm": 1.9478274146180403, "learning_rate": 1.673772634115554e-05, "loss": 2.6966, "step": 25430 }, { "epoch": 0.57, "grad_norm": 2.5000365456255684, "learning_rate": 1.6735027385114438e-05, "loss": 2.637, "step": 25440 }, { "epoch": 0.57, "grad_norm": 1.5696608982886588, "learning_rate": 1.6732327530880576e-05, "loss": 2.714, "step": 25450 }, { "epoch": 0.57, "grad_norm": 1.6936549377988246, "learning_rate": 1.6729626778814006e-05, "loss": 2.745, "step": 25460 }, { "epoch": 0.57, "grad_norm": 1.6449291994521509, "learning_rate": 1.6726925129274913e-05, "loss": 2.722, "step": 25470 }, { "epoch": 0.57, "grad_norm": 1.8265976429794677, "learning_rate": 1.6724222582623588e-05, "loss": 2.761, "step": 25480 }, { "epoch": 0.57, "grad_norm": 1.7095341542661506, "learning_rate": 1.6721519139220444e-05, "loss": 2.76, "step": 25490 }, { "epoch": 0.58, "grad_norm": 2.0522801178047985, "learning_rate": 1.6718814799426018e-05, "loss": 2.5865, "step": 25500 }, { "epoch": 0.58, "grad_norm": 1.8619729538968441, "learning_rate": 1.6716109563600963e-05, "loss": 2.6462, "step": 25510 }, { "epoch": 0.58, "grad_norm": 1.7930187021738095, "learning_rate": 1.6713403432106062e-05, "loss": 2.6481, "step": 25520 }, { "epoch": 0.58, "grad_norm": 1.919440712576137, "learning_rate": 1.6710696405302197e-05, "loss": 2.7321, "step": 25530 }, { "epoch": 0.58, "grad_norm": 1.5383180412686532, "learning_rate": 1.6707988483550385e-05, "loss": 2.6343, "step": 25540 }, { "epoch": 0.58, "grad_norm": 1.8291800714948345, "learning_rate": 1.670527966721176e-05, "loss": 2.7419, "step": 25550 }, { "epoch": 0.58, "grad_norm": 1.6936245766762175, "learning_rate": 1.670256995664757e-05, "loss": 2.7338, "step": 25560 }, { "epoch": 0.58, "grad_norm": 2.0838693412630915, "learning_rate": 1.6699859352219187e-05, "loss": 2.5395, "step": 25570 }, { "epoch": 0.58, "grad_norm": 1.6011090951117843, "learning_rate": 1.6697147854288108e-05, "loss": 2.6625, "step": 25580 }, { "epoch": 0.58, "grad_norm": 1.5608648710790543, "learning_rate": 1.6694435463215932e-05, "loss": 2.6636, "step": 25590 }, { "epoch": 0.58, "grad_norm": 1.660050736526065, "learning_rate": 1.6691722179364392e-05, "loss": 2.7028, "step": 25600 }, { "epoch": 0.58, "grad_norm": 1.60182372399414, "learning_rate": 1.6689008003095334e-05, "loss": 2.6763, "step": 25610 }, { "epoch": 0.58, "grad_norm": 1.7891939087742017, "learning_rate": 1.668629293477073e-05, "loss": 2.5181, "step": 25620 }, { "epoch": 0.58, "grad_norm": 1.5530210238491953, "learning_rate": 1.6683576974752656e-05, "loss": 2.6243, "step": 25630 }, { "epoch": 0.58, "grad_norm": 1.7585378853727114, "learning_rate": 1.668086012340332e-05, "loss": 2.7096, "step": 25640 }, { "epoch": 0.58, "grad_norm": 1.6245294204601126, "learning_rate": 1.6678142381085046e-05, "loss": 2.6621, "step": 25650 }, { "epoch": 0.58, "grad_norm": 1.7759070057947288, "learning_rate": 1.667542374816028e-05, "loss": 2.7302, "step": 25660 }, { "epoch": 0.58, "grad_norm": 1.6107657205724575, "learning_rate": 1.6672704224991576e-05, "loss": 2.631, "step": 25670 }, { "epoch": 0.58, "grad_norm": 1.6883569280366002, "learning_rate": 1.6669983811941617e-05, "loss": 2.6009, "step": 25680 }, { "epoch": 0.58, "grad_norm": 1.600959727301415, "learning_rate": 1.6667262509373202e-05, "loss": 2.6681, "step": 25690 }, { "epoch": 0.58, "grad_norm": 1.9596311999243914, "learning_rate": 1.6664540317649246e-05, "loss": 2.5647, "step": 25700 }, { "epoch": 0.58, "grad_norm": 2.6498759241634198, "learning_rate": 1.666181723713278e-05, "loss": 2.7294, "step": 25710 }, { "epoch": 0.58, "grad_norm": 1.6692889202654886, "learning_rate": 1.6659093268186964e-05, "loss": 2.6501, "step": 25720 }, { "epoch": 0.58, "grad_norm": 1.833351540258264, "learning_rate": 1.665636841117507e-05, "loss": 2.656, "step": 25730 }, { "epoch": 0.58, "grad_norm": 1.5715191332967564, "learning_rate": 1.665364266646049e-05, "loss": 2.5134, "step": 25740 }, { "epoch": 0.58, "grad_norm": 1.7558461631711366, "learning_rate": 1.6650916034406726e-05, "loss": 2.7204, "step": 25750 }, { "epoch": 0.58, "grad_norm": 1.7009714479657807, "learning_rate": 1.664818851537741e-05, "loss": 2.7072, "step": 25760 }, { "epoch": 0.58, "grad_norm": 1.983362844817508, "learning_rate": 1.664546010973629e-05, "loss": 2.6912, "step": 25770 }, { "epoch": 0.58, "grad_norm": 1.6426118943112897, "learning_rate": 1.6642730817847225e-05, "loss": 2.6019, "step": 25780 }, { "epoch": 0.58, "grad_norm": 1.846286404042226, "learning_rate": 1.6640000640074206e-05, "loss": 2.7432, "step": 25790 }, { "epoch": 0.58, "grad_norm": 1.9647950106945997, "learning_rate": 1.663726957678132e-05, "loss": 2.6152, "step": 25800 }, { "epoch": 0.58, "grad_norm": 1.818333600086506, "learning_rate": 1.6634537628332797e-05, "loss": 2.6331, "step": 25810 }, { "epoch": 0.58, "grad_norm": 1.9158990450839408, "learning_rate": 1.6631804795092968e-05, "loss": 2.5474, "step": 25820 }, { "epoch": 0.58, "grad_norm": 1.7934193558462102, "learning_rate": 1.662907107742629e-05, "loss": 2.6071, "step": 25830 }, { "epoch": 0.58, "grad_norm": 1.8536518624141451, "learning_rate": 1.6626336475697334e-05, "loss": 2.7406, "step": 25840 }, { "epoch": 0.58, "grad_norm": 1.503462651250082, "learning_rate": 1.662360099027079e-05, "loss": 2.7708, "step": 25850 }, { "epoch": 0.58, "grad_norm": 1.8065426102011244, "learning_rate": 1.662086462151147e-05, "loss": 2.6594, "step": 25860 }, { "epoch": 0.58, "grad_norm": 2.0927380078155715, "learning_rate": 1.6618127369784295e-05, "loss": 2.6994, "step": 25870 }, { "epoch": 0.58, "grad_norm": 1.9245458725262776, "learning_rate": 1.6615389235454305e-05, "loss": 2.6298, "step": 25880 }, { "epoch": 0.58, "grad_norm": 2.062709202550023, "learning_rate": 1.6612650218886674e-05, "loss": 2.6623, "step": 25890 }, { "epoch": 0.58, "grad_norm": 1.808637351204118, "learning_rate": 1.6609910320446675e-05, "loss": 2.6097, "step": 25900 }, { "epoch": 0.58, "grad_norm": 1.5344562433683047, "learning_rate": 1.66071695404997e-05, "loss": 2.781, "step": 25910 }, { "epoch": 0.58, "grad_norm": 1.7482565206794984, "learning_rate": 1.660442787941127e-05, "loss": 2.6212, "step": 25920 }, { "epoch": 0.58, "grad_norm": 1.5586660252193723, "learning_rate": 1.660168533754701e-05, "loss": 2.6246, "step": 25930 }, { "epoch": 0.58, "grad_norm": 1.9134482187798545, "learning_rate": 1.6598941915272676e-05, "loss": 2.6163, "step": 25940 }, { "epoch": 0.59, "grad_norm": 1.8101468840948742, "learning_rate": 1.6596197612954132e-05, "loss": 2.6228, "step": 25950 }, { "epoch": 0.59, "grad_norm": 1.5220208194728748, "learning_rate": 1.659345243095736e-05, "loss": 2.7669, "step": 25960 }, { "epoch": 0.59, "grad_norm": 1.6046738302902668, "learning_rate": 1.6590706369648465e-05, "loss": 2.8321, "step": 25970 }, { "epoch": 0.59, "grad_norm": 1.846319236012948, "learning_rate": 1.6587959429393665e-05, "loss": 2.8153, "step": 25980 }, { "epoch": 0.59, "grad_norm": 2.229970991690324, "learning_rate": 1.6585211610559294e-05, "loss": 2.6415, "step": 25990 }, { "epoch": 0.59, "grad_norm": 1.9652553281283134, "learning_rate": 1.65824629135118e-05, "loss": 2.6186, "step": 26000 }, { "epoch": 0.59, "grad_norm": 1.672046684657116, "learning_rate": 1.657971333861777e-05, "loss": 2.7178, "step": 26010 }, { "epoch": 0.59, "grad_norm": 1.8907603493643372, "learning_rate": 1.657696288624387e-05, "loss": 2.7785, "step": 26020 }, { "epoch": 0.59, "grad_norm": 1.6955133816605783, "learning_rate": 1.6574211556756923e-05, "loss": 2.8463, "step": 26030 }, { "epoch": 0.59, "grad_norm": 2.43426445670222, "learning_rate": 1.6571459350523836e-05, "loss": 2.7253, "step": 26040 }, { "epoch": 0.59, "grad_norm": 1.9038254873341331, "learning_rate": 1.6568706267911655e-05, "loss": 2.7432, "step": 26050 }, { "epoch": 0.59, "grad_norm": 1.7694151210514555, "learning_rate": 1.656595230928753e-05, "loss": 2.6325, "step": 26060 }, { "epoch": 0.59, "grad_norm": 1.7408870751386927, "learning_rate": 1.6563197475018736e-05, "loss": 2.6358, "step": 26070 }, { "epoch": 0.59, "grad_norm": 1.6716187001918934, "learning_rate": 1.656044176547266e-05, "loss": 2.5644, "step": 26080 }, { "epoch": 0.59, "grad_norm": 1.9647417365074382, "learning_rate": 1.655768518101681e-05, "loss": 2.7132, "step": 26090 }, { "epoch": 0.59, "grad_norm": 1.8205669490525236, "learning_rate": 1.6554927722018806e-05, "loss": 2.6897, "step": 26100 }, { "epoch": 0.59, "grad_norm": 1.8141963878127072, "learning_rate": 1.6552169388846388e-05, "loss": 2.7351, "step": 26110 }, { "epoch": 0.59, "grad_norm": 1.7652636889402544, "learning_rate": 1.6549410181867408e-05, "loss": 2.665, "step": 26120 }, { "epoch": 0.59, "grad_norm": 1.7368603871122017, "learning_rate": 1.6546650101449843e-05, "loss": 2.6704, "step": 26130 }, { "epoch": 0.59, "grad_norm": 1.7564377425954263, "learning_rate": 1.6543889147961774e-05, "loss": 2.7678, "step": 26140 }, { "epoch": 0.59, "grad_norm": 1.9464664523310886, "learning_rate": 1.654112732177141e-05, "loss": 2.6609, "step": 26150 }, { "epoch": 0.59, "grad_norm": 1.9754845979580657, "learning_rate": 1.6538364623247072e-05, "loss": 2.5757, "step": 26160 }, { "epoch": 0.59, "grad_norm": 1.838409747870078, "learning_rate": 1.65356010527572e-05, "loss": 2.6073, "step": 26170 }, { "epoch": 0.59, "grad_norm": 1.7595851177136284, "learning_rate": 1.6532836610670342e-05, "loss": 2.7479, "step": 26180 }, { "epoch": 0.59, "grad_norm": 1.82242625786526, "learning_rate": 1.6530071297355172e-05, "loss": 2.6015, "step": 26190 }, { "epoch": 0.59, "grad_norm": 2.28603870873394, "learning_rate": 1.6527305113180475e-05, "loss": 2.6952, "step": 26200 }, { "epoch": 0.59, "grad_norm": 2.1312667634685076, "learning_rate": 1.652453805851515e-05, "loss": 2.782, "step": 26210 }, { "epoch": 0.59, "grad_norm": 1.8393039390303707, "learning_rate": 1.652177013372822e-05, "loss": 2.677, "step": 26220 }, { "epoch": 0.59, "grad_norm": 2.308096890831493, "learning_rate": 1.6519001339188823e-05, "loss": 2.7108, "step": 26230 }, { "epoch": 0.59, "grad_norm": 1.9488334145883728, "learning_rate": 1.6516231675266196e-05, "loss": 2.6667, "step": 26240 }, { "epoch": 0.59, "grad_norm": 1.9527697929871088, "learning_rate": 1.6513461142329718e-05, "loss": 2.6269, "step": 26250 }, { "epoch": 0.59, "grad_norm": 1.7692591813730878, "learning_rate": 1.6510689740748867e-05, "loss": 2.5499, "step": 26260 }, { "epoch": 0.59, "grad_norm": 1.7554205839978758, "learning_rate": 1.6507917470893237e-05, "loss": 2.668, "step": 26270 }, { "epoch": 0.59, "grad_norm": 1.7458756497729724, "learning_rate": 1.650514433313255e-05, "loss": 2.6681, "step": 26280 }, { "epoch": 0.59, "grad_norm": 1.9799970547035841, "learning_rate": 1.6502370327836626e-05, "loss": 2.6881, "step": 26290 }, { "epoch": 0.59, "grad_norm": 1.5072448682343906, "learning_rate": 1.6499595455375416e-05, "loss": 2.6208, "step": 26300 }, { "epoch": 0.59, "grad_norm": 1.9782306545536257, "learning_rate": 1.649681971611898e-05, "loss": 2.6466, "step": 26310 }, { "epoch": 0.59, "grad_norm": 1.7947818057490232, "learning_rate": 1.6494043110437497e-05, "loss": 2.6924, "step": 26320 }, { "epoch": 0.59, "grad_norm": 1.5292416040650019, "learning_rate": 1.6491265638701257e-05, "loss": 2.7353, "step": 26330 }, { "epoch": 0.59, "grad_norm": 1.8043774483743973, "learning_rate": 1.6488487301280665e-05, "loss": 2.6769, "step": 26340 }, { "epoch": 0.59, "grad_norm": 1.671902880369789, "learning_rate": 1.6485708098546248e-05, "loss": 2.5864, "step": 26350 }, { "epoch": 0.59, "grad_norm": 1.70804421523666, "learning_rate": 1.648292803086864e-05, "loss": 2.6988, "step": 26360 }, { "epoch": 0.59, "grad_norm": 2.254097567512537, "learning_rate": 1.6480147098618596e-05, "loss": 2.6455, "step": 26370 }, { "epoch": 0.59, "grad_norm": 1.7111347455019548, "learning_rate": 1.6477365302166993e-05, "loss": 2.6375, "step": 26380 }, { "epoch": 0.6, "grad_norm": 1.970657501415533, "learning_rate": 1.6474582641884804e-05, "loss": 2.6186, "step": 26390 }, { "epoch": 0.6, "grad_norm": 1.8237509867708734, "learning_rate": 1.6471799118143133e-05, "loss": 2.609, "step": 26400 }, { "epoch": 0.6, "grad_norm": 1.7068471512511854, "learning_rate": 1.6469014731313196e-05, "loss": 2.707, "step": 26410 }, { "epoch": 0.6, "grad_norm": 1.6388007365397326, "learning_rate": 1.646622948176632e-05, "loss": 2.7015, "step": 26420 }, { "epoch": 0.6, "grad_norm": 2.283626896423344, "learning_rate": 1.6463443369873955e-05, "loss": 2.5537, "step": 26430 }, { "epoch": 0.6, "grad_norm": 1.6370316034840653, "learning_rate": 1.6460656396007655e-05, "loss": 2.7933, "step": 26440 }, { "epoch": 0.6, "grad_norm": 1.6360703079915642, "learning_rate": 1.6457868560539096e-05, "loss": 2.6227, "step": 26450 }, { "epoch": 0.6, "grad_norm": 1.668004673918294, "learning_rate": 1.645507986384007e-05, "loss": 2.7641, "step": 26460 }, { "epoch": 0.6, "grad_norm": 1.461883458536426, "learning_rate": 1.645229030628248e-05, "loss": 2.8146, "step": 26470 }, { "epoch": 0.6, "grad_norm": 1.6853386351221975, "learning_rate": 1.6449499888238345e-05, "loss": 2.6322, "step": 26480 }, { "epoch": 0.6, "grad_norm": 1.8923560654399552, "learning_rate": 1.6446708610079798e-05, "loss": 2.6668, "step": 26490 }, { "epoch": 0.6, "grad_norm": 1.8705638244070995, "learning_rate": 1.6443916472179093e-05, "loss": 2.6372, "step": 26500 }, { "epoch": 0.6, "grad_norm": 2.3111277509490677, "learning_rate": 1.644112347490859e-05, "loss": 2.7068, "step": 26510 }, { "epoch": 0.6, "grad_norm": 1.5791965581180727, "learning_rate": 1.6438329618640763e-05, "loss": 2.6208, "step": 26520 }, { "epoch": 0.6, "grad_norm": 1.7976695084493144, "learning_rate": 1.6435534903748212e-05, "loss": 2.6571, "step": 26530 }, { "epoch": 0.6, "grad_norm": 1.5905681655178807, "learning_rate": 1.643273933060364e-05, "loss": 2.7064, "step": 26540 }, { "epoch": 0.6, "grad_norm": 1.599819643912167, "learning_rate": 1.6429942899579868e-05, "loss": 2.7572, "step": 26550 }, { "epoch": 0.6, "grad_norm": 1.6892851901839028, "learning_rate": 1.6427145611049836e-05, "loss": 2.6796, "step": 26560 }, { "epoch": 0.6, "grad_norm": 1.8115546064530637, "learning_rate": 1.6424347465386586e-05, "loss": 2.625, "step": 26570 }, { "epoch": 0.6, "grad_norm": 1.6344738070402491, "learning_rate": 1.6421548462963296e-05, "loss": 2.6405, "step": 26580 }, { "epoch": 0.6, "grad_norm": 1.7833261276778662, "learning_rate": 1.6418748604153233e-05, "loss": 2.7332, "step": 26590 }, { "epoch": 0.6, "grad_norm": 1.519915990691494, "learning_rate": 1.6415947889329793e-05, "loss": 2.6532, "step": 26600 }, { "epoch": 0.6, "grad_norm": 1.6737244490255323, "learning_rate": 1.6413146318866488e-05, "loss": 2.7089, "step": 26610 }, { "epoch": 0.6, "grad_norm": 1.6282684951181794, "learning_rate": 1.6410343893136932e-05, "loss": 2.6551, "step": 26620 }, { "epoch": 0.6, "grad_norm": 1.6135572534450813, "learning_rate": 1.6407540612514865e-05, "loss": 2.7113, "step": 26630 }, { "epoch": 0.6, "grad_norm": 1.7347691335100381, "learning_rate": 1.6404736477374136e-05, "loss": 2.6478, "step": 26640 }, { "epoch": 0.6, "grad_norm": 2.099307290675984, "learning_rate": 1.640193148808871e-05, "loss": 2.7326, "step": 26650 }, { "epoch": 0.6, "grad_norm": 1.8362249562766042, "learning_rate": 1.6399125645032664e-05, "loss": 2.6955, "step": 26660 }, { "epoch": 0.6, "grad_norm": 1.627765973758732, "learning_rate": 1.6396318948580187e-05, "loss": 2.6578, "step": 26670 }, { "epoch": 0.6, "grad_norm": 1.658685849950947, "learning_rate": 1.6393511399105585e-05, "loss": 2.6282, "step": 26680 }, { "epoch": 0.6, "grad_norm": 1.7248730641397507, "learning_rate": 1.6390702996983276e-05, "loss": 2.6151, "step": 26690 }, { "epoch": 0.6, "grad_norm": 1.7883179877627693, "learning_rate": 1.6387893742587788e-05, "loss": 2.559, "step": 26700 }, { "epoch": 0.6, "grad_norm": 1.7383187010055923, "learning_rate": 1.6385083636293776e-05, "loss": 2.7032, "step": 26710 }, { "epoch": 0.6, "grad_norm": 1.7588626725293053, "learning_rate": 1.6382272678476e-05, "loss": 2.604, "step": 26720 }, { "epoch": 0.6, "grad_norm": 1.6896675320047663, "learning_rate": 1.6379460869509324e-05, "loss": 2.5562, "step": 26730 }, { "epoch": 0.6, "grad_norm": 1.5546584309729965, "learning_rate": 1.6376648209768742e-05, "loss": 2.6841, "step": 26740 }, { "epoch": 0.6, "grad_norm": 1.5758511823835282, "learning_rate": 1.6373834699629352e-05, "loss": 2.693, "step": 26750 }, { "epoch": 0.6, "grad_norm": 1.6840172002379468, "learning_rate": 1.6371020339466368e-05, "loss": 2.5202, "step": 26760 }, { "epoch": 0.6, "grad_norm": 1.863444416517163, "learning_rate": 1.636820512965512e-05, "loss": 2.5857, "step": 26770 }, { "epoch": 0.6, "grad_norm": 1.7554096446095406, "learning_rate": 1.6365389070571042e-05, "loss": 2.5937, "step": 26780 }, { "epoch": 0.6, "grad_norm": 1.63218223543499, "learning_rate": 1.636257216258969e-05, "loss": 2.7026, "step": 26790 }, { "epoch": 0.6, "grad_norm": 1.879843479617524, "learning_rate": 1.6359754406086734e-05, "loss": 2.5999, "step": 26800 }, { "epoch": 0.6, "grad_norm": 1.8157912890696035, "learning_rate": 1.635693580143795e-05, "loss": 2.7234, "step": 26810 }, { "epoch": 0.6, "grad_norm": 1.7276697575373325, "learning_rate": 1.6354116349019235e-05, "loss": 2.7374, "step": 26820 }, { "epoch": 0.61, "grad_norm": 1.8703880586685764, "learning_rate": 1.6351296049206593e-05, "loss": 2.6193, "step": 26830 }, { "epoch": 0.61, "grad_norm": 1.6425621714582224, "learning_rate": 1.6348474902376142e-05, "loss": 2.5793, "step": 26840 }, { "epoch": 0.61, "grad_norm": 1.6153648658050501, "learning_rate": 1.6345652908904117e-05, "loss": 2.6435, "step": 26850 }, { "epoch": 0.61, "grad_norm": 1.5888834209394846, "learning_rate": 1.6342830069166858e-05, "loss": 2.6692, "step": 26860 }, { "epoch": 0.61, "grad_norm": 1.987882473761171, "learning_rate": 1.634000638354083e-05, "loss": 2.59, "step": 26870 }, { "epoch": 0.61, "grad_norm": 1.7429500205728814, "learning_rate": 1.63371818524026e-05, "loss": 2.6796, "step": 26880 }, { "epoch": 0.61, "grad_norm": 1.589248329313276, "learning_rate": 1.633435647612885e-05, "loss": 2.6262, "step": 26890 }, { "epoch": 0.61, "grad_norm": 1.8342520518999392, "learning_rate": 1.6331530255096376e-05, "loss": 2.7015, "step": 26900 }, { "epoch": 0.61, "grad_norm": 1.5365515176658855, "learning_rate": 1.632870318968209e-05, "loss": 2.6773, "step": 26910 }, { "epoch": 0.61, "grad_norm": 1.5228591735597767, "learning_rate": 1.6325875280263013e-05, "loss": 2.676, "step": 26920 }, { "epoch": 0.61, "grad_norm": 1.941440319283841, "learning_rate": 1.6323046527216276e-05, "loss": 2.6232, "step": 26930 }, { "epoch": 0.61, "grad_norm": 1.8736045542988915, "learning_rate": 1.632021693091913e-05, "loss": 2.6144, "step": 26940 }, { "epoch": 0.61, "grad_norm": 1.9620008334558021, "learning_rate": 1.631738649174893e-05, "loss": 2.6265, "step": 26950 }, { "epoch": 0.61, "grad_norm": 1.760965233724303, "learning_rate": 1.6314555210083146e-05, "loss": 2.7072, "step": 26960 }, { "epoch": 0.61, "grad_norm": 1.6259729114022523, "learning_rate": 1.6311723086299367e-05, "loss": 2.6832, "step": 26970 }, { "epoch": 0.61, "grad_norm": 1.6400484813930873, "learning_rate": 1.6308890120775285e-05, "loss": 2.6888, "step": 26980 }, { "epoch": 0.61, "grad_norm": 1.9101382940567095, "learning_rate": 1.6306056313888715e-05, "loss": 2.5906, "step": 26990 }, { "epoch": 0.61, "grad_norm": 2.1601547605692963, "learning_rate": 1.6303221666017565e-05, "loss": 2.7331, "step": 27000 }, { "epoch": 0.61, "grad_norm": 1.6190082828440653, "learning_rate": 1.630038617753988e-05, "loss": 2.5178, "step": 27010 }, { "epoch": 0.61, "grad_norm": 2.038444520974166, "learning_rate": 1.6297549848833795e-05, "loss": 2.6481, "step": 27020 }, { "epoch": 0.61, "grad_norm": 2.7174541611009997, "learning_rate": 1.6294712680277576e-05, "loss": 2.7068, "step": 27030 }, { "epoch": 0.61, "grad_norm": 2.013139829774394, "learning_rate": 1.6291874672249587e-05, "loss": 2.694, "step": 27040 }, { "epoch": 0.61, "grad_norm": 1.963419374288473, "learning_rate": 1.6289035825128308e-05, "loss": 2.6285, "step": 27050 }, { "epoch": 0.61, "grad_norm": 1.8343054478429242, "learning_rate": 1.628619613929233e-05, "loss": 2.6801, "step": 27060 }, { "epoch": 0.61, "grad_norm": 1.7969334161489177, "learning_rate": 1.6283355615120362e-05, "loss": 2.6252, "step": 27070 }, { "epoch": 0.61, "grad_norm": 2.4067539015167965, "learning_rate": 1.628051425299122e-05, "loss": 2.7293, "step": 27080 }, { "epoch": 0.61, "grad_norm": 1.6222006589984765, "learning_rate": 1.627767205328383e-05, "loss": 2.7329, "step": 27090 }, { "epoch": 0.61, "grad_norm": 1.72464734877612, "learning_rate": 1.627482901637723e-05, "loss": 2.5906, "step": 27100 }, { "epoch": 0.61, "grad_norm": 1.615578229024067, "learning_rate": 1.6271985142650572e-05, "loss": 2.648, "step": 27110 }, { "epoch": 0.61, "grad_norm": 1.7302105664077418, "learning_rate": 1.6269140432483123e-05, "loss": 2.6706, "step": 27120 }, { "epoch": 0.61, "grad_norm": 1.5024986744480664, "learning_rate": 1.626629488625426e-05, "loss": 2.7304, "step": 27130 }, { "epoch": 0.61, "grad_norm": 1.867091216941638, "learning_rate": 1.626344850434346e-05, "loss": 2.7317, "step": 27140 }, { "epoch": 0.61, "grad_norm": 2.960737225639018, "learning_rate": 1.6260601287130323e-05, "loss": 2.6386, "step": 27150 }, { "epoch": 0.61, "grad_norm": 1.610992020001633, "learning_rate": 1.6257753234994563e-05, "loss": 2.6487, "step": 27160 }, { "epoch": 0.61, "grad_norm": 2.178218683071549, "learning_rate": 1.6254904348315994e-05, "loss": 2.6697, "step": 27170 }, { "epoch": 0.61, "grad_norm": 1.8074742475507564, "learning_rate": 1.6252054627474554e-05, "loss": 2.6014, "step": 27180 }, { "epoch": 0.61, "grad_norm": 1.7015966866749612, "learning_rate": 1.624920407285028e-05, "loss": 2.6635, "step": 27190 }, { "epoch": 0.61, "grad_norm": 1.5075850616517519, "learning_rate": 1.624635268482333e-05, "loss": 2.7005, "step": 27200 }, { "epoch": 0.61, "grad_norm": 2.0010076657602274, "learning_rate": 1.6243500463773965e-05, "loss": 2.7353, "step": 27210 }, { "epoch": 0.61, "grad_norm": 2.0148256234252537, "learning_rate": 1.624064741008257e-05, "loss": 2.6401, "step": 27220 }, { "epoch": 0.61, "grad_norm": 1.7440258893545433, "learning_rate": 1.623779352412962e-05, "loss": 2.7654, "step": 27230 }, { "epoch": 0.61, "grad_norm": 1.7040155361524116, "learning_rate": 1.6234938806295718e-05, "loss": 2.6957, "step": 27240 }, { "epoch": 0.61, "grad_norm": 1.6640567857915927, "learning_rate": 1.6232083256961577e-05, "loss": 2.6222, "step": 27250 }, { "epoch": 0.61, "grad_norm": 1.8342837585906067, "learning_rate": 1.6229226876508015e-05, "loss": 2.6293, "step": 27260 }, { "epoch": 0.61, "grad_norm": 1.664423909801853, "learning_rate": 1.6226369665315963e-05, "loss": 2.5031, "step": 27270 }, { "epoch": 0.62, "grad_norm": 1.72015374361296, "learning_rate": 1.622351162376646e-05, "loss": 2.7008, "step": 27280 }, { "epoch": 0.62, "grad_norm": 1.586698301615575, "learning_rate": 1.622065275224066e-05, "loss": 2.6734, "step": 27290 }, { "epoch": 0.62, "grad_norm": 1.6927375689710435, "learning_rate": 1.621779305111983e-05, "loss": 2.6055, "step": 27300 }, { "epoch": 0.62, "grad_norm": 1.6699391648734696, "learning_rate": 1.621493252078534e-05, "loss": 2.6668, "step": 27310 }, { "epoch": 0.62, "grad_norm": 2.0579994315438794, "learning_rate": 1.6212071161618675e-05, "loss": 2.6029, "step": 27320 }, { "epoch": 0.62, "grad_norm": 1.5355269397968125, "learning_rate": 1.6209208974001427e-05, "loss": 2.7238, "step": 27330 }, { "epoch": 0.62, "grad_norm": 1.6469421373356103, "learning_rate": 1.6206345958315307e-05, "loss": 2.6553, "step": 27340 }, { "epoch": 0.62, "grad_norm": 1.6115283589979246, "learning_rate": 1.6203482114942126e-05, "loss": 2.7167, "step": 27350 }, { "epoch": 0.62, "grad_norm": 1.6101539877960744, "learning_rate": 1.6200617444263812e-05, "loss": 2.5711, "step": 27360 }, { "epoch": 0.62, "grad_norm": 1.5982072756476258, "learning_rate": 1.6197751946662404e-05, "loss": 2.7607, "step": 27370 }, { "epoch": 0.62, "grad_norm": 1.5553034288049854, "learning_rate": 1.6194885622520046e-05, "loss": 2.684, "step": 27380 }, { "epoch": 0.62, "grad_norm": 1.7112529993842647, "learning_rate": 1.6192018472218994e-05, "loss": 2.4905, "step": 27390 }, { "epoch": 0.62, "grad_norm": 1.44237471279644, "learning_rate": 1.6189150496141618e-05, "loss": 2.7098, "step": 27400 }, { "epoch": 0.62, "grad_norm": 1.9555607007322793, "learning_rate": 1.6186281694670395e-05, "loss": 2.6204, "step": 27410 }, { "epoch": 0.62, "grad_norm": 1.7055495550467885, "learning_rate": 1.6183412068187912e-05, "loss": 2.6359, "step": 27420 }, { "epoch": 0.62, "grad_norm": 1.7180822210669355, "learning_rate": 1.6180541617076862e-05, "loss": 2.727, "step": 27430 }, { "epoch": 0.62, "grad_norm": 1.556051632941122, "learning_rate": 1.617767034172006e-05, "loss": 2.6569, "step": 27440 }, { "epoch": 0.62, "grad_norm": 3.202244818684371, "learning_rate": 1.6174798242500418e-05, "loss": 2.747, "step": 27450 }, { "epoch": 0.62, "grad_norm": 1.8255540603006775, "learning_rate": 1.6171925319800964e-05, "loss": 2.5883, "step": 27460 }, { "epoch": 0.62, "grad_norm": 2.115521860259845, "learning_rate": 1.616905157400484e-05, "loss": 2.7045, "step": 27470 }, { "epoch": 0.62, "grad_norm": 2.392935526832627, "learning_rate": 1.6166177005495284e-05, "loss": 2.5666, "step": 27480 }, { "epoch": 0.62, "grad_norm": 1.648888956895753, "learning_rate": 1.616330161465566e-05, "loss": 2.6647, "step": 27490 }, { "epoch": 0.62, "grad_norm": 1.8232381503124313, "learning_rate": 1.6160425401869432e-05, "loss": 2.6682, "step": 27500 }, { "epoch": 0.62, "grad_norm": 1.7134888830378563, "learning_rate": 1.6157548367520176e-05, "loss": 2.7438, "step": 27510 }, { "epoch": 0.62, "grad_norm": 2.0053251110276777, "learning_rate": 1.6154670511991573e-05, "loss": 2.7136, "step": 27520 }, { "epoch": 0.62, "grad_norm": 1.8319304195630735, "learning_rate": 1.6151791835667424e-05, "loss": 2.6392, "step": 27530 }, { "epoch": 0.62, "grad_norm": 1.8967170284443273, "learning_rate": 1.6148912338931628e-05, "loss": 2.6419, "step": 27540 }, { "epoch": 0.62, "grad_norm": 1.6900137169896128, "learning_rate": 1.6146032022168206e-05, "loss": 2.7715, "step": 27550 }, { "epoch": 0.62, "grad_norm": 1.7744618493820705, "learning_rate": 1.6143150885761274e-05, "loss": 2.5952, "step": 27560 }, { "epoch": 0.62, "grad_norm": 1.7241452510627024, "learning_rate": 1.6140268930095066e-05, "loss": 2.6228, "step": 27570 }, { "epoch": 0.62, "grad_norm": 2.0195809930076467, "learning_rate": 1.6137386155553925e-05, "loss": 2.6736, "step": 27580 }, { "epoch": 0.62, "grad_norm": 1.7017483718742688, "learning_rate": 1.6134502562522303e-05, "loss": 2.6861, "step": 27590 }, { "epoch": 0.62, "grad_norm": 1.777738150390671, "learning_rate": 1.6131618151384756e-05, "loss": 2.6355, "step": 27600 }, { "epoch": 0.62, "grad_norm": 1.5070648334189436, "learning_rate": 1.612873292252596e-05, "loss": 2.7524, "step": 27610 }, { "epoch": 0.62, "grad_norm": 1.6848546853446817, "learning_rate": 1.612584687633069e-05, "loss": 2.6533, "step": 27620 }, { "epoch": 0.62, "grad_norm": 1.635674922671185, "learning_rate": 1.612296001318383e-05, "loss": 2.6588, "step": 27630 }, { "epoch": 0.62, "grad_norm": 1.6439510672743887, "learning_rate": 1.6120072333470375e-05, "loss": 2.7821, "step": 27640 }, { "epoch": 0.62, "grad_norm": 1.7838510175330944, "learning_rate": 1.6117183837575437e-05, "loss": 2.6783, "step": 27650 }, { "epoch": 0.62, "grad_norm": 1.9704443822429185, "learning_rate": 1.6114294525884227e-05, "loss": 2.6379, "step": 27660 }, { "epoch": 0.62, "grad_norm": 1.5246805703271213, "learning_rate": 1.6111404398782065e-05, "loss": 2.6212, "step": 27670 }, { "epoch": 0.62, "grad_norm": 1.7086511142783127, "learning_rate": 1.6108513456654383e-05, "loss": 2.5951, "step": 27680 }, { "epoch": 0.62, "grad_norm": 2.0833631492324485, "learning_rate": 1.6105621699886727e-05, "loss": 2.6027, "step": 27690 }, { "epoch": 0.62, "grad_norm": 1.6565471922252264, "learning_rate": 1.6102729128864737e-05, "loss": 2.6702, "step": 27700 }, { "epoch": 0.62, "grad_norm": 1.640839776110814, "learning_rate": 1.6099835743974178e-05, "loss": 2.7186, "step": 27710 }, { "epoch": 0.63, "grad_norm": 1.906975258784713, "learning_rate": 1.609694154560091e-05, "loss": 2.5743, "step": 27720 }, { "epoch": 0.63, "grad_norm": 1.4758077767944415, "learning_rate": 1.609404653413091e-05, "loss": 2.7051, "step": 27730 }, { "epoch": 0.63, "grad_norm": 1.8502590131700876, "learning_rate": 1.6091150709950266e-05, "loss": 2.6939, "step": 27740 }, { "epoch": 0.63, "grad_norm": 1.6810283869495557, "learning_rate": 1.608825407344516e-05, "loss": 2.5152, "step": 27750 }, { "epoch": 0.63, "grad_norm": 1.5945446068264877, "learning_rate": 1.6085356625001895e-05, "loss": 2.6946, "step": 27760 }, { "epoch": 0.63, "grad_norm": 1.8382235157333102, "learning_rate": 1.6082458365006876e-05, "loss": 2.7515, "step": 27770 }, { "epoch": 0.63, "grad_norm": 1.547254156366644, "learning_rate": 1.6079559293846628e-05, "loss": 2.644, "step": 27780 }, { "epoch": 0.63, "grad_norm": 1.7542828913342852, "learning_rate": 1.6076659411907768e-05, "loss": 2.6777, "step": 27790 }, { "epoch": 0.63, "grad_norm": 1.6608576079812674, "learning_rate": 1.6073758719577027e-05, "loss": 2.6565, "step": 27800 }, { "epoch": 0.63, "grad_norm": 2.007509802036945, "learning_rate": 1.6070857217241252e-05, "loss": 2.5661, "step": 27810 }, { "epoch": 0.63, "grad_norm": 1.735444645590817, "learning_rate": 1.6067954905287388e-05, "loss": 2.7942, "step": 27820 }, { "epoch": 0.63, "grad_norm": 1.7884156166977716, "learning_rate": 1.6065051784102488e-05, "loss": 2.5657, "step": 27830 }, { "epoch": 0.63, "grad_norm": 1.950123953969983, "learning_rate": 1.606214785407372e-05, "loss": 2.6651, "step": 27840 }, { "epoch": 0.63, "grad_norm": 1.666524426155253, "learning_rate": 1.605924311558836e-05, "loss": 2.7363, "step": 27850 }, { "epoch": 0.63, "grad_norm": 1.672083514137542, "learning_rate": 1.6056337569033778e-05, "loss": 2.6338, "step": 27860 }, { "epoch": 0.63, "grad_norm": 2.249356905987987, "learning_rate": 1.605343121479747e-05, "loss": 2.7515, "step": 27870 }, { "epoch": 0.63, "grad_norm": 1.8436827201813957, "learning_rate": 1.6050524053267033e-05, "loss": 2.6, "step": 27880 }, { "epoch": 0.63, "grad_norm": 1.9889947127911631, "learning_rate": 1.6047616084830162e-05, "loss": 2.564, "step": 27890 }, { "epoch": 0.63, "grad_norm": 1.8210617658715806, "learning_rate": 1.6044707309874673e-05, "loss": 2.6027, "step": 27900 }, { "epoch": 0.63, "grad_norm": 1.7349043627709448, "learning_rate": 1.6041797728788484e-05, "loss": 2.6587, "step": 27910 }, { "epoch": 0.63, "grad_norm": 1.8325864074963096, "learning_rate": 1.603888734195962e-05, "loss": 2.5972, "step": 27920 }, { "epoch": 0.63, "grad_norm": 1.8163086929961978, "learning_rate": 1.6035976149776213e-05, "loss": 2.5317, "step": 27930 }, { "epoch": 0.63, "grad_norm": 1.858179392517054, "learning_rate": 1.6033064152626506e-05, "loss": 2.7329, "step": 27940 }, { "epoch": 0.63, "grad_norm": 1.7162041801909098, "learning_rate": 1.603015135089885e-05, "loss": 2.6196, "step": 27950 }, { "epoch": 0.63, "grad_norm": 2.169853095878204, "learning_rate": 1.602723774498169e-05, "loss": 2.7162, "step": 27960 }, { "epoch": 0.63, "grad_norm": 1.471842873500496, "learning_rate": 1.6024323335263597e-05, "loss": 2.6963, "step": 27970 }, { "epoch": 0.63, "grad_norm": 1.7701250840625582, "learning_rate": 1.602140812213324e-05, "loss": 2.6435, "step": 27980 }, { "epoch": 0.63, "grad_norm": 1.8103732505362526, "learning_rate": 1.6018492105979394e-05, "loss": 2.7204, "step": 27990 }, { "epoch": 0.63, "grad_norm": 1.8851182516077871, "learning_rate": 1.6015575287190947e-05, "loss": 2.719, "step": 28000 }, { "epoch": 0.63, "grad_norm": 1.9350106984696152, "learning_rate": 1.6012657666156887e-05, "loss": 2.7032, "step": 28010 }, { "epoch": 0.63, "grad_norm": 1.8465343931154252, "learning_rate": 1.6009739243266308e-05, "loss": 2.6565, "step": 28020 }, { "epoch": 0.63, "grad_norm": 1.743168101257181, "learning_rate": 1.600682001890842e-05, "loss": 2.707, "step": 28030 }, { "epoch": 0.63, "grad_norm": 1.8657863275833508, "learning_rate": 1.6003899993472535e-05, "loss": 2.7637, "step": 28040 }, { "epoch": 0.63, "grad_norm": 2.2741414258227577, "learning_rate": 1.6000979167348073e-05, "loss": 2.6145, "step": 28050 }, { "epoch": 0.63, "grad_norm": 1.8811187437521757, "learning_rate": 1.5998057540924555e-05, "loss": 2.7047, "step": 28060 }, { "epoch": 0.63, "grad_norm": 1.6669089785472273, "learning_rate": 1.5995135114591612e-05, "loss": 2.7064, "step": 28070 }, { "epoch": 0.63, "grad_norm": 1.5267855908221044, "learning_rate": 1.5992211888738992e-05, "loss": 2.6963, "step": 28080 }, { "epoch": 0.63, "grad_norm": 1.5883290011916202, "learning_rate": 1.5989287863756533e-05, "loss": 2.687, "step": 28090 }, { "epoch": 0.63, "grad_norm": 1.6511528579220358, "learning_rate": 1.5986363040034188e-05, "loss": 2.769, "step": 28100 }, { "epoch": 0.63, "grad_norm": 1.5797054545705242, "learning_rate": 1.598343741796202e-05, "loss": 2.6467, "step": 28110 }, { "epoch": 0.63, "grad_norm": 1.7678401009702924, "learning_rate": 1.5980510997930184e-05, "loss": 2.6137, "step": 28120 }, { "epoch": 0.63, "grad_norm": 2.1374005303197054, "learning_rate": 1.597758378032896e-05, "loss": 2.6841, "step": 28130 }, { "epoch": 0.63, "grad_norm": 1.7943762967570511, "learning_rate": 1.597465576554873e-05, "loss": 2.6802, "step": 28140 }, { "epoch": 0.63, "grad_norm": 1.9137100652391636, "learning_rate": 1.5971726953979967e-05, "loss": 2.6972, "step": 28150 }, { "epoch": 0.64, "grad_norm": 1.6391308594597667, "learning_rate": 1.5968797346013268e-05, "loss": 2.7257, "step": 28160 }, { "epoch": 0.64, "grad_norm": 1.4248967719830952, "learning_rate": 1.596586694203933e-05, "loss": 2.6369, "step": 28170 }, { "epoch": 0.64, "grad_norm": 1.7265764793251255, "learning_rate": 1.596293574244895e-05, "loss": 2.7365, "step": 28180 }, { "epoch": 0.64, "grad_norm": 1.7162103069573373, "learning_rate": 1.5960003747633043e-05, "loss": 2.5426, "step": 28190 }, { "epoch": 0.64, "grad_norm": 1.6742145760173666, "learning_rate": 1.5957070957982623e-05, "loss": 2.6625, "step": 28200 }, { "epoch": 0.64, "grad_norm": 1.7854459127831344, "learning_rate": 1.5954137373888808e-05, "loss": 2.7545, "step": 28210 }, { "epoch": 0.64, "grad_norm": 1.784830270968913, "learning_rate": 1.5951202995742828e-05, "loss": 2.691, "step": 28220 }, { "epoch": 0.64, "grad_norm": 1.9105460989125895, "learning_rate": 1.5948267823936013e-05, "loss": 2.531, "step": 28230 }, { "epoch": 0.64, "grad_norm": 1.7751058452097228, "learning_rate": 1.5945331858859807e-05, "loss": 2.7755, "step": 28240 }, { "epoch": 0.64, "grad_norm": 1.8794991575418991, "learning_rate": 1.594239510090575e-05, "loss": 2.5287, "step": 28250 }, { "epoch": 0.64, "grad_norm": 1.8373470353956616, "learning_rate": 1.593945755046549e-05, "loss": 2.6394, "step": 28260 }, { "epoch": 0.64, "grad_norm": 1.8445885636920822, "learning_rate": 1.5936519207930787e-05, "loss": 2.7572, "step": 28270 }, { "epoch": 0.64, "grad_norm": 1.9026552430963786, "learning_rate": 1.5933580073693504e-05, "loss": 2.6412, "step": 28280 }, { "epoch": 0.64, "grad_norm": 1.5686407344275854, "learning_rate": 1.5930640148145604e-05, "loss": 2.6785, "step": 28290 }, { "epoch": 0.64, "grad_norm": 1.542181702970103, "learning_rate": 1.592769943167916e-05, "loss": 2.6624, "step": 28300 }, { "epoch": 0.64, "grad_norm": 1.5129456460058617, "learning_rate": 1.5924757924686352e-05, "loss": 2.6143, "step": 28310 }, { "epoch": 0.64, "grad_norm": 2.048124405724157, "learning_rate": 1.592181562755946e-05, "loss": 2.6719, "step": 28320 }, { "epoch": 0.64, "grad_norm": 1.993344650201737, "learning_rate": 1.5918872540690883e-05, "loss": 2.5857, "step": 28330 }, { "epoch": 0.64, "grad_norm": 2.6207093318044414, "learning_rate": 1.5915928664473103e-05, "loss": 2.6321, "step": 28340 }, { "epoch": 0.64, "grad_norm": 1.6578163379989868, "learning_rate": 1.5912983999298722e-05, "loss": 2.7039, "step": 28350 }, { "epoch": 0.64, "grad_norm": 1.7054179417695594, "learning_rate": 1.591003854556045e-05, "loss": 2.539, "step": 28360 }, { "epoch": 0.64, "grad_norm": 1.623329771693777, "learning_rate": 1.5907092303651093e-05, "loss": 2.586, "step": 28370 }, { "epoch": 0.64, "grad_norm": 1.5488705574472976, "learning_rate": 1.590414527396357e-05, "loss": 2.6925, "step": 28380 }, { "epoch": 0.64, "grad_norm": 1.6179622304687433, "learning_rate": 1.5901197456890894e-05, "loss": 2.6047, "step": 28390 }, { "epoch": 0.64, "grad_norm": 1.8386661243116997, "learning_rate": 1.5898248852826196e-05, "loss": 2.6975, "step": 28400 }, { "epoch": 0.64, "grad_norm": 1.6530917185128817, "learning_rate": 1.58952994621627e-05, "loss": 2.6799, "step": 28410 }, { "epoch": 0.64, "grad_norm": 2.121088040288626, "learning_rate": 1.589234928529375e-05, "loss": 2.7582, "step": 28420 }, { "epoch": 0.64, "grad_norm": 1.6880388512117672, "learning_rate": 1.5889398322612778e-05, "loss": 2.6399, "step": 28430 }, { "epoch": 0.64, "grad_norm": 1.594811806177464, "learning_rate": 1.588644657451333e-05, "loss": 2.6906, "step": 28440 }, { "epoch": 0.64, "grad_norm": 1.924544977148648, "learning_rate": 1.588349404138906e-05, "loss": 2.587, "step": 28450 }, { "epoch": 0.64, "grad_norm": 1.606765135924115, "learning_rate": 1.5880540723633716e-05, "loss": 2.6293, "step": 28460 }, { "epoch": 0.64, "grad_norm": 1.5636858840852719, "learning_rate": 1.587758662164116e-05, "loss": 2.6092, "step": 28470 }, { "epoch": 0.64, "grad_norm": 1.5979747169454344, "learning_rate": 1.587463173580535e-05, "loss": 2.6441, "step": 28480 }, { "epoch": 0.64, "grad_norm": 2.054845136773674, "learning_rate": 1.5871676066520364e-05, "loss": 2.6507, "step": 28490 }, { "epoch": 0.64, "grad_norm": 2.1745126447569874, "learning_rate": 1.586871961418037e-05, "loss": 2.6623, "step": 28500 }, { "epoch": 0.64, "grad_norm": 1.923284159938859, "learning_rate": 1.5865762379179638e-05, "loss": 2.663, "step": 28510 }, { "epoch": 0.64, "grad_norm": 1.8880338598661899, "learning_rate": 1.5862804361912554e-05, "loss": 2.6571, "step": 28520 }, { "epoch": 0.64, "grad_norm": 1.9027372670367237, "learning_rate": 1.5859845562773603e-05, "loss": 2.7153, "step": 28530 }, { "epoch": 0.64, "grad_norm": 1.8270869075771938, "learning_rate": 1.585688598215738e-05, "loss": 2.6661, "step": 28540 }, { "epoch": 0.64, "grad_norm": 1.6817227437878854, "learning_rate": 1.585392562045857e-05, "loss": 2.6897, "step": 28550 }, { "epoch": 0.64, "grad_norm": 1.566641128758595, "learning_rate": 1.5850964478071975e-05, "loss": 2.6343, "step": 28560 }, { "epoch": 0.64, "grad_norm": 1.8034722932447316, "learning_rate": 1.5848002555392494e-05, "loss": 2.6568, "step": 28570 }, { "epoch": 0.64, "grad_norm": 1.8932216219115443, "learning_rate": 1.5845039852815138e-05, "loss": 2.7247, "step": 28580 }, { "epoch": 0.64, "grad_norm": 1.5440730837461196, "learning_rate": 1.5842076370735018e-05, "loss": 2.754, "step": 28590 }, { "epoch": 0.64, "grad_norm": 1.6329568703527322, "learning_rate": 1.5839112109547342e-05, "loss": 2.6263, "step": 28600 }, { "epoch": 0.65, "grad_norm": 1.684971564782271, "learning_rate": 1.5836147069647434e-05, "loss": 2.6564, "step": 28610 }, { "epoch": 0.65, "grad_norm": 1.5301747233991485, "learning_rate": 1.5833181251430713e-05, "loss": 2.6558, "step": 28620 }, { "epoch": 0.65, "grad_norm": 1.7668285228256477, "learning_rate": 1.5830214655292702e-05, "loss": 2.7438, "step": 28630 }, { "epoch": 0.65, "grad_norm": 1.7879909483762977, "learning_rate": 1.5827247281629038e-05, "loss": 2.6343, "step": 28640 }, { "epoch": 0.65, "grad_norm": 1.9209816578829364, "learning_rate": 1.582427913083545e-05, "loss": 2.7139, "step": 28650 }, { "epoch": 0.65, "grad_norm": 1.6457254544829623, "learning_rate": 1.5821310203307773e-05, "loss": 2.652, "step": 28660 }, { "epoch": 0.65, "grad_norm": 1.5538295371610795, "learning_rate": 1.581834049944195e-05, "loss": 2.7425, "step": 28670 }, { "epoch": 0.65, "grad_norm": 1.8648544850063755, "learning_rate": 1.581537001963402e-05, "loss": 2.5869, "step": 28680 }, { "epoch": 0.65, "grad_norm": 1.5983612424865983, "learning_rate": 1.5812398764280135e-05, "loss": 2.7073, "step": 28690 }, { "epoch": 0.65, "grad_norm": 1.6626965532200526, "learning_rate": 1.5809426733776544e-05, "loss": 2.7378, "step": 28700 }, { "epoch": 0.65, "grad_norm": 1.8320483161921228, "learning_rate": 1.580645392851961e-05, "loss": 2.7271, "step": 28710 }, { "epoch": 0.65, "grad_norm": 1.7896693933967596, "learning_rate": 1.5803480348905774e-05, "loss": 2.6917, "step": 28720 }, { "epoch": 0.65, "grad_norm": 1.6126295617109239, "learning_rate": 1.580050599533161e-05, "loss": 2.6962, "step": 28730 }, { "epoch": 0.65, "grad_norm": 1.7150315314631899, "learning_rate": 1.5797530868193773e-05, "loss": 2.7212, "step": 28740 }, { "epoch": 0.65, "grad_norm": 1.8419902919759055, "learning_rate": 1.579455496788904e-05, "loss": 2.624, "step": 28750 }, { "epoch": 0.65, "grad_norm": 1.7272360949386054, "learning_rate": 1.5791578294814276e-05, "loss": 2.6325, "step": 28760 }, { "epoch": 0.65, "grad_norm": 1.6062533819637252, "learning_rate": 1.578860084936645e-05, "loss": 2.7879, "step": 28770 }, { "epoch": 0.65, "grad_norm": 1.5859460378875383, "learning_rate": 1.5785622631942646e-05, "loss": 2.6415, "step": 28780 }, { "epoch": 0.65, "grad_norm": 1.840315226583566, "learning_rate": 1.5782643642940035e-05, "loss": 2.7061, "step": 28790 }, { "epoch": 0.65, "grad_norm": 1.6687940155395988, "learning_rate": 1.5779663882755907e-05, "loss": 2.6134, "step": 28800 }, { "epoch": 0.65, "grad_norm": 1.8761989017279714, "learning_rate": 1.5776683351787644e-05, "loss": 2.693, "step": 28810 }, { "epoch": 0.65, "grad_norm": 1.8646041967127127, "learning_rate": 1.5773702050432735e-05, "loss": 2.6704, "step": 28820 }, { "epoch": 0.65, "grad_norm": 1.6782644595364518, "learning_rate": 1.5770719979088765e-05, "loss": 2.6175, "step": 28830 }, { "epoch": 0.65, "grad_norm": 1.756384307879483, "learning_rate": 1.5767737138153435e-05, "loss": 2.6441, "step": 28840 }, { "epoch": 0.65, "grad_norm": 1.9281051740280333, "learning_rate": 1.576475352802454e-05, "loss": 2.6644, "step": 28850 }, { "epoch": 0.65, "grad_norm": 1.73845073040378, "learning_rate": 1.576176914909997e-05, "loss": 2.6125, "step": 28860 }, { "epoch": 0.65, "grad_norm": 1.665972793469683, "learning_rate": 1.5758784001777738e-05, "loss": 2.7431, "step": 28870 }, { "epoch": 0.65, "grad_norm": 1.7224709639458078, "learning_rate": 1.575579808645594e-05, "loss": 2.688, "step": 28880 }, { "epoch": 0.65, "grad_norm": 2.0685177325294744, "learning_rate": 1.575281140353278e-05, "loss": 2.7969, "step": 28890 }, { "epoch": 0.65, "grad_norm": 1.9645133179639478, "learning_rate": 1.5749823953406573e-05, "loss": 2.6253, "step": 28900 }, { "epoch": 0.65, "grad_norm": 1.6811574029961174, "learning_rate": 1.5746835736475722e-05, "loss": 2.6269, "step": 28910 }, { "epoch": 0.65, "grad_norm": 1.882294530007446, "learning_rate": 1.5743846753138744e-05, "loss": 2.5567, "step": 28920 }, { "epoch": 0.65, "grad_norm": 1.4730924836788637, "learning_rate": 1.574085700379426e-05, "loss": 2.6008, "step": 28930 }, { "epoch": 0.65, "grad_norm": 1.573285473134896, "learning_rate": 1.5737866488840973e-05, "loss": 2.7954, "step": 28940 }, { "epoch": 0.65, "grad_norm": 1.7304549557933866, "learning_rate": 1.573487520867772e-05, "loss": 2.5835, "step": 28950 }, { "epoch": 0.65, "grad_norm": 1.650927635583022, "learning_rate": 1.5731883163703403e-05, "loss": 2.679, "step": 28960 }, { "epoch": 0.65, "grad_norm": 1.6738967913841156, "learning_rate": 1.572889035431706e-05, "loss": 2.525, "step": 28970 }, { "epoch": 0.65, "grad_norm": 1.885529720083742, "learning_rate": 1.5725896780917815e-05, "loss": 2.5345, "step": 28980 }, { "epoch": 0.65, "grad_norm": 1.7771925070082244, "learning_rate": 1.5722902443904886e-05, "loss": 2.7088, "step": 28990 }, { "epoch": 0.65, "grad_norm": 1.7452355202432914, "learning_rate": 1.571990734367761e-05, "loss": 2.6636, "step": 29000 }, { "epoch": 0.65, "grad_norm": 1.586962210019698, "learning_rate": 1.5716911480635414e-05, "loss": 2.7051, "step": 29010 }, { "epoch": 0.65, "grad_norm": 2.254257388061446, "learning_rate": 1.5713914855177834e-05, "loss": 2.6582, "step": 29020 }, { "epoch": 0.65, "grad_norm": 1.8584330391820207, "learning_rate": 1.57109174677045e-05, "loss": 2.5394, "step": 29030 }, { "epoch": 0.65, "grad_norm": 1.837002099302173, "learning_rate": 1.570791931861515e-05, "loss": 2.7112, "step": 29040 }, { "epoch": 0.66, "grad_norm": 1.7045381191751237, "learning_rate": 1.570492040830962e-05, "loss": 2.6546, "step": 29050 }, { "epoch": 0.66, "grad_norm": 1.711046052405394, "learning_rate": 1.5701920737187852e-05, "loss": 2.7679, "step": 29060 }, { "epoch": 0.66, "grad_norm": 1.6452854923518019, "learning_rate": 1.569892030564989e-05, "loss": 2.8464, "step": 29070 }, { "epoch": 0.66, "grad_norm": 1.6097678211061008, "learning_rate": 1.5695919114095862e-05, "loss": 2.8017, "step": 29080 }, { "epoch": 0.66, "grad_norm": 1.6811136835539564, "learning_rate": 1.569291716292602e-05, "loss": 2.5625, "step": 29090 }, { "epoch": 0.66, "grad_norm": 1.6493630825629628, "learning_rate": 1.5689914452540713e-05, "loss": 2.7872, "step": 29100 }, { "epoch": 0.66, "grad_norm": 1.760586648433301, "learning_rate": 1.568691098334038e-05, "loss": 2.6865, "step": 29110 }, { "epoch": 0.66, "grad_norm": 1.9828674857964217, "learning_rate": 1.5683906755725568e-05, "loss": 2.6927, "step": 29120 }, { "epoch": 0.66, "grad_norm": 2.1538127110422836, "learning_rate": 1.5680901770096928e-05, "loss": 2.8559, "step": 29130 }, { "epoch": 0.66, "grad_norm": 1.7870874637056975, "learning_rate": 1.5677896026855207e-05, "loss": 2.7692, "step": 29140 }, { "epoch": 0.66, "grad_norm": 1.7584936050014004, "learning_rate": 1.567488952640126e-05, "loss": 2.5891, "step": 29150 }, { "epoch": 0.66, "grad_norm": 1.8964686703338114, "learning_rate": 1.567188226913603e-05, "loss": 2.5385, "step": 29160 }, { "epoch": 0.66, "grad_norm": 1.586524557617644, "learning_rate": 1.566887425546057e-05, "loss": 2.5505, "step": 29170 }, { "epoch": 0.66, "grad_norm": 1.4940147357668794, "learning_rate": 1.566586548577604e-05, "loss": 2.5211, "step": 29180 }, { "epoch": 0.66, "grad_norm": 1.5702159637290305, "learning_rate": 1.5662855960483692e-05, "loss": 2.6707, "step": 29190 }, { "epoch": 0.66, "grad_norm": 1.5223023840389545, "learning_rate": 1.565984567998488e-05, "loss": 2.6041, "step": 29200 }, { "epoch": 0.66, "grad_norm": 1.6005134754340502, "learning_rate": 1.565683464468105e-05, "loss": 2.6931, "step": 29210 }, { "epoch": 0.66, "grad_norm": 1.6993172726932866, "learning_rate": 1.565382285497377e-05, "loss": 2.6238, "step": 29220 }, { "epoch": 0.66, "grad_norm": 1.6719938733378596, "learning_rate": 1.565081031126469e-05, "loss": 2.6871, "step": 29230 }, { "epoch": 0.66, "grad_norm": 2.025393782557794, "learning_rate": 1.564779701395557e-05, "loss": 2.6567, "step": 29240 }, { "epoch": 0.66, "grad_norm": 1.5854089010767465, "learning_rate": 1.5644782963448268e-05, "loss": 2.6466, "step": 29250 }, { "epoch": 0.66, "grad_norm": 1.7687852423928423, "learning_rate": 1.564176816014474e-05, "loss": 2.6975, "step": 29260 }, { "epoch": 0.66, "grad_norm": 2.1231353864051896, "learning_rate": 1.5638752604447043e-05, "loss": 2.7907, "step": 29270 }, { "epoch": 0.66, "grad_norm": 1.702335945531781, "learning_rate": 1.563573629675734e-05, "loss": 2.6579, "step": 29280 }, { "epoch": 0.66, "grad_norm": 1.8893214836374341, "learning_rate": 1.563271923747789e-05, "loss": 2.5466, "step": 29290 }, { "epoch": 0.66, "grad_norm": 1.7676719445722298, "learning_rate": 1.562970142701105e-05, "loss": 2.6917, "step": 29300 }, { "epoch": 0.66, "grad_norm": 2.0997256428416557, "learning_rate": 1.562668286575928e-05, "loss": 2.6503, "step": 29310 }, { "epoch": 0.66, "grad_norm": 1.6620294560282287, "learning_rate": 1.5623663554125135e-05, "loss": 2.6767, "step": 29320 }, { "epoch": 0.66, "grad_norm": 1.4464705692897804, "learning_rate": 1.5620643492511283e-05, "loss": 2.5876, "step": 29330 }, { "epoch": 0.66, "grad_norm": 1.6280028711554484, "learning_rate": 1.561762268132048e-05, "loss": 2.6583, "step": 29340 }, { "epoch": 0.66, "grad_norm": 1.7489220633860558, "learning_rate": 1.5614601120955584e-05, "loss": 2.6311, "step": 29350 }, { "epoch": 0.66, "grad_norm": 1.7213066640246317, "learning_rate": 1.5611578811819556e-05, "loss": 2.6878, "step": 29360 }, { "epoch": 0.66, "grad_norm": 1.6103149644568409, "learning_rate": 1.5608555754315452e-05, "loss": 2.6612, "step": 29370 }, { "epoch": 0.66, "grad_norm": 1.5083793372816068, "learning_rate": 1.5605531948846443e-05, "loss": 2.6515, "step": 29380 }, { "epoch": 0.66, "grad_norm": 1.7146297337681606, "learning_rate": 1.560250739581577e-05, "loss": 2.5597, "step": 29390 }, { "epoch": 0.66, "grad_norm": 1.9219754706545638, "learning_rate": 1.5599482095626807e-05, "loss": 2.5961, "step": 29400 }, { "epoch": 0.66, "grad_norm": 1.9866912787487645, "learning_rate": 1.5596456048683007e-05, "loss": 2.5374, "step": 29410 }, { "epoch": 0.66, "grad_norm": 1.5282801292383708, "learning_rate": 1.5593429255387923e-05, "loss": 2.6702, "step": 29420 }, { "epoch": 0.66, "grad_norm": 2.0476166491872045, "learning_rate": 1.5590401716145224e-05, "loss": 2.7957, "step": 29430 }, { "epoch": 0.66, "grad_norm": 1.7888013613638454, "learning_rate": 1.558737343135865e-05, "loss": 2.615, "step": 29440 }, { "epoch": 0.66, "grad_norm": 1.6426351361849916, "learning_rate": 1.5584344401432075e-05, "loss": 2.6803, "step": 29450 }, { "epoch": 0.66, "grad_norm": 1.8719969386621278, "learning_rate": 1.5581314626769444e-05, "loss": 2.7508, "step": 29460 }, { "epoch": 0.66, "grad_norm": 1.7846397300560632, "learning_rate": 1.557828410777481e-05, "loss": 2.582, "step": 29470 }, { "epoch": 0.66, "grad_norm": 2.5109843966864567, "learning_rate": 1.557525284485234e-05, "loss": 2.6847, "step": 29480 }, { "epoch": 0.67, "grad_norm": 2.222020642506224, "learning_rate": 1.5572220838406277e-05, "loss": 2.6567, "step": 29490 }, { "epoch": 0.67, "grad_norm": 1.8239195467322926, "learning_rate": 1.5569188088840977e-05, "loss": 2.6957, "step": 29500 }, { "epoch": 0.67, "grad_norm": 1.4771279880276895, "learning_rate": 1.5566154596560888e-05, "loss": 2.6937, "step": 29510 }, { "epoch": 0.67, "grad_norm": 1.530627179085994, "learning_rate": 1.5563120361970567e-05, "loss": 2.7651, "step": 29520 }, { "epoch": 0.67, "grad_norm": 1.4141835827401883, "learning_rate": 1.5560085385474658e-05, "loss": 2.7573, "step": 29530 }, { "epoch": 0.67, "grad_norm": 1.6804964784559593, "learning_rate": 1.555704966747791e-05, "loss": 2.6297, "step": 29540 }, { "epoch": 0.67, "grad_norm": 1.673311359823016, "learning_rate": 1.5554013208385178e-05, "loss": 2.6065, "step": 29550 }, { "epoch": 0.67, "grad_norm": 1.6768505418413355, "learning_rate": 1.55509760086014e-05, "loss": 2.6326, "step": 29560 }, { "epoch": 0.67, "grad_norm": 1.6136548766056142, "learning_rate": 1.5547938068531628e-05, "loss": 2.6443, "step": 29570 }, { "epoch": 0.67, "grad_norm": 1.6887181507587579, "learning_rate": 1.5544899388581004e-05, "loss": 2.6821, "step": 29580 }, { "epoch": 0.67, "grad_norm": 1.8221196419926418, "learning_rate": 1.5541859969154765e-05, "loss": 2.6719, "step": 29590 }, { "epoch": 0.67, "grad_norm": 1.8409558400761648, "learning_rate": 1.553881981065826e-05, "loss": 2.6142, "step": 29600 }, { "epoch": 0.67, "grad_norm": 1.7803106983560653, "learning_rate": 1.5535778913496923e-05, "loss": 2.4994, "step": 29610 }, { "epoch": 0.67, "grad_norm": 1.6987841528844616, "learning_rate": 1.5532737278076294e-05, "loss": 2.5951, "step": 29620 }, { "epoch": 0.67, "grad_norm": 1.564920690828256, "learning_rate": 1.5529694904802016e-05, "loss": 2.7275, "step": 29630 }, { "epoch": 0.67, "grad_norm": 1.5458572283742709, "learning_rate": 1.5526651794079813e-05, "loss": 2.6283, "step": 29640 }, { "epoch": 0.67, "grad_norm": 1.6696962474691355, "learning_rate": 1.5523607946315523e-05, "loss": 2.5212, "step": 29650 }, { "epoch": 0.67, "grad_norm": 1.6893289115137062, "learning_rate": 1.5520563361915085e-05, "loss": 2.5446, "step": 29660 }, { "epoch": 0.67, "grad_norm": 1.4530639657228615, "learning_rate": 1.551751804128452e-05, "loss": 2.6761, "step": 29670 }, { "epoch": 0.67, "grad_norm": 1.6853993283419324, "learning_rate": 1.551447198482996e-05, "loss": 2.6525, "step": 29680 }, { "epoch": 0.67, "grad_norm": 1.9018646508806776, "learning_rate": 1.551142519295763e-05, "loss": 2.6763, "step": 29690 }, { "epoch": 0.67, "grad_norm": 1.6203356048944337, "learning_rate": 1.5508377666073858e-05, "loss": 2.6343, "step": 29700 }, { "epoch": 0.67, "grad_norm": 2.1903857754751797, "learning_rate": 1.5505329404585062e-05, "loss": 2.5975, "step": 29710 }, { "epoch": 0.67, "grad_norm": 1.7774533705947932, "learning_rate": 1.5502280408897763e-05, "loss": 2.5931, "step": 29720 }, { "epoch": 0.67, "grad_norm": 1.5673149740578645, "learning_rate": 1.5499230679418586e-05, "loss": 2.5773, "step": 29730 }, { "epoch": 0.67, "grad_norm": 1.7912783405648696, "learning_rate": 1.5496180216554237e-05, "loss": 2.6147, "step": 29740 }, { "epoch": 0.67, "grad_norm": 1.6662088832163462, "learning_rate": 1.5493129020711536e-05, "loss": 2.6784, "step": 29750 }, { "epoch": 0.67, "grad_norm": 1.7139804263537939, "learning_rate": 1.5490077092297392e-05, "loss": 2.6286, "step": 29760 }, { "epoch": 0.67, "grad_norm": 1.6950611391045902, "learning_rate": 1.5487024431718817e-05, "loss": 2.5448, "step": 29770 }, { "epoch": 0.67, "grad_norm": 1.950251682519063, "learning_rate": 1.548397103938292e-05, "loss": 2.5809, "step": 29780 }, { "epoch": 0.67, "grad_norm": 1.9419056847965792, "learning_rate": 1.54809169156969e-05, "loss": 2.662, "step": 29790 }, { "epoch": 0.67, "grad_norm": 1.5728730106015794, "learning_rate": 1.5477862061068062e-05, "loss": 2.6324, "step": 29800 }, { "epoch": 0.67, "grad_norm": 1.7834264212660558, "learning_rate": 1.547480647590381e-05, "loss": 2.5853, "step": 29810 }, { "epoch": 0.67, "grad_norm": 1.7290219289977073, "learning_rate": 1.5471750160611633e-05, "loss": 2.667, "step": 29820 }, { "epoch": 0.67, "grad_norm": 1.8021766062198317, "learning_rate": 1.546869311559913e-05, "loss": 2.718, "step": 29830 }, { "epoch": 0.67, "grad_norm": 1.7314228128325275, "learning_rate": 1.5465635341273998e-05, "loss": 2.8146, "step": 29840 }, { "epoch": 0.67, "grad_norm": 1.8950340461981927, "learning_rate": 1.5462576838044014e-05, "loss": 2.6801, "step": 29850 }, { "epoch": 0.67, "grad_norm": 1.6460644624690937, "learning_rate": 1.5459517606317073e-05, "loss": 2.6875, "step": 29860 }, { "epoch": 0.67, "grad_norm": 1.6186399140474186, "learning_rate": 1.545645764650116e-05, "loss": 2.6809, "step": 29870 }, { "epoch": 0.67, "grad_norm": 1.810930397610869, "learning_rate": 1.5453396959004346e-05, "loss": 2.6553, "step": 29880 }, { "epoch": 0.67, "grad_norm": 1.8608361302872203, "learning_rate": 1.545033554423482e-05, "loss": 2.6617, "step": 29890 }, { "epoch": 0.67, "grad_norm": 1.5499750928268978, "learning_rate": 1.5447273402600848e-05, "loss": 2.7589, "step": 29900 }, { "epoch": 0.67, "grad_norm": 1.8076655187759183, "learning_rate": 1.5444210534510805e-05, "loss": 2.6461, "step": 29910 }, { "epoch": 0.67, "grad_norm": 1.4896134710479354, "learning_rate": 1.544114694037316e-05, "loss": 2.6585, "step": 29920 }, { "epoch": 0.67, "grad_norm": 1.7419241912910806, "learning_rate": 1.5438082620596474e-05, "loss": 2.579, "step": 29930 }, { "epoch": 0.68, "grad_norm": 1.7716778230185286, "learning_rate": 1.5435017575589415e-05, "loss": 2.6297, "step": 29940 }, { "epoch": 0.68, "grad_norm": 1.6773268597992539, "learning_rate": 1.5431951805760738e-05, "loss": 2.6936, "step": 29950 }, { "epoch": 0.68, "grad_norm": 1.665141074461371, "learning_rate": 1.54288853115193e-05, "loss": 2.6332, "step": 29960 }, { "epoch": 0.68, "grad_norm": 1.9140121287962544, "learning_rate": 1.5425818093274055e-05, "loss": 2.6956, "step": 29970 }, { "epoch": 0.68, "grad_norm": 1.7285891181582655, "learning_rate": 1.5422750151434045e-05, "loss": 2.6373, "step": 29980 }, { "epoch": 0.68, "grad_norm": 1.6495287682890267, "learning_rate": 1.541968148640842e-05, "loss": 2.567, "step": 29990 }, { "epoch": 0.68, "grad_norm": 1.5331486010494004, "learning_rate": 1.5416612098606423e-05, "loss": 2.731, "step": 30000 }, { "epoch": 0.68, "grad_norm": 1.983613740108673, "learning_rate": 1.541354198843739e-05, "loss": 2.642, "step": 30010 }, { "epoch": 0.68, "grad_norm": 1.7790242150841662, "learning_rate": 1.541047115631075e-05, "loss": 2.6556, "step": 30020 }, { "epoch": 0.68, "grad_norm": 1.9905547951254865, "learning_rate": 1.5407399602636042e-05, "loss": 2.5913, "step": 30030 }, { "epoch": 0.68, "grad_norm": 1.5425218701038008, "learning_rate": 1.5404327327822887e-05, "loss": 2.5945, "step": 30040 }, { "epoch": 0.68, "grad_norm": 1.4823951041366847, "learning_rate": 1.5401254332281012e-05, "loss": 2.7269, "step": 30050 }, { "epoch": 0.68, "grad_norm": 1.5816042424989074, "learning_rate": 1.539818061642023e-05, "loss": 2.5297, "step": 30060 }, { "epoch": 0.68, "grad_norm": 1.69689357637165, "learning_rate": 1.5395106180650465e-05, "loss": 2.6409, "step": 30070 }, { "epoch": 0.68, "grad_norm": 1.8397518198224914, "learning_rate": 1.539203102538172e-05, "loss": 2.5493, "step": 30080 }, { "epoch": 0.68, "grad_norm": 1.9189292874194361, "learning_rate": 1.538895515102411e-05, "loss": 2.6917, "step": 30090 }, { "epoch": 0.68, "grad_norm": 1.5714452238121763, "learning_rate": 1.5385878557987824e-05, "loss": 2.6431, "step": 30100 }, { "epoch": 0.68, "grad_norm": 1.7387762103628062, "learning_rate": 1.5382801246683174e-05, "loss": 2.5491, "step": 30110 }, { "epoch": 0.68, "grad_norm": 1.7209313685369498, "learning_rate": 1.5379723217520552e-05, "loss": 2.656, "step": 30120 }, { "epoch": 0.68, "grad_norm": 1.5325235348290491, "learning_rate": 1.5376644470910445e-05, "loss": 2.7494, "step": 30130 }, { "epoch": 0.68, "grad_norm": 1.664947955313078, "learning_rate": 1.537356500726344e-05, "loss": 2.6424, "step": 30140 }, { "epoch": 0.68, "grad_norm": 1.7304421667050696, "learning_rate": 1.537048482699022e-05, "loss": 2.5674, "step": 30150 }, { "epoch": 0.68, "grad_norm": 1.7278355399308918, "learning_rate": 1.5367403930501566e-05, "loss": 2.594, "step": 30160 }, { "epoch": 0.68, "grad_norm": 1.674926786648839, "learning_rate": 1.536432231820834e-05, "loss": 2.5448, "step": 30170 }, { "epoch": 0.68, "grad_norm": 1.98094343842693, "learning_rate": 1.5361239990521522e-05, "loss": 2.5695, "step": 30180 }, { "epoch": 0.68, "grad_norm": 1.6598096921674161, "learning_rate": 1.5358156947852166e-05, "loss": 2.5845, "step": 30190 }, { "epoch": 0.68, "grad_norm": 1.553422470244519, "learning_rate": 1.5355073190611433e-05, "loss": 2.5901, "step": 30200 }, { "epoch": 0.68, "grad_norm": 1.6120773528537344, "learning_rate": 1.535198871921058e-05, "loss": 2.6613, "step": 30210 }, { "epoch": 0.68, "grad_norm": 1.4385416268914835, "learning_rate": 1.534890353406096e-05, "loss": 2.6683, "step": 30220 }, { "epoch": 0.68, "grad_norm": 1.514654843593007, "learning_rate": 1.5345817635574012e-05, "loss": 2.6971, "step": 30230 }, { "epoch": 0.68, "grad_norm": 1.996642115860304, "learning_rate": 1.5342731024161275e-05, "loss": 2.6695, "step": 30240 }, { "epoch": 0.68, "grad_norm": 1.7341783092562533, "learning_rate": 1.5339643700234388e-05, "loss": 2.5656, "step": 30250 }, { "epoch": 0.68, "grad_norm": 1.6103352216330855, "learning_rate": 1.5336555664205075e-05, "loss": 2.5953, "step": 30260 }, { "epoch": 0.68, "grad_norm": 1.8348396800961269, "learning_rate": 1.5333466916485168e-05, "loss": 2.6896, "step": 30270 }, { "epoch": 0.68, "grad_norm": 1.5870627067122378, "learning_rate": 1.5330377457486583e-05, "loss": 2.6684, "step": 30280 }, { "epoch": 0.68, "grad_norm": 2.092715750308254, "learning_rate": 1.5327287287621332e-05, "loss": 2.6442, "step": 30290 }, { "epoch": 0.68, "grad_norm": 1.643533793937663, "learning_rate": 1.532419640730153e-05, "loss": 2.621, "step": 30300 }, { "epoch": 0.68, "grad_norm": 1.6434803741397956, "learning_rate": 1.5321104816939377e-05, "loss": 2.6447, "step": 30310 }, { "epoch": 0.68, "grad_norm": 1.5763872518311268, "learning_rate": 1.5318012516947173e-05, "loss": 2.5795, "step": 30320 }, { "epoch": 0.68, "grad_norm": 1.7079973136742548, "learning_rate": 1.5314919507737315e-05, "loss": 2.6265, "step": 30330 }, { "epoch": 0.68, "grad_norm": 1.8934618733591566, "learning_rate": 1.5311825789722288e-05, "loss": 2.6511, "step": 30340 }, { "epoch": 0.68, "grad_norm": 1.6566544524287374, "learning_rate": 1.5308731363314674e-05, "loss": 2.602, "step": 30350 }, { "epoch": 0.68, "grad_norm": 2.0631746693539714, "learning_rate": 1.5305636228927144e-05, "loss": 2.5215, "step": 30360 }, { "epoch": 0.68, "grad_norm": 1.4514260427863332, "learning_rate": 1.5302540386972485e-05, "loss": 2.6834, "step": 30370 }, { "epoch": 0.69, "grad_norm": 1.6515107486911564, "learning_rate": 1.5299443837863554e-05, "loss": 2.6121, "step": 30380 }, { "epoch": 0.69, "grad_norm": 1.8599274310830247, "learning_rate": 1.5296346582013303e-05, "loss": 2.6215, "step": 30390 }, { "epoch": 0.69, "grad_norm": 1.6368661139780913, "learning_rate": 1.52932486198348e-05, "loss": 2.7746, "step": 30400 }, { "epoch": 0.69, "grad_norm": 1.6849003399918094, "learning_rate": 1.5290149951741187e-05, "loss": 2.5712, "step": 30410 }, { "epoch": 0.69, "grad_norm": 1.570618637879382, "learning_rate": 1.5287050578145707e-05, "loss": 2.7138, "step": 30420 }, { "epoch": 0.69, "grad_norm": 1.7768136916836486, "learning_rate": 1.5283950499461703e-05, "loss": 2.5948, "step": 30430 }, { "epoch": 0.69, "grad_norm": 1.5978491322356365, "learning_rate": 1.5280849716102594e-05, "loss": 2.6107, "step": 30440 }, { "epoch": 0.69, "grad_norm": 1.841081338779279, "learning_rate": 1.5277748228481915e-05, "loss": 2.6054, "step": 30450 }, { "epoch": 0.69, "grad_norm": 1.807516717099785, "learning_rate": 1.527464603701328e-05, "loss": 2.6413, "step": 30460 }, { "epoch": 0.69, "grad_norm": 1.816247941455089, "learning_rate": 1.5271543142110403e-05, "loss": 2.6496, "step": 30470 }, { "epoch": 0.69, "grad_norm": 1.6683515041772228, "learning_rate": 1.526843954418709e-05, "loss": 2.5546, "step": 30480 }, { "epoch": 0.69, "grad_norm": 2.147022628373837, "learning_rate": 1.526533524365724e-05, "loss": 2.6906, "step": 30490 }, { "epoch": 0.69, "grad_norm": 1.8020958241400644, "learning_rate": 1.5262230240934852e-05, "loss": 2.584, "step": 30500 }, { "epoch": 0.69, "grad_norm": 1.890366142179004, "learning_rate": 1.525912453643401e-05, "loss": 2.5959, "step": 30510 }, { "epoch": 0.69, "grad_norm": 1.5991587445100628, "learning_rate": 1.5256018130568892e-05, "loss": 2.7338, "step": 30520 }, { "epoch": 0.69, "grad_norm": 1.8210705961017857, "learning_rate": 1.5252911023753776e-05, "loss": 2.696, "step": 30530 }, { "epoch": 0.69, "grad_norm": 1.8163682073251972, "learning_rate": 1.5249803216403033e-05, "loss": 2.668, "step": 30540 }, { "epoch": 0.69, "grad_norm": 2.127577408579455, "learning_rate": 1.524669470893112e-05, "loss": 2.6326, "step": 30550 }, { "epoch": 0.69, "grad_norm": 1.5779433867416577, "learning_rate": 1.5243585501752592e-05, "loss": 2.6152, "step": 30560 }, { "epoch": 0.69, "grad_norm": 4.221018652389527, "learning_rate": 1.52404755952821e-05, "loss": 2.6168, "step": 30570 }, { "epoch": 0.69, "grad_norm": 1.7563789365022817, "learning_rate": 1.5237364989934382e-05, "loss": 2.4688, "step": 30580 }, { "epoch": 0.69, "grad_norm": 1.7454083015978838, "learning_rate": 1.5234253686124279e-05, "loss": 2.7029, "step": 30590 }, { "epoch": 0.69, "grad_norm": 1.747583249742786, "learning_rate": 1.5231141684266713e-05, "loss": 2.684, "step": 30600 }, { "epoch": 0.69, "grad_norm": 1.6795642260012957, "learning_rate": 1.5228028984776705e-05, "loss": 2.6771, "step": 30610 }, { "epoch": 0.69, "grad_norm": 1.6203273478653935, "learning_rate": 1.5224915588069373e-05, "loss": 2.6686, "step": 30620 }, { "epoch": 0.69, "grad_norm": 1.6651626570442808, "learning_rate": 1.522180149455992e-05, "loss": 2.6118, "step": 30630 }, { "epoch": 0.69, "grad_norm": 1.7180138682410577, "learning_rate": 1.5218686704663648e-05, "loss": 2.6794, "step": 30640 }, { "epoch": 0.69, "grad_norm": 1.775395918955345, "learning_rate": 1.5215571218795952e-05, "loss": 2.5787, "step": 30650 }, { "epoch": 0.69, "grad_norm": 1.558527192595447, "learning_rate": 1.521245503737231e-05, "loss": 2.6562, "step": 30660 }, { "epoch": 0.69, "grad_norm": 1.7836353355148697, "learning_rate": 1.5209338160808311e-05, "loss": 2.6355, "step": 30670 }, { "epoch": 0.69, "grad_norm": 1.6067183928645672, "learning_rate": 1.5206220589519619e-05, "loss": 2.5723, "step": 30680 }, { "epoch": 0.69, "grad_norm": 1.582837546875691, "learning_rate": 1.5203102323921996e-05, "loss": 2.7044, "step": 30690 }, { "epoch": 0.69, "grad_norm": 1.655771772699115, "learning_rate": 1.5199983364431306e-05, "loss": 2.6322, "step": 30700 }, { "epoch": 0.69, "grad_norm": 1.6352842964534986, "learning_rate": 1.519686371146349e-05, "loss": 2.6041, "step": 30710 }, { "epoch": 0.69, "grad_norm": 1.5684839986772954, "learning_rate": 1.5193743365434595e-05, "loss": 2.5935, "step": 30720 }, { "epoch": 0.69, "grad_norm": 1.6276737391812146, "learning_rate": 1.5190622326760749e-05, "loss": 2.6323, "step": 30730 }, { "epoch": 0.69, "grad_norm": 1.5430037341944234, "learning_rate": 1.5187500595858183e-05, "loss": 2.6494, "step": 30740 }, { "epoch": 0.69, "grad_norm": 2.8066240764694474, "learning_rate": 1.5184378173143212e-05, "loss": 2.6565, "step": 30750 }, { "epoch": 0.69, "grad_norm": 1.531174248295073, "learning_rate": 1.5181255059032252e-05, "loss": 2.5869, "step": 30760 }, { "epoch": 0.69, "grad_norm": 1.6519782285185396, "learning_rate": 1.5178131253941803e-05, "loss": 2.613, "step": 30770 }, { "epoch": 0.69, "grad_norm": 1.7802097637871834, "learning_rate": 1.5175006758288458e-05, "loss": 2.6274, "step": 30780 }, { "epoch": 0.69, "grad_norm": 1.533134980438779, "learning_rate": 1.5171881572488904e-05, "loss": 2.644, "step": 30790 }, { "epoch": 0.69, "grad_norm": 1.4965040223947166, "learning_rate": 1.5168755696959925e-05, "loss": 2.6339, "step": 30800 }, { "epoch": 0.69, "grad_norm": 1.7144404100691963, "learning_rate": 1.5165629132118387e-05, "loss": 2.6027, "step": 30810 }, { "epoch": 0.7, "grad_norm": 1.6224142786239961, "learning_rate": 1.5162501878381257e-05, "loss": 2.6383, "step": 30820 }, { "epoch": 0.7, "grad_norm": 1.6860535605386384, "learning_rate": 1.5159373936165586e-05, "loss": 2.6203, "step": 30830 }, { "epoch": 0.7, "grad_norm": 1.551645453837549, "learning_rate": 1.5156245305888524e-05, "loss": 2.6406, "step": 30840 }, { "epoch": 0.7, "grad_norm": 1.6312245834909562, "learning_rate": 1.515311598796731e-05, "loss": 2.6585, "step": 30850 }, { "epoch": 0.7, "grad_norm": 1.871358951287494, "learning_rate": 1.5149985982819269e-05, "loss": 2.6881, "step": 30860 }, { "epoch": 0.7, "grad_norm": 1.8862447633120032, "learning_rate": 1.514685529086183e-05, "loss": 2.6997, "step": 30870 }, { "epoch": 0.7, "grad_norm": 1.5668931304378864, "learning_rate": 1.5143723912512504e-05, "loss": 2.6594, "step": 30880 }, { "epoch": 0.7, "grad_norm": 1.7129367361945782, "learning_rate": 1.5140591848188892e-05, "loss": 2.6376, "step": 30890 }, { "epoch": 0.7, "grad_norm": 1.7066007288121854, "learning_rate": 1.5137459098308695e-05, "loss": 2.5943, "step": 30900 }, { "epoch": 0.7, "grad_norm": 1.6401541046360306, "learning_rate": 1.51343256632897e-05, "loss": 2.6592, "step": 30910 }, { "epoch": 0.7, "grad_norm": 1.5656537060564342, "learning_rate": 1.5131191543549786e-05, "loss": 2.5785, "step": 30920 }, { "epoch": 0.7, "grad_norm": 1.830398465768271, "learning_rate": 1.5128056739506924e-05, "loss": 2.5308, "step": 30930 }, { "epoch": 0.7, "grad_norm": 1.7248483092730498, "learning_rate": 1.5124921251579174e-05, "loss": 2.6241, "step": 30940 }, { "epoch": 0.7, "grad_norm": 1.6711352831782842, "learning_rate": 1.5121785080184693e-05, "loss": 2.5963, "step": 30950 }, { "epoch": 0.7, "grad_norm": 1.6115122380950437, "learning_rate": 1.5118648225741722e-05, "loss": 2.579, "step": 30960 }, { "epoch": 0.7, "grad_norm": 1.493302243918812, "learning_rate": 1.5115510688668598e-05, "loss": 2.5925, "step": 30970 }, { "epoch": 0.7, "grad_norm": 1.8618222457684839, "learning_rate": 1.5112372469383744e-05, "loss": 2.6438, "step": 30980 }, { "epoch": 0.7, "grad_norm": 1.886686602268506, "learning_rate": 1.510923356830568e-05, "loss": 2.6611, "step": 30990 }, { "epoch": 0.7, "grad_norm": 1.5911841443858874, "learning_rate": 1.5106093985853014e-05, "loss": 2.7509, "step": 31000 }, { "epoch": 0.7, "grad_norm": 2.00649281986536, "learning_rate": 1.5102953722444449e-05, "loss": 2.7661, "step": 31010 }, { "epoch": 0.7, "grad_norm": 1.860489804228584, "learning_rate": 1.5099812778498768e-05, "loss": 2.694, "step": 31020 }, { "epoch": 0.7, "grad_norm": 1.8180767788518157, "learning_rate": 1.5096671154434858e-05, "loss": 2.6322, "step": 31030 }, { "epoch": 0.7, "grad_norm": 1.6638208450299274, "learning_rate": 1.5093528850671686e-05, "loss": 2.6699, "step": 31040 }, { "epoch": 0.7, "grad_norm": 1.759540339837325, "learning_rate": 1.5090385867628317e-05, "loss": 2.5972, "step": 31050 }, { "epoch": 0.7, "grad_norm": 1.678946632419876, "learning_rate": 1.5087242205723898e-05, "loss": 2.7392, "step": 31060 }, { "epoch": 0.7, "grad_norm": 2.1672767909531045, "learning_rate": 1.5084097865377679e-05, "loss": 2.676, "step": 31070 }, { "epoch": 0.7, "grad_norm": 1.7215643998328374, "learning_rate": 1.5080952847008994e-05, "loss": 2.5735, "step": 31080 }, { "epoch": 0.7, "grad_norm": 1.7799332276630717, "learning_rate": 1.5077807151037258e-05, "loss": 2.6993, "step": 31090 }, { "epoch": 0.7, "grad_norm": 1.8707723482271224, "learning_rate": 1.5074660777881996e-05, "loss": 2.5735, "step": 31100 }, { "epoch": 0.7, "grad_norm": 1.54628820324489, "learning_rate": 1.5071513727962805e-05, "loss": 2.6718, "step": 31110 }, { "epoch": 0.7, "grad_norm": 1.7448176638894293, "learning_rate": 1.5068366001699385e-05, "loss": 2.6741, "step": 31120 }, { "epoch": 0.7, "grad_norm": 1.7730583388591734, "learning_rate": 1.5065217599511522e-05, "loss": 2.7137, "step": 31130 }, { "epoch": 0.7, "grad_norm": 2.3674016552305934, "learning_rate": 1.5062068521819086e-05, "loss": 2.6662, "step": 31140 }, { "epoch": 0.7, "grad_norm": 1.6710932326456582, "learning_rate": 1.5058918769042046e-05, "loss": 2.6403, "step": 31150 }, { "epoch": 0.7, "grad_norm": 1.6188561885597084, "learning_rate": 1.5055768341600456e-05, "loss": 2.5092, "step": 31160 }, { "epoch": 0.7, "grad_norm": 1.6000391891782124, "learning_rate": 1.5052617239914464e-05, "loss": 2.7297, "step": 31170 }, { "epoch": 0.7, "grad_norm": 1.6278378902585149, "learning_rate": 1.5049465464404302e-05, "loss": 2.5769, "step": 31180 }, { "epoch": 0.7, "grad_norm": 1.499330164546318, "learning_rate": 1.5046313015490302e-05, "loss": 2.7348, "step": 31190 }, { "epoch": 0.7, "grad_norm": 1.6563552112184037, "learning_rate": 1.5043159893592867e-05, "loss": 2.724, "step": 31200 }, { "epoch": 0.7, "grad_norm": 1.739301236501938, "learning_rate": 1.5040006099132514e-05, "loss": 2.6236, "step": 31210 }, { "epoch": 0.7, "grad_norm": 1.5287799149660535, "learning_rate": 1.503685163252983e-05, "loss": 2.6428, "step": 31220 }, { "epoch": 0.7, "grad_norm": 1.9974447377810025, "learning_rate": 1.50336964942055e-05, "loss": 2.6963, "step": 31230 }, { "epoch": 0.7, "grad_norm": 1.6967856388757445, "learning_rate": 1.50305406845803e-05, "loss": 2.6665, "step": 31240 }, { "epoch": 0.7, "grad_norm": 1.7697070932956847, "learning_rate": 1.5027384204075096e-05, "loss": 2.729, "step": 31250 }, { "epoch": 0.7, "grad_norm": 1.595222436054869, "learning_rate": 1.502422705311083e-05, "loss": 2.8175, "step": 31260 }, { "epoch": 0.71, "grad_norm": 1.6280607795523125, "learning_rate": 1.5021069232108559e-05, "loss": 2.813, "step": 31270 }, { "epoch": 0.71, "grad_norm": 1.6809877053200661, "learning_rate": 1.5017910741489401e-05, "loss": 2.6014, "step": 31280 }, { "epoch": 0.71, "grad_norm": 1.7186635500531051, "learning_rate": 1.5014751581674584e-05, "loss": 2.6356, "step": 31290 }, { "epoch": 0.71, "grad_norm": 1.7458271837396757, "learning_rate": 1.5011591753085415e-05, "loss": 2.7012, "step": 31300 }, { "epoch": 0.71, "grad_norm": 1.6138688005261776, "learning_rate": 1.5008431256143297e-05, "loss": 2.6102, "step": 31310 }, { "epoch": 0.71, "grad_norm": 2.0542177520032197, "learning_rate": 1.5005270091269714e-05, "loss": 2.5968, "step": 31320 }, { "epoch": 0.71, "grad_norm": 1.7264643364105634, "learning_rate": 1.5002108258886245e-05, "loss": 2.5227, "step": 31330 }, { "epoch": 0.71, "grad_norm": 1.7063292696030756, "learning_rate": 1.4998945759414558e-05, "loss": 2.5579, "step": 31340 }, { "epoch": 0.71, "grad_norm": 1.9251669890439769, "learning_rate": 1.4995782593276408e-05, "loss": 2.7248, "step": 31350 }, { "epoch": 0.71, "grad_norm": 1.604908167332338, "learning_rate": 1.4992618760893636e-05, "loss": 2.6369, "step": 31360 }, { "epoch": 0.71, "grad_norm": 1.7119926454582244, "learning_rate": 1.4989454262688178e-05, "loss": 2.5342, "step": 31370 }, { "epoch": 0.71, "grad_norm": 1.7033127025542778, "learning_rate": 1.498628909908206e-05, "loss": 2.6903, "step": 31380 }, { "epoch": 0.71, "grad_norm": 1.6613723165420702, "learning_rate": 1.498312327049738e-05, "loss": 2.6767, "step": 31390 }, { "epoch": 0.71, "grad_norm": 1.7808439064445383, "learning_rate": 1.4979956777356353e-05, "loss": 2.6408, "step": 31400 }, { "epoch": 0.71, "grad_norm": 1.9833768330353003, "learning_rate": 1.4976789620081255e-05, "loss": 2.6341, "step": 31410 }, { "epoch": 0.71, "grad_norm": 1.9325361921019806, "learning_rate": 1.4973621799094468e-05, "loss": 2.7401, "step": 31420 }, { "epoch": 0.71, "grad_norm": 1.5897123276076521, "learning_rate": 1.4970453314818457e-05, "loss": 2.7436, "step": 31430 }, { "epoch": 0.71, "grad_norm": 2.1368387924534176, "learning_rate": 1.4967284167675775e-05, "loss": 2.7332, "step": 31440 }, { "epoch": 0.71, "grad_norm": 1.509893207458225, "learning_rate": 1.4964114358089063e-05, "loss": 2.7058, "step": 31450 }, { "epoch": 0.71, "grad_norm": 1.7365808432659458, "learning_rate": 1.4960943886481052e-05, "loss": 2.6483, "step": 31460 }, { "epoch": 0.71, "grad_norm": 1.866923954708618, "learning_rate": 1.4957772753274561e-05, "loss": 2.5588, "step": 31470 }, { "epoch": 0.71, "grad_norm": 1.9640038910865587, "learning_rate": 1.4954600958892499e-05, "loss": 2.6023, "step": 31480 }, { "epoch": 0.71, "grad_norm": 1.5673179565356274, "learning_rate": 1.4951428503757852e-05, "loss": 2.6022, "step": 31490 }, { "epoch": 0.71, "grad_norm": 1.7745832573442715, "learning_rate": 1.4948255388293714e-05, "loss": 2.6385, "step": 31500 }, { "epoch": 0.71, "grad_norm": 1.6952752008054965, "learning_rate": 1.494508161292325e-05, "loss": 2.6512, "step": 31510 }, { "epoch": 0.71, "grad_norm": 1.5552987833177592, "learning_rate": 1.4941907178069721e-05, "loss": 2.6791, "step": 31520 }, { "epoch": 0.71, "grad_norm": 1.7881391425026951, "learning_rate": 1.4938732084156475e-05, "loss": 2.7286, "step": 31530 }, { "epoch": 0.71, "grad_norm": 1.538469487097419, "learning_rate": 1.4935556331606944e-05, "loss": 2.5612, "step": 31540 }, { "epoch": 0.71, "grad_norm": 1.8755986918127867, "learning_rate": 1.4932379920844656e-05, "loss": 2.6878, "step": 31550 }, { "epoch": 0.71, "grad_norm": 1.774472640293821, "learning_rate": 1.4929202852293215e-05, "loss": 2.6611, "step": 31560 }, { "epoch": 0.71, "grad_norm": 1.4639472553003912, "learning_rate": 1.4926025126376324e-05, "loss": 2.672, "step": 31570 }, { "epoch": 0.71, "grad_norm": 1.5336491755276513, "learning_rate": 1.4922846743517767e-05, "loss": 2.6678, "step": 31580 }, { "epoch": 0.71, "grad_norm": 1.8929230474172636, "learning_rate": 1.4919667704141417e-05, "loss": 2.6235, "step": 31590 }, { "epoch": 0.71, "grad_norm": 1.9164229812255604, "learning_rate": 1.491648800867124e-05, "loss": 2.5519, "step": 31600 }, { "epoch": 0.71, "grad_norm": 1.8630961538360449, "learning_rate": 1.4913307657531278e-05, "loss": 2.639, "step": 31610 }, { "epoch": 0.71, "grad_norm": 1.6280512240610563, "learning_rate": 1.4910126651145671e-05, "loss": 2.609, "step": 31620 }, { "epoch": 0.71, "grad_norm": 1.8697363850785165, "learning_rate": 1.4906944989938644e-05, "loss": 2.7288, "step": 31630 }, { "epoch": 0.71, "grad_norm": 1.623819004269523, "learning_rate": 1.4903762674334504e-05, "loss": 2.7072, "step": 31640 }, { "epoch": 0.71, "grad_norm": 1.6795492302749275, "learning_rate": 1.490057970475765e-05, "loss": 2.658, "step": 31650 }, { "epoch": 0.71, "grad_norm": 2.125720549375477, "learning_rate": 1.4897396081632566e-05, "loss": 2.6213, "step": 31660 }, { "epoch": 0.71, "grad_norm": 1.5485541272215855, "learning_rate": 1.4894211805383833e-05, "loss": 2.7055, "step": 31670 }, { "epoch": 0.71, "grad_norm": 2.0531224090574502, "learning_rate": 1.4891026876436102e-05, "loss": 2.6727, "step": 31680 }, { "epoch": 0.71, "grad_norm": 1.6068065957768036, "learning_rate": 1.4887841295214117e-05, "loss": 2.605, "step": 31690 }, { "epoch": 0.71, "grad_norm": 1.5215648162132136, "learning_rate": 1.4884655062142718e-05, "loss": 2.586, "step": 31700 }, { "epoch": 0.72, "grad_norm": 1.9432497396462425, "learning_rate": 1.4881468177646828e-05, "loss": 2.7017, "step": 31710 }, { "epoch": 0.72, "grad_norm": 1.9766309478841202, "learning_rate": 1.4878280642151445e-05, "loss": 2.5933, "step": 31720 }, { "epoch": 0.72, "grad_norm": 1.5678772981381828, "learning_rate": 1.4875092456081672e-05, "loss": 2.7375, "step": 31730 }, { "epoch": 0.72, "grad_norm": 1.8858727064656042, "learning_rate": 1.4871903619862686e-05, "loss": 2.6183, "step": 31740 }, { "epoch": 0.72, "grad_norm": 1.5529649436003512, "learning_rate": 1.4868714133919755e-05, "loss": 2.7488, "step": 31750 }, { "epoch": 0.72, "grad_norm": 1.747728600097973, "learning_rate": 1.4865523998678231e-05, "loss": 2.6741, "step": 31760 }, { "epoch": 0.72, "grad_norm": 1.7891519135201923, "learning_rate": 1.486233321456356e-05, "loss": 2.7213, "step": 31770 }, { "epoch": 0.72, "grad_norm": 1.7031579276567976, "learning_rate": 1.485914178200127e-05, "loss": 2.5312, "step": 31780 }, { "epoch": 0.72, "grad_norm": 1.8410905759109968, "learning_rate": 1.4855949701416966e-05, "loss": 2.716, "step": 31790 }, { "epoch": 0.72, "grad_norm": 1.716480348821857, "learning_rate": 1.4852756973236358e-05, "loss": 2.6854, "step": 31800 }, { "epoch": 0.72, "grad_norm": 1.5700419028473362, "learning_rate": 1.4849563597885226e-05, "loss": 2.6739, "step": 31810 }, { "epoch": 0.72, "grad_norm": 1.8873859260172476, "learning_rate": 1.484636957578945e-05, "loss": 2.6, "step": 31820 }, { "epoch": 0.72, "grad_norm": 1.9028789431183273, "learning_rate": 1.4843174907374983e-05, "loss": 2.6223, "step": 31830 }, { "epoch": 0.72, "grad_norm": 1.7218700726196334, "learning_rate": 1.4839979593067873e-05, "loss": 2.7548, "step": 31840 }, { "epoch": 0.72, "grad_norm": 1.5634842386037435, "learning_rate": 1.4836783633294251e-05, "loss": 2.6383, "step": 31850 }, { "epoch": 0.72, "grad_norm": 1.925366309300092, "learning_rate": 1.4833587028480337e-05, "loss": 2.5661, "step": 31860 }, { "epoch": 0.72, "grad_norm": 1.7310431087692333, "learning_rate": 1.483038977905243e-05, "loss": 2.604, "step": 31870 }, { "epoch": 0.72, "grad_norm": 1.644858488488186, "learning_rate": 1.4827191885436924e-05, "loss": 2.6123, "step": 31880 }, { "epoch": 0.72, "grad_norm": 1.829913604198183, "learning_rate": 1.4823993348060295e-05, "loss": 2.6011, "step": 31890 }, { "epoch": 0.72, "grad_norm": 1.6842675263424693, "learning_rate": 1.48207941673491e-05, "loss": 2.5509, "step": 31900 }, { "epoch": 0.72, "grad_norm": 1.6122624957182388, "learning_rate": 1.481759434372999e-05, "loss": 2.7553, "step": 31910 }, { "epoch": 0.72, "grad_norm": 1.9233136005805997, "learning_rate": 1.4814393877629698e-05, "loss": 2.6864, "step": 31920 }, { "epoch": 0.72, "grad_norm": 1.7508535121074755, "learning_rate": 1.481119276947504e-05, "loss": 2.5791, "step": 31930 }, { "epoch": 0.72, "grad_norm": 1.7735536918777848, "learning_rate": 1.4807991019692925e-05, "loss": 2.7053, "step": 31940 }, { "epoch": 0.72, "grad_norm": 1.7772185020721232, "learning_rate": 1.4804788628710335e-05, "loss": 2.7586, "step": 31950 }, { "epoch": 0.72, "grad_norm": 1.5551312461041855, "learning_rate": 1.4801585596954353e-05, "loss": 2.6808, "step": 31960 }, { "epoch": 0.72, "grad_norm": 1.570731514582289, "learning_rate": 1.4798381924852136e-05, "loss": 2.5737, "step": 31970 }, { "epoch": 0.72, "grad_norm": 1.7879511800389563, "learning_rate": 1.4795177612830933e-05, "loss": 2.7316, "step": 31980 }, { "epoch": 0.72, "grad_norm": 1.7684789048381002, "learning_rate": 1.479197266131807e-05, "loss": 2.6952, "step": 31990 }, { "epoch": 0.72, "grad_norm": 1.7162539454933903, "learning_rate": 1.4788767070740974e-05, "loss": 2.6043, "step": 32000 }, { "epoch": 0.72, "grad_norm": 1.950351818322781, "learning_rate": 1.4785560841527138e-05, "loss": 2.5639, "step": 32010 }, { "epoch": 0.72, "grad_norm": 1.588376853117495, "learning_rate": 1.4782353974104148e-05, "loss": 2.6273, "step": 32020 }, { "epoch": 0.72, "grad_norm": 1.7479047932234697, "learning_rate": 1.4779146468899685e-05, "loss": 2.5556, "step": 32030 }, { "epoch": 0.72, "grad_norm": 1.7502143660485938, "learning_rate": 1.4775938326341499e-05, "loss": 2.6738, "step": 32040 }, { "epoch": 0.72, "grad_norm": 1.584487386268952, "learning_rate": 1.4772729546857438e-05, "loss": 2.6435, "step": 32050 }, { "epoch": 0.72, "grad_norm": 1.7129998993927296, "learning_rate": 1.4769520130875426e-05, "loss": 2.6352, "step": 32060 }, { "epoch": 0.72, "grad_norm": 1.6345693912937864, "learning_rate": 1.4766310078823475e-05, "loss": 2.6161, "step": 32070 }, { "epoch": 0.72, "grad_norm": 1.7110494650609407, "learning_rate": 1.4763099391129683e-05, "loss": 2.6481, "step": 32080 }, { "epoch": 0.72, "grad_norm": 1.6098440673421837, "learning_rate": 1.475988806822223e-05, "loss": 2.6674, "step": 32090 }, { "epoch": 0.72, "grad_norm": 1.7623353568717197, "learning_rate": 1.4756676110529387e-05, "loss": 2.6592, "step": 32100 }, { "epoch": 0.72, "grad_norm": 1.7896372734782524, "learning_rate": 1.4753463518479503e-05, "loss": 2.6457, "step": 32110 }, { "epoch": 0.72, "grad_norm": 1.8736117916020136, "learning_rate": 1.4750250292501009e-05, "loss": 2.6961, "step": 32120 }, { "epoch": 0.72, "grad_norm": 1.6169406093215493, "learning_rate": 1.4747036433022434e-05, "loss": 2.7509, "step": 32130 }, { "epoch": 0.72, "grad_norm": 1.6387914620966233, "learning_rate": 1.4743821940472376e-05, "loss": 2.6495, "step": 32140 }, { "epoch": 0.73, "grad_norm": 2.391232204184153, "learning_rate": 1.4740606815279527e-05, "loss": 2.6525, "step": 32150 }, { "epoch": 0.73, "grad_norm": 1.9276369748275624, "learning_rate": 1.4737391057872663e-05, "loss": 2.6273, "step": 32160 }, { "epoch": 0.73, "grad_norm": 1.525342372277814, "learning_rate": 1.4734174668680639e-05, "loss": 2.6924, "step": 32170 }, { "epoch": 0.73, "grad_norm": 1.627672207296931, "learning_rate": 1.4730957648132395e-05, "loss": 2.6634, "step": 32180 }, { "epoch": 0.73, "grad_norm": 1.6676734715293313, "learning_rate": 1.472773999665696e-05, "loss": 2.6664, "step": 32190 }, { "epoch": 0.73, "grad_norm": 1.6051521153986958, "learning_rate": 1.4724521714683446e-05, "loss": 2.7163, "step": 32200 }, { "epoch": 0.73, "grad_norm": 1.6055372921226296, "learning_rate": 1.472130280264105e-05, "loss": 2.5759, "step": 32210 }, { "epoch": 0.73, "grad_norm": 1.7967976772491885, "learning_rate": 1.4718083260959039e-05, "loss": 2.5758, "step": 32220 }, { "epoch": 0.73, "grad_norm": 1.565831741324795, "learning_rate": 1.4714863090066786e-05, "loss": 2.7047, "step": 32230 }, { "epoch": 0.73, "grad_norm": 1.7403460671341302, "learning_rate": 1.4711642290393738e-05, "loss": 2.543, "step": 32240 }, { "epoch": 0.73, "grad_norm": 1.648775185675745, "learning_rate": 1.4708420862369418e-05, "loss": 2.6392, "step": 32250 }, { "epoch": 0.73, "grad_norm": 1.9652064414775179, "learning_rate": 1.4705198806423446e-05, "loss": 2.636, "step": 32260 }, { "epoch": 0.73, "grad_norm": 1.588312584239321, "learning_rate": 1.4701976122985522e-05, "loss": 2.6281, "step": 32270 }, { "epoch": 0.73, "grad_norm": 2.209754781399736, "learning_rate": 1.4698752812485422e-05, "loss": 2.6139, "step": 32280 }, { "epoch": 0.73, "grad_norm": 1.452299019711653, "learning_rate": 1.4695528875353011e-05, "loss": 2.6326, "step": 32290 }, { "epoch": 0.73, "grad_norm": 1.7053062264604504, "learning_rate": 1.4692304312018241e-05, "loss": 2.6811, "step": 32300 }, { "epoch": 0.73, "grad_norm": 1.709825182826489, "learning_rate": 1.4689079122911146e-05, "loss": 2.7206, "step": 32310 }, { "epoch": 0.73, "grad_norm": 2.037008039923484, "learning_rate": 1.4685853308461838e-05, "loss": 2.6175, "step": 32320 }, { "epoch": 0.73, "grad_norm": 1.8451179081501063, "learning_rate": 1.4682626869100519e-05, "loss": 2.5978, "step": 32330 }, { "epoch": 0.73, "grad_norm": 1.5673352211387774, "learning_rate": 1.4679399805257466e-05, "loss": 2.7529, "step": 32340 }, { "epoch": 0.73, "grad_norm": 1.488890702251891, "learning_rate": 1.467617211736305e-05, "loss": 2.564, "step": 32350 }, { "epoch": 0.73, "grad_norm": 1.73752725160353, "learning_rate": 1.4672943805847717e-05, "loss": 2.7469, "step": 32360 }, { "epoch": 0.73, "grad_norm": 1.6392540971530554, "learning_rate": 1.4669714871142005e-05, "loss": 2.7568, "step": 32370 }, { "epoch": 0.73, "grad_norm": 1.849506691791111, "learning_rate": 1.4666485313676524e-05, "loss": 2.7897, "step": 32380 }, { "epoch": 0.73, "grad_norm": 1.5826450610146403, "learning_rate": 1.4663255133881973e-05, "loss": 2.7892, "step": 32390 }, { "epoch": 0.73, "grad_norm": 1.4734491419105518, "learning_rate": 1.4660024332189133e-05, "loss": 2.5782, "step": 32400 }, { "epoch": 0.73, "grad_norm": 1.9199282824852193, "learning_rate": 1.4656792909028869e-05, "loss": 2.6461, "step": 32410 }, { "epoch": 0.73, "grad_norm": 1.6773595841620674, "learning_rate": 1.4653560864832129e-05, "loss": 2.695, "step": 32420 }, { "epoch": 0.73, "grad_norm": 1.593280408674265, "learning_rate": 1.4650328200029943e-05, "loss": 2.5782, "step": 32430 }, { "epoch": 0.73, "grad_norm": 1.8360893053113934, "learning_rate": 1.4647094915053424e-05, "loss": 2.6292, "step": 32440 }, { "epoch": 0.73, "grad_norm": 1.6082463108776788, "learning_rate": 1.4643861010333762e-05, "loss": 2.6243, "step": 32450 }, { "epoch": 0.73, "grad_norm": 1.4514448005079392, "learning_rate": 1.4640626486302243e-05, "loss": 2.6013, "step": 32460 }, { "epoch": 0.73, "grad_norm": 1.4885694498387034, "learning_rate": 1.4637391343390223e-05, "loss": 2.6902, "step": 32470 }, { "epoch": 0.73, "grad_norm": 1.8201349416389292, "learning_rate": 1.4634155582029149e-05, "loss": 2.5909, "step": 32480 }, { "epoch": 0.73, "grad_norm": 1.7758353177969735, "learning_rate": 1.4630919202650542e-05, "loss": 2.5318, "step": 32490 }, { "epoch": 0.73, "grad_norm": 1.6817773027907592, "learning_rate": 1.4627682205686013e-05, "loss": 2.7053, "step": 32500 }, { "epoch": 0.73, "grad_norm": 1.5710059488039916, "learning_rate": 1.462444459156725e-05, "loss": 2.6201, "step": 32510 }, { "epoch": 0.73, "grad_norm": 1.7652583012049476, "learning_rate": 1.4621206360726028e-05, "loss": 2.5105, "step": 32520 }, { "epoch": 0.73, "grad_norm": 1.630596979049906, "learning_rate": 1.4617967513594202e-05, "loss": 2.624, "step": 32530 }, { "epoch": 0.73, "grad_norm": 1.6733657054908413, "learning_rate": 1.4614728050603712e-05, "loss": 2.6816, "step": 32540 }, { "epoch": 0.73, "grad_norm": 1.6004545787676976, "learning_rate": 1.4611487972186571e-05, "loss": 2.6185, "step": 32550 }, { "epoch": 0.73, "grad_norm": 1.6620381512738371, "learning_rate": 1.4608247278774883e-05, "loss": 2.569, "step": 32560 }, { "epoch": 0.73, "grad_norm": 1.6156109949574164, "learning_rate": 1.4605005970800834e-05, "loss": 2.6974, "step": 32570 }, { "epoch": 0.73, "grad_norm": 1.6539955072753636, "learning_rate": 1.4601764048696688e-05, "loss": 2.724, "step": 32580 }, { "epoch": 0.73, "grad_norm": 1.711219293433376, "learning_rate": 1.459852151289479e-05, "loss": 2.5696, "step": 32590 }, { "epoch": 0.74, "grad_norm": 1.7038036799350755, "learning_rate": 1.4595278363827572e-05, "loss": 2.6916, "step": 32600 }, { "epoch": 0.74, "grad_norm": 1.4636190316883888, "learning_rate": 1.4592034601927543e-05, "loss": 2.5991, "step": 32610 }, { "epoch": 0.74, "grad_norm": 1.419933873886794, "learning_rate": 1.4588790227627297e-05, "loss": 2.7343, "step": 32620 }, { "epoch": 0.74, "grad_norm": 1.5633131232299107, "learning_rate": 1.4585545241359507e-05, "loss": 2.6061, "step": 32630 }, { "epoch": 0.74, "grad_norm": 1.7275168581586096, "learning_rate": 1.4582299643556929e-05, "loss": 2.6712, "step": 32640 }, { "epoch": 0.74, "grad_norm": 1.7797325160654902, "learning_rate": 1.4579053434652404e-05, "loss": 2.7197, "step": 32650 }, { "epoch": 0.74, "grad_norm": 1.8663264572052742, "learning_rate": 1.4575806615078846e-05, "loss": 2.5292, "step": 32660 }, { "epoch": 0.74, "grad_norm": 1.4351525341415485, "learning_rate": 1.4572559185269261e-05, "loss": 2.5571, "step": 32670 }, { "epoch": 0.74, "grad_norm": 1.6450668194027946, "learning_rate": 1.4569311145656725e-05, "loss": 2.603, "step": 32680 }, { "epoch": 0.74, "grad_norm": 1.6207110643452862, "learning_rate": 1.4566062496674405e-05, "loss": 2.6479, "step": 32690 }, { "epoch": 0.74, "grad_norm": 1.5979961543530603, "learning_rate": 1.4562813238755545e-05, "loss": 2.4942, "step": 32700 }, { "epoch": 0.74, "grad_norm": 1.472173990228211, "learning_rate": 1.4559563372333467e-05, "loss": 2.6104, "step": 32710 }, { "epoch": 0.74, "grad_norm": 1.6378140110525197, "learning_rate": 1.4556312897841581e-05, "loss": 2.6348, "step": 32720 }, { "epoch": 0.74, "grad_norm": 1.5897543553908529, "learning_rate": 1.4553061815713377e-05, "loss": 2.7006, "step": 32730 }, { "epoch": 0.74, "grad_norm": 1.5724887780726011, "learning_rate": 1.4549810126382418e-05, "loss": 2.7205, "step": 32740 }, { "epoch": 0.74, "grad_norm": 1.5023877114249509, "learning_rate": 1.4546557830282359e-05, "loss": 2.6265, "step": 32750 }, { "epoch": 0.74, "grad_norm": 1.7174753927629691, "learning_rate": 1.454330492784693e-05, "loss": 2.6585, "step": 32760 }, { "epoch": 0.74, "grad_norm": 1.7070713808984632, "learning_rate": 1.4540051419509943e-05, "loss": 2.6278, "step": 32770 }, { "epoch": 0.74, "grad_norm": 1.7154786042676153, "learning_rate": 1.4536797305705283e-05, "loss": 2.5909, "step": 32780 }, { "epoch": 0.74, "grad_norm": 1.567190749840816, "learning_rate": 1.4533542586866934e-05, "loss": 2.5807, "step": 32790 }, { "epoch": 0.74, "grad_norm": 1.7103283668188674, "learning_rate": 1.4530287263428944e-05, "loss": 2.6855, "step": 32800 }, { "epoch": 0.74, "grad_norm": 1.6433328912338676, "learning_rate": 1.4527031335825453e-05, "loss": 2.687, "step": 32810 }, { "epoch": 0.74, "grad_norm": 1.6074891690168598, "learning_rate": 1.4523774804490665e-05, "loss": 2.6636, "step": 32820 }, { "epoch": 0.74, "grad_norm": 1.8259251478979421, "learning_rate": 1.4520517669858885e-05, "loss": 2.5401, "step": 32830 }, { "epoch": 0.74, "grad_norm": 1.4735874295383664, "learning_rate": 1.4517259932364487e-05, "loss": 2.498, "step": 32840 }, { "epoch": 0.74, "grad_norm": 1.5891459038101132, "learning_rate": 1.451400159244193e-05, "loss": 2.5603, "step": 32850 }, { "epoch": 0.74, "grad_norm": 1.7088896181168212, "learning_rate": 1.4510742650525745e-05, "loss": 2.5775, "step": 32860 }, { "epoch": 0.74, "grad_norm": 2.4477854753365564, "learning_rate": 1.4507483107050551e-05, "loss": 2.6281, "step": 32870 }, { "epoch": 0.74, "grad_norm": 1.711131906857718, "learning_rate": 1.4504222962451047e-05, "loss": 2.6413, "step": 32880 }, { "epoch": 0.74, "grad_norm": 2.0322029869290272, "learning_rate": 1.4500962217162009e-05, "loss": 2.7224, "step": 32890 }, { "epoch": 0.74, "grad_norm": 1.7971213569716233, "learning_rate": 1.4497700871618294e-05, "loss": 2.5803, "step": 32900 }, { "epoch": 0.74, "grad_norm": 1.8180802880707743, "learning_rate": 1.4494438926254846e-05, "loss": 2.7033, "step": 32910 }, { "epoch": 0.74, "grad_norm": 1.8813650023470265, "learning_rate": 1.4491176381506672e-05, "loss": 2.5729, "step": 32920 }, { "epoch": 0.74, "grad_norm": 1.67176861903026, "learning_rate": 1.4487913237808879e-05, "loss": 2.6161, "step": 32930 }, { "epoch": 0.74, "grad_norm": 1.774749110558396, "learning_rate": 1.448464949559664e-05, "loss": 2.6079, "step": 32940 }, { "epoch": 0.74, "grad_norm": 1.9150331275064187, "learning_rate": 1.4481385155305213e-05, "loss": 2.671, "step": 32950 }, { "epoch": 0.74, "grad_norm": 1.694508093802192, "learning_rate": 1.4478120217369936e-05, "loss": 2.4599, "step": 32960 }, { "epoch": 0.74, "grad_norm": 1.7829856191302094, "learning_rate": 1.4474854682226225e-05, "loss": 2.5755, "step": 32970 }, { "epoch": 0.74, "grad_norm": 1.7070324789185018, "learning_rate": 1.4471588550309572e-05, "loss": 2.516, "step": 32980 }, { "epoch": 0.74, "grad_norm": 1.6515879177540538, "learning_rate": 1.446832182205556e-05, "loss": 2.6629, "step": 32990 }, { "epoch": 0.74, "grad_norm": 1.5063516082188158, "learning_rate": 1.4465054497899841e-05, "loss": 2.687, "step": 33000 }, { "epoch": 0.74, "grad_norm": 1.6934583973158808, "learning_rate": 1.4461786578278152e-05, "loss": 2.6704, "step": 33010 }, { "epoch": 0.74, "grad_norm": 1.6879636793496857, "learning_rate": 1.4458518063626303e-05, "loss": 2.5924, "step": 33020 }, { "epoch": 0.74, "grad_norm": 1.8287163918611942, "learning_rate": 1.4455248954380192e-05, "loss": 2.5337, "step": 33030 }, { "epoch": 0.75, "grad_norm": 1.5698030378878438, "learning_rate": 1.4451979250975792e-05, "loss": 2.7993, "step": 33040 }, { "epoch": 0.75, "grad_norm": 1.7228024404242321, "learning_rate": 1.4448708953849151e-05, "loss": 2.723, "step": 33050 }, { "epoch": 0.75, "grad_norm": 1.8266357293256164, "learning_rate": 1.4445438063436403e-05, "loss": 2.5784, "step": 33060 }, { "epoch": 0.75, "grad_norm": 1.5696675155339384, "learning_rate": 1.4442166580173763e-05, "loss": 2.6307, "step": 33070 }, { "epoch": 0.75, "grad_norm": 1.5960575842577225, "learning_rate": 1.4438894504497514e-05, "loss": 2.6204, "step": 33080 }, { "epoch": 0.75, "grad_norm": 1.6879219524469788, "learning_rate": 1.4435621836844027e-05, "loss": 2.7458, "step": 33090 }, { "epoch": 0.75, "grad_norm": 1.6215340865458205, "learning_rate": 1.4432348577649752e-05, "loss": 2.6645, "step": 33100 }, { "epoch": 0.75, "grad_norm": 1.6791568112971675, "learning_rate": 1.442907472735121e-05, "loss": 2.6217, "step": 33110 }, { "epoch": 0.75, "grad_norm": 1.550088165119701, "learning_rate": 1.4425800286385007e-05, "loss": 2.6158, "step": 33120 }, { "epoch": 0.75, "grad_norm": 1.860234571043855, "learning_rate": 1.4422525255187837e-05, "loss": 2.664, "step": 33130 }, { "epoch": 0.75, "grad_norm": 1.7916198442070148, "learning_rate": 1.4419249634196453e-05, "loss": 2.5795, "step": 33140 }, { "epoch": 0.75, "grad_norm": 1.7868419142869483, "learning_rate": 1.4415973423847694e-05, "loss": 2.6709, "step": 33150 }, { "epoch": 0.75, "grad_norm": 1.9260639951153677, "learning_rate": 1.4412696624578488e-05, "loss": 2.5128, "step": 33160 }, { "epoch": 0.75, "grad_norm": 1.6806460417117326, "learning_rate": 1.440941923682583e-05, "loss": 2.7566, "step": 33170 }, { "epoch": 0.75, "grad_norm": 1.5605199385952133, "learning_rate": 1.44061412610268e-05, "loss": 2.6241, "step": 33180 }, { "epoch": 0.75, "grad_norm": 1.5651691705710513, "learning_rate": 1.440286269761855e-05, "loss": 2.5794, "step": 33190 }, { "epoch": 0.75, "grad_norm": 2.0196953035162477, "learning_rate": 1.4399583547038311e-05, "loss": 2.7068, "step": 33200 }, { "epoch": 0.75, "grad_norm": 1.8545691314483628, "learning_rate": 1.4396303809723402e-05, "loss": 2.6585, "step": 33210 }, { "epoch": 0.75, "grad_norm": 1.567396198107839, "learning_rate": 1.4393023486111211e-05, "loss": 2.5641, "step": 33220 }, { "epoch": 0.75, "grad_norm": 1.6734921744814795, "learning_rate": 1.4389742576639202e-05, "loss": 2.6646, "step": 33230 }, { "epoch": 0.75, "grad_norm": 2.4440777346190794, "learning_rate": 1.4386461081744934e-05, "loss": 2.5256, "step": 33240 }, { "epoch": 0.75, "grad_norm": 1.6151190698611508, "learning_rate": 1.4383179001866014e-05, "loss": 2.6686, "step": 33250 }, { "epoch": 0.75, "grad_norm": 1.5915371028108396, "learning_rate": 1.4379896337440157e-05, "loss": 2.7302, "step": 33260 }, { "epoch": 0.75, "grad_norm": 1.9097551330278841, "learning_rate": 1.4376613088905145e-05, "loss": 2.66, "step": 33270 }, { "epoch": 0.75, "grad_norm": 1.8526320760159252, "learning_rate": 1.437332925669883e-05, "loss": 2.6122, "step": 33280 }, { "epoch": 0.75, "grad_norm": 1.4917462756876623, "learning_rate": 1.437004484125915e-05, "loss": 2.7031, "step": 33290 }, { "epoch": 0.75, "grad_norm": 1.5879229614372414, "learning_rate": 1.4366759843024122e-05, "loss": 2.7012, "step": 33300 }, { "epoch": 0.75, "grad_norm": 1.638256148847692, "learning_rate": 1.4363474262431835e-05, "loss": 2.7699, "step": 33310 }, { "epoch": 0.75, "grad_norm": 1.701749449383197, "learning_rate": 1.436018809992046e-05, "loss": 2.7463, "step": 33320 }, { "epoch": 0.75, "grad_norm": 1.5536286390498832, "learning_rate": 1.4356901355928246e-05, "loss": 2.6384, "step": 33330 }, { "epoch": 0.75, "grad_norm": 1.7823611596346776, "learning_rate": 1.4353614030893515e-05, "loss": 2.6192, "step": 33340 }, { "epoch": 0.75, "grad_norm": 1.7659444176827008, "learning_rate": 1.435032612525467e-05, "loss": 2.6438, "step": 33350 }, { "epoch": 0.75, "grad_norm": 1.8253847374042995, "learning_rate": 1.4347037639450194e-05, "loss": 2.5949, "step": 33360 }, { "epoch": 0.75, "grad_norm": 1.7744317865676773, "learning_rate": 1.4343748573918638e-05, "loss": 2.6786, "step": 33370 }, { "epoch": 0.75, "grad_norm": 1.7457986001884562, "learning_rate": 1.4340458929098642e-05, "loss": 2.7007, "step": 33380 }, { "epoch": 0.75, "grad_norm": 1.534061214040574, "learning_rate": 1.4337168705428915e-05, "loss": 2.7581, "step": 33390 }, { "epoch": 0.75, "grad_norm": 1.7002311884370753, "learning_rate": 1.4333877903348245e-05, "loss": 2.5516, "step": 33400 }, { "epoch": 0.75, "grad_norm": 1.6192627893837597, "learning_rate": 1.4330586523295503e-05, "loss": 2.6333, "step": 33410 }, { "epoch": 0.75, "grad_norm": 1.9956531430546343, "learning_rate": 1.4327294565709624e-05, "loss": 2.7128, "step": 33420 }, { "epoch": 0.75, "grad_norm": 1.6929329133748465, "learning_rate": 1.4324002031029633e-05, "loss": 2.5508, "step": 33430 }, { "epoch": 0.75, "grad_norm": 1.5928971599431054, "learning_rate": 1.4320708919694629e-05, "loss": 2.6395, "step": 33440 }, { "epoch": 0.75, "grad_norm": 1.6536104739962039, "learning_rate": 1.4317415232143783e-05, "loss": 2.5486, "step": 33450 }, { "epoch": 0.75, "grad_norm": 1.785948363647976, "learning_rate": 1.4314120968816348e-05, "loss": 2.629, "step": 33460 }, { "epoch": 0.75, "grad_norm": 1.6632894761401207, "learning_rate": 1.4310826130151645e-05, "loss": 2.7301, "step": 33470 }, { "epoch": 0.76, "grad_norm": 1.8676632248681906, "learning_rate": 1.4307530716589086e-05, "loss": 2.5747, "step": 33480 }, { "epoch": 0.76, "grad_norm": 1.8235591235728643, "learning_rate": 1.430423472856815e-05, "loss": 2.6158, "step": 33490 }, { "epoch": 0.76, "grad_norm": 1.8641698826982445, "learning_rate": 1.4300938166528393e-05, "loss": 2.6741, "step": 33500 }, { "epoch": 0.76, "grad_norm": 2.039936432014585, "learning_rate": 1.4297641030909451e-05, "loss": 2.5935, "step": 33510 }, { "epoch": 0.76, "grad_norm": 1.5159642863693734, "learning_rate": 1.4294343322151034e-05, "loss": 2.7109, "step": 33520 }, { "epoch": 0.76, "grad_norm": 1.8090133559173545, "learning_rate": 1.4291045040692929e-05, "loss": 2.5372, "step": 33530 }, { "epoch": 0.76, "grad_norm": 1.5216112128381687, "learning_rate": 1.4287746186974998e-05, "loss": 2.7632, "step": 33540 }, { "epoch": 0.76, "grad_norm": 1.6041242667533566, "learning_rate": 1.4284446761437182e-05, "loss": 2.6562, "step": 33550 }, { "epoch": 0.76, "grad_norm": 1.6694641946789117, "learning_rate": 1.42811467645195e-05, "loss": 2.6955, "step": 33560 }, { "epoch": 0.76, "grad_norm": 1.6652916113796108, "learning_rate": 1.427784619666204e-05, "loss": 2.6897, "step": 33570 }, { "epoch": 0.76, "grad_norm": 2.0426021877372764, "learning_rate": 1.427454505830497e-05, "loss": 2.6281, "step": 33580 }, { "epoch": 0.76, "grad_norm": 1.532653976128416, "learning_rate": 1.4271243349888537e-05, "loss": 2.5566, "step": 33590 }, { "epoch": 0.76, "grad_norm": 1.6617079940793997, "learning_rate": 1.4267941071853063e-05, "loss": 2.6734, "step": 33600 }, { "epoch": 0.76, "grad_norm": 1.8260164884903556, "learning_rate": 1.4264638224638941e-05, "loss": 2.7381, "step": 33610 }, { "epoch": 0.76, "grad_norm": 2.2634668142576606, "learning_rate": 1.4261334808686642e-05, "loss": 2.6183, "step": 33620 }, { "epoch": 0.76, "grad_norm": 1.921882981316056, "learning_rate": 1.4258030824436722e-05, "loss": 2.7831, "step": 33630 }, { "epoch": 0.76, "grad_norm": 1.6718464167686622, "learning_rate": 1.4254726272329795e-05, "loss": 2.5898, "step": 33640 }, { "epoch": 0.76, "grad_norm": 2.068325220865104, "learning_rate": 1.4251421152806568e-05, "loss": 2.537, "step": 33650 }, { "epoch": 0.76, "grad_norm": 1.824409796923841, "learning_rate": 1.4248115466307812e-05, "loss": 2.655, "step": 33660 }, { "epoch": 0.76, "grad_norm": 1.617427723641503, "learning_rate": 1.4244809213274384e-05, "loss": 2.5329, "step": 33670 }, { "epoch": 0.76, "grad_norm": 1.7042643646329367, "learning_rate": 1.4241502394147205e-05, "loss": 2.6548, "step": 33680 }, { "epoch": 0.76, "grad_norm": 1.6094663630870099, "learning_rate": 1.4238195009367276e-05, "loss": 2.5948, "step": 33690 }, { "epoch": 0.76, "grad_norm": 1.4129301886247523, "learning_rate": 1.4234887059375679e-05, "loss": 2.7484, "step": 33700 }, { "epoch": 0.76, "grad_norm": 1.8548032664772445, "learning_rate": 1.4231578544613566e-05, "loss": 2.6549, "step": 33710 }, { "epoch": 0.76, "grad_norm": 1.6806187861201236, "learning_rate": 1.4228269465522165e-05, "loss": 2.6309, "step": 33720 }, { "epoch": 0.76, "grad_norm": 1.8282218921209277, "learning_rate": 1.4224959822542777e-05, "loss": 2.503, "step": 33730 }, { "epoch": 0.76, "grad_norm": 1.478736779608924, "learning_rate": 1.4221649616116782e-05, "loss": 2.6655, "step": 33740 }, { "epoch": 0.76, "grad_norm": 1.5972040917545807, "learning_rate": 1.4218338846685633e-05, "loss": 2.4967, "step": 33750 }, { "epoch": 0.76, "grad_norm": 1.5831430118272913, "learning_rate": 1.421502751469086e-05, "loss": 2.6164, "step": 33760 }, { "epoch": 0.76, "grad_norm": 1.625271677856054, "learning_rate": 1.421171562057407e-05, "loss": 2.5218, "step": 33770 }, { "epoch": 0.76, "grad_norm": 1.8180771025653395, "learning_rate": 1.4208403164776935e-05, "loss": 2.8114, "step": 33780 }, { "epoch": 0.76, "grad_norm": 1.6111104779957146, "learning_rate": 1.4205090147741216e-05, "loss": 2.6823, "step": 33790 }, { "epoch": 0.76, "grad_norm": 1.4537042639106068, "learning_rate": 1.4201776569908739e-05, "loss": 2.6128, "step": 33800 }, { "epoch": 0.76, "grad_norm": 1.6544897193566022, "learning_rate": 1.4198462431721405e-05, "loss": 2.6564, "step": 33810 }, { "epoch": 0.76, "grad_norm": 1.4292574361599395, "learning_rate": 1.4195147733621193e-05, "loss": 2.6013, "step": 33820 }, { "epoch": 0.76, "grad_norm": 1.9819792558125124, "learning_rate": 1.4191832476050158e-05, "loss": 2.7008, "step": 33830 }, { "epoch": 0.76, "grad_norm": 1.8449611950937208, "learning_rate": 1.4188516659450427e-05, "loss": 2.6925, "step": 33840 }, { "epoch": 0.76, "grad_norm": 1.564905393599673, "learning_rate": 1.4185200284264195e-05, "loss": 2.6531, "step": 33850 }, { "epoch": 0.76, "grad_norm": 1.8305005605408748, "learning_rate": 1.4181883350933752e-05, "loss": 2.7466, "step": 33860 }, { "epoch": 0.76, "grad_norm": 1.562759061300756, "learning_rate": 1.4178565859901437e-05, "loss": 2.6143, "step": 33870 }, { "epoch": 0.76, "grad_norm": 1.753944853620913, "learning_rate": 1.417524781160968e-05, "loss": 2.687, "step": 33880 }, { "epoch": 0.76, "grad_norm": 1.5226371784653026, "learning_rate": 1.4171929206500983e-05, "loss": 2.7336, "step": 33890 }, { "epoch": 0.76, "grad_norm": 1.9694109292093096, "learning_rate": 1.4168610045017915e-05, "loss": 2.675, "step": 33900 }, { "epoch": 0.76, "grad_norm": 1.687232757346313, "learning_rate": 1.4165290327603127e-05, "loss": 2.643, "step": 33910 }, { "epoch": 0.76, "grad_norm": 1.687225801597828, "learning_rate": 1.4161970054699338e-05, "loss": 2.6737, "step": 33920 }, { "epoch": 0.77, "grad_norm": 1.4985596743926102, "learning_rate": 1.415864922674935e-05, "loss": 2.5596, "step": 33930 }, { "epoch": 0.77, "grad_norm": 1.6323460364895088, "learning_rate": 1.415532784419603e-05, "loss": 2.6144, "step": 33940 }, { "epoch": 0.77, "grad_norm": 1.6507791451915235, "learning_rate": 1.415200590748232e-05, "loss": 2.5914, "step": 33950 }, { "epoch": 0.77, "grad_norm": 1.509755299248661, "learning_rate": 1.4148683417051241e-05, "loss": 2.6887, "step": 33960 }, { "epoch": 0.77, "grad_norm": 1.5884513340157937, "learning_rate": 1.4145360373345886e-05, "loss": 2.5628, "step": 33970 }, { "epoch": 0.77, "grad_norm": 1.8375153777190056, "learning_rate": 1.4142036776809418e-05, "loss": 2.7006, "step": 33980 }, { "epoch": 0.77, "grad_norm": 2.009705308008569, "learning_rate": 1.4138712627885078e-05, "loss": 2.6954, "step": 33990 }, { "epoch": 0.77, "grad_norm": 1.598566960201946, "learning_rate": 1.4135387927016183e-05, "loss": 2.682, "step": 34000 }, { "epoch": 0.77, "grad_norm": 1.5861984650884497, "learning_rate": 1.4132062674646107e-05, "loss": 2.7643, "step": 34010 }, { "epoch": 0.77, "grad_norm": 1.6911894026914234, "learning_rate": 1.4128736871218326e-05, "loss": 2.5788, "step": 34020 }, { "epoch": 0.77, "grad_norm": 1.671442025412853, "learning_rate": 1.4125410517176365e-05, "loss": 2.5813, "step": 34030 }, { "epoch": 0.77, "grad_norm": 1.665687716327142, "learning_rate": 1.4122083612963832e-05, "loss": 2.5307, "step": 34040 }, { "epoch": 0.77, "grad_norm": 1.5945036768030008, "learning_rate": 1.411875615902441e-05, "loss": 2.6882, "step": 34050 }, { "epoch": 0.77, "grad_norm": 1.586622230257966, "learning_rate": 1.4115428155801853e-05, "loss": 2.6727, "step": 34060 }, { "epoch": 0.77, "grad_norm": 1.6147447701037654, "learning_rate": 1.4112099603739988e-05, "loss": 2.6359, "step": 34070 }, { "epoch": 0.77, "grad_norm": 1.631341567794707, "learning_rate": 1.4108770503282712e-05, "loss": 2.6333, "step": 34080 }, { "epoch": 0.77, "grad_norm": 1.8110492695062277, "learning_rate": 1.4105440854873999e-05, "loss": 2.5329, "step": 34090 }, { "epoch": 0.77, "grad_norm": 1.643186768036196, "learning_rate": 1.4102110658957903e-05, "loss": 2.6114, "step": 34100 }, { "epoch": 0.77, "grad_norm": 1.7116208274105753, "learning_rate": 1.4098779915978536e-05, "loss": 2.6147, "step": 34110 }, { "epoch": 0.77, "grad_norm": 1.625259262853048, "learning_rate": 1.409544862638009e-05, "loss": 2.7181, "step": 34120 }, { "epoch": 0.77, "grad_norm": 1.3950579613078216, "learning_rate": 1.4092116790606838e-05, "loss": 2.6293, "step": 34130 }, { "epoch": 0.77, "grad_norm": 1.8181300508734224, "learning_rate": 1.4088784409103111e-05, "loss": 2.5861, "step": 34140 }, { "epoch": 0.77, "grad_norm": 1.8308840078340116, "learning_rate": 1.408545148231332e-05, "loss": 2.5913, "step": 34150 }, { "epoch": 0.77, "grad_norm": 1.5702044129045933, "learning_rate": 1.4082118010681957e-05, "loss": 2.6131, "step": 34160 }, { "epoch": 0.77, "grad_norm": 1.5496864229422107, "learning_rate": 1.4078783994653571e-05, "loss": 2.6632, "step": 34170 }, { "epoch": 0.77, "grad_norm": 1.592681591438288, "learning_rate": 1.4075449434672792e-05, "loss": 2.6208, "step": 34180 }, { "epoch": 0.77, "grad_norm": 1.6986329779872054, "learning_rate": 1.4072114331184323e-05, "loss": 2.6909, "step": 34190 }, { "epoch": 0.77, "grad_norm": 1.6173310665390068, "learning_rate": 1.406877868463294e-05, "loss": 2.6533, "step": 34200 }, { "epoch": 0.77, "grad_norm": 1.6113118636125474, "learning_rate": 1.4065442495463483e-05, "loss": 2.6701, "step": 34210 }, { "epoch": 0.77, "grad_norm": 1.6521488187355975, "learning_rate": 1.4062105764120879e-05, "loss": 2.4879, "step": 34220 }, { "epoch": 0.77, "grad_norm": 1.34533804355013, "learning_rate": 1.4058768491050115e-05, "loss": 2.6345, "step": 34230 }, { "epoch": 0.77, "grad_norm": 2.114329635836571, "learning_rate": 1.4055430676696256e-05, "loss": 2.6507, "step": 34240 }, { "epoch": 0.77, "grad_norm": 1.7707147967557444, "learning_rate": 1.4052092321504432e-05, "loss": 2.6315, "step": 34250 }, { "epoch": 0.77, "grad_norm": 1.7509622633389466, "learning_rate": 1.4048753425919862e-05, "loss": 2.6519, "step": 34260 }, { "epoch": 0.77, "grad_norm": 1.6468332690286607, "learning_rate": 1.4045413990387815e-05, "loss": 2.5886, "step": 34270 }, { "epoch": 0.77, "grad_norm": 1.9593217585888776, "learning_rate": 1.4042074015353646e-05, "loss": 2.7014, "step": 34280 }, { "epoch": 0.77, "grad_norm": 1.4760751686722249, "learning_rate": 1.4038733501262782e-05, "loss": 2.7012, "step": 34290 }, { "epoch": 0.77, "grad_norm": 1.5024474167338129, "learning_rate": 1.4035392448560716e-05, "loss": 2.5941, "step": 34300 }, { "epoch": 0.77, "grad_norm": 1.7653127872595207, "learning_rate": 1.4032050857693015e-05, "loss": 2.5231, "step": 34310 }, { "epoch": 0.77, "grad_norm": 1.7146529693443457, "learning_rate": 1.402870872910532e-05, "loss": 2.5193, "step": 34320 }, { "epoch": 0.77, "grad_norm": 1.859856377071696, "learning_rate": 1.4025366063243344e-05, "loss": 2.5975, "step": 34330 }, { "epoch": 0.77, "grad_norm": 1.7381260810760828, "learning_rate": 1.4022022860552863e-05, "loss": 2.6265, "step": 34340 }, { "epoch": 0.77, "grad_norm": 1.5970833881755861, "learning_rate": 1.4018679121479735e-05, "loss": 2.5543, "step": 34350 }, { "epoch": 0.77, "grad_norm": 1.7055409914091508, "learning_rate": 1.4015334846469888e-05, "loss": 2.7843, "step": 34360 }, { "epoch": 0.78, "grad_norm": 1.5906164341752316, "learning_rate": 1.4011990035969318e-05, "loss": 2.6919, "step": 34370 }, { "epoch": 0.78, "grad_norm": 1.5424984887730366, "learning_rate": 1.400864469042409e-05, "loss": 2.6137, "step": 34380 }, { "epoch": 0.78, "grad_norm": 1.6970229202437195, "learning_rate": 1.4005298810280348e-05, "loss": 2.6667, "step": 34390 }, { "epoch": 0.78, "grad_norm": 1.4934699498243251, "learning_rate": 1.4001952395984302e-05, "loss": 2.6219, "step": 34400 }, { "epoch": 0.78, "grad_norm": 1.5495730662316693, "learning_rate": 1.3998605447982235e-05, "loss": 2.6153, "step": 34410 }, { "epoch": 0.78, "grad_norm": 1.6196347452831312, "learning_rate": 1.3995257966720497e-05, "loss": 2.5998, "step": 34420 }, { "epoch": 0.78, "grad_norm": 2.0717365900786193, "learning_rate": 1.3991909952645522e-05, "loss": 2.7731, "step": 34430 }, { "epoch": 0.78, "grad_norm": 2.233459186499427, "learning_rate": 1.3988561406203796e-05, "loss": 2.5465, "step": 34440 }, { "epoch": 0.78, "grad_norm": 1.8089819360526416, "learning_rate": 1.398521232784189e-05, "loss": 2.6253, "step": 34450 }, { "epoch": 0.78, "grad_norm": 1.6938372333367409, "learning_rate": 1.3981862718006443e-05, "loss": 2.7143, "step": 34460 }, { "epoch": 0.78, "grad_norm": 1.5882401949788392, "learning_rate": 1.397851257714416e-05, "loss": 2.6543, "step": 34470 }, { "epoch": 0.78, "grad_norm": 1.4893037399222084, "learning_rate": 1.3975161905701827e-05, "loss": 2.7353, "step": 34480 }, { "epoch": 0.78, "grad_norm": 1.6595025096659426, "learning_rate": 1.3971810704126288e-05, "loss": 2.5464, "step": 34490 }, { "epoch": 0.78, "grad_norm": 1.6245878200005746, "learning_rate": 1.3968458972864464e-05, "loss": 2.4707, "step": 34500 }, { "epoch": 0.78, "grad_norm": 1.5998751179918838, "learning_rate": 1.3965106712363353e-05, "loss": 2.5913, "step": 34510 }, { "epoch": 0.78, "grad_norm": 1.5318643745774505, "learning_rate": 1.396175392307001e-05, "loss": 2.5552, "step": 34520 }, { "epoch": 0.78, "grad_norm": 1.7183765976500316, "learning_rate": 1.3958400605431575e-05, "loss": 2.6046, "step": 34530 }, { "epoch": 0.78, "grad_norm": 1.4915307412637002, "learning_rate": 1.3955046759895241e-05, "loss": 2.5782, "step": 34540 }, { "epoch": 0.78, "grad_norm": 1.9875653101586506, "learning_rate": 1.3951692386908291e-05, "loss": 2.6533, "step": 34550 }, { "epoch": 0.78, "grad_norm": 1.704857918102041, "learning_rate": 1.3948337486918061e-05, "loss": 2.6127, "step": 34560 }, { "epoch": 0.78, "grad_norm": 1.4851005324289184, "learning_rate": 1.3944982060371977e-05, "loss": 2.7139, "step": 34570 }, { "epoch": 0.78, "grad_norm": 1.7558250437462544, "learning_rate": 1.3941626107717507e-05, "loss": 2.5818, "step": 34580 }, { "epoch": 0.78, "grad_norm": 1.6683120251412602, "learning_rate": 1.3938269629402223e-05, "loss": 2.5254, "step": 34590 }, { "epoch": 0.78, "grad_norm": 1.6951788193091983, "learning_rate": 1.3934912625873738e-05, "loss": 2.5677, "step": 34600 }, { "epoch": 0.78, "grad_norm": 1.5868581960196382, "learning_rate": 1.3931555097579748e-05, "loss": 2.6858, "step": 34610 }, { "epoch": 0.78, "grad_norm": 1.6834517520349197, "learning_rate": 1.3928197044968022e-05, "loss": 2.597, "step": 34620 }, { "epoch": 0.78, "grad_norm": 1.6714182862097093, "learning_rate": 1.3924838468486393e-05, "loss": 2.6378, "step": 34630 }, { "epoch": 0.78, "grad_norm": 1.6720504119162183, "learning_rate": 1.3921479368582761e-05, "loss": 2.6665, "step": 34640 }, { "epoch": 0.78, "grad_norm": 1.7539723693572036, "learning_rate": 1.3918119745705108e-05, "loss": 2.6011, "step": 34650 }, { "epoch": 0.78, "grad_norm": 1.7613935239059497, "learning_rate": 1.3914759600301473e-05, "loss": 2.4789, "step": 34660 }, { "epoch": 0.78, "grad_norm": 1.5722309710945443, "learning_rate": 1.3911398932819973e-05, "loss": 2.6491, "step": 34670 }, { "epoch": 0.78, "grad_norm": 1.7146318533698095, "learning_rate": 1.3908037743708784e-05, "loss": 2.6605, "step": 34680 }, { "epoch": 0.78, "grad_norm": 1.789970190509087, "learning_rate": 1.3904676033416168e-05, "loss": 2.7086, "step": 34690 }, { "epoch": 0.78, "grad_norm": 1.5853944039053458, "learning_rate": 1.3901313802390446e-05, "loss": 2.5966, "step": 34700 }, { "epoch": 0.78, "grad_norm": 1.919350799603943, "learning_rate": 1.389795105108e-05, "loss": 2.5798, "step": 34710 }, { "epoch": 0.78, "grad_norm": 1.4892972318143411, "learning_rate": 1.3894587779933306e-05, "loss": 2.7553, "step": 34720 }, { "epoch": 0.78, "grad_norm": 1.7418620259241264, "learning_rate": 1.3891223989398884e-05, "loss": 2.6662, "step": 34730 }, { "epoch": 0.78, "grad_norm": 1.7872281186473011, "learning_rate": 1.3887859679925338e-05, "loss": 2.5626, "step": 34740 }, { "epoch": 0.78, "grad_norm": 1.7100075214638573, "learning_rate": 1.3884494851961337e-05, "loss": 2.6931, "step": 34750 }, { "epoch": 0.78, "grad_norm": 1.5927075121815621, "learning_rate": 1.3881129505955618e-05, "loss": 2.7486, "step": 34760 }, { "epoch": 0.78, "grad_norm": 1.7460440414952247, "learning_rate": 1.387776364235699e-05, "loss": 2.6653, "step": 34770 }, { "epoch": 0.78, "grad_norm": 1.6599244070862003, "learning_rate": 1.3874397261614322e-05, "loss": 2.601, "step": 34780 }, { "epoch": 0.78, "grad_norm": 1.659141907036551, "learning_rate": 1.3871030364176573e-05, "loss": 2.6498, "step": 34790 }, { "epoch": 0.78, "grad_norm": 1.61855255328463, "learning_rate": 1.3867662950492747e-05, "loss": 2.7888, "step": 34800 }, { "epoch": 0.78, "grad_norm": 1.6327573182108621, "learning_rate": 1.3864295021011928e-05, "loss": 2.6132, "step": 34810 }, { "epoch": 0.79, "grad_norm": 1.5471203120556942, "learning_rate": 1.386092657618327e-05, "loss": 2.6224, "step": 34820 }, { "epoch": 0.79, "grad_norm": 1.6682712443504493, "learning_rate": 1.3857557616455995e-05, "loss": 2.6813, "step": 34830 }, { "epoch": 0.79, "grad_norm": 1.6191420971465387, "learning_rate": 1.385418814227939e-05, "loss": 2.652, "step": 34840 }, { "epoch": 0.79, "grad_norm": 1.4551369137595822, "learning_rate": 1.3850818154102807e-05, "loss": 2.7099, "step": 34850 }, { "epoch": 0.79, "grad_norm": 1.7276708070082354, "learning_rate": 1.3847447652375688e-05, "loss": 2.6114, "step": 34860 }, { "epoch": 0.79, "grad_norm": 1.6032572228286384, "learning_rate": 1.3844076637547514e-05, "loss": 2.6763, "step": 34870 }, { "epoch": 0.79, "grad_norm": 1.724986854717723, "learning_rate": 1.3840705110067849e-05, "loss": 2.6574, "step": 34880 }, { "epoch": 0.79, "grad_norm": 1.6313925309898394, "learning_rate": 1.3837333070386335e-05, "loss": 2.5456, "step": 34890 }, { "epoch": 0.79, "grad_norm": 1.6424874153430076, "learning_rate": 1.383396051895266e-05, "loss": 2.6788, "step": 34900 }, { "epoch": 0.79, "grad_norm": 2.4769263959490417, "learning_rate": 1.38305874562166e-05, "loss": 2.5668, "step": 34910 }, { "epoch": 0.79, "grad_norm": 1.8485260894645652, "learning_rate": 1.382721388262799e-05, "loss": 2.7572, "step": 34920 }, { "epoch": 0.79, "grad_norm": 1.9054709363967777, "learning_rate": 1.3823839798636732e-05, "loss": 2.5942, "step": 34930 }, { "epoch": 0.79, "grad_norm": 1.663375342892562, "learning_rate": 1.38204652046928e-05, "loss": 2.579, "step": 34940 }, { "epoch": 0.79, "grad_norm": 1.684758688219877, "learning_rate": 1.381709010124623e-05, "loss": 2.5814, "step": 34950 }, { "epoch": 0.79, "grad_norm": 1.5959293310513059, "learning_rate": 1.3813714488747141e-05, "loss": 2.6391, "step": 34960 }, { "epoch": 0.79, "grad_norm": 1.6468347481681038, "learning_rate": 1.3810338367645706e-05, "loss": 2.536, "step": 34970 }, { "epoch": 0.79, "grad_norm": 1.6428768791035466, "learning_rate": 1.380696173839216e-05, "loss": 2.4854, "step": 34980 }, { "epoch": 0.79, "grad_norm": 1.7391057558733587, "learning_rate": 1.3803584601436826e-05, "loss": 2.637, "step": 34990 }, { "epoch": 0.79, "grad_norm": 1.4721264292221996, "learning_rate": 1.3800206957230077e-05, "loss": 2.6435, "step": 35000 }, { "epoch": 0.79, "grad_norm": 1.706667126897512, "learning_rate": 1.3796828806222364e-05, "loss": 2.4823, "step": 35010 }, { "epoch": 0.79, "grad_norm": 1.4808676764633038, "learning_rate": 1.3793450148864205e-05, "loss": 2.6771, "step": 35020 }, { "epoch": 0.79, "grad_norm": 1.666449325713113, "learning_rate": 1.3790070985606177e-05, "loss": 2.6454, "step": 35030 }, { "epoch": 0.79, "grad_norm": 1.4839183904531896, "learning_rate": 1.3786691316898929e-05, "loss": 2.6777, "step": 35040 }, { "epoch": 0.79, "grad_norm": 1.6363358906981786, "learning_rate": 1.3783311143193184e-05, "loss": 2.6723, "step": 35050 }, { "epoch": 0.79, "grad_norm": 1.6521727335831013, "learning_rate": 1.3779930464939719e-05, "loss": 2.7962, "step": 35060 }, { "epoch": 0.79, "grad_norm": 1.8154788156202726, "learning_rate": 1.3776549282589395e-05, "loss": 2.6518, "step": 35070 }, { "epoch": 0.79, "grad_norm": 1.7257964485496784, "learning_rate": 1.3773167596593127e-05, "loss": 2.7826, "step": 35080 }, { "epoch": 0.79, "grad_norm": 2.0743267219733137, "learning_rate": 1.3769785407401901e-05, "loss": 2.6737, "step": 35090 }, { "epoch": 0.79, "grad_norm": 1.8682106681811905, "learning_rate": 1.3766402715466768e-05, "loss": 2.5295, "step": 35100 }, { "epoch": 0.79, "grad_norm": 1.7627387072207719, "learning_rate": 1.3763019521238853e-05, "loss": 2.6477, "step": 35110 }, { "epoch": 0.79, "grad_norm": 1.7390602194525655, "learning_rate": 1.3759635825169343e-05, "loss": 2.5848, "step": 35120 }, { "epoch": 0.79, "grad_norm": 1.7477652256760612, "learning_rate": 1.3756251627709497e-05, "loss": 2.5888, "step": 35130 }, { "epoch": 0.79, "grad_norm": 1.5074576419985022, "learning_rate": 1.3752866929310621e-05, "loss": 2.6806, "step": 35140 }, { "epoch": 0.79, "grad_norm": 1.5862627793984532, "learning_rate": 1.374948173042412e-05, "loss": 2.6679, "step": 35150 }, { "epoch": 0.79, "grad_norm": 1.6278007837386814, "learning_rate": 1.374609603150144e-05, "loss": 2.5886, "step": 35160 }, { "epoch": 0.79, "grad_norm": 1.564995789637278, "learning_rate": 1.3742709832994104e-05, "loss": 2.5772, "step": 35170 }, { "epoch": 0.79, "grad_norm": 1.7212041539165983, "learning_rate": 1.3739323135353699e-05, "loss": 2.7809, "step": 35180 }, { "epoch": 0.79, "grad_norm": 1.7226597549625289, "learning_rate": 1.3735935939031887e-05, "loss": 2.6286, "step": 35190 }, { "epoch": 0.79, "grad_norm": 1.7667415218716271, "learning_rate": 1.3732548244480379e-05, "loss": 2.7281, "step": 35200 }, { "epoch": 0.79, "grad_norm": 2.1071630340810494, "learning_rate": 1.3729160052150968e-05, "loss": 2.6808, "step": 35210 }, { "epoch": 0.79, "grad_norm": 1.5457141897758397, "learning_rate": 1.3725771362495508e-05, "loss": 2.6118, "step": 35220 }, { "epoch": 0.79, "grad_norm": 1.6380574470201439, "learning_rate": 1.3722382175965923e-05, "loss": 2.7262, "step": 35230 }, { "epoch": 0.79, "grad_norm": 1.4587571945033844, "learning_rate": 1.3718992493014189e-05, "loss": 2.6801, "step": 35240 }, { "epoch": 0.79, "grad_norm": 1.554508056987462, "learning_rate": 1.3715602314092368e-05, "loss": 2.6059, "step": 35250 }, { "epoch": 0.8, "grad_norm": 1.6648663207550838, "learning_rate": 1.3712211639652579e-05, "loss": 2.6442, "step": 35260 }, { "epoch": 0.8, "grad_norm": 1.8084407875000945, "learning_rate": 1.3708820470147006e-05, "loss": 2.6283, "step": 35270 }, { "epoch": 0.8, "grad_norm": 1.6853657278480796, "learning_rate": 1.3705428806027893e-05, "loss": 2.7529, "step": 35280 }, { "epoch": 0.8, "grad_norm": 1.6856078803274306, "learning_rate": 1.3702036647747573e-05, "loss": 2.5963, "step": 35290 }, { "epoch": 0.8, "grad_norm": 1.4723557684556126, "learning_rate": 1.3698643995758411e-05, "loss": 2.5333, "step": 35300 }, { "epoch": 0.8, "grad_norm": 1.5096258020909148, "learning_rate": 1.3695250850512864e-05, "loss": 2.6571, "step": 35310 }, { "epoch": 0.8, "grad_norm": 1.672686909552171, "learning_rate": 1.369185721246345e-05, "loss": 2.522, "step": 35320 }, { "epoch": 0.8, "grad_norm": 1.51850246020141, "learning_rate": 1.3688463082062746e-05, "loss": 2.5544, "step": 35330 }, { "epoch": 0.8, "grad_norm": 2.0126695770940235, "learning_rate": 1.3685068459763398e-05, "loss": 2.5115, "step": 35340 }, { "epoch": 0.8, "grad_norm": 1.5791999687527465, "learning_rate": 1.3681673346018118e-05, "loss": 2.6798, "step": 35350 }, { "epoch": 0.8, "grad_norm": 1.5315907007852838, "learning_rate": 1.3678277741279682e-05, "loss": 2.5441, "step": 35360 }, { "epoch": 0.8, "grad_norm": 1.866821081722806, "learning_rate": 1.3674881646000932e-05, "loss": 2.5874, "step": 35370 }, { "epoch": 0.8, "grad_norm": 1.923211364836656, "learning_rate": 1.3671485060634777e-05, "loss": 2.6988, "step": 35380 }, { "epoch": 0.8, "grad_norm": 1.7582382264620067, "learning_rate": 1.3668087985634195e-05, "loss": 2.5246, "step": 35390 }, { "epoch": 0.8, "grad_norm": 1.930613328804806, "learning_rate": 1.366469042145222e-05, "loss": 2.6122, "step": 35400 }, { "epoch": 0.8, "grad_norm": 1.707359979692713, "learning_rate": 1.3661292368541954e-05, "loss": 2.6266, "step": 35410 }, { "epoch": 0.8, "grad_norm": 1.5443236736279218, "learning_rate": 1.365789382735657e-05, "loss": 2.5813, "step": 35420 }, { "epoch": 0.8, "grad_norm": 1.499924087947674, "learning_rate": 1.36544947983493e-05, "loss": 2.5948, "step": 35430 }, { "epoch": 0.8, "grad_norm": 1.3498897941094719, "learning_rate": 1.3651095281973448e-05, "loss": 2.6733, "step": 35440 }, { "epoch": 0.8, "grad_norm": 1.6240169001935014, "learning_rate": 1.3647695278682372e-05, "loss": 2.6707, "step": 35450 }, { "epoch": 0.8, "grad_norm": 1.5954868633184691, "learning_rate": 1.3644294788929504e-05, "loss": 2.6052, "step": 35460 }, { "epoch": 0.8, "grad_norm": 1.3886740824137138, "learning_rate": 1.3640893813168339e-05, "loss": 2.487, "step": 35470 }, { "epoch": 0.8, "grad_norm": 1.6193600219338788, "learning_rate": 1.3637492351852432e-05, "loss": 2.649, "step": 35480 }, { "epoch": 0.8, "grad_norm": 1.9182865052485263, "learning_rate": 1.3634090405435412e-05, "loss": 2.5309, "step": 35490 }, { "epoch": 0.8, "grad_norm": 1.400706803347509, "learning_rate": 1.3630687974370967e-05, "loss": 2.6527, "step": 35500 }, { "epoch": 0.8, "grad_norm": 1.5062625834484986, "learning_rate": 1.3627285059112842e-05, "loss": 2.6145, "step": 35510 }, { "epoch": 0.8, "grad_norm": 1.6547068222016106, "learning_rate": 1.3623881660114866e-05, "loss": 2.6136, "step": 35520 }, { "epoch": 0.8, "grad_norm": 1.6001508276398622, "learning_rate": 1.3620477777830914e-05, "loss": 2.6306, "step": 35530 }, { "epoch": 0.8, "grad_norm": 1.5656907727927807, "learning_rate": 1.3617073412714932e-05, "loss": 2.6579, "step": 35540 }, { "epoch": 0.8, "grad_norm": 1.5800967005879334, "learning_rate": 1.3613668565220936e-05, "loss": 2.5265, "step": 35550 }, { "epoch": 0.8, "grad_norm": 1.5301935752046663, "learning_rate": 1.3610263235802998e-05, "loss": 2.58, "step": 35560 }, { "epoch": 0.8, "grad_norm": 1.7027685869510685, "learning_rate": 1.3606857424915258e-05, "loss": 2.6352, "step": 35570 }, { "epoch": 0.8, "grad_norm": 1.787442641943851, "learning_rate": 1.3603451133011919e-05, "loss": 2.5587, "step": 35580 }, { "epoch": 0.8, "grad_norm": 1.6345696648409358, "learning_rate": 1.360004436054725e-05, "loss": 2.5678, "step": 35590 }, { "epoch": 0.8, "grad_norm": 1.467382563818802, "learning_rate": 1.3596637107975583e-05, "loss": 2.5592, "step": 35600 }, { "epoch": 0.8, "grad_norm": 1.6616309895121602, "learning_rate": 1.3593229375751313e-05, "loss": 2.5524, "step": 35610 }, { "epoch": 0.8, "grad_norm": 1.5995025193906414, "learning_rate": 1.3589821164328902e-05, "loss": 2.599, "step": 35620 }, { "epoch": 0.8, "grad_norm": 1.9327922094858867, "learning_rate": 1.3586412474162871e-05, "loss": 2.6007, "step": 35630 }, { "epoch": 0.8, "grad_norm": 1.5254312457189512, "learning_rate": 1.358300330570781e-05, "loss": 2.7279, "step": 35640 }, { "epoch": 0.8, "grad_norm": 1.9148620813532438, "learning_rate": 1.3579593659418367e-05, "loss": 2.6113, "step": 35650 }, { "epoch": 0.8, "grad_norm": 1.458847526842213, "learning_rate": 1.3576183535749263e-05, "loss": 2.6376, "step": 35660 }, { "epoch": 0.8, "grad_norm": 1.5317238890511937, "learning_rate": 1.3572772935155273e-05, "loss": 2.6678, "step": 35670 }, { "epoch": 0.8, "grad_norm": 1.627806572866699, "learning_rate": 1.3569361858091241e-05, "loss": 2.5163, "step": 35680 }, { "epoch": 0.8, "grad_norm": 1.671693380250628, "learning_rate": 1.3565950305012072e-05, "loss": 2.5077, "step": 35690 }, { "epoch": 0.81, "grad_norm": 1.665671641985802, "learning_rate": 1.3562538276372736e-05, "loss": 2.6152, "step": 35700 }, { "epoch": 0.81, "grad_norm": 1.6239270039483822, "learning_rate": 1.3559125772628266e-05, "loss": 2.6717, "step": 35710 }, { "epoch": 0.81, "grad_norm": 1.6511558338935013, "learning_rate": 1.355571279423376e-05, "loss": 2.6637, "step": 35720 }, { "epoch": 0.81, "grad_norm": 1.8287105531390886, "learning_rate": 1.3552299341644374e-05, "loss": 2.5464, "step": 35730 }, { "epoch": 0.81, "grad_norm": 1.6413008954510049, "learning_rate": 1.3548885415315332e-05, "loss": 2.5602, "step": 35740 }, { "epoch": 0.81, "grad_norm": 1.984066113102726, "learning_rate": 1.354547101570192e-05, "loss": 2.665, "step": 35750 }, { "epoch": 0.81, "grad_norm": 1.9800957360891038, "learning_rate": 1.354205614325949e-05, "loss": 2.6509, "step": 35760 }, { "epoch": 0.81, "grad_norm": 1.5889215620135715, "learning_rate": 1.3538640798443452e-05, "loss": 2.6475, "step": 35770 }, { "epoch": 0.81, "grad_norm": 1.5724551465203733, "learning_rate": 1.3535224981709281e-05, "loss": 2.6943, "step": 35780 }, { "epoch": 0.81, "grad_norm": 2.2173618798223873, "learning_rate": 1.3531808693512517e-05, "loss": 2.735, "step": 35790 }, { "epoch": 0.81, "grad_norm": 1.785589623523377, "learning_rate": 1.3528391934308758e-05, "loss": 2.6159, "step": 35800 }, { "epoch": 0.81, "grad_norm": 1.6366047993206219, "learning_rate": 1.3524974704553668e-05, "loss": 2.5324, "step": 35810 }, { "epoch": 0.81, "grad_norm": 1.9633227061344483, "learning_rate": 1.3521557004702976e-05, "loss": 2.6299, "step": 35820 }, { "epoch": 0.81, "grad_norm": 1.8844330546086292, "learning_rate": 1.3518138835212473e-05, "loss": 2.6595, "step": 35830 }, { "epoch": 0.81, "grad_norm": 1.8040644627224247, "learning_rate": 1.3514720196538003e-05, "loss": 2.6281, "step": 35840 }, { "epoch": 0.81, "grad_norm": 1.609263006739261, "learning_rate": 1.351130108913549e-05, "loss": 2.5796, "step": 35850 }, { "epoch": 0.81, "grad_norm": 1.5590869259703244, "learning_rate": 1.3507881513460903e-05, "loss": 2.5686, "step": 35860 }, { "epoch": 0.81, "grad_norm": 2.4745610429622755, "learning_rate": 1.3504461469970287e-05, "loss": 2.5551, "step": 35870 }, { "epoch": 0.81, "grad_norm": 1.596297543033049, "learning_rate": 1.3501040959119743e-05, "loss": 2.5896, "step": 35880 }, { "epoch": 0.81, "grad_norm": 1.5330269529953326, "learning_rate": 1.3497619981365433e-05, "loss": 2.6516, "step": 35890 }, { "epoch": 0.81, "grad_norm": 1.952924706524933, "learning_rate": 1.3494198537163584e-05, "loss": 2.6585, "step": 35900 }, { "epoch": 0.81, "grad_norm": 1.7936584931087522, "learning_rate": 1.3490776626970483e-05, "loss": 2.6599, "step": 35910 }, { "epoch": 0.81, "grad_norm": 1.964330981698533, "learning_rate": 1.3487354251242484e-05, "loss": 2.6123, "step": 35920 }, { "epoch": 0.81, "grad_norm": 1.6562716131458435, "learning_rate": 1.3483931410436e-05, "loss": 2.6422, "step": 35930 }, { "epoch": 0.81, "grad_norm": 1.526110733357947, "learning_rate": 1.3480508105007504e-05, "loss": 2.6826, "step": 35940 }, { "epoch": 0.81, "grad_norm": 1.8264939261328226, "learning_rate": 1.3477084335413532e-05, "loss": 2.5358, "step": 35950 }, { "epoch": 0.81, "grad_norm": 1.5594872708677532, "learning_rate": 1.3473660102110686e-05, "loss": 2.5775, "step": 35960 }, { "epoch": 0.81, "grad_norm": 1.6273367264421572, "learning_rate": 1.3470235405555625e-05, "loss": 2.7529, "step": 35970 }, { "epoch": 0.81, "grad_norm": 1.7112166841224963, "learning_rate": 1.3466810246205068e-05, "loss": 2.7423, "step": 35980 }, { "epoch": 0.81, "grad_norm": 1.7021825580521153, "learning_rate": 1.346338462451581e-05, "loss": 2.4778, "step": 35990 }, { "epoch": 0.81, "grad_norm": 1.6341477233263142, "learning_rate": 1.3459958540944681e-05, "loss": 2.6433, "step": 36000 }, { "epoch": 0.81, "grad_norm": 1.7133202388537268, "learning_rate": 1.3456531995948599e-05, "loss": 2.5879, "step": 36010 }, { "epoch": 0.81, "grad_norm": 1.5617753213999448, "learning_rate": 1.3453104989984529e-05, "loss": 2.6935, "step": 36020 }, { "epoch": 0.81, "grad_norm": 1.6208217542575616, "learning_rate": 1.3449677523509504e-05, "loss": 2.6641, "step": 36030 }, { "epoch": 0.81, "grad_norm": 1.7403582604253443, "learning_rate": 1.3446249596980614e-05, "loss": 2.6106, "step": 36040 }, { "epoch": 0.81, "grad_norm": 1.8636582831259247, "learning_rate": 1.3442821210855012e-05, "loss": 2.5114, "step": 36050 }, { "epoch": 0.81, "grad_norm": 1.651826978223427, "learning_rate": 1.3439392365589913e-05, "loss": 2.6382, "step": 36060 }, { "epoch": 0.81, "grad_norm": 1.7199458337767068, "learning_rate": 1.3435963061642593e-05, "loss": 2.6018, "step": 36070 }, { "epoch": 0.81, "grad_norm": 1.5953182561155765, "learning_rate": 1.3432533299470385e-05, "loss": 2.6342, "step": 36080 }, { "epoch": 0.81, "grad_norm": 1.6156036099161994, "learning_rate": 1.3429103079530695e-05, "loss": 2.6108, "step": 36090 }, { "epoch": 0.81, "grad_norm": 1.7116381619848142, "learning_rate": 1.3425672402280974e-05, "loss": 2.6073, "step": 36100 }, { "epoch": 0.81, "grad_norm": 1.5797363720382667, "learning_rate": 1.3422241268178748e-05, "loss": 2.6106, "step": 36110 }, { "epoch": 0.81, "grad_norm": 1.958978064703157, "learning_rate": 1.3418809677681592e-05, "loss": 2.5757, "step": 36120 }, { "epoch": 0.81, "grad_norm": 1.571736889457023, "learning_rate": 1.3415377631247153e-05, "loss": 2.6619, "step": 36130 }, { "epoch": 0.81, "grad_norm": 1.8308773329913322, "learning_rate": 1.3411945129333128e-05, "loss": 2.618, "step": 36140 }, { "epoch": 0.82, "grad_norm": 1.9902306252973194, "learning_rate": 1.3408512172397288e-05, "loss": 2.5921, "step": 36150 }, { "epoch": 0.82, "grad_norm": 1.920495614112821, "learning_rate": 1.3405078760897452e-05, "loss": 2.6466, "step": 36160 }, { "epoch": 0.82, "grad_norm": 1.6900647513425457, "learning_rate": 1.3401644895291502e-05, "loss": 2.4595, "step": 36170 }, { "epoch": 0.82, "grad_norm": 1.615863470304394, "learning_rate": 1.3398210576037389e-05, "loss": 2.6391, "step": 36180 }, { "epoch": 0.82, "grad_norm": 1.6403235342265832, "learning_rate": 1.3394775803593114e-05, "loss": 2.5986, "step": 36190 }, { "epoch": 0.82, "grad_norm": 1.668455496640931, "learning_rate": 1.3391340578416747e-05, "loss": 2.7883, "step": 36200 }, { "epoch": 0.82, "grad_norm": 1.5486918323753867, "learning_rate": 1.3387904900966414e-05, "loss": 2.4246, "step": 36210 }, { "epoch": 0.82, "grad_norm": 1.511118205227325, "learning_rate": 1.3384468771700298e-05, "loss": 2.5861, "step": 36220 }, { "epoch": 0.82, "grad_norm": 1.7413545070631666, "learning_rate": 1.3381032191076649e-05, "loss": 2.5548, "step": 36230 }, { "epoch": 0.82, "grad_norm": 1.6544822799975858, "learning_rate": 1.3377595159553772e-05, "loss": 2.6337, "step": 36240 }, { "epoch": 0.82, "grad_norm": 1.7295389445888802, "learning_rate": 1.3374157677590038e-05, "loss": 2.6152, "step": 36250 }, { "epoch": 0.82, "grad_norm": 1.7579636594743295, "learning_rate": 1.3370719745643873e-05, "loss": 2.5522, "step": 36260 }, { "epoch": 0.82, "grad_norm": 1.5571798114921631, "learning_rate": 1.3367281364173763e-05, "loss": 2.5892, "step": 36270 }, { "epoch": 0.82, "grad_norm": 1.8686742372070437, "learning_rate": 1.3363842533638259e-05, "loss": 2.5601, "step": 36280 }, { "epoch": 0.82, "grad_norm": 1.5744499441135937, "learning_rate": 1.3360403254495962e-05, "loss": 2.6933, "step": 36290 }, { "epoch": 0.82, "grad_norm": 1.6521671146818817, "learning_rate": 1.3356963527205544e-05, "loss": 2.5743, "step": 36300 }, { "epoch": 0.82, "grad_norm": 1.6854144980030503, "learning_rate": 1.3353523352225732e-05, "loss": 2.686, "step": 36310 }, { "epoch": 0.82, "grad_norm": 1.5032163200158892, "learning_rate": 1.335008273001531e-05, "loss": 2.7381, "step": 36320 }, { "epoch": 0.82, "grad_norm": 2.1679180401725877, "learning_rate": 1.3346641661033125e-05, "loss": 2.6322, "step": 36330 }, { "epoch": 0.82, "grad_norm": 1.548723628765049, "learning_rate": 1.3343200145738085e-05, "loss": 2.6355, "step": 36340 }, { "epoch": 0.82, "grad_norm": 1.4511864314052754, "learning_rate": 1.3339758184589151e-05, "loss": 2.4782, "step": 36350 }, { "epoch": 0.82, "grad_norm": 1.545718831530955, "learning_rate": 1.3336315778045355e-05, "loss": 2.687, "step": 36360 }, { "epoch": 0.82, "grad_norm": 1.5773953279262667, "learning_rate": 1.3332872926565774e-05, "loss": 2.7186, "step": 36370 }, { "epoch": 0.82, "grad_norm": 1.7139623128149708, "learning_rate": 1.3329429630609554e-05, "loss": 2.5346, "step": 36380 }, { "epoch": 0.82, "grad_norm": 1.7053318928395786, "learning_rate": 1.33259858906359e-05, "loss": 2.7737, "step": 36390 }, { "epoch": 0.82, "grad_norm": 1.575162165369109, "learning_rate": 1.332254170710407e-05, "loss": 2.6225, "step": 36400 }, { "epoch": 0.82, "grad_norm": 1.523992661766608, "learning_rate": 1.3319097080473386e-05, "loss": 2.7619, "step": 36410 }, { "epoch": 0.82, "grad_norm": 1.6318687718645362, "learning_rate": 1.3315652011203233e-05, "loss": 2.6983, "step": 36420 }, { "epoch": 0.82, "grad_norm": 1.4889669398995113, "learning_rate": 1.3312206499753045e-05, "loss": 2.6171, "step": 36430 }, { "epoch": 0.82, "grad_norm": 1.8964360483962035, "learning_rate": 1.330876054658232e-05, "loss": 2.5701, "step": 36440 }, { "epoch": 0.82, "grad_norm": 1.7604695200994902, "learning_rate": 1.3305314152150622e-05, "loss": 2.6392, "step": 36450 }, { "epoch": 0.82, "grad_norm": 1.615161961470705, "learning_rate": 1.3301867316917559e-05, "loss": 2.7161, "step": 36460 }, { "epoch": 0.82, "grad_norm": 1.516793318627129, "learning_rate": 1.329842004134281e-05, "loss": 2.5736, "step": 36470 }, { "epoch": 0.82, "grad_norm": 1.5896237422720259, "learning_rate": 1.3294972325886107e-05, "loss": 2.6771, "step": 36480 }, { "epoch": 0.82, "grad_norm": 1.5323679372635106, "learning_rate": 1.3291524171007243e-05, "loss": 2.6035, "step": 36490 }, { "epoch": 0.82, "grad_norm": 1.5530621338868644, "learning_rate": 1.328807557716607e-05, "loss": 2.6477, "step": 36500 }, { "epoch": 0.82, "grad_norm": 1.6591396115556276, "learning_rate": 1.3284626544822494e-05, "loss": 2.7062, "step": 36510 }, { "epoch": 0.82, "grad_norm": 2.793262070948556, "learning_rate": 1.3281177074436484e-05, "loss": 2.6056, "step": 36520 }, { "epoch": 0.82, "grad_norm": 1.470968145913293, "learning_rate": 1.3277727166468073e-05, "loss": 2.6912, "step": 36530 }, { "epoch": 0.82, "grad_norm": 1.7398671290145504, "learning_rate": 1.3274276821377333e-05, "loss": 2.6202, "step": 36540 }, { "epoch": 0.82, "grad_norm": 1.709346190043201, "learning_rate": 1.3270826039624416e-05, "loss": 2.7557, "step": 36550 }, { "epoch": 0.82, "grad_norm": 1.6792747242962498, "learning_rate": 1.3267374821669521e-05, "loss": 2.6988, "step": 36560 }, { "epoch": 0.82, "grad_norm": 1.570434345516859, "learning_rate": 1.3263923167972905e-05, "loss": 2.6085, "step": 36570 }, { "epoch": 0.82, "grad_norm": 1.7201183365533352, "learning_rate": 1.3260471078994893e-05, "loss": 2.7152, "step": 36580 }, { "epoch": 0.83, "grad_norm": 1.488494955721631, "learning_rate": 1.3257018555195851e-05, "loss": 2.5569, "step": 36590 }, { "epoch": 0.83, "grad_norm": 1.5692969899142133, "learning_rate": 1.3253565597036213e-05, "loss": 2.5724, "step": 36600 }, { "epoch": 0.83, "grad_norm": 1.757425100841872, "learning_rate": 1.3250112204976478e-05, "loss": 2.6991, "step": 36610 }, { "epoch": 0.83, "grad_norm": 1.6229895644269845, "learning_rate": 1.324665837947719e-05, "loss": 2.7238, "step": 36620 }, { "epoch": 0.83, "grad_norm": 1.727910735581579, "learning_rate": 1.3243204120998956e-05, "loss": 2.5885, "step": 36630 }, { "epoch": 0.83, "grad_norm": 1.8881184740081154, "learning_rate": 1.3239749430002442e-05, "loss": 2.4498, "step": 36640 }, { "epoch": 0.83, "grad_norm": 1.6150061904058157, "learning_rate": 1.3236294306948371e-05, "loss": 2.5246, "step": 36650 }, { "epoch": 0.83, "grad_norm": 1.5479711879659805, "learning_rate": 1.3232838752297523e-05, "loss": 2.597, "step": 36660 }, { "epoch": 0.83, "grad_norm": 1.6435043062262622, "learning_rate": 1.322938276651073e-05, "loss": 2.6623, "step": 36670 }, { "epoch": 0.83, "grad_norm": 1.9010886683571357, "learning_rate": 1.32259263500489e-05, "loss": 2.6936, "step": 36680 }, { "epoch": 0.83, "grad_norm": 1.629430059610173, "learning_rate": 1.3222469503372972e-05, "loss": 2.6637, "step": 36690 }, { "epoch": 0.83, "grad_norm": 1.8684733737390589, "learning_rate": 1.3219012226943963e-05, "loss": 2.4602, "step": 36700 }, { "epoch": 0.83, "grad_norm": 1.7007411726788266, "learning_rate": 1.3215554521222939e-05, "loss": 2.514, "step": 36710 }, { "epoch": 0.83, "grad_norm": 1.9537835505647263, "learning_rate": 1.3212096386671026e-05, "loss": 2.6016, "step": 36720 }, { "epoch": 0.83, "grad_norm": 1.6781916927190956, "learning_rate": 1.3208637823749401e-05, "loss": 2.5625, "step": 36730 }, { "epoch": 0.83, "grad_norm": 1.6020853081239563, "learning_rate": 1.3205178832919305e-05, "loss": 2.5518, "step": 36740 }, { "epoch": 0.83, "grad_norm": 1.748142993610591, "learning_rate": 1.320171941464204e-05, "loss": 2.6721, "step": 36750 }, { "epoch": 0.83, "grad_norm": 1.6271911406178108, "learning_rate": 1.3198259569378952e-05, "loss": 2.5618, "step": 36760 }, { "epoch": 0.83, "grad_norm": 1.427336008522268, "learning_rate": 1.319479929759145e-05, "loss": 2.6333, "step": 36770 }, { "epoch": 0.83, "grad_norm": 1.5319363934449883, "learning_rate": 1.3191338599741007e-05, "loss": 2.7219, "step": 36780 }, { "epoch": 0.83, "grad_norm": 1.8388748230123297, "learning_rate": 1.3187877476289139e-05, "loss": 2.7072, "step": 36790 }, { "epoch": 0.83, "grad_norm": 1.9815848230493143, "learning_rate": 1.3184415927697431e-05, "loss": 2.5808, "step": 36800 }, { "epoch": 0.83, "grad_norm": 1.6224338661288102, "learning_rate": 1.3180953954427522e-05, "loss": 2.6797, "step": 36810 }, { "epoch": 0.83, "grad_norm": 1.7547330708590854, "learning_rate": 1.3177491556941103e-05, "loss": 2.638, "step": 36820 }, { "epoch": 0.83, "grad_norm": 1.8766143957930992, "learning_rate": 1.317402873569992e-05, "loss": 2.5416, "step": 36830 }, { "epoch": 0.83, "grad_norm": 1.6613296995656055, "learning_rate": 1.3170565491165785e-05, "loss": 2.6028, "step": 36840 }, { "epoch": 0.83, "grad_norm": 1.5210391764116675, "learning_rate": 1.3167101823800564e-05, "loss": 2.5248, "step": 36850 }, { "epoch": 0.83, "grad_norm": 1.6224194381333807, "learning_rate": 1.3163637734066169e-05, "loss": 2.5833, "step": 36860 }, { "epoch": 0.83, "grad_norm": 1.6012188060148733, "learning_rate": 1.3160173222424576e-05, "loss": 2.6138, "step": 36870 }, { "epoch": 0.83, "grad_norm": 1.6093118064889358, "learning_rate": 1.3156708289337824e-05, "loss": 2.4626, "step": 36880 }, { "epoch": 0.83, "grad_norm": 1.5945115464010773, "learning_rate": 1.3153242935268e-05, "loss": 2.5919, "step": 36890 }, { "epoch": 0.83, "grad_norm": 1.8210735152941866, "learning_rate": 1.3149777160677247e-05, "loss": 2.4755, "step": 36900 }, { "epoch": 0.83, "grad_norm": 1.5020811850063398, "learning_rate": 1.3146310966027762e-05, "loss": 2.5452, "step": 36910 }, { "epoch": 0.83, "grad_norm": 1.8673000192223184, "learning_rate": 1.3142844351781805e-05, "loss": 2.6537, "step": 36920 }, { "epoch": 0.83, "grad_norm": 1.7349294635508674, "learning_rate": 1.313937731840169e-05, "loss": 2.6735, "step": 36930 }, { "epoch": 0.83, "grad_norm": 1.6152652848295417, "learning_rate": 1.3135909866349781e-05, "loss": 2.6901, "step": 36940 }, { "epoch": 0.83, "grad_norm": 2.0771983704586163, "learning_rate": 1.3132441996088505e-05, "loss": 2.6079, "step": 36950 }, { "epoch": 0.83, "grad_norm": 1.6422324018414238, "learning_rate": 1.3128973708080345e-05, "loss": 2.5735, "step": 36960 }, { "epoch": 0.83, "grad_norm": 1.5149861821519657, "learning_rate": 1.3125505002787833e-05, "loss": 2.51, "step": 36970 }, { "epoch": 0.83, "grad_norm": 1.453421332689358, "learning_rate": 1.312203588067356e-05, "loss": 2.7474, "step": 36980 }, { "epoch": 0.83, "grad_norm": 1.699730845793278, "learning_rate": 1.3118566342200176e-05, "loss": 2.7327, "step": 36990 }, { "epoch": 0.83, "grad_norm": 1.5926597450168474, "learning_rate": 1.311509638783038e-05, "loss": 2.6388, "step": 37000 }, { "epoch": 0.83, "grad_norm": 1.6072210782432266, "learning_rate": 1.3111626018026932e-05, "loss": 2.6111, "step": 37010 }, { "epoch": 0.83, "grad_norm": 1.838258046458815, "learning_rate": 1.3108155233252648e-05, "loss": 2.5741, "step": 37020 }, { "epoch": 0.84, "grad_norm": 1.7782848837514, "learning_rate": 1.3104684033970392e-05, "loss": 2.8068, "step": 37030 }, { "epoch": 0.84, "grad_norm": 1.9373162979344636, "learning_rate": 1.3101212420643091e-05, "loss": 2.5242, "step": 37040 }, { "epoch": 0.84, "grad_norm": 1.5764327007963759, "learning_rate": 1.3097740393733722e-05, "loss": 2.6942, "step": 37050 }, { "epoch": 0.84, "grad_norm": 1.49627833086934, "learning_rate": 1.309426795370532e-05, "loss": 2.56, "step": 37060 }, { "epoch": 0.84, "grad_norm": 1.5031683179150737, "learning_rate": 1.3090795101020976e-05, "loss": 2.6575, "step": 37070 }, { "epoch": 0.84, "grad_norm": 1.9000344855581595, "learning_rate": 1.3087321836143836e-05, "loss": 2.6722, "step": 37080 }, { "epoch": 0.84, "grad_norm": 1.4660028182382656, "learning_rate": 1.3083848159537098e-05, "loss": 2.8326, "step": 37090 }, { "epoch": 0.84, "grad_norm": 1.9822133763844674, "learning_rate": 1.308037407166401e-05, "loss": 2.6505, "step": 37100 }, { "epoch": 0.84, "grad_norm": 1.5007557534015528, "learning_rate": 1.3076899572987891e-05, "loss": 2.6882, "step": 37110 }, { "epoch": 0.84, "grad_norm": 2.0169338292554038, "learning_rate": 1.3073424663972104e-05, "loss": 2.6734, "step": 37120 }, { "epoch": 0.84, "grad_norm": 1.5633503820508388, "learning_rate": 1.3069949345080058e-05, "loss": 2.5223, "step": 37130 }, { "epoch": 0.84, "grad_norm": 1.7597533430027792, "learning_rate": 1.3066473616775237e-05, "loss": 2.6193, "step": 37140 }, { "epoch": 0.84, "grad_norm": 1.9018229342214001, "learning_rate": 1.3062997479521164e-05, "loss": 2.6268, "step": 37150 }, { "epoch": 0.84, "grad_norm": 1.7155786451109662, "learning_rate": 1.3059520933781427e-05, "loss": 2.5931, "step": 37160 }, { "epoch": 0.84, "grad_norm": 1.640369554345249, "learning_rate": 1.3056043980019652e-05, "loss": 2.4712, "step": 37170 }, { "epoch": 0.84, "grad_norm": 1.6748707377838552, "learning_rate": 1.3052566618699544e-05, "loss": 2.5996, "step": 37180 }, { "epoch": 0.84, "grad_norm": 1.7950288352627337, "learning_rate": 1.304908885028484e-05, "loss": 2.6811, "step": 37190 }, { "epoch": 0.84, "grad_norm": 1.5159289787165002, "learning_rate": 1.304561067523934e-05, "loss": 2.7176, "step": 37200 }, { "epoch": 0.84, "grad_norm": 1.5345297409569378, "learning_rate": 1.3042132094026902e-05, "loss": 2.6506, "step": 37210 }, { "epoch": 0.84, "grad_norm": 1.6299191105599151, "learning_rate": 1.3038653107111432e-05, "loss": 2.5933, "step": 37220 }, { "epoch": 0.84, "grad_norm": 1.6059617686561183, "learning_rate": 1.3035173714956897e-05, "loss": 2.6402, "step": 37230 }, { "epoch": 0.84, "grad_norm": 1.5891089451882643, "learning_rate": 1.3031693918027309e-05, "loss": 2.6012, "step": 37240 }, { "epoch": 0.84, "grad_norm": 1.6410252491040234, "learning_rate": 1.302821371678674e-05, "loss": 2.6396, "step": 37250 }, { "epoch": 0.84, "grad_norm": 1.5791002944755739, "learning_rate": 1.3024733111699313e-05, "loss": 2.6619, "step": 37260 }, { "epoch": 0.84, "grad_norm": 2.068762906085575, "learning_rate": 1.3021252103229207e-05, "loss": 2.4767, "step": 37270 }, { "epoch": 0.84, "grad_norm": 1.642890589404591, "learning_rate": 1.3017770691840657e-05, "loss": 2.5826, "step": 37280 }, { "epoch": 0.84, "grad_norm": 1.527667448453053, "learning_rate": 1.301428887799795e-05, "loss": 2.5661, "step": 37290 }, { "epoch": 0.84, "grad_norm": 1.5753326131788468, "learning_rate": 1.3010806662165419e-05, "loss": 2.8115, "step": 37300 }, { "epoch": 0.84, "grad_norm": 1.5708363782428796, "learning_rate": 1.3007324044807459e-05, "loss": 2.5092, "step": 37310 }, { "epoch": 0.84, "grad_norm": 1.6146663714495795, "learning_rate": 1.3003841026388523e-05, "loss": 2.5934, "step": 37320 }, { "epoch": 0.84, "grad_norm": 1.8979428538507686, "learning_rate": 1.3000357607373107e-05, "loss": 2.6806, "step": 37330 }, { "epoch": 0.84, "grad_norm": 1.5526540328792902, "learning_rate": 1.299687378822576e-05, "loss": 2.6159, "step": 37340 }, { "epoch": 0.84, "grad_norm": 1.715839923124289, "learning_rate": 1.2993389569411099e-05, "loss": 2.5622, "step": 37350 }, { "epoch": 0.84, "grad_norm": 1.8893321395247604, "learning_rate": 1.2989904951393778e-05, "loss": 2.6128, "step": 37360 }, { "epoch": 0.84, "grad_norm": 1.8721843056139218, "learning_rate": 1.2986419934638507e-05, "loss": 2.6167, "step": 37370 }, { "epoch": 0.84, "grad_norm": 1.77557870020789, "learning_rate": 1.2982934519610061e-05, "loss": 2.5283, "step": 37380 }, { "epoch": 0.84, "grad_norm": 1.6296401740152926, "learning_rate": 1.2979448706773259e-05, "loss": 2.5316, "step": 37390 }, { "epoch": 0.84, "grad_norm": 1.7174713337413952, "learning_rate": 1.2975962496592966e-05, "loss": 2.5983, "step": 37400 }, { "epoch": 0.84, "grad_norm": 1.6045275555765974, "learning_rate": 1.2972475889534114e-05, "loss": 2.5586, "step": 37410 }, { "epoch": 0.84, "grad_norm": 1.4843932204826837, "learning_rate": 1.2968988886061684e-05, "loss": 2.7077, "step": 37420 }, { "epoch": 0.84, "grad_norm": 1.5346365014922785, "learning_rate": 1.2965501486640704e-05, "loss": 2.6342, "step": 37430 }, { "epoch": 0.84, "grad_norm": 1.6401459450656377, "learning_rate": 1.2962013691736254e-05, "loss": 2.6739, "step": 37440 }, { "epoch": 0.84, "grad_norm": 1.780387966025886, "learning_rate": 1.2958525501813483e-05, "loss": 2.6228, "step": 37450 }, { "epoch": 0.84, "grad_norm": 1.731322141990189, "learning_rate": 1.2955036917337572e-05, "loss": 2.6082, "step": 37460 }, { "epoch": 0.84, "grad_norm": 1.7010390668957442, "learning_rate": 1.2951547938773763e-05, "loss": 2.7196, "step": 37470 }, { "epoch": 0.85, "grad_norm": 1.8088206023137845, "learning_rate": 1.2948058566587356e-05, "loss": 2.7137, "step": 37480 }, { "epoch": 0.85, "grad_norm": 1.7826961900635823, "learning_rate": 1.2944568801243697e-05, "loss": 2.5667, "step": 37490 }, { "epoch": 0.85, "grad_norm": 1.8942411595292363, "learning_rate": 1.2941078643208184e-05, "loss": 2.5629, "step": 37500 }, { "epoch": 0.85, "grad_norm": 1.691234332723233, "learning_rate": 1.2937588092946274e-05, "loss": 2.5618, "step": 37510 }, { "epoch": 0.85, "grad_norm": 1.4376614193133845, "learning_rate": 1.2934097150923469e-05, "loss": 2.7131, "step": 37520 }, { "epoch": 0.85, "grad_norm": 1.7665899489300596, "learning_rate": 1.2930605817605323e-05, "loss": 2.4966, "step": 37530 }, { "epoch": 0.85, "grad_norm": 1.7194383256152679, "learning_rate": 1.2927114093457447e-05, "loss": 2.7057, "step": 37540 }, { "epoch": 0.85, "grad_norm": 1.4883551993944102, "learning_rate": 1.2923621978945507e-05, "loss": 2.6534, "step": 37550 }, { "epoch": 0.85, "grad_norm": 1.5314462119700125, "learning_rate": 1.292012947453521e-05, "loss": 2.6476, "step": 37560 }, { "epoch": 0.85, "grad_norm": 1.5232928668805865, "learning_rate": 1.2916636580692322e-05, "loss": 2.5714, "step": 37570 }, { "epoch": 0.85, "grad_norm": 1.545893278521504, "learning_rate": 1.2913143297882662e-05, "loss": 2.6273, "step": 37580 }, { "epoch": 0.85, "grad_norm": 1.5705755368364096, "learning_rate": 1.2909649626572102e-05, "loss": 2.6725, "step": 37590 }, { "epoch": 0.85, "grad_norm": 1.570862415953368, "learning_rate": 1.2906155567226557e-05, "loss": 2.6583, "step": 37600 }, { "epoch": 0.85, "grad_norm": 1.5592638055182484, "learning_rate": 1.2902661120312005e-05, "loss": 2.6305, "step": 37610 }, { "epoch": 0.85, "grad_norm": 1.7496185599442813, "learning_rate": 1.2899166286294468e-05, "loss": 2.6805, "step": 37620 }, { "epoch": 0.85, "grad_norm": 1.557518383981101, "learning_rate": 1.289567106564002e-05, "loss": 2.705, "step": 37630 }, { "epoch": 0.85, "grad_norm": 2.064058428409677, "learning_rate": 1.2892175458814793e-05, "loss": 2.5934, "step": 37640 }, { "epoch": 0.85, "grad_norm": 1.5158236410290142, "learning_rate": 1.2888679466284963e-05, "loss": 2.6903, "step": 37650 }, { "epoch": 0.85, "grad_norm": 1.5749738723379718, "learning_rate": 1.2885183088516762e-05, "loss": 2.6718, "step": 37660 }, { "epoch": 0.85, "grad_norm": 1.615173966636303, "learning_rate": 1.288168632597647e-05, "loss": 2.6069, "step": 37670 }, { "epoch": 0.85, "grad_norm": 1.594250411313882, "learning_rate": 1.287818917913042e-05, "loss": 2.5875, "step": 37680 }, { "epoch": 0.85, "grad_norm": 1.6011044507069019, "learning_rate": 1.2874691648445001e-05, "loss": 2.6606, "step": 37690 }, { "epoch": 0.85, "grad_norm": 1.7711285627596347, "learning_rate": 1.287119373438664e-05, "loss": 2.4445, "step": 37700 }, { "epoch": 0.85, "grad_norm": 1.9124106943196588, "learning_rate": 1.2867695437421834e-05, "loss": 2.5799, "step": 37710 }, { "epoch": 0.85, "grad_norm": 1.7408898202069538, "learning_rate": 1.2864196758017115e-05, "loss": 2.5268, "step": 37720 }, { "epoch": 0.85, "grad_norm": 1.9837188600475524, "learning_rate": 1.286069769663907e-05, "loss": 2.7136, "step": 37730 }, { "epoch": 0.85, "grad_norm": 1.7479112786712452, "learning_rate": 1.2857198253754346e-05, "loss": 2.6539, "step": 37740 }, { "epoch": 0.85, "grad_norm": 1.8140165969455981, "learning_rate": 1.2853698429829627e-05, "loss": 2.6188, "step": 37750 }, { "epoch": 0.85, "grad_norm": 1.4953421337460517, "learning_rate": 1.2850198225331656e-05, "loss": 2.5237, "step": 37760 }, { "epoch": 0.85, "grad_norm": 1.8418084620152921, "learning_rate": 1.2846697640727226e-05, "loss": 2.4476, "step": 37770 }, { "epoch": 0.85, "grad_norm": 1.5959018733148194, "learning_rate": 1.284319667648318e-05, "loss": 2.5731, "step": 37780 }, { "epoch": 0.85, "grad_norm": 1.3969181991410815, "learning_rate": 1.283969533306641e-05, "loss": 2.5146, "step": 37790 }, { "epoch": 0.85, "grad_norm": 1.7835121542564518, "learning_rate": 1.2836193610943864e-05, "loss": 2.6793, "step": 37800 }, { "epoch": 0.85, "grad_norm": 1.4815274459296923, "learning_rate": 1.2832691510582532e-05, "loss": 2.5531, "step": 37810 }, { "epoch": 0.85, "grad_norm": 1.7975192498098789, "learning_rate": 1.2829189032449464e-05, "loss": 2.6275, "step": 37820 }, { "epoch": 0.85, "grad_norm": 1.5816034086226303, "learning_rate": 1.2825686177011748e-05, "loss": 2.5201, "step": 37830 }, { "epoch": 0.85, "grad_norm": 1.7459064000477564, "learning_rate": 1.2822182944736534e-05, "loss": 2.6525, "step": 37840 }, { "epoch": 0.85, "grad_norm": 1.7689946667247511, "learning_rate": 1.2818679336091022e-05, "loss": 2.7661, "step": 37850 }, { "epoch": 0.85, "grad_norm": 1.522948829617737, "learning_rate": 1.2815175351542453e-05, "loss": 2.6116, "step": 37860 }, { "epoch": 0.85, "grad_norm": 1.7580689820981275, "learning_rate": 1.2811670991558123e-05, "loss": 2.5986, "step": 37870 }, { "epoch": 0.85, "grad_norm": 1.6682263762046692, "learning_rate": 1.2808166256605383e-05, "loss": 2.6602, "step": 37880 }, { "epoch": 0.85, "grad_norm": 1.8227028109122725, "learning_rate": 1.2804661147151625e-05, "loss": 2.5442, "step": 37890 }, { "epoch": 0.85, "grad_norm": 1.8605442500155065, "learning_rate": 1.2801155663664296e-05, "loss": 2.5448, "step": 37900 }, { "epoch": 0.85, "grad_norm": 1.838179134930112, "learning_rate": 1.2797649806610895e-05, "loss": 2.6058, "step": 37910 }, { "epoch": 0.86, "grad_norm": 1.7343987906405034, "learning_rate": 1.2794143576458965e-05, "loss": 2.5888, "step": 37920 }, { "epoch": 0.86, "grad_norm": 1.566379886206936, "learning_rate": 1.2790636973676102e-05, "loss": 2.555, "step": 37930 }, { "epoch": 0.86, "grad_norm": 1.6847469472824537, "learning_rate": 1.2787129998729954e-05, "loss": 2.5729, "step": 37940 }, { "epoch": 0.86, "grad_norm": 1.626073267135847, "learning_rate": 1.2783622652088214e-05, "loss": 2.6157, "step": 37950 }, { "epoch": 0.86, "grad_norm": 1.6137550906299867, "learning_rate": 1.2780114934218628e-05, "loss": 2.6102, "step": 37960 }, { "epoch": 0.86, "grad_norm": 1.6455599459747092, "learning_rate": 1.2776606845588985e-05, "loss": 2.5985, "step": 37970 }, { "epoch": 0.86, "grad_norm": 1.57549586185013, "learning_rate": 1.2773098386667138e-05, "loss": 2.7985, "step": 37980 }, { "epoch": 0.86, "grad_norm": 1.7375462250755052, "learning_rate": 1.2769589557920975e-05, "loss": 2.5563, "step": 37990 }, { "epoch": 0.86, "grad_norm": 1.7323085365848707, "learning_rate": 1.2766080359818437e-05, "loss": 2.5362, "step": 38000 }, { "epoch": 0.86, "grad_norm": 1.514421545483348, "learning_rate": 1.2762570792827518e-05, "loss": 2.5829, "step": 38010 }, { "epoch": 0.86, "grad_norm": 2.6684264202715995, "learning_rate": 1.2759060857416258e-05, "loss": 2.6214, "step": 38020 }, { "epoch": 0.86, "grad_norm": 2.141422834976873, "learning_rate": 1.2755550554052747e-05, "loss": 2.6252, "step": 38030 }, { "epoch": 0.86, "grad_norm": 1.6375862135315324, "learning_rate": 1.2752039883205123e-05, "loss": 2.6816, "step": 38040 }, { "epoch": 0.86, "grad_norm": 1.7566191755280793, "learning_rate": 1.2748528845341577e-05, "loss": 2.6441, "step": 38050 }, { "epoch": 0.86, "grad_norm": 1.644383411284264, "learning_rate": 1.274501744093034e-05, "loss": 2.5125, "step": 38060 }, { "epoch": 0.86, "grad_norm": 1.5218481263834325, "learning_rate": 1.2741505670439702e-05, "loss": 2.7723, "step": 38070 }, { "epoch": 0.86, "grad_norm": 1.758663533197599, "learning_rate": 1.2737993534338e-05, "loss": 2.5916, "step": 38080 }, { "epoch": 0.86, "grad_norm": 1.5203800201735684, "learning_rate": 1.2734481033093617e-05, "loss": 2.6271, "step": 38090 }, { "epoch": 0.86, "grad_norm": 2.000618435524465, "learning_rate": 1.2730968167174978e-05, "loss": 2.6668, "step": 38100 }, { "epoch": 0.86, "grad_norm": 1.5112124189415912, "learning_rate": 1.272745493705057e-05, "loss": 2.652, "step": 38110 }, { "epoch": 0.86, "grad_norm": 1.8095520600521446, "learning_rate": 1.2723941343188921e-05, "loss": 2.5948, "step": 38120 }, { "epoch": 0.86, "grad_norm": 1.7572110944896486, "learning_rate": 1.272042738605861e-05, "loss": 2.6505, "step": 38130 }, { "epoch": 0.86, "grad_norm": 1.7353181031936116, "learning_rate": 1.2716913066128261e-05, "loss": 2.4665, "step": 38140 }, { "epoch": 0.86, "grad_norm": 1.6393416188591179, "learning_rate": 1.2713398383866554e-05, "loss": 2.6566, "step": 38150 }, { "epoch": 0.86, "grad_norm": 1.623716362475052, "learning_rate": 1.2709883339742201e-05, "loss": 2.6227, "step": 38160 }, { "epoch": 0.86, "grad_norm": 1.4292138880188985, "learning_rate": 1.2706367934223982e-05, "loss": 2.6806, "step": 38170 }, { "epoch": 0.86, "grad_norm": 1.4924285532535653, "learning_rate": 1.2702852167780717e-05, "loss": 2.5629, "step": 38180 }, { "epoch": 0.86, "grad_norm": 1.4989849502712889, "learning_rate": 1.2699336040881272e-05, "loss": 2.5484, "step": 38190 }, { "epoch": 0.86, "grad_norm": 1.7214347325754766, "learning_rate": 1.2695819553994557e-05, "loss": 2.5582, "step": 38200 }, { "epoch": 0.86, "grad_norm": 1.5002671824269618, "learning_rate": 1.2692302707589546e-05, "loss": 2.6797, "step": 38210 }, { "epoch": 0.86, "grad_norm": 1.589156897782339, "learning_rate": 1.2688785502135241e-05, "loss": 2.5643, "step": 38220 }, { "epoch": 0.86, "grad_norm": 1.5257277891589784, "learning_rate": 1.2685267938100705e-05, "loss": 2.5749, "step": 38230 }, { "epoch": 0.86, "grad_norm": 1.8614604778856467, "learning_rate": 1.2681750015955046e-05, "loss": 2.5924, "step": 38240 }, { "epoch": 0.86, "grad_norm": 1.3854235790203644, "learning_rate": 1.2678231736167422e-05, "loss": 2.5935, "step": 38250 }, { "epoch": 0.86, "grad_norm": 1.6656128901745904, "learning_rate": 1.267471309920703e-05, "loss": 2.5562, "step": 38260 }, { "epoch": 0.86, "grad_norm": 1.5829461483331277, "learning_rate": 1.2671194105543123e-05, "loss": 2.6503, "step": 38270 }, { "epoch": 0.86, "grad_norm": 1.4348285051354082, "learning_rate": 1.2667674755645e-05, "loss": 2.5206, "step": 38280 }, { "epoch": 0.86, "grad_norm": 1.3981971741954806, "learning_rate": 1.2664155049982005e-05, "loss": 2.755, "step": 38290 }, { "epoch": 0.86, "grad_norm": 1.7362137962318795, "learning_rate": 1.266063498902353e-05, "loss": 2.5829, "step": 38300 }, { "epoch": 0.86, "grad_norm": 1.7130731327458444, "learning_rate": 1.265711457323902e-05, "loss": 2.519, "step": 38310 }, { "epoch": 0.86, "grad_norm": 1.6999668757926871, "learning_rate": 1.265359380309796e-05, "loss": 2.6505, "step": 38320 }, { "epoch": 0.86, "grad_norm": 1.4471806389559336, "learning_rate": 1.2650072679069878e-05, "loss": 2.5982, "step": 38330 }, { "epoch": 0.86, "grad_norm": 1.8425435133172654, "learning_rate": 1.2646551201624365e-05, "loss": 2.6133, "step": 38340 }, { "epoch": 0.86, "grad_norm": 1.7056051262610636, "learning_rate": 1.2643029371231048e-05, "loss": 2.5135, "step": 38350 }, { "epoch": 0.87, "grad_norm": 1.4238302289929132, "learning_rate": 1.2639507188359603e-05, "loss": 2.6495, "step": 38360 }, { "epoch": 0.87, "grad_norm": 1.631969151690994, "learning_rate": 1.2635984653479754e-05, "loss": 2.5592, "step": 38370 }, { "epoch": 0.87, "grad_norm": 1.5884605791614823, "learning_rate": 1.2632461767061268e-05, "loss": 2.6876, "step": 38380 }, { "epoch": 0.87, "grad_norm": 1.636330101261771, "learning_rate": 1.2628938529573965e-05, "loss": 2.6181, "step": 38390 }, { "epoch": 0.87, "grad_norm": 1.6667477817292558, "learning_rate": 1.2625414941487706e-05, "loss": 2.5417, "step": 38400 }, { "epoch": 0.87, "grad_norm": 2.2204534527715762, "learning_rate": 1.2621891003272406e-05, "loss": 2.5867, "step": 38410 }, { "epoch": 0.87, "grad_norm": 1.4572638368271429, "learning_rate": 1.2618366715398023e-05, "loss": 2.6756, "step": 38420 }, { "epoch": 0.87, "grad_norm": 1.630395776853344, "learning_rate": 1.2614842078334555e-05, "loss": 2.5745, "step": 38430 }, { "epoch": 0.87, "grad_norm": 1.599798334355677, "learning_rate": 1.2611317092552055e-05, "loss": 2.6002, "step": 38440 }, { "epoch": 0.87, "grad_norm": 1.660298324789749, "learning_rate": 1.2607791758520625e-05, "loss": 2.6549, "step": 38450 }, { "epoch": 0.87, "grad_norm": 1.504012957968365, "learning_rate": 1.2604266076710401e-05, "loss": 2.5894, "step": 38460 }, { "epoch": 0.87, "grad_norm": 1.6234840322722779, "learning_rate": 1.2600740047591577e-05, "loss": 2.6825, "step": 38470 }, { "epoch": 0.87, "grad_norm": 1.6313605770983055, "learning_rate": 1.259721367163439e-05, "loss": 2.5601, "step": 38480 }, { "epoch": 0.87, "grad_norm": 1.7301780927322972, "learning_rate": 1.2593686949309119e-05, "loss": 2.6255, "step": 38490 }, { "epoch": 0.87, "grad_norm": 1.4521768274395455, "learning_rate": 1.2590159881086094e-05, "loss": 2.5489, "step": 38500 }, { "epoch": 0.87, "grad_norm": 1.7561332560316005, "learning_rate": 1.2586632467435693e-05, "loss": 2.5178, "step": 38510 }, { "epoch": 0.87, "grad_norm": 1.886698068355517, "learning_rate": 1.2583104708828333e-05, "loss": 2.5581, "step": 38520 }, { "epoch": 0.87, "grad_norm": 1.7161923814928899, "learning_rate": 1.2579576605734482e-05, "loss": 2.4975, "step": 38530 }, { "epoch": 0.87, "grad_norm": 1.6312506452678868, "learning_rate": 1.2576048158624657e-05, "loss": 2.6288, "step": 38540 }, { "epoch": 0.87, "grad_norm": 1.6164151237306386, "learning_rate": 1.2572519367969409e-05, "loss": 2.5974, "step": 38550 }, { "epoch": 0.87, "grad_norm": 1.5200446914228565, "learning_rate": 1.2568990234239347e-05, "loss": 2.5578, "step": 38560 }, { "epoch": 0.87, "grad_norm": 1.4585307215420424, "learning_rate": 1.2565460757905119e-05, "loss": 2.597, "step": 38570 }, { "epoch": 0.87, "grad_norm": 1.6863771502892597, "learning_rate": 1.2561930939437426e-05, "loss": 2.5767, "step": 38580 }, { "epoch": 0.87, "grad_norm": 1.568387478842752, "learning_rate": 1.2558400779307003e-05, "loss": 2.5363, "step": 38590 }, { "epoch": 0.87, "grad_norm": 1.6194995510012864, "learning_rate": 1.2554870277984641e-05, "loss": 2.6455, "step": 38600 }, { "epoch": 0.87, "grad_norm": 1.5789463700561963, "learning_rate": 1.255133943594117e-05, "loss": 2.3972, "step": 38610 }, { "epoch": 0.87, "grad_norm": 1.4657240178917814, "learning_rate": 1.2547808253647473e-05, "loss": 2.6851, "step": 38620 }, { "epoch": 0.87, "grad_norm": 1.8214340161052847, "learning_rate": 1.254427673157447e-05, "loss": 2.4859, "step": 38630 }, { "epoch": 0.87, "grad_norm": 1.9440650530250836, "learning_rate": 1.2540744870193129e-05, "loss": 2.6517, "step": 38640 }, { "epoch": 0.87, "grad_norm": 1.8557762163543872, "learning_rate": 1.2537212669974466e-05, "loss": 2.629, "step": 38650 }, { "epoch": 0.87, "grad_norm": 1.7491182091671886, "learning_rate": 1.2533680131389537e-05, "loss": 2.6099, "step": 38660 }, { "epoch": 0.87, "grad_norm": 1.6789877116204037, "learning_rate": 1.253014725490945e-05, "loss": 2.6166, "step": 38670 }, { "epoch": 0.87, "grad_norm": 1.5822073911867571, "learning_rate": 1.2526614041005354e-05, "loss": 2.5606, "step": 38680 }, { "epoch": 0.87, "grad_norm": 1.6737627758809643, "learning_rate": 1.2523080490148443e-05, "loss": 2.6587, "step": 38690 }, { "epoch": 0.87, "grad_norm": 1.43660800323104, "learning_rate": 1.2519546602809954e-05, "loss": 2.502, "step": 38700 }, { "epoch": 0.87, "grad_norm": 1.611154140671375, "learning_rate": 1.2516012379461173e-05, "loss": 2.6414, "step": 38710 }, { "epoch": 0.87, "grad_norm": 1.6319582985887813, "learning_rate": 1.2512477820573432e-05, "loss": 2.5958, "step": 38720 }, { "epoch": 0.87, "grad_norm": 1.5806920410049086, "learning_rate": 1.2508942926618098e-05, "loss": 2.605, "step": 38730 }, { "epoch": 0.87, "grad_norm": 1.6882217124631151, "learning_rate": 1.2505407698066596e-05, "loss": 2.5849, "step": 38740 }, { "epoch": 0.87, "grad_norm": 1.5877146169766019, "learning_rate": 1.2501872135390385e-05, "loss": 2.6147, "step": 38750 }, { "epoch": 0.87, "grad_norm": 1.8050059081131913, "learning_rate": 1.2498336239060974e-05, "loss": 2.5759, "step": 38760 }, { "epoch": 0.87, "grad_norm": 1.6078493605283124, "learning_rate": 1.2494800009549915e-05, "loss": 2.5248, "step": 38770 }, { "epoch": 0.87, "grad_norm": 1.4860950431979525, "learning_rate": 1.2491263447328803e-05, "loss": 2.5965, "step": 38780 }, { "epoch": 0.87, "grad_norm": 1.5080185770311547, "learning_rate": 1.2487726552869284e-05, "loss": 2.6453, "step": 38790 }, { "epoch": 0.87, "grad_norm": 1.589146221971707, "learning_rate": 1.2484189326643036e-05, "loss": 2.5143, "step": 38800 }, { "epoch": 0.88, "grad_norm": 1.4889646322053125, "learning_rate": 1.2480651769121796e-05, "loss": 2.4905, "step": 38810 }, { "epoch": 0.88, "grad_norm": 1.6258346266680446, "learning_rate": 1.2477113880777332e-05, "loss": 2.4952, "step": 38820 }, { "epoch": 0.88, "grad_norm": 1.9178194504845096, "learning_rate": 1.2473575662081464e-05, "loss": 2.5856, "step": 38830 }, { "epoch": 0.88, "grad_norm": 1.5065742455560145, "learning_rate": 1.2470037113506053e-05, "loss": 2.4459, "step": 38840 }, { "epoch": 0.88, "grad_norm": 1.5435412569225966, "learning_rate": 1.2466498235523008e-05, "loss": 2.5778, "step": 38850 }, { "epoch": 0.88, "grad_norm": 1.6473303603726053, "learning_rate": 1.246295902860427e-05, "loss": 2.5898, "step": 38860 }, { "epoch": 0.88, "grad_norm": 1.9503716170312402, "learning_rate": 1.2459419493221846e-05, "loss": 2.6266, "step": 38870 }, { "epoch": 0.88, "grad_norm": 1.7675263678367523, "learning_rate": 1.2455879629847763e-05, "loss": 2.5875, "step": 38880 }, { "epoch": 0.88, "grad_norm": 1.7666642246177264, "learning_rate": 1.2452339438954104e-05, "loss": 2.6337, "step": 38890 }, { "epoch": 0.88, "grad_norm": 1.586258323040874, "learning_rate": 1.2448798921012996e-05, "loss": 2.7166, "step": 38900 }, { "epoch": 0.88, "grad_norm": 1.6386408651239222, "learning_rate": 1.2445258076496609e-05, "loss": 2.5413, "step": 38910 }, { "epoch": 0.88, "grad_norm": 1.6109062838087465, "learning_rate": 1.2441716905877153e-05, "loss": 2.5042, "step": 38920 }, { "epoch": 0.88, "grad_norm": 1.6168204910006663, "learning_rate": 1.2438175409626879e-05, "loss": 2.6015, "step": 38930 }, { "epoch": 0.88, "grad_norm": 1.7853478727342151, "learning_rate": 1.2434633588218093e-05, "loss": 2.6187, "step": 38940 }, { "epoch": 0.88, "grad_norm": 1.5226293851804862, "learning_rate": 1.2431091442123134e-05, "loss": 2.5262, "step": 38950 }, { "epoch": 0.88, "grad_norm": 1.5783620777404768, "learning_rate": 1.2427548971814387e-05, "loss": 2.7001, "step": 38960 }, { "epoch": 0.88, "grad_norm": 1.4349624398426033, "learning_rate": 1.2424006177764285e-05, "loss": 2.6341, "step": 38970 }, { "epoch": 0.88, "grad_norm": 1.6693312143242387, "learning_rate": 1.2420463060445296e-05, "loss": 2.6609, "step": 38980 }, { "epoch": 0.88, "grad_norm": 1.5980736878962463, "learning_rate": 1.2416919620329936e-05, "loss": 2.5734, "step": 38990 }, { "epoch": 0.88, "grad_norm": 1.471530416009596, "learning_rate": 1.241337585789076e-05, "loss": 2.5867, "step": 39000 }, { "epoch": 0.88, "grad_norm": 1.5211044461241514, "learning_rate": 1.2409831773600375e-05, "loss": 2.6562, "step": 39010 }, { "epoch": 0.88, "grad_norm": 1.5902246838663474, "learning_rate": 1.2406287367931425e-05, "loss": 2.6734, "step": 39020 }, { "epoch": 0.88, "grad_norm": 1.5848851366735714, "learning_rate": 1.240274264135659e-05, "loss": 2.6149, "step": 39030 }, { "epoch": 0.88, "grad_norm": 1.6818040335552962, "learning_rate": 1.2399197594348606e-05, "loss": 2.6027, "step": 39040 }, { "epoch": 0.88, "grad_norm": 1.6543671349783071, "learning_rate": 1.2395652227380244e-05, "loss": 2.4892, "step": 39050 }, { "epoch": 0.88, "grad_norm": 1.6186100782271744, "learning_rate": 1.239210654092432e-05, "loss": 2.6183, "step": 39060 }, { "epoch": 0.88, "grad_norm": 1.5989285471940937, "learning_rate": 1.238856053545369e-05, "loss": 2.637, "step": 39070 }, { "epoch": 0.88, "grad_norm": 1.602987788285078, "learning_rate": 1.2385014211441252e-05, "loss": 2.6184, "step": 39080 }, { "epoch": 0.88, "grad_norm": 1.545892210060524, "learning_rate": 1.2381467569359955e-05, "loss": 2.6013, "step": 39090 }, { "epoch": 0.88, "grad_norm": 1.6585518283931333, "learning_rate": 1.2377920609682779e-05, "loss": 2.6647, "step": 39100 }, { "epoch": 0.88, "grad_norm": 1.7514830758547602, "learning_rate": 1.2374373332882755e-05, "loss": 2.5809, "step": 39110 }, { "epoch": 0.88, "grad_norm": 2.044911713772067, "learning_rate": 1.2370825739432953e-05, "loss": 2.6106, "step": 39120 }, { "epoch": 0.88, "grad_norm": 1.564255989772877, "learning_rate": 1.2367277829806477e-05, "loss": 2.5897, "step": 39130 }, { "epoch": 0.88, "grad_norm": 1.7431601870147042, "learning_rate": 1.2363729604476492e-05, "loss": 2.6533, "step": 39140 }, { "epoch": 0.88, "grad_norm": 1.4953303081464484, "learning_rate": 1.2360181063916188e-05, "loss": 2.5507, "step": 39150 }, { "epoch": 0.88, "grad_norm": 1.610661193359639, "learning_rate": 1.2356632208598804e-05, "loss": 2.559, "step": 39160 }, { "epoch": 0.88, "grad_norm": 1.6047494090169132, "learning_rate": 1.2353083038997626e-05, "loss": 2.6178, "step": 39170 }, { "epoch": 0.88, "grad_norm": 1.707865218044789, "learning_rate": 1.234953355558597e-05, "loss": 2.5615, "step": 39180 }, { "epoch": 0.88, "grad_norm": 1.8211263056139857, "learning_rate": 1.23459837588372e-05, "loss": 2.5882, "step": 39190 }, { "epoch": 0.88, "grad_norm": 1.570667225139642, "learning_rate": 1.2342433649224727e-05, "loss": 2.5837, "step": 39200 }, { "epoch": 0.88, "grad_norm": 1.61823190300801, "learning_rate": 1.2338883227221994e-05, "loss": 2.662, "step": 39210 }, { "epoch": 0.88, "grad_norm": 1.6829697401960613, "learning_rate": 1.2335332493302493e-05, "loss": 2.5442, "step": 39220 }, { "epoch": 0.88, "grad_norm": 1.6886366676656481, "learning_rate": 1.2331781447939753e-05, "loss": 2.5938, "step": 39230 }, { "epoch": 0.88, "grad_norm": 2.117356049312476, "learning_rate": 1.2328230091607347e-05, "loss": 2.4705, "step": 39240 }, { "epoch": 0.89, "grad_norm": 1.4751774611356765, "learning_rate": 1.2324678424778889e-05, "loss": 2.5311, "step": 39250 }, { "epoch": 0.89, "grad_norm": 2.0107518760287975, "learning_rate": 1.2321126447928033e-05, "loss": 2.5044, "step": 39260 }, { "epoch": 0.89, "grad_norm": 1.6355698372278338, "learning_rate": 1.2317574161528478e-05, "loss": 2.5984, "step": 39270 }, { "epoch": 0.89, "grad_norm": 1.7285741376106631, "learning_rate": 1.2314021566053963e-05, "loss": 2.5155, "step": 39280 }, { "epoch": 0.89, "grad_norm": 1.6711902788803175, "learning_rate": 1.2310468661978264e-05, "loss": 2.6927, "step": 39290 }, { "epoch": 0.89, "grad_norm": 1.4842887076652282, "learning_rate": 1.2306915449775201e-05, "loss": 2.6496, "step": 39300 }, { "epoch": 0.89, "grad_norm": 1.6145937766160137, "learning_rate": 1.2303361929918636e-05, "loss": 2.5909, "step": 39310 }, { "epoch": 0.89, "grad_norm": 1.8993942763150862, "learning_rate": 1.2299808102882476e-05, "loss": 2.5209, "step": 39320 }, { "epoch": 0.89, "grad_norm": 1.4717937876109004, "learning_rate": 1.2296253969140655e-05, "loss": 2.5229, "step": 39330 }, { "epoch": 0.89, "grad_norm": 2.0736608980180913, "learning_rate": 1.229269952916717e-05, "loss": 2.6048, "step": 39340 }, { "epoch": 0.89, "grad_norm": 1.7329979030060758, "learning_rate": 1.2289144783436034e-05, "loss": 2.607, "step": 39350 }, { "epoch": 0.89, "grad_norm": 1.8037645778480176, "learning_rate": 1.2285589732421317e-05, "loss": 2.4727, "step": 39360 }, { "epoch": 0.89, "grad_norm": 1.518385992490353, "learning_rate": 1.2282034376597127e-05, "loss": 2.6158, "step": 39370 }, { "epoch": 0.89, "grad_norm": 1.4494907327641318, "learning_rate": 1.2278478716437613e-05, "loss": 2.6915, "step": 39380 }, { "epoch": 0.89, "grad_norm": 1.6546439576821315, "learning_rate": 1.227492275241696e-05, "loss": 2.6038, "step": 39390 }, { "epoch": 0.89, "grad_norm": 1.6111886727186642, "learning_rate": 1.2271366485009396e-05, "loss": 2.649, "step": 39400 }, { "epoch": 0.89, "grad_norm": 1.3984402756635073, "learning_rate": 1.2267809914689195e-05, "loss": 2.523, "step": 39410 }, { "epoch": 0.89, "grad_norm": 1.8684210007988629, "learning_rate": 1.2264253041930659e-05, "loss": 2.6524, "step": 39420 }, { "epoch": 0.89, "grad_norm": 1.88162472264548, "learning_rate": 1.226069586720814e-05, "loss": 2.5785, "step": 39430 }, { "epoch": 0.89, "grad_norm": 1.6849576336686456, "learning_rate": 1.225713839099603e-05, "loss": 2.613, "step": 39440 }, { "epoch": 0.89, "grad_norm": 1.4783054657190622, "learning_rate": 1.2253580613768757e-05, "loss": 2.6647, "step": 39450 }, { "epoch": 0.89, "grad_norm": 1.6425135781750155, "learning_rate": 1.225002253600079e-05, "loss": 2.7253, "step": 39460 }, { "epoch": 0.89, "grad_norm": 1.5395930479276874, "learning_rate": 1.2246464158166642e-05, "loss": 2.7329, "step": 39470 }, { "epoch": 0.89, "grad_norm": 1.706325509220927, "learning_rate": 1.2242905480740864e-05, "loss": 2.591, "step": 39480 }, { "epoch": 0.89, "grad_norm": 1.5371542465623296, "learning_rate": 1.2239346504198045e-05, "loss": 2.6324, "step": 39490 }, { "epoch": 0.89, "grad_norm": 1.75065121563172, "learning_rate": 1.2235787229012811e-05, "loss": 2.6533, "step": 39500 }, { "epoch": 0.89, "grad_norm": 1.7414461302345132, "learning_rate": 1.223222765565984e-05, "loss": 2.6462, "step": 39510 }, { "epoch": 0.89, "grad_norm": 1.8236659789467644, "learning_rate": 1.2228667784613838e-05, "loss": 2.5314, "step": 39520 }, { "epoch": 0.89, "grad_norm": 1.6168546213814008, "learning_rate": 1.2225107616349548e-05, "loss": 2.5906, "step": 39530 }, { "epoch": 0.89, "grad_norm": 1.5976041757262105, "learning_rate": 1.2221547151341769e-05, "loss": 2.6706, "step": 39540 }, { "epoch": 0.89, "grad_norm": 1.4914545083129125, "learning_rate": 1.2217986390065325e-05, "loss": 2.4733, "step": 39550 }, { "epoch": 0.89, "grad_norm": 1.656009086563263, "learning_rate": 1.2214425332995084e-05, "loss": 2.6381, "step": 39560 }, { "epoch": 0.89, "grad_norm": 1.5366483956327068, "learning_rate": 1.2210863980605954e-05, "loss": 2.5938, "step": 39570 }, { "epoch": 0.89, "grad_norm": 1.4860260563306724, "learning_rate": 1.2207302333372884e-05, "loss": 2.5388, "step": 39580 }, { "epoch": 0.89, "grad_norm": 1.794965111800645, "learning_rate": 1.2203740391770857e-05, "loss": 2.5374, "step": 39590 }, { "epoch": 0.89, "grad_norm": 1.5369119132040192, "learning_rate": 1.2200178156274902e-05, "loss": 2.6651, "step": 39600 }, { "epoch": 0.89, "grad_norm": 1.659620697831278, "learning_rate": 1.2196615627360082e-05, "loss": 2.7915, "step": 39610 }, { "epoch": 0.89, "grad_norm": 1.5976416581140251, "learning_rate": 1.21930528055015e-05, "loss": 2.597, "step": 39620 }, { "epoch": 0.89, "grad_norm": 1.554266688619721, "learning_rate": 1.2189489691174296e-05, "loss": 2.6231, "step": 39630 }, { "epoch": 0.89, "grad_norm": 1.5739867253956434, "learning_rate": 1.2185926284853659e-05, "loss": 2.6656, "step": 39640 }, { "epoch": 0.89, "grad_norm": 1.6589720869294362, "learning_rate": 1.2182362587014808e-05, "loss": 2.5796, "step": 39650 }, { "epoch": 0.89, "grad_norm": 1.6009352454638153, "learning_rate": 1.2178798598133e-05, "loss": 2.5814, "step": 39660 }, { "epoch": 0.89, "grad_norm": 1.8737817127875083, "learning_rate": 1.2175234318683535e-05, "loss": 2.5383, "step": 39670 }, { "epoch": 0.89, "grad_norm": 1.6481699766029065, "learning_rate": 1.217166974914175e-05, "loss": 2.6818, "step": 39680 }, { "epoch": 0.9, "grad_norm": 1.6988299562331048, "learning_rate": 1.216810488998302e-05, "loss": 2.6444, "step": 39690 }, { "epoch": 0.9, "grad_norm": 1.4525014715547815, "learning_rate": 1.2164539741682764e-05, "loss": 2.5338, "step": 39700 }, { "epoch": 0.9, "grad_norm": 1.68055144152865, "learning_rate": 1.2160974304716433e-05, "loss": 2.5994, "step": 39710 }, { "epoch": 0.9, "grad_norm": 1.6368392058157228, "learning_rate": 1.2157408579559512e-05, "loss": 2.4986, "step": 39720 }, { "epoch": 0.9, "grad_norm": 1.5221215759712559, "learning_rate": 1.2153842566687544e-05, "loss": 2.5771, "step": 39730 }, { "epoch": 0.9, "grad_norm": 1.9067343573432132, "learning_rate": 1.2150276266576087e-05, "loss": 2.5809, "step": 39740 }, { "epoch": 0.9, "grad_norm": 1.7183707661869803, "learning_rate": 1.2146709679700751e-05, "loss": 2.5847, "step": 39750 }, { "epoch": 0.9, "grad_norm": 1.552845890511578, "learning_rate": 1.214314280653718e-05, "loss": 2.5717, "step": 39760 }, { "epoch": 0.9, "grad_norm": 1.57780192605063, "learning_rate": 1.2139575647561064e-05, "loss": 2.8023, "step": 39770 }, { "epoch": 0.9, "grad_norm": 1.6953072822252178, "learning_rate": 1.2136008203248118e-05, "loss": 2.473, "step": 39780 }, { "epoch": 0.9, "grad_norm": 1.6586223508210836, "learning_rate": 1.2132440474074097e-05, "loss": 2.6019, "step": 39790 }, { "epoch": 0.9, "grad_norm": 1.5492665143997086, "learning_rate": 1.2128872460514807e-05, "loss": 2.7083, "step": 39800 }, { "epoch": 0.9, "grad_norm": 1.7809132488454325, "learning_rate": 1.2125304163046082e-05, "loss": 2.6291, "step": 39810 }, { "epoch": 0.9, "grad_norm": 1.6875369955549124, "learning_rate": 1.212173558214379e-05, "loss": 2.6303, "step": 39820 }, { "epoch": 0.9, "grad_norm": 1.876573880785748, "learning_rate": 1.2118166718283848e-05, "loss": 2.5673, "step": 39830 }, { "epoch": 0.9, "grad_norm": 1.7491701187062554, "learning_rate": 1.2114597571942202e-05, "loss": 2.4639, "step": 39840 }, { "epoch": 0.9, "grad_norm": 1.4996301676335801, "learning_rate": 1.2111028143594836e-05, "loss": 2.6179, "step": 39850 }, { "epoch": 0.9, "grad_norm": 1.9214374305882975, "learning_rate": 1.2107458433717778e-05, "loss": 2.6783, "step": 39860 }, { "epoch": 0.9, "grad_norm": 1.86842355930924, "learning_rate": 1.2103888442787091e-05, "loss": 2.6148, "step": 39870 }, { "epoch": 0.9, "grad_norm": 1.754807246721506, "learning_rate": 1.2100318171278867e-05, "loss": 2.6552, "step": 39880 }, { "epoch": 0.9, "grad_norm": 1.5619244810612243, "learning_rate": 1.2096747619669244e-05, "loss": 2.5569, "step": 39890 }, { "epoch": 0.9, "grad_norm": 1.577539185736928, "learning_rate": 1.2093176788434404e-05, "loss": 2.5821, "step": 39900 }, { "epoch": 0.9, "grad_norm": 1.585842798602665, "learning_rate": 1.208960567805055e-05, "loss": 2.5755, "step": 39910 }, { "epoch": 0.9, "grad_norm": 2.098091938858193, "learning_rate": 1.2086034288993934e-05, "loss": 2.5557, "step": 39920 }, { "epoch": 0.9, "grad_norm": 1.7958512038584598, "learning_rate": 1.2082462621740838e-05, "loss": 2.6407, "step": 39930 }, { "epoch": 0.9, "grad_norm": 1.5397001223787794, "learning_rate": 1.207889067676759e-05, "loss": 2.561, "step": 39940 }, { "epoch": 0.9, "grad_norm": 1.4369332743491132, "learning_rate": 1.2075318454550545e-05, "loss": 2.6877, "step": 39950 }, { "epoch": 0.9, "grad_norm": 1.477988832115339, "learning_rate": 1.2071745955566096e-05, "loss": 2.6238, "step": 39960 }, { "epoch": 0.9, "grad_norm": 1.4813785831130815, "learning_rate": 1.2068173180290687e-05, "loss": 2.7233, "step": 39970 }, { "epoch": 0.9, "grad_norm": 1.7911922477160824, "learning_rate": 1.2064600129200785e-05, "loss": 2.6091, "step": 39980 }, { "epoch": 0.9, "grad_norm": 1.603558262023106, "learning_rate": 1.2061026802772889e-05, "loss": 2.7034, "step": 39990 }, { "epoch": 0.9, "grad_norm": 1.4142183719668813, "learning_rate": 1.2057453201483553e-05, "loss": 2.5444, "step": 40000 }, { "epoch": 0.9, "grad_norm": 1.875073920465265, "learning_rate": 1.2053879325809352e-05, "loss": 2.5314, "step": 40010 }, { "epoch": 0.9, "grad_norm": 1.8549676286715797, "learning_rate": 1.2050305176226905e-05, "loss": 2.7207, "step": 40020 }, { "epoch": 0.9, "grad_norm": 1.697519533289966, "learning_rate": 1.2046730753212862e-05, "loss": 2.6337, "step": 40030 }, { "epoch": 0.9, "grad_norm": 1.7624711589354636, "learning_rate": 1.2043156057243921e-05, "loss": 2.4931, "step": 40040 }, { "epoch": 0.9, "grad_norm": 1.5265063767814444, "learning_rate": 1.20395810887968e-05, "loss": 2.5915, "step": 40050 }, { "epoch": 0.9, "grad_norm": 1.6676002635531593, "learning_rate": 1.2036005848348267e-05, "loss": 2.635, "step": 40060 }, { "epoch": 0.9, "grad_norm": 1.5324309971738173, "learning_rate": 1.203243033637512e-05, "loss": 2.5944, "step": 40070 }, { "epoch": 0.9, "grad_norm": 1.534263085687828, "learning_rate": 1.2028854553354195e-05, "loss": 2.5736, "step": 40080 }, { "epoch": 0.9, "grad_norm": 1.4827068858372667, "learning_rate": 1.2025278499762361e-05, "loss": 2.5231, "step": 40090 }, { "epoch": 0.9, "grad_norm": 1.7538696937201366, "learning_rate": 1.2021702176076526e-05, "loss": 2.6522, "step": 40100 }, { "epoch": 0.9, "grad_norm": 1.5115780733862136, "learning_rate": 1.2018125582773636e-05, "loss": 2.4899, "step": 40110 }, { "epoch": 0.9, "grad_norm": 1.6293377309021269, "learning_rate": 1.201454872033067e-05, "loss": 2.6752, "step": 40120 }, { "epoch": 0.9, "grad_norm": 1.717938193689583, "learning_rate": 1.201097158922464e-05, "loss": 2.5965, "step": 40130 }, { "epoch": 0.91, "grad_norm": 1.6819800848097057, "learning_rate": 1.2007394189932605e-05, "loss": 2.6436, "step": 40140 }, { "epoch": 0.91, "grad_norm": 1.573441571023158, "learning_rate": 1.2003816522931642e-05, "loss": 2.57, "step": 40150 }, { "epoch": 0.91, "grad_norm": 1.5332814993516448, "learning_rate": 1.2000238588698879e-05, "loss": 2.6064, "step": 40160 }, { "epoch": 0.91, "grad_norm": 1.661872225153804, "learning_rate": 1.1996660387711473e-05, "loss": 2.7484, "step": 40170 }, { "epoch": 0.91, "grad_norm": 1.516453099825259, "learning_rate": 1.1993081920446621e-05, "loss": 2.5949, "step": 40180 }, { "epoch": 0.91, "grad_norm": 1.862224169967771, "learning_rate": 1.1989503187381548e-05, "loss": 2.6756, "step": 40190 }, { "epoch": 0.91, "grad_norm": 1.7016197019534345, "learning_rate": 1.1985924188993527e-05, "loss": 2.6081, "step": 40200 }, { "epoch": 0.91, "grad_norm": 1.4719566031985316, "learning_rate": 1.1982344925759848e-05, "loss": 2.6196, "step": 40210 }, { "epoch": 0.91, "grad_norm": 1.6961622577714002, "learning_rate": 1.197876539815785e-05, "loss": 2.6296, "step": 40220 }, { "epoch": 0.91, "grad_norm": 1.5805999995481756, "learning_rate": 1.1975185606664905e-05, "loss": 2.6243, "step": 40230 }, { "epoch": 0.91, "grad_norm": 1.5507801578691904, "learning_rate": 1.197160555175842e-05, "loss": 2.6473, "step": 40240 }, { "epoch": 0.91, "grad_norm": 1.6419554637089517, "learning_rate": 1.1968025233915839e-05, "loss": 2.7141, "step": 40250 }, { "epoch": 0.91, "grad_norm": 1.6064816811781066, "learning_rate": 1.1964444653614631e-05, "loss": 2.6165, "step": 40260 }, { "epoch": 0.91, "grad_norm": 1.6793689184032672, "learning_rate": 1.196086381133231e-05, "loss": 2.5474, "step": 40270 }, { "epoch": 0.91, "grad_norm": 1.8208157258177433, "learning_rate": 1.1957282707546423e-05, "loss": 2.7045, "step": 40280 }, { "epoch": 0.91, "grad_norm": 1.493027404593929, "learning_rate": 1.195370134273455e-05, "loss": 2.6724, "step": 40290 }, { "epoch": 0.91, "grad_norm": 1.8389852021113928, "learning_rate": 1.1950119717374309e-05, "loss": 2.5776, "step": 40300 }, { "epoch": 0.91, "grad_norm": 1.5861722569189807, "learning_rate": 1.1946537831943353e-05, "loss": 2.7308, "step": 40310 }, { "epoch": 0.91, "grad_norm": 1.7453316884594812, "learning_rate": 1.1942955686919357e-05, "loss": 2.5495, "step": 40320 }, { "epoch": 0.91, "grad_norm": 1.9432360604877732, "learning_rate": 1.1939373282780051e-05, "loss": 2.6752, "step": 40330 }, { "epoch": 0.91, "grad_norm": 2.2738688599695553, "learning_rate": 1.1935790620003185e-05, "loss": 2.6292, "step": 40340 }, { "epoch": 0.91, "grad_norm": 1.6618575041176997, "learning_rate": 1.1932207699066552e-05, "loss": 2.5045, "step": 40350 }, { "epoch": 0.91, "grad_norm": 1.6073923487288435, "learning_rate": 1.1928624520447968e-05, "loss": 2.7119, "step": 40360 }, { "epoch": 0.91, "grad_norm": 1.7265133050289967, "learning_rate": 1.1925041084625297e-05, "loss": 2.6744, "step": 40370 }, { "epoch": 0.91, "grad_norm": 1.56266895260536, "learning_rate": 1.192145739207643e-05, "loss": 2.5857, "step": 40380 }, { "epoch": 0.91, "grad_norm": 1.530610955259697, "learning_rate": 1.1917873443279292e-05, "loss": 2.6434, "step": 40390 }, { "epoch": 0.91, "grad_norm": 1.9468690110872846, "learning_rate": 1.1914289238711842e-05, "loss": 2.5917, "step": 40400 }, { "epoch": 0.91, "grad_norm": 1.5160979182044916, "learning_rate": 1.1910704778852082e-05, "loss": 2.5797, "step": 40410 }, { "epoch": 0.91, "grad_norm": 1.7857763223778884, "learning_rate": 1.1907120064178031e-05, "loss": 2.4805, "step": 40420 }, { "epoch": 0.91, "grad_norm": 1.4585933079517288, "learning_rate": 1.1903535095167756e-05, "loss": 2.5753, "step": 40430 }, { "epoch": 0.91, "grad_norm": 1.4458521383294776, "learning_rate": 1.1899949872299352e-05, "loss": 2.5365, "step": 40440 }, { "epoch": 0.91, "grad_norm": 1.654191745461678, "learning_rate": 1.1896364396050954e-05, "loss": 2.63, "step": 40450 }, { "epoch": 0.91, "grad_norm": 1.6435898757187861, "learning_rate": 1.1892778666900719e-05, "loss": 2.6565, "step": 40460 }, { "epoch": 0.91, "grad_norm": 1.6192422527160353, "learning_rate": 1.1889192685326854e-05, "loss": 2.6798, "step": 40470 }, { "epoch": 0.91, "grad_norm": 1.9384037516069041, "learning_rate": 1.1885606451807582e-05, "loss": 2.598, "step": 40480 }, { "epoch": 0.91, "grad_norm": 1.7434616693751768, "learning_rate": 1.1882019966821173e-05, "loss": 2.6056, "step": 40490 }, { "epoch": 0.91, "grad_norm": 1.6348449810205166, "learning_rate": 1.1878433230845923e-05, "loss": 2.5656, "step": 40500 }, { "epoch": 0.91, "grad_norm": 1.8254250685713136, "learning_rate": 1.1874846244360165e-05, "loss": 2.5503, "step": 40510 }, { "epoch": 0.91, "grad_norm": 1.5131407353547381, "learning_rate": 1.1871259007842267e-05, "loss": 2.5921, "step": 40520 }, { "epoch": 0.91, "grad_norm": 1.432094201408284, "learning_rate": 1.1867671521770622e-05, "loss": 2.5649, "step": 40530 }, { "epoch": 0.91, "grad_norm": 1.7285598629488796, "learning_rate": 1.186408378662367e-05, "loss": 2.6005, "step": 40540 }, { "epoch": 0.91, "grad_norm": 1.6425656481802282, "learning_rate": 1.1860495802879872e-05, "loss": 2.6795, "step": 40550 }, { "epoch": 0.91, "grad_norm": 1.5653974701665085, "learning_rate": 1.1856907571017725e-05, "loss": 2.6101, "step": 40560 }, { "epoch": 0.91, "grad_norm": 1.5335459552812951, "learning_rate": 1.1853319091515765e-05, "loss": 2.6465, "step": 40570 }, { "epoch": 0.92, "grad_norm": 1.5113569404414817, "learning_rate": 1.1849730364852557e-05, "loss": 2.6674, "step": 40580 }, { "epoch": 0.92, "grad_norm": 1.7278995190930764, "learning_rate": 1.1846141391506689e-05, "loss": 2.6509, "step": 40590 }, { "epoch": 0.92, "grad_norm": 1.5661402230791612, "learning_rate": 1.1842552171956802e-05, "loss": 2.6273, "step": 40600 }, { "epoch": 0.92, "grad_norm": 1.7768577629408175, "learning_rate": 1.1838962706681559e-05, "loss": 2.6338, "step": 40610 }, { "epoch": 0.92, "grad_norm": 1.5470554921950188, "learning_rate": 1.1835372996159649e-05, "loss": 2.5856, "step": 40620 }, { "epoch": 0.92, "grad_norm": 1.6488067272520244, "learning_rate": 1.1831783040869808e-05, "loss": 2.5426, "step": 40630 }, { "epoch": 0.92, "grad_norm": 1.8547399922150178, "learning_rate": 1.1828192841290793e-05, "loss": 2.4401, "step": 40640 }, { "epoch": 0.92, "grad_norm": 1.6094902099671635, "learning_rate": 1.1824602397901402e-05, "loss": 2.6364, "step": 40650 }, { "epoch": 0.92, "grad_norm": 1.614189014313251, "learning_rate": 1.1821011711180454e-05, "loss": 2.5564, "step": 40660 }, { "epoch": 0.92, "grad_norm": 1.5929076199591738, "learning_rate": 1.1817420781606819e-05, "loss": 2.5904, "step": 40670 }, { "epoch": 0.92, "grad_norm": 1.4104892877871824, "learning_rate": 1.1813829609659382e-05, "loss": 2.5221, "step": 40680 }, { "epoch": 0.92, "grad_norm": 1.6458351840940257, "learning_rate": 1.1810238195817063e-05, "loss": 2.7554, "step": 40690 }, { "epoch": 0.92, "grad_norm": 1.7535433512706307, "learning_rate": 1.1806646540558826e-05, "loss": 2.6544, "step": 40700 }, { "epoch": 0.92, "grad_norm": 2.199622619395622, "learning_rate": 1.1803054644363657e-05, "loss": 2.6189, "step": 40710 }, { "epoch": 0.92, "grad_norm": 1.6643999564551257, "learning_rate": 1.179946250771057e-05, "loss": 2.5243, "step": 40720 }, { "epoch": 0.92, "grad_norm": 1.625996312559519, "learning_rate": 1.1795870131078626e-05, "loss": 2.4958, "step": 40730 }, { "epoch": 0.92, "grad_norm": 1.575021529703568, "learning_rate": 1.179227751494691e-05, "loss": 2.5571, "step": 40740 }, { "epoch": 0.92, "grad_norm": 1.7071474671318472, "learning_rate": 1.1788684659794528e-05, "loss": 2.6066, "step": 40750 }, { "epoch": 0.92, "grad_norm": 1.7435351538956543, "learning_rate": 1.1785091566100637e-05, "loss": 2.667, "step": 40760 }, { "epoch": 0.92, "grad_norm": 1.5444888871573152, "learning_rate": 1.1781498234344415e-05, "loss": 2.6634, "step": 40770 }, { "epoch": 0.92, "grad_norm": 1.543168683865853, "learning_rate": 1.1777904665005071e-05, "loss": 2.5821, "step": 40780 }, { "epoch": 0.92, "grad_norm": 1.6981188633047626, "learning_rate": 1.1774310858561854e-05, "loss": 2.5626, "step": 40790 }, { "epoch": 0.92, "grad_norm": 1.8038286192413782, "learning_rate": 1.1770716815494034e-05, "loss": 2.5635, "step": 40800 }, { "epoch": 0.92, "grad_norm": 1.6346704740762976, "learning_rate": 1.1767122536280921e-05, "loss": 2.6168, "step": 40810 }, { "epoch": 0.92, "grad_norm": 1.6222330548728727, "learning_rate": 1.176352802140185e-05, "loss": 2.604, "step": 40820 }, { "epoch": 0.92, "grad_norm": 1.6525126318391952, "learning_rate": 1.1759933271336192e-05, "loss": 2.6387, "step": 40830 }, { "epoch": 0.92, "grad_norm": 1.905171251718224, "learning_rate": 1.1756338286563352e-05, "loss": 2.528, "step": 40840 }, { "epoch": 0.92, "grad_norm": 1.6950160834545904, "learning_rate": 1.1752743067562756e-05, "loss": 2.5966, "step": 40850 }, { "epoch": 0.92, "grad_norm": 1.6040020360154235, "learning_rate": 1.174914761481387e-05, "loss": 2.5354, "step": 40860 }, { "epoch": 0.92, "grad_norm": 1.3677198120672265, "learning_rate": 1.1745551928796188e-05, "loss": 2.556, "step": 40870 }, { "epoch": 0.92, "grad_norm": 1.8564201015134743, "learning_rate": 1.1741956009989239e-05, "loss": 2.6066, "step": 40880 }, { "epoch": 0.92, "grad_norm": 1.56214017727886, "learning_rate": 1.1738359858872573e-05, "loss": 2.6265, "step": 40890 }, { "epoch": 0.92, "grad_norm": 1.5287207951377897, "learning_rate": 1.1734763475925787e-05, "loss": 2.5045, "step": 40900 }, { "epoch": 0.92, "grad_norm": 1.407314667306428, "learning_rate": 1.1731166861628493e-05, "loss": 2.5787, "step": 40910 }, { "epoch": 0.92, "grad_norm": 1.5523912304606908, "learning_rate": 1.172757001646034e-05, "loss": 2.6194, "step": 40920 }, { "epoch": 0.92, "grad_norm": 1.7600887840292576, "learning_rate": 1.1723972940901011e-05, "loss": 2.6792, "step": 40930 }, { "epoch": 0.92, "grad_norm": 1.5234658767937759, "learning_rate": 1.172037563543022e-05, "loss": 2.5934, "step": 40940 }, { "epoch": 0.92, "grad_norm": 1.676580988884804, "learning_rate": 1.1716778100527704e-05, "loss": 2.6933, "step": 40950 }, { "epoch": 0.92, "grad_norm": 1.6335053315121084, "learning_rate": 1.1713180336673235e-05, "loss": 2.6604, "step": 40960 }, { "epoch": 0.92, "grad_norm": 1.633496838748432, "learning_rate": 1.170958234434662e-05, "loss": 2.4401, "step": 40970 }, { "epoch": 0.92, "grad_norm": 1.8085111357394141, "learning_rate": 1.170598412402769e-05, "loss": 2.5536, "step": 40980 }, { "epoch": 0.92, "grad_norm": 1.896033414816088, "learning_rate": 1.1702385676196304e-05, "loss": 2.6371, "step": 40990 }, { "epoch": 0.92, "grad_norm": 1.658481113986106, "learning_rate": 1.1698787001332366e-05, "loss": 2.5531, "step": 41000 }, { "epoch": 0.92, "grad_norm": 1.4739542459230306, "learning_rate": 1.1695188099915795e-05, "loss": 2.6848, "step": 41010 }, { "epoch": 0.93, "grad_norm": 1.5876228549634792, "learning_rate": 1.1691588972426545e-05, "loss": 2.6412, "step": 41020 }, { "epoch": 0.93, "grad_norm": 1.7702622455475514, "learning_rate": 1.16879896193446e-05, "loss": 2.6247, "step": 41030 }, { "epoch": 0.93, "grad_norm": 1.6273359699040135, "learning_rate": 1.1684390041149977e-05, "loss": 2.661, "step": 41040 }, { "epoch": 0.93, "grad_norm": 1.4914931751945613, "learning_rate": 1.168079023832272e-05, "loss": 2.5752, "step": 41050 }, { "epoch": 0.93, "grad_norm": 1.6462246904831295, "learning_rate": 1.1677190211342906e-05, "loss": 2.6482, "step": 41060 }, { "epoch": 0.93, "grad_norm": 1.6000846818631234, "learning_rate": 1.1673589960690636e-05, "loss": 2.4385, "step": 41070 }, { "epoch": 0.93, "grad_norm": 1.7632393456765947, "learning_rate": 1.1669989486846047e-05, "loss": 2.6104, "step": 41080 }, { "epoch": 0.93, "grad_norm": 1.5356993512640613, "learning_rate": 1.1666388790289298e-05, "loss": 2.5616, "step": 41090 }, { "epoch": 0.93, "grad_norm": 1.6012221578164278, "learning_rate": 1.1662787871500592e-05, "loss": 2.4629, "step": 41100 }, { "epoch": 0.93, "grad_norm": 1.7434129765020854, "learning_rate": 1.1659186730960147e-05, "loss": 2.5954, "step": 41110 }, { "epoch": 0.93, "grad_norm": 1.666231660543158, "learning_rate": 1.1655585369148217e-05, "loss": 2.5038, "step": 41120 }, { "epoch": 0.93, "grad_norm": 1.9748979819317611, "learning_rate": 1.1651983786545084e-05, "loss": 2.5229, "step": 41130 }, { "epoch": 0.93, "grad_norm": 1.4895437696791503, "learning_rate": 1.1648381983631064e-05, "loss": 2.5464, "step": 41140 }, { "epoch": 0.93, "grad_norm": 1.6845882567667285, "learning_rate": 1.1644779960886491e-05, "loss": 2.6982, "step": 41150 }, { "epoch": 0.93, "grad_norm": 1.7399601331814978, "learning_rate": 1.164117771879174e-05, "loss": 2.6867, "step": 41160 }, { "epoch": 0.93, "grad_norm": 1.643670616456929, "learning_rate": 1.1637575257827219e-05, "loss": 2.4641, "step": 41170 }, { "epoch": 0.93, "grad_norm": 1.7003941619558105, "learning_rate": 1.1633972578473344e-05, "loss": 2.4561, "step": 41180 }, { "epoch": 0.93, "grad_norm": 1.4362103706363119, "learning_rate": 1.1630369681210578e-05, "loss": 2.6109, "step": 41190 }, { "epoch": 0.93, "grad_norm": 1.4173088049328484, "learning_rate": 1.1626766566519412e-05, "loss": 2.6499, "step": 41200 }, { "epoch": 0.93, "grad_norm": 1.7301363129120038, "learning_rate": 1.1623163234880359e-05, "loss": 2.6755, "step": 41210 }, { "epoch": 0.93, "grad_norm": 1.72231818305408, "learning_rate": 1.1619559686773967e-05, "loss": 2.6299, "step": 41220 }, { "epoch": 0.93, "grad_norm": 1.9604361051860828, "learning_rate": 1.1615955922680808e-05, "loss": 2.6048, "step": 41230 }, { "epoch": 0.93, "grad_norm": 1.4773491533206193, "learning_rate": 1.1612351943081486e-05, "loss": 2.5869, "step": 41240 }, { "epoch": 0.93, "grad_norm": 1.6974447274285027, "learning_rate": 1.160874774845663e-05, "loss": 2.5001, "step": 41250 }, { "epoch": 0.93, "grad_norm": 2.001125434635482, "learning_rate": 1.1605143339286903e-05, "loss": 2.4569, "step": 41260 }, { "epoch": 0.93, "grad_norm": 1.5678827583691142, "learning_rate": 1.1601538716052998e-05, "loss": 2.5634, "step": 41270 }, { "epoch": 0.93, "grad_norm": 1.644093516579096, "learning_rate": 1.1597933879235627e-05, "loss": 2.5494, "step": 41280 }, { "epoch": 0.93, "grad_norm": 1.5685718450346022, "learning_rate": 1.159432882931554e-05, "loss": 2.5965, "step": 41290 }, { "epoch": 0.93, "grad_norm": 1.4166049482687615, "learning_rate": 1.1590723566773508e-05, "loss": 2.6122, "step": 41300 }, { "epoch": 0.93, "grad_norm": 1.7775268632948173, "learning_rate": 1.1587118092090336e-05, "loss": 2.5211, "step": 41310 }, { "epoch": 0.93, "grad_norm": 1.550069700625255, "learning_rate": 1.1583512405746849e-05, "loss": 2.7244, "step": 41320 }, { "epoch": 0.93, "grad_norm": 1.79711456922465, "learning_rate": 1.157990650822392e-05, "loss": 2.6912, "step": 41330 }, { "epoch": 0.93, "grad_norm": 3.4958568999916304, "learning_rate": 1.1576300400002425e-05, "loss": 2.6637, "step": 41340 }, { "epoch": 0.93, "grad_norm": 1.4643424898147646, "learning_rate": 1.157269408156328e-05, "loss": 2.6353, "step": 41350 }, { "epoch": 0.93, "grad_norm": 1.5339162902418015, "learning_rate": 1.1569087553387436e-05, "loss": 2.6266, "step": 41360 }, { "epoch": 0.93, "grad_norm": 1.6111090757727984, "learning_rate": 1.156548081595586e-05, "loss": 2.5242, "step": 41370 }, { "epoch": 0.93, "grad_norm": 1.584379821047226, "learning_rate": 1.1561873869749554e-05, "loss": 2.6258, "step": 41380 }, { "epoch": 0.93, "grad_norm": 1.5154853625385472, "learning_rate": 1.1558266715249541e-05, "loss": 2.5136, "step": 41390 }, { "epoch": 0.93, "grad_norm": 1.3673965052797656, "learning_rate": 1.1554659352936882e-05, "loss": 2.564, "step": 41400 }, { "epoch": 0.93, "grad_norm": 1.9487432088157808, "learning_rate": 1.1551051783292652e-05, "loss": 2.7157, "step": 41410 }, { "epoch": 0.93, "grad_norm": 1.5101928044490824, "learning_rate": 1.1547444006797967e-05, "loss": 2.5383, "step": 41420 }, { "epoch": 0.93, "grad_norm": 1.7094717243990247, "learning_rate": 1.1543836023933968e-05, "loss": 2.5123, "step": 41430 }, { "epoch": 0.93, "grad_norm": 1.499201851375213, "learning_rate": 1.1540227835181819e-05, "loss": 2.557, "step": 41440 }, { "epoch": 0.93, "grad_norm": 1.7245605150384824, "learning_rate": 1.1536619441022708e-05, "loss": 2.5415, "step": 41450 }, { "epoch": 0.93, "grad_norm": 1.8116200981181907, "learning_rate": 1.1533010841937858e-05, "loss": 2.6245, "step": 41460 }, { "epoch": 0.94, "grad_norm": 1.5933455685711053, "learning_rate": 1.152940203840852e-05, "loss": 2.5647, "step": 41470 }, { "epoch": 0.94, "grad_norm": 1.6317807850757633, "learning_rate": 1.1525793030915966e-05, "loss": 2.5488, "step": 41480 }, { "epoch": 0.94, "grad_norm": 1.4791313644281427, "learning_rate": 1.15221838199415e-05, "loss": 2.5071, "step": 41490 }, { "epoch": 0.94, "grad_norm": 1.5803044139138678, "learning_rate": 1.1518574405966451e-05, "loss": 2.6022, "step": 41500 }, { "epoch": 0.94, "grad_norm": 1.6717474862093713, "learning_rate": 1.1514964789472173e-05, "loss": 2.5997, "step": 41510 }, { "epoch": 0.94, "grad_norm": 2.1137559457018558, "learning_rate": 1.1511354970940054e-05, "loss": 2.6755, "step": 41520 }, { "epoch": 0.94, "grad_norm": 1.6612207500346052, "learning_rate": 1.15077449508515e-05, "loss": 2.6076, "step": 41530 }, { "epoch": 0.94, "grad_norm": 1.7716867135758207, "learning_rate": 1.1504134729687954e-05, "loss": 2.5794, "step": 41540 }, { "epoch": 0.94, "grad_norm": 1.8745097635284407, "learning_rate": 1.1500524307930876e-05, "loss": 2.6044, "step": 41550 }, { "epoch": 0.94, "grad_norm": 1.4903053970796736, "learning_rate": 1.1496913686061756e-05, "loss": 2.6506, "step": 41560 }, { "epoch": 0.94, "grad_norm": 1.7170222708711953, "learning_rate": 1.1493302864562117e-05, "loss": 2.5995, "step": 41570 }, { "epoch": 0.94, "grad_norm": 2.008142952692572, "learning_rate": 1.14896918439135e-05, "loss": 2.5824, "step": 41580 }, { "epoch": 0.94, "grad_norm": 1.5832454987745943, "learning_rate": 1.148608062459747e-05, "loss": 2.5947, "step": 41590 }, { "epoch": 0.94, "grad_norm": 1.5798094712417496, "learning_rate": 1.1482469207095636e-05, "loss": 2.5775, "step": 41600 }, { "epoch": 0.94, "grad_norm": 1.510876287001521, "learning_rate": 1.1478857591889612e-05, "loss": 2.7129, "step": 41610 }, { "epoch": 0.94, "grad_norm": 1.7527580794386561, "learning_rate": 1.1475245779461051e-05, "loss": 2.5508, "step": 41620 }, { "epoch": 0.94, "grad_norm": 1.5798141347734704, "learning_rate": 1.1471633770291632e-05, "loss": 2.5654, "step": 41630 }, { "epoch": 0.94, "grad_norm": 1.5183922494068514, "learning_rate": 1.1468021564863057e-05, "loss": 2.6107, "step": 41640 }, { "epoch": 0.94, "grad_norm": 1.6363670157601085, "learning_rate": 1.1464409163657051e-05, "loss": 2.5952, "step": 41650 }, { "epoch": 0.94, "grad_norm": 1.6650935237839963, "learning_rate": 1.1460796567155375e-05, "loss": 2.6579, "step": 41660 }, { "epoch": 0.94, "grad_norm": 1.6021073052508257, "learning_rate": 1.1457183775839802e-05, "loss": 2.5985, "step": 41670 }, { "epoch": 0.94, "grad_norm": 1.5730765136365747, "learning_rate": 1.1453570790192145e-05, "loss": 2.5558, "step": 41680 }, { "epoch": 0.94, "grad_norm": 1.5469983680547528, "learning_rate": 1.1449957610694236e-05, "loss": 2.5873, "step": 41690 }, { "epoch": 0.94, "grad_norm": 1.7230562400725014, "learning_rate": 1.1446344237827933e-05, "loss": 2.4765, "step": 41700 }, { "epoch": 0.94, "grad_norm": 1.5047881729214057, "learning_rate": 1.1442730672075123e-05, "loss": 2.3946, "step": 41710 }, { "epoch": 0.94, "grad_norm": 1.6946048741266335, "learning_rate": 1.1439116913917709e-05, "loss": 2.6739, "step": 41720 }, { "epoch": 0.94, "grad_norm": 1.982929702284005, "learning_rate": 1.1435502963837633e-05, "loss": 2.5444, "step": 41730 }, { "epoch": 0.94, "grad_norm": 1.7098424053076537, "learning_rate": 1.1431888822316856e-05, "loss": 2.6916, "step": 41740 }, { "epoch": 0.94, "grad_norm": 1.6131162541557755, "learning_rate": 1.1428274489837364e-05, "loss": 2.5825, "step": 41750 }, { "epoch": 0.94, "grad_norm": 1.6974374217789134, "learning_rate": 1.142465996688117e-05, "loss": 2.6418, "step": 41760 }, { "epoch": 0.94, "grad_norm": 1.6690194770637867, "learning_rate": 1.1421045253930313e-05, "loss": 2.6473, "step": 41770 }, { "epoch": 0.94, "grad_norm": 1.7185133737574068, "learning_rate": 1.1417430351466849e-05, "loss": 2.6374, "step": 41780 }, { "epoch": 0.94, "grad_norm": 1.4841060571204132, "learning_rate": 1.1413815259972879e-05, "loss": 2.5995, "step": 41790 }, { "epoch": 0.94, "grad_norm": 1.4450719574338644, "learning_rate": 1.1410199979930506e-05, "loss": 2.5196, "step": 41800 }, { "epoch": 0.94, "grad_norm": 1.7232978409598976, "learning_rate": 1.1406584511821874e-05, "loss": 2.5087, "step": 41810 }, { "epoch": 0.94, "grad_norm": 1.7281969469749734, "learning_rate": 1.1402968856129146e-05, "loss": 2.5338, "step": 41820 }, { "epoch": 0.94, "grad_norm": 1.7181691391661646, "learning_rate": 1.139935301333451e-05, "loss": 2.6415, "step": 41830 }, { "epoch": 0.94, "grad_norm": 1.448981514232008, "learning_rate": 1.139573698392018e-05, "loss": 2.595, "step": 41840 }, { "epoch": 0.94, "grad_norm": 1.7078262062667275, "learning_rate": 1.1392120768368395e-05, "loss": 2.5821, "step": 41850 }, { "epoch": 0.94, "grad_norm": 1.6229038615769087, "learning_rate": 1.1388504367161422e-05, "loss": 2.5929, "step": 41860 }, { "epoch": 0.94, "grad_norm": 2.2235493373406268, "learning_rate": 1.1384887780781546e-05, "loss": 2.5472, "step": 41870 }, { "epoch": 0.94, "grad_norm": 1.8520547477872442, "learning_rate": 1.1381271009711076e-05, "loss": 2.5192, "step": 41880 }, { "epoch": 0.94, "grad_norm": 1.7445170673833454, "learning_rate": 1.1377654054432359e-05, "loss": 2.6325, "step": 41890 }, { "epoch": 0.94, "grad_norm": 1.5009864196779652, "learning_rate": 1.137403691542775e-05, "loss": 2.6257, "step": 41900 }, { "epoch": 0.95, "grad_norm": 1.552446922068622, "learning_rate": 1.1370419593179639e-05, "loss": 2.5645, "step": 41910 }, { "epoch": 0.95, "grad_norm": 1.598610970178863, "learning_rate": 1.1366802088170434e-05, "loss": 2.6184, "step": 41920 }, { "epoch": 0.95, "grad_norm": 1.6802980148650462, "learning_rate": 1.1363184400882577e-05, "loss": 2.5302, "step": 41930 }, { "epoch": 0.95, "grad_norm": 1.6711059619184958, "learning_rate": 1.135956653179852e-05, "loss": 2.5632, "step": 41940 }, { "epoch": 0.95, "grad_norm": 1.5874979462412602, "learning_rate": 1.135594848140075e-05, "loss": 2.6181, "step": 41950 }, { "epoch": 0.95, "grad_norm": 1.9008032484100552, "learning_rate": 1.1352330250171779e-05, "loss": 2.4924, "step": 41960 }, { "epoch": 0.95, "grad_norm": 1.6911057420510904, "learning_rate": 1.1348711838594138e-05, "loss": 2.6669, "step": 41970 }, { "epoch": 0.95, "grad_norm": 2.1611660281628926, "learning_rate": 1.134509324715038e-05, "loss": 2.5222, "step": 41980 }, { "epoch": 0.95, "grad_norm": 1.648272758941773, "learning_rate": 1.134147447632309e-05, "loss": 2.6941, "step": 41990 }, { "epoch": 0.95, "grad_norm": 1.801674912893682, "learning_rate": 1.1337855526594868e-05, "loss": 2.6659, "step": 42000 }, { "epoch": 0.95, "grad_norm": 1.832512243100829, "learning_rate": 1.1334236398448341e-05, "loss": 2.5912, "step": 42010 }, { "epoch": 0.95, "grad_norm": 1.9415003474770032, "learning_rate": 1.1330617092366167e-05, "loss": 2.4759, "step": 42020 }, { "epoch": 0.95, "grad_norm": 1.892072066963302, "learning_rate": 1.1326997608831021e-05, "loss": 2.5184, "step": 42030 }, { "epoch": 0.95, "grad_norm": 1.63765368811253, "learning_rate": 1.13233779483256e-05, "loss": 2.7032, "step": 42040 }, { "epoch": 0.95, "grad_norm": 1.6200998896645098, "learning_rate": 1.1319758111332625e-05, "loss": 2.539, "step": 42050 }, { "epoch": 0.95, "grad_norm": 1.70551794175109, "learning_rate": 1.1316138098334846e-05, "loss": 2.6001, "step": 42060 }, { "epoch": 0.95, "grad_norm": 1.6219228057660278, "learning_rate": 1.1312517909815033e-05, "loss": 2.5461, "step": 42070 }, { "epoch": 0.95, "grad_norm": 1.7619275686723146, "learning_rate": 1.1308897546255979e-05, "loss": 2.6719, "step": 42080 }, { "epoch": 0.95, "grad_norm": 1.5187175075169856, "learning_rate": 1.1305277008140499e-05, "loss": 2.657, "step": 42090 }, { "epoch": 0.95, "grad_norm": 2.0004082468691515, "learning_rate": 1.1301656295951435e-05, "loss": 2.6322, "step": 42100 }, { "epoch": 0.95, "grad_norm": 1.4978229973623156, "learning_rate": 1.129803541017165e-05, "loss": 2.6076, "step": 42110 }, { "epoch": 0.95, "grad_norm": 1.4454745258335062, "learning_rate": 1.1294414351284026e-05, "loss": 2.65, "step": 42120 }, { "epoch": 0.95, "grad_norm": 1.497208766440747, "learning_rate": 1.1290793119771482e-05, "loss": 2.535, "step": 42130 }, { "epoch": 0.95, "grad_norm": 1.6084414118370556, "learning_rate": 1.1287171716116948e-05, "loss": 2.62, "step": 42140 }, { "epoch": 0.95, "grad_norm": 1.6288272721100638, "learning_rate": 1.1283550140803367e-05, "loss": 2.544, "step": 42150 }, { "epoch": 0.95, "grad_norm": 1.6357557252103399, "learning_rate": 1.1279928394313736e-05, "loss": 2.5589, "step": 42160 }, { "epoch": 0.95, "grad_norm": 2.1112735157770715, "learning_rate": 1.1276306477131042e-05, "loss": 2.548, "step": 42170 }, { "epoch": 0.95, "grad_norm": 1.5610150952727546, "learning_rate": 1.1272684389738317e-05, "loss": 2.6466, "step": 42180 }, { "epoch": 0.95, "grad_norm": 1.8766315062947423, "learning_rate": 1.1269062132618605e-05, "loss": 2.4652, "step": 42190 }, { "epoch": 0.95, "grad_norm": 1.5389361013455667, "learning_rate": 1.1265439706254974e-05, "loss": 2.5215, "step": 42200 }, { "epoch": 0.95, "grad_norm": 1.4984186213364972, "learning_rate": 1.1261817111130517e-05, "loss": 2.662, "step": 42210 }, { "epoch": 0.95, "grad_norm": 1.5430428022304912, "learning_rate": 1.1258194347728348e-05, "loss": 2.5317, "step": 42220 }, { "epoch": 0.95, "grad_norm": 1.811101901041831, "learning_rate": 1.1254571416531607e-05, "loss": 2.6611, "step": 42230 }, { "epoch": 0.95, "grad_norm": 1.7602402399063444, "learning_rate": 1.1250948318023449e-05, "loss": 2.6982, "step": 42240 }, { "epoch": 0.95, "grad_norm": 1.7114844265336846, "learning_rate": 1.1247325052687059e-05, "loss": 2.6372, "step": 42250 }, { "epoch": 0.95, "grad_norm": 1.5594031718609858, "learning_rate": 1.1243701621005639e-05, "loss": 2.6364, "step": 42260 }, { "epoch": 0.95, "grad_norm": 1.4926367809935586, "learning_rate": 1.1240078023462414e-05, "loss": 2.6354, "step": 42270 }, { "epoch": 0.95, "grad_norm": 1.524186505472746, "learning_rate": 1.123645426054063e-05, "loss": 2.6269, "step": 42280 }, { "epoch": 0.95, "grad_norm": 1.589111741230315, "learning_rate": 1.1232830332723563e-05, "loss": 2.5769, "step": 42290 }, { "epoch": 0.95, "grad_norm": 1.4467093372130018, "learning_rate": 1.1229206240494506e-05, "loss": 2.5132, "step": 42300 }, { "epoch": 0.95, "grad_norm": 1.5627187684031683, "learning_rate": 1.1225581984336764e-05, "loss": 2.6116, "step": 42310 }, { "epoch": 0.95, "grad_norm": 1.6478492744485382, "learning_rate": 1.122195756473368e-05, "loss": 2.6655, "step": 42320 }, { "epoch": 0.95, "grad_norm": 1.5752943185089991, "learning_rate": 1.1218332982168612e-05, "loss": 2.5801, "step": 42330 }, { "epoch": 0.95, "grad_norm": 1.7908517739907681, "learning_rate": 1.1214708237124935e-05, "loss": 2.6011, "step": 42340 }, { "epoch": 0.96, "grad_norm": 1.4770391459548862, "learning_rate": 1.1211083330086053e-05, "loss": 2.6124, "step": 42350 }, { "epoch": 0.96, "grad_norm": 1.8103692330804733, "learning_rate": 1.120745826153539e-05, "loss": 2.6615, "step": 42360 }, { "epoch": 0.96, "grad_norm": 1.8521064640641098, "learning_rate": 1.1203833031956388e-05, "loss": 2.5177, "step": 42370 }, { "epoch": 0.96, "grad_norm": 1.556042933100201, "learning_rate": 1.120020764183251e-05, "loss": 2.6132, "step": 42380 }, { "epoch": 0.96, "grad_norm": 1.748132482200065, "learning_rate": 1.119658209164725e-05, "loss": 2.4717, "step": 42390 }, { "epoch": 0.96, "grad_norm": 1.628663799575796, "learning_rate": 1.1192956381884112e-05, "loss": 2.658, "step": 42400 }, { "epoch": 0.96, "grad_norm": 1.5775691726856609, "learning_rate": 1.1189330513026626e-05, "loss": 2.6653, "step": 42410 }, { "epoch": 0.96, "grad_norm": 1.5728223550461826, "learning_rate": 1.1185704485558347e-05, "loss": 2.5776, "step": 42420 }, { "epoch": 0.96, "grad_norm": 1.69754985142995, "learning_rate": 1.118207829996284e-05, "loss": 2.6466, "step": 42430 }, { "epoch": 0.96, "grad_norm": 1.7163612799293972, "learning_rate": 1.1178451956723705e-05, "loss": 2.5463, "step": 42440 }, { "epoch": 0.96, "grad_norm": 1.7164028514957912, "learning_rate": 1.117482545632455e-05, "loss": 2.5376, "step": 42450 }, { "epoch": 0.96, "grad_norm": 1.9259135892860144, "learning_rate": 1.1171198799249017e-05, "loss": 2.6695, "step": 42460 }, { "epoch": 0.96, "grad_norm": 1.3756264414345476, "learning_rate": 1.1167571985980756e-05, "loss": 2.6347, "step": 42470 }, { "epoch": 0.96, "grad_norm": 1.5532825617699904, "learning_rate": 1.1163945017003446e-05, "loss": 2.6916, "step": 42480 }, { "epoch": 0.96, "grad_norm": 1.9186420522085919, "learning_rate": 1.1160317892800787e-05, "loss": 2.5603, "step": 42490 }, { "epoch": 0.96, "grad_norm": 1.5909081440647879, "learning_rate": 1.1156690613856495e-05, "loss": 2.4821, "step": 42500 }, { "epoch": 0.96, "grad_norm": 1.6388812205428225, "learning_rate": 1.1153063180654312e-05, "loss": 2.4741, "step": 42510 }, { "epoch": 0.96, "grad_norm": 1.6920913350447668, "learning_rate": 1.1149435593677992e-05, "loss": 2.5907, "step": 42520 }, { "epoch": 0.96, "grad_norm": 1.4661626080286743, "learning_rate": 1.114580785341132e-05, "loss": 2.5563, "step": 42530 }, { "epoch": 0.96, "grad_norm": 1.791031011149403, "learning_rate": 1.1142179960338097e-05, "loss": 2.5538, "step": 42540 }, { "epoch": 0.96, "grad_norm": 1.8584751850701229, "learning_rate": 1.113855191494214e-05, "loss": 2.624, "step": 42550 }, { "epoch": 0.96, "grad_norm": 1.8310468197291225, "learning_rate": 1.1134923717707291e-05, "loss": 2.4969, "step": 42560 }, { "epoch": 0.96, "grad_norm": 1.5421383650084148, "learning_rate": 1.1131295369117417e-05, "loss": 2.591, "step": 42570 }, { "epoch": 0.96, "grad_norm": 1.5205549216103127, "learning_rate": 1.112766686965639e-05, "loss": 2.6381, "step": 42580 }, { "epoch": 0.96, "grad_norm": 1.5963551885901153, "learning_rate": 1.112403821980812e-05, "loss": 2.498, "step": 42590 }, { "epoch": 0.96, "grad_norm": 1.549672743063187, "learning_rate": 1.1120409420056526e-05, "loss": 2.6071, "step": 42600 }, { "epoch": 0.96, "grad_norm": 1.9030740005545022, "learning_rate": 1.1116780470885551e-05, "loss": 2.6862, "step": 42610 }, { "epoch": 0.96, "grad_norm": 1.5303388322301468, "learning_rate": 1.1113151372779153e-05, "loss": 2.6114, "step": 42620 }, { "epoch": 0.96, "grad_norm": 1.6157126951789753, "learning_rate": 1.1109522126221319e-05, "loss": 2.6683, "step": 42630 }, { "epoch": 0.96, "grad_norm": 1.6149243240027282, "learning_rate": 1.1105892731696048e-05, "loss": 2.5213, "step": 42640 }, { "epoch": 0.96, "grad_norm": 1.6834512481646038, "learning_rate": 1.110226318968736e-05, "loss": 2.5952, "step": 42650 }, { "epoch": 0.96, "grad_norm": 1.7457239786627208, "learning_rate": 1.1098633500679294e-05, "loss": 2.5489, "step": 42660 }, { "epoch": 0.96, "grad_norm": 1.44617346436342, "learning_rate": 1.1095003665155916e-05, "loss": 2.602, "step": 42670 }, { "epoch": 0.96, "grad_norm": 1.7532131553613752, "learning_rate": 1.1091373683601306e-05, "loss": 2.5455, "step": 42680 }, { "epoch": 0.96, "grad_norm": 1.5612092497172818, "learning_rate": 1.1087743556499557e-05, "loss": 2.6004, "step": 42690 }, { "epoch": 0.96, "grad_norm": 1.7080559034366425, "learning_rate": 1.1084113284334795e-05, "loss": 2.5761, "step": 42700 }, { "epoch": 0.96, "grad_norm": 1.5347392796591122, "learning_rate": 1.1080482867591151e-05, "loss": 2.4871, "step": 42710 }, { "epoch": 0.96, "grad_norm": 1.9099879870049195, "learning_rate": 1.1076852306752786e-05, "loss": 2.6393, "step": 42720 }, { "epoch": 0.96, "grad_norm": 1.3501682111518611, "learning_rate": 1.1073221602303884e-05, "loss": 2.6022, "step": 42730 }, { "epoch": 0.96, "grad_norm": 1.8602275178230285, "learning_rate": 1.1069590754728627e-05, "loss": 2.5514, "step": 42740 }, { "epoch": 0.96, "grad_norm": 1.7789568459996594, "learning_rate": 1.1065959764511238e-05, "loss": 2.5564, "step": 42750 }, { "epoch": 0.96, "grad_norm": 1.4300721395088416, "learning_rate": 1.1062328632135951e-05, "loss": 2.6304, "step": 42760 }, { "epoch": 0.96, "grad_norm": 1.5500264621388928, "learning_rate": 1.1058697358087019e-05, "loss": 2.6441, "step": 42770 }, { "epoch": 0.96, "grad_norm": 1.931654849687138, "learning_rate": 1.1055065942848709e-05, "loss": 2.5035, "step": 42780 }, { "epoch": 0.96, "grad_norm": 1.5406470487204555, "learning_rate": 1.105143438690532e-05, "loss": 2.6666, "step": 42790 }, { "epoch": 0.97, "grad_norm": 1.569461767587571, "learning_rate": 1.1047802690741154e-05, "loss": 2.6071, "step": 42800 }, { "epoch": 0.97, "grad_norm": 1.5832942050619347, "learning_rate": 1.1044170854840541e-05, "loss": 2.5346, "step": 42810 }, { "epoch": 0.97, "grad_norm": 1.4374886122229937, "learning_rate": 1.104053887968783e-05, "loss": 2.4362, "step": 42820 }, { "epoch": 0.97, "grad_norm": 1.7208143877540993, "learning_rate": 1.1036906765767386e-05, "loss": 2.6018, "step": 42830 }, { "epoch": 0.97, "grad_norm": 1.5298846142884046, "learning_rate": 1.1033274513563593e-05, "loss": 2.5863, "step": 42840 }, { "epoch": 0.97, "grad_norm": 1.6062770255231262, "learning_rate": 1.1029642123560848e-05, "loss": 2.5731, "step": 42850 }, { "epoch": 0.97, "grad_norm": 1.6156683780914405, "learning_rate": 1.1026009596243578e-05, "loss": 2.6093, "step": 42860 }, { "epoch": 0.97, "grad_norm": 1.6811046694865788, "learning_rate": 1.1022376932096222e-05, "loss": 2.5781, "step": 42870 }, { "epoch": 0.97, "grad_norm": 1.8668410329434897, "learning_rate": 1.101874413160323e-05, "loss": 2.6073, "step": 42880 }, { "epoch": 0.97, "grad_norm": 1.6359853083525246, "learning_rate": 1.1015111195249084e-05, "loss": 2.5804, "step": 42890 }, { "epoch": 0.97, "grad_norm": 1.4692466803690944, "learning_rate": 1.101147812351828e-05, "loss": 2.785, "step": 42900 }, { "epoch": 0.97, "grad_norm": 1.3241537066692164, "learning_rate": 1.100784491689532e-05, "loss": 2.5329, "step": 42910 }, { "epoch": 0.97, "grad_norm": 1.846133066764917, "learning_rate": 1.100421157586474e-05, "loss": 2.592, "step": 42920 }, { "epoch": 0.97, "grad_norm": 1.5944480727255468, "learning_rate": 1.100057810091109e-05, "loss": 2.4414, "step": 42930 }, { "epoch": 0.97, "grad_norm": 1.5610152196656002, "learning_rate": 1.099694449251893e-05, "loss": 2.6273, "step": 42940 }, { "epoch": 0.97, "grad_norm": 1.4939157016122844, "learning_rate": 1.0993310751172846e-05, "loss": 2.5844, "step": 42950 }, { "epoch": 0.97, "grad_norm": 1.7347191393527999, "learning_rate": 1.0989676877357435e-05, "loss": 2.6146, "step": 42960 }, { "epoch": 0.97, "grad_norm": 1.3995011469633776, "learning_rate": 1.098604287155732e-05, "loss": 2.5927, "step": 42970 }, { "epoch": 0.97, "grad_norm": 1.6710863750660605, "learning_rate": 1.0982408734257134e-05, "loss": 2.5836, "step": 42980 }, { "epoch": 0.97, "grad_norm": 1.56697542803747, "learning_rate": 1.0978774465941535e-05, "loss": 2.5557, "step": 42990 }, { "epoch": 0.97, "grad_norm": 1.5823744793579737, "learning_rate": 1.0975140067095191e-05, "loss": 2.5336, "step": 43000 }, { "epoch": 0.97, "grad_norm": 1.6484507126331955, "learning_rate": 1.0971505538202788e-05, "loss": 2.6829, "step": 43010 }, { "epoch": 0.97, "grad_norm": 1.4537042898215422, "learning_rate": 1.096787087974904e-05, "loss": 2.6162, "step": 43020 }, { "epoch": 0.97, "grad_norm": 1.6347013168557558, "learning_rate": 1.0964236092218662e-05, "loss": 2.5465, "step": 43030 }, { "epoch": 0.97, "grad_norm": 1.6418603583660845, "learning_rate": 1.0960601176096397e-05, "loss": 2.701, "step": 43040 }, { "epoch": 0.97, "grad_norm": 1.687209286836953, "learning_rate": 1.0956966131867001e-05, "loss": 2.5289, "step": 43050 }, { "epoch": 0.97, "grad_norm": 1.6973527270463407, "learning_rate": 1.0953330960015256e-05, "loss": 2.6359, "step": 43060 }, { "epoch": 0.97, "grad_norm": 1.6812975739723262, "learning_rate": 1.0949695661025946e-05, "loss": 2.5487, "step": 43070 }, { "epoch": 0.97, "grad_norm": 1.5910284410042679, "learning_rate": 1.094606023538388e-05, "loss": 2.6311, "step": 43080 }, { "epoch": 0.97, "grad_norm": 1.5700223401051208, "learning_rate": 1.0942424683573887e-05, "loss": 2.5706, "step": 43090 }, { "epoch": 0.97, "grad_norm": 1.4851538703811569, "learning_rate": 1.0938789006080808e-05, "loss": 2.4986, "step": 43100 }, { "epoch": 0.97, "grad_norm": 1.4935463134528926, "learning_rate": 1.0935153203389502e-05, "loss": 2.6351, "step": 43110 }, { "epoch": 0.97, "grad_norm": 1.5155495151007983, "learning_rate": 1.0931517275984844e-05, "loss": 2.618, "step": 43120 }, { "epoch": 0.97, "grad_norm": 1.4808849769310355, "learning_rate": 1.0927881224351727e-05, "loss": 2.6405, "step": 43130 }, { "epoch": 0.97, "grad_norm": 1.682329134519495, "learning_rate": 1.092424504897506e-05, "loss": 2.644, "step": 43140 }, { "epoch": 0.97, "grad_norm": 1.6698934309728188, "learning_rate": 1.0920608750339764e-05, "loss": 2.5722, "step": 43150 }, { "epoch": 0.97, "grad_norm": 1.5306520697198711, "learning_rate": 1.091697232893079e-05, "loss": 2.5783, "step": 43160 }, { "epoch": 0.97, "grad_norm": 1.8305104247500408, "learning_rate": 1.0913335785233095e-05, "loss": 2.6646, "step": 43170 }, { "epoch": 0.97, "grad_norm": 1.598187003602288, "learning_rate": 1.0909699119731645e-05, "loss": 2.6121, "step": 43180 }, { "epoch": 0.97, "grad_norm": 1.7788362152396875, "learning_rate": 1.0906062332911438e-05, "loss": 2.5588, "step": 43190 }, { "epoch": 0.97, "grad_norm": 1.7921961624767035, "learning_rate": 1.090242542525748e-05, "loss": 2.7379, "step": 43200 }, { "epoch": 0.97, "grad_norm": 1.536055732929284, "learning_rate": 1.0898788397254792e-05, "loss": 2.6009, "step": 43210 }, { "epoch": 0.97, "grad_norm": 1.7556192025603745, "learning_rate": 1.0895151249388414e-05, "loss": 2.5292, "step": 43220 }, { "epoch": 0.97, "grad_norm": 2.0323429928841676, "learning_rate": 1.0891513982143403e-05, "loss": 2.6432, "step": 43230 }, { "epoch": 0.98, "grad_norm": 1.6502540347376182, "learning_rate": 1.0887876596004827e-05, "loss": 2.6498, "step": 43240 }, { "epoch": 0.98, "grad_norm": 1.5445346510032931, "learning_rate": 1.0884239091457773e-05, "loss": 2.5472, "step": 43250 }, { "epoch": 0.98, "grad_norm": 1.5759770123135968, "learning_rate": 1.0880601468987348e-05, "loss": 2.6877, "step": 43260 }, { "epoch": 0.98, "grad_norm": 1.8731161884859162, "learning_rate": 1.0876963729078671e-05, "loss": 2.5394, "step": 43270 }, { "epoch": 0.98, "grad_norm": 1.6722219526024782, "learning_rate": 1.0873325872216866e-05, "loss": 2.614, "step": 43280 }, { "epoch": 0.98, "grad_norm": 1.4427683306388404, "learning_rate": 1.0869687898887093e-05, "loss": 2.5207, "step": 43290 }, { "epoch": 0.98, "grad_norm": 1.709132671643762, "learning_rate": 1.0866049809574517e-05, "loss": 2.4385, "step": 43300 }, { "epoch": 0.98, "grad_norm": 1.6615958656923733, "learning_rate": 1.086241160476431e-05, "loss": 2.4675, "step": 43310 }, { "epoch": 0.98, "grad_norm": 1.5927644615931493, "learning_rate": 1.0858773284941676e-05, "loss": 2.6814, "step": 43320 }, { "epoch": 0.98, "grad_norm": 1.7653259228636602, "learning_rate": 1.085513485059183e-05, "loss": 2.5715, "step": 43330 }, { "epoch": 0.98, "grad_norm": 1.607461946620211, "learning_rate": 1.0851496302199987e-05, "loss": 2.5726, "step": 43340 }, { "epoch": 0.98, "grad_norm": 1.5907484174336366, "learning_rate": 1.0847857640251396e-05, "loss": 2.543, "step": 43350 }, { "epoch": 0.98, "grad_norm": 1.774650474529637, "learning_rate": 1.0844218865231312e-05, "loss": 2.5585, "step": 43360 }, { "epoch": 0.98, "grad_norm": 1.4146579583669516, "learning_rate": 1.0840579977625013e-05, "loss": 2.5445, "step": 43370 }, { "epoch": 0.98, "grad_norm": 1.6632821337399335, "learning_rate": 1.0836940977917778e-05, "loss": 2.7054, "step": 43380 }, { "epoch": 0.98, "grad_norm": 1.5421321637243093, "learning_rate": 1.0833301866594912e-05, "loss": 2.6389, "step": 43390 }, { "epoch": 0.98, "grad_norm": 1.5953513931263292, "learning_rate": 1.0829662644141734e-05, "loss": 2.5982, "step": 43400 }, { "epoch": 0.98, "grad_norm": 1.6581273991463557, "learning_rate": 1.0826023311043572e-05, "loss": 2.4925, "step": 43410 }, { "epoch": 0.98, "grad_norm": 1.5504700209072786, "learning_rate": 1.0822383867785778e-05, "loss": 2.6217, "step": 43420 }, { "epoch": 0.98, "grad_norm": 1.506576565003693, "learning_rate": 1.0818744314853709e-05, "loss": 2.4865, "step": 43430 }, { "epoch": 0.98, "grad_norm": 1.5601750698541572, "learning_rate": 1.0815104652732744e-05, "loss": 2.6589, "step": 43440 }, { "epoch": 0.98, "grad_norm": 1.643215065024222, "learning_rate": 1.081146488190827e-05, "loss": 2.443, "step": 43450 }, { "epoch": 0.98, "grad_norm": 1.7427763925359148, "learning_rate": 1.0807825002865696e-05, "loss": 2.6611, "step": 43460 }, { "epoch": 0.98, "grad_norm": 1.8598259866097095, "learning_rate": 1.0804185016090436e-05, "loss": 2.5295, "step": 43470 }, { "epoch": 0.98, "grad_norm": 1.8610192243073596, "learning_rate": 1.0800544922067926e-05, "loss": 2.5911, "step": 43480 }, { "epoch": 0.98, "grad_norm": 1.6545572898925784, "learning_rate": 1.0796904721283621e-05, "loss": 2.6094, "step": 43490 }, { "epoch": 0.98, "grad_norm": 1.801564660991024, "learning_rate": 1.0793264414222972e-05, "loss": 2.6096, "step": 43500 }, { "epoch": 0.98, "grad_norm": 1.817499314468928, "learning_rate": 1.0789624001371461e-05, "loss": 2.5561, "step": 43510 }, { "epoch": 0.98, "grad_norm": 1.782886256268502, "learning_rate": 1.0785983483214581e-05, "loss": 2.6367, "step": 43520 }, { "epoch": 0.98, "grad_norm": 1.7735003712420114, "learning_rate": 1.0782342860237833e-05, "loss": 2.6433, "step": 43530 }, { "epoch": 0.98, "grad_norm": 1.5022910248015868, "learning_rate": 1.0778702132926737e-05, "loss": 2.6094, "step": 43540 }, { "epoch": 0.98, "grad_norm": 1.5895811898119683, "learning_rate": 1.0775061301766825e-05, "loss": 2.544, "step": 43550 }, { "epoch": 0.98, "grad_norm": 1.4971768601694087, "learning_rate": 1.0771420367243643e-05, "loss": 2.6965, "step": 43560 }, { "epoch": 0.98, "grad_norm": 1.6973958023700235, "learning_rate": 1.0767779329842754e-05, "loss": 2.6286, "step": 43570 }, { "epoch": 0.98, "grad_norm": 1.5854079903396814, "learning_rate": 1.0764138190049724e-05, "loss": 2.5929, "step": 43580 }, { "epoch": 0.98, "grad_norm": 1.4097781150426076, "learning_rate": 1.076049694835015e-05, "loss": 2.6334, "step": 43590 }, { "epoch": 0.98, "grad_norm": 1.546403728349576, "learning_rate": 1.0756855605229632e-05, "loss": 2.5958, "step": 43600 }, { "epoch": 0.98, "grad_norm": 1.5740494378895937, "learning_rate": 1.0753214161173777e-05, "loss": 2.6498, "step": 43610 }, { "epoch": 0.98, "grad_norm": 1.6359186648914952, "learning_rate": 1.074957261666822e-05, "loss": 2.502, "step": 43620 }, { "epoch": 0.98, "grad_norm": 1.750858219709535, "learning_rate": 1.0745930972198601e-05, "loss": 2.5844, "step": 43630 }, { "epoch": 0.98, "grad_norm": 1.3659926987829645, "learning_rate": 1.0742289228250575e-05, "loss": 2.5938, "step": 43640 }, { "epoch": 0.98, "grad_norm": 1.5094202647054475, "learning_rate": 1.073864738530981e-05, "loss": 2.5761, "step": 43650 }, { "epoch": 0.98, "grad_norm": 1.5358869967297388, "learning_rate": 1.0735005443861985e-05, "loss": 2.6284, "step": 43660 }, { "epoch": 0.98, "grad_norm": 1.5913085711728143, "learning_rate": 1.0731363404392798e-05, "loss": 2.6566, "step": 43670 }, { "epoch": 0.99, "grad_norm": 1.6924224872857467, "learning_rate": 1.0727721267387952e-05, "loss": 2.5028, "step": 43680 }, { "epoch": 0.99, "grad_norm": 1.4565117506626135, "learning_rate": 1.0724079033333173e-05, "loss": 2.629, "step": 43690 }, { "epoch": 0.99, "grad_norm": 1.4464544483053718, "learning_rate": 1.0720436702714195e-05, "loss": 2.5561, "step": 43700 }, { "epoch": 0.99, "grad_norm": 1.6996083606874435, "learning_rate": 1.0716794276016754e-05, "loss": 2.6376, "step": 43710 }, { "epoch": 0.99, "grad_norm": 1.5113423632990748, "learning_rate": 1.071315175372662e-05, "loss": 2.5242, "step": 43720 }, { "epoch": 0.99, "grad_norm": 1.5277111541593775, "learning_rate": 1.070950913632956e-05, "loss": 2.5888, "step": 43730 }, { "epoch": 0.99, "grad_norm": 1.599779813553909, "learning_rate": 1.0705866424311364e-05, "loss": 2.6574, "step": 43740 }, { "epoch": 0.99, "grad_norm": 1.461711496789906, "learning_rate": 1.0702223618157818e-05, "loss": 2.5954, "step": 43750 }, { "epoch": 0.99, "grad_norm": 1.5839528025239726, "learning_rate": 1.0698580718354748e-05, "loss": 2.6025, "step": 43760 }, { "epoch": 0.99, "grad_norm": 1.5169465228485348, "learning_rate": 1.0694937725387962e-05, "loss": 2.7038, "step": 43770 }, { "epoch": 0.99, "grad_norm": 1.6027184303782829, "learning_rate": 1.0691294639743298e-05, "loss": 2.5012, "step": 43780 }, { "epoch": 0.99, "grad_norm": 1.5026579136152158, "learning_rate": 1.0687651461906607e-05, "loss": 2.681, "step": 43790 }, { "epoch": 0.99, "grad_norm": 1.5394400263255075, "learning_rate": 1.0684008192363745e-05, "loss": 2.5392, "step": 43800 }, { "epoch": 0.99, "grad_norm": 1.5297087603318045, "learning_rate": 1.0680364831600586e-05, "loss": 2.637, "step": 43810 }, { "epoch": 0.99, "grad_norm": 1.859702432207734, "learning_rate": 1.0676721380103014e-05, "loss": 2.5545, "step": 43820 }, { "epoch": 0.99, "grad_norm": 1.7308603475657671, "learning_rate": 1.0673077838356923e-05, "loss": 2.5746, "step": 43830 }, { "epoch": 0.99, "grad_norm": 1.6661939063282651, "learning_rate": 1.0669434206848217e-05, "loss": 2.6745, "step": 43840 }, { "epoch": 0.99, "grad_norm": 1.6462015553050635, "learning_rate": 1.0665790486062824e-05, "loss": 2.6389, "step": 43850 }, { "epoch": 0.99, "grad_norm": 1.4815650321617437, "learning_rate": 1.0662146676486672e-05, "loss": 2.5785, "step": 43860 }, { "epoch": 0.99, "grad_norm": 1.562041810929011, "learning_rate": 1.0658502778605704e-05, "loss": 2.5331, "step": 43870 }, { "epoch": 0.99, "grad_norm": 1.6335409472033326, "learning_rate": 1.0654858792905875e-05, "loss": 2.6604, "step": 43880 }, { "epoch": 0.99, "grad_norm": 1.635858477068659, "learning_rate": 1.0651214719873153e-05, "loss": 2.6979, "step": 43890 }, { "epoch": 0.99, "grad_norm": 1.6448472762420063, "learning_rate": 1.0647570559993518e-05, "loss": 2.5839, "step": 43900 }, { "epoch": 0.99, "grad_norm": 1.6892085843059788, "learning_rate": 1.0643926313752955e-05, "loss": 2.5373, "step": 43910 }, { "epoch": 0.99, "grad_norm": 1.7758932461336534, "learning_rate": 1.0640281981637475e-05, "loss": 2.5214, "step": 43920 }, { "epoch": 0.99, "grad_norm": 1.4984638969190325, "learning_rate": 1.0636637564133083e-05, "loss": 2.519, "step": 43930 }, { "epoch": 0.99, "grad_norm": 1.7156698961767813, "learning_rate": 1.0632993061725805e-05, "loss": 2.5586, "step": 43940 }, { "epoch": 0.99, "grad_norm": 1.585002396473216, "learning_rate": 1.0629348474901681e-05, "loss": 2.6226, "step": 43950 }, { "epoch": 0.99, "grad_norm": 1.8387650505356183, "learning_rate": 1.0625703804146757e-05, "loss": 2.5224, "step": 43960 }, { "epoch": 0.99, "grad_norm": 2.5562840038504016, "learning_rate": 1.0622059049947089e-05, "loss": 2.5286, "step": 43970 }, { "epoch": 0.99, "grad_norm": 1.5966227051957385, "learning_rate": 1.0618414212788749e-05, "loss": 2.6014, "step": 43980 }, { "epoch": 0.99, "grad_norm": 1.719840880511191, "learning_rate": 1.0614769293157814e-05, "loss": 2.6288, "step": 43990 }, { "epoch": 0.99, "grad_norm": 1.4527251004675406, "learning_rate": 1.061112429154038e-05, "loss": 2.6645, "step": 44000 }, { "epoch": 0.99, "grad_norm": 1.6817415792672274, "learning_rate": 1.0607479208422544e-05, "loss": 2.5168, "step": 44010 }, { "epoch": 0.99, "grad_norm": 1.606513688453301, "learning_rate": 1.0603834044290426e-05, "loss": 2.4599, "step": 44020 }, { "epoch": 0.99, "grad_norm": 1.8659633357456809, "learning_rate": 1.060018879963015e-05, "loss": 2.4873, "step": 44030 }, { "epoch": 0.99, "grad_norm": 1.5077896797322703, "learning_rate": 1.0596543474927845e-05, "loss": 2.6228, "step": 44040 }, { "epoch": 0.99, "grad_norm": 1.554862330273774, "learning_rate": 1.0592898070669662e-05, "loss": 2.6389, "step": 44050 }, { "epoch": 0.99, "grad_norm": 1.628926079244387, "learning_rate": 1.0589252587341754e-05, "loss": 2.6328, "step": 44060 }, { "epoch": 0.99, "grad_norm": 1.6099970600990146, "learning_rate": 1.0585607025430289e-05, "loss": 2.6018, "step": 44070 }, { "epoch": 0.99, "grad_norm": 1.6202343016448901, "learning_rate": 1.0581961385421446e-05, "loss": 2.5286, "step": 44080 }, { "epoch": 0.99, "grad_norm": 1.8354590778342508, "learning_rate": 1.0578315667801409e-05, "loss": 2.7114, "step": 44090 }, { "epoch": 0.99, "grad_norm": 1.6509798776519329, "learning_rate": 1.057466987305638e-05, "loss": 2.7218, "step": 44100 }, { "epoch": 0.99, "grad_norm": 1.491038559511952, "learning_rate": 1.0571024001672561e-05, "loss": 2.5804, "step": 44110 }, { "epoch": 0.99, "grad_norm": 1.8931396207701308, "learning_rate": 1.056737805413618e-05, "loss": 2.6107, "step": 44120 }, { "epoch": 1.0, "grad_norm": 1.5533805663530293, "learning_rate": 1.0563732030933458e-05, "loss": 2.6235, "step": 44130 }, { "epoch": 1.0, "grad_norm": 1.8675053612453172, "learning_rate": 1.0560085932550637e-05, "loss": 2.5373, "step": 44140 }, { "epoch": 1.0, "grad_norm": 1.4943852662916084, "learning_rate": 1.0556439759473967e-05, "loss": 2.5064, "step": 44150 }, { "epoch": 1.0, "grad_norm": 1.6241467386949489, "learning_rate": 1.0552793512189703e-05, "loss": 2.641, "step": 44160 }, { "epoch": 1.0, "grad_norm": 1.699512188366334, "learning_rate": 1.0549147191184117e-05, "loss": 2.488, "step": 44170 }, { "epoch": 1.0, "grad_norm": 1.5700513529702183, "learning_rate": 1.0545500796943484e-05, "loss": 2.6823, "step": 44180 }, { "epoch": 1.0, "grad_norm": 1.3793820095763205, "learning_rate": 1.05418543299541e-05, "loss": 2.5239, "step": 44190 }, { "epoch": 1.0, "grad_norm": 1.5088953276037094, "learning_rate": 1.0538207790702254e-05, "loss": 2.5369, "step": 44200 }, { "epoch": 1.0, "grad_norm": 1.736880559339674, "learning_rate": 1.0534561179674254e-05, "loss": 2.6004, "step": 44210 }, { "epoch": 1.0, "grad_norm": 1.858290154465672, "learning_rate": 1.0530914497356424e-05, "loss": 2.4753, "step": 44220 }, { "epoch": 1.0, "grad_norm": 1.3924783821397488, "learning_rate": 1.0527267744235087e-05, "loss": 2.7326, "step": 44230 }, { "epoch": 1.0, "grad_norm": 1.8145999681415848, "learning_rate": 1.052362092079658e-05, "loss": 2.5143, "step": 44240 }, { "epoch": 1.0, "grad_norm": 1.6764593501675518, "learning_rate": 1.0519974027527246e-05, "loss": 2.589, "step": 44250 }, { "epoch": 1.0, "grad_norm": 1.5124680404761688, "learning_rate": 1.0516327064913442e-05, "loss": 2.651, "step": 44260 }, { "epoch": 1.0, "grad_norm": 1.4814963201979148, "learning_rate": 1.0512680033441534e-05, "loss": 2.549, "step": 44270 }, { "epoch": 1.0, "grad_norm": 1.6539974029650109, "learning_rate": 1.050903293359789e-05, "loss": 2.7579, "step": 44280 }, { "epoch": 1.0, "grad_norm": 1.4884569051270518, "learning_rate": 1.0505385765868893e-05, "loss": 2.5755, "step": 44290 }, { "epoch": 1.0, "grad_norm": 1.6631053799033735, "learning_rate": 1.0501738530740943e-05, "loss": 2.7213, "step": 44300 }, { "epoch": 1.0, "grad_norm": 1.6311468937380789, "learning_rate": 1.0498091228700431e-05, "loss": 2.4957, "step": 44310 }, { "epoch": 1.0, "grad_norm": 1.6236876488137575, "learning_rate": 1.049444386023377e-05, "loss": 2.596, "step": 44320 }, { "epoch": 1.0, "grad_norm": 1.7126764248961415, "learning_rate": 1.0490796425827378e-05, "loss": 2.5991, "step": 44330 }, { "epoch": 1.0, "grad_norm": 1.6073422973500369, "learning_rate": 1.0487148925967678e-05, "loss": 2.5704, "step": 44340 }, { "epoch": 1.0, "grad_norm": 1.587676119301676, "learning_rate": 1.0483501361141118e-05, "loss": 2.5673, "step": 44350 }, { "epoch": 1.0, "grad_norm": 1.6741912216509078, "learning_rate": 1.0479853731834126e-05, "loss": 2.4848, "step": 44360 }, { "epoch": 1.0, "grad_norm": 1.635368605498256, "learning_rate": 1.0476206038533165e-05, "loss": 2.3726, "step": 44370 }, { "epoch": 1.0, "grad_norm": 1.549552029203337, "learning_rate": 1.0472558281724694e-05, "loss": 2.4339, "step": 44380 }, { "epoch": 1.0, "grad_norm": 1.5446129854689283, "learning_rate": 1.0468910461895185e-05, "loss": 2.5203, "step": 44390 }, { "epoch": 1.0, "grad_norm": 1.682256633921907, "learning_rate": 1.0465262579531115e-05, "loss": 2.4103, "step": 44400 }, { "epoch": 1.0, "grad_norm": 1.8736688992635457, "learning_rate": 1.0461614635118971e-05, "loss": 2.4577, "step": 44410 }, { "epoch": 1.0, "grad_norm": 1.741529442093047, "learning_rate": 1.0457966629145247e-05, "loss": 2.4534, "step": 44420 }, { "epoch": 1.0, "grad_norm": 1.5584003543104548, "learning_rate": 1.0454318562096444e-05, "loss": 2.3738, "step": 44430 }, { "epoch": 1.0, "grad_norm": 1.7387650281648959, "learning_rate": 1.0450670434459076e-05, "loss": 2.4066, "step": 44440 }, { "epoch": 1.0, "grad_norm": 1.8172708947315246, "learning_rate": 1.0447022246719668e-05, "loss": 2.4253, "step": 44450 }, { "epoch": 1.0, "grad_norm": 1.5204701885816485, "learning_rate": 1.0443373999364738e-05, "loss": 2.4326, "step": 44460 }, { "epoch": 1.0, "grad_norm": 1.612959240117167, "learning_rate": 1.0439725692880826e-05, "loss": 2.3591, "step": 44470 }, { "epoch": 1.0, "grad_norm": 1.5229759963286478, "learning_rate": 1.0436077327754475e-05, "loss": 2.4014, "step": 44480 }, { "epoch": 1.0, "grad_norm": 1.6511523541806314, "learning_rate": 1.0432428904472235e-05, "loss": 2.4884, "step": 44490 }, { "epoch": 1.0, "grad_norm": 1.712582120627928, "learning_rate": 1.0428780423520668e-05, "loss": 2.3362, "step": 44500 }, { "epoch": 1.0, "grad_norm": 1.626983812127314, "learning_rate": 1.042513188538634e-05, "loss": 2.4176, "step": 44510 }, { "epoch": 1.0, "grad_norm": 1.7711047040663124, "learning_rate": 1.042148329055582e-05, "loss": 2.3419, "step": 44520 }, { "epoch": 1.0, "grad_norm": 1.5484360687065588, "learning_rate": 1.0417834639515694e-05, "loss": 2.3192, "step": 44530 }, { "epoch": 1.0, "grad_norm": 2.00785434438587, "learning_rate": 1.041418593275255e-05, "loss": 2.3609, "step": 44540 }, { "epoch": 1.0, "grad_norm": 1.5350231543160344, "learning_rate": 1.0410537170752985e-05, "loss": 2.573, "step": 44550 }, { "epoch": 1.0, "grad_norm": 1.7186893415431002, "learning_rate": 1.0406888354003606e-05, "loss": 2.4923, "step": 44560 }, { "epoch": 1.01, "grad_norm": 1.6818073929360005, "learning_rate": 1.0403239482991018e-05, "loss": 2.5222, "step": 44570 }, { "epoch": 1.01, "grad_norm": 1.69863623599516, "learning_rate": 1.0399590558201846e-05, "loss": 2.4359, "step": 44580 }, { "epoch": 1.01, "grad_norm": 1.720321410471403, "learning_rate": 1.0395941580122713e-05, "loss": 2.3103, "step": 44590 }, { "epoch": 1.01, "grad_norm": 1.6877458998193402, "learning_rate": 1.0392292549240253e-05, "loss": 2.4253, "step": 44600 }, { "epoch": 1.01, "grad_norm": 1.8043371329565074, "learning_rate": 1.0388643466041099e-05, "loss": 2.4037, "step": 44610 }, { "epoch": 1.01, "grad_norm": 1.668275243876723, "learning_rate": 1.038499433101191e-05, "loss": 2.3789, "step": 44620 }, { "epoch": 1.01, "grad_norm": 1.9368828910005964, "learning_rate": 1.0381345144639332e-05, "loss": 2.5387, "step": 44630 }, { "epoch": 1.01, "grad_norm": 1.6373780021304778, "learning_rate": 1.0377695907410025e-05, "loss": 2.5607, "step": 44640 }, { "epoch": 1.01, "grad_norm": 1.5348691426355823, "learning_rate": 1.037404661981066e-05, "loss": 2.5217, "step": 44650 }, { "epoch": 1.01, "grad_norm": 1.641871961326748, "learning_rate": 1.0370397282327911e-05, "loss": 2.4394, "step": 44660 }, { "epoch": 1.01, "grad_norm": 1.6243520732986856, "learning_rate": 1.0366747895448455e-05, "loss": 2.5251, "step": 44670 }, { "epoch": 1.01, "grad_norm": 1.6768195427367218, "learning_rate": 1.0363098459658985e-05, "loss": 2.4956, "step": 44680 }, { "epoch": 1.01, "grad_norm": 1.4892417626167407, "learning_rate": 1.0359448975446189e-05, "loss": 2.4679, "step": 44690 }, { "epoch": 1.01, "grad_norm": 1.585072239824107, "learning_rate": 1.035579944329677e-05, "loss": 2.4605, "step": 44700 }, { "epoch": 1.01, "grad_norm": 1.5172959121818084, "learning_rate": 1.0352149863697436e-05, "loss": 2.4239, "step": 44710 }, { "epoch": 1.01, "grad_norm": 1.5632445341915333, "learning_rate": 1.0348500237134898e-05, "loss": 2.4813, "step": 44720 }, { "epoch": 1.01, "grad_norm": 1.647390238444216, "learning_rate": 1.034485056409588e-05, "loss": 2.5684, "step": 44730 }, { "epoch": 1.01, "grad_norm": 1.568961682103632, "learning_rate": 1.03412008450671e-05, "loss": 2.4573, "step": 44740 }, { "epoch": 1.01, "grad_norm": 1.5455013179954233, "learning_rate": 1.0337551080535295e-05, "loss": 2.4794, "step": 44750 }, { "epoch": 1.01, "grad_norm": 1.698159758539337, "learning_rate": 1.03339012709872e-05, "loss": 2.3809, "step": 44760 }, { "epoch": 1.01, "grad_norm": 1.6015905545415705, "learning_rate": 1.0330251416909562e-05, "loss": 2.4578, "step": 44770 }, { "epoch": 1.01, "grad_norm": 1.5395277695584748, "learning_rate": 1.0326601518789128e-05, "loss": 2.4322, "step": 44780 }, { "epoch": 1.01, "grad_norm": 1.6455336685637385, "learning_rate": 1.0322951577112656e-05, "loss": 2.4881, "step": 44790 }, { "epoch": 1.01, "grad_norm": 1.615985634074058, "learning_rate": 1.0319301592366904e-05, "loss": 2.4562, "step": 44800 }, { "epoch": 1.01, "grad_norm": 1.5492376100452285, "learning_rate": 1.0315651565038638e-05, "loss": 2.5526, "step": 44810 }, { "epoch": 1.01, "grad_norm": 1.8770449932207496, "learning_rate": 1.0312001495614639e-05, "loss": 2.6145, "step": 44820 }, { "epoch": 1.01, "grad_norm": 1.7527569394562716, "learning_rate": 1.0308351384581676e-05, "loss": 2.355, "step": 44830 }, { "epoch": 1.01, "grad_norm": 1.6296905236688306, "learning_rate": 1.030470123242654e-05, "loss": 2.5257, "step": 44840 }, { "epoch": 1.01, "grad_norm": 1.542442877068603, "learning_rate": 1.0301051039636017e-05, "loss": 2.5015, "step": 44850 }, { "epoch": 1.01, "grad_norm": 1.6983809873221691, "learning_rate": 1.0297400806696902e-05, "loss": 2.4156, "step": 44860 }, { "epoch": 1.01, "grad_norm": 2.1471522667267986, "learning_rate": 1.0293750534095992e-05, "loss": 2.5003, "step": 44870 }, { "epoch": 1.01, "grad_norm": 1.6094393025414975, "learning_rate": 1.02901002223201e-05, "loss": 2.5119, "step": 44880 }, { "epoch": 1.01, "grad_norm": 1.6239048416472321, "learning_rate": 1.0286449871856035e-05, "loss": 2.4273, "step": 44890 }, { "epoch": 1.01, "grad_norm": 1.5689770233710605, "learning_rate": 1.0282799483190607e-05, "loss": 2.4359, "step": 44900 }, { "epoch": 1.01, "grad_norm": 1.7924550613934875, "learning_rate": 1.027914905681064e-05, "loss": 2.453, "step": 44910 }, { "epoch": 1.01, "grad_norm": 1.6000996212562026, "learning_rate": 1.0275498593202961e-05, "loss": 2.4511, "step": 44920 }, { "epoch": 1.01, "grad_norm": 1.9024595686815808, "learning_rate": 1.0271848092854405e-05, "loss": 2.5529, "step": 44930 }, { "epoch": 1.01, "grad_norm": 1.5189850046125617, "learning_rate": 1.0268197556251798e-05, "loss": 2.5386, "step": 44940 }, { "epoch": 1.01, "grad_norm": 1.5583185943346674, "learning_rate": 1.026454698388199e-05, "loss": 2.3171, "step": 44950 }, { "epoch": 1.01, "grad_norm": 1.7757387364806367, "learning_rate": 1.0260896376231821e-05, "loss": 2.4524, "step": 44960 }, { "epoch": 1.01, "grad_norm": 1.5969224939148536, "learning_rate": 1.0257245733788141e-05, "loss": 2.4917, "step": 44970 }, { "epoch": 1.01, "grad_norm": 1.4197445684899919, "learning_rate": 1.025359505703781e-05, "loss": 2.4787, "step": 44980 }, { "epoch": 1.01, "grad_norm": 1.4988133570898508, "learning_rate": 1.0249944346467684e-05, "loss": 2.4249, "step": 44990 }, { "epoch": 1.01, "grad_norm": 1.6617523534909349, "learning_rate": 1.0246293602564625e-05, "loss": 2.482, "step": 45000 }, { "epoch": 1.02, "grad_norm": 1.7313402335741288, "learning_rate": 1.0242642825815504e-05, "loss": 2.4435, "step": 45010 }, { "epoch": 1.02, "grad_norm": 1.7838995786973875, "learning_rate": 1.0238992016707196e-05, "loss": 2.4369, "step": 45020 }, { "epoch": 1.02, "grad_norm": 1.6787608570488597, "learning_rate": 1.0235341175726574e-05, "loss": 2.4318, "step": 45030 }, { "epoch": 1.02, "grad_norm": 1.6154323448140533, "learning_rate": 1.0231690303360518e-05, "loss": 2.4753, "step": 45040 }, { "epoch": 1.02, "grad_norm": 1.9049827527226069, "learning_rate": 1.0228039400095919e-05, "loss": 2.4192, "step": 45050 }, { "epoch": 1.02, "grad_norm": 1.691401707592586, "learning_rate": 1.0224388466419664e-05, "loss": 2.4235, "step": 45060 }, { "epoch": 1.02, "grad_norm": 1.8443521040416901, "learning_rate": 1.0220737502818642e-05, "loss": 2.3751, "step": 45070 }, { "epoch": 1.02, "grad_norm": 1.5929307815631308, "learning_rate": 1.0217086509779758e-05, "loss": 2.405, "step": 45080 }, { "epoch": 1.02, "grad_norm": 1.6244259885318464, "learning_rate": 1.0213435487789912e-05, "loss": 2.5717, "step": 45090 }, { "epoch": 1.02, "grad_norm": 1.7891418910439656, "learning_rate": 1.0209784437336006e-05, "loss": 2.5181, "step": 45100 }, { "epoch": 1.02, "grad_norm": 1.5640782055290912, "learning_rate": 1.020613335890495e-05, "loss": 2.6093, "step": 45110 }, { "epoch": 1.02, "grad_norm": 1.8116581228974422, "learning_rate": 1.020248225298366e-05, "loss": 2.3376, "step": 45120 }, { "epoch": 1.02, "grad_norm": 1.613002082618152, "learning_rate": 1.0198831120059052e-05, "loss": 2.5524, "step": 45130 }, { "epoch": 1.02, "grad_norm": 1.7788483005049516, "learning_rate": 1.0195179960618042e-05, "loss": 2.3557, "step": 45140 }, { "epoch": 1.02, "grad_norm": 2.1012873945237445, "learning_rate": 1.0191528775147559e-05, "loss": 2.4157, "step": 45150 }, { "epoch": 1.02, "grad_norm": 1.737798596970365, "learning_rate": 1.0187877564134529e-05, "loss": 2.4468, "step": 45160 }, { "epoch": 1.02, "grad_norm": 1.7656417514133838, "learning_rate": 1.018422632806588e-05, "loss": 2.3274, "step": 45170 }, { "epoch": 1.02, "grad_norm": 1.511310163504966, "learning_rate": 1.018057506742855e-05, "loss": 2.4397, "step": 45180 }, { "epoch": 1.02, "grad_norm": 1.5165582598094642, "learning_rate": 1.017692378270947e-05, "loss": 2.4822, "step": 45190 }, { "epoch": 1.02, "grad_norm": 1.7796835395583623, "learning_rate": 1.017327247439559e-05, "loss": 2.4533, "step": 45200 }, { "epoch": 1.02, "grad_norm": 1.4128704004040507, "learning_rate": 1.0169621142973843e-05, "loss": 2.418, "step": 45210 }, { "epoch": 1.02, "grad_norm": 1.5176802861225556, "learning_rate": 1.0165969788931186e-05, "loss": 2.5206, "step": 45220 }, { "epoch": 1.02, "grad_norm": 1.483923487952668, "learning_rate": 1.0162318412754561e-05, "loss": 2.467, "step": 45230 }, { "epoch": 1.02, "grad_norm": 1.5709439966895449, "learning_rate": 1.015866701493092e-05, "loss": 2.4277, "step": 45240 }, { "epoch": 1.02, "grad_norm": 1.7642586574612151, "learning_rate": 1.0155015595947226e-05, "loss": 2.4004, "step": 45250 }, { "epoch": 1.02, "grad_norm": 1.7158422482598923, "learning_rate": 1.0151364156290432e-05, "loss": 2.2726, "step": 45260 }, { "epoch": 1.02, "grad_norm": 1.5060020026187175, "learning_rate": 1.0147712696447503e-05, "loss": 2.4591, "step": 45270 }, { "epoch": 1.02, "grad_norm": 1.762675594024552, "learning_rate": 1.0144061216905397e-05, "loss": 2.3867, "step": 45280 }, { "epoch": 1.02, "grad_norm": 1.7287274337505667, "learning_rate": 1.0140409718151086e-05, "loss": 2.406, "step": 45290 }, { "epoch": 1.02, "grad_norm": 1.5824004811770422, "learning_rate": 1.0136758200671535e-05, "loss": 2.4917, "step": 45300 }, { "epoch": 1.02, "grad_norm": 1.615742590607622, "learning_rate": 1.0133106664953716e-05, "loss": 2.5331, "step": 45310 }, { "epoch": 1.02, "grad_norm": 1.8592923749061294, "learning_rate": 1.0129455111484607e-05, "loss": 2.4099, "step": 45320 }, { "epoch": 1.02, "grad_norm": 1.6948174643502365, "learning_rate": 1.0125803540751182e-05, "loss": 2.464, "step": 45330 }, { "epoch": 1.02, "grad_norm": 1.6332401018237679, "learning_rate": 1.0122151953240415e-05, "loss": 2.4831, "step": 45340 }, { "epoch": 1.02, "grad_norm": 1.467811918950131, "learning_rate": 1.0118500349439295e-05, "loss": 2.4345, "step": 45350 }, { "epoch": 1.02, "grad_norm": 1.8300389149891636, "learning_rate": 1.01148487298348e-05, "loss": 2.3971, "step": 45360 }, { "epoch": 1.02, "grad_norm": 1.6739483549525946, "learning_rate": 1.0111197094913915e-05, "loss": 2.352, "step": 45370 }, { "epoch": 1.02, "grad_norm": 1.5398723901558742, "learning_rate": 1.0107545445163631e-05, "loss": 2.4061, "step": 45380 }, { "epoch": 1.02, "grad_norm": 1.694788929304555, "learning_rate": 1.0103893781070935e-05, "loss": 2.2988, "step": 45390 }, { "epoch": 1.02, "grad_norm": 1.6570606656495208, "learning_rate": 1.0100242103122812e-05, "loss": 2.5041, "step": 45400 }, { "epoch": 1.02, "grad_norm": 1.6089468198280228, "learning_rate": 1.0096590411806267e-05, "loss": 2.522, "step": 45410 }, { "epoch": 1.02, "grad_norm": 1.612835451074862, "learning_rate": 1.0092938707608287e-05, "loss": 2.4399, "step": 45420 }, { "epoch": 1.02, "grad_norm": 1.6772106374817848, "learning_rate": 1.0089286991015867e-05, "loss": 2.4826, "step": 45430 }, { "epoch": 1.02, "grad_norm": 1.6460479874700944, "learning_rate": 1.0085635262516011e-05, "loss": 2.5785, "step": 45440 }, { "epoch": 1.02, "grad_norm": 1.7940350773602485, "learning_rate": 1.0081983522595713e-05, "loss": 2.4018, "step": 45450 }, { "epoch": 1.03, "grad_norm": 1.5581854452612218, "learning_rate": 1.0078331771741979e-05, "loss": 2.5167, "step": 45460 }, { "epoch": 1.03, "grad_norm": 1.5073079106234266, "learning_rate": 1.0074680010441806e-05, "loss": 2.4381, "step": 45470 }, { "epoch": 1.03, "grad_norm": 1.648127058276349, "learning_rate": 1.0071028239182203e-05, "loss": 2.4196, "step": 45480 }, { "epoch": 1.03, "grad_norm": 1.591084106178039, "learning_rate": 1.0067376458450178e-05, "loss": 2.421, "step": 45490 }, { "epoch": 1.03, "grad_norm": 1.647353251960103, "learning_rate": 1.0063724668732726e-05, "loss": 2.469, "step": 45500 }, { "epoch": 1.03, "grad_norm": 1.5776907795085708, "learning_rate": 1.0060072870516867e-05, "loss": 2.4968, "step": 45510 }, { "epoch": 1.03, "grad_norm": 1.6265821914365195, "learning_rate": 1.0056421064289602e-05, "loss": 2.5166, "step": 45520 }, { "epoch": 1.03, "grad_norm": 1.8013735466181613, "learning_rate": 1.0052769250537948e-05, "loss": 2.6143, "step": 45530 }, { "epoch": 1.03, "grad_norm": 1.6212351990554668, "learning_rate": 1.004911742974891e-05, "loss": 2.485, "step": 45540 }, { "epoch": 1.03, "grad_norm": 1.7569464346548382, "learning_rate": 1.0045465602409503e-05, "loss": 2.3722, "step": 45550 }, { "epoch": 1.03, "grad_norm": 1.5627737519475127, "learning_rate": 1.0041813769006739e-05, "loss": 2.5306, "step": 45560 }, { "epoch": 1.03, "grad_norm": 1.763417052823606, "learning_rate": 1.003816193002763e-05, "loss": 2.5758, "step": 45570 }, { "epoch": 1.03, "grad_norm": 1.6632236388761799, "learning_rate": 1.0034510085959193e-05, "loss": 2.5501, "step": 45580 }, { "epoch": 1.03, "grad_norm": 1.878266248505935, "learning_rate": 1.0030858237288446e-05, "loss": 2.4896, "step": 45590 }, { "epoch": 1.03, "grad_norm": 1.7762153291175902, "learning_rate": 1.0027206384502394e-05, "loss": 2.4423, "step": 45600 }, { "epoch": 1.03, "grad_norm": 2.047228320363083, "learning_rate": 1.0023554528088064e-05, "loss": 2.4877, "step": 45610 }, { "epoch": 1.03, "grad_norm": 1.801311599608496, "learning_rate": 1.001990266853247e-05, "loss": 2.4456, "step": 45620 }, { "epoch": 1.03, "grad_norm": 1.552906700602952, "learning_rate": 1.0016250806322627e-05, "loss": 2.4172, "step": 45630 }, { "epoch": 1.03, "grad_norm": 1.8666606185416734, "learning_rate": 1.0012598941945551e-05, "loss": 2.415, "step": 45640 }, { "epoch": 1.03, "grad_norm": 1.596230968195213, "learning_rate": 1.0008947075888267e-05, "loss": 2.4636, "step": 45650 }, { "epoch": 1.03, "grad_norm": 1.526241910427546, "learning_rate": 1.0005295208637786e-05, "loss": 2.3715, "step": 45660 }, { "epoch": 1.03, "grad_norm": 1.5235321759181462, "learning_rate": 1.0001643340681126e-05, "loss": 2.3993, "step": 45670 }, { "epoch": 1.03, "grad_norm": 1.844355882967999, "learning_rate": 9.997991472505312e-06, "loss": 2.5069, "step": 45680 }, { "epoch": 1.03, "grad_norm": 1.6663997648802638, "learning_rate": 9.994339604597358e-06, "loss": 2.5223, "step": 45690 }, { "epoch": 1.03, "grad_norm": 1.5899524280610637, "learning_rate": 9.990687737444279e-06, "loss": 2.4026, "step": 45700 }, { "epoch": 1.03, "grad_norm": 1.6303075628968566, "learning_rate": 9.987035871533098e-06, "loss": 2.4675, "step": 45710 }, { "epoch": 1.03, "grad_norm": 1.5826485201252605, "learning_rate": 9.983384007350831e-06, "loss": 2.3914, "step": 45720 }, { "epoch": 1.03, "grad_norm": 1.580382455040318, "learning_rate": 9.979732145384497e-06, "loss": 2.4043, "step": 45730 }, { "epoch": 1.03, "grad_norm": 1.5715952157431305, "learning_rate": 9.97608028612111e-06, "loss": 2.4282, "step": 45740 }, { "epoch": 1.03, "grad_norm": 1.833476022644402, "learning_rate": 9.972428430047693e-06, "loss": 2.4327, "step": 45750 }, { "epoch": 1.03, "grad_norm": 1.511806466966608, "learning_rate": 9.968776577651258e-06, "loss": 2.512, "step": 45760 }, { "epoch": 1.03, "grad_norm": 1.7119655741127986, "learning_rate": 9.965124729418826e-06, "loss": 2.3436, "step": 45770 }, { "epoch": 1.03, "grad_norm": 1.5084017253966808, "learning_rate": 9.961472885837408e-06, "loss": 2.4321, "step": 45780 }, { "epoch": 1.03, "grad_norm": 1.7867987861099925, "learning_rate": 9.95782104739402e-06, "loss": 2.492, "step": 45790 }, { "epoch": 1.03, "grad_norm": 1.6612746259397884, "learning_rate": 9.954169214575672e-06, "loss": 2.5235, "step": 45800 }, { "epoch": 1.03, "grad_norm": 1.5557158399498061, "learning_rate": 9.950517387869387e-06, "loss": 2.4879, "step": 45810 }, { "epoch": 1.03, "grad_norm": 1.6865133237568843, "learning_rate": 9.946865567762174e-06, "loss": 2.3476, "step": 45820 }, { "epoch": 1.03, "grad_norm": 1.8532916459494335, "learning_rate": 9.94321375474104e-06, "loss": 2.4403, "step": 45830 }, { "epoch": 1.03, "grad_norm": 1.5557801769397657, "learning_rate": 9.939561949293001e-06, "loss": 2.3586, "step": 45840 }, { "epoch": 1.03, "grad_norm": 1.679853248473294, "learning_rate": 9.935910151905066e-06, "loss": 2.4825, "step": 45850 }, { "epoch": 1.03, "grad_norm": 1.6155686652118386, "learning_rate": 9.932258363064244e-06, "loss": 2.4007, "step": 45860 }, { "epoch": 1.03, "grad_norm": 1.5719050190622894, "learning_rate": 9.928606583257541e-06, "loss": 2.4039, "step": 45870 }, { "epoch": 1.03, "grad_norm": 2.1351464279774133, "learning_rate": 9.924954812971967e-06, "loss": 2.375, "step": 45880 }, { "epoch": 1.03, "grad_norm": 1.708389273635108, "learning_rate": 9.921303052694529e-06, "loss": 2.6287, "step": 45890 }, { "epoch": 1.04, "grad_norm": 1.693455195218001, "learning_rate": 9.91765130291222e-06, "loss": 2.5315, "step": 45900 }, { "epoch": 1.04, "grad_norm": 1.6778823193496366, "learning_rate": 9.913999564112053e-06, "loss": 2.3813, "step": 45910 }, { "epoch": 1.04, "grad_norm": 1.6119058344471258, "learning_rate": 9.910347836781025e-06, "loss": 2.5633, "step": 45920 }, { "epoch": 1.04, "grad_norm": 1.6195158419287061, "learning_rate": 9.906696121406134e-06, "loss": 2.5137, "step": 45930 }, { "epoch": 1.04, "grad_norm": 1.5889470824329024, "learning_rate": 9.90304441847438e-06, "loss": 2.4084, "step": 45940 }, { "epoch": 1.04, "grad_norm": 1.6083377993039343, "learning_rate": 9.899392728472762e-06, "loss": 2.5451, "step": 45950 }, { "epoch": 1.04, "grad_norm": 1.5212403909201406, "learning_rate": 9.89574105188827e-06, "loss": 2.5057, "step": 45960 }, { "epoch": 1.04, "grad_norm": 1.4981612756964282, "learning_rate": 9.892089389207897e-06, "loss": 2.4255, "step": 45970 }, { "epoch": 1.04, "grad_norm": 1.4917564144154973, "learning_rate": 9.888437740918638e-06, "loss": 2.6013, "step": 45980 }, { "epoch": 1.04, "grad_norm": 2.0076557315088386, "learning_rate": 9.884786107507479e-06, "loss": 2.3671, "step": 45990 }, { "epoch": 1.04, "grad_norm": 1.675526494344982, "learning_rate": 9.88113448946141e-06, "loss": 2.4352, "step": 46000 }, { "epoch": 1.04, "grad_norm": 1.6519169979166328, "learning_rate": 9.87748288726741e-06, "loss": 2.4488, "step": 46010 }, { "epoch": 1.04, "grad_norm": 1.6689657223892034, "learning_rate": 9.873831301412465e-06, "loss": 2.5218, "step": 46020 }, { "epoch": 1.04, "grad_norm": 1.6350652526424365, "learning_rate": 9.870179732383555e-06, "loss": 2.5541, "step": 46030 }, { "epoch": 1.04, "grad_norm": 2.5465232238553104, "learning_rate": 9.86652818066766e-06, "loss": 2.5142, "step": 46040 }, { "epoch": 1.04, "grad_norm": 1.4807737980278206, "learning_rate": 9.862876646751757e-06, "loss": 2.5824, "step": 46050 }, { "epoch": 1.04, "grad_norm": 1.7305811467193652, "learning_rate": 9.859225131122817e-06, "loss": 2.4627, "step": 46060 }, { "epoch": 1.04, "grad_norm": 1.438663386051892, "learning_rate": 9.85557363426781e-06, "loss": 2.3168, "step": 46070 }, { "epoch": 1.04, "grad_norm": 1.4119920312844343, "learning_rate": 9.851922156673709e-06, "loss": 2.5256, "step": 46080 }, { "epoch": 1.04, "grad_norm": 1.5320303920438083, "learning_rate": 9.848270698827477e-06, "loss": 2.5341, "step": 46090 }, { "epoch": 1.04, "grad_norm": 1.5225239130274293, "learning_rate": 9.844619261216082e-06, "loss": 2.4506, "step": 46100 }, { "epoch": 1.04, "grad_norm": 1.787857966851382, "learning_rate": 9.840967844326483e-06, "loss": 2.4879, "step": 46110 }, { "epoch": 1.04, "grad_norm": 1.6947860342102292, "learning_rate": 9.837316448645632e-06, "loss": 2.4592, "step": 46120 }, { "epoch": 1.04, "grad_norm": 1.8577181636626636, "learning_rate": 9.833665074660489e-06, "loss": 2.4818, "step": 46130 }, { "epoch": 1.04, "grad_norm": 1.8548645176544671, "learning_rate": 9.830013722858008e-06, "loss": 2.3543, "step": 46140 }, { "epoch": 1.04, "grad_norm": 1.6884444257974085, "learning_rate": 9.826362393725138e-06, "loss": 2.5961, "step": 46150 }, { "epoch": 1.04, "grad_norm": 1.4117988567885325, "learning_rate": 9.822711087748824e-06, "loss": 2.4125, "step": 46160 }, { "epoch": 1.04, "grad_norm": 1.7099297140247511, "learning_rate": 9.819059805416006e-06, "loss": 2.4732, "step": 46170 }, { "epoch": 1.04, "grad_norm": 1.8087798500066532, "learning_rate": 9.815408547213633e-06, "loss": 2.6114, "step": 46180 }, { "epoch": 1.04, "grad_norm": 1.7356434145900381, "learning_rate": 9.811757313628633e-06, "loss": 2.4427, "step": 46190 }, { "epoch": 1.04, "grad_norm": 1.830717881597614, "learning_rate": 9.808106105147946e-06, "loss": 2.3969, "step": 46200 }, { "epoch": 1.04, "grad_norm": 1.9004183366097658, "learning_rate": 9.8044549222585e-06, "loss": 2.5457, "step": 46210 }, { "epoch": 1.04, "grad_norm": 1.5542035865292496, "learning_rate": 9.800803765447221e-06, "loss": 2.4841, "step": 46220 }, { "epoch": 1.04, "grad_norm": 1.423009941050254, "learning_rate": 9.79715263520103e-06, "loss": 2.4744, "step": 46230 }, { "epoch": 1.04, "grad_norm": 1.7598465779549866, "learning_rate": 9.793501532006852e-06, "loss": 2.47, "step": 46240 }, { "epoch": 1.04, "grad_norm": 1.715449755763527, "learning_rate": 9.789850456351603e-06, "loss": 2.4209, "step": 46250 }, { "epoch": 1.04, "grad_norm": 1.5096321549133735, "learning_rate": 9.786199408722193e-06, "loss": 2.5268, "step": 46260 }, { "epoch": 1.04, "grad_norm": 1.8026409643844223, "learning_rate": 9.782548389605528e-06, "loss": 2.3931, "step": 46270 }, { "epoch": 1.04, "grad_norm": 1.7893834261523263, "learning_rate": 9.77889739948852e-06, "loss": 2.4529, "step": 46280 }, { "epoch": 1.04, "grad_norm": 1.6184765316815701, "learning_rate": 9.775246438858067e-06, "loss": 2.4522, "step": 46290 }, { "epoch": 1.04, "grad_norm": 1.700129463502441, "learning_rate": 9.771595508201065e-06, "loss": 2.3426, "step": 46300 }, { "epoch": 1.04, "grad_norm": 1.5948412666612808, "learning_rate": 9.767944608004409e-06, "loss": 2.4698, "step": 46310 }, { "epoch": 1.04, "grad_norm": 1.5808524526651226, "learning_rate": 9.764293738754992e-06, "loss": 2.4872, "step": 46320 }, { "epoch": 1.04, "grad_norm": 1.562371592185144, "learning_rate": 9.760642900939689e-06, "loss": 2.4611, "step": 46330 }, { "epoch": 1.05, "grad_norm": 1.6136465386901206, "learning_rate": 9.756992095045389e-06, "loss": 2.4977, "step": 46340 }, { "epoch": 1.05, "grad_norm": 1.9576902988697698, "learning_rate": 9.753341321558966e-06, "loss": 2.4318, "step": 46350 }, { "epoch": 1.05, "grad_norm": 1.5254692543205843, "learning_rate": 9.749690580967293e-06, "loss": 2.5399, "step": 46360 }, { "epoch": 1.05, "grad_norm": 1.783797679670848, "learning_rate": 9.746039873757234e-06, "loss": 2.3946, "step": 46370 }, { "epoch": 1.05, "grad_norm": 1.462547182862599, "learning_rate": 9.74238920041566e-06, "loss": 2.498, "step": 46380 }, { "epoch": 1.05, "grad_norm": 1.6313837218128995, "learning_rate": 9.738738561429425e-06, "loss": 2.4374, "step": 46390 }, { "epoch": 1.05, "grad_norm": 1.663360485247097, "learning_rate": 9.735087957285382e-06, "loss": 2.4545, "step": 46400 }, { "epoch": 1.05, "grad_norm": 1.6886996240945935, "learning_rate": 9.731437388470385e-06, "loss": 2.4748, "step": 46410 }, { "epoch": 1.05, "grad_norm": 1.5250493775964709, "learning_rate": 9.727786855471278e-06, "loss": 2.3267, "step": 46420 }, { "epoch": 1.05, "grad_norm": 1.6906527968224079, "learning_rate": 9.7241363587749e-06, "loss": 2.4803, "step": 46430 }, { "epoch": 1.05, "grad_norm": 1.6488131098449654, "learning_rate": 9.720485898868087e-06, "loss": 2.5137, "step": 46440 }, { "epoch": 1.05, "grad_norm": 1.533434360428047, "learning_rate": 9.716835476237668e-06, "loss": 2.4782, "step": 46450 }, { "epoch": 1.05, "grad_norm": 1.639583235544346, "learning_rate": 9.713185091370468e-06, "loss": 2.4646, "step": 46460 }, { "epoch": 1.05, "grad_norm": 1.7991664864894912, "learning_rate": 9.709534744753311e-06, "loss": 2.2936, "step": 46470 }, { "epoch": 1.05, "grad_norm": 1.5544242076542532, "learning_rate": 9.70588443687301e-06, "loss": 2.4379, "step": 46480 }, { "epoch": 1.05, "grad_norm": 1.7823681721843256, "learning_rate": 9.702234168216377e-06, "loss": 2.4599, "step": 46490 }, { "epoch": 1.05, "grad_norm": 1.6484464381434312, "learning_rate": 9.698583939270215e-06, "loss": 2.3132, "step": 46500 }, { "epoch": 1.05, "grad_norm": 1.670081042459067, "learning_rate": 9.694933750521323e-06, "loss": 2.5044, "step": 46510 }, { "epoch": 1.05, "grad_norm": 1.644835888132913, "learning_rate": 9.691283602456499e-06, "loss": 2.461, "step": 46520 }, { "epoch": 1.05, "grad_norm": 1.605908680880999, "learning_rate": 9.687633495562528e-06, "loss": 2.493, "step": 46530 }, { "epoch": 1.05, "grad_norm": 1.4630065875533576, "learning_rate": 9.683983430326198e-06, "loss": 2.4629, "step": 46540 }, { "epoch": 1.05, "grad_norm": 1.6979328460112395, "learning_rate": 9.680333407234283e-06, "loss": 2.4058, "step": 46550 }, { "epoch": 1.05, "grad_norm": 1.5902790701481107, "learning_rate": 9.676683426773553e-06, "loss": 2.6271, "step": 46560 }, { "epoch": 1.05, "grad_norm": 1.7300822209413587, "learning_rate": 9.673033489430779e-06, "loss": 2.4829, "step": 46570 }, { "epoch": 1.05, "grad_norm": 1.5606622008457238, "learning_rate": 9.669383595692721e-06, "loss": 2.4521, "step": 46580 }, { "epoch": 1.05, "grad_norm": 1.621227430630219, "learning_rate": 9.665733746046134e-06, "loss": 2.443, "step": 46590 }, { "epoch": 1.05, "grad_norm": 1.6312831156604737, "learning_rate": 9.662083940977761e-06, "loss": 2.464, "step": 46600 }, { "epoch": 1.05, "grad_norm": 1.516141320208107, "learning_rate": 9.658434180974358e-06, "loss": 2.5284, "step": 46610 }, { "epoch": 1.05, "grad_norm": 1.5509619740810043, "learning_rate": 9.654784466522651e-06, "loss": 2.4677, "step": 46620 }, { "epoch": 1.05, "grad_norm": 1.7034422502140645, "learning_rate": 9.651134798109377e-06, "loss": 2.4628, "step": 46630 }, { "epoch": 1.05, "grad_norm": 1.7538370801831595, "learning_rate": 9.647485176221257e-06, "loss": 2.3992, "step": 46640 }, { "epoch": 1.05, "grad_norm": 1.7933873007471832, "learning_rate": 9.643835601345017e-06, "loss": 2.4819, "step": 46650 }, { "epoch": 1.05, "grad_norm": 1.729608848123409, "learning_rate": 9.640186073967357e-06, "loss": 2.4503, "step": 46660 }, { "epoch": 1.05, "grad_norm": 1.662414089804157, "learning_rate": 9.636536594574993e-06, "loss": 2.6286, "step": 46670 }, { "epoch": 1.05, "grad_norm": 1.6495002479068055, "learning_rate": 9.632887163654622e-06, "loss": 2.3593, "step": 46680 }, { "epoch": 1.05, "grad_norm": 1.6350922288176755, "learning_rate": 9.629237781692938e-06, "loss": 2.4732, "step": 46690 }, { "epoch": 1.05, "grad_norm": 1.5944706561806308, "learning_rate": 9.625588449176622e-06, "loss": 2.607, "step": 46700 }, { "epoch": 1.05, "grad_norm": 1.6165189548975174, "learning_rate": 9.621939166592363e-06, "loss": 2.4891, "step": 46710 }, { "epoch": 1.05, "grad_norm": 1.5917493146588033, "learning_rate": 9.618289934426831e-06, "loss": 2.5625, "step": 46720 }, { "epoch": 1.05, "grad_norm": 1.8412188753762972, "learning_rate": 9.614640753166691e-06, "loss": 2.4392, "step": 46730 }, { "epoch": 1.05, "grad_norm": 1.540564488188346, "learning_rate": 9.610991623298603e-06, "loss": 2.4584, "step": 46740 }, { "epoch": 1.05, "grad_norm": 1.6099406965038887, "learning_rate": 9.607342545309227e-06, "loss": 2.5231, "step": 46750 }, { "epoch": 1.05, "grad_norm": 1.4453906192033583, "learning_rate": 9.603693519685198e-06, "loss": 2.4799, "step": 46760 }, { "epoch": 1.05, "grad_norm": 1.6874206096547997, "learning_rate": 9.600044546913162e-06, "loss": 2.4644, "step": 46770 }, { "epoch": 1.05, "grad_norm": 1.719388891177356, "learning_rate": 9.59639562747975e-06, "loss": 2.5352, "step": 46780 }, { "epoch": 1.06, "grad_norm": 1.6415264465881239, "learning_rate": 9.592746761871587e-06, "loss": 2.4478, "step": 46790 }, { "epoch": 1.06, "grad_norm": 1.608024855308731, "learning_rate": 9.58909795057529e-06, "loss": 2.5457, "step": 46800 }, { "epoch": 1.06, "grad_norm": 1.7179280549760119, "learning_rate": 9.585449194077471e-06, "loss": 2.3853, "step": 46810 }, { "epoch": 1.06, "grad_norm": 1.7120114038861156, "learning_rate": 9.581800492864732e-06, "loss": 2.5257, "step": 46820 }, { "epoch": 1.06, "grad_norm": 1.8214642724061008, "learning_rate": 9.578151847423672e-06, "loss": 2.4132, "step": 46830 }, { "epoch": 1.06, "grad_norm": 1.5784153935920193, "learning_rate": 9.574503258240872e-06, "loss": 2.4815, "step": 46840 }, { "epoch": 1.06, "grad_norm": 1.4620325598968142, "learning_rate": 9.570854725802923e-06, "loss": 2.5383, "step": 46850 }, { "epoch": 1.06, "grad_norm": 1.6116854563484968, "learning_rate": 9.567206250596395e-06, "loss": 2.4577, "step": 46860 }, { "epoch": 1.06, "grad_norm": 1.9169810380838226, "learning_rate": 9.563557833107851e-06, "loss": 2.3954, "step": 46870 }, { "epoch": 1.06, "grad_norm": 1.7642320679431833, "learning_rate": 9.559909473823848e-06, "loss": 2.4228, "step": 46880 }, { "epoch": 1.06, "grad_norm": 1.6281290702971354, "learning_rate": 9.556261173230942e-06, "loss": 2.4692, "step": 46890 }, { "epoch": 1.06, "grad_norm": 1.8759398144456207, "learning_rate": 9.552612931815668e-06, "loss": 2.4893, "step": 46900 }, { "epoch": 1.06, "grad_norm": 1.587831137869456, "learning_rate": 9.548964750064569e-06, "loss": 2.4749, "step": 46910 }, { "epoch": 1.06, "grad_norm": 1.6647328885900659, "learning_rate": 9.545316628464166e-06, "loss": 2.433, "step": 46920 }, { "epoch": 1.06, "grad_norm": 1.732603434778542, "learning_rate": 9.541668567500978e-06, "loss": 2.5065, "step": 46930 }, { "epoch": 1.06, "grad_norm": 1.6234105673014723, "learning_rate": 9.538020567661519e-06, "loss": 2.654, "step": 46940 }, { "epoch": 1.06, "grad_norm": 1.5923784736536435, "learning_rate": 9.534372629432289e-06, "loss": 2.5094, "step": 46950 }, { "epoch": 1.06, "grad_norm": 1.8613596023733872, "learning_rate": 9.530724753299782e-06, "loss": 2.376, "step": 46960 }, { "epoch": 1.06, "grad_norm": 1.6953782126894341, "learning_rate": 9.527076939750488e-06, "loss": 2.3402, "step": 46970 }, { "epoch": 1.06, "grad_norm": 1.4798925947743862, "learning_rate": 9.523429189270878e-06, "loss": 2.5136, "step": 46980 }, { "epoch": 1.06, "grad_norm": 1.585499858187726, "learning_rate": 9.519781502347422e-06, "loss": 2.3941, "step": 46990 }, { "epoch": 1.06, "grad_norm": 1.6671787015604953, "learning_rate": 9.516133879466586e-06, "loss": 2.4227, "step": 47000 }, { "epoch": 1.06, "grad_norm": 1.659586561775528, "learning_rate": 9.512486321114818e-06, "loss": 2.4637, "step": 47010 }, { "epoch": 1.06, "grad_norm": 1.4213049804910187, "learning_rate": 9.508838827778562e-06, "loss": 2.4861, "step": 47020 }, { "epoch": 1.06, "grad_norm": 1.5147865501032134, "learning_rate": 9.505191399944252e-06, "loss": 2.5844, "step": 47030 }, { "epoch": 1.06, "grad_norm": 1.540886612499032, "learning_rate": 9.501544038098316e-06, "loss": 2.4223, "step": 47040 }, { "epoch": 1.06, "grad_norm": 1.668089640783824, "learning_rate": 9.497896742727172e-06, "loss": 2.4458, "step": 47050 }, { "epoch": 1.06, "grad_norm": 1.5929215503291505, "learning_rate": 9.494249514317229e-06, "loss": 2.5276, "step": 47060 }, { "epoch": 1.06, "grad_norm": 1.6476471231221956, "learning_rate": 9.49060235335488e-06, "loss": 2.3803, "step": 47070 }, { "epoch": 1.06, "grad_norm": 1.7441932562679168, "learning_rate": 9.486955260326528e-06, "loss": 2.55, "step": 47080 }, { "epoch": 1.06, "grad_norm": 1.6092970222303238, "learning_rate": 9.48330823571854e-06, "loss": 2.3491, "step": 47090 }, { "epoch": 1.06, "grad_norm": 1.6870327379611956, "learning_rate": 9.479661280017297e-06, "loss": 2.4304, "step": 47100 }, { "epoch": 1.06, "grad_norm": 1.6120272920979637, "learning_rate": 9.476014393709161e-06, "loss": 2.4953, "step": 47110 }, { "epoch": 1.06, "grad_norm": 1.5248751083544803, "learning_rate": 9.472367577280486e-06, "loss": 2.4571, "step": 47120 }, { "epoch": 1.06, "grad_norm": 1.5118926391965046, "learning_rate": 9.468720831217612e-06, "loss": 2.597, "step": 47130 }, { "epoch": 1.06, "grad_norm": 1.7526459541023376, "learning_rate": 9.46507415600688e-06, "loss": 2.4149, "step": 47140 }, { "epoch": 1.06, "grad_norm": 1.8501280305173167, "learning_rate": 9.461427552134614e-06, "loss": 2.5155, "step": 47150 }, { "epoch": 1.06, "grad_norm": 1.6472768211108617, "learning_rate": 9.457781020087131e-06, "loss": 2.5757, "step": 47160 }, { "epoch": 1.06, "grad_norm": 1.740144347706952, "learning_rate": 9.454134560350734e-06, "loss": 2.4523, "step": 47170 }, { "epoch": 1.06, "grad_norm": 1.6332355805727807, "learning_rate": 9.45048817341173e-06, "loss": 2.4749, "step": 47180 }, { "epoch": 1.06, "grad_norm": 1.5283232723819902, "learning_rate": 9.44684185975639e-06, "loss": 2.4729, "step": 47190 }, { "epoch": 1.06, "grad_norm": 1.640158736858523, "learning_rate": 9.443195619871006e-06, "loss": 2.408, "step": 47200 }, { "epoch": 1.06, "grad_norm": 1.7259382481023804, "learning_rate": 9.439549454241837e-06, "loss": 2.3967, "step": 47210 }, { "epoch": 1.06, "grad_norm": 1.6946087703659594, "learning_rate": 9.435903363355146e-06, "loss": 2.3529, "step": 47220 }, { "epoch": 1.07, "grad_norm": 1.9020839353150736, "learning_rate": 9.432257347697176e-06, "loss": 2.4036, "step": 47230 }, { "epoch": 1.07, "grad_norm": 1.858340362400242, "learning_rate": 9.42861140775417e-06, "loss": 2.408, "step": 47240 }, { "epoch": 1.07, "grad_norm": 1.8077578483090424, "learning_rate": 9.424965544012353e-06, "loss": 2.3822, "step": 47250 }, { "epoch": 1.07, "grad_norm": 1.8499517340828506, "learning_rate": 9.421319756957943e-06, "loss": 2.4717, "step": 47260 }, { "epoch": 1.07, "grad_norm": 1.5465891447975608, "learning_rate": 9.417674047077146e-06, "loss": 2.5209, "step": 47270 }, { "epoch": 1.07, "grad_norm": 1.5120475154832174, "learning_rate": 9.414028414856161e-06, "loss": 2.4534, "step": 47280 }, { "epoch": 1.07, "grad_norm": 1.5911659466472832, "learning_rate": 9.410382860781176e-06, "loss": 2.4219, "step": 47290 }, { "epoch": 1.07, "grad_norm": 1.9205816725667122, "learning_rate": 9.406737385338365e-06, "loss": 2.3282, "step": 47300 }, { "epoch": 1.07, "grad_norm": 1.6307247513336176, "learning_rate": 9.403091989013892e-06, "loss": 2.4842, "step": 47310 }, { "epoch": 1.07, "grad_norm": 1.624593433943227, "learning_rate": 9.399446672293914e-06, "loss": 2.4825, "step": 47320 }, { "epoch": 1.07, "grad_norm": 1.7871847959941656, "learning_rate": 9.395801435664575e-06, "loss": 2.4293, "step": 47330 }, { "epoch": 1.07, "grad_norm": 1.5349841438254317, "learning_rate": 9.39215627961201e-06, "loss": 2.4935, "step": 47340 }, { "epoch": 1.07, "grad_norm": 1.6168171193963685, "learning_rate": 9.388511204622342e-06, "loss": 2.446, "step": 47350 }, { "epoch": 1.07, "grad_norm": 1.5786207525768248, "learning_rate": 9.384866211181683e-06, "loss": 2.443, "step": 47360 }, { "epoch": 1.07, "grad_norm": 1.57662153161442, "learning_rate": 9.381221299776132e-06, "loss": 2.4473, "step": 47370 }, { "epoch": 1.07, "grad_norm": 1.8155319801224286, "learning_rate": 9.377576470891786e-06, "loss": 2.4279, "step": 47380 }, { "epoch": 1.07, "grad_norm": 1.5717200277185701, "learning_rate": 9.373931725014719e-06, "loss": 2.4917, "step": 47390 }, { "epoch": 1.07, "grad_norm": 1.592186040596812, "learning_rate": 9.370287062631004e-06, "loss": 2.4962, "step": 47400 }, { "epoch": 1.07, "grad_norm": 1.6442300999360175, "learning_rate": 9.366642484226692e-06, "loss": 2.4836, "step": 47410 }, { "epoch": 1.07, "grad_norm": 1.5398592032084384, "learning_rate": 9.362997990287835e-06, "loss": 2.3703, "step": 47420 }, { "epoch": 1.07, "grad_norm": 1.6490224794193262, "learning_rate": 9.35935358130046e-06, "loss": 2.5646, "step": 47430 }, { "epoch": 1.07, "grad_norm": 1.4724170238178365, "learning_rate": 9.3557092577506e-06, "loss": 2.4949, "step": 47440 }, { "epoch": 1.07, "grad_norm": 1.5150967539338733, "learning_rate": 9.352065020124262e-06, "loss": 2.3148, "step": 47450 }, { "epoch": 1.07, "grad_norm": 1.4651791566973185, "learning_rate": 9.348420868907446e-06, "loss": 2.4447, "step": 47460 }, { "epoch": 1.07, "grad_norm": 1.8244388373420053, "learning_rate": 9.344776804586144e-06, "loss": 2.3926, "step": 47470 }, { "epoch": 1.07, "grad_norm": 1.5342162912580053, "learning_rate": 9.341132827646334e-06, "loss": 2.5562, "step": 47480 }, { "epoch": 1.07, "grad_norm": 1.7256379365322372, "learning_rate": 9.33748893857398e-06, "loss": 2.5338, "step": 47490 }, { "epoch": 1.07, "grad_norm": 1.8189365670083695, "learning_rate": 9.333845137855032e-06, "loss": 2.5257, "step": 47500 }, { "epoch": 1.07, "grad_norm": 1.6930960229364036, "learning_rate": 9.330201425975446e-06, "loss": 2.5451, "step": 47510 }, { "epoch": 1.07, "grad_norm": 1.4740983893556105, "learning_rate": 9.326557803421137e-06, "loss": 2.3998, "step": 47520 }, { "epoch": 1.07, "grad_norm": 1.8318159900224604, "learning_rate": 9.32291427067803e-06, "loss": 2.4565, "step": 47530 }, { "epoch": 1.07, "grad_norm": 1.8264502534098959, "learning_rate": 9.319270828232035e-06, "loss": 2.5287, "step": 47540 }, { "epoch": 1.07, "grad_norm": 1.6562462593740017, "learning_rate": 9.31562747656904e-06, "loss": 2.3775, "step": 47550 }, { "epoch": 1.07, "grad_norm": 1.599632372566099, "learning_rate": 9.31198421617493e-06, "loss": 2.4569, "step": 47560 }, { "epoch": 1.07, "grad_norm": 1.7481445058597116, "learning_rate": 9.30834104753558e-06, "loss": 2.468, "step": 47570 }, { "epoch": 1.07, "grad_norm": 1.5684683302090343, "learning_rate": 9.304697971136843e-06, "loss": 2.4173, "step": 47580 }, { "epoch": 1.07, "grad_norm": 1.694102598205877, "learning_rate": 9.301054987464566e-06, "loss": 2.5246, "step": 47590 }, { "epoch": 1.07, "grad_norm": 1.7353668881089213, "learning_rate": 9.297412097004581e-06, "loss": 2.4586, "step": 47600 }, { "epoch": 1.07, "grad_norm": 1.982929483304043, "learning_rate": 9.293769300242711e-06, "loss": 2.3178, "step": 47610 }, { "epoch": 1.07, "grad_norm": 1.789376922105352, "learning_rate": 9.29012659766477e-06, "loss": 2.3985, "step": 47620 }, { "epoch": 1.07, "grad_norm": 1.5412726015631153, "learning_rate": 9.286483989756542e-06, "loss": 2.3884, "step": 47630 }, { "epoch": 1.07, "grad_norm": 1.791251006151301, "learning_rate": 9.282841477003818e-06, "loss": 2.5221, "step": 47640 }, { "epoch": 1.07, "grad_norm": 1.669173162906022, "learning_rate": 9.279199059892368e-06, "loss": 2.598, "step": 47650 }, { "epoch": 1.07, "grad_norm": 1.9524987480567277, "learning_rate": 9.275556738907947e-06, "loss": 2.3951, "step": 47660 }, { "epoch": 1.08, "grad_norm": 1.7119730439382765, "learning_rate": 9.271914514536303e-06, "loss": 2.5853, "step": 47670 }, { "epoch": 1.08, "grad_norm": 1.5624454843156783, "learning_rate": 9.268272387263168e-06, "loss": 2.4743, "step": 47680 }, { "epoch": 1.08, "grad_norm": 1.5022312942108824, "learning_rate": 9.26463035757426e-06, "loss": 2.5167, "step": 47690 }, { "epoch": 1.08, "grad_norm": 1.5964691625449805, "learning_rate": 9.260988425955284e-06, "loss": 2.4176, "step": 47700 }, { "epoch": 1.08, "grad_norm": 1.6586596488406824, "learning_rate": 9.257346592891936e-06, "loss": 2.3033, "step": 47710 }, { "epoch": 1.08, "grad_norm": 1.641661614229297, "learning_rate": 9.253704858869899e-06, "loss": 2.4761, "step": 47720 }, { "epoch": 1.08, "grad_norm": 1.6914055428411618, "learning_rate": 9.250063224374832e-06, "loss": 2.3976, "step": 47730 }, { "epoch": 1.08, "grad_norm": 1.6160540512118002, "learning_rate": 9.24642168989239e-06, "loss": 2.6011, "step": 47740 }, { "epoch": 1.08, "grad_norm": 1.9010034871177222, "learning_rate": 9.242780255908219e-06, "loss": 2.5465, "step": 47750 }, { "epoch": 1.08, "grad_norm": 1.7161696012960561, "learning_rate": 9.23913892290794e-06, "loss": 2.3721, "step": 47760 }, { "epoch": 1.08, "grad_norm": 1.5747242113021873, "learning_rate": 9.235497691377167e-06, "loss": 2.4547, "step": 47770 }, { "epoch": 1.08, "grad_norm": 1.7153523579236691, "learning_rate": 9.231856561801504e-06, "loss": 2.4963, "step": 47780 }, { "epoch": 1.08, "grad_norm": 1.7157358031647725, "learning_rate": 9.228215534666534e-06, "loss": 2.537, "step": 47790 }, { "epoch": 1.08, "grad_norm": 1.7020992365504453, "learning_rate": 9.224574610457828e-06, "loss": 2.4258, "step": 47800 }, { "epoch": 1.08, "grad_norm": 1.8433282449768775, "learning_rate": 9.220933789660948e-06, "loss": 2.5049, "step": 47810 }, { "epoch": 1.08, "grad_norm": 1.6084534867517892, "learning_rate": 9.217293072761438e-06, "loss": 2.36, "step": 47820 }, { "epoch": 1.08, "grad_norm": 1.8556183026488153, "learning_rate": 9.21365246024483e-06, "loss": 2.5479, "step": 47830 }, { "epoch": 1.08, "grad_norm": 1.7129322357997503, "learning_rate": 9.210011952596637e-06, "loss": 2.5227, "step": 47840 }, { "epoch": 1.08, "grad_norm": 1.644646736307288, "learning_rate": 9.206371550302369e-06, "loss": 2.5679, "step": 47850 }, { "epoch": 1.08, "grad_norm": 1.456959332158986, "learning_rate": 9.202731253847506e-06, "loss": 2.503, "step": 47860 }, { "epoch": 1.08, "grad_norm": 1.5898727905800007, "learning_rate": 9.199091063717533e-06, "loss": 2.5988, "step": 47870 }, { "epoch": 1.08, "grad_norm": 1.7473387127318565, "learning_rate": 9.195450980397906e-06, "loss": 2.4528, "step": 47880 }, { "epoch": 1.08, "grad_norm": 1.689290640541624, "learning_rate": 9.191811004374072e-06, "loss": 2.4759, "step": 47890 }, { "epoch": 1.08, "grad_norm": 1.830487049004698, "learning_rate": 9.188171136131461e-06, "loss": 2.505, "step": 47900 }, { "epoch": 1.08, "grad_norm": 1.5701989441820823, "learning_rate": 9.184531376155495e-06, "loss": 2.4775, "step": 47910 }, { "epoch": 1.08, "grad_norm": 1.5963311077558278, "learning_rate": 9.180891724931577e-06, "loss": 2.354, "step": 47920 }, { "epoch": 1.08, "grad_norm": 1.5256949170581933, "learning_rate": 9.177252182945091e-06, "loss": 2.5085, "step": 47930 }, { "epoch": 1.08, "grad_norm": 1.5510756624810471, "learning_rate": 9.173612750681424e-06, "loss": 2.4617, "step": 47940 }, { "epoch": 1.08, "grad_norm": 1.7877334776450375, "learning_rate": 9.169973428625922e-06, "loss": 2.4566, "step": 47950 }, { "epoch": 1.08, "grad_norm": 1.6481623803840135, "learning_rate": 9.166334217263933e-06, "loss": 2.4901, "step": 47960 }, { "epoch": 1.08, "grad_norm": 1.5381318516358315, "learning_rate": 9.162695117080793e-06, "loss": 2.5148, "step": 47970 }, { "epoch": 1.08, "grad_norm": 1.730500240773963, "learning_rate": 9.159056128561813e-06, "loss": 2.4428, "step": 47980 }, { "epoch": 1.08, "grad_norm": 1.5668898200298411, "learning_rate": 9.155417252192292e-06, "loss": 2.4115, "step": 47990 }, { "epoch": 1.08, "grad_norm": 1.6693809393118952, "learning_rate": 9.151778488457522e-06, "loss": 2.4574, "step": 48000 }, { "epoch": 1.08, "grad_norm": 1.7249628084329949, "learning_rate": 9.148139837842772e-06, "loss": 2.4243, "step": 48010 }, { "epoch": 1.08, "grad_norm": 1.7983628242766714, "learning_rate": 9.144501300833294e-06, "loss": 2.3866, "step": 48020 }, { "epoch": 1.08, "grad_norm": 1.7232604019815307, "learning_rate": 9.140862877914327e-06, "loss": 2.4184, "step": 48030 }, { "epoch": 1.08, "grad_norm": 1.6223347235950247, "learning_rate": 9.137224569571104e-06, "loss": 2.419, "step": 48040 }, { "epoch": 1.08, "grad_norm": 1.989727800171548, "learning_rate": 9.13358637628883e-06, "loss": 2.4504, "step": 48050 }, { "epoch": 1.08, "grad_norm": 1.5581879938627217, "learning_rate": 9.1299482985527e-06, "loss": 2.4251, "step": 48060 }, { "epoch": 1.08, "grad_norm": 1.6523430230057714, "learning_rate": 9.126310336847891e-06, "loss": 2.3774, "step": 48070 }, { "epoch": 1.08, "grad_norm": 1.6898856517192504, "learning_rate": 9.122672491659573e-06, "loss": 2.5595, "step": 48080 }, { "epoch": 1.08, "grad_norm": 1.549752017447302, "learning_rate": 9.119034763472884e-06, "loss": 2.3791, "step": 48090 }, { "epoch": 1.08, "grad_norm": 1.7174678672812456, "learning_rate": 9.115397152772967e-06, "loss": 2.5632, "step": 48100 }, { "epoch": 1.08, "grad_norm": 1.638502936127987, "learning_rate": 9.111759660044935e-06, "loss": 2.556, "step": 48110 }, { "epoch": 1.09, "grad_norm": 1.796826620032338, "learning_rate": 9.108122285773887e-06, "loss": 2.4155, "step": 48120 }, { "epoch": 1.09, "grad_norm": 1.763608417814166, "learning_rate": 9.104485030444908e-06, "loss": 2.5019, "step": 48130 }, { "epoch": 1.09, "grad_norm": 1.6820657147285307, "learning_rate": 9.100847894543074e-06, "loss": 2.4831, "step": 48140 }, { "epoch": 1.09, "grad_norm": 1.9041576995128089, "learning_rate": 9.097210878553432e-06, "loss": 2.4783, "step": 48150 }, { "epoch": 1.09, "grad_norm": 1.6039922670183435, "learning_rate": 9.093573982961024e-06, "loss": 2.4628, "step": 48160 }, { "epoch": 1.09, "grad_norm": 1.590799513027298, "learning_rate": 9.089937208250867e-06, "loss": 2.4762, "step": 48170 }, { "epoch": 1.09, "grad_norm": 1.7040439374210214, "learning_rate": 9.08630055490797e-06, "loss": 2.3564, "step": 48180 }, { "epoch": 1.09, "grad_norm": 1.627396503647987, "learning_rate": 9.082664023417319e-06, "loss": 2.5015, "step": 48190 }, { "epoch": 1.09, "grad_norm": 1.5743646910378104, "learning_rate": 9.079027614263888e-06, "loss": 2.3917, "step": 48200 }, { "epoch": 1.09, "grad_norm": 1.4159553484173413, "learning_rate": 9.075391327932639e-06, "loss": 2.4601, "step": 48210 }, { "epoch": 1.09, "grad_norm": 1.7865342761200005, "learning_rate": 9.071755164908504e-06, "loss": 2.3854, "step": 48220 }, { "epoch": 1.09, "grad_norm": 1.8303494925844848, "learning_rate": 9.068119125676412e-06, "loss": 2.4287, "step": 48230 }, { "epoch": 1.09, "grad_norm": 1.4360179957160577, "learning_rate": 9.064483210721268e-06, "loss": 2.4814, "step": 48240 }, { "epoch": 1.09, "grad_norm": 1.6579705008975265, "learning_rate": 9.060847420527965e-06, "loss": 2.2883, "step": 48250 }, { "epoch": 1.09, "grad_norm": 1.7563570925957193, "learning_rate": 9.057211755581379e-06, "loss": 2.4058, "step": 48260 }, { "epoch": 1.09, "grad_norm": 1.699520182773203, "learning_rate": 9.05357621636636e-06, "loss": 2.2931, "step": 48270 }, { "epoch": 1.09, "grad_norm": 1.7241150381023709, "learning_rate": 9.049940803367756e-06, "loss": 2.3798, "step": 48280 }, { "epoch": 1.09, "grad_norm": 1.7397144906478643, "learning_rate": 9.046305517070382e-06, "loss": 2.5048, "step": 48290 }, { "epoch": 1.09, "grad_norm": 1.687091522592572, "learning_rate": 9.042670357959056e-06, "loss": 2.3281, "step": 48300 }, { "epoch": 1.09, "grad_norm": 1.6880393905406883, "learning_rate": 9.03903532651856e-06, "loss": 2.4119, "step": 48310 }, { "epoch": 1.09, "grad_norm": 1.6236766205388602, "learning_rate": 9.035400423233671e-06, "loss": 2.3658, "step": 48320 }, { "epoch": 1.09, "grad_norm": 1.6559432068491826, "learning_rate": 9.031765648589141e-06, "loss": 2.5593, "step": 48330 }, { "epoch": 1.09, "grad_norm": 1.685637644611691, "learning_rate": 9.028131003069714e-06, "loss": 2.4453, "step": 48340 }, { "epoch": 1.09, "grad_norm": 1.685507261449373, "learning_rate": 9.024496487160107e-06, "loss": 2.4982, "step": 48350 }, { "epoch": 1.09, "grad_norm": 1.7764457769532545, "learning_rate": 9.020862101345025e-06, "loss": 2.3705, "step": 48360 }, { "epoch": 1.09, "grad_norm": 1.5294182501059568, "learning_rate": 9.01722784610916e-06, "loss": 2.4379, "step": 48370 }, { "epoch": 1.09, "grad_norm": 1.7950016362625485, "learning_rate": 9.013593721937174e-06, "loss": 2.4459, "step": 48380 }, { "epoch": 1.09, "grad_norm": 1.4656023857043556, "learning_rate": 9.009959729313719e-06, "loss": 2.3621, "step": 48390 }, { "epoch": 1.09, "grad_norm": 1.487406335748959, "learning_rate": 9.006325868723435e-06, "loss": 2.5059, "step": 48400 }, { "epoch": 1.09, "grad_norm": 1.6566350384445003, "learning_rate": 9.002692140650936e-06, "loss": 2.4274, "step": 48410 }, { "epoch": 1.09, "grad_norm": 1.5558610253465313, "learning_rate": 8.999058545580821e-06, "loss": 2.5267, "step": 48420 }, { "epoch": 1.09, "grad_norm": 1.5187013092931596, "learning_rate": 8.99542508399767e-06, "loss": 2.4699, "step": 48430 }, { "epoch": 1.09, "grad_norm": 1.7402059524958673, "learning_rate": 8.991791756386047e-06, "loss": 2.3461, "step": 48440 }, { "epoch": 1.09, "grad_norm": 1.6232914124481015, "learning_rate": 8.988158563230503e-06, "loss": 2.5406, "step": 48450 }, { "epoch": 1.09, "grad_norm": 1.5937647221974702, "learning_rate": 8.984525505015556e-06, "loss": 2.4173, "step": 48460 }, { "epoch": 1.09, "grad_norm": 1.5201135640936116, "learning_rate": 8.980892582225725e-06, "loss": 2.4866, "step": 48470 }, { "epoch": 1.09, "grad_norm": 1.6143946083567517, "learning_rate": 8.977259795345503e-06, "loss": 2.507, "step": 48480 }, { "epoch": 1.09, "grad_norm": 1.53764443619822, "learning_rate": 8.97362714485935e-06, "loss": 2.4988, "step": 48490 }, { "epoch": 1.09, "grad_norm": 1.753050497405038, "learning_rate": 8.969994631251733e-06, "loss": 2.4668, "step": 48500 }, { "epoch": 1.09, "grad_norm": 1.7698811524153637, "learning_rate": 8.966362255007087e-06, "loss": 2.3124, "step": 48510 }, { "epoch": 1.09, "grad_norm": 1.7138717758177215, "learning_rate": 8.962730016609827e-06, "loss": 2.4987, "step": 48520 }, { "epoch": 1.09, "grad_norm": 1.6600354943860247, "learning_rate": 8.95909791654436e-06, "loss": 2.4372, "step": 48530 }, { "epoch": 1.09, "grad_norm": 1.5708786343909884, "learning_rate": 8.955465955295063e-06, "loss": 2.4813, "step": 48540 }, { "epoch": 1.09, "grad_norm": 1.568566711024439, "learning_rate": 8.9518341333463e-06, "loss": 2.5356, "step": 48550 }, { "epoch": 1.1, "grad_norm": 1.6402494522241262, "learning_rate": 8.948202451182417e-06, "loss": 2.4996, "step": 48560 }, { "epoch": 1.1, "grad_norm": 1.7608789780021366, "learning_rate": 8.944570909287741e-06, "loss": 2.5734, "step": 48570 }, { "epoch": 1.1, "grad_norm": 1.551693386709776, "learning_rate": 8.940939508146577e-06, "loss": 2.4664, "step": 48580 }, { "epoch": 1.1, "grad_norm": 1.63403240569359, "learning_rate": 8.937308248243219e-06, "loss": 2.3549, "step": 48590 }, { "epoch": 1.1, "grad_norm": 1.4379453371572193, "learning_rate": 8.933677130061932e-06, "loss": 2.4719, "step": 48600 }, { "epoch": 1.1, "grad_norm": 1.706554906226081, "learning_rate": 8.930046154086967e-06, "loss": 2.4853, "step": 48610 }, { "epoch": 1.1, "grad_norm": 1.7324445186922712, "learning_rate": 8.926415320802556e-06, "loss": 2.4587, "step": 48620 }, { "epoch": 1.1, "grad_norm": 1.7099078893863888, "learning_rate": 8.922784630692916e-06, "loss": 2.3286, "step": 48630 }, { "epoch": 1.1, "grad_norm": 1.7237254255178394, "learning_rate": 8.919154084242236e-06, "loss": 2.3744, "step": 48640 }, { "epoch": 1.1, "grad_norm": 1.5853500520406554, "learning_rate": 8.915523681934695e-06, "loss": 2.3498, "step": 48650 }, { "epoch": 1.1, "grad_norm": 1.693392402517623, "learning_rate": 8.911893424254445e-06, "loss": 2.5056, "step": 48660 }, { "epoch": 1.1, "grad_norm": 1.6105470028595077, "learning_rate": 8.908263311685627e-06, "loss": 2.3365, "step": 48670 }, { "epoch": 1.1, "grad_norm": 1.551670703615949, "learning_rate": 8.904633344712353e-06, "loss": 2.4465, "step": 48680 }, { "epoch": 1.1, "grad_norm": 1.581159113921046, "learning_rate": 8.901003523818724e-06, "loss": 2.5117, "step": 48690 }, { "epoch": 1.1, "grad_norm": 1.5515341794682107, "learning_rate": 8.897373849488818e-06, "loss": 2.4773, "step": 48700 }, { "epoch": 1.1, "grad_norm": 1.6749126401710566, "learning_rate": 8.893744322206691e-06, "loss": 2.3836, "step": 48710 }, { "epoch": 1.1, "grad_norm": 1.764200415077657, "learning_rate": 8.890114942456379e-06, "loss": 2.483, "step": 48720 }, { "epoch": 1.1, "grad_norm": 1.667010254291631, "learning_rate": 8.88648571072191e-06, "loss": 2.5217, "step": 48730 }, { "epoch": 1.1, "grad_norm": 1.537868313690628, "learning_rate": 8.882856627487276e-06, "loss": 2.5565, "step": 48740 }, { "epoch": 1.1, "grad_norm": 1.4667542926527517, "learning_rate": 8.87922769323646e-06, "loss": 2.2359, "step": 48750 }, { "epoch": 1.1, "grad_norm": 1.8437875698621642, "learning_rate": 8.87559890845342e-06, "loss": 2.4978, "step": 48760 }, { "epoch": 1.1, "grad_norm": 1.7245470821076632, "learning_rate": 8.871970273622098e-06, "loss": 2.348, "step": 48770 }, { "epoch": 1.1, "grad_norm": 1.5915406483382322, "learning_rate": 8.868341789226413e-06, "loss": 2.4839, "step": 48780 }, { "epoch": 1.1, "grad_norm": 1.879797681187919, "learning_rate": 8.864713455750266e-06, "loss": 2.4251, "step": 48790 }, { "epoch": 1.1, "grad_norm": 1.628812571032276, "learning_rate": 8.861085273677536e-06, "loss": 2.415, "step": 48800 }, { "epoch": 1.1, "grad_norm": 1.7793754392968282, "learning_rate": 8.85745724349208e-06, "loss": 2.4315, "step": 48810 }, { "epoch": 1.1, "grad_norm": 1.8417414464529485, "learning_rate": 8.853829365677738e-06, "loss": 2.4135, "step": 48820 }, { "epoch": 1.1, "grad_norm": 1.6294082541847135, "learning_rate": 8.85020164071833e-06, "loss": 2.3993, "step": 48830 }, { "epoch": 1.1, "grad_norm": 1.4156393108490208, "learning_rate": 8.846574069097656e-06, "loss": 2.466, "step": 48840 }, { "epoch": 1.1, "grad_norm": 1.4482930316867737, "learning_rate": 8.842946651299492e-06, "loss": 2.4384, "step": 48850 }, { "epoch": 1.1, "grad_norm": 1.5769093836105523, "learning_rate": 8.839319387807593e-06, "loss": 2.452, "step": 48860 }, { "epoch": 1.1, "grad_norm": 1.6217104741061628, "learning_rate": 8.835692279105702e-06, "loss": 2.4914, "step": 48870 }, { "epoch": 1.1, "grad_norm": 1.484616517022606, "learning_rate": 8.832065325677533e-06, "loss": 2.4792, "step": 48880 }, { "epoch": 1.1, "grad_norm": 1.6030054505932063, "learning_rate": 8.82843852800678e-06, "loss": 2.5525, "step": 48890 }, { "epoch": 1.1, "grad_norm": 1.5425608733189322, "learning_rate": 8.824811886577118e-06, "loss": 2.3829, "step": 48900 }, { "epoch": 1.1, "grad_norm": 1.609021481507743, "learning_rate": 8.821185401872207e-06, "loss": 2.6031, "step": 48910 }, { "epoch": 1.1, "grad_norm": 1.6269549856767802, "learning_rate": 8.81755907437567e-06, "loss": 2.3801, "step": 48920 }, { "epoch": 1.1, "grad_norm": 1.5077465880077552, "learning_rate": 8.813932904571126e-06, "loss": 2.5269, "step": 48930 }, { "epoch": 1.1, "grad_norm": 1.5823122964809098, "learning_rate": 8.810306892942163e-06, "loss": 2.3822, "step": 48940 }, { "epoch": 1.1, "grad_norm": 1.5152875653229432, "learning_rate": 8.806681039972353e-06, "loss": 2.7256, "step": 48950 }, { "epoch": 1.1, "grad_norm": 1.662418387692082, "learning_rate": 8.803055346145242e-06, "loss": 2.4373, "step": 48960 }, { "epoch": 1.1, "grad_norm": 1.642837816575875, "learning_rate": 8.799429811944364e-06, "loss": 2.3816, "step": 48970 }, { "epoch": 1.1, "grad_norm": 1.779116876463714, "learning_rate": 8.795804437853218e-06, "loss": 2.5451, "step": 48980 }, { "epoch": 1.1, "grad_norm": 1.3770211629791596, "learning_rate": 8.792179224355293e-06, "loss": 2.4695, "step": 48990 }, { "epoch": 1.1, "grad_norm": 1.6278478927400564, "learning_rate": 8.788554171934051e-06, "loss": 2.4757, "step": 49000 }, { "epoch": 1.11, "grad_norm": 1.5453457531942287, "learning_rate": 8.784929281072937e-06, "loss": 2.4256, "step": 49010 }, { "epoch": 1.11, "grad_norm": 1.748375577364553, "learning_rate": 8.781304552255372e-06, "loss": 2.3574, "step": 49020 }, { "epoch": 1.11, "grad_norm": 1.612078449879406, "learning_rate": 8.77767998596475e-06, "loss": 2.4404, "step": 49030 }, { "epoch": 1.11, "grad_norm": 1.7224140475856542, "learning_rate": 8.77405558268445e-06, "loss": 2.3991, "step": 49040 }, { "epoch": 1.11, "grad_norm": 1.6827881276076975, "learning_rate": 8.770431342897826e-06, "loss": 2.4759, "step": 49050 }, { "epoch": 1.11, "grad_norm": 1.791103028563089, "learning_rate": 8.766807267088216e-06, "loss": 2.4161, "step": 49060 }, { "epoch": 1.11, "grad_norm": 1.7601643175125425, "learning_rate": 8.763183355738931e-06, "loss": 2.3695, "step": 49070 }, { "epoch": 1.11, "grad_norm": 1.4422470313544018, "learning_rate": 8.759559609333259e-06, "loss": 2.4488, "step": 49080 }, { "epoch": 1.11, "grad_norm": 1.6218158784327326, "learning_rate": 8.755936028354466e-06, "loss": 2.3634, "step": 49090 }, { "epoch": 1.11, "grad_norm": 2.086201966344517, "learning_rate": 8.752312613285803e-06, "loss": 2.5455, "step": 49100 }, { "epoch": 1.11, "grad_norm": 1.660189921822672, "learning_rate": 8.748689364610492e-06, "loss": 2.3558, "step": 49110 }, { "epoch": 1.11, "grad_norm": 1.5195787715376163, "learning_rate": 8.745066282811733e-06, "loss": 2.4299, "step": 49120 }, { "epoch": 1.11, "grad_norm": 1.6711041360144185, "learning_rate": 8.741443368372711e-06, "loss": 2.4268, "step": 49130 }, { "epoch": 1.11, "grad_norm": 1.8425894440190114, "learning_rate": 8.737820621776574e-06, "loss": 2.3674, "step": 49140 }, { "epoch": 1.11, "grad_norm": 1.7632403745315608, "learning_rate": 8.73419804350646e-06, "loss": 2.4318, "step": 49150 }, { "epoch": 1.11, "grad_norm": 1.6196976724519698, "learning_rate": 8.730575634045482e-06, "loss": 2.5193, "step": 49160 }, { "epoch": 1.11, "grad_norm": 1.6605686684685819, "learning_rate": 8.726953393876729e-06, "loss": 2.4273, "step": 49170 }, { "epoch": 1.11, "grad_norm": 1.4305036591393026, "learning_rate": 8.72333132348327e-06, "loss": 2.5406, "step": 49180 }, { "epoch": 1.11, "grad_norm": 1.8913991207374732, "learning_rate": 8.719709423348146e-06, "loss": 2.5073, "step": 49190 }, { "epoch": 1.11, "grad_norm": 1.7756208036464576, "learning_rate": 8.716087693954384e-06, "loss": 2.3545, "step": 49200 }, { "epoch": 1.11, "grad_norm": 1.41129828534336, "learning_rate": 8.712466135784979e-06, "loss": 2.5268, "step": 49210 }, { "epoch": 1.11, "grad_norm": 1.7641539828479884, "learning_rate": 8.708844749322907e-06, "loss": 2.3935, "step": 49220 }, { "epoch": 1.11, "grad_norm": 1.7904356035742406, "learning_rate": 8.705223535051125e-06, "loss": 2.446, "step": 49230 }, { "epoch": 1.11, "grad_norm": 1.7219972448513634, "learning_rate": 8.701602493452557e-06, "loss": 2.4242, "step": 49240 }, { "epoch": 1.11, "grad_norm": 1.5368017947322097, "learning_rate": 8.697981625010113e-06, "loss": 2.3867, "step": 49250 }, { "epoch": 1.11, "grad_norm": 1.7295970271027048, "learning_rate": 8.694360930206678e-06, "loss": 2.4032, "step": 49260 }, { "epoch": 1.11, "grad_norm": 1.582733088487381, "learning_rate": 8.690740409525115e-06, "loss": 2.3324, "step": 49270 }, { "epoch": 1.11, "grad_norm": 1.65487464641681, "learning_rate": 8.687120063448258e-06, "loss": 2.3896, "step": 49280 }, { "epoch": 1.11, "grad_norm": 1.545502752948974, "learning_rate": 8.683499892458919e-06, "loss": 2.4341, "step": 49290 }, { "epoch": 1.11, "grad_norm": 1.6701264165935916, "learning_rate": 8.679879897039898e-06, "loss": 2.3467, "step": 49300 }, { "epoch": 1.11, "grad_norm": 1.6266316517419277, "learning_rate": 8.676260077673957e-06, "loss": 2.3374, "step": 49310 }, { "epoch": 1.11, "grad_norm": 1.7912721726597882, "learning_rate": 8.67264043484384e-06, "loss": 2.4922, "step": 49320 }, { "epoch": 1.11, "grad_norm": 1.7105019850186443, "learning_rate": 8.669020969032267e-06, "loss": 2.3491, "step": 49330 }, { "epoch": 1.11, "grad_norm": 1.6213119331694499, "learning_rate": 8.665401680721943e-06, "loss": 2.5111, "step": 49340 }, { "epoch": 1.11, "grad_norm": 1.5698386952637244, "learning_rate": 8.661782570395529e-06, "loss": 2.4699, "step": 49350 }, { "epoch": 1.11, "grad_norm": 1.559410326194137, "learning_rate": 8.658163638535682e-06, "loss": 2.5173, "step": 49360 }, { "epoch": 1.11, "grad_norm": 1.5513969686784608, "learning_rate": 8.654544885625026e-06, "loss": 2.479, "step": 49370 }, { "epoch": 1.11, "grad_norm": 1.4832956116179534, "learning_rate": 8.650926312146164e-06, "loss": 2.3649, "step": 49380 }, { "epoch": 1.11, "grad_norm": 1.6393215547382654, "learning_rate": 8.647307918581671e-06, "loss": 2.3539, "step": 49390 }, { "epoch": 1.11, "grad_norm": 1.607395775707731, "learning_rate": 8.643689705414107e-06, "loss": 2.4665, "step": 49400 }, { "epoch": 1.11, "grad_norm": 1.562006281997388, "learning_rate": 8.640071673126e-06, "loss": 2.5677, "step": 49410 }, { "epoch": 1.11, "grad_norm": 1.8167421907899584, "learning_rate": 8.636453822199851e-06, "loss": 2.515, "step": 49420 }, { "epoch": 1.11, "grad_norm": 1.793841424016182, "learning_rate": 8.632836153118146e-06, "loss": 2.5262, "step": 49430 }, { "epoch": 1.11, "grad_norm": 1.5152383915725187, "learning_rate": 8.629218666363343e-06, "loss": 2.5658, "step": 49440 }, { "epoch": 1.12, "grad_norm": 1.9632995400296829, "learning_rate": 8.625601362417877e-06, "loss": 2.3417, "step": 49450 }, { "epoch": 1.12, "grad_norm": 1.7430888691663076, "learning_rate": 8.62198424176415e-06, "loss": 2.5021, "step": 49460 }, { "epoch": 1.12, "grad_norm": 1.7653460701162815, "learning_rate": 8.61836730488455e-06, "loss": 2.3952, "step": 49470 }, { "epoch": 1.12, "grad_norm": 1.541229952827859, "learning_rate": 8.614750552261437e-06, "loss": 2.4244, "step": 49480 }, { "epoch": 1.12, "grad_norm": 1.6884436005001102, "learning_rate": 8.611133984377144e-06, "loss": 2.4752, "step": 49490 }, { "epoch": 1.12, "grad_norm": 2.2189624152441723, "learning_rate": 8.607517601713986e-06, "loss": 2.3773, "step": 49500 }, { "epoch": 1.12, "grad_norm": 1.5804737881613549, "learning_rate": 8.603901404754247e-06, "loss": 2.4102, "step": 49510 }, { "epoch": 1.12, "grad_norm": 1.6783322883304475, "learning_rate": 8.600285393980184e-06, "loss": 2.507, "step": 49520 }, { "epoch": 1.12, "grad_norm": 1.7644676918087014, "learning_rate": 8.596669569874042e-06, "loss": 2.4697, "step": 49530 }, { "epoch": 1.12, "grad_norm": 1.4540861552606295, "learning_rate": 8.593053932918025e-06, "loss": 2.4324, "step": 49540 }, { "epoch": 1.12, "grad_norm": 1.8676863008604832, "learning_rate": 8.589438483594322e-06, "loss": 2.5209, "step": 49550 }, { "epoch": 1.12, "grad_norm": 1.5296691724500735, "learning_rate": 8.585823222385097e-06, "loss": 2.4685, "step": 49560 }, { "epoch": 1.12, "grad_norm": 1.6536134929586677, "learning_rate": 8.58220814977248e-06, "loss": 2.3746, "step": 49570 }, { "epoch": 1.12, "grad_norm": 1.5843474169446667, "learning_rate": 8.578593266238583e-06, "loss": 2.4973, "step": 49580 }, { "epoch": 1.12, "grad_norm": 1.488455685411321, "learning_rate": 8.5749785722655e-06, "loss": 2.4654, "step": 49590 }, { "epoch": 1.12, "grad_norm": 1.727096412547899, "learning_rate": 8.571364068335283e-06, "loss": 2.5281, "step": 49600 }, { "epoch": 1.12, "grad_norm": 1.626650067574198, "learning_rate": 8.567749754929974e-06, "loss": 2.5827, "step": 49610 }, { "epoch": 1.12, "grad_norm": 1.7799642672502602, "learning_rate": 8.564135632531576e-06, "loss": 2.5574, "step": 49620 }, { "epoch": 1.12, "grad_norm": 1.714797425389643, "learning_rate": 8.560521701622079e-06, "loss": 2.5271, "step": 49630 }, { "epoch": 1.12, "grad_norm": 1.661705385855994, "learning_rate": 8.556907962683441e-06, "loss": 2.4584, "step": 49640 }, { "epoch": 1.12, "grad_norm": 1.6157307157005998, "learning_rate": 8.553294416197596e-06, "loss": 2.5475, "step": 49650 }, { "epoch": 1.12, "grad_norm": 1.831168225849708, "learning_rate": 8.549681062646445e-06, "loss": 2.4309, "step": 49660 }, { "epoch": 1.12, "grad_norm": 1.9202202619365014, "learning_rate": 8.546067902511882e-06, "loss": 2.5188, "step": 49670 }, { "epoch": 1.12, "grad_norm": 1.571733518181548, "learning_rate": 8.54245493627575e-06, "loss": 2.4597, "step": 49680 }, { "epoch": 1.12, "grad_norm": 1.592192484576376, "learning_rate": 8.538842164419886e-06, "loss": 2.4997, "step": 49690 }, { "epoch": 1.12, "grad_norm": 1.7947467517349869, "learning_rate": 8.535229587426097e-06, "loss": 2.4355, "step": 49700 }, { "epoch": 1.12, "grad_norm": 1.7810052722166922, "learning_rate": 8.531617205776158e-06, "loss": 2.4354, "step": 49710 }, { "epoch": 1.12, "grad_norm": 1.7381187096950756, "learning_rate": 8.528005019951818e-06, "loss": 2.4216, "step": 49720 }, { "epoch": 1.12, "grad_norm": 1.4785999305955835, "learning_rate": 8.524393030434807e-06, "loss": 2.5107, "step": 49730 }, { "epoch": 1.12, "grad_norm": 1.7098494656698007, "learning_rate": 8.520781237706828e-06, "loss": 2.4373, "step": 49740 }, { "epoch": 1.12, "grad_norm": 1.7907547923968836, "learning_rate": 8.517169642249548e-06, "loss": 2.5261, "step": 49750 }, { "epoch": 1.12, "grad_norm": 1.721862894801573, "learning_rate": 8.513558244544618e-06, "loss": 2.4393, "step": 49760 }, { "epoch": 1.12, "grad_norm": 1.840654685115557, "learning_rate": 8.509947045073667e-06, "loss": 2.4262, "step": 49770 }, { "epoch": 1.12, "grad_norm": 1.5862410291879971, "learning_rate": 8.506336044318274e-06, "loss": 2.4526, "step": 49780 }, { "epoch": 1.12, "grad_norm": 1.8001320013710778, "learning_rate": 8.502725242760015e-06, "loss": 2.3674, "step": 49790 }, { "epoch": 1.12, "grad_norm": 1.687305655674626, "learning_rate": 8.499114640880433e-06, "loss": 2.4233, "step": 49800 }, { "epoch": 1.12, "grad_norm": 1.7456243254684969, "learning_rate": 8.495504239161041e-06, "loss": 2.4253, "step": 49810 }, { "epoch": 1.12, "grad_norm": 1.6855128384996214, "learning_rate": 8.491894038083326e-06, "loss": 2.3508, "step": 49820 }, { "epoch": 1.12, "grad_norm": 1.9166010889020464, "learning_rate": 8.488284038128754e-06, "loss": 2.4995, "step": 49830 }, { "epoch": 1.12, "grad_norm": 1.5468847226113953, "learning_rate": 8.484674239778755e-06, "loss": 2.3941, "step": 49840 }, { "epoch": 1.12, "grad_norm": 1.6820598148206312, "learning_rate": 8.48106464351474e-06, "loss": 2.5288, "step": 49850 }, { "epoch": 1.12, "grad_norm": 1.7135363412203206, "learning_rate": 8.477455249818087e-06, "loss": 2.4017, "step": 49860 }, { "epoch": 1.12, "grad_norm": 1.6323228352425927, "learning_rate": 8.473846059170152e-06, "loss": 2.3378, "step": 49870 }, { "epoch": 1.12, "grad_norm": 1.7748517815288563, "learning_rate": 8.470237072052265e-06, "loss": 2.3602, "step": 49880 }, { "epoch": 1.13, "grad_norm": 1.559479044776533, "learning_rate": 8.466628288945718e-06, "loss": 2.5042, "step": 49890 }, { "epoch": 1.13, "grad_norm": 1.6893147088700067, "learning_rate": 8.463019710331788e-06, "loss": 2.4509, "step": 49900 }, { "epoch": 1.13, "grad_norm": 1.7004468177976062, "learning_rate": 8.459411336691718e-06, "loss": 2.5053, "step": 49910 }, { "epoch": 1.13, "grad_norm": 1.6230139771658758, "learning_rate": 8.455803168506727e-06, "loss": 2.428, "step": 49920 }, { "epoch": 1.13, "grad_norm": 1.6368519504554586, "learning_rate": 8.452195206258008e-06, "loss": 2.44, "step": 49930 }, { "epoch": 1.13, "grad_norm": 1.6998818309521513, "learning_rate": 8.448587450426719e-06, "loss": 2.4981, "step": 49940 }, { "epoch": 1.13, "grad_norm": 1.6283356160465803, "learning_rate": 8.444979901494e-06, "loss": 2.5139, "step": 49950 }, { "epoch": 1.13, "grad_norm": 1.540197601978774, "learning_rate": 8.441372559940952e-06, "loss": 2.3167, "step": 49960 }, { "epoch": 1.13, "grad_norm": 1.757591192544961, "learning_rate": 8.437765426248665e-06, "loss": 2.478, "step": 49970 }, { "epoch": 1.13, "grad_norm": 1.6287139918800004, "learning_rate": 8.434158500898185e-06, "loss": 2.5464, "step": 49980 }, { "epoch": 1.13, "grad_norm": 1.6613161857698486, "learning_rate": 8.430551784370539e-06, "loss": 2.4962, "step": 49990 }, { "epoch": 1.13, "grad_norm": 1.6589906990151604, "learning_rate": 8.426945277146721e-06, "loss": 2.529, "step": 50000 }, { "epoch": 1.13, "grad_norm": 1.7827726934420665, "learning_rate": 8.423338979707702e-06, "loss": 2.3434, "step": 50010 }, { "epoch": 1.13, "grad_norm": 1.7633995032950667, "learning_rate": 8.419732892534421e-06, "loss": 2.2874, "step": 50020 }, { "epoch": 1.13, "grad_norm": 1.3267020139144527, "learning_rate": 8.416127016107795e-06, "loss": 2.4154, "step": 50030 }, { "epoch": 1.13, "grad_norm": 1.679259488896468, "learning_rate": 8.412521350908705e-06, "loss": 2.4466, "step": 50040 }, { "epoch": 1.13, "grad_norm": 1.598810415709261, "learning_rate": 8.408915897418007e-06, "loss": 2.4393, "step": 50050 }, { "epoch": 1.13, "grad_norm": 1.657022191125893, "learning_rate": 8.405310656116533e-06, "loss": 2.398, "step": 50060 }, { "epoch": 1.13, "grad_norm": 1.6173084658244035, "learning_rate": 8.401705627485082e-06, "loss": 2.3624, "step": 50070 }, { "epoch": 1.13, "grad_norm": 1.629978069453322, "learning_rate": 8.398100812004423e-06, "loss": 2.5147, "step": 50080 }, { "epoch": 1.13, "grad_norm": 1.549318380819077, "learning_rate": 8.394496210155301e-06, "loss": 2.377, "step": 50090 }, { "epoch": 1.13, "grad_norm": 1.704078960363473, "learning_rate": 8.390891822418439e-06, "loss": 2.436, "step": 50100 }, { "epoch": 1.13, "grad_norm": 1.7454228961139255, "learning_rate": 8.387287649274508e-06, "loss": 2.5376, "step": 50110 }, { "epoch": 1.13, "grad_norm": 1.7332199597694007, "learning_rate": 8.383683691204171e-06, "loss": 2.6195, "step": 50120 }, { "epoch": 1.13, "grad_norm": 1.853037386870641, "learning_rate": 8.380079948688064e-06, "loss": 2.5212, "step": 50130 }, { "epoch": 1.13, "grad_norm": 1.6366373138988344, "learning_rate": 8.376476422206779e-06, "loss": 2.2669, "step": 50140 }, { "epoch": 1.13, "grad_norm": 1.9107077649971314, "learning_rate": 8.372873112240888e-06, "loss": 2.548, "step": 50150 }, { "epoch": 1.13, "grad_norm": 1.7759938398430408, "learning_rate": 8.36927001927094e-06, "loss": 2.4719, "step": 50160 }, { "epoch": 1.13, "grad_norm": 1.6974839374419548, "learning_rate": 8.365667143777444e-06, "loss": 2.5355, "step": 50170 }, { "epoch": 1.13, "grad_norm": 1.6740981220444533, "learning_rate": 8.362064486240885e-06, "loss": 2.4499, "step": 50180 }, { "epoch": 1.13, "grad_norm": 1.637462667604503, "learning_rate": 8.358462047141716e-06, "loss": 2.3733, "step": 50190 }, { "epoch": 1.13, "grad_norm": 1.5632936443307994, "learning_rate": 8.354859826960367e-06, "loss": 2.4381, "step": 50200 }, { "epoch": 1.13, "grad_norm": 1.9026030126313014, "learning_rate": 8.351257826177238e-06, "loss": 2.5172, "step": 50210 }, { "epoch": 1.13, "grad_norm": 1.6755100933073992, "learning_rate": 8.34765604527269e-06, "loss": 2.4175, "step": 50220 }, { "epoch": 1.13, "grad_norm": 1.6406050381447999, "learning_rate": 8.344054484727065e-06, "loss": 2.618, "step": 50230 }, { "epoch": 1.13, "grad_norm": 1.6757695950439329, "learning_rate": 8.34045314502067e-06, "loss": 2.4825, "step": 50240 }, { "epoch": 1.13, "grad_norm": 1.6697336991881262, "learning_rate": 8.336852026633787e-06, "loss": 2.5461, "step": 50250 }, { "epoch": 1.13, "grad_norm": 1.6972502984197122, "learning_rate": 8.333251130046667e-06, "loss": 2.5359, "step": 50260 }, { "epoch": 1.13, "grad_norm": 1.6178919938231648, "learning_rate": 8.329650455739529e-06, "loss": 2.4394, "step": 50270 }, { "epoch": 1.13, "grad_norm": 1.5448393050865465, "learning_rate": 8.326050004192564e-06, "loss": 2.4885, "step": 50280 }, { "epoch": 1.13, "grad_norm": 1.833339396426478, "learning_rate": 8.322449775885933e-06, "loss": 2.476, "step": 50290 }, { "epoch": 1.13, "grad_norm": 1.5790508657937723, "learning_rate": 8.318849771299768e-06, "loss": 2.3546, "step": 50300 }, { "epoch": 1.13, "grad_norm": 1.7579083036158076, "learning_rate": 8.315249990914175e-06, "loss": 2.5889, "step": 50310 }, { "epoch": 1.13, "grad_norm": 1.5668469303111772, "learning_rate": 8.311650435209218e-06, "loss": 2.333, "step": 50320 }, { "epoch": 1.13, "grad_norm": 1.7514855114750294, "learning_rate": 8.308051104664944e-06, "loss": 2.403, "step": 50330 }, { "epoch": 1.14, "grad_norm": 1.667133905363491, "learning_rate": 8.304451999761362e-06, "loss": 2.4167, "step": 50340 }, { "epoch": 1.14, "grad_norm": 2.0148482505562506, "learning_rate": 8.300853120978453e-06, "loss": 2.4327, "step": 50350 }, { "epoch": 1.14, "grad_norm": 1.6335993031388976, "learning_rate": 8.297254468796173e-06, "loss": 2.41, "step": 50360 }, { "epoch": 1.14, "grad_norm": 1.7200024354277446, "learning_rate": 8.293656043694443e-06, "loss": 2.5473, "step": 50370 }, { "epoch": 1.14, "grad_norm": 1.9048861937636012, "learning_rate": 8.29005784615315e-06, "loss": 2.451, "step": 50380 }, { "epoch": 1.14, "grad_norm": 1.511354922986199, "learning_rate": 8.286459876652155e-06, "loss": 2.4797, "step": 50390 }, { "epoch": 1.14, "grad_norm": 1.7660852492850634, "learning_rate": 8.282862135671293e-06, "loss": 2.4418, "step": 50400 }, { "epoch": 1.14, "grad_norm": 1.6867296425288716, "learning_rate": 8.27926462369036e-06, "loss": 2.4568, "step": 50410 }, { "epoch": 1.14, "grad_norm": 1.512736818282505, "learning_rate": 8.275667341189127e-06, "loss": 2.4903, "step": 50420 }, { "epoch": 1.14, "grad_norm": 1.7393806813920207, "learning_rate": 8.272070288647332e-06, "loss": 2.4137, "step": 50430 }, { "epoch": 1.14, "grad_norm": 1.5660117292786948, "learning_rate": 8.26847346654468e-06, "loss": 2.4128, "step": 50440 }, { "epoch": 1.14, "grad_norm": 1.7609344331929067, "learning_rate": 8.264876875360852e-06, "loss": 2.3639, "step": 50450 }, { "epoch": 1.14, "grad_norm": 1.6683192695454616, "learning_rate": 8.261280515575493e-06, "loss": 2.6461, "step": 50460 }, { "epoch": 1.14, "grad_norm": 1.5864217831340977, "learning_rate": 8.25768438766822e-06, "loss": 2.5245, "step": 50470 }, { "epoch": 1.14, "grad_norm": 1.9434876414768885, "learning_rate": 8.254088492118617e-06, "loss": 2.5112, "step": 50480 }, { "epoch": 1.14, "grad_norm": 1.6865376443804276, "learning_rate": 8.250492829406235e-06, "loss": 2.4648, "step": 50490 }, { "epoch": 1.14, "grad_norm": 1.8315922130692828, "learning_rate": 8.246897400010602e-06, "loss": 2.3526, "step": 50500 }, { "epoch": 1.14, "grad_norm": 1.893337063077462, "learning_rate": 8.243302204411207e-06, "loss": 2.4512, "step": 50510 }, { "epoch": 1.14, "grad_norm": 1.601106909721551, "learning_rate": 8.239707243087506e-06, "loss": 2.3548, "step": 50520 }, { "epoch": 1.14, "grad_norm": 1.56093112669382, "learning_rate": 8.236112516518941e-06, "loss": 2.4689, "step": 50530 }, { "epoch": 1.14, "grad_norm": 1.6390886306915997, "learning_rate": 8.232518025184895e-06, "loss": 2.5627, "step": 50540 }, { "epoch": 1.14, "grad_norm": 1.5582488093168512, "learning_rate": 8.22892376956474e-06, "loss": 2.4739, "step": 50550 }, { "epoch": 1.14, "grad_norm": 1.7276685957232782, "learning_rate": 8.225329750137813e-06, "loss": 2.4739, "step": 50560 }, { "epoch": 1.14, "grad_norm": 1.6093509657657377, "learning_rate": 8.221735967383416e-06, "loss": 2.3112, "step": 50570 }, { "epoch": 1.14, "grad_norm": 1.727879427360055, "learning_rate": 8.218142421780817e-06, "loss": 2.3328, "step": 50580 }, { "epoch": 1.14, "grad_norm": 1.6549801642818316, "learning_rate": 8.214549113809264e-06, "loss": 2.4615, "step": 50590 }, { "epoch": 1.14, "grad_norm": 1.6329301236732834, "learning_rate": 8.210956043947962e-06, "loss": 2.5017, "step": 50600 }, { "epoch": 1.14, "grad_norm": 1.7080118147930365, "learning_rate": 8.207363212676088e-06, "loss": 2.5374, "step": 50610 }, { "epoch": 1.14, "grad_norm": 1.6084527064227256, "learning_rate": 8.203770620472783e-06, "loss": 2.443, "step": 50620 }, { "epoch": 1.14, "grad_norm": 1.623809380091574, "learning_rate": 8.200178267817167e-06, "loss": 2.4484, "step": 50630 }, { "epoch": 1.14, "grad_norm": 1.9445018061093569, "learning_rate": 8.19658615518832e-06, "loss": 2.435, "step": 50640 }, { "epoch": 1.14, "grad_norm": 1.603062045815061, "learning_rate": 8.192994283065288e-06, "loss": 2.3936, "step": 50650 }, { "epoch": 1.14, "grad_norm": 1.6537604428154953, "learning_rate": 8.189402651927089e-06, "loss": 2.4282, "step": 50660 }, { "epoch": 1.14, "grad_norm": 1.54927380913895, "learning_rate": 8.185811262252709e-06, "loss": 2.4016, "step": 50670 }, { "epoch": 1.14, "grad_norm": 1.691236321417446, "learning_rate": 8.182220114521098e-06, "loss": 2.5395, "step": 50680 }, { "epoch": 1.14, "grad_norm": 1.687097012621019, "learning_rate": 8.17862920921118e-06, "loss": 2.356, "step": 50690 }, { "epoch": 1.14, "grad_norm": 1.6116442503309945, "learning_rate": 8.175038546801842e-06, "loss": 2.3994, "step": 50700 }, { "epoch": 1.14, "grad_norm": 1.7732806867452406, "learning_rate": 8.171448127771941e-06, "loss": 2.3563, "step": 50710 }, { "epoch": 1.14, "grad_norm": 1.6526902860806312, "learning_rate": 8.167857952600296e-06, "loss": 2.3927, "step": 50720 }, { "epoch": 1.14, "grad_norm": 1.6504128384997852, "learning_rate": 8.164268021765704e-06, "loss": 2.3792, "step": 50730 }, { "epoch": 1.14, "grad_norm": 1.6665168098499779, "learning_rate": 8.16067833574692e-06, "loss": 2.4733, "step": 50740 }, { "epoch": 1.14, "grad_norm": 1.5464162393114476, "learning_rate": 8.157088895022672e-06, "loss": 2.4078, "step": 50750 }, { "epoch": 1.14, "grad_norm": 1.8028815320872642, "learning_rate": 8.15349970007165e-06, "loss": 2.3846, "step": 50760 }, { "epoch": 1.14, "grad_norm": 1.748951687876915, "learning_rate": 8.149910751372513e-06, "loss": 2.3391, "step": 50770 }, { "epoch": 1.15, "grad_norm": 1.6258196397278488, "learning_rate": 8.146322049403888e-06, "loss": 2.4036, "step": 50780 }, { "epoch": 1.15, "grad_norm": 1.7582379151419467, "learning_rate": 8.142733594644374e-06, "loss": 2.3788, "step": 50790 }, { "epoch": 1.15, "grad_norm": 1.897460961159037, "learning_rate": 8.13914538757253e-06, "loss": 2.4423, "step": 50800 }, { "epoch": 1.15, "grad_norm": 1.5876913544179652, "learning_rate": 8.135557428666884e-06, "loss": 2.5664, "step": 50810 }, { "epoch": 1.15, "grad_norm": 1.739189759563084, "learning_rate": 8.13196971840593e-06, "loss": 2.3729, "step": 50820 }, { "epoch": 1.15, "grad_norm": 1.8053895253943453, "learning_rate": 8.128382257268134e-06, "loss": 2.4346, "step": 50830 }, { "epoch": 1.15, "grad_norm": 1.847468799500869, "learning_rate": 8.12479504573192e-06, "loss": 2.5059, "step": 50840 }, { "epoch": 1.15, "grad_norm": 1.5722136121914119, "learning_rate": 8.121208084275692e-06, "loss": 2.3322, "step": 50850 }, { "epoch": 1.15, "grad_norm": 1.4876736659118122, "learning_rate": 8.117621373377802e-06, "loss": 2.425, "step": 50860 }, { "epoch": 1.15, "grad_norm": 1.7217779936390964, "learning_rate": 8.114034913516582e-06, "loss": 2.3996, "step": 50870 }, { "epoch": 1.15, "grad_norm": 1.4879156128977793, "learning_rate": 8.110448705170328e-06, "loss": 2.4696, "step": 50880 }, { "epoch": 1.15, "grad_norm": 1.7997572661695989, "learning_rate": 8.106862748817305e-06, "loss": 2.5048, "step": 50890 }, { "epoch": 1.15, "grad_norm": 1.631186988755308, "learning_rate": 8.103277044935736e-06, "loss": 2.4245, "step": 50900 }, { "epoch": 1.15, "grad_norm": 1.619127171757374, "learning_rate": 8.099691594003821e-06, "loss": 2.4497, "step": 50910 }, { "epoch": 1.15, "grad_norm": 1.709680477738078, "learning_rate": 8.096106396499714e-06, "loss": 2.3512, "step": 50920 }, { "epoch": 1.15, "grad_norm": 1.6131033737684435, "learning_rate": 8.09252145290155e-06, "loss": 2.5335, "step": 50930 }, { "epoch": 1.15, "grad_norm": 1.5862429512329714, "learning_rate": 8.088936763687416e-06, "loss": 2.3518, "step": 50940 }, { "epoch": 1.15, "grad_norm": 1.624759393485588, "learning_rate": 8.085352329335372e-06, "loss": 2.4413, "step": 50950 }, { "epoch": 1.15, "grad_norm": 1.8788273220894098, "learning_rate": 8.081768150323447e-06, "loss": 2.3356, "step": 50960 }, { "epoch": 1.15, "grad_norm": 1.939350479401338, "learning_rate": 8.078184227129631e-06, "loss": 2.4015, "step": 50970 }, { "epoch": 1.15, "grad_norm": 1.5469018611781378, "learning_rate": 8.074600560231874e-06, "loss": 2.4273, "step": 50980 }, { "epoch": 1.15, "grad_norm": 1.8753128264760592, "learning_rate": 8.071017150108109e-06, "loss": 2.5742, "step": 50990 }, { "epoch": 1.15, "grad_norm": 1.5929314275567248, "learning_rate": 8.067433997236218e-06, "loss": 2.4176, "step": 51000 }, { "epoch": 1.15, "grad_norm": 1.5314357741115558, "learning_rate": 8.063851102094059e-06, "loss": 2.5661, "step": 51010 }, { "epoch": 1.15, "grad_norm": 1.7152254436854546, "learning_rate": 8.060268465159446e-06, "loss": 2.4784, "step": 51020 }, { "epoch": 1.15, "grad_norm": 1.6890275075233916, "learning_rate": 8.056686086910173e-06, "loss": 2.3985, "step": 51030 }, { "epoch": 1.15, "grad_norm": 1.5482336369896212, "learning_rate": 8.053103967823988e-06, "loss": 2.541, "step": 51040 }, { "epoch": 1.15, "grad_norm": 1.756679553488535, "learning_rate": 8.049522108378602e-06, "loss": 2.4861, "step": 51050 }, { "epoch": 1.15, "grad_norm": 1.7448585953483653, "learning_rate": 8.045940509051705e-06, "loss": 2.4019, "step": 51060 }, { "epoch": 1.15, "grad_norm": 1.8593332659662256, "learning_rate": 8.042359170320944e-06, "loss": 2.4995, "step": 51070 }, { "epoch": 1.15, "grad_norm": 1.6363697387110014, "learning_rate": 8.038778092663922e-06, "loss": 2.4965, "step": 51080 }, { "epoch": 1.15, "grad_norm": 1.5265476446351045, "learning_rate": 8.035197276558225e-06, "loss": 2.4113, "step": 51090 }, { "epoch": 1.15, "grad_norm": 1.7254631404586795, "learning_rate": 8.03161672248139e-06, "loss": 2.4997, "step": 51100 }, { "epoch": 1.15, "grad_norm": 1.776146101651609, "learning_rate": 8.028036430910927e-06, "loss": 2.4412, "step": 51110 }, { "epoch": 1.15, "grad_norm": 1.7213517669097582, "learning_rate": 8.024456402324315e-06, "loss": 2.488, "step": 51120 }, { "epoch": 1.15, "grad_norm": 1.4754096429550805, "learning_rate": 8.020876637198984e-06, "loss": 2.4518, "step": 51130 }, { "epoch": 1.15, "grad_norm": 1.6461197502773781, "learning_rate": 8.017297136012338e-06, "loss": 2.5094, "step": 51140 }, { "epoch": 1.15, "grad_norm": 1.6823116371477225, "learning_rate": 8.013717899241743e-06, "loss": 2.4934, "step": 51150 }, { "epoch": 1.15, "grad_norm": 1.6778228620343207, "learning_rate": 8.010138927364536e-06, "loss": 2.4189, "step": 51160 }, { "epoch": 1.15, "grad_norm": 1.7174958557009619, "learning_rate": 8.006560220858011e-06, "loss": 2.4722, "step": 51170 }, { "epoch": 1.15, "grad_norm": 1.7581211208685086, "learning_rate": 8.00298178019943e-06, "loss": 2.4569, "step": 51180 }, { "epoch": 1.15, "grad_norm": 1.683419158344231, "learning_rate": 7.999403605866018e-06, "loss": 2.4559, "step": 51190 }, { "epoch": 1.15, "grad_norm": 1.763442995607547, "learning_rate": 7.995825698334964e-06, "loss": 2.3765, "step": 51200 }, { "epoch": 1.15, "grad_norm": 1.553611796606995, "learning_rate": 7.992248058083423e-06, "loss": 2.3693, "step": 51210 }, { "epoch": 1.16, "grad_norm": 1.7565990442902735, "learning_rate": 7.988670685588517e-06, "loss": 2.3673, "step": 51220 }, { "epoch": 1.16, "grad_norm": 1.7533391949969224, "learning_rate": 7.985093581327328e-06, "loss": 2.471, "step": 51230 }, { "epoch": 1.16, "grad_norm": 1.7182076770574244, "learning_rate": 7.981516745776904e-06, "loss": 2.4386, "step": 51240 }, { "epoch": 1.16, "grad_norm": 1.5763356450499804, "learning_rate": 7.977940179414255e-06, "loss": 2.3068, "step": 51250 }, { "epoch": 1.16, "grad_norm": 1.6463456706648933, "learning_rate": 7.974363882716358e-06, "loss": 2.5921, "step": 51260 }, { "epoch": 1.16, "grad_norm": 1.7416289424684739, "learning_rate": 7.970787856160155e-06, "loss": 2.4673, "step": 51270 }, { "epoch": 1.16, "grad_norm": 1.7463417623338706, "learning_rate": 7.967212100222551e-06, "loss": 2.3807, "step": 51280 }, { "epoch": 1.16, "grad_norm": 1.7632373677726383, "learning_rate": 7.963636615380408e-06, "loss": 2.4318, "step": 51290 }, { "epoch": 1.16, "grad_norm": 1.7817782512381295, "learning_rate": 7.960061402110562e-06, "loss": 2.3982, "step": 51300 }, { "epoch": 1.16, "grad_norm": 1.4964533506151687, "learning_rate": 7.956486460889804e-06, "loss": 2.3389, "step": 51310 }, { "epoch": 1.16, "grad_norm": 1.5595383122176227, "learning_rate": 7.952911792194899e-06, "loss": 2.3726, "step": 51320 }, { "epoch": 1.16, "grad_norm": 1.7715982255591325, "learning_rate": 7.949337396502568e-06, "loss": 2.5113, "step": 51330 }, { "epoch": 1.16, "grad_norm": 1.6034464103450528, "learning_rate": 7.945763274289497e-06, "loss": 2.5204, "step": 51340 }, { "epoch": 1.16, "grad_norm": 1.504036568253478, "learning_rate": 7.942189426032332e-06, "loss": 2.3805, "step": 51350 }, { "epoch": 1.16, "grad_norm": 1.5288807876163806, "learning_rate": 7.938615852207695e-06, "loss": 2.512, "step": 51360 }, { "epoch": 1.16, "grad_norm": 1.6034792864965226, "learning_rate": 7.935042553292156e-06, "loss": 2.4555, "step": 51370 }, { "epoch": 1.16, "grad_norm": 1.549277162656117, "learning_rate": 7.93146952976226e-06, "loss": 2.4571, "step": 51380 }, { "epoch": 1.16, "grad_norm": 1.673882511019284, "learning_rate": 7.927896782094507e-06, "loss": 2.37, "step": 51390 }, { "epoch": 1.16, "grad_norm": 1.611172570740646, "learning_rate": 7.924324310765364e-06, "loss": 2.3805, "step": 51400 }, { "epoch": 1.16, "grad_norm": 1.6070252145546746, "learning_rate": 7.920752116251258e-06, "loss": 2.2981, "step": 51410 }, { "epoch": 1.16, "grad_norm": 1.8485735094670943, "learning_rate": 7.917180199028588e-06, "loss": 2.5528, "step": 51420 }, { "epoch": 1.16, "grad_norm": 1.552086726785364, "learning_rate": 7.913608559573707e-06, "loss": 2.4884, "step": 51430 }, { "epoch": 1.16, "grad_norm": 1.8999092070381953, "learning_rate": 7.910037198362936e-06, "loss": 2.5298, "step": 51440 }, { "epoch": 1.16, "grad_norm": 1.5686630817640776, "learning_rate": 7.90646611587255e-06, "loss": 2.4982, "step": 51450 }, { "epoch": 1.16, "grad_norm": 1.674308670464022, "learning_rate": 7.902895312578801e-06, "loss": 2.4467, "step": 51460 }, { "epoch": 1.16, "grad_norm": 1.5817637672316887, "learning_rate": 7.899324788957895e-06, "loss": 2.4445, "step": 51470 }, { "epoch": 1.16, "grad_norm": 1.6272389341101188, "learning_rate": 7.895754545486e-06, "loss": 2.4478, "step": 51480 }, { "epoch": 1.16, "grad_norm": 1.7244414982843734, "learning_rate": 7.892184582639248e-06, "loss": 2.4725, "step": 51490 }, { "epoch": 1.16, "grad_norm": 1.647985594035388, "learning_rate": 7.88861490089374e-06, "loss": 2.421, "step": 51500 }, { "epoch": 1.16, "grad_norm": 1.5800805732456662, "learning_rate": 7.885045500725527e-06, "loss": 2.3371, "step": 51510 }, { "epoch": 1.16, "grad_norm": 1.5729747651338029, "learning_rate": 7.881476382610632e-06, "loss": 2.5201, "step": 51520 }, { "epoch": 1.16, "grad_norm": 1.6507080844759738, "learning_rate": 7.877907547025037e-06, "loss": 2.4976, "step": 51530 }, { "epoch": 1.16, "grad_norm": 1.7116257142750875, "learning_rate": 7.874338994444687e-06, "loss": 2.4543, "step": 51540 }, { "epoch": 1.16, "grad_norm": 1.4159797946269657, "learning_rate": 7.87077072534549e-06, "loss": 2.4549, "step": 51550 }, { "epoch": 1.16, "grad_norm": 1.6598992589156796, "learning_rate": 7.867202740203314e-06, "loss": 2.4396, "step": 51560 }, { "epoch": 1.16, "grad_norm": 1.6063253034887792, "learning_rate": 7.863635039493995e-06, "loss": 2.4865, "step": 51570 }, { "epoch": 1.16, "grad_norm": 1.395803118363865, "learning_rate": 7.860067623693318e-06, "loss": 2.4639, "step": 51580 }, { "epoch": 1.16, "grad_norm": 1.5333334432180183, "learning_rate": 7.856500493277048e-06, "loss": 2.4777, "step": 51590 }, { "epoch": 1.16, "grad_norm": 1.539960879212446, "learning_rate": 7.852933648720897e-06, "loss": 2.476, "step": 51600 }, { "epoch": 1.16, "grad_norm": 1.706387009259533, "learning_rate": 7.849367090500549e-06, "loss": 2.4283, "step": 51610 }, { "epoch": 1.16, "grad_norm": 1.6235528397969143, "learning_rate": 7.84580081909164e-06, "loss": 2.4564, "step": 51620 }, { "epoch": 1.16, "grad_norm": 1.584908411674806, "learning_rate": 7.842234834969774e-06, "loss": 2.5442, "step": 51630 }, { "epoch": 1.16, "grad_norm": 1.7743692221238687, "learning_rate": 7.838669138610515e-06, "loss": 2.4384, "step": 51640 }, { "epoch": 1.16, "grad_norm": 1.6327448654044792, "learning_rate": 7.835103730489393e-06, "loss": 2.4546, "step": 51650 }, { "epoch": 1.16, "grad_norm": 1.578822790624881, "learning_rate": 7.831538611081894e-06, "loss": 2.3283, "step": 51660 }, { "epoch": 1.17, "grad_norm": 1.6668365331266934, "learning_rate": 7.827973780863467e-06, "loss": 2.4916, "step": 51670 }, { "epoch": 1.17, "grad_norm": 1.7411097006919596, "learning_rate": 7.824409240309522e-06, "loss": 2.4294, "step": 51680 }, { "epoch": 1.17, "grad_norm": 1.5947337253351157, "learning_rate": 7.820844989895433e-06, "loss": 2.484, "step": 51690 }, { "epoch": 1.17, "grad_norm": 1.6655934914117263, "learning_rate": 7.817281030096533e-06, "loss": 2.4089, "step": 51700 }, { "epoch": 1.17, "grad_norm": 1.8716304743934762, "learning_rate": 7.813717361388117e-06, "loss": 2.4466, "step": 51710 }, { "epoch": 1.17, "grad_norm": 1.8152068603368288, "learning_rate": 7.81015398424544e-06, "loss": 2.4426, "step": 51720 }, { "epoch": 1.17, "grad_norm": 1.5157152442043185, "learning_rate": 7.80659089914372e-06, "loss": 2.5778, "step": 51730 }, { "epoch": 1.17, "grad_norm": 1.6019812750604099, "learning_rate": 7.803028106558132e-06, "loss": 2.5549, "step": 51740 }, { "epoch": 1.17, "grad_norm": 1.742236951115679, "learning_rate": 7.799465606963818e-06, "loss": 2.3464, "step": 51750 }, { "epoch": 1.17, "grad_norm": 1.5879562848091304, "learning_rate": 7.795903400835878e-06, "loss": 2.3952, "step": 51760 }, { "epoch": 1.17, "grad_norm": 1.768063041284617, "learning_rate": 7.792341488649372e-06, "loss": 2.5185, "step": 51770 }, { "epoch": 1.17, "grad_norm": 1.8338796682671323, "learning_rate": 7.78877987087932e-06, "loss": 2.4512, "step": 51780 }, { "epoch": 1.17, "grad_norm": 1.4940254374970436, "learning_rate": 7.785218548000709e-06, "loss": 2.4024, "step": 51790 }, { "epoch": 1.17, "grad_norm": 1.64407648823665, "learning_rate": 7.781657520488478e-06, "loss": 2.4201, "step": 51800 }, { "epoch": 1.17, "grad_norm": 1.5974477140262406, "learning_rate": 7.778096788817532e-06, "loss": 2.4964, "step": 51810 }, { "epoch": 1.17, "grad_norm": 1.556921651957691, "learning_rate": 7.774536353462736e-06, "loss": 2.3928, "step": 51820 }, { "epoch": 1.17, "grad_norm": 1.618299041566736, "learning_rate": 7.770976214898913e-06, "loss": 2.4927, "step": 51830 }, { "epoch": 1.17, "grad_norm": 1.6123532198193855, "learning_rate": 7.767416373600847e-06, "loss": 2.5421, "step": 51840 }, { "epoch": 1.17, "grad_norm": 1.7542904347043848, "learning_rate": 7.763856830043287e-06, "loss": 2.3908, "step": 51850 }, { "epoch": 1.17, "grad_norm": 1.853603714211233, "learning_rate": 7.760297584700937e-06, "loss": 2.5142, "step": 51860 }, { "epoch": 1.17, "grad_norm": 1.5240443391696767, "learning_rate": 7.756738638048464e-06, "loss": 2.4902, "step": 51870 }, { "epoch": 1.17, "grad_norm": 1.541653453429103, "learning_rate": 7.75317999056049e-06, "loss": 2.5382, "step": 51880 }, { "epoch": 1.17, "grad_norm": 1.537386522917413, "learning_rate": 7.749621642711607e-06, "loss": 2.4107, "step": 51890 }, { "epoch": 1.17, "grad_norm": 1.5916003665851064, "learning_rate": 7.746063594976358e-06, "loss": 2.4363, "step": 51900 }, { "epoch": 1.17, "grad_norm": 1.7192492686450833, "learning_rate": 7.742505847829251e-06, "loss": 2.5519, "step": 51910 }, { "epoch": 1.17, "grad_norm": 1.6927362137549582, "learning_rate": 7.73894840174475e-06, "loss": 2.3888, "step": 51920 }, { "epoch": 1.17, "grad_norm": 1.6620707161291726, "learning_rate": 7.735391257197288e-06, "loss": 2.4852, "step": 51930 }, { "epoch": 1.17, "grad_norm": 1.7578976334053333, "learning_rate": 7.731834414661238e-06, "loss": 2.4328, "step": 51940 }, { "epoch": 1.17, "grad_norm": 1.5588139682670543, "learning_rate": 7.728277874610955e-06, "loss": 2.5397, "step": 51950 }, { "epoch": 1.17, "grad_norm": 1.6866546426924744, "learning_rate": 7.724721637520742e-06, "loss": 2.4396, "step": 51960 }, { "epoch": 1.17, "grad_norm": 1.6219031734578129, "learning_rate": 7.721165703864864e-06, "loss": 2.3855, "step": 51970 }, { "epoch": 1.17, "grad_norm": 1.7329394446160167, "learning_rate": 7.71761007411754e-06, "loss": 2.4377, "step": 51980 }, { "epoch": 1.17, "grad_norm": 1.5922398713732855, "learning_rate": 7.714054748752963e-06, "loss": 2.3876, "step": 51990 }, { "epoch": 1.17, "grad_norm": 1.5510985397603863, "learning_rate": 7.710499728245273e-06, "loss": 2.4559, "step": 52000 }, { "epoch": 1.17, "grad_norm": 1.8823927446874853, "learning_rate": 7.70694501306857e-06, "loss": 2.4389, "step": 52010 }, { "epoch": 1.17, "grad_norm": 1.727953302618443, "learning_rate": 7.703390603696916e-06, "loss": 2.2995, "step": 52020 }, { "epoch": 1.17, "grad_norm": 1.653765531961778, "learning_rate": 7.699836500604336e-06, "loss": 2.4197, "step": 52030 }, { "epoch": 1.17, "grad_norm": 1.922149864133498, "learning_rate": 7.696282704264809e-06, "loss": 2.4235, "step": 52040 }, { "epoch": 1.17, "grad_norm": 1.4515363236058587, "learning_rate": 7.69272921515227e-06, "loss": 2.3453, "step": 52050 }, { "epoch": 1.17, "grad_norm": 1.762256626271492, "learning_rate": 7.689176033740623e-06, "loss": 2.3809, "step": 52060 }, { "epoch": 1.17, "grad_norm": 1.7169544870828974, "learning_rate": 7.685623160503719e-06, "loss": 2.4473, "step": 52070 }, { "epoch": 1.17, "grad_norm": 1.6190286381865626, "learning_rate": 7.682070595915378e-06, "loss": 2.4792, "step": 52080 }, { "epoch": 1.17, "grad_norm": 1.6828339524885667, "learning_rate": 7.678518340449376e-06, "loss": 2.389, "step": 52090 }, { "epoch": 1.17, "grad_norm": 1.6833751461682707, "learning_rate": 7.674966394579445e-06, "loss": 2.4757, "step": 52100 }, { "epoch": 1.18, "grad_norm": 2.0132400439370097, "learning_rate": 7.671414758779276e-06, "loss": 2.495, "step": 52110 }, { "epoch": 1.18, "grad_norm": 1.5014474975852525, "learning_rate": 7.667863433522525e-06, "loss": 2.4474, "step": 52120 }, { "epoch": 1.18, "grad_norm": 1.7645190225907028, "learning_rate": 7.6643124192828e-06, "loss": 2.3302, "step": 52130 }, { "epoch": 1.18, "grad_norm": 1.6639630001378574, "learning_rate": 7.660761716533667e-06, "loss": 2.4609, "step": 52140 }, { "epoch": 1.18, "grad_norm": 1.5353780312759688, "learning_rate": 7.657211325748655e-06, "loss": 2.4799, "step": 52150 }, { "epoch": 1.18, "grad_norm": 1.8339431418079404, "learning_rate": 7.653661247401247e-06, "loss": 2.4367, "step": 52160 }, { "epoch": 1.18, "grad_norm": 1.7361031110469705, "learning_rate": 7.650111481964886e-06, "loss": 2.3576, "step": 52170 }, { "epoch": 1.18, "grad_norm": 1.792254043901839, "learning_rate": 7.646562029912975e-06, "loss": 2.4735, "step": 52180 }, { "epoch": 1.18, "grad_norm": 1.621531539095002, "learning_rate": 7.64301289171888e-06, "loss": 2.3421, "step": 52190 }, { "epoch": 1.18, "grad_norm": 1.719965653882792, "learning_rate": 7.639464067855908e-06, "loss": 2.5106, "step": 52200 }, { "epoch": 1.18, "grad_norm": 1.5368236793650027, "learning_rate": 7.635915558797339e-06, "loss": 2.4347, "step": 52210 }, { "epoch": 1.18, "grad_norm": 1.827105372537391, "learning_rate": 7.632367365016412e-06, "loss": 2.4336, "step": 52220 }, { "epoch": 1.18, "grad_norm": 1.5152109353689558, "learning_rate": 7.628819486986315e-06, "loss": 2.5179, "step": 52230 }, { "epoch": 1.18, "grad_norm": 1.6542460559671412, "learning_rate": 7.6252719251802e-06, "loss": 2.4092, "step": 52240 }, { "epoch": 1.18, "grad_norm": 1.9005446507301718, "learning_rate": 7.621724680071169e-06, "loss": 2.4235, "step": 52250 }, { "epoch": 1.18, "grad_norm": 1.7679695687500643, "learning_rate": 7.6181777521323006e-06, "loss": 2.4326, "step": 52260 }, { "epoch": 1.18, "grad_norm": 1.7138933735717206, "learning_rate": 7.6146311418366035e-06, "loss": 2.4647, "step": 52270 }, { "epoch": 1.18, "grad_norm": 1.7154141392332127, "learning_rate": 7.611084849657067e-06, "loss": 2.3837, "step": 52280 }, { "epoch": 1.18, "grad_norm": 1.800756912984294, "learning_rate": 7.607538876066627e-06, "loss": 2.3392, "step": 52290 }, { "epoch": 1.18, "grad_norm": 1.8404518165578576, "learning_rate": 7.60399322153818e-06, "loss": 2.4033, "step": 52300 }, { "epoch": 1.18, "grad_norm": 1.5257922397415096, "learning_rate": 7.600447886544578e-06, "loss": 2.5102, "step": 52310 }, { "epoch": 1.18, "grad_norm": 1.5575096186880866, "learning_rate": 7.596902871558636e-06, "loss": 2.3712, "step": 52320 }, { "epoch": 1.18, "grad_norm": 1.7218858768388963, "learning_rate": 7.593358177053119e-06, "loss": 2.377, "step": 52330 }, { "epoch": 1.18, "grad_norm": 1.8514913713425476, "learning_rate": 7.589813803500754e-06, "loss": 2.4058, "step": 52340 }, { "epoch": 1.18, "grad_norm": 1.711199566174259, "learning_rate": 7.58626975137422e-06, "loss": 2.4271, "step": 52350 }, { "epoch": 1.18, "grad_norm": 1.7086317592809257, "learning_rate": 7.582726021146166e-06, "loss": 2.5108, "step": 52360 }, { "epoch": 1.18, "grad_norm": 1.6756749991547903, "learning_rate": 7.579182613289175e-06, "loss": 2.4409, "step": 52370 }, { "epoch": 1.18, "grad_norm": 1.7004585216852186, "learning_rate": 7.575639528275814e-06, "loss": 2.4352, "step": 52380 }, { "epoch": 1.18, "grad_norm": 1.7138156433059317, "learning_rate": 7.572096766578585e-06, "loss": 2.417, "step": 52390 }, { "epoch": 1.18, "grad_norm": 1.7295631205869144, "learning_rate": 7.568554328669959e-06, "loss": 2.4746, "step": 52400 }, { "epoch": 1.18, "grad_norm": 1.7771563817093743, "learning_rate": 7.5650122150223595e-06, "loss": 2.4745, "step": 52410 }, { "epoch": 1.18, "grad_norm": 1.6418725526664255, "learning_rate": 7.561470426108169e-06, "loss": 2.5165, "step": 52420 }, { "epoch": 1.18, "grad_norm": 1.651229839982824, "learning_rate": 7.5579289623997275e-06, "loss": 2.4189, "step": 52430 }, { "epoch": 1.18, "grad_norm": 1.5895608124715914, "learning_rate": 7.554387824369326e-06, "loss": 2.4879, "step": 52440 }, { "epoch": 1.18, "grad_norm": 1.8473235126866092, "learning_rate": 7.5508470124892145e-06, "loss": 2.4392, "step": 52450 }, { "epoch": 1.18, "grad_norm": 1.5963449110257695, "learning_rate": 7.547306527231606e-06, "loss": 2.4495, "step": 52460 }, { "epoch": 1.18, "grad_norm": 1.6318571561760244, "learning_rate": 7.543766369068664e-06, "loss": 2.4652, "step": 52470 }, { "epoch": 1.18, "grad_norm": 1.7191008505701504, "learning_rate": 7.5402265384725035e-06, "loss": 2.3946, "step": 52480 }, { "epoch": 1.18, "grad_norm": 1.4408452779918477, "learning_rate": 7.5366870359152046e-06, "loss": 2.4315, "step": 52490 }, { "epoch": 1.18, "grad_norm": 1.6252838335247195, "learning_rate": 7.533147861868801e-06, "loss": 2.3328, "step": 52500 }, { "epoch": 1.18, "grad_norm": 1.6645593088765143, "learning_rate": 7.529609016805278e-06, "loss": 2.4145, "step": 52510 }, { "epoch": 1.18, "grad_norm": 1.7104620287966439, "learning_rate": 7.526070501196587e-06, "loss": 2.3692, "step": 52520 }, { "epoch": 1.18, "grad_norm": 1.5731606825580233, "learning_rate": 7.522532315514627e-06, "loss": 2.4392, "step": 52530 }, { "epoch": 1.18, "grad_norm": 1.957763715266679, "learning_rate": 7.5189944602312546e-06, "loss": 2.223, "step": 52540 }, { "epoch": 1.19, "grad_norm": 1.6939461709804535, "learning_rate": 7.515456935818282e-06, "loss": 2.5185, "step": 52550 }, { "epoch": 1.19, "grad_norm": 1.6459650254260432, "learning_rate": 7.511919742747481e-06, "loss": 2.4769, "step": 52560 }, { "epoch": 1.19, "grad_norm": 1.672271792730935, "learning_rate": 7.508382881490577e-06, "loss": 2.4559, "step": 52570 }, { "epoch": 1.19, "grad_norm": 1.5592306161808482, "learning_rate": 7.504846352519251e-06, "loss": 2.462, "step": 52580 }, { "epoch": 1.19, "grad_norm": 2.1539136159462045, "learning_rate": 7.501310156305135e-06, "loss": 2.5915, "step": 52590 }, { "epoch": 1.19, "grad_norm": 1.769053469043141, "learning_rate": 7.497774293319826e-06, "loss": 2.4763, "step": 52600 }, { "epoch": 1.19, "grad_norm": 1.6426511120095113, "learning_rate": 7.494238764034867e-06, "loss": 2.4275, "step": 52610 }, { "epoch": 1.19, "grad_norm": 1.4874268227456595, "learning_rate": 7.490703568921765e-06, "loss": 2.4217, "step": 52620 }, { "epoch": 1.19, "grad_norm": 1.5315791786125457, "learning_rate": 7.48716870845198e-06, "loss": 2.4201, "step": 52630 }, { "epoch": 1.19, "grad_norm": 1.5736308318142787, "learning_rate": 7.48363418309692e-06, "loss": 2.4712, "step": 52640 }, { "epoch": 1.19, "grad_norm": 1.6255560675766747, "learning_rate": 7.480099993327959e-06, "loss": 2.4537, "step": 52650 }, { "epoch": 1.19, "grad_norm": 1.5753171655247584, "learning_rate": 7.476566139616421e-06, "loss": 2.4237, "step": 52660 }, { "epoch": 1.19, "grad_norm": 1.6903624090932965, "learning_rate": 7.473032622433587e-06, "loss": 2.5052, "step": 52670 }, { "epoch": 1.19, "grad_norm": 1.6045960639923533, "learning_rate": 7.469499442250686e-06, "loss": 2.4304, "step": 52680 }, { "epoch": 1.19, "grad_norm": 1.6054753790621685, "learning_rate": 7.465966599538919e-06, "loss": 2.4213, "step": 52690 }, { "epoch": 1.19, "grad_norm": 1.5977564549081906, "learning_rate": 7.462434094769418e-06, "loss": 2.484, "step": 52700 }, { "epoch": 1.19, "grad_norm": 1.802738323337742, "learning_rate": 7.4589019284132895e-06, "loss": 2.4009, "step": 52710 }, { "epoch": 1.19, "grad_norm": 1.541103267716756, "learning_rate": 7.455370100941588e-06, "loss": 2.452, "step": 52720 }, { "epoch": 1.19, "grad_norm": 1.7381100628437753, "learning_rate": 7.451838612825322e-06, "loss": 2.3006, "step": 52730 }, { "epoch": 1.19, "grad_norm": 1.5237556768185467, "learning_rate": 7.448307464535454e-06, "loss": 2.4752, "step": 52740 }, { "epoch": 1.19, "grad_norm": 1.593038125932624, "learning_rate": 7.444776656542907e-06, "loss": 2.4899, "step": 52750 }, { "epoch": 1.19, "grad_norm": 1.6069368830665953, "learning_rate": 7.441246189318552e-06, "loss": 2.4715, "step": 52760 }, { "epoch": 1.19, "grad_norm": 1.532311875011228, "learning_rate": 7.437716063333218e-06, "loss": 2.5677, "step": 52770 }, { "epoch": 1.19, "grad_norm": 1.66059495383901, "learning_rate": 7.4341862790576844e-06, "loss": 2.4426, "step": 52780 }, { "epoch": 1.19, "grad_norm": 1.7256536192416367, "learning_rate": 7.430656836962695e-06, "loss": 2.4685, "step": 52790 }, { "epoch": 1.19, "grad_norm": 1.8113531622230834, "learning_rate": 7.427127737518938e-06, "loss": 2.3802, "step": 52800 }, { "epoch": 1.19, "grad_norm": 1.8066564648914099, "learning_rate": 7.423598981197056e-06, "loss": 2.5165, "step": 52810 }, { "epoch": 1.19, "grad_norm": 1.8284058114712871, "learning_rate": 7.420070568467652e-06, "loss": 2.347, "step": 52820 }, { "epoch": 1.19, "grad_norm": 1.6339540848163445, "learning_rate": 7.416542499801278e-06, "loss": 2.3276, "step": 52830 }, { "epoch": 1.19, "grad_norm": 1.6922002655443211, "learning_rate": 7.413014775668444e-06, "loss": 2.4633, "step": 52840 }, { "epoch": 1.19, "grad_norm": 1.6088080643676985, "learning_rate": 7.409487396539612e-06, "loss": 2.4366, "step": 52850 }, { "epoch": 1.19, "grad_norm": 1.9008267860996857, "learning_rate": 7.405960362885198e-06, "loss": 2.413, "step": 52860 }, { "epoch": 1.19, "grad_norm": 1.6844150500585928, "learning_rate": 7.402433675175571e-06, "loss": 2.3402, "step": 52870 }, { "epoch": 1.19, "grad_norm": 1.5736418381953392, "learning_rate": 7.3989073338810555e-06, "loss": 2.4453, "step": 52880 }, { "epoch": 1.19, "grad_norm": 1.574067273279354, "learning_rate": 7.395381339471932e-06, "loss": 2.4234, "step": 52890 }, { "epoch": 1.19, "grad_norm": 1.5713530997856817, "learning_rate": 7.391855692418433e-06, "loss": 2.4504, "step": 52900 }, { "epoch": 1.19, "grad_norm": 1.7532260622776135, "learning_rate": 7.388330393190739e-06, "loss": 2.3887, "step": 52910 }, { "epoch": 1.19, "grad_norm": 1.54516123505659, "learning_rate": 7.38480544225899e-06, "loss": 2.4642, "step": 52920 }, { "epoch": 1.19, "grad_norm": 1.839784370729532, "learning_rate": 7.3812808400932786e-06, "loss": 2.6264, "step": 52930 }, { "epoch": 1.19, "grad_norm": 1.7445451321477627, "learning_rate": 7.377756587163651e-06, "loss": 2.4017, "step": 52940 }, { "epoch": 1.19, "grad_norm": 1.7989702062236015, "learning_rate": 7.374232683940108e-06, "loss": 2.3957, "step": 52950 }, { "epoch": 1.19, "grad_norm": 1.5456154627944567, "learning_rate": 7.370709130892602e-06, "loss": 2.4617, "step": 52960 }, { "epoch": 1.19, "grad_norm": 1.44097131696555, "learning_rate": 7.367185928491039e-06, "loss": 2.4623, "step": 52970 }, { "epoch": 1.19, "grad_norm": 1.894651207401699, "learning_rate": 7.363663077205274e-06, "loss": 2.4282, "step": 52980 }, { "epoch": 1.19, "grad_norm": 1.630025730856033, "learning_rate": 7.360140577505125e-06, "loss": 2.388, "step": 52990 }, { "epoch": 1.2, "grad_norm": 1.8198803272385227, "learning_rate": 7.356618429860356e-06, "loss": 2.4814, "step": 53000 }, { "epoch": 1.2, "grad_norm": 1.8052751323969187, "learning_rate": 7.3530966347406865e-06, "loss": 2.5455, "step": 53010 }, { "epoch": 1.2, "grad_norm": 1.7358999796780743, "learning_rate": 7.349575192615785e-06, "loss": 2.3978, "step": 53020 }, { "epoch": 1.2, "grad_norm": 1.5947913823300728, "learning_rate": 7.346054103955278e-06, "loss": 2.2915, "step": 53030 }, { "epoch": 1.2, "grad_norm": 1.5910228636987658, "learning_rate": 7.34253336922874e-06, "loss": 2.5486, "step": 53040 }, { "epoch": 1.2, "grad_norm": 1.6097332901723065, "learning_rate": 7.339012988905706e-06, "loss": 2.3802, "step": 53050 }, { "epoch": 1.2, "grad_norm": 1.5832195538222544, "learning_rate": 7.3354929634556575e-06, "loss": 2.481, "step": 53060 }, { "epoch": 1.2, "grad_norm": 1.6507175233534983, "learning_rate": 7.331973293348027e-06, "loss": 2.4253, "step": 53070 }, { "epoch": 1.2, "grad_norm": 1.4999287333993236, "learning_rate": 7.328453979052204e-06, "loss": 2.3433, "step": 53080 }, { "epoch": 1.2, "grad_norm": 1.7295000566600969, "learning_rate": 7.324935021037531e-06, "loss": 2.4586, "step": 53090 }, { "epoch": 1.2, "grad_norm": 1.6121733254022128, "learning_rate": 7.3214164197733015e-06, "loss": 2.539, "step": 53100 }, { "epoch": 1.2, "grad_norm": 1.5522233934532201, "learning_rate": 7.317898175728757e-06, "loss": 2.4143, "step": 53110 }, { "epoch": 1.2, "grad_norm": 1.608594481179648, "learning_rate": 7.314380289373104e-06, "loss": 2.5235, "step": 53120 }, { "epoch": 1.2, "grad_norm": 1.5954271601514525, "learning_rate": 7.3108627611754846e-06, "loss": 2.4095, "step": 53130 }, { "epoch": 1.2, "grad_norm": 1.7262083523427294, "learning_rate": 7.307345591605e-06, "loss": 2.3918, "step": 53140 }, { "epoch": 1.2, "grad_norm": 1.4556107419688293, "learning_rate": 7.303828781130713e-06, "loss": 2.5056, "step": 53150 }, { "epoch": 1.2, "grad_norm": 1.71890310542134, "learning_rate": 7.3003123302216235e-06, "loss": 2.4809, "step": 53160 }, { "epoch": 1.2, "grad_norm": 2.088916630458351, "learning_rate": 7.296796239346693e-06, "loss": 2.4235, "step": 53170 }, { "epoch": 1.2, "grad_norm": 1.5728554742610164, "learning_rate": 7.293280508974833e-06, "loss": 2.4393, "step": 53180 }, { "epoch": 1.2, "grad_norm": 1.5758422443888362, "learning_rate": 7.289765139574907e-06, "loss": 2.4548, "step": 53190 }, { "epoch": 1.2, "grad_norm": 1.6966367888013771, "learning_rate": 7.2862501316157265e-06, "loss": 2.3963, "step": 53200 }, { "epoch": 1.2, "grad_norm": 1.5093315503017224, "learning_rate": 7.282735485566059e-06, "loss": 2.3856, "step": 53210 }, { "epoch": 1.2, "grad_norm": 1.5932908459120751, "learning_rate": 7.279221201894625e-06, "loss": 2.5582, "step": 53220 }, { "epoch": 1.2, "grad_norm": 1.554361916287439, "learning_rate": 7.275707281070095e-06, "loss": 2.4452, "step": 53230 }, { "epoch": 1.2, "grad_norm": 1.6719161617776948, "learning_rate": 7.2721937235610854e-06, "loss": 2.4759, "step": 53240 }, { "epoch": 1.2, "grad_norm": 1.5668114554036412, "learning_rate": 7.268680529836172e-06, "loss": 2.4112, "step": 53250 }, { "epoch": 1.2, "grad_norm": 1.9184559912812202, "learning_rate": 7.26516770036388e-06, "loss": 2.3434, "step": 53260 }, { "epoch": 1.2, "grad_norm": 1.7416272513647548, "learning_rate": 7.261655235612682e-06, "loss": 2.4582, "step": 53270 }, { "epoch": 1.2, "grad_norm": 1.7207021887878444, "learning_rate": 7.258143136051009e-06, "loss": 2.4794, "step": 53280 }, { "epoch": 1.2, "grad_norm": 1.6665634674510252, "learning_rate": 7.254631402147238e-06, "loss": 2.3783, "step": 53290 }, { "epoch": 1.2, "grad_norm": 1.7915511458011963, "learning_rate": 7.251120034369701e-06, "loss": 2.4474, "step": 53300 }, { "epoch": 1.2, "grad_norm": 1.5219268340211243, "learning_rate": 7.247609033186673e-06, "loss": 2.4594, "step": 53310 }, { "epoch": 1.2, "grad_norm": 1.6661110692468397, "learning_rate": 7.2440983990663934e-06, "loss": 2.3974, "step": 53320 }, { "epoch": 1.2, "grad_norm": 1.5918295762113417, "learning_rate": 7.240588132477044e-06, "loss": 2.3429, "step": 53330 }, { "epoch": 1.2, "grad_norm": 1.7019072396653316, "learning_rate": 7.237078233886755e-06, "loss": 2.3614, "step": 53340 }, { "epoch": 1.2, "grad_norm": 1.927649452637633, "learning_rate": 7.233568703763614e-06, "loss": 2.3434, "step": 53350 }, { "epoch": 1.2, "grad_norm": 1.5699214137926947, "learning_rate": 7.230059542575653e-06, "loss": 2.4672, "step": 53360 }, { "epoch": 1.2, "grad_norm": 1.780038716982047, "learning_rate": 7.2265507507908635e-06, "loss": 2.4322, "step": 53370 }, { "epoch": 1.2, "grad_norm": 1.7933435601547185, "learning_rate": 7.223042328877181e-06, "loss": 2.508, "step": 53380 }, { "epoch": 1.2, "grad_norm": 1.6422348910426252, "learning_rate": 7.219534277302496e-06, "loss": 2.4953, "step": 53390 }, { "epoch": 1.2, "grad_norm": 1.5410964695370362, "learning_rate": 7.216026596534644e-06, "loss": 2.3206, "step": 53400 }, { "epoch": 1.2, "grad_norm": 1.7585158421890765, "learning_rate": 7.212519287041413e-06, "loss": 2.3855, "step": 53410 }, { "epoch": 1.2, "grad_norm": 1.5131676290172171, "learning_rate": 7.209012349290548e-06, "loss": 2.4426, "step": 53420 }, { "epoch": 1.2, "grad_norm": 1.6084494550470698, "learning_rate": 7.205505783749737e-06, "loss": 2.3045, "step": 53430 }, { "epoch": 1.21, "grad_norm": 1.8246059315569108, "learning_rate": 7.201999590886621e-06, "loss": 2.4755, "step": 53440 }, { "epoch": 1.21, "grad_norm": 1.8625561984746652, "learning_rate": 7.198493771168788e-06, "loss": 2.403, "step": 53450 }, { "epoch": 1.21, "grad_norm": 1.9794489267707307, "learning_rate": 7.194988325063781e-06, "loss": 2.3661, "step": 53460 }, { "epoch": 1.21, "grad_norm": 1.7199594076728018, "learning_rate": 7.191483253039089e-06, "loss": 2.3742, "step": 53470 }, { "epoch": 1.21, "grad_norm": 1.7796270778505778, "learning_rate": 7.187978555562156e-06, "loss": 2.3482, "step": 53480 }, { "epoch": 1.21, "grad_norm": 1.5790200107581125, "learning_rate": 7.184474233100374e-06, "loss": 2.4301, "step": 53490 }, { "epoch": 1.21, "grad_norm": 1.8154173643787226, "learning_rate": 7.180970286121083e-06, "loss": 2.3692, "step": 53500 }, { "epoch": 1.21, "grad_norm": 1.632408352998536, "learning_rate": 7.177466715091572e-06, "loss": 2.4986, "step": 53510 }, { "epoch": 1.21, "grad_norm": 1.67464327236302, "learning_rate": 7.173963520479088e-06, "loss": 2.4033, "step": 53520 }, { "epoch": 1.21, "grad_norm": 1.624785428199028, "learning_rate": 7.170460702750817e-06, "loss": 2.4314, "step": 53530 }, { "epoch": 1.21, "grad_norm": 1.7428158552448438, "learning_rate": 7.166958262373902e-06, "loss": 2.4715, "step": 53540 }, { "epoch": 1.21, "grad_norm": 1.6873648524199074, "learning_rate": 7.163456199815435e-06, "loss": 2.449, "step": 53550 }, { "epoch": 1.21, "grad_norm": 1.6842126698108064, "learning_rate": 7.159954515542452e-06, "loss": 2.3366, "step": 53560 }, { "epoch": 1.21, "grad_norm": 1.5858270318317937, "learning_rate": 7.156453210021942e-06, "loss": 2.4171, "step": 53570 }, { "epoch": 1.21, "grad_norm": 1.6596279941125172, "learning_rate": 7.1529522837208485e-06, "loss": 2.4559, "step": 53580 }, { "epoch": 1.21, "grad_norm": 1.4708730778555517, "learning_rate": 7.149451737106057e-06, "loss": 2.4862, "step": 53590 }, { "epoch": 1.21, "grad_norm": 1.7145162847899278, "learning_rate": 7.1459515706444076e-06, "loss": 2.3819, "step": 53600 }, { "epoch": 1.21, "grad_norm": 1.6315476476079454, "learning_rate": 7.142451784802683e-06, "loss": 2.4396, "step": 53610 }, { "epoch": 1.21, "grad_norm": 1.5794089452035376, "learning_rate": 7.1389523800476255e-06, "loss": 2.4784, "step": 53620 }, { "epoch": 1.21, "grad_norm": 1.5858678579543268, "learning_rate": 7.135453356845916e-06, "loss": 2.5224, "step": 53630 }, { "epoch": 1.21, "grad_norm": 1.6023559150281235, "learning_rate": 7.131954715664191e-06, "loss": 2.4051, "step": 53640 }, { "epoch": 1.21, "grad_norm": 1.593493762914199, "learning_rate": 7.128456456969035e-06, "loss": 2.4328, "step": 53650 }, { "epoch": 1.21, "grad_norm": 1.6955544641433582, "learning_rate": 7.124958581226984e-06, "loss": 2.5531, "step": 53660 }, { "epoch": 1.21, "grad_norm": 2.0273100142450335, "learning_rate": 7.121461088904511e-06, "loss": 2.5244, "step": 53670 }, { "epoch": 1.21, "grad_norm": 1.5805082298945692, "learning_rate": 7.117963980468051e-06, "loss": 2.5052, "step": 53680 }, { "epoch": 1.21, "grad_norm": 1.4632101685811132, "learning_rate": 7.114467256383985e-06, "loss": 2.4816, "step": 53690 }, { "epoch": 1.21, "grad_norm": 1.90426610185343, "learning_rate": 7.110970917118636e-06, "loss": 2.3807, "step": 53700 }, { "epoch": 1.21, "grad_norm": 1.5477495697973265, "learning_rate": 7.107474963138287e-06, "loss": 2.4434, "step": 53710 }, { "epoch": 1.21, "grad_norm": 1.5983896354469385, "learning_rate": 7.10397939490916e-06, "loss": 2.4651, "step": 53720 }, { "epoch": 1.21, "grad_norm": 1.7806738804539584, "learning_rate": 7.10048421289743e-06, "loss": 2.473, "step": 53730 }, { "epoch": 1.21, "grad_norm": 1.6659255381228226, "learning_rate": 7.0969894175692155e-06, "loss": 2.3489, "step": 53740 }, { "epoch": 1.21, "grad_norm": 1.4681421916531174, "learning_rate": 7.093495009390594e-06, "loss": 2.3108, "step": 53750 }, { "epoch": 1.21, "grad_norm": 1.6148163335751309, "learning_rate": 7.090000988827581e-06, "loss": 2.5443, "step": 53760 }, { "epoch": 1.21, "grad_norm": 1.5700674030067199, "learning_rate": 7.0865073563461475e-06, "loss": 2.3422, "step": 53770 }, { "epoch": 1.21, "grad_norm": 1.6609213826742082, "learning_rate": 7.083014112412203e-06, "loss": 2.4463, "step": 53780 }, { "epoch": 1.21, "grad_norm": 1.5570741504623586, "learning_rate": 7.079521257491615e-06, "loss": 2.4365, "step": 53790 }, { "epoch": 1.21, "grad_norm": 1.9064533190766197, "learning_rate": 7.0760287920501926e-06, "loss": 2.3496, "step": 53800 }, { "epoch": 1.21, "grad_norm": 1.5944836735041559, "learning_rate": 7.072536716553701e-06, "loss": 2.4875, "step": 53810 }, { "epoch": 1.21, "grad_norm": 1.653228686642044, "learning_rate": 7.069045031467845e-06, "loss": 2.4207, "step": 53820 }, { "epoch": 1.21, "grad_norm": 1.851915634710487, "learning_rate": 7.065553737258283e-06, "loss": 2.3818, "step": 53830 }, { "epoch": 1.21, "grad_norm": 1.6223042617538748, "learning_rate": 7.062062834390615e-06, "loss": 2.5207, "step": 53840 }, { "epoch": 1.21, "grad_norm": 1.6978079289298318, "learning_rate": 7.058572323330396e-06, "loss": 2.3999, "step": 53850 }, { "epoch": 1.21, "grad_norm": 1.6673936015287052, "learning_rate": 7.055082204543126e-06, "loss": 2.3914, "step": 53860 }, { "epoch": 1.21, "grad_norm": 1.6114741903289522, "learning_rate": 7.051592478494251e-06, "loss": 2.3506, "step": 53870 }, { "epoch": 1.22, "grad_norm": 1.6993049236520954, "learning_rate": 7.048103145649165e-06, "loss": 2.3316, "step": 53880 }, { "epoch": 1.22, "grad_norm": 1.6414891343541838, "learning_rate": 7.0446142064732106e-06, "loss": 2.51, "step": 53890 }, { "epoch": 1.22, "grad_norm": 1.7799080325500747, "learning_rate": 7.041125661431676e-06, "loss": 2.5738, "step": 53900 }, { "epoch": 1.22, "grad_norm": 1.5235614339831085, "learning_rate": 7.037637510989803e-06, "loss": 2.416, "step": 53910 }, { "epoch": 1.22, "grad_norm": 1.7253804164031668, "learning_rate": 7.034149755612773e-06, "loss": 2.2818, "step": 53920 }, { "epoch": 1.22, "grad_norm": 1.6035933820457484, "learning_rate": 7.030662395765718e-06, "loss": 2.5451, "step": 53930 }, { "epoch": 1.22, "grad_norm": 1.6035137273969242, "learning_rate": 7.027175431913718e-06, "loss": 2.5737, "step": 53940 }, { "epoch": 1.22, "grad_norm": 1.6756393351275274, "learning_rate": 7.023688864521799e-06, "loss": 2.4611, "step": 53950 }, { "epoch": 1.22, "grad_norm": 1.592117619206558, "learning_rate": 7.020202694054936e-06, "loss": 2.3982, "step": 53960 }, { "epoch": 1.22, "grad_norm": 1.8818512124612248, "learning_rate": 7.016716920978049e-06, "loss": 2.4371, "step": 53970 }, { "epoch": 1.22, "grad_norm": 1.8295740737571258, "learning_rate": 7.013231545756007e-06, "loss": 2.3993, "step": 53980 }, { "epoch": 1.22, "grad_norm": 1.8468082681933733, "learning_rate": 7.0097465688536215e-06, "loss": 2.2758, "step": 53990 }, { "epoch": 1.22, "grad_norm": 1.4768191490584097, "learning_rate": 7.006261990735651e-06, "loss": 2.3195, "step": 54000 }, { "epoch": 1.22, "grad_norm": 1.8109033486805248, "learning_rate": 7.002777811866812e-06, "loss": 2.485, "step": 54010 }, { "epoch": 1.22, "grad_norm": 1.8093287165897225, "learning_rate": 6.999294032711755e-06, "loss": 2.2615, "step": 54020 }, { "epoch": 1.22, "grad_norm": 1.5829920742166386, "learning_rate": 6.99581065373508e-06, "loss": 2.3944, "step": 54030 }, { "epoch": 1.22, "grad_norm": 1.479242141539316, "learning_rate": 6.992327675401338e-06, "loss": 2.3735, "step": 54040 }, { "epoch": 1.22, "grad_norm": 1.8307229354138006, "learning_rate": 6.988845098175025e-06, "loss": 2.5083, "step": 54050 }, { "epoch": 1.22, "grad_norm": 1.5839131776468742, "learning_rate": 6.98536292252058e-06, "loss": 2.4908, "step": 54060 }, { "epoch": 1.22, "grad_norm": 1.7441015701421536, "learning_rate": 6.981881148902392e-06, "loss": 2.5922, "step": 54070 }, { "epoch": 1.22, "grad_norm": 1.5546871989450652, "learning_rate": 6.9783997777847924e-06, "loss": 2.4561, "step": 54080 }, { "epoch": 1.22, "grad_norm": 1.711416289821683, "learning_rate": 6.974918809632071e-06, "loss": 2.4569, "step": 54090 }, { "epoch": 1.22, "grad_norm": 1.6330709812596955, "learning_rate": 6.971438244908442e-06, "loss": 2.3856, "step": 54100 }, { "epoch": 1.22, "grad_norm": 1.685885456610255, "learning_rate": 6.9679580840780855e-06, "loss": 2.4531, "step": 54110 }, { "epoch": 1.22, "grad_norm": 2.2299042512972704, "learning_rate": 6.96447832760512e-06, "loss": 2.4174, "step": 54120 }, { "epoch": 1.22, "grad_norm": 1.7819673592805438, "learning_rate": 6.96099897595361e-06, "loss": 2.364, "step": 54130 }, { "epoch": 1.22, "grad_norm": 1.530216520928981, "learning_rate": 6.957520029587565e-06, "loss": 2.4301, "step": 54140 }, { "epoch": 1.22, "grad_norm": 1.5744284543634144, "learning_rate": 6.954041488970946e-06, "loss": 2.3826, "step": 54150 }, { "epoch": 1.22, "grad_norm": 1.5297041516675471, "learning_rate": 6.950563354567654e-06, "loss": 2.4445, "step": 54160 }, { "epoch": 1.22, "grad_norm": 1.7473907213328048, "learning_rate": 6.9470856268415375e-06, "loss": 2.3753, "step": 54170 }, { "epoch": 1.22, "grad_norm": 1.6079411565151724, "learning_rate": 6.943608306256393e-06, "loss": 2.4913, "step": 54180 }, { "epoch": 1.22, "grad_norm": 1.448809945250618, "learning_rate": 6.94013139327596e-06, "loss": 2.3725, "step": 54190 }, { "epoch": 1.22, "grad_norm": 1.787465188503513, "learning_rate": 6.936654888363926e-06, "loss": 2.4311, "step": 54200 }, { "epoch": 1.22, "grad_norm": 1.9653488963660632, "learning_rate": 6.933178791983919e-06, "loss": 2.4937, "step": 54210 }, { "epoch": 1.22, "grad_norm": 1.776091156822788, "learning_rate": 6.929703104599517e-06, "loss": 2.5102, "step": 54220 }, { "epoch": 1.22, "grad_norm": 1.621840992476866, "learning_rate": 6.926227826674243e-06, "loss": 2.4959, "step": 54230 }, { "epoch": 1.22, "grad_norm": 1.5947308717029673, "learning_rate": 6.922752958671568e-06, "loss": 2.431, "step": 54240 }, { "epoch": 1.22, "grad_norm": 1.5359353580956396, "learning_rate": 6.919278501054901e-06, "loss": 2.4292, "step": 54250 }, { "epoch": 1.22, "grad_norm": 1.4933717627091554, "learning_rate": 6.9158044542876035e-06, "loss": 2.4671, "step": 54260 }, { "epoch": 1.22, "grad_norm": 1.6879019732131038, "learning_rate": 6.912330818832975e-06, "loss": 2.453, "step": 54270 }, { "epoch": 1.22, "grad_norm": 1.8299539928081707, "learning_rate": 6.9088575951542714e-06, "loss": 2.3979, "step": 54280 }, { "epoch": 1.22, "grad_norm": 1.5296111743351617, "learning_rate": 6.905384783714682e-06, "loss": 2.3463, "step": 54290 }, { "epoch": 1.22, "grad_norm": 1.4954367472864911, "learning_rate": 6.901912384977347e-06, "loss": 2.4803, "step": 54300 }, { "epoch": 1.22, "grad_norm": 1.812527049729758, "learning_rate": 6.8984403994053514e-06, "loss": 2.5096, "step": 54310 }, { "epoch": 1.22, "grad_norm": 1.752637866961347, "learning_rate": 6.894968827461721e-06, "loss": 2.4842, "step": 54320 }, { "epoch": 1.23, "grad_norm": 1.7712841840157985, "learning_rate": 6.891497669609431e-06, "loss": 2.568, "step": 54330 }, { "epoch": 1.23, "grad_norm": 1.969207194905346, "learning_rate": 6.888026926311399e-06, "loss": 2.5113, "step": 54340 }, { "epoch": 1.23, "grad_norm": 1.7272818208074328, "learning_rate": 6.884556598030493e-06, "loss": 2.5759, "step": 54350 }, { "epoch": 1.23, "grad_norm": 1.652334159368157, "learning_rate": 6.881086685229516e-06, "loss": 2.4841, "step": 54360 }, { "epoch": 1.23, "grad_norm": 1.6724559885824932, "learning_rate": 6.87761718837122e-06, "loss": 2.4501, "step": 54370 }, { "epoch": 1.23, "grad_norm": 1.427710730272385, "learning_rate": 6.874148107918306e-06, "loss": 2.5061, "step": 54380 }, { "epoch": 1.23, "grad_norm": 1.4490639257768427, "learning_rate": 6.870679444333412e-06, "loss": 2.4805, "step": 54390 }, { "epoch": 1.23, "grad_norm": 1.5005102033410613, "learning_rate": 6.867211198079127e-06, "loss": 2.432, "step": 54400 }, { "epoch": 1.23, "grad_norm": 1.645825214366632, "learning_rate": 6.863743369617981e-06, "loss": 2.5692, "step": 54410 }, { "epoch": 1.23, "grad_norm": 1.6036643261602792, "learning_rate": 6.860275959412445e-06, "loss": 2.5278, "step": 54420 }, { "epoch": 1.23, "grad_norm": 1.6736360857600545, "learning_rate": 6.856808967924937e-06, "loss": 2.5045, "step": 54430 }, { "epoch": 1.23, "grad_norm": 1.5471597280859304, "learning_rate": 6.853342395617825e-06, "loss": 2.5125, "step": 54440 }, { "epoch": 1.23, "grad_norm": 1.678648765741852, "learning_rate": 6.849876242953413e-06, "loss": 2.3556, "step": 54450 }, { "epoch": 1.23, "grad_norm": 1.6525433011601474, "learning_rate": 6.846410510393953e-06, "loss": 2.4862, "step": 54460 }, { "epoch": 1.23, "grad_norm": 1.673431705860325, "learning_rate": 6.842945198401637e-06, "loss": 2.5652, "step": 54470 }, { "epoch": 1.23, "grad_norm": 1.7153406432267067, "learning_rate": 6.839480307438608e-06, "loss": 2.3387, "step": 54480 }, { "epoch": 1.23, "grad_norm": 1.5473592528225655, "learning_rate": 6.836015837966949e-06, "loss": 2.4891, "step": 54490 }, { "epoch": 1.23, "grad_norm": 1.798615411407398, "learning_rate": 6.832551790448684e-06, "loss": 2.4544, "step": 54500 }, { "epoch": 1.23, "grad_norm": 1.800180099810111, "learning_rate": 6.829088165345781e-06, "loss": 2.3499, "step": 54510 }, { "epoch": 1.23, "grad_norm": 1.6449217113812238, "learning_rate": 6.825624963120163e-06, "loss": 2.4055, "step": 54520 }, { "epoch": 1.23, "grad_norm": 1.564380534286689, "learning_rate": 6.822162184233676e-06, "loss": 2.5568, "step": 54530 }, { "epoch": 1.23, "grad_norm": 1.8473768530690315, "learning_rate": 6.81869982914813e-06, "loss": 2.4572, "step": 54540 }, { "epoch": 1.23, "grad_norm": 1.6482868463105824, "learning_rate": 6.8152378983252645e-06, "loss": 2.4758, "step": 54550 }, { "epoch": 1.23, "grad_norm": 1.580444238667102, "learning_rate": 6.811776392226767e-06, "loss": 2.4197, "step": 54560 }, { "epoch": 1.23, "grad_norm": 1.6060142213386568, "learning_rate": 6.808315311314273e-06, "loss": 2.4412, "step": 54570 }, { "epoch": 1.23, "grad_norm": 1.7420696322234932, "learning_rate": 6.804854656049355e-06, "loss": 2.3628, "step": 54580 }, { "epoch": 1.23, "grad_norm": 1.5910651986651756, "learning_rate": 6.8013944268935305e-06, "loss": 2.5671, "step": 54590 }, { "epoch": 1.23, "grad_norm": 1.6553665731812577, "learning_rate": 6.7979346243082615e-06, "loss": 2.5332, "step": 54600 }, { "epoch": 1.23, "grad_norm": 1.7178213151250215, "learning_rate": 6.794475248754949e-06, "loss": 2.5003, "step": 54610 }, { "epoch": 1.23, "grad_norm": 1.5968364671673503, "learning_rate": 6.791016300694946e-06, "loss": 2.4333, "step": 54620 }, { "epoch": 1.23, "grad_norm": 1.5628499081563438, "learning_rate": 6.78755778058954e-06, "loss": 2.4961, "step": 54630 }, { "epoch": 1.23, "grad_norm": 1.5681275375207986, "learning_rate": 6.784099688899963e-06, "loss": 2.5251, "step": 54640 }, { "epoch": 1.23, "grad_norm": 1.7163726517929423, "learning_rate": 6.7806420260873915e-06, "loss": 2.5156, "step": 54650 }, { "epoch": 1.23, "grad_norm": 1.775551886938836, "learning_rate": 6.777184792612944e-06, "loss": 2.5679, "step": 54660 }, { "epoch": 1.23, "grad_norm": 1.6923741539187882, "learning_rate": 6.77372798893768e-06, "loss": 2.4242, "step": 54670 }, { "epoch": 1.23, "grad_norm": 1.5337141902129234, "learning_rate": 6.770271615522609e-06, "loss": 2.5848, "step": 54680 }, { "epoch": 1.23, "grad_norm": 1.705108643755293, "learning_rate": 6.766815672828675e-06, "loss": 2.416, "step": 54690 }, { "epoch": 1.23, "grad_norm": 1.5611125708864055, "learning_rate": 6.763360161316765e-06, "loss": 2.4307, "step": 54700 }, { "epoch": 1.23, "grad_norm": 1.6115791175813432, "learning_rate": 6.759905081447715e-06, "loss": 2.3837, "step": 54710 }, { "epoch": 1.23, "grad_norm": 1.718337121984102, "learning_rate": 6.7564504336822974e-06, "loss": 2.3911, "step": 54720 }, { "epoch": 1.23, "grad_norm": 1.6965799499372023, "learning_rate": 6.752996218481232e-06, "loss": 2.4765, "step": 54730 }, { "epoch": 1.23, "grad_norm": 1.7664355854640224, "learning_rate": 6.749542436305175e-06, "loss": 2.4802, "step": 54740 }, { "epoch": 1.23, "grad_norm": 1.8459045354599823, "learning_rate": 6.746089087614724e-06, "loss": 2.4752, "step": 54750 }, { "epoch": 1.23, "grad_norm": 1.7953391409771227, "learning_rate": 6.742636172870426e-06, "loss": 2.4033, "step": 54760 }, { "epoch": 1.24, "grad_norm": 1.6194684907089594, "learning_rate": 6.739183692532768e-06, "loss": 2.5153, "step": 54770 }, { "epoch": 1.24, "grad_norm": 1.7321359341623583, "learning_rate": 6.7357316470621755e-06, "loss": 2.489, "step": 54780 }, { "epoch": 1.24, "grad_norm": 1.4650692688232723, "learning_rate": 6.732280036919021e-06, "loss": 2.5111, "step": 54790 }, { "epoch": 1.24, "grad_norm": 1.5554893842806548, "learning_rate": 6.728828862563611e-06, "loss": 2.4186, "step": 54800 }, { "epoch": 1.24, "grad_norm": 1.7464057794013295, "learning_rate": 6.725378124456203e-06, "loss": 2.4562, "step": 54810 }, { "epoch": 1.24, "grad_norm": 1.7165300358500502, "learning_rate": 6.721927823056993e-06, "loss": 2.4887, "step": 54820 }, { "epoch": 1.24, "grad_norm": 1.7215289638801579, "learning_rate": 6.718477958826116e-06, "loss": 2.3698, "step": 54830 }, { "epoch": 1.24, "grad_norm": 1.674036434952293, "learning_rate": 6.7150285322236496e-06, "loss": 2.368, "step": 54840 }, { "epoch": 1.24, "grad_norm": 1.7324201111969213, "learning_rate": 6.711579543709621e-06, "loss": 2.5798, "step": 54850 }, { "epoch": 1.24, "grad_norm": 1.6710581156473854, "learning_rate": 6.708130993743982e-06, "loss": 2.4941, "step": 54860 }, { "epoch": 1.24, "grad_norm": 1.828287312840808, "learning_rate": 6.704682882786642e-06, "loss": 2.4993, "step": 54870 }, { "epoch": 1.24, "grad_norm": 1.743673132244394, "learning_rate": 6.701235211297445e-06, "loss": 2.535, "step": 54880 }, { "epoch": 1.24, "grad_norm": 1.5328380754346338, "learning_rate": 6.6977879797361785e-06, "loss": 2.5324, "step": 54890 }, { "epoch": 1.24, "grad_norm": 1.6761538779749052, "learning_rate": 6.694341188562565e-06, "loss": 2.3963, "step": 54900 }, { "epoch": 1.24, "grad_norm": 1.9720528228485403, "learning_rate": 6.690894838236281e-06, "loss": 2.4305, "step": 54910 }, { "epoch": 1.24, "grad_norm": 1.9679382859324128, "learning_rate": 6.687448929216933e-06, "loss": 2.4138, "step": 54920 }, { "epoch": 1.24, "grad_norm": 1.7863697907778455, "learning_rate": 6.684003461964072e-06, "loss": 2.3841, "step": 54930 }, { "epoch": 1.24, "grad_norm": 1.644267792284961, "learning_rate": 6.68055843693719e-06, "loss": 2.4641, "step": 54940 }, { "epoch": 1.24, "grad_norm": 1.6566137889219261, "learning_rate": 6.6771138545957255e-06, "loss": 2.3869, "step": 54950 }, { "epoch": 1.24, "grad_norm": 1.534540646449724, "learning_rate": 6.673669715399043e-06, "loss": 2.5296, "step": 54960 }, { "epoch": 1.24, "grad_norm": 1.8733752999310909, "learning_rate": 6.670226019806465e-06, "loss": 2.4533, "step": 54970 }, { "epoch": 1.24, "grad_norm": 1.5882711732182562, "learning_rate": 6.666782768277247e-06, "loss": 2.3924, "step": 54980 }, { "epoch": 1.24, "grad_norm": 2.0604653163903297, "learning_rate": 6.663339961270584e-06, "loss": 2.5703, "step": 54990 }, { "epoch": 1.24, "grad_norm": 1.5982354347185146, "learning_rate": 6.659897599245613e-06, "loss": 2.3034, "step": 55000 }, { "epoch": 1.24, "grad_norm": 1.4308863578180746, "learning_rate": 6.6564556826614155e-06, "loss": 2.2996, "step": 55010 }, { "epoch": 1.24, "grad_norm": 1.6211436696249162, "learning_rate": 6.653014211977007e-06, "loss": 2.4142, "step": 55020 }, { "epoch": 1.24, "grad_norm": 1.8167092406758525, "learning_rate": 6.649573187651351e-06, "loss": 2.3966, "step": 55030 }, { "epoch": 1.24, "grad_norm": 1.533747300959738, "learning_rate": 6.646132610143341e-06, "loss": 2.3946, "step": 55040 }, { "epoch": 1.24, "grad_norm": 1.4650960022042987, "learning_rate": 6.642692479911823e-06, "loss": 2.5063, "step": 55050 }, { "epoch": 1.24, "grad_norm": 1.6613092165500276, "learning_rate": 6.639252797415579e-06, "loss": 2.2791, "step": 55060 }, { "epoch": 1.24, "grad_norm": 1.4801482080244368, "learning_rate": 6.635813563113322e-06, "loss": 2.4778, "step": 55070 }, { "epoch": 1.24, "grad_norm": 1.5652174052180878, "learning_rate": 6.6323747774637206e-06, "loss": 2.4119, "step": 55080 }, { "epoch": 1.24, "grad_norm": 1.6620351793511134, "learning_rate": 6.628936440925371e-06, "loss": 2.6044, "step": 55090 }, { "epoch": 1.24, "grad_norm": 1.673407284659652, "learning_rate": 6.6254985539568165e-06, "loss": 2.5869, "step": 55100 }, { "epoch": 1.24, "grad_norm": 1.60219878653305, "learning_rate": 6.622061117016541e-06, "loss": 2.3837, "step": 55110 }, { "epoch": 1.24, "grad_norm": 1.6531670088861203, "learning_rate": 6.618624130562963e-06, "loss": 2.5399, "step": 55120 }, { "epoch": 1.24, "grad_norm": 1.7058956315746105, "learning_rate": 6.6151875950544445e-06, "loss": 2.4785, "step": 55130 }, { "epoch": 1.24, "grad_norm": 1.7043092227608652, "learning_rate": 6.611751510949285e-06, "loss": 2.4842, "step": 55140 }, { "epoch": 1.24, "grad_norm": 1.6306480875165688, "learning_rate": 6.6083158787057285e-06, "loss": 2.4696, "step": 55150 }, { "epoch": 1.24, "grad_norm": 1.677792053890769, "learning_rate": 6.604880698781957e-06, "loss": 2.4176, "step": 55160 }, { "epoch": 1.24, "grad_norm": 1.7094395742765447, "learning_rate": 6.601445971636089e-06, "loss": 2.4464, "step": 55170 }, { "epoch": 1.24, "grad_norm": 1.7529597552411045, "learning_rate": 6.598011697726183e-06, "loss": 2.4623, "step": 55180 }, { "epoch": 1.24, "grad_norm": 1.6446055265959831, "learning_rate": 6.5945778775102396e-06, "loss": 2.4141, "step": 55190 }, { "epoch": 1.24, "grad_norm": 1.7741180511626204, "learning_rate": 6.591144511446197e-06, "loss": 2.4407, "step": 55200 }, { "epoch": 1.25, "grad_norm": 1.5857641056696066, "learning_rate": 6.587711599991935e-06, "loss": 2.4434, "step": 55210 }, { "epoch": 1.25, "grad_norm": 1.5826859880233173, "learning_rate": 6.584279143605272e-06, "loss": 2.4241, "step": 55220 }, { "epoch": 1.25, "grad_norm": 1.90607571124872, "learning_rate": 6.580847142743965e-06, "loss": 2.3152, "step": 55230 }, { "epoch": 1.25, "grad_norm": 1.6226709565047688, "learning_rate": 6.57741559786571e-06, "loss": 2.4405, "step": 55240 }, { "epoch": 1.25, "grad_norm": 1.5245649464562552, "learning_rate": 6.573984509428142e-06, "loss": 2.4711, "step": 55250 }, { "epoch": 1.25, "grad_norm": 1.660633071799346, "learning_rate": 6.570553877888839e-06, "loss": 2.4009, "step": 55260 }, { "epoch": 1.25, "grad_norm": 1.5955960270396232, "learning_rate": 6.5671237037053086e-06, "loss": 2.5074, "step": 55270 }, { "epoch": 1.25, "grad_norm": 1.5341958233849775, "learning_rate": 6.563693987335013e-06, "loss": 2.5207, "step": 55280 }, { "epoch": 1.25, "grad_norm": 1.519761824943628, "learning_rate": 6.560264729235335e-06, "loss": 2.4143, "step": 55290 }, { "epoch": 1.25, "grad_norm": 1.7463436996603885, "learning_rate": 6.55683592986361e-06, "loss": 2.491, "step": 55300 }, { "epoch": 1.25, "grad_norm": 1.5149061882339385, "learning_rate": 6.5534075896771056e-06, "loss": 2.4233, "step": 55310 }, { "epoch": 1.25, "grad_norm": 1.6698404204364217, "learning_rate": 6.549979709133031e-06, "loss": 2.4833, "step": 55320 }, { "epoch": 1.25, "grad_norm": 1.6896504485752306, "learning_rate": 6.54655228868853e-06, "loss": 2.4436, "step": 55330 }, { "epoch": 1.25, "grad_norm": 1.6398363537812568, "learning_rate": 6.543125328800695e-06, "loss": 2.5346, "step": 55340 }, { "epoch": 1.25, "grad_norm": 1.5536736788796668, "learning_rate": 6.539698829926547e-06, "loss": 2.4151, "step": 55350 }, { "epoch": 1.25, "grad_norm": 1.831213019321923, "learning_rate": 6.536272792523046e-06, "loss": 2.4754, "step": 55360 }, { "epoch": 1.25, "grad_norm": 1.6188613029287795, "learning_rate": 6.532847217047096e-06, "loss": 2.4426, "step": 55370 }, { "epoch": 1.25, "grad_norm": 1.7821286935032965, "learning_rate": 6.52942210395554e-06, "loss": 2.4861, "step": 55380 }, { "epoch": 1.25, "grad_norm": 1.74441696778847, "learning_rate": 6.525997453705146e-06, "loss": 2.3984, "step": 55390 }, { "epoch": 1.25, "grad_norm": 1.7100364822544987, "learning_rate": 6.522573266752639e-06, "loss": 2.321, "step": 55400 }, { "epoch": 1.25, "grad_norm": 1.6576815163434142, "learning_rate": 6.51914954355467e-06, "loss": 2.4207, "step": 55410 }, { "epoch": 1.25, "grad_norm": 1.5506493023862884, "learning_rate": 6.5157262845678325e-06, "loss": 2.5998, "step": 55420 }, { "epoch": 1.25, "grad_norm": 1.7435506653031134, "learning_rate": 6.512303490248655e-06, "loss": 2.4411, "step": 55430 }, { "epoch": 1.25, "grad_norm": 1.655550301489535, "learning_rate": 6.508881161053608e-06, "loss": 2.3914, "step": 55440 }, { "epoch": 1.25, "grad_norm": 1.7345764737577296, "learning_rate": 6.505459297439101e-06, "loss": 2.4247, "step": 55450 }, { "epoch": 1.25, "grad_norm": 1.695847676232446, "learning_rate": 6.502037899861473e-06, "loss": 2.4684, "step": 55460 }, { "epoch": 1.25, "grad_norm": 1.5712744034836579, "learning_rate": 6.498616968777008e-06, "loss": 2.4147, "step": 55470 }, { "epoch": 1.25, "grad_norm": 1.5287234067068596, "learning_rate": 6.49519650464193e-06, "loss": 2.4442, "step": 55480 }, { "epoch": 1.25, "grad_norm": 1.5522332859332304, "learning_rate": 6.491776507912395e-06, "loss": 2.3769, "step": 55490 }, { "epoch": 1.25, "grad_norm": 1.7305880722333278, "learning_rate": 6.488356979044497e-06, "loss": 2.4176, "step": 55500 }, { "epoch": 1.25, "grad_norm": 1.6029760691192494, "learning_rate": 6.484937918494268e-06, "loss": 2.3448, "step": 55510 }, { "epoch": 1.25, "grad_norm": 1.5528632032339, "learning_rate": 6.481519326717682e-06, "loss": 2.3736, "step": 55520 }, { "epoch": 1.25, "grad_norm": 1.5577515302990157, "learning_rate": 6.478101204170644e-06, "loss": 2.5233, "step": 55530 }, { "epoch": 1.25, "grad_norm": 1.6619582588675441, "learning_rate": 6.474683551309004e-06, "loss": 2.4851, "step": 55540 }, { "epoch": 1.25, "grad_norm": 1.8476821195379765, "learning_rate": 6.4712663685885415e-06, "loss": 2.4241, "step": 55550 }, { "epoch": 1.25, "grad_norm": 1.7353288041576966, "learning_rate": 6.467849656464977e-06, "loss": 2.2705, "step": 55560 }, { "epoch": 1.25, "grad_norm": 1.8803648060623743, "learning_rate": 6.4644334153939695e-06, "loss": 2.3734, "step": 55570 }, { "epoch": 1.25, "grad_norm": 1.6399084005197586, "learning_rate": 6.461017645831114e-06, "loss": 2.3435, "step": 55580 }, { "epoch": 1.25, "grad_norm": 1.489654330415134, "learning_rate": 6.457602348231941e-06, "loss": 2.4247, "step": 55590 }, { "epoch": 1.25, "grad_norm": 1.5404836971165048, "learning_rate": 6.454187523051923e-06, "loss": 2.4481, "step": 55600 }, { "epoch": 1.25, "grad_norm": 1.815128893038291, "learning_rate": 6.45077317074646e-06, "loss": 2.4603, "step": 55610 }, { "epoch": 1.25, "grad_norm": 1.5803929393304592, "learning_rate": 6.4473592917708975e-06, "loss": 2.4691, "step": 55620 }, { "epoch": 1.25, "grad_norm": 1.5928648029656196, "learning_rate": 6.443945886580514e-06, "loss": 2.4297, "step": 55630 }, { "epoch": 1.25, "grad_norm": 1.5801868989954697, "learning_rate": 6.440532955630528e-06, "loss": 2.4553, "step": 55640 }, { "epoch": 1.25, "grad_norm": 1.6698414228631577, "learning_rate": 6.4371204993760925e-06, "loss": 2.3778, "step": 55650 }, { "epoch": 1.26, "grad_norm": 1.7057694652345785, "learning_rate": 6.433708518272299e-06, "loss": 2.4014, "step": 55660 }, { "epoch": 1.26, "grad_norm": 1.5541506733847552, "learning_rate": 6.430297012774169e-06, "loss": 2.4944, "step": 55670 }, { "epoch": 1.26, "grad_norm": 1.5862707266639793, "learning_rate": 6.426885983336672e-06, "loss": 2.4596, "step": 55680 }, { "epoch": 1.26, "grad_norm": 1.7093335458621501, "learning_rate": 6.423475430414704e-06, "loss": 2.4918, "step": 55690 }, { "epoch": 1.26, "grad_norm": 1.685057589101389, "learning_rate": 6.4200653544631e-06, "loss": 2.3382, "step": 55700 }, { "epoch": 1.26, "grad_norm": 1.9416662632107418, "learning_rate": 6.41665575593664e-06, "loss": 2.3285, "step": 55710 }, { "epoch": 1.26, "grad_norm": 2.012841707154507, "learning_rate": 6.413246635290024e-06, "loss": 2.4771, "step": 55720 }, { "epoch": 1.26, "grad_norm": 2.018576995040225, "learning_rate": 6.409837992977898e-06, "loss": 2.436, "step": 55730 }, { "epoch": 1.26, "grad_norm": 1.4869663971016298, "learning_rate": 6.406429829454848e-06, "loss": 2.4489, "step": 55740 }, { "epoch": 1.26, "grad_norm": 1.8965126494233921, "learning_rate": 6.403022145175388e-06, "loss": 2.4307, "step": 55750 }, { "epoch": 1.26, "grad_norm": 1.6841448640442018, "learning_rate": 6.399614940593973e-06, "loss": 2.342, "step": 55760 }, { "epoch": 1.26, "grad_norm": 1.9076545207055517, "learning_rate": 6.396208216164993e-06, "loss": 2.3764, "step": 55770 }, { "epoch": 1.26, "grad_norm": 1.788135709161936, "learning_rate": 6.392801972342774e-06, "loss": 2.4094, "step": 55780 }, { "epoch": 1.26, "grad_norm": 1.6796787944812333, "learning_rate": 6.389396209581576e-06, "loss": 2.3374, "step": 55790 }, { "epoch": 1.26, "grad_norm": 1.5599720233763092, "learning_rate": 6.385990928335595e-06, "loss": 2.3472, "step": 55800 }, { "epoch": 1.26, "grad_norm": 1.9222165805943432, "learning_rate": 6.3825861290589675e-06, "loss": 2.45, "step": 55810 }, { "epoch": 1.26, "grad_norm": 1.7821244032844714, "learning_rate": 6.379181812205763e-06, "loss": 2.4804, "step": 55820 }, { "epoch": 1.26, "grad_norm": 1.7696460638632623, "learning_rate": 6.3757779782299825e-06, "loss": 2.385, "step": 55830 }, { "epoch": 1.26, "grad_norm": 1.5351405268915939, "learning_rate": 6.372374627585567e-06, "loss": 2.4336, "step": 55840 }, { "epoch": 1.26, "grad_norm": 1.529207259212932, "learning_rate": 6.368971760726392e-06, "loss": 2.3481, "step": 55850 }, { "epoch": 1.26, "grad_norm": 1.6325395100059434, "learning_rate": 6.365569378106268e-06, "loss": 2.4967, "step": 55860 }, { "epoch": 1.26, "grad_norm": 1.5518744350887055, "learning_rate": 6.362167480178945e-06, "loss": 2.3946, "step": 55870 }, { "epoch": 1.26, "grad_norm": 1.696382434388868, "learning_rate": 6.358766067398101e-06, "loss": 2.3665, "step": 55880 }, { "epoch": 1.26, "grad_norm": 1.67786558053547, "learning_rate": 6.355365140217356e-06, "loss": 2.4915, "step": 55890 }, { "epoch": 1.26, "grad_norm": 1.4995065514355126, "learning_rate": 6.351964699090259e-06, "loss": 2.4167, "step": 55900 }, { "epoch": 1.26, "grad_norm": 1.8501935302524353, "learning_rate": 6.348564744470303e-06, "loss": 2.3125, "step": 55910 }, { "epoch": 1.26, "grad_norm": 1.9074357613209127, "learning_rate": 6.345165276810909e-06, "loss": 2.3609, "step": 55920 }, { "epoch": 1.26, "grad_norm": 1.7904487103026574, "learning_rate": 6.3417662965654315e-06, "loss": 2.5033, "step": 55930 }, { "epoch": 1.26, "grad_norm": 1.5101801983735699, "learning_rate": 6.3383678041871645e-06, "loss": 2.3624, "step": 55940 }, { "epoch": 1.26, "grad_norm": 1.6906504678348608, "learning_rate": 6.334969800129338e-06, "loss": 2.4811, "step": 55950 }, { "epoch": 1.26, "grad_norm": 1.6198529204478531, "learning_rate": 6.331572284845111e-06, "loss": 2.4453, "step": 55960 }, { "epoch": 1.26, "grad_norm": 1.7934921637716765, "learning_rate": 6.328175258787584e-06, "loss": 2.4049, "step": 55970 }, { "epoch": 1.26, "grad_norm": 1.6269355365417932, "learning_rate": 6.324778722409789e-06, "loss": 2.521, "step": 55980 }, { "epoch": 1.26, "grad_norm": 1.5580278195537147, "learning_rate": 6.3213826761646916e-06, "loss": 2.514, "step": 55990 }, { "epoch": 1.26, "grad_norm": 1.652533402390311, "learning_rate": 6.3179871205051935e-06, "loss": 2.3803, "step": 56000 }, { "epoch": 1.26, "grad_norm": 1.7199454987127902, "learning_rate": 6.3145920558841315e-06, "loss": 2.3235, "step": 56010 }, { "epoch": 1.26, "grad_norm": 1.5819060204908133, "learning_rate": 6.311197482754277e-06, "loss": 2.5446, "step": 56020 }, { "epoch": 1.26, "grad_norm": 1.664608813804814, "learning_rate": 6.307803401568335e-06, "loss": 2.4399, "step": 56030 }, { "epoch": 1.26, "grad_norm": 1.530098134752198, "learning_rate": 6.304409812778942e-06, "loss": 2.4365, "step": 56040 }, { "epoch": 1.26, "grad_norm": 1.5189168579276808, "learning_rate": 6.301016716838675e-06, "loss": 2.488, "step": 56050 }, { "epoch": 1.26, "grad_norm": 1.5999986097191348, "learning_rate": 6.297624114200038e-06, "loss": 2.5919, "step": 56060 }, { "epoch": 1.26, "grad_norm": 1.7326391391796567, "learning_rate": 6.294232005315477e-06, "loss": 2.2811, "step": 56070 }, { "epoch": 1.26, "grad_norm": 1.5297402950434797, "learning_rate": 6.29084039063737e-06, "loss": 2.4242, "step": 56080 }, { "epoch": 1.26, "grad_norm": 1.6081544468585192, "learning_rate": 6.287449270618023e-06, "loss": 2.4651, "step": 56090 }, { "epoch": 1.27, "grad_norm": 1.6246501219469611, "learning_rate": 6.284058645709681e-06, "loss": 2.3526, "step": 56100 }, { "epoch": 1.27, "grad_norm": 1.6931264353799897, "learning_rate": 6.280668516364525e-06, "loss": 2.366, "step": 56110 }, { "epoch": 1.27, "grad_norm": 1.6055825019139438, "learning_rate": 6.277278883034669e-06, "loss": 2.3945, "step": 56120 }, { "epoch": 1.27, "grad_norm": 2.0234775663918416, "learning_rate": 6.273889746172155e-06, "loss": 2.5394, "step": 56130 }, { "epoch": 1.27, "grad_norm": 1.6621931472255238, "learning_rate": 6.270501106228968e-06, "loss": 2.5156, "step": 56140 }, { "epoch": 1.27, "grad_norm": 1.8289189858611092, "learning_rate": 6.267112963657016e-06, "loss": 2.476, "step": 56150 }, { "epoch": 1.27, "grad_norm": 1.740269400488104, "learning_rate": 6.263725318908147e-06, "loss": 2.4485, "step": 56160 }, { "epoch": 1.27, "grad_norm": 1.5524788782889494, "learning_rate": 6.260338172434146e-06, "loss": 2.3801, "step": 56170 }, { "epoch": 1.27, "grad_norm": 1.5645890933196729, "learning_rate": 6.256951524686727e-06, "loss": 2.566, "step": 56180 }, { "epoch": 1.27, "grad_norm": 1.7054073833842762, "learning_rate": 6.253565376117536e-06, "loss": 2.4932, "step": 56190 }, { "epoch": 1.27, "grad_norm": 1.6584431326657614, "learning_rate": 6.250179727178155e-06, "loss": 2.3632, "step": 56200 }, { "epoch": 1.27, "grad_norm": 1.519519891385281, "learning_rate": 6.246794578320101e-06, "loss": 2.4963, "step": 56210 }, { "epoch": 1.27, "grad_norm": 1.8967453421970126, "learning_rate": 6.243409929994819e-06, "loss": 2.4642, "step": 56220 }, { "epoch": 1.27, "grad_norm": 1.5865793100153818, "learning_rate": 6.240025782653693e-06, "loss": 2.4312, "step": 56230 }, { "epoch": 1.27, "grad_norm": 1.6595597636619424, "learning_rate": 6.236642136748037e-06, "loss": 2.4129, "step": 56240 }, { "epoch": 1.27, "grad_norm": 1.6097361750741448, "learning_rate": 6.2332589927291035e-06, "loss": 2.4902, "step": 56250 }, { "epoch": 1.27, "grad_norm": 1.4996518808038835, "learning_rate": 6.229876351048062e-06, "loss": 2.5473, "step": 56260 }, { "epoch": 1.27, "grad_norm": 1.465180740891659, "learning_rate": 6.2264942121560355e-06, "loss": 2.448, "step": 56270 }, { "epoch": 1.27, "grad_norm": 2.766880919382219, "learning_rate": 6.223112576504066e-06, "loss": 2.4043, "step": 56280 }, { "epoch": 1.27, "grad_norm": 1.7470850878949507, "learning_rate": 6.219731444543137e-06, "loss": 2.4375, "step": 56290 }, { "epoch": 1.27, "grad_norm": 1.5624352674693676, "learning_rate": 6.21635081672416e-06, "loss": 2.4284, "step": 56300 }, { "epoch": 1.27, "grad_norm": 1.7850454611547326, "learning_rate": 6.212970693497979e-06, "loss": 2.4285, "step": 56310 }, { "epoch": 1.27, "grad_norm": 1.6672020960349059, "learning_rate": 6.209591075315373e-06, "loss": 2.361, "step": 56320 }, { "epoch": 1.27, "grad_norm": 1.852760300900069, "learning_rate": 6.206211962627051e-06, "loss": 2.4703, "step": 56330 }, { "epoch": 1.27, "grad_norm": 1.715585431467152, "learning_rate": 6.20283335588366e-06, "loss": 2.4488, "step": 56340 }, { "epoch": 1.27, "grad_norm": 1.6815299440168778, "learning_rate": 6.199455255535771e-06, "loss": 2.519, "step": 56350 }, { "epoch": 1.27, "grad_norm": 1.586847246089144, "learning_rate": 6.1960776620338995e-06, "loss": 2.5404, "step": 56360 }, { "epoch": 1.27, "grad_norm": 1.526733001799086, "learning_rate": 6.192700575828477e-06, "loss": 2.5764, "step": 56370 }, { "epoch": 1.27, "grad_norm": 1.7551191970133788, "learning_rate": 6.189323997369882e-06, "loss": 2.4275, "step": 56380 }, { "epoch": 1.27, "grad_norm": 1.513340651209612, "learning_rate": 6.185947927108418e-06, "loss": 2.446, "step": 56390 }, { "epoch": 1.27, "grad_norm": 1.4375743092976676, "learning_rate": 6.182572365494323e-06, "loss": 2.4583, "step": 56400 }, { "epoch": 1.27, "grad_norm": 1.571556553106838, "learning_rate": 6.179197312977767e-06, "loss": 2.4072, "step": 56410 }, { "epoch": 1.27, "grad_norm": 1.720237497169111, "learning_rate": 6.175822770008853e-06, "loss": 2.4035, "step": 56420 }, { "epoch": 1.27, "grad_norm": 1.6246816058242488, "learning_rate": 6.172448737037609e-06, "loss": 2.4941, "step": 56430 }, { "epoch": 1.27, "grad_norm": 1.4726089142868035, "learning_rate": 6.1690752145140076e-06, "loss": 2.4985, "step": 56440 }, { "epoch": 1.27, "grad_norm": 2.1223312136317123, "learning_rate": 6.165702202887944e-06, "loss": 2.4207, "step": 56450 }, { "epoch": 1.27, "grad_norm": 1.7018691193558106, "learning_rate": 6.162329702609251e-06, "loss": 2.4255, "step": 56460 }, { "epoch": 1.27, "grad_norm": 1.6740723241592619, "learning_rate": 6.158957714127685e-06, "loss": 2.4006, "step": 56470 }, { "epoch": 1.27, "grad_norm": 1.7719017962742318, "learning_rate": 6.155586237892939e-06, "loss": 2.3471, "step": 56480 }, { "epoch": 1.27, "grad_norm": 2.0280009824659526, "learning_rate": 6.1522152743546394e-06, "loss": 2.4933, "step": 56490 }, { "epoch": 1.27, "grad_norm": 1.8580660932406463, "learning_rate": 6.148844823962345e-06, "loss": 2.451, "step": 56500 }, { "epoch": 1.27, "grad_norm": 1.5892676936084023, "learning_rate": 6.1454748871655415e-06, "loss": 2.2934, "step": 56510 }, { "epoch": 1.27, "grad_norm": 1.5863497945192682, "learning_rate": 6.142105464413648e-06, "loss": 2.5498, "step": 56520 }, { "epoch": 1.27, "grad_norm": 1.5664841259768723, "learning_rate": 6.138736556156015e-06, "loss": 2.4297, "step": 56530 }, { "epoch": 1.28, "grad_norm": 1.5693377118865453, "learning_rate": 6.135368162841928e-06, "loss": 2.3704, "step": 56540 }, { "epoch": 1.28, "grad_norm": 1.6743531447230078, "learning_rate": 6.132000284920598e-06, "loss": 2.3122, "step": 56550 }, { "epoch": 1.28, "grad_norm": 1.5240114615082523, "learning_rate": 6.128632922841172e-06, "loss": 2.4457, "step": 56560 }, { "epoch": 1.28, "grad_norm": 1.6439793457862721, "learning_rate": 6.125266077052727e-06, "loss": 2.5136, "step": 56570 }, { "epoch": 1.28, "grad_norm": 1.594232199271589, "learning_rate": 6.121899748004265e-06, "loss": 2.4033, "step": 56580 }, { "epoch": 1.28, "grad_norm": 1.7906073415067099, "learning_rate": 6.118533936144727e-06, "loss": 2.4531, "step": 56590 }, { "epoch": 1.28, "grad_norm": 1.6005959231071578, "learning_rate": 6.115168641922984e-06, "loss": 2.5292, "step": 56600 }, { "epoch": 1.28, "grad_norm": 1.5924328145381716, "learning_rate": 6.111803865787837e-06, "loss": 2.511, "step": 56610 }, { "epoch": 1.28, "grad_norm": 1.6896606338764046, "learning_rate": 6.108439608188014e-06, "loss": 2.5084, "step": 56620 }, { "epoch": 1.28, "grad_norm": 1.625848343603316, "learning_rate": 6.1050758695721765e-06, "loss": 2.5759, "step": 56630 }, { "epoch": 1.28, "grad_norm": 1.6123973258465596, "learning_rate": 6.101712650388923e-06, "loss": 2.468, "step": 56640 }, { "epoch": 1.28, "grad_norm": 1.9020095641172567, "learning_rate": 6.098349951086775e-06, "loss": 2.4492, "step": 56650 }, { "epoch": 1.28, "grad_norm": 1.7712901342835516, "learning_rate": 6.0949877721141845e-06, "loss": 2.5227, "step": 56660 }, { "epoch": 1.28, "grad_norm": 1.5084193709791327, "learning_rate": 6.091626113919535e-06, "loss": 2.4251, "step": 56670 }, { "epoch": 1.28, "grad_norm": 1.6107444667783166, "learning_rate": 6.088264976951153e-06, "loss": 2.4938, "step": 56680 }, { "epoch": 1.28, "grad_norm": 1.5972953375824073, "learning_rate": 6.084904361657268e-06, "loss": 2.5509, "step": 56690 }, { "epoch": 1.28, "grad_norm": 1.6297697441533174, "learning_rate": 6.081544268486067e-06, "loss": 2.4118, "step": 56700 }, { "epoch": 1.28, "grad_norm": 1.7086956723171964, "learning_rate": 6.078184697885654e-06, "loss": 2.3405, "step": 56710 }, { "epoch": 1.28, "grad_norm": 1.7532521979868159, "learning_rate": 6.074825650304068e-06, "loss": 2.4991, "step": 56720 }, { "epoch": 1.28, "grad_norm": 1.678171363813406, "learning_rate": 6.071467126189272e-06, "loss": 2.3646, "step": 56730 }, { "epoch": 1.28, "grad_norm": 1.932864640272456, "learning_rate": 6.068109125989168e-06, "loss": 2.36, "step": 56740 }, { "epoch": 1.28, "grad_norm": 1.6564642214992809, "learning_rate": 6.064751650151581e-06, "loss": 2.4617, "step": 56750 }, { "epoch": 1.28, "grad_norm": 1.5111707333434867, "learning_rate": 6.061394699124269e-06, "loss": 2.3631, "step": 56760 }, { "epoch": 1.28, "grad_norm": 1.8601629391174066, "learning_rate": 6.0580382733549226e-06, "loss": 2.4924, "step": 56770 }, { "epoch": 1.28, "grad_norm": 1.6352695881422699, "learning_rate": 6.054682373291157e-06, "loss": 2.4144, "step": 56780 }, { "epoch": 1.28, "grad_norm": 1.9223320745764507, "learning_rate": 6.051326999380521e-06, "loss": 2.3106, "step": 56790 }, { "epoch": 1.28, "grad_norm": 1.7459957673106103, "learning_rate": 6.047972152070489e-06, "loss": 2.403, "step": 56800 }, { "epoch": 1.28, "grad_norm": 1.6017259661937124, "learning_rate": 6.0446178318084715e-06, "loss": 2.3705, "step": 56810 }, { "epoch": 1.28, "grad_norm": 1.6833857226201032, "learning_rate": 6.041264039041801e-06, "loss": 2.5236, "step": 56820 }, { "epoch": 1.28, "grad_norm": 1.5782587347553507, "learning_rate": 6.03791077421775e-06, "loss": 2.4213, "step": 56830 }, { "epoch": 1.28, "grad_norm": 1.806265504069015, "learning_rate": 6.034558037783511e-06, "loss": 2.4261, "step": 56840 }, { "epoch": 1.28, "grad_norm": 1.7845819178692262, "learning_rate": 6.031205830186211e-06, "loss": 2.4337, "step": 56850 }, { "epoch": 1.28, "grad_norm": 1.6577754494440766, "learning_rate": 6.027854151872903e-06, "loss": 2.316, "step": 56860 }, { "epoch": 1.28, "grad_norm": 1.8487830245952708, "learning_rate": 6.0245030032905735e-06, "loss": 2.3628, "step": 56870 }, { "epoch": 1.28, "grad_norm": 1.7756642615352918, "learning_rate": 6.021152384886137e-06, "loss": 2.4669, "step": 56880 }, { "epoch": 1.28, "grad_norm": 1.7256702545886264, "learning_rate": 6.017802297106437e-06, "loss": 2.3939, "step": 56890 }, { "epoch": 1.28, "grad_norm": 1.6447679166263858, "learning_rate": 6.014452740398244e-06, "loss": 2.5027, "step": 56900 }, { "epoch": 1.28, "grad_norm": 1.6712105850505539, "learning_rate": 6.01110371520826e-06, "loss": 2.4417, "step": 56910 }, { "epoch": 1.28, "grad_norm": 1.7602328601600368, "learning_rate": 6.0077552219831134e-06, "loss": 2.4453, "step": 56920 }, { "epoch": 1.28, "grad_norm": 1.7248355746871664, "learning_rate": 6.004407261169369e-06, "loss": 2.4249, "step": 56930 }, { "epoch": 1.28, "grad_norm": 1.6453860306684502, "learning_rate": 6.001059833213513e-06, "loss": 2.2758, "step": 56940 }, { "epoch": 1.28, "grad_norm": 1.640223896063578, "learning_rate": 5.997712938561963e-06, "loss": 2.3498, "step": 56950 }, { "epoch": 1.28, "grad_norm": 1.5864726098736077, "learning_rate": 5.994366577661063e-06, "loss": 2.5006, "step": 56960 }, { "epoch": 1.28, "grad_norm": 1.5977003021423133, "learning_rate": 5.991020750957095e-06, "loss": 2.5122, "step": 56970 }, { "epoch": 1.28, "grad_norm": 1.592540145978629, "learning_rate": 5.987675458896258e-06, "loss": 2.3268, "step": 56980 }, { "epoch": 1.29, "grad_norm": 1.6169042761897068, "learning_rate": 5.984330701924688e-06, "loss": 2.3648, "step": 56990 }, { "epoch": 1.29, "grad_norm": 1.8204102187949143, "learning_rate": 5.980986480488445e-06, "loss": 2.4086, "step": 57000 }, { "epoch": 1.29, "grad_norm": 1.6725531299486294, "learning_rate": 5.9776427950335184e-06, "loss": 2.4005, "step": 57010 }, { "epoch": 1.29, "grad_norm": 1.7118224751333528, "learning_rate": 5.9742996460058245e-06, "loss": 2.4536, "step": 57020 }, { "epoch": 1.29, "grad_norm": 1.5424803004478884, "learning_rate": 5.970957033851214e-06, "loss": 2.4372, "step": 57030 }, { "epoch": 1.29, "grad_norm": 1.7952445404980033, "learning_rate": 5.967614959015464e-06, "loss": 2.4274, "step": 57040 }, { "epoch": 1.29, "grad_norm": 1.6603464737724658, "learning_rate": 5.9642734219442756e-06, "loss": 2.3764, "step": 57050 }, { "epoch": 1.29, "grad_norm": 1.7427019662519232, "learning_rate": 5.960932423083279e-06, "loss": 2.5123, "step": 57060 }, { "epoch": 1.29, "grad_norm": 1.6431639224564565, "learning_rate": 5.9575919628780375e-06, "loss": 2.5387, "step": 57070 }, { "epoch": 1.29, "grad_norm": 1.716068913128833, "learning_rate": 5.95425204177404e-06, "loss": 2.4303, "step": 57080 }, { "epoch": 1.29, "grad_norm": 1.912207805934612, "learning_rate": 5.950912660216701e-06, "loss": 2.436, "step": 57090 }, { "epoch": 1.29, "grad_norm": 1.862304893548876, "learning_rate": 5.947573818651364e-06, "loss": 2.4688, "step": 57100 }, { "epoch": 1.29, "grad_norm": 1.5740693380719426, "learning_rate": 5.944235517523311e-06, "loss": 2.4665, "step": 57110 }, { "epoch": 1.29, "grad_norm": 1.5959743467395833, "learning_rate": 5.940897757277728e-06, "loss": 2.5115, "step": 57120 }, { "epoch": 1.29, "grad_norm": 1.9497125184627164, "learning_rate": 5.937560538359752e-06, "loss": 2.4068, "step": 57130 }, { "epoch": 1.29, "grad_norm": 1.755536751730932, "learning_rate": 5.93422386121444e-06, "loss": 2.4252, "step": 57140 }, { "epoch": 1.29, "grad_norm": 1.7024370277692118, "learning_rate": 5.930887726286774e-06, "loss": 2.3672, "step": 57150 }, { "epoch": 1.29, "grad_norm": 1.8617873906975673, "learning_rate": 5.927552134021661e-06, "loss": 2.2389, "step": 57160 }, { "epoch": 1.29, "grad_norm": 1.5698694045895227, "learning_rate": 5.924217084863948e-06, "loss": 2.4246, "step": 57170 }, { "epoch": 1.29, "grad_norm": 1.575733892978207, "learning_rate": 5.920882579258399e-06, "loss": 2.3564, "step": 57180 }, { "epoch": 1.29, "grad_norm": 1.653021945520733, "learning_rate": 5.917548617649709e-06, "loss": 2.4041, "step": 57190 }, { "epoch": 1.29, "grad_norm": 1.606168605806241, "learning_rate": 5.914215200482495e-06, "loss": 2.3544, "step": 57200 }, { "epoch": 1.29, "grad_norm": 1.6664263521281553, "learning_rate": 5.9108823282013125e-06, "loss": 2.5723, "step": 57210 }, { "epoch": 1.29, "grad_norm": 1.5352445348021015, "learning_rate": 5.907550001250638e-06, "loss": 2.3922, "step": 57220 }, { "epoch": 1.29, "grad_norm": 1.7317092083062557, "learning_rate": 5.904218220074871e-06, "loss": 2.3508, "step": 57230 }, { "epoch": 1.29, "grad_norm": 1.673964007693396, "learning_rate": 5.9008869851183434e-06, "loss": 2.4164, "step": 57240 }, { "epoch": 1.29, "grad_norm": 1.5412718452982768, "learning_rate": 5.897556296825317e-06, "loss": 2.3976, "step": 57250 }, { "epoch": 1.29, "grad_norm": 1.657599820577507, "learning_rate": 5.89422615563997e-06, "loss": 2.3522, "step": 57260 }, { "epoch": 1.29, "grad_norm": 1.6961490145290306, "learning_rate": 5.890896562006423e-06, "loss": 2.4509, "step": 57270 }, { "epoch": 1.29, "grad_norm": 1.6967435753909303, "learning_rate": 5.8875675163687106e-06, "loss": 2.4047, "step": 57280 }, { "epoch": 1.29, "grad_norm": 1.5349082831911154, "learning_rate": 5.884239019170799e-06, "loss": 2.4602, "step": 57290 }, { "epoch": 1.29, "grad_norm": 1.6255764172008178, "learning_rate": 5.8809110708565845e-06, "loss": 2.5307, "step": 57300 }, { "epoch": 1.29, "grad_norm": 1.6104396283821787, "learning_rate": 5.877583671869884e-06, "loss": 2.3944, "step": 57310 }, { "epoch": 1.29, "grad_norm": 1.7309924882669785, "learning_rate": 5.874256822654445e-06, "loss": 2.3412, "step": 57320 }, { "epoch": 1.29, "grad_norm": 1.7243738510463416, "learning_rate": 5.870930523653944e-06, "loss": 2.4008, "step": 57330 }, { "epoch": 1.29, "grad_norm": 1.7047183996778732, "learning_rate": 5.8676047753119745e-06, "loss": 2.4025, "step": 57340 }, { "epoch": 1.29, "grad_norm": 1.5299496306541835, "learning_rate": 5.864279578072063e-06, "loss": 2.4413, "step": 57350 }, { "epoch": 1.29, "grad_norm": 1.606310331667937, "learning_rate": 5.860954932377669e-06, "loss": 2.4885, "step": 57360 }, { "epoch": 1.29, "grad_norm": 1.7487000326948756, "learning_rate": 5.8576308386721705e-06, "loss": 2.4205, "step": 57370 }, { "epoch": 1.29, "grad_norm": 1.5755934900436754, "learning_rate": 5.854307297398869e-06, "loss": 2.3713, "step": 57380 }, { "epoch": 1.29, "grad_norm": 1.580239654978476, "learning_rate": 5.850984309000998e-06, "loss": 2.432, "step": 57390 }, { "epoch": 1.29, "grad_norm": 1.623654345466938, "learning_rate": 5.847661873921718e-06, "loss": 2.3741, "step": 57400 }, { "epoch": 1.29, "grad_norm": 1.7691751804703575, "learning_rate": 5.8443399926041135e-06, "loss": 2.4193, "step": 57410 }, { "epoch": 1.29, "grad_norm": 1.7570740303905648, "learning_rate": 5.841018665491195e-06, "loss": 2.2465, "step": 57420 }, { "epoch": 1.3, "grad_norm": 1.6782997514576725, "learning_rate": 5.8376978930259e-06, "loss": 2.3921, "step": 57430 }, { "epoch": 1.3, "grad_norm": 1.5905766933579357, "learning_rate": 5.834377675651088e-06, "loss": 2.5516, "step": 57440 }, { "epoch": 1.3, "grad_norm": 1.4431390287401036, "learning_rate": 5.831058013809552e-06, "loss": 2.402, "step": 57450 }, { "epoch": 1.3, "grad_norm": 1.6038791324130066, "learning_rate": 5.827738907944003e-06, "loss": 2.415, "step": 57460 }, { "epoch": 1.3, "grad_norm": 1.7963954664813089, "learning_rate": 5.8244203584970805e-06, "loss": 2.3723, "step": 57470 }, { "epoch": 1.3, "grad_norm": 1.6521208524455298, "learning_rate": 5.8211023659113575e-06, "loss": 2.2997, "step": 57480 }, { "epoch": 1.3, "grad_norm": 1.6570191004878914, "learning_rate": 5.817784930629323e-06, "loss": 2.5175, "step": 57490 }, { "epoch": 1.3, "grad_norm": 1.5457488022325698, "learning_rate": 5.814468053093394e-06, "loss": 2.4641, "step": 57500 }, { "epoch": 1.3, "grad_norm": 1.5514165440446042, "learning_rate": 5.8111517337459155e-06, "loss": 2.3169, "step": 57510 }, { "epoch": 1.3, "grad_norm": 1.6138633062607148, "learning_rate": 5.8078359730291545e-06, "loss": 2.3624, "step": 57520 }, { "epoch": 1.3, "grad_norm": 1.9749073131634494, "learning_rate": 5.804520771385307e-06, "loss": 2.4376, "step": 57530 }, { "epoch": 1.3, "grad_norm": 1.7444860604024244, "learning_rate": 5.801206129256493e-06, "loss": 2.4302, "step": 57540 }, { "epoch": 1.3, "grad_norm": 1.7281923864073476, "learning_rate": 5.797892047084756e-06, "loss": 2.4361, "step": 57550 }, { "epoch": 1.3, "grad_norm": 1.7744546580635197, "learning_rate": 5.794578525312069e-06, "loss": 2.415, "step": 57560 }, { "epoch": 1.3, "grad_norm": 1.6034781365515622, "learning_rate": 5.791265564380327e-06, "loss": 2.4569, "step": 57570 }, { "epoch": 1.3, "grad_norm": 1.4333248633640285, "learning_rate": 5.7879531647313505e-06, "loss": 2.3788, "step": 57580 }, { "epoch": 1.3, "grad_norm": 1.7220028583359805, "learning_rate": 5.784641326806888e-06, "loss": 2.3201, "step": 57590 }, { "epoch": 1.3, "grad_norm": 1.4913880932119556, "learning_rate": 5.781330051048608e-06, "loss": 2.4888, "step": 57600 }, { "epoch": 1.3, "grad_norm": 1.7650545898536132, "learning_rate": 5.7780193378981045e-06, "loss": 2.3978, "step": 57610 }, { "epoch": 1.3, "grad_norm": 1.6899071232455347, "learning_rate": 5.774709187796907e-06, "loss": 2.4844, "step": 57620 }, { "epoch": 1.3, "grad_norm": 1.6856110288331378, "learning_rate": 5.771399601186456e-06, "loss": 2.2801, "step": 57630 }, { "epoch": 1.3, "grad_norm": 1.8162624158665657, "learning_rate": 5.768090578508125e-06, "loss": 2.5764, "step": 57640 }, { "epoch": 1.3, "grad_norm": 1.7608317335359998, "learning_rate": 5.764782120203214e-06, "loss": 2.4474, "step": 57650 }, { "epoch": 1.3, "grad_norm": 1.678772804103706, "learning_rate": 5.761474226712933e-06, "loss": 2.4456, "step": 57660 }, { "epoch": 1.3, "grad_norm": 1.7254955847664428, "learning_rate": 5.75816689847843e-06, "loss": 2.4064, "step": 57670 }, { "epoch": 1.3, "grad_norm": 1.787970657477758, "learning_rate": 5.7548601359407805e-06, "loss": 2.3424, "step": 57680 }, { "epoch": 1.3, "grad_norm": 1.6823256731564185, "learning_rate": 5.751553939540978e-06, "loss": 2.415, "step": 57690 }, { "epoch": 1.3, "grad_norm": 1.8753284478794487, "learning_rate": 5.748248309719939e-06, "loss": 2.4179, "step": 57700 }, { "epoch": 1.3, "grad_norm": 1.6046507434240738, "learning_rate": 5.744943246918508e-06, "loss": 2.4881, "step": 57710 }, { "epoch": 1.3, "grad_norm": 1.5534131386468402, "learning_rate": 5.741638751577454e-06, "loss": 2.2848, "step": 57720 }, { "epoch": 1.3, "grad_norm": 1.448723534545617, "learning_rate": 5.738334824137468e-06, "loss": 2.5327, "step": 57730 }, { "epoch": 1.3, "grad_norm": 1.534452184740319, "learning_rate": 5.735031465039161e-06, "loss": 2.4301, "step": 57740 }, { "epoch": 1.3, "grad_norm": 1.525343085186709, "learning_rate": 5.731728674723084e-06, "loss": 2.5197, "step": 57750 }, { "epoch": 1.3, "grad_norm": 1.622136312942272, "learning_rate": 5.728426453629701e-06, "loss": 2.4191, "step": 57760 }, { "epoch": 1.3, "grad_norm": 1.6206005491046327, "learning_rate": 5.725124802199391e-06, "loss": 2.4265, "step": 57770 }, { "epoch": 1.3, "grad_norm": 1.5825216579320156, "learning_rate": 5.721823720872475e-06, "loss": 2.3938, "step": 57780 }, { "epoch": 1.3, "grad_norm": 1.6777589573501732, "learning_rate": 5.7185232100891855e-06, "loss": 2.2991, "step": 57790 }, { "epoch": 1.3, "grad_norm": 1.513159786917011, "learning_rate": 5.715223270289682e-06, "loss": 2.3977, "step": 57800 }, { "epoch": 1.3, "grad_norm": 1.8922454411541285, "learning_rate": 5.711923901914056e-06, "loss": 2.4639, "step": 57810 }, { "epoch": 1.3, "grad_norm": 1.5870677231434087, "learning_rate": 5.7086251054023125e-06, "loss": 2.4208, "step": 57820 }, { "epoch": 1.3, "grad_norm": 1.674959258089455, "learning_rate": 5.705326881194383e-06, "loss": 2.4134, "step": 57830 }, { "epoch": 1.3, "grad_norm": 1.6231123661184164, "learning_rate": 5.702029229730124e-06, "loss": 2.3381, "step": 57840 }, { "epoch": 1.3, "grad_norm": 1.5139361738791735, "learning_rate": 5.698732151449316e-06, "loss": 2.3538, "step": 57850 }, { "epoch": 1.3, "grad_norm": 1.5704266642061173, "learning_rate": 5.695435646791659e-06, "loss": 2.388, "step": 57860 }, { "epoch": 1.31, "grad_norm": 1.5454082153031583, "learning_rate": 5.692139716196781e-06, "loss": 2.3531, "step": 57870 }, { "epoch": 1.31, "grad_norm": 1.5588216153651906, "learning_rate": 5.688844360104233e-06, "loss": 2.5922, "step": 57880 }, { "epoch": 1.31, "grad_norm": 1.792132254950973, "learning_rate": 5.685549578953487e-06, "loss": 2.3193, "step": 57890 }, { "epoch": 1.31, "grad_norm": 1.6608341627217884, "learning_rate": 5.682255373183941e-06, "loss": 2.4825, "step": 57900 }, { "epoch": 1.31, "grad_norm": 1.6161961636031936, "learning_rate": 5.678961743234914e-06, "loss": 2.3516, "step": 57910 }, { "epoch": 1.31, "grad_norm": 1.7011959804608556, "learning_rate": 5.675668689545649e-06, "loss": 2.4271, "step": 57920 }, { "epoch": 1.31, "grad_norm": 1.6059082083288747, "learning_rate": 5.672376212555313e-06, "loss": 2.3161, "step": 57930 }, { "epoch": 1.31, "grad_norm": 1.5430802833510249, "learning_rate": 5.66908431270299e-06, "loss": 2.434, "step": 57940 }, { "epoch": 1.31, "grad_norm": 1.789514952131455, "learning_rate": 5.665792990427703e-06, "loss": 2.4497, "step": 57950 }, { "epoch": 1.31, "grad_norm": 1.5712203190363967, "learning_rate": 5.6625022461683794e-06, "loss": 2.4286, "step": 57960 }, { "epoch": 1.31, "grad_norm": 1.5368438740699555, "learning_rate": 5.659212080363886e-06, "loss": 2.4167, "step": 57970 }, { "epoch": 1.31, "grad_norm": 1.5864021142950118, "learning_rate": 5.655922493452992e-06, "loss": 2.294, "step": 57980 }, { "epoch": 1.31, "grad_norm": 1.8411851769129615, "learning_rate": 5.652633485874409e-06, "loss": 2.3022, "step": 57990 }, { "epoch": 1.31, "grad_norm": 1.6299675333984507, "learning_rate": 5.649345058066757e-06, "loss": 2.364, "step": 58000 }, { "epoch": 1.31, "grad_norm": 1.482596005991219, "learning_rate": 5.646057210468594e-06, "loss": 2.4461, "step": 58010 }, { "epoch": 1.31, "grad_norm": 1.694579141305594, "learning_rate": 5.642769943518389e-06, "loss": 2.3449, "step": 58020 }, { "epoch": 1.31, "grad_norm": 1.5747037730283426, "learning_rate": 5.639483257654534e-06, "loss": 2.4209, "step": 58030 }, { "epoch": 1.31, "grad_norm": 1.9140978461646552, "learning_rate": 5.63619715331535e-06, "loss": 2.5368, "step": 58040 }, { "epoch": 1.31, "grad_norm": 1.6803273633764373, "learning_rate": 5.632911630939073e-06, "loss": 2.3025, "step": 58050 }, { "epoch": 1.31, "grad_norm": 1.754762588589534, "learning_rate": 5.629626690963868e-06, "loss": 2.496, "step": 58060 }, { "epoch": 1.31, "grad_norm": 1.7912563702220232, "learning_rate": 5.626342333827815e-06, "loss": 2.4469, "step": 58070 }, { "epoch": 1.31, "grad_norm": 1.4493215758369262, "learning_rate": 5.623058559968923e-06, "loss": 2.4662, "step": 58080 }, { "epoch": 1.31, "grad_norm": 1.497487311370707, "learning_rate": 5.619775369825123e-06, "loss": 2.5281, "step": 58090 }, { "epoch": 1.31, "grad_norm": 1.720517286791422, "learning_rate": 5.616492763834263e-06, "loss": 2.2995, "step": 58100 }, { "epoch": 1.31, "grad_norm": 1.6394392749425193, "learning_rate": 5.613210742434116e-06, "loss": 2.4561, "step": 58110 }, { "epoch": 1.31, "grad_norm": 1.6273669839869418, "learning_rate": 5.609929306062377e-06, "loss": 2.5582, "step": 58120 }, { "epoch": 1.31, "grad_norm": 1.6267020746821077, "learning_rate": 5.606648455156665e-06, "loss": 2.5529, "step": 58130 }, { "epoch": 1.31, "grad_norm": 1.7295765484073697, "learning_rate": 5.603368190154512e-06, "loss": 2.3213, "step": 58140 }, { "epoch": 1.31, "grad_norm": 1.6315790014242673, "learning_rate": 5.6000885114933885e-06, "loss": 2.4653, "step": 58150 }, { "epoch": 1.31, "grad_norm": 1.6714732739357108, "learning_rate": 5.596809419610673e-06, "loss": 2.4396, "step": 58160 }, { "epoch": 1.31, "grad_norm": 1.77449421986816, "learning_rate": 5.59353091494367e-06, "loss": 2.5308, "step": 58170 }, { "epoch": 1.31, "grad_norm": 1.902660671419295, "learning_rate": 5.590252997929606e-06, "loss": 2.5102, "step": 58180 }, { "epoch": 1.31, "grad_norm": 1.5518646355488872, "learning_rate": 5.58697566900563e-06, "loss": 2.3834, "step": 58190 }, { "epoch": 1.31, "grad_norm": 1.5042901631702323, "learning_rate": 5.5836989286088015e-06, "loss": 2.3044, "step": 58200 }, { "epoch": 1.31, "grad_norm": 1.7434129024428207, "learning_rate": 5.5804227771761235e-06, "loss": 2.5032, "step": 58210 }, { "epoch": 1.31, "grad_norm": 1.5222773547545239, "learning_rate": 5.577147215144502e-06, "loss": 2.4722, "step": 58220 }, { "epoch": 1.31, "grad_norm": 1.5601982460561252, "learning_rate": 5.573872242950773e-06, "loss": 2.4226, "step": 58230 }, { "epoch": 1.31, "grad_norm": 1.6566516978861614, "learning_rate": 5.57059786103169e-06, "loss": 2.3886, "step": 58240 }, { "epoch": 1.31, "grad_norm": 1.8709472342302256, "learning_rate": 5.567324069823931e-06, "loss": 2.3954, "step": 58250 }, { "epoch": 1.31, "grad_norm": 1.6697125059853788, "learning_rate": 5.564050869764091e-06, "loss": 2.4004, "step": 58260 }, { "epoch": 1.31, "grad_norm": 1.6474725892121778, "learning_rate": 5.560778261288685e-06, "loss": 2.4529, "step": 58270 }, { "epoch": 1.31, "grad_norm": 1.8507290517408639, "learning_rate": 5.557506244834162e-06, "loss": 2.3208, "step": 58280 }, { "epoch": 1.31, "grad_norm": 1.6797174140180895, "learning_rate": 5.554234820836878e-06, "loss": 2.3985, "step": 58290 }, { "epoch": 1.31, "grad_norm": 1.5023147375493477, "learning_rate": 5.550963989733118e-06, "loss": 2.3635, "step": 58300 }, { "epoch": 1.31, "grad_norm": 1.7419966298519343, "learning_rate": 5.547693751959078e-06, "loss": 2.3798, "step": 58310 }, { "epoch": 1.32, "grad_norm": 1.6613502236435154, "learning_rate": 5.544424107950885e-06, "loss": 2.5563, "step": 58320 }, { "epoch": 1.32, "grad_norm": 1.677462706859379, "learning_rate": 5.541155058144578e-06, "loss": 2.3972, "step": 58330 }, { "epoch": 1.32, "grad_norm": 1.5845058933292082, "learning_rate": 5.537886602976134e-06, "loss": 2.4532, "step": 58340 }, { "epoch": 1.32, "grad_norm": 1.8409963647219876, "learning_rate": 5.53461874288143e-06, "loss": 2.452, "step": 58350 }, { "epoch": 1.32, "grad_norm": 1.7409729482587835, "learning_rate": 5.531351478296275e-06, "loss": 2.4441, "step": 58360 }, { "epoch": 1.32, "grad_norm": 1.800246851565299, "learning_rate": 5.528084809656395e-06, "loss": 2.5304, "step": 58370 }, { "epoch": 1.32, "grad_norm": 1.7273264853800625, "learning_rate": 5.524818737397439e-06, "loss": 2.3777, "step": 58380 }, { "epoch": 1.32, "grad_norm": 1.7492439230776022, "learning_rate": 5.521553261954975e-06, "loss": 2.4871, "step": 58390 }, { "epoch": 1.32, "grad_norm": 1.6333479354531186, "learning_rate": 5.51828838376449e-06, "loss": 2.3973, "step": 58400 }, { "epoch": 1.32, "grad_norm": 1.6408454418077778, "learning_rate": 5.515024103261393e-06, "loss": 2.3268, "step": 58410 }, { "epoch": 1.32, "grad_norm": 1.61830399171598, "learning_rate": 5.511760420881013e-06, "loss": 2.4458, "step": 58420 }, { "epoch": 1.32, "grad_norm": 1.428678350296599, "learning_rate": 5.5084973370586e-06, "loss": 2.3284, "step": 58430 }, { "epoch": 1.32, "grad_norm": 1.7521340875248852, "learning_rate": 5.505234852229322e-06, "loss": 2.3732, "step": 58440 }, { "epoch": 1.32, "grad_norm": 1.5576953279527181, "learning_rate": 5.501972966828272e-06, "loss": 2.3636, "step": 58450 }, { "epoch": 1.32, "grad_norm": 1.5002440265859769, "learning_rate": 5.498711681290456e-06, "loss": 2.3598, "step": 58460 }, { "epoch": 1.32, "grad_norm": 1.5645917253879669, "learning_rate": 5.4954509960508015e-06, "loss": 2.5053, "step": 58470 }, { "epoch": 1.32, "grad_norm": 1.6709824892681726, "learning_rate": 5.492190911544164e-06, "loss": 2.4256, "step": 58480 }, { "epoch": 1.32, "grad_norm": 1.621857288707758, "learning_rate": 5.488931428205313e-06, "loss": 2.408, "step": 58490 }, { "epoch": 1.32, "grad_norm": 1.630513905929844, "learning_rate": 5.485672546468934e-06, "loss": 2.3971, "step": 58500 }, { "epoch": 1.32, "grad_norm": 1.6538831609486926, "learning_rate": 5.482414266769642e-06, "loss": 2.4414, "step": 58510 }, { "epoch": 1.32, "grad_norm": 1.701509769728939, "learning_rate": 5.479156589541956e-06, "loss": 2.4807, "step": 58520 }, { "epoch": 1.32, "grad_norm": 1.575682095002363, "learning_rate": 5.475899515220327e-06, "loss": 2.3456, "step": 58530 }, { "epoch": 1.32, "grad_norm": 1.7427681982443914, "learning_rate": 5.47264304423913e-06, "loss": 2.4507, "step": 58540 }, { "epoch": 1.32, "grad_norm": 1.8103457147451443, "learning_rate": 5.469387177032648e-06, "loss": 2.3268, "step": 58550 }, { "epoch": 1.32, "grad_norm": 1.5795010685133564, "learning_rate": 5.466131914035086e-06, "loss": 2.4774, "step": 58560 }, { "epoch": 1.32, "grad_norm": 1.6951234738995051, "learning_rate": 5.462877255680576e-06, "loss": 2.3221, "step": 58570 }, { "epoch": 1.32, "grad_norm": 1.7345239946980477, "learning_rate": 5.459623202403159e-06, "loss": 2.3797, "step": 58580 }, { "epoch": 1.32, "grad_norm": 1.5553033683994328, "learning_rate": 5.456369754636801e-06, "loss": 2.4209, "step": 58590 }, { "epoch": 1.32, "grad_norm": 1.6877672471168217, "learning_rate": 5.4531169128153885e-06, "loss": 2.5201, "step": 58600 }, { "epoch": 1.32, "grad_norm": 1.555418507632434, "learning_rate": 5.4498646773727205e-06, "loss": 2.3638, "step": 58610 }, { "epoch": 1.32, "grad_norm": 1.6602146109908507, "learning_rate": 5.446613048742528e-06, "loss": 2.4198, "step": 58620 }, { "epoch": 1.32, "grad_norm": 1.699645659202753, "learning_rate": 5.4433620273584456e-06, "loss": 2.4038, "step": 58630 }, { "epoch": 1.32, "grad_norm": 1.8410846462900425, "learning_rate": 5.440111613654037e-06, "loss": 2.2908, "step": 58640 }, { "epoch": 1.32, "grad_norm": 1.5964891527409508, "learning_rate": 5.436861808062779e-06, "loss": 2.529, "step": 58650 }, { "epoch": 1.32, "grad_norm": 1.727724159063125, "learning_rate": 5.433612611018074e-06, "loss": 2.4441, "step": 58660 }, { "epoch": 1.32, "grad_norm": 1.6958194383969094, "learning_rate": 5.430364022953234e-06, "loss": 2.3612, "step": 58670 }, { "epoch": 1.32, "grad_norm": 1.7103451330027422, "learning_rate": 5.4271160443015025e-06, "loss": 2.5058, "step": 58680 }, { "epoch": 1.32, "grad_norm": 1.7964246405514768, "learning_rate": 5.423868675496032e-06, "loss": 2.4808, "step": 58690 }, { "epoch": 1.32, "grad_norm": 1.5749194976472263, "learning_rate": 5.420621916969895e-06, "loss": 2.4435, "step": 58700 }, { "epoch": 1.32, "grad_norm": 1.6796336640404832, "learning_rate": 5.417375769156083e-06, "loss": 2.3739, "step": 58710 }, { "epoch": 1.32, "grad_norm": 1.7130639378122852, "learning_rate": 5.414130232487509e-06, "loss": 2.4484, "step": 58720 }, { "epoch": 1.32, "grad_norm": 1.5355879884981514, "learning_rate": 5.410885307397002e-06, "loss": 2.4578, "step": 58730 }, { "epoch": 1.32, "grad_norm": 1.8497887841034293, "learning_rate": 5.4076409943173095e-06, "loss": 2.3357, "step": 58740 }, { "epoch": 1.32, "grad_norm": 1.562314597474612, "learning_rate": 5.404397293681096e-06, "loss": 2.3914, "step": 58750 }, { "epoch": 1.33, "grad_norm": 1.5613362250345442, "learning_rate": 5.401154205920949e-06, "loss": 2.3652, "step": 58760 }, { "epoch": 1.33, "grad_norm": 1.7880820076977877, "learning_rate": 5.397911731469369e-06, "loss": 2.3823, "step": 58770 }, { "epoch": 1.33, "grad_norm": 1.4794312166485828, "learning_rate": 5.394669870758778e-06, "loss": 2.4311, "step": 58780 }, { "epoch": 1.33, "grad_norm": 1.4748980178777396, "learning_rate": 5.391428624221513e-06, "loss": 2.5115, "step": 58790 }, { "epoch": 1.33, "grad_norm": 1.553051143319884, "learning_rate": 5.388187992289831e-06, "loss": 2.4258, "step": 58800 }, { "epoch": 1.33, "grad_norm": 1.6669381571293331, "learning_rate": 5.384947975395911e-06, "loss": 2.3262, "step": 58810 }, { "epoch": 1.33, "grad_norm": 1.6023432521739094, "learning_rate": 5.381708573971847e-06, "loss": 2.402, "step": 58820 }, { "epoch": 1.33, "grad_norm": 1.6707728773423556, "learning_rate": 5.378469788449645e-06, "loss": 2.4144, "step": 58830 }, { "epoch": 1.33, "grad_norm": 1.7691830647588422, "learning_rate": 5.375231619261243e-06, "loss": 2.4145, "step": 58840 }, { "epoch": 1.33, "grad_norm": 1.4802966269232012, "learning_rate": 5.371994066838475e-06, "loss": 2.384, "step": 58850 }, { "epoch": 1.33, "grad_norm": 1.5988238528002545, "learning_rate": 5.36875713161311e-06, "loss": 2.4548, "step": 58860 }, { "epoch": 1.33, "grad_norm": 1.4680142297247987, "learning_rate": 5.365520814016836e-06, "loss": 2.5283, "step": 58870 }, { "epoch": 1.33, "grad_norm": 1.5996351753317373, "learning_rate": 5.362285114481248e-06, "loss": 2.4537, "step": 58880 }, { "epoch": 1.33, "grad_norm": 1.5830783636799393, "learning_rate": 5.359050033437866e-06, "loss": 2.3144, "step": 58890 }, { "epoch": 1.33, "grad_norm": 1.6697333535857457, "learning_rate": 5.355815571318122e-06, "loss": 2.4112, "step": 58900 }, { "epoch": 1.33, "grad_norm": 1.8100638386418646, "learning_rate": 5.352581728553371e-06, "loss": 2.461, "step": 58910 }, { "epoch": 1.33, "grad_norm": 1.5513804685976205, "learning_rate": 5.349348505574882e-06, "loss": 2.3763, "step": 58920 }, { "epoch": 1.33, "grad_norm": 1.548547836784938, "learning_rate": 5.346115902813842e-06, "loss": 2.4746, "step": 58930 }, { "epoch": 1.33, "grad_norm": 1.636783212154764, "learning_rate": 5.34288392070135e-06, "loss": 2.4634, "step": 58940 }, { "epoch": 1.33, "grad_norm": 1.5623713906151109, "learning_rate": 5.339652559668442e-06, "loss": 2.446, "step": 58950 }, { "epoch": 1.33, "grad_norm": 1.6168098879883581, "learning_rate": 5.336421820146045e-06, "loss": 2.4436, "step": 58960 }, { "epoch": 1.33, "grad_norm": 1.723317105701288, "learning_rate": 5.333191702565017e-06, "loss": 2.4873, "step": 58970 }, { "epoch": 1.33, "grad_norm": 1.7611458679058163, "learning_rate": 5.329962207356132e-06, "loss": 2.406, "step": 58980 }, { "epoch": 1.33, "grad_norm": 1.6236084990016388, "learning_rate": 5.326733334950081e-06, "loss": 2.5323, "step": 58990 }, { "epoch": 1.33, "grad_norm": 1.5255511999072229, "learning_rate": 5.323505085777465e-06, "loss": 2.4706, "step": 59000 }, { "epoch": 1.33, "grad_norm": 1.6283843410167105, "learning_rate": 5.320277460268817e-06, "loss": 2.3873, "step": 59010 }, { "epoch": 1.33, "grad_norm": 1.7899981392347306, "learning_rate": 5.317050458854575e-06, "loss": 2.2778, "step": 59020 }, { "epoch": 1.33, "grad_norm": 1.7255144212265188, "learning_rate": 5.3138240819650935e-06, "loss": 2.4351, "step": 59030 }, { "epoch": 1.33, "grad_norm": 1.6158590330573601, "learning_rate": 5.31059833003065e-06, "loss": 2.4157, "step": 59040 }, { "epoch": 1.33, "grad_norm": 1.7970422395579206, "learning_rate": 5.307373203481435e-06, "loss": 2.3692, "step": 59050 }, { "epoch": 1.33, "grad_norm": 1.534371969946733, "learning_rate": 5.304148702747549e-06, "loss": 2.5591, "step": 59060 }, { "epoch": 1.33, "grad_norm": 1.582663743324919, "learning_rate": 5.300924828259025e-06, "loss": 2.3626, "step": 59070 }, { "epoch": 1.33, "grad_norm": 1.6540149945893161, "learning_rate": 5.297701580445799e-06, "loss": 2.4125, "step": 59080 }, { "epoch": 1.33, "grad_norm": 1.7269965615907354, "learning_rate": 5.294478959737729e-06, "loss": 2.4914, "step": 59090 }, { "epoch": 1.33, "grad_norm": 1.4738690155741083, "learning_rate": 5.2912569665645886e-06, "loss": 2.3855, "step": 59100 }, { "epoch": 1.33, "grad_norm": 1.7340909720190862, "learning_rate": 5.288035601356066e-06, "loss": 2.3963, "step": 59110 }, { "epoch": 1.33, "grad_norm": 1.6979089832007759, "learning_rate": 5.284814864541769e-06, "loss": 2.4659, "step": 59120 }, { "epoch": 1.33, "grad_norm": 1.4057142395692528, "learning_rate": 5.281594756551218e-06, "loss": 2.4635, "step": 59130 }, { "epoch": 1.33, "grad_norm": 1.704265980749421, "learning_rate": 5.278375277813847e-06, "loss": 2.417, "step": 59140 }, { "epoch": 1.33, "grad_norm": 1.5611356021486906, "learning_rate": 5.275156428759021e-06, "loss": 2.4922, "step": 59150 }, { "epoch": 1.33, "grad_norm": 1.7755263341135084, "learning_rate": 5.271938209816006e-06, "loss": 2.4191, "step": 59160 }, { "epoch": 1.33, "grad_norm": 1.6751768772740963, "learning_rate": 5.268720621413983e-06, "loss": 2.5024, "step": 59170 }, { "epoch": 1.33, "grad_norm": 1.6478102661787817, "learning_rate": 5.265503663982057e-06, "loss": 2.3943, "step": 59180 }, { "epoch": 1.33, "grad_norm": 1.7982788044171751, "learning_rate": 5.262287337949247e-06, "loss": 2.4896, "step": 59190 }, { "epoch": 1.34, "grad_norm": 1.829259696966733, "learning_rate": 5.259071643744483e-06, "loss": 2.3064, "step": 59200 }, { "epoch": 1.34, "grad_norm": 1.513734024608852, "learning_rate": 5.255856581796619e-06, "loss": 2.4631, "step": 59210 }, { "epoch": 1.34, "grad_norm": 1.4983227475382574, "learning_rate": 5.252642152534422e-06, "loss": 2.4315, "step": 59220 }, { "epoch": 1.34, "grad_norm": 1.7010106625161066, "learning_rate": 5.249428356386568e-06, "loss": 2.4865, "step": 59230 }, { "epoch": 1.34, "grad_norm": 1.568544534442258, "learning_rate": 5.246215193781655e-06, "loss": 2.4828, "step": 59240 }, { "epoch": 1.34, "grad_norm": 1.5800723324559756, "learning_rate": 5.2430026651481955e-06, "loss": 2.4101, "step": 59250 }, { "epoch": 1.34, "grad_norm": 1.7064878358561204, "learning_rate": 5.239790770914616e-06, "loss": 2.3117, "step": 59260 }, { "epoch": 1.34, "grad_norm": 1.6772536612741435, "learning_rate": 5.236579511509261e-06, "loss": 2.4613, "step": 59270 }, { "epoch": 1.34, "grad_norm": 1.832117044982822, "learning_rate": 5.233368887360386e-06, "loss": 2.4909, "step": 59280 }, { "epoch": 1.34, "grad_norm": 1.8079869688365988, "learning_rate": 5.230158898896167e-06, "loss": 2.5583, "step": 59290 }, { "epoch": 1.34, "grad_norm": 1.7472850119555388, "learning_rate": 5.22694954654469e-06, "loss": 2.4493, "step": 59300 }, { "epoch": 1.34, "grad_norm": 1.534452031943297, "learning_rate": 5.223740830733961e-06, "loss": 2.3649, "step": 59310 }, { "epoch": 1.34, "grad_norm": 1.7353870027838758, "learning_rate": 5.220532751891896e-06, "loss": 2.5097, "step": 59320 }, { "epoch": 1.34, "grad_norm": 1.7473479112347179, "learning_rate": 5.2173253104463285e-06, "loss": 2.4481, "step": 59330 }, { "epoch": 1.34, "grad_norm": 1.7573975374488477, "learning_rate": 5.214118506825012e-06, "loss": 2.525, "step": 59340 }, { "epoch": 1.34, "grad_norm": 1.5562127727060568, "learning_rate": 5.21091234145561e-06, "loss": 2.5303, "step": 59350 }, { "epoch": 1.34, "grad_norm": 1.7370939902798064, "learning_rate": 5.207706814765699e-06, "loss": 2.2853, "step": 59360 }, { "epoch": 1.34, "grad_norm": 1.6316772971100817, "learning_rate": 5.2045019271827726e-06, "loss": 2.425, "step": 59370 }, { "epoch": 1.34, "grad_norm": 1.6636688078590502, "learning_rate": 5.201297679134241e-06, "loss": 2.445, "step": 59380 }, { "epoch": 1.34, "grad_norm": 1.5676460913260224, "learning_rate": 5.198094071047421e-06, "loss": 2.3161, "step": 59390 }, { "epoch": 1.34, "grad_norm": 1.5388328410114687, "learning_rate": 5.194891103349557e-06, "loss": 2.3931, "step": 59400 }, { "epoch": 1.34, "grad_norm": 1.7629913335753635, "learning_rate": 5.1916887764678e-06, "loss": 2.4392, "step": 59410 }, { "epoch": 1.34, "grad_norm": 1.6044708391685254, "learning_rate": 5.188487090829215e-06, "loss": 2.473, "step": 59420 }, { "epoch": 1.34, "grad_norm": 1.566950300011339, "learning_rate": 5.185286046860786e-06, "loss": 2.3349, "step": 59430 }, { "epoch": 1.34, "grad_norm": 1.7601693693662679, "learning_rate": 5.182085644989406e-06, "loss": 2.3651, "step": 59440 }, { "epoch": 1.34, "grad_norm": 1.907028149660656, "learning_rate": 5.178885885641887e-06, "loss": 2.475, "step": 59450 }, { "epoch": 1.34, "grad_norm": 1.726712581768231, "learning_rate": 5.175686769244952e-06, "loss": 2.345, "step": 59460 }, { "epoch": 1.34, "grad_norm": 1.588620176117486, "learning_rate": 5.172488296225236e-06, "loss": 2.3753, "step": 59470 }, { "epoch": 1.34, "grad_norm": 1.9928104313454529, "learning_rate": 5.1692904670093005e-06, "loss": 2.3681, "step": 59480 }, { "epoch": 1.34, "grad_norm": 1.647357335077868, "learning_rate": 5.166093282023612e-06, "loss": 2.3399, "step": 59490 }, { "epoch": 1.34, "grad_norm": 1.557643258611181, "learning_rate": 5.162896741694545e-06, "loss": 2.4974, "step": 59500 }, { "epoch": 1.34, "grad_norm": 1.5976453515477922, "learning_rate": 5.159700846448397e-06, "loss": 2.3913, "step": 59510 }, { "epoch": 1.34, "grad_norm": 1.7000663954574076, "learning_rate": 5.1565055967113765e-06, "loss": 2.4635, "step": 59520 }, { "epoch": 1.34, "grad_norm": 1.6635805084227646, "learning_rate": 5.1533109929096045e-06, "loss": 2.4071, "step": 59530 }, { "epoch": 1.34, "grad_norm": 1.5952860013919066, "learning_rate": 5.150117035469124e-06, "loss": 2.3617, "step": 59540 }, { "epoch": 1.34, "grad_norm": 1.9754650901963116, "learning_rate": 5.146923724815884e-06, "loss": 2.4226, "step": 59550 }, { "epoch": 1.34, "grad_norm": 1.5410124385142345, "learning_rate": 5.1437310613757475e-06, "loss": 2.3876, "step": 59560 }, { "epoch": 1.34, "grad_norm": 1.8446772116883552, "learning_rate": 5.140539045574491e-06, "loss": 2.3322, "step": 59570 }, { "epoch": 1.34, "grad_norm": 1.7628276526027307, "learning_rate": 5.13734767783781e-06, "loss": 2.4819, "step": 59580 }, { "epoch": 1.34, "grad_norm": 1.6804050474600878, "learning_rate": 5.134156958591305e-06, "loss": 2.4206, "step": 59590 }, { "epoch": 1.34, "grad_norm": 1.4464520426875673, "learning_rate": 5.1309668882605006e-06, "loss": 2.365, "step": 59600 }, { "epoch": 1.34, "grad_norm": 1.557744989899497, "learning_rate": 5.1277774672708235e-06, "loss": 2.5027, "step": 59610 }, { "epoch": 1.34, "grad_norm": 1.7660301410807906, "learning_rate": 5.124588696047623e-06, "loss": 2.4345, "step": 59620 }, { "epoch": 1.34, "grad_norm": 1.5712919018730596, "learning_rate": 5.121400575016157e-06, "loss": 2.3822, "step": 59630 }, { "epoch": 1.34, "grad_norm": 1.6120929549664405, "learning_rate": 5.118213104601597e-06, "loss": 2.5378, "step": 59640 }, { "epoch": 1.35, "grad_norm": 1.6018827736716104, "learning_rate": 5.11502628522903e-06, "loss": 2.3502, "step": 59650 }, { "epoch": 1.35, "grad_norm": 1.6429175385649615, "learning_rate": 5.111840117323454e-06, "loss": 2.4394, "step": 59660 }, { "epoch": 1.35, "grad_norm": 1.7188986740479342, "learning_rate": 5.108654601309777e-06, "loss": 2.4148, "step": 59670 }, { "epoch": 1.35, "grad_norm": 1.8028989206758066, "learning_rate": 5.105469737612833e-06, "loss": 2.439, "step": 59680 }, { "epoch": 1.35, "grad_norm": 1.8261446429294896, "learning_rate": 5.102285526657353e-06, "loss": 2.4357, "step": 59690 }, { "epoch": 1.35, "grad_norm": 1.7189526561690673, "learning_rate": 5.099101968867994e-06, "loss": 2.3667, "step": 59700 }, { "epoch": 1.35, "grad_norm": 1.734119545609725, "learning_rate": 5.095919064669312e-06, "loss": 2.3548, "step": 59710 }, { "epoch": 1.35, "grad_norm": 1.6398686008663215, "learning_rate": 5.092736814485788e-06, "loss": 2.3985, "step": 59720 }, { "epoch": 1.35, "grad_norm": 1.6469041932318276, "learning_rate": 5.089555218741804e-06, "loss": 2.2593, "step": 59730 }, { "epoch": 1.35, "grad_norm": 1.6376365005718334, "learning_rate": 5.086374277861675e-06, "loss": 2.31, "step": 59740 }, { "epoch": 1.35, "grad_norm": 1.6603374334765928, "learning_rate": 5.083193992269607e-06, "loss": 2.4454, "step": 59750 }, { "epoch": 1.35, "grad_norm": 1.596698630438801, "learning_rate": 5.080014362389732e-06, "loss": 2.2607, "step": 59760 }, { "epoch": 1.35, "grad_norm": 1.5573937495034098, "learning_rate": 5.076835388646086e-06, "loss": 2.4215, "step": 59770 }, { "epoch": 1.35, "grad_norm": 1.6758491236007624, "learning_rate": 5.073657071462622e-06, "loss": 2.444, "step": 59780 }, { "epoch": 1.35, "grad_norm": 1.5256390204400236, "learning_rate": 5.070479411263207e-06, "loss": 2.4374, "step": 59790 }, { "epoch": 1.35, "grad_norm": 1.675314017576633, "learning_rate": 5.0673024084716136e-06, "loss": 2.3531, "step": 59800 }, { "epoch": 1.35, "grad_norm": 1.7119278503183017, "learning_rate": 5.064126063511543e-06, "loss": 2.328, "step": 59810 }, { "epoch": 1.35, "grad_norm": 1.599420395300225, "learning_rate": 5.060950376806584e-06, "loss": 2.3024, "step": 59820 }, { "epoch": 1.35, "grad_norm": 1.5708130522473165, "learning_rate": 5.057775348780256e-06, "loss": 2.3807, "step": 59830 }, { "epoch": 1.35, "grad_norm": 1.6613876894299053, "learning_rate": 5.0546009798559845e-06, "loss": 2.4384, "step": 59840 }, { "epoch": 1.35, "grad_norm": 1.616274720274254, "learning_rate": 5.051427270457108e-06, "loss": 2.6397, "step": 59850 }, { "epoch": 1.35, "grad_norm": 1.6717216394421166, "learning_rate": 5.048254221006874e-06, "loss": 2.4886, "step": 59860 }, { "epoch": 1.35, "grad_norm": 1.5129555473829606, "learning_rate": 5.0450818319284515e-06, "loss": 2.504, "step": 59870 }, { "epoch": 1.35, "grad_norm": 1.7483675374449108, "learning_rate": 5.041910103644911e-06, "loss": 2.4069, "step": 59880 }, { "epoch": 1.35, "grad_norm": 1.5102452275376228, "learning_rate": 5.038739036579239e-06, "loss": 2.4423, "step": 59890 }, { "epoch": 1.35, "grad_norm": 1.663493962149726, "learning_rate": 5.035568631154333e-06, "loss": 2.4175, "step": 59900 }, { "epoch": 1.35, "grad_norm": 1.7239151755949238, "learning_rate": 5.032398887793003e-06, "loss": 2.5121, "step": 59910 }, { "epoch": 1.35, "grad_norm": 1.7626088286145614, "learning_rate": 5.02922980691797e-06, "loss": 2.4538, "step": 59920 }, { "epoch": 1.35, "grad_norm": 1.5267697653121095, "learning_rate": 5.026061388951869e-06, "loss": 2.4249, "step": 59930 }, { "epoch": 1.35, "grad_norm": 1.5902944835677033, "learning_rate": 5.022893634317242e-06, "loss": 2.4393, "step": 59940 }, { "epoch": 1.35, "grad_norm": 1.5600149195825763, "learning_rate": 5.019726543436547e-06, "loss": 2.4366, "step": 59950 }, { "epoch": 1.35, "grad_norm": 1.5896509668094956, "learning_rate": 5.016560116732151e-06, "loss": 2.4679, "step": 59960 }, { "epoch": 1.35, "grad_norm": 1.5149770277770964, "learning_rate": 5.013394354626333e-06, "loss": 2.2819, "step": 59970 }, { "epoch": 1.35, "grad_norm": 1.705416783344064, "learning_rate": 5.010229257541285e-06, "loss": 2.4533, "step": 59980 }, { "epoch": 1.35, "grad_norm": 1.5371851925562467, "learning_rate": 5.0070648258991065e-06, "loss": 2.3955, "step": 59990 }, { "epoch": 1.35, "grad_norm": 1.8814266571533178, "learning_rate": 5.00390106012181e-06, "loss": 2.4407, "step": 60000 }, { "epoch": 1.35, "grad_norm": 1.5849470301174016, "learning_rate": 5.000737960631323e-06, "loss": 2.3949, "step": 60010 }, { "epoch": 1.35, "grad_norm": 1.7710069224567666, "learning_rate": 4.997575527849484e-06, "loss": 2.387, "step": 60020 }, { "epoch": 1.35, "grad_norm": 1.66258992661219, "learning_rate": 4.99441376219803e-06, "loss": 2.237, "step": 60030 }, { "epoch": 1.35, "grad_norm": 1.6332817133405568, "learning_rate": 4.991252664098626e-06, "loss": 2.3481, "step": 60040 }, { "epoch": 1.35, "grad_norm": 1.9135447344788137, "learning_rate": 4.988092233972836e-06, "loss": 2.4953, "step": 60050 }, { "epoch": 1.35, "grad_norm": 1.7641392494129542, "learning_rate": 4.9849324722421375e-06, "loss": 2.483, "step": 60060 }, { "epoch": 1.35, "grad_norm": 1.5845487530008822, "learning_rate": 4.98177337932793e-06, "loss": 2.3821, "step": 60070 }, { "epoch": 1.35, "grad_norm": 1.7441839653677254, "learning_rate": 4.978614955651508e-06, "loss": 2.2635, "step": 60080 }, { "epoch": 1.36, "grad_norm": 1.4990467844428246, "learning_rate": 4.975457201634086e-06, "loss": 2.4804, "step": 60090 }, { "epoch": 1.36, "grad_norm": 1.7439702951539808, "learning_rate": 4.972300117696783e-06, "loss": 2.5249, "step": 60100 }, { "epoch": 1.36, "grad_norm": 1.8310545488678411, "learning_rate": 4.969143704260636e-06, "loss": 2.3432, "step": 60110 }, { "epoch": 1.36, "grad_norm": 1.795633677493711, "learning_rate": 4.965987961746587e-06, "loss": 2.5817, "step": 60120 }, { "epoch": 1.36, "grad_norm": 1.6519245967568872, "learning_rate": 4.962832890575488e-06, "loss": 2.403, "step": 60130 }, { "epoch": 1.36, "grad_norm": 2.4639284343909518, "learning_rate": 4.95967849116811e-06, "loss": 2.5783, "step": 60140 }, { "epoch": 1.36, "grad_norm": 1.794106329459119, "learning_rate": 4.956524763945121e-06, "loss": 2.3863, "step": 60150 }, { "epoch": 1.36, "grad_norm": 1.5680024425704193, "learning_rate": 4.953371709327112e-06, "loss": 2.2767, "step": 60160 }, { "epoch": 1.36, "grad_norm": 1.5901755083475277, "learning_rate": 4.950219327734575e-06, "loss": 2.4662, "step": 60170 }, { "epoch": 1.36, "grad_norm": 1.6860564083125642, "learning_rate": 4.947067619587917e-06, "loss": 2.5398, "step": 60180 }, { "epoch": 1.36, "grad_norm": 1.8214486727519172, "learning_rate": 4.943916585307457e-06, "loss": 2.4161, "step": 60190 }, { "epoch": 1.36, "grad_norm": 1.567253976497072, "learning_rate": 4.940766225313415e-06, "loss": 2.4171, "step": 60200 }, { "epoch": 1.36, "grad_norm": 1.7529748463566082, "learning_rate": 4.937616540025935e-06, "loss": 2.4439, "step": 60210 }, { "epoch": 1.36, "grad_norm": 1.7005687941796683, "learning_rate": 4.934467529865059e-06, "loss": 2.4171, "step": 60220 }, { "epoch": 1.36, "grad_norm": 1.6476062778946259, "learning_rate": 4.931319195250746e-06, "loss": 2.1817, "step": 60230 }, { "epoch": 1.36, "grad_norm": 1.615577780734439, "learning_rate": 4.9281715366028644e-06, "loss": 2.4385, "step": 60240 }, { "epoch": 1.36, "grad_norm": 1.8095003387017057, "learning_rate": 4.925024554341183e-06, "loss": 2.3009, "step": 60250 }, { "epoch": 1.36, "grad_norm": 1.6428320394158735, "learning_rate": 4.921878248885389e-06, "loss": 2.2804, "step": 60260 }, { "epoch": 1.36, "grad_norm": 1.6101603903470527, "learning_rate": 4.9187326206550845e-06, "loss": 2.3606, "step": 60270 }, { "epoch": 1.36, "grad_norm": 1.6865462250673453, "learning_rate": 4.9155876700697705e-06, "loss": 2.3748, "step": 60280 }, { "epoch": 1.36, "grad_norm": 1.5111945738377544, "learning_rate": 4.912443397548864e-06, "loss": 2.33, "step": 60290 }, { "epoch": 1.36, "grad_norm": 1.8119791305999562, "learning_rate": 4.909299803511687e-06, "loss": 2.3296, "step": 60300 }, { "epoch": 1.36, "grad_norm": 1.5053517817906377, "learning_rate": 4.906156888377476e-06, "loss": 2.5733, "step": 60310 }, { "epoch": 1.36, "grad_norm": 1.6063931381703327, "learning_rate": 4.903014652565375e-06, "loss": 2.418, "step": 60320 }, { "epoch": 1.36, "grad_norm": 1.6239609933584522, "learning_rate": 4.899873096494429e-06, "loss": 2.4101, "step": 60330 }, { "epoch": 1.36, "grad_norm": 1.5034975554214955, "learning_rate": 4.896732220583613e-06, "loss": 2.2785, "step": 60340 }, { "epoch": 1.36, "grad_norm": 1.6809724979069072, "learning_rate": 4.893592025251797e-06, "loss": 2.309, "step": 60350 }, { "epoch": 1.36, "grad_norm": 1.5943925600693998, "learning_rate": 4.8904525109177525e-06, "loss": 2.3757, "step": 60360 }, { "epoch": 1.36, "grad_norm": 1.5808680739813612, "learning_rate": 4.887313678000175e-06, "loss": 2.4466, "step": 60370 }, { "epoch": 1.36, "grad_norm": 1.8306581415671417, "learning_rate": 4.884175526917664e-06, "loss": 2.3571, "step": 60380 }, { "epoch": 1.36, "grad_norm": 1.634501279208587, "learning_rate": 4.881038058088723e-06, "loss": 2.3598, "step": 60390 }, { "epoch": 1.36, "grad_norm": 1.5647845173738064, "learning_rate": 4.8779012719317765e-06, "loss": 2.4572, "step": 60400 }, { "epoch": 1.36, "grad_norm": 1.7407375822385422, "learning_rate": 4.874765168865149e-06, "loss": 2.4791, "step": 60410 }, { "epoch": 1.36, "grad_norm": 1.685395157072328, "learning_rate": 4.871629749307073e-06, "loss": 2.3435, "step": 60420 }, { "epoch": 1.36, "grad_norm": 1.6775428754161865, "learning_rate": 4.868495013675695e-06, "loss": 2.3752, "step": 60430 }, { "epoch": 1.36, "grad_norm": 1.8258932210151826, "learning_rate": 4.865360962389065e-06, "loss": 2.2923, "step": 60440 }, { "epoch": 1.36, "grad_norm": 1.6897457069551989, "learning_rate": 4.862227595865147e-06, "loss": 2.3164, "step": 60450 }, { "epoch": 1.36, "grad_norm": 1.6357272762031771, "learning_rate": 4.85909491452181e-06, "loss": 2.4108, "step": 60460 }, { "epoch": 1.36, "grad_norm": 1.6618578401446933, "learning_rate": 4.8559629187768335e-06, "loss": 2.4304, "step": 60470 }, { "epoch": 1.36, "grad_norm": 1.7955294129143065, "learning_rate": 4.852831609047904e-06, "loss": 2.3616, "step": 60480 }, { "epoch": 1.36, "grad_norm": 1.5980036380904687, "learning_rate": 4.8497009857526176e-06, "loss": 2.5145, "step": 60490 }, { "epoch": 1.36, "grad_norm": 1.874734833447845, "learning_rate": 4.846571049308479e-06, "loss": 2.3756, "step": 60500 }, { "epoch": 1.36, "grad_norm": 1.8486532838543048, "learning_rate": 4.843441800132901e-06, "loss": 2.4777, "step": 60510 }, { "epoch": 1.36, "grad_norm": 1.4202188133639868, "learning_rate": 4.840313238643204e-06, "loss": 2.3166, "step": 60520 }, { "epoch": 1.37, "grad_norm": 1.685780468255683, "learning_rate": 4.837185365256615e-06, "loss": 2.5718, "step": 60530 }, { "epoch": 1.37, "grad_norm": 1.75682740859833, "learning_rate": 4.8340581803902776e-06, "loss": 2.3114, "step": 60540 }, { "epoch": 1.37, "grad_norm": 1.6760958854987624, "learning_rate": 4.830931684461234e-06, "loss": 2.3645, "step": 60550 }, { "epoch": 1.37, "grad_norm": 1.587980352081453, "learning_rate": 4.827805877886443e-06, "loss": 2.4749, "step": 60560 }, { "epoch": 1.37, "grad_norm": 1.7753955402650952, "learning_rate": 4.824680761082758e-06, "loss": 2.4044, "step": 60570 }, { "epoch": 1.37, "grad_norm": 1.6027385730280972, "learning_rate": 4.821556334466954e-06, "loss": 2.4598, "step": 60580 }, { "epoch": 1.37, "grad_norm": 1.6152258322475657, "learning_rate": 4.818432598455702e-06, "loss": 2.4303, "step": 60590 }, { "epoch": 1.37, "grad_norm": 1.5647825754956899, "learning_rate": 4.815309553465599e-06, "loss": 2.3726, "step": 60600 }, { "epoch": 1.37, "grad_norm": 1.5920776372584242, "learning_rate": 4.812187199913134e-06, "loss": 2.4223, "step": 60610 }, { "epoch": 1.37, "grad_norm": 1.7265613284982968, "learning_rate": 4.809065538214708e-06, "loss": 2.494, "step": 60620 }, { "epoch": 1.37, "grad_norm": 1.5757076591105326, "learning_rate": 4.80594456878663e-06, "loss": 2.3026, "step": 60630 }, { "epoch": 1.37, "grad_norm": 1.7663406083401743, "learning_rate": 4.802824292045118e-06, "loss": 2.4229, "step": 60640 }, { "epoch": 1.37, "grad_norm": 1.6030702672018644, "learning_rate": 4.799704708406294e-06, "loss": 2.3988, "step": 60650 }, { "epoch": 1.37, "grad_norm": 1.5832405458498637, "learning_rate": 4.796585818286194e-06, "loss": 2.3994, "step": 60660 }, { "epoch": 1.37, "grad_norm": 2.0334557965701974, "learning_rate": 4.793467622100753e-06, "loss": 2.3161, "step": 60670 }, { "epoch": 1.37, "grad_norm": 1.6103953776561377, "learning_rate": 4.790350120265822e-06, "loss": 2.4578, "step": 60680 }, { "epoch": 1.37, "grad_norm": 1.6745154269709772, "learning_rate": 4.7872333131971526e-06, "loss": 2.4535, "step": 60690 }, { "epoch": 1.37, "grad_norm": 1.5866600555820483, "learning_rate": 4.7841172013104085e-06, "loss": 2.3688, "step": 60700 }, { "epoch": 1.37, "grad_norm": 1.6565482467206587, "learning_rate": 4.7810017850211585e-06, "loss": 2.3892, "step": 60710 }, { "epoch": 1.37, "grad_norm": 1.593338592261436, "learning_rate": 4.777887064744878e-06, "loss": 2.5005, "step": 60720 }, { "epoch": 1.37, "grad_norm": 1.6621553782962584, "learning_rate": 4.774773040896947e-06, "loss": 2.3665, "step": 60730 }, { "epoch": 1.37, "grad_norm": 1.5723605564269962, "learning_rate": 4.771659713892665e-06, "loss": 2.3676, "step": 60740 }, { "epoch": 1.37, "grad_norm": 2.0522292572818164, "learning_rate": 4.768547084147224e-06, "loss": 2.4782, "step": 60750 }, { "epoch": 1.37, "grad_norm": 1.6656189011282279, "learning_rate": 4.765435152075729e-06, "loss": 2.4996, "step": 60760 }, { "epoch": 1.37, "grad_norm": 1.4419123928382913, "learning_rate": 4.762323918093194e-06, "loss": 2.3981, "step": 60770 }, { "epoch": 1.37, "grad_norm": 1.5034348175340053, "learning_rate": 4.759213382614538e-06, "loss": 2.4057, "step": 60780 }, { "epoch": 1.37, "grad_norm": 1.976843090062695, "learning_rate": 4.756103546054577e-06, "loss": 2.2954, "step": 60790 }, { "epoch": 1.37, "grad_norm": 1.6013586672957727, "learning_rate": 4.752994408828054e-06, "loss": 2.2933, "step": 60800 }, { "epoch": 1.37, "grad_norm": 1.578409912679035, "learning_rate": 4.749885971349604e-06, "loss": 2.3208, "step": 60810 }, { "epoch": 1.37, "grad_norm": 1.7673065794315181, "learning_rate": 4.746778234033773e-06, "loss": 2.3379, "step": 60820 }, { "epoch": 1.37, "grad_norm": 1.619681044209025, "learning_rate": 4.7436711972950135e-06, "loss": 2.4059, "step": 60830 }, { "epoch": 1.37, "grad_norm": 1.519320572161911, "learning_rate": 4.7405648615476825e-06, "loss": 2.4039, "step": 60840 }, { "epoch": 1.37, "grad_norm": 1.6368417224930043, "learning_rate": 4.737459227206048e-06, "loss": 2.3614, "step": 60850 }, { "epoch": 1.37, "grad_norm": 1.4682906093904142, "learning_rate": 4.734354294684277e-06, "loss": 2.365, "step": 60860 }, { "epoch": 1.37, "grad_norm": 1.7172137188280578, "learning_rate": 4.7312500643964555e-06, "loss": 2.3965, "step": 60870 }, { "epoch": 1.37, "grad_norm": 1.6232821614777198, "learning_rate": 4.728146536756563e-06, "loss": 2.3928, "step": 60880 }, { "epoch": 1.37, "grad_norm": 1.7368903524610182, "learning_rate": 4.725043712178494e-06, "loss": 2.393, "step": 60890 }, { "epoch": 1.37, "grad_norm": 1.6188871941748133, "learning_rate": 4.721941591076039e-06, "loss": 2.3316, "step": 60900 }, { "epoch": 1.37, "grad_norm": 1.6019184305869258, "learning_rate": 4.718840173862905e-06, "loss": 2.4214, "step": 60910 }, { "epoch": 1.37, "grad_norm": 1.6453522794653885, "learning_rate": 4.715739460952699e-06, "loss": 2.3137, "step": 60920 }, { "epoch": 1.37, "grad_norm": 1.6659939120288203, "learning_rate": 4.712639452758941e-06, "loss": 2.33, "step": 60930 }, { "epoch": 1.37, "grad_norm": 1.506809976349989, "learning_rate": 4.70954014969505e-06, "loss": 2.4278, "step": 60940 }, { "epoch": 1.37, "grad_norm": 1.6051132831011623, "learning_rate": 4.7064415521743544e-06, "loss": 2.2813, "step": 60950 }, { "epoch": 1.37, "grad_norm": 1.6182054702649578, "learning_rate": 4.7033436606100856e-06, "loss": 2.397, "step": 60960 }, { "epoch": 1.37, "grad_norm": 1.4354738749611329, "learning_rate": 4.700246475415385e-06, "loss": 2.4892, "step": 60970 }, { "epoch": 1.38, "grad_norm": 1.8190277774558048, "learning_rate": 4.697149997003296e-06, "loss": 2.3534, "step": 60980 }, { "epoch": 1.38, "grad_norm": 1.6064267892831356, "learning_rate": 4.694054225786771e-06, "loss": 2.3734, "step": 60990 }, { "epoch": 1.38, "grad_norm": 1.6606232683554674, "learning_rate": 4.690959162178664e-06, "loss": 2.3806, "step": 61000 }, { "epoch": 1.38, "grad_norm": 1.7366893568578976, "learning_rate": 4.687864806591739e-06, "loss": 2.4177, "step": 61010 }, { "epoch": 1.38, "grad_norm": 1.461836230968088, "learning_rate": 4.684771159438663e-06, "loss": 2.4316, "step": 61020 }, { "epoch": 1.38, "grad_norm": 1.6128326265871022, "learning_rate": 4.681678221132008e-06, "loss": 2.4129, "step": 61030 }, { "epoch": 1.38, "grad_norm": 1.7866431194778727, "learning_rate": 4.678585992084255e-06, "loss": 2.5092, "step": 61040 }, { "epoch": 1.38, "grad_norm": 1.8970778195614704, "learning_rate": 4.675494472707786e-06, "loss": 2.3903, "step": 61050 }, { "epoch": 1.38, "grad_norm": 1.6618968035136328, "learning_rate": 4.6724036634148875e-06, "loss": 2.4573, "step": 61060 }, { "epoch": 1.38, "grad_norm": 1.5390995013315028, "learning_rate": 4.669313564617762e-06, "loss": 2.3834, "step": 61070 }, { "epoch": 1.38, "grad_norm": 1.8962944410339007, "learning_rate": 4.666224176728506e-06, "loss": 2.3997, "step": 61080 }, { "epoch": 1.38, "grad_norm": 1.5134679362140504, "learning_rate": 4.663135500159122e-06, "loss": 2.4657, "step": 61090 }, { "epoch": 1.38, "grad_norm": 1.7027177213674325, "learning_rate": 4.6600475353215265e-06, "loss": 2.4633, "step": 61100 }, { "epoch": 1.38, "grad_norm": 1.5804076547462762, "learning_rate": 4.656960282627527e-06, "loss": 2.5106, "step": 61110 }, { "epoch": 1.38, "grad_norm": 1.7818252800867256, "learning_rate": 4.653873742488845e-06, "loss": 2.4693, "step": 61120 }, { "epoch": 1.38, "grad_norm": 1.7855748389827732, "learning_rate": 4.6507879153171096e-06, "loss": 2.4382, "step": 61130 }, { "epoch": 1.38, "grad_norm": 1.7647173715297286, "learning_rate": 4.647702801523852e-06, "loss": 2.4059, "step": 61140 }, { "epoch": 1.38, "grad_norm": 1.7403529962717743, "learning_rate": 4.644618401520503e-06, "loss": 2.5179, "step": 61150 }, { "epoch": 1.38, "grad_norm": 1.6238351421726192, "learning_rate": 4.641534715718406e-06, "loss": 2.5337, "step": 61160 }, { "epoch": 1.38, "grad_norm": 1.87637802811231, "learning_rate": 4.6384517445288035e-06, "loss": 2.4423, "step": 61170 }, { "epoch": 1.38, "grad_norm": 1.7314309726019312, "learning_rate": 4.635369488362846e-06, "loss": 2.5402, "step": 61180 }, { "epoch": 1.38, "grad_norm": 1.5748106779949886, "learning_rate": 4.632287947631588e-06, "loss": 2.3965, "step": 61190 }, { "epoch": 1.38, "grad_norm": 1.6406402527776276, "learning_rate": 4.629207122745983e-06, "loss": 2.2833, "step": 61200 }, { "epoch": 1.38, "grad_norm": 1.6444813243417815, "learning_rate": 4.626127014116905e-06, "loss": 2.4746, "step": 61210 }, { "epoch": 1.38, "grad_norm": 1.6863872632222703, "learning_rate": 4.623047622155113e-06, "loss": 2.3577, "step": 61220 }, { "epoch": 1.38, "grad_norm": 1.6794993225828534, "learning_rate": 4.619968947271279e-06, "loss": 2.4246, "step": 61230 }, { "epoch": 1.38, "grad_norm": 1.6053996448751293, "learning_rate": 4.616890989875984e-06, "loss": 2.4524, "step": 61240 }, { "epoch": 1.38, "grad_norm": 1.5944191440488724, "learning_rate": 4.613813750379705e-06, "loss": 2.4315, "step": 61250 }, { "epoch": 1.38, "grad_norm": 1.502268617446515, "learning_rate": 4.610737229192825e-06, "loss": 2.3957, "step": 61260 }, { "epoch": 1.38, "grad_norm": 1.7430719303820479, "learning_rate": 4.60766142672564e-06, "loss": 2.4938, "step": 61270 }, { "epoch": 1.38, "grad_norm": 1.6937522972658419, "learning_rate": 4.604586343388341e-06, "loss": 2.3805, "step": 61280 }, { "epoch": 1.38, "grad_norm": 1.9041300231441207, "learning_rate": 4.6015119795910234e-06, "loss": 2.2283, "step": 61290 }, { "epoch": 1.38, "grad_norm": 1.8799703632004228, "learning_rate": 4.59843833574369e-06, "loss": 2.5093, "step": 61300 }, { "epoch": 1.38, "grad_norm": 1.440299664943333, "learning_rate": 4.595365412256245e-06, "loss": 2.4278, "step": 61310 }, { "epoch": 1.38, "grad_norm": 1.7046723300581068, "learning_rate": 4.5922932095385e-06, "loss": 2.3939, "step": 61320 }, { "epoch": 1.38, "grad_norm": 1.6254221024589164, "learning_rate": 4.589221728000167e-06, "loss": 2.4302, "step": 61330 }, { "epoch": 1.38, "grad_norm": 1.9851495468575209, "learning_rate": 4.586150968050862e-06, "loss": 2.3557, "step": 61340 }, { "epoch": 1.38, "grad_norm": 1.5343069292217872, "learning_rate": 4.5830809301001086e-06, "loss": 2.4145, "step": 61350 }, { "epoch": 1.38, "grad_norm": 1.7662019323336637, "learning_rate": 4.580011614557329e-06, "loss": 2.3781, "step": 61360 }, { "epoch": 1.38, "grad_norm": 1.5451033706499013, "learning_rate": 4.576943021831852e-06, "loss": 2.4193, "step": 61370 }, { "epoch": 1.38, "grad_norm": 1.991974002801458, "learning_rate": 4.57387515233291e-06, "loss": 2.3705, "step": 61380 }, { "epoch": 1.38, "grad_norm": 2.0333269890903063, "learning_rate": 4.570808006469636e-06, "loss": 2.3628, "step": 61390 }, { "epoch": 1.38, "grad_norm": 1.628658824926218, "learning_rate": 4.567741584651073e-06, "loss": 2.3403, "step": 61400 }, { "epoch": 1.38, "grad_norm": 1.7297557575787057, "learning_rate": 4.564675887286162e-06, "loss": 2.4244, "step": 61410 }, { "epoch": 1.39, "grad_norm": 1.7143555526609633, "learning_rate": 4.561610914783749e-06, "loss": 2.4768, "step": 61420 }, { "epoch": 1.39, "grad_norm": 1.518113501312012, "learning_rate": 4.558546667552584e-06, "loss": 2.3826, "step": 61430 }, { "epoch": 1.39, "grad_norm": 1.6357976338441178, "learning_rate": 4.555483146001315e-06, "loss": 2.4921, "step": 61440 }, { "epoch": 1.39, "grad_norm": 1.7800003349295728, "learning_rate": 4.552420350538495e-06, "loss": 2.4525, "step": 61450 }, { "epoch": 1.39, "grad_norm": 1.5438060169131358, "learning_rate": 4.549358281572593e-06, "loss": 2.4065, "step": 61460 }, { "epoch": 1.39, "grad_norm": 1.498640629498924, "learning_rate": 4.546296939511966e-06, "loss": 2.4994, "step": 61470 }, { "epoch": 1.39, "grad_norm": 1.4387797419765613, "learning_rate": 4.543236324764878e-06, "loss": 2.5997, "step": 61480 }, { "epoch": 1.39, "grad_norm": 1.6826326171818229, "learning_rate": 4.540176437739497e-06, "loss": 2.5585, "step": 61490 }, { "epoch": 1.39, "grad_norm": 1.5492032488364857, "learning_rate": 4.537117278843896e-06, "loss": 2.4929, "step": 61500 }, { "epoch": 1.39, "grad_norm": 1.5982110847237228, "learning_rate": 4.534058848486046e-06, "loss": 2.4837, "step": 61510 }, { "epoch": 1.39, "grad_norm": 1.6360719241747437, "learning_rate": 4.531001147073824e-06, "loss": 2.2958, "step": 61520 }, { "epoch": 1.39, "grad_norm": 1.5275283655447471, "learning_rate": 4.527944175015007e-06, "loss": 2.5302, "step": 61530 }, { "epoch": 1.39, "grad_norm": 1.991702884009907, "learning_rate": 4.524887932717288e-06, "loss": 2.4928, "step": 61540 }, { "epoch": 1.39, "grad_norm": 1.7786169192525934, "learning_rate": 4.521832420588239e-06, "loss": 2.3783, "step": 61550 }, { "epoch": 1.39, "grad_norm": 1.802616550832061, "learning_rate": 4.518777639035353e-06, "loss": 2.4022, "step": 61560 }, { "epoch": 1.39, "grad_norm": 1.5095599047253725, "learning_rate": 4.51572358846602e-06, "loss": 2.4173, "step": 61570 }, { "epoch": 1.39, "grad_norm": 1.9469668642569402, "learning_rate": 4.512670269287531e-06, "loss": 2.2915, "step": 61580 }, { "epoch": 1.39, "grad_norm": 1.9133186901071884, "learning_rate": 4.509617681907078e-06, "loss": 2.4498, "step": 61590 }, { "epoch": 1.39, "grad_norm": 1.5054992774608646, "learning_rate": 4.506565826731767e-06, "loss": 2.4722, "step": 61600 }, { "epoch": 1.39, "grad_norm": 1.6210944154424245, "learning_rate": 4.503514704168592e-06, "loss": 2.4367, "step": 61610 }, { "epoch": 1.39, "grad_norm": 1.7697074014181484, "learning_rate": 4.500464314624455e-06, "loss": 2.3836, "step": 61620 }, { "epoch": 1.39, "grad_norm": 1.505350478693621, "learning_rate": 4.497414658506163e-06, "loss": 2.4814, "step": 61630 }, { "epoch": 1.39, "grad_norm": 1.7877428508398725, "learning_rate": 4.494365736220423e-06, "loss": 2.3378, "step": 61640 }, { "epoch": 1.39, "grad_norm": 1.7044998395034718, "learning_rate": 4.491317548173835e-06, "loss": 2.4197, "step": 61650 }, { "epoch": 1.39, "grad_norm": 2.27101397457591, "learning_rate": 4.4882700947729184e-06, "loss": 2.559, "step": 61660 }, { "epoch": 1.39, "grad_norm": 1.4865850778709893, "learning_rate": 4.485223376424083e-06, "loss": 2.3894, "step": 61670 }, { "epoch": 1.39, "grad_norm": 1.7717635426690566, "learning_rate": 4.482177393533644e-06, "loss": 2.3708, "step": 61680 }, { "epoch": 1.39, "grad_norm": 1.6886500293206788, "learning_rate": 4.479132146507819e-06, "loss": 2.3451, "step": 61690 }, { "epoch": 1.39, "grad_norm": 1.6607927371320121, "learning_rate": 4.476087635752724e-06, "loss": 2.3598, "step": 61700 }, { "epoch": 1.39, "grad_norm": 1.5624684071674593, "learning_rate": 4.473043861674381e-06, "loss": 2.5323, "step": 61710 }, { "epoch": 1.39, "grad_norm": 1.4384638014359927, "learning_rate": 4.4700008246787106e-06, "loss": 2.3327, "step": 61720 }, { "epoch": 1.39, "grad_norm": 1.6921638396556524, "learning_rate": 4.466958525171534e-06, "loss": 2.4073, "step": 61730 }, { "epoch": 1.39, "grad_norm": 1.5499008674002241, "learning_rate": 4.463916963558585e-06, "loss": 2.3481, "step": 61740 }, { "epoch": 1.39, "grad_norm": 1.5798223976574397, "learning_rate": 4.460876140245487e-06, "loss": 2.3978, "step": 61750 }, { "epoch": 1.39, "grad_norm": 1.8645666315509664, "learning_rate": 4.457836055637764e-06, "loss": 2.3198, "step": 61760 }, { "epoch": 1.39, "grad_norm": 1.4695195161267236, "learning_rate": 4.45479671014085e-06, "loss": 2.3665, "step": 61770 }, { "epoch": 1.39, "grad_norm": 1.95756147101199, "learning_rate": 4.451758104160075e-06, "loss": 2.3296, "step": 61780 }, { "epoch": 1.39, "grad_norm": 1.6700526076119837, "learning_rate": 4.4487202381006675e-06, "loss": 2.3604, "step": 61790 }, { "epoch": 1.39, "grad_norm": 1.6801088704424045, "learning_rate": 4.44568311236777e-06, "loss": 2.4711, "step": 61800 }, { "epoch": 1.39, "grad_norm": 1.6632333480920536, "learning_rate": 4.442646727366416e-06, "loss": 2.4683, "step": 61810 }, { "epoch": 1.39, "grad_norm": 1.9800799966293245, "learning_rate": 4.439611083501537e-06, "loss": 2.4056, "step": 61820 }, { "epoch": 1.39, "grad_norm": 1.7696307966201976, "learning_rate": 4.436576181177976e-06, "loss": 2.2518, "step": 61830 }, { "epoch": 1.39, "grad_norm": 1.6718682847174442, "learning_rate": 4.43354202080047e-06, "loss": 2.435, "step": 61840 }, { "epoch": 1.39, "grad_norm": 1.7752977680874167, "learning_rate": 4.430508602773658e-06, "loss": 2.5188, "step": 61850 }, { "epoch": 1.4, "grad_norm": 1.7399741404382056, "learning_rate": 4.427475927502082e-06, "loss": 2.4011, "step": 61860 }, { "epoch": 1.4, "grad_norm": 1.68586732027061, "learning_rate": 4.424443995390183e-06, "loss": 2.366, "step": 61870 }, { "epoch": 1.4, "grad_norm": 1.6901448930126874, "learning_rate": 4.421412806842305e-06, "loss": 2.3811, "step": 61880 }, { "epoch": 1.4, "grad_norm": 1.5318483265811438, "learning_rate": 4.4183823622626896e-06, "loss": 2.3928, "step": 61890 }, { "epoch": 1.4, "grad_norm": 2.058455188337144, "learning_rate": 4.415352662055483e-06, "loss": 2.4665, "step": 61900 }, { "epoch": 1.4, "grad_norm": 1.6619404379138099, "learning_rate": 4.41232370662473e-06, "loss": 2.4527, "step": 61910 }, { "epoch": 1.4, "grad_norm": 1.7854822481687664, "learning_rate": 4.4092954963743725e-06, "loss": 2.4482, "step": 61920 }, { "epoch": 1.4, "grad_norm": 1.6507800063934255, "learning_rate": 4.406268031708264e-06, "loss": 2.3959, "step": 61930 }, { "epoch": 1.4, "grad_norm": 1.6240178935783687, "learning_rate": 4.4032413130301475e-06, "loss": 2.4406, "step": 61940 }, { "epoch": 1.4, "grad_norm": 1.669478877607678, "learning_rate": 4.400215340743671e-06, "loss": 2.3883, "step": 61950 }, { "epoch": 1.4, "grad_norm": 1.5555930208210909, "learning_rate": 4.3971901152523835e-06, "loss": 2.4936, "step": 61960 }, { "epoch": 1.4, "grad_norm": 1.7324707083279973, "learning_rate": 4.394165636959734e-06, "loss": 2.5473, "step": 61970 }, { "epoch": 1.4, "grad_norm": 1.5270518425537642, "learning_rate": 4.391141906269063e-06, "loss": 2.3906, "step": 61980 }, { "epoch": 1.4, "grad_norm": 1.5778812795893182, "learning_rate": 4.38811892358363e-06, "loss": 2.4496, "step": 61990 }, { "epoch": 1.4, "grad_norm": 1.7276798156358812, "learning_rate": 4.385096689306578e-06, "loss": 2.4718, "step": 62000 }, { "epoch": 1.4, "grad_norm": 1.5739180303154607, "learning_rate": 4.3820752038409604e-06, "loss": 2.4739, "step": 62010 }, { "epoch": 1.4, "grad_norm": 1.8646429911716316, "learning_rate": 4.379054467589725e-06, "loss": 2.3974, "step": 62020 }, { "epoch": 1.4, "grad_norm": 1.4877634574493321, "learning_rate": 4.37603448095572e-06, "loss": 2.5423, "step": 62030 }, { "epoch": 1.4, "grad_norm": 1.6205999921089833, "learning_rate": 4.3730152443416975e-06, "loss": 2.3804, "step": 62040 }, { "epoch": 1.4, "grad_norm": 1.5346875744419815, "learning_rate": 4.369996758150306e-06, "loss": 2.3275, "step": 62050 }, { "epoch": 1.4, "grad_norm": 1.7939180188537356, "learning_rate": 4.366979022784091e-06, "loss": 2.4611, "step": 62060 }, { "epoch": 1.4, "grad_norm": 1.5378321386497016, "learning_rate": 4.363962038645514e-06, "loss": 2.4192, "step": 62070 }, { "epoch": 1.4, "grad_norm": 1.756344931222169, "learning_rate": 4.360945806136912e-06, "loss": 2.3957, "step": 62080 }, { "epoch": 1.4, "grad_norm": 1.8222648173360507, "learning_rate": 4.357930325660538e-06, "loss": 2.2769, "step": 62090 }, { "epoch": 1.4, "grad_norm": 1.9323423959370902, "learning_rate": 4.3549155976185434e-06, "loss": 2.3209, "step": 62100 }, { "epoch": 1.4, "grad_norm": 1.7708789079955947, "learning_rate": 4.351901622412972e-06, "loss": 2.3483, "step": 62110 }, { "epoch": 1.4, "grad_norm": 1.5536769871920697, "learning_rate": 4.348888400445772e-06, "loss": 2.4573, "step": 62120 }, { "epoch": 1.4, "grad_norm": 1.8997919559456213, "learning_rate": 4.345875932118796e-06, "loss": 2.4664, "step": 62130 }, { "epoch": 1.4, "grad_norm": 1.710001618454305, "learning_rate": 4.3428642178337875e-06, "loss": 2.5655, "step": 62140 }, { "epoch": 1.4, "grad_norm": 1.6528853859859751, "learning_rate": 4.339853257992395e-06, "loss": 2.4642, "step": 62150 }, { "epoch": 1.4, "grad_norm": 1.52628820016521, "learning_rate": 4.336843052996161e-06, "loss": 2.4838, "step": 62160 }, { "epoch": 1.4, "grad_norm": 1.6809134652856148, "learning_rate": 4.333833603246533e-06, "loss": 2.4298, "step": 62170 }, { "epoch": 1.4, "grad_norm": 1.7722414523014334, "learning_rate": 4.330824909144856e-06, "loss": 2.3587, "step": 62180 }, { "epoch": 1.4, "grad_norm": 1.6859978241221047, "learning_rate": 4.3278169710923725e-06, "loss": 2.3589, "step": 62190 }, { "epoch": 1.4, "grad_norm": 1.9096672969661619, "learning_rate": 4.324809789490225e-06, "loss": 2.4233, "step": 62200 }, { "epoch": 1.4, "grad_norm": 1.6639064108613313, "learning_rate": 4.321803364739458e-06, "loss": 2.4462, "step": 62210 }, { "epoch": 1.4, "grad_norm": 1.7834890685099005, "learning_rate": 4.3187976972410085e-06, "loss": 2.4102, "step": 62220 }, { "epoch": 1.4, "grad_norm": 1.519040155653183, "learning_rate": 4.31579278739572e-06, "loss": 2.3702, "step": 62230 }, { "epoch": 1.4, "grad_norm": 1.5281472960551667, "learning_rate": 4.312788635604329e-06, "loss": 2.3886, "step": 62240 }, { "epoch": 1.4, "grad_norm": 1.7428518338740149, "learning_rate": 4.3097852422674755e-06, "loss": 2.469, "step": 62250 }, { "epoch": 1.4, "grad_norm": 1.5909777353623011, "learning_rate": 4.306782607785691e-06, "loss": 2.4664, "step": 62260 }, { "epoch": 1.4, "grad_norm": 1.685537671103816, "learning_rate": 4.3037807325594206e-06, "loss": 2.4151, "step": 62270 }, { "epoch": 1.4, "grad_norm": 1.753101073306776, "learning_rate": 4.3007796169889924e-06, "loss": 2.3677, "step": 62280 }, { "epoch": 1.4, "grad_norm": 1.5068236760810594, "learning_rate": 4.297779261474644e-06, "loss": 2.3239, "step": 62290 }, { "epoch": 1.4, "grad_norm": 1.784189446390729, "learning_rate": 4.294779666416501e-06, "loss": 2.3729, "step": 62300 }, { "epoch": 1.41, "grad_norm": 1.9239573347264323, "learning_rate": 4.291780832214595e-06, "loss": 2.2959, "step": 62310 }, { "epoch": 1.41, "grad_norm": 1.8087815593142857, "learning_rate": 4.288782759268853e-06, "loss": 2.3956, "step": 62320 }, { "epoch": 1.41, "grad_norm": 1.6256446333413062, "learning_rate": 4.285785447979108e-06, "loss": 2.3038, "step": 62330 }, { "epoch": 1.41, "grad_norm": 1.6903261489100696, "learning_rate": 4.282788898745084e-06, "loss": 2.3832, "step": 62340 }, { "epoch": 1.41, "grad_norm": 1.6402917863580866, "learning_rate": 4.279793111966404e-06, "loss": 2.4858, "step": 62350 }, { "epoch": 1.41, "grad_norm": 1.6207639860965786, "learning_rate": 4.27679808804259e-06, "loss": 2.3786, "step": 62360 }, { "epoch": 1.41, "grad_norm": 1.6172413489947453, "learning_rate": 4.273803827373062e-06, "loss": 2.284, "step": 62370 }, { "epoch": 1.41, "grad_norm": 1.674030865116004, "learning_rate": 4.27081033035714e-06, "loss": 2.4772, "step": 62380 }, { "epoch": 1.41, "grad_norm": 1.5475358694079602, "learning_rate": 4.267817597394037e-06, "loss": 2.3225, "step": 62390 }, { "epoch": 1.41, "grad_norm": 1.7669037856776253, "learning_rate": 4.264825628882881e-06, "loss": 2.3617, "step": 62400 }, { "epoch": 1.41, "grad_norm": 1.6801985929557637, "learning_rate": 4.261834425222669e-06, "loss": 2.367, "step": 62410 }, { "epoch": 1.41, "grad_norm": 1.5887723690955644, "learning_rate": 4.258843986812321e-06, "loss": 2.3633, "step": 62420 }, { "epoch": 1.41, "grad_norm": 1.7178114184909827, "learning_rate": 4.255854314050644e-06, "loss": 2.3187, "step": 62430 }, { "epoch": 1.41, "grad_norm": 1.587312916899196, "learning_rate": 4.252865407336345e-06, "loss": 2.4728, "step": 62440 }, { "epoch": 1.41, "grad_norm": 1.7561770003172876, "learning_rate": 4.2498772670680265e-06, "loss": 2.3807, "step": 62450 }, { "epoch": 1.41, "grad_norm": 1.559961023937885, "learning_rate": 4.246889893644196e-06, "loss": 2.4033, "step": 62460 }, { "epoch": 1.41, "grad_norm": 1.7585155507467014, "learning_rate": 4.2439032874632515e-06, "loss": 2.4763, "step": 62470 }, { "epoch": 1.41, "grad_norm": 1.8317844694981777, "learning_rate": 4.2409174489234905e-06, "loss": 2.3649, "step": 62480 }, { "epoch": 1.41, "grad_norm": 1.684864131185781, "learning_rate": 4.23793237842311e-06, "loss": 2.3947, "step": 62490 }, { "epoch": 1.41, "grad_norm": 1.9479175332775895, "learning_rate": 4.234948076360201e-06, "loss": 2.4763, "step": 62500 }, { "epoch": 1.41, "grad_norm": 1.6434330156378414, "learning_rate": 4.231964543132761e-06, "loss": 2.4505, "step": 62510 }, { "epoch": 1.41, "grad_norm": 2.2822816600562095, "learning_rate": 4.228981779138664e-06, "loss": 2.315, "step": 62520 }, { "epoch": 1.41, "grad_norm": 1.766821232524001, "learning_rate": 4.2259997847757085e-06, "loss": 2.4221, "step": 62530 }, { "epoch": 1.41, "grad_norm": 1.6770038501073943, "learning_rate": 4.223018560441573e-06, "loss": 2.4037, "step": 62540 }, { "epoch": 1.41, "grad_norm": 1.5198624322417305, "learning_rate": 4.220038106533839e-06, "loss": 2.3576, "step": 62550 }, { "epoch": 1.41, "grad_norm": 1.4985830486013738, "learning_rate": 4.217058423449982e-06, "loss": 2.4154, "step": 62560 }, { "epoch": 1.41, "grad_norm": 1.631070762879451, "learning_rate": 4.214079511587379e-06, "loss": 2.4407, "step": 62570 }, { "epoch": 1.41, "grad_norm": 1.5144839573754825, "learning_rate": 4.2111013713433e-06, "loss": 2.3949, "step": 62580 }, { "epoch": 1.41, "grad_norm": 1.6602256994531996, "learning_rate": 4.2081240031149105e-06, "loss": 2.2486, "step": 62590 }, { "epoch": 1.41, "grad_norm": 1.6351351589977376, "learning_rate": 4.205147407299286e-06, "loss": 2.4431, "step": 62600 }, { "epoch": 1.41, "grad_norm": 1.4990507132342028, "learning_rate": 4.202171584293387e-06, "loss": 2.4008, "step": 62610 }, { "epoch": 1.41, "grad_norm": 1.7606213208984214, "learning_rate": 4.1991965344940674e-06, "loss": 2.2537, "step": 62620 }, { "epoch": 1.41, "grad_norm": 1.537270150947556, "learning_rate": 4.196222258298087e-06, "loss": 2.5505, "step": 62630 }, { "epoch": 1.41, "grad_norm": 1.906290347440603, "learning_rate": 4.193248756102101e-06, "loss": 2.331, "step": 62640 }, { "epoch": 1.41, "grad_norm": 1.7149297193372786, "learning_rate": 4.190276028302654e-06, "loss": 2.4069, "step": 62650 }, { "epoch": 1.41, "grad_norm": 1.6493335989400935, "learning_rate": 4.1873040752962025e-06, "loss": 2.3471, "step": 62660 }, { "epoch": 1.41, "grad_norm": 1.6579583390167196, "learning_rate": 4.1843328974790845e-06, "loss": 2.3385, "step": 62670 }, { "epoch": 1.41, "grad_norm": 1.6258840143290068, "learning_rate": 4.181362495247543e-06, "loss": 2.335, "step": 62680 }, { "epoch": 1.41, "grad_norm": 1.866587962040628, "learning_rate": 4.178392868997713e-06, "loss": 2.5071, "step": 62690 }, { "epoch": 1.41, "grad_norm": 1.6975022596189013, "learning_rate": 4.175424019125629e-06, "loss": 2.4328, "step": 62700 }, { "epoch": 1.41, "grad_norm": 1.4758802478562056, "learning_rate": 4.172455946027221e-06, "loss": 2.3578, "step": 62710 }, { "epoch": 1.41, "grad_norm": 1.7540379402420894, "learning_rate": 4.169488650098315e-06, "loss": 2.289, "step": 62720 }, { "epoch": 1.41, "grad_norm": 1.4733551783548173, "learning_rate": 4.166522131734634e-06, "loss": 2.4403, "step": 62730 }, { "epoch": 1.41, "grad_norm": 1.7725359981060522, "learning_rate": 4.1635563913317965e-06, "loss": 2.3838, "step": 62740 }, { "epoch": 1.42, "grad_norm": 1.9338766198140778, "learning_rate": 4.1605914292853186e-06, "loss": 2.3239, "step": 62750 }, { "epoch": 1.42, "grad_norm": 1.953954972782738, "learning_rate": 4.157627245990613e-06, "loss": 2.5502, "step": 62760 }, { "epoch": 1.42, "grad_norm": 1.5114120664542225, "learning_rate": 4.154663841842985e-06, "loss": 2.3758, "step": 62770 }, { "epoch": 1.42, "grad_norm": 1.4743751065904314, "learning_rate": 4.151701217237639e-06, "loss": 2.4258, "step": 62780 }, { "epoch": 1.42, "grad_norm": 1.6218014920877142, "learning_rate": 4.148739372569673e-06, "loss": 2.3751, "step": 62790 }, { "epoch": 1.42, "grad_norm": 1.5477485481265592, "learning_rate": 4.1457783082340876e-06, "loss": 2.3578, "step": 62800 }, { "epoch": 1.42, "grad_norm": 1.714677166051185, "learning_rate": 4.142818024625772e-06, "loss": 2.4166, "step": 62810 }, { "epoch": 1.42, "grad_norm": 1.577221512161357, "learning_rate": 4.139858522139514e-06, "loss": 2.4537, "step": 62820 }, { "epoch": 1.42, "grad_norm": 1.7258391613108548, "learning_rate": 4.136899801169999e-06, "loss": 2.4111, "step": 62830 }, { "epoch": 1.42, "grad_norm": 1.8733308449400587, "learning_rate": 4.133941862111802e-06, "loss": 2.3288, "step": 62840 }, { "epoch": 1.42, "grad_norm": 1.8771610661448466, "learning_rate": 4.130984705359395e-06, "loss": 2.3987, "step": 62850 }, { "epoch": 1.42, "grad_norm": 1.5334925779069037, "learning_rate": 4.1280283313071575e-06, "loss": 2.5094, "step": 62860 }, { "epoch": 1.42, "grad_norm": 1.6250249337195326, "learning_rate": 4.1250727403493514e-06, "loss": 2.4075, "step": 62870 }, { "epoch": 1.42, "grad_norm": 1.744441619411346, "learning_rate": 4.122117932880139e-06, "loss": 2.3756, "step": 62880 }, { "epoch": 1.42, "grad_norm": 1.7784788539194116, "learning_rate": 4.119163909293579e-06, "loss": 2.3443, "step": 62890 }, { "epoch": 1.42, "grad_norm": 1.9943730034644398, "learning_rate": 4.1162106699836205e-06, "loss": 2.3291, "step": 62900 }, { "epoch": 1.42, "grad_norm": 1.5836613160986441, "learning_rate": 4.113258215344114e-06, "loss": 2.4129, "step": 62910 }, { "epoch": 1.42, "grad_norm": 1.6352128512308088, "learning_rate": 4.1103065457688e-06, "loss": 2.471, "step": 62920 }, { "epoch": 1.42, "grad_norm": 1.7493354017065013, "learning_rate": 4.1073556616513224e-06, "loss": 2.4183, "step": 62930 }, { "epoch": 1.42, "grad_norm": 1.5570358556507817, "learning_rate": 4.104405563385216e-06, "loss": 2.4236, "step": 62940 }, { "epoch": 1.42, "grad_norm": 1.7313619854969764, "learning_rate": 4.101456251363904e-06, "loss": 2.4733, "step": 62950 }, { "epoch": 1.42, "grad_norm": 1.7137586981528266, "learning_rate": 4.098507725980715e-06, "loss": 2.454, "step": 62960 }, { "epoch": 1.42, "grad_norm": 1.600125932189484, "learning_rate": 4.095559987628867e-06, "loss": 2.3225, "step": 62970 }, { "epoch": 1.42, "grad_norm": 1.7151111118126705, "learning_rate": 4.092613036701469e-06, "loss": 2.4367, "step": 62980 }, { "epoch": 1.42, "grad_norm": 1.6585945177294796, "learning_rate": 4.089666873591542e-06, "loss": 2.4061, "step": 62990 }, { "epoch": 1.42, "grad_norm": 1.8571610524435875, "learning_rate": 4.0867214986919845e-06, "loss": 2.5101, "step": 63000 }, { "epoch": 1.42, "grad_norm": 1.6372144248959297, "learning_rate": 4.0837769123955964e-06, "loss": 2.4938, "step": 63010 }, { "epoch": 1.42, "grad_norm": 1.7668213573600484, "learning_rate": 4.080833115095071e-06, "loss": 2.4501, "step": 63020 }, { "epoch": 1.42, "grad_norm": 1.4668437756510286, "learning_rate": 4.077890107182998e-06, "loss": 2.4005, "step": 63030 }, { "epoch": 1.42, "grad_norm": 1.649962754845559, "learning_rate": 4.074947889051861e-06, "loss": 2.3671, "step": 63040 }, { "epoch": 1.42, "grad_norm": 1.8116270811202047, "learning_rate": 4.072006461094039e-06, "loss": 2.4525, "step": 63050 }, { "epoch": 1.42, "grad_norm": 1.6682641324121674, "learning_rate": 4.069065823701804e-06, "loss": 2.3469, "step": 63060 }, { "epoch": 1.42, "grad_norm": 1.6020338977840622, "learning_rate": 4.066125977267323e-06, "loss": 2.4105, "step": 63070 }, { "epoch": 1.42, "grad_norm": 2.0976998033039482, "learning_rate": 4.06318692218266e-06, "loss": 2.4398, "step": 63080 }, { "epoch": 1.42, "grad_norm": 1.81368671558447, "learning_rate": 4.060248658839772e-06, "loss": 2.3894, "step": 63090 }, { "epoch": 1.42, "grad_norm": 1.5105986202611728, "learning_rate": 4.057311187630506e-06, "loss": 2.5144, "step": 63100 }, { "epoch": 1.42, "grad_norm": 1.6998412741919426, "learning_rate": 4.05437450894661e-06, "loss": 2.3445, "step": 63110 }, { "epoch": 1.42, "grad_norm": 1.6243655502897734, "learning_rate": 4.051438623179721e-06, "loss": 2.391, "step": 63120 }, { "epoch": 1.42, "grad_norm": 1.809872857176419, "learning_rate": 4.048503530721378e-06, "loss": 2.4742, "step": 63130 }, { "epoch": 1.42, "grad_norm": 1.8414259297348758, "learning_rate": 4.045569231963007e-06, "loss": 2.518, "step": 63140 }, { "epoch": 1.42, "grad_norm": 1.5073340058379139, "learning_rate": 4.042635727295933e-06, "loss": 2.3663, "step": 63150 }, { "epoch": 1.42, "grad_norm": 1.7162377132722875, "learning_rate": 4.039703017111365e-06, "loss": 2.2977, "step": 63160 }, { "epoch": 1.42, "grad_norm": 1.803738814387565, "learning_rate": 4.0367711018004175e-06, "loss": 2.4466, "step": 63170 }, { "epoch": 1.42, "grad_norm": 1.7565590613585935, "learning_rate": 4.033839981754091e-06, "loss": 2.3533, "step": 63180 }, { "epoch": 1.42, "grad_norm": 1.909561958127335, "learning_rate": 4.030909657363291e-06, "loss": 2.2529, "step": 63190 }, { "epoch": 1.43, "grad_norm": 1.6391873744844472, "learning_rate": 4.0279801290188054e-06, "loss": 2.346, "step": 63200 }, { "epoch": 1.43, "grad_norm": 1.8178032654308562, "learning_rate": 4.0250513971113226e-06, "loss": 2.3415, "step": 63210 }, { "epoch": 1.43, "grad_norm": 1.4054709113256814, "learning_rate": 4.02212346203142e-06, "loss": 2.389, "step": 63220 }, { "epoch": 1.43, "grad_norm": 1.7702541861979082, "learning_rate": 4.019196324169572e-06, "loss": 2.5039, "step": 63230 }, { "epoch": 1.43, "grad_norm": 2.0306016439177763, "learning_rate": 4.016269983916146e-06, "loss": 2.3482, "step": 63240 }, { "epoch": 1.43, "grad_norm": 1.65744503502901, "learning_rate": 4.0133444416614005e-06, "loss": 2.3499, "step": 63250 }, { "epoch": 1.43, "grad_norm": 1.930707114848362, "learning_rate": 4.010419697795495e-06, "loss": 2.4091, "step": 63260 }, { "epoch": 1.43, "grad_norm": 1.6012218288720348, "learning_rate": 4.007495752708474e-06, "loss": 2.3059, "step": 63270 }, { "epoch": 1.43, "grad_norm": 1.6469181737862906, "learning_rate": 4.004572606790278e-06, "loss": 2.4563, "step": 63280 }, { "epoch": 1.43, "grad_norm": 1.7918408897538345, "learning_rate": 4.001650260430745e-06, "loss": 2.4621, "step": 63290 }, { "epoch": 1.43, "grad_norm": 1.6270123258405071, "learning_rate": 3.9987287140196005e-06, "loss": 2.3482, "step": 63300 }, { "epoch": 1.43, "grad_norm": 1.7834824769628737, "learning_rate": 3.9958079679464675e-06, "loss": 2.3644, "step": 63310 }, { "epoch": 1.43, "grad_norm": 1.8612767801986518, "learning_rate": 3.992888022600857e-06, "loss": 2.3235, "step": 63320 }, { "epoch": 1.43, "grad_norm": 1.4954584078716655, "learning_rate": 3.9899688783721845e-06, "loss": 2.3179, "step": 63330 }, { "epoch": 1.43, "grad_norm": 1.448812108911915, "learning_rate": 3.987050535649747e-06, "loss": 2.438, "step": 63340 }, { "epoch": 1.43, "grad_norm": 1.4386135098263702, "learning_rate": 3.984132994822739e-06, "loss": 2.3307, "step": 63350 }, { "epoch": 1.43, "grad_norm": 1.544514951542136, "learning_rate": 3.981216256280247e-06, "loss": 2.4619, "step": 63360 }, { "epoch": 1.43, "grad_norm": 1.6815811834603553, "learning_rate": 3.9783003204112565e-06, "loss": 2.4051, "step": 63370 }, { "epoch": 1.43, "grad_norm": 1.6939685637703796, "learning_rate": 3.97538518760463e-06, "loss": 2.4224, "step": 63380 }, { "epoch": 1.43, "grad_norm": 2.4525874145628346, "learning_rate": 3.972470858249143e-06, "loss": 2.4053, "step": 63390 }, { "epoch": 1.43, "grad_norm": 1.7680163939597382, "learning_rate": 3.9695573327334525e-06, "loss": 2.3493, "step": 63400 }, { "epoch": 1.43, "grad_norm": 1.6189774809619932, "learning_rate": 3.96664461144611e-06, "loss": 2.4119, "step": 63410 }, { "epoch": 1.43, "grad_norm": 1.5491982989470745, "learning_rate": 3.9637326947755595e-06, "loss": 2.3961, "step": 63420 }, { "epoch": 1.43, "grad_norm": 1.6176283510827563, "learning_rate": 3.960821583110138e-06, "loss": 2.4088, "step": 63430 }, { "epoch": 1.43, "grad_norm": 1.5576402256775872, "learning_rate": 3.957911276838078e-06, "loss": 2.3436, "step": 63440 }, { "epoch": 1.43, "grad_norm": 1.6383782741438653, "learning_rate": 3.955001776347494e-06, "loss": 2.2691, "step": 63450 }, { "epoch": 1.43, "grad_norm": 1.6253470860369845, "learning_rate": 3.952093082026414e-06, "loss": 2.3904, "step": 63460 }, { "epoch": 1.43, "grad_norm": 1.668489167184667, "learning_rate": 3.949185194262737e-06, "loss": 2.4346, "step": 63470 }, { "epoch": 1.43, "grad_norm": 1.702011859506553, "learning_rate": 3.946278113444268e-06, "loss": 2.4798, "step": 63480 }, { "epoch": 1.43, "grad_norm": 1.6798466594765282, "learning_rate": 3.943371839958694e-06, "loss": 2.3354, "step": 63490 }, { "epoch": 1.43, "grad_norm": 1.4781941022717484, "learning_rate": 3.940466374193601e-06, "loss": 2.3469, "step": 63500 }, { "epoch": 1.43, "grad_norm": 1.8084742836887346, "learning_rate": 3.937561716536465e-06, "loss": 2.3486, "step": 63510 }, { "epoch": 1.43, "grad_norm": 1.7423991937088188, "learning_rate": 3.934657867374659e-06, "loss": 2.3848, "step": 63520 }, { "epoch": 1.43, "grad_norm": 1.698205640220657, "learning_rate": 3.9317548270954444e-06, "loss": 2.3961, "step": 63530 }, { "epoch": 1.43, "grad_norm": 1.5529197707920068, "learning_rate": 3.928852596085972e-06, "loss": 2.3549, "step": 63540 }, { "epoch": 1.43, "grad_norm": 1.4903669581354786, "learning_rate": 3.925951174733289e-06, "loss": 2.3942, "step": 63550 }, { "epoch": 1.43, "grad_norm": 1.6910326775225353, "learning_rate": 3.923050563424332e-06, "loss": 2.4571, "step": 63560 }, { "epoch": 1.43, "grad_norm": 1.8106381477111544, "learning_rate": 3.920150762545932e-06, "loss": 2.4303, "step": 63570 }, { "epoch": 1.43, "grad_norm": 1.7330011469532112, "learning_rate": 3.917251772484809e-06, "loss": 2.4735, "step": 63580 }, { "epoch": 1.43, "grad_norm": 1.8096411283753784, "learning_rate": 3.9143535936275776e-06, "loss": 2.4171, "step": 63590 }, { "epoch": 1.43, "grad_norm": 1.544977454962405, "learning_rate": 3.911456226360743e-06, "loss": 2.3272, "step": 63600 }, { "epoch": 1.43, "grad_norm": 1.6973011558322972, "learning_rate": 3.908559671070701e-06, "loss": 2.3864, "step": 63610 }, { "epoch": 1.43, "grad_norm": 1.6878938162121284, "learning_rate": 3.90566392814374e-06, "loss": 2.4953, "step": 63620 }, { "epoch": 1.43, "grad_norm": 1.6189284668700155, "learning_rate": 3.902768997966043e-06, "loss": 2.4569, "step": 63630 }, { "epoch": 1.44, "grad_norm": 1.638250346149444, "learning_rate": 3.899874880923681e-06, "loss": 2.5493, "step": 63640 }, { "epoch": 1.44, "grad_norm": 1.785565775576217, "learning_rate": 3.896981577402612e-06, "loss": 2.4281, "step": 63650 }, { "epoch": 1.44, "grad_norm": 1.795819916078236, "learning_rate": 3.894089087788699e-06, "loss": 2.4725, "step": 63660 }, { "epoch": 1.44, "grad_norm": 1.800398406430741, "learning_rate": 3.891197412467686e-06, "loss": 2.4021, "step": 63670 }, { "epoch": 1.44, "grad_norm": 1.6374540926494836, "learning_rate": 3.888306551825212e-06, "loss": 2.4374, "step": 63680 }, { "epoch": 1.44, "grad_norm": 1.7760894359120878, "learning_rate": 3.885416506246806e-06, "loss": 2.4019, "step": 63690 }, { "epoch": 1.44, "grad_norm": 1.717695394056626, "learning_rate": 3.882527276117885e-06, "loss": 2.3821, "step": 63700 }, { "epoch": 1.44, "grad_norm": 1.580795545476818, "learning_rate": 3.879638861823758e-06, "loss": 2.4109, "step": 63710 }, { "epoch": 1.44, "grad_norm": 1.6823508945507946, "learning_rate": 3.8767512637496375e-06, "loss": 2.4353, "step": 63720 }, { "epoch": 1.44, "grad_norm": 1.6321484381896292, "learning_rate": 3.8738644822806135e-06, "loss": 2.3727, "step": 63730 }, { "epoch": 1.44, "grad_norm": 1.7449407575586082, "learning_rate": 3.87097851780167e-06, "loss": 2.3158, "step": 63740 }, { "epoch": 1.44, "grad_norm": 1.7278149712598931, "learning_rate": 3.868093370697684e-06, "loss": 2.5533, "step": 63750 }, { "epoch": 1.44, "grad_norm": 1.6887217131655683, "learning_rate": 3.865209041353424e-06, "loss": 2.5107, "step": 63760 }, { "epoch": 1.44, "grad_norm": 1.6533284292107886, "learning_rate": 3.862325530153547e-06, "loss": 2.4026, "step": 63770 }, { "epoch": 1.44, "grad_norm": 1.6399932011417369, "learning_rate": 3.859442837482602e-06, "loss": 2.3541, "step": 63780 }, { "epoch": 1.44, "grad_norm": 1.7919508844698486, "learning_rate": 3.856560963725027e-06, "loss": 2.355, "step": 63790 }, { "epoch": 1.44, "grad_norm": 1.690019364021139, "learning_rate": 3.85367990926516e-06, "loss": 2.4819, "step": 63800 }, { "epoch": 1.44, "grad_norm": 1.747277943661644, "learning_rate": 3.850799674487216e-06, "loss": 2.5463, "step": 63810 }, { "epoch": 1.44, "grad_norm": 1.71837602006672, "learning_rate": 3.847920259775307e-06, "loss": 2.3842, "step": 63820 }, { "epoch": 1.44, "grad_norm": 1.6222775884478176, "learning_rate": 3.845041665513438e-06, "loss": 2.5135, "step": 63830 }, { "epoch": 1.44, "grad_norm": 1.8613237508259883, "learning_rate": 3.842163892085503e-06, "loss": 2.381, "step": 63840 }, { "epoch": 1.44, "grad_norm": 1.7483531165528734, "learning_rate": 3.839286939875281e-06, "loss": 2.4319, "step": 63850 }, { "epoch": 1.44, "grad_norm": 1.5791248968237588, "learning_rate": 3.836410809266453e-06, "loss": 2.331, "step": 63860 }, { "epoch": 1.44, "grad_norm": 1.6983021599340926, "learning_rate": 3.8335355006425815e-06, "loss": 2.5274, "step": 63870 }, { "epoch": 1.44, "grad_norm": 1.6874228644157068, "learning_rate": 3.830661014387122e-06, "loss": 2.3042, "step": 63880 }, { "epoch": 1.44, "grad_norm": 1.8414269247836672, "learning_rate": 3.827787350883417e-06, "loss": 2.4401, "step": 63890 }, { "epoch": 1.44, "grad_norm": 1.512896029605137, "learning_rate": 3.824914510514707e-06, "loss": 2.224, "step": 63900 }, { "epoch": 1.44, "grad_norm": 1.9888494575476712, "learning_rate": 3.822042493664115e-06, "loss": 2.4654, "step": 63910 }, { "epoch": 1.44, "grad_norm": 1.8765901037817203, "learning_rate": 3.819171300714658e-06, "loss": 2.3379, "step": 63920 }, { "epoch": 1.44, "grad_norm": 1.498331525072119, "learning_rate": 3.816300932049243e-06, "loss": 2.387, "step": 63930 }, { "epoch": 1.44, "grad_norm": 1.541538223406838, "learning_rate": 3.8134313880506646e-06, "loss": 2.5945, "step": 63940 }, { "epoch": 1.44, "grad_norm": 1.5894964813320367, "learning_rate": 3.8105626691016118e-06, "loss": 2.4062, "step": 63950 }, { "epoch": 1.44, "grad_norm": 1.757051225709816, "learning_rate": 3.8076947755846583e-06, "loss": 2.4651, "step": 63960 }, { "epoch": 1.44, "grad_norm": 1.6403869870551717, "learning_rate": 3.8048277078822714e-06, "loss": 2.4058, "step": 63970 }, { "epoch": 1.44, "grad_norm": 1.9125389766453134, "learning_rate": 3.8019614663768055e-06, "loss": 2.3966, "step": 63980 }, { "epoch": 1.44, "grad_norm": 1.4940209948090042, "learning_rate": 3.7990960514505114e-06, "loss": 2.2766, "step": 63990 }, { "epoch": 1.44, "grad_norm": 1.5491792396619402, "learning_rate": 3.7962314634855223e-06, "loss": 2.3676, "step": 64000 }, { "epoch": 1.44, "grad_norm": 1.7615255740330273, "learning_rate": 3.793367702863864e-06, "loss": 2.331, "step": 64010 }, { "epoch": 1.44, "grad_norm": 1.584672231074274, "learning_rate": 3.7905047699674537e-06, "loss": 2.4893, "step": 64020 }, { "epoch": 1.44, "grad_norm": 1.6569064927834862, "learning_rate": 3.787642665178091e-06, "loss": 2.4258, "step": 64030 }, { "epoch": 1.44, "grad_norm": 1.579947178374656, "learning_rate": 3.784781388877472e-06, "loss": 2.4446, "step": 64040 }, { "epoch": 1.44, "grad_norm": 1.8938272389994981, "learning_rate": 3.781920941447179e-06, "loss": 2.4375, "step": 64050 }, { "epoch": 1.44, "grad_norm": 1.7837577518270145, "learning_rate": 3.779061323268691e-06, "loss": 2.4008, "step": 64060 }, { "epoch": 1.44, "grad_norm": 1.5726138199453192, "learning_rate": 3.776202534723368e-06, "loss": 2.4706, "step": 64070 }, { "epoch": 1.45, "grad_norm": 1.5970252516446035, "learning_rate": 3.7733445761924614e-06, "loss": 2.4157, "step": 64080 }, { "epoch": 1.45, "grad_norm": 1.8385339560998175, "learning_rate": 3.770487448057112e-06, "loss": 2.4149, "step": 64090 }, { "epoch": 1.45, "grad_norm": 1.4645612701653548, "learning_rate": 3.7676311506983523e-06, "loss": 2.4153, "step": 64100 }, { "epoch": 1.45, "grad_norm": 1.619730624441822, "learning_rate": 3.7647756844971006e-06, "loss": 2.2913, "step": 64110 }, { "epoch": 1.45, "grad_norm": 1.7528218177490393, "learning_rate": 3.7619210498341665e-06, "loss": 2.2568, "step": 64120 }, { "epoch": 1.45, "grad_norm": 1.5168265918984596, "learning_rate": 3.759067247090248e-06, "loss": 2.585, "step": 64130 }, { "epoch": 1.45, "grad_norm": 1.706574146503106, "learning_rate": 3.756214276645932e-06, "loss": 2.374, "step": 64140 }, { "epoch": 1.45, "grad_norm": 1.5339140478884556, "learning_rate": 3.7533621388816944e-06, "loss": 2.4944, "step": 64150 }, { "epoch": 1.45, "grad_norm": 1.5015896908395203, "learning_rate": 3.7505108341779015e-06, "loss": 2.5903, "step": 64160 }, { "epoch": 1.45, "grad_norm": 1.6563851932118188, "learning_rate": 3.7476603629148067e-06, "loss": 2.3509, "step": 64170 }, { "epoch": 1.45, "grad_norm": 1.7360786473158323, "learning_rate": 3.7448107254725495e-06, "loss": 2.4518, "step": 64180 }, { "epoch": 1.45, "grad_norm": 1.643313974714535, "learning_rate": 3.7419619222311676e-06, "loss": 2.5335, "step": 64190 }, { "epoch": 1.45, "grad_norm": 1.686845970569432, "learning_rate": 3.73911395357058e-06, "loss": 2.4156, "step": 64200 }, { "epoch": 1.45, "grad_norm": 1.4617844821764012, "learning_rate": 3.7362668198705933e-06, "loss": 2.2828, "step": 64210 }, { "epoch": 1.45, "grad_norm": 1.6229550146355556, "learning_rate": 3.7334205215109077e-06, "loss": 2.4245, "step": 64220 }, { "epoch": 1.45, "grad_norm": 1.843524251974469, "learning_rate": 3.730575058871111e-06, "loss": 2.4638, "step": 64230 }, { "epoch": 1.45, "grad_norm": 1.5897202287529029, "learning_rate": 3.7277304323306683e-06, "loss": 2.47, "step": 64240 }, { "epoch": 1.45, "grad_norm": 1.602948764682336, "learning_rate": 3.7248866422689543e-06, "loss": 2.5502, "step": 64250 }, { "epoch": 1.45, "grad_norm": 1.5334043527231564, "learning_rate": 3.7220436890652165e-06, "loss": 2.5915, "step": 64260 }, { "epoch": 1.45, "grad_norm": 1.7087144300430204, "learning_rate": 3.7192015730985955e-06, "loss": 2.4777, "step": 64270 }, { "epoch": 1.45, "grad_norm": 1.6810229765886835, "learning_rate": 3.7163602947481193e-06, "loss": 2.4294, "step": 64280 }, { "epoch": 1.45, "grad_norm": 1.640764362405917, "learning_rate": 3.7135198543927064e-06, "loss": 2.3505, "step": 64290 }, { "epoch": 1.45, "grad_norm": 1.5293258434078403, "learning_rate": 3.71068025241116e-06, "loss": 2.4181, "step": 64300 }, { "epoch": 1.45, "grad_norm": 1.8056985902326603, "learning_rate": 3.7078414891821747e-06, "loss": 2.471, "step": 64310 }, { "epoch": 1.45, "grad_norm": 1.6362438500821237, "learning_rate": 3.7050035650843274e-06, "loss": 2.4214, "step": 64320 }, { "epoch": 1.45, "grad_norm": 1.5514482676096801, "learning_rate": 3.7021664804960965e-06, "loss": 2.238, "step": 64330 }, { "epoch": 1.45, "grad_norm": 1.5472994262365147, "learning_rate": 3.699330235795837e-06, "loss": 2.3636, "step": 64340 }, { "epoch": 1.45, "grad_norm": 1.5947037889710691, "learning_rate": 3.69649483136179e-06, "loss": 2.4097, "step": 64350 }, { "epoch": 1.45, "grad_norm": 1.3896842133333736, "learning_rate": 3.693660267572091e-06, "loss": 2.315, "step": 64360 }, { "epoch": 1.45, "grad_norm": 1.5689899523112303, "learning_rate": 3.690826544804762e-06, "loss": 2.4636, "step": 64370 }, { "epoch": 1.45, "grad_norm": 1.5402733544433642, "learning_rate": 3.687993663437709e-06, "loss": 2.4058, "step": 64380 }, { "epoch": 1.45, "grad_norm": 1.6052429064751468, "learning_rate": 3.685161623848735e-06, "loss": 2.4685, "step": 64390 }, { "epoch": 1.45, "grad_norm": 1.6581219852962477, "learning_rate": 3.6823304264155225e-06, "loss": 2.3478, "step": 64400 }, { "epoch": 1.45, "grad_norm": 1.7800403565309482, "learning_rate": 3.6795000715156428e-06, "loss": 2.4174, "step": 64410 }, { "epoch": 1.45, "grad_norm": 1.7795992106688667, "learning_rate": 3.676670559526557e-06, "loss": 2.3243, "step": 64420 }, { "epoch": 1.45, "grad_norm": 1.735855776755463, "learning_rate": 3.6738418908256136e-06, "loss": 2.4737, "step": 64430 }, { "epoch": 1.45, "grad_norm": 1.4308205184149643, "learning_rate": 3.671014065790045e-06, "loss": 2.4145, "step": 64440 }, { "epoch": 1.45, "grad_norm": 1.7153878282916382, "learning_rate": 3.668187084796977e-06, "loss": 2.3436, "step": 64450 }, { "epoch": 1.45, "grad_norm": 1.682426103335766, "learning_rate": 3.665360948223419e-06, "loss": 2.443, "step": 64460 }, { "epoch": 1.45, "grad_norm": 1.745873532746708, "learning_rate": 3.6625356564462668e-06, "loss": 2.5299, "step": 64470 }, { "epoch": 1.45, "grad_norm": 1.8228246731186581, "learning_rate": 3.659711209842307e-06, "loss": 2.3565, "step": 64480 }, { "epoch": 1.45, "grad_norm": 1.5963700808702994, "learning_rate": 3.656887608788211e-06, "loss": 2.4623, "step": 64490 }, { "epoch": 1.45, "grad_norm": 1.7016544653210726, "learning_rate": 3.6540648536605385e-06, "loss": 2.5212, "step": 64500 }, { "epoch": 1.45, "grad_norm": 1.6174113717666805, "learning_rate": 3.6512429448357333e-06, "loss": 2.4536, "step": 64510 }, { "epoch": 1.45, "grad_norm": 1.7700850478411656, "learning_rate": 3.6484218826901354e-06, "loss": 2.361, "step": 64520 }, { "epoch": 1.46, "grad_norm": 1.7409674109101296, "learning_rate": 3.6456016675999616e-06, "loss": 2.5055, "step": 64530 }, { "epoch": 1.46, "grad_norm": 1.5368763129439884, "learning_rate": 3.6427822999413208e-06, "loss": 2.2945, "step": 64540 }, { "epoch": 1.46, "grad_norm": 1.681266159662338, "learning_rate": 3.639963780090209e-06, "loss": 2.4146, "step": 64550 }, { "epoch": 1.46, "grad_norm": 1.61355327270692, "learning_rate": 3.637146108422508e-06, "loss": 2.474, "step": 64560 }, { "epoch": 1.46, "grad_norm": 1.711180024300798, "learning_rate": 3.6343292853139823e-06, "loss": 2.3452, "step": 64570 }, { "epoch": 1.46, "grad_norm": 1.432413653899481, "learning_rate": 3.6315133111402856e-06, "loss": 2.413, "step": 64580 }, { "epoch": 1.46, "grad_norm": 1.6506845622209259, "learning_rate": 3.628698186276969e-06, "loss": 2.3902, "step": 64590 }, { "epoch": 1.46, "grad_norm": 1.6642650536029477, "learning_rate": 3.625883911099457e-06, "loss": 2.34, "step": 64600 }, { "epoch": 1.46, "grad_norm": 1.5340961949116525, "learning_rate": 3.6230704859830657e-06, "loss": 2.5188, "step": 64610 }, { "epoch": 1.46, "grad_norm": 1.7085748704515296, "learning_rate": 3.6202579113029966e-06, "loss": 2.4167, "step": 64620 }, { "epoch": 1.46, "grad_norm": 1.5657311726897643, "learning_rate": 3.6174461874343392e-06, "loss": 2.3675, "step": 64630 }, { "epoch": 1.46, "grad_norm": 1.7344595800664024, "learning_rate": 3.6146353147520696e-06, "loss": 2.4118, "step": 64640 }, { "epoch": 1.46, "grad_norm": 1.72345644003134, "learning_rate": 3.611825293631045e-06, "loss": 2.4261, "step": 64650 }, { "epoch": 1.46, "grad_norm": 1.6654441011344725, "learning_rate": 3.6090161244460252e-06, "loss": 2.32, "step": 64660 }, { "epoch": 1.46, "grad_norm": 1.6740791750283162, "learning_rate": 3.6062078075716345e-06, "loss": 2.4441, "step": 64670 }, { "epoch": 1.46, "grad_norm": 1.8331024338585342, "learning_rate": 3.603400343382397e-06, "loss": 2.4111, "step": 64680 }, { "epoch": 1.46, "grad_norm": 1.541727511028615, "learning_rate": 3.6005937322527208e-06, "loss": 2.376, "step": 64690 }, { "epoch": 1.46, "grad_norm": 1.5175626527024797, "learning_rate": 3.597787974556899e-06, "loss": 2.3383, "step": 64700 }, { "epoch": 1.46, "grad_norm": 1.5297012567182298, "learning_rate": 3.594983070669107e-06, "loss": 2.3902, "step": 64710 }, { "epoch": 1.46, "grad_norm": 1.5127219743357982, "learning_rate": 3.592179020963419e-06, "loss": 2.4146, "step": 64720 }, { "epoch": 1.46, "grad_norm": 1.8743732587518118, "learning_rate": 3.5893758258137833e-06, "loss": 2.3659, "step": 64730 }, { "epoch": 1.46, "grad_norm": 1.7156642573386527, "learning_rate": 3.586573485594038e-06, "loss": 2.4449, "step": 64740 }, { "epoch": 1.46, "grad_norm": 1.6278023810731816, "learning_rate": 3.5837720006779075e-06, "loss": 2.3946, "step": 64750 }, { "epoch": 1.46, "grad_norm": 1.6925422310374154, "learning_rate": 3.5809713714390004e-06, "loss": 2.4414, "step": 64760 }, { "epoch": 1.46, "grad_norm": 1.6403064899342645, "learning_rate": 3.5781715982508136e-06, "loss": 2.4247, "step": 64770 }, { "epoch": 1.46, "grad_norm": 1.703251275087864, "learning_rate": 3.5753726814867297e-06, "loss": 2.4786, "step": 64780 }, { "epoch": 1.46, "grad_norm": 1.7843290715366866, "learning_rate": 3.572574621520014e-06, "loss": 2.4564, "step": 64790 }, { "epoch": 1.46, "grad_norm": 1.7483751921777806, "learning_rate": 3.569777418723821e-06, "loss": 2.3392, "step": 64800 }, { "epoch": 1.46, "grad_norm": 1.5470993910922486, "learning_rate": 3.5669810734711898e-06, "loss": 2.2481, "step": 64810 }, { "epoch": 1.46, "grad_norm": 1.5979772988443524, "learning_rate": 3.564185586135044e-06, "loss": 2.3056, "step": 64820 }, { "epoch": 1.46, "grad_norm": 1.8772111083415128, "learning_rate": 3.5613909570881956e-06, "loss": 2.5254, "step": 64830 }, { "epoch": 1.46, "grad_norm": 1.7741348684158331, "learning_rate": 3.5585971867033375e-06, "loss": 2.4957, "step": 64840 }, { "epoch": 1.46, "grad_norm": 1.5228216056640065, "learning_rate": 3.555804275353051e-06, "loss": 2.4056, "step": 64850 }, { "epoch": 1.46, "grad_norm": 1.761248564641559, "learning_rate": 3.5530122234098065e-06, "loss": 2.4867, "step": 64860 }, { "epoch": 1.46, "grad_norm": 1.57745776863294, "learning_rate": 3.550221031245953e-06, "loss": 2.472, "step": 64870 }, { "epoch": 1.46, "grad_norm": 1.6652825471801525, "learning_rate": 3.5474306992337326e-06, "loss": 2.3018, "step": 64880 }, { "epoch": 1.46, "grad_norm": 2.3321382867593488, "learning_rate": 3.5446412277452603e-06, "loss": 2.4357, "step": 64890 }, { "epoch": 1.46, "grad_norm": 1.6204297863814687, "learning_rate": 3.5418526171525467e-06, "loss": 2.3737, "step": 64900 }, { "epoch": 1.46, "grad_norm": 1.596566200868272, "learning_rate": 3.539064867827483e-06, "loss": 2.3716, "step": 64910 }, { "epoch": 1.46, "grad_norm": 1.6370525946328132, "learning_rate": 3.5362779801418523e-06, "loss": 2.4355, "step": 64920 }, { "epoch": 1.46, "grad_norm": 1.5639551785456607, "learning_rate": 3.5334919544673165e-06, "loss": 2.3157, "step": 64930 }, { "epoch": 1.46, "grad_norm": 1.8160305049411205, "learning_rate": 3.530706791175423e-06, "loss": 2.4058, "step": 64940 }, { "epoch": 1.46, "grad_norm": 1.7899586044690916, "learning_rate": 3.527922490637605e-06, "loss": 2.3922, "step": 64950 }, { "epoch": 1.46, "grad_norm": 1.5293288632534505, "learning_rate": 3.5251390532251808e-06, "loss": 2.4259, "step": 64960 }, { "epoch": 1.47, "grad_norm": 1.660140157020562, "learning_rate": 3.5223564793093522e-06, "loss": 2.5287, "step": 64970 }, { "epoch": 1.47, "grad_norm": 1.5549105449906053, "learning_rate": 3.519574769261207e-06, "loss": 2.5044, "step": 64980 }, { "epoch": 1.47, "grad_norm": 1.8514372480706092, "learning_rate": 3.5167939234517257e-06, "loss": 2.2749, "step": 64990 }, { "epoch": 1.47, "grad_norm": 1.8659616594136226, "learning_rate": 3.514013942251756e-06, "loss": 2.2731, "step": 65000 }, { "epoch": 1.47, "grad_norm": 1.8177510092567863, "learning_rate": 3.511234826032044e-06, "loss": 2.3536, "step": 65010 }, { "epoch": 1.47, "grad_norm": 1.6328481766004452, "learning_rate": 3.508456575163216e-06, "loss": 2.5036, "step": 65020 }, { "epoch": 1.47, "grad_norm": 1.6377717672682042, "learning_rate": 3.505679190015784e-06, "loss": 2.2763, "step": 65030 }, { "epoch": 1.47, "grad_norm": 1.747890778605316, "learning_rate": 3.50290267096014e-06, "loss": 2.3623, "step": 65040 }, { "epoch": 1.47, "grad_norm": 1.516469415255501, "learning_rate": 3.5001270183665714e-06, "loss": 2.4296, "step": 65050 }, { "epoch": 1.47, "grad_norm": 1.746794246866489, "learning_rate": 3.497352232605241e-06, "loss": 2.4474, "step": 65060 }, { "epoch": 1.47, "grad_norm": 1.7759469107058126, "learning_rate": 3.4945783140461966e-06, "loss": 2.3934, "step": 65070 }, { "epoch": 1.47, "grad_norm": 1.6579271355591152, "learning_rate": 3.4918052630593722e-06, "loss": 2.3008, "step": 65080 }, { "epoch": 1.47, "grad_norm": 1.7700278389313533, "learning_rate": 3.4890330800145864e-06, "loss": 2.3443, "step": 65090 }, { "epoch": 1.47, "grad_norm": 1.8428588242225754, "learning_rate": 3.4862617652815445e-06, "loss": 2.3703, "step": 65100 }, { "epoch": 1.47, "grad_norm": 1.6402165115772678, "learning_rate": 3.4834913192298235e-06, "loss": 2.4297, "step": 65110 }, { "epoch": 1.47, "grad_norm": 1.6054162759229382, "learning_rate": 3.480721742228902e-06, "loss": 2.3726, "step": 65120 }, { "epoch": 1.47, "grad_norm": 1.7644248117904044, "learning_rate": 3.477953034648134e-06, "loss": 2.3394, "step": 65130 }, { "epoch": 1.47, "grad_norm": 1.5900513166487076, "learning_rate": 3.4751851968567563e-06, "loss": 2.4083, "step": 65140 }, { "epoch": 1.47, "grad_norm": 1.706035374256281, "learning_rate": 3.4724182292238928e-06, "loss": 2.4314, "step": 65150 }, { "epoch": 1.47, "grad_norm": 1.6251461970356156, "learning_rate": 3.4696521321185495e-06, "loss": 2.4689, "step": 65160 }, { "epoch": 1.47, "grad_norm": 1.844465967001549, "learning_rate": 3.4668869059096164e-06, "loss": 2.3897, "step": 65170 }, { "epoch": 1.47, "grad_norm": 1.8009687969603525, "learning_rate": 3.4641225509658673e-06, "loss": 2.4963, "step": 65180 }, { "epoch": 1.47, "grad_norm": 1.870621941774295, "learning_rate": 3.4613590676559638e-06, "loss": 2.4122, "step": 65190 }, { "epoch": 1.47, "grad_norm": 1.4164691504740616, "learning_rate": 3.45859645634845e-06, "loss": 2.3669, "step": 65200 }, { "epoch": 1.47, "grad_norm": 1.8021679254788008, "learning_rate": 3.4558347174117445e-06, "loss": 2.308, "step": 65210 }, { "epoch": 1.47, "grad_norm": 1.7025383078546033, "learning_rate": 3.4530738512141606e-06, "loss": 2.4713, "step": 65220 }, { "epoch": 1.47, "grad_norm": 1.7673418196928403, "learning_rate": 3.4503138581238895e-06, "loss": 2.306, "step": 65230 }, { "epoch": 1.47, "grad_norm": 1.8033091281838858, "learning_rate": 3.447554738509007e-06, "loss": 2.3722, "step": 65240 }, { "epoch": 1.47, "grad_norm": 1.700891356094598, "learning_rate": 3.4447964927374777e-06, "loss": 2.4865, "step": 65250 }, { "epoch": 1.47, "grad_norm": 1.6894731570778538, "learning_rate": 3.442039121177143e-06, "loss": 2.4029, "step": 65260 }, { "epoch": 1.47, "grad_norm": 1.7770461354554334, "learning_rate": 3.439282624195729e-06, "loss": 2.5169, "step": 65270 }, { "epoch": 1.47, "grad_norm": 1.8409839451145509, "learning_rate": 3.4365270021608464e-06, "loss": 2.3811, "step": 65280 }, { "epoch": 1.47, "grad_norm": 1.4721591613669516, "learning_rate": 3.43377225543999e-06, "loss": 2.4718, "step": 65290 }, { "epoch": 1.47, "grad_norm": 1.6298117053567596, "learning_rate": 3.4310183844005352e-06, "loss": 2.3872, "step": 65300 }, { "epoch": 1.47, "grad_norm": 1.7384615727692838, "learning_rate": 3.4282653894097416e-06, "loss": 2.3795, "step": 65310 }, { "epoch": 1.47, "grad_norm": 1.771917333197368, "learning_rate": 3.425513270834755e-06, "loss": 2.3893, "step": 65320 }, { "epoch": 1.47, "grad_norm": 1.8097197470394593, "learning_rate": 3.4227620290426e-06, "loss": 2.3605, "step": 65330 }, { "epoch": 1.47, "grad_norm": 1.6738098751691404, "learning_rate": 3.4200116644001856e-06, "loss": 2.3205, "step": 65340 }, { "epoch": 1.47, "grad_norm": 1.6630585457397988, "learning_rate": 3.4172621772743055e-06, "loss": 2.4752, "step": 65350 }, { "epoch": 1.47, "grad_norm": 1.6633033997902986, "learning_rate": 3.4145135680316356e-06, "loss": 2.4133, "step": 65360 }, { "epoch": 1.47, "grad_norm": 1.7206862948962307, "learning_rate": 3.411765837038733e-06, "loss": 2.3922, "step": 65370 }, { "epoch": 1.47, "grad_norm": 1.8673045713494356, "learning_rate": 3.409018984662036e-06, "loss": 2.4, "step": 65380 }, { "epoch": 1.47, "grad_norm": 1.49887142168669, "learning_rate": 3.4062730112678754e-06, "loss": 2.355, "step": 65390 }, { "epoch": 1.47, "grad_norm": 1.5652961224543045, "learning_rate": 3.4035279172224555e-06, "loss": 2.4987, "step": 65400 }, { "epoch": 1.48, "grad_norm": 1.7439713183037207, "learning_rate": 3.4007837028918644e-06, "loss": 2.435, "step": 65410 }, { "epoch": 1.48, "grad_norm": 1.6184806846034687, "learning_rate": 3.398040368642079e-06, "loss": 2.385, "step": 65420 }, { "epoch": 1.48, "grad_norm": 1.6832191741357803, "learning_rate": 3.3952979148389474e-06, "loss": 2.4116, "step": 65430 }, { "epoch": 1.48, "grad_norm": 1.7769566754472796, "learning_rate": 3.3925563418482076e-06, "loss": 2.3428, "step": 65440 }, { "epoch": 1.48, "grad_norm": 1.80663129479018, "learning_rate": 3.389815650035485e-06, "loss": 2.303, "step": 65450 }, { "epoch": 1.48, "grad_norm": 1.7003559879771206, "learning_rate": 3.3870758397662796e-06, "loss": 2.3562, "step": 65460 }, { "epoch": 1.48, "grad_norm": 1.576240247921019, "learning_rate": 3.3843369114059754e-06, "loss": 2.325, "step": 65470 }, { "epoch": 1.48, "grad_norm": 1.6726466920865777, "learning_rate": 3.3815988653198417e-06, "loss": 2.4281, "step": 65480 }, { "epoch": 1.48, "grad_norm": 1.828280795859826, "learning_rate": 3.378861701873026e-06, "loss": 2.3418, "step": 65490 }, { "epoch": 1.48, "grad_norm": 1.687213837705903, "learning_rate": 3.376125421430563e-06, "loss": 2.4676, "step": 65500 }, { "epoch": 1.48, "grad_norm": 1.5628337537197334, "learning_rate": 3.3733900243573602e-06, "loss": 2.4371, "step": 65510 }, { "epoch": 1.48, "grad_norm": 1.7133630703036586, "learning_rate": 3.3706555110182225e-06, "loss": 2.3196, "step": 65520 }, { "epoch": 1.48, "grad_norm": 1.8129694520155568, "learning_rate": 3.367921881777829e-06, "loss": 2.267, "step": 65530 }, { "epoch": 1.48, "grad_norm": 1.5726037926430219, "learning_rate": 3.365189137000733e-06, "loss": 2.477, "step": 65540 }, { "epoch": 1.48, "grad_norm": 1.7027138090964569, "learning_rate": 3.362457277051381e-06, "loss": 2.387, "step": 65550 }, { "epoch": 1.48, "grad_norm": 1.6556920614439976, "learning_rate": 3.3597263022940963e-06, "loss": 2.3873, "step": 65560 }, { "epoch": 1.48, "grad_norm": 1.8172489238128295, "learning_rate": 3.3569962130930833e-06, "loss": 2.3817, "step": 65570 }, { "epoch": 1.48, "grad_norm": 1.5222566513329199, "learning_rate": 3.3542670098124375e-06, "loss": 2.3903, "step": 65580 }, { "epoch": 1.48, "grad_norm": 1.9214349602922172, "learning_rate": 3.3515386928161266e-06, "loss": 2.4364, "step": 65590 }, { "epoch": 1.48, "grad_norm": 1.7577026877612123, "learning_rate": 3.348811262468e-06, "loss": 2.4413, "step": 65600 }, { "epoch": 1.48, "grad_norm": 1.9523938546254473, "learning_rate": 3.3460847191317947e-06, "loss": 2.321, "step": 65610 }, { "epoch": 1.48, "grad_norm": 1.5676949325478386, "learning_rate": 3.3433590631711243e-06, "loss": 2.386, "step": 65620 }, { "epoch": 1.48, "grad_norm": 1.723401871087602, "learning_rate": 3.3406342949494874e-06, "loss": 2.3347, "step": 65630 }, { "epoch": 1.48, "grad_norm": 1.7232054100308682, "learning_rate": 3.3379104148302623e-06, "loss": 2.4586, "step": 65640 }, { "epoch": 1.48, "grad_norm": 1.8687421944382994, "learning_rate": 3.33518742317671e-06, "loss": 2.3837, "step": 65650 }, { "epoch": 1.48, "grad_norm": 1.7413612096120739, "learning_rate": 3.332465320351972e-06, "loss": 2.3514, "step": 65660 }, { "epoch": 1.48, "grad_norm": 1.7241802962589785, "learning_rate": 3.329744106719072e-06, "loss": 2.4507, "step": 65670 }, { "epoch": 1.48, "grad_norm": 1.7509950769171483, "learning_rate": 3.327023782640915e-06, "loss": 2.3535, "step": 65680 }, { "epoch": 1.48, "grad_norm": 1.8549379847635872, "learning_rate": 3.324304348480286e-06, "loss": 2.3899, "step": 65690 }, { "epoch": 1.48, "grad_norm": 1.6826077331340024, "learning_rate": 3.321585804599855e-06, "loss": 2.3404, "step": 65700 }, { "epoch": 1.48, "grad_norm": 1.8418605111844362, "learning_rate": 3.3188681513621658e-06, "loss": 2.4224, "step": 65710 }, { "epoch": 1.48, "grad_norm": 1.4609906112845654, "learning_rate": 3.3161513891296548e-06, "loss": 2.3854, "step": 65720 }, { "epoch": 1.48, "grad_norm": 1.838035309330548, "learning_rate": 3.3134355182646315e-06, "loss": 2.4573, "step": 65730 }, { "epoch": 1.48, "grad_norm": 1.609793927857842, "learning_rate": 3.3107205391292907e-06, "loss": 2.4316, "step": 65740 }, { "epoch": 1.48, "grad_norm": 1.878820935976372, "learning_rate": 3.308006452085699e-06, "loss": 2.318, "step": 65750 }, { "epoch": 1.48, "grad_norm": 1.5183122739545583, "learning_rate": 3.3052932574958153e-06, "loss": 2.4444, "step": 65760 }, { "epoch": 1.48, "grad_norm": 1.8374695993301435, "learning_rate": 3.3025809557214705e-06, "loss": 2.5107, "step": 65770 }, { "epoch": 1.48, "grad_norm": 1.549222026207497, "learning_rate": 3.299869547124388e-06, "loss": 2.362, "step": 65780 }, { "epoch": 1.48, "grad_norm": 1.640342130431927, "learning_rate": 3.2971590320661617e-06, "loss": 2.2669, "step": 65790 }, { "epoch": 1.48, "grad_norm": 1.6844729926852273, "learning_rate": 3.2944494109082692e-06, "loss": 2.3357, "step": 65800 }, { "epoch": 1.48, "grad_norm": 1.5367129298944031, "learning_rate": 3.2917406840120714e-06, "loss": 2.3595, "step": 65810 }, { "epoch": 1.48, "grad_norm": 2.0004514894618617, "learning_rate": 3.2890328517388046e-06, "loss": 2.4307, "step": 65820 }, { "epoch": 1.48, "grad_norm": 1.5818956230426582, "learning_rate": 3.2863259144495928e-06, "loss": 2.3556, "step": 65830 }, { "epoch": 1.48, "grad_norm": 1.5750955983552664, "learning_rate": 3.2836198725054346e-06, "loss": 2.4256, "step": 65840 }, { "epoch": 1.48, "grad_norm": 1.681766370975022, "learning_rate": 3.2809147262672127e-06, "loss": 2.3427, "step": 65850 }, { "epoch": 1.49, "grad_norm": 1.6434017295653938, "learning_rate": 3.278210476095689e-06, "loss": 2.3576, "step": 65860 }, { "epoch": 1.49, "grad_norm": 1.6459084026467345, "learning_rate": 3.2755071223515043e-06, "loss": 2.4769, "step": 65870 }, { "epoch": 1.49, "grad_norm": 1.5285087604454846, "learning_rate": 3.272804665395185e-06, "loss": 2.2744, "step": 65880 }, { "epoch": 1.49, "grad_norm": 1.8212962208085675, "learning_rate": 3.2701031055871323e-06, "loss": 2.2693, "step": 65890 }, { "epoch": 1.49, "grad_norm": 1.3850568912142638, "learning_rate": 3.2674024432876307e-06, "loss": 2.4474, "step": 65900 }, { "epoch": 1.49, "grad_norm": 1.6700666488276465, "learning_rate": 3.2647026788568403e-06, "loss": 2.5108, "step": 65910 }, { "epoch": 1.49, "grad_norm": 1.6360317515823488, "learning_rate": 3.2620038126548138e-06, "loss": 2.4456, "step": 65920 }, { "epoch": 1.49, "grad_norm": 1.9602715739713301, "learning_rate": 3.2593058450414705e-06, "loss": 2.365, "step": 65930 }, { "epoch": 1.49, "grad_norm": 1.5725089465099822, "learning_rate": 3.256608776376616e-06, "loss": 2.2814, "step": 65940 }, { "epoch": 1.49, "grad_norm": 1.4780223067622216, "learning_rate": 3.2539126070199357e-06, "loss": 2.401, "step": 65950 }, { "epoch": 1.49, "grad_norm": 1.728211977207046, "learning_rate": 3.2512173373309976e-06, "loss": 2.5195, "step": 65960 }, { "epoch": 1.49, "grad_norm": 1.8038854833595652, "learning_rate": 3.2485229676692355e-06, "loss": 2.3478, "step": 65970 }, { "epoch": 1.49, "grad_norm": 1.6323673483115775, "learning_rate": 3.2458294983939863e-06, "loss": 2.448, "step": 65980 }, { "epoch": 1.49, "grad_norm": 1.8246461722171783, "learning_rate": 3.24313692986445e-06, "loss": 2.3365, "step": 65990 }, { "epoch": 1.49, "grad_norm": 1.865386835975501, "learning_rate": 3.240445262439712e-06, "loss": 2.3412, "step": 66000 }, { "epoch": 1.49, "grad_norm": 1.6062493847691524, "learning_rate": 3.2377544964787378e-06, "loss": 2.4487, "step": 66010 }, { "epoch": 1.49, "grad_norm": 1.7531989405991475, "learning_rate": 3.2350646323403702e-06, "loss": 2.3891, "step": 66020 }, { "epoch": 1.49, "grad_norm": 1.6590244252738484, "learning_rate": 3.232375670383333e-06, "loss": 2.4004, "step": 66030 }, { "epoch": 1.49, "grad_norm": 1.945992562065073, "learning_rate": 3.229687610966229e-06, "loss": 2.3378, "step": 66040 }, { "epoch": 1.49, "grad_norm": 1.7957878965670486, "learning_rate": 3.2270004544475455e-06, "loss": 2.4542, "step": 66050 }, { "epoch": 1.49, "grad_norm": 1.7277694259342342, "learning_rate": 3.224314201185644e-06, "loss": 2.4668, "step": 66060 }, { "epoch": 1.49, "grad_norm": 1.7917177279838845, "learning_rate": 3.2216288515387696e-06, "loss": 2.3777, "step": 66070 }, { "epoch": 1.49, "grad_norm": 1.6232207573808939, "learning_rate": 3.2189444058650377e-06, "loss": 2.4644, "step": 66080 }, { "epoch": 1.49, "grad_norm": 1.5396688849463014, "learning_rate": 3.2162608645224524e-06, "loss": 2.3527, "step": 66090 }, { "epoch": 1.49, "grad_norm": 1.7618819591737966, "learning_rate": 3.2135782278688965e-06, "loss": 2.4085, "step": 66100 }, { "epoch": 1.49, "grad_norm": 1.8193510479813448, "learning_rate": 3.210896496262126e-06, "loss": 2.2991, "step": 66110 }, { "epoch": 1.49, "grad_norm": 1.8454720497431738, "learning_rate": 3.2082156700597856e-06, "loss": 2.4164, "step": 66120 }, { "epoch": 1.49, "grad_norm": 1.8621528248091144, "learning_rate": 3.205535749619393e-06, "loss": 2.3538, "step": 66130 }, { "epoch": 1.49, "grad_norm": 1.877776400675028, "learning_rate": 3.202856735298344e-06, "loss": 2.3954, "step": 66140 }, { "epoch": 1.49, "grad_norm": 1.5526435274413224, "learning_rate": 3.200178627453918e-06, "loss": 2.2438, "step": 66150 }, { "epoch": 1.49, "grad_norm": 1.7202393886216008, "learning_rate": 3.197501426443269e-06, "loss": 2.5038, "step": 66160 }, { "epoch": 1.49, "grad_norm": 1.6978533150907407, "learning_rate": 3.194825132623435e-06, "loss": 2.4895, "step": 66170 }, { "epoch": 1.49, "grad_norm": 1.6416652081234608, "learning_rate": 3.1921497463513274e-06, "loss": 2.4373, "step": 66180 }, { "epoch": 1.49, "grad_norm": 1.853449367796415, "learning_rate": 3.1894752679837406e-06, "loss": 2.3517, "step": 66190 }, { "epoch": 1.49, "grad_norm": 1.7101706641654788, "learning_rate": 3.1868016978773477e-06, "loss": 2.3497, "step": 66200 }, { "epoch": 1.49, "grad_norm": 1.5709591667171505, "learning_rate": 3.1841290363886988e-06, "loss": 2.3641, "step": 66210 }, { "epoch": 1.49, "grad_norm": 1.6593308313541604, "learning_rate": 3.181457283874224e-06, "loss": 2.5001, "step": 66220 }, { "epoch": 1.49, "grad_norm": 1.7826693316239128, "learning_rate": 3.1787864406902326e-06, "loss": 2.5384, "step": 66230 }, { "epoch": 1.49, "grad_norm": 1.7731730921904354, "learning_rate": 3.1761165071929078e-06, "loss": 2.3264, "step": 66240 }, { "epoch": 1.49, "grad_norm": 1.5403570219596125, "learning_rate": 3.1734474837383234e-06, "loss": 2.2993, "step": 66250 }, { "epoch": 1.49, "grad_norm": 1.8305022166648155, "learning_rate": 3.17077937068242e-06, "loss": 2.4122, "step": 66260 }, { "epoch": 1.49, "grad_norm": 1.661441252621213, "learning_rate": 3.1681121683810203e-06, "loss": 2.3882, "step": 66270 }, { "epoch": 1.49, "grad_norm": 1.7963615319495254, "learning_rate": 3.165445877189831e-06, "loss": 2.3266, "step": 66280 }, { "epoch": 1.49, "grad_norm": 1.4878452080699693, "learning_rate": 3.162780497464425e-06, "loss": 2.362, "step": 66290 }, { "epoch": 1.5, "grad_norm": 1.6112332973337393, "learning_rate": 3.1601160295602607e-06, "loss": 2.3619, "step": 66300 }, { "epoch": 1.5, "grad_norm": 1.8326018435722502, "learning_rate": 3.1574524738326826e-06, "loss": 2.2739, "step": 66310 }, { "epoch": 1.5, "grad_norm": 1.583407579324507, "learning_rate": 3.1547898306369027e-06, "loss": 2.4986, "step": 66320 }, { "epoch": 1.5, "grad_norm": 2.0347577952890106, "learning_rate": 3.1521281003280158e-06, "loss": 2.4692, "step": 66330 }, { "epoch": 1.5, "grad_norm": 1.809802304567181, "learning_rate": 3.1494672832609917e-06, "loss": 2.3724, "step": 66340 }, { "epoch": 1.5, "grad_norm": 1.7191642766852107, "learning_rate": 3.1468073797906827e-06, "loss": 2.3732, "step": 66350 }, { "epoch": 1.5, "grad_norm": 1.5136467012826331, "learning_rate": 3.144148390271817e-06, "loss": 2.482, "step": 66360 }, { "epoch": 1.5, "grad_norm": 1.7195098690686028, "learning_rate": 3.1414903150590014e-06, "loss": 2.359, "step": 66370 }, { "epoch": 1.5, "grad_norm": 1.5375467366772326, "learning_rate": 3.1388331545067164e-06, "loss": 2.4481, "step": 66380 }, { "epoch": 1.5, "grad_norm": 1.6580137026773867, "learning_rate": 3.1361769089693337e-06, "loss": 2.288, "step": 66390 }, { "epoch": 1.5, "grad_norm": 1.5851408528253583, "learning_rate": 3.1335215788010864e-06, "loss": 2.3732, "step": 66400 }, { "epoch": 1.5, "grad_norm": 1.7170624523317235, "learning_rate": 3.1308671643560963e-06, "loss": 2.3399, "step": 66410 }, { "epoch": 1.5, "grad_norm": 1.5156584441518293, "learning_rate": 3.1282136659883577e-06, "loss": 2.3613, "step": 66420 }, { "epoch": 1.5, "grad_norm": 1.8480469358127278, "learning_rate": 3.1255610840517457e-06, "loss": 2.3496, "step": 66430 }, { "epoch": 1.5, "grad_norm": 1.9225319218917438, "learning_rate": 3.1229094189000096e-06, "loss": 2.4519, "step": 66440 }, { "epoch": 1.5, "grad_norm": 1.544936403766472, "learning_rate": 3.1202586708867865e-06, "loss": 2.3479, "step": 66450 }, { "epoch": 1.5, "grad_norm": 1.7651987429830813, "learning_rate": 3.117608840365579e-06, "loss": 2.3899, "step": 66460 }, { "epoch": 1.5, "grad_norm": 1.5137608249306043, "learning_rate": 3.1149599276897735e-06, "loss": 2.4155, "step": 66470 }, { "epoch": 1.5, "grad_norm": 1.7892335449622165, "learning_rate": 3.1123119332126318e-06, "loss": 2.2747, "step": 66480 }, { "epoch": 1.5, "grad_norm": 1.7605507861720697, "learning_rate": 3.1096648572872945e-06, "loss": 2.5663, "step": 66490 }, { "epoch": 1.5, "grad_norm": 1.6559227804622527, "learning_rate": 3.1070187002667795e-06, "loss": 2.4541, "step": 66500 }, { "epoch": 1.5, "grad_norm": 1.662532715486948, "learning_rate": 3.1043734625039824e-06, "loss": 2.3661, "step": 66510 }, { "epoch": 1.5, "grad_norm": 1.6070284257384357, "learning_rate": 3.101729144351675e-06, "loss": 2.4633, "step": 66520 }, { "epoch": 1.5, "grad_norm": 1.8218193859413738, "learning_rate": 3.099085746162509e-06, "loss": 2.4438, "step": 66530 }, { "epoch": 1.5, "grad_norm": 1.6397098725289774, "learning_rate": 3.096443268289009e-06, "loss": 2.4136, "step": 66540 }, { "epoch": 1.5, "grad_norm": 1.6097429776576577, "learning_rate": 3.0938017110835817e-06, "loss": 2.2785, "step": 66550 }, { "epoch": 1.5, "grad_norm": 1.771998703856754, "learning_rate": 3.0911610748985078e-06, "loss": 2.42, "step": 66560 }, { "epoch": 1.5, "grad_norm": 1.8712464684271193, "learning_rate": 3.088521360085943e-06, "loss": 2.3671, "step": 66570 }, { "epoch": 1.5, "grad_norm": 1.6817107399444085, "learning_rate": 3.085882566997932e-06, "loss": 2.4159, "step": 66580 }, { "epoch": 1.5, "grad_norm": 1.6785480256040084, "learning_rate": 3.0832446959863817e-06, "loss": 2.4916, "step": 66590 }, { "epoch": 1.5, "grad_norm": 1.5963671712424583, "learning_rate": 3.080607747403084e-06, "loss": 2.4111, "step": 66600 }, { "epoch": 1.5, "grad_norm": 1.6656303022244352, "learning_rate": 3.0779717215997097e-06, "loss": 2.4402, "step": 66610 }, { "epoch": 1.5, "grad_norm": 1.9041887805912778, "learning_rate": 3.0753366189277946e-06, "loss": 2.3146, "step": 66620 }, { "epoch": 1.5, "grad_norm": 1.699122267503492, "learning_rate": 3.072702439738765e-06, "loss": 2.3912, "step": 66630 }, { "epoch": 1.5, "grad_norm": 1.8355153987157782, "learning_rate": 3.0700691843839135e-06, "loss": 2.3451, "step": 66640 }, { "epoch": 1.5, "grad_norm": 1.4417373274074974, "learning_rate": 3.0674368532144237e-06, "loss": 2.4278, "step": 66650 }, { "epoch": 1.5, "grad_norm": 1.9411463603403287, "learning_rate": 3.0648054465813415e-06, "loss": 2.3708, "step": 66660 }, { "epoch": 1.5, "grad_norm": 1.84128687585016, "learning_rate": 3.062174964835596e-06, "loss": 2.4865, "step": 66670 }, { "epoch": 1.5, "grad_norm": 1.9255154017667626, "learning_rate": 3.059545408327992e-06, "loss": 2.4518, "step": 66680 }, { "epoch": 1.5, "grad_norm": 1.521401674358215, "learning_rate": 3.056916777409211e-06, "loss": 2.3459, "step": 66690 }, { "epoch": 1.5, "grad_norm": 1.773652767524914, "learning_rate": 3.0542890724298102e-06, "loss": 2.3081, "step": 66700 }, { "epoch": 1.5, "grad_norm": 1.7032286013918356, "learning_rate": 3.0516622937402253e-06, "loss": 2.3358, "step": 66710 }, { "epoch": 1.5, "grad_norm": 1.4765055778838763, "learning_rate": 3.049036441690766e-06, "loss": 2.3629, "step": 66720 }, { "epoch": 1.5, "grad_norm": 1.4752848697143628, "learning_rate": 3.0464115166316198e-06, "loss": 2.3448, "step": 66730 }, { "epoch": 1.51, "grad_norm": 1.5897668612286144, "learning_rate": 3.0437875189128506e-06, "loss": 2.3457, "step": 66740 }, { "epoch": 1.51, "grad_norm": 1.683577344865483, "learning_rate": 3.0411644488843994e-06, "loss": 2.3202, "step": 66750 }, { "epoch": 1.51, "grad_norm": 1.7666784134694642, "learning_rate": 3.0385423068960805e-06, "loss": 2.352, "step": 66760 }, { "epoch": 1.51, "grad_norm": 1.7717348182542654, "learning_rate": 3.0359210932975845e-06, "loss": 2.3768, "step": 66770 }, { "epoch": 1.51, "grad_norm": 1.6058614544396594, "learning_rate": 3.0333008084384876e-06, "loss": 2.4118, "step": 66780 }, { "epoch": 1.51, "grad_norm": 1.8429804238993375, "learning_rate": 3.0306814526682295e-06, "loss": 2.5848, "step": 66790 }, { "epoch": 1.51, "grad_norm": 1.716513089933012, "learning_rate": 3.028063026336132e-06, "loss": 2.3733, "step": 66800 }, { "epoch": 1.51, "grad_norm": 1.6338020164476494, "learning_rate": 3.025445529791392e-06, "loss": 2.3491, "step": 66810 }, { "epoch": 1.51, "grad_norm": 1.6521163058771549, "learning_rate": 3.0228289633830866e-06, "loss": 2.4212, "step": 66820 }, { "epoch": 1.51, "grad_norm": 1.933861863311796, "learning_rate": 3.020213327460154e-06, "loss": 2.4315, "step": 66830 }, { "epoch": 1.51, "grad_norm": 1.8260768108242544, "learning_rate": 3.0175986223714295e-06, "loss": 2.3595, "step": 66840 }, { "epoch": 1.51, "grad_norm": 1.6418370418453905, "learning_rate": 3.01498484846561e-06, "loss": 2.3431, "step": 66850 }, { "epoch": 1.51, "grad_norm": 1.686484767196342, "learning_rate": 3.012372006091272e-06, "loss": 2.3706, "step": 66860 }, { "epoch": 1.51, "grad_norm": 1.8524768710924984, "learning_rate": 3.0097600955968687e-06, "loss": 2.3665, "step": 66870 }, { "epoch": 1.51, "grad_norm": 1.5419036170688318, "learning_rate": 3.0071491173307276e-06, "loss": 2.407, "step": 66880 }, { "epoch": 1.51, "grad_norm": 1.5639238554569332, "learning_rate": 3.004539071641053e-06, "loss": 2.3431, "step": 66890 }, { "epoch": 1.51, "grad_norm": 1.6575772056367728, "learning_rate": 3.001929958875923e-06, "loss": 2.4516, "step": 66900 }, { "epoch": 1.51, "grad_norm": 1.641862009231984, "learning_rate": 2.999321779383292e-06, "loss": 2.4525, "step": 66910 }, { "epoch": 1.51, "grad_norm": 1.6923143595903587, "learning_rate": 2.996714533510994e-06, "loss": 2.4036, "step": 66920 }, { "epoch": 1.51, "grad_norm": 1.732057807637737, "learning_rate": 2.994108221606735e-06, "loss": 2.3556, "step": 66930 }, { "epoch": 1.51, "grad_norm": 1.6616880418480164, "learning_rate": 2.991502844018093e-06, "loss": 2.3675, "step": 66940 }, { "epoch": 1.51, "grad_norm": 1.695245777213209, "learning_rate": 2.9888984010925247e-06, "loss": 2.2993, "step": 66950 }, { "epoch": 1.51, "grad_norm": 1.637899548125182, "learning_rate": 2.986294893177365e-06, "loss": 2.3778, "step": 66960 }, { "epoch": 1.51, "grad_norm": 1.7185409783466299, "learning_rate": 2.9836923206198155e-06, "loss": 2.3357, "step": 66970 }, { "epoch": 1.51, "grad_norm": 1.7766086566595538, "learning_rate": 2.9810906837669683e-06, "loss": 2.4713, "step": 66980 }, { "epoch": 1.51, "grad_norm": 1.6432402423605885, "learning_rate": 2.9784899829657753e-06, "loss": 2.3394, "step": 66990 }, { "epoch": 1.51, "grad_norm": 1.664608092333408, "learning_rate": 2.9758902185630724e-06, "loss": 2.4935, "step": 67000 }, { "epoch": 1.51, "grad_norm": 1.6394296364657648, "learning_rate": 2.973291390905566e-06, "loss": 2.3829, "step": 67010 }, { "epoch": 1.51, "grad_norm": 1.7302443266795626, "learning_rate": 2.9706935003398397e-06, "loss": 2.328, "step": 67020 }, { "epoch": 1.51, "grad_norm": 1.6972310682220528, "learning_rate": 2.9680965472123515e-06, "loss": 2.5892, "step": 67030 }, { "epoch": 1.51, "grad_norm": 1.6393542576281135, "learning_rate": 2.965500531869435e-06, "loss": 2.3125, "step": 67040 }, { "epoch": 1.51, "grad_norm": 1.8791281987583368, "learning_rate": 2.9629054546573e-06, "loss": 2.4068, "step": 67050 }, { "epoch": 1.51, "grad_norm": 1.5753667999482244, "learning_rate": 2.9603113159220275e-06, "loss": 2.3723, "step": 67060 }, { "epoch": 1.51, "grad_norm": 1.6136927699481427, "learning_rate": 2.957718116009577e-06, "loss": 2.4044, "step": 67070 }, { "epoch": 1.51, "grad_norm": 1.7231360600562589, "learning_rate": 2.95512585526578e-06, "loss": 2.3728, "step": 67080 }, { "epoch": 1.51, "grad_norm": 1.5896835834707144, "learning_rate": 2.9525345340363456e-06, "loss": 2.3414, "step": 67090 }, { "epoch": 1.51, "grad_norm": 1.922955371352372, "learning_rate": 2.9499441526668517e-06, "loss": 2.4303, "step": 67100 }, { "epoch": 1.51, "grad_norm": 1.8216290041785173, "learning_rate": 2.9473547115027624e-06, "loss": 2.3647, "step": 67110 }, { "epoch": 1.51, "grad_norm": 2.016928611024718, "learning_rate": 2.9447662108894047e-06, "loss": 2.4836, "step": 67120 }, { "epoch": 1.51, "grad_norm": 1.552457683306896, "learning_rate": 2.9421786511719864e-06, "loss": 2.4587, "step": 67130 }, { "epoch": 1.51, "grad_norm": 1.4501457857592734, "learning_rate": 2.939592032695586e-06, "loss": 2.4208, "step": 67140 }, { "epoch": 1.51, "grad_norm": 1.8048388409019889, "learning_rate": 2.9370063558051644e-06, "loss": 2.3374, "step": 67150 }, { "epoch": 1.51, "grad_norm": 1.592130993494144, "learning_rate": 2.9344216208455425e-06, "loss": 2.4624, "step": 67160 }, { "epoch": 1.51, "grad_norm": 1.7802580630881135, "learning_rate": 2.931837828161426e-06, "loss": 2.2562, "step": 67170 }, { "epoch": 1.51, "grad_norm": 1.7120451975501145, "learning_rate": 2.9292549780973978e-06, "loss": 2.2851, "step": 67180 }, { "epoch": 1.52, "grad_norm": 1.5116679379493583, "learning_rate": 2.9266730709979075e-06, "loss": 2.3916, "step": 67190 }, { "epoch": 1.52, "grad_norm": 1.625096844776035, "learning_rate": 2.9240921072072837e-06, "loss": 2.3721, "step": 67200 }, { "epoch": 1.52, "grad_norm": 1.8884603529937318, "learning_rate": 2.9215120870697254e-06, "loss": 2.2329, "step": 67210 }, { "epoch": 1.52, "grad_norm": 1.7291622643356082, "learning_rate": 2.918933010929308e-06, "loss": 2.4315, "step": 67220 }, { "epoch": 1.52, "grad_norm": 1.7882897692372723, "learning_rate": 2.9163548791299823e-06, "loss": 2.3614, "step": 67230 }, { "epoch": 1.52, "grad_norm": 1.742197026326161, "learning_rate": 2.913777692015567e-06, "loss": 2.4139, "step": 67240 }, { "epoch": 1.52, "grad_norm": 1.5693319463131548, "learning_rate": 2.9112014499297693e-06, "loss": 2.4455, "step": 67250 }, { "epoch": 1.52, "grad_norm": 1.5850111922389565, "learning_rate": 2.9086261532161496e-06, "loss": 2.4713, "step": 67260 }, { "epoch": 1.52, "grad_norm": 1.5361502597608083, "learning_rate": 2.9060518022181596e-06, "loss": 2.3221, "step": 67270 }, { "epoch": 1.52, "grad_norm": 1.598043253480198, "learning_rate": 2.903478397279115e-06, "loss": 2.3774, "step": 67280 }, { "epoch": 1.52, "grad_norm": 1.6701724259816928, "learning_rate": 2.90090593874221e-06, "loss": 2.3645, "step": 67290 }, { "epoch": 1.52, "grad_norm": 1.67801789669423, "learning_rate": 2.8983344269505086e-06, "loss": 2.2936, "step": 67300 }, { "epoch": 1.52, "grad_norm": 1.715103875088444, "learning_rate": 2.8957638622469574e-06, "loss": 2.4338, "step": 67310 }, { "epoch": 1.52, "grad_norm": 1.5566683983086154, "learning_rate": 2.8931942449743665e-06, "loss": 2.3451, "step": 67320 }, { "epoch": 1.52, "grad_norm": 1.6509239814870733, "learning_rate": 2.8906255754754252e-06, "loss": 2.5328, "step": 67330 }, { "epoch": 1.52, "grad_norm": 1.959087974752096, "learning_rate": 2.888057854092693e-06, "loss": 2.2901, "step": 67340 }, { "epoch": 1.52, "grad_norm": 1.6512898045674154, "learning_rate": 2.8854910811686065e-06, "loss": 2.3054, "step": 67350 }, { "epoch": 1.52, "grad_norm": 1.8474386854129097, "learning_rate": 2.8829252570454734e-06, "loss": 2.4363, "step": 67360 }, { "epoch": 1.52, "grad_norm": 1.5089611831118928, "learning_rate": 2.8803603820654757e-06, "loss": 2.5199, "step": 67370 }, { "epoch": 1.52, "grad_norm": 1.559858218929656, "learning_rate": 2.8777964565706695e-06, "loss": 2.4745, "step": 67380 }, { "epoch": 1.52, "grad_norm": 1.747696840647478, "learning_rate": 2.8752334809029824e-06, "loss": 2.3918, "step": 67390 }, { "epoch": 1.52, "grad_norm": 1.729767205730615, "learning_rate": 2.8726714554042168e-06, "loss": 2.4483, "step": 67400 }, { "epoch": 1.52, "grad_norm": 1.5943024696532782, "learning_rate": 2.870110380416048e-06, "loss": 2.4222, "step": 67410 }, { "epoch": 1.52, "grad_norm": 1.6995974984591795, "learning_rate": 2.867550256280024e-06, "loss": 2.3932, "step": 67420 }, { "epoch": 1.52, "grad_norm": 1.6478170792873506, "learning_rate": 2.864991083337568e-06, "loss": 2.4252, "step": 67430 }, { "epoch": 1.52, "grad_norm": 1.6257018658860214, "learning_rate": 2.8624328619299714e-06, "loss": 2.3791, "step": 67440 }, { "epoch": 1.52, "grad_norm": 1.5705910355462058, "learning_rate": 2.8598755923984078e-06, "loss": 2.3584, "step": 67450 }, { "epoch": 1.52, "grad_norm": 1.6678151073326655, "learning_rate": 2.8573192750839163e-06, "loss": 2.3381, "step": 67460 }, { "epoch": 1.52, "grad_norm": 1.6018552958646912, "learning_rate": 2.8547639103274116e-06, "loss": 2.4522, "step": 67470 }, { "epoch": 1.52, "grad_norm": 1.5051511413878946, "learning_rate": 2.852209498469678e-06, "loss": 2.4977, "step": 67480 }, { "epoch": 1.52, "grad_norm": 1.5598494754865175, "learning_rate": 2.849656039851375e-06, "loss": 2.2835, "step": 67490 }, { "epoch": 1.52, "grad_norm": 1.7596422296403256, "learning_rate": 2.847103534813036e-06, "loss": 2.4358, "step": 67500 }, { "epoch": 1.52, "grad_norm": 1.5357091977446238, "learning_rate": 2.8445519836950696e-06, "loss": 2.3879, "step": 67510 }, { "epoch": 1.52, "grad_norm": 1.8662104778230335, "learning_rate": 2.8420013868377527e-06, "loss": 2.4173, "step": 67520 }, { "epoch": 1.52, "grad_norm": 1.689214305056328, "learning_rate": 2.8394517445812373e-06, "loss": 2.4276, "step": 67530 }, { "epoch": 1.52, "grad_norm": 1.8404923537715887, "learning_rate": 2.8369030572655455e-06, "loss": 2.3322, "step": 67540 }, { "epoch": 1.52, "grad_norm": 1.8685363801734889, "learning_rate": 2.8343553252305746e-06, "loss": 2.3427, "step": 67550 }, { "epoch": 1.52, "grad_norm": 1.593633327708858, "learning_rate": 2.831808548816094e-06, "loss": 2.393, "step": 67560 }, { "epoch": 1.52, "grad_norm": 1.617060477084324, "learning_rate": 2.8292627283617424e-06, "loss": 2.411, "step": 67570 }, { "epoch": 1.52, "grad_norm": 1.609643836262976, "learning_rate": 2.8267178642070436e-06, "loss": 2.347, "step": 67580 }, { "epoch": 1.52, "grad_norm": 1.8575657390813853, "learning_rate": 2.824173956691374e-06, "loss": 2.3702, "step": 67590 }, { "epoch": 1.52, "grad_norm": 1.6435536920953375, "learning_rate": 2.8216310061539955e-06, "loss": 2.4137, "step": 67600 }, { "epoch": 1.52, "grad_norm": 1.7547985276168825, "learning_rate": 2.819089012934041e-06, "loss": 2.3711, "step": 67610 }, { "epoch": 1.52, "grad_norm": 1.7726463256993796, "learning_rate": 2.816547977370514e-06, "loss": 2.3929, "step": 67620 }, { "epoch": 1.53, "grad_norm": 1.6401252468142833, "learning_rate": 2.814007899802286e-06, "loss": 2.4899, "step": 67630 }, { "epoch": 1.53, "grad_norm": 1.8410231618460953, "learning_rate": 2.811468780568113e-06, "loss": 2.4949, "step": 67640 }, { "epoch": 1.53, "grad_norm": 1.485400960671335, "learning_rate": 2.8089306200066124e-06, "loss": 2.3795, "step": 67650 }, { "epoch": 1.53, "grad_norm": 1.8250045795300467, "learning_rate": 2.806393418456276e-06, "loss": 2.5833, "step": 67660 }, { "epoch": 1.53, "grad_norm": 1.7014194172788388, "learning_rate": 2.80385717625547e-06, "loss": 2.3803, "step": 67670 }, { "epoch": 1.53, "grad_norm": 1.713471080851077, "learning_rate": 2.8013218937424293e-06, "loss": 2.4361, "step": 67680 }, { "epoch": 1.53, "grad_norm": 1.674065981807929, "learning_rate": 2.798787571255267e-06, "loss": 2.5539, "step": 67690 }, { "epoch": 1.53, "grad_norm": 1.587140266988368, "learning_rate": 2.7962542091319555e-06, "loss": 2.4393, "step": 67700 }, { "epoch": 1.53, "grad_norm": 1.811747564353847, "learning_rate": 2.7937218077103558e-06, "loss": 2.4832, "step": 67710 }, { "epoch": 1.53, "grad_norm": 1.7871355930880524, "learning_rate": 2.791190367328188e-06, "loss": 2.3412, "step": 67720 }, { "epoch": 1.53, "grad_norm": 1.7347936391055485, "learning_rate": 2.788659888323051e-06, "loss": 2.411, "step": 67730 }, { "epoch": 1.53, "grad_norm": 1.7230563743479197, "learning_rate": 2.7861303710324126e-06, "loss": 2.464, "step": 67740 }, { "epoch": 1.53, "grad_norm": 1.5589144635292989, "learning_rate": 2.7836018157936117e-06, "loss": 2.4159, "step": 67750 }, { "epoch": 1.53, "grad_norm": 1.75826070614706, "learning_rate": 2.7810742229438605e-06, "loss": 2.405, "step": 67760 }, { "epoch": 1.53, "grad_norm": 1.6634094766297005, "learning_rate": 2.77854759282024e-06, "loss": 2.3533, "step": 67770 }, { "epoch": 1.53, "grad_norm": 1.5475248574099538, "learning_rate": 2.7760219257597098e-06, "loss": 2.3524, "step": 67780 }, { "epoch": 1.53, "grad_norm": 1.8034601087477329, "learning_rate": 2.7734972220990987e-06, "loss": 2.3443, "step": 67790 }, { "epoch": 1.53, "grad_norm": 1.598962541120397, "learning_rate": 2.7709734821750957e-06, "loss": 2.5604, "step": 67800 }, { "epoch": 1.53, "grad_norm": 1.887603116562434, "learning_rate": 2.768450706324276e-06, "loss": 2.4048, "step": 67810 }, { "epoch": 1.53, "grad_norm": 1.665951241480448, "learning_rate": 2.7659288948830785e-06, "loss": 2.4509, "step": 67820 }, { "epoch": 1.53, "grad_norm": 1.5334887254527425, "learning_rate": 2.7634080481878144e-06, "loss": 2.4169, "step": 67830 }, { "epoch": 1.53, "grad_norm": 1.8707021218843791, "learning_rate": 2.7608881665746722e-06, "loss": 2.4159, "step": 67840 }, { "epoch": 1.53, "grad_norm": 1.5412458372440996, "learning_rate": 2.7583692503797043e-06, "loss": 2.4471, "step": 67850 }, { "epoch": 1.53, "grad_norm": 1.7306227203149196, "learning_rate": 2.7558512999388363e-06, "loss": 2.2909, "step": 67860 }, { "epoch": 1.53, "grad_norm": 1.7580279197547983, "learning_rate": 2.7533343155878667e-06, "loss": 2.3567, "step": 67870 }, { "epoch": 1.53, "grad_norm": 1.6392468514462097, "learning_rate": 2.750818297662462e-06, "loss": 2.3754, "step": 67880 }, { "epoch": 1.53, "grad_norm": 1.6102522216596493, "learning_rate": 2.748303246498165e-06, "loss": 2.4117, "step": 67890 }, { "epoch": 1.53, "grad_norm": 1.6994076671913017, "learning_rate": 2.7457891624303845e-06, "loss": 2.3765, "step": 67900 }, { "epoch": 1.53, "grad_norm": 1.856473050150092, "learning_rate": 2.743276045794402e-06, "loss": 2.3118, "step": 67910 }, { "epoch": 1.53, "grad_norm": 1.7956473835864972, "learning_rate": 2.740763896925371e-06, "loss": 2.4713, "step": 67920 }, { "epoch": 1.53, "grad_norm": 1.514913504087994, "learning_rate": 2.738252716158316e-06, "loss": 2.3395, "step": 67930 }, { "epoch": 1.53, "grad_norm": 3.1284890902903495, "learning_rate": 2.7357425038281306e-06, "loss": 2.3941, "step": 67940 }, { "epoch": 1.53, "grad_norm": 1.6422245725116498, "learning_rate": 2.73323326026958e-06, "loss": 2.4371, "step": 67950 }, { "epoch": 1.53, "grad_norm": 1.73006108122524, "learning_rate": 2.730724985817301e-06, "loss": 2.4063, "step": 67960 }, { "epoch": 1.53, "grad_norm": 1.7975673033067094, "learning_rate": 2.728217680805797e-06, "loss": 2.4105, "step": 67970 }, { "epoch": 1.53, "grad_norm": 1.695770795216884, "learning_rate": 2.725711345569453e-06, "loss": 2.3989, "step": 67980 }, { "epoch": 1.53, "grad_norm": 1.7411017528243315, "learning_rate": 2.723205980442514e-06, "loss": 2.4487, "step": 67990 }, { "epoch": 1.53, "grad_norm": 1.7671591370600015, "learning_rate": 2.720701585759098e-06, "loss": 2.3126, "step": 68000 }, { "epoch": 1.53, "grad_norm": 1.7021760172385334, "learning_rate": 2.718198161853198e-06, "loss": 2.4452, "step": 68010 }, { "epoch": 1.53, "grad_norm": 1.7677201651275871, "learning_rate": 2.715695709058669e-06, "loss": 2.4411, "step": 68020 }, { "epoch": 1.53, "grad_norm": 1.6378424620248688, "learning_rate": 2.7131942277092406e-06, "loss": 2.3831, "step": 68030 }, { "epoch": 1.53, "grad_norm": 1.6029201840062677, "learning_rate": 2.710693718138521e-06, "loss": 2.3881, "step": 68040 }, { "epoch": 1.53, "grad_norm": 1.7167142184468105, "learning_rate": 2.708194180679977e-06, "loss": 2.3728, "step": 68050 }, { "epoch": 1.53, "grad_norm": 1.8065553013497566, "learning_rate": 2.705695615666952e-06, "loss": 2.3422, "step": 68060 }, { "epoch": 1.54, "grad_norm": 1.6736796782533987, "learning_rate": 2.7031980234326584e-06, "loss": 2.3195, "step": 68070 }, { "epoch": 1.54, "grad_norm": 1.8200913051708236, "learning_rate": 2.7007014043101765e-06, "loss": 2.425, "step": 68080 }, { "epoch": 1.54, "grad_norm": 1.755060700486225, "learning_rate": 2.698205758632462e-06, "loss": 2.515, "step": 68090 }, { "epoch": 1.54, "grad_norm": 1.6335310171729167, "learning_rate": 2.695711086732331e-06, "loss": 2.606, "step": 68100 }, { "epoch": 1.54, "grad_norm": 1.453553974355926, "learning_rate": 2.6932173889424862e-06, "loss": 2.4264, "step": 68110 }, { "epoch": 1.54, "grad_norm": 1.8124563998125127, "learning_rate": 2.6907246655954887e-06, "loss": 2.3684, "step": 68120 }, { "epoch": 1.54, "grad_norm": 1.7297865819685903, "learning_rate": 2.6882329170237654e-06, "loss": 2.3419, "step": 68130 }, { "epoch": 1.54, "grad_norm": 1.7202588812877064, "learning_rate": 2.6857421435596233e-06, "loss": 2.376, "step": 68140 }, { "epoch": 1.54, "grad_norm": 1.5787067096843963, "learning_rate": 2.6832523455352342e-06, "loss": 2.3362, "step": 68150 }, { "epoch": 1.54, "grad_norm": 1.6811521751110339, "learning_rate": 2.6807635232826436e-06, "loss": 2.299, "step": 68160 }, { "epoch": 1.54, "grad_norm": 1.5820561989385489, "learning_rate": 2.678275677133758e-06, "loss": 2.3736, "step": 68170 }, { "epoch": 1.54, "grad_norm": 1.5972013672185918, "learning_rate": 2.6757888074203685e-06, "loss": 2.2764, "step": 68180 }, { "epoch": 1.54, "grad_norm": 1.6169481120622853, "learning_rate": 2.6733029144741242e-06, "loss": 2.4634, "step": 68190 }, { "epoch": 1.54, "grad_norm": 2.168939127264286, "learning_rate": 2.670817998626546e-06, "loss": 2.4469, "step": 68200 }, { "epoch": 1.54, "grad_norm": 1.6164443000981656, "learning_rate": 2.668334060209027e-06, "loss": 2.3822, "step": 68210 }, { "epoch": 1.54, "grad_norm": 1.870162993129608, "learning_rate": 2.665851099552832e-06, "loss": 2.3339, "step": 68220 }, { "epoch": 1.54, "grad_norm": 1.6829890167061954, "learning_rate": 2.663369116989082e-06, "loss": 2.402, "step": 68230 }, { "epoch": 1.54, "grad_norm": 1.755549713827103, "learning_rate": 2.6608881128487875e-06, "loss": 2.3322, "step": 68240 }, { "epoch": 1.54, "grad_norm": 1.6657220690039005, "learning_rate": 2.658408087462816e-06, "loss": 2.191, "step": 68250 }, { "epoch": 1.54, "grad_norm": 1.8464969214155316, "learning_rate": 2.6559290411619053e-06, "loss": 2.37, "step": 68260 }, { "epoch": 1.54, "grad_norm": 1.7532480846617886, "learning_rate": 2.6534509742766667e-06, "loss": 2.5213, "step": 68270 }, { "epoch": 1.54, "grad_norm": 1.5898430135833292, "learning_rate": 2.650973887137578e-06, "loss": 2.2312, "step": 68280 }, { "epoch": 1.54, "grad_norm": 1.8308801010371192, "learning_rate": 2.6484977800749867e-06, "loss": 2.4111, "step": 68290 }, { "epoch": 1.54, "grad_norm": 1.7482075209524284, "learning_rate": 2.646022653419107e-06, "loss": 2.3694, "step": 68300 }, { "epoch": 1.54, "grad_norm": 1.600034803207759, "learning_rate": 2.6435485075000324e-06, "loss": 2.3428, "step": 68310 }, { "epoch": 1.54, "grad_norm": 1.6223803575673725, "learning_rate": 2.641075342647713e-06, "loss": 2.3389, "step": 68320 }, { "epoch": 1.54, "grad_norm": 1.5041723154883124, "learning_rate": 2.6386031591919793e-06, "loss": 2.3479, "step": 68330 }, { "epoch": 1.54, "grad_norm": 1.8143150452692072, "learning_rate": 2.6361319574625175e-06, "loss": 2.4462, "step": 68340 }, { "epoch": 1.54, "grad_norm": 1.5817173099470274, "learning_rate": 2.633661737788894e-06, "loss": 2.4608, "step": 68350 }, { "epoch": 1.54, "grad_norm": 1.7321207949278765, "learning_rate": 2.631192500500538e-06, "loss": 2.3556, "step": 68360 }, { "epoch": 1.54, "grad_norm": 1.6962401619952132, "learning_rate": 2.6287242459267558e-06, "loss": 2.3609, "step": 68370 }, { "epoch": 1.54, "grad_norm": 1.663185003379326, "learning_rate": 2.626256974396715e-06, "loss": 2.3655, "step": 68380 }, { "epoch": 1.54, "grad_norm": 1.9329900173792782, "learning_rate": 2.6237906862394546e-06, "loss": 2.2926, "step": 68390 }, { "epoch": 1.54, "grad_norm": 1.7712076291566117, "learning_rate": 2.6213253817838814e-06, "loss": 2.3442, "step": 68400 }, { "epoch": 1.54, "grad_norm": 1.5739040875488113, "learning_rate": 2.618861061358773e-06, "loss": 2.2849, "step": 68410 }, { "epoch": 1.54, "grad_norm": 1.6222364338825326, "learning_rate": 2.616397725292773e-06, "loss": 2.5063, "step": 68420 }, { "epoch": 1.54, "grad_norm": 1.631204269146112, "learning_rate": 2.613935373914397e-06, "loss": 2.3193, "step": 68430 }, { "epoch": 1.54, "grad_norm": 1.7028000061438964, "learning_rate": 2.611474007552027e-06, "loss": 2.468, "step": 68440 }, { "epoch": 1.54, "grad_norm": 1.7239137392567832, "learning_rate": 2.6090136265339137e-06, "loss": 2.3994, "step": 68450 }, { "epoch": 1.54, "grad_norm": 1.5791804995605856, "learning_rate": 2.606554231188179e-06, "loss": 2.3874, "step": 68460 }, { "epoch": 1.54, "grad_norm": 1.8414663635285053, "learning_rate": 2.6040958218428094e-06, "loss": 2.4167, "step": 68470 }, { "epoch": 1.54, "grad_norm": 1.5170733906181668, "learning_rate": 2.601638398825661e-06, "loss": 2.3138, "step": 68480 }, { "epoch": 1.54, "grad_norm": 1.6697115454593885, "learning_rate": 2.599181962464462e-06, "loss": 2.3734, "step": 68490 }, { "epoch": 1.54, "grad_norm": 1.7183548474728365, "learning_rate": 2.5967265130868014e-06, "loss": 2.3784, "step": 68500 }, { "epoch": 1.54, "grad_norm": 2.216297345332204, "learning_rate": 2.594272051020148e-06, "loss": 2.3363, "step": 68510 }, { "epoch": 1.55, "grad_norm": 1.6619711119276315, "learning_rate": 2.591818576591829e-06, "loss": 2.3327, "step": 68520 }, { "epoch": 1.55, "grad_norm": 1.8315971571239842, "learning_rate": 2.589366090129043e-06, "loss": 2.4385, "step": 68530 }, { "epoch": 1.55, "grad_norm": 1.6842950772718646, "learning_rate": 2.5869145919588577e-06, "loss": 2.2784, "step": 68540 }, { "epoch": 1.55, "grad_norm": 1.7624685249656338, "learning_rate": 2.5844640824082114e-06, "loss": 2.4157, "step": 68550 }, { "epoch": 1.55, "grad_norm": 1.7097995225980778, "learning_rate": 2.5820145618038984e-06, "loss": 2.3721, "step": 68560 }, { "epoch": 1.55, "grad_norm": 1.6061274752518413, "learning_rate": 2.579566030472598e-06, "loss": 2.4164, "step": 68570 }, { "epoch": 1.55, "grad_norm": 1.6130134629248947, "learning_rate": 2.5771184887408485e-06, "loss": 2.2822, "step": 68580 }, { "epoch": 1.55, "grad_norm": 1.8470063334680205, "learning_rate": 2.574671936935057e-06, "loss": 2.3613, "step": 68590 }, { "epoch": 1.55, "grad_norm": 1.7074609585036968, "learning_rate": 2.572226375381498e-06, "loss": 2.3627, "step": 68600 }, { "epoch": 1.55, "grad_norm": 1.6668770473861194, "learning_rate": 2.569781804406317e-06, "loss": 2.3935, "step": 68610 }, { "epoch": 1.55, "grad_norm": 2.0319787353409895, "learning_rate": 2.567338224335524e-06, "loss": 2.3501, "step": 68620 }, { "epoch": 1.55, "grad_norm": 1.690560857522365, "learning_rate": 2.564895635494995e-06, "loss": 2.227, "step": 68630 }, { "epoch": 1.55, "grad_norm": 1.7177142582657645, "learning_rate": 2.5624540382104847e-06, "loss": 2.4094, "step": 68640 }, { "epoch": 1.55, "grad_norm": 1.6982956452657623, "learning_rate": 2.5600134328076032e-06, "loss": 2.4147, "step": 68650 }, { "epoch": 1.55, "grad_norm": 1.7261889793384464, "learning_rate": 2.557573819611837e-06, "loss": 2.2973, "step": 68660 }, { "epoch": 1.55, "grad_norm": 1.808071155832171, "learning_rate": 2.555135198948532e-06, "loss": 2.4311, "step": 68670 }, { "epoch": 1.55, "grad_norm": 1.9286597102496463, "learning_rate": 2.552697571142906e-06, "loss": 2.4473, "step": 68680 }, { "epoch": 1.55, "grad_norm": 1.4554095235986781, "learning_rate": 2.5502609365200458e-06, "loss": 2.41, "step": 68690 }, { "epoch": 1.55, "grad_norm": 1.579095519050989, "learning_rate": 2.547825295404901e-06, "loss": 2.3398, "step": 68700 }, { "epoch": 1.55, "grad_norm": 1.7068080824435543, "learning_rate": 2.5453906481223e-06, "loss": 2.3695, "step": 68710 }, { "epoch": 1.55, "grad_norm": 1.5454458227760381, "learning_rate": 2.5429569949969257e-06, "loss": 2.4527, "step": 68720 }, { "epoch": 1.55, "grad_norm": 1.8445810318483982, "learning_rate": 2.5405243363533337e-06, "loss": 2.4042, "step": 68730 }, { "epoch": 1.55, "grad_norm": 1.7717205677339112, "learning_rate": 2.538092672515948e-06, "loss": 2.3191, "step": 68740 }, { "epoch": 1.55, "grad_norm": 1.600724060199032, "learning_rate": 2.5356620038090584e-06, "loss": 2.2503, "step": 68750 }, { "epoch": 1.55, "grad_norm": 1.8103206341776723, "learning_rate": 2.5332323305568207e-06, "loss": 2.3195, "step": 68760 }, { "epoch": 1.55, "grad_norm": 2.031208631535753, "learning_rate": 2.5308036530832625e-06, "loss": 2.3388, "step": 68770 }, { "epoch": 1.55, "grad_norm": 1.652891606799813, "learning_rate": 2.5283759717122725e-06, "loss": 2.1661, "step": 68780 }, { "epoch": 1.55, "grad_norm": 1.6193505387588567, "learning_rate": 2.525949286767612e-06, "loss": 2.4174, "step": 68790 }, { "epoch": 1.55, "grad_norm": 1.6601797483250378, "learning_rate": 2.5235235985729067e-06, "loss": 2.3319, "step": 68800 }, { "epoch": 1.55, "grad_norm": 1.794439903127064, "learning_rate": 2.521098907451649e-06, "loss": 2.2904, "step": 68810 }, { "epoch": 1.55, "grad_norm": 1.476431682033887, "learning_rate": 2.5186752137271984e-06, "loss": 2.5462, "step": 68820 }, { "epoch": 1.55, "grad_norm": 1.7799566650174463, "learning_rate": 2.5162525177227825e-06, "loss": 2.3772, "step": 68830 }, { "epoch": 1.55, "grad_norm": 1.59348401468194, "learning_rate": 2.513830819761498e-06, "loss": 2.3506, "step": 68840 }, { "epoch": 1.55, "grad_norm": 1.8271167107631126, "learning_rate": 2.5114101201663044e-06, "loss": 2.3575, "step": 68850 }, { "epoch": 1.55, "grad_norm": 1.5250603514940648, "learning_rate": 2.508990419260029e-06, "loss": 2.2705, "step": 68860 }, { "epoch": 1.55, "grad_norm": 1.6087608328216945, "learning_rate": 2.5065717173653703e-06, "loss": 2.3817, "step": 68870 }, { "epoch": 1.55, "grad_norm": 1.7228627915048826, "learning_rate": 2.504154014804883e-06, "loss": 2.3886, "step": 68880 }, { "epoch": 1.55, "grad_norm": 1.5862542161534525, "learning_rate": 2.501737311900997e-06, "loss": 2.473, "step": 68890 }, { "epoch": 1.55, "grad_norm": 1.6988379669537355, "learning_rate": 2.4993216089760106e-06, "loss": 2.3523, "step": 68900 }, { "epoch": 1.55, "grad_norm": 1.799406646685786, "learning_rate": 2.4969069063520845e-06, "loss": 2.3856, "step": 68910 }, { "epoch": 1.55, "grad_norm": 2.145193851272632, "learning_rate": 2.494493204351246e-06, "loss": 2.2885, "step": 68920 }, { "epoch": 1.55, "grad_norm": 1.7206749519589573, "learning_rate": 2.49208050329539e-06, "loss": 2.3138, "step": 68930 }, { "epoch": 1.55, "grad_norm": 1.7907199195277144, "learning_rate": 2.4896688035062768e-06, "loss": 2.4533, "step": 68940 }, { "epoch": 1.55, "grad_norm": 1.7946050596235543, "learning_rate": 2.4872581053055356e-06, "loss": 2.4034, "step": 68950 }, { "epoch": 1.56, "grad_norm": 1.908080755525104, "learning_rate": 2.4848484090146597e-06, "loss": 2.3782, "step": 68960 }, { "epoch": 1.56, "grad_norm": 1.6530481804402613, "learning_rate": 2.482439714955006e-06, "loss": 2.4111, "step": 68970 }, { "epoch": 1.56, "grad_norm": 1.8970716061096835, "learning_rate": 2.4800320234478125e-06, "loss": 2.3572, "step": 68980 }, { "epoch": 1.56, "grad_norm": 1.7921169612424261, "learning_rate": 2.477625334814161e-06, "loss": 2.3855, "step": 68990 }, { "epoch": 1.56, "grad_norm": 1.9145983328166152, "learning_rate": 2.4752196493750146e-06, "loss": 2.3027, "step": 69000 }, { "epoch": 1.56, "grad_norm": 1.625234431927508, "learning_rate": 2.4728149674511992e-06, "loss": 2.4524, "step": 69010 }, { "epoch": 1.56, "grad_norm": 1.9144691121578954, "learning_rate": 2.470411289363407e-06, "loss": 2.4349, "step": 69020 }, { "epoch": 1.56, "grad_norm": 1.6697449267487199, "learning_rate": 2.4680086154321926e-06, "loss": 2.4732, "step": 69030 }, { "epoch": 1.56, "grad_norm": 1.634780251195953, "learning_rate": 2.4656069459779853e-06, "loss": 2.3994, "step": 69040 }, { "epoch": 1.56, "grad_norm": 1.494540535922145, "learning_rate": 2.4632062813210723e-06, "loss": 2.2904, "step": 69050 }, { "epoch": 1.56, "grad_norm": 1.9780863357105716, "learning_rate": 2.4608066217816094e-06, "loss": 2.4324, "step": 69060 }, { "epoch": 1.56, "grad_norm": 1.8087376055964197, "learning_rate": 2.45840796767962e-06, "loss": 2.4701, "step": 69070 }, { "epoch": 1.56, "grad_norm": 1.7765449471705543, "learning_rate": 2.4560103193349915e-06, "loss": 2.3957, "step": 69080 }, { "epoch": 1.56, "grad_norm": 1.478717861740319, "learning_rate": 2.453613677067477e-06, "loss": 2.334, "step": 69090 }, { "epoch": 1.56, "grad_norm": 1.6679172273594793, "learning_rate": 2.451218041196696e-06, "loss": 2.4074, "step": 69100 }, { "epoch": 1.56, "grad_norm": 1.7859751862541615, "learning_rate": 2.448823412042134e-06, "loss": 2.4166, "step": 69110 }, { "epoch": 1.56, "grad_norm": 1.666809424800415, "learning_rate": 2.4464297899231426e-06, "loss": 2.3543, "step": 69120 }, { "epoch": 1.56, "grad_norm": 1.8854743318494178, "learning_rate": 2.4440371751589386e-06, "loss": 2.3184, "step": 69130 }, { "epoch": 1.56, "grad_norm": 1.5931274178460995, "learning_rate": 2.4416455680686035e-06, "loss": 2.3154, "step": 69140 }, { "epoch": 1.56, "grad_norm": 1.8028835761469422, "learning_rate": 2.4392549689710875e-06, "loss": 2.4297, "step": 69150 }, { "epoch": 1.56, "grad_norm": 1.8546839273556204, "learning_rate": 2.436865378185199e-06, "loss": 2.4227, "step": 69160 }, { "epoch": 1.56, "grad_norm": 1.6516268541969261, "learning_rate": 2.4344767960296244e-06, "loss": 2.3344, "step": 69170 }, { "epoch": 1.56, "grad_norm": 1.7225473027961546, "learning_rate": 2.4320892228229053e-06, "loss": 2.367, "step": 69180 }, { "epoch": 1.56, "grad_norm": 1.5197491409817205, "learning_rate": 2.4297026588834515e-06, "loss": 2.4027, "step": 69190 }, { "epoch": 1.56, "grad_norm": 1.615637296402952, "learning_rate": 2.427317104529542e-06, "loss": 2.445, "step": 69200 }, { "epoch": 1.56, "grad_norm": 1.8864747369935233, "learning_rate": 2.4249325600793105e-06, "loss": 2.377, "step": 69210 }, { "epoch": 1.56, "grad_norm": 1.469864199138116, "learning_rate": 2.422549025850768e-06, "loss": 2.5003, "step": 69220 }, { "epoch": 1.56, "grad_norm": 1.6151880596638652, "learning_rate": 2.4201665021617816e-06, "loss": 2.2742, "step": 69230 }, { "epoch": 1.56, "grad_norm": 1.8370301507582714, "learning_rate": 2.417784989330095e-06, "loss": 2.3309, "step": 69240 }, { "epoch": 1.56, "grad_norm": 1.5511479814757654, "learning_rate": 2.415404487673306e-06, "loss": 2.3503, "step": 69250 }, { "epoch": 1.56, "grad_norm": 1.865856932696303, "learning_rate": 2.4130249975088825e-06, "loss": 2.3702, "step": 69260 }, { "epoch": 1.56, "grad_norm": 1.707268934274392, "learning_rate": 2.410646519154157e-06, "loss": 2.449, "step": 69270 }, { "epoch": 1.56, "grad_norm": 1.8124922140365345, "learning_rate": 2.4082690529263265e-06, "loss": 2.4101, "step": 69280 }, { "epoch": 1.56, "grad_norm": 1.6778634078858121, "learning_rate": 2.405892599142452e-06, "loss": 2.3768, "step": 69290 }, { "epoch": 1.56, "grad_norm": 1.6367229491910877, "learning_rate": 2.4035171581194614e-06, "loss": 2.4408, "step": 69300 }, { "epoch": 1.56, "grad_norm": 1.7972621032632747, "learning_rate": 2.4011427301741487e-06, "loss": 2.3563, "step": 69310 }, { "epoch": 1.56, "grad_norm": 1.588491104203438, "learning_rate": 2.3987693156231685e-06, "loss": 2.3835, "step": 69320 }, { "epoch": 1.56, "grad_norm": 1.7071724939202806, "learning_rate": 2.3963969147830435e-06, "loss": 2.4539, "step": 69330 }, { "epoch": 1.56, "grad_norm": 1.569432256168127, "learning_rate": 2.394025527970161e-06, "loss": 2.466, "step": 69340 }, { "epoch": 1.56, "grad_norm": 1.6838319814241218, "learning_rate": 2.3916551555007705e-06, "loss": 2.4179, "step": 69350 }, { "epoch": 1.56, "grad_norm": 1.680669447100235, "learning_rate": 2.389285797690989e-06, "loss": 2.4001, "step": 69360 }, { "epoch": 1.56, "grad_norm": 1.6934957558337775, "learning_rate": 2.3869174548567996e-06, "loss": 2.3054, "step": 69370 }, { "epoch": 1.56, "grad_norm": 1.714177537185408, "learning_rate": 2.384550127314047e-06, "loss": 2.3322, "step": 69380 }, { "epoch": 1.56, "grad_norm": 1.5709292305586835, "learning_rate": 2.3821838153784404e-06, "loss": 2.376, "step": 69390 }, { "epoch": 1.57, "grad_norm": 1.6647530948919964, "learning_rate": 2.3798185193655553e-06, "loss": 2.2876, "step": 69400 }, { "epoch": 1.57, "grad_norm": 1.9077503019676747, "learning_rate": 2.377454239590834e-06, "loss": 2.4427, "step": 69410 }, { "epoch": 1.57, "grad_norm": 1.6984507812127931, "learning_rate": 2.3750909763695696e-06, "loss": 2.4128, "step": 69420 }, { "epoch": 1.57, "grad_norm": 1.7791217748533321, "learning_rate": 2.372728730016941e-06, "loss": 2.3367, "step": 69430 }, { "epoch": 1.57, "grad_norm": 1.8724462774630237, "learning_rate": 2.3703675008479766e-06, "loss": 2.3542, "step": 69440 }, { "epoch": 1.57, "grad_norm": 1.8163969934566304, "learning_rate": 2.368007289177574e-06, "loss": 2.3964, "step": 69450 }, { "epoch": 1.57, "grad_norm": 1.6356084468640644, "learning_rate": 2.365648095320494e-06, "loss": 2.3215, "step": 69460 }, { "epoch": 1.57, "grad_norm": 1.8158488380417075, "learning_rate": 2.363289919591362e-06, "loss": 2.4208, "step": 69470 }, { "epoch": 1.57, "grad_norm": 1.9088628508490533, "learning_rate": 2.3609327623046673e-06, "loss": 2.347, "step": 69480 }, { "epoch": 1.57, "grad_norm": 1.7311315706341817, "learning_rate": 2.3585766237747643e-06, "loss": 2.4431, "step": 69490 }, { "epoch": 1.57, "grad_norm": 1.794066785350329, "learning_rate": 2.3562215043158675e-06, "loss": 2.3836, "step": 69500 }, { "epoch": 1.57, "grad_norm": 1.8597213617278432, "learning_rate": 2.353867404242065e-06, "loss": 2.3147, "step": 69510 }, { "epoch": 1.57, "grad_norm": 1.768223114755675, "learning_rate": 2.3515143238673044e-06, "loss": 2.4382, "step": 69520 }, { "epoch": 1.57, "grad_norm": 1.5250302408677696, "learning_rate": 2.349162263505388e-06, "loss": 2.5111, "step": 69530 }, { "epoch": 1.57, "grad_norm": 1.7605596585818355, "learning_rate": 2.3468112234699926e-06, "loss": 2.381, "step": 69540 }, { "epoch": 1.57, "grad_norm": 1.575637593743946, "learning_rate": 2.344461204074657e-06, "loss": 2.4366, "step": 69550 }, { "epoch": 1.57, "grad_norm": 1.811479370741278, "learning_rate": 2.3421122056327816e-06, "loss": 2.325, "step": 69560 }, { "epoch": 1.57, "grad_norm": 1.5643172524161009, "learning_rate": 2.339764228457636e-06, "loss": 2.4138, "step": 69570 }, { "epoch": 1.57, "grad_norm": 1.7373047345442598, "learning_rate": 2.3374172728623467e-06, "loss": 2.3958, "step": 69580 }, { "epoch": 1.57, "grad_norm": 1.609779160741839, "learning_rate": 2.335071339159909e-06, "loss": 2.4574, "step": 69590 }, { "epoch": 1.57, "grad_norm": 1.728410497840085, "learning_rate": 2.3327264276631776e-06, "loss": 2.3309, "step": 69600 }, { "epoch": 1.57, "grad_norm": 1.5208704811915181, "learning_rate": 2.3303825386848754e-06, "loss": 2.4685, "step": 69610 }, { "epoch": 1.57, "grad_norm": 1.6290766167525972, "learning_rate": 2.3280396725375855e-06, "loss": 2.4706, "step": 69620 }, { "epoch": 1.57, "grad_norm": 1.7166984823328821, "learning_rate": 2.325697829533755e-06, "loss": 2.3372, "step": 69630 }, { "epoch": 1.57, "grad_norm": 1.6045436105156397, "learning_rate": 2.323357009985697e-06, "loss": 2.4414, "step": 69640 }, { "epoch": 1.57, "grad_norm": 1.6290075427208472, "learning_rate": 2.3210172142055853e-06, "loss": 2.449, "step": 69650 }, { "epoch": 1.57, "grad_norm": 1.5627473650275054, "learning_rate": 2.31867844250546e-06, "loss": 2.5332, "step": 69660 }, { "epoch": 1.57, "grad_norm": 1.7527625031952738, "learning_rate": 2.316340695197221e-06, "loss": 2.2497, "step": 69670 }, { "epoch": 1.57, "grad_norm": 1.7179468205691795, "learning_rate": 2.3140039725926357e-06, "loss": 2.3868, "step": 69680 }, { "epoch": 1.57, "grad_norm": 1.7662351429807928, "learning_rate": 2.31166827500333e-06, "loss": 2.3609, "step": 69690 }, { "epoch": 1.57, "grad_norm": 1.7300123669889456, "learning_rate": 2.3093336027407953e-06, "loss": 2.3277, "step": 69700 }, { "epoch": 1.57, "grad_norm": 1.9118547767996048, "learning_rate": 2.3069999561163924e-06, "loss": 2.3847, "step": 69710 }, { "epoch": 1.57, "grad_norm": 1.6974655274401773, "learning_rate": 2.304667335441335e-06, "loss": 2.4212, "step": 69720 }, { "epoch": 1.57, "grad_norm": 1.8789976419425714, "learning_rate": 2.302335741026707e-06, "loss": 2.4293, "step": 69730 }, { "epoch": 1.57, "grad_norm": 1.6971771141122025, "learning_rate": 2.300005173183454e-06, "loss": 2.4388, "step": 69740 }, { "epoch": 1.57, "grad_norm": 1.5712057007327518, "learning_rate": 2.2976756322223793e-06, "loss": 2.3948, "step": 69750 }, { "epoch": 1.57, "grad_norm": 1.5576877140086371, "learning_rate": 2.295347118454153e-06, "loss": 2.3685, "step": 69760 }, { "epoch": 1.57, "grad_norm": 1.7728411948484328, "learning_rate": 2.293019632189315e-06, "loss": 2.2608, "step": 69770 }, { "epoch": 1.57, "grad_norm": 1.7708027494984984, "learning_rate": 2.29069317373826e-06, "loss": 2.33, "step": 69780 }, { "epoch": 1.57, "grad_norm": 1.8032639962453663, "learning_rate": 2.288367743411247e-06, "loss": 2.4184, "step": 69790 }, { "epoch": 1.57, "grad_norm": 1.6953783716756854, "learning_rate": 2.286043341518399e-06, "loss": 2.2964, "step": 69800 }, { "epoch": 1.57, "grad_norm": 1.7357546219806188, "learning_rate": 2.283719968369701e-06, "loss": 2.34, "step": 69810 }, { "epoch": 1.57, "grad_norm": 1.6903357293935075, "learning_rate": 2.2813976242750025e-06, "loss": 2.3484, "step": 69820 }, { "epoch": 1.57, "grad_norm": 1.9101379509092558, "learning_rate": 2.2790763095440106e-06, "loss": 2.3522, "step": 69830 }, { "epoch": 1.57, "grad_norm": 1.6990057247258463, "learning_rate": 2.2767560244863075e-06, "loss": 2.3738, "step": 69840 }, { "epoch": 1.58, "grad_norm": 1.672547112328611, "learning_rate": 2.2744367694113222e-06, "loss": 2.3512, "step": 69850 }, { "epoch": 1.58, "grad_norm": 1.7347584736343749, "learning_rate": 2.2721185446283555e-06, "loss": 2.2466, "step": 69860 }, { "epoch": 1.58, "grad_norm": 1.8405254973003993, "learning_rate": 2.26980135044657e-06, "loss": 2.5155, "step": 69870 }, { "epoch": 1.58, "grad_norm": 1.7794882732071187, "learning_rate": 2.2674851871749903e-06, "loss": 2.4676, "step": 69880 }, { "epoch": 1.58, "grad_norm": 1.871659062651817, "learning_rate": 2.2651700551224986e-06, "loss": 2.3169, "step": 69890 }, { "epoch": 1.58, "grad_norm": 1.6984739207977233, "learning_rate": 2.2628559545978524e-06, "loss": 2.5495, "step": 69900 }, { "epoch": 1.58, "grad_norm": 1.8821525361989162, "learning_rate": 2.260542885909659e-06, "loss": 2.3848, "step": 69910 }, { "epoch": 1.58, "grad_norm": 1.8958479022713997, "learning_rate": 2.258230849366392e-06, "loss": 2.4191, "step": 69920 }, { "epoch": 1.58, "grad_norm": 1.556089657390781, "learning_rate": 2.2559198452763887e-06, "loss": 2.4113, "step": 69930 }, { "epoch": 1.58, "grad_norm": 1.9799218817660325, "learning_rate": 2.2536098739478485e-06, "loss": 2.308, "step": 69940 }, { "epoch": 1.58, "grad_norm": 1.6341819958604993, "learning_rate": 2.2513009356888314e-06, "loss": 2.4472, "step": 69950 }, { "epoch": 1.58, "grad_norm": 1.661845050540954, "learning_rate": 2.2489930308072604e-06, "loss": 2.5256, "step": 69960 }, { "epoch": 1.58, "grad_norm": 1.7348462095341874, "learning_rate": 2.2466861596109224e-06, "loss": 2.3604, "step": 69970 }, { "epoch": 1.58, "grad_norm": 1.6367627784557934, "learning_rate": 2.2443803224074635e-06, "loss": 2.4081, "step": 69980 }, { "epoch": 1.58, "grad_norm": 1.6059942498451176, "learning_rate": 2.242075519504394e-06, "loss": 2.3847, "step": 69990 }, { "epoch": 1.58, "grad_norm": 1.9706355520939551, "learning_rate": 2.2397717512090865e-06, "loss": 2.2398, "step": 70000 }, { "epoch": 1.58, "grad_norm": 1.711334167286794, "learning_rate": 2.237469017828774e-06, "loss": 2.3353, "step": 70010 }, { "epoch": 1.58, "grad_norm": 1.6506818574463478, "learning_rate": 2.2351673196705514e-06, "loss": 2.3362, "step": 70020 }, { "epoch": 1.58, "grad_norm": 1.7268093754167055, "learning_rate": 2.232866657041375e-06, "loss": 2.33, "step": 70030 }, { "epoch": 1.58, "grad_norm": 1.5507911253024307, "learning_rate": 2.2305670302480688e-06, "loss": 2.3755, "step": 70040 }, { "epoch": 1.58, "grad_norm": 1.7320839102643841, "learning_rate": 2.228268439597313e-06, "loss": 2.3889, "step": 70050 }, { "epoch": 1.58, "grad_norm": 1.498120456872771, "learning_rate": 2.2259708853956517e-06, "loss": 2.4664, "step": 70060 }, { "epoch": 1.58, "grad_norm": 1.5981276393121007, "learning_rate": 2.2236743679494865e-06, "loss": 2.3147, "step": 70070 }, { "epoch": 1.58, "grad_norm": 1.6624201546906616, "learning_rate": 2.2213788875650864e-06, "loss": 2.497, "step": 70080 }, { "epoch": 1.58, "grad_norm": 1.8347416005829693, "learning_rate": 2.2190844445485758e-06, "loss": 2.3316, "step": 70090 }, { "epoch": 1.58, "grad_norm": 1.678622906116039, "learning_rate": 2.216791039205952e-06, "loss": 2.3343, "step": 70100 }, { "epoch": 1.58, "grad_norm": 1.6984746195255953, "learning_rate": 2.214498671843064e-06, "loss": 2.5462, "step": 70110 }, { "epoch": 1.58, "grad_norm": 1.624708776181235, "learning_rate": 2.2122073427656233e-06, "loss": 2.403, "step": 70120 }, { "epoch": 1.58, "grad_norm": 1.6901826129658537, "learning_rate": 2.209917052279207e-06, "loss": 2.4155, "step": 70130 }, { "epoch": 1.58, "grad_norm": 1.5660919884256168, "learning_rate": 2.2076278006892492e-06, "loss": 2.4239, "step": 70140 }, { "epoch": 1.58, "grad_norm": 1.8059836242487741, "learning_rate": 2.205339588301051e-06, "loss": 2.4358, "step": 70150 }, { "epoch": 1.58, "grad_norm": 1.4976034734403436, "learning_rate": 2.2030524154197654e-06, "loss": 2.2373, "step": 70160 }, { "epoch": 1.58, "grad_norm": 1.7080906229081267, "learning_rate": 2.2007662823504227e-06, "loss": 2.3301, "step": 70170 }, { "epoch": 1.58, "grad_norm": 1.653531476985205, "learning_rate": 2.1984811893978976e-06, "loss": 2.3699, "step": 70180 }, { "epoch": 1.58, "grad_norm": 1.9267153842705742, "learning_rate": 2.1961971368669344e-06, "loss": 2.332, "step": 70190 }, { "epoch": 1.58, "grad_norm": 1.5736942212788503, "learning_rate": 2.193914125062139e-06, "loss": 2.3505, "step": 70200 }, { "epoch": 1.58, "grad_norm": 1.8844408845290395, "learning_rate": 2.191632154287975e-06, "loss": 2.3179, "step": 70210 }, { "epoch": 1.58, "grad_norm": 1.6688397543916331, "learning_rate": 2.1893512248487713e-06, "loss": 2.2667, "step": 70220 }, { "epoch": 1.58, "grad_norm": 1.6759882267679236, "learning_rate": 2.1870713370487127e-06, "loss": 2.4239, "step": 70230 }, { "epoch": 1.58, "grad_norm": 1.6003055889958833, "learning_rate": 2.1847924911918528e-06, "loss": 2.4827, "step": 70240 }, { "epoch": 1.58, "grad_norm": 1.8228168792049229, "learning_rate": 2.1825146875821e-06, "loss": 2.4118, "step": 70250 }, { "epoch": 1.58, "grad_norm": 1.6248403425702262, "learning_rate": 2.1802379265232254e-06, "loss": 2.3178, "step": 70260 }, { "epoch": 1.58, "grad_norm": 2.2744297287328523, "learning_rate": 2.1779622083188633e-06, "loss": 2.4332, "step": 70270 }, { "epoch": 1.58, "grad_norm": 1.924445101107723, "learning_rate": 2.1756875332725013e-06, "loss": 2.3191, "step": 70280 }, { "epoch": 1.59, "grad_norm": 2.0336973072968862, "learning_rate": 2.173413901687493e-06, "loss": 2.3522, "step": 70290 }, { "epoch": 1.59, "grad_norm": 1.76989498841975, "learning_rate": 2.17114131386706e-06, "loss": 2.5123, "step": 70300 }, { "epoch": 1.59, "grad_norm": 1.6558239581590835, "learning_rate": 2.1688697701142723e-06, "loss": 2.3755, "step": 70310 }, { "epoch": 1.59, "grad_norm": 1.5690134431650415, "learning_rate": 2.1665992707320692e-06, "loss": 2.3571, "step": 70320 }, { "epoch": 1.59, "grad_norm": 1.61425659963227, "learning_rate": 2.1643298160232463e-06, "loss": 2.4819, "step": 70330 }, { "epoch": 1.59, "grad_norm": 1.7580709143651887, "learning_rate": 2.162061406290461e-06, "loss": 2.3652, "step": 70340 }, { "epoch": 1.59, "grad_norm": 1.9465782403024794, "learning_rate": 2.159794041836232e-06, "loss": 2.3112, "step": 70350 }, { "epoch": 1.59, "grad_norm": 1.6208403718335316, "learning_rate": 2.1575277229629355e-06, "loss": 2.3081, "step": 70360 }, { "epoch": 1.59, "grad_norm": 1.5611193770042402, "learning_rate": 2.155262449972816e-06, "loss": 2.292, "step": 70370 }, { "epoch": 1.59, "grad_norm": 1.534888977031446, "learning_rate": 2.1529982231679747e-06, "loss": 2.4463, "step": 70380 }, { "epoch": 1.59, "grad_norm": 1.7082725590602814, "learning_rate": 2.150735042850366e-06, "loss": 2.4568, "step": 70390 }, { "epoch": 1.59, "grad_norm": 1.6683745570231654, "learning_rate": 2.1484729093218125e-06, "loss": 2.3506, "step": 70400 }, { "epoch": 1.59, "grad_norm": 1.4671374028493882, "learning_rate": 2.146211822883997e-06, "loss": 2.4143, "step": 70410 }, { "epoch": 1.59, "grad_norm": 1.8429787345088307, "learning_rate": 2.1439517838384583e-06, "loss": 2.4455, "step": 70420 }, { "epoch": 1.59, "grad_norm": 1.6899109684008058, "learning_rate": 2.1416927924866025e-06, "loss": 2.3067, "step": 70430 }, { "epoch": 1.59, "grad_norm": 1.861413063207637, "learning_rate": 2.1394348491296912e-06, "loss": 2.5539, "step": 70440 }, { "epoch": 1.59, "grad_norm": 1.453091139904059, "learning_rate": 2.1371779540688455e-06, "loss": 2.3925, "step": 70450 }, { "epoch": 1.59, "grad_norm": 1.7320228572542464, "learning_rate": 2.1349221076050474e-06, "loss": 2.4067, "step": 70460 }, { "epoch": 1.59, "grad_norm": 1.6988838690433683, "learning_rate": 2.1326673100391415e-06, "loss": 2.3552, "step": 70470 }, { "epoch": 1.59, "grad_norm": 1.5675680455770318, "learning_rate": 2.13041356167183e-06, "loss": 2.3936, "step": 70480 }, { "epoch": 1.59, "grad_norm": 1.5141254781877693, "learning_rate": 2.1281608628036766e-06, "loss": 2.3338, "step": 70490 }, { "epoch": 1.59, "grad_norm": 1.7767412670561338, "learning_rate": 2.1259092137351036e-06, "loss": 2.283, "step": 70500 }, { "epoch": 1.59, "grad_norm": 1.842991946031566, "learning_rate": 2.123658614766394e-06, "loss": 2.4067, "step": 70510 }, { "epoch": 1.59, "grad_norm": 1.7206579939867408, "learning_rate": 2.1214090661976905e-06, "loss": 2.3358, "step": 70520 }, { "epoch": 1.59, "grad_norm": 1.6677551699066537, "learning_rate": 2.1191605683289983e-06, "loss": 2.3899, "step": 70530 }, { "epoch": 1.59, "grad_norm": 1.583518591054289, "learning_rate": 2.116913121460178e-06, "loss": 2.2564, "step": 70540 }, { "epoch": 1.59, "grad_norm": 1.82780466608504, "learning_rate": 2.1146667258909526e-06, "loss": 2.2753, "step": 70550 }, { "epoch": 1.59, "grad_norm": 1.7778268964159316, "learning_rate": 2.112421381920903e-06, "loss": 2.3095, "step": 70560 }, { "epoch": 1.59, "grad_norm": 1.8051088330497649, "learning_rate": 2.1101770898494757e-06, "loss": 2.3628, "step": 70570 }, { "epoch": 1.59, "grad_norm": 1.72326497466945, "learning_rate": 2.1079338499759705e-06, "loss": 2.4541, "step": 70580 }, { "epoch": 1.59, "grad_norm": 1.5627345814048914, "learning_rate": 2.105691662599548e-06, "loss": 2.443, "step": 70590 }, { "epoch": 1.59, "grad_norm": 1.8884669024042853, "learning_rate": 2.1034505280192352e-06, "loss": 2.3808, "step": 70600 }, { "epoch": 1.59, "grad_norm": 1.7589320682621854, "learning_rate": 2.1012104465339034e-06, "loss": 2.3817, "step": 70610 }, { "epoch": 1.59, "grad_norm": 1.8053662740080647, "learning_rate": 2.098971418442296e-06, "loss": 2.3878, "step": 70620 }, { "epoch": 1.59, "grad_norm": 1.8020153552386882, "learning_rate": 2.096733444043018e-06, "loss": 2.3579, "step": 70630 }, { "epoch": 1.59, "grad_norm": 1.7113372665289683, "learning_rate": 2.094496523634525e-06, "loss": 2.4007, "step": 70640 }, { "epoch": 1.59, "grad_norm": 1.7927806433308497, "learning_rate": 2.092260657515136e-06, "loss": 2.3919, "step": 70650 }, { "epoch": 1.59, "grad_norm": 1.56035746838743, "learning_rate": 2.0900258459830304e-06, "loss": 2.4398, "step": 70660 }, { "epoch": 1.59, "grad_norm": 1.5645568232472402, "learning_rate": 2.0877920893362447e-06, "loss": 2.2845, "step": 70670 }, { "epoch": 1.59, "grad_norm": 1.5830653983377019, "learning_rate": 2.0855593878726766e-06, "loss": 2.3893, "step": 70680 }, { "epoch": 1.59, "grad_norm": 1.6042513055648429, "learning_rate": 2.083327741890079e-06, "loss": 2.4309, "step": 70690 }, { "epoch": 1.59, "grad_norm": 1.7234494874430548, "learning_rate": 2.081097151686073e-06, "loss": 2.2872, "step": 70700 }, { "epoch": 1.59, "grad_norm": 1.8092728932518176, "learning_rate": 2.0788676175581336e-06, "loss": 2.3898, "step": 70710 }, { "epoch": 1.59, "grad_norm": 1.6527010335694614, "learning_rate": 2.076639139803589e-06, "loss": 2.4266, "step": 70720 }, { "epoch": 1.6, "grad_norm": 1.612404007812958, "learning_rate": 2.0744117187196345e-06, "loss": 2.3142, "step": 70730 }, { "epoch": 1.6, "grad_norm": 1.768482433735242, "learning_rate": 2.0721853546033224e-06, "loss": 2.4221, "step": 70740 }, { "epoch": 1.6, "grad_norm": 1.8130318313270073, "learning_rate": 2.0699600477515635e-06, "loss": 2.4676, "step": 70750 }, { "epoch": 1.6, "grad_norm": 1.937300264714743, "learning_rate": 2.0677357984611258e-06, "loss": 2.2938, "step": 70760 }, { "epoch": 1.6, "grad_norm": 1.6202917213454817, "learning_rate": 2.0655126070286433e-06, "loss": 2.4229, "step": 70770 }, { "epoch": 1.6, "grad_norm": 1.772666585116522, "learning_rate": 2.063290473750601e-06, "loss": 2.3569, "step": 70780 }, { "epoch": 1.6, "grad_norm": 1.9583814533398842, "learning_rate": 2.0610693989233466e-06, "loss": 2.3158, "step": 70790 }, { "epoch": 1.6, "grad_norm": 1.6582660448652486, "learning_rate": 2.0588493828430844e-06, "loss": 2.3095, "step": 70800 }, { "epoch": 1.6, "grad_norm": 1.5048207668233298, "learning_rate": 2.0566304258058833e-06, "loss": 2.4578, "step": 70810 }, { "epoch": 1.6, "grad_norm": 1.706865323103654, "learning_rate": 2.0544125281076565e-06, "loss": 2.3589, "step": 70820 }, { "epoch": 1.6, "grad_norm": 1.4545292038592956, "learning_rate": 2.0521956900441955e-06, "loss": 2.5706, "step": 70830 }, { "epoch": 1.6, "grad_norm": 1.5785490326775866, "learning_rate": 2.0499799119111376e-06, "loss": 2.333, "step": 70840 }, { "epoch": 1.6, "grad_norm": 1.7779628034257127, "learning_rate": 2.047765194003982e-06, "loss": 2.5482, "step": 70850 }, { "epoch": 1.6, "grad_norm": 1.791713466037124, "learning_rate": 2.045551536618088e-06, "loss": 2.282, "step": 70860 }, { "epoch": 1.6, "grad_norm": 1.7771724709508367, "learning_rate": 2.0433389400486703e-06, "loss": 2.3462, "step": 70870 }, { "epoch": 1.6, "grad_norm": 1.7055574912170248, "learning_rate": 2.0411274045908048e-06, "loss": 2.3144, "step": 70880 }, { "epoch": 1.6, "grad_norm": 1.7541580173407028, "learning_rate": 2.038916930539423e-06, "loss": 2.4326, "step": 70890 }, { "epoch": 1.6, "grad_norm": 1.627929458055418, "learning_rate": 2.03670751818932e-06, "loss": 2.3417, "step": 70900 }, { "epoch": 1.6, "grad_norm": 1.8844920341700047, "learning_rate": 2.0344991678351455e-06, "loss": 2.4527, "step": 70910 }, { "epoch": 1.6, "grad_norm": 1.5519025760434806, "learning_rate": 2.0322918797714097e-06, "loss": 2.3767, "step": 70920 }, { "epoch": 1.6, "grad_norm": 1.7509604081088712, "learning_rate": 2.030085654292475e-06, "loss": 2.4608, "step": 70930 }, { "epoch": 1.6, "grad_norm": 1.7180000864866865, "learning_rate": 2.0278804916925686e-06, "loss": 2.3185, "step": 70940 }, { "epoch": 1.6, "grad_norm": 1.6729117254431651, "learning_rate": 2.025676392265773e-06, "loss": 2.3611, "step": 70950 }, { "epoch": 1.6, "grad_norm": 1.7746808912763443, "learning_rate": 2.023473356306034e-06, "loss": 2.3316, "step": 70960 }, { "epoch": 1.6, "grad_norm": 1.6708767111846525, "learning_rate": 2.0212713841071497e-06, "loss": 2.3087, "step": 70970 }, { "epoch": 1.6, "grad_norm": 1.5377413645207763, "learning_rate": 2.0190704759627766e-06, "loss": 2.544, "step": 70980 }, { "epoch": 1.6, "grad_norm": 1.6256014590578203, "learning_rate": 2.0168706321664334e-06, "loss": 2.3671, "step": 70990 }, { "epoch": 1.6, "grad_norm": 1.8461911606969659, "learning_rate": 2.0146718530114928e-06, "loss": 2.4084, "step": 71000 }, { "epoch": 1.6, "grad_norm": 1.5916385561596378, "learning_rate": 2.0124741387911874e-06, "loss": 2.4003, "step": 71010 }, { "epoch": 1.6, "grad_norm": 1.7647981014998024, "learning_rate": 2.0102774897986077e-06, "loss": 2.4101, "step": 71020 }, { "epoch": 1.6, "grad_norm": 1.686683986985443, "learning_rate": 2.008081906326701e-06, "loss": 2.4224, "step": 71030 }, { "epoch": 1.6, "grad_norm": 1.64862327771433, "learning_rate": 2.005887388668274e-06, "loss": 2.4703, "step": 71040 }, { "epoch": 1.6, "grad_norm": 1.588156626240932, "learning_rate": 2.0036939371159913e-06, "loss": 2.4189, "step": 71050 }, { "epoch": 1.6, "grad_norm": 1.8812990064273025, "learning_rate": 2.0015015519623747e-06, "loss": 2.3976, "step": 71060 }, { "epoch": 1.6, "grad_norm": 1.6954039543934127, "learning_rate": 1.9993102334998026e-06, "loss": 2.4349, "step": 71070 }, { "epoch": 1.6, "grad_norm": 1.5872871603724221, "learning_rate": 1.997119982020513e-06, "loss": 2.4037, "step": 71080 }, { "epoch": 1.6, "grad_norm": 1.8940458248026975, "learning_rate": 1.9949307978165977e-06, "loss": 2.4181, "step": 71090 }, { "epoch": 1.6, "grad_norm": 1.7641114899466124, "learning_rate": 1.9927426811800167e-06, "loss": 2.3336, "step": 71100 }, { "epoch": 1.6, "grad_norm": 1.6805687360439208, "learning_rate": 1.9905556324025753e-06, "loss": 2.3153, "step": 71110 }, { "epoch": 1.6, "grad_norm": 1.7916020094566811, "learning_rate": 1.988369651775942e-06, "loss": 2.4124, "step": 71120 }, { "epoch": 1.6, "grad_norm": 1.7245865291855969, "learning_rate": 1.9861847395916436e-06, "loss": 2.3476, "step": 71130 }, { "epoch": 1.6, "grad_norm": 1.590453777091813, "learning_rate": 1.984000896141064e-06, "loss": 2.4204, "step": 71140 }, { "epoch": 1.6, "grad_norm": 1.669574297399942, "learning_rate": 1.9818181217154364e-06, "loss": 2.1921, "step": 71150 }, { "epoch": 1.6, "grad_norm": 1.8971284883685038, "learning_rate": 1.9796364166058666e-06, "loss": 2.4112, "step": 71160 }, { "epoch": 1.6, "grad_norm": 1.8120399626291301, "learning_rate": 1.977455781103307e-06, "loss": 2.3619, "step": 71170 }, { "epoch": 1.61, "grad_norm": 1.6184685560625014, "learning_rate": 1.975276215498572e-06, "loss": 2.3141, "step": 71180 }, { "epoch": 1.61, "grad_norm": 1.8538650476559484, "learning_rate": 1.9730977200823286e-06, "loss": 2.3825, "step": 71190 }, { "epoch": 1.61, "grad_norm": 1.9946441021378742, "learning_rate": 1.9709202951451057e-06, "loss": 2.5317, "step": 71200 }, { "epoch": 1.61, "grad_norm": 1.6829508749909408, "learning_rate": 1.968743940977288e-06, "loss": 2.2794, "step": 71210 }, { "epoch": 1.61, "grad_norm": 1.5700048604232137, "learning_rate": 1.9665686578691135e-06, "loss": 2.3544, "step": 71220 }, { "epoch": 1.61, "grad_norm": 1.631530530930928, "learning_rate": 1.9643944461106866e-06, "loss": 2.3698, "step": 71230 }, { "epoch": 1.61, "grad_norm": 1.7284341127120817, "learning_rate": 1.9622213059919617e-06, "loss": 2.4069, "step": 71240 }, { "epoch": 1.61, "grad_norm": 1.7360415592514704, "learning_rate": 1.9600492378027535e-06, "loss": 2.2476, "step": 71250 }, { "epoch": 1.61, "grad_norm": 1.7290599920259608, "learning_rate": 1.957878241832727e-06, "loss": 2.27, "step": 71260 }, { "epoch": 1.61, "grad_norm": 1.7246916686206926, "learning_rate": 1.955708318371411e-06, "loss": 2.4108, "step": 71270 }, { "epoch": 1.61, "grad_norm": 1.7317107683341304, "learning_rate": 1.953539467708191e-06, "loss": 2.401, "step": 71280 }, { "epoch": 1.61, "grad_norm": 1.8241517751601002, "learning_rate": 1.9513716901323054e-06, "loss": 2.432, "step": 71290 }, { "epoch": 1.61, "grad_norm": 1.6511101228797704, "learning_rate": 1.9492049859328553e-06, "loss": 2.4066, "step": 71300 }, { "epoch": 1.61, "grad_norm": 1.7405405623501964, "learning_rate": 1.9470393553987953e-06, "loss": 2.5105, "step": 71310 }, { "epoch": 1.61, "grad_norm": 1.749978737621208, "learning_rate": 1.9448747988189365e-06, "loss": 2.3495, "step": 71320 }, { "epoch": 1.61, "grad_norm": 1.5495445234171155, "learning_rate": 1.9427113164819457e-06, "loss": 2.4547, "step": 71330 }, { "epoch": 1.61, "grad_norm": 1.6849436250527057, "learning_rate": 1.9405489086763485e-06, "loss": 2.2767, "step": 71340 }, { "epoch": 1.61, "grad_norm": 1.771366918053845, "learning_rate": 1.938387575690528e-06, "loss": 2.3128, "step": 71350 }, { "epoch": 1.61, "grad_norm": 1.7663857911057796, "learning_rate": 1.936227317812721e-06, "loss": 2.414, "step": 71360 }, { "epoch": 1.61, "grad_norm": 1.698284122202059, "learning_rate": 1.934068135331023e-06, "loss": 2.3907, "step": 71370 }, { "epoch": 1.61, "grad_norm": 1.6005346450343954, "learning_rate": 1.931910028533386e-06, "loss": 2.3279, "step": 71380 }, { "epoch": 1.61, "grad_norm": 1.8141105543470903, "learning_rate": 1.929752997707618e-06, "loss": 2.4342, "step": 71390 }, { "epoch": 1.61, "grad_norm": 1.5232927213328264, "learning_rate": 1.9275970431413837e-06, "loss": 2.2808, "step": 71400 }, { "epoch": 1.61, "grad_norm": 1.902414406858178, "learning_rate": 1.925442165122204e-06, "loss": 2.3641, "step": 71410 }, { "epoch": 1.61, "grad_norm": 1.8173881378061478, "learning_rate": 1.9232883639374542e-06, "loss": 2.404, "step": 71420 }, { "epoch": 1.61, "grad_norm": 1.5420993633725972, "learning_rate": 1.921135639874373e-06, "loss": 2.4723, "step": 71430 }, { "epoch": 1.61, "grad_norm": 1.92568108067463, "learning_rate": 1.9189839932200494e-06, "loss": 2.5444, "step": 71440 }, { "epoch": 1.61, "grad_norm": 1.6620348199233559, "learning_rate": 1.9168334242614284e-06, "loss": 2.2841, "step": 71450 }, { "epoch": 1.61, "grad_norm": 1.9426303117352737, "learning_rate": 1.9146839332853164e-06, "loss": 2.4207, "step": 71460 }, { "epoch": 1.61, "grad_norm": 1.7944574455427595, "learning_rate": 1.912535520578366e-06, "loss": 2.3571, "step": 71470 }, { "epoch": 1.61, "grad_norm": 1.7045234647022407, "learning_rate": 1.9103881864270944e-06, "loss": 2.4279, "step": 71480 }, { "epoch": 1.61, "grad_norm": 1.8210773784428622, "learning_rate": 1.9082419311178777e-06, "loss": 2.4493, "step": 71490 }, { "epoch": 1.61, "grad_norm": 1.7625531962714371, "learning_rate": 1.9060967549369414e-06, "loss": 2.4453, "step": 71500 }, { "epoch": 1.61, "grad_norm": 1.7176395507016098, "learning_rate": 1.9039526581703683e-06, "loss": 2.3874, "step": 71510 }, { "epoch": 1.61, "grad_norm": 1.6159882108812371, "learning_rate": 1.9018096411040987e-06, "loss": 2.3974, "step": 71520 }, { "epoch": 1.61, "grad_norm": 1.6477393771351176, "learning_rate": 1.8996677040239286e-06, "loss": 2.3489, "step": 71530 }, { "epoch": 1.61, "grad_norm": 1.7210355544732654, "learning_rate": 1.8975268472155084e-06, "loss": 2.3164, "step": 71540 }, { "epoch": 1.61, "grad_norm": 1.8959469389129882, "learning_rate": 1.8953870709643474e-06, "loss": 2.3594, "step": 71550 }, { "epoch": 1.61, "grad_norm": 1.7878955520787623, "learning_rate": 1.893248375555805e-06, "loss": 2.4294, "step": 71560 }, { "epoch": 1.61, "grad_norm": 1.6590069766022277, "learning_rate": 1.8911107612751101e-06, "loss": 2.3623, "step": 71570 }, { "epoch": 1.61, "grad_norm": 1.6611168265139056, "learning_rate": 1.8889742284073298e-06, "loss": 2.4487, "step": 71580 }, { "epoch": 1.61, "grad_norm": 1.7578954010671701, "learning_rate": 1.8868387772373975e-06, "loss": 2.4609, "step": 71590 }, { "epoch": 1.61, "grad_norm": 1.868705534445761, "learning_rate": 1.8847044080500986e-06, "loss": 2.3957, "step": 71600 }, { "epoch": 1.61, "grad_norm": 1.8727679283675642, "learning_rate": 1.882571121130078e-06, "loss": 2.4229, "step": 71610 }, { "epoch": 1.62, "grad_norm": 1.794923054183533, "learning_rate": 1.8804389167618299e-06, "loss": 2.2656, "step": 71620 }, { "epoch": 1.62, "grad_norm": 1.9263399437690771, "learning_rate": 1.8783077952297135e-06, "loss": 2.2912, "step": 71630 }, { "epoch": 1.62, "grad_norm": 1.6225143667981266, "learning_rate": 1.8761777568179352e-06, "loss": 2.4834, "step": 71640 }, { "epoch": 1.62, "grad_norm": 1.6646261669924485, "learning_rate": 1.8740488018105606e-06, "loss": 2.3372, "step": 71650 }, { "epoch": 1.62, "grad_norm": 1.7011501958199542, "learning_rate": 1.8719209304915099e-06, "loss": 2.3434, "step": 71660 }, { "epoch": 1.62, "grad_norm": 1.69709166096793, "learning_rate": 1.8697941431445588e-06, "loss": 2.4259, "step": 71670 }, { "epoch": 1.62, "grad_norm": 1.7479725887874449, "learning_rate": 1.8676684400533385e-06, "loss": 2.4044, "step": 71680 }, { "epoch": 1.62, "grad_norm": 1.614542789958942, "learning_rate": 1.8655438215013366e-06, "loss": 2.3856, "step": 71690 }, { "epoch": 1.62, "grad_norm": 1.6266731961166714, "learning_rate": 1.8634202877718944e-06, "loss": 2.397, "step": 71700 }, { "epoch": 1.62, "grad_norm": 1.864030606455131, "learning_rate": 1.8612978391482106e-06, "loss": 2.4466, "step": 71710 }, { "epoch": 1.62, "grad_norm": 1.6381543622210346, "learning_rate": 1.8591764759133357e-06, "loss": 2.3981, "step": 71720 }, { "epoch": 1.62, "grad_norm": 1.567179615529985, "learning_rate": 1.85705619835018e-06, "loss": 2.4044, "step": 71730 }, { "epoch": 1.62, "grad_norm": 1.6262666137985249, "learning_rate": 1.8549370067415051e-06, "loss": 2.2664, "step": 71740 }, { "epoch": 1.62, "grad_norm": 1.568661073460585, "learning_rate": 1.8528189013699305e-06, "loss": 2.3377, "step": 71750 }, { "epoch": 1.62, "grad_norm": 1.6287240520633202, "learning_rate": 1.8507018825179257e-06, "loss": 2.3514, "step": 71760 }, { "epoch": 1.62, "grad_norm": 1.7441033351748458, "learning_rate": 1.8485859504678271e-06, "loss": 2.3624, "step": 71770 }, { "epoch": 1.62, "grad_norm": 1.7453504779327684, "learning_rate": 1.8464711055018125e-06, "loss": 2.4433, "step": 71780 }, { "epoch": 1.62, "grad_norm": 1.8355498073765715, "learning_rate": 1.844357347901926e-06, "loss": 2.2937, "step": 71790 }, { "epoch": 1.62, "grad_norm": 1.8830076655154422, "learning_rate": 1.8422446779500547e-06, "loss": 2.2699, "step": 71800 }, { "epoch": 1.62, "grad_norm": 1.7883572297789727, "learning_rate": 1.8401330959279494e-06, "loss": 2.3955, "step": 71810 }, { "epoch": 1.62, "grad_norm": 1.5989694757837754, "learning_rate": 1.838022602117212e-06, "loss": 2.3969, "step": 71820 }, { "epoch": 1.62, "grad_norm": 1.7828849835538332, "learning_rate": 1.8359131967993072e-06, "loss": 2.4312, "step": 71830 }, { "epoch": 1.62, "grad_norm": 1.6965110435431983, "learning_rate": 1.8338048802555431e-06, "loss": 2.3223, "step": 71840 }, { "epoch": 1.62, "grad_norm": 1.6492971726300374, "learning_rate": 1.8316976527670893e-06, "loss": 2.4292, "step": 71850 }, { "epoch": 1.62, "grad_norm": 1.8187939695315456, "learning_rate": 1.8295915146149679e-06, "loss": 2.3485, "step": 71860 }, { "epoch": 1.62, "grad_norm": 1.7221361262591133, "learning_rate": 1.8274864660800572e-06, "loss": 2.4369, "step": 71870 }, { "epoch": 1.62, "grad_norm": 2.0742116715159598, "learning_rate": 1.8253825074430898e-06, "loss": 2.375, "step": 71880 }, { "epoch": 1.62, "grad_norm": 1.9133133876371666, "learning_rate": 1.8232796389846519e-06, "loss": 2.5265, "step": 71890 }, { "epoch": 1.62, "grad_norm": 1.9599025716111813, "learning_rate": 1.8211778609851838e-06, "loss": 2.2621, "step": 71900 }, { "epoch": 1.62, "grad_norm": 1.903413477019184, "learning_rate": 1.8190771737249846e-06, "loss": 2.4122, "step": 71910 }, { "epoch": 1.62, "grad_norm": 1.5650314369351548, "learning_rate": 1.8169775774842024e-06, "loss": 2.4402, "step": 71920 }, { "epoch": 1.62, "grad_norm": 1.45864973786402, "learning_rate": 1.8148790725428423e-06, "loss": 2.5191, "step": 71930 }, { "epoch": 1.62, "grad_norm": 1.8957726133004515, "learning_rate": 1.8127816591807656e-06, "loss": 2.3359, "step": 71940 }, { "epoch": 1.62, "grad_norm": 1.7828259089687892, "learning_rate": 1.8106853376776822e-06, "loss": 2.2381, "step": 71950 }, { "epoch": 1.62, "grad_norm": 1.682252386748874, "learning_rate": 1.8085901083131674e-06, "loss": 2.3396, "step": 71960 }, { "epoch": 1.62, "grad_norm": 1.7732355491295615, "learning_rate": 1.8064959713666398e-06, "loss": 2.3181, "step": 71970 }, { "epoch": 1.62, "grad_norm": 1.8157252773542518, "learning_rate": 1.8044029271173756e-06, "loss": 2.4239, "step": 71980 }, { "epoch": 1.62, "grad_norm": 1.6861423611898332, "learning_rate": 1.802310975844509e-06, "loss": 2.5374, "step": 71990 }, { "epoch": 1.62, "grad_norm": 1.8344773246095145, "learning_rate": 1.8002201178270262e-06, "loss": 2.4037, "step": 72000 }, { "epoch": 1.62, "grad_norm": 1.4788661515998605, "learning_rate": 1.7981303533437589e-06, "loss": 2.2818, "step": 72010 }, { "epoch": 1.62, "grad_norm": 1.775341823714602, "learning_rate": 1.7960416826734105e-06, "loss": 2.4185, "step": 72020 }, { "epoch": 1.62, "grad_norm": 1.8036759218279321, "learning_rate": 1.7939541060945243e-06, "loss": 2.3134, "step": 72030 }, { "epoch": 1.62, "grad_norm": 1.7269054328084703, "learning_rate": 1.7918676238855049e-06, "loss": 2.3055, "step": 72040 }, { "epoch": 1.62, "grad_norm": 1.6920372453966013, "learning_rate": 1.7897822363246065e-06, "loss": 2.4766, "step": 72050 }, { "epoch": 1.63, "grad_norm": 1.5130762623658716, "learning_rate": 1.78769794368994e-06, "loss": 2.5215, "step": 72060 }, { "epoch": 1.63, "grad_norm": 1.690295198780831, "learning_rate": 1.7856147462594697e-06, "loss": 2.3476, "step": 72070 }, { "epoch": 1.63, "grad_norm": 1.5899874789062556, "learning_rate": 1.7835326443110146e-06, "loss": 2.4271, "step": 72080 }, { "epoch": 1.63, "grad_norm": 1.9269134617642298, "learning_rate": 1.7814516381222424e-06, "loss": 2.4235, "step": 72090 }, { "epoch": 1.63, "grad_norm": 1.648999810496459, "learning_rate": 1.7793717279706857e-06, "loss": 2.2365, "step": 72100 }, { "epoch": 1.63, "grad_norm": 1.816004685993584, "learning_rate": 1.7772929141337226e-06, "loss": 2.3417, "step": 72110 }, { "epoch": 1.63, "grad_norm": 1.7717196355901923, "learning_rate": 1.7752151968885833e-06, "loss": 2.2951, "step": 72120 }, { "epoch": 1.63, "grad_norm": 1.6227480396667582, "learning_rate": 1.7731385765123577e-06, "loss": 2.3903, "step": 72130 }, { "epoch": 1.63, "grad_norm": 1.9019643044400623, "learning_rate": 1.771063053281985e-06, "loss": 2.4756, "step": 72140 }, { "epoch": 1.63, "grad_norm": 1.7639759937256334, "learning_rate": 1.7689886274742585e-06, "loss": 2.2425, "step": 72150 }, { "epoch": 1.63, "grad_norm": 1.64103315271616, "learning_rate": 1.766915299365831e-06, "loss": 2.4441, "step": 72160 }, { "epoch": 1.63, "grad_norm": 1.474789665077329, "learning_rate": 1.7648430692332041e-06, "loss": 2.3217, "step": 72170 }, { "epoch": 1.63, "grad_norm": 1.688085932530252, "learning_rate": 1.7627719373527308e-06, "loss": 2.3706, "step": 72180 }, { "epoch": 1.63, "grad_norm": 1.5512583527516812, "learning_rate": 1.7607019040006211e-06, "loss": 2.5051, "step": 72190 }, { "epoch": 1.63, "grad_norm": 1.7057145693078737, "learning_rate": 1.758632969452937e-06, "loss": 2.4049, "step": 72200 }, { "epoch": 1.63, "grad_norm": 1.6980054310059374, "learning_rate": 1.756565133985596e-06, "loss": 2.4609, "step": 72210 }, { "epoch": 1.63, "grad_norm": 1.607350287299794, "learning_rate": 1.7544983978743668e-06, "loss": 2.3935, "step": 72220 }, { "epoch": 1.63, "grad_norm": 1.9717575271977181, "learning_rate": 1.752432761394871e-06, "loss": 2.4699, "step": 72230 }, { "epoch": 1.63, "grad_norm": 1.7375089519848257, "learning_rate": 1.7503682248225862e-06, "loss": 2.4064, "step": 72240 }, { "epoch": 1.63, "grad_norm": 2.069597721663079, "learning_rate": 1.748304788432842e-06, "loss": 2.4759, "step": 72250 }, { "epoch": 1.63, "grad_norm": 1.6847060200611086, "learning_rate": 1.7462424525008215e-06, "loss": 2.3773, "step": 72260 }, { "epoch": 1.63, "grad_norm": 1.7275561762818432, "learning_rate": 1.7441812173015582e-06, "loss": 2.3905, "step": 72270 }, { "epoch": 1.63, "grad_norm": 1.562622769719216, "learning_rate": 1.7421210831099444e-06, "loss": 2.3783, "step": 72280 }, { "epoch": 1.63, "grad_norm": 1.5618929633201573, "learning_rate": 1.7400620502007193e-06, "loss": 2.4037, "step": 72290 }, { "epoch": 1.63, "grad_norm": 1.612143447819404, "learning_rate": 1.738004118848482e-06, "loss": 2.3014, "step": 72300 }, { "epoch": 1.63, "grad_norm": 1.9536171354269651, "learning_rate": 1.7359472893276797e-06, "loss": 2.376, "step": 72310 }, { "epoch": 1.63, "grad_norm": 1.6069301762683896, "learning_rate": 1.7338915619126162e-06, "loss": 2.4205, "step": 72320 }, { "epoch": 1.63, "grad_norm": 1.9117915047369574, "learning_rate": 1.7318369368774412e-06, "loss": 2.4268, "step": 72330 }, { "epoch": 1.63, "grad_norm": 1.7228202411167126, "learning_rate": 1.7297834144961656e-06, "loss": 2.4239, "step": 72340 }, { "epoch": 1.63, "grad_norm": 1.5750681523041592, "learning_rate": 1.727730995042647e-06, "loss": 2.2742, "step": 72350 }, { "epoch": 1.63, "grad_norm": 1.6111673277987189, "learning_rate": 1.7256796787906039e-06, "loss": 2.31, "step": 72360 }, { "epoch": 1.63, "grad_norm": 1.5115442960425258, "learning_rate": 1.723629466013601e-06, "loss": 2.4224, "step": 72370 }, { "epoch": 1.63, "grad_norm": 1.6470160006486658, "learning_rate": 1.7215803569850564e-06, "loss": 2.327, "step": 72380 }, { "epoch": 1.63, "grad_norm": 1.5980979480389228, "learning_rate": 1.7195323519782425e-06, "loss": 2.3396, "step": 72390 }, { "epoch": 1.63, "grad_norm": 1.6303772827458594, "learning_rate": 1.7174854512662842e-06, "loss": 2.3657, "step": 72400 }, { "epoch": 1.63, "grad_norm": 1.7107093115866803, "learning_rate": 1.7154396551221598e-06, "loss": 2.2793, "step": 72410 }, { "epoch": 1.63, "grad_norm": 1.8147826851193603, "learning_rate": 1.7133949638186963e-06, "loss": 2.3502, "step": 72420 }, { "epoch": 1.63, "grad_norm": 1.724866405850948, "learning_rate": 1.711351377628584e-06, "loss": 2.2221, "step": 72430 }, { "epoch": 1.63, "grad_norm": 1.581481737167423, "learning_rate": 1.709308896824351e-06, "loss": 2.4291, "step": 72440 }, { "epoch": 1.63, "grad_norm": 1.6538872265083018, "learning_rate": 1.7072675216783884e-06, "loss": 2.4215, "step": 72450 }, { "epoch": 1.63, "grad_norm": 1.688413341869409, "learning_rate": 1.7052272524629376e-06, "loss": 2.438, "step": 72460 }, { "epoch": 1.63, "grad_norm": 1.7169397466438714, "learning_rate": 1.7031880894500895e-06, "loss": 2.3725, "step": 72470 }, { "epoch": 1.63, "grad_norm": 1.6914410005410419, "learning_rate": 1.7011500329117892e-06, "loss": 2.395, "step": 72480 }, { "epoch": 1.63, "grad_norm": 1.728629250117134, "learning_rate": 1.6991130831198387e-06, "loss": 2.3884, "step": 72490 }, { "epoch": 1.63, "grad_norm": 1.654336254629587, "learning_rate": 1.6970772403458868e-06, "loss": 2.3734, "step": 72500 }, { "epoch": 1.64, "grad_norm": 1.862262946626104, "learning_rate": 1.695042504861436e-06, "loss": 2.303, "step": 72510 }, { "epoch": 1.64, "grad_norm": 1.664226399113283, "learning_rate": 1.6930088769378416e-06, "loss": 2.3768, "step": 72520 }, { "epoch": 1.64, "grad_norm": 1.8928913570089478, "learning_rate": 1.690976356846311e-06, "loss": 2.3215, "step": 72530 }, { "epoch": 1.64, "grad_norm": 1.7414675190100672, "learning_rate": 1.6889449448579042e-06, "loss": 2.348, "step": 72540 }, { "epoch": 1.64, "grad_norm": 1.5418077385784321, "learning_rate": 1.6869146412435323e-06, "loss": 2.4302, "step": 72550 }, { "epoch": 1.64, "grad_norm": 1.779040468877659, "learning_rate": 1.6848854462739606e-06, "loss": 2.4405, "step": 72560 }, { "epoch": 1.64, "grad_norm": 1.7078344284821516, "learning_rate": 1.682857360219805e-06, "loss": 2.4437, "step": 72570 }, { "epoch": 1.64, "grad_norm": 1.8206207705734108, "learning_rate": 1.6808303833515338e-06, "loss": 2.273, "step": 72580 }, { "epoch": 1.64, "grad_norm": 1.5882557818893719, "learning_rate": 1.6788045159394672e-06, "loss": 2.4236, "step": 72590 }, { "epoch": 1.64, "grad_norm": 1.6338934611405331, "learning_rate": 1.676779758253778e-06, "loss": 2.2522, "step": 72600 }, { "epoch": 1.64, "grad_norm": 1.514231535281391, "learning_rate": 1.6747561105644906e-06, "loss": 2.3104, "step": 72610 }, { "epoch": 1.64, "grad_norm": 1.7927610181158575, "learning_rate": 1.6727335731414795e-06, "loss": 2.2962, "step": 72620 }, { "epoch": 1.64, "grad_norm": 1.976030169643791, "learning_rate": 1.6707121462544785e-06, "loss": 2.419, "step": 72630 }, { "epoch": 1.64, "grad_norm": 1.889433248982311, "learning_rate": 1.6686918301730637e-06, "loss": 2.2573, "step": 72640 }, { "epoch": 1.64, "grad_norm": 1.5871126514629177, "learning_rate": 1.6666726251666709e-06, "loss": 2.258, "step": 72650 }, { "epoch": 1.64, "grad_norm": 1.7049504164332996, "learning_rate": 1.6646545315045803e-06, "loss": 2.4124, "step": 72660 }, { "epoch": 1.64, "grad_norm": 1.77723068474692, "learning_rate": 1.6626375494559277e-06, "loss": 2.3898, "step": 72670 }, { "epoch": 1.64, "grad_norm": 1.7231822465488875, "learning_rate": 1.6606216792897e-06, "loss": 2.4191, "step": 72680 }, { "epoch": 1.64, "grad_norm": 1.9340731892180287, "learning_rate": 1.6586069212747414e-06, "loss": 2.3943, "step": 72690 }, { "epoch": 1.64, "grad_norm": 1.6135290987193316, "learning_rate": 1.65659327567974e-06, "loss": 2.35, "step": 72700 }, { "epoch": 1.64, "grad_norm": 1.4672734065709043, "learning_rate": 1.6545807427732386e-06, "loss": 2.3916, "step": 72710 }, { "epoch": 1.64, "grad_norm": 1.732515444787423, "learning_rate": 1.6525693228236307e-06, "loss": 2.4301, "step": 72720 }, { "epoch": 1.64, "grad_norm": 1.9217742347409812, "learning_rate": 1.6505590160991625e-06, "loss": 2.4135, "step": 72730 }, { "epoch": 1.64, "grad_norm": 1.493641235855679, "learning_rate": 1.648549822867932e-06, "loss": 2.332, "step": 72740 }, { "epoch": 1.64, "grad_norm": 1.5198958519047032, "learning_rate": 1.6465417433978848e-06, "loss": 2.2913, "step": 72750 }, { "epoch": 1.64, "grad_norm": 1.8247852879562123, "learning_rate": 1.6445347779568289e-06, "loss": 2.4102, "step": 72760 }, { "epoch": 1.64, "grad_norm": 1.5489686838806336, "learning_rate": 1.6425289268124078e-06, "loss": 2.3753, "step": 72770 }, { "epoch": 1.64, "grad_norm": 1.800542455443983, "learning_rate": 1.640524190232129e-06, "loss": 2.2714, "step": 72780 }, { "epoch": 1.64, "grad_norm": 1.5757716629389031, "learning_rate": 1.6385205684833461e-06, "loss": 2.2728, "step": 72790 }, { "epoch": 1.64, "grad_norm": 1.7670761408195463, "learning_rate": 1.6365180618332643e-06, "loss": 2.3904, "step": 72800 }, { "epoch": 1.64, "grad_norm": 2.063271879416831, "learning_rate": 1.6345166705489412e-06, "loss": 2.3749, "step": 72810 }, { "epoch": 1.64, "grad_norm": 1.868060291751671, "learning_rate": 1.6325163948972821e-06, "loss": 2.3631, "step": 72820 }, { "epoch": 1.64, "grad_norm": 1.7193066992801203, "learning_rate": 1.6305172351450527e-06, "loss": 2.3619, "step": 72830 }, { "epoch": 1.64, "grad_norm": 1.7846075620931419, "learning_rate": 1.6285191915588605e-06, "loss": 2.3675, "step": 72840 }, { "epoch": 1.64, "grad_norm": 1.5112350084768063, "learning_rate": 1.6265222644051681e-06, "loss": 2.2598, "step": 72850 }, { "epoch": 1.64, "grad_norm": 1.8122648761876126, "learning_rate": 1.6245264539502903e-06, "loss": 2.3967, "step": 72860 }, { "epoch": 1.64, "grad_norm": 1.752878083052038, "learning_rate": 1.6225317604603863e-06, "loss": 2.3315, "step": 72870 }, { "epoch": 1.64, "grad_norm": 1.7688876988622546, "learning_rate": 1.620538184201471e-06, "loss": 2.2832, "step": 72880 }, { "epoch": 1.64, "grad_norm": 1.5295595860429123, "learning_rate": 1.6185457254394166e-06, "loss": 2.4468, "step": 72890 }, { "epoch": 1.64, "grad_norm": 1.7294159170404233, "learning_rate": 1.6165543844399367e-06, "loss": 2.3469, "step": 72900 }, { "epoch": 1.64, "grad_norm": 1.742214694063532, "learning_rate": 1.6145641614685993e-06, "loss": 2.3395, "step": 72910 }, { "epoch": 1.64, "grad_norm": 1.6647690618593616, "learning_rate": 1.6125750567908238e-06, "loss": 2.4248, "step": 72920 }, { "epoch": 1.64, "grad_norm": 1.8888779139475018, "learning_rate": 1.6105870706718807e-06, "loss": 2.4298, "step": 72930 }, { "epoch": 1.64, "grad_norm": 1.6114344615239413, "learning_rate": 1.6086002033768887e-06, "loss": 2.4132, "step": 72940 }, { "epoch": 1.65, "grad_norm": 1.6669260472973173, "learning_rate": 1.6066144551708197e-06, "loss": 2.2395, "step": 72950 }, { "epoch": 1.65, "grad_norm": 1.7677329180435408, "learning_rate": 1.6046298263184978e-06, "loss": 2.3324, "step": 72960 }, { "epoch": 1.65, "grad_norm": 1.5741449599646, "learning_rate": 1.6026463170845973e-06, "loss": 2.3948, "step": 72970 }, { "epoch": 1.65, "grad_norm": 1.9532879777489358, "learning_rate": 1.6006639277336367e-06, "loss": 2.3599, "step": 72980 }, { "epoch": 1.65, "grad_norm": 1.5519393179867558, "learning_rate": 1.598682658529992e-06, "loss": 2.5883, "step": 72990 }, { "epoch": 1.65, "grad_norm": 1.8373479520020193, "learning_rate": 1.5967025097378896e-06, "loss": 2.4387, "step": 73000 }, { "epoch": 1.65, "grad_norm": 1.7192949786708738, "learning_rate": 1.594723481621402e-06, "loss": 2.3741, "step": 73010 }, { "epoch": 1.65, "grad_norm": 1.6698479383228284, "learning_rate": 1.5927455744444587e-06, "loss": 2.427, "step": 73020 }, { "epoch": 1.65, "grad_norm": 1.934362280039281, "learning_rate": 1.5907687884708357e-06, "loss": 2.4469, "step": 73030 }, { "epoch": 1.65, "grad_norm": 1.5424865177631106, "learning_rate": 1.5887931239641586e-06, "loss": 2.4241, "step": 73040 }, { "epoch": 1.65, "grad_norm": 1.5674183003492965, "learning_rate": 1.5868185811879055e-06, "loss": 2.3284, "step": 73050 }, { "epoch": 1.65, "grad_norm": 1.7279118241604579, "learning_rate": 1.584845160405404e-06, "loss": 2.1545, "step": 73060 }, { "epoch": 1.65, "grad_norm": 1.648164859088048, "learning_rate": 1.5828728618798316e-06, "loss": 2.48, "step": 73070 }, { "epoch": 1.65, "grad_norm": 1.9955054328688846, "learning_rate": 1.5809016858742177e-06, "loss": 2.4267, "step": 73080 }, { "epoch": 1.65, "grad_norm": 1.7999224987140394, "learning_rate": 1.5789316326514414e-06, "loss": 2.3362, "step": 73090 }, { "epoch": 1.65, "grad_norm": 1.8389322758347746, "learning_rate": 1.5769627024742306e-06, "loss": 2.3516, "step": 73100 }, { "epoch": 1.65, "grad_norm": 1.7317154773017915, "learning_rate": 1.5749948956051652e-06, "loss": 2.2336, "step": 73110 }, { "epoch": 1.65, "grad_norm": 1.5410731541218015, "learning_rate": 1.5730282123066753e-06, "loss": 2.3332, "step": 73120 }, { "epoch": 1.65, "grad_norm": 1.6382053365615938, "learning_rate": 1.57106265284104e-06, "loss": 2.3007, "step": 73130 }, { "epoch": 1.65, "grad_norm": 1.6862773839452563, "learning_rate": 1.5690982174703895e-06, "loss": 2.55, "step": 73140 }, { "epoch": 1.65, "grad_norm": 1.651247847677353, "learning_rate": 1.5671349064567009e-06, "loss": 2.4162, "step": 73150 }, { "epoch": 1.65, "grad_norm": 1.629903377949148, "learning_rate": 1.5651727200618084e-06, "loss": 2.3769, "step": 73160 }, { "epoch": 1.65, "grad_norm": 1.7575079992991152, "learning_rate": 1.56321165854739e-06, "loss": 2.3527, "step": 73170 }, { "epoch": 1.65, "grad_norm": 1.6221366679564726, "learning_rate": 1.5612517221749756e-06, "loss": 2.3623, "step": 73180 }, { "epoch": 1.65, "grad_norm": 1.6138409511743312, "learning_rate": 1.5592929112059485e-06, "loss": 2.4519, "step": 73190 }, { "epoch": 1.65, "grad_norm": 1.6477852480429946, "learning_rate": 1.5573352259015318e-06, "loss": 2.3373, "step": 73200 }, { "epoch": 1.65, "grad_norm": 1.7636682942924926, "learning_rate": 1.5553786665228066e-06, "loss": 2.2703, "step": 73210 }, { "epoch": 1.65, "grad_norm": 1.7078918874912363, "learning_rate": 1.5534232333307064e-06, "loss": 2.321, "step": 73220 }, { "epoch": 1.65, "grad_norm": 1.7012200353125362, "learning_rate": 1.5514689265860083e-06, "loss": 2.3436, "step": 73230 }, { "epoch": 1.65, "grad_norm": 1.7268464953910383, "learning_rate": 1.5495157465493427e-06, "loss": 2.2945, "step": 73240 }, { "epoch": 1.65, "grad_norm": 1.5938138878413093, "learning_rate": 1.5475636934811866e-06, "loss": 2.3057, "step": 73250 }, { "epoch": 1.65, "grad_norm": 1.6816127143891593, "learning_rate": 1.5456127676418698e-06, "loss": 2.3109, "step": 73260 }, { "epoch": 1.65, "grad_norm": 2.1037714936521246, "learning_rate": 1.5436629692915695e-06, "loss": 2.334, "step": 73270 }, { "epoch": 1.65, "grad_norm": 1.5948643049847622, "learning_rate": 1.5417142986903145e-06, "loss": 2.4047, "step": 73280 }, { "epoch": 1.65, "grad_norm": 1.8089458741439397, "learning_rate": 1.539766756097979e-06, "loss": 2.264, "step": 73290 }, { "epoch": 1.65, "grad_norm": 1.6082416911242476, "learning_rate": 1.5378203417742977e-06, "loss": 2.4254, "step": 73300 }, { "epoch": 1.65, "grad_norm": 1.754578276726143, "learning_rate": 1.5358750559788404e-06, "loss": 2.4661, "step": 73310 }, { "epoch": 1.65, "grad_norm": 1.7884586767219843, "learning_rate": 1.5339308989710355e-06, "loss": 2.2752, "step": 73320 }, { "epoch": 1.65, "grad_norm": 1.469070679283725, "learning_rate": 1.5319878710101577e-06, "loss": 2.4379, "step": 73330 }, { "epoch": 1.65, "grad_norm": 1.7327309256171828, "learning_rate": 1.5300459723553319e-06, "loss": 2.2898, "step": 73340 }, { "epoch": 1.65, "grad_norm": 1.6188239220286498, "learning_rate": 1.5281052032655308e-06, "loss": 2.3631, "step": 73350 }, { "epoch": 1.65, "grad_norm": 2.62032343570094, "learning_rate": 1.5261655639995832e-06, "loss": 2.3246, "step": 73360 }, { "epoch": 1.65, "grad_norm": 1.689308930555695, "learning_rate": 1.5242270548161586e-06, "loss": 2.4, "step": 73370 }, { "epoch": 1.65, "grad_norm": 1.7206946807888182, "learning_rate": 1.5222896759737793e-06, "loss": 2.2517, "step": 73380 }, { "epoch": 1.66, "grad_norm": 1.7560125036228773, "learning_rate": 1.5203534277308186e-06, "loss": 2.2835, "step": 73390 }, { "epoch": 1.66, "grad_norm": 1.7197734038154577, "learning_rate": 1.5184183103454987e-06, "loss": 2.4358, "step": 73400 }, { "epoch": 1.66, "grad_norm": 2.013316939252851, "learning_rate": 1.516484324075881e-06, "loss": 2.4412, "step": 73410 }, { "epoch": 1.66, "grad_norm": 1.6610669064099406, "learning_rate": 1.5145514691798934e-06, "loss": 2.3236, "step": 73420 }, { "epoch": 1.66, "grad_norm": 1.7459000914342744, "learning_rate": 1.5126197459153024e-06, "loss": 2.2619, "step": 73430 }, { "epoch": 1.66, "grad_norm": 1.4874643258001556, "learning_rate": 1.510689154539724e-06, "loss": 2.4068, "step": 73440 }, { "epoch": 1.66, "grad_norm": 1.6071263169862342, "learning_rate": 1.5087596953106255e-06, "loss": 2.3741, "step": 73450 }, { "epoch": 1.66, "grad_norm": 1.6658741738048397, "learning_rate": 1.5068313684853209e-06, "loss": 2.2561, "step": 73460 }, { "epoch": 1.66, "grad_norm": 1.5812263661953307, "learning_rate": 1.5049041743209758e-06, "loss": 2.2605, "step": 73470 }, { "epoch": 1.66, "grad_norm": 1.6587080143601203, "learning_rate": 1.5029781130746003e-06, "loss": 2.3671, "step": 73480 }, { "epoch": 1.66, "grad_norm": 1.8211520856660635, "learning_rate": 1.501053185003063e-06, "loss": 2.4437, "step": 73490 }, { "epoch": 1.66, "grad_norm": 1.669154502861498, "learning_rate": 1.499129390363071e-06, "loss": 2.3448, "step": 73500 }, { "epoch": 1.66, "grad_norm": 1.5280637059163096, "learning_rate": 1.4972067294111868e-06, "loss": 2.3835, "step": 73510 }, { "epoch": 1.66, "grad_norm": 1.7110921495737175, "learning_rate": 1.4952852024038155e-06, "loss": 2.3993, "step": 73520 }, { "epoch": 1.66, "grad_norm": 1.7435523220386373, "learning_rate": 1.4933648095972164e-06, "loss": 2.2368, "step": 73530 }, { "epoch": 1.66, "grad_norm": 1.5956650123846772, "learning_rate": 1.4914455512474924e-06, "loss": 2.3167, "step": 73540 }, { "epoch": 1.66, "grad_norm": 1.6847864054934565, "learning_rate": 1.4895274276106053e-06, "loss": 2.3937, "step": 73550 }, { "epoch": 1.66, "grad_norm": 1.6970334028720182, "learning_rate": 1.4876104389423552e-06, "loss": 2.3442, "step": 73560 }, { "epoch": 1.66, "grad_norm": 1.6850566458867196, "learning_rate": 1.485694585498395e-06, "loss": 2.391, "step": 73570 }, { "epoch": 1.66, "grad_norm": 1.9718340761594206, "learning_rate": 1.4837798675342241e-06, "loss": 2.4658, "step": 73580 }, { "epoch": 1.66, "grad_norm": 1.699909340339733, "learning_rate": 1.4818662853051935e-06, "loss": 2.4944, "step": 73590 }, { "epoch": 1.66, "grad_norm": 1.8774975336799375, "learning_rate": 1.4799538390665014e-06, "loss": 2.3168, "step": 73600 }, { "epoch": 1.66, "grad_norm": 1.658848908021453, "learning_rate": 1.4780425290731937e-06, "loss": 2.4336, "step": 73610 }, { "epoch": 1.66, "grad_norm": 1.7025281342885958, "learning_rate": 1.4761323555801654e-06, "loss": 2.3669, "step": 73620 }, { "epoch": 1.66, "grad_norm": 1.5919898125044554, "learning_rate": 1.4742233188421594e-06, "loss": 2.4055, "step": 73630 }, { "epoch": 1.66, "grad_norm": 1.5383420294131598, "learning_rate": 1.472315419113769e-06, "loss": 2.3791, "step": 73640 }, { "epoch": 1.66, "grad_norm": 1.6655594153439293, "learning_rate": 1.4704086566494324e-06, "loss": 2.4671, "step": 73650 }, { "epoch": 1.66, "grad_norm": 1.7686161080745337, "learning_rate": 1.46850303170344e-06, "loss": 2.3921, "step": 73660 }, { "epoch": 1.66, "grad_norm": 1.7085806223219742, "learning_rate": 1.466598544529928e-06, "loss": 2.4524, "step": 73670 }, { "epoch": 1.66, "grad_norm": 1.7696788545264572, "learning_rate": 1.4646951953828781e-06, "loss": 2.4351, "step": 73680 }, { "epoch": 1.66, "grad_norm": 1.6352721247946196, "learning_rate": 1.4627929845161304e-06, "loss": 2.3523, "step": 73690 }, { "epoch": 1.66, "grad_norm": 1.6377928312867027, "learning_rate": 1.460891912183362e-06, "loss": 2.3578, "step": 73700 }, { "epoch": 1.66, "grad_norm": 1.7196861343121101, "learning_rate": 1.4589919786381047e-06, "loss": 2.275, "step": 73710 }, { "epoch": 1.66, "grad_norm": 1.7284237789385346, "learning_rate": 1.4570931841337343e-06, "loss": 2.3115, "step": 73720 }, { "epoch": 1.66, "grad_norm": 1.537198290217604, "learning_rate": 1.4551955289234797e-06, "loss": 2.4885, "step": 73730 }, { "epoch": 1.66, "grad_norm": 1.7091831439780072, "learning_rate": 1.4532990132604085e-06, "loss": 2.3529, "step": 73740 }, { "epoch": 1.66, "grad_norm": 1.7603659426870735, "learning_rate": 1.4514036373974494e-06, "loss": 2.3435, "step": 73750 }, { "epoch": 1.66, "grad_norm": 1.534988712076634, "learning_rate": 1.449509401587369e-06, "loss": 2.4696, "step": 73760 }, { "epoch": 1.66, "grad_norm": 1.681260236564152, "learning_rate": 1.4476163060827875e-06, "loss": 2.3326, "step": 73770 }, { "epoch": 1.66, "grad_norm": 2.066474496950734, "learning_rate": 1.4457243511361685e-06, "loss": 2.3469, "step": 73780 }, { "epoch": 1.66, "grad_norm": 1.5952221689569226, "learning_rate": 1.4438335369998279e-06, "loss": 2.3295, "step": 73790 }, { "epoch": 1.66, "grad_norm": 1.6947028877441652, "learning_rate": 1.4419438639259253e-06, "loss": 2.3781, "step": 73800 }, { "epoch": 1.66, "grad_norm": 1.8720160933574366, "learning_rate": 1.4400553321664711e-06, "loss": 2.3226, "step": 73810 }, { "epoch": 1.66, "grad_norm": 1.6131885210586152, "learning_rate": 1.4381679419733208e-06, "loss": 2.363, "step": 73820 }, { "epoch": 1.66, "grad_norm": 1.650376250728116, "learning_rate": 1.4362816935981839e-06, "loss": 2.4205, "step": 73830 }, { "epoch": 1.67, "grad_norm": 1.6591280084786815, "learning_rate": 1.4343965872926113e-06, "loss": 2.451, "step": 73840 }, { "epoch": 1.67, "grad_norm": 1.7057920709136443, "learning_rate": 1.432512623308001e-06, "loss": 2.3636, "step": 73850 }, { "epoch": 1.67, "grad_norm": 2.1218838004072027, "learning_rate": 1.4306298018956032e-06, "loss": 2.3339, "step": 73860 }, { "epoch": 1.67, "grad_norm": 1.5773565268921825, "learning_rate": 1.4287481233065126e-06, "loss": 2.4607, "step": 73870 }, { "epoch": 1.67, "grad_norm": 1.8528539291663992, "learning_rate": 1.4268675877916703e-06, "loss": 2.3732, "step": 73880 }, { "epoch": 1.67, "grad_norm": 1.6090928611291893, "learning_rate": 1.424988195601873e-06, "loss": 2.3318, "step": 73890 }, { "epoch": 1.67, "grad_norm": 1.7285881177453768, "learning_rate": 1.4231099469877552e-06, "loss": 2.5132, "step": 73900 }, { "epoch": 1.67, "grad_norm": 1.564045692623387, "learning_rate": 1.4212328421998034e-06, "loss": 2.4544, "step": 73910 }, { "epoch": 1.67, "grad_norm": 1.4817319912311149, "learning_rate": 1.4193568814883519e-06, "loss": 2.3905, "step": 73920 }, { "epoch": 1.67, "grad_norm": 1.6796788709014356, "learning_rate": 1.4174820651035803e-06, "loss": 2.2976, "step": 73930 }, { "epoch": 1.67, "grad_norm": 1.9846951603164829, "learning_rate": 1.4156083932955168e-06, "loss": 2.3717, "step": 73940 }, { "epoch": 1.67, "grad_norm": 2.0084050424955637, "learning_rate": 1.4137358663140377e-06, "loss": 2.3436, "step": 73950 }, { "epoch": 1.67, "grad_norm": 1.9953881428986677, "learning_rate": 1.4118644844088659e-06, "loss": 2.4681, "step": 73960 }, { "epoch": 1.67, "grad_norm": 1.5631027522882623, "learning_rate": 1.4099942478295703e-06, "loss": 2.4759, "step": 73970 }, { "epoch": 1.67, "grad_norm": 1.6710858871244074, "learning_rate": 1.4081251568255682e-06, "loss": 2.3453, "step": 73980 }, { "epoch": 1.67, "grad_norm": 1.6273040111075467, "learning_rate": 1.406257211646126e-06, "loss": 2.3961, "step": 73990 }, { "epoch": 1.67, "grad_norm": 1.8493174729339619, "learning_rate": 1.4043904125403551e-06, "loss": 2.359, "step": 74000 }, { "epoch": 1.67, "grad_norm": 1.8044418660000785, "learning_rate": 1.40252475975721e-06, "loss": 2.3727, "step": 74010 }, { "epoch": 1.67, "grad_norm": 1.673576488964496, "learning_rate": 1.4006602535455037e-06, "loss": 2.2935, "step": 74020 }, { "epoch": 1.67, "grad_norm": 1.6555350544173186, "learning_rate": 1.3987968941538877e-06, "loss": 2.4563, "step": 74030 }, { "epoch": 1.67, "grad_norm": 1.6366207052256838, "learning_rate": 1.3969346818308593e-06, "loss": 2.4326, "step": 74040 }, { "epoch": 1.67, "grad_norm": 1.6061750715862884, "learning_rate": 1.3950736168247702e-06, "loss": 2.4011, "step": 74050 }, { "epoch": 1.67, "grad_norm": 1.821995281569004, "learning_rate": 1.3932136993838098e-06, "loss": 2.4167, "step": 74060 }, { "epoch": 1.67, "grad_norm": 1.5432503143306464, "learning_rate": 1.391354929756018e-06, "loss": 2.336, "step": 74070 }, { "epoch": 1.67, "grad_norm": 1.758610598977753, "learning_rate": 1.3894973081892894e-06, "loss": 2.3398, "step": 74080 }, { "epoch": 1.67, "grad_norm": 1.7947966541586842, "learning_rate": 1.3876408349313552e-06, "loss": 2.3322, "step": 74090 }, { "epoch": 1.67, "grad_norm": 1.5292026496853746, "learning_rate": 1.385785510229799e-06, "loss": 2.5472, "step": 74100 }, { "epoch": 1.67, "grad_norm": 1.8670519239703254, "learning_rate": 1.383931334332048e-06, "loss": 2.421, "step": 74110 }, { "epoch": 1.67, "grad_norm": 1.6031105146868248, "learning_rate": 1.3820783074853783e-06, "loss": 2.4009, "step": 74120 }, { "epoch": 1.67, "grad_norm": 1.6888986884693362, "learning_rate": 1.3802264299369129e-06, "loss": 2.3906, "step": 74130 }, { "epoch": 1.67, "grad_norm": 1.8699811639885204, "learning_rate": 1.3783757019336186e-06, "loss": 2.3286, "step": 74140 }, { "epoch": 1.67, "grad_norm": 1.7317196483666208, "learning_rate": 1.3765261237223114e-06, "loss": 2.3908, "step": 74150 }, { "epoch": 1.67, "grad_norm": 1.5549976372982235, "learning_rate": 1.3746776955496588e-06, "loss": 2.4136, "step": 74160 }, { "epoch": 1.67, "grad_norm": 1.8170810727034723, "learning_rate": 1.3728304176621633e-06, "loss": 2.2851, "step": 74170 }, { "epoch": 1.67, "grad_norm": 1.575370776259165, "learning_rate": 1.3709842903061832e-06, "loss": 2.3297, "step": 74180 }, { "epoch": 1.67, "grad_norm": 1.6227803255824294, "learning_rate": 1.3691393137279208e-06, "loss": 2.3728, "step": 74190 }, { "epoch": 1.67, "grad_norm": 1.5022277569659193, "learning_rate": 1.3672954881734236e-06, "loss": 2.2245, "step": 74200 }, { "epoch": 1.67, "grad_norm": 1.665795091160585, "learning_rate": 1.3654528138885858e-06, "loss": 2.4793, "step": 74210 }, { "epoch": 1.67, "grad_norm": 1.8001647434238064, "learning_rate": 1.3636112911191535e-06, "loss": 2.4844, "step": 74220 }, { "epoch": 1.67, "grad_norm": 1.6989580187973832, "learning_rate": 1.361770920110712e-06, "loss": 2.2952, "step": 74230 }, { "epoch": 1.67, "grad_norm": 1.6368794630422006, "learning_rate": 1.3599317011086955e-06, "loss": 2.39, "step": 74240 }, { "epoch": 1.67, "grad_norm": 1.7185117200697393, "learning_rate": 1.358093634358385e-06, "loss": 2.4655, "step": 74250 }, { "epoch": 1.67, "grad_norm": 1.5414929992966717, "learning_rate": 1.3562567201049082e-06, "loss": 2.4579, "step": 74260 }, { "epoch": 1.67, "grad_norm": 1.7227982191377968, "learning_rate": 1.3544209585932379e-06, "loss": 2.4063, "step": 74270 }, { "epoch": 1.68, "grad_norm": 1.7021810655046938, "learning_rate": 1.3525863500681946e-06, "loss": 2.3288, "step": 74280 }, { "epoch": 1.68, "grad_norm": 1.8402607224723968, "learning_rate": 1.3507528947744442e-06, "loss": 2.3701, "step": 74290 }, { "epoch": 1.68, "grad_norm": 1.7079098862177988, "learning_rate": 1.3489205929564974e-06, "loss": 2.3707, "step": 74300 }, { "epoch": 1.68, "grad_norm": 1.7069153628504747, "learning_rate": 1.3470894448587146e-06, "loss": 2.4081, "step": 74310 }, { "epoch": 1.68, "grad_norm": 1.5896305660628258, "learning_rate": 1.3452594507253003e-06, "loss": 2.3617, "step": 74320 }, { "epoch": 1.68, "grad_norm": 1.732767722697919, "learning_rate": 1.3434306108003026e-06, "loss": 2.4868, "step": 74330 }, { "epoch": 1.68, "grad_norm": 1.5590774708049677, "learning_rate": 1.3416029253276208e-06, "loss": 2.2217, "step": 74340 }, { "epoch": 1.68, "grad_norm": 1.8861394143169439, "learning_rate": 1.3397763945509945e-06, "loss": 2.384, "step": 74350 }, { "epoch": 1.68, "grad_norm": 1.8008110539623878, "learning_rate": 1.3379510187140154e-06, "loss": 2.3591, "step": 74360 }, { "epoch": 1.68, "grad_norm": 1.5696221542542652, "learning_rate": 1.3361267980601211e-06, "loss": 2.3393, "step": 74370 }, { "epoch": 1.68, "grad_norm": 1.8052749614055115, "learning_rate": 1.3343037328325858e-06, "loss": 2.34, "step": 74380 }, { "epoch": 1.68, "grad_norm": 1.5809647708705505, "learning_rate": 1.3324818232745384e-06, "loss": 2.4686, "step": 74390 }, { "epoch": 1.68, "grad_norm": 1.8764719610423397, "learning_rate": 1.3306610696289524e-06, "loss": 2.4698, "step": 74400 }, { "epoch": 1.68, "grad_norm": 1.6132924852895818, "learning_rate": 1.3288414721386423e-06, "loss": 2.3511, "step": 74410 }, { "epoch": 1.68, "grad_norm": 1.5540352383407054, "learning_rate": 1.3270230310462785e-06, "loss": 2.3888, "step": 74420 }, { "epoch": 1.68, "grad_norm": 1.6921156524456482, "learning_rate": 1.3252057465943667e-06, "loss": 2.4651, "step": 74430 }, { "epoch": 1.68, "grad_norm": 1.6761592265529186, "learning_rate": 1.3233896190252636e-06, "loss": 2.4041, "step": 74440 }, { "epoch": 1.68, "grad_norm": 1.7374186948973063, "learning_rate": 1.3215746485811708e-06, "loss": 2.3951, "step": 74450 }, { "epoch": 1.68, "grad_norm": 1.7460937297943395, "learning_rate": 1.3197608355041347e-06, "loss": 2.3136, "step": 74460 }, { "epoch": 1.68, "grad_norm": 1.8142531975103617, "learning_rate": 1.3179481800360472e-06, "loss": 2.2694, "step": 74470 }, { "epoch": 1.68, "grad_norm": 2.0477127673208595, "learning_rate": 1.3161366824186495e-06, "loss": 2.4344, "step": 74480 }, { "epoch": 1.68, "grad_norm": 1.7880364167771494, "learning_rate": 1.3143263428935216e-06, "loss": 2.432, "step": 74490 }, { "epoch": 1.68, "grad_norm": 1.5705936277446388, "learning_rate": 1.312517161702096e-06, "loss": 2.3221, "step": 74500 }, { "epoch": 1.68, "grad_norm": 1.7560549208207397, "learning_rate": 1.3107091390856463e-06, "loss": 2.4939, "step": 74510 }, { "epoch": 1.68, "grad_norm": 1.6469857831674777, "learning_rate": 1.3089022752852932e-06, "loss": 2.3577, "step": 74520 }, { "epoch": 1.68, "grad_norm": 1.668579925538671, "learning_rate": 1.3070965705420024e-06, "loss": 2.4463, "step": 74530 }, { "epoch": 1.68, "grad_norm": 1.7612495368565082, "learning_rate": 1.3052920250965827e-06, "loss": 2.37, "step": 74540 }, { "epoch": 1.68, "grad_norm": 1.9127946976446213, "learning_rate": 1.3034886391896973e-06, "loss": 2.1861, "step": 74550 }, { "epoch": 1.68, "grad_norm": 1.7485911657364737, "learning_rate": 1.3016864130618435e-06, "loss": 2.3105, "step": 74560 }, { "epoch": 1.68, "grad_norm": 1.5574544234946448, "learning_rate": 1.299885346953369e-06, "loss": 2.3526, "step": 74570 }, { "epoch": 1.68, "grad_norm": 1.8403699308723112, "learning_rate": 1.2980854411044685e-06, "loss": 2.4699, "step": 74580 }, { "epoch": 1.68, "grad_norm": 1.7707150479324183, "learning_rate": 1.2962866957551812e-06, "loss": 2.415, "step": 74590 }, { "epoch": 1.68, "grad_norm": 1.7675284723817783, "learning_rate": 1.2944891111453829e-06, "loss": 2.4027, "step": 74600 }, { "epoch": 1.68, "grad_norm": 1.6151722820265693, "learning_rate": 1.2926926875148082e-06, "loss": 2.3949, "step": 74610 }, { "epoch": 1.68, "grad_norm": 1.9383287636182074, "learning_rate": 1.2908974251030315e-06, "loss": 2.4686, "step": 74620 }, { "epoch": 1.68, "grad_norm": 1.9930181831190847, "learning_rate": 1.2891033241494677e-06, "loss": 2.3716, "step": 74630 }, { "epoch": 1.68, "grad_norm": 1.7125459997948156, "learning_rate": 1.2873103848933831e-06, "loss": 2.3339, "step": 74640 }, { "epoch": 1.68, "grad_norm": 1.6817510743726807, "learning_rate": 1.2855186075738867e-06, "loss": 2.3891, "step": 74650 }, { "epoch": 1.68, "grad_norm": 1.8617040991906877, "learning_rate": 1.2837279924299306e-06, "loss": 2.3211, "step": 74660 }, { "epoch": 1.68, "grad_norm": 1.574529049405798, "learning_rate": 1.281938539700316e-06, "loss": 2.3173, "step": 74670 }, { "epoch": 1.68, "grad_norm": 1.5577806362417514, "learning_rate": 1.2801502496236827e-06, "loss": 2.3521, "step": 74680 }, { "epoch": 1.68, "grad_norm": 1.6452431966940204, "learning_rate": 1.278363122438525e-06, "loss": 2.2877, "step": 74690 }, { "epoch": 1.68, "grad_norm": 1.7514865259203523, "learning_rate": 1.2765771583831765e-06, "loss": 2.3622, "step": 74700 }, { "epoch": 1.68, "grad_norm": 1.7949836838701407, "learning_rate": 1.2747923576958121e-06, "loss": 2.4481, "step": 74710 }, { "epoch": 1.69, "grad_norm": 1.6851330220127003, "learning_rate": 1.273008720614456e-06, "loss": 2.3665, "step": 74720 }, { "epoch": 1.69, "grad_norm": 1.6390175936557705, "learning_rate": 1.2712262473769788e-06, "loss": 2.3831, "step": 74730 }, { "epoch": 1.69, "grad_norm": 1.6457676072368184, "learning_rate": 1.2694449382210883e-06, "loss": 2.4154, "step": 74740 }, { "epoch": 1.69, "grad_norm": 1.6092143497673808, "learning_rate": 1.26766479338435e-06, "loss": 2.4068, "step": 74750 }, { "epoch": 1.69, "grad_norm": 1.7773522267184443, "learning_rate": 1.265885813104163e-06, "loss": 2.3744, "step": 74760 }, { "epoch": 1.69, "grad_norm": 1.8001886323944005, "learning_rate": 1.264107997617774e-06, "loss": 2.3433, "step": 74770 }, { "epoch": 1.69, "grad_norm": 1.74961482827626, "learning_rate": 1.2623313471622755e-06, "loss": 2.3572, "step": 74780 }, { "epoch": 1.69, "grad_norm": 1.4700787814671814, "learning_rate": 1.260555861974605e-06, "loss": 2.4807, "step": 74790 }, { "epoch": 1.69, "grad_norm": 1.6304429590267995, "learning_rate": 1.258781542291543e-06, "loss": 2.4879, "step": 74800 }, { "epoch": 1.69, "grad_norm": 1.66125559462391, "learning_rate": 1.2570083883497142e-06, "loss": 2.3038, "step": 74810 }, { "epoch": 1.69, "grad_norm": 1.6397506446639505, "learning_rate": 1.255236400385591e-06, "loss": 2.4542, "step": 74820 }, { "epoch": 1.69, "grad_norm": 1.8140460185860454, "learning_rate": 1.2534655786354865e-06, "loss": 2.3716, "step": 74830 }, { "epoch": 1.69, "grad_norm": 1.53656002512584, "learning_rate": 1.2516959233355597e-06, "loss": 2.3234, "step": 74840 }, { "epoch": 1.69, "grad_norm": 1.6651842910585526, "learning_rate": 1.249927434721817e-06, "loss": 2.3624, "step": 74850 }, { "epoch": 1.69, "grad_norm": 1.76345017706077, "learning_rate": 1.2481601130301036e-06, "loss": 2.3421, "step": 74860 }, { "epoch": 1.69, "grad_norm": 1.726908465090003, "learning_rate": 1.2463939584961137e-06, "loss": 2.4238, "step": 74870 }, { "epoch": 1.69, "grad_norm": 1.7858093467980718, "learning_rate": 1.2446289713553816e-06, "loss": 2.3517, "step": 74880 }, { "epoch": 1.69, "grad_norm": 2.0821818509605605, "learning_rate": 1.2428651518432921e-06, "loss": 2.3109, "step": 74890 }, { "epoch": 1.69, "grad_norm": 1.6537966304533156, "learning_rate": 1.2411025001950694e-06, "loss": 2.4639, "step": 74900 }, { "epoch": 1.69, "grad_norm": 1.5091668643336917, "learning_rate": 1.2393410166457854e-06, "loss": 2.3473, "step": 74910 }, { "epoch": 1.69, "grad_norm": 1.5567313502422402, "learning_rate": 1.2375807014303486e-06, "loss": 2.4103, "step": 74920 }, { "epoch": 1.69, "grad_norm": 1.8341350662389546, "learning_rate": 1.2358215547835207e-06, "loss": 2.4421, "step": 74930 }, { "epoch": 1.69, "grad_norm": 1.8392374524400124, "learning_rate": 1.2340635769399012e-06, "loss": 2.4097, "step": 74940 }, { "epoch": 1.69, "grad_norm": 1.7910183493636556, "learning_rate": 1.2323067681339407e-06, "loss": 2.4694, "step": 74950 }, { "epoch": 1.69, "grad_norm": 1.722228916865833, "learning_rate": 1.2305511285999273e-06, "loss": 2.4441, "step": 74960 }, { "epoch": 1.69, "grad_norm": 1.604200595450097, "learning_rate": 1.2287966585719968e-06, "loss": 2.4854, "step": 74970 }, { "epoch": 1.69, "grad_norm": 1.989900879786658, "learning_rate": 1.2270433582841268e-06, "loss": 2.345, "step": 74980 }, { "epoch": 1.69, "grad_norm": 1.7062769530400375, "learning_rate": 1.225291227970139e-06, "loss": 2.1592, "step": 74990 }, { "epoch": 1.69, "grad_norm": 1.7874842694325868, "learning_rate": 1.223540267863702e-06, "loss": 2.4316, "step": 75000 }, { "epoch": 1.69, "grad_norm": 1.6104054345429344, "learning_rate": 1.2217904781983225e-06, "loss": 2.3916, "step": 75010 }, { "epoch": 1.69, "grad_norm": 1.6828200620774856, "learning_rate": 1.2200418592073626e-06, "loss": 2.3983, "step": 75020 }, { "epoch": 1.69, "grad_norm": 1.7303615551562628, "learning_rate": 1.2182944111240124e-06, "loss": 2.464, "step": 75030 }, { "epoch": 1.69, "grad_norm": 1.6441253615478602, "learning_rate": 1.216548134181319e-06, "loss": 2.2506, "step": 75040 }, { "epoch": 1.69, "grad_norm": 1.7090038418846236, "learning_rate": 1.214803028612166e-06, "loss": 2.4242, "step": 75050 }, { "epoch": 1.69, "grad_norm": 1.637994448804323, "learning_rate": 1.2130590946492837e-06, "loss": 2.4169, "step": 75060 }, { "epoch": 1.69, "grad_norm": 1.5506022975078402, "learning_rate": 1.2113163325252431e-06, "loss": 2.4333, "step": 75070 }, { "epoch": 1.69, "grad_norm": 1.7294314526606267, "learning_rate": 1.2095747424724668e-06, "loss": 2.3777, "step": 75080 }, { "epoch": 1.69, "grad_norm": 1.9033227590177142, "learning_rate": 1.2078343247232138e-06, "loss": 2.4596, "step": 75090 }, { "epoch": 1.69, "grad_norm": 1.4998808343629237, "learning_rate": 1.2060950795095871e-06, "loss": 2.3914, "step": 75100 }, { "epoch": 1.69, "grad_norm": 1.7804229634630808, "learning_rate": 1.2043570070635347e-06, "loss": 2.3073, "step": 75110 }, { "epoch": 1.69, "grad_norm": 1.758615729480028, "learning_rate": 1.2026201076168508e-06, "loss": 2.4796, "step": 75120 }, { "epoch": 1.69, "grad_norm": 1.517069942141568, "learning_rate": 1.2008843814011684e-06, "loss": 2.4551, "step": 75130 }, { "epoch": 1.69, "grad_norm": 1.6589082930998558, "learning_rate": 1.1991498286479686e-06, "loss": 2.3344, "step": 75140 }, { "epoch": 1.69, "grad_norm": 1.808813749092065, "learning_rate": 1.197416449588571e-06, "loss": 2.3952, "step": 75150 }, { "epoch": 1.69, "grad_norm": 1.6252860336202182, "learning_rate": 1.195684244454144e-06, "loss": 2.3728, "step": 75160 }, { "epoch": 1.7, "grad_norm": 1.739420834015086, "learning_rate": 1.193953213475695e-06, "loss": 2.331, "step": 75170 }, { "epoch": 1.7, "grad_norm": 1.775941238742932, "learning_rate": 1.192223356884078e-06, "loss": 2.4301, "step": 75180 }, { "epoch": 1.7, "grad_norm": 1.6387948793369755, "learning_rate": 1.19049467490999e-06, "loss": 2.3853, "step": 75190 }, { "epoch": 1.7, "grad_norm": 1.6012105600406825, "learning_rate": 1.1887671677839675e-06, "loss": 2.4017, "step": 75200 }, { "epoch": 1.7, "grad_norm": 1.7275476055162091, "learning_rate": 1.1870408357363937e-06, "loss": 2.3447, "step": 75210 }, { "epoch": 1.7, "grad_norm": 1.7417957006033482, "learning_rate": 1.1853156789974984e-06, "loss": 2.3257, "step": 75220 }, { "epoch": 1.7, "grad_norm": 1.5759011855436182, "learning_rate": 1.1835916977973484e-06, "loss": 2.3467, "step": 75230 }, { "epoch": 1.7, "grad_norm": 1.5257716081800174, "learning_rate": 1.1818688923658571e-06, "loss": 2.4925, "step": 75240 }, { "epoch": 1.7, "grad_norm": 1.6365910152384902, "learning_rate": 1.1801472629327792e-06, "loss": 2.4425, "step": 75250 }, { "epoch": 1.7, "grad_norm": 1.5932436016223435, "learning_rate": 1.1784268097277141e-06, "loss": 2.3687, "step": 75260 }, { "epoch": 1.7, "grad_norm": 1.6854211695948544, "learning_rate": 1.1767075329801002e-06, "loss": 2.4032, "step": 75270 }, { "epoch": 1.7, "grad_norm": 1.690366878751206, "learning_rate": 1.1749894329192291e-06, "loss": 2.4412, "step": 75280 }, { "epoch": 1.7, "grad_norm": 1.6653637664347591, "learning_rate": 1.173272509774227e-06, "loss": 2.4002, "step": 75290 }, { "epoch": 1.7, "grad_norm": 1.7560672141890084, "learning_rate": 1.1715567637740633e-06, "loss": 2.3824, "step": 75300 }, { "epoch": 1.7, "grad_norm": 1.7083240839102856, "learning_rate": 1.1698421951475536e-06, "loss": 2.3584, "step": 75310 }, { "epoch": 1.7, "grad_norm": 1.5602002445467673, "learning_rate": 1.168128804123355e-06, "loss": 2.365, "step": 75320 }, { "epoch": 1.7, "grad_norm": 1.6041088533761183, "learning_rate": 1.1664165909299685e-06, "loss": 2.3354, "step": 75330 }, { "epoch": 1.7, "grad_norm": 1.699522133081465, "learning_rate": 1.1647055557957353e-06, "loss": 2.2752, "step": 75340 }, { "epoch": 1.7, "grad_norm": 1.7153527303522933, "learning_rate": 1.1629956989488433e-06, "loss": 2.474, "step": 75350 }, { "epoch": 1.7, "grad_norm": 1.5974470540501926, "learning_rate": 1.161287020617321e-06, "loss": 2.3796, "step": 75360 }, { "epoch": 1.7, "grad_norm": 1.590928078955244, "learning_rate": 1.1595795210290395e-06, "loss": 2.3355, "step": 75370 }, { "epoch": 1.7, "grad_norm": 1.527611496673102, "learning_rate": 1.1578732004117143e-06, "loss": 2.3398, "step": 75380 }, { "epoch": 1.7, "grad_norm": 1.9014870821035044, "learning_rate": 1.1561680589929026e-06, "loss": 2.5345, "step": 75390 }, { "epoch": 1.7, "grad_norm": 1.5086167797684684, "learning_rate": 1.1544640970000032e-06, "loss": 2.3862, "step": 75400 }, { "epoch": 1.7, "grad_norm": 1.6839380781591287, "learning_rate": 1.1527613146602578e-06, "loss": 2.3728, "step": 75410 }, { "epoch": 1.7, "grad_norm": 1.5297349487457133, "learning_rate": 1.1510597122007562e-06, "loss": 2.4231, "step": 75420 }, { "epoch": 1.7, "grad_norm": 1.6707883923234563, "learning_rate": 1.149359289848424e-06, "loss": 2.3988, "step": 75430 }, { "epoch": 1.7, "grad_norm": 1.7854123708550906, "learning_rate": 1.1476600478300316e-06, "loss": 2.3326, "step": 75440 }, { "epoch": 1.7, "grad_norm": 1.807860726871894, "learning_rate": 1.1459619863721948e-06, "loss": 2.3503, "step": 75450 }, { "epoch": 1.7, "grad_norm": 1.9052307317677548, "learning_rate": 1.1442651057013665e-06, "loss": 2.2812, "step": 75460 }, { "epoch": 1.7, "grad_norm": 1.6707954984832118, "learning_rate": 1.1425694060438431e-06, "loss": 2.4099, "step": 75470 }, { "epoch": 1.7, "grad_norm": 1.6347394678869187, "learning_rate": 1.1408748876257714e-06, "loss": 2.3653, "step": 75480 }, { "epoch": 1.7, "grad_norm": 1.7801039589408147, "learning_rate": 1.1391815506731308e-06, "loss": 2.3951, "step": 75490 }, { "epoch": 1.7, "grad_norm": 1.4359976714625837, "learning_rate": 1.1374893954117483e-06, "loss": 2.3432, "step": 75500 }, { "epoch": 1.7, "grad_norm": 1.7062820504913332, "learning_rate": 1.1357984220672923e-06, "loss": 2.3614, "step": 75510 }, { "epoch": 1.7, "grad_norm": 1.7332156294342966, "learning_rate": 1.1341086308652726e-06, "loss": 2.4628, "step": 75520 }, { "epoch": 1.7, "grad_norm": 1.5570848733572364, "learning_rate": 1.1324200220310433e-06, "loss": 2.2621, "step": 75530 }, { "epoch": 1.7, "grad_norm": 1.8261524822687163, "learning_rate": 1.1307325957897964e-06, "loss": 2.4603, "step": 75540 }, { "epoch": 1.7, "grad_norm": 1.556941738762828, "learning_rate": 1.129046352366574e-06, "loss": 2.2893, "step": 75550 }, { "epoch": 1.7, "grad_norm": 1.6360441015472749, "learning_rate": 1.1273612919862564e-06, "loss": 2.4168, "step": 75560 }, { "epoch": 1.7, "grad_norm": 1.4545162299581937, "learning_rate": 1.1256774148735604e-06, "loss": 2.3963, "step": 75570 }, { "epoch": 1.7, "grad_norm": 1.643240873146936, "learning_rate": 1.1239947212530534e-06, "loss": 2.3859, "step": 75580 }, { "epoch": 1.7, "grad_norm": 1.6477242389853712, "learning_rate": 1.1223132113491408e-06, "loss": 2.4079, "step": 75590 }, { "epoch": 1.7, "grad_norm": 1.7768390069301372, "learning_rate": 1.1206328853860705e-06, "loss": 2.4703, "step": 75600 }, { "epoch": 1.71, "grad_norm": 1.5896280528019373, "learning_rate": 1.118953743587935e-06, "loss": 2.3781, "step": 75610 }, { "epoch": 1.71, "grad_norm": 1.5525804852491605, "learning_rate": 1.1172757861786675e-06, "loss": 2.3808, "step": 75620 }, { "epoch": 1.71, "grad_norm": 1.680231253758787, "learning_rate": 1.1155990133820416e-06, "loss": 2.3137, "step": 75630 }, { "epoch": 1.71, "grad_norm": 1.5742339729693835, "learning_rate": 1.1139234254216735e-06, "loss": 2.3944, "step": 75640 }, { "epoch": 1.71, "grad_norm": 1.746618740253009, "learning_rate": 1.1122490225210225e-06, "loss": 2.4569, "step": 75650 }, { "epoch": 1.71, "grad_norm": 1.8905081027200006, "learning_rate": 1.11057580490339e-06, "loss": 2.3771, "step": 75660 }, { "epoch": 1.71, "grad_norm": 1.5556840358706616, "learning_rate": 1.1089037727919182e-06, "loss": 2.4073, "step": 75670 }, { "epoch": 1.71, "grad_norm": 1.5662849448885758, "learning_rate": 1.1072329264095904e-06, "loss": 2.4849, "step": 75680 }, { "epoch": 1.71, "grad_norm": 1.8856076027172552, "learning_rate": 1.1055632659792358e-06, "loss": 2.4052, "step": 75690 }, { "epoch": 1.71, "grad_norm": 1.76238837152948, "learning_rate": 1.1038947917235199e-06, "loss": 2.3941, "step": 75700 }, { "epoch": 1.71, "grad_norm": 1.918529291301139, "learning_rate": 1.1022275038649544e-06, "loss": 2.3731, "step": 75710 }, { "epoch": 1.71, "grad_norm": 1.6091222096072642, "learning_rate": 1.1005614026258905e-06, "loss": 2.4016, "step": 75720 }, { "epoch": 1.71, "grad_norm": 1.6225740312098, "learning_rate": 1.0988964882285214e-06, "loss": 2.2678, "step": 75730 }, { "epoch": 1.71, "grad_norm": 1.6618063096092932, "learning_rate": 1.097232760894882e-06, "loss": 2.2843, "step": 75740 }, { "epoch": 1.71, "grad_norm": 1.7735701479100965, "learning_rate": 1.0955702208468522e-06, "loss": 2.4043, "step": 75750 }, { "epoch": 1.71, "grad_norm": 1.6847403725837167, "learning_rate": 1.0939088683061482e-06, "loss": 2.2397, "step": 75760 }, { "epoch": 1.71, "grad_norm": 1.7773783211404734, "learning_rate": 1.0922487034943308e-06, "loss": 2.462, "step": 75770 }, { "epoch": 1.71, "grad_norm": 1.6309857157666052, "learning_rate": 1.0905897266328047e-06, "loss": 2.3742, "step": 75780 }, { "epoch": 1.71, "grad_norm": 1.549307364670353, "learning_rate": 1.0889319379428086e-06, "loss": 2.3235, "step": 75790 }, { "epoch": 1.71, "grad_norm": 1.7728035789022356, "learning_rate": 1.087275337645428e-06, "loss": 2.4446, "step": 75800 }, { "epoch": 1.71, "grad_norm": 1.531599644490502, "learning_rate": 1.0856199259615929e-06, "loss": 2.4348, "step": 75810 }, { "epoch": 1.71, "grad_norm": 1.6557603916622843, "learning_rate": 1.0839657031120698e-06, "loss": 2.3825, "step": 75820 }, { "epoch": 1.71, "grad_norm": 1.6054372717683572, "learning_rate": 1.0823126693174678e-06, "loss": 2.3431, "step": 75830 }, { "epoch": 1.71, "grad_norm": 1.690252291236123, "learning_rate": 1.0806608247982386e-06, "loss": 2.4388, "step": 75840 }, { "epoch": 1.71, "grad_norm": 1.451660844790819, "learning_rate": 1.0790101697746736e-06, "loss": 2.4358, "step": 75850 }, { "epoch": 1.71, "grad_norm": 1.6549608631195025, "learning_rate": 1.0773607044669066e-06, "loss": 2.3927, "step": 75860 }, { "epoch": 1.71, "grad_norm": 1.704506409321036, "learning_rate": 1.0757124290949128e-06, "loss": 2.3822, "step": 75870 }, { "epoch": 1.71, "grad_norm": 1.6820642620154946, "learning_rate": 1.0740653438785066e-06, "loss": 2.3409, "step": 75880 }, { "epoch": 1.71, "grad_norm": 1.7964819950542896, "learning_rate": 1.0724194490373519e-06, "loss": 2.2726, "step": 75890 }, { "epoch": 1.71, "grad_norm": 2.762583460455102, "learning_rate": 1.0707747447909412e-06, "loss": 2.4255, "step": 75900 }, { "epoch": 1.71, "grad_norm": 1.480241351543976, "learning_rate": 1.0691312313586166e-06, "loss": 2.5117, "step": 75910 }, { "epoch": 1.71, "grad_norm": 1.7523998312689162, "learning_rate": 1.0674889089595597e-06, "loss": 2.3518, "step": 75920 }, { "epoch": 1.71, "grad_norm": 1.6041243764278221, "learning_rate": 1.0658477778127919e-06, "loss": 2.3681, "step": 75930 }, { "epoch": 1.71, "grad_norm": 1.519758197403647, "learning_rate": 1.0642078381371757e-06, "loss": 2.4036, "step": 75940 }, { "epoch": 1.71, "grad_norm": 1.6419847579858915, "learning_rate": 1.0625690901514208e-06, "loss": 2.3697, "step": 75950 }, { "epoch": 1.71, "grad_norm": 1.7288153835362112, "learning_rate": 1.0609315340740679e-06, "loss": 2.4285, "step": 75960 }, { "epoch": 1.71, "grad_norm": 1.687754361960726, "learning_rate": 1.0592951701235065e-06, "loss": 2.4271, "step": 75970 }, { "epoch": 1.71, "grad_norm": 1.7809088143744076, "learning_rate": 1.0576599985179636e-06, "loss": 2.3206, "step": 75980 }, { "epoch": 1.71, "grad_norm": 1.683566604121052, "learning_rate": 1.05602601947551e-06, "loss": 2.3804, "step": 75990 }, { "epoch": 1.71, "grad_norm": 1.8070149821928623, "learning_rate": 1.0543932332140494e-06, "loss": 2.3714, "step": 76000 }, { "epoch": 1.71, "grad_norm": 1.678602076359189, "learning_rate": 1.0527616399513374e-06, "loss": 2.4084, "step": 76010 }, { "epoch": 1.71, "grad_norm": 1.7685051586681448, "learning_rate": 1.0511312399049657e-06, "loss": 2.5226, "step": 76020 }, { "epoch": 1.71, "grad_norm": 1.519712738463893, "learning_rate": 1.0495020332923656e-06, "loss": 2.4349, "step": 76030 }, { "epoch": 1.71, "grad_norm": 1.928531550223114, "learning_rate": 1.0478740203308101e-06, "loss": 2.4652, "step": 76040 }, { "epoch": 1.72, "grad_norm": 1.8878855052790182, "learning_rate": 1.0462472012374136e-06, "loss": 2.3619, "step": 76050 }, { "epoch": 1.72, "grad_norm": 1.7362905357152256, "learning_rate": 1.0446215762291311e-06, "loss": 2.3546, "step": 76060 }, { "epoch": 1.72, "grad_norm": 1.671965653771648, "learning_rate": 1.0429971455227573e-06, "loss": 2.3342, "step": 76070 }, { "epoch": 1.72, "grad_norm": 1.5946971018088631, "learning_rate": 1.0413739093349295e-06, "loss": 2.3399, "step": 76080 }, { "epoch": 1.72, "grad_norm": 1.6067295761472333, "learning_rate": 1.0397518678821261e-06, "loss": 2.407, "step": 76090 }, { "epoch": 1.72, "grad_norm": 1.5476381589655135, "learning_rate": 1.0381310213806651e-06, "loss": 2.4305, "step": 76100 }, { "epoch": 1.72, "grad_norm": 1.7374437180720517, "learning_rate": 1.0365113700467021e-06, "loss": 2.2783, "step": 76110 }, { "epoch": 1.72, "grad_norm": 1.7694053949692088, "learning_rate": 1.034892914096236e-06, "loss": 2.3725, "step": 76120 }, { "epoch": 1.72, "grad_norm": 1.7086976720487077, "learning_rate": 1.033275653745106e-06, "loss": 2.4691, "step": 76130 }, { "epoch": 1.72, "grad_norm": 1.8869964828356243, "learning_rate": 1.0316595892089965e-06, "loss": 2.3641, "step": 76140 }, { "epoch": 1.72, "grad_norm": 1.6841824911178298, "learning_rate": 1.0300447207034248e-06, "loss": 2.4084, "step": 76150 }, { "epoch": 1.72, "grad_norm": 1.5901768263949536, "learning_rate": 1.028431048443753e-06, "loss": 2.4145, "step": 76160 }, { "epoch": 1.72, "grad_norm": 1.574133487841402, "learning_rate": 1.026818572645183e-06, "loss": 2.3953, "step": 76170 }, { "epoch": 1.72, "grad_norm": 1.7030006573667673, "learning_rate": 1.0252072935227553e-06, "loss": 2.4048, "step": 76180 }, { "epoch": 1.72, "grad_norm": 1.8369958695472526, "learning_rate": 1.023597211291354e-06, "loss": 2.4086, "step": 76190 }, { "epoch": 1.72, "grad_norm": 1.6660944176395183, "learning_rate": 1.0219883261657016e-06, "loss": 2.3564, "step": 76200 }, { "epoch": 1.72, "grad_norm": 1.9547841473347045, "learning_rate": 1.020380638360361e-06, "loss": 2.4029, "step": 76210 }, { "epoch": 1.72, "grad_norm": 1.538047760337836, "learning_rate": 1.018774148089735e-06, "loss": 2.4362, "step": 76220 }, { "epoch": 1.72, "grad_norm": 1.7436137927055158, "learning_rate": 1.0171688555680681e-06, "loss": 2.4548, "step": 76230 }, { "epoch": 1.72, "grad_norm": 1.7861332919250756, "learning_rate": 1.0155647610094442e-06, "loss": 2.4273, "step": 76240 }, { "epoch": 1.72, "grad_norm": 1.7685675898914315, "learning_rate": 1.0139618646277894e-06, "loss": 2.427, "step": 76250 }, { "epoch": 1.72, "grad_norm": 1.841738436349606, "learning_rate": 1.0123601666368654e-06, "loss": 2.3543, "step": 76260 }, { "epoch": 1.72, "grad_norm": 1.5785565238015993, "learning_rate": 1.0107596672502763e-06, "loss": 2.4088, "step": 76270 }, { "epoch": 1.72, "grad_norm": 1.9180710721127645, "learning_rate": 1.0091603666814709e-06, "loss": 2.3706, "step": 76280 }, { "epoch": 1.72, "grad_norm": 1.4901313587942435, "learning_rate": 1.0075622651437323e-06, "loss": 2.4456, "step": 76290 }, { "epoch": 1.72, "grad_norm": 1.6463911391187867, "learning_rate": 1.0059653628501842e-06, "loss": 2.442, "step": 76300 }, { "epoch": 1.72, "grad_norm": 1.5386708520927221, "learning_rate": 1.0043696600137943e-06, "loss": 2.3146, "step": 76310 }, { "epoch": 1.72, "grad_norm": 1.6912078799829076, "learning_rate": 1.0027751568473676e-06, "loss": 2.3898, "step": 76320 }, { "epoch": 1.72, "grad_norm": 1.8272507566870608, "learning_rate": 1.0011818535635443e-06, "loss": 2.4385, "step": 76330 }, { "epoch": 1.72, "grad_norm": 1.7658143180837955, "learning_rate": 9.995897503748153e-07, "loss": 2.4389, "step": 76340 }, { "epoch": 1.72, "grad_norm": 1.6796374473054987, "learning_rate": 9.979988474935042e-07, "loss": 2.4996, "step": 76350 }, { "epoch": 1.72, "grad_norm": 1.6522267877645622, "learning_rate": 9.964091451317748e-07, "loss": 2.3868, "step": 76360 }, { "epoch": 1.72, "grad_norm": 1.6934259917104386, "learning_rate": 9.948206435016339e-07, "loss": 2.4005, "step": 76370 }, { "epoch": 1.72, "grad_norm": 1.8740384250493278, "learning_rate": 9.932333428149244e-07, "loss": 2.4857, "step": 76380 }, { "epoch": 1.72, "grad_norm": 1.720403532009949, "learning_rate": 9.916472432833324e-07, "loss": 2.3612, "step": 76390 }, { "epoch": 1.72, "grad_norm": 1.718978411384838, "learning_rate": 9.900623451183821e-07, "loss": 2.261, "step": 76400 }, { "epoch": 1.72, "grad_norm": 2.0179258779312796, "learning_rate": 9.884786485314357e-07, "loss": 2.3191, "step": 76410 }, { "epoch": 1.72, "grad_norm": 1.7919967358521338, "learning_rate": 9.868961537336997e-07, "loss": 2.4063, "step": 76420 }, { "epoch": 1.72, "grad_norm": 1.588903484991222, "learning_rate": 9.853148609362207e-07, "loss": 2.416, "step": 76430 }, { "epoch": 1.72, "grad_norm": 1.7360746188424703, "learning_rate": 9.837347703498756e-07, "loss": 2.3757, "step": 76440 }, { "epoch": 1.72, "grad_norm": 1.757414261090898, "learning_rate": 9.821558821853905e-07, "loss": 2.3795, "step": 76450 }, { "epoch": 1.72, "grad_norm": 1.7870474572622839, "learning_rate": 9.80578196653329e-07, "loss": 2.4756, "step": 76460 }, { "epoch": 1.72, "grad_norm": 1.4592704870504571, "learning_rate": 9.790017139640896e-07, "loss": 2.3832, "step": 76470 }, { "epoch": 1.72, "grad_norm": 1.6507723606019968, "learning_rate": 9.774264343279194e-07, "loss": 2.3939, "step": 76480 }, { "epoch": 1.72, "grad_norm": 1.5153338922546673, "learning_rate": 9.758523579548985e-07, "loss": 2.3754, "step": 76490 }, { "epoch": 1.73, "grad_norm": 1.8347978490313563, "learning_rate": 9.742794850549463e-07, "loss": 2.3105, "step": 76500 }, { "epoch": 1.73, "grad_norm": 1.7320440481432888, "learning_rate": 9.727078158378256e-07, "loss": 2.2565, "step": 76510 }, { "epoch": 1.73, "grad_norm": 1.5572422597120146, "learning_rate": 9.71137350513134e-07, "loss": 2.3629, "step": 76520 }, { "epoch": 1.73, "grad_norm": 1.6558975295852392, "learning_rate": 9.695680892903125e-07, "loss": 2.4999, "step": 76530 }, { "epoch": 1.73, "grad_norm": 1.7436176250272097, "learning_rate": 9.680000323786398e-07, "loss": 2.3473, "step": 76540 }, { "epoch": 1.73, "grad_norm": 1.8476858078811504, "learning_rate": 9.66433179987234e-07, "loss": 2.4131, "step": 76550 }, { "epoch": 1.73, "grad_norm": 1.8908876920757063, "learning_rate": 9.648675323250523e-07, "loss": 2.3762, "step": 76560 }, { "epoch": 1.73, "grad_norm": 1.4779380660116914, "learning_rate": 9.633030896008922e-07, "loss": 2.3896, "step": 76570 }, { "epoch": 1.73, "grad_norm": 1.7195620051310705, "learning_rate": 9.6173985202339e-07, "loss": 2.2507, "step": 76580 }, { "epoch": 1.73, "grad_norm": 1.62211509635309, "learning_rate": 9.601778198010215e-07, "loss": 2.4031, "step": 76590 }, { "epoch": 1.73, "grad_norm": 1.5484701185997733, "learning_rate": 9.586169931420986e-07, "loss": 2.3994, "step": 76600 }, { "epoch": 1.73, "grad_norm": 1.7901838769693688, "learning_rate": 9.570573722547804e-07, "loss": 2.381, "step": 76610 }, { "epoch": 1.73, "grad_norm": 1.6329647971521883, "learning_rate": 9.554989573470575e-07, "loss": 2.3192, "step": 76620 }, { "epoch": 1.73, "grad_norm": 1.7470309515942446, "learning_rate": 9.539417486267622e-07, "loss": 2.2989, "step": 76630 }, { "epoch": 1.73, "grad_norm": 1.6741062519020484, "learning_rate": 9.523857463015684e-07, "loss": 2.3304, "step": 76640 }, { "epoch": 1.73, "grad_norm": 1.585280893785019, "learning_rate": 9.508309505789837e-07, "loss": 2.4583, "step": 76650 }, { "epoch": 1.73, "grad_norm": 1.8204808471916367, "learning_rate": 9.492773616663565e-07, "loss": 2.4143, "step": 76660 }, { "epoch": 1.73, "grad_norm": 1.7379660068492897, "learning_rate": 9.477249797708788e-07, "loss": 2.2991, "step": 76670 }, { "epoch": 1.73, "grad_norm": 1.8216245834032296, "learning_rate": 9.461738050995784e-07, "loss": 2.5576, "step": 76680 }, { "epoch": 1.73, "grad_norm": 1.642531010162169, "learning_rate": 9.44623837859322e-07, "loss": 2.4753, "step": 76690 }, { "epoch": 1.73, "grad_norm": 1.8010506094464365, "learning_rate": 9.430750782568132e-07, "loss": 2.2703, "step": 76700 }, { "epoch": 1.73, "grad_norm": 1.6514344091792816, "learning_rate": 9.415275264986001e-07, "loss": 2.4155, "step": 76710 }, { "epoch": 1.73, "grad_norm": 1.8537009719084734, "learning_rate": 9.399811827910632e-07, "loss": 2.4452, "step": 76720 }, { "epoch": 1.73, "grad_norm": 1.7572930065432564, "learning_rate": 9.384360473404275e-07, "loss": 2.3529, "step": 76730 }, { "epoch": 1.73, "grad_norm": 1.6793117995631255, "learning_rate": 9.368921203527514e-07, "loss": 2.3425, "step": 76740 }, { "epoch": 1.73, "grad_norm": 1.8861760888530525, "learning_rate": 9.353494020339404e-07, "loss": 2.3566, "step": 76750 }, { "epoch": 1.73, "grad_norm": 1.683491491036234, "learning_rate": 9.338078925897287e-07, "loss": 2.3421, "step": 76760 }, { "epoch": 1.73, "grad_norm": 1.6385148374782383, "learning_rate": 9.322675922256963e-07, "loss": 2.3806, "step": 76770 }, { "epoch": 1.73, "grad_norm": 1.6734987075477552, "learning_rate": 9.307285011472599e-07, "loss": 2.3699, "step": 76780 }, { "epoch": 1.73, "grad_norm": 2.1703245392447625, "learning_rate": 9.291906195596734e-07, "loss": 2.3926, "step": 76790 }, { "epoch": 1.73, "grad_norm": 1.6478865883290068, "learning_rate": 9.276539476680302e-07, "loss": 2.3766, "step": 76800 }, { "epoch": 1.73, "grad_norm": 1.6029027700747422, "learning_rate": 9.261184856772676e-07, "loss": 2.4456, "step": 76810 }, { "epoch": 1.73, "grad_norm": 1.8238894175587472, "learning_rate": 9.245842337921529e-07, "loss": 2.4325, "step": 76820 }, { "epoch": 1.73, "grad_norm": 1.7601056541708284, "learning_rate": 9.23051192217298e-07, "loss": 2.4134, "step": 76830 }, { "epoch": 1.73, "grad_norm": 1.8489343539368366, "learning_rate": 9.215193611571505e-07, "loss": 2.4265, "step": 76840 }, { "epoch": 1.73, "grad_norm": 1.861406332292784, "learning_rate": 9.199887408159969e-07, "loss": 2.3582, "step": 76850 }, { "epoch": 1.73, "grad_norm": 1.6478090018616178, "learning_rate": 9.18459331397965e-07, "loss": 2.4459, "step": 76860 }, { "epoch": 1.73, "grad_norm": 1.5914372352919237, "learning_rate": 9.169311331070175e-07, "loss": 2.4307, "step": 76870 }, { "epoch": 1.73, "grad_norm": 1.777884103365022, "learning_rate": 9.154041461469576e-07, "loss": 2.395, "step": 76880 }, { "epoch": 1.73, "grad_norm": 1.6006045102414892, "learning_rate": 9.138783707214249e-07, "loss": 2.3396, "step": 76890 }, { "epoch": 1.73, "grad_norm": 1.596819917123227, "learning_rate": 9.123538070339022e-07, "loss": 2.4493, "step": 76900 }, { "epoch": 1.73, "grad_norm": 1.7915515173631587, "learning_rate": 9.108304552877046e-07, "loss": 2.2744, "step": 76910 }, { "epoch": 1.73, "grad_norm": 1.5995384911598123, "learning_rate": 9.093083156859894e-07, "loss": 2.4028, "step": 76920 }, { "epoch": 1.73, "grad_norm": 1.7264645570297943, "learning_rate": 9.077873884317512e-07, "loss": 2.3411, "step": 76930 }, { "epoch": 1.74, "grad_norm": 1.787700348604329, "learning_rate": 9.062676737278208e-07, "loss": 2.3622, "step": 76940 }, { "epoch": 1.74, "grad_norm": 1.6529794399026547, "learning_rate": 9.047491717768741e-07, "loss": 2.2468, "step": 76950 }, { "epoch": 1.74, "grad_norm": 1.8862213224763418, "learning_rate": 9.032318827814202e-07, "loss": 2.311, "step": 76960 }, { "epoch": 1.74, "grad_norm": 1.6386639334384678, "learning_rate": 9.017158069438036e-07, "loss": 2.4453, "step": 76970 }, { "epoch": 1.74, "grad_norm": 1.7493317509881094, "learning_rate": 9.002009444662118e-07, "loss": 2.3523, "step": 76980 }, { "epoch": 1.74, "grad_norm": 1.8099701823573255, "learning_rate": 8.986872955506687e-07, "loss": 2.4078, "step": 76990 }, { "epoch": 1.74, "grad_norm": 1.5960048711366879, "learning_rate": 8.971748603990349e-07, "loss": 2.3458, "step": 77000 }, { "epoch": 1.74, "grad_norm": 1.8285842939047452, "learning_rate": 8.956636392130147e-07, "loss": 2.4188, "step": 77010 }, { "epoch": 1.74, "grad_norm": 1.5798320596125504, "learning_rate": 8.941536321941446e-07, "loss": 2.2778, "step": 77020 }, { "epoch": 1.74, "grad_norm": 1.9087318119177543, "learning_rate": 8.926448395438025e-07, "loss": 2.4748, "step": 77030 }, { "epoch": 1.74, "grad_norm": 1.6926998002695808, "learning_rate": 8.911372614632008e-07, "loss": 2.3115, "step": 77040 }, { "epoch": 1.74, "grad_norm": 1.605851925199921, "learning_rate": 8.896308981533952e-07, "loss": 2.4397, "step": 77050 }, { "epoch": 1.74, "grad_norm": 1.8008972891286568, "learning_rate": 8.88125749815274e-07, "loss": 2.3612, "step": 77060 }, { "epoch": 1.74, "grad_norm": 1.5993391850968233, "learning_rate": 8.866218166495665e-07, "loss": 2.3982, "step": 77070 }, { "epoch": 1.74, "grad_norm": 1.9401401938067533, "learning_rate": 8.851190988568404e-07, "loss": 2.3597, "step": 77080 }, { "epoch": 1.74, "grad_norm": 1.6531557582058345, "learning_rate": 8.836175966374983e-07, "loss": 2.419, "step": 77090 }, { "epoch": 1.74, "grad_norm": 1.7046003358298654, "learning_rate": 8.821173101917846e-07, "loss": 2.4459, "step": 77100 }, { "epoch": 1.74, "grad_norm": 1.6763636090659533, "learning_rate": 8.806182397197793e-07, "loss": 2.2691, "step": 77110 }, { "epoch": 1.74, "grad_norm": 1.7360116544865922, "learning_rate": 8.791203854213992e-07, "loss": 2.3446, "step": 77120 }, { "epoch": 1.74, "grad_norm": 1.723826011386359, "learning_rate": 8.776237474964e-07, "loss": 2.5202, "step": 77130 }, { "epoch": 1.74, "grad_norm": 1.727745924319653, "learning_rate": 8.761283261443787e-07, "loss": 2.3493, "step": 77140 }, { "epoch": 1.74, "grad_norm": 1.8928183175032625, "learning_rate": 8.746341215647636e-07, "loss": 2.2905, "step": 77150 }, { "epoch": 1.74, "grad_norm": 1.7486148426969792, "learning_rate": 8.731411339568263e-07, "loss": 2.4212, "step": 77160 }, { "epoch": 1.74, "grad_norm": 1.8414099464242892, "learning_rate": 8.716493635196721e-07, "loss": 2.3455, "step": 77170 }, { "epoch": 1.74, "grad_norm": 1.620523006705583, "learning_rate": 8.701588104522473e-07, "loss": 2.3126, "step": 77180 }, { "epoch": 1.74, "grad_norm": 2.04715030153717, "learning_rate": 8.686694749533297e-07, "loss": 2.2618, "step": 77190 }, { "epoch": 1.74, "grad_norm": 1.8692348683620486, "learning_rate": 8.671813572215438e-07, "loss": 2.3651, "step": 77200 }, { "epoch": 1.74, "grad_norm": 1.765975642380116, "learning_rate": 8.656944574553461e-07, "loss": 2.4304, "step": 77210 }, { "epoch": 1.74, "grad_norm": 1.6325682722867179, "learning_rate": 8.642087758530305e-07, "loss": 2.4097, "step": 77220 }, { "epoch": 1.74, "grad_norm": 1.6692830807183419, "learning_rate": 8.627243126127316e-07, "loss": 2.3826, "step": 77230 }, { "epoch": 1.74, "grad_norm": 1.7527360712552602, "learning_rate": 8.612410679324168e-07, "loss": 2.2532, "step": 77240 }, { "epoch": 1.74, "grad_norm": 1.6438225914415436, "learning_rate": 8.597590420098956e-07, "loss": 2.278, "step": 77250 }, { "epoch": 1.74, "grad_norm": 2.0171238671545564, "learning_rate": 8.582782350428131e-07, "loss": 2.2376, "step": 77260 }, { "epoch": 1.74, "grad_norm": 2.1187520658233434, "learning_rate": 8.567986472286482e-07, "loss": 2.3278, "step": 77270 }, { "epoch": 1.74, "grad_norm": 1.6609580948801665, "learning_rate": 8.553202787647263e-07, "loss": 2.3724, "step": 77280 }, { "epoch": 1.74, "grad_norm": 1.8393315349804635, "learning_rate": 8.538431298482041e-07, "loss": 2.364, "step": 77290 }, { "epoch": 1.74, "grad_norm": 1.7930057498012943, "learning_rate": 8.52367200676073e-07, "loss": 2.3765, "step": 77300 }, { "epoch": 1.74, "grad_norm": 1.53541258243107, "learning_rate": 8.508924914451666e-07, "loss": 2.3038, "step": 77310 }, { "epoch": 1.74, "grad_norm": 1.8445275573653757, "learning_rate": 8.494190023521542e-07, "loss": 2.4195, "step": 77320 }, { "epoch": 1.74, "grad_norm": 1.574429739878022, "learning_rate": 8.479467335935399e-07, "loss": 2.444, "step": 77330 }, { "epoch": 1.74, "grad_norm": 1.8363709985489343, "learning_rate": 8.464756853656719e-07, "loss": 2.4001, "step": 77340 }, { "epoch": 1.74, "grad_norm": 1.7096972763774743, "learning_rate": 8.450058578647291e-07, "loss": 2.4382, "step": 77350 }, { "epoch": 1.74, "grad_norm": 1.5921557500171999, "learning_rate": 8.43537251286729e-07, "loss": 2.3735, "step": 77360 }, { "epoch": 1.74, "grad_norm": 1.57235997699336, "learning_rate": 8.420698658275295e-07, "loss": 2.4311, "step": 77370 }, { "epoch": 1.74, "grad_norm": 1.7476526560614938, "learning_rate": 8.406037016828217e-07, "loss": 2.4362, "step": 77380 }, { "epoch": 1.75, "grad_norm": 2.3233827031136176, "learning_rate": 8.391387590481337e-07, "loss": 2.2726, "step": 77390 }, { "epoch": 1.75, "grad_norm": 1.8503823989765193, "learning_rate": 8.376750381188359e-07, "loss": 2.4066, "step": 77400 }, { "epoch": 1.75, "grad_norm": 1.8098874962341298, "learning_rate": 8.362125390901287e-07, "loss": 2.313, "step": 77410 }, { "epoch": 1.75, "grad_norm": 1.6353363702670507, "learning_rate": 8.347512621570552e-07, "loss": 2.3181, "step": 77420 }, { "epoch": 1.75, "grad_norm": 1.7853196707026124, "learning_rate": 8.332912075144928e-07, "loss": 2.3512, "step": 77430 }, { "epoch": 1.75, "grad_norm": 1.685034453135321, "learning_rate": 8.318323753571555e-07, "loss": 2.3282, "step": 77440 }, { "epoch": 1.75, "grad_norm": 1.6901740463354238, "learning_rate": 8.303747658795969e-07, "loss": 2.3671, "step": 77450 }, { "epoch": 1.75, "grad_norm": 1.6735500368149032, "learning_rate": 8.289183792762045e-07, "loss": 2.3828, "step": 77460 }, { "epoch": 1.75, "grad_norm": 1.9507511476850476, "learning_rate": 8.274632157412033e-07, "loss": 2.4471, "step": 77470 }, { "epoch": 1.75, "grad_norm": 1.8073558298747932, "learning_rate": 8.260092754686589e-07, "loss": 2.3273, "step": 77480 }, { "epoch": 1.75, "grad_norm": 1.7673142350903788, "learning_rate": 8.245565586524696e-07, "loss": 2.4383, "step": 77490 }, { "epoch": 1.75, "grad_norm": 1.7842875566712575, "learning_rate": 8.231050654863726e-07, "loss": 2.3466, "step": 77500 }, { "epoch": 1.75, "grad_norm": 1.4388777267111301, "learning_rate": 8.216547961639377e-07, "loss": 2.4566, "step": 77510 }, { "epoch": 1.75, "grad_norm": 1.690443910886374, "learning_rate": 8.202057508785777e-07, "loss": 2.3279, "step": 77520 }, { "epoch": 1.75, "grad_norm": 1.7429719853189034, "learning_rate": 8.18757929823536e-07, "loss": 2.4161, "step": 77530 }, { "epoch": 1.75, "grad_norm": 1.755817434045906, "learning_rate": 8.173113331919014e-07, "loss": 2.3213, "step": 77540 }, { "epoch": 1.75, "grad_norm": 1.5224281216008666, "learning_rate": 8.158659611765907e-07, "loss": 2.4786, "step": 77550 }, { "epoch": 1.75, "grad_norm": 1.8319803576292897, "learning_rate": 8.144218139703608e-07, "loss": 2.3165, "step": 77560 }, { "epoch": 1.75, "grad_norm": 1.7376665762275736, "learning_rate": 8.129788917658066e-07, "loss": 2.2753, "step": 77570 }, { "epoch": 1.75, "grad_norm": 1.4869240395928267, "learning_rate": 8.115371947553574e-07, "loss": 2.4013, "step": 77580 }, { "epoch": 1.75, "grad_norm": 1.7987930277941258, "learning_rate": 8.100967231312806e-07, "loss": 2.4746, "step": 77590 }, { "epoch": 1.75, "grad_norm": 1.5154192736898942, "learning_rate": 8.086574770856759e-07, "loss": 2.3179, "step": 77600 }, { "epoch": 1.75, "grad_norm": 1.894606202945964, "learning_rate": 8.072194568104907e-07, "loss": 2.4161, "step": 77610 }, { "epoch": 1.75, "grad_norm": 1.8127137909957836, "learning_rate": 8.057826624974951e-07, "loss": 2.3635, "step": 77620 }, { "epoch": 1.75, "grad_norm": 1.5384577467194482, "learning_rate": 8.043470943383036e-07, "loss": 2.3991, "step": 77630 }, { "epoch": 1.75, "grad_norm": 1.7687636599440428, "learning_rate": 8.029127525243651e-07, "loss": 2.4069, "step": 77640 }, { "epoch": 1.75, "grad_norm": 1.844577344531668, "learning_rate": 8.01479637246968e-07, "loss": 2.2517, "step": 77650 }, { "epoch": 1.75, "grad_norm": 1.8073064304929618, "learning_rate": 8.000477486972291e-07, "loss": 2.352, "step": 77660 }, { "epoch": 1.75, "grad_norm": 1.6189062171813475, "learning_rate": 7.986170870661136e-07, "loss": 2.3647, "step": 77670 }, { "epoch": 1.75, "grad_norm": 1.7538849694878242, "learning_rate": 7.971876525444133e-07, "loss": 2.4173, "step": 77680 }, { "epoch": 1.75, "grad_norm": 1.9120158858974579, "learning_rate": 7.957594453227601e-07, "loss": 2.3102, "step": 77690 }, { "epoch": 1.75, "grad_norm": 1.5735511894696905, "learning_rate": 7.943324655916218e-07, "loss": 2.3438, "step": 77700 }, { "epoch": 1.75, "grad_norm": 1.9356214902843507, "learning_rate": 7.929067135413027e-07, "loss": 2.3047, "step": 77710 }, { "epoch": 1.75, "grad_norm": 1.6205539177684183, "learning_rate": 7.914821893619418e-07, "loss": 2.4353, "step": 77720 }, { "epoch": 1.75, "grad_norm": 2.170598188412539, "learning_rate": 7.900588932435161e-07, "loss": 2.3559, "step": 77730 }, { "epoch": 1.75, "grad_norm": 1.6136110459514488, "learning_rate": 7.886368253758403e-07, "loss": 2.3412, "step": 77740 }, { "epoch": 1.75, "grad_norm": 1.8392159228984406, "learning_rate": 7.872159859485607e-07, "loss": 2.3799, "step": 77750 }, { "epoch": 1.75, "grad_norm": 1.9251192622594364, "learning_rate": 7.857963751511644e-07, "loss": 2.4502, "step": 77760 }, { "epoch": 1.75, "grad_norm": 1.5850440240155483, "learning_rate": 7.843779931729712e-07, "loss": 2.4621, "step": 77770 }, { "epoch": 1.75, "grad_norm": 1.7268161789201177, "learning_rate": 7.829608402031397e-07, "loss": 2.3547, "step": 77780 }, { "epoch": 1.75, "grad_norm": 1.762708182852948, "learning_rate": 7.815449164306632e-07, "loss": 2.3416, "step": 77790 }, { "epoch": 1.75, "grad_norm": 1.5789151016625205, "learning_rate": 7.801302220443685e-07, "loss": 2.3863, "step": 77800 }, { "epoch": 1.75, "grad_norm": 2.191171460728779, "learning_rate": 7.787167572329268e-07, "loss": 2.238, "step": 77810 }, { "epoch": 1.75, "grad_norm": 1.4750501827416447, "learning_rate": 7.773045221848352e-07, "loss": 2.307, "step": 77820 }, { "epoch": 1.76, "grad_norm": 1.5585413663594667, "learning_rate": 7.758935170884363e-07, "loss": 2.421, "step": 77830 }, { "epoch": 1.76, "grad_norm": 1.7023286780950744, "learning_rate": 7.744837421318974e-07, "loss": 2.401, "step": 77840 }, { "epoch": 1.76, "grad_norm": 1.6431332171289608, "learning_rate": 7.730751975032314e-07, "loss": 2.3649, "step": 77850 }, { "epoch": 1.76, "grad_norm": 1.671685107919373, "learning_rate": 7.716678833902824e-07, "loss": 2.3476, "step": 77860 }, { "epoch": 1.76, "grad_norm": 1.6487221631828726, "learning_rate": 7.702617999807338e-07, "loss": 2.3672, "step": 77870 }, { "epoch": 1.76, "grad_norm": 1.5856582429092796, "learning_rate": 7.688569474621022e-07, "loss": 2.354, "step": 77880 }, { "epoch": 1.76, "grad_norm": 1.6550029191238376, "learning_rate": 7.674533260217398e-07, "loss": 2.309, "step": 77890 }, { "epoch": 1.76, "grad_norm": 1.8353970695707433, "learning_rate": 7.660509358468371e-07, "loss": 2.3064, "step": 77900 }, { "epoch": 1.76, "grad_norm": 1.6957165068905646, "learning_rate": 7.646497771244177e-07, "loss": 2.4326, "step": 77910 }, { "epoch": 1.76, "grad_norm": 1.667000411086318, "learning_rate": 7.632498500413421e-07, "loss": 2.4606, "step": 77920 }, { "epoch": 1.76, "grad_norm": 1.5643748513110693, "learning_rate": 7.618511547843066e-07, "loss": 2.3198, "step": 77930 }, { "epoch": 1.76, "grad_norm": 1.5531970371099244, "learning_rate": 7.604536915398442e-07, "loss": 2.2345, "step": 77940 }, { "epoch": 1.76, "grad_norm": 1.5845349398685196, "learning_rate": 7.590574604943202e-07, "loss": 2.3178, "step": 77950 }, { "epoch": 1.76, "grad_norm": 1.8430226086654464, "learning_rate": 7.576624618339401e-07, "loss": 2.3116, "step": 77960 }, { "epoch": 1.76, "grad_norm": 1.891148525558062, "learning_rate": 7.562686957447429e-07, "loss": 2.3335, "step": 77970 }, { "epoch": 1.76, "grad_norm": 1.582561054774657, "learning_rate": 7.548761624126022e-07, "loss": 2.3787, "step": 77980 }, { "epoch": 1.76, "grad_norm": 1.8445622512663928, "learning_rate": 7.534848620232282e-07, "loss": 2.5021, "step": 77990 }, { "epoch": 1.76, "grad_norm": 1.6404251331685697, "learning_rate": 7.52094794762166e-07, "loss": 2.4612, "step": 78000 }, { "epoch": 1.76, "grad_norm": 1.6058261220807437, "learning_rate": 7.507059608147993e-07, "loss": 2.3511, "step": 78010 }, { "epoch": 1.76, "grad_norm": 1.7380858299096609, "learning_rate": 7.493183603663434e-07, "loss": 2.1859, "step": 78020 }, { "epoch": 1.76, "grad_norm": 1.6143967437350357, "learning_rate": 7.479319936018503e-07, "loss": 2.2929, "step": 78030 }, { "epoch": 1.76, "grad_norm": 1.6806624256500393, "learning_rate": 7.46546860706211e-07, "loss": 2.3851, "step": 78040 }, { "epoch": 1.76, "grad_norm": 1.9007041729047973, "learning_rate": 7.451629618641443e-07, "loss": 2.2922, "step": 78050 }, { "epoch": 1.76, "grad_norm": 1.7922079633405328, "learning_rate": 7.437802972602092e-07, "loss": 2.3117, "step": 78060 }, { "epoch": 1.76, "grad_norm": 1.8758624699299662, "learning_rate": 7.42398867078803e-07, "loss": 2.5314, "step": 78070 }, { "epoch": 1.76, "grad_norm": 1.8629442117690835, "learning_rate": 7.410186715041534e-07, "loss": 2.299, "step": 78080 }, { "epoch": 1.76, "grad_norm": 1.7038163659760932, "learning_rate": 7.396397107203257e-07, "loss": 2.4065, "step": 78090 }, { "epoch": 1.76, "grad_norm": 1.9126959118856142, "learning_rate": 7.382619849112205e-07, "loss": 2.3933, "step": 78100 }, { "epoch": 1.76, "grad_norm": 1.6322073751276611, "learning_rate": 7.368854942605719e-07, "loss": 2.3467, "step": 78110 }, { "epoch": 1.76, "grad_norm": 2.0188818908879798, "learning_rate": 7.355102389519509e-07, "loss": 2.3409, "step": 78120 }, { "epoch": 1.76, "grad_norm": 1.7804490595305338, "learning_rate": 7.341362191687618e-07, "loss": 2.2684, "step": 78130 }, { "epoch": 1.76, "grad_norm": 1.5783794607636359, "learning_rate": 7.327634350942503e-07, "loss": 2.3076, "step": 78140 }, { "epoch": 1.76, "grad_norm": 1.6121059382301088, "learning_rate": 7.313918869114923e-07, "loss": 2.4593, "step": 78150 }, { "epoch": 1.76, "grad_norm": 1.6616884812055706, "learning_rate": 7.300215748033956e-07, "loss": 2.3176, "step": 78160 }, { "epoch": 1.76, "grad_norm": 1.842334734764344, "learning_rate": 7.286524989527078e-07, "loss": 2.3858, "step": 78170 }, { "epoch": 1.76, "grad_norm": 1.6667628124543243, "learning_rate": 7.272846595420136e-07, "loss": 2.4193, "step": 78180 }, { "epoch": 1.76, "grad_norm": 1.5368467252968936, "learning_rate": 7.259180567537249e-07, "loss": 2.4935, "step": 78190 }, { "epoch": 1.76, "grad_norm": 1.6959345542895463, "learning_rate": 7.245526907700995e-07, "loss": 2.395, "step": 78200 }, { "epoch": 1.76, "grad_norm": 1.7440337588482004, "learning_rate": 7.231885617732226e-07, "loss": 2.3058, "step": 78210 }, { "epoch": 1.76, "grad_norm": 1.758469766944562, "learning_rate": 7.218256699450155e-07, "loss": 2.3687, "step": 78220 }, { "epoch": 1.76, "grad_norm": 1.8296380431436188, "learning_rate": 7.20464015467236e-07, "loss": 2.2343, "step": 78230 }, { "epoch": 1.76, "grad_norm": 1.6531125906145208, "learning_rate": 7.191035985214756e-07, "loss": 2.2651, "step": 78240 }, { "epoch": 1.76, "grad_norm": 1.7375942672205549, "learning_rate": 7.177444192891624e-07, "loss": 2.1639, "step": 78250 }, { "epoch": 1.76, "grad_norm": 1.6166808872304415, "learning_rate": 7.163864779515572e-07, "loss": 2.5261, "step": 78260 }, { "epoch": 1.77, "grad_norm": 1.760296883053247, "learning_rate": 7.150297746897583e-07, "loss": 2.3064, "step": 78270 }, { "epoch": 1.77, "grad_norm": 1.6666689619695534, "learning_rate": 7.136743096846976e-07, "loss": 2.3749, "step": 78280 }, { "epoch": 1.77, "grad_norm": 1.7386696260165158, "learning_rate": 7.123200831171406e-07, "loss": 2.3651, "step": 78290 }, { "epoch": 1.77, "grad_norm": 1.8124745701731613, "learning_rate": 7.109670951676895e-07, "loss": 2.3127, "step": 78300 }, { "epoch": 1.77, "grad_norm": 1.5306799742563024, "learning_rate": 7.096153460167799e-07, "loss": 2.4064, "step": 78310 }, { "epoch": 1.77, "grad_norm": 1.710382053424115, "learning_rate": 7.082648358446842e-07, "loss": 2.3702, "step": 78320 }, { "epoch": 1.77, "grad_norm": 1.6809909729071815, "learning_rate": 7.069155648315063e-07, "loss": 2.3929, "step": 78330 }, { "epoch": 1.77, "grad_norm": 1.6645954760812178, "learning_rate": 7.055675331571909e-07, "loss": 2.2882, "step": 78340 }, { "epoch": 1.77, "grad_norm": 1.6236375717758669, "learning_rate": 7.042207410015101e-07, "loss": 2.1632, "step": 78350 }, { "epoch": 1.77, "grad_norm": 1.7026081380518454, "learning_rate": 7.028751885440755e-07, "loss": 2.3476, "step": 78360 }, { "epoch": 1.77, "grad_norm": 1.6398815318437072, "learning_rate": 7.015308759643325e-07, "loss": 2.4923, "step": 78370 }, { "epoch": 1.77, "grad_norm": 1.6318255294154258, "learning_rate": 7.001878034415566e-07, "loss": 2.3133, "step": 78380 }, { "epoch": 1.77, "grad_norm": 1.676440106888097, "learning_rate": 6.988459711548645e-07, "loss": 2.5385, "step": 78390 }, { "epoch": 1.77, "grad_norm": 2.089005159234158, "learning_rate": 6.975053792832054e-07, "loss": 2.3335, "step": 78400 }, { "epoch": 1.77, "grad_norm": 1.943005812908789, "learning_rate": 6.961660280053628e-07, "loss": 2.2421, "step": 78410 }, { "epoch": 1.77, "grad_norm": 1.5275315386574333, "learning_rate": 6.948279174999528e-07, "loss": 2.423, "step": 78420 }, { "epoch": 1.77, "grad_norm": 1.9232833768320798, "learning_rate": 6.934910479454293e-07, "loss": 2.3346, "step": 78430 }, { "epoch": 1.77, "grad_norm": 1.844309774785893, "learning_rate": 6.921554195200786e-07, "loss": 2.4027, "step": 78440 }, { "epoch": 1.77, "grad_norm": 1.7999155282833996, "learning_rate": 6.908210324020215e-07, "loss": 2.4068, "step": 78450 }, { "epoch": 1.77, "grad_norm": 1.7581310570950903, "learning_rate": 6.894878867692134e-07, "loss": 2.3124, "step": 78460 }, { "epoch": 1.77, "grad_norm": 1.911279503130346, "learning_rate": 6.881559827994444e-07, "loss": 2.428, "step": 78470 }, { "epoch": 1.77, "grad_norm": 1.8265083365428745, "learning_rate": 6.868253206703424e-07, "loss": 2.4042, "step": 78480 }, { "epoch": 1.77, "grad_norm": 1.788172601277215, "learning_rate": 6.85495900559362e-07, "loss": 2.3186, "step": 78490 }, { "epoch": 1.77, "grad_norm": 1.9832525936589713, "learning_rate": 6.841677226437993e-07, "loss": 2.3565, "step": 78500 }, { "epoch": 1.77, "grad_norm": 1.8183888655172271, "learning_rate": 6.828407871007803e-07, "loss": 2.4162, "step": 78510 }, { "epoch": 1.77, "grad_norm": 1.511325962205788, "learning_rate": 6.815150941072679e-07, "loss": 2.4323, "step": 78520 }, { "epoch": 1.77, "grad_norm": 1.8423344753399868, "learning_rate": 6.801906438400563e-07, "loss": 2.2885, "step": 78530 }, { "epoch": 1.77, "grad_norm": 1.5952254116583693, "learning_rate": 6.788674364757797e-07, "loss": 2.359, "step": 78540 }, { "epoch": 1.77, "grad_norm": 1.6498287882220841, "learning_rate": 6.775454721909014e-07, "loss": 2.2864, "step": 78550 }, { "epoch": 1.77, "grad_norm": 1.7031219169309153, "learning_rate": 6.762247511617204e-07, "loss": 2.3486, "step": 78560 }, { "epoch": 1.77, "grad_norm": 1.8228537563525538, "learning_rate": 6.749052735643691e-07, "loss": 2.3816, "step": 78570 }, { "epoch": 1.77, "grad_norm": 1.6994501713106283, "learning_rate": 6.735870395748178e-07, "loss": 2.3331, "step": 78580 }, { "epoch": 1.77, "grad_norm": 1.7047405170376444, "learning_rate": 6.722700493688617e-07, "loss": 2.3935, "step": 78590 }, { "epoch": 1.77, "grad_norm": 1.8940693725700117, "learning_rate": 6.709543031221433e-07, "loss": 2.4728, "step": 78600 }, { "epoch": 1.77, "grad_norm": 1.8741761942432322, "learning_rate": 6.696398010101279e-07, "loss": 2.3801, "step": 78610 }, { "epoch": 1.77, "grad_norm": 1.735232401241559, "learning_rate": 6.683265432081221e-07, "loss": 2.3181, "step": 78620 }, { "epoch": 1.77, "grad_norm": 1.7179390158783918, "learning_rate": 6.670145298912633e-07, "loss": 2.31, "step": 78630 }, { "epoch": 1.77, "grad_norm": 1.7373748586281865, "learning_rate": 6.657037612345218e-07, "loss": 2.439, "step": 78640 }, { "epoch": 1.77, "grad_norm": 1.8040247346902, "learning_rate": 6.643942374127055e-07, "loss": 2.4268, "step": 78650 }, { "epoch": 1.77, "grad_norm": 1.7105238714674362, "learning_rate": 6.630859586004512e-07, "loss": 2.4683, "step": 78660 }, { "epoch": 1.77, "grad_norm": 1.6670587560902075, "learning_rate": 6.617789249722373e-07, "loss": 2.4895, "step": 78670 }, { "epoch": 1.77, "grad_norm": 1.843203032501078, "learning_rate": 6.604731367023686e-07, "loss": 2.4732, "step": 78680 }, { "epoch": 1.77, "grad_norm": 1.854001726402219, "learning_rate": 6.591685939649905e-07, "loss": 2.3741, "step": 78690 }, { "epoch": 1.77, "grad_norm": 1.6471829669940283, "learning_rate": 6.578652969340737e-07, "loss": 2.4619, "step": 78700 }, { "epoch": 1.77, "grad_norm": 1.7505740524382565, "learning_rate": 6.565632457834304e-07, "loss": 2.4597, "step": 78710 }, { "epoch": 1.78, "grad_norm": 1.7640517976628238, "learning_rate": 6.552624406867014e-07, "loss": 2.3923, "step": 78720 }, { "epoch": 1.78, "grad_norm": 1.7033542542659315, "learning_rate": 6.539628818173682e-07, "loss": 2.3728, "step": 78730 }, { "epoch": 1.78, "grad_norm": 1.707624812437689, "learning_rate": 6.526645693487387e-07, "loss": 2.5258, "step": 78740 }, { "epoch": 1.78, "grad_norm": 1.6520146836178236, "learning_rate": 6.51367503453959e-07, "loss": 2.3565, "step": 78750 }, { "epoch": 1.78, "grad_norm": 1.8605586365017375, "learning_rate": 6.500716843060073e-07, "loss": 2.2578, "step": 78760 }, { "epoch": 1.78, "grad_norm": 1.5819500553598487, "learning_rate": 6.487771120776964e-07, "loss": 2.2447, "step": 78770 }, { "epoch": 1.78, "grad_norm": 1.8024952986477218, "learning_rate": 6.474837869416706e-07, "loss": 2.395, "step": 78780 }, { "epoch": 1.78, "grad_norm": 2.4865467096838483, "learning_rate": 6.461917090704118e-07, "loss": 2.3742, "step": 78790 }, { "epoch": 1.78, "grad_norm": 1.7773068085595831, "learning_rate": 6.44900878636231e-07, "loss": 2.3048, "step": 78800 }, { "epoch": 1.78, "grad_norm": 1.7258928509878628, "learning_rate": 6.436112958112772e-07, "loss": 2.4366, "step": 78810 }, { "epoch": 1.78, "grad_norm": 1.8211458792938235, "learning_rate": 6.423229607675296e-07, "loss": 2.3325, "step": 78820 }, { "epoch": 1.78, "grad_norm": 1.522777420687436, "learning_rate": 6.410358736768041e-07, "loss": 2.2851, "step": 78830 }, { "epoch": 1.78, "grad_norm": 1.6124135395919559, "learning_rate": 6.397500347107466e-07, "loss": 2.3875, "step": 78840 }, { "epoch": 1.78, "grad_norm": 2.064076611217767, "learning_rate": 6.384654440408389e-07, "loss": 2.4238, "step": 78850 }, { "epoch": 1.78, "grad_norm": 1.826319911726033, "learning_rate": 6.371821018383939e-07, "loss": 2.3099, "step": 78860 }, { "epoch": 1.78, "grad_norm": 1.7780625315137486, "learning_rate": 6.359000082745647e-07, "loss": 2.3981, "step": 78870 }, { "epoch": 1.78, "grad_norm": 1.6078517996815893, "learning_rate": 6.346191635203303e-07, "loss": 2.3601, "step": 78880 }, { "epoch": 1.78, "grad_norm": 1.7816215270651927, "learning_rate": 6.333395677465048e-07, "loss": 2.3808, "step": 78890 }, { "epoch": 1.78, "grad_norm": 1.5985291334197627, "learning_rate": 6.320612211237398e-07, "loss": 2.2998, "step": 78900 }, { "epoch": 1.78, "grad_norm": 1.8459073299778528, "learning_rate": 6.307841238225165e-07, "loss": 2.4481, "step": 78910 }, { "epoch": 1.78, "grad_norm": 1.6469925605552733, "learning_rate": 6.295082760131466e-07, "loss": 2.2854, "step": 78920 }, { "epoch": 1.78, "grad_norm": 1.8140727090141688, "learning_rate": 6.282336778657838e-07, "loss": 2.3195, "step": 78930 }, { "epoch": 1.78, "grad_norm": 1.544794041964966, "learning_rate": 6.26960329550409e-07, "loss": 2.4435, "step": 78940 }, { "epoch": 1.78, "grad_norm": 1.7230083753117869, "learning_rate": 6.256882312368373e-07, "loss": 2.3587, "step": 78950 }, { "epoch": 1.78, "grad_norm": 1.4627654236525385, "learning_rate": 6.244173830947175e-07, "loss": 2.3618, "step": 78960 }, { "epoch": 1.78, "grad_norm": 1.6079112471565278, "learning_rate": 6.231477852935308e-07, "loss": 2.3913, "step": 78970 }, { "epoch": 1.78, "grad_norm": 1.7663407923823897, "learning_rate": 6.21879438002595e-07, "loss": 2.3292, "step": 78980 }, { "epoch": 1.78, "grad_norm": 1.5781141111077173, "learning_rate": 6.206123413910559e-07, "loss": 2.394, "step": 78990 }, { "epoch": 1.78, "grad_norm": 1.594654719801637, "learning_rate": 6.19346495627896e-07, "loss": 2.3667, "step": 79000 }, { "epoch": 1.78, "grad_norm": 1.7288259659989509, "learning_rate": 6.180819008819328e-07, "loss": 2.4134, "step": 79010 }, { "epoch": 1.78, "grad_norm": 1.6673825137095717, "learning_rate": 6.168185573218122e-07, "loss": 2.3551, "step": 79020 }, { "epoch": 1.78, "grad_norm": 1.484966713965944, "learning_rate": 6.155564651160151e-07, "loss": 2.3879, "step": 79030 }, { "epoch": 1.78, "grad_norm": 1.536213998137711, "learning_rate": 6.142956244328568e-07, "loss": 2.3381, "step": 79040 }, { "epoch": 1.78, "grad_norm": 1.7805375828187018, "learning_rate": 6.130360354404841e-07, "loss": 2.3005, "step": 79050 }, { "epoch": 1.78, "grad_norm": 1.6203931878148048, "learning_rate": 6.117776983068769e-07, "loss": 2.3305, "step": 79060 }, { "epoch": 1.78, "grad_norm": 1.836532970960088, "learning_rate": 6.105206131998509e-07, "loss": 2.4048, "step": 79070 }, { "epoch": 1.78, "grad_norm": 1.7261944629209793, "learning_rate": 6.092647802870533e-07, "loss": 2.288, "step": 79080 }, { "epoch": 1.78, "grad_norm": 1.6428138144632767, "learning_rate": 6.080101997359611e-07, "loss": 2.2914, "step": 79090 }, { "epoch": 1.78, "grad_norm": 1.7304457878552355, "learning_rate": 6.067568717138883e-07, "loss": 2.3963, "step": 79100 }, { "epoch": 1.78, "grad_norm": 1.5743659832845878, "learning_rate": 6.0550479638798e-07, "loss": 2.2942, "step": 79110 }, { "epoch": 1.78, "grad_norm": 1.6165550931058505, "learning_rate": 6.042539739252151e-07, "loss": 2.3153, "step": 79120 }, { "epoch": 1.78, "grad_norm": 1.7501911806994253, "learning_rate": 6.030044044924055e-07, "loss": 2.3077, "step": 79130 }, { "epoch": 1.78, "grad_norm": 1.5824398150659833, "learning_rate": 6.017560882561946e-07, "loss": 2.4175, "step": 79140 }, { "epoch": 1.78, "grad_norm": 1.5370204252945063, "learning_rate": 6.005090253830603e-07, "loss": 2.49, "step": 79150 }, { "epoch": 1.79, "grad_norm": 1.9272577237880804, "learning_rate": 5.99263216039312e-07, "loss": 2.3647, "step": 79160 }, { "epoch": 1.79, "grad_norm": 1.7231247401923273, "learning_rate": 5.980186603910932e-07, "loss": 2.3412, "step": 79170 }, { "epoch": 1.79, "grad_norm": 1.8963470606984854, "learning_rate": 5.96775358604379e-07, "loss": 2.4261, "step": 79180 }, { "epoch": 1.79, "grad_norm": 1.9450439379557565, "learning_rate": 5.955333108449768e-07, "loss": 2.3426, "step": 79190 }, { "epoch": 1.79, "grad_norm": 1.776873422999953, "learning_rate": 5.942925172785319e-07, "loss": 2.3829, "step": 79200 }, { "epoch": 1.79, "grad_norm": 1.763081839650979, "learning_rate": 5.930529780705141e-07, "loss": 2.415, "step": 79210 }, { "epoch": 1.79, "grad_norm": 1.582926958372157, "learning_rate": 5.918146933862334e-07, "loss": 2.3663, "step": 79220 }, { "epoch": 1.79, "grad_norm": 1.8409373096395225, "learning_rate": 5.905776633908278e-07, "loss": 2.3576, "step": 79230 }, { "epoch": 1.79, "grad_norm": 1.7513694044801293, "learning_rate": 5.893418882492674e-07, "loss": 2.3799, "step": 79240 }, { "epoch": 1.79, "grad_norm": 1.7835047948772487, "learning_rate": 5.881073681263572e-07, "loss": 2.37, "step": 79250 }, { "epoch": 1.79, "grad_norm": 1.5328189234237668, "learning_rate": 5.868741031867375e-07, "loss": 2.4419, "step": 79260 }, { "epoch": 1.79, "grad_norm": 1.688724465339995, "learning_rate": 5.85642093594877e-07, "loss": 2.3566, "step": 79270 }, { "epoch": 1.79, "grad_norm": 1.7092682197999403, "learning_rate": 5.844113395150785e-07, "loss": 2.3441, "step": 79280 }, { "epoch": 1.79, "grad_norm": 1.6902317602926955, "learning_rate": 5.831818411114753e-07, "loss": 2.2805, "step": 79290 }, { "epoch": 1.79, "grad_norm": 1.7735160325472055, "learning_rate": 5.819535985480363e-07, "loss": 2.272, "step": 79300 }, { "epoch": 1.79, "grad_norm": 1.6334327611167745, "learning_rate": 5.807266119885624e-07, "loss": 2.4483, "step": 79310 }, { "epoch": 1.79, "grad_norm": 1.8606495928588709, "learning_rate": 5.795008815966852e-07, "loss": 2.3016, "step": 79320 }, { "epoch": 1.79, "grad_norm": 1.558100631005257, "learning_rate": 5.782764075358682e-07, "loss": 2.3096, "step": 79330 }, { "epoch": 1.79, "grad_norm": 1.7977242740958919, "learning_rate": 5.770531899694142e-07, "loss": 2.3324, "step": 79340 }, { "epoch": 1.79, "grad_norm": 1.8198632539879922, "learning_rate": 5.758312290604485e-07, "loss": 2.3316, "step": 79350 }, { "epoch": 1.79, "grad_norm": 1.7159466558699028, "learning_rate": 5.746105249719347e-07, "loss": 2.2741, "step": 79360 }, { "epoch": 1.79, "grad_norm": 1.6547537696027281, "learning_rate": 5.733910778666674e-07, "loss": 2.3808, "step": 79370 }, { "epoch": 1.79, "grad_norm": 1.7347919894379664, "learning_rate": 5.721728879072752e-07, "loss": 2.3473, "step": 79380 }, { "epoch": 1.79, "grad_norm": 1.6557705071238282, "learning_rate": 5.70955955256215e-07, "loss": 2.2997, "step": 79390 }, { "epoch": 1.79, "grad_norm": 1.698837998631402, "learning_rate": 5.697402800757812e-07, "loss": 2.4628, "step": 79400 }, { "epoch": 1.79, "grad_norm": 1.633475562695627, "learning_rate": 5.685258625280976e-07, "loss": 2.3506, "step": 79410 }, { "epoch": 1.79, "grad_norm": 1.6179123788405623, "learning_rate": 5.673127027751201e-07, "loss": 2.2948, "step": 79420 }, { "epoch": 1.79, "grad_norm": 1.6651641838090914, "learning_rate": 5.661008009786373e-07, "loss": 2.3638, "step": 79430 }, { "epoch": 1.79, "grad_norm": 1.6288245636527887, "learning_rate": 5.648901573002696e-07, "loss": 2.4234, "step": 79440 }, { "epoch": 1.79, "grad_norm": 1.7384680718609855, "learning_rate": 5.636807719014725e-07, "loss": 2.4771, "step": 79450 }, { "epoch": 1.79, "grad_norm": 1.8364231798592392, "learning_rate": 5.624726449435281e-07, "loss": 2.3962, "step": 79460 }, { "epoch": 1.79, "grad_norm": 1.761449312338436, "learning_rate": 5.612657765875562e-07, "loss": 2.3539, "step": 79470 }, { "epoch": 1.79, "grad_norm": 1.723697431914996, "learning_rate": 5.60060166994506e-07, "loss": 2.3249, "step": 79480 }, { "epoch": 1.79, "grad_norm": 1.60308064031907, "learning_rate": 5.58855816325159e-07, "loss": 2.2859, "step": 79490 }, { "epoch": 1.79, "grad_norm": 1.7338594144196506, "learning_rate": 5.576527247401287e-07, "loss": 2.3668, "step": 79500 }, { "epoch": 1.79, "grad_norm": 1.7527166775215566, "learning_rate": 5.564508923998613e-07, "loss": 2.3765, "step": 79510 }, { "epoch": 1.79, "grad_norm": 1.6715512335372413, "learning_rate": 5.552503194646353e-07, "loss": 2.298, "step": 79520 }, { "epoch": 1.79, "grad_norm": 1.7946806501679249, "learning_rate": 5.540510060945592e-07, "loss": 2.3257, "step": 79530 }, { "epoch": 1.79, "grad_norm": 1.6260329331588008, "learning_rate": 5.528529524495785e-07, "loss": 2.1834, "step": 79540 }, { "epoch": 1.79, "grad_norm": 1.5803760147854258, "learning_rate": 5.516561586894664e-07, "loss": 2.3513, "step": 79550 }, { "epoch": 1.79, "grad_norm": 2.0293047581852535, "learning_rate": 5.504606249738264e-07, "loss": 2.3777, "step": 79560 }, { "epoch": 1.79, "grad_norm": 1.6863118021629706, "learning_rate": 5.492663514620988e-07, "loss": 2.3123, "step": 79570 }, { "epoch": 1.79, "grad_norm": 1.7026050921632652, "learning_rate": 5.480733383135517e-07, "loss": 2.4556, "step": 79580 }, { "epoch": 1.79, "grad_norm": 1.6291145039626684, "learning_rate": 5.468815856872877e-07, "loss": 2.4598, "step": 79590 }, { "epoch": 1.8, "grad_norm": 1.56921170665397, "learning_rate": 5.456910937422422e-07, "loss": 2.3622, "step": 79600 }, { "epoch": 1.8, "grad_norm": 1.7423198231177979, "learning_rate": 5.445018626371801e-07, "loss": 2.3586, "step": 79610 }, { "epoch": 1.8, "grad_norm": 1.7824815098676599, "learning_rate": 5.433138925306991e-07, "loss": 2.4024, "step": 79620 }, { "epoch": 1.8, "grad_norm": 1.6596013811838517, "learning_rate": 5.421271835812269e-07, "loss": 2.2963, "step": 79630 }, { "epoch": 1.8, "grad_norm": 1.5641089577437228, "learning_rate": 5.409417359470271e-07, "loss": 2.4545, "step": 79640 }, { "epoch": 1.8, "grad_norm": 1.75575820405234, "learning_rate": 5.397575497861918e-07, "loss": 2.3708, "step": 79650 }, { "epoch": 1.8, "grad_norm": 1.5690998047682434, "learning_rate": 5.385746252566438e-07, "loss": 2.4415, "step": 79660 }, { "epoch": 1.8, "grad_norm": 1.653921621462666, "learning_rate": 5.373929625161423e-07, "loss": 2.3287, "step": 79670 }, { "epoch": 1.8, "grad_norm": 1.7655543460360585, "learning_rate": 5.362125617222746e-07, "loss": 2.3113, "step": 79680 }, { "epoch": 1.8, "grad_norm": 1.6615405116783797, "learning_rate": 5.350334230324605e-07, "loss": 2.4955, "step": 79690 }, { "epoch": 1.8, "grad_norm": 1.615367421531356, "learning_rate": 5.338555466039508e-07, "loss": 2.4058, "step": 79700 }, { "epoch": 1.8, "grad_norm": 1.5568382337147222, "learning_rate": 5.326789325938297e-07, "loss": 2.4749, "step": 79710 }, { "epoch": 1.8, "grad_norm": 1.7252098168550307, "learning_rate": 5.315035811590108e-07, "loss": 2.3595, "step": 79720 }, { "epoch": 1.8, "grad_norm": 1.6282429916983288, "learning_rate": 5.303294924562429e-07, "loss": 2.2665, "step": 79730 }, { "epoch": 1.8, "grad_norm": 1.6197591270655045, "learning_rate": 5.291566666421033e-07, "loss": 2.3541, "step": 79740 }, { "epoch": 1.8, "grad_norm": 1.623294039929234, "learning_rate": 5.279851038730011e-07, "loss": 2.5255, "step": 79750 }, { "epoch": 1.8, "grad_norm": 1.6123608522996598, "learning_rate": 5.268148043051779e-07, "loss": 2.3189, "step": 79760 }, { "epoch": 1.8, "grad_norm": 1.475831870279334, "learning_rate": 5.25645768094708e-07, "loss": 2.3143, "step": 79770 }, { "epoch": 1.8, "grad_norm": 1.5528758157688758, "learning_rate": 5.244779953974922e-07, "loss": 2.3526, "step": 79780 }, { "epoch": 1.8, "grad_norm": 1.7341897349509294, "learning_rate": 5.233114863692679e-07, "loss": 2.3956, "step": 79790 }, { "epoch": 1.8, "grad_norm": 1.8386290739842694, "learning_rate": 5.221462411656042e-07, "loss": 2.4037, "step": 79800 }, { "epoch": 1.8, "grad_norm": 1.689806424424735, "learning_rate": 5.209822599418979e-07, "loss": 2.3804, "step": 79810 }, { "epoch": 1.8, "grad_norm": 1.5152847477629496, "learning_rate": 5.198195428533792e-07, "loss": 2.3821, "step": 79820 }, { "epoch": 1.8, "grad_norm": 1.8282492575699913, "learning_rate": 5.186580900551108e-07, "loss": 2.2376, "step": 79830 }, { "epoch": 1.8, "grad_norm": 1.594658592658697, "learning_rate": 5.174979017019855e-07, "loss": 2.367, "step": 79840 }, { "epoch": 1.8, "grad_norm": 1.7263745597122158, "learning_rate": 5.16338977948726e-07, "loss": 2.3756, "step": 79850 }, { "epoch": 1.8, "grad_norm": 1.7926386909547545, "learning_rate": 5.151813189498888e-07, "loss": 2.3055, "step": 79860 }, { "epoch": 1.8, "grad_norm": 1.81970895400309, "learning_rate": 5.140249248598628e-07, "loss": 2.3174, "step": 79870 }, { "epoch": 1.8, "grad_norm": 1.6711389954177451, "learning_rate": 5.128697958328655e-07, "loss": 2.3164, "step": 79880 }, { "epoch": 1.8, "grad_norm": 1.6915334562470608, "learning_rate": 5.117159320229459e-07, "loss": 2.3174, "step": 79890 }, { "epoch": 1.8, "grad_norm": 1.459836918144899, "learning_rate": 5.105633335839843e-07, "loss": 2.28, "step": 79900 }, { "epoch": 1.8, "grad_norm": 1.5014011026395948, "learning_rate": 5.094120006696934e-07, "loss": 2.4423, "step": 79910 }, { "epoch": 1.8, "grad_norm": 1.4891539044004702, "learning_rate": 5.082619334336147e-07, "loss": 2.4571, "step": 79920 }, { "epoch": 1.8, "grad_norm": 1.7658330739875778, "learning_rate": 5.071131320291267e-07, "loss": 2.3788, "step": 79930 }, { "epoch": 1.8, "grad_norm": 1.791570929644009, "learning_rate": 5.059655966094335e-07, "loss": 2.382, "step": 79940 }, { "epoch": 1.8, "grad_norm": 1.745163380844217, "learning_rate": 5.048193273275714e-07, "loss": 2.4468, "step": 79950 }, { "epoch": 1.8, "grad_norm": 1.8761907830575948, "learning_rate": 5.036743243364095e-07, "loss": 2.2437, "step": 79960 }, { "epoch": 1.8, "grad_norm": 1.5947701667958332, "learning_rate": 5.025305877886466e-07, "loss": 2.3285, "step": 79970 }, { "epoch": 1.8, "grad_norm": 1.766492141592457, "learning_rate": 5.013881178368119e-07, "loss": 2.3634, "step": 79980 }, { "epoch": 1.8, "grad_norm": 1.6275186054881896, "learning_rate": 5.00246914633269e-07, "loss": 2.2132, "step": 79990 }, { "epoch": 1.8, "grad_norm": 1.5868799120036636, "learning_rate": 4.991069783302083e-07, "loss": 2.4402, "step": 80000 }, { "epoch": 1.8, "grad_norm": 1.7074329034919362, "learning_rate": 4.979683090796539e-07, "loss": 2.3844, "step": 80010 }, { "epoch": 1.8, "grad_norm": 1.8079220070539133, "learning_rate": 4.968309070334609e-07, "loss": 2.343, "step": 80020 }, { "epoch": 1.8, "grad_norm": 2.1183887372537478, "learning_rate": 4.956947723433148e-07, "loss": 2.3407, "step": 80030 }, { "epoch": 1.8, "grad_norm": 1.7192608466808743, "learning_rate": 4.945599051607319e-07, "loss": 2.4149, "step": 80040 }, { "epoch": 1.81, "grad_norm": 1.4955783575120956, "learning_rate": 4.934263056370591e-07, "loss": 2.438, "step": 80050 }, { "epoch": 1.81, "grad_norm": 1.8196075711460016, "learning_rate": 4.922939739234745e-07, "loss": 2.4859, "step": 80060 }, { "epoch": 1.81, "grad_norm": 1.6153247533450328, "learning_rate": 4.911629101709891e-07, "loss": 2.3254, "step": 80070 }, { "epoch": 1.81, "grad_norm": 1.56177216674309, "learning_rate": 4.900331145304427e-07, "loss": 2.3111, "step": 80080 }, { "epoch": 1.81, "grad_norm": 1.7287053956485363, "learning_rate": 4.889045871525078e-07, "loss": 2.4435, "step": 80090 }, { "epoch": 1.81, "grad_norm": 1.5482276472363172, "learning_rate": 4.87777328187683e-07, "loss": 2.4452, "step": 80100 }, { "epoch": 1.81, "grad_norm": 1.7009571826206253, "learning_rate": 4.866513377863025e-07, "loss": 2.3497, "step": 80110 }, { "epoch": 1.81, "grad_norm": 1.958281578382322, "learning_rate": 4.855266160985295e-07, "loss": 2.3537, "step": 80120 }, { "epoch": 1.81, "grad_norm": 1.7774603677838923, "learning_rate": 4.844031632743607e-07, "loss": 2.32, "step": 80130 }, { "epoch": 1.81, "grad_norm": 1.509581497625237, "learning_rate": 4.832809794636195e-07, "loss": 2.3761, "step": 80140 }, { "epoch": 1.81, "grad_norm": 1.7104831353160246, "learning_rate": 4.82160064815963e-07, "loss": 2.3678, "step": 80150 }, { "epoch": 1.81, "grad_norm": 1.7106163980493694, "learning_rate": 4.810404194808771e-07, "loss": 2.3786, "step": 80160 }, { "epoch": 1.81, "grad_norm": 1.851549604992328, "learning_rate": 4.79922043607679e-07, "loss": 2.3589, "step": 80170 }, { "epoch": 1.81, "grad_norm": 1.721015006283881, "learning_rate": 4.788049373455184e-07, "loss": 2.2441, "step": 80180 }, { "epoch": 1.81, "grad_norm": 2.135112467113133, "learning_rate": 4.776891008433715e-07, "loss": 2.3503, "step": 80190 }, { "epoch": 1.81, "grad_norm": 1.7276863325694778, "learning_rate": 4.765745342500516e-07, "loss": 2.3371, "step": 80200 }, { "epoch": 1.81, "grad_norm": 1.7305913944529634, "learning_rate": 4.754612377141965e-07, "loss": 2.2952, "step": 80210 }, { "epoch": 1.81, "grad_norm": 1.8386256637673464, "learning_rate": 4.7434921138427625e-07, "loss": 2.3256, "step": 80220 }, { "epoch": 1.81, "grad_norm": 2.013945111154308, "learning_rate": 4.732384554085945e-07, "loss": 2.2408, "step": 80230 }, { "epoch": 1.81, "grad_norm": 1.6870584860010227, "learning_rate": 4.7212896993528044e-07, "loss": 2.3902, "step": 80240 }, { "epoch": 1.81, "grad_norm": 1.6769799552299396, "learning_rate": 4.7102075511229807e-07, "loss": 2.368, "step": 80250 }, { "epoch": 1.81, "grad_norm": 1.6108122915233052, "learning_rate": 4.699138110874413e-07, "loss": 2.3486, "step": 80260 }, { "epoch": 1.81, "grad_norm": 1.7613103472131464, "learning_rate": 4.6880813800833427e-07, "loss": 2.4727, "step": 80270 }, { "epoch": 1.81, "grad_norm": 1.6698972165120767, "learning_rate": 4.6770373602242905e-07, "loss": 2.3099, "step": 80280 }, { "epoch": 1.81, "grad_norm": 1.7649577861755104, "learning_rate": 4.666006052770111e-07, "loss": 2.3959, "step": 80290 }, { "epoch": 1.81, "grad_norm": 2.0278351268031365, "learning_rate": 4.654987459191951e-07, "loss": 2.3868, "step": 80300 }, { "epoch": 1.81, "grad_norm": 1.8395762708414676, "learning_rate": 4.643981580959267e-07, "loss": 2.318, "step": 80310 }, { "epoch": 1.81, "grad_norm": 1.4707036216007112, "learning_rate": 4.63298841953983e-07, "loss": 2.4076, "step": 80320 }, { "epoch": 1.81, "grad_norm": 1.6086788144687136, "learning_rate": 4.6220079763996894e-07, "loss": 2.4444, "step": 80330 }, { "epoch": 1.81, "grad_norm": 1.833916689190466, "learning_rate": 4.6110402530032295e-07, "loss": 2.5053, "step": 80340 }, { "epoch": 1.81, "grad_norm": 1.7594909689407958, "learning_rate": 4.600085250813091e-07, "loss": 2.3006, "step": 80350 }, { "epoch": 1.81, "grad_norm": 1.6188146689090726, "learning_rate": 4.5891429712902836e-07, "loss": 2.3908, "step": 80360 }, { "epoch": 1.81, "grad_norm": 1.8092017518391497, "learning_rate": 4.578213415894062e-07, "loss": 2.4165, "step": 80370 }, { "epoch": 1.81, "grad_norm": 1.5649092980301815, "learning_rate": 4.567296586082015e-07, "loss": 2.4332, "step": 80380 }, { "epoch": 1.81, "grad_norm": 1.7980950324779545, "learning_rate": 4.5563924833100014e-07, "loss": 2.3762, "step": 80390 }, { "epoch": 1.81, "grad_norm": 2.108823444091739, "learning_rate": 4.545501109032258e-07, "loss": 2.3981, "step": 80400 }, { "epoch": 1.81, "grad_norm": 1.6293607250457398, "learning_rate": 4.534622464701244e-07, "loss": 2.3959, "step": 80410 }, { "epoch": 1.81, "grad_norm": 1.5984081029520008, "learning_rate": 4.523756551767766e-07, "loss": 2.464, "step": 80420 }, { "epoch": 1.81, "grad_norm": 1.8932162209595589, "learning_rate": 4.512903371680899e-07, "loss": 2.3114, "step": 80430 }, { "epoch": 1.81, "grad_norm": 1.4206563922800606, "learning_rate": 4.50206292588804e-07, "loss": 2.3975, "step": 80440 }, { "epoch": 1.81, "grad_norm": 1.8989891425801566, "learning_rate": 4.491235215834877e-07, "loss": 2.3019, "step": 80450 }, { "epoch": 1.81, "grad_norm": 1.8188142650256438, "learning_rate": 4.480420242965433e-07, "loss": 2.3015, "step": 80460 }, { "epoch": 1.81, "grad_norm": 1.7913307964728558, "learning_rate": 4.469618008721999e-07, "loss": 2.3311, "step": 80470 }, { "epoch": 1.81, "grad_norm": 1.579285903996608, "learning_rate": 4.458828514545177e-07, "loss": 2.2802, "step": 80480 }, { "epoch": 1.82, "grad_norm": 1.5184614769149454, "learning_rate": 4.448051761873873e-07, "loss": 2.4751, "step": 80490 }, { "epoch": 1.82, "grad_norm": 1.6318455479686638, "learning_rate": 4.4372877521452805e-07, "loss": 2.3212, "step": 80500 }, { "epoch": 1.82, "grad_norm": 1.6195045212760415, "learning_rate": 4.426536486794897e-07, "loss": 2.2608, "step": 80510 }, { "epoch": 1.82, "grad_norm": 2.0389466137609404, "learning_rate": 4.415797967256552e-07, "loss": 2.3292, "step": 80520 }, { "epoch": 1.82, "grad_norm": 1.7404772013869307, "learning_rate": 4.4050721949623234e-07, "loss": 2.3683, "step": 80530 }, { "epoch": 1.82, "grad_norm": 1.731918930933678, "learning_rate": 4.394359171342633e-07, "loss": 2.4592, "step": 80540 }, { "epoch": 1.82, "grad_norm": 1.6620284259530413, "learning_rate": 4.383658897826171e-07, "loss": 2.3151, "step": 80550 }, { "epoch": 1.82, "grad_norm": 1.839564355821082, "learning_rate": 4.3729713758399515e-07, "loss": 2.2436, "step": 80560 }, { "epoch": 1.82, "grad_norm": 1.6560004389798415, "learning_rate": 4.36229660680928e-07, "loss": 2.3009, "step": 80570 }, { "epoch": 1.82, "grad_norm": 1.6377931361544051, "learning_rate": 4.3516345921577385e-07, "loss": 2.376, "step": 80580 }, { "epoch": 1.82, "grad_norm": 1.8207926351847175, "learning_rate": 4.340985333307235e-07, "loss": 2.4142, "step": 80590 }, { "epoch": 1.82, "grad_norm": 1.6906615643725489, "learning_rate": 4.3303488316779885e-07, "loss": 2.2945, "step": 80600 }, { "epoch": 1.82, "grad_norm": 1.9186437598036743, "learning_rate": 4.319725088688498e-07, "loss": 2.3798, "step": 80610 }, { "epoch": 1.82, "grad_norm": 1.5973300793203364, "learning_rate": 4.309114105755541e-07, "loss": 2.4198, "step": 80620 }, { "epoch": 1.82, "grad_norm": 1.799596033775513, "learning_rate": 4.29851588429423e-07, "loss": 2.4569, "step": 80630 }, { "epoch": 1.82, "grad_norm": 1.7420958286262203, "learning_rate": 4.2879304257179344e-07, "loss": 2.3161, "step": 80640 }, { "epoch": 1.82, "grad_norm": 1.5061882946063638, "learning_rate": 4.2773577314383584e-07, "loss": 2.3686, "step": 80650 }, { "epoch": 1.82, "grad_norm": 2.16739609116401, "learning_rate": 4.2667978028655077e-07, "loss": 2.3446, "step": 80660 }, { "epoch": 1.82, "grad_norm": 1.7587565006585304, "learning_rate": 4.2562506414076446e-07, "loss": 2.3896, "step": 80670 }, { "epoch": 1.82, "grad_norm": 1.791508494832204, "learning_rate": 4.2457162484713767e-07, "loss": 2.2843, "step": 80680 }, { "epoch": 1.82, "grad_norm": 1.6517354213484277, "learning_rate": 4.235194625461569e-07, "loss": 2.4067, "step": 80690 }, { "epoch": 1.82, "grad_norm": 1.8773819093726694, "learning_rate": 4.2246857737814115e-07, "loss": 2.1856, "step": 80700 }, { "epoch": 1.82, "grad_norm": 1.7702953646174149, "learning_rate": 4.21418969483236e-07, "loss": 2.2941, "step": 80710 }, { "epoch": 1.82, "grad_norm": 1.5455404475472572, "learning_rate": 4.203706390014195e-07, "loss": 2.3293, "step": 80720 }, { "epoch": 1.82, "grad_norm": 1.7869074508484735, "learning_rate": 4.193235860724998e-07, "loss": 2.3994, "step": 80730 }, { "epoch": 1.82, "grad_norm": 1.6362718930004048, "learning_rate": 4.1827781083611406e-07, "loss": 2.3582, "step": 80740 }, { "epoch": 1.82, "grad_norm": 1.9030537195171873, "learning_rate": 4.172333134317252e-07, "loss": 2.3512, "step": 80750 }, { "epoch": 1.82, "grad_norm": 1.7710448667960377, "learning_rate": 4.1619009399863076e-07, "loss": 2.5168, "step": 80760 }, { "epoch": 1.82, "grad_norm": 1.8721258217194325, "learning_rate": 4.15148152675956e-07, "loss": 2.4333, "step": 80770 }, { "epoch": 1.82, "grad_norm": 1.7116541145878916, "learning_rate": 4.1410748960265313e-07, "loss": 2.2303, "step": 80780 }, { "epoch": 1.82, "grad_norm": 1.7357174999466054, "learning_rate": 4.1306810491751004e-07, "loss": 2.3387, "step": 80790 }, { "epoch": 1.82, "grad_norm": 1.759509578937728, "learning_rate": 4.120299987591403e-07, "loss": 2.3848, "step": 80800 }, { "epoch": 1.82, "grad_norm": 1.6842489128154503, "learning_rate": 4.1099317126598425e-07, "loss": 2.318, "step": 80810 }, { "epoch": 1.82, "grad_norm": 1.4318412203035529, "learning_rate": 4.09957622576318e-07, "loss": 2.3067, "step": 80820 }, { "epoch": 1.82, "grad_norm": 1.7212858607400876, "learning_rate": 4.08923352828241e-07, "loss": 2.4445, "step": 80830 }, { "epoch": 1.82, "grad_norm": 1.6658464751019437, "learning_rate": 4.078903621596864e-07, "loss": 2.307, "step": 80840 }, { "epoch": 1.82, "grad_norm": 1.5267549240761225, "learning_rate": 4.06858650708416e-07, "loss": 2.4291, "step": 80850 }, { "epoch": 1.82, "grad_norm": 1.740795190310852, "learning_rate": 4.058282186120188e-07, "loss": 2.3409, "step": 80860 }, { "epoch": 1.82, "grad_norm": 1.6730463831059734, "learning_rate": 4.0479906600791486e-07, "loss": 2.2459, "step": 80870 }, { "epoch": 1.82, "grad_norm": 1.710012691028477, "learning_rate": 4.037711930333532e-07, "loss": 2.4166, "step": 80880 }, { "epoch": 1.82, "grad_norm": 1.758107399346493, "learning_rate": 4.0274459982541426e-07, "loss": 2.3688, "step": 80890 }, { "epoch": 1.82, "grad_norm": 1.940709851436449, "learning_rate": 4.01719286521004e-07, "loss": 2.4129, "step": 80900 }, { "epoch": 1.82, "grad_norm": 1.7308135049537252, "learning_rate": 4.0069525325686086e-07, "loss": 2.2587, "step": 80910 }, { "epoch": 1.82, "grad_norm": 1.4666852261255956, "learning_rate": 3.9967250016954894e-07, "loss": 2.2782, "step": 80920 }, { "epoch": 1.83, "grad_norm": 1.9320889186427064, "learning_rate": 3.98651027395468e-07, "loss": 2.4013, "step": 80930 }, { "epoch": 1.83, "grad_norm": 1.7427192040637016, "learning_rate": 3.9763083507084023e-07, "loss": 2.2786, "step": 80940 }, { "epoch": 1.83, "grad_norm": 1.5252469544681175, "learning_rate": 3.966119233317212e-07, "loss": 2.4505, "step": 80950 }, { "epoch": 1.83, "grad_norm": 1.5425067817638494, "learning_rate": 3.955942923139955e-07, "loss": 2.3456, "step": 80960 }, { "epoch": 1.83, "grad_norm": 1.5733623085332744, "learning_rate": 3.9457794215337244e-07, "loss": 2.4253, "step": 80970 }, { "epoch": 1.83, "grad_norm": 1.7837224249972718, "learning_rate": 3.9356287298539464e-07, "loss": 2.355, "step": 80980 }, { "epoch": 1.83, "grad_norm": 1.7913652900644745, "learning_rate": 3.9254908494543606e-07, "loss": 2.4056, "step": 80990 }, { "epoch": 1.83, "grad_norm": 1.6146910885451298, "learning_rate": 3.915365781686942e-07, "loss": 2.5487, "step": 81000 }, { "epoch": 1.83, "grad_norm": 1.6290521866774865, "learning_rate": 3.9052535279019975e-07, "loss": 2.3552, "step": 81010 }, { "epoch": 1.83, "grad_norm": 1.5219501745382438, "learning_rate": 3.8951540894481054e-07, "loss": 2.284, "step": 81020 }, { "epoch": 1.83, "grad_norm": 1.5594973540470543, "learning_rate": 3.8850674676721436e-07, "loss": 2.3612, "step": 81030 }, { "epoch": 1.83, "grad_norm": 1.6084667523721423, "learning_rate": 3.87499366391928e-07, "loss": 2.3979, "step": 81040 }, { "epoch": 1.83, "grad_norm": 1.58102709361863, "learning_rate": 3.864932679532973e-07, "loss": 2.4246, "step": 81050 }, { "epoch": 1.83, "grad_norm": 1.7765775956315142, "learning_rate": 3.854884515854962e-07, "loss": 2.2361, "step": 81060 }, { "epoch": 1.83, "grad_norm": 1.9203774676667769, "learning_rate": 3.8448491742252847e-07, "loss": 2.3491, "step": 81070 }, { "epoch": 1.83, "grad_norm": 1.51547811868593, "learning_rate": 3.834826655982271e-07, "loss": 2.298, "step": 81080 }, { "epoch": 1.83, "grad_norm": 1.7946888804415995, "learning_rate": 3.824816962462541e-07, "loss": 2.3664, "step": 81090 }, { "epoch": 1.83, "grad_norm": 1.7174029122711996, "learning_rate": 3.814820095000993e-07, "loss": 2.4057, "step": 81100 }, { "epoch": 1.83, "grad_norm": 1.794847424404667, "learning_rate": 3.8048360549308385e-07, "loss": 2.3539, "step": 81110 }, { "epoch": 1.83, "grad_norm": 1.5486641511594965, "learning_rate": 3.7948648435835346e-07, "loss": 2.5118, "step": 81120 }, { "epoch": 1.83, "grad_norm": 1.6478857735758676, "learning_rate": 3.7849064622888843e-07, "loss": 2.2326, "step": 81130 }, { "epoch": 1.83, "grad_norm": 1.5824038122822255, "learning_rate": 3.7749609123749586e-07, "loss": 2.3748, "step": 81140 }, { "epoch": 1.83, "grad_norm": 1.7796285537102652, "learning_rate": 3.7650281951680744e-07, "loss": 2.3378, "step": 81150 }, { "epoch": 1.83, "grad_norm": 1.8133437422985694, "learning_rate": 3.7551083119929055e-07, "loss": 2.2253, "step": 81160 }, { "epoch": 1.83, "grad_norm": 1.735653633743639, "learning_rate": 3.7452012641723824e-07, "loss": 2.3684, "step": 81170 }, { "epoch": 1.83, "grad_norm": 1.68765963279404, "learning_rate": 3.735307053027681e-07, "loss": 2.4482, "step": 81180 }, { "epoch": 1.83, "grad_norm": 1.644256609195475, "learning_rate": 3.725425679878347e-07, "loss": 2.4343, "step": 81190 }, { "epoch": 1.83, "grad_norm": 1.789963951426997, "learning_rate": 3.715557146042159e-07, "loss": 2.413, "step": 81200 }, { "epoch": 1.83, "grad_norm": 1.6938857271023982, "learning_rate": 3.705701452835209e-07, "loss": 2.393, "step": 81210 }, { "epoch": 1.83, "grad_norm": 1.6997825448994508, "learning_rate": 3.695858601571856e-07, "loss": 2.3959, "step": 81220 }, { "epoch": 1.83, "grad_norm": 1.8307484191358154, "learning_rate": 3.686028593564761e-07, "loss": 2.344, "step": 81230 }, { "epoch": 1.83, "grad_norm": 1.9179031690524033, "learning_rate": 3.676211430124876e-07, "loss": 2.3597, "step": 81240 }, { "epoch": 1.83, "grad_norm": 1.543542609457121, "learning_rate": 3.666407112561399e-07, "loss": 2.4335, "step": 81250 }, { "epoch": 1.83, "grad_norm": 1.59848491916892, "learning_rate": 3.6566156421818824e-07, "loss": 2.4014, "step": 81260 }, { "epoch": 1.83, "grad_norm": 1.748722204985316, "learning_rate": 3.6468370202921267e-07, "loss": 2.3899, "step": 81270 }, { "epoch": 1.83, "grad_norm": 1.641200614846936, "learning_rate": 3.6370712481962224e-07, "loss": 2.3603, "step": 81280 }, { "epoch": 1.83, "grad_norm": 1.7303726291372272, "learning_rate": 3.6273183271965274e-07, "loss": 2.3617, "step": 81290 }, { "epoch": 1.83, "grad_norm": 1.5879029269372458, "learning_rate": 3.6175782585937234e-07, "loss": 2.3875, "step": 81300 }, { "epoch": 1.83, "grad_norm": 1.7388204524625899, "learning_rate": 3.6078510436867386e-07, "loss": 2.2573, "step": 81310 }, { "epoch": 1.83, "grad_norm": 1.6073996843313905, "learning_rate": 3.5981366837728347e-07, "loss": 2.4011, "step": 81320 }, { "epoch": 1.83, "grad_norm": 1.7806071297904127, "learning_rate": 3.5884351801475204e-07, "loss": 2.3518, "step": 81330 }, { "epoch": 1.83, "grad_norm": 1.758162233862239, "learning_rate": 3.5787465341046047e-07, "loss": 2.2963, "step": 81340 }, { "epoch": 1.83, "grad_norm": 1.7406274848468022, "learning_rate": 3.569070746936165e-07, "loss": 2.3056, "step": 81350 }, { "epoch": 1.83, "grad_norm": 1.6371536233095523, "learning_rate": 3.5594078199326034e-07, "loss": 2.479, "step": 81360 }, { "epoch": 1.83, "grad_norm": 2.0009602097246524, "learning_rate": 3.549757754382566e-07, "loss": 2.428, "step": 81370 }, { "epoch": 1.84, "grad_norm": 1.6318446560057325, "learning_rate": 3.540120551573001e-07, "loss": 2.2492, "step": 81380 }, { "epoch": 1.84, "grad_norm": 1.596839309677939, "learning_rate": 3.530496212789136e-07, "loss": 2.3815, "step": 81390 }, { "epoch": 1.84, "grad_norm": 1.805520767240783, "learning_rate": 3.52088473931449e-07, "loss": 2.4012, "step": 81400 }, { "epoch": 1.84, "grad_norm": 1.7539907217785127, "learning_rate": 3.511286132430858e-07, "loss": 2.3597, "step": 81410 }, { "epoch": 1.84, "grad_norm": 1.7027896937753677, "learning_rate": 3.501700393418339e-07, "loss": 2.2825, "step": 81420 }, { "epoch": 1.84, "grad_norm": 1.6355339305068164, "learning_rate": 3.4921275235552775e-07, "loss": 2.2962, "step": 81430 }, { "epoch": 1.84, "grad_norm": 1.8118908777700755, "learning_rate": 3.482567524118341e-07, "loss": 2.3276, "step": 81440 }, { "epoch": 1.84, "grad_norm": 1.7122949150610096, "learning_rate": 3.473020396382454e-07, "loss": 2.2833, "step": 81450 }, { "epoch": 1.84, "grad_norm": 1.5775619125613054, "learning_rate": 3.4634861416208534e-07, "loss": 2.3187, "step": 81460 }, { "epoch": 1.84, "grad_norm": 1.573405432708626, "learning_rate": 3.453964761105033e-07, "loss": 2.3511, "step": 81470 }, { "epoch": 1.84, "grad_norm": 1.7500383677621276, "learning_rate": 3.4444562561047666e-07, "loss": 2.2672, "step": 81480 }, { "epoch": 1.84, "grad_norm": 1.7331491625626863, "learning_rate": 3.4349606278881286e-07, "loss": 2.3531, "step": 81490 }, { "epoch": 1.84, "grad_norm": 1.9804974196913063, "learning_rate": 3.425477877721484e-07, "loss": 2.403, "step": 81500 }, { "epoch": 1.84, "grad_norm": 1.566167965997596, "learning_rate": 3.416008006869431e-07, "loss": 2.3802, "step": 81510 }, { "epoch": 1.84, "grad_norm": 1.654742194959259, "learning_rate": 3.406551016594906e-07, "loss": 2.4275, "step": 81520 }, { "epoch": 1.84, "grad_norm": 1.6147786702099305, "learning_rate": 3.3971069081591203e-07, "loss": 2.3093, "step": 81530 }, { "epoch": 1.84, "grad_norm": 1.6639661237917007, "learning_rate": 3.387675682821534e-07, "loss": 2.3455, "step": 81540 }, { "epoch": 1.84, "grad_norm": 1.8144917051510356, "learning_rate": 3.378257341839908e-07, "loss": 2.2668, "step": 81550 }, { "epoch": 1.84, "grad_norm": 1.6977889485396234, "learning_rate": 3.3688518864703037e-07, "loss": 2.3566, "step": 81560 }, { "epoch": 1.84, "grad_norm": 1.6771580224061207, "learning_rate": 3.3594593179670177e-07, "loss": 2.3444, "step": 81570 }, { "epoch": 1.84, "grad_norm": 1.8537135749460238, "learning_rate": 3.350079637582682e-07, "loss": 2.3781, "step": 81580 }, { "epoch": 1.84, "grad_norm": 1.7151662903521976, "learning_rate": 3.3407128465681616e-07, "loss": 2.2575, "step": 81590 }, { "epoch": 1.84, "grad_norm": 1.6874358938121437, "learning_rate": 3.331358946172669e-07, "loss": 2.4679, "step": 81600 }, { "epoch": 1.84, "grad_norm": 1.8751470015797644, "learning_rate": 3.322017937643596e-07, "loss": 2.3501, "step": 81610 }, { "epoch": 1.84, "grad_norm": 1.7181524770562207, "learning_rate": 3.3126898222267113e-07, "loss": 2.3597, "step": 81620 }, { "epoch": 1.84, "grad_norm": 1.8942076601650342, "learning_rate": 3.3033746011659983e-07, "loss": 2.3903, "step": 81630 }, { "epoch": 1.84, "grad_norm": 1.5932836580308636, "learning_rate": 3.2940722757037746e-07, "loss": 2.3176, "step": 81640 }, { "epoch": 1.84, "grad_norm": 1.6958431002619612, "learning_rate": 3.284782847080592e-07, "loss": 2.3851, "step": 81650 }, { "epoch": 1.84, "grad_norm": 1.795492357309918, "learning_rate": 3.275506316535304e-07, "loss": 2.3901, "step": 81660 }, { "epoch": 1.84, "grad_norm": 1.5326573850770964, "learning_rate": 3.2662426853050655e-07, "loss": 2.3679, "step": 81670 }, { "epoch": 1.84, "grad_norm": 1.7357769630392228, "learning_rate": 3.2569919546252547e-07, "loss": 2.375, "step": 81680 }, { "epoch": 1.84, "grad_norm": 1.5872947506051347, "learning_rate": 3.247754125729585e-07, "loss": 2.4363, "step": 81690 }, { "epoch": 1.84, "grad_norm": 1.6936039722873093, "learning_rate": 3.2385291998500157e-07, "loss": 2.3463, "step": 81700 }, { "epoch": 1.84, "grad_norm": 1.7248402746740852, "learning_rate": 3.2293171782168064e-07, "loss": 2.403, "step": 81710 }, { "epoch": 1.84, "grad_norm": 1.6911101711033958, "learning_rate": 3.2201180620584637e-07, "loss": 2.3654, "step": 81720 }, { "epoch": 1.84, "grad_norm": 1.5796283288978832, "learning_rate": 3.210931852601817e-07, "loss": 2.5314, "step": 81730 }, { "epoch": 1.84, "grad_norm": 1.5902506552867222, "learning_rate": 3.201758551071943e-07, "loss": 2.3278, "step": 81740 }, { "epoch": 1.84, "grad_norm": 1.6875202279096915, "learning_rate": 3.192598158692195e-07, "loss": 2.3992, "step": 81750 }, { "epoch": 1.84, "grad_norm": 1.5538628757946695, "learning_rate": 3.1834506766842413e-07, "loss": 2.4019, "step": 81760 }, { "epoch": 1.84, "grad_norm": 1.7050839238831783, "learning_rate": 3.1743161062679826e-07, "loss": 2.3014, "step": 81770 }, { "epoch": 1.84, "grad_norm": 1.9239654954882899, "learning_rate": 3.1651944486616125e-07, "loss": 2.2718, "step": 81780 }, { "epoch": 1.84, "grad_norm": 1.8211887174266994, "learning_rate": 3.156085705081635e-07, "loss": 2.3501, "step": 81790 }, { "epoch": 1.84, "grad_norm": 1.7320010085174875, "learning_rate": 3.14698987674279e-07, "loss": 2.3846, "step": 81800 }, { "epoch": 1.84, "grad_norm": 1.8208128750716115, "learning_rate": 3.137906964858106e-07, "loss": 2.2852, "step": 81810 }, { "epoch": 1.85, "grad_norm": 1.7008438376337547, "learning_rate": 3.1288369706389153e-07, "loss": 2.3544, "step": 81820 }, { "epoch": 1.85, "grad_norm": 1.67869324217455, "learning_rate": 3.119779895294772e-07, "loss": 2.3346, "step": 81830 }, { "epoch": 1.85, "grad_norm": 1.5975402869562412, "learning_rate": 3.1107357400335434e-07, "loss": 2.4364, "step": 81840 }, { "epoch": 1.85, "grad_norm": 1.764309447065483, "learning_rate": 3.101704506061387e-07, "loss": 2.4152, "step": 81850 }, { "epoch": 1.85, "grad_norm": 1.5545981107917959, "learning_rate": 3.092686194582728e-07, "loss": 2.3898, "step": 81860 }, { "epoch": 1.85, "grad_norm": 1.785315972356485, "learning_rate": 3.0836808068002486e-07, "loss": 2.4453, "step": 81870 }, { "epoch": 1.85, "grad_norm": 1.6128098313142019, "learning_rate": 3.074688343914911e-07, "loss": 2.3826, "step": 81880 }, { "epoch": 1.85, "grad_norm": 1.728617825682801, "learning_rate": 3.065708807125989e-07, "loss": 2.4125, "step": 81890 }, { "epoch": 1.85, "grad_norm": 1.6141902103407804, "learning_rate": 3.05674219763098e-07, "loss": 2.3525, "step": 81900 }, { "epoch": 1.85, "grad_norm": 1.4821016460535255, "learning_rate": 3.0477885166256936e-07, "loss": 2.4325, "step": 81910 }, { "epoch": 1.85, "grad_norm": 1.590403772368682, "learning_rate": 3.038847765304198e-07, "loss": 2.2761, "step": 81920 }, { "epoch": 1.85, "grad_norm": 1.5658635072461249, "learning_rate": 3.0299199448588724e-07, "loss": 2.3367, "step": 81930 }, { "epoch": 1.85, "grad_norm": 1.8108883771524769, "learning_rate": 3.0210050564803194e-07, "loss": 2.3461, "step": 81940 }, { "epoch": 1.85, "grad_norm": 1.7426126064284946, "learning_rate": 3.012103101357433e-07, "loss": 2.2829, "step": 81950 }, { "epoch": 1.85, "grad_norm": 1.763024113329008, "learning_rate": 3.003214080677408e-07, "loss": 2.2696, "step": 81960 }, { "epoch": 1.85, "grad_norm": 1.952399343608842, "learning_rate": 2.994337995625696e-07, "loss": 2.323, "step": 81970 }, { "epoch": 1.85, "grad_norm": 1.8493647662241157, "learning_rate": 2.985474847386005e-07, "loss": 2.4318, "step": 81980 }, { "epoch": 1.85, "grad_norm": 1.5476354989821859, "learning_rate": 2.976624637140357e-07, "loss": 2.3528, "step": 81990 }, { "epoch": 1.85, "grad_norm": 1.433615449210749, "learning_rate": 2.9677873660690303e-07, "loss": 2.2698, "step": 82000 }, { "epoch": 1.85, "grad_norm": 1.7822594600292088, "learning_rate": 2.9589630353505703e-07, "loss": 2.3942, "step": 82010 }, { "epoch": 1.85, "grad_norm": 1.7753431629330125, "learning_rate": 2.950151646161792e-07, "loss": 2.5103, "step": 82020 }, { "epoch": 1.85, "grad_norm": 2.225530229384032, "learning_rate": 2.94135319967781e-07, "loss": 2.4238, "step": 82030 }, { "epoch": 1.85, "grad_norm": 1.771616558480077, "learning_rate": 2.932567697071986e-07, "loss": 2.3437, "step": 82040 }, { "epoch": 1.85, "grad_norm": 1.5761867255822555, "learning_rate": 2.9237951395159727e-07, "loss": 2.412, "step": 82050 }, { "epoch": 1.85, "grad_norm": 1.6131698953307068, "learning_rate": 2.915035528179688e-07, "loss": 2.387, "step": 82060 }, { "epoch": 1.85, "grad_norm": 1.7834180079982862, "learning_rate": 2.9062888642313323e-07, "loss": 2.462, "step": 82070 }, { "epoch": 1.85, "grad_norm": 1.7890399006131812, "learning_rate": 2.897555148837372e-07, "loss": 2.4905, "step": 82080 }, { "epoch": 1.85, "grad_norm": 1.5923601337445905, "learning_rate": 2.888834383162542e-07, "loss": 2.4242, "step": 82090 }, { "epoch": 1.85, "grad_norm": 2.2266526858459397, "learning_rate": 2.8801265683698565e-07, "loss": 2.3201, "step": 82100 }, { "epoch": 1.85, "grad_norm": 1.7108009590366515, "learning_rate": 2.8714317056205975e-07, "loss": 2.3697, "step": 82110 }, { "epoch": 1.85, "grad_norm": 1.7051984355596033, "learning_rate": 2.862749796074327e-07, "loss": 2.37, "step": 82120 }, { "epoch": 1.85, "grad_norm": 2.083184293179534, "learning_rate": 2.8540808408888843e-07, "loss": 2.3, "step": 82130 }, { "epoch": 1.85, "grad_norm": 1.6891281599094994, "learning_rate": 2.84542484122039e-07, "loss": 2.4012, "step": 82140 }, { "epoch": 1.85, "grad_norm": 1.739268396240724, "learning_rate": 2.836781798223187e-07, "loss": 2.3573, "step": 82150 }, { "epoch": 1.85, "grad_norm": 1.977004030762526, "learning_rate": 2.8281517130499316e-07, "loss": 2.3616, "step": 82160 }, { "epoch": 1.85, "grad_norm": 1.7618224586967604, "learning_rate": 2.819534586851547e-07, "loss": 2.3962, "step": 82170 }, { "epoch": 1.85, "grad_norm": 1.5668325236394094, "learning_rate": 2.810930420777225e-07, "loss": 2.3999, "step": 82180 }, { "epoch": 1.85, "grad_norm": 1.8265662686130077, "learning_rate": 2.802339215974437e-07, "loss": 2.3483, "step": 82190 }, { "epoch": 1.85, "grad_norm": 2.011770919370705, "learning_rate": 2.79376097358891e-07, "loss": 2.2993, "step": 82200 }, { "epoch": 1.85, "grad_norm": 1.705189227786008, "learning_rate": 2.785195694764664e-07, "loss": 2.3772, "step": 82210 }, { "epoch": 1.85, "grad_norm": 1.6713224880716453, "learning_rate": 2.7766433806439507e-07, "loss": 2.4332, "step": 82220 }, { "epoch": 1.85, "grad_norm": 1.6729389959409255, "learning_rate": 2.768104032367347e-07, "loss": 2.385, "step": 82230 }, { "epoch": 1.85, "grad_norm": 1.799855611944456, "learning_rate": 2.759577651073653e-07, "loss": 2.3789, "step": 82240 }, { "epoch": 1.85, "grad_norm": 1.6729720770597463, "learning_rate": 2.7510642378999695e-07, "loss": 2.3471, "step": 82250 }, { "epoch": 1.86, "grad_norm": 1.7099827524027964, "learning_rate": 2.7425637939816564e-07, "loss": 2.3698, "step": 82260 }, { "epoch": 1.86, "grad_norm": 1.728253135010296, "learning_rate": 2.734076320452328e-07, "loss": 2.3782, "step": 82270 }, { "epoch": 1.86, "grad_norm": 1.6920453777624125, "learning_rate": 2.725601818443913e-07, "loss": 2.3732, "step": 82280 }, { "epoch": 1.86, "grad_norm": 1.7227816689123616, "learning_rate": 2.7171402890865617e-07, "loss": 2.4949, "step": 82290 }, { "epoch": 1.86, "grad_norm": 1.6947860644901234, "learning_rate": 2.708691733508728e-07, "loss": 2.4396, "step": 82300 }, { "epoch": 1.86, "grad_norm": 1.6698243706995373, "learning_rate": 2.7002561528371106e-07, "loss": 2.2377, "step": 82310 }, { "epoch": 1.86, "grad_norm": 1.7056187777873115, "learning_rate": 2.6918335481966984e-07, "loss": 2.4247, "step": 82320 }, { "epoch": 1.86, "grad_norm": 1.9262635705352955, "learning_rate": 2.6834239207107483e-07, "loss": 2.2928, "step": 82330 }, { "epoch": 1.86, "grad_norm": 1.6209276616258637, "learning_rate": 2.6750272715007743e-07, "loss": 2.1953, "step": 82340 }, { "epoch": 1.86, "grad_norm": 1.8140667954084249, "learning_rate": 2.6666436016865693e-07, "loss": 2.3295, "step": 82350 }, { "epoch": 1.86, "grad_norm": 1.600332853713747, "learning_rate": 2.658272912386195e-07, "loss": 2.4137, "step": 82360 }, { "epoch": 1.86, "grad_norm": 1.6882178289552472, "learning_rate": 2.649915204715947e-07, "loss": 2.4341, "step": 82370 }, { "epoch": 1.86, "grad_norm": 1.7672752712013853, "learning_rate": 2.641570479790445e-07, "loss": 2.4179, "step": 82380 }, { "epoch": 1.86, "grad_norm": 2.0709864658944053, "learning_rate": 2.6332387387225655e-07, "loss": 2.4016, "step": 82390 }, { "epoch": 1.86, "grad_norm": 1.6818607461503186, "learning_rate": 2.6249199826234306e-07, "loss": 2.3202, "step": 82400 }, { "epoch": 1.86, "grad_norm": 1.773014477384001, "learning_rate": 2.616614212602431e-07, "loss": 2.3028, "step": 82410 }, { "epoch": 1.86, "grad_norm": 1.718144176538633, "learning_rate": 2.608321429767247e-07, "loss": 2.4672, "step": 82420 }, { "epoch": 1.86, "grad_norm": 1.8273805675256845, "learning_rate": 2.600041635223804e-07, "loss": 2.4117, "step": 82430 }, { "epoch": 1.86, "grad_norm": 1.636367691269186, "learning_rate": 2.5917748300763325e-07, "loss": 2.4594, "step": 82440 }, { "epoch": 1.86, "grad_norm": 1.5894758926854926, "learning_rate": 2.583521015427271e-07, "loss": 2.3779, "step": 82450 }, { "epoch": 1.86, "grad_norm": 1.790659090181406, "learning_rate": 2.575280192377383e-07, "loss": 2.2068, "step": 82460 }, { "epoch": 1.86, "grad_norm": 1.7071008220999462, "learning_rate": 2.5670523620256793e-07, "loss": 2.5268, "step": 82470 }, { "epoch": 1.86, "grad_norm": 1.9962500815405144, "learning_rate": 2.5588375254694266e-07, "loss": 2.4154, "step": 82480 }, { "epoch": 1.86, "grad_norm": 1.7854623051362732, "learning_rate": 2.550635683804159e-07, "loss": 2.3691, "step": 82490 }, { "epoch": 1.86, "grad_norm": 1.4945807327351572, "learning_rate": 2.542446838123691e-07, "loss": 2.3682, "step": 82500 }, { "epoch": 1.86, "grad_norm": 1.8534258716691094, "learning_rate": 2.5342709895201047e-07, "loss": 2.2621, "step": 82510 }, { "epoch": 1.86, "grad_norm": 1.9321920349764095, "learning_rate": 2.526108139083738e-07, "loss": 2.2665, "step": 82520 }, { "epoch": 1.86, "grad_norm": 1.547107465219893, "learning_rate": 2.5179582879031993e-07, "loss": 2.4513, "step": 82530 }, { "epoch": 1.86, "grad_norm": 1.6067449739969766, "learning_rate": 2.5098214370653737e-07, "loss": 2.4348, "step": 82540 }, { "epoch": 1.86, "grad_norm": 1.7003404386965713, "learning_rate": 2.5016975876553937e-07, "loss": 2.4122, "step": 82550 }, { "epoch": 1.86, "grad_norm": 1.6918995523814933, "learning_rate": 2.4935867407566705e-07, "loss": 2.3461, "step": 82560 }, { "epoch": 1.86, "grad_norm": 1.785817485266401, "learning_rate": 2.485488897450883e-07, "loss": 2.3263, "step": 82570 }, { "epoch": 1.86, "grad_norm": 1.839583271926349, "learning_rate": 2.477404058817956e-07, "loss": 2.3636, "step": 82580 }, { "epoch": 1.86, "grad_norm": 1.8434644209557218, "learning_rate": 2.469332225936105e-07, "loss": 2.4243, "step": 82590 }, { "epoch": 1.86, "grad_norm": 1.6397464661461132, "learning_rate": 2.4612733998818027e-07, "loss": 2.4412, "step": 82600 }, { "epoch": 1.86, "grad_norm": 1.8136709457349935, "learning_rate": 2.453227581729778e-07, "loss": 2.3615, "step": 82610 }, { "epoch": 1.86, "grad_norm": 1.6079502549873477, "learning_rate": 2.44519477255305e-07, "loss": 2.396, "step": 82620 }, { "epoch": 1.86, "grad_norm": 1.7159936921797783, "learning_rate": 2.437174973422862e-07, "loss": 2.3335, "step": 82630 }, { "epoch": 1.86, "grad_norm": 2.010888560060241, "learning_rate": 2.429168185408759e-07, "loss": 2.3436, "step": 82640 }, { "epoch": 1.86, "grad_norm": 1.9115388532770927, "learning_rate": 2.4211744095785305e-07, "loss": 2.3308, "step": 82650 }, { "epoch": 1.86, "grad_norm": 1.9042188322355422, "learning_rate": 2.413193646998246e-07, "loss": 2.2831, "step": 82660 }, { "epoch": 1.86, "grad_norm": 1.851623833199967, "learning_rate": 2.405225898732233e-07, "loss": 2.4821, "step": 82670 }, { "epoch": 1.86, "grad_norm": 1.7143466697113008, "learning_rate": 2.3972711658430737e-07, "loss": 2.262, "step": 82680 }, { "epoch": 1.86, "grad_norm": 1.7212877354464557, "learning_rate": 2.389329449391631e-07, "loss": 2.2827, "step": 82690 }, { "epoch": 1.86, "grad_norm": 1.6578144097827927, "learning_rate": 2.3814007504370128e-07, "loss": 2.3715, "step": 82700 }, { "epoch": 1.87, "grad_norm": 1.534690591045481, "learning_rate": 2.3734850700365963e-07, "loss": 2.2314, "step": 82710 }, { "epoch": 1.87, "grad_norm": 1.658527621302004, "learning_rate": 2.3655824092460477e-07, "loss": 2.3818, "step": 82720 }, { "epoch": 1.87, "grad_norm": 1.747774196070226, "learning_rate": 2.3576927691192575e-07, "loss": 2.3885, "step": 82730 }, { "epoch": 1.87, "grad_norm": 1.6915466297408075, "learning_rate": 2.3498161507084283e-07, "loss": 2.2547, "step": 82740 }, { "epoch": 1.87, "grad_norm": 1.839655775885476, "learning_rate": 2.3419525550639644e-07, "loss": 2.3421, "step": 82750 }, { "epoch": 1.87, "grad_norm": 1.7702018904397854, "learning_rate": 2.3341019832345825e-07, "loss": 2.4326, "step": 82760 }, { "epoch": 1.87, "grad_norm": 1.724254995946681, "learning_rate": 2.326264436267245e-07, "loss": 2.3895, "step": 82770 }, { "epoch": 1.87, "grad_norm": 1.9950656721194517, "learning_rate": 2.31843991520716e-07, "loss": 2.2405, "step": 82780 }, { "epoch": 1.87, "grad_norm": 1.723846400799194, "learning_rate": 2.3106284210978602e-07, "loss": 2.3565, "step": 82790 }, { "epoch": 1.87, "grad_norm": 1.8779022187070555, "learning_rate": 2.3028299549810562e-07, "loss": 2.39, "step": 82800 }, { "epoch": 1.87, "grad_norm": 1.7281396019276238, "learning_rate": 2.2950445178967718e-07, "loss": 2.4014, "step": 82810 }, { "epoch": 1.87, "grad_norm": 1.6989394738285368, "learning_rate": 2.2872721108832984e-07, "loss": 2.3352, "step": 82820 }, { "epoch": 1.87, "grad_norm": 1.7810375610457756, "learning_rate": 2.2795127349771628e-07, "loss": 2.4286, "step": 82830 }, { "epoch": 1.87, "grad_norm": 1.6654372176688044, "learning_rate": 2.2717663912131482e-07, "loss": 2.2851, "step": 82840 }, { "epoch": 1.87, "grad_norm": 1.801694493379081, "learning_rate": 2.2640330806243615e-07, "loss": 2.3417, "step": 82850 }, { "epoch": 1.87, "grad_norm": 1.6462761304195852, "learning_rate": 2.2563128042421e-07, "loss": 2.2623, "step": 82860 }, { "epoch": 1.87, "grad_norm": 2.10887371658468, "learning_rate": 2.2486055630959514e-07, "loss": 2.323, "step": 82870 }, { "epoch": 1.87, "grad_norm": 2.042976978602578, "learning_rate": 2.240911358213771e-07, "loss": 2.4368, "step": 82880 }, { "epoch": 1.87, "grad_norm": 1.6559647141078189, "learning_rate": 2.2332301906216715e-07, "loss": 2.3565, "step": 82890 }, { "epoch": 1.87, "grad_norm": 1.9843055213778338, "learning_rate": 2.225562061344011e-07, "loss": 2.3617, "step": 82900 }, { "epoch": 1.87, "grad_norm": 1.5217254298359488, "learning_rate": 2.2179069714034385e-07, "loss": 2.3159, "step": 82910 }, { "epoch": 1.87, "grad_norm": 1.757784965236303, "learning_rate": 2.2102649218208373e-07, "loss": 2.4269, "step": 82920 }, { "epoch": 1.87, "grad_norm": 1.7506777126661808, "learning_rate": 2.2026359136153586e-07, "loss": 2.4879, "step": 82930 }, { "epoch": 1.87, "grad_norm": 1.6974891443485363, "learning_rate": 2.1950199478044222e-07, "loss": 2.3573, "step": 82940 }, { "epoch": 1.87, "grad_norm": 1.7932547602529478, "learning_rate": 2.187417025403704e-07, "loss": 2.3041, "step": 82950 }, { "epoch": 1.87, "grad_norm": 1.5851398994112063, "learning_rate": 2.1798271474271493e-07, "loss": 2.3536, "step": 82960 }, { "epoch": 1.87, "grad_norm": 1.7358834785838004, "learning_rate": 2.1722503148869367e-07, "loss": 2.3837, "step": 82970 }, { "epoch": 1.87, "grad_norm": 1.7772569208009685, "learning_rate": 2.164686528793536e-07, "loss": 2.3895, "step": 82980 }, { "epoch": 1.87, "grad_norm": 1.753693318600891, "learning_rate": 2.1571357901556622e-07, "loss": 2.3521, "step": 82990 }, { "epoch": 1.87, "grad_norm": 2.0140505977935725, "learning_rate": 2.149598099980299e-07, "loss": 2.4618, "step": 83000 }, { "epoch": 1.87, "grad_norm": 1.6761170371228, "learning_rate": 2.1420734592726755e-07, "loss": 2.4433, "step": 83010 }, { "epoch": 1.87, "grad_norm": 1.7585613611424527, "learning_rate": 2.134561869036289e-07, "loss": 2.2649, "step": 83020 }, { "epoch": 1.87, "grad_norm": 1.6628667100596832, "learning_rate": 2.1270633302728938e-07, "loss": 2.4598, "step": 83030 }, { "epoch": 1.87, "grad_norm": 1.6346833476309737, "learning_rate": 2.1195778439824899e-07, "loss": 2.3765, "step": 83040 }, { "epoch": 1.87, "grad_norm": 1.744856710481178, "learning_rate": 2.11210541116339e-07, "loss": 2.3305, "step": 83050 }, { "epoch": 1.87, "grad_norm": 1.6767873645668971, "learning_rate": 2.1046460328121077e-07, "loss": 2.3642, "step": 83060 }, { "epoch": 1.87, "grad_norm": 1.7908898583431905, "learning_rate": 2.0971997099234364e-07, "loss": 2.4174, "step": 83070 }, { "epoch": 1.87, "grad_norm": 1.748902741250301, "learning_rate": 2.0897664434904262e-07, "loss": 2.4199, "step": 83080 }, { "epoch": 1.87, "grad_norm": 1.7804455801701433, "learning_rate": 2.0823462345043843e-07, "loss": 2.4203, "step": 83090 }, { "epoch": 1.87, "grad_norm": 1.5489700818710743, "learning_rate": 2.0749390839548966e-07, "loss": 2.3274, "step": 83100 }, { "epoch": 1.87, "grad_norm": 1.6677988394147392, "learning_rate": 2.0675449928297842e-07, "loss": 2.394, "step": 83110 }, { "epoch": 1.87, "grad_norm": 1.6763767245476604, "learning_rate": 2.0601639621151248e-07, "loss": 2.4883, "step": 83120 }, { "epoch": 1.87, "grad_norm": 1.5791170739405023, "learning_rate": 2.0527959927952645e-07, "loss": 2.3491, "step": 83130 }, { "epoch": 1.87, "grad_norm": 1.7108223941090646, "learning_rate": 2.0454410858528174e-07, "loss": 2.4037, "step": 83140 }, { "epoch": 1.88, "grad_norm": 1.635953318699248, "learning_rate": 2.038099242268643e-07, "loss": 2.3823, "step": 83150 }, { "epoch": 1.88, "grad_norm": 1.6707741812132906, "learning_rate": 2.0307704630218472e-07, "loss": 2.415, "step": 83160 }, { "epoch": 1.88, "grad_norm": 1.7781920657132346, "learning_rate": 2.0234547490898039e-07, "loss": 2.3706, "step": 83170 }, { "epoch": 1.88, "grad_norm": 1.8090299057543622, "learning_rate": 2.0161521014481655e-07, "loss": 2.4287, "step": 83180 }, { "epoch": 1.88, "grad_norm": 1.7469153503862576, "learning_rate": 2.008862521070809e-07, "loss": 2.5826, "step": 83190 }, { "epoch": 1.88, "grad_norm": 1.7712330546323203, "learning_rate": 2.0015860089299011e-07, "loss": 2.395, "step": 83200 }, { "epoch": 1.88, "grad_norm": 1.9111346064332848, "learning_rate": 1.994322565995821e-07, "loss": 2.4349, "step": 83210 }, { "epoch": 1.88, "grad_norm": 1.9609287944606766, "learning_rate": 1.9870721932372716e-07, "loss": 2.1917, "step": 83220 }, { "epoch": 1.88, "grad_norm": 1.5633663873322148, "learning_rate": 1.979834891621124e-07, "loss": 2.4587, "step": 83230 }, { "epoch": 1.88, "grad_norm": 1.681444057326643, "learning_rate": 1.9726106621125728e-07, "loss": 2.2955, "step": 83240 }, { "epoch": 1.88, "grad_norm": 1.6548423893991158, "learning_rate": 1.9653995056750696e-07, "loss": 2.2498, "step": 83250 }, { "epoch": 1.88, "grad_norm": 2.2461656918074855, "learning_rate": 1.9582014232702894e-07, "loss": 2.472, "step": 83260 }, { "epoch": 1.88, "grad_norm": 2.1113496063622246, "learning_rate": 1.9510164158581756e-07, "loss": 2.3393, "step": 83270 }, { "epoch": 1.88, "grad_norm": 1.5696420791745254, "learning_rate": 1.9438444843969505e-07, "loss": 2.5357, "step": 83280 }, { "epoch": 1.88, "grad_norm": 1.6378173769874074, "learning_rate": 1.9366856298430492e-07, "loss": 2.436, "step": 83290 }, { "epoch": 1.88, "grad_norm": 1.8133465707664442, "learning_rate": 1.9295398531511968e-07, "loss": 2.3793, "step": 83300 }, { "epoch": 1.88, "grad_norm": 1.866338196165498, "learning_rate": 1.9224071552743527e-07, "loss": 2.4197, "step": 83310 }, { "epoch": 1.88, "grad_norm": 1.811977892317761, "learning_rate": 1.915287537163757e-07, "loss": 2.3768, "step": 83320 }, { "epoch": 1.88, "grad_norm": 1.8290991977591649, "learning_rate": 1.908180999768916e-07, "loss": 2.3377, "step": 83330 }, { "epoch": 1.88, "grad_norm": 1.7072452868775392, "learning_rate": 1.9010875440375165e-07, "loss": 2.333, "step": 83340 }, { "epoch": 1.88, "grad_norm": 1.4606992186545353, "learning_rate": 1.8940071709155793e-07, "loss": 2.468, "step": 83350 }, { "epoch": 1.88, "grad_norm": 1.476649305993828, "learning_rate": 1.8869398813473493e-07, "loss": 2.4566, "step": 83360 }, { "epoch": 1.88, "grad_norm": 1.7966410803285868, "learning_rate": 1.8798856762753282e-07, "loss": 2.3779, "step": 83370 }, { "epoch": 1.88, "grad_norm": 1.6259242894807147, "learning_rate": 1.8728445566402852e-07, "loss": 2.3914, "step": 83380 }, { "epoch": 1.88, "grad_norm": 1.5602361270162965, "learning_rate": 1.8658165233812253e-07, "loss": 2.3789, "step": 83390 }, { "epoch": 1.88, "grad_norm": 1.9658627130648036, "learning_rate": 1.8588015774354096e-07, "loss": 2.2881, "step": 83400 }, { "epoch": 1.88, "grad_norm": 1.748903248909708, "learning_rate": 1.8517997197383785e-07, "loss": 2.4461, "step": 83410 }, { "epoch": 1.88, "grad_norm": 1.7145054877531432, "learning_rate": 1.8448109512238964e-07, "loss": 2.2666, "step": 83420 }, { "epoch": 1.88, "grad_norm": 1.7932157752520996, "learning_rate": 1.8378352728240068e-07, "loss": 2.4276, "step": 83430 }, { "epoch": 1.88, "grad_norm": 1.7565550096855218, "learning_rate": 1.8308726854689762e-07, "loss": 2.5058, "step": 83440 }, { "epoch": 1.88, "grad_norm": 1.692282648546628, "learning_rate": 1.8239231900873622e-07, "loss": 2.4242, "step": 83450 }, { "epoch": 1.88, "grad_norm": 1.657785378415348, "learning_rate": 1.8169867876059566e-07, "loss": 2.4027, "step": 83460 }, { "epoch": 1.88, "grad_norm": 1.710708227910333, "learning_rate": 1.8100634789498085e-07, "loss": 2.2911, "step": 83470 }, { "epoch": 1.88, "grad_norm": 1.7495880684601943, "learning_rate": 1.8031532650422125e-07, "loss": 2.4007, "step": 83480 }, { "epoch": 1.88, "grad_norm": 1.8832505256470424, "learning_rate": 1.7962561468047313e-07, "loss": 2.3523, "step": 83490 }, { "epoch": 1.88, "grad_norm": 1.6646715123567666, "learning_rate": 1.789372125157174e-07, "loss": 2.2766, "step": 83500 }, { "epoch": 1.88, "grad_norm": 1.8662208496379065, "learning_rate": 1.7825012010175945e-07, "loss": 2.3316, "step": 83510 }, { "epoch": 1.88, "grad_norm": 1.7507139454085154, "learning_rate": 1.775643375302316e-07, "loss": 2.3584, "step": 83520 }, { "epoch": 1.88, "grad_norm": 1.5779584051346864, "learning_rate": 1.768798648925918e-07, "loss": 2.3091, "step": 83530 }, { "epoch": 1.88, "grad_norm": 1.728481194252765, "learning_rate": 1.7619670228012142e-07, "loss": 2.3744, "step": 83540 }, { "epoch": 1.88, "grad_norm": 1.7618474015639345, "learning_rate": 1.7551484978392764e-07, "loss": 2.4114, "step": 83550 }, { "epoch": 1.88, "grad_norm": 1.8362888543035947, "learning_rate": 1.7483430749494323e-07, "loss": 2.3715, "step": 83560 }, { "epoch": 1.88, "grad_norm": 1.8894101610482321, "learning_rate": 1.7415507550392562e-07, "loss": 2.3915, "step": 83570 }, { "epoch": 1.88, "grad_norm": 1.7067613630492648, "learning_rate": 1.7347715390146015e-07, "loss": 2.4003, "step": 83580 }, { "epoch": 1.89, "grad_norm": 1.615017701141913, "learning_rate": 1.7280054277795444e-07, "loss": 2.3343, "step": 83590 }, { "epoch": 1.89, "grad_norm": 1.7557575131905057, "learning_rate": 1.7212524222364192e-07, "loss": 2.4507, "step": 83600 }, { "epoch": 1.89, "grad_norm": 1.9503242398227745, "learning_rate": 1.714512523285816e-07, "loss": 2.4711, "step": 83610 }, { "epoch": 1.89, "grad_norm": 2.0204217871202044, "learning_rate": 1.7077857318265832e-07, "loss": 2.3663, "step": 83620 }, { "epoch": 1.89, "grad_norm": 1.9256267177284672, "learning_rate": 1.701072048755814e-07, "loss": 2.3554, "step": 83630 }, { "epoch": 1.89, "grad_norm": 1.8003630185885542, "learning_rate": 1.694371474968848e-07, "loss": 2.3011, "step": 83640 }, { "epoch": 1.89, "grad_norm": 1.8551463006274935, "learning_rate": 1.687684011359292e-07, "loss": 2.3367, "step": 83650 }, { "epoch": 1.89, "grad_norm": 1.6756690555001015, "learning_rate": 1.6810096588189884e-07, "loss": 2.2888, "step": 83660 }, { "epoch": 1.89, "grad_norm": 1.4999319668042042, "learning_rate": 1.6743484182380476e-07, "loss": 2.3615, "step": 83670 }, { "epoch": 1.89, "grad_norm": 1.7217126840973684, "learning_rate": 1.6677002905048033e-07, "loss": 2.4176, "step": 83680 }, { "epoch": 1.89, "grad_norm": 1.6979770562092178, "learning_rate": 1.6610652765058798e-07, "loss": 2.389, "step": 83690 }, { "epoch": 1.89, "grad_norm": 1.3618173105950595, "learning_rate": 1.6544433771261249e-07, "loss": 2.4767, "step": 83700 }, { "epoch": 1.89, "grad_norm": 1.8100086013595496, "learning_rate": 1.647834593248643e-07, "loss": 2.406, "step": 83710 }, { "epoch": 1.89, "grad_norm": 1.7952360895880137, "learning_rate": 1.6412389257547845e-07, "loss": 2.2474, "step": 83720 }, { "epoch": 1.89, "grad_norm": 1.6553776093000598, "learning_rate": 1.6346563755241683e-07, "loss": 2.4629, "step": 83730 }, { "epoch": 1.89, "grad_norm": 1.7775909237875778, "learning_rate": 1.6280869434346592e-07, "loss": 2.3235, "step": 83740 }, { "epoch": 1.89, "grad_norm": 1.5619693606665925, "learning_rate": 1.6215306303623445e-07, "loss": 2.4481, "step": 83750 }, { "epoch": 1.89, "grad_norm": 1.9042505571978825, "learning_rate": 1.6149874371816033e-07, "loss": 2.4479, "step": 83760 }, { "epoch": 1.89, "grad_norm": 1.5971283750000882, "learning_rate": 1.6084573647650149e-07, "loss": 2.3404, "step": 83770 }, { "epoch": 1.89, "grad_norm": 1.8565930212462507, "learning_rate": 1.6019404139834716e-07, "loss": 2.3686, "step": 83780 }, { "epoch": 1.89, "grad_norm": 1.8785021696435118, "learning_rate": 1.5954365857060784e-07, "loss": 2.3239, "step": 83790 }, { "epoch": 1.89, "grad_norm": 1.6921435063679513, "learning_rate": 1.5889458808001746e-07, "loss": 2.4264, "step": 83800 }, { "epoch": 1.89, "grad_norm": 2.2646788650743455, "learning_rate": 1.58246830013139e-07, "loss": 2.3616, "step": 83810 }, { "epoch": 1.89, "grad_norm": 1.7267656483628604, "learning_rate": 1.5760038445635785e-07, "loss": 2.3463, "step": 83820 }, { "epoch": 1.89, "grad_norm": 1.665292019058468, "learning_rate": 1.5695525149588387e-07, "loss": 2.244, "step": 83830 }, { "epoch": 1.89, "grad_norm": 1.6983771390161846, "learning_rate": 1.5631143121775383e-07, "loss": 2.3924, "step": 83840 }, { "epoch": 1.89, "grad_norm": 2.169732019657885, "learning_rate": 1.5566892370782904e-07, "loss": 2.3247, "step": 83850 }, { "epoch": 1.89, "grad_norm": 1.9899162126284469, "learning_rate": 1.5502772905179432e-07, "loss": 2.3909, "step": 83860 }, { "epoch": 1.89, "grad_norm": 1.8989162385015523, "learning_rate": 1.5438784733516233e-07, "loss": 2.2969, "step": 83870 }, { "epoch": 1.89, "grad_norm": 1.6085166115648784, "learning_rate": 1.5374927864326484e-07, "loss": 2.3016, "step": 83880 }, { "epoch": 1.89, "grad_norm": 1.6718589971458644, "learning_rate": 1.5311202306126481e-07, "loss": 2.2987, "step": 83890 }, { "epoch": 1.89, "grad_norm": 1.658140010556977, "learning_rate": 1.5247608067414656e-07, "loss": 2.4023, "step": 83900 }, { "epoch": 1.89, "grad_norm": 1.859036009519322, "learning_rate": 1.5184145156672104e-07, "loss": 2.3021, "step": 83910 }, { "epoch": 1.89, "grad_norm": 2.014167740334895, "learning_rate": 1.512081358236228e-07, "loss": 2.3529, "step": 83920 }, { "epoch": 1.89, "grad_norm": 1.6977102302906855, "learning_rate": 1.505761335293121e-07, "loss": 2.3456, "step": 83930 }, { "epoch": 1.89, "grad_norm": 1.6818412665490479, "learning_rate": 1.499454447680737e-07, "loss": 2.3513, "step": 83940 }, { "epoch": 1.89, "grad_norm": 1.6036553075703903, "learning_rate": 1.49316069624017e-07, "loss": 2.3888, "step": 83950 }, { "epoch": 1.89, "grad_norm": 1.595461533237575, "learning_rate": 1.48688008181076e-07, "loss": 2.347, "step": 83960 }, { "epoch": 1.89, "grad_norm": 1.7920102651879453, "learning_rate": 1.4806126052300918e-07, "loss": 2.2699, "step": 83970 }, { "epoch": 1.89, "grad_norm": 1.6445609717119163, "learning_rate": 1.4743582673340306e-07, "loss": 2.4487, "step": 83980 }, { "epoch": 1.89, "grad_norm": 1.5468198679381817, "learning_rate": 1.4681170689566426e-07, "loss": 2.4377, "step": 83990 }, { "epoch": 1.89, "grad_norm": 1.6191370594526326, "learning_rate": 1.4618890109302619e-07, "loss": 2.3959, "step": 84000 }, { "epoch": 1.89, "grad_norm": 1.8918803181415824, "learning_rate": 1.4556740940854797e-07, "loss": 2.3842, "step": 84010 }, { "epoch": 1.89, "grad_norm": 1.901022994970993, "learning_rate": 1.4494723192511218e-07, "loss": 2.3629, "step": 84020 }, { "epoch": 1.89, "grad_norm": 1.5994671946229422, "learning_rate": 1.443283687254271e-07, "loss": 2.324, "step": 84030 }, { "epoch": 1.9, "grad_norm": 1.72563955255716, "learning_rate": 1.437108198920234e-07, "loss": 2.4683, "step": 84040 }, { "epoch": 1.9, "grad_norm": 1.8395681343872692, "learning_rate": 1.4309458550726074e-07, "loss": 2.3233, "step": 84050 }, { "epoch": 1.9, "grad_norm": 1.4768419366671082, "learning_rate": 1.4247966565332005e-07, "loss": 2.3703, "step": 84060 }, { "epoch": 1.9, "grad_norm": 1.8766094575407324, "learning_rate": 1.4186606041220796e-07, "loss": 2.4314, "step": 84070 }, { "epoch": 1.9, "grad_norm": 1.7638629807108064, "learning_rate": 1.4125376986575567e-07, "loss": 2.2959, "step": 84080 }, { "epoch": 1.9, "grad_norm": 1.7073866946667682, "learning_rate": 1.4064279409561897e-07, "loss": 2.4146, "step": 84090 }, { "epoch": 1.9, "grad_norm": 1.54168991999578, "learning_rate": 1.4003313318327826e-07, "loss": 2.3233, "step": 84100 }, { "epoch": 1.9, "grad_norm": 1.7351538139769342, "learning_rate": 1.3942478721003848e-07, "loss": 2.3857, "step": 84110 }, { "epoch": 1.9, "grad_norm": 1.56553236417665, "learning_rate": 1.388177562570303e-07, "loss": 2.5091, "step": 84120 }, { "epoch": 1.9, "grad_norm": 1.5239503760592927, "learning_rate": 1.3821204040520897e-07, "loss": 2.3397, "step": 84130 }, { "epoch": 1.9, "grad_norm": 2.006661587852461, "learning_rate": 1.3760763973535206e-07, "loss": 2.5077, "step": 84140 }, { "epoch": 1.9, "grad_norm": 1.7110088835382111, "learning_rate": 1.3700455432806293e-07, "loss": 2.4105, "step": 84150 }, { "epoch": 1.9, "grad_norm": 1.6661767796598514, "learning_rate": 1.3640278426377164e-07, "loss": 2.2648, "step": 84160 }, { "epoch": 1.9, "grad_norm": 1.764927011296158, "learning_rate": 1.3580232962273065e-07, "loss": 2.3285, "step": 84170 }, { "epoch": 1.9, "grad_norm": 1.6902363692033773, "learning_rate": 1.3520319048501596e-07, "loss": 2.3565, "step": 84180 }, { "epoch": 1.9, "grad_norm": 2.0231895372224145, "learning_rate": 1.346053669305314e-07, "loss": 2.2567, "step": 84190 }, { "epoch": 1.9, "grad_norm": 1.7882508676211415, "learning_rate": 1.3400885903900318e-07, "loss": 2.3999, "step": 84200 }, { "epoch": 1.9, "grad_norm": 1.6372128082942468, "learning_rate": 1.3341366688998102e-07, "loss": 2.4625, "step": 84210 }, { "epoch": 1.9, "grad_norm": 1.7626579471110715, "learning_rate": 1.3281979056284257e-07, "loss": 2.5021, "step": 84220 }, { "epoch": 1.9, "grad_norm": 1.7280471918169036, "learning_rate": 1.3222723013678775e-07, "loss": 2.4423, "step": 84230 }, { "epoch": 1.9, "grad_norm": 1.678462394276784, "learning_rate": 1.316359856908389e-07, "loss": 2.3515, "step": 84240 }, { "epoch": 1.9, "grad_norm": 1.7486827518006784, "learning_rate": 1.3104605730384746e-07, "loss": 2.3484, "step": 84250 }, { "epoch": 1.9, "grad_norm": 1.8689891367698417, "learning_rate": 1.304574450544871e-07, "loss": 2.4176, "step": 84260 }, { "epoch": 1.9, "grad_norm": 1.9244243147551712, "learning_rate": 1.2987014902125615e-07, "loss": 2.2547, "step": 84270 }, { "epoch": 1.9, "grad_norm": 1.855665578609105, "learning_rate": 1.2928416928247534e-07, "loss": 2.3551, "step": 84280 }, { "epoch": 1.9, "grad_norm": 1.7175328522877908, "learning_rate": 1.286995059162932e-07, "loss": 2.3434, "step": 84290 }, { "epoch": 1.9, "grad_norm": 1.5702208024793067, "learning_rate": 1.2811615900068186e-07, "loss": 2.3176, "step": 84300 }, { "epoch": 1.9, "grad_norm": 1.7933563462467088, "learning_rate": 1.2753412861343574e-07, "loss": 2.3982, "step": 84310 }, { "epoch": 1.9, "grad_norm": 1.986119946896553, "learning_rate": 1.2695341483217605e-07, "loss": 2.318, "step": 84320 }, { "epoch": 1.9, "grad_norm": 1.5696483248037205, "learning_rate": 1.2637401773434866e-07, "loss": 2.1862, "step": 84330 }, { "epoch": 1.9, "grad_norm": 1.6264134530242103, "learning_rate": 1.2579593739722063e-07, "loss": 2.3531, "step": 84340 }, { "epoch": 1.9, "grad_norm": 1.5346667565551564, "learning_rate": 1.2521917389788584e-07, "loss": 2.3307, "step": 84350 }, { "epoch": 1.9, "grad_norm": 1.613696274968002, "learning_rate": 1.2464372731326392e-07, "loss": 2.4277, "step": 84360 }, { "epoch": 1.9, "grad_norm": 1.683156377900071, "learning_rate": 1.2406959772009563e-07, "loss": 2.277, "step": 84370 }, { "epoch": 1.9, "grad_norm": 1.6047293834263787, "learning_rate": 1.234967851949498e-07, "loss": 2.3728, "step": 84380 }, { "epoch": 1.9, "grad_norm": 1.5976096762340335, "learning_rate": 1.2292528981421525e-07, "loss": 2.2018, "step": 84390 }, { "epoch": 1.9, "grad_norm": 1.576131737166601, "learning_rate": 1.2235511165410885e-07, "loss": 2.4399, "step": 84400 }, { "epoch": 1.9, "grad_norm": 1.6338055135326577, "learning_rate": 1.2178625079066974e-07, "loss": 2.3756, "step": 84410 }, { "epoch": 1.9, "grad_norm": 1.5652378022935507, "learning_rate": 1.212187072997617e-07, "loss": 2.4052, "step": 84420 }, { "epoch": 1.9, "grad_norm": 1.8374603735502042, "learning_rate": 1.2065248125707308e-07, "loss": 2.4118, "step": 84430 }, { "epoch": 1.9, "grad_norm": 1.66982965390427, "learning_rate": 1.200875727381179e-07, "loss": 2.4781, "step": 84440 }, { "epoch": 1.9, "grad_norm": 1.5962688764164614, "learning_rate": 1.1952398181823143e-07, "loss": 2.3551, "step": 84450 }, { "epoch": 1.9, "grad_norm": 1.7098767240729147, "learning_rate": 1.1896170857257694e-07, "loss": 2.4702, "step": 84460 }, { "epoch": 1.9, "grad_norm": 1.661781867242882, "learning_rate": 1.1840075307613775e-07, "loss": 2.3646, "step": 84470 }, { "epoch": 1.91, "grad_norm": 1.4915443707334688, "learning_rate": 1.1784111540372511e-07, "loss": 2.2774, "step": 84480 }, { "epoch": 1.91, "grad_norm": 1.6644674787307923, "learning_rate": 1.1728279562997268e-07, "loss": 2.3729, "step": 84490 }, { "epoch": 1.91, "grad_norm": 1.6133530006909358, "learning_rate": 1.1672579382933979e-07, "loss": 2.3076, "step": 84500 }, { "epoch": 1.91, "grad_norm": 1.954934900215108, "learning_rate": 1.161701100761059e-07, "loss": 2.3944, "step": 84510 }, { "epoch": 1.91, "grad_norm": 1.9667144674207853, "learning_rate": 1.1561574444438284e-07, "loss": 2.2049, "step": 84520 }, { "epoch": 1.91, "grad_norm": 1.931326284191964, "learning_rate": 1.1506269700809703e-07, "loss": 2.3232, "step": 84530 }, { "epoch": 1.91, "grad_norm": 2.0232054358496887, "learning_rate": 1.1451096784100502e-07, "loss": 2.2881, "step": 84540 }, { "epoch": 1.91, "grad_norm": 1.7212870077410682, "learning_rate": 1.1396055701668685e-07, "loss": 2.4446, "step": 84550 }, { "epoch": 1.91, "grad_norm": 1.736750478397609, "learning_rate": 1.1341146460854601e-07, "loss": 2.4599, "step": 84560 }, { "epoch": 1.91, "grad_norm": 1.7573196491106093, "learning_rate": 1.128636906898084e-07, "loss": 2.422, "step": 84570 }, { "epoch": 1.91, "grad_norm": 1.5855299101167957, "learning_rate": 1.1231723533352889e-07, "loss": 2.4093, "step": 84580 }, { "epoch": 1.91, "grad_norm": 1.6605882214540046, "learning_rate": 1.1177209861258254e-07, "loss": 2.3133, "step": 84590 }, { "epoch": 1.91, "grad_norm": 1.6989032184070685, "learning_rate": 1.1122828059966783e-07, "loss": 2.3391, "step": 84600 }, { "epoch": 1.91, "grad_norm": 1.682611575490649, "learning_rate": 1.1068578136731123e-07, "loss": 2.4008, "step": 84610 }, { "epoch": 1.91, "grad_norm": 1.6532226957830016, "learning_rate": 1.101446009878604e-07, "loss": 2.3187, "step": 84620 }, { "epoch": 1.91, "grad_norm": 1.5794200979380844, "learning_rate": 1.0960473953348761e-07, "loss": 2.2691, "step": 84630 }, { "epoch": 1.91, "grad_norm": 1.5526615860401343, "learning_rate": 1.090661970761897e-07, "loss": 2.3687, "step": 84640 }, { "epoch": 1.91, "grad_norm": 1.603849543950645, "learning_rate": 1.0852897368778815e-07, "loss": 2.3931, "step": 84650 }, { "epoch": 1.91, "grad_norm": 1.652390719865925, "learning_rate": 1.0799306943992671e-07, "loss": 2.3743, "step": 84660 }, { "epoch": 1.91, "grad_norm": 1.6319938733819777, "learning_rate": 1.0745848440407491e-07, "loss": 2.4187, "step": 84670 }, { "epoch": 1.91, "grad_norm": 1.6180167829288912, "learning_rate": 1.0692521865152572e-07, "loss": 2.2873, "step": 84680 }, { "epoch": 1.91, "grad_norm": 1.6463172787821074, "learning_rate": 1.0639327225339668e-07, "loss": 2.3637, "step": 84690 }, { "epoch": 1.91, "grad_norm": 1.7293785939893827, "learning_rate": 1.0586264528062773e-07, "loss": 2.3797, "step": 84700 }, { "epoch": 1.91, "grad_norm": 1.5193655318519315, "learning_rate": 1.0533333780398447e-07, "loss": 2.4085, "step": 84710 }, { "epoch": 1.91, "grad_norm": 1.7031427978791613, "learning_rate": 1.048053498940571e-07, "loss": 2.3877, "step": 84720 }, { "epoch": 1.91, "grad_norm": 1.557689526122035, "learning_rate": 1.0427868162125932e-07, "loss": 2.4375, "step": 84730 }, { "epoch": 1.91, "grad_norm": 1.5881335498785591, "learning_rate": 1.0375333305582602e-07, "loss": 2.4279, "step": 84740 }, { "epoch": 1.91, "grad_norm": 1.59115841617555, "learning_rate": 1.0322930426781896e-07, "loss": 2.3397, "step": 84750 }, { "epoch": 1.91, "grad_norm": 1.7387645219605543, "learning_rate": 1.0270659532712446e-07, "loss": 2.4326, "step": 84760 }, { "epoch": 1.91, "grad_norm": 1.7943681920028174, "learning_rate": 1.0218520630345007e-07, "loss": 2.2438, "step": 84770 }, { "epoch": 1.91, "grad_norm": 1.4854759141550469, "learning_rate": 1.0166513726633131e-07, "loss": 2.2417, "step": 84780 }, { "epoch": 1.91, "grad_norm": 1.4701649391931308, "learning_rate": 1.011463882851238e-07, "loss": 2.4518, "step": 84790 }, { "epoch": 1.91, "grad_norm": 1.7158943510439622, "learning_rate": 1.0062895942900996e-07, "loss": 2.3415, "step": 84800 }, { "epoch": 1.91, "grad_norm": 1.8237334630270778, "learning_rate": 1.001128507669924e-07, "loss": 2.4772, "step": 84810 }, { "epoch": 1.91, "grad_norm": 1.783806179579887, "learning_rate": 9.959806236790159e-08, "loss": 2.3837, "step": 84820 }, { "epoch": 1.91, "grad_norm": 1.8144934086201343, "learning_rate": 9.908459430039153e-08, "loss": 2.4501, "step": 84830 }, { "epoch": 1.91, "grad_norm": 1.7072674837006088, "learning_rate": 9.857244663293741e-08, "loss": 2.3756, "step": 84840 }, { "epoch": 1.91, "grad_norm": 1.7587897903286127, "learning_rate": 9.806161943384018e-08, "loss": 2.419, "step": 84850 }, { "epoch": 1.91, "grad_norm": 1.760978101087646, "learning_rate": 9.755211277122422e-08, "loss": 2.4198, "step": 84860 }, { "epoch": 1.91, "grad_norm": 1.8232521822930827, "learning_rate": 9.704392671303964e-08, "loss": 2.2882, "step": 84870 }, { "epoch": 1.91, "grad_norm": 1.689597011063322, "learning_rate": 9.653706132705665e-08, "loss": 2.478, "step": 84880 }, { "epoch": 1.91, "grad_norm": 1.8993188925004987, "learning_rate": 9.60315166808734e-08, "loss": 2.3763, "step": 84890 }, { "epoch": 1.91, "grad_norm": 1.618445330013946, "learning_rate": 9.552729284190932e-08, "loss": 2.365, "step": 84900 }, { "epoch": 1.91, "grad_norm": 1.8763526230385372, "learning_rate": 9.502438987740837e-08, "loss": 2.3091, "step": 84910 }, { "epoch": 1.92, "grad_norm": 1.8239547845425468, "learning_rate": 9.452280785443802e-08, "loss": 2.5287, "step": 84920 }, { "epoch": 1.92, "grad_norm": 1.6619271083191836, "learning_rate": 9.402254683989032e-08, "loss": 2.3835, "step": 84930 }, { "epoch": 1.92, "grad_norm": 1.8123562406054174, "learning_rate": 9.35236069004819e-08, "loss": 2.3495, "step": 84940 }, { "epoch": 1.92, "grad_norm": 1.6718234391810962, "learning_rate": 9.302598810275065e-08, "loss": 2.3352, "step": 84950 }, { "epoch": 1.92, "grad_norm": 1.6536373290037139, "learning_rate": 9.252969051305905e-08, "loss": 2.3741, "step": 84960 }, { "epoch": 1.92, "grad_norm": 1.753918919833679, "learning_rate": 9.203471419759635e-08, "loss": 2.2427, "step": 84970 }, { "epoch": 1.92, "grad_norm": 1.6480059525697757, "learning_rate": 9.154105922237089e-08, "loss": 2.4195, "step": 84980 }, { "epoch": 1.92, "grad_norm": 1.754439188308646, "learning_rate": 9.104872565321887e-08, "loss": 2.2925, "step": 84990 }, { "epoch": 1.92, "grad_norm": 1.699953245180982, "learning_rate": 9.055771355579889e-08, "loss": 2.2556, "step": 85000 }, { "epoch": 1.92, "grad_norm": 1.6535826360607577, "learning_rate": 9.006802299559081e-08, "loss": 2.2172, "step": 85010 }, { "epoch": 1.92, "grad_norm": 1.5725019939888485, "learning_rate": 8.957965403790237e-08, "loss": 2.4219, "step": 85020 }, { "epoch": 1.92, "grad_norm": 1.7737169257984982, "learning_rate": 8.909260674786368e-08, "loss": 2.2739, "step": 85030 }, { "epoch": 1.92, "grad_norm": 1.8472460324822753, "learning_rate": 8.860688119042504e-08, "loss": 2.2798, "step": 85040 }, { "epoch": 1.92, "grad_norm": 1.6305424418111596, "learning_rate": 8.812247743036572e-08, "loss": 2.3799, "step": 85050 }, { "epoch": 1.92, "grad_norm": 1.5236558883133244, "learning_rate": 8.76393955322874e-08, "loss": 2.4148, "step": 85060 }, { "epoch": 1.92, "grad_norm": 1.7746278084572036, "learning_rate": 8.715763556061297e-08, "loss": 2.3276, "step": 85070 }, { "epoch": 1.92, "grad_norm": 1.7472341913143876, "learning_rate": 8.667719757959103e-08, "loss": 2.4832, "step": 85080 }, { "epoch": 1.92, "grad_norm": 1.5208039406900193, "learning_rate": 8.619808165329369e-08, "loss": 2.3687, "step": 85090 }, { "epoch": 1.92, "grad_norm": 1.7816094966149727, "learning_rate": 8.572028784561537e-08, "loss": 2.3201, "step": 85100 }, { "epoch": 1.92, "grad_norm": 1.6962615508189567, "learning_rate": 8.524381622027622e-08, "loss": 2.315, "step": 85110 }, { "epoch": 1.92, "grad_norm": 1.6508535704586584, "learning_rate": 8.476866684081986e-08, "loss": 2.4344, "step": 85120 }, { "epoch": 1.92, "grad_norm": 1.5324332650474775, "learning_rate": 8.429483977061226e-08, "loss": 2.5012, "step": 85130 }, { "epoch": 1.92, "grad_norm": 1.7437591028643393, "learning_rate": 8.382233507284288e-08, "loss": 2.4903, "step": 85140 }, { "epoch": 1.92, "grad_norm": 1.6310456541061222, "learning_rate": 8.335115281052686e-08, "loss": 2.4528, "step": 85150 }, { "epoch": 1.92, "grad_norm": 1.7223355484159633, "learning_rate": 8.288129304650172e-08, "loss": 2.3375, "step": 85160 }, { "epoch": 1.92, "grad_norm": 1.6462510960412857, "learning_rate": 8.241275584342734e-08, "loss": 2.3621, "step": 85170 }, { "epoch": 1.92, "grad_norm": 1.6772889212655877, "learning_rate": 8.19455412637904e-08, "loss": 2.3712, "step": 85180 }, { "epoch": 1.92, "grad_norm": 1.5490857273625316, "learning_rate": 8.147964936989661e-08, "loss": 2.2688, "step": 85190 }, { "epoch": 1.92, "grad_norm": 1.7336227430500748, "learning_rate": 8.101508022388072e-08, "loss": 2.3597, "step": 85200 }, { "epoch": 1.92, "grad_norm": 1.5312932018649152, "learning_rate": 8.055183388769649e-08, "loss": 2.3848, "step": 85210 }, { "epoch": 1.92, "grad_norm": 1.767968313137457, "learning_rate": 8.008991042312452e-08, "loss": 2.3173, "step": 85220 }, { "epoch": 1.92, "grad_norm": 1.6854783378912706, "learning_rate": 7.962930989176665e-08, "loss": 2.2472, "step": 85230 }, { "epoch": 1.92, "grad_norm": 2.0984272241438524, "learning_rate": 7.917003235504816e-08, "loss": 2.4221, "step": 85240 }, { "epoch": 1.92, "grad_norm": 1.8450090299540567, "learning_rate": 7.87120778742212e-08, "loss": 2.3926, "step": 85250 }, { "epoch": 1.92, "grad_norm": 1.673175930910033, "learning_rate": 7.8255446510358e-08, "loss": 2.4573, "step": 85260 }, { "epoch": 1.92, "grad_norm": 1.774452608877182, "learning_rate": 7.780013832435652e-08, "loss": 2.3315, "step": 85270 }, { "epoch": 1.92, "grad_norm": 1.6803795942496855, "learning_rate": 7.734615337693596e-08, "loss": 2.3851, "step": 85280 }, { "epoch": 1.92, "grad_norm": 1.7658884740674547, "learning_rate": 7.689349172864124e-08, "loss": 2.3632, "step": 85290 }, { "epoch": 1.92, "grad_norm": 3.6026831094853424, "learning_rate": 7.644215343983852e-08, "loss": 2.3818, "step": 85300 }, { "epoch": 1.92, "grad_norm": 1.7997458266771482, "learning_rate": 7.599213857072074e-08, "loss": 2.3447, "step": 85310 }, { "epoch": 1.92, "grad_norm": 2.011506675763839, "learning_rate": 7.554344718130213e-08, "loss": 2.3946, "step": 85320 }, { "epoch": 1.92, "grad_norm": 1.7331148043328122, "learning_rate": 7.509607933142149e-08, "loss": 2.3974, "step": 85330 }, { "epoch": 1.92, "grad_norm": 1.8690350626220056, "learning_rate": 7.465003508073776e-08, "loss": 2.5045, "step": 85340 }, { "epoch": 1.92, "grad_norm": 1.6901332592058758, "learning_rate": 7.42053144887389e-08, "loss": 2.2773, "step": 85350 }, { "epoch": 1.92, "grad_norm": 1.7342521056179416, "learning_rate": 7.376191761473195e-08, "loss": 2.5317, "step": 85360 }, { "epoch": 1.93, "grad_norm": 1.6902597849130956, "learning_rate": 7.331984451784957e-08, "loss": 2.4702, "step": 85370 }, { "epoch": 1.93, "grad_norm": 1.67026374445821, "learning_rate": 7.287909525704684e-08, "loss": 2.4343, "step": 85380 }, { "epoch": 1.93, "grad_norm": 1.8455076598414157, "learning_rate": 7.24396698911034e-08, "loss": 2.2981, "step": 85390 }, { "epoch": 1.93, "grad_norm": 1.9861568093561337, "learning_rate": 7.200156847862017e-08, "loss": 2.4675, "step": 85400 }, { "epoch": 1.93, "grad_norm": 1.6380238967599379, "learning_rate": 7.156479107802373e-08, "loss": 2.3566, "step": 85410 }, { "epoch": 1.93, "grad_norm": 1.6295509321367616, "learning_rate": 7.112933774756414e-08, "loss": 2.303, "step": 85420 }, { "epoch": 1.93, "grad_norm": 1.6993566495548362, "learning_rate": 7.069520854531387e-08, "loss": 2.4388, "step": 85430 }, { "epoch": 1.93, "grad_norm": 1.9534542772942687, "learning_rate": 7.026240352916769e-08, "loss": 2.3349, "step": 85440 }, { "epoch": 1.93, "grad_norm": 1.5687628544991086, "learning_rate": 6.983092275684611e-08, "loss": 2.3936, "step": 85450 }, { "epoch": 1.93, "grad_norm": 1.7264729179796932, "learning_rate": 6.940076628589199e-08, "loss": 2.3506, "step": 85460 }, { "epoch": 1.93, "grad_norm": 1.7717258817249277, "learning_rate": 6.897193417367165e-08, "loss": 2.5745, "step": 85470 }, { "epoch": 1.93, "grad_norm": 1.70070853031238, "learning_rate": 6.854442647737491e-08, "loss": 2.2104, "step": 85480 }, { "epoch": 1.93, "grad_norm": 1.8734064039566514, "learning_rate": 6.811824325401395e-08, "loss": 2.3899, "step": 85490 }, { "epoch": 1.93, "grad_norm": 1.8814077487228322, "learning_rate": 6.769338456042662e-08, "loss": 2.3649, "step": 85500 }, { "epoch": 1.93, "grad_norm": 1.8445935162402685, "learning_rate": 6.726985045327095e-08, "loss": 2.5445, "step": 85510 }, { "epoch": 1.93, "grad_norm": 1.644478754068248, "learning_rate": 6.684764098903063e-08, "loss": 2.329, "step": 85520 }, { "epoch": 1.93, "grad_norm": 1.875723229657148, "learning_rate": 6.642675622401285e-08, "loss": 2.4712, "step": 85530 }, { "epoch": 1.93, "grad_norm": 1.764397642573561, "learning_rate": 6.600719621434715e-08, "loss": 2.311, "step": 85540 }, { "epoch": 1.93, "grad_norm": 1.560817265056549, "learning_rate": 6.558896101598544e-08, "loss": 2.3791, "step": 85550 }, { "epoch": 1.93, "grad_norm": 1.883150214262798, "learning_rate": 6.517205068470645e-08, "loss": 2.3909, "step": 85560 }, { "epoch": 1.93, "grad_norm": 1.6543669804402255, "learning_rate": 6.47564652761068e-08, "loss": 2.3642, "step": 85570 }, { "epoch": 1.93, "grad_norm": 1.600918084307246, "learning_rate": 6.434220484561215e-08, "loss": 2.3239, "step": 85580 }, { "epoch": 1.93, "grad_norm": 1.6178475082284185, "learning_rate": 6.392926944846833e-08, "loss": 2.411, "step": 85590 }, { "epoch": 1.93, "grad_norm": 1.691810377652225, "learning_rate": 6.351765913974461e-08, "loss": 2.362, "step": 85600 }, { "epoch": 1.93, "grad_norm": 1.6006998100490664, "learning_rate": 6.310737397433375e-08, "loss": 2.3625, "step": 85610 }, { "epoch": 1.93, "grad_norm": 1.706314016543792, "learning_rate": 6.269841400695308e-08, "loss": 2.351, "step": 85620 }, { "epoch": 1.93, "grad_norm": 1.6773795932799316, "learning_rate": 6.229077929213901e-08, "loss": 2.3958, "step": 85630 }, { "epoch": 1.93, "grad_norm": 1.662339195746713, "learning_rate": 6.1884469884258e-08, "loss": 2.3068, "step": 85640 }, { "epoch": 1.93, "grad_norm": 1.6326849837146065, "learning_rate": 6.147948583749453e-08, "loss": 2.4391, "step": 85650 }, { "epoch": 1.93, "grad_norm": 1.6349788215669538, "learning_rate": 6.107582720585759e-08, "loss": 2.4521, "step": 85660 }, { "epoch": 1.93, "grad_norm": 1.7466970283995809, "learning_rate": 6.06734940431808e-08, "loss": 2.3158, "step": 85670 }, { "epoch": 1.93, "grad_norm": 1.7041450322385927, "learning_rate": 6.02724864031179e-08, "loss": 2.3021, "step": 85680 }, { "epoch": 1.93, "grad_norm": 1.5982848890117325, "learning_rate": 5.987280433914944e-08, "loss": 2.2602, "step": 85690 }, { "epoch": 1.93, "grad_norm": 1.715880141876358, "learning_rate": 5.9474447904576125e-08, "loss": 2.3367, "step": 85700 }, { "epoch": 1.93, "grad_norm": 1.677599665462105, "learning_rate": 5.9077417152525465e-08, "loss": 2.4382, "step": 85710 }, { "epoch": 1.93, "grad_norm": 1.802073602036883, "learning_rate": 5.868171213594398e-08, "loss": 2.4769, "step": 85720 }, { "epoch": 1.93, "grad_norm": 1.6088618972163347, "learning_rate": 5.828733290760502e-08, "loss": 2.2874, "step": 85730 }, { "epoch": 1.93, "grad_norm": 1.9457310793819527, "learning_rate": 5.789427952010207e-08, "loss": 2.2895, "step": 85740 }, { "epoch": 1.93, "grad_norm": 1.6304863390481354, "learning_rate": 5.750255202585542e-08, "loss": 2.3151, "step": 85750 }, { "epoch": 1.93, "grad_norm": 1.7149933301250686, "learning_rate": 5.7112150477103276e-08, "loss": 2.3577, "step": 85760 }, { "epoch": 1.93, "grad_norm": 1.849821408745715, "learning_rate": 5.672307492591289e-08, "loss": 2.2146, "step": 85770 }, { "epoch": 1.93, "grad_norm": 1.782685228916436, "learning_rate": 5.633532542417053e-08, "loss": 2.4822, "step": 85780 }, { "epoch": 1.93, "grad_norm": 2.1020024740828083, "learning_rate": 5.5948902023588156e-08, "loss": 2.4441, "step": 85790 }, { "epoch": 1.93, "grad_norm": 1.588777379218, "learning_rate": 5.5563804775699006e-08, "loss": 2.3172, "step": 85800 }, { "epoch": 1.94, "grad_norm": 1.6580018884386742, "learning_rate": 5.518003373185976e-08, "loss": 2.4043, "step": 85810 }, { "epoch": 1.94, "grad_norm": 1.7921962365432722, "learning_rate": 5.4797588943251714e-08, "loss": 2.3279, "step": 85820 }, { "epoch": 1.94, "grad_norm": 1.5743342862488252, "learning_rate": 5.441647046087628e-08, "loss": 2.3018, "step": 85830 }, { "epoch": 1.94, "grad_norm": 1.9620342216642788, "learning_rate": 5.4036678335561697e-08, "loss": 2.3057, "step": 85840 }, { "epoch": 1.94, "grad_norm": 1.7532418114176285, "learning_rate": 5.365821261795856e-08, "loss": 2.4322, "step": 85850 }, { "epoch": 1.94, "grad_norm": 2.012996969298492, "learning_rate": 5.328107335853649e-08, "loss": 2.3809, "step": 85860 }, { "epoch": 1.94, "grad_norm": 1.807824792295749, "learning_rate": 5.2905260607594157e-08, "loss": 2.4321, "step": 85870 }, { "epoch": 1.94, "grad_norm": 1.7538587007526245, "learning_rate": 5.253077441524923e-08, "loss": 2.2842, "step": 85880 }, { "epoch": 1.94, "grad_norm": 1.7820801143290015, "learning_rate": 5.215761483144399e-08, "loss": 2.3831, "step": 85890 }, { "epoch": 1.94, "grad_norm": 1.4945129731511297, "learning_rate": 5.178578190594197e-08, "loss": 2.4023, "step": 85900 }, { "epoch": 1.94, "grad_norm": 1.6751102188608724, "learning_rate": 5.14152756883346e-08, "loss": 2.3037, "step": 85910 }, { "epoch": 1.94, "grad_norm": 1.66025964260215, "learning_rate": 5.104609622803125e-08, "loss": 2.3171, "step": 85920 }, { "epoch": 1.94, "grad_norm": 1.5514222770597061, "learning_rate": 5.067824357426476e-08, "loss": 2.304, "step": 85930 }, { "epoch": 1.94, "grad_norm": 1.8642967687050989, "learning_rate": 5.031171777609589e-08, "loss": 2.4606, "step": 85940 }, { "epoch": 1.94, "grad_norm": 1.5696583447447527, "learning_rate": 4.9946518882401095e-08, "loss": 2.3835, "step": 85950 }, { "epoch": 1.94, "grad_norm": 1.6698571142091316, "learning_rate": 4.958264694188697e-08, "loss": 2.3698, "step": 85960 }, { "epoch": 1.94, "grad_norm": 1.5603805482726594, "learning_rate": 4.922010200307803e-08, "loss": 2.2928, "step": 85970 }, { "epoch": 1.94, "grad_norm": 1.7834695051112999, "learning_rate": 4.88588841143256e-08, "loss": 2.3922, "step": 85980 }, { "epoch": 1.94, "grad_norm": 1.7456941890424564, "learning_rate": 4.8498993323801145e-08, "loss": 2.3158, "step": 85990 }, { "epoch": 1.94, "grad_norm": 1.5994341285990616, "learning_rate": 4.814042967949961e-08, "loss": 2.3847, "step": 86000 }, { "epoch": 1.94, "grad_norm": 1.7125680181013359, "learning_rate": 4.7783193229240524e-08, "loss": 2.3989, "step": 86010 }, { "epoch": 1.94, "grad_norm": 1.632095103428874, "learning_rate": 4.7427284020664655e-08, "loss": 2.5103, "step": 86020 }, { "epoch": 1.94, "grad_norm": 1.6327602071903573, "learning_rate": 4.707270210123849e-08, "loss": 2.3345, "step": 86030 }, { "epoch": 1.94, "grad_norm": 1.636631725902034, "learning_rate": 4.671944751824642e-08, "loss": 2.3095, "step": 86040 }, { "epoch": 1.94, "grad_norm": 1.718466068516966, "learning_rate": 4.6367520318801875e-08, "loss": 2.4152, "step": 86050 }, { "epoch": 1.94, "grad_norm": 1.8083949166865951, "learning_rate": 4.60169205498362e-08, "loss": 2.3794, "step": 86060 }, { "epoch": 1.94, "grad_norm": 1.7767638621917294, "learning_rate": 4.5667648258107545e-08, "loss": 2.5156, "step": 86070 }, { "epoch": 1.94, "grad_norm": 1.9266326082252196, "learning_rate": 4.531970349019421e-08, "loss": 2.2854, "step": 86080 }, { "epoch": 1.94, "grad_norm": 1.8393702435574584, "learning_rate": 4.497308629249908e-08, "loss": 2.3196, "step": 86090 }, { "epoch": 1.94, "grad_norm": 1.7232294739621674, "learning_rate": 4.4627796711247394e-08, "loss": 2.3057, "step": 86100 }, { "epoch": 1.94, "grad_norm": 1.5294951248565292, "learning_rate": 4.428383479248788e-08, "loss": 2.3083, "step": 86110 }, { "epoch": 1.94, "grad_norm": 1.6972728437212063, "learning_rate": 4.3941200582091616e-08, "loss": 2.3273, "step": 86120 }, { "epoch": 1.94, "grad_norm": 1.6415575537243947, "learning_rate": 4.359989412575205e-08, "loss": 2.3557, "step": 86130 }, { "epoch": 1.94, "grad_norm": 1.4992274340760212, "learning_rate": 4.325991546898722e-08, "loss": 2.3614, "step": 86140 }, { "epoch": 1.94, "grad_norm": 1.6972028513282853, "learning_rate": 4.292126465713642e-08, "loss": 2.3219, "step": 86150 }, { "epoch": 1.94, "grad_norm": 1.583182497874501, "learning_rate": 4.258394173536351e-08, "loss": 2.4386, "step": 86160 }, { "epoch": 1.94, "grad_norm": 1.8145249276401216, "learning_rate": 4.224794674865362e-08, "loss": 2.1838, "step": 86170 }, { "epoch": 1.94, "grad_norm": 1.680428453004223, "learning_rate": 4.191327974181536e-08, "loss": 2.2928, "step": 86180 }, { "epoch": 1.94, "grad_norm": 1.5541920296058995, "learning_rate": 4.1579940759480796e-08, "loss": 2.3485, "step": 86190 }, { "epoch": 1.94, "grad_norm": 1.8327021147580418, "learning_rate": 4.1247929846104375e-08, "loss": 2.4492, "step": 86200 }, { "epoch": 1.94, "grad_norm": 1.5577629687186807, "learning_rate": 4.0917247045964e-08, "loss": 2.4271, "step": 86210 }, { "epoch": 1.94, "grad_norm": 1.8679009576588896, "learning_rate": 4.058789240315997e-08, "loss": 2.3427, "step": 86220 }, { "epoch": 1.94, "grad_norm": 1.8301475475891757, "learning_rate": 4.02598659616138e-08, "loss": 2.2604, "step": 86230 }, { "epoch": 1.94, "grad_norm": 1.734078162007245, "learning_rate": 3.9933167765073835e-08, "loss": 2.2876, "step": 86240 }, { "epoch": 1.95, "grad_norm": 1.703830052721683, "learning_rate": 3.9607797857108554e-08, "loss": 2.353, "step": 86250 }, { "epoch": 1.95, "grad_norm": 1.6878577435219657, "learning_rate": 3.92837562811077e-08, "loss": 2.416, "step": 86260 }, { "epoch": 1.95, "grad_norm": 1.811844291138538, "learning_rate": 3.896104308028892e-08, "loss": 2.2837, "step": 86270 }, { "epoch": 1.95, "grad_norm": 1.8362737993499019, "learning_rate": 3.8639658297686676e-08, "loss": 2.3211, "step": 86280 }, { "epoch": 1.95, "grad_norm": 1.7165781826102997, "learning_rate": 3.831960197616447e-08, "loss": 2.3727, "step": 86290 }, { "epoch": 1.95, "grad_norm": 1.8160043061988442, "learning_rate": 3.800087415840259e-08, "loss": 2.4306, "step": 86300 }, { "epoch": 1.95, "grad_norm": 1.7350226591945848, "learning_rate": 3.7683474886909266e-08, "loss": 2.4077, "step": 86310 }, { "epoch": 1.95, "grad_norm": 1.784741761394556, "learning_rate": 3.7367404204011746e-08, "loss": 2.4043, "step": 86320 }, { "epoch": 1.95, "grad_norm": 1.5296270353560035, "learning_rate": 3.7052662151861876e-08, "loss": 2.5241, "step": 86330 }, { "epoch": 1.95, "grad_norm": 1.6780964258388198, "learning_rate": 3.6739248772434956e-08, "loss": 2.2313, "step": 86340 }, { "epoch": 1.95, "grad_norm": 1.711973480277794, "learning_rate": 3.642716410752867e-08, "loss": 2.3534, "step": 86350 }, { "epoch": 1.95, "grad_norm": 2.020742863559235, "learning_rate": 3.6116408198760835e-08, "loss": 2.3269, "step": 86360 }, { "epoch": 1.95, "grad_norm": 1.5938371473854305, "learning_rate": 3.5806981087576075e-08, "loss": 2.4428, "step": 86370 }, { "epoch": 1.95, "grad_norm": 1.6157265911982917, "learning_rate": 3.549888281523916e-08, "loss": 2.3282, "step": 86380 }, { "epoch": 1.95, "grad_norm": 1.83469850366318, "learning_rate": 3.519211342283946e-08, "loss": 2.2984, "step": 86390 }, { "epoch": 1.95, "grad_norm": 1.9090944461145647, "learning_rate": 3.488667295128756e-08, "loss": 2.2949, "step": 86400 }, { "epoch": 1.95, "grad_norm": 1.5985738827423697, "learning_rate": 3.458256144131644e-08, "loss": 2.4256, "step": 86410 }, { "epoch": 1.95, "grad_norm": 1.752296497637226, "learning_rate": 3.427977893348478e-08, "loss": 2.4716, "step": 86420 }, { "epoch": 1.95, "grad_norm": 1.7577509418435504, "learning_rate": 3.397832546817137e-08, "loss": 2.3478, "step": 86430 }, { "epoch": 1.95, "grad_norm": 1.6947705761785499, "learning_rate": 3.367820108557851e-08, "loss": 2.3313, "step": 86440 }, { "epoch": 1.95, "grad_norm": 1.6574512486441977, "learning_rate": 3.3379405825729735e-08, "loss": 2.3296, "step": 86450 }, { "epoch": 1.95, "grad_norm": 1.691901212393281, "learning_rate": 3.30819397284754e-08, "loss": 2.3031, "step": 86460 }, { "epoch": 1.95, "grad_norm": 1.73346593373472, "learning_rate": 3.278580283348487e-08, "loss": 2.2455, "step": 86470 }, { "epoch": 1.95, "grad_norm": 1.684266358119523, "learning_rate": 3.2490995180251005e-08, "loss": 2.3954, "step": 86480 }, { "epoch": 1.95, "grad_norm": 1.5918774401725204, "learning_rate": 3.2197516808089025e-08, "loss": 2.2992, "step": 86490 }, { "epoch": 1.95, "grad_norm": 1.6841044548063688, "learning_rate": 3.190536775613873e-08, "loss": 2.3074, "step": 86500 }, { "epoch": 1.95, "grad_norm": 1.8367129718751238, "learning_rate": 3.161454806336117e-08, "loss": 2.2928, "step": 86510 }, { "epoch": 1.95, "grad_norm": 1.527817977226325, "learning_rate": 3.1325057768542e-08, "loss": 2.5209, "step": 86520 }, { "epoch": 1.95, "grad_norm": 1.8937386632464506, "learning_rate": 3.103689691028588e-08, "loss": 2.2699, "step": 86530 }, { "epoch": 1.95, "grad_norm": 1.669968146646098, "learning_rate": 3.0750065527023197e-08, "loss": 2.3553, "step": 86540 }, { "epoch": 1.95, "grad_norm": 1.8937893269290251, "learning_rate": 3.0464563657005564e-08, "loss": 2.3475, "step": 86550 }, { "epoch": 1.95, "grad_norm": 1.649417307614027, "learning_rate": 3.018039133830919e-08, "loss": 2.3612, "step": 86560 }, { "epoch": 1.95, "grad_norm": 1.7130342944855739, "learning_rate": 2.989754860883043e-08, "loss": 2.3225, "step": 86570 }, { "epoch": 1.95, "grad_norm": 1.7142835925433664, "learning_rate": 2.961603550629133e-08, "loss": 2.3356, "step": 86580 }, { "epoch": 1.95, "grad_norm": 1.7348979179906479, "learning_rate": 2.9335852068232972e-08, "loss": 2.4197, "step": 86590 }, { "epoch": 1.95, "grad_norm": 1.710118516024183, "learning_rate": 2.9056998332022135e-08, "loss": 2.2743, "step": 86600 }, { "epoch": 1.95, "grad_norm": 1.9520113704937248, "learning_rate": 2.8779474334845737e-08, "loss": 2.2917, "step": 86610 }, { "epoch": 1.95, "grad_norm": 1.6182857364250949, "learning_rate": 2.850328011371639e-08, "loss": 2.4251, "step": 86620 }, { "epoch": 1.95, "grad_norm": 1.6600796939703855, "learning_rate": 2.822841570546797e-08, "loss": 2.3104, "step": 86630 }, { "epoch": 1.95, "grad_norm": 1.8173260081473575, "learning_rate": 2.7954881146754486e-08, "loss": 2.3861, "step": 86640 }, { "epoch": 1.95, "grad_norm": 1.70293656930199, "learning_rate": 2.7682676474057867e-08, "loss": 2.3541, "step": 86650 }, { "epoch": 1.95, "grad_norm": 1.6784876074135313, "learning_rate": 2.7411801723676856e-08, "loss": 2.4205, "step": 86660 }, { "epoch": 1.95, "grad_norm": 1.8459829910294931, "learning_rate": 2.714225693173811e-08, "loss": 2.4161, "step": 86670 }, { "epoch": 1.95, "grad_norm": 1.8380977975416892, "learning_rate": 2.6874042134187317e-08, "loss": 2.4263, "step": 86680 }, { "epoch": 1.95, "grad_norm": 1.8126570576857615, "learning_rate": 2.6607157366793647e-08, "loss": 2.4053, "step": 86690 }, { "epoch": 1.96, "grad_norm": 1.512247491949622, "learning_rate": 2.6341602665149735e-08, "loss": 2.3802, "step": 86700 }, { "epoch": 1.96, "grad_norm": 1.6794569106334394, "learning_rate": 2.607737806467059e-08, "loss": 2.3416, "step": 86710 }, { "epoch": 1.96, "grad_norm": 1.621429372360857, "learning_rate": 2.5814483600593576e-08, "loss": 2.5117, "step": 86720 }, { "epoch": 1.96, "grad_norm": 1.6306506440445003, "learning_rate": 2.555291930797843e-08, "loss": 2.4087, "step": 86730 }, { "epoch": 1.96, "grad_norm": 1.8713381903241288, "learning_rate": 2.5292685221707247e-08, "loss": 2.432, "step": 86740 }, { "epoch": 1.96, "grad_norm": 1.6125200789029979, "learning_rate": 2.50337813764856e-08, "loss": 2.3838, "step": 86750 }, { "epoch": 1.96, "grad_norm": 2.097972448984028, "learning_rate": 2.4776207806841422e-08, "loss": 2.3101, "step": 86760 }, { "epoch": 1.96, "grad_norm": 1.5684964243105972, "learning_rate": 2.451996454712502e-08, "loss": 2.4148, "step": 86770 }, { "epoch": 1.96, "grad_norm": 1.7154470539810078, "learning_rate": 2.4265051631509052e-08, "loss": 2.3557, "step": 86780 }, { "epoch": 1.96, "grad_norm": 1.864236931304774, "learning_rate": 2.4011469093989657e-08, "loss": 2.3223, "step": 86790 }, { "epoch": 1.96, "grad_norm": 1.6674441544918133, "learning_rate": 2.3759216968385346e-08, "loss": 2.3766, "step": 86800 }, { "epoch": 1.96, "grad_norm": 1.6504590920394322, "learning_rate": 2.350829528833476e-08, "loss": 2.41, "step": 86810 }, { "epoch": 1.96, "grad_norm": 1.6433600325816504, "learning_rate": 2.3258704087303353e-08, "loss": 2.4191, "step": 86820 }, { "epoch": 1.96, "grad_norm": 1.7443315853748331, "learning_rate": 2.301044339857561e-08, "loss": 2.4289, "step": 86830 }, { "epoch": 1.96, "grad_norm": 1.8036338076814296, "learning_rate": 2.2763513255260606e-08, "loss": 2.2291, "step": 86840 }, { "epoch": 1.96, "grad_norm": 1.6874655333051902, "learning_rate": 2.251791369028977e-08, "loss": 2.2288, "step": 86850 }, { "epoch": 1.96, "grad_norm": 1.5673444117937483, "learning_rate": 2.227364473641469e-08, "loss": 2.3329, "step": 86860 }, { "epoch": 1.96, "grad_norm": 1.5856249301775633, "learning_rate": 2.2030706426213744e-08, "loss": 2.1866, "step": 86870 }, { "epoch": 1.96, "grad_norm": 1.6327744831503488, "learning_rate": 2.178909879208435e-08, "loss": 2.3418, "step": 86880 }, { "epoch": 1.96, "grad_norm": 1.5173733392682038, "learning_rate": 2.1548821866247406e-08, "loss": 2.3803, "step": 86890 }, { "epoch": 1.96, "grad_norm": 1.6350939666237738, "learning_rate": 2.130987568074727e-08, "loss": 2.489, "step": 86900 }, { "epoch": 1.96, "grad_norm": 1.7153117613807227, "learning_rate": 2.1072260267450685e-08, "loss": 2.3115, "step": 86910 }, { "epoch": 1.96, "grad_norm": 1.7220833614261615, "learning_rate": 2.0835975658044515e-08, "loss": 2.4651, "step": 86920 }, { "epoch": 1.96, "grad_norm": 1.5734282607081436, "learning_rate": 2.0601021884040227e-08, "loss": 2.4784, "step": 86930 }, { "epoch": 1.96, "grad_norm": 1.6576659430072107, "learning_rate": 2.0367398976773868e-08, "loss": 2.4333, "step": 86940 }, { "epoch": 1.96, "grad_norm": 1.5707189888818422, "learning_rate": 2.0135106967399398e-08, "loss": 2.2329, "step": 86950 }, { "epoch": 1.96, "grad_norm": 1.7316746444237876, "learning_rate": 1.9904145886896486e-08, "loss": 2.3411, "step": 86960 }, { "epoch": 1.96, "grad_norm": 1.6941632369189328, "learning_rate": 1.967451576606605e-08, "loss": 2.3347, "step": 86970 }, { "epoch": 1.96, "grad_norm": 1.6324781447515657, "learning_rate": 1.944621663553359e-08, "loss": 2.3975, "step": 86980 }, { "epoch": 1.96, "grad_norm": 1.8560319510536794, "learning_rate": 1.921924852574253e-08, "loss": 2.3869, "step": 86990 }, { "epoch": 1.96, "grad_norm": 1.575967731287666, "learning_rate": 1.899361146696421e-08, "loss": 2.3771, "step": 87000 }, { "epoch": 1.96, "grad_norm": 1.8141118051943665, "learning_rate": 1.8769305489287903e-08, "loss": 2.4489, "step": 87010 }, { "epoch": 1.96, "grad_norm": 1.76292532415422, "learning_rate": 1.854633062262856e-08, "loss": 2.3779, "step": 87020 }, { "epoch": 1.96, "grad_norm": 1.5735911770985547, "learning_rate": 1.8324686896722398e-08, "loss": 2.4362, "step": 87030 }, { "epoch": 1.96, "grad_norm": 1.8348932856383178, "learning_rate": 1.8104374341127995e-08, "loss": 2.4269, "step": 87040 }, { "epoch": 1.96, "grad_norm": 2.264647042068106, "learning_rate": 1.7885392985227402e-08, "loss": 2.3045, "step": 87050 }, { "epoch": 1.96, "grad_norm": 1.4620866038983833, "learning_rate": 1.7667742858221704e-08, "loss": 2.274, "step": 87060 }, { "epoch": 1.96, "grad_norm": 1.8946778821521928, "learning_rate": 1.7451423989139904e-08, "loss": 2.3213, "step": 87070 }, { "epoch": 1.96, "grad_norm": 1.856775573158624, "learning_rate": 1.7236436406828928e-08, "loss": 2.285, "step": 87080 }, { "epoch": 1.96, "grad_norm": 1.6669307138354803, "learning_rate": 1.702278013995917e-08, "loss": 2.3866, "step": 87090 }, { "epoch": 1.96, "grad_norm": 1.7001357206705066, "learning_rate": 1.6810455217026735e-08, "loss": 2.3396, "step": 87100 }, { "epoch": 1.96, "grad_norm": 1.736692686515119, "learning_rate": 1.659946166634563e-08, "loss": 2.3073, "step": 87110 }, { "epoch": 1.96, "grad_norm": 1.6162932063748188, "learning_rate": 1.638979951605446e-08, "loss": 2.2274, "step": 87120 }, { "epoch": 1.96, "grad_norm": 1.7975368774646765, "learning_rate": 1.6181468794114198e-08, "loss": 2.3931, "step": 87130 }, { "epoch": 1.97, "grad_norm": 1.6243040995767257, "learning_rate": 1.597446952830817e-08, "loss": 2.3618, "step": 87140 }, { "epoch": 1.97, "grad_norm": 1.9469099697491996, "learning_rate": 1.5768801746242068e-08, "loss": 2.2954, "step": 87150 }, { "epoch": 1.97, "grad_norm": 1.8387035100591194, "learning_rate": 1.556446547534507e-08, "loss": 2.4672, "step": 87160 }, { "epoch": 1.97, "grad_norm": 1.6149866762637242, "learning_rate": 1.5361460742865373e-08, "loss": 2.3321, "step": 87170 }, { "epoch": 1.97, "grad_norm": 1.887203983383177, "learning_rate": 1.5159787575877994e-08, "loss": 2.3792, "step": 87180 }, { "epoch": 1.97, "grad_norm": 1.7067421883614624, "learning_rate": 1.495944600127808e-08, "loss": 2.422, "step": 87190 }, { "epoch": 1.97, "grad_norm": 1.88556290497219, "learning_rate": 1.4760436045783144e-08, "loss": 2.4184, "step": 87200 }, { "epoch": 1.97, "grad_norm": 1.5115543842084, "learning_rate": 1.4562757735933075e-08, "loss": 2.2561, "step": 87210 }, { "epoch": 1.97, "grad_norm": 1.7537782296500055, "learning_rate": 1.4366411098091226e-08, "loss": 2.2579, "step": 87220 }, { "epoch": 1.97, "grad_norm": 1.6858595466608197, "learning_rate": 1.4171396158442208e-08, "loss": 2.316, "step": 87230 }, { "epoch": 1.97, "grad_norm": 1.5513017311038542, "learning_rate": 1.3977712942992994e-08, "loss": 2.2958, "step": 87240 }, { "epoch": 1.97, "grad_norm": 1.7054410327478404, "learning_rate": 1.3785361477574034e-08, "loss": 2.455, "step": 87250 }, { "epoch": 1.97, "grad_norm": 2.0720673684372723, "learning_rate": 1.359434178783814e-08, "loss": 2.4946, "step": 87260 }, { "epoch": 1.97, "grad_norm": 1.6950204769497277, "learning_rate": 1.340465389925938e-08, "loss": 2.4063, "step": 87270 }, { "epoch": 1.97, "grad_norm": 1.86426930866758, "learning_rate": 1.3216297837135294e-08, "loss": 2.3528, "step": 87280 }, { "epoch": 1.97, "grad_norm": 2.002844671734016, "learning_rate": 1.3029273626584682e-08, "loss": 2.3959, "step": 87290 }, { "epoch": 1.97, "grad_norm": 1.7003625230732582, "learning_rate": 1.2843581292548702e-08, "loss": 2.4369, "step": 87300 }, { "epoch": 1.97, "grad_norm": 2.0120546809749915, "learning_rate": 1.2659220859793098e-08, "loss": 2.4034, "step": 87310 }, { "epoch": 1.97, "grad_norm": 1.8235135445973234, "learning_rate": 1.247619235290376e-08, "loss": 2.2741, "step": 87320 }, { "epoch": 1.97, "grad_norm": 2.031005637239619, "learning_rate": 1.2294495796288942e-08, "loss": 2.3568, "step": 87330 }, { "epoch": 1.97, "grad_norm": 1.7997168664689898, "learning_rate": 1.211413121418148e-08, "loss": 2.4378, "step": 87340 }, { "epoch": 1.97, "grad_norm": 1.9494144494221246, "learning_rate": 1.1935098630633247e-08, "loss": 2.3709, "step": 87350 }, { "epoch": 1.97, "grad_norm": 1.7060179420777293, "learning_rate": 1.1757398069521808e-08, "loss": 2.3554, "step": 87360 }, { "epoch": 1.97, "grad_norm": 1.6322653890392278, "learning_rate": 1.1581029554544876e-08, "loss": 2.4571, "step": 87370 }, { "epoch": 1.97, "grad_norm": 1.818815887985191, "learning_rate": 1.1405993109223634e-08, "loss": 2.5383, "step": 87380 }, { "epoch": 1.97, "grad_norm": 1.7212582195111028, "learning_rate": 1.1232288756900522e-08, "loss": 2.3461, "step": 87390 }, { "epoch": 1.97, "grad_norm": 1.8596951650565012, "learning_rate": 1.1059916520741454e-08, "loss": 2.4484, "step": 87400 }, { "epoch": 1.97, "grad_norm": 1.6869570748090885, "learning_rate": 1.0888876423733596e-08, "loss": 2.3206, "step": 87410 }, { "epoch": 1.97, "grad_norm": 1.7246396685837049, "learning_rate": 1.0719168488687593e-08, "loss": 2.3137, "step": 87420 }, { "epoch": 1.97, "grad_norm": 1.8278114607924212, "learning_rate": 1.0550792738236448e-08, "loss": 2.374, "step": 87430 }, { "epoch": 1.97, "grad_norm": 1.7507407148477925, "learning_rate": 1.0383749194834424e-08, "loss": 2.2725, "step": 87440 }, { "epoch": 1.97, "grad_norm": 1.6573605873912627, "learning_rate": 1.0218037880759258e-08, "loss": 2.3241, "step": 87450 }, { "epoch": 1.97, "grad_norm": 1.6900399541495883, "learning_rate": 1.0053658818108824e-08, "loss": 2.3247, "step": 87460 }, { "epoch": 1.97, "grad_norm": 1.7037604424584298, "learning_rate": 9.890612028806702e-09, "loss": 2.3836, "step": 87470 }, { "epoch": 1.97, "grad_norm": 1.7520275481194574, "learning_rate": 9.728897534596604e-09, "loss": 2.4718, "step": 87480 }, { "epoch": 1.97, "grad_norm": 1.5941310445751076, "learning_rate": 9.568515357043505e-09, "loss": 2.4342, "step": 87490 }, { "epoch": 1.97, "grad_norm": 1.767696768752388, "learning_rate": 9.409465517539185e-09, "loss": 2.294, "step": 87500 }, { "epoch": 1.97, "grad_norm": 1.71398667833447, "learning_rate": 9.25174803729223e-09, "loss": 2.4616, "step": 87510 }, { "epoch": 1.97, "grad_norm": 1.6360507626581795, "learning_rate": 9.095362937336927e-09, "loss": 2.4444, "step": 87520 }, { "epoch": 1.97, "grad_norm": 1.7442573134508554, "learning_rate": 8.940310238528816e-09, "loss": 2.3815, "step": 87530 }, { "epoch": 1.97, "grad_norm": 1.7152101106647024, "learning_rate": 8.786589961546909e-09, "loss": 2.3514, "step": 87540 }, { "epoch": 1.97, "grad_norm": 1.943504978045317, "learning_rate": 8.634202126889257e-09, "loss": 2.3093, "step": 87550 }, { "epoch": 1.97, "grad_norm": 1.5660162013522123, "learning_rate": 8.483146754880712e-09, "loss": 2.2821, "step": 87560 }, { "epoch": 1.97, "grad_norm": 1.5878887264098664, "learning_rate": 8.333423865666268e-09, "loss": 2.3864, "step": 87570 }, { "epoch": 1.98, "grad_norm": 1.8488361190089633, "learning_rate": 8.18503347921218e-09, "loss": 2.4676, "step": 87580 }, { "epoch": 1.98, "grad_norm": 1.8311141566807672, "learning_rate": 8.03797561530817e-09, "loss": 2.2361, "step": 87590 }, { "epoch": 1.98, "grad_norm": 1.8318170565149574, "learning_rate": 7.892250293565217e-09, "loss": 2.4299, "step": 87600 }, { "epoch": 1.98, "grad_norm": 1.660802068853625, "learning_rate": 7.74785753342e-09, "loss": 2.4377, "step": 87610 }, { "epoch": 1.98, "grad_norm": 1.7955707479829786, "learning_rate": 7.604797354126003e-09, "loss": 2.4395, "step": 87620 }, { "epoch": 1.98, "grad_norm": 1.864066727140129, "learning_rate": 7.463069774763521e-09, "loss": 2.3637, "step": 87630 }, { "epoch": 1.98, "grad_norm": 1.7245676497404228, "learning_rate": 7.3226748142340985e-09, "loss": 2.333, "step": 87640 }, { "epoch": 1.98, "grad_norm": 1.6482080065364335, "learning_rate": 7.183612491260539e-09, "loss": 2.3259, "step": 87650 }, { "epoch": 1.98, "grad_norm": 1.8181660297824882, "learning_rate": 7.045882824386896e-09, "loss": 2.2939, "step": 87660 }, { "epoch": 1.98, "grad_norm": 1.5046421107075034, "learning_rate": 6.90948583198292e-09, "loss": 2.5014, "step": 87670 }, { "epoch": 1.98, "grad_norm": 1.6908231654105046, "learning_rate": 6.774421532237396e-09, "loss": 2.3453, "step": 87680 }, { "epoch": 1.98, "grad_norm": 1.526423507792946, "learning_rate": 6.6406899431636915e-09, "loss": 2.2792, "step": 87690 }, { "epoch": 1.98, "grad_norm": 1.739152123676966, "learning_rate": 6.50829108259532e-09, "loss": 2.3382, "step": 87700 }, { "epoch": 1.98, "grad_norm": 1.8374241606987425, "learning_rate": 6.377224968190376e-09, "loss": 2.2993, "step": 87710 }, { "epoch": 1.98, "grad_norm": 1.8353145500593928, "learning_rate": 6.247491617428214e-09, "loss": 2.5378, "step": 87720 }, { "epoch": 1.98, "grad_norm": 1.6419591395891164, "learning_rate": 6.119091047608327e-09, "loss": 2.1153, "step": 87730 }, { "epoch": 1.98, "grad_norm": 1.6678607319603012, "learning_rate": 5.992023275855907e-09, "loss": 2.2397, "step": 87740 }, { "epoch": 1.98, "grad_norm": 1.886956916578118, "learning_rate": 5.866288319117397e-09, "loss": 2.4074, "step": 87750 }, { "epoch": 1.98, "grad_norm": 1.674939049121307, "learning_rate": 5.741886194159385e-09, "loss": 2.4212, "step": 87760 }, { "epoch": 1.98, "grad_norm": 1.7613141493962061, "learning_rate": 5.618816917573044e-09, "loss": 2.3372, "step": 87770 }, { "epoch": 1.98, "grad_norm": 1.807219029237913, "learning_rate": 5.497080505771912e-09, "loss": 2.414, "step": 87780 }, { "epoch": 1.98, "grad_norm": 1.518275772185729, "learning_rate": 5.376676974989669e-09, "loss": 2.3973, "step": 87790 }, { "epoch": 1.98, "grad_norm": 1.7802516124184662, "learning_rate": 5.257606341283472e-09, "loss": 2.4378, "step": 87800 }, { "epoch": 1.98, "grad_norm": 1.9156587557038514, "learning_rate": 5.13986862053395e-09, "loss": 2.3227, "step": 87810 }, { "epoch": 1.98, "grad_norm": 1.6515199651047265, "learning_rate": 5.023463828441877e-09, "loss": 2.4613, "step": 87820 }, { "epoch": 1.98, "grad_norm": 1.7451292685080857, "learning_rate": 4.908391980531502e-09, "loss": 2.2692, "step": 87830 }, { "epoch": 1.98, "grad_norm": 1.5759643230850173, "learning_rate": 4.794653092148327e-09, "loss": 2.3612, "step": 87840 }, { "epoch": 1.98, "grad_norm": 1.7443649126361527, "learning_rate": 4.68224717846244e-09, "loss": 2.3929, "step": 87850 }, { "epoch": 1.98, "grad_norm": 1.585137212640819, "learning_rate": 4.571174254462962e-09, "loss": 2.4332, "step": 87860 }, { "epoch": 1.98, "grad_norm": 1.5637870075022557, "learning_rate": 4.461434334962489e-09, "loss": 2.3761, "step": 87870 }, { "epoch": 1.98, "grad_norm": 1.7522394606965672, "learning_rate": 4.35302743459598e-09, "loss": 2.3379, "step": 87880 }, { "epoch": 1.98, "grad_norm": 1.933155566428064, "learning_rate": 4.24595356782187e-09, "loss": 2.4476, "step": 87890 }, { "epoch": 1.98, "grad_norm": 1.8318294817928586, "learning_rate": 4.140212748919847e-09, "loss": 2.3394, "step": 87900 }, { "epoch": 1.98, "grad_norm": 1.5919594730753295, "learning_rate": 4.035804991990855e-09, "loss": 2.5089, "step": 87910 }, { "epoch": 1.98, "grad_norm": 1.5854105988436258, "learning_rate": 3.9327303109582e-09, "loss": 2.3443, "step": 87920 }, { "epoch": 1.98, "grad_norm": 1.8056063929551964, "learning_rate": 3.8309887195697725e-09, "loss": 2.3808, "step": 87930 }, { "epoch": 1.98, "grad_norm": 1.783053380265477, "learning_rate": 3.7305802313925e-09, "loss": 2.3207, "step": 87940 }, { "epoch": 1.98, "grad_norm": 1.6057571090772427, "learning_rate": 3.6315048598178915e-09, "loss": 2.4611, "step": 87950 }, { "epoch": 1.98, "grad_norm": 1.6293683984738099, "learning_rate": 3.5337626180587114e-09, "loss": 2.3023, "step": 87960 }, { "epoch": 1.98, "grad_norm": 1.9361633736066384, "learning_rate": 3.4373535191500884e-09, "loss": 2.4, "step": 87970 }, { "epoch": 1.98, "grad_norm": 1.680506223551161, "learning_rate": 3.342277575947295e-09, "loss": 2.4122, "step": 87980 }, { "epoch": 1.98, "grad_norm": 1.7257421448452848, "learning_rate": 3.2485348011335183e-09, "loss": 2.421, "step": 87990 }, { "epoch": 1.98, "grad_norm": 1.6418826828691166, "learning_rate": 3.1561252072065394e-09, "loss": 2.2975, "step": 88000 }, { "epoch": 1.98, "grad_norm": 1.5679323081675733, "learning_rate": 3.065048806493165e-09, "loss": 2.4568, "step": 88010 }, { "epoch": 1.98, "grad_norm": 1.7668087817126341, "learning_rate": 2.975305611138124e-09, "loss": 2.4223, "step": 88020 }, { "epoch": 1.99, "grad_norm": 1.858556196414928, "learning_rate": 2.886895633109621e-09, "loss": 2.3206, "step": 88030 }, { "epoch": 1.99, "grad_norm": 1.6905989977250409, "learning_rate": 2.7998188841982244e-09, "loss": 2.3137, "step": 88040 }, { "epoch": 1.99, "grad_norm": 1.8950695835830071, "learning_rate": 2.7140753760179772e-09, "loss": 2.2963, "step": 88050 }, { "epoch": 1.99, "grad_norm": 1.761059224718693, "learning_rate": 2.629665120000846e-09, "loss": 2.297, "step": 88060 }, { "epoch": 1.99, "grad_norm": 1.5436766149044374, "learning_rate": 2.546588127406713e-09, "loss": 2.4797, "step": 88070 }, { "epoch": 1.99, "grad_norm": 1.74190973823307, "learning_rate": 2.464844409314493e-09, "loss": 2.3472, "step": 88080 }, { "epoch": 1.99, "grad_norm": 1.750154439883096, "learning_rate": 2.3844339766243563e-09, "loss": 2.4272, "step": 88090 }, { "epoch": 1.99, "grad_norm": 2.0113797198517767, "learning_rate": 2.305356840059947e-09, "loss": 2.4326, "step": 88100 }, { "epoch": 1.99, "grad_norm": 1.6690137566061873, "learning_rate": 2.2276130101683837e-09, "loss": 2.2172, "step": 88110 }, { "epoch": 1.99, "grad_norm": 1.63770165774168, "learning_rate": 2.1512024973169286e-09, "loss": 2.3837, "step": 88120 }, { "epoch": 1.99, "grad_norm": 1.7783720261603362, "learning_rate": 2.0761253116952094e-09, "loss": 2.3019, "step": 88130 }, { "epoch": 1.99, "grad_norm": 1.786641222541442, "learning_rate": 2.0023814633163273e-09, "loss": 2.2737, "step": 88140 }, { "epoch": 1.99, "grad_norm": 1.617667814905796, "learning_rate": 1.929970962015748e-09, "loss": 2.2953, "step": 88150 }, { "epoch": 1.99, "grad_norm": 1.8699725451073173, "learning_rate": 1.8588938174490811e-09, "loss": 2.3124, "step": 88160 }, { "epoch": 1.99, "grad_norm": 1.7937850281010832, "learning_rate": 1.7891500390943007e-09, "loss": 2.4301, "step": 88170 }, { "epoch": 1.99, "grad_norm": 1.6783060148031639, "learning_rate": 1.7207396362550755e-09, "loss": 2.3901, "step": 88180 }, { "epoch": 1.99, "grad_norm": 1.7264584867997597, "learning_rate": 1.653662618052998e-09, "loss": 2.3626, "step": 88190 }, { "epoch": 1.99, "grad_norm": 1.489890286682901, "learning_rate": 1.5879189934342453e-09, "loss": 2.4679, "step": 88200 }, { "epoch": 1.99, "grad_norm": 1.5992609488126017, "learning_rate": 1.523508771166249e-09, "loss": 2.3718, "step": 88210 }, { "epoch": 1.99, "grad_norm": 1.8519303659765232, "learning_rate": 1.4604319598376937e-09, "loss": 2.3923, "step": 88220 }, { "epoch": 1.99, "grad_norm": 1.7389329048697961, "learning_rate": 1.3986885678629602e-09, "loss": 2.3182, "step": 88230 }, { "epoch": 1.99, "grad_norm": 1.8035895022664787, "learning_rate": 1.3382786034743522e-09, "loss": 2.2572, "step": 88240 }, { "epoch": 1.99, "grad_norm": 1.9243915806294412, "learning_rate": 1.279202074728758e-09, "loss": 2.3966, "step": 88250 }, { "epoch": 1.99, "grad_norm": 1.702157160679504, "learning_rate": 1.2214589895043206e-09, "loss": 2.426, "step": 88260 }, { "epoch": 1.99, "grad_norm": 1.9188237612055483, "learning_rate": 1.1650493555026565e-09, "loss": 2.2547, "step": 88270 }, { "epoch": 1.99, "grad_norm": 1.636233962606256, "learning_rate": 1.1099731802466373e-09, "loss": 2.3961, "step": 88280 }, { "epoch": 1.99, "grad_norm": 1.619751007945698, "learning_rate": 1.0562304710803883e-09, "loss": 2.3709, "step": 88290 }, { "epoch": 1.99, "grad_norm": 1.585557607513026, "learning_rate": 1.0038212351715093e-09, "loss": 2.3843, "step": 88300 }, { "epoch": 1.99, "grad_norm": 1.7236218348091992, "learning_rate": 9.527454795088542e-10, "loss": 2.3252, "step": 88310 }, { "epoch": 1.99, "grad_norm": 1.6092007899353926, "learning_rate": 9.030032109058617e-10, "loss": 2.2531, "step": 88320 }, { "epoch": 1.99, "grad_norm": 1.9254063585995356, "learning_rate": 8.545944359938941e-10, "loss": 2.434, "step": 88330 }, { "epoch": 1.99, "grad_norm": 1.7247196096527089, "learning_rate": 8.075191612300082e-10, "loss": 2.368, "step": 88340 }, { "epoch": 1.99, "grad_norm": 1.7604587125167317, "learning_rate": 7.617773928914052e-10, "loss": 2.4191, "step": 88350 }, { "epoch": 1.99, "grad_norm": 1.9525907111160266, "learning_rate": 7.173691370798707e-10, "loss": 2.3475, "step": 88360 }, { "epoch": 1.99, "grad_norm": 1.7562176034752912, "learning_rate": 6.742943997162243e-10, "loss": 2.2705, "step": 88370 }, { "epoch": 1.99, "grad_norm": 1.736302591639478, "learning_rate": 6.3255318654587e-10, "loss": 2.3333, "step": 88380 }, { "epoch": 1.99, "grad_norm": 1.707014815946809, "learning_rate": 5.921455031354661e-10, "loss": 2.3137, "step": 88390 }, { "epoch": 1.99, "grad_norm": 1.7162108598051502, "learning_rate": 5.53071354872925e-10, "loss": 2.429, "step": 88400 }, { "epoch": 1.99, "grad_norm": 1.5813829057569808, "learning_rate": 5.153307469696334e-10, "loss": 2.4353, "step": 88410 }, { "epoch": 1.99, "grad_norm": 2.0332034107828596, "learning_rate": 4.789236844593426e-10, "loss": 2.3644, "step": 88420 }, { "epoch": 1.99, "grad_norm": 1.7118267895245194, "learning_rate": 4.4385017219705783e-10, "loss": 2.3405, "step": 88430 }, { "epoch": 1.99, "grad_norm": 1.6680163697707988, "learning_rate": 4.1011021485903857e-10, "loss": 2.3264, "step": 88440 }, { "epoch": 1.99, "grad_norm": 1.911308769902226, "learning_rate": 3.7770381694723914e-10, "loss": 2.3313, "step": 88450 }, { "epoch": 1.99, "grad_norm": 1.7526914455159688, "learning_rate": 3.466309827815373e-10, "loss": 2.2945, "step": 88460 }, { "epoch": 2.0, "grad_norm": 1.88668365546548, "learning_rate": 3.168917165063956e-10, "loss": 2.421, "step": 88470 }, { "epoch": 2.0, "grad_norm": 1.6529941562080663, "learning_rate": 2.8848602208753075e-10, "loss": 2.3783, "step": 88480 }, { "epoch": 2.0, "grad_norm": 1.7715900647041394, "learning_rate": 2.6141390331413385e-10, "loss": 2.243, "step": 88490 }, { "epoch": 2.0, "grad_norm": 1.8090658411454812, "learning_rate": 2.3567536379554e-10, "loss": 2.3486, "step": 88500 }, { "epoch": 2.0, "grad_norm": 1.6708937818216285, "learning_rate": 2.1127040696566903e-10, "loss": 2.3569, "step": 88510 }, { "epoch": 2.0, "grad_norm": 1.7240582002409444, "learning_rate": 1.8819903607747435e-10, "loss": 2.2918, "step": 88520 }, { "epoch": 2.0, "grad_norm": 1.6505565917365816, "learning_rate": 1.6646125420849423e-10, "loss": 2.4559, "step": 88530 }, { "epoch": 2.0, "grad_norm": 1.775733235448224, "learning_rate": 1.460570642586312e-10, "loss": 2.3606, "step": 88540 }, { "epoch": 2.0, "grad_norm": 1.7177827258875766, "learning_rate": 1.2698646894793166e-10, "loss": 2.2779, "step": 88550 }, { "epoch": 2.0, "grad_norm": 1.8468548072349973, "learning_rate": 1.0924947081991655e-10, "loss": 2.3879, "step": 88560 }, { "epoch": 2.0, "grad_norm": 1.6282362562468533, "learning_rate": 9.284607223936093e-11, "loss": 2.4059, "step": 88570 }, { "epoch": 2.0, "grad_norm": 1.7075260188487722, "learning_rate": 7.777627539451437e-11, "loss": 2.3652, "step": 88580 }, { "epoch": 2.0, "grad_norm": 1.5085646129609669, "learning_rate": 6.404008229599079e-11, "loss": 2.3848, "step": 88590 }, { "epoch": 2.0, "grad_norm": 1.6768041297973304, "learning_rate": 5.1637494774547934e-11, "loss": 2.2452, "step": 88600 }, { "epoch": 2.0, "grad_norm": 1.8087929306297494, "learning_rate": 4.056851448441812e-11, "loss": 2.3903, "step": 88610 }, { "epoch": 2.0, "grad_norm": 1.7021501989082424, "learning_rate": 3.083314290108774e-11, "loss": 2.3622, "step": 88620 }, { "epoch": 2.0, "grad_norm": 1.87453838616786, "learning_rate": 2.2431381324627965e-11, "loss": 2.3664, "step": 88630 }, { "epoch": 2.0, "grad_norm": 1.6642416942065026, "learning_rate": 1.5363230874143597e-11, "loss": 2.3548, "step": 88640 }, { "epoch": 2.0, "grad_norm": 1.6058788819357883, "learning_rate": 9.628692492213986e-12, "loss": 2.4296, "step": 88650 }, { "epoch": 2.0, "grad_norm": 1.8069033210519239, "learning_rate": 5.22776694489302e-12, "loss": 2.2742, "step": 88660 }, { "epoch": 2.0, "grad_norm": 1.7668386634522502, "learning_rate": 2.1604548172682314e-12, "loss": 2.3611, "step": 88670 }, { "epoch": 2.0, "grad_norm": 1.792240302068036, "learning_rate": 4.2675651901191717e-13, "loss": 2.3578, "step": 88680 }, { "epoch": 2.0, "step": 88688, "total_flos": 3.272007336853504e+16, "train_loss": 2.5470479068699325, "train_runtime": 297591.5912, "train_samples_per_second": 76.293, "train_steps_per_second": 0.298 } ], "logging_steps": 10, "max_steps": 88688, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "total_flos": 3.272007336853504e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }